diff --git a/.bcc.metadata b/.bcc.metadata index 7891d2b..b4704b4 100644 --- a/.bcc.metadata +++ b/.bcc.metadata @@ -1 +1 @@ -7c555629ea7bbd69fa22827076c11113cf2af104 SOURCES/bcc-0.26.0.tar.gz +8ce0ccb0724da475f127d62acc10a88569956474 SOURCES/bcc-0.28.0.tar.gz diff --git a/.gitignore b/.gitignore index a679cb4..5ea7f97 100644 --- a/.gitignore +++ b/.gitignore @@ -1 +1 @@ -SOURCES/bcc-0.26.0.tar.gz +SOURCES/bcc-0.28.0.tar.gz diff --git a/SOURCES/bcc-0.26.0-killsnoop-add-missing-s-and-T-options-to-the-synopsi.patch b/SOURCES/bcc-0.26.0-killsnoop-add-missing-s-and-T-options-to-the-synopsi.patch deleted file mode 100644 index 79edaf2..0000000 --- a/SOURCES/bcc-0.26.0-killsnoop-add-missing-s-and-T-options-to-the-synopsi.patch +++ /dev/null @@ -1,58 +0,0 @@ -From 64f9c355a62f78000270d025b479b7eeba7349e9 Mon Sep 17 00:00:00 2001 -From: Jerome Marchand -Date: Wed, 11 Jan 2023 16:46:32 +0100 -Subject: [PATCH] killsnoop: add missing -s and -T options to the synopsis - -The -s option is missing from the synopsis of the killsnoop manpage, -example file and the comment on top of the tool itself. -Also, -T option is missing from the example file. - -Signed-off-by: Jerome Marchand ---- - man/man8/killsnoop.8 | 2 +- - tools/killsnoop.py | 2 +- - tools/killsnoop_example.txt | 2 +- - 3 files changed, 3 insertions(+), 3 deletions(-) - -diff --git a/man/man8/killsnoop.8 b/man/man8/killsnoop.8 -index 3f63d2ee..cb2a975e 100644 ---- a/man/man8/killsnoop.8 -+++ b/man/man8/killsnoop.8 -@@ -2,7 +2,7 @@ - .SH NAME - killsnoop \- Trace signals issued by the kill() syscall. Uses Linux eBPF/bcc. - .SH SYNOPSIS --.B killsnoop [\-h] [\-x] [-p PID] [-T PID] -+.B killsnoop [\-h] [\-x] [-p PID] [-T PID] [-s SIGNAL] - .SH DESCRIPTION - killsnoop traces the kill() syscall, to show signals sent via this method. This - may be useful to troubleshoot failing applications, where an unknown mechanism -diff --git a/tools/killsnoop.py b/tools/killsnoop.py -index c0166f1d..9cce8dcc 100755 ---- a/tools/killsnoop.py -+++ b/tools/killsnoop.py -@@ -4,7 +4,7 @@ - # killsnoop Trace signals issued by the kill() syscall. - # For Linux, uses BCC, eBPF. Embedded C. - # --# USAGE: killsnoop [-h] [-x] [-p PID] [-T PID] -+# USAGE: killsnoop [-h] [-x] [-p PID] [-T PID] [-s SIGNAL] - # - # Copyright (c) 2015 Brendan Gregg. - # Licensed under the Apache License, Version 2.0 (the "License") -diff --git a/tools/killsnoop_example.txt b/tools/killsnoop_example.txt -index 904fe6ef..97c3ad70 100644 ---- a/tools/killsnoop_example.txt -+++ b/tools/killsnoop_example.txt -@@ -19,7 +19,7 @@ The second line showed the same signal sent, this time resulting in a -3 - USAGE message: - - # ./killsnoop -h --usage: killsnoop [-h] [-x] [-p PID] -+usage: killsnoop [-h] [-x] [-p PID] [-T PID] [-s SIGNAL] - - Trace signals issued by the kill() syscall - --- -2.39.2 - diff --git a/SOURCES/bcc-0.26.0-sync-with-latest-libbpf-repo.patch b/SOURCES/bcc-0.26.0-sync-with-latest-libbpf-repo.patch deleted file mode 100644 index 3f77c5e..0000000 --- a/SOURCES/bcc-0.26.0-sync-with-latest-libbpf-repo.patch +++ /dev/null @@ -1,363 +0,0 @@ -From 34f77c4aaaa039fd2ef3d51b8b61db30fc34912f Mon Sep 17 00:00:00 2001 -From: Yonghong Song -Date: Wed, 19 Apr 2023 23:46:53 -0700 -Subject: [PATCH] sync with latest libbpf repo - -Sync libbpf submodule up to the following commit: - 44b0bc9ad70a ci: Regenerate latest vmlinux.h for old kernel CI tests. - -Signed-off-by: Yonghong Song ---- - src/cc/compat/linux/virtual_bpf.h | 141 ++++++++++++++++++++++++++---- - src/cc/export/helpers.h | 6 +- - 2 files changed, 126 insertions(+), 21 deletions(-) - -diff --git a/src/cc/compat/linux/virtual_bpf.h b/src/cc/compat/linux/virtual_bpf.h -index be3a4627..a182123e 100644 ---- a/src/cc/compat/linux/virtual_bpf.h -+++ b/src/cc/compat/linux/virtual_bpf.h -@@ -1034,6 +1034,7 @@ enum bpf_attach_type { - BPF_PERF_EVENT, - BPF_TRACE_KPROBE_MULTI, - BPF_LSM_CGROUP, -+ BPF_STRUCT_OPS, - __MAX_BPF_ATTACH_TYPE - }; - -@@ -1109,7 +1110,7 @@ enum bpf_link_type { - */ - #define BPF_F_STRICT_ALIGNMENT (1U << 0) - --/* If BPF_F_ANY_ALIGNMENT is used in BPF_PROF_LOAD command, the -+/* If BPF_F_ANY_ALIGNMENT is used in BPF_PROG_LOAD command, the - * verifier will allow any alignment whatsoever. On platforms - * with strict alignment requirements for loads ands stores (such - * as sparc and mips) the verifier validates that all loads and -@@ -1157,6 +1158,11 @@ enum bpf_link_type { - */ - #define BPF_F_XDP_HAS_FRAGS (1U << 5) - -+/* If BPF_F_XDP_DEV_BOUND_ONLY is used in BPF_PROG_LOAD command, the loaded -+ * program becomes device-bound but can access XDP metadata. -+ */ -+#define BPF_F_XDP_DEV_BOUND_ONLY (1U << 6) -+ - /* link_create.kprobe_multi.flags used in LINK_CREATE command for - * BPF_TRACE_KPROBE_MULTI attach type to create return probe. - */ -@@ -1262,6 +1268,9 @@ enum { - - /* Create a map that is suitable to be an inner map with dynamic max entries */ - BPF_F_INNER_MAP = (1U << 12), -+ -+/* Create a map that will be registered/unregesitered by the backed bpf_link */ -+ BPF_F_LINK = (1U << 13), - }; - - /* Flags for BPF_PROG_QUERY. */ -@@ -1399,6 +1408,11 @@ union bpf_attr { - __aligned_u64 fd_array; /* array of FDs */ - __aligned_u64 core_relos; - __u32 core_relo_rec_size; /* sizeof(struct bpf_core_relo) */ -+ /* output: actual total log contents size (including termintaing zero). -+ * It could be both larger than original log_size (if log was -+ * truncated), or smaller (if log buffer wasn't filled completely). -+ */ -+ __u32 log_true_size; - }; - - struct { /* anonymous struct used by BPF_OBJ_* commands */ -@@ -1484,6 +1498,11 @@ union bpf_attr { - __u32 btf_size; - __u32 btf_log_size; - __u32 btf_log_level; -+ /* output: actual total log contents size (including termintaing zero). -+ * It could be both larger than original log_size (if log was -+ * truncated), or smaller (if log buffer wasn't filled completely). -+ */ -+ __u32 btf_log_true_size; - }; - - struct { -@@ -1503,7 +1522,10 @@ union bpf_attr { - } task_fd_query; - - struct { /* struct used by BPF_LINK_CREATE command */ -- __u32 prog_fd; /* eBPF program to attach */ -+ union { -+ __u32 prog_fd; /* eBPF program to attach */ -+ __u32 map_fd; /* struct_ops to attach */ -+ }; - union { - __u32 target_fd; /* object to attach to */ - __u32 target_ifindex; /* target ifindex */ -@@ -1544,12 +1566,23 @@ union bpf_attr { - - struct { /* struct used by BPF_LINK_UPDATE command */ - __u32 link_fd; /* link fd */ -- /* new program fd to update link with */ -- __u32 new_prog_fd; -+ union { -+ /* new program fd to update link with */ -+ __u32 new_prog_fd; -+ /* new struct_ops map fd to update link with */ -+ __u32 new_map_fd; -+ }; - __u32 flags; /* extra flags */ -- /* expected link's program fd; is specified only if -- * BPF_F_REPLACE flag is set in flags */ -- __u32 old_prog_fd; -+ union { -+ /* expected link's program fd; is specified only if -+ * BPF_F_REPLACE flag is set in flags. -+ */ -+ __u32 old_prog_fd; -+ /* expected link's map fd; is specified only -+ * if BPF_F_REPLACE flag is set. -+ */ -+ __u32 old_map_fd; -+ }; - } link_update; - - struct { -@@ -1643,17 +1676,17 @@ union bpf_attr { - * Description - * This helper is a "printk()-like" facility for debugging. It - * prints a message defined by format *fmt* (of size *fmt_size*) -- * to file *\/sys/kernel/debug/tracing/trace* from DebugFS, if -+ * to file *\/sys/kernel/tracing/trace* from TraceFS, if - * available. It can take up to three additional **u64** - * arguments (as an eBPF helpers, the total number of arguments is - * limited to five). - * - * Each time the helper is called, it appends a line to the trace. -- * Lines are discarded while *\/sys/kernel/debug/tracing/trace* is -- * open, use *\/sys/kernel/debug/tracing/trace_pipe* to avoid this. -+ * Lines are discarded while *\/sys/kernel/tracing/trace* is -+ * open, use *\/sys/kernel/tracing/trace_pipe* to avoid this. - * The format of the trace is customizable, and the exact output - * one will get depends on the options set in -- * *\/sys/kernel/debug/tracing/trace_options* (see also the -+ * *\/sys/kernel/tracing/trace_options* (see also the - * *README* file under the same directory). However, it usually - * defaults to something like: - * -@@ -2002,6 +2035,9 @@ union bpf_attr { - * sending the packet. This flag was added for GRE - * encapsulation, but might be used with other protocols - * as well in the future. -+ * **BPF_F_NO_TUNNEL_KEY** -+ * Add a flag to tunnel metadata indicating that no tunnel -+ * key should be set in the resulting tunnel header. - * - * Here is a typical usage on the transmit path: - * -@@ -2645,6 +2681,11 @@ union bpf_attr { - * Use with BPF_F_ADJ_ROOM_ENCAP_L2 flag to further specify the - * L2 type as Ethernet. - * -+ * * **BPF_F_ADJ_ROOM_DECAP_L3_IPV4**, -+ * **BPF_F_ADJ_ROOM_DECAP_L3_IPV6**: -+ * Indicate the new IP header version after decapsulating the outer -+ * IP header. Used when the inner and outer IP versions are different. -+ * - * A call to this helper is susceptible to change the underlying - * packet buffer. Therefore, at load time, all checks on pointers - * previously done by the verifier are invalidated and must be -@@ -2789,7 +2830,7 @@ union bpf_attr { - * - * long bpf_perf_prog_read_value(struct bpf_perf_event_data *ctx, struct bpf_perf_event_value *buf, u32 buf_size) - * Description -- * For en eBPF program attached to a perf event, retrieve the -+ * For an eBPF program attached to a perf event, retrieve the - * value of the event counter associated to *ctx* and store it in - * the structure pointed by *buf* and of size *buf_size*. Enabled - * and running times are also stored in the structure (see -@@ -3122,6 +3163,11 @@ union bpf_attr { - * **BPF_FIB_LOOKUP_OUTPUT** - * Perform lookup from an egress perspective (default is - * ingress). -+ * **BPF_FIB_LOOKUP_SKIP_NEIGH** -+ * Skip the neighbour table lookup. *params*->dmac -+ * and *params*->smac will not be set as output. A common -+ * use case is to call **bpf_redirect_neigh**\ () after -+ * doing **bpf_fib_lookup**\ (). - * - * *ctx* is either **struct xdp_md** for XDP programs or - * **struct sk_buff** tc cls_act programs. -@@ -4952,6 +4998,12 @@ union bpf_attr { - * different maps if key/value layout matches across maps. - * Every bpf_timer_set_callback() can have different callback_fn. - * -+ * *flags* can be one of: -+ * -+ * **BPF_F_TIMER_ABS** -+ * Start the timer in absolute expire value instead of the -+ * default relative one. -+ * - * Return - * 0 on success. - * **-EINVAL** if *timer* was not initialized with bpf_timer_init() earlier -@@ -5294,7 +5346,7 @@ union bpf_attr { - * Return - * Nothing. Always succeeds. - * -- * long bpf_dynptr_read(void *dst, u32 len, struct bpf_dynptr *src, u32 offset, u64 flags) -+ * long bpf_dynptr_read(void *dst, u32 len, const struct bpf_dynptr *src, u32 offset, u64 flags) - * Description - * Read *len* bytes from *src* into *dst*, starting from *offset* - * into *src*. -@@ -5304,22 +5356,36 @@ union bpf_attr { - * of *src*'s data, -EINVAL if *src* is an invalid dynptr or if - * *flags* is not 0. - * -- * long bpf_dynptr_write(struct bpf_dynptr *dst, u32 offset, void *src, u32 len, u64 flags) -+ * long bpf_dynptr_write(const struct bpf_dynptr *dst, u32 offset, void *src, u32 len, u64 flags) - * Description - * Write *len* bytes from *src* into *dst*, starting from *offset* - * into *dst*. -- * *flags* is currently unused. -+ * -+ * *flags* must be 0 except for skb-type dynptrs. -+ * -+ * For skb-type dynptrs: -+ * * All data slices of the dynptr are automatically -+ * invalidated after **bpf_dynptr_write**\ (). This is -+ * because writing may pull the skb and change the -+ * underlying packet buffer. -+ * -+ * * For *flags*, please see the flags accepted by -+ * **bpf_skb_store_bytes**\ (). - * Return - * 0 on success, -E2BIG if *offset* + *len* exceeds the length - * of *dst*'s data, -EINVAL if *dst* is an invalid dynptr or if *dst* -- * is a read-only dynptr or if *flags* is not 0. -+ * is a read-only dynptr or if *flags* is not correct. For skb-type dynptrs, -+ * other errors correspond to errors returned by **bpf_skb_store_bytes**\ (). - * -- * void *bpf_dynptr_data(struct bpf_dynptr *ptr, u32 offset, u32 len) -+ * void *bpf_dynptr_data(const struct bpf_dynptr *ptr, u32 offset, u32 len) - * Description - * Get a pointer to the underlying dynptr data. - * - * *len* must be a statically known value. The returned data slice - * is invalidated whenever the dynptr is invalidated. -+ * -+ * skb and xdp type dynptrs may not use bpf_dynptr_data. They should -+ * instead use bpf_dynptr_slice and bpf_dynptr_slice_rdwr. - * Return - * Pointer to the underlying dynptr data, NULL if the dynptr is - * read-only, if the dynptr is invalid, or if the offset and length -@@ -5415,7 +5481,7 @@ union bpf_attr { - * Drain samples from the specified user ring buffer, and invoke - * the provided callback for each such sample: - * -- * long (\*callback_fn)(struct bpf_dynptr \*dynptr, void \*ctx); -+ * long (\*callback_fn)(const struct bpf_dynptr \*dynptr, void \*ctx); - * - * If **callback_fn** returns 0, the helper will continue to try - * and drain the next sample, up to a maximum of -@@ -5765,6 +5831,7 @@ enum { - BPF_F_ZERO_CSUM_TX = (1ULL << 1), - BPF_F_DONT_FRAGMENT = (1ULL << 2), - BPF_F_SEQ_NUMBER = (1ULL << 3), -+ BPF_F_NO_TUNNEL_KEY = (1ULL << 4), - }; - - /* BPF_FUNC_skb_get_tunnel_key flags. */ -@@ -5804,6 +5871,8 @@ enum { - BPF_F_ADJ_ROOM_ENCAP_L4_UDP = (1ULL << 4), - BPF_F_ADJ_ROOM_NO_CSUM_RESET = (1ULL << 5), - BPF_F_ADJ_ROOM_ENCAP_L2_ETH = (1ULL << 6), -+ BPF_F_ADJ_ROOM_DECAP_L3_IPV4 = (1ULL << 7), -+ BPF_F_ADJ_ROOM_DECAP_L3_IPV6 = (1ULL << 8), - }; - - enum { -@@ -6339,6 +6408,9 @@ struct bpf_link_info { - struct { - __u32 ifindex; - } xdp; -+ struct { -+ __u32 map_id; -+ } struct_ops; - }; - } __attribute__((aligned(8))); - -@@ -6735,6 +6807,7 @@ struct bpf_raw_tracepoint_args { - enum { - BPF_FIB_LOOKUP_DIRECT = (1U << 0), - BPF_FIB_LOOKUP_OUTPUT = (1U << 1), -+ BPF_FIB_LOOKUP_SKIP_NEIGH = (1U << 2), - }; - - enum { -@@ -6902,6 +6975,21 @@ struct bpf_list_node { - __u64 :64; - } __attribute__((aligned(8))); - -+struct bpf_rb_root { -+ __u64 :64; -+ __u64 :64; -+} __attribute__((aligned(8))); -+ -+struct bpf_rb_node { -+ __u64 :64; -+ __u64 :64; -+ __u64 :64; -+} __attribute__((aligned(8))); -+ -+struct bpf_refcount { -+ __u32 :32; -+} __attribute__((aligned(4))); -+ - struct bpf_sysctl { - __u32 write; /* Sysctl is being read (= 0) or written (= 1). - * Allows 1,2,4-byte read, but no write. -@@ -7051,5 +7139,22 @@ struct bpf_core_relo { - enum bpf_core_relo_kind kind; - }; - -+/* -+ * Flags to control bpf_timer_start() behaviour. -+ * - BPF_F_TIMER_ABS: Timeout passed is absolute time, by default it is -+ * relative to current time. -+ */ -+enum { -+ BPF_F_TIMER_ABS = (1ULL << 0), -+}; -+ -+/* BPF numbers iterator state */ -+struct bpf_iter_num { -+ /* opaque iterator state; having __u64 here allows to preserve correct -+ * alignment requirements in vmlinux.h, generated from BTF -+ */ -+ __u64 __opaque[1]; -+} __attribute__((aligned(8))); -+ - #endif /* _UAPI__LINUX_BPF_H__ */ - )********" -diff --git a/src/cc/export/helpers.h b/src/cc/export/helpers.h -index d7b869e0..e989440a 100644 ---- a/src/cc/export/helpers.h -+++ b/src/cc/export/helpers.h -@@ -1006,13 +1006,13 @@ static void (*bpf_ringbuf_submit_dynptr)(struct bpf_dynptr *ptr, __u64 flags) = - (void *)BPF_FUNC_ringbuf_submit_dynptr; - static void (*bpf_ringbuf_discard_dynptr)(struct bpf_dynptr *ptr, __u64 flags) = - (void *)BPF_FUNC_ringbuf_discard_dynptr; --static long (*bpf_dynptr_read)(void *dst, __u32 len, struct bpf_dynptr *src, __u32 offset, -+static long (*bpf_dynptr_read)(void *dst, __u32 len, const struct bpf_dynptr *src, __u32 offset, - __u64 flags) = - (void *)BPF_FUNC_dynptr_read; --static long (*bpf_dynptr_write)(struct bpf_dynptr *dst, __u32 offset, void *src, __u32 len, -+static long (*bpf_dynptr_write)(const struct bpf_dynptr *dst, __u32 offset, void *src, __u32 len, - __u64 flags) = - (void *)BPF_FUNC_dynptr_write; --static void *(*bpf_dynptr_data)(struct bpf_dynptr *ptr, __u32 offset, __u32 len) = -+static void *(*bpf_dynptr_data)(const struct bpf_dynptr *ptr, __u32 offset, __u32 len) = - (void *)BPF_FUNC_dynptr_data; - static __s64 (*bpf_tcp_raw_gen_syncookie_ipv4)(struct iphdr *iph, struct tcphdr *th, - __u32 th_len) = --- -2.41.0 - diff --git a/SOURCES/bcc-0.26.0-tools-compactsnoop.py-Fix-raw_tracepoint-Invalid-arg.patch b/SOURCES/bcc-0.26.0-tools-compactsnoop.py-Fix-raw_tracepoint-Invalid-arg.patch deleted file mode 100644 index ab540be..0000000 --- a/SOURCES/bcc-0.26.0-tools-compactsnoop.py-Fix-raw_tracepoint-Invalid-arg.patch +++ /dev/null @@ -1,71 +0,0 @@ -From cc35f70515cb0f3b8032b8fb68f9f37a844e74c8 Mon Sep 17 00:00:00 2001 -From: Rong Tao -Date: Fri, 10 Feb 2023 23:28:55 +0800 -Subject: [PATCH] tools/compactsnoop.py: Fix raw_tracepoint Invalid argument - error - -kernel commit abd4349ff9b8("mm: compaction: cleanup the compaction trace -events") change the arguments of 'mm_compaction_begin' from (start_pfn, -migrate_pfn, free_pfn, end_pfn, sync) to (cc, start_pfn, end_pfn, sync), -and change the arguments of 'mm_compaction_end' from (start_pfn, -migrate_pfn, free_pfn, end_pfn, sync, ret) to (cc, start_pfn, end_pfn, -sync, ret). - -Replacing RAW_TRACEPOINT_PROBE with TRACEPOINT_PROBE solves this problem -and guarantees compatibility. - - $ sudo ./compactsnoop.py - bpf_attach_raw_tracepoint (mm_compaction_begin): Invalid argument - Traceback (most recent call last): - File "/home/sdb/Git/bcc/tools/./compactsnoop.py", line 292, in - b = BPF(text=bpf_text) - ^^^^^^^^^^^^^^^^^^ - File "/usr/lib/python3.11/site-packages/bcc/__init__.py", line 483, in __init__ - self._trace_autoload() - File "/usr/lib/python3.11/site-packages/bcc/__init__.py", line 1462, in _trace_autoload - self.attach_raw_tracepoint(tp=tp, fn_name=fn.name) - File "/usr/lib/python3.11/site-packages/bcc/__init__.py", line 1055, in attach_raw_tracepoint - raise Exception("Failed to attach BPF to raw tracepoint") - Exception: Failed to attach BPF to raw tracepoint - -Signed-off-by: Rong Tao ---- - tools/compactsnoop.py | 13 ++++--------- - 1 file changed, 4 insertions(+), 9 deletions(-) - -diff --git a/tools/compactsnoop.py b/tools/compactsnoop.py -index 2643e8ed..2b395dec 100755 ---- a/tools/compactsnoop.py -+++ b/tools/compactsnoop.py -@@ -237,11 +237,9 @@ RAW_TRACEPOINT_PROBE(mm_compaction_suitable) - return 0; - } - --RAW_TRACEPOINT_PROBE(mm_compaction_begin) -+TRACEPOINT_PROBE(compaction, mm_compaction_begin) - { -- // TP_PROTO(unsigned long zone_start, unsigned long migrate_pfn, -- // unsigned long free_pfn, unsigned long zone_end, bool sync) -- bool sync = (bool)ctx->args[4]; -+ bool sync = args->sync; - - u64 id = bpf_get_current_pid_tgid(); - struct val_t *valp = start.lookup(&id); -@@ -255,12 +253,9 @@ RAW_TRACEPOINT_PROBE(mm_compaction_begin) - return 0; - } - --RAW_TRACEPOINT_PROBE(mm_compaction_end) -+TRACEPOINT_PROBE(compaction, mm_compaction_end) - { -- // TP_PROTO(unsigned long zone_start, unsigned long migrate_pfn, -- // unsigned long free_pfn, unsigned long zone_end, bool sync, -- // int status) -- submit_event(ctx, ctx->args[5]); -+ submit_event(args, args->status); - return 0; - } - """ --- -2.39.2 - diff --git a/SOURCES/bcc-0.26.0-tools-deadlock-Add-an-option-to-set-the-maximum-numb.patch b/SOURCES/bcc-0.26.0-tools-deadlock-Add-an-option-to-set-the-maximum-numb.patch deleted file mode 100644 index 70410ac..0000000 --- a/SOURCES/bcc-0.26.0-tools-deadlock-Add-an-option-to-set-the-maximum-numb.patch +++ /dev/null @@ -1,76 +0,0 @@ -From 0e9384ec4c88d2da2d23475f58ec9bff7eb48639 Mon Sep 17 00:00:00 2001 -From: Jerome Marchand -Date: Tue, 25 Apr 2023 16:04:05 +0200 -Subject: [PATCH] tools/deadlock: Add an option to set the maximum number of - stack traces - -Commit 77f5252d ("tools/deadlock: support specifies maxnum of threads -and edge cases (#3455)") allow to set the maximum number of threads -and edge cases to be able to reduce the memory usage of the deadlock -tool. It however let the size of the map of stack traces fixed. It's -current size, 640k (actually rounded up to 1M) takes 1Gb of vmalloced -kernel memory. - -This patch adds an option to make the maximum number of stack traces -user defined. It also set the default value to 64k, in line with the -current default for the number of edge cases and threads. - -It fix the following issue on system with limited memory ressources: -could not open bpf map: stack_traces, error: Cannot allocate memory -Traceback (most recent call last): - File "/tmp/./deadlock.py", line 577, in - main() - File "/tmp/./deadlock.py", line 489, in main - bpf = BPF(text=text) - File "/usr/lib/python3.9/site-packages/bcc/__init__.py", line 479, in __init__ - raise Exception("Failed to compile BPF module %s" % (src_file or "")) -Exception: Failed to compile BPF module - -Signed-off-by: Jerome Marchand ---- - tools/deadlock.c | 2 +- - tools/deadlock.py | 8 ++++++++ - 2 files changed, 9 insertions(+), 1 deletion(-) - -diff --git a/tools/deadlock.c b/tools/deadlock.c -index 006dc121..6ae405ba 100644 ---- a/tools/deadlock.c -+++ b/tools/deadlock.c -@@ -60,7 +60,7 @@ struct thread_created_leaf_t { - BPF_HASH(thread_to_parent, u32, struct thread_created_leaf_t); - - // Stack traces when threads are created and when mutexes are locked/unlocked. --BPF_STACK_TRACE(stack_traces, 655360); -+BPF_STACK_TRACE(stack_traces, MAX_TRACES); - - // The first argument to the user space function we are tracing - // is a pointer to the mutex M held by thread T. -diff --git a/tools/deadlock.py b/tools/deadlock.py -index 12de099f..f7eb4ce0 100755 ---- a/tools/deadlock.py -+++ b/tools/deadlock.py -@@ -467,6 +467,13 @@ import time - help='Specifies the maximum number of edge cases that can be recorded. ' - 'default 65536. Note. 88 bytes per edge case.' - ) -+ parser.add_argument( -+ '-s', '--stacktraces', type=int, default=65536, -+ help='Specifies the maximum number of stack traces that can be recorded. ' -+ 'This number is rounded up to the next power of two.' -+ 'default 65536. Note. 1 kbytes vmalloced per stack trace.' -+ ) -+ - args = parser.parse_args() - if not args.binary: - try: -@@ -479,6 +486,7 @@ import time - text = f.read() - text = text.replace('MAX_THREADS', str(args.threads)); - text = text.replace('MAX_EDGES', str(args.edges)); -+ text = text.replace('MAX_TRACES', str(args.stacktraces)); - bpf = BPF(text=text) - - # Trace where threads are created --- -2.39.2 - diff --git a/SOURCES/bcc-0.26.0-tools-funcslower-fix-printing-of-folded-stacks.patch b/SOURCES/bcc-0.26.0-tools-funcslower-fix-printing-of-folded-stacks.patch deleted file mode 100644 index 1569468..0000000 --- a/SOURCES/bcc-0.26.0-tools-funcslower-fix-printing-of-folded-stacks.patch +++ /dev/null @@ -1,57 +0,0 @@ -From 29f0fa3693d679102680fece9ed5e606e291c5fa Mon Sep 17 00:00:00 2001 -From: Jerome Marchand -Date: Fri, 7 Apr 2023 14:30:54 +0200 -Subject: [PATCH] tools/funcslower: fix printing of folded stacks - -When trying to print folded stack, funcslower tries to join bytes to a -string. Let's perform that operation with bytes only, and decode -before printing. -Also, decode symbols name before printing for the default stack -format, to avoid unsightly b'xxx' output. - -It fixes the following error: - -Exception ignored on calling ctypes callback function: .raw_cb_ at 0x7f200541e5e0> -Traceback (most recent call last): - File "/usr/lib/python3.9/site-packages/bcc/table.py", line 982, in raw_cb_ - callback(cpu, data, size) - File "/usr/share/bcc/tools/funcslower", line 340, in print_event - print_stack(event) - File "/usr/share/bcc/tools/funcslower", line 324, in print_stack - print("%s %d" % (";".join(line), 1)) -TypeError: sequence item 1: expected str instance, bytes found - -Signed-off-by: Jerome Marchand ---- - tools/funcslower.py | 8 ++++---- - 1 file changed, 4 insertions(+), 4 deletions(-) - -diff --git a/tools/funcslower.py b/tools/funcslower.py -index 6df7f24c..4b3798a0 100755 ---- a/tools/funcslower.py -+++ b/tools/funcslower.py -@@ -317,17 +317,17 @@ earliest_ts = 0 - # print folded stack output - user_stack = list(user_stack) - kernel_stack = list(kernel_stack) -- line = [event.comm.decode('utf-8', 'replace')] + \ -+ line = [event.comm] + \ - [b.sym(addr, event.tgid_pid) for addr in reversed(user_stack)] + \ - (do_delimiter and ["-"] or []) + \ - [b.ksym(addr) for addr in reversed(kernel_stack)] -- print("%s %d" % (";".join(line), 1)) -+ print("%s %d" % (b';'.join(line).decode('utf-8', 'replace'), 1)) - else: - # print default multi-line stack output. - for addr in kernel_stack: -- print(" %s" % b.ksym(addr)) -+ print(" %s" % b.ksym(addr).decode('utf-8', 'replace')) - for addr in user_stack: -- print(" %s" % b.sym(addr, event.tgid_pid)) -+ print(" %s" % b.sym(addr, event.tgid_pid).decode('utf-8', 'replace')) - - def print_event(cpu, data, size): - event = b["events"].event(data) --- -2.39.2 - diff --git a/SOURCES/bcc-0.26.0-tools-nfsslower.py-Fix-uninitialized-struct-pad-erro.patch b/SOURCES/bcc-0.26.0-tools-nfsslower.py-Fix-uninitialized-struct-pad-erro.patch deleted file mode 100644 index 863b89c..0000000 --- a/SOURCES/bcc-0.26.0-tools-nfsslower.py-Fix-uninitialized-struct-pad-erro.patch +++ /dev/null @@ -1,85 +0,0 @@ -From 9965f8397950d8aa1bc1a5decbc2250d0627a798 Mon Sep 17 00:00:00 2001 -From: Rong Tao -Date: Fri, 10 Feb 2023 22:16:56 +0800 -Subject: [PATCH] tools/nfsslower.py: Fix uninitialized struct pad error - -The verifier is unhappy, if data struct _pad_ is not initialized, see [0][1]. - - $ sudo ./nfsslower.py - ... - ; bpf_perf_event_output(ctx, (void *)bpf_pseudo_fd(1, -2), CUR_CPU_IDENTIFIER, &data, sizeof(data)); - 83: (79) r1 = *(u64 *)(r10 -144) ; R1_w=ctx(off=0,imm=0) R10=fp0 - 84: (18) r3 = 0xffffffff ; R3_w=4294967295 - 86: (b7) r5 = 96 ; R5_w=96 - 87: (85) call bpf_perf_event_output#25 - invalid indirect read from stack R4 off -136+92 size 96 - processed 84 insns (limit 1000000) max_states_per_insn 0 total_states 4 peak_states 4 mark_read 4 - ... - raise Exception("Failed to load BPF program %s: %s" % - Exception: Failed to load BPF program b'raw_tracepoint__nfs_commit_done': Permission denied - -[0] https://github.com/iovisor/bcc/issues/2623 -[1] https://github.com/iovisor/bcc/pull/4453 - -Signed-off-by: Rong Tao ---- - tools/nfsslower.py | 29 +++++++++++++++++++++-------- - 1 file changed, 21 insertions(+), 8 deletions(-) - -diff --git a/tools/nfsslower.py b/tools/nfsslower.py -index 34756f72..99f63f0f 100755 ---- a/tools/nfsslower.py -+++ b/tools/nfsslower.py -@@ -195,8 +195,11 @@ static int trace_exit(struct pt_regs *ctx, int type) - - // populate output struct - u32 size = PT_REGS_RC(ctx); -- struct data_t data = {.type = type, .size = size, .delta_us = delta_us, -- .pid = pid}; -+ struct data_t data = {}; -+ data.type = type; -+ data.size = size; -+ data.delta_us = delta_us; -+ data.pid = pid; - data.ts_us = ts / 1000; - data.offset = valp->offset; - bpf_get_current_comm(&data.task, sizeof(data.task)); -@@ -280,9 +283,14 @@ RAW_TRACEPOINT_PROBE(nfs_commit_done) - u64 ts = bpf_ktime_get_ns(); - u64 delta_us = (ts - cp->ts) / 1000; - u32 pid = bpf_get_current_pid_tgid() >> 32; -- struct data_t data = {.type = TRACE_COMMIT, .offset = cp->offset, -- .size = cp->count, .ts_us = ts/1000, .delta_us = delta_us, -- .pid = pid}; -+ -+ struct data_t data = {}; -+ data.type = TRACE_COMMIT; -+ data.offset = cp->offset; -+ data.size = cp->count; -+ data.ts_us = ts/1000; -+ data.delta_us = delta_us; -+ data.pid = pid; - - commitinfo.delete(&key); - bpf_get_current_comm(&data.task, sizeof(data.task)); -@@ -325,9 +333,14 @@ int trace_nfs_commit_done(struct pt_regs *ctx, void *task, void *calldata) - u64 ts = bpf_ktime_get_ns(); - u64 delta_us = (ts - cp->ts) / 1000; - u32 pid = bpf_get_current_pid_tgid() >> 32; -- struct data_t data = {.type = TRACE_COMMIT, .offset = cp->offset, -- .size = cp->count, .ts_us = ts/1000, .delta_us = delta_us, -- .pid = pid}; -+ -+ struct data_t data = {}; -+ data.type = TRACE_COMMIT; -+ data.offset = cp->offset; -+ data.size = cp->count; -+ data.ts_us = ts/1000; -+ data.delta_us = delta_us; -+ data.pid = pid; - - commitinfo.delete(&key); - bpf_get_current_comm(&data.task, sizeof(data.task)); --- -2.39.2 - diff --git a/SOURCES/bcc-0.26.0-tools-readahead-Fix-Failed-to-attach-BPF-program-ent.patch b/SOURCES/bcc-0.26.0-tools-readahead-Fix-Failed-to-attach-BPF-program-ent.patch deleted file mode 100644 index 1f1c0be..0000000 --- a/SOURCES/bcc-0.26.0-tools-readahead-Fix-Failed-to-attach-BPF-program-ent.patch +++ /dev/null @@ -1,159 +0,0 @@ -From 02fce045ce02fe81d8649ce63ce81d5cdf3e3a72 Mon Sep 17 00:00:00 2001 -From: Rong Tao -Date: Mon, 30 Jan 2023 17:39:35 +0800 -Subject: [PATCH] tools/readahead: Fix: Failed to attach BPF program - entry__do_page_cache_readahead - -since commit 56a4d67c264e("mm/readahead: Switch to page_cache_ra_order") switch -do_page_cache_ra() to page_cache_ra_order() (v5.17), and commit bb3c579e25e5 -("mm/filemap: Add filemap_alloc_folio") swap __page_cache_alloc() to -filemap_alloc_folio() (since v5.15) - -Reprocude the error(fedora37, 6.1.7-200.fc37.aarch64): - - $ sudo ./readahead.py - cannot attach kprobe, probe entry may not exist - Traceback (most recent call last): - File "/home/rongtao/Git/bcc/tools/./readahead.py", line 159, in - b.attach_kprobe(event=ra_event, fn_name="entry__do_page_cache_readahead") - File "/usr/lib/python3.11/site-packages/bcc/__init__.py", line 840, in attach_kprobe - raise Exception("Failed to attach BPF program %s to kprobe %s" % - Exception: Failed to attach BPF program b'entry__do_page_cache_readahead' to kprobe b'do_page_cache_ra' - -Signed-off-by: Rong Tao ---- - tools/readahead.py | 69 +++++++++++++++++++++++++++++++++++++--------- - 1 file changed, 56 insertions(+), 13 deletions(-) - -diff --git a/tools/readahead.py b/tools/readahead.py -index f2afdcb3..adad2ea8 100755 ---- a/tools/readahead.py -+++ b/tools/readahead.py -@@ -12,6 +12,7 @@ - # - # 20-Aug-2020 Suchakra Sharma Ported from bpftrace to BCC - # 17-Sep-2021 Hengqi Chen Migrated to kfunc -+# 30-Jan-2023 Rong Tao Support more kfunc/kprobe, introduce folio - - from __future__ import print_function - from bcc import BPF -@@ -38,6 +39,7 @@ args = parser.parse_args() - bpf_text = """ - #include - #include -+#include - - BPF_HASH(flag, u32, u8); // used to track if we are in do_page_cache_readahead() - BPF_HASH(birth, struct page*, u64); // used to track timestamps of cache alloc'ed page -@@ -65,7 +67,7 @@ int exit__do_page_cache_readahead(struct pt_regs *ctx) { - int exit__page_cache_alloc(struct pt_regs *ctx) { - u32 pid; - u64 ts; -- struct page *retval = (struct page*) PT_REGS_RC(ctx); -+ struct page *retval = (struct page*) GET_RETVAL_PAGE; - u32 zero = 0; // static key for accessing pages[0] - pid = bpf_get_current_pid_tgid(); - u8 *f = flag.lookup(&pid); -@@ -111,6 +113,23 @@ KRETFUNC_PROBE(RA_FUNC) - return 0; - } - -+KFUNC_PROBE(mark_page_accessed, struct page *arg0) -+{ -+ u64 ts, delta; -+ u32 zero = 0; // static key for accessing pages[0] -+ u64 *bts = birth.lookup(&arg0); -+ -+ if (bts != NULL) { -+ delta = bpf_ktime_get_ns() - *bts; -+ dist.atomic_increment(bpf_log2l(delta/1000000)); -+ pages.atomic_increment(zero, -1); -+ birth.delete(&arg0); // remove the entry from hashmap -+ } -+ return 0; -+} -+""" -+ -+bpf_text_kfunc_cache_alloc_ret_page = """ - KRETFUNC_PROBE(__page_cache_alloc, gfp_t gfp, struct page *retval) - { - u64 ts; -@@ -125,18 +144,22 @@ KRETFUNC_PROBE(__page_cache_alloc, gfp_t gfp, struct page *retval) - } - return 0; - } -+""" - --KFUNC_PROBE(mark_page_accessed, struct page *arg0) -+bpf_text_kfunc_cache_alloc_ret_folio = """ -+KRETFUNC_PROBE(filemap_alloc_folio, gfp_t gfp, unsigned int order, -+ struct folio *retval) - { -- u64 ts, delta; -+ u64 ts; - u32 zero = 0; // static key for accessing pages[0] -- u64 *bts = birth.lookup(&arg0); -+ u32 pid = bpf_get_current_pid_tgid(); -+ u8 *f = flag.lookup(&pid); -+ struct page *page = folio_page(retval, 0); - -- if (bts != NULL) { -- delta = bpf_ktime_get_ns() - *bts; -- dist.atomic_increment(bpf_log2l(delta/1000000)); -- pages.atomic_increment(zero, -1); -- birth.delete(&arg0); // remove the entry from hashmap -+ if (f != NULL && *f == 1) { -+ ts = bpf_ktime_get_ns(); -+ birth.update(&page, &ts); -+ pages.atomic_increment(zero); - } - return 0; - } -@@ -145,20 +168,40 @@ KFUNC_PROBE(mark_page_accessed, struct page *arg0) - if BPF.support_kfunc(): - if BPF.get_kprobe_functions(b"__do_page_cache_readahead"): - ra_func = "__do_page_cache_readahead" -- else: -+ elif BPF.get_kprobe_functions(b"do_page_cache_ra"): - ra_func = "do_page_cache_ra" -+ elif BPF.get_kprobe_functions(b"page_cache_ra_order"): -+ ra_func = "page_cache_ra_order" -+ else: -+ print("Not found any kfunc.") -+ exit() - bpf_text += bpf_text_kfunc.replace("RA_FUNC", ra_func) -+ if BPF.get_kprobe_functions(b"__page_cache_alloc"): -+ bpf_text += bpf_text_kfunc_cache_alloc_ret_page -+ else: -+ bpf_text += bpf_text_kfunc_cache_alloc_ret_folio - b = BPF(text=bpf_text) - else: - bpf_text += bpf_text_kprobe -- b = BPF(text=bpf_text) - if BPF.get_kprobe_functions(b"__do_page_cache_readahead"): - ra_event = "__do_page_cache_readahead" -- else: -+ elif BPF.get_kprobe_functions(b"do_page_cache_ra"): - ra_event = "do_page_cache_ra" -+ elif BPF.get_kprobe_functions(b"page_cache_ra_order"): -+ ra_event = "page_cache_ra_order" -+ else: -+ print("Not found any kprobe.") -+ exit() -+ if BPF.get_kprobe_functions(b"__page_cache_alloc"): -+ cache_func = "__page_cache_alloc" -+ bpf_text = bpf_text.replace('GET_RETVAL_PAGE', 'PT_REGS_RC(ctx)') -+ else: -+ cache_func = "filemap_alloc_folio" -+ bpf_text = bpf_text.replace('GET_RETVAL_PAGE', 'folio_page((struct folio *)PT_REGS_RC(ctx), 0)') -+ b = BPF(text=bpf_text) - b.attach_kprobe(event=ra_event, fn_name="entry__do_page_cache_readahead") - b.attach_kretprobe(event=ra_event, fn_name="exit__do_page_cache_readahead") -- b.attach_kretprobe(event="__page_cache_alloc", fn_name="exit__page_cache_alloc") -+ b.attach_kretprobe(event=cache_func, fn_name="exit__page_cache_alloc") - b.attach_kprobe(event="mark_page_accessed", fn_name="entry_mark_page_accessed") - - # header --- -2.39.1 - diff --git a/SOURCES/bcc-0.26.0-tools-slabratetop-Fix-error-incomplete-definition-of.patch b/SOURCES/bcc-0.26.0-tools-slabratetop-Fix-error-incomplete-definition-of.patch deleted file mode 100644 index 626e849..0000000 --- a/SOURCES/bcc-0.26.0-tools-slabratetop-Fix-error-incomplete-definition-of.patch +++ /dev/null @@ -1,140 +0,0 @@ -From 533db3453a09695f79368792cdd5fbe2ddeaa55e Mon Sep 17 00:00:00 2001 -From: Rong Tao -Date: Sun, 22 Jan 2023 15:44:46 +0800 -Subject: [PATCH] tools/slabratetop: Fix error: incomplete definition of type - 'struct slab' - -kernel commit 40f3bf0cb04c("mm: Convert struct page to struct slab in functions -used by other subsystems") introduce slab_address() function, commit 6e48a966dfd1 -("mm/kasan: Convert to struct folio and struct slab") linux/kasan.h adds a -dependency on the slab struct, This leads to the following problems: - - $ sudo ./slabratetop.py - In file included from /virtual/main.c:13: - include/linux/slub_def.h:162:26: warning: call to undeclared function 'slab_address'; - ISO C99 and later do not support implicit function declarations [-Wimplicit-function-declaration] - void *object = x - (x - slab_address(slab)) % cache->size; - ^ - include/linux/slub_def.h:162:46: error: invalid operands to binary expression ('void *' and 'unsigned int') - void *object = x - (x - slab_address(slab)) % cache->size; - ~~~~~~~~~~~~~~~~~~~~~~~~ ^ ~~~~~~~~~~~ - include/linux/slub_def.h:164:8: error: incomplete definition of type 'struct slab' - (slab->objects - 1) * cache->size; - ~~~~^ - include/linux/kasan.h:13:8: note: forward declaration of 'struct slab' - struct slab; - ^ - ... - -At first, I wanted to fix this with a kernel patch [1], however, bcc as a -downstream project of the kernel, this issue should be solved inside the bcc -project. This is agreed by kernel maintainer and bcc maintainer @yonghong-song. - -This solution is provided by @yonghong-song [0]. - -[0] https://github.com/iovisor/bcc/issues/4438 -[1] https://lore.kernel.org/all/tencent_ABA832E296819D1053D6C625ADCAF76BC706@qq.com/ - -Signed-off-by: Rong Tao -Signed-off-by: Yonghong Song ---- - tools/slabratetop.py | 76 ++++++++++++++++++++++++++++++++++++++++++++ - 1 file changed, 76 insertions(+) - -diff --git a/tools/slabratetop.py b/tools/slabratetop.py -index ac44b2bd..8fbcac5e 100755 ---- a/tools/slabratetop.py -+++ b/tools/slabratetop.py -@@ -14,6 +14,9 @@ - # Licensed under the Apache License, Version 2.0 (the "License") - # - # 15-Oct-2016 Brendan Gregg Created this. -+# 23-Jan-2023 Rong Tao Introduce kernel internal data structure and -+# functions to temporarily solve problem for -+# >=5.16(TODO: fix this workaround) - - from __future__ import print_function - from bcc import BPF -@@ -65,6 +68,79 @@ bpf_text = """ - // 5.9, but it does not hurt to have it here for versions 5.4 to 5.8. - struct memcg_cache_params {}; - -+// introduce kernel interval slab structure and slab_address() function, solved -+// 'undefined' error for >=5.16. TODO: we should fix this workaround if BCC -+// framework support BTF/CO-RE. -+struct slab { -+ unsigned long __page_flags; -+ -+#if defined(CONFIG_SLAB) -+ -+ struct kmem_cache *slab_cache; -+ union { -+ struct { -+ struct list_head slab_list; -+ void *freelist; /* array of free object indexes */ -+ void *s_mem; /* first object */ -+ }; -+ struct rcu_head rcu_head; -+ }; -+ unsigned int active; -+ -+#elif defined(CONFIG_SLUB) -+ -+ struct kmem_cache *slab_cache; -+ union { -+ struct { -+ union { -+ struct list_head slab_list; -+#ifdef CONFIG_SLUB_CPU_PARTIAL -+ struct { -+ struct slab *next; -+ int slabs; /* Nr of slabs left */ -+ }; -+#endif -+ }; -+ /* Double-word boundary */ -+ void *freelist; /* first free object */ -+ union { -+ unsigned long counters; -+ struct { -+ unsigned inuse:16; -+ unsigned objects:15; -+ unsigned frozen:1; -+ }; -+ }; -+ }; -+ struct rcu_head rcu_head; -+ }; -+ unsigned int __unused; -+ -+#elif defined(CONFIG_SLOB) -+ -+ struct list_head slab_list; -+ void *__unused_1; -+ void *freelist; /* first free block */ -+ long units; -+ unsigned int __unused_2; -+ -+#else -+#error "Unexpected slab allocator configured" -+#endif -+ -+ atomic_t __page_refcount; -+#ifdef CONFIG_MEMCG -+ unsigned long memcg_data; -+#endif -+}; -+ -+// slab_address() will not be used, and NULL will be returned directly, which -+// can avoid adaptation of different kernel versions -+static inline void *slab_address(const struct slab *slab) -+{ -+ return NULL; -+} -+ - #ifdef CONFIG_SLUB - #include - #else --- -2.39.1 - diff --git a/SOURCES/bcc-0.28.0-Adding-memory-zones-for-Power-server.patch b/SOURCES/bcc-0.28.0-Adding-memory-zones-for-Power-server.patch new file mode 100644 index 0000000..579857e --- /dev/null +++ b/SOURCES/bcc-0.28.0-Adding-memory-zones-for-Power-server.patch @@ -0,0 +1,66 @@ +From 63808fbdcb70ce2e858db0a42e7e3eeec153d5b6 Mon Sep 17 00:00:00 2001 +From: Abhishek Dubey +Date: Wed, 20 Sep 2023 10:37:38 -0400 +Subject: [PATCH 4/4] Adding memory zones for Power server + +config PPC_BOOK3S_64 skips setting ZONE_DMA for +server processor. NORMAL and MOVABLE zones are +available on Power. + +Signed-off-by: Abhishek Dubey +--- + tools/compactsnoop.py | 28 +++++++++++++++++++--------- + 1 file changed, 19 insertions(+), 9 deletions(-) + +diff --git a/tools/compactsnoop.py b/tools/compactsnoop.py +index 2b395dec..1a476aad 100755 +--- a/tools/compactsnoop.py ++++ b/tools/compactsnoop.py +@@ -260,11 +260,12 @@ TRACEPOINT_PROBE(compaction, mm_compaction_end) + } + """ + +-if platform.machine() != 'x86_64': ++if platform.machine() != 'x86_64' and platform.machine() != 'ppc64le': + print(""" +- Currently only support x86_64 servers, if you want to use it on +- other platforms, please refer include/linux/mmzone.h to modify +- zone_idex_to_str to get the right zone type ++ Currently only support x86_64 and power servers, if you want ++ to use it on other platforms(including power embedded processors), ++ please refer include/linux/mmzone.h to modify zone_idex_to_str to ++ get the right zone type + """) + exit() + +@@ -296,13 +297,22 @@ initial_ts = 0 + # from include/linux/mmzone.h + # NOTICE: consider only x86_64 servers + zone_type = { +- 0: "ZONE_DMA", +- 1: "ZONE_DMA32", +- 2: "ZONE_NORMAL", ++ 'x86_64': ++ { ++ 0: "ZONE_DMA", ++ 1: "ZONE_DMA32", ++ 2: "ZONE_NORMAL" ++ }, ++ # Zones in Power server only ++ 'ppc64le': ++ { ++ 0: "ZONE_NORMAL", ++ 1: "ZONE_MOVABLE" ++ } + } + +- if idx in zone_type: +- return zone_type[idx] ++ if idx in zone_type[platform.machine()]: ++ return zone_type[platform.machine()][idx] + else: + return str(idx) + +-- +2.43.0 + diff --git a/SOURCES/bcc-0.28.0-Fixing-pvalloc-memleak-test.patch b/SOURCES/bcc-0.28.0-Fixing-pvalloc-memleak-test.patch new file mode 100644 index 0000000..6661f99 --- /dev/null +++ b/SOURCES/bcc-0.28.0-Fixing-pvalloc-memleak-test.patch @@ -0,0 +1,45 @@ +From e6493835a28c08c45fd374e70dba7aa66f700d08 Mon Sep 17 00:00:00 2001 +From: Abhishek Dubey +Date: Tue, 14 Nov 2023 03:54:19 -0500 +Subject: [PATCH 2/4] Fixing pvalloc memleak test + +Request to allocate 30K bytes using pvalloc(), results +in allocating 3*64Kb(on 64Kb pagesize system). The assertion +expects leak to be 30Kb, whereas leaked memory is much more +due to pvalloc's implementation for power. + +Signed-off-by: Abhishek Dubey +--- + tests/python/test_tools_memleak.py | 9 ++++++++- + 1 file changed, 8 insertions(+), 1 deletion(-) + +diff --git a/tests/python/test_tools_memleak.py b/tests/python/test_tools_memleak.py +index cae7e35d..4e921a0c 100755 +--- a/tests/python/test_tools_memleak.py ++++ b/tests/python/test_tools_memleak.py +@@ -3,6 +3,7 @@ + from unittest import main, skipUnless, TestCase + from utils import kernel_version_ge + import os ++import platform + import subprocess + import sys + import tempfile +@@ -102,7 +103,13 @@ TOOLS_DIR = "/bcc/tools/" + self.assertEqual(cfg.leaking_amount, self.run_leaker("memalign")) + + def test_pvalloc(self): +- self.assertEqual(cfg.leaking_amount, self.run_leaker("pvalloc")) ++ # pvalloc's implementation for power invokes mmap(), which adjusts the ++ # allocated size to meet pvalloc's constraints. Actual leaked memory ++ # could be more than requested, hence assertLessEqual. ++ if platform.machine() == 'ppc64le': ++ self.assertLessEqual(cfg.leaking_amount, self.run_leaker("pvalloc")) ++ else: ++ self.assertEqual(cfg.leaking_amount, self.run_leaker("pvalloc")) + + def test_aligned_alloc(self): + self.assertEqual(cfg.leaking_amount, self.run_leaker("aligned_alloc")) +-- +2.43.0 + diff --git a/SOURCES/bcc-0.28.0-Skipping-USDT-tests-for-Power-processor.patch b/SOURCES/bcc-0.28.0-Skipping-USDT-tests-for-Power-processor.patch new file mode 100644 index 0000000..f2beb04 --- /dev/null +++ b/SOURCES/bcc-0.28.0-Skipping-USDT-tests-for-Power-processor.patch @@ -0,0 +1,41 @@ +From a5d86850e3bfeaa23ef4c82dccb9288a2cd42a27 Mon Sep 17 00:00:00 2001 +From: Abhishek Dubey +Date: Mon, 11 Sep 2023 05:10:36 -0400 +Subject: [PATCH 3/4] Skipping USDT tests for Power processor + +Support for Power processor in folly package is absent, +so skipping USDT tests having dependency on folly. + +Signed-off-by: Abhishek Dubey +--- + tests/python/CMakeLists.txt | 14 ++++++++------ + 1 file changed, 8 insertions(+), 6 deletions(-) + +diff --git a/tests/python/CMakeLists.txt b/tests/python/CMakeLists.txt +index a42a16ce..81a547f0 100644 +--- a/tests/python/CMakeLists.txt ++++ b/tests/python/CMakeLists.txt +@@ -71,12 +71,14 @@ add_test(NAME py_test_tools_smoke WORKING_DIRECTORY ${CMAKE_CURRENT_SOURCE_DIR} + COMMAND ${TEST_WRAPPER} py_test_tools_smoke sudo ${CMAKE_CURRENT_SOURCE_DIR}/test_tools_smoke.py) + add_test(NAME py_test_tools_memleak WORKING_DIRECTORY ${CMAKE_CURRENT_SOURCE_DIR} + COMMAND ${TEST_WRAPPER} py_test_tools_memleak sudo ${CMAKE_CURRENT_SOURCE_DIR}/test_tools_memleak.py) +-add_test(NAME py_test_usdt WORKING_DIRECTORY ${CMAKE_CURRENT_SOURCE_DIR} +- COMMAND ${TEST_WRAPPER} py_test_usdt sudo ${CMAKE_CURRENT_SOURCE_DIR}/test_usdt.py) +-add_test(NAME py_test_usdt2 WORKING_DIRECTORY ${CMAKE_CURRENT_SOURCE_DIR} +- COMMAND ${TEST_WRAPPER} py_test_usdt2 sudo ${CMAKE_CURRENT_SOURCE_DIR}/test_usdt2.py) +-add_test(NAME py_test_usdt3 WORKING_DIRECTORY ${CMAKE_CURRENT_SOURCE_DIR} +- COMMAND ${TEST_WRAPPER} py_test_usdt3 sudo ${CMAKE_CURRENT_SOURCE_DIR}/test_usdt3.py) ++if(NOT(CMAKE_SYSTEM_PROCESSOR STREQUAL "ppc64le" OR CMAKE_SYSTEM_PROCESSOR STREQUAL "ppc64")) ++ add_test(NAME py_test_usdt WORKING_DIRECTORY ${CMAKE_CURRENT_SOURCE_DIR} ++ COMMAND ${TEST_WRAPPER} py_test_usdt sudo ${CMAKE_CURRENT_SOURCE_DIR}/test_usdt.py) ++ add_test(NAME py_test_usdt2 WORKING_DIRECTORY ${CMAKE_CURRENT_SOURCE_DIR} ++ COMMAND ${TEST_WRAPPER} py_test_usdt2 sudo ${CMAKE_CURRENT_SOURCE_DIR}/test_usdt2.py) ++ add_test(NAME py_test_usdt3 WORKING_DIRECTORY ${CMAKE_CURRENT_SOURCE_DIR} ++ COMMAND ${TEST_WRAPPER} py_test_usdt3 sudo ${CMAKE_CURRENT_SOURCE_DIR}/test_usdt3.py) ++endif() + add_test(NAME py_test_license WORKING_DIRECTORY ${CMAKE_CURRENT_SOURCE_DIR} + COMMAND ${TEST_WRAPPER} py_test_license sudo ${CMAKE_CURRENT_SOURCE_DIR}/test_license.py) + add_test(NAME py_test_free_bcc_memory WORKING_DIRECTORY ${CMAKE_CURRENT_SOURCE_DIR} +-- +2.43.0 + diff --git a/SOURCES/bcc-0.28.0-libbpf-tools-Add-s390x-support.patch b/SOURCES/bcc-0.28.0-libbpf-tools-Add-s390x-support.patch new file mode 100644 index 0000000..f1206b0 --- /dev/null +++ b/SOURCES/bcc-0.28.0-libbpf-tools-Add-s390x-support.patch @@ -0,0 +1,120974 @@ +From ee7ce397d56ec843c50d3f9f5f2d289809d069c4 Mon Sep 17 00:00:00 2001 +From: Ilya Leoshkevich +Date: Wed, 6 Sep 2023 22:34:33 +0200 +Subject: [PATCH 1/4] libbpf-tools: Add s390x support + +Generate vmlinux.h by running the following commands on Fedora 37: + + libbpf-tools$ bpftool btf dump file /usr/lib/debug/usr/lib/modules/6.4.13-100.fc37.s390x/vmlinux format c > s390/vmlinux_6413.h + libbpf-tools$ ln -s vmlinux_6413.h s390/vmlinux.h + +Signed-off-by: Ilya Leoshkevich +Signed-off-by: Jerome Marchand +--- + libbpf-tools/Makefile | 3 +- + libbpf-tools/s390/vmlinux.h | 1 + + libbpf-tools/s390/vmlinux_6413.h | 120923 ++++++++++++++++++++++++++++ + 3 files changed, 120926 insertions(+), 1 deletion(-) + create mode 120000 libbpf-tools/s390/vmlinux.h + create mode 100644 libbpf-tools/s390/vmlinux_6413.h + +diff --git a/libbpf-tools/Makefile b/libbpf-tools/Makefile +index 9171dacd..daf333c2 100644 +--- a/libbpf-tools/Makefile ++++ b/libbpf-tools/Makefile +@@ -15,7 +15,8 @@ INSTALL ?= install + prefix ?= /usr/local + ARCH ?= $(shell uname -m | sed 's/x86_64/x86/' | sed 's/aarch64/arm64/' \ + | sed 's/ppc64le/powerpc/' | sed 's/mips.*/mips/' \ +- | sed 's/riscv64/riscv/' | sed 's/loongarch.*/loongarch/') ++ | sed 's/riscv64/riscv/' | sed 's/loongarch.*/loongarch/' \ ++ | sed 's/s390x/s390/') + BTFHUB_ARCHIVE ?= $(abspath btfhub-archive) + ifeq ($(ARCH),x86) + CARGO ?= $(shell which cargo) +diff --git a/libbpf-tools/s390/vmlinux.h b/libbpf-tools/s390/vmlinux.h +new file mode 120000 +index 00000000..877b2e06 +--- /dev/null ++++ b/libbpf-tools/s390/vmlinux.h +@@ -0,0 +1 @@ ++vmlinux_6413.h +\ No newline at end of file +diff --git a/libbpf-tools/s390/vmlinux_6413.h b/libbpf-tools/s390/vmlinux_6413.h +new file mode 100644 +index 00000000..1ebc0c7b +--- /dev/null ++++ b/libbpf-tools/s390/vmlinux_6413.h +@@ -0,0 +1,120923 @@ ++#ifndef __VMLINUX_H__ ++#define __VMLINUX_H__ ++ ++#ifndef BPF_NO_PRESERVE_ACCESS_INDEX ++#pragma clang attribute push (__attribute__((preserve_access_index)), apply_to = record) ++#endif ++ ++typedef signed char __s8; ++ ++typedef unsigned char __u8; ++ ++typedef short int __s16; ++ ++typedef short unsigned int __u16; ++ ++typedef int __s32; ++ ++typedef unsigned int __u32; ++ ++typedef long long int __s64; ++ ++typedef long long unsigned int __u64; ++ ++typedef __s8 s8; ++ ++typedef __u8 u8; ++ ++typedef __s16 s16; ++ ++typedef __u16 u16; ++ ++typedef __s32 s32; ++ ++typedef __u32 u32; ++ ++typedef __s64 s64; ++ ++typedef __u64 u64; ++ ++typedef struct { ++ union { ++ struct { ++ __u64 high; ++ __u64 low; ++ }; ++ __u32 u[4]; ++ }; ++} __vector128; ++ ++enum { ++ false = 0, ++ true = 1, ++}; ++ ++typedef long unsigned int __kernel_size_t; ++ ++typedef long int __kernel_ssize_t; ++ ++typedef long int __kernel_long_t; ++ ++typedef long unsigned int __kernel_ulong_t; ++ ++typedef int __kernel_pid_t; ++ ++typedef unsigned int __kernel_uid32_t; ++ ++typedef unsigned int __kernel_gid32_t; ++ ++typedef long long int __kernel_loff_t; ++ ++typedef long long int __kernel_time64_t; ++ ++typedef __kernel_long_t __kernel_clock_t; ++ ++typedef int __kernel_timer_t; ++ ++typedef int __kernel_clockid_t; ++ ++typedef __u16 __be16; ++ ++typedef __u32 __be32; ++ ++typedef __u64 __be64; ++ ++typedef __u32 __wsum; ++ ++typedef unsigned int __poll_t; ++ ++typedef u32 __kernel_dev_t; ++ ++typedef __kernel_dev_t dev_t; ++ ++typedef short unsigned int umode_t; ++ ++typedef __kernel_pid_t pid_t; ++ ++typedef __kernel_clockid_t clockid_t; ++ ++typedef _Bool bool; ++ ++typedef __kernel_uid32_t uid_t; ++ ++typedef __kernel_gid32_t gid_t; ++ ++typedef long unsigned int uintptr_t; ++ ++typedef __kernel_loff_t loff_t; ++ ++typedef __kernel_size_t size_t; ++ ++typedef __kernel_ssize_t ssize_t; ++ ++typedef s32 int32_t; ++ ++typedef u32 uint32_t; ++ ++typedef u64 sector_t; ++ ++typedef u64 blkcnt_t; ++ ++typedef unsigned int gfp_t; ++ ++typedef unsigned int slab_flags_t; ++ ++typedef unsigned int fmode_t; ++ ++typedef u64 phys_addr_t; ++ ++typedef struct { ++ int counter; ++} atomic_t; ++ ++typedef struct { ++ s64 counter; ++} atomic64_t; ++ ++typedef struct { ++ atomic_t refcnt; ++} rcuref_t; ++ ++struct list_head { ++ struct list_head *next; ++ struct list_head *prev; ++}; ++ ++struct hlist_node; ++ ++struct hlist_head { ++ struct hlist_node *first; ++}; ++ ++struct hlist_node { ++ struct hlist_node *next; ++ struct hlist_node **pprev; ++}; ++ ++struct callback_head { ++ struct callback_head *next; ++ void (*func)(struct callback_head *); ++}; ++ ++struct cacheline_padding { ++ char x[0]; ++}; ++ ++typedef __s64 time64_t; ++ ++struct __kernel_timespec { ++ __kernel_time64_t tv_sec; ++ long long int tv_nsec; ++}; ++ ++struct timespec64 { ++ time64_t tv_sec; ++ long int tv_nsec; ++}; ++ ++struct kernel_symbol { ++ long unsigned int value; ++ const char *name; ++ const char *namespace; ++}; ++ ++typedef union { ++ float f; ++ double d; ++ __u64 ui; ++ struct { ++ __u32 hi; ++ __u32 lo; ++ } fp; ++} freg_t; ++ ++typedef struct { ++ long unsigned int mask; ++ long unsigned int addr; ++} psw_t; ++ ++typedef struct { ++ long unsigned int args[1]; ++ psw_t psw; ++ long unsigned int gprs[16]; ++} user_pt_regs; ++ ++struct subchannel_id { ++ __u32 cssid: 8; ++ char: 4; ++ __u32 m: 1; ++ __u32 ssid: 2; ++ __u32 one: 1; ++ __u32 sch_no: 16; ++}; ++ ++struct tpi_info { ++ struct subchannel_id schid; ++ u32 intparm; ++ u32 adapter_IO: 1; ++ u32 directed_irq: 1; ++ u32 isc: 3; ++ short: 11; ++ char: 1; ++ u32 type: 3; ++}; ++ ++struct pt_regs { ++ union { ++ user_pt_regs user_regs; ++ struct { ++ long unsigned int args[1]; ++ psw_t psw; ++ long unsigned int gprs[16]; ++ }; ++ }; ++ long unsigned int orig_gpr2; ++ union { ++ struct { ++ unsigned int int_code; ++ unsigned int int_parm; ++ long unsigned int int_parm_long; ++ }; ++ struct tpi_info tpi_info; ++ }; ++ long unsigned int flags; ++ long unsigned int cr1; ++ long unsigned int last_break; ++}; ++ ++struct per_regs { ++ long unsigned int control; ++ long unsigned int start; ++ long unsigned int end; ++}; ++ ++struct per_event { ++ short unsigned int cause; ++ long unsigned int address; ++ unsigned char paid; ++}; ++ ++struct jump_entry { ++ s32 code; ++ s32 target; ++ long int key; ++}; ++ ++struct static_key_mod; ++ ++struct static_key { ++ atomic_t enabled; ++ union { ++ long unsigned int type; ++ struct jump_entry *entries; ++ struct static_key_mod *next; ++ }; ++}; ++ ++struct static_key_true { ++ struct static_key key; ++}; ++ ++struct static_key_false { ++ struct static_key key; ++}; ++ ++struct pgm_tdb { ++ u64 data[32]; ++}; ++ ++struct lowcore { ++ __u8 pad_0x0000[20]; ++ __u32 ipl_parmblock_ptr; ++ __u8 pad_0x0018[104]; ++ __u32 ext_params; ++ union { ++ struct { ++ __u16 ext_cpu_addr; ++ __u16 ext_int_code; ++ }; ++ __u32 ext_int_code_addr; ++ }; ++ __u32 svc_int_code; ++ union { ++ struct { ++ __u16 pgm_ilc; ++ __u16 pgm_code; ++ }; ++ __u32 pgm_int_code; ++ }; ++ __u32 data_exc_code; ++ __u16 mon_class_num; ++ union { ++ struct { ++ __u8 per_code; ++ __u8 per_atmid; ++ }; ++ __u16 per_code_combined; ++ }; ++ __u64 per_address; ++ __u8 exc_access_id; ++ __u8 per_access_id; ++ __u8 op_access_id; ++ __u8 ar_mode_id; ++ __u8 pad_0x00a4[4]; ++ __u64 trans_exc_code; ++ __u64 monitor_code; ++ union { ++ struct { ++ __u16 subchannel_id; ++ __u16 subchannel_nr; ++ __u32 io_int_parm; ++ __u32 io_int_word; ++ }; ++ struct tpi_info tpi_info; ++ }; ++ __u8 pad_0x00c4[4]; ++ __u32 stfl_fac_list; ++ __u8 pad_0x00cc[28]; ++ __u64 mcck_interruption_code; ++ __u8 pad_0x00f0[4]; ++ __u32 external_damage_code; ++ __u64 failing_storage_address; ++ __u8 pad_0x0100[16]; ++ __u64 pgm_last_break; ++ __u8 pad_0x0118[8]; ++ psw_t restart_old_psw; ++ psw_t external_old_psw; ++ psw_t svc_old_psw; ++ psw_t program_old_psw; ++ psw_t mcck_old_psw; ++ psw_t io_old_psw; ++ __u8 pad_0x0180[32]; ++ psw_t restart_psw; ++ psw_t external_new_psw; ++ psw_t svc_new_psw; ++ psw_t program_new_psw; ++ psw_t mcck_new_psw; ++ psw_t io_new_psw; ++ __u64 save_area_sync[8]; ++ __u64 save_area_async[8]; ++ __u64 save_area_restart[1]; ++ __u64 cpu_flags; ++ psw_t return_psw; ++ psw_t return_mcck_psw; ++ __u64 last_break; ++ __u64 sys_enter_timer; ++ __u64 mcck_enter_timer; ++ __u64 exit_timer; ++ __u64 user_timer; ++ __u64 guest_timer; ++ __u64 system_timer; ++ __u64 hardirq_timer; ++ __u64 softirq_timer; ++ __u64 steal_timer; ++ __u64 avg_steal_timer; ++ __u64 last_update_timer; ++ __u64 last_update_clock; ++ __u64 int_clock; ++ __u64 mcck_clock; ++ __u64 clock_comparator; ++ __u64 boot_clock[2]; ++ __u64 current_task; ++ __u64 kernel_stack; ++ __u64 async_stack; ++ __u64 nodat_stack; ++ __u64 restart_stack; ++ __u64 mcck_stack; ++ __u64 restart_fn; ++ __u64 restart_data; ++ __u32 restart_source; ++ __u32 restart_flags; ++ __u64 kernel_asce; ++ __u64 user_asce; ++ __u32 lpp; ++ __u32 current_pid; ++ __u32 cpu_nr; ++ __u32 softirq_pending; ++ __s32 preempt_count; ++ __u32 spinlock_lockval; ++ __u32 spinlock_index; ++ __u32 fpu_flags; ++ __u64 percpu_offset; ++ __u8 pad_0x03c0[8]; ++ __u64 machine_flags; ++ __u64 gmap; ++ __u8 pad_0x03d8[40]; ++ __u32 return_lpswe; ++ __u32 return_mcck_lpswe; ++ __u8 pad_0x040a[2552]; ++ __u64 ipib; ++ __u32 ipib_checksum; ++ __u64 vmcore_info; ++ __u8 pad_0x0e14[4]; ++ __u64 os_info; ++ __u8 pad_0x0e20[912]; ++ __u64 mcesad; ++ __u64 ext_params2; ++ __u8 pad_0x11c0[64]; ++ __u64 floating_pt_save_area[16]; ++ __u64 gpregs_save_area[16]; ++ psw_t psw_save_area; ++ __u8 pad_0x1310[8]; ++ __u32 prefixreg_save_area; ++ __u32 fpt_creg_save_area; ++ __u8 pad_0x1320[4]; ++ __u32 tod_progreg_save_area; ++ __u32 cpu_timer_save_area[2]; ++ __u32 clock_comp_save_area[2]; ++ __u64 last_break_save_area; ++ __u32 access_regs_save_area[16]; ++ __u64 cregs_save_area[16]; ++ __u8 pad_0x1400[256]; ++ __u64 ccd; ++ __u64 aicd; ++ __u8 pad_0x1510[752]; ++ struct pgm_tdb pgm_tdb; ++ __u8 pad_0x1900[1792]; ++} __attribute__((packed)); ++ ++struct lock_class_key {}; ++ ++struct fs_context; ++ ++struct fs_parameter_spec; ++ ++struct dentry; ++ ++struct super_block; ++ ++struct module; ++ ++struct file_system_type { ++ const char *name; ++ int fs_flags; ++ int (*init_fs_context)(struct fs_context *); ++ const struct fs_parameter_spec *parameters; ++ struct dentry * (*mount)(struct file_system_type *, int, const char *, void *); ++ void (*kill_sb)(struct super_block *); ++ struct module *owner; ++ struct file_system_type *next; ++ struct hlist_head fs_supers; ++ struct lock_class_key s_lock_key; ++ struct lock_class_key s_umount_key; ++ struct lock_class_key s_vfs_rename_key; ++ struct lock_class_key s_writers_key[3]; ++ struct lock_class_key i_lock_key; ++ struct lock_class_key i_mutex_key; ++ struct lock_class_key invalidate_lock_key; ++ struct lock_class_key i_mutex_dir_key; ++}; ++ ++struct obs_kernel_param { ++ const char *str; ++ int (*setup_func)(char *); ++ int early; ++}; ++ ++typedef struct { ++ int lock; ++} arch_spinlock_t; ++ ++typedef struct { ++ int cnts; ++ arch_spinlock_t wait; ++} arch_rwlock_t; ++ ++struct lockdep_map {}; ++ ++struct raw_spinlock { ++ arch_spinlock_t raw_lock; ++}; ++ ++typedef struct raw_spinlock raw_spinlock_t; ++ ++struct ratelimit_state { ++ raw_spinlock_t lock; ++ int interval; ++ int burst; ++ int printed; ++ int missed; ++ long unsigned int begin; ++ long unsigned int flags; ++}; ++ ++struct pi_entry { ++ const char *fmt; ++ const char *func; ++ const char *file; ++ unsigned int line; ++ const char *level; ++ const char *subsys_fmt_prefix; ++} __attribute__((packed)); ++ ++struct _ddebug { ++ const char *modname; ++ const char *function; ++ const char *filename; ++ const char *format; ++ unsigned int lineno: 18; ++ unsigned int class_id: 6; ++ unsigned int flags: 8; ++ union { ++ struct static_key_true dd_key_true; ++ struct static_key_false dd_key_false; ++ } key; ++}; ++ ++enum class_map_type { ++ DD_CLASS_TYPE_DISJOINT_BITS = 0, ++ DD_CLASS_TYPE_LEVEL_NUM = 1, ++ DD_CLASS_TYPE_DISJOINT_NAMES = 2, ++ DD_CLASS_TYPE_LEVEL_NAMES = 3, ++}; ++ ++struct ddebug_class_map { ++ struct list_head link; ++ struct module *mod; ++ const char *mod_name; ++ const char **class_names; ++ const int length; ++ const int base; ++ enum class_map_type map_type; ++}; ++ ++enum module_state { ++ MODULE_STATE_LIVE = 0, ++ MODULE_STATE_COMING = 1, ++ MODULE_STATE_GOING = 2, ++ MODULE_STATE_UNFORMED = 3, ++}; ++ ++struct refcount_struct { ++ atomic_t refs; ++}; ++ ++typedef struct refcount_struct refcount_t; ++ ++struct kref { ++ refcount_t refcount; ++}; ++ ++struct kset; ++ ++struct kobj_type; ++ ++struct kernfs_node; ++ ++struct kobject { ++ const char *name; ++ struct list_head entry; ++ struct kobject *parent; ++ struct kset *kset; ++ const struct kobj_type *ktype; ++ struct kernfs_node *sd; ++ struct kref kref; ++ unsigned int state_initialized: 1; ++ unsigned int state_in_sysfs: 1; ++ unsigned int state_add_uevent_sent: 1; ++ unsigned int state_remove_uevent_sent: 1; ++ unsigned int uevent_suppress: 1; ++}; ++ ++struct module_param_attrs; ++ ++struct completion; ++ ++struct module_kobject { ++ struct kobject kobj; ++ struct module *mod; ++ struct kobject *drivers_dir; ++ struct module_param_attrs *mp; ++ struct completion *kobj_completion; ++}; ++ ++typedef atomic64_t atomic_long_t; ++ ++struct optimistic_spin_queue { ++ atomic_t tail; ++}; ++ ++struct mutex { ++ atomic_long_t owner; ++ raw_spinlock_t wait_lock; ++ struct optimistic_spin_queue osq; ++ struct list_head wait_list; ++}; ++ ++struct rb_node { ++ long unsigned int __rb_parent_color; ++ struct rb_node *rb_right; ++ struct rb_node *rb_left; ++}; ++ ++struct latch_tree_node { ++ struct rb_node node[2]; ++}; ++ ++struct mod_tree_node { ++ struct module *mod; ++ struct latch_tree_node node; ++}; ++ ++struct module_memory { ++ void *base; ++ unsigned int size; ++ struct mod_tree_node mtn; ++}; ++ ++struct mod_arch_syminfo; ++ ++struct ftrace_hotpatch_trampoline; ++ ++struct mod_arch_specific { ++ long unsigned int got_offset; ++ long unsigned int plt_offset; ++ long unsigned int got_size; ++ long unsigned int plt_size; ++ int nsyms; ++ struct mod_arch_syminfo *syminfo; ++ struct ftrace_hotpatch_trampoline *trampolines_start; ++ struct ftrace_hotpatch_trampoline *trampolines_end; ++ struct ftrace_hotpatch_trampoline *next_trampoline; ++}; ++ ++struct elf64_sym; ++ ++typedef struct elf64_sym Elf64_Sym; ++ ++struct mod_kallsyms { ++ Elf64_Sym *symtab; ++ unsigned int num_symtab; ++ char *strtab; ++ char *typetab; ++}; ++ ++struct _ddebug_info { ++ struct _ddebug *descs; ++ struct ddebug_class_map *classes; ++ unsigned int num_descs; ++ unsigned int num_classes; ++}; ++ ++struct module_attribute; ++ ++struct kernel_param; ++ ++struct exception_table_entry; ++ ++struct bug_entry; ++ ++struct module_sect_attrs; ++ ++struct module_notes_attrs; ++ ++struct tracepoint; ++ ++typedef struct tracepoint * const tracepoint_ptr_t; ++ ++struct srcu_struct; ++ ++struct bpf_raw_event_map; ++ ++struct trace_event_call; ++ ++struct trace_eval_map; ++ ++struct kunit_suite; ++ ++struct klp_modinfo; ++ ++struct module { ++ enum module_state state; ++ struct list_head list; ++ char name[56]; ++ struct module_kobject mkobj; ++ struct module_attribute *modinfo_attrs; ++ const char *version; ++ const char *srcversion; ++ const char *rhelversion; ++ struct kobject *holders_dir; ++ const struct kernel_symbol *syms; ++ const s32 *crcs; ++ unsigned int num_syms; ++ struct mutex param_lock; ++ struct kernel_param *kp; ++ unsigned int num_kp; ++ unsigned int num_gpl_syms; ++ const struct kernel_symbol *gpl_syms; ++ const s32 *gpl_crcs; ++ bool using_gplonly_symbols; ++ bool sig_ok; ++ bool async_probe_requested; ++ unsigned int num_exentries; ++ struct exception_table_entry *extable; ++ int (*init)(); ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct module_memory mem[7]; ++ struct mod_arch_specific arch; ++ long unsigned int taints; ++ unsigned int num_bugs; ++ struct list_head bug_list; ++ struct bug_entry *bug_table; ++ struct mod_kallsyms *kallsyms; ++ struct mod_kallsyms core_kallsyms; ++ struct module_sect_attrs *sect_attrs; ++ struct module_notes_attrs *notes_attrs; ++ char *args; ++ void *percpu; ++ unsigned int percpu_size; ++ void *noinstr_text_start; ++ unsigned int noinstr_text_size; ++ unsigned int num_tracepoints; ++ tracepoint_ptr_t *tracepoints_ptrs; ++ unsigned int num_srcu_structs; ++ struct srcu_struct **srcu_struct_ptrs; ++ unsigned int num_bpf_raw_events; ++ struct bpf_raw_event_map *bpf_raw_events; ++ unsigned int btf_data_size; ++ void *btf_data; ++ struct jump_entry *jump_entries; ++ unsigned int num_jump_entries; ++ unsigned int num_trace_bprintk_fmt; ++ const char **trace_bprintk_fmt_start; ++ struct trace_event_call **trace_events; ++ unsigned int num_trace_events; ++ struct trace_eval_map **trace_evals; ++ unsigned int num_trace_evals; ++ unsigned int num_ftrace_callsites; ++ long unsigned int *ftrace_callsites; ++ void *kprobes_text_start; ++ unsigned int kprobes_text_size; ++ long unsigned int *kprobe_blacklist; ++ unsigned int num_kprobe_blacklist; ++ int num_kunit_suites; ++ struct kunit_suite **kunit_suites; ++ bool klp; ++ bool klp_alive; ++ struct klp_modinfo *klp_info; ++ unsigned int printk_index_size; ++ struct pi_entry **printk_index_start; ++ struct list_head source_list; ++ struct list_head target_list; ++ void (*exit)(); ++ atomic_t refcnt; ++ struct _ddebug_info dyndbg_info; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct kernel_param_ops { ++ unsigned int flags; ++ int (*set)(const char *, const struct kernel_param *); ++ int (*get)(char *, const struct kernel_param *); ++ void (*free)(void *); ++}; ++ ++typedef void *fl_owner_t; ++ ++struct file; ++ ++struct kiocb; ++ ++struct iov_iter; ++ ++struct io_comp_batch; ++ ++struct dir_context; ++ ++struct poll_table_struct; ++ ++struct vm_area_struct; ++ ++struct inode; ++ ++struct file_lock; ++ ++struct page; ++ ++struct pipe_inode_info; ++ ++struct seq_file; ++ ++struct io_uring_cmd; ++ ++struct file_operations { ++ struct module *owner; ++ loff_t (*llseek)(struct file *, loff_t, int); ++ ssize_t (*read)(struct file *, char *, size_t, loff_t *); ++ ssize_t (*write)(struct file *, const char *, size_t, loff_t *); ++ ssize_t (*read_iter)(struct kiocb *, struct iov_iter *); ++ ssize_t (*write_iter)(struct kiocb *, struct iov_iter *); ++ int (*iopoll)(struct kiocb *, struct io_comp_batch *, unsigned int); ++ int (*iterate)(struct file *, struct dir_context *); ++ int (*iterate_shared)(struct file *, struct dir_context *); ++ __poll_t (*poll)(struct file *, struct poll_table_struct *); ++ long int (*unlocked_ioctl)(struct file *, unsigned int, long unsigned int); ++ long int (*compat_ioctl)(struct file *, unsigned int, long unsigned int); ++ int (*mmap)(struct file *, struct vm_area_struct *); ++ long unsigned int mmap_supported_flags; ++ int (*open)(struct inode *, struct file *); ++ int (*flush)(struct file *, fl_owner_t); ++ int (*release)(struct inode *, struct file *); ++ int (*fsync)(struct file *, loff_t, loff_t, int); ++ int (*fasync)(int, struct file *, int); ++ int (*lock)(struct file *, int, struct file_lock *); ++ ssize_t (*sendpage)(struct file *, struct page *, int, size_t, loff_t *, int); ++ long unsigned int (*get_unmapped_area)(struct file *, long unsigned int, long unsigned int, long unsigned int, long unsigned int); ++ int (*check_flags)(int); ++ int (*flock)(struct file *, int, struct file_lock *); ++ ssize_t (*splice_write)(struct pipe_inode_info *, struct file *, loff_t *, size_t, unsigned int); ++ ssize_t (*splice_read)(struct file *, loff_t *, struct pipe_inode_info *, size_t, unsigned int); ++ int (*setlease)(struct file *, long int, struct file_lock **, void **); ++ long int (*fallocate)(struct file *, int, loff_t, loff_t); ++ void (*show_fdinfo)(struct seq_file *, struct file *); ++ ssize_t (*copy_file_range)(struct file *, loff_t, struct file *, loff_t, size_t, unsigned int); ++ loff_t (*remap_file_range)(struct file *, loff_t, struct file *, loff_t, loff_t, unsigned int); ++ int (*fadvise)(struct file *, loff_t, loff_t, int); ++ int (*uring_cmd)(struct io_uring_cmd *, unsigned int); ++ int (*uring_cmd_iopoll)(struct io_uring_cmd *, struct io_comp_batch *, unsigned int); ++}; ++ ++struct bug_entry { ++ int bug_addr_disp; ++ int file_disp; ++ short unsigned int line; ++ short unsigned int flags; ++}; ++ ++enum timespec_type { ++ TT_NONE = 0, ++ TT_NATIVE = 1, ++ TT_COMPAT = 2, ++}; ++ ++typedef s32 old_time32_t; ++ ++struct old_timespec32 { ++ old_time32_t tv_sec; ++ s32 tv_nsec; ++}; ++ ++struct pollfd { ++ int fd; ++ short int events; ++ short int revents; ++}; ++ ++struct restart_block { ++ long unsigned int arch_data; ++ long int (*fn)(struct restart_block *); ++ union { ++ struct { ++ u32 *uaddr; ++ u32 val; ++ u32 flags; ++ u32 bitset; ++ u64 time; ++ u32 *uaddr2; ++ } futex; ++ struct { ++ clockid_t clockid; ++ enum timespec_type type; ++ union { ++ struct __kernel_timespec *rmtp; ++ struct old_timespec32 *compat_rmtp; ++ }; ++ u64 expires; ++ } nanosleep; ++ struct { ++ struct pollfd *ufds; ++ int nfds; ++ int has_timeout; ++ long unsigned int tv_sec; ++ long unsigned int tv_nsec; ++ } poll; ++ }; ++}; ++ ++typedef struct { ++ long unsigned int pgprot; ++} pgprot_t; ++ ++typedef struct { ++ long unsigned int pte; ++} pte_t; ++ ++typedef struct { ++ long unsigned int pmd; ++} pmd_t; ++ ++typedef struct { ++ long unsigned int pud; ++} pud_t; ++ ++typedef struct { ++ long unsigned int pgd; ++} pgd_t; ++ ++typedef pte_t *pgtable_t; ++ ++struct thread_info { ++ long unsigned int flags; ++ long unsigned int syscall_work; ++ unsigned int cpu; ++}; ++ ++struct llist_node { ++ struct llist_node *next; ++}; ++ ++struct __call_single_node { ++ struct llist_node llist; ++ union { ++ unsigned int u_flags; ++ atomic_t a_flags; ++ }; ++ u16 src; ++ u16 dst; ++}; ++ ++struct load_weight { ++ long unsigned int weight; ++ u32 inv_weight; ++}; ++ ++struct util_est { ++ unsigned int enqueued; ++ unsigned int ewma; ++}; ++ ++struct sched_avg { ++ u64 last_update_time; ++ u64 load_sum; ++ u64 runnable_sum; ++ u32 util_sum; ++ u32 period_contrib; ++ long unsigned int load_avg; ++ long unsigned int runnable_avg; ++ long unsigned int util_avg; ++ struct util_est util_est; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct cfs_rq; ++ ++struct sched_entity { ++ struct load_weight load; ++ struct rb_node run_node; ++ struct list_head group_node; ++ unsigned int on_rq; ++ u64 exec_start; ++ u64 sum_exec_runtime; ++ u64 vruntime; ++ u64 prev_sum_exec_runtime; ++ u64 nr_migrations; ++ int depth; ++ struct sched_entity *parent; ++ struct cfs_rq *cfs_rq; ++ struct cfs_rq *my_q; ++ long unsigned int runnable_weight; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct sched_avg avg; ++}; ++ ++struct sched_rt_entity { ++ struct list_head run_list; ++ long unsigned int timeout; ++ long unsigned int watchdog_stamp; ++ unsigned int time_slice; ++ short unsigned int on_rq; ++ short unsigned int on_list; ++ struct sched_rt_entity *back; ++}; ++ ++typedef s64 ktime_t; ++ ++struct timerqueue_node { ++ struct rb_node node; ++ ktime_t expires; ++}; ++ ++enum hrtimer_restart { ++ HRTIMER_NORESTART = 0, ++ HRTIMER_RESTART = 1, ++}; ++ ++struct hrtimer_clock_base; ++ ++struct hrtimer { ++ struct timerqueue_node node; ++ ktime_t _softexpires; ++ enum hrtimer_restart (*function)(struct hrtimer *); ++ struct hrtimer_clock_base *base; ++ u8 state; ++ u8 is_rel; ++ u8 is_soft; ++ u8 is_hard; ++}; ++ ++struct sched_dl_entity { ++ struct rb_node rb_node; ++ u64 dl_runtime; ++ u64 dl_deadline; ++ u64 dl_period; ++ u64 dl_bw; ++ u64 dl_density; ++ s64 runtime; ++ u64 deadline; ++ unsigned int flags; ++ unsigned int dl_throttled: 1; ++ unsigned int dl_yielded: 1; ++ unsigned int dl_non_contending: 1; ++ unsigned int dl_overrun: 1; ++ struct hrtimer dl_timer; ++ struct hrtimer inactive_timer; ++ struct sched_dl_entity *pi_se; ++}; ++ ++struct sched_statistics { ++ u64 wait_start; ++ u64 wait_max; ++ u64 wait_count; ++ u64 wait_sum; ++ u64 iowait_count; ++ u64 iowait_sum; ++ u64 sleep_start; ++ u64 sleep_max; ++ s64 sum_sleep_runtime; ++ u64 block_start; ++ u64 block_max; ++ s64 sum_block_runtime; ++ u64 exec_max; ++ u64 slice_max; ++ u64 nr_migrations_cold; ++ u64 nr_failed_migrations_affine; ++ u64 nr_failed_migrations_running; ++ u64 nr_failed_migrations_hot; ++ u64 nr_forced_migrations; ++ u64 nr_wakeups; ++ u64 nr_wakeups_sync; ++ u64 nr_wakeups_migrate; ++ u64 nr_wakeups_local; ++ u64 nr_wakeups_remote; ++ u64 nr_wakeups_affine; ++ u64 nr_wakeups_affine_attempts; ++ u64 nr_wakeups_passive; ++ u64 nr_wakeups_idle; ++ u64 core_forceidle_sum; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct cpumask { ++ long unsigned int bits[8]; ++}; ++ ++typedef struct cpumask cpumask_t; ++ ++union rcu_special { ++ struct { ++ u8 blocked; ++ u8 need_qs; ++ u8 exp_hint; ++ u8 need_mb; ++ } b; ++ u32 s; ++}; ++ ++struct sched_info { ++ long unsigned int pcount; ++ long long unsigned int run_delay; ++ long long unsigned int last_arrival; ++ long long unsigned int last_queued; ++}; ++ ++struct plist_node { ++ int prio; ++ struct list_head prio_list; ++ struct list_head node_list; ++}; ++ ++struct prev_cputime {}; ++ ++struct rb_root { ++ struct rb_node *rb_node; ++}; ++ ++struct rb_root_cached { ++ struct rb_root rb_root; ++ struct rb_node *rb_leftmost; ++}; ++ ++struct timerqueue_head { ++ struct rb_root_cached rb_root; ++}; ++ ++struct posix_cputimer_base { ++ u64 nextevt; ++ struct timerqueue_head tqhead; ++}; ++ ++struct posix_cputimers { ++ struct posix_cputimer_base bases[3]; ++ unsigned int timers_active; ++ unsigned int expiry_active; ++}; ++ ++struct sem_undo_list; ++ ++struct sysv_sem { ++ struct sem_undo_list *undo_list; ++}; ++ ++struct sysv_shm { ++ struct list_head shm_clist; ++}; ++ ++typedef struct { ++ long unsigned int sig[1]; ++} sigset_t; ++ ++struct sigpending { ++ struct list_head list; ++ sigset_t signal; ++}; ++ ++typedef struct { ++ uid_t val; ++} kuid_t; ++ ++struct seccomp_filter; ++ ++struct seccomp { ++ int mode; ++ atomic_t filter_count; ++ struct seccomp_filter *filter; ++}; ++ ++struct syscall_user_dispatch { ++ char *selector; ++ long unsigned int offset; ++ long unsigned int len; ++ bool on_dispatch; ++}; ++ ++struct spinlock { ++ union { ++ struct raw_spinlock rlock; ++ }; ++}; ++ ++typedef struct spinlock spinlock_t; ++ ++struct wake_q_node { ++ struct wake_q_node *next; ++}; ++ ++struct task_io_accounting { ++ u64 rchar; ++ u64 wchar; ++ u64 syscr; ++ u64 syscw; ++ u64 read_bytes; ++ u64 write_bytes; ++ u64 cancelled_write_bytes; ++}; ++ ++typedef struct { ++ long unsigned int bits[1]; ++} nodemask_t; ++ ++struct seqcount { ++ unsigned int sequence; ++}; ++ ++typedef struct seqcount seqcount_t; ++ ++struct seqcount_spinlock { ++ seqcount_t seqcount; ++}; ++ ++typedef struct seqcount_spinlock seqcount_spinlock_t; ++ ++struct tlbflush_unmap_batch {}; ++ ++struct page_frag { ++ struct page *page; ++ __u32 offset; ++ __u32 size; ++}; ++ ++struct latency_record { ++ long unsigned int backtrace[12]; ++ unsigned int count; ++ long unsigned int time; ++ long unsigned int max; ++}; ++ ++struct kmap_ctrl {}; ++ ++struct timer_list { ++ struct hlist_node entry; ++ long unsigned int expires; ++ void (*function)(struct timer_list *); ++ u32 flags; ++}; ++ ++struct llist_head { ++ struct llist_node *first; ++}; ++ ++struct da_monitor { ++ bool monitoring; ++ unsigned int curr_state; ++}; ++ ++union rv_task_monitor { ++ struct da_monitor da_mon; ++}; ++ ++typedef long int (*sys_call_ptr_t)(struct pt_regs *); ++ ++struct fpu { ++ __u32 fpc; ++ void *regs; ++ union { ++ freg_t fprs[16]; ++ __vector128 vxrs[32]; ++ }; ++}; ++ ++struct runtime_instr_cb; ++ ++struct gs_cb; ++ ++struct thread_struct { ++ unsigned int acrs[16]; ++ long unsigned int ksp; ++ long unsigned int user_timer; ++ long unsigned int guest_timer; ++ long unsigned int system_timer; ++ long unsigned int hardirq_timer; ++ long unsigned int softirq_timer; ++ const sys_call_ptr_t *sys_call_table; ++ long unsigned int gmap_addr; ++ unsigned int gmap_write_flag; ++ unsigned int gmap_int_code; ++ unsigned int gmap_pfault; ++ struct per_regs per_user; ++ struct per_event per_event; ++ long unsigned int per_flags; ++ unsigned int system_call; ++ long unsigned int last_break; ++ long unsigned int pfault_wait; ++ struct list_head list; ++ struct runtime_instr_cb *ri_cb; ++ struct gs_cb *gs_cb; ++ struct gs_cb *gs_bc_cb; ++ struct pgm_tdb trap_tdb; ++ struct fpu fpu; ++}; ++ ++struct sched_class; ++ ++struct task_group; ++ ++struct mm_struct; ++ ++struct pid; ++ ++struct cred; ++ ++struct key; ++ ++struct nameidata; ++ ++struct fs_struct; ++ ++struct files_struct; ++ ++struct io_uring_task; ++ ++struct nsproxy; ++ ++struct signal_struct; ++ ++struct sighand_struct; ++ ++struct audit_context; ++ ++struct rt_mutex_waiter; ++ ++struct bio_list; ++ ++struct blk_plug; ++ ++struct reclaim_state; ++ ++struct backing_dev_info; ++ ++struct io_context; ++ ++struct capture_control; ++ ++struct kernel_siginfo; ++ ++typedef struct kernel_siginfo kernel_siginfo_t; ++ ++struct css_set; ++ ++struct robust_list_head; ++ ++struct compat_robust_list_head; ++ ++struct futex_pi_state; ++ ++struct perf_event_context; ++ ++struct mempolicy; ++ ++struct numa_group; ++ ++struct rseq; ++ ++struct task_delay_info; ++ ++struct kunit; ++ ++struct ftrace_ret_stack; ++ ++struct mem_cgroup; ++ ++struct gendisk; ++ ++struct uprobe_task; ++ ++struct vm_struct; ++ ++struct bpf_local_storage; ++ ++struct bpf_run_ctx; ++ ++struct task_struct { ++ struct thread_info thread_info; ++ unsigned int __state; ++ void *stack; ++ refcount_t usage; ++ unsigned int flags; ++ unsigned int ptrace; ++ int on_cpu; ++ struct __call_single_node wake_entry; ++ unsigned int wakee_flips; ++ long unsigned int wakee_flip_decay_ts; ++ struct task_struct *last_wakee; ++ int recent_used_cpu; ++ int wake_cpu; ++ int on_rq; ++ int prio; ++ int static_prio; ++ int normal_prio; ++ unsigned int rt_priority; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct sched_entity se; ++ struct sched_rt_entity rt; ++ struct sched_dl_entity dl; ++ const struct sched_class *sched_class; ++ struct rb_node core_node; ++ long unsigned int core_cookie; ++ unsigned int core_occupation; ++ struct task_group *sched_task_group; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct sched_statistics stats; ++ struct hlist_head preempt_notifiers; ++ unsigned int btrace_seq; ++ unsigned int policy; ++ int nr_cpus_allowed; ++ const cpumask_t *cpus_ptr; ++ cpumask_t *user_cpus_ptr; ++ cpumask_t cpus_mask; ++ void *migration_pending; ++ short unsigned int migration_disabled; ++ short unsigned int migration_flags; ++ int trc_reader_nesting; ++ int trc_ipi_to_cpu; ++ union rcu_special trc_reader_special; ++ struct list_head trc_holdout_list; ++ struct list_head trc_blkd_node; ++ int trc_blkd_cpu; ++ struct sched_info sched_info; ++ struct list_head tasks; ++ struct plist_node pushable_tasks; ++ struct rb_node pushable_dl_tasks; ++ struct mm_struct *mm; ++ struct mm_struct *active_mm; ++ int exit_state; ++ int exit_code; ++ int exit_signal; ++ int pdeath_signal; ++ long unsigned int jobctl; ++ unsigned int personality; ++ unsigned int sched_reset_on_fork: 1; ++ unsigned int sched_contributes_to_load: 1; ++ unsigned int sched_migrated: 1; ++ long: 29; ++ unsigned int sched_remote_wakeup: 1; ++ unsigned int in_execve: 1; ++ unsigned int in_iowait: 1; ++ unsigned int in_user_fault: 1; ++ unsigned int in_lru_fault: 1; ++ unsigned int no_cgroup_migration: 1; ++ unsigned int frozen: 1; ++ unsigned int use_memdelay: 1; ++ unsigned int in_memstall: 1; ++ unsigned int in_page_owner: 1; ++ unsigned int in_eventfd: 1; ++ unsigned int in_thrashing: 1; ++ long unsigned int atomic_flags; ++ struct restart_block restart_block; ++ pid_t pid; ++ pid_t tgid; ++ struct task_struct *real_parent; ++ struct task_struct *parent; ++ struct list_head children; ++ struct list_head sibling; ++ struct task_struct *group_leader; ++ struct list_head ptraced; ++ struct list_head ptrace_entry; ++ struct pid *thread_pid; ++ struct hlist_node pid_links[4]; ++ struct list_head thread_group; ++ struct list_head thread_node; ++ struct completion *vfork_done; ++ int *set_child_tid; ++ int *clear_child_tid; ++ void *worker_private; ++ u64 utime; ++ u64 stime; ++ u64 utimescaled; ++ u64 stimescaled; ++ u64 gtime; ++ struct prev_cputime prev_cputime; ++ long unsigned int nvcsw; ++ long unsigned int nivcsw; ++ u64 start_time; ++ u64 start_boottime; ++ long unsigned int min_flt; ++ long unsigned int maj_flt; ++ struct posix_cputimers posix_cputimers; ++ const struct cred *ptracer_cred; ++ const struct cred *real_cred; ++ const struct cred *cred; ++ struct key *cached_requested_key; ++ char comm[16]; ++ struct nameidata *nameidata; ++ struct sysv_sem sysvsem; ++ struct sysv_shm sysvshm; ++ struct fs_struct *fs; ++ struct files_struct *files; ++ struct io_uring_task *io_uring; ++ struct nsproxy *nsproxy; ++ struct signal_struct *signal; ++ struct sighand_struct *sighand; ++ sigset_t blocked; ++ sigset_t real_blocked; ++ sigset_t saved_sigmask; ++ struct sigpending pending; ++ long unsigned int sas_ss_sp; ++ size_t sas_ss_size; ++ unsigned int sas_ss_flags; ++ struct callback_head *task_works; ++ struct audit_context *audit_context; ++ kuid_t loginuid; ++ unsigned int sessionid; ++ struct seccomp seccomp; ++ struct syscall_user_dispatch syscall_dispatch; ++ u64 parent_exec_id; ++ u64 self_exec_id; ++ spinlock_t alloc_lock; ++ raw_spinlock_t pi_lock; ++ struct wake_q_node wake_q; ++ struct rb_root_cached pi_waiters; ++ struct task_struct *pi_top_task; ++ struct rt_mutex_waiter *pi_blocked_on; ++ void *journal_info; ++ struct bio_list *bio_list; ++ struct blk_plug *plug; ++ struct reclaim_state *reclaim_state; ++ struct backing_dev_info *backing_dev_info; ++ struct io_context *io_context; ++ struct capture_control *capture_control; ++ long unsigned int ptrace_message; ++ kernel_siginfo_t *last_siginfo; ++ struct task_io_accounting ioac; ++ unsigned int psi_flags; ++ u64 acct_rss_mem1; ++ u64 acct_vm_mem1; ++ u64 acct_timexpd; ++ nodemask_t mems_allowed; ++ seqcount_spinlock_t mems_allowed_seq; ++ int cpuset_mem_spread_rotor; ++ int cpuset_slab_spread_rotor; ++ struct css_set *cgroups; ++ struct list_head cg_list; ++ struct robust_list_head *robust_list; ++ struct compat_robust_list_head *compat_robust_list; ++ struct list_head pi_state_list; ++ struct futex_pi_state *pi_state_cache; ++ struct mutex futex_exit_mutex; ++ unsigned int futex_state; ++ struct perf_event_context *perf_event_ctxp; ++ struct mutex perf_event_mutex; ++ struct list_head perf_event_list; ++ struct mempolicy *mempolicy; ++ short int il_prev; ++ short int pref_node_fork; ++ int numa_scan_seq; ++ unsigned int numa_scan_period; ++ unsigned int numa_scan_period_max; ++ int numa_preferred_nid; ++ long unsigned int numa_migrate_retry; ++ u64 node_stamp; ++ u64 last_task_numa_placement; ++ u64 last_sum_exec_runtime; ++ struct callback_head numa_work; ++ struct numa_group *numa_group; ++ long unsigned int *numa_faults; ++ long unsigned int total_numa_faults; ++ long unsigned int numa_faults_locality[3]; ++ long unsigned int numa_pages_migrated; ++ struct rseq *rseq; ++ u32 rseq_len; ++ u32 rseq_sig; ++ long unsigned int rseq_event_mask; ++ int mm_cid; ++ int last_mm_cid; ++ int migrate_from_cpu; ++ int mm_cid_active; ++ struct callback_head cid_work; ++ struct tlbflush_unmap_batch tlb_ubc; ++ struct pipe_inode_info *splice_pipe; ++ struct page_frag task_frag; ++ struct task_delay_info *delays; ++ int nr_dirtied; ++ int nr_dirtied_pause; ++ long unsigned int dirty_paused_when; ++ int latency_record_count; ++ struct latency_record latency_record[32]; ++ u64 timer_slack_ns; ++ u64 default_timer_slack_ns; ++ struct kunit *kunit_test; ++ int curr_ret_stack; ++ int curr_ret_depth; ++ struct ftrace_ret_stack *ret_stack; ++ long long unsigned int ftrace_timestamp; ++ atomic_t trace_overrun; ++ atomic_t tracing_graph_pause; ++ long unsigned int trace_recursion; ++ struct mem_cgroup *memcg_in_oom; ++ gfp_t memcg_oom_gfp_mask; ++ int memcg_oom_order; ++ unsigned int memcg_nr_pages_over_high; ++ struct mem_cgroup *active_memcg; ++ struct gendisk *throttle_disk; ++ struct uprobe_task *utask; ++ unsigned int sequential_io; ++ unsigned int sequential_io_avg; ++ struct kmap_ctrl kmap_ctrl; ++ struct callback_head rcu; ++ refcount_t rcu_users; ++ int pagefault_disabled; ++ struct task_struct *oom_reaper_list; ++ struct timer_list oom_reaper_timer; ++ struct vm_struct *stack_vm_area; ++ refcount_t stack_refcount; ++ int patch_state; ++ void *security; ++ struct bpf_local_storage *bpf_storage; ++ struct bpf_run_ctx *bpf_ctx; ++ struct llist_head kretprobe_instances; ++ struct llist_head rethooks; ++ union rv_task_monitor rv[1]; ++ struct thread_struct thread; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct uid_gid_extent { ++ u32 first; ++ u32 lower_first; ++ u32 count; ++}; ++ ++struct uid_gid_map { ++ u32 nr_extents; ++ union { ++ struct uid_gid_extent extent[5]; ++ struct { ++ struct uid_gid_extent *forward; ++ struct uid_gid_extent *reverse; ++ }; ++ }; ++}; ++ ++typedef struct { ++ gid_t val; ++} kgid_t; ++ ++struct proc_ns_operations; ++ ++struct ns_common { ++ atomic_long_t stashed; ++ const struct proc_ns_operations *ops; ++ unsigned int inum; ++ refcount_t count; ++}; ++ ++struct rw_semaphore { ++ atomic_long_t count; ++ atomic_long_t owner; ++ struct optimistic_spin_queue osq; ++ raw_spinlock_t wait_lock; ++ struct list_head wait_list; ++}; ++ ++struct work_struct; ++ ++typedef void (*work_func_t)(struct work_struct *); ++ ++struct work_struct { ++ atomic_long_t data; ++ struct list_head entry; ++ work_func_t func; ++}; ++ ++struct ctl_table; ++ ++struct ctl_table_root; ++ ++struct ctl_table_set; ++ ++struct ctl_dir; ++ ++struct ctl_node; ++ ++struct ctl_table_header { ++ union { ++ struct { ++ struct ctl_table *ctl_table; ++ int used; ++ int count; ++ int nreg; ++ }; ++ struct callback_head rcu; ++ }; ++ struct completion *unregistering; ++ struct ctl_table *ctl_table_arg; ++ struct ctl_table_root *root; ++ struct ctl_table_set *set; ++ struct ctl_dir *parent; ++ struct ctl_node *node; ++ struct hlist_head inodes; ++}; ++ ++struct ctl_dir { ++ struct ctl_table_header header; ++ struct rb_root root; ++}; ++ ++struct ctl_table_set { ++ int (*is_seen)(struct ctl_table_set *); ++ struct ctl_dir dir; ++}; ++ ++struct ucounts; ++ ++struct user_namespace { ++ struct uid_gid_map uid_map; ++ struct uid_gid_map gid_map; ++ struct uid_gid_map projid_map; ++ struct user_namespace *parent; ++ int level; ++ kuid_t owner; ++ kgid_t group; ++ struct ns_common ns; ++ long unsigned int flags; ++ bool parent_could_setfcap; ++ struct list_head keyring_name_list; ++ struct key *user_keyring_register; ++ struct rw_semaphore keyring_sem; ++ struct key *persistent_keyring_register; ++ struct work_struct work; ++ struct ctl_table_set set; ++ struct ctl_table_header *sysctls; ++ struct ucounts *ucounts; ++ long int ucount_max[12]; ++ long int rlimit_max[4]; ++}; ++ ++struct kstat { ++ u32 result_mask; ++ umode_t mode; ++ unsigned int nlink; ++ uint32_t blksize; ++ u64 attributes; ++ u64 attributes_mask; ++ u64 ino; ++ dev_t dev; ++ dev_t rdev; ++ kuid_t uid; ++ kgid_t gid; ++ loff_t size; ++ struct timespec64 atime; ++ struct timespec64 mtime; ++ struct timespec64 ctime; ++ struct timespec64 btime; ++ u64 blocks; ++ u64 mnt_id; ++ u32 dio_mem_align; ++ u32 dio_offset_align; ++ u64 change_cookie; ++}; ++ ++struct static_call_key { ++ void *func; ++}; ++ ++typedef struct cpumask cpumask_var_t[1]; ++ ++struct address_space; ++ ++struct page_pool; ++ ++struct dev_pagemap; ++ ++struct page { ++ long unsigned int flags; ++ union { ++ struct { ++ union { ++ struct list_head lru; ++ struct { ++ void *__filler; ++ unsigned int mlock_count; ++ }; ++ struct list_head buddy_list; ++ struct list_head pcp_list; ++ }; ++ struct address_space *mapping; ++ union { ++ long unsigned int index; ++ long unsigned int share; ++ }; ++ long unsigned int private; ++ }; ++ struct { ++ long unsigned int pp_magic; ++ struct page_pool *pp; ++ long unsigned int _pp_mapping_pad; ++ long unsigned int dma_addr; ++ union { ++ long unsigned int dma_addr_upper; ++ atomic_long_t pp_frag_count; ++ }; ++ }; ++ struct { ++ long unsigned int compound_head; ++ }; ++ struct { ++ long unsigned int _pt_pad_1; ++ pgtable_t pmd_huge_pte; ++ long unsigned int _pt_pad_2; ++ union { ++ struct mm_struct *pt_mm; ++ atomic_t pt_frag_refcount; ++ }; ++ spinlock_t ptl; ++ }; ++ struct { ++ struct dev_pagemap *pgmap; ++ void *zone_device_data; ++ }; ++ struct callback_head callback_head; ++ }; ++ union { ++ atomic_t _mapcount; ++ unsigned int page_type; ++ }; ++ atomic_t _refcount; ++ long unsigned int memcg_data; ++}; ++ ++struct runtime_instr_cb { ++ __u64 rca; ++ __u64 roa; ++ __u64 rla; ++ __u32 v: 1; ++ __u32 s: 1; ++ __u32 k: 1; ++ __u32 h: 1; ++ __u32 a: 1; ++ __u32 reserved1: 3; ++ __u32 ps: 1; ++ __u32 qs: 1; ++ __u32 pc: 1; ++ __u32 qc: 1; ++ __u32 reserved2: 1; ++ __u32 g: 1; ++ __u32 u: 1; ++ __u32 l: 1; ++ __u32 key: 4; ++ __u32 reserved3: 8; ++ __u32 t: 1; ++ __u32 rgs: 3; ++ __u32 m: 4; ++ __u32 n: 1; ++ __u32 mae: 1; ++ __u32 reserved4: 2; ++ __u32 c: 1; ++ __u32 r: 1; ++ __u32 b: 1; ++ __u32 j: 1; ++ __u32 e: 1; ++ __u32 x: 1; ++ __u32 reserved5: 2; ++ __u32 bpxn: 1; ++ __u32 bpxt: 1; ++ __u32 bpti: 1; ++ __u32 bpni: 1; ++ __u32 reserved6: 2; ++ __u32 d: 1; ++ __u32 f: 1; ++ __u32 ic: 4; ++ __u32 dc: 4; ++ __u64 reserved7; ++ __u64 sf; ++ __u64 rsic; ++ __u64 reserved8; ++}; ++ ++struct seq_operations { ++ void * (*start)(struct seq_file *, loff_t *); ++ void (*stop)(struct seq_file *, void *); ++ void * (*next)(struct seq_file *, void *, loff_t *); ++ int (*show)(struct seq_file *, void *); ++}; ++ ++struct gs_cb { ++ __u64 reserved; ++ __u64 gsd; ++ __u64 gssm; ++ __u64 gs_epl_a; ++}; ++ ++typedef struct { ++ arch_rwlock_t raw_lock; ++} rwlock_t; ++ ++typedef struct {} lockdep_map_p; ++ ++struct maple_tree { ++ union { ++ spinlock_t ma_lock; ++ lockdep_map_p ma_external_lock; ++ }; ++ void *ma_root; ++ unsigned int ma_flags; ++}; ++ ++struct wait_queue_entry; ++ ++typedef int (*wait_queue_func_t)(struct wait_queue_entry *, unsigned int, int, void *); ++ ++struct wait_queue_entry { ++ unsigned int flags; ++ void *private; ++ wait_queue_func_t func; ++ struct list_head entry; ++}; ++ ++typedef struct wait_queue_entry wait_queue_entry_t; ++ ++struct wait_queue_head { ++ spinlock_t lock; ++ struct list_head head; ++}; ++ ++typedef struct wait_queue_head wait_queue_head_t; ++ ++struct swait_queue_head { ++ raw_spinlock_t lock; ++ struct list_head task_list; ++}; ++ ++struct completion { ++ unsigned int done; ++ struct swait_queue_head wait; ++}; ++ ++struct percpu_counter { ++ raw_spinlock_t lock; ++ s64 count; ++ struct list_head list; ++ s32 *counters; ++}; ++ ++typedef struct { ++ spinlock_t lock; ++ cpumask_t cpu_attach_mask; ++ atomic_t flush_count; ++ unsigned int flush_mm; ++ struct list_head pgtable_list; ++ struct list_head gmap_list; ++ long unsigned int gmap_asce; ++ long unsigned int asce; ++ long unsigned int asce_limit; ++ long unsigned int vdso_base; ++ atomic_t protected_count; ++ unsigned int alloc_pgste: 1; ++ unsigned int has_pgste: 1; ++ unsigned int uses_skeys: 1; ++ unsigned int uses_cmm: 1; ++ unsigned int allow_gmap_hpage_1m: 1; ++} mm_context_t; ++ ++struct xol_area; ++ ++struct uprobes_state { ++ struct xol_area *xol_area; ++}; ++ ++struct mm_cid; ++ ++struct linux_binfmt; ++ ++struct kioctx_table; ++ ++struct mmu_notifier_subscriptions; ++ ++struct mm_struct { ++ struct { ++ struct { ++ atomic_t mm_count; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ }; ++ struct maple_tree mm_mt; ++ long unsigned int (*get_unmapped_area)(struct file *, long unsigned int, long unsigned int, long unsigned int, long unsigned int); ++ long unsigned int mmap_base; ++ long unsigned int mmap_legacy_base; ++ long unsigned int task_size; ++ pgd_t *pgd; ++ atomic_t membarrier_state; ++ atomic_t mm_users; ++ struct mm_cid *pcpu_cid; ++ long unsigned int mm_cid_next_scan; ++ atomic_long_t pgtables_bytes; ++ int map_count; ++ spinlock_t page_table_lock; ++ struct rw_semaphore mmap_lock; ++ struct list_head mmlist; ++ long unsigned int hiwater_rss; ++ long unsigned int hiwater_vm; ++ long unsigned int total_vm; ++ long unsigned int locked_vm; ++ atomic64_t pinned_vm; ++ long unsigned int data_vm; ++ long unsigned int exec_vm; ++ long unsigned int stack_vm; ++ long unsigned int def_flags; ++ seqcount_t write_protect_seq; ++ spinlock_t arg_lock; ++ long unsigned int start_code; ++ long unsigned int end_code; ++ long unsigned int start_data; ++ long unsigned int end_data; ++ long unsigned int start_brk; ++ long unsigned int brk; ++ long unsigned int start_stack; ++ long unsigned int arg_start; ++ long unsigned int arg_end; ++ long unsigned int env_start; ++ long unsigned int env_end; ++ long unsigned int saved_auxv[48]; ++ struct percpu_counter rss_stat[4]; ++ struct linux_binfmt *binfmt; ++ mm_context_t context; ++ long unsigned int flags; ++ spinlock_t ioctx_lock; ++ struct kioctx_table *ioctx_table; ++ struct task_struct *owner; ++ struct user_namespace *user_ns; ++ struct file *exe_file; ++ struct mmu_notifier_subscriptions *notifier_subscriptions; ++ long unsigned int numa_next_scan; ++ long unsigned int numa_scan_offset; ++ int numa_scan_seq; ++ atomic_t tlb_flush_pending; ++ struct uprobes_state uprobes_state; ++ atomic_long_t hugetlb_usage; ++ struct work_struct async_put_work; ++ long unsigned int ksm_merging_pages; ++ long unsigned int ksm_rmap_items; ++ struct { ++ struct list_head list; ++ long unsigned int bitmap; ++ struct mem_cgroup *memcg; ++ } lru_gen; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ }; ++ long unsigned int cpu_bitmap[0]; ++}; ++ ++struct workqueue_struct; ++ ++struct delayed_work { ++ struct work_struct work; ++ struct timer_list timer; ++ struct workqueue_struct *wq; ++ int cpu; ++}; ++ ++struct rcu_work { ++ struct work_struct work; ++ struct callback_head rcu; ++ struct workqueue_struct *wq; ++}; ++ ++struct rcu_segcblist { ++ struct callback_head *head; ++ struct callback_head **tails[4]; ++ long unsigned int gp_seq[4]; ++ long int len; ++ long int seglen[4]; ++ u8 flags; ++}; ++ ++struct srcu_node; ++ ++struct srcu_data { ++ atomic_long_t srcu_lock_count[2]; ++ atomic_long_t srcu_unlock_count[2]; ++ int srcu_nmi_safety; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ spinlock_t lock; ++ struct rcu_segcblist srcu_cblist; ++ long unsigned int srcu_gp_seq_needed; ++ long unsigned int srcu_gp_seq_needed_exp; ++ bool srcu_cblist_invoking; ++ struct timer_list delay_work; ++ struct work_struct work; ++ struct callback_head srcu_barrier_head; ++ struct srcu_node *mynode; ++ long unsigned int grpmask; ++ int cpu; ++ struct srcu_struct *ssp; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct srcu_node { ++ spinlock_t lock; ++ long unsigned int srcu_have_cbs[4]; ++ long unsigned int srcu_data_have_cbs[4]; ++ long unsigned int srcu_gp_seq_needed_exp; ++ struct srcu_node *srcu_parent; ++ int grplo; ++ int grphi; ++}; ++ ++struct srcu_usage; ++ ++struct srcu_struct { ++ unsigned int srcu_idx; ++ struct srcu_data *sda; ++ struct lockdep_map dep_map; ++ struct srcu_usage *srcu_sup; ++}; ++ ++struct srcu_usage { ++ struct srcu_node *node; ++ struct srcu_node *level[3]; ++ int srcu_size_state; ++ struct mutex srcu_cb_mutex; ++ spinlock_t lock; ++ struct mutex srcu_gp_mutex; ++ long unsigned int srcu_gp_seq; ++ long unsigned int srcu_gp_seq_needed; ++ long unsigned int srcu_gp_seq_needed_exp; ++ long unsigned int srcu_gp_start; ++ long unsigned int srcu_last_gp_end; ++ long unsigned int srcu_size_jiffies; ++ long unsigned int srcu_n_lock_retries; ++ long unsigned int srcu_n_exp_nodelay; ++ bool sda_is_static; ++ long unsigned int srcu_barrier_seq; ++ struct mutex srcu_barrier_mutex; ++ struct completion srcu_barrier_completion; ++ atomic_t srcu_barrier_cpu_cnt; ++ long unsigned int reschedule_jiffies; ++ long unsigned int reschedule_count; ++ struct delayed_work work; ++ struct srcu_struct *srcu_ssp; ++}; ++ ++struct notifier_block; ++ ++typedef int (*notifier_fn_t)(struct notifier_block *, long unsigned int, void *); ++ ++struct notifier_block { ++ notifier_fn_t notifier_call; ++ struct notifier_block *next; ++ int priority; ++}; ++ ++struct blocking_notifier_head { ++ struct rw_semaphore rwsem; ++ struct notifier_block *head; ++}; ++ ++struct raw_notifier_head { ++ struct notifier_block *head; ++}; ++ ++struct arch_uprobe_task {}; ++ ++enum uprobe_task_state { ++ UTASK_RUNNING = 0, ++ UTASK_SSTEP = 1, ++ UTASK_SSTEP_ACK = 2, ++ UTASK_SSTEP_TRAPPED = 3, ++}; ++ ++struct uprobe; ++ ++struct return_instance; ++ ++struct uprobe_task { ++ enum uprobe_task_state state; ++ union { ++ struct { ++ struct arch_uprobe_task autask; ++ long unsigned int vaddr; ++ }; ++ struct { ++ struct callback_head dup_xol_work; ++ long unsigned int dup_xol_addr; ++ }; ++ }; ++ struct uprobe *active_uprobe; ++ long unsigned int xol_vaddr; ++ struct return_instance *return_instances; ++ unsigned int depth; ++}; ++ ++struct return_instance { ++ struct uprobe *uprobe; ++ long unsigned int func; ++ long unsigned int stack; ++ long unsigned int orig_ret_vaddr; ++ bool chained; ++ struct return_instance *next; ++}; ++ ++struct seqcount_raw_spinlock { ++ seqcount_t seqcount; ++}; ++ ++typedef struct seqcount_raw_spinlock seqcount_raw_spinlock_t; ++ ++typedef struct { ++ seqcount_spinlock_t seqcount; ++ spinlock_t lock; ++} seqlock_t; ++ ++struct xarray { ++ spinlock_t xa_lock; ++ gfp_t xa_flags; ++ void *xa_head; ++}; ++ ++typedef u32 errseq_t; ++ ++struct address_space_operations; ++ ++struct address_space { ++ struct inode *host; ++ struct xarray i_pages; ++ struct rw_semaphore invalidate_lock; ++ gfp_t gfp_mask; ++ atomic_t i_mmap_writable; ++ struct rb_root_cached i_mmap; ++ struct rw_semaphore i_mmap_rwsem; ++ long unsigned int nrpages; ++ long unsigned int writeback_index; ++ const struct address_space_operations *a_ops; ++ long unsigned int flags; ++ errseq_t wb_err; ++ spinlock_t private_lock; ++ struct list_head private_list; ++ void *private_data; ++}; ++ ++enum dma_data_direction { ++ DMA_BIDIRECTIONAL = 0, ++ DMA_TO_DEVICE = 1, ++ DMA_FROM_DEVICE = 2, ++ DMA_NONE = 3, ++}; ++ ++struct device; ++ ++struct napi_struct; ++ ++struct page_pool_params { ++ unsigned int flags; ++ unsigned int order; ++ unsigned int pool_size; ++ int nid; ++ struct device *dev; ++ struct napi_struct *napi; ++ enum dma_data_direction dma_dir; ++ unsigned int max_len; ++ unsigned int offset; ++ void (*init_callback)(struct page *, void *); ++ void *init_arg; ++}; ++ ++struct page_pool_alloc_stats { ++ u64 fast; ++ u64 slow; ++ u64 slow_high_order; ++ u64 empty; ++ u64 refill; ++ u64 waive; ++}; ++ ++struct pp_alloc_cache { ++ u32 count; ++ struct page *cache[128]; ++}; ++ ++struct ptr_ring { ++ int producer; ++ spinlock_t producer_lock; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ int consumer_head; ++ int consumer_tail; ++ spinlock_t consumer_lock; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ int size; ++ int batch; ++ void **queue; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct page_pool_recycle_stats; ++ ++struct page_pool { ++ struct page_pool_params p; ++ struct delayed_work release_dw; ++ void (*disconnect)(void *); ++ long unsigned int defer_start; ++ long unsigned int defer_warn; ++ u32 pages_state_hold_cnt; ++ unsigned int frag_offset; ++ struct page *frag_page; ++ long int frag_users; ++ struct page_pool_alloc_stats alloc_stats; ++ u32 xdp_mem_id; ++ struct pp_alloc_cache alloc; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct ptr_ring ring; ++ struct page_pool_recycle_stats *recycle_stats; ++ atomic_t pages_state_release_cnt; ++ refcount_t user_cnt; ++ u64 destroy_cnt; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct vmem_altmap { ++ long unsigned int base_pfn; ++ const long unsigned int end_pfn; ++ const long unsigned int reserve; ++ long unsigned int free; ++ long unsigned int align; ++ long unsigned int alloc; ++}; ++ ++struct percpu_ref_data; ++ ++struct percpu_ref { ++ long unsigned int percpu_count_ptr; ++ struct percpu_ref_data *data; ++}; ++ ++enum memory_type { ++ MEMORY_DEVICE_PRIVATE = 1, ++ MEMORY_DEVICE_COHERENT = 2, ++ MEMORY_DEVICE_FS_DAX = 3, ++ MEMORY_DEVICE_GENERIC = 4, ++ MEMORY_DEVICE_PCI_P2PDMA = 5, ++}; ++ ++struct range { ++ u64 start; ++ u64 end; ++}; ++ ++struct dev_pagemap_ops; ++ ++struct dev_pagemap { ++ struct vmem_altmap altmap; ++ struct percpu_ref ref; ++ struct completion done; ++ enum memory_type type; ++ unsigned int flags; ++ long unsigned int vmemmap_shift; ++ const struct dev_pagemap_ops *ops; ++ void *owner; ++ int nr_range; ++ union { ++ struct range range; ++ struct { ++ struct {} __empty_ranges; ++ struct range ranges[0]; ++ }; ++ }; ++}; ++ ++struct folio { ++ union { ++ struct { ++ long unsigned int flags; ++ union { ++ struct list_head lru; ++ struct { ++ void *__filler; ++ unsigned int mlock_count; ++ }; ++ }; ++ struct address_space *mapping; ++ long unsigned int index; ++ void *private; ++ atomic_t _mapcount; ++ atomic_t _refcount; ++ long unsigned int memcg_data; ++ }; ++ struct page page; ++ }; ++ union { ++ struct { ++ long unsigned int _flags_1; ++ long unsigned int _head_1; ++ unsigned char _folio_dtor; ++ unsigned char _folio_order; ++ atomic_t _entire_mapcount; ++ atomic_t _nr_pages_mapped; ++ atomic_t _pincount; ++ unsigned int _folio_nr_pages; ++ }; ++ struct page __page_1; ++ }; ++ union { ++ struct { ++ long unsigned int _flags_2; ++ long unsigned int _head_2; ++ void *_hugetlb_subpool; ++ void *_hugetlb_cgroup; ++ void *_hugetlb_cgroup_rsvd; ++ void *_hugetlb_hwpoison; ++ }; ++ struct { ++ long unsigned int _flags_2a; ++ long unsigned int _head_2a; ++ struct list_head _deferred_list; ++ }; ++ struct page __page_2; ++ }; ++}; ++ ++typedef long unsigned int vm_flags_t; ++ ++struct vfsmount; ++ ++struct path { ++ struct vfsmount *mnt; ++ struct dentry *dentry; ++}; ++ ++enum pid_type { ++ PIDTYPE_PID = 0, ++ PIDTYPE_TGID = 1, ++ PIDTYPE_PGID = 2, ++ PIDTYPE_SID = 3, ++ PIDTYPE_MAX = 4, ++}; ++ ++struct fown_struct { ++ rwlock_t lock; ++ struct pid *pid; ++ enum pid_type pid_type; ++ kuid_t uid; ++ kuid_t euid; ++ int signum; ++}; ++ ++struct file_ra_state { ++ long unsigned int start; ++ unsigned int size; ++ unsigned int async_size; ++ unsigned int ra_pages; ++ unsigned int mmap_miss; ++ loff_t prev_pos; ++}; ++ ++struct file { ++ union { ++ struct llist_node f_llist; ++ struct callback_head f_rcuhead; ++ unsigned int f_iocb_flags; ++ }; ++ struct path f_path; ++ struct inode *f_inode; ++ const struct file_operations *f_op; ++ spinlock_t f_lock; ++ atomic_long_t f_count; ++ unsigned int f_flags; ++ fmode_t f_mode; ++ struct mutex f_pos_lock; ++ loff_t f_pos; ++ struct fown_struct f_owner; ++ const struct cred *f_cred; ++ struct file_ra_state f_ra; ++ u64 f_version; ++ void *f_security; ++ void *private_data; ++ struct hlist_head *f_ep; ++ struct address_space *f_mapping; ++ errseq_t f_wb_err; ++ errseq_t f_sb_err; ++}; ++ ++struct userfaultfd_ctx; ++ ++struct vm_userfaultfd_ctx { ++ struct userfaultfd_ctx *ctx; ++}; ++ ++struct vma_numab_state { ++ long unsigned int next_scan; ++ long unsigned int next_pid_reset; ++ long unsigned int access_pids[2]; ++}; ++ ++struct anon_vma; ++ ++struct vm_operations_struct; ++ ++struct vm_area_struct { ++ union { ++ struct { ++ long unsigned int vm_start; ++ long unsigned int vm_end; ++ }; ++ }; ++ struct mm_struct *vm_mm; ++ pgprot_t vm_page_prot; ++ union { ++ const vm_flags_t vm_flags; ++ vm_flags_t __vm_flags; ++ }; ++ struct { ++ struct rb_node rb; ++ long unsigned int rb_subtree_last; ++ } shared; ++ struct list_head anon_vma_chain; ++ struct anon_vma *anon_vma; ++ const struct vm_operations_struct *vm_ops; ++ long unsigned int vm_pgoff; ++ struct file *vm_file; ++ void *vm_private_data; ++ atomic_long_t swap_readahead_info; ++ struct mempolicy *vm_policy; ++ struct vma_numab_state *numab_state; ++ struct vm_userfaultfd_ctx vm_userfaultfd_ctx; ++}; ++ ++typedef unsigned int vm_fault_t; ++ ++enum page_entry_size { ++ PE_SIZE_PTE = 0, ++ PE_SIZE_PMD = 1, ++ PE_SIZE_PUD = 2, ++}; ++ ++struct vm_fault; ++ ++struct vm_operations_struct { ++ void (*open)(struct vm_area_struct *); ++ void (*close)(struct vm_area_struct *); ++ int (*may_split)(struct vm_area_struct *, long unsigned int); ++ int (*mremap)(struct vm_area_struct *); ++ int (*mprotect)(struct vm_area_struct *, long unsigned int, long unsigned int, long unsigned int); ++ vm_fault_t (*fault)(struct vm_fault *); ++ vm_fault_t (*huge_fault)(struct vm_fault *, enum page_entry_size); ++ vm_fault_t (*map_pages)(struct vm_fault *, long unsigned int, long unsigned int); ++ long unsigned int (*pagesize)(struct vm_area_struct *); ++ vm_fault_t (*page_mkwrite)(struct vm_fault *); ++ vm_fault_t (*pfn_mkwrite)(struct vm_fault *); ++ int (*access)(struct vm_area_struct *, long unsigned int, void *, int, int); ++ const char * (*name)(struct vm_area_struct *); ++ int (*set_policy)(struct vm_area_struct *, struct mempolicy *); ++ struct mempolicy * (*get_policy)(struct vm_area_struct *, long unsigned int); ++ struct page * (*find_special_page)(struct vm_area_struct *, long unsigned int); ++}; ++ ++struct mempolicy { ++ atomic_t refcnt; ++ short unsigned int mode; ++ short unsigned int flags; ++ nodemask_t nodes; ++ int home_node; ++ union { ++ nodemask_t cpuset_mems_allowed; ++ nodemask_t user_nodemask; ++ } w; ++}; ++ ++struct mm_cid { ++ u64 time; ++ int cid; ++}; ++ ++struct cgroup; ++ ++struct cgroup_subsys; ++ ++struct cgroup_subsys_state { ++ struct cgroup *cgroup; ++ struct cgroup_subsys *ss; ++ struct percpu_ref refcnt; ++ struct list_head sibling; ++ struct list_head children; ++ struct list_head rstat_css_node; ++ int id; ++ unsigned int flags; ++ u64 serial_nr; ++ atomic_t online_cnt; ++ struct work_struct destroy_work; ++ struct rcu_work destroy_rwork; ++ struct cgroup_subsys_state *parent; ++}; ++ ++struct mem_cgroup_id { ++ int id; ++ refcount_t ref; ++}; ++ ++struct page_counter { ++ atomic_long_t usage; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct cacheline_padding _pad1_; ++ long unsigned int emin; ++ atomic_long_t min_usage; ++ atomic_long_t children_min_usage; ++ long unsigned int elow; ++ atomic_long_t low_usage; ++ atomic_long_t children_low_usage; ++ long unsigned int watermark; ++ long unsigned int failcnt; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct cacheline_padding _pad2_; ++ long unsigned int min; ++ long unsigned int low; ++ long unsigned int high; ++ long unsigned int max; ++ struct page_counter *parent; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct vmpressure { ++ long unsigned int scanned; ++ long unsigned int reclaimed; ++ long unsigned int tree_scanned; ++ long unsigned int tree_reclaimed; ++ spinlock_t sr_lock; ++ struct list_head events; ++ struct mutex events_lock; ++ struct work_struct work; ++}; ++ ++struct cgroup_file { ++ struct kernfs_node *kn; ++ long unsigned int notified_at; ++ struct timer_list notify_timer; ++}; ++ ++struct mem_cgroup_threshold_ary; ++ ++struct mem_cgroup_thresholds { ++ struct mem_cgroup_threshold_ary *primary; ++ struct mem_cgroup_threshold_ary *spare; ++}; ++ ++struct fprop_global { ++ struct percpu_counter events; ++ unsigned int period; ++ seqcount_t sequence; ++}; ++ ++struct wb_domain { ++ spinlock_t lock; ++ struct fprop_global completions; ++ struct timer_list period_timer; ++ long unsigned int period_time; ++ long unsigned int dirty_limit_tstamp; ++ long unsigned int dirty_limit; ++}; ++ ++struct wb_completion { ++ atomic_t cnt; ++ wait_queue_head_t *waitq; ++}; ++ ++struct memcg_cgwb_frn { ++ u64 bdi_id; ++ int memcg_id; ++ u64 at; ++ struct wb_completion done; ++}; ++ ++struct lru_gen_mm_list { ++ struct list_head fifo; ++ spinlock_t lock; ++}; ++ ++struct memcg_vmstats; ++ ++struct obj_cgroup; ++ ++struct memcg_vmstats_percpu; ++ ++struct mem_cgroup_per_node; ++ ++struct mem_cgroup { ++ struct cgroup_subsys_state css; ++ struct mem_cgroup_id id; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct page_counter memory; ++ union { ++ struct page_counter swap; ++ struct page_counter memsw; ++ }; ++ struct page_counter kmem; ++ struct page_counter tcpmem; ++ struct work_struct high_work; ++ long unsigned int zswap_max; ++ long unsigned int soft_limit; ++ struct vmpressure vmpressure; ++ bool oom_group; ++ bool oom_lock; ++ int under_oom; ++ int swappiness; ++ int oom_kill_disable; ++ struct cgroup_file events_file; ++ struct cgroup_file events_local_file; ++ struct cgroup_file swap_events_file; ++ struct mutex thresholds_lock; ++ struct mem_cgroup_thresholds thresholds; ++ struct mem_cgroup_thresholds memsw_thresholds; ++ struct list_head oom_notify; ++ long unsigned int move_charge_at_immigrate; ++ spinlock_t move_lock; ++ long unsigned int move_lock_flags; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct cacheline_padding _pad1_; ++ struct memcg_vmstats *vmstats; ++ atomic_long_t memory_events[9]; ++ atomic_long_t memory_events_local[9]; ++ long unsigned int socket_pressure; ++ bool tcpmem_active; ++ int tcpmem_pressure; ++ int kmemcg_id; ++ struct obj_cgroup *objcg; ++ struct list_head objcg_list; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct cacheline_padding _pad2_; ++ atomic_t moving_account; ++ struct task_struct *move_lock_task; ++ struct memcg_vmstats_percpu *vmstats_percpu; ++ struct list_head cgwb_list; ++ struct wb_domain cgwb_domain; ++ struct memcg_cgwb_frn cgwb_frn[4]; ++ struct list_head event_list; ++ spinlock_t event_list_lock; ++ struct lru_gen_mm_list mm_list; ++ struct mem_cgroup_per_node *nodeinfo[0]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum fault_flag { ++ FAULT_FLAG_WRITE = 1, ++ FAULT_FLAG_MKWRITE = 2, ++ FAULT_FLAG_ALLOW_RETRY = 4, ++ FAULT_FLAG_RETRY_NOWAIT = 8, ++ FAULT_FLAG_KILLABLE = 16, ++ FAULT_FLAG_TRIED = 32, ++ FAULT_FLAG_USER = 64, ++ FAULT_FLAG_REMOTE = 128, ++ FAULT_FLAG_INSTRUCTION = 256, ++ FAULT_FLAG_INTERRUPTIBLE = 512, ++ FAULT_FLAG_UNSHARE = 1024, ++ FAULT_FLAG_ORIG_PTE_VALID = 2048, ++ FAULT_FLAG_VMA_LOCK = 4096, ++}; ++ ++struct vm_fault { ++ const struct { ++ struct vm_area_struct *vma; ++ gfp_t gfp_mask; ++ long unsigned int pgoff; ++ long unsigned int address; ++ long unsigned int real_address; ++ }; ++ enum fault_flag flags; ++ pmd_t *pmd; ++ pud_t *pud; ++ union { ++ pte_t orig_pte; ++ pmd_t orig_pmd; ++ }; ++ struct page *cow_page; ++ struct page *page; ++ pte_t *pte; ++ spinlock_t *ptl; ++ pgtable_t prealloc_pte; ++}; ++ ++struct hlist_nulls_node; ++ ++struct hlist_nulls_head { ++ struct hlist_nulls_node *first; ++}; ++ ++struct hlist_nulls_node { ++ struct hlist_nulls_node *next; ++ struct hlist_nulls_node **pprev; ++}; ++ ++enum pageflags { ++ PG_locked = 0, ++ PG_referenced = 1, ++ PG_uptodate = 2, ++ PG_dirty = 3, ++ PG_lru = 4, ++ PG_active = 5, ++ PG_workingset = 6, ++ PG_waiters = 7, ++ PG_error = 8, ++ PG_slab = 9, ++ PG_owner_priv_1 = 10, ++ PG_arch_1 = 11, ++ PG_reserved = 12, ++ PG_private = 13, ++ PG_private_2 = 14, ++ PG_writeback = 15, ++ PG_head = 16, ++ PG_mappedtodisk = 17, ++ PG_reclaim = 18, ++ PG_swapbacked = 19, ++ PG_unevictable = 20, ++ PG_mlocked = 21, ++ PG_young = 22, ++ PG_idle = 23, ++ __NR_PAGEFLAGS = 24, ++ PG_readahead = 18, ++ PG_anon_exclusive = 17, ++ PG_checked = 10, ++ PG_swapcache = 10, ++ PG_fscache = 14, ++ PG_pinned = 10, ++ PG_savepinned = 3, ++ PG_foreign = 10, ++ PG_xen_remapped = 10, ++ PG_isolated = 18, ++ PG_reported = 2, ++ PG_vmemmap_self_hosted = 10, ++}; ++ ++struct free_area { ++ struct list_head free_list[6]; ++ long unsigned int nr_free; ++}; ++ ++struct lru_gen_folio { ++ long unsigned int max_seq; ++ long unsigned int min_seq[2]; ++ long unsigned int timestamps[4]; ++ struct list_head folios[24]; ++ long int nr_pages[24]; ++ long unsigned int avg_refaulted[8]; ++ long unsigned int avg_total[8]; ++ long unsigned int protected[6]; ++ atomic_long_t evicted[8]; ++ atomic_long_t refaulted[8]; ++ bool enabled; ++ u8 gen; ++ u8 seg; ++ struct hlist_nulls_node list; ++}; ++ ++struct lru_gen_mm_state { ++ long unsigned int seq; ++ struct list_head *head; ++ struct list_head *tail; ++ long unsigned int *filters[2]; ++ long unsigned int stats[6]; ++}; ++ ++struct lruvec; ++ ++struct lru_gen_mm_walk { ++ struct lruvec *lruvec; ++ long unsigned int max_seq; ++ long unsigned int next_addr; ++ int nr_pages[24]; ++ int mm_stats[6]; ++ int batched; ++ bool can_swap; ++ bool force_scan; ++}; ++ ++struct pglist_data; ++ ++struct lruvec { ++ struct list_head lists[5]; ++ spinlock_t lru_lock; ++ long unsigned int anon_cost; ++ long unsigned int file_cost; ++ atomic_long_t nonresident_age; ++ long unsigned int refaults[2]; ++ long unsigned int flags; ++ struct lru_gen_folio lrugen; ++ struct lru_gen_mm_state mm_state; ++ struct pglist_data *pgdat; ++}; ++ ++struct lru_gen_memcg { ++ long unsigned int seq; ++ long unsigned int nr_memcgs[2]; ++ struct hlist_nulls_head fifo[16]; ++ spinlock_t lock; ++}; ++ ++struct per_cpu_pages; ++ ++struct per_cpu_zonestat; ++ ++struct zone { ++ long unsigned int _watermark[4]; ++ long unsigned int watermark_boost; ++ long unsigned int nr_reserved_highatomic; ++ long int lowmem_reserve[3]; ++ int node; ++ struct pglist_data *zone_pgdat; ++ struct per_cpu_pages *per_cpu_pageset; ++ struct per_cpu_zonestat *per_cpu_zonestats; ++ int pageset_high; ++ int pageset_batch; ++ long unsigned int zone_start_pfn; ++ atomic_long_t managed_pages; ++ long unsigned int spanned_pages; ++ long unsigned int present_pages; ++ long unsigned int present_early_pages; ++ long unsigned int cma_pages; ++ const char *name; ++ long unsigned int nr_isolate_pageblock; ++ seqlock_t span_seqlock; ++ int initialized; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct cacheline_padding _pad1_; ++ struct free_area free_area[11]; ++ long unsigned int flags; ++ spinlock_t lock; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct cacheline_padding _pad2_; ++ long unsigned int percpu_drift_mark; ++ long unsigned int compact_cached_free_pfn; ++ long unsigned int compact_cached_migrate_pfn[2]; ++ long unsigned int compact_init_migrate_pfn; ++ long unsigned int compact_init_free_pfn; ++ unsigned int compact_considered; ++ unsigned int compact_defer_shift; ++ int compact_order_failed; ++ bool compact_blockskip_flush; ++ bool contiguous; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct cacheline_padding _pad3_; ++ atomic_long_t vm_stat[11]; ++ atomic_long_t vm_numa_event[6]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct zoneref { ++ struct zone *zone; ++ int zone_idx; ++}; ++ ++struct zonelist { ++ struct zoneref _zonerefs[7]; ++}; ++ ++enum zone_type { ++ ZONE_DMA = 0, ++ ZONE_NORMAL = 1, ++ ZONE_MOVABLE = 2, ++ __MAX_NR_ZONES = 3, ++}; ++ ++struct per_cpu_nodestat; ++ ++struct memory_tier; ++ ++struct pglist_data { ++ struct zone node_zones[3]; ++ struct zonelist node_zonelists[2]; ++ int nr_zones; ++ spinlock_t node_size_lock; ++ long unsigned int node_start_pfn; ++ long unsigned int node_present_pages; ++ long unsigned int node_spanned_pages; ++ int node_id; ++ wait_queue_head_t kswapd_wait; ++ wait_queue_head_t pfmemalloc_wait; ++ wait_queue_head_t reclaim_wait[4]; ++ atomic_t nr_writeback_throttled; ++ long unsigned int nr_reclaim_start; ++ struct mutex kswapd_lock; ++ struct task_struct *kswapd; ++ int kswapd_order; ++ enum zone_type kswapd_highest_zoneidx; ++ int kswapd_failures; ++ int kcompactd_max_order; ++ enum zone_type kcompactd_highest_zoneidx; ++ wait_queue_head_t kcompactd_wait; ++ struct task_struct *kcompactd; ++ bool proactive_compact_trigger; ++ long unsigned int totalreserve_pages; ++ long unsigned int min_unmapped_pages; ++ long unsigned int min_slab_pages; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct cacheline_padding _pad1_; ++ unsigned int nbp_rl_start; ++ long unsigned int nbp_rl_nr_cand; ++ unsigned int nbp_threshold; ++ unsigned int nbp_th_start; ++ long unsigned int nbp_th_nr_cand; ++ struct lruvec __lruvec; ++ long unsigned int flags; ++ struct lru_gen_mm_walk mm_walk; ++ struct lru_gen_memcg memcg_lru; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct cacheline_padding _pad2_; ++ struct per_cpu_nodestat *per_cpu_nodestats; ++ atomic_long_t vm_stat[43]; ++ struct memory_tier *memtier; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct per_cpu_pages { ++ spinlock_t lock; ++ int count; ++ int high; ++ int batch; ++ short int free_factor; ++ short int expire; ++ struct list_head lists[12]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct per_cpu_zonestat { ++ s8 vm_stat_diff[11]; ++ s8 stat_threshold; ++ long unsigned int vm_numa_event[6]; ++}; ++ ++struct per_cpu_nodestat { ++ s8 stat_threshold; ++ s8 vm_node_stat_diff[43]; ++}; ++ ++typedef int proc_handler(struct ctl_table *, int, void *, size_t *, loff_t *); ++ ++struct ctl_table_poll; ++ ++struct ctl_table { ++ const char *procname; ++ void *data; ++ int maxlen; ++ umode_t mode; ++ struct ctl_table *child; ++ proc_handler *proc_handler; ++ struct ctl_table_poll *poll; ++ void *extra1; ++ void *extra2; ++}; ++ ++struct ctl_table_poll { ++ atomic_t event; ++ wait_queue_head_t wait; ++}; ++ ++struct ctl_node { ++ struct rb_node node; ++ struct ctl_table_header *header; ++}; ++ ++struct ctl_table_root { ++ struct ctl_table_set default_set; ++ struct ctl_table_set * (*lookup)(struct ctl_table_root *); ++ void (*set_ownership)(struct ctl_table_header *, struct ctl_table *, kuid_t *, kgid_t *); ++ int (*permissions)(struct ctl_table_header *, struct ctl_table *); ++}; ++ ++typedef struct { ++ u64 val; ++} kernel_cap_t; ++ ++struct user_struct; ++ ++struct group_info; ++ ++struct cred { ++ atomic_t usage; ++ kuid_t uid; ++ kgid_t gid; ++ kuid_t suid; ++ kgid_t sgid; ++ kuid_t euid; ++ kgid_t egid; ++ kuid_t fsuid; ++ kgid_t fsgid; ++ unsigned int securebits; ++ kernel_cap_t cap_inheritable; ++ kernel_cap_t cap_permitted; ++ kernel_cap_t cap_effective; ++ kernel_cap_t cap_bset; ++ kernel_cap_t cap_ambient; ++ unsigned char jit_keyring; ++ struct key *session_keyring; ++ struct key *process_keyring; ++ struct key *thread_keyring; ++ struct key *request_key_auth; ++ void *security; ++ struct user_struct *user; ++ struct user_namespace *user_ns; ++ struct ucounts *ucounts; ++ struct group_info *group_info; ++ union { ++ int non_rcu; ++ struct callback_head rcu; ++ }; ++}; ++ ++struct rhash_head { ++ struct rhash_head *next; ++}; ++ ++struct rhashtable; ++ ++struct rhashtable_compare_arg { ++ struct rhashtable *ht; ++ const void *key; ++}; ++ ++typedef u32 (*rht_hashfn_t)(const void *, u32, u32); ++ ++typedef u32 (*rht_obj_hashfn_t)(const void *, u32, u32); ++ ++typedef int (*rht_obj_cmpfn_t)(struct rhashtable_compare_arg *, const void *); ++ ++struct rhashtable_params { ++ u16 nelem_hint; ++ u16 key_len; ++ u16 key_offset; ++ u16 head_offset; ++ unsigned int max_size; ++ u16 min_size; ++ bool automatic_shrinking; ++ rht_hashfn_t hashfn; ++ rht_obj_hashfn_t obj_hashfn; ++ rht_obj_cmpfn_t obj_cmpfn; ++}; ++ ++struct bucket_table; ++ ++struct rhashtable { ++ struct bucket_table *tbl; ++ unsigned int key_len; ++ unsigned int max_elems; ++ struct rhashtable_params p; ++ bool rhlist; ++ struct work_struct run_work; ++ struct mutex mutex; ++ spinlock_t lock; ++ atomic_t nelems; ++}; ++ ++struct iovec { ++ void *iov_base; ++ __kernel_size_t iov_len; ++}; ++ ++struct kvec { ++ void *iov_base; ++ size_t iov_len; ++}; ++ ++struct bio_vec { ++ struct page *bv_page; ++ unsigned int bv_len; ++ unsigned int bv_offset; ++}; ++ ++struct iov_iter { ++ u8 iter_type; ++ bool copy_mc; ++ bool nofault; ++ bool data_source; ++ bool user_backed; ++ union { ++ size_t iov_offset; ++ int last_offset; ++ }; ++ union { ++ struct iovec __ubuf_iovec; ++ struct { ++ union { ++ const struct iovec *__iov; ++ const struct kvec *kvec; ++ const struct bio_vec *bvec; ++ struct xarray *xarray; ++ struct pipe_inode_info *pipe; ++ void *ubuf; ++ }; ++ size_t count; ++ }; ++ }; ++ union { ++ long unsigned int nr_segs; ++ struct { ++ unsigned int head; ++ unsigned int start_head; ++ }; ++ loff_t xarray_start; ++ }; ++}; ++ ++typedef short unsigned int __kernel_sa_family_t; ++ ++typedef __kernel_sa_family_t sa_family_t; ++ ++struct sockaddr { ++ sa_family_t sa_family; ++ union { ++ char sa_data_min[14]; ++ struct { ++ struct {} __empty_sa_data; ++ char sa_data[0]; ++ }; ++ }; ++}; ++ ++struct ubuf_info; ++ ++struct sock; ++ ++struct sk_buff; ++ ++struct msghdr { ++ void *msg_name; ++ int msg_namelen; ++ int msg_inq; ++ struct iov_iter msg_iter; ++ union { ++ void *msg_control; ++ void *msg_control_user; ++ }; ++ bool msg_control_is_user: 1; ++ bool msg_get_inq: 1; ++ unsigned int msg_flags; ++ __kernel_size_t msg_controllen; ++ struct kiocb *msg_iocb; ++ struct ubuf_info *msg_ubuf; ++ int (*sg_from_iter)(struct sock *, struct sk_buff *, struct iov_iter *, size_t); ++}; ++ ++struct wait_page_queue; ++ ++struct kiocb { ++ struct file *ki_filp; ++ loff_t ki_pos; ++ void (*ki_complete)(struct kiocb *, long int); ++ void *private; ++ int ki_flags; ++ u16 ki_ioprio; ++ struct wait_page_queue *ki_waitq; ++}; ++ ++struct ubuf_info { ++ void (*callback)(struct sk_buff *, struct ubuf_info *, bool); ++ refcount_t refcnt; ++ u8 flags; ++}; ++ ++typedef __u64 __addrpair; ++ ++typedef __u32 __portpair; ++ ++struct net; ++ ++typedef struct { ++ struct net *net; ++} possible_net_t; ++ ++struct in6_addr { ++ union { ++ __u8 u6_addr8[16]; ++ __be16 u6_addr16[8]; ++ __be32 u6_addr32[4]; ++ } in6_u; ++}; ++ ++struct proto; ++ ++struct inet_timewait_death_row; ++ ++struct sock_common { ++ union { ++ __addrpair skc_addrpair; ++ struct { ++ __be32 skc_daddr; ++ __be32 skc_rcv_saddr; ++ }; ++ }; ++ union { ++ unsigned int skc_hash; ++ __u16 skc_u16hashes[2]; ++ }; ++ union { ++ __portpair skc_portpair; ++ struct { ++ __be16 skc_dport; ++ __u16 skc_num; ++ }; ++ }; ++ short unsigned int skc_family; ++ volatile unsigned char skc_state; ++ unsigned char skc_reuse: 4; ++ unsigned char skc_reuseport: 1; ++ unsigned char skc_ipv6only: 1; ++ unsigned char skc_net_refcnt: 1; ++ int skc_bound_dev_if; ++ union { ++ struct hlist_node skc_bind_node; ++ struct hlist_node skc_portaddr_node; ++ }; ++ struct proto *skc_prot; ++ possible_net_t skc_net; ++ struct in6_addr skc_v6_daddr; ++ struct in6_addr skc_v6_rcv_saddr; ++ atomic64_t skc_cookie; ++ union { ++ long unsigned int skc_flags; ++ struct sock *skc_listener; ++ struct inet_timewait_death_row *skc_tw_dr; ++ }; ++ int skc_dontcopy_begin[0]; ++ union { ++ struct hlist_node skc_node; ++ struct hlist_nulls_node skc_nulls_node; ++ }; ++ short unsigned int skc_tx_queue_mapping; ++ short unsigned int skc_rx_queue_mapping; ++ union { ++ int skc_incoming_cpu; ++ u32 skc_rcv_wnd; ++ u32 skc_tw_rcv_nxt; ++ }; ++ refcount_t skc_refcnt; ++ int skc_dontcopy_end[0]; ++ union { ++ u32 skc_rxhash; ++ u32 skc_window_clamp; ++ u32 skc_tw_snd_nxt; ++ }; ++}; ++ ++typedef struct { ++ spinlock_t slock; ++ int owned; ++ wait_queue_head_t wq; ++} socket_lock_t; ++ ++struct sk_buff_list { ++ struct sk_buff *next; ++ struct sk_buff *prev; ++}; ++ ++struct sk_buff_head { ++ union { ++ struct { ++ struct sk_buff *next; ++ struct sk_buff *prev; ++ }; ++ struct sk_buff_list list; ++ }; ++ __u32 qlen; ++ spinlock_t lock; ++}; ++ ++typedef u64 netdev_features_t; ++ ++struct sock_cgroup_data { ++ struct cgroup *cgroup; ++ u32 classid; ++ u16 prioidx; ++}; ++ ++typedef struct {} netns_tracker; ++ ++struct dst_entry; ++ ++struct sk_filter; ++ ++struct socket_wq; ++ ++struct xfrm_policy; ++ ++struct socket; ++ ++struct net_device; ++ ++struct sock_reuseport; ++ ++struct sock { ++ struct sock_common __sk_common; ++ struct dst_entry *sk_rx_dst; ++ int sk_rx_dst_ifindex; ++ u32 sk_rx_dst_cookie; ++ socket_lock_t sk_lock; ++ atomic_t sk_drops; ++ int sk_rcvlowat; ++ struct sk_buff_head sk_error_queue; ++ struct sk_buff_head sk_receive_queue; ++ struct { ++ atomic_t rmem_alloc; ++ int len; ++ struct sk_buff *head; ++ struct sk_buff *tail; ++ } sk_backlog; ++ int sk_forward_alloc; ++ u32 sk_reserved_mem; ++ unsigned int sk_ll_usec; ++ unsigned int sk_napi_id; ++ int sk_rcvbuf; ++ int sk_wait_pending; ++ struct sk_filter *sk_filter; ++ union { ++ struct socket_wq *sk_wq; ++ struct socket_wq *sk_wq_raw; ++ }; ++ struct xfrm_policy *sk_policy[2]; ++ struct dst_entry *sk_dst_cache; ++ atomic_t sk_omem_alloc; ++ int sk_sndbuf; ++ int sk_wmem_queued; ++ refcount_t sk_wmem_alloc; ++ long unsigned int sk_tsq_flags; ++ union { ++ struct sk_buff *sk_send_head; ++ struct rb_root tcp_rtx_queue; ++ }; ++ struct sk_buff_head sk_write_queue; ++ __s32 sk_peek_off; ++ int sk_write_pending; ++ __u32 sk_dst_pending_confirm; ++ u32 sk_pacing_status; ++ long int sk_sndtimeo; ++ struct timer_list sk_timer; ++ __u32 sk_priority; ++ __u32 sk_mark; ++ long unsigned int sk_pacing_rate; ++ long unsigned int sk_max_pacing_rate; ++ struct page_frag sk_frag; ++ netdev_features_t sk_route_caps; ++ int sk_gso_type; ++ unsigned int sk_gso_max_size; ++ gfp_t sk_allocation; ++ __u32 sk_txhash; ++ u8 sk_gso_disabled: 1; ++ u8 sk_kern_sock: 1; ++ u8 sk_no_check_tx: 1; ++ u8 sk_no_check_rx: 1; ++ u8 sk_userlocks: 4; ++ u8 sk_pacing_shift; ++ u16 sk_type; ++ u16 sk_protocol; ++ u16 sk_gso_max_segs; ++ long unsigned int sk_lingertime; ++ struct proto *sk_prot_creator; ++ rwlock_t sk_callback_lock; ++ int sk_err; ++ int sk_err_soft; ++ u32 sk_ack_backlog; ++ u32 sk_max_ack_backlog; ++ kuid_t sk_uid; ++ u8 sk_txrehash; ++ u8 sk_prefer_busy_poll; ++ u16 sk_busy_poll_budget; ++ spinlock_t sk_peer_lock; ++ int sk_bind_phc; ++ struct pid *sk_peer_pid; ++ const struct cred *sk_peer_cred; ++ long int sk_rcvtimeo; ++ ktime_t sk_stamp; ++ atomic_t sk_tskey; ++ atomic_t sk_zckey; ++ u32 sk_tsflags; ++ u8 sk_shutdown; ++ u8 sk_clockid; ++ u8 sk_txtime_deadline_mode: 1; ++ u8 sk_txtime_report_errors: 1; ++ u8 sk_txtime_unused: 6; ++ bool sk_use_task_frag; ++ struct socket *sk_socket; ++ void *sk_user_data; ++ void *sk_security; ++ struct sock_cgroup_data sk_cgrp_data; ++ struct mem_cgroup *sk_memcg; ++ void (*sk_state_change)(struct sock *); ++ void (*sk_data_ready)(struct sock *); ++ void (*sk_write_space)(struct sock *); ++ void (*sk_error_report)(struct sock *); ++ int (*sk_backlog_rcv)(struct sock *, struct sk_buff *); ++ struct sk_buff * (*sk_validate_xmit_skb)(struct sock *, struct net_device *, struct sk_buff *); ++ void (*sk_destruct)(struct sock *); ++ struct sock_reuseport *sk_reuseport_cb; ++ struct bpf_local_storage *sk_bpf_storage; ++ struct callback_head sk_rcu; ++ netns_tracker ns_tracker; ++ struct hlist_node sk_bind2_node; ++}; ++ ++typedef unsigned int sk_buff_data_t; ++ ++struct skb_ext; ++ ++struct sk_buff { ++ union { ++ struct { ++ struct sk_buff *next; ++ struct sk_buff *prev; ++ union { ++ struct net_device *dev; ++ long unsigned int dev_scratch; ++ }; ++ }; ++ struct rb_node rbnode; ++ struct list_head list; ++ struct llist_node ll_node; ++ }; ++ union { ++ struct sock *sk; ++ int ip_defrag_offset; ++ }; ++ union { ++ ktime_t tstamp; ++ u64 skb_mstamp_ns; ++ }; ++ char cb[48]; ++ union { ++ struct { ++ long unsigned int _skb_refdst; ++ void (*destructor)(struct sk_buff *); ++ }; ++ struct list_head tcp_tsorted_anchor; ++ long unsigned int _sk_redir; ++ }; ++ long unsigned int _nfct; ++ unsigned int len; ++ unsigned int data_len; ++ __u16 mac_len; ++ __u16 hdr_len; ++ __u16 queue_mapping; ++ __u8 __cloned_offset[0]; ++ __u8 cloned: 1; ++ __u8 nohdr: 1; ++ __u8 fclone: 2; ++ __u8 peeked: 1; ++ __u8 head_frag: 1; ++ __u8 pfmemalloc: 1; ++ __u8 pp_recycle: 1; ++ __u8 active_extensions; ++ union { ++ struct { ++ __u8 __pkt_type_offset[0]; ++ __u8 pkt_type: 3; ++ __u8 ignore_df: 1; ++ __u8 dst_pending_confirm: 1; ++ __u8 ip_summed: 2; ++ __u8 ooo_okay: 1; ++ __u8 __mono_tc_offset[0]; ++ __u8 mono_delivery_time: 1; ++ __u8 tc_at_ingress: 1; ++ __u8 tc_skip_classify: 1; ++ __u8 remcsum_offload: 1; ++ __u8 csum_complete_sw: 1; ++ __u8 csum_level: 2; ++ __u8 inner_protocol_type: 1; ++ __u8 l4_hash: 1; ++ __u8 sw_hash: 1; ++ __u8 no_fcs: 1; ++ __u8 encapsulation: 1; ++ __u8 encap_hdr_csum: 1; ++ __u8 csum_valid: 1; ++ __u8 ndisc_nodetype: 2; ++ __u8 ipvs_property: 1; ++ __u8 nf_trace: 1; ++ __u8 offload_fwd_mark: 1; ++ __u8 offload_l3_fwd_mark: 1; ++ __u8 redirected: 1; ++ __u8 from_ingress: 1; ++ __u8 nf_skip_egress: 1; ++ __u8 decrypted: 1; ++ __u8 slow_gro: 1; ++ __u8 csum_not_inet: 1; ++ __u16 tc_index; ++ u16 alloc_cpu; ++ union { ++ __wsum csum; ++ struct { ++ __u16 csum_start; ++ __u16 csum_offset; ++ }; ++ }; ++ __u32 priority; ++ int skb_iif; ++ __u32 hash; ++ union { ++ u32 vlan_all; ++ struct { ++ __be16 vlan_proto; ++ __u16 vlan_tci; ++ }; ++ }; ++ union { ++ unsigned int napi_id; ++ unsigned int sender_cpu; ++ }; ++ __u32 secmark; ++ union { ++ __u32 mark; ++ __u32 reserved_tailroom; ++ }; ++ union { ++ __be16 inner_protocol; ++ __u8 inner_ipproto; ++ }; ++ __u16 inner_transport_header; ++ __u16 inner_network_header; ++ __u16 inner_mac_header; ++ __be16 protocol; ++ __u16 transport_header; ++ __u16 network_header; ++ __u16 mac_header; ++ }; ++ struct { ++ __u8 __pkt_type_offset[0]; ++ __u8 pkt_type: 3; ++ __u8 ignore_df: 1; ++ __u8 dst_pending_confirm: 1; ++ __u8 ip_summed: 2; ++ __u8 ooo_okay: 1; ++ __u8 __mono_tc_offset[0]; ++ __u8 mono_delivery_time: 1; ++ __u8 tc_at_ingress: 1; ++ __u8 tc_skip_classify: 1; ++ __u8 remcsum_offload: 1; ++ __u8 csum_complete_sw: 1; ++ __u8 csum_level: 2; ++ __u8 inner_protocol_type: 1; ++ __u8 l4_hash: 1; ++ __u8 sw_hash: 1; ++ __u8 no_fcs: 1; ++ __u8 encapsulation: 1; ++ __u8 encap_hdr_csum: 1; ++ __u8 csum_valid: 1; ++ __u8 ndisc_nodetype: 2; ++ __u8 ipvs_property: 1; ++ __u8 nf_trace: 1; ++ __u8 offload_fwd_mark: 1; ++ __u8 offload_l3_fwd_mark: 1; ++ __u8 redirected: 1; ++ __u8 from_ingress: 1; ++ __u8 nf_skip_egress: 1; ++ __u8 decrypted: 1; ++ __u8 slow_gro: 1; ++ __u8 csum_not_inet: 1; ++ __u16 tc_index; ++ u16 alloc_cpu; ++ union { ++ __wsum csum; ++ struct { ++ __u16 csum_start; ++ __u16 csum_offset; ++ }; ++ }; ++ __u32 priority; ++ int skb_iif; ++ __u32 hash; ++ union { ++ u32 vlan_all; ++ struct { ++ __be16 vlan_proto; ++ __u16 vlan_tci; ++ }; ++ }; ++ union { ++ unsigned int napi_id; ++ unsigned int sender_cpu; ++ }; ++ __u32 secmark; ++ union { ++ __u32 mark; ++ __u32 reserved_tailroom; ++ }; ++ union { ++ __be16 inner_protocol; ++ __u8 inner_ipproto; ++ }; ++ __u16 inner_transport_header; ++ __u16 inner_network_header; ++ __u16 inner_mac_header; ++ __be16 protocol; ++ __u16 transport_header; ++ __u16 network_header; ++ __u16 mac_header; ++ } headers; ++ }; ++ sk_buff_data_t tail; ++ sk_buff_data_t end; ++ unsigned char *head; ++ unsigned char *data; ++ unsigned int truesize; ++ refcount_t users; ++ struct skb_ext *extensions; ++}; ++ ++typedef struct { ++ unsigned int clock_rate; ++ unsigned int clock_type; ++ short unsigned int loopback; ++} sync_serial_settings; ++ ++typedef struct { ++ unsigned int clock_rate; ++ unsigned int clock_type; ++ short unsigned int loopback; ++ unsigned int slot_map; ++} te1_settings; ++ ++typedef struct { ++ short unsigned int encoding; ++ short unsigned int parity; ++} raw_hdlc_proto; ++ ++typedef struct { ++ unsigned int t391; ++ unsigned int t392; ++ unsigned int n391; ++ unsigned int n392; ++ unsigned int n393; ++ short unsigned int lmi; ++ short unsigned int dce; ++} fr_proto; ++ ++typedef struct { ++ unsigned int dlci; ++} fr_proto_pvc; ++ ++typedef struct { ++ unsigned int dlci; ++ char master[16]; ++} fr_proto_pvc_info; ++ ++typedef struct { ++ unsigned int interval; ++ unsigned int timeout; ++} cisco_proto; ++ ++typedef struct { ++ short unsigned int dce; ++ unsigned int modulo; ++ unsigned int window; ++ unsigned int t1; ++ unsigned int t2; ++ unsigned int n2; ++} x25_hdlc_proto; ++ ++struct ifmap { ++ long unsigned int mem_start; ++ long unsigned int mem_end; ++ short unsigned int base_addr; ++ unsigned char irq; ++ unsigned char dma; ++ unsigned char port; ++}; ++ ++struct if_settings { ++ unsigned int type; ++ unsigned int size; ++ union { ++ raw_hdlc_proto *raw_hdlc; ++ cisco_proto *cisco; ++ fr_proto *fr; ++ fr_proto_pvc *fr_pvc; ++ fr_proto_pvc_info *fr_pvc_info; ++ x25_hdlc_proto *x25; ++ sync_serial_settings *sync; ++ te1_settings *te1; ++ } ifs_ifsu; ++}; ++ ++struct ifreq { ++ union { ++ char ifrn_name[16]; ++ } ifr_ifrn; ++ union { ++ struct sockaddr ifru_addr; ++ struct sockaddr ifru_dstaddr; ++ struct sockaddr ifru_broadaddr; ++ struct sockaddr ifru_netmask; ++ struct sockaddr ifru_hwaddr; ++ short int ifru_flags; ++ int ifru_ivalue; ++ int ifru_mtu; ++ struct ifmap ifru_map; ++ char ifru_slave[16]; ++ char ifru_newname[16]; ++ void *ifru_data; ++ struct if_settings ifru_settings; ++ } ifr_ifru; ++}; ++ ++struct hlist_bl_node; ++ ++struct hlist_bl_head { ++ struct hlist_bl_node *first; ++}; ++ ++struct hlist_bl_node { ++ struct hlist_bl_node *next; ++ struct hlist_bl_node **pprev; ++}; ++ ++struct lockref { ++ union { ++ __u64 lock_count; ++ struct { ++ spinlock_t lock; ++ int count; ++ }; ++ }; ++}; ++ ++struct qstr { ++ union { ++ struct { ++ u32 len; ++ u32 hash; ++ }; ++ u64 hash_len; ++ }; ++ const unsigned char *name; ++}; ++ ++struct dentry_operations; ++ ++struct dentry { ++ unsigned int d_flags; ++ seqcount_spinlock_t d_seq; ++ struct hlist_bl_node d_hash; ++ struct dentry *d_parent; ++ struct qstr d_name; ++ struct inode *d_inode; ++ unsigned char d_iname[32]; ++ struct lockref d_lockref; ++ const struct dentry_operations *d_op; ++ struct super_block *d_sb; ++ long unsigned int d_time; ++ void *d_fsdata; ++ union { ++ struct list_head d_lru; ++ wait_queue_head_t *d_wait; ++ }; ++ struct list_head d_child; ++ struct list_head d_subdirs; ++ union { ++ struct hlist_node d_alias; ++ struct hlist_bl_node d_in_lookup_hash; ++ struct callback_head d_rcu; ++ } d_u; ++}; ++ ++struct posix_acl; ++ ++struct inode_operations; ++ ++struct bdi_writeback; ++ ++struct file_lock_context; ++ ++struct cdev; ++ ++struct fsnotify_mark_connector; ++ ++struct fscrypt_info; ++ ++struct fsverity_info; ++ ++struct inode { ++ umode_t i_mode; ++ short unsigned int i_opflags; ++ kuid_t i_uid; ++ kgid_t i_gid; ++ unsigned int i_flags; ++ struct posix_acl *i_acl; ++ struct posix_acl *i_default_acl; ++ const struct inode_operations *i_op; ++ struct super_block *i_sb; ++ struct address_space *i_mapping; ++ void *i_security; ++ long unsigned int i_ino; ++ union { ++ const unsigned int i_nlink; ++ unsigned int __i_nlink; ++ }; ++ dev_t i_rdev; ++ loff_t i_size; ++ struct timespec64 i_atime; ++ struct timespec64 i_mtime; ++ struct timespec64 i_ctime; ++ spinlock_t i_lock; ++ short unsigned int i_bytes; ++ u8 i_blkbits; ++ u8 i_write_hint; ++ blkcnt_t i_blocks; ++ long unsigned int i_state; ++ struct rw_semaphore i_rwsem; ++ long unsigned int dirtied_when; ++ long unsigned int dirtied_time_when; ++ struct hlist_node i_hash; ++ struct list_head i_io_list; ++ struct bdi_writeback *i_wb; ++ int i_wb_frn_winner; ++ u16 i_wb_frn_avg_time; ++ u16 i_wb_frn_history; ++ struct list_head i_lru; ++ struct list_head i_sb_list; ++ struct list_head i_wb_list; ++ union { ++ struct hlist_head i_dentry; ++ struct callback_head i_rcu; ++ }; ++ atomic64_t i_version; ++ atomic64_t i_sequence; ++ atomic_t i_count; ++ atomic_t i_dio_count; ++ atomic_t i_writecount; ++ atomic_t i_readcount; ++ union { ++ const struct file_operations *i_fop; ++ void (*free_inode)(struct inode *); ++ }; ++ struct file_lock_context *i_flctx; ++ struct address_space i_data; ++ struct list_head i_devices; ++ union { ++ struct pipe_inode_info *i_pipe; ++ struct cdev *i_cdev; ++ char *i_link; ++ unsigned int i_dir_seq; ++ }; ++ __u32 i_generation; ++ __u32 i_fsnotify_mask; ++ struct fsnotify_mark_connector *i_fsnotify_marks; ++ struct fscrypt_info *i_crypt_info; ++ struct fsverity_info *i_verity_info; ++ void *i_private; ++}; ++ ++struct dentry_operations { ++ int (*d_revalidate)(struct dentry *, unsigned int); ++ int (*d_weak_revalidate)(struct dentry *, unsigned int); ++ int (*d_hash)(const struct dentry *, struct qstr *); ++ int (*d_compare)(const struct dentry *, unsigned int, const char *, const struct qstr *); ++ int (*d_delete)(const struct dentry *); ++ int (*d_init)(struct dentry *); ++ void (*d_release)(struct dentry *); ++ void (*d_prune)(struct dentry *); ++ void (*d_iput)(struct dentry *, struct inode *); ++ char * (*d_dname)(struct dentry *, char *, int); ++ struct vfsmount * (*d_automount)(struct path *); ++ int (*d_manage)(const struct path *, bool); ++ struct dentry * (*d_real)(struct dentry *, const struct inode *); ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct mtd_info; ++ ++typedef long long int qsize_t; ++ ++struct quota_format_type; ++ ++struct mem_dqinfo { ++ struct quota_format_type *dqi_format; ++ int dqi_fmt_id; ++ struct list_head dqi_dirty_list; ++ long unsigned int dqi_flags; ++ unsigned int dqi_bgrace; ++ unsigned int dqi_igrace; ++ qsize_t dqi_max_spc_limit; ++ qsize_t dqi_max_ino_limit; ++ void *dqi_priv; ++}; ++ ++struct quota_format_ops; ++ ++struct quota_info { ++ unsigned int flags; ++ struct rw_semaphore dqio_sem; ++ struct inode *files[3]; ++ struct mem_dqinfo info[3]; ++ const struct quota_format_ops *ops[3]; ++}; ++ ++struct rcu_sync { ++ int gp_state; ++ int gp_count; ++ wait_queue_head_t gp_wait; ++ struct callback_head cb_head; ++}; ++ ++struct rcuwait { ++ struct task_struct *task; ++}; ++ ++struct percpu_rw_semaphore { ++ struct rcu_sync rss; ++ unsigned int *read_count; ++ struct rcuwait writer; ++ wait_queue_head_t waiters; ++ atomic_t block; ++}; ++ ++struct sb_writers { ++ int frozen; ++ wait_queue_head_t wait_unfrozen; ++ struct percpu_rw_semaphore rw_sem[3]; ++}; ++ ++typedef struct { ++ __u8 b[16]; ++} uuid_t; ++ ++struct shrink_control; ++ ++struct shrinker { ++ long unsigned int (*count_objects)(struct shrinker *, struct shrink_control *); ++ long unsigned int (*scan_objects)(struct shrinker *, struct shrink_control *); ++ long int batch; ++ int seeks; ++ unsigned int flags; ++ struct list_head list; ++ int id; ++ atomic_long_t *nr_deferred; ++}; ++ ++struct list_lru_node; ++ ++struct list_lru { ++ struct list_lru_node *node; ++ struct list_head list; ++ int shrinker_id; ++ bool memcg_aware; ++ struct xarray xa; ++}; ++ ++struct super_operations; ++ ++struct dquot_operations; ++ ++struct quotactl_ops; ++ ++struct export_operations; ++ ++struct xattr_handler; ++ ++struct fscrypt_operations; ++ ++struct fscrypt_keyring; ++ ++struct fsverity_operations; ++ ++struct unicode_map; ++ ++struct block_device; ++ ++struct super_block { ++ struct list_head s_list; ++ dev_t s_dev; ++ unsigned char s_blocksize_bits; ++ long unsigned int s_blocksize; ++ loff_t s_maxbytes; ++ struct file_system_type *s_type; ++ const struct super_operations *s_op; ++ const struct dquot_operations *dq_op; ++ const struct quotactl_ops *s_qcop; ++ const struct export_operations *s_export_op; ++ long unsigned int s_flags; ++ long unsigned int s_iflags; ++ long unsigned int s_magic; ++ struct dentry *s_root; ++ struct rw_semaphore s_umount; ++ int s_count; ++ atomic_t s_active; ++ void *s_security; ++ const struct xattr_handler **s_xattr; ++ const struct fscrypt_operations *s_cop; ++ struct fscrypt_keyring *s_master_keys; ++ const struct fsverity_operations *s_vop; ++ struct unicode_map *s_encoding; ++ __u16 s_encoding_flags; ++ struct hlist_bl_head s_roots; ++ struct list_head s_mounts; ++ struct block_device *s_bdev; ++ struct backing_dev_info *s_bdi; ++ struct mtd_info *s_mtd; ++ struct hlist_node s_instances; ++ unsigned int s_quota_types; ++ struct quota_info s_dquot; ++ struct sb_writers s_writers; ++ void *s_fs_info; ++ u32 s_time_gran; ++ time64_t s_time_min; ++ time64_t s_time_max; ++ __u32 s_fsnotify_mask; ++ struct fsnotify_mark_connector *s_fsnotify_marks; ++ char s_id[32]; ++ uuid_t s_uuid; ++ unsigned int s_max_links; ++ fmode_t s_mode; ++ struct mutex s_vfs_rename_mutex; ++ const char *s_subtype; ++ const struct dentry_operations *s_d_op; ++ struct shrinker s_shrink; ++ atomic_long_t s_remove_count; ++ atomic_long_t s_fsnotify_connectors; ++ int s_readonly_remount; ++ errseq_t s_wb_err; ++ struct workqueue_struct *s_dio_done_wq; ++ struct hlist_head s_pins; ++ struct user_namespace *s_user_ns; ++ struct list_lru s_dentry_lru; ++ struct list_lru s_inode_lru; ++ struct callback_head rcu; ++ struct work_struct destroy_work; ++ struct mutex s_sync_lock; ++ int s_stack_depth; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ spinlock_t s_inode_list_lock; ++ struct list_head s_inodes; ++ spinlock_t s_inode_wblist_lock; ++ struct list_head s_inodes_wb; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct mnt_idmap; ++ ++struct vfsmount { ++ struct dentry *mnt_root; ++ struct super_block *mnt_sb; ++ int mnt_flags; ++ struct mnt_idmap *mnt_idmap; ++}; ++ ++struct shrink_control { ++ gfp_t gfp_mask; ++ int nid; ++ long unsigned int nr_to_scan; ++ long unsigned int nr_scanned; ++ struct mem_cgroup *memcg; ++}; ++ ++struct pid_namespace; ++ ++struct upid { ++ int nr; ++ struct pid_namespace *ns; ++}; ++ ++struct idr { ++ struct xarray idr_rt; ++ unsigned int idr_base; ++ unsigned int idr_next; ++}; ++ ++struct kmem_cache; ++ ++struct fs_pin; ++ ++struct pid_namespace { ++ struct idr idr; ++ struct callback_head rcu; ++ unsigned int pid_allocated; ++ struct task_struct *child_reaper; ++ struct kmem_cache *pid_cachep; ++ unsigned int level; ++ struct pid_namespace *parent; ++ struct fs_pin *bacct; ++ struct user_namespace *user_ns; ++ struct ucounts *ucounts; ++ int reboot; ++ struct ns_common ns; ++ int memfd_noexec_scope; ++}; ++ ++struct pid { ++ refcount_t count; ++ unsigned int level; ++ spinlock_t lock; ++ struct hlist_head tasks[4]; ++ struct hlist_head inodes; ++ wait_queue_head_t wait_pidfd; ++ struct callback_head rcu; ++ struct upid numbers[1]; ++}; ++ ++struct hrtimer_cpu_base; ++ ++struct hrtimer_clock_base { ++ struct hrtimer_cpu_base *cpu_base; ++ unsigned int index; ++ clockid_t clockid; ++ seqcount_raw_spinlock_t seq; ++ struct hrtimer *running; ++ struct timerqueue_head active; ++ ktime_t (*get_time)(); ++ ktime_t offset; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct hrtimer_cpu_base { ++ raw_spinlock_t lock; ++ unsigned int cpu; ++ unsigned int active_bases; ++ unsigned int clock_was_set_seq; ++ unsigned int hres_active: 1; ++ unsigned int in_hrtirq: 1; ++ unsigned int hang_detected: 1; ++ unsigned int softirq_activated: 1; ++ unsigned int nr_events; ++ short unsigned int nr_retries; ++ short unsigned int nr_hangs; ++ unsigned int max_hang_time; ++ ktime_t expires_next; ++ struct hrtimer *next_timer; ++ ktime_t softirq_expires_next; ++ struct hrtimer *softirq_next_timer; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct hrtimer_clock_base clock_base[8]; ++}; ++ ++struct rlimit { ++ __kernel_ulong_t rlim_cur; ++ __kernel_ulong_t rlim_max; ++}; ++ ++struct task_cputime { ++ u64 stime; ++ u64 utime; ++ long long unsigned int sum_exec_runtime; ++}; ++ ++typedef void __signalfn_t(int); ++ ++typedef __signalfn_t *__sighandler_t; ++ ++typedef void __restorefn_t(); ++ ++typedef __restorefn_t *__sigrestore_t; ++ ++union sigval { ++ int sival_int; ++ void *sival_ptr; ++}; ++ ++typedef union sigval sigval_t; ++ ++union __sifields { ++ struct { ++ __kernel_pid_t _pid; ++ __kernel_uid32_t _uid; ++ } _kill; ++ struct { ++ __kernel_timer_t _tid; ++ int _overrun; ++ sigval_t _sigval; ++ int _sys_private; ++ } _timer; ++ struct { ++ __kernel_pid_t _pid; ++ __kernel_uid32_t _uid; ++ sigval_t _sigval; ++ } _rt; ++ struct { ++ __kernel_pid_t _pid; ++ __kernel_uid32_t _uid; ++ int _status; ++ __kernel_clock_t _utime; ++ __kernel_clock_t _stime; ++ } _sigchld; ++ struct { ++ void *_addr; ++ union { ++ int _trapno; ++ short int _addr_lsb; ++ struct { ++ char _dummy_bnd[8]; ++ void *_lower; ++ void *_upper; ++ } _addr_bnd; ++ struct { ++ char _dummy_pkey[8]; ++ __u32 _pkey; ++ } _addr_pkey; ++ struct { ++ long unsigned int _data; ++ __u32 _type; ++ __u32 _flags; ++ } _perf; ++ }; ++ } _sigfault; ++ struct { ++ long int _band; ++ int _fd; ++ } _sigpoll; ++ struct { ++ void *_call_addr; ++ int _syscall; ++ unsigned int _arch; ++ } _sigsys; ++}; ++ ++struct kernel_siginfo { ++ struct { ++ int si_signo; ++ int si_errno; ++ int si_code; ++ union __sifields _sifields; ++ }; ++}; ++ ++struct ucounts { ++ struct hlist_node node; ++ struct user_namespace *ns; ++ kuid_t uid; ++ atomic_t count; ++ atomic_long_t ucount[12]; ++ atomic_long_t rlimit[4]; ++}; ++ ++struct sigaction { ++ __sighandler_t sa_handler; ++ long unsigned int sa_flags; ++ __sigrestore_t sa_restorer; ++ sigset_t sa_mask; ++}; ++ ++struct k_sigaction { ++ struct sigaction sa; ++}; ++ ++struct cpu_itimer { ++ u64 expires; ++ u64 incr; ++}; ++ ++struct task_cputime_atomic { ++ atomic64_t utime; ++ atomic64_t stime; ++ atomic64_t sum_exec_runtime; ++}; ++ ++struct thread_group_cputimer { ++ struct task_cputime_atomic cputime_atomic; ++}; ++ ++struct pacct_struct { ++ int ac_flag; ++ long int ac_exitcode; ++ long unsigned int ac_mem; ++ u64 ac_utime; ++ u64 ac_stime; ++ long unsigned int ac_minflt; ++ long unsigned int ac_majflt; ++}; ++ ++struct core_state; ++ ++struct tty_struct; ++ ++struct autogroup; ++ ++struct taskstats; ++ ++struct tty_audit_buf; ++ ++struct signal_struct { ++ refcount_t sigcnt; ++ atomic_t live; ++ int nr_threads; ++ int quick_threads; ++ struct list_head thread_head; ++ wait_queue_head_t wait_chldexit; ++ struct task_struct *curr_target; ++ struct sigpending shared_pending; ++ struct hlist_head multiprocess; ++ int group_exit_code; ++ int notify_count; ++ struct task_struct *group_exec_task; ++ int group_stop_count; ++ unsigned int flags; ++ struct core_state *core_state; ++ unsigned int is_child_subreaper: 1; ++ unsigned int has_child_subreaper: 1; ++ unsigned int next_posix_timer_id; ++ struct list_head posix_timers; ++ struct hrtimer real_timer; ++ ktime_t it_real_incr; ++ struct cpu_itimer it[2]; ++ struct thread_group_cputimer cputimer; ++ struct posix_cputimers posix_cputimers; ++ struct pid *pids[4]; ++ struct pid *tty_old_pgrp; ++ int leader; ++ struct tty_struct *tty; ++ struct autogroup *autogroup; ++ seqlock_t stats_lock; ++ u64 utime; ++ u64 stime; ++ u64 cutime; ++ u64 cstime; ++ u64 gtime; ++ u64 cgtime; ++ struct prev_cputime prev_cputime; ++ long unsigned int nvcsw; ++ long unsigned int nivcsw; ++ long unsigned int cnvcsw; ++ long unsigned int cnivcsw; ++ long unsigned int min_flt; ++ long unsigned int maj_flt; ++ long unsigned int cmin_flt; ++ long unsigned int cmaj_flt; ++ long unsigned int inblock; ++ long unsigned int oublock; ++ long unsigned int cinblock; ++ long unsigned int coublock; ++ long unsigned int maxrss; ++ long unsigned int cmaxrss; ++ struct task_io_accounting ioac; ++ long long unsigned int sum_sched_runtime; ++ struct rlimit rlim[16]; ++ struct pacct_struct pacct; ++ struct taskstats *stats; ++ unsigned int audit_tty; ++ struct tty_audit_buf *tty_audit_buf; ++ bool oom_flag_origin; ++ short int oom_score_adj; ++ short int oom_score_adj_min; ++ struct mm_struct *oom_mm; ++ struct mutex cred_guard_mutex; ++ struct rw_semaphore exec_update_lock; ++}; ++ ++struct rseq { ++ __u32 cpu_id_start; ++ __u32 cpu_id; ++ __u64 rseq_cs; ++ __u32 flags; ++ __u32 node_id; ++ __u32 mm_cid; ++ char end[0]; ++}; ++ ++struct rq; ++ ++struct rq_flags; ++ ++struct affinity_context; ++ ++struct sched_class { ++ void (*enqueue_task)(struct rq *, struct task_struct *, int); ++ void (*dequeue_task)(struct rq *, struct task_struct *, int); ++ void (*yield_task)(struct rq *); ++ bool (*yield_to_task)(struct rq *, struct task_struct *); ++ void (*check_preempt_curr)(struct rq *, struct task_struct *, int); ++ struct task_struct * (*pick_next_task)(struct rq *); ++ void (*put_prev_task)(struct rq *, struct task_struct *); ++ void (*set_next_task)(struct rq *, struct task_struct *, bool); ++ int (*balance)(struct rq *, struct task_struct *, struct rq_flags *); ++ int (*select_task_rq)(struct task_struct *, int, int); ++ struct task_struct * (*pick_task)(struct rq *); ++ void (*migrate_task_rq)(struct task_struct *, int); ++ void (*task_woken)(struct rq *, struct task_struct *); ++ void (*set_cpus_allowed)(struct task_struct *, struct affinity_context *); ++ void (*rq_online)(struct rq *); ++ void (*rq_offline)(struct rq *); ++ struct rq * (*find_lock_rq)(struct task_struct *, struct rq *); ++ void (*task_tick)(struct rq *, struct task_struct *, int); ++ void (*task_fork)(struct task_struct *); ++ void (*task_dead)(struct task_struct *); ++ void (*switched_from)(struct rq *, struct task_struct *); ++ void (*switched_to)(struct rq *, struct task_struct *); ++ void (*prio_changed)(struct rq *, struct task_struct *, int); ++ unsigned int (*get_rr_interval)(struct rq *, struct task_struct *); ++ void (*update_curr)(struct rq *); ++ void (*task_change_group)(struct task_struct *); ++ int (*task_is_throttled)(struct task_struct *, int); ++}; ++ ++typedef int32_t key_serial_t; ++ ++typedef uint32_t key_perm_t; ++ ++struct key_type; ++ ++struct key_tag; ++ ++struct keyring_index_key { ++ long unsigned int hash; ++ union { ++ struct { ++ char desc[6]; ++ u16 desc_len; ++ }; ++ long unsigned int x; ++ }; ++ struct key_type *type; ++ struct key_tag *domain_tag; ++ const char *description; ++}; ++ ++union key_payload { ++ void *rcu_data0; ++ void *data[4]; ++}; ++ ++struct assoc_array_ptr; ++ ++struct assoc_array { ++ struct assoc_array_ptr *root; ++ long unsigned int nr_leaves_on_tree; ++}; ++ ++struct watch_list; ++ ++struct key_user; ++ ++struct key_restriction; ++ ++struct key { ++ refcount_t usage; ++ key_serial_t serial; ++ union { ++ struct list_head graveyard_link; ++ struct rb_node serial_node; ++ }; ++ struct watch_list *watchers; ++ struct rw_semaphore sem; ++ struct key_user *user; ++ void *security; ++ union { ++ time64_t expiry; ++ time64_t revoked_at; ++ }; ++ time64_t last_used_at; ++ kuid_t uid; ++ kgid_t gid; ++ key_perm_t perm; ++ short unsigned int quotalen; ++ short unsigned int datalen; ++ short int state; ++ long unsigned int flags; ++ union { ++ struct keyring_index_key index_key; ++ struct { ++ long unsigned int hash; ++ long unsigned int len_desc; ++ struct key_type *type; ++ struct key_tag *domain_tag; ++ char *description; ++ }; ++ }; ++ union { ++ union key_payload payload; ++ struct { ++ struct list_head name_link; ++ struct assoc_array keys; ++ }; ++ }; ++ struct key_restriction *restrict_link; ++}; ++ ++struct fs_struct { ++ int users; ++ spinlock_t lock; ++ seqcount_spinlock_t seq; ++ int umask; ++ int in_exec; ++ struct path root; ++ struct path pwd; ++}; ++ ++struct uts_namespace; ++ ++struct ipc_namespace; ++ ++struct mnt_namespace; ++ ++struct time_namespace; ++ ++struct cgroup_namespace; ++ ++struct nsproxy { ++ atomic_t count; ++ struct uts_namespace *uts_ns; ++ struct ipc_namespace *ipc_ns; ++ struct mnt_namespace *mnt_ns; ++ struct pid_namespace *pid_ns_for_children; ++ struct net *net_ns; ++ struct time_namespace *time_ns; ++ struct time_namespace *time_ns_for_children; ++ struct cgroup_namespace *cgroup_ns; ++}; ++ ++struct sighand_struct { ++ spinlock_t siglock; ++ refcount_t count; ++ wait_queue_head_t signalfd_wqh; ++ struct k_sigaction action[64]; ++}; ++ ++struct bio; ++ ++struct bio_list { ++ struct bio *head; ++ struct bio *tail; ++}; ++ ++struct request; ++ ++struct blk_plug { ++ struct request *mq_list; ++ struct request *cached_rq; ++ short unsigned int nr_ios; ++ short unsigned int rq_count; ++ bool multiple_queues; ++ bool has_elevator; ++ bool nowait; ++ struct list_head cb_list; ++}; ++ ++struct reclaim_state { ++ long unsigned int reclaimed; ++ struct lru_gen_mm_walk *mm_walk; ++}; ++ ++struct fprop_local_percpu { ++ struct percpu_counter events; ++ unsigned int period; ++ raw_spinlock_t lock; ++}; ++ ++enum wb_reason { ++ WB_REASON_BACKGROUND = 0, ++ WB_REASON_VMSCAN = 1, ++ WB_REASON_SYNC = 2, ++ WB_REASON_PERIODIC = 3, ++ WB_REASON_LAPTOP_TIMER = 4, ++ WB_REASON_FS_FREE_SPACE = 5, ++ WB_REASON_FORKER_THREAD = 6, ++ WB_REASON_FOREIGN_FLUSH = 7, ++ WB_REASON_MAX = 8, ++}; ++ ++struct bdi_writeback { ++ struct backing_dev_info *bdi; ++ long unsigned int state; ++ long unsigned int last_old_flush; ++ struct list_head b_dirty; ++ struct list_head b_io; ++ struct list_head b_more_io; ++ struct list_head b_dirty_time; ++ spinlock_t list_lock; ++ atomic_t writeback_inodes; ++ struct percpu_counter stat[4]; ++ long unsigned int bw_time_stamp; ++ long unsigned int dirtied_stamp; ++ long unsigned int written_stamp; ++ long unsigned int write_bandwidth; ++ long unsigned int avg_write_bandwidth; ++ long unsigned int dirty_ratelimit; ++ long unsigned int balanced_dirty_ratelimit; ++ struct fprop_local_percpu completions; ++ int dirty_exceeded; ++ enum wb_reason start_all_reason; ++ spinlock_t work_lock; ++ struct list_head work_list; ++ struct delayed_work dwork; ++ struct delayed_work bw_dwork; ++ long unsigned int dirty_sleep; ++ struct list_head bdi_node; ++ struct percpu_ref refcnt; ++ struct fprop_local_percpu memcg_completions; ++ struct cgroup_subsys_state *memcg_css; ++ struct cgroup_subsys_state *blkcg_css; ++ struct list_head memcg_node; ++ struct list_head blkcg_node; ++ struct list_head b_attached; ++ struct list_head offline_node; ++ union { ++ struct work_struct release_work; ++ struct callback_head rcu; ++ }; ++}; ++ ++struct backing_dev_info { ++ u64 id; ++ struct rb_node rb_node; ++ struct list_head bdi_list; ++ long unsigned int ra_pages; ++ long unsigned int io_pages; ++ struct kref refcnt; ++ unsigned int capabilities; ++ unsigned int min_ratio; ++ unsigned int max_ratio; ++ unsigned int max_prop_frac; ++ atomic_long_t tot_write_bandwidth; ++ struct bdi_writeback wb; ++ struct list_head wb_list; ++ struct xarray cgwb_tree; ++ struct mutex cgwb_release_mutex; ++ struct rw_semaphore wb_switch_rwsem; ++ wait_queue_head_t wb_waitq; ++ struct device *dev; ++ char dev_name[64]; ++ struct device *owner; ++ struct timer_list laptop_mode_wb_timer; ++ struct dentry *debug_dir; ++}; ++ ++struct io_cq; ++ ++struct io_context { ++ atomic_long_t refcount; ++ atomic_t active_ref; ++ short unsigned int ioprio; ++ spinlock_t lock; ++ struct xarray icq_tree; ++ struct io_cq *icq_hint; ++ struct hlist_head icq_list; ++ struct work_struct release_work; ++}; ++ ++struct css_set { ++ struct cgroup_subsys_state *subsys[13]; ++ refcount_t refcount; ++ struct css_set *dom_cset; ++ struct cgroup *dfl_cgrp; ++ int nr_tasks; ++ struct list_head tasks; ++ struct list_head mg_tasks; ++ struct list_head dying_tasks; ++ struct list_head task_iters; ++ struct list_head e_cset_node[13]; ++ struct list_head threaded_csets; ++ struct list_head threaded_csets_node; ++ struct hlist_node hlist; ++ struct list_head cgrp_links; ++ struct list_head mg_src_preload_node; ++ struct list_head mg_dst_preload_node; ++ struct list_head mg_node; ++ struct cgroup *mg_src_cgrp; ++ struct cgroup *mg_dst_cgrp; ++ struct css_set *mg_dst_cset; ++ bool dead; ++ struct callback_head callback_head; ++}; ++ ++typedef u32 compat_uptr_t; ++ ++struct compat_robust_list { ++ compat_uptr_t next; ++}; ++ ++typedef s32 compat_long_t; ++ ++struct compat_robust_list_head { ++ struct compat_robust_list list; ++ compat_long_t futex_offset; ++ compat_uptr_t list_op_pending; ++}; ++ ++struct perf_event_groups { ++ struct rb_root tree; ++ u64 index; ++}; ++ ++typedef struct { ++ atomic_long_t a; ++} local_t; ++ ++struct perf_event_context { ++ raw_spinlock_t lock; ++ struct mutex mutex; ++ struct list_head pmu_ctx_list; ++ struct perf_event_groups pinned_groups; ++ struct perf_event_groups flexible_groups; ++ struct list_head event_list; ++ int nr_events; ++ int nr_user; ++ int is_active; ++ int nr_task_data; ++ int nr_stat; ++ int nr_freq; ++ int rotate_disable; ++ refcount_t refcount; ++ struct task_struct *task; ++ u64 time; ++ u64 timestamp; ++ u64 timeoffset; ++ struct perf_event_context *parent_ctx; ++ u64 parent_gen; ++ u64 generation; ++ int pin_count; ++ int nr_cgroups; ++ struct callback_head callback_head; ++ local_t nr_pending; ++}; ++ ++struct ftrace_ret_stack { ++ long unsigned int ret; ++ long unsigned int func; ++ long long unsigned int calltime; ++ long long unsigned int subtime; ++ long unsigned int *retp; ++}; ++ ++typedef void *mempool_alloc_t(gfp_t, void *); ++ ++typedef void mempool_free_t(void *, void *); ++ ++struct mempool_s { ++ spinlock_t lock; ++ int min_nr; ++ int curr_nr; ++ void **elements; ++ void *pool_data; ++ mempool_alloc_t *alloc; ++ mempool_free_t *free; ++ wait_queue_head_t wait; ++}; ++ ++typedef struct mempool_s mempool_t; ++ ++struct bio_alloc_cache; ++ ++struct bio_set { ++ struct kmem_cache *bio_slab; ++ unsigned int front_pad; ++ struct bio_alloc_cache *cache; ++ mempool_t bio_pool; ++ mempool_t bvec_pool; ++ mempool_t bio_integrity_pool; ++ mempool_t bvec_integrity_pool; ++ unsigned int back_pad; ++ spinlock_t rescue_lock; ++ struct bio_list rescue_list; ++ struct work_struct rescue_work; ++ struct workqueue_struct *rescue_workqueue; ++ struct hlist_node cpuhp_dead; ++}; ++ ++struct block_device_operations; ++ ++struct request_queue; ++ ++struct timer_rand_state; ++ ++struct disk_events; ++ ++struct cdrom_device_info; ++ ++struct badblocks; ++ ++struct blk_independent_access_ranges; ++ ++struct gendisk { ++ int major; ++ int first_minor; ++ int minors; ++ char disk_name[32]; ++ short unsigned int events; ++ short unsigned int event_flags; ++ struct xarray part_tbl; ++ struct block_device *part0; ++ const struct block_device_operations *fops; ++ struct request_queue *queue; ++ void *private_data; ++ struct bio_set bio_split; ++ int flags; ++ long unsigned int state; ++ struct mutex open_mutex; ++ unsigned int open_partitions; ++ struct backing_dev_info *bdi; ++ struct kobject queue_kobj; ++ struct kobject *slave_dir; ++ struct list_head slave_bdevs; ++ struct timer_rand_state *random; ++ atomic_t sync_io; ++ struct disk_events *ev; ++ unsigned int nr_zones; ++ unsigned int max_open_zones; ++ unsigned int max_active_zones; ++ long unsigned int *conv_zones_bitmap; ++ long unsigned int *seq_zones_wlock; ++ struct cdrom_device_info *cdi; ++ int node_id; ++ struct badblocks *bb; ++ struct lockdep_map lockdep_map; ++ u64 diskseq; ++ struct blk_independent_access_ranges *ia_ranges; ++}; ++ ++struct vm_struct { ++ struct vm_struct *next; ++ void *addr; ++ long unsigned int size; ++ long unsigned int flags; ++ struct page **pages; ++ unsigned int nr_pages; ++ phys_addr_t phys_addr; ++ const void *caller; ++}; ++ ++struct list_lru_one { ++ struct list_head list; ++ long int nr_items; ++}; ++ ++struct list_lru_node { ++ spinlock_t lock; ++ struct list_lru_one lru; ++ long int nr_items; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum migrate_mode { ++ MIGRATE_ASYNC = 0, ++ MIGRATE_SYNC_LIGHT = 1, ++ MIGRATE_SYNC = 2, ++ MIGRATE_SYNC_NO_COPY = 3, ++}; ++ ++struct exception_table_entry { ++ int insn; ++ int fixup; ++ short int type; ++ short int data; ++}; ++ ++struct cgroup_base_stat { ++ struct task_cputime cputime; ++ u64 forceidle_sum; ++}; ++ ++struct bpf_prog_array; ++ ++struct cgroup_bpf { ++ struct bpf_prog_array *effective[33]; ++ struct hlist_head progs[33]; ++ u8 flags[33]; ++ struct list_head storages; ++ struct bpf_prog_array *inactive; ++ struct percpu_ref refcnt; ++ struct work_struct release_work; ++}; ++ ++struct cgroup_freezer_state { ++ bool freeze; ++ int e_freeze; ++ int nr_frozen_descendants; ++ int nr_frozen_tasks; ++}; ++ ++struct cgroup_root; ++ ++struct cgroup_rstat_cpu; ++ ++struct psi_group; ++ ++struct cgroup { ++ struct cgroup_subsys_state self; ++ long unsigned int flags; ++ int level; ++ int max_depth; ++ int nr_descendants; ++ int nr_dying_descendants; ++ int max_descendants; ++ int nr_populated_csets; ++ int nr_populated_domain_children; ++ int nr_populated_threaded_children; ++ int nr_threaded_children; ++ struct kernfs_node *kn; ++ struct cgroup_file procs_file; ++ struct cgroup_file events_file; ++ struct cgroup_file psi_files[3]; ++ u16 subtree_control; ++ u16 subtree_ss_mask; ++ u16 old_subtree_control; ++ u16 old_subtree_ss_mask; ++ struct cgroup_subsys_state *subsys[13]; ++ struct cgroup_root *root; ++ struct list_head cset_links; ++ struct list_head e_csets[13]; ++ struct cgroup *dom_cgrp; ++ struct cgroup *old_dom_cgrp; ++ struct cgroup_rstat_cpu *rstat_cpu; ++ struct list_head rstat_css_list; ++ struct cgroup_base_stat last_bstat; ++ struct cgroup_base_stat bstat; ++ struct prev_cputime prev_cputime; ++ struct list_head pidlists; ++ struct mutex pidlist_mutex; ++ wait_queue_head_t offline_waitq; ++ struct work_struct release_agent_work; ++ struct psi_group *psi; ++ struct cgroup_bpf bpf; ++ atomic_t congestion_count; ++ struct cgroup_freezer_state freezer; ++ struct bpf_local_storage *bpf_cgrp_storage; ++ struct cgroup *ancestors[0]; ++}; ++ ++struct key_tag { ++ struct callback_head rcu; ++ refcount_t usage; ++ bool removed; ++}; ++ ++typedef int (*request_key_actor_t)(struct key *, void *); ++ ++struct key_preparsed_payload; ++ ++struct key_match_data; ++ ++struct kernel_pkey_params; ++ ++struct kernel_pkey_query; ++ ++struct key_type { ++ const char *name; ++ size_t def_datalen; ++ unsigned int flags; ++ int (*vet_description)(const char *); ++ int (*preparse)(struct key_preparsed_payload *); ++ void (*free_preparse)(struct key_preparsed_payload *); ++ int (*instantiate)(struct key *, struct key_preparsed_payload *); ++ int (*update)(struct key *, struct key_preparsed_payload *); ++ int (*match_preparse)(struct key_match_data *); ++ void (*match_free)(struct key_match_data *); ++ void (*revoke)(struct key *); ++ void (*destroy)(struct key *); ++ void (*describe)(const struct key *, struct seq_file *); ++ long int (*read)(const struct key *, char *, size_t); ++ request_key_actor_t request_key; ++ struct key_restriction * (*lookup_restriction)(const char *); ++ int (*asym_query)(const struct kernel_pkey_params *, struct kernel_pkey_query *); ++ int (*asym_eds_op)(struct kernel_pkey_params *, const void *, void *); ++ int (*asym_verify_signature)(struct kernel_pkey_params *, const void *, const void *); ++ struct list_head link; ++ struct lock_class_key lock_class; ++}; ++ ++typedef int (*key_restrict_link_func_t)(struct key *, const struct key_type *, const union key_payload *, struct key *); ++ ++struct key_restriction { ++ key_restrict_link_func_t check; ++ struct key *key; ++ struct key_type *keytype; ++}; ++ ++struct user_struct { ++ refcount_t __count; ++ struct percpu_counter epoll_watches; ++ long unsigned int unix_inflight; ++ atomic_long_t pipe_bufs; ++ struct hlist_node uidhash_node; ++ kuid_t uid; ++ atomic_long_t locked_vm; ++ atomic_t nr_watches; ++ struct ratelimit_state ratelimit; ++}; ++ ++struct group_info { ++ atomic_t usage; ++ int ngroups; ++ kgid_t gid[0]; ++}; ++ ++struct core_thread { ++ struct task_struct *task; ++ struct core_thread *next; ++}; ++ ++struct core_state { ++ atomic_t nr_threads; ++ struct core_thread dumper; ++ struct completion startup; ++}; ++ ++struct ld_semaphore { ++ atomic_long_t count; ++ raw_spinlock_t wait_lock; ++ unsigned int wait_readers; ++ struct list_head read_wait; ++ struct list_head write_wait; ++}; ++ ++typedef unsigned int tcflag_t; ++ ++typedef unsigned char cc_t; ++ ++typedef unsigned int speed_t; ++ ++struct ktermios { ++ tcflag_t c_iflag; ++ tcflag_t c_oflag; ++ tcflag_t c_cflag; ++ tcflag_t c_lflag; ++ cc_t c_line; ++ cc_t c_cc[19]; ++ speed_t c_ispeed; ++ speed_t c_ospeed; ++}; ++ ++struct winsize { ++ short unsigned int ws_row; ++ short unsigned int ws_col; ++ short unsigned int ws_xpixel; ++ short unsigned int ws_ypixel; ++}; ++ ++struct tty_driver; ++ ++struct tty_operations; ++ ++struct tty_ldisc; ++ ++struct fasync_struct; ++ ++struct tty_port; ++ ++struct tty_struct { ++ struct kref kref; ++ struct device *dev; ++ struct tty_driver *driver; ++ const struct tty_operations *ops; ++ int index; ++ struct ld_semaphore ldisc_sem; ++ struct tty_ldisc *ldisc; ++ struct mutex atomic_write_lock; ++ struct mutex legacy_mutex; ++ struct mutex throttle_mutex; ++ struct rw_semaphore termios_rwsem; ++ struct mutex winsize_mutex; ++ struct ktermios termios; ++ struct ktermios termios_locked; ++ char name[64]; ++ long unsigned int flags; ++ int count; ++ struct winsize winsize; ++ struct { ++ spinlock_t lock; ++ bool stopped; ++ bool tco_stopped; ++ long unsigned int unused[0]; ++ } flow; ++ struct { ++ spinlock_t lock; ++ struct pid *pgrp; ++ struct pid *session; ++ unsigned char pktstatus; ++ bool packet; ++ long unsigned int unused[0]; ++ } ctrl; ++ bool hw_stopped; ++ unsigned int receive_room; ++ int flow_change; ++ struct tty_struct *link; ++ struct fasync_struct *fasync; ++ wait_queue_head_t write_wait; ++ wait_queue_head_t read_wait; ++ struct work_struct hangup_work; ++ void *disc_data; ++ void *driver_data; ++ spinlock_t files_lock; ++ struct list_head tty_files; ++ int closing; ++ unsigned char *write_buf; ++ int write_cnt; ++ struct work_struct SAK_work; ++ struct tty_port *port; ++}; ++ ++struct taskstats { ++ __u16 version; ++ __u32 ac_exitcode; ++ __u8 ac_flag; ++ __u8 ac_nice; ++ __u64 cpu_count; ++ __u64 cpu_delay_total; ++ __u64 blkio_count; ++ __u64 blkio_delay_total; ++ __u64 swapin_count; ++ __u64 swapin_delay_total; ++ __u64 cpu_run_real_total; ++ __u64 cpu_run_virtual_total; ++ char ac_comm[32]; ++ __u8 ac_sched; ++ __u8 ac_pad[3]; ++ long: 0; ++ __u32 ac_uid; ++ __u32 ac_gid; ++ __u32 ac_pid; ++ __u32 ac_ppid; ++ __u32 ac_btime; ++ __u64 ac_etime; ++ __u64 ac_utime; ++ __u64 ac_stime; ++ __u64 ac_minflt; ++ __u64 ac_majflt; ++ __u64 coremem; ++ __u64 virtmem; ++ __u64 hiwater_rss; ++ __u64 hiwater_vm; ++ __u64 read_char; ++ __u64 write_char; ++ __u64 read_syscalls; ++ __u64 write_syscalls; ++ __u64 read_bytes; ++ __u64 write_bytes; ++ __u64 cancelled_write_bytes; ++ __u64 nvcsw; ++ __u64 nivcsw; ++ __u64 ac_utimescaled; ++ __u64 ac_stimescaled; ++ __u64 cpu_scaled_run_real_total; ++ __u64 freepages_count; ++ __u64 freepages_delay_total; ++ __u64 thrashing_count; ++ __u64 thrashing_delay_total; ++ __u64 ac_btime64; ++ __u64 compact_count; ++ __u64 compact_delay_total; ++ __u32 ac_tgid; ++ __u64 ac_tgetime; ++ __u64 ac_exe_dev; ++ __u64 ac_exe_inode; ++ __u64 wpcopy_count; ++ __u64 wpcopy_delay_total; ++ __u64 irq_count; ++ __u64 irq_delay_total; ++}; ++ ++struct delayed_call { ++ void (*fn)(void *); ++ void *arg; ++}; ++ ++struct io_cq { ++ struct request_queue *q; ++ struct io_context *ioc; ++ union { ++ struct list_head q_node; ++ struct kmem_cache *__rcu_icq_cache; ++ }; ++ union { ++ struct hlist_node ioc_node; ++ struct callback_head __rcu_head; ++ }; ++ unsigned int flags; ++}; ++ ++struct blk_integrity_profile; ++ ++struct blk_integrity { ++ const struct blk_integrity_profile *profile; ++ unsigned char flags; ++ unsigned char tuple_size; ++ unsigned char interval_exp; ++ unsigned char tag_size; ++}; ++ ++enum blk_bounce { ++ BLK_BOUNCE_NONE = 0, ++ BLK_BOUNCE_HIGH = 1, ++}; ++ ++enum blk_zoned_model { ++ BLK_ZONED_NONE = 0, ++ BLK_ZONED_HA = 1, ++ BLK_ZONED_HM = 2, ++}; ++ ++struct queue_limits { ++ enum blk_bounce bounce; ++ long unsigned int seg_boundary_mask; ++ long unsigned int virt_boundary_mask; ++ unsigned int max_hw_sectors; ++ unsigned int max_dev_sectors; ++ unsigned int chunk_sectors; ++ unsigned int max_sectors; ++ unsigned int max_user_sectors; ++ unsigned int max_segment_size; ++ unsigned int physical_block_size; ++ unsigned int logical_block_size; ++ unsigned int alignment_offset; ++ unsigned int io_min; ++ unsigned int io_opt; ++ unsigned int max_discard_sectors; ++ unsigned int max_hw_discard_sectors; ++ unsigned int max_secure_erase_sectors; ++ unsigned int max_write_zeroes_sectors; ++ unsigned int max_zone_append_sectors; ++ unsigned int discard_granularity; ++ unsigned int discard_alignment; ++ unsigned int zone_write_granularity; ++ short unsigned int max_segments; ++ short unsigned int max_integrity_segments; ++ short unsigned int max_discard_segments; ++ unsigned char misaligned; ++ unsigned char discard_misaligned; ++ unsigned char raid_partial_stripes_expensive; ++ enum blk_zoned_model zoned; ++ unsigned int dma_alignment; ++}; ++ ++struct elevator_queue; ++ ++struct blk_queue_stats; ++ ++struct rq_qos; ++ ++struct blk_mq_ops; ++ ++struct blk_mq_ctx; ++ ++struct blk_crypto_profile; ++ ++struct blk_mq_tags; ++ ++struct blkcg_gq; ++ ++struct blk_trace; ++ ++struct blk_flush_queue; ++ ++struct throtl_data; ++ ++struct blk_mq_tag_set; ++ ++struct request_queue { ++ struct request *last_merge; ++ struct elevator_queue *elevator; ++ struct percpu_ref q_usage_counter; ++ struct blk_queue_stats *stats; ++ struct rq_qos *rq_qos; ++ struct mutex rq_qos_mutex; ++ const struct blk_mq_ops *mq_ops; ++ struct blk_mq_ctx *queue_ctx; ++ unsigned int queue_depth; ++ struct xarray hctx_table; ++ unsigned int nr_hw_queues; ++ void *queuedata; ++ long unsigned int queue_flags; ++ atomic_t pm_only; ++ int id; ++ spinlock_t queue_lock; ++ struct gendisk *disk; ++ refcount_t refs; ++ struct kobject *mq_kobj; ++ struct blk_integrity integrity; ++ long unsigned int nr_requests; ++ unsigned int dma_pad_mask; ++ struct blk_crypto_profile *crypto_profile; ++ struct kobject *crypto_kobject; ++ unsigned int rq_timeout; ++ struct timer_list timeout; ++ struct work_struct timeout_work; ++ atomic_t nr_active_requests_shared_tags; ++ struct blk_mq_tags *sched_shared_tags; ++ struct list_head icq_list; ++ long unsigned int blkcg_pols[1]; ++ struct blkcg_gq *root_blkg; ++ struct list_head blkg_list; ++ struct mutex blkcg_mutex; ++ struct queue_limits limits; ++ unsigned int required_elevator_features; ++ int node; ++ struct blk_trace *blk_trace; ++ struct blk_flush_queue *fq; ++ struct list_head requeue_list; ++ spinlock_t requeue_lock; ++ struct delayed_work requeue_work; ++ struct mutex sysfs_lock; ++ struct mutex sysfs_dir_lock; ++ struct list_head unused_hctx_list; ++ spinlock_t unused_hctx_lock; ++ int mq_freeze_depth; ++ struct throtl_data *td; ++ struct callback_head callback_head; ++ wait_queue_head_t mq_freeze_wq; ++ struct mutex mq_freeze_lock; ++ int quiesce_depth; ++ struct blk_mq_tag_set *tag_set; ++ struct list_head tag_set_list; ++ struct dentry *debugfs_dir; ++ struct dentry *sched_debugfs_dir; ++ struct dentry *rqos_debugfs_dir; ++ struct mutex debugfs_mutex; ++ bool mq_sysfs_init_done; ++}; ++ ++typedef struct { ++ uid_t val; ++} vfsuid_t; ++ ++typedef struct { ++ gid_t val; ++} vfsgid_t; ++ ++typedef void percpu_ref_func_t(struct percpu_ref *); ++ ++struct percpu_ref_data { ++ atomic_long_t count; ++ percpu_ref_func_t *release; ++ percpu_ref_func_t *confirm_switch; ++ bool force_atomic: 1; ++ bool allow_reinit: 1; ++ struct callback_head rcu; ++ struct percpu_ref *ref; ++}; ++ ++struct iattr { ++ unsigned int ia_valid; ++ umode_t ia_mode; ++ union { ++ kuid_t ia_uid; ++ vfsuid_t ia_vfsuid; ++ }; ++ union { ++ kgid_t ia_gid; ++ vfsgid_t ia_vfsgid; ++ }; ++ loff_t ia_size; ++ struct timespec64 ia_atime; ++ struct timespec64 ia_mtime; ++ struct timespec64 ia_ctime; ++ struct file *ia_file; ++}; ++ ++typedef __kernel_uid32_t projid_t; ++ ++typedef struct { ++ projid_t val; ++} kprojid_t; ++ ++enum quota_type { ++ USRQUOTA = 0, ++ GRPQUOTA = 1, ++ PRJQUOTA = 2, ++}; ++ ++struct kqid { ++ union { ++ kuid_t uid; ++ kgid_t gid; ++ kprojid_t projid; ++ }; ++ enum quota_type type; ++}; ++ ++struct mem_dqblk { ++ qsize_t dqb_bhardlimit; ++ qsize_t dqb_bsoftlimit; ++ qsize_t dqb_curspace; ++ qsize_t dqb_rsvspace; ++ qsize_t dqb_ihardlimit; ++ qsize_t dqb_isoftlimit; ++ qsize_t dqb_curinodes; ++ time64_t dqb_btime; ++ time64_t dqb_itime; ++}; ++ ++struct dquot { ++ struct hlist_node dq_hash; ++ struct list_head dq_inuse; ++ struct list_head dq_free; ++ struct list_head dq_dirty; ++ struct mutex dq_lock; ++ spinlock_t dq_dqb_lock; ++ atomic_t dq_count; ++ struct super_block *dq_sb; ++ struct kqid dq_id; ++ loff_t dq_off; ++ long unsigned int dq_flags; ++ struct mem_dqblk dq_dqb; ++}; ++ ++struct quota_format_type { ++ int qf_fmt_id; ++ const struct quota_format_ops *qf_ops; ++ struct module *qf_owner; ++ struct quota_format_type *qf_next; ++}; ++ ++struct quota_format_ops { ++ int (*check_quota_file)(struct super_block *, int); ++ int (*read_file_info)(struct super_block *, int); ++ int (*write_file_info)(struct super_block *, int); ++ int (*free_file_info)(struct super_block *, int); ++ int (*read_dqblk)(struct dquot *); ++ int (*commit_dqblk)(struct dquot *); ++ int (*release_dqblk)(struct dquot *); ++ int (*get_next_id)(struct super_block *, struct kqid *); ++}; ++ ++struct dquot_operations { ++ int (*write_dquot)(struct dquot *); ++ struct dquot * (*alloc_dquot)(struct super_block *, int); ++ void (*destroy_dquot)(struct dquot *); ++ int (*acquire_dquot)(struct dquot *); ++ int (*release_dquot)(struct dquot *); ++ int (*mark_dirty)(struct dquot *); ++ int (*write_info)(struct super_block *, int); ++ qsize_t * (*get_reserved_space)(struct inode *); ++ int (*get_projid)(struct inode *, kprojid_t *); ++ int (*get_inode_usage)(struct inode *, qsize_t *); ++ int (*get_next_id)(struct super_block *, struct kqid *); ++}; ++ ++struct qc_dqblk { ++ int d_fieldmask; ++ u64 d_spc_hardlimit; ++ u64 d_spc_softlimit; ++ u64 d_ino_hardlimit; ++ u64 d_ino_softlimit; ++ u64 d_space; ++ u64 d_ino_count; ++ s64 d_ino_timer; ++ s64 d_spc_timer; ++ int d_ino_warns; ++ int d_spc_warns; ++ u64 d_rt_spc_hardlimit; ++ u64 d_rt_spc_softlimit; ++ u64 d_rt_space; ++ s64 d_rt_spc_timer; ++ int d_rt_spc_warns; ++}; ++ ++struct qc_type_state { ++ unsigned int flags; ++ unsigned int spc_timelimit; ++ unsigned int ino_timelimit; ++ unsigned int rt_spc_timelimit; ++ unsigned int spc_warnlimit; ++ unsigned int ino_warnlimit; ++ unsigned int rt_spc_warnlimit; ++ long long unsigned int ino; ++ blkcnt_t blocks; ++ blkcnt_t nextents; ++}; ++ ++struct qc_state { ++ unsigned int s_incoredqs; ++ struct qc_type_state s_state[3]; ++}; ++ ++struct qc_info { ++ int i_fieldmask; ++ unsigned int i_flags; ++ unsigned int i_spc_timelimit; ++ unsigned int i_ino_timelimit; ++ unsigned int i_rt_spc_timelimit; ++ unsigned int i_spc_warnlimit; ++ unsigned int i_ino_warnlimit; ++ unsigned int i_rt_spc_warnlimit; ++}; ++ ++struct quotactl_ops { ++ int (*quota_on)(struct super_block *, int, int, const struct path *); ++ int (*quota_off)(struct super_block *, int); ++ int (*quota_enable)(struct super_block *, unsigned int); ++ int (*quota_disable)(struct super_block *, unsigned int); ++ int (*quota_sync)(struct super_block *, int); ++ int (*set_info)(struct super_block *, int, struct qc_info *); ++ int (*get_dqblk)(struct super_block *, struct kqid, struct qc_dqblk *); ++ int (*get_nextdqblk)(struct super_block *, struct kqid *, struct qc_dqblk *); ++ int (*set_dqblk)(struct super_block *, struct kqid, struct qc_dqblk *); ++ int (*get_state)(struct super_block *, struct qc_state *); ++ int (*rm_xquota)(struct super_block *, unsigned int); ++}; ++ ++struct wait_page_queue { ++ struct folio *folio; ++ int bit_nr; ++ wait_queue_entry_t wait; ++}; ++ ++struct writeback_control; ++ ++struct readahead_control; ++ ++struct swap_info_struct; ++ ++struct address_space_operations { ++ int (*writepage)(struct page *, struct writeback_control *); ++ int (*read_folio)(struct file *, struct folio *); ++ int (*writepages)(struct address_space *, struct writeback_control *); ++ bool (*dirty_folio)(struct address_space *, struct folio *); ++ void (*readahead)(struct readahead_control *); ++ int (*write_begin)(struct file *, struct address_space *, loff_t, unsigned int, struct page **, void **); ++ int (*write_end)(struct file *, struct address_space *, loff_t, unsigned int, unsigned int, struct page *, void *); ++ sector_t (*bmap)(struct address_space *, sector_t); ++ void (*invalidate_folio)(struct folio *, size_t, size_t); ++ bool (*release_folio)(struct folio *, gfp_t); ++ void (*free_folio)(struct folio *); ++ ssize_t (*direct_IO)(struct kiocb *, struct iov_iter *); ++ int (*migrate_folio)(struct address_space *, struct folio *, struct folio *, enum migrate_mode); ++ int (*launder_folio)(struct folio *); ++ bool (*is_partially_uptodate)(struct folio *, size_t, size_t); ++ void (*is_dirty_writeback)(struct folio *, bool *, bool *); ++ int (*error_remove_page)(struct address_space *, struct page *); ++ int (*swap_activate)(struct swap_info_struct *, struct file *, sector_t *); ++ void (*swap_deactivate)(struct file *); ++ int (*swap_rw)(struct kiocb *, struct iov_iter *); ++}; ++ ++enum writeback_sync_modes { ++ WB_SYNC_NONE = 0, ++ WB_SYNC_ALL = 1, ++}; ++ ++struct swap_iocb; ++ ++struct writeback_control { ++ long int nr_to_write; ++ long int pages_skipped; ++ loff_t range_start; ++ loff_t range_end; ++ enum writeback_sync_modes sync_mode; ++ unsigned int for_kupdate: 1; ++ unsigned int for_background: 1; ++ unsigned int tagged_writepages: 1; ++ unsigned int for_reclaim: 1; ++ unsigned int range_cyclic: 1; ++ unsigned int for_sync: 1; ++ unsigned int unpinned_fscache_wb: 1; ++ unsigned int no_cgroup_owner: 1; ++ struct swap_iocb **swap_plug; ++ struct bdi_writeback *wb; ++ struct inode *inode; ++ int wb_id; ++ int wb_lcand_id; ++ int wb_tcand_id; ++ size_t wb_bytes; ++ size_t wb_lcand_bytes; ++ size_t wb_tcand_bytes; ++}; ++ ++struct readahead_control { ++ struct file *file; ++ struct address_space *mapping; ++ struct file_ra_state *ra; ++ long unsigned int _index; ++ unsigned int _nr_pages; ++ unsigned int _batch_count; ++ bool _workingset; ++ long unsigned int _pflags; ++}; ++ ++struct swap_cluster_info { ++ spinlock_t lock; ++ unsigned int data: 24; ++ unsigned int flags: 8; ++}; ++ ++struct swap_cluster_list { ++ struct swap_cluster_info head; ++ struct swap_cluster_info tail; ++}; ++ ++struct percpu_cluster; ++ ++struct swap_info_struct { ++ struct percpu_ref users; ++ long unsigned int flags; ++ short int prio; ++ struct plist_node list; ++ signed char type; ++ unsigned int max; ++ unsigned char *swap_map; ++ struct swap_cluster_info *cluster_info; ++ struct swap_cluster_list free_clusters; ++ unsigned int lowest_bit; ++ unsigned int highest_bit; ++ unsigned int pages; ++ unsigned int inuse_pages; ++ unsigned int cluster_next; ++ unsigned int cluster_nr; ++ unsigned int *cluster_next_cpu; ++ struct percpu_cluster *percpu_cluster; ++ struct rb_root swap_extent_root; ++ struct block_device *bdev; ++ struct file *swap_file; ++ unsigned int old_block_size; ++ struct completion comp; ++ long unsigned int *frontswap_map; ++ atomic_t frontswap_pages; ++ spinlock_t lock; ++ spinlock_t cont_lock; ++ struct work_struct discard_work; ++ struct swap_cluster_list discard_clusters; ++ struct plist_node avail_lists[0]; ++}; ++ ++struct cdev { ++ struct kobject kobj; ++ struct module *owner; ++ const struct file_operations *ops; ++ struct list_head list; ++ dev_t dev; ++ unsigned int count; ++}; ++ ++struct posix_acl_entry { ++ short int e_tag; ++ short unsigned int e_perm; ++ union { ++ kuid_t e_uid; ++ kgid_t e_gid; ++ }; ++}; ++ ++struct posix_acl { ++ refcount_t a_refcount; ++ struct callback_head a_rcu; ++ unsigned int a_count; ++ struct posix_acl_entry a_entries[0]; ++}; ++ ++struct fiemap_extent_info; ++ ++struct fileattr; ++ ++struct inode_operations { ++ struct dentry * (*lookup)(struct inode *, struct dentry *, unsigned int); ++ const char * (*get_link)(struct dentry *, struct inode *, struct delayed_call *); ++ int (*permission)(struct mnt_idmap *, struct inode *, int); ++ struct posix_acl * (*get_inode_acl)(struct inode *, int, bool); ++ int (*readlink)(struct dentry *, char *, int); ++ int (*create)(struct mnt_idmap *, struct inode *, struct dentry *, umode_t, bool); ++ int (*link)(struct dentry *, struct inode *, struct dentry *); ++ int (*unlink)(struct inode *, struct dentry *); ++ int (*symlink)(struct mnt_idmap *, struct inode *, struct dentry *, const char *); ++ int (*mkdir)(struct mnt_idmap *, struct inode *, struct dentry *, umode_t); ++ int (*rmdir)(struct inode *, struct dentry *); ++ int (*mknod)(struct mnt_idmap *, struct inode *, struct dentry *, umode_t, dev_t); ++ int (*rename)(struct mnt_idmap *, struct inode *, struct dentry *, struct inode *, struct dentry *, unsigned int); ++ int (*setattr)(struct mnt_idmap *, struct dentry *, struct iattr *); ++ int (*getattr)(struct mnt_idmap *, const struct path *, struct kstat *, u32, unsigned int); ++ ssize_t (*listxattr)(struct dentry *, char *, size_t); ++ int (*fiemap)(struct inode *, struct fiemap_extent_info *, u64, u64); ++ int (*update_time)(struct inode *, struct timespec64 *, int); ++ int (*atomic_open)(struct inode *, struct dentry *, struct file *, unsigned int, umode_t); ++ int (*tmpfile)(struct mnt_idmap *, struct inode *, struct file *, umode_t); ++ struct posix_acl * (*get_acl)(struct mnt_idmap *, struct dentry *, int); ++ int (*set_acl)(struct mnt_idmap *, struct dentry *, struct posix_acl *, int); ++ int (*fileattr_set)(struct mnt_idmap *, struct dentry *, struct fileattr *); ++ int (*fileattr_get)(struct dentry *, struct fileattr *); ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct fasync_struct { ++ rwlock_t fa_lock; ++ int magic; ++ int fa_fd; ++ struct fasync_struct *fa_next; ++ struct file *fa_file; ++ struct callback_head fa_rcu; ++}; ++ ++struct kstatfs; ++ ++struct super_operations { ++ struct inode * (*alloc_inode)(struct super_block *); ++ void (*destroy_inode)(struct inode *); ++ void (*free_inode)(struct inode *); ++ void (*dirty_inode)(struct inode *, int); ++ int (*write_inode)(struct inode *, struct writeback_control *); ++ int (*drop_inode)(struct inode *); ++ void (*evict_inode)(struct inode *); ++ void (*put_super)(struct super_block *); ++ int (*sync_fs)(struct super_block *, int); ++ int (*freeze_super)(struct super_block *); ++ int (*freeze_fs)(struct super_block *); ++ int (*thaw_super)(struct super_block *); ++ int (*unfreeze_fs)(struct super_block *); ++ int (*statfs)(struct dentry *, struct kstatfs *); ++ int (*remount_fs)(struct super_block *, int *, char *); ++ void (*umount_begin)(struct super_block *); ++ int (*show_options)(struct seq_file *, struct dentry *); ++ int (*show_devname)(struct seq_file *, struct dentry *); ++ int (*show_path)(struct seq_file *, struct dentry *); ++ int (*show_stats)(struct seq_file *, struct dentry *); ++ ssize_t (*quota_read)(struct super_block *, int, char *, size_t, loff_t); ++ ssize_t (*quota_write)(struct super_block *, int, const char *, size_t, loff_t); ++ struct dquot ** (*get_dquots)(struct inode *); ++ long int (*nr_cached_objects)(struct super_block *, struct shrink_control *); ++ long int (*free_cached_objects)(struct super_block *, struct shrink_control *); ++}; ++ ++struct fid; ++ ++struct iomap; ++ ++struct export_operations { ++ int (*encode_fh)(struct inode *, __u32 *, int *, struct inode *); ++ struct dentry * (*fh_to_dentry)(struct super_block *, struct fid *, int, int); ++ struct dentry * (*fh_to_parent)(struct super_block *, struct fid *, int, int); ++ int (*get_name)(struct dentry *, char *, struct dentry *); ++ struct dentry * (*get_parent)(struct dentry *); ++ int (*commit_metadata)(struct inode *); ++ int (*get_uuid)(struct super_block *, u8 *, u32 *, u64 *); ++ int (*map_blocks)(struct inode *, loff_t, u64, struct iomap *, bool, u32 *); ++ int (*commit_blocks)(struct inode *, struct iomap *, int, struct iattr *); ++ long unsigned int flags; ++}; ++ ++struct xattr_handler { ++ const char *name; ++ const char *prefix; ++ int flags; ++ bool (*list)(struct dentry *); ++ int (*get)(const struct xattr_handler *, struct dentry *, struct inode *, const char *, void *, size_t); ++ int (*set)(const struct xattr_handler *, struct mnt_idmap *, struct dentry *, struct inode *, const char *, const void *, size_t, int); ++}; ++ ++union fscrypt_policy; ++ ++struct fscrypt_operations { ++ unsigned int flags; ++ const char *key_prefix; ++ int (*get_context)(struct inode *, void *, size_t); ++ int (*set_context)(struct inode *, const void *, size_t, void *); ++ const union fscrypt_policy * (*get_dummy_policy)(struct super_block *); ++ bool (*empty_dir)(struct inode *); ++ bool (*has_stable_inodes)(struct super_block *); ++ void (*get_ino_and_lblk_bits)(struct super_block *, int *, int *); ++ struct block_device ** (*get_devices)(struct super_block *, unsigned int *); ++}; ++ ++struct fsverity_operations { ++ int (*begin_enable_verity)(struct file *); ++ int (*end_enable_verity)(struct file *, const void *, size_t, u64); ++ int (*get_verity_descriptor)(struct inode *, void *, size_t); ++ struct page * (*read_merkle_tree_page)(struct inode *, long unsigned int, long unsigned int); ++ int (*write_merkle_tree_block)(struct inode *, const void *, u64, unsigned int); ++}; ++ ++enum dl_dev_state { ++ DL_DEV_NO_DRIVER = 0, ++ DL_DEV_PROBING = 1, ++ DL_DEV_DRIVER_BOUND = 2, ++ DL_DEV_UNBINDING = 3, ++}; ++ ++struct dev_links_info { ++ struct list_head suppliers; ++ struct list_head consumers; ++ struct list_head defer_sync; ++ enum dl_dev_state status; ++}; ++ ++struct pm_message { ++ int event; ++}; ++ ++typedef struct pm_message pm_message_t; ++ ++struct pm_subsys_data; ++ ++struct dev_pm_qos; ++ ++struct dev_pm_info { ++ pm_message_t power_state; ++ unsigned int can_wakeup: 1; ++ unsigned int async_suspend: 1; ++ bool in_dpm_list: 1; ++ bool is_prepared: 1; ++ bool is_suspended: 1; ++ bool is_noirq_suspended: 1; ++ bool is_late_suspended: 1; ++ bool no_pm: 1; ++ bool early_init: 1; ++ bool direct_complete: 1; ++ u32 driver_flags; ++ spinlock_t lock; ++ unsigned int should_wakeup: 1; ++ struct pm_subsys_data *subsys_data; ++ void (*set_latency_tolerance)(struct device *, s32); ++ struct dev_pm_qos *qos; ++}; ++ ++struct irq_domain; ++ ++struct msi_device_data; ++ ++struct dev_msi_info { ++ struct irq_domain *domain; ++ struct msi_device_data *data; ++}; ++ ++struct dev_archdata {}; ++ ++enum device_removable { ++ DEVICE_REMOVABLE_NOT_SUPPORTED = 0, ++ DEVICE_REMOVABLE_UNKNOWN = 1, ++ DEVICE_FIXED = 2, ++ DEVICE_REMOVABLE = 3, ++}; ++ ++struct device_private; ++ ++struct device_type; ++ ++struct bus_type; ++ ++struct device_driver; ++ ++struct dev_pm_domain; ++ ++struct dma_map_ops; ++ ++struct bus_dma_region; ++ ++struct device_dma_parameters; ++ ++struct io_tlb_mem; ++ ++struct device_node; ++ ++struct fwnode_handle; ++ ++struct class; ++ ++struct attribute_group; ++ ++struct iommu_group; ++ ++struct dev_iommu; ++ ++struct device_physical_location; ++ ++struct device { ++ struct kobject kobj; ++ struct device *parent; ++ struct device_private *p; ++ const char *init_name; ++ const struct device_type *type; ++ const struct bus_type *bus; ++ struct device_driver *driver; ++ void *platform_data; ++ void *driver_data; ++ struct mutex mutex; ++ struct dev_links_info links; ++ struct dev_pm_info power; ++ struct dev_pm_domain *pm_domain; ++ struct dev_msi_info msi; ++ const struct dma_map_ops *dma_ops; ++ u64 *dma_mask; ++ u64 coherent_dma_mask; ++ u64 bus_dma_limit; ++ const struct bus_dma_region *dma_range_map; ++ struct device_dma_parameters *dma_parms; ++ struct list_head dma_pools; ++ struct io_tlb_mem *dma_io_tlb_mem; ++ struct dev_archdata archdata; ++ struct device_node *of_node; ++ struct fwnode_handle *fwnode; ++ int numa_node; ++ dev_t devt; ++ u32 id; ++ spinlock_t devres_lock; ++ struct list_head devres_head; ++ const struct class *class; ++ const struct attribute_group **groups; ++ void (*release)(struct device *); ++ struct iommu_group *iommu_group; ++ struct dev_iommu *iommu; ++ struct device_physical_location *physical_location; ++ enum device_removable removable; ++ bool offline_disabled: 1; ++ bool offline: 1; ++ bool of_node_reused: 1; ++ bool state_synced: 1; ++ bool can_match: 1; ++}; ++ ++struct disk_stats; ++ ++struct partition_meta_info; ++ ++struct block_device { ++ sector_t bd_start_sect; ++ sector_t bd_nr_sectors; ++ struct gendisk *bd_disk; ++ struct request_queue *bd_queue; ++ struct disk_stats *bd_stats; ++ long unsigned int bd_stamp; ++ bool bd_read_only; ++ u8 bd_partno; ++ bool bd_write_holder; ++ bool bd_has_submit_bio; ++ dev_t bd_dev; ++ atomic_t bd_openers; ++ spinlock_t bd_size_lock; ++ struct inode *bd_inode; ++ struct super_block *bd_super; ++ void *bd_claiming; ++ void *bd_holder; ++ int bd_fsfreeze_count; ++ int bd_holders; ++ struct kobject *bd_holder_dir; ++ struct mutex bd_fsfreeze_mutex; ++ struct super_block *bd_fsfreeze_sb; ++ struct partition_meta_info *bd_meta_info; ++ struct device bd_device; ++}; ++ ++typedef bool (*filldir_t)(struct dir_context *, const char *, int, loff_t, u64, unsigned int); ++ ++struct dir_context { ++ filldir_t actor; ++ loff_t pos; ++}; ++ ++struct io_comp_batch { ++ struct request *req_list; ++ bool need_ts; ++ void (*complete)(struct io_comp_batch *); ++}; ++ ++typedef void (*poll_queue_proc)(struct file *, wait_queue_head_t *, struct poll_table_struct *); ++ ++struct poll_table_struct { ++ poll_queue_proc _qproc; ++ __poll_t _key; ++}; ++ ++struct seq_file { ++ char *buf; ++ size_t size; ++ size_t from; ++ size_t count; ++ size_t pad_until; ++ loff_t index; ++ loff_t read_pos; ++ struct mutex lock; ++ const struct seq_operations *op; ++ int poll_event; ++ const struct file *file; ++ void *private; ++}; ++ ++struct fc_log; ++ ++struct p_log { ++ const char *prefix; ++ struct fc_log *log; ++}; ++ ++enum fs_context_purpose { ++ FS_CONTEXT_FOR_MOUNT = 0, ++ FS_CONTEXT_FOR_SUBMOUNT = 1, ++ FS_CONTEXT_FOR_RECONFIGURE = 2, ++}; ++ ++enum fs_context_phase { ++ FS_CONTEXT_CREATE_PARAMS = 0, ++ FS_CONTEXT_CREATING = 1, ++ FS_CONTEXT_AWAITING_MOUNT = 2, ++ FS_CONTEXT_AWAITING_RECONF = 3, ++ FS_CONTEXT_RECONF_PARAMS = 4, ++ FS_CONTEXT_RECONFIGURING = 5, ++ FS_CONTEXT_FAILED = 6, ++}; ++ ++struct fs_context_operations; ++ ++struct fs_context { ++ const struct fs_context_operations *ops; ++ struct mutex uapi_mutex; ++ struct file_system_type *fs_type; ++ void *fs_private; ++ void *sget_key; ++ struct dentry *root; ++ struct user_namespace *user_ns; ++ struct net *net_ns; ++ const struct cred *cred; ++ struct p_log log; ++ const char *source; ++ void *security; ++ void *s_fs_info; ++ unsigned int sb_flags; ++ unsigned int sb_flags_mask; ++ unsigned int s_iflags; ++ enum fs_context_purpose purpose: 8; ++ enum fs_context_phase phase: 8; ++ bool need_free: 1; ++ bool global: 1; ++ bool oldapi: 1; ++}; ++ ++struct fs_parameter; ++ ++struct fs_parse_result; ++ ++typedef int fs_param_type(struct p_log *, const struct fs_parameter_spec *, struct fs_parameter *, struct fs_parse_result *); ++ ++struct fs_parameter_spec { ++ const char *name; ++ fs_param_type *type; ++ u8 opt; ++ short unsigned int flags; ++ const void *data; ++}; ++ ++struct audit_names; ++ ++struct filename { ++ const char *name; ++ const char *uptr; ++ int refcnt; ++ struct audit_names *aname; ++ const char iname[0]; ++}; ++ ++typedef __u32 blk_opf_t; ++ ++typedef u8 blk_status_t; ++ ++struct bvec_iter { ++ sector_t bi_sector; ++ unsigned int bi_size; ++ unsigned int bi_idx; ++ unsigned int bi_bvec_done; ++} __attribute__((packed)); ++ ++typedef unsigned int blk_qc_t; ++ ++typedef void bio_end_io_t(struct bio *); ++ ++struct bio_issue { ++ u64 value; ++}; ++ ++struct bio_crypt_ctx; ++ ++struct bio_integrity_payload; ++ ++struct bio { ++ struct bio *bi_next; ++ struct block_device *bi_bdev; ++ blk_opf_t bi_opf; ++ short unsigned int bi_flags; ++ short unsigned int bi_ioprio; ++ blk_status_t bi_status; ++ atomic_t __bi_remaining; ++ struct bvec_iter bi_iter; ++ blk_qc_t bi_cookie; ++ bio_end_io_t *bi_end_io; ++ void *bi_private; ++ struct blkcg_gq *bi_blkg; ++ struct bio_issue bi_issue; ++ u64 bi_iocost_cost; ++ struct bio_crypt_ctx *bi_crypt_context; ++ union { ++ struct bio_integrity_payload *bi_integrity; ++ }; ++ short unsigned int bi_vcnt; ++ short unsigned int bi_max_vecs; ++ atomic_t __bi_cnt; ++ struct bio_vec *bi_io_vec; ++ struct bio_set *bi_pool; ++ struct bio_vec bi_inline_vecs[0]; ++}; ++ ++struct tracepoint_func { ++ void *func; ++ void *data; ++ int prio; ++}; ++ ++struct tracepoint { ++ const char *name; ++ struct static_key key; ++ struct static_call_key *static_call_key; ++ void *static_call_tramp; ++ void *iterator; ++ int (*regfunc)(); ++ void (*unregfunc)(); ++ struct tracepoint_func *funcs; ++}; ++ ++struct bpf_raw_event_map { ++ struct tracepoint *tp; ++ void *bpf_func; ++ u32 num_args; ++ u32 writable_size; ++ long: 64; ++}; ++ ++struct dev_pagemap_ops { ++ void (*page_free)(struct page *); ++ vm_fault_t (*migrate_to_ram)(struct vm_fault *); ++ int (*memory_failure)(struct dev_pagemap *, long unsigned int, long unsigned int, int); ++}; ++ ++struct attribute; ++ ++struct bin_attribute; ++ ++struct attribute_group { ++ const char *name; ++ umode_t (*is_visible)(struct kobject *, struct attribute *, int); ++ umode_t (*is_bin_visible)(struct kobject *, struct bin_attribute *, int); ++ struct attribute **attrs; ++ struct bin_attribute **bin_attrs; ++}; ++ ++typedef __u64 Elf64_Addr; ++ ++typedef __u16 Elf64_Half; ++ ++typedef __u64 Elf64_Off; ++ ++typedef __u32 Elf64_Word; ++ ++typedef __u64 Elf64_Xword; ++ ++struct elf64_sym { ++ Elf64_Word st_name; ++ unsigned char st_info; ++ unsigned char st_other; ++ Elf64_Half st_shndx; ++ Elf64_Addr st_value; ++ Elf64_Xword st_size; ++}; ++ ++struct elf64_hdr { ++ unsigned char e_ident[16]; ++ Elf64_Half e_type; ++ Elf64_Half e_machine; ++ Elf64_Word e_version; ++ Elf64_Addr e_entry; ++ Elf64_Off e_phoff; ++ Elf64_Off e_shoff; ++ Elf64_Word e_flags; ++ Elf64_Half e_ehsize; ++ Elf64_Half e_phentsize; ++ Elf64_Half e_phnum; ++ Elf64_Half e_shentsize; ++ Elf64_Half e_shnum; ++ Elf64_Half e_shstrndx; ++}; ++ ++typedef struct elf64_hdr Elf64_Ehdr; ++ ++struct elf64_shdr { ++ Elf64_Word sh_name; ++ Elf64_Word sh_type; ++ Elf64_Xword sh_flags; ++ Elf64_Addr sh_addr; ++ Elf64_Off sh_offset; ++ Elf64_Xword sh_size; ++ Elf64_Word sh_link; ++ Elf64_Word sh_info; ++ Elf64_Xword sh_addralign; ++ Elf64_Xword sh_entsize; ++}; ++ ++typedef struct elf64_shdr Elf64_Shdr; ++ ++struct kernfs_root; ++ ++struct kernfs_elem_dir { ++ long unsigned int subdirs; ++ struct rb_root children; ++ struct kernfs_root *root; ++ long unsigned int rev; ++}; ++ ++struct kernfs_elem_symlink { ++ struct kernfs_node *target_kn; ++}; ++ ++struct kernfs_ops; ++ ++struct kernfs_open_node; ++ ++struct kernfs_elem_attr { ++ const struct kernfs_ops *ops; ++ struct kernfs_open_node *open; ++ loff_t size; ++ struct kernfs_node *notify_next; ++}; ++ ++struct kernfs_iattrs; ++ ++struct kernfs_node { ++ atomic_t count; ++ atomic_t active; ++ struct kernfs_node *parent; ++ const char *name; ++ struct rb_node rb; ++ const void *ns; ++ unsigned int hash; ++ union { ++ struct kernfs_elem_dir dir; ++ struct kernfs_elem_symlink symlink; ++ struct kernfs_elem_attr attr; ++ }; ++ void *priv; ++ u64 id; ++ short unsigned int flags; ++ umode_t mode; ++ struct kernfs_iattrs *iattr; ++}; ++ ++struct kernfs_open_file; ++ ++struct kernfs_ops { ++ int (*open)(struct kernfs_open_file *); ++ void (*release)(struct kernfs_open_file *); ++ int (*seq_show)(struct seq_file *, void *); ++ void * (*seq_start)(struct seq_file *, loff_t *); ++ void * (*seq_next)(struct seq_file *, void *, loff_t *); ++ void (*seq_stop)(struct seq_file *, void *); ++ ssize_t (*read)(struct kernfs_open_file *, char *, size_t, loff_t); ++ size_t atomic_write_len; ++ bool prealloc; ++ ssize_t (*write)(struct kernfs_open_file *, char *, size_t, loff_t); ++ __poll_t (*poll)(struct kernfs_open_file *, struct poll_table_struct *); ++ int (*mmap)(struct kernfs_open_file *, struct vm_area_struct *); ++}; ++ ++struct kernfs_open_file { ++ struct kernfs_node *kn; ++ struct file *file; ++ struct seq_file *seq_file; ++ void *priv; ++ struct mutex mutex; ++ struct mutex prealloc_mutex; ++ int event; ++ struct list_head list; ++ char *prealloc_buf; ++ size_t atomic_write_len; ++ bool mmapped: 1; ++ bool released: 1; ++ const struct vm_operations_struct *vm_ops; ++}; ++ ++enum kobj_ns_type { ++ KOBJ_NS_TYPE_NONE = 0, ++ KOBJ_NS_TYPE_NET = 1, ++ KOBJ_NS_TYPES = 2, ++}; ++ ++struct kobj_ns_type_operations { ++ enum kobj_ns_type type; ++ bool (*current_may_mount)(); ++ void * (*grab_current_ns)(); ++ const void * (*netlink_ns)(struct sock *); ++ const void * (*initial_ns)(); ++ void (*drop_ns)(void *); ++}; ++ ++struct attribute { ++ const char *name; ++ umode_t mode; ++}; ++ ++struct bin_attribute { ++ struct attribute attr; ++ size_t size; ++ void *private; ++ struct address_space * (*f_mapping)(); ++ ssize_t (*read)(struct file *, struct kobject *, struct bin_attribute *, char *, loff_t, size_t); ++ ssize_t (*write)(struct file *, struct kobject *, struct bin_attribute *, char *, loff_t, size_t); ++ int (*mmap)(struct file *, struct kobject *, struct bin_attribute *, struct vm_area_struct *); ++}; ++ ++struct sysfs_ops { ++ ssize_t (*show)(struct kobject *, struct attribute *, char *); ++ ssize_t (*store)(struct kobject *, struct attribute *, const char *, size_t); ++}; ++ ++struct kset_uevent_ops; ++ ++struct kset { ++ struct list_head list; ++ spinlock_t list_lock; ++ struct kobject kobj; ++ const struct kset_uevent_ops *uevent_ops; ++}; ++ ++struct kobj_type { ++ void (*release)(struct kobject *); ++ const struct sysfs_ops *sysfs_ops; ++ const struct attribute_group **default_groups; ++ const struct kobj_ns_type_operations * (*child_ns_type)(const struct kobject *); ++ const void * (*namespace)(const struct kobject *); ++ void (*get_ownership)(const struct kobject *, kuid_t *, kgid_t *); ++}; ++ ++struct kobj_uevent_env { ++ char *argv[3]; ++ char *envp[64]; ++ int envp_idx; ++ char buf[2048]; ++ int buflen; ++}; ++ ++struct kset_uevent_ops { ++ int (* const filter)(const struct kobject *); ++ const char * (* const name)(const struct kobject *); ++ int (* const uevent)(const struct kobject *, struct kobj_uevent_env *); ++}; ++ ++struct kparam_string; ++ ++struct kparam_array; ++ ++struct kernel_param { ++ const char *name; ++ struct module *mod; ++ const struct kernel_param_ops *ops; ++ const u16 perm; ++ s8 level; ++ u8 flags; ++ union { ++ void *arg; ++ const struct kparam_string *str; ++ const struct kparam_array *arr; ++ }; ++}; ++ ++struct kparam_string { ++ unsigned int maxlen; ++ char *string; ++}; ++ ++struct kparam_array { ++ unsigned int max; ++ unsigned int elemsize; ++ unsigned int *num; ++ const struct kernel_param_ops *ops; ++ void *elem; ++}; ++ ++struct mod_arch_syminfo { ++ long unsigned int got_offset; ++ long unsigned int plt_offset; ++ int got_initialized; ++ int plt_initialized; ++}; ++ ++struct module_attribute { ++ struct attribute attr; ++ ssize_t (*show)(struct module_attribute *, struct module_kobject *, char *); ++ ssize_t (*store)(struct module_attribute *, struct module_kobject *, const char *, size_t); ++ void (*setup)(struct module *, const char *); ++ int (*test)(struct module *); ++ void (*free)(struct module *); ++}; ++ ++struct klp_modinfo { ++ Elf64_Ehdr hdr; ++ Elf64_Shdr *sechdrs; ++ char *secstrings; ++ unsigned int symndx; ++}; ++ ++struct trace_event_functions; ++ ++struct trace_event { ++ struct hlist_node node; ++ int type; ++ struct trace_event_functions *funcs; ++}; ++ ++struct trace_event_class; ++ ++struct event_filter; ++ ++struct perf_event; ++ ++struct trace_event_call { ++ struct list_head list; ++ struct trace_event_class *class; ++ union { ++ char *name; ++ struct tracepoint *tp; ++ }; ++ struct trace_event event; ++ char *print_fmt; ++ struct event_filter *filter; ++ union { ++ void *module; ++ atomic_t refcnt; ++ }; ++ void *data; ++ int flags; ++ int perf_refcount; ++ struct hlist_head *perf_events; ++ struct bpf_prog_array *prog_array; ++ int (*perf_perm)(struct trace_event_call *, struct perf_event *); ++}; ++ ++struct trace_eval_map { ++ const char *system; ++ const char *eval_string; ++ long unsigned int eval_value; ++}; ++ ++struct tty_buffer { ++ union { ++ struct tty_buffer *next; ++ struct llist_node free; ++ }; ++ int used; ++ int size; ++ int commit; ++ int lookahead; ++ int read; ++ bool flags; ++ long unsigned int data[0]; ++}; ++ ++struct tty_bufhead { ++ struct tty_buffer *head; ++ struct work_struct work; ++ struct mutex lock; ++ atomic_t priority; ++ struct tty_buffer sentinel; ++ struct llist_head free; ++ atomic_t mem_used; ++ int mem_limit; ++ struct tty_buffer *tail; ++}; ++ ++struct dev_pm_ops { ++ int (*prepare)(struct device *); ++ void (*complete)(struct device *); ++ int (*suspend)(struct device *); ++ int (*resume)(struct device *); ++ int (*freeze)(struct device *); ++ int (*thaw)(struct device *); ++ int (*poweroff)(struct device *); ++ int (*restore)(struct device *); ++ int (*suspend_late)(struct device *); ++ int (*resume_early)(struct device *); ++ int (*freeze_late)(struct device *); ++ int (*thaw_early)(struct device *); ++ int (*poweroff_late)(struct device *); ++ int (*restore_early)(struct device *); ++ int (*suspend_noirq)(struct device *); ++ int (*resume_noirq)(struct device *); ++ int (*freeze_noirq)(struct device *); ++ int (*thaw_noirq)(struct device *); ++ int (*poweroff_noirq)(struct device *); ++ int (*restore_noirq)(struct device *); ++ int (*runtime_suspend)(struct device *); ++ int (*runtime_resume)(struct device *); ++ int (*runtime_idle)(struct device *); ++}; ++ ++struct pm_subsys_data { ++ spinlock_t lock; ++ unsigned int refcount; ++}; ++ ++struct dev_pm_domain { ++ struct dev_pm_ops ops; ++ int (*start)(struct device *); ++ void (*detach)(struct device *, bool); ++ int (*activate)(struct device *); ++ void (*sync)(struct device *); ++ void (*dismiss)(struct device *); ++}; ++ ++struct iommu_ops; ++ ++struct bus_type { ++ const char *name; ++ const char *dev_name; ++ const struct attribute_group **bus_groups; ++ const struct attribute_group **dev_groups; ++ const struct attribute_group **drv_groups; ++ int (*match)(struct device *, struct device_driver *); ++ int (*uevent)(const struct device *, struct kobj_uevent_env *); ++ int (*probe)(struct device *); ++ void (*sync_state)(struct device *); ++ void (*remove)(struct device *); ++ void (*shutdown)(struct device *); ++ int (*online)(struct device *); ++ int (*offline)(struct device *); ++ int (*suspend)(struct device *, pm_message_t); ++ int (*resume)(struct device *); ++ int (*num_vf)(struct device *); ++ int (*dma_configure)(struct device *); ++ void (*dma_cleanup)(struct device *); ++ const struct dev_pm_ops *pm; ++ const struct iommu_ops *iommu_ops; ++ bool need_parent_lock; ++}; ++ ++enum probe_type { ++ PROBE_DEFAULT_STRATEGY = 0, ++ PROBE_PREFER_ASYNCHRONOUS = 1, ++ PROBE_FORCE_SYNCHRONOUS = 2, ++}; ++ ++struct of_device_id; ++ ++struct acpi_device_id; ++ ++struct driver_private; ++ ++struct device_driver { ++ const char *name; ++ const struct bus_type *bus; ++ struct module *owner; ++ const char *mod_name; ++ bool suppress_bind_attrs; ++ enum probe_type probe_type; ++ const struct of_device_id *of_match_table; ++ const struct acpi_device_id *acpi_match_table; ++ int (*probe)(struct device *); ++ void (*sync_state)(struct device *); ++ int (*remove)(struct device *); ++ void (*shutdown)(struct device *); ++ int (*suspend)(struct device *, pm_message_t); ++ int (*resume)(struct device *); ++ const struct attribute_group **groups; ++ const struct attribute_group **dev_groups; ++ const struct dev_pm_ops *pm; ++ void (*coredump)(struct device *); ++ struct driver_private *p; ++}; ++ ++enum iommu_cap { ++ IOMMU_CAP_CACHE_COHERENCY = 0, ++ IOMMU_CAP_NOEXEC = 1, ++ IOMMU_CAP_PRE_BOOT_PROTECTION = 2, ++ IOMMU_CAP_ENFORCE_CACHE_COHERENCY = 3, ++}; ++ ++enum iommu_dev_features { ++ IOMMU_DEV_FEAT_SVA = 0, ++ IOMMU_DEV_FEAT_IOPF = 1, ++}; ++ ++typedef unsigned int ioasid_t; ++ ++struct iommu_domain; ++ ++struct iommu_device; ++ ++struct of_phandle_args; ++ ++struct iommu_fault_event; ++ ++struct iommu_page_response; ++ ++struct iommu_domain_ops; ++ ++struct iommu_ops { ++ bool (*capable)(struct device *, enum iommu_cap); ++ struct iommu_domain * (*domain_alloc)(unsigned int); ++ struct iommu_device * (*probe_device)(struct device *); ++ void (*release_device)(struct device *); ++ void (*probe_finalize)(struct device *); ++ void (*set_platform_dma_ops)(struct device *); ++ struct iommu_group * (*device_group)(struct device *); ++ void (*get_resv_regions)(struct device *, struct list_head *); ++ int (*of_xlate)(struct device *, struct of_phandle_args *); ++ bool (*is_attach_deferred)(struct device *); ++ int (*dev_enable_feat)(struct device *, enum iommu_dev_features); ++ int (*dev_disable_feat)(struct device *, enum iommu_dev_features); ++ int (*page_response)(struct device *, struct iommu_fault_event *, struct iommu_page_response *); ++ int (*def_domain_type)(struct device *); ++ void (*remove_dev_pasid)(struct device *, ioasid_t); ++ const struct iommu_domain_ops *default_domain_ops; ++ long unsigned int pgsize_bitmap; ++ struct module *owner; ++}; ++ ++struct class { ++ const char *name; ++ const struct attribute_group **class_groups; ++ const struct attribute_group **dev_groups; ++ int (*dev_uevent)(const struct device *, struct kobj_uevent_env *); ++ char * (*devnode)(const struct device *, umode_t *); ++ void (*class_release)(const struct class *); ++ void (*dev_release)(struct device *); ++ int (*shutdown_pre)(struct device *); ++ const struct kobj_ns_type_operations *ns_type; ++ const void * (*namespace)(const struct device *); ++ void (*get_ownership)(const struct device *, kuid_t *, kgid_t *); ++ const struct dev_pm_ops *pm; ++}; ++ ++struct device_type { ++ const char *name; ++ const struct attribute_group **groups; ++ int (*uevent)(const struct device *, struct kobj_uevent_env *); ++ char * (*devnode)(const struct device *, umode_t *, kuid_t *, kgid_t *); ++ void (*release)(struct device *); ++ const struct dev_pm_ops *pm; ++}; ++ ++struct of_device_id { ++ char name[32]; ++ char type[32]; ++ char compatible[128]; ++ const void *data; ++}; ++ ++typedef long unsigned int kernel_ulong_t; ++ ++struct acpi_device_id { ++ __u8 id[16]; ++ kernel_ulong_t driver_data; ++ __u32 cls; ++ __u32 cls_msk; ++}; ++ ++struct device_dma_parameters { ++ unsigned int max_segment_size; ++ unsigned int min_align_mask; ++ long unsigned int segment_boundary_mask; ++}; ++ ++enum device_physical_location_panel { ++ DEVICE_PANEL_TOP = 0, ++ DEVICE_PANEL_BOTTOM = 1, ++ DEVICE_PANEL_LEFT = 2, ++ DEVICE_PANEL_RIGHT = 3, ++ DEVICE_PANEL_FRONT = 4, ++ DEVICE_PANEL_BACK = 5, ++ DEVICE_PANEL_UNKNOWN = 6, ++}; ++ ++enum device_physical_location_vertical_position { ++ DEVICE_VERT_POS_UPPER = 0, ++ DEVICE_VERT_POS_CENTER = 1, ++ DEVICE_VERT_POS_LOWER = 2, ++}; ++ ++enum device_physical_location_horizontal_position { ++ DEVICE_HORI_POS_LEFT = 0, ++ DEVICE_HORI_POS_CENTER = 1, ++ DEVICE_HORI_POS_RIGHT = 2, ++}; ++ ++struct device_physical_location { ++ enum device_physical_location_panel panel; ++ enum device_physical_location_vertical_position vertical_position; ++ enum device_physical_location_horizontal_position horizontal_position; ++ bool dock; ++ bool lid; ++}; ++ ++typedef u64 dma_addr_t; ++ ++struct sg_table; ++ ++struct scatterlist; ++ ++struct dma_map_ops { ++ unsigned int flags; ++ void * (*alloc)(struct device *, size_t, dma_addr_t *, gfp_t, long unsigned int); ++ void (*free)(struct device *, size_t, void *, dma_addr_t, long unsigned int); ++ struct page * (*alloc_pages)(struct device *, size_t, dma_addr_t *, enum dma_data_direction, gfp_t); ++ void (*free_pages)(struct device *, size_t, struct page *, dma_addr_t, enum dma_data_direction); ++ struct sg_table * (*alloc_noncontiguous)(struct device *, size_t, enum dma_data_direction, gfp_t, long unsigned int); ++ void (*free_noncontiguous)(struct device *, size_t, struct sg_table *, enum dma_data_direction); ++ int (*mmap)(struct device *, struct vm_area_struct *, void *, dma_addr_t, size_t, long unsigned int); ++ int (*get_sgtable)(struct device *, struct sg_table *, void *, dma_addr_t, size_t, long unsigned int); ++ dma_addr_t (*map_page)(struct device *, struct page *, long unsigned int, size_t, enum dma_data_direction, long unsigned int); ++ void (*unmap_page)(struct device *, dma_addr_t, size_t, enum dma_data_direction, long unsigned int); ++ int (*map_sg)(struct device *, struct scatterlist *, int, enum dma_data_direction, long unsigned int); ++ void (*unmap_sg)(struct device *, struct scatterlist *, int, enum dma_data_direction, long unsigned int); ++ dma_addr_t (*map_resource)(struct device *, phys_addr_t, size_t, enum dma_data_direction, long unsigned int); ++ void (*unmap_resource)(struct device *, dma_addr_t, size_t, enum dma_data_direction, long unsigned int); ++ void (*sync_single_for_cpu)(struct device *, dma_addr_t, size_t, enum dma_data_direction); ++ void (*sync_single_for_device)(struct device *, dma_addr_t, size_t, enum dma_data_direction); ++ void (*sync_sg_for_cpu)(struct device *, struct scatterlist *, int, enum dma_data_direction); ++ void (*sync_sg_for_device)(struct device *, struct scatterlist *, int, enum dma_data_direction); ++ void (*cache_sync)(struct device *, void *, size_t, enum dma_data_direction); ++ int (*dma_supported)(struct device *, u64); ++ u64 (*get_required_mask)(struct device *); ++ size_t (*max_mapping_size)(struct device *); ++ size_t (*opt_mapping_size)(); ++ long unsigned int (*get_merge_boundary)(struct device *); ++}; ++ ++struct bus_dma_region { ++ phys_addr_t cpu_start; ++ dma_addr_t dma_start; ++ u64 size; ++ u64 offset; ++}; ++ ++struct fwnode_operations; ++ ++struct fwnode_handle { ++ struct fwnode_handle *secondary; ++ const struct fwnode_operations *ops; ++ struct device *dev; ++ struct list_head suppliers; ++ struct list_head consumers; ++ u8 flags; ++}; ++ ++struct serial_icounter_struct; ++ ++struct serial_struct; ++ ++struct tty_operations { ++ struct tty_struct * (*lookup)(struct tty_driver *, struct file *, int); ++ int (*install)(struct tty_driver *, struct tty_struct *); ++ void (*remove)(struct tty_driver *, struct tty_struct *); ++ int (*open)(struct tty_struct *, struct file *); ++ void (*close)(struct tty_struct *, struct file *); ++ void (*shutdown)(struct tty_struct *); ++ void (*cleanup)(struct tty_struct *); ++ int (*write)(struct tty_struct *, const unsigned char *, int); ++ int (*put_char)(struct tty_struct *, unsigned char); ++ void (*flush_chars)(struct tty_struct *); ++ unsigned int (*write_room)(struct tty_struct *); ++ unsigned int (*chars_in_buffer)(struct tty_struct *); ++ int (*ioctl)(struct tty_struct *, unsigned int, long unsigned int); ++ long int (*compat_ioctl)(struct tty_struct *, unsigned int, long unsigned int); ++ void (*set_termios)(struct tty_struct *, const struct ktermios *); ++ void (*throttle)(struct tty_struct *); ++ void (*unthrottle)(struct tty_struct *); ++ void (*stop)(struct tty_struct *); ++ void (*start)(struct tty_struct *); ++ void (*hangup)(struct tty_struct *); ++ int (*break_ctl)(struct tty_struct *, int); ++ void (*flush_buffer)(struct tty_struct *); ++ void (*set_ldisc)(struct tty_struct *); ++ void (*wait_until_sent)(struct tty_struct *, int); ++ void (*send_xchar)(struct tty_struct *, char); ++ int (*tiocmget)(struct tty_struct *); ++ int (*tiocmset)(struct tty_struct *, unsigned int, unsigned int); ++ int (*resize)(struct tty_struct *, struct winsize *); ++ int (*get_icount)(struct tty_struct *, struct serial_icounter_struct *); ++ int (*get_serial)(struct tty_struct *, struct serial_struct *); ++ int (*set_serial)(struct tty_struct *, struct serial_struct *); ++ void (*show_fdinfo)(struct tty_struct *, struct seq_file *); ++ int (*proc_show)(struct seq_file *, void *); ++}; ++ ++struct proc_dir_entry; ++ ++struct tty_driver { ++ struct kref kref; ++ struct cdev **cdevs; ++ struct module *owner; ++ const char *driver_name; ++ const char *name; ++ int name_base; ++ int major; ++ int minor_start; ++ unsigned int num; ++ short int type; ++ short int subtype; ++ struct ktermios init_termios; ++ long unsigned int flags; ++ struct proc_dir_entry *proc_entry; ++ struct tty_driver *other; ++ struct tty_struct **ttys; ++ struct tty_port **ports; ++ struct ktermios **termios; ++ void *driver_state; ++ const struct tty_operations *ops; ++ struct list_head tty_drivers; ++}; ++ ++struct __kfifo { ++ unsigned int in; ++ unsigned int out; ++ unsigned int mask; ++ unsigned int esize; ++ void *data; ++}; ++ ++struct tty_port_operations; ++ ++struct tty_port_client_operations; ++ ++struct tty_port { ++ struct tty_bufhead buf; ++ struct tty_struct *tty; ++ struct tty_struct *itty; ++ const struct tty_port_operations *ops; ++ const struct tty_port_client_operations *client_ops; ++ spinlock_t lock; ++ int blocked_open; ++ int count; ++ wait_queue_head_t open_wait; ++ wait_queue_head_t delta_msr_wait; ++ long unsigned int flags; ++ long unsigned int iflags; ++ unsigned char console: 1; ++ struct mutex mutex; ++ struct mutex buf_mutex; ++ unsigned char *xmit_buf; ++ struct { ++ union { ++ struct __kfifo kfifo; ++ unsigned char *type; ++ const unsigned char *const_type; ++ char (*rectype)[0]; ++ unsigned char *ptr; ++ const unsigned char *ptr_const; ++ }; ++ unsigned char buf[0]; ++ } xmit_fifo; ++ unsigned int close_delay; ++ unsigned int closing_wait; ++ int drain_delay; ++ struct kref kref; ++ void *client_data; ++}; ++ ++struct tty_ldisc_ops { ++ char *name; ++ int num; ++ int (*open)(struct tty_struct *); ++ void (*close)(struct tty_struct *); ++ void (*flush_buffer)(struct tty_struct *); ++ ssize_t (*read)(struct tty_struct *, struct file *, unsigned char *, size_t, void **, long unsigned int); ++ ssize_t (*write)(struct tty_struct *, struct file *, const unsigned char *, size_t); ++ int (*ioctl)(struct tty_struct *, unsigned int, long unsigned int); ++ int (*compat_ioctl)(struct tty_struct *, unsigned int, long unsigned int); ++ void (*set_termios)(struct tty_struct *, const struct ktermios *); ++ __poll_t (*poll)(struct tty_struct *, struct file *, struct poll_table_struct *); ++ void (*hangup)(struct tty_struct *); ++ void (*receive_buf)(struct tty_struct *, const unsigned char *, const char *, int); ++ void (*write_wakeup)(struct tty_struct *); ++ void (*dcd_change)(struct tty_struct *, bool); ++ int (*receive_buf2)(struct tty_struct *, const unsigned char *, const char *, int); ++ void (*lookahead_buf)(struct tty_struct *, const unsigned char *, const unsigned char *, unsigned int); ++ struct module *owner; ++}; ++ ++struct tty_ldisc { ++ struct tty_ldisc_ops *ops; ++ struct tty_struct *tty; ++}; ++ ++enum dev_dma_attr { ++ DEV_DMA_NOT_SUPPORTED = 0, ++ DEV_DMA_NON_COHERENT = 1, ++ DEV_DMA_COHERENT = 2, ++}; ++ ++struct fwnode_reference_args; ++ ++struct fwnode_endpoint; ++ ++struct fwnode_operations { ++ struct fwnode_handle * (*get)(struct fwnode_handle *); ++ void (*put)(struct fwnode_handle *); ++ bool (*device_is_available)(const struct fwnode_handle *); ++ const void * (*device_get_match_data)(const struct fwnode_handle *, const struct device *); ++ bool (*device_dma_supported)(const struct fwnode_handle *); ++ enum dev_dma_attr (*device_get_dma_attr)(const struct fwnode_handle *); ++ bool (*property_present)(const struct fwnode_handle *, const char *); ++ int (*property_read_int_array)(const struct fwnode_handle *, const char *, unsigned int, void *, size_t); ++ int (*property_read_string_array)(const struct fwnode_handle *, const char *, const char **, size_t); ++ const char * (*get_name)(const struct fwnode_handle *); ++ const char * (*get_name_prefix)(const struct fwnode_handle *); ++ struct fwnode_handle * (*get_parent)(const struct fwnode_handle *); ++ struct fwnode_handle * (*get_next_child_node)(const struct fwnode_handle *, struct fwnode_handle *); ++ struct fwnode_handle * (*get_named_child_node)(const struct fwnode_handle *, const char *); ++ int (*get_reference_args)(const struct fwnode_handle *, const char *, const char *, unsigned int, unsigned int, struct fwnode_reference_args *); ++ struct fwnode_handle * (*graph_get_next_endpoint)(const struct fwnode_handle *, struct fwnode_handle *); ++ struct fwnode_handle * (*graph_get_remote_endpoint)(const struct fwnode_handle *); ++ struct fwnode_handle * (*graph_get_port_parent)(struct fwnode_handle *); ++ int (*graph_parse_endpoint)(const struct fwnode_handle *, struct fwnode_endpoint *); ++ void * (*iomap)(struct fwnode_handle *, int); ++ int (*irq_get)(const struct fwnode_handle *, unsigned int); ++ int (*add_links)(struct fwnode_handle *); ++}; ++ ++struct fwnode_endpoint { ++ unsigned int port; ++ unsigned int id; ++ const struct fwnode_handle *local_fwnode; ++}; ++ ++struct fwnode_reference_args { ++ struct fwnode_handle *fwnode; ++ unsigned int nargs; ++ u64 args[8]; ++}; ++ ++struct tty_port_operations { ++ bool (*carrier_raised)(struct tty_port *); ++ void (*dtr_rts)(struct tty_port *, bool); ++ void (*shutdown)(struct tty_port *); ++ int (*activate)(struct tty_port *, struct tty_struct *); ++ void (*destruct)(struct tty_port *); ++}; ++ ++struct tty_port_client_operations { ++ int (*receive_buf)(struct tty_port *, const unsigned char *, const unsigned char *, size_t); ++ void (*lookahead_buf)(struct tty_port *, const unsigned char *, const unsigned char *, unsigned int); ++ void (*write_wakeup)(struct tty_port *); ++}; ++ ++struct nsset; ++ ++struct proc_ns_operations { ++ const char *name; ++ const char *real_ns_name; ++ int type; ++ struct ns_common * (*get)(struct task_struct *); ++ void (*put)(struct ns_common *); ++ int (*install)(struct nsset *, struct ns_common *); ++ struct user_namespace * (*owner)(struct ns_common *); ++ struct ns_common * (*get_parent)(struct ns_common *); ++}; ++ ++struct new_utsname { ++ char sysname[65]; ++ char nodename[65]; ++ char release[65]; ++ char version[65]; ++ char machine[65]; ++ char domainname[65]; ++}; ++ ++struct uts_namespace { ++ struct new_utsname name; ++ struct user_namespace *user_ns; ++ struct ucounts *ucounts; ++ struct ns_common ns; ++}; ++ ++struct ref_tracker_dir {}; ++ ++struct prot_inuse; ++ ++struct netns_core { ++ struct ctl_table_header *sysctl_hdr; ++ int sysctl_somaxconn; ++ u8 sysctl_txrehash; ++ struct prot_inuse *prot_inuse; ++ struct cpumask *rps_default_mask; ++}; ++ ++struct ipstats_mib; ++ ++struct tcp_mib; ++ ++struct linux_mib; ++ ++struct udp_mib; ++ ++struct linux_xfrm_mib; ++ ++struct linux_tls_mib; ++ ++struct mptcp_mib; ++ ++struct icmp_mib; ++ ++struct icmpmsg_mib; ++ ++struct icmpv6_mib; ++ ++struct icmpv6msg_mib; ++ ++struct netns_mib { ++ struct ipstats_mib *ip_statistics; ++ struct ipstats_mib *ipv6_statistics; ++ struct tcp_mib *tcp_statistics; ++ struct linux_mib *net_statistics; ++ struct udp_mib *udp_statistics; ++ struct udp_mib *udp_stats_in6; ++ struct linux_xfrm_mib *xfrm_statistics; ++ struct linux_tls_mib *tls_statistics; ++ struct mptcp_mib *mptcp_statistics; ++ struct udp_mib *udplite_statistics; ++ struct udp_mib *udplite_stats_in6; ++ struct icmp_mib *icmp_statistics; ++ struct icmpmsg_mib *icmpmsg_statistics; ++ struct icmpv6_mib *icmpv6_statistics; ++ struct icmpv6msg_mib *icmpv6msg_statistics; ++ struct proc_dir_entry *proc_net_devsnmp6; ++}; ++ ++struct netns_packet { ++ struct mutex sklist_lock; ++ struct hlist_head sklist; ++}; ++ ++struct unix_table { ++ spinlock_t *locks; ++ struct hlist_head *buckets; ++}; ++ ++struct netns_unix { ++ struct unix_table table; ++ int sysctl_max_dgram_qlen; ++ struct ctl_table_header *ctl; ++}; ++ ++struct netns_nexthop { ++ struct rb_root rb_root; ++ struct hlist_head *devhash; ++ unsigned int seq; ++ u32 last_id_allocated; ++ struct blocking_notifier_head notifier_chain; ++}; ++ ++struct inet_hashinfo; ++ ++struct inet_timewait_death_row { ++ refcount_t tw_refcount; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct inet_hashinfo *hashinfo; ++ int sysctl_max_tw_buckets; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct local_ports { ++ seqlock_t lock; ++ int range[2]; ++ bool warned; ++}; ++ ++struct ping_group_range { ++ seqlock_t lock; ++ kgid_t range[2]; ++}; ++ ++typedef struct { ++ u64 key[2]; ++} siphash_key_t; ++ ++struct udp_table; ++ ++struct ipv4_devconf; ++ ++struct ip_ra_chain; ++ ++struct fib_rules_ops; ++ ++struct fib_table; ++ ++struct inet_peer_base; ++ ++struct fqdir; ++ ++struct tcp_congestion_ops; ++ ++struct tcp_fastopen_context; ++ ++struct fib_notifier_ops; ++ ++struct netns_ipv4 { ++ struct inet_timewait_death_row tcp_death_row; ++ struct udp_table *udp_table; ++ struct ctl_table_header *forw_hdr; ++ struct ctl_table_header *frags_hdr; ++ struct ctl_table_header *ipv4_hdr; ++ struct ctl_table_header *route_hdr; ++ struct ctl_table_header *xfrm4_hdr; ++ struct ipv4_devconf *devconf_all; ++ struct ipv4_devconf *devconf_dflt; ++ struct ip_ra_chain *ra_chain; ++ struct mutex ra_mutex; ++ struct fib_rules_ops *rules_ops; ++ struct fib_table *fib_main; ++ struct fib_table *fib_default; ++ unsigned int fib_rules_require_fldissect; ++ bool fib_has_custom_rules; ++ bool fib_has_custom_local_routes; ++ bool fib_offload_disabled; ++ atomic_t fib_num_tclassid_users; ++ struct hlist_head *fib_table_hash; ++ struct sock *fibnl; ++ struct sock *mc_autojoin_sk; ++ struct inet_peer_base *peers; ++ struct fqdir *fqdir; ++ u8 sysctl_icmp_echo_ignore_all; ++ u8 sysctl_icmp_echo_enable_probe; ++ u8 sysctl_icmp_echo_ignore_broadcasts; ++ u8 sysctl_icmp_ignore_bogus_error_responses; ++ u8 sysctl_icmp_errors_use_inbound_ifaddr; ++ int sysctl_icmp_ratelimit; ++ int sysctl_icmp_ratemask; ++ u32 ip_rt_min_pmtu; ++ int ip_rt_mtu_expires; ++ int ip_rt_min_advmss; ++ struct local_ports ip_local_ports; ++ u8 sysctl_tcp_ecn; ++ u8 sysctl_tcp_ecn_fallback; ++ u8 sysctl_ip_default_ttl; ++ u8 sysctl_ip_no_pmtu_disc; ++ u8 sysctl_ip_fwd_use_pmtu; ++ u8 sysctl_ip_fwd_update_priority; ++ u8 sysctl_ip_nonlocal_bind; ++ u8 sysctl_ip_autobind_reuse; ++ u8 sysctl_ip_dynaddr; ++ u8 sysctl_ip_early_demux; ++ u8 sysctl_raw_l3mdev_accept; ++ u8 sysctl_tcp_early_demux; ++ u8 sysctl_udp_early_demux; ++ u8 sysctl_nexthop_compat_mode; ++ u8 sysctl_fwmark_reflect; ++ u8 sysctl_tcp_fwmark_accept; ++ u8 sysctl_tcp_l3mdev_accept; ++ u8 sysctl_tcp_mtu_probing; ++ int sysctl_tcp_mtu_probe_floor; ++ int sysctl_tcp_base_mss; ++ int sysctl_tcp_min_snd_mss; ++ int sysctl_tcp_probe_threshold; ++ u32 sysctl_tcp_probe_interval; ++ int sysctl_tcp_keepalive_time; ++ int sysctl_tcp_keepalive_intvl; ++ u8 sysctl_tcp_keepalive_probes; ++ u8 sysctl_tcp_syn_retries; ++ u8 sysctl_tcp_synack_retries; ++ u8 sysctl_tcp_syncookies; ++ u8 sysctl_tcp_migrate_req; ++ u8 sysctl_tcp_comp_sack_nr; ++ int sysctl_tcp_reordering; ++ u8 sysctl_tcp_retries1; ++ u8 sysctl_tcp_retries2; ++ u8 sysctl_tcp_orphan_retries; ++ u8 sysctl_tcp_tw_reuse; ++ int sysctl_tcp_fin_timeout; ++ unsigned int sysctl_tcp_notsent_lowat; ++ u8 sysctl_tcp_sack; ++ u8 sysctl_tcp_window_scaling; ++ u8 sysctl_tcp_timestamps; ++ u8 sysctl_tcp_early_retrans; ++ u8 sysctl_tcp_recovery; ++ u8 sysctl_tcp_thin_linear_timeouts; ++ u8 sysctl_tcp_slow_start_after_idle; ++ u8 sysctl_tcp_retrans_collapse; ++ u8 sysctl_tcp_stdurg; ++ u8 sysctl_tcp_rfc1337; ++ u8 sysctl_tcp_abort_on_overflow; ++ u8 sysctl_tcp_fack; ++ int sysctl_tcp_max_reordering; ++ int sysctl_tcp_adv_win_scale; ++ u8 sysctl_tcp_dsack; ++ u8 sysctl_tcp_app_win; ++ u8 sysctl_tcp_frto; ++ u8 sysctl_tcp_nometrics_save; ++ u8 sysctl_tcp_no_ssthresh_metrics_save; ++ u8 sysctl_tcp_moderate_rcvbuf; ++ u8 sysctl_tcp_tso_win_divisor; ++ u8 sysctl_tcp_workaround_signed_windows; ++ int sysctl_tcp_limit_output_bytes; ++ int sysctl_tcp_challenge_ack_limit; ++ int sysctl_tcp_min_rtt_wlen; ++ u8 sysctl_tcp_min_tso_segs; ++ u8 sysctl_tcp_tso_rtt_log; ++ u8 sysctl_tcp_autocorking; ++ u8 sysctl_tcp_reflect_tos; ++ int sysctl_tcp_invalid_ratelimit; ++ int sysctl_tcp_pacing_ss_ratio; ++ int sysctl_tcp_pacing_ca_ratio; ++ int sysctl_tcp_wmem[3]; ++ int sysctl_tcp_rmem[3]; ++ unsigned int sysctl_tcp_child_ehash_entries; ++ long unsigned int sysctl_tcp_comp_sack_delay_ns; ++ long unsigned int sysctl_tcp_comp_sack_slack_ns; ++ int sysctl_max_syn_backlog; ++ int sysctl_tcp_fastopen; ++ const struct tcp_congestion_ops *tcp_congestion_control; ++ struct tcp_fastopen_context *tcp_fastopen_ctx; ++ unsigned int sysctl_tcp_fastopen_blackhole_timeout; ++ atomic_t tfo_active_disable_times; ++ long unsigned int tfo_active_disable_stamp; ++ u32 tcp_challenge_timestamp; ++ u32 tcp_challenge_count; ++ u8 sysctl_tcp_plb_enabled; ++ u8 sysctl_tcp_plb_idle_rehash_rounds; ++ u8 sysctl_tcp_plb_rehash_rounds; ++ u8 sysctl_tcp_plb_suspend_rto_sec; ++ int sysctl_tcp_plb_cong_thresh; ++ int sysctl_udp_wmem_min; ++ int sysctl_udp_rmem_min; ++ u8 sysctl_fib_notify_on_flag_change; ++ u8 sysctl_udp_l3mdev_accept; ++ u8 sysctl_igmp_llm_reports; ++ int sysctl_igmp_max_memberships; ++ int sysctl_igmp_max_msf; ++ int sysctl_igmp_qrv; ++ struct ping_group_range ping_group_range; ++ atomic_t dev_addr_genid; ++ unsigned int sysctl_udp_child_hash_entries; ++ long unsigned int *sysctl_local_reserved_ports; ++ int sysctl_ip_prot_sock; ++ struct list_head mr_tables; ++ struct fib_rules_ops *mr_rules_ops; ++ u32 sysctl_fib_multipath_hash_fields; ++ u8 sysctl_fib_multipath_use_neigh; ++ u8 sysctl_fib_multipath_hash_policy; ++ struct fib_notifier_ops *notifier_ops; ++ unsigned int fib_seq; ++ struct fib_notifier_ops *ipmr_notifier_ops; ++ unsigned int ipmr_seq; ++ atomic_t rt_genid; ++ siphash_key_t ip_id_key; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct neighbour; ++ ++struct dst_ops { ++ short unsigned int family; ++ unsigned int gc_thresh; ++ void (*gc)(struct dst_ops *); ++ struct dst_entry * (*check)(struct dst_entry *, __u32); ++ unsigned int (*default_advmss)(const struct dst_entry *); ++ unsigned int (*mtu)(const struct dst_entry *); ++ u32 * (*cow_metrics)(struct dst_entry *, long unsigned int); ++ void (*destroy)(struct dst_entry *); ++ void (*ifdown)(struct dst_entry *, struct net_device *, int); ++ struct dst_entry * (*negative_advice)(struct dst_entry *); ++ void (*link_failure)(struct sk_buff *); ++ void (*update_pmtu)(struct dst_entry *, struct sock *, struct sk_buff *, u32, bool); ++ void (*redirect)(struct dst_entry *, struct sock *, struct sk_buff *); ++ int (*local_out)(struct net *, struct sock *, struct sk_buff *); ++ struct neighbour * (*neigh_lookup)(const struct dst_entry *, struct sk_buff *, const void *); ++ void (*confirm_neigh)(const struct dst_entry *, const void *); ++ struct kmem_cache *kmem_cachep; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct percpu_counter pcpuc_entries; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct netns_sysctl_ipv6 { ++ struct ctl_table_header *hdr; ++ struct ctl_table_header *route_hdr; ++ struct ctl_table_header *icmp_hdr; ++ struct ctl_table_header *frags_hdr; ++ struct ctl_table_header *xfrm6_hdr; ++ int flush_delay; ++ int ip6_rt_max_size; ++ int ip6_rt_gc_min_interval; ++ int ip6_rt_gc_timeout; ++ int ip6_rt_gc_interval; ++ int ip6_rt_gc_elasticity; ++ int ip6_rt_mtu_expires; ++ int ip6_rt_min_advmss; ++ u32 multipath_hash_fields; ++ u8 multipath_hash_policy; ++ u8 bindv6only; ++ u8 flowlabel_consistency; ++ u8 auto_flowlabels; ++ int icmpv6_time; ++ u8 icmpv6_echo_ignore_all; ++ u8 icmpv6_echo_ignore_multicast; ++ u8 icmpv6_echo_ignore_anycast; ++ long unsigned int icmpv6_ratemask[4]; ++ long unsigned int *icmpv6_ratemask_ptr; ++ u8 anycast_src_echo_reply; ++ u8 ip_nonlocal_bind; ++ u8 fwmark_reflect; ++ u8 flowlabel_state_ranges; ++ int idgen_retries; ++ int idgen_delay; ++ int flowlabel_reflect; ++ int max_dst_opts_cnt; ++ int max_hbh_opts_cnt; ++ int max_dst_opts_len; ++ int max_hbh_opts_len; ++ int seg6_flowlabel; ++ u32 ioam6_id; ++ u64 ioam6_id_wide; ++ u8 skip_notify_on_dev_down; ++ u8 fib_notify_on_flag_change; ++ u8 icmpv6_error_anycast_as_unicast; ++}; ++ ++struct ipv6_devconf; ++ ++struct fib6_info; ++ ++struct rt6_info; ++ ++struct rt6_statistics; ++ ++struct fib6_table; ++ ++struct seg6_pernet_data; ++ ++struct ioam6_pernet_data; ++ ++struct netns_ipv6 { ++ struct dst_ops ip6_dst_ops; ++ struct netns_sysctl_ipv6 sysctl; ++ struct ipv6_devconf *devconf_all; ++ struct ipv6_devconf *devconf_dflt; ++ struct inet_peer_base *peers; ++ struct fqdir *fqdir; ++ struct fib6_info *fib6_null_entry; ++ struct rt6_info *ip6_null_entry; ++ struct rt6_statistics *rt6_stats; ++ struct timer_list ip6_fib_timer; ++ struct hlist_head *fib_table_hash; ++ struct fib6_table *fib6_main_tbl; ++ struct list_head fib6_walkers; ++ rwlock_t fib6_walker_lock; ++ spinlock_t fib6_gc_lock; ++ atomic_t ip6_rt_gc_expire; ++ long unsigned int ip6_rt_last_gc; ++ unsigned char flowlabel_has_excl; ++ bool fib6_has_custom_rules; ++ unsigned int fib6_rules_require_fldissect; ++ unsigned int fib6_routes_require_src; ++ struct rt6_info *ip6_prohibit_entry; ++ struct rt6_info *ip6_blk_hole_entry; ++ struct fib6_table *fib6_local_tbl; ++ struct fib_rules_ops *fib6_rules_ops; ++ struct sock *ndisc_sk; ++ struct sock *tcp_sk; ++ struct sock *igmp_sk; ++ struct sock *mc_autojoin_sk; ++ struct hlist_head *inet6_addr_lst; ++ spinlock_t addrconf_hash_lock; ++ struct delayed_work addr_chk_work; ++ struct list_head mr6_tables; ++ struct fib_rules_ops *mr6_rules_ops; ++ atomic_t dev_addr_genid; ++ atomic_t fib6_sernum; ++ struct seg6_pernet_data *seg6_data; ++ struct fib_notifier_ops *notifier_ops; ++ struct fib_notifier_ops *ip6mr_notifier_ops; ++ unsigned int ipmr_seq; ++ struct { ++ struct hlist_head head; ++ spinlock_t lock; ++ u32 seq; ++ } ip6addrlbl_table; ++ struct ioam6_pernet_data *ioam6_data; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct netns_sysctl_lowpan { ++ struct ctl_table_header *frags_hdr; ++}; ++ ++struct netns_ieee802154_lowpan { ++ struct netns_sysctl_lowpan sysctl; ++ struct fqdir *fqdir; ++}; ++ ++struct sctp_mib; ++ ++struct netns_sctp { ++ struct sctp_mib *sctp_statistics; ++ struct proc_dir_entry *proc_net_sctp; ++ struct ctl_table_header *sysctl_header; ++ struct sock *ctl_sock; ++ struct sock *udp4_sock; ++ struct sock *udp6_sock; ++ int udp_port; ++ int encap_port; ++ struct list_head local_addr_list; ++ struct list_head addr_waitq; ++ struct timer_list addr_wq_timer; ++ struct list_head auto_asconf_splist; ++ spinlock_t addr_wq_lock; ++ spinlock_t local_addr_lock; ++ unsigned int rto_initial; ++ unsigned int rto_min; ++ unsigned int rto_max; ++ int rto_alpha; ++ int rto_beta; ++ int max_burst; ++ int cookie_preserve_enable; ++ char *sctp_hmac_alg; ++ unsigned int valid_cookie_life; ++ unsigned int sack_timeout; ++ unsigned int hb_interval; ++ unsigned int probe_interval; ++ int max_retrans_association; ++ int max_retrans_path; ++ int max_retrans_init; ++ int pf_retrans; ++ int ps_retrans; ++ int pf_enable; ++ int pf_expose; ++ int sndbuf_policy; ++ int rcvbuf_policy; ++ int default_auto_asconf; ++ int addip_enable; ++ int addip_noauth; ++ int prsctp_enable; ++ int reconf_enable; ++ int auth_enable; ++ int intl_enable; ++ int ecn_enable; ++ int scope_policy; ++ int rwnd_upd_shift; ++ long unsigned int max_autoclose; ++ int l3mdev_accept; ++}; ++ ++struct nf_logger; ++ ++struct nf_hook_entries; ++ ++struct netns_nf { ++ struct proc_dir_entry *proc_netfilter; ++ const struct nf_logger *nf_loggers[11]; ++ struct ctl_table_header *nf_log_dir_header; ++ struct nf_hook_entries *hooks_ipv4[5]; ++ struct nf_hook_entries *hooks_ipv6[5]; ++ struct nf_hook_entries *hooks_arp[3]; ++ struct nf_hook_entries *hooks_bridge[5]; ++ unsigned int defrag_ipv4_users; ++ unsigned int defrag_ipv6_users; ++}; ++ ++struct nf_generic_net { ++ unsigned int timeout; ++}; ++ ++struct nf_tcp_net { ++ unsigned int timeouts[14]; ++ u8 tcp_loose; ++ u8 tcp_be_liberal; ++ u8 tcp_max_retrans; ++ u8 tcp_ignore_invalid_rst; ++ unsigned int offload_timeout; ++}; ++ ++struct nf_udp_net { ++ unsigned int timeouts[2]; ++ unsigned int offload_timeout; ++}; ++ ++struct nf_icmp_net { ++ unsigned int timeout; ++}; ++ ++struct nf_dccp_net { ++ u8 dccp_loose; ++ unsigned int dccp_timeout[10]; ++}; ++ ++struct nf_sctp_net { ++ unsigned int timeouts[10]; ++}; ++ ++struct nf_gre_net { ++ struct list_head keymap_list; ++ unsigned int timeouts[2]; ++}; ++ ++struct nf_ip_net { ++ struct nf_generic_net generic; ++ struct nf_tcp_net tcp; ++ struct nf_udp_net udp; ++ struct nf_icmp_net icmp; ++ struct nf_icmp_net icmpv6; ++ struct nf_dccp_net dccp; ++ struct nf_sctp_net sctp; ++ struct nf_gre_net gre; ++}; ++ ++struct ip_conntrack_stat; ++ ++struct nf_ct_event_notifier; ++ ++struct netns_ct { ++ bool ecache_dwork_pending; ++ u8 sysctl_log_invalid; ++ u8 sysctl_events; ++ u8 sysctl_acct; ++ u8 sysctl_tstamp; ++ u8 sysctl_checksum; ++ struct ip_conntrack_stat *stat; ++ struct nf_ct_event_notifier *nf_conntrack_event_cb; ++ struct nf_ip_net nf_ct_proto; ++ unsigned int labels_used; ++}; ++ ++struct netns_nftables { ++ u8 gencursor; ++}; ++ ++struct nf_flow_table_stat; ++ ++struct netns_ft { ++ struct nf_flow_table_stat *stat; ++}; ++ ++struct bpf_prog; ++ ++struct netns_bpf { ++ struct bpf_prog_array *run_array[2]; ++ struct bpf_prog *progs[2]; ++ struct list_head links[2]; ++}; ++ ++struct xfrm_policy_hash { ++ struct hlist_head *table; ++ unsigned int hmask; ++ u8 dbits4; ++ u8 sbits4; ++ u8 dbits6; ++ u8 sbits6; ++}; ++ ++struct xfrm_policy_hthresh { ++ struct work_struct work; ++ seqlock_t lock; ++ u8 lbits4; ++ u8 rbits4; ++ u8 lbits6; ++ u8 rbits6; ++}; ++ ++struct netns_xfrm { ++ struct list_head state_all; ++ struct hlist_head *state_bydst; ++ struct hlist_head *state_bysrc; ++ struct hlist_head *state_byspi; ++ struct hlist_head *state_byseq; ++ unsigned int state_hmask; ++ unsigned int state_num; ++ struct work_struct state_hash_work; ++ struct list_head policy_all; ++ struct hlist_head *policy_byidx; ++ unsigned int policy_idx_hmask; ++ struct hlist_head policy_inexact[3]; ++ struct xfrm_policy_hash policy_bydst[3]; ++ unsigned int policy_count[6]; ++ struct work_struct policy_hash_work; ++ struct xfrm_policy_hthresh policy_hthresh; ++ struct list_head inexact_bins; ++ struct sock *nlsk; ++ struct sock *nlsk_stash; ++ u32 sysctl_aevent_etime; ++ u32 sysctl_aevent_rseqth; ++ int sysctl_larval_drop; ++ u32 sysctl_acq_expires; ++ u8 policy_default[3]; ++ struct ctl_table_header *sysctl_hdr; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct dst_ops xfrm4_dst_ops; ++ struct dst_ops xfrm6_dst_ops; ++ spinlock_t xfrm_state_lock; ++ seqcount_spinlock_t xfrm_state_hash_generation; ++ seqcount_spinlock_t xfrm_policy_hash_generation; ++ spinlock_t xfrm_policy_lock; ++ struct mutex xfrm_cfg_mutex; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct netns_ipvs; ++ ++struct mpls_route; ++ ++struct netns_mpls { ++ int ip_ttl_propagate; ++ int default_ttl; ++ size_t platform_labels; ++ struct mpls_route **platform_label; ++ struct ctl_table_header *ctl; ++}; ++ ++struct netns_xdp { ++ struct mutex lock; ++ struct hlist_head list; ++}; ++ ++struct netns_mctp { ++ struct list_head routes; ++ struct mutex bind_lock; ++ struct hlist_head binds; ++ spinlock_t keys_lock; ++ struct hlist_head keys; ++ unsigned int default_net; ++ struct mutex neigh_lock; ++ struct list_head neighbours; ++}; ++ ++struct smc_stats; ++ ++struct smc_stats_rsn; ++ ++struct netns_smc { ++ struct smc_stats *smc_stats; ++ struct mutex mutex_fback_rsn; ++ struct smc_stats_rsn *fback_rsn; ++ bool limit_smc_hs; ++ struct ctl_table_header *smc_hdr; ++ unsigned int sysctl_autocorking_size; ++ unsigned int sysctl_smcr_buf_type; ++ int sysctl_smcr_testlink_time; ++ int sysctl_wmem; ++ int sysctl_rmem; ++}; ++ ++struct uevent_sock; ++ ++struct net_generic; ++ ++struct net { ++ refcount_t passive; ++ spinlock_t rules_mod_lock; ++ atomic_t dev_unreg_count; ++ unsigned int dev_base_seq; ++ int ifindex; ++ spinlock_t nsid_lock; ++ atomic_t fnhe_genid; ++ struct list_head list; ++ struct list_head exit_list; ++ struct llist_node cleanup_list; ++ struct key_tag *key_domain; ++ struct user_namespace *user_ns; ++ struct ucounts *ucounts; ++ struct idr netns_ids; ++ struct ns_common ns; ++ struct ref_tracker_dir refcnt_tracker; ++ struct ref_tracker_dir notrefcnt_tracker; ++ struct list_head dev_base_head; ++ struct proc_dir_entry *proc_net; ++ struct proc_dir_entry *proc_net_stat; ++ struct ctl_table_set sysctls; ++ struct sock *rtnl; ++ struct sock *genl_sock; ++ struct uevent_sock *uevent_sock; ++ struct hlist_head *dev_name_head; ++ struct hlist_head *dev_index_head; ++ struct raw_notifier_head netdev_chain; ++ u32 hash_mix; ++ struct net_device *loopback_dev; ++ struct list_head rules_ops; ++ struct netns_core core; ++ struct netns_mib mib; ++ struct netns_packet packet; ++ struct netns_unix unx; ++ struct netns_nexthop nexthop; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct netns_ipv4 ipv4; ++ struct netns_ipv6 ipv6; ++ struct netns_ieee802154_lowpan ieee802154_lowpan; ++ struct netns_sctp sctp; ++ struct netns_nf nf; ++ struct netns_ct ct; ++ struct netns_nftables nft; ++ struct netns_ft ft; ++ struct net_generic *gen; ++ struct netns_bpf bpf; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct netns_xfrm xfrm; ++ u64 net_cookie; ++ struct netns_ipvs *ipvs; ++ struct netns_mpls mpls; ++ struct netns_xdp xdp; ++ struct netns_mctp mctp; ++ struct sock *crypto_nlsk; ++ struct sock *diag_nlsk; ++ struct netns_smc smc; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct cgroup_namespace { ++ struct ns_common ns; ++ struct user_namespace *user_ns; ++ struct ucounts *ucounts; ++ struct css_set *root_cset; ++}; ++ ++struct u64_stats_sync {}; ++ ++typedef struct { ++ local_t a; ++} local64_t; ++ ++typedef struct { ++ local64_t v; ++} u64_stats_t; ++ ++struct psi_group_cpu { ++ seqcount_t seq; ++ unsigned int tasks[4]; ++ u32 state_mask; ++ u32 times[7]; ++ u64 state_start; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ u32 times_prev[14]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct psi_group { ++ struct psi_group *parent; ++ bool enabled; ++ struct mutex avgs_lock; ++ struct psi_group_cpu *pcpu; ++ u64 avg_total[6]; ++ u64 avg_last_update; ++ u64 avg_next_update; ++ struct delayed_work avgs_work; ++ struct list_head avg_triggers; ++ u32 avg_nr_triggers[6]; ++ u64 total[12]; ++ long unsigned int avg[18]; ++ struct task_struct *rtpoll_task; ++ struct timer_list rtpoll_timer; ++ wait_queue_head_t rtpoll_wait; ++ atomic_t rtpoll_wakeup; ++ atomic_t rtpoll_scheduled; ++ struct mutex rtpoll_trigger_lock; ++ struct list_head rtpoll_triggers; ++ u32 rtpoll_nr_triggers[6]; ++ u32 rtpoll_states; ++ u64 rtpoll_min_period; ++ u64 rtpoll_total[6]; ++ u64 rtpoll_next_update; ++ u64 rtpoll_until; ++}; ++ ++struct cgroup_taskset; ++ ++struct cftype; ++ ++struct cgroup_subsys { ++ struct cgroup_subsys_state * (*css_alloc)(struct cgroup_subsys_state *); ++ int (*css_online)(struct cgroup_subsys_state *); ++ void (*css_offline)(struct cgroup_subsys_state *); ++ void (*css_released)(struct cgroup_subsys_state *); ++ void (*css_free)(struct cgroup_subsys_state *); ++ void (*css_reset)(struct cgroup_subsys_state *); ++ void (*css_rstat_flush)(struct cgroup_subsys_state *, int); ++ int (*css_extra_stat_show)(struct seq_file *, struct cgroup_subsys_state *); ++ int (*can_attach)(struct cgroup_taskset *); ++ void (*cancel_attach)(struct cgroup_taskset *); ++ void (*attach)(struct cgroup_taskset *); ++ void (*post_attach)(); ++ int (*can_fork)(struct task_struct *, struct css_set *); ++ void (*cancel_fork)(struct task_struct *, struct css_set *); ++ void (*fork)(struct task_struct *); ++ void (*exit)(struct task_struct *); ++ void (*release)(struct task_struct *); ++ void (*bind)(struct cgroup_subsys_state *); ++ bool early_init: 1; ++ bool implicit_on_dfl: 1; ++ bool threaded: 1; ++ int id; ++ const char *name; ++ const char *legacy_name; ++ struct cgroup_root *root; ++ struct idr css_idr; ++ struct list_head cfts; ++ struct cftype *dfl_cftypes; ++ struct cftype *legacy_cftypes; ++ unsigned int depends_on; ++}; ++ ++struct cgroup_rstat_cpu { ++ struct u64_stats_sync bsync; ++ struct cgroup_base_stat bstat; ++ struct cgroup_base_stat last_bstat; ++ struct cgroup *updated_children; ++ struct cgroup *updated_next; ++}; ++ ++struct cgroup_root { ++ struct kernfs_root *kf_root; ++ unsigned int subsys_mask; ++ int hierarchy_id; ++ struct cgroup cgrp; ++ struct cgroup *cgrp_ancestor_storage; ++ atomic_t nr_cgrps; ++ struct list_head root_list; ++ unsigned int flags; ++ char release_agent_path[4096]; ++ char name[64]; ++}; ++ ++struct cftype { ++ char name[64]; ++ long unsigned int private; ++ size_t max_write_len; ++ unsigned int flags; ++ unsigned int file_offset; ++ struct cgroup_subsys *ss; ++ struct list_head node; ++ struct kernfs_ops *kf_ops; ++ int (*open)(struct kernfs_open_file *); ++ void (*release)(struct kernfs_open_file *); ++ u64 (*read_u64)(struct cgroup_subsys_state *, struct cftype *); ++ s64 (*read_s64)(struct cgroup_subsys_state *, struct cftype *); ++ int (*seq_show)(struct seq_file *, void *); ++ void * (*seq_start)(struct seq_file *, loff_t *); ++ void * (*seq_next)(struct seq_file *, void *, loff_t *); ++ void (*seq_stop)(struct seq_file *, void *); ++ int (*write_u64)(struct cgroup_subsys_state *, struct cftype *, u64); ++ int (*write_s64)(struct cgroup_subsys_state *, struct cftype *, s64); ++ ssize_t (*write)(struct kernfs_open_file *, char *, size_t, loff_t); ++ __poll_t (*poll)(struct kernfs_open_file *, struct poll_table_struct *); ++}; ++ ++struct partition_meta_info { ++ char uuid[37]; ++ u8 volname[64]; ++}; ++ ++struct bio_integrity_payload { ++ struct bio *bip_bio; ++ struct bvec_iter bip_iter; ++ short unsigned int bip_vcnt; ++ short unsigned int bip_max_vcnt; ++ short unsigned int bip_flags; ++ struct bvec_iter bio_iter; ++ struct work_struct bip_work; ++ struct bio_vec *bip_vec; ++ struct bio_vec bip_inline_vecs[0]; ++}; ++ ++struct mem_cgroup_reclaim_iter { ++ struct mem_cgroup *position; ++ unsigned int generation; ++}; ++ ++struct shrinker_info { ++ struct callback_head rcu; ++ atomic_long_t *nr_deferred; ++ long unsigned int *map; ++ int map_nr_max; ++}; ++ ++struct lruvec_stats_percpu { ++ long int state[43]; ++ long int state_prev[43]; ++}; ++ ++struct lruvec_stats { ++ long int state[43]; ++ long int state_pending[43]; ++}; ++ ++struct mem_cgroup_per_node { ++ struct lruvec lruvec; ++ struct lruvec_stats_percpu *lruvec_stats_percpu; ++ struct lruvec_stats lruvec_stats; ++ long unsigned int lru_zone_size[15]; ++ struct mem_cgroup_reclaim_iter iter; ++ struct shrinker_info *shrinker_info; ++ struct rb_node tree_node; ++ long unsigned int usage_in_excess; ++ bool on_tree; ++ struct mem_cgroup *memcg; ++}; ++ ++struct eventfd_ctx; ++ ++struct mem_cgroup_threshold { ++ struct eventfd_ctx *eventfd; ++ long unsigned int threshold; ++}; ++ ++struct mem_cgroup_threshold_ary { ++ int current_threshold; ++ unsigned int size; ++ struct mem_cgroup_threshold entries[0]; ++}; ++ ++struct obj_cgroup { ++ struct percpu_ref refcnt; ++ struct mem_cgroup *memcg; ++ atomic_t nr_charged_bytes; ++ union { ++ struct list_head list; ++ struct callback_head rcu; ++ }; ++}; ++ ++struct percpu_cluster { ++ struct swap_cluster_info index; ++ unsigned int next; ++}; ++ ++enum { ++ Root_NFS = 255, ++ Root_CIFS = 254, ++ Root_RAM0 = 1048576, ++ Root_RAM1 = 1048577, ++ Root_FD0 = 2097152, ++ Root_HDA1 = 3145729, ++ Root_HDA2 = 3145730, ++ Root_SDA1 = 8388609, ++ Root_SDA2 = 8388610, ++ Root_HDC1 = 23068673, ++ Root_SR0 = 11534336, ++}; ++ ++typedef struct { ++ union { ++ void *kernel; ++ void *user; ++ }; ++ bool is_kernel: 1; ++} sockptr_t; ++ ++enum fs_value_type { ++ fs_value_is_undefined = 0, ++ fs_value_is_flag = 1, ++ fs_value_is_string = 2, ++ fs_value_is_blob = 3, ++ fs_value_is_filename = 4, ++ fs_value_is_file = 5, ++}; ++ ++struct fs_parameter { ++ const char *key; ++ enum fs_value_type type: 8; ++ union { ++ char *string; ++ void *blob; ++ struct filename *name; ++ struct file *file; ++ }; ++ size_t size; ++ int dirfd; ++}; ++ ++struct fc_log { ++ refcount_t usage; ++ u8 head; ++ u8 tail; ++ u8 need_free; ++ struct module *owner; ++ char *buffer[8]; ++}; ++ ++struct fs_context_operations { ++ void (*free)(struct fs_context *); ++ int (*dup)(struct fs_context *, struct fs_context *); ++ int (*parse_param)(struct fs_context *, struct fs_parameter *); ++ int (*parse_monolithic)(struct fs_context *, void *); ++ int (*get_tree)(struct fs_context *); ++ int (*reconfigure)(struct fs_context *); ++}; ++ ++struct fs_parse_result { ++ bool negated; ++ union { ++ bool boolean; ++ int int_32; ++ unsigned int uint_32; ++ u64 uint_64; ++ }; ++}; ++ ++enum { ++ IPPROTO_IP = 0, ++ IPPROTO_ICMP = 1, ++ IPPROTO_IGMP = 2, ++ IPPROTO_IPIP = 4, ++ IPPROTO_TCP = 6, ++ IPPROTO_EGP = 8, ++ IPPROTO_PUP = 12, ++ IPPROTO_UDP = 17, ++ IPPROTO_IDP = 22, ++ IPPROTO_TP = 29, ++ IPPROTO_DCCP = 33, ++ IPPROTO_IPV6 = 41, ++ IPPROTO_RSVP = 46, ++ IPPROTO_GRE = 47, ++ IPPROTO_ESP = 50, ++ IPPROTO_AH = 51, ++ IPPROTO_MTP = 92, ++ IPPROTO_BEETPH = 94, ++ IPPROTO_ENCAP = 98, ++ IPPROTO_PIM = 103, ++ IPPROTO_COMP = 108, ++ IPPROTO_L2TP = 115, ++ IPPROTO_SCTP = 132, ++ IPPROTO_UDPLITE = 136, ++ IPPROTO_MPLS = 137, ++ IPPROTO_ETHERNET = 143, ++ IPPROTO_RAW = 255, ++ IPPROTO_MPTCP = 262, ++ IPPROTO_MAX = 263, ++}; ++ ++struct flowi_tunnel { ++ __be64 tun_id; ++}; ++ ++struct flowi_common { ++ int flowic_oif; ++ int flowic_iif; ++ int flowic_l3mdev; ++ __u32 flowic_mark; ++ __u8 flowic_tos; ++ __u8 flowic_scope; ++ __u8 flowic_proto; ++ __u8 flowic_flags; ++ __u32 flowic_secid; ++ kuid_t flowic_uid; ++ struct flowi_tunnel flowic_tun_key; ++ __u32 flowic_multipath_hash; ++}; ++ ++union flowi_uli { ++ struct { ++ __be16 dport; ++ __be16 sport; ++ } ports; ++ struct { ++ __u8 type; ++ __u8 code; ++ } icmpt; ++ __be32 gre_key; ++ struct { ++ __u8 type; ++ } mht; ++}; ++ ++struct flowi4 { ++ struct flowi_common __fl_common; ++ __be32 saddr; ++ __be32 daddr; ++ union flowi_uli uli; ++}; ++ ++struct flowi6 { ++ struct flowi_common __fl_common; ++ struct in6_addr daddr; ++ struct in6_addr saddr; ++ __be32 flowlabel; ++ union flowi_uli uli; ++ __u32 mp_hash; ++}; ++ ++struct flowi { ++ union { ++ struct flowi_common __fl_common; ++ struct flowi4 ip4; ++ struct flowi6 ip6; ++ } u; ++}; ++ ++struct prot_inuse { ++ int all; ++ int val[64]; ++}; ++ ++enum { ++ IPSTATS_MIB_NUM = 0, ++ IPSTATS_MIB_INPKTS = 1, ++ IPSTATS_MIB_INOCTETS = 2, ++ IPSTATS_MIB_INDELIVERS = 3, ++ IPSTATS_MIB_OUTFORWDATAGRAMS = 4, ++ IPSTATS_MIB_OUTPKTS = 5, ++ IPSTATS_MIB_OUTOCTETS = 6, ++ IPSTATS_MIB_INHDRERRORS = 7, ++ IPSTATS_MIB_INTOOBIGERRORS = 8, ++ IPSTATS_MIB_INNOROUTES = 9, ++ IPSTATS_MIB_INADDRERRORS = 10, ++ IPSTATS_MIB_INUNKNOWNPROTOS = 11, ++ IPSTATS_MIB_INTRUNCATEDPKTS = 12, ++ IPSTATS_MIB_INDISCARDS = 13, ++ IPSTATS_MIB_OUTDISCARDS = 14, ++ IPSTATS_MIB_OUTNOROUTES = 15, ++ IPSTATS_MIB_REASMTIMEOUT = 16, ++ IPSTATS_MIB_REASMREQDS = 17, ++ IPSTATS_MIB_REASMOKS = 18, ++ IPSTATS_MIB_REASMFAILS = 19, ++ IPSTATS_MIB_FRAGOKS = 20, ++ IPSTATS_MIB_FRAGFAILS = 21, ++ IPSTATS_MIB_FRAGCREATES = 22, ++ IPSTATS_MIB_INMCASTPKTS = 23, ++ IPSTATS_MIB_OUTMCASTPKTS = 24, ++ IPSTATS_MIB_INBCASTPKTS = 25, ++ IPSTATS_MIB_OUTBCASTPKTS = 26, ++ IPSTATS_MIB_INMCASTOCTETS = 27, ++ IPSTATS_MIB_OUTMCASTOCTETS = 28, ++ IPSTATS_MIB_INBCASTOCTETS = 29, ++ IPSTATS_MIB_OUTBCASTOCTETS = 30, ++ IPSTATS_MIB_CSUMERRORS = 31, ++ IPSTATS_MIB_NOECTPKTS = 32, ++ IPSTATS_MIB_ECT1PKTS = 33, ++ IPSTATS_MIB_ECT0PKTS = 34, ++ IPSTATS_MIB_CEPKTS = 35, ++ IPSTATS_MIB_REASM_OVERLAPS = 36, ++ __IPSTATS_MIB_MAX = 37, ++}; ++ ++enum { ++ ICMP_MIB_NUM = 0, ++ ICMP_MIB_INMSGS = 1, ++ ICMP_MIB_INERRORS = 2, ++ ICMP_MIB_INDESTUNREACHS = 3, ++ ICMP_MIB_INTIMEEXCDS = 4, ++ ICMP_MIB_INPARMPROBS = 5, ++ ICMP_MIB_INSRCQUENCHS = 6, ++ ICMP_MIB_INREDIRECTS = 7, ++ ICMP_MIB_INECHOS = 8, ++ ICMP_MIB_INECHOREPS = 9, ++ ICMP_MIB_INTIMESTAMPS = 10, ++ ICMP_MIB_INTIMESTAMPREPS = 11, ++ ICMP_MIB_INADDRMASKS = 12, ++ ICMP_MIB_INADDRMASKREPS = 13, ++ ICMP_MIB_OUTMSGS = 14, ++ ICMP_MIB_OUTERRORS = 15, ++ ICMP_MIB_OUTDESTUNREACHS = 16, ++ ICMP_MIB_OUTTIMEEXCDS = 17, ++ ICMP_MIB_OUTPARMPROBS = 18, ++ ICMP_MIB_OUTSRCQUENCHS = 19, ++ ICMP_MIB_OUTREDIRECTS = 20, ++ ICMP_MIB_OUTECHOS = 21, ++ ICMP_MIB_OUTECHOREPS = 22, ++ ICMP_MIB_OUTTIMESTAMPS = 23, ++ ICMP_MIB_OUTTIMESTAMPREPS = 24, ++ ICMP_MIB_OUTADDRMASKS = 25, ++ ICMP_MIB_OUTADDRMASKREPS = 26, ++ ICMP_MIB_CSUMERRORS = 27, ++ ICMP_MIB_RATELIMITGLOBAL = 28, ++ ICMP_MIB_RATELIMITHOST = 29, ++ __ICMP_MIB_MAX = 30, ++}; ++ ++enum { ++ ICMP6_MIB_NUM = 0, ++ ICMP6_MIB_INMSGS = 1, ++ ICMP6_MIB_INERRORS = 2, ++ ICMP6_MIB_OUTMSGS = 3, ++ ICMP6_MIB_OUTERRORS = 4, ++ ICMP6_MIB_CSUMERRORS = 5, ++ ICMP6_MIB_RATELIMITHOST = 6, ++ __ICMP6_MIB_MAX = 7, ++}; ++ ++enum { ++ TCP_MIB_NUM = 0, ++ TCP_MIB_RTOALGORITHM = 1, ++ TCP_MIB_RTOMIN = 2, ++ TCP_MIB_RTOMAX = 3, ++ TCP_MIB_MAXCONN = 4, ++ TCP_MIB_ACTIVEOPENS = 5, ++ TCP_MIB_PASSIVEOPENS = 6, ++ TCP_MIB_ATTEMPTFAILS = 7, ++ TCP_MIB_ESTABRESETS = 8, ++ TCP_MIB_CURRESTAB = 9, ++ TCP_MIB_INSEGS = 10, ++ TCP_MIB_OUTSEGS = 11, ++ TCP_MIB_RETRANSSEGS = 12, ++ TCP_MIB_INERRS = 13, ++ TCP_MIB_OUTRSTS = 14, ++ TCP_MIB_CSUMERRORS = 15, ++ __TCP_MIB_MAX = 16, ++}; ++ ++enum { ++ UDP_MIB_NUM = 0, ++ UDP_MIB_INDATAGRAMS = 1, ++ UDP_MIB_NOPORTS = 2, ++ UDP_MIB_INERRORS = 3, ++ UDP_MIB_OUTDATAGRAMS = 4, ++ UDP_MIB_RCVBUFERRORS = 5, ++ UDP_MIB_SNDBUFERRORS = 6, ++ UDP_MIB_CSUMERRORS = 7, ++ UDP_MIB_IGNOREDMULTI = 8, ++ UDP_MIB_MEMERRORS = 9, ++ __UDP_MIB_MAX = 10, ++}; ++ ++enum { ++ LINUX_MIB_NUM = 0, ++ LINUX_MIB_SYNCOOKIESSENT = 1, ++ LINUX_MIB_SYNCOOKIESRECV = 2, ++ LINUX_MIB_SYNCOOKIESFAILED = 3, ++ LINUX_MIB_EMBRYONICRSTS = 4, ++ LINUX_MIB_PRUNECALLED = 5, ++ LINUX_MIB_RCVPRUNED = 6, ++ LINUX_MIB_OFOPRUNED = 7, ++ LINUX_MIB_OUTOFWINDOWICMPS = 8, ++ LINUX_MIB_LOCKDROPPEDICMPS = 9, ++ LINUX_MIB_ARPFILTER = 10, ++ LINUX_MIB_TIMEWAITED = 11, ++ LINUX_MIB_TIMEWAITRECYCLED = 12, ++ LINUX_MIB_TIMEWAITKILLED = 13, ++ LINUX_MIB_PAWSACTIVEREJECTED = 14, ++ LINUX_MIB_PAWSESTABREJECTED = 15, ++ LINUX_MIB_DELAYEDACKS = 16, ++ LINUX_MIB_DELAYEDACKLOCKED = 17, ++ LINUX_MIB_DELAYEDACKLOST = 18, ++ LINUX_MIB_LISTENOVERFLOWS = 19, ++ LINUX_MIB_LISTENDROPS = 20, ++ LINUX_MIB_TCPHPHITS = 21, ++ LINUX_MIB_TCPPUREACKS = 22, ++ LINUX_MIB_TCPHPACKS = 23, ++ LINUX_MIB_TCPRENORECOVERY = 24, ++ LINUX_MIB_TCPSACKRECOVERY = 25, ++ LINUX_MIB_TCPSACKRENEGING = 26, ++ LINUX_MIB_TCPSACKREORDER = 27, ++ LINUX_MIB_TCPRENOREORDER = 28, ++ LINUX_MIB_TCPTSREORDER = 29, ++ LINUX_MIB_TCPFULLUNDO = 30, ++ LINUX_MIB_TCPPARTIALUNDO = 31, ++ LINUX_MIB_TCPDSACKUNDO = 32, ++ LINUX_MIB_TCPLOSSUNDO = 33, ++ LINUX_MIB_TCPLOSTRETRANSMIT = 34, ++ LINUX_MIB_TCPRENOFAILURES = 35, ++ LINUX_MIB_TCPSACKFAILURES = 36, ++ LINUX_MIB_TCPLOSSFAILURES = 37, ++ LINUX_MIB_TCPFASTRETRANS = 38, ++ LINUX_MIB_TCPSLOWSTARTRETRANS = 39, ++ LINUX_MIB_TCPTIMEOUTS = 40, ++ LINUX_MIB_TCPLOSSPROBES = 41, ++ LINUX_MIB_TCPLOSSPROBERECOVERY = 42, ++ LINUX_MIB_TCPRENORECOVERYFAIL = 43, ++ LINUX_MIB_TCPSACKRECOVERYFAIL = 44, ++ LINUX_MIB_TCPRCVCOLLAPSED = 45, ++ LINUX_MIB_TCPDSACKOLDSENT = 46, ++ LINUX_MIB_TCPDSACKOFOSENT = 47, ++ LINUX_MIB_TCPDSACKRECV = 48, ++ LINUX_MIB_TCPDSACKOFORECV = 49, ++ LINUX_MIB_TCPABORTONDATA = 50, ++ LINUX_MIB_TCPABORTONCLOSE = 51, ++ LINUX_MIB_TCPABORTONMEMORY = 52, ++ LINUX_MIB_TCPABORTONTIMEOUT = 53, ++ LINUX_MIB_TCPABORTONLINGER = 54, ++ LINUX_MIB_TCPABORTFAILED = 55, ++ LINUX_MIB_TCPMEMORYPRESSURES = 56, ++ LINUX_MIB_TCPMEMORYPRESSURESCHRONO = 57, ++ LINUX_MIB_TCPSACKDISCARD = 58, ++ LINUX_MIB_TCPDSACKIGNOREDOLD = 59, ++ LINUX_MIB_TCPDSACKIGNOREDNOUNDO = 60, ++ LINUX_MIB_TCPSPURIOUSRTOS = 61, ++ LINUX_MIB_TCPMD5NOTFOUND = 62, ++ LINUX_MIB_TCPMD5UNEXPECTED = 63, ++ LINUX_MIB_TCPMD5FAILURE = 64, ++ LINUX_MIB_SACKSHIFTED = 65, ++ LINUX_MIB_SACKMERGED = 66, ++ LINUX_MIB_SACKSHIFTFALLBACK = 67, ++ LINUX_MIB_TCPBACKLOGDROP = 68, ++ LINUX_MIB_PFMEMALLOCDROP = 69, ++ LINUX_MIB_TCPMINTTLDROP = 70, ++ LINUX_MIB_TCPDEFERACCEPTDROP = 71, ++ LINUX_MIB_IPRPFILTER = 72, ++ LINUX_MIB_TCPTIMEWAITOVERFLOW = 73, ++ LINUX_MIB_TCPREQQFULLDOCOOKIES = 74, ++ LINUX_MIB_TCPREQQFULLDROP = 75, ++ LINUX_MIB_TCPRETRANSFAIL = 76, ++ LINUX_MIB_TCPRCVCOALESCE = 77, ++ LINUX_MIB_TCPBACKLOGCOALESCE = 78, ++ LINUX_MIB_TCPOFOQUEUE = 79, ++ LINUX_MIB_TCPOFODROP = 80, ++ LINUX_MIB_TCPOFOMERGE = 81, ++ LINUX_MIB_TCPCHALLENGEACK = 82, ++ LINUX_MIB_TCPSYNCHALLENGE = 83, ++ LINUX_MIB_TCPFASTOPENACTIVE = 84, ++ LINUX_MIB_TCPFASTOPENACTIVEFAIL = 85, ++ LINUX_MIB_TCPFASTOPENPASSIVE = 86, ++ LINUX_MIB_TCPFASTOPENPASSIVEFAIL = 87, ++ LINUX_MIB_TCPFASTOPENLISTENOVERFLOW = 88, ++ LINUX_MIB_TCPFASTOPENCOOKIEREQD = 89, ++ LINUX_MIB_TCPFASTOPENBLACKHOLE = 90, ++ LINUX_MIB_TCPSPURIOUS_RTX_HOSTQUEUES = 91, ++ LINUX_MIB_BUSYPOLLRXPACKETS = 92, ++ LINUX_MIB_TCPAUTOCORKING = 93, ++ LINUX_MIB_TCPFROMZEROWINDOWADV = 94, ++ LINUX_MIB_TCPTOZEROWINDOWADV = 95, ++ LINUX_MIB_TCPWANTZEROWINDOWADV = 96, ++ LINUX_MIB_TCPSYNRETRANS = 97, ++ LINUX_MIB_TCPORIGDATASENT = 98, ++ LINUX_MIB_TCPHYSTARTTRAINDETECT = 99, ++ LINUX_MIB_TCPHYSTARTTRAINCWND = 100, ++ LINUX_MIB_TCPHYSTARTDELAYDETECT = 101, ++ LINUX_MIB_TCPHYSTARTDELAYCWND = 102, ++ LINUX_MIB_TCPACKSKIPPEDSYNRECV = 103, ++ LINUX_MIB_TCPACKSKIPPEDPAWS = 104, ++ LINUX_MIB_TCPACKSKIPPEDSEQ = 105, ++ LINUX_MIB_TCPACKSKIPPEDFINWAIT2 = 106, ++ LINUX_MIB_TCPACKSKIPPEDTIMEWAIT = 107, ++ LINUX_MIB_TCPACKSKIPPEDCHALLENGE = 108, ++ LINUX_MIB_TCPWINPROBE = 109, ++ LINUX_MIB_TCPKEEPALIVE = 110, ++ LINUX_MIB_TCPMTUPFAIL = 111, ++ LINUX_MIB_TCPMTUPSUCCESS = 112, ++ LINUX_MIB_TCPDELIVERED = 113, ++ LINUX_MIB_TCPDELIVEREDCE = 114, ++ LINUX_MIB_TCPACKCOMPRESSED = 115, ++ LINUX_MIB_TCPZEROWINDOWDROP = 116, ++ LINUX_MIB_TCPRCVQDROP = 117, ++ LINUX_MIB_TCPWQUEUETOOBIG = 118, ++ LINUX_MIB_TCPFASTOPENPASSIVEALTKEY = 119, ++ LINUX_MIB_TCPTIMEOUTREHASH = 120, ++ LINUX_MIB_TCPDUPLICATEDATAREHASH = 121, ++ LINUX_MIB_TCPDSACKRECVSEGS = 122, ++ LINUX_MIB_TCPDSACKIGNOREDDUBIOUS = 123, ++ LINUX_MIB_TCPMIGRATEREQSUCCESS = 124, ++ LINUX_MIB_TCPMIGRATEREQFAILURE = 125, ++ LINUX_MIB_TCPPLBREHASH = 126, ++ __LINUX_MIB_MAX = 127, ++}; ++ ++enum { ++ LINUX_MIB_XFRMNUM = 0, ++ LINUX_MIB_XFRMINERROR = 1, ++ LINUX_MIB_XFRMINBUFFERERROR = 2, ++ LINUX_MIB_XFRMINHDRERROR = 3, ++ LINUX_MIB_XFRMINNOSTATES = 4, ++ LINUX_MIB_XFRMINSTATEPROTOERROR = 5, ++ LINUX_MIB_XFRMINSTATEMODEERROR = 6, ++ LINUX_MIB_XFRMINSTATESEQERROR = 7, ++ LINUX_MIB_XFRMINSTATEEXPIRED = 8, ++ LINUX_MIB_XFRMINSTATEMISMATCH = 9, ++ LINUX_MIB_XFRMINSTATEINVALID = 10, ++ LINUX_MIB_XFRMINTMPLMISMATCH = 11, ++ LINUX_MIB_XFRMINNOPOLS = 12, ++ LINUX_MIB_XFRMINPOLBLOCK = 13, ++ LINUX_MIB_XFRMINPOLERROR = 14, ++ LINUX_MIB_XFRMOUTERROR = 15, ++ LINUX_MIB_XFRMOUTBUNDLEGENERROR = 16, ++ LINUX_MIB_XFRMOUTBUNDLECHECKERROR = 17, ++ LINUX_MIB_XFRMOUTNOSTATES = 18, ++ LINUX_MIB_XFRMOUTSTATEPROTOERROR = 19, ++ LINUX_MIB_XFRMOUTSTATEMODEERROR = 20, ++ LINUX_MIB_XFRMOUTSTATESEQERROR = 21, ++ LINUX_MIB_XFRMOUTSTATEEXPIRED = 22, ++ LINUX_MIB_XFRMOUTPOLBLOCK = 23, ++ LINUX_MIB_XFRMOUTPOLDEAD = 24, ++ LINUX_MIB_XFRMOUTPOLERROR = 25, ++ LINUX_MIB_XFRMFWDHDRERROR = 26, ++ LINUX_MIB_XFRMOUTSTATEINVALID = 27, ++ LINUX_MIB_XFRMACQUIREERROR = 28, ++ __LINUX_MIB_XFRMMAX = 29, ++}; ++ ++enum { ++ LINUX_MIB_TLSNUM = 0, ++ LINUX_MIB_TLSCURRTXSW = 1, ++ LINUX_MIB_TLSCURRRXSW = 2, ++ LINUX_MIB_TLSCURRTXDEVICE = 3, ++ LINUX_MIB_TLSCURRRXDEVICE = 4, ++ LINUX_MIB_TLSTXSW = 5, ++ LINUX_MIB_TLSRXSW = 6, ++ LINUX_MIB_TLSTXDEVICE = 7, ++ LINUX_MIB_TLSRXDEVICE = 8, ++ LINUX_MIB_TLSDECRYPTERROR = 9, ++ LINUX_MIB_TLSRXDEVICERESYNC = 10, ++ LINUX_MIB_TLSDECRYPTRETRY = 11, ++ LINUX_MIB_TLSRXNOPADVIOL = 12, ++ __LINUX_MIB_TLSMAX = 13, ++}; ++ ++struct ipstats_mib { ++ u64 mibs[37]; ++ struct u64_stats_sync syncp; ++}; ++ ++struct icmp_mib { ++ long unsigned int mibs[30]; ++}; ++ ++struct icmpmsg_mib { ++ atomic_long_t mibs[512]; ++}; ++ ++struct icmpv6_mib { ++ long unsigned int mibs[7]; ++}; ++ ++struct icmpv6_mib_device { ++ atomic_long_t mibs[7]; ++}; ++ ++struct icmpv6msg_mib { ++ atomic_long_t mibs[512]; ++}; ++ ++struct icmpv6msg_mib_device { ++ atomic_long_t mibs[512]; ++}; ++ ++struct tcp_mib { ++ long unsigned int mibs[16]; ++}; ++ ++struct udp_mib { ++ long unsigned int mibs[10]; ++}; ++ ++struct linux_mib { ++ long unsigned int mibs[127]; ++}; ++ ++struct linux_xfrm_mib { ++ long unsigned int mibs[29]; ++}; ++ ++struct linux_tls_mib { ++ long unsigned int mibs[13]; ++}; ++ ++struct inet_frags; ++ ++struct fqdir { ++ long int high_thresh; ++ long int low_thresh; ++ int timeout; ++ int max_dist; ++ struct inet_frags *f; ++ struct net *net; ++ bool dead; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct rhashtable rhashtable; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ atomic_long_t mem; ++ struct work_struct destroy_work; ++ struct llist_node free_list; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct inet_frag_queue; ++ ++struct inet_frags { ++ unsigned int qsize; ++ void (*constructor)(struct inet_frag_queue *, const void *); ++ void (*destructor)(struct inet_frag_queue *); ++ void (*frag_expire)(struct timer_list *); ++ struct kmem_cache *frags_cachep; ++ const char *frags_cache_name; ++ struct rhashtable_params rhash_params; ++ refcount_t refcnt; ++ struct completion completion; ++}; ++ ++struct frag_v4_compare_key { ++ __be32 saddr; ++ __be32 daddr; ++ u32 user; ++ u32 vif; ++ __be16 id; ++ u16 protocol; ++}; ++ ++struct frag_v6_compare_key { ++ struct in6_addr saddr; ++ struct in6_addr daddr; ++ u32 user; ++ __be32 id; ++ u32 iif; ++}; ++ ++struct inet_frag_queue { ++ struct rhash_head node; ++ union { ++ struct frag_v4_compare_key v4; ++ struct frag_v6_compare_key v6; ++ } key; ++ struct timer_list timer; ++ spinlock_t lock; ++ refcount_t refcnt; ++ struct rb_root rb_fragments; ++ struct sk_buff *fragments_tail; ++ struct sk_buff *last_run_head; ++ ktime_t stamp; ++ int len; ++ int meat; ++ u8 mono_delivery_time; ++ __u8 flags; ++ u16 max_size; ++ struct fqdir *fqdir; ++ struct callback_head rcu; ++}; ++ ++struct fib_rule; ++ ++struct fib_lookup_arg; ++ ++struct fib_rule_hdr; ++ ++struct nlattr; ++ ++struct netlink_ext_ack; ++ ++struct fib_rules_ops { ++ int family; ++ struct list_head list; ++ int rule_size; ++ int addr_size; ++ int unresolved_rules; ++ int nr_goto_rules; ++ unsigned int fib_rules_seq; ++ int (*action)(struct fib_rule *, struct flowi *, int, struct fib_lookup_arg *); ++ bool (*suppress)(struct fib_rule *, int, struct fib_lookup_arg *); ++ int (*match)(struct fib_rule *, struct flowi *, int); ++ int (*configure)(struct fib_rule *, struct sk_buff *, struct fib_rule_hdr *, struct nlattr **, struct netlink_ext_ack *); ++ int (*delete)(struct fib_rule *); ++ int (*compare)(struct fib_rule *, struct fib_rule_hdr *, struct nlattr **); ++ int (*fill)(struct fib_rule *, struct sk_buff *, struct fib_rule_hdr *); ++ size_t (*nlmsg_payload)(struct fib_rule *); ++ void (*flush_cache)(struct fib_rules_ops *); ++ int nlgroup; ++ struct list_head rules_list; ++ struct module *owner; ++ struct net *fro_net; ++ struct callback_head rcu; ++}; ++ ++enum tcp_ca_event { ++ CA_EVENT_TX_START = 0, ++ CA_EVENT_CWND_RESTART = 1, ++ CA_EVENT_COMPLETE_CWR = 2, ++ CA_EVENT_LOSS = 3, ++ CA_EVENT_ECN_NO_CE = 4, ++ CA_EVENT_ECN_IS_CE = 5, ++}; ++ ++struct ack_sample; ++ ++struct rate_sample; ++ ++union tcp_cc_info; ++ ++struct tcp_congestion_ops { ++ u32 (*ssthresh)(struct sock *); ++ void (*cong_avoid)(struct sock *, u32, u32); ++ void (*set_state)(struct sock *, u8); ++ void (*cwnd_event)(struct sock *, enum tcp_ca_event); ++ void (*in_ack_event)(struct sock *, u32); ++ void (*pkts_acked)(struct sock *, const struct ack_sample *); ++ u32 (*min_tso_segs)(struct sock *); ++ void (*cong_control)(struct sock *, const struct rate_sample *); ++ u32 (*undo_cwnd)(struct sock *); ++ u32 (*sndbuf_expand)(struct sock *); ++ size_t (*get_info)(struct sock *, u32, int *, union tcp_cc_info *); ++ char name[16]; ++ struct module *owner; ++ struct list_head list; ++ u32 key; ++ u32 flags; ++ void (*init)(struct sock *); ++ void (*release)(struct sock *); ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct fib_notifier_ops { ++ int family; ++ struct list_head list; ++ unsigned int (*fib_seq_read)(struct net *); ++ int (*fib_dump)(struct net *, struct notifier_block *, struct netlink_ext_ack *); ++ struct module *owner; ++ struct callback_head rcu; ++}; ++ ++typedef struct {} netdevice_tracker; ++ ++struct xfrm_state; ++ ++struct uncached_list; ++ ++struct lwtunnel_state; ++ ++struct dst_entry { ++ struct net_device *dev; ++ struct dst_ops *ops; ++ long unsigned int _metrics; ++ long unsigned int expires; ++ struct xfrm_state *xfrm; ++ int (*input)(struct sk_buff *); ++ int (*output)(struct net *, struct sock *, struct sk_buff *); ++ short unsigned int flags; ++ short int obsolete; ++ short unsigned int header_len; ++ short unsigned int trailer_len; ++ rcuref_t __rcuref; ++ int __use; ++ long unsigned int lastuse; ++ struct callback_head callback_head; ++ short int error; ++ short int __pad; ++ __u32 tclassid; ++ netdevice_tracker dev_tracker; ++ struct list_head rt_uncached; ++ struct uncached_list *rt_uncached_list; ++ struct lwtunnel_state *lwtstate; ++}; ++ ++typedef u32 xdp_features_t; ++ ++struct net_device_stats { ++ union { ++ long unsigned int rx_packets; ++ atomic_long_t __rx_packets; ++ }; ++ union { ++ long unsigned int tx_packets; ++ atomic_long_t __tx_packets; ++ }; ++ union { ++ long unsigned int rx_bytes; ++ atomic_long_t __rx_bytes; ++ }; ++ union { ++ long unsigned int tx_bytes; ++ atomic_long_t __tx_bytes; ++ }; ++ union { ++ long unsigned int rx_errors; ++ atomic_long_t __rx_errors; ++ }; ++ union { ++ long unsigned int tx_errors; ++ atomic_long_t __tx_errors; ++ }; ++ union { ++ long unsigned int rx_dropped; ++ atomic_long_t __rx_dropped; ++ }; ++ union { ++ long unsigned int tx_dropped; ++ atomic_long_t __tx_dropped; ++ }; ++ union { ++ long unsigned int multicast; ++ atomic_long_t __multicast; ++ }; ++ union { ++ long unsigned int collisions; ++ atomic_long_t __collisions; ++ }; ++ union { ++ long unsigned int rx_length_errors; ++ atomic_long_t __rx_length_errors; ++ }; ++ union { ++ long unsigned int rx_over_errors; ++ atomic_long_t __rx_over_errors; ++ }; ++ union { ++ long unsigned int rx_crc_errors; ++ atomic_long_t __rx_crc_errors; ++ }; ++ union { ++ long unsigned int rx_frame_errors; ++ atomic_long_t __rx_frame_errors; ++ }; ++ union { ++ long unsigned int rx_fifo_errors; ++ atomic_long_t __rx_fifo_errors; ++ }; ++ union { ++ long unsigned int rx_missed_errors; ++ atomic_long_t __rx_missed_errors; ++ }; ++ union { ++ long unsigned int tx_aborted_errors; ++ atomic_long_t __tx_aborted_errors; ++ }; ++ union { ++ long unsigned int tx_carrier_errors; ++ atomic_long_t __tx_carrier_errors; ++ }; ++ union { ++ long unsigned int tx_fifo_errors; ++ atomic_long_t __tx_fifo_errors; ++ }; ++ union { ++ long unsigned int tx_heartbeat_errors; ++ atomic_long_t __tx_heartbeat_errors; ++ }; ++ union { ++ long unsigned int tx_window_errors; ++ atomic_long_t __tx_window_errors; ++ }; ++ union { ++ long unsigned int rx_compressed; ++ atomic_long_t __rx_compressed; ++ }; ++ union { ++ long unsigned int tx_compressed; ++ atomic_long_t __tx_compressed; ++ }; ++}; ++ ++struct netdev_hw_addr_list { ++ struct list_head list; ++ int count; ++ struct rb_root tree; ++}; ++ ++struct tipc_bearer; ++ ++struct mpls_dev; ++ ++enum rx_handler_result { ++ RX_HANDLER_CONSUMED = 0, ++ RX_HANDLER_ANOTHER = 1, ++ RX_HANDLER_EXACT = 2, ++ RX_HANDLER_PASS = 3, ++}; ++ ++typedef enum rx_handler_result rx_handler_result_t; ++ ++typedef rx_handler_result_t rx_handler_func_t(struct sk_buff **); ++ ++enum netdev_ml_priv_type { ++ ML_PRIV_NONE = 0, ++ ML_PRIV_CAN = 1, ++}; ++ ++struct pcpu_dstats; ++ ++struct garp_port; ++ ++struct mrp_port; ++ ++struct netdev_tc_txq { ++ u16 count; ++ u16 offset; ++}; ++ ++struct udp_tunnel_nic; ++ ++struct bpf_xdp_link; ++ ++struct bpf_xdp_entity { ++ struct bpf_prog *prog; ++ struct bpf_xdp_link *link; ++}; ++ ++struct netdev_name_node; ++ ++struct dev_ifalias; ++ ++struct net_device_ops; ++ ++struct xdp_metadata_ops; ++ ++struct net_device_core_stats; ++ ++struct ethtool_ops; ++ ++struct l3mdev_ops; ++ ++struct ndisc_ops; ++ ++struct xfrmdev_ops; ++ ++struct tlsdev_ops; ++ ++struct header_ops; ++ ++struct in_device; ++ ++struct inet6_dev; ++ ++struct vlan_info; ++ ++struct dsa_port; ++ ++struct wpan_dev; ++ ++struct mctp_dev; ++ ++struct netdev_rx_queue; ++ ++struct mini_Qdisc; ++ ++struct netdev_queue; ++ ++struct cpu_rmap; ++ ++struct Qdisc; ++ ++struct xdp_dev_bulk_queue; ++ ++struct xps_dev_maps; ++ ++struct netpoll_info; ++ ++struct pcpu_lstats; ++ ++struct pcpu_sw_netstats; ++ ++struct dm_hw_stat_delta; ++ ++struct rtnl_link_ops; ++ ++struct dcbnl_rtnl_ops; ++ ++struct netprio_map; ++ ++struct phy_device; ++ ++struct sfp_bus; ++ ++struct macsec_ops; ++ ++struct udp_tunnel_nic_info; ++ ++struct rtnl_hw_stats64; ++ ++struct devlink_port; ++ ++struct net_device { ++ char name[16]; ++ struct netdev_name_node *name_node; ++ struct dev_ifalias *ifalias; ++ long unsigned int mem_end; ++ long unsigned int mem_start; ++ long unsigned int base_addr; ++ long unsigned int state; ++ struct list_head dev_list; ++ struct list_head napi_list; ++ struct list_head unreg_list; ++ struct list_head close_list; ++ struct list_head ptype_all; ++ struct list_head ptype_specific; ++ struct { ++ struct list_head upper; ++ struct list_head lower; ++ } adj_list; ++ unsigned int flags; ++ xdp_features_t xdp_features; ++ long long unsigned int priv_flags; ++ const struct net_device_ops *netdev_ops; ++ const struct xdp_metadata_ops *xdp_metadata_ops; ++ int ifindex; ++ short unsigned int gflags; ++ short unsigned int hard_header_len; ++ unsigned int mtu; ++ short unsigned int needed_headroom; ++ short unsigned int needed_tailroom; ++ netdev_features_t features; ++ netdev_features_t hw_features; ++ netdev_features_t wanted_features; ++ netdev_features_t vlan_features; ++ netdev_features_t hw_enc_features; ++ netdev_features_t mpls_features; ++ netdev_features_t gso_partial_features; ++ unsigned int min_mtu; ++ unsigned int max_mtu; ++ short unsigned int type; ++ unsigned char min_header_len; ++ unsigned char name_assign_type; ++ int group; ++ struct net_device_stats stats; ++ struct net_device_core_stats *core_stats; ++ atomic_t carrier_up_count; ++ atomic_t carrier_down_count; ++ const struct ethtool_ops *ethtool_ops; ++ const struct l3mdev_ops *l3mdev_ops; ++ const struct ndisc_ops *ndisc_ops; ++ const struct xfrmdev_ops *xfrmdev_ops; ++ const struct tlsdev_ops *tlsdev_ops; ++ const struct header_ops *header_ops; ++ unsigned char operstate; ++ unsigned char link_mode; ++ unsigned char if_port; ++ unsigned char dma; ++ unsigned char perm_addr[32]; ++ unsigned char addr_assign_type; ++ unsigned char addr_len; ++ unsigned char upper_level; ++ unsigned char lower_level; ++ short unsigned int neigh_priv_len; ++ short unsigned int dev_id; ++ short unsigned int dev_port; ++ short unsigned int padded; ++ spinlock_t addr_list_lock; ++ int irq; ++ struct netdev_hw_addr_list uc; ++ struct netdev_hw_addr_list mc; ++ struct netdev_hw_addr_list dev_addrs; ++ struct kset *queues_kset; ++ unsigned int promiscuity; ++ unsigned int allmulti; ++ bool uc_promisc; ++ struct in_device *ip_ptr; ++ struct inet6_dev *ip6_ptr; ++ struct vlan_info *vlan_info; ++ struct dsa_port *dsa_ptr; ++ struct tipc_bearer *tipc_ptr; ++ void *atalk_ptr; ++ struct wpan_dev *ieee802154_ptr; ++ struct mpls_dev *mpls_ptr; ++ struct mctp_dev *mctp_ptr; ++ const unsigned char *dev_addr; ++ struct netdev_rx_queue *_rx; ++ unsigned int num_rx_queues; ++ unsigned int real_num_rx_queues; ++ struct bpf_prog *xdp_prog; ++ long unsigned int gro_flush_timeout; ++ int napi_defer_hard_irqs; ++ unsigned int gro_max_size; ++ unsigned int gro_ipv4_max_size; ++ rx_handler_func_t *rx_handler; ++ void *rx_handler_data; ++ struct mini_Qdisc *miniq_ingress; ++ struct netdev_queue *ingress_queue; ++ struct nf_hook_entries *nf_hooks_ingress; ++ unsigned char broadcast[32]; ++ struct cpu_rmap *rx_cpu_rmap; ++ struct hlist_node index_hlist; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct netdev_queue *_tx; ++ unsigned int num_tx_queues; ++ unsigned int real_num_tx_queues; ++ struct Qdisc *qdisc; ++ unsigned int tx_queue_len; ++ spinlock_t tx_global_lock; ++ struct xdp_dev_bulk_queue *xdp_bulkq; ++ struct xps_dev_maps *xps_maps[2]; ++ struct mini_Qdisc *miniq_egress; ++ struct nf_hook_entries *nf_hooks_egress; ++ struct hlist_head qdisc_hash[16]; ++ struct timer_list watchdog_timer; ++ int watchdog_timeo; ++ u32 proto_down_reason; ++ struct list_head todo_list; ++ int *pcpu_refcnt; ++ struct ref_tracker_dir refcnt_tracker; ++ struct list_head link_watch_list; ++ enum { ++ NETREG_UNINITIALIZED = 0, ++ NETREG_REGISTERED = 1, ++ NETREG_UNREGISTERING = 2, ++ NETREG_UNREGISTERED = 3, ++ NETREG_RELEASED = 4, ++ NETREG_DUMMY = 5, ++ } reg_state: 8; ++ bool dismantle; ++ enum { ++ RTNL_LINK_INITIALIZED = 0, ++ RTNL_LINK_INITIALIZING = 1, ++ } rtnl_link_state: 16; ++ bool needs_free_netdev; ++ void (*priv_destructor)(struct net_device *); ++ struct netpoll_info *npinfo; ++ possible_net_t nd_net; ++ void *ml_priv; ++ enum netdev_ml_priv_type ml_priv_type; ++ union { ++ struct pcpu_lstats *lstats; ++ struct pcpu_sw_netstats *tstats; ++ struct pcpu_dstats *dstats; ++ }; ++ struct garp_port *garp_port; ++ struct mrp_port *mrp_port; ++ struct dm_hw_stat_delta *dm_private; ++ struct device dev; ++ const struct attribute_group *sysfs_groups[4]; ++ const struct attribute_group *sysfs_rx_queue_group; ++ const struct rtnl_link_ops *rtnl_link_ops; ++ unsigned int gso_max_size; ++ unsigned int tso_max_size; ++ u16 gso_max_segs; ++ u16 tso_max_segs; ++ unsigned int gso_ipv4_max_size; ++ const struct dcbnl_rtnl_ops *dcbnl_ops; ++ s16 num_tc; ++ struct netdev_tc_txq tc_to_txq[16]; ++ u8 prio_tc_map[16]; ++ struct netprio_map *priomap; ++ struct phy_device *phydev; ++ struct sfp_bus *sfp_bus; ++ struct lock_class_key *qdisc_tx_busylock; ++ bool proto_down; ++ unsigned int wol_enabled: 1; ++ unsigned int threaded: 1; ++ struct list_head net_notifier_list; ++ const struct macsec_ops *macsec_ops; ++ const struct udp_tunnel_nic_info *udp_tunnel_nic_info; ++ struct udp_tunnel_nic *udp_tunnel_nic; ++ struct bpf_xdp_entity xdp_state[3]; ++ u8 dev_addr_shadow[32]; ++ netdevice_tracker linkwatch_dev_tracker; ++ netdevice_tracker watchdog_dev_tracker; ++ netdevice_tracker dev_registered_tracker; ++ struct rtnl_hw_stats64 *offload_xstats_l3; ++ struct devlink_port *devlink_port; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct hh_cache { ++ unsigned int hh_len; ++ seqlock_t hh_lock; ++ long unsigned int hh_data[4]; ++}; ++ ++struct neigh_table; ++ ++struct neigh_parms; ++ ++struct neigh_ops; ++ ++struct neighbour { ++ struct neighbour *next; ++ struct neigh_table *tbl; ++ struct neigh_parms *parms; ++ long unsigned int confirmed; ++ long unsigned int updated; ++ rwlock_t lock; ++ refcount_t refcnt; ++ unsigned int arp_queue_len_bytes; ++ struct sk_buff_head arp_queue; ++ struct timer_list timer; ++ long unsigned int used; ++ atomic_t probes; ++ u8 nud_state; ++ u8 type; ++ u8 dead; ++ u8 protocol; ++ u32 flags; ++ seqlock_t ha_lock; ++ long: 0; ++ unsigned char ha[32]; ++ struct hh_cache hh; ++ int (*output)(struct neighbour *, struct sk_buff *); ++ const struct neigh_ops *ops; ++ struct list_head gc_list; ++ struct list_head managed_list; ++ struct callback_head rcu; ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ u8 primary_key[0]; ++}; ++ ++struct ipv6_stable_secret { ++ bool initialized; ++ struct in6_addr secret; ++}; ++ ++struct ipv6_devconf { ++ __s32 forwarding; ++ __s32 hop_limit; ++ __s32 mtu6; ++ __s32 accept_ra; ++ __s32 accept_redirects; ++ __s32 autoconf; ++ __s32 dad_transmits; ++ __s32 rtr_solicits; ++ __s32 rtr_solicit_interval; ++ __s32 rtr_solicit_max_interval; ++ __s32 rtr_solicit_delay; ++ __s32 force_mld_version; ++ __s32 mldv1_unsolicited_report_interval; ++ __s32 mldv2_unsolicited_report_interval; ++ __s32 use_tempaddr; ++ __s32 temp_valid_lft; ++ __s32 temp_prefered_lft; ++ __s32 regen_max_retry; ++ __s32 max_desync_factor; ++ __s32 max_addresses; ++ __s32 accept_ra_defrtr; ++ __u32 ra_defrtr_metric; ++ __s32 accept_ra_min_hop_limit; ++ __s32 accept_ra_pinfo; ++ __s32 ignore_routes_with_linkdown; ++ __s32 accept_ra_rtr_pref; ++ __s32 rtr_probe_interval; ++ __s32 accept_ra_rt_info_min_plen; ++ __s32 accept_ra_rt_info_max_plen; ++ __s32 proxy_ndp; ++ __s32 accept_source_route; ++ __s32 accept_ra_from_local; ++ __s32 optimistic_dad; ++ __s32 use_optimistic; ++ atomic_t mc_forwarding; ++ __s32 disable_ipv6; ++ __s32 drop_unicast_in_l2_multicast; ++ __s32 accept_dad; ++ __s32 force_tllao; ++ __s32 ndisc_notify; ++ __s32 suppress_frag_ndisc; ++ __s32 accept_ra_mtu; ++ __s32 drop_unsolicited_na; ++ __s32 accept_untracked_na; ++ struct ipv6_stable_secret stable_secret; ++ __s32 use_oif_addrs_only; ++ __s32 keep_addr_on_down; ++ __s32 seg6_enabled; ++ __s32 seg6_require_hmac; ++ __u32 enhanced_dad; ++ __u32 addr_gen_mode; ++ __s32 disable_policy; ++ __s32 ndisc_tclass; ++ __s32 rpl_seg_enabled; ++ __u32 ioam6_id; ++ __u32 ioam6_id_wide; ++ __u8 ioam6_enabled; ++ __u8 ndisc_evict_nocarrier; ++ struct ctl_table_header *sysctl_header; ++}; ++ ++enum nf_inet_hooks { ++ NF_INET_PRE_ROUTING = 0, ++ NF_INET_LOCAL_IN = 1, ++ NF_INET_FORWARD = 2, ++ NF_INET_LOCAL_OUT = 3, ++ NF_INET_POST_ROUTING = 4, ++ NF_INET_NUMHOOKS = 5, ++ NF_INET_INGRESS = 5, ++}; ++ ++enum { ++ NFPROTO_UNSPEC = 0, ++ NFPROTO_INET = 1, ++ NFPROTO_IPV4 = 2, ++ NFPROTO_ARP = 3, ++ NFPROTO_NETDEV = 5, ++ NFPROTO_BRIDGE = 7, ++ NFPROTO_IPV6 = 10, ++ NFPROTO_NUMPROTO = 11, ++}; ++ ++enum nf_log_type { ++ NF_LOG_TYPE_LOG = 0, ++ NF_LOG_TYPE_ULOG = 1, ++ NF_LOG_TYPE_MAX = 2, ++}; ++ ++typedef u8 u_int8_t; ++ ++struct nf_loginfo; ++ ++typedef void nf_logfn(struct net *, u_int8_t, unsigned int, const struct sk_buff *, const struct net_device *, const struct net_device *, const struct nf_loginfo *, const char *); ++ ++struct nf_logger { ++ char *name; ++ enum nf_log_type type; ++ nf_logfn *logfn; ++ struct module *me; ++}; ++ ++enum tcp_conntrack { ++ TCP_CONNTRACK_NONE = 0, ++ TCP_CONNTRACK_SYN_SENT = 1, ++ TCP_CONNTRACK_SYN_RECV = 2, ++ TCP_CONNTRACK_ESTABLISHED = 3, ++ TCP_CONNTRACK_FIN_WAIT = 4, ++ TCP_CONNTRACK_CLOSE_WAIT = 5, ++ TCP_CONNTRACK_LAST_ACK = 6, ++ TCP_CONNTRACK_TIME_WAIT = 7, ++ TCP_CONNTRACK_CLOSE = 8, ++ TCP_CONNTRACK_LISTEN = 9, ++ TCP_CONNTRACK_MAX = 10, ++ TCP_CONNTRACK_IGNORE = 11, ++ TCP_CONNTRACK_RETRANS = 12, ++ TCP_CONNTRACK_UNACK = 13, ++ TCP_CONNTRACK_TIMEOUT_MAX = 14, ++}; ++ ++enum ct_dccp_states { ++ CT_DCCP_NONE = 0, ++ CT_DCCP_REQUEST = 1, ++ CT_DCCP_RESPOND = 2, ++ CT_DCCP_PARTOPEN = 3, ++ CT_DCCP_OPEN = 4, ++ CT_DCCP_CLOSEREQ = 5, ++ CT_DCCP_CLOSING = 6, ++ CT_DCCP_TIMEWAIT = 7, ++ CT_DCCP_IGNORE = 8, ++ CT_DCCP_INVALID = 9, ++ __CT_DCCP_MAX = 10, ++}; ++ ++struct ip_conntrack_stat { ++ unsigned int found; ++ unsigned int invalid; ++ unsigned int insert; ++ unsigned int insert_failed; ++ unsigned int clash_resolve; ++ unsigned int drop; ++ unsigned int early_drop; ++ unsigned int error; ++ unsigned int expect_new; ++ unsigned int expect_create; ++ unsigned int expect_delete; ++ unsigned int search_restart; ++ unsigned int chaintoolong; ++}; ++ ++enum ip_conntrack_dir { ++ IP_CT_DIR_ORIGINAL = 0, ++ IP_CT_DIR_REPLY = 1, ++ IP_CT_DIR_MAX = 2, ++}; ++ ++enum sctp_conntrack { ++ SCTP_CONNTRACK_NONE = 0, ++ SCTP_CONNTRACK_CLOSED = 1, ++ SCTP_CONNTRACK_COOKIE_WAIT = 2, ++ SCTP_CONNTRACK_COOKIE_ECHOED = 3, ++ SCTP_CONNTRACK_ESTABLISHED = 4, ++ SCTP_CONNTRACK_SHUTDOWN_SENT = 5, ++ SCTP_CONNTRACK_SHUTDOWN_RECD = 6, ++ SCTP_CONNTRACK_SHUTDOWN_ACK_SENT = 7, ++ SCTP_CONNTRACK_HEARTBEAT_SENT = 8, ++ SCTP_CONNTRACK_HEARTBEAT_ACKED = 9, ++ SCTP_CONNTRACK_MAX = 10, ++}; ++ ++enum udp_conntrack { ++ UDP_CT_UNREPLIED = 0, ++ UDP_CT_REPLIED = 1, ++ UDP_CT_MAX = 2, ++}; ++ ++enum gre_conntrack { ++ GRE_CT_UNREPLIED = 0, ++ GRE_CT_REPLIED = 1, ++ GRE_CT_MAX = 2, ++}; ++ ++struct nf_flow_table_stat { ++ unsigned int count_wq_add; ++ unsigned int count_wq_del; ++ unsigned int count_wq_stats; ++}; ++ ++enum { ++ XFRM_POLICY_IN = 0, ++ XFRM_POLICY_OUT = 1, ++ XFRM_POLICY_FWD = 2, ++ XFRM_POLICY_MASK = 3, ++ XFRM_POLICY_MAX = 3, ++}; ++ ++enum netns_bpf_attach_type { ++ NETNS_BPF_INVALID = -1, ++ NETNS_BPF_FLOW_DISSECTOR = 0, ++ NETNS_BPF_SK_LOOKUP = 1, ++ MAX_NETNS_BPF_ATTACH_TYPE = 2, ++}; ++ ++enum flow_dissector_key_id { ++ FLOW_DISSECTOR_KEY_CONTROL = 0, ++ FLOW_DISSECTOR_KEY_BASIC = 1, ++ FLOW_DISSECTOR_KEY_IPV4_ADDRS = 2, ++ FLOW_DISSECTOR_KEY_IPV6_ADDRS = 3, ++ FLOW_DISSECTOR_KEY_PORTS = 4, ++ FLOW_DISSECTOR_KEY_PORTS_RANGE = 5, ++ FLOW_DISSECTOR_KEY_ICMP = 6, ++ FLOW_DISSECTOR_KEY_ETH_ADDRS = 7, ++ FLOW_DISSECTOR_KEY_TIPC = 8, ++ FLOW_DISSECTOR_KEY_ARP = 9, ++ FLOW_DISSECTOR_KEY_VLAN = 10, ++ FLOW_DISSECTOR_KEY_FLOW_LABEL = 11, ++ FLOW_DISSECTOR_KEY_GRE_KEYID = 12, ++ FLOW_DISSECTOR_KEY_MPLS_ENTROPY = 13, ++ FLOW_DISSECTOR_KEY_ENC_KEYID = 14, ++ FLOW_DISSECTOR_KEY_ENC_IPV4_ADDRS = 15, ++ FLOW_DISSECTOR_KEY_ENC_IPV6_ADDRS = 16, ++ FLOW_DISSECTOR_KEY_ENC_CONTROL = 17, ++ FLOW_DISSECTOR_KEY_ENC_PORTS = 18, ++ FLOW_DISSECTOR_KEY_MPLS = 19, ++ FLOW_DISSECTOR_KEY_TCP = 20, ++ FLOW_DISSECTOR_KEY_IP = 21, ++ FLOW_DISSECTOR_KEY_CVLAN = 22, ++ FLOW_DISSECTOR_KEY_ENC_IP = 23, ++ FLOW_DISSECTOR_KEY_ENC_OPTS = 24, ++ FLOW_DISSECTOR_KEY_META = 25, ++ FLOW_DISSECTOR_KEY_CT = 26, ++ FLOW_DISSECTOR_KEY_HASH = 27, ++ FLOW_DISSECTOR_KEY_NUM_OF_VLANS = 28, ++ FLOW_DISSECTOR_KEY_PPPOE = 29, ++ FLOW_DISSECTOR_KEY_L2TPV3 = 30, ++ FLOW_DISSECTOR_KEY_MAX = 31, ++}; ++ ++struct gro_list { ++ struct list_head list; ++ int count; ++}; ++ ++struct napi_struct { ++ struct list_head poll_list; ++ long unsigned int state; ++ int weight; ++ int defer_hard_irqs_count; ++ long unsigned int gro_bitmask; ++ int (*poll)(struct napi_struct *, int); ++ int poll_owner; ++ int list_owner; ++ struct net_device *dev; ++ struct gro_list gro_hash[8]; ++ struct sk_buff *skb; ++ struct list_head rx_list; ++ int rx_count; ++ unsigned int napi_id; ++ struct hrtimer timer; ++ struct task_struct *thread; ++ struct list_head dev_list; ++ struct hlist_node napi_hash_node; ++}; ++ ++struct page_pool_recycle_stats { ++ u64 cached; ++ u64 cache_full; ++ u64 ring; ++ u64 ring_full; ++ u64 released_refcnt; ++}; ++ ++struct skb_shared_hwtstamps { ++ union { ++ ktime_t hwtstamp; ++ void *netdev_data; ++ }; ++}; ++ ++struct skb_ext { ++ refcount_t refcnt; ++ u8 offset[4]; ++ u8 chunks; ++ long: 0; ++ char data[0]; ++}; ++ ++enum skb_ext_id { ++ SKB_EXT_BRIDGE_NF = 0, ++ SKB_EXT_SEC_PATH = 1, ++ TC_SKB_EXT = 2, ++ SKB_EXT_MPTCP = 3, ++ SKB_EXT_NUM = 4, ++}; ++ ++enum rpc_display_format_t { ++ RPC_DISPLAY_ADDR = 0, ++ RPC_DISPLAY_PORT = 1, ++ RPC_DISPLAY_PROTO = 2, ++ RPC_DISPLAY_HEX_ADDR = 3, ++ RPC_DISPLAY_HEX_PORT = 4, ++ RPC_DISPLAY_NETID = 5, ++ RPC_DISPLAY_MAX = 6, ++}; ++ ++struct dql { ++ unsigned int num_queued; ++ unsigned int adj_limit; ++ unsigned int last_obj_cnt; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ unsigned int limit; ++ unsigned int num_completed; ++ unsigned int prev_ovlimit; ++ unsigned int prev_num_queued; ++ unsigned int prev_last_obj_cnt; ++ unsigned int lowest_slack; ++ long unsigned int slack_start_time; ++ unsigned int max_limit; ++ unsigned int min_limit; ++ unsigned int slack_hold_time; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct ieee_ets { ++ __u8 willing; ++ __u8 ets_cap; ++ __u8 cbs; ++ __u8 tc_tx_bw[8]; ++ __u8 tc_rx_bw[8]; ++ __u8 tc_tsa[8]; ++ __u8 prio_tc[8]; ++ __u8 tc_reco_bw[8]; ++ __u8 tc_reco_tsa[8]; ++ __u8 reco_prio_tc[8]; ++}; ++ ++struct ieee_maxrate { ++ __u64 tc_maxrate[8]; ++}; ++ ++struct ieee_qcn { ++ __u8 rpg_enable[8]; ++ __u32 rppp_max_rps[8]; ++ __u32 rpg_time_reset[8]; ++ __u32 rpg_byte_reset[8]; ++ __u32 rpg_threshold[8]; ++ __u32 rpg_max_rate[8]; ++ __u32 rpg_ai_rate[8]; ++ __u32 rpg_hai_rate[8]; ++ __u32 rpg_gd[8]; ++ __u32 rpg_min_dec_fac[8]; ++ __u32 rpg_min_rate[8]; ++ __u32 cndd_state_machine[8]; ++}; ++ ++struct ieee_qcn_stats { ++ __u64 rppp_rp_centiseconds[8]; ++ __u32 rppp_created_rps[8]; ++}; ++ ++struct ieee_pfc { ++ __u8 pfc_cap; ++ __u8 pfc_en; ++ __u8 mbc; ++ __u16 delay; ++ __u64 requests[8]; ++ __u64 indications[8]; ++}; ++ ++struct dcbnl_buffer { ++ __u8 prio2buffer[8]; ++ __u32 buffer_size[8]; ++ __u32 total_size; ++}; ++ ++struct cee_pg { ++ __u8 willing; ++ __u8 error; ++ __u8 pg_en; ++ __u8 tcs_supported; ++ __u8 pg_bw[8]; ++ __u8 prio_pg[8]; ++}; ++ ++struct cee_pfc { ++ __u8 willing; ++ __u8 error; ++ __u8 pfc_en; ++ __u8 tcs_supported; ++}; ++ ++struct dcb_app { ++ __u8 selector; ++ __u8 priority; ++ __u16 protocol; ++}; ++ ++struct dcb_peer_app_info { ++ __u8 willing; ++ __u8 error; ++}; ++ ++struct dcbnl_rtnl_ops { ++ int (*ieee_getets)(struct net_device *, struct ieee_ets *); ++ int (*ieee_setets)(struct net_device *, struct ieee_ets *); ++ int (*ieee_getmaxrate)(struct net_device *, struct ieee_maxrate *); ++ int (*ieee_setmaxrate)(struct net_device *, struct ieee_maxrate *); ++ int (*ieee_getqcn)(struct net_device *, struct ieee_qcn *); ++ int (*ieee_setqcn)(struct net_device *, struct ieee_qcn *); ++ int (*ieee_getqcnstats)(struct net_device *, struct ieee_qcn_stats *); ++ int (*ieee_getpfc)(struct net_device *, struct ieee_pfc *); ++ int (*ieee_setpfc)(struct net_device *, struct ieee_pfc *); ++ int (*ieee_getapp)(struct net_device *, struct dcb_app *); ++ int (*ieee_setapp)(struct net_device *, struct dcb_app *); ++ int (*ieee_delapp)(struct net_device *, struct dcb_app *); ++ int (*ieee_peer_getets)(struct net_device *, struct ieee_ets *); ++ int (*ieee_peer_getpfc)(struct net_device *, struct ieee_pfc *); ++ u8 (*getstate)(struct net_device *); ++ u8 (*setstate)(struct net_device *, u8); ++ void (*getpermhwaddr)(struct net_device *, u8 *); ++ void (*setpgtccfgtx)(struct net_device *, int, u8, u8, u8, u8); ++ void (*setpgbwgcfgtx)(struct net_device *, int, u8); ++ void (*setpgtccfgrx)(struct net_device *, int, u8, u8, u8, u8); ++ void (*setpgbwgcfgrx)(struct net_device *, int, u8); ++ void (*getpgtccfgtx)(struct net_device *, int, u8 *, u8 *, u8 *, u8 *); ++ void (*getpgbwgcfgtx)(struct net_device *, int, u8 *); ++ void (*getpgtccfgrx)(struct net_device *, int, u8 *, u8 *, u8 *, u8 *); ++ void (*getpgbwgcfgrx)(struct net_device *, int, u8 *); ++ void (*setpfccfg)(struct net_device *, int, u8); ++ void (*getpfccfg)(struct net_device *, int, u8 *); ++ u8 (*setall)(struct net_device *); ++ u8 (*getcap)(struct net_device *, int, u8 *); ++ int (*getnumtcs)(struct net_device *, int, u8 *); ++ int (*setnumtcs)(struct net_device *, int, u8); ++ u8 (*getpfcstate)(struct net_device *); ++ void (*setpfcstate)(struct net_device *, u8); ++ void (*getbcncfg)(struct net_device *, int, u32 *); ++ void (*setbcncfg)(struct net_device *, int, u32); ++ void (*getbcnrp)(struct net_device *, int, u8 *); ++ void (*setbcnrp)(struct net_device *, int, u8); ++ int (*setapp)(struct net_device *, u8, u16, u8); ++ int (*getapp)(struct net_device *, u8, u16); ++ u8 (*getfeatcfg)(struct net_device *, int, u8 *); ++ u8 (*setfeatcfg)(struct net_device *, int, u8); ++ u8 (*getdcbx)(struct net_device *); ++ u8 (*setdcbx)(struct net_device *, u8); ++ int (*peer_getappinfo)(struct net_device *, struct dcb_peer_app_info *, u16 *); ++ int (*peer_getapptable)(struct net_device *, struct dcb_app *); ++ int (*cee_peer_getpg)(struct net_device *, struct cee_pg *); ++ int (*cee_peer_getpfc)(struct net_device *, struct cee_pfc *); ++ int (*dcbnl_getbuffer)(struct net_device *, struct dcbnl_buffer *); ++ int (*dcbnl_setbuffer)(struct net_device *, struct dcbnl_buffer *); ++ int (*dcbnl_setapptrust)(struct net_device *, u8 *, int); ++ int (*dcbnl_getapptrust)(struct net_device *, u8 *, int *); ++ int (*dcbnl_setrewr)(struct net_device *, struct dcb_app *); ++ int (*dcbnl_delrewr)(struct net_device *, struct dcb_app *); ++}; ++ ++struct netprio_map { ++ struct callback_head rcu; ++ u32 priomap_len; ++ u32 priomap[0]; ++}; ++ ++struct xdp_mem_info { ++ u32 type; ++ u32 id; ++}; ++ ++struct xdp_rxq_info { ++ struct net_device *dev; ++ u32 queue_index; ++ u32 reg_state; ++ struct xdp_mem_info mem; ++ unsigned int napi_id; ++ u32 frag_size; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct xdp_txq_info { ++ struct net_device *dev; ++}; ++ ++struct xdp_buff { ++ void *data; ++ void *data_end; ++ void *data_meta; ++ void *data_hard_start; ++ struct xdp_rxq_info *rxq; ++ struct xdp_txq_info *txq; ++ u32 frame_sz; ++ u32 flags; ++}; ++ ++struct xdp_frame { ++ void *data; ++ u16 len; ++ u16 headroom; ++ u32 metasize; ++ struct xdp_mem_info mem; ++ struct net_device *dev_rx; ++ u32 frame_sz; ++ u32 flags; ++}; ++ ++enum xdp_rss_hash_type { ++ XDP_RSS_L3_IPV4 = 1, ++ XDP_RSS_L3_IPV6 = 2, ++ XDP_RSS_L3_DYNHDR = 4, ++ XDP_RSS_L4 = 8, ++ XDP_RSS_L4_TCP = 16, ++ XDP_RSS_L4_UDP = 32, ++ XDP_RSS_L4_SCTP = 64, ++ XDP_RSS_L4_IPSEC = 128, ++ XDP_RSS_TYPE_NONE = 0, ++ XDP_RSS_TYPE_L2 = 0, ++ XDP_RSS_TYPE_L3_IPV4 = 1, ++ XDP_RSS_TYPE_L3_IPV6 = 2, ++ XDP_RSS_TYPE_L3_IPV4_OPT = 5, ++ XDP_RSS_TYPE_L3_IPV6_EX = 6, ++ XDP_RSS_TYPE_L4_ANY = 8, ++ XDP_RSS_TYPE_L4_IPV4_TCP = 25, ++ XDP_RSS_TYPE_L4_IPV4_UDP = 41, ++ XDP_RSS_TYPE_L4_IPV4_SCTP = 73, ++ XDP_RSS_TYPE_L4_IPV4_IPSEC = 137, ++ XDP_RSS_TYPE_L4_IPV6_TCP = 26, ++ XDP_RSS_TYPE_L4_IPV6_UDP = 42, ++ XDP_RSS_TYPE_L4_IPV6_SCTP = 74, ++ XDP_RSS_TYPE_L4_IPV6_IPSEC = 138, ++ XDP_RSS_TYPE_L4_IPV6_TCP_EX = 30, ++ XDP_RSS_TYPE_L4_IPV6_UDP_EX = 46, ++ XDP_RSS_TYPE_L4_IPV6_SCTP_EX = 78, ++}; ++ ++typedef enum { ++ SS_FREE = 0, ++ SS_UNCONNECTED = 1, ++ SS_CONNECTING = 2, ++ SS_CONNECTED = 3, ++ SS_DISCONNECTING = 4, ++} socket_state; ++ ++struct socket_wq { ++ wait_queue_head_t wait; ++ struct fasync_struct *fasync_list; ++ long unsigned int flags; ++ struct callback_head rcu; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct proto_ops; ++ ++struct socket { ++ socket_state state; ++ short int type; ++ long unsigned int flags; ++ struct file *file; ++ struct sock *sk; ++ const struct proto_ops *ops; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct socket_wq wq; ++}; ++ ++typedef struct { ++ size_t written; ++ size_t count; ++ union { ++ char *buf; ++ void *data; ++ } arg; ++ int error; ++} read_descriptor_t; ++ ++typedef int (*sk_read_actor_t)(read_descriptor_t *, struct sk_buff *, unsigned int, size_t); ++ ++typedef int (*skb_read_actor_t)(struct sock *, struct sk_buff *); ++ ++struct proto_ops { ++ int family; ++ struct module *owner; ++ int (*release)(struct socket *); ++ int (*bind)(struct socket *, struct sockaddr *, int); ++ int (*connect)(struct socket *, struct sockaddr *, int, int); ++ int (*socketpair)(struct socket *, struct socket *); ++ int (*accept)(struct socket *, struct socket *, int, bool); ++ int (*getname)(struct socket *, struct sockaddr *, int); ++ __poll_t (*poll)(struct file *, struct socket *, struct poll_table_struct *); ++ int (*ioctl)(struct socket *, unsigned int, long unsigned int); ++ int (*compat_ioctl)(struct socket *, unsigned int, long unsigned int); ++ int (*gettstamp)(struct socket *, void *, bool, bool); ++ int (*listen)(struct socket *, int); ++ int (*shutdown)(struct socket *, int); ++ int (*setsockopt)(struct socket *, int, int, sockptr_t, unsigned int); ++ int (*getsockopt)(struct socket *, int, int, char *, int *); ++ void (*show_fdinfo)(struct seq_file *, struct socket *); ++ int (*sendmsg)(struct socket *, struct msghdr *, size_t); ++ int (*recvmsg)(struct socket *, struct msghdr *, size_t, int); ++ int (*mmap)(struct file *, struct socket *, struct vm_area_struct *); ++ ssize_t (*sendpage)(struct socket *, struct page *, int, size_t, int); ++ ssize_t (*splice_read)(struct socket *, loff_t *, struct pipe_inode_info *, size_t, unsigned int); ++ int (*set_peek_off)(struct sock *, int); ++ int (*peek_len)(struct socket *); ++ int (*read_sock)(struct sock *, read_descriptor_t *, sk_read_actor_t); ++ int (*read_skb)(struct sock *, skb_read_actor_t); ++ int (*sendpage_locked)(struct sock *, struct page *, int, size_t, int); ++ int (*sendmsg_locked)(struct sock *, struct msghdr *, size_t); ++ int (*set_rcvlowat)(struct sock *, int); ++}; ++ ++struct nlmsghdr { ++ __u32 nlmsg_len; ++ __u16 nlmsg_type; ++ __u16 nlmsg_flags; ++ __u32 nlmsg_seq; ++ __u32 nlmsg_pid; ++}; ++ ++struct nlattr { ++ __u16 nla_len; ++ __u16 nla_type; ++}; ++ ++struct nla_policy; ++ ++struct netlink_ext_ack { ++ const char *_msg; ++ const struct nlattr *bad_attr; ++ const struct nla_policy *policy; ++ const struct nlattr *miss_nest; ++ u16 miss_type; ++ u8 cookie[20]; ++ u8 cookie_len; ++ char _msg_buf[80]; ++}; ++ ++struct netlink_range_validation; ++ ++struct netlink_range_validation_signed; ++ ++struct nla_policy { ++ u8 type; ++ u8 validation_type; ++ u16 len; ++ union { ++ u16 strict_start_type; ++ const u32 bitfield32_valid; ++ const u32 mask; ++ const char *reject_message; ++ const struct nla_policy *nested_policy; ++ struct netlink_range_validation *range; ++ struct netlink_range_validation_signed *range_signed; ++ struct { ++ s16 min; ++ s16 max; ++ }; ++ int (*validate)(const struct nlattr *, struct netlink_ext_ack *); ++ }; ++}; ++ ++struct netlink_callback { ++ struct sk_buff *skb; ++ const struct nlmsghdr *nlh; ++ int (*dump)(struct sk_buff *, struct netlink_callback *); ++ int (*done)(struct netlink_callback *); ++ void *data; ++ struct module *module; ++ struct netlink_ext_ack *extack; ++ u16 family; ++ u16 answer_flags; ++ u32 min_dump_alloc; ++ unsigned int prev_seq; ++ unsigned int seq; ++ bool strict_check; ++ union { ++ u8 ctx[48]; ++ long int args[6]; ++ }; ++}; ++ ++struct ndmsg { ++ __u8 ndm_family; ++ __u8 ndm_pad1; ++ __u16 ndm_pad2; ++ __s32 ndm_ifindex; ++ __u16 ndm_state; ++ __u8 ndm_flags; ++ __u8 ndm_type; ++}; ++ ++struct rtnl_link_stats64 { ++ __u64 rx_packets; ++ __u64 tx_packets; ++ __u64 rx_bytes; ++ __u64 tx_bytes; ++ __u64 rx_errors; ++ __u64 tx_errors; ++ __u64 rx_dropped; ++ __u64 tx_dropped; ++ __u64 multicast; ++ __u64 collisions; ++ __u64 rx_length_errors; ++ __u64 rx_over_errors; ++ __u64 rx_crc_errors; ++ __u64 rx_frame_errors; ++ __u64 rx_fifo_errors; ++ __u64 rx_missed_errors; ++ __u64 tx_aborted_errors; ++ __u64 tx_carrier_errors; ++ __u64 tx_fifo_errors; ++ __u64 tx_heartbeat_errors; ++ __u64 tx_window_errors; ++ __u64 rx_compressed; ++ __u64 tx_compressed; ++ __u64 rx_nohandler; ++ __u64 rx_otherhost_dropped; ++}; ++ ++struct rtnl_hw_stats64 { ++ __u64 rx_packets; ++ __u64 tx_packets; ++ __u64 rx_bytes; ++ __u64 tx_bytes; ++ __u64 rx_errors; ++ __u64 tx_errors; ++ __u64 rx_dropped; ++ __u64 tx_dropped; ++ __u64 multicast; ++}; ++ ++struct ifla_vf_guid { ++ __u32 vf; ++ __u64 guid; ++}; ++ ++struct ifla_vf_stats { ++ __u64 rx_packets; ++ __u64 tx_packets; ++ __u64 rx_bytes; ++ __u64 tx_bytes; ++ __u64 broadcast; ++ __u64 multicast; ++ __u64 rx_dropped; ++ __u64 tx_dropped; ++}; ++ ++struct ifla_vf_info { ++ __u32 vf; ++ __u8 mac[32]; ++ __u32 vlan; ++ __u32 qos; ++ __u32 spoofchk; ++ __u32 linkstate; ++ __u32 min_tx_rate; ++ __u32 max_tx_rate; ++ __u32 rss_query_en; ++ __u32 trusted; ++ __be16 vlan_proto; ++}; ++ ++enum netdev_tx { ++ __NETDEV_TX_MIN = -2147483648, ++ NETDEV_TX_OK = 0, ++ NETDEV_TX_BUSY = 16, ++}; ++ ++typedef enum netdev_tx netdev_tx_t; ++ ++struct net_device_core_stats { ++ long unsigned int rx_dropped; ++ long unsigned int tx_dropped; ++ long unsigned int rx_nohandler; ++ long unsigned int rx_otherhost_dropped; ++}; ++ ++struct header_ops { ++ int (*create)(struct sk_buff *, struct net_device *, short unsigned int, const void *, const void *, unsigned int); ++ int (*parse)(const struct sk_buff *, unsigned char *); ++ int (*cache)(const struct neighbour *, struct hh_cache *, __be16); ++ void (*cache_update)(struct hh_cache *, const struct net_device *, const unsigned char *); ++ bool (*validate)(const char *, unsigned int); ++ __be16 (*parse_protocol)(const struct sk_buff *); ++}; ++ ++enum { ++ NAPI_STATE_SCHED = 0, ++ NAPI_STATE_MISSED = 1, ++ NAPI_STATE_DISABLE = 2, ++ NAPI_STATE_NPSVC = 3, ++ NAPI_STATE_LISTED = 4, ++ NAPI_STATE_NO_BUSY_POLL = 5, ++ NAPI_STATE_IN_BUSY_POLL = 6, ++ NAPI_STATE_PREFER_BUSY_POLL = 7, ++ NAPI_STATE_THREADED = 8, ++ NAPI_STATE_SCHED_THREADED = 9, ++}; ++ ++struct xsk_buff_pool; ++ ++struct netdev_queue { ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ struct Qdisc *qdisc; ++ struct Qdisc *qdisc_sleeping; ++ struct kobject kobj; ++ int numa_node; ++ long unsigned int tx_maxrate; ++ atomic_long_t trans_timeout; ++ struct net_device *sb_dev; ++ struct xsk_buff_pool *pool; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ spinlock_t _xmit_lock; ++ int xmit_lock_owner; ++ long unsigned int trans_start; ++ long unsigned int state; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct dql dql; ++}; ++ ++struct rps_map { ++ unsigned int len; ++ struct callback_head rcu; ++ u16 cpus[0]; ++}; ++ ++struct rps_dev_flow { ++ u16 cpu; ++ u16 filter; ++ unsigned int last_qtail; ++}; ++ ++struct rps_dev_flow_table { ++ unsigned int mask; ++ struct callback_head rcu; ++ struct rps_dev_flow flows[0]; ++}; ++ ++struct netdev_rx_queue { ++ struct xdp_rxq_info xdp_rxq; ++ struct rps_map *rps_map; ++ struct rps_dev_flow_table *rps_flow_table; ++ struct kobject kobj; ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ struct xsk_buff_pool *pool; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum xps_map_type { ++ XPS_CPUS = 0, ++ XPS_RXQS = 1, ++ XPS_MAPS_MAX = 2, ++}; ++ ++struct xps_map { ++ unsigned int len; ++ unsigned int alloc_len; ++ struct callback_head rcu; ++ u16 queues[0]; ++}; ++ ++struct xps_dev_maps { ++ struct callback_head rcu; ++ unsigned int nr_ids; ++ s16 num_tc; ++ struct xps_map *attr_map[0]; ++}; ++ ++struct netdev_phys_item_id { ++ unsigned char id[32]; ++ unsigned char id_len; ++}; ++ ++enum net_device_path_type { ++ DEV_PATH_ETHERNET = 0, ++ DEV_PATH_VLAN = 1, ++ DEV_PATH_BRIDGE = 2, ++ DEV_PATH_PPPOE = 3, ++ DEV_PATH_DSA = 4, ++ DEV_PATH_MTK_WDMA = 5, ++}; ++ ++struct net_device_path { ++ enum net_device_path_type type; ++ const struct net_device *dev; ++ union { ++ struct { ++ u16 id; ++ __be16 proto; ++ u8 h_dest[6]; ++ } encap; ++ struct { ++ enum { ++ DEV_PATH_BR_VLAN_KEEP = 0, ++ DEV_PATH_BR_VLAN_TAG = 1, ++ DEV_PATH_BR_VLAN_UNTAG = 2, ++ DEV_PATH_BR_VLAN_UNTAG_HW = 3, ++ } vlan_mode; ++ u16 vlan_id; ++ __be16 vlan_proto; ++ } bridge; ++ struct { ++ int port; ++ u16 proto; ++ } dsa; ++ struct { ++ u8 wdma_idx; ++ u8 queue; ++ u16 wcid; ++ u8 bss; ++ } mtk_wdma; ++ }; ++}; ++ ++struct net_device_path_ctx { ++ const struct net_device *dev; ++ u8 daddr[6]; ++ int num_vlans; ++ struct { ++ u16 id; ++ __be16 proto; ++ } vlan[2]; ++}; ++ ++enum tc_setup_type { ++ TC_QUERY_CAPS = 0, ++ TC_SETUP_QDISC_MQPRIO = 1, ++ TC_SETUP_CLSU32 = 2, ++ TC_SETUP_CLSFLOWER = 3, ++ TC_SETUP_CLSMATCHALL = 4, ++ TC_SETUP_CLSBPF = 5, ++ TC_SETUP_BLOCK = 6, ++ TC_SETUP_QDISC_CBS = 7, ++ TC_SETUP_QDISC_RED = 8, ++ TC_SETUP_QDISC_PRIO = 9, ++ TC_SETUP_QDISC_MQ = 10, ++ TC_SETUP_QDISC_ETF = 11, ++ TC_SETUP_ROOT_QDISC = 12, ++ TC_SETUP_QDISC_GRED = 13, ++ TC_SETUP_QDISC_TAPRIO = 14, ++ TC_SETUP_FT = 15, ++ TC_SETUP_QDISC_ETS = 16, ++ TC_SETUP_QDISC_TBF = 17, ++ TC_SETUP_QDISC_FIFO = 18, ++ TC_SETUP_QDISC_HTB = 19, ++ TC_SETUP_ACT = 20, ++}; ++ ++enum bpf_netdev_command { ++ XDP_SETUP_PROG = 0, ++ XDP_SETUP_PROG_HW = 1, ++ BPF_OFFLOAD_MAP_ALLOC = 2, ++ BPF_OFFLOAD_MAP_FREE = 3, ++ XDP_SETUP_XSK_POOL = 4, ++}; ++ ++enum bpf_xdp_mode { ++ XDP_MODE_SKB = 0, ++ XDP_MODE_DRV = 1, ++ XDP_MODE_HW = 2, ++ __MAX_XDP_MODE = 3, ++}; ++ ++struct bpf_offloaded_map; ++ ++struct netdev_bpf { ++ enum bpf_netdev_command command; ++ union { ++ struct { ++ u32 flags; ++ struct bpf_prog *prog; ++ struct netlink_ext_ack *extack; ++ }; ++ struct { ++ struct bpf_offloaded_map *offmap; ++ }; ++ struct { ++ struct xsk_buff_pool *pool; ++ u16 queue_id; ++ } xsk; ++ }; ++}; ++ ++struct xfrmdev_ops { ++ int (*xdo_dev_state_add)(struct xfrm_state *, struct netlink_ext_ack *); ++ void (*xdo_dev_state_delete)(struct xfrm_state *); ++ void (*xdo_dev_state_free)(struct xfrm_state *); ++ bool (*xdo_dev_offload_ok)(struct sk_buff *, struct xfrm_state *); ++ void (*xdo_dev_state_advance_esn)(struct xfrm_state *); ++ void (*xdo_dev_state_update_curlft)(struct xfrm_state *); ++ int (*xdo_dev_policy_add)(struct xfrm_policy *, struct netlink_ext_ack *); ++ void (*xdo_dev_policy_delete)(struct xfrm_policy *); ++ void (*xdo_dev_policy_free)(struct xfrm_policy *); ++}; ++ ++struct dev_ifalias { ++ struct callback_head rcuhead; ++ char ifalias[0]; ++}; ++ ++struct ip_tunnel_parm; ++ ++struct net_device_ops { ++ int (*ndo_init)(struct net_device *); ++ void (*ndo_uninit)(struct net_device *); ++ int (*ndo_open)(struct net_device *); ++ int (*ndo_stop)(struct net_device *); ++ netdev_tx_t (*ndo_start_xmit)(struct sk_buff *, struct net_device *); ++ netdev_features_t (*ndo_features_check)(struct sk_buff *, struct net_device *, netdev_features_t); ++ u16 (*ndo_select_queue)(struct net_device *, struct sk_buff *, struct net_device *); ++ void (*ndo_change_rx_flags)(struct net_device *, int); ++ void (*ndo_set_rx_mode)(struct net_device *); ++ int (*ndo_set_mac_address)(struct net_device *, void *); ++ int (*ndo_validate_addr)(struct net_device *); ++ int (*ndo_do_ioctl)(struct net_device *, struct ifreq *, int); ++ int (*ndo_eth_ioctl)(struct net_device *, struct ifreq *, int); ++ int (*ndo_siocbond)(struct net_device *, struct ifreq *, int); ++ int (*ndo_siocwandev)(struct net_device *, struct if_settings *); ++ int (*ndo_siocdevprivate)(struct net_device *, struct ifreq *, void *, int); ++ int (*ndo_set_config)(struct net_device *, struct ifmap *); ++ int (*ndo_change_mtu)(struct net_device *, int); ++ int (*ndo_neigh_setup)(struct net_device *, struct neigh_parms *); ++ void (*ndo_tx_timeout)(struct net_device *, unsigned int); ++ void (*ndo_get_stats64)(struct net_device *, struct rtnl_link_stats64 *); ++ bool (*ndo_has_offload_stats)(const struct net_device *, int); ++ int (*ndo_get_offload_stats)(int, const struct net_device *, void *); ++ struct net_device_stats * (*ndo_get_stats)(struct net_device *); ++ int (*ndo_vlan_rx_add_vid)(struct net_device *, __be16, u16); ++ int (*ndo_vlan_rx_kill_vid)(struct net_device *, __be16, u16); ++ void (*ndo_poll_controller)(struct net_device *); ++ int (*ndo_netpoll_setup)(struct net_device *, struct netpoll_info *); ++ void (*ndo_netpoll_cleanup)(struct net_device *); ++ int (*ndo_set_vf_mac)(struct net_device *, int, u8 *); ++ int (*ndo_set_vf_vlan)(struct net_device *, int, u16, u8, __be16); ++ int (*ndo_set_vf_rate)(struct net_device *, int, int, int); ++ int (*ndo_set_vf_spoofchk)(struct net_device *, int, bool); ++ int (*ndo_set_vf_trust)(struct net_device *, int, bool); ++ int (*ndo_get_vf_config)(struct net_device *, int, struct ifla_vf_info *); ++ int (*ndo_set_vf_link_state)(struct net_device *, int, int); ++ int (*ndo_get_vf_stats)(struct net_device *, int, struct ifla_vf_stats *); ++ int (*ndo_set_vf_port)(struct net_device *, int, struct nlattr **); ++ int (*ndo_get_vf_port)(struct net_device *, int, struct sk_buff *); ++ int (*ndo_get_vf_guid)(struct net_device *, int, struct ifla_vf_guid *, struct ifla_vf_guid *); ++ int (*ndo_set_vf_guid)(struct net_device *, int, u64, int); ++ int (*ndo_set_vf_rss_query_en)(struct net_device *, int, bool); ++ int (*ndo_setup_tc)(struct net_device *, enum tc_setup_type, void *); ++ int (*ndo_fcoe_get_wwn)(struct net_device *, u64 *, int); ++ int (*ndo_rx_flow_steer)(struct net_device *, const struct sk_buff *, u16, u32); ++ int (*ndo_add_slave)(struct net_device *, struct net_device *, struct netlink_ext_ack *); ++ int (*ndo_del_slave)(struct net_device *, struct net_device *); ++ struct net_device * (*ndo_get_xmit_slave)(struct net_device *, struct sk_buff *, bool); ++ struct net_device * (*ndo_sk_get_lower_dev)(struct net_device *, struct sock *); ++ netdev_features_t (*ndo_fix_features)(struct net_device *, netdev_features_t); ++ int (*ndo_set_features)(struct net_device *, netdev_features_t); ++ int (*ndo_neigh_construct)(struct net_device *, struct neighbour *); ++ void (*ndo_neigh_destroy)(struct net_device *, struct neighbour *); ++ int (*ndo_fdb_add)(struct ndmsg *, struct nlattr **, struct net_device *, const unsigned char *, u16, u16, struct netlink_ext_ack *); ++ int (*ndo_fdb_del)(struct ndmsg *, struct nlattr **, struct net_device *, const unsigned char *, u16, struct netlink_ext_ack *); ++ int (*ndo_fdb_del_bulk)(struct ndmsg *, struct nlattr **, struct net_device *, u16, struct netlink_ext_ack *); ++ int (*ndo_fdb_dump)(struct sk_buff *, struct netlink_callback *, struct net_device *, struct net_device *, int *); ++ int (*ndo_fdb_get)(struct sk_buff *, struct nlattr **, struct net_device *, const unsigned char *, u16, u32, u32, struct netlink_ext_ack *); ++ int (*ndo_mdb_add)(struct net_device *, struct nlattr **, u16, struct netlink_ext_ack *); ++ int (*ndo_mdb_del)(struct net_device *, struct nlattr **, struct netlink_ext_ack *); ++ int (*ndo_mdb_dump)(struct net_device *, struct sk_buff *, struct netlink_callback *); ++ int (*ndo_bridge_setlink)(struct net_device *, struct nlmsghdr *, u16, struct netlink_ext_ack *); ++ int (*ndo_bridge_getlink)(struct sk_buff *, u32, u32, struct net_device *, u32, int); ++ int (*ndo_bridge_dellink)(struct net_device *, struct nlmsghdr *, u16); ++ int (*ndo_change_carrier)(struct net_device *, bool); ++ int (*ndo_get_phys_port_id)(struct net_device *, struct netdev_phys_item_id *); ++ int (*ndo_get_port_parent_id)(struct net_device *, struct netdev_phys_item_id *); ++ int (*ndo_get_phys_port_name)(struct net_device *, char *, size_t); ++ void * (*ndo_dfwd_add_station)(struct net_device *, struct net_device *); ++ void (*ndo_dfwd_del_station)(struct net_device *, void *); ++ int (*ndo_set_tx_maxrate)(struct net_device *, int, u32); ++ int (*ndo_get_iflink)(const struct net_device *); ++ int (*ndo_fill_metadata_dst)(struct net_device *, struct sk_buff *); ++ void (*ndo_set_rx_headroom)(struct net_device *, int); ++ int (*ndo_bpf)(struct net_device *, struct netdev_bpf *); ++ int (*ndo_xdp_xmit)(struct net_device *, int, struct xdp_frame **, u32); ++ struct net_device * (*ndo_xdp_get_xmit_slave)(struct net_device *, struct xdp_buff *); ++ int (*ndo_xsk_wakeup)(struct net_device *, u32, u32); ++ int (*ndo_tunnel_ctl)(struct net_device *, struct ip_tunnel_parm *, int); ++ struct net_device * (*ndo_get_peer_dev)(struct net_device *); ++ int (*ndo_fill_forward_path)(struct net_device_path_ctx *, struct net_device_path *); ++ ktime_t (*ndo_get_tstamp)(struct net_device *, const struct skb_shared_hwtstamps *, bool); ++}; ++ ++struct neigh_parms { ++ possible_net_t net; ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ struct list_head list; ++ int (*neigh_setup)(struct neighbour *); ++ struct neigh_table *tbl; ++ void *sysctl_table; ++ int dead; ++ refcount_t refcnt; ++ struct callback_head callback_head; ++ int reachable_time; ++ u32 qlen; ++ int data[14]; ++ long unsigned int data_state[1]; ++}; ++ ++struct xdp_md; ++ ++struct xdp_metadata_ops { ++ int (*xmo_rx_timestamp)(const struct xdp_md *, u64 *); ++ int (*xmo_rx_hash)(const struct xdp_md *, u32 *, enum xdp_rss_hash_type *); ++}; ++ ++struct xdp_md { ++ __u32 data; ++ __u32 data_end; ++ __u32 data_meta; ++ __u32 ingress_ifindex; ++ __u32 rx_queue_index; ++ __u32 egress_ifindex; ++}; ++ ++struct pcpu_lstats { ++ u64_stats_t packets; ++ u64_stats_t bytes; ++ struct u64_stats_sync syncp; ++}; ++ ++struct pcpu_sw_netstats { ++ u64_stats_t rx_packets; ++ u64_stats_t rx_bytes; ++ u64_stats_t tx_packets; ++ u64_stats_t tx_bytes; ++ struct u64_stats_sync syncp; ++}; ++ ++enum ethtool_phys_id_state { ++ ETHTOOL_ID_INACTIVE = 0, ++ ETHTOOL_ID_ACTIVE = 1, ++ ETHTOOL_ID_ON = 2, ++ ETHTOOL_ID_OFF = 3, ++}; ++ ++struct ethtool_drvinfo; ++ ++struct ethtool_regs; ++ ++struct ethtool_wolinfo; ++ ++struct ethtool_link_ext_state_info; ++ ++struct ethtool_link_ext_stats; ++ ++struct ethtool_eeprom; ++ ++struct ethtool_coalesce; ++ ++struct kernel_ethtool_coalesce; ++ ++struct ethtool_ringparam; ++ ++struct kernel_ethtool_ringparam; ++ ++struct ethtool_pause_stats; ++ ++struct ethtool_pauseparam; ++ ++struct ethtool_test; ++ ++struct ethtool_stats; ++ ++struct ethtool_rxnfc; ++ ++struct ethtool_flash; ++ ++struct ethtool_channels; ++ ++struct ethtool_dump; ++ ++struct ethtool_ts_info; ++ ++struct ethtool_modinfo; ++ ++struct ethtool_eee; ++ ++struct ethtool_tunable; ++ ++struct ethtool_link_ksettings; ++ ++struct ethtool_fec_stats; ++ ++struct ethtool_fecparam; ++ ++struct ethtool_module_eeprom; ++ ++struct ethtool_eth_phy_stats; ++ ++struct ethtool_eth_mac_stats; ++ ++struct ethtool_eth_ctrl_stats; ++ ++struct ethtool_rmon_stats; ++ ++struct ethtool_rmon_hist_range; ++ ++struct ethtool_module_power_mode_params; ++ ++struct ethtool_mm_state; ++ ++struct ethtool_mm_cfg; ++ ++struct ethtool_mm_stats; ++ ++struct ethtool_ops { ++ u32 cap_link_lanes_supported: 1; ++ u32 supported_coalesce_params; ++ u32 supported_ring_params; ++ void (*get_drvinfo)(struct net_device *, struct ethtool_drvinfo *); ++ int (*get_regs_len)(struct net_device *); ++ void (*get_regs)(struct net_device *, struct ethtool_regs *, void *); ++ void (*get_wol)(struct net_device *, struct ethtool_wolinfo *); ++ int (*set_wol)(struct net_device *, struct ethtool_wolinfo *); ++ u32 (*get_msglevel)(struct net_device *); ++ void (*set_msglevel)(struct net_device *, u32); ++ int (*nway_reset)(struct net_device *); ++ u32 (*get_link)(struct net_device *); ++ int (*get_link_ext_state)(struct net_device *, struct ethtool_link_ext_state_info *); ++ void (*get_link_ext_stats)(struct net_device *, struct ethtool_link_ext_stats *); ++ int (*get_eeprom_len)(struct net_device *); ++ int (*get_eeprom)(struct net_device *, struct ethtool_eeprom *, u8 *); ++ int (*set_eeprom)(struct net_device *, struct ethtool_eeprom *, u8 *); ++ int (*get_coalesce)(struct net_device *, struct ethtool_coalesce *, struct kernel_ethtool_coalesce *, struct netlink_ext_ack *); ++ int (*set_coalesce)(struct net_device *, struct ethtool_coalesce *, struct kernel_ethtool_coalesce *, struct netlink_ext_ack *); ++ void (*get_ringparam)(struct net_device *, struct ethtool_ringparam *, struct kernel_ethtool_ringparam *, struct netlink_ext_ack *); ++ int (*set_ringparam)(struct net_device *, struct ethtool_ringparam *, struct kernel_ethtool_ringparam *, struct netlink_ext_ack *); ++ void (*get_pause_stats)(struct net_device *, struct ethtool_pause_stats *); ++ void (*get_pauseparam)(struct net_device *, struct ethtool_pauseparam *); ++ int (*set_pauseparam)(struct net_device *, struct ethtool_pauseparam *); ++ void (*self_test)(struct net_device *, struct ethtool_test *, u64 *); ++ void (*get_strings)(struct net_device *, u32, u8 *); ++ int (*set_phys_id)(struct net_device *, enum ethtool_phys_id_state); ++ void (*get_ethtool_stats)(struct net_device *, struct ethtool_stats *, u64 *); ++ int (*begin)(struct net_device *); ++ void (*complete)(struct net_device *); ++ u32 (*get_priv_flags)(struct net_device *); ++ int (*set_priv_flags)(struct net_device *, u32); ++ int (*get_sset_count)(struct net_device *, int); ++ int (*get_rxnfc)(struct net_device *, struct ethtool_rxnfc *, u32 *); ++ int (*set_rxnfc)(struct net_device *, struct ethtool_rxnfc *); ++ int (*flash_device)(struct net_device *, struct ethtool_flash *); ++ int (*reset)(struct net_device *, u32 *); ++ u32 (*get_rxfh_key_size)(struct net_device *); ++ u32 (*get_rxfh_indir_size)(struct net_device *); ++ int (*get_rxfh)(struct net_device *, u32 *, u8 *, u8 *); ++ int (*set_rxfh)(struct net_device *, const u32 *, const u8 *, const u8); ++ int (*get_rxfh_context)(struct net_device *, u32 *, u8 *, u8 *, u32); ++ int (*set_rxfh_context)(struct net_device *, const u32 *, const u8 *, const u8, u32 *, bool); ++ void (*get_channels)(struct net_device *, struct ethtool_channels *); ++ int (*set_channels)(struct net_device *, struct ethtool_channels *); ++ int (*get_dump_flag)(struct net_device *, struct ethtool_dump *); ++ int (*get_dump_data)(struct net_device *, struct ethtool_dump *, void *); ++ int (*set_dump)(struct net_device *, struct ethtool_dump *); ++ int (*get_ts_info)(struct net_device *, struct ethtool_ts_info *); ++ int (*get_module_info)(struct net_device *, struct ethtool_modinfo *); ++ int (*get_module_eeprom)(struct net_device *, struct ethtool_eeprom *, u8 *); ++ int (*get_eee)(struct net_device *, struct ethtool_eee *); ++ int (*set_eee)(struct net_device *, struct ethtool_eee *); ++ int (*get_tunable)(struct net_device *, const struct ethtool_tunable *, void *); ++ int (*set_tunable)(struct net_device *, const struct ethtool_tunable *, const void *); ++ int (*get_per_queue_coalesce)(struct net_device *, u32, struct ethtool_coalesce *); ++ int (*set_per_queue_coalesce)(struct net_device *, u32, struct ethtool_coalesce *); ++ int (*get_link_ksettings)(struct net_device *, struct ethtool_link_ksettings *); ++ int (*set_link_ksettings)(struct net_device *, const struct ethtool_link_ksettings *); ++ void (*get_fec_stats)(struct net_device *, struct ethtool_fec_stats *); ++ int (*get_fecparam)(struct net_device *, struct ethtool_fecparam *); ++ int (*set_fecparam)(struct net_device *, struct ethtool_fecparam *); ++ void (*get_ethtool_phy_stats)(struct net_device *, struct ethtool_stats *, u64 *); ++ int (*get_phy_tunable)(struct net_device *, const struct ethtool_tunable *, void *); ++ int (*set_phy_tunable)(struct net_device *, const struct ethtool_tunable *, const void *); ++ int (*get_module_eeprom_by_page)(struct net_device *, const struct ethtool_module_eeprom *, struct netlink_ext_ack *); ++ void (*get_eth_phy_stats)(struct net_device *, struct ethtool_eth_phy_stats *); ++ void (*get_eth_mac_stats)(struct net_device *, struct ethtool_eth_mac_stats *); ++ void (*get_eth_ctrl_stats)(struct net_device *, struct ethtool_eth_ctrl_stats *); ++ void (*get_rmon_stats)(struct net_device *, struct ethtool_rmon_stats *, const struct ethtool_rmon_hist_range **); ++ int (*get_module_power_mode)(struct net_device *, struct ethtool_module_power_mode_params *, struct netlink_ext_ack *); ++ int (*set_module_power_mode)(struct net_device *, const struct ethtool_module_power_mode_params *, struct netlink_ext_ack *); ++ int (*get_mm)(struct net_device *, struct ethtool_mm_state *); ++ int (*set_mm)(struct net_device *, struct ethtool_mm_cfg *, struct netlink_ext_ack *); ++ void (*get_mm_stats)(struct net_device *, struct ethtool_mm_stats *); ++}; ++ ++struct l3mdev_ops { ++ u32 (*l3mdev_fib_table)(const struct net_device *); ++ struct sk_buff * (*l3mdev_l3_rcv)(struct net_device *, struct sk_buff *, u16); ++ struct sk_buff * (*l3mdev_l3_out)(struct net_device *, struct sock *, struct sk_buff *, u16); ++ struct dst_entry * (*l3mdev_link_scope_lookup)(const struct net_device *, struct flowi6 *); ++}; ++ ++struct nd_opt_hdr; ++ ++struct ndisc_options; ++ ++struct prefix_info; ++ ++struct ndisc_ops { ++ int (*is_useropt)(u8); ++ int (*parse_options)(const struct net_device *, struct nd_opt_hdr *, struct ndisc_options *); ++ void (*update)(const struct net_device *, struct neighbour *, u32, u8, const struct ndisc_options *); ++ int (*opt_addr_space)(const struct net_device *, u8, struct neighbour *, u8 *, u8 **); ++ void (*fill_addr_option)(const struct net_device *, struct sk_buff *, u8, const u8 *); ++ void (*prefix_rcv_add_addr)(struct net *, struct net_device *, const struct prefix_info *, struct inet6_dev *, struct in6_addr *, int, u32, bool, bool, __u32, u32, bool); ++}; ++ ++enum tls_offload_ctx_dir { ++ TLS_OFFLOAD_CTX_DIR_RX = 0, ++ TLS_OFFLOAD_CTX_DIR_TX = 1, ++}; ++ ++struct tls_crypto_info; ++ ++struct tls_context; ++ ++struct tlsdev_ops { ++ int (*tls_dev_add)(struct net_device *, struct sock *, enum tls_offload_ctx_dir, struct tls_crypto_info *, u32); ++ void (*tls_dev_del)(struct net_device *, struct tls_context *, enum tls_offload_ctx_dir); ++ int (*tls_dev_resync)(struct net_device *, struct sock *, u32, u8 *, enum tls_offload_ctx_dir); ++}; ++ ++struct ipv6_devstat { ++ struct proc_dir_entry *proc_dir_entry; ++ struct ipstats_mib *ipv6; ++ struct icmpv6_mib_device *icmpv6dev; ++ struct icmpv6msg_mib_device *icmpv6msgdev; ++}; ++ ++struct ifmcaddr6; ++ ++struct ifacaddr6; ++ ++struct inet6_dev { ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ struct list_head addr_list; ++ struct ifmcaddr6 *mc_list; ++ struct ifmcaddr6 *mc_tomb; ++ unsigned char mc_qrv; ++ unsigned char mc_gq_running; ++ unsigned char mc_ifc_count; ++ unsigned char mc_dad_count; ++ long unsigned int mc_v1_seen; ++ long unsigned int mc_qi; ++ long unsigned int mc_qri; ++ long unsigned int mc_maxdelay; ++ struct delayed_work mc_gq_work; ++ struct delayed_work mc_ifc_work; ++ struct delayed_work mc_dad_work; ++ struct delayed_work mc_query_work; ++ struct delayed_work mc_report_work; ++ struct sk_buff_head mc_query_queue; ++ struct sk_buff_head mc_report_queue; ++ spinlock_t mc_query_lock; ++ spinlock_t mc_report_lock; ++ struct mutex mc_lock; ++ struct ifacaddr6 *ac_list; ++ rwlock_t lock; ++ refcount_t refcnt; ++ __u32 if_flags; ++ int dead; ++ u32 desync_factor; ++ struct list_head tempaddr_list; ++ struct in6_addr token; ++ struct neigh_parms *nd_parms; ++ struct ipv6_devconf cnf; ++ struct ipv6_devstat stats; ++ struct timer_list rs_timer; ++ __s32 rs_interval; ++ __u8 rs_probes; ++ long unsigned int tstamp; ++ struct callback_head rcu; ++ unsigned int ra_mtu; ++}; ++ ++struct rtnl_link_ops { ++ struct list_head list; ++ const char *kind; ++ size_t priv_size; ++ struct net_device * (*alloc)(struct nlattr **, const char *, unsigned char, unsigned int, unsigned int); ++ void (*setup)(struct net_device *); ++ bool netns_refund; ++ unsigned int maxtype; ++ const struct nla_policy *policy; ++ int (*validate)(struct nlattr **, struct nlattr **, struct netlink_ext_ack *); ++ int (*newlink)(struct net *, struct net_device *, struct nlattr **, struct nlattr **, struct netlink_ext_ack *); ++ int (*changelink)(struct net_device *, struct nlattr **, struct nlattr **, struct netlink_ext_ack *); ++ void (*dellink)(struct net_device *, struct list_head *); ++ size_t (*get_size)(const struct net_device *); ++ int (*fill_info)(struct sk_buff *, const struct net_device *); ++ size_t (*get_xstats_size)(const struct net_device *); ++ int (*fill_xstats)(struct sk_buff *, const struct net_device *); ++ unsigned int (*get_num_tx_queues)(); ++ unsigned int (*get_num_rx_queues)(); ++ unsigned int slave_maxtype; ++ const struct nla_policy *slave_policy; ++ int (*slave_changelink)(struct net_device *, struct net_device *, struct nlattr **, struct nlattr **, struct netlink_ext_ack *); ++ size_t (*get_slave_size)(const struct net_device *, const struct net_device *); ++ int (*fill_slave_info)(struct sk_buff *, const struct net_device *, const struct net_device *); ++ struct net * (*get_link_net)(const struct net_device *); ++ size_t (*get_linkxstats_size)(const struct net_device *, int); ++ int (*fill_linkxstats)(struct sk_buff *, const struct net_device *, int *, int); ++}; ++ ++struct macsec_context; ++ ++struct macsec_ops { ++ int (*mdo_dev_open)(struct macsec_context *); ++ int (*mdo_dev_stop)(struct macsec_context *); ++ int (*mdo_add_secy)(struct macsec_context *); ++ int (*mdo_upd_secy)(struct macsec_context *); ++ int (*mdo_del_secy)(struct macsec_context *); ++ int (*mdo_add_rxsc)(struct macsec_context *); ++ int (*mdo_upd_rxsc)(struct macsec_context *); ++ int (*mdo_del_rxsc)(struct macsec_context *); ++ int (*mdo_add_rxsa)(struct macsec_context *); ++ int (*mdo_upd_rxsa)(struct macsec_context *); ++ int (*mdo_del_rxsa)(struct macsec_context *); ++ int (*mdo_add_txsa)(struct macsec_context *); ++ int (*mdo_upd_txsa)(struct macsec_context *); ++ int (*mdo_del_txsa)(struct macsec_context *); ++ int (*mdo_get_dev_stats)(struct macsec_context *); ++ int (*mdo_get_tx_sc_stats)(struct macsec_context *); ++ int (*mdo_get_tx_sa_stats)(struct macsec_context *); ++ int (*mdo_get_rx_sc_stats)(struct macsec_context *); ++ int (*mdo_get_rx_sa_stats)(struct macsec_context *); ++}; ++ ++struct udp_tunnel_nic_table_info { ++ unsigned int n_entries; ++ unsigned int tunnel_types; ++}; ++ ++struct udp_tunnel_info; ++ ++struct udp_tunnel_nic_shared; ++ ++struct udp_tunnel_nic_info { ++ int (*set_port)(struct net_device *, unsigned int, unsigned int, struct udp_tunnel_info *); ++ int (*unset_port)(struct net_device *, unsigned int, unsigned int, struct udp_tunnel_info *); ++ int (*sync_table)(struct net_device *, unsigned int); ++ struct udp_tunnel_nic_shared *shared; ++ unsigned int flags; ++ struct udp_tunnel_nic_table_info tables[4]; ++}; ++ ++enum { ++ NETIF_MSG_DRV_BIT = 0, ++ NETIF_MSG_PROBE_BIT = 1, ++ NETIF_MSG_LINK_BIT = 2, ++ NETIF_MSG_TIMER_BIT = 3, ++ NETIF_MSG_IFDOWN_BIT = 4, ++ NETIF_MSG_IFUP_BIT = 5, ++ NETIF_MSG_RX_ERR_BIT = 6, ++ NETIF_MSG_TX_ERR_BIT = 7, ++ NETIF_MSG_TX_QUEUED_BIT = 8, ++ NETIF_MSG_INTR_BIT = 9, ++ NETIF_MSG_TX_DONE_BIT = 10, ++ NETIF_MSG_RX_STATUS_BIT = 11, ++ NETIF_MSG_PKTDATA_BIT = 12, ++ NETIF_MSG_HW_BIT = 13, ++ NETIF_MSG_WOL_BIT = 14, ++ NETIF_MSG_CLASS_COUNT = 15, ++}; ++ ++enum { ++ RTAX_UNSPEC = 0, ++ RTAX_LOCK = 1, ++ RTAX_MTU = 2, ++ RTAX_WINDOW = 3, ++ RTAX_RTT = 4, ++ RTAX_RTTVAR = 5, ++ RTAX_SSTHRESH = 6, ++ RTAX_CWND = 7, ++ RTAX_ADVMSS = 8, ++ RTAX_REORDERING = 9, ++ RTAX_HOPLIMIT = 10, ++ RTAX_INITCWND = 11, ++ RTAX_FEATURES = 12, ++ RTAX_RTO_MIN = 13, ++ RTAX_INITRWND = 14, ++ RTAX_QUICKACK = 15, ++ RTAX_CC_ALGO = 16, ++ RTAX_FASTOPEN_NO_COOKIE = 17, ++ __RTAX_MAX = 18, ++}; ++ ++struct netlink_range_validation { ++ u64 min; ++ u64 max; ++}; ++ ++struct netlink_range_validation_signed { ++ s64 min; ++ s64 max; ++}; ++ ++enum { ++ NEIGH_VAR_MCAST_PROBES = 0, ++ NEIGH_VAR_UCAST_PROBES = 1, ++ NEIGH_VAR_APP_PROBES = 2, ++ NEIGH_VAR_MCAST_REPROBES = 3, ++ NEIGH_VAR_RETRANS_TIME = 4, ++ NEIGH_VAR_BASE_REACHABLE_TIME = 5, ++ NEIGH_VAR_DELAY_PROBE_TIME = 6, ++ NEIGH_VAR_INTERVAL_PROBE_TIME_MS = 7, ++ NEIGH_VAR_GC_STALETIME = 8, ++ NEIGH_VAR_QUEUE_LEN_BYTES = 9, ++ NEIGH_VAR_PROXY_QLEN = 10, ++ NEIGH_VAR_ANYCAST_DELAY = 11, ++ NEIGH_VAR_PROXY_DELAY = 12, ++ NEIGH_VAR_LOCKTIME = 13, ++ NEIGH_VAR_QUEUE_LEN = 14, ++ NEIGH_VAR_RETRANS_TIME_MS = 15, ++ NEIGH_VAR_BASE_REACHABLE_TIME_MS = 16, ++ NEIGH_VAR_GC_INTERVAL = 17, ++ NEIGH_VAR_GC_THRESH1 = 18, ++ NEIGH_VAR_GC_THRESH2 = 19, ++ NEIGH_VAR_GC_THRESH3 = 20, ++ NEIGH_VAR_MAX = 21, ++}; ++ ++struct pneigh_entry; ++ ++struct neigh_statistics; ++ ++struct neigh_hash_table; ++ ++struct neigh_table { ++ int family; ++ unsigned int entry_size; ++ unsigned int key_len; ++ __be16 protocol; ++ __u32 (*hash)(const void *, const struct net_device *, __u32 *); ++ bool (*key_eq)(const struct neighbour *, const void *); ++ int (*constructor)(struct neighbour *); ++ int (*pconstructor)(struct pneigh_entry *); ++ void (*pdestructor)(struct pneigh_entry *); ++ void (*proxy_redo)(struct sk_buff *); ++ int (*is_multicast)(const void *); ++ bool (*allow_add)(const struct net_device *, struct netlink_ext_ack *); ++ char *id; ++ struct neigh_parms parms; ++ struct list_head parms_list; ++ int gc_interval; ++ int gc_thresh1; ++ int gc_thresh2; ++ int gc_thresh3; ++ long unsigned int last_flush; ++ struct delayed_work gc_work; ++ struct delayed_work managed_work; ++ struct timer_list proxy_timer; ++ struct sk_buff_head proxy_queue; ++ atomic_t entries; ++ atomic_t gc_entries; ++ struct list_head gc_list; ++ struct list_head managed_list; ++ rwlock_t lock; ++ long unsigned int last_rand; ++ struct neigh_statistics *stats; ++ struct neigh_hash_table *nht; ++ struct pneigh_entry **phash_buckets; ++}; ++ ++struct neigh_statistics { ++ long unsigned int allocs; ++ long unsigned int destroys; ++ long unsigned int hash_grows; ++ long unsigned int res_failed; ++ long unsigned int lookups; ++ long unsigned int hits; ++ long unsigned int rcv_probes_mcast; ++ long unsigned int rcv_probes_ucast; ++ long unsigned int periodic_gc_runs; ++ long unsigned int forced_gc_runs; ++ long unsigned int unres_discards; ++ long unsigned int table_fulls; ++}; ++ ++struct neigh_ops { ++ int family; ++ void (*solicit)(struct neighbour *, struct sk_buff *); ++ void (*error_report)(struct neighbour *, struct sk_buff *); ++ int (*output)(struct neighbour *, struct sk_buff *); ++ int (*connected_output)(struct neighbour *, struct sk_buff *); ++}; ++ ++struct pneigh_entry { ++ struct pneigh_entry *next; ++ possible_net_t net; ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ u32 flags; ++ u8 protocol; ++ u32 key[0]; ++}; ++ ++struct neigh_hash_table { ++ struct neighbour **hash_buckets; ++ unsigned int hash_shift; ++ __u32 hash_rnd[4]; ++ struct callback_head rcu; ++}; ++ ++enum { ++ TCP_ESTABLISHED = 1, ++ TCP_SYN_SENT = 2, ++ TCP_SYN_RECV = 3, ++ TCP_FIN_WAIT1 = 4, ++ TCP_FIN_WAIT2 = 5, ++ TCP_TIME_WAIT = 6, ++ TCP_CLOSE = 7, ++ TCP_CLOSE_WAIT = 8, ++ TCP_LAST_ACK = 9, ++ TCP_LISTEN = 10, ++ TCP_CLOSING = 11, ++ TCP_NEW_SYN_RECV = 12, ++ TCP_MAX_STATES = 13, ++}; ++ ++struct fib_rule_hdr { ++ __u8 family; ++ __u8 dst_len; ++ __u8 src_len; ++ __u8 tos; ++ __u8 table; ++ __u8 res1; ++ __u8 res2; ++ __u8 action; ++ __u32 flags; ++}; ++ ++struct fib_rule_port_range { ++ __u16 start; ++ __u16 end; ++}; ++ ++struct fib_kuid_range { ++ kuid_t start; ++ kuid_t end; ++}; ++ ++struct fib_rule { ++ struct list_head list; ++ int iifindex; ++ int oifindex; ++ u32 mark; ++ u32 mark_mask; ++ u32 flags; ++ u32 table; ++ u8 action; ++ u8 l3mdev; ++ u8 proto; ++ u8 ip_proto; ++ u32 target; ++ __be64 tun_id; ++ struct fib_rule *ctarget; ++ struct net *fr_net; ++ refcount_t refcnt; ++ u32 pref; ++ int suppress_ifgroup; ++ int suppress_prefixlen; ++ char iifname[16]; ++ char oifname[16]; ++ struct fib_kuid_range uid_range; ++ struct fib_rule_port_range sport_range; ++ struct fib_rule_port_range dport_range; ++ struct callback_head rcu; ++}; ++ ++struct fib_lookup_arg { ++ void *lookup_ptr; ++ const void *lookup_data; ++ void *result; ++ struct fib_rule *rule; ++ u32 table; ++ int flags; ++}; ++ ++struct smc_hashinfo; ++ ++struct sk_psock; ++ ++struct request_sock_ops; ++ ++struct timewait_sock_ops; ++ ++struct raw_hashinfo; ++ ++struct proto { ++ void (*close)(struct sock *, long int); ++ int (*pre_connect)(struct sock *, struct sockaddr *, int); ++ int (*connect)(struct sock *, struct sockaddr *, int); ++ int (*disconnect)(struct sock *, int); ++ struct sock * (*accept)(struct sock *, int, int *, bool); ++ int (*ioctl)(struct sock *, int, long unsigned int); ++ int (*init)(struct sock *); ++ void (*destroy)(struct sock *); ++ void (*shutdown)(struct sock *, int); ++ int (*setsockopt)(struct sock *, int, int, sockptr_t, unsigned int); ++ int (*getsockopt)(struct sock *, int, int, char *, int *); ++ void (*keepalive)(struct sock *, int); ++ int (*compat_ioctl)(struct sock *, unsigned int, long unsigned int); ++ int (*sendmsg)(struct sock *, struct msghdr *, size_t); ++ int (*recvmsg)(struct sock *, struct msghdr *, size_t, int, int *); ++ int (*sendpage)(struct sock *, struct page *, int, size_t, int); ++ int (*bind)(struct sock *, struct sockaddr *, int); ++ int (*bind_add)(struct sock *, struct sockaddr *, int); ++ int (*backlog_rcv)(struct sock *, struct sk_buff *); ++ bool (*bpf_bypass_getsockopt)(int, int); ++ void (*release_cb)(struct sock *); ++ int (*hash)(struct sock *); ++ void (*unhash)(struct sock *); ++ void (*rehash)(struct sock *); ++ int (*get_port)(struct sock *, short unsigned int); ++ void (*put_port)(struct sock *); ++ int (*psock_update_sk_prot)(struct sock *, struct sk_psock *, bool); ++ unsigned int inuse_idx; ++ int (*forward_alloc_get)(const struct sock *); ++ bool (*stream_memory_free)(const struct sock *, int); ++ bool (*sock_is_readable)(struct sock *); ++ void (*enter_memory_pressure)(struct sock *); ++ void (*leave_memory_pressure)(struct sock *); ++ atomic_long_t *memory_allocated; ++ int *per_cpu_fw_alloc; ++ struct percpu_counter *sockets_allocated; ++ long unsigned int *memory_pressure; ++ long int *sysctl_mem; ++ int *sysctl_wmem; ++ int *sysctl_rmem; ++ u32 sysctl_wmem_offset; ++ u32 sysctl_rmem_offset; ++ int max_header; ++ bool no_autobind; ++ struct kmem_cache *slab; ++ unsigned int obj_size; ++ slab_flags_t slab_flags; ++ unsigned int useroffset; ++ unsigned int usersize; ++ unsigned int *orphan_count; ++ struct request_sock_ops *rsk_prot; ++ struct timewait_sock_ops *twsk_prot; ++ union { ++ struct inet_hashinfo *hashinfo; ++ struct udp_table *udp_table; ++ struct raw_hashinfo *raw_hash; ++ struct smc_hashinfo *smc_hash; ++ } h; ++ struct module *owner; ++ char name[32]; ++ struct list_head node; ++ int (*diag_destroy)(struct sock *, int); ++}; ++ ++struct request_sock; ++ ++struct request_sock_ops { ++ int family; ++ unsigned int obj_size; ++ struct kmem_cache *slab; ++ char *slab_name; ++ int (*rtx_syn_ack)(const struct sock *, struct request_sock *); ++ void (*send_ack)(const struct sock *, struct sk_buff *, struct request_sock *); ++ void (*send_reset)(const struct sock *, struct sk_buff *); ++ void (*destructor)(struct request_sock *); ++ void (*syn_ack_timeout)(const struct request_sock *); ++}; ++ ++struct timewait_sock_ops { ++ struct kmem_cache *twsk_slab; ++ char *twsk_slab_name; ++ unsigned int twsk_obj_size; ++ int (*twsk_unique)(struct sock *, struct sock *, void *); ++ void (*twsk_destructor)(struct sock *); ++}; ++ ++struct saved_syn; ++ ++struct request_sock { ++ struct sock_common __req_common; ++ struct request_sock *dl_next; ++ u16 mss; ++ u8 num_retrans; ++ u8 syncookie: 1; ++ u8 num_timeout: 7; ++ u32 ts_recent; ++ struct timer_list rsk_timer; ++ const struct request_sock_ops *rsk_ops; ++ struct sock *sk; ++ struct saved_syn *saved_syn; ++ u32 secid; ++ u32 peer_secid; ++ u32 timeout; ++}; ++ ++struct saved_syn { ++ u32 mac_hdrlen; ++ u32 network_hdrlen; ++ u32 tcp_hdrlen; ++ u8 data[0]; ++}; ++ ++enum tsq_enum { ++ TSQ_THROTTLED = 0, ++ TSQ_QUEUED = 1, ++ TCP_TSQ_DEFERRED = 2, ++ TCP_WRITE_TIMER_DEFERRED = 3, ++ TCP_DELACK_TIMER_DEFERRED = 4, ++ TCP_MTU_REDUCED_DEFERRED = 5, ++}; ++ ++struct ip6_sf_list { ++ struct ip6_sf_list *sf_next; ++ struct in6_addr sf_addr; ++ long unsigned int sf_count[2]; ++ unsigned char sf_gsresp; ++ unsigned char sf_oldin; ++ unsigned char sf_crcount; ++ struct callback_head rcu; ++}; ++ ++struct ifmcaddr6 { ++ struct in6_addr mca_addr; ++ struct inet6_dev *idev; ++ struct ifmcaddr6 *next; ++ struct ip6_sf_list *mca_sources; ++ struct ip6_sf_list *mca_tomb; ++ unsigned int mca_sfmode; ++ unsigned char mca_crcount; ++ long unsigned int mca_sfcount[2]; ++ struct delayed_work mca_work; ++ unsigned int mca_flags; ++ int mca_users; ++ refcount_t mca_refcnt; ++ long unsigned int mca_cstamp; ++ long unsigned int mca_tstamp; ++ struct callback_head rcu; ++}; ++ ++struct ifacaddr6 { ++ struct in6_addr aca_addr; ++ struct fib6_info *aca_rt; ++ struct ifacaddr6 *aca_next; ++ struct hlist_node aca_addr_lst; ++ int aca_users; ++ refcount_t aca_refcnt; ++ long unsigned int aca_cstamp; ++ long unsigned int aca_tstamp; ++ struct callback_head rcu; ++}; ++ ++enum nfs_opnum4 { ++ OP_ACCESS = 3, ++ OP_CLOSE = 4, ++ OP_COMMIT = 5, ++ OP_CREATE = 6, ++ OP_DELEGPURGE = 7, ++ OP_DELEGRETURN = 8, ++ OP_GETATTR = 9, ++ OP_GETFH = 10, ++ OP_LINK = 11, ++ OP_LOCK = 12, ++ OP_LOCKT = 13, ++ OP_LOCKU = 14, ++ OP_LOOKUP = 15, ++ OP_LOOKUPP = 16, ++ OP_NVERIFY = 17, ++ OP_OPEN = 18, ++ OP_OPENATTR = 19, ++ OP_OPEN_CONFIRM = 20, ++ OP_OPEN_DOWNGRADE = 21, ++ OP_PUTFH = 22, ++ OP_PUTPUBFH = 23, ++ OP_PUTROOTFH = 24, ++ OP_READ = 25, ++ OP_READDIR = 26, ++ OP_READLINK = 27, ++ OP_REMOVE = 28, ++ OP_RENAME = 29, ++ OP_RENEW = 30, ++ OP_RESTOREFH = 31, ++ OP_SAVEFH = 32, ++ OP_SECINFO = 33, ++ OP_SETATTR = 34, ++ OP_SETCLIENTID = 35, ++ OP_SETCLIENTID_CONFIRM = 36, ++ OP_VERIFY = 37, ++ OP_WRITE = 38, ++ OP_RELEASE_LOCKOWNER = 39, ++ OP_BACKCHANNEL_CTL = 40, ++ OP_BIND_CONN_TO_SESSION = 41, ++ OP_EXCHANGE_ID = 42, ++ OP_CREATE_SESSION = 43, ++ OP_DESTROY_SESSION = 44, ++ OP_FREE_STATEID = 45, ++ OP_GET_DIR_DELEGATION = 46, ++ OP_GETDEVICEINFO = 47, ++ OP_GETDEVICELIST = 48, ++ OP_LAYOUTCOMMIT = 49, ++ OP_LAYOUTGET = 50, ++ OP_LAYOUTRETURN = 51, ++ OP_SECINFO_NO_NAME = 52, ++ OP_SEQUENCE = 53, ++ OP_SET_SSV = 54, ++ OP_TEST_STATEID = 55, ++ OP_WANT_DELEGATION = 56, ++ OP_DESTROY_CLIENTID = 57, ++ OP_RECLAIM_COMPLETE = 58, ++ OP_ALLOCATE = 59, ++ OP_COPY = 60, ++ OP_COPY_NOTIFY = 61, ++ OP_DEALLOCATE = 62, ++ OP_IO_ADVISE = 63, ++ OP_LAYOUTERROR = 64, ++ OP_LAYOUTSTATS = 65, ++ OP_OFFLOAD_CANCEL = 66, ++ OP_OFFLOAD_STATUS = 67, ++ OP_READ_PLUS = 68, ++ OP_SEEK = 69, ++ OP_WRITE_SAME = 70, ++ OP_CLONE = 71, ++ OP_GETXATTR = 72, ++ OP_SETXATTR = 73, ++ OP_LISTXATTRS = 74, ++ OP_REMOVEXATTR = 75, ++ OP_ILLEGAL = 10044, ++}; ++ ++struct blk_zone { ++ __u64 start; ++ __u64 len; ++ __u64 wp; ++ __u8 type; ++ __u8 cond; ++ __u8 non_seq; ++ __u8 reset; ++ __u8 resv[4]; ++ __u64 capacity; ++ __u8 reserved[24]; ++}; ++ ++struct blk_integrity_iter; ++ ++typedef blk_status_t integrity_processing_fn(struct blk_integrity_iter *); ++ ++typedef void integrity_prepare_fn(struct request *); ++ ++typedef void integrity_complete_fn(struct request *, unsigned int); ++ ++struct blk_integrity_profile { ++ integrity_processing_fn *generate_fn; ++ integrity_processing_fn *verify_fn; ++ integrity_prepare_fn *prepare_fn; ++ integrity_complete_fn *complete_fn; ++ const char *name; ++}; ++ ++typedef int (*report_zones_cb)(struct blk_zone *, unsigned int, void *); ++ ++enum blk_unique_id { ++ BLK_UID_T10 = 1, ++ BLK_UID_EUI64 = 2, ++ BLK_UID_NAA = 3, ++}; ++ ++struct hd_geometry; ++ ++struct pr_ops; ++ ++struct block_device_operations { ++ void (*submit_bio)(struct bio *); ++ int (*poll_bio)(struct bio *, struct io_comp_batch *, unsigned int); ++ int (*open)(struct block_device *, fmode_t); ++ void (*release)(struct gendisk *, fmode_t); ++ int (*ioctl)(struct block_device *, fmode_t, unsigned int, long unsigned int); ++ int (*compat_ioctl)(struct block_device *, fmode_t, unsigned int, long unsigned int); ++ unsigned int (*check_events)(struct gendisk *, unsigned int); ++ void (*unlock_native_capacity)(struct gendisk *); ++ int (*getgeo)(struct block_device *, struct hd_geometry *); ++ int (*set_read_only)(struct block_device *, bool); ++ void (*free_disk)(struct gendisk *); ++ void (*swap_slot_free_notify)(struct block_device *, long unsigned int); ++ int (*report_zones)(struct gendisk *, sector_t, unsigned int, report_zones_cb, void *); ++ char * (*devnode)(struct gendisk *, umode_t *); ++ int (*get_unique_id)(struct gendisk *, u8 *, enum blk_unique_id); ++ struct module *owner; ++ const struct pr_ops *pr_ops; ++ int (*alternative_gpt_sector)(struct gendisk *, sector_t *); ++}; ++ ++struct blk_independent_access_range { ++ struct kobject kobj; ++ sector_t sector; ++ sector_t nr_sectors; ++}; ++ ++struct blk_independent_access_ranges { ++ struct kobject kobj; ++ bool sysfs_registered; ++ unsigned int nr_ia_ranges; ++ struct blk_independent_access_range ia_range[0]; ++}; ++ ++enum blk_eh_timer_return { ++ BLK_EH_DONE = 0, ++ BLK_EH_RESET_TIMER = 1, ++}; ++ ++struct blk_mq_hw_ctx; ++ ++struct blk_mq_queue_data; ++ ++struct blk_mq_ops { ++ blk_status_t (*queue_rq)(struct blk_mq_hw_ctx *, const struct blk_mq_queue_data *); ++ void (*commit_rqs)(struct blk_mq_hw_ctx *); ++ void (*queue_rqs)(struct request **); ++ int (*get_budget)(struct request_queue *); ++ void (*put_budget)(struct request_queue *, int); ++ void (*set_rq_budget_token)(struct request *, int); ++ int (*get_rq_budget_token)(struct request *); ++ enum blk_eh_timer_return (*timeout)(struct request *); ++ int (*poll)(struct blk_mq_hw_ctx *, struct io_comp_batch *); ++ void (*complete)(struct request *); ++ int (*init_hctx)(struct blk_mq_hw_ctx *, void *, unsigned int); ++ void (*exit_hctx)(struct blk_mq_hw_ctx *, unsigned int); ++ int (*init_request)(struct blk_mq_tag_set *, struct request *, unsigned int, unsigned int); ++ void (*exit_request)(struct blk_mq_tag_set *, struct request *, unsigned int); ++ void (*cleanup_rq)(struct request *); ++ bool (*busy)(struct request_queue *); ++ void (*map_queues)(struct blk_mq_tag_set *); ++ void (*show_rq)(struct seq_file *, struct request *); ++}; ++ ++enum pr_type { ++ PR_WRITE_EXCLUSIVE = 1, ++ PR_EXCLUSIVE_ACCESS = 2, ++ PR_WRITE_EXCLUSIVE_REG_ONLY = 3, ++ PR_EXCLUSIVE_ACCESS_REG_ONLY = 4, ++ PR_WRITE_EXCLUSIVE_ALL_REGS = 5, ++ PR_EXCLUSIVE_ACCESS_ALL_REGS = 6, ++}; ++ ++struct pr_ops { ++ int (*pr_register)(struct block_device *, u64, u64, u32); ++ int (*pr_reserve)(struct block_device *, u64, enum pr_type, u32); ++ int (*pr_release)(struct block_device *, u64, enum pr_type); ++ int (*pr_preempt)(struct block_device *, u64, u64, enum pr_type, bool); ++ int (*pr_clear)(struct block_device *, u64); ++}; ++ ++enum { ++ UNAME26 = 131072, ++ ADDR_NO_RANDOMIZE = 262144, ++ FDPIC_FUNCPTRS = 524288, ++ MMAP_PAGE_ZERO = 1048576, ++ ADDR_COMPAT_LAYOUT = 2097152, ++ READ_IMPLIES_EXEC = 4194304, ++ ADDR_LIMIT_32BIT = 8388608, ++ SHORT_INODE = 16777216, ++ WHOLE_SECONDS = 33554432, ++ STICKY_TIMEOUTS = 67108864, ++ ADDR_LIMIT_3GB = 134217728, ++}; ++ ++struct seq_buf { ++ char *buffer; ++ size_t size; ++ size_t len; ++ loff_t readpos; ++}; ++ ++struct trace_seq { ++ char buffer[4096]; ++ struct seq_buf seq; ++ int full; ++}; ++ ++enum perf_sw_ids { ++ PERF_COUNT_SW_CPU_CLOCK = 0, ++ PERF_COUNT_SW_TASK_CLOCK = 1, ++ PERF_COUNT_SW_PAGE_FAULTS = 2, ++ PERF_COUNT_SW_CONTEXT_SWITCHES = 3, ++ PERF_COUNT_SW_CPU_MIGRATIONS = 4, ++ PERF_COUNT_SW_PAGE_FAULTS_MIN = 5, ++ PERF_COUNT_SW_PAGE_FAULTS_MAJ = 6, ++ PERF_COUNT_SW_ALIGNMENT_FAULTS = 7, ++ PERF_COUNT_SW_EMULATION_FAULTS = 8, ++ PERF_COUNT_SW_DUMMY = 9, ++ PERF_COUNT_SW_BPF_OUTPUT = 10, ++ PERF_COUNT_SW_CGROUP_SWITCHES = 11, ++ PERF_COUNT_SW_MAX = 12, ++}; ++ ++enum perf_branch_sample_type_shift { ++ PERF_SAMPLE_BRANCH_USER_SHIFT = 0, ++ PERF_SAMPLE_BRANCH_KERNEL_SHIFT = 1, ++ PERF_SAMPLE_BRANCH_HV_SHIFT = 2, ++ PERF_SAMPLE_BRANCH_ANY_SHIFT = 3, ++ PERF_SAMPLE_BRANCH_ANY_CALL_SHIFT = 4, ++ PERF_SAMPLE_BRANCH_ANY_RETURN_SHIFT = 5, ++ PERF_SAMPLE_BRANCH_IND_CALL_SHIFT = 6, ++ PERF_SAMPLE_BRANCH_ABORT_TX_SHIFT = 7, ++ PERF_SAMPLE_BRANCH_IN_TX_SHIFT = 8, ++ PERF_SAMPLE_BRANCH_NO_TX_SHIFT = 9, ++ PERF_SAMPLE_BRANCH_COND_SHIFT = 10, ++ PERF_SAMPLE_BRANCH_CALL_STACK_SHIFT = 11, ++ PERF_SAMPLE_BRANCH_IND_JUMP_SHIFT = 12, ++ PERF_SAMPLE_BRANCH_CALL_SHIFT = 13, ++ PERF_SAMPLE_BRANCH_NO_FLAGS_SHIFT = 14, ++ PERF_SAMPLE_BRANCH_NO_CYCLES_SHIFT = 15, ++ PERF_SAMPLE_BRANCH_TYPE_SAVE_SHIFT = 16, ++ PERF_SAMPLE_BRANCH_HW_INDEX_SHIFT = 17, ++ PERF_SAMPLE_BRANCH_PRIV_SAVE_SHIFT = 18, ++ PERF_SAMPLE_BRANCH_MAX_SHIFT = 19, ++}; ++ ++struct perf_event_attr { ++ __u32 type; ++ __u32 size; ++ __u64 config; ++ union { ++ __u64 sample_period; ++ __u64 sample_freq; ++ }; ++ __u64 sample_type; ++ __u64 read_format; ++ __u64 disabled: 1; ++ __u64 inherit: 1; ++ __u64 pinned: 1; ++ __u64 exclusive: 1; ++ __u64 exclude_user: 1; ++ __u64 exclude_kernel: 1; ++ __u64 exclude_hv: 1; ++ __u64 exclude_idle: 1; ++ __u64 mmap: 1; ++ __u64 comm: 1; ++ __u64 freq: 1; ++ __u64 inherit_stat: 1; ++ __u64 enable_on_exec: 1; ++ __u64 task: 1; ++ __u64 watermark: 1; ++ __u64 precise_ip: 2; ++ __u64 mmap_data: 1; ++ __u64 sample_id_all: 1; ++ __u64 exclude_host: 1; ++ __u64 exclude_guest: 1; ++ __u64 exclude_callchain_kernel: 1; ++ __u64 exclude_callchain_user: 1; ++ __u64 mmap2: 1; ++ __u64 comm_exec: 1; ++ __u64 use_clockid: 1; ++ __u64 context_switch: 1; ++ __u64 write_backward: 1; ++ __u64 namespaces: 1; ++ __u64 ksymbol: 1; ++ __u64 bpf_event: 1; ++ __u64 aux_output: 1; ++ __u64 cgroup: 1; ++ __u64 text_poke: 1; ++ __u64 build_id: 1; ++ __u64 inherit_thread: 1; ++ __u64 remove_on_exec: 1; ++ __u64 sigtrap: 1; ++ __u64 __reserved_1: 26; ++ union { ++ __u32 wakeup_events; ++ __u32 wakeup_watermark; ++ }; ++ __u32 bp_type; ++ union { ++ __u64 bp_addr; ++ __u64 kprobe_func; ++ __u64 uprobe_path; ++ __u64 config1; ++ }; ++ union { ++ __u64 bp_len; ++ __u64 kprobe_addr; ++ __u64 probe_offset; ++ __u64 config2; ++ }; ++ __u64 branch_sample_type; ++ __u64 sample_regs_user; ++ __u32 sample_stack_user; ++ __s32 clockid; ++ __u64 sample_regs_intr; ++ __u32 aux_watermark; ++ __u16 sample_max_stack; ++ __u16 __reserved_2; ++ __u32 aux_sample_size; ++ __u32 __reserved_3; ++ __u64 sig_data; ++ __u64 config3; ++}; ++ ++union perf_mem_data_src { ++ __u64 val; ++ struct { ++ __u64 mem_rsvd: 18; ++ __u64 mem_hops: 3; ++ __u64 mem_blk: 3; ++ __u64 mem_snoopx: 2; ++ __u64 mem_remote: 1; ++ __u64 mem_lvl_num: 4; ++ __u64 mem_dtlb: 7; ++ __u64 mem_lock: 2; ++ __u64 mem_snoop: 5; ++ __u64 mem_lvl: 14; ++ __u64 mem_op: 5; ++ }; ++}; ++ ++struct perf_branch_entry { ++ __u64 from; ++ __u64 to; ++ __u64 mispred: 1; ++ __u64 predicted: 1; ++ __u64 in_tx: 1; ++ __u64 abort: 1; ++ __u64 cycles: 16; ++ __u64 type: 4; ++ __u64 spec: 2; ++ __u64 new_type: 4; ++ __u64 priv: 3; ++ __u64 reserved: 31; ++}; ++ ++union perf_sample_weight { ++ __u64 full; ++ struct { ++ __u16 var3_w; ++ __u16 var2_w; ++ __u32 var1_dw; ++ }; ++}; ++ ++struct ftrace_regs { ++ struct pt_regs regs; ++}; ++ ++struct ftrace_ops; ++ ++typedef void (*ftrace_func_t)(long unsigned int, long unsigned int, struct ftrace_ops *, struct ftrace_regs *); ++ ++struct ftrace_hash; ++ ++struct ftrace_ops_hash { ++ struct ftrace_hash *notrace_hash; ++ struct ftrace_hash *filter_hash; ++ struct mutex regex_lock; ++}; ++ ++enum ftrace_ops_cmd { ++ FTRACE_OPS_CMD_ENABLE_SHARE_IPMODIFY_SELF = 0, ++ FTRACE_OPS_CMD_ENABLE_SHARE_IPMODIFY_PEER = 1, ++ FTRACE_OPS_CMD_DISABLE_SHARE_IPMODIFY_PEER = 2, ++}; ++ ++typedef int (*ftrace_ops_func_t)(struct ftrace_ops *, enum ftrace_ops_cmd); ++ ++struct ftrace_ops { ++ ftrace_func_t func; ++ struct ftrace_ops *next; ++ long unsigned int flags; ++ void *private; ++ ftrace_func_t saved_func; ++ struct ftrace_ops_hash local_hash; ++ struct ftrace_ops_hash *func_hash; ++ struct ftrace_ops_hash old_hash; ++ long unsigned int trampoline; ++ long unsigned int trampoline_size; ++ struct list_head list; ++ ftrace_ops_func_t ops_func; ++ long unsigned int direct_call; ++}; ++ ++struct irq_work { ++ struct __call_single_node node; ++ void (*func)(struct irq_work *); ++ struct rcuwait irqwait; ++}; ++ ++struct perf_regs { ++ __u64 abi; ++ struct pt_regs *regs; ++}; ++ ++struct perf_callchain_entry { ++ __u64 nr; ++ __u64 ip[0]; ++}; ++ ++typedef long unsigned int (*perf_copy_f)(void *, const void *, long unsigned int, long unsigned int); ++ ++struct perf_raw_frag { ++ union { ++ struct perf_raw_frag *next; ++ long unsigned int pad; ++ }; ++ perf_copy_f copy; ++ void *data; ++ u32 size; ++} __attribute__((packed)); ++ ++struct perf_raw_record { ++ struct perf_raw_frag frag; ++ u32 size; ++}; ++ ++struct perf_branch_stack { ++ __u64 nr; ++ __u64 hw_idx; ++ struct perf_branch_entry entries[0]; ++}; ++ ++struct hw_perf_event_extra { ++ u64 config; ++ unsigned int reg; ++ int alloc; ++ int idx; ++}; ++ ++struct hw_perf_event { ++ union { ++ struct { ++ u64 config; ++ u64 last_tag; ++ long unsigned int config_base; ++ long unsigned int event_base; ++ int event_base_rdpmc; ++ int idx; ++ int last_cpu; ++ int flags; ++ struct hw_perf_event_extra extra_reg; ++ struct hw_perf_event_extra branch_reg; ++ }; ++ struct { ++ struct hrtimer hrtimer; ++ }; ++ struct { ++ struct list_head tp_list; ++ }; ++ struct { ++ u64 pwr_acc; ++ u64 ptsc; ++ }; ++ struct { ++ u8 iommu_bank; ++ u8 iommu_cntr; ++ u16 padding; ++ u64 conf; ++ u64 conf1; ++ }; ++ }; ++ struct task_struct *target; ++ void *addr_filters; ++ long unsigned int addr_filters_gen; ++ int state; ++ local64_t prev_count; ++ u64 sample_period; ++ union { ++ struct { ++ u64 last_period; ++ local64_t period_left; ++ }; ++ struct { ++ u64 saved_metric; ++ u64 saved_slots; ++ }; ++ }; ++ u64 interrupts_seq; ++ u64 interrupts; ++ u64 freq_time_stamp; ++ u64 freq_count_stamp; ++}; ++ ++struct perf_cpu_pmu_context; ++ ++struct perf_event_pmu_context; ++ ++struct perf_output_handle; ++ ++struct pmu { ++ struct list_head entry; ++ struct module *module; ++ struct device *dev; ++ const struct attribute_group **attr_groups; ++ const struct attribute_group **attr_update; ++ const char *name; ++ int type; ++ int capabilities; ++ int *pmu_disable_count; ++ struct perf_cpu_pmu_context *cpu_pmu_context; ++ atomic_t exclusive_cnt; ++ int task_ctx_nr; ++ int hrtimer_interval_ms; ++ unsigned int nr_addr_filters; ++ void (*pmu_enable)(struct pmu *); ++ void (*pmu_disable)(struct pmu *); ++ int (*event_init)(struct perf_event *); ++ void (*event_mapped)(struct perf_event *, struct mm_struct *); ++ void (*event_unmapped)(struct perf_event *, struct mm_struct *); ++ int (*add)(struct perf_event *, int); ++ void (*del)(struct perf_event *, int); ++ void (*start)(struct perf_event *, int); ++ void (*stop)(struct perf_event *, int); ++ void (*read)(struct perf_event *); ++ void (*start_txn)(struct pmu *, unsigned int); ++ int (*commit_txn)(struct pmu *); ++ void (*cancel_txn)(struct pmu *); ++ int (*event_idx)(struct perf_event *); ++ void (*sched_task)(struct perf_event_pmu_context *, bool); ++ struct kmem_cache *task_ctx_cache; ++ void (*swap_task_ctx)(struct perf_event_pmu_context *, struct perf_event_pmu_context *); ++ void * (*setup_aux)(struct perf_event *, void **, int, bool); ++ void (*free_aux)(void *); ++ long int (*snapshot_aux)(struct perf_event *, struct perf_output_handle *, long unsigned int); ++ int (*addr_filters_validate)(struct list_head *); ++ void (*addr_filters_sync)(struct perf_event *); ++ int (*aux_output_match)(struct perf_event *); ++ bool (*filter)(struct pmu *, int); ++ int (*check_period)(struct perf_event *, u64); ++}; ++ ++struct perf_event_pmu_context { ++ struct pmu *pmu; ++ struct perf_event_context *ctx; ++ struct list_head pmu_ctx_entry; ++ struct list_head pinned_active; ++ struct list_head flexible_active; ++ unsigned int embedded: 1; ++ unsigned int nr_events; ++ atomic_t refcount; ++ struct callback_head callback_head; ++ void *task_ctx_data; ++ int rotate_necessary; ++}; ++ ++struct perf_cpu_pmu_context { ++ struct perf_event_pmu_context epc; ++ struct perf_event_pmu_context *task_epc; ++ struct list_head sched_cb_entry; ++ int sched_cb_usage; ++ int active_oncpu; ++ int exclusive; ++ raw_spinlock_t hrtimer_lock; ++ struct hrtimer hrtimer; ++ ktime_t hrtimer_interval; ++ unsigned int hrtimer_active; ++}; ++ ++enum perf_event_state { ++ PERF_EVENT_STATE_DEAD = -4, ++ PERF_EVENT_STATE_EXIT = -3, ++ PERF_EVENT_STATE_ERROR = -2, ++ PERF_EVENT_STATE_OFF = -1, ++ PERF_EVENT_STATE_INACTIVE = 0, ++ PERF_EVENT_STATE_ACTIVE = 1, ++}; ++ ++struct perf_addr_filters_head { ++ struct list_head list; ++ raw_spinlock_t lock; ++ unsigned int nr_file_filters; ++}; ++ ++struct perf_sample_data; ++ ++typedef void (*perf_overflow_handler_t)(struct perf_event *, struct perf_sample_data *, struct pt_regs *); ++ ++struct perf_buffer; ++ ++struct perf_addr_filter_range; ++ ++struct perf_cgroup; ++ ++struct perf_event { ++ struct list_head event_entry; ++ struct list_head sibling_list; ++ struct list_head active_list; ++ struct rb_node group_node; ++ u64 group_index; ++ struct list_head migrate_entry; ++ struct hlist_node hlist_entry; ++ struct list_head active_entry; ++ int nr_siblings; ++ int event_caps; ++ int group_caps; ++ struct perf_event *group_leader; ++ struct pmu *pmu; ++ void *pmu_private; ++ enum perf_event_state state; ++ unsigned int attach_state; ++ local64_t count; ++ atomic64_t child_count; ++ u64 total_time_enabled; ++ u64 total_time_running; ++ u64 tstamp; ++ struct perf_event_attr attr; ++ u16 header_size; ++ u16 id_header_size; ++ u16 read_size; ++ struct hw_perf_event hw; ++ struct perf_event_context *ctx; ++ struct perf_event_pmu_context *pmu_ctx; ++ atomic_long_t refcount; ++ atomic64_t child_total_time_enabled; ++ atomic64_t child_total_time_running; ++ struct mutex child_mutex; ++ struct list_head child_list; ++ struct perf_event *parent; ++ int oncpu; ++ int cpu; ++ struct list_head owner_entry; ++ struct task_struct *owner; ++ struct mutex mmap_mutex; ++ atomic_t mmap_count; ++ struct perf_buffer *rb; ++ struct list_head rb_entry; ++ long unsigned int rcu_batches; ++ int rcu_pending; ++ wait_queue_head_t waitq; ++ struct fasync_struct *fasync; ++ unsigned int pending_wakeup; ++ unsigned int pending_kill; ++ unsigned int pending_disable; ++ unsigned int pending_sigtrap; ++ long unsigned int pending_addr; ++ struct irq_work pending_irq; ++ struct callback_head pending_task; ++ unsigned int pending_work; ++ atomic_t event_limit; ++ struct perf_addr_filters_head addr_filters; ++ struct perf_addr_filter_range *addr_filter_ranges; ++ long unsigned int addr_filters_gen; ++ struct perf_event *aux_event; ++ void (*destroy)(struct perf_event *); ++ struct callback_head callback_head; ++ struct pid_namespace *ns; ++ u64 id; ++ atomic64_t lost_samples; ++ u64 (*clock)(); ++ perf_overflow_handler_t overflow_handler; ++ void *overflow_handler_context; ++ perf_overflow_handler_t orig_overflow_handler; ++ struct bpf_prog *prog; ++ u64 bpf_cookie; ++ struct trace_event_call *tp_event; ++ struct event_filter *filter; ++ struct ftrace_ops ftrace_ops; ++ struct perf_cgroup *cgrp; ++ void *security; ++ struct list_head sb_list; ++}; ++ ++struct perf_output_handle { ++ struct perf_event *event; ++ struct perf_buffer *rb; ++ long unsigned int wakeup; ++ long unsigned int size; ++ u64 aux_flags; ++ union { ++ void *addr; ++ long unsigned int head; ++ }; ++ int page; ++}; ++ ++struct perf_addr_filter_range { ++ long unsigned int start; ++ long unsigned int size; ++}; ++ ++struct perf_sample_data { ++ u64 sample_flags; ++ u64 period; ++ u64 dyn_size; ++ u64 type; ++ struct { ++ u32 pid; ++ u32 tid; ++ } tid_entry; ++ u64 time; ++ u64 id; ++ struct { ++ u32 cpu; ++ u32 reserved; ++ } cpu_entry; ++ u64 ip; ++ struct perf_callchain_entry *callchain; ++ struct perf_raw_record *raw; ++ struct perf_branch_stack *br_stack; ++ union perf_sample_weight weight; ++ union perf_mem_data_src data_src; ++ u64 txn; ++ struct perf_regs regs_user; ++ struct perf_regs regs_intr; ++ u64 stack_user_size; ++ u64 stream_id; ++ u64 cgroup; ++ u64 addr; ++ u64 phys_addr; ++ u64 data_page_size; ++ u64 code_page_size; ++ u64 aux_size; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct perf_cgroup_info; ++ ++struct perf_cgroup { ++ struct cgroup_subsys_state css; ++ struct perf_cgroup_info *info; ++}; ++ ++struct perf_cgroup_info { ++ u64 time; ++ u64 timestamp; ++ u64 timeoffset; ++ int active; ++}; ++ ++struct trace_entry { ++ short unsigned int type; ++ unsigned char flags; ++ unsigned char preempt_count; ++ int pid; ++}; ++ ++struct trace_array; ++ ++struct tracer; ++ ++struct array_buffer; ++ ++struct ring_buffer_iter; ++ ++struct trace_iterator { ++ struct trace_array *tr; ++ struct tracer *trace; ++ struct array_buffer *array_buffer; ++ void *private; ++ int cpu_file; ++ struct mutex mutex; ++ struct ring_buffer_iter **buffer_iter; ++ long unsigned int iter_flags; ++ void *temp; ++ unsigned int temp_size; ++ char *fmt; ++ unsigned int fmt_size; ++ long int wait_index; ++ struct trace_seq tmp_seq; ++ cpumask_var_t started; ++ bool snapshot; ++ struct trace_seq seq; ++ struct trace_entry *ent; ++ long unsigned int lost_events; ++ int leftover; ++ int ent_size; ++ int cpu; ++ u64 ts; ++ loff_t pos; ++ long int idx; ++}; ++ ++enum print_line_t { ++ TRACE_TYPE_PARTIAL_LINE = 0, ++ TRACE_TYPE_HANDLED = 1, ++ TRACE_TYPE_UNHANDLED = 2, ++ TRACE_TYPE_NO_CONSUME = 3, ++}; ++ ++typedef enum print_line_t (*trace_print_func)(struct trace_iterator *, int, struct trace_event *); ++ ++struct trace_event_functions { ++ trace_print_func trace; ++ trace_print_func raw; ++ trace_print_func hex; ++ trace_print_func binary; ++}; ++ ++enum trace_reg { ++ TRACE_REG_REGISTER = 0, ++ TRACE_REG_UNREGISTER = 1, ++ TRACE_REG_PERF_REGISTER = 2, ++ TRACE_REG_PERF_UNREGISTER = 3, ++ TRACE_REG_PERF_OPEN = 4, ++ TRACE_REG_PERF_CLOSE = 5, ++ TRACE_REG_PERF_ADD = 6, ++ TRACE_REG_PERF_DEL = 7, ++}; ++ ++struct trace_event_fields { ++ const char *type; ++ union { ++ struct { ++ const char *name; ++ const int size; ++ const int align; ++ const int is_signed; ++ const int filter_type; ++ const int len; ++ }; ++ int (*define_fields)(struct trace_event_call *); ++ }; ++}; ++ ++struct trace_event_class { ++ const char *system; ++ void *probe; ++ void *perf_probe; ++ int (*reg)(struct trace_event_call *, enum trace_reg, void *); ++ struct trace_event_fields *fields_array; ++ struct list_head * (*get_fields)(struct trace_event_call *); ++ struct list_head fields; ++ int (*raw_init)(struct trace_event_call *); ++}; ++ ++enum { ++ TRACE_EVENT_FL_FILTERED_BIT = 0, ++ TRACE_EVENT_FL_CAP_ANY_BIT = 1, ++ TRACE_EVENT_FL_NO_SET_FILTER_BIT = 2, ++ TRACE_EVENT_FL_IGNORE_ENABLE_BIT = 3, ++ TRACE_EVENT_FL_TRACEPOINT_BIT = 4, ++ TRACE_EVENT_FL_DYNAMIC_BIT = 5, ++ TRACE_EVENT_FL_KPROBE_BIT = 6, ++ TRACE_EVENT_FL_UPROBE_BIT = 7, ++ TRACE_EVENT_FL_EPROBE_BIT = 8, ++ TRACE_EVENT_FL_CUSTOM_BIT = 9, ++}; ++ ++enum { ++ EVENT_FILE_FL_ENABLED_BIT = 0, ++ EVENT_FILE_FL_RECORDED_CMD_BIT = 1, ++ EVENT_FILE_FL_RECORDED_TGID_BIT = 2, ++ EVENT_FILE_FL_FILTERED_BIT = 3, ++ EVENT_FILE_FL_NO_SET_FILTER_BIT = 4, ++ EVENT_FILE_FL_SOFT_MODE_BIT = 5, ++ EVENT_FILE_FL_SOFT_DISABLED_BIT = 6, ++ EVENT_FILE_FL_TRIGGER_MODE_BIT = 7, ++ EVENT_FILE_FL_TRIGGER_COND_BIT = 8, ++ EVENT_FILE_FL_PID_FILTER_BIT = 9, ++ EVENT_FILE_FL_WAS_ENABLED_BIT = 10, ++}; ++ ++struct uuidcmp { ++ const char *uuid; ++ int len; ++}; ++ ++typedef int (*initcall_t)(); ++ ++struct subprocess_info { ++ struct work_struct work; ++ struct completion *complete; ++ const char *path; ++ char **argv; ++ char **envp; ++ int wait; ++ int retval; ++ int (*init)(struct subprocess_info *, struct cred *); ++ void (*cleanup)(struct subprocess_info *); ++ void *data; ++}; ++ ++typedef phys_addr_t resource_size_t; ++ ++struct resource { ++ resource_size_t start; ++ resource_size_t end; ++ const char *name; ++ long unsigned int flags; ++ long unsigned int desc; ++ struct resource *parent; ++ struct resource *sibling; ++ struct resource *child; ++}; ++ ++enum umh_disable_depth { ++ UMH_ENABLED = 0, ++ UMH_FREEZING = 1, ++ UMH_DISABLED = 2, ++}; ++ ++enum kmalloc_cache_type { ++ KMALLOC_NORMAL = 0, ++ KMALLOC_RECLAIM = 1, ++ KMALLOC_DMA = 2, ++ KMALLOC_CGROUP = 3, ++ NR_KMALLOC_TYPES = 4, ++}; ++ ++typedef u64 async_cookie_t; ++ ++typedef void (*async_func_t)(void *, async_cookie_t); ++ ++struct async_domain { ++ struct list_head pending; ++ unsigned int registered: 1; ++}; ++ ++struct hash { ++ int ino; ++ int minor; ++ int major; ++ umode_t mode; ++ struct hash *next; ++ char name[4098]; ++}; ++ ++enum state { ++ Start = 0, ++ Collect = 1, ++ GotHeader = 2, ++ SkipIt = 3, ++ GotName = 4, ++ CopyFile = 5, ++ GotSymlink = 6, ++ Reset = 7, ++}; ++ ++typedef int (*decompress_fn)(unsigned char *, long int, long int (*)(void *, long unsigned int), long int (*)(void *, long unsigned int), unsigned char *, long int *, void (*)(char *)); ++ ++enum key_being_used_for { ++ VERIFYING_MODULE_SIGNATURE = 0, ++ VERIFYING_FIRMWARE_SIGNATURE = 1, ++ VERIFYING_KEXEC_PE_SIGNATURE = 2, ++ VERIFYING_KEY_SIGNATURE = 3, ++ VERIFYING_KEY_SELF_SIGNATURE = 4, ++ VERIFYING_UNSPECIFIED_SIGNATURE = 5, ++ NR__KEY_BEING_USED_FOR = 6, ++}; ++ ++enum pcpu_fc { ++ PCPU_FC_AUTO = 0, ++ PCPU_FC_EMBED = 1, ++ PCPU_FC_PAGE = 2, ++ PCPU_FC_NR = 3, ++}; ++ ++enum hrtimer_base_type { ++ HRTIMER_BASE_MONOTONIC = 0, ++ HRTIMER_BASE_REALTIME = 1, ++ HRTIMER_BASE_BOOTTIME = 2, ++ HRTIMER_BASE_TAI = 3, ++ HRTIMER_BASE_MONOTONIC_SOFT = 4, ++ HRTIMER_BASE_REALTIME_SOFT = 5, ++ HRTIMER_BASE_BOOTTIME_SOFT = 6, ++ HRTIMER_BASE_TAI_SOFT = 7, ++ HRTIMER_MAX_CLOCK_BASES = 8, ++}; ++ ++enum node_states { ++ N_POSSIBLE = 0, ++ N_ONLINE = 1, ++ N_NORMAL_MEMORY = 2, ++ N_HIGH_MEMORY = 2, ++ N_MEMORY = 3, ++ N_CPU = 4, ++ N_GENERIC_INITIATOR = 5, ++ NR_NODE_STATES = 6, ++}; ++ ++enum { ++ MM_FILEPAGES = 0, ++ MM_ANONPAGES = 1, ++ MM_SWAPENTS = 2, ++ MM_SHMEMPAGES = 3, ++ NR_MM_COUNTERS = 4, ++}; ++ ++enum rseq_cs_flags_bit { ++ RSEQ_CS_FLAG_NO_RESTART_ON_PREEMPT_BIT = 0, ++ RSEQ_CS_FLAG_NO_RESTART_ON_SIGNAL_BIT = 1, ++ RSEQ_CS_FLAG_NO_RESTART_ON_MIGRATE_BIT = 2, ++}; ++ ++enum { ++ TASK_COMM_LEN = 16, ++}; ++ ++enum rseq_event_mask_bits { ++ RSEQ_EVENT_PREEMPT_BIT = 0, ++ RSEQ_EVENT_SIGNAL_BIT = 1, ++ RSEQ_EVENT_MIGRATE_BIT = 2, ++}; ++ ++enum interruption_class { ++ IRQEXT_CLK = 0, ++ IRQEXT_EXC = 1, ++ IRQEXT_EMS = 2, ++ IRQEXT_TMR = 3, ++ IRQEXT_TLA = 4, ++ IRQEXT_PFL = 5, ++ IRQEXT_DSD = 6, ++ IRQEXT_VRT = 7, ++ IRQEXT_SCP = 8, ++ IRQEXT_IUC = 9, ++ IRQEXT_CMS = 10, ++ IRQEXT_CMC = 11, ++ IRQEXT_FTP = 12, ++ IRQIO_CIO = 13, ++ IRQIO_DAS = 14, ++ IRQIO_C15 = 15, ++ IRQIO_C70 = 16, ++ IRQIO_TAP = 17, ++ IRQIO_VMR = 18, ++ IRQIO_LCS = 19, ++ IRQIO_CTC = 20, ++ IRQIO_ADM = 21, ++ IRQIO_CSC = 22, ++ IRQIO_VIR = 23, ++ IRQIO_QAI = 24, ++ IRQIO_APB = 25, ++ IRQIO_PCF = 26, ++ IRQIO_PCD = 27, ++ IRQIO_MSI = 28, ++ IRQIO_VAI = 29, ++ IRQIO_GAL = 30, ++ NMI_NMI = 31, ++ CPU_RST = 32, ++ NR_ARCH_IRQS = 33, ++}; ++ ++enum { ++ HI_SOFTIRQ = 0, ++ TIMER_SOFTIRQ = 1, ++ NET_TX_SOFTIRQ = 2, ++ NET_RX_SOFTIRQ = 3, ++ BLOCK_SOFTIRQ = 4, ++ IRQ_POLL_SOFTIRQ = 5, ++ TASKLET_SOFTIRQ = 6, ++ SCHED_SOFTIRQ = 7, ++ HRTIMER_SOFTIRQ = 8, ++ RCU_SOFTIRQ = 9, ++ NR_SOFTIRQS = 10, ++}; ++ ++enum migratetype { ++ MIGRATE_UNMOVABLE = 0, ++ MIGRATE_MOVABLE = 1, ++ MIGRATE_RECLAIMABLE = 2, ++ MIGRATE_PCPTYPES = 3, ++ MIGRATE_HIGHATOMIC = 3, ++ MIGRATE_CMA = 4, ++ MIGRATE_ISOLATE = 5, ++ MIGRATE_TYPES = 6, ++}; ++ ++enum numa_stat_item { ++ NUMA_HIT = 0, ++ NUMA_MISS = 1, ++ NUMA_FOREIGN = 2, ++ NUMA_INTERLEAVE_HIT = 3, ++ NUMA_LOCAL = 4, ++ NUMA_OTHER = 5, ++ NR_VM_NUMA_EVENT_ITEMS = 6, ++}; ++ ++enum zone_stat_item { ++ NR_FREE_PAGES = 0, ++ NR_ZONE_LRU_BASE = 1, ++ NR_ZONE_INACTIVE_ANON = 1, ++ NR_ZONE_ACTIVE_ANON = 2, ++ NR_ZONE_INACTIVE_FILE = 3, ++ NR_ZONE_ACTIVE_FILE = 4, ++ NR_ZONE_UNEVICTABLE = 5, ++ NR_ZONE_WRITE_PENDING = 6, ++ NR_MLOCK = 7, ++ NR_BOUNCE = 8, ++ NR_ZSPAGES = 9, ++ NR_FREE_CMA_PAGES = 10, ++ NR_VM_ZONE_STAT_ITEMS = 11, ++}; ++ ++enum node_stat_item { ++ NR_LRU_BASE = 0, ++ NR_INACTIVE_ANON = 0, ++ NR_ACTIVE_ANON = 1, ++ NR_INACTIVE_FILE = 2, ++ NR_ACTIVE_FILE = 3, ++ NR_UNEVICTABLE = 4, ++ NR_SLAB_RECLAIMABLE_B = 5, ++ NR_SLAB_UNRECLAIMABLE_B = 6, ++ NR_ISOLATED_ANON = 7, ++ NR_ISOLATED_FILE = 8, ++ WORKINGSET_NODES = 9, ++ WORKINGSET_REFAULT_BASE = 10, ++ WORKINGSET_REFAULT_ANON = 10, ++ WORKINGSET_REFAULT_FILE = 11, ++ WORKINGSET_ACTIVATE_BASE = 12, ++ WORKINGSET_ACTIVATE_ANON = 12, ++ WORKINGSET_ACTIVATE_FILE = 13, ++ WORKINGSET_RESTORE_BASE = 14, ++ WORKINGSET_RESTORE_ANON = 14, ++ WORKINGSET_RESTORE_FILE = 15, ++ WORKINGSET_NODERECLAIM = 16, ++ NR_ANON_MAPPED = 17, ++ NR_FILE_MAPPED = 18, ++ NR_FILE_PAGES = 19, ++ NR_FILE_DIRTY = 20, ++ NR_WRITEBACK = 21, ++ NR_WRITEBACK_TEMP = 22, ++ NR_SHMEM = 23, ++ NR_SHMEM_THPS = 24, ++ NR_SHMEM_PMDMAPPED = 25, ++ NR_FILE_THPS = 26, ++ NR_FILE_PMDMAPPED = 27, ++ NR_ANON_THPS = 28, ++ NR_VMSCAN_WRITE = 29, ++ NR_VMSCAN_IMMEDIATE = 30, ++ NR_DIRTIED = 31, ++ NR_WRITTEN = 32, ++ NR_THROTTLED_WRITTEN = 33, ++ NR_KERNEL_MISC_RECLAIMABLE = 34, ++ NR_FOLL_PIN_ACQUIRED = 35, ++ NR_FOLL_PIN_RELEASED = 36, ++ NR_KERNEL_STACK_KB = 37, ++ NR_PAGETABLE = 38, ++ NR_SECONDARY_PAGETABLE = 39, ++ NR_SWAPCACHE = 40, ++ PGPROMOTE_SUCCESS = 41, ++ PGPROMOTE_CANDIDATE = 42, ++ NR_VM_NODE_STAT_ITEMS = 43, ++}; ++ ++enum lru_list { ++ LRU_INACTIVE_ANON = 0, ++ LRU_ACTIVE_ANON = 1, ++ LRU_INACTIVE_FILE = 2, ++ LRU_ACTIVE_FILE = 3, ++ LRU_UNEVICTABLE = 4, ++ NR_LRU_LISTS = 5, ++}; ++ ++enum vmscan_throttle_state { ++ VMSCAN_THROTTLE_WRITEBACK = 0, ++ VMSCAN_THROTTLE_ISOLATED = 1, ++ VMSCAN_THROTTLE_NOPROGRESS = 2, ++ VMSCAN_THROTTLE_CONGESTED = 3, ++ NR_VMSCAN_THROTTLE = 4, ++}; ++ ++enum { ++ MM_LEAF_TOTAL = 0, ++ MM_LEAF_OLD = 1, ++ MM_LEAF_YOUNG = 2, ++ MM_NONLEAF_TOTAL = 3, ++ MM_NONLEAF_FOUND = 4, ++ MM_NONLEAF_ADDED = 5, ++ NR_MM_STATS = 6, ++}; ++ ++enum zone_watermarks { ++ WMARK_MIN = 0, ++ WMARK_LOW = 1, ++ WMARK_HIGH = 2, ++ WMARK_PROMO = 3, ++ NR_WMARK = 4, ++}; ++ ++enum { ++ ZONELIST_FALLBACK = 0, ++ ZONELIST_NOFALLBACK = 1, ++ MAX_ZONELISTS = 2, ++}; ++ ++enum { ++ PG_DIRECT_MAP_4K = 0, ++ PG_DIRECT_MAP_1M = 1, ++ PG_DIRECT_MAP_2G = 2, ++ PG_DIRECT_MAP_MAX = 3, ++}; ++ ++enum { ++ DQF_ROOT_SQUASH_B = 0, ++ DQF_SYS_FILE_B = 16, ++ DQF_PRIVATE = 17, ++}; ++ ++enum { ++ DQST_LOOKUPS = 0, ++ DQST_DROPS = 1, ++ DQST_READS = 2, ++ DQST_WRITES = 3, ++ DQST_CACHE_HITS = 4, ++ DQST_ALLOC_DQUOTS = 5, ++ DQST_FREE_DQUOTS = 6, ++ DQST_SYNCS = 7, ++ _DQST_DQSTAT_LAST = 8, ++}; ++ ++enum { ++ SB_UNFROZEN = 0, ++ SB_FREEZE_WRITE = 1, ++ SB_FREEZE_PAGEFAULT = 2, ++ SB_FREEZE_FS = 3, ++ SB_FREEZE_COMPLETE = 4, ++}; ++ ++enum compound_dtor_id { ++ NULL_COMPOUND_DTOR = 0, ++ COMPOUND_PAGE_DTOR = 1, ++ HUGETLB_PAGE_DTOR = 2, ++ NR_COMPOUND_DTORS = 3, ++}; ++ ++enum vm_event_item { ++ PGPGIN = 0, ++ PGPGOUT = 1, ++ PSWPIN = 2, ++ PSWPOUT = 3, ++ PGALLOC_DMA = 4, ++ PGALLOC_NORMAL = 5, ++ PGALLOC_MOVABLE = 6, ++ ALLOCSTALL_DMA = 7, ++ ALLOCSTALL_NORMAL = 8, ++ ALLOCSTALL_MOVABLE = 9, ++ PGSCAN_SKIP_DMA = 10, ++ PGSCAN_SKIP_NORMAL = 11, ++ PGSCAN_SKIP_MOVABLE = 12, ++ PGFREE = 13, ++ PGACTIVATE = 14, ++ PGDEACTIVATE = 15, ++ PGLAZYFREE = 16, ++ PGFAULT = 17, ++ PGMAJFAULT = 18, ++ PGLAZYFREED = 19, ++ PGREFILL = 20, ++ PGREUSE = 21, ++ PGSTEAL_KSWAPD = 22, ++ PGSTEAL_DIRECT = 23, ++ PGSTEAL_KHUGEPAGED = 24, ++ PGDEMOTE_KSWAPD = 25, ++ PGDEMOTE_DIRECT = 26, ++ PGDEMOTE_KHUGEPAGED = 27, ++ PGSCAN_KSWAPD = 28, ++ PGSCAN_DIRECT = 29, ++ PGSCAN_KHUGEPAGED = 30, ++ PGSCAN_DIRECT_THROTTLE = 31, ++ PGSCAN_ANON = 32, ++ PGSCAN_FILE = 33, ++ PGSTEAL_ANON = 34, ++ PGSTEAL_FILE = 35, ++ PGSCAN_ZONE_RECLAIM_FAILED = 36, ++ PGINODESTEAL = 37, ++ SLABS_SCANNED = 38, ++ KSWAPD_INODESTEAL = 39, ++ KSWAPD_LOW_WMARK_HIT_QUICKLY = 40, ++ KSWAPD_HIGH_WMARK_HIT_QUICKLY = 41, ++ PAGEOUTRUN = 42, ++ PGROTATED = 43, ++ DROP_PAGECACHE = 44, ++ DROP_SLAB = 45, ++ OOM_KILL = 46, ++ NUMA_PTE_UPDATES = 47, ++ NUMA_HUGE_PTE_UPDATES = 48, ++ NUMA_HINT_FAULTS = 49, ++ NUMA_HINT_FAULTS_LOCAL = 50, ++ NUMA_PAGE_MIGRATE = 51, ++ PGMIGRATE_SUCCESS = 52, ++ PGMIGRATE_FAIL = 53, ++ THP_MIGRATION_SUCCESS = 54, ++ THP_MIGRATION_FAIL = 55, ++ THP_MIGRATION_SPLIT = 56, ++ COMPACTMIGRATE_SCANNED = 57, ++ COMPACTFREE_SCANNED = 58, ++ COMPACTISOLATED = 59, ++ COMPACTSTALL = 60, ++ COMPACTFAIL = 61, ++ COMPACTSUCCESS = 62, ++ KCOMPACTD_WAKE = 63, ++ KCOMPACTD_MIGRATE_SCANNED = 64, ++ KCOMPACTD_FREE_SCANNED = 65, ++ HTLB_BUDDY_PGALLOC = 66, ++ HTLB_BUDDY_PGALLOC_FAIL = 67, ++ CMA_ALLOC_SUCCESS = 68, ++ CMA_ALLOC_FAIL = 69, ++ UNEVICTABLE_PGCULLED = 70, ++ UNEVICTABLE_PGSCANNED = 71, ++ UNEVICTABLE_PGRESCUED = 72, ++ UNEVICTABLE_PGMLOCKED = 73, ++ UNEVICTABLE_PGMUNLOCKED = 74, ++ UNEVICTABLE_PGCLEARED = 75, ++ UNEVICTABLE_PGSTRANDED = 76, ++ BALLOON_INFLATE = 77, ++ BALLOON_DEFLATE = 78, ++ BALLOON_MIGRATE = 79, ++ SWAP_RA = 80, ++ SWAP_RA_HIT = 81, ++ KSM_SWPIN_COPY = 82, ++ COW_KSM = 83, ++ ZSWPIN = 84, ++ ZSWPOUT = 85, ++ NR_VM_EVENT_ITEMS = 86, ++}; ++ ++enum mod_mem_type { ++ MOD_TEXT = 0, ++ MOD_DATA = 1, ++ MOD_RODATA = 2, ++ MOD_RO_AFTER_INIT = 3, ++ MOD_INIT_TEXT = 4, ++ MOD_INIT_DATA = 5, ++ MOD_INIT_RODATA = 6, ++ MOD_MEM_NUM_TYPES = 7, ++ MOD_INVALID = -1, ++}; ++ ++enum ucount_type { ++ UCOUNT_USER_NAMESPACES = 0, ++ UCOUNT_PID_NAMESPACES = 1, ++ UCOUNT_UTS_NAMESPACES = 2, ++ UCOUNT_IPC_NAMESPACES = 3, ++ UCOUNT_NET_NAMESPACES = 4, ++ UCOUNT_MNT_NAMESPACES = 5, ++ UCOUNT_CGROUP_NAMESPACES = 6, ++ UCOUNT_TIME_NAMESPACES = 7, ++ UCOUNT_INOTIFY_INSTANCES = 8, ++ UCOUNT_INOTIFY_WATCHES = 9, ++ UCOUNT_FANOTIFY_GROUPS = 10, ++ UCOUNT_FANOTIFY_MARKS = 11, ++ UCOUNT_COUNTS = 12, ++}; ++ ++enum rlimit_type { ++ UCOUNT_RLIMIT_NPROC = 0, ++ UCOUNT_RLIMIT_MSGQUEUE = 1, ++ UCOUNT_RLIMIT_SIGPENDING = 2, ++ UCOUNT_RLIMIT_MEMLOCK = 3, ++ UCOUNT_RLIMIT_COUNTS = 4, ++}; ++ ++enum cpu_idle_type { ++ CPU_IDLE = 0, ++ CPU_NOT_IDLE = 1, ++ CPU_NEWLY_IDLE = 2, ++ CPU_MAX_IDLE_TYPES = 3, ++}; ++ ++enum { ++ __SD_BALANCE_NEWIDLE = 0, ++ __SD_BALANCE_EXEC = 1, ++ __SD_BALANCE_FORK = 2, ++ __SD_BALANCE_WAKE = 3, ++ __SD_WAKE_AFFINE = 4, ++ __SD_ASYM_CPUCAPACITY = 5, ++ __SD_ASYM_CPUCAPACITY_FULL = 6, ++ __SD_SHARE_CPUCAPACITY = 7, ++ __SD_SHARE_PKG_RESOURCES = 8, ++ __SD_SERIALIZE = 9, ++ __SD_ASYM_PACKING = 10, ++ __SD_PREFER_SIBLING = 11, ++ __SD_OVERLAP = 12, ++ __SD_NUMA = 13, ++ __SD_FLAG_CNT = 14, ++}; ++ ++enum audit_ntp_type { ++ AUDIT_NTP_OFFSET = 0, ++ AUDIT_NTP_FREQ = 1, ++ AUDIT_NTP_STATUS = 2, ++ AUDIT_NTP_TAI = 3, ++ AUDIT_NTP_TICK = 4, ++ AUDIT_NTP_ADJUST = 5, ++ AUDIT_NTP_NVALS = 6, ++}; ++ ++enum { ++ PSW_BITS_AS_PRIMARY = 0, ++ PSW_BITS_AS_ACCREG = 1, ++ PSW_BITS_AS_SECONDARY = 2, ++ PSW_BITS_AS_HOME = 3, ++}; ++ ++enum bug_trap_type { ++ BUG_TRAP_TYPE_NONE = 0, ++ BUG_TRAP_TYPE_WARN = 1, ++ BUG_TRAP_TYPE_BUG = 2, ++}; ++ ++union oac { ++ unsigned int val; ++ struct { ++ struct { ++ short unsigned int key: 4; ++ char: 4; ++ short unsigned int as: 2; ++ char: 4; ++ short unsigned int k: 1; ++ short unsigned int a: 1; ++ } oac1; ++ struct { ++ short unsigned int key: 4; ++ char: 4; ++ short unsigned int as: 2; ++ char: 4; ++ short unsigned int k: 1; ++ short unsigned int a: 1; ++ } oac2; ++ }; ++}; ++ ++enum die_val { ++ DIE_OOPS = 1, ++ DIE_BPT = 2, ++ DIE_SSTEP = 3, ++ DIE_PANIC = 4, ++ DIE_NMI = 5, ++ DIE_DIE = 6, ++ DIE_NMIWATCHDOG = 7, ++ DIE_KERNELDEBUG = 8, ++ DIE_TRAP = 9, ++ DIE_GPF = 10, ++ DIE_CALL = 11, ++ DIE_NMI_IPI = 12, ++}; ++ ++struct irqentry_state { ++ union { ++ bool exit_rcu; ++ bool lockdep; ++ }; ++}; ++ ++typedef struct irqentry_state irqentry_state_t; ++ ++struct vx_array { ++ __vector128 _[32]; ++}; ++ ++union tod_clock { ++ __int128 unsigned val; ++ struct { ++ __int128 unsigned ei: 8; ++ __int128 unsigned tod: 64; ++ int: 24; ++ short: 16; ++ __int128 unsigned pf: 16; ++ }; ++ struct { ++ __int128 unsigned eitod: 72; ++ }; ++ struct { ++ __int128 unsigned us: 60; ++ __int128 unsigned sus: 12; ++ }; ++}; ++ ++enum diag_stat_enum { ++ DIAG_STAT_X008 = 0, ++ DIAG_STAT_X00C = 1, ++ DIAG_STAT_X010 = 2, ++ DIAG_STAT_X014 = 3, ++ DIAG_STAT_X044 = 4, ++ DIAG_STAT_X064 = 5, ++ DIAG_STAT_X08C = 6, ++ DIAG_STAT_X09C = 7, ++ DIAG_STAT_X0DC = 8, ++ DIAG_STAT_X204 = 9, ++ DIAG_STAT_X210 = 10, ++ DIAG_STAT_X224 = 11, ++ DIAG_STAT_X250 = 12, ++ DIAG_STAT_X258 = 13, ++ DIAG_STAT_X26C = 14, ++ DIAG_STAT_X288 = 15, ++ DIAG_STAT_X2C4 = 16, ++ DIAG_STAT_X2FC = 17, ++ DIAG_STAT_X304 = 18, ++ DIAG_STAT_X308 = 19, ++ DIAG_STAT_X318 = 20, ++ DIAG_STAT_X500 = 21, ++ NR_DIAG_STAT = 22, ++}; ++ ++struct sysinfo_1_1_1 { ++ unsigned char p: 1; ++ char: 6; ++ unsigned char t: 1; ++ short: 0; ++ unsigned char ccr; ++ unsigned char cai; ++ char reserved_0[20]; ++ long unsigned int lic; ++ char manufacturer[16]; ++ char type[4]; ++ char reserved_1[12]; ++ char model_capacity[16]; ++ char sequence[16]; ++ char plant[4]; ++ char model[16]; ++ char model_perm_cap[16]; ++ char model_temp_cap[16]; ++ unsigned int model_cap_rating; ++ unsigned int model_perm_cap_rating; ++ unsigned int model_temp_cap_rating; ++ unsigned char typepct[5]; ++ unsigned char reserved_2[3]; ++ unsigned int ncr; ++ unsigned int npr; ++ unsigned int ntr; ++}; ++ ++struct sysinfo_3_2_2 { ++ char reserved_0[31]; ++ char: 4; ++ unsigned char count: 4; ++ struct { ++ char reserved_0[4]; ++ short unsigned int cpus_total; ++ short unsigned int cpus_configured; ++ short unsigned int cpus_standby; ++ short unsigned int cpus_reserved; ++ char name[8]; ++ unsigned int caf; ++ char cpi[16]; ++ char reserved_1[3]; ++ unsigned char evmne; ++ unsigned int reserved_2; ++ uuid_t uuid; ++ } vm[8]; ++ char reserved_3[1504]; ++ char ext_names[2048]; ++}; ++ ++typedef struct { ++ char _[8]; ++} addrtype; ++ ++typedef struct { ++ int _[16]; ++} acrstype; ++ ++struct plist_head { ++ struct list_head node_list; ++}; ++ ++enum pm_qos_type { ++ PM_QOS_UNITIALIZED = 0, ++ PM_QOS_MAX = 1, ++ PM_QOS_MIN = 2, ++}; ++ ++struct pm_qos_constraints { ++ struct plist_head list; ++ s32 target_value; ++ s32 default_value; ++ s32 no_constraint_value; ++ enum pm_qos_type type; ++ struct blocking_notifier_head *notifiers; ++}; ++ ++struct freq_constraints { ++ struct pm_qos_constraints min_freq; ++ struct blocking_notifier_head min_freq_notifiers; ++ struct pm_qos_constraints max_freq; ++ struct blocking_notifier_head max_freq_notifiers; ++}; ++ ++struct pm_qos_flags { ++ struct list_head list; ++ s32 effective_flags; ++}; ++ ++struct dev_pm_qos_request; ++ ++struct dev_pm_qos { ++ struct pm_qos_constraints resume_latency; ++ struct pm_qos_constraints latency_tolerance; ++ struct freq_constraints freq; ++ struct pm_qos_flags flags; ++ struct dev_pm_qos_request *resume_latency_req; ++ struct dev_pm_qos_request *latency_tolerance_req; ++ struct dev_pm_qos_request *flags_req; ++}; ++ ++struct device_attribute { ++ struct attribute attr; ++ ssize_t (*show)(struct device *, struct device_attribute *, char *); ++ ssize_t (*store)(struct device *, struct device_attribute *, const char *, size_t); ++}; ++ ++typedef u32 phandle; ++ ++struct property; ++ ++struct device_node { ++ const char *name; ++ phandle phandle; ++ const char *full_name; ++ struct fwnode_handle fwnode; ++ struct property *properties; ++ struct property *deadprops; ++ struct device_node *parent; ++ struct device_node *child; ++ struct device_node *sibling; ++ long unsigned int _flags; ++ void *data; ++}; ++ ++struct property { ++ char *name; ++ int length; ++ void *value; ++ struct property *next; ++}; ++ ++struct pm_qos_flags_request { ++ struct list_head node; ++ s32 flags; ++}; ++ ++enum freq_qos_req_type { ++ FREQ_QOS_MIN = 1, ++ FREQ_QOS_MAX = 2, ++}; ++ ++struct freq_qos_request { ++ enum freq_qos_req_type type; ++ struct plist_node pnode; ++ struct freq_constraints *qos; ++}; ++ ++enum dev_pm_qos_req_type { ++ DEV_PM_QOS_RESUME_LATENCY = 1, ++ DEV_PM_QOS_LATENCY_TOLERANCE = 2, ++ DEV_PM_QOS_MIN_FREQUENCY = 3, ++ DEV_PM_QOS_MAX_FREQUENCY = 4, ++ DEV_PM_QOS_FLAGS = 5, ++}; ++ ++struct dev_pm_qos_request { ++ enum dev_pm_qos_req_type type; ++ union { ++ struct plist_node pnode; ++ struct pm_qos_flags_request flr; ++ struct freq_qos_request freq; ++ } data; ++ struct device *dev; ++}; ++ ++enum stcctm_ctr_set { ++ EXTENDED = 0, ++ BASIC = 1, ++ PROBLEM_STATE = 2, ++ CRYPTO_ACTIVITY = 3, ++ MT_DIAG = 5, ++ MT_DIAG_CLEARING = 9, ++}; ++ ++struct s390_idle_data { ++ long unsigned int idle_count; ++ long unsigned int idle_time; ++ long unsigned int clock_idle_enter; ++ long unsigned int timer_idle_enter; ++ long unsigned int mt_cycles_enter[8]; ++}; ++ ++typedef u64 pcp_op_T__; ++ ++union register_pair { ++ __int128 unsigned pair; ++ struct { ++ long unsigned int even; ++ long unsigned int odd; ++ }; ++}; ++ ++struct cpuid { ++ unsigned int version: 8; ++ unsigned int ident: 24; ++ unsigned int machine: 16; ++ unsigned int unused: 16; ++}; ++ ++typedef void (*smp_call_func_t)(void *); ++ ++typedef bool (*smp_cond_func_t)(int, void *); ++ ++struct cpu_topology_s390 { ++ short unsigned int thread_id; ++ short unsigned int core_id; ++ short unsigned int socket_id; ++ short unsigned int book_id; ++ short unsigned int drawer_id; ++ short unsigned int dedicated: 1; ++ int booted_cores; ++ cpumask_t thread_mask; ++ cpumask_t core_mask; ++ cpumask_t book_mask; ++ cpumask_t drawer_mask; ++}; ++ ++enum { ++ HWCAP_NR_ESAN3 = 0, ++ HWCAP_NR_ZARCH = 1, ++ HWCAP_NR_STFLE = 2, ++ HWCAP_NR_MSA = 3, ++ HWCAP_NR_LDISP = 4, ++ HWCAP_NR_EIMM = 5, ++ HWCAP_NR_DFP = 6, ++ HWCAP_NR_HPAGE = 7, ++ HWCAP_NR_ETF3EH = 8, ++ HWCAP_NR_HIGH_GPRS = 9, ++ HWCAP_NR_TE = 10, ++ HWCAP_NR_VXRS = 11, ++ HWCAP_NR_VXRS_BCD = 12, ++ HWCAP_NR_VXRS_EXT = 13, ++ HWCAP_NR_GS = 14, ++ HWCAP_NR_VXRS_EXT2 = 15, ++ HWCAP_NR_VXRS_PDE = 16, ++ HWCAP_NR_SORT = 17, ++ HWCAP_NR_DFLT = 18, ++ HWCAP_NR_VXRS_PDE2 = 19, ++ HWCAP_NR_NNPA = 20, ++ HWCAP_NR_PCI_MIO = 21, ++ HWCAP_NR_SIE = 22, ++ HWCAP_NR_MAX = 23, ++}; ++ ++struct sclp_info { ++ unsigned char has_linemode: 1; ++ unsigned char has_vt220: 1; ++ unsigned char has_siif: 1; ++ unsigned char has_sigpif: 1; ++ unsigned char has_core_type: 1; ++ unsigned char has_sprp: 1; ++ unsigned char has_hvs: 1; ++ unsigned char has_esca: 1; ++ unsigned char has_sief2: 1; ++ unsigned char has_64bscao: 1; ++ unsigned char has_gpere: 1; ++ unsigned char has_cmma: 1; ++ unsigned char has_gsls: 1; ++ unsigned char has_ib: 1; ++ unsigned char has_cei: 1; ++ unsigned char has_pfmfi: 1; ++ unsigned char has_ibs: 1; ++ unsigned char has_skey: 1; ++ unsigned char has_kss: 1; ++ unsigned char has_gisaf: 1; ++ unsigned char has_diag318: 1; ++ unsigned char has_sipl: 1; ++ unsigned char has_sipl_eckd: 1; ++ unsigned char has_dirq: 1; ++ unsigned char has_iplcc: 1; ++ unsigned char has_zpci_lsi: 1; ++ unsigned char has_aisii: 1; ++ unsigned char has_aeni: 1; ++ unsigned char has_aisi: 1; ++ unsigned int ibc; ++ unsigned int mtid; ++ unsigned int mtid_cp; ++ unsigned int mtid_prev; ++ long unsigned int rzm; ++ long unsigned int rnmax; ++ long unsigned int hamax; ++ unsigned int max_cores; ++ long unsigned int hsa_size; ++ long unsigned int facilities; ++ unsigned int hmfai; ++}; ++ ++struct cpu_info { ++ unsigned int cpu_mhz_dynamic; ++ unsigned int cpu_mhz_static; ++ struct cpuid cpu_id; ++}; ++ ++struct psw_bits { ++ char: 1; ++ long unsigned int per: 1; ++ char: 3; ++ long unsigned int dat: 1; ++ long unsigned int io: 1; ++ long unsigned int ext: 1; ++ long unsigned int key: 4; ++ char: 1; ++ long unsigned int mcheck: 1; ++ long unsigned int wait: 1; ++ long unsigned int pstate: 1; ++ long unsigned int as: 2; ++ long unsigned int cc: 2; ++ long unsigned int pm: 4; ++ long unsigned int ri: 1; ++ char: 6; ++ long unsigned int eaba: 2; ++ long: 31; ++ long unsigned int ia: 64; ++}; ++ ++union ctlreg0 { ++ long unsigned int val; ++ struct { ++ char: 8; ++ long unsigned int tcx: 1; ++ long unsigned int pifo: 1; ++ char: 3; ++ long unsigned int ccc: 1; ++ long unsigned int pec: 1; ++ int: 17; ++ char: 3; ++ long unsigned int lap: 1; ++ char: 4; ++ long unsigned int edat: 1; ++ char: 2; ++ long unsigned int iep: 1; ++ char: 1; ++ long unsigned int afp: 1; ++ long unsigned int vx: 1; ++ short: 1; ++ char: 6; ++ long unsigned int sssm: 1; ++ }; ++}; ++ ++union ctlreg2 { ++ long unsigned int val; ++ struct { ++ long: 33; ++ long unsigned int ducto: 25; ++ char: 1; ++ long unsigned int gse: 1; ++ char: 1; ++ long unsigned int tds: 1; ++ long unsigned int tdc: 2; ++ }; ++}; ++ ++struct irq_stat { ++ unsigned int irqs[33]; ++}; ++ ++struct msi_dev_domain { ++ struct xarray store; ++ struct irq_domain *domain; ++}; ++ ++struct platform_msi_priv_data; ++ ++struct msi_device_data { ++ long unsigned int properties; ++ struct platform_msi_priv_data *platform_data; ++ struct mutex mutex; ++ struct msi_dev_domain __domains[2]; ++ long unsigned int __iter_idx; ++}; ++ ++struct iopf_device_param; ++ ++struct iommu_fault_param; ++ ++struct iommu_fwspec; ++ ++struct dev_iommu { ++ struct mutex lock; ++ struct iommu_fault_param *fault_param; ++ struct iopf_device_param *iopf_param; ++ struct iommu_fwspec *fwspec; ++ struct iommu_device *iommu_dev; ++ void *priv; ++ u32 max_pasids; ++ u32 attach_deferred: 1; ++}; ++ ++enum msi_domain_ids { ++ MSI_DEFAULT_DOMAIN = 0, ++ MSI_SECONDARY_DOMAIN = 1, ++ MSI_MAX_DEVICE_IRQDOMAINS = 2, ++}; ++ ++enum { ++ PCI_STD_RESOURCES = 0, ++ PCI_STD_RESOURCE_END = 5, ++ PCI_ROM_RESOURCE = 6, ++ PCI_IOV_RESOURCES = 7, ++ PCI_IOV_RESOURCE_END = 12, ++ PCI_BRIDGE_RESOURCES = 13, ++ PCI_BRIDGE_RESOURCE_END = 16, ++ PCI_NUM_RESOURCES = 17, ++ DEVICE_COUNT_RESOURCE = 17, ++}; ++ ++typedef unsigned int pci_channel_state_t; ++ ++typedef unsigned int pcie_reset_state_t; ++ ++typedef short unsigned int pci_dev_flags_t; ++ ++typedef short unsigned int pci_bus_flags_t; ++ ++typedef unsigned int pci_ers_result_t; ++ ++struct of_phandle_args { ++ struct device_node *np; ++ int args_count; ++ uint32_t args[16]; ++}; ++ ++struct iommu_fault_unrecoverable { ++ __u32 reason; ++ __u32 flags; ++ __u32 pasid; ++ __u32 perm; ++ __u64 addr; ++ __u64 fetch_addr; ++}; ++ ++struct iommu_fault_page_request { ++ __u32 flags; ++ __u32 pasid; ++ __u32 grpid; ++ __u32 perm; ++ __u64 addr; ++ __u64 private_data[2]; ++}; ++ ++struct iommu_fault { ++ __u32 type; ++ __u32 padding; ++ union { ++ struct iommu_fault_unrecoverable event; ++ struct iommu_fault_page_request prm; ++ __u8 padding2[56]; ++ }; ++}; ++ ++enum iommu_page_response_code { ++ IOMMU_PAGE_RESP_SUCCESS = 0, ++ IOMMU_PAGE_RESP_INVALID = 1, ++ IOMMU_PAGE_RESP_FAILURE = 2, ++}; ++ ++struct iommu_page_response { ++ __u32 argsz; ++ __u32 version; ++ __u32 flags; ++ __u32 pasid; ++ __u32 grpid; ++ __u32 code; ++}; ++ ++typedef int (*iommu_fault_handler_t)(struct iommu_domain *, struct device *, long unsigned int, int, void *); ++ ++struct iommu_domain_geometry { ++ dma_addr_t aperture_start; ++ dma_addr_t aperture_end; ++ bool force_aperture; ++}; ++ ++struct iommu_dma_cookie; ++ ++struct iommu_domain { ++ unsigned int type; ++ const struct iommu_domain_ops *ops; ++ long unsigned int pgsize_bitmap; ++ struct iommu_domain_geometry geometry; ++ struct iommu_dma_cookie *iova_cookie; ++ enum iommu_page_response_code (*iopf_handler)(struct iommu_fault *, void *); ++ void *fault_data; ++ union { ++ struct { ++ iommu_fault_handler_t handler; ++ void *handler_token; ++ }; ++ struct { ++ struct mm_struct *mm; ++ int users; ++ }; ++ }; ++}; ++ ++typedef int (*iommu_dev_fault_handler_t)(struct iommu_fault *, void *); ++ ++struct iommu_iotlb_gather; ++ ++struct iommu_domain_ops { ++ int (*attach_dev)(struct iommu_domain *, struct device *); ++ int (*set_dev_pasid)(struct iommu_domain *, struct device *, ioasid_t); ++ int (*map)(struct iommu_domain *, long unsigned int, phys_addr_t, size_t, int, gfp_t); ++ int (*map_pages)(struct iommu_domain *, long unsigned int, phys_addr_t, size_t, size_t, int, gfp_t, size_t *); ++ size_t (*unmap)(struct iommu_domain *, long unsigned int, size_t, struct iommu_iotlb_gather *); ++ size_t (*unmap_pages)(struct iommu_domain *, long unsigned int, size_t, size_t, struct iommu_iotlb_gather *); ++ void (*flush_iotlb_all)(struct iommu_domain *); ++ void (*iotlb_sync_map)(struct iommu_domain *, long unsigned int, size_t); ++ void (*iotlb_sync)(struct iommu_domain *, struct iommu_iotlb_gather *); ++ phys_addr_t (*iova_to_phys)(struct iommu_domain *, dma_addr_t); ++ bool (*enforce_cache_coherency)(struct iommu_domain *); ++ int (*enable_nesting)(struct iommu_domain *); ++ int (*set_pgtable_quirks)(struct iommu_domain *, long unsigned int); ++ void (*free)(struct iommu_domain *); ++}; ++ ++struct iommu_iotlb_gather { ++ long unsigned int start; ++ long unsigned int end; ++ size_t pgsize; ++ struct list_head freelist; ++ bool queued; ++}; ++ ++struct iommu_device { ++ struct list_head list; ++ const struct iommu_ops *ops; ++ struct fwnode_handle *fwnode; ++ struct device *dev; ++ u32 max_pasids; ++}; ++ ++struct iommu_fault_event { ++ struct iommu_fault fault; ++ struct list_head list; ++}; ++ ++struct iommu_fault_param { ++ iommu_dev_fault_handler_t handler; ++ void *data; ++ struct list_head faults; ++ struct mutex lock; ++}; ++ ++struct iommu_fwspec { ++ const struct iommu_ops *ops; ++ struct fwnode_handle *iommu_fwnode; ++ u32 flags; ++ unsigned int num_ids; ++ u32 ids[0]; ++}; ++ ++struct ccw_dev_id { ++ u8 ssid; ++ u16 devno; ++}; ++ ++struct diag210 { ++ u16 vrdcdvno; ++ u16 vrdclen; ++ u8 vrdcvcla; ++ u8 vrdcvtyp; ++ u8 vrdcvsta; ++ u8 vrdcvfla; ++ u8 vrdcrccl; ++ u8 vrdccrty; ++ u8 vrdccrmd; ++ u8 vrdccrft; ++}; ++ ++struct diag8c { ++ u8 flags; ++ u8 num_partitions; ++ u16 width; ++ u16 height; ++ u8 data[0]; ++ long: 0; ++}; ++ ++enum diag26c_sc { ++ DIAG26C_PORT_VNIC = 36, ++ DIAG26C_MAC_SERVICES = 48, ++}; ++ ++struct hypfs_diag0c_entry; ++ ++struct diag_ops { ++ int (*diag210)(struct diag210 *); ++ int (*diag26c)(void *, void *, enum diag26c_sc); ++ int (*diag14)(long unsigned int, long unsigned int, long unsigned int); ++ int (*diag8c)(struct diag8c *, struct ccw_dev_id *, size_t); ++ void (*diag0c)(struct hypfs_diag0c_entry *); ++ void (*diag308_reset)(); ++}; ++ ++struct mcck_volatile_info { ++ __u64 mcic; ++ __u64 failing_storage_address; ++ __u32 ext_damage_code; ++ __u32 reserved; ++}; ++ ++struct kvm_s390_sie_block { ++ atomic_t cpuflags; ++ char: 1; ++ __u32 prefix: 18; ++ char: 1; ++ __u32 ibc: 12; ++ __u8 reserved08[4]; ++ __u32 prog0c; ++ union { ++ __u8 reserved10[16]; ++ struct { ++ __u64 pv_handle_cpu; ++ __u64 pv_handle_config; ++ }; ++ }; ++ atomic_t prog20; ++ __u8 reserved24[4]; ++ __u64 cputm; ++ __u64 ckc; ++ __u64 epoch; ++ __u32 svcc; ++ __u16 lctl; ++ __s16 icpua; ++ __u32 ictl; ++ __u32 eca; ++ __u8 icptcode; ++ __u8 icptstatus; ++ __u16 ihcpu; ++ __u8 reserved54; ++ __u8 iictl; ++ __u16 ipa; ++ __u32 ipb; ++ __u32 scaoh; ++ __u8 fpf; ++ __u8 ecb; ++ __u8 ecb2; ++ __u8 ecb3; ++ __u32 scaol; ++ __u8 sdf; ++ __u8 epdx; ++ __u8 cpnc; ++ __u8 reserved6b; ++ __u32 todpr; ++ __u32 gd; ++ __u8 reserved74[12]; ++ __u64 mso; ++ __u64 msl; ++ psw_t gpsw; ++ __u64 gg14; ++ __u64 gg15; ++ __u8 reservedb0[8]; ++ __u8 hpid; ++ __u8 reservedb9[7]; ++ union { ++ struct { ++ __u32 eiparams; ++ __u16 extcpuaddr; ++ __u16 eic; ++ }; ++ __u64 mcic; ++ }; ++ __u32 reservedc8; ++ union { ++ struct { ++ __u16 pgmilc; ++ __u16 iprcc; ++ }; ++ __u32 edc; ++ }; ++ union { ++ struct { ++ __u32 dxc; ++ __u16 mcn; ++ __u8 perc; ++ __u8 peratmid; ++ }; ++ __u64 faddr; ++ }; ++ __u64 peraddr; ++ __u8 eai; ++ __u8 peraid; ++ __u8 oai; ++ __u8 armid; ++ __u8 reservede4[4]; ++ union { ++ __u64 tecmc; ++ struct { ++ __u16 subchannel_id; ++ __u16 subchannel_nr; ++ __u32 io_int_parm; ++ __u32 io_int_word; ++ }; ++ }; ++ __u8 reservedf4[8]; ++ __u32 crycbd; ++ __u64 gcr[16]; ++ union { ++ __u64 gbea; ++ __u64 sidad; ++ }; ++ __u8 reserved188[8]; ++ __u64 sdnxo; ++ __u8 reserved198[8]; ++ __u32 fac; ++ __u8 reserved1a4[20]; ++ __u64 cbrlo; ++ __u8 reserved1c0[8]; ++ __u32 ecd; ++ __u8 reserved1cc[18]; ++ __u64 pp; ++ __u8 reserved1e6[2]; ++ __u64 itdba; ++ __u64 riccbd; ++ __u64 gvrd; ++} __attribute__((packed)); ++ ++struct kvm_s390_itdb { ++ __u8 data[256]; ++}; ++ ++struct sie_page { ++ struct kvm_s390_sie_block sie_block; ++ struct mcck_volatile_info mcck_info; ++ __u8 reserved218[360]; ++ __u64 pv_grregs[16]; ++ __u8 reserved400[512]; ++ struct kvm_s390_itdb itdb; ++ __u8 reserved700[2304]; ++}; ++ ++enum kvm_bus { ++ KVM_MMIO_BUS = 0, ++ KVM_PIO_BUS = 1, ++ KVM_VIRTIO_CCW_NOTIFY_BUS = 2, ++ KVM_FAST_MMIO_BUS = 3, ++ KVM_NR_BUSES = 4, ++}; ++ ++union mci { ++ long unsigned int val; ++ struct { ++ u64 sd: 1; ++ u64 pd: 1; ++ u64 sr: 1; ++ char: 1; ++ u64 cd: 1; ++ u64 ed: 1; ++ char: 1; ++ u64 dg: 1; ++ u64 w: 1; ++ u64 cp: 1; ++ u64 sp: 1; ++ u64 ck: 1; ++ char: 2; ++ u64 b: 1; ++ short: 1; ++ u64 se: 1; ++ u64 sc: 1; ++ u64 ke: 1; ++ u64 ds: 1; ++ u64 wp: 1; ++ u64 ms: 1; ++ u64 pm: 1; ++ u64 ia: 1; ++ u64 fa: 1; ++ u64 vr: 1; ++ u64 ec: 1; ++ u64 fp: 1; ++ u64 gr: 1; ++ u64 cr: 1; ++ char: 1; ++ u64 st: 1; ++ u64 ie: 1; ++ u64 ar: 1; ++ u64 da: 1; ++ char: 1; ++ u64 gs: 1; ++ char: 3; ++ char: 2; ++ u64 pr: 1; ++ u64 fc: 1; ++ u64 ap: 1; ++ char: 1; ++ u64 ct: 1; ++ u64 cc: 1; ++ }; ++}; ++ ++struct mcesa { ++ u8 vector_save_area[1024]; ++ u8 guarded_storage_save_area[32]; ++}; ++ ++struct mcck_struct { ++ unsigned int kill_task: 1; ++ unsigned int channel_report: 1; ++ unsigned int warning: 1; ++ unsigned int stp_queue: 1; ++ long unsigned int mcck_code; ++}; ++ ++enum { ++ INSTR_E = 0, ++ INSTR_IE_UU = 1, ++ INSTR_MII_UPP = 2, ++ INSTR_RIE_R0IU = 3, ++ INSTR_RIE_R0UU = 4, ++ INSTR_RIE_RRI0 = 5, ++ INSTR_RIE_RRP = 6, ++ INSTR_RIE_RRPU = 7, ++ INSTR_RIE_RRUUU = 8, ++ INSTR_RIE_RUI0 = 9, ++ INSTR_RIE_RUPI = 10, ++ INSTR_RIE_RUPU = 11, ++ INSTR_RIL_RI = 12, ++ INSTR_RIL_RP = 13, ++ INSTR_RIL_RU = 14, ++ INSTR_RIL_UP = 15, ++ INSTR_RIS_RURDI = 16, ++ INSTR_RIS_RURDU = 17, ++ INSTR_RI_RI = 18, ++ INSTR_RI_RP = 19, ++ INSTR_RI_RU = 20, ++ INSTR_RI_UP = 21, ++ INSTR_RRE_00 = 22, ++ INSTR_RRE_AA = 23, ++ INSTR_RRE_AR = 24, ++ INSTR_RRE_F0 = 25, ++ INSTR_RRE_FF = 26, ++ INSTR_RRE_FR = 27, ++ INSTR_RRE_R0 = 28, ++ INSTR_RRE_RA = 29, ++ INSTR_RRE_RF = 30, ++ INSTR_RRE_RR = 31, ++ INSTR_RRF_0UFF = 32, ++ INSTR_RRF_0URF = 33, ++ INSTR_RRF_F0FF = 34, ++ INSTR_RRF_F0FF2 = 35, ++ INSTR_RRF_F0FR = 36, ++ INSTR_RRF_FFRU = 37, ++ INSTR_RRF_FUFF = 38, ++ INSTR_RRF_FUFF2 = 39, ++ INSTR_RRF_R0RR = 40, ++ INSTR_RRF_R0RR2 = 41, ++ INSTR_RRF_RURR = 42, ++ INSTR_RRF_RURR2 = 43, ++ INSTR_RRF_U0FF = 44, ++ INSTR_RRF_U0RF = 45, ++ INSTR_RRF_U0RR = 46, ++ INSTR_RRF_URR = 47, ++ INSTR_RRF_UUFF = 48, ++ INSTR_RRF_UUFR = 49, ++ INSTR_RRF_UURF = 50, ++ INSTR_RRS_RRRDU = 51, ++ INSTR_RR_FF = 52, ++ INSTR_RR_R0 = 53, ++ INSTR_RR_RR = 54, ++ INSTR_RR_U0 = 55, ++ INSTR_RR_UR = 56, ++ INSTR_RSI_RRP = 57, ++ INSTR_RSL_LRDFU = 58, ++ INSTR_RSL_R0RD = 59, ++ INSTR_RSY_AARD = 60, ++ INSTR_RSY_CCRD = 61, ++ INSTR_RSY_RDRU = 62, ++ INSTR_RSY_RRRD = 63, ++ INSTR_RSY_RURD = 64, ++ INSTR_RSY_RURD2 = 65, ++ INSTR_RS_AARD = 66, ++ INSTR_RS_CCRD = 67, ++ INSTR_RS_R0RD = 68, ++ INSTR_RS_RRRD = 69, ++ INSTR_RS_RURD = 70, ++ INSTR_RXE_FRRD = 71, ++ INSTR_RXE_RRRDU = 72, ++ INSTR_RXF_FRRDF = 73, ++ INSTR_RXY_FRRD = 74, ++ INSTR_RXY_RRRD = 75, ++ INSTR_RXY_URRD = 76, ++ INSTR_RX_FRRD = 77, ++ INSTR_RX_RRRD = 78, ++ INSTR_RX_URRD = 79, ++ INSTR_SIL_RDI = 80, ++ INSTR_SIL_RDU = 81, ++ INSTR_SIY_IRD = 82, ++ INSTR_SIY_RD = 83, ++ INSTR_SIY_URD = 84, ++ INSTR_SI_RD = 85, ++ INSTR_SI_URD = 86, ++ INSTR_SMI_U0RDP = 87, ++ INSTR_SSE_RDRD = 88, ++ INSTR_SSF_RRDRD = 89, ++ INSTR_SSF_RRDRD2 = 90, ++ INSTR_SS_L0RDRD = 91, ++ INSTR_SS_L2RDRD = 92, ++ INSTR_SS_LIRDRD = 93, ++ INSTR_SS_LLRDRD = 94, ++ INSTR_SS_RRRDRD = 95, ++ INSTR_SS_RRRDRD2 = 96, ++ INSTR_SS_RRRDRD3 = 97, ++ INSTR_S_00 = 98, ++ INSTR_S_RD = 99, ++ INSTR_VRI_V0IU = 100, ++ INSTR_VRI_V0U = 101, ++ INSTR_VRI_V0UU2 = 102, ++ INSTR_VRI_V0UUU = 103, ++ INSTR_VRI_VR0UU = 104, ++ INSTR_VRI_VVUU = 105, ++ INSTR_VRI_VVUUU = 106, ++ INSTR_VRI_VVUUU2 = 107, ++ INSTR_VRI_VVV0U = 108, ++ INSTR_VRI_VVV0UU = 109, ++ INSTR_VRI_VVV0UU2 = 110, ++ INSTR_VRR_0V = 111, ++ INSTR_VRR_0VV0U = 112, ++ INSTR_VRR_RV0UU = 113, ++ INSTR_VRR_VRR = 114, ++ INSTR_VRR_VV = 115, ++ INSTR_VRR_VV0U = 116, ++ INSTR_VRR_VV0U0U = 117, ++ INSTR_VRR_VV0U2 = 118, ++ INSTR_VRR_VV0UU2 = 119, ++ INSTR_VRR_VV0UUU = 120, ++ INSTR_VRR_VVV = 121, ++ INSTR_VRR_VVV0U = 122, ++ INSTR_VRR_VVV0U0 = 123, ++ INSTR_VRR_VVV0U0U = 124, ++ INSTR_VRR_VVV0UU = 125, ++ INSTR_VRR_VVV0UUU = 126, ++ INSTR_VRR_VVV0V = 127, ++ INSTR_VRR_VVVU0UV = 128, ++ INSTR_VRR_VVVU0V = 129, ++ INSTR_VRR_VVVUU0V = 130, ++ INSTR_VRS_RRDV = 131, ++ INSTR_VRS_RVRDU = 132, ++ INSTR_VRS_VRRD = 133, ++ INSTR_VRS_VRRDU = 134, ++ INSTR_VRS_VVRDU = 135, ++ INSTR_VRV_VVXRDU = 136, ++ INSTR_VRX_VRRDU = 137, ++ INSTR_VRX_VV = 138, ++ INSTR_VSI_URDV = 139, ++}; ++ ++enum { ++ LONG_INSN_ALGHSIK = 0, ++ LONG_INSN_ALHHHR = 1, ++ LONG_INSN_ALHHLR = 2, ++ LONG_INSN_ALHSIK = 3, ++ LONG_INSN_ALSIHN = 4, ++ LONG_INSN_CDFBRA = 5, ++ LONG_INSN_CDGBRA = 6, ++ LONG_INSN_CDGTRA = 7, ++ LONG_INSN_CDLFBR = 8, ++ LONG_INSN_CDLFTR = 9, ++ LONG_INSN_CDLGBR = 10, ++ LONG_INSN_CDLGTR = 11, ++ LONG_INSN_CEFBRA = 12, ++ LONG_INSN_CEGBRA = 13, ++ LONG_INSN_CELFBR = 14, ++ LONG_INSN_CELGBR = 15, ++ LONG_INSN_CFDBRA = 16, ++ LONG_INSN_CFEBRA = 17, ++ LONG_INSN_CFXBRA = 18, ++ LONG_INSN_CGDBRA = 19, ++ LONG_INSN_CGDTRA = 20, ++ LONG_INSN_CGEBRA = 21, ++ LONG_INSN_CGXBRA = 22, ++ LONG_INSN_CGXTRA = 23, ++ LONG_INSN_CLFDBR = 24, ++ LONG_INSN_CLFDTR = 25, ++ LONG_INSN_CLFEBR = 26, ++ LONG_INSN_CLFHSI = 27, ++ LONG_INSN_CLFXBR = 28, ++ LONG_INSN_CLFXTR = 29, ++ LONG_INSN_CLGDBR = 30, ++ LONG_INSN_CLGDTR = 31, ++ LONG_INSN_CLGEBR = 32, ++ LONG_INSN_CLGFRL = 33, ++ LONG_INSN_CLGHRL = 34, ++ LONG_INSN_CLGHSI = 35, ++ LONG_INSN_CLGXBR = 36, ++ LONG_INSN_CLGXTR = 37, ++ LONG_INSN_CLHHSI = 38, ++ LONG_INSN_CXFBRA = 39, ++ LONG_INSN_CXGBRA = 40, ++ LONG_INSN_CXGTRA = 41, ++ LONG_INSN_CXLFBR = 42, ++ LONG_INSN_CXLFTR = 43, ++ LONG_INSN_CXLGBR = 44, ++ LONG_INSN_CXLGTR = 45, ++ LONG_INSN_DFLTCC = 46, ++ LONG_INSN_FIDBRA = 47, ++ LONG_INSN_FIEBRA = 48, ++ LONG_INSN_FIXBRA = 49, ++ LONG_INSN_ILLEGAL = 50, ++ LONG_INSN_LDXBRA = 51, ++ LONG_INSN_LEDBRA = 52, ++ LONG_INSN_LEXBRA = 53, ++ LONG_INSN_LLGFAT = 54, ++ LONG_INSN_LLGFRL = 55, ++ LONG_INSN_LLGFSG = 56, ++ LONG_INSN_LLGHRL = 57, ++ LONG_INSN_LLGTAT = 58, ++ LONG_INSN_LLZRGF = 59, ++ LONG_INSN_LOCFHR = 60, ++ LONG_INSN_LOCGHI = 61, ++ LONG_INSN_LOCHHI = 62, ++ LONG_INSN_LPSWEY = 63, ++ LONG_INSN_MPCIFC = 64, ++ LONG_INSN_MSGRKC = 65, ++ LONG_INSN_PCILGI = 66, ++ LONG_INSN_PCISTB = 67, ++ LONG_INSN_PCISTBI = 68, ++ LONG_INSN_PCISTG = 69, ++ LONG_INSN_PCISTGI = 70, ++ LONG_INSN_POPCNT = 71, ++ LONG_INSN_RIEMIT = 72, ++ LONG_INSN_RINEXT = 73, ++ LONG_INSN_RISBGN = 74, ++ LONG_INSN_RISBHG = 75, ++ LONG_INSN_RISBLG = 76, ++ LONG_INSN_SELFHR = 77, ++ LONG_INSN_SLHHHR = 78, ++ LONG_INSN_SLHHLR = 79, ++ LONG_INSN_STBEAR = 80, ++ LONG_INSN_STCCTM = 81, ++ LONG_INSN_STOCFH = 82, ++ LONG_INSN_STPCIFC = 83, ++ LONG_INSN_TABORT = 84, ++ LONG_INSN_TBEGIN = 85, ++ LONG_INSN_TBEGINC = 86, ++ LONG_INSN_VBPERM = 87, ++ LONG_INSN_VCLFNH = 88, ++ LONG_INSN_VCLFNL = 89, ++ LONG_INSN_VCLZDP = 90, ++ LONG_INSN_VERLLV = 91, ++ LONG_INSN_VESRAV = 92, ++ LONG_INSN_VESRLV = 93, ++ LONG_INSN_VLBRREP = 94, ++ LONG_INSN_VLEBRF = 95, ++ LONG_INSN_VLEBRG = 96, ++ LONG_INSN_VLEBRH = 97, ++ LONG_INSN_VLLEBRZ = 98, ++ LONG_INSN_VPOPCT = 99, ++ LONG_INSN_VSBCBI = 100, ++ LONG_INSN_VSCSHP = 101, ++ LONG_INSN_VSTEBRF = 102, ++ LONG_INSN_VSTEBRG = 103, ++ LONG_INSN_VSTEBRH = 104, ++ LONG_INSN_VSTRLR = 105, ++ LONG_INSN_VUPKZH = 106, ++ LONG_INSN_VUPKZL = 107, ++}; ++ ++struct s390_operand { ++ unsigned char bits; ++ unsigned char shift; ++ short unsigned int flags; ++}; ++ ++struct s390_insn { ++ union { ++ const char name[5]; ++ struct { ++ unsigned char zero; ++ unsigned int offset; ++ } __attribute__((packed)); ++ }; ++ unsigned char opfrag; ++ unsigned char format; ++}; ++ ++struct s390_opcode_offset { ++ unsigned char opcode; ++ unsigned char mask; ++ unsigned char byte; ++ short unsigned int offset; ++ short unsigned int count; ++} __attribute__((packed)); ++ ++enum { ++ UNUSED = 0, ++ A_8 = 1, ++ A_12 = 2, ++ A_24 = 3, ++ A_28 = 4, ++ B_16 = 5, ++ B_32 = 6, ++ C_8 = 7, ++ C_12 = 8, ++ D20_20 = 9, ++ D_20 = 10, ++ D_36 = 11, ++ F_8 = 12, ++ F_12 = 13, ++ F_16 = 14, ++ F_24 = 15, ++ F_28 = 16, ++ F_32 = 17, ++ I8_8 = 18, ++ I8_32 = 19, ++ I16_16 = 20, ++ I16_32 = 21, ++ I32_16 = 22, ++ J12_12 = 23, ++ J16_16 = 24, ++ J16_32 = 25, ++ J24_24 = 26, ++ J32_16 = 27, ++ L4_8 = 28, ++ L4_12 = 29, ++ L8_8 = 30, ++ R_8 = 31, ++ R_12 = 32, ++ R_16 = 33, ++ R_24 = 34, ++ R_28 = 35, ++ U4_8 = 36, ++ U4_12 = 37, ++ U4_16 = 38, ++ U4_20 = 39, ++ U4_24 = 40, ++ U4_28 = 41, ++ U4_32 = 42, ++ U4_36 = 43, ++ U8_8 = 44, ++ U8_16 = 45, ++ U8_24 = 46, ++ U8_28 = 47, ++ U8_32 = 48, ++ U12_16 = 49, ++ U16_16 = 50, ++ U16_32 = 51, ++ U32_16 = 52, ++ VX_12 = 53, ++ V_8 = 54, ++ V_12 = 55, ++ V_16 = 56, ++ V_32 = 57, ++ X_12 = 58, ++}; ++ ++struct timens_offset { ++ s64 sec; ++ u64 nsec; ++}; ++ ++struct linux_binprm { ++ struct vm_area_struct *vma; ++ long unsigned int vma_pages; ++ struct mm_struct *mm; ++ long unsigned int p; ++ long unsigned int argmin; ++ unsigned int have_execfd: 1; ++ unsigned int execfd_creds: 1; ++ unsigned int secureexec: 1; ++ unsigned int point_of_no_return: 1; ++ struct file *executable; ++ struct file *interpreter; ++ struct file *file; ++ struct cred *cred; ++ int unsafe; ++ unsigned int per_clear; ++ int argc; ++ int envc; ++ const char *filename; ++ const char *interp; ++ const char *fdpath; ++ unsigned int interp_flags; ++ int execfd; ++ long unsigned int loader; ++ long unsigned int exec; ++ struct rlimit rlim_stack; ++ char buf[256]; ++}; ++ ++struct coredump_params; ++ ++struct linux_binfmt { ++ struct list_head lh; ++ struct module *module; ++ int (*load_binary)(struct linux_binprm *); ++ int (*load_shlib)(struct file *); ++ int (*core_dump)(struct coredump_params *); ++ long unsigned int min_coredump; ++}; ++ ++struct maple_alloc { ++ long unsigned int total; ++ unsigned char node_count; ++ unsigned int request_count; ++ struct maple_alloc *slot[30]; ++}; ++ ++struct maple_enode; ++ ++struct ma_state { ++ struct maple_tree *tree; ++ long unsigned int index; ++ long unsigned int last; ++ struct maple_enode *node; ++ long unsigned int min; ++ long unsigned int max; ++ struct maple_alloc *alloc; ++ unsigned char depth; ++ unsigned char offset; ++ unsigned char mas_flags; ++}; ++ ++struct vma_iterator { ++ struct ma_state mas; ++}; ++ ++enum vm_fault_reason { ++ VM_FAULT_OOM = 1, ++ VM_FAULT_SIGBUS = 2, ++ VM_FAULT_MAJOR = 4, ++ VM_FAULT_HWPOISON = 16, ++ VM_FAULT_HWPOISON_LARGE = 32, ++ VM_FAULT_SIGSEGV = 64, ++ VM_FAULT_NOPAGE = 256, ++ VM_FAULT_LOCKED = 512, ++ VM_FAULT_RETRY = 1024, ++ VM_FAULT_FALLBACK = 2048, ++ VM_FAULT_DONE_COW = 4096, ++ VM_FAULT_NEEDDSYNC = 8192, ++ VM_FAULT_COMPLETED = 16384, ++ VM_FAULT_HINDEX_MASK = 983040, ++}; ++ ++struct vm_special_mapping { ++ const char *name; ++ struct page **pages; ++ vm_fault_t (*fault)(const struct vm_special_mapping *, struct vm_area_struct *, struct vm_fault *); ++ int (*mremap)(const struct vm_special_mapping *, struct vm_area_struct *); ++}; ++ ++typedef unsigned int zap_flags_t; ++ ++struct zap_details { ++ struct folio *single_folio; ++ bool even_cows; ++ zap_flags_t zap_flags; ++}; ++ ++struct timens_offsets { ++ struct timespec64 monotonic; ++ struct timespec64 boottime; ++}; ++ ++struct time_namespace { ++ struct user_namespace *user_ns; ++ struct ucounts *ucounts; ++ struct ns_common ns; ++ struct timens_offsets offsets; ++ struct page *vvar_page; ++ bool frozen_offsets; ++}; ++ ++struct arch_vdso_data { ++ __s64 tod_steering_delta; ++ __u64 tod_steering_end; ++}; ++ ++struct vdso_timestamp { ++ u64 sec; ++ u64 nsec; ++}; ++ ++struct vdso_data { ++ u32 seq; ++ s32 clock_mode; ++ u64 cycle_last; ++ u64 mask; ++ u32 mult; ++ u32 shift; ++ union { ++ struct vdso_timestamp basetime[12]; ++ struct timens_offset offset[12]; ++ }; ++ s32 tz_minuteswest; ++ s32 tz_dsttime; ++ u32 hrtimer_res; ++ u32 __unused; ++ struct arch_vdso_data arch_data; ++}; ++ ++enum vvar_pages { ++ VVAR_DATA_PAGE_OFFSET = 0, ++ VVAR_TIMENS_PAGE_OFFSET = 1, ++ VVAR_NR_PAGES = 2, ++}; ++ ++struct sysinfo_2_2_2 { ++ char reserved_0[32]; ++ short unsigned int lpar_number; ++ char reserved_1; ++ unsigned char characteristics; ++ short unsigned int cpus_total; ++ short unsigned int cpus_configured; ++ short unsigned int cpus_standby; ++ short unsigned int cpus_reserved; ++ char name[8]; ++ unsigned int caf; ++ char reserved_2[8]; ++ unsigned char mt_installed: 1; ++ char: 2; ++ unsigned char mt_stid: 5; ++ char: 3; ++ unsigned char mt_gtid: 5; ++ char: 3; ++ unsigned char mt_psmtid: 5; ++ char reserved_3[5]; ++ short unsigned int cpus_dedicated; ++ short unsigned int cpus_shared; ++ char reserved_4[3]; ++ unsigned char vsne; ++ uuid_t uuid; ++ char reserved_5[160]; ++ char ext_name[256]; ++}; ++ ++struct __debug_entry { ++ long unsigned int clock: 60; ++ long unsigned int exception: 1; ++ long unsigned int level: 3; ++ void *caller; ++ short unsigned int cpu; ++} __attribute__((packed)); ++ ++typedef struct __debug_entry debug_entry_t; ++ ++struct debug_view; ++ ++struct debug_info { ++ struct debug_info *next; ++ struct debug_info *prev; ++ refcount_t ref_count; ++ spinlock_t lock; ++ int level; ++ int nr_areas; ++ int pages_per_area; ++ int buf_size; ++ int entry_size; ++ debug_entry_t ***areas; ++ int active_area; ++ int *active_pages; ++ int *active_entries; ++ struct dentry *debugfs_root_entry; ++ struct dentry *debugfs_entries[10]; ++ struct debug_view *views[10]; ++ char name[64]; ++ umode_t mode; ++}; ++ ++typedef struct debug_info debug_info_t; ++ ++typedef int debug_prolog_proc_t(debug_info_t *, struct debug_view *, char *); ++ ++typedef int debug_header_proc_t(debug_info_t *, struct debug_view *, int, debug_entry_t *, char *); ++ ++typedef int debug_format_proc_t(debug_info_t *, struct debug_view *, char *, const char *); ++ ++typedef int debug_input_proc_t(debug_info_t *, struct debug_view *, struct file *, const char *, size_t, loff_t *); ++ ++struct debug_view { ++ char name[64]; ++ debug_prolog_proc_t *prolog_proc; ++ debug_header_proc_t *header_proc; ++ debug_format_proc_t *format_proc; ++ debug_input_proc_t *input_proc; ++ void *private_data; ++}; ++ ++struct lgr_info { ++ u64 stfle_fac_list[4]; ++ u32 level; ++ char manufacturer[16]; ++ char type[4]; ++ char sequence[16]; ++ char plant[4]; ++ char model[16]; ++ u16 lpar_number; ++ char name[8]; ++ u8 vm_count; ++ struct { ++ char name[8]; ++ char cpi[16]; ++ } vm[2]; ++}; ++ ++struct kernel_fpu { ++ u32 mask; ++ u32 fpc; ++ union { ++ freg_t fprs[16]; ++ __vector128 vxrs[32]; ++ }; ++}; ++ ++enum { ++ TRACE_EVENT_FL_FILTERED = 1, ++ TRACE_EVENT_FL_CAP_ANY = 2, ++ TRACE_EVENT_FL_NO_SET_FILTER = 4, ++ TRACE_EVENT_FL_IGNORE_ENABLE = 8, ++ TRACE_EVENT_FL_TRACEPOINT = 16, ++ TRACE_EVENT_FL_DYNAMIC = 32, ++ TRACE_EVENT_FL_KPROBE = 64, ++ TRACE_EVENT_FL_UPROBE = 128, ++ TRACE_EVENT_FL_EPROBE = 256, ++ TRACE_EVENT_FL_CUSTOM = 512, ++}; ++ ++struct syscall_metadata { ++ const char *name; ++ int syscall_nr; ++ int nb_args; ++ const char **types; ++ const char **args; ++ struct list_head enter_fields; ++ struct trace_event_call *enter_event; ++ struct trace_event_call *exit_event; ++}; ++ ++enum diag204_sc { ++ DIAG204_SUBC_STIB4 = 4, ++ DIAG204_SUBC_RSI = 5, ++ DIAG204_SUBC_STIB6 = 6, ++ DIAG204_SUBC_STIB7 = 7, ++}; ++ ++enum diag204_format { ++ DIAG204_INFO_SIMPLE = 0, ++ DIAG204_INFO_EXT = 65536, ++}; ++ ++enum diag204_cpu_flags { ++ DIAG204_CPU_ONLINE = 32, ++ DIAG204_CPU_CAPPED = 64, ++}; ++ ++struct diag204_x_info_blk_hdr { ++ __u8 npar; ++ __u8 flags; ++ __u16 tslice; ++ __u16 phys_cpus; ++ __u16 this_part; ++ __u64 curtod1; ++ __u64 curtod2; ++ char reserved[40]; ++}; ++ ++struct diag204_x_part_hdr { ++ __u8 pn; ++ __u8 cpus; ++ __u8 rcpus; ++ __u8 pflag; ++ __u32 mlu; ++ char part_name[8]; ++ char lpc_name[8]; ++ char os_name[8]; ++ __u64 online_cs; ++ __u64 online_es; ++ __u8 upid; ++ __u8 reserved: 3; ++ __u8 mtid: 5; ++ char reserved1[2]; ++ __u32 group_mlu; ++ char group_name[8]; ++ char hardware_group_name[8]; ++ char reserved2[24]; ++}; ++ ++struct diag204_x_cpu_info { ++ __u16 cpu_addr; ++ char reserved1[2]; ++ __u8 ctidx; ++ __u8 cflag; ++ __u16 weight; ++ __u64 acc_time; ++ __u64 lp_time; ++ __u16 min_weight; ++ __u16 cur_weight; ++ __u16 max_weight; ++ char reseved2[2]; ++ __u64 online_time; ++ __u64 wait_time; ++ __u32 pma_weight; ++ __u32 polar_weight; ++ __u32 cpu_type_cap; ++ __u32 group_cpu_type_cap; ++ char reserved3[32]; ++}; ++ ++struct diag204_x_phys_hdr { ++ char reserved1[1]; ++ __u8 cpus; ++ char reserved2[6]; ++ char mgm_name[8]; ++ char reserved3[80]; ++}; ++ ++struct diag204_x_phys_cpu { ++ __u16 cpu_addr; ++ char reserved1[2]; ++ __u8 ctidx; ++ char reserved2[1]; ++ __u16 weight; ++ __u64 mgm_time; ++ char reserved3[80]; ++}; ++ ++struct diag204_x_part_block { ++ struct diag204_x_part_hdr hdr; ++ struct diag204_x_cpu_info cpus[0]; ++}; ++ ++struct diag204_x_phys_block { ++ struct diag204_x_phys_hdr hdr; ++ struct diag204_x_phys_cpu cpus[0]; ++}; ++ ++enum hdr_flags { ++ HDR_NOT_LPAR = 16, ++ HDR_STACK_INCM = 32, ++ HDR_STSI_UNAV = 64, ++ HDR_PERF_UNAV = 128, ++}; ++ ++enum mac_validity { ++ MAC_NAME_VLD = 32, ++ MAC_ID_VLD = 64, ++ MAC_CNT_VLD = 128, ++}; ++ ++enum par_flag { ++ PAR_MT_EN = 128, ++}; ++ ++enum par_validity { ++ PAR_GRP_VLD = 8, ++ PAR_ID_VLD = 16, ++ PAR_ABS_VLD = 32, ++ PAR_WGHT_VLD = 64, ++ PAR_PCNT_VLD = 128, ++}; ++ ++struct hdr_sctn { ++ u8 infhflg1; ++ u8 infhflg2; ++ u8 infhval1; ++ u8 infhval2; ++ u8 reserved[3]; ++ u8 infhygct; ++ u16 infhtotl; ++ u16 infhdln; ++ u16 infmoff; ++ u16 infmlen; ++ u16 infpoff; ++ u16 infplen; ++ u16 infhoff1; ++ u16 infhlen1; ++ u16 infgoff1; ++ u16 infglen1; ++ u16 infhoff2; ++ u16 infhlen2; ++ u16 infgoff2; ++ u16 infglen2; ++ u16 infhoff3; ++ u16 infhlen3; ++ u16 infgoff3; ++ u16 infglen3; ++ u8 reserved2[4]; ++}; ++ ++struct mac_sctn { ++ u8 infmflg1; ++ u8 infmflg2; ++ u8 infmval1; ++ u8 infmval2; ++ u16 infmscps; ++ u16 infmdcps; ++ u16 infmsifl; ++ u16 infmdifl; ++ char infmname[8]; ++ char infmtype[4]; ++ char infmmanu[16]; ++ char infmseq[16]; ++ char infmpman[4]; ++ u8 reserved[4]; ++}; ++ ++struct par_sctn { ++ u8 infpflg1; ++ u8 infpflg2; ++ u8 infpval1; ++ u8 infpval2; ++ u16 infppnum; ++ u16 infpscps; ++ u16 infpdcps; ++ u16 infpsifl; ++ u16 infpdifl; ++ u16 reserved; ++ char infppnam[8]; ++ u32 infpwbcp; ++ u32 infpabcp; ++ u32 infpwbif; ++ u32 infpabif; ++ char infplgnm[8]; ++ u32 infplgcp; ++ u32 infplgif; ++}; ++ ++struct sthyi_sctns { ++ struct hdr_sctn hdr; ++ struct mac_sctn mac; ++ struct par_sctn par; ++}; ++ ++struct cpu_inf { ++ u64 lpar_cap; ++ u64 lpar_grp_cap; ++ u64 lpar_weight; ++ u64 all_weight; ++ int cpu_num_ded; ++ int cpu_num_shd; ++}; ++ ++struct lpar_cpu_inf { ++ struct cpu_inf cp; ++ struct cpu_inf ifl; ++}; ++ ++struct sthyi_info { ++ void *info; ++ long unsigned int end; ++}; ++ ++typedef bool (*stack_trace_consume_fn)(void *, long unsigned int); ++ ++enum stack_type { ++ STACK_TYPE_UNKNOWN = 0, ++ STACK_TYPE_TASK = 1, ++ STACK_TYPE_IRQ = 2, ++ STACK_TYPE_NODAT = 3, ++ STACK_TYPE_RESTART = 4, ++ STACK_TYPE_MCCK = 5, ++}; ++ ++struct stack_info { ++ enum stack_type type; ++ long unsigned int begin; ++ long unsigned int end; ++}; ++ ++struct unwind_state { ++ struct stack_info stack_info; ++ long unsigned int stack_mask; ++ struct task_struct *task; ++ struct pt_regs *regs; ++ long unsigned int sp; ++ long unsigned int ip; ++ int graph_idx; ++ struct llist_node *kr_cur; ++ bool reliable; ++ bool error; ++}; ++ ++enum { ++ WORK_STRUCT_PENDING_BIT = 0, ++ WORK_STRUCT_INACTIVE_BIT = 1, ++ WORK_STRUCT_PWQ_BIT = 2, ++ WORK_STRUCT_LINKED_BIT = 3, ++ WORK_STRUCT_COLOR_SHIFT = 4, ++ WORK_STRUCT_COLOR_BITS = 4, ++ WORK_STRUCT_PENDING = 1, ++ WORK_STRUCT_INACTIVE = 2, ++ WORK_STRUCT_PWQ = 4, ++ WORK_STRUCT_LINKED = 8, ++ WORK_STRUCT_STATIC = 0, ++ WORK_NR_COLORS = 16, ++ WORK_CPU_UNBOUND = 512, ++ WORK_STRUCT_FLAG_BITS = 8, ++ WORK_OFFQ_FLAG_BASE = 4, ++ __WORK_OFFQ_CANCELING = 4, ++ WORK_OFFQ_FLAG_BITS = 1, ++ WORK_OFFQ_POOL_SHIFT = 5, ++ WORK_OFFQ_LEFT = 59, ++ WORK_OFFQ_POOL_BITS = 31, ++ WORK_BUSY_PENDING = 1, ++ WORK_BUSY_RUNNING = 2, ++ WORKER_DESC_LEN = 24, ++}; ++ ++enum { ++ SD_BALANCE_NEWIDLE = 1, ++ SD_BALANCE_EXEC = 2, ++ SD_BALANCE_FORK = 4, ++ SD_BALANCE_WAKE = 8, ++ SD_WAKE_AFFINE = 16, ++ SD_ASYM_CPUCAPACITY = 32, ++ SD_ASYM_CPUCAPACITY_FULL = 64, ++ SD_SHARE_CPUCAPACITY = 128, ++ SD_SHARE_PKG_RESOURCES = 256, ++ SD_SERIALIZE = 512, ++ SD_ASYM_PACKING = 1024, ++ SD_PREFER_SIBLING = 2048, ++ SD_OVERLAP = 4096, ++ SD_NUMA = 8192, ++}; ++ ++struct sched_domain_shared { ++ atomic_t ref; ++ atomic_t nr_busy_cpus; ++ int has_idle_cores; ++ int nr_idle_scan; ++}; ++ ++struct sched_group; ++ ++struct sched_domain { ++ struct sched_domain *parent; ++ struct sched_domain *child; ++ struct sched_group *groups; ++ long unsigned int min_interval; ++ long unsigned int max_interval; ++ unsigned int busy_factor; ++ unsigned int imbalance_pct; ++ unsigned int cache_nice_tries; ++ unsigned int imb_numa_nr; ++ int nohz_idle; ++ int flags; ++ int level; ++ long unsigned int last_balance; ++ unsigned int balance_interval; ++ unsigned int nr_balance_failed; ++ u64 max_newidle_lb_cost; ++ long unsigned int last_decay_max_lb_cost; ++ u64 avg_scan_cost; ++ unsigned int lb_count[3]; ++ unsigned int lb_failed[3]; ++ unsigned int lb_balanced[3]; ++ unsigned int lb_imbalance[3]; ++ unsigned int lb_gained[3]; ++ unsigned int lb_hot_gained[3]; ++ unsigned int lb_nobusyg[3]; ++ unsigned int lb_nobusyq[3]; ++ unsigned int alb_count; ++ unsigned int alb_failed; ++ unsigned int alb_pushed; ++ unsigned int sbe_count; ++ unsigned int sbe_balanced; ++ unsigned int sbe_pushed; ++ unsigned int sbf_count; ++ unsigned int sbf_balanced; ++ unsigned int sbf_pushed; ++ unsigned int ttwu_wake_remote; ++ unsigned int ttwu_move_affine; ++ unsigned int ttwu_move_balance; ++ char *name; ++ union { ++ void *private; ++ struct callback_head rcu; ++ }; ++ struct sched_domain_shared *shared; ++ unsigned int span_weight; ++ long unsigned int span[0]; ++}; ++ ++typedef const struct cpumask * (*sched_domain_mask_f)(int); ++ ++typedef int (*sched_domain_flags_f)(); ++ ++struct sched_group_capacity; ++ ++struct sd_data { ++ struct sched_domain **sd; ++ struct sched_domain_shared **sds; ++ struct sched_group **sg; ++ struct sched_group_capacity **sgc; ++}; ++ ++struct sched_domain_topology_level { ++ sched_domain_mask_f mask; ++ sched_domain_flags_f sd_flags; ++ int flags; ++ int numa_level; ++ struct sd_data data; ++ char *name; ++}; ++ ++enum kobject_action { ++ KOBJ_ADD = 0, ++ KOBJ_REMOVE = 1, ++ KOBJ_CHANGE = 2, ++ KOBJ_MOVE = 3, ++ KOBJ_ONLINE = 4, ++ KOBJ_OFFLINE = 5, ++ KOBJ_BIND = 6, ++ KOBJ_UNBIND = 7, ++}; ++ ++struct cpu { ++ int node_id; ++ int hotpluggable; ++ struct device dev; ++}; ++ ++struct topology_core { ++ unsigned char nl; ++ unsigned char reserved0[3]; ++ char: 5; ++ unsigned char d: 1; ++ unsigned char pp: 2; ++ unsigned char reserved1; ++ short unsigned int origin; ++ long unsigned int mask; ++}; ++ ++struct topology_container { ++ unsigned char nl; ++ unsigned char reserved[6]; ++ unsigned char id; ++}; ++ ++union topology_entry { ++ unsigned char nl; ++ struct topology_core cpu; ++ struct topology_container container; ++}; ++ ++struct sysinfo_15_1_x { ++ unsigned char reserved0[2]; ++ short unsigned int length; ++ unsigned char mag[6]; ++ unsigned char reserved1; ++ unsigned char mnest; ++ unsigned char reserved2[4]; ++ union topology_entry tle[0]; ++}; ++ ++enum { ++ TOPOLOGY_MODE_HW = 0, ++ TOPOLOGY_MODE_SINGLE = 1, ++ TOPOLOGY_MODE_PACKAGE = 2, ++ TOPOLOGY_MODE_UNINITIALIZED = 3, ++}; ++ ++struct mask_info { ++ struct mask_info *next; ++ unsigned char id; ++ cpumask_t mask; ++}; ++ ++enum auditsc_class_t { ++ AUDITSC_NATIVE = 0, ++ AUDITSC_COMPAT = 1, ++ AUDITSC_OPEN = 2, ++ AUDITSC_OPENAT = 3, ++ AUDITSC_SOCKETCALL = 4, ++ AUDITSC_EXECVE = 5, ++ AUDITSC_OPENAT2 = 6, ++ AUDITSC_NVALS = 7, ++}; ++ ++typedef unsigned int uint; ++ ++enum cons_flags { ++ CON_PRINTBUFFER = 1, ++ CON_CONSDEV = 2, ++ CON_ENABLED = 4, ++ CON_BOOT = 8, ++ CON_ANYTIME = 16, ++ CON_BRL = 32, ++ CON_EXTENDED = 64, ++}; ++ ++struct console { ++ char name[16]; ++ void (*write)(struct console *, const char *, unsigned int); ++ int (*read)(struct console *, char *, unsigned int); ++ struct tty_driver * (*device)(struct console *, int *); ++ void (*unblank)(); ++ int (*setup)(struct console *, char *); ++ int (*exit)(struct console *); ++ int (*match)(struct console *, char *, int, char *); ++ short int flags; ++ short int index; ++ int cflag; ++ uint ispeed; ++ uint ospeed; ++ u64 seq; ++ long unsigned int dropped; ++ void *data; ++ struct hlist_node node; ++}; ++ ++typedef u16 kprobe_opcode_t; ++ ++struct arch_specific_insn { ++ kprobe_opcode_t *insn; ++}; ++ ++struct kprobe; ++ ++struct prev_kprobe { ++ struct kprobe *kp; ++ long unsigned int status; ++}; ++ ++typedef int (*kprobe_pre_handler_t)(struct kprobe *, struct pt_regs *); ++ ++typedef void (*kprobe_post_handler_t)(struct kprobe *, struct pt_regs *, long unsigned int); ++ ++struct kprobe { ++ struct hlist_node hlist; ++ struct list_head list; ++ long unsigned int nmissed; ++ kprobe_opcode_t *addr; ++ const char *symbol_name; ++ unsigned int offset; ++ kprobe_pre_handler_t pre_handler; ++ kprobe_post_handler_t post_handler; ++ kprobe_opcode_t opcode; ++ struct arch_specific_insn ainsn; ++ u32 flags; ++}; ++ ++struct kprobe_ctlblk { ++ long unsigned int kprobe_status; ++ long unsigned int kprobe_saved_imask; ++ long unsigned int kprobe_saved_ctl[3]; ++ struct prev_kprobe prev_kprobe; ++}; ++ ++struct kretprobe_blackpoint { ++ const char *name; ++ void *addr; ++}; ++ ++struct kprobe_insn_cache { ++ struct mutex mutex; ++ void * (*alloc)(); ++ void (*free)(void *); ++ const char *sym; ++ struct list_head pages; ++ size_t insn_size; ++ int nr_garbage; ++}; ++ ++typedef int (*cpu_stop_fn_t)(void *); ++ ++struct die_args { ++ struct pt_regs *regs; ++ const char *str; ++ long int err; ++ int trapnr; ++ int signr; ++}; ++ ++enum { ++ _SET_MEMORY_RO_BIT = 0, ++ _SET_MEMORY_RW_BIT = 1, ++ _SET_MEMORY_NX_BIT = 2, ++ _SET_MEMORY_X_BIT = 3, ++ _SET_MEMORY_4K_BIT = 4, ++ _SET_MEMORY_INV_BIT = 5, ++ _SET_MEMORY_DEF_BIT = 6, ++}; ++ ++struct swap_insn_args { ++ struct kprobe *p; ++ unsigned int arm_kprobe: 1; ++}; ++ ++typedef struct { ++ char _[24]; ++} addrtype___2; ++ ++typedef unsigned int __kernel_uid_t; ++ ++typedef unsigned int __kernel_gid_t; ++ ++typedef struct { ++ __u32 fpc; ++ __u32 pad; ++ freg_t fprs[16]; ++} s390_fp_regs; ++ ++typedef struct { ++ psw_t psw; ++ long unsigned int gprs[16]; ++ unsigned int acrs[16]; ++ long unsigned int orig_gpr2; ++} s390_regs; ++ ++struct oldmem_data { ++ long unsigned int start; ++ long unsigned int size; ++}; ++ ++struct __kernel_old_timeval { ++ __kernel_long_t tv_sec; ++ __kernel_long_t tv_usec; ++}; ++ ++typedef s390_fp_regs elf_fpregset_t; ++ ++typedef s390_regs elf_gregset_t; ++ ++struct elf64_phdr { ++ Elf64_Word p_type; ++ Elf64_Word p_flags; ++ Elf64_Off p_offset; ++ Elf64_Addr p_vaddr; ++ Elf64_Addr p_paddr; ++ Elf64_Xword p_filesz; ++ Elf64_Xword p_memsz; ++ Elf64_Xword p_align; ++}; ++ ++typedef struct elf64_phdr Elf64_Phdr; ++ ++struct elf64_note { ++ Elf64_Word n_namesz; ++ Elf64_Word n_descsz; ++ Elf64_Word n_type; ++}; ++ ++typedef struct elf64_note Elf64_Nhdr; ++ ++struct elf_siginfo { ++ int si_signo; ++ int si_code; ++ int si_errno; ++}; ++ ++struct elf_prstatus_common { ++ struct elf_siginfo pr_info; ++ short int pr_cursig; ++ long unsigned int pr_sigpend; ++ long unsigned int pr_sighold; ++ pid_t pr_pid; ++ pid_t pr_ppid; ++ pid_t pr_pgrp; ++ pid_t pr_sid; ++ struct __kernel_old_timeval pr_utime; ++ struct __kernel_old_timeval pr_stime; ++ struct __kernel_old_timeval pr_cutime; ++ struct __kernel_old_timeval pr_cstime; ++}; ++ ++struct elf_prstatus { ++ struct elf_prstatus_common common; ++ elf_gregset_t pr_reg; ++ int pr_fpvalid; ++}; ++ ++struct elf_prpsinfo { ++ char pr_state; ++ char pr_sname; ++ char pr_zomb; ++ char pr_nice; ++ long unsigned int pr_flag; ++ __kernel_uid_t pr_uid; ++ __kernel_gid_t pr_gid; ++ pid_t pr_pid; ++ pid_t pr_ppid; ++ pid_t pr_pgrp; ++ pid_t pr_sid; ++ char pr_fname[16]; ++ char pr_psargs[80]; ++}; ++ ++enum memblock_flags { ++ MEMBLOCK_NONE = 0, ++ MEMBLOCK_HOTPLUG = 1, ++ MEMBLOCK_MIRROR = 2, ++ MEMBLOCK_NOMAP = 4, ++ MEMBLOCK_DRIVER_MANAGED = 8, ++}; ++ ++struct memblock_region { ++ phys_addr_t base; ++ phys_addr_t size; ++ enum memblock_flags flags; ++ int nid; ++}; ++ ++struct memblock_type { ++ long unsigned int cnt; ++ long unsigned int max; ++ phys_addr_t total_size; ++ struct memblock_region *regions; ++ char *name; ++}; ++ ++enum ipl_type { ++ IPL_TYPE_UNKNOWN = 1, ++ IPL_TYPE_CCW = 2, ++ IPL_TYPE_FCP = 4, ++ IPL_TYPE_FCP_DUMP = 8, ++ IPL_TYPE_NSS = 16, ++ IPL_TYPE_NVME = 32, ++ IPL_TYPE_NVME_DUMP = 64, ++ IPL_TYPE_ECKD = 128, ++ IPL_TYPE_ECKD_DUMP = 256, ++}; ++ ++struct ipl_info { ++ enum ipl_type type; ++ union { ++ struct { ++ struct ccw_dev_id dev_id; ++ } ccw; ++ struct { ++ struct ccw_dev_id dev_id; ++ } eckd; ++ struct { ++ struct ccw_dev_id dev_id; ++ u64 wwpn; ++ u64 lun; ++ } fcp; ++ struct { ++ u32 fid; ++ u32 nsid; ++ } nvme; ++ struct { ++ char name[9]; ++ } nss; ++ } data; ++}; ++ ++struct save_area { ++ struct list_head list; ++ u64 psw[2]; ++ u64 ctrs[16]; ++ u64 gprs[16]; ++ u32 acrs[16]; ++ u64 fprs[16]; ++ u32 fpc; ++ u32 prefix; ++ u32 todpreg; ++ u64 timer; ++ u64 todcmp; ++ u64 vxrs_low[16]; ++ __vector128 vxrs_high[16]; ++}; ++ ++enum jump_label_type { ++ JUMP_LABEL_NOP = 0, ++ JUMP_LABEL_JMP = 1, ++}; ++ ++struct insn { ++ u16 opcode; ++ s32 offset; ++} __attribute__((packed)); ++ ++struct parmarea { ++ long unsigned int ipl_device; ++ long unsigned int initrd_start; ++ long unsigned int initrd_size; ++ long unsigned int oldmem_base; ++ long unsigned int oldmem_size; ++ long unsigned int kernel_version; ++ long unsigned int max_command_line_size; ++ char pad1[72]; ++ char command_line[4096]; ++}; ++ ++typedef __s64 Elf64_Sxword; ++ ++struct elf64_rela { ++ Elf64_Addr r_offset; ++ Elf64_Xword r_info; ++ Elf64_Sxword r_addend; ++}; ++ ++typedef struct elf64_rela Elf64_Rela; ++ ++struct ipl_report; ++ ++struct s390_load_data { ++ void *kernel_buf; ++ long unsigned int kernel_mem; ++ struct parmarea *parm; ++ size_t memsz; ++ struct ipl_report *report; ++}; ++ ++struct ipl_parameter_block; ++ ++struct ipl_report { ++ struct ipl_parameter_block *ipib; ++ struct list_head components; ++ struct list_head certificates; ++ size_t size; ++}; ++ ++struct kimage_arch { ++ void *ipl_buf; ++}; ++ ++typedef int kexec_probe_t(const char *, long unsigned int); ++ ++struct kimage; ++ ++typedef void *kexec_load_t(struct kimage *, char *, long unsigned int, char *, long unsigned int, char *, long unsigned int); ++ ++typedef int kexec_cleanup_t(void *); ++ ++typedef int kexec_verify_sig_t(const char *, long unsigned int); ++ ++struct kexec_file_ops { ++ kexec_probe_t *probe; ++ kexec_load_t *load; ++ kexec_cleanup_t *cleanup; ++ kexec_verify_sig_t *verify_sig; ++}; ++ ++typedef long unsigned int kimage_entry_t; ++ ++struct kexec_segment { ++ union { ++ void *buf; ++ void *kbuf; ++ }; ++ size_t bufsz; ++ long unsigned int mem; ++ size_t memsz; ++}; ++ ++struct purgatory_info { ++ const Elf64_Ehdr *ehdr; ++ Elf64_Shdr *sechdrs; ++ void *purgatory_buf; ++}; ++ ++struct kimage { ++ kimage_entry_t head; ++ kimage_entry_t *entry; ++ kimage_entry_t *last_entry; ++ long unsigned int start; ++ struct page *control_code_page; ++ struct page *swap_page; ++ void *vmcoreinfo_data_copy; ++ long unsigned int nr_segments; ++ struct kexec_segment segment[16]; ++ struct list_head control_pages; ++ struct list_head dest_pages; ++ struct list_head unusable_pages; ++ long unsigned int control_page; ++ unsigned int type: 1; ++ unsigned int preserve_context: 1; ++ unsigned int file_mode: 1; ++ struct kimage_arch arch; ++ void *kernel_buf; ++ long unsigned int kernel_buf_len; ++ void *initrd_buf; ++ long unsigned int initrd_buf_len; ++ char *cmdline_buf; ++ long unsigned int cmdline_buf_len; ++ const struct kexec_file_ops *fops; ++ void *image_loader_data; ++ struct purgatory_info purgatory_info; ++ void *elf_headers; ++ long unsigned int elf_headers_sz; ++ long unsigned int elf_load_addr; ++}; ++ ++struct kexec_buf { ++ struct kimage *image; ++ void *buffer; ++ long unsigned int bufsz; ++ long unsigned int mem; ++ long unsigned int memsz; ++ long unsigned int buf_align; ++ long unsigned int buf_min; ++ long unsigned int buf_max; ++ bool top_down; ++}; ++ ++enum pkey_id_type { ++ PKEY_ID_PGP = 0, ++ PKEY_ID_X509 = 1, ++ PKEY_ID_PKCS7 = 2, ++}; ++ ++struct module_signature { ++ u8 algo; ++ u8 hash; ++ u8 id_type; ++ u8 signer_len; ++ u8 key_id_len; ++ u8 __pad[3]; ++ __be32 sig_len; ++}; ++ ++struct ipl_pl_hdr { ++ __u32 len; ++ __u8 flags; ++ __u8 reserved1[2]; ++ __u8 version; ++}; ++ ++struct ipl_pb_hdr { ++ __u32 len; ++ __u8 pbt; ++} __attribute__((packed)); ++ ++struct ipl_pb0_common { ++ __u32 len; ++ __u8 pbt; ++ __u8 flags; ++ __u8 reserved1[2]; ++ __u8 loadparm[8]; ++ __u8 reserved2[84]; ++}; ++ ++struct ipl_pb0_fcp { ++ __u32 len; ++ __u8 pbt; ++ __u8 reserved1[3]; ++ __u8 loadparm[8]; ++ __u8 reserved2[304]; ++ __u8 opt; ++ __u8 reserved3[3]; ++ __u8 cssid; ++ __u8 reserved4[1]; ++ __u16 devno; ++ __u8 reserved5[4]; ++ __u64 wwpn; ++ __u64 lun; ++ __u32 bootprog; ++ __u8 reserved6[12]; ++ __u64 br_lba; ++ __u32 scp_data_len; ++ __u8 reserved7[260]; ++ __u8 scp_data[0]; ++} __attribute__((packed)); ++ ++struct ipl_pb0_nvme { ++ __u32 len; ++ __u8 pbt; ++ __u8 reserved1[3]; ++ __u8 loadparm[8]; ++ __u8 reserved2[304]; ++ __u8 opt; ++ __u8 reserved3[3]; ++ __u32 fid; ++ __u8 reserved4[12]; ++ __u32 nsid; ++ __u8 reserved5[4]; ++ __u32 bootprog; ++ __u8 reserved6[12]; ++ __u64 br_lba; ++ __u32 scp_data_len; ++ __u8 reserved7[260]; ++ __u8 scp_data[0]; ++} __attribute__((packed)); ++ ++struct ipl_pb0_ccw { ++ __u32 len; ++ __u8 pbt; ++ __u8 flags; ++ __u8 reserved1[2]; ++ __u8 loadparm[8]; ++ __u8 reserved2[84]; ++ __u16 reserved3: 13; ++ __u8 ssid: 3; ++ __u16 devno; ++ __u8 vm_flags; ++ __u8 reserved4[3]; ++ __u32 vm_parm_len; ++ __u8 nss_name[8]; ++ __u8 vm_parm[64]; ++ __u8 reserved5[8]; ++}; ++ ++struct ipl_pb0_eckd { ++ __u32 len; ++ __u8 pbt; ++ __u8 reserved1[3]; ++ __u32 reserved2[78]; ++ __u8 opt; ++ __u8 reserved4[4]; ++ __u8 reserved5: 5; ++ __u8 ssid: 3; ++ __u16 devno; ++ __u32 reserved6[5]; ++ __u32 bootprog; ++ __u8 reserved7[12]; ++ struct { ++ __u16 cyl; ++ __u8 head; ++ __u8 record; ++ __u32 reserved; ++ } br_chr; ++ __u32 scp_data_len; ++ __u8 reserved8[260]; ++ __u8 scp_data[0]; ++}; ++ ++struct ipl_parameter_block { ++ struct ipl_pl_hdr hdr; ++ union { ++ struct ipl_pb_hdr pb0_hdr; ++ struct ipl_pb0_common common; ++ struct ipl_pb0_fcp fcp; ++ struct ipl_pb0_ccw ccw; ++ struct ipl_pb0_eckd eckd; ++ struct ipl_pb0_nvme nvme; ++ char raw[4088]; ++ }; ++}; ++ ++struct perf_pmu_events_attr { ++ struct device_attribute attr; ++ u64 id; ++ const char *event_str; ++}; ++ ++struct cpumf_ctr_info { ++ u16 cfvn; ++ u16 auth_ctl; ++ u16 enable_ctl; ++ u16 act_ctl; ++ u16 max_cpu; ++ u16 csvn; ++ u16 max_cg; ++ u16 reserved1; ++ u32 reserved2[12]; ++}; ++ ++enum perf_event_task_context { ++ perf_invalid_context = -1, ++ perf_hw_context = 0, ++ perf_sw_context = 1, ++ perf_nr_task_contexts = 2, ++}; ++ ++enum perf_type_id { ++ PERF_TYPE_HARDWARE = 0, ++ PERF_TYPE_SOFTWARE = 1, ++ PERF_TYPE_TRACEPOINT = 2, ++ PERF_TYPE_HW_CACHE = 3, ++ PERF_TYPE_RAW = 4, ++ PERF_TYPE_BREAKPOINT = 5, ++ PERF_TYPE_MAX = 6, ++}; ++ ++enum perf_event_sample_format { ++ PERF_SAMPLE_IP = 1, ++ PERF_SAMPLE_TID = 2, ++ PERF_SAMPLE_TIME = 4, ++ PERF_SAMPLE_ADDR = 8, ++ PERF_SAMPLE_READ = 16, ++ PERF_SAMPLE_CALLCHAIN = 32, ++ PERF_SAMPLE_ID = 64, ++ PERF_SAMPLE_CPU = 128, ++ PERF_SAMPLE_PERIOD = 256, ++ PERF_SAMPLE_STREAM_ID = 512, ++ PERF_SAMPLE_RAW = 1024, ++ PERF_SAMPLE_BRANCH_STACK = 2048, ++ PERF_SAMPLE_REGS_USER = 4096, ++ PERF_SAMPLE_STACK_USER = 8192, ++ PERF_SAMPLE_WEIGHT = 16384, ++ PERF_SAMPLE_DATA_SRC = 32768, ++ PERF_SAMPLE_IDENTIFIER = 65536, ++ PERF_SAMPLE_TRANSACTION = 131072, ++ PERF_SAMPLE_REGS_INTR = 262144, ++ PERF_SAMPLE_PHYS_ADDR = 524288, ++ PERF_SAMPLE_AUX = 1048576, ++ PERF_SAMPLE_CGROUP = 2097152, ++ PERF_SAMPLE_DATA_PAGE_SIZE = 4194304, ++ PERF_SAMPLE_CODE_PAGE_SIZE = 8388608, ++ PERF_SAMPLE_WEIGHT_STRUCT = 16777216, ++ PERF_SAMPLE_MAX = 33554432, ++}; ++ ++struct qpaci_info_block { ++ u64 header; ++ struct { ++ char: 8; ++ u64 num_cc: 8; ++ short: 9; ++ u64 num_nnpa: 7; ++ }; ++}; ++ ++enum paievt_mode { ++ PAI_MODE_NONE = 0, ++ PAI_MODE_SAMPLING = 1, ++ PAI_MODE_COUNTING = 2, ++}; ++ ++struct pai_userdata { ++ u16 num; ++ u64 value; ++} __attribute__((packed)); ++ ++struct paiext_cb { ++ u64 header; ++ u64 reserved1; ++ u64 acc; ++ u8 reserved2[488]; ++}; ++ ++struct paiext_map { ++ long unsigned int *area; ++ struct pai_userdata *save; ++ enum paievt_mode mode; ++ unsigned int active_events; ++ unsigned int refcnt; ++ struct perf_event *event; ++ struct paiext_cb *paiext_cb; ++}; ++ ++struct paiext_mapptr { ++ struct paiext_map *mapptr; ++}; ++ ++struct paiext_root { ++ int refcnt; ++ struct paiext_mapptr *mapptr; ++}; ++ ++typedef int pcpu_fc_cpu_to_node_fn_t(int); ++ ++typedef int pcpu_fc_cpu_distance_fn_t(unsigned int, unsigned int); ++ ++struct mhp_params { ++ struct vmem_altmap *altmap; ++ pgprot_t pgprot; ++ struct dev_pagemap *pgmap; ++}; ++ ++struct uv_cb_header { ++ u16 len; ++ u16 cmd; ++ u16 rc; ++ u16 rrc; ++}; ++ ++struct uv_cb_share { ++ struct uv_cb_header header; ++ u64 reserved08[3]; ++ u64 paddr; ++ u64 reserved28; ++}; ++ ++enum { ++ TASKSTATS_CMD_UNSPEC = 0, ++ TASKSTATS_CMD_GET = 1, ++ TASKSTATS_CMD_NEW = 2, ++ __TASKSTATS_CMD_MAX = 3, ++}; ++ ++struct irq_affinity { ++ unsigned int pre_vectors; ++ unsigned int post_vectors; ++ unsigned int nr_sets; ++ unsigned int set_size[4]; ++ void (*calc_sets)(struct irq_affinity *, unsigned int); ++ void *priv; ++}; ++ ++enum cpu_usage_stat { ++ CPUTIME_USER = 0, ++ CPUTIME_NICE = 1, ++ CPUTIME_SYSTEM = 2, ++ CPUTIME_SOFTIRQ = 3, ++ CPUTIME_IRQ = 4, ++ CPUTIME_IDLE = 5, ++ CPUTIME_IOWAIT = 6, ++ CPUTIME_STEAL = 7, ++ CPUTIME_GUEST = 8, ++ CPUTIME_GUEST_NICE = 9, ++ CPUTIME_FORCEIDLE = 10, ++ NR_STATS = 11, ++}; ++ ++enum cgroup_bpf_attach_type { ++ CGROUP_BPF_ATTACH_TYPE_INVALID = -1, ++ CGROUP_INET_INGRESS = 0, ++ CGROUP_INET_EGRESS = 1, ++ CGROUP_INET_SOCK_CREATE = 2, ++ CGROUP_SOCK_OPS = 3, ++ CGROUP_DEVICE = 4, ++ CGROUP_INET4_BIND = 5, ++ CGROUP_INET6_BIND = 6, ++ CGROUP_INET4_CONNECT = 7, ++ CGROUP_INET6_CONNECT = 8, ++ CGROUP_INET4_POST_BIND = 9, ++ CGROUP_INET6_POST_BIND = 10, ++ CGROUP_UDP4_SENDMSG = 11, ++ CGROUP_UDP6_SENDMSG = 12, ++ CGROUP_SYSCTL = 13, ++ CGROUP_UDP4_RECVMSG = 14, ++ CGROUP_UDP6_RECVMSG = 15, ++ CGROUP_GETSOCKOPT = 16, ++ CGROUP_SETSOCKOPT = 17, ++ CGROUP_INET4_GETPEERNAME = 18, ++ CGROUP_INET6_GETPEERNAME = 19, ++ CGROUP_INET4_GETSOCKNAME = 20, ++ CGROUP_INET6_GETSOCKNAME = 21, ++ CGROUP_INET_SOCK_RELEASE = 22, ++ CGROUP_LSM_START = 23, ++ CGROUP_LSM_END = 32, ++ MAX_CGROUP_BPF_ATTACH_TYPE = 33, ++}; ++ ++enum psi_task_count { ++ NR_IOWAIT = 0, ++ NR_MEMSTALL = 1, ++ NR_RUNNING = 2, ++ NR_MEMSTALL_RUNNING = 3, ++ NR_PSI_TASK_COUNTS = 4, ++}; ++ ++enum psi_res { ++ PSI_IO = 0, ++ PSI_MEM = 1, ++ PSI_CPU = 2, ++ NR_PSI_RESOURCES = 3, ++}; ++ ++enum psi_states { ++ PSI_IO_SOME = 0, ++ PSI_IO_FULL = 1, ++ PSI_MEM_SOME = 2, ++ PSI_MEM_FULL = 3, ++ PSI_CPU_SOME = 4, ++ PSI_CPU_FULL = 5, ++ PSI_NONIDLE = 6, ++ NR_PSI_STATES = 7, ++}; ++ ++enum psi_aggregators { ++ PSI_AVGS = 0, ++ PSI_POLL = 1, ++ NR_PSI_AGGREGATORS = 2, ++}; ++ ++enum cgroup_subsys_id { ++ cpuset_cgrp_id = 0, ++ cpu_cgrp_id = 1, ++ cpuacct_cgrp_id = 2, ++ io_cgrp_id = 3, ++ memory_cgrp_id = 4, ++ devices_cgrp_id = 5, ++ freezer_cgrp_id = 6, ++ net_cls_cgrp_id = 7, ++ perf_event_cgrp_id = 8, ++ net_prio_cgrp_id = 9, ++ pids_cgrp_id = 10, ++ rdma_cgrp_id = 11, ++ misc_cgrp_id = 12, ++ CGROUP_SUBSYS_COUNT = 13, ++}; ++ ++enum wb_stat_item { ++ WB_RECLAIMABLE = 0, ++ WB_WRITEBACK = 1, ++ WB_DIRTIED = 2, ++ WB_WRITTEN = 3, ++ NR_WB_STAT_ITEMS = 4, ++}; ++ ++struct io_tlb_area; ++ ++struct io_tlb_slot; ++ ++struct io_tlb_mem { ++ phys_addr_t start; ++ phys_addr_t end; ++ void *vaddr; ++ long unsigned int nslabs; ++ long unsigned int used; ++ struct dentry *debugfs; ++ bool late_alloc; ++ bool force_bounce; ++ bool for_alloc; ++ unsigned int nareas; ++ unsigned int area_nslabs; ++ struct io_tlb_area *areas; ++ struct io_tlb_slot *slots; ++ atomic_long_t total_used; ++ atomic_long_t used_hiwater; ++}; ++ ++enum memcg_memory_event { ++ MEMCG_LOW = 0, ++ MEMCG_HIGH = 1, ++ MEMCG_MAX = 2, ++ MEMCG_OOM = 3, ++ MEMCG_OOM_KILL = 4, ++ MEMCG_OOM_GROUP_KILL = 5, ++ MEMCG_SWAP_HIGH = 6, ++ MEMCG_SWAP_MAX = 7, ++ MEMCG_SWAP_FAIL = 8, ++ MEMCG_NR_MEMORY_EVENTS = 9, ++}; ++ ++struct memory_notify { ++ long unsigned int start_pfn; ++ long unsigned int nr_pages; ++ int status_change_nid_normal; ++ int status_change_nid; ++}; ++ ++struct scatterlist { ++ long unsigned int page_link; ++ unsigned int offset; ++ unsigned int length; ++ dma_addr_t dma_address; ++ unsigned int dma_length; ++}; ++ ++struct sg_table { ++ struct scatterlist *sgl; ++ unsigned int nents; ++ unsigned int orig_nents; ++}; ++ ++struct virtio_device_id { ++ __u32 device; ++ __u32 vendor; ++}; ++ ++struct vringh_config_ops; ++ ++struct virtio_config_ops; ++ ++struct virtio_device { ++ int index; ++ bool failed; ++ bool config_enabled; ++ bool config_change_pending; ++ spinlock_t config_lock; ++ spinlock_t vqs_list_lock; ++ struct device dev; ++ struct virtio_device_id id; ++ const struct virtio_config_ops *config; ++ const struct vringh_config_ops *vringh_config; ++ struct list_head vqs; ++ u64 features; ++ void *priv; ++}; ++ ++struct virtqueue { ++ struct list_head list; ++ void (*callback)(struct virtqueue *); ++ const char *name; ++ struct virtio_device *vdev; ++ unsigned int index; ++ unsigned int num_free; ++ unsigned int num_max; ++ bool reset; ++ void *priv; ++}; ++ ++typedef void vq_callback_t(struct virtqueue *); ++ ++struct virtio_shm_region; ++ ++struct virtio_config_ops { ++ void (*get)(struct virtio_device *, unsigned int, void *, unsigned int); ++ void (*set)(struct virtio_device *, unsigned int, const void *, unsigned int); ++ u32 (*generation)(struct virtio_device *); ++ u8 (*get_status)(struct virtio_device *); ++ void (*set_status)(struct virtio_device *, u8); ++ void (*reset)(struct virtio_device *); ++ int (*find_vqs)(struct virtio_device *, unsigned int, struct virtqueue **, vq_callback_t **, const char * const *, const bool *, struct irq_affinity *); ++ void (*del_vqs)(struct virtio_device *); ++ void (*synchronize_cbs)(struct virtio_device *); ++ u64 (*get_features)(struct virtio_device *); ++ int (*finalize_features)(struct virtio_device *); ++ const char * (*bus_name)(struct virtio_device *); ++ int (*set_vq_affinity)(struct virtqueue *, const struct cpumask *); ++ const struct cpumask * (*get_vq_affinity)(struct virtio_device *, int); ++ bool (*get_shm_region)(struct virtio_device *, struct virtio_shm_region *, u8); ++ int (*disable_vq_and_reset)(struct virtqueue *); ++ int (*enable_vq_after_reset)(struct virtqueue *); ++}; ++ ++struct virtio_shm_region { ++ u64 addr; ++ u64 len; ++}; ++ ++struct s390_cma_mem_data { ++ long unsigned int start; ++ long unsigned int end; ++}; ++ ++struct cma_kobject; ++ ++struct cma { ++ long unsigned int base_pfn; ++ long unsigned int count; ++ long unsigned int *bitmap; ++ unsigned int order_per_bit; ++ spinlock_t lock; ++ char name[64]; ++ atomic64_t nr_pages_succeeded; ++ atomic64_t nr_pages_failed; ++ struct cma_kobject *cma_kobj; ++ bool reserve_pages_on_error; ++}; ++ ++typedef struct { ++ long unsigned int p4d; ++} p4d_t; ++ ++typedef struct { ++ long unsigned int pgste; ++} pgste_t; ++ ++struct anon_vma { ++ struct anon_vma *root; ++ struct rw_semaphore rwsem; ++ atomic_t refcount; ++ long unsigned int num_children; ++ long unsigned int num_active_vmas; ++ struct anon_vma *parent; ++ struct rb_root_cached rb_root; ++}; ++ ++typedef struct { ++ long unsigned int val; ++} swp_entry_t; ++ ++typedef struct { ++ int val[2]; ++} __kernel_fsid_t; ++ ++struct kstatfs { ++ long int f_type; ++ long int f_bsize; ++ u64 f_blocks; ++ u64 f_bfree; ++ u64 f_bavail; ++ u64 f_files; ++ u64 f_ffree; ++ __kernel_fsid_t f_fsid; ++ long int f_namelen; ++ long int f_frsize; ++ long int f_flags; ++ long int f_spare[4]; ++}; ++ ++struct hypfs_sb_info { ++ kuid_t uid; ++ kgid_t gid; ++ struct dentry *update_file; ++ time64_t last_update; ++ struct mutex lock; ++}; ++ ++enum { ++ Opt_uid = 0, ++ Opt_gid = 1, ++}; ++ ++struct hypfs_dbfs_file; ++ ++struct hypfs_dbfs_data { ++ void *buf; ++ void *buf_free_ptr; ++ size_t size; ++ struct hypfs_dbfs_file *dbfs_file; ++}; ++ ++struct hypfs_dbfs_file { ++ const char *name; ++ int (*data_create)(void **, void **, size_t *); ++ void (*data_free)(const void *); ++ long int (*unlocked_ioctl)(struct file *, unsigned int, long unsigned int); ++ struct mutex lock; ++ struct dentry *dentry; ++}; ++ ++struct sysinfo { ++ __kernel_long_t uptime; ++ __kernel_ulong_t loads[3]; ++ __kernel_ulong_t totalram; ++ __kernel_ulong_t freeram; ++ __kernel_ulong_t sharedram; ++ __kernel_ulong_t bufferram; ++ __kernel_ulong_t totalswap; ++ __kernel_ulong_t freeswap; ++ __u16 procs; ++ __u16 pad; ++ __kernel_ulong_t totalhigh; ++ __kernel_ulong_t freehigh; ++ __u32 mem_unit; ++ char _f[0]; ++}; ++ ++enum { ++ WQ_UNBOUND = 2, ++ WQ_FREEZABLE = 4, ++ WQ_MEM_RECLAIM = 8, ++ WQ_HIGHPRI = 16, ++ WQ_CPU_INTENSIVE = 32, ++ WQ_SYSFS = 64, ++ WQ_POWER_EFFICIENT = 128, ++ __WQ_DESTROYING = 32768, ++ __WQ_DRAINING = 65536, ++ __WQ_ORDERED = 131072, ++ __WQ_LEGACY = 262144, ++ __WQ_ORDERED_EXPLICIT = 524288, ++ WQ_MAX_ACTIVE = 512, ++ WQ_MAX_UNBOUND_PER_CPU = 4, ++ WQ_DFL_ACTIVE = 256, ++}; ++ ++struct appldata_parameter_list { ++ u16 diag; ++ u8 function; ++ u8 parlist_length; ++ u32 unused01; ++ u16 reserved; ++ u16 buffer_length; ++ u32 unused02; ++ u64 product_id_addr; ++ u64 buffer_addr; ++}; ++ ++struct appldata_product_id { ++ char prod_nr[7]; ++ u16 prod_fn; ++ u8 record_nr; ++ u16 version_nr; ++ u16 release_nr; ++ u16 mod_lvl; ++} __attribute__((packed)); ++ ++struct vtimer_list { ++ struct list_head entry; ++ u64 expires; ++ u64 interval; ++ void (*function)(long unsigned int); ++ long unsigned int data; ++}; ++ ++struct appldata_ops { ++ struct list_head list; ++ struct ctl_table_header *sysctl_header; ++ struct ctl_table *ctl_table; ++ int active; ++ char name[16]; ++ unsigned char record_nr; ++ void (*callback)(void *); ++ void *data; ++ unsigned int size; ++ struct module *owner; ++ char mod_lvl[2]; ++}; ++ ++struct __call_single_data { ++ struct __call_single_node node; ++ smp_call_func_t func; ++ void *info; ++}; ++ ++typedef struct __call_single_data call_single_data_t; ++ ++struct irq_desc; ++ ++typedef void (*irq_flow_handler_t)(struct irq_desc *); ++ ++struct msi_desc; ++ ++struct irq_common_data { ++ unsigned int state_use_accessors; ++ unsigned int node; ++ void *handler_data; ++ struct msi_desc *msi_desc; ++ cpumask_var_t affinity; ++}; ++ ++struct irq_chip; ++ ++struct irq_data { ++ u32 mask; ++ unsigned int irq; ++ long unsigned int hwirq; ++ struct irq_common_data *common; ++ struct irq_chip *chip; ++ struct irq_domain *domain; ++ struct irq_data *parent_data; ++ void *chip_data; ++}; ++ ++struct irqaction; ++ ++struct irq_affinity_notify; ++ ++struct irq_desc { ++ struct irq_common_data irq_common_data; ++ struct irq_data irq_data; ++ unsigned int *kstat_irqs; ++ irq_flow_handler_t handle_irq; ++ struct irqaction *action; ++ unsigned int status_use_accessors; ++ unsigned int core_internal_state__do_not_mess_with_it; ++ unsigned int depth; ++ unsigned int wake_depth; ++ unsigned int tot_count; ++ unsigned int irq_count; ++ long unsigned int last_unhandled; ++ unsigned int irqs_unhandled; ++ atomic_t threads_handled; ++ int threads_handled_last; ++ raw_spinlock_t lock; ++ struct cpumask *percpu_enabled; ++ const struct cpumask *percpu_affinity; ++ const struct cpumask *affinity_hint; ++ struct irq_affinity_notify *affinity_notify; ++ long unsigned int threads_oneshot; ++ atomic_t threads_active; ++ wait_queue_head_t wait_for_threads; ++ struct proc_dir_entry *dir; ++ struct callback_head rcu; ++ struct kobject kobj; ++ struct mutex request_mutex; ++ int parent_irq; ++ struct module *owner; ++ const char *name; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum irqreturn { ++ IRQ_NONE = 0, ++ IRQ_HANDLED = 1, ++ IRQ_WAKE_THREAD = 2, ++}; ++ ++typedef enum irqreturn irqreturn_t; ++ ++struct zpci_fib_fmt0 { ++ char: 1; ++ u32 isc: 3; ++ u32 noi: 12; ++ char: 2; ++ u32 aibvo: 6; ++ u32 sum: 1; ++ char: 1; ++ u32 aisbo: 6; ++ u64 aibv; ++ u64 aisb; ++}; ++ ++struct zpci_fib_fmt1 { ++ char: 4; ++ u32 noi: 12; ++ int: 16; ++ u32 dibvo: 16; ++ long: 64; ++ long: 64; ++}; ++ ++struct zpci_fib { ++ u32 fmt: 8; ++ long: 0; ++ u8 fc; ++ u64 pba; ++ u64 pal; ++ u64 iota; ++ union { ++ struct zpci_fib_fmt0 fmt0; ++ struct zpci_fib_fmt1 fmt1; ++ }; ++ u64 fmb_addr; ++ int: 32; ++ u32 gd; ++}; ++ ++struct zpci_diib { ++ char: 1; ++ u32 isc: 3; ++ int: 0; ++ short: 16; ++ u16 nr_cpus; ++ u64 disb_addr; ++ long: 64; ++ long: 64; ++}; ++ ++struct zpci_cdiib { ++ long: 64; ++ u64 dibv_addr; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct zpci_aipb { ++ u64 faisb; ++ u64 gait; ++ short: 13; ++ u16 afi: 3; ++ int: 0; ++ short: 16; ++ u16 faal; ++}; ++ ++union zpci_sic_iib { ++ struct zpci_diib diib; ++ struct zpci_cdiib cdiib; ++ struct zpci_aipb aipb; ++}; ++ ++enum { ++ IRQ_SET_MASK_OK = 0, ++ IRQ_SET_MASK_OK_NOCOPY = 1, ++ IRQ_SET_MASK_OK_DONE = 2, ++}; ++ ++struct arch_msi_msg_addr_lo { ++ u32 address_lo; ++}; ++ ++typedef struct arch_msi_msg_addr_lo arch_msi_msg_addr_lo_t; ++ ++struct arch_msi_msg_addr_hi { ++ u32 address_hi; ++}; ++ ++typedef struct arch_msi_msg_addr_hi arch_msi_msg_addr_hi_t; ++ ++struct arch_msi_msg_data { ++ u32 data; ++}; ++ ++typedef struct arch_msi_msg_data arch_msi_msg_data_t; ++ ++struct msi_msg { ++ union { ++ u32 address_lo; ++ arch_msi_msg_addr_lo_t arch_addr_lo; ++ }; ++ union { ++ u32 address_hi; ++ arch_msi_msg_addr_hi_t arch_addr_hi; ++ }; ++ union { ++ u32 data; ++ arch_msi_msg_data_t arch_data; ++ }; ++}; ++ ++struct pci_msi_desc { ++ union { ++ u32 msi_mask; ++ u32 msix_ctrl; ++ }; ++ struct { ++ u8 is_msix: 1; ++ u8 multiple: 3; ++ u8 multi_cap: 3; ++ u8 can_mask: 1; ++ u8 is_64: 1; ++ u8 is_virtual: 1; ++ unsigned int default_irq; ++ } msi_attrib; ++ union { ++ u8 mask_pos; ++ void *mask_base; ++ }; ++}; ++ ++union msi_domain_cookie { ++ u64 value; ++ void *ptr; ++ void *iobase; ++}; ++ ++union msi_instance_cookie { ++ u64 value; ++ void *ptr; ++}; ++ ++struct msi_desc_data { ++ union msi_domain_cookie dcookie; ++ union msi_instance_cookie icookie; ++}; ++ ++struct irq_affinity_desc; ++ ++struct msi_desc { ++ unsigned int irq; ++ unsigned int nvec_used; ++ struct device *dev; ++ struct msi_msg msg; ++ struct irq_affinity_desc *affinity; ++ struct device_attribute *sysfs_attrs; ++ void (*write_msi_msg)(struct msi_desc *, void *); ++ void *write_msi_msg_data; ++ u16 msi_index; ++ union { ++ struct pci_msi_desc pci; ++ struct msi_desc_data data; ++ }; ++}; ++ ++enum irqchip_irq_state { ++ IRQCHIP_STATE_PENDING = 0, ++ IRQCHIP_STATE_ACTIVE = 1, ++ IRQCHIP_STATE_MASKED = 2, ++ IRQCHIP_STATE_LINE_LEVEL = 3, ++}; ++ ++struct irq_chip { ++ const char *name; ++ unsigned int (*irq_startup)(struct irq_data *); ++ void (*irq_shutdown)(struct irq_data *); ++ void (*irq_enable)(struct irq_data *); ++ void (*irq_disable)(struct irq_data *); ++ void (*irq_ack)(struct irq_data *); ++ void (*irq_mask)(struct irq_data *); ++ void (*irq_mask_ack)(struct irq_data *); ++ void (*irq_unmask)(struct irq_data *); ++ void (*irq_eoi)(struct irq_data *); ++ int (*irq_set_affinity)(struct irq_data *, const struct cpumask *, bool); ++ int (*irq_retrigger)(struct irq_data *); ++ int (*irq_set_type)(struct irq_data *, unsigned int); ++ int (*irq_set_wake)(struct irq_data *, unsigned int); ++ void (*irq_bus_lock)(struct irq_data *); ++ void (*irq_bus_sync_unlock)(struct irq_data *); ++ void (*irq_suspend)(struct irq_data *); ++ void (*irq_resume)(struct irq_data *); ++ void (*irq_pm_shutdown)(struct irq_data *); ++ void (*irq_calc_mask)(struct irq_data *); ++ void (*irq_print_chip)(struct irq_data *, struct seq_file *); ++ int (*irq_request_resources)(struct irq_data *); ++ void (*irq_release_resources)(struct irq_data *); ++ void (*irq_compose_msi_msg)(struct irq_data *, struct msi_msg *); ++ void (*irq_write_msi_msg)(struct irq_data *, struct msi_msg *); ++ int (*irq_get_irqchip_state)(struct irq_data *, enum irqchip_irq_state, bool *); ++ int (*irq_set_irqchip_state)(struct irq_data *, enum irqchip_irq_state, bool); ++ int (*irq_set_vcpu_affinity)(struct irq_data *, void *); ++ void (*ipi_send_single)(struct irq_data *, unsigned int); ++ void (*ipi_send_mask)(struct irq_data *, const struct cpumask *); ++ int (*irq_nmi_setup)(struct irq_data *); ++ void (*irq_nmi_teardown)(struct irq_data *); ++ long unsigned int flags; ++}; ++ ++typedef irqreturn_t (*irq_handler_t)(int, void *); ++ ++struct irqaction { ++ irq_handler_t handler; ++ void *dev_id; ++ void *percpu_dev_id; ++ struct irqaction *next; ++ irq_handler_t thread_fn; ++ struct task_struct *thread; ++ struct irqaction *secondary; ++ unsigned int irq; ++ unsigned int flags; ++ long unsigned int thread_flags; ++ long unsigned int thread_mask; ++ const char *name; ++ struct proc_dir_entry *dir; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct irq_affinity_notify { ++ unsigned int irq; ++ struct kref kref; ++ struct work_struct work; ++ void (*notify)(struct irq_affinity_notify *, const cpumask_t *); ++ void (*release)(struct kref *); ++}; ++ ++struct irq_affinity_desc { ++ struct cpumask mask; ++ unsigned int is_managed: 1; ++}; ++ ++enum msi_desc_filter { ++ MSI_DESC_ALL = 0, ++ MSI_DESC_NOTASSOCIATED = 1, ++ MSI_DESC_ASSOCIATED = 2, ++}; ++ ++struct pci_device_id { ++ __u32 vendor; ++ __u32 device; ++ __u32 subvendor; ++ __u32 subdevice; ++ __u32 class; ++ __u32 class_mask; ++ kernel_ulong_t driver_data; ++ __u32 override_only; ++}; ++ ++struct pci_bus; ++ ++struct hotplug_slot; ++ ++struct pci_slot { ++ struct pci_bus *bus; ++ struct list_head list; ++ struct hotplug_slot *hotplug; ++ unsigned char number; ++ struct kobject kobj; ++}; ++ ++struct pci_dev; ++ ++struct pci_ops; ++ ++struct pci_bus { ++ struct list_head node; ++ struct pci_bus *parent; ++ struct list_head children; ++ struct list_head devices; ++ struct pci_dev *self; ++ struct list_head slots; ++ struct resource *resource[4]; ++ struct list_head resources; ++ struct resource busn_res; ++ struct pci_ops *ops; ++ void *sysdata; ++ struct proc_dir_entry *procdir; ++ unsigned char number; ++ unsigned char primary; ++ unsigned char max_bus_speed; ++ unsigned char cur_bus_speed; ++ char name[48]; ++ short unsigned int bridge_ctl; ++ pci_bus_flags_t bus_flags; ++ struct device *bridge; ++ struct device dev; ++ struct bin_attribute *legacy_io; ++ struct bin_attribute *legacy_mem; ++ unsigned int is_added: 1; ++ unsigned int unsafe_warn: 1; ++}; ++ ++struct hotplug_slot_ops; ++ ++struct hotplug_slot { ++ const struct hotplug_slot_ops *ops; ++ struct list_head slot_list; ++ struct pci_slot *pci_slot; ++ struct module *owner; ++ const char *mod_name; ++}; ++ ++typedef int pci_power_t; ++ ++enum pci_bus_speed { ++ PCI_SPEED_33MHz = 0, ++ PCI_SPEED_66MHz = 1, ++ PCI_SPEED_66MHz_PCIX = 2, ++ PCI_SPEED_100MHz_PCIX = 3, ++ PCI_SPEED_133MHz_PCIX = 4, ++ PCI_SPEED_66MHz_PCIX_ECC = 5, ++ PCI_SPEED_100MHz_PCIX_ECC = 6, ++ PCI_SPEED_133MHz_PCIX_ECC = 7, ++ PCI_SPEED_66MHz_PCIX_266 = 9, ++ PCI_SPEED_100MHz_PCIX_266 = 10, ++ PCI_SPEED_133MHz_PCIX_266 = 11, ++ AGP_UNKNOWN = 12, ++ AGP_1X = 13, ++ AGP_2X = 14, ++ AGP_4X = 15, ++ AGP_8X = 16, ++ PCI_SPEED_66MHz_PCIX_533 = 17, ++ PCI_SPEED_100MHz_PCIX_533 = 18, ++ PCI_SPEED_133MHz_PCIX_533 = 19, ++ PCIE_SPEED_2_5GT = 20, ++ PCIE_SPEED_5_0GT = 21, ++ PCIE_SPEED_8_0GT = 22, ++ PCIE_SPEED_16_0GT = 23, ++ PCIE_SPEED_32_0GT = 24, ++ PCIE_SPEED_64_0GT = 25, ++ PCI_SPEED_UNKNOWN = 255, ++}; ++ ++struct pci_vpd { ++ struct mutex lock; ++ unsigned int len; ++ u8 cap; ++}; ++ ++struct aer_stats; ++ ++struct rcec_ea; ++ ++struct pci_driver; ++ ++struct pcie_link_state; ++ ++struct pci_sriov; ++ ++struct pci_dev { ++ struct list_head bus_list; ++ struct pci_bus *bus; ++ struct pci_bus *subordinate; ++ void *sysdata; ++ struct proc_dir_entry *procent; ++ struct pci_slot *slot; ++ unsigned int devfn; ++ short unsigned int vendor; ++ short unsigned int device; ++ short unsigned int subsystem_vendor; ++ short unsigned int subsystem_device; ++ unsigned int class; ++ u8 revision; ++ u8 hdr_type; ++ u16 aer_cap; ++ struct aer_stats *aer_stats; ++ struct rcec_ea *rcec_ea; ++ struct pci_dev *rcec; ++ u32 devcap; ++ u8 pcie_cap; ++ u8 msi_cap; ++ u8 msix_cap; ++ u8 pcie_mpss: 3; ++ u8 rom_base_reg; ++ u8 pin; ++ u16 pcie_flags_reg; ++ long unsigned int *dma_alias_mask; ++ struct pci_driver *driver; ++ u64 dma_mask; ++ struct device_dma_parameters dma_parms; ++ pci_power_t current_state; ++ unsigned int imm_ready: 1; ++ u8 pm_cap; ++ unsigned int pme_support: 5; ++ unsigned int pme_poll: 1; ++ unsigned int d1_support: 1; ++ unsigned int d2_support: 1; ++ unsigned int no_d1d2: 1; ++ unsigned int no_d3cold: 1; ++ unsigned int bridge_d3: 1; ++ unsigned int d3cold_allowed: 1; ++ unsigned int mmio_always_on: 1; ++ unsigned int wakeup_prepared: 1; ++ unsigned int skip_bus_pm: 1; ++ unsigned int ignore_hotplug: 1; ++ unsigned int hotplug_user_indicators: 1; ++ unsigned int clear_retrain_link: 1; ++ unsigned int d3hot_delay; ++ unsigned int d3cold_delay; ++ struct pcie_link_state *link_state; ++ unsigned int ltr_path: 1; ++ u16 l1ss; ++ unsigned int pasid_no_tlp: 1; ++ unsigned int eetlp_prefix_path: 1; ++ pci_channel_state_t error_state; ++ struct device dev; ++ int cfg_size; ++ unsigned int irq; ++ struct resource resource[17]; ++ struct resource driver_exclusive_resource; ++ bool match_driver; ++ unsigned int transparent: 1; ++ unsigned int io_window: 1; ++ unsigned int pref_window: 1; ++ unsigned int pref_64_window: 1; ++ unsigned int multifunction: 1; ++ unsigned int is_busmaster: 1; ++ unsigned int no_msi: 1; ++ unsigned int no_64bit_msi: 1; ++ unsigned int block_cfg_access: 1; ++ unsigned int broken_parity_status: 1; ++ unsigned int irq_reroute_variant: 2; ++ unsigned int msi_enabled: 1; ++ unsigned int msix_enabled: 1; ++ unsigned int ari_enabled: 1; ++ unsigned int ats_enabled: 1; ++ unsigned int pasid_enabled: 1; ++ unsigned int pri_enabled: 1; ++ unsigned int is_managed: 1; ++ unsigned int is_msi_managed: 1; ++ unsigned int needs_freset: 1; ++ unsigned int state_saved: 1; ++ unsigned int is_physfn: 1; ++ unsigned int is_virtfn: 1; ++ unsigned int is_hotplug_bridge: 1; ++ unsigned int shpc_managed: 1; ++ unsigned int is_thunderbolt: 1; ++ unsigned int untrusted: 1; ++ unsigned int external_facing: 1; ++ unsigned int broken_intx_masking: 1; ++ unsigned int io_window_1k: 1; ++ unsigned int irq_managed: 1; ++ unsigned int non_compliant_bars: 1; ++ unsigned int is_probed: 1; ++ unsigned int link_active_reporting: 1; ++ unsigned int no_vf_scan: 1; ++ unsigned int no_command_memory: 1; ++ unsigned int rom_bar_overlap: 1; ++ pci_dev_flags_t dev_flags; ++ atomic_t enable_cnt; ++ u32 saved_config_space[16]; ++ struct hlist_head saved_cap_space; ++ int rom_attr_enabled; ++ struct bin_attribute *res_attr[17]; ++ struct bin_attribute *res_attr_wc[17]; ++ unsigned int broken_cmd_compl: 1; ++ u16 ptm_cap; ++ unsigned int ptm_root: 1; ++ unsigned int ptm_enabled: 1; ++ u8 ptm_granularity; ++ void *msix_base; ++ raw_spinlock_t msi_lock; ++ struct pci_vpd vpd; ++ u16 dpc_cap; ++ unsigned int dpc_rp_extensions: 1; ++ u8 dpc_rp_log_size; ++ union { ++ struct pci_sriov *sriov; ++ struct pci_dev *physfn; ++ }; ++ u16 ats_cap; ++ u8 ats_stu; ++ u16 pri_cap; ++ u32 pri_reqs_alloc; ++ unsigned int pasid_required: 1; ++ u16 pasid_cap; ++ u16 pasid_features; ++ struct xarray doe_mbs; ++ u16 acs_cap; ++ phys_addr_t rom; ++ size_t romlen; ++ const char *driver_override; ++ long unsigned int priv_flags; ++ u8 reset_methods[7]; ++}; ++ ++struct pci_dynids { ++ spinlock_t lock; ++ struct list_head list; ++}; ++ ++struct pci_error_handlers; ++ ++struct pci_driver { ++ struct list_head node; ++ const char *name; ++ const struct pci_device_id *id_table; ++ int (*probe)(struct pci_dev *, const struct pci_device_id *); ++ void (*remove)(struct pci_dev *); ++ int (*suspend)(struct pci_dev *, pm_message_t); ++ int (*resume)(struct pci_dev *); ++ void (*shutdown)(struct pci_dev *); ++ int (*sriov_configure)(struct pci_dev *, int); ++ int (*sriov_set_msix_vec_count)(struct pci_dev *, int); ++ u32 (*sriov_get_vf_total_msix)(struct pci_dev *); ++ const struct pci_error_handlers *err_handler; ++ const struct attribute_group **groups; ++ const struct attribute_group **dev_groups; ++ struct device_driver driver; ++ struct pci_dynids dynids; ++ bool driver_managed_dma; ++}; ++ ++struct pci_ops { ++ int (*add_bus)(struct pci_bus *); ++ void (*remove_bus)(struct pci_bus *); ++ void * (*map_bus)(struct pci_bus *, unsigned int, int); ++ int (*read)(struct pci_bus *, unsigned int, int, int, u32 *); ++ int (*write)(struct pci_bus *, unsigned int, int, int, u32); ++}; ++ ++struct pci_error_handlers { ++ pci_ers_result_t (*error_detected)(struct pci_dev *, pci_channel_state_t); ++ pci_ers_result_t (*mmio_enabled)(struct pci_dev *); ++ pci_ers_result_t (*slot_reset)(struct pci_dev *); ++ void (*reset_prepare)(struct pci_dev *); ++ void (*reset_done)(struct pci_dev *); ++ void (*resume)(struct pci_dev *); ++ void (*cor_error_detected)(struct pci_dev *); ++}; ++ ++struct hotplug_slot_ops { ++ int (*enable_slot)(struct hotplug_slot *); ++ int (*disable_slot)(struct hotplug_slot *); ++ int (*set_attention_status)(struct hotplug_slot *, u8); ++ int (*hardware_test)(struct hotplug_slot *, u32); ++ int (*get_power_status)(struct hotplug_slot *, u8 *); ++ int (*get_attention_status)(struct hotplug_slot *, u8 *); ++ int (*get_latch_status)(struct hotplug_slot *, u8 *); ++ int (*get_adapter_status)(struct hotplug_slot *, u8 *); ++ int (*reset_slot)(struct hotplug_slot *, bool); ++}; ++ ++struct zpci_fmb_fmt0 { ++ u64 dma_rbytes; ++ u64 dma_wbytes; ++}; ++ ++struct zpci_fmb_fmt1 { ++ u64 rx_bytes; ++ u64 rx_packets; ++ u64 tx_bytes; ++ u64 tx_packets; ++}; ++ ++struct zpci_fmb_fmt2 { ++ u64 consumed_work_units; ++ u64 max_work_units; ++}; ++ ++struct zpci_fmb_fmt3 { ++ u64 tx_bytes; ++}; ++ ++struct zpci_fmb { ++ u32 format: 8; ++ u32 fmt_ind: 24; ++ u32 samples; ++ u64 last_update; ++ u64 ld_ops; ++ u64 st_ops; ++ u64 stb_ops; ++ u64 rpcit_ops; ++ union { ++ struct zpci_fmb_fmt0 fmt0; ++ struct zpci_fmb_fmt1 fmt1; ++ struct zpci_fmb_fmt2 fmt2; ++ struct zpci_fmb_fmt3 fmt3; ++ }; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum zpci_state { ++ ZPCI_FN_STATE_STANDBY = 0, ++ ZPCI_FN_STATE_CONFIGURED = 1, ++ ZPCI_FN_STATE_RESERVED = 2, ++}; ++ ++struct zpci_bar_struct { ++ struct resource *res; ++ void *mio_wb; ++ void *mio_wt; ++ u32 val; ++ u16 map_idx; ++ u8 size; ++}; ++ ++struct zpci_dev; ++ ++struct zpci_bus { ++ struct kref kref; ++ struct pci_bus *bus; ++ struct zpci_dev *function[256]; ++ struct list_head resources; ++ struct list_head bus_next; ++ struct resource bus_resource; ++ int pchid; ++ int domain_nr; ++ bool multifunction; ++ enum pci_bus_speed max_bus_speed; ++}; ++ ++struct kvm_zdev; ++ ++struct airq_iv; ++ ++struct s390_domain; ++ ++struct zpci_dev { ++ struct zpci_bus *zbus; ++ struct list_head entry; ++ struct list_head iommu_list; ++ struct kref kref; ++ struct callback_head rcu; ++ struct hotplug_slot hotplug_slot; ++ enum zpci_state state; ++ u32 fid; ++ u32 fh; ++ u32 gisa; ++ u16 vfn; ++ u16 pchid; ++ u16 maxstbl; ++ u8 pfgid; ++ u8 pft; ++ u8 port; ++ u8 dtsm; ++ u8 rid_available: 1; ++ u8 has_hp_slot: 1; ++ u8 has_resources: 1; ++ u8 is_physfn: 1; ++ u8 util_str_avail: 1; ++ u8 irqs_registered: 1; ++ u8 reserved: 2; ++ unsigned int devfn; ++ struct mutex lock; ++ u8 pfip[4]; ++ u32 uid; ++ u8 util_str[64]; ++ u64 msi_addr; ++ unsigned int max_msi; ++ unsigned int msi_first_bit; ++ unsigned int msi_nr_irqs; ++ struct airq_iv *aibv; ++ long unsigned int aisb; ++ long unsigned int *dma_table; ++ int tlb_refresh; ++ spinlock_t iommu_bitmap_lock; ++ long unsigned int *iommu_bitmap; ++ long unsigned int *lazy_bitmap; ++ long unsigned int iommu_size; ++ long unsigned int iommu_pages; ++ unsigned int next_bit; ++ struct iommu_device iommu_dev; ++ char res_name[16]; ++ bool mio_capable; ++ struct zpci_bar_struct bars[6]; ++ u64 start_dma; ++ u64 end_dma; ++ u64 dma_mask; ++ struct zpci_fmb *fmb; ++ u16 fmb_update; ++ u16 fmb_length; ++ atomic64_t allocated_pages; ++ atomic64_t mapped_pages; ++ atomic64_t unmapped_pages; ++ u8 version; ++ enum pci_bus_speed max_bus_speed; ++ struct dentry *debugfs_dev; ++ struct s390_domain *s390_domain; ++ struct kvm_zdev *kzdev; ++ struct mutex kzdev_lock; ++}; ++ ++struct airq_iv { ++ long unsigned int *vector; ++ dma_addr_t vector_dma; ++ long unsigned int *avail; ++ long unsigned int *bitlock; ++ long unsigned int *ptr; ++ unsigned int *data; ++ long unsigned int bits; ++ long unsigned int end; ++ long unsigned int flags; ++ spinlock_t lock; ++}; ++ ++struct airq_struct { ++ struct hlist_node list; ++ void (*handler)(struct airq_struct *, struct tpi_info *); ++ u8 *lsi_ptr; ++ u8 lsi_mask; ++ u8 isc; ++ u8 flags; ++}; ++ ++enum { ++ FLOATING = 0, ++ DIRECTED = 1, ++}; ++ ++struct cpu_irq_data { ++ call_single_data_t csd; ++ atomic_t scheduled; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum refcount_saturation_type { ++ REFCOUNT_ADD_NOT_ZERO_OVF = 0, ++ REFCOUNT_ADD_OVF = 1, ++ REFCOUNT_ADD_UAF = 2, ++ REFCOUNT_SUB_UAF = 3, ++ REFCOUNT_DEC_LEAK = 4, ++}; ++ ++enum { ++ pci_channel_io_normal = 1, ++ pci_channel_io_frozen = 2, ++ pci_channel_io_perm_failure = 3, ++}; ++ ++enum pci_ers_result { ++ PCI_ERS_RESULT_NONE = 1, ++ PCI_ERS_RESULT_CAN_RECOVER = 2, ++ PCI_ERS_RESULT_NEED_RESET = 3, ++ PCI_ERS_RESULT_DISCONNECT = 4, ++ PCI_ERS_RESULT_RECOVERED = 5, ++ PCI_ERS_RESULT_NO_AER_DRIVER = 6, ++}; ++ ++struct zpci_ccdf_err { ++ u32 reserved1; ++ u32 fh; ++ u32 fid; ++ u32 ett: 4; ++ u32 mvn: 12; ++ u32 dmaas: 8; ++ char: 6; ++ u32 q: 1; ++ u32 rw: 1; ++ u64 faddr; ++ u32 reserved3; ++ u16 reserved4; ++ u16 pec; ++}; ++ ++struct zpci_ccdf_avail { ++ u32 reserved1; ++ u32 fh; ++ u32 fid; ++ u32 reserved2; ++ u32 reserved3; ++ u32 reserved4; ++ u32 reserved5; ++ u16 reserved6; ++ u16 pec; ++}; ++ ++struct preempt_notifier; ++ ++struct preempt_ops { ++ void (*sched_in)(struct preempt_notifier *, int); ++ void (*sched_out)(struct preempt_notifier *, struct task_struct *); ++}; ++ ++struct preempt_notifier { ++ struct hlist_node link; ++ struct preempt_ops *ops; ++}; ++ ++struct gmap { ++ struct list_head list; ++ struct list_head crst_list; ++ struct mm_struct *mm; ++ struct xarray guest_to_host; ++ struct xarray host_to_guest; ++ spinlock_t guest_table_lock; ++ refcount_t ref_count; ++ long unsigned int *table; ++ long unsigned int asce; ++ long unsigned int asce_end; ++ void *private; ++ bool pfault_enabled; ++ long unsigned int guest_handle; ++ struct xarray host_to_rmap; ++ struct list_head children; ++ struct list_head pt_list; ++ spinlock_t shadow_lock; ++ struct gmap *parent; ++ long unsigned int orig_asce; ++ int edat_level; ++ bool removed; ++ bool initialized; ++}; ++ ++struct interval_tree_node { ++ struct rb_node rb; ++ long unsigned int start; ++ long unsigned int last; ++ long unsigned int __subtree_last; ++}; ++ ++enum mmu_notifier_event { ++ MMU_NOTIFY_UNMAP = 0, ++ MMU_NOTIFY_CLEAR = 1, ++ MMU_NOTIFY_PROTECTION_VMA = 2, ++ MMU_NOTIFY_PROTECTION_PAGE = 3, ++ MMU_NOTIFY_SOFT_DIRTY = 4, ++ MMU_NOTIFY_RELEASE = 5, ++ MMU_NOTIFY_MIGRATE = 6, ++ MMU_NOTIFY_EXCLUSIVE = 7, ++}; ++ ++struct mmu_notifier; ++ ++struct mmu_notifier_range; ++ ++struct mmu_notifier_ops { ++ void (*release)(struct mmu_notifier *, struct mm_struct *); ++ int (*clear_flush_young)(struct mmu_notifier *, struct mm_struct *, long unsigned int, long unsigned int); ++ int (*clear_young)(struct mmu_notifier *, struct mm_struct *, long unsigned int, long unsigned int); ++ int (*test_young)(struct mmu_notifier *, struct mm_struct *, long unsigned int); ++ void (*change_pte)(struct mmu_notifier *, struct mm_struct *, long unsigned int, pte_t); ++ int (*invalidate_range_start)(struct mmu_notifier *, const struct mmu_notifier_range *); ++ void (*invalidate_range_end)(struct mmu_notifier *, const struct mmu_notifier_range *); ++ void (*invalidate_range)(struct mmu_notifier *, struct mm_struct *, long unsigned int, long unsigned int); ++ struct mmu_notifier * (*alloc_notifier)(struct mm_struct *); ++ void (*free_notifier)(struct mmu_notifier *); ++}; ++ ++struct mmu_notifier { ++ struct hlist_node hlist; ++ const struct mmu_notifier_ops *ops; ++ struct mm_struct *mm; ++ struct callback_head rcu; ++ unsigned int users; ++}; ++ ++struct mmu_notifier_range { ++ struct mm_struct *mm; ++ long unsigned int start; ++ long unsigned int end; ++ unsigned int flags; ++ enum mmu_notifier_event event; ++ void *owner; ++}; ++ ++struct kvm_s390_vm_cpu_subfunc { ++ __u8 plo[32]; ++ __u8 ptff[16]; ++ __u8 kmac[16]; ++ __u8 kmc[16]; ++ __u8 km[16]; ++ __u8 kimd[16]; ++ __u8 klmd[16]; ++ __u8 pckmo[16]; ++ __u8 kmctr[16]; ++ __u8 kmf[16]; ++ __u8 kmo[16]; ++ __u8 pcc[16]; ++ __u8 ppno[16]; ++ __u8 kma[16]; ++ __u8 kdsa[16]; ++ __u8 sortl[32]; ++ __u8 dfltcc[32]; ++ __u8 reserved[1728]; ++}; ++ ++struct kvm_debug_exit_arch { ++ __u64 addr; ++ __u8 type; ++ __u8 pad[7]; ++}; ++ ++struct kvm_sync_regs { ++ __u64 prefix; ++ __u64 gprs[16]; ++ __u32 acrs[16]; ++ __u64 crs[16]; ++ __u64 todpr; ++ __u64 cputm; ++ __u64 ckc; ++ __u64 pp; ++ __u64 gbea; ++ __u64 pft; ++ __u64 pfs; ++ __u64 pfc; ++ union { ++ __u64 vrs[64]; ++ __u64 fprs[16]; ++ }; ++ __u8 reserved[512]; ++ __u32 fpc; ++ __u8 bpbc: 1; ++ __u8 reserved2: 7; ++ __u8 padding1[51]; ++ __u8 riccb[64]; ++ __u64 diag318; ++ __u8 padding2[184]; ++ union { ++ __u8 sdnx[256]; ++ struct { ++ __u64 reserved1[2]; ++ __u64 gscb[4]; ++ __u64 etoken; ++ __u64 etoken_extension; ++ }; ++ }; ++}; ++ ++struct kvm_hyperv_exit { ++ __u32 type; ++ __u32 pad1; ++ union { ++ struct { ++ __u32 msr; ++ __u32 pad2; ++ __u64 control; ++ __u64 evt_page; ++ __u64 msg_page; ++ } synic; ++ struct { ++ __u64 input; ++ __u64 result; ++ __u64 params[2]; ++ } hcall; ++ struct { ++ __u32 msr; ++ __u32 pad2; ++ __u64 control; ++ __u64 status; ++ __u64 send_page; ++ __u64 recv_page; ++ __u64 pending_page; ++ } syndbg; ++ } u; ++}; ++ ++struct kvm_xen_exit { ++ __u32 type; ++ union { ++ struct { ++ __u32 longmode; ++ __u32 cpl; ++ __u64 input; ++ __u64 result; ++ __u64 params[6]; ++ } hcall; ++ } u; ++}; ++ ++struct kvm_run { ++ __u8 request_interrupt_window; ++ __u8 immediate_exit; ++ __u8 padding1[6]; ++ __u32 exit_reason; ++ __u8 ready_for_interrupt_injection; ++ __u8 if_flag; ++ __u16 flags; ++ __u64 cr8; ++ __u64 apic_base; ++ __u64 psw_mask; ++ __u64 psw_addr; ++ union { ++ struct { ++ __u64 hardware_exit_reason; ++ } hw; ++ struct { ++ __u64 hardware_entry_failure_reason; ++ __u32 cpu; ++ } fail_entry; ++ struct { ++ __u32 exception; ++ __u32 error_code; ++ } ex; ++ struct { ++ __u8 direction; ++ __u8 size; ++ __u16 port; ++ __u32 count; ++ __u64 data_offset; ++ } io; ++ struct { ++ struct kvm_debug_exit_arch arch; ++ } debug; ++ struct { ++ __u64 phys_addr; ++ __u8 data[8]; ++ __u32 len; ++ __u8 is_write; ++ } mmio; ++ struct { ++ __u64 nr; ++ __u64 args[6]; ++ __u64 ret; ++ union { ++ __u64 flags; ++ }; ++ } hypercall; ++ struct { ++ __u64 rip; ++ __u32 is_write; ++ __u32 pad; ++ } tpr_access; ++ struct { ++ __u8 icptcode; ++ __u16 ipa; ++ __u32 ipb; ++ } s390_sieic; ++ __u64 s390_reset_flags; ++ struct { ++ __u64 trans_exc_code; ++ __u32 pgm_code; ++ } s390_ucontrol; ++ struct { ++ __u32 dcrn; ++ __u32 data; ++ __u8 is_write; ++ } dcr; ++ struct { ++ __u32 suberror; ++ __u32 ndata; ++ __u64 data[16]; ++ } internal; ++ struct { ++ __u32 suberror; ++ __u32 ndata; ++ __u64 flags; ++ union { ++ struct { ++ __u8 insn_size; ++ __u8 insn_bytes[15]; ++ }; ++ }; ++ } emulation_failure; ++ struct { ++ __u64 gprs[32]; ++ } osi; ++ struct { ++ __u64 nr; ++ __u64 ret; ++ __u64 args[9]; ++ } papr_hcall; ++ struct { ++ __u16 subchannel_id; ++ __u16 subchannel_nr; ++ __u32 io_int_parm; ++ __u32 io_int_word; ++ __u32 ipb; ++ __u8 dequeued; ++ } s390_tsch; ++ struct { ++ __u32 epr; ++ } epr; ++ struct { ++ __u32 type; ++ __u32 ndata; ++ union { ++ __u64 data[16]; ++ }; ++ } system_event; ++ struct { ++ __u64 addr; ++ __u8 ar; ++ __u8 reserved; ++ __u8 fc; ++ __u8 sel1; ++ __u16 sel2; ++ } s390_stsi; ++ struct { ++ __u8 vector; ++ } eoi; ++ struct kvm_hyperv_exit hyperv; ++ struct { ++ __u64 esr_iss; ++ __u64 fault_ipa; ++ } arm_nisv; ++ struct { ++ __u8 error; ++ __u8 pad[7]; ++ __u32 reason; ++ __u32 index; ++ __u64 data; ++ } msr; ++ struct kvm_xen_exit xen; ++ struct { ++ long unsigned int extension_id; ++ long unsigned int function_id; ++ long unsigned int args[6]; ++ long unsigned int ret[2]; ++ } riscv_sbi; ++ struct { ++ long unsigned int csr_num; ++ long unsigned int new_value; ++ long unsigned int write_mask; ++ long unsigned int ret_value; ++ } riscv_csr; ++ struct { ++ __u32 flags; ++ } notify; ++ char padding[256]; ++ }; ++ __u64 kvm_valid_regs; ++ __u64 kvm_dirty_regs; ++ union { ++ struct kvm_sync_regs regs; ++ char padding[2048]; ++ } s; ++}; ++ ++struct kvm_s390_io_info { ++ __u16 subchannel_id; ++ __u16 subchannel_nr; ++ __u32 io_int_parm; ++ __u32 io_int_word; ++}; ++ ++struct kvm_s390_ext_info { ++ __u32 ext_params; ++ __u32 pad; ++ __u64 ext_params2; ++}; ++ ++struct kvm_s390_pgm_info { ++ __u64 trans_exc_code; ++ __u64 mon_code; ++ __u64 per_address; ++ __u32 data_exc_code; ++ __u16 code; ++ __u16 mon_class_nr; ++ __u8 per_code; ++ __u8 per_atmid; ++ __u8 exc_access_id; ++ __u8 per_access_id; ++ __u8 op_access_id; ++ __u8 flags; ++ __u8 pad[2]; ++}; ++ ++struct kvm_s390_prefix_info { ++ __u32 address; ++}; ++ ++struct kvm_s390_extcall_info { ++ __u16 code; ++}; ++ ++struct kvm_s390_emerg_info { ++ __u16 code; ++}; ++ ++struct kvm_s390_stop_info { ++ __u32 flags; ++}; ++ ++struct kvm_s390_mchk_info { ++ __u64 cr14; ++ __u64 mcic; ++ __u64 failing_storage_address; ++ __u32 ext_damage_code; ++ __u32 pad; ++ __u8 fixed_logout[16]; ++}; ++ ++struct kvm_device_attr { ++ __u32 flags; ++ __u32 group; ++ __u64 attr; ++ __u64 addr; ++}; ++ ++struct kvm_dirty_gfn { ++ __u32 flags; ++ __u32 slot; ++ __u64 offset; ++}; ++ ++struct kvm_stats_desc { ++ __u32 flags; ++ __s16 exponent; ++ __u16 size; ++ __u32 offset; ++ __u32 bucket_size; ++ char name[0]; ++}; ++ ++union diag318_info { ++ long unsigned int val; ++ struct { ++ long unsigned int cpnc: 8; ++ long unsigned int cpvc: 56; ++ }; ++}; ++ ++typedef u64 gpa_t; ++ ++typedef u64 gfn_t; ++ ++struct kvm_arch_memory_slot {}; ++ ++struct kvm_memory_slot { ++ struct hlist_node id_node[2]; ++ struct interval_tree_node hva_node[2]; ++ struct rb_node gfn_node[2]; ++ gfn_t base_gfn; ++ long unsigned int npages; ++ long unsigned int *dirty_bitmap; ++ struct kvm_arch_memory_slot arch; ++ long unsigned int userspace_addr; ++ u32 flags; ++ short int id; ++ u16 as_id; ++}; ++ ++struct kvm_memslots { ++ u64 generation; ++ atomic_long_t last_used_slot; ++ struct rb_root_cached hva_tree; ++ struct rb_root gfn_tree; ++ struct hlist_head id_hash[128]; ++ int node_idx; ++}; ++ ++struct kvm_vm_stat_generic { ++ u64 remote_tlb_flush; ++ u64 remote_tlb_flush_requests; ++}; ++ ++struct kvm_vm_stat { ++ struct kvm_vm_stat_generic generic; ++ u64 inject_io; ++ u64 inject_float_mchk; ++ u64 inject_pfault_done; ++ u64 inject_service_signal; ++ u64 inject_virtio; ++ u64 aen_forward; ++}; ++ ++struct kvm_s390_float_interrupt { ++ long unsigned int pending_irqs; ++ long unsigned int masked_irqs; ++ spinlock_t lock; ++ struct list_head lists[10]; ++ int counters[4]; ++ struct kvm_s390_mchk_info mchk; ++ struct kvm_s390_ext_info srv_signal; ++ int next_rr_cpu; ++ struct mutex ais_lock; ++ u8 simm; ++ u8 nimm; ++}; ++ ++struct kvm_s390_cpu_model { ++ __u64 fac_mask[256]; ++ struct kvm_s390_vm_cpu_subfunc subfuncs; ++ __u64 *fac_list; ++ u64 cpuid; ++ short unsigned int ibc; ++}; ++ ++struct kvm_vcpu; ++ ++typedef int (*crypto_hook)(struct kvm_vcpu *); ++ ++struct kvm_s390_crypto_cb; ++ ++struct kvm_s390_crypto { ++ struct kvm_s390_crypto_cb *crycb; ++ struct rw_semaphore pqap_hook_rwsem; ++ crypto_hook *pqap_hook; ++ __u32 crycbd; ++ __u8 aes_kw; ++ __u8 dea_kw; ++ __u8 apie; ++}; ++ ++struct kvm_s390_vsie { ++ struct mutex mutex; ++ struct xarray addr_to_page; ++ int page_count; ++ int next; ++ struct page *pages[255]; ++}; ++ ++struct kvm_s390_gisa_iam { ++ u8 mask; ++ spinlock_t ref_lock; ++ u32 ref_count[8]; ++}; ++ ++struct kvm_s390_gisa; ++ ++struct kvm_s390_gisa_interrupt { ++ struct kvm_s390_gisa *origin; ++ struct kvm_s390_gisa_iam alert; ++ struct hrtimer timer; ++ u64 expires; ++ long unsigned int kicked_mask[4]; ++}; ++ ++struct kvm_s390_pv { ++ u64 handle; ++ u64 guest_len; ++ long unsigned int stor_base; ++ void *stor_var; ++ bool dumping; ++ void *set_aside; ++ struct list_head need_cleanup; ++ struct mmu_notifier mmu_notifier; ++}; ++ ++struct kvm_device; ++ ++struct s390_io_adapter; ++ ++struct sie_page2; ++ ++struct kvm_arch { ++ void *sca; ++ int use_esca; ++ rwlock_t sca_lock; ++ debug_info_t *dbf; ++ struct kvm_s390_float_interrupt float_int; ++ struct kvm_device *flic; ++ struct gmap *gmap; ++ long unsigned int mem_limit; ++ int css_support; ++ int use_irqchip; ++ int use_cmma; ++ int use_pfmfi; ++ int use_skf; ++ int use_zpci_interp; ++ int user_cpu_state_ctrl; ++ int user_sigp; ++ int user_stsi; ++ int user_instr0; ++ struct s390_io_adapter *adapters[64]; ++ wait_queue_head_t ipte_wq; ++ int ipte_lock_count; ++ struct mutex ipte_mutex; ++ spinlock_t start_stop_lock; ++ struct sie_page2 *sie_page2; ++ struct kvm_s390_cpu_model model; ++ struct kvm_s390_crypto crypto; ++ struct kvm_s390_vsie vsie; ++ u8 epdx; ++ u64 epoch; ++ int migration_mode; ++ atomic64_t cmma_dirty_pages; ++ long unsigned int cpu_feat[16]; ++ long unsigned int idle_mask[4]; ++ struct kvm_s390_gisa_interrupt gisa_int; ++ struct kvm_s390_pv pv; ++ struct list_head kzdev_list; ++ spinlock_t kzdev_list_lock; ++}; ++ ++struct kvm_io_bus; ++ ++struct kvm_irq_routing_table; ++ ++struct kvm_stat_data; ++ ++struct kvm { ++ spinlock_t mmu_lock; ++ struct mutex slots_lock; ++ struct mutex slots_arch_lock; ++ struct mm_struct *mm; ++ long unsigned int nr_memslot_pages; ++ struct kvm_memslots __memslots[2]; ++ struct kvm_memslots *memslots[1]; ++ struct xarray vcpu_array; ++ atomic_t nr_memslots_dirty_logging; ++ spinlock_t mn_invalidate_lock; ++ long unsigned int mn_active_invalidate_count; ++ struct rcuwait mn_memslots_update_rcuwait; ++ spinlock_t gpc_lock; ++ struct list_head gpc_list; ++ atomic_t online_vcpus; ++ int max_vcpus; ++ int created_vcpus; ++ int last_boosted_vcpu; ++ struct list_head vm_list; ++ struct mutex lock; ++ struct kvm_io_bus *buses[4]; ++ struct { ++ spinlock_t lock; ++ struct list_head items; ++ struct list_head resampler_list; ++ struct mutex resampler_lock; ++ } irqfds; ++ struct list_head ioeventfds; ++ struct kvm_vm_stat stat; ++ struct kvm_arch arch; ++ refcount_t users_count; ++ struct mutex irq_lock; ++ struct kvm_irq_routing_table *irq_routing; ++ struct hlist_head irq_ack_notifier_list; ++ struct list_head devices; ++ u64 manual_dirty_log_protect; ++ struct dentry *debugfs_dentry; ++ struct kvm_stat_data **debugfs_stat_data; ++ struct srcu_struct srcu; ++ struct srcu_struct irq_srcu; ++ pid_t userspace_pid; ++ bool override_halt_poll_ns; ++ unsigned int max_halt_poll_ns; ++ u32 dirty_ring_size; ++ bool dirty_ring_with_bitmap; ++ bool vm_bugged; ++ bool vm_dead; ++ char stats_id[48]; ++}; ++ ++struct kvm_mmio_fragment { ++ gpa_t gpa; ++ void *data; ++ unsigned int len; ++}; ++ ++struct kvm_s390_irq_payload { ++ struct kvm_s390_io_info io; ++ struct kvm_s390_ext_info ext; ++ struct kvm_s390_pgm_info pgm; ++ struct kvm_s390_emerg_info emerg; ++ struct kvm_s390_extcall_info extcall; ++ struct kvm_s390_prefix_info prefix; ++ struct kvm_s390_stop_info stop; ++ struct kvm_s390_mchk_info mchk; ++}; ++ ++struct kvm_s390_local_interrupt { ++ spinlock_t lock; ++ long unsigned int sigp_emerg_pending[4]; ++ struct kvm_s390_irq_payload irq; ++ long unsigned int pending_irqs; ++}; ++ ++struct kvm_hw_bp_info_arch; ++ ++struct kvm_hw_wp_info_arch; ++ ++struct kvm_guestdbg_info_arch { ++ long unsigned int cr0; ++ long unsigned int cr9; ++ long unsigned int cr10; ++ long unsigned int cr11; ++ struct kvm_hw_bp_info_arch *hw_bp_info; ++ struct kvm_hw_wp_info_arch *hw_wp_info; ++ int nr_hw_bp; ++ int nr_hw_wp; ++ long unsigned int last_bp; ++}; ++ ++struct kvm_s390_pv_vcpu { ++ u64 handle; ++ long unsigned int stor_base; ++}; ++ ++struct kvm_vcpu_arch { ++ struct kvm_s390_sie_block *sie_block; ++ struct kvm_s390_sie_block *vsie_block; ++ unsigned int host_acrs[16]; ++ struct gs_cb *host_gscb; ++ struct fpu host_fpregs; ++ struct kvm_s390_local_interrupt local_int; ++ struct hrtimer ckc_timer; ++ struct kvm_s390_pgm_info pgm; ++ struct gmap *gmap; ++ struct gmap *enabled_gmap; ++ struct kvm_guestdbg_info_arch guestdbg; ++ long unsigned int pfault_token; ++ long unsigned int pfault_select; ++ long unsigned int pfault_compare; ++ bool cputm_enabled; ++ seqcount_t cputm_seqcount; ++ __u64 cputm_start; ++ bool gs_enabled; ++ bool skey_enabled; ++ struct kvm_s390_pv_vcpu pv; ++ union diag318_info diag318_info; ++}; ++ ++struct kvm_vcpu_stat_generic { ++ u64 halt_successful_poll; ++ u64 halt_attempted_poll; ++ u64 halt_poll_invalid; ++ u64 halt_wakeup; ++ u64 halt_poll_success_ns; ++ u64 halt_poll_fail_ns; ++ u64 halt_wait_ns; ++ u64 halt_poll_success_hist[32]; ++ u64 halt_poll_fail_hist[32]; ++ u64 halt_wait_hist[32]; ++ u64 blocking; ++}; ++ ++struct kvm_vcpu_stat { ++ struct kvm_vcpu_stat_generic generic; ++ u64 exit_userspace; ++ u64 exit_null; ++ u64 exit_external_request; ++ u64 exit_io_request; ++ u64 exit_external_interrupt; ++ u64 exit_stop_request; ++ u64 exit_validity; ++ u64 exit_instruction; ++ u64 exit_pei; ++ u64 halt_no_poll_steal; ++ u64 instruction_lctl; ++ u64 instruction_lctlg; ++ u64 instruction_stctl; ++ u64 instruction_stctg; ++ u64 exit_program_interruption; ++ u64 exit_instr_and_program; ++ u64 exit_operation_exception; ++ u64 deliver_ckc; ++ u64 deliver_cputm; ++ u64 deliver_external_call; ++ u64 deliver_emergency_signal; ++ u64 deliver_service_signal; ++ u64 deliver_virtio; ++ u64 deliver_stop_signal; ++ u64 deliver_prefix_signal; ++ u64 deliver_restart_signal; ++ u64 deliver_program; ++ u64 deliver_io; ++ u64 deliver_machine_check; ++ u64 exit_wait_state; ++ u64 inject_ckc; ++ u64 inject_cputm; ++ u64 inject_external_call; ++ u64 inject_emergency_signal; ++ u64 inject_mchk; ++ u64 inject_pfault_init; ++ u64 inject_program; ++ u64 inject_restart; ++ u64 inject_set_prefix; ++ u64 inject_stop_signal; ++ u64 instruction_epsw; ++ u64 instruction_gs; ++ u64 instruction_io_other; ++ u64 instruction_lpsw; ++ u64 instruction_lpswe; ++ u64 instruction_pfmf; ++ u64 instruction_ptff; ++ u64 instruction_sck; ++ u64 instruction_sckpf; ++ u64 instruction_stidp; ++ u64 instruction_spx; ++ u64 instruction_stpx; ++ u64 instruction_stap; ++ u64 instruction_iske; ++ u64 instruction_ri; ++ u64 instruction_rrbe; ++ u64 instruction_sske; ++ u64 instruction_ipte_interlock; ++ u64 instruction_stsi; ++ u64 instruction_stfl; ++ u64 instruction_tb; ++ u64 instruction_tpi; ++ u64 instruction_tprot; ++ u64 instruction_tsch; ++ u64 instruction_sie; ++ u64 instruction_essa; ++ u64 instruction_sthyi; ++ u64 instruction_sigp_sense; ++ u64 instruction_sigp_sense_running; ++ u64 instruction_sigp_external_call; ++ u64 instruction_sigp_emergency; ++ u64 instruction_sigp_cond_emergency; ++ u64 instruction_sigp_start; ++ u64 instruction_sigp_stop; ++ u64 instruction_sigp_stop_store_status; ++ u64 instruction_sigp_store_status; ++ u64 instruction_sigp_store_adtl_status; ++ u64 instruction_sigp_arch; ++ u64 instruction_sigp_prefix; ++ u64 instruction_sigp_restart; ++ u64 instruction_sigp_init_cpu_reset; ++ u64 instruction_sigp_cpu_reset; ++ u64 instruction_sigp_unknown; ++ u64 instruction_diagnose_10; ++ u64 instruction_diagnose_44; ++ u64 instruction_diagnose_9c; ++ u64 diag_9c_ignored; ++ u64 diag_9c_forward; ++ u64 instruction_diagnose_258; ++ u64 instruction_diagnose_308; ++ u64 instruction_diagnose_500; ++ u64 instruction_diagnose_other; ++ u64 pfault_sync; ++}; ++ ++struct kvm_dirty_ring { ++ u32 dirty_index; ++ u32 reset_index; ++ u32 size; ++ u32 soft_limit; ++ struct kvm_dirty_gfn *dirty_gfns; ++ int index; ++}; ++ ++struct kvm_vcpu { ++ struct kvm *kvm; ++ struct preempt_notifier preempt_notifier; ++ int cpu; ++ int vcpu_id; ++ int vcpu_idx; ++ int ____srcu_idx; ++ int mode; ++ u64 requests; ++ long unsigned int guest_debug; ++ struct mutex mutex; ++ struct kvm_run *run; ++ struct rcuwait wait; ++ struct pid *pid; ++ int sigset_active; ++ sigset_t sigset; ++ unsigned int halt_poll_ns; ++ bool valid_wakeup; ++ int mmio_needed; ++ int mmio_read_completed; ++ int mmio_is_write; ++ int mmio_cur_fragment; ++ int mmio_nr_fragments; ++ struct kvm_mmio_fragment mmio_fragments[2]; ++ struct { ++ u32 queued; ++ struct list_head queue; ++ struct list_head done; ++ spinlock_t lock; ++ } async_pf; ++ struct { ++ bool in_spin_loop; ++ bool dy_eligible; ++ } spin_loop; ++ bool preempted; ++ bool ready; ++ struct kvm_vcpu_arch arch; ++ struct kvm_vcpu_stat stat; ++ char stats_id[48]; ++ struct kvm_dirty_ring dirty_ring; ++ struct kvm_memory_slot *last_used_slot; ++ u64 last_used_slot_gen; ++}; ++ ++struct kvm_hw_wp_info_arch { ++ long unsigned int addr; ++ long unsigned int phys_addr; ++ int len; ++ char *old_data; ++}; ++ ++struct kvm_hw_bp_info_arch { ++ long unsigned int addr; ++ int len; ++}; ++ ++struct s390_io_adapter { ++ unsigned int id; ++ int isc; ++ bool maskable; ++ bool masked; ++ bool swap; ++ bool suppressible; ++}; ++ ++struct kvm_s390_apcb0 { ++ __u64 apm[1]; ++ __u64 aqm[1]; ++ __u64 adm[1]; ++ __u64 reserved18; ++}; ++ ++struct kvm_s390_apcb1 { ++ __u64 apm[4]; ++ __u64 aqm[4]; ++ __u64 adm[4]; ++ __u64 reserved60[4]; ++}; ++ ++struct kvm_s390_crypto_cb { ++ struct kvm_s390_apcb0 apcb0; ++ __u8 reserved20[40]; ++ __u8 dea_wrapping_key_mask[24]; ++ __u8 aes_wrapping_key_mask[32]; ++ struct kvm_s390_apcb1 apcb1; ++}; ++ ++struct kvm_s390_gisa { ++ union { ++ struct { ++ u32 next_alert; ++ u8 ipm; ++ u8 reserved01[2]; ++ u8 iam; ++ }; ++ struct { ++ u32 next_alert; ++ u8 ipm; ++ u8 reserved01; ++ char: 6; ++ u8 g: 1; ++ u8 c: 1; ++ u8 iam; ++ u8 reserved02[4]; ++ u32 airq_count; ++ } g0; ++ struct { ++ u32 next_alert; ++ u8 ipm; ++ u8 simm; ++ u8 nimm; ++ u8 iam; ++ u8 aism[8]; ++ char: 6; ++ u8 g: 1; ++ u8 c: 1; ++ u8 reserved03[11]; ++ u32 airq_count; ++ } g1; ++ struct { ++ u64 word[4]; ++ } u64; ++ }; ++}; ++ ++struct sie_page2 { ++ __u64 fac_list[256]; ++ struct kvm_s390_crypto_cb crycb; ++ struct kvm_s390_gisa gisa; ++ struct kvm *kvm; ++ u8 reserved928[1752]; ++}; ++ ++struct kvm_device_ops; ++ ++struct kvm_device { ++ const struct kvm_device_ops *ops; ++ struct kvm *kvm; ++ void *private; ++ struct list_head vm_node; ++}; ++ ++struct zpci_kvm_hook { ++ int (*kvm_register)(void *, struct kvm *); ++ void (*kvm_unregister)(void *); ++}; ++ ++struct kvm_io_device; ++ ++struct kvm_io_range { ++ gpa_t addr; ++ int len; ++ struct kvm_io_device *dev; ++}; ++ ++struct kvm_io_bus { ++ int dev_count; ++ int ioeventfd_count; ++ struct kvm_io_range range[0]; ++}; ++ ++struct kvm_irq_routing_table { ++ int chip[1]; ++ u32 nr_rt_entries; ++ struct hlist_head map[0]; ++}; ++ ++enum kvm_stat_kind { ++ KVM_STAT_VM = 0, ++ KVM_STAT_VCPU = 1, ++}; ++ ++struct _kvm_stats_desc; ++ ++struct kvm_stat_data { ++ struct kvm *kvm; ++ const struct _kvm_stats_desc *desc; ++ enum kvm_stat_kind kind; ++}; ++ ++struct _kvm_stats_desc { ++ struct kvm_stats_desc desc; ++ char name[48]; ++}; ++ ++struct kvm_device_ops { ++ const char *name; ++ int (*create)(struct kvm_device *, u32); ++ void (*init)(struct kvm_device *); ++ void (*destroy)(struct kvm_device *); ++ void (*release)(struct kvm_device *); ++ int (*set_attr)(struct kvm_device *, struct kvm_device_attr *); ++ int (*get_attr)(struct kvm_device *, struct kvm_device_attr *); ++ int (*has_attr)(struct kvm_device *, struct kvm_device_attr *); ++ long int (*ioctl)(struct kvm_device *, unsigned int, long unsigned int); ++ int (*mmap)(struct kvm_device *, struct vm_area_struct *); ++}; ++ ++struct taint_flag { ++ char c_true; ++ char c_false; ++ bool module; ++}; ++ ++enum lockdep_ok { ++ LOCKDEP_STILL_OK = 0, ++ LOCKDEP_NOW_UNRELIABLE = 1, ++}; ++ ++struct __va_list_tag { ++ long int __gpr; ++ long int __fpr; ++ void *__overflow_arg_area; ++ void *__reg_save_area; ++}; ++ ++typedef __builtin_va_list va_list; ++ ++enum ftrace_dump_mode { ++ DUMP_NONE = 0, ++ DUMP_ALL = 1, ++ DUMP_ORIG = 2, ++}; ++ ++enum ctx_state { ++ CONTEXT_DISABLED = -1, ++ CONTEXT_KERNEL = 0, ++ CONTEXT_IDLE = 1, ++ CONTEXT_USER = 2, ++ CONTEXT_GUEST = 3, ++ CONTEXT_MAX = 4, ++}; ++ ++struct context_tracking { ++ atomic_t state; ++ long int dynticks_nesting; ++ long int dynticks_nmi_nesting; ++}; ++ ++struct atomic_notifier_head { ++ spinlock_t lock; ++ struct notifier_block *head; ++}; ++ ++struct kobj_attribute { ++ struct attribute attr; ++ ssize_t (*show)(struct kobject *, struct kobj_attribute *, char *); ++ ssize_t (*store)(struct kobject *, struct kobj_attribute *, const char *, size_t); ++}; ++ ++enum kmsg_dump_reason { ++ KMSG_DUMP_UNDEF = 0, ++ KMSG_DUMP_PANIC = 1, ++ KMSG_DUMP_OOPS = 2, ++ KMSG_DUMP_EMERG = 3, ++ KMSG_DUMP_SHUTDOWN = 4, ++ KMSG_DUMP_MAX = 5, ++}; ++ ++enum reboot_mode { ++ REBOOT_UNDEFINED = -1, ++ REBOOT_COLD = 0, ++ REBOOT_WARM = 1, ++ REBOOT_HARD = 2, ++ REBOOT_SOFT = 3, ++ REBOOT_GPIO = 4, ++}; ++ ++enum con_flush_mode { ++ CONSOLE_FLUSH_PENDING = 0, ++ CONSOLE_REPLAY_ALL = 1, ++}; ++ ++enum error_detector { ++ ERROR_DETECTOR_KFENCE = 0, ++ ERROR_DETECTOR_KASAN = 1, ++ ERROR_DETECTOR_WARN = 2, ++}; ++ ++struct warn_args { ++ const char *fmt; ++ va_list args; ++}; ++ ++typedef void (*rcu_callback_t)(struct callback_head *); ++ ++struct old_timeval32 { ++ old_time32_t tv_sec; ++ s32 tv_usec; ++}; ++ ++struct rusage { ++ struct __kernel_old_timeval ru_utime; ++ struct __kernel_old_timeval ru_stime; ++ __kernel_long_t ru_maxrss; ++ __kernel_long_t ru_ixrss; ++ __kernel_long_t ru_idrss; ++ __kernel_long_t ru_isrss; ++ __kernel_long_t ru_minflt; ++ __kernel_long_t ru_majflt; ++ __kernel_long_t ru_nswap; ++ __kernel_long_t ru_inblock; ++ __kernel_long_t ru_oublock; ++ __kernel_long_t ru_msgsnd; ++ __kernel_long_t ru_msgrcv; ++ __kernel_long_t ru_nsignals; ++ __kernel_long_t ru_nvcsw; ++ __kernel_long_t ru_nivcsw; ++}; ++ ++struct siginfo { ++ union { ++ struct { ++ int si_signo; ++ int si_errno; ++ int si_code; ++ union __sifields _sifields; ++ }; ++ int _si_pad[32]; ++ }; ++}; ++ ++struct fdtable { ++ unsigned int max_fds; ++ struct file **fd; ++ long unsigned int *close_on_exec; ++ long unsigned int *open_fds; ++ long unsigned int *full_fds_bits; ++ struct callback_head rcu; ++}; ++ ++struct files_struct { ++ atomic_t count; ++ bool resize_in_progress; ++ wait_queue_head_t resize_wait; ++ struct fdtable *fdt; ++ struct fdtable fdtab; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ spinlock_t file_lock; ++ unsigned int next_fd; ++ long unsigned int close_on_exec_init[1]; ++ long unsigned int open_fds_init[1]; ++ long unsigned int full_fds_bits_init[1]; ++ struct file *fd_array[64]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct robust_list { ++ struct robust_list *next; ++}; ++ ++struct robust_list_head { ++ struct robust_list list; ++ long int futex_offset; ++ struct robust_list *list_op_pending; ++}; ++ ++struct pipe_buffer; ++ ++struct watch_queue; ++ ++struct pipe_inode_info { ++ struct mutex mutex; ++ wait_queue_head_t rd_wait; ++ wait_queue_head_t wr_wait; ++ unsigned int head; ++ unsigned int tail; ++ unsigned int max_usage; ++ unsigned int ring_size; ++ bool note_loss; ++ unsigned int nr_accounted; ++ unsigned int readers; ++ unsigned int writers; ++ unsigned int files; ++ unsigned int r_counter; ++ unsigned int w_counter; ++ bool poll_usage; ++ struct page *tmp_page; ++ struct fasync_struct *fasync_readers; ++ struct fasync_struct *fasync_writers; ++ struct pipe_buffer *bufs; ++ struct user_struct *user; ++ struct watch_queue *watch_queue; ++}; ++ ++struct task_delay_info { ++ raw_spinlock_t lock; ++ u64 blkio_start; ++ u64 blkio_delay; ++ u64 swapin_start; ++ u64 swapin_delay; ++ u32 blkio_count; ++ u32 swapin_count; ++ u64 freepages_start; ++ u64 freepages_delay; ++ u64 thrashing_start; ++ u64 thrashing_delay; ++ u64 compact_start; ++ u64 compact_delay; ++ u64 wpcopy_start; ++ u64 wpcopy_delay; ++ u64 irq_delay; ++ u32 freepages_count; ++ u32 thrashing_count; ++ u32 compact_count; ++ u32 wpcopy_count; ++ u32 irq_count; ++}; ++ ++typedef int __kernel_rwf_t; ++ ++struct io_uring_sqe; ++ ++struct io_uring_cmd { ++ struct file *file; ++ const struct io_uring_sqe *sqe; ++ union { ++ void (*task_work_cb)(struct io_uring_cmd *, unsigned int); ++ void *cookie; ++ }; ++ u32 cmd_op; ++ u32 flags; ++ u8 pdu[32]; ++}; ++ ++typedef s32 compat_clock_t; ++ ++typedef s32 compat_pid_t; ++ ++typedef s32 compat_timer_t; ++ ++typedef s32 compat_int_t; ++ ++typedef u32 compat_uint_t; ++ ++typedef u32 compat_ulong_t; ++ ++typedef u32 __compat_uid32_t; ++ ++union compat_sigval { ++ compat_int_t sival_int; ++ compat_uptr_t sival_ptr; ++}; ++ ++typedef union compat_sigval compat_sigval_t; ++ ++struct compat_siginfo { ++ int si_signo; ++ int si_errno; ++ int si_code; ++ union { ++ int _pad[29]; ++ struct { ++ compat_pid_t _pid; ++ __compat_uid32_t _uid; ++ } _kill; ++ struct { ++ compat_timer_t _tid; ++ int _overrun; ++ compat_sigval_t _sigval; ++ } _timer; ++ struct { ++ compat_pid_t _pid; ++ __compat_uid32_t _uid; ++ compat_sigval_t _sigval; ++ } _rt; ++ struct { ++ compat_pid_t _pid; ++ __compat_uid32_t _uid; ++ int _status; ++ compat_clock_t _utime; ++ compat_clock_t _stime; ++ } _sigchld; ++ struct { ++ compat_uptr_t _addr; ++ union { ++ int _trapno; ++ short int _addr_lsb; ++ struct { ++ char _dummy_bnd[4]; ++ compat_uptr_t _lower; ++ compat_uptr_t _upper; ++ } _addr_bnd; ++ struct { ++ char _dummy_pkey[4]; ++ u32 _pkey; ++ } _addr_pkey; ++ struct { ++ compat_ulong_t _data; ++ u32 _type; ++ u32 _flags; ++ } _perf; ++ }; ++ } _sigfault; ++ struct { ++ compat_long_t _band; ++ int _fd; ++ } _sigpoll; ++ struct { ++ compat_uptr_t _call_addr; ++ int _syscall; ++ unsigned int _arch; ++ } _sigsys; ++ } _sifields; ++}; ++ ++struct compat_rusage { ++ struct old_timeval32 ru_utime; ++ struct old_timeval32 ru_stime; ++ compat_long_t ru_maxrss; ++ compat_long_t ru_ixrss; ++ compat_long_t ru_idrss; ++ compat_long_t ru_isrss; ++ compat_long_t ru_minflt; ++ compat_long_t ru_majflt; ++ compat_long_t ru_nswap; ++ compat_long_t ru_inblock; ++ compat_long_t ru_oublock; ++ compat_long_t ru_msgsnd; ++ compat_long_t ru_msgrcv; ++ compat_long_t ru_nsignals; ++ compat_long_t ru_nvcsw; ++ compat_long_t ru_nivcsw; ++}; ++ ++struct pipe_buf_operations; ++ ++struct pipe_buffer { ++ struct page *page; ++ unsigned int offset; ++ unsigned int len; ++ const struct pipe_buf_operations *ops; ++ unsigned int flags; ++ long unsigned int private; ++}; ++ ++struct pipe_buf_operations { ++ int (*confirm)(struct pipe_inode_info *, struct pipe_buffer *); ++ void (*release)(struct pipe_inode_info *, struct pipe_buffer *); ++ bool (*try_steal)(struct pipe_inode_info *, struct pipe_buffer *); ++ bool (*get)(struct pipe_inode_info *, struct pipe_buffer *); ++}; ++ ++struct io_uring_sqe { ++ __u8 opcode; ++ __u8 flags; ++ __u16 ioprio; ++ __s32 fd; ++ union { ++ __u64 off; ++ __u64 addr2; ++ struct { ++ __u32 cmd_op; ++ __u32 __pad1; ++ }; ++ }; ++ union { ++ __u64 addr; ++ __u64 splice_off_in; ++ }; ++ __u32 len; ++ union { ++ __kernel_rwf_t rw_flags; ++ __u32 fsync_flags; ++ __u16 poll_events; ++ __u32 poll32_events; ++ __u32 sync_range_flags; ++ __u32 msg_flags; ++ __u32 timeout_flags; ++ __u32 accept_flags; ++ __u32 cancel_flags; ++ __u32 open_flags; ++ __u32 statx_flags; ++ __u32 fadvise_advice; ++ __u32 splice_flags; ++ __u32 rename_flags; ++ __u32 unlink_flags; ++ __u32 hardlink_flags; ++ __u32 xattr_flags; ++ __u32 msg_ring_flags; ++ __u32 uring_cmd_flags; ++ }; ++ __u64 user_data; ++ union { ++ __u16 buf_index; ++ __u16 buf_group; ++ }; ++ __u16 personality; ++ union { ++ __s32 splice_fd_in; ++ __u32 file_index; ++ struct { ++ __u16 addr_len; ++ __u16 __pad3[1]; ++ }; ++ }; ++ union { ++ struct { ++ __u64 addr3; ++ __u64 __pad2[1]; ++ }; ++ __u8 cmd[0]; ++ }; ++}; ++ ++struct waitid_info { ++ pid_t pid; ++ uid_t uid; ++ int status; ++ int cause; ++}; ++ ++struct wait_opts { ++ enum pid_type wo_type; ++ int wo_flags; ++ struct pid *wo_pid; ++ struct waitid_info *wo_info; ++ int wo_stat; ++ struct rusage *wo_rusage; ++ wait_queue_entry_t child_wait; ++ int notask_error; ++}; ++ ++enum syscall_work_bit { ++ SYSCALL_WORK_BIT_SECCOMP = 0, ++ SYSCALL_WORK_BIT_SYSCALL_TRACEPOINT = 1, ++ SYSCALL_WORK_BIT_SYSCALL_TRACE = 2, ++ SYSCALL_WORK_BIT_SYSCALL_EMU = 3, ++ SYSCALL_WORK_BIT_SYSCALL_AUDIT = 4, ++ SYSCALL_WORK_BIT_SYSCALL_USER_DISPATCH = 5, ++ SYSCALL_WORK_BIT_SYSCALL_EXIT_TRAP = 6, ++}; ++ ++typedef struct siginfo siginfo_t; ++ ++struct sigqueue { ++ struct list_head list; ++ int flags; ++ kernel_siginfo_t info; ++ struct ucounts *ucounts; ++}; ++ ++enum { ++ FOLL_WRITE = 1, ++ FOLL_GET = 2, ++ FOLL_DUMP = 4, ++ FOLL_FORCE = 8, ++ FOLL_NOWAIT = 16, ++ FOLL_NOFAULT = 32, ++ FOLL_HWPOISON = 64, ++ FOLL_ANON = 128, ++ FOLL_LONGTERM = 256, ++ FOLL_SPLIT_PMD = 512, ++ FOLL_PCI_P2PDMA = 1024, ++ FOLL_INTERRUPTIBLE = 2048, ++ FOLL_HONOR_NUMA_FAULT = 4096, ++}; ++ ++struct wait_bit_key { ++ void *flags; ++ int bit_nr; ++ long unsigned int timeout; ++}; ++ ++typedef int wait_bit_action_f(struct wait_bit_key *, int); ++ ++struct ptrace_peeksiginfo_args { ++ __u64 off; ++ __u32 flags; ++ __s32 nr; ++}; ++ ++struct ptrace_syscall_info { ++ __u8 op; ++ __u8 pad[3]; ++ __u32 arch; ++ __u64 instruction_pointer; ++ __u64 stack_pointer; ++ union { ++ struct { ++ __u64 nr; ++ __u64 args[6]; ++ } entry; ++ struct { ++ __s64 rval; ++ __u8 is_error; ++ } exit; ++ struct { ++ __u64 nr; ++ __u64 args[6]; ++ __u32 ret_data; ++ } seccomp; ++ }; ++}; ++ ++struct ptrace_rseq_configuration { ++ __u64 rseq_abi_pointer; ++ __u32 rseq_abi_size; ++ __u32 signature; ++ __u32 flags; ++ __u32 pad; ++}; ++ ++typedef u32 compat_size_t; ++ ++struct compat_iovec { ++ compat_uptr_t iov_base; ++ compat_size_t iov_len; ++}; ++ ++typedef struct compat_siginfo compat_siginfo_t; ++ ++struct membuf { ++ void *p; ++ size_t left; ++}; ++ ++struct user_regset; ++ ++typedef int user_regset_active_fn(struct task_struct *, const struct user_regset *); ++ ++typedef int user_regset_get2_fn(struct task_struct *, const struct user_regset *, struct membuf); ++ ++typedef int user_regset_set_fn(struct task_struct *, const struct user_regset *, unsigned int, unsigned int, const void *, const void *); ++ ++typedef int user_regset_writeback_fn(struct task_struct *, const struct user_regset *, int); ++ ++struct user_regset { ++ user_regset_get2_fn *regset_get; ++ user_regset_set_fn *set; ++ user_regset_active_fn *active; ++ user_regset_writeback_fn *writeback; ++ unsigned int n; ++ unsigned int size; ++ unsigned int align; ++ unsigned int bias; ++ unsigned int core_note_type; ++}; ++ ++struct user_regset_view { ++ const char *name; ++ const struct user_regset *regsets; ++ unsigned int n; ++ u32 e_flags; ++ u16 e_machine; ++ u8 ei_osabi; ++}; ++ ++typedef void (*task_work_func_t)(struct callback_head *); ++ ++enum task_work_notify_mode { ++ TWA_NONE = 0, ++ TWA_RESUME = 1, ++ TWA_SIGNAL = 2, ++ TWA_SIGNAL_NO_IPI = 3, ++}; ++ ++struct sched_param { ++ int sched_priority; ++}; ++ ++enum { ++ __PERCPU_REF_ATOMIC = 1, ++ __PERCPU_REF_DEAD = 2, ++ __PERCPU_REF_ATOMIC_DEAD = 3, ++ __PERCPU_REF_FLAG_BITS = 2, ++}; ++ ++struct kthread_work; ++ ++typedef void (*kthread_work_func_t)(struct kthread_work *); ++ ++struct kthread_worker; ++ ++struct kthread_work { ++ struct list_head node; ++ kthread_work_func_t func; ++ struct kthread_worker *worker; ++ int canceling; ++}; ++ ++enum { ++ KTW_FREEZABLE = 1, ++}; ++ ++struct kthread_worker { ++ unsigned int flags; ++ raw_spinlock_t lock; ++ struct list_head work_list; ++ struct list_head delayed_work_list; ++ struct task_struct *task; ++ struct kthread_work *current_work; ++}; ++ ++struct kthread_delayed_work { ++ struct kthread_work work; ++ struct timer_list timer; ++}; ++ ++enum { ++ CSS_NO_REF = 1, ++ CSS_ONLINE = 2, ++ CSS_RELEASED = 4, ++ CSS_VISIBLE = 8, ++ CSS_DYING = 16, ++}; ++ ++enum hk_type { ++ HK_TYPE_TIMER = 0, ++ HK_TYPE_RCU = 1, ++ HK_TYPE_MISC = 2, ++ HK_TYPE_SCHED = 3, ++ HK_TYPE_TICK = 4, ++ HK_TYPE_DOMAIN = 5, ++ HK_TYPE_WQ = 6, ++ HK_TYPE_MANAGED_IRQ = 7, ++ HK_TYPE_KTHREAD = 8, ++ HK_TYPE_MAX = 9, ++}; ++ ++struct kthread_create_info { ++ char *full_name; ++ int (*threadfn)(void *); ++ void *data; ++ int node; ++ struct task_struct *result; ++ struct completion *done; ++ struct list_head list; ++}; ++ ++struct kthread { ++ long unsigned int flags; ++ unsigned int cpu; ++ int result; ++ int (*threadfn)(void *); ++ void *data; ++ struct completion parked; ++ struct completion exited; ++ struct cgroup_subsys_state *blkcg_css; ++ char *full_name; ++}; ++ ++enum KTHREAD_BITS { ++ KTHREAD_IS_PER_CPU = 0, ++ KTHREAD_SHOULD_STOP = 1, ++ KTHREAD_SHOULD_PARK = 2, ++}; ++ ++struct kthread_flush_work { ++ struct kthread_work work; ++ struct completion done; ++}; ++ ++typedef long unsigned int pcp_op_T_____2; ++ ++enum system_states { ++ SYSTEM_BOOTING = 0, ++ SYSTEM_SCHEDULING = 1, ++ SYSTEM_FREEING_INITMEM = 2, ++ SYSTEM_RUNNING = 3, ++ SYSTEM_HALT = 4, ++ SYSTEM_POWER_OFF = 5, ++ SYSTEM_RESTART = 6, ++ SYSTEM_SUSPEND = 7, ++}; ++ ++struct srcu_notifier_head { ++ struct mutex mutex; ++ struct srcu_usage srcuu; ++ struct srcu_struct srcu; ++ struct notifier_block *head; ++}; ++ ++struct ring_buffer_event { ++ u32 type_len: 5; ++ u32 time_delta: 27; ++ u32 array[0]; ++}; ++ ++struct trace_buffer; ++ ++struct trace_event_file; ++ ++struct trace_event_buffer { ++ struct trace_buffer *buffer; ++ struct ring_buffer_event *event; ++ struct trace_event_file *trace_file; ++ void *entry; ++ unsigned int trace_ctx; ++ struct pt_regs *regs; ++}; ++ ++struct trace_subsystem_dir; ++ ++struct trace_event_file { ++ struct list_head list; ++ struct trace_event_call *event_call; ++ struct event_filter *filter; ++ struct dentry *dir; ++ struct trace_array *tr; ++ struct trace_subsystem_dir *system; ++ struct list_head triggers; ++ long unsigned int flags; ++ atomic_t sm_ref; ++ atomic_t tm_ref; ++}; ++ ++enum { ++ EVENT_FILE_FL_ENABLED = 1, ++ EVENT_FILE_FL_RECORDED_CMD = 2, ++ EVENT_FILE_FL_RECORDED_TGID = 4, ++ EVENT_FILE_FL_FILTERED = 8, ++ EVENT_FILE_FL_NO_SET_FILTER = 16, ++ EVENT_FILE_FL_SOFT_MODE = 32, ++ EVENT_FILE_FL_SOFT_DISABLED = 64, ++ EVENT_FILE_FL_TRIGGER_MODE = 128, ++ EVENT_FILE_FL_TRIGGER_COND = 256, ++ EVENT_FILE_FL_PID_FILTER = 512, ++ EVENT_FILE_FL_WAS_ENABLED = 1024, ++}; ++ ++enum { ++ FILTER_OTHER = 0, ++ FILTER_STATIC_STRING = 1, ++ FILTER_DYN_STRING = 2, ++ FILTER_RDYN_STRING = 3, ++ FILTER_PTR_STRING = 4, ++ FILTER_TRACE_FN = 5, ++ FILTER_COMM = 6, ++ FILTER_CPU = 7, ++ FILTER_STACKTRACE = 8, ++}; ++ ++struct trace_event_raw_notifier_info { ++ struct trace_entry ent; ++ void *cb; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_notifier_info {}; ++ ++typedef void (*btf_trace_notifier_register)(void *, void *); ++ ++typedef void (*btf_trace_notifier_unregister)(void *, void *); ++ ++typedef void (*btf_trace_notifier_run)(void *, void *); ++ ++enum reboot_type { ++ BOOT_TRIPLE = 116, ++ BOOT_KBD = 107, ++ BOOT_BIOS = 98, ++ BOOT_ACPI = 97, ++ BOOT_EFI = 101, ++ BOOT_CF9_FORCE = 112, ++ BOOT_CF9_SAFE = 113, ++}; ++ ++enum sys_off_mode { ++ SYS_OFF_MODE_POWER_OFF_PREPARE = 0, ++ SYS_OFF_MODE_POWER_OFF = 1, ++ SYS_OFF_MODE_RESTART_PREPARE = 2, ++ SYS_OFF_MODE_RESTART = 3, ++}; ++ ++struct sys_off_data { ++ int mode; ++ void *cb_data; ++ const char *cmd; ++}; ++ ++typedef void (*dr_release_t)(struct device *, void *); ++ ++struct sys_off_handler { ++ struct notifier_block nb; ++ int (*sys_off_cb)(struct sys_off_data *); ++ void *cb_data; ++ enum sys_off_mode mode; ++ bool blocking; ++ void *list; ++}; ++ ++struct smp_hotplug_thread { ++ struct task_struct **store; ++ struct list_head list; ++ int (*thread_should_run)(unsigned int); ++ void (*thread_fn)(unsigned int); ++ void (*create)(unsigned int); ++ void (*setup)(unsigned int); ++ void (*cleanup)(unsigned int, bool); ++ void (*park)(unsigned int); ++ void (*unpark)(unsigned int); ++ bool selfparking; ++ const char *thread_comm; ++}; ++ ++struct smpboot_thread_data { ++ unsigned int cpu; ++ unsigned int status; ++ struct smp_hotplug_thread *ht; ++}; ++ ++enum { ++ HP_THREAD_NONE = 0, ++ HP_THREAD_ACTIVE = 1, ++ HP_THREAD_PARKED = 2, ++}; ++ ++struct ksignal { ++ struct k_sigaction ka; ++ kernel_siginfo_t info; ++ int sig; ++}; ++ ++struct kernel_clone_args { ++ u64 flags; ++ int *pidfd; ++ int *child_tid; ++ int *parent_tid; ++ const char *name; ++ int exit_signal; ++ u32 kthread: 1; ++ u32 io_thread: 1; ++ u32 user_worker: 1; ++ u32 no_files: 1; ++ long unsigned int stack; ++ long unsigned int stack_size; ++ long unsigned int tls; ++ pid_t *set_tid; ++ size_t set_tid_size; ++ int cgroup; ++ int idle; ++ int (*fn)(void *); ++ void *fn_arg; ++ struct cgroup *cgrp; ++ struct css_set *cset; ++}; ++ ++enum vhost_task_flags { ++ VHOST_TASK_FLAGS_STOP = 0, ++}; ++ ++struct vhost_task { ++ bool (*fn)(void *); ++ void *data; ++ struct completion exited; ++ long unsigned int flags; ++ struct task_struct *task; ++}; ++ ++typedef u64 uint64_t; ++ ++struct pin_cookie {}; ++ ++enum { ++ CSD_FLAG_LOCK = 1, ++ IRQ_WORK_PENDING = 1, ++ IRQ_WORK_BUSY = 2, ++ IRQ_WORK_LAZY = 4, ++ IRQ_WORK_HARD_IRQ = 8, ++ IRQ_WORK_CLAIMED = 3, ++ CSD_TYPE_ASYNC = 0, ++ CSD_TYPE_SYNC = 16, ++ CSD_TYPE_IRQ_WORK = 32, ++ CSD_TYPE_TTWU = 48, ++ CSD_FLAG_TYPE_MASK = 240, ++}; ++ ++typedef int (*task_call_f)(struct task_struct *, void *); ++ ++struct wait_bit_queue_entry { ++ struct wait_bit_key key; ++ struct wait_queue_entry wq_entry; ++}; ++ ++enum mm_cid_state { ++ MM_CID_UNSET = 4294967295, ++ MM_CID_LAZY_PUT = 2147483648, ++}; ++ ++enum hrtimer_mode { ++ HRTIMER_MODE_ABS = 0, ++ HRTIMER_MODE_REL = 1, ++ HRTIMER_MODE_PINNED = 2, ++ HRTIMER_MODE_SOFT = 4, ++ HRTIMER_MODE_HARD = 8, ++ HRTIMER_MODE_ABS_PINNED = 2, ++ HRTIMER_MODE_REL_PINNED = 3, ++ HRTIMER_MODE_ABS_SOFT = 4, ++ HRTIMER_MODE_REL_SOFT = 5, ++ HRTIMER_MODE_ABS_PINNED_SOFT = 6, ++ HRTIMER_MODE_REL_PINNED_SOFT = 7, ++ HRTIMER_MODE_ABS_HARD = 8, ++ HRTIMER_MODE_REL_HARD = 9, ++ HRTIMER_MODE_ABS_PINNED_HARD = 10, ++ HRTIMER_MODE_REL_PINNED_HARD = 11, ++}; ++ ++struct dl_bw { ++ raw_spinlock_t lock; ++ u64 bw; ++ u64 total_bw; ++}; ++ ++struct cpudl_item; ++ ++struct cpudl { ++ raw_spinlock_t lock; ++ int size; ++ cpumask_var_t free_cpus; ++ struct cpudl_item *elements; ++}; ++ ++struct cpupri_vec { ++ atomic_t count; ++ cpumask_var_t mask; ++}; ++ ++struct cpupri { ++ struct cpupri_vec pri_to_cpu[101]; ++ int *cpu_to_pri; ++}; ++ ++struct perf_domain; ++ ++struct root_domain { ++ atomic_t refcount; ++ atomic_t rto_count; ++ struct callback_head rcu; ++ cpumask_var_t span; ++ cpumask_var_t online; ++ int overload; ++ int overutilized; ++ cpumask_var_t dlo_mask; ++ atomic_t dlo_count; ++ struct dl_bw dl_bw; ++ struct cpudl cpudl; ++ u64 visit_gen; ++ struct irq_work rto_push_work; ++ raw_spinlock_t rto_lock; ++ int rto_loop; ++ int rto_cpu; ++ atomic_t rto_loop_next; ++ atomic_t rto_loop_start; ++ cpumask_var_t rto_mask; ++ struct cpupri cpupri; ++ long unsigned int max_cpu_capacity; ++ struct perf_domain *pd; ++}; ++ ++struct cfs_rq { ++ struct load_weight load; ++ unsigned int nr_running; ++ unsigned int h_nr_running; ++ unsigned int idle_nr_running; ++ unsigned int idle_h_nr_running; ++ u64 exec_clock; ++ u64 min_vruntime; ++ unsigned int forceidle_seq; ++ u64 min_vruntime_fi; ++ struct rb_root_cached tasks_timeline; ++ struct sched_entity *curr; ++ struct sched_entity *next; ++ struct sched_entity *last; ++ struct sched_entity *skip; ++ unsigned int nr_spread_over; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct sched_avg avg; ++ struct { ++ raw_spinlock_t lock; ++ int nr; ++ long unsigned int load_avg; ++ long unsigned int util_avg; ++ long unsigned int runnable_avg; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ } removed; ++ long unsigned int tg_load_avg_contrib; ++ long int propagate; ++ long int prop_runnable_sum; ++ long unsigned int h_load; ++ u64 last_h_load_update; ++ struct sched_entity *h_load_next; ++ struct rq *rq; ++ int on_list; ++ struct list_head leaf_cfs_rq_list; ++ struct task_group *tg; ++ int idle; ++ int runtime_enabled; ++ s64 runtime_remaining; ++ u64 throttled_pelt_idle; ++ u64 throttled_clock; ++ u64 throttled_clock_pelt; ++ u64 throttled_clock_pelt_time; ++ int throttled; ++ int throttle_count; ++ struct list_head throttled_list; ++ struct list_head throttled_csd_list; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct cfs_bandwidth { ++ raw_spinlock_t lock; ++ ktime_t period; ++ u64 quota; ++ u64 runtime; ++ u64 burst; ++ u64 runtime_snap; ++ s64 hierarchical_quota; ++ u8 idle; ++ u8 period_active; ++ u8 slack_started; ++ struct hrtimer period_timer; ++ struct hrtimer slack_timer; ++ struct list_head throttled_cfs_rq; ++ int nr_periods; ++ int nr_throttled; ++ int nr_burst; ++ u64 throttled_time; ++ u64 burst_time; ++}; ++ ++struct task_group { ++ struct cgroup_subsys_state css; ++ struct sched_entity **se; ++ struct cfs_rq **cfs_rq; ++ long unsigned int shares; ++ int idle; ++ long: 64; ++ long: 64; ++ long: 64; ++ atomic_long_t load_avg; ++ struct callback_head rcu; ++ struct list_head list; ++ struct task_group *parent; ++ struct list_head siblings; ++ struct list_head children; ++ struct autogroup *autogroup; ++ struct cfs_bandwidth cfs_bandwidth; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct io_ring_ctx; ++ ++struct io_wq; ++ ++struct io_uring_task { ++ int cached_refs; ++ const struct io_ring_ctx *last; ++ struct io_wq *io_wq; ++ struct file *registered_rings[16]; ++ struct xarray xa; ++ struct wait_queue_head wait; ++ atomic_t in_cancel; ++ atomic_t inflight_tracked; ++ struct percpu_counter inflight; ++ long: 64; ++ long: 64; ++ struct { ++ struct llist_head task_list; ++ struct callback_head task_work; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ }; ++}; ++ ++enum { ++ MEMBARRIER_STATE_PRIVATE_EXPEDITED_READY = 1, ++ MEMBARRIER_STATE_PRIVATE_EXPEDITED = 2, ++ MEMBARRIER_STATE_GLOBAL_EXPEDITED_READY = 4, ++ MEMBARRIER_STATE_GLOBAL_EXPEDITED = 8, ++ MEMBARRIER_STATE_PRIVATE_EXPEDITED_SYNC_CORE_READY = 16, ++ MEMBARRIER_STATE_PRIVATE_EXPEDITED_SYNC_CORE = 32, ++ MEMBARRIER_STATE_PRIVATE_EXPEDITED_RSEQ_READY = 64, ++ MEMBARRIER_STATE_PRIVATE_EXPEDITED_RSEQ = 128, ++}; ++ ++struct autogroup { ++ struct kref kref; ++ struct task_group *tg; ++ struct rw_semaphore lock; ++ long unsigned int id; ++ int nice; ++}; ++ ++struct trace_print_flags { ++ long unsigned int mask; ++ const char *name; ++}; ++ ++struct sched_domain_attr { ++ int relax_domain_level; ++}; ++ ++struct sched_group { ++ struct sched_group *next; ++ atomic_t ref; ++ unsigned int group_weight; ++ struct sched_group_capacity *sgc; ++ int asym_prefer_cpu; ++ int flags; ++ long unsigned int cpumask[0]; ++}; ++ ++struct sched_group_capacity { ++ atomic_t ref; ++ long unsigned int capacity; ++ long unsigned int min_capacity; ++ long unsigned int max_capacity; ++ long unsigned int next_update; ++ int imbalance; ++ int id; ++ long unsigned int cpumask[0]; ++}; ++ ++struct em_perf_state { ++ long unsigned int frequency; ++ long unsigned int power; ++ long unsigned int cost; ++ long unsigned int flags; ++}; ++ ++struct em_perf_domain { ++ struct em_perf_state *table; ++ int nr_perf_states; ++ long unsigned int flags; ++ long unsigned int cpus[0]; ++}; ++ ++struct kernel_cpustat { ++ u64 cpustat[11]; ++}; ++ ++struct kernel_stat { ++ long unsigned int irqs_sum; ++ unsigned int softirqs[10]; ++}; ++ ++enum { ++ CFTYPE_ONLY_ON_ROOT = 1, ++ CFTYPE_NOT_ON_ROOT = 2, ++ CFTYPE_NS_DELEGATABLE = 4, ++ CFTYPE_NO_PREFIX = 8, ++ CFTYPE_WORLD_WRITABLE = 16, ++ CFTYPE_DEBUG = 32, ++ __CFTYPE_ONLY_ON_DFL = 65536, ++ __CFTYPE_NOT_ON_DFL = 131072, ++ __CFTYPE_ADDED = 262144, ++}; ++ ++struct wake_q_head { ++ struct wake_q_node *first; ++ struct wake_q_node **lastp; ++}; ++ ++struct sched_attr { ++ __u32 size; ++ __u32 sched_policy; ++ __u64 sched_flags; ++ __s32 sched_nice; ++ __u32 sched_priority; ++ __u64 sched_runtime; ++ __u64 sched_deadline; ++ __u64 sched_period; ++ __u32 sched_util_min; ++ __u32 sched_util_max; ++}; ++ ++struct trace_event_raw_sched_kthread_stop { ++ struct trace_entry ent; ++ char comm[16]; ++ pid_t pid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_kthread_stop_ret { ++ struct trace_entry ent; ++ int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_kthread_work_queue_work { ++ struct trace_entry ent; ++ void *work; ++ void *function; ++ void *worker; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_kthread_work_execute_start { ++ struct trace_entry ent; ++ void *work; ++ void *function; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_kthread_work_execute_end { ++ struct trace_entry ent; ++ void *work; ++ void *function; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_wakeup_template { ++ struct trace_entry ent; ++ char comm[16]; ++ pid_t pid; ++ int prio; ++ int target_cpu; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_switch { ++ struct trace_entry ent; ++ char prev_comm[16]; ++ pid_t prev_pid; ++ int prev_prio; ++ long int prev_state; ++ char next_comm[16]; ++ pid_t next_pid; ++ int next_prio; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_migrate_task { ++ struct trace_entry ent; ++ char comm[16]; ++ pid_t pid; ++ int prio; ++ int orig_cpu; ++ int dest_cpu; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_process_template { ++ struct trace_entry ent; ++ char comm[16]; ++ pid_t pid; ++ int prio; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_process_wait { ++ struct trace_entry ent; ++ char comm[16]; ++ pid_t pid; ++ int prio; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_process_fork { ++ struct trace_entry ent; ++ char parent_comm[16]; ++ pid_t parent_pid; ++ char child_comm[16]; ++ pid_t child_pid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_process_exec { ++ struct trace_entry ent; ++ u32 __data_loc_filename; ++ pid_t pid; ++ pid_t old_pid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_stat_template { ++ struct trace_entry ent; ++ char comm[16]; ++ pid_t pid; ++ u64 delay; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_stat_runtime { ++ struct trace_entry ent; ++ char comm[16]; ++ pid_t pid; ++ u64 runtime; ++ u64 vruntime; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_pi_setprio { ++ struct trace_entry ent; ++ char comm[16]; ++ pid_t pid; ++ int oldprio; ++ int newprio; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_move_numa { ++ struct trace_entry ent; ++ pid_t pid; ++ pid_t tgid; ++ pid_t ngid; ++ int src_cpu; ++ int src_nid; ++ int dst_cpu; ++ int dst_nid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_numa_pair_template { ++ struct trace_entry ent; ++ pid_t src_pid; ++ pid_t src_tgid; ++ pid_t src_ngid; ++ int src_cpu; ++ int src_nid; ++ pid_t dst_pid; ++ pid_t dst_tgid; ++ pid_t dst_ngid; ++ int dst_cpu; ++ int dst_nid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sched_wake_idle_without_ipi { ++ struct trace_entry ent; ++ int cpu; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_sched_kthread_stop {}; ++ ++struct trace_event_data_offsets_sched_kthread_stop_ret {}; ++ ++struct trace_event_data_offsets_sched_kthread_work_queue_work {}; ++ ++struct trace_event_data_offsets_sched_kthread_work_execute_start {}; ++ ++struct trace_event_data_offsets_sched_kthread_work_execute_end {}; ++ ++struct trace_event_data_offsets_sched_wakeup_template {}; ++ ++struct trace_event_data_offsets_sched_switch {}; ++ ++struct trace_event_data_offsets_sched_migrate_task {}; ++ ++struct trace_event_data_offsets_sched_process_template {}; ++ ++struct trace_event_data_offsets_sched_process_wait {}; ++ ++struct trace_event_data_offsets_sched_process_fork {}; ++ ++struct trace_event_data_offsets_sched_process_exec { ++ u32 filename; ++}; ++ ++struct trace_event_data_offsets_sched_stat_template {}; ++ ++struct trace_event_data_offsets_sched_stat_runtime {}; ++ ++struct trace_event_data_offsets_sched_pi_setprio {}; ++ ++struct trace_event_data_offsets_sched_move_numa {}; ++ ++struct trace_event_data_offsets_sched_numa_pair_template {}; ++ ++struct trace_event_data_offsets_sched_wake_idle_without_ipi {}; ++ ++typedef void (*btf_trace_sched_kthread_stop)(void *, struct task_struct *); ++ ++typedef void (*btf_trace_sched_kthread_stop_ret)(void *, int); ++ ++typedef void (*btf_trace_sched_kthread_work_queue_work)(void *, struct kthread_worker *, struct kthread_work *); ++ ++typedef void (*btf_trace_sched_kthread_work_execute_start)(void *, struct kthread_work *); ++ ++typedef void (*btf_trace_sched_kthread_work_execute_end)(void *, struct kthread_work *, kthread_work_func_t); ++ ++typedef void (*btf_trace_sched_waking)(void *, struct task_struct *); ++ ++typedef void (*btf_trace_sched_wakeup)(void *, struct task_struct *); ++ ++typedef void (*btf_trace_sched_wakeup_new)(void *, struct task_struct *); ++ ++typedef void (*btf_trace_sched_switch)(void *, bool, struct task_struct *, struct task_struct *, unsigned int); ++ ++typedef void (*btf_trace_sched_migrate_task)(void *, struct task_struct *, int); ++ ++typedef void (*btf_trace_sched_process_free)(void *, struct task_struct *); ++ ++typedef void (*btf_trace_sched_process_exit)(void *, struct task_struct *); ++ ++typedef void (*btf_trace_sched_wait_task)(void *, struct task_struct *); ++ ++typedef void (*btf_trace_sched_process_wait)(void *, struct pid *); ++ ++typedef void (*btf_trace_sched_process_fork)(void *, struct task_struct *, struct task_struct *); ++ ++typedef void (*btf_trace_sched_process_exec)(void *, struct task_struct *, pid_t, struct linux_binprm *); ++ ++typedef void (*btf_trace_sched_stat_wait)(void *, struct task_struct *, u64); ++ ++typedef void (*btf_trace_sched_stat_sleep)(void *, struct task_struct *, u64); ++ ++typedef void (*btf_trace_sched_stat_iowait)(void *, struct task_struct *, u64); ++ ++typedef void (*btf_trace_sched_stat_blocked)(void *, struct task_struct *, u64); ++ ++typedef void (*btf_trace_sched_stat_runtime)(void *, struct task_struct *, u64, u64); ++ ++typedef void (*btf_trace_sched_pi_setprio)(void *, struct task_struct *, struct task_struct *); ++ ++typedef void (*btf_trace_sched_move_numa)(void *, struct task_struct *, int, int); ++ ++typedef void (*btf_trace_sched_stick_numa)(void *, struct task_struct *, int, struct task_struct *, int); ++ ++typedef void (*btf_trace_sched_swap_numa)(void *, struct task_struct *, int, struct task_struct *, int); ++ ++typedef void (*btf_trace_sched_wake_idle_without_ipi)(void *, int); ++ ++typedef void (*btf_trace_pelt_cfs_tp)(void *, struct cfs_rq *); ++ ++typedef void (*btf_trace_pelt_rt_tp)(void *, struct rq *); ++ ++struct rt_prio_array { ++ long unsigned int bitmap[2]; ++ struct list_head queue[100]; ++}; ++ ++struct rt_rq { ++ struct rt_prio_array active; ++ unsigned int rt_nr_running; ++ unsigned int rr_nr_running; ++ struct { ++ int curr; ++ int next; ++ } highest_prio; ++ unsigned int rt_nr_migratory; ++ unsigned int rt_nr_total; ++ int overloaded; ++ struct plist_head pushable_tasks; ++ int rt_queued; ++ int rt_throttled; ++ u64 rt_time; ++ u64 rt_runtime; ++ raw_spinlock_t rt_runtime_lock; ++}; ++ ++struct dl_rq { ++ struct rb_root_cached root; ++ unsigned int dl_nr_running; ++ struct { ++ u64 curr; ++ u64 next; ++ } earliest_dl; ++ unsigned int dl_nr_migratory; ++ int overloaded; ++ struct rb_root_cached pushable_dl_tasks_root; ++ u64 running_bw; ++ u64 this_bw; ++ u64 extra_bw; ++ u64 bw_ratio; ++}; ++ ++struct cpu_stop_done; ++ ++struct cpu_stop_work { ++ struct list_head list; ++ cpu_stop_fn_t fn; ++ long unsigned int caller; ++ void *arg; ++ struct cpu_stop_done *done; ++}; ++ ++struct balance_callback; ++ ++struct rq { ++ raw_spinlock_t __lock; ++ unsigned int nr_running; ++ unsigned int nr_numa_running; ++ unsigned int nr_preferred_running; ++ unsigned int numa_migrate_on; ++ long unsigned int last_blocked_load_update_tick; ++ unsigned int has_blocked_load; ++ long: 64; ++ long: 64; ++ long: 64; ++ call_single_data_t nohz_csd; ++ unsigned int nohz_tick_stopped; ++ atomic_t nohz_flags; ++ unsigned int ttwu_pending; ++ u64 nr_switches; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct cfs_rq cfs; ++ struct rt_rq rt; ++ struct dl_rq dl; ++ struct list_head leaf_cfs_rq_list; ++ struct list_head *tmp_alone_branch; ++ unsigned int nr_uninterruptible; ++ struct task_struct *curr; ++ struct task_struct *idle; ++ struct task_struct *stop; ++ long unsigned int next_balance; ++ struct mm_struct *prev_mm; ++ unsigned int clock_update_flags; ++ u64 clock; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ u64 clock_task; ++ u64 clock_pelt; ++ long unsigned int lost_idle_time; ++ u64 clock_pelt_idle; ++ u64 clock_idle; ++ atomic_t nr_iowait; ++ u64 last_seen_need_resched_ns; ++ int ticks_without_resched; ++ int membarrier_state; ++ struct root_domain *rd; ++ struct sched_domain *sd; ++ long unsigned int cpu_capacity; ++ long unsigned int cpu_capacity_orig; ++ struct balance_callback *balance_callback; ++ unsigned char nohz_idle_balance; ++ unsigned char idle_balance; ++ long unsigned int misfit_task_load; ++ int active_balance; ++ int push_cpu; ++ struct cpu_stop_work active_balance_work; ++ int cpu; ++ int online; ++ struct list_head cfs_tasks; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct sched_avg avg_rt; ++ struct sched_avg avg_dl; ++ u64 idle_stamp; ++ u64 avg_idle; ++ long unsigned int wake_stamp; ++ u64 wake_avg_idle; ++ u64 max_idle_balance_cost; ++ struct rcuwait hotplug_wait; ++ long unsigned int calc_load_update; ++ long int calc_load_active; ++ call_single_data_t hrtick_csd; ++ struct hrtimer hrtick_timer; ++ ktime_t hrtick_time; ++ struct sched_info rq_sched_info; ++ long long unsigned int rq_cpu_time; ++ unsigned int yld_count; ++ unsigned int sched_count; ++ unsigned int sched_goidle; ++ unsigned int ttwu_count; ++ unsigned int ttwu_local; ++ unsigned int nr_pinned; ++ unsigned int push_busy; ++ struct cpu_stop_work push_work; ++ struct rq *core; ++ struct task_struct *core_pick; ++ unsigned int core_enabled; ++ unsigned int core_sched_seq; ++ struct rb_root core_tree; ++ unsigned int core_task_seq; ++ unsigned int core_pick_seq; ++ long unsigned int core_cookie; ++ unsigned int core_forceidle_count; ++ unsigned int core_forceidle_seq; ++ unsigned int core_forceidle_occupation; ++ u64 core_forceidle_start; ++ cpumask_var_t scratch_mask; ++ long: 64; ++ long: 64; ++ long: 64; ++ call_single_data_t cfsb_csd; ++ struct list_head cfsb_csd_list; ++ long: 64; ++ long: 64; ++}; ++ ++typedef void (*btf_trace_pelt_dl_tp)(void *, struct rq *); ++ ++typedef void (*btf_trace_pelt_thermal_tp)(void *, struct rq *); ++ ++typedef void (*btf_trace_pelt_irq_tp)(void *, struct rq *); ++ ++typedef void (*btf_trace_pelt_se_tp)(void *, struct sched_entity *); ++ ++typedef void (*btf_trace_sched_cpu_capacity_tp)(void *, struct rq *); ++ ++typedef void (*btf_trace_sched_overutilized_tp)(void *, struct root_domain *, bool); ++ ++typedef void (*btf_trace_sched_util_est_cfs_tp)(void *, struct cfs_rq *); ++ ++typedef void (*btf_trace_sched_util_est_se_tp)(void *, struct sched_entity *); ++ ++typedef void (*btf_trace_sched_update_nr_running_tp)(void *, struct rq *, int); ++ ++struct trace_event_raw_ipi_raise { ++ struct trace_entry ent; ++ u32 __data_loc_target_cpus; ++ const char *reason; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ipi_send_cpu { ++ struct trace_entry ent; ++ unsigned int cpu; ++ void *callsite; ++ void *callback; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ipi_send_cpumask { ++ struct trace_entry ent; ++ u32 __data_loc_cpumask; ++ void *callsite; ++ void *callback; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ipi_handler { ++ struct trace_entry ent; ++ const char *reason; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_ipi_raise { ++ u32 target_cpus; ++}; ++ ++struct trace_event_data_offsets_ipi_send_cpu {}; ++ ++struct trace_event_data_offsets_ipi_send_cpumask { ++ u32 cpumask; ++}; ++ ++struct trace_event_data_offsets_ipi_handler {}; ++ ++typedef void (*btf_trace_ipi_raise)(void *, const struct cpumask *, const char *); ++ ++typedef void (*btf_trace_ipi_send_cpu)(void *, const unsigned int, long unsigned int, void *); ++ ++typedef void (*btf_trace_ipi_send_cpumask)(void *, const struct cpumask *, long unsigned int, void *); ++ ++typedef void (*btf_trace_ipi_entry)(void *, const char *); ++ ++typedef void (*btf_trace_ipi_exit)(void *, const char *); ++ ++struct cpudl_item { ++ u64 dl; ++ int cpu; ++ int idx; ++}; ++ ++struct rt_bandwidth { ++ raw_spinlock_t rt_runtime_lock; ++ ktime_t rt_period; ++ u64 rt_runtime; ++ struct hrtimer rt_period_timer; ++ unsigned int rt_period_active; ++}; ++ ++typedef int (*tg_visitor)(struct task_group *, void *); ++ ++struct perf_domain { ++ struct em_perf_domain *em_pd; ++ struct perf_domain *next; ++ struct callback_head rcu; ++}; ++ ++struct balance_callback { ++ struct balance_callback *next; ++ void (*func)(struct rq *); ++}; ++ ++struct rq_flags { ++ long unsigned int flags; ++ struct pin_cookie cookie; ++ unsigned int clock_update_flags; ++}; ++ ++struct sched_entity_stats { ++ struct sched_entity se; ++ struct sched_statistics stats; ++}; ++ ++enum { ++ __SCHED_FEAT_GENTLE_FAIR_SLEEPERS = 0, ++ __SCHED_FEAT_START_DEBIT = 1, ++ __SCHED_FEAT_NEXT_BUDDY = 2, ++ __SCHED_FEAT_LAST_BUDDY = 3, ++ __SCHED_FEAT_CACHE_HOT_BUDDY = 4, ++ __SCHED_FEAT_WAKEUP_PREEMPTION = 5, ++ __SCHED_FEAT_HRTICK = 6, ++ __SCHED_FEAT_HRTICK_DL = 7, ++ __SCHED_FEAT_DOUBLE_TICK = 8, ++ __SCHED_FEAT_NONTASK_CAPACITY = 9, ++ __SCHED_FEAT_TTWU_QUEUE = 10, ++ __SCHED_FEAT_SIS_PROP = 11, ++ __SCHED_FEAT_SIS_UTIL = 12, ++ __SCHED_FEAT_WARN_DOUBLE_CLOCK = 13, ++ __SCHED_FEAT_RT_PUSH_IPI = 14, ++ __SCHED_FEAT_RT_RUNTIME_SHARE = 15, ++ __SCHED_FEAT_LB_MIN = 16, ++ __SCHED_FEAT_ATTACH_AGE_LOAD = 17, ++ __SCHED_FEAT_WA_IDLE = 18, ++ __SCHED_FEAT_WA_WEIGHT = 19, ++ __SCHED_FEAT_WA_BIAS = 20, ++ __SCHED_FEAT_UTIL_EST = 21, ++ __SCHED_FEAT_UTIL_EST_FASTUP = 22, ++ __SCHED_FEAT_LATENCY_WARN = 23, ++ __SCHED_FEAT_ALT_PERIOD = 24, ++ __SCHED_FEAT_BASE_SLICE = 25, ++ __SCHED_FEAT_NR = 26, ++}; ++ ++struct affinity_context { ++ const struct cpumask *new_mask; ++ struct cpumask *user_mask; ++ unsigned int flags; ++}; ++ ++enum cpu_util_type { ++ FREQUENCY_UTIL = 0, ++ ENERGY_UTIL = 1, ++}; ++ ++enum { ++ IOSQE_FIXED_FILE_BIT = 0, ++ IOSQE_IO_DRAIN_BIT = 1, ++ IOSQE_IO_LINK_BIT = 2, ++ IOSQE_IO_HARDLINK_BIT = 3, ++ IOSQE_ASYNC_BIT = 4, ++ IOSQE_BUFFER_SELECT_BIT = 5, ++ IOSQE_CQE_SKIP_SUCCESS_BIT = 6, ++}; ++ ++enum io_uring_op { ++ IORING_OP_NOP = 0, ++ IORING_OP_READV = 1, ++ IORING_OP_WRITEV = 2, ++ IORING_OP_FSYNC = 3, ++ IORING_OP_READ_FIXED = 4, ++ IORING_OP_WRITE_FIXED = 5, ++ IORING_OP_POLL_ADD = 6, ++ IORING_OP_POLL_REMOVE = 7, ++ IORING_OP_SYNC_FILE_RANGE = 8, ++ IORING_OP_SENDMSG = 9, ++ IORING_OP_RECVMSG = 10, ++ IORING_OP_TIMEOUT = 11, ++ IORING_OP_TIMEOUT_REMOVE = 12, ++ IORING_OP_ACCEPT = 13, ++ IORING_OP_ASYNC_CANCEL = 14, ++ IORING_OP_LINK_TIMEOUT = 15, ++ IORING_OP_CONNECT = 16, ++ IORING_OP_FALLOCATE = 17, ++ IORING_OP_OPENAT = 18, ++ IORING_OP_CLOSE = 19, ++ IORING_OP_FILES_UPDATE = 20, ++ IORING_OP_STATX = 21, ++ IORING_OP_READ = 22, ++ IORING_OP_WRITE = 23, ++ IORING_OP_FADVISE = 24, ++ IORING_OP_MADVISE = 25, ++ IORING_OP_SEND = 26, ++ IORING_OP_RECV = 27, ++ IORING_OP_OPENAT2 = 28, ++ IORING_OP_EPOLL_CTL = 29, ++ IORING_OP_SPLICE = 30, ++ IORING_OP_PROVIDE_BUFFERS = 31, ++ IORING_OP_REMOVE_BUFFERS = 32, ++ IORING_OP_TEE = 33, ++ IORING_OP_SHUTDOWN = 34, ++ IORING_OP_RENAMEAT = 35, ++ IORING_OP_UNLINKAT = 36, ++ IORING_OP_MKDIRAT = 37, ++ IORING_OP_SYMLINKAT = 38, ++ IORING_OP_LINKAT = 39, ++ IORING_OP_MSG_RING = 40, ++ IORING_OP_FSETXATTR = 41, ++ IORING_OP_SETXATTR = 42, ++ IORING_OP_FGETXATTR = 43, ++ IORING_OP_GETXATTR = 44, ++ IORING_OP_SOCKET = 45, ++ IORING_OP_URING_CMD = 46, ++ IORING_OP_SEND_ZC = 47, ++ IORING_OP_SENDMSG_ZC = 48, ++ IORING_OP_LAST = 49, ++}; ++ ++struct io_uring_cqe { ++ __u64 user_data; ++ __s32 res; ++ __u32 flags; ++ __u64 big_cqe[0]; ++}; ++ ++enum { ++ IORING_REGISTER_BUFFERS = 0, ++ IORING_UNREGISTER_BUFFERS = 1, ++ IORING_REGISTER_FILES = 2, ++ IORING_UNREGISTER_FILES = 3, ++ IORING_REGISTER_EVENTFD = 4, ++ IORING_UNREGISTER_EVENTFD = 5, ++ IORING_REGISTER_FILES_UPDATE = 6, ++ IORING_REGISTER_EVENTFD_ASYNC = 7, ++ IORING_REGISTER_PROBE = 8, ++ IORING_REGISTER_PERSONALITY = 9, ++ IORING_UNREGISTER_PERSONALITY = 10, ++ IORING_REGISTER_RESTRICTIONS = 11, ++ IORING_REGISTER_ENABLE_RINGS = 12, ++ IORING_REGISTER_FILES2 = 13, ++ IORING_REGISTER_FILES_UPDATE2 = 14, ++ IORING_REGISTER_BUFFERS2 = 15, ++ IORING_REGISTER_BUFFERS_UPDATE = 16, ++ IORING_REGISTER_IOWQ_AFF = 17, ++ IORING_UNREGISTER_IOWQ_AFF = 18, ++ IORING_REGISTER_IOWQ_MAX_WORKERS = 19, ++ IORING_REGISTER_RING_FDS = 20, ++ IORING_UNREGISTER_RING_FDS = 21, ++ IORING_REGISTER_PBUF_RING = 22, ++ IORING_UNREGISTER_PBUF_RING = 23, ++ IORING_REGISTER_SYNC_CANCEL = 24, ++ IORING_REGISTER_FILE_ALLOC_RANGE = 25, ++ IORING_REGISTER_LAST = 26, ++ IORING_REGISTER_USE_REGISTERED_RING = 2147483648, ++}; ++ ++struct io_wq_work_node { ++ struct io_wq_work_node *next; ++}; ++ ++struct io_wq_work_list { ++ struct io_wq_work_node *first; ++ struct io_wq_work_node *last; ++}; ++ ++struct io_wq_work { ++ struct io_wq_work_node list; ++ unsigned int flags; ++ int cancel_seq; ++}; ++ ++struct io_fixed_file { ++ long unsigned int file_ptr; ++}; ++ ++struct io_file_table { ++ struct io_fixed_file *files; ++ long unsigned int *bitmap; ++ unsigned int alloc_hint; ++}; ++ ++struct io_hash_bucket { ++ spinlock_t lock; ++ struct hlist_head list; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct io_hash_table { ++ struct io_hash_bucket *hbs; ++ unsigned int hash_bits; ++}; ++ ++struct io_kiocb; ++ ++struct io_submit_link { ++ struct io_kiocb *head; ++ struct io_kiocb *last; ++}; ++ ++struct io_submit_state { ++ struct io_wq_work_node free_list; ++ struct io_wq_work_list compl_reqs; ++ struct io_submit_link link; ++ bool plug_started; ++ bool need_plug; ++ short unsigned int submit_nr; ++ unsigned int cqes_count; ++ struct blk_plug plug; ++ struct io_uring_cqe cqes[16]; ++}; ++ ++struct io_alloc_cache { ++ struct io_wq_work_node list; ++ unsigned int nr_cached; ++ unsigned int max_cached; ++ size_t elem_size; ++}; ++ ++struct io_restriction { ++ long unsigned int register_op[1]; ++ long unsigned int sqe_op[1]; ++ u8 sqe_flags_allowed; ++ u8 sqe_flags_required; ++ bool registered; ++}; ++ ++struct io_rings; ++ ++struct io_rsrc_node; ++ ++struct io_mapped_ubuf; ++ ++struct io_buffer_list; ++ ++struct io_sq_data; ++ ++struct io_ev_fd; ++ ++struct io_rsrc_data; ++ ++struct io_wq_hash; ++ ++struct io_ring_ctx { ++ struct { ++ unsigned int flags; ++ unsigned int drain_next: 1; ++ unsigned int restricted: 1; ++ unsigned int off_timeout_used: 1; ++ unsigned int drain_active: 1; ++ unsigned int has_evfd: 1; ++ unsigned int task_complete: 1; ++ unsigned int syscall_iopoll: 1; ++ unsigned int poll_activated: 1; ++ unsigned int drain_disabled: 1; ++ unsigned int compat: 1; ++ enum task_work_notify_mode notify_method; ++ struct io_rings *rings; ++ struct task_struct *submitter_task; ++ struct percpu_ref refs; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ }; ++ struct { ++ struct mutex uring_lock; ++ u32 *sq_array; ++ struct io_uring_sqe *sq_sqes; ++ unsigned int cached_sq_head; ++ unsigned int sq_entries; ++ struct io_rsrc_node *rsrc_node; ++ atomic_t cancel_seq; ++ struct io_file_table file_table; ++ unsigned int nr_user_files; ++ unsigned int nr_user_bufs; ++ struct io_mapped_ubuf **user_bufs; ++ struct io_submit_state submit_state; ++ struct io_buffer_list *io_bl; ++ struct xarray io_bl_xa; ++ struct list_head io_buffers_cache; ++ struct io_hash_table cancel_table_locked; ++ struct list_head cq_overflow_list; ++ struct io_alloc_cache apoll_cache; ++ struct io_alloc_cache netmsg_cache; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ }; ++ struct io_wq_work_list locked_free_list; ++ unsigned int locked_free_nr; ++ const struct cred *sq_creds; ++ struct io_sq_data *sq_data; ++ struct wait_queue_head sqo_sq_wait; ++ struct list_head sqd_list; ++ long unsigned int check_cq; ++ unsigned int file_alloc_start; ++ unsigned int file_alloc_end; ++ struct xarray personalities; ++ u32 pers_next; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct { ++ struct io_uring_cqe *cqe_cached; ++ struct io_uring_cqe *cqe_sentinel; ++ unsigned int cached_cq_tail; ++ unsigned int cq_entries; ++ struct io_ev_fd *io_ev_fd; ++ struct wait_queue_head cq_wait; ++ unsigned int cq_extra; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ }; ++ struct { ++ spinlock_t completion_lock; ++ bool poll_multi_queue; ++ atomic_t cq_wait_nr; ++ struct io_wq_work_list iopoll_list; ++ struct io_hash_table cancel_table; ++ struct llist_head work_llist; ++ struct list_head io_buffers_comp; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ }; ++ struct { ++ spinlock_t timeout_lock; ++ atomic_t cq_timeouts; ++ struct list_head timeout_list; ++ struct list_head ltimeout_list; ++ unsigned int cq_last_tm_flush; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ }; ++ struct wait_queue_head poll_wq; ++ struct io_restriction restrictions; ++ struct io_mapped_ubuf *dummy_ubuf; ++ struct io_rsrc_data *file_data; ++ struct io_rsrc_data *buf_data; ++ struct list_head rsrc_ref_list; ++ struct io_alloc_cache rsrc_node_cache; ++ struct wait_queue_head rsrc_quiesce_wq; ++ unsigned int rsrc_quiesce; ++ struct list_head io_buffers_pages; ++ struct socket *ring_sock; ++ struct io_wq_hash *hash_map; ++ struct user_struct *user; ++ struct mm_struct *mm_account; ++ struct llist_head fallback_llist; ++ struct delayed_work fallback_work; ++ struct work_struct exit_work; ++ struct list_head tctx_list; ++ struct completion ref_comp; ++ u32 iowq_limits[2]; ++ bool iowq_limits_set; ++ struct callback_head poll_wq_task_work; ++ struct list_head defer_list; ++ unsigned int sq_thread_idle; ++ unsigned int evfd_last_cq_tail; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct io_uring { ++ u32 head; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ u32 tail; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct io_rings { ++ struct io_uring sq; ++ struct io_uring cq; ++ u32 sq_ring_mask; ++ u32 cq_ring_mask; ++ u32 sq_ring_entries; ++ u32 cq_ring_entries; ++ u32 sq_dropped; ++ atomic_t sq_flags; ++ u32 cq_flags; ++ u32 cq_overflow; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct io_uring_cqe cqes[0]; ++}; ++ ++struct io_cmd_data { ++ struct file *file; ++ __u8 data[56]; ++}; ++ ++struct io_cqe { ++ __u64 user_data; ++ __s32 res; ++ union { ++ __u32 flags; ++ int fd; ++ }; ++}; ++ ++struct io_tw_state; ++ ++typedef void (*io_req_tw_func_t)(struct io_kiocb *, struct io_tw_state *); ++ ++struct io_task_work { ++ struct llist_node node; ++ io_req_tw_func_t func; ++}; ++ ++struct io_buffer; ++ ++struct async_poll; ++ ++struct io_kiocb { ++ union { ++ struct file *file; ++ struct io_cmd_data cmd; ++ }; ++ u8 opcode; ++ u8 iopoll_completed; ++ u16 buf_index; ++ unsigned int flags; ++ struct io_cqe cqe; ++ struct io_ring_ctx *ctx; ++ struct task_struct *task; ++ struct io_rsrc_node *rsrc_node; ++ union { ++ struct io_mapped_ubuf *imu; ++ struct io_buffer *kbuf; ++ struct io_buffer_list *buf_list; ++ }; ++ union { ++ struct io_wq_work_node comp_list; ++ __poll_t apoll_events; ++ }; ++ atomic_t refs; ++ atomic_t poll_refs; ++ struct io_task_work io_task_work; ++ unsigned int nr_tw; ++ union { ++ struct hlist_node hash_node; ++ struct { ++ u64 extra1; ++ u64 extra2; ++ }; ++ }; ++ struct async_poll *apoll; ++ void *async_data; ++ struct io_kiocb *link; ++ const struct cred *creds; ++ struct io_wq_work work; ++}; ++ ++struct io_ev_fd { ++ struct eventfd_ctx *cq_ev_fd; ++ unsigned int eventfd_async: 1; ++ struct callback_head rcu; ++ atomic_t refs; ++ atomic_t ops; ++}; ++ ++struct io_wq_hash { ++ refcount_t refs; ++ long unsigned int map; ++ struct wait_queue_head wait; ++}; ++ ++struct io_tw_state { ++ bool locked; ++}; ++ ++enum { ++ REQ_F_FIXED_FILE_BIT = 0, ++ REQ_F_IO_DRAIN_BIT = 1, ++ REQ_F_LINK_BIT = 2, ++ REQ_F_HARDLINK_BIT = 3, ++ REQ_F_FORCE_ASYNC_BIT = 4, ++ REQ_F_BUFFER_SELECT_BIT = 5, ++ REQ_F_CQE_SKIP_BIT = 6, ++ REQ_F_FAIL_BIT = 8, ++ REQ_F_INFLIGHT_BIT = 9, ++ REQ_F_CUR_POS_BIT = 10, ++ REQ_F_NOWAIT_BIT = 11, ++ REQ_F_LINK_TIMEOUT_BIT = 12, ++ REQ_F_NEED_CLEANUP_BIT = 13, ++ REQ_F_POLLED_BIT = 14, ++ REQ_F_BUFFER_SELECTED_BIT = 15, ++ REQ_F_BUFFER_RING_BIT = 16, ++ REQ_F_REISSUE_BIT = 17, ++ REQ_F_CREDS_BIT = 18, ++ REQ_F_REFCOUNT_BIT = 19, ++ REQ_F_ARM_LTIMEOUT_BIT = 20, ++ REQ_F_ASYNC_DATA_BIT = 21, ++ REQ_F_SKIP_LINK_CQES_BIT = 22, ++ REQ_F_SINGLE_POLL_BIT = 23, ++ REQ_F_DOUBLE_POLL_BIT = 24, ++ REQ_F_PARTIAL_IO_BIT = 25, ++ REQ_F_CQE32_INIT_BIT = 26, ++ REQ_F_APOLL_MULTISHOT_BIT = 27, ++ REQ_F_CLEAR_POLLIN_BIT = 28, ++ REQ_F_HASH_LOCKED_BIT = 29, ++ REQ_F_SUPPORT_NOWAIT_BIT = 30, ++ REQ_F_ISREG_BIT = 31, ++ __REQ_F_LAST_BIT = 32, ++}; ++ ++struct set_affinity_pending; ++ ++struct migration_arg { ++ struct task_struct *task; ++ int dest_cpu; ++ struct set_affinity_pending *pending; ++}; ++ ++struct set_affinity_pending { ++ refcount_t refs; ++ unsigned int stop_pending; ++ struct completion done; ++ struct cpu_stop_work stop_work; ++ struct migration_arg arg; ++}; ++ ++struct migration_swap_arg { ++ struct task_struct *src_task; ++ struct task_struct *dst_task; ++ int src_cpu; ++ int dst_cpu; ++}; ++ ++struct cfs_schedulable_data { ++ struct task_group *tg; ++ u64 period; ++ u64 quota; ++}; ++ ++enum { ++ cpuset = 0, ++ possible = 1, ++ fail = 2, ++}; ++ ++union cpumask_rcuhead { ++ cpumask_t cpumask; ++ struct callback_head rcu; ++}; ++ ++struct semaphore { ++ raw_spinlock_t lock; ++ unsigned int count; ++ struct list_head wait_list; ++}; ++ ++struct semaphore_waiter { ++ struct list_head list; ++ struct task_struct *task; ++ bool up; ++}; ++ ++typedef unsigned int pcp_op_T_____3; ++ ++struct hrtimer_sleeper { ++ struct hrtimer timer; ++ struct task_struct *task; ++}; ++ ++struct rt_waiter_node { ++ struct rb_node entry; ++ int prio; ++ u64 deadline; ++}; ++ ++struct rt_mutex_base; ++ ++struct ww_acquire_ctx; ++ ++struct rt_mutex_waiter { ++ struct rt_waiter_node tree; ++ struct rt_waiter_node pi_tree; ++ struct task_struct *task; ++ struct rt_mutex_base *lock; ++ unsigned int wake_state; ++ struct ww_acquire_ctx *ww_ctx; ++}; ++ ++struct rt_mutex_base { ++ raw_spinlock_t wait_lock; ++ struct rb_root_cached waiters; ++ struct task_struct *owner; ++}; ++ ++struct rt_mutex { ++ struct rt_mutex_base rtmutex; ++}; ++ ++struct ww_mutex { ++ struct mutex base; ++ struct ww_acquire_ctx *ctx; ++}; ++ ++struct ww_acquire_ctx { ++ struct task_struct *task; ++ long unsigned int stamp; ++ unsigned int acquired; ++ short unsigned int wounded; ++ short unsigned int is_wait_die; ++}; ++ ++struct rt_wake_q_head { ++ struct wake_q_head head; ++ struct task_struct *rtlock_task; ++}; ++ ++enum rtmutex_chainwalk { ++ RT_MUTEX_MIN_CHAINWALK = 0, ++ RT_MUTEX_FULL_CHAINWALK = 1, ++}; ++ ++struct sysrq_key_op { ++ void (* const handler)(int); ++ const char * const help_msg; ++ const char * const action_msg; ++ const int enable_mask; ++}; ++ ++typedef initcall_t initcall_entry_t; ++ ++typedef struct { ++ seqcount_t seqcount; ++} seqcount_latch_t; ++ ++struct dev_printk_info { ++ char subsystem[16]; ++ char device[48]; ++}; ++ ++enum cpuhp_state { ++ CPUHP_INVALID = -1, ++ CPUHP_OFFLINE = 0, ++ CPUHP_CREATE_THREADS = 1, ++ CPUHP_PERF_PREPARE = 2, ++ CPUHP_PERF_X86_PREPARE = 3, ++ CPUHP_PERF_X86_AMD_UNCORE_PREP = 4, ++ CPUHP_PERF_POWER = 5, ++ CPUHP_PERF_SUPERH = 6, ++ CPUHP_X86_HPET_DEAD = 7, ++ CPUHP_X86_APB_DEAD = 8, ++ CPUHP_X86_MCE_DEAD = 9, ++ CPUHP_VIRT_NET_DEAD = 10, ++ CPUHP_IBMVNIC_DEAD = 11, ++ CPUHP_SLUB_DEAD = 12, ++ CPUHP_DEBUG_OBJ_DEAD = 13, ++ CPUHP_MM_WRITEBACK_DEAD = 14, ++ CPUHP_MM_DEMOTION_DEAD = 15, ++ CPUHP_MM_VMSTAT_DEAD = 16, ++ CPUHP_SOFTIRQ_DEAD = 17, ++ CPUHP_NET_MVNETA_DEAD = 18, ++ CPUHP_CPUIDLE_DEAD = 19, ++ CPUHP_ARM64_FPSIMD_DEAD = 20, ++ CPUHP_ARM_OMAP_WAKE_DEAD = 21, ++ CPUHP_IRQ_POLL_DEAD = 22, ++ CPUHP_BLOCK_SOFTIRQ_DEAD = 23, ++ CPUHP_BIO_DEAD = 24, ++ CPUHP_ACPI_CPUDRV_DEAD = 25, ++ CPUHP_S390_PFAULT_DEAD = 26, ++ CPUHP_BLK_MQ_DEAD = 27, ++ CPUHP_FS_BUFF_DEAD = 28, ++ CPUHP_PRINTK_DEAD = 29, ++ CPUHP_MM_MEMCQ_DEAD = 30, ++ CPUHP_XFS_DEAD = 31, ++ CPUHP_PERCPU_CNT_DEAD = 32, ++ CPUHP_RADIX_DEAD = 33, ++ CPUHP_PAGE_ALLOC = 34, ++ CPUHP_NET_DEV_DEAD = 35, ++ CPUHP_PCI_XGENE_DEAD = 36, ++ CPUHP_IOMMU_IOVA_DEAD = 37, ++ CPUHP_LUSTRE_CFS_DEAD = 38, ++ CPUHP_AP_ARM_CACHE_B15_RAC_DEAD = 39, ++ CPUHP_PADATA_DEAD = 40, ++ CPUHP_AP_DTPM_CPU_DEAD = 41, ++ CPUHP_RANDOM_PREPARE = 42, ++ CPUHP_WORKQUEUE_PREP = 43, ++ CPUHP_POWER_NUMA_PREPARE = 44, ++ CPUHP_HRTIMERS_PREPARE = 45, ++ CPUHP_PROFILE_PREPARE = 46, ++ CPUHP_X2APIC_PREPARE = 47, ++ CPUHP_SMPCFD_PREPARE = 48, ++ CPUHP_RELAY_PREPARE = 49, ++ CPUHP_SLAB_PREPARE = 50, ++ CPUHP_MD_RAID5_PREPARE = 51, ++ CPUHP_RCUTREE_PREP = 52, ++ CPUHP_CPUIDLE_COUPLED_PREPARE = 53, ++ CPUHP_POWERPC_PMAC_PREPARE = 54, ++ CPUHP_POWERPC_MMU_CTX_PREPARE = 55, ++ CPUHP_XEN_PREPARE = 56, ++ CPUHP_XEN_EVTCHN_PREPARE = 57, ++ CPUHP_ARM_SHMOBILE_SCU_PREPARE = 58, ++ CPUHP_SH_SH3X_PREPARE = 59, ++ CPUHP_NET_FLOW_PREPARE = 60, ++ CPUHP_TOPOLOGY_PREPARE = 61, ++ CPUHP_NET_IUCV_PREPARE = 62, ++ CPUHP_ARM_BL_PREPARE = 63, ++ CPUHP_TRACE_RB_PREPARE = 64, ++ CPUHP_MM_ZS_PREPARE = 65, ++ CPUHP_MM_ZSWP_MEM_PREPARE = 66, ++ CPUHP_MM_ZSWP_POOL_PREPARE = 67, ++ CPUHP_KVM_PPC_BOOK3S_PREPARE = 68, ++ CPUHP_ZCOMP_PREPARE = 69, ++ CPUHP_TIMERS_PREPARE = 70, ++ CPUHP_MIPS_SOC_PREPARE = 71, ++ CPUHP_BP_PREPARE_DYN = 72, ++ CPUHP_BP_PREPARE_DYN_END = 92, ++ CPUHP_BRINGUP_CPU = 93, ++ CPUHP_AP_IDLE_DEAD = 94, ++ CPUHP_AP_OFFLINE = 95, ++ CPUHP_AP_CACHECTRL_STARTING = 96, ++ CPUHP_AP_SCHED_STARTING = 97, ++ CPUHP_AP_RCUTREE_DYING = 98, ++ CPUHP_AP_CPU_PM_STARTING = 99, ++ CPUHP_AP_IRQ_GIC_STARTING = 100, ++ CPUHP_AP_IRQ_HIP04_STARTING = 101, ++ CPUHP_AP_IRQ_APPLE_AIC_STARTING = 102, ++ CPUHP_AP_IRQ_ARMADA_XP_STARTING = 103, ++ CPUHP_AP_IRQ_BCM2836_STARTING = 104, ++ CPUHP_AP_IRQ_MIPS_GIC_STARTING = 105, ++ CPUHP_AP_IRQ_RISCV_STARTING = 106, ++ CPUHP_AP_IRQ_LOONGARCH_STARTING = 107, ++ CPUHP_AP_IRQ_SIFIVE_PLIC_STARTING = 108, ++ CPUHP_AP_ARM_MVEBU_COHERENCY = 109, ++ CPUHP_AP_MICROCODE_LOADER = 110, ++ CPUHP_AP_PERF_X86_AMD_UNCORE_STARTING = 111, ++ CPUHP_AP_PERF_X86_STARTING = 112, ++ CPUHP_AP_PERF_X86_AMD_IBS_STARTING = 113, ++ CPUHP_AP_PERF_X86_CQM_STARTING = 114, ++ CPUHP_AP_PERF_X86_CSTATE_STARTING = 115, ++ CPUHP_AP_PERF_XTENSA_STARTING = 116, ++ CPUHP_AP_MIPS_OP_LOONGSON3_STARTING = 117, ++ CPUHP_AP_ARM_VFP_STARTING = 118, ++ CPUHP_AP_ARM64_DEBUG_MONITORS_STARTING = 119, ++ CPUHP_AP_PERF_ARM_HW_BREAKPOINT_STARTING = 120, ++ CPUHP_AP_PERF_ARM_ACPI_STARTING = 121, ++ CPUHP_AP_PERF_ARM_STARTING = 122, ++ CPUHP_AP_PERF_RISCV_STARTING = 123, ++ CPUHP_AP_ARM_L2X0_STARTING = 124, ++ CPUHP_AP_EXYNOS4_MCT_TIMER_STARTING = 125, ++ CPUHP_AP_ARM_ARCH_TIMER_STARTING = 126, ++ CPUHP_AP_ARM_GLOBAL_TIMER_STARTING = 127, ++ CPUHP_AP_JCORE_TIMER_STARTING = 128, ++ CPUHP_AP_ARM_TWD_STARTING = 129, ++ CPUHP_AP_QCOM_TIMER_STARTING = 130, ++ CPUHP_AP_TEGRA_TIMER_STARTING = 131, ++ CPUHP_AP_ARMADA_TIMER_STARTING = 132, ++ CPUHP_AP_MARCO_TIMER_STARTING = 133, ++ CPUHP_AP_MIPS_GIC_TIMER_STARTING = 134, ++ CPUHP_AP_ARC_TIMER_STARTING = 135, ++ CPUHP_AP_RISCV_TIMER_STARTING = 136, ++ CPUHP_AP_CLINT_TIMER_STARTING = 137, ++ CPUHP_AP_CSKY_TIMER_STARTING = 138, ++ CPUHP_AP_TI_GP_TIMER_STARTING = 139, ++ CPUHP_AP_HYPERV_TIMER_STARTING = 140, ++ CPUHP_AP_DUMMY_TIMER_STARTING = 141, ++ CPUHP_AP_ARM_XEN_STARTING = 142, ++ CPUHP_AP_ARM_CORESIGHT_STARTING = 143, ++ CPUHP_AP_ARM_CORESIGHT_CTI_STARTING = 144, ++ CPUHP_AP_ARM64_ISNDEP_STARTING = 145, ++ CPUHP_AP_SMPCFD_DYING = 146, ++ CPUHP_AP_X86_TBOOT_DYING = 147, ++ CPUHP_AP_ARM_CACHE_B15_RAC_DYING = 148, ++ CPUHP_AP_ONLINE = 149, ++ CPUHP_TEARDOWN_CPU = 150, ++ CPUHP_AP_ONLINE_IDLE = 151, ++ CPUHP_AP_HYPERV_ONLINE = 152, ++ CPUHP_AP_KVM_ONLINE = 153, ++ CPUHP_AP_SCHED_WAIT_EMPTY = 154, ++ CPUHP_AP_SMPBOOT_THREADS = 155, ++ CPUHP_AP_X86_VDSO_VMA_ONLINE = 156, ++ CPUHP_AP_IRQ_AFFINITY_ONLINE = 157, ++ CPUHP_AP_BLK_MQ_ONLINE = 158, ++ CPUHP_AP_ARM_MVEBU_SYNC_CLOCKS = 159, ++ CPUHP_AP_X86_INTEL_EPB_ONLINE = 160, ++ CPUHP_AP_PERF_ONLINE = 161, ++ CPUHP_AP_PERF_X86_ONLINE = 162, ++ CPUHP_AP_PERF_X86_UNCORE_ONLINE = 163, ++ CPUHP_AP_PERF_X86_AMD_UNCORE_ONLINE = 164, ++ CPUHP_AP_PERF_X86_AMD_POWER_ONLINE = 165, ++ CPUHP_AP_PERF_X86_RAPL_ONLINE = 166, ++ CPUHP_AP_PERF_X86_CQM_ONLINE = 167, ++ CPUHP_AP_PERF_X86_CSTATE_ONLINE = 168, ++ CPUHP_AP_PERF_X86_IDXD_ONLINE = 169, ++ CPUHP_AP_PERF_S390_CF_ONLINE = 170, ++ CPUHP_AP_PERF_S390_SF_ONLINE = 171, ++ CPUHP_AP_PERF_ARM_CCI_ONLINE = 172, ++ CPUHP_AP_PERF_ARM_CCN_ONLINE = 173, ++ CPUHP_AP_PERF_ARM_HISI_CPA_ONLINE = 174, ++ CPUHP_AP_PERF_ARM_HISI_DDRC_ONLINE = 175, ++ CPUHP_AP_PERF_ARM_HISI_HHA_ONLINE = 176, ++ CPUHP_AP_PERF_ARM_HISI_L3_ONLINE = 177, ++ CPUHP_AP_PERF_ARM_HISI_PA_ONLINE = 178, ++ CPUHP_AP_PERF_ARM_HISI_SLLC_ONLINE = 179, ++ CPUHP_AP_PERF_ARM_HISI_PCIE_PMU_ONLINE = 180, ++ CPUHP_AP_PERF_ARM_HNS3_PMU_ONLINE = 181, ++ CPUHP_AP_PERF_ARM_L2X0_ONLINE = 182, ++ CPUHP_AP_PERF_ARM_QCOM_L2_ONLINE = 183, ++ CPUHP_AP_PERF_ARM_QCOM_L3_ONLINE = 184, ++ CPUHP_AP_PERF_ARM_APM_XGENE_ONLINE = 185, ++ CPUHP_AP_PERF_ARM_CAVIUM_TX2_UNCORE_ONLINE = 186, ++ CPUHP_AP_PERF_ARM_MARVELL_CN10K_DDR_ONLINE = 187, ++ CPUHP_AP_PERF_POWERPC_NEST_IMC_ONLINE = 188, ++ CPUHP_AP_PERF_POWERPC_CORE_IMC_ONLINE = 189, ++ CPUHP_AP_PERF_POWERPC_THREAD_IMC_ONLINE = 190, ++ CPUHP_AP_PERF_POWERPC_TRACE_IMC_ONLINE = 191, ++ CPUHP_AP_PERF_POWERPC_HV_24x7_ONLINE = 192, ++ CPUHP_AP_PERF_POWERPC_HV_GPCI_ONLINE = 193, ++ CPUHP_AP_PERF_CSKY_ONLINE = 194, ++ CPUHP_AP_WATCHDOG_ONLINE = 195, ++ CPUHP_AP_WORKQUEUE_ONLINE = 196, ++ CPUHP_AP_RANDOM_ONLINE = 197, ++ CPUHP_AP_RCUTREE_ONLINE = 198, ++ CPUHP_AP_BASE_CACHEINFO_ONLINE = 199, ++ CPUHP_AP_ONLINE_DYN = 200, ++ CPUHP_AP_ONLINE_DYN_END = 230, ++ CPUHP_AP_MM_DEMOTION_ONLINE = 231, ++ CPUHP_AP_X86_HPET_ONLINE = 232, ++ CPUHP_AP_X86_KVM_CLK_ONLINE = 233, ++ CPUHP_AP_ACTIVE = 234, ++ CPUHP_ONLINE = 235, ++}; ++ ++typedef struct poll_table_struct poll_table; ++ ++struct kmsg_dump_iter { ++ u64 cur_seq; ++ u64 next_seq; ++}; ++ ++struct kmsg_dumper { ++ struct list_head list; ++ void (*dump)(struct kmsg_dumper *, enum kmsg_dump_reason); ++ enum kmsg_dump_reason max_reason; ++ bool registered; ++}; ++ ++struct trace_event_raw_console { ++ struct trace_entry ent; ++ u32 __data_loc_msg; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_console { ++ u32 msg; ++}; ++ ++typedef void (*btf_trace_console)(void *, const char *, size_t); ++ ++struct printk_info { ++ u64 seq; ++ u64 ts_nsec; ++ u16 text_len; ++ u8 facility; ++ u8 flags: 5; ++ u8 level: 3; ++ u32 caller_id; ++ struct dev_printk_info dev_info; ++}; ++ ++struct printk_record { ++ struct printk_info *info; ++ char *text_buf; ++ unsigned int text_buf_size; ++}; ++ ++struct prb_data_blk_lpos { ++ long unsigned int begin; ++ long unsigned int next; ++}; ++ ++struct prb_desc { ++ atomic_long_t state_var; ++ struct prb_data_blk_lpos text_blk_lpos; ++}; ++ ++struct prb_data_ring { ++ unsigned int size_bits; ++ char *data; ++ atomic_long_t head_lpos; ++ atomic_long_t tail_lpos; ++}; ++ ++struct prb_desc_ring { ++ unsigned int count_bits; ++ struct prb_desc *descs; ++ struct printk_info *infos; ++ atomic_long_t head_id; ++ atomic_long_t tail_id; ++ atomic_long_t last_finalized_id; ++}; ++ ++struct printk_ringbuffer { ++ struct prb_desc_ring desc_ring; ++ struct prb_data_ring text_data_ring; ++ atomic_long_t fail; ++}; ++ ++struct prb_reserved_entry { ++ struct printk_ringbuffer *rb; ++ long unsigned int irqflags; ++ long unsigned int id; ++ unsigned int text_space; ++}; ++ ++enum desc_state { ++ desc_miss = -1, ++ desc_reserved = 0, ++ desc_committed = 1, ++ desc_finalized = 2, ++ desc_reusable = 3, ++}; ++ ++struct console_cmdline { ++ char name[16]; ++ int index; ++ bool user_specified; ++ char *options; ++}; ++ ++enum printk_info_flags { ++ LOG_NEWLINE = 2, ++ LOG_CONT = 8, ++}; ++ ++struct printk_buffers { ++ char outbuf[2048]; ++ char scratchbuf[1024]; ++}; ++ ++struct printk_message { ++ struct printk_buffers *pbufs; ++ unsigned int outbuf_len; ++ u64 seq; ++ long unsigned int dropped; ++}; ++ ++enum devkmsg_log_bits { ++ __DEVKMSG_LOG_BIT_ON = 0, ++ __DEVKMSG_LOG_BIT_OFF = 1, ++ __DEVKMSG_LOG_BIT_LOCK = 2, ++}; ++ ++enum devkmsg_log_masks { ++ DEVKMSG_LOG_MASK_ON = 1, ++ DEVKMSG_LOG_MASK_OFF = 2, ++ DEVKMSG_LOG_MASK_LOCK = 4, ++}; ++ ++enum con_msg_format_flags { ++ MSG_FORMAT_DEFAULT = 0, ++ MSG_FORMAT_SYSLOG = 1, ++}; ++ ++struct latched_seq { ++ seqcount_latch_t latch; ++ u64 val[2]; ++}; ++ ++struct devkmsg_user { ++ atomic64_t seq; ++ struct ratelimit_state rs; ++ struct mutex lock; ++ struct printk_buffers pbufs; ++}; ++ ++typedef int pcp_op_T_____4; ++ ++typedef long unsigned int irq_hw_number_t; ++ ++enum { ++ IRQ_TYPE_NONE = 0, ++ IRQ_TYPE_EDGE_RISING = 1, ++ IRQ_TYPE_EDGE_FALLING = 2, ++ IRQ_TYPE_EDGE_BOTH = 3, ++ IRQ_TYPE_LEVEL_HIGH = 4, ++ IRQ_TYPE_LEVEL_LOW = 8, ++ IRQ_TYPE_LEVEL_MASK = 12, ++ IRQ_TYPE_SENSE_MASK = 15, ++ IRQ_TYPE_DEFAULT = 15, ++ IRQ_TYPE_PROBE = 16, ++ IRQ_LEVEL = 256, ++ IRQ_PER_CPU = 512, ++ IRQ_NOPROBE = 1024, ++ IRQ_NOREQUEST = 2048, ++ IRQ_NOAUTOEN = 4096, ++ IRQ_NO_BALANCING = 8192, ++ IRQ_MOVE_PCNTXT = 16384, ++ IRQ_NESTED_THREAD = 32768, ++ IRQ_NOTHREAD = 65536, ++ IRQ_PER_CPU_DEVID = 131072, ++ IRQ_IS_POLLED = 262144, ++ IRQ_DISABLE_UNLAZY = 524288, ++ IRQ_HIDDEN = 1048576, ++ IRQ_NO_DEBUG = 2097152, ++}; ++ ++enum irq_domain_bus_token { ++ DOMAIN_BUS_ANY = 0, ++ DOMAIN_BUS_WIRED = 1, ++ DOMAIN_BUS_GENERIC_MSI = 2, ++ DOMAIN_BUS_PCI_MSI = 3, ++ DOMAIN_BUS_PLATFORM_MSI = 4, ++ DOMAIN_BUS_NEXUS = 5, ++ DOMAIN_BUS_IPI = 6, ++ DOMAIN_BUS_FSL_MC_MSI = 7, ++ DOMAIN_BUS_TI_SCI_INTA_MSI = 8, ++ DOMAIN_BUS_WAKEUP = 9, ++ DOMAIN_BUS_VMD_MSI = 10, ++ DOMAIN_BUS_PCI_DEVICE_MSI = 11, ++ DOMAIN_BUS_PCI_DEVICE_MSIX = 12, ++ DOMAIN_BUS_DMAR = 13, ++ DOMAIN_BUS_AMDVI = 14, ++ DOMAIN_BUS_PCI_DEVICE_IMS = 15, ++}; ++ ++struct irq_domain_ops; ++ ++struct irq_domain_chip_generic; ++ ++struct msi_parent_ops; ++ ++struct irq_domain { ++ struct list_head link; ++ const char *name; ++ const struct irq_domain_ops *ops; ++ void *host_data; ++ unsigned int flags; ++ unsigned int mapcount; ++ struct mutex mutex; ++ struct irq_domain *root; ++ struct fwnode_handle *fwnode; ++ enum irq_domain_bus_token bus_token; ++ struct irq_domain_chip_generic *gc; ++ struct device *dev; ++ struct device *pm_dev; ++ struct irq_domain *parent; ++ const struct msi_parent_ops *msi_parent_ops; ++ irq_hw_number_t hwirq_max; ++ unsigned int revmap_size; ++ struct xarray revmap_tree; ++ struct irq_data *revmap[0]; ++}; ++ ++enum { ++ IRQD_TRIGGER_MASK = 15, ++ IRQD_SETAFFINITY_PENDING = 256, ++ IRQD_ACTIVATED = 512, ++ IRQD_NO_BALANCING = 1024, ++ IRQD_PER_CPU = 2048, ++ IRQD_AFFINITY_SET = 4096, ++ IRQD_LEVEL = 8192, ++ IRQD_WAKEUP_STATE = 16384, ++ IRQD_MOVE_PCNTXT = 32768, ++ IRQD_IRQ_DISABLED = 65536, ++ IRQD_IRQ_MASKED = 131072, ++ IRQD_IRQ_INPROGRESS = 262144, ++ IRQD_WAKEUP_ARMED = 524288, ++ IRQD_FORWARDED_TO_VCPU = 1048576, ++ IRQD_AFFINITY_MANAGED = 2097152, ++ IRQD_IRQ_STARTED = 4194304, ++ IRQD_MANAGED_SHUTDOWN = 8388608, ++ IRQD_SINGLE_TARGET = 16777216, ++ IRQD_DEFAULT_TRIGGER_SET = 33554432, ++ IRQD_CAN_RESERVE = 67108864, ++ IRQD_MSI_NOMASK_QUIRK = 134217728, ++ IRQD_HANDLE_ENFORCE_IRQCTX = 268435456, ++ IRQD_AFFINITY_ON_ACTIVATE = 536870912, ++ IRQD_IRQ_ENABLED_ON_SUSPEND = 1073741824, ++}; ++ ++enum { ++ IRQCHIP_SET_TYPE_MASKED = 1, ++ IRQCHIP_EOI_IF_HANDLED = 2, ++ IRQCHIP_MASK_ON_SUSPEND = 4, ++ IRQCHIP_ONOFFLINE_ENABLED = 8, ++ IRQCHIP_SKIP_SET_WAKE = 16, ++ IRQCHIP_ONESHOT_SAFE = 32, ++ IRQCHIP_EOI_THREADED = 64, ++ IRQCHIP_SUPPORTS_LEVEL_MSI = 128, ++ IRQCHIP_SUPPORTS_NMI = 256, ++ IRQCHIP_ENABLE_WAKEUP_ON_SUSPEND = 512, ++ IRQCHIP_AFFINITY_PRE_STARTUP = 1024, ++ IRQCHIP_IMMUTABLE = 2048, ++}; ++ ++struct msi_alloc_info { ++ struct msi_desc *desc; ++ irq_hw_number_t hwirq; ++ long unsigned int flags; ++ union { ++ long unsigned int ul; ++ void *ptr; ++ } scratchpad[2]; ++}; ++ ++typedef struct msi_alloc_info msi_alloc_info_t; ++ ++struct msi_domain_info; ++ ++struct msi_domain_ops { ++ irq_hw_number_t (*get_hwirq)(struct msi_domain_info *, msi_alloc_info_t *); ++ int (*msi_init)(struct irq_domain *, struct msi_domain_info *, unsigned int, irq_hw_number_t, msi_alloc_info_t *); ++ void (*msi_free)(struct irq_domain *, struct msi_domain_info *, unsigned int); ++ int (*msi_prepare)(struct irq_domain *, struct device *, int, msi_alloc_info_t *); ++ void (*prepare_desc)(struct irq_domain *, msi_alloc_info_t *, struct msi_desc *); ++ void (*set_desc)(msi_alloc_info_t *, struct msi_desc *); ++ int (*domain_alloc_irqs)(struct irq_domain *, struct device *, int); ++ void (*domain_free_irqs)(struct irq_domain *, struct device *); ++ void (*msi_post_free)(struct irq_domain *, struct device *); ++}; ++ ++struct msi_domain_info { ++ u32 flags; ++ enum irq_domain_bus_token bus_token; ++ unsigned int hwsize; ++ struct msi_domain_ops *ops; ++ struct irq_chip *chip; ++ void *chip_data; ++ irq_flow_handler_t handler; ++ void *handler_data; ++ const char *handler_name; ++ void *data; ++}; ++ ++struct msi_parent_ops { ++ u32 supported_flags; ++ const char *prefix; ++ bool (*init_dev_msi_info)(struct device *, struct irq_domain *, struct irq_domain *, struct msi_domain_info *); ++}; ++ ++enum { ++ IRQC_IS_HARDIRQ = 0, ++ IRQC_IS_NESTED = 1, ++}; ++ ++struct irq_chip_regs { ++ long unsigned int enable; ++ long unsigned int disable; ++ long unsigned int mask; ++ long unsigned int ack; ++ long unsigned int eoi; ++ long unsigned int type; ++ long unsigned int polarity; ++}; ++ ++struct irq_chip_type { ++ struct irq_chip chip; ++ struct irq_chip_regs regs; ++ irq_flow_handler_t handler; ++ u32 type; ++ u32 mask_cache_priv; ++ u32 *mask_cache; ++}; ++ ++struct irq_chip_generic { ++ raw_spinlock_t lock; ++ void *reg_base; ++ u32 (*reg_readl)(void *); ++ void (*reg_writel)(u32, void *); ++ void (*suspend)(struct irq_chip_generic *); ++ void (*resume)(struct irq_chip_generic *); ++ unsigned int irq_base; ++ unsigned int irq_cnt; ++ u32 mask_cache; ++ u32 type_cache; ++ u32 polarity_cache; ++ u32 wake_enabled; ++ u32 wake_active; ++ unsigned int num_ct; ++ void *private; ++ long unsigned int installed; ++ long unsigned int unused; ++ struct irq_domain *domain; ++ struct list_head list; ++ struct irq_chip_type chip_types[0]; ++}; ++ ++enum irq_gc_flags { ++ IRQ_GC_INIT_MASK_CACHE = 1, ++ IRQ_GC_INIT_NESTED_LOCK = 2, ++ IRQ_GC_MASK_CACHE_PER_TYPE = 4, ++ IRQ_GC_NO_MASK = 8, ++ IRQ_GC_BE_IO = 16, ++}; ++ ++struct irq_domain_chip_generic { ++ unsigned int irqs_per_chip; ++ unsigned int num_chips; ++ unsigned int irq_flags_to_clear; ++ unsigned int irq_flags_to_set; ++ enum irq_gc_flags gc_flags; ++ struct irq_chip_generic *gc[0]; ++}; ++ ++struct irq_fwspec { ++ struct fwnode_handle *fwnode; ++ int param_count; ++ u32 param[16]; ++}; ++ ++struct irq_domain_ops { ++ int (*match)(struct irq_domain *, struct device_node *, enum irq_domain_bus_token); ++ int (*select)(struct irq_domain *, struct irq_fwspec *, enum irq_domain_bus_token); ++ int (*map)(struct irq_domain *, unsigned int, irq_hw_number_t); ++ void (*unmap)(struct irq_domain *, unsigned int); ++ int (*xlate)(struct irq_domain *, struct device_node *, const u32 *, unsigned int, long unsigned int *, unsigned int *); ++ int (*alloc)(struct irq_domain *, unsigned int, unsigned int, void *); ++ void (*free)(struct irq_domain *, unsigned int, unsigned int); ++ int (*activate)(struct irq_domain *, struct irq_data *, bool); ++ void (*deactivate)(struct irq_domain *, struct irq_data *); ++ int (*translate)(struct irq_domain *, struct irq_fwspec *, long unsigned int *, unsigned int *); ++}; ++ ++enum { ++ IRQTF_RUNTHREAD = 0, ++ IRQTF_WARNED = 1, ++ IRQTF_AFFINITY = 2, ++ IRQTF_FORCED_THREAD = 3, ++ IRQTF_READY = 4, ++}; ++ ++enum { ++ IRQS_AUTODETECT = 1, ++ IRQS_SPURIOUS_DISABLED = 2, ++ IRQS_POLL_INPROGRESS = 8, ++ IRQS_ONESHOT = 32, ++ IRQS_REPLAY = 64, ++ IRQS_WAITING = 128, ++ IRQS_PENDING = 512, ++ IRQS_SUSPENDED = 2048, ++ IRQS_TIMINGS = 4096, ++ IRQS_NMI = 8192, ++ IRQS_SYSFS = 16384, ++}; ++ ++enum { ++ _IRQ_DEFAULT_INIT_FLAGS = 0, ++ _IRQ_PER_CPU = 512, ++ _IRQ_LEVEL = 256, ++ _IRQ_NOPROBE = 1024, ++ _IRQ_NOREQUEST = 2048, ++ _IRQ_NOTHREAD = 65536, ++ _IRQ_NOAUTOEN = 4096, ++ _IRQ_MOVE_PCNTXT = 16384, ++ _IRQ_NO_BALANCING = 8192, ++ _IRQ_NESTED_THREAD = 32768, ++ _IRQ_PER_CPU_DEVID = 131072, ++ _IRQ_IS_POLLED = 262144, ++ _IRQ_DISABLE_UNLAZY = 524288, ++ _IRQ_HIDDEN = 1048576, ++ _IRQ_NO_DEBUG = 2097152, ++ _IRQF_MODIFY_MASK = 2096911, ++}; ++ ++struct proc_ops { ++ unsigned int proc_flags; ++ int (*proc_open)(struct inode *, struct file *); ++ ssize_t (*proc_read)(struct file *, char *, size_t, loff_t *); ++ ssize_t (*proc_read_iter)(struct kiocb *, struct iov_iter *); ++ ssize_t (*proc_write)(struct file *, const char *, size_t, loff_t *); ++ loff_t (*proc_lseek)(struct file *, loff_t, int); ++ int (*proc_release)(struct inode *, struct file *); ++ __poll_t (*proc_poll)(struct file *, struct poll_table_struct *); ++ long int (*proc_ioctl)(struct file *, unsigned int, long unsigned int); ++ long int (*proc_compat_ioctl)(struct file *, unsigned int, long unsigned int); ++ int (*proc_mmap)(struct file *, struct vm_area_struct *); ++ long unsigned int (*proc_get_unmapped_area)(struct file *, long unsigned int, long unsigned int, long unsigned int, long unsigned int); ++}; ++ ++enum { ++ AFFINITY = 0, ++ AFFINITY_LIST = 1, ++ EFFECTIVE = 2, ++ EFFECTIVE_LIST = 3, ++}; ++ ++enum { ++ GP_IDLE = 0, ++ GP_ENTER = 1, ++ GP_PASSED = 2, ++ GP_EXIT = 3, ++ GP_REPLAY = 4, ++}; ++ ++typedef long unsigned int ulong; ++ ++struct rcu_synchronize { ++ struct callback_head head; ++ struct completion completion; ++}; ++ ++struct rcu_cblist { ++ struct callback_head *head; ++ struct callback_head **tail; ++ long int len; ++}; ++ ++enum rcutorture_type { ++ RCU_FLAVOR = 0, ++ RCU_TASKS_FLAVOR = 1, ++ RCU_TASKS_RUDE_FLAVOR = 2, ++ RCU_TASKS_TRACING_FLAVOR = 3, ++ RCU_TRIVIAL_FLAVOR = 4, ++ SRCU_FLAVOR = 5, ++ INVALID_RCU_FLAVOR = 6, ++}; ++ ++typedef s64 pcp_op_T_____5; ++ ++enum { ++ TRACE_FTRACE_BIT = 0, ++ TRACE_FTRACE_NMI_BIT = 1, ++ TRACE_FTRACE_IRQ_BIT = 2, ++ TRACE_FTRACE_SIRQ_BIT = 3, ++ TRACE_FTRACE_TRANSITION_BIT = 4, ++ TRACE_INTERNAL_BIT = 5, ++ TRACE_INTERNAL_NMI_BIT = 6, ++ TRACE_INTERNAL_IRQ_BIT = 7, ++ TRACE_INTERNAL_SIRQ_BIT = 8, ++ TRACE_INTERNAL_TRANSITION_BIT = 9, ++ TRACE_BRANCH_BIT = 10, ++ TRACE_IRQ_BIT = 11, ++ TRACE_GRAPH_BIT = 12, ++ TRACE_GRAPH_DEPTH_START_BIT = 13, ++ TRACE_GRAPH_DEPTH_END_BIT = 14, ++ TRACE_GRAPH_NOTRACE_BIT = 15, ++ TRACE_RECORD_RECURSION_BIT = 16, ++}; ++ ++enum { ++ TRACE_CTX_NMI = 0, ++ TRACE_CTX_IRQ = 1, ++ TRACE_CTX_SOFTIRQ = 2, ++ TRACE_CTX_NORMAL = 3, ++ TRACE_CTX_TRANSITION = 4, ++}; ++ ++enum { ++ FTRACE_OPS_FL_ENABLED = 1, ++ FTRACE_OPS_FL_DYNAMIC = 2, ++ FTRACE_OPS_FL_SAVE_REGS = 4, ++ FTRACE_OPS_FL_SAVE_REGS_IF_SUPPORTED = 8, ++ FTRACE_OPS_FL_RECURSION = 16, ++ FTRACE_OPS_FL_STUB = 32, ++ FTRACE_OPS_FL_INITIALIZED = 64, ++ FTRACE_OPS_FL_DELETED = 128, ++ FTRACE_OPS_FL_ADDING = 256, ++ FTRACE_OPS_FL_REMOVING = 512, ++ FTRACE_OPS_FL_MODIFYING = 1024, ++ FTRACE_OPS_FL_ALLOC_TRAMP = 2048, ++ FTRACE_OPS_FL_IPMODIFY = 4096, ++ FTRACE_OPS_FL_PID = 8192, ++ FTRACE_OPS_FL_RCU = 16384, ++ FTRACE_OPS_FL_TRACE_ARRAY = 32768, ++ FTRACE_OPS_FL_PERMANENT = 65536, ++ FTRACE_OPS_FL_DIRECT = 131072, ++}; ++ ++struct klp_func { ++ const char *old_name; ++ void *new_func; ++ long unsigned int old_sympos; ++ void *old_func; ++ struct kobject kobj; ++ struct list_head node; ++ struct list_head stack_node; ++ long unsigned int old_size; ++ long unsigned int new_size; ++ bool nop; ++ bool patched; ++ bool transition; ++}; ++ ++struct klp_object; ++ ++struct klp_callbacks { ++ int (*pre_patch)(struct klp_object *); ++ void (*post_patch)(struct klp_object *); ++ void (*pre_unpatch)(struct klp_object *); ++ void (*post_unpatch)(struct klp_object *); ++ bool post_unpatch_enabled; ++}; ++ ++struct klp_object { ++ const char *name; ++ struct klp_func *funcs; ++ struct klp_callbacks callbacks; ++ struct kobject kobj; ++ struct list_head func_list; ++ struct list_head node; ++ struct module *mod; ++ bool dynamic; ++ bool patched; ++}; ++ ++struct klp_state { ++ long unsigned int id; ++ unsigned int version; ++ void *data; ++}; ++ ++struct klp_patch { ++ struct module *mod; ++ struct klp_object *objs; ++ struct klp_state *states; ++ bool replace; ++ struct list_head list; ++ struct kobject kobj; ++ struct list_head obj_list; ++ bool enabled; ++ bool forced; ++ struct work_struct free_work; ++ struct completion finish; ++}; ++ ++struct klp_ops { ++ struct list_head node; ++ struct list_head func_stack; ++ struct ftrace_ops fops; ++}; ++ ++typedef int (*dr_match_t)(struct device *, void *, void *); ++ ++struct dma_sgt_handle { ++ struct sg_table sgt; ++ struct page **pages; ++}; ++ ++struct dma_devres { ++ size_t size; ++ void *vaddr; ++ dma_addr_t dma_handle; ++ long unsigned int attrs; ++}; ++ ++struct seccomp_data { ++ int nr; ++ __u32 arch; ++ __u64 instruction_pointer; ++ __u64 args[6]; ++}; ++ ++struct trace_event_raw_sys_enter { ++ struct trace_entry ent; ++ long int id; ++ long unsigned int args[6]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sys_exit { ++ struct trace_entry ent; ++ long int id; ++ long int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_sys_enter {}; ++ ++struct trace_event_data_offsets_sys_exit {}; ++ ++typedef void (*btf_trace_sys_enter)(void *, struct pt_regs *, long int); ++ ++typedef void (*btf_trace_sys_exit)(void *, struct pt_regs *, long int); ++ ++struct load_info { ++ const char *name; ++ struct module *mod; ++ Elf64_Ehdr *hdr; ++ long unsigned int len; ++ Elf64_Shdr *sechdrs; ++ char *secstrings; ++ char *strtab; ++ long unsigned int symoffs; ++ long unsigned int stroffs; ++ long unsigned int init_typeoffs; ++ long unsigned int core_typeoffs; ++ bool sig_ok; ++ long unsigned int mod_kallsyms_init_off; ++ struct { ++ unsigned int sym; ++ unsigned int str; ++ unsigned int mod; ++ unsigned int vers; ++ unsigned int info; ++ unsigned int pcpu; ++ } index; ++}; ++ ++struct module_use { ++ struct list_head source_list; ++ struct list_head target_list; ++ struct module *source; ++ struct module *target; ++}; ++ ++struct module_sect_attr { ++ struct bin_attribute battr; ++ long unsigned int address; ++}; ++ ++struct module_sect_attrs { ++ struct attribute_group grp; ++ unsigned int nsections; ++ struct module_sect_attr attrs[0]; ++}; ++ ++struct module_notes_attrs { ++ struct kobject *dir; ++ unsigned int notes; ++ struct bin_attribute attrs[0]; ++}; ++ ++struct profile_hit { ++ u32 pc; ++ u32 hits; ++}; ++ ++struct itimerspec64 { ++ struct timespec64 it_interval; ++ struct timespec64 it_value; ++}; ++ ++struct softirq_action { ++ void (*action)(struct softirq_action *); ++}; ++ ++enum tick_dep_bits { ++ TICK_DEP_BIT_POSIX_TIMER = 0, ++ TICK_DEP_BIT_PERF_EVENTS = 1, ++ TICK_DEP_BIT_SCHED = 2, ++ TICK_DEP_BIT_CLOCK_UNSTABLE = 3, ++ TICK_DEP_BIT_RCU = 4, ++ TICK_DEP_BIT_RCU_EXP = 5, ++}; ++ ++struct trace_event_raw_timer_class { ++ struct trace_entry ent; ++ void *timer; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_timer_start { ++ struct trace_entry ent; ++ void *timer; ++ void *function; ++ long unsigned int expires; ++ long unsigned int now; ++ unsigned int flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_timer_expire_entry { ++ struct trace_entry ent; ++ void *timer; ++ long unsigned int now; ++ void *function; ++ long unsigned int baseclk; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_hrtimer_init { ++ struct trace_entry ent; ++ void *hrtimer; ++ clockid_t clockid; ++ enum hrtimer_mode mode; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_hrtimer_start { ++ struct trace_entry ent; ++ void *hrtimer; ++ void *function; ++ s64 expires; ++ s64 softexpires; ++ enum hrtimer_mode mode; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_hrtimer_expire_entry { ++ struct trace_entry ent; ++ void *hrtimer; ++ s64 now; ++ void *function; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_hrtimer_class { ++ struct trace_entry ent; ++ void *hrtimer; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_itimer_state { ++ struct trace_entry ent; ++ int which; ++ long long unsigned int expires; ++ long int value_sec; ++ long int value_nsec; ++ long int interval_sec; ++ long int interval_nsec; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_itimer_expire { ++ struct trace_entry ent; ++ int which; ++ pid_t pid; ++ long long unsigned int now; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_tick_stop { ++ struct trace_entry ent; ++ int success; ++ int dependency; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_timer_class {}; ++ ++struct trace_event_data_offsets_timer_start {}; ++ ++struct trace_event_data_offsets_timer_expire_entry {}; ++ ++struct trace_event_data_offsets_hrtimer_init {}; ++ ++struct trace_event_data_offsets_hrtimer_start {}; ++ ++struct trace_event_data_offsets_hrtimer_expire_entry {}; ++ ++struct trace_event_data_offsets_hrtimer_class {}; ++ ++struct trace_event_data_offsets_itimer_state {}; ++ ++struct trace_event_data_offsets_itimer_expire {}; ++ ++struct trace_event_data_offsets_tick_stop {}; ++ ++typedef void (*btf_trace_timer_init)(void *, struct timer_list *); ++ ++typedef void (*btf_trace_timer_start)(void *, struct timer_list *, long unsigned int, unsigned int); ++ ++typedef void (*btf_trace_timer_expire_entry)(void *, struct timer_list *, long unsigned int); ++ ++typedef void (*btf_trace_timer_expire_exit)(void *, struct timer_list *); ++ ++typedef void (*btf_trace_timer_cancel)(void *, struct timer_list *); ++ ++typedef void (*btf_trace_hrtimer_init)(void *, struct hrtimer *, clockid_t, enum hrtimer_mode); ++ ++typedef void (*btf_trace_hrtimer_start)(void *, struct hrtimer *, enum hrtimer_mode); ++ ++typedef void (*btf_trace_hrtimer_expire_entry)(void *, struct hrtimer *, ktime_t *); ++ ++typedef void (*btf_trace_hrtimer_expire_exit)(void *, struct hrtimer *); ++ ++typedef void (*btf_trace_hrtimer_cancel)(void *, struct hrtimer *); ++ ++typedef void (*btf_trace_itimer_state)(void *, int, const struct itimerspec64 * const, long long unsigned int); ++ ++typedef void (*btf_trace_itimer_expire)(void *, int, struct pid *, long long unsigned int); ++ ++typedef void (*btf_trace_tick_stop)(void *, int, int); ++ ++struct timer_base { ++ raw_spinlock_t lock; ++ struct timer_list *running_timer; ++ long unsigned int clk; ++ long unsigned int next_expiry; ++ unsigned int cpu; ++ bool next_expiry_recalc; ++ bool is_idle; ++ bool timers_pending; ++ long unsigned int pending_map[8]; ++ struct hlist_head vectors[512]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct process_timer { ++ struct timer_list timer; ++ struct task_struct *task; ++}; ++ ++typedef __kernel_timer_t timer_t; ++ ++struct __kernel_itimerspec { ++ struct __kernel_timespec it_interval; ++ struct __kernel_timespec it_value; ++}; ++ ++struct timezone { ++ int tz_minuteswest; ++ int tz_dsttime; ++}; ++ ++struct __kernel_timex_timeval { ++ __kernel_time64_t tv_sec; ++ long long int tv_usec; ++}; ++ ++struct __kernel_timex { ++ unsigned int modes; ++ long long int offset; ++ long long int freq; ++ long long int maxerror; ++ long long int esterror; ++ int status; ++ long long int constant; ++ long long int precision; ++ long long int tolerance; ++ struct __kernel_timex_timeval time; ++ long long int tick; ++ long long int ppsfreq; ++ long long int jitter; ++ int shift; ++ long long int stabil; ++ long long int jitcnt; ++ long long int calcnt; ++ long long int errcnt; ++ long long int stbcnt; ++ int tai; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct old_itimerspec32 { ++ struct old_timespec32 it_interval; ++ struct old_timespec32 it_value; ++}; ++ ++struct old_timex32 { ++ u32 modes; ++ s32 offset; ++ s32 freq; ++ s32 maxerror; ++ s32 esterror; ++ s32 status; ++ s32 constant; ++ s32 precision; ++ s32 tolerance; ++ struct old_timeval32 time; ++ s32 tick; ++ s32 ppsfreq; ++ s32 jitter; ++ s32 shift; ++ s32 stabil; ++ s32 jitcnt; ++ s32 calcnt; ++ s32 errcnt; ++ s32 stbcnt; ++ s32 tai; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum tk_offsets { ++ TK_OFFS_REAL = 0, ++ TK_OFFS_BOOT = 1, ++ TK_OFFS_TAI = 2, ++ TK_OFFS_MAX = 3, ++}; ++ ++struct sigevent { ++ sigval_t sigev_value; ++ int sigev_signo; ++ int sigev_notify; ++ union { ++ int _pad[12]; ++ int _tid; ++ struct { ++ void (*_function)(sigval_t); ++ void *_attribute; ++ } _sigev_thread; ++ } _sigev_un; ++}; ++ ++typedef struct sigevent sigevent_t; ++ ++enum alarmtimer_type { ++ ALARM_REALTIME = 0, ++ ALARM_BOOTTIME = 1, ++ ALARM_NUMTYPE = 2, ++ ALARM_REALTIME_FREEZER = 3, ++ ALARM_BOOTTIME_FREEZER = 4, ++}; ++ ++enum alarmtimer_restart { ++ ALARMTIMER_NORESTART = 0, ++ ALARMTIMER_RESTART = 1, ++}; ++ ++struct alarm { ++ struct timerqueue_node node; ++ struct hrtimer timer; ++ enum alarmtimer_restart (*function)(struct alarm *, ktime_t); ++ enum alarmtimer_type type; ++ int state; ++ void *data; ++}; ++ ++struct cpu_timer { ++ struct timerqueue_node node; ++ struct timerqueue_head *head; ++ struct pid *pid; ++ struct list_head elist; ++ int firing; ++ struct task_struct *handling; ++}; ++ ++struct k_clock; ++ ++struct k_itimer { ++ struct list_head list; ++ struct hlist_node t_hash; ++ spinlock_t it_lock; ++ const struct k_clock *kclock; ++ clockid_t it_clock; ++ timer_t it_id; ++ int it_active; ++ s64 it_overrun; ++ s64 it_overrun_last; ++ int it_requeue_pending; ++ int it_sigev_notify; ++ ktime_t it_interval; ++ struct signal_struct *it_signal; ++ union { ++ struct pid *it_pid; ++ struct task_struct *it_process; ++ }; ++ struct sigqueue *sigq; ++ union { ++ struct { ++ struct hrtimer timer; ++ } real; ++ struct cpu_timer cpu; ++ struct { ++ struct alarm alarmtimer; ++ } alarm; ++ } it; ++ struct callback_head rcu; ++}; ++ ++struct k_clock { ++ int (*clock_getres)(const clockid_t, struct timespec64 *); ++ int (*clock_set)(const clockid_t, const struct timespec64 *); ++ int (*clock_get_timespec)(const clockid_t, struct timespec64 *); ++ ktime_t (*clock_get_ktime)(const clockid_t); ++ int (*clock_adj)(const clockid_t, struct __kernel_timex *); ++ int (*timer_create)(struct k_itimer *); ++ int (*nsleep)(const clockid_t, int, const struct timespec64 *); ++ int (*timer_set)(struct k_itimer *, int, struct itimerspec64 *, struct itimerspec64 *); ++ int (*timer_del)(struct k_itimer *); ++ void (*timer_get)(struct k_itimer *, struct itimerspec64 *); ++ void (*timer_rearm)(struct k_itimer *); ++ s64 (*timer_forward)(struct k_itimer *, ktime_t); ++ ktime_t (*timer_remaining)(struct k_itimer *, ktime_t); ++ int (*timer_try_to_cancel)(struct k_itimer *); ++ void (*timer_arm)(struct k_itimer *, ktime_t, bool, bool); ++ void (*timer_wait_running)(struct k_itimer *); ++}; ++ ++struct compat_sigevent { ++ compat_sigval_t sigev_value; ++ compat_int_t sigev_signo; ++ compat_int_t sigev_notify; ++ union { ++ compat_int_t _pad[13]; ++ compat_int_t _tid; ++ struct { ++ compat_uptr_t _function; ++ compat_uptr_t _attribute; ++ } _sigev_thread; ++ } _sigev_un; ++}; ++ ++enum tick_device_mode { ++ TICKDEV_MODE_PERIODIC = 0, ++ TICKDEV_MODE_ONESHOT = 1, ++}; ++ ++struct clock_event_device; ++ ++struct tick_device { ++ struct clock_event_device *evtdev; ++ enum tick_device_mode mode; ++}; ++ ++enum clock_event_state { ++ CLOCK_EVT_STATE_DETACHED = 0, ++ CLOCK_EVT_STATE_SHUTDOWN = 1, ++ CLOCK_EVT_STATE_PERIODIC = 2, ++ CLOCK_EVT_STATE_ONESHOT = 3, ++ CLOCK_EVT_STATE_ONESHOT_STOPPED = 4, ++}; ++ ++struct clock_event_device { ++ void (*event_handler)(struct clock_event_device *); ++ int (*set_next_event)(long unsigned int, struct clock_event_device *); ++ int (*set_next_ktime)(ktime_t, struct clock_event_device *); ++ ktime_t next_event; ++ u64 max_delta_ns; ++ u64 min_delta_ns; ++ u32 mult; ++ u32 shift; ++ enum clock_event_state state_use_accessors; ++ unsigned int features; ++ long unsigned int retries; ++ int (*set_state_periodic)(struct clock_event_device *); ++ int (*set_state_oneshot)(struct clock_event_device *); ++ int (*set_state_oneshot_stopped)(struct clock_event_device *); ++ int (*set_state_shutdown)(struct clock_event_device *); ++ int (*tick_resume)(struct clock_event_device *); ++ void (*broadcast)(const struct cpumask *); ++ void (*suspend)(struct clock_event_device *); ++ void (*resume)(struct clock_event_device *); ++ long unsigned int min_delta_ticks; ++ long unsigned int max_delta_ticks; ++ const char *name; ++ int rating; ++ int irq; ++ int bound_on; ++ const struct cpumask *cpumask; ++ struct list_head list; ++ struct module *owner; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum tick_broadcast_state { ++ TICK_BROADCAST_EXIT = 0, ++ TICK_BROADCAST_ENTER = 1, ++}; ++ ++struct nsset { ++ unsigned int flags; ++ struct nsproxy *nsproxy; ++ struct fs_struct *fs; ++ const struct cred *cred; ++}; ++ ++struct proc_timens_offset { ++ int clockid; ++ struct timespec64 val; ++}; ++ ++enum vdso_clock_mode { ++ VDSO_CLOCKMODE_NONE = 0, ++ VDSO_CLOCKMODE_TOD = 1, ++ VDSO_CLOCKMODE_MAX = 2, ++ VDSO_CLOCKMODE_TIMENS = 2147483647, ++}; ++ ++enum { ++ PROC_ROOT_INO = 1, ++ PROC_IPC_INIT_INO = 4026531839, ++ PROC_UTS_INIT_INO = 4026531838, ++ PROC_USER_INIT_INO = 4026531837, ++ PROC_PID_INIT_INO = 4026531836, ++ PROC_CGROUP_INIT_INO = 4026531835, ++ PROC_TIME_INIT_INO = 4026531834, ++}; ++ ++union futex_key { ++ struct { ++ u64 i_seq; ++ long unsigned int pgoff; ++ unsigned int offset; ++ } shared; ++ struct { ++ union { ++ struct mm_struct *mm; ++ u64 __tmp; ++ }; ++ long unsigned int address; ++ unsigned int offset; ++ } private; ++ struct { ++ u64 ptr; ++ long unsigned int word; ++ unsigned int offset; ++ } both; ++}; ++ ++struct futex_pi_state { ++ struct list_head list; ++ struct rt_mutex_base pi_mutex; ++ struct task_struct *owner; ++ refcount_t refcount; ++ union futex_key key; ++}; ++ ++enum { ++ FUTEX_STATE_OK = 0, ++ FUTEX_STATE_EXITING = 1, ++ FUTEX_STATE_DEAD = 2, ++}; ++ ++struct futex_hash_bucket { ++ atomic_t waiters; ++ spinlock_t lock; ++ struct plist_head chain; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct futex_q { ++ struct plist_node list; ++ struct task_struct *task; ++ spinlock_t *lock_ptr; ++ union futex_key key; ++ struct futex_pi_state *pi_state; ++ struct rt_mutex_waiter *rt_waiter; ++ union futex_key *requeue_pi_key; ++ u32 bitset; ++ atomic_t requeue_state; ++}; ++ ++enum futex_access { ++ FUTEX_READ = 0, ++ FUTEX_WRITE = 1, ++}; ++ ++struct call_function_data { ++ call_single_data_t *csd; ++ cpumask_var_t cpumask; ++ cpumask_var_t cpumask_ipi; ++}; ++ ++struct smp_call_on_cpu_struct { ++ struct work_struct work; ++ struct completion done; ++ int (*func)(void *); ++ void *data; ++ int ret; ++ int cpu; ++}; ++ ++typedef __u16 comp_t; ++ ++struct acct_v3 { ++ char ac_flag; ++ char ac_version; ++ __u16 ac_tty; ++ __u32 ac_exitcode; ++ __u32 ac_uid; ++ __u32 ac_gid; ++ __u32 ac_pid; ++ __u32 ac_ppid; ++ __u32 ac_btime; ++ __u32 ac_etime; ++ comp_t ac_utime; ++ comp_t ac_stime; ++ comp_t ac_mem; ++ comp_t ac_io; ++ comp_t ac_rw; ++ comp_t ac_minflt; ++ comp_t ac_majflt; ++ comp_t ac_swaps; ++ char ac_comm[16]; ++}; ++ ++typedef struct acct_v3 acct_t; ++ ++struct fs_pin { ++ wait_queue_head_t wait; ++ int done; ++ struct hlist_node s_list; ++ struct hlist_node m_list; ++ void (*kill)(struct fs_pin *); ++}; ++ ++struct bsd_acct_struct { ++ struct fs_pin pin; ++ atomic_long_t count; ++ struct callback_head rcu; ++ struct mutex lock; ++ int active; ++ long unsigned int needcheck; ++ struct file *file; ++ struct pid_namespace *ns; ++ struct work_struct work; ++ struct completion done; ++}; ++ ++enum lockdown_reason { ++ LOCKDOWN_NONE = 0, ++ LOCKDOWN_MODULE_SIGNATURE = 1, ++ LOCKDOWN_DEV_MEM = 2, ++ LOCKDOWN_EFI_TEST = 3, ++ LOCKDOWN_KEXEC = 4, ++ LOCKDOWN_HIBERNATION = 5, ++ LOCKDOWN_PCI_ACCESS = 6, ++ LOCKDOWN_IOPORT = 7, ++ LOCKDOWN_MSR = 8, ++ LOCKDOWN_ACPI_TABLES = 9, ++ LOCKDOWN_DEVICE_TREE = 10, ++ LOCKDOWN_PCMCIA_CIS = 11, ++ LOCKDOWN_TIOCSSERIAL = 12, ++ LOCKDOWN_MODULE_PARAMETERS = 13, ++ LOCKDOWN_MMIOTRACE = 14, ++ LOCKDOWN_DEBUGFS = 15, ++ LOCKDOWN_XMON_WR = 16, ++ LOCKDOWN_BPF_WRITE_USER = 17, ++ LOCKDOWN_DBG_WRITE_KERNEL = 18, ++ LOCKDOWN_RTAS_ERROR_INJECTION = 19, ++ LOCKDOWN_INTEGRITY_MAX = 20, ++ LOCKDOWN_KCORE = 21, ++ LOCKDOWN_KPROBES = 22, ++ LOCKDOWN_BPF_READ_KERNEL = 23, ++ LOCKDOWN_DBG_READ_KERNEL = 24, ++ LOCKDOWN_PERF = 25, ++ LOCKDOWN_TRACEFS = 26, ++ LOCKDOWN_XMON_RW = 27, ++ LOCKDOWN_XFRM_SECRET = 28, ++ LOCKDOWN_CONFIDENTIALITY_MAX = 29, ++}; ++ ++enum kernel_load_data_id { ++ LOADING_UNKNOWN = 0, ++ LOADING_FIRMWARE = 1, ++ LOADING_MODULE = 2, ++ LOADING_KEXEC_IMAGE = 3, ++ LOADING_KEXEC_INITRAMFS = 4, ++ LOADING_POLICY = 5, ++ LOADING_X509_CERTIFICATE = 6, ++ LOADING_MAX_ID = 7, ++}; ++ ++struct compat_kexec_segment { ++ compat_uptr_t buf; ++ compat_size_t bufsz; ++ compat_ulong_t mem; ++ compat_size_t memsz; ++}; ++ ++struct cgroup_taskset { ++ struct list_head src_csets; ++ struct list_head dst_csets; ++ int nr_tasks; ++ int ssid; ++ struct list_head *csets; ++ struct css_set *cur_cset; ++ struct task_struct *cur_task; ++}; ++ ++enum { ++ CGRP_NOTIFY_ON_RELEASE = 0, ++ CGRP_CPUSET_CLONE_CHILDREN = 1, ++ CGRP_FREEZE = 2, ++ CGRP_FROZEN = 3, ++ CGRP_KILL = 4, ++}; ++ ++struct css_task_iter { ++ struct cgroup_subsys *ss; ++ unsigned int flags; ++ struct list_head *cset_pos; ++ struct list_head *cset_head; ++ struct list_head *tcset_pos; ++ struct list_head *tcset_head; ++ struct list_head *task_pos; ++ struct list_head *cur_tasks_head; ++ struct css_set *cur_cset; ++ struct css_set *cur_dcset; ++ struct task_struct *cur_task; ++ struct list_head iters_node; ++}; ++ ++typedef struct { ++ char *from; ++ char *to; ++} substring_t; ++ ++enum rdmacg_resource_type { ++ RDMACG_RESOURCE_HCA_HANDLE = 0, ++ RDMACG_RESOURCE_HCA_OBJECT = 1, ++ RDMACG_RESOURCE_MAX = 2, ++}; ++ ++struct rdma_cgroup { ++ struct cgroup_subsys_state css; ++ struct list_head rpools; ++}; ++ ++struct rdmacg_device { ++ struct list_head dev_node; ++ struct list_head rpools; ++ char *name; ++}; ++ ++enum rdmacg_file_type { ++ RDMACG_RESOURCE_TYPE_MAX = 0, ++ RDMACG_RESOURCE_TYPE_STAT = 1, ++}; ++ ++struct rdmacg_resource { ++ int max; ++ int usage; ++}; ++ ++struct rdmacg_resource_pool { ++ struct rdmacg_device *device; ++ struct rdmacg_resource resources[2]; ++ struct list_head cg_node; ++ struct list_head dev_node; ++ u64 usage_sum; ++ int num_max_cnt; ++}; ++ ++typedef int __kernel_key_t; ++ ++typedef int __kernel_mqd_t; ++ ++typedef __kernel_key_t key_t; ++ ++typedef __kernel_mqd_t mqd_t; ++ ++struct kern_ipc_perm { ++ spinlock_t lock; ++ bool deleted; ++ int id; ++ key_t key; ++ kuid_t uid; ++ kgid_t gid; ++ kuid_t cuid; ++ kgid_t cgid; ++ umode_t mode; ++ long unsigned int seq; ++ void *security; ++ struct rhash_head khtnode; ++ struct callback_head rcu; ++ refcount_t refcount; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum audit_state { ++ AUDIT_STATE_DISABLED = 0, ++ AUDIT_STATE_BUILD = 1, ++ AUDIT_STATE_RECORD = 2, ++}; ++ ++struct audit_cap_data { ++ kernel_cap_t permitted; ++ kernel_cap_t inheritable; ++ union { ++ unsigned int fE; ++ kernel_cap_t effective; ++ }; ++ kernel_cap_t ambient; ++ kuid_t rootid; ++}; ++ ++struct audit_names { ++ struct list_head list; ++ struct filename *name; ++ int name_len; ++ bool hidden; ++ long unsigned int ino; ++ dev_t dev; ++ umode_t mode; ++ kuid_t uid; ++ kgid_t gid; ++ dev_t rdev; ++ u32 osid; ++ struct audit_cap_data fcap; ++ unsigned int fcap_ver; ++ unsigned char type; ++ bool should_free; ++}; ++ ++struct mq_attr { ++ __kernel_long_t mq_flags; ++ __kernel_long_t mq_maxmsg; ++ __kernel_long_t mq_msgsize; ++ __kernel_long_t mq_curmsgs; ++ __kernel_long_t __reserved[4]; ++}; ++ ++struct open_how { ++ __u64 flags; ++ __u64 mode; ++ __u64 resolve; ++}; ++ ++struct audit_ntp_val { ++ long long int oldval; ++ long long int newval; ++}; ++ ++struct audit_ntp_data { ++ struct audit_ntp_val vals[6]; ++}; ++ ++struct audit_proctitle { ++ int len; ++ char *value; ++}; ++ ++struct audit_aux_data; ++ ++struct __kernel_sockaddr_storage; ++ ++struct audit_tree_refs; ++ ++struct audit_context { ++ int dummy; ++ enum { ++ AUDIT_CTX_UNUSED = 0, ++ AUDIT_CTX_SYSCALL = 1, ++ AUDIT_CTX_URING = 2, ++ } context; ++ enum audit_state state; ++ enum audit_state current_state; ++ unsigned int serial; ++ int major; ++ int uring_op; ++ struct timespec64 ctime; ++ long unsigned int argv[4]; ++ long int return_code; ++ u64 prio; ++ int return_valid; ++ struct audit_names preallocated_names[5]; ++ int name_count; ++ struct list_head names_list; ++ char *filterkey; ++ struct path pwd; ++ struct audit_aux_data *aux; ++ struct audit_aux_data *aux_pids; ++ struct __kernel_sockaddr_storage *sockaddr; ++ size_t sockaddr_len; ++ pid_t ppid; ++ kuid_t uid; ++ kuid_t euid; ++ kuid_t suid; ++ kuid_t fsuid; ++ kgid_t gid; ++ kgid_t egid; ++ kgid_t sgid; ++ kgid_t fsgid; ++ long unsigned int personality; ++ int arch; ++ pid_t target_pid; ++ kuid_t target_auid; ++ kuid_t target_uid; ++ unsigned int target_sessionid; ++ u32 target_sid; ++ char target_comm[16]; ++ struct audit_tree_refs *trees; ++ struct audit_tree_refs *first_trees; ++ struct list_head killed_trees; ++ int tree_count; ++ int type; ++ union { ++ struct { ++ int nargs; ++ long int args[6]; ++ } socketcall; ++ struct { ++ kuid_t uid; ++ kgid_t gid; ++ umode_t mode; ++ u32 osid; ++ int has_perm; ++ uid_t perm_uid; ++ gid_t perm_gid; ++ umode_t perm_mode; ++ long unsigned int qbytes; ++ } ipc; ++ struct { ++ mqd_t mqdes; ++ struct mq_attr mqstat; ++ } mq_getsetattr; ++ struct { ++ mqd_t mqdes; ++ int sigev_signo; ++ } mq_notify; ++ struct { ++ mqd_t mqdes; ++ size_t msg_len; ++ unsigned int msg_prio; ++ struct timespec64 abs_timeout; ++ } mq_sendrecv; ++ struct { ++ int oflag; ++ umode_t mode; ++ struct mq_attr attr; ++ } mq_open; ++ struct { ++ pid_t pid; ++ struct audit_cap_data cap; ++ } capset; ++ struct { ++ int fd; ++ int flags; ++ } mmap; ++ struct open_how openat2; ++ struct { ++ int argc; ++ } execve; ++ struct { ++ char *name; ++ } module; ++ struct { ++ struct audit_ntp_data ntp_data; ++ struct timespec64 tk_injoffset; ++ } time; ++ }; ++ int fds[2]; ++ struct audit_proctitle proctitle; ++}; ++ ++struct cpu_vfs_cap_data { ++ __u32 magic_etc; ++ kuid_t rootid; ++ kernel_cap_t permitted; ++ kernel_cap_t inheritable; ++}; ++ ++typedef struct fsnotify_mark_connector *fsnotify_connp_t; ++ ++struct fsnotify_mark_connector { ++ spinlock_t lock; ++ short unsigned int type; ++ short unsigned int flags; ++ __kernel_fsid_t fsid; ++ union { ++ fsnotify_connp_t *obj; ++ struct fsnotify_mark_connector *destroy_next; ++ }; ++ struct hlist_head list; ++}; ++ ++struct __kernel_sockaddr_storage { ++ union { ++ struct { ++ __kernel_sa_family_t ss_family; ++ char __data[126]; ++ }; ++ void *__align; ++ }; ++}; ++ ++enum { ++ Audit_equal = 0, ++ Audit_not_equal = 1, ++ Audit_bitmask = 2, ++ Audit_bittest = 3, ++ Audit_lt = 4, ++ Audit_gt = 5, ++ Audit_le = 6, ++ Audit_ge = 7, ++ Audit_bad = 8, ++}; ++ ++struct fanotify_response_info_header { ++ __u8 type; ++ __u8 pad; ++ __u16 len; ++}; ++ ++struct fanotify_response_info_audit_rule { ++ struct fanotify_response_info_header hdr; ++ __u32 rule_number; ++ __u32 subj_trust; ++ __u32 obj_trust; ++}; ++ ++struct audit_field; ++ ++struct audit_watch; ++ ++struct audit_tree; ++ ++struct audit_fsnotify_mark; ++ ++struct audit_krule { ++ u32 pflags; ++ u32 flags; ++ u32 listnr; ++ u32 action; ++ u32 mask[64]; ++ u32 buflen; ++ u32 field_count; ++ char *filterkey; ++ struct audit_field *fields; ++ struct audit_field *arch_f; ++ struct audit_field *inode_f; ++ struct audit_watch *watch; ++ struct audit_tree *tree; ++ struct audit_fsnotify_mark *exe; ++ struct list_head rlist; ++ struct list_head list; ++ u64 prio; ++}; ++ ++struct audit_field { ++ u32 type; ++ union { ++ u32 val; ++ kuid_t uid; ++ kgid_t gid; ++ struct { ++ char *lsm_str; ++ void *lsm_rule; ++ }; ++ }; ++ u32 op; ++}; ++ ++enum audit_nfcfgop { ++ AUDIT_XT_OP_REGISTER = 0, ++ AUDIT_XT_OP_REPLACE = 1, ++ AUDIT_XT_OP_UNREGISTER = 2, ++ AUDIT_NFT_OP_TABLE_REGISTER = 3, ++ AUDIT_NFT_OP_TABLE_UNREGISTER = 4, ++ AUDIT_NFT_OP_CHAIN_REGISTER = 5, ++ AUDIT_NFT_OP_CHAIN_UNREGISTER = 6, ++ AUDIT_NFT_OP_RULE_REGISTER = 7, ++ AUDIT_NFT_OP_RULE_UNREGISTER = 8, ++ AUDIT_NFT_OP_SET_REGISTER = 9, ++ AUDIT_NFT_OP_SET_UNREGISTER = 10, ++ AUDIT_NFT_OP_SETELEM_REGISTER = 11, ++ AUDIT_NFT_OP_SETELEM_UNREGISTER = 12, ++ AUDIT_NFT_OP_GEN_REGISTER = 13, ++ AUDIT_NFT_OP_OBJ_REGISTER = 14, ++ AUDIT_NFT_OP_OBJ_UNREGISTER = 15, ++ AUDIT_NFT_OP_OBJ_RESET = 16, ++ AUDIT_NFT_OP_FLOWTABLE_REGISTER = 17, ++ AUDIT_NFT_OP_FLOWTABLE_UNREGISTER = 18, ++ AUDIT_NFT_OP_INVALID = 19, ++}; ++ ++enum { ++ PER_LINUX = 0, ++ PER_LINUX_32BIT = 8388608, ++ PER_LINUX_FDPIC = 524288, ++ PER_SVR4 = 68157441, ++ PER_SVR3 = 83886082, ++ PER_SCOSVR3 = 117440515, ++ PER_OSR5 = 100663299, ++ PER_WYSEV386 = 83886084, ++ PER_ISCR4 = 67108869, ++ PER_BSD = 6, ++ PER_SUNOS = 67108870, ++ PER_XENIX = 83886087, ++ PER_LINUX32 = 8, ++ PER_LINUX32_3GB = 134217736, ++ PER_IRIX32 = 67108873, ++ PER_IRIXN32 = 67108874, ++ PER_IRIX64 = 67108875, ++ PER_RISCOS = 12, ++ PER_SOLARIS = 67108877, ++ PER_UW7 = 68157454, ++ PER_OSF4 = 15, ++ PER_HPUX = 16, ++ PER_MASK = 255, ++}; ++ ++enum fsnotify_iter_type { ++ FSNOTIFY_ITER_TYPE_INODE = 0, ++ FSNOTIFY_ITER_TYPE_VFSMOUNT = 1, ++ FSNOTIFY_ITER_TYPE_SB = 2, ++ FSNOTIFY_ITER_TYPE_PARENT = 3, ++ FSNOTIFY_ITER_TYPE_INODE2 = 4, ++ FSNOTIFY_ITER_TYPE_COUNT = 5, ++}; ++ ++struct audit_entry { ++ struct list_head list; ++ struct callback_head rcu; ++ struct audit_krule rule; ++}; ++ ++struct audit_aux_data { ++ struct audit_aux_data *next; ++ int type; ++}; ++ ++struct audit_chunk; ++ ++struct audit_tree_refs { ++ struct audit_tree_refs *next; ++ struct audit_chunk *c[31]; ++}; ++ ++struct audit_aux_data_pids { ++ struct audit_aux_data d; ++ pid_t target_pid[16]; ++ kuid_t target_auid[16]; ++ kuid_t target_uid[16]; ++ unsigned int target_sessionid[16]; ++ u32 target_sid[16]; ++ char target_comm[256]; ++ int pid_count; ++}; ++ ++struct audit_aux_data_bprm_fcaps { ++ struct audit_aux_data d; ++ struct audit_cap_data fcap; ++ unsigned int fcap_ver; ++ struct audit_cap_data old_pcap; ++ struct audit_cap_data new_pcap; ++}; ++ ++struct audit_nfcfgop_tab { ++ enum audit_nfcfgop op; ++ const char *s; ++}; ++ ++struct freelist_node { ++ atomic_t refs; ++ struct freelist_node *next; ++}; ++ ++struct freelist_head { ++ struct freelist_node *head; ++}; ++ ++struct rethook_node; ++ ++typedef void (*rethook_handler_t)(struct rethook_node *, void *, struct pt_regs *); ++ ++struct rethook; ++ ++struct rethook_node { ++ union { ++ struct freelist_node freelist; ++ struct callback_head rcu; ++ }; ++ struct llist_node llist; ++ struct rethook *rethook; ++ long unsigned int ret_addr; ++ long unsigned int frame; ++}; ++ ++struct rethook { ++ void *data; ++ rethook_handler_t handler; ++ struct freelist_head pool; ++ refcount_t ref; ++ struct callback_head rcu; ++}; ++ ++struct kretprobe_instance; ++ ++typedef int (*kretprobe_handler_t)(struct kretprobe_instance *, struct pt_regs *); ++ ++struct kretprobe_instance { ++ struct rethook_node node; ++ char data[0]; ++}; ++ ++struct kretprobe { ++ struct kprobe kp; ++ kretprobe_handler_t handler; ++ kretprobe_handler_t entry_handler; ++ int maxactive; ++ int nmissed; ++ size_t data_size; ++ struct rethook *rh; ++}; ++ ++struct kprobe_blacklist_entry { ++ struct list_head list; ++ long unsigned int start_addr; ++ long unsigned int end_addr; ++}; ++ ++enum perf_record_ksymbol_type { ++ PERF_RECORD_KSYMBOL_TYPE_UNKNOWN = 0, ++ PERF_RECORD_KSYMBOL_TYPE_BPF = 1, ++ PERF_RECORD_KSYMBOL_TYPE_OOL = 2, ++ PERF_RECORD_KSYMBOL_TYPE_MAX = 3, ++}; ++ ++struct kprobe_insn_page { ++ struct list_head list; ++ kprobe_opcode_t *insns; ++ struct kprobe_insn_cache *cache; ++ int nused; ++ int ngarbage; ++ char slot_used[0]; ++}; ++ ++enum kprobe_slot_state { ++ SLOT_CLEAN = 0, ++ SLOT_DIRTY = 1, ++ SLOT_USED = 2, ++}; ++ ++enum { ++ TASKSTATS_TYPE_UNSPEC = 0, ++ TASKSTATS_TYPE_PID = 1, ++ TASKSTATS_TYPE_TGID = 2, ++ TASKSTATS_TYPE_STATS = 3, ++ TASKSTATS_TYPE_AGGR_PID = 4, ++ TASKSTATS_TYPE_AGGR_TGID = 5, ++ TASKSTATS_TYPE_NULL = 6, ++ __TASKSTATS_TYPE_MAX = 7, ++}; ++ ++enum { ++ TASKSTATS_CMD_ATTR_UNSPEC = 0, ++ TASKSTATS_CMD_ATTR_PID = 1, ++ TASKSTATS_CMD_ATTR_TGID = 2, ++ TASKSTATS_CMD_ATTR_REGISTER_CPUMASK = 3, ++ TASKSTATS_CMD_ATTR_DEREGISTER_CPUMASK = 4, ++ __TASKSTATS_CMD_ATTR_MAX = 5, ++}; ++ ++struct cgroupstats { ++ __u64 nr_sleeping; ++ __u64 nr_running; ++ __u64 nr_stopped; ++ __u64 nr_uninterruptible; ++ __u64 nr_io_wait; ++}; ++ ++enum { ++ CGROUPSTATS_CMD_UNSPEC = 3, ++ CGROUPSTATS_CMD_GET = 4, ++ CGROUPSTATS_CMD_NEW = 5, ++ __CGROUPSTATS_CMD_MAX = 6, ++}; ++ ++enum { ++ CGROUPSTATS_TYPE_UNSPEC = 0, ++ CGROUPSTATS_TYPE_CGROUP_STATS = 1, ++ __CGROUPSTATS_TYPE_MAX = 2, ++}; ++ ++enum { ++ CGROUPSTATS_CMD_ATTR_UNSPEC = 0, ++ CGROUPSTATS_CMD_ATTR_FD = 1, ++ __CGROUPSTATS_CMD_ATTR_MAX = 2, ++}; ++ ++struct fd { ++ struct file *file; ++ unsigned int flags; ++}; ++ ++enum skb_drop_reason { ++ SKB_NOT_DROPPED_YET = 0, ++ SKB_CONSUMED = 1, ++ SKB_DROP_REASON_NOT_SPECIFIED = 2, ++ SKB_DROP_REASON_NO_SOCKET = 3, ++ SKB_DROP_REASON_PKT_TOO_SMALL = 4, ++ SKB_DROP_REASON_TCP_CSUM = 5, ++ SKB_DROP_REASON_SOCKET_FILTER = 6, ++ SKB_DROP_REASON_UDP_CSUM = 7, ++ SKB_DROP_REASON_NETFILTER_DROP = 8, ++ SKB_DROP_REASON_OTHERHOST = 9, ++ SKB_DROP_REASON_IP_CSUM = 10, ++ SKB_DROP_REASON_IP_INHDR = 11, ++ SKB_DROP_REASON_IP_RPFILTER = 12, ++ SKB_DROP_REASON_UNICAST_IN_L2_MULTICAST = 13, ++ SKB_DROP_REASON_XFRM_POLICY = 14, ++ SKB_DROP_REASON_IP_NOPROTO = 15, ++ SKB_DROP_REASON_SOCKET_RCVBUFF = 16, ++ SKB_DROP_REASON_PROTO_MEM = 17, ++ SKB_DROP_REASON_TCP_MD5NOTFOUND = 18, ++ SKB_DROP_REASON_TCP_MD5UNEXPECTED = 19, ++ SKB_DROP_REASON_TCP_MD5FAILURE = 20, ++ SKB_DROP_REASON_SOCKET_BACKLOG = 21, ++ SKB_DROP_REASON_TCP_FLAGS = 22, ++ SKB_DROP_REASON_TCP_ZEROWINDOW = 23, ++ SKB_DROP_REASON_TCP_OLD_DATA = 24, ++ SKB_DROP_REASON_TCP_OVERWINDOW = 25, ++ SKB_DROP_REASON_TCP_OFOMERGE = 26, ++ SKB_DROP_REASON_TCP_RFC7323_PAWS = 27, ++ SKB_DROP_REASON_TCP_INVALID_SEQUENCE = 28, ++ SKB_DROP_REASON_TCP_RESET = 29, ++ SKB_DROP_REASON_TCP_INVALID_SYN = 30, ++ SKB_DROP_REASON_TCP_CLOSE = 31, ++ SKB_DROP_REASON_TCP_FASTOPEN = 32, ++ SKB_DROP_REASON_TCP_OLD_ACK = 33, ++ SKB_DROP_REASON_TCP_TOO_OLD_ACK = 34, ++ SKB_DROP_REASON_TCP_ACK_UNSENT_DATA = 35, ++ SKB_DROP_REASON_TCP_OFO_QUEUE_PRUNE = 36, ++ SKB_DROP_REASON_TCP_OFO_DROP = 37, ++ SKB_DROP_REASON_IP_OUTNOROUTES = 38, ++ SKB_DROP_REASON_BPF_CGROUP_EGRESS = 39, ++ SKB_DROP_REASON_IPV6DISABLED = 40, ++ SKB_DROP_REASON_NEIGH_CREATEFAIL = 41, ++ SKB_DROP_REASON_NEIGH_FAILED = 42, ++ SKB_DROP_REASON_NEIGH_QUEUEFULL = 43, ++ SKB_DROP_REASON_NEIGH_DEAD = 44, ++ SKB_DROP_REASON_TC_EGRESS = 45, ++ SKB_DROP_REASON_QDISC_DROP = 46, ++ SKB_DROP_REASON_CPU_BACKLOG = 47, ++ SKB_DROP_REASON_XDP = 48, ++ SKB_DROP_REASON_TC_INGRESS = 49, ++ SKB_DROP_REASON_UNHANDLED_PROTO = 50, ++ SKB_DROP_REASON_SKB_CSUM = 51, ++ SKB_DROP_REASON_SKB_GSO_SEG = 52, ++ SKB_DROP_REASON_SKB_UCOPY_FAULT = 53, ++ SKB_DROP_REASON_DEV_HDR = 54, ++ SKB_DROP_REASON_DEV_READY = 55, ++ SKB_DROP_REASON_FULL_RING = 56, ++ SKB_DROP_REASON_NOMEM = 57, ++ SKB_DROP_REASON_HDR_TRUNC = 58, ++ SKB_DROP_REASON_TAP_FILTER = 59, ++ SKB_DROP_REASON_TAP_TXFILTER = 60, ++ SKB_DROP_REASON_ICMP_CSUM = 61, ++ SKB_DROP_REASON_INVALID_PROTO = 62, ++ SKB_DROP_REASON_IP_INADDRERRORS = 63, ++ SKB_DROP_REASON_IP_INNOROUTES = 64, ++ SKB_DROP_REASON_PKT_TOO_BIG = 65, ++ SKB_DROP_REASON_DUP_FRAG = 66, ++ SKB_DROP_REASON_FRAG_REASM_TIMEOUT = 67, ++ SKB_DROP_REASON_FRAG_TOO_FAR = 68, ++ SKB_DROP_REASON_TCP_MINTTL = 69, ++ SKB_DROP_REASON_IPV6_BAD_EXTHDR = 70, ++ SKB_DROP_REASON_IPV6_NDISC_FRAG = 71, ++ SKB_DROP_REASON_IPV6_NDISC_HOP_LIMIT = 72, ++ SKB_DROP_REASON_IPV6_NDISC_BAD_CODE = 73, ++ SKB_DROP_REASON_IPV6_NDISC_BAD_OPTIONS = 74, ++ SKB_DROP_REASON_IPV6_NDISC_NS_OTHERHOST = 75, ++ SKB_DROP_REASON_MAX = 76, ++ SKB_DROP_REASON_SUBSYS_MASK = 4294901760, ++}; ++ ++struct genlmsghdr { ++ __u8 cmd; ++ __u8 version; ++ __u16 reserved; ++}; ++ ++enum { ++ NLA_UNSPEC = 0, ++ NLA_U8 = 1, ++ NLA_U16 = 2, ++ NLA_U32 = 3, ++ NLA_U64 = 4, ++ NLA_STRING = 5, ++ NLA_FLAG = 6, ++ NLA_MSECS = 7, ++ NLA_NESTED = 8, ++ NLA_NESTED_ARRAY = 9, ++ NLA_NUL_STRING = 10, ++ NLA_BINARY = 11, ++ NLA_S8 = 12, ++ NLA_S16 = 13, ++ NLA_S32 = 14, ++ NLA_S64 = 15, ++ NLA_BITFIELD32 = 16, ++ NLA_REJECT = 17, ++ NLA_BE16 = 18, ++ NLA_BE32 = 19, ++ __NLA_TYPE_MAX = 20, ++}; ++ ++struct genl_multicast_group { ++ char name[16]; ++ u8 flags; ++}; ++ ++struct genl_split_ops; ++ ++struct genl_info; ++ ++struct genl_ops; ++ ++struct genl_small_ops; ++ ++struct genl_family { ++ unsigned int hdrsize; ++ char name[16]; ++ unsigned int version; ++ unsigned int maxattr; ++ u8 netnsok: 1; ++ u8 parallel_ops: 1; ++ u8 n_ops; ++ u8 n_small_ops; ++ u8 n_split_ops; ++ u8 n_mcgrps; ++ u8 resv_start_op; ++ const struct nla_policy *policy; ++ int (*pre_doit)(const struct genl_split_ops *, struct sk_buff *, struct genl_info *); ++ void (*post_doit)(const struct genl_split_ops *, struct sk_buff *, struct genl_info *); ++ const struct genl_ops *ops; ++ const struct genl_small_ops *small_ops; ++ const struct genl_split_ops *split_ops; ++ const struct genl_multicast_group *mcgrps; ++ struct module *module; ++ int id; ++ unsigned int mcgrp_offset; ++}; ++ ++struct genl_split_ops { ++ union { ++ struct { ++ int (*pre_doit)(const struct genl_split_ops *, struct sk_buff *, struct genl_info *); ++ int (*doit)(struct sk_buff *, struct genl_info *); ++ void (*post_doit)(const struct genl_split_ops *, struct sk_buff *, struct genl_info *); ++ }; ++ struct { ++ int (*start)(struct netlink_callback *); ++ int (*dumpit)(struct sk_buff *, struct netlink_callback *); ++ int (*done)(struct netlink_callback *); ++ }; ++ }; ++ const struct nla_policy *policy; ++ unsigned int maxattr; ++ u8 cmd; ++ u8 internal_flags; ++ u8 flags; ++ u8 validate; ++}; ++ ++struct genl_info { ++ u32 snd_seq; ++ u32 snd_portid; ++ struct nlmsghdr *nlhdr; ++ struct genlmsghdr *genlhdr; ++ void *userhdr; ++ struct nlattr **attrs; ++ possible_net_t _net; ++ void *user_ptr[2]; ++ struct netlink_ext_ack *extack; ++}; ++ ++struct genl_ops { ++ int (*doit)(struct sk_buff *, struct genl_info *); ++ int (*start)(struct netlink_callback *); ++ int (*dumpit)(struct sk_buff *, struct netlink_callback *); ++ int (*done)(struct netlink_callback *); ++ const struct nla_policy *policy; ++ unsigned int maxattr; ++ u8 cmd; ++ u8 internal_flags; ++ u8 flags; ++ u8 validate; ++}; ++ ++struct genl_small_ops { ++ int (*doit)(struct sk_buff *, struct genl_info *); ++ int (*dumpit)(struct sk_buff *, struct netlink_callback *); ++ u8 cmd; ++ u8 internal_flags; ++ u8 flags; ++ u8 validate; ++}; ++ ++enum genl_validate_flags { ++ GENL_DONT_VALIDATE_STRICT = 1, ++ GENL_DONT_VALIDATE_DUMP = 2, ++ GENL_DONT_VALIDATE_DUMP_STRICT = 4, ++}; ++ ++struct listener { ++ struct list_head list; ++ pid_t pid; ++ char valid; ++}; ++ ++struct listener_list { ++ struct rw_semaphore sem; ++ struct list_head list; ++}; ++ ++enum actions { ++ REGISTER = 0, ++ DEREGISTER = 1, ++ CPU_DONT_CARE = 2, ++}; ++ ++typedef __u32 pcp_op_T_____6; ++ ++enum ring_buffer_flags { ++ RB_FL_OVERWRITE = 1, ++}; ++ ++typedef struct vfsmount * (*debugfs_automount_t)(struct dentry *, void *); ++ ++struct ftrace_hash { ++ long unsigned int size_bits; ++ struct hlist_head *buckets; ++ long unsigned int count; ++ long unsigned int flags; ++ struct callback_head rcu; ++}; ++ ++struct partial_page { ++ unsigned int offset; ++ unsigned int len; ++ long unsigned int private; ++}; ++ ++struct splice_pipe_desc { ++ struct page **pages; ++ struct partial_page *partial; ++ int nr_pages; ++ unsigned int nr_pages_max; ++ const struct pipe_buf_operations *ops; ++ void (*spd_release)(struct splice_pipe_desc *, unsigned int); ++}; ++ ++struct trace_export { ++ struct trace_export *next; ++ void (*write)(struct trace_export *, const void *, unsigned int); ++ int flags; ++}; ++ ++enum fsnotify_data_type { ++ FSNOTIFY_EVENT_NONE = 0, ++ FSNOTIFY_EVENT_PATH = 1, ++ FSNOTIFY_EVENT_INODE = 2, ++ FSNOTIFY_EVENT_DENTRY = 3, ++ FSNOTIFY_EVENT_ERROR = 4, ++}; ++ ++struct prog_entry; ++ ++struct event_filter { ++ struct prog_entry *prog; ++ char *filter_string; ++}; ++ ++struct trace_array_cpu; ++ ++struct array_buffer { ++ struct trace_array *tr; ++ struct trace_buffer *buffer; ++ struct trace_array_cpu *data; ++ u64 time_start; ++ int cpu; ++}; ++ ++struct trace_pid_list; ++ ++struct trace_options; ++ ++struct cond_snapshot; ++ ++struct trace_func_repeats; ++ ++struct trace_array { ++ struct list_head list; ++ char *name; ++ struct array_buffer array_buffer; ++ struct array_buffer max_buffer; ++ bool allocated_snapshot; ++ long unsigned int max_latency; ++ struct dentry *d_max_latency; ++ struct work_struct fsnotify_work; ++ struct irq_work fsnotify_irqwork; ++ struct trace_pid_list *filtered_pids; ++ struct trace_pid_list *filtered_no_pids; ++ arch_spinlock_t max_lock; ++ int buffer_disabled; ++ int sys_refcount_enter; ++ int sys_refcount_exit; ++ struct trace_event_file *enter_syscall_files[451]; ++ struct trace_event_file *exit_syscall_files[451]; ++ int stop_count; ++ int clock_id; ++ int nr_topts; ++ bool clear_trace; ++ int buffer_percent; ++ unsigned int n_err_log_entries; ++ struct tracer *current_trace; ++ unsigned int trace_flags; ++ unsigned char trace_flags_index[32]; ++ unsigned int flags; ++ raw_spinlock_t start_lock; ++ struct list_head err_log; ++ struct dentry *dir; ++ struct dentry *options; ++ struct dentry *percpu_dir; ++ struct dentry *event_dir; ++ struct trace_options *topts; ++ struct list_head systems; ++ struct list_head events; ++ struct trace_event_file *trace_marker_file; ++ cpumask_var_t tracing_cpumask; ++ int ref; ++ int trace_ref; ++ struct ftrace_ops *ops; ++ struct trace_pid_list *function_pids; ++ struct trace_pid_list *function_no_pids; ++ struct list_head func_probes; ++ struct list_head mod_trace; ++ struct list_head mod_notrace; ++ int function_enabled; ++ int no_filter_buffering_ref; ++ struct list_head hist_vars; ++ struct cond_snapshot *cond_snapshot; ++ struct trace_func_repeats *last_func_repeats; ++}; ++ ++struct tracer_flags; ++ ++struct tracer { ++ const char *name; ++ int (*init)(struct trace_array *); ++ void (*reset)(struct trace_array *); ++ void (*start)(struct trace_array *); ++ void (*stop)(struct trace_array *); ++ int (*update_thresh)(struct trace_array *); ++ void (*open)(struct trace_iterator *); ++ void (*pipe_open)(struct trace_iterator *); ++ void (*close)(struct trace_iterator *); ++ void (*pipe_close)(struct trace_iterator *); ++ ssize_t (*read)(struct trace_iterator *, struct file *, char *, size_t, loff_t *); ++ ssize_t (*splice_read)(struct trace_iterator *, struct file *, loff_t *, struct pipe_inode_info *, size_t, unsigned int); ++ void (*print_header)(struct seq_file *); ++ enum print_line_t (*print_line)(struct trace_iterator *); ++ int (*set_flag)(struct trace_array *, u32, u32, int); ++ int (*flag_changed)(struct trace_array *, u32, int); ++ struct tracer *next; ++ struct tracer_flags *flags; ++ int enabled; ++ bool print_max; ++ bool allow_instances; ++ bool use_max_tr; ++ bool noboot; ++}; ++ ++enum trace_iter_flags { ++ TRACE_FILE_LAT_FMT = 1, ++ TRACE_FILE_ANNOTATE = 2, ++ TRACE_FILE_TIME_IN_NS = 4, ++}; ++ ++enum trace_flag_type { ++ TRACE_FLAG_IRQS_OFF = 1, ++ TRACE_FLAG_IRQS_NOSUPPORT = 2, ++ TRACE_FLAG_NEED_RESCHED = 4, ++ TRACE_FLAG_HARDIRQ = 8, ++ TRACE_FLAG_SOFTIRQ = 16, ++ TRACE_FLAG_PREEMPT_RESCHED = 32, ++ TRACE_FLAG_NMI = 64, ++ TRACE_FLAG_BH_OFF = 128, ++}; ++ ++struct event_subsystem; ++ ++struct trace_subsystem_dir { ++ struct list_head list; ++ struct event_subsystem *subsystem; ++ struct trace_array *tr; ++ struct dentry *entry; ++ int ref_count; ++ int nr_events; ++}; ++ ++enum event_trigger_type { ++ ETT_NONE = 0, ++ ETT_TRACE_ONOFF = 1, ++ ETT_SNAPSHOT = 2, ++ ETT_STACKTRACE = 4, ++ ETT_EVENT_ENABLE = 8, ++ ETT_EVENT_HIST = 16, ++ ETT_HIST_ENABLE = 32, ++ ETT_EVENT_EPROBE = 64, ++}; ++ ++union lower_chunk { ++ union lower_chunk *next; ++ long unsigned int data[256]; ++}; ++ ++union upper_chunk { ++ union upper_chunk *next; ++ union lower_chunk *data[256]; ++}; ++ ++struct trace_pid_list { ++ raw_spinlock_t lock; ++ struct irq_work refill_irqwork; ++ union upper_chunk *upper[256]; ++ union upper_chunk *upper_list; ++ union lower_chunk *lower_list; ++ int free_upper_chunks; ++ int free_lower_chunks; ++}; ++ ++enum trace_type { ++ __TRACE_FIRST_TYPE = 0, ++ TRACE_FN = 1, ++ TRACE_CTX = 2, ++ TRACE_WAKE = 3, ++ TRACE_STACK = 4, ++ TRACE_PRINT = 5, ++ TRACE_BPRINT = 6, ++ TRACE_MMIO_RW = 7, ++ TRACE_MMIO_MAP = 8, ++ TRACE_BRANCH = 9, ++ TRACE_GRAPH_RET = 10, ++ TRACE_GRAPH_ENT = 11, ++ TRACE_USER_STACK = 12, ++ TRACE_BLK = 13, ++ TRACE_BPUTS = 14, ++ TRACE_HWLAT = 15, ++ TRACE_OSNOISE = 16, ++ TRACE_TIMERLAT = 17, ++ TRACE_RAW_DATA = 18, ++ TRACE_FUNC_REPEATS = 19, ++ __TRACE_LAST_TYPE = 20, ++}; ++ ++struct ftrace_entry { ++ struct trace_entry ent; ++ long unsigned int ip; ++ long unsigned int parent_ip; ++}; ++ ++struct stack_entry { ++ struct trace_entry ent; ++ int size; ++ long unsigned int caller[8]; ++}; ++ ++struct bprint_entry { ++ struct trace_entry ent; ++ long unsigned int ip; ++ const char *fmt; ++ u32 buf[0]; ++}; ++ ++struct print_entry { ++ struct trace_entry ent; ++ long unsigned int ip; ++ char buf[0]; ++}; ++ ++struct raw_data_entry { ++ struct trace_entry ent; ++ unsigned int id; ++ char buf[0]; ++}; ++ ++struct bputs_entry { ++ struct trace_entry ent; ++ long unsigned int ip; ++ const char *str; ++}; ++ ++struct func_repeats_entry { ++ struct trace_entry ent; ++ long unsigned int ip; ++ long unsigned int parent_ip; ++ u16 count; ++ u16 top_delta_ts; ++ u32 bottom_delta_ts; ++}; ++ ++struct trace_array_cpu { ++ atomic_t disabled; ++ void *buffer_page; ++ long unsigned int entries; ++ long unsigned int saved_latency; ++ long unsigned int critical_start; ++ long unsigned int critical_end; ++ long unsigned int critical_sequence; ++ long unsigned int nice; ++ long unsigned int policy; ++ long unsigned int rt_priority; ++ long unsigned int skipped_entries; ++ u64 preempt_timestamp; ++ pid_t pid; ++ kuid_t uid; ++ char comm[16]; ++ int ftrace_ignore_pid; ++ bool ignore_pid; ++}; ++ ++struct trace_option_dentry; ++ ++struct trace_options { ++ struct tracer *tracer; ++ struct trace_option_dentry *topts; ++}; ++ ++struct tracer_opt; ++ ++struct trace_option_dentry { ++ struct tracer_opt *opt; ++ struct tracer_flags *flags; ++ struct trace_array *tr; ++ struct dentry *entry; ++}; ++ ++typedef bool (*cond_update_fn_t)(struct trace_array *, void *); ++ ++struct cond_snapshot { ++ void *cond_data; ++ cond_update_fn_t update; ++}; ++ ++struct trace_func_repeats { ++ long unsigned int ip; ++ long unsigned int parent_ip; ++ long unsigned int count; ++ u64 ts_last_call; ++}; ++ ++enum { ++ TRACE_ARRAY_FL_GLOBAL = 1, ++}; ++ ++struct tracer_opt { ++ const char *name; ++ u32 bit; ++}; ++ ++struct tracer_flags { ++ u32 val; ++ struct tracer_opt *opts; ++ struct tracer *trace; ++}; ++ ++struct ftrace_func_command { ++ struct list_head list; ++ char *name; ++ int (*func)(struct trace_array *, struct ftrace_hash *, char *, char *, char *, int); ++}; ++ ++struct ftrace_probe_ops { ++ void (*func)(long unsigned int, long unsigned int, struct trace_array *, struct ftrace_probe_ops *, void *); ++ int (*init)(struct ftrace_probe_ops *, struct trace_array *, long unsigned int, void *, void **); ++ void (*free)(struct ftrace_probe_ops *, struct trace_array *, long unsigned int, void *); ++ int (*print)(struct seq_file *, long unsigned int, struct ftrace_probe_ops *, void *); ++}; ++ ++typedef int (*ftrace_mapper_func)(void *); ++ ++struct trace_parser { ++ bool cont; ++ char *buffer; ++ unsigned int idx; ++ unsigned int size; ++}; ++ ++enum trace_iterator_bits { ++ TRACE_ITER_PRINT_PARENT_BIT = 0, ++ TRACE_ITER_SYM_OFFSET_BIT = 1, ++ TRACE_ITER_SYM_ADDR_BIT = 2, ++ TRACE_ITER_VERBOSE_BIT = 3, ++ TRACE_ITER_RAW_BIT = 4, ++ TRACE_ITER_HEX_BIT = 5, ++ TRACE_ITER_BIN_BIT = 6, ++ TRACE_ITER_BLOCK_BIT = 7, ++ TRACE_ITER_FIELDS_BIT = 8, ++ TRACE_ITER_PRINTK_BIT = 9, ++ TRACE_ITER_ANNOTATE_BIT = 10, ++ TRACE_ITER_USERSTACKTRACE_BIT = 11, ++ TRACE_ITER_SYM_USEROBJ_BIT = 12, ++ TRACE_ITER_PRINTK_MSGONLY_BIT = 13, ++ TRACE_ITER_CONTEXT_INFO_BIT = 14, ++ TRACE_ITER_LATENCY_FMT_BIT = 15, ++ TRACE_ITER_RECORD_CMD_BIT = 16, ++ TRACE_ITER_RECORD_TGID_BIT = 17, ++ TRACE_ITER_OVERWRITE_BIT = 18, ++ TRACE_ITER_STOP_ON_FREE_BIT = 19, ++ TRACE_ITER_IRQ_INFO_BIT = 20, ++ TRACE_ITER_MARKERS_BIT = 21, ++ TRACE_ITER_EVENT_FORK_BIT = 22, ++ TRACE_ITER_PAUSE_ON_TRACE_BIT = 23, ++ TRACE_ITER_HASH_PTR_BIT = 24, ++ TRACE_ITER_FUNCTION_BIT = 25, ++ TRACE_ITER_FUNC_FORK_BIT = 26, ++ TRACE_ITER_DISPLAY_GRAPH_BIT = 27, ++ TRACE_ITER_STACKTRACE_BIT = 28, ++ TRACE_ITER_LAST_BIT = 29, ++}; ++ ++enum trace_iterator_flags { ++ TRACE_ITER_PRINT_PARENT = 1, ++ TRACE_ITER_SYM_OFFSET = 2, ++ TRACE_ITER_SYM_ADDR = 4, ++ TRACE_ITER_VERBOSE = 8, ++ TRACE_ITER_RAW = 16, ++ TRACE_ITER_HEX = 32, ++ TRACE_ITER_BIN = 64, ++ TRACE_ITER_BLOCK = 128, ++ TRACE_ITER_FIELDS = 256, ++ TRACE_ITER_PRINTK = 512, ++ TRACE_ITER_ANNOTATE = 1024, ++ TRACE_ITER_USERSTACKTRACE = 2048, ++ TRACE_ITER_SYM_USEROBJ = 4096, ++ TRACE_ITER_PRINTK_MSGONLY = 8192, ++ TRACE_ITER_CONTEXT_INFO = 16384, ++ TRACE_ITER_LATENCY_FMT = 32768, ++ TRACE_ITER_RECORD_CMD = 65536, ++ TRACE_ITER_RECORD_TGID = 131072, ++ TRACE_ITER_OVERWRITE = 262144, ++ TRACE_ITER_STOP_ON_FREE = 524288, ++ TRACE_ITER_IRQ_INFO = 1048576, ++ TRACE_ITER_MARKERS = 2097152, ++ TRACE_ITER_EVENT_FORK = 4194304, ++ TRACE_ITER_PAUSE_ON_TRACE = 8388608, ++ TRACE_ITER_HASH_PTR = 16777216, ++ TRACE_ITER_FUNCTION = 33554432, ++ TRACE_ITER_FUNC_FORK = 67108864, ++ TRACE_ITER_DISPLAY_GRAPH = 134217728, ++ TRACE_ITER_STACKTRACE = 268435456, ++}; ++ ++struct event_subsystem { ++ struct list_head list; ++ const char *name; ++ struct event_filter *filter; ++ int ref_count; ++}; ++ ++struct trace_min_max_param { ++ struct mutex *lock; ++ u64 *val; ++ u64 *min; ++ u64 *max; ++}; ++ ++struct trace_eval_map_head { ++ struct module *mod; ++ long unsigned int length; ++}; ++ ++union trace_eval_map_item; ++ ++struct trace_eval_map_tail { ++ union trace_eval_map_item *next; ++ const char *end; ++}; ++ ++union trace_eval_map_item { ++ struct trace_eval_map map; ++ struct trace_eval_map_head head; ++ struct trace_eval_map_tail tail; ++}; ++ ++struct saved_cmdlines_buffer { ++ unsigned int map_pid_to_cmdline[32769]; ++ unsigned int *map_cmdline_to_pid; ++ unsigned int cmdline_num; ++ int cmdline_idx; ++ char *saved_cmdlines; ++}; ++ ++struct ftrace_stack { ++ long unsigned int calls[1024]; ++}; ++ ++struct ftrace_stacks { ++ struct ftrace_stack stacks[4]; ++}; ++ ++struct trace_buffer_struct { ++ int nesting; ++ char buffer[4096]; ++}; ++ ++struct ftrace_buffer_info { ++ struct trace_iterator iter; ++ void *spare; ++ unsigned int spare_cpu; ++ unsigned int read; ++}; ++ ++struct err_info { ++ const char **errs; ++ u8 type; ++ u16 pos; ++ u64 ts; ++}; ++ ++struct tracing_log_err { ++ struct list_head list; ++ struct err_info info; ++ char loc[128]; ++ char *cmd; ++}; ++ ++struct buffer_ref { ++ struct trace_buffer *buffer; ++ void *page; ++ int cpu; ++ refcount_t refcount; ++}; ++ ++struct ftrace_func_entry { ++ struct hlist_node hlist; ++ long unsigned int ip; ++ long unsigned int direct; ++}; ++ ++struct ftrace_graph_ent { ++ long unsigned int func; ++ int depth; ++} __attribute__((packed)); ++ ++struct ftrace_graph_ret { ++ long unsigned int func; ++ int depth; ++ unsigned int overrun; ++ long long unsigned int calltime; ++ long long unsigned int rettime; ++}; ++ ++typedef void (*trace_func_graph_ret_t)(struct ftrace_graph_ret *); ++ ++typedef int (*trace_func_graph_ent_t)(struct ftrace_graph_ent *); ++ ++struct fgraph_ops { ++ trace_func_graph_ent_t entryfunc; ++ trace_func_graph_ret_t retfunc; ++}; ++ ++struct ctx_switch_entry { ++ struct trace_entry ent; ++ unsigned int prev_pid; ++ unsigned int next_pid; ++ unsigned int next_cpu; ++ unsigned char prev_prio; ++ unsigned char prev_state; ++ unsigned char next_prio; ++ unsigned char next_state; ++}; ++ ++enum { ++ FTRACE_HASH_FL_MOD = 1, ++}; ++ ++enum { ++ TRACE_NOP_OPT_ACCEPT = 1, ++ TRACE_NOP_OPT_REFUSE = 2, ++}; ++ ++struct blk_crypto_key; ++ ++struct bio_crypt_ctx { ++ const struct blk_crypto_key *bc_key; ++ u64 bc_dun[4]; ++}; ++ ++enum req_op { ++ REQ_OP_READ = 0, ++ REQ_OP_WRITE = 1, ++ REQ_OP_FLUSH = 2, ++ REQ_OP_DISCARD = 3, ++ REQ_OP_SECURE_ERASE = 5, ++ REQ_OP_WRITE_ZEROES = 9, ++ REQ_OP_ZONE_OPEN = 10, ++ REQ_OP_ZONE_CLOSE = 11, ++ REQ_OP_ZONE_FINISH = 12, ++ REQ_OP_ZONE_APPEND = 13, ++ REQ_OP_ZONE_RESET = 15, ++ REQ_OP_ZONE_RESET_ALL = 17, ++ REQ_OP_DRV_IN = 34, ++ REQ_OP_DRV_OUT = 35, ++ REQ_OP_LAST = 36, ++}; ++ ++enum req_flag_bits { ++ __REQ_FAILFAST_DEV = 8, ++ __REQ_FAILFAST_TRANSPORT = 9, ++ __REQ_FAILFAST_DRIVER = 10, ++ __REQ_SYNC = 11, ++ __REQ_META = 12, ++ __REQ_PRIO = 13, ++ __REQ_NOMERGE = 14, ++ __REQ_IDLE = 15, ++ __REQ_INTEGRITY = 16, ++ __REQ_FUA = 17, ++ __REQ_PREFLUSH = 18, ++ __REQ_RAHEAD = 19, ++ __REQ_BACKGROUND = 20, ++ __REQ_NOWAIT = 21, ++ __REQ_POLLED = 22, ++ __REQ_ALLOC_CACHE = 23, ++ __REQ_SWAP = 24, ++ __REQ_DRV = 25, ++ __REQ_FS_PRIVATE = 26, ++ __REQ_NOUNMAP = 27, ++ __REQ_NR_BITS = 28, ++}; ++ ++struct sbitmap_word { ++ long unsigned int word; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long unsigned int cleared; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct sbitmap { ++ unsigned int depth; ++ unsigned int shift; ++ unsigned int map_nr; ++ bool round_robin; ++ struct sbitmap_word *map; ++ unsigned int *alloc_hint; ++}; ++ ++struct sbq_wait_state { ++ wait_queue_head_t wait; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct sbitmap_queue { ++ struct sbitmap sb; ++ unsigned int wake_batch; ++ atomic_t wake_index; ++ struct sbq_wait_state *ws; ++ atomic_t ws_active; ++ unsigned int min_shallow_depth; ++ atomic_t completion_cnt; ++ atomic_t wakeup_cnt; ++}; ++ ++typedef __u32 req_flags_t; ++ ++enum mq_rq_state { ++ MQ_RQ_IDLE = 0, ++ MQ_RQ_IN_FLIGHT = 1, ++ MQ_RQ_COMPLETE = 2, ++}; ++ ++enum rq_end_io_ret { ++ RQ_END_IO_NONE = 0, ++ RQ_END_IO_FREE = 1, ++}; ++ ++typedef enum rq_end_io_ret rq_end_io_fn(struct request *, blk_status_t); ++ ++struct blk_crypto_keyslot; ++ ++struct request { ++ struct request_queue *q; ++ struct blk_mq_ctx *mq_ctx; ++ struct blk_mq_hw_ctx *mq_hctx; ++ blk_opf_t cmd_flags; ++ req_flags_t rq_flags; ++ int tag; ++ int internal_tag; ++ unsigned int timeout; ++ unsigned int __data_len; ++ sector_t __sector; ++ struct bio *bio; ++ struct bio *biotail; ++ union { ++ struct list_head queuelist; ++ struct request *rq_next; ++ }; ++ struct block_device *part; ++ u64 alloc_time_ns; ++ u64 start_time_ns; ++ u64 io_start_time_ns; ++ short unsigned int wbt_flags; ++ short unsigned int stats_sectors; ++ short unsigned int nr_phys_segments; ++ short unsigned int nr_integrity_segments; ++ struct bio_crypt_ctx *crypt_ctx; ++ struct blk_crypto_keyslot *crypt_keyslot; ++ short unsigned int ioprio; ++ enum mq_rq_state state; ++ atomic_t ref; ++ long unsigned int deadline; ++ union { ++ struct hlist_node hash; ++ struct llist_node ipi_list; ++ }; ++ union { ++ struct rb_node rb_node; ++ struct bio_vec special_vec; ++ void *completion_data; ++ }; ++ union { ++ struct { ++ struct io_cq *icq; ++ void *priv[2]; ++ } elv; ++ struct { ++ unsigned int seq; ++ struct list_head list; ++ rq_end_io_fn *saved_end_io; ++ } flush; ++ }; ++ union { ++ struct __call_single_data csd; ++ u64 fifo_time; ++ }; ++ rq_end_io_fn *end_io; ++ void *end_io_data; ++}; ++ ++struct blk_mq_tags { ++ unsigned int nr_tags; ++ unsigned int nr_reserved_tags; ++ unsigned int active_queues; ++ struct sbitmap_queue bitmap_tags; ++ struct sbitmap_queue breserved_tags; ++ struct request **rqs; ++ struct request **static_rqs; ++ struct list_head page_list; ++ spinlock_t lock; ++}; ++ ++struct rchan; ++ ++struct blk_trace { ++ int trace_state; ++ struct rchan *rchan; ++ long unsigned int *sequence; ++ unsigned char *msg_data; ++ u16 act_mask; ++ u64 start_lba; ++ u64 end_lba; ++ u32 pid; ++ u32 dev; ++ struct dentry *dir; ++ struct list_head running_list; ++ atomic_t dropped; ++}; ++ ++struct blk_flush_queue { ++ unsigned int flush_pending_idx: 1; ++ unsigned int flush_running_idx: 1; ++ blk_status_t rq_status; ++ long unsigned int flush_pending_since; ++ struct list_head flush_queue[2]; ++ struct list_head flush_data_in_flight; ++ struct request *flush_rq; ++ spinlock_t mq_flush_lock; ++}; ++ ++struct blk_mq_queue_map { ++ unsigned int *mq_map; ++ unsigned int nr_queues; ++ unsigned int queue_offset; ++}; ++ ++struct blk_mq_tag_set { ++ const struct blk_mq_ops *ops; ++ struct blk_mq_queue_map map[3]; ++ unsigned int nr_maps; ++ unsigned int nr_hw_queues; ++ unsigned int queue_depth; ++ unsigned int reserved_tags; ++ unsigned int cmd_size; ++ int numa_node; ++ unsigned int timeout; ++ unsigned int flags; ++ void *driver_data; ++ struct blk_mq_tags **tags; ++ struct blk_mq_tags *shared_tags; ++ struct mutex tag_list_lock; ++ struct list_head tag_list; ++ struct srcu_struct *srcu; ++}; ++ ++struct blk_mq_hw_ctx { ++ struct { ++ spinlock_t lock; ++ struct list_head dispatch; ++ long unsigned int state; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ }; ++ struct delayed_work run_work; ++ cpumask_var_t cpumask; ++ int next_cpu; ++ int next_cpu_batch; ++ long unsigned int flags; ++ void *sched_data; ++ struct request_queue *queue; ++ struct blk_flush_queue *fq; ++ void *driver_data; ++ struct sbitmap ctx_map; ++ struct blk_mq_ctx *dispatch_from; ++ unsigned int dispatch_busy; ++ short unsigned int type; ++ short unsigned int nr_ctx; ++ struct blk_mq_ctx **ctxs; ++ spinlock_t dispatch_wait_lock; ++ wait_queue_entry_t dispatch_wait; ++ atomic_t wait_index; ++ struct blk_mq_tags *tags; ++ struct blk_mq_tags *sched_tags; ++ long unsigned int queued; ++ long unsigned int run; ++ unsigned int numa_node; ++ unsigned int queue_num; ++ atomic_t nr_active; ++ struct hlist_node cpuhp_online; ++ struct hlist_node cpuhp_dead; ++ struct kobject kobj; ++ struct dentry *debugfs_dir; ++ struct dentry *sched_debugfs_dir; ++ struct list_head hctx_list; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct blk_mq_queue_data { ++ struct request *rq; ++ bool last; ++}; ++ ++struct rchan_buf { ++ void *start; ++ void *data; ++ size_t offset; ++ size_t subbufs_produced; ++ size_t subbufs_consumed; ++ struct rchan *chan; ++ wait_queue_head_t read_wait; ++ struct irq_work wakeup_work; ++ struct dentry *dentry; ++ struct kref kref; ++ struct page **page_array; ++ unsigned int page_count; ++ unsigned int finalized; ++ size_t *padding; ++ size_t prev_padding; ++ size_t bytes_consumed; ++ size_t early_bytes; ++ unsigned int cpu; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct rchan_callbacks; ++ ++struct rchan { ++ u32 version; ++ size_t subbuf_size; ++ size_t n_subbufs; ++ size_t alloc_size; ++ const struct rchan_callbacks *cb; ++ struct kref kref; ++ void *private_data; ++ size_t last_toobig; ++ struct rchan_buf **buf; ++ int is_global; ++ struct list_head list; ++ struct dentry *parent; ++ int has_base_filename; ++ char base_filename[255]; ++}; ++ ++struct rchan_callbacks { ++ int (*subbuf_start)(struct rchan_buf *, void *, void *, size_t); ++ struct dentry * (*create_buf_file)(const char *, struct dentry *, umode_t, struct rchan_buf *, int *); ++ int (*remove_buf_file)(struct dentry *); ++}; ++ ++enum blktrace_cat { ++ BLK_TC_READ = 1, ++ BLK_TC_WRITE = 2, ++ BLK_TC_FLUSH = 4, ++ BLK_TC_SYNC = 8, ++ BLK_TC_SYNCIO = 8, ++ BLK_TC_QUEUE = 16, ++ BLK_TC_REQUEUE = 32, ++ BLK_TC_ISSUE = 64, ++ BLK_TC_COMPLETE = 128, ++ BLK_TC_FS = 256, ++ BLK_TC_PC = 512, ++ BLK_TC_NOTIFY = 1024, ++ BLK_TC_AHEAD = 2048, ++ BLK_TC_META = 4096, ++ BLK_TC_DISCARD = 8192, ++ BLK_TC_DRV_DATA = 16384, ++ BLK_TC_FUA = 32768, ++ BLK_TC_END = 32768, ++}; ++ ++enum blktrace_act { ++ __BLK_TA_QUEUE = 1, ++ __BLK_TA_BACKMERGE = 2, ++ __BLK_TA_FRONTMERGE = 3, ++ __BLK_TA_GETRQ = 4, ++ __BLK_TA_SLEEPRQ = 5, ++ __BLK_TA_REQUEUE = 6, ++ __BLK_TA_ISSUE = 7, ++ __BLK_TA_COMPLETE = 8, ++ __BLK_TA_PLUG = 9, ++ __BLK_TA_UNPLUG_IO = 10, ++ __BLK_TA_UNPLUG_TIMER = 11, ++ __BLK_TA_INSERT = 12, ++ __BLK_TA_SPLIT = 13, ++ __BLK_TA_BOUNCE = 14, ++ __BLK_TA_REMAP = 15, ++ __BLK_TA_ABORT = 16, ++ __BLK_TA_DRV_DATA = 17, ++ __BLK_TA_CGROUP = 256, ++}; ++ ++enum blktrace_notify { ++ __BLK_TN_PROCESS = 0, ++ __BLK_TN_TIMESTAMP = 1, ++ __BLK_TN_MESSAGE = 2, ++ __BLK_TN_CGROUP = 256, ++}; ++ ++struct blk_io_trace { ++ __u32 magic; ++ __u32 sequence; ++ __u64 time; ++ __u64 sector; ++ __u32 bytes; ++ __u32 action; ++ __u32 pid; ++ __u32 device; ++ __u32 cpu; ++ __u16 error; ++ __u16 pdu_len; ++}; ++ ++struct blk_io_trace_remap { ++ __be32 device_from; ++ __be32 device_to; ++ __be64 sector_from; ++}; ++ ++enum { ++ Blktrace_setup = 1, ++ Blktrace_running = 2, ++ Blktrace_stopped = 3, ++}; ++ ++struct blk_user_trace_setup { ++ char name[32]; ++ __u16 act_mask; ++ __u32 buf_size; ++ __u32 buf_nr; ++ __u64 start_lba; ++ __u64 end_lba; ++ __u32 pid; ++}; ++ ++enum blk_crypto_mode_num { ++ BLK_ENCRYPTION_MODE_INVALID = 0, ++ BLK_ENCRYPTION_MODE_AES_256_XTS = 1, ++ BLK_ENCRYPTION_MODE_AES_128_CBC_ESSIV = 2, ++ BLK_ENCRYPTION_MODE_ADIANTUM = 3, ++ BLK_ENCRYPTION_MODE_SM4_XTS = 4, ++ BLK_ENCRYPTION_MODE_MAX = 5, ++}; ++ ++struct blk_crypto_config { ++ enum blk_crypto_mode_num crypto_mode; ++ unsigned int data_unit_size; ++ unsigned int dun_bytes; ++}; ++ ++struct blk_crypto_key { ++ struct blk_crypto_config crypto_cfg; ++ unsigned int data_unit_size_bits; ++ unsigned int size; ++ u8 raw[64]; ++}; ++ ++typedef void blk_log_action_t(struct trace_iterator *, const char *, bool); ++ ++struct syscall_trace_enter { ++ struct trace_entry ent; ++ int nr; ++ long unsigned int args[0]; ++}; ++ ++struct syscall_trace_exit { ++ struct trace_entry ent; ++ int nr; ++ long int ret; ++}; ++ ++struct syscall_tp_t { ++ long long unsigned int regs; ++ long unsigned int syscall_nr; ++ long unsigned int ret; ++}; ++ ++struct syscall_tp_t___2 { ++ long long unsigned int regs; ++ long unsigned int syscall_nr; ++ long unsigned int args[6]; ++}; ++ ++struct ftrace_event_field { ++ struct list_head link; ++ const char *name; ++ const char *type; ++ int filter_type; ++ int offset; ++ int size; ++ int is_signed; ++ int len; ++}; ++ ++struct filter_pred; ++ ++struct prog_entry { ++ int target; ++ int when_to_branch; ++ struct filter_pred *pred; ++}; ++ ++struct regex; ++ ++typedef int (*regex_match_func)(char *, struct regex *, int); ++ ++struct regex { ++ char pattern[256]; ++ int len; ++ int field_len; ++ regex_match_func match; ++}; ++ ++enum regex_type { ++ MATCH_FULL = 0, ++ MATCH_FRONT_ONLY = 1, ++ MATCH_MIDDLE_ONLY = 2, ++ MATCH_END_ONLY = 3, ++ MATCH_GLOB = 4, ++ MATCH_INDEX = 5, ++}; ++ ++enum filter_op_ids { ++ OP_GLOB = 0, ++ OP_NE = 1, ++ OP_EQ = 2, ++ OP_LE = 3, ++ OP_LT = 4, ++ OP_GE = 5, ++ OP_GT = 6, ++ OP_BAND = 7, ++ OP_MAX = 8, ++}; ++ ++enum filter_pred_fn { ++ FILTER_PRED_FN_NOP = 0, ++ FILTER_PRED_FN_64 = 1, ++ FILTER_PRED_FN_S64 = 2, ++ FILTER_PRED_FN_U64 = 3, ++ FILTER_PRED_FN_32 = 4, ++ FILTER_PRED_FN_S32 = 5, ++ FILTER_PRED_FN_U32 = 6, ++ FILTER_PRED_FN_16 = 7, ++ FILTER_PRED_FN_S16 = 8, ++ FILTER_PRED_FN_U16 = 9, ++ FILTER_PRED_FN_8 = 10, ++ FILTER_PRED_FN_S8 = 11, ++ FILTER_PRED_FN_U8 = 12, ++ FILTER_PRED_FN_COMM = 13, ++ FILTER_PRED_FN_STRING = 14, ++ FILTER_PRED_FN_STRLOC = 15, ++ FILTER_PRED_FN_STRRELLOC = 16, ++ FILTER_PRED_FN_PCHAR_USER = 17, ++ FILTER_PRED_FN_PCHAR = 18, ++ FILTER_PRED_FN_CPU = 19, ++ FILTER_PRED_FN_FUNCTION = 20, ++ FILTER_PRED_FN_ = 21, ++ FILTER_PRED_TEST_VISITED = 22, ++}; ++ ++struct filter_pred { ++ enum filter_pred_fn fn_num; ++ u64 val; ++ u64 val2; ++ struct regex regex; ++ short unsigned int *ops; ++ struct ftrace_event_field *field; ++ int offset; ++ int not; ++ int op; ++}; ++ ++enum { ++ FILT_ERR_NONE = 0, ++ FILT_ERR_INVALID_OP = 1, ++ FILT_ERR_TOO_MANY_OPEN = 2, ++ FILT_ERR_TOO_MANY_CLOSE = 3, ++ FILT_ERR_MISSING_QUOTE = 4, ++ FILT_ERR_OPERAND_TOO_LONG = 5, ++ FILT_ERR_EXPECT_STRING = 6, ++ FILT_ERR_EXPECT_DIGIT = 7, ++ FILT_ERR_ILLEGAL_FIELD_OP = 8, ++ FILT_ERR_FIELD_NOT_FOUND = 9, ++ FILT_ERR_ILLEGAL_INTVAL = 10, ++ FILT_ERR_BAD_SUBSYS_FILTER = 11, ++ FILT_ERR_TOO_MANY_PREDS = 12, ++ FILT_ERR_INVALID_FILTER = 13, ++ FILT_ERR_IP_FIELD_ONLY = 14, ++ FILT_ERR_INVALID_VALUE = 15, ++ FILT_ERR_NO_FUNCTION = 16, ++ FILT_ERR_ERRNO = 17, ++ FILT_ERR_NO_FILTER = 18, ++}; ++ ++struct filter_parse_error { ++ int lasterr; ++ int lasterr_pos; ++}; ++ ++typedef int (*parse_pred_fn)(const char *, void *, int, struct filter_parse_error *, struct filter_pred **); ++ ++enum { ++ INVERT = 1, ++ PROCESS_AND = 2, ++ PROCESS_OR = 4, ++}; ++ ++struct ustring_buffer { ++ char buffer[1024]; ++}; ++ ++enum { ++ TOO_MANY_CLOSE = -1, ++ TOO_MANY_OPEN = -2, ++ MISSING_QUOTE = -3, ++}; ++ ++struct filter_list { ++ struct list_head list; ++ struct event_filter *filter; ++}; ++ ++struct function_filter_data { ++ struct ftrace_ops *ops; ++ int first_filter; ++ int first_notrace; ++}; ++ ++struct dyn_event_operations; ++ ++struct dyn_event { ++ struct list_head list; ++ struct dyn_event_operations *ops; ++}; ++ ++struct synth_field; ++ ++struct synth_event { ++ struct dyn_event devent; ++ int ref; ++ char *name; ++ struct synth_field **fields; ++ unsigned int n_fields; ++ struct synth_field **dynamic_fields; ++ unsigned int n_dynamic_fields; ++ unsigned int n_u64; ++ struct trace_event_class class; ++ struct trace_event_call call; ++ struct tracepoint *tp; ++ struct module *mod; ++}; ++ ++typedef int (*tracing_map_cmp_fn_t)(void *, void *); ++ ++struct tracing_map_field { ++ tracing_map_cmp_fn_t cmp_fn; ++ union { ++ atomic64_t sum; ++ unsigned int offset; ++ }; ++}; ++ ++struct tracing_map; ++ ++struct tracing_map_elt { ++ struct tracing_map *map; ++ struct tracing_map_field *fields; ++ atomic64_t *vars; ++ bool *var_set; ++ void *key; ++ void *private_data; ++}; ++ ++struct tracing_map_sort_key { ++ unsigned int field_idx; ++ bool descending; ++}; ++ ++struct tracing_map_array; ++ ++struct tracing_map_ops; ++ ++struct tracing_map { ++ unsigned int key_size; ++ unsigned int map_bits; ++ unsigned int map_size; ++ unsigned int max_elts; ++ atomic_t next_elt; ++ struct tracing_map_array *elts; ++ struct tracing_map_array *map; ++ const struct tracing_map_ops *ops; ++ void *private_data; ++ struct tracing_map_field fields[6]; ++ unsigned int n_fields; ++ int key_idx[3]; ++ unsigned int n_keys; ++ struct tracing_map_sort_key sort_key; ++ unsigned int n_vars; ++ atomic64_t hits; ++ atomic64_t drops; ++}; ++ ++struct tracing_map_sort_entry { ++ void *key; ++ struct tracing_map_elt *elt; ++ bool elt_copied; ++ bool dup; ++}; ++ ++struct tracing_map_array { ++ unsigned int entries_per_page; ++ unsigned int entry_size_shift; ++ unsigned int entry_shift; ++ unsigned int entry_mask; ++ unsigned int n_pages; ++ void **pages; ++}; ++ ++struct tracing_map_ops { ++ int (*elt_alloc)(struct tracing_map_elt *); ++ void (*elt_free)(struct tracing_map_elt *); ++ void (*elt_clear)(struct tracing_map_elt *); ++ void (*elt_init)(struct tracing_map_elt *); ++}; ++ ++struct event_trigger_ops; ++ ++struct event_command; ++ ++struct event_trigger_data { ++ long unsigned int count; ++ int ref; ++ int flags; ++ struct event_trigger_ops *ops; ++ struct event_command *cmd_ops; ++ struct event_filter *filter; ++ char *filter_str; ++ void *private_data; ++ bool paused; ++ bool paused_tmp; ++ struct list_head list; ++ char *name; ++ struct list_head named_list; ++ struct event_trigger_data *named_data; ++}; ++ ++struct event_trigger_ops { ++ void (*trigger)(struct event_trigger_data *, struct trace_buffer *, void *, struct ring_buffer_event *); ++ int (*init)(struct event_trigger_data *); ++ void (*free)(struct event_trigger_data *); ++ int (*print)(struct seq_file *, struct event_trigger_data *); ++}; ++ ++struct event_command { ++ struct list_head list; ++ char *name; ++ enum event_trigger_type trigger_type; ++ int flags; ++ int (*parse)(struct event_command *, struct trace_event_file *, char *, char *, char *); ++ int (*reg)(char *, struct event_trigger_data *, struct trace_event_file *); ++ void (*unreg)(char *, struct event_trigger_data *, struct trace_event_file *); ++ void (*unreg_all)(struct trace_event_file *); ++ int (*set_filter)(char *, struct event_trigger_data *, struct trace_event_file *); ++ struct event_trigger_ops * (*get_trigger_ops)(char *, char *); ++}; ++ ++struct enable_trigger_data { ++ struct trace_event_file *file; ++ bool enable; ++ bool hist; ++}; ++ ++enum event_command_flags { ++ EVENT_CMD_FL_POST_TRIGGER = 1, ++ EVENT_CMD_FL_NEEDS_REC = 2, ++}; ++ ++struct dyn_event_operations { ++ struct list_head list; ++ int (*create)(const char *); ++ int (*show)(struct seq_file *, struct dyn_event *); ++ bool (*is_busy)(struct dyn_event *); ++ int (*free)(struct dyn_event *); ++ bool (*match)(const char *, const char *, int, const char **, struct dyn_event *); ++}; ++ ++struct synth_field { ++ char *type; ++ char *name; ++ size_t size; ++ unsigned int offset; ++ unsigned int field_pos; ++ bool is_signed; ++ bool is_string; ++ bool is_dynamic; ++ bool is_stack; ++}; ++ ++enum { ++ HIST_ERR_NONE = 0, ++ HIST_ERR_DUPLICATE_VAR = 1, ++ HIST_ERR_VAR_NOT_UNIQUE = 2, ++ HIST_ERR_TOO_MANY_VARS = 3, ++ HIST_ERR_MALFORMED_ASSIGNMENT = 4, ++ HIST_ERR_NAMED_MISMATCH = 5, ++ HIST_ERR_TRIGGER_EEXIST = 6, ++ HIST_ERR_TRIGGER_ENOENT_CLEAR = 7, ++ HIST_ERR_SET_CLOCK_FAIL = 8, ++ HIST_ERR_BAD_FIELD_MODIFIER = 9, ++ HIST_ERR_TOO_MANY_SUBEXPR = 10, ++ HIST_ERR_TIMESTAMP_MISMATCH = 11, ++ HIST_ERR_TOO_MANY_FIELD_VARS = 12, ++ HIST_ERR_EVENT_FILE_NOT_FOUND = 13, ++ HIST_ERR_HIST_NOT_FOUND = 14, ++ HIST_ERR_HIST_CREATE_FAIL = 15, ++ HIST_ERR_SYNTH_VAR_NOT_FOUND = 16, ++ HIST_ERR_SYNTH_EVENT_NOT_FOUND = 17, ++ HIST_ERR_SYNTH_TYPE_MISMATCH = 18, ++ HIST_ERR_SYNTH_COUNT_MISMATCH = 19, ++ HIST_ERR_FIELD_VAR_PARSE_FAIL = 20, ++ HIST_ERR_VAR_CREATE_FIND_FAIL = 21, ++ HIST_ERR_ONX_NOT_VAR = 22, ++ HIST_ERR_ONX_VAR_NOT_FOUND = 23, ++ HIST_ERR_ONX_VAR_CREATE_FAIL = 24, ++ HIST_ERR_FIELD_VAR_CREATE_FAIL = 25, ++ HIST_ERR_TOO_MANY_PARAMS = 26, ++ HIST_ERR_PARAM_NOT_FOUND = 27, ++ HIST_ERR_INVALID_PARAM = 28, ++ HIST_ERR_ACTION_NOT_FOUND = 29, ++ HIST_ERR_NO_SAVE_PARAMS = 30, ++ HIST_ERR_TOO_MANY_SAVE_ACTIONS = 31, ++ HIST_ERR_ACTION_MISMATCH = 32, ++ HIST_ERR_NO_CLOSING_PAREN = 33, ++ HIST_ERR_SUBSYS_NOT_FOUND = 34, ++ HIST_ERR_INVALID_SUBSYS_EVENT = 35, ++ HIST_ERR_INVALID_REF_KEY = 36, ++ HIST_ERR_VAR_NOT_FOUND = 37, ++ HIST_ERR_FIELD_NOT_FOUND = 38, ++ HIST_ERR_EMPTY_ASSIGNMENT = 39, ++ HIST_ERR_INVALID_SORT_MODIFIER = 40, ++ HIST_ERR_EMPTY_SORT_FIELD = 41, ++ HIST_ERR_TOO_MANY_SORT_FIELDS = 42, ++ HIST_ERR_INVALID_SORT_FIELD = 43, ++ HIST_ERR_INVALID_STR_OPERAND = 44, ++ HIST_ERR_EXPECT_NUMBER = 45, ++ HIST_ERR_UNARY_MINUS_SUBEXPR = 46, ++ HIST_ERR_DIVISION_BY_ZERO = 47, ++ HIST_ERR_NEED_NOHC_VAL = 48, ++}; ++ ++enum hist_field_fn { ++ HIST_FIELD_FN_NOP = 0, ++ HIST_FIELD_FN_VAR_REF = 1, ++ HIST_FIELD_FN_COUNTER = 2, ++ HIST_FIELD_FN_CONST = 3, ++ HIST_FIELD_FN_LOG2 = 4, ++ HIST_FIELD_FN_BUCKET = 5, ++ HIST_FIELD_FN_TIMESTAMP = 6, ++ HIST_FIELD_FN_CPU = 7, ++ HIST_FIELD_FN_STRING = 8, ++ HIST_FIELD_FN_DYNSTRING = 9, ++ HIST_FIELD_FN_RELDYNSTRING = 10, ++ HIST_FIELD_FN_PSTRING = 11, ++ HIST_FIELD_FN_S64 = 12, ++ HIST_FIELD_FN_U64 = 13, ++ HIST_FIELD_FN_S32 = 14, ++ HIST_FIELD_FN_U32 = 15, ++ HIST_FIELD_FN_S16 = 16, ++ HIST_FIELD_FN_U16 = 17, ++ HIST_FIELD_FN_S8 = 18, ++ HIST_FIELD_FN_U8 = 19, ++ HIST_FIELD_FN_UMINUS = 20, ++ HIST_FIELD_FN_MINUS = 21, ++ HIST_FIELD_FN_PLUS = 22, ++ HIST_FIELD_FN_DIV = 23, ++ HIST_FIELD_FN_MULT = 24, ++ HIST_FIELD_FN_DIV_POWER2 = 25, ++ HIST_FIELD_FN_DIV_NOT_POWER2 = 26, ++ HIST_FIELD_FN_DIV_MULT_SHIFT = 27, ++ HIST_FIELD_FN_EXECNAME = 28, ++ HIST_FIELD_FN_STACK = 29, ++}; ++ ++struct hist_trigger_data; ++ ++struct hist_var { ++ char *name; ++ struct hist_trigger_data *hist_data; ++ unsigned int idx; ++}; ++ ++enum field_op_id { ++ FIELD_OP_NONE = 0, ++ FIELD_OP_PLUS = 1, ++ FIELD_OP_MINUS = 2, ++ FIELD_OP_UNARY_MINUS = 3, ++ FIELD_OP_DIV = 4, ++ FIELD_OP_MULT = 5, ++}; ++ ++struct hist_field { ++ struct ftrace_event_field *field; ++ long unsigned int flags; ++ long unsigned int buckets; ++ const char *type; ++ struct hist_field *operands[2]; ++ struct hist_trigger_data *hist_data; ++ enum hist_field_fn fn_num; ++ unsigned int ref; ++ unsigned int size; ++ unsigned int offset; ++ unsigned int is_signed; ++ struct hist_var var; ++ enum field_op_id operator; ++ char *system; ++ char *event_name; ++ char *name; ++ unsigned int var_ref_idx; ++ bool read_once; ++ unsigned int var_str_idx; ++ u64 constant; ++ u64 div_multiplier; ++}; ++ ++struct hist_trigger_attrs; ++ ++struct action_data; ++ ++struct field_var; ++ ++struct field_var_hist; ++ ++struct hist_trigger_data { ++ struct hist_field *fields[22]; ++ unsigned int n_vals; ++ unsigned int n_keys; ++ unsigned int n_fields; ++ unsigned int n_vars; ++ unsigned int n_var_str; ++ unsigned int key_size; ++ struct tracing_map_sort_key sort_keys[2]; ++ unsigned int n_sort_keys; ++ struct trace_event_file *event_file; ++ struct hist_trigger_attrs *attrs; ++ struct tracing_map *map; ++ bool enable_timestamps; ++ bool remove; ++ struct hist_field *var_refs[16]; ++ unsigned int n_var_refs; ++ struct action_data *actions[8]; ++ unsigned int n_actions; ++ struct field_var *field_vars[64]; ++ unsigned int n_field_vars; ++ unsigned int n_field_var_str; ++ struct field_var_hist *field_var_hists[64]; ++ unsigned int n_field_var_hists; ++ struct field_var *save_vars[64]; ++ unsigned int n_save_vars; ++ unsigned int n_save_var_str; ++}; ++ ++enum hist_field_flags { ++ HIST_FIELD_FL_HITCOUNT = 1, ++ HIST_FIELD_FL_KEY = 2, ++ HIST_FIELD_FL_STRING = 4, ++ HIST_FIELD_FL_HEX = 8, ++ HIST_FIELD_FL_SYM = 16, ++ HIST_FIELD_FL_SYM_OFFSET = 32, ++ HIST_FIELD_FL_EXECNAME = 64, ++ HIST_FIELD_FL_SYSCALL = 128, ++ HIST_FIELD_FL_STACKTRACE = 256, ++ HIST_FIELD_FL_LOG2 = 512, ++ HIST_FIELD_FL_TIMESTAMP = 1024, ++ HIST_FIELD_FL_TIMESTAMP_USECS = 2048, ++ HIST_FIELD_FL_VAR = 4096, ++ HIST_FIELD_FL_EXPR = 8192, ++ HIST_FIELD_FL_VAR_REF = 16384, ++ HIST_FIELD_FL_CPU = 32768, ++ HIST_FIELD_FL_ALIAS = 65536, ++ HIST_FIELD_FL_BUCKET = 131072, ++ HIST_FIELD_FL_CONST = 262144, ++ HIST_FIELD_FL_PERCENT = 524288, ++ HIST_FIELD_FL_GRAPH = 1048576, ++}; ++ ++struct var_defs { ++ unsigned int n_vars; ++ char *name[16]; ++ char *expr[16]; ++}; ++ ++struct hist_trigger_attrs { ++ char *keys_str; ++ char *vals_str; ++ char *sort_key_str; ++ char *name; ++ char *clock; ++ bool pause; ++ bool cont; ++ bool clear; ++ bool ts_in_usecs; ++ bool no_hitcount; ++ unsigned int map_bits; ++ char *assignment_str[16]; ++ unsigned int n_assignments; ++ char *action_str[8]; ++ unsigned int n_actions; ++ struct var_defs var_defs; ++}; ++ ++struct field_var { ++ struct hist_field *var; ++ struct hist_field *val; ++}; ++ ++struct field_var_hist { ++ struct hist_trigger_data *hist_data; ++ char *cmd; ++}; ++ ++enum handler_id { ++ HANDLER_ONMATCH = 1, ++ HANDLER_ONMAX = 2, ++ HANDLER_ONCHANGE = 3, ++}; ++ ++enum action_id { ++ ACTION_SAVE = 1, ++ ACTION_TRACE = 2, ++ ACTION_SNAPSHOT = 3, ++}; ++ ++typedef void (*action_fn_t)(struct hist_trigger_data *, struct tracing_map_elt *, struct trace_buffer *, void *, struct ring_buffer_event *, void *, struct action_data *, u64 *); ++ ++typedef bool (*check_track_val_fn_t)(u64, u64); ++ ++struct action_data { ++ enum handler_id handler; ++ enum action_id action; ++ char *action_name; ++ action_fn_t fn; ++ unsigned int n_params; ++ char *params[64]; ++ unsigned int var_ref_idx[64]; ++ struct synth_event *synth_event; ++ bool use_trace_keyword; ++ char *synth_event_name; ++ union { ++ struct { ++ char *event; ++ char *event_system; ++ } match_data; ++ struct { ++ char *var_str; ++ struct hist_field *var_ref; ++ struct hist_field *track_var; ++ check_track_val_fn_t check_val; ++ action_fn_t save_data; ++ } track_data; ++ }; ++}; ++ ++struct track_data { ++ u64 track_val; ++ bool updated; ++ unsigned int key_len; ++ void *key; ++ struct tracing_map_elt elt; ++ struct action_data *action_data; ++ struct hist_trigger_data *hist_data; ++}; ++ ++struct hist_elt_data { ++ char *comm; ++ u64 *var_ref_vals; ++ char **field_var_str; ++ int n_field_var_str; ++}; ++ ++struct snapshot_context { ++ struct tracing_map_elt *elt; ++ void *key; ++}; ++ ++typedef void (*synth_probe_func_t)(void *, u64 *, unsigned int *); ++ ++struct hist_var_data { ++ struct list_head list; ++ struct hist_trigger_data *hist_data; ++}; ++ ++struct hist_val_stat { ++ u64 max; ++ u64 total; ++}; ++ ++struct trace_event_raw_error_report_template { ++ struct trace_entry ent; ++ enum error_detector error_detector; ++ long unsigned int id; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_error_report_template {}; ++ ++typedef void (*btf_trace_error_report_end)(void *, enum error_detector, long unsigned int); ++ ++enum dynevent_type { ++ DYNEVENT_TYPE_SYNTH = 1, ++ DYNEVENT_TYPE_KPROBE = 2, ++ DYNEVENT_TYPE_NONE = 3, ++}; ++ ++struct dynevent_cmd; ++ ++typedef int (*dynevent_create_fn_t)(struct dynevent_cmd *); ++ ++struct dynevent_cmd { ++ struct seq_buf seq; ++ const char *event_name; ++ unsigned int n_fields; ++ enum dynevent_type type; ++ dynevent_create_fn_t run_command; ++ void *private_data; ++}; ++ ++typedef int (*dynevent_check_arg_fn_t)(void *); ++ ++struct dynevent_arg { ++ const char *str; ++ char separator; ++}; ++ ++struct dynevent_arg_pair { ++ const char *lhs; ++ const char *rhs; ++ char operator; ++ char separator; ++}; ++ ++typedef u16 uint16_t; ++ ++struct xbc_node { ++ uint16_t next; ++ uint16_t child; ++ uint16_t parent; ++ uint16_t data; ++}; ++ ++typedef void (*exitcall_t)(); ++ ++struct rv_reactor { ++ const char *name; ++ const char *description; ++ void (*react)(char *); ++}; ++ ++typedef void (*swap_func_t)(void *, void *, int); ++ ++typedef int (*cmp_func_t)(const void *, const void *); ++ ++struct btf_type { ++ __u32 name_off; ++ __u32 info; ++ union { ++ __u32 size; ++ __u32 type; ++ }; ++}; ++ ++enum { ++ BTF_KIND_UNKN = 0, ++ BTF_KIND_INT = 1, ++ BTF_KIND_PTR = 2, ++ BTF_KIND_ARRAY = 3, ++ BTF_KIND_STRUCT = 4, ++ BTF_KIND_UNION = 5, ++ BTF_KIND_ENUM = 6, ++ BTF_KIND_FWD = 7, ++ BTF_KIND_TYPEDEF = 8, ++ BTF_KIND_VOLATILE = 9, ++ BTF_KIND_CONST = 10, ++ BTF_KIND_RESTRICT = 11, ++ BTF_KIND_FUNC = 12, ++ BTF_KIND_FUNC_PROTO = 13, ++ BTF_KIND_VAR = 14, ++ BTF_KIND_DATASEC = 15, ++ BTF_KIND_FLOAT = 16, ++ BTF_KIND_DECL_TAG = 17, ++ BTF_KIND_TYPE_TAG = 18, ++ BTF_KIND_ENUM64 = 19, ++ NR_BTF_KINDS = 20, ++ BTF_KIND_MAX = 19, ++}; ++ ++struct btf_array { ++ __u32 type; ++ __u32 index_type; ++ __u32 nelems; ++}; ++ ++struct btf_member { ++ __u32 name_off; ++ __u32 type; ++ __u32 offset; ++}; ++ ++struct btf_param { ++ __u32 name_off; ++ __u32 type; ++}; ++ ++enum btf_func_linkage { ++ BTF_FUNC_STATIC = 0, ++ BTF_FUNC_GLOBAL = 1, ++ BTF_FUNC_EXTERN = 2, ++}; ++ ++struct btf_var_secinfo { ++ __u32 type; ++ __u32 offset; ++ __u32 size; ++}; ++ ++enum { ++ BPF_REG_0 = 0, ++ BPF_REG_1 = 1, ++ BPF_REG_2 = 2, ++ BPF_REG_3 = 3, ++ BPF_REG_4 = 4, ++ BPF_REG_5 = 5, ++ BPF_REG_6 = 6, ++ BPF_REG_7 = 7, ++ BPF_REG_8 = 8, ++ BPF_REG_9 = 9, ++ BPF_REG_10 = 10, ++ __MAX_BPF_REG = 11, ++}; ++ ++struct bpf_insn { ++ __u8 code; ++ __u8 dst_reg: 4; ++ __u8 src_reg: 4; ++ __s16 off; ++ __s32 imm; ++}; ++ ++struct bpf_cgroup_storage_key { ++ __u64 cgroup_inode_id; ++ __u32 attach_type; ++}; ++ ++enum bpf_cgroup_iter_order { ++ BPF_CGROUP_ITER_ORDER_UNSPEC = 0, ++ BPF_CGROUP_ITER_SELF_ONLY = 1, ++ BPF_CGROUP_ITER_DESCENDANTS_PRE = 2, ++ BPF_CGROUP_ITER_DESCENDANTS_POST = 3, ++ BPF_CGROUP_ITER_ANCESTORS_UP = 4, ++}; ++ ++enum bpf_map_type { ++ BPF_MAP_TYPE_UNSPEC = 0, ++ BPF_MAP_TYPE_HASH = 1, ++ BPF_MAP_TYPE_ARRAY = 2, ++ BPF_MAP_TYPE_PROG_ARRAY = 3, ++ BPF_MAP_TYPE_PERF_EVENT_ARRAY = 4, ++ BPF_MAP_TYPE_PERCPU_HASH = 5, ++ BPF_MAP_TYPE_PERCPU_ARRAY = 6, ++ BPF_MAP_TYPE_STACK_TRACE = 7, ++ BPF_MAP_TYPE_CGROUP_ARRAY = 8, ++ BPF_MAP_TYPE_LRU_HASH = 9, ++ BPF_MAP_TYPE_LRU_PERCPU_HASH = 10, ++ BPF_MAP_TYPE_LPM_TRIE = 11, ++ BPF_MAP_TYPE_ARRAY_OF_MAPS = 12, ++ BPF_MAP_TYPE_HASH_OF_MAPS = 13, ++ BPF_MAP_TYPE_DEVMAP = 14, ++ BPF_MAP_TYPE_SOCKMAP = 15, ++ BPF_MAP_TYPE_CPUMAP = 16, ++ BPF_MAP_TYPE_XSKMAP = 17, ++ BPF_MAP_TYPE_SOCKHASH = 18, ++ BPF_MAP_TYPE_CGROUP_STORAGE_DEPRECATED = 19, ++ BPF_MAP_TYPE_CGROUP_STORAGE = 19, ++ BPF_MAP_TYPE_REUSEPORT_SOCKARRAY = 20, ++ BPF_MAP_TYPE_PERCPU_CGROUP_STORAGE = 21, ++ BPF_MAP_TYPE_QUEUE = 22, ++ BPF_MAP_TYPE_STACK = 23, ++ BPF_MAP_TYPE_SK_STORAGE = 24, ++ BPF_MAP_TYPE_DEVMAP_HASH = 25, ++ BPF_MAP_TYPE_STRUCT_OPS = 26, ++ BPF_MAP_TYPE_RINGBUF = 27, ++ BPF_MAP_TYPE_INODE_STORAGE = 28, ++ BPF_MAP_TYPE_TASK_STORAGE = 29, ++ BPF_MAP_TYPE_BLOOM_FILTER = 30, ++ BPF_MAP_TYPE_USER_RINGBUF = 31, ++ BPF_MAP_TYPE_CGRP_STORAGE = 32, ++}; ++ ++enum bpf_prog_type { ++ BPF_PROG_TYPE_UNSPEC = 0, ++ BPF_PROG_TYPE_SOCKET_FILTER = 1, ++ BPF_PROG_TYPE_KPROBE = 2, ++ BPF_PROG_TYPE_SCHED_CLS = 3, ++ BPF_PROG_TYPE_SCHED_ACT = 4, ++ BPF_PROG_TYPE_TRACEPOINT = 5, ++ BPF_PROG_TYPE_XDP = 6, ++ BPF_PROG_TYPE_PERF_EVENT = 7, ++ BPF_PROG_TYPE_CGROUP_SKB = 8, ++ BPF_PROG_TYPE_CGROUP_SOCK = 9, ++ BPF_PROG_TYPE_LWT_IN = 10, ++ BPF_PROG_TYPE_LWT_OUT = 11, ++ BPF_PROG_TYPE_LWT_XMIT = 12, ++ BPF_PROG_TYPE_SOCK_OPS = 13, ++ BPF_PROG_TYPE_SK_SKB = 14, ++ BPF_PROG_TYPE_CGROUP_DEVICE = 15, ++ BPF_PROG_TYPE_SK_MSG = 16, ++ BPF_PROG_TYPE_RAW_TRACEPOINT = 17, ++ BPF_PROG_TYPE_CGROUP_SOCK_ADDR = 18, ++ BPF_PROG_TYPE_LWT_SEG6LOCAL = 19, ++ BPF_PROG_TYPE_LIRC_MODE2 = 20, ++ BPF_PROG_TYPE_SK_REUSEPORT = 21, ++ BPF_PROG_TYPE_FLOW_DISSECTOR = 22, ++ BPF_PROG_TYPE_CGROUP_SYSCTL = 23, ++ BPF_PROG_TYPE_RAW_TRACEPOINT_WRITABLE = 24, ++ BPF_PROG_TYPE_CGROUP_SOCKOPT = 25, ++ BPF_PROG_TYPE_TRACING = 26, ++ BPF_PROG_TYPE_STRUCT_OPS = 27, ++ BPF_PROG_TYPE_EXT = 28, ++ BPF_PROG_TYPE_LSM = 29, ++ BPF_PROG_TYPE_SK_LOOKUP = 30, ++ BPF_PROG_TYPE_SYSCALL = 31, ++ BPF_PROG_TYPE_NETFILTER = 32, ++}; ++ ++enum bpf_attach_type { ++ BPF_CGROUP_INET_INGRESS = 0, ++ BPF_CGROUP_INET_EGRESS = 1, ++ BPF_CGROUP_INET_SOCK_CREATE = 2, ++ BPF_CGROUP_SOCK_OPS = 3, ++ BPF_SK_SKB_STREAM_PARSER = 4, ++ BPF_SK_SKB_STREAM_VERDICT = 5, ++ BPF_CGROUP_DEVICE = 6, ++ BPF_SK_MSG_VERDICT = 7, ++ BPF_CGROUP_INET4_BIND = 8, ++ BPF_CGROUP_INET6_BIND = 9, ++ BPF_CGROUP_INET4_CONNECT = 10, ++ BPF_CGROUP_INET6_CONNECT = 11, ++ BPF_CGROUP_INET4_POST_BIND = 12, ++ BPF_CGROUP_INET6_POST_BIND = 13, ++ BPF_CGROUP_UDP4_SENDMSG = 14, ++ BPF_CGROUP_UDP6_SENDMSG = 15, ++ BPF_LIRC_MODE2 = 16, ++ BPF_FLOW_DISSECTOR = 17, ++ BPF_CGROUP_SYSCTL = 18, ++ BPF_CGROUP_UDP4_RECVMSG = 19, ++ BPF_CGROUP_UDP6_RECVMSG = 20, ++ BPF_CGROUP_GETSOCKOPT = 21, ++ BPF_CGROUP_SETSOCKOPT = 22, ++ BPF_TRACE_RAW_TP = 23, ++ BPF_TRACE_FENTRY = 24, ++ BPF_TRACE_FEXIT = 25, ++ BPF_MODIFY_RETURN = 26, ++ BPF_LSM_MAC = 27, ++ BPF_TRACE_ITER = 28, ++ BPF_CGROUP_INET4_GETPEERNAME = 29, ++ BPF_CGROUP_INET6_GETPEERNAME = 30, ++ BPF_CGROUP_INET4_GETSOCKNAME = 31, ++ BPF_CGROUP_INET6_GETSOCKNAME = 32, ++ BPF_XDP_DEVMAP = 33, ++ BPF_CGROUP_INET_SOCK_RELEASE = 34, ++ BPF_XDP_CPUMAP = 35, ++ BPF_SK_LOOKUP = 36, ++ BPF_XDP = 37, ++ BPF_SK_SKB_VERDICT = 38, ++ BPF_SK_REUSEPORT_SELECT = 39, ++ BPF_SK_REUSEPORT_SELECT_OR_MIGRATE = 40, ++ BPF_PERF_EVENT = 41, ++ BPF_TRACE_KPROBE_MULTI = 42, ++ BPF_LSM_CGROUP = 43, ++ BPF_STRUCT_OPS = 44, ++ BPF_NETFILTER = 45, ++ __MAX_BPF_ATTACH_TYPE = 46, ++}; ++ ++enum { ++ BPF_F_NO_PREALLOC = 1, ++ BPF_F_NO_COMMON_LRU = 2, ++ BPF_F_NUMA_NODE = 4, ++ BPF_F_RDONLY = 8, ++ BPF_F_WRONLY = 16, ++ BPF_F_STACK_BUILD_ID = 32, ++ BPF_F_ZERO_SEED = 64, ++ BPF_F_RDONLY_PROG = 128, ++ BPF_F_WRONLY_PROG = 256, ++ BPF_F_CLONE = 512, ++ BPF_F_MMAPABLE = 1024, ++ BPF_F_PRESERVE_ELEMS = 2048, ++ BPF_F_INNER_MAP = 4096, ++ BPF_F_LINK = 8192, ++}; ++ ++union bpf_attr { ++ struct { ++ __u32 map_type; ++ __u32 key_size; ++ __u32 value_size; ++ __u32 max_entries; ++ __u32 map_flags; ++ __u32 inner_map_fd; ++ __u32 numa_node; ++ char map_name[16]; ++ __u32 map_ifindex; ++ __u32 btf_fd; ++ __u32 btf_key_type_id; ++ __u32 btf_value_type_id; ++ __u32 btf_vmlinux_value_type_id; ++ __u64 map_extra; ++ }; ++ struct { ++ __u32 map_fd; ++ __u64 key; ++ union { ++ __u64 value; ++ __u64 next_key; ++ }; ++ __u64 flags; ++ }; ++ struct { ++ __u64 in_batch; ++ __u64 out_batch; ++ __u64 keys; ++ __u64 values; ++ __u32 count; ++ __u32 map_fd; ++ __u64 elem_flags; ++ __u64 flags; ++ } batch; ++ struct { ++ __u32 prog_type; ++ __u32 insn_cnt; ++ __u64 insns; ++ __u64 license; ++ __u32 log_level; ++ __u32 log_size; ++ __u64 log_buf; ++ __u32 kern_version; ++ __u32 prog_flags; ++ char prog_name[16]; ++ __u32 prog_ifindex; ++ __u32 expected_attach_type; ++ __u32 prog_btf_fd; ++ __u32 func_info_rec_size; ++ __u64 func_info; ++ __u32 func_info_cnt; ++ __u32 line_info_rec_size; ++ __u64 line_info; ++ __u32 line_info_cnt; ++ __u32 attach_btf_id; ++ union { ++ __u32 attach_prog_fd; ++ __u32 attach_btf_obj_fd; ++ }; ++ __u32 core_relo_cnt; ++ __u64 fd_array; ++ __u64 core_relos; ++ __u32 core_relo_rec_size; ++ __u32 log_true_size; ++ }; ++ struct { ++ __u64 pathname; ++ __u32 bpf_fd; ++ __u32 file_flags; ++ }; ++ struct { ++ __u32 target_fd; ++ __u32 attach_bpf_fd; ++ __u32 attach_type; ++ __u32 attach_flags; ++ __u32 replace_bpf_fd; ++ }; ++ struct { ++ __u32 prog_fd; ++ __u32 retval; ++ __u32 data_size_in; ++ __u32 data_size_out; ++ __u64 data_in; ++ __u64 data_out; ++ __u32 repeat; ++ __u32 duration; ++ __u32 ctx_size_in; ++ __u32 ctx_size_out; ++ __u64 ctx_in; ++ __u64 ctx_out; ++ __u32 flags; ++ __u32 cpu; ++ __u32 batch_size; ++ } test; ++ struct { ++ union { ++ __u32 start_id; ++ __u32 prog_id; ++ __u32 map_id; ++ __u32 btf_id; ++ __u32 link_id; ++ }; ++ __u32 next_id; ++ __u32 open_flags; ++ }; ++ struct { ++ __u32 bpf_fd; ++ __u32 info_len; ++ __u64 info; ++ } info; ++ struct { ++ __u32 target_fd; ++ __u32 attach_type; ++ __u32 query_flags; ++ __u32 attach_flags; ++ __u64 prog_ids; ++ __u32 prog_cnt; ++ __u64 prog_attach_flags; ++ } query; ++ struct { ++ __u64 name; ++ __u32 prog_fd; ++ } raw_tracepoint; ++ struct { ++ __u64 btf; ++ __u64 btf_log_buf; ++ __u32 btf_size; ++ __u32 btf_log_size; ++ __u32 btf_log_level; ++ __u32 btf_log_true_size; ++ }; ++ struct { ++ __u32 pid; ++ __u32 fd; ++ __u32 flags; ++ __u32 buf_len; ++ __u64 buf; ++ __u32 prog_id; ++ __u32 fd_type; ++ __u64 probe_offset; ++ __u64 probe_addr; ++ } task_fd_query; ++ struct { ++ union { ++ __u32 prog_fd; ++ __u32 map_fd; ++ }; ++ union { ++ __u32 target_fd; ++ __u32 target_ifindex; ++ }; ++ __u32 attach_type; ++ __u32 flags; ++ union { ++ __u32 target_btf_id; ++ struct { ++ __u64 iter_info; ++ __u32 iter_info_len; ++ }; ++ struct { ++ __u64 bpf_cookie; ++ } perf_event; ++ struct { ++ __u32 flags; ++ __u32 cnt; ++ __u64 syms; ++ __u64 addrs; ++ __u64 cookies; ++ } kprobe_multi; ++ struct { ++ __u32 target_btf_id; ++ __u64 cookie; ++ } tracing; ++ struct { ++ __u32 pf; ++ __u32 hooknum; ++ __s32 priority; ++ __u32 flags; ++ } netfilter; ++ }; ++ } link_create; ++ struct { ++ __u32 link_fd; ++ union { ++ __u32 new_prog_fd; ++ __u32 new_map_fd; ++ }; ++ __u32 flags; ++ union { ++ __u32 old_prog_fd; ++ __u32 old_map_fd; ++ }; ++ } link_update; ++ struct { ++ __u32 link_fd; ++ } link_detach; ++ struct { ++ __u32 type; ++ } enable_stats; ++ struct { ++ __u32 link_fd; ++ __u32 flags; ++ } iter_create; ++ struct { ++ __u32 prog_fd; ++ __u32 map_fd; ++ __u32 flags; ++ } prog_bind_map; ++}; ++ ++enum bpf_func_id { ++ BPF_FUNC_unspec = 0, ++ BPF_FUNC_map_lookup_elem = 1, ++ BPF_FUNC_map_update_elem = 2, ++ BPF_FUNC_map_delete_elem = 3, ++ BPF_FUNC_probe_read = 4, ++ BPF_FUNC_ktime_get_ns = 5, ++ BPF_FUNC_trace_printk = 6, ++ BPF_FUNC_get_prandom_u32 = 7, ++ BPF_FUNC_get_smp_processor_id = 8, ++ BPF_FUNC_skb_store_bytes = 9, ++ BPF_FUNC_l3_csum_replace = 10, ++ BPF_FUNC_l4_csum_replace = 11, ++ BPF_FUNC_tail_call = 12, ++ BPF_FUNC_clone_redirect = 13, ++ BPF_FUNC_get_current_pid_tgid = 14, ++ BPF_FUNC_get_current_uid_gid = 15, ++ BPF_FUNC_get_current_comm = 16, ++ BPF_FUNC_get_cgroup_classid = 17, ++ BPF_FUNC_skb_vlan_push = 18, ++ BPF_FUNC_skb_vlan_pop = 19, ++ BPF_FUNC_skb_get_tunnel_key = 20, ++ BPF_FUNC_skb_set_tunnel_key = 21, ++ BPF_FUNC_perf_event_read = 22, ++ BPF_FUNC_redirect = 23, ++ BPF_FUNC_get_route_realm = 24, ++ BPF_FUNC_perf_event_output = 25, ++ BPF_FUNC_skb_load_bytes = 26, ++ BPF_FUNC_get_stackid = 27, ++ BPF_FUNC_csum_diff = 28, ++ BPF_FUNC_skb_get_tunnel_opt = 29, ++ BPF_FUNC_skb_set_tunnel_opt = 30, ++ BPF_FUNC_skb_change_proto = 31, ++ BPF_FUNC_skb_change_type = 32, ++ BPF_FUNC_skb_under_cgroup = 33, ++ BPF_FUNC_get_hash_recalc = 34, ++ BPF_FUNC_get_current_task = 35, ++ BPF_FUNC_probe_write_user = 36, ++ BPF_FUNC_current_task_under_cgroup = 37, ++ BPF_FUNC_skb_change_tail = 38, ++ BPF_FUNC_skb_pull_data = 39, ++ BPF_FUNC_csum_update = 40, ++ BPF_FUNC_set_hash_invalid = 41, ++ BPF_FUNC_get_numa_node_id = 42, ++ BPF_FUNC_skb_change_head = 43, ++ BPF_FUNC_xdp_adjust_head = 44, ++ BPF_FUNC_probe_read_str = 45, ++ BPF_FUNC_get_socket_cookie = 46, ++ BPF_FUNC_get_socket_uid = 47, ++ BPF_FUNC_set_hash = 48, ++ BPF_FUNC_setsockopt = 49, ++ BPF_FUNC_skb_adjust_room = 50, ++ BPF_FUNC_redirect_map = 51, ++ BPF_FUNC_sk_redirect_map = 52, ++ BPF_FUNC_sock_map_update = 53, ++ BPF_FUNC_xdp_adjust_meta = 54, ++ BPF_FUNC_perf_event_read_value = 55, ++ BPF_FUNC_perf_prog_read_value = 56, ++ BPF_FUNC_getsockopt = 57, ++ BPF_FUNC_override_return = 58, ++ BPF_FUNC_sock_ops_cb_flags_set = 59, ++ BPF_FUNC_msg_redirect_map = 60, ++ BPF_FUNC_msg_apply_bytes = 61, ++ BPF_FUNC_msg_cork_bytes = 62, ++ BPF_FUNC_msg_pull_data = 63, ++ BPF_FUNC_bind = 64, ++ BPF_FUNC_xdp_adjust_tail = 65, ++ BPF_FUNC_skb_get_xfrm_state = 66, ++ BPF_FUNC_get_stack = 67, ++ BPF_FUNC_skb_load_bytes_relative = 68, ++ BPF_FUNC_fib_lookup = 69, ++ BPF_FUNC_sock_hash_update = 70, ++ BPF_FUNC_msg_redirect_hash = 71, ++ BPF_FUNC_sk_redirect_hash = 72, ++ BPF_FUNC_lwt_push_encap = 73, ++ BPF_FUNC_lwt_seg6_store_bytes = 74, ++ BPF_FUNC_lwt_seg6_adjust_srh = 75, ++ BPF_FUNC_lwt_seg6_action = 76, ++ BPF_FUNC_rc_repeat = 77, ++ BPF_FUNC_rc_keydown = 78, ++ BPF_FUNC_skb_cgroup_id = 79, ++ BPF_FUNC_get_current_cgroup_id = 80, ++ BPF_FUNC_get_local_storage = 81, ++ BPF_FUNC_sk_select_reuseport = 82, ++ BPF_FUNC_skb_ancestor_cgroup_id = 83, ++ BPF_FUNC_sk_lookup_tcp = 84, ++ BPF_FUNC_sk_lookup_udp = 85, ++ BPF_FUNC_sk_release = 86, ++ BPF_FUNC_map_push_elem = 87, ++ BPF_FUNC_map_pop_elem = 88, ++ BPF_FUNC_map_peek_elem = 89, ++ BPF_FUNC_msg_push_data = 90, ++ BPF_FUNC_msg_pop_data = 91, ++ BPF_FUNC_rc_pointer_rel = 92, ++ BPF_FUNC_spin_lock = 93, ++ BPF_FUNC_spin_unlock = 94, ++ BPF_FUNC_sk_fullsock = 95, ++ BPF_FUNC_tcp_sock = 96, ++ BPF_FUNC_skb_ecn_set_ce = 97, ++ BPF_FUNC_get_listener_sock = 98, ++ BPF_FUNC_skc_lookup_tcp = 99, ++ BPF_FUNC_tcp_check_syncookie = 100, ++ BPF_FUNC_sysctl_get_name = 101, ++ BPF_FUNC_sysctl_get_current_value = 102, ++ BPF_FUNC_sysctl_get_new_value = 103, ++ BPF_FUNC_sysctl_set_new_value = 104, ++ BPF_FUNC_strtol = 105, ++ BPF_FUNC_strtoul = 106, ++ BPF_FUNC_sk_storage_get = 107, ++ BPF_FUNC_sk_storage_delete = 108, ++ BPF_FUNC_send_signal = 109, ++ BPF_FUNC_tcp_gen_syncookie = 110, ++ BPF_FUNC_skb_output = 111, ++ BPF_FUNC_probe_read_user = 112, ++ BPF_FUNC_probe_read_kernel = 113, ++ BPF_FUNC_probe_read_user_str = 114, ++ BPF_FUNC_probe_read_kernel_str = 115, ++ BPF_FUNC_tcp_send_ack = 116, ++ BPF_FUNC_send_signal_thread = 117, ++ BPF_FUNC_jiffies64 = 118, ++ BPF_FUNC_read_branch_records = 119, ++ BPF_FUNC_get_ns_current_pid_tgid = 120, ++ BPF_FUNC_xdp_output = 121, ++ BPF_FUNC_get_netns_cookie = 122, ++ BPF_FUNC_get_current_ancestor_cgroup_id = 123, ++ BPF_FUNC_sk_assign = 124, ++ BPF_FUNC_ktime_get_boot_ns = 125, ++ BPF_FUNC_seq_printf = 126, ++ BPF_FUNC_seq_write = 127, ++ BPF_FUNC_sk_cgroup_id = 128, ++ BPF_FUNC_sk_ancestor_cgroup_id = 129, ++ BPF_FUNC_ringbuf_output = 130, ++ BPF_FUNC_ringbuf_reserve = 131, ++ BPF_FUNC_ringbuf_submit = 132, ++ BPF_FUNC_ringbuf_discard = 133, ++ BPF_FUNC_ringbuf_query = 134, ++ BPF_FUNC_csum_level = 135, ++ BPF_FUNC_skc_to_tcp6_sock = 136, ++ BPF_FUNC_skc_to_tcp_sock = 137, ++ BPF_FUNC_skc_to_tcp_timewait_sock = 138, ++ BPF_FUNC_skc_to_tcp_request_sock = 139, ++ BPF_FUNC_skc_to_udp6_sock = 140, ++ BPF_FUNC_get_task_stack = 141, ++ BPF_FUNC_load_hdr_opt = 142, ++ BPF_FUNC_store_hdr_opt = 143, ++ BPF_FUNC_reserve_hdr_opt = 144, ++ BPF_FUNC_inode_storage_get = 145, ++ BPF_FUNC_inode_storage_delete = 146, ++ BPF_FUNC_d_path = 147, ++ BPF_FUNC_copy_from_user = 148, ++ BPF_FUNC_snprintf_btf = 149, ++ BPF_FUNC_seq_printf_btf = 150, ++ BPF_FUNC_skb_cgroup_classid = 151, ++ BPF_FUNC_redirect_neigh = 152, ++ BPF_FUNC_per_cpu_ptr = 153, ++ BPF_FUNC_this_cpu_ptr = 154, ++ BPF_FUNC_redirect_peer = 155, ++ BPF_FUNC_task_storage_get = 156, ++ BPF_FUNC_task_storage_delete = 157, ++ BPF_FUNC_get_current_task_btf = 158, ++ BPF_FUNC_bprm_opts_set = 159, ++ BPF_FUNC_ktime_get_coarse_ns = 160, ++ BPF_FUNC_ima_inode_hash = 161, ++ BPF_FUNC_sock_from_file = 162, ++ BPF_FUNC_check_mtu = 163, ++ BPF_FUNC_for_each_map_elem = 164, ++ BPF_FUNC_snprintf = 165, ++ BPF_FUNC_sys_bpf = 166, ++ BPF_FUNC_btf_find_by_name_kind = 167, ++ BPF_FUNC_sys_close = 168, ++ BPF_FUNC_timer_init = 169, ++ BPF_FUNC_timer_set_callback = 170, ++ BPF_FUNC_timer_start = 171, ++ BPF_FUNC_timer_cancel = 172, ++ BPF_FUNC_get_func_ip = 173, ++ BPF_FUNC_get_attach_cookie = 174, ++ BPF_FUNC_task_pt_regs = 175, ++ BPF_FUNC_get_branch_snapshot = 176, ++ BPF_FUNC_trace_vprintk = 177, ++ BPF_FUNC_skc_to_unix_sock = 178, ++ BPF_FUNC_kallsyms_lookup_name = 179, ++ BPF_FUNC_find_vma = 180, ++ BPF_FUNC_loop = 181, ++ BPF_FUNC_strncmp = 182, ++ BPF_FUNC_get_func_arg = 183, ++ BPF_FUNC_get_func_ret = 184, ++ BPF_FUNC_get_func_arg_cnt = 185, ++ BPF_FUNC_get_retval = 186, ++ BPF_FUNC_set_retval = 187, ++ BPF_FUNC_xdp_get_buff_len = 188, ++ BPF_FUNC_xdp_load_bytes = 189, ++ BPF_FUNC_xdp_store_bytes = 190, ++ BPF_FUNC_copy_from_user_task = 191, ++ BPF_FUNC_skb_set_tstamp = 192, ++ BPF_FUNC_ima_file_hash = 193, ++ BPF_FUNC_kptr_xchg = 194, ++ BPF_FUNC_map_lookup_percpu_elem = 195, ++ BPF_FUNC_skc_to_mptcp_sock = 196, ++ BPF_FUNC_dynptr_from_mem = 197, ++ BPF_FUNC_ringbuf_reserve_dynptr = 198, ++ BPF_FUNC_ringbuf_submit_dynptr = 199, ++ BPF_FUNC_ringbuf_discard_dynptr = 200, ++ BPF_FUNC_dynptr_read = 201, ++ BPF_FUNC_dynptr_write = 202, ++ BPF_FUNC_dynptr_data = 203, ++ BPF_FUNC_tcp_raw_gen_syncookie_ipv4 = 204, ++ BPF_FUNC_tcp_raw_gen_syncookie_ipv6 = 205, ++ BPF_FUNC_tcp_raw_check_syncookie_ipv4 = 206, ++ BPF_FUNC_tcp_raw_check_syncookie_ipv6 = 207, ++ BPF_FUNC_ktime_get_tai_ns = 208, ++ BPF_FUNC_user_ringbuf_drain = 209, ++ BPF_FUNC_cgrp_storage_get = 210, ++ BPF_FUNC_cgrp_storage_delete = 211, ++ __BPF_FUNC_MAX_ID = 212, ++}; ++ ++enum sk_action { ++ SK_DROP = 0, ++ SK_PASS = 1, ++}; ++ ++struct bpf_func_info { ++ __u32 insn_off; ++ __u32 type_id; ++}; ++ ++struct bpf_line_info { ++ __u32 insn_off; ++ __u32 file_name_off; ++ __u32 line_off; ++ __u32 line_col; ++}; ++ ++enum bpf_core_relo_kind { ++ BPF_CORE_FIELD_BYTE_OFFSET = 0, ++ BPF_CORE_FIELD_BYTE_SIZE = 1, ++ BPF_CORE_FIELD_EXISTS = 2, ++ BPF_CORE_FIELD_SIGNED = 3, ++ BPF_CORE_FIELD_LSHIFT_U64 = 4, ++ BPF_CORE_FIELD_RSHIFT_U64 = 5, ++ BPF_CORE_TYPE_ID_LOCAL = 6, ++ BPF_CORE_TYPE_ID_TARGET = 7, ++ BPF_CORE_TYPE_EXISTS = 8, ++ BPF_CORE_TYPE_SIZE = 9, ++ BPF_CORE_ENUMVAL_EXISTS = 10, ++ BPF_CORE_ENUMVAL_VALUE = 11, ++ BPF_CORE_TYPE_MATCHES = 12, ++}; ++ ++struct bpf_core_relo { ++ __u32 insn_off; ++ __u32 type_id; ++ __u32 access_str_off; ++ enum bpf_core_relo_kind kind; ++}; ++ ++struct sock_filter { ++ __u16 code; ++ __u8 jt; ++ __u8 jf; ++ __u32 k; ++}; ++ ++struct bpf_run_ctx {}; ++ ++typedef sockptr_t bpfptr_t; ++ ++struct btf_id_set { ++ u32 cnt; ++ u32 ids[0]; ++}; ++ ++enum { ++ BTF_SOCK_TYPE_INET = 0, ++ BTF_SOCK_TYPE_INET_CONN = 1, ++ BTF_SOCK_TYPE_INET_REQ = 2, ++ BTF_SOCK_TYPE_INET_TW = 3, ++ BTF_SOCK_TYPE_REQ = 4, ++ BTF_SOCK_TYPE_SOCK = 5, ++ BTF_SOCK_TYPE_SOCK_COMMON = 6, ++ BTF_SOCK_TYPE_TCP = 7, ++ BTF_SOCK_TYPE_TCP_REQ = 8, ++ BTF_SOCK_TYPE_TCP_TW = 9, ++ BTF_SOCK_TYPE_TCP6 = 10, ++ BTF_SOCK_TYPE_UDP = 11, ++ BTF_SOCK_TYPE_UDP6 = 12, ++ BTF_SOCK_TYPE_UNIX = 13, ++ BTF_SOCK_TYPE_MPTCP = 14, ++ BTF_SOCK_TYPE_SOCKET = 15, ++ MAX_BTF_SOCK_TYPE = 16, ++}; ++ ++enum { ++ BTF_TRACING_TYPE_TASK = 0, ++ BTF_TRACING_TYPE_FILE = 1, ++ BTF_TRACING_TYPE_VMA = 2, ++ MAX_BTF_TRACING_TYPE = 3, ++}; ++ ++struct btf_record; ++ ++struct btf_struct_meta { ++ u32 btf_id; ++ struct btf_record *record; ++}; ++ ++enum btf_field_type { ++ BPF_SPIN_LOCK = 1, ++ BPF_TIMER = 2, ++ BPF_KPTR_UNREF = 4, ++ BPF_KPTR_REF = 8, ++ BPF_KPTR = 12, ++ BPF_LIST_HEAD = 16, ++ BPF_LIST_NODE = 32, ++ BPF_RB_ROOT = 64, ++ BPF_RB_NODE = 128, ++ BPF_GRAPH_NODE_OR_ROOT = 240, ++ BPF_REFCOUNT = 256, ++}; ++ ++typedef void (*btf_dtor_kfunc_t)(void *); ++ ++struct btf; ++ ++struct btf_field_kptr { ++ struct btf *btf; ++ struct module *module; ++ btf_dtor_kfunc_t dtor; ++ u32 btf_id; ++}; ++ ++struct btf_field_graph_root { ++ struct btf *btf; ++ u32 value_btf_id; ++ u32 node_offset; ++ struct btf_record *value_rec; ++}; ++ ++struct btf_field { ++ u32 offset; ++ u32 size; ++ enum btf_field_type type; ++ union { ++ struct btf_field_kptr kptr; ++ struct btf_field_graph_root graph_root; ++ }; ++}; ++ ++struct btf_record { ++ u32 cnt; ++ u32 field_mask; ++ int spin_lock_off; ++ int timer_off; ++ int refcount_off; ++ struct btf_field fields[0]; ++}; ++ ++struct bpf_cgroup_storage; ++ ++struct bpf_prog_array_item { ++ struct bpf_prog *prog; ++ union { ++ struct bpf_cgroup_storage *cgroup_storage[2]; ++ u64 bpf_cookie; ++ }; ++}; ++ ++struct bpf_prog_array { ++ struct callback_head rcu; ++ struct bpf_prog_array_item items[0]; ++}; ++ ++typedef u64 (*bpf_callback_t)(u64, u64, u64, u64, u64); ++ ++struct bpf_iter_aux_info; ++ ++typedef int (*bpf_iter_init_seq_priv_t)(void *, struct bpf_iter_aux_info *); ++ ++enum bpf_iter_task_type { ++ BPF_TASK_ITER_ALL = 0, ++ BPF_TASK_ITER_TID = 1, ++ BPF_TASK_ITER_TGID = 2, ++}; ++ ++struct bpf_map; ++ ++struct bpf_iter_aux_info { ++ struct bpf_map *map; ++ struct { ++ struct cgroup *start; ++ enum bpf_cgroup_iter_order order; ++ } cgroup; ++ struct { ++ enum bpf_iter_task_type type; ++ u32 pid; ++ } task; ++}; ++ ++typedef void (*bpf_iter_fini_seq_priv_t)(void *); ++ ++struct bpf_iter_seq_info { ++ const struct seq_operations *seq_ops; ++ bpf_iter_init_seq_priv_t init_seq_private; ++ bpf_iter_fini_seq_priv_t fini_seq_private; ++ u32 seq_priv_size; ++}; ++ ++struct bpf_prog_aux; ++ ++struct bpf_local_storage_map; ++ ++struct bpf_verifier_env; ++ ++struct bpf_func_state; ++ ++struct bpf_map_ops { ++ int (*map_alloc_check)(union bpf_attr *); ++ struct bpf_map * (*map_alloc)(union bpf_attr *); ++ void (*map_release)(struct bpf_map *, struct file *); ++ void (*map_free)(struct bpf_map *); ++ int (*map_get_next_key)(struct bpf_map *, void *, void *); ++ void (*map_release_uref)(struct bpf_map *); ++ void * (*map_lookup_elem_sys_only)(struct bpf_map *, void *); ++ int (*map_lookup_batch)(struct bpf_map *, const union bpf_attr *, union bpf_attr *); ++ int (*map_lookup_and_delete_elem)(struct bpf_map *, void *, void *, u64); ++ int (*map_lookup_and_delete_batch)(struct bpf_map *, const union bpf_attr *, union bpf_attr *); ++ int (*map_update_batch)(struct bpf_map *, struct file *, const union bpf_attr *, union bpf_attr *); ++ int (*map_delete_batch)(struct bpf_map *, const union bpf_attr *, union bpf_attr *); ++ void * (*map_lookup_elem)(struct bpf_map *, void *); ++ long int (*map_update_elem)(struct bpf_map *, void *, void *, u64); ++ long int (*map_delete_elem)(struct bpf_map *, void *); ++ long int (*map_push_elem)(struct bpf_map *, void *, u64); ++ long int (*map_pop_elem)(struct bpf_map *, void *); ++ long int (*map_peek_elem)(struct bpf_map *, void *); ++ void * (*map_lookup_percpu_elem)(struct bpf_map *, void *, u32); ++ void * (*map_fd_get_ptr)(struct bpf_map *, struct file *, int); ++ void (*map_fd_put_ptr)(void *); ++ int (*map_gen_lookup)(struct bpf_map *, struct bpf_insn *); ++ u32 (*map_fd_sys_lookup_elem)(void *); ++ void (*map_seq_show_elem)(struct bpf_map *, void *, struct seq_file *); ++ int (*map_check_btf)(const struct bpf_map *, const struct btf *, const struct btf_type *, const struct btf_type *); ++ int (*map_poke_track)(struct bpf_map *, struct bpf_prog_aux *); ++ void (*map_poke_untrack)(struct bpf_map *, struct bpf_prog_aux *); ++ void (*map_poke_run)(struct bpf_map *, u32, struct bpf_prog *, struct bpf_prog *); ++ int (*map_direct_value_addr)(const struct bpf_map *, u64 *, u32); ++ int (*map_direct_value_meta)(const struct bpf_map *, u64, u32 *); ++ int (*map_mmap)(struct bpf_map *, struct vm_area_struct *); ++ __poll_t (*map_poll)(struct bpf_map *, struct file *, struct poll_table_struct *); ++ int (*map_local_storage_charge)(struct bpf_local_storage_map *, void *, u32); ++ void (*map_local_storage_uncharge)(struct bpf_local_storage_map *, void *, u32); ++ struct bpf_local_storage ** (*map_owner_storage_ptr)(void *); ++ long int (*map_redirect)(struct bpf_map *, u64, u64); ++ bool (*map_meta_equal)(const struct bpf_map *, const struct bpf_map *); ++ int (*map_set_for_each_callback_args)(struct bpf_verifier_env *, struct bpf_func_state *, struct bpf_func_state *); ++ long int (*map_for_each_callback)(struct bpf_map *, bpf_callback_t, void *, u64); ++ u64 (*map_mem_usage)(const struct bpf_map *); ++ int *map_btf_id; ++ const struct bpf_iter_seq_info *iter_seq_info; ++}; ++ ++struct bpf_map { ++ const struct bpf_map_ops *ops; ++ struct bpf_map *inner_map_meta; ++ void *security; ++ enum bpf_map_type map_type; ++ u32 key_size; ++ u32 value_size; ++ u32 max_entries; ++ u64 map_extra; ++ u32 map_flags; ++ u32 id; ++ struct btf_record *record; ++ int numa_node; ++ u32 btf_key_type_id; ++ u32 btf_value_type_id; ++ u32 btf_vmlinux_value_type_id; ++ struct btf *btf; ++ struct obj_cgroup *objcg; ++ char name[16]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ atomic64_t refcnt; ++ atomic64_t usercnt; ++ struct work_struct work; ++ struct mutex freeze_mutex; ++ atomic64_t writecnt; ++ struct { ++ spinlock_t lock; ++ enum bpf_prog_type type; ++ bool jited; ++ bool xdp_has_frags; ++ } owner; ++ bool bypass_spec_v1; ++ bool frozen; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct btf_header { ++ __u16 magic; ++ __u8 version; ++ __u8 flags; ++ __u32 hdr_len; ++ __u32 type_off; ++ __u32 type_len; ++ __u32 str_off; ++ __u32 str_len; ++}; ++ ++struct btf_kfunc_set_tab; ++ ++struct btf_id_dtor_kfunc_tab; ++ ++struct btf_struct_metas; ++ ++struct btf { ++ void *data; ++ struct btf_type **types; ++ u32 *resolved_ids; ++ u32 *resolved_sizes; ++ const char *strings; ++ void *nohdr_data; ++ struct btf_header hdr; ++ u32 nr_types; ++ u32 types_size; ++ u32 data_size; ++ refcount_t refcnt; ++ u32 id; ++ struct callback_head rcu; ++ struct btf_kfunc_set_tab *kfunc_set_tab; ++ struct btf_id_dtor_kfunc_tab *dtor_kfunc_tab; ++ struct btf_struct_metas *struct_meta_tab; ++ struct btf *base_btf; ++ u32 start_id; ++ u32 start_str_off; ++ char name[56]; ++ bool kernel_btf; ++}; ++ ++struct bpf_ksym { ++ long unsigned int start; ++ long unsigned int end; ++ char name[512]; ++ struct list_head lnode; ++ struct latch_tree_node tnode; ++ bool prog; ++}; ++ ++struct bpf_ctx_arg_aux; ++ ++struct bpf_trampoline; ++ ++struct bpf_jit_poke_descriptor; ++ ++struct bpf_kfunc_desc_tab; ++ ++struct bpf_kfunc_btf_tab; ++ ++struct bpf_prog_ops; ++ ++struct btf_mod_pair; ++ ++struct bpf_prog_offload; ++ ++struct bpf_func_info_aux; ++ ++struct bpf_prog_aux { ++ atomic64_t refcnt; ++ u32 used_map_cnt; ++ u32 used_btf_cnt; ++ u32 max_ctx_offset; ++ u32 max_pkt_offset; ++ u32 max_tp_access; ++ u32 stack_depth; ++ u32 id; ++ u32 func_cnt; ++ u32 func_idx; ++ u32 attach_btf_id; ++ u32 ctx_arg_info_size; ++ u32 max_rdonly_access; ++ u32 max_rdwr_access; ++ struct btf *attach_btf; ++ const struct bpf_ctx_arg_aux *ctx_arg_info; ++ struct mutex dst_mutex; ++ struct bpf_prog *dst_prog; ++ struct bpf_trampoline *dst_trampoline; ++ enum bpf_prog_type saved_dst_prog_type; ++ enum bpf_attach_type saved_dst_attach_type; ++ bool verifier_zext; ++ bool dev_bound; ++ bool offload_requested; ++ bool attach_btf_trace; ++ bool func_proto_unreliable; ++ bool sleepable; ++ bool tail_call_reachable; ++ bool xdp_has_frags; ++ const struct btf_type *attach_func_proto; ++ const char *attach_func_name; ++ struct bpf_prog **func; ++ void *jit_data; ++ struct bpf_jit_poke_descriptor *poke_tab; ++ struct bpf_kfunc_desc_tab *kfunc_tab; ++ struct bpf_kfunc_btf_tab *kfunc_btf_tab; ++ u32 size_poke_tab; ++ struct bpf_ksym ksym; ++ const struct bpf_prog_ops *ops; ++ struct bpf_map **used_maps; ++ struct mutex used_maps_mutex; ++ struct btf_mod_pair *used_btfs; ++ struct bpf_prog *prog; ++ struct user_struct *user; ++ u64 load_time; ++ u32 verified_insns; ++ int cgroup_atype; ++ struct bpf_map *cgroup_storage[2]; ++ char name[16]; ++ void *security; ++ struct bpf_prog_offload *offload; ++ struct btf *btf; ++ struct bpf_func_info *func_info; ++ struct bpf_func_info_aux *func_info_aux; ++ struct bpf_line_info *linfo; ++ void **jited_linfo; ++ u32 func_info_cnt; ++ u32 nr_linfo; ++ u32 linfo_idx; ++ struct module *mod; ++ u32 num_exentries; ++ struct exception_table_entry *extable; ++ union { ++ struct work_struct work; ++ struct callback_head rcu; ++ }; ++}; ++ ++struct bpf_prog_stats; ++ ++struct sock_fprog_kern; ++ ++struct bpf_prog { ++ u16 pages; ++ u16 jited: 1; ++ u16 jit_requested: 1; ++ u16 gpl_compatible: 1; ++ u16 cb_access: 1; ++ u16 dst_needed: 1; ++ u16 blinding_requested: 1; ++ u16 blinded: 1; ++ u16 is_func: 1; ++ u16 kprobe_override: 1; ++ u16 has_callchain_buf: 1; ++ u16 enforce_expected_attach_type: 1; ++ u16 call_get_stack: 1; ++ u16 call_get_func_ip: 1; ++ u16 tstamp_type_access: 1; ++ enum bpf_prog_type type; ++ enum bpf_attach_type expected_attach_type; ++ u32 len; ++ u32 jited_len; ++ u8 tag[8]; ++ struct bpf_prog_stats *stats; ++ int *active; ++ unsigned int (*bpf_func)(const void *, const struct bpf_insn *); ++ struct bpf_prog_aux *aux; ++ struct sock_fprog_kern *orig_prog; ++ union { ++ struct { ++ struct {} __empty_insns; ++ struct sock_filter insns[0]; ++ }; ++ struct { ++ struct {} __empty_insnsi; ++ struct bpf_insn insnsi[0]; ++ }; ++ }; ++}; ++ ++struct btf_mod_pair { ++ struct btf *btf; ++ struct module *module; ++}; ++ ++struct bpf_verifier_log { ++ u64 start_pos; ++ u64 end_pos; ++ char *ubuf; ++ u32 level; ++ u32 len_total; ++ u32 len_max; ++ char kbuf[1024]; ++}; ++ ++struct bpf_subprog_info { ++ u32 start; ++ u32 linfo_idx; ++ u16 stack_depth; ++ bool has_tail_call; ++ bool tail_call_reachable; ++ bool has_ld_abs; ++ bool is_async_cb; ++}; ++ ++struct bpf_id_pair { ++ u32 old; ++ u32 cur; ++}; ++ ++struct bpf_idmap { ++ u32 tmp_id_gen; ++ struct bpf_id_pair map[600]; ++}; ++ ++struct bpf_idset { ++ u32 count; ++ u32 ids[600]; ++}; ++ ++struct backtrack_state { ++ struct bpf_verifier_env *env; ++ u32 frame; ++ u32 reg_masks[8]; ++ u64 stack_masks[8]; ++}; ++ ++struct bpf_verifier_ops; ++ ++struct bpf_verifier_stack_elem; ++ ++struct bpf_verifier_state; ++ ++struct bpf_verifier_state_list; ++ ++struct bpf_insn_aux_data; ++ ++struct bpf_verifier_env { ++ u32 insn_idx; ++ u32 prev_insn_idx; ++ struct bpf_prog *prog; ++ const struct bpf_verifier_ops *ops; ++ struct bpf_verifier_stack_elem *head; ++ int stack_size; ++ bool strict_alignment; ++ bool test_state_freq; ++ struct bpf_verifier_state *cur_state; ++ struct bpf_verifier_state_list **explored_states; ++ struct bpf_verifier_state_list *free_list; ++ struct bpf_map *used_maps[64]; ++ struct btf_mod_pair used_btfs[64]; ++ u32 used_map_cnt; ++ u32 used_btf_cnt; ++ u32 id_gen; ++ bool explore_alu_limits; ++ bool allow_ptr_leaks; ++ bool allow_uninit_stack; ++ bool bpf_capable; ++ bool bypass_spec_v1; ++ bool bypass_spec_v4; ++ bool seen_direct_write; ++ struct bpf_insn_aux_data *insn_aux_data; ++ const struct bpf_line_info *prev_linfo; ++ struct bpf_verifier_log log; ++ struct bpf_subprog_info subprog_info[257]; ++ union { ++ struct bpf_idmap idmap_scratch; ++ struct bpf_idset idset_scratch; ++ }; ++ struct { ++ int *insn_state; ++ int *insn_stack; ++ int cur_stack; ++ } cfg; ++ struct backtrack_state bt; ++ u32 pass_cnt; ++ u32 subprog_cnt; ++ u32 prev_insn_processed; ++ u32 insn_processed; ++ u32 prev_jmps_processed; ++ u32 jmps_processed; ++ u64 verification_time; ++ u32 max_states_per_insn; ++ u32 total_states; ++ u32 peak_states; ++ u32 longest_mark_read_walk; ++ bpfptr_t fd_array; ++ u32 scratched_regs; ++ u64 scratched_stack_slots; ++ u64 prev_log_pos; ++ u64 prev_insn_print_pos; ++ char tmp_str_buf[320]; ++}; ++ ++enum bpf_reg_type { ++ NOT_INIT = 0, ++ SCALAR_VALUE = 1, ++ PTR_TO_CTX = 2, ++ CONST_PTR_TO_MAP = 3, ++ PTR_TO_MAP_VALUE = 4, ++ PTR_TO_MAP_KEY = 5, ++ PTR_TO_STACK = 6, ++ PTR_TO_PACKET_META = 7, ++ PTR_TO_PACKET = 8, ++ PTR_TO_PACKET_END = 9, ++ PTR_TO_FLOW_KEYS = 10, ++ PTR_TO_SOCKET = 11, ++ PTR_TO_SOCK_COMMON = 12, ++ PTR_TO_TCP_SOCK = 13, ++ PTR_TO_TP_BUFFER = 14, ++ PTR_TO_XDP_SOCK = 15, ++ PTR_TO_BTF_ID = 16, ++ PTR_TO_MEM = 17, ++ PTR_TO_BUF = 18, ++ PTR_TO_FUNC = 19, ++ CONST_PTR_TO_DYNPTR = 20, ++ __BPF_REG_TYPE_MAX = 21, ++ PTR_TO_MAP_VALUE_OR_NULL = 260, ++ PTR_TO_SOCKET_OR_NULL = 267, ++ PTR_TO_SOCK_COMMON_OR_NULL = 268, ++ PTR_TO_TCP_SOCK_OR_NULL = 269, ++ PTR_TO_BTF_ID_OR_NULL = 272, ++ __BPF_REG_TYPE_LIMIT = 33554431, ++}; ++ ++enum bpf_dynptr_type { ++ BPF_DYNPTR_TYPE_INVALID = 0, ++ BPF_DYNPTR_TYPE_LOCAL = 1, ++ BPF_DYNPTR_TYPE_RINGBUF = 2, ++ BPF_DYNPTR_TYPE_SKB = 3, ++ BPF_DYNPTR_TYPE_XDP = 4, ++}; ++ ++enum bpf_iter_state { ++ BPF_ITER_STATE_INVALID = 0, ++ BPF_ITER_STATE_ACTIVE = 1, ++ BPF_ITER_STATE_DRAINED = 2, ++}; ++ ++struct tnum { ++ u64 value; ++ u64 mask; ++}; ++ ++enum bpf_reg_liveness { ++ REG_LIVE_NONE = 0, ++ REG_LIVE_READ32 = 1, ++ REG_LIVE_READ64 = 2, ++ REG_LIVE_READ = 3, ++ REG_LIVE_WRITTEN = 4, ++ REG_LIVE_DONE = 8, ++}; ++ ++struct bpf_reg_state { ++ enum bpf_reg_type type; ++ s32 off; ++ union { ++ int range; ++ struct { ++ struct bpf_map *map_ptr; ++ u32 map_uid; ++ }; ++ struct { ++ struct btf *btf; ++ u32 btf_id; ++ }; ++ struct { ++ u32 mem_size; ++ u32 dynptr_id; ++ }; ++ struct { ++ enum bpf_dynptr_type type; ++ bool first_slot; ++ } dynptr; ++ struct { ++ struct btf *btf; ++ u32 btf_id; ++ enum bpf_iter_state state: 2; ++ int depth: 30; ++ } iter; ++ struct { ++ long unsigned int raw1; ++ long unsigned int raw2; ++ } raw; ++ u32 subprogno; ++ }; ++ struct tnum var_off; ++ s64 smin_value; ++ s64 smax_value; ++ u64 umin_value; ++ u64 umax_value; ++ s32 s32_min_value; ++ s32 s32_max_value; ++ u32 u32_min_value; ++ u32 u32_max_value; ++ u32 id; ++ u32 ref_obj_id; ++ struct bpf_reg_state *parent; ++ u32 frameno; ++ s32 subreg_def; ++ enum bpf_reg_liveness live; ++ bool precise; ++}; ++ ++struct bpf_reference_state; ++ ++struct bpf_stack_state; ++ ++struct bpf_func_state { ++ struct bpf_reg_state regs[11]; ++ int callsite; ++ u32 frameno; ++ u32 subprogno; ++ u32 async_entry_cnt; ++ bool in_callback_fn; ++ struct tnum callback_ret_range; ++ bool in_async_callback_fn; ++ int acquired_refs; ++ struct bpf_reference_state *refs; ++ int allocated_stack; ++ struct bpf_stack_state *stack; ++}; ++ ++struct bpf_map_dev_ops { ++ int (*map_get_next_key)(struct bpf_offloaded_map *, void *, void *); ++ int (*map_lookup_elem)(struct bpf_offloaded_map *, void *, void *); ++ int (*map_update_elem)(struct bpf_offloaded_map *, void *, void *, u64); ++ int (*map_delete_elem)(struct bpf_offloaded_map *, void *); ++}; ++ ++struct bpf_offloaded_map { ++ struct bpf_map map; ++ struct net_device *netdev; ++ const struct bpf_map_dev_ops *dev_ops; ++ void *dev_priv; ++ struct list_head offloads; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum bpf_type_flag { ++ PTR_MAYBE_NULL = 256, ++ MEM_RDONLY = 512, ++ MEM_RINGBUF = 1024, ++ MEM_USER = 2048, ++ MEM_PERCPU = 4096, ++ OBJ_RELEASE = 8192, ++ PTR_UNTRUSTED = 16384, ++ MEM_UNINIT = 32768, ++ DYNPTR_TYPE_LOCAL = 65536, ++ DYNPTR_TYPE_RINGBUF = 131072, ++ MEM_FIXED_SIZE = 262144, ++ MEM_ALLOC = 524288, ++ PTR_TRUSTED = 1048576, ++ MEM_RCU = 2097152, ++ NON_OWN_REF = 4194304, ++ DYNPTR_TYPE_SKB = 8388608, ++ DYNPTR_TYPE_XDP = 16777216, ++ __BPF_TYPE_FLAG_MAX = 16777217, ++ __BPF_TYPE_LAST_FLAG = 16777216, ++}; ++ ++enum bpf_arg_type { ++ ARG_DONTCARE = 0, ++ ARG_CONST_MAP_PTR = 1, ++ ARG_PTR_TO_MAP_KEY = 2, ++ ARG_PTR_TO_MAP_VALUE = 3, ++ ARG_PTR_TO_MEM = 4, ++ ARG_CONST_SIZE = 5, ++ ARG_CONST_SIZE_OR_ZERO = 6, ++ ARG_PTR_TO_CTX = 7, ++ ARG_ANYTHING = 8, ++ ARG_PTR_TO_SPIN_LOCK = 9, ++ ARG_PTR_TO_SOCK_COMMON = 10, ++ ARG_PTR_TO_INT = 11, ++ ARG_PTR_TO_LONG = 12, ++ ARG_PTR_TO_SOCKET = 13, ++ ARG_PTR_TO_BTF_ID = 14, ++ ARG_PTR_TO_RINGBUF_MEM = 15, ++ ARG_CONST_ALLOC_SIZE_OR_ZERO = 16, ++ ARG_PTR_TO_BTF_ID_SOCK_COMMON = 17, ++ ARG_PTR_TO_PERCPU_BTF_ID = 18, ++ ARG_PTR_TO_FUNC = 19, ++ ARG_PTR_TO_STACK = 20, ++ ARG_PTR_TO_CONST_STR = 21, ++ ARG_PTR_TO_TIMER = 22, ++ ARG_PTR_TO_KPTR = 23, ++ ARG_PTR_TO_DYNPTR = 24, ++ __BPF_ARG_TYPE_MAX = 25, ++ ARG_PTR_TO_MAP_VALUE_OR_NULL = 259, ++ ARG_PTR_TO_MEM_OR_NULL = 260, ++ ARG_PTR_TO_CTX_OR_NULL = 263, ++ ARG_PTR_TO_SOCKET_OR_NULL = 269, ++ ARG_PTR_TO_STACK_OR_NULL = 276, ++ ARG_PTR_TO_BTF_ID_OR_NULL = 270, ++ ARG_PTR_TO_UNINIT_MEM = 32772, ++ ARG_PTR_TO_FIXED_SIZE_MEM = 262148, ++ __BPF_ARG_TYPE_LIMIT = 33554431, ++}; ++ ++enum bpf_return_type { ++ RET_INTEGER = 0, ++ RET_VOID = 1, ++ RET_PTR_TO_MAP_VALUE = 2, ++ RET_PTR_TO_SOCKET = 3, ++ RET_PTR_TO_TCP_SOCK = 4, ++ RET_PTR_TO_SOCK_COMMON = 5, ++ RET_PTR_TO_MEM = 6, ++ RET_PTR_TO_MEM_OR_BTF_ID = 7, ++ RET_PTR_TO_BTF_ID = 8, ++ __BPF_RET_TYPE_MAX = 9, ++ RET_PTR_TO_MAP_VALUE_OR_NULL = 258, ++ RET_PTR_TO_SOCKET_OR_NULL = 259, ++ RET_PTR_TO_TCP_SOCK_OR_NULL = 260, ++ RET_PTR_TO_SOCK_COMMON_OR_NULL = 261, ++ RET_PTR_TO_RINGBUF_MEM_OR_NULL = 1286, ++ RET_PTR_TO_DYNPTR_MEM_OR_NULL = 262, ++ RET_PTR_TO_BTF_ID_OR_NULL = 264, ++ RET_PTR_TO_BTF_ID_TRUSTED = 1048584, ++ __BPF_RET_TYPE_LIMIT = 33554431, ++}; ++ ++struct bpf_func_proto { ++ u64 (*func)(u64, u64, u64, u64, u64); ++ bool gpl_only; ++ bool pkt_access; ++ bool might_sleep; ++ enum bpf_return_type ret_type; ++ union { ++ struct { ++ enum bpf_arg_type arg1_type; ++ enum bpf_arg_type arg2_type; ++ enum bpf_arg_type arg3_type; ++ enum bpf_arg_type arg4_type; ++ enum bpf_arg_type arg5_type; ++ }; ++ enum bpf_arg_type arg_type[5]; ++ }; ++ union { ++ struct { ++ u32 *arg1_btf_id; ++ u32 *arg2_btf_id; ++ u32 *arg3_btf_id; ++ u32 *arg4_btf_id; ++ u32 *arg5_btf_id; ++ }; ++ u32 *arg_btf_id[5]; ++ struct { ++ size_t arg1_size; ++ size_t arg2_size; ++ size_t arg3_size; ++ size_t arg4_size; ++ size_t arg5_size; ++ }; ++ size_t arg_size[5]; ++ }; ++ int *ret_btf_id; ++ bool (*allowed)(const struct bpf_prog *); ++}; ++ ++enum bpf_access_type { ++ BPF_READ = 1, ++ BPF_WRITE = 2, ++}; ++ ++struct bpf_insn_access_aux { ++ enum bpf_reg_type reg_type; ++ union { ++ int ctx_field_size; ++ struct { ++ struct btf *btf; ++ u32 btf_id; ++ }; ++ }; ++ struct bpf_verifier_log *log; ++}; ++ ++struct bpf_prog_ops { ++ int (*test_run)(struct bpf_prog *, const union bpf_attr *, union bpf_attr *); ++}; ++ ++struct bpf_verifier_ops { ++ const struct bpf_func_proto * (*get_func_proto)(enum bpf_func_id, const struct bpf_prog *); ++ bool (*is_valid_access)(int, int, enum bpf_access_type, const struct bpf_prog *, struct bpf_insn_access_aux *); ++ int (*gen_prologue)(struct bpf_insn *, bool, const struct bpf_prog *); ++ int (*gen_ld_abs)(const struct bpf_insn *, struct bpf_insn *); ++ u32 (*convert_ctx_access)(enum bpf_access_type, const struct bpf_insn *, struct bpf_insn *, struct bpf_prog *, u32 *); ++ int (*btf_struct_access)(struct bpf_verifier_log *, const struct bpf_reg_state *, int, int); ++}; ++ ++struct bpf_offload_dev; ++ ++struct bpf_prog_offload { ++ struct bpf_prog *prog; ++ struct net_device *netdev; ++ struct bpf_offload_dev *offdev; ++ void *dev_priv; ++ struct list_head offloads; ++ bool dev_state; ++ bool opt_failed; ++ void *jited_image; ++ u32 jited_len; ++}; ++ ++struct btf_func_model { ++ u8 ret_size; ++ u8 ret_flags; ++ u8 nr_args; ++ u8 arg_size[12]; ++ u8 arg_flags[12]; ++}; ++ ++struct bpf_tramp_image { ++ void *image; ++ struct bpf_ksym ksym; ++ struct percpu_ref pcref; ++ void *ip_after_call; ++ void *ip_epilogue; ++ union { ++ struct callback_head rcu; ++ struct work_struct work; ++ }; ++}; ++ ++struct bpf_trampoline { ++ struct hlist_node hlist; ++ struct ftrace_ops *fops; ++ struct mutex mutex; ++ refcount_t refcnt; ++ u32 flags; ++ u64 key; ++ struct { ++ struct btf_func_model model; ++ void *addr; ++ bool ftrace_managed; ++ } func; ++ struct bpf_prog *extension_prog; ++ struct hlist_head progs_hlist[3]; ++ int progs_cnt[3]; ++ struct bpf_tramp_image *cur_image; ++ struct module *mod; ++}; ++ ++struct bpf_attach_target_info { ++ struct btf_func_model fmodel; ++ long int tgt_addr; ++ struct module *tgt_mod; ++ const char *tgt_name; ++ const struct btf_type *tgt_type; ++}; ++ ++struct bpf_dynptr_kern { ++ void *data; ++ u32 size; ++ u32 offset; ++}; ++ ++struct bpf_func_info_aux { ++ u16 linkage; ++ bool unreliable; ++}; ++ ++enum bpf_jit_poke_reason { ++ BPF_POKE_REASON_TAIL_CALL = 0, ++}; ++ ++struct bpf_jit_poke_descriptor { ++ void *tailcall_target; ++ void *tailcall_bypass; ++ void *bypass_addr; ++ void *aux; ++ union { ++ struct { ++ struct bpf_map *map; ++ u32 key; ++ } tail_call; ++ }; ++ bool tailcall_target_stable; ++ u8 adj_off; ++ u16 reason; ++ u32 insn_idx; ++}; ++ ++struct bpf_ctx_arg_aux { ++ u32 offset; ++ enum bpf_reg_type reg_type; ++ u32 btf_id; ++}; ++ ++struct bpf_kfunc_desc { ++ struct btf_func_model func_model; ++ u32 func_id; ++ s32 imm; ++ u16 offset; ++ long unsigned int addr; ++}; ++ ++struct bpf_kfunc_desc_tab { ++ struct bpf_kfunc_desc descs[256]; ++ u32 nr_descs; ++}; ++ ++struct bpf_kfunc_btf { ++ struct btf *btf; ++ struct module *module; ++ u16 offset; ++}; ++ ++struct bpf_kfunc_btf_tab { ++ struct bpf_kfunc_btf descs[256]; ++ u32 nr_descs; ++}; ++ ++struct bpf_prog_stats { ++ u64_stats_t cnt; ++ u64_stats_t nsecs; ++ u64_stats_t misses; ++ struct u64_stats_sync syncp; ++ long: 64; ++}; ++ ++struct sock_fprog_kern { ++ u16 len; ++ struct sock_filter *filter; ++}; ++ ++struct bpf_array_aux { ++ struct list_head poke_progs; ++ struct bpf_map *map; ++ struct mutex poke_mutex; ++ struct work_struct work; ++}; ++ ++struct bpf_struct_ops { ++ const struct bpf_verifier_ops *verifier_ops; ++ int (*init)(struct btf *); ++ int (*check_member)(const struct btf_type *, const struct btf_member *, const struct bpf_prog *); ++ int (*init_member)(const struct btf_type *, const struct btf_member *, void *, const void *); ++ int (*reg)(void *); ++ void (*unreg)(void *); ++ int (*update)(void *, void *); ++ int (*validate)(void *); ++ const struct btf_type *type; ++ const struct btf_type *value_type; ++ const char *name; ++ struct btf_func_model func_models[64]; ++ u32 type_id; ++ u32 value_id; ++}; ++ ++struct bpf_array { ++ struct bpf_map map; ++ u32 elem_size; ++ u32 index_mask; ++ struct bpf_array_aux *aux; ++ union { ++ struct { ++ struct {} __empty_value; ++ char value[0]; ++ }; ++ struct { ++ struct {} __empty_ptrs; ++ void *ptrs[0]; ++ }; ++ struct { ++ struct {} __empty_pptrs; ++ void *pptrs[0]; ++ }; ++ }; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum { ++ BPF_MAX_LOOPS = 8388608, ++}; ++ ++typedef u32 (*bpf_convert_ctx_access_t)(enum bpf_access_type, const struct bpf_insn *, struct bpf_insn *, struct bpf_prog *, u32 *); ++ ++struct bpf_storage_buffer; ++ ++struct bpf_cgroup_storage_map; ++ ++struct bpf_cgroup_storage { ++ union { ++ struct bpf_storage_buffer *buf; ++ void *percpu_buf; ++ }; ++ struct bpf_cgroup_storage_map *map; ++ struct bpf_cgroup_storage_key key; ++ struct list_head list_map; ++ struct list_head list_cg; ++ struct rb_node node; ++ struct callback_head rcu; ++}; ++ ++struct bpf_iter_meta { ++ union { ++ struct seq_file *seq; ++ }; ++ u64 session_id; ++ u64 seq_num; ++}; ++ ++struct bpf_core_ctx { ++ struct bpf_verifier_log *log; ++ const struct btf *btf; ++}; ++ ++struct bpf_bprintf_data { ++ u32 *bin_args; ++ char *buf; ++ bool get_bin_args; ++ bool get_buf; ++}; ++ ++struct tc_stats { ++ __u64 bytes; ++ __u32 packets; ++ __u32 drops; ++ __u32 overlimits; ++ __u32 bps; ++ __u32 pps; ++ __u32 qlen; ++ __u32 backlog; ++}; ++ ++struct tc_sizespec { ++ unsigned char cell_log; ++ unsigned char size_log; ++ short int cell_align; ++ int overhead; ++ unsigned int linklayer; ++ unsigned int mpu; ++ unsigned int mtu; ++ unsigned int tsize; ++}; ++ ++struct qdisc_skb_head { ++ struct sk_buff *head; ++ struct sk_buff *tail; ++ __u32 qlen; ++ spinlock_t lock; ++}; ++ ++struct gnet_stats_basic_sync { ++ u64_stats_t bytes; ++ u64_stats_t packets; ++ struct u64_stats_sync syncp; ++}; ++ ++struct gnet_stats_queue { ++ __u32 qlen; ++ __u32 backlog; ++ __u32 drops; ++ __u32 requeues; ++ __u32 overlimits; ++}; ++ ++struct Qdisc_ops; ++ ++struct qdisc_size_table; ++ ++struct net_rate_estimator; ++ ++struct Qdisc { ++ int (*enqueue)(struct sk_buff *, struct Qdisc *, struct sk_buff **); ++ struct sk_buff * (*dequeue)(struct Qdisc *); ++ unsigned int flags; ++ u32 limit; ++ const struct Qdisc_ops *ops; ++ struct qdisc_size_table *stab; ++ struct hlist_node hash; ++ u32 handle; ++ u32 parent; ++ struct netdev_queue *dev_queue; ++ struct net_rate_estimator *rate_est; ++ struct gnet_stats_basic_sync *cpu_bstats; ++ struct gnet_stats_queue *cpu_qstats; ++ int pad; ++ refcount_t refcnt; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct sk_buff_head gso_skb; ++ struct qdisc_skb_head q; ++ struct gnet_stats_basic_sync bstats; ++ struct gnet_stats_queue qstats; ++ long unsigned int state; ++ long unsigned int state2; ++ struct Qdisc *next_sched; ++ struct sk_buff_head skb_bad_txq; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ spinlock_t busylock; ++ spinlock_t seqlock; ++ struct callback_head rcu; ++ netdevice_tracker dev_tracker; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long int privdata[0]; ++}; ++ ++struct tcf_proto; ++ ++struct tcf_block; ++ ++struct mini_Qdisc { ++ struct tcf_proto *filter_list; ++ struct tcf_block *block; ++ struct gnet_stats_basic_sync *cpu_bstats; ++ struct gnet_stats_queue *cpu_qstats; ++ long unsigned int rcu_state; ++}; ++ ++struct tcmsg { ++ unsigned char tcm_family; ++ unsigned char tcm__pad1; ++ short unsigned int tcm__pad2; ++ int tcm_ifindex; ++ __u32 tcm_handle; ++ __u32 tcm_parent; ++ __u32 tcm_info; ++}; ++ ++struct sk_filter { ++ refcount_t refcnt; ++ struct callback_head rcu; ++ struct bpf_prog *prog; ++}; ++ ++struct bpf_storage_buffer { ++ struct callback_head rcu; ++ char data[0]; ++}; ++ ++struct gnet_dump { ++ spinlock_t *lock; ++ struct sk_buff *skb; ++ struct nlattr *tail; ++ int compat_tc_stats; ++ int compat_xstats; ++ int padattr; ++ void *xstats; ++ int xstats_len; ++ struct tc_stats tc_stats; ++}; ++ ++enum flow_action_hw_stats_bit { ++ FLOW_ACTION_HW_STATS_IMMEDIATE_BIT = 0, ++ FLOW_ACTION_HW_STATS_DELAYED_BIT = 1, ++ FLOW_ACTION_HW_STATS_DISABLED_BIT = 2, ++ FLOW_ACTION_HW_STATS_NUM_BITS = 3, ++}; ++ ++struct flow_block { ++ struct list_head cb_list; ++}; ++ ++typedef int flow_setup_cb_t(enum tc_setup_type, void *, void *); ++ ++struct qdisc_size_table { ++ struct callback_head rcu; ++ struct list_head list; ++ struct tc_sizespec szopts; ++ int refcnt; ++ u16 data[0]; ++}; ++ ++struct Qdisc_class_ops; ++ ++struct Qdisc_ops { ++ struct Qdisc_ops *next; ++ const struct Qdisc_class_ops *cl_ops; ++ char id[16]; ++ int priv_size; ++ unsigned int static_flags; ++ int (*enqueue)(struct sk_buff *, struct Qdisc *, struct sk_buff **); ++ struct sk_buff * (*dequeue)(struct Qdisc *); ++ struct sk_buff * (*peek)(struct Qdisc *); ++ int (*init)(struct Qdisc *, struct nlattr *, struct netlink_ext_ack *); ++ void (*reset)(struct Qdisc *); ++ void (*destroy)(struct Qdisc *); ++ int (*change)(struct Qdisc *, struct nlattr *, struct netlink_ext_ack *); ++ void (*attach)(struct Qdisc *); ++ int (*change_tx_queue_len)(struct Qdisc *, unsigned int); ++ void (*change_real_num_tx)(struct Qdisc *, unsigned int); ++ int (*dump)(struct Qdisc *, struct sk_buff *); ++ int (*dump_stats)(struct Qdisc *, struct gnet_dump *); ++ void (*ingress_block_set)(struct Qdisc *, u32); ++ void (*egress_block_set)(struct Qdisc *, u32); ++ u32 (*ingress_block_get)(struct Qdisc *); ++ u32 (*egress_block_get)(struct Qdisc *); ++ struct module *owner; ++}; ++ ++struct qdisc_walker; ++ ++struct Qdisc_class_ops { ++ unsigned int flags; ++ struct netdev_queue * (*select_queue)(struct Qdisc *, struct tcmsg *); ++ int (*graft)(struct Qdisc *, long unsigned int, struct Qdisc *, struct Qdisc **, struct netlink_ext_ack *); ++ struct Qdisc * (*leaf)(struct Qdisc *, long unsigned int); ++ void (*qlen_notify)(struct Qdisc *, long unsigned int); ++ long unsigned int (*find)(struct Qdisc *, u32); ++ int (*change)(struct Qdisc *, u32, u32, struct nlattr **, long unsigned int *, struct netlink_ext_ack *); ++ int (*delete)(struct Qdisc *, long unsigned int, struct netlink_ext_ack *); ++ void (*walk)(struct Qdisc *, struct qdisc_walker *); ++ struct tcf_block * (*tcf_block)(struct Qdisc *, long unsigned int, struct netlink_ext_ack *); ++ long unsigned int (*bind_tcf)(struct Qdisc *, long unsigned int, u32); ++ void (*unbind_tcf)(struct Qdisc *, long unsigned int); ++ int (*dump)(struct Qdisc *, long unsigned int, struct sk_buff *, struct tcmsg *); ++ int (*dump_stats)(struct Qdisc *, long unsigned int, struct gnet_dump *); ++}; ++ ++struct tcf_chain; ++ ++struct tcf_block { ++ struct mutex lock; ++ struct list_head chain_list; ++ u32 index; ++ u32 classid; ++ refcount_t refcnt; ++ struct net *net; ++ struct Qdisc *q; ++ struct rw_semaphore cb_lock; ++ struct flow_block flow_block; ++ struct list_head owner_list; ++ bool keep_dst; ++ atomic_t offloadcnt; ++ unsigned int nooffloaddevcnt; ++ unsigned int lockeddevcnt; ++ struct { ++ struct tcf_chain *chain; ++ struct list_head filter_chain_list; ++ } chain0; ++ struct callback_head rcu; ++ struct hlist_head proto_destroy_ht[128]; ++ struct mutex proto_destroy_lock; ++}; ++ ++struct tcf_result; ++ ++struct tcf_proto_ops; ++ ++struct tcf_proto { ++ struct tcf_proto *next; ++ void *root; ++ int (*classify)(struct sk_buff *, const struct tcf_proto *, struct tcf_result *); ++ __be16 protocol; ++ u32 prio; ++ void *data; ++ const struct tcf_proto_ops *ops; ++ struct tcf_chain *chain; ++ spinlock_t lock; ++ bool deleting; ++ refcount_t refcnt; ++ struct callback_head rcu; ++ struct hlist_node destroy_ht_node; ++}; ++ ++struct tcf_result { ++ union { ++ struct { ++ long unsigned int class; ++ u32 classid; ++ }; ++ const struct tcf_proto *goto_tp; ++ }; ++}; ++ ++struct tcf_walker; ++ ++struct tcf_exts; ++ ++struct tcf_proto_ops { ++ struct list_head head; ++ char kind[16]; ++ int (*classify)(struct sk_buff *, const struct tcf_proto *, struct tcf_result *); ++ int (*init)(struct tcf_proto *); ++ void (*destroy)(struct tcf_proto *, bool, struct netlink_ext_ack *); ++ void * (*get)(struct tcf_proto *, u32); ++ void (*put)(struct tcf_proto *, void *); ++ int (*change)(struct net *, struct sk_buff *, struct tcf_proto *, long unsigned int, u32, struct nlattr **, void **, u32, struct netlink_ext_ack *); ++ int (*delete)(struct tcf_proto *, void *, bool *, bool, struct netlink_ext_ack *); ++ bool (*delete_empty)(struct tcf_proto *); ++ void (*walk)(struct tcf_proto *, struct tcf_walker *, bool); ++ int (*reoffload)(struct tcf_proto *, bool, flow_setup_cb_t *, void *, struct netlink_ext_ack *); ++ void (*hw_add)(struct tcf_proto *, void *); ++ void (*hw_del)(struct tcf_proto *, void *); ++ void (*bind_class)(void *, u32, long unsigned int, void *, long unsigned int); ++ void * (*tmplt_create)(struct net *, struct tcf_chain *, struct nlattr **, struct netlink_ext_ack *); ++ void (*tmplt_destroy)(void *); ++ struct tcf_exts * (*get_exts)(const struct tcf_proto *, u32); ++ int (*dump)(struct net *, struct tcf_proto *, void *, struct sk_buff *, struct tcmsg *, bool); ++ int (*terse_dump)(struct net *, struct tcf_proto *, void *, struct sk_buff *, struct tcmsg *, bool); ++ int (*tmplt_dump)(struct sk_buff *, struct net *, void *); ++ struct module *owner; ++ int flags; ++}; ++ ++struct tcf_chain { ++ struct mutex filter_chain_lock; ++ struct tcf_proto *filter_chain; ++ struct list_head list; ++ struct tcf_block *block; ++ u32 index; ++ unsigned int refcnt; ++ unsigned int action_refcnt; ++ bool explicitly_created; ++ bool flushing; ++ const struct tcf_proto_ops *tmplt_ops; ++ void *tmplt_priv; ++ struct callback_head rcu; ++}; ++ ++struct bpf_active_lock { ++ void *ptr; ++ u32 id; ++}; ++ ++enum bpf_stack_slot_type { ++ STACK_INVALID = 0, ++ STACK_SPILL = 1, ++ STACK_MISC = 2, ++ STACK_ZERO = 3, ++ STACK_DYNPTR = 4, ++ STACK_ITER = 5, ++}; ++ ++struct bpf_stack_state { ++ struct bpf_reg_state spilled_ptr; ++ u8 slot_type[8]; ++}; ++ ++struct bpf_reference_state { ++ int id; ++ int insn_idx; ++ int callback_ref; ++}; ++ ++struct bpf_idx_pair { ++ u32 prev_idx; ++ u32 idx; ++}; ++ ++struct bpf_verifier_state { ++ struct bpf_func_state *frame[8]; ++ struct bpf_verifier_state *parent; ++ u32 branches; ++ u32 insn_idx; ++ u32 curframe; ++ struct bpf_active_lock active_lock; ++ bool speculative; ++ bool active_rcu_lock; ++ u32 first_insn_idx; ++ u32 last_insn_idx; ++ struct bpf_idx_pair *jmp_history; ++ u32 jmp_history_cnt; ++}; ++ ++struct bpf_verifier_state_list { ++ struct bpf_verifier_state state; ++ struct bpf_verifier_state_list *next; ++ int miss_cnt; ++ int hit_cnt; ++}; ++ ++struct bpf_loop_inline_state { ++ unsigned int initialized: 1; ++ unsigned int fit_for_inline: 1; ++ u32 callback_subprogno; ++}; ++ ++struct bpf_insn_aux_data { ++ union { ++ enum bpf_reg_type ptr_type; ++ long unsigned int map_ptr_state; ++ s32 call_imm; ++ u32 alu_limit; ++ struct { ++ u32 map_index; ++ u32 map_off; ++ }; ++ struct { ++ enum bpf_reg_type reg_type; ++ union { ++ struct { ++ struct btf *btf; ++ u32 btf_id; ++ }; ++ u32 mem_size; ++ }; ++ } btf_var; ++ struct bpf_loop_inline_state loop_inline_state; ++ }; ++ union { ++ u64 obj_new_size; ++ u64 insert_off; ++ }; ++ struct btf_struct_meta *kptr_struct_meta; ++ u64 map_key_state; ++ int ctx_field_size; ++ u32 seen; ++ bool sanitize_stack_spill; ++ bool zext_dst; ++ bool storage_get_func_atomic; ++ bool is_iter_next; ++ u8 alu_state; ++ unsigned int orig_idx; ++ bool jmp_point; ++ bool prune_point; ++ bool force_checkpoint; ++}; ++ ++struct bpf_verifier_stack_elem { ++ struct bpf_verifier_state st; ++ int insn_idx; ++ int prev_insn_idx; ++ struct bpf_verifier_stack_elem *next; ++ u32 log_pos; ++}; ++ ++typedef void (*bpf_insn_print_t)(void *, const char *, ...); ++ ++typedef const char * (*bpf_insn_revmap_call_t)(void *, const struct bpf_insn *); ++ ++typedef const char * (*bpf_insn_print_imm_t)(void *, const struct bpf_insn *, __u64); ++ ++struct bpf_insn_cbs { ++ bpf_insn_print_t cb_print; ++ bpf_insn_revmap_call_t cb_call; ++ bpf_insn_print_imm_t cb_imm; ++ void *private_data; ++}; ++ ++struct bpf_call_arg_meta { ++ struct bpf_map *map_ptr; ++ bool raw_mode; ++ bool pkt_access; ++ u8 release_regno; ++ int regno; ++ int access_size; ++ int mem_size; ++ u64 msize_max_value; ++ int ref_obj_id; ++ int dynptr_id; ++ int map_uid; ++ int func_id; ++ struct btf *btf; ++ u32 btf_id; ++ struct btf *ret_btf; ++ u32 ret_btf_id; ++ u32 subprogno; ++ struct btf_field *kptr_field; ++}; ++ ++struct bpf_kfunc_call_arg_meta { ++ struct btf *btf; ++ u32 func_id; ++ u32 kfunc_flags; ++ const struct btf_type *func_proto; ++ const char *func_name; ++ u32 ref_obj_id; ++ u8 release_regno; ++ bool r0_rdonly; ++ u32 ret_btf_id; ++ u64 r0_size; ++ u32 subprogno; ++ struct { ++ u64 value; ++ bool found; ++ } arg_constant; ++ struct btf *arg_btf; ++ u32 arg_btf_id; ++ bool arg_owning_ref; ++ struct { ++ struct btf_field *field; ++ } arg_list_head; ++ struct { ++ struct btf_field *field; ++ } arg_rbtree_root; ++ struct { ++ enum bpf_dynptr_type type; ++ u32 id; ++ } initialized_dynptr; ++ struct { ++ u8 spi; ++ u8 frameno; ++ } iter; ++ u64 mem_size; ++}; ++ ++enum reg_arg_type { ++ SRC_OP = 0, ++ DST_OP = 1, ++ DST_OP_NO_MARK = 2, ++}; ++ ++enum bpf_access_src { ++ ACCESS_DIRECT = 1, ++ ACCESS_HELPER = 2, ++}; ++ ++struct task_struct__safe_rcu { ++ const cpumask_t *cpus_ptr; ++ struct css_set *cgroups; ++ struct task_struct *real_parent; ++ struct task_struct *group_leader; ++}; ++ ++struct cgroup__safe_rcu { ++ struct kernfs_node *kn; ++}; ++ ++struct css_set__safe_rcu { ++ struct cgroup *dfl_cgrp; ++}; ++ ++struct mm_struct__safe_rcu_or_null { ++ struct file *exe_file; ++}; ++ ++struct sk_buff__safe_rcu_or_null { ++ struct sock *sk; ++}; ++ ++struct request_sock__safe_rcu_or_null { ++ struct sock *sk; ++}; ++ ++struct bpf_iter_meta__safe_trusted { ++ struct seq_file *seq; ++}; ++ ++struct bpf_iter__task__safe_trusted { ++ struct bpf_iter_meta *meta; ++ struct task_struct *task; ++}; ++ ++struct linux_binprm__safe_trusted { ++ struct file *file; ++}; ++ ++struct file__safe_trusted { ++ struct inode *f_inode; ++}; ++ ++struct dentry__safe_trusted { ++ struct inode *d_inode; ++}; ++ ++struct socket__safe_trusted { ++ struct sock *sk; ++}; ++ ++struct bpf_reg_types { ++ const enum bpf_reg_type types[10]; ++ u32 *btf_id; ++}; ++ ++enum { ++ AT_PKT_END = -1, ++ BEYOND_PKT_END = -2, ++}; ++ ++typedef int (*set_callee_state_fn)(struct bpf_verifier_env *, struct bpf_func_state *, struct bpf_func_state *, int); ++ ++enum { ++ KF_ARG_DYNPTR_ID = 0, ++ KF_ARG_LIST_HEAD_ID = 1, ++ KF_ARG_LIST_NODE_ID = 2, ++ KF_ARG_RB_ROOT_ID = 3, ++ KF_ARG_RB_NODE_ID = 4, ++}; ++ ++enum kfunc_ptr_arg_type { ++ KF_ARG_PTR_TO_CTX = 0, ++ KF_ARG_PTR_TO_ALLOC_BTF_ID = 1, ++ KF_ARG_PTR_TO_REFCOUNTED_KPTR = 2, ++ KF_ARG_PTR_TO_DYNPTR = 3, ++ KF_ARG_PTR_TO_ITER = 4, ++ KF_ARG_PTR_TO_LIST_HEAD = 5, ++ KF_ARG_PTR_TO_LIST_NODE = 6, ++ KF_ARG_PTR_TO_BTF_ID = 7, ++ KF_ARG_PTR_TO_MEM = 8, ++ KF_ARG_PTR_TO_MEM_SIZE = 9, ++ KF_ARG_PTR_TO_CALLBACK = 10, ++ KF_ARG_PTR_TO_RB_ROOT = 11, ++ KF_ARG_PTR_TO_RB_NODE = 12, ++}; ++ ++enum special_kfunc_type { ++ KF_bpf_obj_new_impl = 0, ++ KF_bpf_obj_drop_impl = 1, ++ KF_bpf_refcount_acquire_impl = 2, ++ KF_bpf_list_push_front_impl = 3, ++ KF_bpf_list_push_back_impl = 4, ++ KF_bpf_list_pop_front = 5, ++ KF_bpf_list_pop_back = 6, ++ KF_bpf_cast_to_kern_ctx = 7, ++ KF_bpf_rdonly_cast = 8, ++ KF_bpf_rcu_read_lock = 9, ++ KF_bpf_rcu_read_unlock = 10, ++ KF_bpf_rbtree_remove = 11, ++ KF_bpf_rbtree_add_impl = 12, ++ KF_bpf_rbtree_first = 13, ++ KF_bpf_dynptr_from_skb = 14, ++ KF_bpf_dynptr_from_xdp = 15, ++ KF_bpf_dynptr_slice = 16, ++ KF_bpf_dynptr_slice_rdwr = 17, ++}; ++ ++enum { ++ REASON_BOUNDS = -1, ++ REASON_TYPE = -2, ++ REASON_PATHS = -3, ++ REASON_LIMIT = -4, ++ REASON_STACK = -5, ++}; ++ ++struct bpf_sanitize_info { ++ struct bpf_insn_aux_data aux; ++ bool mask_to_left; ++}; ++ ++enum { ++ DISCOVERED = 16, ++ EXPLORED = 32, ++ FALLTHROUGH = 1, ++ BRANCH = 2, ++}; ++ ++enum { ++ DONE_EXPLORING = 0, ++ KEEP_EXPLORING = 1, ++}; ++ ++struct bpf_iter; ++ ++union bpf_iter_link_info { ++ struct { ++ __u32 map_fd; ++ } map; ++ struct { ++ enum bpf_cgroup_iter_order order; ++ __u32 cgroup_fd; ++ __u64 cgroup_id; ++ } cgroup; ++ struct { ++ __u32 tid; ++ __u32 pid; ++ __u32 pid_fd; ++ } task; ++}; ++ ++enum bpf_link_type { ++ BPF_LINK_TYPE_UNSPEC = 0, ++ BPF_LINK_TYPE_RAW_TRACEPOINT = 1, ++ BPF_LINK_TYPE_TRACING = 2, ++ BPF_LINK_TYPE_CGROUP = 3, ++ BPF_LINK_TYPE_ITER = 4, ++ BPF_LINK_TYPE_NETNS = 5, ++ BPF_LINK_TYPE_XDP = 6, ++ BPF_LINK_TYPE_PERF_EVENT = 7, ++ BPF_LINK_TYPE_KPROBE_MULTI = 8, ++ BPF_LINK_TYPE_STRUCT_OPS = 9, ++ BPF_LINK_TYPE_NETFILTER = 10, ++ MAX_BPF_LINK_TYPE = 11, ++}; ++ ++struct bpf_link_info { ++ __u32 type; ++ __u32 id; ++ __u32 prog_id; ++ union { ++ struct { ++ __u64 tp_name; ++ __u32 tp_name_len; ++ } raw_tracepoint; ++ struct { ++ __u32 attach_type; ++ __u32 target_obj_id; ++ __u32 target_btf_id; ++ } tracing; ++ struct { ++ __u64 cgroup_id; ++ __u32 attach_type; ++ } cgroup; ++ struct { ++ __u64 target_name; ++ __u32 target_name_len; ++ union { ++ struct { ++ __u32 map_id; ++ } map; ++ }; ++ union { ++ struct { ++ __u64 cgroup_id; ++ __u32 order; ++ } cgroup; ++ struct { ++ __u32 tid; ++ __u32 pid; ++ } task; ++ }; ++ } iter; ++ struct { ++ __u32 netns_ino; ++ __u32 attach_type; ++ } netns; ++ struct { ++ __u32 ifindex; ++ } xdp; ++ struct { ++ __u32 map_id; ++ } struct_ops; ++ struct { ++ __u32 pf; ++ __u32 hooknum; ++ __s32 priority; ++ __u32 flags; ++ } netfilter; ++ }; ++}; ++ ++struct bpf_iter_num { ++ __u64 __opaque[1]; ++}; ++ ++typedef unsigned int (*bpf_func_t)(const void *, const struct bpf_insn *); ++ ++struct bpf_link_ops; ++ ++struct bpf_link { ++ atomic64_t refcnt; ++ u32 id; ++ enum bpf_link_type type; ++ const struct bpf_link_ops *ops; ++ struct bpf_prog *prog; ++ struct work_struct work; ++}; ++ ++struct bpf_link_ops { ++ void (*release)(struct bpf_link *); ++ void (*dealloc)(struct bpf_link *); ++ int (*detach)(struct bpf_link *); ++ int (*update_prog)(struct bpf_link *, struct bpf_prog *, struct bpf_prog *); ++ void (*show_fdinfo)(const struct bpf_link *, struct seq_file *); ++ int (*fill_link_info)(const struct bpf_link *, struct bpf_link_info *); ++ int (*update_map)(struct bpf_link *, struct bpf_map *, struct bpf_map *); ++}; ++ ++struct bpf_link_primer { ++ struct bpf_link *link; ++ struct file *file; ++ int fd; ++ u32 id; ++}; ++ ++typedef int (*bpf_iter_attach_target_t)(struct bpf_prog *, union bpf_iter_link_info *, struct bpf_iter_aux_info *); ++ ++typedef void (*bpf_iter_detach_target_t)(struct bpf_iter_aux_info *); ++ ++typedef void (*bpf_iter_show_fdinfo_t)(const struct bpf_iter_aux_info *, struct seq_file *); ++ ++typedef int (*bpf_iter_fill_link_info_t)(const struct bpf_iter_aux_info *, struct bpf_link_info *); ++ ++typedef const struct bpf_func_proto * (*bpf_iter_get_func_proto_t)(enum bpf_func_id, const struct bpf_prog *); ++ ++enum bpf_iter_feature { ++ BPF_ITER_RESCHED = 1, ++}; ++ ++struct bpf_iter_reg { ++ const char *target; ++ bpf_iter_attach_target_t attach_target; ++ bpf_iter_detach_target_t detach_target; ++ bpf_iter_show_fdinfo_t show_fdinfo; ++ bpf_iter_fill_link_info_t fill_link_info; ++ bpf_iter_get_func_proto_t get_func_proto; ++ u32 ctx_arg_info_size; ++ u32 feature; ++ struct bpf_ctx_arg_aux ctx_arg_info[2]; ++ const struct bpf_iter_seq_info *seq_info; ++}; ++ ++typedef unsigned int (*bpf_dispatcher_fn)(const void *, const struct bpf_insn *, unsigned int (*)(const void *, const struct bpf_insn *)); ++ ++struct bpf_iter_target_info { ++ struct list_head list; ++ const struct bpf_iter_reg *reg_info; ++ u32 btf_id; ++}; ++ ++struct bpf_iter_link { ++ struct bpf_link link; ++ struct bpf_iter_aux_info aux; ++ struct bpf_iter_target_info *tinfo; ++}; ++ ++struct bpf_iter_priv_data { ++ struct bpf_iter_target_info *tinfo; ++ const struct bpf_iter_seq_info *seq_info; ++ struct bpf_prog *prog; ++ u64 session_id; ++ u64 seq_num; ++ bool done_stop; ++ long: 0; ++ u8 target_private[0]; ++}; ++ ++typedef u64 (*btf_bpf_for_each_map_elem)(struct bpf_map *, void *, void *, u64); ++ ++typedef u64 (*btf_bpf_loop)(u32, void *, void *, u64); ++ ++struct bpf_iter_num_kern { ++ int cur; ++ int end; ++}; ++ ++struct btf_iter_num; ++ ++struct bpf_iter_seq_link_info { ++ u32 link_id; ++}; ++ ++struct bpf_iter__bpf_link { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct bpf_link *link; ++ }; ++}; ++ ++struct pcpu_freelist_node; ++ ++struct pcpu_freelist_head { ++ struct pcpu_freelist_node *first; ++ raw_spinlock_t lock; ++}; ++ ++struct pcpu_freelist_node { ++ struct pcpu_freelist_node *next; ++}; ++ ++struct pcpu_freelist { ++ struct pcpu_freelist_head *freelist; ++ struct pcpu_freelist_head extralist; ++}; ++ ++enum bpf_lru_list_type { ++ BPF_LRU_LIST_T_ACTIVE = 0, ++ BPF_LRU_LIST_T_INACTIVE = 1, ++ BPF_LRU_LIST_T_FREE = 2, ++ BPF_LRU_LOCAL_LIST_T_FREE = 3, ++ BPF_LRU_LOCAL_LIST_T_PENDING = 4, ++}; ++ ++struct bpf_lru_node { ++ struct list_head list; ++ u16 cpu; ++ u8 type; ++ u8 ref; ++}; ++ ++struct bpf_lru_list { ++ struct list_head lists[3]; ++ unsigned int counts[2]; ++ struct list_head *next_inactive_rotation; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ raw_spinlock_t lock; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_lru_locallist { ++ struct list_head lists[2]; ++ u16 next_steal; ++ raw_spinlock_t lock; ++}; ++ ++struct bpf_common_lru { ++ struct bpf_lru_list lru_list; ++ struct bpf_lru_locallist *local_list; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++typedef bool (*del_from_htab_func)(void *, struct bpf_lru_node *); ++ ++struct bpf_lru { ++ union { ++ struct bpf_common_lru common_lru; ++ struct bpf_lru_list *percpu_lru; ++ }; ++ del_from_htab_func del_from_htab; ++ void *del_arg; ++ unsigned int hash_offset; ++ unsigned int nr_scans; ++ bool percpu; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_lpm_trie_key { ++ __u32 prefixlen; ++ __u8 data[0]; ++}; ++ ++enum { ++ BPF_ANY = 0, ++ BPF_NOEXIST = 1, ++ BPF_EXIST = 2, ++ BPF_F_LOCK = 4, ++}; ++ ++struct lpm_trie_node { ++ struct callback_head rcu; ++ struct lpm_trie_node *child[2]; ++ u32 prefixlen; ++ u32 flags; ++ u8 data[0]; ++}; ++ ++struct lpm_trie { ++ struct bpf_map map; ++ struct lpm_trie_node *root; ++ size_t n_entries; ++ size_t max_prefixlen; ++ size_t data_size; ++ spinlock_t lock; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum bpf_cgroup_storage_type { ++ BPF_CGROUP_STORAGE_SHARED = 0, ++ BPF_CGROUP_STORAGE_PERCPU = 1, ++ __BPF_CGROUP_STORAGE_MAX = 2, ++}; ++ ++enum { ++ BPF_MAX_TRAMP_LINKS = 27, ++}; ++ ++enum bpf_tramp_prog_type { ++ BPF_TRAMP_FENTRY = 0, ++ BPF_TRAMP_FEXIT = 1, ++ BPF_TRAMP_MODIFY_RETURN = 2, ++ BPF_TRAMP_MAX = 3, ++ BPF_TRAMP_REPLACE = 4, ++}; ++ ++struct bpf_queue_stack { ++ struct bpf_map map; ++ raw_spinlock_t lock; ++ u32 head; ++ u32 tail; ++ u32 size; ++ char elements[0]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_local_storage_data; ++ ++struct bpf_local_storage { ++ struct bpf_local_storage_data *cache[16]; ++ struct bpf_local_storage_map *smap; ++ struct hlist_head list; ++ void *owner; ++ struct callback_head rcu; ++ raw_spinlock_t lock; ++}; ++ ++struct bpf_mem_caches; ++ ++struct bpf_mem_cache; ++ ++struct bpf_mem_alloc { ++ struct bpf_mem_caches *caches; ++ struct bpf_mem_cache *cache; ++ struct work_struct work; ++}; ++ ++struct bpf_local_storage_map_bucket; ++ ++struct bpf_local_storage_map { ++ struct bpf_map map; ++ struct bpf_local_storage_map_bucket *buckets; ++ u32 bucket_log; ++ u16 elem_size; ++ u16 cache_idx; ++ struct bpf_mem_alloc selem_ma; ++ struct bpf_mem_alloc storage_ma; ++ bool bpf_ma; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_local_storage_map_bucket { ++ struct hlist_head list; ++ raw_spinlock_t lock; ++}; ++ ++struct bpf_local_storage_data { ++ struct bpf_local_storage_map *smap; ++ u8 data[0]; ++}; ++ ++struct bpf_local_storage_elem { ++ struct hlist_node map_node; ++ struct hlist_node snode; ++ struct bpf_local_storage *local_storage; ++ struct callback_head rcu; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct bpf_local_storage_data sdata; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_local_storage_cache { ++ spinlock_t idx_lock; ++ u64 idx_usage_counts[16]; ++}; ++ ++typedef void (*swap_r_func_t)(void *, void *, int, const void *); ++ ++typedef int (*cmp_r_func_t)(const void *, const void *, const void *); ++ ++struct btf_enum { ++ __u32 name_off; ++ __s32 val; ++}; ++ ++enum { ++ BTF_VAR_STATIC = 0, ++ BTF_VAR_GLOBAL_ALLOCATED = 1, ++ BTF_VAR_GLOBAL_EXTERN = 2, ++}; ++ ++struct btf_var { ++ __u32 linkage; ++}; ++ ++struct btf_decl_tag { ++ __s32 component_idx; ++}; ++ ++struct btf_enum64 { ++ __u32 name_off; ++ __u32 val_lo32; ++ __u32 val_hi32; ++}; ++ ++struct bpf_flow_keys { ++ __u16 nhoff; ++ __u16 thoff; ++ __u16 addr_proto; ++ __u8 is_frag; ++ __u8 is_first_frag; ++ __u8 is_encap; ++ __u8 ip_proto; ++ __be16 n_proto; ++ __be16 sport; ++ __be16 dport; ++ union { ++ struct { ++ __be32 ipv4_src; ++ __be32 ipv4_dst; ++ }; ++ struct { ++ __u32 ipv6_src[4]; ++ __u32 ipv6_dst[4]; ++ }; ++ }; ++ __u32 flags; ++ __be32 flow_label; ++}; ++ ++struct bpf_sock { ++ __u32 bound_dev_if; ++ __u32 family; ++ __u32 type; ++ __u32 protocol; ++ __u32 mark; ++ __u32 priority; ++ __u32 src_ip4; ++ __u32 src_ip6[4]; ++ __u32 src_port; ++ __be16 dst_port; ++ __u32 dst_ip4; ++ __u32 dst_ip6[4]; ++ __u32 state; ++ __s32 rx_queue_mapping; ++}; ++ ++struct __sk_buff { ++ __u32 len; ++ __u32 pkt_type; ++ __u32 mark; ++ __u32 queue_mapping; ++ __u32 protocol; ++ __u32 vlan_present; ++ __u32 vlan_tci; ++ __u32 vlan_proto; ++ __u32 priority; ++ __u32 ingress_ifindex; ++ __u32 ifindex; ++ __u32 tc_index; ++ __u32 cb[5]; ++ __u32 hash; ++ __u32 tc_classid; ++ __u32 data; ++ __u32 data_end; ++ __u32 napi_id; ++ __u32 family; ++ __u32 remote_ip4; ++ __u32 local_ip4; ++ __u32 remote_ip6[4]; ++ __u32 local_ip6[4]; ++ __u32 remote_port; ++ __u32 local_port; ++ __u32 data_meta; ++ union { ++ struct bpf_flow_keys *flow_keys; ++ }; ++ __u64 tstamp; ++ __u32 wire_len; ++ __u32 gso_segs; ++ union { ++ struct bpf_sock *sk; ++ }; ++ __u32 gso_size; ++ __u8 tstamp_type; ++ __u64 hwtstamp; ++}; ++ ++struct sk_msg_md { ++ union { ++ void *data; ++ }; ++ union { ++ void *data_end; ++ }; ++ __u32 family; ++ __u32 remote_ip4; ++ __u32 local_ip4; ++ __u32 remote_ip6[4]; ++ __u32 local_ip6[4]; ++ __u32 remote_port; ++ __u32 local_port; ++ __u32 size; ++ union { ++ struct bpf_sock *sk; ++ }; ++}; ++ ++struct sk_reuseport_md { ++ union { ++ void *data; ++ }; ++ union { ++ void *data_end; ++ }; ++ __u32 len; ++ __u32 eth_protocol; ++ __u32 ip_protocol; ++ __u32 bind_inany; ++ __u32 hash; ++ union { ++ struct bpf_sock *sk; ++ }; ++ union { ++ struct bpf_sock *migrating_sk; ++ }; ++}; ++ ++struct bpf_btf_info { ++ __u64 btf; ++ __u32 btf_size; ++ __u32 id; ++ __u64 name; ++ __u32 name_len; ++ __u32 kernel_btf; ++}; ++ ++struct bpf_sock_addr { ++ __u32 user_family; ++ __u32 user_ip4; ++ __u32 user_ip6[4]; ++ __u32 user_port; ++ __u32 family; ++ __u32 type; ++ __u32 protocol; ++ __u32 msg_src_ip4; ++ __u32 msg_src_ip6[4]; ++ union { ++ struct bpf_sock *sk; ++ }; ++}; ++ ++struct bpf_sock_ops { ++ __u32 op; ++ union { ++ __u32 args[4]; ++ __u32 reply; ++ __u32 replylong[4]; ++ }; ++ __u32 family; ++ __u32 remote_ip4; ++ __u32 local_ip4; ++ __u32 remote_ip6[4]; ++ __u32 local_ip6[4]; ++ __u32 remote_port; ++ __u32 local_port; ++ __u32 is_fullsock; ++ __u32 snd_cwnd; ++ __u32 srtt_us; ++ __u32 bpf_sock_ops_cb_flags; ++ __u32 state; ++ __u32 rtt_min; ++ __u32 snd_ssthresh; ++ __u32 rcv_nxt; ++ __u32 snd_nxt; ++ __u32 snd_una; ++ __u32 mss_cache; ++ __u32 ecn_flags; ++ __u32 rate_delivered; ++ __u32 rate_interval_us; ++ __u32 packets_out; ++ __u32 retrans_out; ++ __u32 total_retrans; ++ __u32 segs_in; ++ __u32 data_segs_in; ++ __u32 segs_out; ++ __u32 data_segs_out; ++ __u32 lost_out; ++ __u32 sacked_out; ++ __u32 sk_txhash; ++ __u64 bytes_received; ++ __u64 bytes_acked; ++ union { ++ struct bpf_sock *sk; ++ }; ++ union { ++ void *skb_data; ++ }; ++ union { ++ void *skb_data_end; ++ }; ++ __u32 skb_len; ++ __u32 skb_tcp_flags; ++ __u64 skb_hwtstamp; ++}; ++ ++struct bpf_cgroup_dev_ctx { ++ __u32 access_type; ++ __u32 major; ++ __u32 minor; ++}; ++ ++struct bpf_raw_tracepoint_args { ++ __u64 args[0]; ++}; ++ ++struct bpf_sysctl { ++ __u32 write; ++ __u32 file_pos; ++}; ++ ++struct bpf_sockopt { ++ union { ++ struct bpf_sock *sk; ++ }; ++ union { ++ void *optval; ++ }; ++ union { ++ void *optval_end; ++ }; ++ __s32 level; ++ __s32 optname; ++ __s32 optlen; ++ __s32 retval; ++}; ++ ++struct bpf_sk_lookup { ++ union { ++ union { ++ struct bpf_sock *sk; ++ }; ++ __u64 cookie; ++ }; ++ __u32 family; ++ __u32 protocol; ++ __u32 remote_ip4; ++ __u32 remote_ip6[4]; ++ __be16 remote_port; ++ __u32 local_ip4; ++ __u32 local_ip6[4]; ++ __u32 local_port; ++ __u32 ingress_ifindex; ++}; ++ ++enum { ++ BTF_F_COMPACT = 1, ++ BTF_F_NONAME = 2, ++ BTF_F_PTR_RAW = 4, ++ BTF_F_ZERO = 8, ++}; ++ ++typedef user_pt_regs bpf_user_pt_regs_t; ++ ++struct bpf_perf_event_data { ++ bpf_user_pt_regs_t regs; ++ __u64 sample_period; ++ __u64 addr; ++}; ++ ++typedef struct {} local_lock_t; ++ ++struct xa_node { ++ unsigned char shift; ++ unsigned char offset; ++ unsigned char count; ++ unsigned char nr_values; ++ struct xa_node *parent; ++ struct xarray *array; ++ union { ++ struct list_head private_list; ++ struct callback_head callback_head; ++ }; ++ void *slots[64]; ++ union { ++ long unsigned int tags[3]; ++ long unsigned int marks[3]; ++ }; ++}; ++ ++struct radix_tree_preload { ++ local_lock_t lock; ++ unsigned int nr; ++ struct xa_node *nodes; ++}; ++ ++struct btf_id_set8 { ++ u32 cnt; ++ u32 flags; ++ struct { ++ u32 id; ++ u32 flags; ++ } pairs[0]; ++}; ++ ++struct btf_kfunc_id_set { ++ struct module *owner; ++ struct btf_id_set8 *set; ++}; ++ ++struct btf_id_dtor_kfunc { ++ u32 btf_id; ++ u32 kfunc_btf_id; ++}; ++ ++struct btf_struct_metas { ++ u32 cnt; ++ struct btf_struct_meta types[0]; ++}; ++ ++enum { ++ BTF_FIELDS_MAX = 10, ++}; ++ ++struct sk_reuseport_kern { ++ struct sk_buff *skb; ++ struct sock *sk; ++ struct sock *selected_sk; ++ struct sock *migrating_sk; ++ void *data_end; ++ u32 hash; ++ u32 reuseport_id; ++ bool bind_inany; ++}; ++ ++struct bpf_flow_dissector { ++ struct bpf_flow_keys *flow_keys; ++ const struct sk_buff *skb; ++ const void *data; ++ const void *data_end; ++}; ++ ++struct inet_ehash_bucket; ++ ++struct inet_bind_hashbucket; ++ ++struct inet_listen_hashbucket; ++ ++struct inet_hashinfo { ++ struct inet_ehash_bucket *ehash; ++ spinlock_t *ehash_locks; ++ unsigned int ehash_mask; ++ unsigned int ehash_locks_mask; ++ struct kmem_cache *bind_bucket_cachep; ++ struct inet_bind_hashbucket *bhash; ++ struct kmem_cache *bind2_bucket_cachep; ++ struct inet_bind_hashbucket *bhash2; ++ unsigned int bhash_size; ++ unsigned int lhash2_mask; ++ struct inet_listen_hashbucket *lhash2; ++ bool pernet; ++}; ++ ++struct ip_ra_chain { ++ struct ip_ra_chain *next; ++ struct sock *sk; ++ union { ++ void (*destructor)(struct sock *); ++ struct sock *saved_sk; ++ }; ++ struct callback_head rcu; ++}; ++ ++struct fib_table { ++ struct hlist_node tb_hlist; ++ u32 tb_id; ++ int tb_num_default; ++ struct callback_head rcu; ++ long unsigned int *tb_data; ++ long unsigned int __data[0]; ++}; ++ ++struct inet_peer_base { ++ struct rb_root rb_root; ++ seqlock_t lock; ++ int total; ++}; ++ ++struct tcp_fastopen_context { ++ siphash_key_t key[2]; ++ int num; ++ struct callback_head rcu; ++}; ++ ++struct __una_u32 { ++ u32 x; ++}; ++ ++struct bpf_sock_addr_kern { ++ struct sock *sk; ++ struct sockaddr *uaddr; ++ u64 tmp_reg; ++ void *t_ctx; ++}; ++ ++struct bpf_sock_ops_kern { ++ struct sock *sk; ++ union { ++ u32 args[4]; ++ u32 reply; ++ u32 replylong[4]; ++ }; ++ struct sk_buff *syn_skb; ++ struct sk_buff *skb; ++ void *skb_data_end; ++ u8 op; ++ u8 is_fullsock; ++ u8 remaining_opt_len; ++ u64 temp; ++}; ++ ++struct bpf_sysctl_kern { ++ struct ctl_table_header *head; ++ struct ctl_table *table; ++ void *cur_val; ++ size_t cur_len; ++ void *new_val; ++ size_t new_len; ++ int new_updated; ++ int write; ++ loff_t *ppos; ++ u64 tmp_reg; ++}; ++ ++struct bpf_sockopt_kern { ++ struct sock *sk; ++ u8 *optval; ++ u8 *optval_end; ++ s32 level; ++ s32 optname; ++ s32 optlen; ++ struct task_struct *current_task; ++ u64 tmp_reg; ++}; ++ ++struct bpf_sk_lookup_kern { ++ u16 family; ++ u16 protocol; ++ __be16 sport; ++ u16 dport; ++ struct { ++ __be32 saddr; ++ __be32 daddr; ++ } v4; ++ struct { ++ const struct in6_addr *saddr; ++ const struct in6_addr *daddr; ++ } v6; ++ struct sock *selected_sk; ++ u32 ingress_ifindex; ++ bool no_reuseport; ++}; ++ ++struct lwtunnel_state { ++ __u16 type; ++ __u16 flags; ++ __u16 headroom; ++ atomic_t refcnt; ++ int (*orig_output)(struct net *, struct sock *, struct sk_buff *); ++ int (*orig_input)(struct sk_buff *); ++ struct callback_head rcu; ++ __u8 data[0]; ++}; ++ ++struct sock_reuseport { ++ struct callback_head rcu; ++ u16 max_socks; ++ u16 num_socks; ++ u16 num_closed_socks; ++ u16 incoming_cpu; ++ unsigned int synq_overflow_ts; ++ unsigned int reuseport_id; ++ unsigned int bind_inany: 1; ++ unsigned int has_conns: 1; ++ struct bpf_prog *prog; ++ struct sock *socks[0]; ++}; ++ ++struct sk_psock_progs { ++ struct bpf_prog *msg_parser; ++ struct bpf_prog *stream_parser; ++ struct bpf_prog *stream_verdict; ++ struct bpf_prog *skb_verdict; ++}; ++ ++struct strp_stats { ++ long long unsigned int msgs; ++ long long unsigned int bytes; ++ unsigned int mem_fail; ++ unsigned int need_more_hdr; ++ unsigned int msg_too_big; ++ unsigned int msg_timeouts; ++ unsigned int bad_hdr_len; ++}; ++ ++struct strparser; ++ ++struct strp_callbacks { ++ int (*parse_msg)(struct strparser *, struct sk_buff *); ++ void (*rcv_msg)(struct strparser *, struct sk_buff *); ++ int (*read_sock_done)(struct strparser *, int); ++ void (*abort_parser)(struct strparser *, int); ++ void (*lock)(struct strparser *); ++ void (*unlock)(struct strparser *); ++}; ++ ++struct strparser { ++ struct sock *sk; ++ u32 stopped: 1; ++ u32 paused: 1; ++ u32 aborted: 1; ++ u32 interrupted: 1; ++ u32 unrecov_intr: 1; ++ struct sk_buff **skb_nextp; ++ struct sk_buff *skb_head; ++ unsigned int need_bytes; ++ struct delayed_work msg_timer_work; ++ struct work_struct work; ++ struct strp_stats stats; ++ struct strp_callbacks cb; ++}; ++ ++struct sk_psock_work_state { ++ u32 len; ++ u32 off; ++}; ++ ++struct sk_msg; ++ ++struct sk_psock { ++ struct sock *sk; ++ struct sock *sk_redir; ++ u32 apply_bytes; ++ u32 cork_bytes; ++ u32 eval; ++ bool redir_ingress; ++ struct sk_msg *cork; ++ struct sk_psock_progs progs; ++ struct strparser strp; ++ struct sk_buff_head ingress_skb; ++ struct list_head ingress_msg; ++ spinlock_t ingress_lock; ++ long unsigned int state; ++ struct list_head link; ++ spinlock_t link_lock; ++ refcount_t refcnt; ++ void (*saved_unhash)(struct sock *); ++ void (*saved_destroy)(struct sock *); ++ void (*saved_close)(struct sock *, long int); ++ void (*saved_write_space)(struct sock *); ++ void (*saved_data_ready)(struct sock *); ++ int (*psock_update_sk_prot)(struct sock *, struct sk_psock *, bool); ++ struct proto *sk_proto; ++ struct mutex work_mutex; ++ struct sk_psock_work_state work_state; ++ struct delayed_work work; ++ struct rcu_work rwork; ++}; ++ ++enum { ++ __ND_OPT_PREFIX_INFO_END = 0, ++ ND_OPT_SOURCE_LL_ADDR = 1, ++ ND_OPT_TARGET_LL_ADDR = 2, ++ ND_OPT_PREFIX_INFO = 3, ++ ND_OPT_REDIRECT_HDR = 4, ++ ND_OPT_MTU = 5, ++ ND_OPT_NONCE = 14, ++ __ND_OPT_ARRAY_MAX = 15, ++ ND_OPT_ROUTE_INFO = 24, ++ ND_OPT_RDNSS = 25, ++ ND_OPT_DNSSL = 31, ++ ND_OPT_6CO = 34, ++ ND_OPT_CAPTIVE_PORTAL = 37, ++ ND_OPT_PREF64 = 38, ++ __ND_OPT_MAX = 39, ++}; ++ ++struct nd_opt_hdr { ++ __u8 nd_opt_type; ++ __u8 nd_opt_len; ++}; ++ ++struct ndisc_options { ++ struct nd_opt_hdr *nd_opt_array[15]; ++ struct nd_opt_hdr *nd_opts_ri; ++ struct nd_opt_hdr *nd_opts_ri_end; ++ struct nd_opt_hdr *nd_useropts; ++ struct nd_opt_hdr *nd_useropts_end; ++ struct nd_opt_hdr *nd_802154_opt_array[3]; ++}; ++ ++struct prefix_info { ++ __u8 type; ++ __u8 length; ++ __u8 prefix_len; ++ __u8 onlink: 1; ++ __u8 autoconf: 1; ++ __u8 reserved: 6; ++ __be32 valid; ++ __be32 prefered; ++ __be32 reserved2; ++ struct in6_addr prefix; ++}; ++ ++struct inet_ehash_bucket { ++ struct hlist_nulls_head chain; ++}; ++ ++struct inet_bind_hashbucket { ++ spinlock_t lock; ++ struct hlist_head chain; ++}; ++ ++struct inet_listen_hashbucket { ++ spinlock_t lock; ++ struct hlist_nulls_head nulls_head; ++}; ++ ++struct ack_sample { ++ u32 pkts_acked; ++ s32 rtt_us; ++ u32 in_flight; ++}; ++ ++struct rate_sample { ++ u64 prior_mstamp; ++ u32 prior_delivered; ++ u32 prior_delivered_ce; ++ s32 delivered; ++ s32 delivered_ce; ++ long int interval_us; ++ u32 snd_interval_us; ++ u32 rcv_interval_us; ++ long int rtt_us; ++ int losses; ++ u32 acked_sacked; ++ u32 prior_in_flight; ++ u32 last_end_seq; ++ bool is_app_limited; ++ bool is_retrans; ++ bool is_ack_delayed; ++}; ++ ++struct sk_msg_sg { ++ u32 start; ++ u32 curr; ++ u32 end; ++ u32 size; ++ u32 copybreak; ++ long unsigned int copy[1]; ++ struct scatterlist data[19]; ++}; ++ ++struct sk_msg { ++ struct sk_msg_sg sg; ++ void *data; ++ void *data_end; ++ u32 apply_bytes; ++ u32 cork_bytes; ++ u32 flags; ++ struct sk_buff *skb; ++ struct sock *sk_redir; ++ struct sock *sk; ++ struct list_head list; ++}; ++ ++struct bpf_perf_event_data_kern { ++ bpf_user_pt_regs_t *regs; ++ struct perf_sample_data *data; ++ struct perf_event *event; ++}; ++ ++struct nf_hook_state; ++ ++struct bpf_nf_ctx { ++ const struct nf_hook_state *state; ++ struct sk_buff *skb; ++}; ++ ++struct nf_hook_state { ++ u8 hook; ++ u8 pf; ++ struct net_device *in; ++ struct net_device *out; ++ struct sock *sk; ++ struct net *net; ++ int (*okfn)(struct net *, struct sock *, struct sk_buff *); ++}; ++ ++struct bpf_core_cand { ++ const struct btf *btf; ++ __u32 id; ++}; ++ ++struct bpf_core_cand_list { ++ struct bpf_core_cand *cands; ++ int len; ++}; ++ ++struct bpf_core_accessor { ++ __u32 type_id; ++ __u32 idx; ++ const char *name; ++}; ++ ++struct bpf_core_spec { ++ const struct btf *btf; ++ struct bpf_core_accessor spec[64]; ++ __u32 root_type_id; ++ enum bpf_core_relo_kind relo_kind; ++ int len; ++ int raw_spec[64]; ++ int raw_len; ++ __u32 bit_offset; ++}; ++ ++struct bpf_core_relo_res { ++ __u64 orig_val; ++ __u64 new_val; ++ bool poison; ++ bool validate; ++ bool fail_memsz_adjust; ++ __u32 orig_sz; ++ __u32 orig_type_id; ++ __u32 new_sz; ++ __u32 new_type_id; ++}; ++ ++enum btf_kfunc_hook { ++ BTF_KFUNC_HOOK_COMMON = 0, ++ BTF_KFUNC_HOOK_XDP = 1, ++ BTF_KFUNC_HOOK_TC = 2, ++ BTF_KFUNC_HOOK_STRUCT_OPS = 3, ++ BTF_KFUNC_HOOK_TRACING = 4, ++ BTF_KFUNC_HOOK_SYSCALL = 5, ++ BTF_KFUNC_HOOK_FMODRET = 6, ++ BTF_KFUNC_HOOK_CGROUP_SKB = 7, ++ BTF_KFUNC_HOOK_SCHED_ACT = 8, ++ BTF_KFUNC_HOOK_SK_SKB = 9, ++ BTF_KFUNC_HOOK_SOCKET_FILTER = 10, ++ BTF_KFUNC_HOOK_LWT = 11, ++ BTF_KFUNC_HOOK_NETFILTER = 12, ++ BTF_KFUNC_HOOK_MAX = 13, ++}; ++ ++enum { ++ BTF_KFUNC_SET_MAX_CNT = 256, ++ BTF_DTOR_KFUNC_MAX_CNT = 256, ++}; ++ ++struct btf_kfunc_set_tab { ++ struct btf_id_set8 *sets[13]; ++}; ++ ++struct btf_id_dtor_kfunc_tab { ++ u32 cnt; ++ struct btf_id_dtor_kfunc dtors[0]; ++}; ++ ++enum verifier_phase { ++ CHECK_META = 0, ++ CHECK_TYPE = 1, ++}; ++ ++struct resolve_vertex { ++ const struct btf_type *t; ++ u32 type_id; ++ u16 next_member; ++}; ++ ++enum visit_state { ++ NOT_VISITED = 0, ++ VISITED = 1, ++ RESOLVED = 2, ++}; ++ ++enum resolve_mode { ++ RESOLVE_TBD = 0, ++ RESOLVE_PTR = 1, ++ RESOLVE_STRUCT_OR_ARRAY = 2, ++}; ++ ++struct btf_sec_info { ++ u32 off; ++ u32 len; ++}; ++ ++struct btf_verifier_env { ++ struct btf *btf; ++ u8 *visit_states; ++ struct resolve_vertex stack[32]; ++ struct bpf_verifier_log log; ++ u32 log_type_id; ++ u32 top_stack; ++ enum verifier_phase phase; ++ enum resolve_mode resolve_mode; ++}; ++ ++struct btf_show { ++ u64 flags; ++ void *target; ++ void (*showfn)(struct btf_show *, const char *, struct __va_list_tag *); ++ const struct btf *btf; ++ struct { ++ u8 depth; ++ u8 depth_to_show; ++ u8 depth_check; ++ u8 array_member: 1; ++ u8 array_terminated: 1; ++ u16 array_encoding; ++ u32 type_id; ++ int status; ++ const struct btf_type *type; ++ const struct btf_member *member; ++ char name[80]; ++ } state; ++ struct { ++ u32 size; ++ void *head; ++ void *data; ++ u8 safe[32]; ++ } obj; ++}; ++ ++struct btf_kind_operations { ++ s32 (*check_meta)(struct btf_verifier_env *, const struct btf_type *, u32); ++ int (*resolve)(struct btf_verifier_env *, const struct resolve_vertex *); ++ int (*check_member)(struct btf_verifier_env *, const struct btf_type *, const struct btf_member *, const struct btf_type *); ++ int (*check_kflag_member)(struct btf_verifier_env *, const struct btf_type *, const struct btf_member *, const struct btf_type *); ++ void (*log_details)(struct btf_verifier_env *, const struct btf_type *); ++ void (*show)(const struct btf *, const struct btf_type *, u32, void *, u8, struct btf_show *); ++}; ++ ++enum { ++ BTF_FIELD_IGNORE = 0, ++ BTF_FIELD_FOUND = 1, ++}; ++ ++struct btf_field_info { ++ enum btf_field_type type; ++ u32 off; ++ union { ++ struct { ++ u32 type_id; ++ } kptr; ++ struct { ++ const char *node_name; ++ u32 value_btf_id; ++ } graph_root; ++ }; ++}; ++ ++struct bpf_ctx_convert { ++ struct __sk_buff BPF_PROG_TYPE_SOCKET_FILTER_prog; ++ struct sk_buff BPF_PROG_TYPE_SOCKET_FILTER_kern; ++ struct __sk_buff BPF_PROG_TYPE_SCHED_CLS_prog; ++ struct sk_buff BPF_PROG_TYPE_SCHED_CLS_kern; ++ struct __sk_buff BPF_PROG_TYPE_SCHED_ACT_prog; ++ struct sk_buff BPF_PROG_TYPE_SCHED_ACT_kern; ++ struct xdp_md BPF_PROG_TYPE_XDP_prog; ++ struct xdp_buff BPF_PROG_TYPE_XDP_kern; ++ struct __sk_buff BPF_PROG_TYPE_CGROUP_SKB_prog; ++ struct sk_buff BPF_PROG_TYPE_CGROUP_SKB_kern; ++ struct bpf_sock BPF_PROG_TYPE_CGROUP_SOCK_prog; ++ struct sock BPF_PROG_TYPE_CGROUP_SOCK_kern; ++ struct bpf_sock_addr BPF_PROG_TYPE_CGROUP_SOCK_ADDR_prog; ++ struct bpf_sock_addr_kern BPF_PROG_TYPE_CGROUP_SOCK_ADDR_kern; ++ struct __sk_buff BPF_PROG_TYPE_LWT_IN_prog; ++ struct sk_buff BPF_PROG_TYPE_LWT_IN_kern; ++ struct __sk_buff BPF_PROG_TYPE_LWT_OUT_prog; ++ struct sk_buff BPF_PROG_TYPE_LWT_OUT_kern; ++ struct __sk_buff BPF_PROG_TYPE_LWT_XMIT_prog; ++ struct sk_buff BPF_PROG_TYPE_LWT_XMIT_kern; ++ struct __sk_buff BPF_PROG_TYPE_LWT_SEG6LOCAL_prog; ++ struct sk_buff BPF_PROG_TYPE_LWT_SEG6LOCAL_kern; ++ struct bpf_sock_ops BPF_PROG_TYPE_SOCK_OPS_prog; ++ struct bpf_sock_ops_kern BPF_PROG_TYPE_SOCK_OPS_kern; ++ struct __sk_buff BPF_PROG_TYPE_SK_SKB_prog; ++ struct sk_buff BPF_PROG_TYPE_SK_SKB_kern; ++ struct sk_msg_md BPF_PROG_TYPE_SK_MSG_prog; ++ struct sk_msg BPF_PROG_TYPE_SK_MSG_kern; ++ struct __sk_buff BPF_PROG_TYPE_FLOW_DISSECTOR_prog; ++ struct bpf_flow_dissector BPF_PROG_TYPE_FLOW_DISSECTOR_kern; ++ bpf_user_pt_regs_t BPF_PROG_TYPE_KPROBE_prog; ++ struct pt_regs BPF_PROG_TYPE_KPROBE_kern; ++ __u64 BPF_PROG_TYPE_TRACEPOINT_prog; ++ u64 BPF_PROG_TYPE_TRACEPOINT_kern; ++ struct bpf_perf_event_data BPF_PROG_TYPE_PERF_EVENT_prog; ++ struct bpf_perf_event_data_kern BPF_PROG_TYPE_PERF_EVENT_kern; ++ struct bpf_raw_tracepoint_args BPF_PROG_TYPE_RAW_TRACEPOINT_prog; ++ u64 BPF_PROG_TYPE_RAW_TRACEPOINT_kern; ++ struct bpf_raw_tracepoint_args BPF_PROG_TYPE_RAW_TRACEPOINT_WRITABLE_prog; ++ u64 BPF_PROG_TYPE_RAW_TRACEPOINT_WRITABLE_kern; ++ void *BPF_PROG_TYPE_TRACING_prog; ++ void *BPF_PROG_TYPE_TRACING_kern; ++ struct bpf_cgroup_dev_ctx BPF_PROG_TYPE_CGROUP_DEVICE_prog; ++ struct bpf_cgroup_dev_ctx BPF_PROG_TYPE_CGROUP_DEVICE_kern; ++ struct bpf_sysctl BPF_PROG_TYPE_CGROUP_SYSCTL_prog; ++ struct bpf_sysctl_kern BPF_PROG_TYPE_CGROUP_SYSCTL_kern; ++ struct bpf_sockopt BPF_PROG_TYPE_CGROUP_SOCKOPT_prog; ++ struct bpf_sockopt_kern BPF_PROG_TYPE_CGROUP_SOCKOPT_kern; ++ __u32 BPF_PROG_TYPE_LIRC_MODE2_prog; ++ u32 BPF_PROG_TYPE_LIRC_MODE2_kern; ++ struct sk_reuseport_md BPF_PROG_TYPE_SK_REUSEPORT_prog; ++ struct sk_reuseport_kern BPF_PROG_TYPE_SK_REUSEPORT_kern; ++ struct bpf_sk_lookup BPF_PROG_TYPE_SK_LOOKUP_prog; ++ struct bpf_sk_lookup_kern BPF_PROG_TYPE_SK_LOOKUP_kern; ++ void *BPF_PROG_TYPE_STRUCT_OPS_prog; ++ void *BPF_PROG_TYPE_STRUCT_OPS_kern; ++ void *BPF_PROG_TYPE_EXT_prog; ++ void *BPF_PROG_TYPE_EXT_kern; ++ void *BPF_PROG_TYPE_LSM_prog; ++ void *BPF_PROG_TYPE_LSM_kern; ++ void *BPF_PROG_TYPE_SYSCALL_prog; ++ void *BPF_PROG_TYPE_SYSCALL_kern; ++ struct bpf_nf_ctx BPF_PROG_TYPE_NETFILTER_prog; ++ struct bpf_nf_ctx BPF_PROG_TYPE_NETFILTER_kern; ++}; ++ ++enum { ++ __ctx_convertBPF_PROG_TYPE_SOCKET_FILTER = 0, ++ __ctx_convertBPF_PROG_TYPE_SCHED_CLS = 1, ++ __ctx_convertBPF_PROG_TYPE_SCHED_ACT = 2, ++ __ctx_convertBPF_PROG_TYPE_XDP = 3, ++ __ctx_convertBPF_PROG_TYPE_CGROUP_SKB = 4, ++ __ctx_convertBPF_PROG_TYPE_CGROUP_SOCK = 5, ++ __ctx_convertBPF_PROG_TYPE_CGROUP_SOCK_ADDR = 6, ++ __ctx_convertBPF_PROG_TYPE_LWT_IN = 7, ++ __ctx_convertBPF_PROG_TYPE_LWT_OUT = 8, ++ __ctx_convertBPF_PROG_TYPE_LWT_XMIT = 9, ++ __ctx_convertBPF_PROG_TYPE_LWT_SEG6LOCAL = 10, ++ __ctx_convertBPF_PROG_TYPE_SOCK_OPS = 11, ++ __ctx_convertBPF_PROG_TYPE_SK_SKB = 12, ++ __ctx_convertBPF_PROG_TYPE_SK_MSG = 13, ++ __ctx_convertBPF_PROG_TYPE_FLOW_DISSECTOR = 14, ++ __ctx_convertBPF_PROG_TYPE_KPROBE = 15, ++ __ctx_convertBPF_PROG_TYPE_TRACEPOINT = 16, ++ __ctx_convertBPF_PROG_TYPE_PERF_EVENT = 17, ++ __ctx_convertBPF_PROG_TYPE_RAW_TRACEPOINT = 18, ++ __ctx_convertBPF_PROG_TYPE_RAW_TRACEPOINT_WRITABLE = 19, ++ __ctx_convertBPF_PROG_TYPE_TRACING = 20, ++ __ctx_convertBPF_PROG_TYPE_CGROUP_DEVICE = 21, ++ __ctx_convertBPF_PROG_TYPE_CGROUP_SYSCTL = 22, ++ __ctx_convertBPF_PROG_TYPE_CGROUP_SOCKOPT = 23, ++ __ctx_convertBPF_PROG_TYPE_LIRC_MODE2 = 24, ++ __ctx_convertBPF_PROG_TYPE_SK_REUSEPORT = 25, ++ __ctx_convertBPF_PROG_TYPE_SK_LOOKUP = 26, ++ __ctx_convertBPF_PROG_TYPE_STRUCT_OPS = 27, ++ __ctx_convertBPF_PROG_TYPE_EXT = 28, ++ __ctx_convertBPF_PROG_TYPE_LSM = 29, ++ __ctx_convertBPF_PROG_TYPE_SYSCALL = 30, ++ __ctx_convertBPF_PROG_TYPE_NETFILTER = 31, ++ __ctx_convert_unused = 32, ++}; ++ ++enum bpf_struct_walk_result { ++ WALK_SCALAR = 0, ++ WALK_PTR = 1, ++ WALK_STRUCT = 2, ++}; ++ ++struct btf_show_snprintf { ++ struct btf_show show; ++ int len_left; ++ int len; ++}; ++ ++enum { ++ BTF_MODULE_F_LIVE = 1, ++}; ++ ++struct btf_module { ++ struct list_head list; ++ struct module *module; ++ struct btf *btf; ++ struct bin_attribute *sysfs_attr; ++ int flags; ++}; ++ ++typedef u64 (*btf_bpf_btf_find_by_name_kind)(char *, int, u32, int); ++ ++struct bpf_cand_cache { ++ const char *name; ++ u32 name_len; ++ u16 kind; ++ u16 cnt; ++ struct { ++ const struct btf *btf; ++ u32 id; ++ } cands[0]; ++}; ++ ++enum bpf_stack_build_id_status { ++ BPF_STACK_BUILD_ID_EMPTY = 0, ++ BPF_STACK_BUILD_ID_VALID = 1, ++ BPF_STACK_BUILD_ID_IP = 2, ++}; ++ ++struct bpf_stack_build_id { ++ __s32 status; ++ unsigned char build_id[20]; ++ union { ++ __u64 offset; ++ __u64 ip; ++ }; ++}; ++ ++enum { ++ BPF_F_SKIP_FIELD_MASK = 255, ++ BPF_F_USER_STACK = 256, ++ BPF_F_FAST_STACK_CMP = 512, ++ BPF_F_REUSE_STACKID = 1024, ++ BPF_F_USER_BUILD_ID = 2048, ++}; ++ ++enum perf_callchain_context { ++ PERF_CONTEXT_HV = 18446744073709551584ULL, ++ PERF_CONTEXT_KERNEL = 18446744073709551488ULL, ++ PERF_CONTEXT_USER = 18446744073709551104ULL, ++ PERF_CONTEXT_GUEST = 18446744073709549568ULL, ++ PERF_CONTEXT_GUEST_KERNEL = 18446744073709549440ULL, ++ PERF_CONTEXT_GUEST_USER = 18446744073709549056ULL, ++ PERF_CONTEXT_MAX = 18446744073709547521ULL, ++}; ++ ++struct mmap_unlock_irq_work { ++ struct irq_work irq_work; ++ struct mm_struct *mm; ++}; ++ ++struct stack_map_bucket { ++ struct pcpu_freelist_node fnode; ++ u32 hash; ++ u32 nr; ++ u64 data[0]; ++}; ++ ++struct bpf_stack_map { ++ struct bpf_map map; ++ void *elems; ++ struct pcpu_freelist freelist; ++ u32 n_buckets; ++ struct stack_map_bucket *buckets[0]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++typedef u64 (*btf_bpf_get_stackid)(struct pt_regs *, struct bpf_map *, u64); ++ ++typedef u64 (*btf_bpf_get_stackid_pe)(struct bpf_perf_event_data_kern *, struct bpf_map *, u64); ++ ++typedef u64 (*btf_bpf_get_stack)(struct pt_regs *, void *, u32, u64); ++ ++typedef u64 (*btf_bpf_get_task_stack)(struct task_struct *, void *, u32, u64); ++ ++typedef u64 (*btf_bpf_get_stack_pe)(struct bpf_perf_event_data_kern *, void *, u32, u64); ++ ++enum sock_type { ++ SOCK_STREAM = 1, ++ SOCK_DGRAM = 2, ++ SOCK_RAW = 3, ++ SOCK_RDM = 4, ++ SOCK_SEQPACKET = 5, ++ SOCK_DCCP = 6, ++ SOCK_PACKET = 10, ++}; ++ ++enum sock_flags { ++ SOCK_DEAD = 0, ++ SOCK_DONE = 1, ++ SOCK_URGINLINE = 2, ++ SOCK_KEEPOPEN = 3, ++ SOCK_LINGER = 4, ++ SOCK_DESTROY = 5, ++ SOCK_BROADCAST = 6, ++ SOCK_TIMESTAMP = 7, ++ SOCK_ZAPPED = 8, ++ SOCK_USE_WRITE_QUEUE = 9, ++ SOCK_DBG = 10, ++ SOCK_RCVTSTAMP = 11, ++ SOCK_RCVTSTAMPNS = 12, ++ SOCK_LOCALROUTE = 13, ++ SOCK_MEMALLOC = 14, ++ SOCK_TIMESTAMPING_RX_SOFTWARE = 15, ++ SOCK_FASYNC = 16, ++ SOCK_RXQ_OVFL = 17, ++ SOCK_ZEROCOPY = 18, ++ SOCK_WIFI_STATUS = 19, ++ SOCK_NOFCS = 20, ++ SOCK_FILTER_LOCKED = 21, ++ SOCK_SELECT_ERR_QUEUE = 22, ++ SOCK_RCU_FREE = 23, ++ SOCK_TXTIME = 24, ++ SOCK_XDP = 25, ++ SOCK_TSTAMP_NEW = 26, ++ SOCK_RCVMARK = 27, ++}; ++ ++struct reuseport_array { ++ struct bpf_map map; ++ struct sock *ptrs[0]; ++}; ++ ++enum { ++ BPF_F_BPRM_SECUREEXEC = 1, ++}; ++ ++struct sembuf { ++ short unsigned int sem_num; ++ short int sem_op; ++ short int sem_flg; ++}; ++ ++enum key_need_perm { ++ KEY_NEED_UNSPECIFIED = 0, ++ KEY_NEED_VIEW = 1, ++ KEY_NEED_READ = 2, ++ KEY_NEED_WRITE = 3, ++ KEY_NEED_SEARCH = 4, ++ KEY_NEED_LINK = 5, ++ KEY_NEED_SETATTR = 6, ++ KEY_NEED_UNLINK = 7, ++ KEY_SYSADMIN_OVERRIDE = 8, ++ KEY_AUTHTOKEN_OVERRIDE = 9, ++ KEY_DEFER_PERM_CHECK = 10, ++}; ++ ++struct __key_reference_with_attributes; ++ ++typedef struct __key_reference_with_attributes *key_ref_t; ++ ++struct bpf_trace_run_ctx { ++ struct bpf_run_ctx run_ctx; ++ u64 bpf_cookie; ++}; ++ ++struct xfrm_sec_ctx { ++ __u8 ctx_doi; ++ __u8 ctx_alg; ++ __u16 ctx_len; ++ __u32 ctx_sid; ++ char ctx_str[0]; ++}; ++ ++struct xfrm_user_sec_ctx { ++ __u16 len; ++ __u16 exttype; ++ __u8 ctx_alg; ++ __u8 ctx_doi; ++ __u16 ctx_len; ++}; ++ ++enum kernel_read_file_id { ++ READING_UNKNOWN = 0, ++ READING_FIRMWARE = 1, ++ READING_MODULE = 2, ++ READING_KEXEC_IMAGE = 3, ++ READING_KEXEC_INITRAMFS = 4, ++ READING_POLICY = 5, ++ READING_X509_CERTIFICATE = 6, ++ READING_MAX_ID = 7, ++}; ++ ++enum hash_algo { ++ HASH_ALGO_MD4 = 0, ++ HASH_ALGO_MD5 = 1, ++ HASH_ALGO_SHA1 = 2, ++ HASH_ALGO_RIPE_MD_160 = 3, ++ HASH_ALGO_SHA256 = 4, ++ HASH_ALGO_SHA384 = 5, ++ HASH_ALGO_SHA512 = 6, ++ HASH_ALGO_SHA224 = 7, ++ HASH_ALGO_RIPE_MD_128 = 8, ++ HASH_ALGO_RIPE_MD_256 = 9, ++ HASH_ALGO_RIPE_MD_320 = 10, ++ HASH_ALGO_WP_256 = 11, ++ HASH_ALGO_WP_384 = 12, ++ HASH_ALGO_WP_512 = 13, ++ HASH_ALGO_TGR_128 = 14, ++ HASH_ALGO_TGR_160 = 15, ++ HASH_ALGO_TGR_192 = 16, ++ HASH_ALGO_SM3_256 = 17, ++ HASH_ALGO_STREEBOG_256 = 18, ++ HASH_ALGO_STREEBOG_512 = 19, ++ HASH_ALGO__LAST = 20, ++}; ++ ++typedef u64 (*btf_bpf_bprm_opts_set)(struct linux_binprm *, u64); ++ ++typedef u64 (*btf_bpf_ima_inode_hash)(struct inode *, void *, u32); ++ ++typedef u64 (*btf_bpf_ima_file_hash)(struct file *, void *, u32); ++ ++typedef u64 (*btf_bpf_get_attach_cookie)(void *); ++ ++struct perf_callchain_entry_ctx { ++ struct perf_callchain_entry *entry; ++ u32 max_stack; ++ u32 nr; ++ short int contexts; ++ bool contexts_maxed; ++}; ++ ++struct callchain_cpus_entries { ++ struct callback_head callback_head; ++ struct perf_callchain_entry *cpu_entries[0]; ++}; ++ ++struct workqueue_attrs { ++ int nice; ++ cpumask_var_t cpumask; ++ bool no_numa; ++}; ++ ++struct parallel_data; ++ ++struct padata_priv { ++ struct list_head list; ++ struct parallel_data *pd; ++ int cb_cpu; ++ unsigned int seq_nr; ++ int info; ++ void (*parallel)(struct padata_priv *); ++ void (*serial)(struct padata_priv *); ++}; ++ ++struct padata_cpumask { ++ cpumask_var_t pcpu; ++ cpumask_var_t cbcpu; ++}; ++ ++struct padata_shell; ++ ++struct padata_list; ++ ++struct padata_serial_queue; ++ ++struct parallel_data { ++ struct padata_shell *ps; ++ struct padata_list *reorder_list; ++ struct padata_serial_queue *squeue; ++ refcount_t refcnt; ++ unsigned int seq_nr; ++ unsigned int processed; ++ int cpu; ++ struct padata_cpumask cpumask; ++ struct work_struct reorder_work; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ spinlock_t lock; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct padata_list { ++ struct list_head list; ++ spinlock_t lock; ++}; ++ ++struct padata_serial_queue { ++ struct padata_list serial; ++ struct work_struct work; ++ struct parallel_data *pd; ++}; ++ ++struct padata_instance; ++ ++struct padata_shell { ++ struct padata_instance *pinst; ++ struct parallel_data *pd; ++ struct parallel_data *opd; ++ struct list_head list; ++}; ++ ++struct padata_instance { ++ struct hlist_node cpu_online_node; ++ struct hlist_node cpu_dead_node; ++ struct workqueue_struct *parallel_wq; ++ struct workqueue_struct *serial_wq; ++ struct list_head pslist; ++ struct padata_cpumask cpumask; ++ struct kobject kobj; ++ struct mutex lock; ++ u8 flags; ++}; ++ ++struct padata_mt_job { ++ void (*thread_fn)(long unsigned int, long unsigned int, void *); ++ void *fn_arg; ++ long unsigned int start; ++ long unsigned int size; ++ long unsigned int align; ++ long unsigned int min_chunk; ++ int max_threads; ++}; ++ ++struct padata_work { ++ struct work_struct pw_work; ++ struct list_head pw_list; ++ void *pw_data; ++}; ++ ++struct padata_mt_job_state { ++ spinlock_t lock; ++ struct completion completion; ++ struct padata_mt_job *job; ++ int nworks; ++ int nworks_fini; ++ long unsigned int chunk_size; ++}; ++ ++struct padata_sysfs_entry { ++ struct attribute attr; ++ ssize_t (*show)(struct padata_instance *, struct attribute *, char *); ++ ssize_t (*store)(struct padata_instance *, struct attribute *, const char *, size_t); ++}; ++ ++struct mem_section_usage { ++ long unsigned int subsection_map[2]; ++ long unsigned int pageblock_flags[0]; ++}; ++ ++struct page_ext; ++ ++struct mem_section { ++ long unsigned int section_mem_map; ++ struct mem_section_usage *usage; ++ struct page_ext *page_ext; ++ long unsigned int pad; ++}; ++ ++struct page_ext { ++ long unsigned int flags; ++}; ++ ++enum { ++ SECTION_MARKED_PRESENT_BIT = 0, ++ SECTION_HAS_MEM_MAP_BIT = 1, ++ SECTION_IS_ONLINE_BIT = 2, ++ SECTION_IS_EARLY_BIT = 3, ++ SECTION_MAP_LAST_BIT = 4, ++}; ++ ++enum { ++ IORES_DESC_NONE = 0, ++ IORES_DESC_CRASH_KERNEL = 1, ++ IORES_DESC_ACPI_TABLES = 2, ++ IORES_DESC_ACPI_NV_STORAGE = 3, ++ IORES_DESC_PERSISTENT_MEMORY = 4, ++ IORES_DESC_PERSISTENT_MEMORY_LEGACY = 5, ++ IORES_DESC_DEVICE_PRIVATE_MEMORY = 6, ++ IORES_DESC_RESERVED = 7, ++ IORES_DESC_SOFT_RESERVED = 8, ++ IORES_DESC_CXL = 9, ++}; ++ ++enum { ++ REGION_INTERSECTS = 0, ++ REGION_DISJOINT = 1, ++ REGION_MIXED = 2, ++}; ++ ++enum { ++ MEMREMAP_WB = 1, ++ MEMREMAP_WT = 2, ++ MEMREMAP_WC = 4, ++ MEMREMAP_ENC = 8, ++ MEMREMAP_DEC = 16, ++}; ++ ++struct watch; ++ ++struct watch_list { ++ struct callback_head rcu; ++ struct hlist_head watchers; ++ void (*release_watch)(struct watch *); ++ spinlock_t lock; ++}; ++ ++enum watch_notification_type { ++ WATCH_TYPE_META = 0, ++ WATCH_TYPE_KEY_NOTIFY = 1, ++ WATCH_TYPE__NR = 2, ++}; ++ ++enum watch_meta_notification_subtype { ++ WATCH_META_REMOVAL_NOTIFICATION = 0, ++ WATCH_META_LOSS_NOTIFICATION = 1, ++}; ++ ++struct watch_notification { ++ __u32 type: 24; ++ __u32 subtype: 8; ++ __u32 info; ++}; ++ ++struct watch_notification_type_filter { ++ __u32 type; ++ __u32 info_filter; ++ __u32 info_mask; ++ __u32 subtype_filter[8]; ++}; ++ ++struct watch_notification_filter { ++ __u32 nr_filters; ++ __u32 __reserved; ++ struct watch_notification_type_filter filters[0]; ++}; ++ ++struct watch_notification_removal { ++ struct watch_notification watch; ++ __u64 id; ++}; ++ ++struct watch_type_filter { ++ enum watch_notification_type type; ++ __u32 subtype_filter[1]; ++ __u32 info_filter; ++ __u32 info_mask; ++}; ++ ++struct watch_filter { ++ union { ++ struct callback_head rcu; ++ long unsigned int type_filter[1]; ++ }; ++ u32 nr_filters; ++ struct watch_type_filter filters[0]; ++}; ++ ++struct watch_queue { ++ struct callback_head rcu; ++ struct watch_filter *filter; ++ struct pipe_inode_info *pipe; ++ struct hlist_head watches; ++ struct page **notes; ++ long unsigned int *notes_bitmap; ++ struct kref usage; ++ spinlock_t lock; ++ unsigned int nr_notes; ++ unsigned int nr_pages; ++}; ++ ++struct watch { ++ union { ++ struct callback_head rcu; ++ u32 info_id; ++ }; ++ struct watch_queue *queue; ++ struct hlist_node queue_node; ++ struct watch_list *watch_list; ++ struct hlist_node list_node; ++ const struct cred *cred; ++ void *private; ++ u64 id; ++ struct kref usage; ++}; ++ ++struct reciprocal_value { ++ u32 m; ++ u8 sh1; ++ u8 sh2; ++}; ++ ++struct kmem_cache_order_objects { ++ unsigned int x; ++}; ++ ++struct kmem_cache_cpu; ++ ++struct kmem_cache_node; ++ ++struct kmem_cache { ++ struct kmem_cache_cpu *cpu_slab; ++ slab_flags_t flags; ++ long unsigned int min_partial; ++ unsigned int size; ++ unsigned int object_size; ++ struct reciprocal_value reciprocal_size; ++ unsigned int offset; ++ unsigned int cpu_partial; ++ unsigned int cpu_partial_slabs; ++ struct kmem_cache_order_objects oo; ++ struct kmem_cache_order_objects min; ++ gfp_t allocflags; ++ int refcount; ++ void (*ctor)(void *); ++ unsigned int inuse; ++ unsigned int align; ++ unsigned int red_left_pad; ++ const char *name; ++ struct list_head list; ++ struct kobject kobj; ++ long unsigned int random; ++ unsigned int remote_node_defrag_ratio; ++ unsigned int *random_seq; ++ unsigned int useroffset; ++ unsigned int usersize; ++ struct kmem_cache_node *node[2]; ++}; ++ ++struct compact_control; ++ ++struct capture_control { ++ struct compact_control *cc; ++ struct page *page; ++}; ++ ++typedef struct pglist_data pg_data_t; ++ ++struct vm_event_state { ++ long unsigned int event[86]; ++}; ++ ++enum oom_constraint { ++ CONSTRAINT_NONE = 0, ++ CONSTRAINT_CPUSET = 1, ++ CONSTRAINT_MEMORY_POLICY = 2, ++ CONSTRAINT_MEMCG = 3, ++}; ++ ++struct oom_control { ++ struct zonelist *zonelist; ++ nodemask_t *nodemask; ++ struct mem_cgroup *memcg; ++ const gfp_t gfp_mask; ++ const int order; ++ long unsigned int totalpages; ++ struct task_struct *chosen; ++ long int chosen_points; ++ enum oom_constraint constraint; ++}; ++ ++enum { ++ CGRP_ROOT_NOPREFIX = 2, ++ CGRP_ROOT_XATTR = 4, ++ CGRP_ROOT_NS_DELEGATE = 8, ++ CGRP_ROOT_FAVOR_DYNMODS = 16, ++ CGRP_ROOT_CPUSET_V2_MODE = 65536, ++ CGRP_ROOT_MEMORY_LOCAL_EVENTS = 131072, ++ CGRP_ROOT_MEMORY_RECURSIVE_PROT = 262144, ++}; ++ ++struct mmu_table_batch { ++ struct callback_head rcu; ++ unsigned int nr; ++ void *tables[0]; ++}; ++ ++struct mmu_gather { ++ struct mm_struct *mm; ++ struct mmu_table_batch *batch; ++ long unsigned int start; ++ long unsigned int end; ++ unsigned int fullmm: 1; ++ unsigned int need_flush_all: 1; ++ unsigned int freed_tables: 1; ++ unsigned int delayed_rmap: 1; ++ unsigned int cleared_ptes: 1; ++ unsigned int cleared_pmds: 1; ++ unsigned int cleared_puds: 1; ++ unsigned int cleared_p4ds: 1; ++ unsigned int vma_exec: 1; ++ unsigned int vma_huge: 1; ++ unsigned int vma_pfn: 1; ++ unsigned int batch_count; ++}; ++ ++struct compact_control { ++ struct list_head freepages; ++ struct list_head migratepages; ++ unsigned int nr_freepages; ++ unsigned int nr_migratepages; ++ long unsigned int free_pfn; ++ long unsigned int migrate_pfn; ++ long unsigned int fast_start_pfn; ++ struct zone *zone; ++ long unsigned int total_migrate_scanned; ++ long unsigned int total_free_scanned; ++ short unsigned int fast_search_fail; ++ short int search_order; ++ const gfp_t gfp_mask; ++ int order; ++ int migratetype; ++ const unsigned int alloc_flags; ++ const int highest_zoneidx; ++ enum migrate_mode mode; ++ bool ignore_skip_hint; ++ bool no_set_skip_hint; ++ bool ignore_block_suitable; ++ bool direct_compaction; ++ bool proactive_compaction; ++ bool whole_zone; ++ bool contended; ++ bool finish_pageblock; ++ bool alloc_contig; ++}; ++ ++struct slab { ++ long unsigned int __page_flags; ++ struct kmem_cache *slab_cache; ++ union { ++ struct { ++ union { ++ struct list_head slab_list; ++ struct { ++ struct slab *next; ++ int slabs; ++ }; ++ }; ++ void *freelist; ++ union { ++ long unsigned int counters; ++ struct { ++ unsigned int inuse: 16; ++ unsigned int objects: 15; ++ unsigned int frozen: 1; ++ }; ++ }; ++ }; ++ struct callback_head callback_head; ++ }; ++ unsigned int __unused; ++ atomic_t __page_refcount; ++ long unsigned int memcg_data; ++}; ++ ++struct kmem_cache_cpu { ++ void **freelist; ++ long unsigned int tid; ++ struct slab *slab; ++ struct slab *partial; ++ local_lock_t lock; ++}; ++ ++struct kmem_cache_node { ++ spinlock_t list_lock; ++ long unsigned int nr_partial; ++ struct list_head partial; ++ atomic_long_t nr_slabs; ++ atomic_long_t total_objects; ++ struct list_head full; ++}; ++ ++enum compact_priority { ++ COMPACT_PRIO_SYNC_FULL = 0, ++ MIN_COMPACT_PRIORITY = 0, ++ COMPACT_PRIO_SYNC_LIGHT = 1, ++ MIN_COMPACT_COSTLY_PRIORITY = 1, ++ DEF_COMPACT_PRIORITY = 1, ++ COMPACT_PRIO_ASYNC = 2, ++ INIT_COMPACT_PRIORITY = 2, ++}; ++ ++enum compact_result { ++ COMPACT_NOT_SUITABLE_ZONE = 0, ++ COMPACT_SKIPPED = 1, ++ COMPACT_DEFERRED = 2, ++ COMPACT_NO_SUITABLE_PAGE = 3, ++ COMPACT_CONTINUE = 4, ++ COMPACT_COMPLETE = 5, ++ COMPACT_PARTIAL_SKIPPED = 6, ++ COMPACT_CONTENDED = 7, ++ COMPACT_SUCCESS = 8, ++}; ++ ++struct trace_event_raw_oom_score_adj_update { ++ struct trace_entry ent; ++ pid_t pid; ++ char comm[16]; ++ short int oom_score_adj; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_reclaim_retry_zone { ++ struct trace_entry ent; ++ int node; ++ int zone_idx; ++ int order; ++ long unsigned int reclaimable; ++ long unsigned int available; ++ long unsigned int min_wmark; ++ int no_progress_loops; ++ bool wmark_check; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mark_victim { ++ struct trace_entry ent; ++ int pid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_wake_reaper { ++ struct trace_entry ent; ++ int pid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_start_task_reaping { ++ struct trace_entry ent; ++ int pid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_finish_task_reaping { ++ struct trace_entry ent; ++ int pid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_skip_task_reaping { ++ struct trace_entry ent; ++ int pid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_compact_retry { ++ struct trace_entry ent; ++ int order; ++ int priority; ++ int result; ++ int retries; ++ int max_retries; ++ bool ret; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_oom_score_adj_update {}; ++ ++struct trace_event_data_offsets_reclaim_retry_zone {}; ++ ++struct trace_event_data_offsets_mark_victim {}; ++ ++struct trace_event_data_offsets_wake_reaper {}; ++ ++struct trace_event_data_offsets_start_task_reaping {}; ++ ++struct trace_event_data_offsets_finish_task_reaping {}; ++ ++struct trace_event_data_offsets_skip_task_reaping {}; ++ ++struct trace_event_data_offsets_compact_retry {}; ++ ++typedef void (*btf_trace_oom_score_adj_update)(void *, struct task_struct *); ++ ++typedef void (*btf_trace_reclaim_retry_zone)(void *, struct zoneref *, int, long unsigned int, long unsigned int, long unsigned int, int, bool); ++ ++typedef void (*btf_trace_mark_victim)(void *, int); ++ ++typedef void (*btf_trace_wake_reaper)(void *, int); ++ ++typedef void (*btf_trace_start_task_reaping)(void *, int); ++ ++typedef void (*btf_trace_finish_task_reaping)(void *, int); ++ ++typedef void (*btf_trace_skip_task_reaping)(void *, int); ++ ++typedef void (*btf_trace_compact_retry)(void *, int, enum compact_priority, enum compact_result, int, int, bool); ++ ++enum migrate_reason { ++ MR_COMPACTION = 0, ++ MR_MEMORY_FAILURE = 1, ++ MR_MEMORY_HOTPLUG = 2, ++ MR_SYSCALL = 3, ++ MR_MEMPOLICY_MBIND = 4, ++ MR_NUMA_MISPLACED = 5, ++ MR_CONTIG_RANGE = 6, ++ MR_LONGTERM_PIN = 7, ++ MR_DEMOTION = 8, ++ MR_TYPES = 9, ++}; ++ ++enum { ++ LRU_GEN_ANON = 0, ++ LRU_GEN_FILE = 1, ++}; ++ ++enum { ++ LRU_GEN_CORE = 0, ++ LRU_GEN_MM_WALK = 1, ++ LRU_GEN_NONLEAF_YOUNG = 2, ++ NR_LRU_GEN_CAPS = 3, ++}; ++ ++struct encoded_page; ++ ++typedef union { ++} release_pages_arg; ++ ++enum page_memcg_data_flags { ++ MEMCG_DATA_OBJCGS = 1, ++ MEMCG_DATA_KMEM = 2, ++ __NR_MEMCG_DATA_FLAGS = 4, ++}; ++ ++enum mapping_flags { ++ AS_EIO = 0, ++ AS_ENOSPC = 1, ++ AS_MM_ALL_LOCKS = 2, ++ AS_UNEVICTABLE = 3, ++ AS_EXITING = 4, ++ AS_NO_WRITEBACK_TAGS = 5, ++ AS_LARGE_FOLIO_SUPPORT = 6, ++}; ++ ++struct pagevec { ++ unsigned char nr; ++ bool percpu_pvec_drained; ++ struct page *pages[15]; ++}; ++ ++struct folio_batch { ++ unsigned char nr; ++ bool percpu_pvec_drained; ++ struct folio *folios[15]; ++}; ++ ++struct trace_event_raw_mm_lru_insertion { ++ struct trace_entry ent; ++ struct folio *folio; ++ long unsigned int pfn; ++ enum lru_list lru; ++ long unsigned int flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_lru_activate { ++ struct trace_entry ent; ++ struct folio *folio; ++ long unsigned int pfn; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_mm_lru_insertion {}; ++ ++struct trace_event_data_offsets_mm_lru_activate {}; ++ ++typedef void (*btf_trace_mm_lru_insertion)(void *, struct folio *); ++ ++typedef void (*btf_trace_mm_lru_activate)(void *, struct folio *); ++ ++struct lru_rotate { ++ local_lock_t lock; ++ struct folio_batch fbatch; ++}; ++ ++struct cpu_fbatches { ++ local_lock_t lock; ++ struct folio_batch lru_add; ++ struct folio_batch lru_deactivate_file; ++ struct folio_batch lru_deactivate; ++ struct folio_batch lru_lazyfree; ++ struct folio_batch activate; ++}; ++ ++typedef void (*move_fn_t)(struct lruvec *, struct folio *); ++ ++struct radix_tree_iter { ++ long unsigned int index; ++ long unsigned int next_index; ++ long unsigned int tags; ++ struct xa_node *node; ++}; ++ ++enum { ++ RADIX_TREE_ITER_TAG_MASK = 15, ++ RADIX_TREE_ITER_TAGGED = 16, ++ RADIX_TREE_ITER_CONTIG = 32, ++}; ++ ++enum wb_state { ++ WB_registered = 0, ++ WB_writeback_running = 1, ++ WB_has_dirty_io = 2, ++ WB_start_all = 3, ++}; ++ ++enum pageblock_bits { ++ PB_migrate = 0, ++ PB_migrate_end = 2, ++ PB_migrate_skip = 3, ++ NR_PAGEBLOCK_BITS = 4, ++}; ++ ++typedef unsigned int isolate_mode_t; ++ ++struct node { ++ struct device dev; ++ struct list_head access_list; ++}; ++ ++typedef struct page *new_page_t(struct page *, long unsigned int); ++ ++typedef void free_page_t(struct page *, long unsigned int); ++ ++struct movable_operations { ++ bool (*isolate_page)(struct page *, isolate_mode_t); ++ int (*migrate_page)(struct page *, struct page *, enum migrate_mode); ++ void (*putback_page)(struct page *); ++}; ++ ++struct alloc_context { ++ struct zonelist *zonelist; ++ nodemask_t *nodemask; ++ struct zoneref *preferred_zoneref; ++ int migratetype; ++ enum zone_type highest_zoneidx; ++ bool spread_dirty_pages; ++}; ++ ++struct trace_event_raw_mm_compaction_isolate_template { ++ struct trace_entry ent; ++ long unsigned int start_pfn; ++ long unsigned int end_pfn; ++ long unsigned int nr_scanned; ++ long unsigned int nr_taken; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_compaction_migratepages { ++ struct trace_entry ent; ++ long unsigned int nr_migrated; ++ long unsigned int nr_failed; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_compaction_begin { ++ struct trace_entry ent; ++ long unsigned int zone_start; ++ long unsigned int migrate_pfn; ++ long unsigned int free_pfn; ++ long unsigned int zone_end; ++ bool sync; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_compaction_end { ++ struct trace_entry ent; ++ long unsigned int zone_start; ++ long unsigned int migrate_pfn; ++ long unsigned int free_pfn; ++ long unsigned int zone_end; ++ bool sync; ++ int status; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_compaction_try_to_compact_pages { ++ struct trace_entry ent; ++ int order; ++ long unsigned int gfp_mask; ++ int prio; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_compaction_suitable_template { ++ struct trace_entry ent; ++ int nid; ++ enum zone_type idx; ++ int order; ++ int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_compaction_defer_template { ++ struct trace_entry ent; ++ int nid; ++ enum zone_type idx; ++ int order; ++ unsigned int considered; ++ unsigned int defer_shift; ++ int order_failed; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_compaction_kcompactd_sleep { ++ struct trace_entry ent; ++ int nid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_kcompactd_wake_template { ++ struct trace_entry ent; ++ int nid; ++ int order; ++ enum zone_type highest_zoneidx; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_mm_compaction_isolate_template {}; ++ ++struct trace_event_data_offsets_mm_compaction_migratepages {}; ++ ++struct trace_event_data_offsets_mm_compaction_begin {}; ++ ++struct trace_event_data_offsets_mm_compaction_end {}; ++ ++struct trace_event_data_offsets_mm_compaction_try_to_compact_pages {}; ++ ++struct trace_event_data_offsets_mm_compaction_suitable_template {}; ++ ++struct trace_event_data_offsets_mm_compaction_defer_template {}; ++ ++struct trace_event_data_offsets_mm_compaction_kcompactd_sleep {}; ++ ++struct trace_event_data_offsets_kcompactd_wake_template {}; ++ ++typedef void (*btf_trace_mm_compaction_isolate_migratepages)(void *, long unsigned int, long unsigned int, long unsigned int, long unsigned int); ++ ++typedef void (*btf_trace_mm_compaction_isolate_freepages)(void *, long unsigned int, long unsigned int, long unsigned int, long unsigned int); ++ ++typedef void (*btf_trace_mm_compaction_migratepages)(void *, struct compact_control *, unsigned int); ++ ++typedef void (*btf_trace_mm_compaction_begin)(void *, struct compact_control *, long unsigned int, long unsigned int, bool); ++ ++typedef void (*btf_trace_mm_compaction_end)(void *, struct compact_control *, long unsigned int, long unsigned int, bool, int); ++ ++typedef void (*btf_trace_mm_compaction_try_to_compact_pages)(void *, int, gfp_t, int); ++ ++typedef void (*btf_trace_mm_compaction_finished)(void *, struct zone *, int, int); ++ ++typedef void (*btf_trace_mm_compaction_suitable)(void *, struct zone *, int, int); ++ ++typedef void (*btf_trace_mm_compaction_deferred)(void *, struct zone *, int); ++ ++typedef void (*btf_trace_mm_compaction_defer_compaction)(void *, struct zone *, int); ++ ++typedef void (*btf_trace_mm_compaction_defer_reset)(void *, struct zone *, int); ++ ++typedef void (*btf_trace_mm_compaction_kcompactd_sleep)(void *, int); ++ ++typedef void (*btf_trace_mm_compaction_wakeup_kcompactd)(void *, int, int, enum zone_type); ++ ++typedef void (*btf_trace_mm_compaction_kcompactd_wake)(void *, int, int, enum zone_type); ++ ++typedef enum { ++ ISOLATE_ABORT = 0, ++ ISOLATE_NONE = 1, ++ ISOLATE_SUCCESS = 2, ++} isolate_migrate_t; ++ ++enum { ++ DUMP_PREFIX_NONE = 0, ++ DUMP_PREFIX_ADDRESS = 1, ++ DUMP_PREFIX_OFFSET = 2, ++}; ++ ++typedef struct { ++ u64 val; ++} pfn_t; ++ ++typedef unsigned int pgtbl_mod_mask; ++ ++typedef int (*pte_fn_t)(pte_t *, long unsigned int, void *); ++ ++enum { ++ SWP_USED = 1, ++ SWP_WRITEOK = 2, ++ SWP_DISCARDABLE = 4, ++ SWP_DISCARDING = 8, ++ SWP_SOLIDSTATE = 16, ++ SWP_CONTINUED = 32, ++ SWP_BLKDEV = 64, ++ SWP_ACTIVATED = 128, ++ SWP_FS_OPS = 256, ++ SWP_AREA_DISCARD = 512, ++ SWP_PAGE_DISCARD = 1024, ++ SWP_STABLE_WRITES = 2048, ++ SWP_SYNCHRONOUS_IO = 4096, ++ SWP_SCANNING = 16384, ++}; ++ ++typedef long unsigned int pte_marker; ++ ++typedef int rmap_t; ++ ++struct copy_subpage_arg { ++ struct page *dst; ++ struct page *src; ++ struct vm_area_struct *vma; ++}; ++ ++struct anon_vma_name { ++ struct kref kref; ++ char name[0]; ++}; ++ ++struct hstate; ++ ++struct hugepage_subpool { ++ spinlock_t lock; ++ long int count; ++ long int max_hpages; ++ long int used_hpages; ++ struct hstate *hstate; ++ long int min_hpages; ++ long int rsv_hpages; ++}; ++ ++struct hstate { ++ struct mutex resize_lock; ++ int next_nid_to_alloc; ++ int next_nid_to_free; ++ unsigned int order; ++ unsigned int demote_order; ++ long unsigned int mask; ++ long unsigned int max_huge_pages; ++ long unsigned int nr_huge_pages; ++ long unsigned int free_huge_pages; ++ long unsigned int resv_huge_pages; ++ long unsigned int surplus_huge_pages; ++ long unsigned int nr_overcommit_huge_pages; ++ struct list_head hugepage_activelist; ++ struct list_head hugepage_freelists[2]; ++ unsigned int max_huge_pages_node[2]; ++ unsigned int nr_huge_pages_node[2]; ++ unsigned int free_huge_pages_node[2]; ++ unsigned int surplus_huge_pages_node[2]; ++ char name[32]; ++}; ++ ++struct hugetlbfs_sb_info { ++ long int max_inodes; ++ long int free_inodes; ++ spinlock_t stat_lock; ++ struct hstate *hstate; ++ struct hugepage_subpool *spool; ++ kuid_t uid; ++ kgid_t gid; ++ umode_t mode; ++}; ++ ++enum pgt_entry { ++ NORMAL_PMD = 0, ++ HPAGE_PMD = 1, ++ NORMAL_PUD = 2, ++ HPAGE_PUD = 3, ++}; ++ ++enum tlb_flush_reason { ++ TLB_FLUSH_ON_TASK_SWITCH = 0, ++ TLB_REMOTE_SHOOTDOWN = 1, ++ TLB_LOCAL_SHOOTDOWN = 2, ++ TLB_LOCAL_MM_SHOOTDOWN = 3, ++ TLB_REMOTE_SEND_IPI = 4, ++ NR_TLB_FLUSH_REASONS = 5, ++}; ++ ++struct anon_vma_chain { ++ struct vm_area_struct *vma; ++ struct anon_vma *anon_vma; ++ struct list_head same_vma; ++ struct rb_node rb; ++ long unsigned int rb_subtree_last; ++}; ++ ++enum ttu_flags { ++ TTU_SPLIT_HUGE_PMD = 4, ++ TTU_IGNORE_MLOCK = 8, ++ TTU_SYNC = 16, ++ TTU_HWPOISON = 32, ++ TTU_BATCH_FLUSH = 64, ++ TTU_RMAP_LOCKED = 128, ++}; ++ ++struct page_vma_mapped_walk { ++ long unsigned int pfn; ++ long unsigned int nr_pages; ++ long unsigned int pgoff; ++ struct vm_area_struct *vma; ++ long unsigned int address; ++ pmd_t *pmd; ++ pte_t *pte; ++ spinlock_t *ptl; ++ unsigned int flags; ++}; ++ ++struct rmap_walk_control { ++ void *arg; ++ bool try_lock; ++ bool contended; ++ bool (*rmap_one)(struct folio *, struct vm_area_struct *, long unsigned int, void *); ++ int (*done)(struct folio *); ++ struct anon_vma * (*anon_lock)(struct folio *, struct rmap_walk_control *); ++ bool (*invalid_vma)(struct vm_area_struct *, void *); ++}; ++ ++enum hugetlb_page_flags { ++ HPG_restore_reserve = 0, ++ HPG_migratable = 1, ++ HPG_temporary = 2, ++ HPG_freed = 3, ++ HPG_vmemmap_optimized = 4, ++ HPG_raw_hwp_unreliable = 5, ++ __NR_HPAGEFLAGS = 6, ++}; ++ ++struct trace_event_raw_tlb_flush { ++ struct trace_entry ent; ++ int reason; ++ long unsigned int pages; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_tlb_flush {}; ++ ++typedef void (*btf_trace_tlb_flush)(void *, int, long unsigned int); ++ ++struct trace_event_raw_mm_migrate_pages { ++ struct trace_entry ent; ++ long unsigned int succeeded; ++ long unsigned int failed; ++ long unsigned int thp_succeeded; ++ long unsigned int thp_failed; ++ long unsigned int thp_split; ++ enum migrate_mode mode; ++ int reason; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_migrate_pages_start { ++ struct trace_entry ent; ++ enum migrate_mode mode; ++ int reason; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_migration_pte { ++ struct trace_entry ent; ++ long unsigned int addr; ++ long unsigned int pte; ++ int order; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_mm_migrate_pages {}; ++ ++struct trace_event_data_offsets_mm_migrate_pages_start {}; ++ ++struct trace_event_data_offsets_migration_pte {}; ++ ++typedef void (*btf_trace_mm_migrate_pages)(void *, long unsigned int, long unsigned int, long unsigned int, long unsigned int, long unsigned int, enum migrate_mode, int); ++ ++typedef void (*btf_trace_mm_migrate_pages_start)(void *, enum migrate_mode, int); ++ ++typedef void (*btf_trace_set_migration_pte)(void *, long unsigned int, long unsigned int, int); ++ ++typedef void (*btf_trace_remove_migration_pte)(void *, long unsigned int, long unsigned int, int); ++ ++struct folio_referenced_arg { ++ int mapcount; ++ int referenced; ++ long unsigned int vm_flags; ++ struct mem_cgroup *memcg; ++}; ++ ++struct va_format { ++ const char *fmt; ++ va_list *va; ++}; ++ ++struct page_frag_cache { ++ void *va; ++ __u16 offset; ++ __u16 size; ++ unsigned int pagecnt_bias; ++ bool pfmemalloc; ++}; ++ ++enum zone_flags { ++ ZONE_BOOSTED_WATERMARK = 0, ++ ZONE_RECLAIM_ACTIVE = 1, ++}; ++ ++typedef void compound_page_dtor(struct page *); ++ ++struct migration_target_control { ++ int nid; ++ nodemask_t *nmask; ++ gfp_t gfp_mask; ++}; ++ ++typedef int fpi_t; ++ ++enum xa_lock_type { ++ XA_LOCK_IRQ = 1, ++ XA_LOCK_BH = 2, ++}; ++ ++typedef void (*xa_update_node_t)(struct xa_node *); ++ ++struct xa_state { ++ struct xarray *xa; ++ long unsigned int xa_index; ++ unsigned char xa_shift; ++ unsigned char xa_sibs; ++ unsigned char xa_offset; ++ unsigned char xa_pad; ++ struct xa_node *xa_node; ++ struct xa_node *xa_alloc; ++ xa_update_node_t xa_update; ++ struct list_lru *xa_lru; ++}; ++ ++struct vma_swap_readahead { ++ short unsigned int win; ++ short unsigned int offset; ++ short unsigned int nr_pte; ++ pte_t *ptes; ++}; ++ ++struct frontswap_ops { ++ void (*init)(unsigned int); ++ int (*store)(unsigned int, long unsigned int, struct page *); ++ int (*load)(unsigned int, long unsigned int, struct page *); ++ void (*invalidate_page)(unsigned int, long unsigned int); ++ void (*invalidate_area)(unsigned int); ++}; ++ ++typedef void (*crypto_completion_t)(void *, int); ++ ++struct crypto_tfm; ++ ++struct crypto_async_request { ++ struct list_head list; ++ crypto_completion_t complete; ++ void *data; ++ struct crypto_tfm *tfm; ++ u32 flags; ++}; ++ ++struct crypto_alg; ++ ++struct crypto_tfm { ++ refcount_t refcnt; ++ u32 crt_flags; ++ int node; ++ void (*exit)(struct crypto_tfm *); ++ struct crypto_alg *__crt_alg; ++ void *__crt_ctx[0]; ++}; ++ ++struct cipher_alg { ++ unsigned int cia_min_keysize; ++ unsigned int cia_max_keysize; ++ int (*cia_setkey)(struct crypto_tfm *, const u8 *, unsigned int); ++ void (*cia_encrypt)(struct crypto_tfm *, u8 *, const u8 *); ++ void (*cia_decrypt)(struct crypto_tfm *, u8 *, const u8 *); ++}; ++ ++struct compress_alg { ++ int (*coa_compress)(struct crypto_tfm *, const u8 *, unsigned int, u8 *, unsigned int *); ++ int (*coa_decompress)(struct crypto_tfm *, const u8 *, unsigned int, u8 *, unsigned int *); ++}; ++ ++struct crypto_type; ++ ++struct crypto_alg { ++ struct list_head cra_list; ++ struct list_head cra_users; ++ u32 cra_flags; ++ unsigned int cra_blocksize; ++ unsigned int cra_ctxsize; ++ unsigned int cra_alignmask; ++ int cra_priority; ++ refcount_t cra_refcnt; ++ char cra_name[128]; ++ char cra_driver_name[128]; ++ const struct crypto_type *cra_type; ++ union { ++ struct cipher_alg cipher; ++ struct compress_alg compress; ++ } cra_u; ++ int (*cra_init)(struct crypto_tfm *); ++ void (*cra_exit)(struct crypto_tfm *); ++ void (*cra_destroy)(struct crypto_alg *); ++ struct module *cra_module; ++}; ++ ++struct crypto_instance; ++ ++struct crypto_type { ++ unsigned int (*ctxsize)(struct crypto_alg *, u32, u32); ++ unsigned int (*extsize)(struct crypto_alg *); ++ int (*init)(struct crypto_tfm *, u32, u32); ++ int (*init_tfm)(struct crypto_tfm *); ++ void (*show)(struct seq_file *, struct crypto_alg *); ++ int (*report)(struct sk_buff *, struct crypto_alg *); ++ void (*free)(struct crypto_instance *); ++ unsigned int type; ++ unsigned int maskclear; ++ unsigned int maskset; ++ unsigned int tfmsize; ++}; ++ ++struct crypto_wait { ++ struct completion completion; ++ int err; ++}; ++ ++struct zpool; ++ ++struct zpool_ops { ++ int (*evict)(struct zpool *, long unsigned int); ++}; ++ ++enum zpool_mapmode { ++ ZPOOL_MM_RW = 0, ++ ZPOOL_MM_RO = 1, ++ ZPOOL_MM_WO = 2, ++ ZPOOL_MM_DEFAULT = 0, ++}; ++ ++struct acomp_req { ++ struct crypto_async_request base; ++ struct scatterlist *src; ++ struct scatterlist *dst; ++ unsigned int slen; ++ unsigned int dlen; ++ u32 flags; ++ void *__ctx[0]; ++}; ++ ++struct crypto_acomp { ++ int (*compress)(struct acomp_req *); ++ int (*decompress)(struct acomp_req *); ++ void (*dst_free)(struct scatterlist *); ++ unsigned int reqsize; ++ struct crypto_tfm base; ++}; ++ ++struct crypto_istat_compress { ++ atomic64_t compress_cnt; ++ atomic64_t compress_tlen; ++ atomic64_t decompress_cnt; ++ atomic64_t decompress_tlen; ++ atomic64_t err_cnt; ++}; ++ ++struct comp_alg_common { ++ struct crypto_alg base; ++}; ++ ++struct crypto_acomp_ctx { ++ struct crypto_acomp *acomp; ++ struct acomp_req *req; ++ struct crypto_wait wait; ++ u8 *dstmem; ++ struct mutex *mutex; ++}; ++ ++struct zswap_pool { ++ struct zpool *zpool; ++ struct crypto_acomp_ctx *acomp_ctx; ++ struct kref kref; ++ struct list_head list; ++ struct work_struct release_work; ++ struct work_struct shrink_work; ++ struct hlist_node node; ++ char tfm_name[128]; ++}; ++ ++struct zswap_entry { ++ struct rb_node rbnode; ++ long unsigned int offset; ++ int refcount; ++ unsigned int length; ++ struct zswap_pool *pool; ++ union { ++ long unsigned int handle; ++ long unsigned int value; ++ }; ++ struct obj_cgroup *objcg; ++}; ++ ++struct zswap_header { ++ swp_entry_t swpentry; ++}; ++ ++struct zswap_tree { ++ struct rb_root rbroot; ++ spinlock_t lock; ++}; ++ ++enum zswap_init_type { ++ ZSWAP_UNINIT = 0, ++ ZSWAP_INIT_SUCCEED = 1, ++ ZSWAP_INIT_FAILED = 2, ++}; ++ ++enum zswap_get_swap_ret { ++ ZSWAP_SWAPCACHE_NEW = 0, ++ ZSWAP_SWAPCACHE_EXIST = 1, ++ ZSWAP_SWAPCACHE_FAIL = 2, ++}; ++ ++struct nodemask_scratch { ++ nodemask_t mask1; ++ nodemask_t mask2; ++}; ++ ++enum { ++ MPOL_DEFAULT = 0, ++ MPOL_PREFERRED = 1, ++ MPOL_BIND = 2, ++ MPOL_INTERLEAVE = 3, ++ MPOL_LOCAL = 4, ++ MPOL_PREFERRED_MANY = 5, ++ MPOL_MAX = 6, ++}; ++ ++struct sp_node { ++ struct rb_node nd; ++ long unsigned int start; ++ long unsigned int end; ++ struct mempolicy *policy; ++}; ++ ++struct shared_policy { ++ struct rb_root root; ++ rwlock_t lock; ++}; ++ ++enum page_walk_lock { ++ PGWALK_RDLOCK = 0, ++ PGWALK_WRLOCK = 1, ++ PGWALK_WRLOCK_VERIFY = 2, ++}; ++ ++struct mm_walk; ++ ++struct mm_walk_ops { ++ int (*pgd_entry)(pgd_t *, long unsigned int, long unsigned int, struct mm_walk *); ++ int (*p4d_entry)(p4d_t *, long unsigned int, long unsigned int, struct mm_walk *); ++ int (*pud_entry)(pud_t *, long unsigned int, long unsigned int, struct mm_walk *); ++ int (*pmd_entry)(pmd_t *, long unsigned int, long unsigned int, struct mm_walk *); ++ int (*pte_entry)(pte_t *, long unsigned int, long unsigned int, struct mm_walk *); ++ int (*pte_hole)(long unsigned int, long unsigned int, int, struct mm_walk *); ++ int (*hugetlb_entry)(pte_t *, long unsigned int, long unsigned int, long unsigned int, struct mm_walk *); ++ int (*test_walk)(long unsigned int, long unsigned int, struct mm_walk *); ++ int (*pre_vma)(long unsigned int, long unsigned int, struct mm_walk *); ++ void (*post_vma)(struct mm_walk *); ++ enum page_walk_lock walk_lock; ++}; ++ ++enum page_walk_action { ++ ACTION_SUBTREE = 0, ++ ACTION_CONTINUE = 1, ++ ACTION_AGAIN = 2, ++}; ++ ++struct mm_walk { ++ const struct mm_walk_ops *ops; ++ struct mm_struct *mm; ++ pgd_t *pgd; ++ struct vm_area_struct *vma; ++ enum page_walk_action action; ++ bool no_vma; ++ void *private; ++}; ++ ++struct mempolicy_operations { ++ int (*create)(struct mempolicy *, const nodemask_t *); ++ void (*rebind)(struct mempolicy *, const nodemask_t *); ++}; ++ ++struct queue_pages { ++ struct list_head *pagelist; ++ long unsigned int flags; ++ nodemask_t *nmask; ++ long unsigned int start; ++ long unsigned int end; ++ struct vm_area_struct *first; ++}; ++ ++typedef long unsigned int cycles_t; ++ ++typedef u32 depot_stack_handle_t; ++ ++typedef void (*kunit_try_catch_func_t)(void *); ++ ++struct kunit_try_catch { ++ struct kunit *test; ++ struct completion *try_completion; ++ int try_result; ++ kunit_try_catch_func_t try; ++ kunit_try_catch_func_t catch; ++ void *context; ++}; ++ ++enum kunit_status { ++ KUNIT_SUCCESS = 0, ++ KUNIT_FAILURE = 1, ++ KUNIT_SKIPPED = 2, ++}; ++ ++struct kunit { ++ void *priv; ++ const char *name; ++ char *log; ++ struct kunit_try_catch try_catch; ++ const void *param_value; ++ int param_index; ++ spinlock_t lock; ++ enum kunit_status status; ++ struct list_head resources; ++ char status_comment[256]; ++}; ++ ++struct kunit_case; ++ ++struct kunit_suite { ++ const char name[256]; ++ int (*suite_init)(struct kunit_suite *); ++ void (*suite_exit)(struct kunit_suite *); ++ int (*init)(struct kunit *); ++ void (*exit)(struct kunit *); ++ struct kunit_case *test_cases; ++ char status_comment[256]; ++ struct dentry *debugfs; ++ char *log; ++ int suite_init_err; ++}; ++ ++enum stat_item { ++ ALLOC_FASTPATH = 0, ++ ALLOC_SLOWPATH = 1, ++ FREE_FASTPATH = 2, ++ FREE_SLOWPATH = 3, ++ FREE_FROZEN = 4, ++ FREE_ADD_PARTIAL = 5, ++ FREE_REMOVE_PARTIAL = 6, ++ ALLOC_FROM_PARTIAL = 7, ++ ALLOC_SLAB = 8, ++ ALLOC_REFILL = 9, ++ ALLOC_NODE_MISMATCH = 10, ++ FREE_SLAB = 11, ++ CPUSLAB_FLUSH = 12, ++ DEACTIVATE_FULL = 13, ++ DEACTIVATE_EMPTY = 14, ++ DEACTIVATE_TO_HEAD = 15, ++ DEACTIVATE_TO_TAIL = 16, ++ DEACTIVATE_REMOTE_FREES = 17, ++ DEACTIVATE_BYPASS = 18, ++ ORDER_FALLBACK = 19, ++ CMPXCHG_DOUBLE_CPU_FAIL = 20, ++ CMPXCHG_DOUBLE_FAIL = 21, ++ CPU_PARTIAL_ALLOC = 22, ++ CPU_PARTIAL_FREE = 23, ++ CPU_PARTIAL_NODE = 24, ++ CPU_PARTIAL_DRAIN = 25, ++ NR_SLUB_STAT_ITEMS = 26, ++}; ++ ++enum slab_state { ++ DOWN = 0, ++ PARTIAL = 1, ++ PARTIAL_NODE = 2, ++ UP = 3, ++ FULL = 4, ++}; ++ ++struct slabinfo { ++ long unsigned int active_objs; ++ long unsigned int num_objs; ++ long unsigned int active_slabs; ++ long unsigned int num_slabs; ++ long unsigned int shared_avail; ++ unsigned int limit; ++ unsigned int batchcount; ++ unsigned int shared; ++ unsigned int objects_per_slab; ++ unsigned int cache_order; ++}; ++ ++struct kmem_obj_info { ++ void *kp_ptr; ++ struct slab *kp_slab; ++ void *kp_objp; ++ long unsigned int kp_data_offset; ++ struct kmem_cache *kp_slab_cache; ++ void *kp_ret; ++ void *kp_stack[16]; ++ void *kp_free_stack[16]; ++}; ++ ++struct kunit_case { ++ void (*run_case)(struct kunit *); ++ const char *name; ++ const void * (*generate_params)(const void *, char *); ++ enum kunit_status status; ++ char *log; ++}; ++ ++struct kunit_resource; ++ ++typedef void (*kunit_resource_free_t)(struct kunit_resource *); ++ ++struct kunit_resource { ++ void *data; ++ const char *name; ++ kunit_resource_free_t free; ++ struct kref refcount; ++ struct list_head node; ++ bool should_kfree; ++}; ++ ++typedef bool (*kunit_resource_match_t)(struct kunit *, struct kunit_resource *, void *); ++ ++struct partial_context { ++ struct slab **slab; ++ gfp_t flags; ++ unsigned int orig_size; ++}; ++ ++struct track { ++ long unsigned int addr; ++ depot_stack_handle_t handle; ++ int cpu; ++ int pid; ++ long unsigned int when; ++}; ++ ++enum track_item { ++ TRACK_ALLOC = 0, ++ TRACK_FREE = 1, ++}; ++ ++struct slub_flush_work { ++ struct work_struct work; ++ struct kmem_cache *s; ++ bool skip; ++}; ++ ++struct detached_freelist { ++ struct slab *slab; ++ void *tail; ++ void *freelist; ++ int cnt; ++ struct kmem_cache *s; ++}; ++ ++struct location { ++ depot_stack_handle_t handle; ++ long unsigned int count; ++ long unsigned int addr; ++ long unsigned int waste; ++ long long int sum_time; ++ long int min_time; ++ long int max_time; ++ long int min_pid; ++ long int max_pid; ++ long unsigned int cpus[8]; ++ nodemask_t nodes; ++}; ++ ++struct loc_track { ++ long unsigned int max; ++ long unsigned int count; ++ struct location *loc; ++ loff_t idx; ++}; ++ ++enum slab_stat_type { ++ SL_ALL = 0, ++ SL_PARTIAL = 1, ++ SL_CPU = 2, ++ SL_OBJECTS = 3, ++ SL_TOTAL = 4, ++}; ++ ++struct slab_attribute { ++ struct attribute attr; ++ ssize_t (*show)(struct kmem_cache *, char *); ++ ssize_t (*store)(struct kmem_cache *, const char *, size_t); ++}; ++ ++struct saved_alias { ++ struct kmem_cache *s; ++ const char *name; ++ struct saved_alias *next; ++}; ++ ++enum slab_modes { ++ M_NONE = 0, ++ M_PARTIAL = 1, ++ M_FREE = 2, ++ M_FULL_NOLIST = 3, ++}; ++ ++struct swap_cgroup_ctrl { ++ struct page **map; ++ long unsigned int length; ++ spinlock_t lock; ++}; ++ ++struct swap_cgroup { ++ short unsigned int id; ++}; ++ ++struct trace_event_raw_test_pages_isolated { ++ struct trace_entry ent; ++ long unsigned int start_pfn; ++ long unsigned int end_pfn; ++ long unsigned int fin_pfn; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_test_pages_isolated {}; ++ ++typedef void (*btf_trace_test_pages_isolated)(void *, long unsigned int, long unsigned int, long unsigned int); ++ ++struct zpool_driver { ++ char *type; ++ struct module *owner; ++ atomic_t refcount; ++ struct list_head list; ++ void * (*create)(const char *, gfp_t, const struct zpool_ops *, struct zpool *); ++ void (*destroy)(void *); ++ bool malloc_support_movable; ++ int (*malloc)(void *, size_t, gfp_t, long unsigned int *); ++ void (*free)(void *, long unsigned int); ++ int (*shrink)(void *, unsigned int, unsigned int *); ++ bool sleep_mapped; ++ void * (*map)(void *, long unsigned int, enum zpool_mapmode); ++ void (*unmap)(void *, long unsigned int); ++ u64 (*total_size)(void *); ++}; ++ ++enum buddy { ++ HEADLESS = 0, ++ FIRST = 1, ++ MIDDLE = 2, ++ LAST = 3, ++ BUDDIES_MAX = 3, ++}; ++ ++struct z3fold_buddy_slots { ++ long unsigned int slot[4]; ++ long unsigned int pool; ++ rwlock_t lock; ++}; ++ ++struct z3fold_pool; ++ ++struct z3fold_header { ++ struct list_head buddy; ++ spinlock_t page_lock; ++ struct kref refcount; ++ struct work_struct work; ++ struct z3fold_buddy_slots *slots; ++ struct z3fold_pool *pool; ++ short int cpu; ++ short unsigned int first_chunks; ++ short unsigned int middle_chunks; ++ short unsigned int last_chunks; ++ short unsigned int start_middle; ++ short unsigned int first_num: 2; ++ short unsigned int mapped_count: 2; ++ short unsigned int foreign_handles: 2; ++}; ++ ++struct z3fold_pool { ++ const char *name; ++ spinlock_t lock; ++ spinlock_t stale_lock; ++ struct list_head *unbuddied; ++ struct list_head lru; ++ struct list_head stale; ++ atomic64_t pages_nr; ++ struct kmem_cache *c_handle; ++ struct zpool *zpool; ++ const struct zpool_ops *zpool_ops; ++ struct workqueue_struct *compact_wq; ++ struct workqueue_struct *release_wq; ++ struct work_struct work; ++}; ++ ++enum z3fold_page_flags { ++ PAGE_HEADLESS = 0, ++ MIDDLE_CHUNK_MAPPED = 1, ++ NEEDS_COMPACTING = 2, ++ PAGE_STALE = 3, ++ PAGE_CLAIMED = 4, ++ PAGE_MIGRATED = 5, ++}; ++ ++enum z3fold_handle_flags { ++ HANDLES_NOFREE = 0, ++}; ++ ++struct cma_kobject { ++ struct kobject kobj; ++ struct cma *cma; ++}; ++ ++struct damon_addr_range { ++ long unsigned int start; ++ long unsigned int end; ++}; ++ ++struct damon_region { ++ struct damon_addr_range ar; ++ long unsigned int sampling_addr; ++ unsigned int nr_accesses; ++ struct list_head list; ++ unsigned int age; ++ unsigned int last_nr_accesses; ++}; ++ ++struct damon_target { ++ struct pid *pid; ++ unsigned int nr_regions; ++ struct list_head regions_list; ++ struct list_head list; ++}; ++ ++enum damos_action { ++ DAMOS_WILLNEED = 0, ++ DAMOS_COLD = 1, ++ DAMOS_PAGEOUT = 2, ++ DAMOS_HUGEPAGE = 3, ++ DAMOS_NOHUGEPAGE = 4, ++ DAMOS_LRU_PRIO = 5, ++ DAMOS_LRU_DEPRIO = 6, ++ DAMOS_STAT = 7, ++ NR_DAMOS_ACTIONS = 8, ++}; ++ ++struct damos_quota { ++ long unsigned int ms; ++ long unsigned int sz; ++ long unsigned int reset_interval; ++ unsigned int weight_sz; ++ unsigned int weight_nr_accesses; ++ unsigned int weight_age; ++ long unsigned int total_charged_sz; ++ long unsigned int total_charged_ns; ++ long unsigned int esz; ++ long unsigned int charged_sz; ++ long unsigned int charged_from; ++ struct damon_target *charge_target_from; ++ long unsigned int charge_addr_from; ++ long unsigned int histogram[100]; ++ unsigned int min_score; ++}; ++ ++enum damos_wmark_metric { ++ DAMOS_WMARK_NONE = 0, ++ DAMOS_WMARK_FREE_MEM_RATE = 1, ++ NR_DAMOS_WMARK_METRICS = 2, ++}; ++ ++struct damos_watermarks { ++ enum damos_wmark_metric metric; ++ long unsigned int interval; ++ long unsigned int high; ++ long unsigned int mid; ++ long unsigned int low; ++ bool activated; ++}; ++ ++struct damos_stat { ++ long unsigned int nr_tried; ++ long unsigned int sz_tried; ++ long unsigned int nr_applied; ++ long unsigned int sz_applied; ++ long unsigned int qt_exceeds; ++}; ++ ++enum damos_filter_type { ++ DAMOS_FILTER_TYPE_ANON = 0, ++ DAMOS_FILTER_TYPE_MEMCG = 1, ++ NR_DAMOS_FILTER_TYPES = 2, ++}; ++ ++struct damos_filter { ++ enum damos_filter_type type; ++ bool matching; ++ union { ++ short unsigned int memcg_id; ++ }; ++ struct list_head list; ++}; ++ ++struct damos_access_pattern { ++ long unsigned int min_sz_region; ++ long unsigned int max_sz_region; ++ unsigned int min_nr_accesses; ++ unsigned int max_nr_accesses; ++ unsigned int min_age_region; ++ unsigned int max_age_region; ++}; ++ ++struct damos { ++ struct damos_access_pattern pattern; ++ enum damos_action action; ++ struct damos_quota quota; ++ struct damos_watermarks wmarks; ++ struct list_head filters; ++ struct damos_stat stat; ++ struct list_head list; ++}; ++ ++enum damon_ops_id { ++ DAMON_OPS_VADDR = 0, ++ DAMON_OPS_FVADDR = 1, ++ DAMON_OPS_PADDR = 2, ++ NR_DAMON_OPS = 3, ++}; ++ ++struct damon_ctx; ++ ++struct damon_operations { ++ enum damon_ops_id id; ++ void (*init)(struct damon_ctx *); ++ void (*update)(struct damon_ctx *); ++ void (*prepare_access_checks)(struct damon_ctx *); ++ unsigned int (*check_accesses)(struct damon_ctx *); ++ void (*reset_aggregated)(struct damon_ctx *); ++ int (*get_scheme_score)(struct damon_ctx *, struct damon_target *, struct damon_region *, struct damos *); ++ long unsigned int (*apply_scheme)(struct damon_ctx *, struct damon_target *, struct damon_region *, struct damos *); ++ bool (*target_valid)(struct damon_target *); ++ void (*cleanup)(struct damon_ctx *); ++}; ++ ++struct damon_attrs { ++ long unsigned int sample_interval; ++ long unsigned int aggr_interval; ++ long unsigned int ops_update_interval; ++ long unsigned int min_nr_regions; ++ long unsigned int max_nr_regions; ++}; ++ ++struct damon_callback { ++ void *private; ++ int (*before_start)(struct damon_ctx *); ++ int (*after_wmarks_check)(struct damon_ctx *); ++ int (*after_sampling)(struct damon_ctx *); ++ int (*after_aggregation)(struct damon_ctx *); ++ int (*before_damos_apply)(struct damon_ctx *, struct damon_target *, struct damon_region *, struct damos *); ++ void (*before_terminate)(struct damon_ctx *); ++}; ++ ++struct damon_ctx { ++ struct damon_attrs attrs; ++ struct timespec64 last_aggregation; ++ struct timespec64 last_ops_update; ++ struct task_struct *kdamond; ++ struct mutex kdamond_lock; ++ struct damon_operations ops; ++ struct damon_callback callback; ++ struct list_head adaptive_targets; ++ struct list_head schemes; ++}; ++ ++struct trace_event_raw_damon_aggregated { ++ struct trace_entry ent; ++ long unsigned int target_id; ++ unsigned int nr_regions; ++ long unsigned int start; ++ long unsigned int end; ++ unsigned int nr_accesses; ++ unsigned int age; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_damon_aggregated {}; ++ ++typedef void (*btf_trace_damon_aggregated)(void *, struct damon_target *, unsigned int, struct damon_region *, unsigned int); ++ ++struct damon_system_ram_region { ++ long unsigned int start; ++ long unsigned int end; ++}; ++ ++struct damon_sysfs_ul_range { ++ struct kobject kobj; ++ long unsigned int min; ++ long unsigned int max; ++}; ++ ++typedef unsigned int xa_mark_t; ++ ++enum { ++ XA_CHECK_SCHED = 4096, ++}; ++ ++enum { ++ HUGETLB_SHMFS_INODE = 1, ++ HUGETLB_ANONHUGE_INODE = 2, ++}; ++ ++struct hugetlbfs_inode_info { ++ struct shared_policy policy; ++ struct inode vfs_inode; ++ unsigned int seals; ++}; ++ ++struct simple_xattrs { ++ struct rb_root rb_root; ++ rwlock_t lock; ++}; ++ ++struct shmem_inode_info { ++ spinlock_t lock; ++ unsigned int seals; ++ long unsigned int flags; ++ long unsigned int alloced; ++ long unsigned int swapped; ++ long unsigned int fallocend; ++ struct list_head shrinklist; ++ struct list_head swaplist; ++ struct shared_policy policy; ++ struct simple_xattrs xattrs; ++ atomic_t stop_eviction; ++ struct timespec64 i_crtime; ++ unsigned int fsflags; ++ struct inode vfs_inode; ++}; ++ ++typedef __kernel_long_t __kernel_off_t; ++ ++typedef __kernel_off_t off_t; ++ ++enum iter_type { ++ ITER_IOVEC = 0, ++ ITER_KVEC = 1, ++ ITER_BVEC = 2, ++ ITER_PIPE = 3, ++ ITER_XARRAY = 4, ++ ITER_DISCARD = 5, ++ ITER_UBUF = 6, ++}; ++ ++typedef __kernel_rwf_t rwf_t; ++ ++typedef s32 compat_off_t; ++ ++typedef s64 compat_loff_t; ++ ++typedef short unsigned int ushort; ++ ++struct core_vma_metadata; ++ ++struct coredump_params { ++ const kernel_siginfo_t *siginfo; ++ struct file *file; ++ long unsigned int limit; ++ long unsigned int mm_flags; ++ int cpu; ++ loff_t written; ++ loff_t pos; ++ loff_t to_skip; ++ int vma_count; ++ size_t vma_data_size; ++ struct core_vma_metadata *vma_meta; ++}; ++ ++struct core_vma_metadata { ++ long unsigned int start; ++ long unsigned int end; ++ long unsigned int flags; ++ long unsigned int dump_size; ++ long unsigned int pgoff; ++ struct file *file; ++}; ++ ++struct open_flags { ++ int open_flag; ++ umode_t mode; ++ int acc_mode; ++ int intent; ++ int lookup_flags; ++}; ++ ++struct user_arg_ptr { ++ bool is_compat; ++ union { ++ const char * const *native; ++ const compat_uptr_t *compat; ++ } ptr; ++}; ++ ++struct f_owner_ex { ++ int type; ++ __kernel_pid_t pid; ++}; ++ ++struct flock { ++ short int l_type; ++ short int l_whence; ++ __kernel_off_t l_start; ++ __kernel_off_t l_len; ++ __kernel_pid_t l_pid; ++}; ++ ++enum rw_hint { ++ WRITE_LIFE_NOT_SET = 0, ++ WRITE_LIFE_NONE = 1, ++ WRITE_LIFE_SHORT = 2, ++ WRITE_LIFE_MEDIUM = 3, ++ WRITE_LIFE_LONG = 4, ++ WRITE_LIFE_EXTREME = 5, ++}; ++ ++struct file_lock_context { ++ spinlock_t flc_lock; ++ struct list_head flc_flock; ++ struct list_head flc_posix; ++ struct list_head flc_lease; ++}; ++ ++struct nlm_lockowner; ++ ++struct nfs_lock_info { ++ u32 state; ++ struct nlm_lockowner *owner; ++ struct list_head list; ++}; ++ ++struct nfs4_lock_state; ++ ++struct nfs4_lock_info { ++ struct nfs4_lock_state *owner; ++}; ++ ++struct file_lock_operations; ++ ++struct lock_manager_operations; ++ ++struct file_lock { ++ struct file_lock *fl_blocker; ++ struct list_head fl_list; ++ struct hlist_node fl_link; ++ struct list_head fl_blocked_requests; ++ struct list_head fl_blocked_member; ++ fl_owner_t fl_owner; ++ unsigned int fl_flags; ++ unsigned char fl_type; ++ unsigned int fl_pid; ++ int fl_link_cpu; ++ wait_queue_head_t fl_wait; ++ struct file *fl_file; ++ loff_t fl_start; ++ loff_t fl_end; ++ struct fasync_struct *fl_fasync; ++ long unsigned int fl_break_time; ++ long unsigned int fl_downgrade_time; ++ const struct file_lock_operations *fl_ops; ++ const struct lock_manager_operations *fl_lmops; ++ union { ++ struct nfs_lock_info nfs_fl; ++ struct nfs4_lock_info nfs4_fl; ++ struct { ++ struct list_head link; ++ int state; ++ unsigned int debug_id; ++ } afs; ++ struct { ++ struct inode *inode; ++ } ceph; ++ } fl_u; ++}; ++ ++struct compat_flock { ++ short int l_type; ++ short int l_whence; ++ compat_off_t l_start; ++ compat_off_t l_len; ++ compat_pid_t l_pid; ++}; ++ ++struct compat_flock64 { ++ short int l_type; ++ short int l_whence; ++ compat_loff_t l_start; ++ compat_loff_t l_len; ++ compat_pid_t l_pid; ++}; ++ ++struct file_lock_operations { ++ void (*fl_copy_lock)(struct file_lock *, struct file_lock *); ++ void (*fl_release_private)(struct file_lock *); ++}; ++ ++struct lock_manager_operations { ++ void *lm_mod_owner; ++ fl_owner_t (*lm_get_owner)(fl_owner_t); ++ void (*lm_put_owner)(fl_owner_t); ++ void (*lm_notify)(struct file_lock *); ++ int (*lm_grant)(struct file_lock *, int); ++ bool (*lm_break)(struct file_lock *); ++ int (*lm_change)(struct file_lock *, int, struct list_head *); ++ void (*lm_setup)(struct file_lock *, void **); ++ bool (*lm_breaker_owns_lease)(struct file_lock *); ++ bool (*lm_lock_expirable)(struct file_lock *); ++ void (*lm_expire_lock)(); ++}; ++ ++struct mnt_idmap { ++ struct user_namespace *owner; ++ refcount_t count; ++}; ++ ++struct ida { ++ struct xarray xa; ++}; ++ ++struct mount; ++ ++struct mnt_namespace { ++ struct ns_common ns; ++ struct mount *root; ++ struct list_head list; ++ spinlock_t ns_lock; ++ struct user_namespace *user_ns; ++ struct ucounts *ucounts; ++ u64 seq; ++ wait_queue_head_t poll; ++ u64 event; ++ unsigned int mounts; ++ unsigned int pending_mounts; ++}; ++ ++struct mount_attr { ++ __u64 attr_set; ++ __u64 attr_clr; ++ __u64 propagation; ++ __u64 userns_fd; ++}; ++ ++struct mnt_pcp; ++ ++struct mountpoint; ++ ++struct mount { ++ struct hlist_node mnt_hash; ++ struct mount *mnt_parent; ++ struct dentry *mnt_mountpoint; ++ struct vfsmount mnt; ++ union { ++ struct callback_head mnt_rcu; ++ struct llist_node mnt_llist; ++ }; ++ struct mnt_pcp *mnt_pcp; ++ struct list_head mnt_mounts; ++ struct list_head mnt_child; ++ struct list_head mnt_instance; ++ const char *mnt_devname; ++ struct list_head mnt_list; ++ struct list_head mnt_expire; ++ struct list_head mnt_share; ++ struct list_head mnt_slave_list; ++ struct list_head mnt_slave; ++ struct mount *mnt_master; ++ struct mnt_namespace *mnt_ns; ++ struct mountpoint *mnt_mp; ++ union { ++ struct hlist_node mnt_mp_list; ++ struct hlist_node mnt_umount; ++ }; ++ struct list_head mnt_umounting; ++ struct fsnotify_mark_connector *mnt_fsnotify_marks; ++ __u32 mnt_fsnotify_mask; ++ int mnt_id; ++ int mnt_group_id; ++ int mnt_expiry_mark; ++ struct hlist_head mnt_pins; ++ struct hlist_head mnt_stuck_children; ++}; ++ ++struct mnt_pcp { ++ int mnt_count; ++ int mnt_writers; ++}; ++ ++struct mountpoint { ++ struct hlist_node m_hash; ++ struct dentry *m_dentry; ++ struct hlist_head m_list; ++ int m_count; ++}; ++ ++struct proc_mounts { ++ struct mnt_namespace *ns; ++ struct path root; ++ int (*show)(struct seq_file *, struct vfsmount *); ++ struct mount cursor; ++}; ++ ++struct mount_kattr { ++ unsigned int attr_set; ++ unsigned int attr_clr; ++ unsigned int propagation; ++ unsigned int lookup_flags; ++ bool recurse; ++ struct user_namespace *mnt_userns; ++ struct mnt_idmap *mnt_idmap; ++}; ++ ++enum umount_tree_flags { ++ UMOUNT_SYNC = 1, ++ UMOUNT_PROPAGATE = 2, ++ UMOUNT_CONNECTED = 4, ++}; ++ ++typedef int __kernel_daddr_t; ++ ++struct ustat { ++ __kernel_daddr_t f_tfree; ++ unsigned int f_tinode; ++ char f_fname[6]; ++ char f_fpack[6]; ++}; ++ ++typedef u32 compat_ino_t; ++ ++typedef s32 compat_daddr_t; ++ ++typedef __kernel_fsid_t compat_fsid_t; ++ ++struct compat_statfs { ++ u32 f_type; ++ u32 f_bsize; ++ u32 f_blocks; ++ u32 f_bfree; ++ u32 f_bavail; ++ u32 f_files; ++ u32 f_ffree; ++ compat_fsid_t f_fsid; ++ u32 f_namelen; ++ u32 f_frsize; ++ u32 f_flags; ++ u32 f_spare[4]; ++}; ++ ++struct compat_statfs64 { ++ u32 f_type; ++ u32 f_bsize; ++ u64 f_blocks; ++ u64 f_bfree; ++ u64 f_bavail; ++ u64 f_files; ++ u64 f_ffree; ++ compat_fsid_t f_fsid; ++ u32 f_namelen; ++ u32 f_frsize; ++ u32 f_flags; ++ u32 f_spare[5]; ++}; ++ ++struct compat_ustat { ++ compat_daddr_t f_tfree; ++ compat_ino_t f_tinode; ++ char f_fname[6]; ++ char f_fpack[6]; ++}; ++ ++struct statfs { ++ unsigned int f_type; ++ unsigned int f_bsize; ++ long unsigned int f_blocks; ++ long unsigned int f_bfree; ++ long unsigned int f_bavail; ++ long unsigned int f_files; ++ long unsigned int f_ffree; ++ __kernel_fsid_t f_fsid; ++ unsigned int f_namelen; ++ unsigned int f_frsize; ++ unsigned int f_flags; ++ unsigned int f_spare[5]; ++}; ++ ++struct statfs64 { ++ unsigned int f_type; ++ unsigned int f_bsize; ++ long long unsigned int f_blocks; ++ long long unsigned int f_bfree; ++ long long unsigned int f_bavail; ++ long long unsigned int f_files; ++ long long unsigned int f_ffree; ++ __kernel_fsid_t f_fsid; ++ unsigned int f_namelen; ++ unsigned int f_frsize; ++ unsigned int f_flags; ++ unsigned int f_spare[5]; ++}; ++ ++struct constant_table { ++ const char *name; ++ int value; ++}; ++ ++enum legacy_fs_param { ++ LEGACY_FS_UNSET_PARAMS = 0, ++ LEGACY_FS_MONOLITHIC_PARAMS = 1, ++ LEGACY_FS_INDIVIDUAL_PARAMS = 2, ++}; ++ ++struct legacy_fs_context { ++ char *legacy_data; ++ size_t data_size; ++ enum legacy_fs_param param_type; ++}; ++ ++struct buffer_head; ++ ++typedef int get_block_t(struct inode *, sector_t, struct buffer_head *, int); ++ ++typedef void bh_end_io_t(struct buffer_head *, int); ++ ++struct buffer_head { ++ long unsigned int b_state; ++ struct buffer_head *b_this_page; ++ union { ++ struct page *b_page; ++ struct folio *b_folio; ++ }; ++ sector_t b_blocknr; ++ size_t b_size; ++ char *b_data; ++ struct block_device *b_bdev; ++ bh_end_io_t *b_end_io; ++ void *b_private; ++ struct list_head b_assoc_buffers; ++ struct address_space *b_assoc_map; ++ atomic_t b_count; ++ spinlock_t b_uptodate_lock; ++}; ++ ++enum { ++ BIO_NO_PAGE_REF = 0, ++ BIO_CLONED = 1, ++ BIO_BOUNCED = 2, ++ BIO_QUIET = 3, ++ BIO_CHAIN = 4, ++ BIO_REFFED = 5, ++ BIO_BPS_THROTTLED = 6, ++ BIO_TRACE_COMPLETION = 7, ++ BIO_CGROUP_ACCT = 8, ++ BIO_QOS_THROTTLED = 9, ++ BIO_QOS_MERGED = 10, ++ BIO_REMAPPED = 11, ++ BIO_ZONE_WRITE_LOCKED = 12, ++ BIO_FLAG_LAST = 13, ++}; ++ ++struct folio_iter { ++ struct folio *folio; ++ size_t offset; ++ size_t length; ++ struct folio *_next; ++ size_t _seg_count; ++ int _i; ++}; ++ ++enum bh_state_bits { ++ BH_Uptodate = 0, ++ BH_Dirty = 1, ++ BH_Lock = 2, ++ BH_Req = 3, ++ BH_Mapped = 4, ++ BH_New = 5, ++ BH_Async_Read = 6, ++ BH_Async_Write = 7, ++ BH_Delay = 8, ++ BH_Boundary = 9, ++ BH_Write_EIO = 10, ++ BH_Unwritten = 11, ++ BH_Quiet = 12, ++ BH_Meta = 13, ++ BH_Prio = 14, ++ BH_Defer_Completion = 15, ++ BH_PrivateStart = 16, ++}; ++ ++typedef int (*writepage_t)(struct folio *, struct writeback_control *, void *); ++ ++struct mpage_readpage_args { ++ struct bio *bio; ++ struct folio *folio; ++ unsigned int nr_pages; ++ bool is_readahead; ++ sector_t last_block_in_bio; ++ struct buffer_head map_bh; ++ long unsigned int first_logical_block; ++ get_block_t *get_block; ++}; ++ ++struct mpage_data { ++ struct bio *bio; ++ sector_t last_block_in_bio; ++ get_block_t *get_block; ++}; ++ ++enum dentry_d_lock_class { ++ DENTRY_D_LOCK_NORMAL = 0, ++ DENTRY_D_LOCK_NESTED = 1, ++}; ++ ++struct name_snapshot { ++ struct qstr name; ++ unsigned char inline_name[32]; ++}; ++ ++struct fsnotify_group; ++ ++struct fsnotify_iter_info; ++ ++struct fsnotify_mark; ++ ++struct fsnotify_event; ++ ++struct fsnotify_ops { ++ int (*handle_event)(struct fsnotify_group *, u32, const void *, int, struct inode *, const struct qstr *, u32, struct fsnotify_iter_info *); ++ int (*handle_inode_event)(struct fsnotify_mark *, u32, struct inode *, struct inode *, const struct qstr *, u32); ++ void (*free_group_priv)(struct fsnotify_group *); ++ void (*freeing_mark)(struct fsnotify_mark *, struct fsnotify_group *); ++ void (*free_event)(struct fsnotify_group *, struct fsnotify_event *); ++ void (*free_mark)(struct fsnotify_mark *); ++}; ++ ++struct inotify_group_private_data { ++ spinlock_t idr_lock; ++ struct idr idr; ++ struct ucounts *ucounts; ++}; ++ ++struct fanotify_group_private_data { ++ struct hlist_head *merge_hash; ++ struct list_head access_list; ++ wait_queue_head_t access_waitq; ++ int flags; ++ int f_flags; ++ struct ucounts *ucounts; ++ mempool_t error_events_pool; ++}; ++ ++struct fsnotify_group { ++ const struct fsnotify_ops *ops; ++ refcount_t refcnt; ++ spinlock_t notification_lock; ++ struct list_head notification_list; ++ wait_queue_head_t notification_waitq; ++ unsigned int q_len; ++ unsigned int max_events; ++ unsigned int priority; ++ bool shutdown; ++ int flags; ++ unsigned int owner_flags; ++ struct mutex mark_mutex; ++ atomic_t user_waits; ++ struct list_head marks_list; ++ struct fasync_struct *fsn_fa; ++ struct fsnotify_event *overflow_event; ++ struct mem_cgroup *memcg; ++ union { ++ void *private; ++ struct inotify_group_private_data inotify_data; ++ struct fanotify_group_private_data fanotify_data; ++ }; ++}; ++ ++struct fsnotify_iter_info { ++ struct fsnotify_mark *marks[5]; ++ struct fsnotify_group *current_group; ++ unsigned int report_mask; ++ int srcu_idx; ++}; ++ ++struct fsnotify_mark { ++ __u32 mask; ++ refcount_t refcnt; ++ struct fsnotify_group *group; ++ struct list_head g_list; ++ spinlock_t lock; ++ struct hlist_node obj_list; ++ struct fsnotify_mark_connector *connector; ++ __u32 ignore_mask; ++ unsigned int flags; ++}; ++ ++struct fsnotify_event { ++ struct list_head list; ++}; ++ ++struct fs_error_report { ++ int error; ++ struct inode *inode; ++ struct super_block *sb; ++}; ++ ++enum fsnotify_obj_type { ++ FSNOTIFY_OBJ_TYPE_ANY = -1, ++ FSNOTIFY_OBJ_TYPE_INODE = 0, ++ FSNOTIFY_OBJ_TYPE_VFSMOUNT = 1, ++ FSNOTIFY_OBJ_TYPE_SB = 2, ++ FSNOTIFY_OBJ_TYPE_COUNT = 3, ++ FSNOTIFY_OBJ_TYPE_DETACHED = 3, ++}; ++ ++struct inotify_event_info { ++ struct fsnotify_event fse; ++ u32 mask; ++ int wd; ++ u32 sync_cookie; ++ int name_len; ++ char name[0]; ++}; ++ ++struct inotify_inode_mark { ++ struct fsnotify_mark fsn_mark; ++ int wd; ++}; ++ ++enum fid_type { ++ FILEID_ROOT = 0, ++ FILEID_INO32_GEN = 1, ++ FILEID_INO32_GEN_PARENT = 2, ++ FILEID_BTRFS_WITHOUT_PARENT = 77, ++ FILEID_BTRFS_WITH_PARENT = 78, ++ FILEID_BTRFS_WITH_PARENT_ROOT = 79, ++ FILEID_UDF_WITHOUT_PARENT = 81, ++ FILEID_UDF_WITH_PARENT = 82, ++ FILEID_NILFS_WITHOUT_PARENT = 97, ++ FILEID_NILFS_WITH_PARENT = 98, ++ FILEID_FAT_WITHOUT_PARENT = 113, ++ FILEID_FAT_WITH_PARENT = 114, ++ FILEID_LUSTRE = 151, ++ FILEID_KERNFS = 254, ++ FILEID_INVALID = 255, ++}; ++ ++struct fid { ++ union { ++ struct { ++ u32 ino; ++ u32 gen; ++ u32 parent_ino; ++ u32 parent_gen; ++ } i32; ++ struct { ++ u32 block; ++ u16 partref; ++ u16 parent_partref; ++ u32 generation; ++ u32 parent_block; ++ u32 parent_generation; ++ } udf; ++ struct { ++ struct {} __empty_raw; ++ __u32 raw[0]; ++ }; ++ }; ++}; ++ ++enum { ++ FAN_EVENT_INIT = 0, ++ FAN_EVENT_REPORTED = 1, ++ FAN_EVENT_ANSWERED = 2, ++ FAN_EVENT_CANCELED = 3, ++}; ++ ++struct fanotify_fh { ++ u8 type; ++ u8 len; ++ u8 flags; ++ u8 pad; ++ unsigned char buf[0]; ++}; ++ ++struct fanotify_info { ++ u8 dir_fh_totlen; ++ u8 dir2_fh_totlen; ++ u8 file_fh_totlen; ++ u8 name_len; ++ u8 name2_len; ++ u8 pad[3]; ++ unsigned char buf[0]; ++}; ++ ++enum fanotify_event_type { ++ FANOTIFY_EVENT_TYPE_FID = 0, ++ FANOTIFY_EVENT_TYPE_FID_NAME = 1, ++ FANOTIFY_EVENT_TYPE_PATH = 2, ++ FANOTIFY_EVENT_TYPE_PATH_PERM = 3, ++ FANOTIFY_EVENT_TYPE_OVERFLOW = 4, ++ FANOTIFY_EVENT_TYPE_FS_ERROR = 5, ++ __FANOTIFY_EVENT_TYPE_NUM = 6, ++}; ++ ++struct fanotify_event { ++ struct fsnotify_event fse; ++ struct hlist_node merge_list; ++ u32 mask; ++ struct { ++ unsigned int type: 3; ++ unsigned int hash: 29; ++ }; ++ struct pid *pid; ++}; ++ ++struct fanotify_fid_event { ++ struct fanotify_event fae; ++ __kernel_fsid_t fsid; ++ struct { ++ struct fanotify_fh object_fh; ++ unsigned char _inline_fh_buf[12]; ++ }; ++}; ++ ++struct fanotify_name_event { ++ struct fanotify_event fae; ++ __kernel_fsid_t fsid; ++ struct fanotify_info info; ++}; ++ ++struct fanotify_error_event { ++ struct fanotify_event fae; ++ s32 error; ++ u32 err_count; ++ __kernel_fsid_t fsid; ++ struct { ++ struct fanotify_fh object_fh; ++ unsigned char _inline_fh_buf[128]; ++ }; ++}; ++ ++struct fanotify_path_event { ++ struct fanotify_event fae; ++ struct path path; ++}; ++ ++struct fanotify_perm_event { ++ struct fanotify_event fae; ++ struct path path; ++ u32 response; ++ short unsigned int state; ++ int fd; ++ union { ++ struct fanotify_response_info_header hdr; ++ struct fanotify_response_info_audit_rule audit_rule; ++ }; ++}; ++ ++struct pseudo_fs_context { ++ const struct super_operations *ops; ++ const struct xattr_handler **xattr; ++ const struct dentry_operations *dops; ++ long unsigned int magic; ++}; ++ ++struct timerfd_ctx { ++ union { ++ struct hrtimer tmr; ++ struct alarm alarm; ++ } t; ++ ktime_t tintv; ++ ktime_t moffs; ++ wait_queue_head_t wqh; ++ u64 ticks; ++ int clockid; ++ short unsigned int expired; ++ short unsigned int settime_flags; ++ struct callback_head rcu; ++ struct list_head clist; ++ spinlock_t cancel_lock; ++ bool might_cancel; ++}; ++ ++typedef long unsigned int dax_entry_t; ++ ++enum dax_access_mode { ++ DAX_ACCESS = 0, ++ DAX_RECOVERY_WRITE = 1, ++}; ++ ++struct dax_device; ++ ++struct iomap_folio_ops; ++ ++struct iomap { ++ u64 addr; ++ loff_t offset; ++ u64 length; ++ u16 type; ++ u16 flags; ++ struct block_device *bdev; ++ struct dax_device *dax_dev; ++ void *inline_data; ++ void *private; ++ const struct iomap_folio_ops *folio_ops; ++ u64 validity_cookie; ++}; ++ ++struct iomap_iter; ++ ++struct iomap_folio_ops { ++ struct folio * (*get_folio)(struct iomap_iter *, loff_t, unsigned int); ++ void (*put_folio)(struct inode *, loff_t, unsigned int, struct folio *); ++ bool (*iomap_valid)(struct inode *, const struct iomap *); ++}; ++ ++struct iomap_iter { ++ struct inode *inode; ++ loff_t pos; ++ u64 len; ++ s64 processed; ++ unsigned int flags; ++ struct iomap iomap; ++ struct iomap srcmap; ++ void *private; ++}; ++ ++struct iomap_ops { ++ int (*iomap_begin)(struct inode *, loff_t, loff_t, unsigned int, struct iomap *, struct iomap *); ++ int (*iomap_end)(struct inode *, loff_t, loff_t, ssize_t, unsigned int, struct iomap *); ++}; ++ ++struct trace_event_raw_dax_pmd_fault_class { ++ struct trace_entry ent; ++ long unsigned int ino; ++ long unsigned int vm_start; ++ long unsigned int vm_end; ++ long unsigned int vm_flags; ++ long unsigned int address; ++ long unsigned int pgoff; ++ long unsigned int max_pgoff; ++ dev_t dev; ++ unsigned int flags; ++ int result; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_dax_pmd_load_hole_class { ++ struct trace_entry ent; ++ long unsigned int ino; ++ long unsigned int vm_flags; ++ long unsigned int address; ++ struct page *zero_page; ++ void *radix_entry; ++ dev_t dev; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_dax_pmd_insert_mapping_class { ++ struct trace_entry ent; ++ long unsigned int ino; ++ long unsigned int vm_flags; ++ long unsigned int address; ++ long int length; ++ u64 pfn_val; ++ void *radix_entry; ++ dev_t dev; ++ int write; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_dax_pte_fault_class { ++ struct trace_entry ent; ++ long unsigned int ino; ++ long unsigned int vm_flags; ++ long unsigned int address; ++ long unsigned int pgoff; ++ dev_t dev; ++ unsigned int flags; ++ int result; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_dax_insert_mapping { ++ struct trace_entry ent; ++ long unsigned int ino; ++ long unsigned int vm_flags; ++ long unsigned int address; ++ void *radix_entry; ++ dev_t dev; ++ int write; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_dax_writeback_range_class { ++ struct trace_entry ent; ++ long unsigned int ino; ++ long unsigned int start_index; ++ long unsigned int end_index; ++ dev_t dev; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_dax_writeback_one { ++ struct trace_entry ent; ++ long unsigned int ino; ++ long unsigned int pgoff; ++ long unsigned int pglen; ++ dev_t dev; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_dax_pmd_fault_class {}; ++ ++struct trace_event_data_offsets_dax_pmd_load_hole_class {}; ++ ++struct trace_event_data_offsets_dax_pmd_insert_mapping_class {}; ++ ++struct trace_event_data_offsets_dax_pte_fault_class {}; ++ ++struct trace_event_data_offsets_dax_insert_mapping {}; ++ ++struct trace_event_data_offsets_dax_writeback_range_class {}; ++ ++struct trace_event_data_offsets_dax_writeback_one {}; ++ ++typedef void (*btf_trace_dax_pmd_fault)(void *, struct inode *, struct vm_fault *, long unsigned int, int); ++ ++typedef void (*btf_trace_dax_pmd_fault_done)(void *, struct inode *, struct vm_fault *, long unsigned int, int); ++ ++typedef void (*btf_trace_dax_pmd_load_hole)(void *, struct inode *, struct vm_fault *, struct page *, void *); ++ ++typedef void (*btf_trace_dax_pmd_load_hole_fallback)(void *, struct inode *, struct vm_fault *, struct page *, void *); ++ ++typedef void (*btf_trace_dax_pmd_insert_mapping)(void *, struct inode *, struct vm_fault *, long int, pfn_t, void *); ++ ++typedef void (*btf_trace_dax_pte_fault)(void *, struct inode *, struct vm_fault *, int); ++ ++typedef void (*btf_trace_dax_pte_fault_done)(void *, struct inode *, struct vm_fault *, int); ++ ++typedef void (*btf_trace_dax_load_hole)(void *, struct inode *, struct vm_fault *, int); ++ ++typedef void (*btf_trace_dax_insert_pfn_mkwrite_no_entry)(void *, struct inode *, struct vm_fault *, int); ++ ++typedef void (*btf_trace_dax_insert_pfn_mkwrite)(void *, struct inode *, struct vm_fault *, int); ++ ++typedef void (*btf_trace_dax_insert_mapping)(void *, struct inode *, struct vm_fault *, void *); ++ ++typedef void (*btf_trace_dax_writeback_range)(void *, struct inode *, long unsigned int, long unsigned int); ++ ++typedef void (*btf_trace_dax_writeback_range_done)(void *, struct inode *, long unsigned int, long unsigned int); ++ ++typedef void (*btf_trace_dax_writeback_one)(void *, struct inode *, long unsigned int, long unsigned int); ++ ++struct exceptional_entry_key { ++ struct xarray *xa; ++ long unsigned int entry_start; ++}; ++ ++struct wait_exceptional_entry_queue { ++ wait_queue_entry_t wait; ++ struct exceptional_entry_key key; ++}; ++ ++enum dax_wake_mode { ++ WAKE_ALL = 0, ++ WAKE_NEXT = 1, ++}; ++ ++struct crypto_skcipher { ++ unsigned int reqsize; ++ struct crypto_tfm base; ++}; ++ ++enum key_state { ++ KEY_IS_UNINSTANTIATED = 0, ++ KEY_IS_POSITIVE = 1, ++}; ++ ++struct key_preparsed_payload { ++ const char *orig_description; ++ char *description; ++ union key_payload payload; ++ const void *data; ++ size_t datalen; ++ size_t quotalen; ++ time64_t expiry; ++}; ++ ++struct key_match_data { ++ bool (*cmp)(const struct key *, const struct key_match_data *); ++ const void *raw_data; ++ void *preparsed; ++ unsigned int lookup_type; ++}; ++ ++enum kernel_pkey_operation { ++ kernel_pkey_encrypt = 0, ++ kernel_pkey_decrypt = 1, ++ kernel_pkey_sign = 2, ++ kernel_pkey_verify = 3, ++}; ++ ++struct kernel_pkey_params { ++ struct key *key; ++ const char *encoding; ++ const char *hash_algo; ++ char *info; ++ __u32 in_len; ++ union { ++ __u32 out_len; ++ __u32 in2_len; ++ }; ++ enum kernel_pkey_operation op: 8; ++}; ++ ++struct fscrypt_prepared_key { ++ struct crypto_skcipher *tfm; ++ struct blk_crypto_key *blk_key; ++}; ++ ++struct fscrypt_policy_v1 { ++ __u8 version; ++ __u8 contents_encryption_mode; ++ __u8 filenames_encryption_mode; ++ __u8 flags; ++ __u8 master_key_descriptor[8]; ++}; ++ ++struct fscrypt_policy_v2 { ++ __u8 version; ++ __u8 contents_encryption_mode; ++ __u8 filenames_encryption_mode; ++ __u8 flags; ++ __u8 __reserved[4]; ++ __u8 master_key_identifier[16]; ++}; ++ ++union fscrypt_policy { ++ u8 version; ++ struct fscrypt_policy_v1 v1; ++ struct fscrypt_policy_v2 v2; ++}; ++ ++struct fscrypt_mode; ++ ++struct fscrypt_master_key; ++ ++struct fscrypt_direct_key; ++ ++struct fscrypt_info { ++ struct fscrypt_prepared_key ci_enc_key; ++ bool ci_owns_key; ++ bool ci_inlinecrypt; ++ struct fscrypt_mode *ci_mode; ++ struct inode *ci_inode; ++ struct fscrypt_master_key *ci_master_key; ++ struct list_head ci_master_key_link; ++ struct fscrypt_direct_key *ci_direct_key; ++ siphash_key_t ci_dirhash_key; ++ bool ci_dirhash_key_initialized; ++ union fscrypt_policy ci_policy; ++ u8 ci_nonce[16]; ++ u32 ci_hashed_ino; ++}; ++ ++struct fscrypt_keyring { ++ spinlock_t lock; ++ struct hlist_head key_hashtable[128]; ++}; ++ ++struct fscrypt_key_specifier { ++ __u32 type; ++ __u32 __reserved; ++ union { ++ __u8 __reserved[32]; ++ __u8 descriptor[8]; ++ __u8 identifier[16]; ++ } u; ++}; ++ ++struct fscrypt_provisioning_key_payload { ++ __u32 type; ++ __u32 __reserved; ++ __u8 raw[0]; ++}; ++ ++struct fscrypt_add_key_arg { ++ struct fscrypt_key_specifier key_spec; ++ __u32 raw_size; ++ __u32 key_id; ++ __u32 __reserved[8]; ++ __u8 raw[0]; ++}; ++ ++struct fscrypt_remove_key_arg { ++ struct fscrypt_key_specifier key_spec; ++ __u32 removal_status_flags; ++ __u32 __reserved[5]; ++}; ++ ++struct fscrypt_get_key_status_arg { ++ struct fscrypt_key_specifier key_spec; ++ __u32 __reserved[6]; ++ __u32 status; ++ __u32 status_flags; ++ __u32 user_count; ++ __u32 __out_reserved[13]; ++}; ++ ++struct crypto_shash { ++ unsigned int descsize; ++ struct crypto_tfm base; ++}; ++ ++struct fscrypt_mode { ++ const char *friendly_name; ++ const char *cipher_str; ++ int keysize; ++ int security_strength; ++ int ivsize; ++ int logged_cryptoapi_impl; ++ int logged_blk_crypto_native; ++ int logged_blk_crypto_fallback; ++ enum blk_crypto_mode_num blk_crypto_mode; ++}; ++ ++struct fscrypt_hkdf { ++ struct crypto_shash *hmac_tfm; ++}; ++ ++struct fscrypt_master_key_secret { ++ struct fscrypt_hkdf hkdf; ++ u32 size; ++ u8 raw[64]; ++}; ++ ++struct fscrypt_master_key { ++ struct hlist_node mk_node; ++ struct rw_semaphore mk_sem; ++ refcount_t mk_active_refs; ++ refcount_t mk_struct_refs; ++ struct callback_head mk_rcu_head; ++ struct fscrypt_master_key_secret mk_secret; ++ struct fscrypt_key_specifier mk_spec; ++ struct key *mk_users; ++ struct list_head mk_decrypted_inodes; ++ spinlock_t mk_decrypted_inodes_lock; ++ struct fscrypt_prepared_key mk_direct_keys[11]; ++ struct fscrypt_prepared_key mk_iv_ino_lblk_64_keys[11]; ++ struct fscrypt_prepared_key mk_iv_ino_lblk_32_keys[11]; ++ siphash_key_t mk_ino_hash_key; ++ bool mk_ino_hash_key_initialized; ++}; ++ ++typedef enum { ++ FS_DECRYPT = 0, ++ FS_ENCRYPT = 1, ++} fscrypt_direction_t; ++ ++struct fsverity_hash_alg; ++ ++struct merkle_tree_params { ++ struct fsverity_hash_alg *hash_alg; ++ const u8 *hashstate; ++ unsigned int digest_size; ++ unsigned int block_size; ++ unsigned int hashes_per_block; ++ unsigned int blocks_per_page; ++ u8 log_digestsize; ++ u8 log_blocksize; ++ u8 log_arity; ++ u8 log_blocks_per_page; ++ unsigned int num_levels; ++ u64 tree_size; ++ long unsigned int tree_pages; ++ long unsigned int level_start[8]; ++}; ++ ++struct fsverity_info { ++ struct merkle_tree_params tree_params; ++ u8 root_hash[64]; ++ u8 file_digest[64]; ++ const struct inode *inode; ++ long unsigned int *hash_block_verified; ++ spinlock_t hash_page_init_lock; ++}; ++ ++struct fsverity_hash_alg { ++ struct crypto_shash *tfm; ++ const char *name; ++ unsigned int digest_size; ++ unsigned int block_size; ++ enum hash_algo algo_id; ++}; ++ ++typedef __u32 __le32; ++ ++typedef __u64 __le64; ++ ++struct fsverity_descriptor { ++ __u8 version; ++ __u8 hash_algorithm; ++ __u8 log_blocksize; ++ __u8 salt_size; ++ __le32 sig_size; ++ __le64 data_size; ++ __u8 root_hash[64]; ++ __u8 salt[32]; ++ __u8 __reserved[144]; ++ __u8 signature[0]; ++}; ++ ++struct flock64 { ++ short int l_type; ++ short int l_whence; ++ __kernel_loff_t l_start; ++ __kernel_loff_t l_len; ++ __kernel_pid_t l_pid; ++}; ++ ++struct trace_event_raw_locks_get_lock_context { ++ struct trace_entry ent; ++ long unsigned int i_ino; ++ dev_t s_dev; ++ unsigned char type; ++ struct file_lock_context *ctx; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_filelock_lock { ++ struct trace_entry ent; ++ struct file_lock *fl; ++ long unsigned int i_ino; ++ dev_t s_dev; ++ struct file_lock *fl_blocker; ++ fl_owner_t fl_owner; ++ unsigned int fl_pid; ++ unsigned int fl_flags; ++ unsigned char fl_type; ++ loff_t fl_start; ++ loff_t fl_end; ++ int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_filelock_lease { ++ struct trace_entry ent; ++ struct file_lock *fl; ++ long unsigned int i_ino; ++ dev_t s_dev; ++ struct file_lock *fl_blocker; ++ fl_owner_t fl_owner; ++ unsigned int fl_flags; ++ unsigned char fl_type; ++ long unsigned int fl_break_time; ++ long unsigned int fl_downgrade_time; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_generic_add_lease { ++ struct trace_entry ent; ++ long unsigned int i_ino; ++ int wcount; ++ int rcount; ++ int icount; ++ dev_t s_dev; ++ fl_owner_t fl_owner; ++ unsigned int fl_flags; ++ unsigned char fl_type; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_leases_conflict { ++ struct trace_entry ent; ++ void *lease; ++ void *breaker; ++ unsigned int l_fl_flags; ++ unsigned int b_fl_flags; ++ unsigned char l_fl_type; ++ unsigned char b_fl_type; ++ bool conflict; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_locks_get_lock_context {}; ++ ++struct trace_event_data_offsets_filelock_lock {}; ++ ++struct trace_event_data_offsets_filelock_lease {}; ++ ++struct trace_event_data_offsets_generic_add_lease {}; ++ ++struct trace_event_data_offsets_leases_conflict {}; ++ ++typedef void (*btf_trace_locks_get_lock_context)(void *, struct inode *, int, struct file_lock_context *); ++ ++typedef void (*btf_trace_posix_lock_inode)(void *, struct inode *, struct file_lock *, int); ++ ++typedef void (*btf_trace_fcntl_setlk)(void *, struct inode *, struct file_lock *, int); ++ ++typedef void (*btf_trace_locks_remove_posix)(void *, struct inode *, struct file_lock *, int); ++ ++typedef void (*btf_trace_flock_lock_inode)(void *, struct inode *, struct file_lock *, int); ++ ++typedef void (*btf_trace_break_lease_noblock)(void *, struct inode *, struct file_lock *); ++ ++typedef void (*btf_trace_break_lease_block)(void *, struct inode *, struct file_lock *); ++ ++typedef void (*btf_trace_break_lease_unblock)(void *, struct inode *, struct file_lock *); ++ ++typedef void (*btf_trace_generic_delete_lease)(void *, struct inode *, struct file_lock *); ++ ++typedef void (*btf_trace_time_out_leases)(void *, struct inode *, struct file_lock *); ++ ++typedef void (*btf_trace_generic_add_lease)(void *, struct inode *, struct file_lock *); ++ ++typedef void (*btf_trace_leases_conflict)(void *, bool, struct file_lock *, struct file_lock *); ++ ++struct file_lock_list_struct { ++ spinlock_t lock; ++ struct hlist_head hlist; ++}; ++ ++enum proc_hidepid { ++ HIDEPID_OFF = 0, ++ HIDEPID_NO_ACCESS = 1, ++ HIDEPID_INVISIBLE = 2, ++ HIDEPID_NOT_PTRACEABLE = 4, ++}; ++ ++enum proc_pidonly { ++ PROC_PIDONLY_OFF = 0, ++ PROC_PIDONLY_ON = 1, ++}; ++ ++struct proc_fs_info { ++ struct pid_namespace *pid_ns; ++ struct dentry *proc_self; ++ struct dentry *proc_thread_self; ++ kgid_t pid_gid; ++ enum proc_hidepid hide_pid; ++ enum proc_pidonly pidonly; ++}; ++ ++struct locks_iterator { ++ int li_cpu; ++ loff_t li_pos; ++}; ++ ++typedef __u16 __le16; ++ ++struct posix_acl_xattr_entry { ++ __le16 e_tag; ++ __le16 e_perm; ++ __le32 e_id; ++}; ++ ++struct posix_acl_xattr_header { ++ __le32 a_version; ++}; ++ ++typedef __kernel_ulong_t ino_t; ++ ++struct trace_event_raw_iomap_readpage_class { ++ struct trace_entry ent; ++ dev_t dev; ++ u64 ino; ++ int nr_pages; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_iomap_range_class { ++ struct trace_entry ent; ++ dev_t dev; ++ u64 ino; ++ loff_t size; ++ loff_t offset; ++ u64 length; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_iomap_class { ++ struct trace_entry ent; ++ dev_t dev; ++ u64 ino; ++ u64 addr; ++ loff_t offset; ++ u64 length; ++ u16 type; ++ u16 flags; ++ dev_t bdev; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_iomap_iter { ++ struct trace_entry ent; ++ dev_t dev; ++ u64 ino; ++ loff_t pos; ++ u64 length; ++ unsigned int flags; ++ const void *ops; ++ long unsigned int caller; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_iomap_dio_rw_begin { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ loff_t isize; ++ loff_t pos; ++ size_t count; ++ size_t done_before; ++ int ki_flags; ++ unsigned int dio_flags; ++ bool aio; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_iomap_dio_complete { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ loff_t isize; ++ loff_t pos; ++ int ki_flags; ++ bool aio; ++ int error; ++ ssize_t ret; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_iomap_readpage_class {}; ++ ++struct trace_event_data_offsets_iomap_range_class {}; ++ ++struct trace_event_data_offsets_iomap_class {}; ++ ++struct trace_event_data_offsets_iomap_iter {}; ++ ++struct trace_event_data_offsets_iomap_dio_rw_begin {}; ++ ++struct trace_event_data_offsets_iomap_dio_complete {}; ++ ++typedef void (*btf_trace_iomap_readpage)(void *, struct inode *, int); ++ ++typedef void (*btf_trace_iomap_readahead)(void *, struct inode *, int); ++ ++typedef void (*btf_trace_iomap_writepage)(void *, struct inode *, loff_t, u64); ++ ++typedef void (*btf_trace_iomap_release_folio)(void *, struct inode *, loff_t, u64); ++ ++typedef void (*btf_trace_iomap_invalidate_folio)(void *, struct inode *, loff_t, u64); ++ ++typedef void (*btf_trace_iomap_dio_invalidate_fail)(void *, struct inode *, loff_t, u64); ++ ++typedef void (*btf_trace_iomap_dio_rw_queued)(void *, struct inode *, loff_t, u64); ++ ++typedef void (*btf_trace_iomap_iter_dstmap)(void *, struct inode *, struct iomap *); ++ ++typedef void (*btf_trace_iomap_iter_srcmap)(void *, struct inode *, struct iomap *); ++ ++typedef void (*btf_trace_iomap_writepage_map)(void *, struct inode *, struct iomap *); ++ ++typedef void (*btf_trace_iomap_iter)(void *, struct iomap_iter *, const void *, long unsigned int); ++ ++typedef void (*btf_trace_iomap_dio_rw_begin)(void *, struct kiocb *, struct iov_iter *, unsigned int, size_t); ++ ++typedef void (*btf_trace_iomap_dio_complete)(void *, struct kiocb *, int, ssize_t); ++ ++struct fiemap_extent; ++ ++struct fiemap_extent_info { ++ unsigned int fi_flags; ++ unsigned int fi_extents_mapped; ++ unsigned int fi_extents_max; ++ struct fiemap_extent *fi_extents_start; ++}; ++ ++struct fiemap_extent { ++ __u64 fe_logical; ++ __u64 fe_physical; ++ __u64 fe_length; ++ __u64 fe_reserved64[2]; ++ __u32 fe_flags; ++ __u32 fe_reserved[3]; ++}; ++ ++enum { ++ QIF_BLIMITS_B = 0, ++ QIF_SPACE_B = 1, ++ QIF_ILIMITS_B = 2, ++ QIF_INODES_B = 3, ++ QIF_BTIME_B = 4, ++ QIF_ITIME_B = 5, ++}; ++ ++typedef __kernel_uid32_t qid_t; ++ ++enum { ++ DQF_INFO_DIRTY_B = 17, ++}; ++ ++struct dqstats { ++ long unsigned int stat[8]; ++ struct percpu_counter counter[8]; ++}; ++ ++enum { ++ _DQUOT_USAGE_ENABLED = 0, ++ _DQUOT_LIMITS_ENABLED = 1, ++ _DQUOT_SUSPENDED = 2, ++ _DQUOT_STATE_FLAGS = 3, ++}; ++ ++struct quota_module_name { ++ int qm_fmt_id; ++ char *qm_mod_name; ++}; ++ ++struct dquot_warn { ++ struct super_block *w_sb; ++ struct kqid w_dq_id; ++ short int w_type; ++}; ++ ++typedef u32 nlink_t; ++ ++typedef int (*proc_write_t)(struct file *, char *, size_t); ++ ++struct proc_dir_entry { ++ atomic_t in_use; ++ refcount_t refcnt; ++ struct list_head pde_openers; ++ spinlock_t pde_unload_lock; ++ struct completion *pde_unload_completion; ++ const struct inode_operations *proc_iops; ++ union { ++ const struct proc_ops *proc_ops; ++ const struct file_operations *proc_dir_ops; ++ }; ++ const struct dentry_operations *proc_dops; ++ union { ++ const struct seq_operations *seq_ops; ++ int (*single_show)(struct seq_file *, void *); ++ }; ++ proc_write_t write; ++ void *data; ++ unsigned int state_size; ++ unsigned int low_ino; ++ nlink_t nlink; ++ kuid_t uid; ++ kgid_t gid; ++ loff_t size; ++ struct proc_dir_entry *parent; ++ struct rb_root subdir; ++ struct rb_node subdir_node; ++ char *name; ++ umode_t mode; ++ u8 flags; ++ u8 namelen; ++ char inline_name[0]; ++}; ++ ++union proc_op { ++ int (*proc_get_link)(struct dentry *, struct path *); ++ int (*proc_show)(struct seq_file *, struct pid_namespace *, struct pid *, struct task_struct *); ++ const char *lsm; ++}; ++ ++struct proc_inode { ++ struct pid *pid; ++ unsigned int fd; ++ union proc_op op; ++ struct proc_dir_entry *pde; ++ struct ctl_table_header *sysctl; ++ struct ctl_table *sysctl_entry; ++ struct hlist_node sibling_inodes; ++ const struct proc_ns_operations *ns_ops; ++ struct inode vfs_inode; ++}; ++ ++struct proc_maps_private { ++ struct inode *inode; ++ struct task_struct *task; ++ struct mm_struct *mm; ++ struct vma_iterator iter; ++ struct mempolicy *task_mempolicy; ++}; ++ ++struct mem_size_stats { ++ long unsigned int resident; ++ long unsigned int shared_clean; ++ long unsigned int shared_dirty; ++ long unsigned int private_clean; ++ long unsigned int private_dirty; ++ long unsigned int referenced; ++ long unsigned int anonymous; ++ long unsigned int lazyfree; ++ long unsigned int anonymous_thp; ++ long unsigned int shmem_thp; ++ long unsigned int file_thp; ++ long unsigned int swap; ++ long unsigned int shared_hugetlb; ++ long unsigned int private_hugetlb; ++ u64 pss; ++ u64 pss_anon; ++ u64 pss_file; ++ u64 pss_shmem; ++ u64 pss_dirty; ++ u64 pss_locked; ++ u64 swap_pss; ++}; ++ ++enum clear_refs_types { ++ CLEAR_REFS_ALL = 1, ++ CLEAR_REFS_ANON = 2, ++ CLEAR_REFS_MAPPED = 3, ++ CLEAR_REFS_SOFT_DIRTY = 4, ++ CLEAR_REFS_MM_HIWATER_RSS = 5, ++ CLEAR_REFS_LAST = 6, ++}; ++ ++struct clear_refs_private { ++ enum clear_refs_types type; ++}; ++ ++typedef struct { ++ u64 pme; ++} pagemap_entry_t; ++ ++struct pagemapread { ++ int pos; ++ int len; ++ pagemap_entry_t *buffer; ++ bool show_pfn; ++}; ++ ++struct numa_maps { ++ long unsigned int pages; ++ long unsigned int anon; ++ long unsigned int active; ++ long unsigned int writeback; ++ long unsigned int mapcount_max; ++ long unsigned int dirty; ++ long unsigned int swapcache; ++ long unsigned int node[2]; ++}; ++ ++struct numa_maps_private { ++ struct proc_maps_private proc_maps; ++ struct numa_maps md; ++}; ++ ++enum { ++ PROC_ENTRY_PERMANENT = 1, ++}; ++ ++enum kcore_type { ++ KCORE_TEXT = 0, ++ KCORE_VMALLOC = 1, ++ KCORE_RAM = 2, ++ KCORE_VMEMMAP = 3, ++ KCORE_USER = 4, ++}; ++ ++struct kcore_list { ++ struct list_head list; ++ long unsigned int addr; ++ size_t size; ++ int type; ++}; ++ ++enum kernfs_node_type { ++ KERNFS_DIR = 1, ++ KERNFS_FILE = 2, ++ KERNFS_LINK = 4, ++}; ++ ++enum kernfs_node_flag { ++ KERNFS_ACTIVATED = 16, ++ KERNFS_NS = 32, ++ KERNFS_HAS_SEQ_SHOW = 64, ++ KERNFS_HAS_MMAP = 128, ++ KERNFS_LOCKDEP = 256, ++ KERNFS_HIDDEN = 512, ++ KERNFS_SUICIDAL = 1024, ++ KERNFS_SUICIDED = 2048, ++ KERNFS_EMPTY_DIR = 4096, ++ KERNFS_HAS_RELEASE = 8192, ++ KERNFS_REMOVING = 16384, ++}; ++ ++enum kernfs_root_flag { ++ KERNFS_ROOT_CREATE_DEACTIVATED = 1, ++ KERNFS_ROOT_EXTRA_OPEN_PERM_CHECK = 2, ++ KERNFS_ROOT_SUPPORT_EXPORTOP = 4, ++ KERNFS_ROOT_SUPPORT_USER_XATTR = 8, ++}; ++ ++struct kernfs_syscall_ops; ++ ++struct kernfs_root { ++ struct kernfs_node *kn; ++ unsigned int flags; ++ struct idr ino_idr; ++ u32 last_id_lowbits; ++ u32 id_highbits; ++ struct kernfs_syscall_ops *syscall_ops; ++ struct list_head supers; ++ wait_queue_head_t deactivate_waitq; ++ struct rw_semaphore kernfs_rwsem; ++ struct rw_semaphore kernfs_iattr_rwsem; ++ struct rw_semaphore kernfs_supers_rwsem; ++}; ++ ++struct kernfs_iattrs { ++ kuid_t ia_uid; ++ kgid_t ia_gid; ++ struct timespec64 ia_atime; ++ struct timespec64 ia_mtime; ++ struct timespec64 ia_ctime; ++ struct simple_xattrs xattrs; ++ atomic_t nr_user_xattrs; ++ atomic_t user_xattr_size; ++}; ++ ++struct kernfs_syscall_ops { ++ int (*show_options)(struct seq_file *, struct kernfs_root *); ++ int (*mkdir)(struct kernfs_node *, const char *, umode_t); ++ int (*rmdir)(struct kernfs_node *); ++ int (*rename)(struct kernfs_node *, struct kernfs_node *, const char *); ++ int (*show_path)(struct seq_file *, struct kernfs_node *, struct kernfs_root *); ++}; ++ ++struct config_group; ++ ++struct config_item_type; ++ ++struct config_item { ++ char *ci_name; ++ char ci_namebuf[20]; ++ struct kref ci_kref; ++ struct list_head ci_entry; ++ struct config_item *ci_parent; ++ struct config_group *ci_group; ++ const struct config_item_type *ci_type; ++ struct dentry *ci_dentry; ++}; ++ ++struct configfs_subsystem; ++ ++struct config_group { ++ struct config_item cg_item; ++ struct list_head cg_children; ++ struct configfs_subsystem *cg_subsys; ++ struct list_head default_groups; ++ struct list_head group_entry; ++}; ++ ++struct configfs_item_operations; ++ ++struct configfs_group_operations; ++ ++struct configfs_attribute; ++ ++struct configfs_bin_attribute; ++ ++struct config_item_type { ++ struct module *ct_owner; ++ struct configfs_item_operations *ct_item_ops; ++ struct configfs_group_operations *ct_group_ops; ++ struct configfs_attribute **ct_attrs; ++ struct configfs_bin_attribute **ct_bin_attrs; ++}; ++ ++struct configfs_item_operations { ++ void (*release)(struct config_item *); ++ int (*allow_link)(struct config_item *, struct config_item *); ++ void (*drop_link)(struct config_item *, struct config_item *); ++}; ++ ++struct configfs_group_operations { ++ struct config_item * (*make_item)(struct config_group *, const char *); ++ struct config_group * (*make_group)(struct config_group *, const char *); ++ void (*disconnect_notify)(struct config_group *, struct config_item *); ++ void (*drop_item)(struct config_group *, struct config_item *); ++}; ++ ++struct configfs_attribute { ++ const char *ca_name; ++ struct module *ca_owner; ++ umode_t ca_mode; ++ ssize_t (*show)(struct config_item *, char *); ++ ssize_t (*store)(struct config_item *, const char *, size_t); ++}; ++ ++struct configfs_bin_attribute { ++ struct configfs_attribute cb_attr; ++ void *cb_private; ++ size_t cb_max_size; ++ ssize_t (*read)(struct config_item *, void *, size_t); ++ ssize_t (*write)(struct config_item *, const void *, size_t); ++}; ++ ++struct configfs_subsystem { ++ struct config_group su_group; ++ struct mutex su_mutex; ++}; ++ ++struct configfs_fragment { ++ atomic_t frag_count; ++ struct rw_semaphore frag_sem; ++ bool frag_dead; ++}; ++ ++struct configfs_dirent { ++ atomic_t s_count; ++ int s_dependent_count; ++ struct list_head s_sibling; ++ struct list_head s_children; ++ int s_links; ++ void *s_element; ++ int s_type; ++ umode_t s_mode; ++ struct dentry *s_dentry; ++ struct iattr *s_iattr; ++ struct configfs_fragment *s_frag; ++}; ++ ++typedef unsigned int tid_t; ++ ++struct transaction_chp_stats_s { ++ long unsigned int cs_chp_time; ++ __u32 cs_forced_to_close; ++ __u32 cs_written; ++ __u32 cs_dropped; ++}; ++ ++struct journal_s; ++ ++typedef struct journal_s journal_t; ++ ++struct journal_head; ++ ++struct transaction_s; ++ ++typedef struct transaction_s transaction_t; ++ ++struct transaction_s { ++ journal_t *t_journal; ++ tid_t t_tid; ++ enum { ++ T_RUNNING = 0, ++ T_LOCKED = 1, ++ T_SWITCH = 2, ++ T_FLUSH = 3, ++ T_COMMIT = 4, ++ T_COMMIT_DFLUSH = 5, ++ T_COMMIT_JFLUSH = 6, ++ T_COMMIT_CALLBACK = 7, ++ T_FINISHED = 8, ++ } t_state; ++ long unsigned int t_log_start; ++ int t_nr_buffers; ++ struct journal_head *t_reserved_list; ++ struct journal_head *t_buffers; ++ struct journal_head *t_forget; ++ struct journal_head *t_checkpoint_list; ++ struct journal_head *t_shadow_list; ++ struct list_head t_inode_list; ++ spinlock_t t_handle_lock; ++ long unsigned int t_max_wait; ++ long unsigned int t_start; ++ long unsigned int t_requested; ++ struct transaction_chp_stats_s t_chp_stats; ++ atomic_t t_updates; ++ atomic_t t_outstanding_credits; ++ atomic_t t_outstanding_revokes; ++ atomic_t t_handle_count; ++ transaction_t *t_cpnext; ++ transaction_t *t_cpprev; ++ long unsigned int t_expires; ++ ktime_t t_start_time; ++ unsigned int t_synchronous_commit: 1; ++ int t_need_data_flush; ++ struct list_head t_private_list; ++}; ++ ++struct jbd2_buffer_trigger_type; ++ ++struct journal_head { ++ struct buffer_head *b_bh; ++ spinlock_t b_state_lock; ++ int b_jcount; ++ unsigned int b_jlist; ++ unsigned int b_modified; ++ char *b_frozen_data; ++ char *b_committed_data; ++ transaction_t *b_transaction; ++ transaction_t *b_next_transaction; ++ struct journal_head *b_tnext; ++ struct journal_head *b_tprev; ++ transaction_t *b_cp_transaction; ++ struct journal_head *b_cpnext; ++ struct journal_head *b_cpprev; ++ struct jbd2_buffer_trigger_type *b_triggers; ++ struct jbd2_buffer_trigger_type *b_frozen_triggers; ++}; ++ ++struct jbd2_buffer_trigger_type { ++ void (*t_frozen)(struct jbd2_buffer_trigger_type *, struct buffer_head *, void *, size_t); ++ void (*t_abort)(struct jbd2_buffer_trigger_type *, struct buffer_head *); ++}; ++ ++struct shash_desc { ++ struct crypto_shash *tfm; ++ void *__ctx[0]; ++}; ++ ++struct transaction_run_stats_s { ++ long unsigned int rs_wait; ++ long unsigned int rs_request_delay; ++ long unsigned int rs_running; ++ long unsigned int rs_locked; ++ long unsigned int rs_flushing; ++ long unsigned int rs_logging; ++ __u32 rs_handle_count; ++ __u32 rs_blocks; ++ __u32 rs_blocks_logged; ++}; ++ ++struct transaction_stats_s { ++ long unsigned int ts_tid; ++ long unsigned int ts_requested; ++ struct transaction_run_stats_s run; ++}; ++ ++enum passtype { ++ PASS_SCAN = 0, ++ PASS_REVOKE = 1, ++ PASS_REPLAY = 2, ++}; ++ ++struct journal_superblock_s; ++ ++typedef struct journal_superblock_s journal_superblock_t; ++ ++struct jbd2_revoke_table_s; ++ ++struct jbd2_inode; ++ ++struct journal_s { ++ long unsigned int j_flags; ++ long unsigned int j_atomic_flags; ++ int j_errno; ++ struct mutex j_abort_mutex; ++ struct buffer_head *j_sb_buffer; ++ journal_superblock_t *j_superblock; ++ int j_format_version; ++ rwlock_t j_state_lock; ++ int j_barrier_count; ++ struct mutex j_barrier; ++ transaction_t *j_running_transaction; ++ transaction_t *j_committing_transaction; ++ transaction_t *j_checkpoint_transactions; ++ wait_queue_head_t j_wait_transaction_locked; ++ wait_queue_head_t j_wait_done_commit; ++ wait_queue_head_t j_wait_commit; ++ wait_queue_head_t j_wait_updates; ++ wait_queue_head_t j_wait_reserved; ++ wait_queue_head_t j_fc_wait; ++ struct mutex j_checkpoint_mutex; ++ struct buffer_head *j_chkpt_bhs[64]; ++ struct shrinker j_shrinker; ++ struct percpu_counter j_checkpoint_jh_count; ++ transaction_t *j_shrink_transaction; ++ long unsigned int j_head; ++ long unsigned int j_tail; ++ long unsigned int j_free; ++ long unsigned int j_first; ++ long unsigned int j_last; ++ long unsigned int j_fc_first; ++ long unsigned int j_fc_off; ++ long unsigned int j_fc_last; ++ struct block_device *j_dev; ++ int j_blocksize; ++ long long unsigned int j_blk_offset; ++ char j_devname[56]; ++ struct block_device *j_fs_dev; ++ unsigned int j_total_len; ++ atomic_t j_reserved_credits; ++ spinlock_t j_list_lock; ++ struct inode *j_inode; ++ tid_t j_tail_sequence; ++ tid_t j_transaction_sequence; ++ tid_t j_commit_sequence; ++ tid_t j_commit_request; ++ __u8 j_uuid[16]; ++ struct task_struct *j_task; ++ int j_max_transaction_buffers; ++ int j_revoke_records_per_block; ++ long unsigned int j_commit_interval; ++ struct timer_list j_commit_timer; ++ spinlock_t j_revoke_lock; ++ struct jbd2_revoke_table_s *j_revoke; ++ struct jbd2_revoke_table_s *j_revoke_table[2]; ++ struct buffer_head **j_wbuf; ++ struct buffer_head **j_fc_wbuf; ++ int j_wbufsize; ++ int j_fc_wbufsize; ++ pid_t j_last_sync_writer; ++ u64 j_average_commit_time; ++ u32 j_min_batch_time; ++ u32 j_max_batch_time; ++ void (*j_commit_callback)(journal_t *, transaction_t *); ++ int (*j_submit_inode_data_buffers)(struct jbd2_inode *); ++ int (*j_finish_inode_data_buffers)(struct jbd2_inode *); ++ spinlock_t j_history_lock; ++ struct proc_dir_entry *j_proc_entry; ++ struct transaction_stats_s j_stats; ++ unsigned int j_failed_commit; ++ void *j_private; ++ struct crypto_shash *j_chksum_driver; ++ __u32 j_csum_seed; ++ void (*j_fc_cleanup_callback)(struct journal_s *, int, tid_t); ++ int (*j_fc_replay_callback)(struct journal_s *, struct buffer_head *, enum passtype, int, tid_t); ++ int (*j_bmap)(struct journal_s *, sector_t *); ++}; ++ ++struct journal_header_s { ++ __be32 h_magic; ++ __be32 h_blocktype; ++ __be32 h_sequence; ++}; ++ ++typedef struct journal_header_s journal_header_t; ++ ++struct journal_superblock_s { ++ journal_header_t s_header; ++ __be32 s_blocksize; ++ __be32 s_maxlen; ++ __be32 s_first; ++ __be32 s_sequence; ++ __be32 s_start; ++ __be32 s_errno; ++ __be32 s_feature_compat; ++ __be32 s_feature_incompat; ++ __be32 s_feature_ro_compat; ++ __u8 s_uuid[16]; ++ __be32 s_nr_users; ++ __be32 s_dynsuper; ++ __be32 s_max_transaction; ++ __be32 s_max_trans_data; ++ __u8 s_checksum_type; ++ __u8 s_padding2[3]; ++ __be32 s_num_fc_blks; ++ __u32 s_padding[41]; ++ __be32 s_checksum; ++ __u8 s_users[768]; ++}; ++ ++struct jbd2_inode { ++ transaction_t *i_transaction; ++ transaction_t *i_next_transaction; ++ struct list_head i_list; ++ struct inode *i_vfs_inode; ++ long unsigned int i_flags; ++ loff_t i_dirty_start; ++ loff_t i_dirty_end; ++}; ++ ++struct bgl_lock { ++ spinlock_t lock; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct blockgroup_lock { ++ struct bgl_lock locks[128]; ++}; ++ ++struct fscrypt_dummy_policy { ++ const union fscrypt_policy *policy; ++}; ++ ++typedef int ext4_grpblk_t; ++ ++typedef long long unsigned int ext4_fsblk_t; ++ ++typedef __u32 ext4_lblk_t; ++ ++typedef unsigned int ext4_group_t; ++ ++struct ext4_system_blocks { ++ struct rb_root root; ++ struct callback_head rcu; ++}; ++ ++struct ext4_group_desc { ++ __le32 bg_block_bitmap_lo; ++ __le32 bg_inode_bitmap_lo; ++ __le32 bg_inode_table_lo; ++ __le16 bg_free_blocks_count_lo; ++ __le16 bg_free_inodes_count_lo; ++ __le16 bg_used_dirs_count_lo; ++ __le16 bg_flags; ++ __le32 bg_exclude_bitmap_lo; ++ __le16 bg_block_bitmap_csum_lo; ++ __le16 bg_inode_bitmap_csum_lo; ++ __le16 bg_itable_unused_lo; ++ __le16 bg_checksum; ++ __le32 bg_block_bitmap_hi; ++ __le32 bg_inode_bitmap_hi; ++ __le32 bg_inode_table_hi; ++ __le16 bg_free_blocks_count_hi; ++ __le16 bg_free_inodes_count_hi; ++ __le16 bg_used_dirs_count_hi; ++ __le16 bg_itable_unused_hi; ++ __le32 bg_exclude_bitmap_hi; ++ __le16 bg_block_bitmap_csum_hi; ++ __le16 bg_inode_bitmap_csum_hi; ++ __u32 bg_reserved; ++}; ++ ++struct flex_groups { ++ atomic64_t free_clusters; ++ atomic_t free_inodes; ++ atomic_t used_dirs; ++}; ++ ++struct ext4_es_stats { ++ long unsigned int es_stats_shrunk; ++ struct percpu_counter es_stats_cache_hits; ++ struct percpu_counter es_stats_cache_misses; ++ u64 es_stats_scan_time; ++ u64 es_stats_max_scan_time; ++ struct percpu_counter es_stats_all_cnt; ++ struct percpu_counter es_stats_shk_cnt; ++}; ++ ++struct ext4_fc_stats { ++ unsigned int fc_ineligible_reason_count[10]; ++ long unsigned int fc_num_commits; ++ long unsigned int fc_ineligible_commits; ++ long unsigned int fc_failed_commits; ++ long unsigned int fc_skipped_commits; ++ long unsigned int fc_numblks; ++ u64 s_fc_avg_commit_time; ++}; ++ ++struct ext4_fc_alloc_region { ++ ext4_lblk_t lblk; ++ ext4_fsblk_t pblk; ++ int ino; ++ int len; ++}; ++ ++struct ext4_fc_replay_state { ++ int fc_replay_num_tags; ++ int fc_replay_expected_off; ++ int fc_current_pass; ++ int fc_cur_tag; ++ int fc_crc; ++ struct ext4_fc_alloc_region *fc_regions; ++ int fc_regions_size; ++ int fc_regions_used; ++ int fc_regions_valid; ++ int *fc_modified_inodes; ++ int fc_modified_inodes_used; ++ int fc_modified_inodes_size; ++}; ++ ++struct ext4_super_block { ++ __le32 s_inodes_count; ++ __le32 s_blocks_count_lo; ++ __le32 s_r_blocks_count_lo; ++ __le32 s_free_blocks_count_lo; ++ __le32 s_free_inodes_count; ++ __le32 s_first_data_block; ++ __le32 s_log_block_size; ++ __le32 s_log_cluster_size; ++ __le32 s_blocks_per_group; ++ __le32 s_clusters_per_group; ++ __le32 s_inodes_per_group; ++ __le32 s_mtime; ++ __le32 s_wtime; ++ __le16 s_mnt_count; ++ __le16 s_max_mnt_count; ++ __le16 s_magic; ++ __le16 s_state; ++ __le16 s_errors; ++ __le16 s_minor_rev_level; ++ __le32 s_lastcheck; ++ __le32 s_checkinterval; ++ __le32 s_creator_os; ++ __le32 s_rev_level; ++ __le16 s_def_resuid; ++ __le16 s_def_resgid; ++ __le32 s_first_ino; ++ __le16 s_inode_size; ++ __le16 s_block_group_nr; ++ __le32 s_feature_compat; ++ __le32 s_feature_incompat; ++ __le32 s_feature_ro_compat; ++ __u8 s_uuid[16]; ++ char s_volume_name[16]; ++ char s_last_mounted[64]; ++ __le32 s_algorithm_usage_bitmap; ++ __u8 s_prealloc_blocks; ++ __u8 s_prealloc_dir_blocks; ++ __le16 s_reserved_gdt_blocks; ++ __u8 s_journal_uuid[16]; ++ __le32 s_journal_inum; ++ __le32 s_journal_dev; ++ __le32 s_last_orphan; ++ __le32 s_hash_seed[4]; ++ __u8 s_def_hash_version; ++ __u8 s_jnl_backup_type; ++ __le16 s_desc_size; ++ __le32 s_default_mount_opts; ++ __le32 s_first_meta_bg; ++ __le32 s_mkfs_time; ++ __le32 s_jnl_blocks[17]; ++ __le32 s_blocks_count_hi; ++ __le32 s_r_blocks_count_hi; ++ __le32 s_free_blocks_count_hi; ++ __le16 s_min_extra_isize; ++ __le16 s_want_extra_isize; ++ __le32 s_flags; ++ __le16 s_raid_stride; ++ __le16 s_mmp_update_interval; ++ __le64 s_mmp_block; ++ __le32 s_raid_stripe_width; ++ __u8 s_log_groups_per_flex; ++ __u8 s_checksum_type; ++ __u8 s_encryption_level; ++ __u8 s_reserved_pad; ++ __le64 s_kbytes_written; ++ __le32 s_snapshot_inum; ++ __le32 s_snapshot_id; ++ __le64 s_snapshot_r_blocks_count; ++ __le32 s_snapshot_list; ++ __le32 s_error_count; ++ __le32 s_first_error_time; ++ __le32 s_first_error_ino; ++ __le64 s_first_error_block; ++ __u8 s_first_error_func[32]; ++ __le32 s_first_error_line; ++ __le32 s_last_error_time; ++ __le32 s_last_error_ino; ++ __le32 s_last_error_line; ++ __le64 s_last_error_block; ++ __u8 s_last_error_func[32]; ++ __u8 s_mount_opts[64]; ++ __le32 s_usr_quota_inum; ++ __le32 s_grp_quota_inum; ++ __le32 s_overhead_clusters; ++ __le32 s_backup_bgs[2]; ++ __u8 s_encrypt_algos[4]; ++ __u8 s_encrypt_pw_salt[16]; ++ __le32 s_lpf_ino; ++ __le32 s_prj_quota_inum; ++ __le32 s_checksum_seed; ++ __u8 s_wtime_hi; ++ __u8 s_mtime_hi; ++ __u8 s_mkfs_time_hi; ++ __u8 s_lastcheck_hi; ++ __u8 s_first_error_time_hi; ++ __u8 s_last_error_time_hi; ++ __u8 s_first_error_errcode; ++ __u8 s_last_error_errcode; ++ __le16 s_encoding; ++ __le16 s_encoding_flags; ++ __le32 s_orphan_file_inum; ++ __le32 s_reserved[94]; ++ __le32 s_checksum; ++}; ++ ++struct ext4_journal_trigger { ++ struct jbd2_buffer_trigger_type tr_triggers; ++ struct super_block *sb; ++}; ++ ++struct ext4_orphan_block { ++ atomic_t ob_free_entries; ++ struct buffer_head *ob_bh; ++}; ++ ++struct ext4_orphan_info { ++ int of_blocks; ++ __u32 of_csum_seed; ++ struct ext4_orphan_block *of_binfo; ++}; ++ ++struct ext4_group_info; ++ ++struct ext4_locality_group; ++ ++struct ext4_li_request; ++ ++struct mb_cache; ++ ++struct ext4_sb_info { ++ long unsigned int s_desc_size; ++ long unsigned int s_inodes_per_block; ++ long unsigned int s_blocks_per_group; ++ long unsigned int s_clusters_per_group; ++ long unsigned int s_inodes_per_group; ++ long unsigned int s_itb_per_group; ++ long unsigned int s_gdb_count; ++ long unsigned int s_desc_per_block; ++ ext4_group_t s_groups_count; ++ ext4_group_t s_blockfile_groups; ++ long unsigned int s_overhead; ++ unsigned int s_cluster_ratio; ++ unsigned int s_cluster_bits; ++ loff_t s_bitmap_maxbytes; ++ struct buffer_head *s_sbh; ++ struct ext4_super_block *s_es; ++ struct buffer_head **s_group_desc; ++ unsigned int s_mount_opt; ++ unsigned int s_mount_opt2; ++ long unsigned int s_mount_flags; ++ unsigned int s_def_mount_opt; ++ unsigned int s_def_mount_opt2; ++ ext4_fsblk_t s_sb_block; ++ atomic64_t s_resv_clusters; ++ kuid_t s_resuid; ++ kgid_t s_resgid; ++ short unsigned int s_mount_state; ++ short unsigned int s_pad; ++ int s_addr_per_block_bits; ++ int s_desc_per_block_bits; ++ int s_inode_size; ++ int s_first_ino; ++ unsigned int s_inode_readahead_blks; ++ unsigned int s_inode_goal; ++ u32 s_hash_seed[4]; ++ int s_def_hash_version; ++ int s_hash_unsigned; ++ struct percpu_counter s_freeclusters_counter; ++ struct percpu_counter s_freeinodes_counter; ++ struct percpu_counter s_dirs_counter; ++ struct percpu_counter s_dirtyclusters_counter; ++ struct percpu_counter s_sra_exceeded_retry_limit; ++ struct blockgroup_lock *s_blockgroup_lock; ++ struct proc_dir_entry *s_proc; ++ struct kobject s_kobj; ++ struct completion s_kobj_unregister; ++ struct super_block *s_sb; ++ struct buffer_head *s_mmp_bh; ++ struct journal_s *s_journal; ++ long unsigned int s_ext4_flags; ++ struct mutex s_orphan_lock; ++ struct list_head s_orphan; ++ struct ext4_orphan_info s_orphan_info; ++ long unsigned int s_commit_interval; ++ u32 s_max_batch_time; ++ u32 s_min_batch_time; ++ struct block_device *s_journal_bdev; ++ char *s_qf_names[3]; ++ int s_jquota_fmt; ++ unsigned int s_want_extra_isize; ++ struct ext4_system_blocks *s_system_blks; ++ struct ext4_group_info ***s_group_info; ++ struct inode *s_buddy_cache; ++ spinlock_t s_md_lock; ++ short unsigned int *s_mb_offsets; ++ unsigned int *s_mb_maxs; ++ unsigned int s_group_info_size; ++ unsigned int s_mb_free_pending; ++ struct list_head s_freed_data_list; ++ struct list_head s_discard_list; ++ struct work_struct s_discard_work; ++ atomic_t s_retry_alloc_pending; ++ struct list_head *s_mb_avg_fragment_size; ++ rwlock_t *s_mb_avg_fragment_size_locks; ++ struct list_head *s_mb_largest_free_orders; ++ rwlock_t *s_mb_largest_free_orders_locks; ++ long unsigned int s_stripe; ++ unsigned int s_mb_max_linear_groups; ++ unsigned int s_mb_stream_request; ++ unsigned int s_mb_max_to_scan; ++ unsigned int s_mb_min_to_scan; ++ unsigned int s_mb_stats; ++ unsigned int s_mb_order2_reqs; ++ unsigned int s_mb_group_prealloc; ++ unsigned int s_max_dir_size_kb; ++ long unsigned int s_mb_last_group; ++ long unsigned int s_mb_last_start; ++ unsigned int s_mb_prefetch; ++ unsigned int s_mb_prefetch_limit; ++ atomic_t s_bal_reqs; ++ atomic_t s_bal_success; ++ atomic_t s_bal_allocated; ++ atomic_t s_bal_ex_scanned; ++ atomic_t s_bal_groups_scanned; ++ atomic_t s_bal_goals; ++ atomic_t s_bal_breaks; ++ atomic_t s_bal_2orders; ++ atomic_t s_bal_cr0_bad_suggestions; ++ atomic_t s_bal_cr1_bad_suggestions; ++ atomic64_t s_bal_cX_groups_considered[4]; ++ atomic64_t s_bal_cX_hits[4]; ++ atomic64_t s_bal_cX_failed[4]; ++ atomic_t s_mb_buddies_generated; ++ atomic64_t s_mb_generation_time; ++ atomic_t s_mb_lost_chunks; ++ atomic_t s_mb_preallocated; ++ atomic_t s_mb_discarded; ++ atomic_t s_lock_busy; ++ struct ext4_locality_group *s_locality_groups; ++ long unsigned int s_sectors_written_start; ++ u64 s_kbytes_written; ++ unsigned int s_extent_max_zeroout_kb; ++ unsigned int s_log_groups_per_flex; ++ struct flex_groups **s_flex_groups; ++ ext4_group_t s_flex_groups_allocated; ++ struct workqueue_struct *rsv_conversion_wq; ++ struct timer_list s_err_report; ++ struct ext4_li_request *s_li_request; ++ unsigned int s_li_wait_mult; ++ struct task_struct *s_mmp_tsk; ++ long unsigned int s_last_trim_minblks; ++ struct crypto_shash *s_chksum_driver; ++ __u32 s_csum_seed; ++ struct shrinker s_es_shrinker; ++ struct list_head s_es_list; ++ long int s_es_nr_inode; ++ struct ext4_es_stats s_es_stats; ++ struct mb_cache *s_ea_block_cache; ++ struct mb_cache *s_ea_inode_cache; ++ long: 64; ++ long: 64; ++ long: 64; ++ spinlock_t s_es_lock; ++ struct ext4_journal_trigger s_journal_triggers[1]; ++ struct ratelimit_state s_err_ratelimit_state; ++ struct ratelimit_state s_warning_ratelimit_state; ++ struct ratelimit_state s_msg_ratelimit_state; ++ atomic_t s_warning_count; ++ atomic_t s_msg_count; ++ struct fscrypt_dummy_policy s_dummy_enc_policy; ++ struct percpu_rw_semaphore s_writepages_rwsem; ++ struct dax_device *s_daxdev; ++ u64 s_dax_part_off; ++ errseq_t s_bdev_wb_err; ++ spinlock_t s_bdev_wb_lock; ++ spinlock_t s_error_lock; ++ int s_add_error_count; ++ int s_first_error_code; ++ __u32 s_first_error_line; ++ __u32 s_first_error_ino; ++ __u64 s_first_error_block; ++ const char *s_first_error_func; ++ time64_t s_first_error_time; ++ int s_last_error_code; ++ __u32 s_last_error_line; ++ __u32 s_last_error_ino; ++ __u64 s_last_error_block; ++ const char *s_last_error_func; ++ time64_t s_last_error_time; ++ struct work_struct s_error_work; ++ atomic_t s_fc_subtid; ++ struct list_head s_fc_q[2]; ++ struct list_head s_fc_dentry_q[2]; ++ unsigned int s_fc_bytes; ++ spinlock_t s_fc_lock; ++ struct buffer_head *s_fc_bh; ++ struct ext4_fc_stats s_fc_stats; ++ tid_t s_fc_ineligible_tid; ++ struct ext4_fc_replay_state s_fc_replay_state; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct ext4_group_info { ++ long unsigned int bb_state; ++ struct rb_root bb_free_root; ++ ext4_grpblk_t bb_first_free; ++ ext4_grpblk_t bb_free; ++ ext4_grpblk_t bb_fragments; ++ int bb_avg_fragment_size_order; ++ ext4_grpblk_t bb_largest_free_order; ++ ext4_group_t bb_group; ++ struct list_head bb_prealloc_list; ++ struct rw_semaphore alloc_sem; ++ struct list_head bb_avg_fragment_size_node; ++ struct list_head bb_largest_free_order_node; ++ ext4_grpblk_t bb_counters[0]; ++}; ++ ++enum ext4_li_mode { ++ EXT4_LI_MODE_PREFETCH_BBITMAP = 0, ++ EXT4_LI_MODE_ITABLE = 1, ++}; ++ ++struct ext4_li_request { ++ struct super_block *lr_super; ++ enum ext4_li_mode lr_mode; ++ ext4_group_t lr_first_not_zeroed; ++ ext4_group_t lr_next_group; ++ struct list_head lr_request; ++ long unsigned int lr_next_sched; ++ long unsigned int lr_timeout; ++}; ++ ++struct utf8data; ++ ++struct utf8data_table; ++ ++struct unicode_map { ++ unsigned int version; ++ const struct utf8data *ntab[2]; ++ const struct utf8data_table *tables; ++}; ++ ++enum utf8_normalization { ++ UTF8_NFDI = 0, ++ UTF8_NFDICF = 1, ++ UTF8_NMAX = 2, ++}; ++ ++struct utf8data { ++ unsigned int maxage; ++ unsigned int offset; ++}; ++ ++struct utf8data_table { ++ const unsigned int *utf8agetab; ++ int utf8agetab_size; ++ const struct utf8data *utf8nfdicfdata; ++ int utf8nfdicfdata_size; ++ const struct utf8data *utf8nfdidata; ++ int utf8nfdidata_size; ++ const unsigned char *utf8data; ++}; ++ ++struct jbd2_journal_handle; ++ ++typedef struct jbd2_journal_handle handle_t; ++ ++struct jbd2_journal_handle { ++ union { ++ transaction_t *h_transaction; ++ journal_t *h_journal; ++ }; ++ handle_t *h_rsv_handle; ++ int h_total_credits; ++ int h_revoke_credits; ++ int h_revoke_credits_requested; ++ int h_ref; ++ int h_err; ++ unsigned int h_sync: 1; ++ unsigned int h_jdata: 1; ++ unsigned int h_reserved: 1; ++ unsigned int h_aborted: 1; ++ unsigned int h_type: 8; ++ unsigned int h_line_no: 16; ++ long unsigned int h_start_jiffies; ++ unsigned int h_requested_credits; ++ unsigned int saved_alloc_context; ++}; ++ ++enum jbd_state_bits { ++ BH_JBD = 16, ++ BH_JWrite = 17, ++ BH_Freed = 18, ++ BH_Revoked = 19, ++ BH_RevokeValid = 20, ++ BH_JBDDirty = 21, ++ BH_JournalHead = 22, ++ BH_Shadow = 23, ++ BH_Verified = 24, ++ BH_JBDPrivateStart = 25, ++}; ++ ++struct fscrypt_str { ++ unsigned char *name; ++ u32 len; ++}; ++ ++struct ext4_map_blocks { ++ ext4_fsblk_t m_pblk; ++ ext4_lblk_t m_lblk; ++ unsigned int m_len; ++ unsigned int m_flags; ++}; ++ ++enum { ++ EXT4_INODE_SECRM = 0, ++ EXT4_INODE_UNRM = 1, ++ EXT4_INODE_COMPR = 2, ++ EXT4_INODE_SYNC = 3, ++ EXT4_INODE_IMMUTABLE = 4, ++ EXT4_INODE_APPEND = 5, ++ EXT4_INODE_NODUMP = 6, ++ EXT4_INODE_NOATIME = 7, ++ EXT4_INODE_DIRTY = 8, ++ EXT4_INODE_COMPRBLK = 9, ++ EXT4_INODE_NOCOMPR = 10, ++ EXT4_INODE_ENCRYPT = 11, ++ EXT4_INODE_INDEX = 12, ++ EXT4_INODE_IMAGIC = 13, ++ EXT4_INODE_JOURNAL_DATA = 14, ++ EXT4_INODE_NOTAIL = 15, ++ EXT4_INODE_DIRSYNC = 16, ++ EXT4_INODE_TOPDIR = 17, ++ EXT4_INODE_HUGE_FILE = 18, ++ EXT4_INODE_EXTENTS = 19, ++ EXT4_INODE_VERITY = 20, ++ EXT4_INODE_EA_INODE = 21, ++ EXT4_INODE_DAX = 25, ++ EXT4_INODE_INLINE_DATA = 28, ++ EXT4_INODE_PROJINHERIT = 29, ++ EXT4_INODE_CASEFOLD = 30, ++ EXT4_INODE_RESERVED = 31, ++}; ++ ++struct extent_status { ++ struct rb_node rb_node; ++ ext4_lblk_t es_lblk; ++ ext4_lblk_t es_len; ++ ext4_fsblk_t es_pblk; ++}; ++ ++struct ext4_es_tree { ++ struct rb_root root; ++ struct extent_status *cache_es; ++}; ++ ++struct ext4_pending_tree { ++ struct rb_root root; ++}; ++ ++enum { ++ EXT4_FC_REASON_XATTR = 0, ++ EXT4_FC_REASON_CROSS_RENAME = 1, ++ EXT4_FC_REASON_JOURNAL_FLAG_CHANGE = 2, ++ EXT4_FC_REASON_NOMEM = 3, ++ EXT4_FC_REASON_SWAP_BOOT = 4, ++ EXT4_FC_REASON_RESIZE = 5, ++ EXT4_FC_REASON_RENAME_DIR = 6, ++ EXT4_FC_REASON_FALLOC_RANGE = 7, ++ EXT4_FC_REASON_INODE_JOURNAL_DATA = 8, ++ EXT4_FC_REASON_ENCRYPTED_FILENAME = 9, ++ EXT4_FC_REASON_MAX = 10, ++}; ++ ++struct ext4_inode_info { ++ __le32 i_data[15]; ++ __u32 i_dtime; ++ ext4_fsblk_t i_file_acl; ++ ext4_group_t i_block_group; ++ ext4_lblk_t i_dir_start_lookup; ++ long unsigned int i_flags; ++ struct rw_semaphore xattr_sem; ++ union { ++ struct list_head i_orphan; ++ unsigned int i_orphan_idx; ++ }; ++ struct list_head i_fc_dilist; ++ struct list_head i_fc_list; ++ ext4_lblk_t i_fc_lblk_start; ++ ext4_lblk_t i_fc_lblk_len; ++ atomic_t i_fc_updates; ++ wait_queue_head_t i_fc_wait; ++ struct mutex i_fc_lock; ++ loff_t i_disksize; ++ struct rw_semaphore i_data_sem; ++ struct inode vfs_inode; ++ struct jbd2_inode *jinode; ++ spinlock_t i_raw_lock; ++ struct timespec64 i_crtime; ++ atomic_t i_prealloc_active; ++ struct rb_root i_prealloc_node; ++ rwlock_t i_prealloc_lock; ++ struct ext4_es_tree i_es_tree; ++ rwlock_t i_es_lock; ++ struct list_head i_es_list; ++ unsigned int i_es_all_nr; ++ unsigned int i_es_shk_nr; ++ ext4_lblk_t i_es_shrink_lblk; ++ ext4_group_t i_last_alloc_group; ++ unsigned int i_reserved_data_blocks; ++ struct ext4_pending_tree i_pending_tree; ++ __u16 i_extra_isize; ++ u16 i_inline_off; ++ u16 i_inline_size; ++ qsize_t i_reserved_quota; ++ spinlock_t i_completed_io_lock; ++ struct list_head i_rsv_conversion_list; ++ struct work_struct i_rsv_conversion_work; ++ atomic_t i_unwritten; ++ spinlock_t i_block_reservation_lock; ++ tid_t i_sync_tid; ++ tid_t i_datasync_tid; ++ struct dquot *i_dquot[3]; ++ __u32 i_csum_seed; ++ kprojid_t i_projid; ++}; ++ ++enum ext4_journal_trigger_type { ++ EXT4_JTR_ORPHAN_FILE = 0, ++ EXT4_JTR_NONE = 1, ++}; ++ ++struct ext4_dir_entry_hash { ++ __le32 hash; ++ __le32 minor_hash; ++}; ++ ++struct ext4_dir_entry_2 { ++ __le32 inode; ++ __le16 rec_len; ++ __u8 name_len; ++ __u8 file_type; ++ char name[255]; ++}; ++ ++struct fname; ++ ++struct dir_private_info { ++ struct rb_root root; ++ struct rb_node *curr_node; ++ struct fname *extra_fname; ++ loff_t last_pos; ++ __u32 curr_hash; ++ __u32 curr_minor_hash; ++ __u32 next_hash; ++}; ++ ++struct fname { ++ __u32 hash; ++ __u32 minor_hash; ++ struct rb_node rb_hash; ++ struct fname *next; ++ __u32 inode; ++ __u8 name_len; ++ __u8 file_type; ++ char name[0]; ++}; ++ ++enum { ++ ES_WRITTEN_B = 0, ++ ES_UNWRITTEN_B = 1, ++ ES_DELAYED_B = 2, ++ ES_HOLE_B = 3, ++ ES_REFERENCED_B = 4, ++ ES_FLAGS = 5, ++}; ++ ++struct pending_reservation { ++ struct rb_node rb_node; ++ ext4_lblk_t lclu; ++}; ++ ++enum { ++ EXT4_STATE_NEW = 0, ++ EXT4_STATE_XATTR = 1, ++ EXT4_STATE_NO_EXPAND = 2, ++ EXT4_STATE_DA_ALLOC_CLOSE = 3, ++ EXT4_STATE_EXT_MIGRATE = 4, ++ EXT4_STATE_NEWENTRY = 5, ++ EXT4_STATE_MAY_INLINE_DATA = 6, ++ EXT4_STATE_EXT_PRECACHED = 7, ++ EXT4_STATE_LUSTRE_EA_INODE = 8, ++ EXT4_STATE_VERITY_IN_PROGRESS = 9, ++ EXT4_STATE_FC_COMMITTING = 10, ++ EXT4_STATE_ORPHAN_FILE = 11, ++}; ++ ++struct rsvd_count { ++ int ndelonly; ++ bool first_do_lblk_found; ++ ext4_lblk_t first_do_lblk; ++ ext4_lblk_t last_do_lblk; ++ struct extent_status *left_es; ++ bool partial; ++ ext4_lblk_t lclu; ++}; ++ ++struct ext4_locality_group { ++ struct mutex lg_mutex; ++ struct list_head lg_prealloc_list[10]; ++ spinlock_t lg_prealloc_lock; ++}; ++ ++struct fsmap { ++ __u32 fmr_device; ++ __u32 fmr_flags; ++ __u64 fmr_physical; ++ __u64 fmr_owner; ++ __u64 fmr_offset; ++ __u64 fmr_length; ++ __u64 fmr_reserved[3]; ++}; ++ ++struct ext4_fsmap { ++ struct list_head fmr_list; ++ dev_t fmr_device; ++ uint32_t fmr_flags; ++ uint64_t fmr_physical; ++ uint64_t fmr_owner; ++ uint64_t fmr_length; ++}; ++ ++struct ext4_fsmap_head { ++ uint32_t fmh_iflags; ++ uint32_t fmh_oflags; ++ unsigned int fmh_count; ++ unsigned int fmh_entries; ++ struct ext4_fsmap fmh_keys[2]; ++}; ++ ++typedef int (*ext4_fsmap_format_t)(struct ext4_fsmap *, void *); ++ ++typedef int (*ext4_mballoc_query_range_fn)(struct super_block *, ext4_group_t, ext4_grpblk_t, ext4_grpblk_t, void *); ++ ++typedef int (*list_cmp_func_t)(void *, const struct list_head *, const struct list_head *); ++ ++struct ext4_getfsmap_info { ++ struct ext4_fsmap_head *gfi_head; ++ ext4_fsmap_format_t gfi_formatter; ++ void *gfi_format_arg; ++ ext4_fsblk_t gfi_next_fsblk; ++ u32 gfi_dev; ++ ext4_group_t gfi_agno; ++ struct ext4_fsmap gfi_low; ++ struct ext4_fsmap gfi_high; ++ struct ext4_fsmap gfi_lastfree; ++ struct list_head gfi_meta_list; ++ bool gfi_last; ++}; ++ ++struct ext4_getfsmap_dev { ++ int (*gfd_fn)(struct super_block *, struct ext4_fsmap *, struct ext4_getfsmap_info *); ++ u32 gfd_dev; ++}; ++ ++struct ext4_allocation_request { ++ struct inode *inode; ++ unsigned int len; ++ ext4_lblk_t logical; ++ ext4_lblk_t lleft; ++ ext4_lblk_t lright; ++ ext4_fsblk_t goal; ++ ext4_fsblk_t pleft; ++ ext4_fsblk_t pright; ++ unsigned int flags; ++}; ++ ++typedef struct { ++ __le32 *p; ++ __le32 key; ++ struct buffer_head *bh; ++} Indirect; ++ ++struct fstrim_range { ++ __u64 start; ++ __u64 len; ++ __u64 minlen; ++}; ++ ++struct fileattr { ++ u32 flags; ++ u32 fsx_xflags; ++ u32 fsx_extsize; ++ u32 fsx_nextents; ++ u32 fsx_projid; ++ u32 fsx_cowextsize; ++ bool flags_valid: 1; ++ bool fsx_valid: 1; ++}; ++ ++typedef u64 compat_u64; ++ ++struct fiemap { ++ __u64 fm_start; ++ __u64 fm_length; ++ __u32 fm_flags; ++ __u32 fm_mapped_extents; ++ __u32 fm_extent_count; ++ __u32 fm_reserved; ++ struct fiemap_extent fm_extents[0]; ++}; ++ ++struct fsuuid { ++ __u32 fsu_len; ++ __u32 fsu_flags; ++ __u8 fsu_uuid[0]; ++}; ++ ++struct move_extent { ++ __u32 reserved; ++ __u32 donor_fd; ++ __u64 orig_start; ++ __u64 donor_start; ++ __u64 len; ++ __u64 moved_len; ++}; ++ ++struct ext4_new_group_input { ++ __u32 group; ++ __u64 block_bitmap; ++ __u64 inode_bitmap; ++ __u64 inode_table; ++ __u32 blocks_count; ++ __u16 reserved_blocks; ++ __u16 unused; ++}; ++ ++struct compat_ext4_new_group_input { ++ u32 group; ++ compat_u64 block_bitmap; ++ compat_u64 inode_bitmap; ++ compat_u64 inode_table; ++ u32 blocks_count; ++ u16 reserved_blocks; ++ u16 unused; ++}; ++ ++struct ext4_new_group_data { ++ __u32 group; ++ __u64 block_bitmap; ++ __u64 inode_bitmap; ++ __u64 inode_table; ++ __u32 blocks_count; ++ __u16 reserved_blocks; ++ __u16 mdata_blocks; ++ __u32 free_clusters_count; ++}; ++ ++struct ext4_inode { ++ __le16 i_mode; ++ __le16 i_uid; ++ __le32 i_size_lo; ++ __le32 i_atime; ++ __le32 i_ctime; ++ __le32 i_mtime; ++ __le32 i_dtime; ++ __le16 i_gid; ++ __le16 i_links_count; ++ __le32 i_blocks_lo; ++ __le32 i_flags; ++ union { ++ struct { ++ __le32 l_i_version; ++ } linux1; ++ struct { ++ __u32 h_i_translator; ++ } hurd1; ++ struct { ++ __u32 m_i_reserved1; ++ } masix1; ++ } osd1; ++ __le32 i_block[15]; ++ __le32 i_generation; ++ __le32 i_file_acl_lo; ++ __le32 i_size_high; ++ __le32 i_obso_faddr; ++ union { ++ struct { ++ __le16 l_i_blocks_high; ++ __le16 l_i_file_acl_high; ++ __le16 l_i_uid_high; ++ __le16 l_i_gid_high; ++ __le16 l_i_checksum_lo; ++ __le16 l_i_reserved; ++ } linux2; ++ struct { ++ __le16 h_i_reserved1; ++ __u16 h_i_mode_high; ++ __u16 h_i_uid_high; ++ __u16 h_i_gid_high; ++ __u32 h_i_author; ++ } hurd2; ++ struct { ++ __le16 h_i_reserved1; ++ __le16 m_i_file_acl_high; ++ __u32 m_i_reserved2[2]; ++ } masix2; ++ } osd2; ++ __le16 i_extra_isize; ++ __le16 i_checksum_hi; ++ __le32 i_ctime_extra; ++ __le32 i_mtime_extra; ++ __le32 i_atime_extra; ++ __le32 i_crtime; ++ __le32 i_crtime_extra; ++ __le32 i_version_hi; ++ __le32 i_projid; ++}; ++ ++struct ext4_iloc { ++ struct buffer_head *bh; ++ long unsigned int offset; ++ ext4_group_t block_group; ++}; ++ ++typedef enum { ++ EXT4_IGET_NORMAL = 0, ++ EXT4_IGET_SPECIAL = 1, ++ EXT4_IGET_HANDLE = 2, ++ EXT4_IGET_BAD = 4, ++ EXT4_IGET_EA_INODE = 8, ++} ext4_iget_flags; ++ ++struct fsmap_head { ++ __u32 fmh_iflags; ++ __u32 fmh_oflags; ++ __u32 fmh_count; ++ __u32 fmh_entries; ++ __u64 fmh_reserved[6]; ++ struct fsmap fmh_keys[2]; ++ struct fsmap fmh_recs[0]; ++}; ++ ++typedef void ext4_update_sb_callback(struct ext4_super_block *, const void *); ++ ++struct getfsmap_info { ++ struct super_block *gi_sb; ++ struct fsmap_head *gi_data; ++ unsigned int gi_idx; ++ __u32 gi_last_flags; ++}; ++ ++struct mmp_struct { ++ __le32 mmp_magic; ++ __le32 mmp_seq; ++ __le64 mmp_time; ++ char mmp_nodename[64]; ++ char mmp_bdevname[32]; ++ __le16 mmp_check_interval; ++ __le16 mmp_pad1; ++ __le32 mmp_pad2[226]; ++ __le32 mmp_checksum; ++}; ++ ++struct fscrypt_name { ++ const struct qstr *usr_fname; ++ struct fscrypt_str disk_name; ++ u32 hash; ++ u32 minor_hash; ++ struct fscrypt_str crypto_buf; ++ bool is_nokey_name; ++}; ++ ++enum { ++ EXT4_MF_MNTDIR_SAMPLED = 0, ++ EXT4_MF_FS_ABORTED = 1, ++ EXT4_MF_FC_INELIGIBLE = 2, ++}; ++ ++struct ext4_dir_entry { ++ __le32 inode; ++ __le16 rec_len; ++ __le16 name_len; ++ char name[255]; ++}; ++ ++struct ext4_dir_entry_tail { ++ __le32 det_reserved_zero1; ++ __le16 det_rec_len; ++ __u8 det_reserved_zero2; ++ __u8 det_reserved_ft; ++ __le32 det_checksum; ++}; ++ ++struct dx_hash_info { ++ u32 hash; ++ u32 minor_hash; ++ int hash_version; ++ u32 *seed; ++}; ++ ++struct ext4_filename { ++ const struct qstr *usr_fname; ++ struct fscrypt_str disk_name; ++ struct dx_hash_info hinfo; ++ struct fscrypt_str crypto_buf; ++ struct fscrypt_str cf_name; ++}; ++ ++typedef enum { ++ EITHER = 0, ++ INDEX = 1, ++ DIRENT = 2, ++ DIRENT_HTREE = 3, ++} dirblock_type_t; ++ ++struct fake_dirent { ++ __le32 inode; ++ __le16 rec_len; ++ u8 name_len; ++ u8 file_type; ++}; ++ ++struct dx_countlimit { ++ __le16 limit; ++ __le16 count; ++}; ++ ++struct dx_entry { ++ __le32 hash; ++ __le32 block; ++}; ++ ++struct dx_root_info { ++ __le32 reserved_zero; ++ u8 hash_version; ++ u8 info_length; ++ u8 indirect_levels; ++ u8 unused_flags; ++}; ++ ++struct dx_root { ++ struct fake_dirent dot; ++ char dot_name[4]; ++ struct fake_dirent dotdot; ++ char dotdot_name[4]; ++ struct dx_root_info info; ++ struct dx_entry entries[0]; ++}; ++ ++struct dx_node { ++ struct fake_dirent fake; ++ struct dx_entry entries[0]; ++}; ++ ++struct dx_frame { ++ struct buffer_head *bh; ++ struct dx_entry *entries; ++ struct dx_entry *at; ++}; ++ ++struct dx_map_entry { ++ u32 hash; ++ u16 offs; ++ u16 size; ++}; ++ ++struct dx_tail { ++ u32 dt_reserved; ++ __le32 dt_checksum; ++}; ++ ++struct ext4_renament { ++ struct inode *dir; ++ struct dentry *dentry; ++ struct inode *inode; ++ bool is_dir; ++ int dir_nlink_delta; ++ struct buffer_head *bh; ++ struct ext4_dir_entry_2 *de; ++ int inlined; ++ struct buffer_head *dir_bh; ++ struct ext4_dir_entry_2 *parent_de; ++ int dir_inlined; ++}; ++ ++enum bio_post_read_step { ++ STEP_INITIAL = 0, ++ STEP_DECRYPT = 1, ++ STEP_VERITY = 2, ++ STEP_MAX = 3, ++}; ++ ++struct bio_post_read_ctx { ++ struct bio *bio; ++ struct work_struct work; ++ unsigned int cur_step; ++ unsigned int enabled_steps; ++}; ++ ++enum { ++ MBE_REFERENCED_B = 0, ++ MBE_REUSABLE_B = 1, ++}; ++ ++struct mb_cache_entry { ++ struct list_head e_list; ++ struct hlist_bl_node e_hash_list; ++ atomic_t e_refcnt; ++ u32 e_key; ++ long unsigned int e_flags; ++ u64 e_value; ++}; ++ ++struct ext4_xattr_header { ++ __le32 h_magic; ++ __le32 h_refcount; ++ __le32 h_blocks; ++ __le32 h_hash; ++ __le32 h_checksum; ++ __u32 h_reserved[3]; ++}; ++ ++struct ext4_xattr_ibody_header { ++ __le32 h_magic; ++}; ++ ++struct ext4_xattr_entry { ++ __u8 e_name_len; ++ __u8 e_name_index; ++ __le16 e_value_offs; ++ __le32 e_value_inum; ++ __le32 e_value_size; ++ __le32 e_hash; ++ char e_name[0]; ++}; ++ ++struct ext4_xattr_info { ++ const char *name; ++ const void *value; ++ size_t value_len; ++ int name_index; ++ int in_inode; ++}; ++ ++struct ext4_xattr_search { ++ struct ext4_xattr_entry *first; ++ void *base; ++ void *end; ++ struct ext4_xattr_entry *here; ++ int not_found; ++}; ++ ++struct ext4_xattr_ibody_find { ++ struct ext4_xattr_search s; ++ struct ext4_iloc iloc; ++}; ++ ++struct ext4_xattr_inode_array { ++ unsigned int count; ++ struct inode *inodes[0]; ++}; ++ ++struct ext4_xattr_block_find { ++ struct ext4_xattr_search s; ++ struct buffer_head *bh; ++}; ++ ++struct ext4_orphan_block_tail { ++ __le32 ob_magic; ++ __le32 ob_checksum; ++}; ++ ++struct xattr; ++ ++typedef int (*initxattrs)(struct inode *, const struct xattr *, void *); ++ ++struct xattr { ++ const char *name; ++ void *value; ++ size_t value_len; ++}; ++ ++struct commit_header { ++ __be32 h_magic; ++ __be32 h_blocktype; ++ __be32 h_sequence; ++ unsigned char h_chksum_type; ++ unsigned char h_chksum_size; ++ unsigned char h_padding[2]; ++ __be32 h_chksum[8]; ++ __be64 h_commit_sec; ++ __be32 h_commit_nsec; ++}; ++ ++struct journal_block_tag3_s { ++ __be32 t_blocknr; ++ __be32 t_flags; ++ __be32 t_blocknr_high; ++ __be32 t_checksum; ++}; ++ ++typedef struct journal_block_tag3_s journal_block_tag3_t; ++ ++struct journal_block_tag_s { ++ __be32 t_blocknr; ++ __be16 t_checksum; ++ __be16 t_flags; ++ __be32 t_blocknr_high; ++}; ++ ++typedef struct journal_block_tag_s journal_block_tag_t; ++ ++struct jbd2_journal_block_tail { ++ __be32 t_checksum; ++}; ++ ++struct trace_event_raw_jbd2_checkpoint { ++ struct trace_entry ent; ++ dev_t dev; ++ int result; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_jbd2_commit { ++ struct trace_entry ent; ++ dev_t dev; ++ char sync_commit; ++ tid_t transaction; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_jbd2_end_commit { ++ struct trace_entry ent; ++ dev_t dev; ++ char sync_commit; ++ tid_t transaction; ++ tid_t head; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_jbd2_submit_inode_data { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_jbd2_handle_start_class { ++ struct trace_entry ent; ++ dev_t dev; ++ tid_t tid; ++ unsigned int type; ++ unsigned int line_no; ++ int requested_blocks; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_jbd2_handle_extend { ++ struct trace_entry ent; ++ dev_t dev; ++ tid_t tid; ++ unsigned int type; ++ unsigned int line_no; ++ int buffer_credits; ++ int requested_blocks; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_jbd2_handle_stats { ++ struct trace_entry ent; ++ dev_t dev; ++ tid_t tid; ++ unsigned int type; ++ unsigned int line_no; ++ int interval; ++ int sync; ++ int requested_blocks; ++ int dirtied_blocks; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_jbd2_run_stats { ++ struct trace_entry ent; ++ dev_t dev; ++ tid_t tid; ++ long unsigned int wait; ++ long unsigned int request_delay; ++ long unsigned int running; ++ long unsigned int locked; ++ long unsigned int flushing; ++ long unsigned int logging; ++ __u32 handle_count; ++ __u32 blocks; ++ __u32 blocks_logged; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_jbd2_checkpoint_stats { ++ struct trace_entry ent; ++ dev_t dev; ++ tid_t tid; ++ long unsigned int chp_time; ++ __u32 forced_to_close; ++ __u32 written; ++ __u32 dropped; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_jbd2_update_log_tail { ++ struct trace_entry ent; ++ dev_t dev; ++ tid_t tail_sequence; ++ tid_t first_tid; ++ long unsigned int block_nr; ++ long unsigned int freed; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_jbd2_write_superblock { ++ struct trace_entry ent; ++ dev_t dev; ++ blk_opf_t write_flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_jbd2_lock_buffer_stall { ++ struct trace_entry ent; ++ dev_t dev; ++ long unsigned int stall_ms; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_jbd2_journal_shrink { ++ struct trace_entry ent; ++ dev_t dev; ++ long unsigned int nr_to_scan; ++ long unsigned int count; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_jbd2_shrink_scan_exit { ++ struct trace_entry ent; ++ dev_t dev; ++ long unsigned int nr_to_scan; ++ long unsigned int nr_shrunk; ++ long unsigned int count; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_jbd2_shrink_checkpoint_list { ++ struct trace_entry ent; ++ dev_t dev; ++ tid_t first_tid; ++ tid_t tid; ++ tid_t last_tid; ++ long unsigned int nr_freed; ++ tid_t next_tid; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_jbd2_checkpoint {}; ++ ++struct trace_event_data_offsets_jbd2_commit {}; ++ ++struct trace_event_data_offsets_jbd2_end_commit {}; ++ ++struct trace_event_data_offsets_jbd2_submit_inode_data {}; ++ ++struct trace_event_data_offsets_jbd2_handle_start_class {}; ++ ++struct trace_event_data_offsets_jbd2_handle_extend {}; ++ ++struct trace_event_data_offsets_jbd2_handle_stats {}; ++ ++struct trace_event_data_offsets_jbd2_run_stats {}; ++ ++struct trace_event_data_offsets_jbd2_checkpoint_stats {}; ++ ++struct trace_event_data_offsets_jbd2_update_log_tail {}; ++ ++struct trace_event_data_offsets_jbd2_write_superblock {}; ++ ++struct trace_event_data_offsets_jbd2_lock_buffer_stall {}; ++ ++struct trace_event_data_offsets_jbd2_journal_shrink {}; ++ ++struct trace_event_data_offsets_jbd2_shrink_scan_exit {}; ++ ++struct trace_event_data_offsets_jbd2_shrink_checkpoint_list {}; ++ ++typedef void (*btf_trace_jbd2_checkpoint)(void *, journal_t *, int); ++ ++typedef void (*btf_trace_jbd2_start_commit)(void *, journal_t *, transaction_t *); ++ ++typedef void (*btf_trace_jbd2_commit_locking)(void *, journal_t *, transaction_t *); ++ ++typedef void (*btf_trace_jbd2_commit_flushing)(void *, journal_t *, transaction_t *); ++ ++typedef void (*btf_trace_jbd2_commit_logging)(void *, journal_t *, transaction_t *); ++ ++typedef void (*btf_trace_jbd2_drop_transaction)(void *, journal_t *, transaction_t *); ++ ++typedef void (*btf_trace_jbd2_end_commit)(void *, journal_t *, transaction_t *); ++ ++typedef void (*btf_trace_jbd2_submit_inode_data)(void *, struct inode *); ++ ++typedef void (*btf_trace_jbd2_handle_start)(void *, dev_t, tid_t, unsigned int, unsigned int, int); ++ ++typedef void (*btf_trace_jbd2_handle_restart)(void *, dev_t, tid_t, unsigned int, unsigned int, int); ++ ++typedef void (*btf_trace_jbd2_handle_extend)(void *, dev_t, tid_t, unsigned int, unsigned int, int, int); ++ ++typedef void (*btf_trace_jbd2_handle_stats)(void *, dev_t, tid_t, unsigned int, unsigned int, int, int, int, int); ++ ++typedef void (*btf_trace_jbd2_run_stats)(void *, dev_t, tid_t, struct transaction_run_stats_s *); ++ ++typedef void (*btf_trace_jbd2_checkpoint_stats)(void *, dev_t, tid_t, struct transaction_chp_stats_s *); ++ ++typedef void (*btf_trace_jbd2_update_log_tail)(void *, journal_t *, tid_t, long unsigned int, long unsigned int); ++ ++typedef void (*btf_trace_jbd2_write_superblock)(void *, journal_t *, blk_opf_t); ++ ++typedef void (*btf_trace_jbd2_lock_buffer_stall)(void *, dev_t, long unsigned int); ++ ++typedef void (*btf_trace_jbd2_shrink_count)(void *, journal_t *, long unsigned int, long unsigned int); ++ ++typedef void (*btf_trace_jbd2_shrink_scan_enter)(void *, journal_t *, long unsigned int, long unsigned int); ++ ++typedef void (*btf_trace_jbd2_shrink_scan_exit)(void *, journal_t *, long unsigned int, long unsigned int, long unsigned int); ++ ++typedef void (*btf_trace_jbd2_shrink_checkpoint_list)(void *, journal_t *, tid_t, tid_t, tid_t, long unsigned int, tid_t); ++ ++struct jbd2_stats_proc_session { ++ journal_t *journal; ++ struct transaction_stats_s *stats; ++ int start; ++ int max; ++}; ++ ++typedef u16 wchar_t; ++ ++typedef u32 unicode_t; ++ ++struct nls_table { ++ const char *charset; ++ const char *alias; ++ int (*uni2char)(wchar_t, unsigned char *, int); ++ int (*char2uni)(const unsigned char *, int, wchar_t *); ++ const unsigned char *charset2lower; ++ const unsigned char *charset2upper; ++ struct module *owner; ++ struct nls_table *next; ++}; ++ ++enum utf16_endian { ++ UTF16_HOST_ENDIAN = 0, ++ UTF16_LITTLE_ENDIAN = 1, ++ UTF16_BIG_ENDIAN = 2, ++}; ++ ++struct utf8_table { ++ int cmask; ++ int cval; ++ int shift; ++ long int lmask; ++ long int lval; ++}; ++ ++struct utf8cursor { ++ const struct unicode_map *um; ++ enum utf8_normalization n; ++ const char *s; ++ const char *p; ++ const char *ss; ++ const char *sp; ++ unsigned int len; ++ unsigned int slen; ++ short int ccc; ++ short int nccc; ++ unsigned char hangul[12]; ++}; ++ ++typedef const unsigned char utf8trie_t; ++ ++typedef const unsigned char utf8leaf_t; ++ ++struct match_token { ++ int token; ++ const char *pattern; ++}; ++ ++enum { ++ MAX_OPT_ARGS = 3, ++}; ++ ++typedef unsigned int autofs_wqt_t; ++ ++struct autofs_sb_info; ++ ++struct autofs_info { ++ struct dentry *dentry; ++ int flags; ++ struct completion expire_complete; ++ struct list_head active; ++ struct list_head expiring; ++ struct autofs_sb_info *sbi; ++ long unsigned int last_used; ++ int count; ++ kuid_t uid; ++ kgid_t gid; ++ struct callback_head rcu; ++}; ++ ++struct autofs_wait_queue; ++ ++struct autofs_sb_info { ++ u32 magic; ++ int pipefd; ++ struct file *pipe; ++ struct pid *oz_pgrp; ++ int version; ++ int sub_version; ++ int min_proto; ++ int max_proto; ++ unsigned int flags; ++ long unsigned int exp_timeout; ++ unsigned int type; ++ struct super_block *sb; ++ struct mutex wq_mutex; ++ struct mutex pipe_mutex; ++ spinlock_t fs_lock; ++ struct autofs_wait_queue *queues; ++ spinlock_t lookup_lock; ++ struct list_head active_list; ++ struct list_head expiring_list; ++ struct callback_head rcu; ++}; ++ ++struct autofs_wait_queue { ++ wait_queue_head_t queue; ++ struct autofs_wait_queue *next; ++ autofs_wqt_t wait_queue_token; ++ struct qstr name; ++ u32 offset; ++ u32 dev; ++ u64 ino; ++ kuid_t uid; ++ kgid_t gid; ++ pid_t pid; ++ pid_t tgid; ++ int status; ++ unsigned int wait_ctr; ++}; ++ ++enum { ++ Opt_err = 0, ++ Opt_fd = 1, ++ Opt_uid___2 = 2, ++ Opt_gid___2 = 3, ++ Opt_pgrp = 4, ++ Opt_minproto = 5, ++ Opt_maxproto = 6, ++ Opt_indirect = 7, ++ Opt_direct = 8, ++ Opt_offset = 9, ++ Opt_strictexpire = 10, ++ Opt_ignore = 11, ++}; ++ ++struct autofs_packet_hdr { ++ int proto_version; ++ int type; ++}; ++ ++struct autofs_packet_missing { ++ struct autofs_packet_hdr hdr; ++ autofs_wqt_t wait_queue_token; ++ int len; ++ char name[256]; ++}; ++ ++struct autofs_packet_expire { ++ struct autofs_packet_hdr hdr; ++ int len; ++ char name[256]; ++}; ++ ++enum autofs_notify { ++ NFY_NONE = 0, ++ NFY_MOUNT = 1, ++ NFY_EXPIRE = 2, ++}; ++ ++struct autofs_packet_expire_multi { ++ struct autofs_packet_hdr hdr; ++ autofs_wqt_t wait_queue_token; ++ int len; ++ char name[256]; ++}; ++ ++union autofs_packet_union { ++ struct autofs_packet_hdr hdr; ++ struct autofs_packet_missing missing; ++ struct autofs_packet_expire expire; ++ struct autofs_packet_expire_multi expire_multi; ++}; ++ ++struct autofs_v5_packet { ++ struct autofs_packet_hdr hdr; ++ autofs_wqt_t wait_queue_token; ++ __u32 dev; ++ __u64 ino; ++ __u32 uid; ++ __u32 gid; ++ __u32 pid; ++ __u32 tgid; ++ __u32 len; ++ char name[256]; ++}; ++ ++typedef struct autofs_v5_packet autofs_packet_missing_indirect_t; ++ ++typedef struct autofs_v5_packet autofs_packet_expire_indirect_t; ++ ++typedef struct autofs_v5_packet autofs_packet_missing_direct_t; ++ ++typedef struct autofs_v5_packet autofs_packet_expire_direct_t; ++ ++union autofs_v5_packet_union { ++ struct autofs_packet_hdr hdr; ++ struct autofs_v5_packet v5_packet; ++ autofs_packet_missing_indirect_t missing_indirect; ++ autofs_packet_expire_indirect_t expire_indirect; ++ autofs_packet_missing_direct_t missing_direct; ++ autofs_packet_expire_direct_t expire_direct; ++}; ++ ++struct debugfs_blob_wrapper { ++ void *data; ++ long unsigned int size; ++}; ++ ++struct debugfs_reg32 { ++ char *name; ++ long unsigned int offset; ++}; ++ ++struct debugfs_regset32 { ++ const struct debugfs_reg32 *regs; ++ int nregs; ++ void *base; ++ struct device *dev; ++}; ++ ++struct debugfs_u32_array { ++ u32 *array; ++ u32 n_elements; ++}; ++ ++struct debugfs_fsdata { ++ const struct file_operations *real_fops; ++ refcount_t active_users; ++ struct completion active_users_drained; ++}; ++ ++struct debugfs_devm_entry { ++ int (*read)(struct seq_file *, void *); ++ struct device *dev; ++}; ++ ++struct miscdevice { ++ int minor; ++ const char *name; ++ const struct file_operations *fops; ++ struct list_head list; ++ struct device *parent; ++ struct device *this_device; ++ const struct attribute_group **groups; ++ const char *nodename; ++ umode_t mode; ++}; ++ ++struct btrfs_ioctl_vol_args { ++ __s64 fd; ++ char name[4088]; ++}; ++ ++struct btrfs_qgroup_limit { ++ __u64 flags; ++ __u64 max_rfer; ++ __u64 max_excl; ++ __u64 rsv_rfer; ++ __u64 rsv_excl; ++}; ++ ++struct btrfs_qgroup_inherit { ++ __u64 flags; ++ __u64 num_qgroups; ++ __u64 num_ref_copies; ++ __u64 num_excl_copies; ++ struct btrfs_qgroup_limit lim; ++ __u64 qgroups[0]; ++}; ++ ++struct btrfs_scrub_progress { ++ __u64 data_extents_scrubbed; ++ __u64 tree_extents_scrubbed; ++ __u64 data_bytes_scrubbed; ++ __u64 tree_bytes_scrubbed; ++ __u64 read_errors; ++ __u64 csum_errors; ++ __u64 verify_errors; ++ __u64 no_csum; ++ __u64 csum_discards; ++ __u64 super_errors; ++ __u64 malloc_errors; ++ __u64 uncorrectable_errors; ++ __u64 corrected_errors; ++ __u64 last_physical; ++ __u64 unverified_errors; ++}; ++ ++struct btrfs_balance_args { ++ __u64 profiles; ++ union { ++ __u64 usage; ++ struct { ++ __u32 usage_min; ++ __u32 usage_max; ++ }; ++ }; ++ __u64 devid; ++ __u64 pstart; ++ __u64 pend; ++ __u64 vstart; ++ __u64 vend; ++ __u64 target; ++ __u64 flags; ++ union { ++ __u64 limit; ++ struct { ++ __u32 limit_min; ++ __u32 limit_max; ++ }; ++ }; ++ __u32 stripes_min; ++ __u32 stripes_max; ++ __u64 unused[6]; ++}; ++ ++struct btrfs_balance_progress { ++ __u64 expected; ++ __u64 considered; ++ __u64 completed; ++}; ++ ++enum btrfs_dev_stat_values { ++ BTRFS_DEV_STAT_WRITE_ERRS = 0, ++ BTRFS_DEV_STAT_READ_ERRS = 1, ++ BTRFS_DEV_STAT_FLUSH_ERRS = 2, ++ BTRFS_DEV_STAT_CORRUPTION_ERRS = 3, ++ BTRFS_DEV_STAT_GENERATION_ERRS = 4, ++ BTRFS_DEV_STAT_VALUES_MAX = 5, ++}; ++ ++struct btrfs_disk_key { ++ __le64 objectid; ++ __u8 type; ++ __le64 offset; ++} __attribute__((packed)); ++ ++struct btrfs_key { ++ __u64 objectid; ++ __u8 type; ++ __u64 offset; ++} __attribute__((packed)); ++ ++struct btrfs_header { ++ __u8 csum[32]; ++ __u8 fsid[16]; ++ __le64 bytenr; ++ __le64 flags; ++ __u8 chunk_tree_uuid[16]; ++ __le64 generation; ++ __le64 owner; ++ __le32 nritems; ++ __u8 level; ++} __attribute__((packed)); ++ ++struct btrfs_root_backup { ++ __le64 tree_root; ++ __le64 tree_root_gen; ++ __le64 chunk_root; ++ __le64 chunk_root_gen; ++ __le64 extent_root; ++ __le64 extent_root_gen; ++ __le64 fs_root; ++ __le64 fs_root_gen; ++ __le64 dev_root; ++ __le64 dev_root_gen; ++ __le64 csum_root; ++ __le64 csum_root_gen; ++ __le64 total_bytes; ++ __le64 bytes_used; ++ __le64 num_devices; ++ __le64 unused_64[4]; ++ __u8 tree_root_level; ++ __u8 chunk_root_level; ++ __u8 extent_root_level; ++ __u8 fs_root_level; ++ __u8 dev_root_level; ++ __u8 csum_root_level; ++ __u8 unused_8[10]; ++}; ++ ++struct btrfs_item { ++ struct btrfs_disk_key key; ++ __le32 offset; ++ __le32 size; ++} __attribute__((packed)); ++ ++struct btrfs_dev_item { ++ __le64 devid; ++ __le64 total_bytes; ++ __le64 bytes_used; ++ __le32 io_align; ++ __le32 io_width; ++ __le32 sector_size; ++ __le64 type; ++ __le64 generation; ++ __le64 start_offset; ++ __le32 dev_group; ++ __u8 seek_speed; ++ __u8 bandwidth; ++ __u8 uuid[16]; ++ __u8 fsid[16]; ++} __attribute__((packed)); ++ ++struct btrfs_super_block { ++ __u8 csum[32]; ++ __u8 fsid[16]; ++ __le64 bytenr; ++ __le64 flags; ++ __le64 magic; ++ __le64 generation; ++ __le64 root; ++ __le64 chunk_root; ++ __le64 log_root; ++ __le64 __unused_log_root_transid; ++ __le64 total_bytes; ++ __le64 bytes_used; ++ __le64 root_dir_objectid; ++ __le64 num_devices; ++ __le32 sectorsize; ++ __le32 nodesize; ++ __le32 __unused_leafsize; ++ __le32 stripesize; ++ __le32 sys_chunk_array_size; ++ __le64 chunk_root_generation; ++ __le64 compat_flags; ++ __le64 compat_ro_flags; ++ __le64 incompat_flags; ++ __le16 csum_type; ++ __u8 root_level; ++ __u8 chunk_root_level; ++ __u8 log_root_level; ++ struct btrfs_dev_item dev_item; ++ char label[256]; ++ __le64 cache_generation; ++ __le64 uuid_tree_generation; ++ __u8 metadata_uuid[16]; ++ __u64 nr_global_roots; ++ __le64 reserved[27]; ++ __u8 sys_chunk_array[2048]; ++ struct btrfs_root_backup super_roots[4]; ++ __u8 padding[565]; ++} __attribute__((packed)); ++ ++struct btrfs_inode_ref { ++ __le64 index; ++ __le16 name_len; ++} __attribute__((packed)); ++ ++struct btrfs_timespec { ++ __le64 sec; ++ __le32 nsec; ++} __attribute__((packed)); ++ ++struct btrfs_inode_item { ++ __le64 generation; ++ __le64 transid; ++ __le64 size; ++ __le64 nbytes; ++ __le64 block_group; ++ __le32 nlink; ++ __le32 uid; ++ __le32 gid; ++ __le32 mode; ++ __le64 rdev; ++ __le64 flags; ++ __le64 sequence; ++ __le64 reserved[4]; ++ struct btrfs_timespec atime; ++ struct btrfs_timespec ctime; ++ struct btrfs_timespec mtime; ++ struct btrfs_timespec otime; ++}; ++ ++struct btrfs_dir_item { ++ struct btrfs_disk_key location; ++ __le64 transid; ++ __le16 data_len; ++ __le16 name_len; ++ __u8 type; ++} __attribute__((packed)); ++ ++struct btrfs_root_item { ++ struct btrfs_inode_item inode; ++ __le64 generation; ++ __le64 root_dirid; ++ __le64 bytenr; ++ __le64 byte_limit; ++ __le64 bytes_used; ++ __le64 last_snapshot; ++ __le64 flags; ++ __le32 refs; ++ struct btrfs_disk_key drop_progress; ++ __u8 drop_level; ++ __u8 level; ++ __le64 generation_v2; ++ __u8 uuid[16]; ++ __u8 parent_uuid[16]; ++ __u8 received_uuid[16]; ++ __le64 ctransid; ++ __le64 otransid; ++ __le64 stransid; ++ __le64 rtransid; ++ struct btrfs_timespec ctime; ++ struct btrfs_timespec otime; ++ struct btrfs_timespec stime; ++ struct btrfs_timespec rtime; ++ __le64 reserved[8]; ++} __attribute__((packed)); ++ ++struct btrfs_root_ref { ++ __le64 dirid; ++ __le64 sequence; ++ __le16 name_len; ++} __attribute__((packed)); ++ ++enum { ++ BTRFS_FILE_EXTENT_INLINE = 0, ++ BTRFS_FILE_EXTENT_REG = 1, ++ BTRFS_FILE_EXTENT_PREALLOC = 2, ++ BTRFS_NR_FILE_EXTENT_TYPES = 3, ++}; ++ ++struct btrfs_file_extent_item { ++ __le64 generation; ++ __le64 ram_bytes; ++ __u8 compression; ++ __u8 encryption; ++ __le16 other_encoding; ++ __u8 type; ++ __le64 disk_bytenr; ++ __le64 disk_num_bytes; ++ __le64 offset; ++ __le64 num_bytes; ++} __attribute__((packed)); ++ ++enum { ++ __EXTENT_DIRTY_BIT = 0, ++ EXTENT_DIRTY = 1, ++ __EXTENT_DIRTY_SEQ = 0, ++ __EXTENT_UPTODATE_BIT = 1, ++ EXTENT_UPTODATE = 2, ++ __EXTENT_UPTODATE_SEQ = 1, ++ __EXTENT_LOCKED_BIT = 2, ++ EXTENT_LOCKED = 4, ++ __EXTENT_LOCKED_SEQ = 2, ++ __EXTENT_NEW_BIT = 3, ++ EXTENT_NEW = 8, ++ __EXTENT_NEW_SEQ = 3, ++ __EXTENT_DELALLOC_BIT = 4, ++ EXTENT_DELALLOC = 16, ++ __EXTENT_DELALLOC_SEQ = 4, ++ __EXTENT_DEFRAG_BIT = 5, ++ EXTENT_DEFRAG = 32, ++ __EXTENT_DEFRAG_SEQ = 5, ++ __EXTENT_BOUNDARY_BIT = 6, ++ EXTENT_BOUNDARY = 64, ++ __EXTENT_BOUNDARY_SEQ = 6, ++ __EXTENT_NODATASUM_BIT = 7, ++ EXTENT_NODATASUM = 128, ++ __EXTENT_NODATASUM_SEQ = 7, ++ __EXTENT_CLEAR_META_RESV_BIT = 8, ++ EXTENT_CLEAR_META_RESV = 256, ++ __EXTENT_CLEAR_META_RESV_SEQ = 8, ++ __EXTENT_NEED_WAIT_BIT = 9, ++ EXTENT_NEED_WAIT = 512, ++ __EXTENT_NEED_WAIT_SEQ = 9, ++ __EXTENT_NORESERVE_BIT = 10, ++ EXTENT_NORESERVE = 1024, ++ __EXTENT_NORESERVE_SEQ = 10, ++ __EXTENT_QGROUP_RESERVED_BIT = 11, ++ EXTENT_QGROUP_RESERVED = 2048, ++ __EXTENT_QGROUP_RESERVED_SEQ = 11, ++ __EXTENT_CLEAR_DATA_RESV_BIT = 12, ++ EXTENT_CLEAR_DATA_RESV = 4096, ++ __EXTENT_CLEAR_DATA_RESV_SEQ = 12, ++ __EXTENT_DELALLOC_NEW_BIT = 13, ++ EXTENT_DELALLOC_NEW = 8192, ++ __EXTENT_DELALLOC_NEW_SEQ = 13, ++ __EXTENT_ADD_INODE_BYTES_BIT = 14, ++ EXTENT_ADD_INODE_BYTES = 16384, ++ __EXTENT_ADD_INODE_BYTES_SEQ = 14, ++ __EXTENT_CLEAR_ALL_BITS_BIT = 15, ++ EXTENT_CLEAR_ALL_BITS = 32768, ++ __EXTENT_CLEAR_ALL_BITS_SEQ = 15, ++}; ++ ++enum { ++ IO_TREE_FS_PINNED_EXTENTS = 0, ++ IO_TREE_FS_EXCLUDED_EXTENTS = 1, ++ IO_TREE_BTREE_INODE_IO = 2, ++ IO_TREE_INODE_IO = 3, ++ IO_TREE_RELOC_BLOCKS = 4, ++ IO_TREE_TRANS_DIRTY_PAGES = 5, ++ IO_TREE_ROOT_DIRTY_LOG_PAGES = 6, ++ IO_TREE_INODE_FILE_EXTENT = 7, ++ IO_TREE_LOG_CSUM_RANGE = 8, ++ IO_TREE_SELFTEST = 9, ++ IO_TREE_DEVICE_ALLOC_STATE = 10, ++}; ++ ++struct btrfs_fs_info; ++ ++struct btrfs_inode; ++ ++struct extent_io_tree { ++ struct rb_root state; ++ struct btrfs_fs_info *fs_info; ++ struct btrfs_inode *inode; ++ u8 owner; ++ spinlock_t lock; ++}; ++ ++struct extent_map_tree { ++ struct rb_root_cached map; ++ struct list_head modified_extents; ++ rwlock_t lock; ++}; ++ ++enum btrfs_rsv_type { ++ BTRFS_BLOCK_RSV_GLOBAL = 0, ++ BTRFS_BLOCK_RSV_DELALLOC = 1, ++ BTRFS_BLOCK_RSV_TRANS = 2, ++ BTRFS_BLOCK_RSV_CHUNK = 3, ++ BTRFS_BLOCK_RSV_DELOPS = 4, ++ BTRFS_BLOCK_RSV_DELREFS = 5, ++ BTRFS_BLOCK_RSV_EMPTY = 6, ++ BTRFS_BLOCK_RSV_TEMP = 7, ++}; ++ ++struct btrfs_space_info; ++ ++struct btrfs_block_rsv { ++ u64 size; ++ u64 reserved; ++ struct btrfs_space_info *space_info; ++ spinlock_t lock; ++ bool full; ++ bool failfast; ++ enum btrfs_rsv_type type: 8; ++ u64 qgroup_rsv_size; ++ u64 qgroup_rsv_reserved; ++}; ++ ++struct btrfs_block_group; ++ ++struct btrfs_free_cluster { ++ spinlock_t lock; ++ spinlock_t refill_lock; ++ struct rb_root root; ++ u64 max_size; ++ u64 window_start; ++ bool fragmented; ++ struct btrfs_block_group *block_group; ++ struct list_head block_group_list; ++}; ++ ++struct btrfs_discard_ctl { ++ struct workqueue_struct *discard_workers; ++ struct delayed_work work; ++ spinlock_t lock; ++ struct btrfs_block_group *block_group; ++ struct list_head discard_list[3]; ++ u64 prev_discard; ++ u64 prev_discard_time; ++ atomic_t discardable_extents; ++ atomic64_t discardable_bytes; ++ u64 max_discard_size; ++ u64 delay_ms; ++ u32 iops_limit; ++ u32 kbps_limit; ++ u64 discard_extent_bytes; ++ u64 discard_bitmap_bytes; ++ atomic64_t discard_bytes_saved; ++}; ++ ++struct btrfs_work; ++ ++typedef void (*btrfs_func_t)(struct btrfs_work *); ++ ++struct btrfs_workqueue; ++ ++struct btrfs_work { ++ btrfs_func_t func; ++ btrfs_func_t ordered_func; ++ btrfs_func_t ordered_free; ++ struct work_struct normal_work; ++ struct list_head ordered_list; ++ struct btrfs_workqueue *wq; ++ long unsigned int flags; ++}; ++ ++struct btrfs_device; ++ ++struct btrfs_dev_replace { ++ u64 replace_state; ++ time64_t time_started; ++ time64_t time_stopped; ++ atomic64_t num_write_errors; ++ atomic64_t num_uncorrectable_read_errors; ++ u64 cursor_left; ++ u64 committed_cursor_left; ++ u64 cursor_left_last_write_of_item; ++ u64 cursor_right; ++ u64 cont_reading_from_srcdev_mode; ++ int is_valid; ++ int item_needs_writeback; ++ struct btrfs_device *srcdev; ++ struct btrfs_device *tgtdev; ++ struct mutex lock_finishing_cancel_unmount; ++ struct rw_semaphore rwsem; ++ struct btrfs_scrub_progress scrub_progress; ++ struct percpu_counter bio_counter; ++ wait_queue_head_t replace_wait; ++}; ++ ++enum btrfs_exclusive_operation { ++ BTRFS_EXCLOP_NONE = 0, ++ BTRFS_EXCLOP_BALANCE_PAUSED = 1, ++ BTRFS_EXCLOP_BALANCE = 2, ++ BTRFS_EXCLOP_DEV_ADD = 3, ++ BTRFS_EXCLOP_DEV_REMOVE = 4, ++ BTRFS_EXCLOP_DEV_REPLACE = 5, ++ BTRFS_EXCLOP_RESIZE = 6, ++ BTRFS_EXCLOP_SWAP_ACTIVATE = 7, ++}; ++ ++struct btrfs_commit_stats { ++ u64 commit_count; ++ u64 max_commit_dur; ++ u64 last_commit_dur; ++ u64 total_commit_dur; ++}; ++ ++struct btrfs_root; ++ ++struct btrfs_transaction; ++ ++struct btrfs_stripe_hash_table; ++ ++struct btrfs_fs_devices; ++ ++struct reloc_control; ++ ++struct btrfs_balance_control; ++ ++struct btrfs_subpage_info; ++ ++struct ulist; ++ ++struct btrfs_delayed_root; ++ ++struct btrfs_fs_info { ++ u8 chunk_tree_uuid[16]; ++ long unsigned int flags; ++ struct btrfs_root *tree_root; ++ struct btrfs_root *chunk_root; ++ struct btrfs_root *dev_root; ++ struct btrfs_root *fs_root; ++ struct btrfs_root *quota_root; ++ struct btrfs_root *uuid_root; ++ struct btrfs_root *data_reloc_root; ++ struct btrfs_root *block_group_root; ++ struct btrfs_root *log_root_tree; ++ rwlock_t global_root_lock; ++ struct rb_root global_root_tree; ++ spinlock_t fs_roots_radix_lock; ++ struct xarray fs_roots_radix; ++ rwlock_t block_group_cache_lock; ++ struct rb_root_cached block_group_cache_tree; ++ atomic64_t free_chunk_space; ++ struct extent_io_tree excluded_extents; ++ struct extent_map_tree mapping_tree; ++ struct btrfs_block_rsv global_block_rsv; ++ struct btrfs_block_rsv trans_block_rsv; ++ struct btrfs_block_rsv chunk_block_rsv; ++ struct btrfs_block_rsv delayed_block_rsv; ++ struct btrfs_block_rsv delayed_refs_rsv; ++ struct btrfs_block_rsv empty_block_rsv; ++ u64 generation; ++ u64 last_trans_committed; ++ u64 last_reloc_trans; ++ u64 last_trans_log_full_commit; ++ long unsigned int mount_opt; ++ long unsigned int compress_type: 4; ++ unsigned int compress_level; ++ u32 commit_interval; ++ u64 max_inline; ++ struct btrfs_transaction *running_transaction; ++ wait_queue_head_t transaction_throttle; ++ wait_queue_head_t transaction_wait; ++ wait_queue_head_t transaction_blocked_wait; ++ wait_queue_head_t async_submit_wait; ++ spinlock_t super_lock; ++ struct btrfs_super_block *super_copy; ++ struct btrfs_super_block *super_for_commit; ++ struct super_block *sb; ++ struct inode *btree_inode; ++ struct mutex tree_log_mutex; ++ struct mutex transaction_kthread_mutex; ++ struct mutex cleaner_mutex; ++ struct mutex chunk_mutex; ++ struct mutex ro_block_group_mutex; ++ struct btrfs_stripe_hash_table *stripe_hash_table; ++ struct mutex ordered_operations_mutex; ++ struct rw_semaphore commit_root_sem; ++ struct rw_semaphore cleanup_work_sem; ++ struct rw_semaphore subvol_sem; ++ spinlock_t trans_lock; ++ struct mutex reloc_mutex; ++ struct list_head trans_list; ++ struct list_head dead_roots; ++ struct list_head caching_block_groups; ++ spinlock_t delayed_iput_lock; ++ struct list_head delayed_iputs; ++ atomic_t nr_delayed_iputs; ++ wait_queue_head_t delayed_iputs_wait; ++ atomic64_t tree_mod_seq; ++ rwlock_t tree_mod_log_lock; ++ struct rb_root tree_mod_log; ++ struct list_head tree_mod_seq_list; ++ atomic_t async_delalloc_pages; ++ spinlock_t ordered_root_lock; ++ struct list_head ordered_roots; ++ struct mutex delalloc_root_mutex; ++ spinlock_t delalloc_root_lock; ++ struct list_head delalloc_roots; ++ struct btrfs_workqueue *workers; ++ struct btrfs_workqueue *hipri_workers; ++ struct btrfs_workqueue *delalloc_workers; ++ struct btrfs_workqueue *flush_workers; ++ struct workqueue_struct *endio_workers; ++ struct workqueue_struct *endio_meta_workers; ++ struct workqueue_struct *rmw_workers; ++ struct workqueue_struct *compressed_write_workers; ++ struct btrfs_workqueue *endio_write_workers; ++ struct btrfs_workqueue *endio_freespace_worker; ++ struct btrfs_workqueue *caching_workers; ++ struct btrfs_workqueue *fixup_workers; ++ struct btrfs_workqueue *delayed_workers; ++ struct task_struct *transaction_kthread; ++ struct task_struct *cleaner_kthread; ++ u32 thread_pool_size; ++ struct kobject *space_info_kobj; ++ struct kobject *qgroups_kobj; ++ struct kobject *discard_kobj; ++ struct percpu_counter dirty_metadata_bytes; ++ struct percpu_counter delalloc_bytes; ++ struct percpu_counter ordered_bytes; ++ s32 dirty_metadata_batch; ++ s32 delalloc_batch; ++ struct list_head dirty_cowonly_roots; ++ struct btrfs_fs_devices *fs_devices; ++ struct list_head space_info; ++ struct btrfs_space_info *data_sinfo; ++ struct reloc_control *reloc_ctl; ++ struct btrfs_free_cluster data_alloc_cluster; ++ struct btrfs_free_cluster meta_alloc_cluster; ++ spinlock_t defrag_inodes_lock; ++ struct rb_root defrag_inodes; ++ atomic_t defrag_running; ++ seqlock_t profiles_lock; ++ u64 avail_data_alloc_bits; ++ u64 avail_metadata_alloc_bits; ++ u64 avail_system_alloc_bits; ++ spinlock_t balance_lock; ++ struct mutex balance_mutex; ++ atomic_t balance_pause_req; ++ atomic_t balance_cancel_req; ++ struct btrfs_balance_control *balance_ctl; ++ wait_queue_head_t balance_wait_q; ++ atomic_t reloc_cancel_req; ++ u32 data_chunk_allocations; ++ u32 metadata_ratio; ++ void *bdev_holder; ++ struct mutex scrub_lock; ++ atomic_t scrubs_running; ++ atomic_t scrub_pause_req; ++ atomic_t scrubs_paused; ++ atomic_t scrub_cancel_req; ++ wait_queue_head_t scrub_pause_wait; ++ refcount_t scrub_workers_refcnt; ++ struct workqueue_struct *scrub_workers; ++ struct workqueue_struct *scrub_wr_completion_workers; ++ struct btrfs_subpage_info *subpage_info; ++ struct btrfs_discard_ctl discard_ctl; ++ u64 qgroup_flags; ++ struct rb_root qgroup_tree; ++ spinlock_t qgroup_lock; ++ struct ulist *qgroup_ulist; ++ struct mutex qgroup_ioctl_lock; ++ struct list_head dirty_qgroups; ++ u64 qgroup_seq; ++ struct mutex qgroup_rescan_lock; ++ struct btrfs_key qgroup_rescan_progress; ++ struct btrfs_workqueue *qgroup_rescan_workers; ++ struct completion qgroup_rescan_completion; ++ struct btrfs_work qgroup_rescan_work; ++ bool qgroup_rescan_running; ++ u8 qgroup_drop_subtree_thres; ++ long unsigned int fs_state; ++ struct btrfs_delayed_root *delayed_root; ++ spinlock_t buffer_lock; ++ struct xarray buffer_radix; ++ int backup_root_index; ++ struct btrfs_dev_replace dev_replace; ++ struct semaphore uuid_tree_rescan_sem; ++ struct work_struct async_reclaim_work; ++ struct work_struct async_data_reclaim_work; ++ struct work_struct preempt_reclaim_work; ++ struct work_struct reclaim_bgs_work; ++ struct list_head reclaim_bgs; ++ int bg_reclaim_threshold; ++ spinlock_t unused_bgs_lock; ++ struct list_head unused_bgs; ++ struct mutex unused_bg_unpin_mutex; ++ struct mutex reclaim_bgs_lock; ++ u32 nodesize; ++ u32 sectorsize; ++ u32 sectorsize_bits; ++ u32 csum_size; ++ u32 csums_per_leaf; ++ u32 stripesize; ++ u64 max_extent_size; ++ spinlock_t swapfile_pins_lock; ++ struct rb_root swapfile_pins; ++ struct crypto_shash *csum_shash; ++ enum btrfs_exclusive_operation exclusive_operation; ++ u64 zone_size; ++ struct queue_limits limits; ++ u64 max_zone_append_size; ++ struct mutex zoned_meta_io_lock; ++ spinlock_t treelog_bg_lock; ++ u64 treelog_bg; ++ spinlock_t relocation_bg_lock; ++ u64 data_reloc_bg; ++ struct mutex zoned_data_reloc_io_lock; ++ u64 nr_global_roots; ++ spinlock_t zone_active_bgs_lock; ++ struct list_head zone_active_bgs; ++ struct btrfs_commit_stats commit_stats; ++ u64 last_root_drop_gen; ++ struct lockdep_map btrfs_trans_num_writers_map; ++ struct lockdep_map btrfs_trans_num_extwriters_map; ++ struct lockdep_map btrfs_state_change_map[4]; ++ struct lockdep_map btrfs_trans_pending_ordered_map; ++ struct lockdep_map btrfs_ordered_extent_map; ++}; ++ ++struct btrfs_ordered_inode_tree { ++ spinlock_t lock; ++ struct rb_root tree; ++ struct rb_node *last; ++}; ++ ++struct btrfs_delayed_node; ++ ++struct btrfs_inode { ++ struct btrfs_root *root; ++ struct btrfs_key location; ++ spinlock_t lock; ++ struct extent_map_tree extent_tree; ++ struct extent_io_tree io_tree; ++ struct extent_io_tree file_extent_tree; ++ struct mutex log_mutex; ++ struct btrfs_ordered_inode_tree ordered_tree; ++ struct list_head delalloc_inodes; ++ struct rb_node rb_node; ++ long unsigned int runtime_flags; ++ atomic_t sync_writers; ++ u64 generation; ++ u64 last_trans; ++ u64 logged_trans; ++ int last_sub_trans; ++ int last_log_commit; ++ union { ++ u64 delalloc_bytes; ++ u64 first_dir_index_to_log; ++ }; ++ union { ++ u64 new_delalloc_bytes; ++ u64 last_dir_index_offset; ++ }; ++ u64 defrag_bytes; ++ u64 disk_i_size; ++ u64 index_cnt; ++ u64 dir_index; ++ u64 last_unlink_trans; ++ u64 last_reflink_trans; ++ u64 csum_bytes; ++ u32 flags; ++ u32 ro_flags; ++ unsigned int outstanding_extents; ++ struct btrfs_block_rsv block_rsv; ++ unsigned int prop_compress; ++ unsigned int defrag_compress; ++ struct btrfs_delayed_node *delayed_node; ++ struct timespec64 i_otime; ++ struct list_head delayed_iput; ++ struct rw_semaphore i_mmap_lock; ++ struct inode vfs_inode; ++}; ++ ++struct extent_state { ++ u64 start; ++ u64 end; ++ struct rb_node rb_node; ++ wait_queue_head_t wq; ++ refcount_t refs; ++ u32 state; ++}; ++ ++enum btrfs_compression_type { ++ BTRFS_COMPRESS_NONE = 0, ++ BTRFS_COMPRESS_ZLIB = 1, ++ BTRFS_COMPRESS_LZO = 2, ++ BTRFS_COMPRESS_ZSTD = 3, ++ BTRFS_NR_COMPRESS_TYPES = 4, ++}; ++ ++struct ulist { ++ long unsigned int nnodes; ++ struct list_head nodes; ++ struct rb_root root; ++}; ++ ++struct extent_buffer { ++ u64 start; ++ long unsigned int len; ++ long unsigned int bflags; ++ struct btrfs_fs_info *fs_info; ++ spinlock_t refs_lock; ++ atomic_t refs; ++ atomic_t io_pages; ++ int read_mirror; ++ struct callback_head callback_head; ++ pid_t lock_owner; ++ s8 log_index; ++ struct rw_semaphore lock; ++ struct page *pages[16]; ++ struct list_head release_list; ++}; ++ ++enum { ++ EXTENT_FLAG_PINNED = 0, ++ EXTENT_FLAG_COMPRESSED = 1, ++ EXTENT_FLAG_PREALLOC = 2, ++ EXTENT_FLAG_LOGGING = 3, ++ EXTENT_FLAG_FILLING = 4, ++ EXTENT_FLAG_FS_MAPPING = 5, ++ EXTENT_FLAG_MERGED = 6, ++}; ++ ++struct map_lookup; ++ ++struct extent_map { ++ struct rb_node rb_node; ++ u64 start; ++ u64 len; ++ u64 mod_start; ++ u64 mod_len; ++ u64 orig_start; ++ u64 orig_block_len; ++ u64 ram_bytes; ++ u64 block_start; ++ u64 block_len; ++ u64 generation; ++ long unsigned int flags; ++ struct map_lookup *map_lookup; ++ refcount_t refs; ++ unsigned int compress_type; ++ struct list_head list; ++}; ++ ++struct btrfs_io_context; ++ ++struct btrfs_io_stripe { ++ struct btrfs_device *dev; ++ union { ++ u64 physical; ++ struct btrfs_io_context *bioc; ++ }; ++}; ++ ++struct map_lookup { ++ u64 type; ++ int io_align; ++ int io_width; ++ int num_stripes; ++ int sub_stripes; ++ int verified_stripes; ++ struct btrfs_io_stripe stripes[0]; ++}; ++ ++struct btrfs_workqueue { ++ struct workqueue_struct *normal_wq; ++ struct btrfs_fs_info *fs_info; ++ struct list_head ordered_list; ++ spinlock_t list_lock; ++ atomic_t pending; ++ int limit_active; ++ int current_active; ++ int thresh; ++ unsigned int count; ++ spinlock_t thres_lock; ++}; ++ ++struct btrfs_space_info { ++ spinlock_t lock; ++ u64 total_bytes; ++ u64 bytes_used; ++ u64 bytes_pinned; ++ u64 bytes_reserved; ++ u64 bytes_may_use; ++ u64 bytes_readonly; ++ u64 bytes_zone_unusable; ++ u64 max_extent_size; ++ u64 chunk_size; ++ int bg_reclaim_threshold; ++ int clamp; ++ unsigned int full: 1; ++ unsigned int chunk_alloc: 1; ++ unsigned int flush: 1; ++ unsigned int force_alloc; ++ u64 disk_used; ++ u64 disk_total; ++ u64 flags; ++ struct list_head list; ++ struct list_head ro_bgs; ++ struct list_head priority_tickets; ++ struct list_head tickets; ++ u64 reclaim_size; ++ u64 tickets_id; ++ struct rw_semaphore groups_sem; ++ struct list_head block_groups[9]; ++ struct kobject kobj; ++ struct kobject *block_group_kobjs[9]; ++}; ++ ++enum btrfs_lock_nesting { ++ BTRFS_NESTING_NORMAL = 0, ++ BTRFS_NESTING_COW = 1, ++ BTRFS_NESTING_LEFT = 2, ++ BTRFS_NESTING_RIGHT = 3, ++ BTRFS_NESTING_LEFT_COW = 4, ++ BTRFS_NESTING_RIGHT_COW = 5, ++ BTRFS_NESTING_SPLIT = 6, ++ BTRFS_NESTING_NEW_ROOT = 7, ++ BTRFS_NESTING_MAX = 8, ++}; ++ ++struct btrfs_drew_lock { ++ atomic_t readers; ++ atomic_t writers; ++ wait_queue_head_t pending_writers; ++ wait_queue_head_t pending_readers; ++}; ++ ++enum { ++ BTRFS_FS_STATE_ERROR = 0, ++ BTRFS_FS_STATE_REMOUNTING = 1, ++ BTRFS_FS_STATE_RO = 2, ++ BTRFS_FS_STATE_TRANS_ABORTED = 3, ++ BTRFS_FS_STATE_DEV_REPLACING = 4, ++ BTRFS_FS_STATE_DUMMY_FS_INFO = 5, ++ BTRFS_FS_STATE_NO_CSUMS = 6, ++ BTRFS_FS_STATE_LOG_CLEANUP_ERROR = 7, ++ BTRFS_FS_STATE_COUNT = 8, ++}; ++ ++enum { ++ BTRFS_FS_CLOSING_START = 0, ++ BTRFS_FS_CLOSING_DONE = 1, ++ BTRFS_FS_LOG_RECOVERING = 2, ++ BTRFS_FS_OPEN = 3, ++ BTRFS_FS_QUOTA_ENABLED = 4, ++ BTRFS_FS_UPDATE_UUID_TREE_GEN = 5, ++ BTRFS_FS_CREATING_FREE_SPACE_TREE = 6, ++ BTRFS_FS_BTREE_ERR = 7, ++ BTRFS_FS_LOG1_ERR = 8, ++ BTRFS_FS_LOG2_ERR = 9, ++ BTRFS_FS_QUOTA_OVERRIDE = 10, ++ BTRFS_FS_FROZEN = 11, ++ BTRFS_FS_BALANCE_RUNNING = 12, ++ BTRFS_FS_RELOC_RUNNING = 13, ++ BTRFS_FS_CLEANER_RUNNING = 14, ++ BTRFS_FS_CSUM_IMPL_FAST = 15, ++ BTRFS_FS_DISCARD_RUNNING = 16, ++ BTRFS_FS_CLEANUP_SPACE_CACHE_V1 = 17, ++ BTRFS_FS_FREE_SPACE_TREE_UNTRUSTED = 18, ++ BTRFS_FS_TREE_MOD_LOG_USERS = 19, ++ BTRFS_FS_COMMIT_TRANS = 20, ++ BTRFS_FS_UNFINISHED_DROPS = 21, ++ BTRFS_FS_NEED_ZONE_FINISH = 22, ++ BTRFS_FS_NEED_TRANS_COMMIT = 23, ++ BTRFS_FS_ACTIVE_ZONE_TRACKING = 24, ++ BTRFS_FS_FEATURE_CHANGED = 25, ++}; ++ ++enum { ++ BTRFS_MOUNT_NODATASUM = 1, ++ BTRFS_MOUNT_NODATACOW = 2, ++ BTRFS_MOUNT_NOBARRIER = 4, ++ BTRFS_MOUNT_SSD = 8, ++ BTRFS_MOUNT_DEGRADED = 16, ++ BTRFS_MOUNT_COMPRESS = 32, ++ BTRFS_MOUNT_NOTREELOG = 64, ++ BTRFS_MOUNT_FLUSHONCOMMIT = 128, ++ BTRFS_MOUNT_SSD_SPREAD = 256, ++ BTRFS_MOUNT_NOSSD = 512, ++ BTRFS_MOUNT_DISCARD_SYNC = 1024, ++ BTRFS_MOUNT_FORCE_COMPRESS = 2048, ++ BTRFS_MOUNT_SPACE_CACHE = 4096, ++ BTRFS_MOUNT_CLEAR_CACHE = 8192, ++ BTRFS_MOUNT_USER_SUBVOL_RM_ALLOWED = 16384, ++ BTRFS_MOUNT_ENOSPC_DEBUG = 32768, ++ BTRFS_MOUNT_AUTO_DEFRAG = 65536, ++ BTRFS_MOUNT_USEBACKUPROOT = 131072, ++ BTRFS_MOUNT_SKIP_BALANCE = 262144, ++ BTRFS_MOUNT_CHECK_INTEGRITY = 524288, ++ BTRFS_MOUNT_CHECK_INTEGRITY_DATA = 1048576, ++ BTRFS_MOUNT_PANIC_ON_FATAL_ERROR = 2097152, ++ BTRFS_MOUNT_RESCAN_UUID_TREE = 4194304, ++ BTRFS_MOUNT_FRAGMENT_DATA = 8388608, ++ BTRFS_MOUNT_FRAGMENT_METADATA = 16777216, ++ BTRFS_MOUNT_FREE_SPACE_TREE = 33554432, ++ BTRFS_MOUNT_NOLOGREPLAY = 67108864, ++ BTRFS_MOUNT_REF_VERIFY = 134217728, ++ BTRFS_MOUNT_DISCARD_ASYNC = 268435456, ++ BTRFS_MOUNT_IGNOREBADROOTS = 536870912, ++ BTRFS_MOUNT_IGNOREDATACSUMS = 1073741824, ++ BTRFS_MOUNT_NODISCARD = 2147483648, ++}; ++ ++struct rcu_string; ++ ++struct btrfs_zoned_device_info; ++ ++struct scrub_ctx; ++ ++struct btrfs_device { ++ struct list_head dev_list; ++ struct list_head dev_alloc_list; ++ struct list_head post_commit_list; ++ struct btrfs_fs_devices *fs_devices; ++ struct btrfs_fs_info *fs_info; ++ struct rcu_string *name; ++ u64 generation; ++ struct block_device *bdev; ++ struct btrfs_zoned_device_info *zone_info; ++ fmode_t mode; ++ dev_t devt; ++ long unsigned int dev_state; ++ blk_status_t last_flush_error; ++ u64 devid; ++ u64 total_bytes; ++ u64 disk_total_bytes; ++ u64 bytes_used; ++ u32 io_align; ++ u32 io_width; ++ u64 type; ++ u32 sector_size; ++ u8 uuid[16]; ++ u64 commit_total_bytes; ++ u64 commit_bytes_used; ++ struct bio flush_bio; ++ struct completion flush_wait; ++ struct scrub_ctx *scrub_ctx; ++ int dev_stats_valid; ++ atomic_t dev_stats_ccnt; ++ atomic_t dev_stat_values[5]; ++ struct extent_io_tree alloc_state; ++ struct completion kobj_unregister; ++ struct kobject devid_kobj; ++ u64 scrub_speed_max; ++}; ++ ++enum btrfs_discard_state { ++ BTRFS_DISCARD_EXTENTS = 0, ++ BTRFS_DISCARD_BITMAPS = 1, ++ BTRFS_DISCARD_RESET_CURSOR = 2, ++}; ++ ++struct btrfs_io_ctl { ++ void *cur; ++ void *orig; ++ struct page *page; ++ struct page **pages; ++ struct btrfs_fs_info *fs_info; ++ struct inode *inode; ++ long unsigned int size; ++ int index; ++ int num_pages; ++ int entries; ++ int bitmaps; ++}; ++ ++enum btrfs_block_group_size_class { ++ BTRFS_BG_SZ_NONE = 0, ++ BTRFS_BG_SZ_SMALL = 1, ++ BTRFS_BG_SZ_MEDIUM = 2, ++ BTRFS_BG_SZ_LARGE = 3, ++}; ++ ++struct btrfs_caching_control; ++ ++struct btrfs_free_space_ctl; ++ ++struct btrfs_block_group { ++ struct btrfs_fs_info *fs_info; ++ struct inode *inode; ++ spinlock_t lock; ++ u64 start; ++ u64 length; ++ u64 pinned; ++ u64 reserved; ++ u64 used; ++ u64 delalloc_bytes; ++ u64 bytes_super; ++ u64 flags; ++ u64 cache_generation; ++ u64 global_root_id; ++ u64 commit_used; ++ u32 bitmap_high_thresh; ++ u32 bitmap_low_thresh; ++ struct rw_semaphore data_rwsem; ++ long unsigned int full_stripe_len; ++ long unsigned int runtime_flags; ++ unsigned int ro; ++ int disk_cache_state; ++ int cached; ++ struct btrfs_caching_control *caching_ctl; ++ struct btrfs_space_info *space_info; ++ struct btrfs_free_space_ctl *free_space_ctl; ++ struct rb_node cache_node; ++ struct list_head list; ++ refcount_t refs; ++ struct list_head cluster_list; ++ struct list_head bg_list; ++ struct list_head ro_list; ++ atomic_t frozen; ++ struct list_head discard_list; ++ int discard_index; ++ u64 discard_eligible_time; ++ u64 discard_cursor; ++ enum btrfs_discard_state discard_state; ++ struct list_head dirty_list; ++ struct list_head io_list; ++ struct btrfs_io_ctl io_ctl; ++ atomic_t reservations; ++ atomic_t nocow_writers; ++ struct mutex free_space_lock; ++ int swap_extents; ++ u64 alloc_offset; ++ u64 zone_unusable; ++ u64 zone_capacity; ++ u64 meta_write_pointer; ++ struct map_lookup *physical_map; ++ struct list_head active_bg_list; ++ struct work_struct zone_finish_work; ++ struct extent_buffer *last_eb; ++ enum btrfs_block_group_size_class size_class; ++}; ++ ++struct btrfs_qgroup_swapped_blocks { ++ spinlock_t lock; ++ bool swapped; ++ struct rb_root blocks[8]; ++}; ++ ++struct btrfs_root { ++ struct rb_node rb_node; ++ struct extent_buffer *node; ++ struct extent_buffer *commit_root; ++ struct btrfs_root *log_root; ++ struct btrfs_root *reloc_root; ++ long unsigned int state; ++ struct btrfs_root_item root_item; ++ struct btrfs_key root_key; ++ struct btrfs_fs_info *fs_info; ++ struct extent_io_tree dirty_log_pages; ++ struct mutex objectid_mutex; ++ spinlock_t accounting_lock; ++ struct btrfs_block_rsv *block_rsv; ++ struct mutex log_mutex; ++ wait_queue_head_t log_writer_wait; ++ wait_queue_head_t log_commit_wait[2]; ++ struct list_head log_ctxs[2]; ++ atomic_t log_writers; ++ atomic_t log_commit[2]; ++ atomic_t log_batch; ++ int log_transid; ++ int log_transid_committed; ++ int last_log_commit; ++ pid_t log_start_pid; ++ u64 last_trans; ++ u32 type; ++ u64 free_objectid; ++ struct btrfs_key defrag_progress; ++ struct btrfs_key defrag_max; ++ struct list_head dirty_list; ++ struct list_head root_list; ++ spinlock_t log_extents_lock[2]; ++ struct list_head logged_list[2]; ++ spinlock_t inode_lock; ++ struct rb_root inode_tree; ++ struct xarray delayed_nodes_tree; ++ dev_t anon_dev; ++ spinlock_t root_item_lock; ++ refcount_t refs; ++ struct mutex delalloc_mutex; ++ spinlock_t delalloc_lock; ++ struct list_head delalloc_inodes; ++ struct list_head delalloc_root; ++ u64 nr_delalloc_inodes; ++ struct mutex ordered_extent_mutex; ++ spinlock_t ordered_extent_lock; ++ struct list_head ordered_extents; ++ struct list_head ordered_root; ++ u64 nr_ordered_extents; ++ struct list_head reloc_dirty_list; ++ int send_in_progress; ++ int dedupe_in_progress; ++ struct btrfs_drew_lock snapshot_lock; ++ atomic_t snapshot_force_cow; ++ spinlock_t qgroup_meta_rsv_lock; ++ u64 qgroup_meta_rsv_pertrans; ++ u64 qgroup_meta_rsv_prealloc; ++ wait_queue_head_t qgroup_flush_wait; ++ atomic_t nr_swapfiles; ++ struct btrfs_qgroup_swapped_blocks swapped_blocks; ++ struct extent_io_tree log_csum_range; ++}; ++ ++enum btrfs_trans_state { ++ TRANS_STATE_RUNNING = 0, ++ TRANS_STATE_COMMIT_START = 1, ++ TRANS_STATE_COMMIT_DOING = 2, ++ TRANS_STATE_UNBLOCKED = 3, ++ TRANS_STATE_SUPER_COMMITTED = 4, ++ TRANS_STATE_COMPLETED = 5, ++ TRANS_STATE_MAX = 6, ++}; ++ ++struct btrfs_delayed_ref_root { ++ struct rb_root_cached href_root; ++ struct rb_root dirty_extent_root; ++ spinlock_t lock; ++ atomic_t num_entries; ++ long unsigned int num_heads; ++ long unsigned int num_heads_ready; ++ u64 pending_csums; ++ long unsigned int flags; ++ u64 run_delayed_start; ++ u64 qgroup_to_skip; ++}; ++ ++struct btrfs_transaction { ++ u64 transid; ++ atomic_t num_extwriters; ++ atomic_t num_writers; ++ refcount_t use_count; ++ long unsigned int flags; ++ enum btrfs_trans_state state; ++ int aborted; ++ struct list_head list; ++ struct extent_io_tree dirty_pages; ++ time64_t start_time; ++ wait_queue_head_t writer_wait; ++ wait_queue_head_t commit_wait; ++ struct list_head pending_snapshots; ++ struct list_head dev_update_list; ++ struct list_head switch_commits; ++ struct list_head dirty_bgs; ++ struct list_head io_bgs; ++ struct list_head dropped_roots; ++ struct extent_io_tree pinned_extents; ++ struct mutex cache_write_mutex; ++ spinlock_t dirty_bgs_lock; ++ struct list_head deleted_bgs; ++ spinlock_t dropped_roots_lock; ++ struct btrfs_delayed_ref_root delayed_refs; ++ struct btrfs_fs_info *fs_info; ++ atomic_t pending_ordered; ++ wait_queue_head_t pending_wait; ++ spinlock_t releasing_ebs_lock; ++ struct list_head releasing_ebs; ++}; ++ ++enum btrfs_chunk_allocation_policy { ++ BTRFS_CHUNK_ALLOC_REGULAR = 0, ++ BTRFS_CHUNK_ALLOC_ZONED = 1, ++}; ++ ++enum btrfs_read_policy { ++ BTRFS_READ_POLICY_PID = 0, ++ BTRFS_NR_READ_POLICY = 1, ++}; ++ ++struct btrfs_fs_devices { ++ u8 fsid[16]; ++ u8 metadata_uuid[16]; ++ bool fsid_change; ++ struct list_head fs_list; ++ u64 num_devices; ++ u64 open_devices; ++ u64 rw_devices; ++ u64 missing_devices; ++ u64 total_rw_bytes; ++ u64 total_devices; ++ u64 latest_generation; ++ struct btrfs_device *latest_dev; ++ struct mutex device_list_mutex; ++ struct list_head devices; ++ struct list_head alloc_list; ++ struct list_head seed_list; ++ bool seeding; ++ int opened; ++ bool rotating; ++ bool discardable; ++ struct btrfs_fs_info *fs_info; ++ struct kobject fsid_kobj; ++ struct kobject *devices_kobj; ++ struct kobject *devinfo_kobj; ++ struct completion kobj_unregister; ++ enum btrfs_chunk_allocation_policy chunk_alloc_policy; ++ enum btrfs_read_policy read_policy; ++}; ++ ++struct btrfs_balance_control { ++ struct btrfs_balance_args data; ++ struct btrfs_balance_args meta; ++ struct btrfs_balance_args sys; ++ u64 flags; ++ struct btrfs_balance_progress stat; ++}; ++ ++struct btrfs_delayed_root { ++ spinlock_t lock; ++ struct list_head node_list; ++ struct list_head prepare_list; ++ atomic_t items; ++ atomic_t items_seq; ++ int nodes; ++ wait_queue_head_t wait; ++}; ++ ++struct btrfs_path { ++ struct extent_buffer *nodes[8]; ++ int slots[8]; ++ u8 locks[8]; ++ u8 reada; ++ u8 lowest_level; ++ unsigned int search_for_split: 1; ++ unsigned int keep_locks: 1; ++ unsigned int skip_locking: 1; ++ unsigned int search_commit_root: 1; ++ unsigned int need_commit_sem: 1; ++ unsigned int skip_release_on_error: 1; ++ unsigned int search_for_extension: 1; ++ unsigned int nowait: 1; ++}; ++ ++struct btrfs_delayed_node { ++ u64 inode_id; ++ u64 bytes_reserved; ++ struct btrfs_root *root; ++ struct list_head n_list; ++ struct list_head p_list; ++ struct rb_root_cached ins_root; ++ struct rb_root_cached del_root; ++ struct mutex mutex; ++ struct btrfs_inode_item inode_item; ++ refcount_t refs; ++ u64 index_cnt; ++ long unsigned int flags; ++ int count; ++ u32 curr_index_batch_size; ++ u32 index_item_leaves; ++}; ++ ++enum { ++ BTRFS_ORDERED_REGULAR = 0, ++ BTRFS_ORDERED_NOCOW = 1, ++ BTRFS_ORDERED_PREALLOC = 2, ++ BTRFS_ORDERED_COMPRESSED = 3, ++ BTRFS_ORDERED_DIRECT = 4, ++ BTRFS_ORDERED_IO_DONE = 5, ++ BTRFS_ORDERED_COMPLETE = 6, ++ BTRFS_ORDERED_IOERR = 7, ++ BTRFS_ORDERED_TRUNCATED = 8, ++ BTRFS_ORDERED_LOGGED = 9, ++ BTRFS_ORDERED_LOGGED_CSUM = 10, ++ BTRFS_ORDERED_PENDING = 11, ++ BTRFS_ORDERED_ENCODED = 12, ++}; ++ ++struct btrfs_ordered_extent { ++ u64 file_offset; ++ u64 num_bytes; ++ u64 ram_bytes; ++ u64 disk_bytenr; ++ u64 disk_num_bytes; ++ u64 offset; ++ u64 bytes_left; ++ u64 outstanding_isize; ++ u64 truncated_len; ++ long unsigned int flags; ++ int compress_type; ++ int qgroup_rsv; ++ refcount_t refs; ++ struct inode *inode; ++ struct list_head list; ++ struct list_head log_list; ++ wait_queue_head_t wait; ++ struct rb_node rb_node; ++ struct list_head root_extent_list; ++ struct btrfs_work work; ++ struct completion completion; ++ struct btrfs_work flush_work; ++ struct list_head work_list; ++ u64 physical; ++}; ++ ++struct btrfs_delayed_ref_node { ++ struct rb_node ref_node; ++ struct list_head add_list; ++ u64 bytenr; ++ u64 num_bytes; ++ u64 seq; ++ refcount_t refs; ++ int ref_mod; ++ unsigned int action: 8; ++ unsigned int type: 8; ++ unsigned int is_head: 1; ++ unsigned int in_tree: 1; ++}; ++ ++struct btrfs_delayed_extent_op { ++ struct btrfs_disk_key key; ++ u8 level; ++ bool update_key; ++ bool update_flags; ++ u64 flags_to_set; ++}; ++ ++struct btrfs_delayed_ref_head { ++ u64 bytenr; ++ u64 num_bytes; ++ refcount_t refs; ++ struct mutex mutex; ++ spinlock_t lock; ++ struct rb_root_cached ref_tree; ++ struct list_head ref_add_list; ++ struct rb_node href_node; ++ struct btrfs_delayed_extent_op *extent_op; ++ int total_ref_mod; ++ int ref_mod; ++ unsigned int must_insert_reserved: 1; ++ unsigned int is_data: 1; ++ unsigned int is_system: 1; ++ unsigned int processing: 1; ++}; ++ ++struct btrfs_delayed_tree_ref { ++ struct btrfs_delayed_ref_node node; ++ u64 root; ++ u64 parent; ++ int level; ++}; ++ ++struct btrfs_delayed_data_ref { ++ struct btrfs_delayed_ref_node node; ++ u64 root; ++ u64 parent; ++ u64 objectid; ++ u64 offset; ++}; ++ ++struct btrfs_pending_snapshot; ++ ++struct btrfs_trans_handle { ++ u64 transid; ++ u64 bytes_reserved; ++ u64 chunk_bytes_reserved; ++ long unsigned int delayed_ref_updates; ++ struct btrfs_transaction *transaction; ++ struct btrfs_block_rsv *block_rsv; ++ struct btrfs_block_rsv *orig_rsv; ++ struct btrfs_pending_snapshot *pending_snapshot; ++ refcount_t use_count; ++ unsigned int type; ++ short int aborted; ++ bool adding_csums; ++ bool allocating_chunk; ++ bool removing_chunk; ++ bool reloc_reserved; ++ bool in_fsync; ++ struct btrfs_fs_info *fs_info; ++ struct list_head new_bgs; ++}; ++ ++struct btrfs_pending_snapshot { ++ struct dentry *dentry; ++ struct inode *dir; ++ struct btrfs_root *root; ++ struct btrfs_root_item *root_item; ++ struct btrfs_root *snap; ++ struct btrfs_qgroup_inherit *inherit; ++ struct btrfs_path *path; ++ struct btrfs_block_rsv block_rsv; ++ int error; ++ dev_t anon_dev; ++ bool readonly; ++ struct list_head list; ++}; ++ ++struct rcu_string { ++ struct callback_head rcu; ++ char str[0]; ++}; ++ ++enum btrfs_trim_state { ++ BTRFS_TRIM_STATE_UNTRIMMED = 0, ++ BTRFS_TRIM_STATE_TRIMMED = 1, ++ BTRFS_TRIM_STATE_TRIMMING = 2, ++}; ++ ++struct btrfs_free_space { ++ struct rb_node offset_index; ++ struct rb_node bytes_index; ++ u64 offset; ++ u64 bytes; ++ u64 max_extent_size; ++ long unsigned int *bitmap; ++ struct list_head list; ++ enum btrfs_trim_state trim_state; ++ s32 bitmap_extents; ++}; ++ ++enum { ++ BTRFS_STAT_CURR = 0, ++ BTRFS_STAT_PREV = 1, ++ BTRFS_STAT_NR_ENTRIES = 2, ++}; ++ ++struct btrfs_free_space_op; ++ ++struct btrfs_free_space_ctl { ++ spinlock_t tree_lock; ++ struct rb_root free_space_offset; ++ struct rb_root_cached free_space_bytes; ++ u64 free_space; ++ int extents_thresh; ++ int free_extents; ++ int total_bitmaps; ++ int unit; ++ u64 start; ++ s32 discardable_extents[2]; ++ s64 discardable_bytes[2]; ++ const struct btrfs_free_space_op *op; ++ struct btrfs_block_group *block_group; ++ struct mutex cache_writeout_mutex; ++ struct list_head trimming_ranges; ++}; ++ ++struct btrfs_free_space_op { ++ bool (*use_bitmap)(struct btrfs_free_space_ctl *, struct btrfs_free_space *); ++}; ++ ++struct extent_inode_elem; ++ ++struct prelim_ref { ++ struct rb_node rbnode; ++ u64 root_id; ++ struct btrfs_key key_for_search; ++ int level; ++ int count; ++ struct extent_inode_elem *inode_list; ++ u64 parent; ++ u64 wanted_disk_byte; ++}; ++ ++enum btrfs_raid_types { ++ BTRFS_RAID_SINGLE = 0, ++ BTRFS_RAID_RAID0 = 1, ++ BTRFS_RAID_RAID1 = 2, ++ BTRFS_RAID_DUP = 3, ++ BTRFS_RAID_RAID10 = 4, ++ BTRFS_RAID_RAID5 = 5, ++ BTRFS_RAID_RAID6 = 6, ++ BTRFS_RAID_RAID1C3 = 7, ++ BTRFS_RAID_RAID1C4 = 8, ++ BTRFS_NR_RAID_TYPES = 9, ++}; ++ ++struct btrfs_zoned_device_info { ++ u64 zone_size; ++ u8 zone_size_shift; ++ u32 nr_zones; ++ unsigned int max_active_zones; ++ atomic_t active_zones_left; ++ long unsigned int *seq_zones; ++ long unsigned int *empty_zones; ++ long unsigned int *active_zones; ++ struct blk_zone *zone_cache; ++ struct blk_zone sb_zones[6]; ++}; ++ ++struct btrfs_io_context { ++ refcount_t refs; ++ struct btrfs_fs_info *fs_info; ++ u64 map_type; ++ struct bio *orig_bio; ++ atomic_t error; ++ u16 max_errors; ++ u16 num_stripes; ++ u16 mirror_num; ++ u16 replace_nr_stripes; ++ s16 replace_stripe_src; ++ u64 full_stripe_logical; ++ struct btrfs_io_stripe stripes[0]; ++}; ++ ++struct btrfs_device_info { ++ struct btrfs_device *dev; ++ u64 dev_offset; ++ u64 max_avail; ++ u64 total_avail; ++}; ++ ++struct btrfs_raid_attr { ++ u8 sub_stripes; ++ u8 dev_stripes; ++ u8 devs_max; ++ u8 devs_min; ++ u8 tolerated_failures; ++ u8 devs_increment; ++ u8 ncopies; ++ u8 nparity; ++ u8 mindev_error; ++ const char raid_name[8]; ++ u64 bg_flag; ++}; ++ ++enum btrfs_reserve_flush_enum { ++ BTRFS_RESERVE_NO_FLUSH = 0, ++ BTRFS_RESERVE_FLUSH_LIMIT = 1, ++ BTRFS_RESERVE_FLUSH_EVICT = 2, ++ BTRFS_RESERVE_FLUSH_DATA = 3, ++ BTRFS_RESERVE_FLUSH_FREE_SPACE_INODE = 4, ++ BTRFS_RESERVE_FLUSH_ALL = 5, ++ BTRFS_RESERVE_FLUSH_ALL_STEAL = 6, ++ BTRFS_RESERVE_FLUSH_EMERGENCY = 7, ++}; ++ ++enum btrfs_flush_state { ++ FLUSH_DELAYED_ITEMS_NR = 1, ++ FLUSH_DELAYED_ITEMS = 2, ++ FLUSH_DELAYED_REFS_NR = 3, ++ FLUSH_DELAYED_REFS = 4, ++ FLUSH_DELALLOC = 5, ++ FLUSH_DELALLOC_WAIT = 6, ++ FLUSH_DELALLOC_FULL = 7, ++ ALLOC_CHUNK = 8, ++ ALLOC_CHUNK_FORCE = 9, ++ RUN_DELAYED_IPUTS = 10, ++ COMMIT_TRANS = 11, ++}; ++ ++struct btrfs_caching_control { ++ struct list_head list; ++ struct mutex mutex; ++ wait_queue_head_t wait; ++ struct btrfs_work work; ++ struct btrfs_block_group *block_group; ++ atomic_t progress; ++ refcount_t count; ++}; ++ ++struct btrfs_qgroup_extent_record { ++ struct rb_node node; ++ u64 bytenr; ++ u64 num_bytes; ++ u32 data_rsv; ++ u64 data_rsv_refroot; ++ struct ulist *old_roots; ++}; ++ ++enum btrfs_qgroup_rsv_type { ++ BTRFS_QGROUP_RSV_DATA = 0, ++ BTRFS_QGROUP_RSV_META_PERTRANS = 1, ++ BTRFS_QGROUP_RSV_META_PREALLOC = 2, ++ BTRFS_QGROUP_RSV_LAST = 3, ++}; ++ ++struct btrfs_qgroup_rsv { ++ u64 values[3]; ++}; ++ ++struct btrfs_qgroup { ++ u64 qgroupid; ++ u64 rfer; ++ u64 rfer_cmpr; ++ u64 excl; ++ u64 excl_cmpr; ++ u64 lim_flags; ++ u64 max_rfer; ++ u64 max_excl; ++ u64 rsv_rfer; ++ u64 rsv_excl; ++ struct btrfs_qgroup_rsv rsv; ++ struct list_head groups; ++ struct list_head members; ++ struct list_head dirty; ++ struct rb_node node; ++ u64 old_refcnt; ++ u64 new_refcnt; ++ struct kobject kobj; ++}; ++ ++enum { ++ __QGROUP_RESERVE_BIT = 0, ++ QGROUP_RESERVE = 1, ++ __QGROUP_RESERVE_SEQ = 0, ++ __QGROUP_RELEASE_BIT = 1, ++ QGROUP_RELEASE = 2, ++ __QGROUP_RELEASE_SEQ = 1, ++ __QGROUP_FREE_BIT = 2, ++ QGROUP_FREE = 4, ++ __QGROUP_FREE_SEQ = 2, ++}; ++ ++enum btrfs_rbio_ops { ++ BTRFS_RBIO_WRITE = 0, ++ BTRFS_RBIO_READ_REBUILD = 1, ++ BTRFS_RBIO_PARITY_SCRUB = 2, ++ BTRFS_RBIO_REBUILD_MISSING = 3, ++}; ++ ++struct sector_ptr; ++ ++struct btrfs_raid_bio { ++ struct btrfs_io_context *bioc; ++ struct list_head hash_list; ++ struct list_head stripe_cache; ++ struct work_struct work; ++ struct bio_list bio_list; ++ spinlock_t bio_list_lock; ++ struct list_head plug_list; ++ long unsigned int flags; ++ enum btrfs_rbio_ops operation; ++ u16 nr_pages; ++ u16 nr_sectors; ++ u8 nr_data; ++ u8 real_stripes; ++ u8 stripe_npages; ++ u8 stripe_nsectors; ++ u8 scrubp; ++ int bio_list_bytes; ++ refcount_t refs; ++ atomic_t stripes_pending; ++ wait_queue_head_t io_wait; ++ long unsigned int dbitmap; ++ long unsigned int finish_pbitmap; ++ struct page **stripe_pages; ++ struct sector_ptr *bio_sectors; ++ struct sector_ptr *stripe_sectors; ++ void **finish_pointers; ++ long unsigned int *error_bitmap; ++ u8 *csum_buf; ++ long unsigned int *csum_bitmap; ++}; ++ ++struct raid56_bio_trace_info { ++ u64 devid; ++ u32 offset; ++ u8 stripe_nr; ++}; ++ ++enum btrfs_extent_allocation_policy { ++ BTRFS_EXTENT_ALLOC_CLUSTERED = 0, ++ BTRFS_EXTENT_ALLOC_ZONED = 1, ++}; ++ ++struct find_free_extent_ctl { ++ u64 ram_bytes; ++ u64 num_bytes; ++ u64 min_alloc_size; ++ u64 empty_size; ++ u64 flags; ++ int delalloc; ++ u64 search_start; ++ u64 empty_cluster; ++ struct btrfs_free_cluster *last_ptr; ++ bool use_cluster; ++ bool have_caching_bg; ++ bool orig_have_caching_bg; ++ bool for_treelog; ++ bool for_data_reloc; ++ int index; ++ int loop; ++ bool retry_clustered; ++ bool retry_unclustered; ++ int cached; ++ u64 max_extent_size; ++ u64 total_free_space; ++ u64 found_offset; ++ u64 hint_byte; ++ enum btrfs_extent_allocation_policy policy; ++ bool hinted; ++ enum btrfs_block_group_size_class size_class; ++}; ++ ++struct trace_event_raw_btrfs_transaction_commit { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 generation; ++ u64 root_objectid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs__inode { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 ino; ++ u64 blocks; ++ u64 disk_i_size; ++ u64 generation; ++ u64 last_trans; ++ u64 logged_trans; ++ u64 root_objectid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_get_extent { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 root_objectid; ++ u64 ino; ++ u64 start; ++ u64 len; ++ u64 orig_start; ++ u64 block_start; ++ u64 block_len; ++ long unsigned int flags; ++ int refs; ++ unsigned int compress_type; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_handle_em_exist { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 e_start; ++ u64 e_len; ++ u64 map_start; ++ u64 map_len; ++ u64 start; ++ u64 len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs__file_extent_item_regular { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 root_obj; ++ u64 ino; ++ loff_t isize; ++ u64 disk_isize; ++ u64 num_bytes; ++ u64 ram_bytes; ++ u64 disk_bytenr; ++ u64 disk_num_bytes; ++ u64 extent_offset; ++ u8 extent_type; ++ u8 compression; ++ u64 extent_start; ++ u64 extent_end; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs__file_extent_item_inline { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 root_obj; ++ u64 ino; ++ loff_t isize; ++ u64 disk_isize; ++ u8 extent_type; ++ u8 compression; ++ u64 extent_start; ++ u64 extent_end; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs__ordered_extent { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 ino; ++ u64 file_offset; ++ u64 start; ++ u64 len; ++ u64 disk_len; ++ u64 bytes_left; ++ long unsigned int flags; ++ int compress_type; ++ int refs; ++ u64 root_objectid; ++ u64 truncated_len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs__writepage { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 ino; ++ long unsigned int index; ++ long int nr_to_write; ++ long int pages_skipped; ++ loff_t range_start; ++ loff_t range_end; ++ char for_kupdate; ++ char for_reclaim; ++ char range_cyclic; ++ long unsigned int writeback_index; ++ u64 root_objectid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_writepage_end_io_hook { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 ino; ++ u64 start; ++ u64 end; ++ int uptodate; ++ u64 root_objectid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_sync_file { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 ino; ++ u64 parent; ++ int datasync; ++ u64 root_objectid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_sync_fs { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ int wait; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_add_block_group { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 offset; ++ u64 size; ++ u64 flags; ++ u64 bytes_used; ++ u64 bytes_super; ++ int create; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_delayed_tree_ref { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 bytenr; ++ u64 num_bytes; ++ int action; ++ u64 parent; ++ u64 ref_root; ++ int level; ++ int type; ++ u64 seq; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_delayed_data_ref { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 bytenr; ++ u64 num_bytes; ++ int action; ++ u64 parent; ++ u64 ref_root; ++ u64 owner; ++ u64 offset; ++ int type; ++ u64 seq; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_delayed_ref_head { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 bytenr; ++ u64 num_bytes; ++ int action; ++ int is_data; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs__chunk { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ int num_stripes; ++ u64 type; ++ int sub_stripes; ++ u64 offset; ++ u64 size; ++ u64 root_objectid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_cow_block { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 root_objectid; ++ u64 buf_start; ++ int refs; ++ u64 cow_start; ++ int buf_level; ++ int cow_level; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_space_reservation { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u32 __data_loc_type; ++ u64 val; ++ u64 bytes; ++ int reserve; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_trigger_flush { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 flags; ++ u64 bytes; ++ int flush; ++ u32 __data_loc_reason; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_flush_space { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 flags; ++ u64 num_bytes; ++ int state; ++ int ret; ++ bool for_preempt; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs__reserved_extent { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 start; ++ u64 len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_find_free_extent { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 root_objectid; ++ u64 num_bytes; ++ u64 empty_size; ++ u64 flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_find_free_extent_search_loop { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 root_objectid; ++ u64 num_bytes; ++ u64 empty_size; ++ u64 flags; ++ u64 loop; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_find_free_extent_have_block_group { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 root_objectid; ++ u64 num_bytes; ++ u64 empty_size; ++ u64 flags; ++ u64 loop; ++ bool hinted; ++ u64 bg_start; ++ u64 bg_flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs__reserve_extent { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 bg_objectid; ++ u64 flags; ++ int bg_size_class; ++ u64 start; ++ u64 len; ++ u64 loop; ++ bool hinted; ++ int size_class; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_find_cluster { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 bg_objectid; ++ u64 flags; ++ u64 start; ++ u64 bytes; ++ u64 empty_size; ++ u64 min_bytes; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_failed_cluster_setup { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 bg_objectid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_setup_cluster { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 bg_objectid; ++ u64 flags; ++ u64 start; ++ u64 max_size; ++ u64 size; ++ int bitmap; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_alloc_extent_state { ++ struct trace_entry ent; ++ const struct extent_state *state; ++ long unsigned int mask; ++ const void *ip; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_free_extent_state { ++ struct trace_entry ent; ++ const struct extent_state *state; ++ const void *ip; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs__work { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ const void *work; ++ const void *wq; ++ const void *func; ++ const void *ordered_func; ++ const void *ordered_free; ++ const void *normal_work; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs__work__done { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ const void *wtag; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_workqueue { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ const void *wq; ++ u32 __data_loc_name; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_workqueue_done { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ const void *wq; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs__qgroup_rsv_data { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 rootid; ++ u64 ino; ++ u64 start; ++ u64 len; ++ u64 reserved; ++ int op; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_qgroup_extent { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 bytenr; ++ u64 num_bytes; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_qgroup_num_dirty_extents { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 transid; ++ u64 num_dirty_extents; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_qgroup_account_extent { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 transid; ++ u64 bytenr; ++ u64 num_bytes; ++ u64 nr_old_roots; ++ u64 nr_new_roots; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_qgroup_update_counters { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 qgid; ++ u64 old_rfer; ++ u64 old_excl; ++ u64 cur_old_count; ++ u64 cur_new_count; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_qgroup_update_reserve { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 qgid; ++ u64 cur_reserved; ++ s64 diff; ++ int type; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_qgroup_meta_reserve { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 refroot; ++ s64 diff; ++ int type; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_qgroup_meta_convert { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 refroot; ++ s64 diff; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_qgroup_meta_free_all_pertrans { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 refroot; ++ s64 diff; ++ int type; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs__prelim_ref { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 root_id; ++ u64 objectid; ++ u8 type; ++ u64 offset; ++ int level; ++ int old_count; ++ u64 parent; ++ u64 bytenr; ++ int mod_count; ++ u64 tree_size; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_inode_mod_outstanding_extents { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 root_objectid; ++ u64 ino; ++ int mod; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs__block_group { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 bytenr; ++ u64 len; ++ u64 used; ++ u64 flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_set_extent_bit { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ unsigned int owner; ++ u64 ino; ++ u64 rootid; ++ u64 start; ++ u64 len; ++ unsigned int set_bits; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_clear_extent_bit { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ unsigned int owner; ++ u64 ino; ++ u64 rootid; ++ u64 start; ++ u64 len; ++ unsigned int clear_bits; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_convert_extent_bit { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ unsigned int owner; ++ u64 ino; ++ u64 rootid; ++ u64 start; ++ u64 len; ++ unsigned int set_bits; ++ unsigned int clear_bits; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_dump_space_info { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 flags; ++ u64 total_bytes; ++ u64 bytes_used; ++ u64 bytes_pinned; ++ u64 bytes_reserved; ++ u64 bytes_may_use; ++ u64 bytes_readonly; ++ u64 reclaim_size; ++ int clamp; ++ u64 global_reserved; ++ u64 trans_reserved; ++ u64 delayed_refs_reserved; ++ u64 delayed_reserved; ++ u64 free_chunk_space; ++ u64 delalloc_bytes; ++ u64 ordered_bytes; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_reserve_ticket { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 flags; ++ u64 bytes; ++ u64 start_ns; ++ int flush; ++ int error; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_sleep_tree_lock { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 block; ++ u64 generation; ++ u64 start_ns; ++ u64 end_ns; ++ u64 diff_ns; ++ u64 owner; ++ int is_log_tree; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_locking_events { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 block; ++ u64 generation; ++ u64 owner; ++ int is_log_tree; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs__space_info_update { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 type; ++ u64 old; ++ s64 diff; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_btrfs_raid56_bio { ++ struct trace_entry ent; ++ u8 fsid[16]; ++ u64 full_stripe; ++ u64 physical; ++ u64 devid; ++ u32 offset; ++ u32 len; ++ u8 opf; ++ u8 total_stripes; ++ u8 real_stripes; ++ u8 nr_data; ++ u8 stripe_nr; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_btrfs_transaction_commit {}; ++ ++struct trace_event_data_offsets_btrfs__inode {}; ++ ++struct trace_event_data_offsets_btrfs_get_extent {}; ++ ++struct trace_event_data_offsets_btrfs_handle_em_exist {}; ++ ++struct trace_event_data_offsets_btrfs__file_extent_item_regular {}; ++ ++struct trace_event_data_offsets_btrfs__file_extent_item_inline {}; ++ ++struct trace_event_data_offsets_btrfs__ordered_extent {}; ++ ++struct trace_event_data_offsets_btrfs__writepage {}; ++ ++struct trace_event_data_offsets_btrfs_writepage_end_io_hook {}; ++ ++struct trace_event_data_offsets_btrfs_sync_file {}; ++ ++struct trace_event_data_offsets_btrfs_sync_fs {}; ++ ++struct trace_event_data_offsets_btrfs_add_block_group {}; ++ ++struct trace_event_data_offsets_btrfs_delayed_tree_ref {}; ++ ++struct trace_event_data_offsets_btrfs_delayed_data_ref {}; ++ ++struct trace_event_data_offsets_btrfs_delayed_ref_head {}; ++ ++struct trace_event_data_offsets_btrfs__chunk {}; ++ ++struct trace_event_data_offsets_btrfs_cow_block {}; ++ ++struct trace_event_data_offsets_btrfs_space_reservation { ++ u32 type; ++}; ++ ++struct trace_event_data_offsets_btrfs_trigger_flush { ++ u32 reason; ++}; ++ ++struct trace_event_data_offsets_btrfs_flush_space {}; ++ ++struct trace_event_data_offsets_btrfs__reserved_extent {}; ++ ++struct trace_event_data_offsets_find_free_extent {}; ++ ++struct trace_event_data_offsets_find_free_extent_search_loop {}; ++ ++struct trace_event_data_offsets_find_free_extent_have_block_group {}; ++ ++struct trace_event_data_offsets_btrfs__reserve_extent {}; ++ ++struct trace_event_data_offsets_btrfs_find_cluster {}; ++ ++struct trace_event_data_offsets_btrfs_failed_cluster_setup {}; ++ ++struct trace_event_data_offsets_btrfs_setup_cluster {}; ++ ++struct trace_event_data_offsets_alloc_extent_state {}; ++ ++struct trace_event_data_offsets_free_extent_state {}; ++ ++struct trace_event_data_offsets_btrfs__work {}; ++ ++struct trace_event_data_offsets_btrfs__work__done {}; ++ ++struct trace_event_data_offsets_btrfs_workqueue { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_btrfs_workqueue_done {}; ++ ++struct trace_event_data_offsets_btrfs__qgroup_rsv_data {}; ++ ++struct trace_event_data_offsets_btrfs_qgroup_extent {}; ++ ++struct trace_event_data_offsets_qgroup_num_dirty_extents {}; ++ ++struct trace_event_data_offsets_btrfs_qgroup_account_extent {}; ++ ++struct trace_event_data_offsets_qgroup_update_counters {}; ++ ++struct trace_event_data_offsets_qgroup_update_reserve {}; ++ ++struct trace_event_data_offsets_qgroup_meta_reserve {}; ++ ++struct trace_event_data_offsets_qgroup_meta_convert {}; ++ ++struct trace_event_data_offsets_qgroup_meta_free_all_pertrans {}; ++ ++struct trace_event_data_offsets_btrfs__prelim_ref {}; ++ ++struct trace_event_data_offsets_btrfs_inode_mod_outstanding_extents {}; ++ ++struct trace_event_data_offsets_btrfs__block_group {}; ++ ++struct trace_event_data_offsets_btrfs_set_extent_bit {}; ++ ++struct trace_event_data_offsets_btrfs_clear_extent_bit {}; ++ ++struct trace_event_data_offsets_btrfs_convert_extent_bit {}; ++ ++struct trace_event_data_offsets_btrfs_dump_space_info {}; ++ ++struct trace_event_data_offsets_btrfs_reserve_ticket {}; ++ ++struct trace_event_data_offsets_btrfs_sleep_tree_lock {}; ++ ++struct trace_event_data_offsets_btrfs_locking_events {}; ++ ++struct trace_event_data_offsets_btrfs__space_info_update {}; ++ ++struct trace_event_data_offsets_btrfs_raid56_bio {}; ++ ++typedef void (*btf_trace_btrfs_transaction_commit)(void *, const struct btrfs_fs_info *); ++ ++typedef void (*btf_trace_btrfs_inode_new)(void *, const struct inode *); ++ ++typedef void (*btf_trace_btrfs_inode_request)(void *, const struct inode *); ++ ++typedef void (*btf_trace_btrfs_inode_evict)(void *, const struct inode *); ++ ++typedef void (*btf_trace_btrfs_get_extent)(void *, const struct btrfs_root *, const struct btrfs_inode *, const struct extent_map *); ++ ++typedef void (*btf_trace_btrfs_handle_em_exist)(void *, const struct btrfs_fs_info *, const struct extent_map *, const struct extent_map *, u64, u64); ++ ++typedef void (*btf_trace_btrfs_get_extent_show_fi_regular)(void *, const struct btrfs_inode *, const struct extent_buffer *, const struct btrfs_file_extent_item *, u64); ++ ++typedef void (*btf_trace_btrfs_truncate_show_fi_regular)(void *, const struct btrfs_inode *, const struct extent_buffer *, const struct btrfs_file_extent_item *, u64); ++ ++typedef void (*btf_trace_btrfs_get_extent_show_fi_inline)(void *, const struct btrfs_inode *, const struct extent_buffer *, const struct btrfs_file_extent_item *, int, u64); ++ ++typedef void (*btf_trace_btrfs_truncate_show_fi_inline)(void *, const struct btrfs_inode *, const struct extent_buffer *, const struct btrfs_file_extent_item *, int, u64); ++ ++typedef void (*btf_trace_btrfs_ordered_extent_add)(void *, const struct btrfs_inode *, const struct btrfs_ordered_extent *); ++ ++typedef void (*btf_trace_btrfs_ordered_extent_remove)(void *, const struct btrfs_inode *, const struct btrfs_ordered_extent *); ++ ++typedef void (*btf_trace_btrfs_ordered_extent_start)(void *, const struct btrfs_inode *, const struct btrfs_ordered_extent *); ++ ++typedef void (*btf_trace_btrfs_ordered_extent_put)(void *, const struct btrfs_inode *, const struct btrfs_ordered_extent *); ++ ++typedef void (*btf_trace_btrfs_ordered_extent_lookup)(void *, const struct btrfs_inode *, const struct btrfs_ordered_extent *); ++ ++typedef void (*btf_trace_btrfs_ordered_extent_lookup_range)(void *, const struct btrfs_inode *, const struct btrfs_ordered_extent *); ++ ++typedef void (*btf_trace_btrfs_ordered_extent_lookup_first_range)(void *, const struct btrfs_inode *, const struct btrfs_ordered_extent *); ++ ++typedef void (*btf_trace_btrfs_ordered_extent_lookup_for_logging)(void *, const struct btrfs_inode *, const struct btrfs_ordered_extent *); ++ ++typedef void (*btf_trace_btrfs_ordered_extent_lookup_first)(void *, const struct btrfs_inode *, const struct btrfs_ordered_extent *); ++ ++typedef void (*btf_trace_btrfs_ordered_extent_split)(void *, const struct btrfs_inode *, const struct btrfs_ordered_extent *); ++ ++typedef void (*btf_trace_btrfs_ordered_extent_dec_test_pending)(void *, const struct btrfs_inode *, const struct btrfs_ordered_extent *); ++ ++typedef void (*btf_trace_btrfs_ordered_extent_mark_finished)(void *, const struct btrfs_inode *, const struct btrfs_ordered_extent *); ++ ++typedef void (*btf_trace___extent_writepage)(void *, const struct page *, const struct inode *, const struct writeback_control *); ++ ++typedef void (*btf_trace_btrfs_writepage_end_io_hook)(void *, const struct btrfs_inode *, u64, u64, int); ++ ++typedef void (*btf_trace_btrfs_sync_file)(void *, const struct file *, int); ++ ++typedef void (*btf_trace_btrfs_sync_fs)(void *, const struct btrfs_fs_info *, int); ++ ++typedef void (*btf_trace_btrfs_add_block_group)(void *, const struct btrfs_fs_info *, const struct btrfs_block_group *, int); ++ ++typedef void (*btf_trace_add_delayed_tree_ref)(void *, const struct btrfs_fs_info *, const struct btrfs_delayed_ref_node *, const struct btrfs_delayed_tree_ref *, int); ++ ++typedef void (*btf_trace_run_delayed_tree_ref)(void *, const struct btrfs_fs_info *, const struct btrfs_delayed_ref_node *, const struct btrfs_delayed_tree_ref *, int); ++ ++typedef void (*btf_trace_add_delayed_data_ref)(void *, const struct btrfs_fs_info *, const struct btrfs_delayed_ref_node *, const struct btrfs_delayed_data_ref *, int); ++ ++typedef void (*btf_trace_run_delayed_data_ref)(void *, const struct btrfs_fs_info *, const struct btrfs_delayed_ref_node *, const struct btrfs_delayed_data_ref *, int); ++ ++typedef void (*btf_trace_add_delayed_ref_head)(void *, const struct btrfs_fs_info *, const struct btrfs_delayed_ref_head *, int); ++ ++typedef void (*btf_trace_run_delayed_ref_head)(void *, const struct btrfs_fs_info *, const struct btrfs_delayed_ref_head *, int); ++ ++typedef void (*btf_trace_btrfs_chunk_alloc)(void *, const struct btrfs_fs_info *, const struct map_lookup *, u64, u64); ++ ++typedef void (*btf_trace_btrfs_chunk_free)(void *, const struct btrfs_fs_info *, const struct map_lookup *, u64, u64); ++ ++typedef void (*btf_trace_btrfs_cow_block)(void *, const struct btrfs_root *, const struct extent_buffer *, const struct extent_buffer *); ++ ++typedef void (*btf_trace_btrfs_space_reservation)(void *, const struct btrfs_fs_info *, const char *, u64, u64, int); ++ ++typedef void (*btf_trace_btrfs_trigger_flush)(void *, const struct btrfs_fs_info *, u64, u64, int, const char *); ++ ++typedef void (*btf_trace_btrfs_flush_space)(void *, const struct btrfs_fs_info *, u64, u64, int, int, bool); ++ ++typedef void (*btf_trace_btrfs_reserved_extent_alloc)(void *, const struct btrfs_fs_info *, u64, u64); ++ ++typedef void (*btf_trace_btrfs_reserved_extent_free)(void *, const struct btrfs_fs_info *, u64, u64); ++ ++typedef void (*btf_trace_find_free_extent)(void *, const struct btrfs_root *, const struct find_free_extent_ctl *); ++ ++typedef void (*btf_trace_find_free_extent_search_loop)(void *, const struct btrfs_root *, const struct find_free_extent_ctl *); ++ ++typedef void (*btf_trace_find_free_extent_have_block_group)(void *, const struct btrfs_root *, const struct find_free_extent_ctl *, const struct btrfs_block_group *); ++ ++typedef void (*btf_trace_btrfs_reserve_extent)(void *, const struct btrfs_block_group *, const struct find_free_extent_ctl *); ++ ++typedef void (*btf_trace_btrfs_reserve_extent_cluster)(void *, const struct btrfs_block_group *, const struct find_free_extent_ctl *); ++ ++typedef void (*btf_trace_btrfs_find_cluster)(void *, const struct btrfs_block_group *, u64, u64, u64, u64); ++ ++typedef void (*btf_trace_btrfs_failed_cluster_setup)(void *, const struct btrfs_block_group *); ++ ++typedef void (*btf_trace_btrfs_setup_cluster)(void *, const struct btrfs_block_group *, const struct btrfs_free_cluster *, u64, int); ++ ++typedef void (*btf_trace_alloc_extent_state)(void *, const struct extent_state *, gfp_t, long unsigned int); ++ ++typedef void (*btf_trace_free_extent_state)(void *, const struct extent_state *, long unsigned int); ++ ++typedef void (*btf_trace_btrfs_work_queued)(void *, const struct btrfs_work *); ++ ++typedef void (*btf_trace_btrfs_work_sched)(void *, const struct btrfs_work *); ++ ++typedef void (*btf_trace_btrfs_all_work_done)(void *, const struct btrfs_fs_info *, const void *); ++ ++typedef void (*btf_trace_btrfs_ordered_sched)(void *, const struct btrfs_work *); ++ ++typedef void (*btf_trace_btrfs_workqueue_alloc)(void *, const struct btrfs_workqueue *, const char *); ++ ++typedef void (*btf_trace_btrfs_workqueue_destroy)(void *, const struct btrfs_workqueue *); ++ ++typedef void (*btf_trace_btrfs_qgroup_reserve_data)(void *, const struct inode *, u64, u64, u64, int); ++ ++typedef void (*btf_trace_btrfs_qgroup_release_data)(void *, const struct inode *, u64, u64, u64, int); ++ ++typedef void (*btf_trace_btrfs_qgroup_account_extents)(void *, const struct btrfs_fs_info *, const struct btrfs_qgroup_extent_record *); ++ ++typedef void (*btf_trace_btrfs_qgroup_trace_extent)(void *, const struct btrfs_fs_info *, const struct btrfs_qgroup_extent_record *); ++ ++typedef void (*btf_trace_qgroup_num_dirty_extents)(void *, const struct btrfs_fs_info *, u64, u64); ++ ++typedef void (*btf_trace_btrfs_qgroup_account_extent)(void *, const struct btrfs_fs_info *, u64, u64, u64, u64, u64); ++ ++typedef void (*btf_trace_qgroup_update_counters)(void *, const struct btrfs_fs_info *, const struct btrfs_qgroup *, u64, u64); ++ ++typedef void (*btf_trace_qgroup_update_reserve)(void *, struct btrfs_fs_info *, struct btrfs_qgroup *, s64, int); ++ ++typedef void (*btf_trace_qgroup_meta_reserve)(void *, struct btrfs_root *, s64, int); ++ ++typedef void (*btf_trace_qgroup_meta_convert)(void *, struct btrfs_root *, s64); ++ ++typedef void (*btf_trace_qgroup_meta_free_all_pertrans)(void *, struct btrfs_root *); ++ ++typedef void (*btf_trace_btrfs_prelim_ref_merge)(void *, const struct btrfs_fs_info *, const struct prelim_ref *, const struct prelim_ref *, u64); ++ ++typedef void (*btf_trace_btrfs_prelim_ref_insert)(void *, const struct btrfs_fs_info *, const struct prelim_ref *, const struct prelim_ref *, u64); ++ ++typedef void (*btf_trace_btrfs_inode_mod_outstanding_extents)(void *, const struct btrfs_root *, u64, int); ++ ++typedef void (*btf_trace_btrfs_remove_block_group)(void *, const struct btrfs_block_group *); ++ ++typedef void (*btf_trace_btrfs_add_unused_block_group)(void *, const struct btrfs_block_group *); ++ ++typedef void (*btf_trace_btrfs_add_reclaim_block_group)(void *, const struct btrfs_block_group *); ++ ++typedef void (*btf_trace_btrfs_reclaim_block_group)(void *, const struct btrfs_block_group *); ++ ++typedef void (*btf_trace_btrfs_skip_unused_block_group)(void *, const struct btrfs_block_group *); ++ ++typedef void (*btf_trace_btrfs_set_extent_bit)(void *, const struct extent_io_tree *, u64, u64, unsigned int); ++ ++typedef void (*btf_trace_btrfs_clear_extent_bit)(void *, const struct extent_io_tree *, u64, u64, unsigned int); ++ ++typedef void (*btf_trace_btrfs_convert_extent_bit)(void *, const struct extent_io_tree *, u64, u64, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_btrfs_done_preemptive_reclaim)(void *, struct btrfs_fs_info *, const struct btrfs_space_info *); ++ ++typedef void (*btf_trace_btrfs_fail_all_tickets)(void *, struct btrfs_fs_info *, const struct btrfs_space_info *); ++ ++typedef void (*btf_trace_btrfs_reserve_ticket)(void *, const struct btrfs_fs_info *, u64, u64, u64, int, int); ++ ++typedef void (*btf_trace_btrfs_tree_read_lock)(void *, const struct extent_buffer *, u64); ++ ++typedef void (*btf_trace_btrfs_tree_lock)(void *, const struct extent_buffer *, u64); ++ ++typedef void (*btf_trace_btrfs_tree_unlock)(void *, const struct extent_buffer *); ++ ++typedef void (*btf_trace_btrfs_tree_read_unlock)(void *, const struct extent_buffer *); ++ ++typedef void (*btf_trace_btrfs_tree_read_unlock_blocking)(void *, const struct extent_buffer *); ++ ++typedef void (*btf_trace_btrfs_set_lock_blocking_read)(void *, const struct extent_buffer *); ++ ++typedef void (*btf_trace_btrfs_set_lock_blocking_write)(void *, const struct extent_buffer *); ++ ++typedef void (*btf_trace_btrfs_try_tree_read_lock)(void *, const struct extent_buffer *); ++ ++typedef void (*btf_trace_btrfs_try_tree_write_lock)(void *, const struct extent_buffer *); ++ ++typedef void (*btf_trace_btrfs_tree_read_lock_atomic)(void *, const struct extent_buffer *); ++ ++typedef void (*btf_trace_update_bytes_may_use)(void *, const struct btrfs_fs_info *, const struct btrfs_space_info *, u64, s64); ++ ++typedef void (*btf_trace_update_bytes_pinned)(void *, const struct btrfs_fs_info *, const struct btrfs_space_info *, u64, s64); ++ ++typedef void (*btf_trace_raid56_read_partial)(void *, const struct btrfs_raid_bio *, const struct bio *, const struct raid56_bio_trace_info *); ++ ++typedef void (*btf_trace_raid56_write_stripe)(void *, const struct btrfs_raid_bio *, const struct bio *, const struct raid56_bio_trace_info *); ++ ++typedef void (*btf_trace_raid56_scrub_write_stripe)(void *, const struct btrfs_raid_bio *, const struct bio *, const struct raid56_bio_trace_info *); ++ ++typedef void (*btf_trace_raid56_scrub_read)(void *, const struct btrfs_raid_bio *, const struct bio *, const struct raid56_bio_trace_info *); ++ ++typedef void (*btf_trace_raid56_scrub_read_recover)(void *, const struct btrfs_raid_bio *, const struct bio *, const struct raid56_bio_trace_info *); ++ ++enum { ++ Opt_acl = 0, ++ Opt_noacl = 1, ++ Opt_clear_cache = 2, ++ Opt_commit_interval = 3, ++ Opt_compress = 4, ++ Opt_compress_force = 5, ++ Opt_compress_force_type = 6, ++ Opt_compress_type = 7, ++ Opt_degraded = 8, ++ Opt_device = 9, ++ Opt_fatal_errors = 10, ++ Opt_flushoncommit = 11, ++ Opt_noflushoncommit = 12, ++ Opt_max_inline = 13, ++ Opt_barrier = 14, ++ Opt_nobarrier = 15, ++ Opt_datacow = 16, ++ Opt_nodatacow = 17, ++ Opt_datasum = 18, ++ Opt_nodatasum = 19, ++ Opt_defrag = 20, ++ Opt_nodefrag = 21, ++ Opt_discard = 22, ++ Opt_nodiscard = 23, ++ Opt_discard_mode = 24, ++ Opt_norecovery = 25, ++ Opt_ratio = 26, ++ Opt_rescan_uuid_tree = 27, ++ Opt_skip_balance = 28, ++ Opt_space_cache = 29, ++ Opt_no_space_cache = 30, ++ Opt_space_cache_version = 31, ++ Opt_ssd = 32, ++ Opt_nossd = 33, ++ Opt_ssd_spread = 34, ++ Opt_nossd_spread = 35, ++ Opt_subvol = 36, ++ Opt_subvol_empty = 37, ++ Opt_subvolid = 38, ++ Opt_thread_pool = 39, ++ Opt_treelog = 40, ++ Opt_notreelog = 41, ++ Opt_user_subvol_rm_allowed = 42, ++ Opt_rescue = 43, ++ Opt_usebackuproot = 44, ++ Opt_nologreplay = 45, ++ Opt_ignorebadroots = 46, ++ Opt_ignoredatacsums = 47, ++ Opt_rescue_all = 48, ++ Opt_recovery = 49, ++ Opt_inode_cache = 50, ++ Opt_noinode_cache = 51, ++ Opt_check_integrity = 52, ++ Opt_check_integrity_including_extent_data = 53, ++ Opt_check_integrity_print_mask = 54, ++ Opt_enospc_debug = 55, ++ Opt_noenospc_debug = 56, ++ Opt_err___2 = 57, ++}; ++ ++struct init_sequence { ++ int (*init_func)(); ++ void (*exit_func)(); ++}; ++ ++struct btrfs_item_batch { ++ const struct btrfs_key *keys; ++ const u32 *data_sizes; ++ u32 total_data_size; ++ int nr; ++}; ++ ++struct extent_changeset { ++ u64 bytes_changed; ++ struct ulist range_changed; ++}; ++ ++enum { ++ BTRFS_ROOT_IN_TRANS_SETUP = 0, ++ BTRFS_ROOT_SHAREABLE = 1, ++ BTRFS_ROOT_TRACK_DIRTY = 2, ++ BTRFS_ROOT_IN_RADIX = 3, ++ BTRFS_ROOT_ORPHAN_ITEM_INSERTED = 4, ++ BTRFS_ROOT_DEFRAG_RUNNING = 5, ++ BTRFS_ROOT_FORCE_COW = 6, ++ BTRFS_ROOT_MULTI_LOG_TASKS = 7, ++ BTRFS_ROOT_DIRTY = 8, ++ BTRFS_ROOT_DELETING = 9, ++ BTRFS_ROOT_DEAD_RELOC_TREE = 10, ++ BTRFS_ROOT_DEAD_TREE = 11, ++ BTRFS_ROOT_HAS_LOG_TREE = 12, ++ BTRFS_ROOT_QGROUP_FLUSHING = 13, ++ BTRFS_ROOT_ORPHAN_CLEANUP = 14, ++ BTRFS_ROOT_UNFINISHED_DROP = 15, ++ BTRFS_ROOT_RESET_LOCKDEP_CLASS = 16, ++}; ++ ++enum btrfs_delayed_ref_flags { ++ BTRFS_DELAYED_REFS_FLUSHING = 0, ++}; ++ ++enum { ++ ____TRANS_FREEZABLE_BIT = 0, ++ __TRANS_FREEZABLE = 1, ++ ____TRANS_FREEZABLE_SEQ = 0, ++ ____TRANS_START_BIT = 1, ++ __TRANS_START = 2, ++ ____TRANS_START_SEQ = 1, ++ ____TRANS_ATTACH_BIT = 2, ++ __TRANS_ATTACH = 4, ++ ____TRANS_ATTACH_SEQ = 2, ++ ____TRANS_JOIN_BIT = 3, ++ __TRANS_JOIN = 8, ++ ____TRANS_JOIN_SEQ = 3, ++ ____TRANS_JOIN_NOLOCK_BIT = 4, ++ __TRANS_JOIN_NOLOCK = 16, ++ ____TRANS_JOIN_NOLOCK_SEQ = 4, ++ ____TRANS_DUMMY_BIT = 5, ++ __TRANS_DUMMY = 32, ++ ____TRANS_DUMMY_SEQ = 5, ++ ____TRANS_JOIN_NOSTART_BIT = 6, ++ __TRANS_JOIN_NOSTART = 64, ++ ____TRANS_JOIN_NOSTART_SEQ = 6, ++}; ++ ++enum btrfs_chunk_alloc_enum { ++ CHUNK_ALLOC_NO_FORCE = 0, ++ CHUNK_ALLOC_LIMITED = 1, ++ CHUNK_ALLOC_FORCE = 2, ++ CHUNK_ALLOC_FORCE_FOR_EXTENT = 3, ++}; ++ ++struct btrfs_ioctl_defrag_range_args { ++ __u64 start; ++ __u64 len; ++ __u64 flags; ++ __u32 extent_thresh; ++ __u32 compress_type; ++ __u32 unused[4]; ++}; ++ ++struct btrfs_subpage_info { ++ unsigned int bitmap_nr_bits; ++ unsigned int total_nr_bits; ++ unsigned int uptodate_offset; ++ unsigned int error_offset; ++ unsigned int dirty_offset; ++ unsigned int writeback_offset; ++ unsigned int ordered_offset; ++ unsigned int checked_offset; ++}; ++ ++enum { ++ BTRFS_INODE_FLUSH_ON_CLOSE = 0, ++ BTRFS_INODE_DUMMY = 1, ++ BTRFS_INODE_IN_DEFRAG = 2, ++ BTRFS_INODE_HAS_ASYNC_EXTENT = 3, ++ BTRFS_INODE_NEEDS_FULL_SYNC = 4, ++ BTRFS_INODE_COPY_EVERYTHING = 5, ++ BTRFS_INODE_IN_DELALLOC_LIST = 6, ++ BTRFS_INODE_HAS_PROPS = 7, ++ BTRFS_INODE_SNAPSHOT_FLUSH = 8, ++ BTRFS_INODE_NO_XATTRS = 9, ++ BTRFS_INODE_NO_DELALLOC_FLUSH = 10, ++ BTRFS_INODE_VERITY_IN_PROGRESS = 11, ++ BTRFS_INODE_FREE_SPACE_INODE = 12, ++}; ++ ++struct inode_defrag { ++ struct rb_node rb_node; ++ u64 ino; ++ u64 transid; ++ u64 root; ++ u32 extent_thresh; ++}; ++ ++struct defrag_target_range { ++ struct list_head list; ++ u64 start; ++ u64 len; ++}; ++ ++struct btrfs_key_ptr { ++ struct btrfs_disk_key key; ++ __le64 blockptr; ++ __le64 generation; ++} __attribute__((packed)); ++ ++struct btrfs_map_token { ++ struct extent_buffer *eb; ++ char *kaddr; ++ long unsigned int offset; ++}; ++ ++struct btrfs_ordered_sum { ++ u64 bytenr; ++ int len; ++ struct list_head list; ++ u8 sums[0]; ++}; ++ ++enum { ++ WORK_DONE_BIT = 0, ++ WORK_ORDER_DONE_BIT = 1, ++}; ++ ++struct btrfs_inode_extref { ++ __le64 parent_objectid; ++ __le64 index; ++ __le16 name_len; ++ __u8 name[0]; ++} __attribute__((packed)); ++ ++struct btrfs_dir_log_item { ++ __le64 end; ++}; ++ ++struct btrfs_tree_parent_check { ++ u64 owner_root; ++ u64 transid; ++ struct btrfs_key first_key; ++ bool has_first_key; ++ u8 level; ++}; ++ ++struct btrfs_drop_extents_args { ++ struct btrfs_path *path; ++ u64 start; ++ u64 end; ++ bool drop_cache; ++ bool replace_extent; ++ u32 extent_item_size; ++ u64 drop_end; ++ u64 bytes_found; ++ bool extent_inserted; ++}; ++ ++enum btrfs_delayed_item_type { ++ BTRFS_DELAYED_INSERTION_ITEM = 0, ++ BTRFS_DELAYED_DELETION_ITEM = 1, ++}; ++ ++struct btrfs_delayed_item { ++ struct rb_node rb_node; ++ u64 index; ++ struct list_head tree_list; ++ struct list_head readdir_list; ++ struct list_head log_list; ++ u64 bytes_reserved; ++ struct btrfs_delayed_node *delayed_node; ++ refcount_t refs; ++ enum btrfs_delayed_item_type type: 8; ++ bool logged; ++ u16 data_len; ++ char data[0]; ++}; ++ ++enum btrfs_ref_type { ++ BTRFS_REF_NOT_SET = 0, ++ BTRFS_REF_DATA = 1, ++ BTRFS_REF_METADATA = 2, ++ BTRFS_REF_LAST = 3, ++}; ++ ++struct btrfs_data_ref { ++ u64 owning_root; ++ u64 ino; ++ u64 offset; ++}; ++ ++struct btrfs_tree_ref { ++ int level; ++ u64 owning_root; ++}; ++ ++struct btrfs_ref { ++ enum btrfs_ref_type type; ++ int action; ++ bool skip_qgroup; ++ u64 bytenr; ++ u64 len; ++ u64 parent; ++ union { ++ struct btrfs_data_ref data_ref; ++ struct btrfs_tree_ref tree_ref; ++ }; ++}; ++ ++struct btrfs_log_ctx { ++ int log_ret; ++ int log_transid; ++ bool log_new_dentries; ++ bool logging_new_name; ++ bool logging_new_delayed_dentries; ++ bool logged_before; ++ struct inode *inode; ++ struct list_head list; ++ struct list_head ordered_extents; ++ struct list_head conflict_inodes; ++ int num_conflict_inodes; ++ bool logging_conflict_inodes; ++}; ++ ++struct btrfs_truncate_control { ++ struct btrfs_inode *inode; ++ u64 new_size; ++ u64 extents_found; ++ u64 last_size; ++ u64 sub_bytes; ++ u64 ino; ++ u32 min_type; ++ bool skip_ref_updates; ++ bool clear_extent_range; ++}; ++ ++enum { ++ LOG_INODE_ALL = 0, ++ LOG_INODE_EXISTS = 1, ++}; ++ ++enum { ++ LOG_WALK_PIN_ONLY = 0, ++ LOG_WALK_REPLAY_INODES = 1, ++ LOG_WALK_REPLAY_DIR_INDEX = 2, ++ LOG_WALK_REPLAY_ALL = 3, ++}; ++ ++struct walk_control { ++ int free; ++ int pin; ++ int stage; ++ bool ignore_cur_inode; ++ struct btrfs_root *replay_dest; ++ struct btrfs_trans_handle *trans; ++ int (*process_func)(struct btrfs_root *, struct extent_buffer *, struct walk_control *, u64, int); ++}; ++ ++struct btrfs_dir_list { ++ u64 ino; ++ struct list_head list; ++}; ++ ++struct btrfs_ino_list { ++ u64 ino; ++ u64 parent; ++ struct list_head list; ++}; ++ ++enum { ++ BIOSET_NEED_BVECS = 1, ++ BIOSET_NEED_RESCUER = 2, ++ BIOSET_PERCPU_CACHE = 4, ++}; ++ ++struct btrfs_bio; ++ ++typedef void (*btrfs_bio_end_io_t)(struct btrfs_bio *); ++ ++struct btrfs_bio { ++ struct btrfs_inode *inode; ++ u64 file_offset; ++ union { ++ struct { ++ u8 *csum; ++ u8 csum_inline[64]; ++ struct bvec_iter saved_iter; ++ }; ++ struct btrfs_tree_parent_check parent_check; ++ }; ++ btrfs_bio_end_io_t end_io; ++ void *private; ++ unsigned int mirror_num; ++ atomic_t pending_ios; ++ struct work_struct end_io_work; ++ struct btrfs_fs_info *fs_info; ++ struct bio bio; ++}; ++ ++struct compressed_bio { ++ unsigned int nr_pages; ++ struct page **compressed_pages; ++ u64 start; ++ unsigned int len; ++ unsigned int compressed_len; ++ u8 compress_type; ++ bool writeback; ++ union { ++ struct btrfs_bio *orig_bbio; ++ struct work_struct write_end_work; ++ }; ++ struct btrfs_bio bbio; ++}; ++ ++struct workspace_manager { ++ struct list_head idle_ws; ++ spinlock_t ws_lock; ++ int free_ws; ++ atomic_t total_ws; ++ wait_queue_head_t ws_wait; ++}; ++ ++struct btrfs_compress_op { ++ struct workspace_manager *workspace_manager; ++ unsigned int max_level; ++ unsigned int default_level; ++}; ++ ++struct bucket_item { ++ u32 count; ++}; ++ ++struct heuristic_ws { ++ u8 *sample; ++ u32 sample_size; ++ struct bucket_item *bucket; ++ struct bucket_item *bucket_b; ++ struct list_head list; ++}; ++ ++struct hash_alg_common { ++ unsigned int digestsize; ++ unsigned int statesize; ++ struct crypto_alg base; ++}; ++ ++struct shash_alg { ++ int (*init)(struct shash_desc *); ++ int (*update)(struct shash_desc *, const u8 *, unsigned int); ++ int (*final)(struct shash_desc *, u8 *); ++ int (*finup)(struct shash_desc *, const u8 *, unsigned int, u8 *); ++ int (*digest)(struct shash_desc *, const u8 *, unsigned int, u8 *); ++ int (*export)(struct shash_desc *, void *); ++ int (*import)(struct shash_desc *, const void *); ++ int (*setkey)(struct crypto_shash *, const u8 *, unsigned int); ++ int (*init_tfm)(struct crypto_shash *); ++ void (*exit_tfm)(struct crypto_shash *); ++ int (*clone_tfm)(struct crypto_shash *, struct crypto_shash *); ++ unsigned int descsize; ++ union { ++ struct { ++ unsigned int digestsize; ++ unsigned int statesize; ++ struct crypto_alg base; ++ }; ++ struct hash_alg_common halg; ++ }; ++}; ++ ++struct btrfs_data_container { ++ __u32 bytes_left; ++ __u32 bytes_missing; ++ __u32 elem_cnt; ++ __u32 elem_missed; ++ __u64 val[0]; ++}; ++ ++struct btrfs_extent_item { ++ __le64 refs; ++ __le64 generation; ++ __le64 flags; ++}; ++ ++struct btrfs_dev_extent { ++ __le64 chunk_tree; ++ __le64 chunk_objectid; ++ __le64 chunk_offset; ++ __le64 length; ++ __u8 chunk_tree_uuid[16]; ++}; ++ ++enum { ++ READA_NONE = 0, ++ READA_BACK = 1, ++ READA_FORWARD = 2, ++ READA_FORWARD_ALWAYS = 3, ++}; ++ ++struct scrub_sector_verification; ++ ++struct scrub_stripe { ++ struct scrub_ctx *sctx; ++ struct btrfs_block_group *bg; ++ struct page *pages[16]; ++ struct scrub_sector_verification *sectors; ++ struct btrfs_device *dev; ++ u64 logical; ++ u64 physical; ++ u16 mirror_num; ++ u16 nr_sectors; ++ u16 nr_data_extents; ++ u16 nr_meta_extents; ++ atomic_t pending_io; ++ wait_queue_head_t io_wait; ++ wait_queue_head_t repair_wait; ++ long unsigned int state; ++ long unsigned int extent_sector_bitmap; ++ long unsigned int init_error_bitmap; ++ unsigned int init_nr_io_errors; ++ unsigned int init_nr_csum_errors; ++ unsigned int init_nr_meta_errors; ++ long unsigned int error_bitmap; ++ long unsigned int io_error_bitmap; ++ long unsigned int csum_error_bitmap; ++ long unsigned int meta_error_bitmap; ++ long unsigned int write_error_bitmap; ++ spinlock_t write_error_lock; ++ u8 *csums; ++ struct work_struct work; ++}; ++ ++struct scrub_ctx { ++ struct scrub_stripe stripes[8]; ++ struct scrub_stripe *raid56_data_stripes; ++ struct btrfs_fs_info *fs_info; ++ int first_free; ++ int cur_stripe; ++ struct list_head csum_list; ++ atomic_t cancel_req; ++ int readonly; ++ int sectors_per_bio; ++ ktime_t throttle_deadline; ++ u64 throttle_sent; ++ int is_dev_replace; ++ u64 write_pointer; ++ struct mutex wr_lock; ++ struct btrfs_device *wr_tgtdev; ++ struct btrfs_scrub_progress stat; ++ spinlock_t stat_lock; ++ refcount_t refs; ++}; ++ ++struct btrfs_dev_lookup_args { ++ u64 devid; ++ u8 *uuid; ++ u8 *fsid; ++ bool missing; ++}; ++ ++enum btrfs_map_op { ++ BTRFS_MAP_READ = 0, ++ BTRFS_MAP_WRITE = 1, ++ BTRFS_MAP_DISCARD = 2, ++ BTRFS_MAP_GET_READ_MIRRORS = 3, ++}; ++ ++typedef int iterate_extent_inodes_t(u64, u64, u64, u64, void *); ++ ++struct btrfs_backref_walk_ctx { ++ u64 bytenr; ++ u64 extent_item_pos; ++ bool ignore_extent_item_pos; ++ bool skip_inode_ref_list; ++ struct btrfs_trans_handle *trans; ++ struct btrfs_fs_info *fs_info; ++ u64 time_seq; ++ struct ulist *refs; ++ struct ulist *roots; ++ bool (*cache_lookup)(u64, void *, const u64 **, int *); ++ void (*cache_store)(u64, const struct ulist *, void *); ++ iterate_extent_inodes_t *indirect_ref_iterator; ++ int (*check_extent_item)(u64, const struct btrfs_extent_item *, const struct extent_buffer *, void *); ++ bool (*skip_data_ref)(u64, u64, u64, void *); ++ void *user_ctx; ++}; ++ ++struct inode_fs_paths { ++ struct btrfs_path *btrfs_path; ++ struct btrfs_root *fs_root; ++ struct btrfs_data_container *fspath; ++}; ++ ++enum btrfs_block_group_flags { ++ BLOCK_GROUP_FLAG_IREF = 0, ++ BLOCK_GROUP_FLAG_REMOVED = 1, ++ BLOCK_GROUP_FLAG_TO_COPY = 2, ++ BLOCK_GROUP_FLAG_RELOCATING_REPAIR = 3, ++ BLOCK_GROUP_FLAG_CHUNK_ITEM_INSERTED = 4, ++ BLOCK_GROUP_FLAG_ZONE_IS_ACTIVE = 5, ++ BLOCK_GROUP_FLAG_ZONED_DATA_RELOC = 6, ++ BLOCK_GROUP_FLAG_NEEDS_FREE_SPACE = 7, ++ BLOCK_GROUP_FLAG_SEQUENTIAL_ZONE = 8, ++ BLOCK_GROUP_FLAG_NEW = 9, ++}; ++ ++struct scrub_sector_verification { ++ bool is_metadata; ++ union { ++ u8 *csum; ++ u64 generation; ++ }; ++}; ++ ++enum scrub_stripe_flags { ++ SCRUB_STRIPE_FLAG_INITIALIZED = 0, ++ SCRUB_STRIPE_FLAG_REPAIR_DONE = 1, ++ SCRUB_STRIPE_FLAG_NO_REPORT = 2, ++}; ++ ++struct scrub_warning { ++ struct btrfs_path *path; ++ u64 extent_item_size; ++ const char *errstr; ++ u64 physical; ++ u64 logical; ++ struct btrfs_device *dev; ++}; ++ ++struct btrfs_ioctl_dev_replace_start_params { ++ __u64 srcdevid; ++ __u64 cont_reading_from_srcdev_mode; ++ __u8 srcdev_name[1025]; ++ __u8 tgtdev_name[1025]; ++}; ++ ++struct btrfs_ioctl_dev_replace_status_params { ++ __u64 replace_state; ++ __u64 progress_1000; ++ __u64 time_started; ++ __u64 time_stopped; ++ __u64 num_write_errors; ++ __u64 num_uncorrectable_read_errors; ++}; ++ ++struct btrfs_ioctl_dev_replace_args { ++ __u64 cmd; ++ __u64 result; ++ union { ++ struct btrfs_ioctl_dev_replace_start_params start; ++ struct btrfs_ioctl_dev_replace_status_params status; ++ }; ++ __u64 spare[64]; ++}; ++ ++struct btrfs_dev_replace_item { ++ __le64 src_devid; ++ __le64 cursor_left; ++ __le64 cursor_right; ++ __le64 cont_reading_from_srcdev_mode; ++ __le64 replace_state; ++ __le64 time_started; ++ __le64 time_stopped; ++ __le64 num_write_errors; ++ __le64 num_uncorrectable_read_errors; ++}; ++ ++struct btrfs_stripe { ++ __le64 devid; ++ __le64 offset; ++ __u8 dev_uuid[16]; ++}; ++ ++struct btrfs_chunk { ++ __le64 length; ++ __le64 owner; ++ __le64 stripe_len; ++ __le64 type; ++ __le32 io_align; ++ __le32 io_width; ++ __le32 sector_size; ++ __le16 num_stripes; ++ __le16 sub_stripes; ++ struct btrfs_stripe stripe; ++}; ++ ++struct btrfs_tree_block_info { ++ struct btrfs_disk_key key; ++ __u8 level; ++}; ++ ++struct btrfs_extent_data_ref { ++ __le64 root; ++ __le64 objectid; ++ __le64 offset; ++ __le32 count; ++} __attribute__((packed)); ++ ++struct btrfs_shared_data_ref { ++ __le32 count; ++}; ++ ++struct btrfs_extent_inline_ref { ++ __u8 type; ++ __le64 offset; ++} __attribute__((packed)); ++ ++struct btrfs_block_group_item { ++ __le64 used; ++ __le64 chunk_objectid; ++ __le64 flags; ++}; ++ ++struct btrfs_replace_extent_info { ++ u64 disk_offset; ++ u64 disk_len; ++ u64 data_offset; ++ u64 data_len; ++ u64 file_offset; ++ char *extent_buf; ++ bool is_new_extent; ++ bool update_times; ++ int qgroup_reserved; ++ int insertions; ++}; ++ ++enum btrfs_ilock_type { ++ __BTRFS_ILOCK_SHARED_BIT = 0, ++ BTRFS_ILOCK_SHARED = 1, ++ __BTRFS_ILOCK_SHARED_SEQ = 0, ++ __BTRFS_ILOCK_TRY_BIT = 1, ++ BTRFS_ILOCK_TRY = 2, ++ __BTRFS_ILOCK_TRY_SEQ = 1, ++ __BTRFS_ILOCK_MMAP_BIT = 2, ++ BTRFS_ILOCK_MMAP = 4, ++ __BTRFS_ILOCK_MMAP_SEQ = 2, ++}; ++ ++enum { ++ EXTENT_BUFFER_UPTODATE = 0, ++ EXTENT_BUFFER_DIRTY = 1, ++ EXTENT_BUFFER_CORRUPT = 2, ++ EXTENT_BUFFER_READAHEAD = 3, ++ EXTENT_BUFFER_TREE_REF = 4, ++ EXTENT_BUFFER_STALE = 5, ++ EXTENT_BUFFER_WRITEBACK = 6, ++ EXTENT_BUFFER_READ_ERR = 7, ++ EXTENT_BUFFER_UNMAPPED = 8, ++ EXTENT_BUFFER_IN_TREE = 9, ++ EXTENT_BUFFER_WRITE_ERR = 10, ++ EXTENT_BUFFER_NO_CHECK = 11, ++}; ++ ++struct btrfs_seq_list { ++ struct list_head list; ++ u64 seq; ++}; ++ ++enum btrfs_mod_log_op { ++ BTRFS_MOD_LOG_KEY_REPLACE = 0, ++ BTRFS_MOD_LOG_KEY_ADD = 1, ++ BTRFS_MOD_LOG_KEY_REMOVE = 2, ++ BTRFS_MOD_LOG_KEY_REMOVE_WHILE_FREEING = 3, ++ BTRFS_MOD_LOG_KEY_REMOVE_WHILE_MOVING = 4, ++ BTRFS_MOD_LOG_MOVE_KEYS = 5, ++ BTRFS_MOD_LOG_ROOT_REPLACE = 6, ++}; ++ ++struct tree_mod_root { ++ u64 logical; ++ u8 level; ++}; ++ ++struct tree_mod_elem { ++ struct rb_node node; ++ u64 logical; ++ u64 seq; ++ enum btrfs_mod_log_op op; ++ int slot; ++ u64 generation; ++ struct btrfs_disk_key key; ++ u64 blockptr; ++ struct { ++ int dst_slot; ++ int nr_items; ++ } move; ++ struct tree_mod_root old_root; ++}; ++ ++struct btrfs_lru_cache_entry { ++ struct list_head lru_list; ++ u64 key; ++ u64 gen; ++ struct list_head list; ++}; ++ ++struct btrfs_lru_cache { ++ struct list_head lru_list; ++ struct maple_tree entries; ++ unsigned int size; ++ unsigned int max_size; ++}; ++ ++struct btrfs_verity_descriptor_item { ++ __le64 size; ++ __le64 reserved[2]; ++ __u8 encryption; ++} __attribute__((packed)); ++ ++typedef short unsigned int __kernel_old_uid_t; ++ ++typedef short unsigned int __kernel_old_gid_t; ++ ++typedef unsigned int __kernel_mode_t; ++ ++typedef __kernel_old_uid_t old_uid_t; ++ ++typedef __kernel_old_gid_t old_gid_t; ++ ++struct ipc64_perm { ++ __kernel_key_t key; ++ __kernel_uid32_t uid; ++ __kernel_gid32_t gid; ++ __kernel_uid32_t cuid; ++ __kernel_gid32_t cgid; ++ __kernel_mode_t mode; ++ short unsigned int __pad1; ++ short unsigned int seq; ++ long unsigned int __unused1; ++ long unsigned int __unused2; ++}; ++ ++typedef u16 compat_mode_t; ++ ++typedef u16 __compat_uid_t; ++ ++typedef u16 __compat_gid_t; ++ ++typedef s32 compat_key_t; ++ ++typedef u16 compat_ushort_t; ++ ++typedef u32 __compat_gid32_t; ++ ++struct compat_ipc64_perm { ++ compat_key_t key; ++ __compat_uid32_t uid; ++ __compat_gid32_t gid; ++ __compat_uid32_t cuid; ++ __compat_gid32_t cgid; ++ compat_mode_t mode; ++ unsigned char __pad1[2]; ++ compat_ushort_t seq; ++ compat_ushort_t __pad2; ++ compat_ulong_t unused1; ++ compat_ulong_t unused2; ++}; ++ ++struct compat_ipc_perm { ++ key_t key; ++ __compat_uid_t uid; ++ __compat_gid_t gid; ++ __compat_uid_t cuid; ++ __compat_gid_t cgid; ++ compat_mode_t mode; ++ short unsigned int seq; ++}; ++ ++struct rhlist_head { ++ struct rhash_head rhead; ++ struct rhlist_head *next; ++}; ++ ++struct rhash_lock_head; ++ ++struct bucket_table { ++ unsigned int size; ++ unsigned int nest; ++ u32 hash_rnd; ++ struct list_head walkers; ++ struct callback_head rcu; ++ struct bucket_table *future_tbl; ++ struct lockdep_map dep_map; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct rhash_lock_head *buckets[0]; ++}; ++ ++struct ipc_perm { ++ __kernel_key_t key; ++ __kernel_uid_t uid; ++ __kernel_gid_t gid; ++ __kernel_uid_t cuid; ++ __kernel_gid_t cgid; ++ __kernel_mode_t mode; ++ short unsigned int seq; ++}; ++ ++struct ipc_ids { ++ int in_use; ++ short unsigned int seq; ++ struct rw_semaphore rwsem; ++ struct idr ipcs_idr; ++ int max_idx; ++ int last_idx; ++ int next_id; ++ struct rhashtable key_ht; ++}; ++ ++struct ipc_namespace { ++ struct ipc_ids ids[3]; ++ int sem_ctls[4]; ++ int used_sems; ++ unsigned int msg_ctlmax; ++ unsigned int msg_ctlmnb; ++ unsigned int msg_ctlmni; ++ struct percpu_counter percpu_msg_bytes; ++ struct percpu_counter percpu_msg_hdrs; ++ size_t shm_ctlmax; ++ size_t shm_ctlall; ++ long unsigned int shm_tot; ++ int shm_ctlmni; ++ int shm_rmid_forced; ++ struct notifier_block ipcns_nb; ++ struct vfsmount *mq_mnt; ++ unsigned int mq_queues_count; ++ unsigned int mq_queues_max; ++ unsigned int mq_msg_max; ++ unsigned int mq_msgsize_max; ++ unsigned int mq_msg_default; ++ unsigned int mq_msgsize_default; ++ struct ctl_table_set mq_set; ++ struct ctl_table_header *mq_sysctls; ++ struct ctl_table_set ipc_set; ++ struct ctl_table_header *ipc_sysctls; ++ struct user_namespace *user_ns; ++ struct ucounts *ucounts; ++ struct llist_node mnt_llist; ++ struct ns_common ns; ++}; ++ ++struct rhash_lock_head {}; ++ ++struct ipc_params { ++ key_t key; ++ int flg; ++ union { ++ size_t size; ++ int nsems; ++ } u; ++}; ++ ++struct ipc_ops { ++ int (*getnew)(struct ipc_namespace *, struct ipc_params *); ++ int (*associate)(struct kern_ipc_perm *, int); ++ int (*more_checks)(struct kern_ipc_perm *, struct ipc_params *); ++}; ++ ++struct ipc_proc_iface { ++ const char *path; ++ const char *header; ++ int ids; ++ int (*show)(struct seq_file *, void *); ++}; ++ ++struct ipc_proc_iter { ++ struct ipc_namespace *ns; ++ struct pid_namespace *pid_ns; ++ struct ipc_proc_iface *iface; ++}; ++ ++typedef int __kernel_ipc_pid_t; ++ ++typedef __kernel_long_t __kernel_old_time_t; ++ ++struct shmid_ds { ++ struct ipc_perm shm_perm; ++ int shm_segsz; ++ __kernel_old_time_t shm_atime; ++ __kernel_old_time_t shm_dtime; ++ __kernel_old_time_t shm_ctime; ++ __kernel_ipc_pid_t shm_cpid; ++ __kernel_ipc_pid_t shm_lpid; ++ short unsigned int shm_nattch; ++ short unsigned int shm_unused; ++ void *shm_unused2; ++ void *shm_unused3; ++}; ++ ++struct shmid64_ds { ++ struct ipc64_perm shm_perm; ++ __kernel_size_t shm_segsz; ++ long int shm_atime; ++ long int shm_dtime; ++ long int shm_ctime; ++ __kernel_pid_t shm_cpid; ++ __kernel_pid_t shm_lpid; ++ long unsigned int shm_nattch; ++ long unsigned int __unused4; ++ long unsigned int __unused5; ++}; ++ ++struct shminfo64 { ++ long unsigned int shmmax; ++ long unsigned int shmmin; ++ long unsigned int shmmni; ++ long unsigned int shmseg; ++ long unsigned int shmall; ++ long unsigned int __unused1; ++ long unsigned int __unused2; ++ long unsigned int __unused3; ++ long unsigned int __unused4; ++}; ++ ++struct shminfo { ++ int shmmax; ++ int shmmin; ++ int shmmni; ++ int shmseg; ++ int shmall; ++}; ++ ++struct shm_info { ++ int used_ids; ++ __kernel_ulong_t shm_tot; ++ __kernel_ulong_t shm_rss; ++ __kernel_ulong_t shm_swp; ++ __kernel_ulong_t swap_attempts; ++ __kernel_ulong_t swap_successes; ++}; ++ ++typedef u16 compat_ipc_pid_t; ++ ++struct compat_shmid64_ds { ++ struct compat_ipc64_perm shm_perm; ++ compat_size_t shm_segsz; ++ compat_ulong_t shm_atime; ++ compat_ulong_t shm_atime_high; ++ compat_ulong_t shm_dtime; ++ compat_ulong_t shm_dtime_high; ++ compat_ulong_t shm_ctime; ++ compat_ulong_t shm_ctime_high; ++ compat_pid_t shm_cpid; ++ compat_pid_t shm_lpid; ++ compat_ulong_t shm_nattch; ++ compat_ulong_t __unused4; ++ compat_ulong_t __unused5; ++}; ++ ++struct shmid_kernel { ++ struct kern_ipc_perm shm_perm; ++ struct file *shm_file; ++ long unsigned int shm_nattch; ++ long unsigned int shm_segsz; ++ time64_t shm_atim; ++ time64_t shm_dtim; ++ time64_t shm_ctim; ++ struct pid *shm_cprid; ++ struct pid *shm_lprid; ++ struct ucounts *mlock_ucounts; ++ struct task_struct *shm_creator; ++ struct list_head shm_clist; ++ struct ipc_namespace *ns; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct shm_file_data { ++ int id; ++ struct ipc_namespace *ns; ++ struct file *file; ++ const struct vm_operations_struct *vm_ops; ++}; ++ ++struct compat_shmid_ds { ++ struct compat_ipc_perm shm_perm; ++ int shm_segsz; ++ old_time32_t shm_atime; ++ old_time32_t shm_dtime; ++ old_time32_t shm_ctime; ++ compat_ipc_pid_t shm_cpid; ++ compat_ipc_pid_t shm_lpid; ++ short unsigned int shm_nattch; ++ short unsigned int shm_unused; ++ compat_uptr_t shm_unused2; ++ compat_uptr_t shm_unused3; ++}; ++ ++struct compat_shminfo64 { ++ compat_ulong_t shmmax; ++ compat_ulong_t shmmin; ++ compat_ulong_t shmmni; ++ compat_ulong_t shmseg; ++ compat_ulong_t shmall; ++ compat_ulong_t __unused1; ++ compat_ulong_t __unused2; ++ compat_ulong_t __unused3; ++ compat_ulong_t __unused4; ++}; ++ ++struct compat_shm_info { ++ compat_int_t used_ids; ++ compat_ulong_t shm_tot; ++ compat_ulong_t shm_rss; ++ compat_ulong_t shm_swp; ++ compat_ulong_t swap_attempts; ++ compat_ulong_t swap_successes; ++}; ++ ++enum inode_i_mutex_lock_class { ++ I_MUTEX_NORMAL = 0, ++ I_MUTEX_PARENT = 1, ++ I_MUTEX_CHILD = 2, ++ I_MUTEX_XATTR = 3, ++ I_MUTEX_NONDIR2 = 4, ++ I_MUTEX_PARENT2 = 5, ++}; ++ ++struct msg_msgseg; ++ ++struct msg_msg { ++ struct list_head m_list; ++ long int m_type; ++ size_t m_ts; ++ struct msg_msgseg *next; ++ void *security; ++}; ++ ++struct mqueue_fs_context { ++ struct ipc_namespace *ipc_ns; ++ bool newns; ++}; ++ ++struct posix_msg_tree_node { ++ struct rb_node rb_node; ++ struct list_head msg_list; ++ int priority; ++}; ++ ++struct ext_wait_queue { ++ struct task_struct *task; ++ struct list_head list; ++ struct msg_msg *msg; ++ int state; ++}; ++ ++struct mqueue_inode_info { ++ spinlock_t lock; ++ struct inode vfs_inode; ++ wait_queue_head_t wait_q; ++ struct rb_root msg_tree; ++ struct rb_node *msg_tree_rightmost; ++ struct posix_msg_tree_node *node_cache; ++ struct mq_attr attr; ++ struct sigevent notify; ++ struct pid *notify_owner; ++ u32 notify_self_exec_id; ++ struct user_namespace *notify_user_ns; ++ struct ucounts *ucounts; ++ struct sock *notify_sock; ++ struct sk_buff *notify_cookie; ++ struct ext_wait_queue e_wait_q[2]; ++ long unsigned int qsize; ++}; ++ ++struct compat_mq_attr { ++ compat_long_t mq_flags; ++ compat_long_t mq_maxmsg; ++ compat_long_t mq_msgsize; ++ compat_long_t mq_curmsgs; ++ compat_long_t __reserved[4]; ++}; ++ ++struct key_user { ++ struct rb_node node; ++ struct mutex cons_lock; ++ spinlock_t lock; ++ refcount_t usage; ++ atomic_t nkeys; ++ atomic_t nikeys; ++ kuid_t uid; ++ int qnkeys; ++ int qnbytes; ++}; ++ ++struct kernel_pkey_query { ++ __u32 supported_ops; ++ __u32 key_size; ++ __u16 max_data_size; ++ __u16 max_sig_size; ++ __u16 max_enc_size; ++ __u16 max_dec_size; ++}; ++ ++struct keyring_search_context { ++ struct keyring_index_key index_key; ++ const struct cred *cred; ++ struct key_match_data match_data; ++ unsigned int flags; ++ int (*iterator)(const void *, void *); ++ int skipped_ret; ++ bool possessed; ++ key_ref_t result; ++ time64_t now; ++}; ++ ++struct request_key_auth { ++ struct callback_head rcu; ++ struct key *target_key; ++ struct key *dest_keyring; ++ const struct cred *cred; ++ void *callout_info; ++ size_t callout_len; ++ pid_t pid; ++ char op[8]; ++}; ++ ++struct compat_keyctl_kdf_params { ++ compat_uptr_t hashname; ++ compat_uptr_t otherinfo; ++ __u32 otherinfolen; ++ __u32 __spare[8]; ++}; ++ ++struct keyctl_dh_params { ++ union { ++ __s32 private; ++ __s32 priv; ++ }; ++ __s32 prime; ++ __s32 base; ++}; ++ ++struct keyctl_pkey_query { ++ __u32 supported_ops; ++ __u32 key_size; ++ __u16 max_data_size; ++ __u16 max_sig_size; ++ __u16 max_enc_size; ++ __u16 max_dec_size; ++ __u32 __spare[10]; ++}; ++ ++struct keyctl_pkey_params { ++ __s32 key_id; ++ __u32 in_len; ++ union { ++ __u32 out_len; ++ __u32 in2_len; ++ }; ++ __u32 __spare[7]; ++}; ++ ++struct keyctl_kdf_params { ++ char *hashname; ++ char *otherinfo; ++ __u32 otherinfolen; ++ __u32 __spare[8]; ++}; ++ ++enum key_lookup_flag { ++ KEY_LOOKUP_CREATE = 1, ++ KEY_LOOKUP_PARTIAL = 2, ++ KEY_LOOKUP_ALL = 3, ++}; ++ ++enum tpm_duration { ++ TPM_SHORT = 0, ++ TPM_MEDIUM = 1, ++ TPM_LONG = 2, ++ TPM_LONG_LONG = 3, ++ TPM_UNDEFINED = 4, ++ TPM_NUM_DURATIONS = 4, ++}; ++ ++struct trusted_key_payload { ++ struct callback_head rcu; ++ unsigned int key_len; ++ unsigned int blob_len; ++ unsigned char migratable; ++ unsigned char old_format; ++ unsigned char key[129]; ++ unsigned char blob[512]; ++}; ++ ++struct trusted_key_ops { ++ unsigned char migratable; ++ int (*init)(); ++ int (*seal)(struct trusted_key_payload *, char *); ++ int (*unseal)(struct trusted_key_payload *, char *); ++ int (*get_random)(unsigned char *, size_t); ++ void (*exit)(); ++}; ++ ++struct trusted_key_source { ++ char *name; ++ struct trusted_key_ops *ops; ++}; ++ ++enum { ++ Opt_err___3 = 0, ++ Opt_new = 1, ++ Opt_load = 2, ++ Opt_update = 3, ++}; ++ ++typedef int (*asn1_action_t)(void *, size_t, unsigned char, const void *, size_t); ++ ++struct asn1_decoder { ++ const unsigned char *machine; ++ size_t machlen; ++ const asn1_action_t *actions; ++}; ++ ++enum OID { ++ OID_id_dsa_with_sha1 = 0, ++ OID_id_dsa = 1, ++ OID_id_ecPublicKey = 2, ++ OID_id_prime192v1 = 3, ++ OID_id_prime256v1 = 4, ++ OID_id_ecdsa_with_sha1 = 5, ++ OID_id_ecdsa_with_sha224 = 6, ++ OID_id_ecdsa_with_sha256 = 7, ++ OID_id_ecdsa_with_sha384 = 8, ++ OID_id_ecdsa_with_sha512 = 9, ++ OID_rsaEncryption = 10, ++ OID_md2WithRSAEncryption = 11, ++ OID_md3WithRSAEncryption = 12, ++ OID_md4WithRSAEncryption = 13, ++ OID_sha1WithRSAEncryption = 14, ++ OID_sha256WithRSAEncryption = 15, ++ OID_sha384WithRSAEncryption = 16, ++ OID_sha512WithRSAEncryption = 17, ++ OID_sha224WithRSAEncryption = 18, ++ OID_data = 19, ++ OID_signed_data = 20, ++ OID_email_address = 21, ++ OID_contentType = 22, ++ OID_messageDigest = 23, ++ OID_signingTime = 24, ++ OID_smimeCapabilites = 25, ++ OID_smimeAuthenticatedAttrs = 26, ++ OID_md2 = 27, ++ OID_md4 = 28, ++ OID_md5 = 29, ++ OID_mskrb5 = 30, ++ OID_krb5 = 31, ++ OID_krb5u2u = 32, ++ OID_msIndirectData = 33, ++ OID_msStatementType = 34, ++ OID_msSpOpusInfo = 35, ++ OID_msPeImageDataObjId = 36, ++ OID_msIndividualSPKeyPurpose = 37, ++ OID_msOutlookExpress = 38, ++ OID_ntlmssp = 39, ++ OID_spnego = 40, ++ OID_IAKerb = 41, ++ OID_PKU2U = 42, ++ OID_Scram = 43, ++ OID_certAuthInfoAccess = 44, ++ OID_sha1 = 45, ++ OID_id_ansip384r1 = 46, ++ OID_sha256 = 47, ++ OID_sha384 = 48, ++ OID_sha512 = 49, ++ OID_sha224 = 50, ++ OID_commonName = 51, ++ OID_surname = 52, ++ OID_countryName = 53, ++ OID_locality = 54, ++ OID_stateOrProvinceName = 55, ++ OID_organizationName = 56, ++ OID_organizationUnitName = 57, ++ OID_title = 58, ++ OID_description = 59, ++ OID_name = 60, ++ OID_givenName = 61, ++ OID_initials = 62, ++ OID_generationalQualifier = 63, ++ OID_subjectKeyIdentifier = 64, ++ OID_keyUsage = 65, ++ OID_subjectAltName = 66, ++ OID_issuerAltName = 67, ++ OID_basicConstraints = 68, ++ OID_crlDistributionPoints = 69, ++ OID_certPolicies = 70, ++ OID_authorityKeyIdentifier = 71, ++ OID_extKeyUsage = 72, ++ OID_NetlogonMechanism = 73, ++ OID_appleLocalKdcSupported = 74, ++ OID_gostCPSignA = 75, ++ OID_gostCPSignB = 76, ++ OID_gostCPSignC = 77, ++ OID_gost2012PKey256 = 78, ++ OID_gost2012PKey512 = 79, ++ OID_gost2012Digest256 = 80, ++ OID_gost2012Digest512 = 81, ++ OID_gost2012Signature256 = 82, ++ OID_gost2012Signature512 = 83, ++ OID_gostTC26Sign256A = 84, ++ OID_gostTC26Sign256B = 85, ++ OID_gostTC26Sign256C = 86, ++ OID_gostTC26Sign256D = 87, ++ OID_gostTC26Sign512A = 88, ++ OID_gostTC26Sign512B = 89, ++ OID_gostTC26Sign512C = 90, ++ OID_sm2 = 91, ++ OID_sm3 = 92, ++ OID_SM2_with_SM3 = 93, ++ OID_sm3WithRSAEncryption = 94, ++ OID_TPMLoadableKey = 95, ++ OID_TPMImportableKey = 96, ++ OID_TPMSealedData = 97, ++ OID__NR = 98, ++}; ++ ++struct hwrng { ++ const char *name; ++ int (*init)(struct hwrng *); ++ void (*cleanup)(struct hwrng *); ++ int (*data_present)(struct hwrng *, int); ++ int (*data_read)(struct hwrng *, u32 *); ++ int (*read)(struct hwrng *, void *, size_t, bool); ++ long unsigned int priv; ++ short unsigned int quality; ++ struct list_head list; ++ struct kref ref; ++ struct completion cleanup_done; ++ struct completion dying; ++}; ++ ++enum tpm_algorithms { ++ TPM_ALG_ERROR = 0, ++ TPM_ALG_SHA1 = 4, ++ TPM_ALG_KEYEDHASH = 8, ++ TPM_ALG_SHA256 = 11, ++ TPM_ALG_SHA384 = 12, ++ TPM_ALG_SHA512 = 13, ++ TPM_ALG_NULL = 16, ++ TPM_ALG_SM3_256 = 18, ++}; ++ ++struct tpm_bank_info { ++ u16 alg_id; ++ u16 digest_size; ++ u16 crypto_id; ++}; ++ ++struct tpm_chip; ++ ++struct tpm_class_ops { ++ unsigned int flags; ++ const u8 req_complete_mask; ++ const u8 req_complete_val; ++ bool (*req_canceled)(struct tpm_chip *, u8); ++ int (*recv)(struct tpm_chip *, u8 *, size_t); ++ int (*send)(struct tpm_chip *, u8 *, size_t); ++ void (*cancel)(struct tpm_chip *); ++ u8 (*status)(struct tpm_chip *); ++ void (*update_timeouts)(struct tpm_chip *, long unsigned int *); ++ void (*update_durations)(struct tpm_chip *, long unsigned int *); ++ int (*go_idle)(struct tpm_chip *); ++ int (*cmd_ready)(struct tpm_chip *); ++ int (*request_locality)(struct tpm_chip *, int); ++ int (*relinquish_locality)(struct tpm_chip *, int); ++ void (*clk_enable)(struct tpm_chip *, bool); ++}; ++ ++struct tpm_bios_log { ++ void *bios_event_log; ++ void *bios_event_log_end; ++}; ++ ++struct tpm_chip_seqops { ++ struct tpm_chip *chip; ++ const struct seq_operations *seqops; ++}; ++ ++struct tpm_space { ++ u32 context_tbl[3]; ++ u8 *context_buf; ++ u32 session_tbl[3]; ++ u8 *session_buf; ++ u32 buf_size; ++}; ++ ++struct tpm_chip { ++ struct device dev; ++ struct device devs; ++ struct cdev cdev; ++ struct cdev cdevs; ++ struct rw_semaphore ops_sem; ++ const struct tpm_class_ops *ops; ++ struct tpm_bios_log log; ++ struct tpm_chip_seqops bin_log_seqops; ++ struct tpm_chip_seqops ascii_log_seqops; ++ unsigned int flags; ++ int dev_num; ++ long unsigned int is_open; ++ char hwrng_name[64]; ++ struct hwrng hwrng; ++ struct mutex tpm_mutex; ++ long unsigned int timeout_a; ++ long unsigned int timeout_b; ++ long unsigned int timeout_c; ++ long unsigned int timeout_d; ++ bool timeout_adjusted; ++ long unsigned int duration[4]; ++ bool duration_adjusted; ++ struct dentry *bios_dir[3]; ++ const struct attribute_group *groups[8]; ++ unsigned int groups_cnt; ++ u32 nr_allocated_banks; ++ struct tpm_bank_info *allocated_banks; ++ struct tpm_space work_space; ++ u32 last_cc; ++ u32 nr_commands; ++ u32 *cc_attrs_tbl; ++ int locality; ++}; ++ ++enum tpm2_structures { ++ TPM2_ST_NO_SESSIONS = 32769, ++ TPM2_ST_SESSIONS = 32770, ++}; ++ ++enum tpm2_return_codes { ++ TPM2_RC_SUCCESS = 0, ++ TPM2_RC_HASH = 131, ++ TPM2_RC_HANDLE = 139, ++ TPM2_RC_INITIALIZE = 256, ++ TPM2_RC_FAILURE = 257, ++ TPM2_RC_DISABLED = 288, ++ TPM2_RC_UPGRADE = 301, ++ TPM2_RC_COMMAND_CODE = 323, ++ TPM2_RC_TESTING = 2314, ++ TPM2_RC_REFERENCE_H0 = 2320, ++ TPM2_RC_RETRY = 2338, ++}; ++ ++enum tpm2_command_codes { ++ TPM2_CC_FIRST = 287, ++ TPM2_CC_HIERARCHY_CONTROL = 289, ++ TPM2_CC_HIERARCHY_CHANGE_AUTH = 297, ++ TPM2_CC_CREATE_PRIMARY = 305, ++ TPM2_CC_SEQUENCE_COMPLETE = 318, ++ TPM2_CC_SELF_TEST = 323, ++ TPM2_CC_STARTUP = 324, ++ TPM2_CC_SHUTDOWN = 325, ++ TPM2_CC_NV_READ = 334, ++ TPM2_CC_CREATE = 339, ++ TPM2_CC_LOAD = 343, ++ TPM2_CC_SEQUENCE_UPDATE = 348, ++ TPM2_CC_UNSEAL = 350, ++ TPM2_CC_CONTEXT_LOAD = 353, ++ TPM2_CC_CONTEXT_SAVE = 354, ++ TPM2_CC_FLUSH_CONTEXT = 357, ++ TPM2_CC_VERIFY_SIGNATURE = 375, ++ TPM2_CC_GET_CAPABILITY = 378, ++ TPM2_CC_GET_RANDOM = 379, ++ TPM2_CC_PCR_READ = 382, ++ TPM2_CC_PCR_EXTEND = 386, ++ TPM2_CC_EVENT_SEQUENCE_COMPLETE = 389, ++ TPM2_CC_HASH_SEQUENCE_START = 390, ++ TPM2_CC_CREATE_LOADED = 401, ++ TPM2_CC_LAST = 403, ++}; ++ ++enum tpm2_permanent_handles { ++ TPM2_RS_PW = 1073741833, ++}; ++ ++struct tpm_header { ++ __be16 tag; ++ __be32 length; ++ union { ++ __be32 ordinal; ++ __be32 return_code; ++ }; ++} __attribute__((packed)); ++ ++enum tpm_buf_flags { ++ TPM_BUF_OVERFLOW = 1, ++}; ++ ++struct tpm_buf { ++ unsigned int flags; ++ u8 *data; ++}; ++ ++enum tpm2_object_attributes { ++ TPM2_OA_FIXED_TPM = 2, ++ TPM2_OA_FIXED_PARENT = 16, ++ TPM2_OA_USER_WITH_AUTH = 64, ++}; ++ ++enum tpm2_session_attributes { ++ TPM2_SA_CONTINUE_SESSION = 1, ++}; ++ ++struct tpm2_hash { ++ unsigned int crypto_id; ++ unsigned int tpm_id; ++}; ++ ++struct trusted_key_options { ++ uint16_t keytype; ++ uint32_t keyhandle; ++ unsigned char keyauth[20]; ++ uint32_t blobauth_len; ++ unsigned char blobauth[20]; ++ uint32_t pcrinfo_len; ++ unsigned char pcrinfo[64]; ++ int pcrlock; ++ uint32_t hash; ++ uint32_t policydigest_len; ++ unsigned char policydigest[64]; ++ uint32_t policyhandle; ++}; ++ ++struct tpm2_key_context { ++ u32 parent; ++ const u8 *pub; ++ u32 pub_len; ++ const u8 *priv; ++ u32 priv_len; ++}; ++ ++enum lsm_event { ++ LSM_POLICY_CHANGE = 0, ++}; ++ ++struct sctp_association; ++ ++union security_list_options { ++ int (*binder_set_context_mgr)(const struct cred *); ++ int (*binder_transaction)(const struct cred *, const struct cred *); ++ int (*binder_transfer_binder)(const struct cred *, const struct cred *); ++ int (*binder_transfer_file)(const struct cred *, const struct cred *, struct file *); ++ int (*ptrace_access_check)(struct task_struct *, unsigned int); ++ int (*ptrace_traceme)(struct task_struct *); ++ int (*capget)(struct task_struct *, kernel_cap_t *, kernel_cap_t *, kernel_cap_t *); ++ int (*capset)(struct cred *, const struct cred *, const kernel_cap_t *, const kernel_cap_t *, const kernel_cap_t *); ++ int (*capable)(const struct cred *, struct user_namespace *, int, unsigned int); ++ int (*quotactl)(int, int, int, struct super_block *); ++ int (*quota_on)(struct dentry *); ++ int (*syslog)(int); ++ int (*settime)(const struct timespec64 *, const struct timezone *); ++ int (*vm_enough_memory)(struct mm_struct *, long int); ++ int (*bprm_creds_for_exec)(struct linux_binprm *); ++ int (*bprm_creds_from_file)(struct linux_binprm *, struct file *); ++ int (*bprm_check_security)(struct linux_binprm *); ++ void (*bprm_committing_creds)(struct linux_binprm *); ++ void (*bprm_committed_creds)(struct linux_binprm *); ++ int (*fs_context_dup)(struct fs_context *, struct fs_context *); ++ int (*fs_context_parse_param)(struct fs_context *, struct fs_parameter *); ++ int (*sb_alloc_security)(struct super_block *); ++ void (*sb_delete)(struct super_block *); ++ void (*sb_free_security)(struct super_block *); ++ void (*sb_free_mnt_opts)(void *); ++ int (*sb_eat_lsm_opts)(char *, void **); ++ int (*sb_mnt_opts_compat)(struct super_block *, void *); ++ int (*sb_remount)(struct super_block *, void *); ++ int (*sb_kern_mount)(struct super_block *); ++ int (*sb_show_options)(struct seq_file *, struct super_block *); ++ int (*sb_statfs)(struct dentry *); ++ int (*sb_mount)(const char *, const struct path *, const char *, long unsigned int, void *); ++ int (*sb_umount)(struct vfsmount *, int); ++ int (*sb_pivotroot)(const struct path *, const struct path *); ++ int (*sb_set_mnt_opts)(struct super_block *, void *, long unsigned int, long unsigned int *); ++ int (*sb_clone_mnt_opts)(const struct super_block *, struct super_block *, long unsigned int, long unsigned int *); ++ int (*move_mount)(const struct path *, const struct path *); ++ int (*dentry_init_security)(struct dentry *, int, const struct qstr *, const char **, void **, u32 *); ++ int (*dentry_create_files_as)(struct dentry *, int, struct qstr *, const struct cred *, struct cred *); ++ int (*path_unlink)(const struct path *, struct dentry *); ++ int (*path_mkdir)(const struct path *, struct dentry *, umode_t); ++ int (*path_rmdir)(const struct path *, struct dentry *); ++ int (*path_mknod)(const struct path *, struct dentry *, umode_t, unsigned int); ++ int (*path_truncate)(const struct path *); ++ int (*path_symlink)(const struct path *, struct dentry *, const char *); ++ int (*path_link)(struct dentry *, const struct path *, struct dentry *); ++ int (*path_rename)(const struct path *, struct dentry *, const struct path *, struct dentry *, unsigned int); ++ int (*path_chmod)(const struct path *, umode_t); ++ int (*path_chown)(const struct path *, kuid_t, kgid_t); ++ int (*path_chroot)(const struct path *); ++ int (*path_notify)(const struct path *, u64, unsigned int); ++ int (*inode_alloc_security)(struct inode *); ++ void (*inode_free_security)(struct inode *); ++ int (*inode_init_security)(struct inode *, struct inode *, const struct qstr *, const char **, void **, size_t *); ++ int (*inode_init_security_anon)(struct inode *, const struct qstr *, const struct inode *); ++ int (*inode_create)(struct inode *, struct dentry *, umode_t); ++ int (*inode_link)(struct dentry *, struct inode *, struct dentry *); ++ int (*inode_unlink)(struct inode *, struct dentry *); ++ int (*inode_symlink)(struct inode *, struct dentry *, const char *); ++ int (*inode_mkdir)(struct inode *, struct dentry *, umode_t); ++ int (*inode_rmdir)(struct inode *, struct dentry *); ++ int (*inode_mknod)(struct inode *, struct dentry *, umode_t, dev_t); ++ int (*inode_rename)(struct inode *, struct dentry *, struct inode *, struct dentry *); ++ int (*inode_readlink)(struct dentry *); ++ int (*inode_follow_link)(struct dentry *, struct inode *, bool); ++ int (*inode_permission)(struct inode *, int); ++ int (*inode_setattr)(struct dentry *, struct iattr *); ++ int (*inode_getattr)(const struct path *); ++ int (*inode_setxattr)(struct mnt_idmap *, struct dentry *, const char *, const void *, size_t, int); ++ void (*inode_post_setxattr)(struct dentry *, const char *, const void *, size_t, int); ++ int (*inode_getxattr)(struct dentry *, const char *); ++ int (*inode_listxattr)(struct dentry *); ++ int (*inode_removexattr)(struct mnt_idmap *, struct dentry *, const char *); ++ int (*inode_set_acl)(struct mnt_idmap *, struct dentry *, const char *, struct posix_acl *); ++ int (*inode_get_acl)(struct mnt_idmap *, struct dentry *, const char *); ++ int (*inode_remove_acl)(struct mnt_idmap *, struct dentry *, const char *); ++ int (*inode_need_killpriv)(struct dentry *); ++ int (*inode_killpriv)(struct mnt_idmap *, struct dentry *); ++ int (*inode_getsecurity)(struct mnt_idmap *, struct inode *, const char *, void **, bool); ++ int (*inode_setsecurity)(struct inode *, const char *, const void *, size_t, int); ++ int (*inode_listsecurity)(struct inode *, char *, size_t); ++ void (*inode_getsecid)(struct inode *, u32 *); ++ int (*inode_copy_up)(struct dentry *, struct cred **); ++ int (*inode_copy_up_xattr)(const char *); ++ int (*kernfs_init_security)(struct kernfs_node *, struct kernfs_node *); ++ int (*file_permission)(struct file *, int); ++ int (*file_alloc_security)(struct file *); ++ void (*file_free_security)(struct file *); ++ int (*file_ioctl)(struct file *, unsigned int, long unsigned int); ++ int (*mmap_addr)(long unsigned int); ++ int (*mmap_file)(struct file *, long unsigned int, long unsigned int, long unsigned int); ++ int (*file_mprotect)(struct vm_area_struct *, long unsigned int, long unsigned int); ++ int (*file_lock)(struct file *, unsigned int); ++ int (*file_fcntl)(struct file *, unsigned int, long unsigned int); ++ void (*file_set_fowner)(struct file *); ++ int (*file_send_sigiotask)(struct task_struct *, struct fown_struct *, int); ++ int (*file_receive)(struct file *); ++ int (*file_open)(struct file *); ++ int (*file_truncate)(struct file *); ++ int (*task_alloc)(struct task_struct *, long unsigned int); ++ void (*task_free)(struct task_struct *); ++ int (*cred_alloc_blank)(struct cred *, gfp_t); ++ void (*cred_free)(struct cred *); ++ int (*cred_prepare)(struct cred *, const struct cred *, gfp_t); ++ void (*cred_transfer)(struct cred *, const struct cred *); ++ void (*cred_getsecid)(const struct cred *, u32 *); ++ int (*kernel_act_as)(struct cred *, u32); ++ int (*kernel_create_files_as)(struct cred *, struct inode *); ++ int (*kernel_module_request)(char *); ++ int (*kernel_load_data)(enum kernel_load_data_id, bool); ++ int (*kernel_post_load_data)(char *, loff_t, enum kernel_load_data_id, char *); ++ int (*kernel_read_file)(struct file *, enum kernel_read_file_id, bool); ++ int (*kernel_post_read_file)(struct file *, char *, loff_t, enum kernel_read_file_id); ++ int (*task_fix_setuid)(struct cred *, const struct cred *, int); ++ int (*task_fix_setgid)(struct cred *, const struct cred *, int); ++ int (*task_fix_setgroups)(struct cred *, const struct cred *); ++ int (*task_setpgid)(struct task_struct *, pid_t); ++ int (*task_getpgid)(struct task_struct *); ++ int (*task_getsid)(struct task_struct *); ++ void (*current_getsecid_subj)(u32 *); ++ void (*task_getsecid_obj)(struct task_struct *, u32 *); ++ int (*task_setnice)(struct task_struct *, int); ++ int (*task_setioprio)(struct task_struct *, int); ++ int (*task_getioprio)(struct task_struct *); ++ int (*task_prlimit)(const struct cred *, const struct cred *, unsigned int); ++ int (*task_setrlimit)(struct task_struct *, unsigned int, struct rlimit *); ++ int (*task_setscheduler)(struct task_struct *); ++ int (*task_getscheduler)(struct task_struct *); ++ int (*task_movememory)(struct task_struct *); ++ int (*task_kill)(struct task_struct *, struct kernel_siginfo *, int, const struct cred *); ++ int (*task_prctl)(int, long unsigned int, long unsigned int, long unsigned int, long unsigned int); ++ void (*task_to_inode)(struct task_struct *, struct inode *); ++ int (*userns_create)(const struct cred *); ++ int (*ipc_permission)(struct kern_ipc_perm *, short int); ++ void (*ipc_getsecid)(struct kern_ipc_perm *, u32 *); ++ int (*msg_msg_alloc_security)(struct msg_msg *); ++ void (*msg_msg_free_security)(struct msg_msg *); ++ int (*msg_queue_alloc_security)(struct kern_ipc_perm *); ++ void (*msg_queue_free_security)(struct kern_ipc_perm *); ++ int (*msg_queue_associate)(struct kern_ipc_perm *, int); ++ int (*msg_queue_msgctl)(struct kern_ipc_perm *, int); ++ int (*msg_queue_msgsnd)(struct kern_ipc_perm *, struct msg_msg *, int); ++ int (*msg_queue_msgrcv)(struct kern_ipc_perm *, struct msg_msg *, struct task_struct *, long int, int); ++ int (*shm_alloc_security)(struct kern_ipc_perm *); ++ void (*shm_free_security)(struct kern_ipc_perm *); ++ int (*shm_associate)(struct kern_ipc_perm *, int); ++ int (*shm_shmctl)(struct kern_ipc_perm *, int); ++ int (*shm_shmat)(struct kern_ipc_perm *, char *, int); ++ int (*sem_alloc_security)(struct kern_ipc_perm *); ++ void (*sem_free_security)(struct kern_ipc_perm *); ++ int (*sem_associate)(struct kern_ipc_perm *, int); ++ int (*sem_semctl)(struct kern_ipc_perm *, int); ++ int (*sem_semop)(struct kern_ipc_perm *, struct sembuf *, unsigned int, int); ++ int (*netlink_send)(struct sock *, struct sk_buff *); ++ void (*d_instantiate)(struct dentry *, struct inode *); ++ int (*getprocattr)(struct task_struct *, const char *, char **); ++ int (*setprocattr)(const char *, void *, size_t); ++ int (*ismaclabel)(const char *); ++ int (*secid_to_secctx)(u32, char **, u32 *); ++ int (*secctx_to_secid)(const char *, u32, u32 *); ++ void (*release_secctx)(char *, u32); ++ void (*inode_invalidate_secctx)(struct inode *); ++ int (*inode_notifysecctx)(struct inode *, void *, u32); ++ int (*inode_setsecctx)(struct dentry *, void *, u32); ++ int (*inode_getsecctx)(struct inode *, void **, u32 *); ++ int (*post_notification)(const struct cred *, const struct cred *, struct watch_notification *); ++ int (*watch_key)(struct key *); ++ int (*unix_stream_connect)(struct sock *, struct sock *, struct sock *); ++ int (*unix_may_send)(struct socket *, struct socket *); ++ int (*socket_create)(int, int, int, int); ++ int (*socket_post_create)(struct socket *, int, int, int, int); ++ int (*socket_socketpair)(struct socket *, struct socket *); ++ int (*socket_bind)(struct socket *, struct sockaddr *, int); ++ int (*socket_connect)(struct socket *, struct sockaddr *, int); ++ int (*socket_listen)(struct socket *, int); ++ int (*socket_accept)(struct socket *, struct socket *); ++ int (*socket_sendmsg)(struct socket *, struct msghdr *, int); ++ int (*socket_recvmsg)(struct socket *, struct msghdr *, int, int); ++ int (*socket_getsockname)(struct socket *); ++ int (*socket_getpeername)(struct socket *); ++ int (*socket_getsockopt)(struct socket *, int, int); ++ int (*socket_setsockopt)(struct socket *, int, int); ++ int (*socket_shutdown)(struct socket *, int); ++ int (*socket_sock_rcv_skb)(struct sock *, struct sk_buff *); ++ int (*socket_getpeersec_stream)(struct socket *, sockptr_t, sockptr_t, unsigned int); ++ int (*socket_getpeersec_dgram)(struct socket *, struct sk_buff *, u32 *); ++ int (*sk_alloc_security)(struct sock *, int, gfp_t); ++ void (*sk_free_security)(struct sock *); ++ void (*sk_clone_security)(const struct sock *, struct sock *); ++ void (*sk_getsecid)(struct sock *, u32 *); ++ void (*sock_graft)(struct sock *, struct socket *); ++ int (*inet_conn_request)(const struct sock *, struct sk_buff *, struct request_sock *); ++ void (*inet_csk_clone)(struct sock *, const struct request_sock *); ++ void (*inet_conn_established)(struct sock *, struct sk_buff *); ++ int (*secmark_relabel_packet)(u32); ++ void (*secmark_refcount_inc)(); ++ void (*secmark_refcount_dec)(); ++ void (*req_classify_flow)(const struct request_sock *, struct flowi_common *); ++ int (*tun_dev_alloc_security)(void **); ++ void (*tun_dev_free_security)(void *); ++ int (*tun_dev_create)(); ++ int (*tun_dev_attach_queue)(void *); ++ int (*tun_dev_attach)(struct sock *, void *); ++ int (*tun_dev_open)(void *); ++ int (*sctp_assoc_request)(struct sctp_association *, struct sk_buff *); ++ int (*sctp_bind_connect)(struct sock *, int, struct sockaddr *, int); ++ void (*sctp_sk_clone)(struct sctp_association *, struct sock *, struct sock *); ++ int (*sctp_assoc_established)(struct sctp_association *, struct sk_buff *); ++ int (*ib_pkey_access)(void *, u64, u16); ++ int (*ib_endport_manage_subnet)(void *, const char *, u8); ++ int (*ib_alloc_security)(void **); ++ void (*ib_free_security)(void *); ++ int (*xfrm_policy_alloc_security)(struct xfrm_sec_ctx **, struct xfrm_user_sec_ctx *, gfp_t); ++ int (*xfrm_policy_clone_security)(struct xfrm_sec_ctx *, struct xfrm_sec_ctx **); ++ void (*xfrm_policy_free_security)(struct xfrm_sec_ctx *); ++ int (*xfrm_policy_delete_security)(struct xfrm_sec_ctx *); ++ int (*xfrm_state_alloc)(struct xfrm_state *, struct xfrm_user_sec_ctx *); ++ int (*xfrm_state_alloc_acquire)(struct xfrm_state *, struct xfrm_sec_ctx *, u32); ++ void (*xfrm_state_free_security)(struct xfrm_state *); ++ int (*xfrm_state_delete_security)(struct xfrm_state *); ++ int (*xfrm_policy_lookup)(struct xfrm_sec_ctx *, u32); ++ int (*xfrm_state_pol_flow_match)(struct xfrm_state *, struct xfrm_policy *, const struct flowi_common *); ++ int (*xfrm_decode_session)(struct sk_buff *, u32 *, int); ++ int (*key_alloc)(struct key *, const struct cred *, long unsigned int); ++ void (*key_free)(struct key *); ++ int (*key_permission)(key_ref_t, const struct cred *, enum key_need_perm); ++ int (*key_getsecurity)(struct key *, char **); ++ int (*audit_rule_init)(u32, u32, char *, void **); ++ int (*audit_rule_known)(struct audit_krule *); ++ int (*audit_rule_match)(u32, u32, u32, void *); ++ void (*audit_rule_free)(void *); ++ int (*bpf)(int, union bpf_attr *, unsigned int); ++ int (*bpf_map)(struct bpf_map *, fmode_t); ++ int (*bpf_prog)(struct bpf_prog *); ++ int (*bpf_map_alloc_security)(struct bpf_map *); ++ void (*bpf_map_free_security)(struct bpf_map *); ++ int (*bpf_prog_alloc_security)(struct bpf_prog_aux *); ++ void (*bpf_prog_free_security)(struct bpf_prog_aux *); ++ int (*locked_down)(enum lockdown_reason); ++ int (*lock_kernel_down)(const char *, enum lockdown_reason); ++ int (*perf_event_open)(struct perf_event_attr *, int); ++ int (*perf_event_alloc)(struct perf_event *); ++ void (*perf_event_free)(struct perf_event *); ++ int (*perf_event_read)(struct perf_event *); ++ int (*perf_event_write)(struct perf_event *); ++ int (*uring_override_creds)(const struct cred *); ++ int (*uring_sqpoll)(); ++ int (*uring_cmd)(struct io_uring_cmd *); ++}; ++ ++struct security_hook_heads { ++ struct hlist_head binder_set_context_mgr; ++ struct hlist_head binder_transaction; ++ struct hlist_head binder_transfer_binder; ++ struct hlist_head binder_transfer_file; ++ struct hlist_head ptrace_access_check; ++ struct hlist_head ptrace_traceme; ++ struct hlist_head capget; ++ struct hlist_head capset; ++ struct hlist_head capable; ++ struct hlist_head quotactl; ++ struct hlist_head quota_on; ++ struct hlist_head syslog; ++ struct hlist_head settime; ++ struct hlist_head vm_enough_memory; ++ struct hlist_head bprm_creds_for_exec; ++ struct hlist_head bprm_creds_from_file; ++ struct hlist_head bprm_check_security; ++ struct hlist_head bprm_committing_creds; ++ struct hlist_head bprm_committed_creds; ++ struct hlist_head fs_context_dup; ++ struct hlist_head fs_context_parse_param; ++ struct hlist_head sb_alloc_security; ++ struct hlist_head sb_delete; ++ struct hlist_head sb_free_security; ++ struct hlist_head sb_free_mnt_opts; ++ struct hlist_head sb_eat_lsm_opts; ++ struct hlist_head sb_mnt_opts_compat; ++ struct hlist_head sb_remount; ++ struct hlist_head sb_kern_mount; ++ struct hlist_head sb_show_options; ++ struct hlist_head sb_statfs; ++ struct hlist_head sb_mount; ++ struct hlist_head sb_umount; ++ struct hlist_head sb_pivotroot; ++ struct hlist_head sb_set_mnt_opts; ++ struct hlist_head sb_clone_mnt_opts; ++ struct hlist_head move_mount; ++ struct hlist_head dentry_init_security; ++ struct hlist_head dentry_create_files_as; ++ struct hlist_head path_unlink; ++ struct hlist_head path_mkdir; ++ struct hlist_head path_rmdir; ++ struct hlist_head path_mknod; ++ struct hlist_head path_truncate; ++ struct hlist_head path_symlink; ++ struct hlist_head path_link; ++ struct hlist_head path_rename; ++ struct hlist_head path_chmod; ++ struct hlist_head path_chown; ++ struct hlist_head path_chroot; ++ struct hlist_head path_notify; ++ struct hlist_head inode_alloc_security; ++ struct hlist_head inode_free_security; ++ struct hlist_head inode_init_security; ++ struct hlist_head inode_init_security_anon; ++ struct hlist_head inode_create; ++ struct hlist_head inode_link; ++ struct hlist_head inode_unlink; ++ struct hlist_head inode_symlink; ++ struct hlist_head inode_mkdir; ++ struct hlist_head inode_rmdir; ++ struct hlist_head inode_mknod; ++ struct hlist_head inode_rename; ++ struct hlist_head inode_readlink; ++ struct hlist_head inode_follow_link; ++ struct hlist_head inode_permission; ++ struct hlist_head inode_setattr; ++ struct hlist_head inode_getattr; ++ struct hlist_head inode_setxattr; ++ struct hlist_head inode_post_setxattr; ++ struct hlist_head inode_getxattr; ++ struct hlist_head inode_listxattr; ++ struct hlist_head inode_removexattr; ++ struct hlist_head inode_set_acl; ++ struct hlist_head inode_get_acl; ++ struct hlist_head inode_remove_acl; ++ struct hlist_head inode_need_killpriv; ++ struct hlist_head inode_killpriv; ++ struct hlist_head inode_getsecurity; ++ struct hlist_head inode_setsecurity; ++ struct hlist_head inode_listsecurity; ++ struct hlist_head inode_getsecid; ++ struct hlist_head inode_copy_up; ++ struct hlist_head inode_copy_up_xattr; ++ struct hlist_head kernfs_init_security; ++ struct hlist_head file_permission; ++ struct hlist_head file_alloc_security; ++ struct hlist_head file_free_security; ++ struct hlist_head file_ioctl; ++ struct hlist_head mmap_addr; ++ struct hlist_head mmap_file; ++ struct hlist_head file_mprotect; ++ struct hlist_head file_lock; ++ struct hlist_head file_fcntl; ++ struct hlist_head file_set_fowner; ++ struct hlist_head file_send_sigiotask; ++ struct hlist_head file_receive; ++ struct hlist_head file_open; ++ struct hlist_head file_truncate; ++ struct hlist_head task_alloc; ++ struct hlist_head task_free; ++ struct hlist_head cred_alloc_blank; ++ struct hlist_head cred_free; ++ struct hlist_head cred_prepare; ++ struct hlist_head cred_transfer; ++ struct hlist_head cred_getsecid; ++ struct hlist_head kernel_act_as; ++ struct hlist_head kernel_create_files_as; ++ struct hlist_head kernel_module_request; ++ struct hlist_head kernel_load_data; ++ struct hlist_head kernel_post_load_data; ++ struct hlist_head kernel_read_file; ++ struct hlist_head kernel_post_read_file; ++ struct hlist_head task_fix_setuid; ++ struct hlist_head task_fix_setgid; ++ struct hlist_head task_fix_setgroups; ++ struct hlist_head task_setpgid; ++ struct hlist_head task_getpgid; ++ struct hlist_head task_getsid; ++ struct hlist_head current_getsecid_subj; ++ struct hlist_head task_getsecid_obj; ++ struct hlist_head task_setnice; ++ struct hlist_head task_setioprio; ++ struct hlist_head task_getioprio; ++ struct hlist_head task_prlimit; ++ struct hlist_head task_setrlimit; ++ struct hlist_head task_setscheduler; ++ struct hlist_head task_getscheduler; ++ struct hlist_head task_movememory; ++ struct hlist_head task_kill; ++ struct hlist_head task_prctl; ++ struct hlist_head task_to_inode; ++ struct hlist_head userns_create; ++ struct hlist_head ipc_permission; ++ struct hlist_head ipc_getsecid; ++ struct hlist_head msg_msg_alloc_security; ++ struct hlist_head msg_msg_free_security; ++ struct hlist_head msg_queue_alloc_security; ++ struct hlist_head msg_queue_free_security; ++ struct hlist_head msg_queue_associate; ++ struct hlist_head msg_queue_msgctl; ++ struct hlist_head msg_queue_msgsnd; ++ struct hlist_head msg_queue_msgrcv; ++ struct hlist_head shm_alloc_security; ++ struct hlist_head shm_free_security; ++ struct hlist_head shm_associate; ++ struct hlist_head shm_shmctl; ++ struct hlist_head shm_shmat; ++ struct hlist_head sem_alloc_security; ++ struct hlist_head sem_free_security; ++ struct hlist_head sem_associate; ++ struct hlist_head sem_semctl; ++ struct hlist_head sem_semop; ++ struct hlist_head netlink_send; ++ struct hlist_head d_instantiate; ++ struct hlist_head getprocattr; ++ struct hlist_head setprocattr; ++ struct hlist_head ismaclabel; ++ struct hlist_head secid_to_secctx; ++ struct hlist_head secctx_to_secid; ++ struct hlist_head release_secctx; ++ struct hlist_head inode_invalidate_secctx; ++ struct hlist_head inode_notifysecctx; ++ struct hlist_head inode_setsecctx; ++ struct hlist_head inode_getsecctx; ++ struct hlist_head post_notification; ++ struct hlist_head watch_key; ++ struct hlist_head unix_stream_connect; ++ struct hlist_head unix_may_send; ++ struct hlist_head socket_create; ++ struct hlist_head socket_post_create; ++ struct hlist_head socket_socketpair; ++ struct hlist_head socket_bind; ++ struct hlist_head socket_connect; ++ struct hlist_head socket_listen; ++ struct hlist_head socket_accept; ++ struct hlist_head socket_sendmsg; ++ struct hlist_head socket_recvmsg; ++ struct hlist_head socket_getsockname; ++ struct hlist_head socket_getpeername; ++ struct hlist_head socket_getsockopt; ++ struct hlist_head socket_setsockopt; ++ struct hlist_head socket_shutdown; ++ struct hlist_head socket_sock_rcv_skb; ++ struct hlist_head socket_getpeersec_stream; ++ struct hlist_head socket_getpeersec_dgram; ++ struct hlist_head sk_alloc_security; ++ struct hlist_head sk_free_security; ++ struct hlist_head sk_clone_security; ++ struct hlist_head sk_getsecid; ++ struct hlist_head sock_graft; ++ struct hlist_head inet_conn_request; ++ struct hlist_head inet_csk_clone; ++ struct hlist_head inet_conn_established; ++ struct hlist_head secmark_relabel_packet; ++ struct hlist_head secmark_refcount_inc; ++ struct hlist_head secmark_refcount_dec; ++ struct hlist_head req_classify_flow; ++ struct hlist_head tun_dev_alloc_security; ++ struct hlist_head tun_dev_free_security; ++ struct hlist_head tun_dev_create; ++ struct hlist_head tun_dev_attach_queue; ++ struct hlist_head tun_dev_attach; ++ struct hlist_head tun_dev_open; ++ struct hlist_head sctp_assoc_request; ++ struct hlist_head sctp_bind_connect; ++ struct hlist_head sctp_sk_clone; ++ struct hlist_head sctp_assoc_established; ++ struct hlist_head ib_pkey_access; ++ struct hlist_head ib_endport_manage_subnet; ++ struct hlist_head ib_alloc_security; ++ struct hlist_head ib_free_security; ++ struct hlist_head xfrm_policy_alloc_security; ++ struct hlist_head xfrm_policy_clone_security; ++ struct hlist_head xfrm_policy_free_security; ++ struct hlist_head xfrm_policy_delete_security; ++ struct hlist_head xfrm_state_alloc; ++ struct hlist_head xfrm_state_alloc_acquire; ++ struct hlist_head xfrm_state_free_security; ++ struct hlist_head xfrm_state_delete_security; ++ struct hlist_head xfrm_policy_lookup; ++ struct hlist_head xfrm_state_pol_flow_match; ++ struct hlist_head xfrm_decode_session; ++ struct hlist_head key_alloc; ++ struct hlist_head key_free; ++ struct hlist_head key_permission; ++ struct hlist_head key_getsecurity; ++ struct hlist_head audit_rule_init; ++ struct hlist_head audit_rule_known; ++ struct hlist_head audit_rule_match; ++ struct hlist_head audit_rule_free; ++ struct hlist_head bpf; ++ struct hlist_head bpf_map; ++ struct hlist_head bpf_prog; ++ struct hlist_head bpf_map_alloc_security; ++ struct hlist_head bpf_map_free_security; ++ struct hlist_head bpf_prog_alloc_security; ++ struct hlist_head bpf_prog_free_security; ++ struct hlist_head locked_down; ++ struct hlist_head lock_kernel_down; ++ struct hlist_head perf_event_open; ++ struct hlist_head perf_event_alloc; ++ struct hlist_head perf_event_free; ++ struct hlist_head perf_event_read; ++ struct hlist_head perf_event_write; ++ struct hlist_head uring_override_creds; ++ struct hlist_head uring_sqpoll; ++ struct hlist_head uring_cmd; ++}; ++ ++struct security_hook_list { ++ struct hlist_node list; ++ struct hlist_head *head; ++ union security_list_options hook; ++ const char *lsm; ++}; ++ ++struct lsm_blob_sizes { ++ int lbs_cred; ++ int lbs_file; ++ int lbs_inode; ++ int lbs_superblock; ++ int lbs_ipc; ++ int lbs_msg_msg; ++ int lbs_task; ++}; ++ ++enum lsm_order { ++ LSM_ORDER_FIRST = -1, ++ LSM_ORDER_MUTABLE = 0, ++ LSM_ORDER_LAST = 1, ++}; ++ ++struct lsm_info { ++ const char *name; ++ enum lsm_order order; ++ long unsigned int flags; ++ int *enabled; ++ int (*init)(); ++ struct lsm_blob_sizes *blobs; ++}; ++ ++struct tree_descr { ++ const char *name; ++ const struct file_operations *ops; ++ int mode; ++}; ++ ++struct ethhdr { ++ unsigned char h_dest[6]; ++ unsigned char h_source[6]; ++ __be16 h_proto; ++}; ++ ++struct ethtool_drvinfo { ++ __u32 cmd; ++ char driver[32]; ++ char version[32]; ++ char fw_version[32]; ++ char bus_info[32]; ++ char erom_version[32]; ++ char reserved2[12]; ++ __u32 n_priv_flags; ++ __u32 n_stats; ++ __u32 testinfo_len; ++ __u32 eedump_len; ++ __u32 regdump_len; ++}; ++ ++struct ethtool_wolinfo { ++ __u32 cmd; ++ __u32 supported; ++ __u32 wolopts; ++ __u8 sopass[6]; ++}; ++ ++struct ethtool_tunable { ++ __u32 cmd; ++ __u32 id; ++ __u32 type_id; ++ __u32 len; ++ void *data[0]; ++}; ++ ++struct ethtool_regs { ++ __u32 cmd; ++ __u32 version; ++ __u32 len; ++ __u8 data[0]; ++}; ++ ++struct ethtool_eeprom { ++ __u32 cmd; ++ __u32 magic; ++ __u32 offset; ++ __u32 len; ++ __u8 data[0]; ++}; ++ ++struct ethtool_eee { ++ __u32 cmd; ++ __u32 supported; ++ __u32 advertised; ++ __u32 lp_advertised; ++ __u32 eee_active; ++ __u32 eee_enabled; ++ __u32 tx_lpi_enabled; ++ __u32 tx_lpi_timer; ++ __u32 reserved[2]; ++}; ++ ++struct ethtool_modinfo { ++ __u32 cmd; ++ __u32 type; ++ __u32 eeprom_len; ++ __u32 reserved[8]; ++}; ++ ++struct ethtool_coalesce { ++ __u32 cmd; ++ __u32 rx_coalesce_usecs; ++ __u32 rx_max_coalesced_frames; ++ __u32 rx_coalesce_usecs_irq; ++ __u32 rx_max_coalesced_frames_irq; ++ __u32 tx_coalesce_usecs; ++ __u32 tx_max_coalesced_frames; ++ __u32 tx_coalesce_usecs_irq; ++ __u32 tx_max_coalesced_frames_irq; ++ __u32 stats_block_coalesce_usecs; ++ __u32 use_adaptive_rx_coalesce; ++ __u32 use_adaptive_tx_coalesce; ++ __u32 pkt_rate_low; ++ __u32 rx_coalesce_usecs_low; ++ __u32 rx_max_coalesced_frames_low; ++ __u32 tx_coalesce_usecs_low; ++ __u32 tx_max_coalesced_frames_low; ++ __u32 pkt_rate_high; ++ __u32 rx_coalesce_usecs_high; ++ __u32 rx_max_coalesced_frames_high; ++ __u32 tx_coalesce_usecs_high; ++ __u32 tx_max_coalesced_frames_high; ++ __u32 rate_sample_interval; ++}; ++ ++struct ethtool_ringparam { ++ __u32 cmd; ++ __u32 rx_max_pending; ++ __u32 rx_mini_max_pending; ++ __u32 rx_jumbo_max_pending; ++ __u32 tx_max_pending; ++ __u32 rx_pending; ++ __u32 rx_mini_pending; ++ __u32 rx_jumbo_pending; ++ __u32 tx_pending; ++}; ++ ++struct ethtool_channels { ++ __u32 cmd; ++ __u32 max_rx; ++ __u32 max_tx; ++ __u32 max_other; ++ __u32 max_combined; ++ __u32 rx_count; ++ __u32 tx_count; ++ __u32 other_count; ++ __u32 combined_count; ++}; ++ ++struct ethtool_pauseparam { ++ __u32 cmd; ++ __u32 autoneg; ++ __u32 rx_pause; ++ __u32 tx_pause; ++}; ++ ++enum ethtool_link_ext_state { ++ ETHTOOL_LINK_EXT_STATE_AUTONEG = 0, ++ ETHTOOL_LINK_EXT_STATE_LINK_TRAINING_FAILURE = 1, ++ ETHTOOL_LINK_EXT_STATE_LINK_LOGICAL_MISMATCH = 2, ++ ETHTOOL_LINK_EXT_STATE_BAD_SIGNAL_INTEGRITY = 3, ++ ETHTOOL_LINK_EXT_STATE_NO_CABLE = 4, ++ ETHTOOL_LINK_EXT_STATE_CABLE_ISSUE = 5, ++ ETHTOOL_LINK_EXT_STATE_EEPROM_ISSUE = 6, ++ ETHTOOL_LINK_EXT_STATE_CALIBRATION_FAILURE = 7, ++ ETHTOOL_LINK_EXT_STATE_POWER_BUDGET_EXCEEDED = 8, ++ ETHTOOL_LINK_EXT_STATE_OVERHEAT = 9, ++ ETHTOOL_LINK_EXT_STATE_MODULE = 10, ++}; ++ ++enum ethtool_link_ext_substate_autoneg { ++ ETHTOOL_LINK_EXT_SUBSTATE_AN_NO_PARTNER_DETECTED = 1, ++ ETHTOOL_LINK_EXT_SUBSTATE_AN_ACK_NOT_RECEIVED = 2, ++ ETHTOOL_LINK_EXT_SUBSTATE_AN_NEXT_PAGE_EXCHANGE_FAILED = 3, ++ ETHTOOL_LINK_EXT_SUBSTATE_AN_NO_PARTNER_DETECTED_FORCE_MODE = 4, ++ ETHTOOL_LINK_EXT_SUBSTATE_AN_FEC_MISMATCH_DURING_OVERRIDE = 5, ++ ETHTOOL_LINK_EXT_SUBSTATE_AN_NO_HCD = 6, ++}; ++ ++enum ethtool_link_ext_substate_link_training { ++ ETHTOOL_LINK_EXT_SUBSTATE_LT_KR_FRAME_LOCK_NOT_ACQUIRED = 1, ++ ETHTOOL_LINK_EXT_SUBSTATE_LT_KR_LINK_INHIBIT_TIMEOUT = 2, ++ ETHTOOL_LINK_EXT_SUBSTATE_LT_KR_LINK_PARTNER_DID_NOT_SET_RECEIVER_READY = 3, ++ ETHTOOL_LINK_EXT_SUBSTATE_LT_REMOTE_FAULT = 4, ++}; ++ ++enum ethtool_link_ext_substate_link_logical_mismatch { ++ ETHTOOL_LINK_EXT_SUBSTATE_LLM_PCS_DID_NOT_ACQUIRE_BLOCK_LOCK = 1, ++ ETHTOOL_LINK_EXT_SUBSTATE_LLM_PCS_DID_NOT_ACQUIRE_AM_LOCK = 2, ++ ETHTOOL_LINK_EXT_SUBSTATE_LLM_PCS_DID_NOT_GET_ALIGN_STATUS = 3, ++ ETHTOOL_LINK_EXT_SUBSTATE_LLM_FC_FEC_IS_NOT_LOCKED = 4, ++ ETHTOOL_LINK_EXT_SUBSTATE_LLM_RS_FEC_IS_NOT_LOCKED = 5, ++}; ++ ++enum ethtool_link_ext_substate_bad_signal_integrity { ++ ETHTOOL_LINK_EXT_SUBSTATE_BSI_LARGE_NUMBER_OF_PHYSICAL_ERRORS = 1, ++ ETHTOOL_LINK_EXT_SUBSTATE_BSI_UNSUPPORTED_RATE = 2, ++ ETHTOOL_LINK_EXT_SUBSTATE_BSI_SERDES_REFERENCE_CLOCK_LOST = 3, ++ ETHTOOL_LINK_EXT_SUBSTATE_BSI_SERDES_ALOS = 4, ++}; ++ ++enum ethtool_link_ext_substate_cable_issue { ++ ETHTOOL_LINK_EXT_SUBSTATE_CI_UNSUPPORTED_CABLE = 1, ++ ETHTOOL_LINK_EXT_SUBSTATE_CI_CABLE_TEST_FAILURE = 2, ++}; ++ ++enum ethtool_link_ext_substate_module { ++ ETHTOOL_LINK_EXT_SUBSTATE_MODULE_CMIS_NOT_READY = 1, ++}; ++ ++enum ethtool_mac_stats_src { ++ ETHTOOL_MAC_STATS_SRC_AGGREGATE = 0, ++ ETHTOOL_MAC_STATS_SRC_EMAC = 1, ++ ETHTOOL_MAC_STATS_SRC_PMAC = 2, ++}; ++ ++enum ethtool_module_power_mode_policy { ++ ETHTOOL_MODULE_POWER_MODE_POLICY_HIGH = 1, ++ ETHTOOL_MODULE_POWER_MODE_POLICY_AUTO = 2, ++}; ++ ++enum ethtool_module_power_mode { ++ ETHTOOL_MODULE_POWER_MODE_LOW = 1, ++ ETHTOOL_MODULE_POWER_MODE_HIGH = 2, ++}; ++ ++enum ethtool_mm_verify_status { ++ ETHTOOL_MM_VERIFY_STATUS_UNKNOWN = 0, ++ ETHTOOL_MM_VERIFY_STATUS_INITIAL = 1, ++ ETHTOOL_MM_VERIFY_STATUS_VERIFYING = 2, ++ ETHTOOL_MM_VERIFY_STATUS_SUCCEEDED = 3, ++ ETHTOOL_MM_VERIFY_STATUS_FAILED = 4, ++ ETHTOOL_MM_VERIFY_STATUS_DISABLED = 5, ++}; ++ ++struct ethtool_test { ++ __u32 cmd; ++ __u32 flags; ++ __u32 reserved; ++ __u32 len; ++ __u64 data[0]; ++}; ++ ++struct ethtool_stats { ++ __u32 cmd; ++ __u32 n_stats; ++ __u64 data[0]; ++}; ++ ++struct ethtool_tcpip4_spec { ++ __be32 ip4src; ++ __be32 ip4dst; ++ __be16 psrc; ++ __be16 pdst; ++ __u8 tos; ++}; ++ ++struct ethtool_ah_espip4_spec { ++ __be32 ip4src; ++ __be32 ip4dst; ++ __be32 spi; ++ __u8 tos; ++}; ++ ++struct ethtool_usrip4_spec { ++ __be32 ip4src; ++ __be32 ip4dst; ++ __be32 l4_4_bytes; ++ __u8 tos; ++ __u8 ip_ver; ++ __u8 proto; ++}; ++ ++struct ethtool_tcpip6_spec { ++ __be32 ip6src[4]; ++ __be32 ip6dst[4]; ++ __be16 psrc; ++ __be16 pdst; ++ __u8 tclass; ++}; ++ ++struct ethtool_ah_espip6_spec { ++ __be32 ip6src[4]; ++ __be32 ip6dst[4]; ++ __be32 spi; ++ __u8 tclass; ++}; ++ ++struct ethtool_usrip6_spec { ++ __be32 ip6src[4]; ++ __be32 ip6dst[4]; ++ __be32 l4_4_bytes; ++ __u8 tclass; ++ __u8 l4_proto; ++}; ++ ++union ethtool_flow_union { ++ struct ethtool_tcpip4_spec tcp_ip4_spec; ++ struct ethtool_tcpip4_spec udp_ip4_spec; ++ struct ethtool_tcpip4_spec sctp_ip4_spec; ++ struct ethtool_ah_espip4_spec ah_ip4_spec; ++ struct ethtool_ah_espip4_spec esp_ip4_spec; ++ struct ethtool_usrip4_spec usr_ip4_spec; ++ struct ethtool_tcpip6_spec tcp_ip6_spec; ++ struct ethtool_tcpip6_spec udp_ip6_spec; ++ struct ethtool_tcpip6_spec sctp_ip6_spec; ++ struct ethtool_ah_espip6_spec ah_ip6_spec; ++ struct ethtool_ah_espip6_spec esp_ip6_spec; ++ struct ethtool_usrip6_spec usr_ip6_spec; ++ struct ethhdr ether_spec; ++ __u8 hdata[52]; ++}; ++ ++struct ethtool_flow_ext { ++ __u8 padding[2]; ++ unsigned char h_dest[6]; ++ __be16 vlan_etype; ++ __be16 vlan_tci; ++ __be32 data[2]; ++}; ++ ++struct ethtool_rx_flow_spec { ++ __u32 flow_type; ++ union ethtool_flow_union h_u; ++ struct ethtool_flow_ext h_ext; ++ union ethtool_flow_union m_u; ++ struct ethtool_flow_ext m_ext; ++ __u64 ring_cookie; ++ __u32 location; ++}; ++ ++struct ethtool_rxnfc { ++ __u32 cmd; ++ __u32 flow_type; ++ __u64 data; ++ struct ethtool_rx_flow_spec fs; ++ union { ++ __u32 rule_cnt; ++ __u32 rss_context; ++ }; ++ __u32 rule_locs[0]; ++}; ++ ++struct ethtool_flash { ++ __u32 cmd; ++ __u32 region; ++ char data[128]; ++}; ++ ++struct ethtool_dump { ++ __u32 cmd; ++ __u32 version; ++ __u32 flag; ++ __u32 len; ++ __u8 data[0]; ++}; ++ ++struct ethtool_ts_info { ++ __u32 cmd; ++ __u32 so_timestamping; ++ __s32 phc_index; ++ __u32 tx_types; ++ __u32 tx_reserved[3]; ++ __u32 rx_filters; ++ __u32 rx_reserved[3]; ++}; ++ ++struct ethtool_fecparam { ++ __u32 cmd; ++ __u32 active_fec; ++ __u32 fec; ++ __u32 reserved; ++}; ++ ++enum ethtool_link_mode_bit_indices { ++ ETHTOOL_LINK_MODE_10baseT_Half_BIT = 0, ++ ETHTOOL_LINK_MODE_10baseT_Full_BIT = 1, ++ ETHTOOL_LINK_MODE_100baseT_Half_BIT = 2, ++ ETHTOOL_LINK_MODE_100baseT_Full_BIT = 3, ++ ETHTOOL_LINK_MODE_1000baseT_Half_BIT = 4, ++ ETHTOOL_LINK_MODE_1000baseT_Full_BIT = 5, ++ ETHTOOL_LINK_MODE_Autoneg_BIT = 6, ++ ETHTOOL_LINK_MODE_TP_BIT = 7, ++ ETHTOOL_LINK_MODE_AUI_BIT = 8, ++ ETHTOOL_LINK_MODE_MII_BIT = 9, ++ ETHTOOL_LINK_MODE_FIBRE_BIT = 10, ++ ETHTOOL_LINK_MODE_BNC_BIT = 11, ++ ETHTOOL_LINK_MODE_10000baseT_Full_BIT = 12, ++ ETHTOOL_LINK_MODE_Pause_BIT = 13, ++ ETHTOOL_LINK_MODE_Asym_Pause_BIT = 14, ++ ETHTOOL_LINK_MODE_2500baseX_Full_BIT = 15, ++ ETHTOOL_LINK_MODE_Backplane_BIT = 16, ++ ETHTOOL_LINK_MODE_1000baseKX_Full_BIT = 17, ++ ETHTOOL_LINK_MODE_10000baseKX4_Full_BIT = 18, ++ ETHTOOL_LINK_MODE_10000baseKR_Full_BIT = 19, ++ ETHTOOL_LINK_MODE_10000baseR_FEC_BIT = 20, ++ ETHTOOL_LINK_MODE_20000baseMLD2_Full_BIT = 21, ++ ETHTOOL_LINK_MODE_20000baseKR2_Full_BIT = 22, ++ ETHTOOL_LINK_MODE_40000baseKR4_Full_BIT = 23, ++ ETHTOOL_LINK_MODE_40000baseCR4_Full_BIT = 24, ++ ETHTOOL_LINK_MODE_40000baseSR4_Full_BIT = 25, ++ ETHTOOL_LINK_MODE_40000baseLR4_Full_BIT = 26, ++ ETHTOOL_LINK_MODE_56000baseKR4_Full_BIT = 27, ++ ETHTOOL_LINK_MODE_56000baseCR4_Full_BIT = 28, ++ ETHTOOL_LINK_MODE_56000baseSR4_Full_BIT = 29, ++ ETHTOOL_LINK_MODE_56000baseLR4_Full_BIT = 30, ++ ETHTOOL_LINK_MODE_25000baseCR_Full_BIT = 31, ++ ETHTOOL_LINK_MODE_25000baseKR_Full_BIT = 32, ++ ETHTOOL_LINK_MODE_25000baseSR_Full_BIT = 33, ++ ETHTOOL_LINK_MODE_50000baseCR2_Full_BIT = 34, ++ ETHTOOL_LINK_MODE_50000baseKR2_Full_BIT = 35, ++ ETHTOOL_LINK_MODE_100000baseKR4_Full_BIT = 36, ++ ETHTOOL_LINK_MODE_100000baseSR4_Full_BIT = 37, ++ ETHTOOL_LINK_MODE_100000baseCR4_Full_BIT = 38, ++ ETHTOOL_LINK_MODE_100000baseLR4_ER4_Full_BIT = 39, ++ ETHTOOL_LINK_MODE_50000baseSR2_Full_BIT = 40, ++ ETHTOOL_LINK_MODE_1000baseX_Full_BIT = 41, ++ ETHTOOL_LINK_MODE_10000baseCR_Full_BIT = 42, ++ ETHTOOL_LINK_MODE_10000baseSR_Full_BIT = 43, ++ ETHTOOL_LINK_MODE_10000baseLR_Full_BIT = 44, ++ ETHTOOL_LINK_MODE_10000baseLRM_Full_BIT = 45, ++ ETHTOOL_LINK_MODE_10000baseER_Full_BIT = 46, ++ ETHTOOL_LINK_MODE_2500baseT_Full_BIT = 47, ++ ETHTOOL_LINK_MODE_5000baseT_Full_BIT = 48, ++ ETHTOOL_LINK_MODE_FEC_NONE_BIT = 49, ++ ETHTOOL_LINK_MODE_FEC_RS_BIT = 50, ++ ETHTOOL_LINK_MODE_FEC_BASER_BIT = 51, ++ ETHTOOL_LINK_MODE_50000baseKR_Full_BIT = 52, ++ ETHTOOL_LINK_MODE_50000baseSR_Full_BIT = 53, ++ ETHTOOL_LINK_MODE_50000baseCR_Full_BIT = 54, ++ ETHTOOL_LINK_MODE_50000baseLR_ER_FR_Full_BIT = 55, ++ ETHTOOL_LINK_MODE_50000baseDR_Full_BIT = 56, ++ ETHTOOL_LINK_MODE_100000baseKR2_Full_BIT = 57, ++ ETHTOOL_LINK_MODE_100000baseSR2_Full_BIT = 58, ++ ETHTOOL_LINK_MODE_100000baseCR2_Full_BIT = 59, ++ ETHTOOL_LINK_MODE_100000baseLR2_ER2_FR2_Full_BIT = 60, ++ ETHTOOL_LINK_MODE_100000baseDR2_Full_BIT = 61, ++ ETHTOOL_LINK_MODE_200000baseKR4_Full_BIT = 62, ++ ETHTOOL_LINK_MODE_200000baseSR4_Full_BIT = 63, ++ ETHTOOL_LINK_MODE_200000baseLR4_ER4_FR4_Full_BIT = 64, ++ ETHTOOL_LINK_MODE_200000baseDR4_Full_BIT = 65, ++ ETHTOOL_LINK_MODE_200000baseCR4_Full_BIT = 66, ++ ETHTOOL_LINK_MODE_100baseT1_Full_BIT = 67, ++ ETHTOOL_LINK_MODE_1000baseT1_Full_BIT = 68, ++ ETHTOOL_LINK_MODE_400000baseKR8_Full_BIT = 69, ++ ETHTOOL_LINK_MODE_400000baseSR8_Full_BIT = 70, ++ ETHTOOL_LINK_MODE_400000baseLR8_ER8_FR8_Full_BIT = 71, ++ ETHTOOL_LINK_MODE_400000baseDR8_Full_BIT = 72, ++ ETHTOOL_LINK_MODE_400000baseCR8_Full_BIT = 73, ++ ETHTOOL_LINK_MODE_FEC_LLRS_BIT = 74, ++ ETHTOOL_LINK_MODE_100000baseKR_Full_BIT = 75, ++ ETHTOOL_LINK_MODE_100000baseSR_Full_BIT = 76, ++ ETHTOOL_LINK_MODE_100000baseLR_ER_FR_Full_BIT = 77, ++ ETHTOOL_LINK_MODE_100000baseCR_Full_BIT = 78, ++ ETHTOOL_LINK_MODE_100000baseDR_Full_BIT = 79, ++ ETHTOOL_LINK_MODE_200000baseKR2_Full_BIT = 80, ++ ETHTOOL_LINK_MODE_200000baseSR2_Full_BIT = 81, ++ ETHTOOL_LINK_MODE_200000baseLR2_ER2_FR2_Full_BIT = 82, ++ ETHTOOL_LINK_MODE_200000baseDR2_Full_BIT = 83, ++ ETHTOOL_LINK_MODE_200000baseCR2_Full_BIT = 84, ++ ETHTOOL_LINK_MODE_400000baseKR4_Full_BIT = 85, ++ ETHTOOL_LINK_MODE_400000baseSR4_Full_BIT = 86, ++ ETHTOOL_LINK_MODE_400000baseLR4_ER4_FR4_Full_BIT = 87, ++ ETHTOOL_LINK_MODE_400000baseDR4_Full_BIT = 88, ++ ETHTOOL_LINK_MODE_400000baseCR4_Full_BIT = 89, ++ ETHTOOL_LINK_MODE_100baseFX_Half_BIT = 90, ++ ETHTOOL_LINK_MODE_100baseFX_Full_BIT = 91, ++ ETHTOOL_LINK_MODE_10baseT1L_Full_BIT = 92, ++ ETHTOOL_LINK_MODE_800000baseCR8_Full_BIT = 93, ++ ETHTOOL_LINK_MODE_800000baseKR8_Full_BIT = 94, ++ ETHTOOL_LINK_MODE_800000baseDR8_Full_BIT = 95, ++ ETHTOOL_LINK_MODE_800000baseDR8_2_Full_BIT = 96, ++ ETHTOOL_LINK_MODE_800000baseSR8_Full_BIT = 97, ++ ETHTOOL_LINK_MODE_800000baseVR8_Full_BIT = 98, ++ ETHTOOL_LINK_MODE_10baseT1S_Full_BIT = 99, ++ ETHTOOL_LINK_MODE_10baseT1S_Half_BIT = 100, ++ ETHTOOL_LINK_MODE_10baseT1S_P2MP_Half_BIT = 101, ++ __ETHTOOL_LINK_MODE_MASK_NBITS = 102, ++}; ++ ++struct ethtool_link_settings { ++ __u32 cmd; ++ __u32 speed; ++ __u8 duplex; ++ __u8 port; ++ __u8 phy_address; ++ __u8 autoneg; ++ __u8 mdio_support; ++ __u8 eth_tp_mdix; ++ __u8 eth_tp_mdix_ctrl; ++ __s8 link_mode_masks_nwords; ++ __u8 transceiver; ++ __u8 master_slave_cfg; ++ __u8 master_slave_state; ++ __u8 rate_matching; ++ __u32 reserved[7]; ++ __u32 link_mode_masks[0]; ++}; ++ ++struct kernel_ethtool_ringparam { ++ u32 rx_buf_len; ++ u8 tcp_data_split; ++ u8 tx_push; ++ u8 rx_push; ++ u32 cqe_size; ++ u32 tx_push_buf_len; ++ u32 tx_push_buf_max_len; ++}; ++ ++struct ethtool_link_ext_state_info { ++ enum ethtool_link_ext_state link_ext_state; ++ union { ++ enum ethtool_link_ext_substate_autoneg autoneg; ++ enum ethtool_link_ext_substate_link_training link_training; ++ enum ethtool_link_ext_substate_link_logical_mismatch link_logical_mismatch; ++ enum ethtool_link_ext_substate_bad_signal_integrity bad_signal_integrity; ++ enum ethtool_link_ext_substate_cable_issue cable_issue; ++ enum ethtool_link_ext_substate_module module; ++ u32 __link_ext_substate; ++ }; ++}; ++ ++struct ethtool_link_ext_stats { ++ u64 link_down_events; ++}; ++ ++struct ethtool_link_ksettings { ++ struct ethtool_link_settings base; ++ struct { ++ long unsigned int supported[2]; ++ long unsigned int advertising[2]; ++ long unsigned int lp_advertising[2]; ++ } link_modes; ++ u32 lanes; ++}; ++ ++struct kernel_ethtool_coalesce { ++ u8 use_cqe_mode_tx; ++ u8 use_cqe_mode_rx; ++ u32 tx_aggr_max_bytes; ++ u32 tx_aggr_max_frames; ++ u32 tx_aggr_time_usecs; ++}; ++ ++struct ethtool_eth_mac_stats { ++ enum ethtool_mac_stats_src src; ++ union { ++ struct { ++ u64 FramesTransmittedOK; ++ u64 SingleCollisionFrames; ++ u64 MultipleCollisionFrames; ++ u64 FramesReceivedOK; ++ u64 FrameCheckSequenceErrors; ++ u64 AlignmentErrors; ++ u64 OctetsTransmittedOK; ++ u64 FramesWithDeferredXmissions; ++ u64 LateCollisions; ++ u64 FramesAbortedDueToXSColls; ++ u64 FramesLostDueToIntMACXmitError; ++ u64 CarrierSenseErrors; ++ u64 OctetsReceivedOK; ++ u64 FramesLostDueToIntMACRcvError; ++ u64 MulticastFramesXmittedOK; ++ u64 BroadcastFramesXmittedOK; ++ u64 FramesWithExcessiveDeferral; ++ u64 MulticastFramesReceivedOK; ++ u64 BroadcastFramesReceivedOK; ++ u64 InRangeLengthErrors; ++ u64 OutOfRangeLengthField; ++ u64 FrameTooLongErrors; ++ }; ++ struct { ++ u64 FramesTransmittedOK; ++ u64 SingleCollisionFrames; ++ u64 MultipleCollisionFrames; ++ u64 FramesReceivedOK; ++ u64 FrameCheckSequenceErrors; ++ u64 AlignmentErrors; ++ u64 OctetsTransmittedOK; ++ u64 FramesWithDeferredXmissions; ++ u64 LateCollisions; ++ u64 FramesAbortedDueToXSColls; ++ u64 FramesLostDueToIntMACXmitError; ++ u64 CarrierSenseErrors; ++ u64 OctetsReceivedOK; ++ u64 FramesLostDueToIntMACRcvError; ++ u64 MulticastFramesXmittedOK; ++ u64 BroadcastFramesXmittedOK; ++ u64 FramesWithExcessiveDeferral; ++ u64 MulticastFramesReceivedOK; ++ u64 BroadcastFramesReceivedOK; ++ u64 InRangeLengthErrors; ++ u64 OutOfRangeLengthField; ++ u64 FrameTooLongErrors; ++ } stats; ++ }; ++}; ++ ++struct ethtool_eth_phy_stats { ++ enum ethtool_mac_stats_src src; ++ union { ++ struct { ++ u64 SymbolErrorDuringCarrier; ++ }; ++ struct { ++ u64 SymbolErrorDuringCarrier; ++ } stats; ++ }; ++}; ++ ++struct ethtool_eth_ctrl_stats { ++ enum ethtool_mac_stats_src src; ++ union { ++ struct { ++ u64 MACControlFramesTransmitted; ++ u64 MACControlFramesReceived; ++ u64 UnsupportedOpcodesReceived; ++ }; ++ struct { ++ u64 MACControlFramesTransmitted; ++ u64 MACControlFramesReceived; ++ u64 UnsupportedOpcodesReceived; ++ } stats; ++ }; ++}; ++ ++struct ethtool_pause_stats { ++ enum ethtool_mac_stats_src src; ++ union { ++ struct { ++ u64 tx_pause_frames; ++ u64 rx_pause_frames; ++ }; ++ struct { ++ u64 tx_pause_frames; ++ u64 rx_pause_frames; ++ } stats; ++ }; ++}; ++ ++struct ethtool_fec_stat { ++ u64 total; ++ u64 lanes[8]; ++}; ++ ++struct ethtool_fec_stats { ++ struct ethtool_fec_stat corrected_blocks; ++ struct ethtool_fec_stat uncorrectable_blocks; ++ struct ethtool_fec_stat corrected_bits; ++}; ++ ++struct ethtool_rmon_hist_range { ++ u16 low; ++ u16 high; ++}; ++ ++struct ethtool_rmon_stats { ++ enum ethtool_mac_stats_src src; ++ union { ++ struct { ++ u64 undersize_pkts; ++ u64 oversize_pkts; ++ u64 fragments; ++ u64 jabbers; ++ u64 hist[10]; ++ u64 hist_tx[10]; ++ }; ++ struct { ++ u64 undersize_pkts; ++ u64 oversize_pkts; ++ u64 fragments; ++ u64 jabbers; ++ u64 hist[10]; ++ u64 hist_tx[10]; ++ } stats; ++ }; ++}; ++ ++struct ethtool_module_eeprom { ++ u32 offset; ++ u32 length; ++ u8 page; ++ u8 bank; ++ u8 i2c_address; ++ u8 *data; ++}; ++ ++struct ethtool_module_power_mode_params { ++ enum ethtool_module_power_mode_policy policy; ++ enum ethtool_module_power_mode mode; ++}; ++ ++struct ethtool_mm_state { ++ u32 verify_time; ++ u32 max_verify_time; ++ enum ethtool_mm_verify_status verify_status; ++ bool tx_enabled; ++ bool tx_active; ++ bool pmac_enabled; ++ bool verify_enabled; ++ u32 tx_min_frag_size; ++ u32 rx_min_frag_size; ++}; ++ ++struct ethtool_mm_cfg { ++ u32 verify_time; ++ bool verify_enabled; ++ bool tx_enabled; ++ bool pmac_enabled; ++ u32 tx_min_frag_size; ++}; ++ ++struct ethtool_mm_stats { ++ u64 MACMergeFrameAssErrorCount; ++ u64 MACMergeFrameSmdErrorCount; ++ u64 MACMergeFrameAssOkCount; ++ u64 MACMergeFragCountRx; ++ u64 MACMergeFragCountTx; ++ u64 MACMergeHoldCount; ++}; ++ ++enum ib_uverbs_write_cmds { ++ IB_USER_VERBS_CMD_GET_CONTEXT = 0, ++ IB_USER_VERBS_CMD_QUERY_DEVICE = 1, ++ IB_USER_VERBS_CMD_QUERY_PORT = 2, ++ IB_USER_VERBS_CMD_ALLOC_PD = 3, ++ IB_USER_VERBS_CMD_DEALLOC_PD = 4, ++ IB_USER_VERBS_CMD_CREATE_AH = 5, ++ IB_USER_VERBS_CMD_MODIFY_AH = 6, ++ IB_USER_VERBS_CMD_QUERY_AH = 7, ++ IB_USER_VERBS_CMD_DESTROY_AH = 8, ++ IB_USER_VERBS_CMD_REG_MR = 9, ++ IB_USER_VERBS_CMD_REG_SMR = 10, ++ IB_USER_VERBS_CMD_REREG_MR = 11, ++ IB_USER_VERBS_CMD_QUERY_MR = 12, ++ IB_USER_VERBS_CMD_DEREG_MR = 13, ++ IB_USER_VERBS_CMD_ALLOC_MW = 14, ++ IB_USER_VERBS_CMD_BIND_MW = 15, ++ IB_USER_VERBS_CMD_DEALLOC_MW = 16, ++ IB_USER_VERBS_CMD_CREATE_COMP_CHANNEL = 17, ++ IB_USER_VERBS_CMD_CREATE_CQ = 18, ++ IB_USER_VERBS_CMD_RESIZE_CQ = 19, ++ IB_USER_VERBS_CMD_DESTROY_CQ = 20, ++ IB_USER_VERBS_CMD_POLL_CQ = 21, ++ IB_USER_VERBS_CMD_PEEK_CQ = 22, ++ IB_USER_VERBS_CMD_REQ_NOTIFY_CQ = 23, ++ IB_USER_VERBS_CMD_CREATE_QP = 24, ++ IB_USER_VERBS_CMD_QUERY_QP = 25, ++ IB_USER_VERBS_CMD_MODIFY_QP = 26, ++ IB_USER_VERBS_CMD_DESTROY_QP = 27, ++ IB_USER_VERBS_CMD_POST_SEND = 28, ++ IB_USER_VERBS_CMD_POST_RECV = 29, ++ IB_USER_VERBS_CMD_ATTACH_MCAST = 30, ++ IB_USER_VERBS_CMD_DETACH_MCAST = 31, ++ IB_USER_VERBS_CMD_CREATE_SRQ = 32, ++ IB_USER_VERBS_CMD_MODIFY_SRQ = 33, ++ IB_USER_VERBS_CMD_QUERY_SRQ = 34, ++ IB_USER_VERBS_CMD_DESTROY_SRQ = 35, ++ IB_USER_VERBS_CMD_POST_SRQ_RECV = 36, ++ IB_USER_VERBS_CMD_OPEN_XRCD = 37, ++ IB_USER_VERBS_CMD_CLOSE_XRCD = 38, ++ IB_USER_VERBS_CMD_CREATE_XSRQ = 39, ++ IB_USER_VERBS_CMD_OPEN_QP = 40, ++}; ++ ++enum ib_uverbs_wc_opcode { ++ IB_UVERBS_WC_SEND = 0, ++ IB_UVERBS_WC_RDMA_WRITE = 1, ++ IB_UVERBS_WC_RDMA_READ = 2, ++ IB_UVERBS_WC_COMP_SWAP = 3, ++ IB_UVERBS_WC_FETCH_ADD = 4, ++ IB_UVERBS_WC_BIND_MW = 5, ++ IB_UVERBS_WC_LOCAL_INV = 6, ++ IB_UVERBS_WC_TSO = 7, ++ IB_UVERBS_WC_FLUSH = 8, ++ IB_UVERBS_WC_ATOMIC_WRITE = 9, ++}; ++ ++enum ib_uverbs_create_qp_mask { ++ IB_UVERBS_CREATE_QP_MASK_IND_TABLE = 1, ++}; ++ ++enum ib_uverbs_wr_opcode { ++ IB_UVERBS_WR_RDMA_WRITE = 0, ++ IB_UVERBS_WR_RDMA_WRITE_WITH_IMM = 1, ++ IB_UVERBS_WR_SEND = 2, ++ IB_UVERBS_WR_SEND_WITH_IMM = 3, ++ IB_UVERBS_WR_RDMA_READ = 4, ++ IB_UVERBS_WR_ATOMIC_CMP_AND_SWP = 5, ++ IB_UVERBS_WR_ATOMIC_FETCH_AND_ADD = 6, ++ IB_UVERBS_WR_LOCAL_INV = 7, ++ IB_UVERBS_WR_BIND_MW = 8, ++ IB_UVERBS_WR_SEND_WITH_INV = 9, ++ IB_UVERBS_WR_TSO = 10, ++ IB_UVERBS_WR_RDMA_READ_WITH_INV = 11, ++ IB_UVERBS_WR_MASKED_ATOMIC_CMP_AND_SWP = 12, ++ IB_UVERBS_WR_MASKED_ATOMIC_FETCH_AND_ADD = 13, ++ IB_UVERBS_WR_FLUSH = 14, ++ IB_UVERBS_WR_ATOMIC_WRITE = 15, ++}; ++ ++enum ib_uverbs_device_cap_flags { ++ IB_UVERBS_DEVICE_RESIZE_MAX_WR = 1ULL, ++ IB_UVERBS_DEVICE_BAD_PKEY_CNTR = 2ULL, ++ IB_UVERBS_DEVICE_BAD_QKEY_CNTR = 4ULL, ++ IB_UVERBS_DEVICE_RAW_MULTI = 8ULL, ++ IB_UVERBS_DEVICE_AUTO_PATH_MIG = 16ULL, ++ IB_UVERBS_DEVICE_CHANGE_PHY_PORT = 32ULL, ++ IB_UVERBS_DEVICE_UD_AV_PORT_ENFORCE = 64ULL, ++ IB_UVERBS_DEVICE_CURR_QP_STATE_MOD = 128ULL, ++ IB_UVERBS_DEVICE_SHUTDOWN_PORT = 256ULL, ++ IB_UVERBS_DEVICE_PORT_ACTIVE_EVENT = 1024ULL, ++ IB_UVERBS_DEVICE_SYS_IMAGE_GUID = 2048ULL, ++ IB_UVERBS_DEVICE_RC_RNR_NAK_GEN = 4096ULL, ++ IB_UVERBS_DEVICE_SRQ_RESIZE = 8192ULL, ++ IB_UVERBS_DEVICE_N_NOTIFY_CQ = 16384ULL, ++ IB_UVERBS_DEVICE_MEM_WINDOW = 131072ULL, ++ IB_UVERBS_DEVICE_UD_IP_CSUM = 262144ULL, ++ IB_UVERBS_DEVICE_XRC = 1048576ULL, ++ IB_UVERBS_DEVICE_MEM_MGT_EXTENSIONS = 2097152ULL, ++ IB_UVERBS_DEVICE_MEM_WINDOW_TYPE_2A = 8388608ULL, ++ IB_UVERBS_DEVICE_MEM_WINDOW_TYPE_2B = 16777216ULL, ++ IB_UVERBS_DEVICE_RC_IP_CSUM = 33554432ULL, ++ IB_UVERBS_DEVICE_RAW_IP_CSUM = 67108864ULL, ++ IB_UVERBS_DEVICE_MANAGED_FLOW_STEERING = 536870912ULL, ++ IB_UVERBS_DEVICE_RAW_SCATTER_FCS = 17179869184ULL, ++ IB_UVERBS_DEVICE_PCI_WRITE_END_PADDING = 68719476736ULL, ++ IB_UVERBS_DEVICE_FLUSH_GLOBAL = 274877906944ULL, ++ IB_UVERBS_DEVICE_FLUSH_PERSISTENT = 549755813888ULL, ++ IB_UVERBS_DEVICE_ATOMIC_WRITE = 1099511627776ULL, ++}; ++ ++enum ib_uverbs_raw_packet_caps { ++ IB_UVERBS_RAW_PACKET_CAP_CVLAN_STRIPPING = 1, ++ IB_UVERBS_RAW_PACKET_CAP_SCATTER_FCS = 2, ++ IB_UVERBS_RAW_PACKET_CAP_IP_CSUM = 4, ++ IB_UVERBS_RAW_PACKET_CAP_DELAY_DROP = 8, ++}; ++ ++enum ib_uverbs_access_flags { ++ IB_UVERBS_ACCESS_LOCAL_WRITE = 1, ++ IB_UVERBS_ACCESS_REMOTE_WRITE = 2, ++ IB_UVERBS_ACCESS_REMOTE_READ = 4, ++ IB_UVERBS_ACCESS_REMOTE_ATOMIC = 8, ++ IB_UVERBS_ACCESS_MW_BIND = 16, ++ IB_UVERBS_ACCESS_ZERO_BASED = 32, ++ IB_UVERBS_ACCESS_ON_DEMAND = 64, ++ IB_UVERBS_ACCESS_HUGETLB = 128, ++ IB_UVERBS_ACCESS_FLUSH_GLOBAL = 256, ++ IB_UVERBS_ACCESS_FLUSH_PERSISTENT = 512, ++ IB_UVERBS_ACCESS_RELAXED_ORDERING = 1048576, ++ IB_UVERBS_ACCESS_OPTIONAL_RANGE = 1072693248, ++}; ++ ++enum ib_uverbs_srq_type { ++ IB_UVERBS_SRQT_BASIC = 0, ++ IB_UVERBS_SRQT_XRC = 1, ++ IB_UVERBS_SRQT_TM = 2, ++}; ++ ++enum ib_uverbs_wq_type { ++ IB_UVERBS_WQT_RQ = 0, ++}; ++ ++enum ib_uverbs_wq_flags { ++ IB_UVERBS_WQ_FLAGS_CVLAN_STRIPPING = 1, ++ IB_UVERBS_WQ_FLAGS_SCATTER_FCS = 2, ++ IB_UVERBS_WQ_FLAGS_DELAY_DROP = 4, ++ IB_UVERBS_WQ_FLAGS_PCI_WRITE_END_PADDING = 8, ++}; ++ ++enum ib_uverbs_qp_type { ++ IB_UVERBS_QPT_RC = 2, ++ IB_UVERBS_QPT_UC = 3, ++ IB_UVERBS_QPT_UD = 4, ++ IB_UVERBS_QPT_RAW_PACKET = 8, ++ IB_UVERBS_QPT_XRC_INI = 9, ++ IB_UVERBS_QPT_XRC_TGT = 10, ++ IB_UVERBS_QPT_DRIVER = 255, ++}; ++ ++enum ib_uverbs_qp_create_flags { ++ IB_UVERBS_QP_CREATE_BLOCK_MULTICAST_LOOPBACK = 2, ++ IB_UVERBS_QP_CREATE_SCATTER_FCS = 256, ++ IB_UVERBS_QP_CREATE_CVLAN_STRIPPING = 512, ++ IB_UVERBS_QP_CREATE_PCI_WRITE_END_PADDING = 2048, ++ IB_UVERBS_QP_CREATE_SQ_SIG_ALL = 4096, ++}; ++ ++enum ib_uverbs_gid_type { ++ IB_UVERBS_GID_TYPE_IB = 0, ++ IB_UVERBS_GID_TYPE_ROCE_V1 = 1, ++ IB_UVERBS_GID_TYPE_ROCE_V2 = 2, ++}; ++ ++enum ib_poll_context { ++ IB_POLL_SOFTIRQ = 0, ++ IB_POLL_WORKQUEUE = 1, ++ IB_POLL_UNBOUND_WORKQUEUE = 2, ++ IB_POLL_LAST_POOL_TYPE = 2, ++ IB_POLL_DIRECT = 3, ++}; ++ ++struct lsm_network_audit { ++ int netif; ++ const struct sock *sk; ++ u16 family; ++ __be16 dport; ++ __be16 sport; ++ union { ++ struct { ++ __be32 daddr; ++ __be32 saddr; ++ } v4; ++ struct { ++ struct in6_addr daddr; ++ struct in6_addr saddr; ++ } v6; ++ } fam; ++}; ++ ++struct lsm_ioctlop_audit { ++ struct path path; ++ u16 cmd; ++}; ++ ++struct lsm_ibpkey_audit { ++ u64 subnet_prefix; ++ u16 pkey; ++}; ++ ++struct lsm_ibendport_audit { ++ const char *dev_name; ++ u8 port; ++}; ++ ++struct selinux_audit_data { ++ u32 ssid; ++ u32 tsid; ++ u16 tclass; ++ u32 requested; ++ u32 audited; ++ u32 denied; ++ int result; ++}; ++ ++struct common_audit_data { ++ char type; ++ union { ++ struct path path; ++ struct dentry *dentry; ++ struct inode *inode; ++ struct lsm_network_audit *net; ++ int cap; ++ int ipc_id; ++ struct task_struct *tsk; ++ struct { ++ key_serial_t key; ++ char *key_desc; ++ } key_struct; ++ char *kmod_name; ++ struct lsm_ioctlop_audit *op; ++ struct file *file; ++ struct lsm_ibpkey_audit *ibpkey; ++ struct lsm_ibendport_audit *ibendport; ++ int reason; ++ const char *anonclass; ++ } u; ++ union { ++ struct selinux_audit_data *selinux_audit_data; ++ }; ++}; ++ ++enum { ++ POLICYDB_CAP_NETPEER = 0, ++ POLICYDB_CAP_OPENPERM = 1, ++ POLICYDB_CAP_EXTSOCKCLASS = 2, ++ POLICYDB_CAP_ALWAYSNETWORK = 3, ++ POLICYDB_CAP_CGROUPSECLABEL = 4, ++ POLICYDB_CAP_NNP_NOSUID_TRANSITION = 5, ++ POLICYDB_CAP_GENFS_SECLABEL_SYMLINKS = 6, ++ POLICYDB_CAP_IOCTL_SKIP_CLOEXEC = 7, ++ __POLICYDB_CAP_MAX = 8, ++}; ++ ++struct selinux_policy; ++ ++struct selinux_state { ++ bool enforcing; ++ bool initialized; ++ bool policycap[8]; ++ struct page *status_page; ++ struct mutex status_lock; ++ struct selinux_policy *policy; ++ struct mutex policy_mutex; ++}; ++ ++struct selinux_policy_convert_data; ++ ++struct selinux_load_state { ++ struct selinux_policy *policy; ++ struct selinux_policy_convert_data *convert_data; ++}; ++ ++struct av_decision { ++ u32 allowed; ++ u32 auditallow; ++ u32 auditdeny; ++ u32 seqno; ++ u32 flags; ++}; ++ ++struct avc_cache_stats { ++ unsigned int lookups; ++ unsigned int misses; ++ unsigned int allocations; ++ unsigned int reclaims; ++ unsigned int frees; ++}; ++ ++struct task_security_struct { ++ u32 osid; ++ u32 sid; ++ u32 exec_sid; ++ u32 create_sid; ++ u32 keycreate_sid; ++ u32 sockcreate_sid; ++}; ++ ++enum label_initialized { ++ LABEL_INVALID = 0, ++ LABEL_INITIALIZED = 1, ++ LABEL_PENDING = 2, ++}; ++ ++struct inode_security_struct { ++ struct inode *inode; ++ struct list_head list; ++ u32 task_sid; ++ u32 sid; ++ u16 sclass; ++ unsigned char initialized; ++ spinlock_t lock; ++}; ++ ++enum sel_inos { ++ SEL_ROOT_INO = 2, ++ SEL_LOAD = 3, ++ SEL_ENFORCE = 4, ++ SEL_CONTEXT = 5, ++ SEL_ACCESS = 6, ++ SEL_CREATE = 7, ++ SEL_RELABEL = 8, ++ SEL_USER = 9, ++ SEL_POLICYVERS = 10, ++ SEL_COMMIT_BOOLS = 11, ++ SEL_MLS = 12, ++ SEL_DISABLE = 13, ++ SEL_MEMBER = 14, ++ SEL_CHECKREQPROT = 15, ++ SEL_COMPAT_NET = 16, ++ SEL_REJECT_UNKNOWN = 17, ++ SEL_DENY_UNKNOWN = 18, ++ SEL_STATUS = 19, ++ SEL_POLICY = 20, ++ SEL_VALIDATE_TRANS = 21, ++ SEL_INO_NEXT = 22, ++}; ++ ++struct selinux_fs_info { ++ struct dentry *bool_dir; ++ unsigned int bool_num; ++ char **bool_pending_names; ++ int *bool_pending_values; ++ struct dentry *class_dir; ++ long unsigned int last_class_ino; ++ bool policy_opened; ++ struct dentry *policycap_dir; ++ long unsigned int last_ino; ++ struct super_block *sb; ++}; ++ ++struct policy_load_memory { ++ size_t len; ++ void *data; ++}; ++ ++struct netnode_security_struct { ++ union { ++ __be32 ipv4; ++ struct in6_addr ipv6; ++ } addr; ++ u32 sid; ++ u16 family; ++}; ++ ++struct sel_netnode_bkt { ++ unsigned int size; ++ struct list_head list; ++}; ++ ++struct sel_netnode { ++ struct netnode_security_struct nsec; ++ struct list_head list; ++ struct callback_head rcu; ++}; ++ ++struct selinux_kernel_status { ++ u32 version; ++ u32 sequence; ++ u32 enforcing; ++ u32 policyload; ++ u32 deny_unknown; ++}; ++ ++typedef __u16 __sum16; ++ ++typedef u8 uint8_t; ++ ++struct sockaddr_in6 { ++ short unsigned int sin6_family; ++ __be16 sin6_port; ++ __be32 sin6_flowinfo; ++ struct in6_addr sin6_addr; ++ __u32 sin6_scope_id; ++}; ++ ++struct iphdr { ++ __u8 version: 4; ++ __u8 ihl: 4; ++ __u8 tos; ++ __be16 tot_len; ++ __be16 id; ++ __be16 frag_off; ++ __u8 ttl; ++ __u8 protocol; ++ __sum16 check; ++ union { ++ struct { ++ __be32 saddr; ++ __be32 daddr; ++ }; ++ struct { ++ __be32 saddr; ++ __be32 daddr; ++ } addrs; ++ }; ++}; ++ ++struct in6_pktinfo { ++ struct in6_addr ipi6_addr; ++ int ipi6_ifindex; ++}; ++ ++struct ipv6_rt_hdr { ++ __u8 nexthdr; ++ __u8 hdrlen; ++ __u8 type; ++ __u8 segments_left; ++}; ++ ++struct ipv6_opt_hdr { ++ __u8 nexthdr; ++ __u8 hdrlen; ++}; ++ ++struct ipv6hdr { ++ __u8 version: 4; ++ __u8 priority: 4; ++ __u8 flow_lbl[3]; ++ __be16 payload_len; ++ __u8 nexthdr; ++ __u8 hop_limit; ++ union { ++ struct { ++ struct in6_addr saddr; ++ struct in6_addr daddr; ++ }; ++ struct { ++ struct in6_addr saddr; ++ struct in6_addr daddr; ++ } addrs; ++ }; ++}; ++ ++struct in_addr { ++ __be32 s_addr; ++}; ++ ++struct sockaddr_in { ++ __kernel_sa_family_t sin_family; ++ __be16 sin_port; ++ struct in_addr sin_addr; ++ unsigned char __pad[8]; ++}; ++ ++struct ip_options { ++ __be32 faddr; ++ __be32 nexthop; ++ unsigned char optlen; ++ unsigned char srr; ++ unsigned char rr; ++ unsigned char ts; ++ unsigned char is_strictroute: 1; ++ unsigned char srr_is_hit: 1; ++ unsigned char is_changed: 1; ++ unsigned char rr_needaddr: 1; ++ unsigned char ts_needtime: 1; ++ unsigned char ts_needaddr: 1; ++ unsigned char router_alert; ++ unsigned char cipso; ++ unsigned char __pad2; ++ unsigned char __data[0]; ++}; ++ ++struct ip_options_rcu { ++ struct callback_head rcu; ++ struct ip_options opt; ++}; ++ ++struct ipv6_txoptions { ++ refcount_t refcnt; ++ int tot_len; ++ __u16 opt_flen; ++ __u16 opt_nflen; ++ struct ipv6_opt_hdr *hopopt; ++ struct ipv6_opt_hdr *dst0opt; ++ struct ipv6_rt_hdr *srcrt; ++ struct ipv6_opt_hdr *dst1opt; ++ struct callback_head rcu; ++}; ++ ++struct inet_cork { ++ unsigned int flags; ++ __be32 addr; ++ struct ip_options *opt; ++ unsigned int fragsize; ++ int length; ++ struct dst_entry *dst; ++ u8 tx_flags; ++ __u8 ttl; ++ __s16 tos; ++ char priority; ++ __u16 gso_size; ++ u64 transmit_time; ++ u32 mark; ++}; ++ ++struct inet_cork_full { ++ struct inet_cork base; ++ struct flowi fl; ++}; ++ ++struct ipv6_pinfo; ++ ++struct ip_mc_socklist; ++ ++struct inet_sock { ++ struct sock sk; ++ struct ipv6_pinfo *pinet6; ++ __be32 inet_saddr; ++ __s16 uc_ttl; ++ __u16 cmsg_flags; ++ struct ip_options_rcu *inet_opt; ++ atomic_t inet_id; ++ __be16 inet_sport; ++ __u8 tos; ++ __u8 min_ttl; ++ __u8 mc_ttl; ++ __u8 pmtudisc; ++ __u8 recverr: 1; ++ __u8 is_icsk: 1; ++ __u8 freebind: 1; ++ __u8 hdrincl: 1; ++ __u8 mc_loop: 1; ++ __u8 transparent: 1; ++ __u8 mc_all: 1; ++ __u8 nodefrag: 1; ++ __u8 bind_address_no_port: 1; ++ __u8 recverr_rfc4884: 1; ++ __u8 defer_connect: 1; ++ __u8 rcv_tos; ++ __u8 convert_csum; ++ int uc_index; ++ int mc_index; ++ __be32 mc_addr; ++ struct ip_mc_socklist *mc_list; ++ struct inet_cork_full cork; ++ struct { ++ __u16 lo; ++ __u16 hi; ++ } local_port_range; ++}; ++ ++struct inet6_cork { ++ struct ipv6_txoptions *opt; ++ u8 hop_limit; ++ u8 tclass; ++}; ++ ++struct ipv6_mc_socklist; ++ ++struct ipv6_ac_socklist; ++ ++struct ipv6_fl_socklist; ++ ++struct ipv6_pinfo { ++ struct in6_addr saddr; ++ struct in6_pktinfo sticky_pktinfo; ++ const struct in6_addr *daddr_cache; ++ const struct in6_addr *saddr_cache; ++ __be32 flow_label; ++ __u32 frag_size; ++ __s16 hop_limit: 9; ++ __u16 __unused_1: 7; ++ __s16 mcast_hops: 9; ++ __u16 __unused_2: 6; ++ __u16 mc_loop: 1; ++ int ucast_oif; ++ int mcast_oif; ++ union { ++ struct { ++ __u16 srcrt: 1; ++ __u16 osrcrt: 1; ++ __u16 rxinfo: 1; ++ __u16 rxoinfo: 1; ++ __u16 rxhlim: 1; ++ __u16 rxohlim: 1; ++ __u16 hopopts: 1; ++ __u16 ohopopts: 1; ++ __u16 dstopts: 1; ++ __u16 odstopts: 1; ++ __u16 rxflow: 1; ++ __u16 rxtclass: 1; ++ __u16 rxpmtu: 1; ++ __u16 rxorigdstaddr: 1; ++ __u16 recvfragsize: 1; ++ } bits; ++ __u16 all; ++ } rxopt; ++ __u16 recverr: 1; ++ __u16 sndflow: 1; ++ __u16 repflow: 1; ++ __u16 pmtudisc: 3; ++ __u16 padding: 1; ++ __u16 srcprefs: 3; ++ __u16 dontfrag: 1; ++ __u16 autoflowlabel: 1; ++ __u16 autoflowlabel_set: 1; ++ __u16 mc_all: 1; ++ __u16 recverr_rfc4884: 1; ++ __u16 rtalert_isolate: 1; ++ __u8 min_hopcount; ++ __u8 tclass; ++ __be32 rcv_flowinfo; ++ __u32 dst_cookie; ++ struct ipv6_mc_socklist *ipv6_mc_list; ++ struct ipv6_ac_socklist *ipv6_ac_list; ++ struct ipv6_fl_socklist *ipv6_fl_list; ++ struct ipv6_txoptions *opt; ++ struct sk_buff *pktoptions; ++ struct sk_buff *rxpmtu; ++ struct inet6_cork cork; ++}; ++ ++struct ip6_sf_socklist; ++ ++struct ipv6_mc_socklist { ++ struct in6_addr addr; ++ int ifindex; ++ unsigned int sfmode; ++ struct ipv6_mc_socklist *next; ++ struct ip6_sf_socklist *sflist; ++ struct callback_head rcu; ++}; ++ ++struct ipv6_ac_socklist { ++ struct in6_addr acl_addr; ++ int acl_ifindex; ++ struct ipv6_ac_socklist *acl_next; ++}; ++ ++struct ip6_flowlabel; ++ ++struct ipv6_fl_socklist { ++ struct ipv6_fl_socklist *next; ++ struct ip6_flowlabel *fl; ++ struct callback_head rcu; ++}; ++ ++struct netlbl_lsm_cache { ++ refcount_t refcount; ++ void (*free)(const void *); ++ void *data; ++}; ++ ++struct netlbl_lsm_catmap { ++ u32 startbit; ++ u64 bitmap[4]; ++ struct netlbl_lsm_catmap *next; ++}; ++ ++struct netlbl_lsm_secattr { ++ u32 flags; ++ u32 type; ++ char *domain; ++ struct netlbl_lsm_cache *cache; ++ struct { ++ struct { ++ struct netlbl_lsm_catmap *cat; ++ u32 lvl; ++ } mls; ++ u32 secid; ++ } attr; ++}; ++ ++struct ip6_sf_socklist { ++ unsigned int sl_max; ++ unsigned int sl_count; ++ struct callback_head rcu; ++ struct in6_addr sl_addr[0]; ++}; ++ ++struct ip6_flowlabel { ++ struct ip6_flowlabel *next; ++ __be32 label; ++ atomic_t users; ++ struct in6_addr dst; ++ struct ipv6_txoptions *opt; ++ long unsigned int linger; ++ struct callback_head rcu; ++ u8 share; ++ union { ++ struct pid *pid; ++ kuid_t uid; ++ } owner; ++ long unsigned int lastuse; ++ long unsigned int expires; ++ struct net *fl_net; ++}; ++ ++enum sctp_endpoint_type { ++ SCTP_EP_TYPE_SOCKET = 0, ++ SCTP_EP_TYPE_ASSOCIATION = 1, ++}; ++ ++struct sctp_chunk; ++ ++struct sctp_inq { ++ struct list_head in_chunk_list; ++ struct sctp_chunk *in_progress; ++ struct work_struct immediate; ++}; ++ ++struct sctp_bind_addr { ++ __u16 port; ++ struct list_head address_list; ++}; ++ ++struct sctp_ep_common { ++ enum sctp_endpoint_type type; ++ refcount_t refcnt; ++ bool dead; ++ struct sock *sk; ++ struct net *net; ++ struct sctp_inq inqueue; ++ struct sctp_bind_addr bind_addr; ++}; ++ ++typedef __s32 sctp_assoc_t; ++ ++union sctp_addr { ++ struct sockaddr_in v4; ++ struct sockaddr_in6 v6; ++ struct sockaddr sa; ++}; ++ ++struct sctp_cookie { ++ __u32 my_vtag; ++ __u32 peer_vtag; ++ __u32 my_ttag; ++ __u32 peer_ttag; ++ ktime_t expiration; ++ __u16 sinit_num_ostreams; ++ __u16 sinit_max_instreams; ++ __u32 initial_tsn; ++ union sctp_addr peer_addr; ++ __u16 my_port; ++ __u8 prsctp_capable; ++ __u8 padding; ++ __u32 adaptation_ind; ++ __u8 auth_random[36]; ++ __u8 auth_hmacs[10]; ++ __u8 auth_chunks[20]; ++ __u32 raw_addr_list_len; ++}; ++ ++struct sctp_tsnmap { ++ long unsigned int *tsn_map; ++ __u32 base_tsn; ++ __u32 cumulative_tsn_ack_point; ++ __u32 max_tsn_seen; ++ __u16 len; ++ __u16 pending_data; ++ __u16 num_dup_tsns; ++ __be32 dup_tsns[16]; ++}; ++ ++struct sctp_inithdr_host { ++ __u32 init_tag; ++ __u32 a_rwnd; ++ __u16 num_outbound_streams; ++ __u16 num_inbound_streams; ++ __u32 initial_tsn; ++}; ++ ++enum sctp_state { ++ SCTP_STATE_CLOSED = 0, ++ SCTP_STATE_COOKIE_WAIT = 1, ++ SCTP_STATE_COOKIE_ECHOED = 2, ++ SCTP_STATE_ESTABLISHED = 3, ++ SCTP_STATE_SHUTDOWN_PENDING = 4, ++ SCTP_STATE_SHUTDOWN_SENT = 5, ++ SCTP_STATE_SHUTDOWN_RECEIVED = 6, ++ SCTP_STATE_SHUTDOWN_ACK_SENT = 7, ++}; ++ ++struct genradix_root; ++ ++struct __genradix { ++ struct genradix_root *root; ++}; ++ ++struct sctp_stream_out_ext; ++ ++struct sctp_stream_out { ++ union { ++ __u32 mid; ++ __u16 ssn; ++ }; ++ __u32 mid_uo; ++ struct sctp_stream_out_ext *ext; ++ __u8 state; ++}; ++ ++struct sctp_stream_in { ++ union { ++ __u32 mid; ++ __u16 ssn; ++ }; ++ __u32 mid_uo; ++ __u32 fsn; ++ __u32 fsn_uo; ++ char pd_mode; ++ char pd_mode_uo; ++}; ++ ++struct sctp_stream_interleave; ++ ++struct sctp_stream { ++ struct { ++ struct __genradix tree; ++ struct sctp_stream_out type[0]; ++ } out; ++ struct { ++ struct __genradix tree; ++ struct sctp_stream_in type[0]; ++ } in; ++ __u16 outcnt; ++ __u16 incnt; ++ struct sctp_stream_out *out_curr; ++ union { ++ struct { ++ struct list_head prio_list; ++ }; ++ struct { ++ struct list_head rr_list; ++ struct sctp_stream_out_ext *rr_next; ++ }; ++ struct { ++ struct list_head fc_list; ++ }; ++ }; ++ struct sctp_stream_interleave *si; ++}; ++ ++struct sctp_sched_ops; ++ ++struct sctp_outq { ++ struct sctp_association *asoc; ++ struct list_head out_chunk_list; ++ struct sctp_sched_ops *sched; ++ unsigned int out_qlen; ++ unsigned int error; ++ struct list_head control_chunk_list; ++ struct list_head sacked; ++ struct list_head retransmit; ++ struct list_head abandoned; ++ __u32 outstanding_bytes; ++ char fast_rtx; ++ char cork; ++}; ++ ++struct sctp_ulpq { ++ char pd_mode; ++ struct sctp_association *asoc; ++ struct sk_buff_head reasm; ++ struct sk_buff_head reasm_uo; ++ struct sk_buff_head lobby; ++}; ++ ++struct sctp_priv_assoc_stats { ++ struct __kernel_sockaddr_storage obs_rto_ipaddr; ++ __u64 max_obs_rto; ++ __u64 isacks; ++ __u64 osacks; ++ __u64 opackets; ++ __u64 ipackets; ++ __u64 rtxchunks; ++ __u64 outofseqtsns; ++ __u64 idupchunks; ++ __u64 gapcnt; ++ __u64 ouodchunks; ++ __u64 iuodchunks; ++ __u64 oodchunks; ++ __u64 iodchunks; ++ __u64 octrlchunks; ++ __u64 ictrlchunks; ++}; ++ ++struct sctp_endpoint; ++ ++struct sctp_transport; ++ ++struct sctp_random_param; ++ ++struct sctp_chunks_param; ++ ++struct sctp_hmac_algo_param; ++ ++struct sctp_auth_bytes; ++ ++struct sctp_shared_key; ++ ++struct sctp_association { ++ struct sctp_ep_common base; ++ struct list_head asocs; ++ sctp_assoc_t assoc_id; ++ struct sctp_endpoint *ep; ++ struct sctp_cookie c; ++ struct { ++ struct list_head transport_addr_list; ++ __u32 rwnd; ++ __u16 transport_count; ++ __u16 port; ++ struct sctp_transport *primary_path; ++ union sctp_addr primary_addr; ++ struct sctp_transport *active_path; ++ struct sctp_transport *retran_path; ++ struct sctp_transport *last_sent_to; ++ struct sctp_transport *last_data_from; ++ struct sctp_tsnmap tsn_map; ++ __be16 addip_disabled_mask; ++ __u16 ecn_capable: 1; ++ __u16 ipv4_address: 1; ++ __u16 ipv6_address: 1; ++ __u16 asconf_capable: 1; ++ __u16 prsctp_capable: 1; ++ __u16 reconf_capable: 1; ++ __u16 intl_capable: 1; ++ __u16 auth_capable: 1; ++ __u16 sack_needed: 1; ++ __u16 sack_generation: 1; ++ __u16 zero_window_announced: 1; ++ __u32 sack_cnt; ++ __u32 adaptation_ind; ++ struct sctp_inithdr_host i; ++ void *cookie; ++ int cookie_len; ++ __u32 addip_serial; ++ struct sctp_random_param *peer_random; ++ struct sctp_chunks_param *peer_chunks; ++ struct sctp_hmac_algo_param *peer_hmacs; ++ } peer; ++ enum sctp_state state; ++ int overall_error_count; ++ ktime_t cookie_life; ++ long unsigned int rto_initial; ++ long unsigned int rto_max; ++ long unsigned int rto_min; ++ int max_burst; ++ int max_retrans; ++ __u16 pf_retrans; ++ __u16 ps_retrans; ++ __u16 max_init_attempts; ++ __u16 init_retries; ++ long unsigned int max_init_timeo; ++ long unsigned int hbinterval; ++ long unsigned int probe_interval; ++ __be16 encap_port; ++ __u16 pathmaxrxt; ++ __u32 flowlabel; ++ __u8 dscp; ++ __u8 pmtu_pending; ++ __u32 pathmtu; ++ __u32 param_flags; ++ __u32 sackfreq; ++ long unsigned int sackdelay; ++ long unsigned int timeouts[12]; ++ struct timer_list timers[12]; ++ struct sctp_transport *shutdown_last_sent_to; ++ struct sctp_transport *init_last_sent_to; ++ int shutdown_retries; ++ __u32 next_tsn; ++ __u32 ctsn_ack_point; ++ __u32 adv_peer_ack_point; ++ __u32 highest_sacked; ++ __u32 fast_recovery_exit; ++ __u8 fast_recovery; ++ __u16 unack_data; ++ __u32 rtx_data_chunks; ++ __u32 rwnd; ++ __u32 a_rwnd; ++ __u32 rwnd_over; ++ __u32 rwnd_press; ++ int sndbuf_used; ++ atomic_t rmem_alloc; ++ wait_queue_head_t wait; ++ __u32 frag_point; ++ __u32 user_frag; ++ int init_err_counter; ++ int init_cycle; ++ __u16 default_stream; ++ __u16 default_flags; ++ __u32 default_ppid; ++ __u32 default_context; ++ __u32 default_timetolive; ++ __u32 default_rcv_context; ++ struct sctp_stream stream; ++ struct sctp_outq outqueue; ++ struct sctp_ulpq ulpq; ++ __u32 last_ecne_tsn; ++ __u32 last_cwr_tsn; ++ int numduptsns; ++ struct sctp_chunk *addip_last_asconf; ++ struct list_head asconf_ack_list; ++ struct list_head addip_chunk_list; ++ __u32 addip_serial; ++ int src_out_of_asoc_ok; ++ union sctp_addr *asconf_addr_del_pending; ++ struct sctp_transport *new_transport; ++ struct list_head endpoint_shared_keys; ++ struct sctp_auth_bytes *asoc_shared_key; ++ struct sctp_shared_key *shkey; ++ __u16 default_hmac_id; ++ __u16 active_key_id; ++ __u8 need_ecne: 1; ++ __u8 temp: 1; ++ __u8 pf_expose: 2; ++ __u8 force_delay: 1; ++ __u8 strreset_enable; ++ __u8 strreset_outstanding; ++ __u32 strreset_outseq; ++ __u32 strreset_inseq; ++ __u32 strreset_result[2]; ++ struct sctp_chunk *strreset_chunk; ++ struct sctp_priv_assoc_stats stats; ++ int sent_cnt_removable; ++ __u16 subscribe; ++ __u64 abandoned_unsent[3]; ++ __u64 abandoned_sent[3]; ++ u32 secid; ++ u32 peer_secid; ++ struct callback_head rcu; ++}; ++ ++struct sk_security_struct { ++ enum { ++ NLBL_UNSET = 0, ++ NLBL_REQUIRE = 1, ++ NLBL_LABELED = 2, ++ NLBL_REQSKB = 3, ++ NLBL_CONNLABELED = 4, ++ } nlbl_state; ++ struct netlbl_lsm_secattr *nlbl_secattr; ++ u32 sid; ++ u32 peer_sid; ++ u16 sclass; ++ enum { ++ SCTP_ASSOC_UNSET = 0, ++ SCTP_ASSOC_SET = 1, ++ } sctp_assoc_state; ++}; ++ ++struct sctp_initmsg { ++ __u16 sinit_num_ostreams; ++ __u16 sinit_max_instreams; ++ __u16 sinit_max_attempts; ++ __u16 sinit_max_init_timeo; ++}; ++ ++struct sctp_sndrcvinfo { ++ __u16 sinfo_stream; ++ __u16 sinfo_ssn; ++ __u16 sinfo_flags; ++ __u32 sinfo_ppid; ++ __u32 sinfo_context; ++ __u32 sinfo_timetolive; ++ __u32 sinfo_tsn; ++ __u32 sinfo_cumtsn; ++ sctp_assoc_t sinfo_assoc_id; ++}; ++ ++struct sctp_rtoinfo { ++ sctp_assoc_t srto_assoc_id; ++ __u32 srto_initial; ++ __u32 srto_max; ++ __u32 srto_min; ++}; ++ ++struct sctp_assocparams { ++ sctp_assoc_t sasoc_assoc_id; ++ __u16 sasoc_asocmaxrxt; ++ __u16 sasoc_number_peer_destinations; ++ __u32 sasoc_peer_rwnd; ++ __u32 sasoc_local_rwnd; ++ __u32 sasoc_cookie_life; ++}; ++ ++struct sctp_paddrparams { ++ sctp_assoc_t spp_assoc_id; ++ struct __kernel_sockaddr_storage spp_address; ++ __u32 spp_hbinterval; ++ __u16 spp_pathmaxrxt; ++ __u32 spp_pathmtu; ++ __u32 spp_sackdelay; ++ __u32 spp_flags; ++ __u32 spp_ipv6_flowlabel; ++ __u8 spp_dscp; ++ int: 0; ++} __attribute__((packed)); ++ ++struct sctphdr { ++ __be16 source; ++ __be16 dest; ++ __be32 vtag; ++ __le32 checksum; ++}; ++ ++struct sctp_chunkhdr { ++ __u8 type; ++ __u8 flags; ++ __be16 length; ++}; ++ ++struct sctp_paramhdr { ++ __be16 type; ++ __be16 length; ++}; ++ ++struct sctp_datahdr { ++ __be32 tsn; ++ __be16 stream; ++ __be16 ssn; ++ __u32 ppid; ++}; ++ ++struct sctp_idatahdr { ++ __be32 tsn; ++ __be16 stream; ++ __be16 reserved; ++ __be32 mid; ++ union { ++ __u32 ppid; ++ __be32 fsn; ++ }; ++ __u8 payload[0]; ++}; ++ ++struct sctp_inithdr { ++ __be32 init_tag; ++ __be32 a_rwnd; ++ __be16 num_outbound_streams; ++ __be16 num_inbound_streams; ++ __be32 initial_tsn; ++}; ++ ++struct sctp_ipv4addr_param { ++ struct sctp_paramhdr param_hdr; ++ struct in_addr addr; ++}; ++ ++struct sctp_ipv6addr_param { ++ struct sctp_paramhdr param_hdr; ++ struct in6_addr addr; ++}; ++ ++struct sctp_cookie_preserve_param { ++ struct sctp_paramhdr param_hdr; ++ __be32 lifespan_increment; ++}; ++ ++struct sctp_hostname_param { ++ struct sctp_paramhdr param_hdr; ++ uint8_t hostname[0]; ++}; ++ ++struct sctp_supported_addrs_param { ++ struct sctp_paramhdr param_hdr; ++ __be16 types[0]; ++}; ++ ++struct sctp_adaptation_ind_param { ++ struct sctp_paramhdr param_hdr; ++ __be32 adaptation_ind; ++}; ++ ++struct sctp_supported_ext_param { ++ struct sctp_paramhdr param_hdr; ++ __u8 chunks[0]; ++}; ++ ++struct sctp_random_param { ++ struct sctp_paramhdr param_hdr; ++ __u8 random_val[0]; ++}; ++ ++struct sctp_chunks_param { ++ struct sctp_paramhdr param_hdr; ++ __u8 chunks[0]; ++}; ++ ++struct sctp_hmac_algo_param { ++ struct sctp_paramhdr param_hdr; ++ __be16 hmac_ids[0]; ++}; ++ ++struct sctp_cookie_param { ++ struct sctp_paramhdr p; ++ __u8 body[0]; ++}; ++ ++struct sctp_sackhdr { ++ __be32 cum_tsn_ack; ++ __be32 a_rwnd; ++ __be16 num_gap_ack_blocks; ++ __be16 num_dup_tsns; ++}; ++ ++struct sctp_heartbeathdr { ++ struct sctp_paramhdr info; ++}; ++ ++struct sctp_shutdownhdr { ++ __be32 cum_tsn_ack; ++}; ++ ++struct sctp_errhdr { ++ __be16 cause; ++ __be16 length; ++}; ++ ++struct sctp_ecnehdr { ++ __be32 lowest_tsn; ++}; ++ ++struct sctp_cwrhdr { ++ __be32 lowest_tsn; ++}; ++ ++struct sctp_fwdtsn_hdr { ++ __be32 new_cum_tsn; ++}; ++ ++struct sctp_ifwdtsn_hdr { ++ __be32 new_cum_tsn; ++}; ++ ++struct sctp_addip_param { ++ struct sctp_paramhdr param_hdr; ++ __be32 crr_id; ++}; ++ ++struct sctp_addiphdr { ++ __be32 serial; ++}; ++ ++struct sctp_authhdr { ++ __be16 shkey_id; ++ __be16 hmac_id; ++}; ++ ++struct sctp_auth_bytes { ++ refcount_t refcnt; ++ __u32 len; ++ __u8 data[0]; ++}; ++ ++struct sctp_shared_key { ++ struct list_head key_list; ++ struct sctp_auth_bytes *key; ++ refcount_t refcnt; ++ __u16 key_id; ++ __u8 deactivated; ++}; ++ ++enum sctp_scope { ++ SCTP_SCOPE_GLOBAL = 0, ++ SCTP_SCOPE_PRIVATE = 1, ++ SCTP_SCOPE_LINK = 2, ++ SCTP_SCOPE_LOOPBACK = 3, ++ SCTP_SCOPE_UNUSABLE = 4, ++}; ++ ++struct sctp_ulpevent { ++ struct sctp_association *asoc; ++ struct sctp_chunk *chunk; ++ unsigned int rmem_len; ++ union { ++ __u32 mid; ++ __u16 ssn; ++ }; ++ union { ++ __u32 ppid; ++ __u32 fsn; ++ }; ++ __u32 tsn; ++ __u32 cumtsn; ++ __u16 stream; ++ __u16 flags; ++ __u16 msg_flags; ++} __attribute__((packed)); ++ ++union sctp_addr_param; ++ ++union sctp_params { ++ void *v; ++ struct sctp_paramhdr *p; ++ struct sctp_cookie_preserve_param *life; ++ struct sctp_hostname_param *dns; ++ struct sctp_cookie_param *cookie; ++ struct sctp_supported_addrs_param *sat; ++ struct sctp_ipv4addr_param *v4; ++ struct sctp_ipv6addr_param *v6; ++ union sctp_addr_param *addr; ++ struct sctp_adaptation_ind_param *aind; ++ struct sctp_supported_ext_param *ext; ++ struct sctp_random_param *random; ++ struct sctp_chunks_param *chunks; ++ struct sctp_hmac_algo_param *hmac_algo; ++ struct sctp_addip_param *addip; ++}; ++ ++struct sctp_sender_hb_info; ++ ++struct sctp_signed_cookie; ++ ++struct sctp_datamsg; ++ ++struct sctp_chunk { ++ struct list_head list; ++ refcount_t refcnt; ++ int sent_count; ++ union { ++ struct list_head transmitted_list; ++ struct list_head stream_list; ++ }; ++ struct list_head frag_list; ++ struct sk_buff *skb; ++ union { ++ struct sk_buff *head_skb; ++ struct sctp_shared_key *shkey; ++ }; ++ union sctp_params param_hdr; ++ union { ++ __u8 *v; ++ struct sctp_datahdr *data_hdr; ++ struct sctp_inithdr *init_hdr; ++ struct sctp_sackhdr *sack_hdr; ++ struct sctp_heartbeathdr *hb_hdr; ++ struct sctp_sender_hb_info *hbs_hdr; ++ struct sctp_shutdownhdr *shutdown_hdr; ++ struct sctp_signed_cookie *cookie_hdr; ++ struct sctp_ecnehdr *ecne_hdr; ++ struct sctp_cwrhdr *ecn_cwr_hdr; ++ struct sctp_errhdr *err_hdr; ++ struct sctp_addiphdr *addip_hdr; ++ struct sctp_fwdtsn_hdr *fwdtsn_hdr; ++ struct sctp_authhdr *auth_hdr; ++ struct sctp_idatahdr *idata_hdr; ++ struct sctp_ifwdtsn_hdr *ifwdtsn_hdr; ++ } subh; ++ __u8 *chunk_end; ++ struct sctp_chunkhdr *chunk_hdr; ++ struct sctphdr *sctp_hdr; ++ struct sctp_sndrcvinfo sinfo; ++ struct sctp_association *asoc; ++ struct sctp_ep_common *rcvr; ++ long unsigned int sent_at; ++ union sctp_addr source; ++ union sctp_addr dest; ++ struct sctp_datamsg *msg; ++ struct sctp_transport *transport; ++ struct sk_buff *auth_chunk; ++ __u16 rtt_in_progress: 1; ++ __u16 has_tsn: 1; ++ __u16 has_ssn: 1; ++ __u16 singleton: 1; ++ __u16 end_of_packet: 1; ++ __u16 ecn_ce_done: 1; ++ __u16 pdiscard: 1; ++ __u16 tsn_gap_acked: 1; ++ __u16 data_accepted: 1; ++ __u16 auth: 1; ++ __u16 has_asconf: 1; ++ __u16 pmtu_probe: 1; ++ __u16 tsn_missing_report: 2; ++ __u16 fast_retransmit: 2; ++}; ++ ++struct sctp_stream_interleave { ++ __u16 data_chunk_len; ++ __u16 ftsn_chunk_len; ++ struct sctp_chunk * (*make_datafrag)(const struct sctp_association *, const struct sctp_sndrcvinfo *, int, __u8, gfp_t); ++ void (*assign_number)(struct sctp_chunk *); ++ bool (*validate_data)(struct sctp_chunk *); ++ int (*ulpevent_data)(struct sctp_ulpq *, struct sctp_chunk *, gfp_t); ++ int (*enqueue_event)(struct sctp_ulpq *, struct sctp_ulpevent *); ++ void (*renege_events)(struct sctp_ulpq *, struct sctp_chunk *, gfp_t); ++ void (*start_pd)(struct sctp_ulpq *, gfp_t); ++ void (*abort_pd)(struct sctp_ulpq *, gfp_t); ++ void (*generate_ftsn)(struct sctp_outq *, __u32); ++ bool (*validate_ftsn)(struct sctp_chunk *); ++ void (*report_ftsn)(struct sctp_ulpq *, __u32); ++ void (*handle_ftsn)(struct sctp_ulpq *, struct sctp_chunk *); ++}; ++ ++struct sctp_bind_bucket { ++ short unsigned int port; ++ signed char fastreuse; ++ signed char fastreuseport; ++ kuid_t fastuid; ++ struct hlist_node node; ++ struct hlist_head owner; ++ struct net *net; ++}; ++ ++enum sctp_socket_type { ++ SCTP_SOCKET_UDP = 0, ++ SCTP_SOCKET_UDP_HIGH_BANDWIDTH = 1, ++ SCTP_SOCKET_TCP = 2, ++}; ++ ++struct sctp_pf; ++ ++struct sctp_sock { ++ struct inet_sock inet; ++ enum sctp_socket_type type; ++ struct sctp_pf *pf; ++ struct crypto_shash *hmac; ++ char *sctp_hmac_alg; ++ struct sctp_endpoint *ep; ++ struct sctp_bind_bucket *bind_hash; ++ __u16 default_stream; ++ __u32 default_ppid; ++ __u16 default_flags; ++ __u32 default_context; ++ __u32 default_timetolive; ++ __u32 default_rcv_context; ++ int max_burst; ++ __u32 hbinterval; ++ __u32 probe_interval; ++ __be16 udp_port; ++ __be16 encap_port; ++ __u16 pathmaxrxt; ++ __u32 flowlabel; ++ __u8 dscp; ++ __u16 pf_retrans; ++ __u16 ps_retrans; ++ __u32 pathmtu; ++ __u32 sackdelay; ++ __u32 sackfreq; ++ __u32 param_flags; ++ __u32 default_ss; ++ struct sctp_rtoinfo rtoinfo; ++ struct sctp_paddrparams paddrparam; ++ struct sctp_assocparams assocparams; ++ __u16 subscribe; ++ struct sctp_initmsg initmsg; ++ int user_frag; ++ __u32 autoclose; ++ __u32 adaptation_ind; ++ __u32 pd_point; ++ __u16 nodelay: 1; ++ __u16 pf_expose: 2; ++ __u16 reuse: 1; ++ __u16 disable_fragments: 1; ++ __u16 v4mapped: 1; ++ __u16 frag_interleave: 1; ++ __u16 recvrcvinfo: 1; ++ __u16 recvnxtinfo: 1; ++ __u16 data_ready_signalled: 1; ++ atomic_t pd_mode; ++ struct sk_buff_head pd_lobby; ++ struct list_head auto_asconf_list; ++ int do_auto_asconf; ++}; ++ ++struct sctp_af; ++ ++struct sctp_pf { ++ void (*event_msgname)(struct sctp_ulpevent *, char *, int *); ++ void (*skb_msgname)(struct sk_buff *, char *, int *); ++ int (*af_supported)(sa_family_t, struct sctp_sock *); ++ int (*cmp_addr)(const union sctp_addr *, const union sctp_addr *, struct sctp_sock *); ++ int (*bind_verify)(struct sctp_sock *, union sctp_addr *); ++ int (*send_verify)(struct sctp_sock *, union sctp_addr *); ++ int (*supported_addrs)(const struct sctp_sock *, __be16 *); ++ struct sock * (*create_accept_sk)(struct sock *, struct sctp_association *, bool); ++ int (*addr_to_user)(struct sctp_sock *, union sctp_addr *); ++ void (*to_sk_saddr)(union sctp_addr *, struct sock *); ++ void (*to_sk_daddr)(union sctp_addr *, struct sock *); ++ void (*copy_ip_options)(struct sock *, struct sock *); ++ struct sctp_af *af; ++}; ++ ++struct sctp_endpoint { ++ struct sctp_ep_common base; ++ struct hlist_node node; ++ int hashent; ++ struct list_head asocs; ++ __u8 secret_key[32]; ++ __u8 *digest; ++ __u32 sndbuf_policy; ++ __u32 rcvbuf_policy; ++ struct crypto_shash **auth_hmacs; ++ struct sctp_hmac_algo_param *auth_hmacs_list; ++ struct sctp_chunks_param *auth_chunk_list; ++ struct list_head endpoint_shared_keys; ++ __u16 active_key_id; ++ __u8 ecn_enable: 1; ++ __u8 auth_enable: 1; ++ __u8 intl_enable: 1; ++ __u8 prsctp_enable: 1; ++ __u8 asconf_enable: 1; ++ __u8 reconf_enable: 1; ++ __u8 strreset_enable; ++ struct callback_head rcu; ++}; ++ ++struct sctp_signed_cookie { ++ __u8 signature[32]; ++ __u32 __pad; ++ struct sctp_cookie c; ++} __attribute__((packed)); ++ ++union sctp_addr_param { ++ struct sctp_paramhdr p; ++ struct sctp_ipv4addr_param v4; ++ struct sctp_ipv6addr_param v6; ++}; ++ ++struct sctp_sender_hb_info { ++ struct sctp_paramhdr param_hdr; ++ union sctp_addr daddr; ++ long unsigned int sent_at; ++ __u64 hb_nonce; ++ __u32 probe_size; ++}; ++ ++struct sctp_af { ++ int (*sctp_xmit)(struct sk_buff *, struct sctp_transport *); ++ int (*setsockopt)(struct sock *, int, int, sockptr_t, unsigned int); ++ int (*getsockopt)(struct sock *, int, int, char *, int *); ++ void (*get_dst)(struct sctp_transport *, union sctp_addr *, struct flowi *, struct sock *); ++ void (*get_saddr)(struct sctp_sock *, struct sctp_transport *, struct flowi *); ++ void (*copy_addrlist)(struct list_head *, struct net_device *); ++ int (*cmp_addr)(const union sctp_addr *, const union sctp_addr *); ++ void (*addr_copy)(union sctp_addr *, union sctp_addr *); ++ void (*from_skb)(union sctp_addr *, struct sk_buff *, int); ++ void (*from_sk)(union sctp_addr *, struct sock *); ++ bool (*from_addr_param)(union sctp_addr *, union sctp_addr_param *, __be16, int); ++ int (*to_addr_param)(const union sctp_addr *, union sctp_addr_param *); ++ int (*addr_valid)(union sctp_addr *, struct sctp_sock *, const struct sk_buff *); ++ enum sctp_scope (*scope)(union sctp_addr *); ++ void (*inaddr_any)(union sctp_addr *, __be16); ++ int (*is_any)(const union sctp_addr *); ++ int (*available)(union sctp_addr *, struct sctp_sock *); ++ int (*skb_iif)(const struct sk_buff *); ++ int (*skb_sdif)(const struct sk_buff *); ++ int (*is_ce)(const struct sk_buff *); ++ void (*seq_dump_addr)(struct seq_file *, union sctp_addr *); ++ void (*ecn_capable)(struct sock *); ++ __u16 net_header_len; ++ int sockaddr_len; ++ int (*ip_options_len)(struct sock *); ++ sa_family_t sa_family; ++ struct list_head list; ++}; ++ ++struct sctp_packet { ++ __u16 source_port; ++ __u16 destination_port; ++ __u32 vtag; ++ struct list_head chunk_list; ++ size_t overhead; ++ size_t size; ++ size_t max_size; ++ struct sctp_transport *transport; ++ struct sctp_chunk *auth; ++ u8 has_cookie_echo: 1; ++ u8 has_sack: 1; ++ u8 has_auth: 1; ++ u8 has_data: 1; ++ u8 ipfragok: 1; ++}; ++ ++struct sctp_transport { ++ struct list_head transports; ++ struct rhlist_head node; ++ refcount_t refcnt; ++ __u32 rto_pending: 1; ++ __u32 hb_sent: 1; ++ __u32 pmtu_pending: 1; ++ __u32 dst_pending_confirm: 1; ++ __u32 sack_generation: 1; ++ u32 dst_cookie; ++ struct flowi fl; ++ union sctp_addr ipaddr; ++ struct sctp_af *af_specific; ++ struct sctp_association *asoc; ++ long unsigned int rto; ++ __u32 rtt; ++ __u32 rttvar; ++ __u32 srtt; ++ __u32 cwnd; ++ __u32 ssthresh; ++ __u32 partial_bytes_acked; ++ __u32 flight_size; ++ __u32 burst_limited; ++ struct dst_entry *dst; ++ union sctp_addr saddr; ++ long unsigned int hbinterval; ++ long unsigned int probe_interval; ++ long unsigned int sackdelay; ++ __u32 sackfreq; ++ atomic_t mtu_info; ++ ktime_t last_time_heard; ++ long unsigned int last_time_sent; ++ long unsigned int last_time_ecne_reduced; ++ __be16 encap_port; ++ __u16 pathmaxrxt; ++ __u32 flowlabel; ++ __u8 dscp; ++ __u16 pf_retrans; ++ __u16 ps_retrans; ++ __u32 pathmtu; ++ __u32 param_flags; ++ int init_sent_count; ++ int state; ++ short unsigned int error_count; ++ struct timer_list T3_rtx_timer; ++ struct timer_list hb_timer; ++ struct timer_list proto_unreach_timer; ++ struct timer_list reconf_timer; ++ struct timer_list probe_timer; ++ struct list_head transmitted; ++ struct sctp_packet packet; ++ struct list_head send_ready; ++ struct { ++ __u32 next_tsn_at_change; ++ char changeover_active; ++ char cycling_changeover; ++ char cacc_saw_newack; ++ } cacc; ++ struct { ++ __u16 pmtu; ++ __u16 probe_size; ++ __u16 probe_high; ++ __u8 probe_count; ++ __u8 state; ++ } pl; ++ __u64 hb_nonce; ++ struct callback_head rcu; ++}; ++ ++struct sctp_datamsg { ++ struct list_head chunks; ++ refcount_t refcnt; ++ long unsigned int expires_at; ++ int send_error; ++ u8 send_failed: 1; ++ u8 can_delay: 1; ++ u8 abandoned: 1; ++}; ++ ++struct sctp_stream_priorities { ++ struct list_head prio_sched; ++ struct list_head active; ++ struct sctp_stream_out_ext *next; ++ __u16 prio; ++ __u16 users; ++}; ++ ++struct sctp_stream_out_ext { ++ __u64 abandoned_unsent[3]; ++ __u64 abandoned_sent[3]; ++ struct list_head outq; ++ union { ++ struct { ++ struct list_head prio_list; ++ struct sctp_stream_priorities *prio_head; ++ }; ++ struct { ++ struct list_head rr_list; ++ }; ++ struct { ++ struct list_head fc_list; ++ __u32 fc_length; ++ __u16 fc_weight; ++ }; ++ }; ++}; ++ ++enum devcg_behavior { ++ DEVCG_DEFAULT_NONE = 0, ++ DEVCG_DEFAULT_ALLOW = 1, ++ DEVCG_DEFAULT_DENY = 2, ++}; ++ ++struct dev_exception_item { ++ u32 major; ++ u32 minor; ++ short int type; ++ short int access; ++ struct list_head list; ++ struct callback_head rcu; ++}; ++ ++struct dev_cgroup { ++ struct cgroup_subsys_state css; ++ struct list_head exceptions; ++ enum devcg_behavior behavior; ++}; ++ ++struct landlock_object; ++ ++struct landlock_object_underops { ++ void (*release)(struct landlock_object * const); ++}; ++ ++struct landlock_object { ++ refcount_t usage; ++ spinlock_t lock; ++ void *underobj; ++ union { ++ struct callback_head rcu_free; ++ const struct landlock_object_underops *underops; ++ }; ++}; ++ ++typedef u16 access_mask_t; ++ ++typedef u16 layer_mask_t; ++ ++struct landlock_layer { ++ u16 level; ++ access_mask_t access; ++}; ++ ++struct landlock_rule { ++ struct rb_node node; ++ struct landlock_object *object; ++ u32 num_layers; ++ struct landlock_layer layers[0]; ++}; ++ ++struct landlock_hierarchy { ++ struct landlock_hierarchy *parent; ++ refcount_t usage; ++}; ++ ++struct landlock_ruleset { ++ struct rb_root root; ++ struct landlock_hierarchy *hierarchy; ++ union { ++ struct work_struct work_free; ++ struct { ++ struct mutex lock; ++ refcount_t usage; ++ u32 num_rules; ++ u32 num_layers; ++ access_mask_t fs_access_masks[0]; ++ }; ++ }; ++}; ++ ++struct landlock_cred_security { ++ struct landlock_ruleset *domain; ++}; ++ ++struct landlock_inode_security { ++ struct landlock_object *object; ++}; ++ ++struct landlock_file_security { ++ access_mask_t allowed_access; ++}; ++ ++struct landlock_superblock_security { ++ atomic_long_t inode_refs; ++}; ++ ++struct public_key { ++ void *key; ++ u32 keylen; ++ enum OID algo; ++ void *params; ++ u32 paramlen; ++ bool key_is_private; ++ const char *id_type; ++ const char *pkey_algo; ++ long unsigned int key_eflags; ++}; ++ ++struct asymmetric_key_id; ++ ++struct public_key_signature { ++ struct asymmetric_key_id *auth_ids[3]; ++ u8 *s; ++ u8 *digest; ++ u32 s_size; ++ u32 digest_size; ++ const char *pkey_algo; ++ const char *hash_algo; ++ const char *encoding; ++ const void *data; ++ unsigned int data_size; ++}; ++ ++struct asymmetric_key_id { ++ short unsigned int len; ++ unsigned char data[0]; ++}; ++ ++enum asymmetric_payload_bits { ++ asym_crypto = 0, ++ asym_subtype = 1, ++ asym_key_ids = 2, ++ asym_auth = 3, ++}; ++ ++struct signature_v2_hdr { ++ uint8_t type; ++ uint8_t version; ++ uint8_t hash_algo; ++ __be32 keyid; ++ __be16 sig_size; ++ uint8_t sig[0]; ++} __attribute__((packed)); ++ ++struct tpm_digest { ++ u16 alg_id; ++ u8 digest[64]; ++}; ++ ++enum integrity_status { ++ INTEGRITY_PASS = 0, ++ INTEGRITY_PASS_IMMUTABLE = 1, ++ INTEGRITY_FAIL = 2, ++ INTEGRITY_FAIL_IMMUTABLE = 3, ++ INTEGRITY_NOLABEL = 4, ++ INTEGRITY_NOXATTRS = 5, ++ INTEGRITY_UNKNOWN = 6, ++}; ++ ++struct evm_ima_xattr_data { ++ u8 type; ++ u8 data[0]; ++}; ++ ++struct ima_digest_data { ++ u8 algo; ++ u8 length; ++ union { ++ struct { ++ u8 unused; ++ u8 type; ++ } sha1; ++ struct { ++ u8 type; ++ u8 algo; ++ } ng; ++ u8 data[2]; ++ } xattr; ++ u8 digest[0]; ++}; ++ ++struct integrity_iint_cache { ++ struct rb_node rb_node; ++ struct mutex mutex; ++ struct inode *inode; ++ u64 version; ++ long unsigned int flags; ++ long unsigned int measured_pcrs; ++ long unsigned int atomic_flags; ++ enum integrity_status ima_file_status: 4; ++ enum integrity_status ima_mmap_status: 4; ++ enum integrity_status ima_bprm_status: 4; ++ enum integrity_status ima_read_status: 4; ++ enum integrity_status ima_creds_status: 4; ++ enum integrity_status evm_status: 4; ++ struct ima_digest_data *ima_hash; ++}; ++ ++enum ima_show_type { ++ IMA_SHOW_BINARY = 0, ++ IMA_SHOW_BINARY_NO_FIELD_LEN = 1, ++ IMA_SHOW_BINARY_OLD_STRING_FMT = 2, ++ IMA_SHOW_ASCII = 3, ++}; ++ ++struct modsig; ++ ++struct ima_event_data { ++ struct integrity_iint_cache *iint; ++ struct file *file; ++ const unsigned char *filename; ++ struct evm_ima_xattr_data *xattr_value; ++ int xattr_len; ++ const struct modsig *modsig; ++ const char *violation; ++ const void *buf; ++ int buf_len; ++}; ++ ++struct pkcs7_message; ++ ++struct modsig { ++ struct pkcs7_message *pkcs7_msg; ++ enum hash_algo hash_algo; ++ const u8 *digest; ++ u32 digest_size; ++ int raw_pkcs7_len; ++ u8 raw_pkcs7[0]; ++}; ++ ++struct ima_field_data { ++ u8 *data; ++ u32 len; ++}; ++ ++struct ima_template_field { ++ const char field_id[16]; ++ int (*field_init)(struct ima_event_data *, struct ima_field_data *); ++ void (*field_show)(struct seq_file *, enum ima_show_type, struct ima_field_data *); ++}; ++ ++struct ima_template_desc { ++ struct list_head list; ++ char *name; ++ char *fmt; ++ int num_fields; ++ const struct ima_template_field **fields; ++}; ++ ++struct ima_template_entry { ++ int pcr; ++ struct tpm_digest *digests; ++ struct ima_template_desc *template_desc; ++ u32 template_data_len; ++ struct ima_field_data template_data[0]; ++}; ++ ++struct ima_queue_entry { ++ struct hlist_node hnext; ++ struct list_head later; ++ struct ima_template_entry *entry; ++}; ++ ++struct ima_h_table { ++ atomic_long_t len; ++ atomic_long_t violations; ++ struct hlist_head queue[1024]; ++}; ++ ++enum ima_fs_flags { ++ IMA_FS_BUSY = 0, ++}; ++ ++struct ahash_request { ++ struct crypto_async_request base; ++ unsigned int nbytes; ++ struct scatterlist *src; ++ u8 *result; ++ void *priv; ++ void *__ctx[0]; ++}; ++ ++struct crypto_ahash { ++ int (*init)(struct ahash_request *); ++ int (*update)(struct ahash_request *); ++ int (*final)(struct ahash_request *); ++ int (*finup)(struct ahash_request *); ++ int (*digest)(struct ahash_request *); ++ int (*export)(struct ahash_request *, void *); ++ int (*import)(struct ahash_request *, const void *); ++ int (*setkey)(struct crypto_ahash *, const u8 *, unsigned int); ++ unsigned int reqsize; ++ struct crypto_tfm base; ++}; ++ ++enum tpm_pcrs { ++ TPM_PCR0 = 0, ++ TPM_PCR8 = 8, ++ TPM_PCR10 = 10, ++}; ++ ++struct ima_algo_desc { ++ struct crypto_shash *tfm; ++ enum hash_algo algo; ++}; ++ ++struct ima_kexec_hdr { ++ u16 version; ++ u16 _reserved0; ++ u32 _reserved1; ++ u64 buffer_size; ++ u64 count; ++}; ++ ++enum header_fields { ++ HDR_PCR = 0, ++ HDR_DIGEST = 1, ++ HDR_TEMPLATE_NAME = 2, ++ HDR_TEMPLATE_DATA = 3, ++ HDR__LAST = 4, ++}; ++ ++enum ima_hooks { ++ NONE = 0, ++ FILE_CHECK = 1, ++ MMAP_CHECK = 2, ++ MMAP_CHECK_REQPROT = 3, ++ BPRM_CHECK = 4, ++ CREDS_CHECK = 5, ++ POST_SETATTR = 6, ++ MODULE_CHECK = 7, ++ FIRMWARE_CHECK = 8, ++ KEXEC_KERNEL_CHECK = 9, ++ KEXEC_INITRAMFS_CHECK = 10, ++ POLICY_CHECK = 11, ++ KEXEC_CMDLINE = 12, ++ KEY_CHECK = 13, ++ CRITICAL_DATA = 14, ++ SETXATTR_CHECK = 15, ++ MAX_CHECK = 16, ++}; ++ ++struct encrypted_key_payload { ++ struct callback_head rcu; ++ char *format; ++ char *master_desc; ++ char *datalen; ++ u8 *iv; ++ u8 *encrypted_data; ++ short unsigned int datablob_len; ++ short unsigned int decrypted_datalen; ++ short unsigned int payload_datalen; ++ short unsigned int encrypted_key_format; ++ u8 *decrypted_data; ++ u8 payload_data[0]; ++}; ++ ++enum evm_ima_xattr_type { ++ IMA_XATTR_DIGEST = 1, ++ EVM_XATTR_HMAC = 2, ++ EVM_IMA_XATTR_DIGSIG = 3, ++ IMA_XATTR_DIGEST_NG = 4, ++ EVM_XATTR_PORTABLE_DIGSIG = 5, ++ IMA_VERITY_DIGSIG = 6, ++ IMA_XATTR_LAST = 7, ++}; ++ ++struct xattr_list { ++ struct list_head list; ++ char *name; ++ bool enabled; ++}; ++ ++struct evm_digest { ++ struct ima_digest_data hdr; ++ char digest[64]; ++}; ++ ++struct h_misc { ++ long unsigned int ino; ++ __u32 generation; ++ uid_t uid; ++ gid_t gid; ++ umode_t mode; ++}; ++ ++struct crypto_comp { ++ struct crypto_tfm base; ++}; ++ ++struct crypto_template; ++ ++struct crypto_spawn; ++ ++struct crypto_instance { ++ struct crypto_alg alg; ++ struct crypto_template *tmpl; ++ union { ++ struct hlist_node list; ++ struct crypto_spawn *spawns; ++ }; ++ void *__ctx[0]; ++}; ++ ++struct crypto_spawn { ++ struct list_head list; ++ struct crypto_alg *alg; ++ union { ++ struct crypto_instance *inst; ++ struct crypto_spawn *next; ++ }; ++ const struct crypto_type *frontend; ++ u32 mask; ++ bool dead; ++ bool registered; ++}; ++ ++struct rtattr; ++ ++struct crypto_template { ++ struct list_head list; ++ struct hlist_head instances; ++ struct module *module; ++ int (*create)(struct crypto_template *, struct rtattr **); ++ char name[128]; ++}; ++ ++struct aead_request { ++ struct crypto_async_request base; ++ unsigned int assoclen; ++ unsigned int cryptlen; ++ u8 *iv; ++ struct scatterlist *src; ++ struct scatterlist *dst; ++ void *__ctx[0]; ++}; ++ ++struct crypto_aead; ++ ++struct aead_alg { ++ int (*setkey)(struct crypto_aead *, const u8 *, unsigned int); ++ int (*setauthsize)(struct crypto_aead *, unsigned int); ++ int (*encrypt)(struct aead_request *); ++ int (*decrypt)(struct aead_request *); ++ int (*init)(struct crypto_aead *); ++ void (*exit)(struct crypto_aead *); ++ unsigned int ivsize; ++ unsigned int maxauthsize; ++ unsigned int chunksize; ++ struct crypto_alg base; ++}; ++ ++struct crypto_aead { ++ unsigned int authsize; ++ unsigned int reqsize; ++ struct crypto_tfm base; ++}; ++ ++struct rtattr { ++ short unsigned int rta_len; ++ short unsigned int rta_type; ++}; ++ ++struct aead_instance { ++ void (*free)(struct aead_instance *); ++ union { ++ struct { ++ char head[64]; ++ struct crypto_instance base; ++ } s; ++ struct aead_alg alg; ++ }; ++}; ++ ++struct crypto_aead_spawn { ++ struct crypto_spawn base; ++}; ++ ++struct crypto_sync_skcipher; ++ ++struct aead_geniv_ctx { ++ spinlock_t lock; ++ struct crypto_aead *child; ++ struct crypto_sync_skcipher *sknull; ++ u8 salt[0]; ++}; ++ ++struct crypto_istat_rng { ++ atomic64_t generate_cnt; ++ atomic64_t generate_tlen; ++ atomic64_t seed_cnt; ++ atomic64_t err_cnt; ++}; ++ ++struct crypto_rng; ++ ++struct rng_alg { ++ int (*generate)(struct crypto_rng *, const u8 *, unsigned int, u8 *, unsigned int); ++ int (*seed)(struct crypto_rng *, const u8 *, unsigned int); ++ void (*set_ent)(struct crypto_rng *, const u8 *, unsigned int); ++ unsigned int seedsize; ++ struct crypto_alg base; ++}; ++ ++struct crypto_rng { ++ struct crypto_tfm base; ++}; ++ ++struct akcipher_request { ++ struct crypto_async_request base; ++ struct scatterlist *src; ++ struct scatterlist *dst; ++ unsigned int src_len; ++ unsigned int dst_len; ++ void *__ctx[0]; ++}; ++ ++struct crypto_akcipher { ++ unsigned int reqsize; ++ struct crypto_tfm base; ++}; ++ ++struct crypto_istat_akcipher { ++ atomic64_t encrypt_cnt; ++ atomic64_t encrypt_tlen; ++ atomic64_t decrypt_cnt; ++ atomic64_t decrypt_tlen; ++ atomic64_t verify_cnt; ++ atomic64_t sign_cnt; ++ atomic64_t err_cnt; ++}; ++ ++struct akcipher_alg { ++ int (*sign)(struct akcipher_request *); ++ int (*verify)(struct akcipher_request *); ++ int (*encrypt)(struct akcipher_request *); ++ int (*decrypt)(struct akcipher_request *); ++ int (*set_pub_key)(struct crypto_akcipher *, const void *, unsigned int); ++ int (*set_priv_key)(struct crypto_akcipher *, const void *, unsigned int); ++ unsigned int (*max_size)(struct crypto_akcipher *); ++ int (*init)(struct crypto_akcipher *); ++ void (*exit)(struct crypto_akcipher *); ++ struct crypto_alg base; ++}; ++ ++struct akcipher_instance { ++ void (*free)(struct akcipher_instance *); ++ union { ++ struct { ++ char head[72]; ++ struct crypto_instance base; ++ } s; ++ struct akcipher_alg alg; ++ }; ++}; ++ ++struct crypto_akcipher_spawn { ++ struct crypto_spawn base; ++}; ++ ++enum crypto_attr_type_t { ++ CRYPTOCFGA_UNSPEC = 0, ++ CRYPTOCFGA_PRIORITY_VAL = 1, ++ CRYPTOCFGA_REPORT_LARVAL = 2, ++ CRYPTOCFGA_REPORT_HASH = 3, ++ CRYPTOCFGA_REPORT_BLKCIPHER = 4, ++ CRYPTOCFGA_REPORT_AEAD = 5, ++ CRYPTOCFGA_REPORT_COMPRESS = 6, ++ CRYPTOCFGA_REPORT_RNG = 7, ++ CRYPTOCFGA_REPORT_CIPHER = 8, ++ CRYPTOCFGA_REPORT_AKCIPHER = 9, ++ CRYPTOCFGA_REPORT_KPP = 10, ++ CRYPTOCFGA_REPORT_ACOMP = 11, ++ CRYPTOCFGA_STAT_LARVAL = 12, ++ CRYPTOCFGA_STAT_HASH = 13, ++ CRYPTOCFGA_STAT_BLKCIPHER = 14, ++ CRYPTOCFGA_STAT_AEAD = 15, ++ CRYPTOCFGA_STAT_COMPRESS = 16, ++ CRYPTOCFGA_STAT_RNG = 17, ++ CRYPTOCFGA_STAT_CIPHER = 18, ++ CRYPTOCFGA_STAT_AKCIPHER = 19, ++ CRYPTOCFGA_STAT_KPP = 20, ++ CRYPTOCFGA_STAT_ACOMP = 21, ++ __CRYPTOCFGA_MAX = 22, ++}; ++ ++struct crypto_report_akcipher { ++ char type[64]; ++}; ++ ++struct kpp_request { ++ struct crypto_async_request base; ++ struct scatterlist *src; ++ struct scatterlist *dst; ++ unsigned int src_len; ++ unsigned int dst_len; ++ void *__ctx[0]; ++}; ++ ++struct crypto_kpp { ++ unsigned int reqsize; ++ struct crypto_tfm base; ++}; ++ ++struct kpp_alg { ++ int (*set_secret)(struct crypto_kpp *, const void *, unsigned int); ++ int (*generate_public_key)(struct kpp_request *); ++ int (*compute_shared_secret)(struct kpp_request *); ++ unsigned int (*max_size)(struct crypto_kpp *); ++ int (*init)(struct crypto_kpp *); ++ void (*exit)(struct crypto_kpp *); ++ struct crypto_alg base; ++}; ++ ++struct kpp_instance { ++ void (*free)(struct kpp_instance *); ++ union { ++ struct { ++ char head[48]; ++ struct crypto_instance base; ++ } s; ++ struct kpp_alg alg; ++ }; ++}; ++ ++struct crypto_kpp_spawn { ++ struct crypto_spawn base; ++}; ++ ++struct dh { ++ const void *key; ++ const void *p; ++ const void *g; ++ unsigned int key_size; ++ unsigned int p_size; ++ unsigned int g_size; ++}; ++ ++typedef long unsigned int mpi_limb_t; ++ ++struct gcry_mpi { ++ int alloced; ++ int nlimbs; ++ int nbits; ++ int sign; ++ unsigned int flags; ++ mpi_limb_t *d; ++}; ++ ++typedef struct gcry_mpi *MPI; ++ ++struct dh_ctx { ++ MPI p; ++ MPI g; ++ MPI xa; ++}; ++ ++struct dh_safe_prime { ++ unsigned int max_strength; ++ unsigned int p_size; ++ const char *p; ++}; ++ ++struct dh_safe_prime_instance_ctx { ++ struct crypto_kpp_spawn dh_spawn; ++ const struct dh_safe_prime *safe_prime; ++}; ++ ++struct dh_safe_prime_tfm_ctx { ++ struct crypto_kpp *dh_tfm; ++}; ++ ++enum asn1_class { ++ ASN1_UNIV = 0, ++ ASN1_APPL = 1, ++ ASN1_CONT = 2, ++ ASN1_PRIV = 3, ++}; ++ ++enum asn1_method { ++ ASN1_PRIM = 0, ++ ASN1_CONS = 1, ++}; ++ ++enum asn1_tag { ++ ASN1_EOC = 0, ++ ASN1_BOOL = 1, ++ ASN1_INT = 2, ++ ASN1_BTS = 3, ++ ASN1_OTS = 4, ++ ASN1_NULL = 5, ++ ASN1_OID = 6, ++ ASN1_ODE = 7, ++ ASN1_EXT = 8, ++ ASN1_REAL = 9, ++ ASN1_ENUM = 10, ++ ASN1_EPDV = 11, ++ ASN1_UTF8STR = 12, ++ ASN1_RELOID = 13, ++ ASN1_SEQ = 16, ++ ASN1_SET = 17, ++ ASN1_NUMSTR = 18, ++ ASN1_PRNSTR = 19, ++ ASN1_TEXSTR = 20, ++ ASN1_VIDSTR = 21, ++ ASN1_IA5STR = 22, ++ ASN1_UNITIM = 23, ++ ASN1_GENTIM = 24, ++ ASN1_GRASTR = 25, ++ ASN1_VISSTR = 26, ++ ASN1_GENSTR = 27, ++ ASN1_UNISTR = 28, ++ ASN1_CHRSTR = 29, ++ ASN1_BMPSTR = 30, ++ ASN1_LONG_TAG = 31, ++}; ++ ++enum asn1_opcode { ++ ASN1_OP_MATCH = 0, ++ ASN1_OP_MATCH_OR_SKIP = 1, ++ ASN1_OP_MATCH_ACT = 2, ++ ASN1_OP_MATCH_ACT_OR_SKIP = 3, ++ ASN1_OP_MATCH_JUMP = 4, ++ ASN1_OP_MATCH_JUMP_OR_SKIP = 5, ++ ASN1_OP_MATCH_ANY = 8, ++ ASN1_OP_MATCH_ANY_OR_SKIP = 9, ++ ASN1_OP_MATCH_ANY_ACT = 10, ++ ASN1_OP_MATCH_ANY_ACT_OR_SKIP = 11, ++ ASN1_OP_COND_MATCH_OR_SKIP = 17, ++ ASN1_OP_COND_MATCH_ACT_OR_SKIP = 19, ++ ASN1_OP_COND_MATCH_JUMP_OR_SKIP = 21, ++ ASN1_OP_COND_MATCH_ANY = 24, ++ ASN1_OP_COND_MATCH_ANY_OR_SKIP = 25, ++ ASN1_OP_COND_MATCH_ANY_ACT = 26, ++ ASN1_OP_COND_MATCH_ANY_ACT_OR_SKIP = 27, ++ ASN1_OP_COND_FAIL = 28, ++ ASN1_OP_COMPLETE = 29, ++ ASN1_OP_ACT = 30, ++ ASN1_OP_MAYBE_ACT = 31, ++ ASN1_OP_END_SEQ = 32, ++ ASN1_OP_END_SET = 33, ++ ASN1_OP_END_SEQ_OF = 34, ++ ASN1_OP_END_SET_OF = 35, ++ ASN1_OP_END_SEQ_ACT = 36, ++ ASN1_OP_END_SET_ACT = 37, ++ ASN1_OP_END_SEQ_OF_ACT = 38, ++ ASN1_OP_END_SET_OF_ACT = 39, ++ ASN1_OP_RETURN = 40, ++ ASN1_OP__NR = 41, ++}; ++ ++enum ecdsasignature_actions { ++ ACT_ecdsa_get_signature_r = 0, ++ ACT_ecdsa_get_signature_s = 1, ++ NR__ecdsasignature_actions = 2, ++}; ++ ++struct crypto_scomp { ++ struct crypto_tfm base; ++}; ++ ++struct scomp_alg { ++ void * (*alloc_ctx)(struct crypto_scomp *); ++ void (*free_ctx)(struct crypto_scomp *, void *); ++ int (*compress)(struct crypto_scomp *, const u8 *, unsigned int, u8 *, unsigned int *, void *); ++ int (*decompress)(struct crypto_scomp *, const u8 *, unsigned int, u8 *, unsigned int *, void *); ++ union { ++ struct { ++ struct crypto_alg base; ++ }; ++ struct comp_alg_common calg; ++ }; ++}; ++ ++struct crypto_report_comp { ++ char type[64]; ++}; ++ ++struct scomp_scratch { ++ spinlock_t lock; ++ void *src; ++ void *dst; ++}; ++ ++struct crypto_cipher { ++ struct crypto_tfm base; ++}; ++ ++struct crypto_cipher_spawn { ++ struct crypto_spawn base; ++}; ++ ++struct shash_instance { ++ void (*free)(struct shash_instance *); ++ union { ++ struct { ++ char head[104]; ++ struct crypto_instance base; ++ } s; ++ struct shash_alg alg; ++ }; ++}; ++ ++struct cmac_tfm_ctx { ++ struct crypto_cipher *child; ++ u8 ctx[0]; ++}; ++ ++struct cmac_desc_ctx { ++ unsigned int len; ++ u8 ctx[0]; ++}; ++ ++struct scatter_walk { ++ struct scatterlist *sg; ++ unsigned int offset; ++}; ++ ++struct skcipher_request { ++ unsigned int cryptlen; ++ u8 *iv; ++ struct scatterlist *src; ++ struct scatterlist *dst; ++ struct crypto_async_request base; ++ void *__ctx[0]; ++}; ++ ++struct crypto_sync_skcipher { ++ struct crypto_skcipher base; ++}; ++ ++struct skcipher_alg { ++ int (*setkey)(struct crypto_skcipher *, const u8 *, unsigned int); ++ int (*encrypt)(struct skcipher_request *); ++ int (*decrypt)(struct skcipher_request *); ++ int (*init)(struct crypto_skcipher *); ++ void (*exit)(struct crypto_skcipher *); ++ unsigned int min_keysize; ++ unsigned int max_keysize; ++ unsigned int ivsize; ++ unsigned int chunksize; ++ unsigned int walksize; ++ struct crypto_alg base; ++}; ++ ++struct skcipher_walk { ++ union { ++ struct { ++ struct page *page; ++ long unsigned int offset; ++ } phys; ++ struct { ++ u8 *page; ++ void *addr; ++ } virt; ++ } src; ++ union { ++ struct { ++ struct page *page; ++ long unsigned int offset; ++ } phys; ++ struct { ++ u8 *page; ++ void *addr; ++ } virt; ++ } dst; ++ struct scatter_walk in; ++ unsigned int nbytes; ++ struct scatter_walk out; ++ unsigned int total; ++ struct list_head buffers; ++ u8 *page; ++ u8 *buffer; ++ u8 *oiv; ++ void *iv; ++ unsigned int ivsize; ++ int flags; ++ unsigned int blocksize; ++ unsigned int stride; ++ unsigned int alignmask; ++}; ++ ++struct sha1_state { ++ u32 state[5]; ++ u64 count; ++ u8 buffer[64]; ++}; ++ ++typedef void sha1_block_fn(struct sha1_state *, const u8 *, int); ++ ++enum blake2b_lengths { ++ BLAKE2B_BLOCK_SIZE = 128, ++ BLAKE2B_HASH_SIZE = 64, ++ BLAKE2B_KEY_SIZE = 64, ++ BLAKE2B_160_HASH_SIZE = 20, ++ BLAKE2B_256_HASH_SIZE = 32, ++ BLAKE2B_384_HASH_SIZE = 48, ++ BLAKE2B_512_HASH_SIZE = 64, ++}; ++ ++struct blake2b_state { ++ u64 h[8]; ++ u64 t[2]; ++ u64 f[2]; ++ u8 buf[128]; ++ unsigned int buflen; ++ unsigned int outlen; ++}; ++ ++enum blake2b_iv { ++ BLAKE2B_IV0 = 7640891576956012808ULL, ++ BLAKE2B_IV1 = 13503953896175478587ULL, ++ BLAKE2B_IV2 = 4354685564936845355ULL, ++ BLAKE2B_IV3 = 11912009170470909681ULL, ++ BLAKE2B_IV4 = 5840696475078001361ULL, ++ BLAKE2B_IV5 = 11170449401992604703ULL, ++ BLAKE2B_IV6 = 2270897969802886507ULL, ++ BLAKE2B_IV7 = 6620516959819538809ULL, ++}; ++ ++typedef void (*blake2b_compress_t)(struct blake2b_state *, const u8 *, size_t, u32); ++ ++struct blake2b_tfm_ctx { ++ u8 key[64]; ++ unsigned int keylen; ++}; ++ ++struct skcipher_instance { ++ void (*free)(struct skcipher_instance *); ++ union { ++ struct { ++ char head[64]; ++ struct crypto_instance base; ++ } s; ++ struct skcipher_alg alg; ++ }; ++}; ++ ++struct skcipher_ctx_simple { ++ struct crypto_cipher *cipher; ++}; ++ ++struct crypto_skcipher_spawn { ++ struct crypto_spawn base; ++}; ++ ++struct crypto_rfc3686_ctx { ++ struct crypto_skcipher *child; ++ u8 nonce[4]; ++}; ++ ++struct crypto_rfc3686_req_ctx { ++ u8 iv[16]; ++ struct skcipher_request subreq; ++}; ++ ++typedef unsigned char Byte; ++ ++typedef long unsigned int uLong; ++ ++struct internal_state; ++ ++struct z_stream_s { ++ const Byte *next_in; ++ uLong avail_in; ++ uLong total_in; ++ Byte *next_out; ++ uLong avail_out; ++ uLong total_out; ++ char *msg; ++ struct internal_state *state; ++ void *workspace; ++ int data_type; ++ uLong adler; ++ uLong reserved; ++}; ++ ++struct internal_state { ++ int dummy; ++}; ++ ++typedef struct z_stream_s z_stream; ++ ++typedef z_stream *z_streamp; ++ ++struct deflate_ctx { ++ struct z_stream_s comp_stream; ++ struct z_stream_s decomp_stream; ++}; ++ ++struct lzo_ctx { ++ void *lzo_comp_mem; ++}; ++ ++struct crypto_report_rng { ++ char type[64]; ++ unsigned int seedsize; ++}; ++ ++struct af_alg_type; ++ ++struct alg_sock { ++ struct sock sk; ++ struct sock *parent; ++ atomic_t refcnt; ++ atomic_t nokey_refcnt; ++ const struct af_alg_type *type; ++ void *private; ++}; ++ ++struct af_alg_type { ++ void * (*bind)(const char *, u32, u32); ++ void (*release)(void *); ++ int (*setkey)(void *, const u8 *, unsigned int); ++ int (*setentropy)(void *, sockptr_t, unsigned int); ++ int (*accept)(void *, struct sock *); ++ int (*accept_nokey)(void *, struct sock *); ++ int (*setauthsize)(void *, unsigned int); ++ struct proto_ops *ops; ++ struct proto_ops *ops_nokey; ++ struct module *owner; ++ char name[14]; ++}; ++ ++struct af_alg_sgl { ++ struct scatterlist sg[17]; ++ struct page *pages[16]; ++ unsigned int npages; ++}; ++ ++struct hash_ctx { ++ struct af_alg_sgl sgl; ++ u8 *result; ++ struct crypto_wait wait; ++ unsigned int len; ++ bool more; ++ struct ahash_request req; ++}; ++ ++struct af_alg_tsgl { ++ struct list_head list; ++ unsigned int cur; ++ struct scatterlist sg[0]; ++}; ++ ++struct af_alg_rsgl { ++ struct af_alg_sgl sgl; ++ struct list_head list; ++ size_t sg_num_bytes; ++}; ++ ++struct af_alg_async_req { ++ struct kiocb *iocb; ++ struct sock *sk; ++ struct af_alg_rsgl first_rsgl; ++ struct af_alg_rsgl *last_rsgl; ++ struct list_head rsgl_list; ++ struct scatterlist *tsgl; ++ unsigned int tsgl_entries; ++ unsigned int outlen; ++ unsigned int areqlen; ++ union { ++ struct aead_request aead_req; ++ struct skcipher_request skcipher_req; ++ } cra_u; ++}; ++ ++struct af_alg_ctx { ++ struct list_head tsgl_list; ++ void *iv; ++ size_t aead_assoclen; ++ struct crypto_wait wait; ++ size_t used; ++ atomic_t rcvused; ++ bool more; ++ bool merge; ++ bool enc; ++ bool init; ++ unsigned int len; ++}; ++ ++struct aead_tfm { ++ struct crypto_aead *aead; ++ struct crypto_sync_skcipher *null_tfm; ++}; ++ ++enum x509_actions { ++ ACT_x509_extract_key_data = 0, ++ ACT_x509_extract_name_segment = 1, ++ ACT_x509_note_OID = 2, ++ ACT_x509_note_issuer = 3, ++ ACT_x509_note_not_after = 4, ++ ACT_x509_note_not_before = 5, ++ ACT_x509_note_params = 6, ++ ACT_x509_note_serial = 7, ++ ACT_x509_note_sig_algo = 8, ++ ACT_x509_note_signature = 9, ++ ACT_x509_note_subject = 10, ++ ACT_x509_note_tbs_certificate = 11, ++ ACT_x509_process_extension = 12, ++ NR__x509_actions = 13, ++}; ++ ++struct x509_certificate { ++ struct x509_certificate *next; ++ struct x509_certificate *signer; ++ struct public_key *pub; ++ struct public_key_signature *sig; ++ char *issuer; ++ char *subject; ++ struct asymmetric_key_id *id; ++ struct asymmetric_key_id *skid; ++ time64_t valid_from; ++ time64_t valid_to; ++ const void *tbs; ++ unsigned int tbs_size; ++ unsigned int raw_sig_size; ++ const void *raw_sig; ++ const void *raw_serial; ++ unsigned int raw_serial_size; ++ unsigned int raw_issuer_size; ++ const void *raw_issuer; ++ const void *raw_subject; ++ unsigned int raw_subject_size; ++ unsigned int raw_skid_size; ++ const void *raw_skid; ++ unsigned int index; ++ bool seen; ++ bool verified; ++ bool self_signed; ++ bool unsupported_sig; ++ bool blacklisted; ++}; ++ ++struct x509_parse_context { ++ struct x509_certificate *cert; ++ long unsigned int data; ++ const void *key; ++ size_t key_size; ++ const void *params; ++ size_t params_size; ++ enum OID key_algo; ++ enum OID last_oid; ++ enum OID sig_algo; ++ u8 o_size; ++ u8 cn_size; ++ u8 email_size; ++ u16 o_offset; ++ u16 cn_offset; ++ u16 email_offset; ++ unsigned int raw_akid_size; ++ const void *raw_akid; ++ const void *akid_raw_issuer; ++ unsigned int akid_raw_issuer_size; ++}; ++ ++enum pkcs7_actions { ++ ACT_pkcs7_check_content_type = 0, ++ ACT_pkcs7_extract_cert = 1, ++ ACT_pkcs7_note_OID = 2, ++ ACT_pkcs7_note_certificate_list = 3, ++ ACT_pkcs7_note_content = 4, ++ ACT_pkcs7_note_data = 5, ++ ACT_pkcs7_note_signed_info = 6, ++ ACT_pkcs7_note_signeddata_version = 7, ++ ACT_pkcs7_note_signerinfo_version = 8, ++ ACT_pkcs7_sig_note_authenticated_attr = 9, ++ ACT_pkcs7_sig_note_digest_algo = 10, ++ ACT_pkcs7_sig_note_issuer = 11, ++ ACT_pkcs7_sig_note_pkey_algo = 12, ++ ACT_pkcs7_sig_note_serial = 13, ++ ACT_pkcs7_sig_note_set_of_authattrs = 14, ++ ACT_pkcs7_sig_note_signature = 15, ++ ACT_pkcs7_sig_note_skid = 16, ++ NR__pkcs7_actions = 17, ++}; ++ ++struct pkcs7_signed_info { ++ struct pkcs7_signed_info *next; ++ struct x509_certificate *signer; ++ unsigned int index; ++ bool unsupported_crypto; ++ bool blacklisted; ++ const void *msgdigest; ++ unsigned int msgdigest_len; ++ unsigned int authattrs_len; ++ const void *authattrs; ++ long unsigned int aa_set; ++ time64_t signing_time; ++ struct public_key_signature *sig; ++}; ++ ++struct pkcs7_message { ++ struct x509_certificate *certs; ++ struct x509_certificate *crl; ++ struct pkcs7_signed_info *signed_infos; ++ u8 version; ++ bool have_authattrs; ++ enum OID data_type; ++ size_t data_len; ++ size_t data_hdrlen; ++ const void *data; ++}; ++ ++struct pkcs7_parse_context { ++ struct pkcs7_message *msg; ++ struct pkcs7_signed_info *sinfo; ++ struct pkcs7_signed_info **ppsinfo; ++ struct x509_certificate *certs; ++ struct x509_certificate **ppcerts; ++ long unsigned int data; ++ enum OID last_oid; ++ unsigned int x509_index; ++ unsigned int sinfo_index; ++ const void *raw_serial; ++ unsigned int raw_serial_size; ++ unsigned int raw_issuer_size; ++ const void *raw_issuer; ++ const void *raw_skid; ++ unsigned int raw_skid_size; ++ bool expect_skid; ++}; ++ ++struct mz_hdr { ++ uint16_t magic; ++ uint16_t lbsize; ++ uint16_t blocks; ++ uint16_t relocs; ++ uint16_t hdrsize; ++ uint16_t min_extra_pps; ++ uint16_t max_extra_pps; ++ uint16_t ss; ++ uint16_t sp; ++ uint16_t checksum; ++ uint16_t ip; ++ uint16_t cs; ++ uint16_t reloc_table_offset; ++ uint16_t overlay_num; ++ uint16_t reserved0[4]; ++ uint16_t oem_id; ++ uint16_t oem_info; ++ uint16_t reserved1[10]; ++ uint32_t peaddr; ++ char message[0]; ++}; ++ ++struct pe_hdr { ++ uint32_t magic; ++ uint16_t machine; ++ uint16_t sections; ++ uint32_t timestamp; ++ uint32_t symbol_table; ++ uint32_t symbols; ++ uint16_t opt_hdr_size; ++ uint16_t flags; ++}; ++ ++struct pe32_opt_hdr { ++ uint16_t magic; ++ uint8_t ld_major; ++ uint8_t ld_minor; ++ uint32_t text_size; ++ uint32_t data_size; ++ uint32_t bss_size; ++ uint32_t entry_point; ++ uint32_t code_base; ++ uint32_t data_base; ++ uint32_t image_base; ++ uint32_t section_align; ++ uint32_t file_align; ++ uint16_t os_major; ++ uint16_t os_minor; ++ uint16_t image_major; ++ uint16_t image_minor; ++ uint16_t subsys_major; ++ uint16_t subsys_minor; ++ uint32_t win32_version; ++ uint32_t image_size; ++ uint32_t header_size; ++ uint32_t csum; ++ uint16_t subsys; ++ uint16_t dll_flags; ++ uint32_t stack_size_req; ++ uint32_t stack_size; ++ uint32_t heap_size_req; ++ uint32_t heap_size; ++ uint32_t loader_flags; ++ uint32_t data_dirs; ++}; ++ ++struct pe32plus_opt_hdr { ++ uint16_t magic; ++ uint8_t ld_major; ++ uint8_t ld_minor; ++ uint32_t text_size; ++ uint32_t data_size; ++ uint32_t bss_size; ++ uint32_t entry_point; ++ uint32_t code_base; ++ uint64_t image_base; ++ uint32_t section_align; ++ uint32_t file_align; ++ uint16_t os_major; ++ uint16_t os_minor; ++ uint16_t image_major; ++ uint16_t image_minor; ++ uint16_t subsys_major; ++ uint16_t subsys_minor; ++ uint32_t win32_version; ++ uint32_t image_size; ++ uint32_t header_size; ++ uint32_t csum; ++ uint16_t subsys; ++ uint16_t dll_flags; ++ uint64_t stack_size_req; ++ uint64_t stack_size; ++ uint64_t heap_size_req; ++ uint64_t heap_size; ++ uint32_t loader_flags; ++ uint32_t data_dirs; ++}; ++ ++struct data_dirent { ++ uint32_t virtual_address; ++ uint32_t size; ++}; ++ ++struct data_directory { ++ struct data_dirent exports; ++ struct data_dirent imports; ++ struct data_dirent resources; ++ struct data_dirent exceptions; ++ struct data_dirent certs; ++ struct data_dirent base_relocations; ++ struct data_dirent debug; ++ struct data_dirent arch; ++ struct data_dirent global_ptr; ++ struct data_dirent tls; ++ struct data_dirent load_config; ++ struct data_dirent bound_imports; ++ struct data_dirent import_addrs; ++ struct data_dirent delay_imports; ++ struct data_dirent clr_runtime_hdr; ++ struct data_dirent reserved; ++}; ++ ++struct section_header { ++ char name[8]; ++ uint32_t virtual_size; ++ uint32_t virtual_address; ++ uint32_t raw_data_size; ++ uint32_t data_addr; ++ uint32_t relocs; ++ uint32_t line_numbers; ++ uint16_t num_relocs; ++ uint16_t num_lin_numbers; ++ uint32_t flags; ++}; ++ ++struct win_certificate { ++ uint32_t length; ++ uint16_t revision; ++ uint16_t cert_type; ++}; ++ ++struct pefile_context { ++ unsigned int header_size; ++ unsigned int image_checksum_offset; ++ unsigned int cert_dirent_offset; ++ unsigned int n_data_dirents; ++ unsigned int n_sections; ++ unsigned int certs_size; ++ unsigned int sig_offset; ++ unsigned int sig_len; ++ const struct section_header *secs; ++ const void *digest; ++ unsigned int digest_len; ++ const char *digest_algo; ++}; ++ ++struct bvec_iter_all { ++ struct bio_vec bv; ++ int idx; ++ unsigned int done; ++}; ++ ++typedef unsigned int iov_iter_extraction_t; ++ ++struct blkg_iostat { ++ u64 bytes[3]; ++ u64 ios[3]; ++}; ++ ++struct blkg_iostat_set { ++ struct u64_stats_sync sync; ++ struct blkcg_gq *blkg; ++ struct llist_node lnode; ++ int lqueued; ++ struct blkg_iostat cur; ++ struct blkg_iostat last; ++}; ++ ++struct blkcg; ++ ++struct blkg_policy_data; ++ ++struct blkcg_gq { ++ struct request_queue *q; ++ struct list_head q_node; ++ struct hlist_node blkcg_node; ++ struct blkcg *blkcg; ++ struct blkcg_gq *parent; ++ struct percpu_ref refcnt; ++ bool online; ++ struct blkg_iostat_set *iostat_cpu; ++ struct blkg_iostat_set iostat; ++ struct blkg_policy_data *pd[6]; ++ spinlock_t async_bio_lock; ++ struct bio_list async_bios; ++ union { ++ struct work_struct async_bio_work; ++ struct work_struct free_work; ++ }; ++ atomic_t use_delay; ++ atomic64_t delay_nsec; ++ atomic64_t delay_start; ++ u64 last_delay; ++ int last_use; ++ struct callback_head callback_head; ++}; ++ ++typedef __u32 blk_mq_req_flags_t; ++ ++struct bio_alloc_cache { ++ struct bio *free_list; ++ struct bio *free_list_irq; ++ unsigned int nr; ++ unsigned int nr_irq; ++}; ++ ++enum rq_qos_id { ++ RQ_QOS_WBT = 0, ++ RQ_QOS_LATENCY = 1, ++ RQ_QOS_COST = 2, ++}; ++ ++struct rq_qos_ops; ++ ++struct rq_qos { ++ const struct rq_qos_ops *ops; ++ struct gendisk *disk; ++ enum rq_qos_id id; ++ struct rq_qos *next; ++ struct dentry *debugfs_dir; ++}; ++ ++enum { ++ sysctl_hung_task_timeout_secs = 0, ++}; ++ ++enum hctx_type { ++ HCTX_TYPE_DEFAULT = 0, ++ HCTX_TYPE_READ = 1, ++ HCTX_TYPE_POLL = 2, ++ HCTX_MAX_TYPES = 3, ++}; ++ ++enum xen_domain_type { ++ XEN_NATIVE = 0, ++ XEN_PV_DOMAIN = 1, ++ XEN_HVM_DOMAIN = 2, ++}; ++ ++struct blk_mq_debugfs_attr { ++ const char *name; ++ umode_t mode; ++ int (*show)(void *, struct seq_file *); ++ ssize_t (*write)(void *, const char *, size_t, loff_t *); ++ const struct seq_operations *seq_ops; ++}; ++ ++struct rq_qos_ops { ++ void (*throttle)(struct rq_qos *, struct bio *); ++ void (*track)(struct rq_qos *, struct request *, struct bio *); ++ void (*merge)(struct rq_qos *, struct request *, struct bio *); ++ void (*issue)(struct rq_qos *, struct request *); ++ void (*requeue)(struct rq_qos *, struct request *); ++ void (*done)(struct rq_qos *, struct request *); ++ void (*done_bio)(struct rq_qos *, struct bio *); ++ void (*cleanup)(struct rq_qos *, struct bio *); ++ void (*queue_depth_changed)(struct rq_qos *); ++ void (*exit)(struct rq_qos *); ++ const struct blk_mq_debugfs_attr *debugfs_attrs; ++}; ++ ++enum blkg_iostat_type { ++ BLKG_IOSTAT_READ = 0, ++ BLKG_IOSTAT_WRITE = 1, ++ BLKG_IOSTAT_DISCARD = 2, ++ BLKG_IOSTAT_NR = 3, ++}; ++ ++struct blkcg_policy_data; ++ ++struct blkcg { ++ struct cgroup_subsys_state css; ++ spinlock_t lock; ++ refcount_t online_pin; ++ struct xarray blkg_tree; ++ struct blkcg_gq *blkg_hint; ++ struct hlist_head blkg_list; ++ struct blkcg_policy_data *cpd[6]; ++ struct list_head all_blkcgs_node; ++ struct llist_head *lhead; ++ char fc_app_id[129]; ++ struct list_head cgwb_list; ++}; ++ ++struct blkg_policy_data { ++ struct blkcg_gq *blkg; ++ int plid; ++ bool online; ++}; ++ ++struct blkcg_policy_data { ++ struct blkcg *blkcg; ++ int plid; ++}; ++ ++struct biovec_slab { ++ int nr_vecs; ++ char *name; ++ struct kmem_cache *slab; ++}; ++ ++struct bio_slab { ++ struct kmem_cache *slab; ++ unsigned int slab_ref; ++ unsigned int slab_size; ++ char name[8]; ++}; ++ ++struct disk_stats { ++ u64 nsecs[4]; ++ long unsigned int sectors[4]; ++ long unsigned int ios[4]; ++ long unsigned int merges[4]; ++ long unsigned int io_ticks; ++ local_t in_flight[2]; ++}; ++ ++enum stat_group { ++ STAT_READ = 0, ++ STAT_WRITE = 1, ++ STAT_DISCARD = 2, ++ STAT_FLUSH = 3, ++ NR_STAT_GROUPS = 4, ++}; ++ ++struct elevator_type; ++ ++struct elevator_queue { ++ struct elevator_type *type; ++ void *elevator_data; ++ struct kobject kobj; ++ struct mutex sysfs_lock; ++ long unsigned int flags; ++ struct hlist_head hash[64]; ++}; ++ ++struct blk_mq_ctxs; ++ ++struct blk_mq_ctx { ++ struct { ++ spinlock_t lock; ++ struct list_head rq_lists[3]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ }; ++ unsigned int cpu; ++ short unsigned int index_hw[3]; ++ struct blk_mq_hw_ctx *hctxs[3]; ++ struct request_queue *queue; ++ struct blk_mq_ctxs *ctxs; ++ struct kobject kobj; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum { ++ BLK_MQ_F_SHOULD_MERGE = 1, ++ BLK_MQ_F_TAG_QUEUE_SHARED = 2, ++ BLK_MQ_F_STACKING = 4, ++ BLK_MQ_F_TAG_HCTX_SHARED = 8, ++ BLK_MQ_F_BLOCKING = 32, ++ BLK_MQ_F_NO_SCHED = 64, ++ BLK_MQ_F_NO_SCHED_BY_DEFAULT = 128, ++ BLK_MQ_F_ALLOC_POLICY_START_BIT = 8, ++ BLK_MQ_F_ALLOC_POLICY_BITS = 1, ++ BLK_MQ_S_STOPPED = 0, ++ BLK_MQ_S_TAG_ACTIVE = 1, ++ BLK_MQ_S_SCHED_RESTART = 2, ++ BLK_MQ_S_INACTIVE = 3, ++ BLK_MQ_MAX_DEPTH = 10240, ++ BLK_MQ_CPU_WORK_BATCH = 8, ++}; ++ ++struct blk_mq_ctxs { ++ struct kobject kobj; ++ struct blk_mq_ctx *queue_ctx; ++}; ++ ++enum { ++ BLK_MQ_NO_TAG = 4294967295, ++ BLK_MQ_TAG_MIN = 1, ++ BLK_MQ_TAG_MAX = 4294967294, ++}; ++ ++typedef unsigned int blk_insert_t; ++ ++struct blk_mq_alloc_data { ++ struct request_queue *q; ++ blk_mq_req_flags_t flags; ++ unsigned int shallow_depth; ++ blk_opf_t cmd_flags; ++ req_flags_t rq_flags; ++ unsigned int nr_tags; ++ struct request **cached_rq; ++ struct blk_mq_ctx *ctx; ++ struct blk_mq_hw_ctx *hctx; ++}; ++ ++enum elv_merge { ++ ELEVATOR_NO_MERGE = 0, ++ ELEVATOR_FRONT_MERGE = 1, ++ ELEVATOR_BACK_MERGE = 2, ++ ELEVATOR_DISCARD_MERGE = 3, ++}; ++ ++struct elevator_mq_ops { ++ int (*init_sched)(struct request_queue *, struct elevator_type *); ++ void (*exit_sched)(struct elevator_queue *); ++ int (*init_hctx)(struct blk_mq_hw_ctx *, unsigned int); ++ void (*exit_hctx)(struct blk_mq_hw_ctx *, unsigned int); ++ void (*depth_updated)(struct blk_mq_hw_ctx *); ++ bool (*allow_merge)(struct request_queue *, struct request *, struct bio *); ++ bool (*bio_merge)(struct request_queue *, struct bio *, unsigned int); ++ int (*request_merge)(struct request_queue *, struct request **, struct bio *); ++ void (*request_merged)(struct request_queue *, struct request *, enum elv_merge); ++ void (*requests_merged)(struct request_queue *, struct request *, struct request *); ++ void (*limit_depth)(blk_opf_t, struct blk_mq_alloc_data *); ++ void (*prepare_request)(struct request *); ++ void (*finish_request)(struct request *); ++ void (*insert_requests)(struct blk_mq_hw_ctx *, struct list_head *, blk_insert_t); ++ struct request * (*dispatch_request)(struct blk_mq_hw_ctx *); ++ bool (*has_work)(struct blk_mq_hw_ctx *); ++ void (*completed_request)(struct request *, u64); ++ void (*requeue_request)(struct request *); ++ struct request * (*former_request)(struct request_queue *, struct request *); ++ struct request * (*next_request)(struct request_queue *, struct request *); ++ void (*init_icq)(struct io_cq *); ++ void (*exit_icq)(struct io_cq *); ++}; ++ ++struct elv_fs_entry; ++ ++struct elevator_type { ++ struct kmem_cache *icq_cache; ++ struct elevator_mq_ops ops; ++ size_t icq_size; ++ size_t icq_align; ++ struct elv_fs_entry *elevator_attrs; ++ const char *elevator_name; ++ const char *elevator_alias; ++ const unsigned int elevator_features; ++ struct module *elevator_owner; ++ const struct blk_mq_debugfs_attr *queue_debugfs_attrs; ++ const struct blk_mq_debugfs_attr *hctx_debugfs_attrs; ++ char icq_cache_name[22]; ++ struct list_head list; ++}; ++ ++struct elv_fs_entry { ++ struct attribute attr; ++ ssize_t (*show)(struct elevator_queue *, char *); ++ ssize_t (*store)(struct elevator_queue *, const char *, size_t); ++}; ++ ++enum { ++ REQ_FSEQ_PREFLUSH = 1, ++ REQ_FSEQ_DATA = 2, ++ REQ_FSEQ_POSTFLUSH = 4, ++ REQ_FSEQ_DONE = 8, ++ REQ_FSEQ_ACTIONS = 7, ++ FLUSH_PENDING_TIMEOUT = 500, ++}; ++ ++struct rq_map_data { ++ struct page **pages; ++ long unsigned int offset; ++ short unsigned int page_order; ++ short unsigned int nr_entries; ++ bool null_mapped; ++ bool from_user; ++}; ++ ++struct bio_map_data { ++ bool is_our_pages: 1; ++ bool is_null_mapped: 1; ++ struct iov_iter iter; ++ struct iovec iov[0]; ++}; ++ ++enum { ++ IOPRIO_CLASS_NONE = 0, ++ IOPRIO_CLASS_RT = 1, ++ IOPRIO_CLASS_BE = 2, ++ IOPRIO_CLASS_IDLE = 3, ++}; ++ ++enum rpm_status { ++ RPM_INVALID = -1, ++ RPM_ACTIVE = 0, ++ RPM_RESUMING = 1, ++ RPM_SUSPENDED = 2, ++ RPM_SUSPENDING = 3, ++}; ++ ++typedef bool (*sb_for_each_fn)(struct sbitmap *, unsigned int, void *); ++ ++typedef bool busy_tag_iter_fn(struct request *, void *); ++ ++enum { ++ BLK_MQ_REQ_NOWAIT = 1, ++ BLK_MQ_REQ_RESERVED = 2, ++ BLK_MQ_REQ_PM = 4, ++}; ++ ++enum { ++ BLK_MQ_UNIQUE_TAG_BITS = 16, ++ BLK_MQ_UNIQUE_TAG_MASK = 65535, ++}; ++ ++struct blk_integrity_iter { ++ void *prot_buf; ++ void *data_buf; ++ sector_t seed; ++ unsigned int data_size; ++ short unsigned int interval; ++ unsigned char tuple_size; ++ const char *disk_name; ++}; ++ ++struct mq_inflight { ++ struct block_device *part; ++ unsigned int inflight[2]; ++}; ++ ++struct blk_rq_wait { ++ struct completion done; ++ blk_status_t ret; ++}; ++ ++struct blk_expired_data { ++ bool has_timedout_rq; ++ long unsigned int next; ++ long unsigned int timeout_start; ++}; ++ ++struct flush_busy_ctx_data { ++ struct blk_mq_hw_ctx *hctx; ++ struct list_head *list; ++}; ++ ++struct dispatch_rq_data { ++ struct blk_mq_hw_ctx *hctx; ++ struct request *rq; ++}; ++ ++enum prep_dispatch { ++ PREP_DISPATCH_OK = 0, ++ PREP_DISPATCH_NO_TAG = 1, ++ PREP_DISPATCH_NO_BUDGET = 2, ++}; ++ ++struct rq_iter_data { ++ struct blk_mq_hw_ctx *hctx; ++ bool has_rq; ++}; ++ ++struct blk_mq_qe_pair { ++ struct list_head node; ++ struct request_queue *q; ++ struct elevator_type *type; ++}; ++ ++struct badblocks { ++ struct device *dev; ++ int count; ++ int unacked_exist; ++ int shift; ++ u64 *page; ++ int changed; ++ seqlock_t lock; ++ sector_t sector; ++ sector_t size; ++}; ++ ++struct parsed_partitions { ++ struct gendisk *disk; ++ char name[32]; ++ struct { ++ sector_t from; ++ sector_t size; ++ int flags; ++ bool has_info; ++ struct partition_meta_info info; ++ } *parts; ++ int next; ++ int limit; ++ bool access_beyond_eod; ++ char *pp_buf; ++}; ++ ++typedef struct { ++ struct folio *v; ++} Sector; ++ ++struct lvm_rec { ++ char lvm_id[4]; ++ char reserved4[16]; ++ __be32 lvmarea_len; ++ __be32 vgda_len; ++ __be32 vgda_psn[2]; ++ char reserved36[10]; ++ __be16 pp_size; ++ char reserved46[12]; ++ __be16 version; ++}; ++ ++struct vgda { ++ __be32 secs; ++ __be32 usec; ++ char reserved8[16]; ++ __be16 numlvs; ++ __be16 maxlvs; ++ __be16 pp_size; ++ __be16 numpvs; ++ __be16 total_vgdas; ++ __be16 vgda_size; ++}; ++ ++struct lvd { ++ __be16 lv_ix; ++ __be16 res2; ++ __be16 res4; ++ __be16 maxsize; ++ __be16 lv_state; ++ __be16 mirror; ++ __be16 mirror_policy; ++ __be16 num_lps; ++ __be16 res10[8]; ++}; ++ ++struct lvname { ++ char name[64]; ++}; ++ ++struct ppe { ++ __be16 lv_ix; ++ short unsigned int res2; ++ short unsigned int res4; ++ __be16 lp_ix; ++ short unsigned int res8[12]; ++}; ++ ++struct pvd { ++ char reserved0[16]; ++ __be16 pp_count; ++ char reserved18[2]; ++ __be32 psn_part1; ++ char reserved24[8]; ++ struct ppe ppe[1016]; ++}; ++ ++struct lv_info { ++ short unsigned int pps_per_lv; ++ short unsigned int pps_found; ++ unsigned char lv_is_contiguous; ++}; ++ ++typedef struct { ++ __u8 b[16]; ++} guid_t; ++ ++typedef guid_t efi_guid_t; ++ ++struct _gpt_header { ++ __le64 signature; ++ __le32 revision; ++ __le32 header_size; ++ __le32 header_crc32; ++ __le32 reserved1; ++ __le64 my_lba; ++ __le64 alternate_lba; ++ __le64 first_usable_lba; ++ __le64 last_usable_lba; ++ efi_guid_t disk_guid; ++ __le64 partition_entry_lba; ++ __le32 num_partition_entries; ++ __le32 sizeof_partition_entry; ++ __le32 partition_entry_array_crc32; ++} __attribute__((packed)); ++ ++typedef struct _gpt_header gpt_header; ++ ++struct _gpt_entry_attributes { ++ u64 required_to_function: 1; ++ u64 reserved: 47; ++ u64 type_guid_specific: 16; ++}; ++ ++typedef struct _gpt_entry_attributes gpt_entry_attributes; ++ ++struct _gpt_entry { ++ efi_guid_t partition_type_guid; ++ efi_guid_t unique_partition_guid; ++ __le64 starting_lba; ++ __le64 ending_lba; ++ gpt_entry_attributes attributes; ++ __le16 partition_name[36]; ++}; ++ ++typedef struct _gpt_entry gpt_entry; ++ ++struct _gpt_mbr_record { ++ u8 boot_indicator; ++ u8 start_head; ++ u8 start_sector; ++ u8 start_track; ++ u8 os_type; ++ u8 end_head; ++ u8 end_sector; ++ u8 end_track; ++ __le32 starting_lba; ++ __le32 size_in_lba; ++}; ++ ++typedef struct _gpt_mbr_record gpt_mbr_record; ++ ++struct _legacy_mbr { ++ u8 boot_code[440]; ++ __le32 unique_mbr_signature; ++ __le16 unknown; ++ gpt_mbr_record partition_record[4]; ++ __le16 signature; ++} __attribute__((packed)); ++ ++typedef struct _legacy_mbr legacy_mbr; ++ ++struct blk_ia_range_sysfs_entry { ++ struct attribute attr; ++ ssize_t (*show)(struct blk_independent_access_range *, char *); ++}; ++ ++struct klist_node; ++ ++struct klist { ++ spinlock_t k_lock; ++ struct list_head k_list; ++ void (*get)(struct klist_node *); ++ void (*put)(struct klist_node *); ++}; ++ ++struct klist_node { ++ void *n_klist; ++ struct list_head n_node; ++ struct kref n_ref; ++}; ++ ++struct klist_iter { ++ struct klist *i_klist; ++ struct klist_node *i_cur; ++}; ++ ++struct subsys_private; ++ ++struct class_dev_iter { ++ struct klist_iter ki; ++ const struct device_type *type; ++ struct subsys_private *sp; ++}; ++ ++typedef struct blkcg_policy_data *blkcg_pol_alloc_cpd_fn(gfp_t); ++ ++typedef void blkcg_pol_free_cpd_fn(struct blkcg_policy_data *); ++ ++typedef struct blkg_policy_data *blkcg_pol_alloc_pd_fn(struct gendisk *, struct blkcg *, gfp_t); ++ ++typedef void blkcg_pol_init_pd_fn(struct blkg_policy_data *); ++ ++typedef void blkcg_pol_online_pd_fn(struct blkg_policy_data *); ++ ++typedef void blkcg_pol_offline_pd_fn(struct blkg_policy_data *); ++ ++typedef void blkcg_pol_free_pd_fn(struct blkg_policy_data *); ++ ++typedef void blkcg_pol_reset_pd_stats_fn(struct blkg_policy_data *); ++ ++typedef void blkcg_pol_stat_pd_fn(struct blkg_policy_data *, struct seq_file *); ++ ++struct blkcg_policy { ++ int plid; ++ struct cftype *dfl_cftypes; ++ struct cftype *legacy_cftypes; ++ blkcg_pol_alloc_cpd_fn *cpd_alloc_fn; ++ blkcg_pol_free_cpd_fn *cpd_free_fn; ++ blkcg_pol_alloc_pd_fn *pd_alloc_fn; ++ blkcg_pol_init_pd_fn *pd_init_fn; ++ blkcg_pol_online_pd_fn *pd_online_fn; ++ blkcg_pol_offline_pd_fn *pd_offline_fn; ++ blkcg_pol_free_pd_fn *pd_free_fn; ++ blkcg_pol_reset_pd_stats_fn *pd_reset_stats_fn; ++ blkcg_pol_stat_pd_fn *pd_stat_fn; ++}; ++ ++struct blkg_conf_ctx { ++ char *input; ++ char *body; ++ struct block_device *bdev; ++ struct blkcg_gq *blkg; ++}; ++ ++enum blkg_rwstat_type { ++ BLKG_RWSTAT_READ = 0, ++ BLKG_RWSTAT_WRITE = 1, ++ BLKG_RWSTAT_SYNC = 2, ++ BLKG_RWSTAT_ASYNC = 3, ++ BLKG_RWSTAT_DISCARD = 4, ++ BLKG_RWSTAT_NR = 5, ++ BLKG_RWSTAT_TOTAL = 5, ++}; ++ ++enum { ++ LIMIT_LOW = 0, ++ LIMIT_MAX = 1, ++ LIMIT_CNT = 2, ++}; ++ ++enum { ++ MILLION = 1000000, ++ MIN_PERIOD = 1000, ++ MAX_PERIOD = 1000000, ++ MARGIN_MIN_PCT = 10, ++ MARGIN_LOW_PCT = 20, ++ MARGIN_TARGET_PCT = 50, ++ INUSE_ADJ_STEP_PCT = 25, ++ TIMER_SLACK_PCT = 1, ++ WEIGHT_ONE = 65536, ++}; ++ ++enum { ++ VTIME_PER_SEC_SHIFT = 37ULL, ++ VTIME_PER_SEC = 137438953472ULL, ++ VTIME_PER_USEC = 137438ULL, ++ VTIME_PER_NSEC = 137ULL, ++ VRATE_MIN_PPM = 10000ULL, ++ VRATE_MAX_PPM = 100000000ULL, ++ VRATE_MIN = 1374ULL, ++ VRATE_CLAMP_ADJ_PCT = 4ULL, ++ AUTOP_CYCLE_NSEC = 10000000000ULL, ++}; ++ ++enum { ++ RQ_WAIT_BUSY_PCT = 5, ++ UNBUSY_THR_PCT = 75, ++ MIN_DELAY_THR_PCT = 500, ++ MAX_DELAY_THR_PCT = 25000, ++ MIN_DELAY = 250, ++ MAX_DELAY = 250000, ++ DFGV_USAGE_PCT = 50, ++ DFGV_PERIOD = 100000, ++ MAX_LAGGING_PERIODS = 10, ++ IOC_PAGE_SHIFT = 12, ++ IOC_PAGE_SIZE = 4096, ++ IOC_SECT_TO_PAGE_SHIFT = 3, ++ LCOEF_RANDIO_PAGES = 4096, ++}; ++ ++enum ioc_running { ++ IOC_IDLE = 0, ++ IOC_RUNNING = 1, ++ IOC_STOP = 2, ++}; ++ ++enum { ++ QOS_ENABLE = 0, ++ QOS_CTRL = 1, ++ NR_QOS_CTRL_PARAMS = 2, ++}; ++ ++enum { ++ QOS_RPPM = 0, ++ QOS_RLAT = 1, ++ QOS_WPPM = 2, ++ QOS_WLAT = 3, ++ QOS_MIN = 4, ++ QOS_MAX = 5, ++ NR_QOS_PARAMS = 6, ++}; ++ ++enum { ++ COST_CTRL = 0, ++ COST_MODEL = 1, ++ NR_COST_CTRL_PARAMS = 2, ++}; ++ ++enum { ++ I_LCOEF_RBPS = 0, ++ I_LCOEF_RSEQIOPS = 1, ++ I_LCOEF_RRANDIOPS = 2, ++ I_LCOEF_WBPS = 3, ++ I_LCOEF_WSEQIOPS = 4, ++ I_LCOEF_WRANDIOPS = 5, ++ NR_I_LCOEFS = 6, ++}; ++ ++enum { ++ LCOEF_RPAGE = 0, ++ LCOEF_RSEQIO = 1, ++ LCOEF_RRANDIO = 2, ++ LCOEF_WPAGE = 3, ++ LCOEF_WSEQIO = 4, ++ LCOEF_WRANDIO = 5, ++ NR_LCOEFS = 6, ++}; ++ ++enum { ++ AUTOP_INVALID = 0, ++ AUTOP_HDD = 1, ++ AUTOP_SSD_QD1 = 2, ++ AUTOP_SSD_DFL = 3, ++ AUTOP_SSD_FAST = 4, ++}; ++ ++struct ioc_params { ++ u32 qos[6]; ++ u64 i_lcoefs[6]; ++ u64 lcoefs[6]; ++ u32 too_fast_vrate_pct; ++ u32 too_slow_vrate_pct; ++}; ++ ++struct ioc_margins { ++ s64 min; ++ s64 low; ++ s64 target; ++}; ++ ++struct ioc_missed { ++ local_t nr_met; ++ local_t nr_missed; ++ u32 last_met; ++ u32 last_missed; ++}; ++ ++struct ioc_pcpu_stat { ++ struct ioc_missed missed[2]; ++ local64_t rq_wait_ns; ++ u64 last_rq_wait_ns; ++}; ++ ++struct ioc { ++ struct rq_qos rqos; ++ bool enabled; ++ struct ioc_params params; ++ struct ioc_margins margins; ++ u32 period_us; ++ u32 timer_slack_ns; ++ u64 vrate_min; ++ u64 vrate_max; ++ spinlock_t lock; ++ struct timer_list timer; ++ struct list_head active_iocgs; ++ struct ioc_pcpu_stat *pcpu_stat; ++ enum ioc_running running; ++ atomic64_t vtime_rate; ++ u64 vtime_base_rate; ++ s64 vtime_err; ++ seqcount_spinlock_t period_seqcount; ++ u64 period_at; ++ u64 period_at_vtime; ++ atomic64_t cur_period; ++ int busy_level; ++ bool weights_updated; ++ atomic_t hweight_gen; ++ u64 dfgv_period_at; ++ u64 dfgv_period_rem; ++ u64 dfgv_usage_us_sum; ++ u64 autop_too_fast_at; ++ u64 autop_too_slow_at; ++ int autop_idx; ++ bool user_qos_params: 1; ++ bool user_cost_model: 1; ++}; ++ ++struct iocg_pcpu_stat { ++ local64_t abs_vusage; ++}; ++ ++struct iocg_stat { ++ u64 usage_us; ++ u64 wait_us; ++ u64 indebt_us; ++ u64 indelay_us; ++}; ++ ++struct ioc_gq { ++ struct blkg_policy_data pd; ++ struct ioc *ioc; ++ u32 cfg_weight; ++ u32 weight; ++ u32 active; ++ u32 inuse; ++ u32 last_inuse; ++ s64 saved_margin; ++ sector_t cursor; ++ atomic64_t vtime; ++ atomic64_t done_vtime; ++ u64 abs_vdebt; ++ u64 delay; ++ u64 delay_at; ++ atomic64_t active_period; ++ struct list_head active_list; ++ u64 child_active_sum; ++ u64 child_inuse_sum; ++ u64 child_adjusted_sum; ++ int hweight_gen; ++ u32 hweight_active; ++ u32 hweight_inuse; ++ u32 hweight_donating; ++ u32 hweight_after_donation; ++ struct list_head walk_list; ++ struct list_head surplus_list; ++ struct wait_queue_head waitq; ++ struct hrtimer waitq_timer; ++ u64 activated_at; ++ struct iocg_pcpu_stat *pcpu_stat; ++ struct iocg_stat stat; ++ struct iocg_stat last_stat; ++ u64 last_stat_abs_vusage; ++ u64 usage_delta_us; ++ u64 wait_since; ++ u64 indebt_since; ++ u64 indelay_since; ++ int level; ++ struct ioc_gq *ancestors[0]; ++}; ++ ++struct ioc_cgrp { ++ struct blkcg_policy_data cpd; ++ unsigned int dfl_weight; ++}; ++ ++struct ioc_now { ++ u64 now_ns; ++ u64 now; ++ u64 vnow; ++}; ++ ++struct iocg_wait { ++ struct wait_queue_entry wait; ++ struct bio *bio; ++ u64 abs_cost; ++ bool committed; ++}; ++ ++struct iocg_wake_ctx { ++ struct ioc_gq *iocg; ++ u32 hw_inuse; ++ s64 vbudget; ++}; ++ ++struct trace_event_raw_iocost_iocg_state { ++ struct trace_entry ent; ++ u32 __data_loc_devname; ++ u32 __data_loc_cgroup; ++ u64 now; ++ u64 vnow; ++ u64 vrate; ++ u64 last_period; ++ u64 cur_period; ++ u64 vtime; ++ u32 weight; ++ u32 inuse; ++ u64 hweight_active; ++ u64 hweight_inuse; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_iocg_inuse_update { ++ struct trace_entry ent; ++ u32 __data_loc_devname; ++ u32 __data_loc_cgroup; ++ u64 now; ++ u32 old_inuse; ++ u32 new_inuse; ++ u64 old_hweight_inuse; ++ u64 new_hweight_inuse; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_iocost_ioc_vrate_adj { ++ struct trace_entry ent; ++ u32 __data_loc_devname; ++ u64 old_vrate; ++ u64 new_vrate; ++ int busy_level; ++ u32 read_missed_ppm; ++ u32 write_missed_ppm; ++ u32 rq_wait_pct; ++ int nr_lagging; ++ int nr_shortages; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_iocost_iocg_forgive_debt { ++ struct trace_entry ent; ++ u32 __data_loc_devname; ++ u32 __data_loc_cgroup; ++ u64 now; ++ u64 vnow; ++ u32 usage_pct; ++ u64 old_debt; ++ u64 new_debt; ++ u64 old_delay; ++ u64 new_delay; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_iocost_iocg_state { ++ u32 devname; ++ u32 cgroup; ++}; ++ ++struct trace_event_data_offsets_iocg_inuse_update { ++ u32 devname; ++ u32 cgroup; ++}; ++ ++struct trace_event_data_offsets_iocost_ioc_vrate_adj { ++ u32 devname; ++}; ++ ++struct trace_event_data_offsets_iocost_iocg_forgive_debt { ++ u32 devname; ++ u32 cgroup; ++}; ++ ++typedef void (*btf_trace_iocost_iocg_activate)(void *, struct ioc_gq *, const char *, struct ioc_now *, u64, u64, u64); ++ ++typedef void (*btf_trace_iocost_iocg_idle)(void *, struct ioc_gq *, const char *, struct ioc_now *, u64, u64, u64); ++ ++typedef void (*btf_trace_iocost_inuse_shortage)(void *, struct ioc_gq *, const char *, struct ioc_now *, u32, u32, u64, u64); ++ ++typedef void (*btf_trace_iocost_inuse_transfer)(void *, struct ioc_gq *, const char *, struct ioc_now *, u32, u32, u64, u64); ++ ++typedef void (*btf_trace_iocost_inuse_adjust)(void *, struct ioc_gq *, const char *, struct ioc_now *, u32, u32, u64, u64); ++ ++typedef void (*btf_trace_iocost_ioc_vrate_adj)(void *, struct ioc *, u64, u32 *, u32, int, int); ++ ++typedef void (*btf_trace_iocost_iocg_forgive_debt)(void *, struct ioc_gq *, const char *, struct ioc_now *, u32, u64, u64, u64, u64); ++ ++struct blkg_rwstat { ++ struct percpu_counter cpu_cnt[5]; ++ atomic64_t aux_cnt[5]; ++}; ++ ++struct blkg_rwstat_sample { ++ u64 cnt[5]; ++}; ++ ++struct bfq_entity; ++ ++struct bfq_service_tree { ++ struct rb_root active; ++ struct rb_root idle; ++ struct bfq_entity *first_idle; ++ struct bfq_entity *last_idle; ++ u64 vtime; ++ long unsigned int wsum; ++}; ++ ++struct bfq_sched_data; ++ ++struct bfq_queue; ++ ++struct bfq_entity { ++ struct rb_node rb_node; ++ bool on_st_or_in_serv; ++ u64 start; ++ u64 finish; ++ struct rb_root *tree; ++ u64 min_start; ++ int service; ++ int budget; ++ int allocated; ++ int dev_weight; ++ int weight; ++ int new_weight; ++ int orig_weight; ++ struct bfq_entity *parent; ++ struct bfq_sched_data *my_sched_data; ++ struct bfq_sched_data *sched_data; ++ int prio_changed; ++ bool in_groups_with_pending_reqs; ++ struct bfq_queue *last_bfqq_created; ++}; ++ ++struct bfq_sched_data { ++ struct bfq_entity *in_service_entity; ++ struct bfq_entity *next_in_service; ++ struct bfq_service_tree service_tree[3]; ++ long unsigned int bfq_class_idle_last_service; ++}; ++ ++struct bfq_weight_counter { ++ unsigned int weight; ++ unsigned int num_active; ++ struct rb_node weights_node; ++}; ++ ++struct bfq_ttime { ++ u64 last_end_request; ++ u64 ttime_total; ++ long unsigned int ttime_samples; ++ u64 ttime_mean; ++}; ++ ++struct bfq_data; ++ ++struct bfq_io_cq; ++ ++struct bfq_queue { ++ int ref; ++ int stable_ref; ++ struct bfq_data *bfqd; ++ short unsigned int ioprio; ++ short unsigned int ioprio_class; ++ short unsigned int new_ioprio; ++ short unsigned int new_ioprio_class; ++ u64 last_serv_time_ns; ++ unsigned int inject_limit; ++ long unsigned int decrease_time_jif; ++ struct bfq_queue *new_bfqq; ++ struct rb_node pos_node; ++ struct rb_root *pos_root; ++ struct rb_root sort_list; ++ struct request *next_rq; ++ int queued[2]; ++ int meta_pending; ++ struct list_head fifo; ++ struct bfq_entity entity; ++ struct bfq_weight_counter *weight_counter; ++ int max_budget; ++ long unsigned int budget_timeout; ++ int dispatched; ++ long unsigned int flags; ++ struct list_head bfqq_list; ++ struct bfq_ttime ttime; ++ u64 io_start_time; ++ u64 tot_idle_time; ++ u32 seek_history; ++ struct hlist_node burst_list_node; ++ sector_t last_request_pos; ++ unsigned int requests_within_timer; ++ pid_t pid; ++ struct bfq_io_cq *bic; ++ long unsigned int wr_cur_max_time; ++ long unsigned int soft_rt_next_start; ++ long unsigned int last_wr_start_finish; ++ unsigned int wr_coeff; ++ long unsigned int last_idle_bklogged; ++ long unsigned int service_from_backlogged; ++ long unsigned int service_from_wr; ++ long unsigned int wr_start_at_switch_to_srt; ++ long unsigned int split_time; ++ long unsigned int first_IO_time; ++ long unsigned int creation_time; ++ struct bfq_queue *waker_bfqq; ++ struct bfq_queue *tentative_waker_bfqq; ++ unsigned int num_waker_detections; ++ u64 waker_detection_started; ++ struct hlist_node woken_list_node; ++ struct hlist_head woken_list; ++ unsigned int actuator_idx; ++}; ++ ++struct bfq_group; ++ ++struct bfq_data { ++ struct request_queue *queue; ++ struct list_head dispatch; ++ struct bfq_group *root_group; ++ struct rb_root_cached queue_weights_tree; ++ unsigned int num_groups_with_pending_reqs; ++ unsigned int busy_queues[3]; ++ int wr_busy_queues; ++ int queued; ++ int tot_rq_in_driver; ++ int rq_in_driver[8]; ++ bool nonrot_with_queueing; ++ int max_rq_in_driver; ++ int hw_tag_samples; ++ int hw_tag; ++ int budgets_assigned; ++ struct hrtimer idle_slice_timer; ++ struct bfq_queue *in_service_queue; ++ sector_t last_position; ++ sector_t in_serv_last_pos; ++ u64 last_completion; ++ struct bfq_queue *last_completed_rq_bfqq; ++ struct bfq_queue *last_bfqq_created; ++ u64 last_empty_occupied_ns; ++ bool wait_dispatch; ++ struct request *waited_rq; ++ bool rqs_injected; ++ u64 first_dispatch; ++ u64 last_dispatch; ++ ktime_t last_budget_start; ++ ktime_t last_idling_start; ++ long unsigned int last_idling_start_jiffies; ++ int peak_rate_samples; ++ u32 sequential_samples; ++ u64 tot_sectors_dispatched; ++ u32 last_rq_max_size; ++ u64 delta_from_first; ++ u32 peak_rate; ++ int bfq_max_budget; ++ struct list_head active_list[8]; ++ struct list_head idle_list; ++ u64 bfq_fifo_expire[2]; ++ unsigned int bfq_back_penalty; ++ unsigned int bfq_back_max; ++ u32 bfq_slice_idle; ++ int bfq_user_max_budget; ++ unsigned int bfq_timeout; ++ bool strict_guarantees; ++ long unsigned int last_ins_in_burst; ++ long unsigned int bfq_burst_interval; ++ int burst_size; ++ struct bfq_entity *burst_parent_entity; ++ long unsigned int bfq_large_burst_thresh; ++ bool large_burst; ++ struct hlist_head burst_list; ++ bool low_latency; ++ unsigned int bfq_wr_coeff; ++ unsigned int bfq_wr_rt_max_time; ++ unsigned int bfq_wr_min_idle_time; ++ long unsigned int bfq_wr_min_inter_arr_async; ++ unsigned int bfq_wr_max_softrt_rate; ++ u64 rate_dur_prod; ++ struct bfq_queue oom_bfqq; ++ spinlock_t lock; ++ struct bfq_io_cq *bio_bic; ++ struct bfq_queue *bio_bfqq; ++ unsigned int word_depths[4]; ++ unsigned int full_depth_shift; ++ unsigned int num_actuators; ++ sector_t sector[8]; ++ sector_t nr_sectors[8]; ++ struct blk_independent_access_range ia_ranges[8]; ++ unsigned int actuator_load_threshold; ++}; ++ ++struct bfq_iocq_bfqq_data { ++ bool saved_has_short_ttime; ++ bool saved_IO_bound; ++ u64 saved_io_start_time; ++ u64 saved_tot_idle_time; ++ bool saved_in_large_burst; ++ bool was_in_burst_list; ++ unsigned int saved_weight; ++ long unsigned int saved_wr_coeff; ++ long unsigned int saved_last_wr_start_finish; ++ long unsigned int saved_service_from_wr; ++ long unsigned int saved_wr_start_at_switch_to_srt; ++ unsigned int saved_wr_cur_max_time; ++ struct bfq_ttime saved_ttime; ++ u64 saved_last_serv_time_ns; ++ unsigned int saved_inject_limit; ++ long unsigned int saved_decrease_time_jif; ++ struct bfq_queue *stable_merge_bfqq; ++ bool stably_merged; ++}; ++ ++struct bfq_io_cq { ++ struct io_cq icq; ++ struct bfq_queue *bfqq[16]; ++ int ioprio; ++ uint64_t blkcg_serial_nr; ++ struct bfq_iocq_bfqq_data bfqq_data[8]; ++ unsigned int requests; ++}; ++ ++struct bfqg_stats { ++ struct blkg_rwstat bytes; ++ struct blkg_rwstat ios; ++}; ++ ++struct bfq_group { ++ struct blkg_policy_data pd; ++ char blkg_path[128]; ++ refcount_t ref; ++ struct bfq_entity entity; ++ struct bfq_sched_data sched_data; ++ struct bfq_data *bfqd; ++ struct bfq_queue *async_bfqq[128]; ++ struct bfq_queue *async_idle_bfqq[8]; ++ struct bfq_entity *my_entity; ++ int active_entities; ++ int num_queues_with_pending_reqs; ++ struct rb_root rq_pos_tree; ++ struct bfqg_stats stats; ++}; ++ ++enum bfqq_expiration { ++ BFQQE_TOO_IDLE = 0, ++ BFQQE_BUDGET_TIMEOUT = 1, ++ BFQQE_BUDGET_EXHAUSTED = 2, ++ BFQQE_NO_MORE_REQUESTS = 3, ++ BFQQE_PREEMPTED = 4, ++}; ++ ++struct bfq_group_data { ++ struct blkcg_policy_data pd; ++ unsigned int weight; ++}; ++ ++struct blk_rq_stat { ++ u64 mean; ++ u64 min; ++ u64 max; ++ u32 nr_samples; ++ u64 batch; ++}; ++ ++struct blk_stat_callback { ++ struct list_head list; ++ struct timer_list timer; ++ struct blk_rq_stat *cpu_stat; ++ int (*bucket_fn)(const struct request *); ++ unsigned int buckets; ++ struct blk_rq_stat *stat; ++ void (*timer_fn)(struct blk_stat_callback *); ++ void *data; ++ struct callback_head rcu; ++}; ++ ++struct rq_wait { ++ wait_queue_head_t wait; ++ atomic_t inflight; ++}; ++ ++struct rq_depth { ++ unsigned int max_depth; ++ int scale_step; ++ bool scaled_max; ++ unsigned int queue_depth; ++ unsigned int default_depth; ++}; ++ ++typedef bool acquire_inflight_cb_t(struct rq_wait *, void *); ++ ++typedef void cleanup_cb_t(struct rq_wait *, void *); ++ ++struct trace_event_raw_wbt_stat { ++ struct trace_entry ent; ++ char name[32]; ++ s64 rmean; ++ u64 rmin; ++ u64 rmax; ++ s64 rnr_samples; ++ s64 rtime; ++ s64 wmean; ++ u64 wmin; ++ u64 wmax; ++ s64 wnr_samples; ++ s64 wtime; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_wbt_lat { ++ struct trace_entry ent; ++ char name[32]; ++ long unsigned int lat; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_wbt_step { ++ struct trace_entry ent; ++ char name[32]; ++ const char *msg; ++ int step; ++ long unsigned int window; ++ unsigned int bg; ++ unsigned int normal; ++ unsigned int max; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_wbt_timer { ++ struct trace_entry ent; ++ char name[32]; ++ unsigned int status; ++ int step; ++ unsigned int inflight; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_wbt_stat {}; ++ ++struct trace_event_data_offsets_wbt_lat {}; ++ ++struct trace_event_data_offsets_wbt_step {}; ++ ++struct trace_event_data_offsets_wbt_timer {}; ++ ++typedef void (*btf_trace_wbt_stat)(void *, struct backing_dev_info *, struct blk_rq_stat *); ++ ++typedef void (*btf_trace_wbt_lat)(void *, struct backing_dev_info *, long unsigned int); ++ ++typedef void (*btf_trace_wbt_step)(void *, struct backing_dev_info *, const char *, int, long unsigned int, unsigned int, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_wbt_timer)(void *, struct backing_dev_info *, unsigned int, int, unsigned int); ++ ++enum wbt_flags { ++ WBT_TRACKED = 1, ++ WBT_READ = 2, ++ WBT_KSWAPD = 4, ++ WBT_DISCARD = 8, ++ WBT_NR_BITS = 4, ++}; ++ ++enum { ++ WBT_RWQ_BG = 0, ++ WBT_RWQ_KSWAPD = 1, ++ WBT_RWQ_DISCARD = 2, ++ WBT_NUM_RWQ = 3, ++}; ++ ++enum { ++ WBT_STATE_ON_DEFAULT = 1, ++ WBT_STATE_ON_MANUAL = 2, ++ WBT_STATE_OFF_DEFAULT = 3, ++ WBT_STATE_OFF_MANUAL = 4, ++}; ++ ++struct rq_wb { ++ unsigned int wb_background; ++ unsigned int wb_normal; ++ short int enable_state; ++ unsigned int unknown_cnt; ++ u64 win_nsec; ++ u64 cur_win_nsec; ++ struct blk_stat_callback *cb; ++ u64 sync_issue; ++ void *sync_cookie; ++ unsigned int wc; ++ long unsigned int last_issue; ++ long unsigned int last_comp; ++ long unsigned int min_lat_nsec; ++ struct rq_qos rqos; ++ struct rq_wait rq_wait[3]; ++ struct rq_depth rq_depth; ++}; ++ ++enum { ++ RWB_DEF_DEPTH = 16, ++ RWB_WINDOW_NSEC = 100000000, ++ RWB_MIN_WRITE_SAMPLES = 3, ++ RWB_UNKNOWN_BUMP = 5, ++}; ++ ++enum { ++ LAT_OK = 1, ++ LAT_UNKNOWN = 2, ++ LAT_UNKNOWN_WRITES = 3, ++ LAT_EXCEEDED = 4, ++}; ++ ++struct wbt_wait_data { ++ struct rq_wb *rwb; ++ enum wbt_flags wb_acct; ++ blk_opf_t opf; ++}; ++ ++struct blk_crypto_ll_ops { ++ int (*keyslot_program)(struct blk_crypto_profile *, const struct blk_crypto_key *, unsigned int); ++ int (*keyslot_evict)(struct blk_crypto_profile *, const struct blk_crypto_key *, unsigned int); ++}; ++ ++struct blk_crypto_profile { ++ struct blk_crypto_ll_ops ll_ops; ++ unsigned int max_dun_bytes_supported; ++ unsigned int modes_supported[5]; ++ struct device *dev; ++ unsigned int num_slots; ++ struct rw_semaphore lock; ++ struct lock_class_key lockdep_key; ++ wait_queue_head_t idle_slots_wait_queue; ++ struct list_head idle_slots; ++ spinlock_t idle_slots_lock; ++ struct hlist_head *slot_hashtable; ++ unsigned int log_slot_ht_size; ++ struct blk_crypto_keyslot *slots; ++}; ++ ++struct blk_crypto_mode { ++ const char *name; ++ const char *cipher_str; ++ unsigned int keysize; ++ unsigned int ivsize; ++}; ++ ++enum io_uring_cmd_flags { ++ IO_URING_F_COMPLETE_DEFER = 1, ++ IO_URING_F_UNLOCKED = 2, ++ IO_URING_F_MULTISHOT = 4, ++ IO_URING_F_IOWQ = 8, ++ IO_URING_F_NONBLOCK = -2147483648, ++ IO_URING_F_SQE128 = 256, ++ IO_URING_F_CQE32 = 512, ++ IO_URING_F_IOPOLL = 1024, ++}; ++ ++struct xattr_name { ++ char name[256]; ++}; ++ ++struct xattr_ctx { ++ union { ++ const void *cvalue; ++ void *value; ++ }; ++ void *kvalue; ++ size_t size; ++ struct xattr_name *kname; ++ unsigned int flags; ++}; ++ ++enum { ++ REQ_F_FIXED_FILE = 1, ++ REQ_F_IO_DRAIN = 2, ++ REQ_F_LINK = 4, ++ REQ_F_HARDLINK = 8, ++ REQ_F_FORCE_ASYNC = 16, ++ REQ_F_BUFFER_SELECT = 32, ++ REQ_F_CQE_SKIP = 64, ++ REQ_F_FAIL = 256, ++ REQ_F_INFLIGHT = 512, ++ REQ_F_CUR_POS = 1024, ++ REQ_F_NOWAIT = 2048, ++ REQ_F_LINK_TIMEOUT = 4096, ++ REQ_F_NEED_CLEANUP = 8192, ++ REQ_F_POLLED = 16384, ++ REQ_F_BUFFER_SELECTED = 32768, ++ REQ_F_BUFFER_RING = 65536, ++ REQ_F_REISSUE = 131072, ++ REQ_F_SUPPORT_NOWAIT = 1073741824, ++ REQ_F_ISREG = 2147483648, ++ REQ_F_CREDS = 262144, ++ REQ_F_REFCOUNT = 524288, ++ REQ_F_ARM_LTIMEOUT = 1048576, ++ REQ_F_ASYNC_DATA = 2097152, ++ REQ_F_SKIP_LINK_CQES = 4194304, ++ REQ_F_SINGLE_POLL = 8388608, ++ REQ_F_DOUBLE_POLL = 16777216, ++ REQ_F_PARTIAL_IO = 33554432, ++ REQ_F_APOLL_MULTISHOT = 134217728, ++ REQ_F_CQE32_INIT = 67108864, ++ REQ_F_CLEAR_POLLIN = 268435456, ++ REQ_F_HASH_LOCKED = 536870912, ++}; ++ ++enum { ++ IOU_OK = 0, ++ IOU_ISSUE_SKIP_COMPLETE = -529, ++ IOU_STOP_MULTISHOT = -125, ++}; ++ ++struct io_xattr { ++ struct file *file; ++ struct xattr_ctx ctx; ++ struct filename *filename; ++}; ++ ++struct io_splice { ++ struct file *file_out; ++ loff_t off_out; ++ loff_t off_in; ++ u64 len; ++ int splice_fd_in; ++ unsigned int flags; ++}; ++ ++struct io_fadvise { ++ struct file *file; ++ u64 offset; ++ u32 len; ++ u32 advice; ++}; ++ ++struct io_madvise { ++ struct file *file; ++ u64 addr; ++ u32 len; ++ u32 advice; ++}; ++ ++struct io_cache_entry { ++ struct io_wq_work_node node; ++}; ++ ++struct io_rsrc_put { ++ u64 tag; ++ union { ++ void *rsrc; ++ struct file *file; ++ struct io_mapped_ubuf *buf; ++ }; ++}; ++ ++struct io_rsrc_node { ++ union { ++ struct io_cache_entry cache; ++ struct io_ring_ctx *ctx; ++ }; ++ int refs; ++ bool empty; ++ u16 type; ++ struct list_head node; ++ struct io_rsrc_put item; ++}; ++ ++struct io_mapped_ubuf { ++ u64 ubuf; ++ u64 ubuf_end; ++ unsigned int nr_bvecs; ++ long unsigned int acct_pages; ++ struct bio_vec bvec[0]; ++}; ++ ++struct io_rsrc_data { ++ struct io_ring_ctx *ctx; ++ u64 **tags; ++ unsigned int nr; ++ u16 rsrc_type; ++ bool quiesce; ++}; ++ ++struct io_open { ++ struct file *file; ++ int dfd; ++ u32 file_slot; ++ struct filename *filename; ++ struct open_how how; ++ long unsigned int nofile; ++}; ++ ++struct io_close { ++ struct file *file; ++ int fd; ++ u32 file_slot; ++}; ++ ++struct statx_timestamp { ++ __s64 tv_sec; ++ __u32 tv_nsec; ++ __s32 __reserved; ++}; ++ ++struct statx { ++ __u32 stx_mask; ++ __u32 stx_blksize; ++ __u64 stx_attributes; ++ __u32 stx_nlink; ++ __u32 stx_uid; ++ __u32 stx_gid; ++ __u16 stx_mode; ++ __u16 __spare0[1]; ++ __u64 stx_ino; ++ __u64 stx_size; ++ __u64 stx_blocks; ++ __u64 stx_attributes_mask; ++ struct statx_timestamp stx_atime; ++ struct statx_timestamp stx_btime; ++ struct statx_timestamp stx_ctime; ++ struct statx_timestamp stx_mtime; ++ __u32 stx_rdev_major; ++ __u32 stx_rdev_minor; ++ __u32 stx_dev_major; ++ __u32 stx_dev_minor; ++ __u64 stx_mnt_id; ++ __u32 stx_dio_mem_align; ++ __u32 stx_dio_offset_align; ++ __u64 __spare3[12]; ++}; ++ ++struct io_statx { ++ struct file *file; ++ int dfd; ++ unsigned int mask; ++ unsigned int flags; ++ struct filename *filename; ++ struct statx *buffer; ++}; ++ ++enum { ++ IORING_MSG_DATA = 0, ++ IORING_MSG_SEND_FD = 1, ++}; ++ ++struct io_msg { ++ struct file *file; ++ struct file *src_file; ++ struct callback_head tw; ++ u64 user_data; ++ u32 len; ++ u32 cmd; ++ u32 src_fd; ++ union { ++ u32 dst_fd; ++ u32 cqe_flags; ++ }; ++ u32 flags; ++}; ++ ++struct io_uring_rsrc_update { ++ __u32 offset; ++ __u32 resv; ++ __u64 data; ++}; ++ ++typedef struct io_wq_work *free_work_fn(struct io_wq_work *); ++ ++typedef void io_wq_work_fn(struct io_wq_work *); ++ ++struct io_wq_data { ++ struct io_wq_hash *hash; ++ struct task_struct *task; ++ io_wq_work_fn *do_work; ++ free_work_fn *free_work; ++}; ++ ++struct io_tctx_node { ++ struct list_head ctx_node; ++ struct task_struct *task; ++ struct io_ring_ctx *ctx; ++}; ++ ++enum { ++ IORING_CQE_BUFFER_SHIFT = 16, ++}; ++ ++struct io_uring_buf { ++ __u64 addr; ++ __u32 len; ++ __u16 bid; ++ __u16 resv; ++}; ++ ++struct io_uring_buf_ring { ++ union { ++ struct { ++ __u64 resv1; ++ __u32 resv2; ++ __u16 resv3; ++ __u16 tail; ++ }; ++ struct { ++ struct {} __empty_bufs; ++ struct io_uring_buf bufs[0]; ++ }; ++ }; ++}; ++ ++enum { ++ IOU_PBUF_RING_MMAP = 1, ++}; ++ ++struct io_uring_buf_reg { ++ __u64 ring_addr; ++ __u32 ring_entries; ++ __u16 bgid; ++ __u16 flags; ++ __u64 resv[3]; ++}; ++ ++struct io_buffer_list { ++ union { ++ struct list_head buf_list; ++ struct { ++ struct page **buf_pages; ++ struct io_uring_buf_ring *buf_ring; ++ }; ++ }; ++ __u16 bgid; ++ __u16 buf_nr_pages; ++ __u16 nr_entries; ++ __u16 head; ++ __u16 mask; ++ __u8 is_mapped; ++ __u8 is_mmap; ++}; ++ ++struct io_buffer { ++ struct list_head list; ++ __u64 addr; ++ __u32 len; ++ __u16 bid; ++ __u16 bgid; ++}; ++ ++struct io_provide_buf { ++ struct file *file; ++ __u64 addr; ++ __u32 len; ++ __u32 bgid; ++ __u16 nbufs; ++ __u16 bid; ++}; ++ ++struct io_sq_data { ++ refcount_t refs; ++ atomic_t park_pending; ++ struct mutex lock; ++ struct list_head ctx_list; ++ struct task_struct *thread; ++ struct wait_queue_head wait; ++ unsigned int sq_thread_idle; ++ int sq_cpu; ++ pid_t task_pid; ++ pid_t task_tgid; ++ long unsigned int state; ++ struct completion exited; ++}; ++ ++struct io_poll { ++ struct file *file; ++ struct wait_queue_head *head; ++ __poll_t events; ++ int retries; ++ struct wait_queue_entry wait; ++}; ++ ++struct async_poll { ++ union { ++ struct io_poll poll; ++ struct io_cache_entry cache; ++ }; ++ struct io_poll *double_poll; ++}; ++ ++struct io_issue_def { ++ unsigned int needs_file: 1; ++ unsigned int plug: 1; ++ unsigned int hash_reg_file: 1; ++ unsigned int unbound_nonreg_file: 1; ++ unsigned int pollin: 1; ++ unsigned int pollout: 1; ++ unsigned int poll_exclusive: 1; ++ unsigned int buffer_select: 1; ++ unsigned int not_supported: 1; ++ unsigned int audit_skip: 1; ++ unsigned int ioprio: 1; ++ unsigned int iopoll: 1; ++ unsigned int iopoll_queue: 1; ++ unsigned int manual_alloc: 1; ++ int (*issue)(struct io_kiocb *, unsigned int); ++ int (*prep)(struct io_kiocb *, const struct io_uring_sqe *); ++}; ++ ++struct io_cold_def { ++ short unsigned int async_size; ++ const char *name; ++ int (*prep_async)(struct io_kiocb *); ++ void (*cleanup)(struct io_kiocb *); ++ void (*fail)(struct io_kiocb *); ++}; ++ ++struct wrapper { ++ cmp_func_t cmp; ++ swap_func_t swap; ++}; ++ ++struct sg_append_table { ++ struct sg_table sgt; ++ struct scatterlist *prv; ++ unsigned int total_nents; ++}; ++ ++typedef struct scatterlist *sg_alloc_fn(unsigned int, gfp_t); ++ ++typedef void sg_free_fn(struct scatterlist *, unsigned int); ++ ++struct sg_page_iter { ++ struct scatterlist *sg; ++ unsigned int sg_pgoffset; ++ unsigned int __nents; ++ int __pg_advance; ++}; ++ ++struct sg_dma_page_iter { ++ struct sg_page_iter base; ++}; ++ ++struct sg_mapping_iter { ++ struct page *page; ++ void *addr; ++ size_t length; ++ size_t consumed; ++ struct sg_page_iter piter; ++ unsigned int __offset; ++ unsigned int __remaining; ++ unsigned int __flags; ++}; ++ ++struct iov_iter_state { ++ size_t iov_offset; ++ size_t count; ++ long unsigned int nr_segs; ++}; ++ ++struct csum_state { ++ __wsum csum; ++ size_t off; ++}; ++ ++typedef s32 compat_ssize_t; ++ ++struct test_fail { ++ const char *str; ++ unsigned int base; ++}; ++ ++struct test_s8 { ++ const char *str; ++ unsigned int base; ++ s8 expected_res; ++}; ++ ++struct test_u8 { ++ const char *str; ++ unsigned int base; ++ u8 expected_res; ++}; ++ ++struct test_s16 { ++ const char *str; ++ unsigned int base; ++ s16 expected_res; ++}; ++ ++struct test_u16 { ++ const char *str; ++ unsigned int base; ++ u16 expected_res; ++}; ++ ++struct test_s32 { ++ const char *str; ++ unsigned int base; ++ s32 expected_res; ++}; ++ ++struct test_u32 { ++ const char *str; ++ unsigned int base; ++ u32 expected_res; ++}; ++ ++struct test_s64 { ++ const char *str; ++ unsigned int base; ++ s64 expected_res; ++}; ++ ++struct test_u64 { ++ const char *str; ++ unsigned int base; ++ u64 expected_res; ++}; ++ ++struct test_ll { ++ const char *str; ++ unsigned int base; ++ long long int expected_res; ++}; ++ ++struct test_ull { ++ const char *str; ++ unsigned int base; ++ long long unsigned int expected_res; ++}; ++ ++typedef struct { ++ u64 a; ++ u64 b; ++} u128; ++ ++typedef struct { ++ __be64 a; ++ __be64 b; ++} be128; ++ ++typedef struct { ++ __le64 b; ++ __le64 a; ++} le128; ++ ++struct gf128mul_4k { ++ be128 t[256]; ++}; ++ ++struct gf128mul_64k { ++ struct gf128mul_4k *t[16]; ++}; ++ ++enum blake2s_lengths { ++ BLAKE2S_BLOCK_SIZE = 64, ++ BLAKE2S_HASH_SIZE = 32, ++ BLAKE2S_KEY_SIZE = 32, ++ BLAKE2S_128_HASH_SIZE = 16, ++ BLAKE2S_160_HASH_SIZE = 20, ++ BLAKE2S_224_HASH_SIZE = 28, ++ BLAKE2S_256_HASH_SIZE = 32, ++}; ++ ++struct blake2s_state { ++ u32 h[8]; ++ u32 t[2]; ++ u32 f[2]; ++ u8 buf[64]; ++ unsigned int buflen; ++ unsigned int outlen; ++}; ++ ++enum blake2s_iv { ++ BLAKE2S_IV0 = 1779033703, ++ BLAKE2S_IV1 = 3144134277, ++ BLAKE2S_IV2 = 1013904242, ++ BLAKE2S_IV3 = 2773480762, ++ BLAKE2S_IV4 = 1359893119, ++ BLAKE2S_IV5 = 2600822924, ++ BLAKE2S_IV6 = 528734635, ++ BLAKE2S_IV7 = 1541459225, ++}; ++ ++enum chacha20poly1305_lengths { ++ XCHACHA20POLY1305_NONCE_SIZE = 24, ++ CHACHA20POLY1305_KEY_SIZE = 32, ++ CHACHA20POLY1305_AUTHTAG_SIZE = 16, ++}; ++ ++struct poly1305_key { ++ union { ++ u32 r[5]; ++ u64 r64[3]; ++ }; ++}; ++ ++struct poly1305_core_key { ++ struct poly1305_key key; ++ struct poly1305_key precomputed_s; ++}; ++ ++struct poly1305_state { ++ union { ++ u32 h[5]; ++ u64 h64[3]; ++ }; ++}; ++ ++struct poly1305_desc_ctx { ++ u8 buf[16]; ++ unsigned int buflen; ++ short unsigned int rset; ++ bool sset; ++ u32 s[4]; ++ struct poly1305_state h; ++ union { ++ struct poly1305_key opaque_r[1]; ++ struct poly1305_core_key core_r; ++ }; ++}; ++ ++struct chacha20poly1305_testvec { ++ const u8 *input; ++ const u8 *output; ++ const u8 *assoc; ++ const u8 *nonce; ++ const u8 *key; ++ size_t ilen; ++ size_t alen; ++ size_t nlen; ++ bool failure; ++}; ++ ++enum { ++ MAXIMUM_TEST_BUFFER_LEN = 4096, ++}; ++ ++struct assoc_array_ops { ++ long unsigned int (*get_key_chunk)(const void *, int); ++ long unsigned int (*get_object_key_chunk)(const void *, int); ++ bool (*compare_object)(const void *, const void *); ++ int (*diff_objects)(const void *, const void *); ++ void (*free_object)(void *); ++}; ++ ++struct assoc_array_node { ++ struct assoc_array_ptr *back_pointer; ++ u8 parent_slot; ++ struct assoc_array_ptr *slots[16]; ++ long unsigned int nr_leaves_on_branch; ++}; ++ ++struct assoc_array_shortcut { ++ struct assoc_array_ptr *back_pointer; ++ int parent_slot; ++ int skip_to_level; ++ struct assoc_array_ptr *next_node; ++ long unsigned int index_key[0]; ++}; ++ ++struct assoc_array_edit { ++ struct callback_head rcu; ++ struct assoc_array *array; ++ const struct assoc_array_ops *ops; ++ const struct assoc_array_ops *ops_for_excised_subtree; ++ struct assoc_array_ptr *leaf; ++ struct assoc_array_ptr **leaf_p; ++ struct assoc_array_ptr *dead_leaf; ++ struct assoc_array_ptr *new_meta[3]; ++ struct assoc_array_ptr *excised_meta[1]; ++ struct assoc_array_ptr *excised_subtree; ++ struct assoc_array_ptr **set_backpointers[16]; ++ struct assoc_array_ptr *set_backpointers_to; ++ struct assoc_array_node *adjust_count_on; ++ long int adjust_count_by; ++ struct { ++ struct assoc_array_ptr **ptr; ++ struct assoc_array_ptr *to; ++ } set[2]; ++ struct { ++ u8 *p; ++ u8 to; ++ } set_parent_slot[1]; ++ u8 segment_cache[17]; ++}; ++ ++enum assoc_array_walk_status { ++ assoc_array_walk_tree_empty = 0, ++ assoc_array_walk_found_terminal_node = 1, ++ assoc_array_walk_found_wrong_shortcut = 2, ++}; ++ ++struct assoc_array_walk_result { ++ struct { ++ struct assoc_array_node *node; ++ int level; ++ int slot; ++ } terminal_node; ++ struct { ++ struct assoc_array_shortcut *shortcut; ++ int level; ++ int sc_level; ++ long unsigned int sc_segments; ++ long unsigned int dissimilarity; ++ } wrong_shortcut; ++}; ++ ++struct assoc_array_delete_collapse_context { ++ struct assoc_array_node *node; ++ const void *skip_leaf; ++ int slot; ++}; ++ ++enum { ++ CRYPTO_MSG_ALG_REQUEST = 0, ++ CRYPTO_MSG_ALG_REGISTER = 1, ++ CRYPTO_MSG_ALG_LOADED = 2, ++}; ++ ++struct sw842_param { ++ u8 *in; ++ u8 bit; ++ u64 ilen; ++ u8 *out; ++ u8 *ostart; ++ u64 olen; ++}; ++ ++typedef long int __kernel_ptrdiff_t; ++ ++typedef __kernel_ptrdiff_t ptrdiff_t; ++ ++typedef uint8_t BYTE; ++ ++typedef uint16_t U16; ++ ++typedef uint32_t U32; ++ ++typedef uint64_t U64; ++ ++typedef struct { ++ size_t bitContainer; ++ unsigned int bitPos; ++ char *startPtr; ++ char *ptr; ++ char *endPtr; ++} BIT_CStream_t; ++ ++typedef unsigned int FSE_CTable; ++ ++typedef struct { ++ ptrdiff_t value; ++ const void *stateTable; ++ const void *symbolTT; ++ unsigned int stateLog; ++} FSE_CState_t; ++ ++typedef struct { ++ int deltaFindState; ++ U32 deltaNbBits; ++} FSE_symbolCompressionTransform; ++ ++typedef s16 int16_t; ++ ++typedef uint8_t U8; ++ ++typedef int16_t S16; ++ ++typedef enum { ++ FSE_repeat_none = 0, ++ FSE_repeat_check = 1, ++ FSE_repeat_valid = 2, ++} FSE_repeat; ++ ++typedef size_t HUF_CElt; ++ ++typedef enum { ++ HUF_repeat_none = 0, ++ HUF_repeat_check = 1, ++ HUF_repeat_valid = 2, ++} HUF_repeat; ++ ++typedef struct { ++ U32 f1c; ++ U32 f1d; ++ U32 f7b; ++ U32 f7c; ++} ZSTD_cpuid_t; ++ ++typedef enum { ++ ZSTDcs_created = 0, ++ ZSTDcs_init = 1, ++ ZSTDcs_ongoing = 2, ++ ZSTDcs_ending = 3, ++} ZSTD_compressionStage_e; ++ ++typedef enum { ++ ZSTD_f_zstd1 = 0, ++ ZSTD_f_zstd1_magicless = 1, ++} ZSTD_format_e; ++ ++typedef enum { ++ ZSTD_fast = 1, ++ ZSTD_dfast = 2, ++ ZSTD_greedy = 3, ++ ZSTD_lazy = 4, ++ ZSTD_lazy2 = 5, ++ ZSTD_btlazy2 = 6, ++ ZSTD_btopt = 7, ++ ZSTD_btultra = 8, ++ ZSTD_btultra2 = 9, ++} ZSTD_strategy; ++ ++typedef struct { ++ unsigned int windowLog; ++ unsigned int chainLog; ++ unsigned int hashLog; ++ unsigned int searchLog; ++ unsigned int minMatch; ++ unsigned int targetLength; ++ ZSTD_strategy strategy; ++} ZSTD_compressionParameters; ++ ++typedef struct { ++ int contentSizeFlag; ++ int checksumFlag; ++ int noDictIDFlag; ++} ZSTD_frameParameters; ++ ++typedef enum { ++ ZSTD_dictDefaultAttach = 0, ++ ZSTD_dictForceAttach = 1, ++ ZSTD_dictForceCopy = 2, ++ ZSTD_dictForceLoad = 3, ++} ZSTD_dictAttachPref_e; ++ ++typedef enum { ++ ZSTD_ps_auto = 0, ++ ZSTD_ps_enable = 1, ++ ZSTD_ps_disable = 2, ++} ZSTD_paramSwitch_e; ++ ++typedef struct { ++ ZSTD_paramSwitch_e enableLdm; ++ U32 hashLog; ++ U32 bucketSizeLog; ++ U32 minMatchLength; ++ U32 hashRateLog; ++ U32 windowLog; ++} ldmParams_t; ++ ++typedef enum { ++ ZSTD_bm_buffered = 0, ++ ZSTD_bm_stable = 1, ++} ZSTD_bufferMode_e; ++ ++typedef enum { ++ ZSTD_sf_noBlockDelimiters = 0, ++ ZSTD_sf_explicitBlockDelimiters = 1, ++} ZSTD_sequenceFormat_e; ++ ++typedef void * (*ZSTD_allocFunction)(void *, size_t); ++ ++typedef void (*ZSTD_freeFunction)(void *, void *); ++ ++typedef struct { ++ ZSTD_allocFunction customAlloc; ++ ZSTD_freeFunction customFree; ++ void *opaque; ++} ZSTD_customMem; ++ ++struct ZSTD_CCtx_params_s { ++ ZSTD_format_e format; ++ ZSTD_compressionParameters cParams; ++ ZSTD_frameParameters fParams; ++ int compressionLevel; ++ int forceWindow; ++ size_t targetCBlockSize; ++ int srcSizeHint; ++ ZSTD_dictAttachPref_e attachDictPref; ++ ZSTD_paramSwitch_e literalCompressionMode; ++ int nbWorkers; ++ size_t jobSize; ++ int overlapLog; ++ int rsyncable; ++ ldmParams_t ldmParams; ++ int enableDedicatedDictSearch; ++ ZSTD_bufferMode_e inBufferMode; ++ ZSTD_bufferMode_e outBufferMode; ++ ZSTD_sequenceFormat_e blockDelimiters; ++ int validateSequences; ++ ZSTD_paramSwitch_e useBlockSplitter; ++ ZSTD_paramSwitch_e useRowMatchFinder; ++ int deterministicRefPrefix; ++ ZSTD_customMem customMem; ++}; ++ ++typedef struct ZSTD_CCtx_params_s ZSTD_CCtx_params; ++ ++typedef enum { ++ ZSTD_cwksp_alloc_objects = 0, ++ ZSTD_cwksp_alloc_buffers = 1, ++ ZSTD_cwksp_alloc_aligned = 2, ++} ZSTD_cwksp_alloc_phase_e; ++ ++typedef enum { ++ ZSTD_cwksp_dynamic_alloc = 0, ++ ZSTD_cwksp_static_alloc = 1, ++} ZSTD_cwksp_static_alloc_e; ++ ++typedef struct { ++ void *workspace; ++ void *workspaceEnd; ++ void *objectEnd; ++ void *tableEnd; ++ void *tableValidEnd; ++ void *allocStart; ++ BYTE allocFailed; ++ int workspaceOversizedDuration; ++ ZSTD_cwksp_alloc_phase_e phase; ++ ZSTD_cwksp_static_alloc_e isStatic; ++} ZSTD_cwksp; ++ ++struct xxh64_state { ++ uint64_t total_len; ++ uint64_t v1; ++ uint64_t v2; ++ uint64_t v3; ++ uint64_t v4; ++ uint64_t mem64[4]; ++ uint32_t memsize; ++}; ++ ++struct POOL_ctx_s; ++ ++typedef struct POOL_ctx_s ZSTD_threadPool; ++ ++typedef struct { ++ unsigned int offset; ++ unsigned int litLength; ++ unsigned int matchLength; ++ unsigned int rep; ++} ZSTD_Sequence; ++ ++typedef struct { ++ int collectSequences; ++ ZSTD_Sequence *seqStart; ++ size_t seqIndex; ++ size_t maxSequences; ++} SeqCollector; ++ ++typedef enum { ++ ZSTD_llt_none = 0, ++ ZSTD_llt_literalLength = 1, ++ ZSTD_llt_matchLength = 2, ++} ZSTD_longLengthType_e; ++ ++struct seqDef_s; ++ ++typedef struct seqDef_s seqDef; ++ ++typedef struct { ++ seqDef *sequencesStart; ++ seqDef *sequences; ++ BYTE *litStart; ++ BYTE *lit; ++ BYTE *llCode; ++ BYTE *mlCode; ++ BYTE *ofCode; ++ size_t maxNbSeq; ++ size_t maxNbLit; ++ ZSTD_longLengthType_e longLengthType; ++ U32 longLengthPos; ++} seqStore_t; ++ ++typedef struct { ++ const BYTE *nextSrc; ++ const BYTE *base; ++ const BYTE *dictBase; ++ U32 dictLimit; ++ U32 lowLimit; ++ U32 nbOverflowCorrections; ++} ZSTD_window_t; ++ ++typedef struct { ++ U32 offset; ++ U32 checksum; ++} ldmEntry_t; ++ ++typedef struct { ++ const BYTE *split; ++ U32 hash; ++ U32 checksum; ++ ldmEntry_t *bucket; ++} ldmMatchCandidate_t; ++ ++typedef struct { ++ ZSTD_window_t window; ++ ldmEntry_t *hashTable; ++ U32 loadedDictEnd; ++ BYTE *bucketOffsets; ++ size_t splitIndices[64]; ++ ldmMatchCandidate_t matchCandidates[64]; ++} ldmState_t; ++ ++typedef struct { ++ U32 offset; ++ U32 litLength; ++ U32 matchLength; ++} rawSeq; ++ ++typedef struct { ++ rawSeq *seq; ++ size_t pos; ++ size_t posInSequence; ++ size_t size; ++ size_t capacity; ++} rawSeqStore_t; ++ ++typedef struct { ++ HUF_CElt CTable[257]; ++ HUF_repeat repeatMode; ++} ZSTD_hufCTables_t; ++ ++typedef struct { ++ FSE_CTable offcodeCTable[193]; ++ FSE_CTable matchlengthCTable[363]; ++ FSE_CTable litlengthCTable[329]; ++ FSE_repeat offcode_repeatMode; ++ FSE_repeat matchlength_repeatMode; ++ FSE_repeat litlength_repeatMode; ++} ZSTD_fseCTables_t; ++ ++typedef struct { ++ ZSTD_hufCTables_t huf; ++ ZSTD_fseCTables_t fse; ++} ZSTD_entropyCTables_t; ++ ++typedef struct { ++ ZSTD_entropyCTables_t entropy; ++ U32 rep[3]; ++} ZSTD_compressedBlockState_t; ++ ++typedef struct { ++ U32 off; ++ U32 len; ++} ZSTD_match_t; ++ ++typedef struct { ++ int price; ++ U32 off; ++ U32 mlen; ++ U32 litlen; ++ U32 rep[3]; ++} ZSTD_optimal_t; ++ ++typedef enum { ++ zop_dynamic = 0, ++ zop_predef = 1, ++} ZSTD_OptPrice_e; ++ ++typedef struct { ++ unsigned int *litFreq; ++ unsigned int *litLengthFreq; ++ unsigned int *matchLengthFreq; ++ unsigned int *offCodeFreq; ++ ZSTD_match_t *matchTable; ++ ZSTD_optimal_t *priceTable; ++ U32 litSum; ++ U32 litLengthSum; ++ U32 matchLengthSum; ++ U32 offCodeSum; ++ U32 litSumBasePrice; ++ U32 litLengthSumBasePrice; ++ U32 matchLengthSumBasePrice; ++ U32 offCodeSumBasePrice; ++ ZSTD_OptPrice_e priceType; ++ const ZSTD_entropyCTables_t *symbolCosts; ++ ZSTD_paramSwitch_e literalCompressionMode; ++} optState_t; ++ ++struct ZSTD_matchState_t; ++ ++typedef struct ZSTD_matchState_t ZSTD_matchState_t; ++ ++struct ZSTD_matchState_t { ++ ZSTD_window_t window; ++ U32 loadedDictEnd; ++ U32 nextToUpdate; ++ U32 hashLog3; ++ U32 rowHashLog; ++ U16 *tagTable; ++ U32 hashCache[8]; ++ U32 *hashTable; ++ U32 *hashTable3; ++ U32 *chainTable; ++ U32 forceNonContiguous; ++ int dedicatedDictSearch; ++ optState_t opt; ++ const ZSTD_matchState_t *dictMatchState; ++ ZSTD_compressionParameters cParams; ++ const rawSeqStore_t *ldmSeqStore; ++}; ++ ++typedef struct { ++ ZSTD_compressedBlockState_t *prevCBlock; ++ ZSTD_compressedBlockState_t *nextCBlock; ++ ZSTD_matchState_t matchState; ++} ZSTD_blockState_t; ++ ++typedef enum { ++ ZSTDb_not_buffered = 0, ++ ZSTDb_buffered = 1, ++} ZSTD_buffered_policy_e; ++ ++typedef enum { ++ zcss_init = 0, ++ zcss_load = 1, ++ zcss_flush = 2, ++} ZSTD_cStreamStage; ++ ++struct ZSTD_inBuffer_s { ++ const void *src; ++ size_t size; ++ size_t pos; ++}; ++ ++typedef struct ZSTD_inBuffer_s ZSTD_inBuffer; ++ ++typedef enum { ++ ZSTD_dct_auto = 0, ++ ZSTD_dct_rawContent = 1, ++ ZSTD_dct_fullDict = 2, ++} ZSTD_dictContentType_e; ++ ++struct ZSTD_CDict_s; ++ ++typedef struct ZSTD_CDict_s ZSTD_CDict; ++ ++typedef struct { ++ void *dictBuffer; ++ const void *dict; ++ size_t dictSize; ++ ZSTD_dictContentType_e dictContentType; ++ ZSTD_CDict *cdict; ++} ZSTD_localDict; ++ ++struct ZSTD_prefixDict_s { ++ const void *dict; ++ size_t dictSize; ++ ZSTD_dictContentType_e dictContentType; ++}; ++ ++typedef struct ZSTD_prefixDict_s ZSTD_prefixDict; ++ ++typedef enum { ++ set_basic = 0, ++ set_rle = 1, ++ set_compressed = 2, ++ set_repeat = 3, ++} symbolEncodingType_e; ++ ++typedef struct { ++ symbolEncodingType_e hType; ++ BYTE hufDesBuffer[128]; ++ size_t hufDesSize; ++} ZSTD_hufCTablesMetadata_t; ++ ++typedef struct { ++ symbolEncodingType_e llType; ++ symbolEncodingType_e ofType; ++ symbolEncodingType_e mlType; ++ BYTE fseTablesBuffer[133]; ++ size_t fseTablesSize; ++ size_t lastCountSize; ++} ZSTD_fseCTablesMetadata_t; ++ ++typedef struct { ++ ZSTD_hufCTablesMetadata_t hufMetadata; ++ ZSTD_fseCTablesMetadata_t fseMetadata; ++} ZSTD_entropyCTablesMetadata_t; ++ ++typedef struct { ++ seqStore_t fullSeqStoreChunk; ++ seqStore_t firstHalfSeqStore; ++ seqStore_t secondHalfSeqStore; ++ seqStore_t currSeqStore; ++ seqStore_t nextSeqStore; ++ U32 partitions[196]; ++ ZSTD_entropyCTablesMetadata_t entropyMetadata; ++} ZSTD_blockSplitCtx; ++ ++struct ZSTD_CCtx_s { ++ ZSTD_compressionStage_e stage; ++ int cParamsChanged; ++ int bmi2; ++ ZSTD_CCtx_params requestedParams; ++ ZSTD_CCtx_params appliedParams; ++ ZSTD_CCtx_params simpleApiParams; ++ U32 dictID; ++ size_t dictContentSize; ++ ZSTD_cwksp workspace; ++ size_t blockSize; ++ long long unsigned int pledgedSrcSizePlusOne; ++ long long unsigned int consumedSrcSize; ++ long long unsigned int producedCSize; ++ struct xxh64_state xxhState; ++ ZSTD_customMem customMem; ++ ZSTD_threadPool *pool; ++ size_t staticSize; ++ SeqCollector seqCollector; ++ int isFirstBlock; ++ int initialized; ++ seqStore_t seqStore; ++ ldmState_t ldmState; ++ rawSeq *ldmSequences; ++ size_t maxNbLdmSequences; ++ rawSeqStore_t externSeqStore; ++ ZSTD_blockState_t blockState; ++ U32 *entropyWorkspace; ++ ZSTD_buffered_policy_e bufferedPolicy; ++ char *inBuff; ++ size_t inBuffSize; ++ size_t inToCompress; ++ size_t inBuffPos; ++ size_t inBuffTarget; ++ char *outBuff; ++ size_t outBuffSize; ++ size_t outBuffContentSize; ++ size_t outBuffFlushedSize; ++ ZSTD_cStreamStage streamStage; ++ U32 frameEnded; ++ ZSTD_inBuffer expectedInBuffer; ++ size_t expectedOutBufferSize; ++ ZSTD_localDict localDict; ++ const ZSTD_CDict *cdict; ++ ZSTD_prefixDict prefixDict; ++ ZSTD_blockSplitCtx blockSplitCtx; ++}; ++ ++typedef struct ZSTD_CCtx_s ZSTD_CCtx; ++ ++typedef enum { ++ ZSTD_c_compressionLevel = 100, ++ ZSTD_c_windowLog = 101, ++ ZSTD_c_hashLog = 102, ++ ZSTD_c_chainLog = 103, ++ ZSTD_c_searchLog = 104, ++ ZSTD_c_minMatch = 105, ++ ZSTD_c_targetLength = 106, ++ ZSTD_c_strategy = 107, ++ ZSTD_c_enableLongDistanceMatching = 160, ++ ZSTD_c_ldmHashLog = 161, ++ ZSTD_c_ldmMinMatch = 162, ++ ZSTD_c_ldmBucketSizeLog = 163, ++ ZSTD_c_ldmHashRateLog = 164, ++ ZSTD_c_contentSizeFlag = 200, ++ ZSTD_c_checksumFlag = 201, ++ ZSTD_c_dictIDFlag = 202, ++ ZSTD_c_nbWorkers = 400, ++ ZSTD_c_jobSize = 401, ++ ZSTD_c_overlapLog = 402, ++ ZSTD_c_experimentalParam1 = 500, ++ ZSTD_c_experimentalParam2 = 10, ++ ZSTD_c_experimentalParam3 = 1000, ++ ZSTD_c_experimentalParam4 = 1001, ++ ZSTD_c_experimentalParam5 = 1002, ++ ZSTD_c_experimentalParam6 = 1003, ++ ZSTD_c_experimentalParam7 = 1004, ++ ZSTD_c_experimentalParam8 = 1005, ++ ZSTD_c_experimentalParam9 = 1006, ++ ZSTD_c_experimentalParam10 = 1007, ++ ZSTD_c_experimentalParam11 = 1008, ++ ZSTD_c_experimentalParam12 = 1009, ++ ZSTD_c_experimentalParam13 = 1010, ++ ZSTD_c_experimentalParam14 = 1011, ++ ZSTD_c_experimentalParam15 = 1012, ++} ZSTD_cParameter; ++ ++typedef struct { ++ size_t error; ++ int lowerBound; ++ int upperBound; ++} ZSTD_bounds; ++ ++typedef enum { ++ ZSTD_reset_session_only = 1, ++ ZSTD_reset_parameters = 2, ++ ZSTD_reset_session_and_parameters = 3, ++} ZSTD_ResetDirective; ++ ++struct ZSTD_outBuffer_s { ++ void *dst; ++ size_t size; ++ size_t pos; ++}; ++ ++typedef struct ZSTD_outBuffer_s ZSTD_outBuffer; ++ ++typedef ZSTD_CCtx ZSTD_CStream; ++ ++typedef enum { ++ ZSTD_e_continue = 0, ++ ZSTD_e_flush = 1, ++ ZSTD_e_end = 2, ++} ZSTD_EndDirective; ++ ++struct ZSTD_CDict_s { ++ const void *dictContent; ++ size_t dictContentSize; ++ ZSTD_dictContentType_e dictContentType; ++ U32 *entropyWorkspace; ++ ZSTD_cwksp workspace; ++ ZSTD_matchState_t matchState; ++ ZSTD_compressedBlockState_t cBlockState; ++ ZSTD_customMem customMem; ++ U32 dictID; ++ int compressionLevel; ++ ZSTD_paramSwitch_e useRowMatchFinder; ++}; ++ ++typedef struct { ++ ZSTD_compressionParameters cParams; ++ ZSTD_frameParameters fParams; ++} ZSTD_parameters; ++ ++typedef enum { ++ ZSTD_dlm_byCopy = 0, ++ ZSTD_dlm_byRef = 1, ++} ZSTD_dictLoadMethod_e; ++ ++typedef struct { ++ long long unsigned int ingested; ++ long long unsigned int consumed; ++ long long unsigned int produced; ++ long long unsigned int flushed; ++ unsigned int currentJobID; ++ unsigned int nbActiveWorkers; ++} ZSTD_frameProgression; ++ ++typedef enum { ++ ZSTD_no_overlap = 0, ++ ZSTD_overlap_src_before_dst = 1, ++} ZSTD_overlap_e; ++ ++struct seqDef_s { ++ U32 offBase; ++ U16 litLength; ++ U16 mlBase; ++}; ++ ++typedef enum { ++ ZSTD_dtlm_fast = 0, ++ ZSTD_dtlm_full = 1, ++} ZSTD_dictTableLoadMethod_e; ++ ++typedef enum { ++ ZSTD_noDict = 0, ++ ZSTD_extDict = 1, ++ ZSTD_dictMatchState = 2, ++ ZSTD_dedicatedDictSearch = 3, ++} ZSTD_dictMode_e; ++ ++typedef enum { ++ ZSTD_cpm_noAttachDict = 0, ++ ZSTD_cpm_attachDict = 1, ++ ZSTD_cpm_createCDict = 2, ++ ZSTD_cpm_unknown = 3, ++} ZSTD_cParamMode_e; ++ ++typedef size_t (*ZSTD_blockCompressor)(ZSTD_matchState_t *, seqStore_t *, U32 *, const void *, size_t); ++ ++struct repcodes_s { ++ U32 rep[3]; ++}; ++ ++typedef struct repcodes_s repcodes_t; ++ ++typedef enum { ++ ZSTD_defaultDisallowed = 0, ++ ZSTD_defaultAllowed = 1, ++} ZSTD_defaultPolicy_e; ++ ++typedef enum { ++ ZSTDcrp_makeClean = 0, ++ ZSTDcrp_leaveDirty = 1, ++} ZSTD_compResetPolicy_e; ++ ++typedef enum { ++ ZSTDirp_continue = 0, ++ ZSTDirp_reset = 1, ++} ZSTD_indexResetPolicy_e; ++ ++typedef enum { ++ ZSTD_resetTarget_CDict = 0, ++ ZSTD_resetTarget_CCtx = 1, ++} ZSTD_resetTarget_e; ++ ++typedef struct { ++ U32 LLtype; ++ U32 Offtype; ++ U32 MLtype; ++ size_t size; ++ size_t lastCountSize; ++} ZSTD_symbolEncodingTypeStats_t; ++ ++enum { ++ ZSTDbss_compress = 0, ++ ZSTDbss_noCompress = 1, ++}; ++ ++typedef struct { ++ U32 *splitLocations; ++ size_t idx; ++} seqStoreSplits; ++ ++typedef struct { ++ U32 idx; ++ U32 posInSequence; ++ size_t posInSrc; ++} ZSTD_sequencePosition; ++ ++typedef size_t (*ZSTD_sequenceCopier)(ZSTD_CCtx *, ZSTD_sequencePosition *, const ZSTD_Sequence * const, size_t, const void *, size_t); ++ ++typedef enum { ++ ZSTD_error_no_error = 0, ++ ZSTD_error_GENERIC = 1, ++ ZSTD_error_prefix_unknown = 10, ++ ZSTD_error_version_unsupported = 12, ++ ZSTD_error_frameParameter_unsupported = 14, ++ ZSTD_error_frameParameter_windowTooLarge = 16, ++ ZSTD_error_corruption_detected = 20, ++ ZSTD_error_checksum_wrong = 22, ++ ZSTD_error_dictionary_corrupted = 30, ++ ZSTD_error_dictionary_wrong = 32, ++ ZSTD_error_dictionaryCreation_failed = 34, ++ ZSTD_error_parameter_unsupported = 40, ++ ZSTD_error_parameter_outOfBound = 42, ++ ZSTD_error_tableLog_tooLarge = 44, ++ ZSTD_error_maxSymbolValue_tooLarge = 46, ++ ZSTD_error_maxSymbolValue_tooSmall = 48, ++ ZSTD_error_stage_wrong = 60, ++ ZSTD_error_init_missing = 62, ++ ZSTD_error_memory_allocation = 64, ++ ZSTD_error_workSpace_tooSmall = 66, ++ ZSTD_error_dstSize_tooSmall = 70, ++ ZSTD_error_srcSize_wrong = 72, ++ ZSTD_error_dstBuffer_null = 74, ++ ZSTD_error_frameIndex_tooLarge = 100, ++ ZSTD_error_seekableIO = 102, ++ ZSTD_error_dstBuffer_wrong = 104, ++ ZSTD_error_srcBuffer_wrong = 105, ++ ZSTD_error_maxCode = 120, ++} ZSTD_ErrorCode; ++ ++typedef struct { ++ U16 nextState; ++ BYTE nbAdditionalBits; ++ BYTE nbBits; ++ U32 baseValue; ++} ZSTD_seqSymbol; ++ ++typedef U32 HUF_DTable; ++ ++typedef struct { ++ ZSTD_seqSymbol LLTable[513]; ++ ZSTD_seqSymbol OFTable[257]; ++ ZSTD_seqSymbol MLTable[513]; ++ HUF_DTable hufTable[4097]; ++ U32 rep[3]; ++ U32 workspace[157]; ++} ZSTD_entropyDTables_t; ++ ++typedef enum { ++ ZSTD_frame = 0, ++ ZSTD_skippableFrame = 1, ++} ZSTD_frameType_e; ++ ++typedef struct { ++ long long unsigned int frameContentSize; ++ long long unsigned int windowSize; ++ unsigned int blockSizeMax; ++ ZSTD_frameType_e frameType; ++ unsigned int headerSize; ++ unsigned int dictID; ++ unsigned int checksumFlag; ++} ZSTD_frameHeader; ++ ++typedef enum { ++ bt_raw = 0, ++ bt_rle = 1, ++ bt_compressed = 2, ++ bt_reserved = 3, ++} blockType_e; ++ ++typedef enum { ++ ZSTDds_getFrameHeaderSize = 0, ++ ZSTDds_decodeFrameHeader = 1, ++ ZSTDds_decodeBlockHeader = 2, ++ ZSTDds_decompressBlock = 3, ++ ZSTDds_decompressLastBlock = 4, ++ ZSTDds_checkChecksum = 5, ++ ZSTDds_decodeSkippableHeader = 6, ++ ZSTDds_skipFrame = 7, ++} ZSTD_dStage; ++ ++typedef enum { ++ ZSTD_d_validateChecksum = 0, ++ ZSTD_d_ignoreChecksum = 1, ++} ZSTD_forceIgnoreChecksum_e; ++ ++typedef enum { ++ ZSTD_use_indefinitely = -1, ++ ZSTD_dont_use = 0, ++ ZSTD_use_once = 1, ++} ZSTD_dictUses_e; ++ ++struct ZSTD_DDict_s; ++ ++typedef struct ZSTD_DDict_s ZSTD_DDict; ++ ++typedef struct { ++ const ZSTD_DDict **ddictPtrTable; ++ size_t ddictPtrTableSize; ++ size_t ddictPtrCount; ++} ZSTD_DDictHashSet; ++ ++typedef enum { ++ ZSTD_rmd_refSingleDDict = 0, ++ ZSTD_rmd_refMultipleDDicts = 1, ++} ZSTD_refMultipleDDicts_e; ++ ++typedef enum { ++ zdss_init = 0, ++ zdss_loadHeader = 1, ++ zdss_read = 2, ++ zdss_load = 3, ++ zdss_flush = 4, ++} ZSTD_dStreamStage; ++ ++typedef enum { ++ ZSTD_not_in_dst = 0, ++ ZSTD_in_dst = 1, ++ ZSTD_split = 2, ++} ZSTD_litLocation_e; ++ ++struct ZSTD_DCtx_s { ++ const ZSTD_seqSymbol *LLTptr; ++ const ZSTD_seqSymbol *MLTptr; ++ const ZSTD_seqSymbol *OFTptr; ++ const HUF_DTable *HUFptr; ++ ZSTD_entropyDTables_t entropy; ++ U32 workspace[640]; ++ const void *previousDstEnd; ++ const void *prefixStart; ++ const void *virtualStart; ++ const void *dictEnd; ++ size_t expected; ++ ZSTD_frameHeader fParams; ++ U64 processedCSize; ++ U64 decodedSize; ++ blockType_e bType; ++ ZSTD_dStage stage; ++ U32 litEntropy; ++ U32 fseEntropy; ++ struct xxh64_state xxhState; ++ size_t headerSize; ++ ZSTD_format_e format; ++ ZSTD_forceIgnoreChecksum_e forceIgnoreChecksum; ++ U32 validateChecksum; ++ const BYTE *litPtr; ++ ZSTD_customMem customMem; ++ size_t litSize; ++ size_t rleSize; ++ size_t staticSize; ++ ZSTD_DDict *ddictLocal; ++ const ZSTD_DDict *ddict; ++ U32 dictID; ++ int ddictIsCold; ++ ZSTD_dictUses_e dictUses; ++ ZSTD_DDictHashSet *ddictSet; ++ ZSTD_refMultipleDDicts_e refMultipleDDicts; ++ ZSTD_dStreamStage streamStage; ++ char *inBuff; ++ size_t inBuffSize; ++ size_t inPos; ++ size_t maxWindowSize; ++ char *outBuff; ++ size_t outBuffSize; ++ size_t outStart; ++ size_t outEnd; ++ size_t lhSize; ++ U32 hostageByte; ++ int noForwardProgress; ++ ZSTD_bufferMode_e outBufferMode; ++ ZSTD_outBuffer expectedOutBuffer; ++ BYTE *litBuffer; ++ const BYTE *litBufferEnd; ++ ZSTD_litLocation_e litBufferLocation; ++ BYTE litExtraBuffer[65568]; ++ BYTE headerBuffer[18]; ++ size_t oversizedDuration; ++}; ++ ++typedef struct ZSTD_DCtx_s ZSTD_DCtx; ++ ++typedef ZSTD_DCtx ZSTD_DStream; ++ ++typedef ZSTD_ErrorCode zstd_error_code; ++ ++typedef ZSTD_DCtx zstd_dctx; ++ ++typedef ZSTD_inBuffer zstd_in_buffer; ++ ++typedef ZSTD_outBuffer zstd_out_buffer; ++ ++typedef ZSTD_DStream zstd_dstream; ++ ++typedef ZSTD_frameHeader zstd_frame_header; ++ ++typedef struct { ++ size_t bitContainer; ++ unsigned int bitsConsumed; ++ const char *ptr; ++ const char *start; ++ const char *limitPtr; ++} BIT_DStream_t; ++ ++typedef enum { ++ BIT_DStream_unfinished = 0, ++ BIT_DStream_endOfBuffer = 1, ++ BIT_DStream_completed = 2, ++ BIT_DStream_overflow = 3, ++} BIT_DStream_status; ++ ++typedef struct { ++ BYTE maxTableLog; ++ BYTE tableType; ++ BYTE tableLog; ++ BYTE reserved; ++} DTableDesc; ++ ++typedef struct { ++ BYTE nbBits; ++ BYTE byte; ++} HUF_DEltX1; ++ ++typedef struct { ++ U32 rankVal[13]; ++ U32 rankStart[13]; ++ U32 statsWksp[218]; ++ BYTE symbols[256]; ++ BYTE huffWeight[256]; ++} HUF_ReadDTableX1_Workspace; ++ ++typedef struct { ++ U16 sequence; ++ BYTE nbBits; ++ BYTE length; ++} HUF_DEltX2; ++ ++typedef struct { ++ BYTE symbol; ++} sortedSymbol_t; ++ ++typedef U32 rankValCol_t[13]; ++ ++typedef struct { ++ U32 rankVal[156]; ++ U32 rankStats[13]; ++ U32 rankStart0[15]; ++ sortedSymbol_t sortedSymbol[256]; ++ BYTE weightList[256]; ++ U32 calleeWksp[218]; ++} HUF_ReadDTableX2_Workspace; ++ ++typedef struct { ++ U32 tableTime; ++ U32 decode256Time; ++} algo_time_t; ++ ++struct ZSTD_DDict_s { ++ void *dictBuffer; ++ const void *dictContent; ++ size_t dictSize; ++ ZSTD_entropyDTables_t entropy; ++ U32 dictID; ++ U32 entropyPresent; ++ ZSTD_customMem cMem; ++}; ++ ++typedef enum { ++ ZSTD_d_windowLogMax = 100, ++ ZSTD_d_experimentalParam1 = 1000, ++ ZSTD_d_experimentalParam2 = 1001, ++ ZSTD_d_experimentalParam3 = 1002, ++ ZSTD_d_experimentalParam4 = 1003, ++} ZSTD_dParameter; ++ ++typedef enum { ++ ZSTDnit_frameHeader = 0, ++ ZSTDnit_blockHeader = 1, ++ ZSTDnit_block = 2, ++ ZSTDnit_lastBlock = 3, ++ ZSTDnit_checksum = 4, ++ ZSTDnit_skippableFrame = 5, ++} ZSTD_nextInputType_e; ++ ++typedef struct { ++ size_t compressedSize; ++ long long unsigned int decompressedBound; ++} ZSTD_frameSizeInfo; ++ ++typedef struct { ++ blockType_e blockType; ++ U32 lastBlock; ++ U32 origSize; ++} blockProperties_t; ++ ++typedef enum { ++ not_streaming = 0, ++ is_streaming = 1, ++} streaming_operation; ++ ++typedef ZSTD_ErrorCode ERR_enum; ++ ++typedef unsigned int FSE_DTable; ++ ++typedef struct { ++ size_t state; ++ const void *table; ++} FSE_DState_t; ++ ++typedef struct { ++ U16 tableLog; ++ U16 fastMode; ++} FSE_DTableHeader; ++ ++typedef struct { ++ short unsigned int newState; ++ unsigned char symbol; ++ unsigned char nbBits; ++} FSE_decode_t; ++ ++typedef struct { ++ short int ncount[256]; ++ FSE_DTable dtable[1]; ++} FSE_DecompressWksp; ++ ++enum xz_mode { ++ XZ_SINGLE = 0, ++ XZ_PREALLOC = 1, ++ XZ_DYNALLOC = 2, ++}; ++ ++enum xz_ret { ++ XZ_OK = 0, ++ XZ_STREAM_END = 1, ++ XZ_UNSUPPORTED_CHECK = 2, ++ XZ_MEM_ERROR = 3, ++ XZ_MEMLIMIT_ERROR = 4, ++ XZ_FORMAT_ERROR = 5, ++ XZ_OPTIONS_ERROR = 6, ++ XZ_DATA_ERROR = 7, ++ XZ_BUF_ERROR = 8, ++}; ++ ++struct xz_buf { ++ const uint8_t *in; ++ size_t in_pos; ++ size_t in_size; ++ uint8_t *out; ++ size_t out_pos; ++ size_t out_size; ++}; ++ ++typedef uint64_t vli_type; ++ ++enum xz_check { ++ XZ_CHECK_NONE = 0, ++ XZ_CHECK_CRC32 = 1, ++ XZ_CHECK_CRC64 = 4, ++ XZ_CHECK_SHA256 = 10, ++}; ++ ++struct xz_dec_hash { ++ vli_type unpadded; ++ vli_type uncompressed; ++ uint32_t crc32; ++}; ++ ++struct xz_dec_lzma2; ++ ++struct xz_dec_bcj; ++ ++struct xz_dec { ++ enum { ++ SEQ_STREAM_HEADER = 0, ++ SEQ_BLOCK_START = 1, ++ SEQ_BLOCK_HEADER = 2, ++ SEQ_BLOCK_UNCOMPRESS = 3, ++ SEQ_BLOCK_PADDING = 4, ++ SEQ_BLOCK_CHECK = 5, ++ SEQ_INDEX = 6, ++ SEQ_INDEX_PADDING = 7, ++ SEQ_INDEX_CRC32 = 8, ++ SEQ_STREAM_FOOTER = 9, ++ } sequence; ++ uint32_t pos; ++ vli_type vli; ++ size_t in_start; ++ size_t out_start; ++ uint32_t crc32; ++ enum xz_check check_type; ++ enum xz_mode mode; ++ bool allow_buf_error; ++ struct { ++ vli_type compressed; ++ vli_type uncompressed; ++ uint32_t size; ++ } block_header; ++ struct { ++ vli_type compressed; ++ vli_type uncompressed; ++ vli_type count; ++ struct xz_dec_hash hash; ++ } block; ++ struct { ++ enum { ++ SEQ_INDEX_COUNT = 0, ++ SEQ_INDEX_UNPADDED = 1, ++ SEQ_INDEX_UNCOMPRESSED = 2, ++ } sequence; ++ vli_type size; ++ vli_type count; ++ struct xz_dec_hash hash; ++ } index; ++ struct { ++ size_t pos; ++ size_t size; ++ uint8_t buf[1024]; ++ } temp; ++ struct xz_dec_lzma2 *lzma2; ++ struct xz_dec_bcj *bcj; ++ bool bcj_active; ++}; ++ ++enum lzma_state { ++ STATE_LIT_LIT = 0, ++ STATE_MATCH_LIT_LIT = 1, ++ STATE_REP_LIT_LIT = 2, ++ STATE_SHORTREP_LIT_LIT = 3, ++ STATE_MATCH_LIT = 4, ++ STATE_REP_LIT = 5, ++ STATE_SHORTREP_LIT = 6, ++ STATE_LIT_MATCH = 7, ++ STATE_LIT_LONGREP = 8, ++ STATE_LIT_SHORTREP = 9, ++ STATE_NONLIT_MATCH = 10, ++ STATE_NONLIT_REP = 11, ++}; ++ ++struct dictionary { ++ uint8_t *buf; ++ size_t start; ++ size_t pos; ++ size_t full; ++ size_t limit; ++ size_t end; ++ uint32_t size; ++ uint32_t size_max; ++ uint32_t allocated; ++ enum xz_mode mode; ++}; ++ ++struct rc_dec { ++ uint32_t range; ++ uint32_t code; ++ uint32_t init_bytes_left; ++ const uint8_t *in; ++ size_t in_pos; ++ size_t in_limit; ++}; ++ ++struct lzma_len_dec { ++ uint16_t choice; ++ uint16_t choice2; ++ uint16_t low[128]; ++ uint16_t mid[128]; ++ uint16_t high[256]; ++}; ++ ++struct lzma_dec { ++ uint32_t rep0; ++ uint32_t rep1; ++ uint32_t rep2; ++ uint32_t rep3; ++ enum lzma_state state; ++ uint32_t len; ++ uint32_t lc; ++ uint32_t literal_pos_mask; ++ uint32_t pos_mask; ++ uint16_t is_match[192]; ++ uint16_t is_rep[12]; ++ uint16_t is_rep0[12]; ++ uint16_t is_rep1[12]; ++ uint16_t is_rep2[12]; ++ uint16_t is_rep0_long[192]; ++ uint16_t dist_slot[256]; ++ uint16_t dist_special[114]; ++ uint16_t dist_align[16]; ++ struct lzma_len_dec match_len_dec; ++ struct lzma_len_dec rep_len_dec; ++ uint16_t literal[12288]; ++}; ++ ++enum lzma2_seq { ++ SEQ_CONTROL = 0, ++ SEQ_UNCOMPRESSED_1 = 1, ++ SEQ_UNCOMPRESSED_2 = 2, ++ SEQ_COMPRESSED_0 = 3, ++ SEQ_COMPRESSED_1 = 4, ++ SEQ_PROPERTIES = 5, ++ SEQ_LZMA_PREPARE = 6, ++ SEQ_LZMA_RUN = 7, ++ SEQ_COPY = 8, ++}; ++ ++struct lzma2_dec { ++ enum lzma2_seq sequence; ++ enum lzma2_seq next_sequence; ++ uint32_t uncompressed; ++ uint32_t compressed; ++ bool need_dict_reset; ++ bool need_props; ++ bool pedantic_microlzma; ++}; ++ ++struct xz_dec_lzma2 { ++ struct rc_dec rc; ++ struct dictionary dict; ++ struct lzma2_dec lzma2; ++ struct lzma_dec lzma; ++ struct { ++ uint32_t size; ++ uint8_t buf[63]; ++ } temp; ++}; ++ ++struct xz_dec_microlzma { ++ struct xz_dec_lzma2 s; ++}; ++ ++struct xz_dec_bcj { ++ enum { ++ BCJ_X86 = 4, ++ BCJ_POWERPC = 5, ++ BCJ_IA64 = 6, ++ BCJ_ARM = 7, ++ BCJ_ARMTHUMB = 8, ++ BCJ_SPARC = 9, ++ } type; ++ enum xz_ret ret; ++ bool single_call; ++ uint32_t pos; ++ uint32_t x86_prev_mask; ++ uint8_t *out; ++ size_t out_pos; ++ size_t out_size; ++ struct { ++ size_t filtered; ++ size_t size; ++ uint8_t buf[16]; ++ } temp; ++}; ++ ++struct raid6_calls { ++ void (*gen_syndrome)(int, size_t, void **); ++ void (*xor_syndrome)(int, int, int, size_t, void **); ++ int (*valid)(); ++ const char *name; ++ int priority; ++}; ++ ++struct raid6_recov_calls { ++ void (*data2)(int, size_t, int, int, void **); ++ void (*datap)(int, size_t, int, void **); ++ int (*valid)(); ++ const char *name; ++ int priority; ++}; ++ ++typedef u64 unative_t; ++ ++typedef struct { ++ u8 _[256]; ++} addrtype___3; ++ ++struct ts_state { ++ unsigned int offset; ++ char cb[48]; ++}; ++ ++struct ts_config; ++ ++struct ts_ops { ++ const char *name; ++ struct ts_config * (*init)(const void *, unsigned int, gfp_t, int); ++ unsigned int (*find)(struct ts_config *, struct ts_state *); ++ void (*destroy)(struct ts_config *); ++ void * (*get_pattern)(struct ts_config *); ++ unsigned int (*get_pattern_len)(struct ts_config *); ++ struct module *owner; ++ struct list_head list; ++}; ++ ++struct ts_config { ++ struct ts_ops *ops; ++ int flags; ++ unsigned int (*get_next_block)(unsigned int, const u8 **, struct ts_config *, struct ts_state *); ++ void (*finish)(struct ts_config *, struct ts_state *); ++}; ++ ++struct ts_linear_state { ++ unsigned int len; ++ const void *data; ++}; ++ ++enum gcry_mpi_format { ++ GCRYMPI_FMT_NONE = 0, ++ GCRYMPI_FMT_STD = 1, ++ GCRYMPI_FMT_PGP = 2, ++ GCRYMPI_FMT_SSH = 3, ++ GCRYMPI_FMT_HEX = 4, ++ GCRYMPI_FMT_USG = 5, ++ GCRYMPI_FMT_OPAQUE = 8, ++}; ++ ++union handle_parts { ++ depot_stack_handle_t handle; ++ struct { ++ u32 pool_index: 16; ++ u32 offset: 10; ++ u32 valid: 1; ++ u32 extra: 5; ++ }; ++}; ++ ++struct stack_record { ++ struct stack_record *next; ++ u32 hash; ++ u32 size; ++ union handle_parts handle; ++ long unsigned int entries[0]; ++}; ++ ++struct resource_entry { ++ struct list_head node; ++ struct resource *res; ++ resource_size_t offset; ++ struct resource __res; ++}; ++ ++struct pci_sriov { ++ int pos; ++ int nres; ++ u32 cap; ++ u16 ctrl; ++ u16 total_VFs; ++ u16 initial_VFs; ++ u16 num_VFs; ++ u16 offset; ++ u16 stride; ++ u16 vf_device; ++ u32 pgsz; ++ u8 link; ++ u8 max_VF_buses; ++ u16 driver_max_VFs; ++ struct pci_dev *dev; ++ struct pci_dev *self; ++ u32 class; ++ u8 hdr_type; ++ u16 subsystem_vendor; ++ u16 subsystem_device; ++ resource_size_t barsz[6]; ++ bool drivers_autoprobe; ++}; ++ ++struct rcec_ea { ++ u8 nextbusn; ++ u8 lastbusn; ++ u32 bitmap; ++}; ++ ++struct pci_bus_resource { ++ struct list_head list; ++ struct resource *res; ++ unsigned int flags; ++}; ++ ++typedef u64 pci_bus_addr_t; ++ ++struct pci_bus_region { ++ pci_bus_addr_t start; ++ pci_bus_addr_t end; ++}; ++ ++enum pci_fixup_pass { ++ pci_fixup_early = 0, ++ pci_fixup_header = 1, ++ pci_fixup_final = 2, ++ pci_fixup_enable = 3, ++ pci_fixup_resume = 4, ++ pci_fixup_suspend = 5, ++ pci_fixup_resume_early = 6, ++ pci_fixup_suspend_late = 7, ++}; ++ ++struct pci_host_bridge { ++ struct device dev; ++ struct pci_bus *bus; ++ struct pci_ops *ops; ++ struct pci_ops *child_ops; ++ void *sysdata; ++ int busnr; ++ int domain_nr; ++ struct list_head windows; ++ struct list_head dma_ranges; ++ u8 (*swizzle_irq)(struct pci_dev *, u8 *); ++ int (*map_irq)(const struct pci_dev *, u8, u8); ++ void (*release_fn)(struct pci_host_bridge *); ++ void *release_data; ++ unsigned int ignore_reset_delay: 1; ++ unsigned int no_ext_tags: 1; ++ unsigned int no_inc_mrrs: 1; ++ unsigned int native_aer: 1; ++ unsigned int native_pcie_hotplug: 1; ++ unsigned int native_shpc_hotplug: 1; ++ unsigned int native_pme: 1; ++ unsigned int native_ltr: 1; ++ unsigned int native_dpc: 1; ++ unsigned int native_cxl_error: 1; ++ unsigned int preserve_config: 1; ++ unsigned int size_windows: 1; ++ unsigned int msi_domain: 1; ++ resource_size_t (*align_resource)(struct pci_dev *, const struct resource *, resource_size_t, resource_size_t, resource_size_t); ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long unsigned int private[0]; ++}; ++ ++enum pci_dev_flags { ++ PCI_DEV_FLAGS_MSI_INTX_DISABLE_BUG = 1, ++ PCI_DEV_FLAGS_NO_D3 = 2, ++ PCI_DEV_FLAGS_ASSIGNED = 4, ++ PCI_DEV_FLAGS_ACS_ENABLED_QUIRK = 8, ++ PCI_DEV_FLAG_PCIE_BRIDGE_ALIAS = 32, ++ PCI_DEV_FLAGS_NO_BUS_RESET = 64, ++ PCI_DEV_FLAGS_NO_PM_RESET = 128, ++ PCI_DEV_FLAGS_VPD_REF_F0 = 256, ++ PCI_DEV_FLAGS_BRIDGE_XLATE_ROOT = 512, ++ PCI_DEV_FLAGS_NO_FLR_RESET = 1024, ++ PCI_DEV_FLAGS_NO_RELAXED_ORDERING = 2048, ++ PCI_DEV_FLAGS_HAS_MSI_MASKING = 4096, ++}; ++ ++struct pci_cap_saved_data { ++ u16 cap_nr; ++ bool cap_extended; ++ unsigned int size; ++ u32 data[0]; ++}; ++ ++struct pci_cap_saved_state { ++ struct hlist_node next; ++ struct pci_cap_saved_data cap; ++}; ++ ++enum { ++ MSI_FLAG_USE_DEF_DOM_OPS = 1, ++ MSI_FLAG_USE_DEF_CHIP_OPS = 2, ++ MSI_FLAG_ACTIVATE_EARLY = 4, ++ MSI_FLAG_MUST_REACTIVATE = 8, ++ MSI_FLAG_DEV_SYSFS = 16, ++ MSI_FLAG_ALLOC_SIMPLE_MSI_DESCS = 32, ++ MSI_FLAG_FREE_MSI_DESCS = 64, ++ MSI_FLAG_NOMASK_QUIRK = 128, ++ MSI_GENERIC_FLAGS_MASK = 65535, ++ MSI_DOMAIN_FLAGS_MASK = 4294901760, ++ MSI_FLAG_MULTI_PCI_MSI = 65536, ++ MSI_FLAG_PCI_MSIX = 131072, ++ MSI_FLAG_LEVEL_CAPABLE = 262144, ++ MSI_FLAG_MSIX_CONTIGUOUS = 524288, ++ MSI_FLAG_PCI_MSIX_ALLOC_DYN = 1048576, ++ MSI_FLAG_PCI_IMS = 2097152, ++}; ++ ++enum pci_bus_flags { ++ PCI_BUS_FLAGS_NO_MSI = 1, ++ PCI_BUS_FLAGS_NO_MMRBC = 2, ++ PCI_BUS_FLAGS_NO_AERSID = 4, ++ PCI_BUS_FLAGS_NO_EXTCFG = 8, ++}; ++ ++struct msix_entry { ++ u32 vector; ++ u16 entry; ++}; ++ ++enum support_mode { ++ ALLOW_LEGACY = 0, ++ DENY_LEGACY = 1, ++}; ++ ++struct walk_rcec_data { ++ struct pci_dev *rcec; ++ int (*user_callback)(struct pci_dev *, void *); ++ void *user_data; ++}; ++ ++struct slot { ++ u8 number; ++ unsigned int devfn; ++ struct pci_bus *bus; ++ struct pci_dev *dev; ++ unsigned int latch_status: 1; ++ unsigned int adapter_status: 1; ++ unsigned int extracting; ++ struct hotplug_slot hotplug_slot; ++ struct list_head slot_list; ++}; ++ ++struct cpci_hp_controller_ops { ++ int (*query_enum)(); ++ int (*enable_irq)(); ++ int (*disable_irq)(); ++ int (*check_irq)(void *); ++ int (*hardware_test)(struct slot *, u32); ++ u8 (*get_power)(struct slot *); ++ int (*set_power)(struct slot *, int); ++}; ++ ++struct cpci_hp_controller { ++ unsigned int irq; ++ long unsigned int irq_flags; ++ char *devname; ++ void *dev_id; ++ char *name; ++ struct cpci_hp_controller_ops *ops; ++}; ++ ++struct pcie_device { ++ int irq; ++ struct pci_dev *port; ++ u32 service; ++ void *priv_data; ++ struct device device; ++}; ++ ++struct controller { ++ struct pcie_device *pcie; ++ u32 slot_cap; ++ unsigned int inband_presence_disabled: 1; ++ u16 slot_ctrl; ++ struct mutex ctrl_lock; ++ long unsigned int cmd_started; ++ unsigned int cmd_busy: 1; ++ wait_queue_head_t queue; ++ atomic_t pending_events; ++ unsigned int notification_enabled: 1; ++ unsigned int power_fault_detected; ++ struct task_struct *poll_thread; ++ u8 state; ++ struct mutex state_lock; ++ struct delayed_work button_work; ++ struct hotplug_slot hotplug_slot; ++ struct rw_semaphore reset_lock; ++ unsigned int depth; ++ unsigned int ist_running; ++ int request_result; ++ wait_queue_head_t requester; ++}; ++ ++struct dp_sdp_header { ++ u8 HB0; ++ u8 HB1; ++ u8 HB2; ++ u8 HB3; ++}; ++ ++struct dp_sdp { ++ struct dp_sdp_header sdp_header; ++ u8 db[32]; ++}; ++ ++enum hdmi_infoframe_type { ++ HDMI_INFOFRAME_TYPE_VENDOR = 129, ++ HDMI_INFOFRAME_TYPE_AVI = 130, ++ HDMI_INFOFRAME_TYPE_SPD = 131, ++ HDMI_INFOFRAME_TYPE_AUDIO = 132, ++ HDMI_INFOFRAME_TYPE_DRM = 135, ++}; ++ ++struct hdmi_any_infoframe { ++ enum hdmi_infoframe_type type; ++ unsigned char version; ++ unsigned char length; ++}; ++ ++enum hdmi_colorspace { ++ HDMI_COLORSPACE_RGB = 0, ++ HDMI_COLORSPACE_YUV422 = 1, ++ HDMI_COLORSPACE_YUV444 = 2, ++ HDMI_COLORSPACE_YUV420 = 3, ++ HDMI_COLORSPACE_RESERVED4 = 4, ++ HDMI_COLORSPACE_RESERVED5 = 5, ++ HDMI_COLORSPACE_RESERVED6 = 6, ++ HDMI_COLORSPACE_IDO_DEFINED = 7, ++}; ++ ++enum hdmi_scan_mode { ++ HDMI_SCAN_MODE_NONE = 0, ++ HDMI_SCAN_MODE_OVERSCAN = 1, ++ HDMI_SCAN_MODE_UNDERSCAN = 2, ++ HDMI_SCAN_MODE_RESERVED = 3, ++}; ++ ++enum hdmi_colorimetry { ++ HDMI_COLORIMETRY_NONE = 0, ++ HDMI_COLORIMETRY_ITU_601 = 1, ++ HDMI_COLORIMETRY_ITU_709 = 2, ++ HDMI_COLORIMETRY_EXTENDED = 3, ++}; ++ ++enum hdmi_picture_aspect { ++ HDMI_PICTURE_ASPECT_NONE = 0, ++ HDMI_PICTURE_ASPECT_4_3 = 1, ++ HDMI_PICTURE_ASPECT_16_9 = 2, ++ HDMI_PICTURE_ASPECT_64_27 = 3, ++ HDMI_PICTURE_ASPECT_256_135 = 4, ++ HDMI_PICTURE_ASPECT_RESERVED = 5, ++}; ++ ++enum hdmi_active_aspect { ++ HDMI_ACTIVE_ASPECT_16_9_TOP = 2, ++ HDMI_ACTIVE_ASPECT_14_9_TOP = 3, ++ HDMI_ACTIVE_ASPECT_16_9_CENTER = 4, ++ HDMI_ACTIVE_ASPECT_PICTURE = 8, ++ HDMI_ACTIVE_ASPECT_4_3 = 9, ++ HDMI_ACTIVE_ASPECT_16_9 = 10, ++ HDMI_ACTIVE_ASPECT_14_9 = 11, ++ HDMI_ACTIVE_ASPECT_4_3_SP_14_9 = 13, ++ HDMI_ACTIVE_ASPECT_16_9_SP_14_9 = 14, ++ HDMI_ACTIVE_ASPECT_16_9_SP_4_3 = 15, ++}; ++ ++enum hdmi_extended_colorimetry { ++ HDMI_EXTENDED_COLORIMETRY_XV_YCC_601 = 0, ++ HDMI_EXTENDED_COLORIMETRY_XV_YCC_709 = 1, ++ HDMI_EXTENDED_COLORIMETRY_S_YCC_601 = 2, ++ HDMI_EXTENDED_COLORIMETRY_OPYCC_601 = 3, ++ HDMI_EXTENDED_COLORIMETRY_OPRGB = 4, ++ HDMI_EXTENDED_COLORIMETRY_BT2020_CONST_LUM = 5, ++ HDMI_EXTENDED_COLORIMETRY_BT2020 = 6, ++ HDMI_EXTENDED_COLORIMETRY_RESERVED = 7, ++}; ++ ++enum hdmi_quantization_range { ++ HDMI_QUANTIZATION_RANGE_DEFAULT = 0, ++ HDMI_QUANTIZATION_RANGE_LIMITED = 1, ++ HDMI_QUANTIZATION_RANGE_FULL = 2, ++ HDMI_QUANTIZATION_RANGE_RESERVED = 3, ++}; ++ ++enum hdmi_nups { ++ HDMI_NUPS_UNKNOWN = 0, ++ HDMI_NUPS_HORIZONTAL = 1, ++ HDMI_NUPS_VERTICAL = 2, ++ HDMI_NUPS_BOTH = 3, ++}; ++ ++enum hdmi_ycc_quantization_range { ++ HDMI_YCC_QUANTIZATION_RANGE_LIMITED = 0, ++ HDMI_YCC_QUANTIZATION_RANGE_FULL = 1, ++}; ++ ++enum hdmi_content_type { ++ HDMI_CONTENT_TYPE_GRAPHICS = 0, ++ HDMI_CONTENT_TYPE_PHOTO = 1, ++ HDMI_CONTENT_TYPE_CINEMA = 2, ++ HDMI_CONTENT_TYPE_GAME = 3, ++}; ++ ++enum hdmi_metadata_type { ++ HDMI_STATIC_METADATA_TYPE1 = 0, ++}; ++ ++enum hdmi_eotf { ++ HDMI_EOTF_TRADITIONAL_GAMMA_SDR = 0, ++ HDMI_EOTF_TRADITIONAL_GAMMA_HDR = 1, ++ HDMI_EOTF_SMPTE_ST2084 = 2, ++ HDMI_EOTF_BT_2100_HLG = 3, ++}; ++ ++struct hdmi_avi_infoframe { ++ enum hdmi_infoframe_type type; ++ unsigned char version; ++ unsigned char length; ++ enum hdmi_colorspace colorspace; ++ enum hdmi_scan_mode scan_mode; ++ enum hdmi_colorimetry colorimetry; ++ enum hdmi_picture_aspect picture_aspect; ++ enum hdmi_active_aspect active_aspect; ++ bool itc; ++ enum hdmi_extended_colorimetry extended_colorimetry; ++ enum hdmi_quantization_range quantization_range; ++ enum hdmi_nups nups; ++ unsigned char video_code; ++ enum hdmi_ycc_quantization_range ycc_quantization_range; ++ enum hdmi_content_type content_type; ++ unsigned char pixel_repeat; ++ short unsigned int top_bar; ++ short unsigned int bottom_bar; ++ short unsigned int left_bar; ++ short unsigned int right_bar; ++}; ++ ++struct hdmi_drm_infoframe { ++ enum hdmi_infoframe_type type; ++ unsigned char version; ++ unsigned char length; ++ enum hdmi_eotf eotf; ++ enum hdmi_metadata_type metadata_type; ++ struct { ++ u16 x; ++ u16 y; ++ } display_primaries[3]; ++ struct { ++ u16 x; ++ u16 y; ++ } white_point; ++ u16 max_display_mastering_luminance; ++ u16 min_display_mastering_luminance; ++ u16 max_cll; ++ u16 max_fall; ++}; ++ ++enum hdmi_spd_sdi { ++ HDMI_SPD_SDI_UNKNOWN = 0, ++ HDMI_SPD_SDI_DSTB = 1, ++ HDMI_SPD_SDI_DVDP = 2, ++ HDMI_SPD_SDI_DVHS = 3, ++ HDMI_SPD_SDI_HDDVR = 4, ++ HDMI_SPD_SDI_DVC = 5, ++ HDMI_SPD_SDI_DSC = 6, ++ HDMI_SPD_SDI_VCD = 7, ++ HDMI_SPD_SDI_GAME = 8, ++ HDMI_SPD_SDI_PC = 9, ++ HDMI_SPD_SDI_BD = 10, ++ HDMI_SPD_SDI_SACD = 11, ++ HDMI_SPD_SDI_HDDVD = 12, ++ HDMI_SPD_SDI_PMP = 13, ++}; ++ ++struct hdmi_spd_infoframe { ++ enum hdmi_infoframe_type type; ++ unsigned char version; ++ unsigned char length; ++ char vendor[8]; ++ char product[16]; ++ enum hdmi_spd_sdi sdi; ++}; ++ ++enum hdmi_audio_coding_type { ++ HDMI_AUDIO_CODING_TYPE_STREAM = 0, ++ HDMI_AUDIO_CODING_TYPE_PCM = 1, ++ HDMI_AUDIO_CODING_TYPE_AC3 = 2, ++ HDMI_AUDIO_CODING_TYPE_MPEG1 = 3, ++ HDMI_AUDIO_CODING_TYPE_MP3 = 4, ++ HDMI_AUDIO_CODING_TYPE_MPEG2 = 5, ++ HDMI_AUDIO_CODING_TYPE_AAC_LC = 6, ++ HDMI_AUDIO_CODING_TYPE_DTS = 7, ++ HDMI_AUDIO_CODING_TYPE_ATRAC = 8, ++ HDMI_AUDIO_CODING_TYPE_DSD = 9, ++ HDMI_AUDIO_CODING_TYPE_EAC3 = 10, ++ HDMI_AUDIO_CODING_TYPE_DTS_HD = 11, ++ HDMI_AUDIO_CODING_TYPE_MLP = 12, ++ HDMI_AUDIO_CODING_TYPE_DST = 13, ++ HDMI_AUDIO_CODING_TYPE_WMA_PRO = 14, ++ HDMI_AUDIO_CODING_TYPE_CXT = 15, ++}; ++ ++enum hdmi_audio_sample_size { ++ HDMI_AUDIO_SAMPLE_SIZE_STREAM = 0, ++ HDMI_AUDIO_SAMPLE_SIZE_16 = 1, ++ HDMI_AUDIO_SAMPLE_SIZE_20 = 2, ++ HDMI_AUDIO_SAMPLE_SIZE_24 = 3, ++}; ++ ++enum hdmi_audio_sample_frequency { ++ HDMI_AUDIO_SAMPLE_FREQUENCY_STREAM = 0, ++ HDMI_AUDIO_SAMPLE_FREQUENCY_32000 = 1, ++ HDMI_AUDIO_SAMPLE_FREQUENCY_44100 = 2, ++ HDMI_AUDIO_SAMPLE_FREQUENCY_48000 = 3, ++ HDMI_AUDIO_SAMPLE_FREQUENCY_88200 = 4, ++ HDMI_AUDIO_SAMPLE_FREQUENCY_96000 = 5, ++ HDMI_AUDIO_SAMPLE_FREQUENCY_176400 = 6, ++ HDMI_AUDIO_SAMPLE_FREQUENCY_192000 = 7, ++}; ++ ++enum hdmi_audio_coding_type_ext { ++ HDMI_AUDIO_CODING_TYPE_EXT_CT = 0, ++ HDMI_AUDIO_CODING_TYPE_EXT_HE_AAC = 1, ++ HDMI_AUDIO_CODING_TYPE_EXT_HE_AAC_V2 = 2, ++ HDMI_AUDIO_CODING_TYPE_EXT_MPEG_SURROUND = 3, ++ HDMI_AUDIO_CODING_TYPE_EXT_MPEG4_HE_AAC = 4, ++ HDMI_AUDIO_CODING_TYPE_EXT_MPEG4_HE_AAC_V2 = 5, ++ HDMI_AUDIO_CODING_TYPE_EXT_MPEG4_AAC_LC = 6, ++ HDMI_AUDIO_CODING_TYPE_EXT_DRA = 7, ++ HDMI_AUDIO_CODING_TYPE_EXT_MPEG4_HE_AAC_SURROUND = 8, ++ HDMI_AUDIO_CODING_TYPE_EXT_MPEG4_AAC_LC_SURROUND = 10, ++}; ++ ++struct hdmi_audio_infoframe { ++ enum hdmi_infoframe_type type; ++ unsigned char version; ++ unsigned char length; ++ unsigned char channels; ++ enum hdmi_audio_coding_type coding_type; ++ enum hdmi_audio_sample_size sample_size; ++ enum hdmi_audio_sample_frequency sample_frequency; ++ enum hdmi_audio_coding_type_ext coding_type_ext; ++ unsigned char channel_allocation; ++ unsigned char level_shift_value; ++ bool downmix_inhibit; ++}; ++ ++enum hdmi_3d_structure { ++ HDMI_3D_STRUCTURE_INVALID = -1, ++ HDMI_3D_STRUCTURE_FRAME_PACKING = 0, ++ HDMI_3D_STRUCTURE_FIELD_ALTERNATIVE = 1, ++ HDMI_3D_STRUCTURE_LINE_ALTERNATIVE = 2, ++ HDMI_3D_STRUCTURE_SIDE_BY_SIDE_FULL = 3, ++ HDMI_3D_STRUCTURE_L_DEPTH = 4, ++ HDMI_3D_STRUCTURE_L_DEPTH_GFX_GFX_DEPTH = 5, ++ HDMI_3D_STRUCTURE_TOP_AND_BOTTOM = 6, ++ HDMI_3D_STRUCTURE_SIDE_BY_SIDE_HALF = 8, ++}; ++ ++struct hdmi_vendor_infoframe { ++ enum hdmi_infoframe_type type; ++ unsigned char version; ++ unsigned char length; ++ unsigned int oui; ++ u8 vic; ++ enum hdmi_3d_structure s3d_struct; ++ unsigned int s3d_ext_data; ++}; ++ ++union hdmi_vendor_any_infoframe { ++ struct { ++ enum hdmi_infoframe_type type; ++ unsigned char version; ++ unsigned char length; ++ unsigned int oui; ++ } any; ++ struct hdmi_vendor_infoframe hdmi; ++}; ++ ++union hdmi_infoframe { ++ struct hdmi_any_infoframe any; ++ struct hdmi_avi_infoframe avi; ++ struct hdmi_spd_infoframe spd; ++ union hdmi_vendor_any_infoframe vendor; ++ struct hdmi_audio_infoframe audio; ++ struct hdmi_drm_infoframe drm; ++}; ++ ++typedef unsigned int u_int; ++ ++struct fb_fix_screeninfo { ++ char id[16]; ++ long unsigned int smem_start; ++ __u32 smem_len; ++ __u32 type; ++ __u32 type_aux; ++ __u32 visual; ++ __u16 xpanstep; ++ __u16 ypanstep; ++ __u16 ywrapstep; ++ __u32 line_length; ++ long unsigned int mmio_start; ++ __u32 mmio_len; ++ __u32 accel; ++ __u16 capabilities; ++ __u16 reserved[2]; ++}; ++ ++struct fb_bitfield { ++ __u32 offset; ++ __u32 length; ++ __u32 msb_right; ++}; ++ ++struct fb_var_screeninfo { ++ __u32 xres; ++ __u32 yres; ++ __u32 xres_virtual; ++ __u32 yres_virtual; ++ __u32 xoffset; ++ __u32 yoffset; ++ __u32 bits_per_pixel; ++ __u32 grayscale; ++ struct fb_bitfield red; ++ struct fb_bitfield green; ++ struct fb_bitfield blue; ++ struct fb_bitfield transp; ++ __u32 nonstd; ++ __u32 activate; ++ __u32 height; ++ __u32 width; ++ __u32 accel_flags; ++ __u32 pixclock; ++ __u32 left_margin; ++ __u32 right_margin; ++ __u32 upper_margin; ++ __u32 lower_margin; ++ __u32 hsync_len; ++ __u32 vsync_len; ++ __u32 sync; ++ __u32 vmode; ++ __u32 rotate; ++ __u32 colorspace; ++ __u32 reserved[4]; ++}; ++ ++struct fb_cmap { ++ __u32 start; ++ __u32 len; ++ __u16 *red; ++ __u16 *green; ++ __u16 *blue; ++ __u16 *transp; ++}; ++ ++struct fb_copyarea { ++ __u32 dx; ++ __u32 dy; ++ __u32 width; ++ __u32 height; ++ __u32 sx; ++ __u32 sy; ++}; ++ ++struct fb_fillrect { ++ __u32 dx; ++ __u32 dy; ++ __u32 width; ++ __u32 height; ++ __u32 color; ++ __u32 rop; ++}; ++ ++struct fb_image { ++ __u32 dx; ++ __u32 dy; ++ __u32 width; ++ __u32 height; ++ __u32 fg_color; ++ __u32 bg_color; ++ __u8 depth; ++ const char *data; ++ struct fb_cmap cmap; ++}; ++ ++struct fbcurpos { ++ __u16 x; ++ __u16 y; ++}; ++ ++struct fb_cursor { ++ __u16 set; ++ __u16 enable; ++ __u16 rop; ++ const char *mask; ++ struct fbcurpos hot; ++ struct fb_image image; ++}; ++ ++struct fb_chroma { ++ __u32 redx; ++ __u32 greenx; ++ __u32 bluex; ++ __u32 whitex; ++ __u32 redy; ++ __u32 greeny; ++ __u32 bluey; ++ __u32 whitey; ++}; ++ ++struct fb_videomode; ++ ++struct fb_monspecs { ++ struct fb_chroma chroma; ++ struct fb_videomode *modedb; ++ __u8 manufacturer[4]; ++ __u8 monitor[14]; ++ __u8 serial_no[14]; ++ __u8 ascii[14]; ++ __u32 modedb_len; ++ __u32 model; ++ __u32 serial; ++ __u32 year; ++ __u32 week; ++ __u32 hfmin; ++ __u32 hfmax; ++ __u32 dclkmin; ++ __u32 dclkmax; ++ __u16 input; ++ __u16 dpms; ++ __u16 signal; ++ __u16 vfmin; ++ __u16 vfmax; ++ __u16 gamma; ++ __u16 gtf: 1; ++ __u16 misc; ++ __u8 version; ++ __u8 revision; ++ __u8 max_x; ++ __u8 max_y; ++}; ++ ++struct fb_info; ++ ++struct fb_pixmap { ++ u8 *addr; ++ u32 size; ++ u32 offset; ++ u32 buf_align; ++ u32 scan_align; ++ u32 access_align; ++ u32 flags; ++ u32 blit_x; ++ u32 blit_y; ++ void (*writeio)(struct fb_info *, void *, void *, unsigned int); ++ void (*readio)(struct fb_info *, void *, void *, unsigned int); ++}; ++ ++struct fb_deferred_io_pageref; ++ ++struct fb_deferred_io; ++ ++struct fb_ops; ++ ++struct fb_tile_ops; ++ ++struct fb_info { ++ refcount_t count; ++ int node; ++ int flags; ++ int fbcon_rotate_hint; ++ struct mutex lock; ++ struct mutex mm_lock; ++ struct fb_var_screeninfo var; ++ struct fb_fix_screeninfo fix; ++ struct fb_monspecs monspecs; ++ struct fb_pixmap pixmap; ++ struct fb_pixmap sprite; ++ struct fb_cmap cmap; ++ struct list_head modelist; ++ struct fb_videomode *mode; ++ struct delayed_work deferred_work; ++ long unsigned int npagerefs; ++ struct fb_deferred_io_pageref *pagerefs; ++ struct fb_deferred_io *fbdefio; ++ const struct fb_ops *fbops; ++ struct device *device; ++ struct device *dev; ++ int class_flag; ++ struct fb_tile_ops *tileops; ++ union { ++ char *screen_base; ++ char *screen_buffer; ++ }; ++ long unsigned int screen_size; ++ void *pseudo_palette; ++ u32 state; ++ void *fbcon_par; ++ void *par; ++ bool skip_vt_switch; ++}; ++ ++struct fb_videomode { ++ const char *name; ++ u32 refresh; ++ u32 xres; ++ u32 yres; ++ u32 pixclock; ++ u32 left_margin; ++ u32 right_margin; ++ u32 upper_margin; ++ u32 lower_margin; ++ u32 hsync_len; ++ u32 vsync_len; ++ u32 sync; ++ u32 vmode; ++ u32 flag; ++}; ++ ++struct fb_cmap_user { ++ __u32 start; ++ __u32 len; ++ __u16 *red; ++ __u16 *green; ++ __u16 *blue; ++ __u16 *transp; ++}; ++ ++struct fb_blit_caps { ++ u32 x; ++ u32 y; ++ u32 len; ++ u32 flags; ++}; ++ ++struct fb_deferred_io_pageref { ++ struct page *page; ++ long unsigned int offset; ++ struct list_head list; ++}; ++ ++struct fb_deferred_io { ++ long unsigned int delay; ++ bool sort_pagereflist; ++ int open_count; ++ struct mutex lock; ++ struct list_head pagereflist; ++ void (*deferred_io)(struct fb_info *, struct list_head *); ++}; ++ ++struct fb_ops { ++ struct module *owner; ++ int (*fb_open)(struct fb_info *, int); ++ int (*fb_release)(struct fb_info *, int); ++ ssize_t (*fb_read)(struct fb_info *, char *, size_t, loff_t *); ++ ssize_t (*fb_write)(struct fb_info *, const char *, size_t, loff_t *); ++ int (*fb_check_var)(struct fb_var_screeninfo *, struct fb_info *); ++ int (*fb_set_par)(struct fb_info *); ++ int (*fb_setcolreg)(unsigned int, unsigned int, unsigned int, unsigned int, unsigned int, struct fb_info *); ++ int (*fb_setcmap)(struct fb_cmap *, struct fb_info *); ++ int (*fb_blank)(int, struct fb_info *); ++ int (*fb_pan_display)(struct fb_var_screeninfo *, struct fb_info *); ++ void (*fb_fillrect)(struct fb_info *, const struct fb_fillrect *); ++ void (*fb_copyarea)(struct fb_info *, const struct fb_copyarea *); ++ void (*fb_imageblit)(struct fb_info *, const struct fb_image *); ++ int (*fb_cursor)(struct fb_info *, struct fb_cursor *); ++ int (*fb_sync)(struct fb_info *); ++ int (*fb_ioctl)(struct fb_info *, unsigned int, long unsigned int); ++ int (*fb_compat_ioctl)(struct fb_info *, unsigned int, long unsigned int); ++ int (*fb_mmap)(struct fb_info *, struct vm_area_struct *); ++ void (*fb_get_caps)(struct fb_info *, struct fb_blit_caps *, struct fb_var_screeninfo *); ++ void (*fb_destroy)(struct fb_info *); ++ int (*fb_debug_enter)(struct fb_info *); ++ int (*fb_debug_leave)(struct fb_info *); ++}; ++ ++struct fb_tilemap { ++ __u32 width; ++ __u32 height; ++ __u32 depth; ++ __u32 length; ++ const __u8 *data; ++}; ++ ++struct fb_tilerect { ++ __u32 sx; ++ __u32 sy; ++ __u32 width; ++ __u32 height; ++ __u32 index; ++ __u32 fg; ++ __u32 bg; ++ __u32 rop; ++}; ++ ++struct fb_tilearea { ++ __u32 sx; ++ __u32 sy; ++ __u32 dx; ++ __u32 dy; ++ __u32 width; ++ __u32 height; ++}; ++ ++struct fb_tileblit { ++ __u32 sx; ++ __u32 sy; ++ __u32 width; ++ __u32 height; ++ __u32 fg; ++ __u32 bg; ++ __u32 length; ++ __u32 *indices; ++}; ++ ++struct fb_tilecursor { ++ __u32 sx; ++ __u32 sy; ++ __u32 mode; ++ __u32 shape; ++ __u32 fg; ++ __u32 bg; ++}; ++ ++struct fb_tile_ops { ++ void (*fb_settile)(struct fb_info *, struct fb_tilemap *); ++ void (*fb_tilecopy)(struct fb_info *, struct fb_tilearea *); ++ void (*fb_tilefill)(struct fb_info *, struct fb_tilerect *); ++ void (*fb_tileblit)(struct fb_info *, struct fb_tileblit *); ++ void (*fb_tilecursor)(struct fb_info *, struct fb_tilecursor *); ++ int (*fb_get_tilemax)(struct fb_info *); ++}; ++ ++struct dmt_videomode { ++ u32 dmt_id; ++ u32 std_2byte_code; ++ u32 cvt_3byte_code; ++ const struct fb_videomode *mode; ++}; ++ ++struct fb_modelist { ++ struct list_head list; ++ struct fb_videomode mode; ++}; ++ ++typedef unsigned char u_char; ++ ++typedef short unsigned int u_short; ++ ++struct vt_mode { ++ char mode; ++ char waitv; ++ short int relsig; ++ short int acqsig; ++ short int frsig; ++}; ++ ++struct console_font { ++ unsigned int width; ++ unsigned int height; ++ unsigned int charcount; ++ unsigned char *data; ++}; ++ ++enum vc_intensity { ++ VCI_HALF_BRIGHT = 0, ++ VCI_NORMAL = 1, ++ VCI_BOLD = 2, ++ VCI_MASK = 3, ++}; ++ ++struct vc_state { ++ unsigned int x; ++ unsigned int y; ++ unsigned char color; ++ unsigned char Gx_charset[2]; ++ unsigned int charset: 1; ++ enum vc_intensity intensity; ++ bool italic; ++ bool underline; ++ bool blink; ++ bool reverse; ++}; ++ ++struct consw; ++ ++struct uni_pagedict; ++ ++struct vc_data { ++ struct tty_port port; ++ struct vc_state state; ++ struct vc_state saved_state; ++ short unsigned int vc_num; ++ unsigned int vc_cols; ++ unsigned int vc_rows; ++ unsigned int vc_size_row; ++ unsigned int vc_scan_lines; ++ unsigned int vc_cell_height; ++ long unsigned int vc_origin; ++ long unsigned int vc_scr_end; ++ long unsigned int vc_visible_origin; ++ unsigned int vc_top; ++ unsigned int vc_bottom; ++ const struct consw *vc_sw; ++ short unsigned int *vc_screenbuf; ++ unsigned int vc_screenbuf_size; ++ unsigned char vc_mode; ++ unsigned char vc_attr; ++ unsigned char vc_def_color; ++ unsigned char vc_ulcolor; ++ unsigned char vc_itcolor; ++ unsigned char vc_halfcolor; ++ unsigned int vc_cursor_type; ++ short unsigned int vc_complement_mask; ++ short unsigned int vc_s_complement_mask; ++ long unsigned int vc_pos; ++ short unsigned int vc_hi_font_mask; ++ struct console_font vc_font; ++ short unsigned int vc_video_erase_char; ++ unsigned int vc_state; ++ unsigned int vc_npar; ++ unsigned int vc_par[16]; ++ struct vt_mode vt_mode; ++ struct pid *vt_pid; ++ int vt_newvt; ++ wait_queue_head_t paste_wait; ++ unsigned int vc_disp_ctrl: 1; ++ unsigned int vc_toggle_meta: 1; ++ unsigned int vc_decscnm: 1; ++ unsigned int vc_decom: 1; ++ unsigned int vc_decawm: 1; ++ unsigned int vc_deccm: 1; ++ unsigned int vc_decim: 1; ++ unsigned int vc_priv: 3; ++ unsigned int vc_need_wrap: 1; ++ unsigned int vc_can_do_color: 1; ++ unsigned int vc_report_mouse: 2; ++ unsigned char vc_utf: 1; ++ unsigned char vc_utf_count; ++ int vc_utf_char; ++ long unsigned int vc_tab_stop[4]; ++ unsigned char vc_palette[48]; ++ short unsigned int *vc_translate; ++ unsigned int vc_resize_user; ++ unsigned int vc_bell_pitch; ++ unsigned int vc_bell_duration; ++ short unsigned int vc_cur_blink_ms; ++ struct vc_data **vc_display_fg; ++ struct uni_pagedict *uni_pagedict; ++ struct uni_pagedict **uni_pagedict_loc; ++ u32 **vc_uni_lines; ++}; ++ ++enum con_scroll { ++ SM_UP = 0, ++ SM_DOWN = 1, ++}; ++ ++struct consw { ++ struct module *owner; ++ const char * (*con_startup)(); ++ void (*con_init)(struct vc_data *, int); ++ void (*con_deinit)(struct vc_data *); ++ void (*con_clear)(struct vc_data *, int, int, int, int); ++ void (*con_putc)(struct vc_data *, int, int, int); ++ void (*con_putcs)(struct vc_data *, const short unsigned int *, int, int, int); ++ void (*con_cursor)(struct vc_data *, int); ++ bool (*con_scroll)(struct vc_data *, unsigned int, unsigned int, enum con_scroll, unsigned int); ++ int (*con_switch)(struct vc_data *); ++ int (*con_blank)(struct vc_data *, int, int); ++ int (*con_font_set)(struct vc_data *, struct console_font *, unsigned int, unsigned int); ++ int (*con_font_get)(struct vc_data *, struct console_font *, unsigned int); ++ int (*con_font_default)(struct vc_data *, struct console_font *, char *); ++ int (*con_resize)(struct vc_data *, unsigned int, unsigned int, unsigned int); ++ void (*con_set_palette)(struct vc_data *, const unsigned char *); ++ void (*con_scrolldelta)(struct vc_data *, int); ++ int (*con_set_origin)(struct vc_data *); ++ void (*con_save_screen)(struct vc_data *); ++ u8 (*con_build_attr)(struct vc_data *, u8, enum vc_intensity, bool, bool, bool, bool); ++ void (*con_invert_region)(struct vc_data *, u16 *, int); ++ u16 * (*con_screen_pos)(const struct vc_data *, int); ++ long unsigned int (*con_getxy)(struct vc_data *, long unsigned int, int *, int *); ++ void (*con_flush_scrollback)(struct vc_data *); ++ int (*con_debug_enter)(struct vc_data *); ++ int (*con_debug_leave)(struct vc_data *); ++}; ++ ++struct fbcon_display { ++ const u_char *fontdata; ++ int userfont; ++ u_short inverse; ++ short int yscroll; ++ int vrows; ++ int cursor_shape; ++ int con_rotate; ++ u32 xres_virtual; ++ u32 yres_virtual; ++ u32 height; ++ u32 width; ++ u32 bits_per_pixel; ++ u32 grayscale; ++ u32 nonstd; ++ u32 accel_flags; ++ u32 rotate; ++ struct fb_bitfield red; ++ struct fb_bitfield green; ++ struct fb_bitfield blue; ++ struct fb_bitfield transp; ++ const struct fb_videomode *mode; ++}; ++ ++struct fbcon_ops { ++ void (*bmove)(struct vc_data *, struct fb_info *, int, int, int, int, int, int); ++ void (*clear)(struct vc_data *, struct fb_info *, int, int, int, int); ++ void (*putcs)(struct vc_data *, struct fb_info *, const short unsigned int *, int, int, int, int, int); ++ void (*clear_margins)(struct vc_data *, struct fb_info *, int, int); ++ void (*cursor)(struct vc_data *, struct fb_info *, int, int, int); ++ int (*update_start)(struct fb_info *); ++ int (*rotate_font)(struct fb_info *, struct vc_data *); ++ struct fb_var_screeninfo var; ++ struct delayed_work cursor_work; ++ struct fb_cursor cursor_state; ++ struct fbcon_display *p; ++ struct fb_info *info; ++ int currcon; ++ int cur_blink_jiffies; ++ int cursor_flash; ++ int cursor_reset; ++ int blank_state; ++ int graphics; ++ int save_graphics; ++ bool initialized; ++ int rotate; ++ int cur_rotate; ++ char *cursor_data; ++ u8 *fontbuffer; ++ u8 *fontdata; ++ u8 *cursor_src; ++ u32 cursor_size; ++ u32 fd_size; ++}; ++ ++struct clk; ++ ++struct clk_hw; ++ ++struct clk_lookup { ++ struct list_head node; ++ const char *dev_id; ++ const char *con_id; ++ struct clk *clk; ++ struct clk_hw *clk_hw; ++}; ++ ++struct clk_core; ++ ++struct clk_init_data; ++ ++struct clk_hw { ++ struct clk_core *core; ++ struct clk *clk; ++ const struct clk_init_data *init; ++}; ++ ++struct clk_rate_request { ++ struct clk_core *core; ++ long unsigned int rate; ++ long unsigned int min_rate; ++ long unsigned int max_rate; ++ long unsigned int best_parent_rate; ++ struct clk_hw *best_parent_hw; ++}; ++ ++struct clk_duty { ++ unsigned int num; ++ unsigned int den; ++}; ++ ++struct clk_ops { ++ int (*prepare)(struct clk_hw *); ++ void (*unprepare)(struct clk_hw *); ++ int (*is_prepared)(struct clk_hw *); ++ void (*unprepare_unused)(struct clk_hw *); ++ int (*enable)(struct clk_hw *); ++ void (*disable)(struct clk_hw *); ++ int (*is_enabled)(struct clk_hw *); ++ void (*disable_unused)(struct clk_hw *); ++ int (*save_context)(struct clk_hw *); ++ void (*restore_context)(struct clk_hw *); ++ long unsigned int (*recalc_rate)(struct clk_hw *, long unsigned int); ++ long int (*round_rate)(struct clk_hw *, long unsigned int, long unsigned int *); ++ int (*determine_rate)(struct clk_hw *, struct clk_rate_request *); ++ int (*set_parent)(struct clk_hw *, u8); ++ u8 (*get_parent)(struct clk_hw *); ++ int (*set_rate)(struct clk_hw *, long unsigned int, long unsigned int); ++ int (*set_rate_and_parent)(struct clk_hw *, long unsigned int, long unsigned int, u8); ++ long unsigned int (*recalc_accuracy)(struct clk_hw *, long unsigned int); ++ int (*get_phase)(struct clk_hw *); ++ int (*set_phase)(struct clk_hw *, int); ++ int (*get_duty_cycle)(struct clk_hw *, struct clk_duty *); ++ int (*set_duty_cycle)(struct clk_hw *, struct clk_duty *); ++ int (*init)(struct clk_hw *); ++ void (*terminate)(struct clk_hw *); ++ void (*debug_init)(struct clk_hw *, struct dentry *); ++}; ++ ++struct clk_parent_data { ++ const struct clk_hw *hw; ++ const char *fw_name; ++ const char *name; ++ int index; ++}; ++ ++struct clk_init_data { ++ const char *name; ++ const struct clk_ops *ops; ++ const char * const *parent_names; ++ const struct clk_parent_data *parent_data; ++ const struct clk_hw **parent_hws; ++ u8 num_parents; ++ long unsigned int flags; ++}; ++ ++struct clk_lookup_alloc { ++ struct clk_lookup cl; ++ char dev_id[20]; ++ char con_id[16]; ++}; ++ ++struct clk_fixed_rate { ++ struct clk_hw hw; ++ long unsigned int fixed_rate; ++ long unsigned int fixed_accuracy; ++ long unsigned int flags; ++}; ++ ++struct clk_mux { ++ struct clk_hw hw; ++ void *reg; ++ const u32 *table; ++ u32 mask; ++ u8 shift; ++ u8 flags; ++ spinlock_t *lock; ++}; ++ ++typedef s32 dma_cookie_t; ++ ++enum dma_status { ++ DMA_COMPLETE = 0, ++ DMA_IN_PROGRESS = 1, ++ DMA_PAUSED = 2, ++ DMA_ERROR = 3, ++ DMA_OUT_OF_ORDER = 4, ++}; ++ ++enum dma_transaction_type { ++ DMA_MEMCPY = 0, ++ DMA_XOR = 1, ++ DMA_PQ = 2, ++ DMA_XOR_VAL = 3, ++ DMA_PQ_VAL = 4, ++ DMA_MEMSET = 5, ++ DMA_MEMSET_SG = 6, ++ DMA_INTERRUPT = 7, ++ DMA_PRIVATE = 8, ++ DMA_ASYNC_TX = 9, ++ DMA_SLAVE = 10, ++ DMA_CYCLIC = 11, ++ DMA_INTERLEAVE = 12, ++ DMA_COMPLETION_NO_ORDER = 13, ++ DMA_REPEAT = 14, ++ DMA_LOAD_EOT = 15, ++ DMA_TX_TYPE_END = 16, ++}; ++ ++enum dma_transfer_direction { ++ DMA_MEM_TO_MEM = 0, ++ DMA_MEM_TO_DEV = 1, ++ DMA_DEV_TO_MEM = 2, ++ DMA_DEV_TO_DEV = 3, ++ DMA_TRANS_NONE = 4, ++}; ++ ++struct data_chunk { ++ size_t size; ++ size_t icg; ++ size_t dst_icg; ++ size_t src_icg; ++}; ++ ++struct dma_interleaved_template { ++ dma_addr_t src_start; ++ dma_addr_t dst_start; ++ enum dma_transfer_direction dir; ++ bool src_inc; ++ bool dst_inc; ++ bool src_sgl; ++ bool dst_sgl; ++ size_t numf; ++ size_t frame_size; ++ struct data_chunk sgl[0]; ++}; ++ ++enum dma_ctrl_flags { ++ DMA_PREP_INTERRUPT = 1, ++ DMA_CTRL_ACK = 2, ++ DMA_PREP_PQ_DISABLE_P = 4, ++ DMA_PREP_PQ_DISABLE_Q = 8, ++ DMA_PREP_CONTINUE = 16, ++ DMA_PREP_FENCE = 32, ++ DMA_CTRL_REUSE = 64, ++ DMA_PREP_CMD = 128, ++ DMA_PREP_REPEAT = 256, ++ DMA_PREP_LOAD_EOT = 512, ++}; ++ ++enum sum_check_bits { ++ SUM_CHECK_P = 0, ++ SUM_CHECK_Q = 1, ++}; ++ ++enum sum_check_flags { ++ SUM_CHECK_P_RESULT = 1, ++ SUM_CHECK_Q_RESULT = 2, ++}; ++ ++typedef struct { ++ long unsigned int bits[1]; ++} dma_cap_mask_t; ++ ++enum dma_desc_metadata_mode { ++ DESC_METADATA_NONE = 0, ++ DESC_METADATA_CLIENT = 1, ++ DESC_METADATA_ENGINE = 2, ++}; ++ ++struct dma_chan_percpu { ++ long unsigned int memcpy_count; ++ long unsigned int bytes_transferred; ++}; ++ ++struct dma_router { ++ struct device *dev; ++ void (*route_free)(struct device *, void *); ++}; ++ ++struct dma_device; ++ ++struct dma_chan_dev; ++ ++struct dma_chan { ++ struct dma_device *device; ++ struct device *slave; ++ dma_cookie_t cookie; ++ dma_cookie_t completed_cookie; ++ int chan_id; ++ struct dma_chan_dev *dev; ++ const char *name; ++ char *dbg_client_name; ++ struct list_head device_node; ++ struct dma_chan_percpu *local; ++ int client_count; ++ int table_count; ++ struct dma_router *router; ++ void *route_data; ++ void *private; ++}; ++ ++typedef bool (*dma_filter_fn)(struct dma_chan *, void *); ++ ++struct dma_slave_map; ++ ++struct dma_filter { ++ dma_filter_fn fn; ++ int mapcnt; ++ const struct dma_slave_map *map; ++}; ++ ++enum dmaengine_alignment { ++ DMAENGINE_ALIGN_1_BYTE = 0, ++ DMAENGINE_ALIGN_2_BYTES = 1, ++ DMAENGINE_ALIGN_4_BYTES = 2, ++ DMAENGINE_ALIGN_8_BYTES = 3, ++ DMAENGINE_ALIGN_16_BYTES = 4, ++ DMAENGINE_ALIGN_32_BYTES = 5, ++ DMAENGINE_ALIGN_64_BYTES = 6, ++ DMAENGINE_ALIGN_128_BYTES = 7, ++ DMAENGINE_ALIGN_256_BYTES = 8, ++}; ++ ++enum dma_residue_granularity { ++ DMA_RESIDUE_GRANULARITY_DESCRIPTOR = 0, ++ DMA_RESIDUE_GRANULARITY_SEGMENT = 1, ++ DMA_RESIDUE_GRANULARITY_BURST = 2, ++}; ++ ++struct dma_async_tx_descriptor; ++ ++struct dma_slave_caps; ++ ++struct dma_slave_config; ++ ++struct dma_tx_state; ++ ++struct dma_device { ++ struct kref ref; ++ unsigned int chancnt; ++ unsigned int privatecnt; ++ struct list_head channels; ++ struct list_head global_node; ++ struct dma_filter filter; ++ dma_cap_mask_t cap_mask; ++ enum dma_desc_metadata_mode desc_metadata_modes; ++ short unsigned int max_xor; ++ short unsigned int max_pq; ++ enum dmaengine_alignment copy_align; ++ enum dmaengine_alignment xor_align; ++ enum dmaengine_alignment pq_align; ++ enum dmaengine_alignment fill_align; ++ int dev_id; ++ struct device *dev; ++ struct module *owner; ++ struct ida chan_ida; ++ u32 src_addr_widths; ++ u32 dst_addr_widths; ++ u32 directions; ++ u32 min_burst; ++ u32 max_burst; ++ u32 max_sg_burst; ++ bool descriptor_reuse; ++ enum dma_residue_granularity residue_granularity; ++ int (*device_alloc_chan_resources)(struct dma_chan *); ++ int (*device_router_config)(struct dma_chan *); ++ void (*device_free_chan_resources)(struct dma_chan *); ++ struct dma_async_tx_descriptor * (*device_prep_dma_memcpy)(struct dma_chan *, dma_addr_t, dma_addr_t, size_t, long unsigned int); ++ struct dma_async_tx_descriptor * (*device_prep_dma_xor)(struct dma_chan *, dma_addr_t, dma_addr_t *, unsigned int, size_t, long unsigned int); ++ struct dma_async_tx_descriptor * (*device_prep_dma_xor_val)(struct dma_chan *, dma_addr_t *, unsigned int, size_t, enum sum_check_flags *, long unsigned int); ++ struct dma_async_tx_descriptor * (*device_prep_dma_pq)(struct dma_chan *, dma_addr_t *, dma_addr_t *, unsigned int, const unsigned char *, size_t, long unsigned int); ++ struct dma_async_tx_descriptor * (*device_prep_dma_pq_val)(struct dma_chan *, dma_addr_t *, dma_addr_t *, unsigned int, const unsigned char *, size_t, enum sum_check_flags *, long unsigned int); ++ struct dma_async_tx_descriptor * (*device_prep_dma_memset)(struct dma_chan *, dma_addr_t, int, size_t, long unsigned int); ++ struct dma_async_tx_descriptor * (*device_prep_dma_memset_sg)(struct dma_chan *, struct scatterlist *, unsigned int, int, long unsigned int); ++ struct dma_async_tx_descriptor * (*device_prep_dma_interrupt)(struct dma_chan *, long unsigned int); ++ struct dma_async_tx_descriptor * (*device_prep_slave_sg)(struct dma_chan *, struct scatterlist *, unsigned int, enum dma_transfer_direction, long unsigned int, void *); ++ struct dma_async_tx_descriptor * (*device_prep_dma_cyclic)(struct dma_chan *, dma_addr_t, size_t, size_t, enum dma_transfer_direction, long unsigned int); ++ struct dma_async_tx_descriptor * (*device_prep_interleaved_dma)(struct dma_chan *, struct dma_interleaved_template *, long unsigned int); ++ struct dma_async_tx_descriptor * (*device_prep_dma_imm_data)(struct dma_chan *, dma_addr_t, u64, long unsigned int); ++ void (*device_caps)(struct dma_chan *, struct dma_slave_caps *); ++ int (*device_config)(struct dma_chan *, struct dma_slave_config *); ++ int (*device_pause)(struct dma_chan *); ++ int (*device_resume)(struct dma_chan *); ++ int (*device_terminate_all)(struct dma_chan *); ++ void (*device_synchronize)(struct dma_chan *); ++ enum dma_status (*device_tx_status)(struct dma_chan *, dma_cookie_t, struct dma_tx_state *); ++ void (*device_issue_pending)(struct dma_chan *); ++ void (*device_release)(struct dma_device *); ++ void (*dbg_summary_show)(struct seq_file *, struct dma_device *); ++ struct dentry *dbg_dev_root; ++}; ++ ++struct dma_chan_dev { ++ struct dma_chan *chan; ++ struct device device; ++ int dev_id; ++ bool chan_dma_dev; ++}; ++ ++enum dma_slave_buswidth { ++ DMA_SLAVE_BUSWIDTH_UNDEFINED = 0, ++ DMA_SLAVE_BUSWIDTH_1_BYTE = 1, ++ DMA_SLAVE_BUSWIDTH_2_BYTES = 2, ++ DMA_SLAVE_BUSWIDTH_3_BYTES = 3, ++ DMA_SLAVE_BUSWIDTH_4_BYTES = 4, ++ DMA_SLAVE_BUSWIDTH_8_BYTES = 8, ++ DMA_SLAVE_BUSWIDTH_16_BYTES = 16, ++ DMA_SLAVE_BUSWIDTH_32_BYTES = 32, ++ DMA_SLAVE_BUSWIDTH_64_BYTES = 64, ++ DMA_SLAVE_BUSWIDTH_128_BYTES = 128, ++}; ++ ++struct dma_slave_config { ++ enum dma_transfer_direction direction; ++ phys_addr_t src_addr; ++ phys_addr_t dst_addr; ++ enum dma_slave_buswidth src_addr_width; ++ enum dma_slave_buswidth dst_addr_width; ++ u32 src_maxburst; ++ u32 dst_maxburst; ++ u32 src_port_window_size; ++ u32 dst_port_window_size; ++ bool device_fc; ++ void *peripheral_config; ++ size_t peripheral_size; ++}; ++ ++struct dma_slave_caps { ++ u32 src_addr_widths; ++ u32 dst_addr_widths; ++ u32 directions; ++ u32 min_burst; ++ u32 max_burst; ++ u32 max_sg_burst; ++ bool cmd_pause; ++ bool cmd_resume; ++ bool cmd_terminate; ++ enum dma_residue_granularity residue_granularity; ++ bool descriptor_reuse; ++}; ++ ++typedef void (*dma_async_tx_callback)(void *); ++ ++enum dmaengine_tx_result { ++ DMA_TRANS_NOERROR = 0, ++ DMA_TRANS_READ_FAILED = 1, ++ DMA_TRANS_WRITE_FAILED = 2, ++ DMA_TRANS_ABORTED = 3, ++}; ++ ++struct dmaengine_result { ++ enum dmaengine_tx_result result; ++ u32 residue; ++}; ++ ++typedef void (*dma_async_tx_callback_result)(void *, const struct dmaengine_result *); ++ ++struct dmaengine_unmap_data { ++ u8 map_cnt; ++ u8 to_cnt; ++ u8 from_cnt; ++ u8 bidi_cnt; ++ struct device *dev; ++ struct kref kref; ++ size_t len; ++ dma_addr_t addr[0]; ++}; ++ ++struct dma_descriptor_metadata_ops { ++ int (*attach)(struct dma_async_tx_descriptor *, void *, size_t); ++ void * (*get_ptr)(struct dma_async_tx_descriptor *, size_t *, size_t *); ++ int (*set_len)(struct dma_async_tx_descriptor *, size_t); ++}; ++ ++struct dma_async_tx_descriptor { ++ dma_cookie_t cookie; ++ enum dma_ctrl_flags flags; ++ dma_addr_t phys; ++ struct dma_chan *chan; ++ dma_cookie_t (*tx_submit)(struct dma_async_tx_descriptor *); ++ int (*desc_free)(struct dma_async_tx_descriptor *); ++ dma_async_tx_callback callback; ++ dma_async_tx_callback_result callback_result; ++ void *callback_param; ++ struct dmaengine_unmap_data *unmap; ++ enum dma_desc_metadata_mode desc_metadata_mode; ++ struct dma_descriptor_metadata_ops *metadata_ops; ++}; ++ ++struct dma_tx_state { ++ dma_cookie_t last; ++ dma_cookie_t used; ++ u32 residue; ++ u32 in_flight_bytes; ++}; ++ ++struct dma_slave_map { ++ const char *devname; ++ const char *slave; ++ void *param; ++}; ++ ++struct dma_chan_tbl_ent { ++ struct dma_chan *chan; ++}; ++ ++struct dmaengine_unmap_pool { ++ struct kmem_cache *cache; ++ const char *name; ++ mempool_t *pool; ++ size_t size; ++}; ++ ++struct module_version_attribute { ++ struct module_attribute mattr; ++ const char *module_name; ++ const char *version; ++}; ++ ++struct virtio_pci_common_cfg { ++ __le32 device_feature_select; ++ __le32 device_feature; ++ __le32 guest_feature_select; ++ __le32 guest_feature; ++ __le16 msix_config; ++ __le16 num_queues; ++ __u8 device_status; ++ __u8 config_generation; ++ __le16 queue_select; ++ __le16 queue_size; ++ __le16 queue_msix_vector; ++ __le16 queue_enable; ++ __le16 queue_notify_off; ++ __le32 queue_desc_lo; ++ __le32 queue_desc_hi; ++ __le32 queue_avail_lo; ++ __le32 queue_avail_hi; ++ __le32 queue_used_lo; ++ __le32 queue_used_hi; ++}; ++ ++struct virtio_pci_modern_common_cfg { ++ struct virtio_pci_common_cfg cfg; ++ __le16 queue_notify_data; ++ __le16 queue_reset; ++}; ++ ++struct virtio_pci_modern_device { ++ struct pci_dev *pci_dev; ++ struct virtio_pci_common_cfg *common; ++ void *device; ++ void *notify_base; ++ resource_size_t notify_pa; ++ u8 *isr; ++ size_t notify_len; ++ size_t device_len; ++ int notify_map_cap; ++ u32 notify_offset_multiplier; ++ int modern_bars; ++ struct virtio_device_id id; ++}; ++ ++struct virtio_pci_legacy_device { ++ struct pci_dev *pci_dev; ++ u8 *isr; ++ void *ioaddr; ++ struct virtio_device_id id; ++}; ++ ++struct virtio_pci_vq_info { ++ struct virtqueue *vq; ++ struct list_head node; ++ unsigned int msix_vector; ++}; ++ ++struct virtio_pci_device { ++ struct virtio_device vdev; ++ struct pci_dev *pci_dev; ++ struct virtio_pci_legacy_device ldev; ++ struct virtio_pci_modern_device mdev; ++ bool is_legacy; ++ u8 *isr; ++ spinlock_t lock; ++ struct list_head virtqueues; ++ struct virtio_pci_vq_info **vqs; ++ int msix_enabled; ++ int intx_enabled; ++ cpumask_var_t *msix_affinity_masks; ++ char (*msix_names)[256]; ++ unsigned int msix_vectors; ++ unsigned int msix_used_vectors; ++ bool per_vq_vectors; ++ struct virtqueue * (*setup_vq)(struct virtio_pci_device *, struct virtio_pci_vq_info *, unsigned int, void (*)(struct virtqueue *), const char *, bool, u16); ++ void (*del_vq)(struct virtio_pci_vq_info *); ++ u16 (*config_vector)(struct virtio_pci_device *, u16); ++}; ++ ++struct pdev_archdata {}; ++ ++struct mfd_cell; ++ ++struct platform_device_id; ++ ++struct platform_device { ++ const char *name; ++ int id; ++ bool id_auto; ++ struct device dev; ++ u64 platform_dma_mask; ++ struct device_dma_parameters dma_parms; ++ u32 num_resources; ++ struct resource *resource; ++ const struct platform_device_id *id_entry; ++ const char *driver_override; ++ struct mfd_cell *mfd_cell; ++ struct pdev_archdata archdata; ++}; ++ ++struct platform_device_id { ++ char name[20]; ++ kernel_ulong_t driver_data; ++}; ++ ++struct platform_driver { ++ int (*probe)(struct platform_device *); ++ int (*remove)(struct platform_device *); ++ void (*remove_new)(struct platform_device *); ++ void (*shutdown)(struct platform_device *); ++ int (*suspend)(struct platform_device *, pm_message_t); ++ int (*resume)(struct platform_device *); ++ struct device_driver driver; ++ const struct platform_device_id *id_table; ++ bool prevent_deferred_probe; ++ bool driver_managed_dma; ++}; ++ ++struct linear_range { ++ unsigned int min; ++ unsigned int min_sel; ++ unsigned int max_sel; ++ unsigned int step; ++}; ++ ++typedef int suspend_state_t; ++ ++struct regulator_dev; ++ ++struct regulator_ops { ++ int (*list_voltage)(struct regulator_dev *, unsigned int); ++ int (*set_voltage)(struct regulator_dev *, int, int, unsigned int *); ++ int (*map_voltage)(struct regulator_dev *, int, int); ++ int (*set_voltage_sel)(struct regulator_dev *, unsigned int); ++ int (*get_voltage)(struct regulator_dev *); ++ int (*get_voltage_sel)(struct regulator_dev *); ++ int (*set_current_limit)(struct regulator_dev *, int, int); ++ int (*get_current_limit)(struct regulator_dev *); ++ int (*set_input_current_limit)(struct regulator_dev *, int); ++ int (*set_over_current_protection)(struct regulator_dev *, int, int, bool); ++ int (*set_over_voltage_protection)(struct regulator_dev *, int, int, bool); ++ int (*set_under_voltage_protection)(struct regulator_dev *, int, int, bool); ++ int (*set_thermal_protection)(struct regulator_dev *, int, int, bool); ++ int (*set_active_discharge)(struct regulator_dev *, bool); ++ int (*enable)(struct regulator_dev *); ++ int (*disable)(struct regulator_dev *); ++ int (*is_enabled)(struct regulator_dev *); ++ int (*set_mode)(struct regulator_dev *, unsigned int); ++ unsigned int (*get_mode)(struct regulator_dev *); ++ int (*get_error_flags)(struct regulator_dev *, unsigned int *); ++ int (*enable_time)(struct regulator_dev *); ++ int (*set_ramp_delay)(struct regulator_dev *, int); ++ int (*set_voltage_time)(struct regulator_dev *, int, int); ++ int (*set_voltage_time_sel)(struct regulator_dev *, unsigned int, unsigned int); ++ int (*set_soft_start)(struct regulator_dev *); ++ int (*get_status)(struct regulator_dev *); ++ unsigned int (*get_optimum_mode)(struct regulator_dev *, int, int, int); ++ int (*set_load)(struct regulator_dev *, int); ++ int (*set_bypass)(struct regulator_dev *, bool); ++ int (*get_bypass)(struct regulator_dev *, bool *); ++ int (*set_suspend_voltage)(struct regulator_dev *, int); ++ int (*set_suspend_enable)(struct regulator_dev *); ++ int (*set_suspend_disable)(struct regulator_dev *); ++ int (*set_suspend_mode)(struct regulator_dev *, unsigned int); ++ int (*resume)(struct regulator_dev *); ++ int (*set_pull_down)(struct regulator_dev *); ++}; ++ ++struct regulator_coupler; ++ ++struct coupling_desc { ++ struct regulator_dev **coupled_rdevs; ++ struct regulator_coupler *coupler; ++ int n_resolved; ++ int n_coupled; ++}; ++ ++struct regulator_desc; ++ ++struct regulation_constraints; ++ ++struct regulator; ++ ++struct regmap; ++ ++struct regulator_enable_gpio; ++ ++struct regulator_dev { ++ const struct regulator_desc *desc; ++ int exclusive; ++ u32 use_count; ++ u32 open_count; ++ u32 bypass_count; ++ struct list_head list; ++ struct list_head consumer_list; ++ struct coupling_desc coupling_desc; ++ struct blocking_notifier_head notifier; ++ struct ww_mutex mutex; ++ struct task_struct *mutex_owner; ++ int ref_cnt; ++ struct module *owner; ++ struct device dev; ++ struct regulation_constraints *constraints; ++ struct regulator *supply; ++ const char *supply_name; ++ struct regmap *regmap; ++ struct delayed_work disable_work; ++ void *reg_data; ++ struct dentry *debugfs; ++ struct regulator_enable_gpio *ena_pin; ++ unsigned int ena_gpio_state: 1; ++ unsigned int is_switch: 1; ++ ktime_t last_off; ++ int cached_err; ++ bool use_cached_err; ++ spinlock_t err_lock; ++}; ++ ++enum regulator_type { ++ REGULATOR_VOLTAGE = 0, ++ REGULATOR_CURRENT = 1, ++}; ++ ++struct regulator_config; ++ ++struct regulator_desc { ++ const char *name; ++ const char *supply_name; ++ const char *of_match; ++ bool of_match_full_name; ++ const char *regulators_node; ++ int (*of_parse_cb)(struct device_node *, const struct regulator_desc *, struct regulator_config *); ++ int id; ++ unsigned int continuous_voltage_range: 1; ++ unsigned int n_voltages; ++ unsigned int n_current_limits; ++ const struct regulator_ops *ops; ++ int irq; ++ enum regulator_type type; ++ struct module *owner; ++ unsigned int min_uV; ++ unsigned int uV_step; ++ unsigned int linear_min_sel; ++ int fixed_uV; ++ unsigned int ramp_delay; ++ int min_dropout_uV; ++ const struct linear_range *linear_ranges; ++ const unsigned int *linear_range_selectors; ++ int n_linear_ranges; ++ const unsigned int *volt_table; ++ const unsigned int *curr_table; ++ unsigned int vsel_range_reg; ++ unsigned int vsel_range_mask; ++ unsigned int vsel_reg; ++ unsigned int vsel_mask; ++ unsigned int vsel_step; ++ unsigned int csel_reg; ++ unsigned int csel_mask; ++ unsigned int apply_reg; ++ unsigned int apply_bit; ++ unsigned int enable_reg; ++ unsigned int enable_mask; ++ unsigned int enable_val; ++ unsigned int disable_val; ++ bool enable_is_inverted; ++ unsigned int bypass_reg; ++ unsigned int bypass_mask; ++ unsigned int bypass_val_on; ++ unsigned int bypass_val_off; ++ unsigned int active_discharge_on; ++ unsigned int active_discharge_off; ++ unsigned int active_discharge_mask; ++ unsigned int active_discharge_reg; ++ unsigned int soft_start_reg; ++ unsigned int soft_start_mask; ++ unsigned int soft_start_val_on; ++ unsigned int pull_down_reg; ++ unsigned int pull_down_mask; ++ unsigned int pull_down_val_on; ++ unsigned int ramp_reg; ++ unsigned int ramp_mask; ++ const unsigned int *ramp_delay_table; ++ unsigned int n_ramp_values; ++ unsigned int enable_time; ++ unsigned int off_on_delay; ++ unsigned int poll_enabled_time; ++ unsigned int (*of_map_mode)(unsigned int); ++}; ++ ++struct gpio_desc; ++ ++struct regulator_init_data; ++ ++struct regulator_config { ++ struct device *dev; ++ const struct regulator_init_data *init_data; ++ void *driver_data; ++ struct device_node *of_node; ++ struct regmap *regmap; ++ struct gpio_desc *ena_gpiod; ++}; ++ ++struct regulator_state { ++ int uV; ++ int min_uV; ++ int max_uV; ++ unsigned int mode; ++ int enabled; ++ bool changeable; ++}; ++ ++struct notification_limit { ++ int prot; ++ int err; ++ int warn; ++}; ++ ++struct regulation_constraints { ++ const char *name; ++ int min_uV; ++ int max_uV; ++ int uV_offset; ++ int min_uA; ++ int max_uA; ++ int ilim_uA; ++ int system_load; ++ u32 *max_spread; ++ int max_uV_step; ++ unsigned int valid_modes_mask; ++ unsigned int valid_ops_mask; ++ int input_uV; ++ struct regulator_state state_disk; ++ struct regulator_state state_mem; ++ struct regulator_state state_standby; ++ struct notification_limit over_curr_limits; ++ struct notification_limit over_voltage_limits; ++ struct notification_limit under_voltage_limits; ++ struct notification_limit temp_limits; ++ suspend_state_t initial_state; ++ unsigned int initial_mode; ++ unsigned int ramp_delay; ++ unsigned int settling_time; ++ unsigned int settling_time_up; ++ unsigned int settling_time_down; ++ unsigned int enable_time; ++ unsigned int active_discharge; ++ unsigned int always_on: 1; ++ unsigned int boot_on: 1; ++ unsigned int apply_uV: 1; ++ unsigned int ramp_disable: 1; ++ unsigned int soft_start: 1; ++ unsigned int pull_down: 1; ++ unsigned int over_current_protection: 1; ++ unsigned int over_current_detection: 1; ++ unsigned int over_voltage_detection: 1; ++ unsigned int under_voltage_detection: 1; ++ unsigned int over_temp_detection: 1; ++}; ++ ++struct regulator_consumer_supply; ++ ++struct regulator_init_data { ++ const char *supply_regulator; ++ struct regulation_constraints constraints; ++ int num_consumer_supplies; ++ struct regulator_consumer_supply *consumer_supplies; ++ int (*regulator_init)(void *); ++ void *driver_data; ++}; ++ ++struct regulator_consumer_supply { ++ const char *dev_name; ++ const char *supply; ++}; ++ ++struct regulator_bulk_data { ++ const char *supply; ++ int init_load_uA; ++ struct regulator *consumer; ++ int ret; ++}; ++ ++struct regulator_voltage { ++ int min_uV; ++ int max_uV; ++}; ++ ++struct regulator { ++ struct device *dev; ++ struct list_head list; ++ unsigned int always_on: 1; ++ unsigned int bypass: 1; ++ unsigned int device_link: 1; ++ int uA_load; ++ unsigned int enable_count; ++ unsigned int deferred_disables; ++ struct regulator_voltage voltage[5]; ++ const char *supply_name; ++ struct device_attribute dev_attr; ++ struct regulator_dev *rdev; ++ struct dentry *debugfs; ++}; ++ ++struct regulator_err_state { ++ struct regulator_dev *rdev; ++ long unsigned int notifs; ++ long unsigned int errors; ++ int possible_errs; ++}; ++ ++struct regulator_irq_data { ++ struct regulator_err_state *states; ++ int num_states; ++ void *data; ++ long int opaque; ++}; ++ ++struct regulator_irq_desc { ++ const char *name; ++ int fatal_cnt; ++ int reread_ms; ++ int irq_off_ms; ++ bool skip_off; ++ bool high_prio; ++ void *data; ++ int (*die)(struct regulator_irq_data *); ++ int (*map_event)(int, struct regulator_irq_data *, long unsigned int *); ++ int (*renable)(struct regulator_irq_data *); ++}; ++ ++enum regulator_get_type { ++ NORMAL_GET = 0, ++ EXCLUSIVE_GET = 1, ++ OPTIONAL_GET = 2, ++ MAX_GET_TYPE = 3, ++}; ++ ++struct regulator_bulk_devres { ++ struct regulator_bulk_data *consumers; ++ int num_consumers; ++}; ++ ++struct regulator_supply_alias_match { ++ struct device *dev; ++ const char *id; ++}; ++ ++struct regulator_notifier_match { ++ struct regulator *regulator; ++ struct notifier_block *nb; ++}; ++ ++struct termios { ++ tcflag_t c_iflag; ++ tcflag_t c_oflag; ++ tcflag_t c_cflag; ++ tcflag_t c_lflag; ++ cc_t c_line; ++ cc_t c_cc[19]; ++}; ++ ++struct termios2 { ++ tcflag_t c_iflag; ++ tcflag_t c_oflag; ++ tcflag_t c_cflag; ++ tcflag_t c_lflag; ++ cc_t c_line; ++ cc_t c_cc[19]; ++ speed_t c_ispeed; ++ speed_t c_ospeed; ++}; ++ ++struct termio { ++ short unsigned int c_iflag; ++ short unsigned int c_oflag; ++ short unsigned int c_cflag; ++ short unsigned int c_lflag; ++ unsigned char c_line; ++ unsigned char c_cc[8]; ++}; ++ ++struct tty_audit_buf { ++ struct mutex mutex; ++ dev_t dev; ++ unsigned int icanon: 1; ++ size_t valid; ++ unsigned char *data; ++}; ++ ++struct vc { ++ struct vc_data *d; ++ struct work_struct SAK_work; ++}; ++ ++struct vt_notifier_param { ++ struct vc_data *vc; ++ unsigned int c; ++}; ++ ++struct vcs_poll_data { ++ struct notifier_block notifier; ++ unsigned int cons_num; ++ int event; ++ wait_queue_head_t waitq; ++ struct fasync_struct *fasync; ++}; ++ ++struct input_id { ++ __u16 bustype; ++ __u16 vendor; ++ __u16 product; ++ __u16 version; ++}; ++ ++struct input_absinfo { ++ __s32 value; ++ __s32 minimum; ++ __s32 maximum; ++ __s32 fuzz; ++ __s32 flat; ++ __s32 resolution; ++}; ++ ++struct input_keymap_entry { ++ __u8 flags; ++ __u8 len; ++ __u16 index; ++ __u32 keycode; ++ __u8 scancode[32]; ++}; ++ ++struct ff_replay { ++ __u16 length; ++ __u16 delay; ++}; ++ ++struct ff_trigger { ++ __u16 button; ++ __u16 interval; ++}; ++ ++struct ff_envelope { ++ __u16 attack_length; ++ __u16 attack_level; ++ __u16 fade_length; ++ __u16 fade_level; ++}; ++ ++struct ff_constant_effect { ++ __s16 level; ++ struct ff_envelope envelope; ++}; ++ ++struct ff_ramp_effect { ++ __s16 start_level; ++ __s16 end_level; ++ struct ff_envelope envelope; ++}; ++ ++struct ff_condition_effect { ++ __u16 right_saturation; ++ __u16 left_saturation; ++ __s16 right_coeff; ++ __s16 left_coeff; ++ __u16 deadband; ++ __s16 center; ++}; ++ ++struct ff_periodic_effect { ++ __u16 waveform; ++ __u16 period; ++ __s16 magnitude; ++ __s16 offset; ++ __u16 phase; ++ struct ff_envelope envelope; ++ __u32 custom_len; ++ __s16 *custom_data; ++}; ++ ++struct ff_rumble_effect { ++ __u16 strong_magnitude; ++ __u16 weak_magnitude; ++}; ++ ++struct ff_effect { ++ __u16 type; ++ __s16 id; ++ __u16 direction; ++ struct ff_trigger trigger; ++ struct ff_replay replay; ++ union { ++ struct ff_constant_effect constant; ++ struct ff_ramp_effect ramp; ++ struct ff_periodic_effect periodic; ++ struct ff_condition_effect condition[2]; ++ struct ff_rumble_effect rumble; ++ } u; ++}; ++ ++struct input_device_id { ++ kernel_ulong_t flags; ++ __u16 bustype; ++ __u16 vendor; ++ __u16 product; ++ __u16 version; ++ kernel_ulong_t evbit[1]; ++ kernel_ulong_t keybit[12]; ++ kernel_ulong_t relbit[1]; ++ kernel_ulong_t absbit[1]; ++ kernel_ulong_t mscbit[1]; ++ kernel_ulong_t ledbit[1]; ++ kernel_ulong_t sndbit[1]; ++ kernel_ulong_t ffbit[2]; ++ kernel_ulong_t swbit[1]; ++ kernel_ulong_t propbit[1]; ++ kernel_ulong_t driver_info; ++}; ++ ++struct input_value { ++ __u16 type; ++ __u16 code; ++ __s32 value; ++}; ++ ++enum input_clock_type { ++ INPUT_CLK_REAL = 0, ++ INPUT_CLK_MONO = 1, ++ INPUT_CLK_BOOT = 2, ++ INPUT_CLK_MAX = 3, ++}; ++ ++struct ff_device; ++ ++struct input_dev_poller; ++ ++struct input_mt; ++ ++struct input_handle; ++ ++struct input_dev { ++ const char *name; ++ const char *phys; ++ const char *uniq; ++ struct input_id id; ++ long unsigned int propbit[1]; ++ long unsigned int evbit[1]; ++ long unsigned int keybit[12]; ++ long unsigned int relbit[1]; ++ long unsigned int absbit[1]; ++ long unsigned int mscbit[1]; ++ long unsigned int ledbit[1]; ++ long unsigned int sndbit[1]; ++ long unsigned int ffbit[2]; ++ long unsigned int swbit[1]; ++ unsigned int hint_events_per_packet; ++ unsigned int keycodemax; ++ unsigned int keycodesize; ++ void *keycode; ++ int (*setkeycode)(struct input_dev *, const struct input_keymap_entry *, unsigned int *); ++ int (*getkeycode)(struct input_dev *, struct input_keymap_entry *); ++ struct ff_device *ff; ++ struct input_dev_poller *poller; ++ unsigned int repeat_key; ++ struct timer_list timer; ++ int rep[2]; ++ struct input_mt *mt; ++ struct input_absinfo *absinfo; ++ long unsigned int key[12]; ++ long unsigned int led[1]; ++ long unsigned int snd[1]; ++ long unsigned int sw[1]; ++ int (*open)(struct input_dev *); ++ void (*close)(struct input_dev *); ++ int (*flush)(struct input_dev *, struct file *); ++ int (*event)(struct input_dev *, unsigned int, unsigned int, int); ++ struct input_handle *grab; ++ spinlock_t event_lock; ++ struct mutex mutex; ++ unsigned int users; ++ bool going_away; ++ struct device dev; ++ struct list_head h_list; ++ struct list_head node; ++ unsigned int num_vals; ++ unsigned int max_vals; ++ struct input_value *vals; ++ bool devres_managed; ++ ktime_t timestamp[3]; ++ bool inhibited; ++}; ++ ++struct ff_device { ++ int (*upload)(struct input_dev *, struct ff_effect *, struct ff_effect *); ++ int (*erase)(struct input_dev *, int); ++ int (*playback)(struct input_dev *, int, int); ++ void (*set_gain)(struct input_dev *, u16); ++ void (*set_autocenter)(struct input_dev *, u16); ++ void (*destroy)(struct ff_device *); ++ void *private; ++ long unsigned int ffbit[2]; ++ struct mutex mutex; ++ int max_effects; ++ struct ff_effect *effects; ++ struct file *effect_owners[0]; ++}; ++ ++struct input_handler; ++ ++struct input_handle { ++ void *private; ++ int open; ++ const char *name; ++ struct input_dev *dev; ++ struct input_handler *handler; ++ struct list_head d_node; ++ struct list_head h_node; ++}; ++ ++struct input_handler { ++ void *private; ++ void (*event)(struct input_handle *, unsigned int, unsigned int, int); ++ void (*events)(struct input_handle *, const struct input_value *, unsigned int); ++ bool (*filter)(struct input_handle *, unsigned int, unsigned int, int); ++ bool (*match)(struct input_handler *, struct input_dev *); ++ int (*connect)(struct input_handler *, struct input_dev *, const struct input_device_id *); ++ void (*disconnect)(struct input_handle *); ++ void (*start)(struct input_handle *); ++ bool legacy_minors; ++ int minor; ++ const char *name; ++ const struct input_device_id *id_table; ++ struct list_head h_list; ++ struct list_head node; ++}; ++ ++struct kbentry { ++ unsigned char kb_table; ++ unsigned char kb_index; ++ short unsigned int kb_value; ++}; ++ ++struct kbsentry { ++ unsigned char kb_func; ++ unsigned char kb_string[512]; ++}; ++ ++struct kbdiacr { ++ unsigned char diacr; ++ unsigned char base; ++ unsigned char result; ++}; ++ ++struct kbdiacrs { ++ unsigned int kb_cnt; ++ struct kbdiacr kbdiacr[256]; ++}; ++ ++struct kbdiacruc { ++ unsigned int diacr; ++ unsigned int base; ++ unsigned int result; ++}; ++ ++struct kbdiacrsuc { ++ unsigned int kb_cnt; ++ struct kbdiacruc kbdiacruc[256]; ++}; ++ ++struct kbkeycode { ++ unsigned int scancode; ++ unsigned int keycode; ++}; ++ ++struct kbd_repeat { ++ int delay; ++ int period; ++}; ++ ++struct tasklet_struct { ++ struct tasklet_struct *next; ++ long unsigned int state; ++ atomic_t count; ++ bool use_callback; ++ union { ++ void (*func)(long unsigned int); ++ void (*callback)(struct tasklet_struct *); ++ }; ++ long unsigned int data; ++}; ++ ++enum { ++ TASKLET_STATE_SCHED = 0, ++ TASKLET_STATE_RUN = 1, ++}; ++ ++struct keyboard_notifier_param { ++ struct vc_data *vc; ++ int down; ++ int shift; ++ int ledstate; ++ unsigned int value; ++}; ++ ++struct kbd_struct { ++ unsigned char lockstate; ++ unsigned char slockstate; ++ unsigned char ledmode: 1; ++ unsigned char ledflagstate: 4; ++ char: 3; ++ unsigned char default_ledflagstate: 4; ++ unsigned char kbdmode: 3; ++ int: 1; ++ unsigned char modeflags: 5; ++}; ++ ++struct vt_spawn_console { ++ spinlock_t lock; ++ struct pid *pid; ++ int sig; ++}; ++ ++typedef void k_handler_fn(struct vc_data *, unsigned char, char); ++ ++typedef void fn_handler_fn(struct vc_data *); ++ ++struct getset_keycode_data { ++ struct input_keymap_entry ke; ++ int error; ++}; ++ ++struct serial_icounter_struct { ++ int cts; ++ int dsr; ++ int rng; ++ int dcd; ++ int rx; ++ int tx; ++ int frame; ++ int overrun; ++ int parity; ++ int brk; ++ int buf_overrun; ++ int reserved[9]; ++}; ++ ++struct serial_struct { ++ int type; ++ int line; ++ unsigned int port; ++ int irq; ++ int flags; ++ int xmit_fifo_size; ++ int custom_divisor; ++ int baud_base; ++ short unsigned int close_delay; ++ char io_type; ++ char reserved_char[1]; ++ int hub6; ++ short unsigned int closing_wait; ++ short unsigned int closing_wait2; ++ unsigned char *iomem_base; ++ short unsigned int iomem_reg_shift; ++ unsigned int port_high; ++ long unsigned int iomap_base; ++}; ++ ++struct hv_ops; ++ ++struct hvc_struct { ++ struct tty_port port; ++ spinlock_t lock; ++ int index; ++ int do_wakeup; ++ char *outbuf; ++ int outbuf_size; ++ int n_outbuf; ++ uint32_t vtermno; ++ const struct hv_ops *ops; ++ int irq_requested; ++ int data; ++ struct winsize ws; ++ struct work_struct tty_resize; ++ struct list_head next; ++ long unsigned int flags; ++}; ++ ++struct hv_ops { ++ int (*get_chars)(uint32_t, char *, int); ++ int (*put_chars)(uint32_t, const char *, int); ++ int (*flush)(uint32_t, bool); ++ int (*notifier_add)(struct hvc_struct *, int); ++ void (*notifier_del)(struct hvc_struct *, int); ++ void (*notifier_hangup)(struct hvc_struct *, int); ++ int (*tiocmget)(struct hvc_struct *); ++ int (*tiocmset)(struct hvc_struct *, unsigned int, unsigned int); ++ void (*dtr_rts)(struct hvc_struct *, bool); ++}; ++ ++struct splice_desc { ++ size_t total_len; ++ unsigned int len; ++ unsigned int flags; ++ union { ++ void *userptr; ++ struct file *file; ++ void *data; ++ } u; ++ loff_t pos; ++ loff_t *opos; ++ size_t num_spliced; ++ bool need_wakeup; ++}; ++ ++typedef int splice_actor(struct pipe_inode_info *, struct pipe_buffer *, struct splice_desc *); ++ ++struct memdev { ++ const char *name; ++ umode_t mode; ++ const struct file_operations *fops; ++ fmode_t fmode; ++}; ++ ++struct virtio_driver { ++ struct device_driver driver; ++ const struct virtio_device_id *id_table; ++ const unsigned int *feature_table; ++ unsigned int feature_table_size; ++ const unsigned int *feature_table_legacy; ++ unsigned int feature_table_size_legacy; ++ int (*validate)(struct virtio_device *); ++ int (*probe)(struct virtio_device *); ++ void (*scan)(struct virtio_device *); ++ void (*remove)(struct virtio_device *); ++ void (*config_changed)(struct virtio_device *); ++}; ++ ++struct virtrng_info { ++ struct hwrng hwrng; ++ struct virtqueue *vq; ++ char name[25]; ++ int index; ++ bool hwrng_register_done; ++ bool hwrng_removed; ++ struct completion have_data; ++ unsigned int data_avail; ++ unsigned int data_idx; ++ u8 data[256]; ++}; ++ ++struct file_priv { ++ struct tpm_chip *chip; ++ struct tpm_space *space; ++ struct mutex buffer_mutex; ++ struct timer_list user_read_timer; ++ struct work_struct timeout_work; ++ struct work_struct async_work; ++ wait_queue_head_t async_wait; ++ ssize_t response_length; ++ bool response_read; ++ bool command_enqueued; ++ u8 data_buffer[4096]; ++}; ++ ++enum tpm2_const { ++ TPM2_PLATFORM_PCR = 24, ++ TPM2_PCR_SELECT_MIN = 3, ++}; ++ ++enum tpm2_timeouts { ++ TPM2_TIMEOUT_A = 750, ++ TPM2_TIMEOUT_B = 2000, ++ TPM2_TIMEOUT_C = 200, ++ TPM2_TIMEOUT_D = 30, ++ TPM2_DURATION_SHORT = 20, ++ TPM2_DURATION_MEDIUM = 750, ++ TPM2_DURATION_LONG = 2000, ++ TPM2_DURATION_LONG_LONG = 300000, ++ TPM2_DURATION_DEFAULT = 120000, ++}; ++ ++enum tpm2_capabilities { ++ TPM2_CAP_HANDLES = 1, ++ TPM2_CAP_COMMANDS = 2, ++ TPM2_CAP_PCRS = 5, ++ TPM2_CAP_TPM_PROPERTIES = 6, ++}; ++ ++enum tpm2_properties { ++ TPM_PT_TOTAL_COMMANDS = 297, ++}; ++ ++enum tpm2_startup_types { ++ TPM2_SU_CLEAR = 0, ++ TPM2_SU_STATE = 1, ++}; ++ ++enum tpm2_cc_attrs { ++ TPM2_CC_ATTR_CHANDLES = 25, ++ TPM2_CC_ATTR_RHANDLE = 28, ++ TPM2_CC_ATTR_VENDOR = 29, ++}; ++ ++enum tpm_chip_flags { ++ TPM_CHIP_FLAG_BOOTSTRAPPED = 1, ++ TPM_CHIP_FLAG_TPM2 = 2, ++ TPM_CHIP_FLAG_IRQ = 4, ++ TPM_CHIP_FLAG_VIRTUAL = 8, ++ TPM_CHIP_FLAG_HAVE_TIMEOUTS = 16, ++ TPM_CHIP_FLAG_ALWAYS_POWERED = 32, ++ TPM_CHIP_FLAG_FIRMWARE_POWER_MANAGED = 64, ++ TPM_CHIP_FLAG_FIRMWARE_UPGRADE = 128, ++ TPM_CHIP_FLAG_SUSPENDED = 256, ++ TPM_CHIP_FLAG_HWRNG_DISABLED = 512, ++}; ++ ++struct tpm2_pcr_read_out { ++ __be32 update_cnt; ++ __be32 pcr_selects_cnt; ++ __be16 hash_alg; ++ u8 pcr_select_size; ++ u8 pcr_select[3]; ++ __be32 digests_cnt; ++ __be16 digest_size; ++ u8 digest[0]; ++} __attribute__((packed)); ++ ++struct tpm2_null_auth_area { ++ __be32 handle; ++ __be16 nonce_size; ++ u8 attributes; ++ __be16 auth_size; ++} __attribute__((packed)); ++ ++struct tpm2_get_random_out { ++ __be16 size; ++ u8 buffer[128]; ++}; ++ ++struct tpm2_get_cap_out { ++ u8 more_data; ++ __be32 subcap_id; ++ __be32 property_cnt; ++ __be32 property_id; ++ __be32 value; ++} __attribute__((packed)); ++ ++struct tpm2_pcr_selection { ++ __be16 hash_alg; ++ u8 size_of_select; ++ u8 pcr_select[3]; ++}; ++ ++struct tcpa_event { ++ u32 pcr_index; ++ u32 event_type; ++ u8 pcr_value[20]; ++ u32 event_size; ++ u8 event_data[0]; ++}; ++ ++enum tcpa_event_types { ++ PREBOOT = 0, ++ POST_CODE = 1, ++ UNUSED___2 = 2, ++ NO_ACTION = 3, ++ SEPARATOR = 4, ++ ACTION = 5, ++ EVENT_TAG = 6, ++ SCRTM_CONTENTS = 7, ++ SCRTM_VERSION = 8, ++ CPU_MICROCODE = 9, ++ PLATFORM_CONFIG_FLAGS = 10, ++ TABLE_OF_DEVICES = 11, ++ COMPACT_HASH = 12, ++ IPL = 13, ++ IPL_PARTITION_DATA = 14, ++ NONHOST_CODE = 15, ++ NONHOST_CONFIG = 16, ++ NONHOST_INFO = 17, ++}; ++ ++struct tcpa_pc_event { ++ u32 event_id; ++ u32 event_size; ++ u8 event_data[0]; ++}; ++ ++enum tcpa_pc_event_ids { ++ SMBIOS = 1, ++ BIS_CERT = 2, ++ POST_BIOS_ROM = 3, ++ ESCD = 4, ++ CMOS = 5, ++ NVRAM = 6, ++ OPTION_ROM_EXEC = 7, ++ OPTION_ROM_CONFIG = 8, ++ OPTION_ROM_MICROCODE = 10, ++ S_CRTM_VERSION = 11, ++ S_CRTM_CONTENTS = 12, ++ POST_CONTENTS = 13, ++ HOST_TABLE_OF_DEVICES = 14, ++}; ++ ++struct trace_event_raw_iommu_group_event { ++ struct trace_entry ent; ++ int gid; ++ u32 __data_loc_device; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_iommu_device_event { ++ struct trace_entry ent; ++ u32 __data_loc_device; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_map { ++ struct trace_entry ent; ++ u64 iova; ++ u64 paddr; ++ size_t size; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_unmap { ++ struct trace_entry ent; ++ u64 iova; ++ size_t size; ++ size_t unmapped_size; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_iommu_error { ++ struct trace_entry ent; ++ u32 __data_loc_device; ++ u32 __data_loc_driver; ++ u64 iova; ++ int flags; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_iommu_group_event { ++ u32 device; ++}; ++ ++struct trace_event_data_offsets_iommu_device_event { ++ u32 device; ++}; ++ ++struct trace_event_data_offsets_map {}; ++ ++struct trace_event_data_offsets_unmap {}; ++ ++struct trace_event_data_offsets_iommu_error { ++ u32 device; ++ u32 driver; ++}; ++ ++typedef void (*btf_trace_add_device_to_group)(void *, int, struct device *); ++ ++typedef void (*btf_trace_remove_device_from_group)(void *, int, struct device *); ++ ++typedef void (*btf_trace_attach_device_to_domain)(void *, struct device *); ++ ++typedef void (*btf_trace_map)(void *, long unsigned int, phys_addr_t, size_t); ++ ++typedef void (*btf_trace_unmap)(void *, long unsigned int, size_t, size_t); ++ ++typedef void (*btf_trace_io_page_fault)(void *, struct device *, long unsigned int, int); ++ ++struct drm_mode_fb_cmd2 { ++ __u32 fb_id; ++ __u32 width; ++ __u32 height; ++ __u32 pixel_format; ++ __u32 flags; ++ __u32 handles[4]; ++ __u32 pitches[4]; ++ __u32 offsets[4]; ++ __u64 modifier[4]; ++}; ++ ++struct drm_mode_create_dumb { ++ __u32 height; ++ __u32 width; ++ __u32 bpp; ++ __u32 flags; ++ __u32 handle; ++ __u32 pitch; ++ __u64 size; ++}; ++ ++struct drm_device; ++ ++struct drm_master { ++ struct kref refcount; ++ struct drm_device *dev; ++ char *unique; ++ int unique_len; ++ struct idr magic_map; ++ void *driver_priv; ++ struct drm_master *lessor; ++ int lessee_id; ++ struct list_head lessee_list; ++ struct list_head lessees; ++ struct idr leases; ++ struct idr lessee_idr; ++}; ++ ++struct drm_modeset_lock { ++ struct ww_mutex mutex; ++ struct list_head head; ++}; ++ ++struct drm_modeset_acquire_ctx; ++ ++struct drm_mode_config_funcs; ++ ++struct drm_property; ++ ++struct drm_atomic_state; ++ ++struct drm_mode_config_helper_funcs; ++ ++struct drm_mode_config { ++ struct mutex mutex; ++ struct drm_modeset_lock connection_mutex; ++ struct drm_modeset_acquire_ctx *acquire_ctx; ++ struct mutex idr_mutex; ++ struct idr object_idr; ++ struct idr tile_idr; ++ struct mutex fb_lock; ++ int num_fb; ++ struct list_head fb_list; ++ spinlock_t connector_list_lock; ++ int num_connector; ++ struct ida connector_ida; ++ struct list_head connector_list; ++ struct llist_head connector_free_list; ++ struct work_struct connector_free_work; ++ int num_encoder; ++ struct list_head encoder_list; ++ int num_total_plane; ++ struct list_head plane_list; ++ int num_crtc; ++ struct list_head crtc_list; ++ struct list_head property_list; ++ struct list_head privobj_list; ++ int min_width; ++ int min_height; ++ int max_width; ++ int max_height; ++ const struct drm_mode_config_funcs *funcs; ++ bool poll_enabled; ++ bool poll_running; ++ bool delayed_event; ++ struct delayed_work output_poll_work; ++ struct mutex blob_lock; ++ struct list_head property_blob_list; ++ struct drm_property *edid_property; ++ struct drm_property *dpms_property; ++ struct drm_property *path_property; ++ struct drm_property *tile_property; ++ struct drm_property *link_status_property; ++ struct drm_property *plane_type_property; ++ struct drm_property *prop_src_x; ++ struct drm_property *prop_src_y; ++ struct drm_property *prop_src_w; ++ struct drm_property *prop_src_h; ++ struct drm_property *prop_crtc_x; ++ struct drm_property *prop_crtc_y; ++ struct drm_property *prop_crtc_w; ++ struct drm_property *prop_crtc_h; ++ struct drm_property *prop_fb_id; ++ struct drm_property *prop_in_fence_fd; ++ struct drm_property *prop_out_fence_ptr; ++ struct drm_property *prop_crtc_id; ++ struct drm_property *prop_fb_damage_clips; ++ struct drm_property *prop_active; ++ struct drm_property *prop_mode_id; ++ struct drm_property *prop_vrr_enabled; ++ struct drm_property *dvi_i_subconnector_property; ++ struct drm_property *dvi_i_select_subconnector_property; ++ struct drm_property *dp_subconnector_property; ++ struct drm_property *tv_subconnector_property; ++ struct drm_property *tv_select_subconnector_property; ++ struct drm_property *legacy_tv_mode_property; ++ struct drm_property *tv_mode_property; ++ struct drm_property *tv_left_margin_property; ++ struct drm_property *tv_right_margin_property; ++ struct drm_property *tv_top_margin_property; ++ struct drm_property *tv_bottom_margin_property; ++ struct drm_property *tv_brightness_property; ++ struct drm_property *tv_contrast_property; ++ struct drm_property *tv_flicker_reduction_property; ++ struct drm_property *tv_overscan_property; ++ struct drm_property *tv_saturation_property; ++ struct drm_property *tv_hue_property; ++ struct drm_property *scaling_mode_property; ++ struct drm_property *aspect_ratio_property; ++ struct drm_property *content_type_property; ++ struct drm_property *degamma_lut_property; ++ struct drm_property *degamma_lut_size_property; ++ struct drm_property *ctm_property; ++ struct drm_property *gamma_lut_property; ++ struct drm_property *gamma_lut_size_property; ++ struct drm_property *suggested_x_property; ++ struct drm_property *suggested_y_property; ++ struct drm_property *non_desktop_property; ++ struct drm_property *panel_orientation_property; ++ struct drm_property *writeback_fb_id_property; ++ struct drm_property *writeback_pixel_formats_property; ++ struct drm_property *writeback_out_fence_ptr_property; ++ struct drm_property *hdr_output_metadata_property; ++ struct drm_property *content_protection_property; ++ struct drm_property *hdcp_content_type_property; ++ uint32_t preferred_depth; ++ uint32_t prefer_shadow; ++ bool quirk_addfb_prefer_xbgr_30bpp; ++ bool quirk_addfb_prefer_host_byte_order; ++ bool async_page_flip; ++ bool fb_modifiers_not_supported; ++ bool normalize_zpos; ++ struct drm_property *modifiers_property; ++ uint32_t cursor_width; ++ uint32_t cursor_height; ++ struct drm_atomic_state *suspend_state; ++ const struct drm_mode_config_helper_funcs *helper_private; ++}; ++ ++struct drm_vram_mm; ++ ++enum switch_power_state { ++ DRM_SWITCH_POWER_ON = 0, ++ DRM_SWITCH_POWER_OFF = 1, ++ DRM_SWITCH_POWER_CHANGING = 2, ++ DRM_SWITCH_POWER_DYNAMIC_OFF = 3, ++}; ++ ++struct drm_driver; ++ ++struct drm_minor; ++ ++struct drm_vblank_crtc; ++ ++struct drm_vma_offset_manager; ++ ++struct drm_fb_helper; ++ ++struct drm_device { ++ int if_version; ++ struct kref ref; ++ struct device *dev; ++ struct { ++ struct list_head resources; ++ void *final_kfree; ++ spinlock_t lock; ++ } managed; ++ const struct drm_driver *driver; ++ void *dev_private; ++ struct drm_minor *primary; ++ struct drm_minor *render; ++ struct drm_minor *accel; ++ bool registered; ++ struct drm_master *master; ++ u32 driver_features; ++ bool unplugged; ++ struct inode *anon_inode; ++ char *unique; ++ struct mutex struct_mutex; ++ struct mutex master_mutex; ++ atomic_t open_count; ++ struct mutex filelist_mutex; ++ struct list_head filelist; ++ struct list_head filelist_internal; ++ struct mutex clientlist_mutex; ++ struct list_head clientlist; ++ bool vblank_disable_immediate; ++ struct drm_vblank_crtc *vblank; ++ spinlock_t vblank_time_lock; ++ spinlock_t vbl_lock; ++ u32 max_vblank_count; ++ struct list_head vblank_event_list; ++ spinlock_t event_lock; ++ unsigned int num_crtcs; ++ struct drm_mode_config mode_config; ++ struct mutex object_name_lock; ++ struct idr object_name_idr; ++ struct drm_vma_offset_manager *vma_offset_manager; ++ struct drm_vram_mm *vram_mm; ++ enum switch_power_state switch_power_state; ++ struct drm_fb_helper *fb_helper; ++ struct mutex debugfs_mutex; ++ struct list_head debugfs_list; ++}; ++ ++struct drm_modeset_acquire_ctx { ++ struct ww_acquire_ctx ww_ctx; ++ struct drm_modeset_lock *contended; ++ depot_stack_handle_t stack_depot; ++ struct list_head locked; ++ bool trylock_only; ++ bool interruptible; ++}; ++ ++enum drm_mode_status { ++ MODE_OK = 0, ++ MODE_HSYNC = 1, ++ MODE_VSYNC = 2, ++ MODE_H_ILLEGAL = 3, ++ MODE_V_ILLEGAL = 4, ++ MODE_BAD_WIDTH = 5, ++ MODE_NOMODE = 6, ++ MODE_NO_INTERLACE = 7, ++ MODE_NO_DBLESCAN = 8, ++ MODE_NO_VSCAN = 9, ++ MODE_MEM = 10, ++ MODE_VIRTUAL_X = 11, ++ MODE_VIRTUAL_Y = 12, ++ MODE_MEM_VIRT = 13, ++ MODE_NOCLOCK = 14, ++ MODE_CLOCK_HIGH = 15, ++ MODE_CLOCK_LOW = 16, ++ MODE_CLOCK_RANGE = 17, ++ MODE_BAD_HVALUE = 18, ++ MODE_BAD_VVALUE = 19, ++ MODE_BAD_VSCAN = 20, ++ MODE_HSYNC_NARROW = 21, ++ MODE_HSYNC_WIDE = 22, ++ MODE_HBLANK_NARROW = 23, ++ MODE_HBLANK_WIDE = 24, ++ MODE_VSYNC_NARROW = 25, ++ MODE_VSYNC_WIDE = 26, ++ MODE_VBLANK_NARROW = 27, ++ MODE_VBLANK_WIDE = 28, ++ MODE_PANEL = 29, ++ MODE_INTERLACE_WIDTH = 30, ++ MODE_ONE_WIDTH = 31, ++ MODE_ONE_HEIGHT = 32, ++ MODE_ONE_SIZE = 33, ++ MODE_NO_REDUCED = 34, ++ MODE_NO_STEREO = 35, ++ MODE_NO_420 = 36, ++ MODE_STALE = -3, ++ MODE_BAD = -2, ++ MODE_ERROR = -1, ++}; ++ ++struct drm_framebuffer; ++ ++struct drm_file; ++ ++struct drm_format_info; ++ ++struct drm_display_mode; ++ ++struct drm_mode_config_funcs { ++ struct drm_framebuffer * (*fb_create)(struct drm_device *, struct drm_file *, const struct drm_mode_fb_cmd2 *); ++ const struct drm_format_info * (*get_format_info)(const struct drm_mode_fb_cmd2 *); ++ void (*output_poll_changed)(struct drm_device *); ++ enum drm_mode_status (*mode_valid)(struct drm_device *, const struct drm_display_mode *); ++ int (*atomic_check)(struct drm_device *, struct drm_atomic_state *); ++ int (*atomic_commit)(struct drm_device *, struct drm_atomic_state *, bool); ++ struct drm_atomic_state * (*atomic_state_alloc)(struct drm_device *); ++ void (*atomic_state_clear)(struct drm_atomic_state *); ++ void (*atomic_state_free)(struct drm_atomic_state *); ++}; ++ ++struct drm_format_info { ++ u32 format; ++ u8 depth; ++ u8 num_planes; ++ union { ++ u8 cpp[4]; ++ u8 char_per_block[4]; ++ }; ++ u8 block_w[4]; ++ u8 block_h[4]; ++ u8 hsub; ++ u8 vsub; ++ bool has_alpha; ++ bool is_yuv; ++ bool is_color_indexed; ++}; ++ ++struct drm_display_mode { ++ int clock; ++ u16 hdisplay; ++ u16 hsync_start; ++ u16 hsync_end; ++ u16 htotal; ++ u16 hskew; ++ u16 vdisplay; ++ u16 vsync_start; ++ u16 vsync_end; ++ u16 vtotal; ++ u16 vscan; ++ u32 flags; ++ int crtc_clock; ++ u16 crtc_hdisplay; ++ u16 crtc_hblank_start; ++ u16 crtc_hblank_end; ++ u16 crtc_hsync_start; ++ u16 crtc_hsync_end; ++ u16 crtc_htotal; ++ u16 crtc_hskew; ++ u16 crtc_vdisplay; ++ u16 crtc_vblank_start; ++ u16 crtc_vblank_end; ++ u16 crtc_vsync_start; ++ u16 crtc_vsync_end; ++ u16 crtc_vtotal; ++ u16 width_mm; ++ u16 height_mm; ++ u8 type; ++ bool expose_to_userspace; ++ struct list_head head; ++ char name[32]; ++ enum drm_mode_status status; ++ enum hdmi_picture_aspect picture_aspect_ratio; ++}; ++ ++struct drm_mode_config_helper_funcs { ++ void (*atomic_commit_tail)(struct drm_atomic_state *); ++ int (*atomic_commit_setup)(struct drm_atomic_state *); ++}; ++ ++struct drm_gem_object; ++ ++struct dma_buf; ++ ++struct dma_buf_attachment; ++ ++struct drm_ioctl_desc; ++ ++struct drm_driver { ++ int (*load)(struct drm_device *, long unsigned int); ++ int (*open)(struct drm_device *, struct drm_file *); ++ void (*postclose)(struct drm_device *, struct drm_file *); ++ void (*lastclose)(struct drm_device *); ++ void (*unload)(struct drm_device *); ++ void (*release)(struct drm_device *); ++ void (*master_set)(struct drm_device *, struct drm_file *, bool); ++ void (*master_drop)(struct drm_device *, struct drm_file *); ++ void (*debugfs_init)(struct drm_minor *); ++ struct drm_gem_object * (*gem_create_object)(struct drm_device *, size_t); ++ int (*prime_handle_to_fd)(struct drm_device *, struct drm_file *, uint32_t, uint32_t, int *); ++ int (*prime_fd_to_handle)(struct drm_device *, struct drm_file *, int, uint32_t *); ++ struct drm_gem_object * (*gem_prime_import)(struct drm_device *, struct dma_buf *); ++ struct drm_gem_object * (*gem_prime_import_sg_table)(struct drm_device *, struct dma_buf_attachment *, struct sg_table *); ++ int (*gem_prime_mmap)(struct drm_gem_object *, struct vm_area_struct *); ++ int (*dumb_create)(struct drm_file *, struct drm_device *, struct drm_mode_create_dumb *); ++ int (*dumb_map_offset)(struct drm_file *, struct drm_device *, uint32_t, uint64_t *); ++ int major; ++ int minor; ++ int patchlevel; ++ char *name; ++ char *desc; ++ char *date; ++ u32 driver_features; ++ const struct drm_ioctl_desc *ioctls; ++ int num_ioctls; ++ const struct file_operations *fops; ++}; ++ ++enum drm_ioctl_flags { ++ DRM_AUTH = 1, ++ DRM_MASTER = 2, ++ DRM_ROOT_ONLY = 4, ++ DRM_UNLOCKED = 16, ++ DRM_RENDER_ALLOW = 32, ++}; ++ ++typedef int drm_ioctl_t(struct drm_device *, void *, struct drm_file *); ++ ++struct drm_ioctl_desc { ++ unsigned int cmd; ++ enum drm_ioctl_flags flags; ++ drm_ioctl_t *func; ++ const char *name; ++}; ++ ++typedef unsigned int drm_magic_t; ++ ++struct drm_prime_file_private { ++ struct mutex lock; ++ struct rb_root dmabufs; ++ struct rb_root handles; ++}; ++ ++struct drm_file { ++ bool authenticated; ++ bool stereo_allowed; ++ bool universal_planes; ++ bool atomic; ++ bool aspect_ratio_allowed; ++ bool writeback_connectors; ++ bool was_master; ++ bool is_master; ++ struct drm_master *master; ++ spinlock_t master_lookup_lock; ++ struct pid *pid; ++ drm_magic_t magic; ++ struct list_head lhead; ++ struct drm_minor *minor; ++ struct idr object_idr; ++ spinlock_t table_lock; ++ struct idr syncobj_idr; ++ spinlock_t syncobj_table_lock; ++ struct file *filp; ++ void *driver_priv; ++ struct list_head fbs; ++ struct mutex fbs_lock; ++ struct list_head blobs; ++ wait_queue_head_t event_wait; ++ struct list_head pending_event_list; ++ struct list_head event_list; ++ int event_space; ++ struct mutex event_read_lock; ++ struct drm_prime_file_private prime; ++}; ++ ++struct drm_auth { ++ drm_magic_t magic; ++}; ++ ++struct drm_minor { ++ int index; ++ int type; ++ struct device *kdev; ++ struct drm_device *dev; ++ struct dentry *debugfs_root; ++ struct list_head debugfs_list; ++ struct mutex debugfs_lock; ++}; ++ ++struct drm_vblank_crtc { ++ struct drm_device *dev; ++ wait_queue_head_t queue; ++ struct timer_list disable_timer; ++ seqlock_t seqlock; ++ atomic64_t count; ++ ktime_t time; ++ atomic_t refcount; ++ u32 last; ++ u32 max_vblank_count; ++ unsigned int inmodeset; ++ unsigned int pipe; ++ int framedur_ns; ++ int linedur_ns; ++ struct drm_display_mode hwmode; ++ bool enabled; ++ struct kthread_worker *worker; ++ struct list_head pending_work; ++ wait_queue_head_t work_wait_queue; ++}; ++ ++enum drm_driver_feature { ++ DRIVER_GEM = 1, ++ DRIVER_MODESET = 2, ++ DRIVER_RENDER = 8, ++ DRIVER_ATOMIC = 16, ++ DRIVER_SYNCOBJ = 32, ++ DRIVER_SYNCOBJ_TIMELINE = 64, ++ DRIVER_COMPUTE_ACCEL = 128, ++ DRIVER_USE_AGP = 33554432, ++ DRIVER_LEGACY = 67108864, ++ DRIVER_PCI_DMA = 134217728, ++ DRIVER_SG = 268435456, ++ DRIVER_HAVE_DMA = 536870912, ++ DRIVER_HAVE_IRQ = 1073741824, ++}; ++ ++enum drm_debug_category { ++ DRM_UT_CORE = 0, ++ DRM_UT_DRIVER = 1, ++ DRM_UT_KMS = 2, ++ DRM_UT_PRIME = 3, ++ DRM_UT_ATOMIC = 4, ++ DRM_UT_VBL = 5, ++ DRM_UT_STATE = 6, ++ DRM_UT_LEASE = 7, ++ DRM_UT_DP = 8, ++ DRM_UT_DRMRES = 9, ++}; ++ ++struct drm_object_properties; ++ ++struct drm_mode_object { ++ uint32_t id; ++ uint32_t type; ++ struct drm_object_properties *properties; ++ struct kref refcount; ++ void (*free_cb)(struct kref *); ++}; ++ ++struct drm_object_properties { ++ int count; ++ struct drm_property *properties[24]; ++ uint64_t values[24]; ++}; ++ ++struct drm_property { ++ struct list_head head; ++ struct drm_mode_object base; ++ uint32_t flags; ++ char name[32]; ++ uint32_t num_values; ++ uint64_t *values; ++ struct drm_device *dev; ++ struct list_head enum_list; ++}; ++ ++struct hdr_static_metadata { ++ __u8 eotf; ++ __u8 metadata_type; ++ __u16 max_cll; ++ __u16 max_fall; ++ __u16 min_cll; ++}; ++ ++struct hdr_sink_metadata { ++ __u32 metadata_type; ++ union { ++ struct hdr_static_metadata hdmi_type1; ++ }; ++}; ++ ++enum drm_mode_subconnector { ++ DRM_MODE_SUBCONNECTOR_Automatic = 0, ++ DRM_MODE_SUBCONNECTOR_Unknown = 0, ++ DRM_MODE_SUBCONNECTOR_VGA = 1, ++ DRM_MODE_SUBCONNECTOR_DVID = 3, ++ DRM_MODE_SUBCONNECTOR_DVIA = 4, ++ DRM_MODE_SUBCONNECTOR_Composite = 5, ++ DRM_MODE_SUBCONNECTOR_SVIDEO = 6, ++ DRM_MODE_SUBCONNECTOR_Component = 8, ++ DRM_MODE_SUBCONNECTOR_SCART = 9, ++ DRM_MODE_SUBCONNECTOR_DisplayPort = 10, ++ DRM_MODE_SUBCONNECTOR_HDMIA = 11, ++ DRM_MODE_SUBCONNECTOR_Native = 15, ++ DRM_MODE_SUBCONNECTOR_Wireless = 18, ++}; ++ ++enum drm_connector_force { ++ DRM_FORCE_UNSPECIFIED = 0, ++ DRM_FORCE_OFF = 1, ++ DRM_FORCE_ON = 2, ++ DRM_FORCE_ON_DIGITAL = 3, ++}; ++ ++enum drm_connector_status { ++ connector_status_connected = 1, ++ connector_status_disconnected = 2, ++ connector_status_unknown = 3, ++}; ++ ++enum drm_connector_registration_state { ++ DRM_CONNECTOR_INITIALIZING = 0, ++ DRM_CONNECTOR_REGISTERED = 1, ++ DRM_CONNECTOR_UNREGISTERED = 2, ++}; ++ ++enum subpixel_order { ++ SubPixelUnknown = 0, ++ SubPixelHorizontalRGB = 1, ++ SubPixelHorizontalBGR = 2, ++ SubPixelVerticalRGB = 3, ++ SubPixelVerticalBGR = 4, ++ SubPixelNone = 5, ++}; ++ ++enum drm_connector_tv_mode { ++ DRM_MODE_TV_MODE_NTSC = 0, ++ DRM_MODE_TV_MODE_NTSC_443 = 1, ++ DRM_MODE_TV_MODE_NTSC_J = 2, ++ DRM_MODE_TV_MODE_PAL = 3, ++ DRM_MODE_TV_MODE_PAL_M = 4, ++ DRM_MODE_TV_MODE_PAL_N = 5, ++ DRM_MODE_TV_MODE_SECAM = 6, ++ DRM_MODE_TV_MODE_MAX = 7, ++}; ++ ++struct drm_scrambling { ++ bool supported; ++ bool low_rates; ++}; ++ ++struct drm_scdc { ++ bool supported; ++ bool read_request; ++ struct drm_scrambling scrambling; ++}; ++ ++struct drm_hdmi_dsc_cap { ++ bool v_1p2; ++ bool native_420; ++ bool all_bpp; ++ u8 bpc_supported; ++ u8 max_slices; ++ int clk_per_slice; ++ u8 max_lanes; ++ u8 max_frl_rate_per_lane; ++ u8 total_chunk_kbytes; ++}; ++ ++struct drm_hdmi_info { ++ struct drm_scdc scdc; ++ long unsigned int y420_vdb_modes[4]; ++ long unsigned int y420_cmdb_modes[4]; ++ u8 y420_dc_modes; ++ u8 max_frl_rate_per_lane; ++ u8 max_lanes; ++ struct drm_hdmi_dsc_cap dsc_cap; ++}; ++ ++enum drm_link_status { ++ DRM_LINK_STATUS_GOOD = 0, ++ DRM_LINK_STATUS_BAD = 1, ++}; ++ ++enum drm_panel_orientation { ++ DRM_MODE_PANEL_ORIENTATION_UNKNOWN = -1, ++ DRM_MODE_PANEL_ORIENTATION_NORMAL = 0, ++ DRM_MODE_PANEL_ORIENTATION_BOTTOM_UP = 1, ++ DRM_MODE_PANEL_ORIENTATION_LEFT_UP = 2, ++ DRM_MODE_PANEL_ORIENTATION_RIGHT_UP = 3, ++}; ++ ++struct drm_monitor_range_info { ++ u16 min_vfreq; ++ u16 max_vfreq; ++}; ++ ++struct drm_luminance_range_info { ++ u32 min_luminance; ++ u32 max_luminance; ++}; ++ ++enum drm_privacy_screen_status { ++ PRIVACY_SCREEN_DISABLED = 0, ++ PRIVACY_SCREEN_ENABLED = 1, ++ PRIVACY_SCREEN_DISABLED_LOCKED = 2, ++ PRIVACY_SCREEN_ENABLED_LOCKED = 3, ++}; ++ ++struct drm_display_info { ++ unsigned int width_mm; ++ unsigned int height_mm; ++ unsigned int bpc; ++ enum subpixel_order subpixel_order; ++ int panel_orientation; ++ u32 color_formats; ++ const u32 *bus_formats; ++ unsigned int num_bus_formats; ++ u32 bus_flags; ++ int max_tmds_clock; ++ bool dvi_dual; ++ bool is_hdmi; ++ bool has_hdmi_infoframe; ++ bool rgb_quant_range_selectable; ++ u8 edid_hdmi_rgb444_dc_modes; ++ u8 edid_hdmi_ycbcr444_dc_modes; ++ u8 cea_rev; ++ struct drm_hdmi_info hdmi; ++ bool non_desktop; ++ struct drm_monitor_range_info monitor_range; ++ struct drm_luminance_range_info luminance_range; ++ u8 mso_stream_count; ++ u8 mso_pixel_overlap; ++ u32 max_dsc_bpp; ++ u8 *vics; ++ int vics_len; ++ u32 quirks; ++}; ++ ++struct drm_connector_tv_margins { ++ unsigned int bottom; ++ unsigned int left; ++ unsigned int right; ++ unsigned int top; ++}; ++ ++struct drm_tv_connector_state { ++ enum drm_mode_subconnector select_subconnector; ++ enum drm_mode_subconnector subconnector; ++ struct drm_connector_tv_margins margins; ++ unsigned int legacy_mode; ++ unsigned int mode; ++ unsigned int brightness; ++ unsigned int contrast; ++ unsigned int flicker_reduction; ++ unsigned int overscan; ++ unsigned int saturation; ++ unsigned int hue; ++}; ++ ++struct drm_connector; ++ ++struct drm_crtc; ++ ++struct drm_encoder; ++ ++struct drm_crtc_commit; ++ ++struct drm_writeback_job; ++ ++struct drm_property_blob; ++ ++struct drm_connector_state { ++ struct drm_connector *connector; ++ struct drm_crtc *crtc; ++ struct drm_encoder *best_encoder; ++ enum drm_link_status link_status; ++ struct drm_atomic_state *state; ++ struct drm_crtc_commit *commit; ++ struct drm_tv_connector_state tv; ++ bool self_refresh_aware; ++ enum hdmi_picture_aspect picture_aspect_ratio; ++ unsigned int content_type; ++ unsigned int hdcp_content_type; ++ unsigned int scaling_mode; ++ unsigned int content_protection; ++ u32 colorspace; ++ struct drm_writeback_job *writeback_job; ++ u8 max_requested_bpc; ++ u8 max_bpc; ++ enum drm_privacy_screen_status privacy_screen_sw_state; ++ struct drm_property_blob *hdr_output_metadata; ++}; ++ ++struct drm_privacy_screen; ++ ++struct drm_cmdline_mode { ++ char name[32]; ++ bool specified; ++ bool refresh_specified; ++ bool bpp_specified; ++ unsigned int pixel_clock; ++ int xres; ++ int yres; ++ int bpp; ++ int refresh; ++ bool rb; ++ bool interlace; ++ bool cvt; ++ bool margins; ++ enum drm_connector_force force; ++ unsigned int rotation_reflection; ++ enum drm_panel_orientation panel_orientation; ++ struct drm_connector_tv_margins tv_margins; ++ enum drm_connector_tv_mode tv_mode; ++ bool tv_mode_specified; ++}; ++ ++struct drm_connector_funcs; ++ ++struct drm_connector_helper_funcs; ++ ++struct drm_edid; ++ ++struct i2c_adapter; ++ ++struct drm_tile_group; ++ ++struct drm_connector { ++ struct drm_device *dev; ++ struct device *kdev; ++ struct device_attribute *attr; ++ struct fwnode_handle *fwnode; ++ struct list_head head; ++ struct list_head global_connector_list_entry; ++ struct drm_mode_object base; ++ char *name; ++ struct mutex mutex; ++ unsigned int index; ++ int connector_type; ++ int connector_type_id; ++ bool interlace_allowed; ++ bool doublescan_allowed; ++ bool stereo_allowed; ++ bool ycbcr_420_allowed; ++ enum drm_connector_registration_state registration_state; ++ struct list_head modes; ++ enum drm_connector_status status; ++ struct list_head probed_modes; ++ struct drm_display_info display_info; ++ const struct drm_connector_funcs *funcs; ++ struct drm_property_blob *edid_blob_ptr; ++ struct drm_object_properties properties; ++ struct drm_property *scaling_mode_property; ++ struct drm_property *vrr_capable_property; ++ struct drm_property *colorspace_property; ++ struct drm_property_blob *path_blob_ptr; ++ struct drm_property *max_bpc_property; ++ struct drm_privacy_screen *privacy_screen; ++ struct notifier_block privacy_screen_notifier; ++ struct drm_property *privacy_screen_sw_state_property; ++ struct drm_property *privacy_screen_hw_state_property; ++ uint8_t polled; ++ int dpms; ++ const struct drm_connector_helper_funcs *helper_private; ++ struct drm_cmdline_mode cmdline_mode; ++ enum drm_connector_force force; ++ const struct drm_edid *edid_override; ++ struct mutex edid_override_mutex; ++ u64 epoch_counter; ++ u32 possible_encoders; ++ struct drm_encoder *encoder; ++ uint8_t eld[128]; ++ bool latency_present[2]; ++ int video_latency[2]; ++ int audio_latency[2]; ++ struct i2c_adapter *ddc; ++ int null_edid_counter; ++ unsigned int bad_edid_counter; ++ bool edid_corrupt; ++ u8 real_edid_checksum; ++ struct dentry *debugfs_entry; ++ struct drm_connector_state *state; ++ struct drm_property_blob *tile_blob_ptr; ++ bool has_tile; ++ struct drm_tile_group *tile_group; ++ bool tile_is_single_monitor; ++ uint8_t num_h_tile; ++ uint8_t num_v_tile; ++ uint8_t tile_h_loc; ++ uint8_t tile_v_loc; ++ uint16_t tile_h_size; ++ uint16_t tile_v_size; ++ struct llist_node free_node; ++ struct hdr_sink_metadata hdr_sink_metadata; ++}; ++ ++struct drm_crtc_crc_entry; ++ ++struct drm_crtc_crc { ++ spinlock_t lock; ++ const char *source; ++ bool opened; ++ bool overflow; ++ struct drm_crtc_crc_entry *entries; ++ int head; ++ int tail; ++ size_t values_cnt; ++ wait_queue_head_t wq; ++}; ++ ++struct drm_plane; ++ ++struct drm_crtc_funcs; ++ ++struct drm_crtc_helper_funcs; ++ ++struct drm_crtc_state; ++ ++struct drm_self_refresh_data; ++ ++struct drm_crtc { ++ struct drm_device *dev; ++ struct device_node *port; ++ struct list_head head; ++ char *name; ++ struct drm_modeset_lock mutex; ++ struct drm_mode_object base; ++ struct drm_plane *primary; ++ struct drm_plane *cursor; ++ unsigned int index; ++ int cursor_x; ++ int cursor_y; ++ bool enabled; ++ struct drm_display_mode mode; ++ struct drm_display_mode hwmode; ++ int x; ++ int y; ++ const struct drm_crtc_funcs *funcs; ++ uint32_t gamma_size; ++ uint16_t *gamma_store; ++ const struct drm_crtc_helper_funcs *helper_private; ++ struct drm_object_properties properties; ++ struct drm_property *scaling_filter_property; ++ struct drm_crtc_state *state; ++ struct list_head commit_list; ++ spinlock_t commit_lock; ++ struct dentry *debugfs_entry; ++ struct drm_crtc_crc crc; ++ unsigned int fence_context; ++ spinlock_t fence_lock; ++ long unsigned int fence_seqno; ++ char timeline_name[32]; ++ struct drm_self_refresh_data *self_refresh_data; ++}; ++ ++struct drm_encoder_funcs; ++ ++struct drm_encoder_helper_funcs; ++ ++struct drm_encoder { ++ struct drm_device *dev; ++ struct list_head head; ++ struct drm_mode_object base; ++ char *name; ++ int encoder_type; ++ unsigned int index; ++ uint32_t possible_crtcs; ++ uint32_t possible_clones; ++ struct drm_crtc *crtc; ++ struct list_head bridge_chain; ++ const struct drm_encoder_funcs *funcs; ++ const struct drm_encoder_helper_funcs *helper_private; ++}; ++ ++struct __drm_planes_state; ++ ++struct __drm_crtcs_state; ++ ++struct __drm_connnectors_state; ++ ++struct __drm_private_objs_state; ++ ++struct drm_atomic_state { ++ struct kref ref; ++ struct drm_device *dev; ++ bool allow_modeset: 1; ++ bool legacy_cursor_update: 1; ++ bool async_update: 1; ++ bool duplicated: 1; ++ struct __drm_planes_state *planes; ++ struct __drm_crtcs_state *crtcs; ++ int num_connector; ++ struct __drm_connnectors_state *connectors; ++ int num_private_objs; ++ struct __drm_private_objs_state *private_objs; ++ struct drm_modeset_acquire_ctx *acquire_ctx; ++ struct drm_crtc_commit *fake_commit; ++ struct work_struct commit_work; ++}; ++ ++struct drm_pending_vblank_event; ++ ++struct drm_crtc_commit { ++ struct drm_crtc *crtc; ++ struct kref ref; ++ struct completion flip_done; ++ struct completion hw_done; ++ struct completion cleanup_done; ++ struct list_head commit_entry; ++ struct drm_pending_vblank_event *event; ++ bool abort_completion; ++}; ++ ++struct drm_property_blob { ++ struct drm_mode_object base; ++ struct drm_device *dev; ++ struct list_head head_global; ++ struct list_head head_file; ++ size_t length; ++ void *data; ++}; ++ ++struct drm_printer; ++ ++struct drm_connector_funcs { ++ int (*dpms)(struct drm_connector *, int); ++ void (*reset)(struct drm_connector *); ++ enum drm_connector_status (*detect)(struct drm_connector *, bool); ++ void (*force)(struct drm_connector *); ++ int (*fill_modes)(struct drm_connector *, uint32_t, uint32_t); ++ int (*set_property)(struct drm_connector *, struct drm_property *, uint64_t); ++ int (*late_register)(struct drm_connector *); ++ void (*early_unregister)(struct drm_connector *); ++ void (*destroy)(struct drm_connector *); ++ struct drm_connector_state * (*atomic_duplicate_state)(struct drm_connector *); ++ void (*atomic_destroy_state)(struct drm_connector *, struct drm_connector_state *); ++ int (*atomic_set_property)(struct drm_connector *, struct drm_connector_state *, struct drm_property *, uint64_t); ++ int (*atomic_get_property)(struct drm_connector *, const struct drm_connector_state *, struct drm_property *, uint64_t *); ++ void (*atomic_print_state)(struct drm_printer *, const struct drm_connector_state *); ++ void (*oob_hotplug_event)(struct drm_connector *); ++ void (*debugfs_init)(struct drm_connector *, struct dentry *); ++}; ++ ++struct drm_printer { ++ void (*printfn)(struct drm_printer *, struct va_format *); ++ void (*puts)(struct drm_printer *, const char *); ++ void *arg; ++ const char *prefix; ++}; ++ ++struct drm_writeback_connector; ++ ++struct drm_connector_helper_funcs { ++ int (*get_modes)(struct drm_connector *); ++ int (*detect_ctx)(struct drm_connector *, struct drm_modeset_acquire_ctx *, bool); ++ enum drm_mode_status (*mode_valid)(struct drm_connector *, struct drm_display_mode *); ++ int (*mode_valid_ctx)(struct drm_connector *, struct drm_display_mode *, struct drm_modeset_acquire_ctx *, enum drm_mode_status *); ++ struct drm_encoder * (*best_encoder)(struct drm_connector *); ++ struct drm_encoder * (*atomic_best_encoder)(struct drm_connector *, struct drm_atomic_state *); ++ int (*atomic_check)(struct drm_connector *, struct drm_atomic_state *); ++ void (*atomic_commit)(struct drm_connector *, struct drm_atomic_state *); ++ int (*prepare_writeback_job)(struct drm_writeback_connector *, struct drm_writeback_job *); ++ void (*cleanup_writeback_job)(struct drm_writeback_connector *, struct drm_writeback_job *); ++ void (*enable_hpd)(struct drm_connector *); ++ void (*disable_hpd)(struct drm_connector *); ++}; ++ ++struct edid; ++ ++struct drm_edid { ++ size_t size; ++ const struct edid *edid; ++}; ++ ++struct drm_tile_group { ++ struct kref refcount; ++ struct drm_device *dev; ++ int id; ++ u8 group_data[8]; ++}; ++ ++enum drm_color_encoding { ++ DRM_COLOR_YCBCR_BT601 = 0, ++ DRM_COLOR_YCBCR_BT709 = 1, ++ DRM_COLOR_YCBCR_BT2020 = 2, ++ DRM_COLOR_ENCODING_MAX = 3, ++}; ++ ++enum drm_color_range { ++ DRM_COLOR_YCBCR_LIMITED_RANGE = 0, ++ DRM_COLOR_YCBCR_FULL_RANGE = 1, ++ DRM_COLOR_RANGE_MAX = 2, ++}; ++ ++struct drm_rect { ++ int x1; ++ int y1; ++ int x2; ++ int y2; ++}; ++ ++enum drm_scaling_filter { ++ DRM_SCALING_FILTER_DEFAULT = 0, ++ DRM_SCALING_FILTER_NEAREST_NEIGHBOR = 1, ++}; ++ ++struct dma_fence; ++ ++struct drm_plane_state { ++ struct drm_plane *plane; ++ struct drm_crtc *crtc; ++ struct drm_framebuffer *fb; ++ struct dma_fence *fence; ++ int32_t crtc_x; ++ int32_t crtc_y; ++ uint32_t crtc_w; ++ uint32_t crtc_h; ++ uint32_t src_x; ++ uint32_t src_y; ++ uint32_t src_h; ++ uint32_t src_w; ++ u16 alpha; ++ uint16_t pixel_blend_mode; ++ unsigned int rotation; ++ unsigned int zpos; ++ unsigned int normalized_zpos; ++ enum drm_color_encoding color_encoding; ++ enum drm_color_range color_range; ++ struct drm_property_blob *fb_damage_clips; ++ struct drm_rect src; ++ struct drm_rect dst; ++ bool visible; ++ enum drm_scaling_filter scaling_filter; ++ struct drm_crtc_commit *commit; ++ struct drm_atomic_state *state; ++}; ++ ++enum drm_plane_type { ++ DRM_PLANE_TYPE_OVERLAY = 0, ++ DRM_PLANE_TYPE_PRIMARY = 1, ++ DRM_PLANE_TYPE_CURSOR = 2, ++}; ++ ++struct drm_plane_funcs; ++ ++struct drm_plane_helper_funcs; ++ ++struct drm_plane { ++ struct drm_device *dev; ++ struct list_head head; ++ char *name; ++ struct drm_modeset_lock mutex; ++ struct drm_mode_object base; ++ uint32_t possible_crtcs; ++ uint32_t *format_types; ++ unsigned int format_count; ++ bool format_default; ++ uint64_t *modifiers; ++ unsigned int modifier_count; ++ struct drm_crtc *crtc; ++ struct drm_framebuffer *fb; ++ struct drm_framebuffer *old_fb; ++ const struct drm_plane_funcs *funcs; ++ struct drm_object_properties properties; ++ enum drm_plane_type type; ++ unsigned int index; ++ const struct drm_plane_helper_funcs *helper_private; ++ struct drm_plane_state *state; ++ struct drm_property *alpha_property; ++ struct drm_property *zpos_property; ++ struct drm_property *rotation_property; ++ struct drm_property *blend_mode_property; ++ struct drm_property *color_encoding_property; ++ struct drm_property *color_range_property; ++ struct drm_property *scaling_filter_property; ++}; ++ ++struct drm_plane_funcs { ++ int (*update_plane)(struct drm_plane *, struct drm_crtc *, struct drm_framebuffer *, int, int, unsigned int, unsigned int, uint32_t, uint32_t, uint32_t, uint32_t, struct drm_modeset_acquire_ctx *); ++ int (*disable_plane)(struct drm_plane *, struct drm_modeset_acquire_ctx *); ++ void (*destroy)(struct drm_plane *); ++ void (*reset)(struct drm_plane *); ++ int (*set_property)(struct drm_plane *, struct drm_property *, uint64_t); ++ struct drm_plane_state * (*atomic_duplicate_state)(struct drm_plane *); ++ void (*atomic_destroy_state)(struct drm_plane *, struct drm_plane_state *); ++ int (*atomic_set_property)(struct drm_plane *, struct drm_plane_state *, struct drm_property *, uint64_t); ++ int (*atomic_get_property)(struct drm_plane *, const struct drm_plane_state *, struct drm_property *, uint64_t *); ++ int (*late_register)(struct drm_plane *); ++ void (*early_unregister)(struct drm_plane *); ++ void (*atomic_print_state)(struct drm_printer *, const struct drm_plane_state *); ++ bool (*format_mod_supported)(struct drm_plane *, uint32_t, uint64_t); ++}; ++ ++struct drm_plane_helper_funcs { ++ int (*prepare_fb)(struct drm_plane *, struct drm_plane_state *); ++ void (*cleanup_fb)(struct drm_plane *, struct drm_plane_state *); ++ int (*begin_fb_access)(struct drm_plane *, struct drm_plane_state *); ++ void (*end_fb_access)(struct drm_plane *, struct drm_plane_state *); ++ int (*atomic_check)(struct drm_plane *, struct drm_atomic_state *); ++ void (*atomic_update)(struct drm_plane *, struct drm_atomic_state *); ++ void (*atomic_enable)(struct drm_plane *, struct drm_atomic_state *); ++ void (*atomic_disable)(struct drm_plane *, struct drm_atomic_state *); ++ int (*atomic_async_check)(struct drm_plane *, struct drm_atomic_state *); ++ void (*atomic_async_update)(struct drm_plane *, struct drm_atomic_state *); ++}; ++ ++struct drm_crtc_crc_entry { ++ bool has_frame_counter; ++ uint32_t frame; ++ uint32_t crcs[10]; ++}; ++ ++struct drm_crtc_state { ++ struct drm_crtc *crtc; ++ bool enable; ++ bool active; ++ bool planes_changed: 1; ++ bool mode_changed: 1; ++ bool active_changed: 1; ++ bool connectors_changed: 1; ++ bool zpos_changed: 1; ++ bool color_mgmt_changed: 1; ++ bool no_vblank: 1; ++ u32 plane_mask; ++ u32 connector_mask; ++ u32 encoder_mask; ++ struct drm_display_mode adjusted_mode; ++ struct drm_display_mode mode; ++ struct drm_property_blob *mode_blob; ++ struct drm_property_blob *degamma_lut; ++ struct drm_property_blob *ctm; ++ struct drm_property_blob *gamma_lut; ++ u32 target_vblank; ++ bool async_flip; ++ bool vrr_enabled; ++ bool self_refresh_active; ++ enum drm_scaling_filter scaling_filter; ++ struct drm_pending_vblank_event *event; ++ struct drm_crtc_commit *commit; ++ struct drm_atomic_state *state; ++}; ++ ++struct drm_mode_set; ++ ++struct drm_crtc_funcs { ++ void (*reset)(struct drm_crtc *); ++ int (*cursor_set)(struct drm_crtc *, struct drm_file *, uint32_t, uint32_t, uint32_t); ++ int (*cursor_set2)(struct drm_crtc *, struct drm_file *, uint32_t, uint32_t, uint32_t, int32_t, int32_t); ++ int (*cursor_move)(struct drm_crtc *, int, int); ++ int (*gamma_set)(struct drm_crtc *, u16 *, u16 *, u16 *, uint32_t, struct drm_modeset_acquire_ctx *); ++ void (*destroy)(struct drm_crtc *); ++ int (*set_config)(struct drm_mode_set *, struct drm_modeset_acquire_ctx *); ++ int (*page_flip)(struct drm_crtc *, struct drm_framebuffer *, struct drm_pending_vblank_event *, uint32_t, struct drm_modeset_acquire_ctx *); ++ int (*page_flip_target)(struct drm_crtc *, struct drm_framebuffer *, struct drm_pending_vblank_event *, uint32_t, uint32_t, struct drm_modeset_acquire_ctx *); ++ int (*set_property)(struct drm_crtc *, struct drm_property *, uint64_t); ++ struct drm_crtc_state * (*atomic_duplicate_state)(struct drm_crtc *); ++ void (*atomic_destroy_state)(struct drm_crtc *, struct drm_crtc_state *); ++ int (*atomic_set_property)(struct drm_crtc *, struct drm_crtc_state *, struct drm_property *, uint64_t); ++ int (*atomic_get_property)(struct drm_crtc *, const struct drm_crtc_state *, struct drm_property *, uint64_t *); ++ int (*late_register)(struct drm_crtc *); ++ void (*early_unregister)(struct drm_crtc *); ++ int (*set_crc_source)(struct drm_crtc *, const char *); ++ int (*verify_crc_source)(struct drm_crtc *, const char *, size_t *); ++ const char * const * (*get_crc_sources)(struct drm_crtc *, size_t *); ++ void (*atomic_print_state)(struct drm_printer *, const struct drm_crtc_state *); ++ u32 (*get_vblank_counter)(struct drm_crtc *); ++ int (*enable_vblank)(struct drm_crtc *); ++ void (*disable_vblank)(struct drm_crtc *); ++ bool (*get_vblank_timestamp)(struct drm_crtc *, int *, ktime_t *, bool); ++}; ++ ++struct drm_mode_set { ++ struct drm_framebuffer *fb; ++ struct drm_crtc *crtc; ++ struct drm_display_mode *mode; ++ uint32_t x; ++ uint32_t y; ++ struct drm_connector **connectors; ++ size_t num_connectors; ++}; ++ ++enum mode_set_atomic { ++ LEAVE_ATOMIC_MODE_SET = 0, ++ ENTER_ATOMIC_MODE_SET = 1, ++}; ++ ++struct drm_crtc_helper_funcs { ++ void (*dpms)(struct drm_crtc *, int); ++ void (*prepare)(struct drm_crtc *); ++ void (*commit)(struct drm_crtc *); ++ enum drm_mode_status (*mode_valid)(struct drm_crtc *, const struct drm_display_mode *); ++ bool (*mode_fixup)(struct drm_crtc *, const struct drm_display_mode *, struct drm_display_mode *); ++ int (*mode_set)(struct drm_crtc *, struct drm_display_mode *, struct drm_display_mode *, int, int, struct drm_framebuffer *); ++ void (*mode_set_nofb)(struct drm_crtc *); ++ int (*mode_set_base)(struct drm_crtc *, int, int, struct drm_framebuffer *); ++ int (*mode_set_base_atomic)(struct drm_crtc *, struct drm_framebuffer *, int, int, enum mode_set_atomic); ++ void (*disable)(struct drm_crtc *); ++ int (*atomic_check)(struct drm_crtc *, struct drm_atomic_state *); ++ void (*atomic_begin)(struct drm_crtc *, struct drm_atomic_state *); ++ void (*atomic_flush)(struct drm_crtc *, struct drm_atomic_state *); ++ void (*atomic_enable)(struct drm_crtc *, struct drm_atomic_state *); ++ void (*atomic_disable)(struct drm_crtc *, struct drm_atomic_state *); ++ bool (*get_scanout_position)(struct drm_crtc *, bool, int *, int *, ktime_t *, ktime_t *, const struct drm_display_mode *); ++}; ++ ++struct __drm_planes_state { ++ struct drm_plane *ptr; ++ struct drm_plane_state *state; ++ struct drm_plane_state *old_state; ++ struct drm_plane_state *new_state; ++}; ++ ++struct __drm_crtcs_state { ++ struct drm_crtc *ptr; ++ struct drm_crtc_state *state; ++ struct drm_crtc_state *old_state; ++ struct drm_crtc_state *new_state; ++ struct drm_crtc_commit *commit; ++ s32 *out_fence_ptr; ++ u64 last_vblank_count; ++}; ++ ++struct __drm_connnectors_state { ++ struct drm_connector *ptr; ++ struct drm_connector_state *state; ++ struct drm_connector_state *old_state; ++ struct drm_connector_state *new_state; ++ s32 *out_fence_ptr; ++}; ++ ++struct drm_private_state; ++ ++struct drm_private_obj; ++ ++struct drm_private_state_funcs { ++ struct drm_private_state * (*atomic_duplicate_state)(struct drm_private_obj *); ++ void (*atomic_destroy_state)(struct drm_private_obj *, struct drm_private_state *); ++ void (*atomic_print_state)(struct drm_printer *, const struct drm_private_state *); ++}; ++ ++struct drm_private_state { ++ struct drm_atomic_state *state; ++ struct drm_private_obj *obj; ++}; ++ ++struct drm_private_obj { ++ struct list_head head; ++ struct drm_modeset_lock lock; ++ struct drm_private_state *state; ++ const struct drm_private_state_funcs *funcs; ++}; ++ ++struct __drm_private_objs_state { ++ struct drm_private_obj *ptr; ++ struct drm_private_state *state; ++ struct drm_private_state *old_state; ++ struct drm_private_state *new_state; ++}; ++ ++struct drm_bus_cfg { ++ u32 format; ++ u32 flags; ++}; ++ ++struct drm_bridge; ++ ++struct drm_bridge_state { ++ struct drm_private_state base; ++ struct drm_bridge *bridge; ++ struct drm_bus_cfg input_bus_cfg; ++ struct drm_bus_cfg output_bus_cfg; ++}; ++ ++enum drm_bridge_ops { ++ DRM_BRIDGE_OP_DETECT = 1, ++ DRM_BRIDGE_OP_EDID = 2, ++ DRM_BRIDGE_OP_HPD = 4, ++ DRM_BRIDGE_OP_MODES = 8, ++}; ++ ++struct drm_bridge_timings; ++ ++struct drm_bridge_funcs; ++ ++struct drm_bridge { ++ struct drm_private_obj base; ++ struct drm_device *dev; ++ struct drm_encoder *encoder; ++ struct list_head chain_node; ++ struct list_head list; ++ const struct drm_bridge_timings *timings; ++ const struct drm_bridge_funcs *funcs; ++ void *driver_private; ++ enum drm_bridge_ops ops; ++ int type; ++ bool interlace_allowed; ++ bool pre_enable_prev_first; ++ struct i2c_adapter *ddc; ++ struct mutex hpd_mutex; ++ void (*hpd_cb)(void *, enum drm_connector_status); ++ void *hpd_data; ++}; ++ ++struct drm_encoder_funcs { ++ void (*reset)(struct drm_encoder *); ++ void (*destroy)(struct drm_encoder *); ++ int (*late_register)(struct drm_encoder *); ++ void (*early_unregister)(struct drm_encoder *); ++}; ++ ++struct drm_encoder_helper_funcs { ++ void (*dpms)(struct drm_encoder *, int); ++ enum drm_mode_status (*mode_valid)(struct drm_encoder *, const struct drm_display_mode *); ++ bool (*mode_fixup)(struct drm_encoder *, const struct drm_display_mode *, struct drm_display_mode *); ++ void (*prepare)(struct drm_encoder *); ++ void (*commit)(struct drm_encoder *); ++ void (*mode_set)(struct drm_encoder *, struct drm_display_mode *, struct drm_display_mode *); ++ void (*atomic_mode_set)(struct drm_encoder *, struct drm_crtc_state *, struct drm_connector_state *); ++ enum drm_connector_status (*detect)(struct drm_encoder *, struct drm_connector *); ++ void (*atomic_disable)(struct drm_encoder *, struct drm_atomic_state *); ++ void (*atomic_enable)(struct drm_encoder *, struct drm_atomic_state *); ++ void (*disable)(struct drm_encoder *); ++ void (*enable)(struct drm_encoder *); ++ int (*atomic_check)(struct drm_encoder *, struct drm_crtc_state *, struct drm_connector_state *); ++}; ++ ++enum drm_bridge_attach_flags { ++ DRM_BRIDGE_ATTACH_NO_CONNECTOR = 1, ++}; ++ ++struct drm_bridge_funcs { ++ int (*attach)(struct drm_bridge *, enum drm_bridge_attach_flags); ++ void (*detach)(struct drm_bridge *); ++ enum drm_mode_status (*mode_valid)(struct drm_bridge *, const struct drm_display_info *, const struct drm_display_mode *); ++ bool (*mode_fixup)(struct drm_bridge *, const struct drm_display_mode *, struct drm_display_mode *); ++ void (*disable)(struct drm_bridge *); ++ void (*post_disable)(struct drm_bridge *); ++ void (*mode_set)(struct drm_bridge *, const struct drm_display_mode *, const struct drm_display_mode *); ++ void (*pre_enable)(struct drm_bridge *); ++ void (*enable)(struct drm_bridge *); ++ void (*atomic_pre_enable)(struct drm_bridge *, struct drm_bridge_state *); ++ void (*atomic_enable)(struct drm_bridge *, struct drm_bridge_state *); ++ void (*atomic_disable)(struct drm_bridge *, struct drm_bridge_state *); ++ void (*atomic_post_disable)(struct drm_bridge *, struct drm_bridge_state *); ++ struct drm_bridge_state * (*atomic_duplicate_state)(struct drm_bridge *); ++ void (*atomic_destroy_state)(struct drm_bridge *, struct drm_bridge_state *); ++ u32 * (*atomic_get_output_bus_fmts)(struct drm_bridge *, struct drm_bridge_state *, struct drm_crtc_state *, struct drm_connector_state *, unsigned int *); ++ u32 * (*atomic_get_input_bus_fmts)(struct drm_bridge *, struct drm_bridge_state *, struct drm_crtc_state *, struct drm_connector_state *, u32, unsigned int *); ++ int (*atomic_check)(struct drm_bridge *, struct drm_bridge_state *, struct drm_crtc_state *, struct drm_connector_state *); ++ struct drm_bridge_state * (*atomic_reset)(struct drm_bridge *); ++ enum drm_connector_status (*detect)(struct drm_bridge *); ++ int (*get_modes)(struct drm_bridge *, struct drm_connector *); ++ struct edid * (*get_edid)(struct drm_bridge *, struct drm_connector *); ++ void (*hpd_notify)(struct drm_bridge *, enum drm_connector_status); ++ void (*hpd_enable)(struct drm_bridge *); ++ void (*hpd_disable)(struct drm_bridge *); ++ void (*debugfs_init)(struct drm_bridge *, struct dentry *); ++}; ++ ++struct drm_bridge_timings { ++ u32 input_bus_flags; ++ u32 setup_time_ps; ++ u32 hold_time_ps; ++ bool dual_link; ++}; ++ ++struct drm_mode_crtc_lut { ++ __u32 crtc_id; ++ __u32 gamma_size; ++ __u64 red; ++ __u64 green; ++ __u64 blue; ++}; ++ ++struct drm_color_lut { ++ __u16 red; ++ __u16 green; ++ __u16 blue; ++ __u16 reserved; ++}; ++ ++struct drm_prop_enum_list { ++ int type; ++ const char *name; ++}; ++ ++enum drm_color_lut_tests { ++ DRM_COLOR_LUT_EQUAL_CHANNELS = 1, ++ DRM_COLOR_LUT_NON_DECREASING = 2, ++}; ++ ++struct dma_fence_ops; ++ ++struct dma_fence { ++ spinlock_t *lock; ++ const struct dma_fence_ops *ops; ++ union { ++ struct list_head cb_list; ++ ktime_t timestamp; ++ struct callback_head rcu; ++ }; ++ u64 context; ++ u64 seqno; ++ long unsigned int flags; ++ struct kref refcount; ++ int error; ++}; ++ ++struct dma_fence_ops { ++ bool use_64bit_seqno; ++ const char * (*get_driver_name)(struct dma_fence *); ++ const char * (*get_timeline_name)(struct dma_fence *); ++ bool (*enable_signaling)(struct dma_fence *); ++ bool (*signaled)(struct dma_fence *); ++ long int (*wait)(struct dma_fence *, bool, long int); ++ void (*release)(struct dma_fence *); ++ void (*fence_value_str)(struct dma_fence *, char *, int); ++ void (*timeline_value_str)(struct dma_fence *, char *, int); ++ void (*set_deadline)(struct dma_fence *, ktime_t); ++}; ++ ++struct drm_clip_rect { ++ short unsigned int x1; ++ short unsigned int y1; ++ short unsigned int x2; ++ short unsigned int y2; ++}; ++ ++struct drm_event { ++ __u32 type; ++ __u32 length; ++}; ++ ++struct drm_event_vblank { ++ struct drm_event base; ++ __u64 user_data; ++ __u32 tv_sec; ++ __u32 tv_usec; ++ __u32 sequence; ++ __u32 crtc_id; ++}; ++ ++struct drm_event_crtc_sequence { ++ struct drm_event base; ++ __u64 user_data; ++ __s64 time_ns; ++ __u64 sequence; ++}; ++ ++struct drm_mode_modeinfo { ++ __u32 clock; ++ __u16 hdisplay; ++ __u16 hsync_start; ++ __u16 hsync_end; ++ __u16 htotal; ++ __u16 hskew; ++ __u16 vdisplay; ++ __u16 vsync_start; ++ __u16 vsync_end; ++ __u16 vtotal; ++ __u16 vscan; ++ __u32 vrefresh; ++ __u32 flags; ++ __u32 type; ++ char name[32]; ++}; ++ ++struct drm_mode_crtc { ++ __u64 set_connectors_ptr; ++ __u32 count_connectors; ++ __u32 crtc_id; ++ __u32 fb_id; ++ __u32 x; ++ __u32 y; ++ __u32 gamma_size; ++ __u32 mode_valid; ++ struct drm_mode_modeinfo mode; ++}; ++ ++struct drm_framebuffer_funcs; ++ ++struct drm_framebuffer { ++ struct drm_device *dev; ++ struct list_head head; ++ struct drm_mode_object base; ++ char comm[16]; ++ const struct drm_format_info *format; ++ const struct drm_framebuffer_funcs *funcs; ++ unsigned int pitches[4]; ++ unsigned int offsets[4]; ++ uint64_t modifier; ++ unsigned int width; ++ unsigned int height; ++ int flags; ++ int hot_x; ++ int hot_y; ++ struct list_head filp_head; ++ struct drm_gem_object *obj[4]; ++}; ++ ++struct drm_pending_event { ++ struct completion *completion; ++ void (*completion_release)(struct completion *); ++ struct drm_event *event; ++ struct dma_fence *fence; ++ struct drm_file *file_priv; ++ struct list_head link; ++ struct list_head pending_link; ++}; ++ ++struct drm_pending_vblank_event { ++ struct drm_pending_event base; ++ unsigned int pipe; ++ u64 sequence; ++ union { ++ struct drm_event base; ++ struct drm_event_vblank vbl; ++ struct drm_event_crtc_sequence seq; ++ } event; ++}; ++ ++struct drm_framebuffer_funcs { ++ void (*destroy)(struct drm_framebuffer *); ++ int (*create_handle)(struct drm_framebuffer *, struct drm_file *, unsigned int *); ++ int (*dirty)(struct drm_framebuffer *, struct drm_file *, unsigned int, unsigned int, struct drm_clip_rect *, unsigned int); ++}; ++ ++typedef void (*drmres_release_t)(struct drm_device *, void *); ++ ++struct drm_mode_get_encoder { ++ __u32 encoder_id; ++ __u32 encoder_type; ++ __u32 crtc_id; ++ __u32 possible_crtcs; ++ __u32 possible_clones; ++}; ++ ++struct drm_connector_list_iter { ++ struct drm_device *dev; ++ struct drm_connector *conn; ++}; ++ ++struct iosys_map { ++ union { ++ void *vaddr_iomem; ++ void *vaddr; ++ }; ++ bool is_iomem; ++}; ++ ++struct dma_fence_cb; ++ ++typedef void (*dma_fence_func_t)(struct dma_fence *, struct dma_fence_cb *); ++ ++struct dma_fence_cb { ++ struct list_head node; ++ dma_fence_func_t func; ++}; ++ ++struct dma_buf_ops { ++ bool cache_sgt_mapping; ++ int (*attach)(struct dma_buf *, struct dma_buf_attachment *); ++ void (*detach)(struct dma_buf *, struct dma_buf_attachment *); ++ int (*pin)(struct dma_buf_attachment *); ++ void (*unpin)(struct dma_buf_attachment *); ++ struct sg_table * (*map_dma_buf)(struct dma_buf_attachment *, enum dma_data_direction); ++ void (*unmap_dma_buf)(struct dma_buf_attachment *, struct sg_table *, enum dma_data_direction); ++ void (*release)(struct dma_buf *); ++ int (*begin_cpu_access)(struct dma_buf *, enum dma_data_direction); ++ int (*end_cpu_access)(struct dma_buf *, enum dma_data_direction); ++ int (*mmap)(struct dma_buf *, struct vm_area_struct *); ++ int (*vmap)(struct dma_buf *, struct iosys_map *); ++ void (*vunmap)(struct dma_buf *, struct iosys_map *); ++}; ++ ++struct dma_buf_poll_cb_t { ++ struct dma_fence_cb cb; ++ wait_queue_head_t *poll; ++ __poll_t active; ++}; ++ ++struct dma_resv; ++ ++struct dma_buf { ++ size_t size; ++ struct file *file; ++ struct list_head attachments; ++ const struct dma_buf_ops *ops; ++ unsigned int vmapping_counter; ++ struct iosys_map vmap_ptr; ++ const char *exp_name; ++ const char *name; ++ spinlock_t name_lock; ++ struct module *owner; ++ struct list_head list_node; ++ void *priv; ++ struct dma_resv *resv; ++ wait_queue_head_t poll; ++ struct dma_buf_poll_cb_t cb_in; ++ struct dma_buf_poll_cb_t cb_out; ++}; ++ ++struct dma_buf_attach_ops; ++ ++struct dma_buf_attachment { ++ struct dma_buf *dmabuf; ++ struct device *dev; ++ struct list_head node; ++ struct sg_table *sgt; ++ enum dma_data_direction dir; ++ bool peer2peer; ++ const struct dma_buf_attach_ops *importer_ops; ++ void *importer_priv; ++ void *priv; ++}; ++ ++struct dma_resv_list; ++ ++struct dma_resv { ++ struct ww_mutex lock; ++ struct dma_resv_list *fences; ++}; ++ ++struct dma_buf_attach_ops { ++ bool allow_peer2peer; ++ void (*move_notify)(struct dma_buf_attachment *); ++}; ++ ++struct drm_gem_close { ++ __u32 handle; ++ __u32 pad; ++}; ++ ++struct drm_gem_flink { ++ __u32 handle; ++ __u32 name; ++}; ++ ++struct drm_gem_open { ++ __u32 name; ++ __u32 handle; ++ __u64 size; ++}; ++ ++struct ww_class { ++ atomic_long_t stamp; ++ struct lock_class_key acquire_key; ++ struct lock_class_key mutex_key; ++ const char *acquire_name; ++ const char *mutex_name; ++ unsigned int is_wait_die; ++}; ++ ++struct drm_mm; ++ ++struct drm_mm_node { ++ long unsigned int color; ++ u64 start; ++ u64 size; ++ struct drm_mm *mm; ++ struct list_head node_list; ++ struct list_head hole_stack; ++ struct rb_node rb; ++ struct rb_node rb_hole_size; ++ struct rb_node rb_hole_addr; ++ u64 __subtree_last; ++ u64 hole_size; ++ u64 subtree_max_hole; ++ long unsigned int flags; ++}; ++ ++struct drm_mm { ++ void (*color_adjust)(const struct drm_mm_node *, long unsigned int, u64 *, u64 *); ++ struct list_head hole_stack; ++ struct drm_mm_node head_node; ++ struct rb_root_cached interval_tree; ++ struct rb_root_cached holes_size; ++ struct rb_root holes_addr; ++ long unsigned int scan_active; ++}; ++ ++struct drm_vma_offset_manager { ++ rwlock_t vm_lock; ++ struct drm_mm vm_addr_space_mm; ++}; ++ ++struct drm_vma_offset_node { ++ rwlock_t vm_lock; ++ struct drm_mm_node vm_node; ++ struct rb_root vm_files; ++ void *driver_private; ++}; ++ ++struct drm_gem_object_funcs; ++ ++struct drm_gem_lru; ++ ++struct drm_gem_object { ++ struct kref refcount; ++ unsigned int handle_count; ++ struct drm_device *dev; ++ struct file *filp; ++ struct drm_vma_offset_node vma_node; ++ size_t size; ++ int name; ++ struct dma_buf *dma_buf; ++ struct dma_buf_attachment *import_attach; ++ struct dma_resv *resv; ++ struct dma_resv _resv; ++ const struct drm_gem_object_funcs *funcs; ++ struct list_head lru_node; ++ struct drm_gem_lru *lru; ++}; ++ ++enum dma_resv_usage { ++ DMA_RESV_USAGE_KERNEL = 0, ++ DMA_RESV_USAGE_WRITE = 1, ++ DMA_RESV_USAGE_READ = 2, ++ DMA_RESV_USAGE_BOOKKEEP = 3, ++}; ++ ++struct drm_gem_object_funcs { ++ void (*free)(struct drm_gem_object *); ++ int (*open)(struct drm_gem_object *, struct drm_file *); ++ void (*close)(struct drm_gem_object *, struct drm_file *); ++ void (*print_info)(struct drm_printer *, unsigned int, const struct drm_gem_object *); ++ struct dma_buf * (*export)(struct drm_gem_object *, int); ++ int (*pin)(struct drm_gem_object *); ++ void (*unpin)(struct drm_gem_object *); ++ struct sg_table * (*get_sg_table)(struct drm_gem_object *); ++ int (*vmap)(struct drm_gem_object *, struct iosys_map *); ++ void (*vunmap)(struct drm_gem_object *, struct iosys_map *); ++ int (*mmap)(struct drm_gem_object *, struct vm_area_struct *); ++ int (*evict)(struct drm_gem_object *); ++ const struct vm_operations_struct *vm_ops; ++}; ++ ++struct drm_gem_lru { ++ struct mutex *lock; ++ long int count; ++ struct list_head list; ++}; ++ ++struct i2c_msg { ++ __u16 addr; ++ __u16 flags; ++ __u16 len; ++ __u8 *buf; ++}; ++ ++union i2c_smbus_data { ++ __u8 byte; ++ __u16 word; ++ __u8 block[34]; ++}; ++ ++struct i2c_algorithm; ++ ++struct i2c_lock_operations; ++ ++struct i2c_bus_recovery_info; ++ ++struct i2c_adapter_quirks; ++ ++struct i2c_adapter { ++ struct module *owner; ++ unsigned int class; ++ const struct i2c_algorithm *algo; ++ void *algo_data; ++ const struct i2c_lock_operations *lock_ops; ++ struct rt_mutex bus_lock; ++ struct rt_mutex mux_lock; ++ int timeout; ++ int retries; ++ struct device dev; ++ long unsigned int locked_flags; ++ int nr; ++ char name[48]; ++ struct completion dev_released; ++ struct mutex userspace_clients_lock; ++ struct list_head userspace_clients; ++ struct i2c_bus_recovery_info *bus_recovery_info; ++ const struct i2c_adapter_quirks *quirks; ++ struct irq_domain *host_notify_domain; ++ struct regulator *bus_regulator; ++}; ++ ++struct i2c_algorithm { ++ int (*master_xfer)(struct i2c_adapter *, struct i2c_msg *, int); ++ int (*master_xfer_atomic)(struct i2c_adapter *, struct i2c_msg *, int); ++ int (*smbus_xfer)(struct i2c_adapter *, u16, short unsigned int, char, u8, int, union i2c_smbus_data *); ++ int (*smbus_xfer_atomic)(struct i2c_adapter *, u16, short unsigned int, char, u8, int, union i2c_smbus_data *); ++ u32 (*functionality)(struct i2c_adapter *); ++}; ++ ++struct i2c_lock_operations { ++ void (*lock_bus)(struct i2c_adapter *, unsigned int); ++ int (*trylock_bus)(struct i2c_adapter *, unsigned int); ++ void (*unlock_bus)(struct i2c_adapter *, unsigned int); ++}; ++ ++struct pinctrl; ++ ++struct pinctrl_state; ++ ++struct i2c_bus_recovery_info { ++ int (*recover_bus)(struct i2c_adapter *); ++ int (*get_scl)(struct i2c_adapter *); ++ void (*set_scl)(struct i2c_adapter *, int); ++ int (*get_sda)(struct i2c_adapter *); ++ void (*set_sda)(struct i2c_adapter *, int); ++ int (*get_bus_free)(struct i2c_adapter *); ++ void (*prepare_recovery)(struct i2c_adapter *); ++ void (*unprepare_recovery)(struct i2c_adapter *); ++ struct gpio_desc *scl_gpiod; ++ struct gpio_desc *sda_gpiod; ++ struct pinctrl *pinctrl; ++ struct pinctrl_state *pins_default; ++ struct pinctrl_state *pins_gpio; ++}; ++ ++struct i2c_adapter_quirks { ++ u64 flags; ++ int max_num_msgs; ++ u16 max_write_len; ++ u16 max_read_len; ++ u16 max_comb_1st_msg_len; ++ u16 max_comb_2nd_msg_len; ++}; ++ ++enum drm_mode_analog { ++ DRM_MODE_ANALOG_NTSC = 0, ++ DRM_MODE_ANALOG_PAL = 1, ++}; ++ ++struct analog_param_field { ++ unsigned int even; ++ unsigned int odd; ++}; ++ ++struct analog_param_range { ++ unsigned int min; ++ unsigned int typ; ++ unsigned int max; ++}; ++ ++struct analog_parameters { ++ unsigned int num_lines; ++ unsigned int line_duration_ns; ++ struct analog_param_range hact_ns; ++ struct analog_param_range hfp_ns; ++ struct analog_param_range hslen_ns; ++ struct analog_param_range hbp_ns; ++ struct analog_param_range hblk_ns; ++ unsigned int bt601_hfp; ++ struct analog_param_field vfp_lines; ++ struct analog_param_field vslen_lines; ++ struct analog_param_field vbp_lines; ++}; ++ ++struct drm_named_mode { ++ const char *name; ++ unsigned int pixel_clock_khz; ++ unsigned int xres; ++ unsigned int yres; ++ unsigned int flags; ++ unsigned int tv_mode; ++}; ++ ++struct dma_fence_unwrap { ++ struct dma_fence *chain; ++ struct dma_fence *array; ++ unsigned int index; ++}; ++ ++typedef s64 int64_t; ++ ++enum dma_fence_flag_bits { ++ DMA_FENCE_FLAG_SIGNALED_BIT = 0, ++ DMA_FENCE_FLAG_TIMESTAMP_BIT = 1, ++ DMA_FENCE_FLAG_ENABLE_SIGNAL_BIT = 2, ++ DMA_FENCE_FLAG_USER_BITS = 3, ++}; ++ ++struct sync_file { ++ struct file *file; ++ char user_name[32]; ++ struct list_head sync_file_list; ++ wait_queue_head_t wq; ++ long unsigned int flags; ++ struct dma_fence *fence; ++ struct dma_fence_cb cb; ++}; ++ ++struct drm_syncobj_create { ++ __u32 handle; ++ __u32 flags; ++}; ++ ++struct drm_syncobj_destroy { ++ __u32 handle; ++ __u32 pad; ++}; ++ ++struct drm_syncobj_handle { ++ __u32 handle; ++ __u32 flags; ++ __s32 fd; ++ __u32 pad; ++}; ++ ++struct drm_syncobj_transfer { ++ __u32 src_handle; ++ __u32 dst_handle; ++ __u64 src_point; ++ __u64 dst_point; ++ __u32 flags; ++ __u32 pad; ++}; ++ ++struct drm_syncobj_wait { ++ __u64 handles; ++ __s64 timeout_nsec; ++ __u32 count_handles; ++ __u32 flags; ++ __u32 first_signaled; ++ __u32 pad; ++}; ++ ++struct drm_syncobj_timeline_wait { ++ __u64 handles; ++ __u64 points; ++ __s64 timeout_nsec; ++ __u32 count_handles; ++ __u32 flags; ++ __u32 first_signaled; ++ __u32 pad; ++}; ++ ++struct drm_syncobj_array { ++ __u64 handles; ++ __u32 count_handles; ++ __u32 pad; ++}; ++ ++struct drm_syncobj_timeline_array { ++ __u64 handles; ++ __u64 points; ++ __u32 count_handles; ++ __u32 flags; ++}; ++ ++struct dma_fence_chain { ++ struct dma_fence base; ++ struct dma_fence *prev; ++ u64 prev_seqno; ++ struct dma_fence *fence; ++ union { ++ struct dma_fence_cb cb; ++ struct irq_work work; ++ }; ++ spinlock_t lock; ++}; ++ ++struct drm_syncobj { ++ struct kref refcount; ++ struct dma_fence *fence; ++ struct list_head cb_list; ++ spinlock_t lock; ++ struct file *file; ++}; ++ ++struct syncobj_wait_entry { ++ struct list_head node; ++ struct task_struct *task; ++ struct dma_fence *fence; ++ struct dma_fence_cb fence_cb; ++ u64 point; ++}; ++ ++enum drm_mm_insert_mode { ++ DRM_MM_INSERT_BEST = 0, ++ DRM_MM_INSERT_LOW = 1, ++ DRM_MM_INSERT_HIGH = 2, ++ DRM_MM_INSERT_EVICT = 3, ++ DRM_MM_INSERT_ONCE = 2147483648, ++ DRM_MM_INSERT_HIGHEST = 2147483650, ++ DRM_MM_INSERT_LOWEST = 2147483649, ++}; ++ ++struct drm_vma_offset_file { ++ struct rb_node vm_rb; ++ struct drm_file *vm_tag; ++ long unsigned int vm_count; ++}; ++ ++struct rnd_state { ++ __u32 s1; ++ __u32 s2; ++ __u32 s3; ++ __u32 s4; ++}; ++ ++struct drm_version { ++ int version_major; ++ int version_minor; ++ int version_patchlevel; ++ __kernel_size_t name_len; ++ char *name; ++ __kernel_size_t date_len; ++ char *date; ++ __kernel_size_t desc_len; ++ char *desc; ++}; ++ ++struct drm_unique { ++ __kernel_size_t unique_len; ++ char *unique; ++}; ++ ++struct drm_client { ++ int idx; ++ int auth; ++ long unsigned int pid; ++ long unsigned int uid; ++ long unsigned int magic; ++ long unsigned int iocs; ++}; ++ ++enum drm_stat_type { ++ _DRM_STAT_LOCK = 0, ++ _DRM_STAT_OPENS = 1, ++ _DRM_STAT_CLOSES = 2, ++ _DRM_STAT_IOCTLS = 3, ++ _DRM_STAT_LOCKS = 4, ++ _DRM_STAT_UNLOCKS = 5, ++ _DRM_STAT_VALUE = 6, ++ _DRM_STAT_BYTE = 7, ++ _DRM_STAT_COUNT = 8, ++ _DRM_STAT_IRQ = 9, ++ _DRM_STAT_PRIMARY = 10, ++ _DRM_STAT_SECONDARY = 11, ++ _DRM_STAT_DMA = 12, ++ _DRM_STAT_SPECIAL = 13, ++ _DRM_STAT_MISSED = 14, ++}; ++ ++enum drm_vblank_seq_type { ++ _DRM_VBLANK_ABSOLUTE = 0, ++ _DRM_VBLANK_RELATIVE = 1, ++ _DRM_VBLANK_HIGH_CRTC_MASK = 62, ++ _DRM_VBLANK_EVENT = 67108864, ++ _DRM_VBLANK_FLIP = 134217728, ++ _DRM_VBLANK_NEXTONMISS = 268435456, ++ _DRM_VBLANK_SECONDARY = 536870912, ++ _DRM_VBLANK_SIGNAL = 1073741824, ++}; ++ ++struct drm_wait_vblank_request { ++ enum drm_vblank_seq_type type; ++ unsigned int sequence; ++ long unsigned int signal; ++}; ++ ++struct drm_wait_vblank_reply { ++ enum drm_vblank_seq_type type; ++ unsigned int sequence; ++ long int tval_sec; ++ long int tval_usec; ++}; ++ ++union drm_wait_vblank { ++ struct drm_wait_vblank_request request; ++ struct drm_wait_vblank_reply reply; ++}; ++ ++typedef int drm_ioctl_compat_t(struct file *, unsigned int, long unsigned int); ++ ++struct drm_version_32 { ++ int version_major; ++ int version_minor; ++ int version_patchlevel; ++ u32 name_len; ++ u32 name; ++ u32 date_len; ++ u32 date; ++ u32 desc_len; ++ u32 desc; ++}; ++ ++typedef struct drm_version_32 drm_version32_t; ++ ++struct drm_unique32 { ++ u32 unique_len; ++ u32 unique; ++}; ++ ++typedef struct drm_unique32 drm_unique32_t; ++ ++struct drm_client32 { ++ int idx; ++ int auth; ++ u32 pid; ++ u32 uid; ++ u32 magic; ++ u32 iocs; ++}; ++ ++typedef struct drm_client32 drm_client32_t; ++ ++struct drm_stats32 { ++ u32 count; ++ struct { ++ u32 value; ++ enum drm_stat_type type; ++ } data[15]; ++}; ++ ++typedef struct drm_stats32 drm_stats32_t; ++ ++struct drm_wait_vblank_request32 { ++ enum drm_vblank_seq_type type; ++ unsigned int sequence; ++ u32 signal; ++}; ++ ++struct drm_wait_vblank_reply32 { ++ enum drm_vblank_seq_type type; ++ unsigned int sequence; ++ s32 tval_sec; ++ s32 tval_usec; ++}; ++ ++union drm_wait_vblank32 { ++ struct drm_wait_vblank_request32 request; ++ struct drm_wait_vblank_reply32 reply; ++}; ++ ++typedef union drm_wait_vblank32 drm_wait_vblank32_t; ++ ++struct drm_info_list { ++ const char *name; ++ int (*show)(struct seq_file *, void *); ++ u32 driver_features; ++ void *data; ++}; ++ ++struct drm_info_node { ++ struct drm_minor *minor; ++ const struct drm_info_list *info_ent; ++ struct list_head list; ++ struct dentry *dent; ++}; ++ ++struct drm_debugfs_info { ++ const char *name; ++ int (*show)(struct seq_file *, void *); ++ u32 driver_features; ++ void *data; ++}; ++ ++struct drm_debugfs_entry { ++ struct drm_device *dev; ++ struct drm_debugfs_info file; ++ struct list_head list; ++}; ++ ++struct est_timings { ++ u8 t1; ++ u8 t2; ++ u8 mfg_rsvd; ++}; ++ ++struct std_timing { ++ u8 hsize; ++ u8 vfreq_aspect; ++}; ++ ++struct detailed_pixel_timing { ++ u8 hactive_lo; ++ u8 hblank_lo; ++ u8 hactive_hblank_hi; ++ u8 vactive_lo; ++ u8 vblank_lo; ++ u8 vactive_vblank_hi; ++ u8 hsync_offset_lo; ++ u8 hsync_pulse_width_lo; ++ u8 vsync_offset_pulse_width_lo; ++ u8 hsync_vsync_offset_pulse_width_hi; ++ u8 width_mm_lo; ++ u8 height_mm_lo; ++ u8 width_height_mm_hi; ++ u8 hborder; ++ u8 vborder; ++ u8 misc; ++}; ++ ++struct detailed_data_string { ++ u8 str[13]; ++}; ++ ++struct detailed_data_monitor_range { ++ u8 min_vfreq; ++ u8 max_vfreq; ++ u8 min_hfreq_khz; ++ u8 max_hfreq_khz; ++ u8 pixel_clock_mhz; ++ u8 flags; ++ union { ++ struct { ++ u8 reserved; ++ u8 hfreq_start_khz; ++ u8 c; ++ __le16 m; ++ u8 k; ++ u8 j; ++ } __attribute__((packed)) gtf2; ++ struct { ++ u8 version; ++ u8 data1; ++ u8 data2; ++ u8 supported_aspects; ++ u8 flags; ++ u8 supported_scalings; ++ u8 preferred_refresh; ++ } cvt; ++ } formula; ++}; ++ ++struct detailed_data_wpindex { ++ u8 white_yx_lo; ++ u8 white_x_hi; ++ u8 white_y_hi; ++ u8 gamma; ++}; ++ ++struct cvt_timing { ++ u8 code[3]; ++}; ++ ++struct detailed_non_pixel { ++ u8 pad1; ++ u8 type; ++ u8 pad2; ++ union { ++ struct detailed_data_string str; ++ struct detailed_data_monitor_range range; ++ struct detailed_data_wpindex color; ++ struct std_timing timings[6]; ++ struct cvt_timing cvt[4]; ++ } data; ++}; ++ ++struct detailed_timing { ++ __le16 pixel_clock; ++ union { ++ struct detailed_pixel_timing pixel_data; ++ struct detailed_non_pixel other_data; ++ } data; ++}; ++ ++struct edid { ++ u8 header[8]; ++ u8 mfg_id[2]; ++ u8 prod_code[2]; ++ u32 serial; ++ u8 mfg_week; ++ u8 mfg_year; ++ u8 version; ++ u8 revision; ++ u8 input; ++ u8 width_cm; ++ u8 height_cm; ++ u8 gamma; ++ u8 features; ++ u8 red_green_lo; ++ u8 blue_white_lo; ++ u8 red_x; ++ u8 red_y; ++ u8 green_x; ++ u8 green_y; ++ u8 blue_x; ++ u8 blue_y; ++ u8 white_x; ++ u8 white_y; ++ struct est_timings established_timings; ++ struct std_timing standard_timings[8]; ++ struct detailed_timing detailed_timings[4]; ++ u8 extensions; ++ u8 checksum; ++}; ++ ++struct drm_bridge_connector { ++ struct drm_connector base; ++ struct drm_encoder *encoder; ++ struct drm_bridge *bridge_edid; ++ struct drm_bridge *bridge_hpd; ++ struct drm_bridge *bridge_detect; ++ struct drm_bridge *bridge_modes; ++}; ++ ++struct drm_flip_work; ++ ++typedef void (*drm_flip_func_t)(struct drm_flip_work *, void *); ++ ++struct drm_flip_work { ++ const char *name; ++ drm_flip_func_t func; ++ struct work_struct worker; ++ struct list_head queued; ++ struct list_head commited; ++ spinlock_t lock; ++}; ++ ++struct drm_flip_task { ++ struct list_head node; ++ void *data; ++}; ++ ++struct drm_client_funcs; ++ ++struct drm_client_dev { ++ struct drm_device *dev; ++ const char *name; ++ struct list_head list; ++ const struct drm_client_funcs *funcs; ++ struct drm_file *file; ++ struct mutex modeset_mutex; ++ struct drm_mode_set *modesets; ++ bool hotplug_failed; ++}; ++ ++struct drm_client_buffer; ++ ++struct drm_fb_helper_funcs; ++ ++struct drm_fb_helper { ++ struct drm_client_dev client; ++ struct drm_client_buffer *buffer; ++ struct drm_framebuffer *fb; ++ struct drm_device *dev; ++ const struct drm_fb_helper_funcs *funcs; ++ struct fb_info *info; ++ u32 pseudo_palette[17]; ++ struct drm_clip_rect damage_clip; ++ spinlock_t damage_lock; ++ struct work_struct damage_work; ++ struct work_struct resume_work; ++ struct mutex lock; ++ struct list_head kernel_fb_list; ++ bool delayed_hotplug; ++ bool deferred_setup; ++ int preferred_bpp; ++ struct fb_deferred_io fbdefio; ++}; ++ ++struct drm_client_funcs { ++ struct module *owner; ++ void (*unregister)(struct drm_client_dev *); ++ int (*restore)(struct drm_client_dev *); ++ int (*hotplug)(struct drm_client_dev *); ++}; ++ ++struct drm_client_buffer { ++ struct drm_client_dev *client; ++ u32 pitch; ++ struct drm_gem_object *gem; ++ struct iosys_map map; ++ struct drm_framebuffer *fb; ++}; ++ ++struct drm_fb_helper_surface_size { ++ u32 fb_width; ++ u32 fb_height; ++ u32 surface_width; ++ u32 surface_height; ++ u32 surface_bpp; ++ u32 surface_depth; ++}; ++ ++struct drm_fb_helper_funcs { ++ int (*fb_probe)(struct drm_fb_helper *, struct drm_fb_helper_surface_size *); ++ int (*fb_dirty)(struct drm_fb_helper *, struct drm_clip_rect *); ++}; ++ ++struct scm_creds { ++ u32 pid; ++ kuid_t uid; ++ kgid_t gid; ++}; ++ ++struct netlink_skb_parms { ++ struct scm_creds creds; ++ __u32 portid; ++ __u32 dst_group; ++ __u32 flags; ++ struct sock *sk; ++ bool nsid_is_set; ++ int nsid; ++}; ++ ++struct netlink_kernel_cfg { ++ unsigned int groups; ++ unsigned int flags; ++ void (*input)(struct sk_buff *); ++ struct mutex *cb_mutex; ++ int (*bind)(struct net *, int); ++ void (*unbind)(struct net *, int); ++}; ++ ++struct cb_id { ++ __u32 idx; ++ __u32 val; ++}; ++ ++struct cn_msg { ++ struct cb_id id; ++ __u32 seq; ++ __u32 ack; ++ __u16 len; ++ __u16 flags; ++ __u8 data[0]; ++}; ++ ++struct cn_queue_dev { ++ atomic_t refcnt; ++ unsigned char name[32]; ++ struct list_head queue_list; ++ spinlock_t queue_lock; ++ struct sock *nls; ++}; ++ ++struct cn_callback_id { ++ unsigned char name[32]; ++ struct cb_id id; ++}; ++ ++struct cn_callback_entry { ++ struct list_head callback_entry; ++ refcount_t refcnt; ++ struct cn_queue_dev *pdev; ++ struct cn_callback_id id; ++ void (*callback)(struct cn_msg *, struct netlink_skb_parms *); ++ u32 seq; ++ u32 group; ++}; ++ ++struct cn_dev { ++ struct cb_id id; ++ u32 seq; ++ u32 groups; ++ struct sock *nls; ++ struct cn_queue_dev *cbdev; ++}; ++ ++enum bus_notifier_event { ++ BUS_NOTIFY_ADD_DEVICE = 0, ++ BUS_NOTIFY_DEL_DEVICE = 1, ++ BUS_NOTIFY_REMOVED_DEVICE = 2, ++ BUS_NOTIFY_BIND_DRIVER = 3, ++ BUS_NOTIFY_BOUND_DRIVER = 4, ++ BUS_NOTIFY_UNBIND_DRIVER = 5, ++ BUS_NOTIFY_UNBOUND_DRIVER = 6, ++ BUS_NOTIFY_DRIVER_NOT_BOUND = 7, ++}; ++ ++struct driver_private { ++ struct kobject kobj; ++ struct klist klist_devices; ++ struct klist_node knode_bus; ++ struct module_kobject *mkobj; ++ struct device_driver *driver; ++}; ++ ++struct device_private { ++ struct klist klist_children; ++ struct klist_node knode_parent; ++ struct klist_node knode_driver; ++ struct klist_node knode_bus; ++ struct klist_node knode_class; ++ struct list_head deferred_probe; ++ struct device_driver *async_driver; ++ char *deferred_probe_reason; ++ struct device *device; ++ u8 dead: 1; ++}; ++ ++struct device_attach_data { ++ struct device *dev; ++ bool check_async; ++ bool want_async; ++ bool have_async; ++}; ++ ++struct subsys_private { ++ struct kset subsys; ++ struct kset *devices_kset; ++ struct list_head interfaces; ++ struct mutex mutex; ++ struct kset *drivers_kset; ++ struct klist klist_devices; ++ struct klist klist_drivers; ++ struct blocking_notifier_head bus_notifier; ++ unsigned int drivers_autoprobe: 1; ++ const struct bus_type *bus; ++ struct device *dev_root; ++ struct kset glue_dirs; ++ const struct class *class; ++ struct lock_class_key lock_key; ++}; ++ ++struct class_attribute { ++ struct attribute attr; ++ ssize_t (*show)(const struct class *, const struct class_attribute *, char *); ++ ssize_t (*store)(const struct class *, const struct class_attribute *, const char *, size_t); ++}; ++ ++struct class_attribute_string { ++ struct class_attribute attr; ++ char *str; ++}; ++ ++struct class_interface { ++ struct list_head node; ++ const struct class *class; ++ int (*add_dev)(struct device *); ++ void (*remove_dev)(struct device *); ++}; ++ ++struct class_compat { ++ struct kobject *kobj; ++}; ++ ++enum { ++ S390_CPU_FEATURE_MSA = 0, ++ S390_CPU_FEATURE_VXRS = 1, ++ S390_CPU_FEATURE_UV = 2, ++ MAX_CPU_FEATURES = 3, ++}; ++ ++typedef u32 note_buf_t[92]; ++ ++struct cpu_attr { ++ struct device_attribute attr; ++ const struct cpumask * const map; ++}; ++ ++struct attribute_container { ++ struct list_head node; ++ struct klist containers; ++ struct class *class; ++ const struct attribute_group *grp; ++ struct device_attribute **attrs; ++ int (*match)(struct attribute_container *, struct device *); ++ long unsigned int flags; ++}; ++ ++struct internal_container { ++ struct klist_node node; ++ struct attribute_container *cont; ++ struct device classdev; ++}; ++ ++enum cache_type { ++ CACHE_TYPE_NOCACHE = 0, ++ CACHE_TYPE_INST = 1, ++ CACHE_TYPE_DATA = 2, ++ CACHE_TYPE_SEPARATE = 3, ++ CACHE_TYPE_UNIFIED = 4, ++}; ++ ++struct cacheinfo { ++ unsigned int id; ++ enum cache_type type; ++ unsigned int level; ++ unsigned int coherency_line_size; ++ unsigned int number_of_sets; ++ unsigned int ways_of_associativity; ++ unsigned int physical_line_partition; ++ unsigned int size; ++ cpumask_t shared_cpu_map; ++ unsigned int attributes; ++ void *fw_token; ++ bool disable_sysfs; ++ void *priv; ++}; ++ ++struct cpu_cacheinfo { ++ struct cacheinfo *info_list; ++ unsigned int num_levels; ++ unsigned int num_leaves; ++ bool cpu_map_populated; ++ bool early_ci_levels; ++}; ++ ++struct req { ++ struct req *next; ++ struct completion done; ++ int err; ++ const char *name; ++ umode_t mode; ++ kuid_t uid; ++ kgid_t gid; ++ struct device *dev; ++}; ++ ++struct firmware { ++ size_t size; ++ const u8 *data; ++ void *priv; ++}; ++ ++enum fw_opt { ++ FW_OPT_UEVENT = 1, ++ FW_OPT_NOWAIT = 2, ++ FW_OPT_USERHELPER = 4, ++ FW_OPT_NO_WARN = 8, ++ FW_OPT_NOCACHE = 16, ++ FW_OPT_NOFALLBACK_SYSFS = 32, ++ FW_OPT_FALLBACK_PLATFORM = 64, ++ FW_OPT_PARTIAL = 128, ++}; ++ ++enum fw_status { ++ FW_STATUS_UNKNOWN = 0, ++ FW_STATUS_LOADING = 1, ++ FW_STATUS_DONE = 2, ++ FW_STATUS_ABORTED = 3, ++}; ++ ++struct fw_state { ++ struct completion completion; ++ enum fw_status status; ++}; ++ ++struct firmware_cache; ++ ++struct fw_priv { ++ struct kref ref; ++ struct list_head list; ++ struct firmware_cache *fwc; ++ struct fw_state fw_st; ++ void *data; ++ size_t size; ++ size_t allocated_size; ++ size_t offset; ++ u32 opt_flags; ++ bool is_paged_buf; ++ struct page **pages; ++ int nr_pages; ++ int page_array_size; ++ bool need_uevent; ++ struct list_head pending_list; ++ const char *fw_name; ++}; ++ ++struct firmware_fallback_config { ++ unsigned int force_sysfs_fallback; ++ unsigned int ignore_sysfs_fallback; ++ int old_timeout; ++ int loading_timeout; ++}; ++ ++struct fw_sysfs { ++ bool nowait; ++ struct device dev; ++ struct fw_priv *fw_priv; ++ struct firmware *fw; ++ void *fw_upload_priv; ++}; ++ ++struct builtin_fw { ++ char *name; ++ void *data; ++ long unsigned int size; ++}; ++ ++enum { ++ MMOP_OFFLINE = 0, ++ MMOP_ONLINE = 1, ++ MMOP_ONLINE_KERNEL = 2, ++ MMOP_ONLINE_MOVABLE = 3, ++}; ++ ++enum meminit_context { ++ MEMINIT_EARLY = 0, ++ MEMINIT_HOTPLUG = 1, ++}; ++ ++struct xa_limit { ++ u32 max; ++ u32 min; ++}; ++ ++struct memory_group { ++ int nid; ++ struct list_head memory_blocks; ++ long unsigned int present_kernel_pages; ++ long unsigned int present_movable_pages; ++ bool is_dynamic; ++ union { ++ struct { ++ long unsigned int max_pages; ++ } s; ++ struct { ++ long unsigned int unit_pages; ++ } d; ++ }; ++}; ++ ++struct memory_block { ++ long unsigned int start_section_nr; ++ long unsigned int state; ++ int online_type; ++ int nid; ++ struct zone *zone; ++ struct device dev; ++ long unsigned int nr_vmemmap_pages; ++ struct memory_group *group; ++ struct list_head group_next; ++}; ++ ++typedef int (*walk_memory_blocks_func_t)(struct memory_block *, void *); ++ ++typedef int (*walk_memory_groups_func_t)(struct memory_group *, void *); ++ ++struct for_each_memory_block_cb_data { ++ walk_memory_blocks_func_t func; ++ void *arg; ++}; ++ ++enum regcache_type { ++ REGCACHE_NONE = 0, ++ REGCACHE_RBTREE = 1, ++ REGCACHE_FLAT = 2, ++ REGCACHE_MAPLE = 3, ++}; ++ ++struct reg_default { ++ unsigned int reg; ++ unsigned int def; ++}; ++ ++struct reg_sequence { ++ unsigned int reg; ++ unsigned int def; ++ unsigned int delay_us; ++}; ++ ++enum regmap_endian { ++ REGMAP_ENDIAN_DEFAULT = 0, ++ REGMAP_ENDIAN_BIG = 1, ++ REGMAP_ENDIAN_LITTLE = 2, ++ REGMAP_ENDIAN_NATIVE = 3, ++}; ++ ++struct regmap_range { ++ unsigned int range_min; ++ unsigned int range_max; ++}; ++ ++struct regmap_access_table { ++ const struct regmap_range *yes_ranges; ++ unsigned int n_yes_ranges; ++ const struct regmap_range *no_ranges; ++ unsigned int n_no_ranges; ++}; ++ ++typedef void (*regmap_lock)(void *); ++ ++typedef void (*regmap_unlock)(void *); ++ ++struct regmap_range_cfg; ++ ++struct regmap_config { ++ const char *name; ++ int reg_bits; ++ int reg_stride; ++ int reg_shift; ++ unsigned int reg_base; ++ int pad_bits; ++ int val_bits; ++ bool (*writeable_reg)(struct device *, unsigned int); ++ bool (*readable_reg)(struct device *, unsigned int); ++ bool (*volatile_reg)(struct device *, unsigned int); ++ bool (*precious_reg)(struct device *, unsigned int); ++ bool (*writeable_noinc_reg)(struct device *, unsigned int); ++ bool (*readable_noinc_reg)(struct device *, unsigned int); ++ bool disable_locking; ++ regmap_lock lock; ++ regmap_unlock unlock; ++ void *lock_arg; ++ int (*reg_read)(void *, unsigned int, unsigned int *); ++ int (*reg_write)(void *, unsigned int, unsigned int); ++ int (*reg_update_bits)(void *, unsigned int, unsigned int, unsigned int); ++ int (*read)(void *, const void *, size_t, void *, size_t); ++ int (*write)(void *, const void *, size_t); ++ size_t max_raw_read; ++ size_t max_raw_write; ++ bool fast_io; ++ bool io_port; ++ unsigned int max_register; ++ const struct regmap_access_table *wr_table; ++ const struct regmap_access_table *rd_table; ++ const struct regmap_access_table *volatile_table; ++ const struct regmap_access_table *precious_table; ++ const struct regmap_access_table *wr_noinc_table; ++ const struct regmap_access_table *rd_noinc_table; ++ const struct reg_default *reg_defaults; ++ unsigned int num_reg_defaults; ++ enum regcache_type cache_type; ++ const void *reg_defaults_raw; ++ unsigned int num_reg_defaults_raw; ++ long unsigned int read_flag_mask; ++ long unsigned int write_flag_mask; ++ bool zero_flag_mask; ++ bool use_single_read; ++ bool use_single_write; ++ bool use_relaxed_mmio; ++ bool can_multi_write; ++ enum regmap_endian reg_format_endian; ++ enum regmap_endian val_format_endian; ++ const struct regmap_range_cfg *ranges; ++ unsigned int num_ranges; ++ bool use_hwlock; ++ bool use_raw_spinlock; ++ unsigned int hwlock_id; ++ unsigned int hwlock_mode; ++ bool can_sleep; ++}; ++ ++struct regmap_range_cfg { ++ const char *name; ++ unsigned int range_min; ++ unsigned int range_max; ++ unsigned int selector_reg; ++ unsigned int selector_mask; ++ int selector_shift; ++ unsigned int window_start; ++ unsigned int window_len; ++}; ++ ++typedef int (*regmap_hw_write)(void *, const void *, size_t); ++ ++typedef int (*regmap_hw_gather_write)(void *, const void *, size_t, const void *, size_t); ++ ++struct regmap_async; ++ ++typedef int (*regmap_hw_async_write)(void *, const void *, size_t, const void *, size_t, struct regmap_async *); ++ ++struct regmap_async { ++ struct list_head list; ++ struct regmap *map; ++ void *work_buf; ++}; ++ ++typedef int (*regmap_hw_read)(void *, const void *, size_t, void *, size_t); ++ ++typedef int (*regmap_hw_reg_read)(void *, unsigned int, unsigned int *); ++ ++typedef int (*regmap_hw_reg_noinc_read)(void *, unsigned int, void *, size_t); ++ ++typedef int (*regmap_hw_reg_write)(void *, unsigned int, unsigned int); ++ ++typedef int (*regmap_hw_reg_noinc_write)(void *, unsigned int, const void *, size_t); ++ ++typedef int (*regmap_hw_reg_update_bits)(void *, unsigned int, unsigned int, unsigned int); ++ ++typedef struct regmap_async * (*regmap_hw_async_alloc)(); ++ ++typedef void (*regmap_hw_free_context)(void *); ++ ++struct regmap_bus { ++ bool fast_io; ++ bool free_on_exit; ++ regmap_hw_write write; ++ regmap_hw_gather_write gather_write; ++ regmap_hw_async_write async_write; ++ regmap_hw_reg_write reg_write; ++ regmap_hw_reg_noinc_write reg_noinc_write; ++ regmap_hw_reg_update_bits reg_update_bits; ++ regmap_hw_read read; ++ regmap_hw_reg_read reg_read; ++ regmap_hw_reg_noinc_read reg_noinc_read; ++ regmap_hw_free_context free_context; ++ regmap_hw_async_alloc async_alloc; ++ u8 read_flag_mask; ++ enum regmap_endian reg_format_endian_default; ++ enum regmap_endian val_format_endian_default; ++ size_t max_raw_read; ++ size_t max_raw_write; ++}; ++ ++struct reg_field { ++ unsigned int reg; ++ unsigned int lsb; ++ unsigned int msb; ++ unsigned int id_size; ++ unsigned int id_offset; ++}; ++ ++struct regmap_format { ++ size_t buf_size; ++ size_t reg_bytes; ++ size_t pad_bytes; ++ size_t val_bytes; ++ s8 reg_shift; ++ void (*format_write)(struct regmap *, unsigned int, unsigned int); ++ void (*format_reg)(void *, unsigned int, unsigned int); ++ void (*format_val)(void *, unsigned int, unsigned int); ++ unsigned int (*parse_val)(const void *); ++ void (*parse_inplace)(void *); ++}; ++ ++struct regcache_ops; ++ ++struct hwspinlock; ++ ++struct regmap { ++ union { ++ struct mutex mutex; ++ struct { ++ spinlock_t spinlock; ++ long unsigned int spinlock_flags; ++ }; ++ struct { ++ raw_spinlock_t raw_spinlock; ++ long unsigned int raw_spinlock_flags; ++ }; ++ }; ++ regmap_lock lock; ++ regmap_unlock unlock; ++ void *lock_arg; ++ gfp_t alloc_flags; ++ unsigned int reg_base; ++ struct device *dev; ++ void *work_buf; ++ struct regmap_format format; ++ const struct regmap_bus *bus; ++ void *bus_context; ++ const char *name; ++ bool async; ++ spinlock_t async_lock; ++ wait_queue_head_t async_waitq; ++ struct list_head async_list; ++ struct list_head async_free; ++ int async_ret; ++ bool debugfs_disable; ++ struct dentry *debugfs; ++ const char *debugfs_name; ++ unsigned int debugfs_reg_len; ++ unsigned int debugfs_val_len; ++ unsigned int debugfs_tot_len; ++ struct list_head debugfs_off_cache; ++ struct mutex cache_lock; ++ unsigned int max_register; ++ bool (*writeable_reg)(struct device *, unsigned int); ++ bool (*readable_reg)(struct device *, unsigned int); ++ bool (*volatile_reg)(struct device *, unsigned int); ++ bool (*precious_reg)(struct device *, unsigned int); ++ bool (*writeable_noinc_reg)(struct device *, unsigned int); ++ bool (*readable_noinc_reg)(struct device *, unsigned int); ++ const struct regmap_access_table *wr_table; ++ const struct regmap_access_table *rd_table; ++ const struct regmap_access_table *volatile_table; ++ const struct regmap_access_table *precious_table; ++ const struct regmap_access_table *wr_noinc_table; ++ const struct regmap_access_table *rd_noinc_table; ++ int (*reg_read)(void *, unsigned int, unsigned int *); ++ int (*reg_write)(void *, unsigned int, unsigned int); ++ int (*reg_update_bits)(void *, unsigned int, unsigned int, unsigned int); ++ int (*read)(void *, const void *, size_t, void *, size_t); ++ int (*write)(void *, const void *, size_t); ++ bool defer_caching; ++ long unsigned int read_flag_mask; ++ long unsigned int write_flag_mask; ++ int reg_shift; ++ int reg_stride; ++ int reg_stride_order; ++ const struct regcache_ops *cache_ops; ++ enum regcache_type cache_type; ++ unsigned int cache_size_raw; ++ unsigned int cache_word_size; ++ unsigned int num_reg_defaults; ++ unsigned int num_reg_defaults_raw; ++ bool cache_only; ++ bool cache_bypass; ++ bool cache_free; ++ struct reg_default *reg_defaults; ++ const void *reg_defaults_raw; ++ void *cache; ++ bool cache_dirty; ++ bool no_sync_defaults; ++ struct reg_sequence *patch; ++ int patch_regs; ++ bool use_single_read; ++ bool use_single_write; ++ bool can_multi_write; ++ size_t max_raw_read; ++ size_t max_raw_write; ++ struct rb_root range_tree; ++ void *selector_work_buf; ++ struct hwspinlock *hwlock; ++ bool can_sleep; ++}; ++ ++struct regcache_ops { ++ const char *name; ++ enum regcache_type type; ++ int (*init)(struct regmap *); ++ int (*exit)(struct regmap *); ++ void (*debugfs_init)(struct regmap *); ++ int (*read)(struct regmap *, unsigned int, unsigned int *); ++ int (*write)(struct regmap *, unsigned int, unsigned int); ++ int (*sync)(struct regmap *, unsigned int, unsigned int); ++ int (*drop)(struct regmap *, unsigned int, unsigned int); ++}; ++ ++struct regmap_range_node { ++ struct rb_node node; ++ const char *name; ++ struct regmap *map; ++ unsigned int range_min; ++ unsigned int range_max; ++ unsigned int selector_reg; ++ unsigned int selector_mask; ++ int selector_shift; ++ unsigned int window_start; ++ unsigned int window_len; ++}; ++ ++struct regmap_field { ++ struct regmap *regmap; ++ unsigned int mask; ++ unsigned int shift; ++ unsigned int reg; ++ unsigned int id_size; ++ unsigned int id_offset; ++}; ++ ++struct trace_event_raw_regmap_reg { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ unsigned int reg; ++ unsigned int val; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_regmap_bulk { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ unsigned int reg; ++ u32 __data_loc_buf; ++ int val_len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_regmap_block { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ unsigned int reg; ++ int count; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_regcache_sync { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ u32 __data_loc_status; ++ u32 __data_loc_type; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_regmap_bool { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ int flag; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_regmap_async { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_regcache_drop_region { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ unsigned int from; ++ unsigned int to; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_regmap_reg { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_regmap_bulk { ++ u32 name; ++ u32 buf; ++}; ++ ++struct trace_event_data_offsets_regmap_block { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_regcache_sync { ++ u32 name; ++ u32 status; ++ u32 type; ++}; ++ ++struct trace_event_data_offsets_regmap_bool { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_regmap_async { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_regcache_drop_region { ++ u32 name; ++}; ++ ++typedef void (*btf_trace_regmap_reg_write)(void *, struct regmap *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_regmap_reg_read)(void *, struct regmap *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_regmap_reg_read_cache)(void *, struct regmap *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_regmap_bulk_write)(void *, struct regmap *, unsigned int, const void *, int); ++ ++typedef void (*btf_trace_regmap_bulk_read)(void *, struct regmap *, unsigned int, const void *, int); ++ ++typedef void (*btf_trace_regmap_hw_read_start)(void *, struct regmap *, unsigned int, int); ++ ++typedef void (*btf_trace_regmap_hw_read_done)(void *, struct regmap *, unsigned int, int); ++ ++typedef void (*btf_trace_regmap_hw_write_start)(void *, struct regmap *, unsigned int, int); ++ ++typedef void (*btf_trace_regmap_hw_write_done)(void *, struct regmap *, unsigned int, int); ++ ++typedef void (*btf_trace_regcache_sync)(void *, struct regmap *, const char *, const char *); ++ ++typedef void (*btf_trace_regmap_cache_only)(void *, struct regmap *, bool); ++ ++typedef void (*btf_trace_regmap_cache_bypass)(void *, struct regmap *, bool); ++ ++typedef void (*btf_trace_regmap_async_write_start)(void *, struct regmap *, unsigned int, int); ++ ++typedef void (*btf_trace_regmap_async_io_complete)(void *, struct regmap *); ++ ++typedef void (*btf_trace_regmap_async_complete_start)(void *, struct regmap *); ++ ++typedef void (*btf_trace_regmap_async_complete_done)(void *, struct regmap *); ++ ++typedef void (*btf_trace_regcache_drop_region)(void *, struct regmap *, unsigned int, unsigned int); ++ ++struct regmap_debugfs_off_cache { ++ struct list_head list; ++ off_t min; ++ off_t max; ++ unsigned int base_reg; ++ unsigned int max_reg; ++}; ++ ++struct regmap_debugfs_node { ++ struct regmap *map; ++ struct list_head link; ++}; ++ ++struct soc_device_attribute { ++ const char *machine; ++ const char *family; ++ const char *revision; ++ const char *serial_number; ++ const char *soc_id; ++ const void *data; ++ const struct attribute_group *custom_attr_group; ++}; ++ ++struct soc_device { ++ struct device dev; ++ struct soc_device_attribute *attr; ++ int soc_dev_num; ++}; ++ ++struct trace_event_raw_devres { ++ struct trace_entry ent; ++ u32 __data_loc_devname; ++ struct device *dev; ++ const char *op; ++ void *node; ++ const char *name; ++ size_t size; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_devres { ++ u32 devname; ++}; ++ ++typedef void (*btf_trace_devres_log)(void *, struct device *, const char *, void *, const char *, size_t); ++ ++struct driver_attribute { ++ struct attribute attr; ++ ssize_t (*show)(struct device_driver *, char *); ++ ssize_t (*store)(struct device_driver *, const char *, size_t); ++}; ++ ++struct dax_operations { ++ long int (*direct_access)(struct dax_device *, long unsigned int, long int, enum dax_access_mode, void **, pfn_t *); ++ bool (*dax_supported)(struct dax_device *, struct block_device *, int, sector_t, sector_t); ++ int (*zero_page_range)(struct dax_device *, long unsigned int, size_t); ++ size_t (*recovery_write)(struct dax_device *, long unsigned int, void *, size_t, struct iov_iter *); ++}; ++ ++struct dax_region { ++ int id; ++ int target_node; ++ struct kref kref; ++ struct device *dev; ++ unsigned int align; ++ struct ida ida; ++ struct resource res; ++ struct device *seed; ++ struct device *youngest; ++}; ++ ++struct dax_mapping { ++ struct device dev; ++ int range_id; ++ int id; ++}; ++ ++struct dev_dax_range { ++ long unsigned int pgoff; ++ struct range range; ++ struct dax_mapping *mapping; ++}; ++ ++struct dev_dax { ++ struct dax_region *region; ++ struct dax_device *dax_dev; ++ unsigned int align; ++ int target_node; ++ bool dyn_id; ++ int id; ++ struct ida ida; ++ struct device dev; ++ struct dev_pagemap *pgmap; ++ int nr_range; ++ struct dev_dax_range *ranges; ++}; ++ ++struct dev_dax_data { ++ struct dax_region *dax_region; ++ struct dev_pagemap *pgmap; ++ resource_size_t size; ++ int id; ++}; ++ ++enum dax_driver_type { ++ DAXDRV_KMEM_TYPE = 0, ++ DAXDRV_DEVICE_TYPE = 1, ++}; ++ ++struct dax_device_driver { ++ struct device_driver drv; ++ struct list_head ids; ++ enum dax_driver_type type; ++ int (*probe)(struct dev_dax *); ++ void (*remove)(struct dev_dax *); ++}; ++ ++struct dax_id { ++ struct list_head list; ++ char dev_name[30]; ++}; ++ ++enum id_action { ++ ID_REMOVE = 0, ++ ID_ADD = 1, ++}; ++ ++struct dma_heap; ++ ++struct dma_heap_ops { ++ struct dma_buf * (*allocate)(struct dma_heap *, long unsigned int, long unsigned int, long unsigned int); ++}; ++ ++struct dma_heap { ++ const char *name; ++ const struct dma_heap_ops *ops; ++ void *priv; ++ dev_t heap_devt; ++ struct list_head list; ++ struct cdev heap_cdev; ++}; ++ ++struct dma_heap_export_info { ++ const char *name; ++ const struct dma_heap_ops *ops; ++ void *priv; ++}; ++ ++struct dma_heap_allocation_data { ++ __u64 len; ++ __u32 fd; ++ __u32 fd_flags; ++ __u64 heap_flags; ++}; ++ ++enum sam_status { ++ SAM_STAT_GOOD = 0, ++ SAM_STAT_CHECK_CONDITION = 2, ++ SAM_STAT_CONDITION_MET = 4, ++ SAM_STAT_BUSY = 8, ++ SAM_STAT_INTERMEDIATE = 16, ++ SAM_STAT_INTERMEDIATE_CONDITION_MET = 20, ++ SAM_STAT_RESERVATION_CONFLICT = 24, ++ SAM_STAT_COMMAND_TERMINATED = 34, ++ SAM_STAT_TASK_SET_FULL = 40, ++ SAM_STAT_ACA_ACTIVE = 48, ++ SAM_STAT_TASK_ABORTED = 64, ++}; ++ ++struct scsi_sense_hdr { ++ u8 response_code; ++ u8 sense_key; ++ u8 asc; ++ u8 ascq; ++ u8 byte4; ++ u8 byte5; ++ u8 byte6; ++ u8 additional_length; ++}; ++ ++enum scsi_host_status { ++ DID_OK = 0, ++ DID_NO_CONNECT = 1, ++ DID_BUS_BUSY = 2, ++ DID_TIME_OUT = 3, ++ DID_BAD_TARGET = 4, ++ DID_ABORT = 5, ++ DID_PARITY = 6, ++ DID_ERROR = 7, ++ DID_RESET = 8, ++ DID_BAD_INTR = 9, ++ DID_PASSTHROUGH = 10, ++ DID_SOFT_ERROR = 11, ++ DID_IMM_RETRY = 12, ++ DID_REQUEUE = 13, ++ DID_TRANSPORT_DISRUPTED = 14, ++ DID_TRANSPORT_FAILFAST = 15, ++ DID_TRANSPORT_MARGINAL = 20, ++}; ++ ++enum scsi_disposition { ++ NEEDS_RETRY = 8193, ++ SUCCESS = 8194, ++ FAILED = 8195, ++ QUEUED = 8196, ++ SOFT_ERROR = 8197, ++ ADD_TO_MLQUEUE = 8198, ++ TIMEOUT_ERROR = 8199, ++ SCSI_RETURN_NOT_HANDLED = 8200, ++ FAST_IO_FAIL = 8201, ++}; ++ ++typedef __u64 blist_flags_t; ++ ++enum scsi_device_state { ++ SDEV_CREATED = 1, ++ SDEV_RUNNING = 2, ++ SDEV_CANCEL = 3, ++ SDEV_DEL = 4, ++ SDEV_QUIESCE = 5, ++ SDEV_OFFLINE = 6, ++ SDEV_TRANSPORT_OFFLINE = 7, ++ SDEV_BLOCK = 8, ++ SDEV_CREATED_BLOCK = 9, ++}; ++ ++enum scsi_device_event { ++ SDEV_EVT_MEDIA_CHANGE = 1, ++ SDEV_EVT_INQUIRY_CHANGE_REPORTED = 2, ++ SDEV_EVT_CAPACITY_CHANGE_REPORTED = 3, ++ SDEV_EVT_SOFT_THRESHOLD_REACHED_REPORTED = 4, ++ SDEV_EVT_MODE_PARAMETER_CHANGE_REPORTED = 5, ++ SDEV_EVT_LUN_CHANGE_REPORTED = 6, ++ SDEV_EVT_ALUA_STATE_CHANGE_REPORTED = 7, ++ SDEV_EVT_POWER_ON_RESET_OCCURRED = 8, ++ SDEV_EVT_FIRST = 1, ++ SDEV_EVT_LAST = 8, ++ SDEV_EVT_MAXBITS = 9, ++}; ++ ++struct scsi_vpd { ++ struct callback_head rcu; ++ int len; ++ unsigned char data[0]; ++}; ++ ++enum scsi_vpd_parameters { ++ SCSI_VPD_HEADER_SIZE = 4, ++}; ++ ++struct Scsi_Host; ++ ++struct scsi_target; ++ ++struct scsi_device_handler; ++ ++struct bsg_device; ++ ++struct scsi_device { ++ struct Scsi_Host *host; ++ struct request_queue *request_queue; ++ struct list_head siblings; ++ struct list_head same_target_siblings; ++ struct sbitmap budget_map; ++ atomic_t device_blocked; ++ atomic_t restarts; ++ spinlock_t list_lock; ++ struct list_head starved_entry; ++ short unsigned int queue_depth; ++ short unsigned int max_queue_depth; ++ short unsigned int last_queue_full_depth; ++ short unsigned int last_queue_full_count; ++ long unsigned int last_queue_full_time; ++ long unsigned int queue_ramp_up_period; ++ long unsigned int last_queue_ramp_up; ++ unsigned int id; ++ unsigned int channel; ++ u64 lun; ++ unsigned int manufacturer; ++ unsigned int sector_size; ++ void *hostdata; ++ unsigned char type; ++ char scsi_level; ++ char inq_periph_qual; ++ struct mutex inquiry_mutex; ++ unsigned char inquiry_len; ++ unsigned char *inquiry; ++ const char *vendor; ++ const char *model; ++ const char *rev; ++ struct scsi_vpd *vpd_pg0; ++ struct scsi_vpd *vpd_pg83; ++ struct scsi_vpd *vpd_pg80; ++ struct scsi_vpd *vpd_pg89; ++ struct scsi_vpd *vpd_pgb0; ++ struct scsi_vpd *vpd_pgb1; ++ struct scsi_vpd *vpd_pgb2; ++ struct scsi_target *sdev_target; ++ blist_flags_t sdev_bflags; ++ unsigned int eh_timeout; ++ unsigned int removable: 1; ++ unsigned int changed: 1; ++ unsigned int busy: 1; ++ unsigned int lockable: 1; ++ unsigned int locked: 1; ++ unsigned int borken: 1; ++ unsigned int disconnect: 1; ++ unsigned int soft_reset: 1; ++ unsigned int sdtr: 1; ++ unsigned int wdtr: 1; ++ unsigned int ppr: 1; ++ unsigned int tagged_supported: 1; ++ unsigned int simple_tags: 1; ++ unsigned int was_reset: 1; ++ unsigned int expecting_cc_ua: 1; ++ unsigned int use_10_for_rw: 1; ++ unsigned int use_10_for_ms: 1; ++ unsigned int set_dbd_for_ms: 1; ++ unsigned int no_report_opcodes: 1; ++ unsigned int no_write_same: 1; ++ unsigned int use_16_for_rw: 1; ++ unsigned int use_16_for_sync: 1; ++ unsigned int skip_ms_page_8: 1; ++ unsigned int skip_ms_page_3f: 1; ++ unsigned int skip_vpd_pages: 1; ++ unsigned int try_vpd_pages: 1; ++ unsigned int use_192_bytes_for_3f: 1; ++ unsigned int no_start_on_add: 1; ++ unsigned int allow_restart: 1; ++ unsigned int manage_start_stop: 1; ++ unsigned int start_stop_pwr_cond: 1; ++ unsigned int no_uld_attach: 1; ++ unsigned int select_no_atn: 1; ++ unsigned int fix_capacity: 1; ++ unsigned int guess_capacity: 1; ++ unsigned int retry_hwerror: 1; ++ unsigned int last_sector_bug: 1; ++ unsigned int no_read_disc_info: 1; ++ unsigned int no_read_capacity_16: 1; ++ unsigned int try_rc_10_first: 1; ++ unsigned int security_supported: 1; ++ unsigned int is_visible: 1; ++ unsigned int wce_default_on: 1; ++ unsigned int no_dif: 1; ++ unsigned int broken_fua: 1; ++ unsigned int lun_in_cdb: 1; ++ unsigned int unmap_limit_for_ws: 1; ++ unsigned int rpm_autosuspend: 1; ++ unsigned int ignore_media_change: 1; ++ unsigned int silence_suspend: 1; ++ unsigned int no_vpd_size: 1; ++ unsigned int queue_stopped; ++ bool offline_already; ++ atomic_t disk_events_disable_depth; ++ long unsigned int supported_events[1]; ++ long unsigned int pending_events[1]; ++ struct list_head event_list; ++ struct work_struct event_work; ++ unsigned int max_device_blocked; ++ atomic_t iorequest_cnt; ++ atomic_t iodone_cnt; ++ atomic_t ioerr_cnt; ++ atomic_t iotmo_cnt; ++ struct device sdev_gendev; ++ struct device sdev_dev; ++ struct work_struct requeue_work; ++ struct scsi_device_handler *handler; ++ void *handler_data; ++ size_t dma_drain_len; ++ void *dma_drain_buf; ++ unsigned int sg_timeout; ++ unsigned int sg_reserved_size; ++ struct bsg_device *bsg_dev; ++ unsigned char access_state; ++ struct mutex state_mutex; ++ enum scsi_device_state sdev_state; ++ struct task_struct *quiesced_by; ++ long unsigned int sdev_data[0]; ++}; ++ ++enum scsi_host_state { ++ SHOST_CREATED = 1, ++ SHOST_RUNNING = 2, ++ SHOST_CANCEL = 3, ++ SHOST_DEL = 4, ++ SHOST_RECOVERY = 5, ++ SHOST_CANCEL_RECOVERY = 6, ++ SHOST_DEL_RECOVERY = 7, ++}; ++ ++struct scsi_host_template; ++ ++struct scsi_transport_template; ++ ++struct Scsi_Host { ++ struct list_head __devices; ++ struct list_head __targets; ++ struct list_head starved_list; ++ spinlock_t default_lock; ++ spinlock_t *host_lock; ++ struct mutex scan_mutex; ++ struct list_head eh_abort_list; ++ struct list_head eh_cmd_q; ++ struct task_struct *ehandler; ++ struct completion *eh_action; ++ wait_queue_head_t host_wait; ++ const struct scsi_host_template *hostt; ++ struct scsi_transport_template *transportt; ++ struct kref tagset_refcnt; ++ struct completion tagset_freed; ++ struct blk_mq_tag_set tag_set; ++ atomic_t host_blocked; ++ unsigned int host_failed; ++ unsigned int host_eh_scheduled; ++ unsigned int host_no; ++ int eh_deadline; ++ long unsigned int last_reset; ++ unsigned int max_channel; ++ unsigned int max_id; ++ u64 max_lun; ++ unsigned int unique_id; ++ short unsigned int max_cmd_len; ++ int this_id; ++ int can_queue; ++ short int cmd_per_lun; ++ short unsigned int sg_tablesize; ++ short unsigned int sg_prot_tablesize; ++ unsigned int max_sectors; ++ unsigned int opt_sectors; ++ unsigned int max_segment_size; ++ long unsigned int dma_boundary; ++ long unsigned int virt_boundary_mask; ++ unsigned int nr_hw_queues; ++ unsigned int nr_maps; ++ unsigned int active_mode: 2; ++ unsigned int host_self_blocked: 1; ++ unsigned int reverse_ordering: 1; ++ unsigned int tmf_in_progress: 1; ++ unsigned int async_scan: 1; ++ unsigned int eh_noresume: 1; ++ unsigned int no_write_same: 1; ++ unsigned int host_tagset: 1; ++ unsigned int short_inquiry: 1; ++ unsigned int no_scsi2_lun_in_cdb: 1; ++ char work_q_name[20]; ++ struct workqueue_struct *work_q; ++ struct workqueue_struct *tmf_work_q; ++ unsigned int max_host_blocked; ++ unsigned int prot_capabilities; ++ unsigned char prot_guard_type; ++ long unsigned int base; ++ long unsigned int io_port; ++ unsigned char n_io_port; ++ unsigned char dma_channel; ++ unsigned int irq; ++ enum scsi_host_state shost_state; ++ struct device shost_gendev; ++ struct device shost_dev; ++ void *shost_data; ++ struct device *dma_dev; ++ long unsigned int hostdata[0]; ++}; ++ ++enum scsi_target_state { ++ STARGET_CREATED = 1, ++ STARGET_RUNNING = 2, ++ STARGET_REMOVE = 3, ++ STARGET_CREATED_REMOVE = 4, ++ STARGET_DEL = 5, ++}; ++ ++struct scsi_target { ++ struct scsi_device *starget_sdev_user; ++ struct list_head siblings; ++ struct list_head devices; ++ struct device dev; ++ struct kref reap_ref; ++ unsigned int channel; ++ unsigned int id; ++ unsigned int create: 1; ++ unsigned int single_lun: 1; ++ unsigned int pdt_1f_for_no_lun: 1; ++ unsigned int no_report_luns: 1; ++ unsigned int expecting_lun_change: 1; ++ atomic_t target_busy; ++ atomic_t target_blocked; ++ unsigned int can_queue; ++ unsigned int max_target_blocked; ++ char scsi_level; ++ enum scsi_target_state state; ++ void *hostdata; ++ long unsigned int starget_data[0]; ++}; ++ ++struct scsi_exec_args { ++ unsigned char *sense; ++ unsigned int sense_len; ++ struct scsi_sense_hdr *sshdr; ++ blk_mq_req_flags_t req_flags; ++ int scmd_flags; ++ int *resid; ++}; ++ ++struct scsi_data_buffer { ++ struct sg_table table; ++ unsigned int length; ++}; ++ ++enum scsi_cmnd_submitter { ++ SUBMITTED_BY_BLOCK_LAYER = 0, ++ SUBMITTED_BY_SCSI_ERROR_HANDLER = 1, ++ SUBMITTED_BY_SCSI_RESET_IOCTL = 2, ++} __attribute__((mode(byte))); ++ ++struct scsi_cmnd { ++ struct scsi_device *device; ++ struct list_head eh_entry; ++ struct delayed_work abort_work; ++ struct callback_head rcu; ++ int eh_eflags; ++ int budget_token; ++ long unsigned int jiffies_at_alloc; ++ int retries; ++ int allowed; ++ unsigned char prot_op; ++ unsigned char prot_type; ++ unsigned char prot_flags; ++ enum scsi_cmnd_submitter submitter; ++ short unsigned int cmd_len; ++ enum dma_data_direction sc_data_direction; ++ unsigned char cmnd[32]; ++ struct scsi_data_buffer sdb; ++ struct scsi_data_buffer *prot_sdb; ++ unsigned int underflow; ++ unsigned int transfersize; ++ unsigned int resid_len; ++ unsigned int sense_len; ++ unsigned char *sense_buffer; ++ int flags; ++ long unsigned int state; ++ unsigned int extra_len; ++ unsigned char *host_scribble; ++ int result; ++}; ++ ++enum scsi_prot_operations { ++ SCSI_PROT_NORMAL = 0, ++ SCSI_PROT_READ_INSERT = 1, ++ SCSI_PROT_WRITE_STRIP = 2, ++ SCSI_PROT_READ_STRIP = 3, ++ SCSI_PROT_WRITE_INSERT = 4, ++ SCSI_PROT_READ_PASS = 5, ++ SCSI_PROT_WRITE_PASS = 6, ++}; ++ ++struct scsi_driver { ++ struct device_driver gendrv; ++ void (*rescan)(struct device *); ++ blk_status_t (*init_command)(struct scsi_cmnd *); ++ void (*uninit_command)(struct scsi_cmnd *); ++ int (*done)(struct scsi_cmnd *); ++ int (*eh_action)(struct scsi_cmnd *, int); ++ void (*eh_reset)(struct scsi_cmnd *); ++}; ++ ++enum scsi_timeout_action { ++ SCSI_EH_DONE = 0, ++ SCSI_EH_RESET_TIMER = 1, ++ SCSI_EH_NOT_HANDLED = 2, ++}; ++ ++struct scsi_host_template { ++ unsigned int cmd_size; ++ int (*queuecommand)(struct Scsi_Host *, struct scsi_cmnd *); ++ void (*commit_rqs)(struct Scsi_Host *, u16); ++ struct module *module; ++ const char *name; ++ const char * (*info)(struct Scsi_Host *); ++ int (*ioctl)(struct scsi_device *, unsigned int, void *); ++ int (*compat_ioctl)(struct scsi_device *, unsigned int, void *); ++ int (*init_cmd_priv)(struct Scsi_Host *, struct scsi_cmnd *); ++ int (*exit_cmd_priv)(struct Scsi_Host *, struct scsi_cmnd *); ++ int (*eh_abort_handler)(struct scsi_cmnd *); ++ int (*eh_device_reset_handler)(struct scsi_cmnd *); ++ int (*eh_target_reset_handler)(struct scsi_cmnd *); ++ int (*eh_bus_reset_handler)(struct scsi_cmnd *); ++ int (*eh_host_reset_handler)(struct scsi_cmnd *); ++ int (*slave_alloc)(struct scsi_device *); ++ int (*slave_configure)(struct scsi_device *); ++ void (*slave_destroy)(struct scsi_device *); ++ int (*target_alloc)(struct scsi_target *); ++ void (*target_destroy)(struct scsi_target *); ++ int (*scan_finished)(struct Scsi_Host *, long unsigned int); ++ void (*scan_start)(struct Scsi_Host *); ++ int (*change_queue_depth)(struct scsi_device *, int); ++ void (*map_queues)(struct Scsi_Host *); ++ int (*mq_poll)(struct Scsi_Host *, unsigned int); ++ bool (*dma_need_drain)(struct request *); ++ int (*bios_param)(struct scsi_device *, struct block_device *, sector_t, int *); ++ void (*unlock_native_capacity)(struct scsi_device *); ++ int (*show_info)(struct seq_file *, struct Scsi_Host *); ++ int (*write_info)(struct Scsi_Host *, char *, int); ++ enum scsi_timeout_action (*eh_timed_out)(struct scsi_cmnd *); ++ bool (*eh_should_retry_cmd)(struct scsi_cmnd *); ++ int (*host_reset)(struct Scsi_Host *, int); ++ const char *proc_name; ++ int can_queue; ++ int this_id; ++ short unsigned int sg_tablesize; ++ short unsigned int sg_prot_tablesize; ++ unsigned int max_sectors; ++ unsigned int max_segment_size; ++ long unsigned int dma_boundary; ++ long unsigned int virt_boundary_mask; ++ short int cmd_per_lun; ++ int tag_alloc_policy; ++ unsigned int track_queue_depth: 1; ++ unsigned int supported_mode: 2; ++ unsigned int emulated: 1; ++ unsigned int skip_settle_delay: 1; ++ unsigned int no_write_same: 1; ++ unsigned int host_tagset: 1; ++ unsigned int max_host_blocked; ++ const struct attribute_group **shost_groups; ++ const struct attribute_group **sdev_groups; ++ u64 vendor_id; ++ int rpm_autosuspend_delay; ++}; ++ ++struct trace_event_raw_scsi_dispatch_cmd_start { ++ struct trace_entry ent; ++ unsigned int host_no; ++ unsigned int channel; ++ unsigned int id; ++ unsigned int lun; ++ unsigned int opcode; ++ unsigned int cmd_len; ++ int driver_tag; ++ int scheduler_tag; ++ unsigned int data_sglen; ++ unsigned int prot_sglen; ++ unsigned char prot_op; ++ u32 __data_loc_cmnd; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_scsi_dispatch_cmd_error { ++ struct trace_entry ent; ++ unsigned int host_no; ++ unsigned int channel; ++ unsigned int id; ++ unsigned int lun; ++ int rtn; ++ unsigned int opcode; ++ unsigned int cmd_len; ++ int driver_tag; ++ int scheduler_tag; ++ unsigned int data_sglen; ++ unsigned int prot_sglen; ++ unsigned char prot_op; ++ u32 __data_loc_cmnd; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_scsi_cmd_done_timeout_template { ++ struct trace_entry ent; ++ unsigned int host_no; ++ unsigned int channel; ++ unsigned int id; ++ unsigned int lun; ++ int result; ++ unsigned int opcode; ++ unsigned int cmd_len; ++ int driver_tag; ++ int scheduler_tag; ++ unsigned int data_sglen; ++ unsigned int prot_sglen; ++ unsigned char prot_op; ++ u32 __data_loc_cmnd; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_scsi_eh_wakeup { ++ struct trace_entry ent; ++ unsigned int host_no; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_scsi_dispatch_cmd_start { ++ u32 cmnd; ++}; ++ ++struct trace_event_data_offsets_scsi_dispatch_cmd_error { ++ u32 cmnd; ++}; ++ ++struct trace_event_data_offsets_scsi_cmd_done_timeout_template { ++ u32 cmnd; ++}; ++ ++struct trace_event_data_offsets_scsi_eh_wakeup {}; ++ ++typedef void (*btf_trace_scsi_dispatch_cmd_start)(void *, struct scsi_cmnd *); ++ ++typedef void (*btf_trace_scsi_dispatch_cmd_error)(void *, struct scsi_cmnd *, int); ++ ++typedef void (*btf_trace_scsi_dispatch_cmd_done)(void *, struct scsi_cmnd *); ++ ++typedef void (*btf_trace_scsi_dispatch_cmd_timeout)(void *, struct scsi_cmnd *); ++ ++typedef void (*btf_trace_scsi_eh_wakeup)(void *, struct Scsi_Host *); ++ ++typedef void (*activate_complete)(void *, int); ++ ++struct scsi_device_handler { ++ struct list_head list; ++ struct module *module; ++ const char *name; ++ enum scsi_disposition (*check_sense)(struct scsi_device *, struct scsi_sense_hdr *); ++ int (*attach)(struct scsi_device *); ++ void (*detach)(struct scsi_device *); ++ int (*activate)(struct scsi_device *, activate_complete, void *); ++ blk_status_t (*prep_fn)(struct scsi_device *, struct request *); ++ int (*set_params)(struct scsi_device *, const char *); ++ void (*rescan)(struct scsi_device *); ++}; ++ ++struct scsi_eh_save { ++ int result; ++ unsigned int resid_len; ++ int eh_eflags; ++ enum dma_data_direction data_direction; ++ unsigned int underflow; ++ unsigned char cmd_len; ++ unsigned char prot_op; ++ unsigned char cmnd[32]; ++ struct scsi_data_buffer sdb; ++ struct scatterlist sense_sgl; ++}; ++ ++struct transport_container { ++ struct attribute_container ac; ++ const struct attribute_group *statistics; ++}; ++ ++struct scsi_transport_template { ++ struct transport_container host_attrs; ++ struct transport_container target_attrs; ++ struct transport_container device_attrs; ++ int (*user_scan)(struct Scsi_Host *, uint, uint, u64); ++ int device_size; ++ int device_private_offset; ++ int target_size; ++ int target_private_offset; ++ int host_size; ++ unsigned int create_work_queue: 1; ++ void (*eh_strategy_handler)(struct Scsi_Host *); ++}; ++ ++enum scsi_ml_status { ++ SCSIML_STAT_OK = 0, ++ SCSIML_STAT_RESV_CONFLICT = 1, ++ SCSIML_STAT_NOSPC = 2, ++ SCSIML_STAT_MED_ERROR = 3, ++ SCSIML_STAT_TGT_FAILURE = 4, ++}; ++ ++struct scsi_lun { ++ __u8 scsi_lun[8]; ++}; ++ ++enum scsi_timeouts { ++ SCSI_DEFAULT_EH_TIMEOUT = 1000, ++}; ++ ++enum scsi_scan_mode { ++ SCSI_SCAN_INITIAL = 0, ++ SCSI_SCAN_RESCAN = 1, ++ SCSI_SCAN_MANUAL = 2, ++}; ++ ++struct async_scan_data { ++ struct list_head list; ++ struct Scsi_Host *shost; ++ struct completion prev_finished; ++}; ++ ++struct scsi_nl_hdr { ++ __u8 version; ++ __u8 transport; ++ __u16 magic; ++ __u16 msgtype; ++ __u16 msglen; ++}; ++ ++struct sg_io_v4 { ++ __s32 guard; ++ __u32 protocol; ++ __u32 subprotocol; ++ __u32 request_len; ++ __u64 request; ++ __u64 request_tag; ++ __u32 request_attr; ++ __u32 request_priority; ++ __u32 request_extra; ++ __u32 max_response_len; ++ __u64 response; ++ __u32 dout_iovec_count; ++ __u32 dout_xfer_len; ++ __u32 din_iovec_count; ++ __u32 din_xfer_len; ++ __u64 dout_xferp; ++ __u64 din_xferp; ++ __u32 timeout; ++ __u32 flags; ++ __u64 usr_ptr; ++ __u32 spare_in; ++ __u32 driver_status; ++ __u32 transport_status; ++ __u32 device_status; ++ __u32 retry_delay; ++ __u32 info; ++ __u32 duration; ++ __u32 response_len; ++ __s32 din_resid; ++ __s32 dout_resid; ++ __u64 generated_tag; ++ __u32 spare_out; ++ __u32 padding; ++}; ++ ++typedef int bsg_sg_io_fn(struct request_queue *, struct sg_io_v4 *, fmode_t, unsigned int); ++ ++enum blk_integrity_flags { ++ BLK_INTEGRITY_VERIFY = 1, ++ BLK_INTEGRITY_GENERATE = 2, ++ BLK_INTEGRITY_DEVICE_CAPABLE = 4, ++ BLK_INTEGRITY_IP_CHECKSUM = 8, ++}; ++ ++enum t10_dif_type { ++ T10_PI_TYPE0_PROTECTION = 0, ++ T10_PI_TYPE1_PROTECTION = 1, ++ T10_PI_TYPE2_PROTECTION = 2, ++ T10_PI_TYPE3_PROTECTION = 3, ++}; ++ ++enum scsi_host_prot_capabilities { ++ SHOST_DIF_TYPE1_PROTECTION = 1, ++ SHOST_DIF_TYPE2_PROTECTION = 2, ++ SHOST_DIF_TYPE3_PROTECTION = 4, ++ SHOST_DIX_TYPE0_PROTECTION = 8, ++ SHOST_DIX_TYPE1_PROTECTION = 16, ++ SHOST_DIX_TYPE2_PROTECTION = 32, ++ SHOST_DIX_TYPE3_PROTECTION = 64, ++}; ++ ++enum scsi_host_guard_type { ++ SHOST_DIX_GUARD_CRC = 1, ++ SHOST_DIX_GUARD_IP = 2, ++}; ++ ++struct zoned_disk_info { ++ u32 nr_zones; ++ u32 zone_blocks; ++}; ++ ++struct opal_dev; ++ ++struct scsi_disk { ++ struct scsi_device *device; ++ struct device disk_dev; ++ struct gendisk *disk; ++ struct opal_dev *opal_dev; ++ struct zoned_disk_info early_zone_info; ++ struct zoned_disk_info zone_info; ++ u32 zones_optimal_open; ++ u32 zones_optimal_nonseq; ++ u32 zones_max_open; ++ u32 zone_starting_lba_gran; ++ u32 *zones_wp_offset; ++ spinlock_t zones_wp_offset_lock; ++ u32 *rev_wp_offset; ++ struct mutex rev_mutex; ++ struct work_struct zone_wp_offset_work; ++ char *zone_wp_update_buf; ++ atomic_t openers; ++ sector_t capacity; ++ int max_retries; ++ u32 min_xfer_blocks; ++ u32 max_xfer_blocks; ++ u32 opt_xfer_blocks; ++ u32 max_ws_blocks; ++ u32 max_unmap_blocks; ++ u32 unmap_granularity; ++ u32 unmap_alignment; ++ u32 index; ++ unsigned int physical_block_size; ++ unsigned int max_medium_access_timeouts; ++ unsigned int medium_access_timed_out; ++ u8 media_present; ++ u8 write_prot; ++ u8 protection_type; ++ u8 provisioning_mode; ++ u8 zeroing_mode; ++ u8 nr_actuators; ++ unsigned int ATO: 1; ++ unsigned int cache_override: 1; ++ unsigned int WCE: 1; ++ unsigned int RCD: 1; ++ unsigned int DPOFUA: 1; ++ unsigned int first_scan: 1; ++ unsigned int lbpme: 1; ++ unsigned int lbprz: 1; ++ unsigned int lbpu: 1; ++ unsigned int lbpws: 1; ++ unsigned int lbpws10: 1; ++ unsigned int lbpvpd: 1; ++ unsigned int ws10: 1; ++ unsigned int ws16: 1; ++ unsigned int rc_basis: 2; ++ unsigned int zoned: 2; ++ unsigned int urswrz: 1; ++ unsigned int security: 1; ++ unsigned int ignore_medium_access_errors: 1; ++}; ++ ++struct cdrom_device_ops; ++ ++struct cdrom_device_info { ++ const struct cdrom_device_ops *ops; ++ struct list_head list; ++ struct gendisk *disk; ++ void *handle; ++ int mask; ++ int speed; ++ int capacity; ++ unsigned int options: 30; ++ unsigned int mc_flags: 2; ++ unsigned int vfs_events; ++ unsigned int ioctl_events; ++ int use_count; ++ char name[20]; ++ __u8 sanyo_slot: 2; ++ __u8 keeplocked: 1; ++ __u8 reserved: 5; ++ int cdda_method; ++ __u8 last_sense; ++ __u8 media_written; ++ short unsigned int mmc3_profile; ++ int for_data; ++ int (*exit)(struct cdrom_device_info *); ++ int mrw_mode_page; ++ __s64 last_media_change_ms; ++}; ++ ++struct cdrom_msf0 { ++ __u8 minute; ++ __u8 second; ++ __u8 frame; ++}; ++ ++union cdrom_addr { ++ struct cdrom_msf0 msf; ++ int lba; ++}; ++ ++struct cdrom_ti { ++ __u8 cdti_trk0; ++ __u8 cdti_ind0; ++ __u8 cdti_trk1; ++ __u8 cdti_ind1; ++}; ++ ++struct cdrom_tochdr { ++ __u8 cdth_trk0; ++ __u8 cdth_trk1; ++}; ++ ++struct cdrom_tocentry { ++ __u8 cdte_track; ++ __u8 cdte_adr: 4; ++ __u8 cdte_ctrl: 4; ++ __u8 cdte_format; ++ union cdrom_addr cdte_addr; ++ __u8 cdte_datamode; ++}; ++ ++struct cdrom_multisession { ++ union cdrom_addr addr; ++ __u8 xa_flag; ++ __u8 addr_format; ++}; ++ ++struct cdrom_mcn { ++ __u8 medium_catalog_number[14]; ++}; ++ ++struct packet_command { ++ unsigned char cmd[12]; ++ unsigned char *buffer; ++ unsigned int buflen; ++ int stat; ++ struct scsi_sense_hdr *sshdr; ++ unsigned char data_direction; ++ int quiet; ++ int timeout; ++ void *reserved[1]; ++}; ++ ++struct cdrom_device_ops { ++ int (*open)(struct cdrom_device_info *, int); ++ void (*release)(struct cdrom_device_info *); ++ int (*drive_status)(struct cdrom_device_info *, int); ++ unsigned int (*check_events)(struct cdrom_device_info *, unsigned int, int); ++ int (*tray_move)(struct cdrom_device_info *, int); ++ int (*lock_door)(struct cdrom_device_info *, int); ++ int (*select_speed)(struct cdrom_device_info *, int); ++ int (*get_last_session)(struct cdrom_device_info *, struct cdrom_multisession *); ++ int (*get_mcn)(struct cdrom_device_info *, struct cdrom_mcn *); ++ int (*reset)(struct cdrom_device_info *); ++ int (*audio_ioctl)(struct cdrom_device_info *, unsigned int, void *); ++ int (*generic_packet)(struct cdrom_device_info *, struct packet_command *); ++ int (*read_cdda_bpc)(struct cdrom_device_info *, void *, u32, u32, u8 *); ++ const int capability; ++}; ++ ++struct media_event_desc { ++ __u8 reserved1: 4; ++ __u8 media_event_code: 4; ++ __u8 reserved2: 6; ++ __u8 media_present: 1; ++ __u8 door_open: 1; ++ __u8 start_slot; ++ __u8 end_slot; ++}; ++ ++struct scsi_cd { ++ unsigned int capacity; ++ struct scsi_device *device; ++ unsigned int vendor; ++ long unsigned int ms_offset; ++ unsigned int writeable: 1; ++ unsigned int use: 1; ++ unsigned int xa_flag: 1; ++ unsigned int readcd_known: 1; ++ unsigned int readcd_cdda: 1; ++ unsigned int media_present: 1; ++ int tur_mismatch; ++ bool tur_changed: 1; ++ bool get_event_changed: 1; ++ bool ignore_get_event: 1; ++ struct cdrom_device_info cdi; ++ struct mutex lock; ++ struct gendisk *disk; ++}; ++ ++typedef struct scsi_cd Scsi_CD; ++ ++struct execute_work { ++ struct work_struct work; ++}; ++ ++enum scsi_msg_byte { ++ COMMAND_COMPLETE = 0, ++ EXTENDED_MESSAGE = 1, ++ SAVE_POINTERS = 2, ++ RESTORE_POINTERS = 3, ++ DISCONNECT = 4, ++ INITIATOR_ERROR = 5, ++ ABORT_TASK_SET = 6, ++ MESSAGE_REJECT = 7, ++ NOP = 8, ++ MSG_PARITY_ERROR = 9, ++ LINKED_CMD_COMPLETE = 10, ++ LINKED_FLG_CMD_COMPLETE = 11, ++ TARGET_RESET = 12, ++ ABORT_TASK = 13, ++ CLEAR_TASK_SET = 14, ++ INITIATE_RECOVERY = 15, ++ RELEASE_RECOVERY = 16, ++ TERMINATE_IO_PROC = 17, ++ CLEAR_ACA = 22, ++ LOGICAL_UNIT_RESET = 23, ++ SIMPLE_QUEUE_TAG = 32, ++ HEAD_OF_QUEUE_TAG = 33, ++ ORDERED_QUEUE_TAG = 34, ++ IGNORE_WIDE_RESIDUE = 35, ++ ACA = 36, ++ QAS_REQUEST = 85, ++ BUS_DEVICE_RESET = 12, ++ ABORT = 6, ++}; ++ ++struct sg_io_hdr { ++ int interface_id; ++ int dxfer_direction; ++ unsigned char cmd_len; ++ unsigned char mx_sb_len; ++ short unsigned int iovec_count; ++ unsigned int dxfer_len; ++ void *dxferp; ++ unsigned char *cmdp; ++ void *sbp; ++ unsigned int timeout; ++ unsigned int flags; ++ int pack_id; ++ void *usr_ptr; ++ unsigned char status; ++ unsigned char masked_status; ++ unsigned char msg_status; ++ unsigned char sb_len_wr; ++ short unsigned int host_status; ++ short unsigned int driver_status; ++ int resid; ++ unsigned int duration; ++ unsigned int info; ++}; ++ ++typedef struct sg_io_hdr sg_io_hdr_t; ++ ++struct compat_sg_io_hdr { ++ compat_int_t interface_id; ++ compat_int_t dxfer_direction; ++ unsigned char cmd_len; ++ unsigned char mx_sb_len; ++ short unsigned int iovec_count; ++ compat_uint_t dxfer_len; ++ compat_uint_t dxferp; ++ compat_uptr_t cmdp; ++ compat_uptr_t sbp; ++ compat_uint_t timeout; ++ compat_uint_t flags; ++ compat_int_t pack_id; ++ compat_uptr_t usr_ptr; ++ unsigned char status; ++ unsigned char masked_status; ++ unsigned char msg_status; ++ unsigned char sb_len_wr; ++ short unsigned int host_status; ++ short unsigned int driver_status; ++ compat_int_t resid; ++ compat_uint_t duration; ++ compat_uint_t info; ++}; ++ ++struct sg_scsi_id { ++ int host_no; ++ int channel; ++ int scsi_id; ++ int lun; ++ int scsi_type; ++ short int h_cmd_per_lun; ++ short int d_queue_depth; ++ int unused[2]; ++}; ++ ++typedef struct sg_scsi_id sg_scsi_id_t; ++ ++struct sg_req_info { ++ char req_state; ++ char orphan; ++ char sg_io_owned; ++ char problem; ++ int pack_id; ++ void *usr_ptr; ++ unsigned int duration; ++ int unused; ++}; ++ ++typedef struct sg_req_info sg_req_info_t; ++ ++struct sg_header { ++ int pack_len; ++ int reply_len; ++ int pack_id; ++ int result; ++ unsigned int twelve_byte: 1; ++ unsigned int target_status: 5; ++ unsigned int host_status: 8; ++ unsigned int driver_status: 8; ++ unsigned int other_flags: 10; ++ unsigned char sense_buffer[16]; ++}; ++ ++struct sg_scatter_hold { ++ short unsigned int k_use_sg; ++ unsigned int sglist_len; ++ unsigned int bufflen; ++ struct page **pages; ++ int page_order; ++ char dio_in_use; ++ unsigned char cmd_opcode; ++}; ++ ++typedef struct sg_scatter_hold Sg_scatter_hold; ++ ++struct sg_fd; ++ ++struct sg_request { ++ struct list_head entry; ++ struct sg_fd *parentfp; ++ Sg_scatter_hold data; ++ sg_io_hdr_t header; ++ unsigned char sense_b[96]; ++ char res_used; ++ char orphan; ++ char sg_io_owned; ++ char done; ++ struct request *rq; ++ struct bio *bio; ++ struct execute_work ew; ++}; ++ ++typedef struct sg_request Sg_request; ++ ++struct sg_device; ++ ++struct sg_fd { ++ struct list_head sfd_siblings; ++ struct sg_device *parentdp; ++ wait_queue_head_t read_wait; ++ rwlock_t rq_list_lock; ++ struct mutex f_mutex; ++ int timeout; ++ int timeout_user; ++ Sg_scatter_hold reserve; ++ struct list_head rq_list; ++ struct fasync_struct *async_qp; ++ Sg_request req_arr[16]; ++ char force_packid; ++ char cmd_q; ++ unsigned char next_cmd_len; ++ char keep_orphan; ++ char mmap_called; ++ char res_in_use; ++ struct kref f_ref; ++ struct execute_work ew; ++}; ++ ++struct sg_device { ++ struct scsi_device *device; ++ wait_queue_head_t open_wait; ++ struct mutex open_rel_lock; ++ int sg_tablesize; ++ u32 index; ++ struct list_head sfds; ++ rwlock_t sfd_lock; ++ atomic_t detaching; ++ bool exclude; ++ int open_cnt; ++ char sgdebug; ++ char name[32]; ++ struct cdev *cdev; ++ struct kref d_ref; ++}; ++ ++typedef struct sg_fd Sg_fd; ++ ++typedef struct sg_device Sg_device; ++ ++struct compat_sg_req_info { ++ char req_state; ++ char orphan; ++ char sg_io_owned; ++ char problem; ++ int pack_id; ++ compat_uptr_t usr_ptr; ++ unsigned int duration; ++ int unused; ++}; ++ ++struct sg_proc_deviter { ++ loff_t index; ++ size_t max; ++}; ++ ++enum { ++ ATA_MAX_DEVICES = 2, ++ ATA_MAX_PRD = 256, ++ ATA_SECT_SIZE = 512, ++ ATA_MAX_SECTORS_128 = 128, ++ ATA_MAX_SECTORS = 256, ++ ATA_MAX_SECTORS_1024 = 1024, ++ ATA_MAX_SECTORS_LBA48 = 65535, ++ ATA_MAX_SECTORS_TAPE = 65535, ++ ATA_MAX_TRIM_RNUM = 64, ++ ATA_ID_WORDS = 256, ++ ATA_ID_CONFIG = 0, ++ ATA_ID_CYLS = 1, ++ ATA_ID_HEADS = 3, ++ ATA_ID_SECTORS = 6, ++ ATA_ID_SERNO = 10, ++ ATA_ID_BUF_SIZE = 21, ++ ATA_ID_FW_REV = 23, ++ ATA_ID_PROD = 27, ++ ATA_ID_MAX_MULTSECT = 47, ++ ATA_ID_DWORD_IO = 48, ++ ATA_ID_TRUSTED = 48, ++ ATA_ID_CAPABILITY = 49, ++ ATA_ID_OLD_PIO_MODES = 51, ++ ATA_ID_OLD_DMA_MODES = 52, ++ ATA_ID_FIELD_VALID = 53, ++ ATA_ID_CUR_CYLS = 54, ++ ATA_ID_CUR_HEADS = 55, ++ ATA_ID_CUR_SECTORS = 56, ++ ATA_ID_MULTSECT = 59, ++ ATA_ID_LBA_CAPACITY = 60, ++ ATA_ID_SWDMA_MODES = 62, ++ ATA_ID_MWDMA_MODES = 63, ++ ATA_ID_PIO_MODES = 64, ++ ATA_ID_EIDE_DMA_MIN = 65, ++ ATA_ID_EIDE_DMA_TIME = 66, ++ ATA_ID_EIDE_PIO = 67, ++ ATA_ID_EIDE_PIO_IORDY = 68, ++ ATA_ID_ADDITIONAL_SUPP = 69, ++ ATA_ID_QUEUE_DEPTH = 75, ++ ATA_ID_SATA_CAPABILITY = 76, ++ ATA_ID_SATA_CAPABILITY_2 = 77, ++ ATA_ID_FEATURE_SUPP = 78, ++ ATA_ID_MAJOR_VER = 80, ++ ATA_ID_COMMAND_SET_1 = 82, ++ ATA_ID_COMMAND_SET_2 = 83, ++ ATA_ID_CFSSE = 84, ++ ATA_ID_CFS_ENABLE_1 = 85, ++ ATA_ID_CFS_ENABLE_2 = 86, ++ ATA_ID_CSF_DEFAULT = 87, ++ ATA_ID_UDMA_MODES = 88, ++ ATA_ID_HW_CONFIG = 93, ++ ATA_ID_SPG = 98, ++ ATA_ID_LBA_CAPACITY_2 = 100, ++ ATA_ID_SECTOR_SIZE = 106, ++ ATA_ID_WWN = 108, ++ ATA_ID_LOGICAL_SECTOR_SIZE = 117, ++ ATA_ID_COMMAND_SET_3 = 119, ++ ATA_ID_COMMAND_SET_4 = 120, ++ ATA_ID_LAST_LUN = 126, ++ ATA_ID_DLF = 128, ++ ATA_ID_CSFO = 129, ++ ATA_ID_CFA_POWER = 160, ++ ATA_ID_CFA_KEY_MGMT = 162, ++ ATA_ID_CFA_MODES = 163, ++ ATA_ID_DATA_SET_MGMT = 169, ++ ATA_ID_SCT_CMD_XPORT = 206, ++ ATA_ID_ROT_SPEED = 217, ++ ATA_ID_PIO4 = 2, ++ ATA_ID_SERNO_LEN = 20, ++ ATA_ID_FW_REV_LEN = 8, ++ ATA_ID_PROD_LEN = 40, ++ ATA_ID_WWN_LEN = 8, ++ ATA_PCI_CTL_OFS = 2, ++ ATA_PIO0 = 1, ++ ATA_PIO1 = 3, ++ ATA_PIO2 = 7, ++ ATA_PIO3 = 15, ++ ATA_PIO4 = 31, ++ ATA_PIO5 = 63, ++ ATA_PIO6 = 127, ++ ATA_PIO4_ONLY = 16, ++ ATA_SWDMA0 = 1, ++ ATA_SWDMA1 = 3, ++ ATA_SWDMA2 = 7, ++ ATA_SWDMA2_ONLY = 4, ++ ATA_MWDMA0 = 1, ++ ATA_MWDMA1 = 3, ++ ATA_MWDMA2 = 7, ++ ATA_MWDMA3 = 15, ++ ATA_MWDMA4 = 31, ++ ATA_MWDMA12_ONLY = 6, ++ ATA_MWDMA2_ONLY = 4, ++ ATA_UDMA0 = 1, ++ ATA_UDMA1 = 3, ++ ATA_UDMA2 = 7, ++ ATA_UDMA3 = 15, ++ ATA_UDMA4 = 31, ++ ATA_UDMA5 = 63, ++ ATA_UDMA6 = 127, ++ ATA_UDMA7 = 255, ++ ATA_UDMA24_ONLY = 20, ++ ATA_UDMA_MASK_40C = 7, ++ ATA_PRD_SZ = 8, ++ ATA_PRD_TBL_SZ = 2048, ++ ATA_PRD_EOT = -2147483648, ++ ATA_DMA_TABLE_OFS = 4, ++ ATA_DMA_STATUS = 2, ++ ATA_DMA_CMD = 0, ++ ATA_DMA_WR = 8, ++ ATA_DMA_START = 1, ++ ATA_DMA_INTR = 4, ++ ATA_DMA_ERR = 2, ++ ATA_DMA_ACTIVE = 1, ++ ATA_HOB = 128, ++ ATA_NIEN = 2, ++ ATA_LBA = 64, ++ ATA_DEV1 = 16, ++ ATA_DEVICE_OBS = 160, ++ ATA_DEVCTL_OBS = 8, ++ ATA_BUSY = 128, ++ ATA_DRDY = 64, ++ ATA_DF = 32, ++ ATA_DSC = 16, ++ ATA_DRQ = 8, ++ ATA_CORR = 4, ++ ATA_SENSE = 2, ++ ATA_ERR = 1, ++ ATA_SRST = 4, ++ ATA_ICRC = 128, ++ ATA_BBK = 128, ++ ATA_UNC = 64, ++ ATA_MC = 32, ++ ATA_IDNF = 16, ++ ATA_MCR = 8, ++ ATA_ABORTED = 4, ++ ATA_TRK0NF = 2, ++ ATA_AMNF = 1, ++ ATAPI_LFS = 240, ++ ATAPI_EOM = 2, ++ ATAPI_ILI = 1, ++ ATAPI_IO = 2, ++ ATAPI_COD = 1, ++ ATA_REG_DATA = 0, ++ ATA_REG_ERR = 1, ++ ATA_REG_NSECT = 2, ++ ATA_REG_LBAL = 3, ++ ATA_REG_LBAM = 4, ++ ATA_REG_LBAH = 5, ++ ATA_REG_DEVICE = 6, ++ ATA_REG_STATUS = 7, ++ ATA_REG_FEATURE = 1, ++ ATA_REG_CMD = 7, ++ ATA_REG_BYTEL = 4, ++ ATA_REG_BYTEH = 5, ++ ATA_REG_DEVSEL = 6, ++ ATA_REG_IRQ = 2, ++ ATA_CMD_DEV_RESET = 8, ++ ATA_CMD_CHK_POWER = 229, ++ ATA_CMD_STANDBY = 226, ++ ATA_CMD_IDLE = 227, ++ ATA_CMD_EDD = 144, ++ ATA_CMD_DOWNLOAD_MICRO = 146, ++ ATA_CMD_DOWNLOAD_MICRO_DMA = 147, ++ ATA_CMD_NOP = 0, ++ ATA_CMD_FLUSH = 231, ++ ATA_CMD_FLUSH_EXT = 234, ++ ATA_CMD_ID_ATA = 236, ++ ATA_CMD_ID_ATAPI = 161, ++ ATA_CMD_SERVICE = 162, ++ ATA_CMD_READ = 200, ++ ATA_CMD_READ_EXT = 37, ++ ATA_CMD_READ_QUEUED = 38, ++ ATA_CMD_READ_STREAM_EXT = 43, ++ ATA_CMD_READ_STREAM_DMA_EXT = 42, ++ ATA_CMD_WRITE = 202, ++ ATA_CMD_WRITE_EXT = 53, ++ ATA_CMD_WRITE_QUEUED = 54, ++ ATA_CMD_WRITE_STREAM_EXT = 59, ++ ATA_CMD_WRITE_STREAM_DMA_EXT = 58, ++ ATA_CMD_WRITE_FUA_EXT = 61, ++ ATA_CMD_WRITE_QUEUED_FUA_EXT = 62, ++ ATA_CMD_FPDMA_READ = 96, ++ ATA_CMD_FPDMA_WRITE = 97, ++ ATA_CMD_NCQ_NON_DATA = 99, ++ ATA_CMD_FPDMA_SEND = 100, ++ ATA_CMD_FPDMA_RECV = 101, ++ ATA_CMD_PIO_READ = 32, ++ ATA_CMD_PIO_READ_EXT = 36, ++ ATA_CMD_PIO_WRITE = 48, ++ ATA_CMD_PIO_WRITE_EXT = 52, ++ ATA_CMD_READ_MULTI = 196, ++ ATA_CMD_READ_MULTI_EXT = 41, ++ ATA_CMD_WRITE_MULTI = 197, ++ ATA_CMD_WRITE_MULTI_EXT = 57, ++ ATA_CMD_WRITE_MULTI_FUA_EXT = 206, ++ ATA_CMD_SET_FEATURES = 239, ++ ATA_CMD_SET_MULTI = 198, ++ ATA_CMD_PACKET = 160, ++ ATA_CMD_VERIFY = 64, ++ ATA_CMD_VERIFY_EXT = 66, ++ ATA_CMD_WRITE_UNCORR_EXT = 69, ++ ATA_CMD_STANDBYNOW1 = 224, ++ ATA_CMD_IDLEIMMEDIATE = 225, ++ ATA_CMD_SLEEP = 230, ++ ATA_CMD_INIT_DEV_PARAMS = 145, ++ ATA_CMD_READ_NATIVE_MAX = 248, ++ ATA_CMD_READ_NATIVE_MAX_EXT = 39, ++ ATA_CMD_SET_MAX = 249, ++ ATA_CMD_SET_MAX_EXT = 55, ++ ATA_CMD_READ_LOG_EXT = 47, ++ ATA_CMD_WRITE_LOG_EXT = 63, ++ ATA_CMD_READ_LOG_DMA_EXT = 71, ++ ATA_CMD_WRITE_LOG_DMA_EXT = 87, ++ ATA_CMD_TRUSTED_NONDATA = 91, ++ ATA_CMD_TRUSTED_RCV = 92, ++ ATA_CMD_TRUSTED_RCV_DMA = 93, ++ ATA_CMD_TRUSTED_SND = 94, ++ ATA_CMD_TRUSTED_SND_DMA = 95, ++ ATA_CMD_PMP_READ = 228, ++ ATA_CMD_PMP_READ_DMA = 233, ++ ATA_CMD_PMP_WRITE = 232, ++ ATA_CMD_PMP_WRITE_DMA = 235, ++ ATA_CMD_CONF_OVERLAY = 177, ++ ATA_CMD_SEC_SET_PASS = 241, ++ ATA_CMD_SEC_UNLOCK = 242, ++ ATA_CMD_SEC_ERASE_PREP = 243, ++ ATA_CMD_SEC_ERASE_UNIT = 244, ++ ATA_CMD_SEC_FREEZE_LOCK = 245, ++ ATA_CMD_SEC_DISABLE_PASS = 246, ++ ATA_CMD_CONFIG_STREAM = 81, ++ ATA_CMD_SMART = 176, ++ ATA_CMD_MEDIA_LOCK = 222, ++ ATA_CMD_MEDIA_UNLOCK = 223, ++ ATA_CMD_DSM = 6, ++ ATA_CMD_CHK_MED_CRD_TYP = 209, ++ ATA_CMD_CFA_REQ_EXT_ERR = 3, ++ ATA_CMD_CFA_WRITE_NE = 56, ++ ATA_CMD_CFA_TRANS_SECT = 135, ++ ATA_CMD_CFA_ERASE = 192, ++ ATA_CMD_CFA_WRITE_MULT_NE = 205, ++ ATA_CMD_REQ_SENSE_DATA = 11, ++ ATA_CMD_SANITIZE_DEVICE = 180, ++ ATA_CMD_ZAC_MGMT_IN = 74, ++ ATA_CMD_ZAC_MGMT_OUT = 159, ++ ATA_CMD_RESTORE = 16, ++ ATA_SUBCMD_FPDMA_RECV_RD_LOG_DMA_EXT = 1, ++ ATA_SUBCMD_FPDMA_RECV_ZAC_MGMT_IN = 2, ++ ATA_SUBCMD_FPDMA_SEND_DSM = 0, ++ ATA_SUBCMD_FPDMA_SEND_WR_LOG_DMA_EXT = 2, ++ ATA_SUBCMD_NCQ_NON_DATA_ABORT_QUEUE = 0, ++ ATA_SUBCMD_NCQ_NON_DATA_SET_FEATURES = 5, ++ ATA_SUBCMD_NCQ_NON_DATA_ZERO_EXT = 6, ++ ATA_SUBCMD_NCQ_NON_DATA_ZAC_MGMT_OUT = 7, ++ ATA_SUBCMD_ZAC_MGMT_IN_REPORT_ZONES = 0, ++ ATA_SUBCMD_ZAC_MGMT_OUT_CLOSE_ZONE = 1, ++ ATA_SUBCMD_ZAC_MGMT_OUT_FINISH_ZONE = 2, ++ ATA_SUBCMD_ZAC_MGMT_OUT_OPEN_ZONE = 3, ++ ATA_SUBCMD_ZAC_MGMT_OUT_RESET_WRITE_POINTER = 4, ++ ATA_LOG_DIRECTORY = 0, ++ ATA_LOG_SATA_NCQ = 16, ++ ATA_LOG_NCQ_NON_DATA = 18, ++ ATA_LOG_NCQ_SEND_RECV = 19, ++ ATA_LOG_IDENTIFY_DEVICE = 48, ++ ATA_LOG_CONCURRENT_POSITIONING_RANGES = 71, ++ ATA_LOG_SECURITY = 6, ++ ATA_LOG_SATA_SETTINGS = 8, ++ ATA_LOG_ZONED_INFORMATION = 9, ++ ATA_LOG_DEVSLP_OFFSET = 48, ++ ATA_LOG_DEVSLP_SIZE = 8, ++ ATA_LOG_DEVSLP_MDAT = 0, ++ ATA_LOG_DEVSLP_MDAT_MASK = 31, ++ ATA_LOG_DEVSLP_DETO = 1, ++ ATA_LOG_DEVSLP_VALID = 7, ++ ATA_LOG_DEVSLP_VALID_MASK = 128, ++ ATA_LOG_NCQ_PRIO_OFFSET = 9, ++ ATA_LOG_NCQ_SEND_RECV_SUBCMDS_OFFSET = 0, ++ ATA_LOG_NCQ_SEND_RECV_SUBCMDS_DSM = 1, ++ ATA_LOG_NCQ_SEND_RECV_DSM_OFFSET = 4, ++ ATA_LOG_NCQ_SEND_RECV_DSM_TRIM = 1, ++ ATA_LOG_NCQ_SEND_RECV_RD_LOG_OFFSET = 8, ++ ATA_LOG_NCQ_SEND_RECV_RD_LOG_SUPPORTED = 1, ++ ATA_LOG_NCQ_SEND_RECV_WR_LOG_OFFSET = 12, ++ ATA_LOG_NCQ_SEND_RECV_WR_LOG_SUPPORTED = 1, ++ ATA_LOG_NCQ_SEND_RECV_ZAC_MGMT_OFFSET = 16, ++ ATA_LOG_NCQ_SEND_RECV_ZAC_MGMT_OUT_SUPPORTED = 1, ++ ATA_LOG_NCQ_SEND_RECV_ZAC_MGMT_IN_SUPPORTED = 2, ++ ATA_LOG_NCQ_SEND_RECV_SIZE = 20, ++ ATA_LOG_NCQ_NON_DATA_SUBCMDS_OFFSET = 0, ++ ATA_LOG_NCQ_NON_DATA_ABORT_OFFSET = 0, ++ ATA_LOG_NCQ_NON_DATA_ABORT_NCQ = 1, ++ ATA_LOG_NCQ_NON_DATA_ABORT_ALL = 2, ++ ATA_LOG_NCQ_NON_DATA_ABORT_STREAMING = 4, ++ ATA_LOG_NCQ_NON_DATA_ABORT_NON_STREAMING = 8, ++ ATA_LOG_NCQ_NON_DATA_ABORT_SELECTED = 16, ++ ATA_LOG_NCQ_NON_DATA_ZAC_MGMT_OFFSET = 28, ++ ATA_LOG_NCQ_NON_DATA_ZAC_MGMT_OUT = 1, ++ ATA_LOG_NCQ_NON_DATA_SIZE = 64, ++ ATA_CMD_READ_LONG = 34, ++ ATA_CMD_READ_LONG_ONCE = 35, ++ ATA_CMD_WRITE_LONG = 50, ++ ATA_CMD_WRITE_LONG_ONCE = 51, ++ SETFEATURES_XFER = 3, ++ XFER_UDMA_7 = 71, ++ XFER_UDMA_6 = 70, ++ XFER_UDMA_5 = 69, ++ XFER_UDMA_4 = 68, ++ XFER_UDMA_3 = 67, ++ XFER_UDMA_2 = 66, ++ XFER_UDMA_1 = 65, ++ XFER_UDMA_0 = 64, ++ XFER_MW_DMA_4 = 36, ++ XFER_MW_DMA_3 = 35, ++ XFER_MW_DMA_2 = 34, ++ XFER_MW_DMA_1 = 33, ++ XFER_MW_DMA_0 = 32, ++ XFER_SW_DMA_2 = 18, ++ XFER_SW_DMA_1 = 17, ++ XFER_SW_DMA_0 = 16, ++ XFER_PIO_6 = 14, ++ XFER_PIO_5 = 13, ++ XFER_PIO_4 = 12, ++ XFER_PIO_3 = 11, ++ XFER_PIO_2 = 10, ++ XFER_PIO_1 = 9, ++ XFER_PIO_0 = 8, ++ XFER_PIO_SLOW = 0, ++ SETFEATURES_WC_ON = 2, ++ SETFEATURES_WC_OFF = 130, ++ SETFEATURES_RA_ON = 170, ++ SETFEATURES_RA_OFF = 85, ++ SETFEATURES_AAM_ON = 66, ++ SETFEATURES_AAM_OFF = 194, ++ SETFEATURES_SPINUP = 7, ++ SETFEATURES_SPINUP_TIMEOUT = 30000, ++ SETFEATURES_SATA_ENABLE = 16, ++ SETFEATURES_SATA_DISABLE = 144, ++ SATA_FPDMA_OFFSET = 1, ++ SATA_FPDMA_AA = 2, ++ SATA_DIPM = 3, ++ SATA_FPDMA_IN_ORDER = 4, ++ SATA_AN = 5, ++ SATA_SSP = 6, ++ SATA_DEVSLP = 9, ++ SETFEATURE_SENSE_DATA = 195, ++ ATA_SET_MAX_ADDR = 0, ++ ATA_SET_MAX_PASSWD = 1, ++ ATA_SET_MAX_LOCK = 2, ++ ATA_SET_MAX_UNLOCK = 3, ++ ATA_SET_MAX_FREEZE_LOCK = 4, ++ ATA_SET_MAX_PASSWD_DMA = 5, ++ ATA_SET_MAX_UNLOCK_DMA = 6, ++ ATA_DCO_RESTORE = 192, ++ ATA_DCO_FREEZE_LOCK = 193, ++ ATA_DCO_IDENTIFY = 194, ++ ATA_DCO_SET = 195, ++ ATA_SMART_ENABLE = 216, ++ ATA_SMART_READ_VALUES = 208, ++ ATA_SMART_READ_THRESHOLDS = 209, ++ ATA_DSM_TRIM = 1, ++ ATA_SMART_LBAM_PASS = 79, ++ ATA_SMART_LBAH_PASS = 194, ++ ATAPI_PKT_DMA = 1, ++ ATAPI_DMADIR = 4, ++ ATAPI_CDB_LEN = 16, ++ SATA_PMP_MAX_PORTS = 15, ++ SATA_PMP_CTRL_PORT = 15, ++ SATA_PMP_GSCR_DWORDS = 128, ++ SATA_PMP_GSCR_PROD_ID = 0, ++ SATA_PMP_GSCR_REV = 1, ++ SATA_PMP_GSCR_PORT_INFO = 2, ++ SATA_PMP_GSCR_ERROR = 32, ++ SATA_PMP_GSCR_ERROR_EN = 33, ++ SATA_PMP_GSCR_FEAT = 64, ++ SATA_PMP_GSCR_FEAT_EN = 96, ++ SATA_PMP_PSCR_STATUS = 0, ++ SATA_PMP_PSCR_ERROR = 1, ++ SATA_PMP_PSCR_CONTROL = 2, ++ SATA_PMP_FEAT_BIST = 1, ++ SATA_PMP_FEAT_PMREQ = 2, ++ SATA_PMP_FEAT_DYNSSC = 4, ++ SATA_PMP_FEAT_NOTIFY = 8, ++ ATA_CBL_NONE = 0, ++ ATA_CBL_PATA40 = 1, ++ ATA_CBL_PATA80 = 2, ++ ATA_CBL_PATA40_SHORT = 3, ++ ATA_CBL_PATA_UNK = 4, ++ ATA_CBL_PATA_IGN = 5, ++ ATA_CBL_SATA = 6, ++ SCR_STATUS = 0, ++ SCR_ERROR = 1, ++ SCR_CONTROL = 2, ++ SCR_ACTIVE = 3, ++ SCR_NOTIFICATION = 4, ++ SERR_DATA_RECOVERED = 1, ++ SERR_COMM_RECOVERED = 2, ++ SERR_DATA = 256, ++ SERR_PERSISTENT = 512, ++ SERR_PROTOCOL = 1024, ++ SERR_INTERNAL = 2048, ++ SERR_PHYRDY_CHG = 65536, ++ SERR_PHY_INT_ERR = 131072, ++ SERR_COMM_WAKE = 262144, ++ SERR_10B_8B_ERR = 524288, ++ SERR_DISPARITY = 1048576, ++ SERR_CRC = 2097152, ++ SERR_HANDSHAKE = 4194304, ++ SERR_LINK_SEQ_ERR = 8388608, ++ SERR_TRANS_ST_ERROR = 16777216, ++ SERR_UNRECOG_FIS = 33554432, ++ SERR_DEV_XCHG = 67108864, ++}; ++ ++enum ata_prot_flags { ++ ATA_PROT_FLAG_PIO = 1, ++ ATA_PROT_FLAG_DMA = 2, ++ ATA_PROT_FLAG_NCQ = 4, ++ ATA_PROT_FLAG_ATAPI = 8, ++ ATA_PROT_UNKNOWN = 255, ++ ATA_PROT_NODATA = 0, ++ ATA_PROT_PIO = 1, ++ ATA_PROT_DMA = 2, ++ ATA_PROT_NCQ_NODATA = 4, ++ ATA_PROT_NCQ = 6, ++ ATAPI_PROT_NODATA = 8, ++ ATAPI_PROT_PIO = 9, ++ ATAPI_PROT_DMA = 10, ++}; ++ ++struct ata_bmdma_prd { ++ __le32 addr; ++ __le32 flags_len; ++}; ++ ++enum { ++ LIBATA_MAX_PRD = 128, ++ LIBATA_DUMB_MAX_PRD = 64, ++ ATA_DEF_QUEUE = 1, ++ ATA_MAX_QUEUE = 32, ++ ATA_TAG_INTERNAL = 32, ++ ATA_SHORT_PAUSE = 16, ++ ATAPI_MAX_DRAIN = 16384, ++ ATA_ALL_DEVICES = 3, ++ ATA_SHT_EMULATED = 1, ++ ATA_SHT_THIS_ID = -1, ++ ATA_TFLAG_LBA48 = 1, ++ ATA_TFLAG_ISADDR = 2, ++ ATA_TFLAG_DEVICE = 4, ++ ATA_TFLAG_WRITE = 8, ++ ATA_TFLAG_LBA = 16, ++ ATA_TFLAG_FUA = 32, ++ ATA_TFLAG_POLLING = 64, ++ ATA_DFLAG_LBA = 1, ++ ATA_DFLAG_LBA48 = 2, ++ ATA_DFLAG_CDB_INTR = 4, ++ ATA_DFLAG_NCQ = 8, ++ ATA_DFLAG_FLUSH_EXT = 16, ++ ATA_DFLAG_ACPI_PENDING = 32, ++ ATA_DFLAG_ACPI_FAILED = 64, ++ ATA_DFLAG_AN = 128, ++ ATA_DFLAG_TRUSTED = 256, ++ ATA_DFLAG_FUA = 512, ++ ATA_DFLAG_DMADIR = 1024, ++ ATA_DFLAG_NCQ_SEND_RECV = 2048, ++ ATA_DFLAG_NCQ_PRIO = 4096, ++ ATA_DFLAG_CFG_MASK = 8191, ++ ATA_DFLAG_PIO = 8192, ++ ATA_DFLAG_NCQ_OFF = 16384, ++ ATA_DFLAG_SLEEPING = 32768, ++ ATA_DFLAG_DUBIOUS_XFER = 65536, ++ ATA_DFLAG_NO_UNLOAD = 131072, ++ ATA_DFLAG_UNLOCK_HPA = 262144, ++ ATA_DFLAG_INIT_MASK = 524287, ++ ATA_DFLAG_NCQ_PRIO_ENABLED = 524288, ++ ATA_DFLAG_DETACH = 16777216, ++ ATA_DFLAG_DETACHED = 33554432, ++ ATA_DFLAG_DA = 67108864, ++ ATA_DFLAG_DEVSLP = 134217728, ++ ATA_DFLAG_ACPI_DISABLED = 268435456, ++ ATA_DFLAG_D_SENSE = 536870912, ++ ATA_DFLAG_ZAC = 1073741824, ++ ATA_DFLAG_FEATURES_MASK = 201333504, ++ ATA_DEV_UNKNOWN = 0, ++ ATA_DEV_ATA = 1, ++ ATA_DEV_ATA_UNSUP = 2, ++ ATA_DEV_ATAPI = 3, ++ ATA_DEV_ATAPI_UNSUP = 4, ++ ATA_DEV_PMP = 5, ++ ATA_DEV_PMP_UNSUP = 6, ++ ATA_DEV_SEMB = 7, ++ ATA_DEV_SEMB_UNSUP = 8, ++ ATA_DEV_ZAC = 9, ++ ATA_DEV_ZAC_UNSUP = 10, ++ ATA_DEV_NONE = 11, ++ ATA_LFLAG_NO_HRST = 2, ++ ATA_LFLAG_NO_SRST = 4, ++ ATA_LFLAG_ASSUME_ATA = 8, ++ ATA_LFLAG_ASSUME_SEMB = 16, ++ ATA_LFLAG_ASSUME_CLASS = 24, ++ ATA_LFLAG_NO_RETRY = 32, ++ ATA_LFLAG_DISABLED = 64, ++ ATA_LFLAG_SW_ACTIVITY = 128, ++ ATA_LFLAG_NO_LPM = 256, ++ ATA_LFLAG_RST_ONCE = 512, ++ ATA_LFLAG_CHANGED = 1024, ++ ATA_LFLAG_NO_DEBOUNCE_DELAY = 2048, ++ ATA_FLAG_SLAVE_POSS = 1, ++ ATA_FLAG_SATA = 2, ++ ATA_FLAG_NO_LPM = 4, ++ ATA_FLAG_NO_LOG_PAGE = 32, ++ ATA_FLAG_NO_ATAPI = 64, ++ ATA_FLAG_PIO_DMA = 128, ++ ATA_FLAG_PIO_LBA48 = 256, ++ ATA_FLAG_PIO_POLLING = 512, ++ ATA_FLAG_NCQ = 1024, ++ ATA_FLAG_NO_POWEROFF_SPINDOWN = 2048, ++ ATA_FLAG_NO_HIBERNATE_SPINDOWN = 4096, ++ ATA_FLAG_DEBUGMSG = 8192, ++ ATA_FLAG_FPDMA_AA = 16384, ++ ATA_FLAG_IGN_SIMPLEX = 32768, ++ ATA_FLAG_NO_IORDY = 65536, ++ ATA_FLAG_ACPI_SATA = 131072, ++ ATA_FLAG_AN = 262144, ++ ATA_FLAG_PMP = 524288, ++ ATA_FLAG_FPDMA_AUX = 1048576, ++ ATA_FLAG_EM = 2097152, ++ ATA_FLAG_SW_ACTIVITY = 4194304, ++ ATA_FLAG_NO_DIPM = 8388608, ++ ATA_FLAG_SAS_HOST = 16777216, ++ ATA_PFLAG_EH_PENDING = 1, ++ ATA_PFLAG_EH_IN_PROGRESS = 2, ++ ATA_PFLAG_FROZEN = 4, ++ ATA_PFLAG_RECOVERED = 8, ++ ATA_PFLAG_LOADING = 16, ++ ATA_PFLAG_SCSI_HOTPLUG = 64, ++ ATA_PFLAG_INITIALIZING = 128, ++ ATA_PFLAG_RESETTING = 256, ++ ATA_PFLAG_UNLOADING = 512, ++ ATA_PFLAG_UNLOADED = 1024, ++ ATA_PFLAG_SUSPENDED = 131072, ++ ATA_PFLAG_PM_PENDING = 262144, ++ ATA_PFLAG_INIT_GTM_VALID = 524288, ++ ATA_PFLAG_PIO32 = 1048576, ++ ATA_PFLAG_PIO32CHANGE = 2097152, ++ ATA_PFLAG_EXTERNAL = 4194304, ++ ATA_QCFLAG_ACTIVE = 1, ++ ATA_QCFLAG_DMAMAP = 2, ++ ATA_QCFLAG_RTF_FILLED = 4, ++ ATA_QCFLAG_IO = 8, ++ ATA_QCFLAG_RESULT_TF = 16, ++ ATA_QCFLAG_CLEAR_EXCL = 32, ++ ATA_QCFLAG_QUIET = 64, ++ ATA_QCFLAG_RETRY = 128, ++ ATA_QCFLAG_EH = 65536, ++ ATA_QCFLAG_SENSE_VALID = 131072, ++ ATA_QCFLAG_EH_SCHEDULED = 262144, ++ ATA_HOST_SIMPLEX = 1, ++ ATA_HOST_STARTED = 2, ++ ATA_HOST_PARALLEL_SCAN = 4, ++ ATA_HOST_IGNORE_ATA = 8, ++ ATA_TMOUT_BOOT = 30000, ++ ATA_TMOUT_BOOT_QUICK = 7000, ++ ATA_TMOUT_INTERNAL_QUICK = 5000, ++ ATA_TMOUT_MAX_PARK = 30000, ++ ATA_TMOUT_FF_WAIT_LONG = 2000, ++ ATA_TMOUT_FF_WAIT = 800, ++ ATA_WAIT_AFTER_RESET = 150, ++ ATA_TMOUT_PMP_SRST_WAIT = 5000, ++ ATA_TMOUT_SPURIOUS_PHY = 10000, ++ BUS_UNKNOWN = 0, ++ BUS_DMA = 1, ++ BUS_IDLE = 2, ++ BUS_NOINTR = 3, ++ BUS_NODATA = 4, ++ BUS_TIMER = 5, ++ BUS_PIO = 6, ++ BUS_EDD = 7, ++ BUS_IDENTIFY = 8, ++ BUS_PACKET = 9, ++ PORT_UNKNOWN = 0, ++ PORT_ENABLED = 1, ++ PORT_DISABLED = 2, ++ ATA_NR_PIO_MODES = 7, ++ ATA_NR_MWDMA_MODES = 5, ++ ATA_NR_UDMA_MODES = 8, ++ ATA_SHIFT_PIO = 0, ++ ATA_SHIFT_MWDMA = 7, ++ ATA_SHIFT_UDMA = 12, ++ ATA_SHIFT_PRIO = 6, ++ ATA_PRIO_HIGH = 2, ++ ATA_DMA_PAD_SZ = 4, ++ ATA_ERING_SIZE = 32, ++ ATA_DEFER_LINK = 1, ++ ATA_DEFER_PORT = 2, ++ ATA_EH_DESC_LEN = 80, ++ ATA_EH_REVALIDATE = 1, ++ ATA_EH_SOFTRESET = 2, ++ ATA_EH_HARDRESET = 4, ++ ATA_EH_RESET = 6, ++ ATA_EH_ENABLE_LINK = 8, ++ ATA_EH_PARK = 32, ++ ATA_EH_PERDEV_MASK = 33, ++ ATA_EH_ALL_ACTIONS = 15, ++ ATA_EHI_HOTPLUGGED = 1, ++ ATA_EHI_NO_AUTOPSY = 4, ++ ATA_EHI_QUIET = 8, ++ ATA_EHI_NO_RECOVERY = 16, ++ ATA_EHI_DID_SOFTRESET = 65536, ++ ATA_EHI_DID_HARDRESET = 131072, ++ ATA_EHI_PRINTINFO = 262144, ++ ATA_EHI_SETMODE = 524288, ++ ATA_EHI_POST_SETMODE = 1048576, ++ ATA_EHI_DID_RESET = 196608, ++ ATA_EHI_TO_SLAVE_MASK = 12, ++ ATA_EH_MAX_TRIES = 5, ++ ATA_LINK_RESUME_TRIES = 5, ++ ATA_PROBE_MAX_TRIES = 3, ++ ATA_EH_DEV_TRIES = 3, ++ ATA_EH_PMP_TRIES = 5, ++ ATA_EH_PMP_LINK_TRIES = 3, ++ SATA_PMP_RW_TIMEOUT = 3000, ++ ATA_EH_CMD_TIMEOUT_TABLE_SIZE = 7, ++ ATA_HORKAGE_DIAGNOSTIC = 1, ++ ATA_HORKAGE_NODMA = 2, ++ ATA_HORKAGE_NONCQ = 4, ++ ATA_HORKAGE_MAX_SEC_128 = 8, ++ ATA_HORKAGE_BROKEN_HPA = 16, ++ ATA_HORKAGE_DISABLE = 32, ++ ATA_HORKAGE_HPA_SIZE = 64, ++ ATA_HORKAGE_IVB = 256, ++ ATA_HORKAGE_STUCK_ERR = 512, ++ ATA_HORKAGE_BRIDGE_OK = 1024, ++ ATA_HORKAGE_ATAPI_MOD16_DMA = 2048, ++ ATA_HORKAGE_FIRMWARE_WARN = 4096, ++ ATA_HORKAGE_1_5_GBPS = 8192, ++ ATA_HORKAGE_NOSETXFER = 16384, ++ ATA_HORKAGE_BROKEN_FPDMA_AA = 32768, ++ ATA_HORKAGE_DUMP_ID = 65536, ++ ATA_HORKAGE_MAX_SEC_LBA48 = 131072, ++ ATA_HORKAGE_ATAPI_DMADIR = 262144, ++ ATA_HORKAGE_NO_NCQ_TRIM = 524288, ++ ATA_HORKAGE_NOLPM = 1048576, ++ ATA_HORKAGE_WD_BROKEN_LPM = 2097152, ++ ATA_HORKAGE_ZERO_AFTER_TRIM = 4194304, ++ ATA_HORKAGE_NO_DMA_LOG = 8388608, ++ ATA_HORKAGE_NOTRIM = 16777216, ++ ATA_HORKAGE_MAX_SEC_1024 = 33554432, ++ ATA_HORKAGE_MAX_TRIM_128M = 67108864, ++ ATA_HORKAGE_NO_NCQ_ON_ATI = 134217728, ++ ATA_HORKAGE_NO_ID_DEV_LOG = 268435456, ++ ATA_HORKAGE_NO_LOG_DIR = 536870912, ++ ATA_HORKAGE_NO_FUA = 1073741824, ++ ATA_DMA_MASK_ATA = 1, ++ ATA_DMA_MASK_ATAPI = 2, ++ ATA_DMA_MASK_CFA = 4, ++ ATAPI_READ = 0, ++ ATAPI_WRITE = 1, ++ ATAPI_READ_CD = 2, ++ ATAPI_PASS_THRU = 3, ++ ATAPI_MISC = 4, ++ ATA_TIMING_SETUP = 1, ++ ATA_TIMING_ACT8B = 2, ++ ATA_TIMING_REC8B = 4, ++ ATA_TIMING_CYC8B = 8, ++ ATA_TIMING_8BIT = 14, ++ ATA_TIMING_ACTIVE = 16, ++ ATA_TIMING_RECOVER = 32, ++ ATA_TIMING_DMACK_HOLD = 64, ++ ATA_TIMING_CYCLE = 128, ++ ATA_TIMING_UDMA = 256, ++ ATA_TIMING_ALL = 511, ++ ATA_ACPI_FILTER_SETXFER = 1, ++ ATA_ACPI_FILTER_LOCK = 2, ++ ATA_ACPI_FILTER_DIPM = 4, ++ ATA_ACPI_FILTER_FPDMA_OFFSET = 8, ++ ATA_ACPI_FILTER_FPDMA_AA = 16, ++ ATA_ACPI_FILTER_DEFAULT = 7, ++}; ++ ++enum ata_completion_errors { ++ AC_ERR_OK = 0, ++ AC_ERR_DEV = 1, ++ AC_ERR_HSM = 2, ++ AC_ERR_TIMEOUT = 4, ++ AC_ERR_MEDIA = 8, ++ AC_ERR_ATA_BUS = 16, ++ AC_ERR_HOST_BUS = 32, ++ AC_ERR_SYSTEM = 64, ++ AC_ERR_INVALID = 128, ++ AC_ERR_OTHER = 256, ++ AC_ERR_NODEV_HINT = 512, ++ AC_ERR_NCQ = 1024, ++}; ++ ++enum ata_lpm_policy { ++ ATA_LPM_UNKNOWN = 0, ++ ATA_LPM_MAX_POWER = 1, ++ ATA_LPM_MED_POWER = 2, ++ ATA_LPM_MED_POWER_WITH_DIPM = 3, ++ ATA_LPM_MIN_POWER_WITH_PARTIAL = 4, ++ ATA_LPM_MIN_POWER = 5, ++}; ++ ++enum ata_lpm_hints { ++ ATA_LPM_EMPTY = 1, ++ ATA_LPM_HIPM = 2, ++ ATA_LPM_WAKE_ONLY = 4, ++}; ++ ++struct ata_queued_cmd; ++ ++typedef void (*ata_qc_cb_t)(struct ata_queued_cmd *); ++ ++struct ata_taskfile { ++ long unsigned int flags; ++ u8 protocol; ++ u8 ctl; ++ u8 hob_feature; ++ u8 hob_nsect; ++ u8 hob_lbal; ++ u8 hob_lbam; ++ u8 hob_lbah; ++ union { ++ u8 error; ++ u8 feature; ++ }; ++ u8 nsect; ++ u8 lbal; ++ u8 lbam; ++ u8 lbah; ++ u8 device; ++ union { ++ u8 status; ++ u8 command; ++ }; ++ u32 auxiliary; ++}; ++ ++struct ata_port; ++ ++struct ata_device; ++ ++struct ata_queued_cmd { ++ struct ata_port *ap; ++ struct ata_device *dev; ++ struct scsi_cmnd *scsicmd; ++ void (*scsidone)(struct scsi_cmnd *); ++ struct ata_taskfile tf; ++ u8 cdb[16]; ++ long unsigned int flags; ++ unsigned int tag; ++ unsigned int hw_tag; ++ unsigned int n_elem; ++ unsigned int orig_n_elem; ++ int dma_dir; ++ unsigned int sect_size; ++ unsigned int nbytes; ++ unsigned int extrabytes; ++ unsigned int curbytes; ++ struct scatterlist sgent; ++ struct scatterlist *sg; ++ struct scatterlist *cursg; ++ unsigned int cursg_ofs; ++ unsigned int err_mask; ++ struct ata_taskfile result_tf; ++ ata_qc_cb_t complete_fn; ++ void *private_data; ++ void *lldd_task; ++}; ++ ++struct ata_link; ++ ++typedef int (*ata_prereset_fn_t)(struct ata_link *, long unsigned int); ++ ++struct ata_eh_info { ++ struct ata_device *dev; ++ u32 serror; ++ unsigned int err_mask; ++ unsigned int action; ++ unsigned int dev_action[2]; ++ unsigned int flags; ++ unsigned int probe_mask; ++ char desc[80]; ++ int desc_len; ++}; ++ ++struct ata_eh_context { ++ struct ata_eh_info i; ++ int tries[2]; ++ int cmd_timeout_idx[14]; ++ unsigned int classes[2]; ++ unsigned int did_probe_mask; ++ unsigned int unloaded_mask; ++ unsigned int saved_ncq_enabled; ++ u8 saved_xfer_mode[2]; ++ long unsigned int last_reset; ++}; ++ ++struct ata_ering_entry { ++ unsigned int eflags; ++ unsigned int err_mask; ++ u64 timestamp; ++}; ++ ++struct ata_ering { ++ int cursor; ++ struct ata_ering_entry ring[32]; ++}; ++ ++struct ata_cpr_log; ++ ++struct ata_device { ++ struct ata_link *link; ++ unsigned int devno; ++ unsigned int horkage; ++ long unsigned int flags; ++ struct scsi_device *sdev; ++ void *private_data; ++ struct device tdev; ++ u64 n_sectors; ++ u64 n_native_sectors; ++ unsigned int class; ++ long unsigned int unpark_deadline; ++ u8 pio_mode; ++ u8 dma_mode; ++ u8 xfer_mode; ++ unsigned int xfer_shift; ++ unsigned int multi_count; ++ unsigned int max_sectors; ++ unsigned int cdb_len; ++ unsigned int pio_mask; ++ unsigned int mwdma_mask; ++ unsigned int udma_mask; ++ u16 cylinders; ++ u16 heads; ++ u16 sectors; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ union { ++ u16 id[256]; ++ u32 gscr[128]; ++ }; ++ u8 devslp_timing[8]; ++ u8 ncq_send_recv_cmds[20]; ++ u8 ncq_non_data_cmds[64]; ++ u32 zac_zoned_cap; ++ u32 zac_zones_optimal_open; ++ u32 zac_zones_optimal_nonseq; ++ u32 zac_zones_max_open; ++ struct ata_cpr_log *cpr_log; ++ int spdn_cnt; ++ struct ata_ering ering; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct ata_link { ++ struct ata_port *ap; ++ int pmp; ++ struct device tdev; ++ unsigned int active_tag; ++ u32 sactive; ++ unsigned int flags; ++ u32 saved_scontrol; ++ unsigned int hw_sata_spd_limit; ++ unsigned int sata_spd_limit; ++ unsigned int sata_spd; ++ enum ata_lpm_policy lpm_policy; ++ struct ata_eh_info eh_info; ++ struct ata_eh_context eh_context; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct ata_device device[2]; ++ long unsigned int last_lpm_change; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++typedef int (*ata_reset_fn_t)(struct ata_link *, unsigned int *, long unsigned int); ++ ++typedef void (*ata_postreset_fn_t)(struct ata_link *, unsigned int *); ++ ++enum sw_activity { ++ OFF = 0, ++ BLINK_ON = 1, ++ BLINK_OFF = 2, ++}; ++ ++struct ata_ioports { ++ void *cmd_addr; ++ void *data_addr; ++ void *error_addr; ++ void *feature_addr; ++ void *nsect_addr; ++ void *lbal_addr; ++ void *lbam_addr; ++ void *lbah_addr; ++ void *device_addr; ++ void *status_addr; ++ void *command_addr; ++ void *altstatus_addr; ++ void *ctl_addr; ++ void *bmdma_addr; ++ void *scr_addr; ++}; ++ ++struct ata_port_operations; ++ ++struct ata_host { ++ spinlock_t lock; ++ struct device *dev; ++ void * const *iomap; ++ unsigned int n_ports; ++ unsigned int n_tags; ++ void *private_data; ++ struct ata_port_operations *ops; ++ long unsigned int flags; ++ struct kref kref; ++ struct mutex eh_mutex; ++ struct task_struct *eh_owner; ++ struct ata_port *simplex_claimed; ++ struct ata_port *ports[0]; ++}; ++ ++struct ata_port_operations { ++ int (*qc_defer)(struct ata_queued_cmd *); ++ int (*check_atapi_dma)(struct ata_queued_cmd *); ++ enum ata_completion_errors (*qc_prep)(struct ata_queued_cmd *); ++ unsigned int (*qc_issue)(struct ata_queued_cmd *); ++ void (*qc_fill_rtf)(struct ata_queued_cmd *); ++ void (*qc_ncq_fill_rtf)(struct ata_port *, u64); ++ int (*cable_detect)(struct ata_port *); ++ unsigned int (*mode_filter)(struct ata_device *, unsigned int); ++ void (*set_piomode)(struct ata_port *, struct ata_device *); ++ void (*set_dmamode)(struct ata_port *, struct ata_device *); ++ int (*set_mode)(struct ata_link *, struct ata_device **); ++ unsigned int (*read_id)(struct ata_device *, struct ata_taskfile *, __le16 *); ++ void (*dev_config)(struct ata_device *); ++ void (*freeze)(struct ata_port *); ++ void (*thaw)(struct ata_port *); ++ ata_prereset_fn_t prereset; ++ ata_reset_fn_t softreset; ++ ata_reset_fn_t hardreset; ++ ata_postreset_fn_t postreset; ++ ata_prereset_fn_t pmp_prereset; ++ ata_reset_fn_t pmp_softreset; ++ ata_reset_fn_t pmp_hardreset; ++ ata_postreset_fn_t pmp_postreset; ++ void (*error_handler)(struct ata_port *); ++ void (*lost_interrupt)(struct ata_port *); ++ void (*post_internal_cmd)(struct ata_queued_cmd *); ++ void (*sched_eh)(struct ata_port *); ++ void (*end_eh)(struct ata_port *); ++ int (*scr_read)(struct ata_link *, unsigned int, u32 *); ++ int (*scr_write)(struct ata_link *, unsigned int, u32); ++ void (*pmp_attach)(struct ata_port *); ++ void (*pmp_detach)(struct ata_port *); ++ int (*set_lpm)(struct ata_link *, enum ata_lpm_policy, unsigned int); ++ int (*port_suspend)(struct ata_port *, pm_message_t); ++ int (*port_resume)(struct ata_port *); ++ int (*port_start)(struct ata_port *); ++ void (*port_stop)(struct ata_port *); ++ void (*host_stop)(struct ata_host *); ++ void (*sff_dev_select)(struct ata_port *, unsigned int); ++ void (*sff_set_devctl)(struct ata_port *, u8); ++ u8 (*sff_check_status)(struct ata_port *); ++ u8 (*sff_check_altstatus)(struct ata_port *); ++ void (*sff_tf_load)(struct ata_port *, const struct ata_taskfile *); ++ void (*sff_tf_read)(struct ata_port *, struct ata_taskfile *); ++ void (*sff_exec_command)(struct ata_port *, const struct ata_taskfile *); ++ unsigned int (*sff_data_xfer)(struct ata_queued_cmd *, unsigned char *, unsigned int, int); ++ void (*sff_irq_on)(struct ata_port *); ++ bool (*sff_irq_check)(struct ata_port *); ++ void (*sff_irq_clear)(struct ata_port *); ++ void (*sff_drain_fifo)(struct ata_queued_cmd *); ++ void (*bmdma_setup)(struct ata_queued_cmd *); ++ void (*bmdma_start)(struct ata_queued_cmd *); ++ void (*bmdma_stop)(struct ata_queued_cmd *); ++ u8 (*bmdma_status)(struct ata_port *); ++ ssize_t (*em_show)(struct ata_port *, char *); ++ ssize_t (*em_store)(struct ata_port *, const char *, size_t); ++ ssize_t (*sw_activity_show)(struct ata_device *, char *); ++ ssize_t (*sw_activity_store)(struct ata_device *, enum sw_activity); ++ ssize_t (*transmit_led_message)(struct ata_port *, u32, ssize_t); ++ void (*phy_reset)(struct ata_port *); ++ void (*eng_timeout)(struct ata_port *); ++ const struct ata_port_operations *inherits; ++}; ++ ++struct ata_port_stats { ++ long unsigned int unhandled_irq; ++ long unsigned int idle_irq; ++ long unsigned int rw_reqbuf; ++}; ++ ++struct ata_port { ++ struct Scsi_Host *scsi_host; ++ struct ata_port_operations *ops; ++ spinlock_t *lock; ++ long unsigned int flags; ++ unsigned int pflags; ++ unsigned int print_id; ++ unsigned int local_port_no; ++ unsigned int port_no; ++ struct ata_ioports ioaddr; ++ u8 ctl; ++ u8 last_ctl; ++ struct ata_link *sff_pio_task_link; ++ struct delayed_work sff_pio_task; ++ struct ata_bmdma_prd *bmdma_prd; ++ dma_addr_t bmdma_prd_dma; ++ unsigned int pio_mask; ++ unsigned int mwdma_mask; ++ unsigned int udma_mask; ++ unsigned int cbl; ++ struct ata_queued_cmd qcmd[33]; ++ u64 qc_active; ++ int nr_active_links; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct ata_link link; ++ struct ata_link *slave_link; ++ int nr_pmp_links; ++ struct ata_link *pmp_link; ++ struct ata_link *excl_link; ++ struct ata_port_stats stats; ++ struct ata_host *host; ++ struct device *dev; ++ struct device tdev; ++ struct mutex scsi_scan_mutex; ++ struct delayed_work hotplug_task; ++ struct delayed_work scsi_rescan_task; ++ unsigned int hsm_task_state; ++ struct list_head eh_done_q; ++ wait_queue_head_t eh_wait_q; ++ int eh_tries; ++ struct completion park_req_pending; ++ pm_message_t pm_mesg; ++ enum ata_lpm_policy target_lpm_policy; ++ struct timer_list fastdrain_timer; ++ unsigned int fastdrain_cnt; ++ async_cookie_t cookie; ++ int em_message_type; ++ void *private_data; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ u8 sector_buf[512]; ++}; ++ ++struct ata_cpr { ++ u8 num; ++ u8 num_storage_elements; ++ u64 start_lba; ++ u64 num_lbas; ++}; ++ ++struct ata_cpr_log { ++ u8 nr_cpr; ++ struct ata_cpr cpr[0]; ++}; ++ ++enum ata_link_iter_mode { ++ ATA_LITER_EDGE = 0, ++ ATA_LITER_HOST_FIRST = 1, ++ ATA_LITER_PMP_FIRST = 2, ++}; ++ ++enum ata_dev_iter_mode { ++ ATA_DITER_ENABLED = 0, ++ ATA_DITER_ENABLED_REVERSE = 1, ++ ATA_DITER_ALL = 2, ++ ATA_DITER_ALL_REVERSE = 3, ++}; ++ ++enum { ++ ATA_READID_POSTRESET = 1, ++ ATA_DNXFER_PIO = 0, ++ ATA_DNXFER_DMA = 1, ++ ATA_DNXFER_40C = 2, ++ ATA_DNXFER_FORCE_PIO = 3, ++ ATA_DNXFER_FORCE_PIO0 = 4, ++ ATA_DNXFER_QUIET = -2147483648, ++}; ++ ++enum { ++ ATA_EH_SPDN_NCQ_OFF = 1, ++ ATA_EH_SPDN_SPEED_DOWN = 2, ++ ATA_EH_SPDN_FALLBACK_TO_PIO = 4, ++ ATA_EH_SPDN_KEEP_ERRORS = 8, ++ ATA_EFLAG_IS_IO = 1, ++ ATA_EFLAG_DUBIOUS_XFER = 2, ++ ATA_EFLAG_OLD_ER = -2147483648, ++ ATA_ECAT_NONE = 0, ++ ATA_ECAT_ATA_BUS = 1, ++ ATA_ECAT_TOUT_HSM = 2, ++ ATA_ECAT_UNK_DEV = 3, ++ ATA_ECAT_DUBIOUS_NONE = 4, ++ ATA_ECAT_DUBIOUS_ATA_BUS = 5, ++ ATA_ECAT_DUBIOUS_TOUT_HSM = 6, ++ ATA_ECAT_DUBIOUS_UNK_DEV = 7, ++ ATA_ECAT_NR = 8, ++ ATA_EH_CMD_DFL_TIMEOUT = 5000, ++ ATA_EH_RESET_COOL_DOWN = 5000, ++ ATA_EH_PRERESET_TIMEOUT = 10000, ++ ATA_EH_FASTDRAIN_INTERVAL = 3000, ++ ATA_EH_UA_TRIES = 5, ++ ATA_EH_PROBE_TRIAL_INTERVAL = 60000, ++ ATA_EH_PROBE_TRIALS = 2, ++}; ++ ++struct ata_eh_cmd_timeout_ent { ++ const u8 *commands; ++ const unsigned int *timeouts; ++}; ++ ++struct speed_down_verdict_arg { ++ u64 since; ++ int xfer_ok; ++ int nr_errors[8]; ++}; ++ ++enum hsm_task_states { ++ HSM_ST_IDLE = 0, ++ HSM_ST_FIRST = 1, ++ HSM_ST = 2, ++ HSM_ST_LAST = 3, ++ HSM_ST_ERR = 4, ++}; ++ ++struct ata_port_info { ++ long unsigned int flags; ++ long unsigned int link_flags; ++ unsigned int pio_mask; ++ unsigned int mwdma_mask; ++ unsigned int udma_mask; ++ struct ata_port_operations *port_ops; ++ void *private_data; ++}; ++ ++enum dmi_field { ++ DMI_NONE = 0, ++ DMI_BIOS_VENDOR = 1, ++ DMI_BIOS_VERSION = 2, ++ DMI_BIOS_DATE = 3, ++ DMI_BIOS_RELEASE = 4, ++ DMI_EC_FIRMWARE_RELEASE = 5, ++ DMI_SYS_VENDOR = 6, ++ DMI_PRODUCT_NAME = 7, ++ DMI_PRODUCT_VERSION = 8, ++ DMI_PRODUCT_SERIAL = 9, ++ DMI_PRODUCT_UUID = 10, ++ DMI_PRODUCT_SKU = 11, ++ DMI_PRODUCT_FAMILY = 12, ++ DMI_BOARD_VENDOR = 13, ++ DMI_BOARD_NAME = 14, ++ DMI_BOARD_VERSION = 15, ++ DMI_BOARD_SERIAL = 16, ++ DMI_BOARD_ASSET_TAG = 17, ++ DMI_CHASSIS_VENDOR = 18, ++ DMI_CHASSIS_TYPE = 19, ++ DMI_CHASSIS_VERSION = 20, ++ DMI_CHASSIS_SERIAL = 21, ++ DMI_CHASSIS_ASSET_TAG = 22, ++ DMI_STRING_MAX = 23, ++ DMI_OEM_STRING = 24, ++}; ++ ++struct dmi_strmatch { ++ unsigned char slot: 7; ++ unsigned char exact_match: 1; ++ char substr[79]; ++}; ++ ++struct dmi_system_id { ++ int (*callback)(const struct dmi_system_id *); ++ const char *ident; ++ struct dmi_strmatch matches[4]; ++ void *driver_data; ++}; ++ ++struct pci_bits { ++ unsigned int reg; ++ unsigned int width; ++ long unsigned int mask; ++ long unsigned int val; ++}; ++ ++enum { ++ PIIX_IOCFG = 84, ++ ICH5_PMR = 144, ++ ICH5_PCS = 146, ++ PIIX_SIDPR_BAR = 5, ++ PIIX_SIDPR_LEN = 16, ++ PIIX_SIDPR_IDX = 0, ++ PIIX_SIDPR_DATA = 4, ++ PIIX_FLAG_CHECKINTR = 268435456, ++ PIIX_FLAG_SIDPR = 536870912, ++ PIIX_PATA_FLAGS = 1, ++ PIIX_SATA_FLAGS = 268435458, ++ PIIX_FLAG_PIO16 = 1073741824, ++ PIIX_80C_PRI = 48, ++ PIIX_80C_SEC = 192, ++ P0 = 0, ++ P1 = 1, ++ P2 = 2, ++ P3 = 3, ++ IDE = -1, ++ NA = -2, ++ RV = -3, ++ PIIX_AHCI_DEVICE = 6, ++ PIIX_HOST_BROKEN_SUSPEND = 16777216, ++}; ++ ++enum piix_controller_ids { ++ piix_pata_mwdma = 0, ++ piix_pata_33 = 1, ++ ich_pata_33 = 2, ++ ich_pata_66 = 3, ++ ich_pata_100 = 4, ++ ich_pata_100_nomwdma1 = 5, ++ ich5_sata = 6, ++ ich6_sata = 7, ++ ich6m_sata = 8, ++ ich8_sata = 9, ++ ich8_2port_sata = 10, ++ ich8m_apple_sata = 11, ++ tolapai_sata = 12, ++ piix_pata_vmw = 13, ++ ich8_sata_snb = 14, ++ ich8_2port_sata_snb = 15, ++ ich8_2port_sata_byt = 16, ++}; ++ ++struct piix_map_db { ++ const u32 mask; ++ const u16 port_enable; ++ const int map[0]; ++}; ++ ++struct piix_host_priv { ++ const int *map; ++ u32 saved_iocfg; ++ void *sidpr; ++}; ++ ++struct ich_laptop { ++ u16 device; ++ u16 subvendor; ++ u16 subdevice; ++}; ++ ++struct reset_control; ++ ++struct mii_bus; ++ ++struct mdio_device { ++ struct device dev; ++ struct mii_bus *bus; ++ char modalias[32]; ++ int (*bus_match)(struct device *, struct device_driver *); ++ void (*device_free)(struct mdio_device *); ++ void (*device_remove)(struct mdio_device *); ++ int addr; ++ int flags; ++ struct gpio_desc *reset_gpio; ++ struct reset_control *reset_ctrl; ++ unsigned int reset_assert_delay; ++ unsigned int reset_deassert_delay; ++}; ++ ++struct mdio_bus_stats { ++ u64_stats_t transfers; ++ u64_stats_t errors; ++ u64_stats_t writes; ++ u64_stats_t reads; ++ struct u64_stats_sync syncp; ++}; ++ ++struct phy_package_shared; ++ ++struct mii_bus { ++ struct module *owner; ++ const char *name; ++ char id[61]; ++ void *priv; ++ int (*read)(struct mii_bus *, int, int); ++ int (*write)(struct mii_bus *, int, int, u16); ++ int (*read_c45)(struct mii_bus *, int, int, int); ++ int (*write_c45)(struct mii_bus *, int, int, int, u16); ++ int (*reset)(struct mii_bus *); ++ struct mdio_bus_stats stats[32]; ++ struct mutex mdio_lock; ++ struct device *parent; ++ enum { ++ MDIOBUS_ALLOCATED = 1, ++ MDIOBUS_REGISTERED = 2, ++ MDIOBUS_UNREGISTERED = 3, ++ MDIOBUS_RELEASED = 4, ++ } state; ++ struct device dev; ++ struct mdio_device *mdio_map[32]; ++ u32 phy_mask; ++ u32 phy_ignore_ta_mask; ++ int irq[32]; ++ int reset_delay_us; ++ int reset_post_delay_us; ++ struct gpio_desc *reset_gpiod; ++ struct mutex shared_lock; ++ struct phy_package_shared *shared[32]; ++}; ++ ++struct phy_package_shared { ++ int addr; ++ refcount_t refcnt; ++ long unsigned int flags; ++ size_t priv_size; ++ void *priv; ++}; ++ ++struct mdio_board_info { ++ const char *bus_id; ++ char modalias[32]; ++ int mdio_addr; ++ const void *platform_data; ++}; ++ ++struct mdio_board_entry { ++ struct list_head list; ++ struct mdio_board_info board_info; ++}; ++ ++enum device_link_state { ++ DL_STATE_NONE = -1, ++ DL_STATE_DORMANT = 0, ++ DL_STATE_AVAILABLE = 1, ++ DL_STATE_CONSUMER_PROBE = 2, ++ DL_STATE_ACTIVE = 3, ++ DL_STATE_SUPPLIER_UNBIND = 4, ++}; ++ ++struct device_link { ++ struct device *supplier; ++ struct list_head s_node; ++ struct device *consumer; ++ struct list_head c_node; ++ struct device link_dev; ++ enum device_link_state status; ++ u32 flags; ++ refcount_t rpm_active; ++ struct kref kref; ++ struct work_struct rm_work; ++ bool supplier_preactivated; ++}; ++ ++struct phy_c45_device_ids { ++ u32 devices_in_package; ++ u32 mmds_present; ++ u32 device_ids[32]; ++}; ++ ++enum phy_state { ++ PHY_DOWN = 0, ++ PHY_READY = 1, ++ PHY_HALTED = 2, ++ PHY_UP = 3, ++ PHY_RUNNING = 4, ++ PHY_NOLINK = 5, ++ PHY_CABLETEST = 6, ++}; ++ ++typedef enum { ++ PHY_INTERFACE_MODE_NA = 0, ++ PHY_INTERFACE_MODE_INTERNAL = 1, ++ PHY_INTERFACE_MODE_MII = 2, ++ PHY_INTERFACE_MODE_GMII = 3, ++ PHY_INTERFACE_MODE_SGMII = 4, ++ PHY_INTERFACE_MODE_TBI = 5, ++ PHY_INTERFACE_MODE_REVMII = 6, ++ PHY_INTERFACE_MODE_RMII = 7, ++ PHY_INTERFACE_MODE_REVRMII = 8, ++ PHY_INTERFACE_MODE_RGMII = 9, ++ PHY_INTERFACE_MODE_RGMII_ID = 10, ++ PHY_INTERFACE_MODE_RGMII_RXID = 11, ++ PHY_INTERFACE_MODE_RGMII_TXID = 12, ++ PHY_INTERFACE_MODE_RTBI = 13, ++ PHY_INTERFACE_MODE_SMII = 14, ++ PHY_INTERFACE_MODE_XGMII = 15, ++ PHY_INTERFACE_MODE_XLGMII = 16, ++ PHY_INTERFACE_MODE_MOCA = 17, ++ PHY_INTERFACE_MODE_QSGMII = 18, ++ PHY_INTERFACE_MODE_TRGMII = 19, ++ PHY_INTERFACE_MODE_100BASEX = 20, ++ PHY_INTERFACE_MODE_1000BASEX = 21, ++ PHY_INTERFACE_MODE_2500BASEX = 22, ++ PHY_INTERFACE_MODE_5GBASER = 23, ++ PHY_INTERFACE_MODE_RXAUI = 24, ++ PHY_INTERFACE_MODE_XAUI = 25, ++ PHY_INTERFACE_MODE_10GBASER = 26, ++ PHY_INTERFACE_MODE_25GBASER = 27, ++ PHY_INTERFACE_MODE_USXGMII = 28, ++ PHY_INTERFACE_MODE_10GKR = 29, ++ PHY_INTERFACE_MODE_QUSGMII = 30, ++ PHY_INTERFACE_MODE_1000BASEKX = 31, ++ PHY_INTERFACE_MODE_MAX = 32, ++} phy_interface_t; ++ ++struct phylink; ++ ++struct pse_control; ++ ++struct phy_driver; ++ ++struct mii_timestamper; ++ ++struct phy_device { ++ struct mdio_device mdio; ++ struct phy_driver *drv; ++ struct device_link *devlink; ++ u32 phy_id; ++ struct phy_c45_device_ids c45_ids; ++ unsigned int is_c45: 1; ++ unsigned int is_internal: 1; ++ unsigned int is_pseudo_fixed_link: 1; ++ unsigned int is_gigabit_capable: 1; ++ unsigned int has_fixups: 1; ++ unsigned int suspended: 1; ++ unsigned int suspended_by_mdio_bus: 1; ++ unsigned int sysfs_links: 1; ++ unsigned int loopback_enabled: 1; ++ unsigned int downshifted_rate: 1; ++ unsigned int is_on_sfp_module: 1; ++ unsigned int mac_managed_pm: 1; ++ unsigned int autoneg: 1; ++ unsigned int link: 1; ++ unsigned int autoneg_complete: 1; ++ unsigned int interrupts: 1; ++ unsigned int irq_suspended: 1; ++ unsigned int irq_rerun: 1; ++ int rate_matching; ++ enum phy_state state; ++ u32 dev_flags; ++ phy_interface_t interface; ++ int speed; ++ int duplex; ++ int port; ++ int pause; ++ int asym_pause; ++ u8 master_slave_get; ++ u8 master_slave_set; ++ u8 master_slave_state; ++ long unsigned int supported[2]; ++ long unsigned int advertising[2]; ++ long unsigned int lp_advertising[2]; ++ long unsigned int adv_old[2]; ++ long unsigned int supported_eee[2]; ++ long unsigned int advertising_eee[2]; ++ bool eee_enabled; ++ long unsigned int host_interfaces[1]; ++ u32 eee_broken_modes; ++ struct list_head leds; ++ int irq; ++ void *priv; ++ struct phy_package_shared *shared; ++ struct sk_buff *skb; ++ void *ehdr; ++ struct nlattr *nest; ++ struct delayed_work state_queue; ++ struct mutex lock; ++ bool sfp_bus_attached; ++ struct sfp_bus *sfp_bus; ++ struct phylink *phylink; ++ struct net_device *attached_dev; ++ struct mii_timestamper *mii_ts; ++ struct pse_control *psec; ++ u8 mdix; ++ u8 mdix_ctrl; ++ int pma_extable; ++ unsigned int link_down_events; ++ void (*phy_link_change)(struct phy_device *, bool); ++ void (*adjust_link)(struct net_device *); ++ const struct macsec_ops *macsec_ops; ++}; ++ ++struct phy_plca_cfg { ++ int version; ++ int enabled; ++ int node_id; ++ int node_cnt; ++ int to_tmr; ++ int burst_cnt; ++ int burst_tmr; ++}; ++ ++struct phy_plca_status { ++ bool pst; ++}; ++ ++struct phy_tdr_config { ++ u32 first; ++ u32 last; ++ u32 step; ++ s8 pair; ++}; ++ ++struct mdio_driver_common { ++ struct device_driver driver; ++ int flags; ++}; ++ ++enum led_brightness { ++ LED_OFF = 0, ++ LED_ON = 1, ++ LED_HALF = 127, ++ LED_FULL = 255, ++}; ++ ++struct mii_timestamper { ++ bool (*rxtstamp)(struct mii_timestamper *, struct sk_buff *, int); ++ void (*txtstamp)(struct mii_timestamper *, struct sk_buff *, int); ++ int (*hwtstamp)(struct mii_timestamper *, struct ifreq *); ++ void (*link_state)(struct mii_timestamper *, struct phy_device *); ++ int (*ts_info)(struct mii_timestamper *, struct ethtool_ts_info *); ++ struct device *device; ++}; ++ ++struct phy_driver { ++ struct mdio_driver_common mdiodrv; ++ u32 phy_id; ++ char *name; ++ u32 phy_id_mask; ++ const long unsigned int * const features; ++ u32 flags; ++ const void *driver_data; ++ int (*soft_reset)(struct phy_device *); ++ int (*config_init)(struct phy_device *); ++ int (*probe)(struct phy_device *); ++ int (*get_features)(struct phy_device *); ++ int (*get_rate_matching)(struct phy_device *, phy_interface_t); ++ int (*suspend)(struct phy_device *); ++ int (*resume)(struct phy_device *); ++ int (*config_aneg)(struct phy_device *); ++ int (*aneg_done)(struct phy_device *); ++ int (*read_status)(struct phy_device *); ++ int (*config_intr)(struct phy_device *); ++ irqreturn_t (*handle_interrupt)(struct phy_device *); ++ void (*remove)(struct phy_device *); ++ int (*match_phy_device)(struct phy_device *); ++ int (*set_wol)(struct phy_device *, struct ethtool_wolinfo *); ++ void (*get_wol)(struct phy_device *, struct ethtool_wolinfo *); ++ void (*link_change_notify)(struct phy_device *); ++ int (*read_mmd)(struct phy_device *, int, u16); ++ int (*write_mmd)(struct phy_device *, int, u16, u16); ++ int (*read_page)(struct phy_device *); ++ int (*write_page)(struct phy_device *, int); ++ int (*module_info)(struct phy_device *, struct ethtool_modinfo *); ++ int (*module_eeprom)(struct phy_device *, struct ethtool_eeprom *, u8 *); ++ int (*cable_test_start)(struct phy_device *); ++ int (*cable_test_tdr_start)(struct phy_device *, const struct phy_tdr_config *); ++ int (*cable_test_get_status)(struct phy_device *, bool *); ++ int (*get_sset_count)(struct phy_device *); ++ void (*get_strings)(struct phy_device *, u8 *); ++ void (*get_stats)(struct phy_device *, struct ethtool_stats *, u64 *); ++ int (*get_tunable)(struct phy_device *, struct ethtool_tunable *, void *); ++ int (*set_tunable)(struct phy_device *, struct ethtool_tunable *, const void *); ++ int (*set_loopback)(struct phy_device *, bool); ++ int (*get_sqi)(struct phy_device *); ++ int (*get_sqi_max)(struct phy_device *); ++ int (*get_plca_cfg)(struct phy_device *, struct phy_plca_cfg *); ++ int (*set_plca_cfg)(struct phy_device *, const struct phy_plca_cfg *); ++ int (*get_plca_status)(struct phy_device *, struct phy_plca_status *); ++ int (*led_brightness_set)(struct phy_device *, u8, enum led_brightness); ++ int (*led_blink_set)(struct phy_device *, u8, long unsigned int *, long unsigned int *); ++}; ++ ++struct phy_setting { ++ u32 speed; ++ u8 duplex; ++ u8 bit; ++}; ++ ++struct mdio_driver { ++ struct mdio_driver_common mdiodrv; ++ int (*probe)(struct mdio_device *); ++ void (*remove)(struct mdio_device *); ++ void (*shutdown)(struct mdio_device *); ++}; ++ ++struct fixed_phy_status { ++ int link; ++ int speed; ++ int duplex; ++ int pause; ++ int asym_pause; ++}; ++ ++struct swmii_regs { ++ u16 bmsr; ++ u16 lpa; ++ u16 lpagb; ++ u16 estat; ++}; ++ ++enum { ++ SWMII_SPEED_10 = 0, ++ SWMII_SPEED_100 = 1, ++ SWMII_SPEED_1000 = 2, ++ SWMII_DUPLEX_HALF = 0, ++ SWMII_DUPLEX_FULL = 1, ++}; ++ ++struct mii_timestamping_ctrl { ++ struct mii_timestamper * (*probe_channel)(struct device *, unsigned int); ++ void (*release_channel)(struct device *, struct mii_timestamper *); ++}; ++ ++struct mii_timestamping_desc { ++ struct list_head list; ++ struct mii_timestamping_ctrl *ctrl; ++ struct device *device; ++}; ++ ++struct property_entry; ++ ++struct platform_device_info { ++ struct device *parent; ++ struct fwnode_handle *fwnode; ++ bool of_node_reused; ++ const char *name; ++ int id; ++ const struct resource *res; ++ unsigned int num_res; ++ const void *data; ++ size_t size_data; ++ u64 dma_mask; ++ const struct property_entry *properties; ++}; ++ ++enum dev_prop_type { ++ DEV_PROP_U8 = 0, ++ DEV_PROP_U16 = 1, ++ DEV_PROP_U32 = 2, ++ DEV_PROP_U64 = 3, ++ DEV_PROP_STRING = 4, ++ DEV_PROP_REF = 5, ++}; ++ ++struct property_entry { ++ const char *name; ++ size_t length; ++ bool is_inline; ++ enum dev_prop_type type; ++ union { ++ const void *pointer; ++ union { ++ u8 u8_data[8]; ++ u16 u16_data[4]; ++ u32 u32_data[2]; ++ u64 u64_data[1]; ++ const char *str[1]; ++ } value; ++ }; ++}; ++ ++struct fixed_mdio_bus { ++ struct mii_bus *mii_bus; ++ struct list_head phys; ++}; ++ ++struct fixed_phy { ++ int addr; ++ struct phy_device *phydev; ++ struct fixed_phy_status status; ++ bool no_carrier; ++ int (*link_update)(struct net_device *, struct fixed_phy_status *); ++ struct list_head node; ++ struct gpio_desc *link_gpiod; ++}; ++ ++struct input_mt_slot { ++ int abs[14]; ++ unsigned int frame; ++ unsigned int key; ++}; ++ ++struct input_mt { ++ int trkid; ++ int num_slots; ++ int slot; ++ unsigned int flags; ++ unsigned int frame; ++ int *red; ++ struct input_mt_slot slots[0]; ++}; ++ ++union input_seq_state { ++ struct { ++ short unsigned int pos; ++ bool mutex_acquired; ++ }; ++ void *p; ++}; ++ ++struct input_devres { ++ struct input_dev *input; ++}; ++ ++struct input_mt_pos { ++ s16 x; ++ s16 y; ++}; ++ ++struct touchscreen_properties { ++ unsigned int max_x; ++ unsigned int max_y; ++ bool invert_x; ++ bool invert_y; ++ bool swap_x_y; ++}; ++ ++struct i2c_device_id { ++ char name[20]; ++ kernel_ulong_t driver_data; ++}; ++ ++struct software_node { ++ const char *name; ++ const struct software_node *parent; ++ const struct property_entry *properties; ++}; ++ ++enum gpiod_flags { ++ GPIOD_ASIS = 0, ++ GPIOD_IN = 1, ++ GPIOD_OUT_LOW = 3, ++ GPIOD_OUT_HIGH = 7, ++ GPIOD_OUT_LOW_OPEN_DRAIN = 11, ++ GPIOD_OUT_HIGH_OPEN_DRAIN = 15, ++}; ++ ++struct i2c_client { ++ short unsigned int flags; ++ short unsigned int addr; ++ char name[20]; ++ struct i2c_adapter *adapter; ++ struct device dev; ++ int init_irq; ++ int irq; ++ struct list_head detected; ++ void *devres_group_id; ++}; ++ ++struct i2c_device_identity { ++ u16 manufacturer_id; ++ u16 part_id; ++ u8 die_revision; ++}; ++ ++enum i2c_alert_protocol { ++ I2C_PROTOCOL_SMBUS_ALERT = 0, ++ I2C_PROTOCOL_SMBUS_HOST_NOTIFY = 1, ++}; ++ ++struct i2c_board_info; ++ ++struct i2c_driver { ++ unsigned int class; ++ union { ++ int (*probe)(struct i2c_client *); ++ int (*probe_new)(struct i2c_client *); ++ }; ++ void (*remove)(struct i2c_client *); ++ void (*shutdown)(struct i2c_client *); ++ void (*alert)(struct i2c_client *, enum i2c_alert_protocol, unsigned int); ++ int (*command)(struct i2c_client *, unsigned int, void *); ++ struct device_driver driver; ++ const struct i2c_device_id *id_table; ++ int (*detect)(struct i2c_client *, struct i2c_board_info *); ++ const short unsigned int *address_list; ++ struct list_head clients; ++ u32 flags; ++}; ++ ++struct i2c_board_info { ++ char type[20]; ++ short unsigned int flags; ++ short unsigned int addr; ++ const char *dev_name; ++ void *platform_data; ++ struct device_node *of_node; ++ struct fwnode_handle *fwnode; ++ const struct software_node *swnode; ++ const struct resource *resources; ++ unsigned int num_resources; ++ int irq; ++}; ++ ++struct i2c_timings { ++ u32 bus_freq_hz; ++ u32 scl_rise_ns; ++ u32 scl_fall_ns; ++ u32 scl_int_delay_ns; ++ u32 sda_fall_ns; ++ u32 sda_hold_ns; ++ u32 digital_filter_width_ns; ++ u32 analog_filter_cutoff_freq_hz; ++}; ++ ++struct i2c_devinfo { ++ struct list_head list; ++ int busnum; ++ struct i2c_board_info board_info; ++}; ++ ++struct trace_event_raw_i2c_write { ++ struct trace_entry ent; ++ int adapter_nr; ++ __u16 msg_nr; ++ __u16 addr; ++ __u16 flags; ++ __u16 len; ++ u32 __data_loc_buf; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_i2c_read { ++ struct trace_entry ent; ++ int adapter_nr; ++ __u16 msg_nr; ++ __u16 addr; ++ __u16 flags; ++ __u16 len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_i2c_reply { ++ struct trace_entry ent; ++ int adapter_nr; ++ __u16 msg_nr; ++ __u16 addr; ++ __u16 flags; ++ __u16 len; ++ u32 __data_loc_buf; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_i2c_result { ++ struct trace_entry ent; ++ int adapter_nr; ++ __u16 nr_msgs; ++ __s16 ret; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_i2c_write { ++ u32 buf; ++}; ++ ++struct trace_event_data_offsets_i2c_read {}; ++ ++struct trace_event_data_offsets_i2c_reply { ++ u32 buf; ++}; ++ ++struct trace_event_data_offsets_i2c_result {}; ++ ++typedef void (*btf_trace_i2c_write)(void *, const struct i2c_adapter *, const struct i2c_msg *, int); ++ ++typedef void (*btf_trace_i2c_read)(void *, const struct i2c_adapter *, const struct i2c_msg *, int); ++ ++typedef void (*btf_trace_i2c_reply)(void *, const struct i2c_adapter *, const struct i2c_msg *, int); ++ ++typedef void (*btf_trace_i2c_result)(void *, const struct i2c_adapter *, int, int); ++ ++struct i2c_cmd_arg { ++ unsigned int cmd; ++ void *arg; ++}; ++ ++struct acpi_device; ++ ++enum rc_proto { ++ RC_PROTO_UNKNOWN = 0, ++ RC_PROTO_OTHER = 1, ++ RC_PROTO_RC5 = 2, ++ RC_PROTO_RC5X_20 = 3, ++ RC_PROTO_RC5_SZ = 4, ++ RC_PROTO_JVC = 5, ++ RC_PROTO_SONY12 = 6, ++ RC_PROTO_SONY15 = 7, ++ RC_PROTO_SONY20 = 8, ++ RC_PROTO_NEC = 9, ++ RC_PROTO_NECX = 10, ++ RC_PROTO_NEC32 = 11, ++ RC_PROTO_SANYO = 12, ++ RC_PROTO_MCIR2_KBD = 13, ++ RC_PROTO_MCIR2_MSE = 14, ++ RC_PROTO_RC6_0 = 15, ++ RC_PROTO_RC6_6A_20 = 16, ++ RC_PROTO_RC6_6A_24 = 17, ++ RC_PROTO_RC6_6A_32 = 18, ++ RC_PROTO_RC6_MCE = 19, ++ RC_PROTO_SHARP = 20, ++ RC_PROTO_XMP = 21, ++ RC_PROTO_CEC = 22, ++ RC_PROTO_IMON = 23, ++ RC_PROTO_RCMM12 = 24, ++ RC_PROTO_RCMM24 = 25, ++ RC_PROTO_RCMM32 = 26, ++ RC_PROTO_XBOX_DVD = 27, ++ RC_PROTO_MAX = 27, ++}; ++ ++struct rc_map_table { ++ u64 scancode; ++ u32 keycode; ++}; ++ ++struct rc_map { ++ struct rc_map_table *scan; ++ unsigned int size; ++ unsigned int len; ++ unsigned int alloc; ++ enum rc_proto rc_proto; ++ const char *name; ++ spinlock_t lock; ++}; ++ ++struct rc_map_list { ++ struct list_head list; ++ struct rc_map map; ++}; ++ ++typedef u32 (*bpf_prog_run_fn)(const struct bpf_prog *, const void *); ++ ++enum rc_driver_type { ++ RC_DRIVER_SCANCODE = 0, ++ RC_DRIVER_IR_RAW = 1, ++ RC_DRIVER_IR_RAW_TX = 2, ++}; ++ ++struct rc_scancode_filter { ++ u32 data; ++ u32 mask; ++}; ++ ++struct ir_raw_event_ctrl; ++ ++struct rc_dev { ++ struct device dev; ++ bool managed_alloc; ++ const struct attribute_group *sysfs_groups[5]; ++ const char *device_name; ++ const char *input_phys; ++ struct input_id input_id; ++ const char *driver_name; ++ const char *map_name; ++ struct rc_map rc_map; ++ struct mutex lock; ++ unsigned int minor; ++ struct ir_raw_event_ctrl *raw; ++ struct input_dev *input_dev; ++ enum rc_driver_type driver_type; ++ bool idle; ++ bool encode_wakeup; ++ u64 allowed_protocols; ++ u64 enabled_protocols; ++ u64 allowed_wakeup_protocols; ++ enum rc_proto wakeup_protocol; ++ struct rc_scancode_filter scancode_filter; ++ struct rc_scancode_filter scancode_wakeup_filter; ++ u32 scancode_mask; ++ u32 users; ++ void *priv; ++ spinlock_t keylock; ++ bool keypressed; ++ long unsigned int keyup_jiffies; ++ struct timer_list timer_keyup; ++ struct timer_list timer_repeat; ++ u32 last_keycode; ++ enum rc_proto last_protocol; ++ u64 last_scancode; ++ u8 last_toggle; ++ u32 timeout; ++ u32 min_timeout; ++ u32 max_timeout; ++ u32 rx_resolution; ++ u32 tx_resolution; ++ struct device lirc_dev; ++ struct cdev lirc_cdev; ++ ktime_t gap_start; ++ spinlock_t lirc_fh_lock; ++ struct list_head lirc_fh; ++ bool registered; ++ int (*change_protocol)(struct rc_dev *, u64 *); ++ int (*open)(struct rc_dev *); ++ void (*close)(struct rc_dev *); ++ int (*s_tx_mask)(struct rc_dev *, u32); ++ int (*s_tx_carrier)(struct rc_dev *, u32); ++ int (*s_tx_duty_cycle)(struct rc_dev *, u32); ++ int (*s_rx_carrier_range)(struct rc_dev *, u32, u32); ++ int (*tx_ir)(struct rc_dev *, unsigned int *, unsigned int); ++ void (*s_idle)(struct rc_dev *, bool); ++ int (*s_wideband_receiver)(struct rc_dev *, int); ++ int (*s_carrier_report)(struct rc_dev *, int); ++ int (*s_filter)(struct rc_dev *, struct rc_scancode_filter *); ++ int (*s_wakeup_filter)(struct rc_dev *, struct rc_scancode_filter *); ++ int (*s_timeout)(struct rc_dev *, unsigned int); ++}; ++ ++struct ir_raw_event { ++ union { ++ u32 duration; ++ u32 carrier; ++ }; ++ u8 duty_cycle; ++ unsigned int pulse: 1; ++ unsigned int overflow: 1; ++ unsigned int timeout: 1; ++ unsigned int carrier_report: 1; ++}; ++ ++struct nec_dec { ++ int state; ++ unsigned int count; ++ u32 bits; ++ bool is_nec_x; ++ bool necx_repeat; ++}; ++ ++struct rc5_dec { ++ int state; ++ u32 bits; ++ unsigned int count; ++ bool is_rc5x; ++}; ++ ++struct rc6_dec { ++ int state; ++ u8 header; ++ u32 body; ++ bool toggle; ++ unsigned int count; ++ unsigned int wanted_bits; ++}; ++ ++struct sony_dec { ++ int state; ++ u32 bits; ++ unsigned int count; ++}; ++ ++struct jvc_dec { ++ int state; ++ u16 bits; ++ u16 old_bits; ++ unsigned int count; ++ bool first; ++ bool toggle; ++}; ++ ++struct sanyo_dec { ++ int state; ++ unsigned int count; ++ u64 bits; ++}; ++ ++struct sharp_dec { ++ int state; ++ unsigned int count; ++ u32 bits; ++ unsigned int pulse_len; ++}; ++ ++struct mce_kbd_dec { ++ spinlock_t keylock; ++ struct timer_list rx_timeout; ++ int state; ++ u8 header; ++ u32 body; ++ unsigned int count; ++ unsigned int wanted_bits; ++}; ++ ++struct xmp_dec { ++ int state; ++ unsigned int count; ++ u32 durations[16]; ++}; ++ ++struct imon_dec { ++ int state; ++ int count; ++ int last_chk; ++ unsigned int bits; ++ bool stick_keyboard; ++}; ++ ++struct rcmm_dec { ++ int state; ++ unsigned int count; ++ u32 bits; ++}; ++ ++struct ir_raw_event_ctrl { ++ struct list_head list; ++ struct task_struct *thread; ++ struct { ++ union { ++ struct __kfifo kfifo; ++ struct ir_raw_event *type; ++ const struct ir_raw_event *const_type; ++ char (*rectype)[0]; ++ struct ir_raw_event *ptr; ++ const struct ir_raw_event *ptr_const; ++ }; ++ struct ir_raw_event buf[512]; ++ } kfifo; ++ ktime_t last_event; ++ struct rc_dev *dev; ++ spinlock_t edge_spinlock; ++ struct timer_list edge_handle; ++ struct ir_raw_event prev_ev; ++ struct ir_raw_event this_ev; ++ u32 bpf_sample; ++ struct bpf_prog_array *progs; ++ struct nec_dec nec; ++ struct rc5_dec rc5; ++ struct rc6_dec rc6; ++ struct sony_dec sony; ++ struct jvc_dec jvc; ++ struct sanyo_dec sanyo; ++ struct sharp_dec sharp; ++ struct mce_kbd_dec mce_kbd; ++ struct xmp_dec xmp; ++ struct imon_dec imon; ++ struct rcmm_dec rcmm; ++}; ++ ++typedef u64 (*btf_bpf_rc_repeat)(u32 *); ++ ++typedef u64 (*btf_bpf_rc_keydown)(u32 *, u32, u64, u32); ++ ++typedef u64 (*btf_bpf_rc_pointer_rel)(u32 *, s32, s32); ++ ++struct system_device_crosststamp { ++ ktime_t device; ++ ktime_t sys_realtime; ++ ktime_t sys_monoraw; ++}; ++ ++struct posix_clock; ++ ++struct posix_clock_operations { ++ struct module *owner; ++ int (*clock_adjtime)(struct posix_clock *, struct __kernel_timex *); ++ int (*clock_gettime)(struct posix_clock *, struct timespec64 *); ++ int (*clock_getres)(struct posix_clock *, struct timespec64 *); ++ int (*clock_settime)(struct posix_clock *, const struct timespec64 *); ++ long int (*ioctl)(struct posix_clock *, unsigned int, long unsigned int); ++ int (*open)(struct posix_clock *, fmode_t); ++ __poll_t (*poll)(struct posix_clock *, struct file *, poll_table *); ++ int (*release)(struct posix_clock *); ++ ssize_t (*read)(struct posix_clock *, uint, char *, size_t); ++}; ++ ++struct posix_clock { ++ struct posix_clock_operations ops; ++ struct cdev cdev; ++ struct device *dev; ++ struct rw_semaphore rwsem; ++ bool zombie; ++}; ++ ++struct ptp_clock_time { ++ __s64 sec; ++ __u32 nsec; ++ __u32 reserved; ++}; ++ ++struct ptp_extts_request { ++ unsigned int index; ++ unsigned int flags; ++ unsigned int rsv[2]; ++}; ++ ++struct ptp_perout_request { ++ union { ++ struct ptp_clock_time start; ++ struct ptp_clock_time phase; ++ }; ++ struct ptp_clock_time period; ++ unsigned int index; ++ unsigned int flags; ++ union { ++ struct ptp_clock_time on; ++ unsigned int rsv[4]; ++ }; ++}; ++ ++enum ptp_pin_function { ++ PTP_PF_NONE = 0, ++ PTP_PF_EXTTS = 1, ++ PTP_PF_PEROUT = 2, ++ PTP_PF_PHYSYNC = 3, ++}; ++ ++struct ptp_pin_desc { ++ char name[64]; ++ unsigned int index; ++ unsigned int func; ++ unsigned int chan; ++ unsigned int rsv[5]; ++}; ++ ++struct ptp_extts_event { ++ struct ptp_clock_time t; ++ unsigned int index; ++ unsigned int flags; ++ unsigned int rsv[2]; ++}; ++ ++struct pps_ktime { ++ __s64 sec; ++ __s32 nsec; ++ __u32 flags; ++}; ++ ++struct pps_kparams { ++ int api_version; ++ int mode; ++ struct pps_ktime assert_off_tu; ++ struct pps_ktime clear_off_tu; ++}; ++ ++struct pps_device; ++ ++struct pps_source_info { ++ char name[32]; ++ char path[32]; ++ int mode; ++ void (*echo)(struct pps_device *, int, void *); ++ struct module *owner; ++ struct device *dev; ++}; ++ ++struct pps_device { ++ struct pps_source_info info; ++ struct pps_kparams params; ++ __u32 assert_sequence; ++ __u32 clear_sequence; ++ struct pps_ktime assert_tu; ++ struct pps_ktime clear_tu; ++ int current_mode; ++ unsigned int last_ev; ++ wait_queue_head_t queue; ++ unsigned int id; ++ const void *lookup_cookie; ++ struct cdev cdev; ++ struct device *dev; ++ struct fasync_struct *async_queue; ++ spinlock_t lock; ++}; ++ ++struct cyclecounter { ++ u64 (*read)(const struct cyclecounter *); ++ u64 mask; ++ u32 mult; ++ u32 shift; ++}; ++ ++struct timecounter { ++ const struct cyclecounter *cc; ++ u64 cycle_last; ++ u64 nsec; ++ u64 mask; ++ u64 frac; ++}; ++ ++struct ptp_clock_request { ++ enum { ++ PTP_CLK_REQ_EXTTS = 0, ++ PTP_CLK_REQ_PEROUT = 1, ++ PTP_CLK_REQ_PPS = 2, ++ } type; ++ union { ++ struct ptp_extts_request extts; ++ struct ptp_perout_request perout; ++ }; ++}; ++ ++struct ptp_system_timestamp { ++ struct timespec64 pre_ts; ++ struct timespec64 post_ts; ++}; ++ ++struct ptp_clock_info { ++ struct module *owner; ++ char name[32]; ++ s32 max_adj; ++ int n_alarm; ++ int n_ext_ts; ++ int n_per_out; ++ int n_pins; ++ int pps; ++ struct ptp_pin_desc *pin_config; ++ int (*adjfine)(struct ptp_clock_info *, long int); ++ int (*adjphase)(struct ptp_clock_info *, s32); ++ int (*adjtime)(struct ptp_clock_info *, s64); ++ int (*gettime64)(struct ptp_clock_info *, struct timespec64 *); ++ int (*gettimex64)(struct ptp_clock_info *, struct timespec64 *, struct ptp_system_timestamp *); ++ int (*getcrosststamp)(struct ptp_clock_info *, struct system_device_crosststamp *); ++ int (*settime64)(struct ptp_clock_info *, const struct timespec64 *); ++ int (*getcycles64)(struct ptp_clock_info *, struct timespec64 *); ++ int (*getcyclesx64)(struct ptp_clock_info *, struct timespec64 *, struct ptp_system_timestamp *); ++ int (*getcrosscycles)(struct ptp_clock_info *, struct system_device_crosststamp *); ++ int (*enable)(struct ptp_clock_info *, struct ptp_clock_request *, int); ++ int (*verify)(struct ptp_clock_info *, unsigned int, enum ptp_pin_function, unsigned int); ++ long int (*do_aux_work)(struct ptp_clock_info *); ++}; ++ ++struct timestamp_event_queue { ++ struct ptp_extts_event buf[128]; ++ int head; ++ int tail; ++ spinlock_t lock; ++}; ++ ++struct ptp_clock { ++ struct posix_clock clock; ++ struct device dev; ++ struct ptp_clock_info *info; ++ dev_t devid; ++ int index; ++ struct pps_device *pps_source; ++ long int dialed_frequency; ++ struct timestamp_event_queue tsevq; ++ struct mutex tsevq_mux; ++ struct mutex pincfg_mux; ++ wait_queue_head_t tsev_wq; ++ int defunct; ++ struct device_attribute *pin_dev_attr; ++ struct attribute **pin_attr; ++ struct attribute_group pin_attr_group; ++ const struct attribute_group *pin_attr_groups[2]; ++ struct kthread_worker *kworker; ++ struct kthread_delayed_work aux_work; ++ unsigned int max_vclocks; ++ unsigned int n_vclocks; ++ int *vclock_index; ++ struct mutex n_vclocks_mux; ++ bool is_virtual_clock; ++ bool has_cycles; ++}; ++ ++struct ptp_vclock { ++ struct ptp_clock *pclock; ++ struct ptp_clock_info info; ++ struct ptp_clock *clock; ++ struct hlist_node vclock_hash_node; ++ struct cyclecounter cc; ++ struct timecounter tc; ++ struct mutex lock; ++}; ++ ++enum { ++ POWER_SUPPLY_TECHNOLOGY_UNKNOWN = 0, ++ POWER_SUPPLY_TECHNOLOGY_NiMH = 1, ++ POWER_SUPPLY_TECHNOLOGY_LION = 2, ++ POWER_SUPPLY_TECHNOLOGY_LIPO = 3, ++ POWER_SUPPLY_TECHNOLOGY_LiFe = 4, ++ POWER_SUPPLY_TECHNOLOGY_NiCd = 5, ++ POWER_SUPPLY_TECHNOLOGY_LiMn = 6, ++}; ++ ++struct power_supply_maintenance_charge_table; ++ ++struct power_supply_battery_ocv_table; ++ ++struct power_supply_resistance_temp_table; ++ ++struct power_supply_vbat_ri_table; ++ ++struct power_supply_battery_info { ++ unsigned int technology; ++ int energy_full_design_uwh; ++ int charge_full_design_uah; ++ int voltage_min_design_uv; ++ int voltage_max_design_uv; ++ int tricklecharge_current_ua; ++ int precharge_current_ua; ++ int precharge_voltage_max_uv; ++ int charge_term_current_ua; ++ int charge_restart_voltage_uv; ++ int overvoltage_limit_uv; ++ int constant_charge_current_max_ua; ++ int constant_charge_voltage_max_uv; ++ struct power_supply_maintenance_charge_table *maintenance_charge; ++ int maintenance_charge_size; ++ int alert_low_temp_charge_current_ua; ++ int alert_low_temp_charge_voltage_uv; ++ int alert_high_temp_charge_current_ua; ++ int alert_high_temp_charge_voltage_uv; ++ int factory_internal_resistance_uohm; ++ int factory_internal_resistance_charging_uohm; ++ int ocv_temp[20]; ++ int temp_ambient_alert_min; ++ int temp_ambient_alert_max; ++ int temp_alert_min; ++ int temp_alert_max; ++ int temp_min; ++ int temp_max; ++ struct power_supply_battery_ocv_table *ocv_table[20]; ++ int ocv_table_size[20]; ++ struct power_supply_resistance_temp_table *resist_table; ++ int resist_table_size; ++ struct power_supply_vbat_ri_table *vbat2ri_discharging; ++ int vbat2ri_discharging_size; ++ struct power_supply_vbat_ri_table *vbat2ri_charging; ++ int vbat2ri_charging_size; ++ int bti_resistance_ohm; ++ int bti_resistance_tolerance; ++}; ++ ++struct power_supply_battery_ocv_table { ++ int ocv; ++ int capacity; ++}; ++ ++struct power_supply_resistance_temp_table { ++ int temp; ++ int resistance; ++}; ++ ++struct power_supply_vbat_ri_table { ++ int vbat_uv; ++ int ri_uohm; ++}; ++ ++struct power_supply_maintenance_charge_table { ++ int charge_current_max_ua; ++ int charge_voltage_max_uv; ++ int charge_safety_timer_minutes; ++}; ++ ++struct samsung_sdi_battery { ++ char *compatible; ++ char *name; ++ struct power_supply_battery_info info; ++}; ++ ++enum { ++ PERCPU_REF_INIT_ATOMIC = 1, ++ PERCPU_REF_INIT_DEAD = 2, ++ PERCPU_REF_ALLOW_REINIT = 4, ++}; ++ ++enum { ++ DISK_EVENT_MEDIA_CHANGE = 1, ++ DISK_EVENT_EJECT_REQUEST = 2, ++}; ++ ++struct hd_geometry { ++ unsigned char heads; ++ unsigned char sectors; ++ short unsigned int cylinders; ++ long unsigned int start; ++}; ++ ++struct mdp_device_descriptor_s { ++ __u32 number; ++ __u32 major; ++ __u32 minor; ++ __u32 raid_disk; ++ __u32 state; ++ __u32 reserved[27]; ++}; ++ ++typedef struct mdp_device_descriptor_s mdp_disk_t; ++ ++struct mdp_superblock_s { ++ __u32 md_magic; ++ __u32 major_version; ++ __u32 minor_version; ++ __u32 patch_version; ++ __u32 gvalid_words; ++ __u32 set_uuid0; ++ __u32 ctime; ++ __u32 level; ++ __u32 size; ++ __u32 nr_disks; ++ __u32 raid_disks; ++ __u32 md_minor; ++ __u32 not_persistent; ++ __u32 set_uuid1; ++ __u32 set_uuid2; ++ __u32 set_uuid3; ++ __u32 gstate_creserved[16]; ++ __u32 utime; ++ __u32 state; ++ __u32 active_disks; ++ __u32 working_disks; ++ __u32 failed_disks; ++ __u32 spare_disks; ++ __u32 sb_csum; ++ __u32 events_hi; ++ __u32 events_lo; ++ __u32 cp_events_hi; ++ __u32 cp_events_lo; ++ __u32 recovery_cp; ++ __u64 reshape_position; ++ __u32 new_level; ++ __u32 delta_disks; ++ __u32 new_layout; ++ __u32 new_chunk; ++ __u32 gstate_sreserved[14]; ++ __u32 layout; ++ __u32 chunk_size; ++ __u32 root_pv; ++ __u32 root_block; ++ __u32 pstate_reserved[60]; ++ mdp_disk_t disks[27]; ++ __u32 reserved[0]; ++ mdp_disk_t this_disk; ++}; ++ ++typedef struct mdp_superblock_s mdp_super_t; ++ ++struct mdp_superblock_1 { ++ __le32 magic; ++ __le32 major_version; ++ __le32 feature_map; ++ __le32 pad0; ++ __u8 set_uuid[16]; ++ char set_name[32]; ++ __le64 ctime; ++ __le32 level; ++ __le32 layout; ++ __le64 size; ++ __le32 chunksize; ++ __le32 raid_disks; ++ union { ++ __le32 bitmap_offset; ++ struct { ++ __le16 offset; ++ __le16 size; ++ } ppl; ++ }; ++ __le32 new_level; ++ __le64 reshape_position; ++ __le32 delta_disks; ++ __le32 new_layout; ++ __le32 new_chunk; ++ __le32 new_offset; ++ __le64 data_offset; ++ __le64 data_size; ++ __le64 super_offset; ++ union { ++ __le64 recovery_offset; ++ __le64 journal_tail; ++ }; ++ __le32 dev_number; ++ __le32 cnt_corrected_read; ++ __u8 device_uuid[16]; ++ __u8 devflags; ++ __u8 bblog_shift; ++ __le16 bblog_size; ++ __le32 bblog_offset; ++ __le64 utime; ++ __le64 events; ++ __le64 resync_offset; ++ __le32 sb_csum; ++ __le32 max_dev; ++ __u8 pad3[32]; ++ __le16 dev_roles[0]; ++}; ++ ++struct mdu_version_s { ++ int major; ++ int minor; ++ int patchlevel; ++}; ++ ++typedef struct mdu_version_s mdu_version_t; ++ ++struct mdu_array_info_s { ++ int major_version; ++ int minor_version; ++ int patch_version; ++ unsigned int ctime; ++ int level; ++ int size; ++ int nr_disks; ++ int raid_disks; ++ int md_minor; ++ int not_persistent; ++ unsigned int utime; ++ int state; ++ int active_disks; ++ int working_disks; ++ int failed_disks; ++ int spare_disks; ++ int layout; ++ int chunk_size; ++}; ++ ++typedef struct mdu_array_info_s mdu_array_info_t; ++ ++struct mdu_disk_info_s { ++ int number; ++ int major; ++ int minor; ++ int raid_disk; ++ int state; ++}; ++ ++typedef struct mdu_disk_info_s mdu_disk_info_t; ++ ++struct mdu_bitmap_file_s { ++ char pathname[4096]; ++}; ++ ++typedef struct mdu_bitmap_file_s mdu_bitmap_file_t; ++ ++struct mddev; ++ ++struct md_rdev; ++ ++struct md_cluster_operations { ++ int (*join)(struct mddev *, int); ++ int (*leave)(struct mddev *); ++ int (*slot_number)(struct mddev *); ++ int (*resync_info_update)(struct mddev *, sector_t, sector_t); ++ void (*resync_info_get)(struct mddev *, sector_t *, sector_t *); ++ int (*metadata_update_start)(struct mddev *); ++ int (*metadata_update_finish)(struct mddev *); ++ void (*metadata_update_cancel)(struct mddev *); ++ int (*resync_start)(struct mddev *); ++ int (*resync_finish)(struct mddev *); ++ int (*area_resyncing)(struct mddev *, int, sector_t, sector_t); ++ int (*add_new_disk)(struct mddev *, struct md_rdev *); ++ void (*add_new_disk_cancel)(struct mddev *); ++ int (*new_disk_ack)(struct mddev *, bool); ++ int (*remove_disk)(struct mddev *, struct md_rdev *); ++ void (*load_bitmaps)(struct mddev *, int); ++ int (*gather_bitmaps)(struct md_rdev *); ++ int (*resize_bitmaps)(struct mddev *, sector_t, sector_t); ++ int (*lock_all_bitmaps)(struct mddev *); ++ void (*unlock_all_bitmaps)(struct mddev *); ++ void (*update_size)(struct mddev *, sector_t); ++}; ++ ++struct md_cluster_info; ++ ++struct md_personality; ++ ++struct md_thread; ++ ++struct bitmap; ++ ++struct mddev { ++ void *private; ++ struct md_personality *pers; ++ dev_t unit; ++ int md_minor; ++ struct list_head disks; ++ long unsigned int flags; ++ long unsigned int sb_flags; ++ int suspended; ++ struct percpu_ref active_io; ++ int ro; ++ int sysfs_active; ++ struct gendisk *gendisk; ++ struct kobject kobj; ++ int hold_active; ++ int major_version; ++ int minor_version; ++ int patch_version; ++ int persistent; ++ int external; ++ char metadata_type[17]; ++ int chunk_sectors; ++ time64_t ctime; ++ time64_t utime; ++ int level; ++ int layout; ++ char clevel[16]; ++ int raid_disks; ++ int max_disks; ++ sector_t dev_sectors; ++ sector_t array_sectors; ++ int external_size; ++ __u64 events; ++ int can_decrease_events; ++ char uuid[16]; ++ sector_t reshape_position; ++ int delta_disks; ++ int new_level; ++ int new_layout; ++ int new_chunk_sectors; ++ int reshape_backwards; ++ struct md_thread *thread; ++ struct md_thread *sync_thread; ++ char *last_sync_action; ++ sector_t curr_resync; ++ sector_t curr_resync_completed; ++ long unsigned int resync_mark; ++ sector_t resync_mark_cnt; ++ sector_t curr_mark_cnt; ++ sector_t resync_max_sectors; ++ atomic64_t resync_mismatches; ++ sector_t suspend_lo; ++ sector_t suspend_hi; ++ int sync_speed_min; ++ int sync_speed_max; ++ int parallel_resync; ++ int ok_start_degraded; ++ long unsigned int recovery; ++ int recovery_disabled; ++ int in_sync; ++ struct mutex open_mutex; ++ struct mutex reconfig_mutex; ++ atomic_t active; ++ atomic_t openers; ++ int changed; ++ int degraded; ++ atomic_t recovery_active; ++ wait_queue_head_t recovery_wait; ++ sector_t recovery_cp; ++ sector_t resync_min; ++ sector_t resync_max; ++ struct kernfs_node *sysfs_state; ++ struct kernfs_node *sysfs_action; ++ struct kernfs_node *sysfs_completed; ++ struct kernfs_node *sysfs_degraded; ++ struct kernfs_node *sysfs_level; ++ struct work_struct del_work; ++ spinlock_t lock; ++ wait_queue_head_t sb_wait; ++ atomic_t pending_writes; ++ unsigned int safemode; ++ unsigned int safemode_delay; ++ struct timer_list safemode_timer; ++ struct percpu_ref writes_pending; ++ int sync_checkers; ++ struct request_queue *queue; ++ struct bitmap *bitmap; ++ struct { ++ struct file *file; ++ loff_t offset; ++ long unsigned int space; ++ loff_t default_offset; ++ long unsigned int default_space; ++ struct mutex mutex; ++ long unsigned int chunksize; ++ long unsigned int daemon_sleep; ++ long unsigned int max_write_behind; ++ int external; ++ int nodes; ++ char cluster_name[64]; ++ } bitmap_info; ++ atomic_t max_corr_read_errors; ++ struct list_head all_mddevs; ++ const struct attribute_group *to_remove; ++ struct bio_set bio_set; ++ struct bio_set sync_set; ++ struct bio_set io_acct_set; ++ struct bio *flush_bio; ++ atomic_t flush_pending; ++ ktime_t start_flush; ++ ktime_t prev_flush_start; ++ struct work_struct flush_work; ++ struct work_struct event_work; ++ mempool_t *serial_info_pool; ++ void (*sync_super)(struct mddev *, struct md_rdev *); ++ struct md_cluster_info *cluster_info; ++ unsigned int good_device_nr; ++ unsigned int noio_flag; ++ bool has_superblocks: 1; ++ bool fail_last_dev: 1; ++ bool serialize_policy: 1; ++}; ++ ++struct serial_in_rdev; ++ ++struct md_rdev { ++ struct list_head same_set; ++ sector_t sectors; ++ struct mddev *mddev; ++ int last_events; ++ struct block_device *meta_bdev; ++ struct block_device *bdev; ++ struct page *sb_page; ++ struct page *bb_page; ++ int sb_loaded; ++ __u64 sb_events; ++ sector_t data_offset; ++ sector_t new_data_offset; ++ sector_t sb_start; ++ int sb_size; ++ int preferred_minor; ++ struct kobject kobj; ++ long unsigned int flags; ++ wait_queue_head_t blocked_wait; ++ int desc_nr; ++ int raid_disk; ++ int new_raid_disk; ++ int saved_raid_disk; ++ union { ++ sector_t recovery_offset; ++ sector_t journal_tail; ++ }; ++ atomic_t nr_pending; ++ atomic_t read_errors; ++ time64_t last_read_error; ++ atomic_t corrected_errors; ++ struct serial_in_rdev *serial; ++ struct work_struct del_work; ++ struct kernfs_node *sysfs_state; ++ struct kernfs_node *sysfs_unack_badblocks; ++ struct kernfs_node *sysfs_badblocks; ++ struct badblocks badblocks; ++ struct { ++ short int offset; ++ unsigned int size; ++ sector_t sector; ++ } ppl; ++}; ++ ++struct serial_in_rdev { ++ struct rb_root_cached serial_rb; ++ spinlock_t serial_lock; ++ wait_queue_head_t serial_io_wait; ++}; ++ ++enum flag_bits { ++ Faulty = 0, ++ In_sync = 1, ++ Bitmap_sync = 2, ++ WriteMostly = 3, ++ AutoDetected = 4, ++ Blocked = 5, ++ WriteErrorSeen = 6, ++ FaultRecorded = 7, ++ BlockedBadBlocks = 8, ++ WantReplacement = 9, ++ Replacement = 10, ++ Candidate = 11, ++ Journal = 12, ++ ClusterRemove = 13, ++ RemoveSynchronized = 14, ++ ExternalBbl = 15, ++ FailFast = 16, ++ LastDev = 17, ++ CollisionCheck = 18, ++}; ++ ++enum mddev_flags { ++ MD_ARRAY_FIRST_USE = 0, ++ MD_CLOSING = 1, ++ MD_JOURNAL_CLEAN = 2, ++ MD_HAS_JOURNAL = 3, ++ MD_CLUSTER_RESYNC_LOCKED = 4, ++ MD_FAILFAST_SUPPORTED = 5, ++ MD_HAS_PPL = 6, ++ MD_HAS_MULTIPLE_PPLS = 7, ++ MD_ALLOW_SB_UPDATE = 8, ++ MD_UPDATING_SB = 9, ++ MD_NOT_READY = 10, ++ MD_BROKEN = 11, ++ MD_DELETED = 12, ++}; ++ ++enum mddev_sb_flags { ++ MD_SB_CHANGE_DEVS = 0, ++ MD_SB_CHANGE_CLEAN = 1, ++ MD_SB_CHANGE_PENDING = 2, ++ MD_SB_NEED_REWRITE = 3, ++}; ++ ++enum { ++ MD_RESYNC_NONE = 0, ++ MD_RESYNC_YIELDED = 1, ++ MD_RESYNC_DELAYED = 2, ++ MD_RESYNC_ACTIVE = 3, ++}; ++ ++struct md_personality { ++ char *name; ++ int level; ++ struct list_head list; ++ struct module *owner; ++ bool (*make_request)(struct mddev *, struct bio *); ++ int (*run)(struct mddev *); ++ int (*start)(struct mddev *); ++ void (*free)(struct mddev *, void *); ++ void (*status)(struct seq_file *, struct mddev *); ++ void (*error_handler)(struct mddev *, struct md_rdev *); ++ int (*hot_add_disk)(struct mddev *, struct md_rdev *); ++ int (*hot_remove_disk)(struct mddev *, struct md_rdev *); ++ int (*spare_active)(struct mddev *); ++ sector_t (*sync_request)(struct mddev *, sector_t, int *); ++ int (*resize)(struct mddev *, sector_t); ++ sector_t (*size)(struct mddev *, sector_t, int); ++ int (*check_reshape)(struct mddev *); ++ int (*start_reshape)(struct mddev *); ++ void (*finish_reshape)(struct mddev *); ++ void (*update_reshape_pos)(struct mddev *); ++ void (*quiesce)(struct mddev *, int); ++ void * (*takeover)(struct mddev *); ++ int (*change_consistency_policy)(struct mddev *, const char *); ++}; ++ ++struct md_thread { ++ void (*run)(struct md_thread *); ++ struct mddev *mddev; ++ wait_queue_head_t wqueue; ++ long unsigned int flags; ++ struct task_struct *tsk; ++ long unsigned int timeout; ++ void *private; ++}; ++ ++struct bitmap_page; ++ ++struct bitmap_counts { ++ spinlock_t lock; ++ struct bitmap_page *bp; ++ long unsigned int pages; ++ long unsigned int missing_pages; ++ long unsigned int chunkshift; ++ long unsigned int chunks; ++}; ++ ++struct bitmap_storage { ++ struct file *file; ++ struct page *sb_page; ++ struct page **filemap; ++ long unsigned int *filemap_attr; ++ long unsigned int file_pages; ++ long unsigned int bytes; ++}; ++ ++struct bitmap { ++ struct bitmap_counts counts; ++ struct mddev *mddev; ++ __u64 events_cleared; ++ int need_sync; ++ struct bitmap_storage storage; ++ long unsigned int flags; ++ int allclean; ++ atomic_t behind_writes; ++ long unsigned int behind_writes_used; ++ long unsigned int daemon_lastrun; ++ long unsigned int last_end_sync; ++ atomic_t pending_writes; ++ wait_queue_head_t write_wait; ++ wait_queue_head_t overflow_wait; ++ wait_queue_head_t behind_wait; ++ struct kernfs_node *sysfs_can_clear; ++ int cluster_slot; ++}; ++ ++enum recovery_flags { ++ MD_RECOVERY_RUNNING = 0, ++ MD_RECOVERY_SYNC = 1, ++ MD_RECOVERY_RECOVER = 2, ++ MD_RECOVERY_INTR = 3, ++ MD_RECOVERY_DONE = 4, ++ MD_RECOVERY_NEEDED = 5, ++ MD_RECOVERY_REQUESTED = 6, ++ MD_RECOVERY_CHECK = 7, ++ MD_RECOVERY_RESHAPE = 8, ++ MD_RECOVERY_FROZEN = 9, ++ MD_RECOVERY_ERROR = 10, ++ MD_RECOVERY_WAIT = 11, ++ MD_RESYNCING_REMOTE = 12, ++}; ++ ++struct md_sysfs_entry { ++ struct attribute attr; ++ ssize_t (*show)(struct mddev *, char *); ++ ssize_t (*store)(struct mddev *, const char *, size_t); ++}; ++ ++struct md_io_acct { ++ struct mddev *mddev; ++ struct bio *orig_bio; ++ long unsigned int start_time; ++ struct bio bio_clone; ++}; ++ ++struct bitmap_page { ++ char *map; ++ unsigned int hijacked: 1; ++ unsigned int pending: 1; ++ unsigned int count: 30; ++}; ++ ++enum md_ro_state { ++ MD_RDWR = 0, ++ MD_RDONLY = 1, ++ MD_AUTO_READ = 2, ++ MD_MAX_STATE = 3, ++}; ++ ++struct super_type { ++ char *name; ++ struct module *owner; ++ int (*load_super)(struct md_rdev *, struct md_rdev *, int); ++ int (*validate_super)(struct mddev *, struct md_rdev *); ++ void (*sync_super)(struct mddev *, struct md_rdev *); ++ long long unsigned int (*rdev_size_change)(struct md_rdev *, sector_t); ++ int (*allow_new_offset)(struct md_rdev *, long long unsigned int); ++}; ++ ++struct rdev_sysfs_entry { ++ struct attribute attr; ++ ssize_t (*show)(struct md_rdev *, char *); ++ ssize_t (*store)(struct md_rdev *, const char *, size_t); ++}; ++ ++enum array_state { ++ clear = 0, ++ inactive = 1, ++ suspended = 2, ++ readonly = 3, ++ read_auto = 4, ++ clean = 5, ++ active = 6, ++ write_pending = 7, ++ active_idle = 8, ++ broken = 9, ++ bad_word = 10, ++}; ++ ++struct detected_devices_node { ++ struct list_head list; ++ dev_t dev; ++}; ++ ++typedef u16 blk_short_t; ++ ++enum dm_queue_mode { ++ DM_TYPE_NONE = 0, ++ DM_TYPE_BIO_BASED = 1, ++ DM_TYPE_REQUEST_BASED = 2, ++ DM_TYPE_DAX_BIO_BASED = 3, ++}; ++ ++typedef enum { ++ STATUSTYPE_INFO = 0, ++ STATUSTYPE_TABLE = 1, ++ STATUSTYPE_IMA = 2, ++} status_type_t; ++ ++union map_info { ++ void *ptr; ++}; ++ ++struct dm_target; ++ ++typedef int (*dm_ctr_fn)(struct dm_target *, unsigned int, char **); ++ ++struct dm_table; ++ ++struct target_type; ++ ++struct dm_target { ++ struct dm_table *table; ++ struct target_type *type; ++ sector_t begin; ++ sector_t len; ++ uint32_t max_io_len; ++ unsigned int num_flush_bios; ++ unsigned int num_discard_bios; ++ unsigned int num_secure_erase_bios; ++ unsigned int num_write_zeroes_bios; ++ unsigned int per_io_data_size; ++ void *private; ++ char *error; ++ bool flush_supported: 1; ++ bool discards_supported: 1; ++ bool max_discard_granularity: 1; ++ bool max_secure_erase_granularity: 1; ++ bool max_write_zeroes_granularity: 1; ++ bool limit_swap_bios: 1; ++ bool emulate_zone_append: 1; ++ bool accounts_remapped_io: 1; ++ bool needs_bio_set_dev: 1; ++}; ++ ++typedef void (*dm_dtr_fn)(struct dm_target *); ++ ++typedef int (*dm_map_fn)(struct dm_target *, struct bio *); ++ ++typedef int (*dm_clone_and_map_request_fn)(struct dm_target *, struct request *, union map_info *, struct request **); ++ ++typedef void (*dm_release_clone_request_fn)(struct request *, union map_info *); ++ ++typedef int (*dm_endio_fn)(struct dm_target *, struct bio *, blk_status_t *); ++ ++typedef int (*dm_request_endio_fn)(struct dm_target *, struct request *, blk_status_t, union map_info *); ++ ++typedef void (*dm_presuspend_fn)(struct dm_target *); ++ ++typedef void (*dm_presuspend_undo_fn)(struct dm_target *); ++ ++typedef void (*dm_postsuspend_fn)(struct dm_target *); ++ ++typedef int (*dm_preresume_fn)(struct dm_target *); ++ ++typedef void (*dm_resume_fn)(struct dm_target *); ++ ++typedef void (*dm_status_fn)(struct dm_target *, status_type_t, unsigned int, char *, unsigned int); ++ ++typedef int (*dm_message_fn)(struct dm_target *, unsigned int, char **, char *, unsigned int); ++ ++typedef int (*dm_prepare_ioctl_fn)(struct dm_target *, struct block_device **); ++ ++struct dm_report_zones_args; ++ ++typedef int (*dm_report_zones_fn)(struct dm_target *, struct dm_report_zones_args *, unsigned int); ++ ++struct dm_report_zones_args { ++ struct dm_target *tgt; ++ sector_t next_sector; ++ void *orig_data; ++ report_zones_cb orig_cb; ++ unsigned int zone_idx; ++ sector_t start; ++}; ++ ++struct dm_dev; ++ ++typedef int (*iterate_devices_callout_fn)(struct dm_target *, struct dm_dev *, sector_t, sector_t, void *); ++ ++struct dm_dev { ++ struct block_device *bdev; ++ struct dax_device *dax_dev; ++ fmode_t mode; ++ char name[16]; ++}; ++ ++typedef int (*dm_iterate_devices_fn)(struct dm_target *, iterate_devices_callout_fn, void *); ++ ++typedef void (*dm_io_hints_fn)(struct dm_target *, struct queue_limits *); ++ ++typedef int (*dm_busy_fn)(struct dm_target *); ++ ++typedef long int (*dm_dax_direct_access_fn)(struct dm_target *, long unsigned int, long int, enum dax_access_mode, void **, pfn_t *); ++ ++typedef int (*dm_dax_zero_page_range_fn)(struct dm_target *, long unsigned int, size_t); ++ ++typedef size_t (*dm_dax_recovery_write_fn)(struct dm_target *, long unsigned int, void *, size_t, struct iov_iter *); ++ ++struct target_type { ++ uint64_t features; ++ const char *name; ++ struct module *module; ++ unsigned int version[3]; ++ dm_ctr_fn ctr; ++ dm_dtr_fn dtr; ++ dm_map_fn map; ++ dm_clone_and_map_request_fn clone_and_map_rq; ++ dm_release_clone_request_fn release_clone_rq; ++ dm_endio_fn end_io; ++ dm_request_endio_fn rq_end_io; ++ dm_presuspend_fn presuspend; ++ dm_presuspend_undo_fn presuspend_undo; ++ dm_postsuspend_fn postsuspend; ++ dm_preresume_fn preresume; ++ dm_resume_fn resume; ++ dm_status_fn status; ++ dm_message_fn message; ++ dm_prepare_ioctl_fn prepare_ioctl; ++ dm_report_zones_fn report_zones; ++ dm_busy_fn busy; ++ dm_iterate_devices_fn iterate_devices; ++ dm_io_hints_fn io_hints; ++ dm_dax_direct_access_fn direct_access; ++ dm_dax_zero_page_range_fn dax_zero_page_range; ++ dm_dax_recovery_write_fn dax_recovery_write; ++ struct list_head list; ++}; ++ ++struct mapped_device; ++ ++struct dm_md_mempools; ++ ++struct dm_table { ++ struct mapped_device *md; ++ enum dm_queue_mode type; ++ unsigned int depth; ++ unsigned int counts[16]; ++ sector_t *index[16]; ++ unsigned int num_targets; ++ unsigned int num_allocated; ++ sector_t *highs; ++ struct dm_target *targets; ++ struct target_type *immutable_target_type; ++ bool integrity_supported: 1; ++ bool singleton: 1; ++ unsigned int integrity_added: 1; ++ fmode_t mode; ++ struct list_head devices; ++ void (*event_fn)(void *); ++ void *event_context; ++ struct dm_md_mempools *mempools; ++ struct blk_crypto_profile *crypto_profile; ++}; ++ ++struct dm_stats_last_position; ++ ++struct dm_stats { ++ struct mutex mutex; ++ struct list_head list; ++ struct dm_stats_last_position *last; ++ bool precise_timestamps; ++}; ++ ++struct dm_stats_aux { ++ bool merged; ++ long long unsigned int duration_ns; ++}; ++ ++struct dm_ima_device_table_metadata { ++ char *device_metadata; ++ unsigned int device_metadata_len; ++ unsigned int num_targets; ++ char *hash; ++ unsigned int hash_len; ++}; ++ ++struct dm_ima_measurements { ++ struct dm_ima_device_table_metadata active_table; ++ struct dm_ima_device_table_metadata inactive_table; ++ unsigned int dm_version_str_len; ++}; ++ ++struct dm_kobject_holder { ++ struct kobject kobj; ++ struct completion completion; ++}; ++ ++struct dm_md_mempools { ++ struct bio_set bs; ++ struct bio_set io_bs; ++}; ++ ++struct dm_io; ++ ++struct mapped_device { ++ struct mutex suspend_lock; ++ struct mutex table_devices_lock; ++ struct list_head table_devices; ++ void *map; ++ long unsigned int flags; ++ struct mutex type_lock; ++ enum dm_queue_mode type; ++ int numa_node_id; ++ struct request_queue *queue; ++ atomic_t holders; ++ atomic_t open_count; ++ struct dm_target *immutable_target; ++ struct target_type *immutable_target_type; ++ char name[16]; ++ struct gendisk *disk; ++ struct dax_device *dax_dev; ++ wait_queue_head_t wait; ++ long unsigned int *pending_io; ++ struct hd_geometry geometry; ++ struct workqueue_struct *wq; ++ struct work_struct work; ++ spinlock_t deferred_lock; ++ struct bio_list deferred; ++ struct work_struct requeue_work; ++ struct dm_io *requeue_list; ++ void *interface_ptr; ++ wait_queue_head_t eventq; ++ atomic_t event_nr; ++ atomic_t uevent_seq; ++ struct list_head uevent_list; ++ spinlock_t uevent_lock; ++ bool init_tio_pdu: 1; ++ struct blk_mq_tag_set *tag_set; ++ struct dm_stats stats; ++ unsigned int internal_suspend_count; ++ int swap_bios; ++ struct semaphore swap_bios_semaphore; ++ struct mutex swap_bios_lock; ++ struct dm_md_mempools *mempools; ++ struct dm_kobject_holder kobj_holder; ++ struct srcu_struct io_barrier; ++ unsigned int nr_zones; ++ unsigned int *zwp_offset; ++ struct dm_ima_measurements ima; ++}; ++ ++struct dm_target_io { ++ short unsigned int magic; ++ blk_short_t flags; ++ unsigned int target_bio_nr; ++ struct dm_io *io; ++ struct dm_target *ti; ++ unsigned int *len_ptr; ++ sector_t old_sector; ++ struct bio clone; ++}; ++ ++struct dm_io { ++ short unsigned int magic; ++ blk_short_t flags; ++ spinlock_t lock; ++ long unsigned int start_time; ++ void *data; ++ struct dm_io *next; ++ struct dm_stats_aux stats_aux; ++ blk_status_t status; ++ atomic_t io_count; ++ struct mapped_device *md; ++ struct bio *orig_bio; ++ unsigned int sector_offset; ++ unsigned int sectors; ++ struct dm_target_io tio; ++}; ++ ++struct stripe { ++ struct dm_dev *dev; ++ sector_t physical_start; ++ atomic_t error_count; ++}; ++ ++struct stripe_c { ++ uint32_t stripes; ++ int stripes_shift; ++ sector_t stripe_width; ++ uint32_t chunk_size; ++ int chunk_size_shift; ++ struct dm_target *ti; ++ struct work_struct trigger_event; ++ struct stripe stripe[0]; ++}; ++ ++struct dm_io_region { ++ struct block_device *bdev; ++ sector_t sector; ++ sector_t count; ++}; ++ ++struct page_list { ++ struct page_list *next; ++ struct page *page; ++}; ++ ++typedef void (*io_notify_fn)(long unsigned int, void *); ++ ++enum dm_io_mem_type { ++ DM_IO_PAGE_LIST = 0, ++ DM_IO_BIO = 1, ++ DM_IO_VMA = 2, ++ DM_IO_KMEM = 3, ++}; ++ ++struct dm_io_memory { ++ enum dm_io_mem_type type; ++ unsigned int offset; ++ union { ++ struct page_list *pl; ++ struct bio *bio; ++ void *vma; ++ void *addr; ++ } ptr; ++}; ++ ++struct dm_io_notify { ++ io_notify_fn fn; ++ void *context; ++}; ++ ++struct dm_io_client; ++ ++struct dm_io_request { ++ blk_opf_t bi_opf; ++ struct dm_io_memory mem; ++ struct dm_io_notify notify; ++ struct dm_io_client *client; ++}; ++ ++struct dm_kcopyd_throttle { ++ unsigned int throttle; ++ unsigned int num_io_jobs; ++ unsigned int io_period; ++ unsigned int total_period; ++ unsigned int last_jiffies; ++}; ++ ++typedef void (*dm_kcopyd_notify_fn)(int, long unsigned int, void *); ++ ++struct dm_kcopyd_client { ++ struct page_list *pages; ++ unsigned int nr_reserved_pages; ++ unsigned int nr_free_pages; ++ unsigned int sub_job_size; ++ struct dm_io_client *io_client; ++ wait_queue_head_t destroyq; ++ mempool_t job_pool; ++ struct workqueue_struct *kcopyd_wq; ++ struct work_struct kcopyd_work; ++ struct dm_kcopyd_throttle *throttle; ++ atomic_t nr_jobs; ++ spinlock_t job_lock; ++ struct list_head callback_jobs; ++ struct list_head complete_jobs; ++ struct list_head io_jobs; ++ struct list_head pages_jobs; ++}; ++ ++struct kcopyd_job { ++ struct dm_kcopyd_client *kc; ++ struct list_head list; ++ unsigned int flags; ++ int read_err; ++ long unsigned int write_err; ++ enum req_op op; ++ struct dm_io_region source; ++ unsigned int num_dests; ++ struct dm_io_region dests[8]; ++ struct page_list *pages; ++ dm_kcopyd_notify_fn fn; ++ void *context; ++ struct mutex lock; ++ atomic_t sub_jobs; ++ sector_t progress; ++ sector_t write_offset; ++ struct kcopyd_job *master_job; ++}; ++ ++struct dm_rq_target_io; ++ ++struct dm_rq_clone_bio_info { ++ struct bio *orig; ++ struct dm_rq_target_io *tio; ++ struct bio clone; ++}; ++ ++struct dm_rq_target_io { ++ struct mapped_device *md; ++ struct dm_target *ti; ++ struct request *orig; ++ struct request *clone; ++ struct kthread_work work; ++ blk_status_t error; ++ union map_info info; ++ struct dm_stats_aux stats_aux; ++ long unsigned int duration_jiffies; ++ unsigned int n_sectors; ++ unsigned int completed; ++}; ++ ++struct dm_arg_set { ++ unsigned int argc; ++ char **argv; ++}; ++ ++struct dm_arg { ++ unsigned int min; ++ unsigned int max; ++ char *error; ++}; ++ ++typedef sector_t chunk_t; ++ ++struct dm_exception { ++ struct hlist_bl_node hash_list; ++ chunk_t old_chunk; ++ chunk_t new_chunk; ++}; ++ ++struct dm_exception_store; ++ ++struct dm_exception_store_type { ++ const char *name; ++ struct module *module; ++ int (*ctr)(struct dm_exception_store *, char *); ++ void (*dtr)(struct dm_exception_store *); ++ int (*read_metadata)(struct dm_exception_store *, int (*)(void *, chunk_t, chunk_t), void *); ++ int (*prepare_exception)(struct dm_exception_store *, struct dm_exception *); ++ void (*commit_exception)(struct dm_exception_store *, struct dm_exception *, int, void (*)(void *, int), void *); ++ int (*prepare_merge)(struct dm_exception_store *, chunk_t *, chunk_t *); ++ int (*commit_merge)(struct dm_exception_store *, int); ++ void (*drop_snapshot)(struct dm_exception_store *); ++ unsigned int (*status)(struct dm_exception_store *, status_type_t, char *, unsigned int); ++ void (*usage)(struct dm_exception_store *, sector_t *, sector_t *, sector_t *); ++ struct list_head list; ++}; ++ ++struct dm_snapshot; ++ ++struct dm_exception_store { ++ struct dm_exception_store_type *type; ++ struct dm_snapshot *snap; ++ unsigned int chunk_size; ++ unsigned int chunk_mask; ++ unsigned int chunk_shift; ++ void *context; ++ bool userspace_supports_overflow; ++}; ++ ++struct dm_exception_table { ++ uint32_t hash_mask; ++ unsigned int hash_shift; ++ struct hlist_bl_head *table; ++}; ++ ++struct dm_snapshot { ++ struct rw_semaphore lock; ++ struct dm_dev *origin; ++ struct dm_dev *cow; ++ struct dm_target *ti; ++ struct list_head list; ++ int valid; ++ int snapshot_overflowed; ++ int active; ++ atomic_t pending_exceptions_count; ++ spinlock_t pe_allocation_lock; ++ sector_t exception_start_sequence; ++ sector_t exception_complete_sequence; ++ struct rb_root out_of_order_tree; ++ mempool_t pending_pool; ++ struct dm_exception_table pending; ++ struct dm_exception_table complete; ++ spinlock_t pe_lock; ++ spinlock_t tracked_chunk_lock; ++ struct hlist_head tracked_chunk_hash[16]; ++ struct dm_exception_store *store; ++ unsigned int in_progress; ++ struct wait_queue_head in_progress_wait; ++ struct dm_kcopyd_client *kcopyd_client; ++ long unsigned int state_bits; ++ chunk_t first_merging_chunk; ++ int num_merging_chunks; ++ bool merge_failed: 1; ++ bool discard_zeroes_cow: 1; ++ bool discard_passdown_origin: 1; ++ struct bio_list bios_queued_during_merge; ++}; ++ ++struct dm_snap_pending_exception { ++ struct dm_exception e; ++ struct bio_list origin_bios; ++ struct bio_list snapshot_bios; ++ struct dm_snapshot *snap; ++ int started; ++ int copy_error; ++ sector_t exception_sequence; ++ struct rb_node out_of_order_node; ++ struct bio *full_bio; ++ bio_end_io_t *full_bio_end_io; ++}; ++ ++struct dm_snap_tracked_chunk { ++ struct hlist_node node; ++ chunk_t chunk; ++}; ++ ++struct origin { ++ struct block_device *bdev; ++ struct list_head hash_list; ++ struct list_head snapshots; ++}; ++ ++struct dm_origin { ++ struct dm_dev *dev; ++ struct dm_target *ti; ++ unsigned int split_boundary; ++ struct list_head hash_list; ++}; ++ ++struct dm_exception_table_lock { ++ struct hlist_bl_head *complete_slot; ++ struct hlist_bl_head *pending_slot; ++}; ++ ++struct disk_header { ++ __le32 magic; ++ __le32 valid; ++ __le32 version; ++ __le32 chunk_size; ++}; ++ ++struct disk_exception { ++ __le64 old_chunk; ++ __le64 new_chunk; ++}; ++ ++struct core_exception { ++ uint64_t old_chunk; ++ uint64_t new_chunk; ++}; ++ ++struct commit_callback { ++ void (*callback)(void *, int); ++ void *context; ++}; ++ ++struct pstore { ++ struct dm_exception_store *store; ++ int version; ++ int valid; ++ uint32_t exceptions_per_area; ++ void *area; ++ void *zero_area; ++ void *header_area; ++ chunk_t current_area; ++ chunk_t next_free; ++ uint32_t current_committed; ++ atomic_t pending_count; ++ uint32_t callback_count; ++ struct commit_callback *callbacks; ++ struct dm_io_client *io_client; ++ struct workqueue_struct *metadata_wq; ++}; ++ ++struct mdata_req { ++ struct dm_io_region *where; ++ struct dm_io_request *io_req; ++ struct work_struct work; ++ int result; ++}; ++ ++typedef sector_t region_t; ++ ++struct dm_dirty_log_type; ++ ++struct dm_dirty_log { ++ struct dm_dirty_log_type *type; ++ int (*flush_callback_fn)(struct dm_target *); ++ void *context; ++}; ++ ++struct dm_dirty_log_type { ++ const char *name; ++ struct module *module; ++ struct list_head list; ++ int (*ctr)(struct dm_dirty_log *, struct dm_target *, unsigned int, char **); ++ void (*dtr)(struct dm_dirty_log *); ++ int (*presuspend)(struct dm_dirty_log *); ++ int (*postsuspend)(struct dm_dirty_log *); ++ int (*resume)(struct dm_dirty_log *); ++ uint32_t (*get_region_size)(struct dm_dirty_log *); ++ int (*is_clean)(struct dm_dirty_log *, region_t); ++ int (*in_sync)(struct dm_dirty_log *, region_t, int); ++ int (*flush)(struct dm_dirty_log *); ++ void (*mark_region)(struct dm_dirty_log *, region_t); ++ void (*clear_region)(struct dm_dirty_log *, region_t); ++ int (*get_resync_work)(struct dm_dirty_log *, region_t *); ++ void (*set_region_sync)(struct dm_dirty_log *, region_t, int); ++ region_t (*get_sync_count)(struct dm_dirty_log *); ++ int (*status)(struct dm_dirty_log *, status_type_t, char *, unsigned int); ++ int (*is_remote_recovering)(struct dm_dirty_log *, region_t); ++}; ++ ++enum dm_rh_region_states { ++ DM_RH_CLEAN = 1, ++ DM_RH_DIRTY = 2, ++ DM_RH_NOSYNC = 4, ++ DM_RH_RECOVERING = 8, ++}; ++ ++struct dm_region_hash { ++ uint32_t region_size; ++ unsigned int region_shift; ++ struct dm_dirty_log *log; ++ rwlock_t hash_lock; ++ unsigned int mask; ++ unsigned int nr_buckets; ++ unsigned int prime; ++ unsigned int shift; ++ struct list_head *buckets; ++ int flush_failure; ++ unsigned int max_recovery; ++ spinlock_t region_lock; ++ atomic_t recovery_in_flight; ++ struct list_head clean_regions; ++ struct list_head quiesced_regions; ++ struct list_head recovered_regions; ++ struct list_head failed_recovered_regions; ++ struct semaphore recovery_count; ++ mempool_t region_pool; ++ void *context; ++ sector_t target_begin; ++ void (*dispatch_bios)(void *, struct bio_list *); ++ void (*wakeup_workers)(void *); ++ void (*wakeup_all_recovery_waiters)(void *); ++}; ++ ++struct dm_region { ++ struct dm_region_hash *rh; ++ region_t key; ++ int state; ++ struct list_head hash_list; ++ struct list_head list; ++ atomic_t pending; ++ struct bio_list delayed_bios; ++}; ++ ++struct resource_table { ++ u32 ver; ++ u32 num; ++ u32 reserved[2]; ++ u32 offset[0]; ++}; ++ ++struct fw_rsc_hdr { ++ u32 type; ++ u8 data[0]; ++}; ++ ++enum fw_resource_type { ++ RSC_CARVEOUT = 0, ++ RSC_DEVMEM = 1, ++ RSC_TRACE = 2, ++ RSC_VDEV = 3, ++ RSC_LAST = 4, ++ RSC_VENDOR_START = 128, ++ RSC_VENDOR_END = 512, ++}; ++ ++struct fw_rsc_carveout { ++ u32 da; ++ u32 pa; ++ u32 len; ++ u32 flags; ++ u32 reserved; ++ u8 name[32]; ++}; ++ ++struct fw_rsc_devmem { ++ u32 da; ++ u32 pa; ++ u32 len; ++ u32 flags; ++ u32 reserved; ++ u8 name[32]; ++}; ++ ++struct fw_rsc_trace { ++ u32 da; ++ u32 len; ++ u32 reserved; ++ u8 name[32]; ++}; ++ ++struct fw_rsc_vdev_vring { ++ u32 da; ++ u32 align; ++ u32 num; ++ u32 notifyid; ++ u32 pa; ++}; ++ ++struct fw_rsc_vdev { ++ u32 id; ++ u32 notifyid; ++ u32 dfeatures; ++ u32 gfeatures; ++ u32 config_len; ++ u8 status; ++ u8 num_of_vrings; ++ u8 reserved[2]; ++ struct fw_rsc_vdev_vring vring[0]; ++}; ++ ++struct rproc; ++ ++struct rproc_mem_entry { ++ void *va; ++ bool is_iomem; ++ dma_addr_t dma; ++ size_t len; ++ u32 da; ++ void *priv; ++ char name[32]; ++ struct list_head node; ++ u32 rsc_offset; ++ u32 flags; ++ u32 of_resm_idx; ++ int (*alloc)(struct rproc *, struct rproc_mem_entry *); ++ int (*release)(struct rproc *, struct rproc_mem_entry *); ++}; ++ ++enum rproc_dump_mechanism { ++ RPROC_COREDUMP_DISABLED = 0, ++ RPROC_COREDUMP_ENABLED = 1, ++ RPROC_COREDUMP_INLINE = 2, ++}; ++ ++struct rproc_ops; ++ ++struct rproc { ++ struct list_head node; ++ struct iommu_domain *domain; ++ const char *name; ++ const char *firmware; ++ void *priv; ++ struct rproc_ops *ops; ++ struct device dev; ++ atomic_t power; ++ unsigned int state; ++ enum rproc_dump_mechanism dump_conf; ++ struct mutex lock; ++ struct dentry *dbg_dir; ++ struct list_head traces; ++ int num_traces; ++ struct list_head carveouts; ++ struct list_head mappings; ++ u64 bootaddr; ++ struct list_head rvdevs; ++ struct list_head subdevs; ++ struct idr notifyids; ++ int index; ++ struct work_struct crash_handler; ++ unsigned int crash_cnt; ++ bool recovery_disabled; ++ int max_notifyid; ++ struct resource_table *table_ptr; ++ struct resource_table *clean_table; ++ struct resource_table *cached_table; ++ size_t table_sz; ++ bool has_iommu; ++ bool auto_boot; ++ bool sysfs_read_only; ++ struct list_head dump_segments; ++ int nb_vdev; ++ u8 elf_class; ++ u16 elf_machine; ++ struct cdev cdev; ++ bool cdev_put_on_release; ++ long unsigned int features[1]; ++}; ++ ++enum rsc_handling_status { ++ RSC_HANDLED = 0, ++ RSC_IGNORED = 1, ++}; ++ ++struct rproc_ops { ++ int (*prepare)(struct rproc *); ++ int (*unprepare)(struct rproc *); ++ int (*start)(struct rproc *); ++ int (*stop)(struct rproc *); ++ int (*attach)(struct rproc *); ++ int (*detach)(struct rproc *); ++ void (*kick)(struct rproc *, int); ++ void * (*da_to_va)(struct rproc *, u64, size_t, bool *); ++ int (*parse_fw)(struct rproc *, const struct firmware *); ++ int (*handle_rsc)(struct rproc *, u32, void *, int, int); ++ struct resource_table * (*find_loaded_rsc_table)(struct rproc *, const struct firmware *); ++ struct resource_table * (*get_loaded_rsc_table)(struct rproc *, size_t *); ++ int (*load)(struct rproc *, const struct firmware *); ++ int (*sanity_check)(struct rproc *, const struct firmware *); ++ u64 (*get_boot_addr)(struct rproc *, const struct firmware *); ++ long unsigned int (*panic)(struct rproc *); ++ void (*coredump)(struct rproc *); ++}; ++ ++enum rproc_state { ++ RPROC_OFFLINE = 0, ++ RPROC_SUSPENDED = 1, ++ RPROC_RUNNING = 2, ++ RPROC_CRASHED = 3, ++ RPROC_DELETED = 4, ++ RPROC_ATTACHED = 5, ++ RPROC_DETACHED = 6, ++ RPROC_LAST = 7, ++}; ++ ++enum rproc_crash_type { ++ RPROC_MMUFAULT = 0, ++ RPROC_WATCHDOG = 1, ++ RPROC_FATAL_ERROR = 2, ++}; ++ ++enum rproc_features { ++ RPROC_FEAT_ATTACH_ON_RECOVERY = 0, ++ RPROC_MAX_FEATURES = 1, ++}; ++ ++struct rproc_subdev { ++ struct list_head node; ++ int (*prepare)(struct rproc_subdev *); ++ int (*start)(struct rproc_subdev *); ++ void (*stop)(struct rproc_subdev *, bool); ++ void (*unprepare)(struct rproc_subdev *); ++}; ++ ++struct rproc_vdev; ++ ++struct rproc_vring { ++ void *va; ++ int num; ++ u32 da; ++ u32 align; ++ int notifyid; ++ struct rproc_vdev *rvdev; ++ struct virtqueue *vq; ++}; ++ ++struct rproc_vdev { ++ struct rproc_subdev subdev; ++ struct platform_device *pdev; ++ unsigned int id; ++ struct list_head node; ++ struct rproc *rproc; ++ struct rproc_vring vring[2]; ++ u32 rsc_offset; ++ u32 index; ++}; ++ ++struct rproc_debug_trace { ++ struct rproc *rproc; ++ struct dentry *tfile; ++ struct list_head node; ++ struct rproc_mem_entry trace_mem; ++}; ++ ++struct rproc_vdev_data { ++ u32 rsc_offset; ++ unsigned int id; ++ u32 index; ++ struct fw_rsc_vdev *rsc; ++}; ++ ++typedef int (*rproc_handle_resource_t)(struct rproc *, void *, int, int); ++ ++struct extcon_cable; ++ ++struct extcon_dev { ++ const char *name; ++ const unsigned int *supported_cable; ++ const u32 *mutually_exclusive; ++ struct device dev; ++ struct raw_notifier_head nh_all; ++ struct raw_notifier_head *nh; ++ struct list_head entry; ++ int max_supported; ++ spinlock_t lock; ++ u32 state; ++ struct device_type extcon_dev_type; ++ struct extcon_cable *cables; ++ struct attribute_group attr_g_muex; ++ struct attribute **attrs_muex; ++ struct device_attribute *d_attrs_muex; ++}; ++ ++struct extcon_dev_notifier_devres { ++ struct extcon_dev *edev; ++ unsigned int id; ++ struct notifier_block *nb; ++}; ++ ++enum tb_cfg_pkg_type { ++ TB_CFG_PKG_READ = 1, ++ TB_CFG_PKG_WRITE = 2, ++ TB_CFG_PKG_ERROR = 3, ++ TB_CFG_PKG_NOTIFY_ACK = 4, ++ TB_CFG_PKG_EVENT = 5, ++ TB_CFG_PKG_XDOMAIN_REQ = 6, ++ TB_CFG_PKG_XDOMAIN_RESP = 7, ++ TB_CFG_PKG_OVERRIDE = 8, ++ TB_CFG_PKG_RESET = 9, ++ TB_CFG_PKG_ICM_EVENT = 10, ++ TB_CFG_PKG_ICM_CMD = 11, ++ TB_CFG_PKG_ICM_RESP = 12, ++ TB_CFG_PKG_PREPARE_TO_SLEEP = 13, ++}; ++ ++enum tb_security_level { ++ TB_SECURITY_NONE = 0, ++ TB_SECURITY_USER = 1, ++ TB_SECURITY_SECURE = 2, ++ TB_SECURITY_DPONLY = 3, ++ TB_SECURITY_USBONLY = 4, ++ TB_SECURITY_NOPCIE = 5, ++}; ++ ++struct tb_nhi; ++ ++struct tb_ctl; ++ ++struct tb_switch; ++ ++struct tb_cm_ops; ++ ++struct tb { ++ struct device dev; ++ struct mutex lock; ++ struct tb_nhi *nhi; ++ struct tb_ctl *ctl; ++ struct workqueue_struct *wq; ++ struct tb_switch *root_switch; ++ const struct tb_cm_ops *cm_ops; ++ int index; ++ enum tb_security_level security_level; ++ size_t nboot_acl; ++ long unsigned int privdata[0]; ++}; ++ ++struct tb_nhi_ops; ++ ++struct tb_ring; ++ ++struct tb_nhi { ++ spinlock_t lock; ++ struct pci_dev *pdev; ++ const struct tb_nhi_ops *ops; ++ void *iobase; ++ struct tb_ring **tx_rings; ++ struct tb_ring **rx_rings; ++ struct ida msix_ida; ++ bool going_away; ++ bool iommu_dma_protection; ++ struct work_struct interrupt_work; ++ u32 hop_count; ++ long unsigned int quirks; ++}; ++ ++struct tb_regs_switch_header { ++ u16 vendor_id; ++ u16 device_id; ++ u32 first_cap_offset: 8; ++ u32 upstream_port_number: 6; ++ u32 max_port_number: 6; ++ u32 depth: 3; ++ u32 __unknown1: 1; ++ u32 revision: 8; ++ u32 route_lo; ++ u32 route_hi: 31; ++ bool enabled: 1; ++ u32 plug_events_delay: 8; ++ u32 cmuv: 8; ++ u32 __unknown4: 8; ++ u32 thunderbolt_version: 8; ++}; ++ ++enum tb_switch_tmu_rate { ++ TB_SWITCH_TMU_RATE_OFF = 0, ++ TB_SWITCH_TMU_RATE_HIFI = 16, ++ TB_SWITCH_TMU_RATE_NORMAL = 1000, ++}; ++ ++struct tb_switch_tmu { ++ int cap; ++ bool has_ucap; ++ enum tb_switch_tmu_rate rate; ++ bool unidirectional; ++ bool unidirectional_request; ++ enum tb_switch_tmu_rate rate_request; ++}; ++ ++enum tb_clx { ++ TB_CLX_DISABLE = 0, ++ TB_CL1 = 1, ++ TB_CL2 = 2, ++}; ++ ++struct tb_port; ++ ++struct tb_dma_port; ++ ++struct tb_nvm; ++ ++struct tb_switch { ++ struct device dev; ++ struct tb_regs_switch_header config; ++ struct tb_port *ports; ++ struct tb_dma_port *dma_port; ++ struct tb_switch_tmu tmu; ++ struct tb *tb; ++ u64 uid; ++ uuid_t *uuid; ++ u16 vendor; ++ u16 device; ++ const char *vendor_name; ++ const char *device_name; ++ unsigned int link_speed; ++ unsigned int link_width; ++ bool link_usb4; ++ unsigned int generation; ++ int cap_plug_events; ++ int cap_vsec_tmu; ++ int cap_lc; ++ int cap_lp; ++ bool is_unplugged; ++ u8 *drom; ++ struct tb_nvm *nvm; ++ bool no_nvm_upgrade; ++ bool safe_mode; ++ bool boot; ++ bool rpm; ++ unsigned int authorized; ++ enum tb_security_level security_level; ++ struct dentry *debugfs_dir; ++ u8 *key; ++ u8 connection_id; ++ u8 connection_key; ++ u8 link; ++ u8 depth; ++ struct completion rpm_complete; ++ long unsigned int quirks; ++ bool credit_allocation; ++ unsigned int max_usb3_credits; ++ unsigned int min_dp_aux_credits; ++ unsigned int min_dp_main_credits; ++ unsigned int max_pcie_credits; ++ unsigned int max_dma_credits; ++ enum tb_clx clx; ++}; ++ ++struct tb_xdomain; ++ ++struct tb_cm_ops { ++ int (*driver_ready)(struct tb *); ++ int (*start)(struct tb *); ++ void (*stop)(struct tb *); ++ int (*suspend_noirq)(struct tb *); ++ int (*resume_noirq)(struct tb *); ++ int (*suspend)(struct tb *); ++ int (*freeze_noirq)(struct tb *); ++ int (*thaw_noirq)(struct tb *); ++ void (*complete)(struct tb *); ++ int (*runtime_suspend)(struct tb *); ++ int (*runtime_resume)(struct tb *); ++ int (*runtime_suspend_switch)(struct tb_switch *); ++ int (*runtime_resume_switch)(struct tb_switch *); ++ void (*handle_event)(struct tb *, enum tb_cfg_pkg_type, const void *, size_t); ++ int (*get_boot_acl)(struct tb *, uuid_t *, size_t); ++ int (*set_boot_acl)(struct tb *, const uuid_t *, size_t); ++ int (*disapprove_switch)(struct tb *, struct tb_switch *); ++ int (*approve_switch)(struct tb *, struct tb_switch *); ++ int (*add_switch_key)(struct tb *, struct tb_switch *); ++ int (*challenge_switch_key)(struct tb *, struct tb_switch *, const u8 *, u8 *); ++ int (*disconnect_pcie_paths)(struct tb *); ++ int (*approve_xdomain_paths)(struct tb *, struct tb_xdomain *, int, int, int, int); ++ int (*disconnect_xdomain_paths)(struct tb *, struct tb_xdomain *, int, int, int, int); ++ int (*usb4_switch_op)(struct tb_switch *, u16, u32 *, u8 *, const void *, size_t, void *, size_t); ++ int (*usb4_switch_nvm_authenticate_status)(struct tb_switch *, u32 *); ++}; ++ ++struct tb_property_dir { ++ const uuid_t *uuid; ++ struct list_head properties; ++}; ++ ++struct tb_xdomain { ++ struct device dev; ++ struct tb *tb; ++ uuid_t *remote_uuid; ++ const uuid_t *local_uuid; ++ u64 route; ++ u16 vendor; ++ u16 device; ++ unsigned int local_max_hopid; ++ unsigned int remote_max_hopid; ++ struct mutex lock; ++ const char *vendor_name; ++ const char *device_name; ++ unsigned int link_speed; ++ unsigned int link_width; ++ bool link_usb4; ++ bool is_unplugged; ++ bool needs_uuid; ++ struct ida service_ids; ++ struct ida in_hopids; ++ struct ida out_hopids; ++ u32 *local_property_block; ++ u32 local_property_block_gen; ++ u32 local_property_block_len; ++ struct tb_property_dir *remote_properties; ++ u32 remote_property_block_gen; ++ int state; ++ struct delayed_work state_work; ++ int state_retries; ++ struct delayed_work properties_changed_work; ++ int properties_changed_retries; ++ bool bonding_possible; ++ u8 target_link_width; ++ u8 link; ++ u8 depth; ++}; ++ ++struct tb_nhi_ops { ++ int (*init)(struct tb_nhi *); ++ int (*suspend_noirq)(struct tb_nhi *, bool); ++ int (*resume_noirq)(struct tb_nhi *); ++ int (*runtime_suspend)(struct tb_nhi *); ++ int (*runtime_resume)(struct tb_nhi *); ++ void (*shutdown)(struct tb_nhi *); ++}; ++ ++struct ring_desc; ++ ++struct tb_ring { ++ spinlock_t lock; ++ struct tb_nhi *nhi; ++ int size; ++ int hop; ++ int head; ++ int tail; ++ struct ring_desc *descriptors; ++ dma_addr_t descriptors_dma; ++ struct list_head queue; ++ struct list_head in_flight; ++ struct work_struct work; ++ bool is_tx: 1; ++ bool running: 1; ++ int irq; ++ u8 vector; ++ unsigned int flags; ++ int e2e_tx_hop; ++ u16 sof_mask; ++ u16 eof_mask; ++ void (*start_poll)(void *); ++ void *poll_data; ++}; ++ ++enum ring_desc_flags { ++ RING_DESC_ISOCH = 1, ++ RING_DESC_CRC_ERROR = 1, ++ RING_DESC_COMPLETED = 2, ++ RING_DESC_POSTED = 4, ++ RING_DESC_BUFFER_OVERRUN = 4, ++ RING_DESC_INTERRUPT = 8, ++}; ++ ++struct ring_desc { ++ u64 phys; ++ u32 length: 12; ++ u32 eof: 4; ++ u32 sof: 4; ++ enum ring_desc_flags flags: 12; ++ u32 time; ++}; ++ ++struct ring_frame; ++ ++typedef void (*ring_cb)(struct tb_ring *, struct ring_frame *, bool); ++ ++struct ring_frame { ++ dma_addr_t buffer_phy; ++ ring_cb callback; ++ struct list_head list; ++ u32 size: 12; ++ u32 flags: 12; ++ u32 eof: 4; ++ u32 sof: 4; ++}; ++ ++enum nhi_fw_mode { ++ NHI_FW_SAFE_MODE = 0, ++ NHI_FW_AUTH_MODE = 1, ++ NHI_FW_EP_MODE = 2, ++ NHI_FW_CM_MODE = 3, ++}; ++ ++enum nhi_mailbox_cmd { ++ NHI_MAILBOX_SAVE_DEVS = 5, ++ NHI_MAILBOX_DISCONNECT_PCIE_PATHS = 6, ++ NHI_MAILBOX_DRV_UNLOADS = 7, ++ NHI_MAILBOX_DISCONNECT_PA = 16, ++ NHI_MAILBOX_DISCONNECT_PB = 17, ++ NHI_MAILBOX_ALLOW_ALL_DEVS = 35, ++}; ++ ++enum ring_flags { ++ RING_FLAG_ISOCH_ENABLE = 134217728, ++ RING_FLAG_E2E_FLOW_CONTROL = 268435456, ++ RING_FLAG_PCI_NO_SNOOP = 536870912, ++ RING_FLAG_RAW = 1073741824, ++ RING_FLAG_ENABLE = -2147483648, ++}; ++ ++enum tb_port_type { ++ TB_TYPE_INACTIVE = 0, ++ TB_TYPE_PORT = 1, ++ TB_TYPE_NHI = 2, ++ TB_TYPE_DP_HDMI_IN = 917761, ++ TB_TYPE_DP_HDMI_OUT = 917762, ++ TB_TYPE_PCIE_DOWN = 1048833, ++ TB_TYPE_PCIE_UP = 1048834, ++ TB_TYPE_USB3_DOWN = 2097409, ++ TB_TYPE_USB3_UP = 2097410, ++}; ++ ++struct tb_regs_port_header { ++ u16 vendor_id; ++ u16 device_id; ++ u32 first_cap_offset: 8; ++ u32 max_counters: 11; ++ u32 counters_support: 1; ++ u32 __unknown1: 4; ++ u32 revision: 8; ++ enum tb_port_type type: 24; ++ u32 thunderbolt_version: 8; ++ u32 __unknown2: 20; ++ u32 port_number: 6; ++ u32 __unknown3: 6; ++ u32 nfc_credits; ++ u32 max_in_hop_id: 11; ++ u32 max_out_hop_id: 11; ++ u32 __unknown4: 10; ++ u32 __unknown5; ++ u32 __unknown6; ++}; ++ ++struct nvmem_device; ++ ++struct tb_nvm_vendor_ops; ++ ++struct tb_nvm { ++ struct device *dev; ++ u32 major; ++ u32 minor; ++ int id; ++ struct nvmem_device *active; ++ size_t active_size; ++ struct nvmem_device *non_active; ++ void *buf; ++ void *buf_data_start; ++ size_t buf_data_size; ++ bool authenticating; ++ bool flushed; ++ const struct tb_nvm_vendor_ops *vops; ++}; ++ ++struct tb_nvm_vendor_ops { ++ int (*read_version)(struct tb_nvm *); ++ int (*validate)(struct tb_nvm *); ++ int (*write_headers)(struct tb_nvm *); ++}; ++ ++struct usb4_port; ++ ++struct tb_bandwidth_group; ++ ++struct tb_port { ++ struct tb_regs_port_header config; ++ struct tb_switch *sw; ++ struct tb_port *remote; ++ struct tb_xdomain *xdomain; ++ int cap_phy; ++ int cap_tmu; ++ int cap_adap; ++ int cap_usb4; ++ struct usb4_port *usb4; ++ u8 port; ++ bool disabled; ++ bool bonded; ++ struct tb_port *dual_link_port; ++ u8 link_nr: 1; ++ struct ida in_hopids; ++ struct ida out_hopids; ++ struct list_head list; ++ unsigned int total_credits; ++ unsigned int ctl_credits; ++ unsigned int dma_credits; ++ struct tb_bandwidth_group *group; ++ struct list_head group_list; ++ unsigned int max_bw; ++}; ++ ++struct tb_bandwidth_group { ++ struct tb *tb; ++ int index; ++ struct list_head ports; ++}; ++ ++struct usb4_port { ++ struct device dev; ++ struct tb_port *port; ++ bool can_offline; ++ bool offline; ++}; ++ ++enum tb_switch_cap { ++ TB_SWITCH_CAP_TMU = 3, ++ TB_SWITCH_CAP_VSE = 5, ++}; ++ ++enum tb_switch_vse_cap { ++ TB_VSE_CAP_PLUG_EVENTS = 1, ++ TB_VSE_CAP_TIME2 = 3, ++ TB_VSE_CAP_CP_LP = 4, ++ TB_VSE_CAP_LINK_CONTROLLER = 6, ++}; ++ ++enum tb_port_cap { ++ TB_PORT_CAP_PHY = 1, ++ TB_PORT_CAP_POWER = 2, ++ TB_PORT_CAP_TIME1 = 3, ++ TB_PORT_CAP_ADAP = 4, ++ TB_PORT_CAP_VSE = 5, ++ TB_PORT_CAP_USB4 = 6, ++}; ++ ++struct tb_cap_basic { ++ u8 next; ++ u8 cap; ++}; ++ ++struct tb_cap_extended_short { ++ u8 next; ++ u8 cap; ++ u8 vsec_id; ++ u8 length; ++}; ++ ++struct tb_cap_extended_long { ++ u8 zero1; ++ u8 cap; ++ u8 vsec_id; ++ u8 zero2; ++ u16 next; ++ u16 length; ++}; ++ ++struct tb_cap_any { ++ union { ++ struct tb_cap_basic basic; ++ struct tb_cap_extended_short extended_short; ++ struct tb_cap_extended_long extended_long; ++ }; ++}; ++ ++enum tb_cfg_space { ++ TB_CFG_HOPS = 0, ++ TB_CFG_PORT = 1, ++ TB_CFG_SWITCH = 2, ++ TB_CFG_COUNTERS = 3, ++}; ++ ++struct tb_path_hop { ++ struct tb_port *in_port; ++ struct tb_port *out_port; ++ int in_hop_index; ++ int in_counter_index; ++ int next_hop_index; ++ unsigned int initial_credits; ++ unsigned int nfc_credits; ++}; ++ ++enum tb_path_port { ++ TB_PATH_NONE = 0, ++ TB_PATH_SOURCE = 1, ++ TB_PATH_INTERNAL = 2, ++ TB_PATH_DESTINATION = 4, ++ TB_PATH_ALL = 7, ++}; ++ ++struct tb_path { ++ struct tb *tb; ++ const char *name; ++ enum tb_path_port ingress_shared_buffer; ++ enum tb_path_port egress_shared_buffer; ++ enum tb_path_port ingress_fc_enable; ++ enum tb_path_port egress_fc_enable; ++ unsigned int priority: 3; ++ int weight: 4; ++ bool drop_packages; ++ bool activated; ++ bool clear_fc; ++ struct tb_path_hop *hops; ++ int path_length; ++ bool alloc_hopid; ++}; ++ ++enum tb_tunnel_type { ++ TB_TUNNEL_PCI = 0, ++ TB_TUNNEL_DP = 1, ++ TB_TUNNEL_DMA = 2, ++ TB_TUNNEL_USB3 = 3, ++}; ++ ++struct tb_tunnel { ++ struct tb *tb; ++ struct tb_port *src_port; ++ struct tb_port *dst_port; ++ struct tb_path **paths; ++ size_t npaths; ++ int (*init)(struct tb_tunnel *); ++ void (*deinit)(struct tb_tunnel *); ++ int (*activate)(struct tb_tunnel *, bool); ++ int (*maximum_bandwidth)(struct tb_tunnel *, int *, int *); ++ int (*allocated_bandwidth)(struct tb_tunnel *, int *, int *); ++ int (*alloc_bandwidth)(struct tb_tunnel *, int *, int *); ++ int (*consumed_bandwidth)(struct tb_tunnel *, int *, int *); ++ int (*release_unused_bandwidth)(struct tb_tunnel *); ++ void (*reclaim_available_bandwidth)(struct tb_tunnel *, int *, int *); ++ struct list_head list; ++ enum tb_tunnel_type type; ++ int max_up; ++ int max_down; ++ int allocated_up; ++ int allocated_down; ++ bool bw_mode; ++}; ++ ++enum tb_cfg_error { ++ TB_CFG_ERROR_PORT_NOT_CONNECTED = 0, ++ TB_CFG_ERROR_LINK_ERROR = 1, ++ TB_CFG_ERROR_INVALID_CONFIG_SPACE = 2, ++ TB_CFG_ERROR_NO_SUCH_PORT = 4, ++ TB_CFG_ERROR_ACK_PLUG_EVENT = 7, ++ TB_CFG_ERROR_LOOP = 8, ++ TB_CFG_ERROR_HEC_ERROR_DETECTED = 12, ++ TB_CFG_ERROR_FLOW_CONTROL_ERROR = 13, ++ TB_CFG_ERROR_LOCK = 15, ++ TB_CFG_ERROR_DP_BW = 32, ++}; ++ ++struct tb_cfg_header { ++ u32 route_hi: 22; ++ u32 unknown: 10; ++ u32 route_lo; ++}; ++ ++struct tb_cfg_address { ++ u32 offset: 13; ++ u32 length: 6; ++ u32 port: 6; ++ enum tb_cfg_space space: 2; ++ u32 seq: 2; ++ u32 zero: 3; ++}; ++ ++struct cfg_read_pkg { ++ struct tb_cfg_header header; ++ struct tb_cfg_address addr; ++}; ++ ++struct cfg_write_pkg { ++ struct tb_cfg_header header; ++ struct tb_cfg_address addr; ++ u32 data[64]; ++}; ++ ++struct tb_cfg_result { ++ u64 response_route; ++ u32 response_port; ++ int err; ++ enum tb_cfg_error tb_error; ++}; ++ ++struct ctl_pkg { ++ struct tb_ctl *ctl; ++ void *buffer; ++ struct ring_frame frame; ++}; ++ ++struct tb_cfg_request { ++ struct kref kref; ++ struct tb_ctl *ctl; ++ const void *request; ++ size_t request_size; ++ enum tb_cfg_pkg_type request_type; ++ void *response; ++ size_t response_size; ++ enum tb_cfg_pkg_type response_type; ++ size_t npackets; ++ bool (*match)(const struct tb_cfg_request *, const struct ctl_pkg *); ++ bool (*copy)(struct tb_cfg_request *, const struct ctl_pkg *); ++ void (*callback)(void *); ++ void *callback_data; ++ long unsigned int flags; ++ struct work_struct work; ++ struct tb_cfg_result result; ++ struct list_head list; ++}; ++ ++struct tb_dma_port { ++ struct tb_switch *sw; ++ u8 port; ++ u32 base; ++ u8 *buf; ++}; ++ ++typedef int (*read_block_fn)(void *, unsigned int, void *, size_t); ++ ++typedef int (*write_block_fn)(void *, unsigned int, const void *, size_t); ++ ++struct component_ops { ++ int (*bind)(struct device *, struct device *, void *); ++ void (*unbind)(struct device *, struct device *, void *); ++}; ++ ++struct tb_quirk { ++ u16 hw_vendor_id; ++ u16 hw_device_id; ++ u16 vendor; ++ u16 device; ++ void (*hook)(struct tb_switch *); ++}; ++ ++struct icc_path; ++ ++struct icc_bulk_data { ++ struct icc_path *path; ++ const char *name; ++ u32 avg_bw; ++ u32 peak_bw; ++}; ++ ++struct icc_bulk_devres { ++ struct icc_bulk_data *paths; ++ int num_paths; ++}; ++ ++struct __kernel_old_timespec { ++ __kernel_old_time_t tv_sec; ++ long int tv_nsec; ++}; ++ ++struct __kernel_sock_timeval { ++ __s64 tv_sec; ++ __s64 tv_usec; ++}; ++ ++struct user_msghdr { ++ void *msg_name; ++ int msg_namelen; ++ struct iovec *msg_iov; ++ __kernel_size_t msg_iovlen; ++ void *msg_control; ++ __kernel_size_t msg_controllen; ++ unsigned int msg_flags; ++}; ++ ++struct mmsghdr { ++ struct user_msghdr msg_hdr; ++ unsigned int msg_len; ++}; ++ ++struct scm_timestamping_internal { ++ struct timespec64 ts[3]; ++}; ++ ++struct ifconf { ++ int ifc_len; ++ union { ++ char *ifcu_buf; ++ struct ifreq *ifcu_req; ++ } ifc_ifcu; ++}; ++ ++typedef u32 compat_caddr_t; ++ ++struct compat_ifmap { ++ compat_ulong_t mem_start; ++ compat_ulong_t mem_end; ++ short unsigned int base_addr; ++ unsigned char irq; ++ unsigned char dma; ++ unsigned char port; ++}; ++ ++struct compat_if_settings { ++ unsigned int type; ++ unsigned int size; ++ compat_uptr_t ifs_ifsu; ++}; ++ ++struct compat_ifreq { ++ union { ++ char ifrn_name[16]; ++ } ifr_ifrn; ++ union { ++ struct sockaddr ifru_addr; ++ struct sockaddr ifru_dstaddr; ++ struct sockaddr ifru_broadaddr; ++ struct sockaddr ifru_netmask; ++ struct sockaddr ifru_hwaddr; ++ short int ifru_flags; ++ compat_int_t ifru_ivalue; ++ compat_int_t ifru_mtu; ++ struct compat_ifmap ifru_map; ++ char ifru_slave[16]; ++ char ifru_newname[16]; ++ compat_caddr_t ifru_data; ++ struct compat_if_settings ifru_settings; ++ } ifr_ifru; ++}; ++ ++struct bpf_empty_prog_array { ++ struct bpf_prog_array hdr; ++ struct bpf_prog *null_prog; ++}; ++ ++typedef unsigned int nf_hookfn(void *, struct sk_buff *, const struct nf_hook_state *); ++ ++struct nf_hook_entry { ++ nf_hookfn *hook; ++ void *priv; ++}; ++ ++struct nf_hook_entries { ++ u16 num_hook_entries; ++ struct nf_hook_entry hooks[0]; ++}; ++ ++typedef struct bio_vec skb_frag_t; ++ ++enum { ++ SKBTX_HW_TSTAMP = 1, ++ SKBTX_SW_TSTAMP = 2, ++ SKBTX_IN_PROGRESS = 4, ++ SKBTX_HW_TSTAMP_USE_CYCLES = 8, ++ SKBTX_WIFI_STATUS = 16, ++ SKBTX_HW_TSTAMP_NETDEV = 32, ++ SKBTX_SCHED_TSTAMP = 64, ++}; ++ ++struct skb_shared_info { ++ __u8 flags; ++ __u8 meta_len; ++ __u8 nr_frags; ++ __u8 tx_flags; ++ short unsigned int gso_size; ++ short unsigned int gso_segs; ++ struct sk_buff *frag_list; ++ struct skb_shared_hwtstamps hwtstamps; ++ unsigned int gso_type; ++ u32 tskey; ++ atomic_t dataref; ++ unsigned int xdp_frags_size; ++ void *destructor_arg; ++ skb_frag_t frags[17]; ++}; ++ ++enum sock_shutdown_cmd { ++ SHUT_RD = 0, ++ SHUT_WR = 1, ++ SHUT_RDWR = 2, ++}; ++ ++struct net_proto_family { ++ int family; ++ int (*create)(struct net *, struct socket *, int, int); ++ struct module *owner; ++}; ++ ++enum { ++ SOCK_WAKE_IO = 0, ++ SOCK_WAKE_WAITD = 1, ++ SOCK_WAKE_SPACE = 2, ++ SOCK_WAKE_URG = 3, ++}; ++ ++enum { ++ TCPF_ESTABLISHED = 2, ++ TCPF_SYN_SENT = 4, ++ TCPF_SYN_RECV = 8, ++ TCPF_FIN_WAIT1 = 16, ++ TCPF_FIN_WAIT2 = 32, ++ TCPF_TIME_WAIT = 64, ++ TCPF_CLOSE = 128, ++ TCPF_CLOSE_WAIT = 256, ++ TCPF_LAST_ACK = 512, ++ TCPF_LISTEN = 1024, ++ TCPF_CLOSING = 2048, ++ TCPF_NEW_SYN_RECV = 4096, ++}; ++ ++enum { ++ SOF_TIMESTAMPING_TX_HARDWARE = 1, ++ SOF_TIMESTAMPING_TX_SOFTWARE = 2, ++ SOF_TIMESTAMPING_RX_HARDWARE = 4, ++ SOF_TIMESTAMPING_RX_SOFTWARE = 8, ++ SOF_TIMESTAMPING_SOFTWARE = 16, ++ SOF_TIMESTAMPING_SYS_HARDWARE = 32, ++ SOF_TIMESTAMPING_RAW_HARDWARE = 64, ++ SOF_TIMESTAMPING_OPT_ID = 128, ++ SOF_TIMESTAMPING_TX_SCHED = 256, ++ SOF_TIMESTAMPING_TX_ACK = 512, ++ SOF_TIMESTAMPING_OPT_CMSG = 1024, ++ SOF_TIMESTAMPING_OPT_TSONLY = 2048, ++ SOF_TIMESTAMPING_OPT_STATS = 4096, ++ SOF_TIMESTAMPING_OPT_PKTINFO = 8192, ++ SOF_TIMESTAMPING_OPT_TX_SWHW = 16384, ++ SOF_TIMESTAMPING_BIND_PHC = 32768, ++ SOF_TIMESTAMPING_OPT_ID_TCP = 65536, ++ SOF_TIMESTAMPING_LAST = 65536, ++ SOF_TIMESTAMPING_MASK = 131071, ++}; ++ ++struct scm_ts_pktinfo { ++ __u32 if_index; ++ __u32 pkt_length; ++ __u32 reserved[2]; ++}; ++ ++struct socket_alloc { ++ struct socket socket; ++ struct inode vfs_inode; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct sock_skb_cb { ++ u32 dropcount; ++}; ++ ++struct compat_msghdr { ++ compat_uptr_t msg_name; ++ compat_int_t msg_namelen; ++ compat_uptr_t msg_iov; ++ compat_size_t msg_iovlen; ++ compat_uptr_t msg_control; ++ compat_size_t msg_controllen; ++ compat_uint_t msg_flags; ++}; ++ ++struct compat_mmsghdr { ++ struct compat_msghdr msg_hdr; ++ compat_uint_t msg_len; ++}; ++ ++struct inet6_skb_parm { ++ int iif; ++ __be16 ra; ++ __u16 dst0; ++ __u16 srcrt; ++ __u16 dst1; ++ __u16 lastopt; ++ __u16 nhoff; ++ __u16 flags; ++ __u16 dsthao; ++ __u16 frag_max_size; ++ __u16 srhoff; ++}; ++ ++struct inet_skb_parm { ++ int iif; ++ struct ip_options opt; ++ u16 flags; ++ u16 frag_max_size; ++}; ++ ++struct sock_ee_data_rfc4884 { ++ __u16 len; ++ __u8 flags; ++ __u8 reserved; ++}; ++ ++struct sock_extended_err { ++ __u32 ee_errno; ++ __u8 ee_origin; ++ __u8 ee_type; ++ __u8 ee_code; ++ __u8 ee_pad; ++ __u32 ee_info; ++ union { ++ __u32 ee_data; ++ struct sock_ee_data_rfc4884 ee_rfc4884; ++ }; ++}; ++ ++struct sock_exterr_skb { ++ union { ++ struct inet_skb_parm h4; ++ struct inet6_skb_parm h6; ++ } header; ++ struct sock_extended_err ee; ++ u16 addr_offset; ++ __be16 port; ++ u8 opt_stats: 1; ++ u8 unused: 7; ++}; ++ ++struct used_address { ++ struct __kernel_sockaddr_storage name; ++ unsigned int name_len; ++}; ++ ++struct rt6key { ++ struct in6_addr addr; ++ int plen; ++}; ++ ++struct rtable; ++ ++struct fnhe_hash_bucket; ++ ++struct fib_nh_common { ++ struct net_device *nhc_dev; ++ netdevice_tracker nhc_dev_tracker; ++ int nhc_oif; ++ unsigned char nhc_scope; ++ u8 nhc_family; ++ u8 nhc_gw_family; ++ unsigned char nhc_flags; ++ struct lwtunnel_state *nhc_lwtstate; ++ union { ++ __be32 ipv4; ++ struct in6_addr ipv6; ++ } nhc_gw; ++ int nhc_weight; ++ atomic_t nhc_upper_bound; ++ struct rtable **nhc_pcpu_rth_output; ++ struct rtable *nhc_rth_input; ++ struct fnhe_hash_bucket *nhc_exceptions; ++}; ++ ++struct rt6_exception_bucket; ++ ++struct fib6_nh { ++ struct fib_nh_common nh_common; ++ long unsigned int last_probe; ++ struct rt6_info **rt6i_pcpu; ++ struct rt6_exception_bucket *rt6i_exception_bucket; ++}; ++ ++struct fib6_node; ++ ++struct dst_metrics; ++ ++struct nexthop; ++ ++struct fib6_info { ++ struct fib6_table *fib6_table; ++ struct fib6_info *fib6_next; ++ struct fib6_node *fib6_node; ++ union { ++ struct list_head fib6_siblings; ++ struct list_head nh_list; ++ }; ++ unsigned int fib6_nsiblings; ++ refcount_t fib6_ref; ++ long unsigned int expires; ++ struct dst_metrics *fib6_metrics; ++ struct rt6key fib6_dst; ++ u32 fib6_flags; ++ struct rt6key fib6_src; ++ struct rt6key fib6_prefsrc; ++ u32 fib6_metric; ++ u8 fib6_protocol; ++ u8 fib6_type; ++ u8 offload; ++ u8 trap; ++ u8 offload_failed; ++ u8 should_flush: 1; ++ u8 dst_nocount: 1; ++ u8 dst_nopolicy: 1; ++ u8 fib6_destroying: 1; ++ u8 unused: 4; ++ struct callback_head rcu; ++ struct nexthop *nh; ++ struct fib6_nh fib6_nh[0]; ++}; ++ ++struct rt6_info { ++ struct dst_entry dst; ++ struct fib6_info *from; ++ int sernum; ++ struct rt6key rt6i_dst; ++ struct rt6key rt6i_src; ++ struct in6_addr rt6i_gateway; ++ struct inet6_dev *rt6i_idev; ++ u32 rt6i_flags; ++ short unsigned int rt6i_nfheader_len; ++}; ++ ++struct rt6_statistics { ++ __u32 fib_nodes; ++ __u32 fib_route_nodes; ++ __u32 fib_rt_entries; ++ __u32 fib_rt_cache; ++ __u32 fib_discarded_routes; ++ atomic_t fib_rt_alloc; ++}; ++ ++struct fib6_node { ++ struct fib6_node *parent; ++ struct fib6_node *left; ++ struct fib6_node *right; ++ struct fib6_node *subtree; ++ struct fib6_info *leaf; ++ __u16 fn_bit; ++ __u16 fn_flags; ++ int fn_sernum; ++ struct fib6_info *rr_ptr; ++ struct callback_head rcu; ++}; ++ ++struct fib6_table { ++ struct hlist_node tb6_hlist; ++ u32 tb6_id; ++ spinlock_t tb6_lock; ++ struct fib6_node tb6_root; ++ struct inet_peer_base tb6_peers; ++ unsigned int flags; ++ unsigned int fib_seq; ++}; ++ ++struct nf_conntrack { ++ refcount_t use; ++}; ++ ++typedef union { ++ __be32 a4; ++ __be32 a6[4]; ++ struct in6_addr in6; ++} xfrm_address_t; ++ ++struct xfrm_id { ++ xfrm_address_t daddr; ++ __be32 spi; ++ __u8 proto; ++}; ++ ++struct xfrm_selector { ++ xfrm_address_t daddr; ++ xfrm_address_t saddr; ++ __be16 dport; ++ __be16 dport_mask; ++ __be16 sport; ++ __be16 sport_mask; ++ __u16 family; ++ __u8 prefixlen_d; ++ __u8 prefixlen_s; ++ __u8 proto; ++ int ifindex; ++ __kernel_uid32_t user; ++}; ++ ++struct xfrm_lifetime_cfg { ++ __u64 soft_byte_limit; ++ __u64 hard_byte_limit; ++ __u64 soft_packet_limit; ++ __u64 hard_packet_limit; ++ __u64 soft_add_expires_seconds; ++ __u64 hard_add_expires_seconds; ++ __u64 soft_use_expires_seconds; ++ __u64 hard_use_expires_seconds; ++}; ++ ++struct xfrm_lifetime_cur { ++ __u64 bytes; ++ __u64 packets; ++ __u64 add_time; ++ __u64 use_time; ++}; ++ ++struct xfrm_replay_state { ++ __u32 oseq; ++ __u32 seq; ++ __u32 bitmap; ++}; ++ ++struct xfrm_replay_state_esn { ++ unsigned int bmp_len; ++ __u32 oseq; ++ __u32 seq; ++ __u32 oseq_hi; ++ __u32 seq_hi; ++ __u32 replay_window; ++ __u32 bmp[0]; ++}; ++ ++struct xfrm_algo { ++ char alg_name[64]; ++ unsigned int alg_key_len; ++ char alg_key[0]; ++}; ++ ++struct xfrm_algo_auth { ++ char alg_name[64]; ++ unsigned int alg_key_len; ++ unsigned int alg_trunc_len; ++ char alg_key[0]; ++}; ++ ++struct xfrm_algo_aead { ++ char alg_name[64]; ++ unsigned int alg_key_len; ++ unsigned int alg_icv_len; ++ char alg_key[0]; ++}; ++ ++struct xfrm_stats { ++ __u32 replay_window; ++ __u32 replay; ++ __u32 integrity_failed; ++}; ++ ++enum { ++ XFRM_POLICY_TYPE_MAIN = 0, ++ XFRM_POLICY_TYPE_SUB = 1, ++ XFRM_POLICY_TYPE_MAX = 2, ++ XFRM_POLICY_TYPE_ANY = 255, ++}; ++ ++enum { ++ XFRM_MSG_BASE = 16, ++ XFRM_MSG_NEWSA = 16, ++ XFRM_MSG_DELSA = 17, ++ XFRM_MSG_GETSA = 18, ++ XFRM_MSG_NEWPOLICY = 19, ++ XFRM_MSG_DELPOLICY = 20, ++ XFRM_MSG_GETPOLICY = 21, ++ XFRM_MSG_ALLOCSPI = 22, ++ XFRM_MSG_ACQUIRE = 23, ++ XFRM_MSG_EXPIRE = 24, ++ XFRM_MSG_UPDPOLICY = 25, ++ XFRM_MSG_UPDSA = 26, ++ XFRM_MSG_POLEXPIRE = 27, ++ XFRM_MSG_FLUSHSA = 28, ++ XFRM_MSG_FLUSHPOLICY = 29, ++ XFRM_MSG_NEWAE = 30, ++ XFRM_MSG_GETAE = 31, ++ XFRM_MSG_REPORT = 32, ++ XFRM_MSG_MIGRATE = 33, ++ XFRM_MSG_NEWSADINFO = 34, ++ XFRM_MSG_GETSADINFO = 35, ++ XFRM_MSG_NEWSPDINFO = 36, ++ XFRM_MSG_GETSPDINFO = 37, ++ XFRM_MSG_MAPPING = 38, ++ XFRM_MSG_SETDEFAULT = 39, ++ XFRM_MSG_GETDEFAULT = 40, ++ __XFRM_MSG_MAX = 41, ++}; ++ ++struct xfrm_encap_tmpl { ++ __u16 encap_type; ++ __be16 encap_sport; ++ __be16 encap_dport; ++ xfrm_address_t encap_oa; ++}; ++ ++enum xfrm_attr_type_t { ++ XFRMA_UNSPEC = 0, ++ XFRMA_ALG_AUTH = 1, ++ XFRMA_ALG_CRYPT = 2, ++ XFRMA_ALG_COMP = 3, ++ XFRMA_ENCAP = 4, ++ XFRMA_TMPL = 5, ++ XFRMA_SA = 6, ++ XFRMA_POLICY = 7, ++ XFRMA_SEC_CTX = 8, ++ XFRMA_LTIME_VAL = 9, ++ XFRMA_REPLAY_VAL = 10, ++ XFRMA_REPLAY_THRESH = 11, ++ XFRMA_ETIMER_THRESH = 12, ++ XFRMA_SRCADDR = 13, ++ XFRMA_COADDR = 14, ++ XFRMA_LASTUSED = 15, ++ XFRMA_POLICY_TYPE = 16, ++ XFRMA_MIGRATE = 17, ++ XFRMA_ALG_AEAD = 18, ++ XFRMA_KMADDRESS = 19, ++ XFRMA_ALG_AUTH_TRUNC = 20, ++ XFRMA_MARK = 21, ++ XFRMA_TFCPAD = 22, ++ XFRMA_REPLAY_ESN_VAL = 23, ++ XFRMA_SA_EXTRA_FLAGS = 24, ++ XFRMA_PROTO = 25, ++ XFRMA_ADDRESS_FILTER = 26, ++ XFRMA_PAD = 27, ++ XFRMA_OFFLOAD_DEV = 28, ++ XFRMA_SET_MARK = 29, ++ XFRMA_SET_MARK_MASK = 30, ++ XFRMA_IF_ID = 31, ++ XFRMA_MTIMER_THRESH = 32, ++ __XFRMA_MAX = 33, ++}; ++ ++struct xfrm_mark { ++ __u32 v; ++ __u32 m; ++}; ++ ++struct xfrm_address_filter { ++ xfrm_address_t saddr; ++ xfrm_address_t daddr; ++ __u16 family; ++ __u8 splen; ++ __u8 dplen; ++}; ++ ++enum { ++ NETIF_F_SG_BIT = 0, ++ NETIF_F_IP_CSUM_BIT = 1, ++ __UNUSED_NETIF_F_1 = 2, ++ NETIF_F_HW_CSUM_BIT = 3, ++ NETIF_F_IPV6_CSUM_BIT = 4, ++ NETIF_F_HIGHDMA_BIT = 5, ++ NETIF_F_FRAGLIST_BIT = 6, ++ NETIF_F_HW_VLAN_CTAG_TX_BIT = 7, ++ NETIF_F_HW_VLAN_CTAG_RX_BIT = 8, ++ NETIF_F_HW_VLAN_CTAG_FILTER_BIT = 9, ++ NETIF_F_VLAN_CHALLENGED_BIT = 10, ++ NETIF_F_GSO_BIT = 11, ++ NETIF_F_LLTX_BIT = 12, ++ NETIF_F_NETNS_LOCAL_BIT = 13, ++ NETIF_F_GRO_BIT = 14, ++ NETIF_F_LRO_BIT = 15, ++ NETIF_F_GSO_SHIFT = 16, ++ NETIF_F_TSO_BIT = 16, ++ NETIF_F_GSO_ROBUST_BIT = 17, ++ NETIF_F_TSO_ECN_BIT = 18, ++ NETIF_F_TSO_MANGLEID_BIT = 19, ++ NETIF_F_TSO6_BIT = 20, ++ NETIF_F_FSO_BIT = 21, ++ NETIF_F_GSO_GRE_BIT = 22, ++ NETIF_F_GSO_GRE_CSUM_BIT = 23, ++ NETIF_F_GSO_IPXIP4_BIT = 24, ++ NETIF_F_GSO_IPXIP6_BIT = 25, ++ NETIF_F_GSO_UDP_TUNNEL_BIT = 26, ++ NETIF_F_GSO_UDP_TUNNEL_CSUM_BIT = 27, ++ NETIF_F_GSO_PARTIAL_BIT = 28, ++ NETIF_F_GSO_TUNNEL_REMCSUM_BIT = 29, ++ NETIF_F_GSO_SCTP_BIT = 30, ++ NETIF_F_GSO_ESP_BIT = 31, ++ NETIF_F_GSO_UDP_BIT = 32, ++ NETIF_F_GSO_UDP_L4_BIT = 33, ++ NETIF_F_GSO_FRAGLIST_BIT = 34, ++ NETIF_F_GSO_LAST = 34, ++ NETIF_F_FCOE_CRC_BIT = 35, ++ NETIF_F_SCTP_CRC_BIT = 36, ++ NETIF_F_FCOE_MTU_BIT = 37, ++ NETIF_F_NTUPLE_BIT = 38, ++ NETIF_F_RXHASH_BIT = 39, ++ NETIF_F_RXCSUM_BIT = 40, ++ NETIF_F_NOCACHE_COPY_BIT = 41, ++ NETIF_F_LOOPBACK_BIT = 42, ++ NETIF_F_RXFCS_BIT = 43, ++ NETIF_F_RXALL_BIT = 44, ++ NETIF_F_HW_VLAN_STAG_TX_BIT = 45, ++ NETIF_F_HW_VLAN_STAG_RX_BIT = 46, ++ NETIF_F_HW_VLAN_STAG_FILTER_BIT = 47, ++ NETIF_F_HW_L2FW_DOFFLOAD_BIT = 48, ++ NETIF_F_HW_TC_BIT = 49, ++ NETIF_F_HW_ESP_BIT = 50, ++ NETIF_F_HW_ESP_TX_CSUM_BIT = 51, ++ NETIF_F_RX_UDP_TUNNEL_PORT_BIT = 52, ++ NETIF_F_HW_TLS_TX_BIT = 53, ++ NETIF_F_HW_TLS_RX_BIT = 54, ++ NETIF_F_GRO_HW_BIT = 55, ++ NETIF_F_HW_TLS_RECORD_BIT = 56, ++ NETIF_F_GRO_FRAGLIST_BIT = 57, ++ NETIF_F_HW_MACSEC_BIT = 58, ++ NETIF_F_GRO_UDP_FWD_BIT = 59, ++ NETIF_F_HW_HSR_TAG_INS_BIT = 60, ++ NETIF_F_HW_HSR_TAG_RM_BIT = 61, ++ NETIF_F_HW_HSR_FWD_BIT = 62, ++ NETIF_F_HW_HSR_DUP_BIT = 63, ++ NETDEV_FEATURE_COUNT = 64, ++}; ++ ++enum { ++ SKBFL_ZEROCOPY_ENABLE = 1, ++ SKBFL_SHARED_FRAG = 2, ++ SKBFL_PURE_ZEROCOPY = 4, ++ SKBFL_DONT_ORPHAN = 8, ++ SKBFL_MANAGED_FRAG_REFS = 16, ++}; ++ ++struct mmpin { ++ struct user_struct *user; ++ unsigned int num_pg; ++}; ++ ++struct ubuf_info_msgzc { ++ struct ubuf_info ubuf; ++ union { ++ struct { ++ long unsigned int desc; ++ void *ctx; ++ }; ++ struct { ++ u32 id; ++ u16 len; ++ u16 zerocopy: 1; ++ u32 bytelen; ++ }; ++ }; ++ struct mmpin mmp; ++}; ++ ++enum { ++ SKB_FCLONE_UNAVAILABLE = 0, ++ SKB_FCLONE_ORIG = 1, ++ SKB_FCLONE_CLONE = 2, ++}; ++ ++enum { ++ SKB_GSO_TCPV4 = 1, ++ SKB_GSO_DODGY = 2, ++ SKB_GSO_TCP_ECN = 4, ++ SKB_GSO_TCP_FIXEDID = 8, ++ SKB_GSO_TCPV6 = 16, ++ SKB_GSO_FCOE = 32, ++ SKB_GSO_GRE = 64, ++ SKB_GSO_GRE_CSUM = 128, ++ SKB_GSO_IPXIP4 = 256, ++ SKB_GSO_IPXIP6 = 512, ++ SKB_GSO_UDP_TUNNEL = 1024, ++ SKB_GSO_UDP_TUNNEL_CSUM = 2048, ++ SKB_GSO_PARTIAL = 4096, ++ SKB_GSO_TUNNEL_REMCSUM = 8192, ++ SKB_GSO_SCTP = 16384, ++ SKB_GSO_ESP = 32768, ++ SKB_GSO_UDP = 65536, ++ SKB_GSO_UDP_L4 = 131072, ++ SKB_GSO_FRAGLIST = 262144, ++}; ++ ++struct sk_buff_fclones { ++ struct sk_buff skb1; ++ struct sk_buff skb2; ++ refcount_t fclone_ref; ++}; ++ ++struct skb_seq_state { ++ __u32 lower_offset; ++ __u32 upper_offset; ++ __u32 frag_idx; ++ __u32 stepped_offset; ++ struct sk_buff *root_skb; ++ struct sk_buff *cur_skb; ++ __u8 *frag_data; ++ __u32 frag_off; ++}; ++ ++struct skb_checksum_ops { ++ __wsum (*update)(const void *, int, __wsum); ++ __wsum (*combine)(__wsum, __wsum, int, int); ++}; ++ ++struct xfrm_state_walk { ++ struct list_head all; ++ u8 state; ++ u8 dying; ++ u8 proto; ++ u32 seq; ++ struct xfrm_address_filter *filter; ++}; ++ ++enum xfrm_replay_mode { ++ XFRM_REPLAY_MODE_LEGACY = 0, ++ XFRM_REPLAY_MODE_BMP = 1, ++ XFRM_REPLAY_MODE_ESN = 2, ++}; ++ ++struct xfrm_dev_offload { ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ struct net_device *real_dev; ++ long unsigned int offload_handle; ++ u8 dir: 2; ++ u8 type: 2; ++ u8 flags: 2; ++}; ++ ++struct xfrm_mode { ++ u8 encap; ++ u8 family; ++ u8 flags; ++}; ++ ++struct xfrm_type; ++ ++struct xfrm_type_offload; ++ ++struct xfrm_state { ++ possible_net_t xs_net; ++ union { ++ struct hlist_node gclist; ++ struct hlist_node bydst; ++ }; ++ struct hlist_node bysrc; ++ struct hlist_node byspi; ++ struct hlist_node byseq; ++ refcount_t refcnt; ++ spinlock_t lock; ++ struct xfrm_id id; ++ struct xfrm_selector sel; ++ struct xfrm_mark mark; ++ u32 if_id; ++ u32 tfcpad; ++ u32 genid; ++ struct xfrm_state_walk km; ++ struct { ++ u32 reqid; ++ u8 mode; ++ u8 replay_window; ++ u8 aalgo; ++ u8 ealgo; ++ u8 calgo; ++ u8 flags; ++ u16 family; ++ xfrm_address_t saddr; ++ int header_len; ++ int trailer_len; ++ u32 extra_flags; ++ struct xfrm_mark smark; ++ } props; ++ struct xfrm_lifetime_cfg lft; ++ struct xfrm_algo_auth *aalg; ++ struct xfrm_algo *ealg; ++ struct xfrm_algo *calg; ++ struct xfrm_algo_aead *aead; ++ const char *geniv; ++ __be16 new_mapping_sport; ++ u32 new_mapping; ++ u32 mapping_maxage; ++ struct xfrm_encap_tmpl *encap; ++ struct sock *encap_sk; ++ xfrm_address_t *coaddr; ++ struct xfrm_state *tunnel; ++ atomic_t tunnel_users; ++ struct xfrm_replay_state replay; ++ struct xfrm_replay_state_esn *replay_esn; ++ struct xfrm_replay_state preplay; ++ struct xfrm_replay_state_esn *preplay_esn; ++ enum xfrm_replay_mode repl_mode; ++ u32 xflags; ++ u32 replay_maxage; ++ u32 replay_maxdiff; ++ struct timer_list rtimer; ++ struct xfrm_stats stats; ++ struct xfrm_lifetime_cur curlft; ++ struct hrtimer mtimer; ++ struct xfrm_dev_offload xso; ++ long int saved_tmo; ++ time64_t lastused; ++ struct page_frag xfrag; ++ const struct xfrm_type *type; ++ struct xfrm_mode inner_mode; ++ struct xfrm_mode inner_mode_iaf; ++ struct xfrm_mode outer_mode; ++ const struct xfrm_type_offload *type_offload; ++ struct xfrm_sec_ctx *security; ++ void *data; ++}; ++ ++struct xfrm_policy_walk_entry { ++ struct list_head all; ++ u8 dead; ++}; ++ ++struct xfrm_policy_queue { ++ struct sk_buff_head hold_queue; ++ struct timer_list hold_timer; ++ long unsigned int timeout; ++}; ++ ++struct xfrm_tmpl { ++ struct xfrm_id id; ++ xfrm_address_t saddr; ++ short unsigned int encap_family; ++ u32 reqid; ++ u8 mode; ++ u8 share; ++ u8 optional; ++ u8 allalgs; ++ u32 aalgos; ++ u32 ealgos; ++ u32 calgos; ++}; ++ ++struct xfrm_policy { ++ possible_net_t xp_net; ++ struct hlist_node bydst; ++ struct hlist_node byidx; ++ rwlock_t lock; ++ refcount_t refcnt; ++ u32 pos; ++ struct timer_list timer; ++ atomic_t genid; ++ u32 priority; ++ u32 index; ++ u32 if_id; ++ struct xfrm_mark mark; ++ struct xfrm_selector selector; ++ struct xfrm_lifetime_cfg lft; ++ struct xfrm_lifetime_cur curlft; ++ struct xfrm_policy_walk_entry walk; ++ struct xfrm_policy_queue polq; ++ bool bydst_reinsert; ++ u8 type; ++ u8 action; ++ u8 flags; ++ u8 xfrm_nr; ++ u16 family; ++ struct xfrm_sec_ctx *security; ++ struct xfrm_tmpl xfrm_vec[6]; ++ struct hlist_node bydst_inexact_list; ++ struct callback_head rcu; ++ struct xfrm_dev_offload xdo; ++}; ++ ++struct netdev_name_node { ++ struct hlist_node hlist; ++ struct list_head list; ++ struct net_device *dev; ++ const char *name; ++}; ++ ++struct sd_flow_limit; ++ ++struct softnet_data { ++ struct list_head poll_list; ++ struct sk_buff_head process_queue; ++ unsigned int processed; ++ unsigned int time_squeeze; ++ struct softnet_data *rps_ipi_list; ++ bool in_net_rx_action; ++ bool in_napi_threaded_poll; ++ struct sd_flow_limit *flow_limit; ++ struct Qdisc *output_queue; ++ struct Qdisc **output_queue_tailp; ++ struct sk_buff *completion_queue; ++ struct sk_buff_head xfrm_backlog; ++ struct { ++ u16 recursion; ++ u8 more; ++ u8 skip_txqueue; ++ } xmit; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ unsigned int input_queue_head; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ call_single_data_t csd; ++ struct softnet_data *rps_ipi_next; ++ unsigned int cpu; ++ unsigned int input_queue_tail; ++ unsigned int received_rps; ++ unsigned int dropped; ++ struct sk_buff_head input_pkt_queue; ++ struct napi_struct backlog; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ spinlock_t defer_lock; ++ int defer_count; ++ int defer_ipi_scheduled; ++ struct sk_buff *defer_list; ++ long: 64; ++ call_single_data_t defer_csd; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct sd_flow_limit { ++ u64 count; ++ unsigned int num_buckets; ++ unsigned int history_head; ++ u16 history[128]; ++ u8 buckets[0]; ++}; ++ ++enum { ++ TCA_UNSPEC = 0, ++ TCA_KIND = 1, ++ TCA_OPTIONS = 2, ++ TCA_STATS = 3, ++ TCA_XSTATS = 4, ++ TCA_RATE = 5, ++ TCA_FCNT = 6, ++ TCA_STATS2 = 7, ++ TCA_STAB = 8, ++ TCA_PAD = 9, ++ TCA_DUMP_INVISIBLE = 10, ++ TCA_CHAIN = 11, ++ TCA_HW_OFFLOAD = 12, ++ TCA_INGRESS_BLOCK = 13, ++ TCA_EGRESS_BLOCK = 14, ++ TCA_DUMP_FLAGS = 15, ++ TCA_EXT_WARN_MSG = 16, ++ __TCA_MAX = 17, ++}; ++ ++struct dst_metrics { ++ u32 metrics[17]; ++ refcount_t refcnt; ++}; ++ ++struct tcphdr { ++ __be16 source; ++ __be16 dest; ++ __be32 seq; ++ __be32 ack_seq; ++ __u16 doff: 4; ++ __u16 res1: 4; ++ __u16 cwr: 1; ++ __u16 ece: 1; ++ __u16 urg: 1; ++ __u16 ack: 1; ++ __u16 psh: 1; ++ __u16 rst: 1; ++ __u16 syn: 1; ++ __u16 fin: 1; ++ __be16 window; ++ __sum16 check; ++ __be16 urg_ptr; ++}; ++ ++struct udphdr { ++ __be16 source; ++ __be16 dest; ++ __be16 len; ++ __sum16 check; ++}; ++ ++struct vlan_hdr { ++ __be16 h_vlan_TCI; ++ __be16 h_vlan_encapsulated_proto; ++}; ++ ++struct vlan_ethhdr { ++ union { ++ struct { ++ unsigned char h_dest[6]; ++ unsigned char h_source[6]; ++ }; ++ struct { ++ unsigned char h_dest[6]; ++ unsigned char h_source[6]; ++ } addrs; ++ }; ++ __be16 h_vlan_proto; ++ __be16 h_vlan_TCI; ++ __be16 h_vlan_encapsulated_proto; ++}; ++ ++struct qdisc_walker { ++ int stop; ++ int skip; ++ int count; ++ int (*fn)(struct Qdisc *, long unsigned int, struct qdisc_walker *); ++}; ++ ++struct ip_auth_hdr { ++ __u8 nexthdr; ++ __u8 hdrlen; ++ __be16 reserved; ++ __be32 spi; ++ __be32 seq_no; ++ __u8 auth_data[0]; ++}; ++ ++struct frag_hdr { ++ __u8 nexthdr; ++ __u8 reserved; ++ __be16 frag_off; ++ __be32 identification; ++}; ++ ++struct fib_nh_exception { ++ struct fib_nh_exception *fnhe_next; ++ int fnhe_genid; ++ __be32 fnhe_daddr; ++ u32 fnhe_pmtu; ++ bool fnhe_mtu_locked; ++ __be32 fnhe_gw; ++ long unsigned int fnhe_expires; ++ struct rtable *fnhe_rth_input; ++ struct rtable *fnhe_rth_output; ++ long unsigned int fnhe_stamp; ++ struct callback_head rcu; ++}; ++ ++struct rtable { ++ struct dst_entry dst; ++ int rt_genid; ++ unsigned int rt_flags; ++ __u16 rt_type; ++ __u8 rt_is_input; ++ __u8 rt_uses_gateway; ++ int rt_iif; ++ u8 rt_gw_family; ++ union { ++ __be32 rt_gw4; ++ struct in6_addr rt_gw6; ++ }; ++ u32 rt_mtu_locked: 1; ++ u32 rt_pmtu: 31; ++}; ++ ++struct fnhe_hash_bucket { ++ struct fib_nh_exception *chain; ++}; ++ ++enum { ++ SCM_TSTAMP_SND = 0, ++ SCM_TSTAMP_SCHED = 1, ++ SCM_TSTAMP_ACK = 2, ++}; ++ ++struct skb_gso_cb { ++ union { ++ int mac_offset; ++ int data_offset; ++ }; ++ int encap_level; ++ __wsum csum; ++ __u16 csum_start; ++}; ++ ++struct rt6_exception_bucket { ++ struct hlist_head chain; ++ int depth; ++}; ++ ++struct xfrm_type { ++ struct module *owner; ++ u8 proto; ++ u8 flags; ++ int (*init_state)(struct xfrm_state *, struct netlink_ext_ack *); ++ void (*destructor)(struct xfrm_state *); ++ int (*input)(struct xfrm_state *, struct sk_buff *); ++ int (*output)(struct xfrm_state *, struct sk_buff *); ++ int (*reject)(struct xfrm_state *, struct sk_buff *, const struct flowi *); ++}; ++ ++struct xfrm_type_offload { ++ struct module *owner; ++ u8 proto; ++ void (*encap)(struct xfrm_state *, struct sk_buff *); ++ int (*input_tail)(struct xfrm_state *, struct sk_buff *); ++ int (*xmit)(struct xfrm_state *, struct sk_buff *, netdev_features_t); ++}; ++ ++struct xfrm_offload { ++ struct { ++ __u32 low; ++ __u32 hi; ++ } seq; ++ __u32 flags; ++ __u32 status; ++ __u8 proto; ++ __u8 inner_ipproto; ++}; ++ ++struct sec_path { ++ int len; ++ int olen; ++ int verified_cnt; ++ struct xfrm_state *xvec[6]; ++ struct xfrm_offload ovec[1]; ++}; ++ ++struct mpls_shim_hdr { ++ __be32 label_stack_entry; ++}; ++ ++enum skb_drop_reason_subsys { ++ SKB_DROP_REASON_SUBSYS_CORE = 0, ++ SKB_DROP_REASON_SUBSYS_MAC80211_UNUSABLE = 1, ++ SKB_DROP_REASON_SUBSYS_MAC80211_MONITOR = 2, ++ SKB_DROP_REASON_SUBSYS_NUM = 3, ++}; ++ ++struct drop_reason_list { ++ const char * const *reasons; ++ size_t n_reasons; ++}; ++ ++struct page_frag_1k { ++ void *va; ++ u16 offset; ++ bool pfmemalloc; ++}; ++ ++struct napi_alloc_cache { ++ struct page_frag_cache page; ++ struct page_frag_1k page_small; ++ unsigned int skb_count; ++ void *skb_cache[64]; ++}; ++ ++struct skb_free_array { ++ unsigned int skb_count; ++ void *skb_array[16]; ++}; ++ ++typedef int (*sendmsg_func)(struct sock *, struct msghdr *, struct kvec *, size_t, size_t); ++ ++typedef int (*sendpage_func)(struct sock *, struct page *, int, size_t, int); ++ ++struct gnet_stats_rate_est64 { ++ __u64 bps; ++ __u64 pps; ++}; ++ ++struct gnet_estimator { ++ signed char interval; ++ unsigned char ewma_log; ++}; ++ ++struct net_rate_estimator { ++ struct gnet_stats_basic_sync *bstats; ++ spinlock_t *stats_lock; ++ bool running; ++ struct gnet_stats_basic_sync *cpu_bstats; ++ u8 ewma_log; ++ u8 intvl_log; ++ seqcount_t seq; ++ u64 last_packets; ++ u64 last_bytes; ++ u64 avpps; ++ u64 avbps; ++ long unsigned int next_jiffies; ++ struct timer_list timer; ++ struct callback_head rcu; ++}; ++ ++struct net_generic { ++ union { ++ struct { ++ unsigned int len; ++ struct callback_head rcu; ++ } s; ++ struct { ++ struct {} __empty_ptr; ++ void *ptr[0]; ++ }; ++ }; ++}; ++ ++struct pernet_operations { ++ struct list_head list; ++ int (*init)(struct net *); ++ void (*pre_exit)(struct net *); ++ void (*exit)(struct net *); ++ void (*exit_batch)(struct list_head *); ++ unsigned int *id; ++ size_t size; ++}; ++ ++enum { ++ RTM_BASE = 16, ++ RTM_NEWLINK = 16, ++ RTM_DELLINK = 17, ++ RTM_GETLINK = 18, ++ RTM_SETLINK = 19, ++ RTM_NEWADDR = 20, ++ RTM_DELADDR = 21, ++ RTM_GETADDR = 22, ++ RTM_NEWROUTE = 24, ++ RTM_DELROUTE = 25, ++ RTM_GETROUTE = 26, ++ RTM_NEWNEIGH = 28, ++ RTM_DELNEIGH = 29, ++ RTM_GETNEIGH = 30, ++ RTM_NEWRULE = 32, ++ RTM_DELRULE = 33, ++ RTM_GETRULE = 34, ++ RTM_NEWQDISC = 36, ++ RTM_DELQDISC = 37, ++ RTM_GETQDISC = 38, ++ RTM_NEWTCLASS = 40, ++ RTM_DELTCLASS = 41, ++ RTM_GETTCLASS = 42, ++ RTM_NEWTFILTER = 44, ++ RTM_DELTFILTER = 45, ++ RTM_GETTFILTER = 46, ++ RTM_NEWACTION = 48, ++ RTM_DELACTION = 49, ++ RTM_GETACTION = 50, ++ RTM_NEWPREFIX = 52, ++ RTM_GETMULTICAST = 58, ++ RTM_GETANYCAST = 62, ++ RTM_NEWNEIGHTBL = 64, ++ RTM_GETNEIGHTBL = 66, ++ RTM_SETNEIGHTBL = 67, ++ RTM_NEWNDUSEROPT = 68, ++ RTM_NEWADDRLABEL = 72, ++ RTM_DELADDRLABEL = 73, ++ RTM_GETADDRLABEL = 74, ++ RTM_GETDCB = 78, ++ RTM_SETDCB = 79, ++ RTM_NEWNETCONF = 80, ++ RTM_DELNETCONF = 81, ++ RTM_GETNETCONF = 82, ++ RTM_NEWMDB = 84, ++ RTM_DELMDB = 85, ++ RTM_GETMDB = 86, ++ RTM_NEWNSID = 88, ++ RTM_DELNSID = 89, ++ RTM_GETNSID = 90, ++ RTM_NEWSTATS = 92, ++ RTM_GETSTATS = 94, ++ RTM_SETSTATS = 95, ++ RTM_NEWCACHEREPORT = 96, ++ RTM_NEWCHAIN = 100, ++ RTM_DELCHAIN = 101, ++ RTM_GETCHAIN = 102, ++ RTM_NEWNEXTHOP = 104, ++ RTM_DELNEXTHOP = 105, ++ RTM_GETNEXTHOP = 106, ++ RTM_NEWLINKPROP = 108, ++ RTM_DELLINKPROP = 109, ++ RTM_GETLINKPROP = 110, ++ RTM_NEWVLAN = 112, ++ RTM_DELVLAN = 113, ++ RTM_GETVLAN = 114, ++ RTM_NEWNEXTHOPBUCKET = 116, ++ RTM_DELNEXTHOPBUCKET = 117, ++ RTM_GETNEXTHOPBUCKET = 118, ++ RTM_NEWTUNNEL = 120, ++ RTM_DELTUNNEL = 121, ++ RTM_GETTUNNEL = 122, ++ __RTM_MAX = 123, ++}; ++ ++struct rtgenmsg { ++ unsigned char rtgen_family; ++}; ++ ++enum rtnetlink_groups { ++ RTNLGRP_NONE = 0, ++ RTNLGRP_LINK = 1, ++ RTNLGRP_NOTIFY = 2, ++ RTNLGRP_NEIGH = 3, ++ RTNLGRP_TC = 4, ++ RTNLGRP_IPV4_IFADDR = 5, ++ RTNLGRP_IPV4_MROUTE = 6, ++ RTNLGRP_IPV4_ROUTE = 7, ++ RTNLGRP_IPV4_RULE = 8, ++ RTNLGRP_IPV6_IFADDR = 9, ++ RTNLGRP_IPV6_MROUTE = 10, ++ RTNLGRP_IPV6_ROUTE = 11, ++ RTNLGRP_IPV6_IFINFO = 12, ++ RTNLGRP_DECnet_IFADDR = 13, ++ RTNLGRP_NOP2 = 14, ++ RTNLGRP_DECnet_ROUTE = 15, ++ RTNLGRP_DECnet_RULE = 16, ++ RTNLGRP_NOP4 = 17, ++ RTNLGRP_IPV6_PREFIX = 18, ++ RTNLGRP_IPV6_RULE = 19, ++ RTNLGRP_ND_USEROPT = 20, ++ RTNLGRP_PHONET_IFADDR = 21, ++ RTNLGRP_PHONET_ROUTE = 22, ++ RTNLGRP_DCB = 23, ++ RTNLGRP_IPV4_NETCONF = 24, ++ RTNLGRP_IPV6_NETCONF = 25, ++ RTNLGRP_MDB = 26, ++ RTNLGRP_MPLS_ROUTE = 27, ++ RTNLGRP_NSID = 28, ++ RTNLGRP_MPLS_NETCONF = 29, ++ RTNLGRP_IPV4_MROUTE_R = 30, ++ RTNLGRP_IPV6_MROUTE_R = 31, ++ RTNLGRP_NEXTHOP = 32, ++ RTNLGRP_BRVLAN = 33, ++ RTNLGRP_MCTP_IFADDR = 34, ++ RTNLGRP_TUNNEL = 35, ++ RTNLGRP_STATS = 36, ++ __RTNLGRP_MAX = 37, ++}; ++ ++enum { ++ NETNSA_NONE = 0, ++ NETNSA_NSID = 1, ++ NETNSA_PID = 2, ++ NETNSA_FD = 3, ++ NETNSA_TARGET_NSID = 4, ++ NETNSA_CURRENT_NSID = 5, ++ __NETNSA_MAX = 6, ++}; ++ ++struct pcpu_gen_cookie { ++ local_t nesting; ++ u64 last; ++}; ++ ++struct gen_cookie { ++ struct pcpu_gen_cookie *local; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ atomic64_t forward_last; ++ atomic64_t reverse_last; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum netlink_validation { ++ NL_VALIDATE_LIBERAL = 0, ++ NL_VALIDATE_TRAILING = 1, ++ NL_VALIDATE_MAXTYPE = 2, ++ NL_VALIDATE_UNSPEC = 4, ++ NL_VALIDATE_STRICT_ATTRS = 8, ++ NL_VALIDATE_NESTED = 16, ++}; ++ ++typedef int (*rtnl_doit_func)(struct sk_buff *, struct nlmsghdr *, struct netlink_ext_ack *); ++ ++typedef int (*rtnl_dumpit_func)(struct sk_buff *, struct netlink_callback *); ++ ++enum rtnl_link_flags { ++ RTNL_FLAG_DOIT_UNLOCKED = 1, ++ RTNL_FLAG_BULK_DEL_SUPPORTED = 2, ++}; ++ ++struct net_fill_args { ++ u32 portid; ++ u32 seq; ++ int flags; ++ int cmd; ++ int nsid; ++ bool add_ref; ++ int ref_nsid; ++}; ++ ++struct rtnl_net_dump_cb { ++ struct net *tgt_net; ++ struct net *ref_net; ++ struct sk_buff *skb; ++ struct net_fill_args fillargs; ++ int idx; ++ int s_idx; ++}; ++ ++struct rps_sock_flow_table { ++ u32 mask; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ u32 ents[0]; ++}; ++ ++struct netdev_hw_addr { ++ struct list_head list; ++ struct rb_node node; ++ unsigned char addr[32]; ++ unsigned char type; ++ bool global_use; ++ int sync_cnt; ++ int refcount; ++ int synced; ++ struct callback_head callback_head; ++}; ++ ++enum netdev_cmd { ++ NETDEV_UP = 1, ++ NETDEV_DOWN = 2, ++ NETDEV_REBOOT = 3, ++ NETDEV_CHANGE = 4, ++ NETDEV_REGISTER = 5, ++ NETDEV_UNREGISTER = 6, ++ NETDEV_CHANGEMTU = 7, ++ NETDEV_CHANGEADDR = 8, ++ NETDEV_PRE_CHANGEADDR = 9, ++ NETDEV_GOING_DOWN = 10, ++ NETDEV_CHANGENAME = 11, ++ NETDEV_FEAT_CHANGE = 12, ++ NETDEV_BONDING_FAILOVER = 13, ++ NETDEV_PRE_UP = 14, ++ NETDEV_PRE_TYPE_CHANGE = 15, ++ NETDEV_POST_TYPE_CHANGE = 16, ++ NETDEV_POST_INIT = 17, ++ NETDEV_PRE_UNINIT = 18, ++ NETDEV_RELEASE = 19, ++ NETDEV_NOTIFY_PEERS = 20, ++ NETDEV_JOIN = 21, ++ NETDEV_CHANGEUPPER = 22, ++ NETDEV_RESEND_IGMP = 23, ++ NETDEV_PRECHANGEMTU = 24, ++ NETDEV_CHANGEINFODATA = 25, ++ NETDEV_BONDING_INFO = 26, ++ NETDEV_PRECHANGEUPPER = 27, ++ NETDEV_CHANGELOWERSTATE = 28, ++ NETDEV_UDP_TUNNEL_PUSH_INFO = 29, ++ NETDEV_UDP_TUNNEL_DROP_INFO = 30, ++ NETDEV_CHANGE_TX_QUEUE_LEN = 31, ++ NETDEV_CVLAN_FILTER_PUSH_INFO = 32, ++ NETDEV_CVLAN_FILTER_DROP_INFO = 33, ++ NETDEV_SVLAN_FILTER_PUSH_INFO = 34, ++ NETDEV_SVLAN_FILTER_DROP_INFO = 35, ++ NETDEV_OFFLOAD_XSTATS_ENABLE = 36, ++ NETDEV_OFFLOAD_XSTATS_DISABLE = 37, ++ NETDEV_OFFLOAD_XSTATS_REPORT_USED = 38, ++ NETDEV_OFFLOAD_XSTATS_REPORT_DELTA = 39, ++ NETDEV_XDP_FEAT_CHANGE = 40, ++}; ++ ++enum macsec_validation_type { ++ MACSEC_VALIDATE_DISABLED = 0, ++ MACSEC_VALIDATE_CHECK = 1, ++ MACSEC_VALIDATE_STRICT = 2, ++ __MACSEC_VALIDATE_END = 3, ++ MACSEC_VALIDATE_MAX = 2, ++}; ++ ++enum macsec_offload { ++ MACSEC_OFFLOAD_OFF = 0, ++ MACSEC_OFFLOAD_PHY = 1, ++ MACSEC_OFFLOAD_MAC = 2, ++ __MACSEC_OFFLOAD_END = 3, ++ MACSEC_OFFLOAD_MAX = 2, ++}; ++ ++struct ip_tunnel_parm { ++ char name[16]; ++ int link; ++ __be16 i_flags; ++ __be16 o_flags; ++ __be32 i_key; ++ __be32 o_key; ++ struct iphdr iph; ++}; ++ ++struct fib_info; ++ ++struct fib_nh { ++ struct fib_nh_common nh_common; ++ struct hlist_node nh_hash; ++ struct fib_info *nh_parent; ++ __u32 nh_tclassid; ++ __be32 nh_saddr; ++ int nh_saddr_genid; ++}; ++ ++struct fib_info { ++ struct hlist_node fib_hash; ++ struct hlist_node fib_lhash; ++ struct list_head nh_list; ++ struct net *fib_net; ++ refcount_t fib_treeref; ++ refcount_t fib_clntref; ++ unsigned int fib_flags; ++ unsigned char fib_dead; ++ unsigned char fib_protocol; ++ unsigned char fib_scope; ++ unsigned char fib_type; ++ __be32 fib_prefsrc; ++ u32 fib_tb_id; ++ u32 fib_priority; ++ struct dst_metrics *fib_metrics; ++ int fib_nhs; ++ bool fib_nh_is_v6; ++ bool nh_updated; ++ struct nexthop *nh; ++ struct callback_head rcu; ++ struct fib_nh fib_nh[0]; ++}; ++ ++struct nh_info; ++ ++struct nh_group; ++ ++struct nexthop { ++ struct rb_node rb_node; ++ struct list_head fi_list; ++ struct list_head f6i_list; ++ struct list_head fdb_list; ++ struct list_head grp_list; ++ struct net *net; ++ u32 id; ++ u8 protocol; ++ u8 nh_flags; ++ bool is_group; ++ refcount_t refcnt; ++ struct callback_head rcu; ++ union { ++ struct nh_info *nh_info; ++ struct nh_group *nh_grp; ++ }; ++}; ++ ++struct xfrm_dst { ++ union { ++ struct dst_entry dst; ++ struct rtable rt; ++ struct rt6_info rt6; ++ } u; ++ struct dst_entry *route; ++ struct dst_entry *child; ++ struct dst_entry *path; ++ struct xfrm_policy *pols[2]; ++ int num_pols; ++ int num_xfrms; ++ u32 xfrm_genid; ++ u32 policy_genid; ++ u32 route_mtu_cached; ++ u32 child_mtu_cached; ++ u32 route_cookie; ++ u32 path_cookie; ++}; ++ ++enum metadata_type { ++ METADATA_IP_TUNNEL = 0, ++ METADATA_HW_PORT_MUX = 1, ++ METADATA_MACSEC = 2, ++ METADATA_XFRM = 3, ++}; ++ ++struct ip_tunnel_key { ++ __be64 tun_id; ++ union { ++ struct { ++ __be32 src; ++ __be32 dst; ++ } ipv4; ++ struct { ++ struct in6_addr src; ++ struct in6_addr dst; ++ } ipv6; ++ } u; ++ __be16 tun_flags; ++ u8 tos; ++ u8 ttl; ++ __be32 label; ++ __be16 tp_src; ++ __be16 tp_dst; ++ __u8 flow_flags; ++}; ++ ++struct ip_tunnel_encap { ++ u16 type; ++ u16 flags; ++ __be16 sport; ++ __be16 dport; ++}; ++ ++struct dst_cache_pcpu; ++ ++struct dst_cache { ++ struct dst_cache_pcpu *cache; ++ long unsigned int reset_ts; ++}; ++ ++struct ip_tunnel_info { ++ struct ip_tunnel_key key; ++ struct ip_tunnel_encap encap; ++ struct dst_cache dst_cache; ++ u8 options_len; ++ u8 mode; ++}; ++ ++struct hw_port_info { ++ struct net_device *lower_dev; ++ u32 port_id; ++}; ++ ++typedef u64 sci_t; ++ ++struct macsec_info { ++ sci_t sci; ++}; ++ ++struct xfrm_md_info { ++ u32 if_id; ++ int link; ++ struct dst_entry *dst_orig; ++}; ++ ++struct metadata_dst { ++ struct dst_entry dst; ++ enum metadata_type type; ++ union { ++ struct ip_tunnel_info tun_info; ++ struct hw_port_info port_info; ++ struct macsec_info macsec_info; ++ struct xfrm_md_info xfrm_info; ++ } u; ++}; ++ ++struct nh_info { ++ struct hlist_node dev_hash; ++ struct nexthop *nh_parent; ++ u8 family; ++ bool reject_nh; ++ bool fdb_nh; ++ union { ++ struct fib_nh_common fib_nhc; ++ struct fib_nh fib_nh; ++ struct fib6_nh fib6_nh; ++ }; ++}; ++ ++struct nh_grp_entry; ++ ++struct nh_res_bucket { ++ struct nh_grp_entry *nh_entry; ++ atomic_long_t used_time; ++ long unsigned int migrated_time; ++ bool occupied; ++ u8 nh_flags; ++}; ++ ++struct nh_grp_entry { ++ struct nexthop *nh; ++ u8 weight; ++ union { ++ struct { ++ atomic_t upper_bound; ++ } hthr; ++ struct { ++ struct list_head uw_nh_entry; ++ u16 count_buckets; ++ u16 wants_buckets; ++ } res; ++ }; ++ struct list_head nh_list; ++ struct nexthop *nh_parent; ++}; ++ ++struct nh_res_table { ++ struct net *net; ++ u32 nhg_id; ++ struct delayed_work upkeep_dw; ++ struct list_head uw_nh_entries; ++ long unsigned int unbalanced_since; ++ u32 idle_timer; ++ u32 unbalanced_timer; ++ u16 num_nh_buckets; ++ struct nh_res_bucket nh_buckets[0]; ++}; ++ ++struct nh_group { ++ struct nh_group *spare; ++ u16 num_nh; ++ bool is_multipath; ++ bool hash_threshold; ++ bool resilient; ++ bool fdb_nh; ++ bool has_v4; ++ struct nh_res_table *res_table; ++ struct nh_grp_entry nh_entries[0]; ++}; ++ ++typedef u32 ssci_t; ++ ++union salt { ++ struct { ++ u32 ssci; ++ u64 pn; ++ } __attribute__((packed)); ++ u8 bytes[12]; ++}; ++ ++typedef union salt salt_t; ++ ++union pn { ++ struct { ++ u32 upper; ++ u32 lower; ++ }; ++ u64 full64; ++}; ++ ++typedef union pn pn_t; ++ ++struct macsec_key { ++ u8 id[16]; ++ struct crypto_aead *tfm; ++ salt_t salt; ++}; ++ ++struct macsec_rx_sc_stats { ++ __u64 InOctetsValidated; ++ __u64 InOctetsDecrypted; ++ __u64 InPktsUnchecked; ++ __u64 InPktsDelayed; ++ __u64 InPktsOK; ++ __u64 InPktsInvalid; ++ __u64 InPktsLate; ++ __u64 InPktsNotValid; ++ __u64 InPktsNotUsingSA; ++ __u64 InPktsUnusedSA; ++}; ++ ++struct macsec_rx_sa_stats { ++ __u32 InPktsOK; ++ __u32 InPktsInvalid; ++ __u32 InPktsNotValid; ++ __u32 InPktsNotUsingSA; ++ __u32 InPktsUnusedSA; ++}; ++ ++struct macsec_tx_sa_stats { ++ __u32 OutPktsProtected; ++ __u32 OutPktsEncrypted; ++}; ++ ++struct macsec_tx_sc_stats { ++ __u64 OutPktsProtected; ++ __u64 OutPktsEncrypted; ++ __u64 OutOctetsProtected; ++ __u64 OutOctetsEncrypted; ++}; ++ ++struct macsec_dev_stats { ++ __u64 OutPktsUntagged; ++ __u64 InPktsUntagged; ++ __u64 OutPktsTooLong; ++ __u64 InPktsNoTag; ++ __u64 InPktsBadTag; ++ __u64 InPktsUnknownSCI; ++ __u64 InPktsNoSCI; ++ __u64 InPktsOverrun; ++}; ++ ++struct macsec_rx_sc; ++ ++struct macsec_rx_sa { ++ struct macsec_key key; ++ ssci_t ssci; ++ spinlock_t lock; ++ union { ++ pn_t next_pn_halves; ++ u64 next_pn; ++ }; ++ refcount_t refcnt; ++ bool active; ++ struct macsec_rx_sa_stats *stats; ++ struct macsec_rx_sc *sc; ++ struct callback_head rcu; ++}; ++ ++struct pcpu_rx_sc_stats; ++ ++struct macsec_rx_sc { ++ struct macsec_rx_sc *next; ++ sci_t sci; ++ bool active; ++ struct macsec_rx_sa *sa[4]; ++ struct pcpu_rx_sc_stats *stats; ++ refcount_t refcnt; ++ struct callback_head callback_head; ++}; ++ ++struct pcpu_rx_sc_stats { ++ struct macsec_rx_sc_stats stats; ++ struct u64_stats_sync syncp; ++}; ++ ++struct pcpu_tx_sc_stats { ++ struct macsec_tx_sc_stats stats; ++ struct u64_stats_sync syncp; ++}; ++ ++struct macsec_tx_sa { ++ struct macsec_key key; ++ ssci_t ssci; ++ spinlock_t lock; ++ union { ++ pn_t next_pn_halves; ++ u64 next_pn; ++ }; ++ refcount_t refcnt; ++ bool active; ++ struct macsec_tx_sa_stats *stats; ++ struct callback_head rcu; ++}; ++ ++struct macsec_tx_sc { ++ bool active; ++ u8 encoding_sa; ++ bool encrypt; ++ bool send_sci; ++ bool end_station; ++ bool scb; ++ struct macsec_tx_sa *sa[4]; ++ struct pcpu_tx_sc_stats *stats; ++ struct metadata_dst *md_dst; ++}; ++ ++struct macsec_secy { ++ struct net_device *netdev; ++ unsigned int n_rx_sc; ++ sci_t sci; ++ u16 key_len; ++ u16 icv_len; ++ enum macsec_validation_type validate_frames; ++ bool xpn; ++ bool operational; ++ bool protect_frames; ++ bool replay_protect; ++ u32 replay_window; ++ struct macsec_tx_sc tx_sc; ++ struct macsec_rx_sc *rx_sc; ++}; ++ ++struct macsec_context { ++ union { ++ struct net_device *netdev; ++ struct phy_device *phydev; ++ }; ++ enum macsec_offload offload; ++ struct macsec_secy *secy; ++ struct macsec_rx_sc *rx_sc; ++ struct { ++ unsigned char assoc_num; ++ u8 key[128]; ++ union { ++ struct macsec_rx_sa *rx_sa; ++ struct macsec_tx_sa *tx_sa; ++ }; ++ } sa; ++ union { ++ struct macsec_tx_sc_stats *tx_sc_stats; ++ struct macsec_tx_sa_stats *tx_sa_stats; ++ struct macsec_rx_sc_stats *rx_sc_stats; ++ struct macsec_rx_sa_stats *rx_sa_stats; ++ struct macsec_dev_stats *dev_stats; ++ } stats; ++}; ++ ++typedef __kernel_clock_t clock_t; ++ ++enum net_device_flags { ++ IFF_UP = 1, ++ IFF_BROADCAST = 2, ++ IFF_DEBUG = 4, ++ IFF_LOOPBACK = 8, ++ IFF_POINTOPOINT = 16, ++ IFF_NOTRAILERS = 32, ++ IFF_RUNNING = 64, ++ IFF_NOARP = 128, ++ IFF_PROMISC = 256, ++ IFF_ALLMULTI = 512, ++ IFF_MASTER = 1024, ++ IFF_SLAVE = 2048, ++ IFF_MULTICAST = 4096, ++ IFF_PORTSEL = 8192, ++ IFF_AUTOMEDIA = 16384, ++ IFF_DYNAMIC = 32768, ++ IFF_LOWER_UP = 65536, ++ IFF_DORMANT = 131072, ++ IFF_ECHO = 262144, ++}; ++ ++enum { ++ IF_OPER_UNKNOWN = 0, ++ IF_OPER_NOTPRESENT = 1, ++ IF_OPER_DOWN = 2, ++ IF_OPER_LOWERLAYERDOWN = 3, ++ IF_OPER_TESTING = 4, ++ IF_OPER_DORMANT = 5, ++ IF_OPER_UP = 6, ++}; ++ ++struct netlink_dump_control { ++ int (*start)(struct netlink_callback *); ++ int (*dump)(struct sk_buff *, struct netlink_callback *); ++ int (*done)(struct netlink_callback *); ++ void *data; ++ struct module *module; ++ u32 min_dump_alloc; ++}; ++ ++struct udp_hslot; ++ ++struct udp_table { ++ struct udp_hslot *hash; ++ struct udp_hslot *hash2; ++ unsigned int mask; ++ unsigned int log; ++}; ++ ++enum { ++ NDA_UNSPEC = 0, ++ NDA_DST = 1, ++ NDA_LLADDR = 2, ++ NDA_CACHEINFO = 3, ++ NDA_PROBES = 4, ++ NDA_VLAN = 5, ++ NDA_PORT = 6, ++ NDA_VNI = 7, ++ NDA_IFINDEX = 8, ++ NDA_MASTER = 9, ++ NDA_LINK_NETNSID = 10, ++ NDA_SRC_VNI = 11, ++ NDA_PROTOCOL = 12, ++ NDA_NH_ID = 13, ++ NDA_FDB_EXT_ATTRS = 14, ++ NDA_FLAGS_EXT = 15, ++ NDA_NDM_STATE_MASK = 16, ++ NDA_NDM_FLAGS_MASK = 17, ++ __NDA_MAX = 18, ++}; ++ ++struct rtnl_link_stats { ++ __u32 rx_packets; ++ __u32 tx_packets; ++ __u32 rx_bytes; ++ __u32 tx_bytes; ++ __u32 rx_errors; ++ __u32 tx_errors; ++ __u32 rx_dropped; ++ __u32 tx_dropped; ++ __u32 multicast; ++ __u32 collisions; ++ __u32 rx_length_errors; ++ __u32 rx_over_errors; ++ __u32 rx_crc_errors; ++ __u32 rx_frame_errors; ++ __u32 rx_fifo_errors; ++ __u32 rx_missed_errors; ++ __u32 tx_aborted_errors; ++ __u32 tx_carrier_errors; ++ __u32 tx_fifo_errors; ++ __u32 tx_heartbeat_errors; ++ __u32 tx_window_errors; ++ __u32 rx_compressed; ++ __u32 tx_compressed; ++ __u32 rx_nohandler; ++}; ++ ++struct rtnl_link_ifmap { ++ __u64 mem_start; ++ __u64 mem_end; ++ __u64 base_addr; ++ __u16 irq; ++ __u8 dma; ++ __u8 port; ++}; ++ ++enum { ++ IFLA_UNSPEC = 0, ++ IFLA_ADDRESS = 1, ++ IFLA_BROADCAST = 2, ++ IFLA_IFNAME = 3, ++ IFLA_MTU = 4, ++ IFLA_LINK = 5, ++ IFLA_QDISC = 6, ++ IFLA_STATS = 7, ++ IFLA_COST = 8, ++ IFLA_PRIORITY = 9, ++ IFLA_MASTER = 10, ++ IFLA_WIRELESS = 11, ++ IFLA_PROTINFO = 12, ++ IFLA_TXQLEN = 13, ++ IFLA_MAP = 14, ++ IFLA_WEIGHT = 15, ++ IFLA_OPERSTATE = 16, ++ IFLA_LINKMODE = 17, ++ IFLA_LINKINFO = 18, ++ IFLA_NET_NS_PID = 19, ++ IFLA_IFALIAS = 20, ++ IFLA_NUM_VF = 21, ++ IFLA_VFINFO_LIST = 22, ++ IFLA_STATS64 = 23, ++ IFLA_VF_PORTS = 24, ++ IFLA_PORT_SELF = 25, ++ IFLA_AF_SPEC = 26, ++ IFLA_GROUP = 27, ++ IFLA_NET_NS_FD = 28, ++ IFLA_EXT_MASK = 29, ++ IFLA_PROMISCUITY = 30, ++ IFLA_NUM_TX_QUEUES = 31, ++ IFLA_NUM_RX_QUEUES = 32, ++ IFLA_CARRIER = 33, ++ IFLA_PHYS_PORT_ID = 34, ++ IFLA_CARRIER_CHANGES = 35, ++ IFLA_PHYS_SWITCH_ID = 36, ++ IFLA_LINK_NETNSID = 37, ++ IFLA_PHYS_PORT_NAME = 38, ++ IFLA_PROTO_DOWN = 39, ++ IFLA_GSO_MAX_SEGS = 40, ++ IFLA_GSO_MAX_SIZE = 41, ++ IFLA_PAD = 42, ++ IFLA_XDP = 43, ++ IFLA_EVENT = 44, ++ IFLA_NEW_NETNSID = 45, ++ IFLA_IF_NETNSID = 46, ++ IFLA_TARGET_NETNSID = 46, ++ IFLA_CARRIER_UP_COUNT = 47, ++ IFLA_CARRIER_DOWN_COUNT = 48, ++ IFLA_NEW_IFINDEX = 49, ++ IFLA_MIN_MTU = 50, ++ IFLA_MAX_MTU = 51, ++ IFLA_PROP_LIST = 52, ++ IFLA_ALT_IFNAME = 53, ++ IFLA_PERM_ADDRESS = 54, ++ IFLA_PROTO_DOWN_REASON = 55, ++ IFLA_PARENT_DEV_NAME = 56, ++ IFLA_PARENT_DEV_BUS_NAME = 57, ++ IFLA_GRO_MAX_SIZE = 58, ++ IFLA_TSO_MAX_SIZE = 59, ++ IFLA_TSO_MAX_SEGS = 60, ++ IFLA_ALLMULTI = 61, ++ IFLA_DEVLINK_PORT = 62, ++ IFLA_GSO_IPV4_MAX_SIZE = 63, ++ IFLA_GRO_IPV4_MAX_SIZE = 64, ++ __IFLA_MAX = 65, ++}; ++ ++enum { ++ IFLA_PROTO_DOWN_REASON_UNSPEC = 0, ++ IFLA_PROTO_DOWN_REASON_MASK = 1, ++ IFLA_PROTO_DOWN_REASON_VALUE = 2, ++ __IFLA_PROTO_DOWN_REASON_CNT = 3, ++ IFLA_PROTO_DOWN_REASON_MAX = 2, ++}; ++ ++enum { ++ IFLA_BRPORT_UNSPEC = 0, ++ IFLA_BRPORT_STATE = 1, ++ IFLA_BRPORT_PRIORITY = 2, ++ IFLA_BRPORT_COST = 3, ++ IFLA_BRPORT_MODE = 4, ++ IFLA_BRPORT_GUARD = 5, ++ IFLA_BRPORT_PROTECT = 6, ++ IFLA_BRPORT_FAST_LEAVE = 7, ++ IFLA_BRPORT_LEARNING = 8, ++ IFLA_BRPORT_UNICAST_FLOOD = 9, ++ IFLA_BRPORT_PROXYARP = 10, ++ IFLA_BRPORT_LEARNING_SYNC = 11, ++ IFLA_BRPORT_PROXYARP_WIFI = 12, ++ IFLA_BRPORT_ROOT_ID = 13, ++ IFLA_BRPORT_BRIDGE_ID = 14, ++ IFLA_BRPORT_DESIGNATED_PORT = 15, ++ IFLA_BRPORT_DESIGNATED_COST = 16, ++ IFLA_BRPORT_ID = 17, ++ IFLA_BRPORT_NO = 18, ++ IFLA_BRPORT_TOPOLOGY_CHANGE_ACK = 19, ++ IFLA_BRPORT_CONFIG_PENDING = 20, ++ IFLA_BRPORT_MESSAGE_AGE_TIMER = 21, ++ IFLA_BRPORT_FORWARD_DELAY_TIMER = 22, ++ IFLA_BRPORT_HOLD_TIMER = 23, ++ IFLA_BRPORT_FLUSH = 24, ++ IFLA_BRPORT_MULTICAST_ROUTER = 25, ++ IFLA_BRPORT_PAD = 26, ++ IFLA_BRPORT_MCAST_FLOOD = 27, ++ IFLA_BRPORT_MCAST_TO_UCAST = 28, ++ IFLA_BRPORT_VLAN_TUNNEL = 29, ++ IFLA_BRPORT_BCAST_FLOOD = 30, ++ IFLA_BRPORT_GROUP_FWD_MASK = 31, ++ IFLA_BRPORT_NEIGH_SUPPRESS = 32, ++ IFLA_BRPORT_ISOLATED = 33, ++ IFLA_BRPORT_BACKUP_PORT = 34, ++ IFLA_BRPORT_MRP_RING_OPEN = 35, ++ IFLA_BRPORT_MRP_IN_OPEN = 36, ++ IFLA_BRPORT_MCAST_EHT_HOSTS_LIMIT = 37, ++ IFLA_BRPORT_MCAST_EHT_HOSTS_CNT = 38, ++ IFLA_BRPORT_LOCKED = 39, ++ IFLA_BRPORT_MAB = 40, ++ IFLA_BRPORT_MCAST_N_GROUPS = 41, ++ IFLA_BRPORT_MCAST_MAX_GROUPS = 42, ++ IFLA_BRPORT_NEIGH_VLAN_SUPPRESS = 43, ++ __IFLA_BRPORT_MAX = 44, ++}; ++ ++enum { ++ IFLA_INFO_UNSPEC = 0, ++ IFLA_INFO_KIND = 1, ++ IFLA_INFO_DATA = 2, ++ IFLA_INFO_XSTATS = 3, ++ IFLA_INFO_SLAVE_KIND = 4, ++ IFLA_INFO_SLAVE_DATA = 5, ++ __IFLA_INFO_MAX = 6, ++}; ++ ++enum { ++ IFLA_VF_INFO_UNSPEC = 0, ++ IFLA_VF_INFO = 1, ++ __IFLA_VF_INFO_MAX = 2, ++}; ++ ++enum { ++ IFLA_VF_UNSPEC = 0, ++ IFLA_VF_MAC = 1, ++ IFLA_VF_VLAN = 2, ++ IFLA_VF_TX_RATE = 3, ++ IFLA_VF_SPOOFCHK = 4, ++ IFLA_VF_LINK_STATE = 5, ++ IFLA_VF_RATE = 6, ++ IFLA_VF_RSS_QUERY_EN = 7, ++ IFLA_VF_STATS = 8, ++ IFLA_VF_TRUST = 9, ++ IFLA_VF_IB_NODE_GUID = 10, ++ IFLA_VF_IB_PORT_GUID = 11, ++ IFLA_VF_VLAN_LIST = 12, ++ IFLA_VF_BROADCAST = 13, ++ __IFLA_VF_MAX = 14, ++}; ++ ++struct ifla_vf_mac { ++ __u32 vf; ++ __u8 mac[32]; ++}; ++ ++struct ifla_vf_broadcast { ++ __u8 broadcast[32]; ++}; ++ ++struct ifla_vf_vlan { ++ __u32 vf; ++ __u32 vlan; ++ __u32 qos; ++}; ++ ++enum { ++ IFLA_VF_VLAN_INFO_UNSPEC = 0, ++ IFLA_VF_VLAN_INFO = 1, ++ __IFLA_VF_VLAN_INFO_MAX = 2, ++}; ++ ++struct ifla_vf_vlan_info { ++ __u32 vf; ++ __u32 vlan; ++ __u32 qos; ++ __be16 vlan_proto; ++}; ++ ++struct ifla_vf_tx_rate { ++ __u32 vf; ++ __u32 rate; ++}; ++ ++struct ifla_vf_rate { ++ __u32 vf; ++ __u32 min_tx_rate; ++ __u32 max_tx_rate; ++}; ++ ++struct ifla_vf_spoofchk { ++ __u32 vf; ++ __u32 setting; ++}; ++ ++struct ifla_vf_link_state { ++ __u32 vf; ++ __u32 link_state; ++}; ++ ++struct ifla_vf_rss_query_en { ++ __u32 vf; ++ __u32 setting; ++}; ++ ++enum { ++ IFLA_VF_STATS_RX_PACKETS = 0, ++ IFLA_VF_STATS_TX_PACKETS = 1, ++ IFLA_VF_STATS_RX_BYTES = 2, ++ IFLA_VF_STATS_TX_BYTES = 3, ++ IFLA_VF_STATS_BROADCAST = 4, ++ IFLA_VF_STATS_MULTICAST = 5, ++ IFLA_VF_STATS_PAD = 6, ++ IFLA_VF_STATS_RX_DROPPED = 7, ++ IFLA_VF_STATS_TX_DROPPED = 8, ++ __IFLA_VF_STATS_MAX = 9, ++}; ++ ++struct ifla_vf_trust { ++ __u32 vf; ++ __u32 setting; ++}; ++ ++enum { ++ IFLA_VF_PORT_UNSPEC = 0, ++ IFLA_VF_PORT = 1, ++ __IFLA_VF_PORT_MAX = 2, ++}; ++ ++enum { ++ IFLA_PORT_UNSPEC = 0, ++ IFLA_PORT_VF = 1, ++ IFLA_PORT_PROFILE = 2, ++ IFLA_PORT_VSI_TYPE = 3, ++ IFLA_PORT_INSTANCE_UUID = 4, ++ IFLA_PORT_HOST_UUID = 5, ++ IFLA_PORT_REQUEST = 6, ++ IFLA_PORT_RESPONSE = 7, ++ __IFLA_PORT_MAX = 8, ++}; ++ ++struct if_stats_msg { ++ __u8 family; ++ __u8 pad1; ++ __u16 pad2; ++ __u32 ifindex; ++ __u32 filter_mask; ++}; ++ ++enum { ++ IFLA_STATS_UNSPEC = 0, ++ IFLA_STATS_LINK_64 = 1, ++ IFLA_STATS_LINK_XSTATS = 2, ++ IFLA_STATS_LINK_XSTATS_SLAVE = 3, ++ IFLA_STATS_LINK_OFFLOAD_XSTATS = 4, ++ IFLA_STATS_AF_SPEC = 5, ++ __IFLA_STATS_MAX = 6, ++}; ++ ++enum { ++ IFLA_STATS_GETSET_UNSPEC = 0, ++ IFLA_STATS_GET_FILTERS = 1, ++ IFLA_STATS_SET_OFFLOAD_XSTATS_L3_STATS = 2, ++ __IFLA_STATS_GETSET_MAX = 3, ++}; ++ ++enum { ++ IFLA_OFFLOAD_XSTATS_UNSPEC = 0, ++ IFLA_OFFLOAD_XSTATS_CPU_HIT = 1, ++ IFLA_OFFLOAD_XSTATS_HW_S_INFO = 2, ++ IFLA_OFFLOAD_XSTATS_L3_STATS = 3, ++ __IFLA_OFFLOAD_XSTATS_MAX = 4, ++}; ++ ++enum { ++ IFLA_OFFLOAD_XSTATS_HW_S_INFO_UNSPEC = 0, ++ IFLA_OFFLOAD_XSTATS_HW_S_INFO_REQUEST = 1, ++ IFLA_OFFLOAD_XSTATS_HW_S_INFO_USED = 2, ++ __IFLA_OFFLOAD_XSTATS_HW_S_INFO_MAX = 3, ++}; ++ ++enum { ++ XDP_ATTACHED_NONE = 0, ++ XDP_ATTACHED_DRV = 1, ++ XDP_ATTACHED_SKB = 2, ++ XDP_ATTACHED_HW = 3, ++ XDP_ATTACHED_MULTI = 4, ++}; ++ ++enum { ++ IFLA_XDP_UNSPEC = 0, ++ IFLA_XDP_FD = 1, ++ IFLA_XDP_ATTACHED = 2, ++ IFLA_XDP_FLAGS = 3, ++ IFLA_XDP_PROG_ID = 4, ++ IFLA_XDP_DRV_PROG_ID = 5, ++ IFLA_XDP_SKB_PROG_ID = 6, ++ IFLA_XDP_HW_PROG_ID = 7, ++ IFLA_XDP_EXPECTED_FD = 8, ++ __IFLA_XDP_MAX = 9, ++}; ++ ++enum { ++ IFLA_EVENT_NONE = 0, ++ IFLA_EVENT_REBOOT = 1, ++ IFLA_EVENT_FEATURES = 2, ++ IFLA_EVENT_BONDING_FAILOVER = 3, ++ IFLA_EVENT_NOTIFY_PEERS = 4, ++ IFLA_EVENT_IGMP_RESEND = 5, ++ IFLA_EVENT_BONDING_OPTIONS = 6, ++}; ++ ++enum netdev_state_t { ++ __LINK_STATE_START = 0, ++ __LINK_STATE_PRESENT = 1, ++ __LINK_STATE_NOCARRIER = 2, ++ __LINK_STATE_LINKWATCH_PENDING = 3, ++ __LINK_STATE_DORMANT = 4, ++ __LINK_STATE_TESTING = 5, ++}; ++ ++enum netdev_priv_flags { ++ IFF_802_1Q_VLAN = 1ULL, ++ IFF_EBRIDGE = 2ULL, ++ IFF_BONDING = 4ULL, ++ IFF_ISATAP = 8ULL, ++ IFF_WAN_HDLC = 16ULL, ++ IFF_XMIT_DST_RELEASE = 32ULL, ++ IFF_DONT_BRIDGE = 64ULL, ++ IFF_DISABLE_NETPOLL = 128ULL, ++ IFF_MACVLAN_PORT = 256ULL, ++ IFF_BRIDGE_PORT = 512ULL, ++ IFF_OVS_DATAPATH = 1024ULL, ++ IFF_TX_SKB_SHARING = 2048ULL, ++ IFF_UNICAST_FLT = 4096ULL, ++ IFF_TEAM_PORT = 8192ULL, ++ IFF_SUPP_NOFCS = 16384ULL, ++ IFF_LIVE_ADDR_CHANGE = 32768ULL, ++ IFF_MACVLAN = 65536ULL, ++ IFF_XMIT_DST_RELEASE_PERM = 131072ULL, ++ IFF_L3MDEV_MASTER = 262144ULL, ++ IFF_NO_QUEUE = 524288ULL, ++ IFF_OPENVSWITCH = 1048576ULL, ++ IFF_L3MDEV_SLAVE = 2097152ULL, ++ IFF_TEAM = 4194304ULL, ++ IFF_RXFH_CONFIGURED = 8388608ULL, ++ IFF_PHONY_HEADROOM = 16777216ULL, ++ IFF_MACSEC = 33554432ULL, ++ IFF_NO_RX_HANDLER = 67108864ULL, ++ IFF_FAILOVER = 134217728ULL, ++ IFF_FAILOVER_SLAVE = 268435456ULL, ++ IFF_L3MDEV_RX_HANDLER = 536870912ULL, ++ IFF_NO_ADDRCONF = 1073741824ULL, ++ IFF_TX_SKB_NO_LINEAR = 2147483648ULL, ++ IFF_CHANGE_PROTO_DOWN = 4294967296ULL, ++}; ++ ++enum devlink_port_type { ++ DEVLINK_PORT_TYPE_NOTSET = 0, ++ DEVLINK_PORT_TYPE_AUTO = 1, ++ DEVLINK_PORT_TYPE_ETH = 2, ++ DEVLINK_PORT_TYPE_IB = 3, ++}; ++ ++enum devlink_port_flavour { ++ DEVLINK_PORT_FLAVOUR_PHYSICAL = 0, ++ DEVLINK_PORT_FLAVOUR_CPU = 1, ++ DEVLINK_PORT_FLAVOUR_DSA = 2, ++ DEVLINK_PORT_FLAVOUR_PCI_PF = 3, ++ DEVLINK_PORT_FLAVOUR_PCI_VF = 4, ++ DEVLINK_PORT_FLAVOUR_VIRTUAL = 5, ++ DEVLINK_PORT_FLAVOUR_UNUSED = 6, ++ DEVLINK_PORT_FLAVOUR_PCI_SF = 7, ++}; ++ ++struct devlink_port_phys_attrs { ++ u32 port_number; ++ u32 split_subport_number; ++}; ++ ++struct devlink_port_pci_pf_attrs { ++ u32 controller; ++ u16 pf; ++ u8 external: 1; ++}; ++ ++struct devlink_port_pci_vf_attrs { ++ u32 controller; ++ u16 pf; ++ u16 vf; ++ u8 external: 1; ++}; ++ ++struct devlink_port_pci_sf_attrs { ++ u32 controller; ++ u32 sf; ++ u16 pf; ++ u8 external: 1; ++}; ++ ++struct devlink_port_attrs { ++ u8 split: 1; ++ u8 splittable: 1; ++ u32 lanes; ++ enum devlink_port_flavour flavour; ++ struct netdev_phys_item_id switch_id; ++ union { ++ struct devlink_port_phys_attrs phys; ++ struct devlink_port_pci_pf_attrs pci_pf; ++ struct devlink_port_pci_vf_attrs pci_vf; ++ struct devlink_port_pci_sf_attrs pci_sf; ++ }; ++}; ++ ++struct devlink; ++ ++struct ib_device; ++ ++struct devlink_rate; ++ ++struct devlink_linecard; ++ ++struct devlink_port { ++ struct list_head list; ++ struct list_head region_list; ++ struct devlink *devlink; ++ unsigned int index; ++ spinlock_t type_lock; ++ enum devlink_port_type type; ++ enum devlink_port_type desired_type; ++ union { ++ struct { ++ struct net_device *netdev; ++ int ifindex; ++ char ifname[16]; ++ } type_eth; ++ struct { ++ struct ib_device *ibdev; ++ } type_ib; ++ }; ++ struct devlink_port_attrs attrs; ++ u8 attrs_set: 1; ++ u8 switch_port: 1; ++ u8 registered: 1; ++ u8 initialized: 1; ++ struct delayed_work type_warn_dw; ++ struct list_head reporter_list; ++ struct devlink_rate *devlink_rate; ++ struct devlink_linecard *linecard; ++}; ++ ++struct netdev_notifier_info { ++ struct net_device *dev; ++ struct netlink_ext_ack *extack; ++}; ++ ++enum netdev_offload_xstats_type { ++ NETDEV_OFFLOAD_XSTATS_TYPE_L3 = 1, ++}; ++ ++enum { ++ IFLA_BRIDGE_FLAGS = 0, ++ IFLA_BRIDGE_MODE = 1, ++ IFLA_BRIDGE_VLAN_INFO = 2, ++ IFLA_BRIDGE_VLAN_TUNNEL_INFO = 3, ++ IFLA_BRIDGE_MRP = 4, ++ IFLA_BRIDGE_CFM = 5, ++ IFLA_BRIDGE_MST = 6, ++ __IFLA_BRIDGE_MAX = 7, ++}; ++ ++struct br_port_msg { ++ __u8 family; ++ __u32 ifindex; ++}; ++ ++struct br_mdb_entry { ++ __u32 ifindex; ++ __u8 state; ++ __u8 flags; ++ __u16 vid; ++ struct { ++ union { ++ __be32 ip4; ++ struct in6_addr ip6; ++ unsigned char mac_addr[6]; ++ } u; ++ __be16 proto; ++ } addr; ++}; ++ ++enum { ++ MDBA_SET_ENTRY_UNSPEC = 0, ++ MDBA_SET_ENTRY = 1, ++ MDBA_SET_ENTRY_ATTRS = 2, ++ __MDBA_SET_ENTRY_MAX = 3, ++}; ++ ++enum { ++ BR_MCAST_DIR_RX = 0, ++ BR_MCAST_DIR_TX = 1, ++ BR_MCAST_DIR_SIZE = 2, ++}; ++ ++enum rtattr_type_t { ++ RTA_UNSPEC = 0, ++ RTA_DST = 1, ++ RTA_SRC = 2, ++ RTA_IIF = 3, ++ RTA_OIF = 4, ++ RTA_GATEWAY = 5, ++ RTA_PRIORITY = 6, ++ RTA_PREFSRC = 7, ++ RTA_METRICS = 8, ++ RTA_MULTIPATH = 9, ++ RTA_PROTOINFO = 10, ++ RTA_FLOW = 11, ++ RTA_CACHEINFO = 12, ++ RTA_SESSION = 13, ++ RTA_MP_ALGO = 14, ++ RTA_TABLE = 15, ++ RTA_MARK = 16, ++ RTA_MFC_STATS = 17, ++ RTA_VIA = 18, ++ RTA_NEWDST = 19, ++ RTA_PREF = 20, ++ RTA_ENCAP_TYPE = 21, ++ RTA_ENCAP = 22, ++ RTA_EXPIRES = 23, ++ RTA_PAD = 24, ++ RTA_UID = 25, ++ RTA_TTL_PROPAGATE = 26, ++ RTA_IP_PROTO = 27, ++ RTA_SPORT = 28, ++ RTA_DPORT = 29, ++ RTA_NH_ID = 30, ++ __RTA_MAX = 31, ++}; ++ ++struct rta_cacheinfo { ++ __u32 rta_clntref; ++ __u32 rta_lastuse; ++ __s32 rta_expires; ++ __u32 rta_error; ++ __u32 rta_used; ++ __u32 rta_id; ++ __u32 rta_ts; ++ __u32 rta_tsage; ++}; ++ ++struct ifinfomsg { ++ unsigned char ifi_family; ++ unsigned char __ifi_pad; ++ short unsigned int ifi_type; ++ int ifi_index; ++ unsigned int ifi_flags; ++ unsigned int ifi_change; ++}; ++ ++enum nla_policy_validation { ++ NLA_VALIDATE_NONE = 0, ++ NLA_VALIDATE_RANGE = 1, ++ NLA_VALIDATE_RANGE_WARN_TOO_LONG = 2, ++ NLA_VALIDATE_MIN = 3, ++ NLA_VALIDATE_MAX = 4, ++ NLA_VALIDATE_MASK = 5, ++ NLA_VALIDATE_RANGE_PTR = 6, ++ NLA_VALIDATE_FUNCTION = 7, ++}; ++ ++enum rtnl_kinds { ++ RTNL_KIND_NEW = 0, ++ RTNL_KIND_DEL = 1, ++ RTNL_KIND_GET = 2, ++ RTNL_KIND_SET = 3, ++}; ++ ++struct rtnl_af_ops { ++ struct list_head list; ++ int family; ++ int (*fill_link_af)(struct sk_buff *, const struct net_device *, u32); ++ size_t (*get_link_af_size)(const struct net_device *, u32); ++ int (*validate_link_af)(const struct net_device *, const struct nlattr *, struct netlink_ext_ack *); ++ int (*set_link_af)(struct net_device *, const struct nlattr *, struct netlink_ext_ack *); ++ int (*fill_stats_af)(struct sk_buff *, const struct net_device *); ++ size_t (*get_stats_af_size)(const struct net_device *); ++}; ++ ++struct udp_hslot { ++ struct hlist_head head; ++ int count; ++ spinlock_t lock; ++}; ++ ++enum devlink_rate_type { ++ DEVLINK_RATE_TYPE_LEAF = 0, ++ DEVLINK_RATE_TYPE_NODE = 1, ++}; ++ ++struct devlink_rate { ++ struct list_head list; ++ enum devlink_rate_type type; ++ struct devlink *devlink; ++ void *priv; ++ u64 tx_share; ++ u64 tx_max; ++ struct devlink_rate *parent; ++ union { ++ struct devlink_port *devlink_port; ++ struct { ++ char *name; ++ refcount_t refcnt; ++ }; ++ }; ++ u32 tx_priority; ++ u32 tx_weight; ++}; ++ ++struct rtnl_link { ++ rtnl_doit_func doit; ++ rtnl_dumpit_func dumpit; ++ struct module *owner; ++ unsigned int flags; ++ struct callback_head rcu; ++}; ++ ++struct rtnl_newlink_tbs { ++ struct nlattr *tb[65]; ++ struct nlattr *attr[51]; ++ struct nlattr *slave_attr[44]; ++}; ++ ++struct rtnl_offload_xstats_request_used { ++ bool request; ++ bool used; ++}; ++ ++struct rtnl_stats_dump_filters { ++ u32 mask[6]; ++}; ++ ++struct rtnl_mdb_dump_ctx { ++ long int idx; ++}; ++ ++struct ipv4_devconf { ++ void *sysctl; ++ int data[33]; ++ long unsigned int state[1]; ++}; ++ ++struct compat_ifconf { ++ compat_int_t ifc_len; ++ compat_caddr_t ifcbuf; ++}; ++ ++struct flow_dissector { ++ unsigned int used_keys; ++ short unsigned int offset[31]; ++}; ++ ++struct in_ifaddr; ++ ++struct ip_mc_list; ++ ++struct in_device { ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ refcount_t refcnt; ++ int dead; ++ struct in_ifaddr *ifa_list; ++ struct ip_mc_list *mc_list; ++ struct ip_mc_list **mc_hash; ++ int mc_count; ++ spinlock_t mc_tomb_lock; ++ struct ip_mc_list *mc_tomb; ++ long unsigned int mr_v1_seen; ++ long unsigned int mr_v2_seen; ++ long unsigned int mr_maxdelay; ++ long unsigned int mr_qi; ++ long unsigned int mr_qri; ++ unsigned char mr_qrv; ++ unsigned char mr_gq_running; ++ u32 mr_ifc_count; ++ struct timer_list mr_gq_timer; ++ struct timer_list mr_ifc_timer; ++ struct neigh_parms *arp_parms; ++ struct ipv4_devconf cnf; ++ struct callback_head callback_head; ++}; ++ ++enum phylink_op_type { ++ PHYLINK_NETDEV = 0, ++ PHYLINK_DEV = 1, ++}; ++ ++struct phylink_link_state; ++ ++struct phylink_config { ++ struct device *dev; ++ enum phylink_op_type type; ++ bool legacy_pre_march2020; ++ bool poll_fixed_state; ++ bool mac_managed_pm; ++ bool ovr_an_inband; ++ void (*get_fixed_state)(struct phylink_config *, struct phylink_link_state *); ++ long unsigned int supported_interfaces[1]; ++ long unsigned int mac_capabilities; ++}; ++ ++struct dsa_device_ops; ++ ++struct dsa_switch_tree; ++ ++struct dsa_switch; ++ ++struct dsa_bridge; ++ ++struct dsa_lag; ++ ++struct dsa_port { ++ union { ++ struct net_device *master; ++ struct net_device *slave; ++ }; ++ const struct dsa_device_ops *tag_ops; ++ struct dsa_switch_tree *dst; ++ struct sk_buff * (*rcv)(struct sk_buff *, struct net_device *); ++ struct dsa_switch *ds; ++ unsigned int index; ++ enum { ++ DSA_PORT_TYPE_UNUSED = 0, ++ DSA_PORT_TYPE_CPU = 1, ++ DSA_PORT_TYPE_DSA = 2, ++ DSA_PORT_TYPE_USER = 3, ++ } type; ++ const char *name; ++ struct dsa_port *cpu_dp; ++ u8 mac[6]; ++ u8 stp_state; ++ u8 vlan_filtering: 1; ++ u8 learning: 1; ++ u8 lag_tx_enabled: 1; ++ u8 master_admin_up: 1; ++ u8 master_oper_up: 1; ++ u8 cpu_port_in_lag: 1; ++ u8 setup: 1; ++ struct device_node *dn; ++ unsigned int ageing_time; ++ struct dsa_bridge *bridge; ++ struct devlink_port devlink_port; ++ struct phylink *pl; ++ struct phylink_config pl_config; ++ struct dsa_lag *lag; ++ struct net_device *hsr_dev; ++ struct list_head list; ++ const struct ethtool_ops *orig_ethtool_ops; ++ struct mutex addr_lists_lock; ++ struct list_head fdbs; ++ struct list_head mdbs; ++ struct mutex vlans_lock; ++ union { ++ struct list_head vlans; ++ struct list_head user_vlans; ++ }; ++}; ++ ++enum netdev_lag_tx_type { ++ NETDEV_LAG_TX_TYPE_UNKNOWN = 0, ++ NETDEV_LAG_TX_TYPE_RANDOM = 1, ++ NETDEV_LAG_TX_TYPE_BROADCAST = 2, ++ NETDEV_LAG_TX_TYPE_ROUNDROBIN = 3, ++ NETDEV_LAG_TX_TYPE_ACTIVEBACKUP = 4, ++ NETDEV_LAG_TX_TYPE_HASH = 5, ++}; ++ ++enum netdev_lag_hash { ++ NETDEV_LAG_HASH_NONE = 0, ++ NETDEV_LAG_HASH_L2 = 1, ++ NETDEV_LAG_HASH_L34 = 2, ++ NETDEV_LAG_HASH_L23 = 3, ++ NETDEV_LAG_HASH_E23 = 4, ++ NETDEV_LAG_HASH_E34 = 5, ++ NETDEV_LAG_HASH_VLAN_SRCMAC = 6, ++ NETDEV_LAG_HASH_UNKNOWN = 7, ++}; ++ ++struct netdev_lag_upper_info { ++ enum netdev_lag_tx_type tx_type; ++ enum netdev_lag_hash hash_type; ++}; ++ ++struct netdev_notifier_changeupper_info { ++ struct netdev_notifier_info info; ++ struct net_device *upper_dev; ++ bool master; ++ bool linking; ++ void *upper_info; ++}; ++ ++struct in_ifaddr { ++ struct hlist_node hash; ++ struct in_ifaddr *ifa_next; ++ struct in_device *ifa_dev; ++ struct callback_head callback_head; ++ __be32 ifa_local; ++ __be32 ifa_address; ++ __be32 ifa_mask; ++ __u32 ifa_rt_priority; ++ __be32 ifa_broadcast; ++ unsigned char ifa_scope; ++ unsigned char ifa_prefixlen; ++ unsigned char ifa_proto; ++ __u32 ifa_flags; ++ char ifa_label[16]; ++ __u32 ifa_valid_lft; ++ __u32 ifa_preferred_lft; ++ long unsigned int ifa_cstamp; ++ long unsigned int ifa_tstamp; ++}; ++ ++struct hwtstamp_config { ++ int flags; ++ int tx_type; ++ int rx_filter; ++}; ++ ++enum hwtstamp_flags { ++ HWTSTAMP_FLAG_BONDED_PHC_INDEX = 1, ++ HWTSTAMP_FLAG_LAST = 1, ++ HWTSTAMP_FLAG_MASK = 1, ++}; ++ ++enum hwtstamp_tx_types { ++ HWTSTAMP_TX_OFF = 0, ++ HWTSTAMP_TX_ON = 1, ++ HWTSTAMP_TX_ONESTEP_SYNC = 2, ++ HWTSTAMP_TX_ONESTEP_P2P = 3, ++ __HWTSTAMP_TX_CNT = 4, ++}; ++ ++enum hwtstamp_rx_filters { ++ HWTSTAMP_FILTER_NONE = 0, ++ HWTSTAMP_FILTER_ALL = 1, ++ HWTSTAMP_FILTER_SOME = 2, ++ HWTSTAMP_FILTER_PTP_V1_L4_EVENT = 3, ++ HWTSTAMP_FILTER_PTP_V1_L4_SYNC = 4, ++ HWTSTAMP_FILTER_PTP_V1_L4_DELAY_REQ = 5, ++ HWTSTAMP_FILTER_PTP_V2_L4_EVENT = 6, ++ HWTSTAMP_FILTER_PTP_V2_L4_SYNC = 7, ++ HWTSTAMP_FILTER_PTP_V2_L4_DELAY_REQ = 8, ++ HWTSTAMP_FILTER_PTP_V2_L2_EVENT = 9, ++ HWTSTAMP_FILTER_PTP_V2_L2_SYNC = 10, ++ HWTSTAMP_FILTER_PTP_V2_L2_DELAY_REQ = 11, ++ HWTSTAMP_FILTER_PTP_V2_EVENT = 12, ++ HWTSTAMP_FILTER_PTP_V2_SYNC = 13, ++ HWTSTAMP_FILTER_PTP_V2_DELAY_REQ = 14, ++ HWTSTAMP_FILTER_NTP_ALL = 15, ++ __HWTSTAMP_FILTER_CNT = 16, ++}; ++ ++struct kernel_hwtstamp_config { ++ int flags; ++ int tx_type; ++ int rx_filter; ++}; ++ ++struct flow_match { ++ struct flow_dissector *dissector; ++ void *mask; ++ void *key; ++}; ++ ++enum flow_action_id { ++ FLOW_ACTION_ACCEPT = 0, ++ FLOW_ACTION_DROP = 1, ++ FLOW_ACTION_TRAP = 2, ++ FLOW_ACTION_GOTO = 3, ++ FLOW_ACTION_REDIRECT = 4, ++ FLOW_ACTION_MIRRED = 5, ++ FLOW_ACTION_REDIRECT_INGRESS = 6, ++ FLOW_ACTION_MIRRED_INGRESS = 7, ++ FLOW_ACTION_VLAN_PUSH = 8, ++ FLOW_ACTION_VLAN_POP = 9, ++ FLOW_ACTION_VLAN_MANGLE = 10, ++ FLOW_ACTION_TUNNEL_ENCAP = 11, ++ FLOW_ACTION_TUNNEL_DECAP = 12, ++ FLOW_ACTION_MANGLE = 13, ++ FLOW_ACTION_ADD = 14, ++ FLOW_ACTION_CSUM = 15, ++ FLOW_ACTION_MARK = 16, ++ FLOW_ACTION_PTYPE = 17, ++ FLOW_ACTION_PRIORITY = 18, ++ FLOW_ACTION_RX_QUEUE_MAPPING = 19, ++ FLOW_ACTION_WAKE = 20, ++ FLOW_ACTION_QUEUE = 21, ++ FLOW_ACTION_SAMPLE = 22, ++ FLOW_ACTION_POLICE = 23, ++ FLOW_ACTION_CT = 24, ++ FLOW_ACTION_CT_METADATA = 25, ++ FLOW_ACTION_MPLS_PUSH = 26, ++ FLOW_ACTION_MPLS_POP = 27, ++ FLOW_ACTION_MPLS_MANGLE = 28, ++ FLOW_ACTION_GATE = 29, ++ FLOW_ACTION_PPPOE_PUSH = 30, ++ FLOW_ACTION_JUMP = 31, ++ FLOW_ACTION_PIPE = 32, ++ FLOW_ACTION_VLAN_PUSH_ETH = 33, ++ FLOW_ACTION_VLAN_POP_ETH = 34, ++ FLOW_ACTION_CONTINUE = 35, ++ NUM_FLOW_ACTIONS = 36, ++}; ++ ++enum flow_action_hw_stats { ++ FLOW_ACTION_HW_STATS_IMMEDIATE = 1, ++ FLOW_ACTION_HW_STATS_DELAYED = 2, ++ FLOW_ACTION_HW_STATS_ANY = 3, ++ FLOW_ACTION_HW_STATS_DISABLED = 4, ++ FLOW_ACTION_HW_STATS_DONT_CARE = 7, ++}; ++ ++typedef void (*action_destr)(void *); ++ ++enum flow_action_mangle_base { ++ FLOW_ACT_MANGLE_UNSPEC = 0, ++ FLOW_ACT_MANGLE_HDR_TYPE_ETH = 1, ++ FLOW_ACT_MANGLE_HDR_TYPE_IP4 = 2, ++ FLOW_ACT_MANGLE_HDR_TYPE_IP6 = 3, ++ FLOW_ACT_MANGLE_HDR_TYPE_TCP = 4, ++ FLOW_ACT_MANGLE_HDR_TYPE_UDP = 5, ++}; ++ ++struct nf_flowtable; ++ ++struct psample_group; ++ ++struct action_gate_entry; ++ ++struct flow_action_cookie; ++ ++struct flow_action_entry { ++ enum flow_action_id id; ++ u32 hw_index; ++ long unsigned int cookie; ++ u64 miss_cookie; ++ enum flow_action_hw_stats hw_stats; ++ action_destr destructor; ++ void *destructor_priv; ++ union { ++ u32 chain_index; ++ struct net_device *dev; ++ struct { ++ u16 vid; ++ __be16 proto; ++ u8 prio; ++ } vlan; ++ struct { ++ unsigned char dst[6]; ++ unsigned char src[6]; ++ } vlan_push_eth; ++ struct { ++ enum flow_action_mangle_base htype; ++ u32 offset; ++ u32 mask; ++ u32 val; ++ } mangle; ++ struct ip_tunnel_info *tunnel; ++ u32 csum_flags; ++ u32 mark; ++ u16 ptype; ++ u16 rx_queue; ++ u32 priority; ++ struct { ++ u32 ctx; ++ u32 index; ++ u8 vf; ++ } queue; ++ struct { ++ struct psample_group *psample_group; ++ u32 rate; ++ u32 trunc_size; ++ bool truncate; ++ } sample; ++ struct { ++ u32 burst; ++ u64 rate_bytes_ps; ++ u64 peakrate_bytes_ps; ++ u32 avrate; ++ u16 overhead; ++ u64 burst_pkt; ++ u64 rate_pkt_ps; ++ u32 mtu; ++ struct { ++ enum flow_action_id act_id; ++ u32 extval; ++ } exceed; ++ struct { ++ enum flow_action_id act_id; ++ u32 extval; ++ } notexceed; ++ } police; ++ struct { ++ int action; ++ u16 zone; ++ struct nf_flowtable *flow_table; ++ } ct; ++ struct { ++ long unsigned int cookie; ++ u32 mark; ++ u32 labels[4]; ++ bool orig_dir; ++ } ct_metadata; ++ struct { ++ u32 label; ++ __be16 proto; ++ u8 tc; ++ u8 bos; ++ u8 ttl; ++ } mpls_push; ++ struct { ++ __be16 proto; ++ } mpls_pop; ++ struct { ++ u32 label; ++ u8 tc; ++ u8 bos; ++ u8 ttl; ++ } mpls_mangle; ++ struct { ++ s32 prio; ++ u64 basetime; ++ u64 cycletime; ++ u64 cycletimeext; ++ u32 num_entries; ++ struct action_gate_entry *entries; ++ } gate; ++ struct { ++ u16 sid; ++ } pppoe; ++ }; ++ struct flow_action_cookie *user_cookie; ++}; ++ ++struct flow_action { ++ unsigned int num_entries; ++ struct flow_action_entry entries[0]; ++}; ++ ++struct flow_rule { ++ struct flow_match match; ++ struct flow_action action; ++}; ++ ++struct dsa_chip_data { ++ struct device *host_dev; ++ int sw_addr; ++ struct device *netdev[12]; ++ int eeprom_len; ++ struct device_node *of_node; ++ char *port_names[12]; ++ struct device_node *port_dn[12]; ++ s8 rtable[4]; ++}; ++ ++struct dsa_platform_data { ++ struct device *netdev; ++ struct net_device *of_netdev; ++ int nr_chips; ++ struct dsa_chip_data *chip; ++}; ++ ++struct phylink_link_state { ++ long unsigned int advertising[2]; ++ long unsigned int lp_advertising[2]; ++ phy_interface_t interface; ++ int speed; ++ int duplex; ++ int pause; ++ int rate_matching; ++ unsigned int link: 1; ++ unsigned int an_complete: 1; ++}; ++ ++struct phylink_pcs_ops; ++ ++struct phylink_pcs { ++ const struct phylink_pcs_ops *ops; ++ bool poll; ++}; ++ ++struct phylink_pcs_ops { ++ int (*pcs_validate)(struct phylink_pcs *, long unsigned int *, const struct phylink_link_state *); ++ void (*pcs_get_state)(struct phylink_pcs *, struct phylink_link_state *); ++ int (*pcs_config)(struct phylink_pcs *, unsigned int, phy_interface_t, const long unsigned int *, bool); ++ void (*pcs_an_restart)(struct phylink_pcs *); ++ void (*pcs_link_up)(struct phylink_pcs *, unsigned int, phy_interface_t, int, int); ++}; ++ ++struct flow_action_cookie { ++ u32 cookie_len; ++ u8 cookie[0]; ++}; ++ ++struct flow_stats { ++ u64 pkts; ++ u64 bytes; ++ u64 drops; ++ u64 lastused; ++ enum flow_action_hw_stats used_hw_stats; ++ bool used_hw_stats_valid; ++}; ++ ++enum flow_cls_command { ++ FLOW_CLS_REPLACE = 0, ++ FLOW_CLS_DESTROY = 1, ++ FLOW_CLS_STATS = 2, ++ FLOW_CLS_TMPLT_CREATE = 3, ++ FLOW_CLS_TMPLT_DESTROY = 4, ++}; ++ ++struct flow_cls_common_offload { ++ u32 chain_index; ++ __be16 protocol; ++ u32 prio; ++ struct netlink_ext_ack *extack; ++}; ++ ++struct flow_cls_offload { ++ struct flow_cls_common_offload common; ++ enum flow_cls_command command; ++ bool use_act_stats; ++ long unsigned int cookie; ++ struct flow_rule *rule; ++ struct flow_stats stats; ++ u32 classid; ++}; ++ ++enum devlink_sb_pool_type { ++ DEVLINK_SB_POOL_TYPE_INGRESS = 0, ++ DEVLINK_SB_POOL_TYPE_EGRESS = 1, ++}; ++ ++enum devlink_sb_threshold_type { ++ DEVLINK_SB_THRESHOLD_TYPE_STATIC = 0, ++ DEVLINK_SB_THRESHOLD_TYPE_DYNAMIC = 1, ++}; ++ ++enum devlink_param_cmode { ++ DEVLINK_PARAM_CMODE_RUNTIME = 0, ++ DEVLINK_PARAM_CMODE_DRIVERINIT = 1, ++ DEVLINK_PARAM_CMODE_PERMANENT = 2, ++ __DEVLINK_PARAM_CMODE_MAX = 3, ++ DEVLINK_PARAM_CMODE_MAX = 2, ++}; ++ ++struct devlink_sb_pool_info { ++ enum devlink_sb_pool_type pool_type; ++ u32 size; ++ enum devlink_sb_threshold_type threshold_type; ++ u32 cell_size; ++}; ++ ++union devlink_param_value { ++ u8 vu8; ++ u16 vu16; ++ u32 vu32; ++ char vstr[32]; ++ bool vbool; ++}; ++ ++struct devlink_param_gset_ctx { ++ union devlink_param_value val; ++ enum devlink_param_cmode cmode; ++}; ++ ++struct switchdev_mst_state { ++ u16 msti; ++ u8 state; ++}; ++ ++struct switchdev_brport_flags { ++ long unsigned int val; ++ long unsigned int mask; ++}; ++ ++struct switchdev_vlan_msti { ++ u16 vid; ++ u16 msti; ++}; ++ ++enum switchdev_obj_id { ++ SWITCHDEV_OBJ_ID_UNDEFINED = 0, ++ SWITCHDEV_OBJ_ID_PORT_VLAN = 1, ++ SWITCHDEV_OBJ_ID_PORT_MDB = 2, ++ SWITCHDEV_OBJ_ID_HOST_MDB = 3, ++ SWITCHDEV_OBJ_ID_MRP = 4, ++ SWITCHDEV_OBJ_ID_RING_TEST_MRP = 5, ++ SWITCHDEV_OBJ_ID_RING_ROLE_MRP = 6, ++ SWITCHDEV_OBJ_ID_RING_STATE_MRP = 7, ++ SWITCHDEV_OBJ_ID_IN_TEST_MRP = 8, ++ SWITCHDEV_OBJ_ID_IN_ROLE_MRP = 9, ++ SWITCHDEV_OBJ_ID_IN_STATE_MRP = 10, ++}; ++ ++struct switchdev_obj { ++ struct list_head list; ++ struct net_device *orig_dev; ++ enum switchdev_obj_id id; ++ u32 flags; ++ void *complete_priv; ++ void (*complete)(struct net_device *, int, void *); ++}; ++ ++struct switchdev_obj_port_vlan { ++ struct switchdev_obj obj; ++ u16 flags; ++ u16 vid; ++ bool changed; ++}; ++ ++struct switchdev_obj_port_mdb { ++ struct switchdev_obj obj; ++ unsigned char addr[6]; ++ u16 vid; ++}; ++ ++struct switchdev_obj_mrp { ++ struct switchdev_obj obj; ++ struct net_device *p_port; ++ struct net_device *s_port; ++ u32 ring_id; ++ u16 prio; ++}; ++ ++struct switchdev_obj_ring_role_mrp { ++ struct switchdev_obj obj; ++ u8 ring_role; ++ u32 ring_id; ++ u8 sw_backup; ++}; ++ ++enum dsa_tag_protocol { ++ DSA_TAG_PROTO_NONE = 0, ++ DSA_TAG_PROTO_BRCM = 1, ++ DSA_TAG_PROTO_BRCM_LEGACY = 22, ++ DSA_TAG_PROTO_BRCM_PREPEND = 2, ++ DSA_TAG_PROTO_DSA = 3, ++ DSA_TAG_PROTO_EDSA = 4, ++ DSA_TAG_PROTO_GSWIP = 5, ++ DSA_TAG_PROTO_KSZ9477 = 6, ++ DSA_TAG_PROTO_KSZ9893 = 7, ++ DSA_TAG_PROTO_LAN9303 = 8, ++ DSA_TAG_PROTO_MTK = 9, ++ DSA_TAG_PROTO_QCA = 10, ++ DSA_TAG_PROTO_TRAILER = 11, ++ DSA_TAG_PROTO_8021Q = 12, ++ DSA_TAG_PROTO_SJA1105 = 13, ++ DSA_TAG_PROTO_KSZ8795 = 14, ++ DSA_TAG_PROTO_OCELOT = 15, ++ DSA_TAG_PROTO_AR9331 = 16, ++ DSA_TAG_PROTO_RTL4_A = 17, ++ DSA_TAG_PROTO_HELLCREEK = 18, ++ DSA_TAG_PROTO_XRS700X = 19, ++ DSA_TAG_PROTO_OCELOT_8021Q = 20, ++ DSA_TAG_PROTO_SEVILLE = 21, ++ DSA_TAG_PROTO_SJA1110 = 23, ++ DSA_TAG_PROTO_RTL8_4 = 24, ++ DSA_TAG_PROTO_RTL8_4T = 25, ++ DSA_TAG_PROTO_RZN1_A5PSW = 26, ++ DSA_TAG_PROTO_LAN937X = 27, ++}; ++ ++struct dsa_device_ops { ++ struct sk_buff * (*xmit)(struct sk_buff *, struct net_device *); ++ struct sk_buff * (*rcv)(struct sk_buff *, struct net_device *); ++ void (*flow_dissect)(const struct sk_buff *, __be16 *, int *); ++ int (*connect)(struct dsa_switch *); ++ void (*disconnect)(struct dsa_switch *); ++ unsigned int needed_headroom; ++ unsigned int needed_tailroom; ++ const char *name; ++ enum dsa_tag_protocol proto; ++ bool promisc_on_master; ++}; ++ ++struct dsa_8021q_context; ++ ++struct dsa_switch_ops; ++ ++struct dsa_switch { ++ struct device *dev; ++ struct dsa_switch_tree *dst; ++ unsigned int index; ++ u32 setup: 1; ++ u32 vlan_filtering_is_global: 1; ++ u32 needs_standalone_vlan_filtering: 1; ++ u32 configure_vlan_while_not_filtering: 1; ++ u32 untag_bridge_pvid: 1; ++ u32 assisted_learning_on_cpu_port: 1; ++ u32 vlan_filtering: 1; ++ u32 mtu_enforcement_ingress: 1; ++ u32 fdb_isolation: 1; ++ struct notifier_block nb; ++ void *priv; ++ void *tagger_data; ++ struct dsa_chip_data *cd; ++ const struct dsa_switch_ops *ops; ++ u32 phys_mii_mask; ++ struct mii_bus *slave_mii_bus; ++ unsigned int ageing_time_min; ++ unsigned int ageing_time_max; ++ struct dsa_8021q_context *tag_8021q_ctx; ++ struct devlink *devlink; ++ unsigned int num_tx_queues; ++ unsigned int num_lag_ids; ++ unsigned int max_num_bridges; ++ unsigned int num_ports; ++}; ++ ++struct dsa_lag { ++ struct net_device *dev; ++ unsigned int id; ++ struct mutex fdb_lock; ++ struct list_head fdbs; ++ refcount_t refcount; ++}; ++ ++struct dsa_switch_tree { ++ struct list_head list; ++ struct list_head ports; ++ struct raw_notifier_head nh; ++ unsigned int index; ++ struct kref refcount; ++ struct dsa_lag **lags; ++ const struct dsa_device_ops *tag_ops; ++ enum dsa_tag_protocol default_proto; ++ bool setup; ++ struct dsa_platform_data *pd; ++ struct list_head rtable; ++ unsigned int lags_len; ++ unsigned int last_switch; ++}; ++ ++struct dsa_mall_mirror_tc_entry { ++ u8 to_local_port; ++ bool ingress; ++}; ++ ++struct dsa_mall_policer_tc_entry { ++ u32 burst; ++ u64 rate_bytes_per_sec; ++}; ++ ++struct dsa_bridge { ++ struct net_device *dev; ++ unsigned int num; ++ bool tx_fwd_offload; ++ refcount_t refcount; ++}; ++ ++enum dsa_db_type { ++ DSA_DB_PORT = 0, ++ DSA_DB_LAG = 1, ++ DSA_DB_BRIDGE = 2, ++}; ++ ++struct dsa_db { ++ enum dsa_db_type type; ++ union { ++ const struct dsa_port *dp; ++ struct dsa_lag lag; ++ struct dsa_bridge bridge; ++ }; ++}; ++ ++typedef int dsa_fdb_dump_cb_t(const unsigned char *, u16, bool, void *); ++ ++struct devlink_info_req; ++ ++struct dsa_switch_ops { ++ enum dsa_tag_protocol (*get_tag_protocol)(struct dsa_switch *, int, enum dsa_tag_protocol); ++ int (*change_tag_protocol)(struct dsa_switch *, enum dsa_tag_protocol); ++ int (*connect_tag_protocol)(struct dsa_switch *, enum dsa_tag_protocol); ++ int (*port_change_master)(struct dsa_switch *, int, struct net_device *, struct netlink_ext_ack *); ++ int (*setup)(struct dsa_switch *); ++ void (*teardown)(struct dsa_switch *); ++ int (*port_setup)(struct dsa_switch *, int); ++ void (*port_teardown)(struct dsa_switch *, int); ++ u32 (*get_phy_flags)(struct dsa_switch *, int); ++ int (*phy_read)(struct dsa_switch *, int, int); ++ int (*phy_write)(struct dsa_switch *, int, int, u16); ++ void (*adjust_link)(struct dsa_switch *, int, struct phy_device *); ++ void (*fixed_link_update)(struct dsa_switch *, int, struct fixed_phy_status *); ++ void (*phylink_get_caps)(struct dsa_switch *, int, struct phylink_config *); ++ struct phylink_pcs * (*phylink_mac_select_pcs)(struct dsa_switch *, int, phy_interface_t); ++ int (*phylink_mac_link_state)(struct dsa_switch *, int, struct phylink_link_state *); ++ void (*phylink_mac_config)(struct dsa_switch *, int, unsigned int, const struct phylink_link_state *); ++ void (*phylink_mac_an_restart)(struct dsa_switch *, int); ++ void (*phylink_mac_link_down)(struct dsa_switch *, int, unsigned int, phy_interface_t); ++ void (*phylink_mac_link_up)(struct dsa_switch *, int, unsigned int, phy_interface_t, struct phy_device *, int, int, bool, bool); ++ void (*phylink_fixed_state)(struct dsa_switch *, int, struct phylink_link_state *); ++ void (*get_strings)(struct dsa_switch *, int, u32, uint8_t *); ++ void (*get_ethtool_stats)(struct dsa_switch *, int, uint64_t *); ++ int (*get_sset_count)(struct dsa_switch *, int, int); ++ void (*get_ethtool_phy_stats)(struct dsa_switch *, int, uint64_t *); ++ void (*get_eth_phy_stats)(struct dsa_switch *, int, struct ethtool_eth_phy_stats *); ++ void (*get_eth_mac_stats)(struct dsa_switch *, int, struct ethtool_eth_mac_stats *); ++ void (*get_eth_ctrl_stats)(struct dsa_switch *, int, struct ethtool_eth_ctrl_stats *); ++ void (*get_rmon_stats)(struct dsa_switch *, int, struct ethtool_rmon_stats *, const struct ethtool_rmon_hist_range **); ++ void (*get_stats64)(struct dsa_switch *, int, struct rtnl_link_stats64 *); ++ void (*get_pause_stats)(struct dsa_switch *, int, struct ethtool_pause_stats *); ++ void (*self_test)(struct dsa_switch *, int, struct ethtool_test *, u64 *); ++ void (*get_wol)(struct dsa_switch *, int, struct ethtool_wolinfo *); ++ int (*set_wol)(struct dsa_switch *, int, struct ethtool_wolinfo *); ++ int (*get_ts_info)(struct dsa_switch *, int, struct ethtool_ts_info *); ++ int (*get_mm)(struct dsa_switch *, int, struct ethtool_mm_state *); ++ int (*set_mm)(struct dsa_switch *, int, struct ethtool_mm_cfg *, struct netlink_ext_ack *); ++ void (*get_mm_stats)(struct dsa_switch *, int, struct ethtool_mm_stats *); ++ int (*port_get_default_prio)(struct dsa_switch *, int); ++ int (*port_set_default_prio)(struct dsa_switch *, int, u8); ++ int (*port_get_dscp_prio)(struct dsa_switch *, int, u8); ++ int (*port_add_dscp_prio)(struct dsa_switch *, int, u8, u8); ++ int (*port_del_dscp_prio)(struct dsa_switch *, int, u8, u8); ++ int (*suspend)(struct dsa_switch *); ++ int (*resume)(struct dsa_switch *); ++ int (*port_enable)(struct dsa_switch *, int, struct phy_device *); ++ void (*port_disable)(struct dsa_switch *, int); ++ struct dsa_port * (*preferred_default_local_cpu_port)(struct dsa_switch *); ++ int (*set_mac_eee)(struct dsa_switch *, int, struct ethtool_eee *); ++ int (*get_mac_eee)(struct dsa_switch *, int, struct ethtool_eee *); ++ int (*get_eeprom_len)(struct dsa_switch *); ++ int (*get_eeprom)(struct dsa_switch *, struct ethtool_eeprom *, u8 *); ++ int (*set_eeprom)(struct dsa_switch *, struct ethtool_eeprom *, u8 *); ++ int (*get_regs_len)(struct dsa_switch *, int); ++ void (*get_regs)(struct dsa_switch *, int, struct ethtool_regs *, void *); ++ int (*port_prechangeupper)(struct dsa_switch *, int, struct netdev_notifier_changeupper_info *); ++ int (*set_ageing_time)(struct dsa_switch *, unsigned int); ++ int (*port_bridge_join)(struct dsa_switch *, int, struct dsa_bridge, bool *, struct netlink_ext_ack *); ++ void (*port_bridge_leave)(struct dsa_switch *, int, struct dsa_bridge); ++ void (*port_stp_state_set)(struct dsa_switch *, int, u8); ++ int (*port_mst_state_set)(struct dsa_switch *, int, const struct switchdev_mst_state *); ++ void (*port_fast_age)(struct dsa_switch *, int); ++ int (*port_vlan_fast_age)(struct dsa_switch *, int, u16); ++ int (*port_pre_bridge_flags)(struct dsa_switch *, int, struct switchdev_brport_flags, struct netlink_ext_ack *); ++ int (*port_bridge_flags)(struct dsa_switch *, int, struct switchdev_brport_flags, struct netlink_ext_ack *); ++ void (*port_set_host_flood)(struct dsa_switch *, int, bool, bool); ++ int (*port_vlan_filtering)(struct dsa_switch *, int, bool, struct netlink_ext_ack *); ++ int (*port_vlan_add)(struct dsa_switch *, int, const struct switchdev_obj_port_vlan *, struct netlink_ext_ack *); ++ int (*port_vlan_del)(struct dsa_switch *, int, const struct switchdev_obj_port_vlan *); ++ int (*vlan_msti_set)(struct dsa_switch *, struct dsa_bridge, const struct switchdev_vlan_msti *); ++ int (*port_fdb_add)(struct dsa_switch *, int, const unsigned char *, u16, struct dsa_db); ++ int (*port_fdb_del)(struct dsa_switch *, int, const unsigned char *, u16, struct dsa_db); ++ int (*port_fdb_dump)(struct dsa_switch *, int, dsa_fdb_dump_cb_t *, void *); ++ int (*lag_fdb_add)(struct dsa_switch *, struct dsa_lag, const unsigned char *, u16, struct dsa_db); ++ int (*lag_fdb_del)(struct dsa_switch *, struct dsa_lag, const unsigned char *, u16, struct dsa_db); ++ int (*port_mdb_add)(struct dsa_switch *, int, const struct switchdev_obj_port_mdb *, struct dsa_db); ++ int (*port_mdb_del)(struct dsa_switch *, int, const struct switchdev_obj_port_mdb *, struct dsa_db); ++ int (*get_rxnfc)(struct dsa_switch *, int, struct ethtool_rxnfc *, u32 *); ++ int (*set_rxnfc)(struct dsa_switch *, int, struct ethtool_rxnfc *); ++ int (*cls_flower_add)(struct dsa_switch *, int, struct flow_cls_offload *, bool); ++ int (*cls_flower_del)(struct dsa_switch *, int, struct flow_cls_offload *, bool); ++ int (*cls_flower_stats)(struct dsa_switch *, int, struct flow_cls_offload *, bool); ++ int (*port_mirror_add)(struct dsa_switch *, int, struct dsa_mall_mirror_tc_entry *, bool, struct netlink_ext_ack *); ++ void (*port_mirror_del)(struct dsa_switch *, int, struct dsa_mall_mirror_tc_entry *); ++ int (*port_policer_add)(struct dsa_switch *, int, struct dsa_mall_policer_tc_entry *); ++ void (*port_policer_del)(struct dsa_switch *, int); ++ int (*port_setup_tc)(struct dsa_switch *, int, enum tc_setup_type, void *); ++ int (*crosschip_bridge_join)(struct dsa_switch *, int, int, int, struct dsa_bridge, struct netlink_ext_ack *); ++ void (*crosschip_bridge_leave)(struct dsa_switch *, int, int, int, struct dsa_bridge); ++ int (*crosschip_lag_change)(struct dsa_switch *, int, int); ++ int (*crosschip_lag_join)(struct dsa_switch *, int, int, struct dsa_lag, struct netdev_lag_upper_info *, struct netlink_ext_ack *); ++ int (*crosschip_lag_leave)(struct dsa_switch *, int, int, struct dsa_lag); ++ int (*port_hwtstamp_get)(struct dsa_switch *, int, struct ifreq *); ++ int (*port_hwtstamp_set)(struct dsa_switch *, int, struct ifreq *); ++ void (*port_txtstamp)(struct dsa_switch *, int, struct sk_buff *); ++ bool (*port_rxtstamp)(struct dsa_switch *, int, struct sk_buff *, unsigned int); ++ int (*devlink_param_get)(struct dsa_switch *, u32, struct devlink_param_gset_ctx *); ++ int (*devlink_param_set)(struct dsa_switch *, u32, struct devlink_param_gset_ctx *); ++ int (*devlink_info_get)(struct dsa_switch *, struct devlink_info_req *, struct netlink_ext_ack *); ++ int (*devlink_sb_pool_get)(struct dsa_switch *, unsigned int, u16, struct devlink_sb_pool_info *); ++ int (*devlink_sb_pool_set)(struct dsa_switch *, unsigned int, u16, u32, enum devlink_sb_threshold_type, struct netlink_ext_ack *); ++ int (*devlink_sb_port_pool_get)(struct dsa_switch *, int, unsigned int, u16, u32 *); ++ int (*devlink_sb_port_pool_set)(struct dsa_switch *, int, unsigned int, u16, u32, struct netlink_ext_ack *); ++ int (*devlink_sb_tc_pool_bind_get)(struct dsa_switch *, int, unsigned int, u16, enum devlink_sb_pool_type, u16 *, u32 *); ++ int (*devlink_sb_tc_pool_bind_set)(struct dsa_switch *, int, unsigned int, u16, enum devlink_sb_pool_type, u16, u32, struct netlink_ext_ack *); ++ int (*devlink_sb_occ_snapshot)(struct dsa_switch *, unsigned int); ++ int (*devlink_sb_occ_max_clear)(struct dsa_switch *, unsigned int); ++ int (*devlink_sb_occ_port_pool_get)(struct dsa_switch *, int, unsigned int, u16, u32 *, u32 *); ++ int (*devlink_sb_occ_tc_port_bind_get)(struct dsa_switch *, int, unsigned int, u16, enum devlink_sb_pool_type, u32 *, u32 *); ++ int (*port_change_mtu)(struct dsa_switch *, int, int); ++ int (*port_max_mtu)(struct dsa_switch *, int); ++ int (*port_lag_change)(struct dsa_switch *, int); ++ int (*port_lag_join)(struct dsa_switch *, int, struct dsa_lag, struct netdev_lag_upper_info *, struct netlink_ext_ack *); ++ int (*port_lag_leave)(struct dsa_switch *, int, struct dsa_lag); ++ int (*port_hsr_join)(struct dsa_switch *, int, struct net_device *); ++ int (*port_hsr_leave)(struct dsa_switch *, int, struct net_device *); ++ int (*port_mrp_add)(struct dsa_switch *, int, const struct switchdev_obj_mrp *); ++ int (*port_mrp_del)(struct dsa_switch *, int, const struct switchdev_obj_mrp *); ++ int (*port_mrp_add_ring_role)(struct dsa_switch *, int, const struct switchdev_obj_ring_role_mrp *); ++ int (*port_mrp_del_ring_role)(struct dsa_switch *, int, const struct switchdev_obj_ring_role_mrp *); ++ int (*tag_8021q_vlan_add)(struct dsa_switch *, int, u16, u16); ++ int (*tag_8021q_vlan_del)(struct dsa_switch *, int, u16); ++ void (*master_state_change)(struct dsa_switch *, const struct net_device *, bool); ++}; ++ ++struct dsa_stubs { ++ int (*master_hwtstamp_validate)(struct net_device *, const struct kernel_hwtstamp_config *, struct netlink_ext_ack *); ++}; ++ ++struct rhashtable_walker { ++ struct list_head list; ++ struct bucket_table *tbl; ++}; ++ ++struct rhashtable_iter { ++ struct rhashtable *ht; ++ struct rhash_head *p; ++ struct rhlist_head *list; ++ struct rhashtable_walker walker; ++ unsigned int slot; ++ unsigned int skip; ++ bool end_of_table; ++}; ++ ++enum netdev_xdp_act { ++ NETDEV_XDP_ACT_BASIC = 1, ++ NETDEV_XDP_ACT_REDIRECT = 2, ++ NETDEV_XDP_ACT_NDO_XMIT = 4, ++ NETDEV_XDP_ACT_XSK_ZEROCOPY = 8, ++ NETDEV_XDP_ACT_HW_OFFLOAD = 16, ++ NETDEV_XDP_ACT_RX_SG = 32, ++ NETDEV_XDP_ACT_NDO_XMIT_SG = 64, ++ NETDEV_XDP_ACT_MASK = 127, ++}; ++ ++enum xdp_mem_type { ++ MEM_TYPE_PAGE_SHARED = 0, ++ MEM_TYPE_PAGE_ORDER0 = 1, ++ MEM_TYPE_PAGE_POOL = 2, ++ MEM_TYPE_XSK_BUFF_POOL = 3, ++ MEM_TYPE_MAX = 4, ++}; ++ ++enum xdp_buff_flags { ++ XDP_FLAGS_HAS_FRAGS = 1, ++ XDP_FLAGS_FRAGS_PF_MEMALLOC = 2, ++}; ++ ++struct xdp_frame_bulk { ++ int count; ++ void *xa; ++ void *q[16]; ++}; ++ ++struct xdp_attachment_info { ++ struct bpf_prog *prog; ++ u32 flags; ++}; ++ ++struct xdp_umem; ++ ++struct xsk_queue; ++ ++struct xdp_buff_xsk; ++ ++struct xdp_desc; ++ ++struct xsk_buff_pool { ++ struct device *dev; ++ struct net_device *netdev; ++ struct list_head xsk_tx_list; ++ spinlock_t xsk_tx_list_lock; ++ refcount_t users; ++ struct xdp_umem *umem; ++ struct work_struct work; ++ struct list_head free_list; ++ u32 heads_cnt; ++ u16 queue_id; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct xsk_queue *fq; ++ struct xsk_queue *cq; ++ dma_addr_t *dma_pages; ++ struct xdp_buff_xsk *heads; ++ struct xdp_desc *tx_descs; ++ u64 chunk_mask; ++ u64 addrs_cnt; ++ u32 free_list_cnt; ++ u32 dma_pages_cnt; ++ u32 free_heads_cnt; ++ u32 headroom; ++ u32 chunk_size; ++ u32 chunk_shift; ++ u32 frame_len; ++ u8 cached_need_wakeup; ++ bool uses_need_wakeup; ++ bool dma_need_sync; ++ bool unaligned; ++ void *addrs; ++ spinlock_t cq_lock; ++ struct xdp_buff_xsk *free_heads[0]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_nh_params { ++ u32 nh_family; ++ union { ++ u32 ipv4_nh; ++ struct in6_addr ipv6_nh; ++ }; ++}; ++ ++struct bpf_redirect_info { ++ u64 tgt_index; ++ void *tgt_value; ++ struct bpf_map *map; ++ u32 flags; ++ u32 kern_flags; ++ u32 map_id; ++ enum bpf_map_type map_type; ++ struct bpf_nh_params nh; ++}; ++ ++struct xdp_mem_allocator { ++ struct xdp_mem_info mem; ++ union { ++ void *allocator; ++ struct page_pool *page_pool; ++ }; ++ struct rhash_head node; ++ struct callback_head rcu; ++}; ++ ++struct xdp_desc { ++ __u64 addr; ++ __u32 len; ++ __u32 options; ++}; ++ ++struct xdp_umem { ++ void *addrs; ++ u64 size; ++ u32 headroom; ++ u32 chunk_size; ++ u32 chunks; ++ u32 npgs; ++ struct user_struct *user; ++ refcount_t users; ++ u8 flags; ++ bool zc; ++ struct page **pgs; ++ int id; ++ struct list_head xsk_dma_list; ++ struct work_struct work; ++}; ++ ++struct xdp_buff_xsk { ++ struct xdp_buff xdp; ++ u8 cb[24]; ++ dma_addr_t dma; ++ dma_addr_t frame_dma; ++ struct xsk_buff_pool *pool; ++ u64 orig_addr; ++ struct list_head free_list_node; ++}; ++ ++enum { ++ NETDEV_A_DEV_IFINDEX = 1, ++ NETDEV_A_DEV_PAD = 2, ++ NETDEV_A_DEV_XDP_FEATURES = 3, ++ __NETDEV_A_DEV_MAX = 4, ++ NETDEV_A_DEV_MAX = 3, ++}; ++ ++enum { ++ NETDEV_CMD_DEV_GET = 1, ++ NETDEV_CMD_DEV_ADD_NTF = 2, ++ NETDEV_CMD_DEV_DEL_NTF = 3, ++ NETDEV_CMD_DEV_CHANGE_NTF = 4, ++ __NETDEV_CMD_MAX = 5, ++ NETDEV_CMD_MAX = 4, ++}; ++ ++enum { ++ NETDEV_NLGRP_MGMT = 0, ++}; ++ ++struct page_pool_stats { ++ struct page_pool_alloc_stats alloc_stats; ++ struct page_pool_recycle_stats recycle_stats; ++}; ++ ++enum rt_class_t { ++ RT_TABLE_UNSPEC = 0, ++ RT_TABLE_COMPAT = 252, ++ RT_TABLE_DEFAULT = 253, ++ RT_TABLE_MAIN = 254, ++ RT_TABLE_LOCAL = 255, ++ RT_TABLE_MAX = 4294967295, ++}; ++ ++struct fib_rule_uid_range { ++ __u32 start; ++ __u32 end; ++}; ++ ++enum { ++ FRA_UNSPEC = 0, ++ FRA_DST = 1, ++ FRA_SRC = 2, ++ FRA_IIFNAME = 3, ++ FRA_GOTO = 4, ++ FRA_UNUSED2 = 5, ++ FRA_PRIORITY = 6, ++ FRA_UNUSED3 = 7, ++ FRA_UNUSED4 = 8, ++ FRA_UNUSED5 = 9, ++ FRA_FWMARK = 10, ++ FRA_FLOW = 11, ++ FRA_TUN_ID = 12, ++ FRA_SUPPRESS_IFGROUP = 13, ++ FRA_SUPPRESS_PREFIXLEN = 14, ++ FRA_TABLE = 15, ++ FRA_FWMASK = 16, ++ FRA_OIFNAME = 17, ++ FRA_PAD = 18, ++ FRA_L3MDEV = 19, ++ FRA_UID_RANGE = 20, ++ FRA_PROTOCOL = 21, ++ FRA_IP_PROTO = 22, ++ FRA_SPORT_RANGE = 23, ++ FRA_DPORT_RANGE = 24, ++ __FRA_MAX = 25, ++}; ++ ++enum { ++ FR_ACT_UNSPEC = 0, ++ FR_ACT_TO_TBL = 1, ++ FR_ACT_GOTO = 2, ++ FR_ACT_NOP = 3, ++ FR_ACT_RES3 = 4, ++ FR_ACT_RES4 = 5, ++ FR_ACT_BLACKHOLE = 6, ++ FR_ACT_UNREACHABLE = 7, ++ FR_ACT_PROHIBIT = 8, ++ __FR_ACT_MAX = 9, ++}; ++ ++struct fib_notifier_info { ++ int family; ++ struct netlink_ext_ack *extack; ++}; ++ ++enum fib_event_type { ++ FIB_EVENT_ENTRY_REPLACE = 0, ++ FIB_EVENT_ENTRY_APPEND = 1, ++ FIB_EVENT_ENTRY_ADD = 2, ++ FIB_EVENT_ENTRY_DEL = 3, ++ FIB_EVENT_RULE_ADD = 4, ++ FIB_EVENT_RULE_DEL = 5, ++ FIB_EVENT_NH_ADD = 6, ++ FIB_EVENT_NH_DEL = 7, ++ FIB_EVENT_VIF_ADD = 8, ++ FIB_EVENT_VIF_DEL = 9, ++}; ++ ++struct fib_rule_notifier_info { ++ struct fib_notifier_info info; ++ struct fib_rule *rule; ++}; ++ ++enum ethtool_test_flags { ++ ETH_TEST_FL_OFFLINE = 1, ++ ETH_TEST_FL_FAILED = 2, ++ ETH_TEST_FL_EXTERNAL_LB = 4, ++ ETH_TEST_FL_EXTERNAL_LB_DONE = 8, ++}; ++ ++struct packet_type { ++ __be16 type; ++ bool ignore_outgoing; ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ int (*func)(struct sk_buff *, struct net_device *, struct packet_type *, struct net_device *); ++ void (*list_func)(struct list_head *, struct packet_type *, struct net_device *); ++ bool (*id_match)(struct packet_type *, struct sock *); ++ struct net *af_packet_net; ++ void *af_packet_priv; ++ struct list_head list; ++}; ++ ++struct net_packet_attrs { ++ const unsigned char *src; ++ const unsigned char *dst; ++ u32 ip_src; ++ u32 ip_dst; ++ bool tcp; ++ u16 sport; ++ u16 dport; ++ int timeout; ++ int size; ++ int max_size; ++ u8 id; ++ u16 queue_mapping; ++}; ++ ++struct net_test_priv { ++ struct net_packet_attrs *packet; ++ struct packet_type pt; ++ struct completion comp; ++ int double_vlan; ++ int vlan_id; ++ int ok; ++}; ++ ++struct netsfhdr { ++ __be32 version; ++ __be64 magic; ++ u8 id; ++} __attribute__((packed)); ++ ++struct net_test { ++ char name[32]; ++ int (*fn)(struct net_device *); ++}; ++ ++enum tca_id { ++ TCA_ID_UNSPEC = 0, ++ TCA_ID_POLICE = 1, ++ TCA_ID_GACT = 5, ++ TCA_ID_IPT = 6, ++ TCA_ID_PEDIT = 7, ++ TCA_ID_MIRRED = 8, ++ TCA_ID_NAT = 9, ++ TCA_ID_XT = 10, ++ TCA_ID_SKBEDIT = 11, ++ TCA_ID_VLAN = 12, ++ TCA_ID_BPF = 13, ++ TCA_ID_CONNMARK = 14, ++ TCA_ID_SKBMOD = 15, ++ TCA_ID_CSUM = 16, ++ TCA_ID_TUNNEL_KEY = 17, ++ TCA_ID_SIMP = 22, ++ TCA_ID_IFE = 25, ++ TCA_ID_SAMPLE = 26, ++ TCA_ID_CTINFO = 27, ++ TCA_ID_MPLS = 28, ++ TCA_ID_CT = 29, ++ TCA_ID_GATE = 30, ++ __TCA_ID_MAX = 255, ++}; ++ ++struct tcf_t { ++ __u64 install; ++ __u64 lastuse; ++ __u64 expires; ++ __u64 firstuse; ++}; ++ ++struct tcf_walker { ++ int stop; ++ int skip; ++ int count; ++ bool nonempty; ++ long unsigned int cookie; ++ int (*fn)(struct tcf_proto *, void *, struct tcf_walker *); ++}; ++ ++struct tc_action; ++ ++struct tcf_exts_miss_cookie_node; ++ ++struct tcf_exts { ++ __u32 type; ++ int nr_actions; ++ struct tc_action **actions; ++ struct net *net; ++ netns_tracker ns_tracker; ++ struct tcf_exts_miss_cookie_node *miss_cookie_node; ++ int action; ++ int police; ++}; ++ ++struct tcf_idrinfo { ++ struct mutex lock; ++ struct idr action_idr; ++ struct net *net; ++}; ++ ++struct tc_action_ops; ++ ++struct tc_cookie; ++ ++struct tc_action { ++ const struct tc_action_ops *ops; ++ __u32 type; ++ struct tcf_idrinfo *idrinfo; ++ u32 tcfa_index; ++ refcount_t tcfa_refcnt; ++ atomic_t tcfa_bindcnt; ++ int tcfa_action; ++ struct tcf_t tcfa_tm; ++ long: 64; ++ struct gnet_stats_basic_sync tcfa_bstats; ++ struct gnet_stats_basic_sync tcfa_bstats_hw; ++ struct gnet_stats_queue tcfa_qstats; ++ struct net_rate_estimator *tcfa_rate_est; ++ spinlock_t tcfa_lock; ++ struct gnet_stats_basic_sync *cpu_bstats; ++ struct gnet_stats_basic_sync *cpu_bstats_hw; ++ struct gnet_stats_queue *cpu_qstats; ++ struct tc_cookie *user_cookie; ++ struct tcf_chain *goto_chain; ++ u32 tcfa_flags; ++ u8 hw_stats; ++ u8 used_hw_stats; ++ bool used_hw_stats_valid; ++ u32 in_hw_count; ++}; ++ ++typedef void (*tc_action_priv_destructor)(void *); ++ ++struct tc_action_ops { ++ struct list_head head; ++ char kind[16]; ++ enum tca_id id; ++ unsigned int net_id; ++ size_t size; ++ struct module *owner; ++ int (*act)(struct sk_buff *, const struct tc_action *, struct tcf_result *); ++ int (*dump)(struct sk_buff *, struct tc_action *, int, int); ++ void (*cleanup)(struct tc_action *); ++ int (*lookup)(struct net *, struct tc_action **, u32); ++ int (*init)(struct net *, struct nlattr *, struct nlattr *, struct tc_action **, struct tcf_proto *, u32, struct netlink_ext_ack *); ++ int (*walk)(struct net *, struct sk_buff *, struct netlink_callback *, int, const struct tc_action_ops *, struct netlink_ext_ack *); ++ void (*stats_update)(struct tc_action *, u64, u64, u64, u64, bool); ++ size_t (*get_fill_size)(const struct tc_action *); ++ struct net_device * (*get_dev)(const struct tc_action *, tc_action_priv_destructor *); ++ struct psample_group * (*get_psample_group)(const struct tc_action *, tc_action_priv_destructor *); ++ int (*offload_act_setup)(struct tc_action *, void *, u32 *, bool, struct netlink_ext_ack *); ++}; ++ ++struct tc_cookie { ++ u8 *data; ++ u32 len; ++ struct callback_head rcu; ++}; ++ ++enum bpf_ret_code { ++ BPF_OK = 0, ++ BPF_DROP = 2, ++ BPF_REDIRECT = 7, ++ BPF_LWT_REROUTE = 128, ++ BPF_FLOW_DISSECTOR_CONTINUE = 129, ++}; ++ ++struct nl_info { ++ struct nlmsghdr *nlh; ++ struct net *nl_net; ++ u32 portid; ++ u8 skip_notify: 1; ++ u8 skip_notify_kernel: 1; ++}; ++ ++struct qdisc_skb_cb { ++ struct { ++ unsigned int pkt_len; ++ u16 slave_dev_queue_mapping; ++ u16 tc_classid; ++ }; ++ unsigned char data[20]; ++}; ++ ++struct bpf_skb_data_end { ++ struct qdisc_skb_cb qdisc_cb; ++ void *data_meta; ++ void *data_end; ++}; ++ ++enum lwtunnel_encap_types { ++ LWTUNNEL_ENCAP_NONE = 0, ++ LWTUNNEL_ENCAP_MPLS = 1, ++ LWTUNNEL_ENCAP_IP = 2, ++ LWTUNNEL_ENCAP_ILA = 3, ++ LWTUNNEL_ENCAP_IP6 = 4, ++ LWTUNNEL_ENCAP_SEG6 = 5, ++ LWTUNNEL_ENCAP_BPF = 6, ++ LWTUNNEL_ENCAP_SEG6_LOCAL = 7, ++ LWTUNNEL_ENCAP_RPL = 8, ++ LWTUNNEL_ENCAP_IOAM6 = 9, ++ LWTUNNEL_ENCAP_XFRM = 10, ++ __LWTUNNEL_ENCAP_MAX = 11, ++}; ++ ++enum { ++ LWT_BPF_PROG_UNSPEC = 0, ++ LWT_BPF_PROG_FD = 1, ++ LWT_BPF_PROG_NAME = 2, ++ __LWT_BPF_PROG_MAX = 3, ++}; ++ ++enum { ++ LWT_BPF_UNSPEC = 0, ++ LWT_BPF_IN = 1, ++ LWT_BPF_OUT = 2, ++ LWT_BPF_XMIT = 3, ++ LWT_BPF_XMIT_HEADROOM = 4, ++ __LWT_BPF_MAX = 5, ++}; ++ ++struct fib6_result; ++ ++struct fib6_config; ++ ++struct ipv6_stub { ++ int (*ipv6_sock_mc_join)(struct sock *, int, const struct in6_addr *); ++ int (*ipv6_sock_mc_drop)(struct sock *, int, const struct in6_addr *); ++ struct dst_entry * (*ipv6_dst_lookup_flow)(struct net *, const struct sock *, struct flowi6 *, const struct in6_addr *); ++ int (*ipv6_route_input)(struct sk_buff *); ++ struct fib6_table * (*fib6_get_table)(struct net *, u32); ++ int (*fib6_lookup)(struct net *, int, struct flowi6 *, struct fib6_result *, int); ++ int (*fib6_table_lookup)(struct net *, struct fib6_table *, int, struct flowi6 *, struct fib6_result *, int); ++ void (*fib6_select_path)(const struct net *, struct fib6_result *, struct flowi6 *, int, bool, const struct sk_buff *, int); ++ u32 (*ip6_mtu_from_fib6)(const struct fib6_result *, const struct in6_addr *, const struct in6_addr *); ++ int (*fib6_nh_init)(struct net *, struct fib6_nh *, struct fib6_config *, gfp_t, struct netlink_ext_ack *); ++ void (*fib6_nh_release)(struct fib6_nh *); ++ void (*fib6_nh_release_dsts)(struct fib6_nh *); ++ void (*fib6_update_sernum)(struct net *, struct fib6_info *); ++ int (*ip6_del_rt)(struct net *, struct fib6_info *, bool); ++ void (*fib6_rt_update)(struct net *, struct fib6_info *, struct nl_info *); ++ void (*udpv6_encap_enable)(); ++ void (*ndisc_send_na)(struct net_device *, const struct in6_addr *, const struct in6_addr *, bool, bool, bool, bool); ++ void (*xfrm6_local_rxpmtu)(struct sk_buff *, u32); ++ int (*xfrm6_udp_encap_rcv)(struct sock *, struct sk_buff *); ++ int (*xfrm6_rcv_encap)(struct sk_buff *, int, __be32, int); ++ struct neigh_table *nd_tbl; ++ int (*ipv6_fragment)(struct net *, struct sock *, struct sk_buff *, int (*)(struct net *, struct sock *, struct sk_buff *)); ++ struct net_device * (*ipv6_dev_find)(struct net *, const struct in6_addr *, struct net_device *); ++}; ++ ++struct fib6_result { ++ struct fib6_nh *nh; ++ struct fib6_info *f6i; ++ u32 fib6_flags; ++ u8 fib6_type; ++ struct rt6_info *rt6; ++}; ++ ++struct fib6_config { ++ u32 fc_table; ++ u32 fc_metric; ++ int fc_dst_len; ++ int fc_src_len; ++ int fc_ifindex; ++ u32 fc_flags; ++ u32 fc_protocol; ++ u16 fc_type; ++ u16 fc_delete_all_nh: 1; ++ u16 fc_ignore_dev_down: 1; ++ u16 __unused: 14; ++ u32 fc_nh_id; ++ struct in6_addr fc_dst; ++ struct in6_addr fc_src; ++ struct in6_addr fc_prefsrc; ++ struct in6_addr fc_gateway; ++ long unsigned int fc_expires; ++ struct nlattr *fc_mx; ++ int fc_mx_len; ++ int fc_mp_len; ++ struct nlattr *fc_mp; ++ struct nl_info fc_nlinfo; ++ struct nlattr *fc_encap; ++ u16 fc_encap_type; ++ bool fc_is_fdb; ++}; ++ ++enum { ++ LWTUNNEL_XMIT_DONE = 0, ++ LWTUNNEL_XMIT_CONTINUE = 1, ++}; ++ ++struct lwtunnel_encap_ops { ++ int (*build_state)(struct net *, struct nlattr *, unsigned int, const void *, struct lwtunnel_state **, struct netlink_ext_ack *); ++ void (*destroy_state)(struct lwtunnel_state *); ++ int (*output)(struct net *, struct sock *, struct sk_buff *); ++ int (*input)(struct sk_buff *); ++ int (*fill_encap)(struct sk_buff *, struct lwtunnel_state *); ++ int (*get_encap_size)(struct lwtunnel_state *); ++ int (*cmp_encap)(struct lwtunnel_state *, struct lwtunnel_state *); ++ int (*xmit)(struct sk_buff *); ++ struct module *owner; ++}; ++ ++struct gre_base_hdr { ++ __be16 flags; ++ __be16 protocol; ++}; ++ ++struct bpf_lwt_prog { ++ struct bpf_prog *prog; ++ char *name; ++}; ++ ++struct bpf_lwt { ++ struct bpf_lwt_prog in; ++ struct bpf_lwt_prog out; ++ struct bpf_lwt_prog xmit; ++ int family; ++}; ++ ++enum { ++ BPF_F_INGRESS = 1, ++}; ++ ++enum { ++ BPF_SOCK_OPS_VOID = 0, ++ BPF_SOCK_OPS_TIMEOUT_INIT = 1, ++ BPF_SOCK_OPS_RWND_INIT = 2, ++ BPF_SOCK_OPS_TCP_CONNECT_CB = 3, ++ BPF_SOCK_OPS_ACTIVE_ESTABLISHED_CB = 4, ++ BPF_SOCK_OPS_PASSIVE_ESTABLISHED_CB = 5, ++ BPF_SOCK_OPS_NEEDS_ECN = 6, ++ BPF_SOCK_OPS_BASE_RTT = 7, ++ BPF_SOCK_OPS_RTO_CB = 8, ++ BPF_SOCK_OPS_RETRANS_CB = 9, ++ BPF_SOCK_OPS_STATE_CB = 10, ++ BPF_SOCK_OPS_TCP_LISTEN_CB = 11, ++ BPF_SOCK_OPS_RTT_CB = 12, ++ BPF_SOCK_OPS_PARSE_HDR_OPT_CB = 13, ++ BPF_SOCK_OPS_HDR_OPT_LEN_CB = 14, ++ BPF_SOCK_OPS_WRITE_HDR_OPT_CB = 15, ++}; ++ ++struct sk_psock_link { ++ struct list_head list; ++ struct bpf_map *map; ++ void *link_raw; ++}; ++ ++struct bpf_stab { ++ struct bpf_map map; ++ struct sock **sks; ++ struct sk_psock_progs progs; ++ raw_spinlock_t lock; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++typedef u64 (*btf_bpf_sock_map_update)(struct bpf_sock_ops_kern *, struct bpf_map *, void *, u64); ++ ++typedef u64 (*btf_bpf_sk_redirect_map)(struct sk_buff *, struct bpf_map *, u32, u64); ++ ++typedef u64 (*btf_bpf_msg_redirect_map)(struct sk_msg *, struct bpf_map *, u32, u64); ++ ++struct sock_map_seq_info { ++ struct bpf_map *map; ++ struct sock *sk; ++ u32 index; ++}; ++ ++struct bpf_iter__sockmap { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct bpf_map *map; ++ }; ++ union { ++ void *key; ++ }; ++ union { ++ struct sock *sk; ++ }; ++}; ++ ++struct bpf_shtab_elem { ++ struct callback_head rcu; ++ u32 hash; ++ struct sock *sk; ++ struct hlist_node node; ++ u8 key[0]; ++}; ++ ++struct bpf_shtab_bucket { ++ struct hlist_head head; ++ raw_spinlock_t lock; ++}; ++ ++struct bpf_shtab { ++ struct bpf_map map; ++ struct bpf_shtab_bucket *buckets; ++ u32 buckets_num; ++ u32 elem_size; ++ struct sk_psock_progs progs; ++ atomic_t count; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++typedef u64 (*btf_bpf_sock_hash_update)(struct bpf_sock_ops_kern *, struct bpf_map *, void *, u64); ++ ++typedef u64 (*btf_bpf_sk_redirect_hash)(struct sk_buff *, struct bpf_map *, void *, u64); ++ ++typedef u64 (*btf_bpf_msg_redirect_hash)(struct sk_msg *, struct bpf_map *, void *, u64); ++ ++struct sock_hash_seq_info { ++ struct bpf_map *map; ++ struct bpf_shtab *htab; ++ u32 bucket_id; ++}; ++ ++struct fch_hdr { ++ __u8 daddr[6]; ++ __u8 saddr[6]; ++}; ++ ++struct fcllc { ++ __u8 dsap; ++ __u8 ssap; ++ __u8 llc; ++ __u8 protid[3]; ++ __be16 ethertype; ++}; ++ ++struct tc_ratespec { ++ unsigned char cell_log; ++ __u8 linklayer; ++ short unsigned int overhead; ++ short int cell_align; ++ short unsigned int mpu; ++ __u32 rate; ++}; ++ ++struct tc_prio_qopt { ++ int bands; ++ __u8 priomap[16]; ++}; ++ ++enum netdev_queue_state_t { ++ __QUEUE_STATE_DRV_XOFF = 0, ++ __QUEUE_STATE_STACK_XOFF = 1, ++ __QUEUE_STATE_FROZEN = 2, ++}; ++ ++struct skb_array { ++ struct ptr_ring ring; ++}; ++ ++enum qdisc_state_t { ++ __QDISC_STATE_SCHED = 0, ++ __QDISC_STATE_DEACTIVATED = 1, ++ __QDISC_STATE_MISSED = 2, ++ __QDISC_STATE_DRAINING = 3, ++}; ++ ++enum qdisc_state2_t { ++ __QDISC_STATE2_RUNNING = 0, ++}; ++ ++struct psched_ratecfg { ++ u64 rate_bytes_ps; ++ u32 mult; ++ u16 overhead; ++ u16 mpu; ++ u8 linklayer; ++ u8 shift; ++}; ++ ++struct psched_pktrate { ++ u64 rate_pkts_ps; ++ u32 mult; ++ u8 shift; ++}; ++ ++struct mini_Qdisc_pair { ++ struct mini_Qdisc miniq1; ++ struct mini_Qdisc miniq2; ++ struct mini_Qdisc **p_miniq; ++}; ++ ++struct pfifo_fast_priv { ++ struct skb_array q[3]; ++}; ++ ++struct tc_skb_ext { ++ union { ++ u64 act_miss_cookie; ++ __u32 chain; ++ }; ++ __u16 mru; ++ __u16 zone; ++ u8 post_ct: 1; ++ u8 post_ct_snat: 1; ++ u8 post_ct_dnat: 1; ++ u8 act_miss: 1; ++}; ++ ++struct nla_bitfield32 { ++ __u32 value; ++ __u32 selector; ++}; ++ ++enum { ++ TCA_ACT_UNSPEC = 0, ++ TCA_ACT_KIND = 1, ++ TCA_ACT_OPTIONS = 2, ++ TCA_ACT_INDEX = 3, ++ TCA_ACT_STATS = 4, ++ TCA_ACT_PAD = 5, ++ TCA_ACT_COOKIE = 6, ++ TCA_ACT_FLAGS = 7, ++ TCA_ACT_HW_STATS = 8, ++ TCA_ACT_USED_HW_STATS = 9, ++ TCA_ACT_IN_HW_COUNT = 10, ++ __TCA_ACT_MAX = 11, ++}; ++ ++struct psample_group { ++ struct list_head list; ++ struct net *net; ++ u32 group_num; ++ u32 refcount; ++ u32 seq; ++ struct callback_head rcu; ++}; ++ ++struct action_gate_entry { ++ u8 gate_state; ++ u32 interval; ++ s32 ipv; ++ s32 maxoctets; ++}; ++ ++enum flow_block_command { ++ FLOW_BLOCK_BIND = 0, ++ FLOW_BLOCK_UNBIND = 1, ++}; ++ ++enum flow_block_binder_type { ++ FLOW_BLOCK_BINDER_TYPE_UNSPEC = 0, ++ FLOW_BLOCK_BINDER_TYPE_CLSACT_INGRESS = 1, ++ FLOW_BLOCK_BINDER_TYPE_CLSACT_EGRESS = 2, ++ FLOW_BLOCK_BINDER_TYPE_RED_EARLY_DROP = 3, ++ FLOW_BLOCK_BINDER_TYPE_RED_MARK = 4, ++}; ++ ++struct flow_block_offload { ++ enum flow_block_command command; ++ enum flow_block_binder_type binder_type; ++ bool block_shared; ++ bool unlocked_driver_cb; ++ struct net *net; ++ struct flow_block *block; ++ struct list_head cb_list; ++ struct list_head *driver_block_list; ++ struct netlink_ext_ack *extack; ++ struct Qdisc *sch; ++ struct list_head *cb_list_head; ++}; ++ ++struct flow_block_cb; ++ ++struct flow_block_indr { ++ struct list_head list; ++ struct net_device *dev; ++ struct Qdisc *sch; ++ enum flow_block_binder_type binder_type; ++ void *data; ++ void *cb_priv; ++ void (*cleanup)(struct flow_block_cb *); ++}; ++ ++struct flow_block_cb { ++ struct list_head driver_list; ++ struct list_head list; ++ flow_setup_cb_t *cb; ++ void *cb_ident; ++ void *cb_priv; ++ void (*release)(void *); ++ struct flow_block_indr indr; ++ unsigned int refcnt; ++}; ++ ++enum qdisc_class_ops_flags { ++ QDISC_CLASS_OPS_DOIT_UNLOCKED = 1, ++}; ++ ++enum tcf_proto_ops_flags { ++ TCF_PROTO_OPS_DOIT_UNLOCKED = 1, ++}; ++ ++typedef void tcf_chain_head_change_t(struct tcf_proto *, void *); ++ ++enum net_xmit_qdisc_t { ++ __NET_XMIT_STOLEN = 65536, ++ __NET_XMIT_BYPASS = 131072, ++}; ++ ++struct tc_skb_cb { ++ struct qdisc_skb_cb qdisc_cb; ++ u16 mru; ++ u8 post_ct: 1; ++ u8 post_ct_snat: 1; ++ u8 post_ct_dnat: 1; ++ u16 zone; ++}; ++ ++struct tcf_block_ext_info { ++ enum flow_block_binder_type binder_type; ++ tcf_chain_head_change_t *chain_head_change; ++ void *chain_head_change_priv; ++ u32 block_index; ++}; ++ ++struct tcf_qevent { ++ struct tcf_block *block; ++ struct tcf_block_ext_info info; ++ struct tcf_proto *filter_chain; ++}; ++ ++struct tcf_exts_miss_cookie_node { ++ const struct tcf_chain *chain; ++ const struct tcf_proto *tp; ++ const struct tcf_exts *exts; ++ u32 chain_index; ++ u32 tp_prio; ++ u32 handle; ++ u32 miss_cookie_base; ++ struct callback_head rcu; ++}; ++ ++enum pedit_header_type { ++ TCA_PEDIT_KEY_EX_HDR_TYPE_NETWORK = 0, ++ TCA_PEDIT_KEY_EX_HDR_TYPE_ETH = 1, ++ TCA_PEDIT_KEY_EX_HDR_TYPE_IP4 = 2, ++ TCA_PEDIT_KEY_EX_HDR_TYPE_IP6 = 3, ++ TCA_PEDIT_KEY_EX_HDR_TYPE_TCP = 4, ++ TCA_PEDIT_KEY_EX_HDR_TYPE_UDP = 5, ++ __PEDIT_HDR_TYPE_MAX = 6, ++}; ++ ++enum pedit_cmd { ++ TCA_PEDIT_KEY_EX_CMD_SET = 0, ++ TCA_PEDIT_KEY_EX_CMD_ADD = 1, ++ __PEDIT_CMD_MAX = 2, ++}; ++ ++struct tc_pedit_key { ++ __u32 mask; ++ __u32 val; ++ __u32 off; ++ __u32 at; ++ __u32 offmask; ++ __u32 shift; ++}; ++ ++struct tcf_pedit_key_ex { ++ enum pedit_header_type htype; ++ enum pedit_cmd cmd; ++}; ++ ++struct tcf_pedit_parms { ++ struct tc_pedit_key *tcfp_keys; ++ struct tcf_pedit_key_ex *tcfp_keys_ex; ++ u32 tcfp_off_max_hint; ++ unsigned char tcfp_nkeys; ++ unsigned char tcfp_flags; ++ struct callback_head rcu; ++}; ++ ++struct tcf_pedit { ++ struct tc_action common; ++ struct tcf_pedit_parms *parms; ++ long: 64; ++}; ++ ++enum nf_ct_ext_id { ++ NF_CT_EXT_HELPER = 0, ++ NF_CT_EXT_NAT = 1, ++ NF_CT_EXT_SEQADJ = 2, ++ NF_CT_EXT_ACCT = 3, ++ NF_CT_EXT_ECACHE = 4, ++ NF_CT_EXT_TSTAMP = 5, ++ NF_CT_EXT_LABELS = 6, ++ NF_CT_EXT_SYNPROXY = 7, ++ NF_CT_EXT_ACT_CT = 8, ++ NF_CT_EXT_NUM = 9, ++}; ++ ++union tcf_exts_miss_cookie { ++ struct { ++ u32 miss_cookie_base; ++ u32 act_index; ++ }; ++ u64 miss_cookie; ++}; ++ ++struct tcf_filter_chain_list_item { ++ struct list_head list; ++ tcf_chain_head_change_t *chain_head_change; ++ void *chain_head_change_priv; ++}; ++ ++struct tcf_net { ++ spinlock_t idr_lock; ++ struct idr idr; ++}; ++ ++struct tcf_block_owner_item { ++ struct list_head list; ++ struct Qdisc *q; ++ enum flow_block_binder_type binder_type; ++}; ++ ++struct tcf_chain_info { ++ struct tcf_proto **pprev; ++ struct tcf_proto *next; ++}; ++ ++struct tcf_dump_args { ++ struct tcf_walker w; ++ struct sk_buff *skb; ++ struct netlink_callback *cb; ++ struct tcf_block *block; ++ struct Qdisc *q; ++ u32 parent; ++ bool terse_dump; ++}; ++ ++struct tcf_ematch_tree_hdr { ++ __u16 nmatches; ++ __u16 progid; ++}; ++ ++enum { ++ TCA_EMATCH_TREE_UNSPEC = 0, ++ TCA_EMATCH_TREE_HDR = 1, ++ TCA_EMATCH_TREE_LIST = 2, ++ __TCA_EMATCH_TREE_MAX = 3, ++}; ++ ++struct tcf_ematch_hdr { ++ __u16 matchid; ++ __u16 kind; ++ __u16 flags; ++ __u16 pad; ++}; ++ ++struct tcf_pkt_info { ++ unsigned char *ptr; ++ int nexthdr; ++}; ++ ++struct tcf_ematch_ops; ++ ++struct tcf_ematch { ++ struct tcf_ematch_ops *ops; ++ long unsigned int data; ++ unsigned int datalen; ++ u16 matchid; ++ u16 flags; ++ struct net *net; ++}; ++ ++struct tcf_ematch_ops { ++ int kind; ++ int datalen; ++ int (*change)(struct net *, void *, int, struct tcf_ematch *); ++ int (*match)(struct sk_buff *, struct tcf_ematch *, struct tcf_pkt_info *); ++ void (*destroy)(struct tcf_ematch *); ++ int (*dump)(struct sk_buff *, struct tcf_ematch *); ++ struct module *owner; ++ struct list_head link; ++}; ++ ++struct tcf_ematch_tree { ++ struct tcf_ematch_tree_hdr hdr; ++ struct tcf_ematch *matches; ++}; ++ ++enum { ++ CTRL_CMD_UNSPEC = 0, ++ CTRL_CMD_NEWFAMILY = 1, ++ CTRL_CMD_DELFAMILY = 2, ++ CTRL_CMD_GETFAMILY = 3, ++ CTRL_CMD_NEWOPS = 4, ++ CTRL_CMD_DELOPS = 5, ++ CTRL_CMD_GETOPS = 6, ++ CTRL_CMD_NEWMCAST_GRP = 7, ++ CTRL_CMD_DELMCAST_GRP = 8, ++ CTRL_CMD_GETMCAST_GRP = 9, ++ CTRL_CMD_GETPOLICY = 10, ++ __CTRL_CMD_MAX = 11, ++}; ++ ++enum { ++ CTRL_ATTR_UNSPEC = 0, ++ CTRL_ATTR_FAMILY_ID = 1, ++ CTRL_ATTR_FAMILY_NAME = 2, ++ CTRL_ATTR_VERSION = 3, ++ CTRL_ATTR_HDRSIZE = 4, ++ CTRL_ATTR_MAXATTR = 5, ++ CTRL_ATTR_OPS = 6, ++ CTRL_ATTR_MCAST_GROUPS = 7, ++ CTRL_ATTR_POLICY = 8, ++ CTRL_ATTR_OP_POLICY = 9, ++ CTRL_ATTR_OP = 10, ++ __CTRL_ATTR_MAX = 11, ++}; ++ ++enum { ++ CTRL_ATTR_OP_UNSPEC = 0, ++ CTRL_ATTR_OP_ID = 1, ++ CTRL_ATTR_OP_FLAGS = 2, ++ __CTRL_ATTR_OP_MAX = 3, ++}; ++ ++enum { ++ CTRL_ATTR_MCAST_GRP_UNSPEC = 0, ++ CTRL_ATTR_MCAST_GRP_NAME = 1, ++ CTRL_ATTR_MCAST_GRP_ID = 2, ++ __CTRL_ATTR_MCAST_GRP_MAX = 3, ++}; ++ ++enum { ++ CTRL_ATTR_POLICY_UNSPEC = 0, ++ CTRL_ATTR_POLICY_DO = 1, ++ CTRL_ATTR_POLICY_DUMP = 2, ++ __CTRL_ATTR_POLICY_DUMP_MAX = 3, ++ CTRL_ATTR_POLICY_DUMP_MAX = 2, ++}; ++ ++struct genl_dumpit_info { ++ const struct genl_family *family; ++ struct genl_split_ops op; ++ struct nlattr **attrs; ++}; ++ ++struct genl_op_iter { ++ const struct genl_family *family; ++ struct genl_split_ops doit; ++ struct genl_split_ops dumpit; ++ int cmd_idx; ++ int entry_idx; ++ u32 cmd; ++ u8 flags; ++}; ++ ++struct genl_start_context { ++ const struct genl_family *family; ++ struct nlmsghdr *nlh; ++ struct netlink_ext_ack *extack; ++ const struct genl_split_ops *ops; ++ int hdrlen; ++}; ++ ++struct netlink_policy_dump_state; ++ ++struct ctrl_dump_policy_ctx { ++ struct netlink_policy_dump_state *state; ++ const struct genl_family *rt; ++ struct genl_op_iter *op_iter; ++ u32 op; ++ u16 fam_id; ++ u8 dump_map: 1; ++ u8 single_op: 1; ++}; ++ ++struct bpf_tramp_link; ++ ++struct bpf_tramp_links { ++ struct bpf_tramp_link *links[27]; ++ int nr_links; ++}; ++ ++struct bpf_tramp_link { ++ struct bpf_link link; ++ struct hlist_node tramp_hlist; ++ u64 cookie; ++}; ++ ++struct bpf_dummy_ops_state { ++ int val; ++}; ++ ++typedef int (*dummy_ops_test_ret_fn)(struct bpf_dummy_ops_state *, ...); ++ ++struct bpf_dummy_ops_test_args { ++ u64 args[12]; ++ struct bpf_dummy_ops_state state; ++}; ++ ++enum { ++ ETHTOOL_MSG_USER_NONE = 0, ++ ETHTOOL_MSG_STRSET_GET = 1, ++ ETHTOOL_MSG_LINKINFO_GET = 2, ++ ETHTOOL_MSG_LINKINFO_SET = 3, ++ ETHTOOL_MSG_LINKMODES_GET = 4, ++ ETHTOOL_MSG_LINKMODES_SET = 5, ++ ETHTOOL_MSG_LINKSTATE_GET = 6, ++ ETHTOOL_MSG_DEBUG_GET = 7, ++ ETHTOOL_MSG_DEBUG_SET = 8, ++ ETHTOOL_MSG_WOL_GET = 9, ++ ETHTOOL_MSG_WOL_SET = 10, ++ ETHTOOL_MSG_FEATURES_GET = 11, ++ ETHTOOL_MSG_FEATURES_SET = 12, ++ ETHTOOL_MSG_PRIVFLAGS_GET = 13, ++ ETHTOOL_MSG_PRIVFLAGS_SET = 14, ++ ETHTOOL_MSG_RINGS_GET = 15, ++ ETHTOOL_MSG_RINGS_SET = 16, ++ ETHTOOL_MSG_CHANNELS_GET = 17, ++ ETHTOOL_MSG_CHANNELS_SET = 18, ++ ETHTOOL_MSG_COALESCE_GET = 19, ++ ETHTOOL_MSG_COALESCE_SET = 20, ++ ETHTOOL_MSG_PAUSE_GET = 21, ++ ETHTOOL_MSG_PAUSE_SET = 22, ++ ETHTOOL_MSG_EEE_GET = 23, ++ ETHTOOL_MSG_EEE_SET = 24, ++ ETHTOOL_MSG_TSINFO_GET = 25, ++ ETHTOOL_MSG_CABLE_TEST_ACT = 26, ++ ETHTOOL_MSG_CABLE_TEST_TDR_ACT = 27, ++ ETHTOOL_MSG_TUNNEL_INFO_GET = 28, ++ ETHTOOL_MSG_FEC_GET = 29, ++ ETHTOOL_MSG_FEC_SET = 30, ++ ETHTOOL_MSG_MODULE_EEPROM_GET = 31, ++ ETHTOOL_MSG_STATS_GET = 32, ++ ETHTOOL_MSG_PHC_VCLOCKS_GET = 33, ++ ETHTOOL_MSG_MODULE_GET = 34, ++ ETHTOOL_MSG_MODULE_SET = 35, ++ ETHTOOL_MSG_PSE_GET = 36, ++ ETHTOOL_MSG_PSE_SET = 37, ++ ETHTOOL_MSG_RSS_GET = 38, ++ ETHTOOL_MSG_PLCA_GET_CFG = 39, ++ ETHTOOL_MSG_PLCA_SET_CFG = 40, ++ ETHTOOL_MSG_PLCA_GET_STATUS = 41, ++ ETHTOOL_MSG_MM_GET = 42, ++ ETHTOOL_MSG_MM_SET = 43, ++ __ETHTOOL_MSG_USER_CNT = 44, ++ ETHTOOL_MSG_USER_MAX = 43, ++}; ++ ++enum { ++ ETHTOOL_MSG_KERNEL_NONE = 0, ++ ETHTOOL_MSG_STRSET_GET_REPLY = 1, ++ ETHTOOL_MSG_LINKINFO_GET_REPLY = 2, ++ ETHTOOL_MSG_LINKINFO_NTF = 3, ++ ETHTOOL_MSG_LINKMODES_GET_REPLY = 4, ++ ETHTOOL_MSG_LINKMODES_NTF = 5, ++ ETHTOOL_MSG_LINKSTATE_GET_REPLY = 6, ++ ETHTOOL_MSG_DEBUG_GET_REPLY = 7, ++ ETHTOOL_MSG_DEBUG_NTF = 8, ++ ETHTOOL_MSG_WOL_GET_REPLY = 9, ++ ETHTOOL_MSG_WOL_NTF = 10, ++ ETHTOOL_MSG_FEATURES_GET_REPLY = 11, ++ ETHTOOL_MSG_FEATURES_SET_REPLY = 12, ++ ETHTOOL_MSG_FEATURES_NTF = 13, ++ ETHTOOL_MSG_PRIVFLAGS_GET_REPLY = 14, ++ ETHTOOL_MSG_PRIVFLAGS_NTF = 15, ++ ETHTOOL_MSG_RINGS_GET_REPLY = 16, ++ ETHTOOL_MSG_RINGS_NTF = 17, ++ ETHTOOL_MSG_CHANNELS_GET_REPLY = 18, ++ ETHTOOL_MSG_CHANNELS_NTF = 19, ++ ETHTOOL_MSG_COALESCE_GET_REPLY = 20, ++ ETHTOOL_MSG_COALESCE_NTF = 21, ++ ETHTOOL_MSG_PAUSE_GET_REPLY = 22, ++ ETHTOOL_MSG_PAUSE_NTF = 23, ++ ETHTOOL_MSG_EEE_GET_REPLY = 24, ++ ETHTOOL_MSG_EEE_NTF = 25, ++ ETHTOOL_MSG_TSINFO_GET_REPLY = 26, ++ ETHTOOL_MSG_CABLE_TEST_NTF = 27, ++ ETHTOOL_MSG_CABLE_TEST_TDR_NTF = 28, ++ ETHTOOL_MSG_TUNNEL_INFO_GET_REPLY = 29, ++ ETHTOOL_MSG_FEC_GET_REPLY = 30, ++ ETHTOOL_MSG_FEC_NTF = 31, ++ ETHTOOL_MSG_MODULE_EEPROM_GET_REPLY = 32, ++ ETHTOOL_MSG_STATS_GET_REPLY = 33, ++ ETHTOOL_MSG_PHC_VCLOCKS_GET_REPLY = 34, ++ ETHTOOL_MSG_MODULE_GET_REPLY = 35, ++ ETHTOOL_MSG_MODULE_NTF = 36, ++ ETHTOOL_MSG_PSE_GET_REPLY = 37, ++ ETHTOOL_MSG_RSS_GET_REPLY = 38, ++ ETHTOOL_MSG_PLCA_GET_CFG_REPLY = 39, ++ ETHTOOL_MSG_PLCA_GET_STATUS_REPLY = 40, ++ ETHTOOL_MSG_PLCA_NTF = 41, ++ ETHTOOL_MSG_MM_GET_REPLY = 42, ++ ETHTOOL_MSG_MM_NTF = 43, ++ __ETHTOOL_MSG_KERNEL_CNT = 44, ++ ETHTOOL_MSG_KERNEL_MAX = 43, ++}; ++ ++enum { ++ ETHTOOL_A_HEADER_UNSPEC = 0, ++ ETHTOOL_A_HEADER_DEV_INDEX = 1, ++ ETHTOOL_A_HEADER_DEV_NAME = 2, ++ ETHTOOL_A_HEADER_FLAGS = 3, ++ __ETHTOOL_A_HEADER_CNT = 4, ++ ETHTOOL_A_HEADER_MAX = 3, ++}; ++ ++enum { ++ ETHTOOL_A_STRSET_UNSPEC = 0, ++ ETHTOOL_A_STRSET_HEADER = 1, ++ ETHTOOL_A_STRSET_STRINGSETS = 2, ++ ETHTOOL_A_STRSET_COUNTS_ONLY = 3, ++ __ETHTOOL_A_STRSET_CNT = 4, ++ ETHTOOL_A_STRSET_MAX = 3, ++}; ++ ++enum { ++ ETHTOOL_A_LINKINFO_UNSPEC = 0, ++ ETHTOOL_A_LINKINFO_HEADER = 1, ++ ETHTOOL_A_LINKINFO_PORT = 2, ++ ETHTOOL_A_LINKINFO_PHYADDR = 3, ++ ETHTOOL_A_LINKINFO_TP_MDIX = 4, ++ ETHTOOL_A_LINKINFO_TP_MDIX_CTRL = 5, ++ ETHTOOL_A_LINKINFO_TRANSCEIVER = 6, ++ __ETHTOOL_A_LINKINFO_CNT = 7, ++ ETHTOOL_A_LINKINFO_MAX = 6, ++}; ++ ++enum { ++ ETHTOOL_A_LINKMODES_UNSPEC = 0, ++ ETHTOOL_A_LINKMODES_HEADER = 1, ++ ETHTOOL_A_LINKMODES_AUTONEG = 2, ++ ETHTOOL_A_LINKMODES_OURS = 3, ++ ETHTOOL_A_LINKMODES_PEER = 4, ++ ETHTOOL_A_LINKMODES_SPEED = 5, ++ ETHTOOL_A_LINKMODES_DUPLEX = 6, ++ ETHTOOL_A_LINKMODES_MASTER_SLAVE_CFG = 7, ++ ETHTOOL_A_LINKMODES_MASTER_SLAVE_STATE = 8, ++ ETHTOOL_A_LINKMODES_LANES = 9, ++ ETHTOOL_A_LINKMODES_RATE_MATCHING = 10, ++ __ETHTOOL_A_LINKMODES_CNT = 11, ++ ETHTOOL_A_LINKMODES_MAX = 10, ++}; ++ ++enum { ++ ETHTOOL_A_LINKSTATE_UNSPEC = 0, ++ ETHTOOL_A_LINKSTATE_HEADER = 1, ++ ETHTOOL_A_LINKSTATE_LINK = 2, ++ ETHTOOL_A_LINKSTATE_SQI = 3, ++ ETHTOOL_A_LINKSTATE_SQI_MAX = 4, ++ ETHTOOL_A_LINKSTATE_EXT_STATE = 5, ++ ETHTOOL_A_LINKSTATE_EXT_SUBSTATE = 6, ++ ETHTOOL_A_LINKSTATE_EXT_DOWN_CNT = 7, ++ __ETHTOOL_A_LINKSTATE_CNT = 8, ++ ETHTOOL_A_LINKSTATE_MAX = 7, ++}; ++ ++enum { ++ ETHTOOL_A_DEBUG_UNSPEC = 0, ++ ETHTOOL_A_DEBUG_HEADER = 1, ++ ETHTOOL_A_DEBUG_MSGMASK = 2, ++ __ETHTOOL_A_DEBUG_CNT = 3, ++ ETHTOOL_A_DEBUG_MAX = 2, ++}; ++ ++enum { ++ ETHTOOL_A_WOL_UNSPEC = 0, ++ ETHTOOL_A_WOL_HEADER = 1, ++ ETHTOOL_A_WOL_MODES = 2, ++ ETHTOOL_A_WOL_SOPASS = 3, ++ __ETHTOOL_A_WOL_CNT = 4, ++ ETHTOOL_A_WOL_MAX = 3, ++}; ++ ++enum { ++ ETHTOOL_A_FEATURES_UNSPEC = 0, ++ ETHTOOL_A_FEATURES_HEADER = 1, ++ ETHTOOL_A_FEATURES_HW = 2, ++ ETHTOOL_A_FEATURES_WANTED = 3, ++ ETHTOOL_A_FEATURES_ACTIVE = 4, ++ ETHTOOL_A_FEATURES_NOCHANGE = 5, ++ __ETHTOOL_A_FEATURES_CNT = 6, ++ ETHTOOL_A_FEATURES_MAX = 5, ++}; ++ ++enum { ++ ETHTOOL_A_PRIVFLAGS_UNSPEC = 0, ++ ETHTOOL_A_PRIVFLAGS_HEADER = 1, ++ ETHTOOL_A_PRIVFLAGS_FLAGS = 2, ++ __ETHTOOL_A_PRIVFLAGS_CNT = 3, ++ ETHTOOL_A_PRIVFLAGS_MAX = 2, ++}; ++ ++enum { ++ ETHTOOL_A_RINGS_UNSPEC = 0, ++ ETHTOOL_A_RINGS_HEADER = 1, ++ ETHTOOL_A_RINGS_RX_MAX = 2, ++ ETHTOOL_A_RINGS_RX_MINI_MAX = 3, ++ ETHTOOL_A_RINGS_RX_JUMBO_MAX = 4, ++ ETHTOOL_A_RINGS_TX_MAX = 5, ++ ETHTOOL_A_RINGS_RX = 6, ++ ETHTOOL_A_RINGS_RX_MINI = 7, ++ ETHTOOL_A_RINGS_RX_JUMBO = 8, ++ ETHTOOL_A_RINGS_TX = 9, ++ ETHTOOL_A_RINGS_RX_BUF_LEN = 10, ++ ETHTOOL_A_RINGS_TCP_DATA_SPLIT = 11, ++ ETHTOOL_A_RINGS_CQE_SIZE = 12, ++ ETHTOOL_A_RINGS_TX_PUSH = 13, ++ ETHTOOL_A_RINGS_RX_PUSH = 14, ++ ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN = 15, ++ ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN_MAX = 16, ++ __ETHTOOL_A_RINGS_CNT = 17, ++ ETHTOOL_A_RINGS_MAX = 16, ++}; ++ ++enum { ++ ETHTOOL_A_CHANNELS_UNSPEC = 0, ++ ETHTOOL_A_CHANNELS_HEADER = 1, ++ ETHTOOL_A_CHANNELS_RX_MAX = 2, ++ ETHTOOL_A_CHANNELS_TX_MAX = 3, ++ ETHTOOL_A_CHANNELS_OTHER_MAX = 4, ++ ETHTOOL_A_CHANNELS_COMBINED_MAX = 5, ++ ETHTOOL_A_CHANNELS_RX_COUNT = 6, ++ ETHTOOL_A_CHANNELS_TX_COUNT = 7, ++ ETHTOOL_A_CHANNELS_OTHER_COUNT = 8, ++ ETHTOOL_A_CHANNELS_COMBINED_COUNT = 9, ++ __ETHTOOL_A_CHANNELS_CNT = 10, ++ ETHTOOL_A_CHANNELS_MAX = 9, ++}; ++ ++enum { ++ ETHTOOL_A_COALESCE_UNSPEC = 0, ++ ETHTOOL_A_COALESCE_HEADER = 1, ++ ETHTOOL_A_COALESCE_RX_USECS = 2, ++ ETHTOOL_A_COALESCE_RX_MAX_FRAMES = 3, ++ ETHTOOL_A_COALESCE_RX_USECS_IRQ = 4, ++ ETHTOOL_A_COALESCE_RX_MAX_FRAMES_IRQ = 5, ++ ETHTOOL_A_COALESCE_TX_USECS = 6, ++ ETHTOOL_A_COALESCE_TX_MAX_FRAMES = 7, ++ ETHTOOL_A_COALESCE_TX_USECS_IRQ = 8, ++ ETHTOOL_A_COALESCE_TX_MAX_FRAMES_IRQ = 9, ++ ETHTOOL_A_COALESCE_STATS_BLOCK_USECS = 10, ++ ETHTOOL_A_COALESCE_USE_ADAPTIVE_RX = 11, ++ ETHTOOL_A_COALESCE_USE_ADAPTIVE_TX = 12, ++ ETHTOOL_A_COALESCE_PKT_RATE_LOW = 13, ++ ETHTOOL_A_COALESCE_RX_USECS_LOW = 14, ++ ETHTOOL_A_COALESCE_RX_MAX_FRAMES_LOW = 15, ++ ETHTOOL_A_COALESCE_TX_USECS_LOW = 16, ++ ETHTOOL_A_COALESCE_TX_MAX_FRAMES_LOW = 17, ++ ETHTOOL_A_COALESCE_PKT_RATE_HIGH = 18, ++ ETHTOOL_A_COALESCE_RX_USECS_HIGH = 19, ++ ETHTOOL_A_COALESCE_RX_MAX_FRAMES_HIGH = 20, ++ ETHTOOL_A_COALESCE_TX_USECS_HIGH = 21, ++ ETHTOOL_A_COALESCE_TX_MAX_FRAMES_HIGH = 22, ++ ETHTOOL_A_COALESCE_RATE_SAMPLE_INTERVAL = 23, ++ ETHTOOL_A_COALESCE_USE_CQE_MODE_TX = 24, ++ ETHTOOL_A_COALESCE_USE_CQE_MODE_RX = 25, ++ ETHTOOL_A_COALESCE_TX_AGGR_MAX_BYTES = 26, ++ ETHTOOL_A_COALESCE_TX_AGGR_MAX_FRAMES = 27, ++ ETHTOOL_A_COALESCE_TX_AGGR_TIME_USECS = 28, ++ __ETHTOOL_A_COALESCE_CNT = 29, ++ ETHTOOL_A_COALESCE_MAX = 28, ++}; ++ ++enum { ++ ETHTOOL_A_PAUSE_UNSPEC = 0, ++ ETHTOOL_A_PAUSE_HEADER = 1, ++ ETHTOOL_A_PAUSE_AUTONEG = 2, ++ ETHTOOL_A_PAUSE_RX = 3, ++ ETHTOOL_A_PAUSE_TX = 4, ++ ETHTOOL_A_PAUSE_STATS = 5, ++ ETHTOOL_A_PAUSE_STATS_SRC = 6, ++ __ETHTOOL_A_PAUSE_CNT = 7, ++ ETHTOOL_A_PAUSE_MAX = 6, ++}; ++ ++enum { ++ ETHTOOL_A_EEE_UNSPEC = 0, ++ ETHTOOL_A_EEE_HEADER = 1, ++ ETHTOOL_A_EEE_MODES_OURS = 2, ++ ETHTOOL_A_EEE_MODES_PEER = 3, ++ ETHTOOL_A_EEE_ACTIVE = 4, ++ ETHTOOL_A_EEE_ENABLED = 5, ++ ETHTOOL_A_EEE_TX_LPI_ENABLED = 6, ++ ETHTOOL_A_EEE_TX_LPI_TIMER = 7, ++ __ETHTOOL_A_EEE_CNT = 8, ++ ETHTOOL_A_EEE_MAX = 7, ++}; ++ ++enum { ++ ETHTOOL_A_TSINFO_UNSPEC = 0, ++ ETHTOOL_A_TSINFO_HEADER = 1, ++ ETHTOOL_A_TSINFO_TIMESTAMPING = 2, ++ ETHTOOL_A_TSINFO_TX_TYPES = 3, ++ ETHTOOL_A_TSINFO_RX_FILTERS = 4, ++ ETHTOOL_A_TSINFO_PHC_INDEX = 5, ++ __ETHTOOL_A_TSINFO_CNT = 6, ++ ETHTOOL_A_TSINFO_MAX = 5, ++}; ++ ++enum { ++ ETHTOOL_A_PHC_VCLOCKS_UNSPEC = 0, ++ ETHTOOL_A_PHC_VCLOCKS_HEADER = 1, ++ ETHTOOL_A_PHC_VCLOCKS_NUM = 2, ++ ETHTOOL_A_PHC_VCLOCKS_INDEX = 3, ++ __ETHTOOL_A_PHC_VCLOCKS_CNT = 4, ++ ETHTOOL_A_PHC_VCLOCKS_MAX = 3, ++}; ++ ++enum { ++ ETHTOOL_A_CABLE_TEST_UNSPEC = 0, ++ ETHTOOL_A_CABLE_TEST_HEADER = 1, ++ __ETHTOOL_A_CABLE_TEST_CNT = 2, ++ ETHTOOL_A_CABLE_TEST_MAX = 1, ++}; ++ ++enum { ++ ETHTOOL_A_CABLE_TEST_TDR_UNSPEC = 0, ++ ETHTOOL_A_CABLE_TEST_TDR_HEADER = 1, ++ ETHTOOL_A_CABLE_TEST_TDR_CFG = 2, ++ __ETHTOOL_A_CABLE_TEST_TDR_CNT = 3, ++ ETHTOOL_A_CABLE_TEST_TDR_MAX = 2, ++}; ++ ++enum { ++ ETHTOOL_A_TUNNEL_INFO_UNSPEC = 0, ++ ETHTOOL_A_TUNNEL_INFO_HEADER = 1, ++ ETHTOOL_A_TUNNEL_INFO_UDP_PORTS = 2, ++ __ETHTOOL_A_TUNNEL_INFO_CNT = 3, ++ ETHTOOL_A_TUNNEL_INFO_MAX = 2, ++}; ++ ++enum { ++ ETHTOOL_A_FEC_UNSPEC = 0, ++ ETHTOOL_A_FEC_HEADER = 1, ++ ETHTOOL_A_FEC_MODES = 2, ++ ETHTOOL_A_FEC_AUTO = 3, ++ ETHTOOL_A_FEC_ACTIVE = 4, ++ ETHTOOL_A_FEC_STATS = 5, ++ __ETHTOOL_A_FEC_CNT = 6, ++ ETHTOOL_A_FEC_MAX = 5, ++}; ++ ++enum { ++ ETHTOOL_A_MODULE_EEPROM_UNSPEC = 0, ++ ETHTOOL_A_MODULE_EEPROM_HEADER = 1, ++ ETHTOOL_A_MODULE_EEPROM_OFFSET = 2, ++ ETHTOOL_A_MODULE_EEPROM_LENGTH = 3, ++ ETHTOOL_A_MODULE_EEPROM_PAGE = 4, ++ ETHTOOL_A_MODULE_EEPROM_BANK = 5, ++ ETHTOOL_A_MODULE_EEPROM_I2C_ADDRESS = 6, ++ ETHTOOL_A_MODULE_EEPROM_DATA = 7, ++ __ETHTOOL_A_MODULE_EEPROM_CNT = 8, ++ ETHTOOL_A_MODULE_EEPROM_MAX = 7, ++}; ++ ++enum { ++ ETHTOOL_A_STATS_UNSPEC = 0, ++ ETHTOOL_A_STATS_PAD = 1, ++ ETHTOOL_A_STATS_HEADER = 2, ++ ETHTOOL_A_STATS_GROUPS = 3, ++ ETHTOOL_A_STATS_GRP = 4, ++ ETHTOOL_A_STATS_SRC = 5, ++ __ETHTOOL_A_STATS_CNT = 6, ++ ETHTOOL_A_STATS_MAX = 5, ++}; ++ ++enum { ++ ETHTOOL_STATS_ETH_PHY = 0, ++ ETHTOOL_STATS_ETH_MAC = 1, ++ ETHTOOL_STATS_ETH_CTRL = 2, ++ ETHTOOL_STATS_RMON = 3, ++ __ETHTOOL_STATS_CNT = 4, ++}; ++ ++enum { ++ ETHTOOL_A_STATS_ETH_PHY_5_SYM_ERR = 0, ++ __ETHTOOL_A_STATS_ETH_PHY_CNT = 1, ++ ETHTOOL_A_STATS_ETH_PHY_MAX = 0, ++}; ++ ++enum { ++ ETHTOOL_A_STATS_ETH_MAC_2_TX_PKT = 0, ++ ETHTOOL_A_STATS_ETH_MAC_3_SINGLE_COL = 1, ++ ETHTOOL_A_STATS_ETH_MAC_4_MULTI_COL = 2, ++ ETHTOOL_A_STATS_ETH_MAC_5_RX_PKT = 3, ++ ETHTOOL_A_STATS_ETH_MAC_6_FCS_ERR = 4, ++ ETHTOOL_A_STATS_ETH_MAC_7_ALIGN_ERR = 5, ++ ETHTOOL_A_STATS_ETH_MAC_8_TX_BYTES = 6, ++ ETHTOOL_A_STATS_ETH_MAC_9_TX_DEFER = 7, ++ ETHTOOL_A_STATS_ETH_MAC_10_LATE_COL = 8, ++ ETHTOOL_A_STATS_ETH_MAC_11_XS_COL = 9, ++ ETHTOOL_A_STATS_ETH_MAC_12_TX_INT_ERR = 10, ++ ETHTOOL_A_STATS_ETH_MAC_13_CS_ERR = 11, ++ ETHTOOL_A_STATS_ETH_MAC_14_RX_BYTES = 12, ++ ETHTOOL_A_STATS_ETH_MAC_15_RX_INT_ERR = 13, ++ ETHTOOL_A_STATS_ETH_MAC_18_TX_MCAST = 14, ++ ETHTOOL_A_STATS_ETH_MAC_19_TX_BCAST = 15, ++ ETHTOOL_A_STATS_ETH_MAC_20_XS_DEFER = 16, ++ ETHTOOL_A_STATS_ETH_MAC_21_RX_MCAST = 17, ++ ETHTOOL_A_STATS_ETH_MAC_22_RX_BCAST = 18, ++ ETHTOOL_A_STATS_ETH_MAC_23_IR_LEN_ERR = 19, ++ ETHTOOL_A_STATS_ETH_MAC_24_OOR_LEN = 20, ++ ETHTOOL_A_STATS_ETH_MAC_25_TOO_LONG_ERR = 21, ++ __ETHTOOL_A_STATS_ETH_MAC_CNT = 22, ++ ETHTOOL_A_STATS_ETH_MAC_MAX = 21, ++}; ++ ++enum { ++ ETHTOOL_A_STATS_ETH_CTRL_3_TX = 0, ++ ETHTOOL_A_STATS_ETH_CTRL_4_RX = 1, ++ ETHTOOL_A_STATS_ETH_CTRL_5_RX_UNSUP = 2, ++ __ETHTOOL_A_STATS_ETH_CTRL_CNT = 3, ++ ETHTOOL_A_STATS_ETH_CTRL_MAX = 2, ++}; ++ ++enum { ++ ETHTOOL_A_STATS_RMON_UNDERSIZE = 0, ++ ETHTOOL_A_STATS_RMON_OVERSIZE = 1, ++ ETHTOOL_A_STATS_RMON_FRAG = 2, ++ ETHTOOL_A_STATS_RMON_JABBER = 3, ++ __ETHTOOL_A_STATS_RMON_CNT = 4, ++ ETHTOOL_A_STATS_RMON_MAX = 3, ++}; ++ ++enum { ++ ETHTOOL_A_MODULE_UNSPEC = 0, ++ ETHTOOL_A_MODULE_HEADER = 1, ++ ETHTOOL_A_MODULE_POWER_MODE_POLICY = 2, ++ ETHTOOL_A_MODULE_POWER_MODE = 3, ++ __ETHTOOL_A_MODULE_CNT = 4, ++ ETHTOOL_A_MODULE_MAX = 3, ++}; ++ ++enum { ++ ETHTOOL_A_PSE_UNSPEC = 0, ++ ETHTOOL_A_PSE_HEADER = 1, ++ ETHTOOL_A_PODL_PSE_ADMIN_STATE = 2, ++ ETHTOOL_A_PODL_PSE_ADMIN_CONTROL = 3, ++ ETHTOOL_A_PODL_PSE_PW_D_STATUS = 4, ++ __ETHTOOL_A_PSE_CNT = 5, ++ ETHTOOL_A_PSE_MAX = 4, ++}; ++ ++enum { ++ ETHTOOL_A_RSS_UNSPEC = 0, ++ ETHTOOL_A_RSS_HEADER = 1, ++ ETHTOOL_A_RSS_CONTEXT = 2, ++ ETHTOOL_A_RSS_HFUNC = 3, ++ ETHTOOL_A_RSS_INDIR = 4, ++ ETHTOOL_A_RSS_HKEY = 5, ++ __ETHTOOL_A_RSS_CNT = 6, ++ ETHTOOL_A_RSS_MAX = 5, ++}; ++ ++enum { ++ ETHTOOL_A_PLCA_UNSPEC = 0, ++ ETHTOOL_A_PLCA_HEADER = 1, ++ ETHTOOL_A_PLCA_VERSION = 2, ++ ETHTOOL_A_PLCA_ENABLED = 3, ++ ETHTOOL_A_PLCA_STATUS = 4, ++ ETHTOOL_A_PLCA_NODE_CNT = 5, ++ ETHTOOL_A_PLCA_NODE_ID = 6, ++ ETHTOOL_A_PLCA_TO_TMR = 7, ++ ETHTOOL_A_PLCA_BURST_CNT = 8, ++ ETHTOOL_A_PLCA_BURST_TMR = 9, ++ __ETHTOOL_A_PLCA_CNT = 10, ++ ETHTOOL_A_PLCA_MAX = 9, ++}; ++ ++enum { ++ ETHTOOL_A_MM_UNSPEC = 0, ++ ETHTOOL_A_MM_HEADER = 1, ++ ETHTOOL_A_MM_PMAC_ENABLED = 2, ++ ETHTOOL_A_MM_TX_ENABLED = 3, ++ ETHTOOL_A_MM_TX_ACTIVE = 4, ++ ETHTOOL_A_MM_TX_MIN_FRAG_SIZE = 5, ++ ETHTOOL_A_MM_RX_MIN_FRAG_SIZE = 6, ++ ETHTOOL_A_MM_VERIFY_ENABLED = 7, ++ ETHTOOL_A_MM_VERIFY_STATUS = 8, ++ ETHTOOL_A_MM_VERIFY_TIME = 9, ++ ETHTOOL_A_MM_MAX_VERIFY_TIME = 10, ++ ETHTOOL_A_MM_STATS = 11, ++ __ETHTOOL_A_MM_CNT = 12, ++ ETHTOOL_A_MM_MAX = 11, ++}; ++ ++enum ethtool_multicast_groups { ++ ETHNL_MCGRP_MONITOR = 0, ++}; ++ ++struct ethnl_req_info { ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ u32 flags; ++}; ++ ++struct ethnl_reply_data { ++ struct net_device *dev; ++}; ++ ++struct ethnl_request_ops { ++ u8 request_cmd; ++ u8 reply_cmd; ++ u16 hdr_attr; ++ unsigned int req_info_size; ++ unsigned int reply_data_size; ++ bool allow_nodev_do; ++ u8 set_ntf_cmd; ++ int (*parse_request)(struct ethnl_req_info *, struct nlattr **, struct netlink_ext_ack *); ++ int (*prepare_data)(const struct ethnl_req_info *, struct ethnl_reply_data *, struct genl_info *); ++ int (*reply_size)(const struct ethnl_req_info *, const struct ethnl_reply_data *); ++ int (*fill_reply)(struct sk_buff *, const struct ethnl_req_info *, const struct ethnl_reply_data *); ++ void (*cleanup_data)(struct ethnl_reply_data *); ++ int (*set_validate)(struct ethnl_req_info *, struct genl_info *); ++ int (*set)(struct ethnl_req_info *, struct genl_info *); ++}; ++ ++struct ethnl_dump_ctx { ++ const struct ethnl_request_ops *ops; ++ struct ethnl_req_info *req_info; ++ struct ethnl_reply_data *reply_data; ++ int pos_hash; ++ int pos_idx; ++}; ++ ++typedef void (*ethnl_notify_handler_t)(struct net_device *, unsigned int, const void *); ++ ++enum tunable_id { ++ ETHTOOL_ID_UNSPEC = 0, ++ ETHTOOL_RX_COPYBREAK = 1, ++ ETHTOOL_TX_COPYBREAK = 2, ++ ETHTOOL_PFC_PREVENTION_TOUT = 3, ++ ETHTOOL_TX_COPYBREAK_BUF_SIZE = 4, ++ __ETHTOOL_TUNABLE_COUNT = 5, ++}; ++ ++enum phy_tunable_id { ++ ETHTOOL_PHY_ID_UNSPEC = 0, ++ ETHTOOL_PHY_DOWNSHIFT = 1, ++ ETHTOOL_PHY_FAST_LINK_DOWN = 2, ++ ETHTOOL_PHY_EDPD = 3, ++ __ETHTOOL_PHY_TUNABLE_COUNT = 4, ++}; ++ ++enum { ++ ETH_RSS_HASH_TOP_BIT = 0, ++ ETH_RSS_HASH_XOR_BIT = 1, ++ ETH_RSS_HASH_CRC32_BIT = 2, ++ ETH_RSS_HASH_FUNCS_COUNT = 3, ++}; ++ ++struct link_mode_info { ++ int speed; ++ u8 lanes; ++ u8 duplex; ++}; ++ ++typedef const char (* const ethnl_string_array_t)[32]; ++ ++struct linkmodes_reply_data { ++ struct ethnl_reply_data base; ++ struct ethtool_link_ksettings ksettings; ++ struct ethtool_link_settings *lsettings; ++ bool peer_empty; ++}; ++ ++struct wol_reply_data { ++ struct ethnl_reply_data base; ++ struct ethtool_wolinfo wol; ++ bool show_sopass; ++}; ++ ++struct channels_reply_data { ++ struct ethnl_reply_data base; ++ struct ethtool_channels channels; ++}; ++ ++struct eee_reply_data { ++ struct ethnl_reply_data base; ++ struct ethtool_eee eee; ++}; ++ ++enum { ++ ETHTOOL_UDP_TUNNEL_TYPE_VXLAN = 0, ++ ETHTOOL_UDP_TUNNEL_TYPE_GENEVE = 1, ++ ETHTOOL_UDP_TUNNEL_TYPE_VXLAN_GPE = 2, ++ __ETHTOOL_UDP_TUNNEL_TYPE_CNT = 3, ++}; ++ ++enum { ++ ETHTOOL_A_TUNNEL_UDP_ENTRY_UNSPEC = 0, ++ ETHTOOL_A_TUNNEL_UDP_ENTRY_PORT = 1, ++ ETHTOOL_A_TUNNEL_UDP_ENTRY_TYPE = 2, ++ __ETHTOOL_A_TUNNEL_UDP_ENTRY_CNT = 3, ++ ETHTOOL_A_TUNNEL_UDP_ENTRY_MAX = 2, ++}; ++ ++enum { ++ ETHTOOL_A_TUNNEL_UDP_TABLE_UNSPEC = 0, ++ ETHTOOL_A_TUNNEL_UDP_TABLE_SIZE = 1, ++ ETHTOOL_A_TUNNEL_UDP_TABLE_TYPES = 2, ++ ETHTOOL_A_TUNNEL_UDP_TABLE_ENTRY = 3, ++ __ETHTOOL_A_TUNNEL_UDP_TABLE_CNT = 4, ++ ETHTOOL_A_TUNNEL_UDP_TABLE_MAX = 3, ++}; ++ ++enum { ++ ETHTOOL_A_TUNNEL_UDP_UNSPEC = 0, ++ ETHTOOL_A_TUNNEL_UDP_TABLE = 1, ++ __ETHTOOL_A_TUNNEL_UDP_CNT = 2, ++ ETHTOOL_A_TUNNEL_UDP_MAX = 1, ++}; ++ ++enum udp_parsable_tunnel_type { ++ UDP_TUNNEL_TYPE_VXLAN = 1, ++ UDP_TUNNEL_TYPE_GENEVE = 2, ++ UDP_TUNNEL_TYPE_VXLAN_GPE = 4, ++}; ++ ++struct udp_tunnel_info { ++ short unsigned int type; ++ sa_family_t sa_family; ++ __be16 port; ++ u8 hw_priv; ++}; ++ ++enum udp_tunnel_nic_info_flags { ++ UDP_TUNNEL_NIC_INFO_MAY_SLEEP = 1, ++ UDP_TUNNEL_NIC_INFO_OPEN_ONLY = 2, ++ UDP_TUNNEL_NIC_INFO_IPV4_ONLY = 4, ++ UDP_TUNNEL_NIC_INFO_STATIC_IANA_VXLAN = 8, ++}; ++ ++struct udp_tunnel_nic_shared { ++ struct udp_tunnel_nic *udp_tunnel_nic_info; ++ struct list_head devices; ++}; ++ ++struct udp_tunnel_nic_ops { ++ void (*get_port)(struct net_device *, unsigned int, unsigned int, struct udp_tunnel_info *); ++ void (*set_port_priv)(struct net_device *, unsigned int, unsigned int, u8); ++ void (*add_port)(struct net_device *, struct udp_tunnel_info *); ++ void (*del_port)(struct net_device *, struct udp_tunnel_info *); ++ void (*reset_ntf)(struct net_device *); ++ size_t (*dump_size)(struct net_device *, unsigned int); ++ int (*dump_write)(struct net_device *, unsigned int, struct sk_buff *); ++}; ++ ++struct ethnl_tunnel_info_dump_ctx { ++ struct ethnl_req_info req_info; ++ int pos_hash; ++ int pos_idx; ++}; ++ ++struct phc_vclocks_reply_data { ++ struct ethnl_reply_data base; ++ int num; ++ int *index; ++}; ++ ++struct ethtool_phy_ops { ++ int (*get_sset_count)(struct phy_device *); ++ int (*get_strings)(struct phy_device *, u8 *); ++ int (*get_stats)(struct phy_device *, struct ethtool_stats *, u64 *); ++ int (*get_plca_cfg)(struct phy_device *, struct phy_plca_cfg *); ++ int (*set_plca_cfg)(struct phy_device *, const struct phy_plca_cfg *, struct netlink_ext_ack *); ++ int (*get_plca_status)(struct phy_device *, struct phy_plca_status *); ++ int (*start_cable_test)(struct phy_device *, struct netlink_ext_ack *); ++ int (*start_cable_test_tdr)(struct phy_device *, struct netlink_ext_ack *, const struct phy_tdr_config *); ++}; ++ ++struct plca_reply_data { ++ struct ethnl_reply_data base; ++ struct phy_plca_cfg plca_cfg; ++ struct phy_plca_status plca_st; ++}; ++ ++struct nf_sockopt_ops { ++ struct list_head list; ++ u_int8_t pf; ++ int set_optmin; ++ int set_optmax; ++ int (*set)(struct sock *, int, sockptr_t, unsigned int); ++ int get_optmin; ++ int get_optmax; ++ int (*get)(struct sock *, int, void *, int *); ++ struct module *owner; ++}; ++ ++typedef u16 u_int16_t; ++ ++struct seq_net_private { ++ struct net *net; ++ netns_tracker ns_tracker; ++}; ++ ++enum nf_hook_ops_type { ++ NF_HOOK_OP_UNDEFINED = 0, ++ NF_HOOK_OP_NF_TABLES = 1, ++ NF_HOOK_OP_BPF = 2, ++}; ++ ++struct nf_hook_ops { ++ nf_hookfn *hook; ++ struct net_device *dev; ++ void *priv; ++ u8 pf; ++ enum nf_hook_ops_type hook_ops_type: 8; ++ unsigned int hooknum; ++ int priority; ++}; ++ ++struct xt_action_param; ++ ++struct xt_mtchk_param; ++ ++struct xt_mtdtor_param; ++ ++struct xt_match { ++ struct list_head list; ++ const char name[29]; ++ u_int8_t revision; ++ bool (*match)(const struct sk_buff *, struct xt_action_param *); ++ int (*checkentry)(const struct xt_mtchk_param *); ++ void (*destroy)(const struct xt_mtdtor_param *); ++ void (*compat_from_user)(void *, const void *); ++ int (*compat_to_user)(void *, const void *); ++ struct module *me; ++ const char *table; ++ unsigned int matchsize; ++ unsigned int usersize; ++ unsigned int compatsize; ++ unsigned int hooks; ++ short unsigned int proto; ++ short unsigned int family; ++}; ++ ++struct xt_entry_match { ++ union { ++ struct { ++ __u16 match_size; ++ char name[29]; ++ __u8 revision; ++ } user; ++ struct { ++ __u16 match_size; ++ struct xt_match *match; ++ } kernel; ++ __u16 match_size; ++ } u; ++ unsigned char data[0]; ++}; ++ ++struct xt_tgchk_param; ++ ++struct xt_tgdtor_param; ++ ++struct xt_target { ++ struct list_head list; ++ const char name[29]; ++ u_int8_t revision; ++ unsigned int (*target)(struct sk_buff *, const struct xt_action_param *); ++ int (*checkentry)(const struct xt_tgchk_param *); ++ void (*destroy)(const struct xt_tgdtor_param *); ++ void (*compat_from_user)(void *, const void *); ++ int (*compat_to_user)(void *, const void *); ++ struct module *me; ++ const char *table; ++ unsigned int targetsize; ++ unsigned int usersize; ++ unsigned int compatsize; ++ unsigned int hooks; ++ short unsigned int proto; ++ short unsigned int family; ++}; ++ ++struct xt_entry_target { ++ union { ++ struct { ++ __u16 target_size; ++ char name[29]; ++ __u8 revision; ++ } user; ++ struct { ++ __u16 target_size; ++ struct xt_target *target; ++ } kernel; ++ __u16 target_size; ++ } u; ++ unsigned char data[0]; ++}; ++ ++struct xt_standard_target { ++ struct xt_entry_target target; ++ int verdict; ++}; ++ ++struct xt_error_target { ++ struct xt_entry_target target; ++ char errorname[30]; ++}; ++ ++struct xt_counters { ++ __u64 pcnt; ++ __u64 bcnt; ++}; ++ ++struct xt_counters_info { ++ char name[32]; ++ unsigned int num_counters; ++ struct xt_counters counters[0]; ++}; ++ ++struct xt_action_param { ++ union { ++ const struct xt_match *match; ++ const struct xt_target *target; ++ }; ++ union { ++ const void *matchinfo; ++ const void *targinfo; ++ }; ++ const struct nf_hook_state *state; ++ unsigned int thoff; ++ u16 fragoff; ++ bool hotdrop; ++}; ++ ++struct xt_mtchk_param { ++ struct net *net; ++ const char *table; ++ const void *entryinfo; ++ const struct xt_match *match; ++ void *matchinfo; ++ unsigned int hook_mask; ++ u_int8_t family; ++ bool nft_compat; ++}; ++ ++struct xt_mtdtor_param { ++ struct net *net; ++ const struct xt_match *match; ++ void *matchinfo; ++ u_int8_t family; ++}; ++ ++struct xt_tgchk_param { ++ struct net *net; ++ const char *table; ++ const void *entryinfo; ++ const struct xt_target *target; ++ void *targinfo; ++ unsigned int hook_mask; ++ u_int8_t family; ++ bool nft_compat; ++}; ++ ++struct xt_tgdtor_param { ++ struct net *net; ++ const struct xt_target *target; ++ void *targinfo; ++ u_int8_t family; ++}; ++ ++struct xt_table_info; ++ ++struct xt_table { ++ struct list_head list; ++ unsigned int valid_hooks; ++ struct xt_table_info *private; ++ struct nf_hook_ops *ops; ++ struct module *me; ++ u_int8_t af; ++ int priority; ++ const char name[32]; ++}; ++ ++struct xt_table_info { ++ unsigned int size; ++ unsigned int number; ++ unsigned int initial_entries; ++ unsigned int hook_entry[5]; ++ unsigned int underflow[5]; ++ unsigned int stacksize; ++ void ***jumpstack; ++ unsigned char entries[0]; ++}; ++ ++struct xt_percpu_counter_alloc_state { ++ unsigned int off; ++ const char *mem; ++}; ++ ++struct compat_xt_entry_match { ++ union { ++ struct { ++ u_int16_t match_size; ++ char name[29]; ++ u_int8_t revision; ++ } user; ++ struct { ++ u_int16_t match_size; ++ compat_uptr_t match; ++ } kernel; ++ u_int16_t match_size; ++ } u; ++ unsigned char data[0]; ++}; ++ ++struct compat_xt_entry_target { ++ union { ++ struct { ++ u_int16_t target_size; ++ char name[29]; ++ u_int8_t revision; ++ } user; ++ struct { ++ u_int16_t target_size; ++ compat_uptr_t target; ++ } kernel; ++ u_int16_t target_size; ++ } u; ++ unsigned char data[0]; ++}; ++ ++struct compat_xt_counters { ++ compat_u64 pcnt; ++ compat_u64 bcnt; ++}; ++ ++struct compat_xt_counters_info { ++ char name[32]; ++ compat_uint_t num_counters; ++ struct compat_xt_counters counters[0]; ++}; ++ ++struct xt_template { ++ struct list_head list; ++ int (*table_init)(struct net *); ++ struct module *me; ++ char name[32]; ++}; ++ ++struct xt_pernet { ++ struct list_head tables[11]; ++}; ++ ++struct compat_delta { ++ unsigned int offset; ++ int delta; ++}; ++ ++struct xt_af { ++ struct mutex mutex; ++ struct list_head match; ++ struct list_head target; ++ struct mutex compat_mutex; ++ struct compat_delta *compat_tab; ++ unsigned int number; ++ unsigned int cur; ++}; ++ ++struct compat_xt_standard_target { ++ struct compat_xt_entry_target t; ++ compat_uint_t verdict; ++}; ++ ++struct compat_xt_error_target { ++ struct compat_xt_entry_target t; ++ char errorname[30]; ++}; ++ ++struct nf_mttg_trav { ++ struct list_head *head; ++ struct list_head *curr; ++ uint8_t class; ++}; ++ ++enum { ++ MTTG_TRAV_INIT = 0, ++ MTTG_TRAV_NFP_UNSPEC = 1, ++ MTTG_TRAV_NFP_SPEC = 2, ++ MTTG_TRAV_DONE = 3, ++}; ++ ++struct ipv4_addr_key { ++ __be32 addr; ++ int vif; ++}; ++ ++struct inetpeer_addr { ++ union { ++ struct ipv4_addr_key a4; ++ struct in6_addr a6; ++ u32 key[4]; ++ }; ++ __u16 family; ++}; ++ ++struct inet_peer { ++ struct rb_node rb_node; ++ struct inetpeer_addr daddr; ++ u32 metrics[17]; ++ u32 rate_tokens; ++ u32 n_redirects; ++ long unsigned int rate_last; ++ union { ++ struct { ++ atomic_t rid; ++ }; ++ struct callback_head rcu; ++ }; ++ __u32 dtime; ++ refcount_t refcnt; ++}; ++ ++enum { ++ IPV4_DEVCONF_FORWARDING = 1, ++ IPV4_DEVCONF_MC_FORWARDING = 2, ++ IPV4_DEVCONF_PROXY_ARP = 3, ++ IPV4_DEVCONF_ACCEPT_REDIRECTS = 4, ++ IPV4_DEVCONF_SECURE_REDIRECTS = 5, ++ IPV4_DEVCONF_SEND_REDIRECTS = 6, ++ IPV4_DEVCONF_SHARED_MEDIA = 7, ++ IPV4_DEVCONF_RP_FILTER = 8, ++ IPV4_DEVCONF_ACCEPT_SOURCE_ROUTE = 9, ++ IPV4_DEVCONF_BOOTP_RELAY = 10, ++ IPV4_DEVCONF_LOG_MARTIANS = 11, ++ IPV4_DEVCONF_TAG = 12, ++ IPV4_DEVCONF_ARPFILTER = 13, ++ IPV4_DEVCONF_MEDIUM_ID = 14, ++ IPV4_DEVCONF_NOXFRM = 15, ++ IPV4_DEVCONF_NOPOLICY = 16, ++ IPV4_DEVCONF_FORCE_IGMP_VERSION = 17, ++ IPV4_DEVCONF_ARP_ANNOUNCE = 18, ++ IPV4_DEVCONF_ARP_IGNORE = 19, ++ IPV4_DEVCONF_PROMOTE_SECONDARIES = 20, ++ IPV4_DEVCONF_ARP_ACCEPT = 21, ++ IPV4_DEVCONF_ARP_NOTIFY = 22, ++ IPV4_DEVCONF_ACCEPT_LOCAL = 23, ++ IPV4_DEVCONF_SRC_VMARK = 24, ++ IPV4_DEVCONF_PROXY_ARP_PVLAN = 25, ++ IPV4_DEVCONF_ROUTE_LOCALNET = 26, ++ IPV4_DEVCONF_IGMPV2_UNSOLICITED_REPORT_INTERVAL = 27, ++ IPV4_DEVCONF_IGMPV3_UNSOLICITED_REPORT_INTERVAL = 28, ++ IPV4_DEVCONF_IGNORE_ROUTES_WITH_LINKDOWN = 29, ++ IPV4_DEVCONF_DROP_UNICAST_IN_L2_MULTICAST = 30, ++ IPV4_DEVCONF_DROP_GRATUITOUS_ARP = 31, ++ IPV4_DEVCONF_BC_FORWARDING = 32, ++ IPV4_DEVCONF_ARP_EVICT_NOCARRIER = 33, ++ __IPV4_DEVCONF_MAX = 34, ++}; ++ ++enum { ++ RTN_UNSPEC = 0, ++ RTN_UNICAST = 1, ++ RTN_LOCAL = 2, ++ RTN_BROADCAST = 3, ++ RTN_ANYCAST = 4, ++ RTN_MULTICAST = 5, ++ RTN_BLACKHOLE = 6, ++ RTN_UNREACHABLE = 7, ++ RTN_PROHIBIT = 8, ++ RTN_THROW = 9, ++ RTN_NAT = 10, ++ RTN_XRESOLVE = 11, ++ __RTN_MAX = 12, ++}; ++ ++struct raw_hashinfo { ++ spinlock_t lock; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct hlist_head ht[256]; ++}; ++ ++struct ip_rt_acct { ++ __u32 o_bytes; ++ __u32 o_packets; ++ __u32 i_bytes; ++ __u32 i_packets; ++}; ++ ++enum ip_defrag_users { ++ IP_DEFRAG_LOCAL_DELIVER = 0, ++ IP_DEFRAG_CALL_RA_CHAIN = 1, ++ IP_DEFRAG_CONNTRACK_IN = 2, ++ __IP_DEFRAG_CONNTRACK_IN_END = 65537, ++ IP_DEFRAG_CONNTRACK_OUT = 65538, ++ __IP_DEFRAG_CONNTRACK_OUT_END = 131073, ++ IP_DEFRAG_CONNTRACK_BRIDGE_IN = 131074, ++ __IP_DEFRAG_CONNTRACK_BRIDGE_IN = 196609, ++ IP_DEFRAG_VS_IN = 196610, ++ IP_DEFRAG_VS_OUT = 196611, ++ IP_DEFRAG_VS_FWD = 196612, ++ IP_DEFRAG_AF_PACKET = 196613, ++ IP_DEFRAG_MACVLAN = 196614, ++}; ++ ++struct net_protocol { ++ int (*handler)(struct sk_buff *); ++ int (*err_handler)(struct sk_buff *, u32); ++ unsigned int no_policy: 1; ++ unsigned int icmp_strict_tag_validation: 1; ++}; ++ ++enum { ++ INET_ECN_NOT_ECT = 0, ++ INET_ECN_ECT_1 = 1, ++ INET_ECN_ECT_0 = 2, ++ INET_ECN_CE = 3, ++ INET_ECN_MASK = 3, ++}; ++ ++enum { ++ XFRM_DEV_OFFLOAD_UNSPECIFIED = 0, ++ XFRM_DEV_OFFLOAD_CRYPTO = 1, ++ XFRM_DEV_OFFLOAD_PACKET = 2, ++}; ++ ++struct fastopen_queue { ++ struct request_sock *rskq_rst_head; ++ struct request_sock *rskq_rst_tail; ++ spinlock_t lock; ++ int qlen; ++ int max_qlen; ++ struct tcp_fastopen_context *ctx; ++}; ++ ++struct request_sock_queue { ++ spinlock_t rskq_lock; ++ u8 rskq_defer_accept; ++ u32 synflood_warned; ++ atomic_t qlen; ++ atomic_t young; ++ struct request_sock *rskq_accept_head; ++ struct request_sock *rskq_accept_tail; ++ struct fastopen_queue fastopenq; ++}; ++ ++struct inet_connection_sock_af_ops { ++ int (*queue_xmit)(struct sock *, struct sk_buff *, struct flowi *); ++ void (*send_check)(struct sock *, struct sk_buff *); ++ int (*rebuild_header)(struct sock *); ++ void (*sk_rx_dst_set)(struct sock *, const struct sk_buff *); ++ int (*conn_request)(struct sock *, struct sk_buff *); ++ struct sock * (*syn_recv_sock)(const struct sock *, struct sk_buff *, struct request_sock *, struct dst_entry *, struct request_sock *, bool *); ++ u16 net_header_len; ++ u16 net_frag_header_len; ++ u16 sockaddr_len; ++ int (*setsockopt)(struct sock *, int, int, sockptr_t, unsigned int); ++ int (*getsockopt)(struct sock *, int, int, char *, int *); ++ void (*addr2sockaddr)(struct sock *, struct sockaddr *); ++ void (*mtu_reduced)(struct sock *); ++}; ++ ++struct inet_bind_bucket; ++ ++struct inet_bind2_bucket; ++ ++struct tcp_ulp_ops; ++ ++struct inet_connection_sock { ++ struct inet_sock icsk_inet; ++ struct request_sock_queue icsk_accept_queue; ++ struct inet_bind_bucket *icsk_bind_hash; ++ struct inet_bind2_bucket *icsk_bind2_hash; ++ long unsigned int icsk_timeout; ++ struct timer_list icsk_retransmit_timer; ++ struct timer_list icsk_delack_timer; ++ __u32 icsk_rto; ++ __u32 icsk_rto_min; ++ __u32 icsk_delack_max; ++ __u32 icsk_pmtu_cookie; ++ const struct tcp_congestion_ops *icsk_ca_ops; ++ const struct inet_connection_sock_af_ops *icsk_af_ops; ++ const struct tcp_ulp_ops *icsk_ulp_ops; ++ void *icsk_ulp_data; ++ void (*icsk_clean_acked)(struct sock *, u32); ++ unsigned int (*icsk_sync_mss)(struct sock *, u32); ++ __u8 icsk_ca_state: 5; ++ __u8 icsk_ca_initialized: 1; ++ __u8 icsk_ca_setsockopt: 1; ++ __u8 icsk_ca_dst_locked: 1; ++ __u8 icsk_retransmits; ++ __u8 icsk_pending; ++ __u8 icsk_backoff; ++ __u8 icsk_syn_retries; ++ __u8 icsk_probes_out; ++ __u16 icsk_ext_hdr_len; ++ struct { ++ __u8 pending; ++ __u8 quick; ++ __u8 pingpong; ++ __u8 retry; ++ __u32 ato; ++ long unsigned int timeout; ++ __u32 lrcvtime; ++ __u16 last_seg_size; ++ __u16 rcv_mss; ++ } icsk_ack; ++ struct { ++ int search_high; ++ int search_low; ++ u32 probe_size: 31; ++ u32 enabled: 1; ++ u32 probe_timestamp; ++ } icsk_mtup; ++ u32 icsk_probes_tstamp; ++ u32 icsk_user_timeout; ++ u64 icsk_ca_priv[13]; ++}; ++ ++struct inet_bind_bucket { ++ possible_net_t ib_net; ++ int l3mdev; ++ short unsigned int port; ++ signed char fastreuse; ++ signed char fastreuseport; ++ kuid_t fastuid; ++ struct in6_addr fast_v6_rcv_saddr; ++ __be32 fast_rcv_saddr; ++ short unsigned int fast_sk_family; ++ bool fast_ipv6_only; ++ struct hlist_node node; ++ struct hlist_head owners; ++}; ++ ++struct inet_bind2_bucket { ++ possible_net_t ib_net; ++ int l3mdev; ++ short unsigned int port; ++ short unsigned int family; ++ union { ++ struct in6_addr v6_rcv_saddr; ++ __be32 rcv_saddr; ++ }; ++ struct hlist_node node; ++ struct hlist_head owners; ++ struct hlist_head deathrow; ++}; ++ ++struct tcp_ulp_ops { ++ struct list_head list; ++ int (*init)(struct sock *); ++ void (*update)(struct sock *, struct proto *, void (*)(struct sock *)); ++ void (*release)(struct sock *); ++ int (*get_info)(const struct sock *, struct sk_buff *); ++ size_t (*get_info_size)(const struct sock *); ++ void (*clone)(const struct request_sock *, struct sock *, const gfp_t); ++ char name[16]; ++ struct module *owner; ++}; ++ ++struct inet_timewait_sock { ++ struct sock_common __tw_common; ++ __u32 tw_mark; ++ volatile unsigned char tw_substate; ++ unsigned char tw_rcv_wscale; ++ __be16 tw_sport; ++ unsigned int tw_transparent: 1; ++ unsigned int tw_flowlabel: 20; ++ unsigned int tw_pad: 3; ++ unsigned int tw_tos: 8; ++ u32 tw_txhash; ++ u32 tw_priority; ++ struct timer_list tw_timer; ++ struct inet_bind_bucket *tw_tb; ++ struct inet_bind2_bucket *tw_tb2; ++ struct hlist_node tw_bind2_node; ++}; ++ ++struct minmax_sample { ++ u32 t; ++ u32 v; ++}; ++ ++struct minmax { ++ struct minmax_sample s[3]; ++}; ++ ++struct inet_request_sock { ++ struct request_sock req; ++ u16 snd_wscale: 4; ++ u16 rcv_wscale: 4; ++ u16 tstamp_ok: 1; ++ u16 sack_ok: 1; ++ u16 wscale_ok: 1; ++ u16 ecn_ok: 1; ++ u16 acked: 1; ++ u16 no_srccheck: 1; ++ u16 smc_ok: 1; ++ u32 ir_mark; ++ union { ++ struct ip_options_rcu *ireq_opt; ++ struct { ++ struct ipv6_txoptions *ipv6_opt; ++ struct sk_buff *pktopts; ++ }; ++ }; ++}; ++ ++enum inet_csk_ack_state_t { ++ ICSK_ACK_SCHED = 1, ++ ICSK_ACK_TIMER = 2, ++ ICSK_ACK_PUSHED = 4, ++ ICSK_ACK_PUSHED2 = 8, ++ ICSK_ACK_NOW = 16, ++}; ++ ++union tcp_word_hdr { ++ struct tcphdr hdr; ++ __be32 words[5]; ++}; ++ ++enum { ++ TCP_FLAG_CWR = 8388608, ++ TCP_FLAG_ECE = 4194304, ++ TCP_FLAG_URG = 2097152, ++ TCP_FLAG_ACK = 1048576, ++ TCP_FLAG_PSH = 524288, ++ TCP_FLAG_RST = 262144, ++ TCP_FLAG_SYN = 131072, ++ TCP_FLAG_FIN = 65536, ++ TCP_RESERVED_BITS = 251658240, ++ TCP_DATA_OFFSET = 4026531840, ++}; ++ ++enum tcp_fastopen_client_fail { ++ TFO_STATUS_UNSPEC = 0, ++ TFO_COOKIE_UNAVAILABLE = 1, ++ TFO_DATA_NOT_ACKED = 2, ++ TFO_SYN_RETRANSMITTED = 3, ++}; ++ ++enum tcp_ca_state { ++ TCP_CA_Open = 0, ++ TCP_CA_Disorder = 1, ++ TCP_CA_CWR = 2, ++ TCP_CA_Recovery = 3, ++ TCP_CA_Loss = 4, ++}; ++ ++struct tcp_fastopen_cookie { ++ __le64 val[2]; ++ s8 len; ++ bool exp; ++}; ++ ++struct tcp_sack_block_wire { ++ __be32 start_seq; ++ __be32 end_seq; ++}; ++ ++struct tcp_sack_block { ++ u32 start_seq; ++ u32 end_seq; ++}; ++ ++struct tcp_options_received { ++ int ts_recent_stamp; ++ u32 ts_recent; ++ u32 rcv_tsval; ++ u32 rcv_tsecr; ++ u16 saw_tstamp: 1; ++ u16 tstamp_ok: 1; ++ u16 dsack: 1; ++ u16 wscale_ok: 1; ++ u16 sack_ok: 3; ++ u16 smc_ok: 1; ++ u16 snd_wscale: 4; ++ u16 rcv_wscale: 4; ++ u8 saw_unknown: 1; ++ u8 unused: 7; ++ u8 num_sacks; ++ u16 user_mss; ++ u16 mss_clamp; ++}; ++ ++struct tcp_request_sock_ops; ++ ++struct tcp_request_sock { ++ struct inet_request_sock req; ++ const struct tcp_request_sock_ops *af_specific; ++ u64 snt_synack; ++ bool tfo_listener; ++ bool is_mptcp; ++ bool drop_req; ++ u32 txhash; ++ u32 rcv_isn; ++ u32 snt_isn; ++ u32 ts_off; ++ u32 last_oow_ack_time; ++ u32 rcv_nxt; ++ u8 syn_tos; ++}; ++ ++enum tcp_synack_type { ++ TCP_SYNACK_NORMAL = 0, ++ TCP_SYNACK_FASTOPEN = 1, ++ TCP_SYNACK_COOKIE = 2, ++}; ++ ++struct tcp_md5sig_key; ++ ++struct tcp_request_sock_ops { ++ u16 mss_clamp; ++ struct tcp_md5sig_key * (*req_md5_lookup)(const struct sock *, const struct sock *); ++ int (*calc_md5_hash)(char *, const struct tcp_md5sig_key *, const struct sock *, const struct sk_buff *); ++ __u32 (*cookie_init_seq)(const struct sk_buff *, __u16 *); ++ struct dst_entry * (*route_req)(const struct sock *, struct sk_buff *, struct flowi *, struct request_sock *); ++ u32 (*init_seq)(const struct sk_buff *); ++ u32 (*init_ts_off)(const struct net *, const struct sk_buff *); ++ int (*send_synack)(const struct sock *, struct dst_entry *, struct flowi *, struct request_sock *, struct tcp_fastopen_cookie *, enum tcp_synack_type, struct sk_buff *); ++}; ++ ++struct tcp_rack { ++ u64 mstamp; ++ u32 rtt_us; ++ u32 end_seq; ++ u32 last_delivered; ++ u8 reo_wnd_steps; ++ u8 reo_wnd_persist: 5; ++ u8 dsack_seen: 1; ++ u8 advanced: 1; ++}; ++ ++struct tcp_sock_af_ops; ++ ++struct tcp_md5sig_info; ++ ++struct tcp_fastopen_request; ++ ++struct tcp_sock { ++ struct inet_connection_sock inet_conn; ++ u16 tcp_header_len; ++ u16 gso_segs; ++ __be32 pred_flags; ++ u64 bytes_received; ++ u32 segs_in; ++ u32 data_segs_in; ++ u32 rcv_nxt; ++ u32 copied_seq; ++ u32 rcv_wup; ++ u32 snd_nxt; ++ u32 segs_out; ++ u32 data_segs_out; ++ u64 bytes_sent; ++ u64 bytes_acked; ++ u32 dsack_dups; ++ u32 snd_una; ++ u32 snd_sml; ++ u32 rcv_tstamp; ++ u32 lsndtime; ++ u32 last_oow_ack_time; ++ u32 compressed_ack_rcv_nxt; ++ u32 tsoffset; ++ struct list_head tsq_node; ++ struct list_head tsorted_sent_queue; ++ u32 snd_wl1; ++ u32 snd_wnd; ++ u32 max_window; ++ u32 mss_cache; ++ u32 window_clamp; ++ u32 rcv_ssthresh; ++ struct tcp_rack rack; ++ u16 advmss; ++ u8 compressed_ack; ++ u8 dup_ack_counter: 2; ++ u8 tlp_retrans: 1; ++ u8 unused: 5; ++ u32 chrono_start; ++ u32 chrono_stat[3]; ++ u8 chrono_type: 2; ++ u8 rate_app_limited: 1; ++ u8 fastopen_connect: 1; ++ u8 fastopen_no_cookie: 1; ++ u8 is_sack_reneg: 1; ++ u8 fastopen_client_fail: 2; ++ u8 nonagle: 4; ++ u8 thin_lto: 1; ++ u8 recvmsg_inq: 1; ++ u8 repair: 1; ++ u8 frto: 1; ++ u8 repair_queue; ++ u8 save_syn: 2; ++ u8 syn_data: 1; ++ u8 syn_fastopen: 1; ++ u8 syn_fastopen_exp: 1; ++ u8 syn_fastopen_ch: 1; ++ u8 syn_data_acked: 1; ++ u8 is_cwnd_limited: 1; ++ u32 tlp_high_seq; ++ u32 tcp_tx_delay; ++ u64 tcp_wstamp_ns; ++ u64 tcp_clock_cache; ++ u64 tcp_mstamp; ++ u32 srtt_us; ++ u32 mdev_us; ++ u32 mdev_max_us; ++ u32 rttvar_us; ++ u32 rtt_seq; ++ struct minmax rtt_min; ++ u32 packets_out; ++ u32 retrans_out; ++ u32 max_packets_out; ++ u32 cwnd_usage_seq; ++ u16 urg_data; ++ u8 ecn_flags; ++ u8 keepalive_probes; ++ u32 reordering; ++ u32 reord_seen; ++ u32 snd_up; ++ struct tcp_options_received rx_opt; ++ u32 snd_ssthresh; ++ u32 snd_cwnd; ++ u32 snd_cwnd_cnt; ++ u32 snd_cwnd_clamp; ++ u32 snd_cwnd_used; ++ u32 snd_cwnd_stamp; ++ u32 prior_cwnd; ++ u32 prr_delivered; ++ u32 prr_out; ++ u32 delivered; ++ u32 delivered_ce; ++ u32 lost; ++ u32 app_limited; ++ u64 first_tx_mstamp; ++ u64 delivered_mstamp; ++ u32 rate_delivered; ++ u32 rate_interval_us; ++ u32 rcv_wnd; ++ u32 write_seq; ++ u32 notsent_lowat; ++ u32 pushed_seq; ++ u32 lost_out; ++ u32 sacked_out; ++ struct hrtimer pacing_timer; ++ struct hrtimer compressed_ack_timer; ++ struct sk_buff *lost_skb_hint; ++ struct sk_buff *retransmit_skb_hint; ++ struct rb_root out_of_order_queue; ++ struct sk_buff *ooo_last_skb; ++ struct tcp_sack_block duplicate_sack[1]; ++ struct tcp_sack_block selective_acks[4]; ++ struct tcp_sack_block recv_sack_cache[4]; ++ struct sk_buff *highest_sack; ++ int lost_cnt_hint; ++ u32 prior_ssthresh; ++ u32 high_seq; ++ u32 retrans_stamp; ++ u32 undo_marker; ++ int undo_retrans; ++ u64 bytes_retrans; ++ u32 total_retrans; ++ u32 urg_seq; ++ unsigned int keepalive_time; ++ unsigned int keepalive_intvl; ++ int linger2; ++ u8 bpf_sock_ops_cb_flags; ++ u8 bpf_chg_cc_inprogress: 1; ++ u16 timeout_rehash; ++ u32 rcv_ooopack; ++ u32 rcv_rtt_last_tsecr; ++ struct { ++ u32 rtt_us; ++ u32 seq; ++ u64 time; ++ } rcv_rtt_est; ++ struct { ++ u32 space; ++ u32 seq; ++ u64 time; ++ } rcvq_space; ++ struct { ++ u32 probe_seq_start; ++ u32 probe_seq_end; ++ } mtu_probe; ++ u32 plb_rehash; ++ u32 mtu_info; ++ bool is_mptcp; ++ bool (*smc_hs_congested)(const struct sock *); ++ bool syn_smc; ++ const struct tcp_sock_af_ops *af_specific; ++ struct tcp_md5sig_info *md5sig_info; ++ struct tcp_fastopen_request *fastopen_req; ++ struct request_sock *fastopen_rsk; ++ struct saved_syn *saved_syn; ++}; ++ ++struct tcp_sock_af_ops { ++ struct tcp_md5sig_key * (*md5_lookup)(const struct sock *, const struct sock *); ++ int (*calc_md5_hash)(char *, const struct tcp_md5sig_key *, const struct sock *, const struct sk_buff *); ++ int (*md5_parse)(struct sock *, int, sockptr_t, int); ++}; ++ ++struct tcp_md5sig_info { ++ struct hlist_head head; ++ struct callback_head rcu; ++}; ++ ++struct tcp_fastopen_request { ++ struct tcp_fastopen_cookie cookie; ++ struct msghdr *data; ++ size_t size; ++ int copied; ++ struct ubuf_info *uarg; ++}; ++ ++union tcp_md5_addr { ++ struct in_addr a4; ++ struct in6_addr a6; ++}; ++ ++struct tcp_md5sig_key { ++ struct hlist_node node; ++ u8 keylen; ++ u8 family; ++ u8 prefixlen; ++ u8 flags; ++ union tcp_md5_addr addr; ++ int l3index; ++ u8 key[80]; ++ struct callback_head rcu; ++}; ++ ++struct static_key_false_deferred { ++ struct static_key_false key; ++ long unsigned int timeout; ++ struct delayed_work work; ++}; ++ ++enum { ++ BPF_SOCK_OPS_RTO_CB_FLAG = 1, ++ BPF_SOCK_OPS_RETRANS_CB_FLAG = 2, ++ BPF_SOCK_OPS_STATE_CB_FLAG = 4, ++ BPF_SOCK_OPS_RTT_CB_FLAG = 8, ++ BPF_SOCK_OPS_PARSE_ALL_HDR_OPT_CB_FLAG = 16, ++ BPF_SOCK_OPS_PARSE_UNKNOWN_HDR_OPT_CB_FLAG = 32, ++ BPF_SOCK_OPS_WRITE_HDR_OPT_CB_FLAG = 64, ++ BPF_SOCK_OPS_ALL_CB_FLAGS = 127, ++}; ++ ++struct mptcp_ext { ++ union { ++ u64 data_ack; ++ u32 data_ack32; ++ }; ++ u64 data_seq; ++ u32 subflow_seq; ++ u16 data_len; ++ __sum16 csum; ++ u8 use_map: 1; ++ u8 dsn64: 1; ++ u8 data_fin: 1; ++ u8 use_ack: 1; ++ u8 ack64: 1; ++ u8 mpc_map: 1; ++ u8 frozen: 1; ++ u8 reset_transient: 1; ++ u8 reset_reason: 4; ++ u8 csum_reqd: 1; ++ u8 infinite_map: 1; ++}; ++ ++enum tcp_queue { ++ TCP_FRAG_IN_WRITE_QUEUE = 0, ++ TCP_FRAG_IN_RTX_QUEUE = 1, ++}; ++ ++struct tcp_skb_cb { ++ __u32 seq; ++ __u32 end_seq; ++ union { ++ __u32 tcp_tw_isn; ++ struct { ++ u16 tcp_gso_segs; ++ u16 tcp_gso_size; ++ }; ++ }; ++ __u8 tcp_flags; ++ __u8 sacked; ++ __u8 ip_dsfield; ++ __u8 txstamp_ack: 1; ++ __u8 eor: 1; ++ __u8 has_rxtstamp: 1; ++ __u8 unused: 5; ++ __u32 ack_seq; ++ union { ++ struct { ++ __u32 is_app_limited: 1; ++ __u32 delivered_ce: 20; ++ __u32 unused: 11; ++ __u32 delivered; ++ u64 first_tx_mstamp; ++ u64 delivered_mstamp; ++ } tx; ++ union { ++ struct inet_skb_parm h4; ++ struct inet6_skb_parm h6; ++ } header; ++ }; ++}; ++ ++enum tcp_ca_ack_event_flags { ++ CA_ACK_SLOWPATH = 1, ++ CA_ACK_WIN_UPDATE = 2, ++ CA_ACK_ECE = 4, ++}; ++ ++enum tcp_chrono { ++ TCP_CHRONO_UNSPEC = 0, ++ TCP_CHRONO_BUSY = 1, ++ TCP_CHRONO_RWND_LIMITED = 2, ++ TCP_CHRONO_SNDBUF_LIMITED = 3, ++ __TCP_CHRONO_MAX = 4, ++}; ++ ++struct tcp_sacktag_state { ++ u64 first_sackt; ++ u64 last_sackt; ++ u32 reord; ++ u32 sack_delivered; ++ int flag; ++ unsigned int mss_now; ++ struct rate_sample *rate; ++}; ++ ++struct tcp_timewait_sock { ++ struct inet_timewait_sock tw_sk; ++ u32 tw_rcv_wnd; ++ u32 tw_ts_offset; ++ u32 tw_ts_recent; ++ u32 tw_last_oow_ack_time; ++ int tw_ts_recent_stamp; ++ u32 tw_tx_delay; ++ struct tcp_md5sig_key *tw_md5_key; ++}; ++ ++enum tcp_tw_status { ++ TCP_TW_SUCCESS = 0, ++ TCP_TW_RST = 1, ++ TCP_TW_ACK = 2, ++ TCP_TW_SYN = 3, ++}; ++ ++struct tcp_plb_state { ++ u8 consec_cong_rounds: 5; ++ u8 unused: 3; ++ u32 pause_until; ++}; ++ ++struct cmsghdr { ++ __kernel_size_t cmsg_len; ++ int cmsg_level; ++ int cmsg_type; ++}; ++ ++struct ip_mreqn { ++ struct in_addr imr_multiaddr; ++ struct in_addr imr_address; ++ int imr_ifindex; ++}; ++ ++enum rt_scope_t { ++ RT_SCOPE_UNIVERSE = 0, ++ RT_SCOPE_SITE = 200, ++ RT_SCOPE_LINK = 253, ++ RT_SCOPE_HOST = 254, ++ RT_SCOPE_NOWHERE = 255, ++}; ++ ++struct sockcm_cookie { ++ u64 transmit_time; ++ u32 mark; ++ u32 tsflags; ++}; ++ ++struct ip_sf_socklist { ++ unsigned int sl_max; ++ unsigned int sl_count; ++ struct callback_head rcu; ++ __be32 sl_addr[0]; ++}; ++ ++struct ip_mc_socklist { ++ struct ip_mc_socklist *next_rcu; ++ struct ip_mreqn multi; ++ unsigned int sfmode; ++ struct ip_sf_socklist *sflist; ++ struct callback_head rcu; ++}; ++ ++struct ip_sf_list { ++ struct ip_sf_list *sf_next; ++ long unsigned int sf_count[2]; ++ __be32 sf_inaddr; ++ unsigned char sf_gsresp; ++ unsigned char sf_oldin; ++ unsigned char sf_crcount; ++}; ++ ++struct ip_mc_list { ++ struct in_device *interface; ++ __be32 multiaddr; ++ unsigned int sfmode; ++ struct ip_sf_list *sources; ++ struct ip_sf_list *tomb; ++ long unsigned int sfcount[2]; ++ union { ++ struct ip_mc_list *next; ++ struct ip_mc_list *next_rcu; ++ }; ++ struct ip_mc_list *next_hash; ++ struct timer_list timer; ++ int users; ++ refcount_t refcnt; ++ spinlock_t lock; ++ char tm_running; ++ char reporter; ++ char unsolicit_count; ++ char loaded; ++ unsigned char gsquery; ++ unsigned char crcount; ++ struct callback_head rcu; ++}; ++ ++struct icmphdr { ++ __u8 type; ++ __u8 code; ++ __sum16 checksum; ++ union { ++ struct { ++ __be16 id; ++ __be16 sequence; ++ } echo; ++ __be32 gateway; ++ struct { ++ __be16 __unused; ++ __be16 mtu; ++ } frag; ++ __u8 reserved[4]; ++ } un; ++}; ++ ++struct ip_options_data { ++ struct ip_options_rcu opt; ++ char data[40]; ++}; ++ ++struct udp_sock { ++ struct inet_sock inet; ++ int pending; ++ unsigned int corkflag; ++ __u8 encap_type; ++ unsigned char no_check6_tx: 1; ++ unsigned char no_check6_rx: 1; ++ unsigned char encap_enabled: 1; ++ unsigned char gro_enabled: 1; ++ unsigned char accept_udp_l4: 1; ++ unsigned char accept_udp_fraglist: 1; ++ __u16 len; ++ __u16 gso_size; ++ __u16 pcslen; ++ __u16 pcrlen; ++ __u8 pcflag; ++ __u8 unused[3]; ++ int (*encap_rcv)(struct sock *, struct sk_buff *); ++ void (*encap_err_rcv)(struct sock *, struct sk_buff *, int, __be16, u32, u8 *); ++ int (*encap_err_lookup)(struct sock *, struct sk_buff *); ++ void (*encap_destroy)(struct sock *); ++ struct sk_buff * (*gro_receive)(struct sock *, struct list_head *, struct sk_buff *); ++ int (*gro_complete)(struct sock *, struct sk_buff *, int); ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct sk_buff_head reader_queue; ++ int forward_deficit; ++ int forward_threshold; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct ipcm_cookie { ++ struct sockcm_cookie sockc; ++ __be32 addr; ++ int oif; ++ struct ip_options_rcu *opt; ++ __u8 protocol; ++ __u8 ttl; ++ __s16 tos; ++ char priority; ++ __u16 gso_size; ++}; ++ ++struct icmp_err { ++ int errno; ++ unsigned int fatal: 1; ++}; ++ ++struct ip_tunnel_encap_ops { ++ size_t (*encap_hlen)(struct ip_tunnel_encap *); ++ int (*build_header)(struct sk_buff *, struct ip_tunnel_encap *, u8 *, struct flowi4 *); ++ int (*err_handler)(struct sk_buff *, u32); ++}; ++ ++struct udp_skb_cb { ++ union { ++ struct inet_skb_parm h4; ++ struct inet6_skb_parm h6; ++ } header; ++ __u16 cscov; ++ __u8 partial_cov; ++}; ++ ++struct udp_dev_scratch { ++ u32 _tsize_state; ++ u16 len; ++ bool is_linear; ++ bool csum_unnecessary; ++}; ++ ++struct udp_seq_afinfo { ++ sa_family_t family; ++ struct udp_table *udp_table; ++}; ++ ++struct udp_iter_state { ++ struct seq_net_private p; ++ int bucket; ++ struct udp_seq_afinfo *bpf_seq_afinfo; ++}; ++ ++struct bpf_iter__udp { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct udp_sock *udp_sk; ++ }; ++ uid_t uid; ++ long: 0; ++ int bucket; ++}; ++ ++struct ifaddrmsg { ++ __u8 ifa_family; ++ __u8 ifa_prefixlen; ++ __u8 ifa_flags; ++ __u8 ifa_scope; ++ __u32 ifa_index; ++}; ++ ++enum { ++ IFA_UNSPEC = 0, ++ IFA_ADDRESS = 1, ++ IFA_LOCAL = 2, ++ IFA_LABEL = 3, ++ IFA_BROADCAST = 4, ++ IFA_ANYCAST = 5, ++ IFA_CACHEINFO = 6, ++ IFA_MULTICAST = 7, ++ IFA_FLAGS = 8, ++ IFA_RT_PRIORITY = 9, ++ IFA_TARGET_NETNSID = 10, ++ IFA_PROTO = 11, ++ __IFA_MAX = 12, ++}; ++ ++struct ifa_cacheinfo { ++ __u32 ifa_prefered; ++ __u32 ifa_valid; ++ __u32 cstamp; ++ __u32 tstamp; ++}; ++ ++enum { ++ IFLA_INET_UNSPEC = 0, ++ IFLA_INET_CONF = 1, ++ __IFLA_INET_MAX = 2, ++}; ++ ++struct in_validator_info { ++ __be32 ivi_addr; ++ struct in_device *ivi_dev; ++ struct netlink_ext_ack *extack; ++}; ++ ++struct netconfmsg { ++ __u8 ncm_family; ++}; ++ ++enum { ++ NETCONFA_UNSPEC = 0, ++ NETCONFA_IFINDEX = 1, ++ NETCONFA_FORWARDING = 2, ++ NETCONFA_RP_FILTER = 3, ++ NETCONFA_MC_FORWARDING = 4, ++ NETCONFA_PROXY_NEIGH = 5, ++ NETCONFA_IGNORE_ROUTES_WITH_LINKDOWN = 6, ++ NETCONFA_INPUT = 7, ++ NETCONFA_BC_FORWARDING = 8, ++ __NETCONFA_MAX = 9, ++}; ++ ++struct fib_result { ++ __be32 prefix; ++ unsigned char prefixlen; ++ unsigned char nh_sel; ++ unsigned char type; ++ unsigned char scope; ++ u32 tclassid; ++ struct fib_nh_common *nhc; ++ struct fib_info *fi; ++ struct fib_table *table; ++ struct hlist_head *fa_head; ++}; ++ ++struct inet_fill_args { ++ u32 portid; ++ u32 seq; ++ int event; ++ unsigned int flags; ++ int netnsid; ++ int ifindex; ++}; ++ ++struct devinet_sysctl_table { ++ struct ctl_table_header *sysctl_header; ++ struct ctl_table devinet_vars[34]; ++}; ++ ++struct flow_dissector_key_control { ++ u16 thoff; ++ u16 addr_type; ++ u32 flags; ++}; ++ ++struct flow_dissector_key_basic { ++ __be16 n_proto; ++ u8 ip_proto; ++ u8 padding; ++}; ++ ++struct flow_dissector_key_tags { ++ u32 flow_label; ++}; ++ ++struct flow_dissector_key_vlan { ++ union { ++ struct { ++ u16 vlan_id: 12; ++ u16 vlan_dei: 1; ++ u16 vlan_priority: 3; ++ }; ++ __be16 vlan_tci; ++ }; ++ __be16 vlan_tpid; ++ __be16 vlan_eth_type; ++ u16 padding; ++}; ++ ++struct flow_dissector_key_keyid { ++ __be32 keyid; ++}; ++ ++struct flow_dissector_key_ipv4_addrs { ++ __be32 src; ++ __be32 dst; ++}; ++ ++struct flow_dissector_key_ipv6_addrs { ++ struct in6_addr src; ++ struct in6_addr dst; ++}; ++ ++struct flow_dissector_key_tipc { ++ __be32 key; ++}; ++ ++struct flow_dissector_key_addrs { ++ union { ++ struct flow_dissector_key_ipv4_addrs v4addrs; ++ struct flow_dissector_key_ipv6_addrs v6addrs; ++ struct flow_dissector_key_tipc tipckey; ++ }; ++}; ++ ++struct flow_dissector_key_ports { ++ union { ++ __be32 ports; ++ struct { ++ __be16 src; ++ __be16 dst; ++ }; ++ }; ++}; ++ ++struct flow_dissector_key_icmp { ++ struct { ++ u8 type; ++ u8 code; ++ }; ++ u16 id; ++}; ++ ++struct flow_keys { ++ struct flow_dissector_key_control control; ++ struct flow_dissector_key_basic basic; ++ struct flow_dissector_key_tags tags; ++ struct flow_dissector_key_vlan vlan; ++ struct flow_dissector_key_vlan cvlan; ++ struct flow_dissector_key_keyid keyid; ++ struct flow_dissector_key_ports ports; ++ struct flow_dissector_key_icmp icmp; ++ struct flow_dissector_key_addrs addrs; ++ long: 0; ++}; ++ ++struct netdev_notifier_info_ext { ++ struct netdev_notifier_info info; ++ union { ++ u32 mtu; ++ } ext; ++}; ++ ++struct rtmsg { ++ unsigned char rtm_family; ++ unsigned char rtm_dst_len; ++ unsigned char rtm_src_len; ++ unsigned char rtm_tos; ++ unsigned char rtm_table; ++ unsigned char rtm_protocol; ++ unsigned char rtm_scope; ++ unsigned char rtm_type; ++ unsigned int rtm_flags; ++}; ++ ++struct rtnexthop { ++ short unsigned int rtnh_len; ++ unsigned char rtnh_flags; ++ unsigned char rtnh_hops; ++ int rtnh_ifindex; ++}; ++ ++struct rtvia { ++ __kernel_sa_family_t rtvia_family; ++ __u8 rtvia_addr[0]; ++}; ++ ++typedef u8 dscp_t; ++ ++struct fib_config { ++ u8 fc_dst_len; ++ dscp_t fc_dscp; ++ u8 fc_protocol; ++ u8 fc_scope; ++ u8 fc_type; ++ u8 fc_gw_family; ++ u32 fc_table; ++ __be32 fc_dst; ++ union { ++ __be32 fc_gw4; ++ struct in6_addr fc_gw6; ++ }; ++ int fc_oif; ++ u32 fc_flags; ++ u32 fc_priority; ++ __be32 fc_prefsrc; ++ u32 fc_nh_id; ++ struct nlattr *fc_mx; ++ struct rtnexthop *fc_mp; ++ int fc_mx_len; ++ int fc_mp_len; ++ u32 fc_flow; ++ u32 fc_nlflags; ++ struct nl_info fc_nlinfo; ++ struct nlattr *fc_encap; ++ u16 fc_encap_type; ++}; ++ ++struct fib_result_nl { ++ __be32 fl_addr; ++ u32 fl_mark; ++ unsigned char fl_tos; ++ unsigned char fl_scope; ++ unsigned char tb_id_in; ++ unsigned char tb_id; ++ unsigned char prefixlen; ++ unsigned char nh_sel; ++ unsigned char type; ++ unsigned char scope; ++ int err; ++}; ++ ++struct fib_dump_filter { ++ u32 table_id; ++ bool filter_set; ++ bool dump_routes; ++ bool dump_exceptions; ++ unsigned char protocol; ++ unsigned char rt_type; ++ unsigned int flags; ++ struct net_device *dev; ++}; ++ ++struct rtentry { ++ long unsigned int rt_pad1; ++ struct sockaddr rt_dst; ++ struct sockaddr rt_gateway; ++ struct sockaddr rt_genmask; ++ short unsigned int rt_flags; ++ short int rt_pad2; ++ long unsigned int rt_pad3; ++ void *rt_pad4; ++ short int rt_metric; ++ char *rt_dev; ++ long unsigned int rt_mtu; ++ long unsigned int rt_window; ++ short unsigned int rt_irtt; ++}; ++ ++struct icmpv6_echo { ++ __be16 identifier; ++ __be16 sequence; ++}; ++ ++struct icmpv6_nd_advt { ++ __u32 router: 1; ++ __u32 solicited: 1; ++ __u32 override: 1; ++ __u32 reserved: 29; ++}; ++ ++struct icmpv6_nd_ra { ++ __u8 hop_limit; ++ __u8 managed: 1; ++ __u8 other: 1; ++ __u8 home_agent: 1; ++ __u8 router_pref: 2; ++ __u8 reserved: 3; ++ __be16 rt_lifetime; ++}; ++ ++struct icmp6hdr { ++ __u8 icmp6_type; ++ __u8 icmp6_code; ++ __sum16 icmp6_cksum; ++ union { ++ __be32 un_data32[1]; ++ __be16 un_data16[2]; ++ __u8 un_data8[4]; ++ struct icmpv6_echo u_echo; ++ struct icmpv6_nd_advt u_nd_advt; ++ struct icmpv6_nd_ra u_nd_ra; ++ } icmp6_dataun; ++}; ++ ++struct pingv6_ops { ++ int (*ipv6_recv_error)(struct sock *, struct msghdr *, int, int *); ++ void (*ip6_datagram_recv_common_ctl)(struct sock *, struct msghdr *, struct sk_buff *); ++ void (*ip6_datagram_recv_specific_ctl)(struct sock *, struct msghdr *, struct sk_buff *); ++ int (*icmpv6_err_convert)(u8, u8, int *); ++ void (*ipv6_icmp_error)(struct sock *, struct sk_buff *, int, __be16, u32, u8 *); ++ int (*ipv6_chk_addr)(struct net *, const struct in6_addr *, const struct net_device *, int); ++}; ++ ++struct ping_iter_state { ++ struct seq_net_private p; ++ int bucket; ++ sa_family_t family; ++}; ++ ++struct pingfakehdr { ++ struct icmphdr icmph; ++ struct msghdr *msg; ++ sa_family_t family; ++ __wsum wcheck; ++}; ++ ++struct ping_table { ++ struct hlist_nulls_head hash[64]; ++ spinlock_t lock; ++}; ++ ++struct snmp_mib { ++ const char *name; ++ int entry; ++}; ++ ++struct inet_diag_sockid { ++ __be16 idiag_sport; ++ __be16 idiag_dport; ++ __be32 idiag_src[4]; ++ __be32 idiag_dst[4]; ++ __u32 idiag_if; ++ __u32 idiag_cookie[2]; ++}; ++ ++struct inet_diag_req_v2 { ++ __u8 sdiag_family; ++ __u8 sdiag_protocol; ++ __u8 idiag_ext; ++ __u8 pad; ++ __u32 idiag_states; ++ struct inet_diag_sockid id; ++}; ++ ++enum { ++ INET_DIAG_REQ_NONE = 0, ++ INET_DIAG_REQ_BYTECODE = 1, ++ INET_DIAG_REQ_SK_BPF_STORAGES = 2, ++ INET_DIAG_REQ_PROTOCOL = 3, ++ __INET_DIAG_REQ_MAX = 4, ++}; ++ ++struct inet_diag_msg { ++ __u8 idiag_family; ++ __u8 idiag_state; ++ __u8 idiag_timer; ++ __u8 idiag_retrans; ++ struct inet_diag_sockid id; ++ __u32 idiag_expires; ++ __u32 idiag_rqueue; ++ __u32 idiag_wqueue; ++ __u32 idiag_uid; ++ __u32 idiag_inode; ++}; ++ ++struct inet_diag_handler { ++ void (*dump)(struct sk_buff *, struct netlink_callback *, const struct inet_diag_req_v2 *); ++ int (*dump_one)(struct netlink_callback *, const struct inet_diag_req_v2 *); ++ void (*idiag_get_info)(struct sock *, struct inet_diag_msg *, void *); ++ int (*idiag_get_aux)(struct sock *, bool, struct sk_buff *); ++ size_t (*idiag_get_aux_size)(struct sock *, bool); ++ int (*destroy)(struct sk_buff *, const struct inet_diag_req_v2 *); ++ __u16 idiag_type; ++ __u16 idiag_info_size; ++}; ++ ++struct bpf_sk_storage_diag; ++ ++struct inet_diag_dump_data { ++ struct nlattr *req_nlas[4]; ++ struct bpf_sk_storage_diag *bpf_stg_diag; ++}; ++ ++enum __sk_action { ++ __SK_DROP = 0, ++ __SK_PASS = 1, ++ __SK_REDIRECT = 2, ++ __SK_NONE = 3, ++}; ++ ++enum sk_psock_state_bits { ++ SK_PSOCK_TX_ENABLED = 0, ++ SK_PSOCK_RX_STRP_ENABLED = 1, ++}; ++ ++struct tls_crypto_info { ++ __u16 version; ++ __u16 cipher_type; ++}; ++ ++struct tls12_crypto_info_aes_gcm_128 { ++ struct tls_crypto_info info; ++ unsigned char iv[8]; ++ unsigned char key[16]; ++ unsigned char salt[4]; ++ unsigned char rec_seq[8]; ++}; ++ ++struct tls12_crypto_info_aes_gcm_256 { ++ struct tls_crypto_info info; ++ unsigned char iv[8]; ++ unsigned char key[32]; ++ unsigned char salt[4]; ++ unsigned char rec_seq[8]; ++}; ++ ++struct tls12_crypto_info_chacha20_poly1305 { ++ struct tls_crypto_info info; ++ unsigned char iv[12]; ++ unsigned char key[32]; ++ unsigned char salt[0]; ++ unsigned char rec_seq[8]; ++}; ++ ++struct tls12_crypto_info_sm4_gcm { ++ struct tls_crypto_info info; ++ unsigned char iv[8]; ++ unsigned char key[16]; ++ unsigned char salt[4]; ++ unsigned char rec_seq[8]; ++}; ++ ++struct tls12_crypto_info_sm4_ccm { ++ struct tls_crypto_info info; ++ unsigned char iv[8]; ++ unsigned char key[16]; ++ unsigned char salt[4]; ++ unsigned char rec_seq[8]; ++}; ++ ++struct tx_work { ++ struct delayed_work work; ++ struct sock *sk; ++}; ++ ++struct tls_rec; ++ ++struct tls_sw_context_tx { ++ struct crypto_aead *aead_send; ++ struct crypto_wait async_wait; ++ struct tx_work tx_work; ++ struct tls_rec *open_rec; ++ struct list_head tx_list; ++ atomic_t encrypt_pending; ++ spinlock_t encrypt_compl_lock; ++ int async_notify; ++ u8 async_capable: 1; ++ long unsigned int tx_bitmask; ++}; ++ ++struct tls_prot_info { ++ u16 version; ++ u16 cipher_type; ++ u16 prepend_size; ++ u16 tag_size; ++ u16 overhead_size; ++ u16 iv_size; ++ u16 salt_size; ++ u16 rec_seq_size; ++ u16 aad_size; ++ u16 tail_size; ++}; ++ ++struct cipher_context { ++ char *iv; ++ char *rec_seq; ++}; ++ ++union tls_crypto_context { ++ struct tls_crypto_info info; ++ union { ++ struct tls12_crypto_info_aes_gcm_128 aes_gcm_128; ++ struct tls12_crypto_info_aes_gcm_256 aes_gcm_256; ++ struct tls12_crypto_info_chacha20_poly1305 chacha20_poly1305; ++ struct tls12_crypto_info_sm4_gcm sm4_gcm; ++ struct tls12_crypto_info_sm4_ccm sm4_ccm; ++ }; ++}; ++ ++struct tls_context { ++ struct tls_prot_info prot_info; ++ u8 tx_conf: 3; ++ u8 rx_conf: 3; ++ u8 zerocopy_sendfile: 1; ++ u8 rx_no_pad: 1; ++ int (*push_pending_record)(struct sock *, int); ++ void (*sk_write_space)(struct sock *); ++ void *priv_ctx_tx; ++ void *priv_ctx_rx; ++ struct net_device *netdev; ++ struct cipher_context tx; ++ struct cipher_context rx; ++ struct scatterlist *partially_sent_record; ++ u16 partially_sent_offset; ++ bool in_tcp_sendpages; ++ bool pending_open_record_frags; ++ struct mutex tx_lock; ++ long unsigned int flags; ++ struct proto *sk_proto; ++ struct sock *sk; ++ void (*sk_destruct)(struct sock *); ++ union tls_crypto_context crypto_send; ++ union tls_crypto_context crypto_recv; ++ struct list_head list; ++ refcount_t refcount; ++ struct callback_head rcu; ++}; ++ ++enum { ++ TCP_BPF_IPV4 = 0, ++ TCP_BPF_IPV6 = 1, ++ TCP_BPF_NUM_PROTS = 2, ++}; ++ ++enum { ++ TCP_BPF_BASE = 0, ++ TCP_BPF_TX = 1, ++ TCP_BPF_RX = 2, ++ TCP_BPF_TXRX = 3, ++ TCP_BPF_NUM_CFGS = 4, ++}; ++ ++struct ip_tunnel; ++ ++struct ip6_tnl; ++ ++struct xfrm_tunnel_skb_cb { ++ union { ++ struct inet_skb_parm h4; ++ struct inet6_skb_parm h6; ++ } header; ++ union { ++ struct ip_tunnel *ip4; ++ struct ip6_tnl *ip6; ++ } tunnel; ++}; ++ ++struct xfrm_mode_skb_cb { ++ struct xfrm_tunnel_skb_cb header; ++ __be16 id; ++ __be16 frag_off; ++ u8 ihl; ++ u8 tos; ++ u8 ttl; ++ u8 protocol; ++ u8 optlen; ++ u8 flow_lbl[3]; ++}; ++ ++struct xfrm_spi_skb_cb { ++ struct xfrm_tunnel_skb_cb header; ++ unsigned int daddroff; ++ unsigned int family; ++ __be32 seq; ++}; ++ ++enum nf_nat_manip_type; ++ ++struct nf_conn; ++ ++struct nf_nat_hook { ++ int (*parse_nat_setup)(struct nf_conn *, enum nf_nat_manip_type, const struct nlattr *); ++ void (*decode_session)(struct sk_buff *, struct flowi *); ++ unsigned int (*manip_pkt)(struct sk_buff *, struct nf_conn *, enum nf_nat_manip_type, enum ip_conntrack_dir); ++ void (*remove_nat_bysrc)(struct nf_conn *); ++}; ++ ++enum { ++ XFRM_LOOKUP_ICMP = 1, ++ XFRM_LOOKUP_QUEUE = 2, ++ XFRM_LOOKUP_KEEP_DST_REF = 4, ++}; ++ ++enum { ++ XFRM_STATE_VOID = 0, ++ XFRM_STATE_ACQ = 1, ++ XFRM_STATE_VALID = 2, ++ XFRM_STATE_ERROR = 3, ++ XFRM_STATE_EXPIRED = 4, ++ XFRM_STATE_DEAD = 5, ++}; ++ ++struct xfrm_if_decode_session_result { ++ struct net *net; ++ u32 if_id; ++}; ++ ++struct xfrm_if_cb { ++ bool (*decode_session)(struct sk_buff *, short unsigned int, struct xfrm_if_decode_session_result *); ++}; ++ ++struct xfrm_policy_afinfo { ++ struct dst_ops *dst_ops; ++ struct dst_entry * (*dst_lookup)(struct net *, int, int, const xfrm_address_t *, const xfrm_address_t *, u32); ++ int (*get_saddr)(struct net *, int, xfrm_address_t *, xfrm_address_t *, u32); ++ int (*fill_dst)(struct xfrm_dst *, struct net_device *, const struct flowi *); ++ struct dst_entry * (*blackhole_route)(struct net *, struct dst_entry *); ++}; ++ ++struct xfrm_state_afinfo { ++ u8 family; ++ u8 proto; ++ const struct xfrm_type_offload *type_offload_esp; ++ const struct xfrm_type *type_esp; ++ const struct xfrm_type *type_ipip; ++ const struct xfrm_type *type_ipip6; ++ const struct xfrm_type *type_comp; ++ const struct xfrm_type *type_ah; ++ const struct xfrm_type *type_routing; ++ const struct xfrm_type *type_dstopts; ++ int (*output)(struct net *, struct sock *, struct sk_buff *); ++ int (*transport_finish)(struct sk_buff *, int); ++ void (*local_error)(struct sk_buff *, u32); ++}; ++ ++struct xfrm_policy_walk { ++ struct xfrm_policy_walk_entry walk; ++ u8 type; ++ u32 seq; ++}; ++ ++struct xfrm_kmaddress { ++ xfrm_address_t local; ++ xfrm_address_t remote; ++ u32 reserved; ++ u16 family; ++}; ++ ++struct xfrm_migrate { ++ xfrm_address_t old_daddr; ++ xfrm_address_t old_saddr; ++ xfrm_address_t new_daddr; ++ xfrm_address_t new_saddr; ++ u8 proto; ++ u8 mode; ++ u16 reserved; ++ u32 reqid; ++ u16 old_family; ++ u16 new_family; ++}; ++ ++struct xfrmk_spdinfo { ++ u32 incnt; ++ u32 outcnt; ++ u32 fwdcnt; ++ u32 inscnt; ++ u32 outscnt; ++ u32 fwdscnt; ++ u32 spdhcnt; ++ u32 spdhmcnt; ++}; ++ ++struct ip6_mh { ++ __u8 ip6mh_proto; ++ __u8 ip6mh_hdrlen; ++ __u8 ip6mh_type; ++ __u8 ip6mh_reserved; ++ __u16 ip6mh_cksum; ++ __u8 data[0]; ++}; ++ ++struct xfrm_flo { ++ struct dst_entry *dst_orig; ++ u8 flags; ++}; ++ ++struct xfrm_pol_inexact_node { ++ struct rb_node node; ++ union { ++ xfrm_address_t addr; ++ struct callback_head rcu; ++ }; ++ u8 prefixlen; ++ struct rb_root root; ++ struct hlist_head hhead; ++}; ++ ++struct xfrm_pol_inexact_key { ++ possible_net_t net; ++ u32 if_id; ++ u16 family; ++ u8 dir; ++ u8 type; ++}; ++ ++struct xfrm_pol_inexact_bin { ++ struct xfrm_pol_inexact_key k; ++ struct rhash_head head; ++ struct hlist_head hhead; ++ seqcount_spinlock_t count; ++ struct rb_root root_d; ++ struct rb_root root_s; ++ struct list_head inexact_bins; ++ struct callback_head rcu; ++}; ++ ++enum xfrm_pol_inexact_candidate_type { ++ XFRM_POL_CAND_BOTH = 0, ++ XFRM_POL_CAND_SADDR = 1, ++ XFRM_POL_CAND_DADDR = 2, ++ XFRM_POL_CAND_ANY = 3, ++ XFRM_POL_CAND_MAX = 4, ++}; ++ ++struct xfrm_pol_inexact_candidates { ++ struct hlist_head *res[4]; ++}; ++ ++struct ucred { ++ __u32 pid; ++ __u32 uid; ++ __u32 gid; ++}; ++ ++struct sockaddr_un { ++ __kernel_sa_family_t sun_family; ++ char sun_path[108]; ++}; ++ ++struct scm_fp_list { ++ short int count; ++ short int max; ++ struct user_struct *user; ++ struct file *fp[253]; ++}; ++ ++struct scm_cookie { ++ struct pid *pid; ++ struct scm_fp_list *fp; ++ struct scm_creds creds; ++ u32 secid; ++}; ++ ++struct unix_address { ++ refcount_t refcnt; ++ int len; ++ struct sockaddr_un name[0]; ++}; ++ ++struct unix_skb_parms { ++ struct pid *pid; ++ kuid_t uid; ++ kgid_t gid; ++ struct scm_fp_list *fp; ++ u32 secid; ++ u32 consumed; ++}; ++ ++struct scm_stat { ++ atomic_t nr_fds; ++}; ++ ++struct unix_sock { ++ struct sock sk; ++ struct unix_address *addr; ++ struct path path; ++ struct mutex iolock; ++ struct mutex bindlock; ++ struct sock *peer; ++ struct list_head link; ++ atomic_long_t inflight; ++ spinlock_t lock; ++ long unsigned int gc_flags; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct socket_wq peer_wq; ++ wait_queue_entry_t peer_wake; ++ struct scm_stat scm_stat; ++ struct sk_buff *oob_skb; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct unix_stream_read_state { ++ int (*recv_actor)(struct sk_buff *, int, int, struct unix_stream_read_state *); ++ struct socket *socket; ++ struct msghdr *msg; ++ struct pipe_inode_info *pipe; ++ size_t size; ++ int flags; ++ unsigned int splice_flags; ++}; ++ ++struct bpf_unix_iter_state { ++ struct seq_net_private p; ++ unsigned int cur_sk; ++ unsigned int end_sk; ++ unsigned int max_sk; ++ struct sock **batch; ++ bool st_bucket_done; ++}; ++ ++struct bpf_iter__unix { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct unix_sock *unix_sk; ++ }; ++ uid_t uid; ++}; ++ ++struct sock_diag_handler { ++ __u8 family; ++ int (*dump)(struct sk_buff *, struct nlmsghdr *); ++ int (*get_info)(struct sk_buff *, struct sock *); ++ int (*destroy)(struct sk_buff *, struct nlmsghdr *); ++}; ++ ++struct unix_diag_req { ++ __u8 sdiag_family; ++ __u8 sdiag_protocol; ++ __u16 pad; ++ __u32 udiag_states; ++ __u32 udiag_ino; ++ __u32 udiag_show; ++ __u32 udiag_cookie[2]; ++}; ++ ++struct unix_diag_msg { ++ __u8 udiag_family; ++ __u8 udiag_type; ++ __u8 udiag_state; ++ __u8 pad; ++ __u32 udiag_ino; ++ __u32 udiag_cookie[2]; ++}; ++ ++enum { ++ UNIX_DIAG_NAME = 0, ++ UNIX_DIAG_VFS = 1, ++ UNIX_DIAG_PEER = 2, ++ UNIX_DIAG_ICONS = 3, ++ UNIX_DIAG_RQLEN = 4, ++ UNIX_DIAG_MEMINFO = 5, ++ UNIX_DIAG_SHUTDOWN = 6, ++ UNIX_DIAG_UID = 7, ++ __UNIX_DIAG_MAX = 8, ++}; ++ ++struct unix_diag_vfs { ++ __u32 udiag_vfs_ino; ++ __u32 udiag_vfs_dev; ++}; ++ ++struct unix_diag_rqlen { ++ __u32 udiag_rqueue; ++ __u32 udiag_wqueue; ++}; ++ ++struct seg6_pernet_data { ++ struct mutex lock; ++ struct in6_addr *tun_src; ++ struct rhashtable hmac_infos; ++}; ++ ++struct ioam6_pernet_data { ++ struct mutex lock; ++ struct rhashtable namespaces; ++ struct rhashtable schemas; ++}; ++ ++struct ipv6_params { ++ __s32 disable_ipv6; ++ __s32 autoconf; ++}; ++ ++enum flowlabel_reflect { ++ FLOWLABEL_REFLECT_ESTABLISHED = 1, ++ FLOWLABEL_REFLECT_TCP_RESET = 2, ++ FLOWLABEL_REFLECT_ICMPV6_ECHO_REPLIES = 4, ++}; ++ ++struct ipv6_bpf_stub { ++ int (*inet6_bind)(struct sock *, struct sockaddr *, int, u32); ++ struct sock * (*udp6_lib_lookup)(struct net *, const struct in6_addr *, __be16, const struct in6_addr *, __be16, int, int, struct udp_table *, struct sk_buff *); ++ int (*ipv6_setsockopt)(struct sock *, int, int, sockptr_t, unsigned int); ++ int (*ipv6_getsockopt)(struct sock *, int, int, sockptr_t, sockptr_t); ++}; ++ ++struct inet_protosw { ++ struct list_head list; ++ short unsigned int type; ++ short unsigned int protocol; ++ struct proto *prot; ++ const struct proto_ops *ops; ++ unsigned char flags; ++}; ++ ++struct in6_rtmsg { ++ struct in6_addr rtmsg_dst; ++ struct in6_addr rtmsg_src; ++ struct in6_addr rtmsg_gateway; ++ __u32 rtmsg_type; ++ __u16 rtmsg_dst_len; ++ __u16 rtmsg_src_len; ++ __u32 rtmsg_metric; ++ long unsigned int rtmsg_info; ++ __u32 rtmsg_flags; ++ int rtmsg_ifindex; ++}; ++ ++struct compat_in6_rtmsg { ++ struct in6_addr rtmsg_dst; ++ struct in6_addr rtmsg_src; ++ struct in6_addr rtmsg_gateway; ++ u32 rtmsg_type; ++ u16 rtmsg_dst_len; ++ u16 rtmsg_src_len; ++ u32 rtmsg_metric; ++ u32 rtmsg_info; ++ u32 rtmsg_flags; ++ s32 rtmsg_ifindex; ++}; ++ ++struct ifaddrlblmsg { ++ __u8 ifal_family; ++ __u8 __ifal_reserved; ++ __u8 ifal_prefixlen; ++ __u8 ifal_flags; ++ __u32 ifal_index; ++ __u32 ifal_seq; ++}; ++ ++enum { ++ IFAL_ADDRESS = 1, ++ IFAL_LABEL = 2, ++ __IFAL_MAX = 3, ++}; ++ ++struct ip6addrlbl_entry { ++ struct in6_addr prefix; ++ int prefixlen; ++ int ifindex; ++ int addrtype; ++ u32 label; ++ struct hlist_node list; ++ struct callback_head rcu; ++}; ++ ++struct ip6addrlbl_init_table { ++ const struct in6_addr *prefix; ++ int prefixlen; ++ u32 label; ++}; ++ ++struct uncached_list { ++ spinlock_t lock; ++ struct list_head head; ++ struct list_head quarantine; ++}; ++ ++struct inet6_ifaddr { ++ struct in6_addr addr; ++ __u32 prefix_len; ++ __u32 rt_priority; ++ __u32 valid_lft; ++ __u32 prefered_lft; ++ refcount_t refcnt; ++ spinlock_t lock; ++ int state; ++ __u32 flags; ++ __u8 dad_probes; ++ __u8 stable_privacy_retry; ++ __u16 scope; ++ __u64 dad_nonce; ++ long unsigned int cstamp; ++ long unsigned int tstamp; ++ struct delayed_work dad_work; ++ struct inet6_dev *idev; ++ struct fib6_info *rt; ++ struct hlist_node addr_lst; ++ struct list_head if_list; ++ struct list_head if_list_aux; ++ struct list_head tmp_list; ++ struct inet6_ifaddr *ifpub; ++ int regen_count; ++ bool tokenized; ++ u8 ifa_proto; ++ struct callback_head rcu; ++ struct in6_addr peer_addr; ++}; ++ ++struct fib6_gc_args { ++ int timeout; ++ int more; ++}; ++ ++struct rt6_exception { ++ struct hlist_node hlist; ++ struct rt6_info *rt6i; ++ long unsigned int stamp; ++ struct callback_head rcu; ++}; ++ ++typedef struct rt6_info * (*pol_lookup_t)(struct net *, struct fib6_table *, struct flowi6 *, const struct sk_buff *, int); ++ ++struct rd_msg { ++ struct icmp6hdr icmph; ++ struct in6_addr target; ++ struct in6_addr dest; ++ __u8 opt[0]; ++}; ++ ++struct route_info { ++ __u8 type; ++ __u8 length; ++ __u8 prefix_len; ++ __u8 reserved_h: 3; ++ __u8 route_pref: 2; ++ __u8 reserved_l: 3; ++ __be32 lifetime; ++ __u8 prefix[0]; ++}; ++ ++struct rt6_rtnl_dump_arg { ++ struct sk_buff *skb; ++ struct netlink_callback *cb; ++ struct net *net; ++ struct fib_dump_filter filter; ++}; ++ ++struct netevent_redirect { ++ struct dst_entry *old; ++ struct dst_entry *new; ++ struct neighbour *neigh; ++ const void *daddr; ++}; ++ ++enum netevent_notif_type { ++ NETEVENT_NEIGH_UPDATE = 1, ++ NETEVENT_REDIRECT = 2, ++ NETEVENT_DELAY_PROBE_TIME_UPDATE = 3, ++ NETEVENT_IPV4_MPATH_HASH_UPDATE = 4, ++ NETEVENT_IPV6_MPATH_HASH_UPDATE = 5, ++ NETEVENT_IPV4_FWD_UPDATE_PRIORITY_UPDATE = 6, ++}; ++ ++struct trace_event_raw_fib6_table_lookup { ++ struct trace_entry ent; ++ u32 tb_id; ++ int err; ++ int oif; ++ int iif; ++ __u8 tos; ++ __u8 scope; ++ __u8 flags; ++ __u8 src[16]; ++ __u8 dst[16]; ++ u16 sport; ++ u16 dport; ++ u8 proto; ++ u8 rt_type; ++ char name[16]; ++ __u8 gw[16]; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_fib6_table_lookup {}; ++ ++typedef void (*btf_trace_fib6_table_lookup)(void *, const struct net *, const struct fib6_result *, struct fib6_table *, const struct flowi6 *); ++ ++enum rt6_nud_state { ++ RT6_NUD_FAIL_HARD = -3, ++ RT6_NUD_FAIL_PROBE = -2, ++ RT6_NUD_FAIL_DO_RR = -1, ++ RT6_NUD_SUCCEED = 1, ++}; ++ ++struct fib6_nh_dm_arg { ++ struct net *net; ++ const struct in6_addr *saddr; ++ int oif; ++ int flags; ++ struct fib6_nh *nh; ++}; ++ ++struct __rt6_probe_work { ++ struct work_struct work; ++ struct in6_addr target; ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++}; ++ ++struct fib6_nh_frl_arg { ++ u32 flags; ++ int oif; ++ int strict; ++ int *mpri; ++ bool *do_rr; ++ struct fib6_nh *nh; ++}; ++ ++struct fib6_nh_excptn_arg { ++ struct rt6_info *rt; ++ int plen; ++}; ++ ++struct fib6_nh_match_arg { ++ const struct net_device *dev; ++ const struct in6_addr *gw; ++ struct fib6_nh *match; ++}; ++ ++struct fib6_nh_age_excptn_arg { ++ struct fib6_gc_args *gc_args; ++ long unsigned int now; ++}; ++ ++struct fib6_nh_rd_arg { ++ struct fib6_result *res; ++ struct flowi6 *fl6; ++ const struct in6_addr *gw; ++ struct rt6_info **ret; ++}; ++ ++struct ip6rd_flowi { ++ struct flowi6 fl6; ++ struct in6_addr gateway; ++}; ++ ++struct fib6_nh_del_cached_rt_arg { ++ struct fib6_config *cfg; ++ struct fib6_info *f6i; ++}; ++ ++struct arg_dev_net_ip { ++ struct net_device *dev; ++ struct net *net; ++ struct in6_addr *addr; ++}; ++ ++struct arg_netdev_event { ++ const struct net_device *dev; ++ union { ++ unsigned char nh_flags; ++ long unsigned int event; ++ }; ++}; ++ ++struct rt6_mtu_change_arg { ++ struct net_device *dev; ++ unsigned int mtu; ++ struct fib6_info *f6i; ++}; ++ ++struct rt6_nh { ++ struct fib6_info *fib6_info; ++ struct fib6_config r_cfg; ++ struct list_head next; ++}; ++ ++struct fib6_nh_exception_dump_walker { ++ struct rt6_rtnl_dump_arg *dump; ++ struct fib6_info *rt; ++ unsigned int flags; ++ unsigned int skip; ++ unsigned int count; ++}; ++ ++struct inet6_protocol { ++ int (*handler)(struct sk_buff *); ++ int (*err_handler)(struct sk_buff *, struct inet6_skb_parm *, u8, u8, int, __be32); ++ unsigned int flags; ++}; ++ ++struct group_source_req { ++ __u32 gsr_interface; ++ struct __kernel_sockaddr_storage gsr_group; ++ struct __kernel_sockaddr_storage gsr_source; ++}; ++ ++struct group_filter { ++ union { ++ struct { ++ __u32 gf_interface_aux; ++ struct __kernel_sockaddr_storage gf_group_aux; ++ __u32 gf_fmode_aux; ++ __u32 gf_numsrc_aux; ++ struct __kernel_sockaddr_storage gf_slist[1]; ++ }; ++ struct { ++ __u32 gf_interface; ++ struct __kernel_sockaddr_storage gf_group; ++ __u32 gf_fmode; ++ __u32 gf_numsrc; ++ struct __kernel_sockaddr_storage gf_slist_flex[0]; ++ }; ++ }; ++}; ++ ++struct mld_msg { ++ struct icmp6hdr mld_hdr; ++ struct in6_addr mld_mca; ++}; ++ ++struct mld2_grec { ++ __u8 grec_type; ++ __u8 grec_auxwords; ++ __be16 grec_nsrcs; ++ struct in6_addr grec_mca; ++ struct in6_addr grec_src[0]; ++}; ++ ++struct mld2_report { ++ struct icmp6hdr mld2r_hdr; ++ struct mld2_grec mld2r_grec[0]; ++}; ++ ++struct mld2_query { ++ struct icmp6hdr mld2q_hdr; ++ struct in6_addr mld2q_mca; ++ __u8 mld2q_resv2: 4; ++ __u8 mld2q_suppress: 1; ++ __u8 mld2q_qrv: 3; ++ __u8 mld2q_qqic; ++ __be16 mld2q_nsrcs; ++ struct in6_addr mld2q_srcs[0]; ++}; ++ ++struct igmp6_mc_iter_state { ++ struct seq_net_private p; ++ struct net_device *dev; ++ struct inet6_dev *idev; ++}; ++ ++struct igmp6_mcf_iter_state { ++ struct seq_net_private p; ++ struct net_device *dev; ++ struct inet6_dev *idev; ++ struct ifmcaddr6 *im; ++}; ++ ++struct rt0_hdr { ++ struct ipv6_rt_hdr rt_hdr; ++ __u32 reserved; ++ struct in6_addr addr[0]; ++}; ++ ++struct ipv6_destopt_hao { ++ __u8 type; ++ __u8 length; ++ struct in6_addr addr; ++} __attribute__((packed)); ++ ++struct ipv6_sr_hdr { ++ __u8 nexthdr; ++ __u8 hdrlen; ++ __u8 type; ++ __u8 segments_left; ++ __u8 first_segment; ++ __u8 flags; ++ __u16 tag; ++ struct in6_addr segments[0]; ++}; ++ ++struct ipv6_rpl_sr_hdr { ++ __u8 nexthdr; ++ __u8 hdrlen; ++ __u8 type; ++ __u8 segments_left; ++ __u32 cmpri: 4; ++ __u32 cmpre: 4; ++ __u32 pad: 4; ++ __u32 reserved: 20; ++ union { ++ struct { ++ struct {} __empty_addr; ++ struct in6_addr addr[0]; ++ }; ++ struct { ++ struct {} __empty_data; ++ __u8 data[0]; ++ }; ++ } segments; ++}; ++ ++struct ioam6_hdr { ++ __u8 opt_type; ++ __u8 opt_len; ++ char: 8; ++ __u8 type; ++}; ++ ++struct ioam6_trace_hdr { ++ __be16 namespace_id; ++ __u8 nodelen: 5; ++ __u8 overflow: 1; ++ char: 2; ++ char: 1; ++ __u8 remlen: 7; ++ union { ++ __be32 type_be32; ++ struct { ++ __u32 bit0: 1; ++ __u32 bit1: 1; ++ __u32 bit2: 1; ++ __u32 bit3: 1; ++ __u32 bit4: 1; ++ __u32 bit5: 1; ++ __u32 bit6: 1; ++ __u32 bit7: 1; ++ __u32 bit8: 1; ++ __u32 bit9: 1; ++ __u32 bit10: 1; ++ __u32 bit11: 1; ++ __u32 bit12: 1; ++ __u32 bit13: 1; ++ __u32 bit14: 1; ++ __u32 bit15: 1; ++ __u32 bit16: 1; ++ __u32 bit17: 1; ++ __u32 bit18: 1; ++ __u32 bit19: 1; ++ __u32 bit20: 1; ++ __u32 bit21: 1; ++ __u32 bit22: 1; ++ __u32 bit23: 1; ++ } type; ++ }; ++ __u8 data[0]; ++}; ++ ++struct ioam6_schema; ++ ++struct ioam6_namespace { ++ struct rhash_head head; ++ struct callback_head rcu; ++ struct ioam6_schema *schema; ++ __be16 id; ++ __be32 data; ++ __be64 data_wide; ++}; ++ ++struct ioam6_schema { ++ struct rhash_head head; ++ struct callback_head rcu; ++ struct ioam6_namespace *ns; ++ u32 id; ++ int len; ++ __be32 hdr; ++ u8 data[0]; ++}; ++ ++struct in6_flowlabel_req { ++ struct in6_addr flr_dst; ++ __be32 flr_label; ++ __u8 flr_action; ++ __u8 flr_share; ++ __u16 flr_flags; ++ __u16 flr_expires; ++ __u16 flr_linger; ++ __u32 __flr_pad; ++}; ++ ++struct ipcm6_cookie { ++ struct sockcm_cookie sockc; ++ __s16 hlimit; ++ __s16 tclass; ++ __u16 gso_size; ++ __s8 dontfrag; ++ struct ipv6_txoptions *opt; ++}; ++ ++struct ip6fl_iter_state { ++ struct seq_net_private p; ++ struct pid_namespace *pid_ns; ++ int bucket; ++}; ++ ++struct rhltable { ++ struct rhashtable ht; ++}; ++ ++struct icmp6_filter { ++ __u32 data[8]; ++}; ++ ++struct nlmsgerr { ++ int error; ++ struct nlmsghdr msg; ++}; ++ ++struct raw6_sock { ++ struct inet_sock inet; ++ __u32 checksum; ++ __u32 offset; ++ struct icmp6_filter filter; ++ __u32 ip6mr_table; ++ struct ipv6_pinfo inet6; ++}; ++ ++enum { ++ PIM_TYPE_HELLO = 0, ++ PIM_TYPE_REGISTER = 1, ++ PIM_TYPE_REGISTER_STOP = 2, ++ PIM_TYPE_JOIN_PRUNE = 3, ++ PIM_TYPE_BOOTSTRAP = 4, ++ PIM_TYPE_ASSERT = 5, ++ PIM_TYPE_GRAFT = 6, ++ PIM_TYPE_GRAFT_ACK = 7, ++ PIM_TYPE_CANDIDATE_RP_ADV = 8, ++}; ++ ++struct pimreghdr { ++ __u8 type; ++ __u8 reserved; ++ __be16 csum; ++ __be32 flags; ++}; ++ ++typedef short unsigned int mifi_t; ++ ++typedef __u32 if_mask; ++ ++struct if_set { ++ if_mask ifs_bits[8]; ++}; ++ ++struct mif6ctl { ++ mifi_t mif6c_mifi; ++ unsigned char mif6c_flags; ++ unsigned char vifc_threshold; ++ __u16 mif6c_pifi; ++ unsigned int vifc_rate_limit; ++}; ++ ++struct mf6cctl { ++ struct sockaddr_in6 mf6cc_origin; ++ struct sockaddr_in6 mf6cc_mcastgrp; ++ mifi_t mf6cc_parent; ++ struct if_set mf6cc_ifset; ++}; ++ ++struct sioc_sg_req6 { ++ struct sockaddr_in6 src; ++ struct sockaddr_in6 grp; ++ long unsigned int pktcnt; ++ long unsigned int bytecnt; ++ long unsigned int wrong_if; ++}; ++ ++struct sioc_mif_req6 { ++ mifi_t mifi; ++ long unsigned int icount; ++ long unsigned int ocount; ++ long unsigned int ibytes; ++ long unsigned int obytes; ++}; ++ ++struct mrt6msg { ++ __u8 im6_mbz; ++ __u8 im6_msgtype; ++ __u16 im6_mif; ++ __u32 im6_pad; ++ struct in6_addr im6_src; ++ struct in6_addr im6_dst; ++}; ++ ++enum { ++ IP6MRA_CREPORT_UNSPEC = 0, ++ IP6MRA_CREPORT_MSGTYPE = 1, ++ IP6MRA_CREPORT_MIF_ID = 2, ++ IP6MRA_CREPORT_SRC_ADDR = 3, ++ IP6MRA_CREPORT_DST_ADDR = 4, ++ IP6MRA_CREPORT_PKT = 5, ++ __IP6MRA_CREPORT_MAX = 6, ++}; ++ ++struct vif_device { ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ long unsigned int bytes_in; ++ long unsigned int bytes_out; ++ long unsigned int pkt_in; ++ long unsigned int pkt_out; ++ long unsigned int rate_limit; ++ unsigned char threshold; ++ short unsigned int flags; ++ int link; ++ struct netdev_phys_item_id dev_parent_id; ++ __be32 local; ++ __be32 remote; ++}; ++ ++struct vif_entry_notifier_info { ++ struct fib_notifier_info info; ++ struct net_device *dev; ++ short unsigned int vif_index; ++ short unsigned int vif_flags; ++ u32 tb_id; ++}; ++ ++enum { ++ MFC_STATIC = 1, ++ MFC_OFFLOAD = 2, ++}; ++ ++struct mr_mfc { ++ struct rhlist_head mnode; ++ short unsigned int mfc_parent; ++ int mfc_flags; ++ union { ++ struct { ++ long unsigned int expires; ++ struct sk_buff_head unresolved; ++ } unres; ++ struct { ++ long unsigned int last_assert; ++ int minvif; ++ int maxvif; ++ long unsigned int bytes; ++ long unsigned int pkt; ++ long unsigned int wrong_if; ++ long unsigned int lastuse; ++ unsigned char ttls[32]; ++ refcount_t refcount; ++ } res; ++ } mfc_un; ++ struct list_head list; ++ struct callback_head rcu; ++ void (*free)(struct callback_head *); ++}; ++ ++struct mfc_entry_notifier_info { ++ struct fib_notifier_info info; ++ struct mr_mfc *mfc; ++ u32 tb_id; ++}; ++ ++struct mr_table_ops { ++ const struct rhashtable_params *rht_params; ++ void *cmparg_any; ++}; ++ ++struct mr_table { ++ struct list_head list; ++ possible_net_t net; ++ struct mr_table_ops ops; ++ u32 id; ++ struct sock *mroute_sk; ++ struct timer_list ipmr_expire_timer; ++ struct list_head mfc_unres_queue; ++ struct vif_device vif_table[32]; ++ struct rhltable mfc_hash; ++ struct list_head mfc_cache_list; ++ int maxvif; ++ atomic_t cache_resolve_queue_len; ++ bool mroute_do_assert; ++ bool mroute_do_pim; ++ bool mroute_do_wrvifwhole; ++ int mroute_reg_vif_num; ++}; ++ ++struct mr_vif_iter { ++ struct seq_net_private p; ++ struct mr_table *mrt; ++ int ct; ++}; ++ ++struct mr_mfc_iter { ++ struct seq_net_private p; ++ struct mr_table *mrt; ++ struct list_head *cache; ++ spinlock_t *lock; ++}; ++ ++struct mfc6_cache_cmp_arg { ++ struct in6_addr mf6c_mcastgrp; ++ struct in6_addr mf6c_origin; ++}; ++ ++struct mfc6_cache { ++ struct mr_mfc _c; ++ union { ++ struct { ++ struct in6_addr mf6c_mcastgrp; ++ struct in6_addr mf6c_origin; ++ }; ++ struct mfc6_cache_cmp_arg cmparg; ++ }; ++}; ++ ++struct ip6mr_result { ++ struct mr_table *mrt; ++}; ++ ++struct compat_sioc_sg_req6 { ++ struct sockaddr_in6 src; ++ struct sockaddr_in6 grp; ++ compat_ulong_t pktcnt; ++ compat_ulong_t bytecnt; ++ compat_ulong_t wrong_if; ++}; ++ ++struct compat_sioc_mif_req6 { ++ mifi_t mifi; ++ compat_ulong_t icount; ++ compat_ulong_t ocount; ++ compat_ulong_t ibytes; ++ compat_ulong_t obytes; ++}; ++ ++typedef u32 u_int32_t; ++ ++struct netpoll; ++ ++struct netpoll_info { ++ refcount_t refcnt; ++ struct semaphore dev_lock; ++ struct sk_buff_head txq; ++ struct delayed_work tx_work; ++ struct netpoll *netpoll; ++ struct callback_head rcu; ++}; ++ ++struct ip6_fraglist_iter { ++ struct ipv6hdr *tmp_hdr; ++ struct sk_buff *frag; ++ int offset; ++ unsigned int hlen; ++ __be32 frag_id; ++ u8 nexthdr; ++}; ++ ++struct ip6_frag_state { ++ u8 *prevhdr; ++ unsigned int hlen; ++ unsigned int mtu; ++ unsigned int left; ++ int offset; ++ int ptr; ++ int hroom; ++ int troom; ++ __be32 frag_id; ++ u8 nexthdr; ++}; ++ ++union inet_addr { ++ __u32 all[4]; ++ __be32 ip; ++ __be32 ip6[4]; ++ struct in_addr in; ++ struct in6_addr in6; ++}; ++ ++struct netpoll { ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ char dev_name[16]; ++ const char *name; ++ union inet_addr local_ip; ++ union inet_addr remote_ip; ++ bool ipv6; ++ u16 local_port; ++ u16 remote_port; ++ u8 remote_mac[6]; ++}; ++ ++struct ip6_rt_info { ++ struct in6_addr daddr; ++ struct in6_addr saddr; ++ u_int32_t mark; ++}; ++ ++struct nf_queue_entry; ++ ++struct nf_ipv6_ops { ++ void (*route_input)(struct sk_buff *); ++ int (*fragment)(struct net *, struct sock *, struct sk_buff *, int (*)(struct net *, struct sock *, struct sk_buff *)); ++ int (*reroute)(struct sk_buff *, const struct nf_queue_entry *); ++}; ++ ++struct nf_queue_entry { ++ struct list_head list; ++ struct sk_buff *skb; ++ unsigned int id; ++ unsigned int hook_index; ++ struct net_device *physin; ++ struct net_device *physout; ++ struct nf_hook_state state; ++ u16 size; ++}; ++ ++struct br_input_skb_cb { ++ struct net_device *brdev; ++ u16 frag_max_size; ++ u8 igmp; ++ u8 mrouters_only: 1; ++ u8 proxyarp_replied: 1; ++ u8 src_port_isolated: 1; ++ u8 vlan_filtered: 1; ++ u8 br_netfilter_broute: 1; ++ u8 tx_fwd_offload: 1; ++ int src_hwdom; ++ long unsigned int fwd_hwdoms; ++}; ++ ++struct nf_bridge_frag_data; ++ ++struct netlbl_audit { ++ u32 secid; ++ kuid_t loginuid; ++ unsigned int sessionid; ++}; ++ ++struct calipso_doi; ++ ++struct netlbl_calipso_ops { ++ int (*doi_add)(struct calipso_doi *, struct netlbl_audit *); ++ void (*doi_free)(struct calipso_doi *); ++ int (*doi_remove)(u32, struct netlbl_audit *); ++ struct calipso_doi * (*doi_getdef)(u32); ++ void (*doi_putdef)(struct calipso_doi *); ++ int (*doi_walk)(u32 *, int (*)(struct calipso_doi *, void *), void *); ++ int (*sock_getattr)(struct sock *, struct netlbl_lsm_secattr *); ++ int (*sock_setattr)(struct sock *, const struct calipso_doi *, const struct netlbl_lsm_secattr *); ++ void (*sock_delattr)(struct sock *); ++ int (*req_setattr)(struct request_sock *, const struct calipso_doi *, const struct netlbl_lsm_secattr *); ++ void (*req_delattr)(struct request_sock *); ++ int (*opt_getattr)(const unsigned char *, struct netlbl_lsm_secattr *); ++ unsigned char * (*skbuff_optptr)(const struct sk_buff *); ++ int (*skbuff_setattr)(struct sk_buff *, const struct calipso_doi *, const struct netlbl_lsm_secattr *); ++ int (*skbuff_delattr)(struct sk_buff *); ++ void (*cache_invalidate)(); ++ int (*cache_add)(const unsigned char *, const struct netlbl_lsm_secattr *); ++}; ++ ++struct calipso_doi { ++ u32 doi; ++ u32 type; ++ refcount_t refcount; ++ struct list_head list; ++ struct callback_head rcu; ++}; ++ ++struct calipso_map_cache_bkt { ++ spinlock_t lock; ++ u32 size; ++ struct list_head list; ++}; ++ ++struct calipso_map_cache_entry { ++ u32 hash; ++ unsigned char *key; ++ size_t key_len; ++ struct netlbl_lsm_cache *lsm_data; ++ u32 activity; ++ struct list_head list; ++}; ++ ++struct sr6_tlv { ++ __u8 type; ++ __u8 len; ++ __u8 data[0]; ++}; ++ ++struct sr6_tlv_hmac { ++ struct sr6_tlv tlvhdr; ++ __u16 reserved; ++ __be32 hmackeyid; ++ __u8 hmac[32]; ++}; ++ ++enum { ++ SEG6_HMAC_ALGO_SHA1 = 1, ++ SEG6_HMAC_ALGO_SHA256 = 2, ++}; ++ ++struct seg6_hmac_info { ++ struct rhash_head node; ++ struct callback_head rcu; ++ u32 hmackeyid; ++ char secret[64]; ++ u8 slen; ++ u8 alg_id; ++}; ++ ++struct seg6_hmac_algo { ++ u8 alg_id; ++ char name[64]; ++ struct crypto_shash **tfms; ++ struct shash_desc **shashs; ++}; ++ ++enum { ++ IP6_FH_F_FRAG = 1, ++ IP6_FH_F_AUTH = 2, ++ IP6_FH_F_SKIP_RH = 4, ++}; ++ ++struct offload_callbacks { ++ struct sk_buff * (*gso_segment)(struct sk_buff *, netdev_features_t); ++ struct sk_buff * (*gro_receive)(struct list_head *, struct sk_buff *); ++ int (*gro_complete)(struct sk_buff *, int); ++}; ++ ++struct net_offload { ++ struct offload_callbacks callbacks; ++ unsigned int flags; ++}; ++ ++struct flow_keys_basic { ++ struct flow_dissector_key_control control; ++ struct flow_dissector_key_basic basic; ++}; ++ ++struct sockaddr_pkt { ++ short unsigned int spkt_family; ++ unsigned char spkt_device[14]; ++ __be16 spkt_protocol; ++}; ++ ++struct sockaddr_ll { ++ short unsigned int sll_family; ++ __be16 sll_protocol; ++ int sll_ifindex; ++ short unsigned int sll_hatype; ++ unsigned char sll_pkttype; ++ unsigned char sll_halen; ++ unsigned char sll_addr[8]; ++}; ++ ++struct tpacket_stats { ++ unsigned int tp_packets; ++ unsigned int tp_drops; ++}; ++ ++struct tpacket_stats_v3 { ++ unsigned int tp_packets; ++ unsigned int tp_drops; ++ unsigned int tp_freeze_q_cnt; ++}; ++ ++struct tpacket_rollover_stats { ++ __u64 tp_all; ++ __u64 tp_huge; ++ __u64 tp_failed; ++}; ++ ++union tpacket_stats_u { ++ struct tpacket_stats stats1; ++ struct tpacket_stats_v3 stats3; ++}; ++ ++struct tpacket_auxdata { ++ __u32 tp_status; ++ __u32 tp_len; ++ __u32 tp_snaplen; ++ __u16 tp_mac; ++ __u16 tp_net; ++ __u16 tp_vlan_tci; ++ __u16 tp_vlan_tpid; ++}; ++ ++struct tpacket_hdr { ++ long unsigned int tp_status; ++ unsigned int tp_len; ++ unsigned int tp_snaplen; ++ short unsigned int tp_mac; ++ short unsigned int tp_net; ++ unsigned int tp_sec; ++ unsigned int tp_usec; ++}; ++ ++struct tpacket2_hdr { ++ __u32 tp_status; ++ __u32 tp_len; ++ __u32 tp_snaplen; ++ __u16 tp_mac; ++ __u16 tp_net; ++ __u32 tp_sec; ++ __u32 tp_nsec; ++ __u16 tp_vlan_tci; ++ __u16 tp_vlan_tpid; ++ __u8 tp_padding[4]; ++}; ++ ++struct tpacket_hdr_variant1 { ++ __u32 tp_rxhash; ++ __u32 tp_vlan_tci; ++ __u16 tp_vlan_tpid; ++ __u16 tp_padding; ++}; ++ ++struct tpacket3_hdr { ++ __u32 tp_next_offset; ++ __u32 tp_sec; ++ __u32 tp_nsec; ++ __u32 tp_snaplen; ++ __u32 tp_len; ++ __u32 tp_status; ++ __u16 tp_mac; ++ __u16 tp_net; ++ union { ++ struct tpacket_hdr_variant1 hv1; ++ }; ++ __u8 tp_padding[8]; ++}; ++ ++struct tpacket_bd_ts { ++ unsigned int ts_sec; ++ union { ++ unsigned int ts_usec; ++ unsigned int ts_nsec; ++ }; ++}; ++ ++struct tpacket_hdr_v1 { ++ __u32 block_status; ++ __u32 num_pkts; ++ __u32 offset_to_first_pkt; ++ __u32 blk_len; ++ __u64 seq_num; ++ struct tpacket_bd_ts ts_first_pkt; ++ struct tpacket_bd_ts ts_last_pkt; ++}; ++ ++union tpacket_bd_header_u { ++ struct tpacket_hdr_v1 bh1; ++}; ++ ++struct tpacket_block_desc { ++ __u32 version; ++ __u32 offset_to_priv; ++ union tpacket_bd_header_u hdr; ++}; ++ ++enum tpacket_versions { ++ TPACKET_V1 = 0, ++ TPACKET_V2 = 1, ++ TPACKET_V3 = 2, ++}; ++ ++struct tpacket_req { ++ unsigned int tp_block_size; ++ unsigned int tp_block_nr; ++ unsigned int tp_frame_size; ++ unsigned int tp_frame_nr; ++}; ++ ++struct tpacket_req3 { ++ unsigned int tp_block_size; ++ unsigned int tp_block_nr; ++ unsigned int tp_frame_size; ++ unsigned int tp_frame_nr; ++ unsigned int tp_retire_blk_tov; ++ unsigned int tp_sizeof_priv; ++ unsigned int tp_feature_req_word; ++}; ++ ++union tpacket_req_u { ++ struct tpacket_req req; ++ struct tpacket_req3 req3; ++}; ++ ++struct fanout_args { ++ __u16 type_flags; ++ __u16 id; ++ __u32 max_num_members; ++}; ++ ++struct sock_fprog { ++ short unsigned int len; ++ struct sock_filter *filter; ++}; ++ ++enum nf_dev_hooks { ++ NF_NETDEV_INGRESS = 0, ++ NF_NETDEV_EGRESS = 1, ++ NF_NETDEV_NUMHOOKS = 2, ++}; ++ ++typedef int (*bpf_aux_classic_check_t)(struct sock_filter *, unsigned int); ++ ++typedef __u16 __virtio16; ++ ++struct virtio_net_hdr { ++ __u8 flags; ++ __u8 gso_type; ++ __virtio16 hdr_len; ++ __virtio16 gso_size; ++ __virtio16 csum_start; ++ __virtio16 csum_offset; ++}; ++ ++struct virtio_net_hdr_mrg_rxbuf { ++ struct virtio_net_hdr hdr; ++ __virtio16 num_buffers; ++}; ++ ++struct packet_mclist { ++ struct packet_mclist *next; ++ int ifindex; ++ int count; ++ short unsigned int type; ++ short unsigned int alen; ++ unsigned char addr[32]; ++}; ++ ++struct pgv; ++ ++struct tpacket_kbdq_core { ++ struct pgv *pkbdq; ++ unsigned int feature_req_word; ++ unsigned int hdrlen; ++ unsigned char reset_pending_on_curr_blk; ++ unsigned char delete_blk_timer; ++ short unsigned int kactive_blk_num; ++ short unsigned int blk_sizeof_priv; ++ short unsigned int last_kactive_blk_num; ++ char *pkblk_start; ++ char *pkblk_end; ++ int kblk_size; ++ unsigned int max_frame_len; ++ unsigned int knum_blocks; ++ uint64_t knxt_seq_num; ++ char *prev; ++ char *nxt_offset; ++ struct sk_buff *skb; ++ rwlock_t blk_fill_in_prog_lock; ++ short unsigned int retire_blk_tov; ++ short unsigned int version; ++ long unsigned int tov_in_jiffies; ++ struct timer_list retire_blk_timer; ++}; ++ ++struct pgv { ++ char *buffer; ++}; ++ ++struct packet_ring_buffer { ++ struct pgv *pg_vec; ++ unsigned int head; ++ unsigned int frames_per_block; ++ unsigned int frame_size; ++ unsigned int frame_max; ++ unsigned int pg_vec_order; ++ unsigned int pg_vec_pages; ++ unsigned int pg_vec_len; ++ unsigned int *pending_refcnt; ++ union { ++ long unsigned int *rx_owner_map; ++ struct tpacket_kbdq_core prb_bdqc; ++ }; ++}; ++ ++struct packet_fanout { ++ possible_net_t net; ++ unsigned int num_members; ++ u32 max_num_members; ++ u16 id; ++ u8 type; ++ u8 flags; ++ union { ++ atomic_t rr_cur; ++ struct bpf_prog *bpf_prog; ++ }; ++ struct list_head list; ++ spinlock_t lock; ++ refcount_t sk_ref; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct packet_type prot_hook; ++ struct sock *arr[0]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct packet_rollover { ++ int sock; ++ atomic_long_t num; ++ atomic_long_t num_huge; ++ atomic_long_t num_failed; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ u32 history[64]; ++}; ++ ++struct packet_sock { ++ struct sock sk; ++ struct packet_fanout *fanout; ++ union tpacket_stats_u stats; ++ struct packet_ring_buffer rx_ring; ++ struct packet_ring_buffer tx_ring; ++ int copy_thresh; ++ spinlock_t bind_lock; ++ struct mutex pg_vec_lock; ++ long unsigned int flags; ++ int ifindex; ++ u8 vnet_hdr_sz; ++ __be16 num; ++ struct packet_rollover *rollover; ++ struct packet_mclist *mclist; ++ atomic_t mapped; ++ enum tpacket_versions tp_version; ++ unsigned int tp_hdrlen; ++ unsigned int tp_reserve; ++ unsigned int tp_tstamp; ++ struct completion skb_completion; ++ struct net_device *cached_dev; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct packet_type prot_hook; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ atomic_t tp_drops; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum packet_sock_flags { ++ PACKET_SOCK_ORIGDEV = 0, ++ PACKET_SOCK_AUXDATA = 1, ++ PACKET_SOCK_TX_HAS_OFF = 2, ++ PACKET_SOCK_TP_LOSS = 3, ++ PACKET_SOCK_RUNNING = 4, ++ PACKET_SOCK_PRESSURE = 5, ++ PACKET_SOCK_QDISC_BYPASS = 6, ++}; ++ ++struct packet_mreq_max { ++ int mr_ifindex; ++ short unsigned int mr_type; ++ short unsigned int mr_alen; ++ unsigned char mr_address[32]; ++}; ++ ++union tpacket_uhdr { ++ struct tpacket_hdr *h1; ++ struct tpacket2_hdr *h2; ++ struct tpacket3_hdr *h3; ++ void *raw; ++}; ++ ++struct packet_skb_cb { ++ union { ++ struct sockaddr_pkt pkt; ++ union { ++ unsigned int origlen; ++ struct sockaddr_ll ll; ++ }; ++ } sa; ++}; ++ ++enum devlink_command { ++ DEVLINK_CMD_UNSPEC = 0, ++ DEVLINK_CMD_GET = 1, ++ DEVLINK_CMD_SET = 2, ++ DEVLINK_CMD_NEW = 3, ++ DEVLINK_CMD_DEL = 4, ++ DEVLINK_CMD_PORT_GET = 5, ++ DEVLINK_CMD_PORT_SET = 6, ++ DEVLINK_CMD_PORT_NEW = 7, ++ DEVLINK_CMD_PORT_DEL = 8, ++ DEVLINK_CMD_PORT_SPLIT = 9, ++ DEVLINK_CMD_PORT_UNSPLIT = 10, ++ DEVLINK_CMD_SB_GET = 11, ++ DEVLINK_CMD_SB_SET = 12, ++ DEVLINK_CMD_SB_NEW = 13, ++ DEVLINK_CMD_SB_DEL = 14, ++ DEVLINK_CMD_SB_POOL_GET = 15, ++ DEVLINK_CMD_SB_POOL_SET = 16, ++ DEVLINK_CMD_SB_POOL_NEW = 17, ++ DEVLINK_CMD_SB_POOL_DEL = 18, ++ DEVLINK_CMD_SB_PORT_POOL_GET = 19, ++ DEVLINK_CMD_SB_PORT_POOL_SET = 20, ++ DEVLINK_CMD_SB_PORT_POOL_NEW = 21, ++ DEVLINK_CMD_SB_PORT_POOL_DEL = 22, ++ DEVLINK_CMD_SB_TC_POOL_BIND_GET = 23, ++ DEVLINK_CMD_SB_TC_POOL_BIND_SET = 24, ++ DEVLINK_CMD_SB_TC_POOL_BIND_NEW = 25, ++ DEVLINK_CMD_SB_TC_POOL_BIND_DEL = 26, ++ DEVLINK_CMD_SB_OCC_SNAPSHOT = 27, ++ DEVLINK_CMD_SB_OCC_MAX_CLEAR = 28, ++ DEVLINK_CMD_ESWITCH_GET = 29, ++ DEVLINK_CMD_ESWITCH_SET = 30, ++ DEVLINK_CMD_DPIPE_TABLE_GET = 31, ++ DEVLINK_CMD_DPIPE_ENTRIES_GET = 32, ++ DEVLINK_CMD_DPIPE_HEADERS_GET = 33, ++ DEVLINK_CMD_DPIPE_TABLE_COUNTERS_SET = 34, ++ DEVLINK_CMD_RESOURCE_SET = 35, ++ DEVLINK_CMD_RESOURCE_DUMP = 36, ++ DEVLINK_CMD_RELOAD = 37, ++ DEVLINK_CMD_PARAM_GET = 38, ++ DEVLINK_CMD_PARAM_SET = 39, ++ DEVLINK_CMD_PARAM_NEW = 40, ++ DEVLINK_CMD_PARAM_DEL = 41, ++ DEVLINK_CMD_REGION_GET = 42, ++ DEVLINK_CMD_REGION_SET = 43, ++ DEVLINK_CMD_REGION_NEW = 44, ++ DEVLINK_CMD_REGION_DEL = 45, ++ DEVLINK_CMD_REGION_READ = 46, ++ DEVLINK_CMD_PORT_PARAM_GET = 47, ++ DEVLINK_CMD_PORT_PARAM_SET = 48, ++ DEVLINK_CMD_PORT_PARAM_NEW = 49, ++ DEVLINK_CMD_PORT_PARAM_DEL = 50, ++ DEVLINK_CMD_INFO_GET = 51, ++ DEVLINK_CMD_HEALTH_REPORTER_GET = 52, ++ DEVLINK_CMD_HEALTH_REPORTER_SET = 53, ++ DEVLINK_CMD_HEALTH_REPORTER_RECOVER = 54, ++ DEVLINK_CMD_HEALTH_REPORTER_DIAGNOSE = 55, ++ DEVLINK_CMD_HEALTH_REPORTER_DUMP_GET = 56, ++ DEVLINK_CMD_HEALTH_REPORTER_DUMP_CLEAR = 57, ++ DEVLINK_CMD_FLASH_UPDATE = 58, ++ DEVLINK_CMD_FLASH_UPDATE_END = 59, ++ DEVLINK_CMD_FLASH_UPDATE_STATUS = 60, ++ DEVLINK_CMD_TRAP_GET = 61, ++ DEVLINK_CMD_TRAP_SET = 62, ++ DEVLINK_CMD_TRAP_NEW = 63, ++ DEVLINK_CMD_TRAP_DEL = 64, ++ DEVLINK_CMD_TRAP_GROUP_GET = 65, ++ DEVLINK_CMD_TRAP_GROUP_SET = 66, ++ DEVLINK_CMD_TRAP_GROUP_NEW = 67, ++ DEVLINK_CMD_TRAP_GROUP_DEL = 68, ++ DEVLINK_CMD_TRAP_POLICER_GET = 69, ++ DEVLINK_CMD_TRAP_POLICER_SET = 70, ++ DEVLINK_CMD_TRAP_POLICER_NEW = 71, ++ DEVLINK_CMD_TRAP_POLICER_DEL = 72, ++ DEVLINK_CMD_HEALTH_REPORTER_TEST = 73, ++ DEVLINK_CMD_RATE_GET = 74, ++ DEVLINK_CMD_RATE_SET = 75, ++ DEVLINK_CMD_RATE_NEW = 76, ++ DEVLINK_CMD_RATE_DEL = 77, ++ DEVLINK_CMD_LINECARD_GET = 78, ++ DEVLINK_CMD_LINECARD_SET = 79, ++ DEVLINK_CMD_LINECARD_NEW = 80, ++ DEVLINK_CMD_LINECARD_DEL = 81, ++ DEVLINK_CMD_SELFTESTS_GET = 82, ++ DEVLINK_CMD_SELFTESTS_RUN = 83, ++ __DEVLINK_CMD_MAX = 84, ++ DEVLINK_CMD_MAX = 83, ++}; ++ ++enum devlink_eswitch_encap_mode { ++ DEVLINK_ESWITCH_ENCAP_MODE_NONE = 0, ++ DEVLINK_ESWITCH_ENCAP_MODE_BASIC = 1, ++}; ++ ++enum devlink_attr_selftest_id { ++ DEVLINK_ATTR_SELFTEST_ID_UNSPEC = 0, ++ DEVLINK_ATTR_SELFTEST_ID_FLASH = 1, ++ __DEVLINK_ATTR_SELFTEST_ID_MAX = 2, ++ DEVLINK_ATTR_SELFTEST_ID_MAX = 1, ++}; ++ ++enum devlink_selftest_status { ++ DEVLINK_SELFTEST_STATUS_SKIP = 0, ++ DEVLINK_SELFTEST_STATUS_PASS = 1, ++ DEVLINK_SELFTEST_STATUS_FAIL = 2, ++}; ++ ++enum devlink_attr_selftest_result { ++ DEVLINK_ATTR_SELFTEST_RESULT_UNSPEC = 0, ++ DEVLINK_ATTR_SELFTEST_RESULT = 1, ++ DEVLINK_ATTR_SELFTEST_RESULT_ID = 2, ++ DEVLINK_ATTR_SELFTEST_RESULT_STATUS = 3, ++ __DEVLINK_ATTR_SELFTEST_RESULT_MAX = 4, ++ DEVLINK_ATTR_SELFTEST_RESULT_MAX = 3, ++}; ++ ++enum devlink_trap_action { ++ DEVLINK_TRAP_ACTION_DROP = 0, ++ DEVLINK_TRAP_ACTION_TRAP = 1, ++ DEVLINK_TRAP_ACTION_MIRROR = 2, ++}; ++ ++enum devlink_trap_type { ++ DEVLINK_TRAP_TYPE_DROP = 0, ++ DEVLINK_TRAP_TYPE_EXCEPTION = 1, ++ DEVLINK_TRAP_TYPE_CONTROL = 2, ++}; ++ ++enum devlink_reload_action { ++ DEVLINK_RELOAD_ACTION_UNSPEC = 0, ++ DEVLINK_RELOAD_ACTION_DRIVER_REINIT = 1, ++ DEVLINK_RELOAD_ACTION_FW_ACTIVATE = 2, ++ __DEVLINK_RELOAD_ACTION_MAX = 3, ++ DEVLINK_RELOAD_ACTION_MAX = 2, ++}; ++ ++enum devlink_reload_limit { ++ DEVLINK_RELOAD_LIMIT_UNSPEC = 0, ++ DEVLINK_RELOAD_LIMIT_NO_RESET = 1, ++ __DEVLINK_RELOAD_LIMIT_MAX = 2, ++ DEVLINK_RELOAD_LIMIT_MAX = 1, ++}; ++ ++enum devlink_attr { ++ DEVLINK_ATTR_UNSPEC = 0, ++ DEVLINK_ATTR_BUS_NAME = 1, ++ DEVLINK_ATTR_DEV_NAME = 2, ++ DEVLINK_ATTR_PORT_INDEX = 3, ++ DEVLINK_ATTR_PORT_TYPE = 4, ++ DEVLINK_ATTR_PORT_DESIRED_TYPE = 5, ++ DEVLINK_ATTR_PORT_NETDEV_IFINDEX = 6, ++ DEVLINK_ATTR_PORT_NETDEV_NAME = 7, ++ DEVLINK_ATTR_PORT_IBDEV_NAME = 8, ++ DEVLINK_ATTR_PORT_SPLIT_COUNT = 9, ++ DEVLINK_ATTR_PORT_SPLIT_GROUP = 10, ++ DEVLINK_ATTR_SB_INDEX = 11, ++ DEVLINK_ATTR_SB_SIZE = 12, ++ DEVLINK_ATTR_SB_INGRESS_POOL_COUNT = 13, ++ DEVLINK_ATTR_SB_EGRESS_POOL_COUNT = 14, ++ DEVLINK_ATTR_SB_INGRESS_TC_COUNT = 15, ++ DEVLINK_ATTR_SB_EGRESS_TC_COUNT = 16, ++ DEVLINK_ATTR_SB_POOL_INDEX = 17, ++ DEVLINK_ATTR_SB_POOL_TYPE = 18, ++ DEVLINK_ATTR_SB_POOL_SIZE = 19, ++ DEVLINK_ATTR_SB_POOL_THRESHOLD_TYPE = 20, ++ DEVLINK_ATTR_SB_THRESHOLD = 21, ++ DEVLINK_ATTR_SB_TC_INDEX = 22, ++ DEVLINK_ATTR_SB_OCC_CUR = 23, ++ DEVLINK_ATTR_SB_OCC_MAX = 24, ++ DEVLINK_ATTR_ESWITCH_MODE = 25, ++ DEVLINK_ATTR_ESWITCH_INLINE_MODE = 26, ++ DEVLINK_ATTR_DPIPE_TABLES = 27, ++ DEVLINK_ATTR_DPIPE_TABLE = 28, ++ DEVLINK_ATTR_DPIPE_TABLE_NAME = 29, ++ DEVLINK_ATTR_DPIPE_TABLE_SIZE = 30, ++ DEVLINK_ATTR_DPIPE_TABLE_MATCHES = 31, ++ DEVLINK_ATTR_DPIPE_TABLE_ACTIONS = 32, ++ DEVLINK_ATTR_DPIPE_TABLE_COUNTERS_ENABLED = 33, ++ DEVLINK_ATTR_DPIPE_ENTRIES = 34, ++ DEVLINK_ATTR_DPIPE_ENTRY = 35, ++ DEVLINK_ATTR_DPIPE_ENTRY_INDEX = 36, ++ DEVLINK_ATTR_DPIPE_ENTRY_MATCH_VALUES = 37, ++ DEVLINK_ATTR_DPIPE_ENTRY_ACTION_VALUES = 38, ++ DEVLINK_ATTR_DPIPE_ENTRY_COUNTER = 39, ++ DEVLINK_ATTR_DPIPE_MATCH = 40, ++ DEVLINK_ATTR_DPIPE_MATCH_VALUE = 41, ++ DEVLINK_ATTR_DPIPE_MATCH_TYPE = 42, ++ DEVLINK_ATTR_DPIPE_ACTION = 43, ++ DEVLINK_ATTR_DPIPE_ACTION_VALUE = 44, ++ DEVLINK_ATTR_DPIPE_ACTION_TYPE = 45, ++ DEVLINK_ATTR_DPIPE_VALUE = 46, ++ DEVLINK_ATTR_DPIPE_VALUE_MASK = 47, ++ DEVLINK_ATTR_DPIPE_VALUE_MAPPING = 48, ++ DEVLINK_ATTR_DPIPE_HEADERS = 49, ++ DEVLINK_ATTR_DPIPE_HEADER = 50, ++ DEVLINK_ATTR_DPIPE_HEADER_NAME = 51, ++ DEVLINK_ATTR_DPIPE_HEADER_ID = 52, ++ DEVLINK_ATTR_DPIPE_HEADER_FIELDS = 53, ++ DEVLINK_ATTR_DPIPE_HEADER_GLOBAL = 54, ++ DEVLINK_ATTR_DPIPE_HEADER_INDEX = 55, ++ DEVLINK_ATTR_DPIPE_FIELD = 56, ++ DEVLINK_ATTR_DPIPE_FIELD_NAME = 57, ++ DEVLINK_ATTR_DPIPE_FIELD_ID = 58, ++ DEVLINK_ATTR_DPIPE_FIELD_BITWIDTH = 59, ++ DEVLINK_ATTR_DPIPE_FIELD_MAPPING_TYPE = 60, ++ DEVLINK_ATTR_PAD = 61, ++ DEVLINK_ATTR_ESWITCH_ENCAP_MODE = 62, ++ DEVLINK_ATTR_RESOURCE_LIST = 63, ++ DEVLINK_ATTR_RESOURCE = 64, ++ DEVLINK_ATTR_RESOURCE_NAME = 65, ++ DEVLINK_ATTR_RESOURCE_ID = 66, ++ DEVLINK_ATTR_RESOURCE_SIZE = 67, ++ DEVLINK_ATTR_RESOURCE_SIZE_NEW = 68, ++ DEVLINK_ATTR_RESOURCE_SIZE_VALID = 69, ++ DEVLINK_ATTR_RESOURCE_SIZE_MIN = 70, ++ DEVLINK_ATTR_RESOURCE_SIZE_MAX = 71, ++ DEVLINK_ATTR_RESOURCE_SIZE_GRAN = 72, ++ DEVLINK_ATTR_RESOURCE_UNIT = 73, ++ DEVLINK_ATTR_RESOURCE_OCC = 74, ++ DEVLINK_ATTR_DPIPE_TABLE_RESOURCE_ID = 75, ++ DEVLINK_ATTR_DPIPE_TABLE_RESOURCE_UNITS = 76, ++ DEVLINK_ATTR_PORT_FLAVOUR = 77, ++ DEVLINK_ATTR_PORT_NUMBER = 78, ++ DEVLINK_ATTR_PORT_SPLIT_SUBPORT_NUMBER = 79, ++ DEVLINK_ATTR_PARAM = 80, ++ DEVLINK_ATTR_PARAM_NAME = 81, ++ DEVLINK_ATTR_PARAM_GENERIC = 82, ++ DEVLINK_ATTR_PARAM_TYPE = 83, ++ DEVLINK_ATTR_PARAM_VALUES_LIST = 84, ++ DEVLINK_ATTR_PARAM_VALUE = 85, ++ DEVLINK_ATTR_PARAM_VALUE_DATA = 86, ++ DEVLINK_ATTR_PARAM_VALUE_CMODE = 87, ++ DEVLINK_ATTR_REGION_NAME = 88, ++ DEVLINK_ATTR_REGION_SIZE = 89, ++ DEVLINK_ATTR_REGION_SNAPSHOTS = 90, ++ DEVLINK_ATTR_REGION_SNAPSHOT = 91, ++ DEVLINK_ATTR_REGION_SNAPSHOT_ID = 92, ++ DEVLINK_ATTR_REGION_CHUNKS = 93, ++ DEVLINK_ATTR_REGION_CHUNK = 94, ++ DEVLINK_ATTR_REGION_CHUNK_DATA = 95, ++ DEVLINK_ATTR_REGION_CHUNK_ADDR = 96, ++ DEVLINK_ATTR_REGION_CHUNK_LEN = 97, ++ DEVLINK_ATTR_INFO_DRIVER_NAME = 98, ++ DEVLINK_ATTR_INFO_SERIAL_NUMBER = 99, ++ DEVLINK_ATTR_INFO_VERSION_FIXED = 100, ++ DEVLINK_ATTR_INFO_VERSION_RUNNING = 101, ++ DEVLINK_ATTR_INFO_VERSION_STORED = 102, ++ DEVLINK_ATTR_INFO_VERSION_NAME = 103, ++ DEVLINK_ATTR_INFO_VERSION_VALUE = 104, ++ DEVLINK_ATTR_SB_POOL_CELL_SIZE = 105, ++ DEVLINK_ATTR_FMSG = 106, ++ DEVLINK_ATTR_FMSG_OBJ_NEST_START = 107, ++ DEVLINK_ATTR_FMSG_PAIR_NEST_START = 108, ++ DEVLINK_ATTR_FMSG_ARR_NEST_START = 109, ++ DEVLINK_ATTR_FMSG_NEST_END = 110, ++ DEVLINK_ATTR_FMSG_OBJ_NAME = 111, ++ DEVLINK_ATTR_FMSG_OBJ_VALUE_TYPE = 112, ++ DEVLINK_ATTR_FMSG_OBJ_VALUE_DATA = 113, ++ DEVLINK_ATTR_HEALTH_REPORTER = 114, ++ DEVLINK_ATTR_HEALTH_REPORTER_NAME = 115, ++ DEVLINK_ATTR_HEALTH_REPORTER_STATE = 116, ++ DEVLINK_ATTR_HEALTH_REPORTER_ERR_COUNT = 117, ++ DEVLINK_ATTR_HEALTH_REPORTER_RECOVER_COUNT = 118, ++ DEVLINK_ATTR_HEALTH_REPORTER_DUMP_TS = 119, ++ DEVLINK_ATTR_HEALTH_REPORTER_GRACEFUL_PERIOD = 120, ++ DEVLINK_ATTR_HEALTH_REPORTER_AUTO_RECOVER = 121, ++ DEVLINK_ATTR_FLASH_UPDATE_FILE_NAME = 122, ++ DEVLINK_ATTR_FLASH_UPDATE_COMPONENT = 123, ++ DEVLINK_ATTR_FLASH_UPDATE_STATUS_MSG = 124, ++ DEVLINK_ATTR_FLASH_UPDATE_STATUS_DONE = 125, ++ DEVLINK_ATTR_FLASH_UPDATE_STATUS_TOTAL = 126, ++ DEVLINK_ATTR_PORT_PCI_PF_NUMBER = 127, ++ DEVLINK_ATTR_PORT_PCI_VF_NUMBER = 128, ++ DEVLINK_ATTR_STATS = 129, ++ DEVLINK_ATTR_TRAP_NAME = 130, ++ DEVLINK_ATTR_TRAP_ACTION = 131, ++ DEVLINK_ATTR_TRAP_TYPE = 132, ++ DEVLINK_ATTR_TRAP_GENERIC = 133, ++ DEVLINK_ATTR_TRAP_METADATA = 134, ++ DEVLINK_ATTR_TRAP_GROUP_NAME = 135, ++ DEVLINK_ATTR_RELOAD_FAILED = 136, ++ DEVLINK_ATTR_HEALTH_REPORTER_DUMP_TS_NS = 137, ++ DEVLINK_ATTR_NETNS_FD = 138, ++ DEVLINK_ATTR_NETNS_PID = 139, ++ DEVLINK_ATTR_NETNS_ID = 140, ++ DEVLINK_ATTR_HEALTH_REPORTER_AUTO_DUMP = 141, ++ DEVLINK_ATTR_TRAP_POLICER_ID = 142, ++ DEVLINK_ATTR_TRAP_POLICER_RATE = 143, ++ DEVLINK_ATTR_TRAP_POLICER_BURST = 144, ++ DEVLINK_ATTR_PORT_FUNCTION = 145, ++ DEVLINK_ATTR_INFO_BOARD_SERIAL_NUMBER = 146, ++ DEVLINK_ATTR_PORT_LANES = 147, ++ DEVLINK_ATTR_PORT_SPLITTABLE = 148, ++ DEVLINK_ATTR_PORT_EXTERNAL = 149, ++ DEVLINK_ATTR_PORT_CONTROLLER_NUMBER = 150, ++ DEVLINK_ATTR_FLASH_UPDATE_STATUS_TIMEOUT = 151, ++ DEVLINK_ATTR_FLASH_UPDATE_OVERWRITE_MASK = 152, ++ DEVLINK_ATTR_RELOAD_ACTION = 153, ++ DEVLINK_ATTR_RELOAD_ACTIONS_PERFORMED = 154, ++ DEVLINK_ATTR_RELOAD_LIMITS = 155, ++ DEVLINK_ATTR_DEV_STATS = 156, ++ DEVLINK_ATTR_RELOAD_STATS = 157, ++ DEVLINK_ATTR_RELOAD_STATS_ENTRY = 158, ++ DEVLINK_ATTR_RELOAD_STATS_LIMIT = 159, ++ DEVLINK_ATTR_RELOAD_STATS_VALUE = 160, ++ DEVLINK_ATTR_REMOTE_RELOAD_STATS = 161, ++ DEVLINK_ATTR_RELOAD_ACTION_INFO = 162, ++ DEVLINK_ATTR_RELOAD_ACTION_STATS = 163, ++ DEVLINK_ATTR_PORT_PCI_SF_NUMBER = 164, ++ DEVLINK_ATTR_RATE_TYPE = 165, ++ DEVLINK_ATTR_RATE_TX_SHARE = 166, ++ DEVLINK_ATTR_RATE_TX_MAX = 167, ++ DEVLINK_ATTR_RATE_NODE_NAME = 168, ++ DEVLINK_ATTR_RATE_PARENT_NODE_NAME = 169, ++ DEVLINK_ATTR_REGION_MAX_SNAPSHOTS = 170, ++ DEVLINK_ATTR_LINECARD_INDEX = 171, ++ DEVLINK_ATTR_LINECARD_STATE = 172, ++ DEVLINK_ATTR_LINECARD_TYPE = 173, ++ DEVLINK_ATTR_LINECARD_SUPPORTED_TYPES = 174, ++ DEVLINK_ATTR_NESTED_DEVLINK = 175, ++ DEVLINK_ATTR_SELFTESTS = 176, ++ DEVLINK_ATTR_RATE_TX_PRIORITY = 177, ++ DEVLINK_ATTR_RATE_TX_WEIGHT = 178, ++ DEVLINK_ATTR_REGION_DIRECT = 179, ++ __DEVLINK_ATTR_MAX = 180, ++ DEVLINK_ATTR_MAX = 179, ++}; ++ ++enum devlink_dpipe_field_mapping_type { ++ DEVLINK_DPIPE_FIELD_MAPPING_TYPE_NONE = 0, ++ DEVLINK_DPIPE_FIELD_MAPPING_TYPE_IFINDEX = 1, ++}; ++ ++enum devlink_port_fn_state { ++ DEVLINK_PORT_FN_STATE_INACTIVE = 0, ++ DEVLINK_PORT_FN_STATE_ACTIVE = 1, ++}; ++ ++enum devlink_port_fn_opstate { ++ DEVLINK_PORT_FN_OPSTATE_DETACHED = 0, ++ DEVLINK_PORT_FN_OPSTATE_ATTACHED = 1, ++}; ++ ++struct devlink_dev_stats { ++ u32 reload_stats[6]; ++ u32 remote_reload_stats[6]; ++}; ++ ++struct devlink_dpipe_headers; ++ ++struct devlink_ops; ++ ++struct devlink { ++ u32 index; ++ struct xarray ports; ++ struct list_head rate_list; ++ struct list_head sb_list; ++ struct list_head dpipe_table_list; ++ struct list_head resource_list; ++ struct xarray params; ++ struct list_head region_list; ++ struct list_head reporter_list; ++ struct devlink_dpipe_headers *dpipe_headers; ++ struct list_head trap_list; ++ struct list_head trap_group_list; ++ struct list_head trap_policer_list; ++ struct list_head linecard_list; ++ const struct devlink_ops *ops; ++ struct xarray snapshot_ids; ++ struct devlink_dev_stats stats; ++ struct device *dev; ++ possible_net_t _net; ++ struct mutex lock; ++ struct lock_class_key lock_key; ++ u8 reload_failed: 1; ++ refcount_t refcount; ++ struct rcu_work rwork; ++ long: 64; ++ long: 64; ++ long: 64; ++ char priv[0]; ++}; ++ ++struct devlink_port_new_attrs { ++ enum devlink_port_flavour flavour; ++ unsigned int port_index; ++ u32 controller; ++ u32 sfnum; ++ u16 pfnum; ++ u8 port_index_valid: 1; ++ u8 controller_valid: 1; ++ u8 sfnum_valid: 1; ++}; ++ ++struct devlink_dpipe_field { ++ const char *name; ++ unsigned int id; ++ unsigned int bitwidth; ++ enum devlink_dpipe_field_mapping_type mapping_type; ++}; ++ ++struct devlink_dpipe_header { ++ const char *name; ++ unsigned int id; ++ struct devlink_dpipe_field *fields; ++ unsigned int fields_count; ++ bool global; ++}; ++ ++struct devlink_dpipe_headers { ++ struct devlink_dpipe_header **headers; ++ unsigned int headers_count; ++}; ++ ++struct devlink_flash_notify { ++ const char *status_msg; ++ const char *component; ++ long unsigned int done; ++ long unsigned int total; ++ long unsigned int timeout; ++}; ++ ++struct devlink_flash_update_params { ++ const struct firmware *fw; ++ const char *component; ++ u32 overwrite_mask; ++}; ++ ++struct devlink_trap_policer { ++ u32 id; ++ u64 init_rate; ++ u64 init_burst; ++ u64 max_rate; ++ u64 min_rate; ++ u64 max_burst; ++ u64 min_burst; ++}; ++ ++struct devlink_trap_group { ++ const char *name; ++ u16 id; ++ bool generic; ++ u32 init_policer_id; ++}; ++ ++struct devlink_trap { ++ enum devlink_trap_type type; ++ enum devlink_trap_action init_action; ++ bool generic; ++ u16 id; ++ const char *name; ++ u16 init_group_id; ++ u32 metadata_cap; ++}; ++ ++struct devlink_ops { ++ u32 supported_flash_update_params; ++ long unsigned int reload_actions; ++ long unsigned int reload_limits; ++ int (*reload_down)(struct devlink *, bool, enum devlink_reload_action, enum devlink_reload_limit, struct netlink_ext_ack *); ++ int (*reload_up)(struct devlink *, enum devlink_reload_action, enum devlink_reload_limit, u32 *, struct netlink_ext_ack *); ++ int (*port_type_set)(struct devlink_port *, enum devlink_port_type); ++ int (*port_split)(struct devlink *, struct devlink_port *, unsigned int, struct netlink_ext_ack *); ++ int (*port_unsplit)(struct devlink *, struct devlink_port *, struct netlink_ext_ack *); ++ int (*sb_pool_get)(struct devlink *, unsigned int, u16, struct devlink_sb_pool_info *); ++ int (*sb_pool_set)(struct devlink *, unsigned int, u16, u32, enum devlink_sb_threshold_type, struct netlink_ext_ack *); ++ int (*sb_port_pool_get)(struct devlink_port *, unsigned int, u16, u32 *); ++ int (*sb_port_pool_set)(struct devlink_port *, unsigned int, u16, u32, struct netlink_ext_ack *); ++ int (*sb_tc_pool_bind_get)(struct devlink_port *, unsigned int, u16, enum devlink_sb_pool_type, u16 *, u32 *); ++ int (*sb_tc_pool_bind_set)(struct devlink_port *, unsigned int, u16, enum devlink_sb_pool_type, u16, u32, struct netlink_ext_ack *); ++ int (*sb_occ_snapshot)(struct devlink *, unsigned int); ++ int (*sb_occ_max_clear)(struct devlink *, unsigned int); ++ int (*sb_occ_port_pool_get)(struct devlink_port *, unsigned int, u16, u32 *, u32 *); ++ int (*sb_occ_tc_port_bind_get)(struct devlink_port *, unsigned int, u16, enum devlink_sb_pool_type, u32 *, u32 *); ++ int (*eswitch_mode_get)(struct devlink *, u16 *); ++ int (*eswitch_mode_set)(struct devlink *, u16, struct netlink_ext_ack *); ++ int (*eswitch_inline_mode_get)(struct devlink *, u8 *); ++ int (*eswitch_inline_mode_set)(struct devlink *, u8, struct netlink_ext_ack *); ++ int (*eswitch_encap_mode_get)(struct devlink *, enum devlink_eswitch_encap_mode *); ++ int (*eswitch_encap_mode_set)(struct devlink *, enum devlink_eswitch_encap_mode, struct netlink_ext_ack *); ++ int (*info_get)(struct devlink *, struct devlink_info_req *, struct netlink_ext_ack *); ++ int (*flash_update)(struct devlink *, struct devlink_flash_update_params *, struct netlink_ext_ack *); ++ int (*trap_init)(struct devlink *, const struct devlink_trap *, void *); ++ void (*trap_fini)(struct devlink *, const struct devlink_trap *, void *); ++ int (*trap_action_set)(struct devlink *, const struct devlink_trap *, enum devlink_trap_action, struct netlink_ext_ack *); ++ int (*trap_group_init)(struct devlink *, const struct devlink_trap_group *); ++ int (*trap_group_set)(struct devlink *, const struct devlink_trap_group *, const struct devlink_trap_policer *, struct netlink_ext_ack *); ++ int (*trap_group_action_set)(struct devlink *, const struct devlink_trap_group *, enum devlink_trap_action, struct netlink_ext_ack *); ++ int (*trap_drop_counter_get)(struct devlink *, const struct devlink_trap *, u64 *); ++ int (*trap_policer_init)(struct devlink *, const struct devlink_trap_policer *); ++ void (*trap_policer_fini)(struct devlink *, const struct devlink_trap_policer *); ++ int (*trap_policer_set)(struct devlink *, const struct devlink_trap_policer *, u64, u64, struct netlink_ext_ack *); ++ int (*trap_policer_counter_get)(struct devlink *, const struct devlink_trap_policer *, u64 *); ++ int (*port_function_hw_addr_get)(struct devlink_port *, u8 *, int *, struct netlink_ext_ack *); ++ int (*port_function_hw_addr_set)(struct devlink_port *, const u8 *, int, struct netlink_ext_ack *); ++ int (*port_fn_roce_get)(struct devlink_port *, bool *, struct netlink_ext_ack *); ++ int (*port_fn_roce_set)(struct devlink_port *, bool, struct netlink_ext_ack *); ++ int (*port_fn_migratable_get)(struct devlink_port *, bool *, struct netlink_ext_ack *); ++ int (*port_fn_migratable_set)(struct devlink_port *, bool, struct netlink_ext_ack *); ++ int (*port_new)(struct devlink *, const struct devlink_port_new_attrs *, struct netlink_ext_ack *, unsigned int *); ++ int (*port_del)(struct devlink *, unsigned int, struct netlink_ext_ack *); ++ int (*port_fn_state_get)(struct devlink_port *, enum devlink_port_fn_state *, enum devlink_port_fn_opstate *, struct netlink_ext_ack *); ++ int (*port_fn_state_set)(struct devlink_port *, enum devlink_port_fn_state, struct netlink_ext_ack *); ++ int (*rate_leaf_tx_share_set)(struct devlink_rate *, void *, u64, struct netlink_ext_ack *); ++ int (*rate_leaf_tx_max_set)(struct devlink_rate *, void *, u64, struct netlink_ext_ack *); ++ int (*rate_leaf_tx_priority_set)(struct devlink_rate *, void *, u32, struct netlink_ext_ack *); ++ int (*rate_leaf_tx_weight_set)(struct devlink_rate *, void *, u32, struct netlink_ext_ack *); ++ int (*rate_node_tx_share_set)(struct devlink_rate *, void *, u64, struct netlink_ext_ack *); ++ int (*rate_node_tx_max_set)(struct devlink_rate *, void *, u64, struct netlink_ext_ack *); ++ int (*rate_node_tx_priority_set)(struct devlink_rate *, void *, u32, struct netlink_ext_ack *); ++ int (*rate_node_tx_weight_set)(struct devlink_rate *, void *, u32, struct netlink_ext_ack *); ++ int (*rate_node_new)(struct devlink_rate *, void **, struct netlink_ext_ack *); ++ int (*rate_node_del)(struct devlink_rate *, void *, struct netlink_ext_ack *); ++ int (*rate_leaf_parent_set)(struct devlink_rate *, struct devlink_rate *, void *, void *, struct netlink_ext_ack *); ++ int (*rate_node_parent_set)(struct devlink_rate *, struct devlink_rate *, void *, void *, struct netlink_ext_ack *); ++ bool (*selftest_check)(struct devlink *, unsigned int, struct netlink_ext_ack *); ++ enum devlink_selftest_status (*selftest_run)(struct devlink *, unsigned int, struct netlink_ext_ack *); ++}; ++ ++enum devlink_info_version_type { ++ DEVLINK_INFO_VERSION_TYPE_NONE = 0, ++ DEVLINK_INFO_VERSION_TYPE_COMPONENT = 1, ++}; ++ ++struct devlink_info_req { ++ struct sk_buff *msg; ++ void (*version_cb)(const char *, enum devlink_info_version_type, void *); ++ void *version_cb_priv; ++}; ++ ++enum devlink_multicast_groups { ++ DEVLINK_MCGRP_CONFIG = 0, ++}; ++ ++struct devlink_cmd { ++ int (*dump_one)(struct sk_buff *, struct devlink *, struct netlink_callback *); ++}; ++ ++struct devlink_reload_combination { ++ enum devlink_reload_action action; ++ enum devlink_reload_limit limit; ++}; ++ ++struct devlink_flash_component_lookup_ctx { ++ const char *lookup_name; ++ bool lookup_name_found; ++}; ++ ++struct vlan_group { ++ unsigned int nr_vlan_devs; ++ struct hlist_node hlist; ++ struct net_device **vlan_devices_arrays[16]; ++}; ++ ++struct vlan_info { ++ struct net_device *real_dev; ++ struct vlan_group grp; ++ struct list_head vid_list; ++ unsigned int nr_vids; ++ struct callback_head rcu; ++}; ++ ++struct packet_offload { ++ __be16 type; ++ u16 priority; ++ struct offload_callbacks callbacks; ++ struct list_head list; ++}; ++ ++enum vlan_flags { ++ VLAN_FLAG_REORDER_HDR = 1, ++ VLAN_FLAG_GVRP = 2, ++ VLAN_FLAG_LOOSE_BINDING = 4, ++ VLAN_FLAG_MVRP = 8, ++ VLAN_FLAG_BRIDGE_BINDING = 16, ++}; ++ ++struct vlan_pcpu_stats { ++ u64_stats_t rx_packets; ++ u64_stats_t rx_bytes; ++ u64_stats_t rx_multicast; ++ u64_stats_t tx_packets; ++ u64_stats_t tx_bytes; ++ struct u64_stats_sync syncp; ++ u32 rx_errors; ++ u32 tx_dropped; ++}; ++ ++struct vlan_priority_tci_mapping { ++ u32 priority; ++ u16 vlan_qos; ++ struct vlan_priority_tci_mapping *next; ++}; ++ ++struct vlan_dev_priv { ++ unsigned int nr_ingress_mappings; ++ u32 ingress_priority_map[8]; ++ unsigned int nr_egress_mappings; ++ struct vlan_priority_tci_mapping *egress_priority_map[16]; ++ __be16 vlan_proto; ++ u16 vlan_id; ++ u16 flags; ++ struct net_device *real_dev; ++ netdevice_tracker dev_tracker; ++ unsigned char real_dev_addr[6]; ++ struct proc_dir_entry *dent; ++ struct vlan_pcpu_stats *vlan_pcpu_stats; ++ struct netpoll *netpoll; ++}; ++ ++struct napi_gro_cb { ++ void *frag0; ++ unsigned int frag0_len; ++ int data_offset; ++ u16 flush; ++ u16 flush_id; ++ u16 count; ++ u16 proto; ++ long unsigned int age; ++ union { ++ struct { ++ u16 gro_remcsum_start; ++ u8 same_flow: 1; ++ u8 encap_mark: 1; ++ u8 csum_valid: 1; ++ u8 csum_cnt: 3; ++ u8 free: 2; ++ u8 is_ipv6: 1; ++ u8 is_fou: 1; ++ u8 is_atomic: 1; ++ u8 recursion_counter: 4; ++ u8 is_flist: 1; ++ }; ++ struct { ++ u16 gro_remcsum_start; ++ u8 same_flow: 1; ++ u8 encap_mark: 1; ++ u8 csum_valid: 1; ++ u8 csum_cnt: 3; ++ u8 free: 2; ++ u8 is_ipv6: 1; ++ u8 is_fou: 1; ++ u8 is_atomic: 1; ++ u8 recursion_counter: 4; ++ u8 is_flist: 1; ++ } zeroed; ++ }; ++ __wsum csum; ++ struct sk_buff *last; ++}; ++ ++enum vlan_protos { ++ VLAN_PROTO_8021Q = 0, ++ VLAN_PROTO_8021AD = 1, ++ VLAN_PROTO_NUM = 2, ++}; ++ ++struct vlan_vid_info { ++ struct list_head list; ++ __be16 proto; ++ u16 vid; ++ int refcount; ++}; ++ ++struct netlbl_af4list { ++ __be32 addr; ++ __be32 mask; ++ u32 valid; ++ struct list_head list; ++}; ++ ++struct netlbl_af6list { ++ struct in6_addr addr; ++ struct in6_addr mask; ++ u32 valid; ++ struct list_head list; ++}; ++ ++struct cipso_v4_std_map_tbl { ++ struct { ++ u32 *cipso; ++ u32 *local; ++ u32 cipso_size; ++ u32 local_size; ++ } lvl; ++ struct { ++ u32 *cipso; ++ u32 *local; ++ u32 cipso_size; ++ u32 local_size; ++ } cat; ++}; ++ ++struct cipso_v4_doi { ++ u32 doi; ++ u32 type; ++ union { ++ struct cipso_v4_std_map_tbl *std; ++ } map; ++ u8 tags[5]; ++ refcount_t refcount; ++ struct list_head list; ++ struct callback_head rcu; ++}; ++ ++struct netlbl_domaddr_map { ++ struct list_head list4; ++ struct list_head list6; ++}; ++ ++struct netlbl_dommap_def { ++ u32 type; ++ union { ++ struct netlbl_domaddr_map *addrsel; ++ struct cipso_v4_doi *cipso; ++ struct calipso_doi *calipso; ++ }; ++}; ++ ++struct netlbl_domaddr4_map { ++ struct netlbl_dommap_def def; ++ struct netlbl_af4list list; ++}; ++ ++struct netlbl_domaddr6_map { ++ struct netlbl_dommap_def def; ++ struct netlbl_af6list list; ++}; ++ ++struct netlbl_dom_map { ++ char *domain; ++ u16 family; ++ struct netlbl_dommap_def def; ++ u32 valid; ++ struct list_head list; ++ struct callback_head rcu; ++}; ++ ++enum { ++ NLBL_MGMT_C_UNSPEC = 0, ++ NLBL_MGMT_C_ADD = 1, ++ NLBL_MGMT_C_REMOVE = 2, ++ NLBL_MGMT_C_LISTALL = 3, ++ NLBL_MGMT_C_ADDDEF = 4, ++ NLBL_MGMT_C_REMOVEDEF = 5, ++ NLBL_MGMT_C_LISTDEF = 6, ++ NLBL_MGMT_C_PROTOCOLS = 7, ++ NLBL_MGMT_C_VERSION = 8, ++ __NLBL_MGMT_C_MAX = 9, ++}; ++ ++enum { ++ NLBL_MGMT_A_UNSPEC = 0, ++ NLBL_MGMT_A_DOMAIN = 1, ++ NLBL_MGMT_A_PROTOCOL = 2, ++ NLBL_MGMT_A_VERSION = 3, ++ NLBL_MGMT_A_CV4DOI = 4, ++ NLBL_MGMT_A_IPV6ADDR = 5, ++ NLBL_MGMT_A_IPV6MASK = 6, ++ NLBL_MGMT_A_IPV4ADDR = 7, ++ NLBL_MGMT_A_IPV4MASK = 8, ++ NLBL_MGMT_A_ADDRSELECTOR = 9, ++ NLBL_MGMT_A_SELECTORLIST = 10, ++ NLBL_MGMT_A_FAMILY = 11, ++ NLBL_MGMT_A_CLPDOI = 12, ++ __NLBL_MGMT_A_MAX = 13, ++}; ++ ++struct netlbl_domhsh_walk_arg { ++ struct netlink_callback *nl_cb; ++ struct sk_buff *skb; ++ u32 seq; ++}; ++ ++typedef long unsigned int addr_t; ++ ++struct ext_code { ++ union { ++ struct { ++ short unsigned int subcode; ++ short unsigned int code; ++ }; ++ unsigned int int_code; ++ }; ++}; ++ ++typedef void (*ext_int_handler_t)(struct ext_code, unsigned int, long unsigned int); ++ ++struct iucv_array { ++ u32 address; ++ u32 length; ++}; ++ ++struct iucv_handler; ++ ++struct iucv_path { ++ u16 pathid; ++ u16 msglim; ++ u8 flags; ++ void *private; ++ struct iucv_handler *handler; ++ struct list_head list; ++}; ++ ++struct iucv_message; ++ ++struct iucv_handler { ++ int (*path_pending)(struct iucv_path *, u8 *, u8 *); ++ void (*path_complete)(struct iucv_path *, u8 *); ++ void (*path_severed)(struct iucv_path *, u8 *); ++ void (*path_quiesced)(struct iucv_path *, u8 *); ++ void (*path_resumed)(struct iucv_path *, u8 *); ++ void (*message_pending)(struct iucv_path *, struct iucv_message *); ++ void (*message_complete)(struct iucv_path *, struct iucv_message *); ++ struct list_head list; ++ struct list_head paths; ++}; ++ ++struct iucv_message { ++ u32 id; ++ u32 audit; ++ u32 class; ++ u32 tag; ++ u32 length; ++ u32 reply_size; ++ u8 rmmsg[8]; ++ u8 flags; ++} __attribute__((packed)); ++ ++struct iucv_interface { ++ int (*message_receive)(struct iucv_path *, struct iucv_message *, u8, void *, size_t, size_t *); ++ int (*__message_receive)(struct iucv_path *, struct iucv_message *, u8, void *, size_t, size_t *); ++ int (*message_reply)(struct iucv_path *, struct iucv_message *, u8, void *, size_t); ++ int (*message_reject)(struct iucv_path *, struct iucv_message *); ++ int (*message_send)(struct iucv_path *, struct iucv_message *, u8, u32, void *, size_t); ++ int (*__message_send)(struct iucv_path *, struct iucv_message *, u8, u32, void *, size_t); ++ int (*message_send2way)(struct iucv_path *, struct iucv_message *, u8, u32, void *, size_t, void *, size_t, size_t *); ++ int (*message_purge)(struct iucv_path *, struct iucv_message *, u32); ++ int (*path_accept)(struct iucv_path *, struct iucv_handler *, u8 *, void *); ++ int (*path_connect)(struct iucv_path *, struct iucv_handler *, u8 *, u8 *, u8 *, void *); ++ int (*path_quiesce)(struct iucv_path *, u8 *); ++ int (*path_resume)(struct iucv_path *, u8 *); ++ int (*path_sever)(struct iucv_path *, u8 *); ++ int (*iucv_register)(struct iucv_handler *, int); ++ void (*iucv_unregister)(struct iucv_handler *, int); ++ struct bus_type *bus; ++ struct device *root; ++}; ++ ++struct iucv_irq_data { ++ u16 ippathid; ++ u8 ipflags1; ++ u8 iptype; ++ u32 res2[9]; ++}; ++ ++struct iucv_irq_list { ++ struct list_head list; ++ struct iucv_irq_data data; ++}; ++ ++enum iucv_command_codes { ++ IUCV_QUERY = 0, ++ IUCV_RETRIEVE_BUFFER = 2, ++ IUCV_SEND = 4, ++ IUCV_RECEIVE = 5, ++ IUCV_REPLY = 6, ++ IUCV_REJECT = 8, ++ IUCV_PURGE = 9, ++ IUCV_ACCEPT = 10, ++ IUCV_CONNECT = 11, ++ IUCV_DECLARE_BUFFER = 12, ++ IUCV_QUIESCE = 13, ++ IUCV_RESUME = 14, ++ IUCV_SEVER = 15, ++ IUCV_SETMASK = 16, ++ IUCV_SETCONTROLMASK = 17, ++}; ++ ++struct iucv_cmd_control { ++ u16 ippathid; ++ u8 ipflags1; ++ u8 iprcode; ++ u16 ipmsglim; ++ u16 res1; ++ u8 ipvmid[8]; ++ u8 ipuser[16]; ++ u8 iptarget[8]; ++}; ++ ++struct iucv_cmd_dpl { ++ u16 ippathid; ++ u8 ipflags1; ++ u8 iprcode; ++ u32 ipmsgid; ++ u32 iptrgcls; ++ u8 iprmmsg[8]; ++ u32 ipsrccls; ++ u32 ipmsgtag; ++ u32 ipbfadr2; ++ u32 ipbfln2f; ++ u32 res; ++}; ++ ++struct iucv_cmd_db { ++ u16 ippathid; ++ u8 ipflags1; ++ u8 iprcode; ++ u32 ipmsgid; ++ u32 iptrgcls; ++ u32 ipbfadr1; ++ u32 ipbfln1f; ++ u32 ipsrccls; ++ u32 ipmsgtag; ++ u32 ipbfadr2; ++ u32 ipbfln2f; ++ u32 res; ++}; ++ ++struct iucv_cmd_purge { ++ u16 ippathid; ++ u8 ipflags1; ++ u8 iprcode; ++ u32 ipmsgid; ++ u8 ipaudit[3]; ++ u8 res1[5]; ++ u32 res2; ++ u32 ipsrccls; ++ u32 ipmsgtag; ++ u32 res3[3]; ++}; ++ ++struct iucv_cmd_set_mask { ++ u8 ipmask; ++ u8 res1[2]; ++ u8 iprcode; ++ u32 res2[9]; ++}; ++ ++union iucv_param { ++ struct iucv_cmd_control ctrl; ++ struct iucv_cmd_dpl dpl; ++ struct iucv_cmd_db db; ++ struct iucv_cmd_purge purge; ++ struct iucv_cmd_set_mask set_mask; ++}; ++ ++struct iucv_path_pending { ++ u16 ippathid; ++ u8 ipflags1; ++ u8 iptype; ++ u16 ipmsglim; ++ u16 res1; ++ u8 ipvmid[8]; ++ u8 ipuser[16]; ++ u32 res3; ++ u8 ippollfg; ++ u8 res4[3]; ++}; ++ ++struct iucv_path_complete { ++ u16 ippathid; ++ u8 ipflags1; ++ u8 iptype; ++ u16 ipmsglim; ++ u16 res1; ++ u8 res2[8]; ++ u8 ipuser[16]; ++ u32 res3; ++ u8 ippollfg; ++ u8 res4[3]; ++}; ++ ++struct iucv_path_severed { ++ u16 ippathid; ++ u8 res1; ++ u8 iptype; ++ u32 res2; ++ u8 res3[8]; ++ u8 ipuser[16]; ++ u32 res4; ++ u8 ippollfg; ++ u8 res5[3]; ++}; ++ ++struct iucv_path_quiesced { ++ u16 ippathid; ++ u8 res1; ++ u8 iptype; ++ u32 res2; ++ u8 res3[8]; ++ u8 ipuser[16]; ++ u32 res4; ++ u8 ippollfg; ++ u8 res5[3]; ++}; ++ ++struct iucv_path_resumed { ++ u16 ippathid; ++ u8 res1; ++ u8 iptype; ++ u32 res2; ++ u8 res3[8]; ++ u8 ipuser[16]; ++ u32 res4; ++ u8 ippollfg; ++ u8 res5[3]; ++}; ++ ++struct iucv_message_complete { ++ u16 ippathid; ++ u8 ipflags1; ++ u8 iptype; ++ u32 ipmsgid; ++ u32 ipaudit; ++ u8 iprmmsg[8]; ++ u32 ipsrccls; ++ u32 ipmsgtag; ++ u32 res; ++ u32 ipbfln2f; ++ u8 ippollfg; ++ u8 res2[3]; ++}; ++ ++struct iucv_message_pending { ++ u16 ippathid; ++ u8 ipflags1; ++ u8 iptype; ++ u32 ipmsgid; ++ u32 iptrgcls; ++ struct { ++ union { ++ u32 iprmmsg1_u32; ++ u8 iprmmsg1[4]; ++ } ln1msg1; ++ union { ++ u32 ipbfln1f; ++ u8 iprmmsg2[4]; ++ } ln1msg2; ++ } rmmsg; ++ u32 res1[3]; ++ u32 ipbfln2f; ++ u8 ippollfg; ++ u8 res2[3]; ++}; ++ ++typedef void iucv_irq_fn(struct iucv_irq_data *); ++ ++struct netdev_nested_priv { ++ unsigned char flags; ++ void *data; ++}; ++ ++enum switchdev_attr_id { ++ SWITCHDEV_ATTR_ID_UNDEFINED = 0, ++ SWITCHDEV_ATTR_ID_PORT_STP_STATE = 1, ++ SWITCHDEV_ATTR_ID_PORT_MST_STATE = 2, ++ SWITCHDEV_ATTR_ID_PORT_BRIDGE_FLAGS = 3, ++ SWITCHDEV_ATTR_ID_PORT_PRE_BRIDGE_FLAGS = 4, ++ SWITCHDEV_ATTR_ID_PORT_MROUTER = 5, ++ SWITCHDEV_ATTR_ID_BRIDGE_AGEING_TIME = 6, ++ SWITCHDEV_ATTR_ID_BRIDGE_VLAN_FILTERING = 7, ++ SWITCHDEV_ATTR_ID_BRIDGE_VLAN_PROTOCOL = 8, ++ SWITCHDEV_ATTR_ID_BRIDGE_MC_DISABLED = 9, ++ SWITCHDEV_ATTR_ID_BRIDGE_MROUTER = 10, ++ SWITCHDEV_ATTR_ID_BRIDGE_MST = 11, ++ SWITCHDEV_ATTR_ID_MRP_PORT_ROLE = 12, ++ SWITCHDEV_ATTR_ID_VLAN_MSTI = 13, ++}; ++ ++struct switchdev_attr { ++ struct net_device *orig_dev; ++ enum switchdev_attr_id id; ++ u32 flags; ++ void *complete_priv; ++ void (*complete)(struct net_device *, int, void *); ++ union { ++ u8 stp_state; ++ struct switchdev_mst_state mst_state; ++ struct switchdev_brport_flags brport_flags; ++ bool mrouter; ++ clock_t ageing_time; ++ bool vlan_filtering; ++ u16 vlan_protocol; ++ bool mst; ++ bool mc_disabled; ++ u8 mrp_port_role; ++ struct switchdev_vlan_msti vlan_msti; ++ } u; ++}; ++ ++struct switchdev_brport { ++ struct net_device *dev; ++ const void *ctx; ++ struct notifier_block *atomic_nb; ++ struct notifier_block *blocking_nb; ++ bool tx_fwd_offload; ++}; ++ ++enum switchdev_notifier_type { ++ SWITCHDEV_FDB_ADD_TO_BRIDGE = 1, ++ SWITCHDEV_FDB_DEL_TO_BRIDGE = 2, ++ SWITCHDEV_FDB_ADD_TO_DEVICE = 3, ++ SWITCHDEV_FDB_DEL_TO_DEVICE = 4, ++ SWITCHDEV_FDB_OFFLOADED = 5, ++ SWITCHDEV_FDB_FLUSH_TO_BRIDGE = 6, ++ SWITCHDEV_PORT_OBJ_ADD = 7, ++ SWITCHDEV_PORT_OBJ_DEL = 8, ++ SWITCHDEV_PORT_ATTR_SET = 9, ++ SWITCHDEV_VXLAN_FDB_ADD_TO_BRIDGE = 10, ++ SWITCHDEV_VXLAN_FDB_DEL_TO_BRIDGE = 11, ++ SWITCHDEV_VXLAN_FDB_ADD_TO_DEVICE = 12, ++ SWITCHDEV_VXLAN_FDB_DEL_TO_DEVICE = 13, ++ SWITCHDEV_VXLAN_FDB_OFFLOADED = 14, ++ SWITCHDEV_BRPORT_OFFLOADED = 15, ++ SWITCHDEV_BRPORT_UNOFFLOADED = 16, ++}; ++ ++struct switchdev_notifier_info { ++ struct net_device *dev; ++ struct netlink_ext_ack *extack; ++ const void *ctx; ++}; ++ ++struct switchdev_notifier_fdb_info { ++ struct switchdev_notifier_info info; ++ const unsigned char *addr; ++ u16 vid; ++ u8 added_by_user: 1; ++ u8 is_local: 1; ++ u8 locked: 1; ++ u8 offloaded: 1; ++}; ++ ++struct switchdev_notifier_port_obj_info { ++ struct switchdev_notifier_info info; ++ const struct switchdev_obj *obj; ++ bool handled; ++}; ++ ++struct switchdev_notifier_port_attr_info { ++ struct switchdev_notifier_info info; ++ const struct switchdev_attr *attr; ++ bool handled; ++}; ++ ++struct switchdev_notifier_brport_info { ++ struct switchdev_notifier_info info; ++ const struct switchdev_brport brport; ++}; ++ ++typedef void switchdev_deferred_func_t(struct net_device *, const void *); ++ ++struct switchdev_deferred_item { ++ struct list_head list; ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ switchdev_deferred_func_t *func; ++ long unsigned int data[0]; ++}; ++ ++struct switchdev_nested_priv { ++ bool (*check_cb)(const struct net_device *); ++ bool (*foreign_dev_check_cb)(const struct net_device *, const struct net_device *); ++ const struct net_device *dev; ++ struct net_device *lower_dev; ++}; ++ ++struct ncsi_dev { ++ int state; ++ int link_up; ++ struct net_device *dev; ++ void (*handler)(struct ncsi_dev *); ++}; ++ ++enum { ++ NCSI_CAP_BASE = 0, ++ NCSI_CAP_GENERIC = 0, ++ NCSI_CAP_BC = 1, ++ NCSI_CAP_MC = 2, ++ NCSI_CAP_BUFFER = 3, ++ NCSI_CAP_AEN = 4, ++ NCSI_CAP_VLAN = 5, ++ NCSI_CAP_MAX = 6, ++}; ++ ++enum { ++ NCSI_MODE_BASE = 0, ++ NCSI_MODE_ENABLE = 0, ++ NCSI_MODE_TX_ENABLE = 1, ++ NCSI_MODE_LINK = 2, ++ NCSI_MODE_VLAN = 3, ++ NCSI_MODE_BC = 4, ++ NCSI_MODE_MC = 5, ++ NCSI_MODE_AEN = 6, ++ NCSI_MODE_FC = 7, ++ NCSI_MODE_MAX = 8, ++}; ++ ++struct ncsi_channel_version { ++ u32 version; ++ u32 alpha2; ++ u8 fw_name[12]; ++ u32 fw_version; ++ u16 pci_ids[4]; ++ u32 mf_id; ++}; ++ ++struct ncsi_channel_cap { ++ u32 index; ++ u32 cap; ++}; ++ ++struct ncsi_channel_mode { ++ u32 index; ++ u32 enable; ++ u32 size; ++ u32 data[8]; ++}; ++ ++struct ncsi_channel_mac_filter { ++ u8 n_uc; ++ u8 n_mc; ++ u8 n_mixed; ++ u64 bitmap; ++ unsigned char *addrs; ++}; ++ ++struct ncsi_channel_vlan_filter { ++ u8 n_vids; ++ u64 bitmap; ++ u16 *vids; ++}; ++ ++struct ncsi_channel_stats { ++ u32 hnc_cnt_hi; ++ u32 hnc_cnt_lo; ++ u32 hnc_rx_bytes; ++ u32 hnc_tx_bytes; ++ u32 hnc_rx_uc_pkts; ++ u32 hnc_rx_mc_pkts; ++ u32 hnc_rx_bc_pkts; ++ u32 hnc_tx_uc_pkts; ++ u32 hnc_tx_mc_pkts; ++ u32 hnc_tx_bc_pkts; ++ u32 hnc_fcs_err; ++ u32 hnc_align_err; ++ u32 hnc_false_carrier; ++ u32 hnc_runt_pkts; ++ u32 hnc_jabber_pkts; ++ u32 hnc_rx_pause_xon; ++ u32 hnc_rx_pause_xoff; ++ u32 hnc_tx_pause_xon; ++ u32 hnc_tx_pause_xoff; ++ u32 hnc_tx_s_collision; ++ u32 hnc_tx_m_collision; ++ u32 hnc_l_collision; ++ u32 hnc_e_collision; ++ u32 hnc_rx_ctl_frames; ++ u32 hnc_rx_64_frames; ++ u32 hnc_rx_127_frames; ++ u32 hnc_rx_255_frames; ++ u32 hnc_rx_511_frames; ++ u32 hnc_rx_1023_frames; ++ u32 hnc_rx_1522_frames; ++ u32 hnc_rx_9022_frames; ++ u32 hnc_tx_64_frames; ++ u32 hnc_tx_127_frames; ++ u32 hnc_tx_255_frames; ++ u32 hnc_tx_511_frames; ++ u32 hnc_tx_1023_frames; ++ u32 hnc_tx_1522_frames; ++ u32 hnc_tx_9022_frames; ++ u32 hnc_rx_valid_bytes; ++ u32 hnc_rx_runt_pkts; ++ u32 hnc_rx_jabber_pkts; ++ u32 ncsi_rx_cmds; ++ u32 ncsi_dropped_cmds; ++ u32 ncsi_cmd_type_errs; ++ u32 ncsi_cmd_csum_errs; ++ u32 ncsi_rx_pkts; ++ u32 ncsi_tx_pkts; ++ u32 ncsi_tx_aen_pkts; ++ u32 pt_tx_pkts; ++ u32 pt_tx_dropped; ++ u32 pt_tx_channel_err; ++ u32 pt_tx_us_err; ++ u32 pt_rx_pkts; ++ u32 pt_rx_dropped; ++ u32 pt_rx_channel_err; ++ u32 pt_rx_us_err; ++ u32 pt_rx_os_err; ++}; ++ ++struct ncsi_package; ++ ++struct ncsi_channel { ++ unsigned char id; ++ int state; ++ bool reconfigure_needed; ++ spinlock_t lock; ++ struct ncsi_package *package; ++ struct ncsi_channel_version version; ++ struct ncsi_channel_cap caps[6]; ++ struct ncsi_channel_mode modes[8]; ++ struct ncsi_channel_mac_filter mac_filter; ++ struct ncsi_channel_vlan_filter vlan_filter; ++ struct ncsi_channel_stats stats; ++ struct { ++ struct timer_list timer; ++ bool enabled; ++ unsigned int state; ++ } monitor; ++ struct list_head node; ++ struct list_head link; ++}; ++ ++struct ncsi_dev_priv; ++ ++struct ncsi_package { ++ unsigned char id; ++ unsigned char uuid[16]; ++ struct ncsi_dev_priv *ndp; ++ spinlock_t lock; ++ unsigned int channel_num; ++ struct list_head channels; ++ struct list_head node; ++ bool multi_channel; ++ u32 channel_whitelist; ++ struct ncsi_channel *preferred_channel; ++}; ++ ++struct ncsi_request { ++ unsigned char id; ++ bool used; ++ unsigned int flags; ++ struct ncsi_dev_priv *ndp; ++ struct sk_buff *cmd; ++ struct sk_buff *rsp; ++ struct timer_list timer; ++ bool enabled; ++ u32 snd_seq; ++ u32 snd_portid; ++ struct nlmsghdr nlhdr; ++}; ++ ++struct ncsi_dev_priv { ++ struct ncsi_dev ndev; ++ unsigned int flags; ++ unsigned int gma_flag; ++ spinlock_t lock; ++ unsigned int package_probe_id; ++ unsigned int package_num; ++ struct list_head packages; ++ struct ncsi_channel *hot_channel; ++ struct ncsi_request requests[256]; ++ unsigned int request_id; ++ unsigned int pending_req_num; ++ struct ncsi_package *active_package; ++ struct ncsi_channel *active_channel; ++ struct list_head channel_queue; ++ struct work_struct work; ++ struct packet_type ptype; ++ struct list_head node; ++ struct list_head vlan_vids; ++ bool multi_package; ++ bool mlx_multi_host; ++ u32 package_whitelist; ++}; ++ ++struct ncsi_pkt_hdr { ++ unsigned char mc_id; ++ unsigned char revision; ++ unsigned char reserved; ++ unsigned char id; ++ unsigned char type; ++ unsigned char channel; ++ __be16 length; ++ __be32 reserved1[2]; ++}; ++ ++struct ncsi_aen_pkt_hdr { ++ struct ncsi_pkt_hdr common; ++ unsigned char reserved2[3]; ++ unsigned char type; ++}; ++ ++struct ncsi_aen_lsc_pkt { ++ struct ncsi_aen_pkt_hdr aen; ++ __be32 status; ++ __be32 oem_status; ++ __be32 checksum; ++ unsigned char pad[14]; ++}; ++ ++struct ncsi_aen_hncdsc_pkt { ++ struct ncsi_aen_pkt_hdr aen; ++ __be32 status; ++ __be32 checksum; ++ unsigned char pad[18]; ++}; ++ ++struct ncsi_aen_handler { ++ unsigned char type; ++ int payload; ++ int (*handler)(struct ncsi_dev_priv *, struct ncsi_aen_pkt_hdr *); ++}; ++ ++struct sockaddr_xdp { ++ __u16 sxdp_family; ++ __u16 sxdp_flags; ++ __u32 sxdp_ifindex; ++ __u32 sxdp_queue_id; ++ __u32 sxdp_shared_umem_fd; ++}; ++ ++struct xdp_ring_offset { ++ __u64 producer; ++ __u64 consumer; ++ __u64 desc; ++ __u64 flags; ++}; ++ ++struct xdp_mmap_offsets { ++ struct xdp_ring_offset rx; ++ struct xdp_ring_offset tx; ++ struct xdp_ring_offset fr; ++ struct xdp_ring_offset cr; ++}; ++ ++struct xdp_umem_reg { ++ __u64 addr; ++ __u64 len; ++ __u32 chunk_size; ++ __u32 headroom; ++ __u32 flags; ++}; ++ ++struct xdp_statistics { ++ __u64 rx_dropped; ++ __u64 rx_invalid_descs; ++ __u64 tx_invalid_descs; ++ __u64 rx_ring_full; ++ __u64 rx_fill_ring_empty_descs; ++ __u64 tx_ring_empty_descs; ++}; ++ ++struct xdp_options { ++ __u32 flags; ++}; ++ ++struct xdp_sock; ++ ++struct xsk_map { ++ struct bpf_map map; ++ spinlock_t lock; ++ atomic_t count; ++ struct xdp_sock *xsk_map[0]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct xdp_sock { ++ struct sock sk; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct xsk_queue *rx; ++ struct net_device *dev; ++ struct xdp_umem *umem; ++ struct list_head flush_node; ++ struct xsk_buff_pool *pool; ++ u16 queue_id; ++ bool zc; ++ enum { ++ XSK_READY = 0, ++ XSK_BOUND = 1, ++ XSK_UNBOUND = 2, ++ } state; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct xsk_queue *tx; ++ struct list_head tx_list; ++ spinlock_t rx_lock; ++ u64 rx_dropped; ++ u64 rx_queue_full; ++ struct list_head map_list; ++ spinlock_t map_list_lock; ++ struct mutex mutex; ++ struct xsk_queue *fq_tmp; ++ struct xsk_queue *cq_tmp; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct xdp_ring; ++ ++struct xsk_queue { ++ u32 ring_mask; ++ u32 nentries; ++ u32 cached_prod; ++ u32 cached_cons; ++ struct xdp_ring *ring; ++ u64 invalid_descs; ++ u64 queue_empty_descs; ++ size_t ring_vmalloc_size; ++}; ++ ++struct xdp_ring_offset_v1 { ++ __u64 producer; ++ __u64 consumer; ++ __u64 desc; ++}; ++ ++struct xdp_mmap_offsets_v1 { ++ struct xdp_ring_offset_v1 rx; ++ struct xdp_ring_offset_v1 tx; ++ struct xdp_ring_offset_v1 fr; ++ struct xdp_ring_offset_v1 cr; ++}; ++ ++struct xsk_map_node { ++ struct list_head node; ++ struct xsk_map *map; ++ struct xdp_sock **map_entry; ++}; ++ ++struct xdp_ring { ++ u32 producer; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ u32 pad1; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ u32 consumer; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ u32 pad2; ++ u32 flags; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ u32 pad3; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct xdp_rxtx_ring { ++ struct xdp_ring ptrs; ++ struct xdp_desc desc[0]; ++}; ++ ++struct xdp_umem_ring { ++ struct xdp_ring ptrs; ++ u64 desc[0]; ++}; ++ ++struct mptcp_mib { ++ long unsigned int mibs[52]; ++}; ++ ++struct mptcp_rm_list { ++ u8 ids[8]; ++ u8 nr; ++}; ++ ++struct mptcp_addr_info { ++ u8 id; ++ sa_family_t family; ++ __be16 port; ++ union { ++ struct in_addr addr; ++ struct in6_addr addr6; ++ }; ++}; ++ ++enum mptcp_event_type { ++ MPTCP_EVENT_UNSPEC = 0, ++ MPTCP_EVENT_CREATED = 1, ++ MPTCP_EVENT_ESTABLISHED = 2, ++ MPTCP_EVENT_CLOSED = 3, ++ MPTCP_EVENT_ANNOUNCED = 6, ++ MPTCP_EVENT_REMOVED = 7, ++ MPTCP_EVENT_SUB_ESTABLISHED = 10, ++ MPTCP_EVENT_SUB_CLOSED = 11, ++ MPTCP_EVENT_SUB_PRIORITY = 13, ++ MPTCP_EVENT_LISTENER_CREATED = 15, ++ MPTCP_EVENT_LISTENER_CLOSED = 16, ++}; ++ ++struct mptcp_skb_cb { ++ u64 map_seq; ++ u64 end_seq; ++ u32 offset; ++ u8 has_rxtstamp: 1; ++}; ++ ++struct mptcp_options_received { ++ u64 sndr_key; ++ u64 rcvr_key; ++ u64 data_ack; ++ u64 data_seq; ++ u32 subflow_seq; ++ u16 data_len; ++ __sum16 csum; ++ u16 suboptions; ++ u32 token; ++ u32 nonce; ++ u16 use_map: 1; ++ u16 dsn64: 1; ++ u16 data_fin: 1; ++ u16 use_ack: 1; ++ u16 ack64: 1; ++ u16 mpc_map: 1; ++ u16 reset_reason: 4; ++ u16 reset_transient: 1; ++ u16 echo: 1; ++ u16 backup: 1; ++ u16 deny_join_id0: 1; ++ u16 __unused: 2; ++ u8 join_id; ++ u64 thmac; ++ u8 hmac[20]; ++ struct mptcp_addr_info addr; ++ struct mptcp_rm_list rm_list; ++ u64 ahmac; ++ u64 fail_seq; ++}; ++ ++struct mptcp_pm_data { ++ struct mptcp_addr_info local; ++ struct mptcp_addr_info remote; ++ struct list_head anno_list; ++ struct list_head userspace_pm_local_addr_list; ++ spinlock_t lock; ++ u8 addr_signal; ++ bool server_side; ++ bool work_pending; ++ bool accept_addr; ++ bool accept_subflow; ++ bool remote_deny_join_id0; ++ u8 add_addr_signaled; ++ u8 add_addr_accepted; ++ u8 local_addr_used; ++ u8 pm_type; ++ u8 subflows; ++ u8 status; ++ long unsigned int id_avail_bitmap[4]; ++ struct mptcp_rm_list rm_list_tx; ++ struct mptcp_rm_list rm_list_rx; ++}; ++ ++struct mptcp_data_frag { ++ struct list_head list; ++ u64 data_seq; ++ u16 data_len; ++ u16 offset; ++ u16 overhead; ++ u16 already_sent; ++ struct page *page; ++}; ++ ++struct mptcp_sock { ++ struct inet_connection_sock sk; ++ u64 local_key; ++ u64 remote_key; ++ u64 write_seq; ++ u64 snd_nxt; ++ u64 ack_seq; ++ atomic64_t rcv_wnd_sent; ++ u64 rcv_data_fin_seq; ++ int rmem_fwd_alloc; ++ struct sock *last_snd; ++ int snd_burst; ++ int old_wspace; ++ u64 recovery_snd_nxt; ++ u64 snd_una; ++ u64 wnd_end; ++ long unsigned int timer_ival; ++ u32 token; ++ int rmem_released; ++ long unsigned int flags; ++ long unsigned int cb_flags; ++ long unsigned int push_pending; ++ bool recovery; ++ bool can_ack; ++ bool fully_established; ++ bool rcv_data_fin; ++ bool snd_data_fin_enable; ++ bool rcv_fastclose; ++ bool use_64bit_ack; ++ bool csum_enabled; ++ bool allow_infinite_fallback; ++ u8 mpc_endpoint_id; ++ u8 recvmsg_inq: 1; ++ u8 cork: 1; ++ u8 nodelay: 1; ++ u8 fastopening: 1; ++ u8 in_accept_queue: 1; ++ struct work_struct work; ++ struct sk_buff *ooo_last_skb; ++ struct rb_root out_of_order_queue; ++ struct sk_buff_head receive_queue; ++ struct list_head conn_list; ++ struct list_head rtx_queue; ++ struct mptcp_data_frag *first_pending; ++ struct list_head join_list; ++ struct socket *subflow; ++ struct sock *first; ++ struct mptcp_pm_data pm; ++ struct { ++ u32 space; ++ u32 copied; ++ u64 time; ++ u64 rtt_us; ++ } rcvq_space; ++ u32 setsockopt_seq; ++ char ca_name[16]; ++}; ++ ++struct mptcp_subflow_request_sock { ++ struct tcp_request_sock sk; ++ u16 mp_capable: 1; ++ u16 mp_join: 1; ++ u16 backup: 1; ++ u16 csum_reqd: 1; ++ u16 allow_join_id0: 1; ++ u8 local_id; ++ u8 remote_id; ++ u64 local_key; ++ u64 idsn; ++ u32 token; ++ u32 ssn_offset; ++ u64 thmac; ++ u32 local_nonce; ++ u32 remote_nonce; ++ struct mptcp_sock *msk; ++ struct hlist_nulls_node token_node; ++}; ++ ++enum mptcp_data_avail { ++ MPTCP_SUBFLOW_NODATA = 0, ++ MPTCP_SUBFLOW_DATA_AVAIL = 1, ++}; ++ ++struct mptcp_delegated_action { ++ struct napi_struct napi; ++ struct list_head head; ++}; ++ ++struct mptcp_subflow_context { ++ struct list_head node; ++ union { ++ struct { ++ long unsigned int avg_pacing_rate; ++ u64 local_key; ++ u64 remote_key; ++ u64 idsn; ++ u64 map_seq; ++ u32 snd_isn; ++ u32 token; ++ u32 rel_write_seq; ++ u32 map_subflow_seq; ++ u32 ssn_offset; ++ u32 map_data_len; ++ __wsum map_data_csum; ++ u32 map_csum_len; ++ u32 request_mptcp: 1; ++ u32 request_join: 1; ++ u32 request_bkup: 1; ++ u32 mp_capable: 1; ++ u32 mp_join: 1; ++ u32 fully_established: 1; ++ u32 pm_notified: 1; ++ u32 conn_finished: 1; ++ u32 map_valid: 1; ++ u32 map_csum_reqd: 1; ++ u32 map_data_fin: 1; ++ u32 mpc_map: 1; ++ u32 backup: 1; ++ u32 send_mp_prio: 1; ++ u32 send_mp_fail: 1; ++ u32 send_fastclose: 1; ++ u32 send_infinite_map: 1; ++ u32 remote_key_valid: 1; ++ u32 disposable: 1; ++ u32 stale: 1; ++ u32 local_id_valid: 1; ++ u32 valid_csum_seen: 1; ++ u32 is_mptfo: 1; ++ u32 __unused: 9; ++ enum mptcp_data_avail data_avail; ++ u32 remote_nonce; ++ u64 thmac; ++ u32 local_nonce; ++ u32 remote_token; ++ union { ++ u8 hmac[20]; ++ u64 iasn; ++ }; ++ u8 local_id; ++ u8 remote_id; ++ u8 reset_seen: 1; ++ u8 reset_transient: 1; ++ u8 reset_reason: 4; ++ u8 stale_count; ++ long int delegated_status; ++ long unsigned int fail_tout; ++ }; ++ struct { ++ long unsigned int avg_pacing_rate; ++ u64 local_key; ++ u64 remote_key; ++ u64 idsn; ++ u64 map_seq; ++ u32 snd_isn; ++ u32 token; ++ u32 rel_write_seq; ++ u32 map_subflow_seq; ++ u32 ssn_offset; ++ u32 map_data_len; ++ __wsum map_data_csum; ++ u32 map_csum_len; ++ u32 request_mptcp: 1; ++ u32 request_join: 1; ++ u32 request_bkup: 1; ++ u32 mp_capable: 1; ++ u32 mp_join: 1; ++ u32 fully_established: 1; ++ u32 pm_notified: 1; ++ u32 conn_finished: 1; ++ u32 map_valid: 1; ++ u32 map_csum_reqd: 1; ++ u32 map_data_fin: 1; ++ u32 mpc_map: 1; ++ u32 backup: 1; ++ u32 send_mp_prio: 1; ++ u32 send_mp_fail: 1; ++ u32 send_fastclose: 1; ++ u32 send_infinite_map: 1; ++ u32 remote_key_valid: 1; ++ u32 disposable: 1; ++ u32 stale: 1; ++ u32 local_id_valid: 1; ++ u32 valid_csum_seen: 1; ++ u32 is_mptfo: 1; ++ u32 __unused: 9; ++ enum mptcp_data_avail data_avail; ++ u32 remote_nonce; ++ u64 thmac; ++ u32 local_nonce; ++ u32 remote_token; ++ union { ++ u8 hmac[20]; ++ u64 iasn; ++ }; ++ u8 local_id; ++ u8 remote_id; ++ u8 reset_seen: 1; ++ u8 reset_transient: 1; ++ u8 reset_reason: 4; ++ u8 stale_count; ++ long int delegated_status; ++ long unsigned int fail_tout; ++ } reset; ++ }; ++ struct list_head delegated_node; ++ u32 setsockopt_seq; ++ u32 stale_rcv_tstamp; ++ struct sock *tcp_sock; ++ struct sock *conn; ++ const struct inet_connection_sock_af_ops *icsk_af_ops; ++ void (*tcp_state_change)(struct sock *); ++ void (*tcp_error_report)(struct sock *); ++ struct callback_head rcu; ++}; ++ ++enum linux_mptcp_mib_field { ++ MPTCP_MIB_NUM = 0, ++ MPTCP_MIB_MPCAPABLEPASSIVE = 1, ++ MPTCP_MIB_MPCAPABLEACTIVE = 2, ++ MPTCP_MIB_MPCAPABLEACTIVEACK = 3, ++ MPTCP_MIB_MPCAPABLEPASSIVEACK = 4, ++ MPTCP_MIB_MPCAPABLEPASSIVEFALLBACK = 5, ++ MPTCP_MIB_MPCAPABLEACTIVEFALLBACK = 6, ++ MPTCP_MIB_TOKENFALLBACKINIT = 7, ++ MPTCP_MIB_RETRANSSEGS = 8, ++ MPTCP_MIB_JOINNOTOKEN = 9, ++ MPTCP_MIB_JOINSYNRX = 10, ++ MPTCP_MIB_JOINSYNACKRX = 11, ++ MPTCP_MIB_JOINSYNACKMAC = 12, ++ MPTCP_MIB_JOINACKRX = 13, ++ MPTCP_MIB_JOINACKMAC = 14, ++ MPTCP_MIB_DSSNOMATCH = 15, ++ MPTCP_MIB_INFINITEMAPTX = 16, ++ MPTCP_MIB_INFINITEMAPRX = 17, ++ MPTCP_MIB_DSSTCPMISMATCH = 18, ++ MPTCP_MIB_DATACSUMERR = 19, ++ MPTCP_MIB_OFOQUEUETAIL = 20, ++ MPTCP_MIB_OFOQUEUE = 21, ++ MPTCP_MIB_OFOMERGE = 22, ++ MPTCP_MIB_NODSSWINDOW = 23, ++ MPTCP_MIB_DUPDATA = 24, ++ MPTCP_MIB_ADDADDR = 25, ++ MPTCP_MIB_ECHOADD = 26, ++ MPTCP_MIB_PORTADD = 27, ++ MPTCP_MIB_ADDADDRDROP = 28, ++ MPTCP_MIB_JOINPORTSYNRX = 29, ++ MPTCP_MIB_JOINPORTSYNACKRX = 30, ++ MPTCP_MIB_JOINPORTACKRX = 31, ++ MPTCP_MIB_MISMATCHPORTSYNRX = 32, ++ MPTCP_MIB_MISMATCHPORTACKRX = 33, ++ MPTCP_MIB_RMADDR = 34, ++ MPTCP_MIB_RMADDRDROP = 35, ++ MPTCP_MIB_RMSUBFLOW = 36, ++ MPTCP_MIB_MPPRIOTX = 37, ++ MPTCP_MIB_MPPRIORX = 38, ++ MPTCP_MIB_MPFAILTX = 39, ++ MPTCP_MIB_MPFAILRX = 40, ++ MPTCP_MIB_MPFASTCLOSETX = 41, ++ MPTCP_MIB_MPFASTCLOSERX = 42, ++ MPTCP_MIB_MPRSTTX = 43, ++ MPTCP_MIB_MPRSTRX = 44, ++ MPTCP_MIB_RCVPRUNED = 45, ++ MPTCP_MIB_SUBFLOWSTALE = 46, ++ MPTCP_MIB_SUBFLOWRECOVER = 47, ++ MPTCP_MIB_SNDWNDSHARED = 48, ++ MPTCP_MIB_RCVWNDSHARED = 49, ++ MPTCP_MIB_RCVWNDCONFLICTUPDATE = 50, ++ MPTCP_MIB_RCVWNDCONFLICT = 51, ++ __MPTCP_MIB_MAX = 52, ++}; ++ ++struct trace_event_raw_mptcp_subflow_get_send { ++ struct trace_entry ent; ++ bool active; ++ bool free; ++ u32 snd_wnd; ++ u32 pace; ++ u8 backup; ++ u64 ratio; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mptcp_dump_mpext { ++ struct trace_entry ent; ++ u64 data_ack; ++ u64 data_seq; ++ u32 subflow_seq; ++ u16 data_len; ++ u16 csum; ++ u8 use_map; ++ u8 dsn64; ++ u8 data_fin; ++ u8 use_ack; ++ u8 ack64; ++ u8 mpc_map; ++ u8 frozen; ++ u8 reset_transient; ++ u8 reset_reason; ++ u8 csum_reqd; ++ u8 infinite_map; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ack_update_msk { ++ struct trace_entry ent; ++ u64 data_ack; ++ u64 old_snd_una; ++ u64 new_snd_una; ++ u64 new_wnd_end; ++ u64 msk_wnd_end; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_subflow_check_data_avail { ++ struct trace_entry ent; ++ u8 status; ++ const void *skb; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_mptcp_subflow_get_send {}; ++ ++struct trace_event_data_offsets_mptcp_dump_mpext {}; ++ ++struct trace_event_data_offsets_ack_update_msk {}; ++ ++struct trace_event_data_offsets_subflow_check_data_avail {}; ++ ++typedef void (*btf_trace_mptcp_subflow_get_send)(void *, struct mptcp_subflow_context *); ++ ++typedef void (*btf_trace_mptcp_sendmsg_frag)(void *, struct mptcp_ext *); ++ ++typedef void (*btf_trace_get_mapping_status)(void *, struct mptcp_ext *); ++ ++typedef void (*btf_trace_ack_update_msk)(void *, u64, u64, u64, u64, u64); ++ ++typedef void (*btf_trace_subflow_check_data_avail)(void *, __u8, struct sk_buff *); ++ ++enum { ++ MPTCP_CMSG_TS = 1, ++ MPTCP_CMSG_INQ = 2, ++}; ++ ++struct mptcp_sendmsg_info { ++ int mss_now; ++ int size_goal; ++ u16 limit; ++ u16 sent; ++ unsigned int flags; ++ bool data_lock_held; ++}; ++ ++struct subflow_send_info { ++ struct sock *ssk; ++ u64 linger_time; ++}; ++ ++enum { ++ INET_ULP_INFO_UNSPEC = 0, ++ INET_ULP_INFO_NAME = 1, ++ INET_ULP_INFO_TLS = 2, ++ INET_ULP_INFO_MPTCP = 3, ++ __INET_ULP_INFO_MAX = 4, ++}; ++ ++struct tcpvegas_info { ++ __u32 tcpv_enabled; ++ __u32 tcpv_rttcnt; ++ __u32 tcpv_rtt; ++ __u32 tcpv_minrtt; ++}; ++ ++struct tcp_dctcp_info { ++ __u16 dctcp_enabled; ++ __u16 dctcp_ce_state; ++ __u32 dctcp_alpha; ++ __u32 dctcp_ab_ecn; ++ __u32 dctcp_ab_tot; ++}; ++ ++struct tcp_bbr_info { ++ __u32 bbr_bw_lo; ++ __u32 bbr_bw_hi; ++ __u32 bbr_min_rtt; ++ __u32 bbr_pacing_gain; ++ __u32 bbr_cwnd_gain; ++}; ++ ++union tcp_cc_info { ++ struct tcpvegas_info vegas; ++ struct tcp_dctcp_info dctcp; ++ struct tcp_bbr_info bbr; ++}; ++ ++enum { ++ MPTCP_SUBFLOW_ATTR_UNSPEC = 0, ++ MPTCP_SUBFLOW_ATTR_TOKEN_REM = 1, ++ MPTCP_SUBFLOW_ATTR_TOKEN_LOC = 2, ++ MPTCP_SUBFLOW_ATTR_RELWRITE_SEQ = 3, ++ MPTCP_SUBFLOW_ATTR_MAP_SEQ = 4, ++ MPTCP_SUBFLOW_ATTR_MAP_SFSEQ = 5, ++ MPTCP_SUBFLOW_ATTR_SSN_OFFSET = 6, ++ MPTCP_SUBFLOW_ATTR_MAP_DATALEN = 7, ++ MPTCP_SUBFLOW_ATTR_FLAGS = 8, ++ MPTCP_SUBFLOW_ATTR_ID_REM = 9, ++ MPTCP_SUBFLOW_ATTR_ID_LOC = 10, ++ MPTCP_SUBFLOW_ATTR_PAD = 11, ++ __MPTCP_SUBFLOW_ATTR_MAX = 12, ++}; ++ ++struct linger { ++ int l_onoff; ++ int l_linger; ++}; ++ ++struct so_timestamping { ++ int flags; ++ int bind_phc; ++}; ++ ++struct tcp_info { ++ __u8 tcpi_state; ++ __u8 tcpi_ca_state; ++ __u8 tcpi_retransmits; ++ __u8 tcpi_probes; ++ __u8 tcpi_backoff; ++ __u8 tcpi_options; ++ __u8 tcpi_snd_wscale: 4; ++ __u8 tcpi_rcv_wscale: 4; ++ __u8 tcpi_delivery_rate_app_limited: 1; ++ __u8 tcpi_fastopen_client_fail: 2; ++ __u32 tcpi_rto; ++ __u32 tcpi_ato; ++ __u32 tcpi_snd_mss; ++ __u32 tcpi_rcv_mss; ++ __u32 tcpi_unacked; ++ __u32 tcpi_sacked; ++ __u32 tcpi_lost; ++ __u32 tcpi_retrans; ++ __u32 tcpi_fackets; ++ __u32 tcpi_last_data_sent; ++ __u32 tcpi_last_ack_sent; ++ __u32 tcpi_last_data_recv; ++ __u32 tcpi_last_ack_recv; ++ __u32 tcpi_pmtu; ++ __u32 tcpi_rcv_ssthresh; ++ __u32 tcpi_rtt; ++ __u32 tcpi_rttvar; ++ __u32 tcpi_snd_ssthresh; ++ __u32 tcpi_snd_cwnd; ++ __u32 tcpi_advmss; ++ __u32 tcpi_reordering; ++ __u32 tcpi_rcv_rtt; ++ __u32 tcpi_rcv_space; ++ __u32 tcpi_total_retrans; ++ __u64 tcpi_pacing_rate; ++ __u64 tcpi_max_pacing_rate; ++ __u64 tcpi_bytes_acked; ++ __u64 tcpi_bytes_received; ++ __u32 tcpi_segs_out; ++ __u32 tcpi_segs_in; ++ __u32 tcpi_notsent_bytes; ++ __u32 tcpi_min_rtt; ++ __u32 tcpi_data_segs_in; ++ __u32 tcpi_data_segs_out; ++ __u64 tcpi_delivery_rate; ++ __u64 tcpi_busy_time; ++ __u64 tcpi_rwnd_limited; ++ __u64 tcpi_sndbuf_limited; ++ __u32 tcpi_delivered; ++ __u32 tcpi_delivered_ce; ++ __u64 tcpi_bytes_sent; ++ __u64 tcpi_bytes_retrans; ++ __u32 tcpi_dsack_dups; ++ __u32 tcpi_reord_seen; ++ __u32 tcpi_rcv_ooopack; ++ __u32 tcpi_snd_wnd; ++ __u32 tcpi_rcv_wnd; ++ __u32 tcpi_rehash; ++}; ++ ++struct mptcp_info { ++ __u8 mptcpi_subflows; ++ __u8 mptcpi_add_addr_signal; ++ __u8 mptcpi_add_addr_accepted; ++ __u8 mptcpi_subflows_max; ++ __u8 mptcpi_add_addr_signal_max; ++ __u8 mptcpi_add_addr_accepted_max; ++ __u32 mptcpi_flags; ++ __u32 mptcpi_token; ++ __u64 mptcpi_write_seq; ++ __u64 mptcpi_snd_una; ++ __u64 mptcpi_rcv_nxt; ++ __u8 mptcpi_local_addr_used; ++ __u8 mptcpi_local_addr_max; ++ __u8 mptcpi_csum_enabled; ++}; ++ ++struct mptcp_subflow_data { ++ __u32 size_subflow_data; ++ __u32 num_subflows; ++ __u32 size_kernel; ++ __u32 size_user; ++}; ++ ++struct mptcp_subflow_addrs { ++ union { ++ __kernel_sa_family_t sa_family; ++ struct sockaddr sa_local; ++ struct sockaddr_in sin_local; ++ struct sockaddr_in6 sin6_local; ++ struct __kernel_sockaddr_storage ss_local; ++ }; ++ union { ++ struct sockaddr sa_remote; ++ struct sockaddr_in sin_remote; ++ struct sockaddr_in6 sin6_remote; ++ struct __kernel_sockaddr_storage ss_remote; ++ }; ++}; ++ ++enum mptcp_pm_type { ++ MPTCP_PM_TYPE_KERNEL = 0, ++ MPTCP_PM_TYPE_USERSPACE = 1, ++ __MPTCP_PM_TYPE_NR = 2, ++ __MPTCP_PM_TYPE_MAX = 1, ++}; ++ ++struct join_entry { ++ u32 token; ++ u32 remote_nonce; ++ u32 local_nonce; ++ u8 join_id; ++ u8 local_id; ++ u8 backup; ++ u8 valid; ++}; ++ ++struct mctp_netdev_ops; ++ ++struct mctp_dev { ++ struct net_device *dev; ++ refcount_t refs; ++ unsigned int net; ++ const struct mctp_netdev_ops *ops; ++ u8 *addrs; ++ size_t num_addrs; ++ spinlock_t addrs_lock; ++ struct callback_head rcu; ++}; ++ ++typedef __u8 mctp_eid_t; ++ ++struct mctp_addr { ++ mctp_eid_t s_addr; ++}; ++ ++struct sockaddr_mctp { ++ __kernel_sa_family_t smctp_family; ++ __u16 __smctp_pad0; ++ unsigned int smctp_network; ++ struct mctp_addr smctp_addr; ++ __u8 smctp_type; ++ __u8 smctp_tag; ++ __u8 __smctp_pad1; ++}; ++ ++struct sockaddr_mctp_ext { ++ struct sockaddr_mctp smctp_base; ++ int smctp_ifindex; ++ __u8 smctp_halen; ++ __u8 __smctp_pad0[3]; ++ __u8 smctp_haddr[32]; ++}; ++ ++struct mctp_ioc_tag_ctl { ++ mctp_eid_t peer_addr; ++ __u8 tag; ++ __u16 flags; ++}; ++ ++struct mctp_hdr { ++ u8 ver; ++ u8 dest; ++ u8 src; ++ u8 flags_seq_tag; ++}; ++ ++struct mctp_sock { ++ struct sock sk; ++ unsigned int bind_net; ++ mctp_eid_t bind_addr; ++ __u8 bind_type; ++ bool addr_ext; ++ struct hlist_head keys; ++ struct timer_list key_expiry; ++}; ++ ++struct mctp_sk_key { ++ mctp_eid_t peer_addr; ++ mctp_eid_t local_addr; ++ __u8 tag; ++ struct sock *sk; ++ struct hlist_node hlist; ++ struct hlist_node sklist; ++ spinlock_t lock; ++ refcount_t refs; ++ struct sk_buff *reasm_head; ++ struct sk_buff **reasm_tailp; ++ bool reasm_dead; ++ u8 last_seq; ++ bool valid; ++ long unsigned int expiry; ++ long unsigned int dev_flow_state; ++ struct mctp_dev *dev; ++ bool manual_alloc; ++}; ++ ++struct mctp_skb_cb { ++ unsigned int magic; ++ unsigned int net; ++ int ifindex; ++ mctp_eid_t src; ++ unsigned char halen; ++ unsigned char haddr[32]; ++}; ++ ++struct mctp_route { ++ mctp_eid_t min; ++ mctp_eid_t max; ++ struct mctp_dev *dev; ++ unsigned int mtu; ++ unsigned char type; ++ int (*output)(struct mctp_route *, struct sk_buff *); ++ struct list_head list; ++ refcount_t refs; ++ struct callback_head rcu; ++}; ++ ++struct mctp_netdev_ops { ++ void (*release_flow)(struct mctp_dev *, struct mctp_sk_key *); ++}; ++ ++enum { ++ MCTP_TRACE_KEY_TIMEOUT = 0, ++ MCTP_TRACE_KEY_REPLIED = 1, ++ MCTP_TRACE_KEY_INVALIDATED = 2, ++ MCTP_TRACE_KEY_CLOSED = 3, ++ MCTP_TRACE_KEY_DROPPED = 4, ++}; ++ ++struct trace_event_raw_mctp_key_acquire { ++ struct trace_entry ent; ++ __u8 paddr; ++ __u8 laddr; ++ __u8 tag; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mctp_key_release { ++ struct trace_entry ent; ++ __u8 paddr; ++ __u8 laddr; ++ __u8 tag; ++ int reason; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_mctp_key_acquire {}; ++ ++struct trace_event_data_offsets_mctp_key_release {}; ++ ++typedef void (*btf_trace_mctp_key_acquire)(void *, const struct mctp_sk_key *); ++ ++typedef void (*btf_trace_mctp_key_release)(void *, const struct mctp_sk_key *, int); ++ ++enum mctp_neigh_source { ++ MCTP_NEIGH_STATIC = 0, ++ MCTP_NEIGH_DISCOVER = 1, ++}; ++ ++struct mctp_neigh { ++ struct mctp_dev *dev; ++ mctp_eid_t eid; ++ enum mctp_neigh_source source; ++ unsigned char ha[32]; ++ struct list_head list; ++ struct callback_head rcu; ++}; ++ ++enum { ++ TLS_NO_KEYRING = 0, ++ TLS_NO_PEERID = 0, ++ TLS_NO_CERT = 0, ++ TLS_NO_PRIVKEY = 0, ++}; ++ ++typedef void (*tls_done_func_t)(void *, int, key_serial_t); ++ ++struct tls_handshake_args { ++ struct socket *ta_sock; ++ tls_done_func_t ta_done; ++ void *ta_data; ++ const char *ta_peername; ++ unsigned int ta_timeout_ms; ++ key_serial_t ta_keyring; ++ key_serial_t ta_my_cert; ++ key_serial_t ta_my_privkey; ++ unsigned int ta_num_peerids; ++ key_serial_t ta_my_peerids[5]; ++}; ++ ++enum handshake_handler_class { ++ HANDSHAKE_HANDLER_CLASS_NONE = 0, ++ HANDSHAKE_HANDLER_CLASS_TLSHD = 1, ++ HANDSHAKE_HANDLER_CLASS_MAX = 2, ++}; ++ ++enum handshake_msg_type { ++ HANDSHAKE_MSG_TYPE_UNSPEC = 0, ++ HANDSHAKE_MSG_TYPE_CLIENTHELLO = 1, ++ HANDSHAKE_MSG_TYPE_SERVERHELLO = 2, ++}; ++ ++enum handshake_auth { ++ HANDSHAKE_AUTH_UNSPEC = 0, ++ HANDSHAKE_AUTH_UNAUTH = 1, ++ HANDSHAKE_AUTH_PSK = 2, ++ HANDSHAKE_AUTH_X509 = 3, ++}; ++ ++enum { ++ HANDSHAKE_A_X509_CERT = 1, ++ HANDSHAKE_A_X509_PRIVKEY = 2, ++ __HANDSHAKE_A_X509_MAX = 3, ++ HANDSHAKE_A_X509_MAX = 2, ++}; ++ ++enum { ++ HANDSHAKE_A_ACCEPT_SOCKFD = 1, ++ HANDSHAKE_A_ACCEPT_HANDLER_CLASS = 2, ++ HANDSHAKE_A_ACCEPT_MESSAGE_TYPE = 3, ++ HANDSHAKE_A_ACCEPT_TIMEOUT = 4, ++ HANDSHAKE_A_ACCEPT_AUTH_MODE = 5, ++ HANDSHAKE_A_ACCEPT_PEER_IDENTITY = 6, ++ HANDSHAKE_A_ACCEPT_CERTIFICATE = 7, ++ HANDSHAKE_A_ACCEPT_PEERNAME = 8, ++ __HANDSHAKE_A_ACCEPT_MAX = 9, ++ HANDSHAKE_A_ACCEPT_MAX = 8, ++}; ++ ++enum { ++ HANDSHAKE_A_DONE_STATUS = 1, ++ HANDSHAKE_A_DONE_SOCKFD = 2, ++ HANDSHAKE_A_DONE_REMOTE_AUTH = 3, ++ __HANDSHAKE_A_DONE_MAX = 4, ++ HANDSHAKE_A_DONE_MAX = 3, ++}; ++ ++struct handshake_proto; ++ ++struct handshake_req { ++ struct list_head hr_list; ++ struct rhash_head hr_rhash; ++ long unsigned int hr_flags; ++ const struct handshake_proto *hr_proto; ++ struct sock *hr_sk; ++ void (*hr_odestruct)(struct sock *); ++ char hr_priv[0]; ++}; ++ ++struct handshake_proto { ++ int hp_handler_class; ++ size_t hp_privsize; ++ long unsigned int hp_flags; ++ int (*hp_accept)(struct handshake_req *, struct genl_info *, int); ++ void (*hp_done)(struct handshake_req *, unsigned int, struct genl_info *); ++ void (*hp_destroy)(struct handshake_req *); ++}; ++ ++enum hp_flags_bits { ++ HANDSHAKE_F_PROTO_NOTIFY = 0, ++}; ++ ++struct tls_handshake_req { ++ void (*th_consumer_done)(void *, int, key_serial_t); ++ void *th_consumer_data; ++ int th_type; ++ unsigned int th_timeout_ms; ++ int th_auth_mode; ++ const char *th_peername; ++ key_serial_t th_keyring; ++ key_serial_t th_certificate; ++ key_serial_t th_privkey; ++ unsigned int th_num_peerids; ++ key_serial_t th_peerid[5]; ++}; ++ ++enum css_eval_cond { ++ CSS_EVAL_NO_PATH = 0, ++ CSS_EVAL_NOT_ONLINE = 1, ++}; ++ ++enum dev_state { ++ DEV_STATE_NOT_OPER = 0, ++ DEV_STATE_SENSE_ID = 1, ++ DEV_STATE_OFFLINE = 2, ++ DEV_STATE_VERIFY = 3, ++ DEV_STATE_ONLINE = 4, ++ DEV_STATE_W4SENSE = 5, ++ DEV_STATE_DISBAND_PGID = 6, ++ DEV_STATE_BOXED = 7, ++ DEV_STATE_TIMEOUT_KILL = 8, ++ DEV_STATE_QUIESCE = 9, ++ DEV_STATE_DISCONNECTED = 10, ++ DEV_STATE_DISCONNECTED_SENSE_ID = 11, ++ DEV_STATE_CMFCHANGE = 12, ++ DEV_STATE_CMFUPDATE = 13, ++ DEV_STATE_STEAL_LOCK = 14, ++ NR_DEV_STATES = 15, ++}; ++ ++enum dev_event { ++ DEV_EVENT_NOTOPER = 0, ++ DEV_EVENT_INTERRUPT = 1, ++ DEV_EVENT_TIMEOUT = 2, ++ DEV_EVENT_VERIFY = 3, ++ NR_DEV_EVENTS = 4, ++}; ++ ++typedef enum { ++ add = 0, ++ free = 1, ++} range_action; ++ ++struct ccwdev_iter { ++ int devno; ++ int ssid; ++ int in_range; ++}; ++ ++struct ccw_device_id { ++ __u16 match_flags; ++ __u16 cu_type; ++ __u16 dev_type; ++ __u8 cu_model; ++ __u8 dev_model; ++ kernel_ulong_t driver_info; ++}; ++ ++struct css_device_id { ++ __u8 match_flags; ++ __u8 type; ++ kernel_ulong_t driver_data; ++}; ++ ++struct chp_id { ++ __u8 reserved1; ++ __u8 cssid; ++ __u8 reserved2; ++ __u8 id; ++}; ++ ++struct gen_pool; ++ ++typedef long unsigned int (*genpool_algo_t)(long unsigned int *, long unsigned int, long unsigned int, unsigned int, void *, struct gen_pool *, long unsigned int); ++ ++struct gen_pool { ++ spinlock_t lock; ++ struct list_head chunks; ++ int min_alloc_order; ++ genpool_algo_t algo; ++ void *data; ++ const char *name; ++}; ++ ++struct css_general_char { ++ short: 12; ++ u64 dynio: 1; ++ short: 3; ++ char: 1; ++ u64 eadm: 1; ++ int: 14; ++ short: 9; ++ u64 aif: 1; ++ char: 3; ++ u64 mcss: 1; ++ u64 fcs: 1; ++ short: 1; ++ u64 ext_mb: 1; ++ char: 7; ++ u64 aif_tdd: 1; ++ char: 1; ++ u64 qebsm: 1; ++ char: 2; ++ u64 aiv: 1; ++ long: 2; ++ char: 3; ++ u64 aif_osa: 1; ++ short: 12; ++ u64 eadm_rf: 1; ++ char: 1; ++ u64 cib: 1; ++ char: 5; ++ u64 fcx: 1; ++ int: 7; ++ short: 12; ++ u64 alt_ssi: 1; ++ char: 1; ++ u64 narf: 1; ++ short: 1; ++ char: 4; ++ u64 enarf: 1; ++ char: 3; ++ char: 3; ++ u64 util_str: 1; ++}; ++ ++struct cmd_scsw { ++ __u32 key: 4; ++ __u32 sctl: 1; ++ __u32 eswf: 1; ++ __u32 cc: 2; ++ __u32 fmt: 1; ++ __u32 pfch: 1; ++ __u32 isic: 1; ++ __u32 alcc: 1; ++ __u32 ssi: 1; ++ __u32 zcc: 1; ++ __u32 ectl: 1; ++ __u32 pno: 1; ++ __u32 res: 1; ++ __u32 fctl: 3; ++ __u32 actl: 7; ++ __u32 stctl: 5; ++ __u32 cpa; ++ __u32 dstat: 8; ++ __u32 cstat: 8; ++ __u32 count: 16; ++}; ++ ++struct tm_scsw { ++ u32 key: 4; ++ char: 1; ++ u32 eswf: 1; ++ u32 cc: 2; ++ u32 fmt: 3; ++ u32 x: 1; ++ u32 q: 1; ++ char: 1; ++ u32 ectl: 1; ++ u32 pno: 1; ++ char: 1; ++ u32 fctl: 3; ++ u32 actl: 7; ++ u32 stctl: 5; ++ u32 tcw; ++ u32 dstat: 8; ++ u32 cstat: 8; ++ u32 fcxs: 8; ++ u32 ifob: 1; ++ u32 sesq: 7; ++}; ++ ++struct eadm_scsw { ++ u32 key: 4; ++ char: 1; ++ u32 eswf: 1; ++ u32 cc: 2; ++ char: 6; ++ u32 ectl: 1; ++ short: 1; ++ char: 1; ++ u32 fctl: 3; ++ u32 actl: 7; ++ u32 stctl: 5; ++ u32 aob; ++ u32 dstat: 8; ++ u32 cstat: 8; ++}; ++ ++union scsw { ++ struct cmd_scsw cmd; ++ struct tm_scsw tm; ++ struct eadm_scsw eadm; ++}; ++ ++struct ccw1 { ++ __u8 cmd_code; ++ __u8 flags; ++ __u16 count; ++ __u32 cda; ++}; ++ ++struct erw { ++ __u32 res0: 3; ++ __u32 auth: 1; ++ __u32 pvrf: 1; ++ __u32 cpt: 1; ++ __u32 fsavf: 1; ++ __u32 cons: 1; ++ __u32 scavf: 1; ++ __u32 fsaf: 1; ++ __u32 scnt: 6; ++ __u32 res16: 16; ++}; ++ ++struct erw_eadm { ++ short: 16; ++ __u32 b: 1; ++ __u32 r: 1; ++}; ++ ++struct sublog { ++ __u32 res0: 1; ++ __u32 esf: 7; ++ __u32 lpum: 8; ++ __u32 arep: 1; ++ __u32 fvf: 5; ++ __u32 sacc: 2; ++ __u32 termc: 2; ++ __u32 devsc: 1; ++ __u32 serr: 1; ++ __u32 ioerr: 1; ++ __u32 seqc: 3; ++}; ++ ++struct esw0 { ++ struct sublog sublog; ++ struct erw erw; ++ __u32 faddr[2]; ++ __u32 saddr; ++}; ++ ++struct esw1 { ++ __u8 zero0; ++ __u8 lpum; ++ __u16 zero16; ++ struct erw erw; ++ __u32 zeros[3]; ++}; ++ ++struct esw2 { ++ __u8 zero0; ++ __u8 lpum; ++ __u16 dcti; ++ struct erw erw; ++ __u32 zeros[3]; ++}; ++ ++struct esw3 { ++ __u8 zero0; ++ __u8 lpum; ++ __u16 res; ++ struct erw erw; ++ __u32 zeros[3]; ++}; ++ ++struct esw_eadm { ++ __u32 sublog; ++ struct erw_eadm erw; ++ long: 64; ++ int: 32; ++}; ++ ++struct irb { ++ union scsw scsw; ++ union { ++ struct esw0 esw0; ++ struct esw1 esw1; ++ struct esw2 esw2; ++ struct esw3 esw3; ++ struct esw_eadm eadm; ++ } esw; ++ __u8 ecw[32]; ++}; ++ ++struct ciw { ++ __u32 et: 2; ++ __u32 reserved: 2; ++ __u32 ct: 4; ++ __u32 cmd: 8; ++ __u32 count: 16; ++}; ++ ++struct tcw { ++ u32 format: 2; ++ char: 6; ++ u32 flags: 24; ++ char: 8; ++ u32 tccbl: 6; ++ u32 r: 1; ++ u32 w: 1; ++ u64 output; ++ u64 input; ++ u64 tsb; ++ u64 tccb; ++ u32 output_count; ++ u32 input_count; ++ long: 64; ++ int: 32; ++ u32 intrg; ++}; ++ ++struct ccw_device_private; ++ ++struct ccw_driver; ++ ++struct ccw_device { ++ spinlock_t *ccwlock; ++ struct ccw_device_private *private; ++ struct mutex reg_mutex; ++ struct ccw_device_id id; ++ struct ccw_driver *drv; ++ struct device dev; ++ int online; ++ void (*handler)(struct ccw_device *, long unsigned int, struct irb *); ++}; ++ ++enum io_status { ++ IO_DONE = 0, ++ IO_RUNNING = 1, ++ IO_STATUS_ERROR = 2, ++ IO_PATH_ERROR = 3, ++ IO_REJECTED = 4, ++ IO_KILLED = 5, ++}; ++ ++struct ccw_request { ++ struct ccw1 *cp; ++ long unsigned int timeout; ++ u16 maxretries; ++ u8 lpm; ++ int (*check)(struct ccw_device *, void *); ++ enum io_status (*filter)(struct ccw_device *, void *, struct irb *, enum io_status); ++ void (*callback)(struct ccw_device *, void *, int); ++ void *data; ++ unsigned int singlepath: 1; ++ unsigned int cancel: 1; ++ unsigned int done: 1; ++ u16 mask; ++ u16 retries; ++ int drc; ++} __attribute__((packed)); ++ ++struct qdio_irq; ++ ++enum cdev_todo { ++ CDEV_TODO_NOTHING = 0, ++ CDEV_TODO_ENABLE_CMF = 1, ++ CDEV_TODO_REBIND = 2, ++ CDEV_TODO_REGISTER = 3, ++ CDEV_TODO_UNREG = 4, ++ CDEV_TODO_UNREG_EVAL = 5, ++}; ++ ++struct subchannel; ++ ++struct ccw_device_dma_area; ++ ++struct ccw_device_private { ++ struct ccw_device *cdev; ++ struct subchannel *sch; ++ int state; ++ atomic_t onoff; ++ struct ccw_dev_id dev_id; ++ struct ccw_request req; ++ int iretry; ++ u8 pgid_valid_mask; ++ u8 pgid_todo_mask; ++ u8 pgid_reset_mask; ++ u8 path_noirq_mask; ++ u8 path_notoper_mask; ++ u8 path_gone_mask; ++ u8 path_new_mask; ++ u8 path_broken_mask; ++ struct { ++ unsigned int fast: 1; ++ unsigned int repall: 1; ++ unsigned int pgroup: 1; ++ unsigned int force: 1; ++ unsigned int mpath: 1; ++ } __attribute__((packed)) options; ++ struct { ++ unsigned int esid: 1; ++ unsigned int dosense: 1; ++ unsigned int doverify: 1; ++ unsigned int donotify: 1; ++ unsigned int recog_done: 1; ++ unsigned int fake_irb: 2; ++ unsigned int pgroup: 1; ++ unsigned int mpath: 1; ++ unsigned int pgid_unknown: 1; ++ unsigned int initialized: 1; ++ } __attribute__((packed)) flags; ++ long unsigned int intparm; ++ struct qdio_irq *qdio_data; ++ int async_kill_io_rc; ++ struct work_struct todo_work; ++ enum cdev_todo todo; ++ wait_queue_head_t wait_q; ++ struct timer_list timer; ++ void *cmb; ++ struct list_head cmb_list; ++ u64 cmb_start_time; ++ void *cmb_wait; ++ struct gen_pool *dma_pool; ++ struct ccw_device_dma_area *dma_area; ++ enum interruption_class int_class; ++}; ++ ++enum uc_todo { ++ UC_TODO_RETRY = 0, ++ UC_TODO_RETRY_ON_NEW_PATH = 1, ++ UC_TODO_STOP = 2, ++}; ++ ++struct ccw_driver { ++ struct ccw_device_id *ids; ++ int (*probe)(struct ccw_device *); ++ void (*remove)(struct ccw_device *); ++ int (*set_online)(struct ccw_device *); ++ int (*set_offline)(struct ccw_device *); ++ int (*notify)(struct ccw_device *, int); ++ void (*path_event)(struct ccw_device *, int *); ++ void (*shutdown)(struct ccw_device *); ++ enum uc_todo (*uc_handler)(struct ccw_device *, struct irb *); ++ struct device_driver driver; ++ enum interruption_class int_class; ++}; ++ ++struct chsc_ssd_info { ++ u8 path_mask; ++ u8 fla_valid_mask; ++ struct chp_id chpid[8]; ++ u16 fla[8]; ++}; ++ ++struct pmcw { ++ u32 intparm; ++ u32 qf: 1; ++ u32 w: 1; ++ u32 isc: 3; ++ u32 res5: 3; ++ u32 ena: 1; ++ u32 lm: 2; ++ u32 mme: 2; ++ u32 mp: 1; ++ u32 tf: 1; ++ u32 dnv: 1; ++ u32 dev: 16; ++ u8 lpm; ++ u8 pnom; ++ u8 lpum; ++ u8 pim; ++ u16 mbi; ++ u8 pom; ++ u8 pam; ++ u8 chpid[8]; ++ u32 unused1: 8; ++ u32 st: 3; ++ u32 unused2: 18; ++ u32 mbfc: 1; ++ u32 xmwme: 1; ++ u32 csense: 1; ++}; ++ ++struct schib_config { ++ u64 mba; ++ u32 intparm; ++ u16 mbi; ++ u32 isc: 3; ++ u32 ena: 1; ++ u32 mme: 2; ++ u32 mp: 1; ++ u32 csense: 1; ++ u32 mbfc: 1; ++}; ++ ++struct schib { ++ struct pmcw pmcw; ++ union scsw scsw; ++ __u64 mba; ++ __u8 mda[4]; ++} __attribute__((packed)); ++ ++enum sch_todo { ++ SCH_TODO_NOTHING = 0, ++ SCH_TODO_EVAL = 1, ++ SCH_TODO_UNREG = 2, ++}; ++ ++struct css_driver; ++ ++struct subchannel { ++ struct subchannel_id schid; ++ spinlock_t *lock; ++ struct mutex reg_mutex; ++ enum { ++ SUBCHANNEL_TYPE_IO = 0, ++ SUBCHANNEL_TYPE_CHSC = 1, ++ SUBCHANNEL_TYPE_MSG = 2, ++ SUBCHANNEL_TYPE_ADM = 3, ++ } st; ++ __u8 vpm; ++ __u8 lpm; ++ __u8 opm; ++ long: 0; ++ struct schib schib; ++ int isc; ++ struct chsc_ssd_info ssd_info; ++ struct device dev; ++ struct css_driver *driver; ++ enum sch_todo todo; ++ struct work_struct todo_work; ++ struct schib_config config; ++ u64 dma_mask; ++ const char *driver_override; ++}; ++ ++struct chp_link; ++ ++struct css_driver { ++ struct css_device_id *subchannel_type; ++ struct device_driver drv; ++ void (*irq)(struct subchannel *); ++ int (*chp_event)(struct subchannel *, struct chp_link *, int); ++ int (*sch_event)(struct subchannel *, int); ++ int (*probe)(struct subchannel *); ++ void (*remove)(struct subchannel *); ++ void (*shutdown)(struct subchannel *); ++ int (*settle)(); ++}; ++ ++struct path_state { ++ __u8 state1: 2; ++ __u8 state2: 2; ++ __u8 state3: 1; ++ __u8 resvd: 3; ++}; ++ ++struct extended_cssid { ++ u8 version; ++ u8 cssid; ++}; ++ ++struct pgid { ++ union { ++ __u8 fc; ++ struct path_state ps; ++ } inf; ++ union { ++ __u32 cpu_addr: 16; ++ struct extended_cssid ext_cssid; ++ } pgid_high; ++ __u32 cpu_id: 24; ++ __u32 cpu_model: 16; ++ __u32 tod_high; ++}; ++ ++struct chp_link { ++ struct chp_id chpid; ++ u32 fla_mask; ++ u16 fla; ++}; ++ ++struct cmd_orb { ++ u32 intparm; ++ u32 key: 4; ++ u32 spnd: 1; ++ u32 res1: 1; ++ u32 mod: 1; ++ u32 sync: 1; ++ u32 fmt: 1; ++ u32 pfch: 1; ++ u32 isic: 1; ++ u32 alcc: 1; ++ u32 ssic: 1; ++ u32 res2: 1; ++ u32 c64: 1; ++ u32 i2k: 1; ++ u32 lpm: 8; ++ u32 ils: 1; ++ u32 zero: 6; ++ u32 orbx: 1; ++ u32 cpa; ++}; ++ ++struct tm_orb { ++ u32 intparm; ++ u32 key: 4; ++ char: 4; ++ char: 5; ++ u32 b: 1; ++ short: 2; ++ u32 lpm: 8; ++ char: 7; ++ u32 x: 1; ++ u32 tcw; ++ u32 prio: 8; ++ short: 8; ++ u32 rsvpgm: 8; ++ long: 64; ++ long: 64; ++}; ++ ++struct eadm_orb { ++ u32 intparm; ++ u32 key: 4; ++ char: 4; ++ u32 compat1: 1; ++ u32 compat2: 1; ++ short: 6; ++ short: 15; ++ u32 x: 1; ++ u32 aob; ++ u32 css_prio: 8; ++ short: 8; ++ u32 scm_prio: 8; ++ long: 8; ++ int: 29; ++ u32 fmt: 3; ++ long: 64; ++}; ++ ++union orb { ++ struct cmd_orb cmd; ++ struct tm_orb tm; ++ struct eadm_orb eadm; ++}; ++ ++struct io_subchannel_dma_area { ++ struct ccw1 sense_ccw; ++}; ++ ++struct io_subchannel_private { ++ union orb orb; ++ struct ccw_device *cdev; ++ struct { ++ unsigned int suspend: 1; ++ unsigned int prefetch: 1; ++ unsigned int inter: 1; ++ } __attribute__((packed)) options; ++ struct io_subchannel_dma_area *dma_area; ++ dma_addr_t dma_area_dma; ++}; ++ ++struct senseid { ++ u8 reserved; ++ u16 cu_type; ++ u8 cu_model; ++ u16 dev_type; ++ u8 dev_model; ++ u8 unused; ++ struct ciw ciw[8]; ++} __attribute__((packed)); ++ ++struct ccw_device_dma_area { ++ struct senseid senseid; ++ struct ccw1 iccws[2]; ++ struct irb irb; ++ struct pgid pgid[8]; ++}; ++ ++struct idset { ++ int num_ssid; ++ int num_id; ++ long unsigned int bitmap[0]; ++}; ++ ++struct tidaw { ++ u32 flags: 8; ++ u32 count; ++ u64 addr; ++}; ++ ++struct tsa_iostat { ++ u32 dev_time; ++ u32 def_time; ++ u32 queue_time; ++ u32 dev_busy_time; ++ u32 dev_act_time; ++ u8 sense[32]; ++}; ++ ++struct tsa_ddpc { ++ int: 24; ++ u32 rc: 8; ++ u8 rcq[16]; ++ u8 sense[32]; ++}; ++ ++struct tsa_intrg { ++ u32 format: 8; ++ u32 flags: 8; ++ u32 cu_state: 8; ++ u32 dev_state: 8; ++ u32 op_state: 8; ++ long: 0; ++ u8 sd_info[12]; ++ u32 dl_id; ++ u8 dd_data[28]; ++}; ++ ++struct tsb { ++ u32 length: 8; ++ u32 flags: 8; ++ u32 dcw_offset: 16; ++ u32 count; ++ int: 32; ++ union { ++ struct tsa_iostat iostat; ++ struct tsa_ddpc ddpc; ++ struct tsa_intrg intrg; ++ } tsa; ++}; ++ ++struct dcw { ++ u32 cmd: 8; ++ u32 flags: 8; ++ char: 8; ++ u32 cd_count: 8; ++ u32 count; ++ u8 cd[0]; ++}; ++ ++struct tccb_tcah { ++ u32 format: 8; ++ int: 24; ++ int: 24; ++ u32 tcal: 8; ++ u32 sac: 16; ++ char: 8; ++ u32 prio: 8; ++ long: 0; ++}; ++ ++struct tccb_tcat { ++ int: 32; ++ u32 count; ++}; ++ ++struct tccb { ++ struct tccb_tcah tcah; ++ u8 tca[0]; ++}; ++ ++struct itcw { ++ struct tcw *tcw; ++ struct tcw *intrg_tcw; ++ int num_tidaws; ++ int max_tidaws; ++ int intrg_num_tidaws; ++ int intrg_max_tidaws; ++}; ++ ++enum dpm_order { ++ DPM_ORDER_NONE = 0, ++ DPM_ORDER_DEV_AFTER_PARENT = 1, ++ DPM_ORDER_PARENT_BEFORE_DEV = 2, ++ DPM_ORDER_DEV_LAST = 3, ++}; ++ ++struct channel_path_desc_fmt0 { ++ u8 flags; ++ u8 lsn; ++ u8 desc; ++ u8 chpid; ++ u8 swla; ++ u8 zeroes; ++ u8 chla; ++ u8 chpp; ++}; ++ ++struct cmg_chars { ++ u32 values[5]; ++}; ++ ++struct channel_path_desc_fmt1 { ++ u8 flags; ++ u8 lsn; ++ u8 desc; ++ u8 chpid; ++ short: 16; ++ u8 esc; ++ u8 chpp; ++ u32 unused[2]; ++ u16 chid; ++ int: 0; ++ u16 mdc; ++ short: 13; ++ u8 r: 1; ++ u8 s: 1; ++ u8 f: 1; ++ u32 zeros[2]; ++}; ++ ++struct channel_path_desc_fmt3 { ++ struct channel_path_desc_fmt1 fmt1_desc; ++ u8 util_str[64]; ++}; ++ ++struct channel_path; ++ ++struct channel_subsystem { ++ u8 cssid; ++ u8 iid; ++ bool id_valid; ++ struct channel_path *chps[256]; ++ struct device device; ++ struct pgid global_pgid; ++ struct mutex mutex; ++ int cm_enabled; ++ void *cub_addr1; ++ void *cub_addr2; ++ struct subchannel *pseudo_subchannel; ++}; ++ ++struct channel_path { ++ struct device dev; ++ struct chp_id chpid; ++ struct mutex lock; ++ int state; ++ struct channel_path_desc_fmt0 desc; ++ struct channel_path_desc_fmt1 desc_fmt1; ++ struct channel_path_desc_fmt3 desc_fmt3; ++ int cmg; ++ int shared; ++ struct cmg_chars cmg_chars; ++}; ++ ++typedef void fsm_func_t(struct ccw_device *, enum dev_event); ++ ++enum io_sch_action { ++ IO_SCH_UNREG = 0, ++ IO_SCH_ORPH_UNREG = 1, ++ IO_SCH_UNREG_CDEV = 2, ++ IO_SCH_ATTACH = 3, ++ IO_SCH_UNREG_ATTACH = 4, ++ IO_SCH_ORPH_ATTACH = 5, ++ IO_SCH_REPROBE = 6, ++ IO_SCH_VERIFY = 7, ++ IO_SCH_DISC = 8, ++ IO_SCH_NOP = 9, ++}; ++ ++struct stlck_data { ++ struct completion done; ++ int rc; ++}; ++ ++struct sysrq_work { ++ int key; ++ struct work_struct work; ++}; ++ ++typedef unsigned int sclp_cmdw_t; ++ ++typedef u64 sccb_mask_t; ++ ++struct sccb_header { ++ u16 length; ++ u8 function_code; ++ u8 control_mask[3]; ++ u16 response_code; ++}; ++ ++struct evbuf_header { ++ u16 length; ++ u8 type; ++ u8 flags; ++ u16 _reserved; ++}; ++ ++struct sclp_req { ++ struct list_head list; ++ sclp_cmdw_t command; ++ void *sccb; ++ char status; ++ int start_count; ++ void (*callback)(struct sclp_req *, void *); ++ void *callback_data; ++ int queue_timeout; ++ long unsigned int queue_expires; ++}; ++ ++struct sclp_register { ++ struct list_head list; ++ sccb_mask_t receive_mask; ++ sccb_mask_t send_mask; ++ sccb_mask_t sclp_receive_mask; ++ sccb_mask_t sclp_send_mask; ++ void (*state_change_fn)(struct sclp_register *); ++ void (*receiver_fn)(struct evbuf_header *); ++}; ++ ++struct mto { ++ u16 length; ++ u16 type; ++ u16 line_type_flags; ++ u8 alarm_control; ++ u8 _reserved[3]; ++}; ++ ++struct go { ++ u16 length; ++ u16 type; ++ u32 domid; ++ u8 hhmmss_time[8]; ++ u8 th_time[3]; ++ u8 reserved_0; ++ u8 dddyyyy_date[7]; ++ u8 _reserved_1; ++ u16 general_msg_flags; ++ u8 _reserved_2[10]; ++ u8 originating_system_name[8]; ++ u8 job_guest_name[8]; ++}; ++ ++struct mdb_header { ++ u16 length; ++ u16 type; ++ u32 tag; ++ u32 revision_code; ++}; ++ ++struct mdb { ++ struct mdb_header header; ++ struct go go; ++ struct mto mto; ++} __attribute__((packed)); ++ ++struct msg_buf { ++ struct evbuf_header header; ++ struct mdb mdb; ++}; ++ ++struct sclp_buffer { ++ struct list_head list; ++ struct sclp_req request; ++ void *sccb; ++ struct msg_buf *current_msg; ++ char *current_line; ++ int current_length; ++ int retry_count; ++ short unsigned int columns; ++ short unsigned int htab; ++ unsigned int char_sum; ++ unsigned int messages; ++ void (*callback)(struct sclp_buffer *, int); ++}; ++ ++typedef int mhp_t; ++ ++struct sclp_chp_info { ++ u8 recognized[32]; ++ u8 standby[32]; ++ u8 configured[32]; ++}; ++ ++struct sclp_core_entry { ++ u8 core_id; ++ u8 reserved0; ++ char: 4; ++ u8 sief2: 1; ++ u8 skey: 1; ++ char: 2; ++ char: 2; ++ u8 gpere: 1; ++ u8 siif: 1; ++ u8 sigpif: 1; ++ u8 reserved2[3]; ++ char: 2; ++ u8 ib: 1; ++ u8 cei: 1; ++ u8 reserved3[6]; ++ u8 type; ++ u8 reserved1; ++}; ++ ++struct sclp_core_info { ++ unsigned int configured; ++ unsigned int standby; ++ unsigned int combined; ++ struct sclp_core_entry core[512]; ++}; ++ ++struct read_cpu_info_sccb { ++ struct sccb_header header; ++ u16 nr_configured; ++ u16 offset_configured; ++ u16 nr_standby; ++ u16 offset_standby; ++ u8 reserved[0]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct read_storage_sccb { ++ struct sccb_header header; ++ u16 max_id; ++ u16 assigned; ++ u16 standby; ++ u32 entries[0]; ++}; ++ ++struct cpu_configure_sccb { ++ struct sccb_header header; ++}; ++ ++struct memory_increment { ++ struct list_head list; ++ u16 rn; ++ int standby; ++}; ++ ++struct assign_storage_sccb { ++ struct sccb_header header; ++ u16 rn; ++}; ++ ++struct attach_storage_sccb { ++ struct sccb_header header; ++ short: 16; ++ u16 assigned; ++ long: 0; ++ u32 entries[0]; ++}; ++ ++struct chp_cfg_sccb { ++ struct sccb_header header; ++ u8 ccm; ++ u8 reserved[6]; ++ u8 cssid; ++}; ++ ++struct chp_info_sccb { ++ struct sccb_header header; ++ u8 recognized[32]; ++ u8 standby[32]; ++ u8 configured[32]; ++ u8 ccm; ++ u8 reserved[6]; ++ u8 cssid; ++}; ++ ++struct gds_subvector { ++ u8 length; ++ u8 key; ++}; ++ ++struct gds_vector { ++ u16 length; ++ u16 gds_id; ++}; ++ ++struct sclp_sd_evbuf { ++ struct evbuf_header hdr; ++ u8 eq; ++ u8 di; ++ u8 rflags; ++ long: 0; ++ u32 id; ++ short: 16; ++ u8 fmt; ++ u8 status; ++ u64 sat; ++ u64 sa; ++ u32 esize; ++ u32 dsize; ++}; ++ ++struct sclp_sd_sccb { ++ struct sccb_header hdr; ++ struct sclp_sd_evbuf evbuf; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct sclp_sd_data { ++ size_t esize_bytes; ++ size_t dsize_bytes; ++ void *data; ++}; ++ ++struct sclp_sd_listener { ++ struct list_head list; ++ u32 id; ++ struct completion completion; ++ struct sclp_sd_evbuf evbuf; ++}; ++ ++struct sclp_sd_file { ++ struct kobject kobj; ++ struct bin_attribute data_attr; ++ struct mutex data_mutex; ++ struct sclp_sd_data data; ++ u8 di; ++}; ++ ++enum raw3215_type { ++ RAW3215_FREE = 0, ++ RAW3215_READ = 1, ++ RAW3215_WRITE = 2, ++}; ++ ++struct raw3215_info; ++ ++struct raw3215_req { ++ enum raw3215_type type; ++ int start; ++ int len; ++ int delayable; ++ int residual; ++ long: 0; ++ struct ccw1 ccws[3]; ++ struct raw3215_info *info; ++ struct raw3215_req *next; ++}; ++ ++struct raw3215_info { ++ struct tty_port port; ++ struct ccw_device *cdev; ++ spinlock_t *lock; ++ int flags; ++ char *buffer; ++ char *inbuf; ++ int head; ++ int count; ++ int written; ++ struct raw3215_req *queued_read; ++ struct raw3215_req *queued_write; ++ wait_queue_head_t empty_wait; ++ struct timer_list timer; ++ int line_pos; ++ char ubuffer[80]; ++}; ++ ++struct sclp_vt220_request { ++ struct list_head list; ++ struct sclp_req sclp_req; ++ int retry_count; ++}; ++ ++struct sclp_vt220_sccb { ++ struct sccb_header header; ++ struct evbuf_header evbuf; ++}; ++ ++struct bus_attribute { ++ struct attribute attr; ++ ssize_t (*show)(const struct bus_type *, char *); ++ ssize_t (*store)(const struct bus_type *, const char *, size_t); ++}; ++ ++struct ap_device_id { ++ __u16 match_flags; ++ __u8 dev_type; ++ kernel_ulong_t driver_info; ++}; ++ ++typedef unsigned int ap_qid_t; ++ ++struct ap_queue_status { ++ unsigned int queue_empty: 1; ++ unsigned int replies_waiting: 1; ++ unsigned int queue_full: 1; ++ char: 3; ++ unsigned int async: 1; ++ unsigned int irq_enabled: 1; ++ unsigned int response_code: 8; ++}; ++ ++union ap_queue_status_reg { ++ long unsigned int value; ++ struct { ++ u32 _pad; ++ struct ap_queue_status status; ++ }; ++}; ++ ++struct ap_tapq_gr2 { ++ union { ++ long unsigned int value; ++ struct { ++ unsigned int fac: 32; ++ unsigned int apinfo: 32; ++ }; ++ struct { ++ unsigned int s: 1; ++ unsigned int m: 1; ++ unsigned int c: 1; ++ unsigned int mode: 3; ++ unsigned int n: 1; ++ char: 1; ++ unsigned int class: 8; ++ unsigned int bs: 2; ++ int: 14; ++ unsigned int at: 8; ++ unsigned int nd: 8; ++ char: 4; ++ unsigned int ml: 4; ++ char: 4; ++ unsigned int qd: 4; ++ }; ++ }; ++}; ++ ++struct ap_config_info { ++ unsigned int apsc: 1; ++ unsigned int apxa: 1; ++ unsigned int qact: 1; ++ unsigned int rc8a: 1; ++ char: 4; ++ unsigned int apsb: 1; ++ int: 0; ++ unsigned char na; ++ unsigned char nd; ++ unsigned char _reserved0[10]; ++ unsigned int apm[8]; ++ unsigned int aqm[8]; ++ unsigned int adm[8]; ++ unsigned char _reserved1[16]; ++}; ++ ++union ap_qact_ap_info { ++ long unsigned int val; ++ struct { ++ char: 3; ++ unsigned int mode: 3; ++ int: 26; ++ unsigned int cat: 8; ++ short: 0; ++ unsigned char ver[2]; ++ }; ++}; ++ ++enum ap_sm_state { ++ AP_SM_STATE_RESET_START = 0, ++ AP_SM_STATE_RESET_WAIT = 1, ++ AP_SM_STATE_SETIRQ_WAIT = 2, ++ AP_SM_STATE_IDLE = 3, ++ AP_SM_STATE_WORKING = 4, ++ AP_SM_STATE_QUEUE_FULL = 5, ++ AP_SM_STATE_ASSOC_WAIT = 6, ++ NR_AP_SM_STATES = 7, ++}; ++ ++enum ap_sm_event { ++ AP_SM_EVENT_POLL = 0, ++ AP_SM_EVENT_TIMEOUT = 1, ++ NR_AP_SM_EVENTS = 2, ++}; ++ ++enum ap_sm_wait { ++ AP_SM_WAIT_AGAIN = 0, ++ AP_SM_WAIT_HIGH_TIMEOUT = 1, ++ AP_SM_WAIT_LOW_TIMEOUT = 2, ++ AP_SM_WAIT_INTERRUPT = 3, ++ AP_SM_WAIT_NONE = 4, ++ NR_AP_SM_WAIT = 5, ++}; ++ ++enum ap_dev_state { ++ AP_DEV_STATE_UNINITIATED = 0, ++ AP_DEV_STATE_OPERATING = 1, ++ AP_DEV_STATE_SHUTDOWN = 2, ++ AP_DEV_STATE_ERROR = 3, ++ NR_AP_DEV_STATES = 4, ++}; ++ ++struct ap_device; ++ ++struct ap_driver { ++ struct device_driver driver; ++ struct ap_device_id *ids; ++ unsigned int flags; ++ int (*probe)(struct ap_device *); ++ void (*remove)(struct ap_device *); ++ int (*in_use)(long unsigned int *, long unsigned int *); ++ void (*on_config_changed)(struct ap_config_info *, struct ap_config_info *); ++ void (*on_scan_complete)(struct ap_config_info *, struct ap_config_info *); ++}; ++ ++struct ap_device { ++ struct device device; ++ int device_type; ++}; ++ ++struct ap_card { ++ struct ap_device ap_dev; ++ int raw_hwtype; ++ unsigned int functions; ++ int queue_depth; ++ int id; ++ unsigned int maxmsgsize; ++ bool config; ++ bool chkstop; ++ atomic64_t total_request_count; ++}; ++ ++struct ap_message; ++ ++struct ap_queue { ++ struct ap_device ap_dev; ++ struct hlist_node hnode; ++ struct ap_card *card; ++ spinlock_t lock; ++ enum ap_dev_state dev_state; ++ bool config; ++ bool chkstop; ++ ap_qid_t qid; ++ bool interrupt; ++ unsigned int assoc_idx; ++ int queue_count; ++ int pendingq_count; ++ int requestq_count; ++ u64 total_request_count; ++ int request_timeout; ++ struct timer_list timeout; ++ struct list_head pendingq; ++ struct list_head requestq; ++ struct ap_message *reply; ++ enum ap_sm_state sm_state; ++ int rapq_fbit; ++ int last_err_rc; ++}; ++ ++struct ap_fi { ++ union { ++ u16 cmd; ++ struct { ++ u8 flags; ++ u8 action; ++ }; ++ }; ++}; ++ ++struct ap_message { ++ struct list_head list; ++ long unsigned int psmid; ++ void *msg; ++ size_t len; ++ size_t bufsize; ++ u16 flags; ++ struct ap_fi fi; ++ int rc; ++ void *private; ++ void (*receive)(struct ap_queue *, struct ap_message *, struct ap_message *); ++}; ++ ++struct ap_perms { ++ long unsigned int ioctlm[4]; ++ long unsigned int apm[4]; ++ long unsigned int aqm[4]; ++ long unsigned int adm[4]; ++}; ++ ++struct __ap_calc_ctrs { ++ unsigned int apqns; ++ unsigned int bound; ++}; ++ ++struct compress_format { ++ unsigned char magic[2]; ++ const char *name; ++ decompress_fn decompressor; ++}; ++ ++struct group_data { ++ int limit[21]; ++ int base[20]; ++ int permute[258]; ++ int minLen; ++ int maxLen; ++}; ++ ++struct bunzip_data { ++ int writeCopies; ++ int writePos; ++ int writeRunCountdown; ++ int writeCount; ++ int writeCurrent; ++ long int (*fill)(void *, long unsigned int); ++ long int inbufCount; ++ long int inbufPos; ++ unsigned char *inbuf; ++ unsigned int inbufBitCount; ++ unsigned int inbufBits; ++ unsigned int crc32Table[256]; ++ unsigned int headerCRC; ++ unsigned int totalCRC; ++ unsigned int writeCRC; ++ unsigned int *dbuf; ++ unsigned int dbufSize; ++ unsigned char selectors[32768]; ++ struct group_data groups[6]; ++ int io_error; ++ int byteCount[256]; ++ unsigned char symToByte[256]; ++ unsigned char mtfSymbol[256]; ++}; ++ ++struct rc { ++ long int (*fill)(void *, long unsigned int); ++ uint8_t *ptr; ++ uint8_t *buffer; ++ uint8_t *buffer_end; ++ long int buffer_size; ++ uint32_t code; ++ uint32_t range; ++ uint32_t bound; ++ void (*error)(char *); ++}; ++ ++struct lzma_header { ++ uint8_t pos; ++ uint32_t dict_size; ++ uint64_t dst_size; ++} __attribute__((packed)); ++ ++struct writer { ++ uint8_t *buffer; ++ uint8_t previous_byte; ++ size_t buffer_pos; ++ int bufsize; ++ size_t global_pos; ++ long int (*flush)(void *, long unsigned int); ++ struct lzma_header *header; ++}; ++ ++struct cstate { ++ int state; ++ uint32_t rep0; ++ uint32_t rep1; ++ uint32_t rep2; ++ uint32_t rep3; ++}; ++ ++struct cpio_data { ++ void *data; ++ size_t size; ++ char name[18]; ++}; ++ ++enum cpio_fields { ++ C_MAGIC = 0, ++ C_INO = 1, ++ C_MODE = 2, ++ C_UID = 3, ++ C_GID = 4, ++ C_NLINK = 5, ++ C_MTIME = 6, ++ C_FILESIZE = 7, ++ C_MAJ = 8, ++ C_MIN = 9, ++ C_RMAJ = 10, ++ C_RMIN = 11, ++ C_NAMESIZE = 12, ++ C_CHKSUM = 13, ++ C_NFIELDS = 14, ++}; ++ ++struct rb_augment_callbacks { ++ void (*propagate)(struct rb_node *, struct rb_node *); ++ void (*copy)(struct rb_node *, struct rb_node *); ++ void (*rotate)(struct rb_node *, struct rb_node *); ++}; ++ ++struct spin_wait { ++ struct spin_wait *next; ++ struct spin_wait *prev; ++ int node_id; ++ long: 64; ++}; ++ ++struct trace_event_raw_initcall_level { ++ struct trace_entry ent; ++ u32 __data_loc_level; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_initcall_start { ++ struct trace_entry ent; ++ initcall_t func; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_initcall_finish { ++ struct trace_entry ent; ++ initcall_t func; ++ int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_initcall_level { ++ u32 level; ++}; ++ ++struct trace_event_data_offsets_initcall_start {}; ++ ++struct trace_event_data_offsets_initcall_finish {}; ++ ++typedef void (*btf_trace_initcall_level)(void *, const char *); ++ ++typedef void (*btf_trace_initcall_start)(void *, initcall_t); ++ ++typedef void (*btf_trace_initcall_finish)(void *, initcall_t, int); ++ ++struct blacklist_entry { ++ struct list_head next; ++ char *buf; ++}; ++ ++struct ptff_qto { ++ long unsigned int physical_clock; ++ long unsigned int tod_offset; ++ long unsigned int logical_tod_offset; ++ long unsigned int tod_epoch_difference; ++}; ++ ++struct ptff_qui { ++ unsigned int tm: 2; ++ unsigned int ts: 2; ++ unsigned int pad_0x04; ++ long unsigned int leap_event; ++ short int old_leap; ++ short int new_leap; ++ unsigned int pad_0x14; ++ long unsigned int prt[5]; ++ long unsigned int cst[3]; ++ unsigned int skew; ++ unsigned int pad_0x5c[41]; ++}; ++ ++enum clocksource_ids { ++ CSID_GENERIC = 0, ++ CSID_ARM_ARCH_COUNTER = 1, ++ CSID_MAX = 2, ++}; ++ ++struct clocksource { ++ u64 (*read)(struct clocksource *); ++ u64 mask; ++ u32 mult; ++ u32 shift; ++ u64 max_idle_ns; ++ u32 maxadj; ++ u32 uncertainty_margin; ++ u64 max_cycles; ++ const char *name; ++ struct list_head list; ++ int rating; ++ enum clocksource_ids id; ++ enum vdso_clock_mode vdso_clock_mode; ++ long unsigned int flags; ++ int (*enable)(struct clocksource *); ++ void (*disable)(struct clocksource *); ++ void (*suspend)(struct clocksource *); ++ void (*resume)(struct clocksource *); ++ void (*mark_unstable)(struct clocksource *); ++ void (*tick_stable)(struct clocksource *); ++ struct module *owner; ++}; ++ ++struct stp_irq_parm { ++ short: 14; ++ u32 tsc: 1; ++ u32 lac: 1; ++ u32 tcpc: 1; ++}; ++ ++struct stp_sstpi { ++ int: 32; ++ u32 tu: 1; ++ u32 lu: 1; ++ char: 6; ++ u32 stratum: 8; ++ u32 vbits: 16; ++ u32 leaps: 16; ++ u32 tmd: 4; ++ u32 ctn: 4; ++ char: 3; ++ u32 c: 1; ++ u32 tst: 4; ++ u32 tzo: 16; ++ u32 dsto: 16; ++ u32 ctrl: 16; ++ int: 0; ++ u32 tto; ++ int: 32; ++ u32 ctnid[3]; ++ int: 32; ++ u64 todoff; ++ u32 rsvd[50]; ++} __attribute__((packed)); ++ ++struct stp_tzib { ++ u32 tzan: 16; ++ int: 16; ++ u32 tzo: 16; ++ u32 dsto: 16; ++ u32 stn; ++ u32 dstn; ++ u64 dst_on_alg; ++ u64 dst_off_alg; ++}; ++ ++struct stp_tcpib { ++ u32 atcode: 4; ++ u32 ntcode: 4; ++ u32 d: 1; ++ s32 tto; ++ struct stp_tzib atzib; ++ struct stp_tzib ntzib; ++ s32 adst_offset: 16; ++ s32 ndst_offset: 16; ++ u32 rsvd1; ++ u64 ntzib_update; ++ u64 ndsto_update; ++}; ++ ++struct stp_lsoib { ++ u32 p: 1; ++ int: 31; ++ s32 also: 16; ++ s32 nlso: 16; ++ u64 nlsout; ++}; ++ ++struct stp_stzi { ++ u32 rsvd0[3]; ++ u64 data_ts; ++ u32 rsvd1[22]; ++ struct stp_tcpib tcpib; ++ struct stp_lsoib lsoib; ++} __attribute__((packed)); ++ ++struct clock_sync_data { ++ atomic_t cpus; ++ int in_sync; ++ long int clock_delta; ++}; ++ ++struct addrtype { ++ char _[32]; ++}; ++ ++struct addrtype___2 { ++ char _[16]; ++}; ++ ++struct addrtype___3 { ++ char _[256]; ++}; ++ ++struct stack_frame { ++ union { ++ long unsigned int empty[9]; ++ struct { ++ long unsigned int sie_control_block; ++ long unsigned int sie_savearea; ++ long unsigned int sie_reason; ++ long unsigned int sie_flags; ++ long unsigned int sie_control_block_phys; ++ }; ++ }; ++ long unsigned int gprs[10]; ++ long unsigned int back_chain; ++}; ++ ++struct fake_frame { ++ struct stack_frame sf; ++ struct pt_regs childregs; ++}; ++ ++struct s390_mmap_arg_struct { ++ long unsigned int addr; ++ long unsigned int len; ++ long unsigned int prot; ++ long unsigned int flags; ++ long unsigned int fd; ++ long unsigned int offset; ++}; ++ ++typedef struct { ++ long unsigned int mask; ++ long unsigned int addr; ++} _psw_t; ++ ++typedef struct { ++ _psw_t psw; ++ long unsigned int gprs[16]; ++ unsigned int acrs[16]; ++} _s390_regs_common; ++ ++typedef struct { ++ unsigned int fpc; ++ unsigned int pad; ++ double fprs[16]; ++} _s390_fp_regs; ++ ++typedef struct { ++ _s390_regs_common regs; ++ _s390_fp_regs fpregs; ++} _sigregs; ++ ++typedef struct { ++ long long unsigned int vxrs_low[16]; ++ __vector128 vxrs_high[16]; ++ unsigned char __reserved[128]; ++} _sigregs_ext; ++ ++struct sigcontext { ++ long unsigned int oldmask[1]; ++ _sigregs *sregs; ++}; ++ ++struct sigaltstack { ++ void *ss_sp; ++ int ss_flags; ++ __kernel_size_t ss_size; ++}; ++ ++typedef struct sigaltstack stack_t; ++ ++struct ucontext; ++ ++struct ucontext_extended { ++ long unsigned int uc_flags; ++ struct ucontext *uc_link; ++ stack_t uc_stack; ++ _sigregs uc_mcontext; ++ sigset_t uc_sigmask; ++ unsigned char __unused[120]; ++ _sigregs_ext uc_mcontext_ext; ++}; ++ ++struct ucontext { ++ long unsigned int uc_flags; ++ struct ucontext *uc_link; ++ stack_t uc_stack; ++ _sigregs uc_mcontext; ++ sigset_t uc_sigmask; ++ unsigned char __unused[120]; ++}; ++ ++struct sigframe { ++ __u8 callee_used_stack[160]; ++ struct sigcontext sc; ++ _sigregs sregs; ++ int signo; ++ _sigregs_ext sregs_ext; ++ __u16 svc_insn; ++}; ++ ++struct rt_sigframe { ++ __u8 callee_used_stack[160]; ++ __u16 svc_insn; ++ struct siginfo info; ++ struct ucontext_extended uc; ++}; ++ ++enum irq_subclass { ++ IRQ_SUBCLASS_MEASUREMENT_ALERT = 5, ++ IRQ_SUBCLASS_SERVICE_SIGNAL = 9, ++}; ++ ++struct irq_class { ++ int irq; ++ char *name; ++ char *desc; ++}; ++ ++struct ext_int_info { ++ ext_int_handler_t handler; ++ struct hlist_node entry; ++ struct callback_head rcu; ++ u16 code; ++}; ++ ++struct diag_stat { ++ unsigned int counter[22]; ++}; ++ ++struct diag_desc { ++ int code; ++ char *name; ++}; ++ ++enum { ++ TYPE_HWCAP = 0, ++ TYPE_FACILITY = 1, ++}; ++ ++struct s390_cpu_feature { ++ unsigned int type: 4; ++ unsigned int num: 28; ++}; ++ ++struct sysinfo_1_2_2 { ++ char format; ++ char reserved_0[1]; ++ short unsigned int acc_offset; ++ unsigned char mt_installed: 1; ++ char: 2; ++ unsigned char mt_stid: 5; ++ char: 3; ++ unsigned char mt_gtid: 5; ++ char reserved_1[18]; ++ unsigned int nominal_cap; ++ unsigned int secondary_cap; ++ unsigned int capability; ++ short unsigned int cpus_total; ++ short unsigned int cpus_configured; ++ short unsigned int cpus_standby; ++ short unsigned int cpus_reserved; ++ short unsigned int adjustment[0]; ++}; ++ ++struct sysinfo_1_2_2_extension { ++ unsigned int alt_capability; ++ short unsigned int alt_adjustment[0]; ++}; ++ ++struct service_level { ++ struct list_head list; ++ void (*seq_print)(struct seq_file *, struct service_level *); ++}; ++ ++struct stsi_file { ++ const struct file_operations *fops; ++ char *name; ++}; ++ ++enum diag308_subcode { ++ DIAG308_CLEAR_RESET = 0, ++ DIAG308_LOAD_NORMAL_RESET = 1, ++ DIAG308_REL_HSA = 2, ++ DIAG308_LOAD_CLEAR = 3, ++ DIAG308_LOAD_NORMAL_DUMP = 4, ++ DIAG308_SET = 5, ++ DIAG308_STORE = 6, ++ DIAG308_LOAD_NORMAL = 7, ++}; ++ ++enum diag308_subcode_flags { ++ DIAG308_FLAG_EI = 65536, ++}; ++ ++typedef void (*relocate_kernel_t)(long unsigned int, long unsigned int, long unsigned int); ++ ++typedef int (*purgatory_t)(int); ++ ++enum { ++ CACHE_SCOPE_NOTEXISTS = 0, ++ CACHE_SCOPE_PRIVATE = 1, ++ CACHE_SCOPE_SHARED = 2, ++ CACHE_SCOPE_RESERVED = 3, ++}; ++ ++enum { ++ CTYPE_SEPARATE = 0, ++ CTYPE_DATA = 1, ++ CTYPE_INSTRUCTION = 2, ++ CTYPE_UNIFIED = 3, ++}; ++ ++enum { ++ EXTRACT_TOPOLOGY = 0, ++ EXTRACT_LINE_SIZE = 1, ++ EXTRACT_SIZE = 2, ++ EXTRACT_ASSOCIATIVITY = 3, ++}; ++ ++enum { ++ CACHE_TI_UNIFIED = 0, ++ CACHE_TI_DATA = 0, ++ CACHE_TI_INSTRUCTION = 1, ++}; ++ ++struct cache_info { ++ char: 4; ++ unsigned char scope: 2; ++ unsigned char type: 2; ++}; ++ ++union cache_topology { ++ struct cache_info ci[8]; ++ long unsigned int raw; ++}; ++ ++enum { ++ ec_schedule = 0, ++ ec_call_function_single = 1, ++ ec_stop_cpu = 2, ++ ec_mcck_pending = 3, ++ ec_irq_work = 4, ++}; ++ ++enum { ++ CPU_STATE_STANDBY = 0, ++ CPU_STATE_CONFIGURED = 1, ++}; ++ ++struct pcpu { ++ long unsigned int ec_mask; ++ long unsigned int ec_clk; ++ signed char state; ++ signed char polarization; ++ u16 address; ++}; ++ ++typedef void pcpu_delegate_fn(void *); ++ ++struct ec_creg_mask_parms { ++ long unsigned int orval; ++ long unsigned int andval; ++ int cr; ++}; ++ ++typedef struct { ++ char _[128]; ++} addrtype___4; ++ ++struct alt_instr { ++ s32 instr_offset; ++ s32 repl_offset; ++ u16 facility; ++ u8 instrlen; ++} __attribute__((packed)); ++ ++typedef struct { ++ unsigned int mask; ++ unsigned int addr; ++} psw_t32; ++ ++typedef u32 compat_sigset_word; ++ ++struct compat_sigaltstack { ++ compat_uptr_t ss_sp; ++ int ss_flags; ++ compat_size_t ss_size; ++}; ++ ++typedef struct compat_sigaltstack compat_stack_t; ++ ++typedef struct { ++ compat_sigset_word sig[2]; ++} compat_sigset_t; ++ ++typedef union { ++ __u64 d; ++ __u32 f; ++} freg_t32; ++ ++typedef struct { ++ unsigned int fpc; ++ unsigned int pad; ++ freg_t32 fprs[16]; ++} _s390_fp_regs32; ++ ++typedef struct { ++ psw_t32 psw; ++ __u32 gprs[16]; ++ __u32 acrs[16]; ++} _s390_regs_common32; ++ ++typedef struct { ++ _s390_regs_common32 regs; ++ _s390_fp_regs32 fpregs; ++} _sigregs32; ++ ++typedef struct { ++ __u32 gprs_high[16]; ++ __u64 vxrs_low[16]; ++ __vector128 vxrs_high[16]; ++ __u8 __reserved[128]; ++} _sigregs_ext32; ++ ++struct sigcontext32 { ++ __u32 oldmask[2]; ++ __u32 sregs; ++}; ++ ++struct ucontext32 { ++ __u32 uc_flags; ++ __u32 uc_link; ++ compat_stack_t uc_stack; ++ _sigregs32 uc_mcontext; ++ compat_sigset_t uc_sigmask; ++ unsigned char __unused[120]; ++ _sigregs_ext32 uc_mcontext_ext; ++}; ++ ++typedef struct { ++ __u8 callee_used_stack[96]; ++ struct sigcontext32 sc; ++ _sigregs32 sregs; ++ int signo; ++ _sigregs_ext32 sregs_ext; ++ __u16 svc_insn; ++} sigframe32; ++ ++typedef struct { ++ __u8 callee_used_stack[96]; ++ __u16 svc_insn; ++ compat_siginfo_t info; ++ struct ucontext32 uc; ++} rt_sigframe32; ++ ++struct ftrace_hotpatch_trampoline { ++ u16 brasl_opc; ++ s32 brasl_disp; ++ long: 0; ++ u64 rest_of_intercepted_function; ++ u64 interceptor; ++} __attribute__((packed)); ++ ++struct dyn_arch_ftrace {}; ++ ++struct dyn_ftrace { ++ long unsigned int ip; ++ long unsigned int flags; ++ struct dyn_arch_ftrace arch; ++}; ++ ++struct ftrace_insn { ++ u16 opc; ++ s32 disp; ++} __attribute__((packed)); ++ ++enum { ++ PSW_BITS_AMODE_24BIT = 0, ++ PSW_BITS_AMODE_31BIT = 1, ++ PSW_BITS_AMODE_64BIT = 3, ++}; ++ ++typedef u16 uprobe_opcode_t; ++ ++struct arch_uprobe { ++ union { ++ uprobe_opcode_t insn[3]; ++ uprobe_opcode_t ixol[3]; ++ }; ++ unsigned int saved_per: 1; ++ unsigned int saved_int_code; ++}; ++ ++enum rp_check { ++ RP_CHECK_CALL = 0, ++ RP_CHECK_CHAIN_CALL = 1, ++ RP_CHECK_RET = 2, ++}; ++ ++struct insn_ril { ++ u8 opc0; ++ u8 reg: 4; ++ u8 opc1: 4; ++ s32 disp; ++} __attribute__((packed)); ++ ++union split_register { ++ u64 u64; ++ u32 u32[2]; ++ u16 u16[4]; ++ s64 s64; ++ s32 s32[2]; ++ s16 s16[4]; ++}; ++ ++struct perf_sf_sde_regs { ++ unsigned char in_guest: 1; ++ long unsigned int reserved: 63; ++}; ++ ++struct hws_qsi_info_block { ++ unsigned int b0_13: 14; ++ unsigned int as: 1; ++ unsigned int ad: 1; ++ unsigned int b16_21: 6; ++ unsigned int es: 1; ++ unsigned int ed: 1; ++ unsigned int b24_29: 6; ++ unsigned int cs: 1; ++ unsigned int cd: 1; ++ unsigned int bsdes: 16; ++ unsigned int dsdes: 16; ++ long unsigned int min_sampl_rate; ++ long unsigned int max_sampl_rate; ++ long unsigned int tear; ++ long unsigned int dear; ++ unsigned int rsvrd0: 24; ++ unsigned int ribm: 8; ++ unsigned int cpu_speed; ++ long long unsigned int rsvrd1; ++ long long unsigned int rsvrd2; ++}; ++ ++enum perf_hw_id { ++ PERF_COUNT_HW_CPU_CYCLES = 0, ++ PERF_COUNT_HW_INSTRUCTIONS = 1, ++ PERF_COUNT_HW_CACHE_REFERENCES = 2, ++ PERF_COUNT_HW_CACHE_MISSES = 3, ++ PERF_COUNT_HW_BRANCH_INSTRUCTIONS = 4, ++ PERF_COUNT_HW_BRANCH_MISSES = 5, ++ PERF_COUNT_HW_BUS_CYCLES = 6, ++ PERF_COUNT_HW_STALLED_CYCLES_FRONTEND = 7, ++ PERF_COUNT_HW_STALLED_CYCLES_BACKEND = 8, ++ PERF_COUNT_HW_REF_CPU_CYCLES = 9, ++ PERF_COUNT_HW_MAX = 10, ++}; ++ ++struct perf_event_header { ++ __u32 type; ++ __u16 misc; ++ __u16 size; ++}; ++ ++struct hws_lsctl_request_block { ++ unsigned int s: 1; ++ unsigned int h: 1; ++ long long unsigned int b2_53: 52; ++ unsigned int es: 1; ++ unsigned int ed: 1; ++ unsigned int b56_61: 6; ++ unsigned int cs: 1; ++ unsigned int cd: 1; ++ long unsigned int interval; ++ long unsigned int tear; ++ long unsigned int dear; ++ long unsigned int rsvrd1; ++ long unsigned int rsvrd2; ++ long unsigned int rsvrd3; ++ long unsigned int rsvrd4; ++}; ++ ++struct hws_basic_entry { ++ unsigned int def: 16; ++ unsigned int R: 4; ++ unsigned int U: 4; ++ unsigned int z: 2; ++ unsigned int T: 1; ++ unsigned int W: 1; ++ unsigned int P: 1; ++ unsigned int AS: 2; ++ unsigned int I: 1; ++ unsigned int CL: 2; ++ unsigned int H: 1; ++ unsigned int LS: 1; ++ short: 12; ++ unsigned int prim_asn: 16; ++ long long unsigned int ia; ++ long long unsigned int gpp; ++ long long unsigned int hpp; ++}; ++ ++union hws_trailer_header { ++ struct { ++ unsigned int f: 1; ++ unsigned int a: 1; ++ unsigned int t: 1; ++ int: 29; ++ unsigned int bsdes: 16; ++ unsigned int dsdes: 16; ++ long long unsigned int overflow; ++ }; ++ __int128 unsigned val; ++}; ++ ++struct hws_trailer_entry { ++ union hws_trailer_header header; ++ unsigned char timestamp[16]; ++ long long unsigned int reserved1; ++ long long unsigned int reserved2; ++ union { ++ struct { ++ unsigned int clock_base: 1; ++ long long unsigned int progusage1: 63; ++ long long unsigned int progusage2; ++ }; ++ long long unsigned int progusage[2]; ++ }; ++}; ++ ++struct sf_buffer { ++ long unsigned int *sdbt; ++ long unsigned int num_sdb; ++ long unsigned int num_sdbt; ++ long unsigned int *tail; ++}; ++ ++struct aux_buffer { ++ struct sf_buffer sfb; ++ long unsigned int head; ++ long unsigned int alert_mark; ++ long unsigned int empty_mark; ++ long unsigned int *sdb_index; ++ long unsigned int *sdbt_index; ++}; ++ ++struct cpu_hw_sf { ++ struct hws_qsi_info_block qsi; ++ struct hws_lsctl_request_block lsctl; ++ struct sf_buffer sfb; ++ unsigned int flags; ++ struct perf_event *event; ++ struct perf_output_handle handle; ++}; ++ ++enum { ++ SF_CYCLES_BASIC_ATTR_IDX = 0, ++ SF_CYCLES_BASIC_DIAG_ATTR_IDX = 1, ++ SF_CYCLES_ATTR_MAX = 2, ++}; ++ ++struct paicrypt_map { ++ long unsigned int *page; ++ struct pai_userdata *save; ++ unsigned int active_events; ++ unsigned int refcnt; ++ enum paievt_mode mode; ++ struct perf_event *event; ++}; ++ ++enum uv_feat_ind { ++ BIT_UV_FEAT_MISC = 0, ++ BIT_UV_FEAT_AIV = 1, ++}; ++ ++struct uv_cb_init { ++ struct uv_cb_header header; ++ u64 reserved08[2]; ++ u64 stor_origin; ++ u64 stor_len; ++ u64 reserved28[4]; ++}; ++ ++struct uv_cb_cts { ++ struct uv_cb_header header; ++ u64 reserved08[2]; ++ u64 guest_handle; ++ u64 gaddr; ++}; ++ ++struct uv_cb_cfs { ++ struct uv_cb_header header; ++ u64 reserved08[2]; ++ u64 paddr; ++}; ++ ++struct uv_info { ++ long unsigned int inst_calls_list[4]; ++ long unsigned int uv_base_stor_len; ++ long unsigned int guest_base_stor_len; ++ long unsigned int guest_virt_base_stor_len; ++ long unsigned int guest_virt_var_stor_len; ++ long unsigned int guest_cpu_stor_len; ++ long unsigned int max_sec_stor_addr; ++ unsigned int max_num_sec_conf; ++ short unsigned int max_guest_cpu_id; ++ long unsigned int uv_feature_indications; ++ long unsigned int supp_se_hdr_ver; ++ long unsigned int supp_se_hdr_pcf; ++ long unsigned int conf_dump_storage_state_len; ++ long unsigned int conf_dump_finalize_len; ++ long unsigned int supp_att_req_hdr_ver; ++ long unsigned int supp_att_pflags; ++}; ++ ++enum fault_type { ++ KERNEL_FAULT = 0, ++ USER_FAULT = 1, ++ GMAP_FAULT = 2, ++}; ++ ++struct pfault_refbk { ++ u16 refdiagc; ++ u16 reffcode; ++ u16 refdwlen; ++ u16 refversn; ++ u64 refgaddr; ++ u64 refselmk; ++ u64 refcmpmk; ++ u64 reserved; ++}; ++ ++struct vm_unmapped_area_info { ++ long unsigned int flags; ++ long unsigned int length; ++ long unsigned int low_limit; ++ long unsigned int high_limit; ++ long unsigned int align_mask; ++ long unsigned int align_offset; ++}; ++ ++struct ptdump_range { ++ long unsigned int start; ++ long unsigned int end; ++}; ++ ++struct ptdump_state { ++ void (*note_page)(struct ptdump_state *, long unsigned int, int, u64); ++ void (*effective_prot)(struct ptdump_state *, int, u64); ++ const struct ptdump_range *range; ++}; ++ ++struct addr_marker { ++ long unsigned int start_address; ++ const char *name; ++}; ++ ++enum address_markers_idx { ++ IDENTITY_BEFORE_NR = 0, ++ IDENTITY_BEFORE_END_NR = 1, ++ AMODE31_START_NR = 2, ++ AMODE31_END_NR = 3, ++ KERNEL_START_NR = 4, ++ KERNEL_END_NR = 5, ++ KFENCE_START_NR = 6, ++ KFENCE_END_NR = 7, ++ IDENTITY_AFTER_NR = 8, ++ IDENTITY_AFTER_END_NR = 9, ++ VMEMMAP_NR = 10, ++ VMEMMAP_END_NR = 11, ++ VMALLOC_NR = 12, ++ VMALLOC_END_NR = 13, ++ MODULES_NR = 14, ++ MODULES_END_NR = 15, ++ ABS_LOWCORE_NR = 16, ++ ABS_LOWCORE_END_NR = 17, ++ MEMCPY_REAL_NR = 18, ++ MEMCPY_REAL_END_NR = 19, ++}; ++ ++struct pg_state { ++ struct ptdump_state ptdump; ++ struct seq_file *seq; ++ int level; ++ unsigned int current_prot; ++ bool check_wx; ++ long unsigned int wx_pages; ++ long unsigned int start_address; ++ const struct addr_marker *marker; ++}; ++ ++struct gmap_rmap { ++ struct gmap_rmap *next; ++ long unsigned int raddr; ++}; ++ ++struct gmap_notifier { ++ struct list_head list; ++ struct callback_head rcu; ++ void (*notifier_call)(struct gmap *, long unsigned int, long unsigned int); ++}; ++ ++struct reset_walk_state { ++ long unsigned int next; ++ long unsigned int count; ++ long unsigned int pfns[32]; ++}; ++ ++struct diag2fc_data { ++ __u32 version; ++ __u32 flags; ++ __u64 used_cpu; ++ __u64 el_time; ++ __u64 mem_min_kb; ++ __u64 mem_max_kb; ++ __u64 mem_share_kb; ++ __u64 mem_used_kb; ++ __u32 pcpus; ++ __u32 lcpus; ++ __u32 vcpus; ++ __u32 ocpus; ++ __u32 cpu_max; ++ __u32 cpu_shares; ++ __u32 cpu_use_samp; ++ __u32 cpu_delay_samp; ++ __u32 page_wait_samp; ++ __u32 idle_samp; ++ __u32 other_samp; ++ __u32 total_samp; ++ char guest_name[8]; ++}; ++ ++struct diag2fc_parm_list { ++ char userid[8]; ++ char aci_grp[8]; ++ __u64 addr; ++ __u32 size; ++ __u32 fmt; ++}; ++ ++struct dbfs_d2fc_hdr { ++ u64 len; ++ u16 version; ++ union tod_clock tod_ext; ++ u64 count; ++ char reserved[30]; ++} __attribute__((packed)); ++ ++struct dbfs_d2fc { ++ struct dbfs_d2fc_hdr hdr; ++ char buf[0]; ++}; ++ ++struct hypfs_diag0c_entry { ++ char date[8]; ++ char time[8]; ++ __u64 virtcpu; ++ __u64 totalproc; ++ __u32 cpu; ++ __u32 reserved; ++}; ++ ++struct hypfs_diag0c_hdr { ++ __u64 len; ++ __u16 version; ++ char reserved1[6]; ++ char tod_ext[16]; ++ __u64 count; ++ char reserved2[24]; ++}; ++ ++struct hypfs_diag0c_data { ++ struct hypfs_diag0c_hdr hdr; ++ struct hypfs_diag0c_entry entry[0]; ++}; ++ ++struct bpf_tramp_run_ctx; ++ ++typedef u64 (*bpf_trampoline_enter_t)(struct bpf_prog *, struct bpf_tramp_run_ctx *); ++ ++struct bpf_tramp_run_ctx { ++ struct bpf_run_ctx run_ctx; ++ u64 bpf_cookie; ++ struct bpf_run_ctx *saved_run_ctx; ++}; ++ ++typedef void (*bpf_trampoline_exit_t)(struct bpf_prog *, u64, struct bpf_tramp_run_ctx *); ++ ++enum bpf_text_poke_type { ++ BPF_MOD_CALL = 0, ++ BPF_MOD_JUMP = 1, ++}; ++ ++struct bpf_binary_header { ++ u32 size; ++ long: 0; ++ u8 image[0]; ++}; ++ ++typedef void (*bpf_jit_fill_hole_t)(void *, unsigned int); ++ ++struct bpf_jit { ++ u32 seen; ++ u32 seen_reg[16]; ++ u32 *addrs; ++ u8 *prg_buf; ++ int size; ++ int size_prg; ++ int prg; ++ int lit32_start; ++ int lit32; ++ int lit64_start; ++ int lit64; ++ int base_ip; ++ int exit_ip; ++ int r1_thunk_ip; ++ int r14_thunk_ip; ++ int tail_call_start; ++ int excnt; ++ int prologue_plt_ret; ++ int prologue_plt; ++}; ++ ++struct s390_jit_data { ++ struct bpf_binary_header *header; ++ struct bpf_jit ctx; ++ int pass; ++}; ++ ++struct bpf_tramp_jit { ++ struct bpf_jit common; ++ int orig_stack_args_off; ++ int stack_size; ++ int stack_args_off; ++ int reg_args_off; ++ int ip_off; ++ int arg_cnt_off; ++ int bpf_args_off; ++ int retval_off; ++ int r7_r8_off; ++ int r14_off; ++ int run_ctx_off; ++ int do_fexit; ++}; ++ ++struct zpci_report_error_header { ++ u8 version; ++ u8 action; ++ u16 length; ++ u8 data[0]; ++}; ++ ++struct zpci_iomap_entry { ++ u32 fh; ++ u8 bar; ++ u16 count; ++}; ++ ++struct zpci_err_insn_data { ++ u8 insn; ++ u8 cc; ++ u8 status; ++ union { ++ struct { ++ u64 req; ++ u64 offset; ++ }; ++ struct { ++ u64 addr; ++ u64 len; ++ }; ++ }; ++} __attribute__((packed)); ++ ++enum cpuhp_smt_control { ++ CPU_SMT_ENABLED = 0, ++ CPU_SMT_DISABLED = 1, ++ CPU_SMT_FORCE_DISABLED = 2, ++ CPU_SMT_NOT_SUPPORTED = 3, ++ CPU_SMT_NOT_IMPLEMENTED = 4, ++}; ++ ++enum cc_attr { ++ CC_ATTR_MEM_ENCRYPT = 0, ++ CC_ATTR_HOST_MEM_ENCRYPT = 1, ++ CC_ATTR_GUEST_MEM_ENCRYPT = 2, ++ CC_ATTR_GUEST_STATE_ENCRYPT = 3, ++ CC_ATTR_GUEST_UNROLL_STRING_IO = 4, ++ CC_ATTR_GUEST_SEV_SNP = 5, ++ CC_ATTR_HOTPLUG_DISABLED = 6, ++}; ++ ++struct trace_event_raw_cpuhp_enter { ++ struct trace_entry ent; ++ unsigned int cpu; ++ int target; ++ int idx; ++ void *fun; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_cpuhp_multi_enter { ++ struct trace_entry ent; ++ unsigned int cpu; ++ int target; ++ int idx; ++ void *fun; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_cpuhp_exit { ++ struct trace_entry ent; ++ unsigned int cpu; ++ int state; ++ int idx; ++ int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_cpuhp_enter {}; ++ ++struct trace_event_data_offsets_cpuhp_multi_enter {}; ++ ++struct trace_event_data_offsets_cpuhp_exit {}; ++ ++typedef void (*btf_trace_cpuhp_enter)(void *, unsigned int, int, int, int (*)(unsigned int)); ++ ++typedef void (*btf_trace_cpuhp_multi_enter)(void *, unsigned int, int, int, int (*)(unsigned int, struct hlist_node *), struct hlist_node *); ++ ++typedef void (*btf_trace_cpuhp_exit)(void *, unsigned int, int, int, int); ++ ++struct cpuhp_cpu_state { ++ enum cpuhp_state state; ++ enum cpuhp_state target; ++ enum cpuhp_state fail; ++ struct task_struct *thread; ++ bool should_run; ++ bool rollback; ++ bool single; ++ bool bringup; ++ struct hlist_node *node; ++ struct hlist_node *last; ++ enum cpuhp_state cb_state; ++ int result; ++ struct completion done_up; ++ struct completion done_down; ++}; ++ ++struct cpuhp_step { ++ const char *name; ++ union { ++ int (*single)(unsigned int); ++ int (*multi)(unsigned int, struct hlist_node *); ++ } startup; ++ union { ++ int (*single)(unsigned int); ++ int (*multi)(unsigned int, struct hlist_node *); ++ } teardown; ++ struct hlist_head list; ++ bool cant_stop; ++ bool multi_instance; ++}; ++ ++enum cpu_mitigations { ++ CPU_MITIGATIONS_OFF = 0, ++ CPU_MITIGATIONS_AUTO = 1, ++ CPU_MITIGATIONS_AUTO_NOSMT = 2, ++}; ++ ++struct trace_event_raw_irq_handler_entry { ++ struct trace_entry ent; ++ int irq; ++ u32 __data_loc_name; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_irq_handler_exit { ++ struct trace_entry ent; ++ int irq; ++ int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_softirq { ++ struct trace_entry ent; ++ unsigned int vec; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_tasklet { ++ struct trace_entry ent; ++ void *tasklet; ++ void *func; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_irq_handler_entry { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_irq_handler_exit {}; ++ ++struct trace_event_data_offsets_softirq {}; ++ ++struct trace_event_data_offsets_tasklet {}; ++ ++typedef void (*btf_trace_irq_handler_entry)(void *, int, struct irqaction *); ++ ++typedef void (*btf_trace_irq_handler_exit)(void *, int, struct irqaction *, int); ++ ++typedef void (*btf_trace_softirq_entry)(void *, unsigned int); ++ ++typedef void (*btf_trace_softirq_exit)(void *, unsigned int); ++ ++typedef void (*btf_trace_softirq_raise)(void *, unsigned int); ++ ++typedef void (*btf_trace_tasklet_entry)(void *, struct tasklet_struct *, void *); ++ ++typedef void (*btf_trace_tasklet_exit)(void *, struct tasklet_struct *, void *); ++ ++struct tasklet_head { ++ struct tasklet_struct *head; ++ struct tasklet_struct **tail; ++}; ++ ++enum sysctl_writes_mode { ++ SYSCTL_WRITES_LEGACY = -1, ++ SYSCTL_WRITES_WARN = 0, ++ SYSCTL_WRITES_STRICT = 1, ++}; ++ ++struct do_proc_dointvec_minmax_conv_param { ++ int *min; ++ int *max; ++}; ++ ++struct do_proc_douintvec_minmax_conv_param { ++ unsigned int *min; ++ unsigned int *max; ++}; ++ ++struct rlimit64 { ++ __u64 rlim_cur; ++ __u64 rlim_max; ++}; ++ ++typedef int (*proc_visitor)(struct task_struct *, void *); ++ ++struct compat_tms { ++ compat_clock_t tms_utime; ++ compat_clock_t tms_stime; ++ compat_clock_t tms_cutime; ++ compat_clock_t tms_cstime; ++}; ++ ++struct compat_rlimit { ++ compat_ulong_t rlim_cur; ++ compat_ulong_t rlim_max; ++}; ++ ++enum uts_proc { ++ UTS_PROC_ARCH = 0, ++ UTS_PROC_OSTYPE = 1, ++ UTS_PROC_OSRELEASE = 2, ++ UTS_PROC_VERSION = 3, ++ UTS_PROC_HOSTNAME = 4, ++ UTS_PROC_DOMAINNAME = 5, ++}; ++ ++struct prctl_mm_map { ++ __u64 start_code; ++ __u64 end_code; ++ __u64 start_data; ++ __u64 end_data; ++ __u64 start_brk; ++ __u64 brk; ++ __u64 start_stack; ++ __u64 arg_start; ++ __u64 arg_end; ++ __u64 env_start; ++ __u64 env_end; ++ __u64 *auxv; ++ __u32 auxv_size; ++ __u32 exe_fd; ++}; ++ ++struct tms { ++ __kernel_clock_t tms_utime; ++ __kernel_clock_t tms_stime; ++ __kernel_clock_t tms_cutime; ++ __kernel_clock_t tms_cstime; ++}; ++ ++struct getcpu_cache { ++ long unsigned int blob[16]; ++}; ++ ++struct compat_sysinfo { ++ s32 uptime; ++ u32 loads[3]; ++ u32 totalram; ++ u32 freeram; ++ u32 sharedram; ++ u32 bufferram; ++ u32 totalswap; ++ u32 freeswap; ++ u16 procs; ++ u16 pad; ++ u32 totalhigh; ++ u32 freehigh; ++ u32 mem_unit; ++ char _f[8]; ++}; ++ ++struct umd_info { ++ const char *driver_name; ++ struct file *pipe_to_umh; ++ struct file *pipe_from_umh; ++ struct path wd; ++ struct pid *tgid; ++}; ++ ++struct swait_queue { ++ struct task_struct *task; ++ struct list_head task_list; ++}; ++ ++enum { ++ MEMBARRIER_FLAG_SYNC_CORE = 1, ++ MEMBARRIER_FLAG_RSEQ = 2, ++}; ++ ++struct sd_flag_debug { ++ unsigned int meta_flags; ++ char *name; ++}; ++ ++enum membarrier_cmd { ++ MEMBARRIER_CMD_QUERY = 0, ++ MEMBARRIER_CMD_GLOBAL = 1, ++ MEMBARRIER_CMD_GLOBAL_EXPEDITED = 2, ++ MEMBARRIER_CMD_REGISTER_GLOBAL_EXPEDITED = 4, ++ MEMBARRIER_CMD_PRIVATE_EXPEDITED = 8, ++ MEMBARRIER_CMD_REGISTER_PRIVATE_EXPEDITED = 16, ++ MEMBARRIER_CMD_PRIVATE_EXPEDITED_SYNC_CORE = 32, ++ MEMBARRIER_CMD_REGISTER_PRIVATE_EXPEDITED_SYNC_CORE = 64, ++ MEMBARRIER_CMD_PRIVATE_EXPEDITED_RSEQ = 128, ++ MEMBARRIER_CMD_REGISTER_PRIVATE_EXPEDITED_RSEQ = 256, ++ MEMBARRIER_CMD_GET_REGISTRATIONS = 512, ++ MEMBARRIER_CMD_SHARED = 1, ++}; ++ ++enum membarrier_cmd_flag { ++ MEMBARRIER_CMD_FLAG_CPU = 1, ++}; ++ ++struct psi_window { ++ u64 size; ++ u64 start_time; ++ u64 start_value; ++ u64 prev_growth; ++}; ++ ++struct psi_trigger { ++ enum psi_states state; ++ u64 threshold; ++ struct list_head node; ++ struct psi_group *group; ++ wait_queue_head_t event_wait; ++ struct kernfs_open_file *of; ++ int event; ++ struct psi_window win; ++ u64 last_event_time; ++ bool pending_event; ++ enum psi_aggregators aggregator; ++}; ++ ++enum sched_tunable_scaling { ++ SCHED_TUNABLESCALING_NONE = 0, ++ SCHED_TUNABLESCALING_LOG = 1, ++ SCHED_TUNABLESCALING_LINEAR = 2, ++ SCHED_TUNABLESCALING_END = 3, ++}; ++ ++enum numa_topology_type { ++ NUMA_DIRECT = 0, ++ NUMA_GLUELESS_MESH = 1, ++ NUMA_BACKPLANE = 2, ++}; ++ ++enum cpuacct_stat_index { ++ CPUACCT_STAT_USER = 0, ++ CPUACCT_STAT_SYSTEM = 1, ++ CPUACCT_STAT_NSTATS = 2, ++}; ++ ++struct cpuacct { ++ struct cgroup_subsys_state css; ++ u64 *cpuusage; ++ struct kernel_cpustat *cpustat; ++}; ++ ++struct s_data { ++ struct sched_domain **sd; ++ struct root_domain *rd; ++}; ++ ++enum s_alloc { ++ sa_rootdomain = 0, ++ sa_sd = 1, ++ sa_sd_storage = 2, ++ sa_none = 3, ++}; ++ ++struct asym_cap_data { ++ struct list_head link; ++ long unsigned int capacity; ++ long unsigned int cpus[0]; ++}; ++ ++struct __cmp_key { ++ const struct cpumask *cpus; ++ struct cpumask ***masks; ++ int node; ++ int cpu; ++ int w; ++}; ++ ++struct sched_core_cookie { ++ refcount_t refcnt; ++}; ++ ++enum hk_flags { ++ HK_FLAG_TIMER = 1, ++ HK_FLAG_RCU = 2, ++ HK_FLAG_MISC = 4, ++ HK_FLAG_SCHED = 8, ++ HK_FLAG_TICK = 16, ++ HK_FLAG_DOMAIN = 32, ++ HK_FLAG_WQ = 64, ++ HK_FLAG_MANAGED_IRQ = 128, ++ HK_FLAG_KTHREAD = 256, ++}; ++ ++struct housekeeping { ++ struct cpumask cpumasks[9]; ++ long unsigned int flags; ++}; ++ ++struct irq_devres { ++ unsigned int irq; ++ void *dev_id; ++}; ++ ++struct irq_desc_devres { ++ unsigned int from; ++ unsigned int cnt; ++}; ++ ++enum { ++ IRQ_DOMAIN_FLAG_HIERARCHY = 1, ++ IRQ_DOMAIN_NAME_ALLOCATED = 2, ++ IRQ_DOMAIN_FLAG_IPI_PER_CPU = 4, ++ IRQ_DOMAIN_FLAG_IPI_SINGLE = 8, ++ IRQ_DOMAIN_FLAG_MSI = 16, ++ IRQ_DOMAIN_FLAG_ISOLATED_MSI = 32, ++ IRQ_DOMAIN_FLAG_NO_MAP = 64, ++ IRQ_DOMAIN_FLAG_MSI_PARENT = 256, ++ IRQ_DOMAIN_FLAG_MSI_DEVICE = 512, ++ IRQ_DOMAIN_FLAG_NONCORE = 65536, ++}; ++ ++enum { ++ IRQCHIP_FWNODE_REAL = 0, ++ IRQCHIP_FWNODE_NAMED = 1, ++ IRQCHIP_FWNODE_NAMED_ID = 2, ++}; ++ ++struct irqchip_fwid { ++ struct fwnode_handle fwnode; ++ unsigned int type; ++ char *name; ++ phys_addr_t *pa; ++}; ++ ++typedef void (*call_rcu_func_t)(struct callback_head *, rcu_callback_t); ++ ++struct trace_event_raw_rcu_utilization { ++ struct trace_entry ent; ++ const char *s; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_rcu_stall_warning { ++ struct trace_entry ent; ++ const char *rcuname; ++ const char *msg; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_rcu_utilization {}; ++ ++struct trace_event_data_offsets_rcu_stall_warning {}; ++ ++typedef void (*btf_trace_rcu_utilization)(void *, const char *); ++ ++typedef void (*btf_trace_rcu_stall_warning)(void *, const char *, const char *); ++ ++struct rcu_tasks; ++ ++typedef void (*rcu_tasks_gp_func_t)(struct rcu_tasks *); ++ ++typedef void (*pregp_func_t)(struct list_head *); ++ ++typedef void (*pertask_func_t)(struct task_struct *, struct list_head *); ++ ++typedef void (*postscan_func_t)(struct list_head *); ++ ++typedef void (*holdouts_func_t)(struct list_head *, bool, bool *); ++ ++typedef void (*postgp_func_t)(struct rcu_tasks *); ++ ++struct rcu_tasks_percpu; ++ ++struct rcu_tasks { ++ struct rcuwait cbs_wait; ++ raw_spinlock_t cbs_gbl_lock; ++ struct mutex tasks_gp_mutex; ++ int gp_state; ++ int gp_sleep; ++ int init_fract; ++ long unsigned int gp_jiffies; ++ long unsigned int gp_start; ++ long unsigned int tasks_gp_seq; ++ long unsigned int n_ipis; ++ long unsigned int n_ipis_fails; ++ struct task_struct *kthread_ptr; ++ rcu_tasks_gp_func_t gp_func; ++ pregp_func_t pregp_func; ++ pertask_func_t pertask_func; ++ postscan_func_t postscan_func; ++ holdouts_func_t holdouts_func; ++ postgp_func_t postgp_func; ++ call_rcu_func_t call_func; ++ struct rcu_tasks_percpu *rtpcpu; ++ int percpu_enqueue_shift; ++ int percpu_enqueue_lim; ++ int percpu_dequeue_lim; ++ long unsigned int percpu_dequeue_gpseq; ++ struct mutex barrier_q_mutex; ++ atomic_t barrier_q_count; ++ struct completion barrier_q_completion; ++ long unsigned int barrier_q_seq; ++ char *name; ++ char *kname; ++}; ++ ++struct rcu_tasks_percpu { ++ struct rcu_segcblist cblist; ++ raw_spinlock_t lock; ++ long unsigned int rtp_jiffies; ++ long unsigned int rtp_n_lock_retries; ++ struct work_struct rtp_work; ++ struct irq_work rtp_irq_work; ++ struct callback_head barrier_q_head; ++ struct list_head rtp_blkd_tasks; ++ int cpu; ++ struct rcu_tasks *rtpp; ++}; ++ ++struct trc_stall_chk_rdr { ++ int nesting; ++ int ipi_to_cpu; ++ u8 needqs; ++}; ++ ++struct klp_find_arg { ++ const char *name; ++ long unsigned int addr; ++ long unsigned int count; ++ long unsigned int pos; ++}; ++ ++typedef int (*klp_shadow_ctor_t)(void *, void *, void *); ++ ++typedef void (*klp_shadow_dtor_t)(void *, void *); ++ ++struct klp_shadow { ++ struct hlist_node node; ++ struct callback_head callback_head; ++ void *obj; ++ long unsigned int id; ++ char data[0]; ++}; ++ ++enum pci_p2pdma_map_type { ++ PCI_P2PDMA_MAP_UNKNOWN = 0, ++ PCI_P2PDMA_MAP_NOT_SUPPORTED = 1, ++ PCI_P2PDMA_MAP_BUS_ADDR = 2, ++ PCI_P2PDMA_MAP_THRU_HOST_BRIDGE = 3, ++}; ++ ++struct pci_p2pdma_map_state { ++ struct dev_pagemap *pgmap; ++ int map; ++ u64 bus_off; ++}; ++ ++struct io_tlb_area { ++ long unsigned int used; ++ unsigned int index; ++ spinlock_t lock; ++}; ++ ++struct io_tlb_slot { ++ phys_addr_t orig_addr; ++ size_t alloc_size; ++ unsigned int list; ++}; ++ ++struct trace_event_raw_swiotlb_bounced { ++ struct trace_entry ent; ++ u32 __data_loc_dev_name; ++ u64 dma_mask; ++ dma_addr_t dev_addr; ++ size_t size; ++ bool force; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_swiotlb_bounced { ++ u32 dev_name; ++}; ++ ++typedef void (*btf_trace_swiotlb_bounced)(void *, struct device *, dma_addr_t, size_t); ++ ++struct ptrace_sud_config { ++ __u64 mode; ++ __u64 selector; ++ __u64 offset; ++ __u64 len; ++}; ++ ++struct latch_tree_root { ++ seqcount_latch_t seq; ++ struct rb_root tree[2]; ++}; ++ ++struct latch_tree_ops { ++ bool (*less)(struct latch_tree_node *, struct latch_tree_node *); ++ int (*comp)(void *, struct latch_tree_node *); ++}; ++ ++struct mod_tree_root { ++ struct latch_tree_root root; ++ long unsigned int addr_min; ++ long unsigned int addr_max; ++}; ++ ++struct mod_unload_taint { ++ struct list_head list; ++ char name[56]; ++ long unsigned int taints; ++ u64 count; ++}; ++ ++typedef __kernel_long_t __kernel_suseconds_t; ++ ++typedef __kernel_suseconds_t suseconds_t; ++ ++typedef __u64 timeu64_t; ++ ++struct ktime_timestamps { ++ u64 mono; ++ u64 boot; ++ u64 real; ++}; ++ ++struct system_time_snapshot { ++ u64 cycles; ++ ktime_t real; ++ ktime_t raw; ++ enum clocksource_ids cs_id; ++ unsigned int clock_was_set_seq; ++ u8 cs_was_changed_seq; ++}; ++ ++struct system_counterval_t { ++ u64 cycles; ++ struct clocksource *cs; ++}; ++ ++struct tk_read_base { ++ struct clocksource *clock; ++ u64 mask; ++ u64 cycle_last; ++ u32 mult; ++ u32 shift; ++ u64 xtime_nsec; ++ ktime_t base; ++ u64 base_real; ++}; ++ ++struct timekeeper { ++ struct tk_read_base tkr_mono; ++ struct tk_read_base tkr_raw; ++ u64 xtime_sec; ++ long unsigned int ktime_sec; ++ struct timespec64 wall_to_monotonic; ++ ktime_t offs_real; ++ ktime_t offs_boot; ++ ktime_t offs_tai; ++ s32 tai_offset; ++ unsigned int clock_was_set_seq; ++ u8 cs_was_changed_seq; ++ ktime_t next_leap_ktime; ++ u64 raw_sec; ++ struct timespec64 monotonic_to_boot; ++ u64 cycle_interval; ++ u64 xtime_interval; ++ s64 xtime_remainder; ++ u64 raw_interval; ++ u64 ntp_tick; ++ s64 ntp_error; ++ u32 ntp_error_shift; ++ u32 ntp_err_mult; ++ u32 skip_second_overflow; ++}; ++ ++struct syscore_ops { ++ struct list_head node; ++ int (*suspend)(); ++ void (*resume)(); ++ void (*shutdown)(); ++}; ++ ++enum timekeeping_adv_mode { ++ TK_ADV_TICK = 0, ++ TK_ADV_FREQ = 1, ++}; ++ ++struct tk_fast { ++ seqcount_latch_t seq; ++ struct tk_read_base base[2]; ++}; ++ ++enum tick_nohz_mode { ++ NOHZ_MODE_INACTIVE = 0, ++ NOHZ_MODE_LOWRES = 1, ++ NOHZ_MODE_HIGHRES = 2, ++}; ++ ++struct tick_sched { ++ unsigned int inidle: 1; ++ unsigned int tick_stopped: 1; ++ unsigned int idle_active: 1; ++ unsigned int do_timer_last: 1; ++ unsigned int got_idle_tick: 1; ++ unsigned int stalled_jiffies; ++ long unsigned int last_tick_jiffies; ++ struct hrtimer sched_timer; ++ ktime_t last_tick; ++ ktime_t next_tick; ++ long unsigned int idle_jiffies; ++ ktime_t idle_waketime; ++ seqcount_t idle_sleeptime_seq; ++ ktime_t idle_entrytime; ++ enum tick_nohz_mode nohz_mode; ++ long unsigned int last_jiffies; ++ u64 timer_expires_base; ++ u64 timer_expires; ++ u64 next_timer; ++ ktime_t idle_expires; ++ long unsigned int idle_calls; ++ long unsigned int idle_sleeps; ++ ktime_t idle_exittime; ++ ktime_t idle_sleeptime; ++ ktime_t iowait_sleeptime; ++ atomic_t tick_dep_mask; ++ long unsigned int check_clocks; ++}; ++ ++struct timer_list_iter { ++ int cpu; ++ bool second_pass; ++ u64 now; ++}; ++ ++struct posix_clock_desc { ++ struct file *fp; ++ struct posix_clock *clk; ++}; ++ ++struct ce_unbind { ++ struct clock_event_device *ce; ++ int res; ++}; ++ ++struct futex_waitv { ++ __u64 val; ++ __u64 uaddr; ++ __u32 flags; ++ __u32 __reserved; ++}; ++ ++struct futex_vector { ++ struct futex_waitv w; ++ struct futex_q q; ++}; ++ ++struct kexec_load_limit { ++ struct mutex mutex; ++ int limit; ++}; ++ ++struct crash_mem { ++ unsigned int max_nr_ranges; ++ unsigned int nr_ranges; ++ struct range ranges[0]; ++}; ++ ++struct kexec_sha_region { ++ long unsigned int start; ++ long unsigned int len; ++}; ++ ++struct kernfs_fs_context { ++ struct kernfs_root *root; ++ void *ns_tag; ++ long unsigned int magic; ++ bool new_sb_created; ++}; ++ ++struct cgroup_fs_context { ++ struct kernfs_fs_context kfc; ++ struct cgroup_root *root; ++ struct cgroup_namespace *ns; ++ unsigned int flags; ++ bool cpuset_clone_children; ++ bool none; ++ bool all_ss; ++ u16 subsys_mask; ++ char *name; ++ char *release_agent; ++}; ++ ++struct cgroup_pidlist; ++ ++struct cgroup_file_ctx { ++ struct cgroup_namespace *ns; ++ struct { ++ void *trigger; ++ } psi; ++ struct { ++ bool started; ++ struct css_task_iter iter; ++ } procs; ++ struct { ++ struct cgroup_pidlist *pidlist; ++ } procs1; ++}; ++ ++struct cgrp_cset_link { ++ struct cgroup *cgrp; ++ struct css_set *cset; ++ struct list_head cset_link; ++ struct list_head cgrp_link; ++}; ++ ++struct cgroup_mgctx { ++ struct list_head preloaded_src_csets; ++ struct list_head preloaded_dst_csets; ++ struct cgroup_taskset tset; ++ u16 ss_mask; ++}; ++ ++struct trace_event_raw_cgroup_root { ++ struct trace_entry ent; ++ int root; ++ u16 ss_mask; ++ u32 __data_loc_name; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_cgroup { ++ struct trace_entry ent; ++ int root; ++ int level; ++ u64 id; ++ u32 __data_loc_path; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_cgroup_migrate { ++ struct trace_entry ent; ++ int dst_root; ++ int dst_level; ++ u64 dst_id; ++ int pid; ++ u32 __data_loc_dst_path; ++ u32 __data_loc_comm; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_cgroup_event { ++ struct trace_entry ent; ++ int root; ++ int level; ++ u64 id; ++ u32 __data_loc_path; ++ int val; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_cgroup_root { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_cgroup { ++ u32 path; ++}; ++ ++struct trace_event_data_offsets_cgroup_migrate { ++ u32 dst_path; ++ u32 comm; ++}; ++ ++struct trace_event_data_offsets_cgroup_event { ++ u32 path; ++}; ++ ++typedef void (*btf_trace_cgroup_setup_root)(void *, struct cgroup_root *); ++ ++typedef void (*btf_trace_cgroup_destroy_root)(void *, struct cgroup_root *); ++ ++typedef void (*btf_trace_cgroup_remount)(void *, struct cgroup_root *); ++ ++typedef void (*btf_trace_cgroup_mkdir)(void *, struct cgroup *, const char *); ++ ++typedef void (*btf_trace_cgroup_rmdir)(void *, struct cgroup *, const char *); ++ ++typedef void (*btf_trace_cgroup_release)(void *, struct cgroup *, const char *); ++ ++typedef void (*btf_trace_cgroup_rename)(void *, struct cgroup *, const char *); ++ ++typedef void (*btf_trace_cgroup_freeze)(void *, struct cgroup *, const char *); ++ ++typedef void (*btf_trace_cgroup_unfreeze)(void *, struct cgroup *, const char *); ++ ++typedef void (*btf_trace_cgroup_attach_task)(void *, struct cgroup *, const char *, struct task_struct *, bool); ++ ++typedef void (*btf_trace_cgroup_transfer_tasks)(void *, struct cgroup *, const char *, struct task_struct *, bool); ++ ++typedef void (*btf_trace_cgroup_notify_populated)(void *, struct cgroup *, const char *, int); ++ ++typedef void (*btf_trace_cgroup_notify_frozen)(void *, struct cgroup *, const char *, int); ++ ++enum cgroup_opt_features { ++ OPT_FEATURE_PRESSURE = 0, ++ OPT_FEATURE_COUNT = 1, ++}; ++ ++enum cgroup2_param { ++ Opt_nsdelegate = 0, ++ Opt_favordynmods = 1, ++ Opt_memory_localevents = 2, ++ Opt_memory_recursiveprot = 3, ++ nr__cgroup2_params = 4, ++}; ++ ++struct pids_cgroup { ++ struct cgroup_subsys_state css; ++ atomic64_t counter; ++ atomic64_t limit; ++ int64_t watermark; ++ struct cgroup_file events_file; ++ atomic64_t events_limit; ++}; ++ ++struct fmeter { ++ int cnt; ++ int val; ++ time64_t time; ++ spinlock_t lock; ++}; ++ ++enum prs_errcode { ++ PERR_NONE = 0, ++ PERR_INVCPUS = 1, ++ PERR_INVPARENT = 2, ++ PERR_NOTPART = 3, ++ PERR_NOTEXCL = 4, ++ PERR_NOCPUS = 5, ++ PERR_HOTPLUG = 6, ++ PERR_CPUSEMPTY = 7, ++}; ++ ++struct cpuset { ++ struct cgroup_subsys_state css; ++ long unsigned int flags; ++ cpumask_var_t cpus_allowed; ++ nodemask_t mems_allowed; ++ cpumask_var_t effective_cpus; ++ nodemask_t effective_mems; ++ cpumask_var_t subparts_cpus; ++ nodemask_t old_mems_allowed; ++ struct fmeter fmeter; ++ int attach_in_progress; ++ int pn; ++ int relax_domain_level; ++ int nr_subparts_cpus; ++ int partition_root_state; ++ int use_parent_ecpus; ++ int child_ecpus_count; ++ int nr_deadline_tasks; ++ int nr_migrate_dl_tasks; ++ u64 sum_migrate_dl_bw; ++ enum prs_errcode prs_err; ++ struct cgroup_file partition_file; ++}; ++ ++struct tmpmasks { ++ cpumask_var_t addmask; ++ cpumask_var_t delmask; ++ cpumask_var_t new_cpus; ++}; ++ ++typedef enum { ++ CS_ONLINE = 0, ++ CS_CPU_EXCLUSIVE = 1, ++ CS_MEM_EXCLUSIVE = 2, ++ CS_MEM_HARDWALL = 3, ++ CS_MEMORY_MIGRATE = 4, ++ CS_SCHED_LOAD_BALANCE = 5, ++ CS_SPREAD_PAGE = 6, ++ CS_SPREAD_SLAB = 7, ++} cpuset_flagbits_t; ++ ++enum subparts_cmd { ++ partcmd_enable = 0, ++ partcmd_disable = 1, ++ partcmd_update = 2, ++ partcmd_invalidate = 3, ++}; ++ ++struct cpuset_migrate_mm_work { ++ struct work_struct work; ++ struct mm_struct *mm; ++ nodemask_t from; ++ nodemask_t to; ++}; ++ ++typedef enum { ++ FILE_MEMORY_MIGRATE = 0, ++ FILE_CPULIST = 1, ++ FILE_MEMLIST = 2, ++ FILE_EFFECTIVE_CPULIST = 3, ++ FILE_EFFECTIVE_MEMLIST = 4, ++ FILE_SUBPARTS_CPULIST = 5, ++ FILE_CPU_EXCLUSIVE = 6, ++ FILE_MEM_EXCLUSIVE = 7, ++ FILE_MEM_HARDWALL = 8, ++ FILE_SCHED_LOAD_BALANCE = 9, ++ FILE_PARTITION_ROOT = 10, ++ FILE_SCHED_RELAX_DOMAIN_LEVEL = 11, ++ FILE_MEMORY_PRESSURE_ENABLED = 12, ++ FILE_MEMORY_PRESSURE = 13, ++ FILE_SPREAD_PAGE = 14, ++ FILE_SPREAD_SLAB = 15, ++} cpuset_filetype_t; ++ ++struct cpu_stop_done { ++ atomic_t nr_todo; ++ int ret; ++ struct completion completion; ++}; ++ ++struct cpu_stopper { ++ struct task_struct *thread; ++ raw_spinlock_t lock; ++ bool enabled; ++ struct list_head works; ++ struct cpu_stop_work stop_work; ++ long unsigned int caller; ++ cpu_stop_fn_t fn; ++}; ++ ++enum multi_stop_state { ++ MULTI_STOP_NONE = 0, ++ MULTI_STOP_PREPARE = 1, ++ MULTI_STOP_DISABLE_IRQ = 2, ++ MULTI_STOP_RUN = 3, ++ MULTI_STOP_EXIT = 4, ++}; ++ ++struct multi_stop_data { ++ cpu_stop_fn_t fn; ++ void *data; ++ unsigned int num_threads; ++ const struct cpumask *active_cpus; ++ enum multi_stop_state state; ++ atomic_t thread_ack; ++}; ++ ++struct audit_rule_data { ++ __u32 flags; ++ __u32 action; ++ __u32 field_count; ++ __u32 mask[64]; ++ __u32 fields[64]; ++ __u32 values[64]; ++ __u32 fieldflags[64]; ++ __u32 buflen; ++ char buf[0]; ++}; ++ ++struct audit_netlink_list { ++ __u32 portid; ++ struct net *net; ++ struct sk_buff_head q; ++}; ++ ++struct audit_parent; ++ ++struct audit_watch { ++ refcount_t count; ++ dev_t dev; ++ char *path; ++ long unsigned int ino; ++ struct audit_parent *parent; ++ struct list_head wlist; ++ struct list_head rules; ++}; ++ ++struct audit_parent { ++ struct list_head watches; ++ struct fsnotify_mark mark; ++}; ++ ++struct audit_tree { ++ refcount_t count; ++ int goner; ++ struct audit_chunk *root; ++ struct list_head chunks; ++ struct list_head rules; ++ struct list_head list; ++ struct list_head same_root; ++ struct callback_head head; ++ char pathname[0]; ++}; ++ ++struct audit_node { ++ struct list_head list; ++ struct audit_tree *owner; ++ unsigned int index; ++}; ++ ++struct audit_chunk { ++ struct list_head hash; ++ long unsigned int key; ++ struct fsnotify_mark *mark; ++ struct list_head trees; ++ int count; ++ atomic_long_t refs; ++ struct callback_head head; ++ struct audit_node owners[0]; ++}; ++ ++struct audit_tree_mark { ++ struct fsnotify_mark mark; ++ struct audit_chunk *chunk; ++}; ++ ++enum { ++ HASH_SIZE = 128, ++}; ++ ++struct rchan_percpu_buf_dispatcher { ++ struct rchan_buf *buf; ++ struct dentry *dentry; ++}; ++ ++struct tp_module { ++ struct list_head list; ++ struct module *mod; ++}; ++ ++enum tp_func_state { ++ TP_FUNC_0 = 0, ++ TP_FUNC_1 = 1, ++ TP_FUNC_2 = 2, ++ TP_FUNC_N = 3, ++}; ++ ++enum tp_transition_sync { ++ TP_TRANSITION_SYNC_1_0_1 = 0, ++ TP_TRANSITION_SYNC_N_2_1 = 1, ++ _NR_TP_TRANSITION_SYNC = 2, ++}; ++ ++struct tp_transition_snapshot { ++ long unsigned int rcu; ++ long unsigned int srcu; ++ bool ongoing; ++}; ++ ++struct tp_probes { ++ struct callback_head rcu; ++ struct tracepoint_func probes[0]; ++}; ++ ++enum ftrace_bug_type { ++ FTRACE_BUG_UNKNOWN = 0, ++ FTRACE_BUG_INIT = 1, ++ FTRACE_BUG_NOP = 2, ++ FTRACE_BUG_CALL = 3, ++ FTRACE_BUG_UPDATE = 4, ++}; ++ ++enum { ++ FTRACE_FL_ENABLED = 2147483648, ++ FTRACE_FL_REGS = 1073741824, ++ FTRACE_FL_REGS_EN = 536870912, ++ FTRACE_FL_TRAMP = 268435456, ++ FTRACE_FL_TRAMP_EN = 134217728, ++ FTRACE_FL_IPMODIFY = 67108864, ++ FTRACE_FL_DISABLED = 33554432, ++ FTRACE_FL_DIRECT = 16777216, ++ FTRACE_FL_DIRECT_EN = 8388608, ++ FTRACE_FL_CALL_OPS = 4194304, ++ FTRACE_FL_CALL_OPS_EN = 2097152, ++ FTRACE_FL_TOUCHED = 1048576, ++ FTRACE_FL_MODIFIED = 524288, ++}; ++ ++enum { ++ FTRACE_UPDATE_CALLS = 1, ++ FTRACE_DISABLE_CALLS = 2, ++ FTRACE_UPDATE_TRACE_FUNC = 4, ++ FTRACE_START_FUNC_RET = 8, ++ FTRACE_STOP_FUNC_RET = 16, ++ FTRACE_MAY_SLEEP = 32, ++}; ++ ++enum { ++ FTRACE_UPDATE_IGNORE = 0, ++ FTRACE_UPDATE_MAKE_CALL = 1, ++ FTRACE_UPDATE_MODIFY_CALL = 2, ++ FTRACE_UPDATE_MAKE_NOP = 3, ++}; ++ ++enum { ++ FTRACE_ITER_FILTER = 1, ++ FTRACE_ITER_NOTRACE = 2, ++ FTRACE_ITER_PRINTALL = 4, ++ FTRACE_ITER_DO_PROBES = 8, ++ FTRACE_ITER_PROBE = 16, ++ FTRACE_ITER_MOD = 32, ++ FTRACE_ITER_ENABLED = 64, ++ FTRACE_ITER_TOUCHED = 128, ++}; ++ ++enum { ++ TRACE_PIDS = 1, ++ TRACE_NO_PIDS = 2, ++}; ++ ++struct ftrace_mod_load { ++ struct list_head list; ++ char *func; ++ char *module; ++ int enable; ++}; ++ ++struct tracer_stat { ++ const char *name; ++ void * (*stat_start)(struct tracer_stat *); ++ void * (*stat_next)(void *, int); ++ cmp_func_t stat_cmp; ++ int (*stat_show)(struct seq_file *, void *); ++ void (*stat_release)(void *); ++ int (*stat_headers)(struct seq_file *); ++}; ++ ++enum { ++ FTRACE_MODIFY_ENABLE_FL = 1, ++ FTRACE_MODIFY_MAY_SLEEP_FL = 2, ++}; ++ ++struct ftrace_profile { ++ struct hlist_node node; ++ long unsigned int ip; ++ long unsigned int counter; ++ long long unsigned int time; ++ long long unsigned int time_squared; ++}; ++ ++struct ftrace_profile_page { ++ struct ftrace_profile_page *next; ++ long unsigned int index; ++ struct ftrace_profile records[0]; ++}; ++ ++struct ftrace_profile_stat { ++ atomic_t disabled; ++ struct hlist_head *hash; ++ struct ftrace_profile_page *pages; ++ struct ftrace_profile_page *start; ++ struct tracer_stat stat; ++}; ++ ++struct ftrace_func_probe { ++ struct ftrace_probe_ops *probe_ops; ++ struct ftrace_ops ops; ++ struct trace_array *tr; ++ struct list_head list; ++ void *data; ++ int ref; ++}; ++ ++struct ftrace_page { ++ struct ftrace_page *next; ++ struct dyn_ftrace *records; ++ int index; ++ int order; ++}; ++ ++struct ftrace_rec_iter { ++ struct ftrace_page *pg; ++ int index; ++}; ++ ++struct ftrace_iterator { ++ loff_t pos; ++ loff_t func_pos; ++ loff_t mod_pos; ++ struct ftrace_page *pg; ++ struct dyn_ftrace *func; ++ struct ftrace_func_probe *probe; ++ struct ftrace_func_entry *probe_entry; ++ struct trace_parser parser; ++ struct ftrace_hash *hash; ++ struct ftrace_ops *ops; ++ struct trace_array *tr; ++ struct list_head *mod_list; ++ int pidx; ++ int idx; ++ unsigned int flags; ++}; ++ ++struct ftrace_glob { ++ char *search; ++ unsigned int len; ++ int type; ++}; ++ ++struct ftrace_func_map { ++ struct ftrace_func_entry entry; ++ void *data; ++}; ++ ++struct ftrace_func_mapper { ++ struct ftrace_hash hash; ++}; ++ ++enum graph_filter_type { ++ GRAPH_FILTER_NOTRACE = 0, ++ GRAPH_FILTER_FUNCTION = 1, ++}; ++ ++struct ftrace_graph_data { ++ struct ftrace_hash *hash; ++ struct ftrace_func_entry *entry; ++ int idx; ++ enum graph_filter_type type; ++ struct ftrace_hash *new_hash; ++ const struct seq_operations *seq_ops; ++ struct trace_parser parser; ++}; ++ ++struct ftrace_mod_func { ++ struct list_head list; ++ char *name; ++ long unsigned int ip; ++ unsigned int size; ++}; ++ ++struct ftrace_mod_map { ++ struct callback_head rcu; ++ struct list_head list; ++ struct module *mod; ++ long unsigned int start_addr; ++ long unsigned int end_addr; ++ struct list_head funcs; ++ unsigned int num_funcs; ++}; ++ ++struct ftrace_init_func { ++ struct list_head list; ++ long unsigned int ip; ++}; ++ ++struct kallsyms_data { ++ long unsigned int *addrs; ++ const char **syms; ++ size_t cnt; ++ size_t found; ++}; ++ ++struct userstack_entry { ++ struct trace_entry ent; ++ unsigned int tgid; ++ long unsigned int caller[8]; ++}; ++ ++struct hwlat_entry { ++ struct trace_entry ent; ++ u64 duration; ++ u64 outer_duration; ++ u64 nmi_total_ts; ++ struct timespec64 timestamp; ++ unsigned int nmi_count; ++ unsigned int seqnum; ++ unsigned int count; ++}; ++ ++struct osnoise_entry { ++ struct trace_entry ent; ++ u64 noise; ++ u64 runtime; ++ u64 max_sample; ++ unsigned int hw_count; ++ unsigned int nmi_count; ++ unsigned int irq_count; ++ unsigned int softirq_count; ++ unsigned int thread_count; ++}; ++ ++struct timerlat_entry { ++ struct trace_entry ent; ++ unsigned int seqnum; ++ int context; ++ u64 timer_latency; ++}; ++ ++struct trace_mark { ++ long long unsigned int val; ++ char sym; ++}; ++ ++struct stat_node { ++ struct rb_node node; ++ void *stat; ++}; ++ ++struct stat_session { ++ struct list_head session_list; ++ struct tracer_stat *ts; ++ struct rb_root stat_root; ++ struct mutex stat_mutex; ++ struct dentry *file; ++}; ++ ++enum { ++ TRACE_FUNC_NO_OPTS = 0, ++ TRACE_FUNC_OPT_STACK = 1, ++ TRACE_FUNC_OPT_NO_REPEATS = 2, ++ TRACE_FUNC_OPT_HIGHEST_BIT = 4, ++}; ++ ++enum { ++ MODE_NONE = 0, ++ MODE_ROUND_ROBIN = 1, ++ MODE_PER_CPU = 2, ++ MODE_MAX = 3, ++}; ++ ++struct hwlat_kthread_data { ++ struct task_struct *kthread; ++ u64 nmi_ts_start; ++ u64 nmi_total_ts; ++ int nmi_count; ++ int nmi_cpu; ++}; ++ ++struct hwlat_sample { ++ u64 seqnum; ++ u64 duration; ++ u64 outer_duration; ++ u64 nmi_total_ts; ++ struct timespec64 timestamp; ++ int nmi_count; ++ int count; ++}; ++ ++struct hwlat_data { ++ struct mutex lock; ++ u64 count; ++ u64 sample_window; ++ u64 sample_width; ++ int thread_mode; ++}; ++ ++struct ftrace_graph_ent_entry { ++ struct trace_entry ent; ++ struct ftrace_graph_ent graph_ent; ++}; ++ ++struct ftrace_graph_ret_entry { ++ struct trace_entry ent; ++ struct ftrace_graph_ret ret; ++}; ++ ++struct fgraph_cpu_data { ++ pid_t last_pid; ++ int depth; ++ int depth_irq; ++ int ignore; ++ long unsigned int enter_funcs[50]; ++}; ++ ++struct fgraph_data { ++ struct fgraph_cpu_data *cpu_data; ++ struct ftrace_graph_ent_entry ent; ++ struct ftrace_graph_ret_entry ret; ++ int failed; ++ int cpu; ++ long: 0; ++} __attribute__((packed)); ++ ++enum { ++ FLAGS_FILL_FULL = 268435456, ++ FLAGS_FILL_START = 536870912, ++ FLAGS_FILL_END = 805306368, ++}; ++ ++typedef long unsigned int perf_trace_t[1024]; ++ ++enum { ++ EVENT_TRIGGER_FL_PROBE = 1, ++}; ++ ++struct eprobe_trace_entry_head { ++ struct trace_entry ent; ++}; ++ ++typedef int (*print_type_func_t)(struct trace_seq *, void *, void *); ++ ++enum fetch_op { ++ FETCH_OP_NOP = 0, ++ FETCH_OP_REG = 1, ++ FETCH_OP_STACK = 2, ++ FETCH_OP_STACKP = 3, ++ FETCH_OP_RETVAL = 4, ++ FETCH_OP_IMM = 5, ++ FETCH_OP_COMM = 6, ++ FETCH_OP_ARG = 7, ++ FETCH_OP_FOFFS = 8, ++ FETCH_OP_DATA = 9, ++ FETCH_OP_DEREF = 10, ++ FETCH_OP_UDEREF = 11, ++ FETCH_OP_ST_RAW = 12, ++ FETCH_OP_ST_MEM = 13, ++ FETCH_OP_ST_UMEM = 14, ++ FETCH_OP_ST_STRING = 15, ++ FETCH_OP_ST_USTRING = 16, ++ FETCH_OP_ST_SYMSTR = 17, ++ FETCH_OP_MOD_BF = 18, ++ FETCH_OP_LP_ARRAY = 19, ++ FETCH_OP_TP_ARG = 20, ++ FETCH_OP_END = 21, ++ FETCH_NOP_SYMBOL = 22, ++}; ++ ++struct fetch_insn { ++ enum fetch_op op; ++ union { ++ unsigned int param; ++ struct { ++ unsigned int size; ++ int offset; ++ }; ++ struct { ++ unsigned char basesize; ++ unsigned char lshift; ++ unsigned char rshift; ++ }; ++ long unsigned int immediate; ++ void *data; ++ }; ++}; ++ ++struct fetch_type { ++ const char *name; ++ size_t size; ++ bool is_signed; ++ bool is_string; ++ print_type_func_t print; ++ const char *fmt; ++ const char *fmttype; ++}; ++ ++struct probe_arg { ++ struct fetch_insn *code; ++ bool dynamic; ++ unsigned int offset; ++ unsigned int count; ++ const char *name; ++ const char *comm; ++ char *fmt; ++ const struct fetch_type *type; ++}; ++ ++struct trace_uprobe_filter { ++ rwlock_t rwlock; ++ int nr_systemwide; ++ struct list_head perf_events; ++}; ++ ++struct trace_probe_event { ++ unsigned int flags; ++ struct trace_event_class class; ++ struct trace_event_call call; ++ struct list_head files; ++ struct list_head probes; ++ struct trace_uprobe_filter filter[0]; ++}; ++ ++struct trace_probe { ++ struct list_head list; ++ struct trace_probe_event *event; ++ ssize_t size; ++ unsigned int nr_args; ++ struct probe_arg args[0]; ++}; ++ ++struct event_file_link { ++ struct trace_event_file *file; ++ struct list_head list; ++}; ++ ++enum probe_print_type { ++ PROBE_PRINT_NORMAL = 0, ++ PROBE_PRINT_RETURN = 1, ++ PROBE_PRINT_EVENT = 2, ++}; ++ ++enum { ++ TP_ERR_FILE_NOT_FOUND = 0, ++ TP_ERR_NO_REGULAR_FILE = 1, ++ TP_ERR_BAD_REFCNT = 2, ++ TP_ERR_REFCNT_OPEN_BRACE = 3, ++ TP_ERR_BAD_REFCNT_SUFFIX = 4, ++ TP_ERR_BAD_UPROBE_OFFS = 5, ++ TP_ERR_MAXACT_NO_KPROBE = 6, ++ TP_ERR_BAD_MAXACT = 7, ++ TP_ERR_MAXACT_TOO_BIG = 8, ++ TP_ERR_BAD_PROBE_ADDR = 9, ++ TP_ERR_BAD_RETPROBE = 10, ++ TP_ERR_BAD_ADDR_SUFFIX = 11, ++ TP_ERR_NO_GROUP_NAME = 12, ++ TP_ERR_GROUP_TOO_LONG = 13, ++ TP_ERR_BAD_GROUP_NAME = 14, ++ TP_ERR_NO_EVENT_NAME = 15, ++ TP_ERR_EVENT_TOO_LONG = 16, ++ TP_ERR_BAD_EVENT_NAME = 17, ++ TP_ERR_EVENT_EXIST = 18, ++ TP_ERR_RETVAL_ON_PROBE = 19, ++ TP_ERR_BAD_STACK_NUM = 20, ++ TP_ERR_BAD_ARG_NUM = 21, ++ TP_ERR_BAD_VAR = 22, ++ TP_ERR_BAD_REG_NAME = 23, ++ TP_ERR_BAD_MEM_ADDR = 24, ++ TP_ERR_BAD_IMM = 25, ++ TP_ERR_IMMSTR_NO_CLOSE = 26, ++ TP_ERR_FILE_ON_KPROBE = 27, ++ TP_ERR_BAD_FILE_OFFS = 28, ++ TP_ERR_SYM_ON_UPROBE = 29, ++ TP_ERR_TOO_MANY_OPS = 30, ++ TP_ERR_DEREF_NEED_BRACE = 31, ++ TP_ERR_BAD_DEREF_OFFS = 32, ++ TP_ERR_DEREF_OPEN_BRACE = 33, ++ TP_ERR_COMM_CANT_DEREF = 34, ++ TP_ERR_BAD_FETCH_ARG = 35, ++ TP_ERR_ARRAY_NO_CLOSE = 36, ++ TP_ERR_BAD_ARRAY_SUFFIX = 37, ++ TP_ERR_BAD_ARRAY_NUM = 38, ++ TP_ERR_ARRAY_TOO_BIG = 39, ++ TP_ERR_BAD_TYPE = 40, ++ TP_ERR_BAD_STRING = 41, ++ TP_ERR_BAD_SYMSTRING = 42, ++ TP_ERR_BAD_BITFIELD = 43, ++ TP_ERR_ARG_NAME_TOO_LONG = 44, ++ TP_ERR_NO_ARG_NAME = 45, ++ TP_ERR_BAD_ARG_NAME = 46, ++ TP_ERR_USED_ARG_NAME = 47, ++ TP_ERR_ARG_TOO_LONG = 48, ++ TP_ERR_NO_ARG_BODY = 49, ++ TP_ERR_BAD_INSN_BNDRY = 50, ++ TP_ERR_FAIL_REG_PROBE = 51, ++ TP_ERR_DIFF_PROBE_TYPE = 52, ++ TP_ERR_DIFF_ARG_TYPE = 53, ++ TP_ERR_SAME_PROBE = 54, ++ TP_ERR_NO_EVENT_INFO = 55, ++ TP_ERR_BAD_ATTACH_EVENT = 56, ++ TP_ERR_BAD_ATTACH_ARG = 57, ++ TP_ERR_NO_EP_FILTER = 58, ++}; ++ ++struct trace_eprobe { ++ const char *event_system; ++ const char *event_name; ++ char *filter_str; ++ struct trace_event_call *event; ++ struct dyn_event devent; ++ struct trace_probe tp; ++}; ++ ++struct eprobe_data { ++ struct trace_event_file *file; ++ struct trace_eprobe *ep; ++}; ++ ++enum { ++ BPF_F_INDEX_MASK = 4294967295ULL, ++ BPF_F_CURRENT_CPU = 4294967295ULL, ++ BPF_F_CTXLEN_MASK = 4503595332403200ULL, ++}; ++ ++enum { ++ BPF_F_GET_BRANCH_RECORDS_SIZE = 1, ++}; ++ ++struct bpf_perf_event_value { ++ __u64 counter; ++ __u64 enabled; ++ __u64 running; ++}; ++ ++enum bpf_task_fd_type { ++ BPF_FD_TYPE_RAW_TRACEPOINT = 0, ++ BPF_FD_TYPE_TRACEPOINT = 1, ++ BPF_FD_TYPE_KPROBE = 2, ++ BPF_FD_TYPE_KRETPROBE = 3, ++ BPF_FD_TYPE_UPROBE = 4, ++ BPF_FD_TYPE_URETPROBE = 5, ++}; ++ ++struct btf_ptr { ++ void *ptr; ++ __u32 type_id; ++ __u32 flags; ++}; ++ ++struct bpf_event_entry { ++ struct perf_event *event; ++ struct file *perf_file; ++ struct file *map_file; ++ struct callback_head rcu; ++}; ++ ++typedef long unsigned int (*bpf_ctx_copy_t)(void *, const void *, long unsigned int, long unsigned int); ++ ++struct bpf_key { ++ struct key *key; ++ bool has_ref; ++}; ++ ++struct perf_event_query_bpf { ++ __u32 ids_len; ++ __u32 prog_cnt; ++ __u32 ids[0]; ++}; ++ ++struct fprobe { ++ struct ftrace_ops ops; ++ long unsigned int nmissed; ++ unsigned int flags; ++ struct rethook *rethook; ++ size_t entry_data_size; ++ int nr_maxactive; ++ int (*entry_handler)(struct fprobe *, long unsigned int, struct pt_regs *, void *); ++ void (*exit_handler)(struct fprobe *, long unsigned int, struct pt_regs *, void *); ++}; ++ ++struct trace_event_raw_bpf_trace_printk { ++ struct trace_entry ent; ++ u32 __data_loc_bpf_string; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_bpf_trace_printk { ++ u32 bpf_string; ++}; ++ ++typedef void (*btf_trace_bpf_trace_printk)(void *, const char *); ++ ++struct bpf_trace_module { ++ struct module *module; ++ struct list_head list; ++}; ++ ++typedef u64 (*btf_bpf_probe_read_user)(void *, u32, const void *); ++ ++typedef u64 (*btf_bpf_probe_read_user_str)(void *, u32, const void *); ++ ++typedef u64 (*btf_bpf_probe_read_kernel)(void *, u32, const void *); ++ ++typedef u64 (*btf_bpf_probe_read_kernel_str)(void *, u32, const void *); ++ ++typedef u64 (*btf_bpf_probe_write_user)(void *, const void *, u32); ++ ++typedef u64 (*btf_bpf_trace_printk)(char *, u32, u64, u64, u64); ++ ++typedef u64 (*btf_bpf_trace_vprintk)(char *, u32, const void *, u32); ++ ++typedef u64 (*btf_bpf_seq_printf)(struct seq_file *, char *, u32, const void *, u32); ++ ++typedef u64 (*btf_bpf_seq_write)(struct seq_file *, const void *, u32); ++ ++typedef u64 (*btf_bpf_seq_printf_btf)(struct seq_file *, struct btf_ptr *, u32, u64); ++ ++typedef u64 (*btf_bpf_perf_event_read)(struct bpf_map *, u64); ++ ++typedef u64 (*btf_bpf_perf_event_read_value)(struct bpf_map *, u64, struct bpf_perf_event_value *, u32); ++ ++struct bpf_trace_sample_data { ++ struct perf_sample_data sds[3]; ++}; ++ ++typedef u64 (*btf_bpf_perf_event_output)(struct pt_regs *, struct bpf_map *, u64, void *, u64); ++ ++struct bpf_nested_pt_regs { ++ struct pt_regs regs[3]; ++}; ++ ++typedef u64 (*btf_bpf_get_current_task)(); ++ ++typedef u64 (*btf_bpf_get_current_task_btf)(); ++ ++typedef u64 (*btf_bpf_task_pt_regs)(struct task_struct *); ++ ++typedef u64 (*btf_bpf_current_task_under_cgroup)(struct bpf_map *, u32); ++ ++struct send_signal_irq_work { ++ struct irq_work irq_work; ++ struct task_struct *task; ++ u32 sig; ++ enum pid_type type; ++}; ++ ++typedef u64 (*btf_bpf_send_signal)(u32); ++ ++typedef u64 (*btf_bpf_send_signal_thread)(u32); ++ ++typedef u64 (*btf_bpf_d_path)(struct path *, char *, u32); ++ ++typedef u64 (*btf_bpf_snprintf_btf)(char *, u32, struct btf_ptr *, u32, u64); ++ ++typedef u64 (*btf_bpf_get_func_ip_tracing)(void *); ++ ++typedef u64 (*btf_bpf_get_func_ip_kprobe)(struct pt_regs *); ++ ++typedef u64 (*btf_bpf_get_func_ip_kprobe_multi)(struct pt_regs *); ++ ++typedef u64 (*btf_bpf_get_attach_cookie_kprobe_multi)(struct pt_regs *); ++ ++typedef u64 (*btf_bpf_get_attach_cookie_trace)(void *); ++ ++typedef u64 (*btf_bpf_get_attach_cookie_pe)(struct bpf_perf_event_data_kern *); ++ ++typedef u64 (*btf_bpf_get_attach_cookie_tracing)(void *); ++ ++typedef u64 (*btf_bpf_get_branch_snapshot)(void *, u32, u64); ++ ++typedef u64 (*btf_get_func_arg)(void *, u32, u64 *); ++ ++typedef u64 (*btf_get_func_ret)(void *, u64 *); ++ ++typedef u64 (*btf_get_func_arg_cnt)(void *); ++ ++typedef u64 (*btf_bpf_perf_event_output_tp)(void *, struct bpf_map *, u64, void *, u64); ++ ++typedef u64 (*btf_bpf_get_stackid_tp)(void *, struct bpf_map *, u64); ++ ++typedef u64 (*btf_bpf_get_stack_tp)(void *, void *, u32, u64); ++ ++typedef u64 (*btf_bpf_perf_prog_read_value)(struct bpf_perf_event_data_kern *, struct bpf_perf_event_value *, u32); ++ ++typedef u64 (*btf_bpf_read_branch_records)(struct bpf_perf_event_data_kern *, void *, u32, u64); ++ ++struct bpf_raw_tp_regs { ++ struct pt_regs regs[3]; ++}; ++ ++typedef u64 (*btf_bpf_perf_event_output_raw_tp)(struct bpf_raw_tracepoint_args *, struct bpf_map *, u64, void *, u64); ++ ++typedef u64 (*btf_bpf_get_stackid_raw_tp)(struct bpf_raw_tracepoint_args *, struct bpf_map *, u64); ++ ++typedef u64 (*btf_bpf_get_stack_raw_tp)(struct bpf_raw_tracepoint_args *, void *, u32, u64); ++ ++struct bpf_kprobe_multi_link { ++ struct bpf_link link; ++ struct fprobe fp; ++ long unsigned int *addrs; ++ u64 *cookies; ++ u32 cnt; ++ u32 mods_cnt; ++ struct module **mods; ++}; ++ ++struct bpf_kprobe_multi_run_ctx { ++ struct bpf_run_ctx run_ctx; ++ struct bpf_kprobe_multi_link *link; ++ long unsigned int entry_ip; ++}; ++ ++struct user_syms { ++ const char **syms; ++ char *buf; ++}; ++ ++struct multi_symbols_sort { ++ const char **funcs; ++ u64 *cookies; ++}; ++ ++struct modules_array { ++ struct module **mods; ++ int mods_cnt; ++ int mods_cap; ++}; ++ ++enum pm_qos_req_action { ++ PM_QOS_ADD_REQ = 0, ++ PM_QOS_UPDATE_REQ = 1, ++ PM_QOS_REMOVE_REQ = 2, ++}; ++ ++enum cpufreq_table_sorting { ++ CPUFREQ_TABLE_UNSORTED = 0, ++ CPUFREQ_TABLE_SORTED_ASCENDING = 1, ++ CPUFREQ_TABLE_SORTED_DESCENDING = 2, ++}; ++ ++struct cpufreq_cpuinfo { ++ unsigned int max_freq; ++ unsigned int min_freq; ++ unsigned int transition_latency; ++}; ++ ++struct cpufreq_stats; ++ ++struct thermal_cooling_device; ++ ++struct cpufreq_governor; ++ ++struct cpufreq_frequency_table; ++ ++struct cpufreq_policy { ++ cpumask_var_t cpus; ++ cpumask_var_t related_cpus; ++ cpumask_var_t real_cpus; ++ unsigned int shared_type; ++ unsigned int cpu; ++ struct clk *clk; ++ struct cpufreq_cpuinfo cpuinfo; ++ unsigned int min; ++ unsigned int max; ++ unsigned int cur; ++ unsigned int suspend_freq; ++ unsigned int policy; ++ unsigned int last_policy; ++ struct cpufreq_governor *governor; ++ void *governor_data; ++ char last_governor[16]; ++ struct work_struct update; ++ struct freq_constraints constraints; ++ struct freq_qos_request *min_freq_req; ++ struct freq_qos_request *max_freq_req; ++ struct cpufreq_frequency_table *freq_table; ++ enum cpufreq_table_sorting freq_table_sorted; ++ struct list_head policy_list; ++ struct kobject kobj; ++ struct completion kobj_unregister; ++ struct rw_semaphore rwsem; ++ bool fast_switch_possible; ++ bool fast_switch_enabled; ++ bool strict_target; ++ bool efficiencies_available; ++ unsigned int transition_delay_us; ++ bool dvfs_possible_from_any_cpu; ++ unsigned int cached_target_freq; ++ unsigned int cached_resolved_idx; ++ bool transition_ongoing; ++ spinlock_t transition_lock; ++ wait_queue_head_t transition_wait; ++ struct task_struct *transition_task; ++ struct cpufreq_stats *stats; ++ void *driver_data; ++ struct thermal_cooling_device *cdev; ++ struct notifier_block nb_min; ++ struct notifier_block nb_max; ++}; ++ ++struct cpufreq_governor { ++ char name[16]; ++ int (*init)(struct cpufreq_policy *); ++ void (*exit)(struct cpufreq_policy *); ++ int (*start)(struct cpufreq_policy *); ++ void (*stop)(struct cpufreq_policy *); ++ void (*limits)(struct cpufreq_policy *); ++ ssize_t (*show_setspeed)(struct cpufreq_policy *, char *); ++ int (*store_setspeed)(struct cpufreq_policy *, unsigned int); ++ struct list_head governor_list; ++ struct module *owner; ++ u8 flags; ++}; ++ ++struct cpufreq_frequency_table { ++ unsigned int flags; ++ unsigned int driver_data; ++ unsigned int frequency; ++}; ++ ++struct trace_event_raw_cpu { ++ struct trace_entry ent; ++ u32 state; ++ u32 cpu_id; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_cpu_idle_miss { ++ struct trace_entry ent; ++ u32 cpu_id; ++ u32 state; ++ bool below; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_powernv_throttle { ++ struct trace_entry ent; ++ int chip_id; ++ u32 __data_loc_reason; ++ int pmax; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_pstate_sample { ++ struct trace_entry ent; ++ u32 core_busy; ++ u32 scaled_busy; ++ u32 from; ++ u32 to; ++ u64 mperf; ++ u64 aperf; ++ u64 tsc; ++ u32 freq; ++ u32 io_boost; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_cpu_frequency_limits { ++ struct trace_entry ent; ++ u32 min_freq; ++ u32 max_freq; ++ u32 cpu_id; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_device_pm_callback_start { ++ struct trace_entry ent; ++ u32 __data_loc_device; ++ u32 __data_loc_driver; ++ u32 __data_loc_parent; ++ u32 __data_loc_pm_ops; ++ int event; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_device_pm_callback_end { ++ struct trace_entry ent; ++ u32 __data_loc_device; ++ u32 __data_loc_driver; ++ int error; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_suspend_resume { ++ struct trace_entry ent; ++ const char *action; ++ int val; ++ bool start; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_wakeup_source { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ u64 state; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_clock { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ u64 state; ++ u64 cpu_id; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_power_domain { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ u64 state; ++ u64 cpu_id; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_cpu_latency_qos_request { ++ struct trace_entry ent; ++ s32 value; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_pm_qos_update { ++ struct trace_entry ent; ++ enum pm_qos_req_action action; ++ int prev_value; ++ int curr_value; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_dev_pm_qos_request { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ enum dev_pm_qos_req_type type; ++ s32 new_value; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_guest_halt_poll_ns { ++ struct trace_entry ent; ++ bool grow; ++ unsigned int new; ++ unsigned int old; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_cpu {}; ++ ++struct trace_event_data_offsets_cpu_idle_miss {}; ++ ++struct trace_event_data_offsets_powernv_throttle { ++ u32 reason; ++}; ++ ++struct trace_event_data_offsets_pstate_sample {}; ++ ++struct trace_event_data_offsets_cpu_frequency_limits {}; ++ ++struct trace_event_data_offsets_device_pm_callback_start { ++ u32 device; ++ u32 driver; ++ u32 parent; ++ u32 pm_ops; ++}; ++ ++struct trace_event_data_offsets_device_pm_callback_end { ++ u32 device; ++ u32 driver; ++}; ++ ++struct trace_event_data_offsets_suspend_resume {}; ++ ++struct trace_event_data_offsets_wakeup_source { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_clock { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_power_domain { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_cpu_latency_qos_request {}; ++ ++struct trace_event_data_offsets_pm_qos_update {}; ++ ++struct trace_event_data_offsets_dev_pm_qos_request { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_guest_halt_poll_ns {}; ++ ++typedef void (*btf_trace_cpu_idle)(void *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_cpu_idle_miss)(void *, unsigned int, unsigned int, bool); ++ ++typedef void (*btf_trace_powernv_throttle)(void *, int, const char *, int); ++ ++typedef void (*btf_trace_pstate_sample)(void *, u32, u32, u32, u32, u64, u64, u64, u32, u32); ++ ++typedef void (*btf_trace_cpu_frequency)(void *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_cpu_frequency_limits)(void *, struct cpufreq_policy *); ++ ++typedef void (*btf_trace_device_pm_callback_start)(void *, struct device *, const char *, int); ++ ++typedef void (*btf_trace_device_pm_callback_end)(void *, struct device *, int); ++ ++typedef void (*btf_trace_suspend_resume)(void *, const char *, int, bool); ++ ++typedef void (*btf_trace_wakeup_source_activate)(void *, const char *, unsigned int); ++ ++typedef void (*btf_trace_wakeup_source_deactivate)(void *, const char *, unsigned int); ++ ++typedef void (*btf_trace_clock_enable)(void *, const char *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_clock_disable)(void *, const char *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_clock_set_rate)(void *, const char *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_power_domain_target)(void *, const char *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_pm_qos_add_request)(void *, s32); ++ ++typedef void (*btf_trace_pm_qos_update_request)(void *, s32); ++ ++typedef void (*btf_trace_pm_qos_remove_request)(void *, s32); ++ ++typedef void (*btf_trace_pm_qos_update_target)(void *, enum pm_qos_req_action, int, int); ++ ++typedef void (*btf_trace_pm_qos_update_flags)(void *, enum pm_qos_req_action, int, int); ++ ++typedef void (*btf_trace_dev_pm_qos_add_request)(void *, const char *, enum dev_pm_qos_req_type, s32); ++ ++typedef void (*btf_trace_dev_pm_qos_update_request)(void *, const char *, enum dev_pm_qos_req_type, s32); ++ ++typedef void (*btf_trace_dev_pm_qos_remove_request)(void *, const char *, enum dev_pm_qos_req_type, s32); ++ ++typedef void (*btf_trace_guest_halt_poll_ns)(void *, bool, unsigned int, unsigned int); ++ ++struct trace_probe_log { ++ const char *subsystem; ++ const char **argv; ++ int argc; ++ int index; ++}; ++ ++struct fprobe_rethook_node { ++ struct rethook_node node; ++ long unsigned int entry_ip; ++ long unsigned int entry_parent_ip; ++ char data[0]; ++}; ++ ++struct rv_monitor { ++ const char *name; ++ const char *description; ++ bool enabled; ++ int (*enable)(); ++ void (*disable)(); ++ void (*reset)(); ++ void (*react)(char *); ++}; ++ ++struct trace_event_raw_event_da_monitor_id { ++ struct trace_entry ent; ++ int id; ++ char state[24]; ++ char event[24]; ++ char next_state[24]; ++ bool final_state; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_error_da_monitor_id { ++ struct trace_entry ent; ++ int id; ++ char state[24]; ++ char event[24]; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_event_da_monitor_id {}; ++ ++struct trace_event_data_offsets_error_da_monitor_id {}; ++ ++typedef void (*btf_trace_event_wwnr)(void *, int, char *, char *, char *, bool); ++ ++typedef void (*btf_trace_error_wwnr)(void *, int, char *, char *); ++ ++struct rv_interface { ++ struct dentry *root_dir; ++ struct dentry *monitors_dir; ++}; ++ ++struct rv_reactor_def { ++ struct list_head list; ++ struct rv_reactor *reactor; ++ int counter; ++}; ++ ++struct rv_monitor_def { ++ struct list_head list; ++ struct rv_monitor *monitor; ++ struct dentry *root_d; ++ struct rv_reactor_def *rdef; ++ bool reacting; ++ bool task_monitor; ++}; ++ ++enum xdp_action { ++ XDP_ABORTED = 0, ++ XDP_DROP = 1, ++ XDP_PASS = 2, ++ XDP_TX = 3, ++ XDP_REDIRECT = 4, ++}; ++ ++struct xdp_cpumap_stats { ++ unsigned int redirect; ++ unsigned int pass; ++ unsigned int drop; ++}; ++ ++struct bpf_prog_pack { ++ struct list_head list; ++ void *ptr; ++ long unsigned int bitmap[0]; ++}; ++ ++struct bpf_prog_dummy { ++ struct bpf_prog prog; ++}; ++ ++typedef u64 (*btf_bpf_user_rnd_u32)(); ++ ++typedef u64 (*btf_bpf_get_raw_cpu_id)(); ++ ++struct _bpf_dtab_netdev { ++ struct net_device *dev; ++}; ++ ++struct trace_event_raw_xdp_exception { ++ struct trace_entry ent; ++ int prog_id; ++ u32 act; ++ int ifindex; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_xdp_bulk_tx { ++ struct trace_entry ent; ++ int ifindex; ++ u32 act; ++ int drops; ++ int sent; ++ int err; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_xdp_redirect_template { ++ struct trace_entry ent; ++ int prog_id; ++ u32 act; ++ int ifindex; ++ int err; ++ int to_ifindex; ++ u32 map_id; ++ int map_index; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_xdp_cpumap_kthread { ++ struct trace_entry ent; ++ int map_id; ++ u32 act; ++ int cpu; ++ unsigned int drops; ++ unsigned int processed; ++ int sched; ++ unsigned int xdp_pass; ++ unsigned int xdp_drop; ++ unsigned int xdp_redirect; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_xdp_cpumap_enqueue { ++ struct trace_entry ent; ++ int map_id; ++ u32 act; ++ int cpu; ++ unsigned int drops; ++ unsigned int processed; ++ int to_cpu; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_xdp_devmap_xmit { ++ struct trace_entry ent; ++ int from_ifindex; ++ u32 act; ++ int to_ifindex; ++ int drops; ++ int sent; ++ int err; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mem_disconnect { ++ struct trace_entry ent; ++ const struct xdp_mem_allocator *xa; ++ u32 mem_id; ++ u32 mem_type; ++ const void *allocator; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mem_connect { ++ struct trace_entry ent; ++ const struct xdp_mem_allocator *xa; ++ u32 mem_id; ++ u32 mem_type; ++ const void *allocator; ++ const struct xdp_rxq_info *rxq; ++ int ifindex; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mem_return_failed { ++ struct trace_entry ent; ++ const struct page *page; ++ u32 mem_id; ++ u32 mem_type; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_xdp_exception {}; ++ ++struct trace_event_data_offsets_xdp_bulk_tx {}; ++ ++struct trace_event_data_offsets_xdp_redirect_template {}; ++ ++struct trace_event_data_offsets_xdp_cpumap_kthread {}; ++ ++struct trace_event_data_offsets_xdp_cpumap_enqueue {}; ++ ++struct trace_event_data_offsets_xdp_devmap_xmit {}; ++ ++struct trace_event_data_offsets_mem_disconnect {}; ++ ++struct trace_event_data_offsets_mem_connect {}; ++ ++struct trace_event_data_offsets_mem_return_failed {}; ++ ++typedef void (*btf_trace_xdp_exception)(void *, const struct net_device *, const struct bpf_prog *, u32); ++ ++typedef void (*btf_trace_xdp_bulk_tx)(void *, const struct net_device *, int, int, int); ++ ++typedef void (*btf_trace_xdp_redirect)(void *, const struct net_device *, const struct bpf_prog *, const void *, int, enum bpf_map_type, u32, u32); ++ ++typedef void (*btf_trace_xdp_redirect_err)(void *, const struct net_device *, const struct bpf_prog *, const void *, int, enum bpf_map_type, u32, u32); ++ ++typedef void (*btf_trace_xdp_redirect_map)(void *, const struct net_device *, const struct bpf_prog *, const void *, int, enum bpf_map_type, u32, u32); ++ ++typedef void (*btf_trace_xdp_redirect_map_err)(void *, const struct net_device *, const struct bpf_prog *, const void *, int, enum bpf_map_type, u32, u32); ++ ++typedef void (*btf_trace_xdp_cpumap_kthread)(void *, int, unsigned int, unsigned int, int, struct xdp_cpumap_stats *); ++ ++typedef void (*btf_trace_xdp_cpumap_enqueue)(void *, int, unsigned int, unsigned int, int); ++ ++typedef void (*btf_trace_xdp_devmap_xmit)(void *, const struct net_device *, const struct net_device *, int, int, int); ++ ++typedef void (*btf_trace_mem_disconnect)(void *, const struct xdp_mem_allocator *); ++ ++typedef void (*btf_trace_mem_connect)(void *, const struct xdp_mem_allocator *, const struct xdp_rxq_info *); ++ ++typedef void (*btf_trace_mem_return_failed)(void *, const struct xdp_mem_info *, const struct page *); ++ ++struct bpf_preload_info { ++ char link_name[16]; ++ struct bpf_link *link; ++}; ++ ++struct bpf_preload_ops { ++ int (*preload)(struct bpf_preload_info *); ++ struct module *owner; ++}; ++ ++enum bpf_type { ++ BPF_TYPE_UNSPEC = 0, ++ BPF_TYPE_PROG = 1, ++ BPF_TYPE_MAP = 2, ++ BPF_TYPE_LINK = 3, ++}; ++ ++struct map_iter { ++ void *key; ++ bool done; ++}; ++ ++enum { ++ OPT_MODE = 0, ++}; ++ ++struct bpf_mount_opts { ++ umode_t mode; ++}; ++ ++struct bpf_spin_lock { ++ __u32 val; ++}; ++ ++struct bpf_timer { ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_dynptr { ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_list_head { ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_list_node { ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_rb_root { ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_rb_node { ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_refcount { ++ int: 32; ++}; ++ ++struct bpf_pidns_info { ++ __u32 pid; ++ __u32 tgid; ++}; ++ ++enum { ++ BPF_F_TIMER_ABS = 1, ++}; ++ ++typedef u64 (*btf_bpf_map_lookup_elem)(struct bpf_map *, void *); ++ ++typedef u64 (*btf_bpf_map_update_elem)(struct bpf_map *, void *, void *, u64); ++ ++typedef u64 (*btf_bpf_map_delete_elem)(struct bpf_map *, void *); ++ ++typedef u64 (*btf_bpf_map_push_elem)(struct bpf_map *, void *, u64); ++ ++typedef u64 (*btf_bpf_map_pop_elem)(struct bpf_map *, void *); ++ ++typedef u64 (*btf_bpf_map_peek_elem)(struct bpf_map *, void *); ++ ++typedef u64 (*btf_bpf_map_lookup_percpu_elem)(struct bpf_map *, void *, u32); ++ ++typedef u64 (*btf_bpf_get_smp_processor_id)(); ++ ++typedef u64 (*btf_bpf_get_numa_node_id)(); ++ ++typedef u64 (*btf_bpf_ktime_get_ns)(); ++ ++typedef u64 (*btf_bpf_ktime_get_boot_ns)(); ++ ++typedef u64 (*btf_bpf_ktime_get_coarse_ns)(); ++ ++typedef u64 (*btf_bpf_ktime_get_tai_ns)(); ++ ++typedef u64 (*btf_bpf_get_current_pid_tgid)(); ++ ++typedef u64 (*btf_bpf_get_current_uid_gid)(); ++ ++typedef u64 (*btf_bpf_get_current_comm)(char *, u32); ++ ++typedef u64 (*btf_bpf_spin_lock)(struct bpf_spin_lock *); ++ ++typedef u64 (*btf_bpf_spin_unlock)(struct bpf_spin_lock *); ++ ++typedef u64 (*btf_bpf_jiffies64)(); ++ ++typedef u64 (*btf_bpf_get_current_cgroup_id)(); ++ ++typedef u64 (*btf_bpf_get_current_ancestor_cgroup_id)(int); ++ ++typedef u64 (*btf_bpf_strtol)(const char *, size_t, u64, long int *); ++ ++typedef u64 (*btf_bpf_strtoul)(const char *, size_t, u64, long unsigned int *); ++ ++typedef u64 (*btf_bpf_strncmp)(const char *, u32, const char *); ++ ++typedef u64 (*btf_bpf_get_ns_current_pid_tgid)(u64, u64, struct bpf_pidns_info *, u32); ++ ++typedef u64 (*btf_bpf_event_output_data)(void *, struct bpf_map *, u64, void *, u64); ++ ++typedef u64 (*btf_bpf_copy_from_user)(void *, u32, const void *); ++ ++typedef u64 (*btf_bpf_copy_from_user_task)(void *, u32, const void *, struct task_struct *, u64); ++ ++typedef u64 (*btf_bpf_per_cpu_ptr)(const void *, u32); ++ ++typedef u64 (*btf_bpf_this_cpu_ptr)(const void *); ++ ++struct bpf_bprintf_buffers { ++ char bin_args[512]; ++ char buf[1024]; ++}; ++ ++typedef u64 (*btf_bpf_snprintf)(char *, u32, char *, const void *, u32); ++ ++struct bpf_hrtimer { ++ struct hrtimer timer; ++ struct bpf_map *map; ++ struct bpf_prog *prog; ++ void *callback_fn; ++ void *value; ++}; ++ ++struct bpf_timer_kern { ++ struct bpf_hrtimer *timer; ++ struct bpf_spin_lock lock; ++}; ++ ++typedef u64 (*btf_bpf_timer_init)(struct bpf_timer_kern *, struct bpf_map *, u64); ++ ++typedef u64 (*btf_bpf_timer_set_callback)(struct bpf_timer_kern *, void *, struct bpf_prog_aux *); ++ ++typedef u64 (*btf_bpf_timer_start)(struct bpf_timer_kern *, u64, u64); ++ ++typedef u64 (*btf_bpf_timer_cancel)(struct bpf_timer_kern *); ++ ++typedef u64 (*btf_bpf_kptr_xchg)(void *, void *); ++ ++typedef u64 (*btf_bpf_dynptr_from_mem)(void *, u32, u64, struct bpf_dynptr_kern *); ++ ++typedef u64 (*btf_bpf_dynptr_read)(void *, u32, const struct bpf_dynptr_kern *, u32, u64); ++ ++typedef u64 (*btf_bpf_dynptr_write)(const struct bpf_dynptr_kern *, u32, void *, u32, u64); ++ ++typedef u64 (*btf_bpf_dynptr_data)(const struct bpf_dynptr_kern *, u32, u32); ++ ++struct bpf_iter_seq_map_info { ++ u32 map_id; ++}; ++ ++struct bpf_iter__bpf_map { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct bpf_map *map; ++ }; ++}; ++ ++struct bpf_iter_seq_prog_info { ++ u32 prog_id; ++}; ++ ++struct bpf_iter__bpf_prog { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct bpf_prog *prog; ++ }; ++}; ++ ++struct bpf_iter__bpf_map_elem { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct bpf_map *map; ++ }; ++ union { ++ void *key; ++ }; ++ union { ++ void *value; ++ }; ++}; ++ ++struct bucket { ++ struct hlist_nulls_head head; ++ raw_spinlock_t raw_lock; ++}; ++ ++struct htab_elem; ++ ++struct bpf_htab { ++ struct bpf_map map; ++ struct bpf_mem_alloc ma; ++ struct bpf_mem_alloc pcpu_ma; ++ struct bucket *buckets; ++ void *elems; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ union { ++ struct pcpu_freelist freelist; ++ struct bpf_lru lru; ++ }; ++ struct htab_elem **extra_elems; ++ struct percpu_counter pcount; ++ atomic_t count; ++ bool use_percpu_counter; ++ u32 n_buckets; ++ u32 elem_size; ++ u32 hashrnd; ++ struct lock_class_key lockdep_key; ++ int *map_locked[8]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct htab_elem { ++ union { ++ struct hlist_nulls_node hash_node; ++ struct { ++ void *padding; ++ union { ++ struct pcpu_freelist_node fnode; ++ struct htab_elem *batch_flink; ++ }; ++ }; ++ }; ++ union { ++ void *ptr_to_pptr; ++ struct bpf_lru_node lru_node; ++ }; ++ u32 hash; ++ long: 0; ++ char key[0]; ++}; ++ ++struct bpf_iter_seq_hash_map_info { ++ struct bpf_map *map; ++ struct bpf_htab *htab; ++ void *percpu_value_buf; ++ u32 bucket_id; ++ u32 skip_elems; ++}; ++ ++struct bpf_bloom_filter { ++ struct bpf_map map; ++ u32 bitset_mask; ++ u32 hash_seed; ++ u32 nr_hash_funcs; ++ long unsigned int bitset[0]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++enum { ++ BPF_RB_NO_WAKEUP = 1, ++ BPF_RB_FORCE_WAKEUP = 2, ++}; ++ ++enum { ++ BPF_RB_AVAIL_DATA = 0, ++ BPF_RB_RING_SIZE = 1, ++ BPF_RB_CONS_POS = 2, ++ BPF_RB_PROD_POS = 3, ++}; ++ ++enum { ++ BPF_RINGBUF_BUSY_BIT = 2147483648, ++ BPF_RINGBUF_DISCARD_BIT = 1073741824, ++ BPF_RINGBUF_HDR_SZ = 8, ++}; ++ ++struct bpf_ringbuf { ++ wait_queue_head_t waitq; ++ struct irq_work work; ++ u64 mask; ++ struct page **pages; ++ int nr_pages; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ spinlock_t spinlock; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ atomic_t busy; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long unsigned int consumer_pos; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long unsigned int producer_pos; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ char data[0]; ++}; ++ ++struct bpf_ringbuf_map { ++ struct bpf_map map; ++ struct bpf_ringbuf *rb; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_ringbuf_hdr { ++ u32 len; ++ u32 pg_off; ++}; ++ ++typedef u64 (*btf_bpf_ringbuf_reserve)(struct bpf_map *, u64, u64); ++ ++typedef u64 (*btf_bpf_ringbuf_submit)(void *, u64); ++ ++typedef u64 (*btf_bpf_ringbuf_discard)(void *, u64); ++ ++typedef u64 (*btf_bpf_ringbuf_output)(struct bpf_map *, void *, u64, u64); ++ ++typedef u64 (*btf_bpf_ringbuf_query)(struct bpf_map *, u64); ++ ++typedef u64 (*btf_bpf_ringbuf_reserve_dynptr)(struct bpf_map *, u32, u64, struct bpf_dynptr_kern *); ++ ++typedef u64 (*btf_bpf_ringbuf_submit_dynptr)(struct bpf_dynptr_kern *, u64); ++ ++typedef u64 (*btf_bpf_ringbuf_discard_dynptr)(struct bpf_dynptr_kern *, u64); ++ ++typedef u64 (*btf_bpf_user_ringbuf_drain)(struct bpf_map *, void *, void *, u64); ++ ++enum { ++ BPF_LOCAL_STORAGE_GET_F_CREATE = 1, ++ BPF_SK_STORAGE_GET_F_CREATE = 1, ++}; ++ ++struct bpf_storage_blob { ++ struct bpf_local_storage *storage; ++}; ++ ++typedef u64 (*btf_bpf_inode_storage_get)(struct bpf_map *, struct inode *, void *, u64, gfp_t); ++ ++typedef u64 (*btf_bpf_inode_storage_delete)(struct bpf_map *, struct inode *); ++ ++struct bpf_shim_tramp_link { ++ struct bpf_tramp_link link; ++ struct bpf_trampoline *trampoline; ++}; ++ ++struct bpf_dispatcher_prog { ++ struct bpf_prog *prog; ++ refcount_t users; ++}; ++ ++struct bpf_dispatcher { ++ struct mutex mutex; ++ void *func; ++ struct bpf_dispatcher_prog progs[48]; ++ int num_progs; ++ void *image; ++ void *rw_image; ++ u32 image_off; ++ struct bpf_ksym ksym; ++}; ++ ++enum { ++ BPF_F_BROADCAST = 8, ++ BPF_F_EXCLUDE_INGRESS = 16, ++}; ++ ++struct bpf_cpumap_val { ++ __u32 qsize; ++ union { ++ int fd; ++ __u32 id; ++ } bpf_prog; ++}; ++ ++struct bpf_cpu_map_entry; ++ ++struct xdp_bulk_queue { ++ void *q[8]; ++ struct list_head flush_node; ++ struct bpf_cpu_map_entry *obj; ++ unsigned int count; ++}; ++ ++struct bpf_cpu_map; ++ ++struct bpf_cpu_map_entry { ++ u32 cpu; ++ int map_id; ++ struct xdp_bulk_queue *bulkq; ++ struct bpf_cpu_map *cmap; ++ struct ptr_ring *queue; ++ struct task_struct *kthread; ++ struct bpf_cpumap_val value; ++ struct bpf_prog *prog; ++ atomic_t refcnt; ++ struct callback_head rcu; ++ struct work_struct kthread_stop_wq; ++ struct completion kthread_running; ++}; ++ ++struct bpf_cpu_map { ++ struct bpf_map map; ++ struct bpf_cpu_map_entry **cpu_map; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_prog_info { ++ __u32 type; ++ __u32 id; ++ __u8 tag[8]; ++ __u32 jited_prog_len; ++ __u32 xlated_prog_len; ++ __u64 jited_prog_insns; ++ __u64 xlated_prog_insns; ++ __u64 load_time; ++ __u32 created_by_uid; ++ __u32 nr_map_ids; ++ __u64 map_ids; ++ char name[16]; ++ __u32 ifindex; ++ __u32 gpl_compatible: 1; ++ __u64 netns_dev; ++ __u64 netns_ino; ++ __u32 nr_jited_ksyms; ++ __u32 nr_jited_func_lens; ++ __u64 jited_ksyms; ++ __u64 jited_func_lens; ++ __u32 btf_id; ++ __u32 func_info_rec_size; ++ __u64 func_info; ++ __u32 nr_func_info; ++ __u32 nr_line_info; ++ __u64 line_info; ++ __u64 jited_line_info; ++ __u32 nr_jited_line_info; ++ __u32 line_info_rec_size; ++ __u32 jited_line_info_rec_size; ++ __u32 nr_prog_tags; ++ __u64 prog_tags; ++ __u64 run_time_ns; ++ __u64 run_cnt; ++ __u64 recursion_misses; ++ __u32 verified_insns; ++ __u32 attach_btf_obj_id; ++ __u32 attach_btf_id; ++}; ++ ++struct bpf_map_info { ++ __u32 type; ++ __u32 id; ++ __u32 key_size; ++ __u32 value_size; ++ __u32 max_entries; ++ __u32 map_flags; ++ char name[16]; ++ __u32 ifindex; ++ __u32 btf_vmlinux_value_type_id; ++ __u64 netns_dev; ++ __u64 netns_ino; ++ __u32 btf_id; ++ __u32 btf_key_type_id; ++ __u32 btf_value_type_id; ++ __u64 map_extra; ++}; ++ ++struct bpf_prog_offload_ops { ++ int (*insn_hook)(struct bpf_verifier_env *, int, int); ++ int (*finalize)(struct bpf_verifier_env *); ++ int (*replace_insn)(struct bpf_verifier_env *, u32, struct bpf_insn *); ++ int (*remove_insns)(struct bpf_verifier_env *, u32, u32); ++ int (*prepare)(struct bpf_prog *); ++ int (*translate)(struct bpf_prog *); ++ void (*destroy)(struct bpf_prog *); ++}; ++ ++struct bpf_offload_dev { ++ const struct bpf_prog_offload_ops *ops; ++ struct list_head netdevs; ++ void *priv; ++}; ++ ++enum { ++ XDP_METADATA_KFUNC_RX_TIMESTAMP = 0, ++ XDP_METADATA_KFUNC_RX_HASH = 1, ++ MAX_XDP_METADATA_KFUNC = 2, ++}; ++ ++typedef struct ns_common *ns_get_path_helper_t(void *); ++ ++struct bpf_offload_netdev { ++ struct rhash_head l; ++ struct net_device *netdev; ++ struct bpf_offload_dev *offdev; ++ struct list_head progs; ++ struct list_head maps; ++ struct list_head offdev_netdevs; ++}; ++ ++struct ns_get_path_bpf_prog_args { ++ struct bpf_prog *prog; ++ struct bpf_prog_info *info; ++}; ++ ++struct ns_get_path_bpf_map_args { ++ struct bpf_offloaded_map *offmap; ++ struct bpf_map_info *info; ++}; ++ ++struct bpf_iter__cgroup { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct cgroup *cgroup; ++ }; ++}; ++ ++struct cgroup_iter_priv { ++ struct cgroup_subsys_state *start_css; ++ bool visited_all; ++ bool terminate; ++ int order; ++}; ++ ++enum { ++ BPF_F_SYSCTL_BASE_NAME = 1, ++}; ++ ++struct bpf_cg_run_ctx { ++ struct bpf_run_ctx run_ctx; ++ const struct bpf_prog_array_item *prog_item; ++ int retval; ++}; ++ ++struct bpf_sockopt_buf { ++ u8 data[32]; ++}; ++ ++struct bpf_cgroup_link { ++ struct bpf_link link; ++ struct cgroup *cgroup; ++ enum bpf_attach_type type; ++}; ++ ++struct bpf_prog_list { ++ struct hlist_node node; ++ struct bpf_prog *prog; ++ struct bpf_cgroup_link *link; ++ struct bpf_cgroup_storage *storage[2]; ++}; ++ ++struct cgroup_lsm_atype { ++ u32 attach_btf_id; ++ int refcnt; ++}; ++ ++typedef u64 (*btf_bpf_get_local_storage)(struct bpf_map *, u64); ++ ++typedef u64 (*btf_bpf_get_retval)(); ++ ++typedef u64 (*btf_bpf_set_retval)(int); ++ ++typedef u64 (*btf_bpf_sysctl_get_name)(struct bpf_sysctl_kern *, char *, size_t, u64); ++ ++typedef u64 (*btf_bpf_sysctl_get_current_value)(struct bpf_sysctl_kern *, char *, size_t); ++ ++typedef u64 (*btf_bpf_sysctl_get_new_value)(struct bpf_sysctl_kern *, char *, size_t); ++ ++typedef u64 (*btf_bpf_sysctl_set_new_value)(struct bpf_sysctl_kern *, const char *, size_t); ++ ++typedef u64 (*btf_bpf_get_netns_cookie_sockopt)(struct bpf_sockopt_kern *); ++ ++struct bpf_dummy_ops { ++ int (*test_1)(struct bpf_dummy_ops_state *); ++ int (*test_2)(struct bpf_dummy_ops_state *, int, short unsigned int, char, long unsigned int); ++ int (*test_sleepable)(struct bpf_dummy_ops_state *); ++}; ++ ++enum bpf_struct_ops_state { ++ BPF_STRUCT_OPS_STATE_INIT = 0, ++ BPF_STRUCT_OPS_STATE_INUSE = 1, ++ BPF_STRUCT_OPS_STATE_TOBEFREE = 2, ++ BPF_STRUCT_OPS_STATE_READY = 3, ++}; ++ ++struct bpf_struct_ops_value { ++ refcount_t refcnt; ++ enum bpf_struct_ops_state state; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ char data[0]; ++}; ++ ++struct bpf_struct_ops_map { ++ struct bpf_map map; ++ struct callback_head rcu; ++ const struct bpf_struct_ops *st_ops; ++ struct mutex lock; ++ struct bpf_link **links; ++ void *image; ++ struct bpf_struct_ops_value *uvalue; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct bpf_struct_ops_value kvalue; ++}; ++ ++struct bpf_struct_ops_link { ++ struct bpf_link link; ++ struct bpf_map *map; ++}; ++ ++struct bpf_struct_ops_bpf_dummy_ops { ++ refcount_t refcnt; ++ enum bpf_struct_ops_state state; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct bpf_dummy_ops data; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_struct_ops_tcp_congestion_ops { ++ refcount_t refcnt; ++ enum bpf_struct_ops_state state; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct tcp_congestion_ops data; ++}; ++ ++enum { ++ BPF_STRUCT_OPS_TYPE_bpf_dummy_ops = 0, ++ BPF_STRUCT_OPS_TYPE_tcp_congestion_ops = 1, ++ __NR_BPF_STRUCT_OPS_TYPE = 2, ++}; ++ ++enum perf_branch_sample_type { ++ PERF_SAMPLE_BRANCH_USER = 1, ++ PERF_SAMPLE_BRANCH_KERNEL = 2, ++ PERF_SAMPLE_BRANCH_HV = 4, ++ PERF_SAMPLE_BRANCH_ANY = 8, ++ PERF_SAMPLE_BRANCH_ANY_CALL = 16, ++ PERF_SAMPLE_BRANCH_ANY_RETURN = 32, ++ PERF_SAMPLE_BRANCH_IND_CALL = 64, ++ PERF_SAMPLE_BRANCH_ABORT_TX = 128, ++ PERF_SAMPLE_BRANCH_IN_TX = 256, ++ PERF_SAMPLE_BRANCH_NO_TX = 512, ++ PERF_SAMPLE_BRANCH_COND = 1024, ++ PERF_SAMPLE_BRANCH_CALL_STACK = 2048, ++ PERF_SAMPLE_BRANCH_IND_JUMP = 4096, ++ PERF_SAMPLE_BRANCH_CALL = 8192, ++ PERF_SAMPLE_BRANCH_NO_FLAGS = 16384, ++ PERF_SAMPLE_BRANCH_NO_CYCLES = 32768, ++ PERF_SAMPLE_BRANCH_TYPE_SAVE = 65536, ++ PERF_SAMPLE_BRANCH_HW_INDEX = 131072, ++ PERF_SAMPLE_BRANCH_PRIV_SAVE = 262144, ++ PERF_SAMPLE_BRANCH_MAX = 524288, ++}; ++ ++enum perf_sample_regs_abi { ++ PERF_SAMPLE_REGS_ABI_NONE = 0, ++ PERF_SAMPLE_REGS_ABI_32 = 1, ++ PERF_SAMPLE_REGS_ABI_64 = 2, ++}; ++ ++enum perf_event_read_format { ++ PERF_FORMAT_TOTAL_TIME_ENABLED = 1, ++ PERF_FORMAT_TOTAL_TIME_RUNNING = 2, ++ PERF_FORMAT_ID = 4, ++ PERF_FORMAT_GROUP = 8, ++ PERF_FORMAT_LOST = 16, ++ PERF_FORMAT_MAX = 32, ++}; ++ ++enum perf_event_ioc_flags { ++ PERF_IOC_FLAG_GROUP = 1, ++}; ++ ++struct perf_event_mmap_page { ++ __u32 version; ++ __u32 compat_version; ++ __u32 lock; ++ __u32 index; ++ __s64 offset; ++ __u64 time_enabled; ++ __u64 time_running; ++ union { ++ __u64 capabilities; ++ struct { ++ __u64 cap_bit0: 1; ++ __u64 cap_bit0_is_deprecated: 1; ++ __u64 cap_user_rdpmc: 1; ++ __u64 cap_user_time: 1; ++ __u64 cap_user_time_zero: 1; ++ __u64 cap_user_time_short: 1; ++ __u64 cap_____res: 58; ++ }; ++ }; ++ __u16 pmc_width; ++ __u16 time_shift; ++ __u32 time_mult; ++ __u64 time_offset; ++ __u64 time_zero; ++ __u32 size; ++ __u32 __reserved_1; ++ __u64 time_cycles; ++ __u64 time_mask; ++ __u8 __reserved[928]; ++ __u64 data_head; ++ __u64 data_tail; ++ __u64 data_offset; ++ __u64 data_size; ++ __u64 aux_head; ++ __u64 aux_tail; ++ __u64 aux_offset; ++ __u64 aux_size; ++}; ++ ++struct perf_ns_link_info { ++ __u64 dev; ++ __u64 ino; ++}; ++ ++enum { ++ NET_NS_INDEX = 0, ++ UTS_NS_INDEX = 1, ++ IPC_NS_INDEX = 2, ++ PID_NS_INDEX = 3, ++ USER_NS_INDEX = 4, ++ MNT_NS_INDEX = 5, ++ CGROUP_NS_INDEX = 6, ++ NR_NAMESPACES = 7, ++}; ++ ++enum perf_event_type { ++ PERF_RECORD_MMAP = 1, ++ PERF_RECORD_LOST = 2, ++ PERF_RECORD_COMM = 3, ++ PERF_RECORD_EXIT = 4, ++ PERF_RECORD_THROTTLE = 5, ++ PERF_RECORD_UNTHROTTLE = 6, ++ PERF_RECORD_FORK = 7, ++ PERF_RECORD_READ = 8, ++ PERF_RECORD_SAMPLE = 9, ++ PERF_RECORD_MMAP2 = 10, ++ PERF_RECORD_AUX = 11, ++ PERF_RECORD_ITRACE_START = 12, ++ PERF_RECORD_LOST_SAMPLES = 13, ++ PERF_RECORD_SWITCH = 14, ++ PERF_RECORD_SWITCH_CPU_WIDE = 15, ++ PERF_RECORD_NAMESPACES = 16, ++ PERF_RECORD_KSYMBOL = 17, ++ PERF_RECORD_BPF_EVENT = 18, ++ PERF_RECORD_CGROUP = 19, ++ PERF_RECORD_TEXT_POKE = 20, ++ PERF_RECORD_AUX_OUTPUT_HW_ID = 21, ++ PERF_RECORD_MAX = 22, ++}; ++ ++enum perf_bpf_event_type { ++ PERF_BPF_EVENT_UNKNOWN = 0, ++ PERF_BPF_EVENT_PROG_LOAD = 1, ++ PERF_BPF_EVENT_PROG_UNLOAD = 2, ++ PERF_BPF_EVENT_MAX = 3, ++}; ++ ++enum perf_addr_filter_action_t { ++ PERF_ADDR_FILTER_ACTION_STOP = 0, ++ PERF_ADDR_FILTER_ACTION_START = 1, ++ PERF_ADDR_FILTER_ACTION_FILTER = 2, ++}; ++ ++struct perf_addr_filter { ++ struct list_head entry; ++ struct path path; ++ long unsigned int offset; ++ long unsigned int size; ++ enum perf_addr_filter_action_t action; ++}; ++ ++struct swevent_hlist { ++ struct hlist_head heads[256]; ++ struct callback_head callback_head; ++}; ++ ++struct pmu_event_list { ++ raw_spinlock_t lock; ++ struct list_head list; ++}; ++ ++struct perf_buffer { ++ refcount_t refcount; ++ struct callback_head callback_head; ++ int nr_pages; ++ int overwrite; ++ int paused; ++ atomic_t poll; ++ local_t head; ++ unsigned int nest; ++ local_t events; ++ local_t wakeup; ++ local_t lost; ++ long int watermark; ++ long int aux_watermark; ++ spinlock_t event_lock; ++ struct list_head event_list; ++ atomic_t mmap_count; ++ long unsigned int mmap_locked; ++ struct user_struct *mmap_user; ++ long int aux_head; ++ unsigned int aux_nest; ++ long int aux_wakeup; ++ long unsigned int aux_pgoff; ++ int aux_nr_pages; ++ int aux_overwrite; ++ atomic_t aux_mmap_count; ++ long unsigned int aux_mmap_locked; ++ void (*free_aux)(void *); ++ refcount_t aux_refcount; ++ int aux_in_sampling; ++ void **aux_pages; ++ void *aux_priv; ++ struct perf_event_mmap_page *user_page; ++ void *data_pages[0]; ++}; ++ ++struct perf_cpu_context { ++ struct perf_event_context ctx; ++ struct perf_event_context *task_ctx; ++ int online; ++ struct perf_cgroup *cgrp; ++ int heap_size; ++ struct perf_event **heap; ++ struct perf_event *heap_default[2]; ++}; ++ ++struct min_heap { ++ void *data; ++ int nr; ++ int size; ++}; ++ ++struct min_heap_callbacks { ++ int elem_size; ++ bool (*less)(const void *, const void *); ++ void (*swp)(void *, void *); ++}; ++ ++typedef int (*remote_function_f)(void *); ++ ++struct remote_function_call { ++ struct task_struct *p; ++ remote_function_f func; ++ void *info; ++ int ret; ++}; ++ ++typedef void (*event_f)(struct perf_event *, struct perf_cpu_context *, struct perf_event_context *, void *); ++ ++struct event_function_struct { ++ struct perf_event *event; ++ event_f func; ++ void *data; ++}; ++ ++enum event_type_t { ++ EVENT_FLEXIBLE = 1, ++ EVENT_PINNED = 2, ++ EVENT_TIME = 4, ++ EVENT_CPU = 8, ++ EVENT_ALL = 3, ++}; ++ ++struct __group_key { ++ int cpu; ++ struct pmu *pmu; ++ struct cgroup *cgroup; ++}; ++ ++struct stop_event_data { ++ struct perf_event *event; ++ unsigned int restart; ++}; ++ ++struct perf_read_data { ++ struct perf_event *event; ++ bool group; ++ int ret; ++}; ++ ++struct perf_read_event { ++ struct perf_event_header header; ++ u32 pid; ++ u32 tid; ++}; ++ ++typedef void perf_iterate_f(struct perf_event *, void *); ++ ++struct remote_output { ++ struct perf_buffer *rb; ++ int err; ++}; ++ ++struct perf_task_event { ++ struct task_struct *task; ++ struct perf_event_context *task_ctx; ++ struct { ++ struct perf_event_header header; ++ u32 pid; ++ u32 ppid; ++ u32 tid; ++ u32 ptid; ++ u64 time; ++ } event_id; ++}; ++ ++struct perf_comm_event { ++ struct task_struct *task; ++ char *comm; ++ int comm_size; ++ struct { ++ struct perf_event_header header; ++ u32 pid; ++ u32 tid; ++ } event_id; ++}; ++ ++struct perf_namespaces_event { ++ struct task_struct *task; ++ struct { ++ struct perf_event_header header; ++ u32 pid; ++ u32 tid; ++ u64 nr_namespaces; ++ struct perf_ns_link_info link_info[7]; ++ } event_id; ++}; ++ ++struct perf_cgroup_event { ++ char *path; ++ int path_size; ++ struct { ++ struct perf_event_header header; ++ u64 id; ++ char path[0]; ++ } event_id; ++}; ++ ++struct perf_mmap_event { ++ struct vm_area_struct *vma; ++ const char *file_name; ++ int file_size; ++ int maj; ++ int min; ++ u64 ino; ++ u64 ino_generation; ++ u32 prot; ++ u32 flags; ++ u8 build_id[20]; ++ u32 build_id_size; ++ struct { ++ struct perf_event_header header; ++ u32 pid; ++ u32 tid; ++ u64 start; ++ u64 len; ++ u64 pgoff; ++ } event_id; ++}; ++ ++struct perf_switch_event { ++ struct task_struct *task; ++ struct task_struct *next_prev; ++ struct { ++ struct perf_event_header header; ++ u32 next_prev_pid; ++ u32 next_prev_tid; ++ } event_id; ++}; ++ ++struct perf_ksymbol_event { ++ const char *name; ++ int name_len; ++ struct { ++ struct perf_event_header header; ++ u64 addr; ++ u32 len; ++ u16 ksym_type; ++ u16 flags; ++ } event_id; ++}; ++ ++struct perf_bpf_event { ++ struct bpf_prog *prog; ++ struct { ++ struct perf_event_header header; ++ u16 type; ++ u16 flags; ++ u32 id; ++ u8 tag[8]; ++ } event_id; ++}; ++ ++struct perf_text_poke_event { ++ const void *old_bytes; ++ const void *new_bytes; ++ size_t pad; ++ u16 old_len; ++ u16 new_len; ++ struct { ++ struct perf_event_header header; ++ u64 addr; ++ } event_id; ++}; ++ ++struct swevent_htable { ++ struct swevent_hlist *swevent_hlist; ++ struct mutex hlist_mutex; ++ int hlist_refcount; ++ int recursion[4]; ++}; ++ ++enum perf_probe_config { ++ PERF_PROBE_CONFIG_IS_RETPROBE = 1, ++ PERF_UPROBE_REF_CTR_OFFSET_BITS = 32, ++ PERF_UPROBE_REF_CTR_OFFSET_SHIFT = 32, ++}; ++ ++enum { ++ IF_ACT_NONE = -1, ++ IF_ACT_FILTER = 0, ++ IF_ACT_START = 1, ++ IF_ACT_STOP = 2, ++ IF_SRC_FILE = 3, ++ IF_SRC_KERNEL = 4, ++ IF_SRC_FILEADDR = 5, ++ IF_SRC_KERNELADDR = 6, ++}; ++ ++enum { ++ IF_STATE_ACTION = 0, ++ IF_STATE_SOURCE = 1, ++ IF_STATE_END = 2, ++}; ++ ++struct perf_aux_event { ++ struct perf_event_header header; ++ u64 hw_id; ++}; ++ ++struct perf_aux_event___2 { ++ struct perf_event_header header; ++ u32 pid; ++ u32 tid; ++}; ++ ++struct perf_aux_event___3 { ++ struct perf_event_header header; ++ u64 offset; ++ u64 size; ++ u64 flags; ++}; ++ ++enum rseq_cpu_id_state { ++ RSEQ_CPU_ID_UNINITIALIZED = -1, ++ RSEQ_CPU_ID_REGISTRATION_FAILED = -2, ++}; ++ ++enum rseq_flags { ++ RSEQ_FLAG_UNREGISTER = 1, ++}; ++ ++enum rseq_cs_flags { ++ RSEQ_CS_FLAG_NO_RESTART_ON_PREEMPT = 1, ++ RSEQ_CS_FLAG_NO_RESTART_ON_SIGNAL = 2, ++ RSEQ_CS_FLAG_NO_RESTART_ON_MIGRATE = 4, ++}; ++ ++struct rseq_cs { ++ __u32 version; ++ __u32 flags; ++ __u64 start_ip; ++ __u64 post_commit_offset; ++ __u64 abort_ip; ++}; ++ ++struct trace_event_raw_rseq_update { ++ struct trace_entry ent; ++ s32 cpu_id; ++ s32 node_id; ++ s32 mm_cid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_rseq_ip_fixup { ++ struct trace_entry ent; ++ long unsigned int regs_ip; ++ long unsigned int start_ip; ++ long unsigned int post_commit_offset; ++ long unsigned int abort_ip; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_rseq_update {}; ++ ++struct trace_event_data_offsets_rseq_ip_fixup {}; ++ ++typedef void (*btf_trace_rseq_update)(void *, struct task_struct *); ++ ++typedef void (*btf_trace_rseq_ip_fixup)(void *, long unsigned int, long unsigned int, long unsigned int, long unsigned int); ++ ++enum blacklist_hash_type { ++ BLACKLIST_HASH_X509_TBS = 1, ++ BLACKLIST_HASH_BINARY = 2, ++}; ++ ++enum positive_aop_returns { ++ AOP_WRITEPAGE_ACTIVATE = 524288, ++ AOP_TRUNCATED_PAGE = 524289, ++}; ++ ++struct wb_lock_cookie { ++ bool locked; ++ long unsigned int flags; ++}; ++ ++struct dirty_throttle_control { ++ struct wb_domain *dom; ++ struct dirty_throttle_control *gdtc; ++ struct bdi_writeback *wb; ++ struct fprop_local_percpu *wb_completions; ++ long unsigned int avail; ++ long unsigned int dirty; ++ long unsigned int thresh; ++ long unsigned int bg_thresh; ++ long unsigned int wb_dirty; ++ long unsigned int wb_thresh; ++ long unsigned int wb_bg_thresh; ++ long unsigned int pos_ratio; ++}; ++ ++struct simple_xattr { ++ struct rb_node rb_node; ++ char *name; ++ size_t size; ++ char value[0]; ++}; ++ ++typedef unsigned int uffd_flags_t; ++ ++enum mfill_atomic_mode { ++ MFILL_ATOMIC_COPY = 0, ++ MFILL_ATOMIC_ZEROPAGE = 1, ++ MFILL_ATOMIC_CONTINUE = 2, ++ NR_MFILL_ATOMIC_MODES = 3, ++}; ++ ++struct shmem_sb_info { ++ long unsigned int max_blocks; ++ struct percpu_counter used_blocks; ++ long unsigned int max_inodes; ++ long unsigned int free_inodes; ++ raw_spinlock_t stat_lock; ++ umode_t mode; ++ unsigned char huge; ++ kuid_t uid; ++ kgid_t gid; ++ bool full_inums; ++ bool noswap; ++ ino_t next_ino; ++ ino_t *ino_batch; ++ struct mempolicy *mpol; ++ spinlock_t shrinklist_lock; ++ struct list_head shrinklist; ++ long unsigned int shrinklist_len; ++}; ++ ++enum sgp_type { ++ SGP_READ = 0, ++ SGP_NOALLOC = 1, ++ SGP_CACHE = 2, ++ SGP_WRITE = 3, ++ SGP_FALLOC = 4, ++}; ++ ++struct shmem_falloc { ++ wait_queue_head_t *waitq; ++ long unsigned int start; ++ long unsigned int next; ++ long unsigned int nr_falloced; ++ long unsigned int nr_unswapped; ++}; ++ ++struct shmem_options { ++ long long unsigned int blocks; ++ long long unsigned int inodes; ++ struct mempolicy *mpol; ++ kuid_t uid; ++ kgid_t gid; ++ umode_t mode; ++ bool full_inums; ++ int huge; ++ int seen; ++ bool noswap; ++}; ++ ++enum shmem_param { ++ Opt_gid___3 = 0, ++ Opt_huge = 1, ++ Opt_mode = 2, ++ Opt_mpol = 3, ++ Opt_nr_blocks = 4, ++ Opt_nr_inodes = 5, ++ Opt_size = 6, ++ Opt_uid___3 = 7, ++ Opt_inode32 = 8, ++ Opt_inode64 = 9, ++ Opt_noswap = 10, ++}; ++ ++enum writeback_stat_item { ++ NR_DIRTY_THRESHOLD = 0, ++ NR_DIRTY_BG_THRESHOLD = 1, ++ NR_VM_WRITEBACK_STAT_ITEMS = 2, ++}; ++ ++struct contig_page_info { ++ long unsigned int free_pages; ++ long unsigned int free_blocks_total; ++ long unsigned int free_blocks_suitable; ++}; ++ ++typedef s8 pcp_op_T_____7; ++ ++struct memblock { ++ bool bottom_up; ++ phys_addr_t current_limit; ++ struct memblock_type memory; ++ struct memblock_type reserved; ++}; ++ ++enum mminit_level { ++ MMINIT_WARNING = 0, ++ MMINIT_VERIFY = 1, ++ MMINIT_TRACE = 2, ++}; ++ ++struct mminit_pfnnid_cache { ++ long unsigned int last_start; ++ long unsigned int last_end; ++ int last_nid; ++}; ++ ++struct pcpu_group_info { ++ int nr_units; ++ long unsigned int base_offset; ++ unsigned int *cpu_map; ++}; ++ ++struct pcpu_alloc_info { ++ size_t static_size; ++ size_t reserved_size; ++ size_t dyn_size; ++ size_t unit_size; ++ size_t atom_size; ++ size_t alloc_size; ++ size_t __ai_size; ++ int nr_groups; ++ struct pcpu_group_info groups[0]; ++}; ++ ++enum memcg_stat_item { ++ MEMCG_SWAP = 43, ++ MEMCG_SOCK = 44, ++ MEMCG_PERCPU_B = 45, ++ MEMCG_VMALLOC = 46, ++ MEMCG_KMEM = 47, ++ MEMCG_ZSWAP_B = 48, ++ MEMCG_ZSWAPPED = 49, ++ MEMCG_NR_STAT = 50, ++}; ++ ++struct trace_event_raw_percpu_alloc_percpu { ++ struct trace_entry ent; ++ long unsigned int call_site; ++ bool reserved; ++ bool is_atomic; ++ size_t size; ++ size_t align; ++ void *base_addr; ++ int off; ++ void *ptr; ++ size_t bytes_alloc; ++ long unsigned int gfp_flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_percpu_free_percpu { ++ struct trace_entry ent; ++ void *base_addr; ++ int off; ++ void *ptr; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_percpu_alloc_percpu_fail { ++ struct trace_entry ent; ++ bool reserved; ++ bool is_atomic; ++ size_t size; ++ size_t align; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_percpu_create_chunk { ++ struct trace_entry ent; ++ void *base_addr; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_percpu_destroy_chunk { ++ struct trace_entry ent; ++ void *base_addr; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_percpu_alloc_percpu {}; ++ ++struct trace_event_data_offsets_percpu_free_percpu {}; ++ ++struct trace_event_data_offsets_percpu_alloc_percpu_fail {}; ++ ++struct trace_event_data_offsets_percpu_create_chunk {}; ++ ++struct trace_event_data_offsets_percpu_destroy_chunk {}; ++ ++typedef void (*btf_trace_percpu_alloc_percpu)(void *, long unsigned int, bool, bool, size_t, size_t, void *, int, void *, size_t, gfp_t); ++ ++typedef void (*btf_trace_percpu_free_percpu)(void *, void *, int, void *); ++ ++typedef void (*btf_trace_percpu_alloc_percpu_fail)(void *, bool, bool, size_t, size_t); ++ ++typedef void (*btf_trace_percpu_create_chunk)(void *, void *); ++ ++typedef void (*btf_trace_percpu_destroy_chunk)(void *, void *); ++ ++struct pcpu_block_md { ++ int scan_hint; ++ int scan_hint_start; ++ int contig_hint; ++ int contig_hint_start; ++ int left_free; ++ int right_free; ++ int first_free; ++ int nr_bits; ++}; ++ ++struct pcpu_chunk { ++ struct list_head list; ++ int free_bytes; ++ struct pcpu_block_md chunk_md; ++ void *base_addr; ++ long unsigned int *alloc_map; ++ long unsigned int *bound_map; ++ struct pcpu_block_md *md_blocks; ++ void *data; ++ bool immutable; ++ bool isolated; ++ int start_offset; ++ int end_offset; ++ struct obj_cgroup **obj_cgroups; ++ int nr_pages; ++ int nr_populated; ++ int nr_empty_pop_pages; ++ long unsigned int populated[0]; ++}; ++ ++enum lru_status { ++ LRU_REMOVED = 0, ++ LRU_REMOVED_RETRY = 1, ++ LRU_ROTATE = 2, ++ LRU_SKIP = 3, ++ LRU_RETRY = 4, ++}; ++ ++typedef enum lru_status (*list_lru_walk_cb)(struct list_head *, struct list_lru_one *, spinlock_t *, void *); ++ ++struct trace_event_raw_mmap_lock { ++ struct trace_entry ent; ++ struct mm_struct *mm; ++ u32 __data_loc_memcg_path; ++ bool write; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mmap_lock_acquire_returned { ++ struct trace_entry ent; ++ struct mm_struct *mm; ++ u32 __data_loc_memcg_path; ++ bool write; ++ bool success; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_mmap_lock { ++ u32 memcg_path; ++}; ++ ++struct trace_event_data_offsets_mmap_lock_acquire_returned { ++ u32 memcg_path; ++}; ++ ++typedef void (*btf_trace_mmap_lock_start_locking)(void *, struct mm_struct *, const char *, bool); ++ ++typedef void (*btf_trace_mmap_lock_released)(void *, struct mm_struct *, const char *, bool); ++ ++typedef void (*btf_trace_mmap_lock_acquire_returned)(void *, struct mm_struct *, const char *, bool, bool); ++ ++struct memcg_path { ++ local_lock_t lock; ++ char *buf; ++ local_t buf_idx; ++}; ++ ++struct trace_event_raw_vm_unmapped_area { ++ struct trace_entry ent; ++ long unsigned int addr; ++ long unsigned int total_vm; ++ long unsigned int flags; ++ long unsigned int length; ++ long unsigned int low_limit; ++ long unsigned int high_limit; ++ long unsigned int align_mask; ++ long unsigned int align_offset; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_vma_mas_szero { ++ struct trace_entry ent; ++ struct maple_tree *mt; ++ long unsigned int start; ++ long unsigned int end; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_vma_store { ++ struct trace_entry ent; ++ struct maple_tree *mt; ++ struct vm_area_struct *vma; ++ long unsigned int vm_start; ++ long unsigned int vm_end; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_exit_mmap { ++ struct trace_entry ent; ++ struct mm_struct *mm; ++ struct maple_tree *mt; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_vm_unmapped_area {}; ++ ++struct trace_event_data_offsets_vma_mas_szero {}; ++ ++struct trace_event_data_offsets_vma_store {}; ++ ++struct trace_event_data_offsets_exit_mmap {}; ++ ++typedef void (*btf_trace_vm_unmapped_area)(void *, long unsigned int, struct vm_unmapped_area_info *); ++ ++typedef void (*btf_trace_vma_mas_szero)(void *, struct maple_tree *, long unsigned int, long unsigned int); ++ ++typedef void (*btf_trace_vma_store)(void *, struct maple_tree *, struct vm_area_struct *); ++ ++typedef void (*btf_trace_exit_mmap)(void *, struct mm_struct *); ++ ++struct vma_prepare { ++ struct vm_area_struct *vma; ++ struct vm_area_struct *adj_next; ++ struct file *file; ++ struct address_space *mapping; ++ struct anon_vma *anon_vma; ++ struct vm_area_struct *insert; ++ struct vm_area_struct *remove; ++ struct vm_area_struct *remove2; ++}; ++ ++struct mmap_arg_struct { ++ long unsigned int addr; ++ long unsigned int len; ++ long unsigned int prot; ++ long unsigned int flags; ++ long unsigned int fd; ++ long unsigned int offset; ++}; ++ ++struct vmap_area { ++ long unsigned int va_start; ++ long unsigned int va_end; ++ struct rb_node rb_node; ++ struct list_head list; ++ union { ++ long unsigned int subtree_max_size; ++ struct vm_struct *vm; ++ }; ++ long unsigned int flags; ++}; ++ ++typedef unsigned int kasan_vmalloc_flags_t; ++ ++struct trace_event_raw_alloc_vmap_area { ++ struct trace_entry ent; ++ long unsigned int addr; ++ long unsigned int size; ++ long unsigned int align; ++ long unsigned int vstart; ++ long unsigned int vend; ++ int failed; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_purge_vmap_area_lazy { ++ struct trace_entry ent; ++ long unsigned int start; ++ long unsigned int end; ++ unsigned int npurged; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_free_vmap_area_noflush { ++ struct trace_entry ent; ++ long unsigned int va_start; ++ long unsigned int nr_lazy; ++ long unsigned int nr_lazy_max; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_alloc_vmap_area {}; ++ ++struct trace_event_data_offsets_purge_vmap_area_lazy {}; ++ ++struct trace_event_data_offsets_free_vmap_area_noflush {}; ++ ++typedef void (*btf_trace_alloc_vmap_area)(void *, long unsigned int, long unsigned int, long unsigned int, long unsigned int, long unsigned int, int); ++ ++typedef void (*btf_trace_purge_vmap_area_lazy)(void *, long unsigned int, long unsigned int, unsigned int); ++ ++typedef void (*btf_trace_free_vmap_area_noflush)(void *, long unsigned int, long unsigned int, long unsigned int); ++ ++struct vfree_deferred { ++ struct llist_head list; ++ struct work_struct wq; ++}; ++ ++enum fit_type { ++ NOTHING_FIT = 0, ++ FL_FIT_TYPE = 1, ++ LE_FIT_TYPE = 2, ++ RE_FIT_TYPE = 3, ++ NE_FIT_TYPE = 4, ++}; ++ ++struct vmap_block_queue { ++ spinlock_t lock; ++ struct list_head free; ++ struct xarray vmap_blocks; ++}; ++ ++struct vmap_block { ++ spinlock_t lock; ++ struct vmap_area *va; ++ long unsigned int free; ++ long unsigned int dirty; ++ long unsigned int used_map[16]; ++ long unsigned int dirty_min; ++ long unsigned int dirty_max; ++ struct list_head free_list; ++ struct callback_head callback_head; ++ struct list_head purge; ++}; ++ ++struct madvise_walk_private { ++ struct mmu_gather *tlb; ++ bool pageout; ++}; ++ ++struct swap_iocb { ++ struct kiocb iocb; ++ struct bio_vec bvec[32]; ++ int pages; ++ int len; ++}; ++ ++struct swap_slots_cache { ++ bool lock_initialized; ++ struct mutex alloc_lock; ++ swp_entry_t *slots; ++ int nr; ++ int cur; ++ spinlock_t free_lock; ++ swp_entry_t *slots_ret; ++ int n_ret; ++}; ++ ++struct dma_block { ++ struct dma_block *next_block; ++ dma_addr_t dma; ++}; ++ ++struct dma_pool { ++ struct list_head page_list; ++ spinlock_t lock; ++ struct dma_block *next_block; ++ size_t nr_blocks; ++ size_t nr_active; ++ size_t nr_pages; ++ struct device *dev; ++ unsigned int size; ++ unsigned int allocation; ++ unsigned int boundary; ++ char name[32]; ++ struct list_head pools; ++}; ++ ++struct dma_page { ++ struct list_head page_list; ++ void *vaddr; ++ dma_addr_t dma; ++}; ++ ++enum string_size_units { ++ STRING_UNITS_10 = 0, ++ STRING_UNITS_2 = 1, ++}; ++ ++struct resv_map { ++ struct kref refs; ++ spinlock_t lock; ++ struct list_head regions; ++ long int adds_in_progress; ++ struct list_head region_cache; ++ long int region_cache_count; ++}; ++ ++struct file_region { ++ struct list_head link; ++ long int from; ++ long int to; ++}; ++ ++struct hugetlb_vma_lock { ++ struct kref refs; ++ struct rw_semaphore rw_sema; ++ struct vm_area_struct *vma; ++}; ++ ++struct huge_bootmem_page { ++ struct list_head list; ++ struct hstate *hstate; ++}; ++ ++enum { ++ FOLL_TOUCH = 65536, ++ FOLL_TRIED = 131072, ++ FOLL_REMOTE = 262144, ++ FOLL_PIN = 524288, ++ FOLL_FAST_ONLY = 1048576, ++ FOLL_UNLOCKABLE = 2097152, ++}; ++ ++enum vma_resv_mode { ++ VMA_NEEDS_RESV = 0, ++ VMA_COMMIT_RESV = 1, ++ VMA_END_RESV = 2, ++ VMA_ADD_RESV = 3, ++ VMA_DEL_RESV = 4, ++}; ++ ++struct node_hstate { ++ struct kobject *hugepages_kobj; ++ struct kobject *hstate_kobjs[2]; ++}; ++ ++struct hugetlb_cgroup; ++ ++struct mmu_notifier_subscriptions { ++ struct hlist_head list; ++ bool has_itree; ++ spinlock_t lock; ++ long unsigned int invalidate_seq; ++ long unsigned int active_invalidate_ranges; ++ struct rb_root_cached itree; ++ wait_queue_head_t wq; ++ struct hlist_head deferred_list; ++}; ++ ++struct mmu_interval_notifier; ++ ++struct mmu_interval_notifier_ops { ++ bool (*invalidate)(struct mmu_interval_notifier *, const struct mmu_notifier_range *, long unsigned int); ++}; ++ ++struct mmu_interval_notifier { ++ struct interval_tree_node interval_tree; ++ const struct mmu_interval_notifier_ops *ops; ++ struct mm_struct *mm; ++ struct hlist_node deferred_item; ++ long unsigned int invalidate_seq; ++}; ++ ++struct kcsan_scoped_access {}; ++ ++enum kfence_object_state { ++ KFENCE_OBJECT_UNUSED = 0, ++ KFENCE_OBJECT_ALLOCATED = 1, ++ KFENCE_OBJECT_FREED = 2, ++}; ++ ++struct kfence_track { ++ pid_t pid; ++ int cpu; ++ u64 ts_nsec; ++ int num_stack_entries; ++ long unsigned int stack_entries[64]; ++}; ++ ++struct kfence_metadata { ++ struct list_head list; ++ struct callback_head callback_head; ++ raw_spinlock_t lock; ++ enum kfence_object_state state; ++ long unsigned int addr; ++ size_t size; ++ struct kmem_cache *cache; ++ long unsigned int unprotected_page; ++ struct kfence_track alloc_track; ++ struct kfence_track free_track; ++ u32 alloc_stack_hash; ++ struct obj_cgroup *objcg; ++}; ++ ++enum kfence_error_type { ++ KFENCE_ERROR_OOB = 0, ++ KFENCE_ERROR_UAF = 1, ++ KFENCE_ERROR_CORRUPTION = 2, ++ KFENCE_ERROR_INVALID = 3, ++ KFENCE_ERROR_INVALID_FREE = 4, ++}; ++ ++enum kfence_counter_id { ++ KFENCE_COUNTER_ALLOCATED = 0, ++ KFENCE_COUNTER_ALLOCS = 1, ++ KFENCE_COUNTER_FREES = 2, ++ KFENCE_COUNTER_ZOMBIES = 3, ++ KFENCE_COUNTER_BUGS = 4, ++ KFENCE_COUNTER_SKIP_INCOMPAT = 5, ++ KFENCE_COUNTER_SKIP_CAPACITY = 6, ++ KFENCE_COUNTER_SKIP_COVERED = 7, ++ KFENCE_COUNTER_COUNT = 8, ++}; ++ ++union migration_ptr { ++ struct anon_vma *anon_vma; ++ struct address_space *mapping; ++}; ++ ++struct migrate_pages_stats { ++ int nr_succeeded; ++ int nr_failed_pages; ++ int nr_thp_succeeded; ++ int nr_thp_failed; ++ int nr_thp_split; ++}; ++ ++struct mem_cgroup_reclaim_cookie { ++ pg_data_t *pgdat; ++ unsigned int generation; ++}; ++ ++enum mem_cgroup_events_target { ++ MEM_CGROUP_TARGET_THRESH = 0, ++ MEM_CGROUP_TARGET_SOFTLIMIT = 1, ++ MEM_CGROUP_NTARGETS = 2, ++}; ++ ++struct memcg_vmstats { ++ long int state[50]; ++ long unsigned int events[17]; ++ long int state_pending[50]; ++ long unsigned int events_pending[17]; ++}; ++ ++struct memcg_vmstats_percpu { ++ long int state[50]; ++ long unsigned int events[17]; ++ long int state_prev[50]; ++ long unsigned int events_prev[17]; ++ long unsigned int nr_page_events; ++ long unsigned int targets[2]; ++}; ++ ++struct mem_cgroup_tree_per_node { ++ struct rb_root rb_root; ++ struct rb_node *rb_rightmost; ++ spinlock_t lock; ++}; ++ ++struct mem_cgroup_tree { ++ struct mem_cgroup_tree_per_node *rb_tree_per_node[2]; ++}; ++ ++struct mem_cgroup_eventfd_list { ++ struct list_head list; ++ struct eventfd_ctx *eventfd; ++}; ++ ++struct mem_cgroup_event { ++ struct mem_cgroup *memcg; ++ struct eventfd_ctx *eventfd; ++ struct list_head list; ++ int (*register_event)(struct mem_cgroup *, struct eventfd_ctx *, const char *); ++ void (*unregister_event)(struct mem_cgroup *, struct eventfd_ctx *); ++ poll_table pt; ++ wait_queue_head_t *wqh; ++ wait_queue_entry_t wait; ++ struct work_struct remove; ++}; ++ ++struct move_charge_struct { ++ spinlock_t lock; ++ struct mm_struct *mm; ++ struct mem_cgroup *from; ++ struct mem_cgroup *to; ++ long unsigned int flags; ++ long unsigned int precharge; ++ long unsigned int moved_charge; ++ long unsigned int moved_swap; ++ struct task_struct *moving_task; ++ wait_queue_head_t waitq; ++}; ++ ++enum res_type { ++ _MEM = 0, ++ _MEMSWAP = 1, ++ _KMEM = 2, ++ _TCP = 3, ++}; ++ ++struct memory_stat { ++ const char *name; ++ unsigned int idx; ++}; ++ ++struct oom_wait_info { ++ struct mem_cgroup *memcg; ++ wait_queue_entry_t wait; ++}; ++ ++struct memcg_stock_pcp { ++ local_lock_t stock_lock; ++ struct mem_cgroup *cached; ++ unsigned int nr_pages; ++ struct obj_cgroup *cached_objcg; ++ struct pglist_data *cached_pgdat; ++ unsigned int nr_bytes; ++ int nr_slab_reclaimable_b; ++ int nr_slab_unreclaimable_b; ++ struct work_struct work; ++ long unsigned int flags; ++}; ++ ++enum { ++ RES_USAGE = 0, ++ RES_LIMIT = 1, ++ RES_MAX_USAGE = 2, ++ RES_FAILCNT = 3, ++ RES_SOFT_LIMIT = 4, ++}; ++ ++union mc_target { ++ struct page *page; ++ swp_entry_t ent; ++}; ++ ++enum mc_target_type { ++ MC_TARGET_NONE = 0, ++ MC_TARGET_PAGE = 1, ++ MC_TARGET_SWAP = 2, ++ MC_TARGET_DEVICE = 3, ++}; ++ ++struct uncharge_gather { ++ struct mem_cgroup *memcg; ++ long unsigned int nr_memory; ++ long unsigned int pgpgout; ++ long unsigned int nr_kmem; ++ int nid; ++}; ++ ++struct numa_stat { ++ const char *name; ++ unsigned int lru_mask; ++}; ++ ++struct zpool { ++ struct zpool_driver *driver; ++ void *pool; ++}; ++ ++enum zs_mapmode { ++ ZS_MM_RW = 0, ++ ZS_MM_RO = 1, ++ ZS_MM_WO = 2, ++}; ++ ++struct zs_pool_stats { ++ atomic_long_t pages_compacted; ++}; ++ ++enum fullness_group { ++ ZS_INUSE_RATIO_0 = 0, ++ ZS_INUSE_RATIO_10 = 1, ++ ZS_INUSE_RATIO_99 = 10, ++ ZS_INUSE_RATIO_100 = 11, ++ NR_FULLNESS_GROUPS = 12, ++}; ++ ++enum class_stat_type { ++ ZS_OBJS_ALLOCATED = 12, ++ ZS_OBJS_INUSE = 13, ++ NR_CLASS_STAT_TYPES = 14, ++}; ++ ++struct zs_size_stat { ++ long unsigned int objs[14]; ++}; ++ ++struct size_class { ++ struct list_head fullness_list[12]; ++ int size; ++ int objs_per_zspage; ++ int pages_per_zspage; ++ unsigned int index; ++ struct zs_size_stat stats; ++}; ++ ++struct link_free { ++ union { ++ long unsigned int next; ++ long unsigned int handle; ++ long unsigned int deferred_handle; ++ }; ++}; ++ ++struct zs_pool { ++ const char *name; ++ struct size_class *size_class[255]; ++ struct kmem_cache *handle_cachep; ++ struct kmem_cache *zspage_cachep; ++ atomic_long_t pages_allocated; ++ struct zs_pool_stats stats; ++ struct shrinker shrinker; ++ struct list_head lru; ++ struct zpool *zpool; ++ const struct zpool_ops *zpool_ops; ++ struct work_struct free_work; ++ spinlock_t lock; ++ atomic_t compaction_in_progress; ++}; ++ ++struct zspage { ++ struct { ++ unsigned int huge: 1; ++ unsigned int fullness: 4; ++ unsigned int class: 9; ++ unsigned int isolated: 5; ++ unsigned int magic: 8; ++ }; ++ unsigned int inuse; ++ unsigned int freeobj; ++ struct page *first_page; ++ struct list_head list; ++ struct list_head lru; ++ bool under_reclaim; ++ struct zs_pool *pool; ++ rwlock_t lock; ++}; ++ ++struct mapping_area { ++ local_lock_t lock; ++ char *vm_buf; ++ char *vm_addr; ++ enum zs_mapmode vm_mm; ++}; ++ ++struct zs_compact_control { ++ struct page *s_page; ++ struct page *d_page; ++ int obj_idx; ++}; ++ ++struct trace_event_raw_cma_release { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ long unsigned int pfn; ++ const struct page *page; ++ long unsigned int count; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_cma_alloc_start { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ long unsigned int count; ++ unsigned int align; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_cma_alloc_finish { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ long unsigned int pfn; ++ const struct page *page; ++ long unsigned int count; ++ unsigned int align; ++ int errorno; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_cma_alloc_busy_retry { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ long unsigned int pfn; ++ const struct page *page; ++ long unsigned int count; ++ unsigned int align; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_cma_release { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_cma_alloc_start { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_cma_alloc_finish { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_cma_alloc_busy_retry { ++ u32 name; ++}; ++ ++typedef void (*btf_trace_cma_release)(void *, const char *, long unsigned int, const struct page *, long unsigned int); ++ ++typedef void (*btf_trace_cma_alloc_start)(void *, const char *, long unsigned int, unsigned int); ++ ++typedef void (*btf_trace_cma_alloc_finish)(void *, const char *, long unsigned int, const struct page *, long unsigned int, unsigned int, int); ++ ++typedef void (*btf_trace_cma_alloc_busy_retry)(void *, const char *, long unsigned int, const struct page *, long unsigned int, unsigned int); ++ ++struct page_ext_operations { ++ size_t offset; ++ size_t size; ++ bool (*need)(); ++ void (*init)(); ++ bool need_shared_flags; ++}; ++ ++struct damon_sysfs_scheme; ++ ++struct damon_sysfs_schemes { ++ struct kobject kobj; ++ struct damon_sysfs_scheme **schemes_arr; ++ int nr; ++}; ++ ++struct damon_sysfs_access_pattern; ++ ++struct damon_sysfs_quotas; ++ ++struct damon_sysfs_watermarks; ++ ++struct damon_sysfs_scheme_filters; ++ ++struct damon_sysfs_stats; ++ ++struct damon_sysfs_scheme_regions; ++ ++struct damon_sysfs_scheme { ++ struct kobject kobj; ++ enum damos_action action; ++ struct damon_sysfs_access_pattern *access_pattern; ++ struct damon_sysfs_quotas *quotas; ++ struct damon_sysfs_watermarks *watermarks; ++ struct damon_sysfs_scheme_filters *filters; ++ struct damon_sysfs_stats *stats; ++ struct damon_sysfs_scheme_regions *tried_regions; ++}; ++ ++struct damon_sysfs_scheme_region { ++ struct kobject kobj; ++ struct damon_addr_range ar; ++ unsigned int nr_accesses; ++ unsigned int age; ++ struct list_head list; ++}; ++ ++struct damon_sysfs_scheme_regions { ++ struct kobject kobj; ++ struct list_head regions_list; ++ int nr_regions; ++}; ++ ++struct damon_sysfs_stats { ++ struct kobject kobj; ++ long unsigned int nr_tried; ++ long unsigned int sz_tried; ++ long unsigned int nr_applied; ++ long unsigned int sz_applied; ++ long unsigned int qt_exceeds; ++}; ++ ++struct damon_sysfs_scheme_filter { ++ struct kobject kobj; ++ enum damos_filter_type type; ++ bool matching; ++ char *memcg_path; ++}; ++ ++struct damon_sysfs_scheme_filters { ++ struct kobject kobj; ++ struct damon_sysfs_scheme_filter **filters_arr; ++ int nr; ++}; ++ ++struct damon_sysfs_watermarks { ++ struct kobject kobj; ++ enum damos_wmark_metric metric; ++ long unsigned int interval_us; ++ long unsigned int high; ++ long unsigned int mid; ++ long unsigned int low; ++}; ++ ++struct damon_sysfs_weights { ++ struct kobject kobj; ++ unsigned int sz; ++ unsigned int nr_accesses; ++ unsigned int age; ++}; ++ ++struct damon_sysfs_quotas { ++ struct kobject kobj; ++ struct damon_sysfs_weights *weights; ++ long unsigned int ms; ++ long unsigned int sz; ++ long unsigned int reset_interval_ms; ++}; ++ ++struct damon_sysfs_access_pattern { ++ struct kobject kobj; ++ struct damon_sysfs_ul_range *sz; ++ struct damon_sysfs_ul_range *nr_accesses; ++ struct damon_sysfs_ul_range *age; ++}; ++ ++enum hmm_pfn_flags { ++ HMM_PFN_VALID = 9223372036854775808ULL, ++ HMM_PFN_WRITE = 4611686018427387904ULL, ++ HMM_PFN_ERROR = 2305843009213693952ULL, ++ HMM_PFN_ORDER_SHIFT = 56ULL, ++ HMM_PFN_REQ_FAULT = 9223372036854775808ULL, ++ HMM_PFN_REQ_WRITE = 4611686018427387904ULL, ++ HMM_PFN_FLAGS = 18374686479671623680ULL, ++}; ++ ++struct hmm_range { ++ struct mmu_interval_notifier *notifier; ++ long unsigned int notifier_seq; ++ long unsigned int start; ++ long unsigned int end; ++ long unsigned int *hmm_pfns; ++ long unsigned int default_flags; ++ long unsigned int pfn_flags_mask; ++ void *dev_private_owner; ++}; ++ ++struct hmm_vma_walk { ++ struct hmm_range *range; ++ long unsigned int last; ++}; ++ ++enum { ++ HMM_NEED_FAULT = 1, ++ HMM_NEED_WRITE_FAULT = 2, ++ HMM_NEED_ALL_BITS = 3, ++}; ++ ++struct page_reporting_dev_info { ++ int (*report)(struct page_reporting_dev_info *, struct scatterlist *, unsigned int); ++ struct delayed_work work; ++ atomic_t state; ++ unsigned int order; ++}; ++ ++enum { ++ PAGE_REPORTING_IDLE = 0, ++ PAGE_REPORTING_REQUESTED = 1, ++ PAGE_REPORTING_ACTIVE = 2, ++}; ++ ++struct files_stat_struct { ++ long unsigned int nr_files; ++ long unsigned int nr_free_files; ++ long unsigned int max_files; ++}; ++ ++typedef struct kobject *kobj_probe_t(dev_t, int *, void *); ++ ++struct char_device_struct { ++ struct char_device_struct *next; ++ unsigned int major; ++ unsigned int baseminor; ++ int minorct; ++ char name[64]; ++ struct cdev *cdev; ++}; ++ ++struct stat { ++ long unsigned int st_dev; ++ long unsigned int st_ino; ++ long unsigned int st_nlink; ++ unsigned int st_mode; ++ unsigned int st_uid; ++ unsigned int st_gid; ++ unsigned int __pad1; ++ long unsigned int st_rdev; ++ long unsigned int st_size; ++ long unsigned int st_atime; ++ long unsigned int st_atime_nsec; ++ long unsigned int st_mtime; ++ long unsigned int st_mtime_nsec; ++ long unsigned int st_ctime; ++ long unsigned int st_ctime_nsec; ++ long unsigned int st_blksize; ++ long int st_blocks; ++ long unsigned int __unused[3]; ++}; ++ ++typedef u16 compat_dev_t; ++ ++typedef u16 compat_nlink_t; ++ ++struct compat_stat { ++ compat_dev_t st_dev; ++ u16 __pad1; ++ compat_ino_t st_ino; ++ compat_mode_t st_mode; ++ compat_nlink_t st_nlink; ++ __compat_uid_t st_uid; ++ __compat_gid_t st_gid; ++ compat_dev_t st_rdev; ++ u16 __pad2; ++ u32 st_size; ++ u32 st_blksize; ++ u32 st_blocks; ++ u32 st_atime; ++ u32 st_atime_nsec; ++ u32 st_mtime; ++ u32 st_mtime_nsec; ++ u32 st_ctime; ++ u32 st_ctime_nsec; ++ u32 __unused4; ++ u32 __unused5; ++}; ++ ++struct saved { ++ struct path link; ++ struct delayed_call done; ++ const char *name; ++ unsigned int seq; ++}; ++ ++struct nameidata { ++ struct path path; ++ struct qstr last; ++ struct path root; ++ struct inode *inode; ++ unsigned int flags; ++ unsigned int state; ++ unsigned int seq; ++ unsigned int next_seq; ++ unsigned int m_seq; ++ unsigned int r_seq; ++ int last_type; ++ unsigned int depth; ++ int total_link_count; ++ struct saved *stack; ++ struct saved internal[2]; ++ struct filename *name; ++ struct nameidata *saved; ++ unsigned int root_seq; ++ int dfd; ++ vfsuid_t dir_vfsuid; ++ umode_t dir_mode; ++}; ++ ++struct renamedata { ++ struct mnt_idmap *old_mnt_idmap; ++ struct inode *old_dir; ++ struct dentry *old_dentry; ++ struct mnt_idmap *new_mnt_idmap; ++ struct inode *new_dir; ++ struct dentry *new_dentry; ++ struct inode **delegated_inode; ++ unsigned int flags; ++}; ++ ++enum { ++ LAST_NORM = 0, ++ LAST_ROOT = 1, ++ LAST_DOT = 2, ++ LAST_DOTDOT = 3, ++}; ++ ++typedef int filler_t(struct file *, struct folio *); ++ ++enum { ++ WALK_TRAILING = 1, ++ WALK_MORE = 2, ++ WALK_NOFOLLOW = 4, ++}; ++ ++typedef struct { ++ long unsigned int fds_bits[16]; ++} __kernel_fd_set; ++ ++typedef __kernel_fd_set fd_set; ++ ++struct poll_table_entry { ++ struct file *filp; ++ __poll_t key; ++ wait_queue_entry_t wait; ++ wait_queue_head_t *wait_address; ++}; ++ ++struct poll_table_page; ++ ++struct poll_wqueues { ++ poll_table pt; ++ struct poll_table_page *table; ++ struct task_struct *polling_task; ++ int triggered; ++ int error; ++ int inline_index; ++ struct poll_table_entry inline_entries[9]; ++}; ++ ++struct poll_table_page { ++ struct poll_table_page *next; ++ struct poll_table_entry *entry; ++ struct poll_table_entry entries[0]; ++}; ++ ++enum poll_time_type { ++ PT_TIMEVAL = 0, ++ PT_OLD_TIMEVAL = 1, ++ PT_TIMESPEC = 2, ++ PT_OLD_TIMESPEC = 3, ++}; ++ ++typedef struct { ++ long unsigned int *in; ++ long unsigned int *out; ++ long unsigned int *ex; ++ long unsigned int *res_in; ++ long unsigned int *res_out; ++ long unsigned int *res_ex; ++} fd_set_bits; ++ ++struct sigset_argpack { ++ sigset_t *p; ++ size_t size; ++}; ++ ++struct poll_list { ++ struct poll_list *next; ++ int len; ++ struct pollfd entries[0]; ++}; ++ ++struct compat_sel_arg_struct { ++ compat_ulong_t n; ++ compat_uptr_t inp; ++ compat_uptr_t outp; ++ compat_uptr_t exp; ++ compat_uptr_t tvp; ++}; ++ ++struct compat_sigset_argpack { ++ compat_uptr_t p; ++ compat_size_t size; ++}; ++ ++struct inodes_stat_t { ++ long int nr_inodes; ++ long int nr_unused; ++ long int dummy[5]; ++}; ++ ++enum file_time_flags { ++ S_ATIME = 1, ++ S_MTIME = 2, ++ S_CTIME = 4, ++ S_VERSION = 8, ++}; ++ ++struct wb_writeback_work { ++ long int nr_pages; ++ struct super_block *sb; ++ enum writeback_sync_modes sync_mode; ++ unsigned int tagged_writepages: 1; ++ unsigned int for_kupdate: 1; ++ unsigned int range_cyclic: 1; ++ unsigned int for_background: 1; ++ unsigned int for_sync: 1; ++ unsigned int auto_free: 1; ++ enum wb_reason reason; ++ struct list_head list; ++ struct wb_completion *done; ++}; ++ ++struct trace_event_raw_writeback_folio_template { ++ struct trace_entry ent; ++ char name[32]; ++ ino_t ino; ++ long unsigned int index; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_writeback_dirty_inode_template { ++ struct trace_entry ent; ++ char name[32]; ++ ino_t ino; ++ long unsigned int state; ++ long unsigned int flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_inode_foreign_history { ++ struct trace_entry ent; ++ char name[32]; ++ ino_t ino; ++ ino_t cgroup_ino; ++ unsigned int history; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_inode_switch_wbs { ++ struct trace_entry ent; ++ char name[32]; ++ ino_t ino; ++ ino_t old_cgroup_ino; ++ ino_t new_cgroup_ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_track_foreign_dirty { ++ struct trace_entry ent; ++ char name[32]; ++ u64 bdi_id; ++ ino_t ino; ++ unsigned int memcg_id; ++ ino_t cgroup_ino; ++ ino_t page_cgroup_ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_flush_foreign { ++ struct trace_entry ent; ++ char name[32]; ++ ino_t cgroup_ino; ++ unsigned int frn_bdi_id; ++ unsigned int frn_memcg_id; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_writeback_write_inode_template { ++ struct trace_entry ent; ++ char name[32]; ++ ino_t ino; ++ int sync_mode; ++ ino_t cgroup_ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_writeback_work_class { ++ struct trace_entry ent; ++ char name[32]; ++ long int nr_pages; ++ dev_t sb_dev; ++ int sync_mode; ++ int for_kupdate; ++ int range_cyclic; ++ int for_background; ++ int reason; ++ ino_t cgroup_ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_writeback_pages_written { ++ struct trace_entry ent; ++ long int pages; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_writeback_class { ++ struct trace_entry ent; ++ char name[32]; ++ ino_t cgroup_ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_writeback_bdi_register { ++ struct trace_entry ent; ++ char name[32]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_wbc_class { ++ struct trace_entry ent; ++ char name[32]; ++ long int nr_to_write; ++ long int pages_skipped; ++ int sync_mode; ++ int for_kupdate; ++ int for_background; ++ int for_reclaim; ++ int range_cyclic; ++ long int range_start; ++ long int range_end; ++ ino_t cgroup_ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_writeback_queue_io { ++ struct trace_entry ent; ++ char name[32]; ++ long unsigned int older; ++ long int age; ++ int moved; ++ int reason; ++ ino_t cgroup_ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_global_dirty_state { ++ struct trace_entry ent; ++ long unsigned int nr_dirty; ++ long unsigned int nr_writeback; ++ long unsigned int background_thresh; ++ long unsigned int dirty_thresh; ++ long unsigned int dirty_limit; ++ long unsigned int nr_dirtied; ++ long unsigned int nr_written; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_bdi_dirty_ratelimit { ++ struct trace_entry ent; ++ char bdi[32]; ++ long unsigned int write_bw; ++ long unsigned int avg_write_bw; ++ long unsigned int dirty_rate; ++ long unsigned int dirty_ratelimit; ++ long unsigned int task_ratelimit; ++ long unsigned int balanced_dirty_ratelimit; ++ ino_t cgroup_ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_balance_dirty_pages { ++ struct trace_entry ent; ++ char bdi[32]; ++ long unsigned int limit; ++ long unsigned int setpoint; ++ long unsigned int dirty; ++ long unsigned int bdi_setpoint; ++ long unsigned int bdi_dirty; ++ long unsigned int dirty_ratelimit; ++ long unsigned int task_ratelimit; ++ unsigned int dirtied; ++ unsigned int dirtied_pause; ++ long unsigned int paused; ++ long int pause; ++ long unsigned int period; ++ long int think; ++ ino_t cgroup_ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_writeback_sb_inodes_requeue { ++ struct trace_entry ent; ++ char name[32]; ++ ino_t ino; ++ long unsigned int state; ++ long unsigned int dirtied_when; ++ ino_t cgroup_ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_writeback_single_inode_template { ++ struct trace_entry ent; ++ char name[32]; ++ ino_t ino; ++ long unsigned int state; ++ long unsigned int dirtied_when; ++ long unsigned int writeback_index; ++ long int nr_to_write; ++ long unsigned int wrote; ++ ino_t cgroup_ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_writeback_inode_template { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ long unsigned int state; ++ __u16 mode; ++ long unsigned int dirtied_when; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_writeback_folio_template {}; ++ ++struct trace_event_data_offsets_writeback_dirty_inode_template {}; ++ ++struct trace_event_data_offsets_inode_foreign_history {}; ++ ++struct trace_event_data_offsets_inode_switch_wbs {}; ++ ++struct trace_event_data_offsets_track_foreign_dirty {}; ++ ++struct trace_event_data_offsets_flush_foreign {}; ++ ++struct trace_event_data_offsets_writeback_write_inode_template {}; ++ ++struct trace_event_data_offsets_writeback_work_class {}; ++ ++struct trace_event_data_offsets_writeback_pages_written {}; ++ ++struct trace_event_data_offsets_writeback_class {}; ++ ++struct trace_event_data_offsets_writeback_bdi_register {}; ++ ++struct trace_event_data_offsets_wbc_class {}; ++ ++struct trace_event_data_offsets_writeback_queue_io {}; ++ ++struct trace_event_data_offsets_global_dirty_state {}; ++ ++struct trace_event_data_offsets_bdi_dirty_ratelimit {}; ++ ++struct trace_event_data_offsets_balance_dirty_pages {}; ++ ++struct trace_event_data_offsets_writeback_sb_inodes_requeue {}; ++ ++struct trace_event_data_offsets_writeback_single_inode_template {}; ++ ++struct trace_event_data_offsets_writeback_inode_template {}; ++ ++typedef void (*btf_trace_writeback_dirty_folio)(void *, struct folio *, struct address_space *); ++ ++typedef void (*btf_trace_folio_wait_writeback)(void *, struct folio *, struct address_space *); ++ ++typedef void (*btf_trace_writeback_mark_inode_dirty)(void *, struct inode *, int); ++ ++typedef void (*btf_trace_writeback_dirty_inode_start)(void *, struct inode *, int); ++ ++typedef void (*btf_trace_writeback_dirty_inode)(void *, struct inode *, int); ++ ++typedef void (*btf_trace_inode_foreign_history)(void *, struct inode *, struct writeback_control *, unsigned int); ++ ++typedef void (*btf_trace_inode_switch_wbs)(void *, struct inode *, struct bdi_writeback *, struct bdi_writeback *); ++ ++typedef void (*btf_trace_track_foreign_dirty)(void *, struct folio *, struct bdi_writeback *); ++ ++typedef void (*btf_trace_flush_foreign)(void *, struct bdi_writeback *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_writeback_write_inode_start)(void *, struct inode *, struct writeback_control *); ++ ++typedef void (*btf_trace_writeback_write_inode)(void *, struct inode *, struct writeback_control *); ++ ++typedef void (*btf_trace_writeback_queue)(void *, struct bdi_writeback *, struct wb_writeback_work *); ++ ++typedef void (*btf_trace_writeback_exec)(void *, struct bdi_writeback *, struct wb_writeback_work *); ++ ++typedef void (*btf_trace_writeback_start)(void *, struct bdi_writeback *, struct wb_writeback_work *); ++ ++typedef void (*btf_trace_writeback_written)(void *, struct bdi_writeback *, struct wb_writeback_work *); ++ ++typedef void (*btf_trace_writeback_wait)(void *, struct bdi_writeback *, struct wb_writeback_work *); ++ ++typedef void (*btf_trace_writeback_pages_written)(void *, long int); ++ ++typedef void (*btf_trace_writeback_wake_background)(void *, struct bdi_writeback *); ++ ++typedef void (*btf_trace_writeback_bdi_register)(void *, struct backing_dev_info *); ++ ++typedef void (*btf_trace_wbc_writepage)(void *, struct writeback_control *, struct backing_dev_info *); ++ ++typedef void (*btf_trace_writeback_queue_io)(void *, struct bdi_writeback *, struct wb_writeback_work *, long unsigned int, int); ++ ++typedef void (*btf_trace_global_dirty_state)(void *, long unsigned int, long unsigned int); ++ ++typedef void (*btf_trace_bdi_dirty_ratelimit)(void *, struct bdi_writeback *, long unsigned int, long unsigned int); ++ ++typedef void (*btf_trace_balance_dirty_pages)(void *, struct bdi_writeback *, long unsigned int, long unsigned int, long unsigned int, long unsigned int, long unsigned int, long unsigned int, long unsigned int, long unsigned int, long unsigned int, long int, long unsigned int); ++ ++typedef void (*btf_trace_writeback_sb_inodes_requeue)(void *, struct inode *); ++ ++typedef void (*btf_trace_writeback_single_inode_start)(void *, struct inode *, struct writeback_control *, long unsigned int); ++ ++typedef void (*btf_trace_writeback_single_inode)(void *, struct inode *, struct writeback_control *, long unsigned int); ++ ++typedef void (*btf_trace_writeback_lazytime)(void *, struct inode *); ++ ++typedef void (*btf_trace_writeback_lazytime_iput)(void *, struct inode *); ++ ++typedef void (*btf_trace_writeback_dirty_inode_enqueue)(void *, struct inode *); ++ ++typedef void (*btf_trace_sb_mark_inode_writeback)(void *, struct inode *); ++ ++typedef void (*btf_trace_sb_clear_inode_writeback)(void *, struct inode *); ++ ++struct inode_switch_wbs_context { ++ struct rcu_work work; ++ struct bdi_writeback *new_wb; ++ struct inode *inodes[0]; ++}; ++ ++struct old_utimbuf32 { ++ old_time32_t actime; ++ old_time32_t modtime; ++}; ++ ++struct utimbuf { ++ __kernel_old_time_t actime; ++ __kernel_old_time_t modtime; ++}; ++ ++struct prepend_buffer { ++ char *buf; ++ int len; ++}; ++ ++struct ns_get_path_task_args { ++ const struct proc_ns_operations *ns_ops; ++ struct task_struct *task; ++}; ++ ++enum fsconfig_command { ++ FSCONFIG_SET_FLAG = 0, ++ FSCONFIG_SET_STRING = 1, ++ FSCONFIG_SET_BINARY = 2, ++ FSCONFIG_SET_PATH = 3, ++ FSCONFIG_SET_PATH_EMPTY = 4, ++ FSCONFIG_SET_FD = 5, ++ FSCONFIG_CMD_CREATE = 6, ++ FSCONFIG_CMD_RECONFIGURE = 7, ++}; ++ ++struct file_dedupe_range_info { ++ __s64 dest_fd; ++ __u64 dest_offset; ++ __u64 bytes_deduped; ++ __s32 status; ++ __u32 reserved; ++}; ++ ++struct file_dedupe_range { ++ __u64 src_offset; ++ __u64 src_length; ++ __u16 dest_count; ++ __u16 reserved1; ++ __u32 reserved2; ++ struct file_dedupe_range_info info[0]; ++}; ++ ++struct proc_fs_opts { ++ int flag; ++ const char *str; ++}; ++ ++typedef int dio_iodone_t(struct kiocb *, loff_t, ssize_t, void *); ++ ++enum { ++ DIO_LOCKING = 1, ++ DIO_SKIP_HOLES = 2, ++}; ++ ++struct dio_submit { ++ struct bio *bio; ++ unsigned int blkbits; ++ unsigned int blkfactor; ++ unsigned int start_zero_done; ++ int pages_in_io; ++ sector_t block_in_file; ++ unsigned int blocks_available; ++ int reap_counter; ++ sector_t final_block_in_request; ++ int boundary; ++ get_block_t *get_block; ++ loff_t logical_offset_in_bio; ++ sector_t final_block_in_bio; ++ sector_t next_block_for_io; ++ struct page *cur_page; ++ unsigned int cur_page_offset; ++ unsigned int cur_page_len; ++ sector_t cur_page_block; ++ loff_t cur_page_fs_offset; ++ struct iov_iter *iter; ++ unsigned int head; ++ unsigned int tail; ++ size_t from; ++ size_t to; ++}; ++ ++struct dio { ++ int flags; ++ blk_opf_t opf; ++ struct gendisk *bio_disk; ++ struct inode *inode; ++ loff_t i_size; ++ dio_iodone_t *end_io; ++ void *private; ++ spinlock_t bio_lock; ++ int page_errors; ++ int is_async; ++ bool defer_completion; ++ bool should_dirty; ++ int io_error; ++ long unsigned int refcount; ++ struct bio *bio_list; ++ struct task_struct *waiter; ++ struct kiocb *iocb; ++ ssize_t result; ++ union { ++ struct page *pages[64]; ++ struct work_struct complete_work; ++ }; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct file_handle { ++ __u32 handle_bytes; ++ int handle_type; ++ unsigned char f_handle[0]; ++}; ++ ++struct dnotify_struct { ++ struct dnotify_struct *dn_next; ++ __u32 dn_mask; ++ int dn_fd; ++ struct file *dn_filp; ++ fl_owner_t dn_owner; ++}; ++ ++struct dnotify_mark { ++ struct fsnotify_mark fsn_mark; ++ struct dnotify_struct *dn; ++}; ++ ++struct inotify_event { ++ __s32 wd; ++ __u32 mask; ++ __u32 cookie; ++ __u32 len; ++ char name[0]; ++}; ++ ++struct fanotify_event_metadata { ++ __u32 event_len; ++ __u8 vers; ++ __u8 reserved; ++ __u16 metadata_len; ++ __u64 mask; ++ __s32 fd; ++ __s32 pid; ++}; ++ ++struct fanotify_event_info_header { ++ __u8 info_type; ++ __u8 pad; ++ __u16 len; ++}; ++ ++struct fanotify_event_info_fid { ++ struct fanotify_event_info_header hdr; ++ __kernel_fsid_t fsid; ++ unsigned char handle[0]; ++}; ++ ++struct fanotify_event_info_pidfd { ++ struct fanotify_event_info_header hdr; ++ __s32 pidfd; ++}; ++ ++struct fanotify_event_info_error { ++ struct fanotify_event_info_header hdr; ++ __s32 error; ++ __u32 error_count; ++}; ++ ++struct fanotify_response { ++ __s32 fd; ++ __u32 response; ++}; ++ ++enum siginfo_layout { ++ SIL_KILL = 0, ++ SIL_TIMER = 1, ++ SIL_POLL = 2, ++ SIL_FAULT = 3, ++ SIL_FAULT_TRAPNO = 4, ++ SIL_FAULT_MCEERR = 5, ++ SIL_FAULT_BNDERR = 6, ++ SIL_FAULT_PKUERR = 7, ++ SIL_FAULT_PERF_EVENT = 8, ++ SIL_CHLD = 9, ++ SIL_RT = 10, ++ SIL_SYS = 11, ++}; ++ ++struct signalfd_siginfo { ++ __u32 ssi_signo; ++ __s32 ssi_errno; ++ __s32 ssi_code; ++ __u32 ssi_pid; ++ __u32 ssi_uid; ++ __s32 ssi_fd; ++ __u32 ssi_tid; ++ __u32 ssi_band; ++ __u32 ssi_overrun; ++ __u32 ssi_trapno; ++ __s32 ssi_status; ++ __s32 ssi_int; ++ __u64 ssi_ptr; ++ __u64 ssi_utime; ++ __u64 ssi_stime; ++ __u64 ssi_addr; ++ __u16 ssi_addr_lsb; ++ __u16 __pad2; ++ __s32 ssi_syscall; ++ __u64 ssi_call_addr; ++ __u32 ssi_arch; ++ __u8 __pad[28]; ++}; ++ ++struct signalfd_ctx { ++ sigset_t sigmask; ++}; ++ ++struct eventfd_ctx { ++ struct kref kref; ++ wait_queue_head_t wqh; ++ __u64 count; ++ unsigned int flags; ++ int id; ++}; ++ ++struct userfaultfd_ctx { ++ wait_queue_head_t fault_pending_wqh; ++ wait_queue_head_t fault_wqh; ++ wait_queue_head_t fd_wqh; ++ wait_queue_head_t event_wqh; ++ seqcount_spinlock_t refile_seq; ++ refcount_t refcount; ++ unsigned int flags; ++ unsigned int features; ++ bool released; ++ atomic_t mmap_changing; ++ struct mm_struct *mm; ++}; ++ ++struct uffd_msg { ++ __u8 event; ++ __u8 reserved1; ++ __u16 reserved2; ++ __u32 reserved3; ++ union { ++ struct { ++ __u64 flags; ++ __u64 address; ++ union { ++ __u32 ptid; ++ } feat; ++ } pagefault; ++ struct { ++ __u32 ufd; ++ } fork; ++ struct { ++ __u64 from; ++ __u64 to; ++ __u64 len; ++ } remap; ++ struct { ++ __u64 start; ++ __u64 end; ++ } remove; ++ struct { ++ __u64 reserved1; ++ __u64 reserved2; ++ __u64 reserved3; ++ } reserved; ++ } arg; ++}; ++ ++struct uffdio_api { ++ __u64 api; ++ __u64 features; ++ __u64 ioctls; ++}; ++ ++struct uffdio_range { ++ __u64 start; ++ __u64 len; ++}; ++ ++struct uffdio_register { ++ struct uffdio_range range; ++ __u64 mode; ++ __u64 ioctls; ++}; ++ ++struct uffdio_copy { ++ __u64 dst; ++ __u64 src; ++ __u64 len; ++ __u64 mode; ++ __s64 copy; ++}; ++ ++struct uffdio_zeropage { ++ struct uffdio_range range; ++ __u64 mode; ++ __s64 zeropage; ++}; ++ ++struct uffdio_writeprotect { ++ struct uffdio_range range; ++ __u64 mode; ++}; ++ ++struct uffdio_continue { ++ struct uffdio_range range; ++ __u64 mode; ++ __s64 mapped; ++}; ++ ++struct userfaultfd_fork_ctx { ++ struct userfaultfd_ctx *orig; ++ struct userfaultfd_ctx *new; ++ struct list_head list; ++}; ++ ++struct userfaultfd_unmap_ctx { ++ struct userfaultfd_ctx *ctx; ++ long unsigned int start; ++ long unsigned int end; ++ struct list_head list; ++}; ++ ++struct userfaultfd_wait_queue { ++ struct uffd_msg msg; ++ wait_queue_entry_t wq; ++ struct userfaultfd_ctx *ctx; ++ bool waken; ++}; ++ ++struct userfaultfd_wake_range { ++ long unsigned int start; ++ long unsigned int len; ++}; ++ ++union fscrypt_iv { ++ struct { ++ __le64 lblk_num; ++ u8 nonce[16]; ++ }; ++ u8 raw[32]; ++ __le64 dun[4]; ++}; ++ ++struct fscrypt_nokey_name { ++ u32 dirhash[2]; ++ u8 bytes[149]; ++ u8 sha256[32]; ++}; ++ ++struct fscrypt_symlink_data { ++ __le16 len; ++ char encrypted_path[1]; ++} __attribute__((packed)); ++ ++struct user_key_payload { ++ struct callback_head rcu; ++ short unsigned int datalen; ++ long: 0; ++ char data[0]; ++}; ++ ++struct fscrypt_key { ++ __u32 mode; ++ __u8 raw[64]; ++ __u32 size; ++}; ++ ++struct fscrypt_direct_key { ++ struct super_block *dk_sb; ++ struct hlist_node dk_node; ++ refcount_t dk_refcount; ++ const struct fscrypt_mode *dk_mode; ++ struct fscrypt_prepared_key dk_key; ++ u8 dk_descriptor[8]; ++ u8 dk_raw[64]; ++}; ++ ++struct fscrypt_get_policy_ex_arg { ++ __u64 policy_size; ++ union { ++ __u8 version; ++ struct fscrypt_policy_v1 v1; ++ struct fscrypt_policy_v2 v2; ++ } policy; ++}; ++ ++struct fscrypt_context_v1 { ++ u8 version; ++ u8 contents_encryption_mode; ++ u8 filenames_encryption_mode; ++ u8 flags; ++ u8 master_key_descriptor[8]; ++ u8 nonce[16]; ++}; ++ ++struct fscrypt_context_v2 { ++ u8 version; ++ u8 contents_encryption_mode; ++ u8 filenames_encryption_mode; ++ u8 flags; ++ u8 __reserved[4]; ++ u8 master_key_identifier[16]; ++ u8 nonce[16]; ++}; ++ ++union fscrypt_context { ++ u8 version; ++ struct fscrypt_context_v1 v1; ++ struct fscrypt_context_v2 v2; ++}; ++ ++struct fsverity_enable_arg { ++ __u32 version; ++ __u32 hash_algorithm; ++ __u32 block_size; ++ __u32 salt_size; ++ __u64 salt_ptr; ++ __u32 sig_size; ++ __u32 __reserved1; ++ __u64 sig_ptr; ++ __u64 __reserved2[11]; ++}; ++ ++struct block_buffer { ++ u32 filled; ++ bool is_root_hash; ++ u8 *data; ++}; ++ ++struct fsverity_digest { ++ __u16 digest_algorithm; ++ __u16 digest_size; ++ __u8 digest[0]; ++}; ++ ++typedef struct { ++ u32 mask; ++ u32 addr; ++} psw_compat_t; ++ ++typedef struct { ++ psw_compat_t psw; ++ u32 gprs[16]; ++ u32 acrs[16]; ++ u32 orig_gpr2; ++ long: 0; ++} s390_compat_regs; ++ ++typedef s390_compat_regs compat_elf_gregset_t; ++ ++struct arch_elf_state { ++ int rc; ++}; ++ ++typedef __u32 Elf32_Addr; ++ ++typedef __u16 Elf32_Half; ++ ++typedef __u32 Elf32_Off; ++ ++typedef __u32 Elf32_Word; ++ ++struct elf32_hdr { ++ unsigned char e_ident[16]; ++ Elf32_Half e_type; ++ Elf32_Half e_machine; ++ Elf32_Word e_version; ++ Elf32_Addr e_entry; ++ Elf32_Off e_phoff; ++ Elf32_Off e_shoff; ++ Elf32_Word e_flags; ++ Elf32_Half e_ehsize; ++ Elf32_Half e_phentsize; ++ Elf32_Half e_phnum; ++ Elf32_Half e_shentsize; ++ Elf32_Half e_shnum; ++ Elf32_Half e_shstrndx; ++}; ++ ++struct elf32_phdr { ++ Elf32_Word p_type; ++ Elf32_Off p_offset; ++ Elf32_Addr p_vaddr; ++ Elf32_Addr p_paddr; ++ Elf32_Word p_filesz; ++ Elf32_Word p_memsz; ++ Elf32_Word p_flags; ++ Elf32_Word p_align; ++}; ++ ++struct elf32_shdr { ++ Elf32_Word sh_name; ++ Elf32_Word sh_type; ++ Elf32_Word sh_flags; ++ Elf32_Addr sh_addr; ++ Elf32_Off sh_offset; ++ Elf32_Word sh_size; ++ Elf32_Word sh_link; ++ Elf32_Word sh_info; ++ Elf32_Word sh_addralign; ++ Elf32_Word sh_entsize; ++}; ++ ++struct elf32_note { ++ Elf32_Word n_namesz; ++ Elf32_Word n_descsz; ++ Elf32_Word n_type; ++}; ++ ++struct compat_elf_siginfo { ++ compat_int_t si_signo; ++ compat_int_t si_code; ++ compat_int_t si_errno; ++}; ++ ++struct compat_elf_prstatus_common { ++ struct compat_elf_siginfo pr_info; ++ short int pr_cursig; ++ compat_ulong_t pr_sigpend; ++ compat_ulong_t pr_sighold; ++ compat_pid_t pr_pid; ++ compat_pid_t pr_ppid; ++ compat_pid_t pr_pgrp; ++ compat_pid_t pr_sid; ++ struct old_timeval32 pr_utime; ++ struct old_timeval32 pr_stime; ++ struct old_timeval32 pr_cutime; ++ struct old_timeval32 pr_cstime; ++}; ++ ++struct compat_elf_prpsinfo { ++ char pr_state; ++ char pr_sname; ++ char pr_zomb; ++ char pr_nice; ++ compat_ulong_t pr_flag; ++ __compat_uid_t pr_uid; ++ __compat_gid_t pr_gid; ++ compat_pid_t pr_pid; ++ compat_pid_t pr_ppid; ++ compat_pid_t pr_pgrp; ++ compat_pid_t pr_sid; ++ char pr_fname[16]; ++ char pr_psargs[80]; ++}; ++ ++struct compat_elf_prstatus { ++ struct compat_elf_prstatus_common common; ++ compat_elf_gregset_t pr_reg; ++ compat_int_t pr_fpvalid; ++ long: 0; ++}; ++ ++struct memelfnote { ++ const char *name; ++ int type; ++ unsigned int datasz; ++ void *data; ++}; ++ ++struct elf_thread_core_info { ++ struct elf_thread_core_info *next; ++ struct task_struct *task; ++ struct compat_elf_prstatus prstatus; ++ struct memelfnote notes[0]; ++}; ++ ++struct elf_note_info { ++ struct elf_thread_core_info *thread; ++ struct memelfnote psinfo; ++ struct memelfnote signote; ++ struct memelfnote auxv; ++ struct memelfnote files; ++ compat_siginfo_t csigdata; ++ size_t size; ++ int thread_notes; ++}; ++ ++struct mb_cache { ++ struct hlist_bl_head *c_hash; ++ int c_bucket_bits; ++ long unsigned int c_max_entries; ++ spinlock_t c_list_lock; ++ struct list_head c_list; ++ long unsigned int c_entry_count; ++ struct shrinker c_shrink; ++ struct work_struct c_shrink_work; ++}; ++ ++struct xdr_netobj { ++ unsigned int len; ++ u8 *data; ++}; ++ ++struct xdr_buf { ++ struct kvec head[1]; ++ struct kvec tail[1]; ++ struct bio_vec *bvec; ++ struct page **pages; ++ unsigned int page_base; ++ unsigned int page_len; ++ unsigned int flags; ++ unsigned int buflen; ++ unsigned int len; ++}; ++ ++struct rpc_rqst; ++ ++struct xdr_stream { ++ __be32 *p; ++ struct xdr_buf *buf; ++ __be32 *end; ++ struct kvec *iov; ++ struct kvec scratch; ++ struct page **page_ptr; ++ unsigned int nwords; ++ struct rpc_rqst *rqst; ++}; ++ ++struct rpc_xprt; ++ ++struct rpc_task; ++ ++struct rpc_cred; ++ ++struct rpc_rqst { ++ struct rpc_xprt *rq_xprt; ++ struct xdr_buf rq_snd_buf; ++ struct xdr_buf rq_rcv_buf; ++ struct rpc_task *rq_task; ++ struct rpc_cred *rq_cred; ++ __be32 rq_xid; ++ int rq_cong; ++ u32 rq_seqno; ++ int rq_enc_pages_num; ++ struct page **rq_enc_pages; ++ void (*rq_release_snd_buf)(struct rpc_rqst *); ++ union { ++ struct list_head rq_list; ++ struct rb_node rq_recv; ++ }; ++ struct list_head rq_xmit; ++ struct list_head rq_xmit2; ++ void *rq_buffer; ++ size_t rq_callsize; ++ void *rq_rbuffer; ++ size_t rq_rcvsize; ++ size_t rq_xmit_bytes_sent; ++ size_t rq_reply_bytes_recvd; ++ struct xdr_buf rq_private_buf; ++ long unsigned int rq_majortimeo; ++ long unsigned int rq_minortimeo; ++ long unsigned int rq_timeout; ++ ktime_t rq_rtt; ++ unsigned int rq_retries; ++ unsigned int rq_connect_cookie; ++ atomic_t rq_pin; ++ u32 rq_bytes_sent; ++ ktime_t rq_xtime; ++ int rq_ntrans; ++ struct list_head rq_bc_list; ++ long unsigned int rq_bc_pa_state; ++ struct list_head rq_bc_pa_list; ++}; ++ ++typedef void (*kxdreproc_t)(struct rpc_rqst *, struct xdr_stream *, const void *); ++ ++typedef int (*kxdrdproc_t)(struct rpc_rqst *, struct xdr_stream *, void *); ++ ++struct rpc_procinfo; ++ ++struct rpc_message { ++ const struct rpc_procinfo *rpc_proc; ++ void *rpc_argp; ++ void *rpc_resp; ++ const struct cred *rpc_cred; ++}; ++ ++struct rpc_procinfo { ++ u32 p_proc; ++ kxdreproc_t p_encode; ++ kxdrdproc_t p_decode; ++ unsigned int p_arglen; ++ unsigned int p_replen; ++ unsigned int p_timer; ++ u32 p_statidx; ++ const char *p_name; ++}; ++ ++struct rpc_wait { ++ struct list_head list; ++ struct list_head links; ++ struct list_head timer_list; ++}; ++ ++struct rpc_wait_queue; ++ ++struct rpc_call_ops; ++ ++struct rpc_clnt; ++ ++struct rpc_task { ++ atomic_t tk_count; ++ int tk_status; ++ struct list_head tk_task; ++ void (*tk_callback)(struct rpc_task *); ++ void (*tk_action)(struct rpc_task *); ++ long unsigned int tk_timeout; ++ long unsigned int tk_runstate; ++ struct rpc_wait_queue *tk_waitqueue; ++ union { ++ struct work_struct tk_work; ++ struct rpc_wait tk_wait; ++ } u; ++ struct rpc_message tk_msg; ++ void *tk_calldata; ++ const struct rpc_call_ops *tk_ops; ++ struct rpc_clnt *tk_client; ++ struct rpc_xprt *tk_xprt; ++ struct rpc_cred *tk_op_cred; ++ struct rpc_rqst *tk_rqstp; ++ struct workqueue_struct *tk_workqueue; ++ ktime_t tk_start; ++ pid_t tk_owner; ++ int tk_rpc_status; ++ short unsigned int tk_flags; ++ short unsigned int tk_timeouts; ++ short unsigned int tk_pid; ++ unsigned char tk_priority: 2; ++ unsigned char tk_garb_retry: 2; ++ unsigned char tk_cred_retry: 2; ++}; ++ ++struct rpc_timer { ++ struct list_head list; ++ long unsigned int expires; ++ struct delayed_work dwork; ++}; ++ ++struct rpc_wait_queue { ++ spinlock_t lock; ++ struct list_head tasks[4]; ++ unsigned char maxpriority; ++ unsigned char priority; ++ unsigned char nr; ++ short unsigned int qlen; ++ struct rpc_timer timer_list; ++ const char *name; ++}; ++ ++struct rpc_call_ops { ++ void (*rpc_call_prepare)(struct rpc_task *, void *); ++ void (*rpc_call_done)(struct rpc_task *, void *); ++ void (*rpc_count_stats)(struct rpc_task *, void *); ++ void (*rpc_release)(void *); ++}; ++ ++struct rpc_iostats; ++ ++struct rpc_pipe_dir_head { ++ struct list_head pdh_entries; ++ struct dentry *pdh_dentry; ++}; ++ ++struct rpc_rtt { ++ long unsigned int timeo; ++ long unsigned int srtt[5]; ++ long unsigned int sdrtt[5]; ++ int ntimeouts[5]; ++}; ++ ++struct rpc_timeout { ++ long unsigned int to_initval; ++ long unsigned int to_maxval; ++ long unsigned int to_increment; ++ unsigned int to_retries; ++ unsigned char to_exponential; ++}; ++ ++struct rpc_sysfs_client; ++ ++struct rpc_xprt_switch; ++ ++struct rpc_xprt_iter_ops; ++ ++struct rpc_xprt_iter { ++ struct rpc_xprt_switch *xpi_xpswitch; ++ struct rpc_xprt *xpi_cursor; ++ const struct rpc_xprt_iter_ops *xpi_ops; ++}; ++ ++struct rpc_auth; ++ ++struct rpc_stat; ++ ++struct rpc_program; ++ ++struct rpc_clnt { ++ refcount_t cl_count; ++ unsigned int cl_clid; ++ struct list_head cl_clients; ++ struct list_head cl_tasks; ++ atomic_t cl_pid; ++ spinlock_t cl_lock; ++ struct rpc_xprt *cl_xprt; ++ const struct rpc_procinfo *cl_procinfo; ++ u32 cl_prog; ++ u32 cl_vers; ++ u32 cl_maxproc; ++ struct rpc_auth *cl_auth; ++ struct rpc_stat *cl_stats; ++ struct rpc_iostats *cl_metrics; ++ unsigned int cl_softrtry: 1; ++ unsigned int cl_softerr: 1; ++ unsigned int cl_discrtry: 1; ++ unsigned int cl_noretranstimeo: 1; ++ unsigned int cl_autobind: 1; ++ unsigned int cl_chatty: 1; ++ struct rpc_rtt *cl_rtt; ++ const struct rpc_timeout *cl_timeout; ++ atomic_t cl_swapper; ++ int cl_nodelen; ++ char cl_nodename[65]; ++ struct rpc_pipe_dir_head cl_pipedir_objects; ++ struct rpc_clnt *cl_parent; ++ struct rpc_rtt cl_rtt_default; ++ struct rpc_timeout cl_timeout_default; ++ const struct rpc_program *cl_program; ++ const char *cl_principal; ++ struct dentry *cl_debugfs; ++ struct rpc_sysfs_client *cl_sysfs; ++ union { ++ struct rpc_xprt_iter cl_xpi; ++ struct work_struct cl_work; ++ }; ++ const struct cred *cl_cred; ++ unsigned int cl_max_connect; ++}; ++ ++struct svc_xprt; ++ ++struct rpc_sysfs_xprt; ++ ++struct rpc_xprt_ops; ++ ++struct svc_serv; ++ ++struct xprt_class; ++ ++struct rpc_xprt { ++ struct kref kref; ++ const struct rpc_xprt_ops *ops; ++ unsigned int id; ++ const struct rpc_timeout *timeout; ++ struct __kernel_sockaddr_storage addr; ++ size_t addrlen; ++ int prot; ++ long unsigned int cong; ++ long unsigned int cwnd; ++ size_t max_payload; ++ struct rpc_wait_queue binding; ++ struct rpc_wait_queue sending; ++ struct rpc_wait_queue pending; ++ struct rpc_wait_queue backlog; ++ struct list_head free; ++ unsigned int max_reqs; ++ unsigned int min_reqs; ++ unsigned int num_reqs; ++ long unsigned int state; ++ unsigned char resvport: 1; ++ unsigned char reuseport: 1; ++ atomic_t swapper; ++ unsigned int bind_index; ++ struct list_head xprt_switch; ++ long unsigned int bind_timeout; ++ long unsigned int reestablish_timeout; ++ unsigned int connect_cookie; ++ struct work_struct task_cleanup; ++ struct timer_list timer; ++ long unsigned int last_used; ++ long unsigned int idle_timeout; ++ long unsigned int connect_timeout; ++ long unsigned int max_reconnect_timeout; ++ atomic_long_t queuelen; ++ spinlock_t transport_lock; ++ spinlock_t reserve_lock; ++ spinlock_t queue_lock; ++ u32 xid; ++ struct rpc_task *snd_task; ++ struct list_head xmit_queue; ++ atomic_long_t xmit_queuelen; ++ struct svc_xprt *bc_xprt; ++ struct svc_serv *bc_serv; ++ unsigned int bc_alloc_max; ++ unsigned int bc_alloc_count; ++ atomic_t bc_slot_count; ++ spinlock_t bc_pa_lock; ++ struct list_head bc_pa_list; ++ struct rb_root recv_queue; ++ struct { ++ long unsigned int bind_count; ++ long unsigned int connect_count; ++ long unsigned int connect_start; ++ long unsigned int connect_time; ++ long unsigned int sends; ++ long unsigned int recvs; ++ long unsigned int bad_xids; ++ long unsigned int max_slots; ++ long long unsigned int req_u; ++ long long unsigned int bklog_u; ++ long long unsigned int sending_u; ++ long long unsigned int pending_u; ++ } stat; ++ struct net *xprt_net; ++ netns_tracker ns_tracker; ++ const char *servername; ++ const char *address_strings[6]; ++ struct dentry *debugfs; ++ struct callback_head rcu; ++ const struct xprt_class *xprt_class; ++ struct rpc_sysfs_xprt *xprt_sysfs; ++ bool main; ++}; ++ ++struct rpc_credops; ++ ++struct rpc_cred { ++ struct hlist_node cr_hash; ++ struct list_head cr_lru; ++ struct callback_head cr_rcu; ++ struct rpc_auth *cr_auth; ++ const struct rpc_credops *cr_ops; ++ long unsigned int cr_expire; ++ long unsigned int cr_flags; ++ refcount_t cr_count; ++ const struct cred *cr_cred; ++}; ++ ++typedef u32 rpc_authflavor_t; ++ ++struct auth_cred { ++ const struct cred *cred; ++ const char *principal; ++}; ++ ++struct rpc_cred_cache; ++ ++struct rpc_authops; ++ ++struct rpc_auth { ++ unsigned int au_cslack; ++ unsigned int au_rslack; ++ unsigned int au_verfsize; ++ unsigned int au_ralign; ++ long unsigned int au_flags; ++ const struct rpc_authops *au_ops; ++ rpc_authflavor_t au_flavor; ++ refcount_t au_count; ++ struct rpc_cred_cache *au_credcache; ++}; ++ ++struct rpc_credops { ++ const char *cr_name; ++ int (*cr_init)(struct rpc_auth *, struct rpc_cred *); ++ void (*crdestroy)(struct rpc_cred *); ++ int (*crmatch)(struct auth_cred *, struct rpc_cred *, int); ++ int (*crmarshal)(struct rpc_task *, struct xdr_stream *); ++ int (*crrefresh)(struct rpc_task *); ++ int (*crvalidate)(struct rpc_task *, struct xdr_stream *); ++ int (*crwrap_req)(struct rpc_task *, struct xdr_stream *); ++ int (*crunwrap_resp)(struct rpc_task *, struct xdr_stream *); ++ int (*crkey_timeout)(struct rpc_cred *); ++ char * (*crstringify_acceptor)(struct rpc_cred *); ++ bool (*crneed_reencode)(struct rpc_task *); ++}; ++ ++struct rpc_auth_create_args; ++ ++struct rpcsec_gss_info; ++ ++struct rpc_authops { ++ struct module *owner; ++ rpc_authflavor_t au_flavor; ++ char *au_name; ++ struct rpc_auth * (*create)(const struct rpc_auth_create_args *, struct rpc_clnt *); ++ void (*destroy)(struct rpc_auth *); ++ int (*hash_cred)(struct auth_cred *, unsigned int); ++ struct rpc_cred * (*lookup_cred)(struct rpc_auth *, struct auth_cred *, int); ++ struct rpc_cred * (*crcreate)(struct rpc_auth *, struct auth_cred *, int, gfp_t); ++ rpc_authflavor_t (*info2flavor)(struct rpcsec_gss_info *); ++ int (*flavor2info)(rpc_authflavor_t, struct rpcsec_gss_info *); ++ int (*key_timeout)(struct rpc_auth *, struct rpc_cred *); ++}; ++ ++struct rpc_auth_create_args { ++ rpc_authflavor_t pseudoflavor; ++ const char *target_name; ++}; ++ ++struct rpcsec_gss_oid { ++ unsigned int len; ++ u8 data[32]; ++}; ++ ++struct rpcsec_gss_info { ++ struct rpcsec_gss_oid oid; ++ u32 qop; ++ u32 service; ++}; ++ ++struct rpc_xprt_ops { ++ void (*set_buffer_size)(struct rpc_xprt *, size_t, size_t); ++ int (*reserve_xprt)(struct rpc_xprt *, struct rpc_task *); ++ void (*release_xprt)(struct rpc_xprt *, struct rpc_task *); ++ void (*alloc_slot)(struct rpc_xprt *, struct rpc_task *); ++ void (*free_slot)(struct rpc_xprt *, struct rpc_rqst *); ++ void (*rpcbind)(struct rpc_task *); ++ void (*set_port)(struct rpc_xprt *, short unsigned int); ++ void (*connect)(struct rpc_xprt *, struct rpc_task *); ++ int (*get_srcaddr)(struct rpc_xprt *, char *, size_t); ++ short unsigned int (*get_srcport)(struct rpc_xprt *); ++ int (*buf_alloc)(struct rpc_task *); ++ void (*buf_free)(struct rpc_task *); ++ int (*prepare_request)(struct rpc_rqst *, struct xdr_buf *); ++ int (*send_request)(struct rpc_rqst *); ++ void (*wait_for_reply_request)(struct rpc_task *); ++ void (*timer)(struct rpc_xprt *, struct rpc_task *); ++ void (*release_request)(struct rpc_task *); ++ void (*close)(struct rpc_xprt *); ++ void (*destroy)(struct rpc_xprt *); ++ void (*set_connect_timeout)(struct rpc_xprt *, long unsigned int, long unsigned int); ++ void (*print_stats)(struct rpc_xprt *, struct seq_file *); ++ int (*enable_swap)(struct rpc_xprt *); ++ void (*disable_swap)(struct rpc_xprt *); ++ void (*inject_disconnect)(struct rpc_xprt *); ++ int (*bc_setup)(struct rpc_xprt *, unsigned int); ++ size_t (*bc_maxpayload)(struct rpc_xprt *); ++ unsigned int (*bc_num_slots)(struct rpc_xprt *); ++ void (*bc_free_rqst)(struct rpc_rqst *); ++ void (*bc_destroy)(struct rpc_xprt *, unsigned int); ++}; ++ ++struct svc_program; ++ ++struct svc_stat; ++ ++struct svc_pool; ++ ++struct svc_serv { ++ struct svc_program *sv_program; ++ struct svc_stat *sv_stats; ++ spinlock_t sv_lock; ++ struct kref sv_refcnt; ++ unsigned int sv_nrthreads; ++ unsigned int sv_maxconn; ++ unsigned int sv_max_payload; ++ unsigned int sv_max_mesg; ++ unsigned int sv_xdrsize; ++ struct list_head sv_permsocks; ++ struct list_head sv_tempsocks; ++ int sv_tmpcnt; ++ struct timer_list sv_temptimer; ++ char *sv_name; ++ unsigned int sv_nrpools; ++ struct svc_pool *sv_pools; ++ int (*sv_threadfn)(void *); ++ struct list_head sv_cb_list; ++ spinlock_t sv_cb_lock; ++ wait_queue_head_t sv_cb_waitq; ++ bool sv_bc_enabled; ++}; ++ ++struct xprt_create; ++ ++struct xprt_class { ++ struct list_head list; ++ int ident; ++ struct rpc_xprt * (*setup)(struct xprt_create *); ++ struct module *owner; ++ char name[32]; ++ const char *netid[0]; ++}; ++ ++struct xprt_create { ++ int ident; ++ struct net *net; ++ struct sockaddr *srcaddr; ++ struct sockaddr *dstaddr; ++ size_t addrlen; ++ const char *servername; ++ struct svc_xprt *bc_xprt; ++ struct rpc_xprt_switch *bc_xps; ++ unsigned int flags; ++}; ++ ++struct rpc_sysfs_xprt_switch; ++ ++struct rpc_xprt_switch { ++ spinlock_t xps_lock; ++ struct kref xps_kref; ++ unsigned int xps_id; ++ unsigned int xps_nxprts; ++ unsigned int xps_nactive; ++ unsigned int xps_nunique_destaddr_xprts; ++ atomic_long_t xps_queuelen; ++ struct list_head xps_xprt_list; ++ struct net *xps_net; ++ const struct rpc_xprt_iter_ops *xps_iter_ops; ++ struct rpc_sysfs_xprt_switch *xps_sysfs; ++ struct callback_head xps_rcu; ++}; ++ ++struct rpc_stat { ++ const struct rpc_program *program; ++ unsigned int netcnt; ++ unsigned int netudpcnt; ++ unsigned int nettcpcnt; ++ unsigned int nettcpconn; ++ unsigned int netreconn; ++ unsigned int rpccnt; ++ unsigned int rpcretrans; ++ unsigned int rpcauthrefresh; ++ unsigned int rpcgarbage; ++}; ++ ++struct rpc_version; ++ ++struct rpc_program { ++ const char *name; ++ u32 number; ++ unsigned int nrvers; ++ const struct rpc_version **version; ++ struct rpc_stat *stats; ++ const char *pipe_dir_name; ++}; ++ ++struct svc_stat { ++ struct svc_program *program; ++ unsigned int netcnt; ++ unsigned int netudpcnt; ++ unsigned int nettcpcnt; ++ unsigned int nettcpconn; ++ unsigned int rpccnt; ++ unsigned int rpcbadfmt; ++ unsigned int rpcbadauth; ++ unsigned int rpcbadclnt; ++}; ++ ++struct svc_version; ++ ++struct svc_rqst; ++ ++struct svc_process_info; ++ ++struct svc_program { ++ struct svc_program *pg_next; ++ u32 pg_prog; ++ unsigned int pg_lovers; ++ unsigned int pg_hivers; ++ unsigned int pg_nvers; ++ const struct svc_version **pg_vers; ++ char *pg_name; ++ char *pg_class; ++ struct svc_stat *pg_stats; ++ int (*pg_authenticate)(struct svc_rqst *); ++ __be32 (*pg_init_request)(struct svc_rqst *, const struct svc_program *, struct svc_process_info *); ++ int (*pg_rpcbind_set)(struct net *, const struct svc_program *, u32, int, short unsigned int, short unsigned int); ++}; ++ ++struct rpc_xprt_iter_ops { ++ void (*xpi_rewind)(struct rpc_xprt_iter *); ++ struct rpc_xprt * (*xpi_xprt)(struct rpc_xprt_iter *); ++ struct rpc_xprt * (*xpi_next)(struct rpc_xprt_iter *); ++}; ++ ++struct rpc_version { ++ u32 number; ++ unsigned int nrprocs; ++ const struct rpc_procinfo *procs; ++ unsigned int *counts; ++}; ++ ++struct nfs_fh { ++ short unsigned int size; ++ unsigned char data[128]; ++}; ++ ++enum nfs3_stable_how { ++ NFS_UNSTABLE = 0, ++ NFS_DATA_SYNC = 1, ++ NFS_FILE_SYNC = 2, ++ NFS_INVALID_STABLE_HOW = -1, ++}; ++ ++struct nfs4_label { ++ uint32_t lfs; ++ uint32_t pi; ++ u32 len; ++ char *label; ++}; ++ ++typedef struct { ++ char data[8]; ++} nfs4_verifier; ++ ++struct nfs4_stateid_struct { ++ union { ++ char data[16]; ++ struct { ++ __be32 seqid; ++ char other[12]; ++ }; ++ }; ++ enum { ++ NFS4_INVALID_STATEID_TYPE = 0, ++ NFS4_SPECIAL_STATEID_TYPE = 1, ++ NFS4_OPEN_STATEID_TYPE = 2, ++ NFS4_LOCK_STATEID_TYPE = 3, ++ NFS4_DELEGATION_STATEID_TYPE = 4, ++ NFS4_LAYOUT_STATEID_TYPE = 5, ++ NFS4_PNFS_DS_STATEID_TYPE = 6, ++ NFS4_REVOKED_STATEID_TYPE = 7, ++ } type; ++}; ++ ++typedef struct nfs4_stateid_struct nfs4_stateid; ++ ++enum nfs4_change_attr_type { ++ NFS4_CHANGE_TYPE_IS_MONOTONIC_INCR = 0, ++ NFS4_CHANGE_TYPE_IS_VERSION_COUNTER = 1, ++ NFS4_CHANGE_TYPE_IS_VERSION_COUNTER_NOPNFS = 2, ++ NFS4_CHANGE_TYPE_IS_TIME_METADATA = 3, ++ NFS4_CHANGE_TYPE_IS_UNDEFINED = 4, ++}; ++ ++struct gss_api_mech; ++ ++struct gss_ctx { ++ struct gss_api_mech *mech_type; ++ void *internal_ctx_id; ++ unsigned int slack; ++ unsigned int align; ++}; ++ ++struct gss_api_ops; ++ ++struct pf_desc; ++ ++struct gss_api_mech { ++ struct list_head gm_list; ++ struct module *gm_owner; ++ struct rpcsec_gss_oid gm_oid; ++ char *gm_name; ++ const struct gss_api_ops *gm_ops; ++ int gm_pf_num; ++ struct pf_desc *gm_pfs; ++ const char *gm_upcall_enctypes; ++}; ++ ++struct auth_domain; ++ ++struct pf_desc { ++ u32 pseudoflavor; ++ u32 qop; ++ u32 service; ++ char *name; ++ char *auth_domain_name; ++ struct auth_domain *domain; ++ bool datatouch; ++}; ++ ++struct auth_ops; ++ ++struct auth_domain { ++ struct kref ref; ++ struct hlist_node hash; ++ char *name; ++ struct auth_ops *flavour; ++ struct callback_head callback_head; ++}; ++ ++struct gss_api_ops { ++ int (*gss_import_sec_context)(const void *, size_t, struct gss_ctx *, time64_t *, gfp_t); ++ u32 (*gss_get_mic)(struct gss_ctx *, struct xdr_buf *, struct xdr_netobj *); ++ u32 (*gss_verify_mic)(struct gss_ctx *, struct xdr_buf *, struct xdr_netobj *); ++ u32 (*gss_wrap)(struct gss_ctx *, int, struct xdr_buf *, struct page **); ++ u32 (*gss_unwrap)(struct gss_ctx *, int, int, struct xdr_buf *); ++ void (*gss_delete_sec_context)(void *); ++}; ++ ++struct nfs4_string { ++ unsigned int len; ++ char *data; ++}; ++ ++struct nfs_fsid { ++ uint64_t major; ++ uint64_t minor; ++}; ++ ++struct nfs4_threshold { ++ __u32 bm; ++ __u32 l_type; ++ __u64 rd_sz; ++ __u64 wr_sz; ++ __u64 rd_io_sz; ++ __u64 wr_io_sz; ++}; ++ ++struct nfs_fattr { ++ unsigned int valid; ++ umode_t mode; ++ __u32 nlink; ++ kuid_t uid; ++ kgid_t gid; ++ dev_t rdev; ++ __u64 size; ++ union { ++ struct { ++ __u32 blocksize; ++ __u32 blocks; ++ } nfs2; ++ struct { ++ __u64 used; ++ } nfs3; ++ } du; ++ struct nfs_fsid fsid; ++ __u64 fileid; ++ __u64 mounted_on_fileid; ++ struct timespec64 atime; ++ struct timespec64 mtime; ++ struct timespec64 ctime; ++ __u64 change_attr; ++ __u64 pre_change_attr; ++ __u64 pre_size; ++ struct timespec64 pre_mtime; ++ struct timespec64 pre_ctime; ++ long unsigned int time_start; ++ long unsigned int gencount; ++ struct nfs4_string *owner_name; ++ struct nfs4_string *group_name; ++ struct nfs4_threshold *mdsthreshold; ++ struct nfs4_label *label; ++}; ++ ++struct nfs_fsinfo { ++ struct nfs_fattr *fattr; ++ __u32 rtmax; ++ __u32 rtpref; ++ __u32 rtmult; ++ __u32 wtmax; ++ __u32 wtpref; ++ __u32 wtmult; ++ __u32 dtpref; ++ __u64 maxfilesize; ++ struct timespec64 time_delta; ++ __u32 lease_time; ++ __u32 nlayouttypes; ++ __u32 layouttype[8]; ++ __u32 blksize; ++ __u32 clone_blksize; ++ enum nfs4_change_attr_type change_attr_type; ++ __u32 xattr_support; ++}; ++ ++struct nfs_fsstat { ++ struct nfs_fattr *fattr; ++ __u64 tbytes; ++ __u64 fbytes; ++ __u64 abytes; ++ __u64 tfiles; ++ __u64 ffiles; ++ __u64 afiles; ++}; ++ ++struct nfs_pathconf { ++ struct nfs_fattr *fattr; ++ __u32 max_link; ++ __u32 max_namelen; ++}; ++ ++struct nfs4_change_info { ++ u32 atomic; ++ u64 before; ++ u64 after; ++}; ++ ++struct nfs4_slot; ++ ++struct nfs4_sequence_args { ++ struct nfs4_slot *sa_slot; ++ u8 sa_cache_this: 1; ++ u8 sa_privileged: 1; ++}; ++ ++struct nfs4_sequence_res { ++ struct nfs4_slot *sr_slot; ++ long unsigned int sr_timestamp; ++ int sr_status; ++ u32 sr_status_flags; ++ u32 sr_highest_slotid; ++ u32 sr_target_highest_slotid; ++}; ++ ++struct nfs_open_context; ++ ++struct nfs_lock_context { ++ refcount_t count; ++ struct list_head list; ++ struct nfs_open_context *open_context; ++ fl_owner_t lockowner; ++ atomic_t io_count; ++ struct callback_head callback_head; ++}; ++ ++struct nfs4_state; ++ ++struct nfs_open_context { ++ struct nfs_lock_context lock_context; ++ fl_owner_t flock_owner; ++ struct dentry *dentry; ++ const struct cred *cred; ++ struct rpc_cred *ll_cred; ++ struct nfs4_state *state; ++ fmode_t mode; ++ long unsigned int flags; ++ int error; ++ struct list_head list; ++ struct nfs4_threshold *mdsthreshold; ++ struct callback_head callback_head; ++}; ++ ++struct nlm_host; ++ ++struct nfs_iostats; ++ ++struct nfs_auth_info { ++ unsigned int flavor_len; ++ rpc_authflavor_t flavors[12]; ++}; ++ ++struct fscache_volume; ++ ++struct pnfs_layoutdriver_type; ++ ++struct nfs_client; ++ ++struct nfs_server { ++ struct nfs_client *nfs_client; ++ struct list_head client_link; ++ struct list_head master_link; ++ struct rpc_clnt *client; ++ struct rpc_clnt *client_acl; ++ struct nlm_host *nlm_host; ++ struct nfs_iostats *io_stats; ++ atomic_long_t writeback; ++ unsigned int write_congested; ++ unsigned int flags; ++ unsigned int fattr_valid; ++ unsigned int caps; ++ unsigned int rsize; ++ unsigned int rpages; ++ unsigned int wsize; ++ unsigned int wpages; ++ unsigned int wtmult; ++ unsigned int dtsize; ++ short unsigned int port; ++ unsigned int bsize; ++ unsigned int gxasize; ++ unsigned int sxasize; ++ unsigned int lxasize; ++ unsigned int acregmin; ++ unsigned int acregmax; ++ unsigned int acdirmin; ++ unsigned int acdirmax; ++ unsigned int namelen; ++ unsigned int options; ++ unsigned int clone_blksize; ++ enum nfs4_change_attr_type change_attr_type; ++ struct nfs_fsid fsid; ++ __u64 maxfilesize; ++ struct timespec64 time_delta; ++ long unsigned int mount_time; ++ struct super_block *super; ++ dev_t s_dev; ++ struct nfs_auth_info auth_info; ++ struct fscache_volume *fscache; ++ char *fscache_uniq; ++ u32 pnfs_blksize; ++ u32 attr_bitmask[3]; ++ u32 attr_bitmask_nl[3]; ++ u32 exclcreat_bitmask[3]; ++ u32 cache_consistency_bitmask[3]; ++ u32 acl_bitmask; ++ u32 fh_expire_type; ++ struct pnfs_layoutdriver_type *pnfs_curr_ld; ++ struct rpc_wait_queue roc_rpcwaitq; ++ void *pnfs_ld_data; ++ struct rb_root state_owners; ++ struct ida openowner_id; ++ struct ida lockowner_id; ++ struct list_head state_owners_lru; ++ struct list_head layouts; ++ struct list_head delegations; ++ struct list_head ss_copies; ++ long unsigned int mig_gen; ++ long unsigned int mig_status; ++ void (*destroy)(struct nfs_server *); ++ atomic_t active; ++ struct __kernel_sockaddr_storage mountd_address; ++ size_t mountd_addrlen; ++ u32 mountd_version; ++ short unsigned int mountd_port; ++ short unsigned int mountd_protocol; ++ struct rpc_wait_queue uoc_rpcwaitq; ++ unsigned int read_hdrsize; ++ const struct cred *cred; ++ bool has_sec_mnt_opts; ++}; ++ ++struct nfs_subversion; ++ ++struct idmap; ++ ++struct nfs4_slot_table; ++ ++struct nfs4_session; ++ ++struct nfs_rpc_ops; ++ ++struct nfs4_minor_version_ops; ++ ++struct nfs41_server_owner; ++ ++struct nfs41_server_scope; ++ ++struct nfs41_impl_id; ++ ++struct nfs_client { ++ refcount_t cl_count; ++ atomic_t cl_mds_count; ++ int cl_cons_state; ++ long unsigned int cl_res_state; ++ long unsigned int cl_flags; ++ struct __kernel_sockaddr_storage cl_addr; ++ size_t cl_addrlen; ++ char *cl_hostname; ++ char *cl_acceptor; ++ struct list_head cl_share_link; ++ struct list_head cl_superblocks; ++ struct rpc_clnt *cl_rpcclient; ++ const struct nfs_rpc_ops *rpc_ops; ++ int cl_proto; ++ struct nfs_subversion *cl_nfs_mod; ++ u32 cl_minorversion; ++ unsigned int cl_nconnect; ++ unsigned int cl_max_connect; ++ const char *cl_principal; ++ struct list_head cl_ds_clients; ++ u64 cl_clientid; ++ nfs4_verifier cl_confirm; ++ long unsigned int cl_state; ++ spinlock_t cl_lock; ++ long unsigned int cl_lease_time; ++ long unsigned int cl_last_renewal; ++ struct delayed_work cl_renewd; ++ struct rpc_wait_queue cl_rpcwaitq; ++ struct idmap *cl_idmap; ++ const char *cl_owner_id; ++ u32 cl_cb_ident; ++ const struct nfs4_minor_version_ops *cl_mvops; ++ long unsigned int cl_mig_gen; ++ struct nfs4_slot_table *cl_slot_tbl; ++ u32 cl_seqid; ++ u32 cl_exchange_flags; ++ struct nfs4_session *cl_session; ++ bool cl_preserve_clid; ++ struct nfs41_server_owner *cl_serverowner; ++ struct nfs41_server_scope *cl_serverscope; ++ struct nfs41_impl_id *cl_implid; ++ long unsigned int cl_sp4_flags; ++ wait_queue_head_t cl_lock_waitq; ++ char cl_ipaddr[48]; ++ struct net *cl_net; ++ struct list_head pending_cb_stateids; ++}; ++ ++struct pnfs_layout_segment; ++ ++struct nfs_seqid_counter; ++ ++struct nfs_seqid { ++ struct nfs_seqid_counter *sequence; ++ struct list_head list; ++ struct rpc_task *task; ++}; ++ ++struct nfs_write_verifier { ++ char data[8]; ++}; ++ ++struct nfs_writeverf { ++ struct nfs_write_verifier verifier; ++ enum nfs3_stable_how committed; ++}; ++ ++struct nfs_pgio_args { ++ struct nfs4_sequence_args seq_args; ++ struct nfs_fh *fh; ++ struct nfs_open_context *context; ++ struct nfs_lock_context *lock_context; ++ nfs4_stateid stateid; ++ __u64 offset; ++ __u32 count; ++ unsigned int pgbase; ++ struct page **pages; ++ union { ++ unsigned int replen; ++ struct { ++ const u32 *bitmask; ++ u32 bitmask_store[3]; ++ enum nfs3_stable_how stable; ++ }; ++ }; ++}; ++ ++struct nfs_pgio_res { ++ struct nfs4_sequence_res seq_res; ++ struct nfs_fattr *fattr; ++ __u64 count; ++ __u32 op_status; ++ union { ++ struct { ++ unsigned int replen; ++ int eof; ++ void *scratch; ++ }; ++ struct { ++ struct nfs_writeverf *verf; ++ const struct nfs_server *server; ++ }; ++ }; ++}; ++ ++struct nfs_commitargs { ++ struct nfs4_sequence_args seq_args; ++ struct nfs_fh *fh; ++ __u64 offset; ++ __u32 count; ++ const u32 *bitmask; ++}; ++ ++struct nfs_commitres { ++ struct nfs4_sequence_res seq_res; ++ __u32 op_status; ++ struct nfs_fattr *fattr; ++ struct nfs_writeverf *verf; ++ const struct nfs_server *server; ++}; ++ ++struct nfs_removeargs { ++ struct nfs4_sequence_args seq_args; ++ const struct nfs_fh *fh; ++ struct qstr name; ++}; ++ ++struct nfs_removeres { ++ struct nfs4_sequence_res seq_res; ++ struct nfs_server *server; ++ struct nfs_fattr *dir_attr; ++ struct nfs4_change_info cinfo; ++}; ++ ++struct nfs_renameargs { ++ struct nfs4_sequence_args seq_args; ++ const struct nfs_fh *old_dir; ++ const struct nfs_fh *new_dir; ++ const struct qstr *old_name; ++ const struct qstr *new_name; ++}; ++ ++struct nfs_renameres { ++ struct nfs4_sequence_res seq_res; ++ struct nfs_server *server; ++ struct nfs4_change_info old_cinfo; ++ struct nfs_fattr *old_fattr; ++ struct nfs4_change_info new_cinfo; ++ struct nfs_fattr *new_fattr; ++}; ++ ++struct nfs_entry { ++ __u64 ino; ++ __u64 cookie; ++ const char *name; ++ unsigned int len; ++ int eof; ++ struct nfs_fh *fh; ++ struct nfs_fattr *fattr; ++ unsigned char d_type; ++ struct nfs_server *server; ++}; ++ ++struct nfs_readdir_arg { ++ struct dentry *dentry; ++ const struct cred *cred; ++ __be32 *verf; ++ u64 cookie; ++ struct page **pages; ++ unsigned int page_len; ++ bool plus; ++}; ++ ++struct nfs_readdir_res { ++ __be32 *verf; ++}; ++ ++struct nfs4_pathname { ++ unsigned int ncomponents; ++ struct nfs4_string components[512]; ++}; ++ ++struct nfs4_fs_location { ++ unsigned int nservers; ++ struct nfs4_string servers[10]; ++ struct nfs4_pathname rootpath; ++}; ++ ++struct nfs4_fs_locations { ++ struct nfs_fattr *fattr; ++ const struct nfs_server *server; ++ struct nfs4_pathname fs_path; ++ int nlocations; ++ struct nfs4_fs_location locations[10]; ++}; ++ ++struct nfstime4 { ++ u64 seconds; ++ u32 nseconds; ++}; ++ ++struct pnfs_commit_ops; ++ ++struct pnfs_ds_commit_info { ++ struct list_head commits; ++ unsigned int nwritten; ++ unsigned int ncommitting; ++ const struct pnfs_commit_ops *ops; ++}; ++ ++struct nfs41_server_owner { ++ uint64_t minor_id; ++ uint32_t major_id_sz; ++ char major_id[1024]; ++}; ++ ++struct nfs41_server_scope { ++ uint32_t server_scope_sz; ++ char server_scope[1024]; ++}; ++ ++struct nfs41_impl_id { ++ char domain[1025]; ++ char name[1025]; ++ struct nfstime4 date; ++}; ++ ++struct nfs_page_array { ++ struct page **pagevec; ++ unsigned int npages; ++ struct page *page_array[8]; ++}; ++ ++struct nfs_page; ++ ++struct nfs_rw_ops; ++ ++struct nfs_io_completion; ++ ++struct nfs_direct_req; ++ ++struct nfs_pgio_completion_ops; ++ ++struct nfs_pgio_header { ++ struct inode *inode; ++ const struct cred *cred; ++ struct list_head pages; ++ struct nfs_page *req; ++ struct nfs_writeverf verf; ++ fmode_t rw_mode; ++ struct pnfs_layout_segment *lseg; ++ loff_t io_start; ++ const struct rpc_call_ops *mds_ops; ++ void (*release)(struct nfs_pgio_header *); ++ const struct nfs_pgio_completion_ops *completion_ops; ++ const struct nfs_rw_ops *rw_ops; ++ struct nfs_io_completion *io_completion; ++ struct nfs_direct_req *dreq; ++ void *netfs; ++ int pnfs_error; ++ int error; ++ unsigned int good_bytes; ++ long unsigned int flags; ++ struct rpc_task task; ++ struct nfs_fattr fattr; ++ struct nfs_pgio_args args; ++ struct nfs_pgio_res res; ++ long unsigned int timestamp; ++ int (*pgio_done_cb)(struct rpc_task *, struct nfs_pgio_header *); ++ __u64 mds_offset; ++ struct nfs_page_array page_array; ++ struct nfs_client *ds_clp; ++ u32 ds_commit_idx; ++ u32 pgio_mirror_idx; ++}; ++ ++struct nfs_pgio_completion_ops { ++ void (*error_cleanup)(struct list_head *, int); ++ void (*init_hdr)(struct nfs_pgio_header *); ++ void (*completion)(struct nfs_pgio_header *); ++ void (*reschedule_io)(struct nfs_pgio_header *); ++}; ++ ++struct nfs_mds_commit_info { ++ atomic_t rpcs_out; ++ atomic_long_t ncommit; ++ struct list_head list; ++}; ++ ++struct nfs_commit_data; ++ ++struct nfs_commit_info; ++ ++struct nfs_commit_completion_ops { ++ void (*completion)(struct nfs_commit_data *); ++ void (*resched_write)(struct nfs_commit_info *, struct nfs_page *); ++}; ++ ++struct nfs_commit_data { ++ struct rpc_task task; ++ struct inode *inode; ++ const struct cred *cred; ++ struct nfs_fattr fattr; ++ struct nfs_writeverf verf; ++ struct list_head pages; ++ struct list_head list; ++ struct nfs_direct_req *dreq; ++ struct nfs_commitargs args; ++ struct nfs_commitres res; ++ struct nfs_open_context *context; ++ struct pnfs_layout_segment *lseg; ++ struct nfs_client *ds_clp; ++ int ds_commit_index; ++ loff_t lwb; ++ const struct rpc_call_ops *mds_ops; ++ const struct nfs_commit_completion_ops *completion_ops; ++ int (*commit_done_cb)(struct rpc_task *, struct nfs_commit_data *); ++ long unsigned int flags; ++}; ++ ++struct nfs_commit_info { ++ struct inode *inode; ++ struct nfs_mds_commit_info *mds; ++ struct pnfs_ds_commit_info *ds; ++ struct nfs_direct_req *dreq; ++ const struct nfs_commit_completion_ops *completion_ops; ++}; ++ ++struct nfs_unlinkdata { ++ struct nfs_removeargs args; ++ struct nfs_removeres res; ++ struct dentry *dentry; ++ wait_queue_head_t wq; ++ const struct cred *cred; ++ struct nfs_fattr dir_attr; ++ long int timeout; ++}; ++ ++struct nfs_renamedata { ++ struct nfs_renameargs args; ++ struct nfs_renameres res; ++ struct rpc_task task; ++ const struct cred *cred; ++ struct inode *old_dir; ++ struct dentry *old_dentry; ++ struct nfs_fattr old_fattr; ++ struct inode *new_dir; ++ struct dentry *new_dentry; ++ struct nfs_fattr new_fattr; ++ void (*complete)(struct rpc_task *, struct nfs_renamedata *); ++ long int timeout; ++ bool cancelled; ++}; ++ ++struct nlmclnt_operations; ++ ++struct nfs_client_initdata; ++ ++struct nfs_access_entry; ++ ++struct nfs_rpc_ops { ++ u32 version; ++ const struct dentry_operations *dentry_ops; ++ const struct inode_operations *dir_inode_ops; ++ const struct inode_operations *file_inode_ops; ++ const struct file_operations *file_ops; ++ const struct nlmclnt_operations *nlmclnt_ops; ++ int (*getroot)(struct nfs_server *, struct nfs_fh *, struct nfs_fsinfo *); ++ int (*submount)(struct fs_context *, struct nfs_server *); ++ int (*try_get_tree)(struct fs_context *); ++ int (*getattr)(struct nfs_server *, struct nfs_fh *, struct nfs_fattr *, struct inode *); ++ int (*setattr)(struct dentry *, struct nfs_fattr *, struct iattr *); ++ int (*lookup)(struct inode *, struct dentry *, struct nfs_fh *, struct nfs_fattr *); ++ int (*lookupp)(struct inode *, struct nfs_fh *, struct nfs_fattr *); ++ int (*access)(struct inode *, struct nfs_access_entry *, const struct cred *); ++ int (*readlink)(struct inode *, struct page *, unsigned int, unsigned int); ++ int (*create)(struct inode *, struct dentry *, struct iattr *, int); ++ int (*remove)(struct inode *, struct dentry *); ++ void (*unlink_setup)(struct rpc_message *, struct dentry *, struct inode *); ++ void (*unlink_rpc_prepare)(struct rpc_task *, struct nfs_unlinkdata *); ++ int (*unlink_done)(struct rpc_task *, struct inode *); ++ void (*rename_setup)(struct rpc_message *, struct dentry *, struct dentry *); ++ void (*rename_rpc_prepare)(struct rpc_task *, struct nfs_renamedata *); ++ int (*rename_done)(struct rpc_task *, struct inode *, struct inode *); ++ int (*link)(struct inode *, struct inode *, const struct qstr *); ++ int (*symlink)(struct inode *, struct dentry *, struct page *, unsigned int, struct iattr *); ++ int (*mkdir)(struct inode *, struct dentry *, struct iattr *); ++ int (*rmdir)(struct inode *, const struct qstr *); ++ int (*readdir)(struct nfs_readdir_arg *, struct nfs_readdir_res *); ++ int (*mknod)(struct inode *, struct dentry *, struct iattr *, dev_t); ++ int (*statfs)(struct nfs_server *, struct nfs_fh *, struct nfs_fsstat *); ++ int (*fsinfo)(struct nfs_server *, struct nfs_fh *, struct nfs_fsinfo *); ++ int (*pathconf)(struct nfs_server *, struct nfs_fh *, struct nfs_pathconf *); ++ int (*set_capabilities)(struct nfs_server *, struct nfs_fh *); ++ int (*decode_dirent)(struct xdr_stream *, struct nfs_entry *, bool); ++ int (*pgio_rpc_prepare)(struct rpc_task *, struct nfs_pgio_header *); ++ void (*read_setup)(struct nfs_pgio_header *, struct rpc_message *); ++ int (*read_done)(struct rpc_task *, struct nfs_pgio_header *); ++ void (*write_setup)(struct nfs_pgio_header *, struct rpc_message *, struct rpc_clnt **); ++ int (*write_done)(struct rpc_task *, struct nfs_pgio_header *); ++ void (*commit_setup)(struct nfs_commit_data *, struct rpc_message *, struct rpc_clnt **); ++ void (*commit_rpc_prepare)(struct rpc_task *, struct nfs_commit_data *); ++ int (*commit_done)(struct rpc_task *, struct nfs_commit_data *); ++ int (*lock)(struct file *, int, struct file_lock *); ++ int (*lock_check_bounds)(const struct file_lock *); ++ void (*clear_acl_cache)(struct inode *); ++ void (*close_context)(struct nfs_open_context *, int); ++ struct inode * (*open_context)(struct inode *, struct nfs_open_context *, int, struct iattr *, int *); ++ int (*have_delegation)(struct inode *, fmode_t); ++ struct nfs_client * (*alloc_client)(const struct nfs_client_initdata *); ++ struct nfs_client * (*init_client)(struct nfs_client *, const struct nfs_client_initdata *); ++ void (*free_client)(struct nfs_client *); ++ struct nfs_server * (*create_server)(struct fs_context *); ++ struct nfs_server * (*clone_server)(struct nfs_server *, struct nfs_fh *, struct nfs_fattr *, rpc_authflavor_t); ++ int (*discover_trunking)(struct nfs_server *, struct nfs_fh *); ++ void (*enable_swap)(struct inode *); ++ void (*disable_swap)(struct inode *); ++}; ++ ++struct nfs_access_entry { ++ struct rb_node rb_node; ++ struct list_head lru; ++ kuid_t fsuid; ++ kgid_t fsgid; ++ struct group_info *group_info; ++ u64 timestamp; ++ __u32 mask; ++ struct callback_head callback_head; ++}; ++ ++struct nfs4_state_recovery_ops; ++ ++struct nfs4_state_maintenance_ops; ++ ++struct nfs4_mig_recovery_ops; ++ ++struct nfs4_minor_version_ops { ++ u32 minor_version; ++ unsigned int init_caps; ++ int (*init_client)(struct nfs_client *); ++ void (*shutdown_client)(struct nfs_client *); ++ bool (*match_stateid)(const nfs4_stateid *, const nfs4_stateid *); ++ int (*find_root_sec)(struct nfs_server *, struct nfs_fh *, struct nfs_fsinfo *); ++ void (*free_lock_state)(struct nfs_server *, struct nfs4_lock_state *); ++ int (*test_and_free_expired)(struct nfs_server *, nfs4_stateid *, const struct cred *); ++ struct nfs_seqid * (*alloc_seqid)(struct nfs_seqid_counter *, gfp_t); ++ void (*session_trunk)(struct rpc_clnt *, struct rpc_xprt *, void *); ++ const struct rpc_call_ops *call_sync_ops; ++ const struct nfs4_state_recovery_ops *reboot_recovery_ops; ++ const struct nfs4_state_recovery_ops *nograce_recovery_ops; ++ const struct nfs4_state_maintenance_ops *state_renewal_ops; ++ const struct nfs4_mig_recovery_ops *mig_recovery_ops; ++}; ++ ++struct nfs4_state_owner; ++ ++struct nfs4_state { ++ struct list_head open_states; ++ struct list_head inode_states; ++ struct list_head lock_states; ++ struct nfs4_state_owner *owner; ++ struct inode *inode; ++ long unsigned int flags; ++ spinlock_t state_lock; ++ seqlock_t seqlock; ++ nfs4_stateid stateid; ++ nfs4_stateid open_stateid; ++ unsigned int n_rdonly; ++ unsigned int n_wronly; ++ unsigned int n_rdwr; ++ fmode_t state; ++ refcount_t count; ++ wait_queue_head_t waitq; ++ struct callback_head callback_head; ++}; ++ ++struct cache_head { ++ struct hlist_node cache_list; ++ time64_t expiry_time; ++ time64_t last_refresh; ++ struct kref ref; ++ long unsigned int flags; ++}; ++ ++struct cache_deferred_req; ++ ++struct cache_req { ++ struct cache_deferred_req * (*defer)(struct cache_req *); ++ long unsigned int thread_wait; ++}; ++ ++struct cache_deferred_req { ++ struct hlist_node hash; ++ struct list_head recent; ++ struct cache_head *item; ++ void *owner; ++ void (*revisit)(struct cache_deferred_req *, int); ++}; ++ ++struct svc_cred { ++ kuid_t cr_uid; ++ kgid_t cr_gid; ++ struct group_info *cr_group_info; ++ u32 cr_flavor; ++ char *cr_raw_principal; ++ char *cr_principal; ++ char *cr_targ_princ; ++ struct gss_api_mech *cr_gss_mech; ++}; ++ ++struct auth_ops { ++ char *name; ++ struct module *owner; ++ int flavour; ++ int (*accept)(struct svc_rqst *); ++ int (*release)(struct svc_rqst *); ++ void (*domain_release)(struct auth_domain *); ++ int (*set_client)(struct svc_rqst *); ++}; ++ ++struct svc_cacherep; ++ ++struct svc_procedure; ++ ++struct svc_deferred_req; ++ ++struct svc_rqst { ++ struct list_head rq_all; ++ struct callback_head rq_rcu_head; ++ struct svc_xprt *rq_xprt; ++ struct __kernel_sockaddr_storage rq_addr; ++ size_t rq_addrlen; ++ struct __kernel_sockaddr_storage rq_daddr; ++ size_t rq_daddrlen; ++ struct svc_serv *rq_server; ++ struct svc_pool *rq_pool; ++ const struct svc_procedure *rq_procinfo; ++ struct auth_ops *rq_authop; ++ struct svc_cred rq_cred; ++ void *rq_xprt_ctxt; ++ struct svc_deferred_req *rq_deferred; ++ struct xdr_buf rq_arg; ++ struct xdr_stream rq_arg_stream; ++ struct xdr_stream rq_res_stream; ++ struct page *rq_scratch_page; ++ struct xdr_buf rq_res; ++ struct page *rq_pages[260]; ++ struct page **rq_respages; ++ struct page **rq_next_page; ++ struct page **rq_page_end; ++ struct pagevec rq_pvec; ++ struct kvec rq_vec[259]; ++ struct bio_vec rq_bvec[259]; ++ __be32 rq_xid; ++ u32 rq_prog; ++ u32 rq_vers; ++ u32 rq_proc; ++ u32 rq_prot; ++ int rq_cachetype; ++ long unsigned int rq_flags; ++ ktime_t rq_qtime; ++ void *rq_argp; ++ void *rq_resp; ++ __be32 *rq_accept_statp; ++ void *rq_auth_data; ++ __be32 rq_auth_stat; ++ int rq_auth_slack; ++ int rq_reserved; ++ ktime_t rq_stime; ++ struct cache_req rq_chandle; ++ struct auth_domain *rq_client; ++ struct auth_domain *rq_gssclient; ++ struct svc_cacherep *rq_cacherep; ++ struct task_struct *rq_task; ++ struct net *rq_bc_net; ++ void **rq_lease_breaker; ++}; ++ ++struct svc_pool { ++ unsigned int sp_id; ++ spinlock_t sp_lock; ++ struct list_head sp_sockets; ++ unsigned int sp_nrthreads; ++ struct list_head sp_all_threads; ++ struct percpu_counter sp_sockets_queued; ++ struct percpu_counter sp_threads_woken; ++ struct percpu_counter sp_threads_timedout; ++ long unsigned int sp_flags; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct svc_procedure { ++ __be32 (*pc_func)(struct svc_rqst *); ++ bool (*pc_decode)(struct svc_rqst *, struct xdr_stream *); ++ bool (*pc_encode)(struct svc_rqst *, struct xdr_stream *); ++ void (*pc_release)(struct svc_rqst *); ++ unsigned int pc_argsize; ++ unsigned int pc_argzero; ++ unsigned int pc_ressize; ++ unsigned int pc_cachetype; ++ unsigned int pc_xdrressize; ++ const char *pc_name; ++}; ++ ++struct svc_deferred_req { ++ u32 prot; ++ struct svc_xprt *xprt; ++ struct __kernel_sockaddr_storage addr; ++ size_t addrlen; ++ struct __kernel_sockaddr_storage daddr; ++ size_t daddrlen; ++ void *xprt_ctxt; ++ struct cache_deferred_req handle; ++ int argslen; ++ __be32 args[0]; ++}; ++ ++struct svc_process_info { ++ union { ++ int (*dispatch)(struct svc_rqst *); ++ struct { ++ unsigned int lovers; ++ unsigned int hivers; ++ } mismatch; ++ }; ++}; ++ ++struct svc_version { ++ u32 vs_vers; ++ u32 vs_nproc; ++ const struct svc_procedure *vs_proc; ++ long unsigned int *vs_count; ++ u32 vs_xdrsize; ++ bool vs_hidden; ++ bool vs_rpcb_optnl; ++ bool vs_need_cong_ctrl; ++ int (*vs_dispatch)(struct svc_rqst *); ++}; ++ ++struct nfs4_ssc_client_ops; ++ ++struct nfs_ssc_client_ops; ++ ++struct nfs_ssc_client_ops_tbl { ++ const struct nfs4_ssc_client_ops *ssc_nfs4_ops; ++ const struct nfs_ssc_client_ops *ssc_nfs_ops; ++}; ++ ++struct nfs4_ssc_client_ops { ++ struct file * (*sco_open)(struct vfsmount *, struct nfs_fh *, nfs4_stateid *); ++ void (*sco_close)(struct file *); ++}; ++ ++struct nfs_ssc_client_ops { ++ void (*sco_sb_deactive)(struct super_block *); ++}; ++ ++struct nfs_seqid_counter { ++ ktime_t create_time; ++ int owner_id; ++ int flags; ++ u32 counter; ++ spinlock_t lock; ++ struct list_head list; ++ struct rpc_wait_queue wait; ++}; ++ ++struct nfs4_lock_state { ++ struct list_head ls_locks; ++ struct nfs4_state *ls_state; ++ long unsigned int ls_flags; ++ struct nfs_seqid_counter ls_seqid; ++ nfs4_stateid ls_stateid; ++ refcount_t ls_count; ++ fl_owner_t ls_owner; ++}; ++ ++struct nfs4_state_recovery_ops { ++ int owner_flag_bit; ++ int state_flag_bit; ++ int (*recover_open)(struct nfs4_state_owner *, struct nfs4_state *); ++ int (*recover_lock)(struct nfs4_state *, struct file_lock *); ++ int (*establish_clid)(struct nfs_client *, const struct cred *); ++ int (*reclaim_complete)(struct nfs_client *, const struct cred *); ++ int (*detect_trunking)(struct nfs_client *, struct nfs_client **, const struct cred *); ++}; ++ ++struct nfs4_state_maintenance_ops { ++ int (*sched_state_renewal)(struct nfs_client *, const struct cred *, unsigned int); ++ const struct cred * (*get_state_renewal_cred)(struct nfs_client *); ++ int (*renew_lease)(struct nfs_client *, const struct cred *); ++}; ++ ++struct nfs4_mig_recovery_ops { ++ int (*get_locations)(struct nfs_server *, struct nfs_fh *, struct nfs4_fs_locations *, struct page *, const struct cred *); ++ int (*fsid_present)(struct inode *, const struct cred *); ++}; ++ ++struct nfs4_state_owner { ++ struct nfs_server *so_server; ++ struct list_head so_lru; ++ long unsigned int so_expires; ++ struct rb_node so_server_node; ++ const struct cred *so_cred; ++ spinlock_t so_lock; ++ atomic_t so_count; ++ long unsigned int so_flags; ++ struct list_head so_states; ++ struct nfs_seqid_counter so_seqid; ++ seqcount_spinlock_t so_reclaim_seqcount; ++ struct mutex so_delegreturn_mutex; ++}; ++ ++struct iomap_dio_ops { ++ int (*end_io)(struct kiocb *, ssize_t, int, unsigned int); ++ void (*submit_io)(const struct iomap_iter *, struct bio *, loff_t); ++ struct bio_set *bio_set; ++}; ++ ++struct iomap_dio { ++ struct kiocb *iocb; ++ const struct iomap_dio_ops *dops; ++ loff_t i_size; ++ loff_t size; ++ atomic_t ref; ++ unsigned int flags; ++ int error; ++ size_t done_before; ++ bool wait_for_completion; ++ union { ++ struct { ++ struct iov_iter *iter; ++ struct task_struct *waiter; ++ struct bio *poll_bio; ++ } submit; ++ struct { ++ struct work_struct work; ++ } aio; ++ }; ++}; ++ ++struct iomap_swapfile_info { ++ struct iomap iomap; ++ struct swap_info_struct *sis; ++ uint64_t lowest_ppage; ++ uint64_t highest_ppage; ++ long unsigned int nr_pages; ++ int nr_extents; ++ struct file *file; ++}; ++ ++struct qtree_fmt_operations { ++ void (*mem2disk_dqblk)(void *, struct dquot *); ++ void (*disk2mem_dqblk)(struct dquot *, void *); ++ int (*is_id)(void *, struct dquot *); ++}; ++ ++struct qtree_mem_dqinfo { ++ struct super_block *dqi_sb; ++ int dqi_type; ++ unsigned int dqi_blocks; ++ unsigned int dqi_free_blk; ++ unsigned int dqi_free_entry; ++ unsigned int dqi_blocksize_bits; ++ unsigned int dqi_entry_size; ++ unsigned int dqi_usable_bs; ++ unsigned int dqi_qtree_depth; ++ const struct qtree_fmt_operations *dqi_ops; ++}; ++ ++struct v2_disk_dqheader { ++ __le32 dqh_magic; ++ __le32 dqh_version; ++}; ++ ++struct v2r0_disk_dqblk { ++ __le32 dqb_id; ++ __le32 dqb_ihardlimit; ++ __le32 dqb_isoftlimit; ++ __le32 dqb_curinodes; ++ __le32 dqb_bhardlimit; ++ __le32 dqb_bsoftlimit; ++ __le64 dqb_curspace; ++ __le64 dqb_btime; ++ __le64 dqb_itime; ++}; ++ ++struct v2r1_disk_dqblk { ++ __le32 dqb_id; ++ __le32 dqb_pad; ++ __le64 dqb_ihardlimit; ++ __le64 dqb_isoftlimit; ++ __le64 dqb_curinodes; ++ __le64 dqb_bhardlimit; ++ __le64 dqb_bsoftlimit; ++ __le64 dqb_curspace; ++ __le64 dqb_btime; ++ __le64 dqb_itime; ++}; ++ ++struct v2_disk_dqinfo { ++ __le32 dqi_bgrace; ++ __le32 dqi_igrace; ++ __le32 dqi_flags; ++ __le32 dqi_blocks; ++ __le32 dqi_free_blk; ++ __le32 dqi_free_entry; ++}; ++ ++struct qt_disk_dqdbheader { ++ __le32 dqdh_next_free; ++ __le32 dqdh_prev_free; ++ __le16 dqdh_entries; ++ __le16 dqdh_pad1; ++ __le32 dqdh_pad2; ++}; ++ ++struct fs_disk_quota { ++ __s8 d_version; ++ __s8 d_flags; ++ __u16 d_fieldmask; ++ __u32 d_id; ++ __u64 d_blk_hardlimit; ++ __u64 d_blk_softlimit; ++ __u64 d_ino_hardlimit; ++ __u64 d_ino_softlimit; ++ __u64 d_bcount; ++ __u64 d_icount; ++ __s32 d_itimer; ++ __s32 d_btimer; ++ __u16 d_iwarns; ++ __u16 d_bwarns; ++ __s8 d_itimer_hi; ++ __s8 d_btimer_hi; ++ __s8 d_rtbtimer_hi; ++ __s8 d_padding2; ++ __u64 d_rtb_hardlimit; ++ __u64 d_rtb_softlimit; ++ __u64 d_rtbcount; ++ __s32 d_rtbtimer; ++ __u16 d_rtbwarns; ++ __s16 d_padding3; ++ char d_padding4[8]; ++}; ++ ++struct fs_qfilestat { ++ __u64 qfs_ino; ++ __u64 qfs_nblks; ++ __u32 qfs_nextents; ++}; ++ ++typedef struct fs_qfilestat fs_qfilestat_t; ++ ++struct fs_quota_stat { ++ __s8 qs_version; ++ __u16 qs_flags; ++ __s8 qs_pad; ++ fs_qfilestat_t qs_uquota; ++ fs_qfilestat_t qs_gquota; ++ __u32 qs_incoredqs; ++ __s32 qs_btimelimit; ++ __s32 qs_itimelimit; ++ __s32 qs_rtbtimelimit; ++ __u16 qs_bwarnlimit; ++ __u16 qs_iwarnlimit; ++}; ++ ++struct fs_qfilestatv { ++ __u64 qfs_ino; ++ __u64 qfs_nblks; ++ __u32 qfs_nextents; ++ __u32 qfs_pad; ++}; ++ ++struct fs_quota_statv { ++ __s8 qs_version; ++ __u8 qs_pad1; ++ __u16 qs_flags; ++ __u32 qs_incoredqs; ++ struct fs_qfilestatv qs_uquota; ++ struct fs_qfilestatv qs_gquota; ++ struct fs_qfilestatv qs_pquota; ++ __s32 qs_btimelimit; ++ __s32 qs_itimelimit; ++ __s32 qs_rtbtimelimit; ++ __u16 qs_bwarnlimit; ++ __u16 qs_iwarnlimit; ++ __u16 qs_rtbwarnlimit; ++ __u16 qs_pad3; ++ __u32 qs_pad4; ++ __u64 qs_pad2[7]; ++}; ++ ++struct if_dqblk { ++ __u64 dqb_bhardlimit; ++ __u64 dqb_bsoftlimit; ++ __u64 dqb_curspace; ++ __u64 dqb_ihardlimit; ++ __u64 dqb_isoftlimit; ++ __u64 dqb_curinodes; ++ __u64 dqb_btime; ++ __u64 dqb_itime; ++ __u32 dqb_valid; ++}; ++ ++struct if_nextdqblk { ++ __u64 dqb_bhardlimit; ++ __u64 dqb_bsoftlimit; ++ __u64 dqb_curspace; ++ __u64 dqb_ihardlimit; ++ __u64 dqb_isoftlimit; ++ __u64 dqb_curinodes; ++ __u64 dqb_btime; ++ __u64 dqb_itime; ++ __u32 dqb_valid; ++ __u32 dqb_id; ++}; ++ ++struct if_dqinfo { ++ __u64 dqi_bgrace; ++ __u64 dqi_igrace; ++ __u32 dqi_flags; ++ __u32 dqi_valid; ++}; ++ ++struct compat_if_dqblk { ++ compat_u64 dqb_bhardlimit; ++ compat_u64 dqb_bsoftlimit; ++ compat_u64 dqb_curspace; ++ compat_u64 dqb_ihardlimit; ++ compat_u64 dqb_isoftlimit; ++ compat_u64 dqb_curinodes; ++ compat_u64 dqb_btime; ++ compat_u64 dqb_itime; ++ compat_uint_t dqb_valid; ++}; ++ ++struct pde_opener { ++ struct list_head lh; ++ struct file *file; ++ bool closing; ++ struct completion *c; ++}; ++ ++enum { ++ BIAS = 2147483648, ++}; ++ ++struct proc_fs_context { ++ struct pid_namespace *pid_ns; ++ unsigned int mask; ++ enum proc_hidepid hidepid; ++ int gid; ++ enum proc_pidonly pidonly; ++}; ++ ++enum proc_param { ++ Opt_gid___4 = 0, ++ Opt_hidepid = 1, ++ Opt_subset = 2, ++}; ++ ++typedef struct dentry *instantiate_t(struct dentry *, struct task_struct *, const void *); ++ ++struct fd_data { ++ fmode_t mode; ++ unsigned int fd; ++}; ++ ++struct vmcore { ++ struct list_head list; ++ long long unsigned int paddr; ++ long long unsigned int size; ++ loff_t offset; ++}; ++ ++struct vmcoredd_node { ++ struct list_head list; ++ void *buf; ++ unsigned int size; ++}; ++ ++typedef struct elf32_hdr Elf32_Ehdr; ++ ++typedef struct elf32_phdr Elf32_Phdr; ++ ++typedef struct elf32_note Elf32_Nhdr; ++ ++struct vmcoredd_header { ++ __u32 n_namesz; ++ __u32 n_descsz; ++ __u32 n_type; ++ __u8 name[8]; ++ __u8 dump_name[44]; ++}; ++ ++struct vmcore_cb { ++ bool (*pfn_is_ram)(struct vmcore_cb *, long unsigned int); ++ struct list_head next; ++}; ++ ++struct vmcoredd_data { ++ char dump_name[44]; ++ unsigned int size; ++ int (*vmcoredd_callback)(struct vmcoredd_data *, void *); ++}; ++ ++struct kernfs_super_info { ++ struct super_block *sb; ++ struct kernfs_root *root; ++ const void *ns; ++ struct list_head node; ++}; ++ ++struct pts_mount_opts { ++ int setuid; ++ int setgid; ++ kuid_t uid; ++ kgid_t gid; ++ umode_t mode; ++ umode_t ptmxmode; ++ int reserve; ++ int max; ++}; ++ ++enum { ++ Opt_uid___4 = 0, ++ Opt_gid___5 = 1, ++ Opt_mode___2 = 2, ++ Opt_ptmxmode = 3, ++ Opt_newinstance = 4, ++ Opt_max = 5, ++ Opt_err___4 = 6, ++}; ++ ++struct pts_fs_info { ++ struct ida allocated_ptys; ++ struct pts_mount_opts mount_opts; ++ struct super_block *sb; ++ struct dentry *ptmx_dentry; ++}; ++ ++struct ext4_system_zone { ++ struct rb_node node; ++ ext4_fsblk_t start_blk; ++ unsigned int count; ++ u32 ino; ++}; ++ ++enum SHIFT_DIRECTION { ++ SHIFT_LEFT = 0, ++ SHIFT_RIGHT = 1, ++}; ++ ++struct ext4_io_end_vec { ++ struct list_head list; ++ loff_t offset; ++ ssize_t size; ++}; ++ ++struct ext4_io_end { ++ struct list_head list; ++ handle_t *handle; ++ struct inode *inode; ++ struct bio *bio; ++ unsigned int flag; ++ refcount_t count; ++ struct list_head list_vec; ++}; ++ ++typedef struct ext4_io_end ext4_io_end_t; ++ ++struct ext4_extent_tail { ++ __le32 et_checksum; ++}; ++ ++struct ext4_extent { ++ __le32 ee_block; ++ __le16 ee_len; ++ __le16 ee_start_hi; ++ __le32 ee_start_lo; ++}; ++ ++struct ext4_extent_idx { ++ __le32 ei_block; ++ __le32 ei_leaf_lo; ++ __le16 ei_leaf_hi; ++ __u16 ei_unused; ++}; ++ ++struct ext4_extent_header { ++ __le16 eh_magic; ++ __le16 eh_entries; ++ __le16 eh_max; ++ __le16 eh_depth; ++ __le32 eh_generation; ++}; ++ ++struct ext4_ext_path { ++ ext4_fsblk_t p_block; ++ __u16 p_depth; ++ __u16 p_maxdepth; ++ struct ext4_extent *p_ext; ++ struct ext4_extent_idx *p_idx; ++ struct ext4_extent_header *p_hdr; ++ struct buffer_head *p_bh; ++}; ++ ++struct partial_cluster { ++ ext4_fsblk_t pclu; ++ ext4_lblk_t lblk; ++ enum { ++ initial = 0, ++ tofree = 1, ++ nofree = 2, ++ } state; ++}; ++ ++struct migrate_struct { ++ ext4_lblk_t first_block; ++ ext4_lblk_t last_block; ++ ext4_lblk_t curr_block; ++ ext4_fsblk_t first_pblock; ++ ext4_fsblk_t last_pblock; ++}; ++ ++struct dax_holder_operations { ++ int (*notify_failure)(struct dax_device *, u64, u64, int); ++}; ++ ++enum { ++ I_DATA_SEM_NORMAL = 0, ++ I_DATA_SEM_OTHER = 1, ++ I_DATA_SEM_QUOTA = 2, ++ I_DATA_SEM_EA = 3, ++}; ++ ++struct ext4_lazy_init { ++ long unsigned int li_state; ++ struct list_head li_request_list; ++ struct mutex li_list_mtx; ++}; ++ ++struct ext4_journal_cb_entry { ++ struct list_head jce_list; ++ void (*jce_func)(struct super_block *, struct ext4_journal_cb_entry *, int); ++}; ++ ++struct ext4_prealloc_space { ++ union { ++ struct rb_node inode_node; ++ struct list_head lg_list; ++ } pa_node; ++ struct list_head pa_group_list; ++ union { ++ struct list_head pa_tmp_list; ++ struct callback_head pa_rcu; ++ } u; ++ spinlock_t pa_lock; ++ atomic_t pa_count; ++ unsigned int pa_deleted; ++ ext4_fsblk_t pa_pstart; ++ ext4_lblk_t pa_lstart; ++ ext4_grpblk_t pa_len; ++ ext4_grpblk_t pa_free; ++ short unsigned int pa_type; ++ union { ++ rwlock_t *inode_lock; ++ spinlock_t *lg_lock; ++ } pa_node_lock; ++ struct inode *pa_inode; ++}; ++ ++struct ext4_free_extent { ++ ext4_lblk_t fe_logical; ++ ext4_grpblk_t fe_start; ++ ext4_group_t fe_group; ++ ext4_grpblk_t fe_len; ++}; ++ ++struct ext4_allocation_context { ++ struct inode *ac_inode; ++ struct super_block *ac_sb; ++ struct ext4_free_extent ac_o_ex; ++ struct ext4_free_extent ac_g_ex; ++ struct ext4_free_extent ac_b_ex; ++ struct ext4_free_extent ac_f_ex; ++ __u32 ac_groups_considered; ++ __u32 ac_flags; ++ __u16 ac_groups_scanned; ++ __u16 ac_groups_linear_remaining; ++ __u16 ac_found; ++ __u16 ac_tail; ++ __u16 ac_buddy; ++ __u8 ac_status; ++ __u8 ac_criteria; ++ __u8 ac_2order; ++ __u8 ac_op; ++ struct page *ac_bitmap_page; ++ struct page *ac_buddy_page; ++ struct ext4_prealloc_space *ac_pa; ++ struct ext4_locality_group *ac_lg; ++}; ++ ++struct trace_event_raw_ext4_other_inode_update_time { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ino_t orig_ino; ++ uid_t uid; ++ gid_t gid; ++ __u16 mode; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_free_inode { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ uid_t uid; ++ gid_t gid; ++ __u64 blocks; ++ __u16 mode; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_request_inode { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t dir; ++ __u16 mode; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_allocate_inode { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ino_t dir; ++ __u16 mode; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_evict_inode { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ int nlink; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_drop_inode { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ int drop; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_nfs_commit_metadata { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_mark_inode_dirty { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ long unsigned int ip; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_begin_ordered_truncate { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ loff_t new_size; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4__write_begin { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ loff_t pos; ++ unsigned int len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4__write_end { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ loff_t pos; ++ unsigned int len; ++ unsigned int copied; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_writepages { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ long int nr_to_write; ++ long int pages_skipped; ++ loff_t range_start; ++ loff_t range_end; ++ long unsigned int writeback_index; ++ int sync_mode; ++ char for_kupdate; ++ char range_cyclic; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_da_write_pages { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ long unsigned int first_page; ++ long int nr_to_write; ++ int sync_mode; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_da_write_pages_extent { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ __u64 lblk; ++ __u32 len; ++ __u32 flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_writepages_result { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ int ret; ++ int pages_written; ++ long int pages_skipped; ++ long unsigned int writeback_index; ++ int sync_mode; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4__page_op { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ long unsigned int index; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_invalidate_folio_op { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ long unsigned int index; ++ size_t offset; ++ size_t length; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_discard_blocks { ++ struct trace_entry ent; ++ dev_t dev; ++ __u64 blk; ++ __u64 count; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4__mb_new_pa { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ __u64 pa_pstart; ++ __u64 pa_lstart; ++ __u32 pa_len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_mb_release_inode_pa { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ __u64 block; ++ __u32 count; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_mb_release_group_pa { ++ struct trace_entry ent; ++ dev_t dev; ++ __u64 pa_pstart; ++ __u32 pa_len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_discard_preallocations { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ unsigned int len; ++ unsigned int needed; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_mb_discard_preallocations { ++ struct trace_entry ent; ++ dev_t dev; ++ int needed; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_request_blocks { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ unsigned int len; ++ __u32 logical; ++ __u32 lleft; ++ __u32 lright; ++ __u64 goal; ++ __u64 pleft; ++ __u64 pright; ++ unsigned int flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_allocate_blocks { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ __u64 block; ++ unsigned int len; ++ __u32 logical; ++ __u32 lleft; ++ __u32 lright; ++ __u64 goal; ++ __u64 pleft; ++ __u64 pright; ++ unsigned int flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_free_blocks { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ __u64 block; ++ long unsigned int count; ++ int flags; ++ __u16 mode; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_sync_file_enter { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ino_t parent; ++ int datasync; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_sync_file_exit { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_sync_fs { ++ struct trace_entry ent; ++ dev_t dev; ++ int wait; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_alloc_da_blocks { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ unsigned int data_blocks; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_mballoc_alloc { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ __u32 orig_logical; ++ int orig_start; ++ __u32 orig_group; ++ int orig_len; ++ __u32 goal_logical; ++ int goal_start; ++ __u32 goal_group; ++ int goal_len; ++ __u32 result_logical; ++ int result_start; ++ __u32 result_group; ++ int result_len; ++ __u16 found; ++ __u16 groups; ++ __u16 buddy; ++ __u16 flags; ++ __u16 tail; ++ __u8 cr; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_mballoc_prealloc { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ __u32 orig_logical; ++ int orig_start; ++ __u32 orig_group; ++ int orig_len; ++ __u32 result_logical; ++ int result_start; ++ __u32 result_group; ++ int result_len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4__mballoc { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ int result_start; ++ __u32 result_group; ++ int result_len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_forget { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ __u64 block; ++ int is_metadata; ++ __u16 mode; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_da_update_reserve_space { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ __u64 i_blocks; ++ int used_blocks; ++ int reserved_data_blocks; ++ int quota_claim; ++ __u16 mode; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_da_reserve_space { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ __u64 i_blocks; ++ int reserved_data_blocks; ++ __u16 mode; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_da_release_space { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ __u64 i_blocks; ++ int freed_blocks; ++ int reserved_data_blocks; ++ __u16 mode; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4__bitmap_load { ++ struct trace_entry ent; ++ dev_t dev; ++ __u32 group; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_read_block_bitmap_load { ++ struct trace_entry ent; ++ dev_t dev; ++ __u32 group; ++ bool prefetch; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4__fallocate_mode { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ loff_t offset; ++ loff_t len; ++ int mode; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_fallocate_exit { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ loff_t pos; ++ unsigned int blocks; ++ int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_unlink_enter { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ino_t parent; ++ loff_t size; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_unlink_exit { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4__truncate { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ __u64 blocks; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_ext_convert_to_initialized_enter { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_lblk_t m_lblk; ++ unsigned int m_len; ++ ext4_lblk_t u_lblk; ++ unsigned int u_len; ++ ext4_fsblk_t u_pblk; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_ext_convert_to_initialized_fastpath { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_lblk_t m_lblk; ++ unsigned int m_len; ++ ext4_lblk_t u_lblk; ++ unsigned int u_len; ++ ext4_fsblk_t u_pblk; ++ ext4_lblk_t i_lblk; ++ unsigned int i_len; ++ ext4_fsblk_t i_pblk; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4__map_blocks_enter { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_lblk_t lblk; ++ unsigned int len; ++ unsigned int flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4__map_blocks_exit { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ unsigned int flags; ++ ext4_fsblk_t pblk; ++ ext4_lblk_t lblk; ++ unsigned int len; ++ unsigned int mflags; ++ int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_ext_load_extent { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_fsblk_t pblk; ++ ext4_lblk_t lblk; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_load_inode { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_journal_start_sb { ++ struct trace_entry ent; ++ dev_t dev; ++ long unsigned int ip; ++ int blocks; ++ int rsv_blocks; ++ int revoke_creds; ++ int type; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_journal_start_inode { ++ struct trace_entry ent; ++ long unsigned int ino; ++ dev_t dev; ++ long unsigned int ip; ++ int blocks; ++ int rsv_blocks; ++ int revoke_creds; ++ int type; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_journal_start_reserved { ++ struct trace_entry ent; ++ dev_t dev; ++ long unsigned int ip; ++ int blocks; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4__trim { ++ struct trace_entry ent; ++ int dev_major; ++ int dev_minor; ++ __u32 group; ++ int start; ++ int len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_ext_handle_unwritten_extents { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ int flags; ++ ext4_lblk_t lblk; ++ ext4_fsblk_t pblk; ++ unsigned int len; ++ unsigned int allocated; ++ ext4_fsblk_t newblk; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_get_implied_cluster_alloc_exit { ++ struct trace_entry ent; ++ dev_t dev; ++ unsigned int flags; ++ ext4_lblk_t lblk; ++ ext4_fsblk_t pblk; ++ unsigned int len; ++ int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_ext_show_extent { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_fsblk_t pblk; ++ ext4_lblk_t lblk; ++ short unsigned int len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_remove_blocks { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_lblk_t from; ++ ext4_lblk_t to; ++ ext4_fsblk_t ee_pblk; ++ ext4_lblk_t ee_lblk; ++ short unsigned int ee_len; ++ ext4_fsblk_t pc_pclu; ++ ext4_lblk_t pc_lblk; ++ int pc_state; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_ext_rm_leaf { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_lblk_t start; ++ ext4_lblk_t ee_lblk; ++ ext4_fsblk_t ee_pblk; ++ short int ee_len; ++ ext4_fsblk_t pc_pclu; ++ ext4_lblk_t pc_lblk; ++ int pc_state; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_ext_rm_idx { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_fsblk_t pblk; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_ext_remove_space { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_lblk_t start; ++ ext4_lblk_t end; ++ int depth; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_ext_remove_space_done { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_lblk_t start; ++ ext4_lblk_t end; ++ int depth; ++ ext4_fsblk_t pc_pclu; ++ ext4_lblk_t pc_lblk; ++ int pc_state; ++ short unsigned int eh_entries; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4__es_extent { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_lblk_t lblk; ++ ext4_lblk_t len; ++ ext4_fsblk_t pblk; ++ char status; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_es_remove_extent { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ loff_t lblk; ++ loff_t len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_es_find_extent_range_enter { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_lblk_t lblk; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_es_find_extent_range_exit { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_lblk_t lblk; ++ ext4_lblk_t len; ++ ext4_fsblk_t pblk; ++ char status; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_es_lookup_extent_enter { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_lblk_t lblk; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_es_lookup_extent_exit { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_lblk_t lblk; ++ ext4_lblk_t len; ++ ext4_fsblk_t pblk; ++ char status; ++ int found; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4__es_shrink_enter { ++ struct trace_entry ent; ++ dev_t dev; ++ int nr_to_scan; ++ int cache_cnt; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_es_shrink_scan_exit { ++ struct trace_entry ent; ++ dev_t dev; ++ int nr_shrunk; ++ int cache_cnt; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_collapse_range { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ loff_t offset; ++ loff_t len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_insert_range { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ loff_t offset; ++ loff_t len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_es_shrink { ++ struct trace_entry ent; ++ dev_t dev; ++ int nr_shrunk; ++ long long unsigned int scan_time; ++ int nr_skipped; ++ int retried; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_es_insert_delayed_block { ++ struct trace_entry ent; ++ dev_t dev; ++ ino_t ino; ++ ext4_lblk_t lblk; ++ ext4_lblk_t len; ++ ext4_fsblk_t pblk; ++ char status; ++ bool allocated; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_fsmap_class { ++ struct trace_entry ent; ++ dev_t dev; ++ dev_t keydev; ++ u32 agno; ++ u64 bno; ++ u64 len; ++ u64 owner; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_getfsmap_class { ++ struct trace_entry ent; ++ dev_t dev; ++ dev_t keydev; ++ u64 block; ++ u64 len; ++ u64 owner; ++ u64 flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_shutdown { ++ struct trace_entry ent; ++ dev_t dev; ++ unsigned int flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_error { ++ struct trace_entry ent; ++ dev_t dev; ++ const char *function; ++ unsigned int line; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_prefetch_bitmaps { ++ struct trace_entry ent; ++ dev_t dev; ++ __u32 group; ++ __u32 next; ++ __u32 ios; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_lazy_itable_init { ++ struct trace_entry ent; ++ dev_t dev; ++ __u32 group; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_fc_replay_scan { ++ struct trace_entry ent; ++ dev_t dev; ++ int error; ++ int off; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_fc_replay { ++ struct trace_entry ent; ++ dev_t dev; ++ int tag; ++ int ino; ++ int priv1; ++ int priv2; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_fc_commit_start { ++ struct trace_entry ent; ++ dev_t dev; ++ tid_t tid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_fc_commit_stop { ++ struct trace_entry ent; ++ dev_t dev; ++ int nblks; ++ int reason; ++ int num_fc; ++ int num_fc_ineligible; ++ int nblks_agg; ++ tid_t tid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_fc_stats { ++ struct trace_entry ent; ++ dev_t dev; ++ unsigned int fc_ineligible_rc[10]; ++ long unsigned int fc_commits; ++ long unsigned int fc_ineligible_commits; ++ long unsigned int fc_numblks; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_fc_track_dentry { ++ struct trace_entry ent; ++ dev_t dev; ++ tid_t t_tid; ++ ino_t i_ino; ++ tid_t i_sync_tid; ++ int error; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_fc_track_inode { ++ struct trace_entry ent; ++ dev_t dev; ++ tid_t t_tid; ++ ino_t i_ino; ++ tid_t i_sync_tid; ++ int error; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_fc_track_range { ++ struct trace_entry ent; ++ dev_t dev; ++ tid_t t_tid; ++ ino_t i_ino; ++ tid_t i_sync_tid; ++ long int start; ++ long int end; ++ int error; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_fc_cleanup { ++ struct trace_entry ent; ++ dev_t dev; ++ int j_fc_off; ++ int full; ++ tid_t tid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ext4_update_sb { ++ struct trace_entry ent; ++ dev_t dev; ++ ext4_fsblk_t fsblk; ++ unsigned int flags; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_ext4_other_inode_update_time {}; ++ ++struct trace_event_data_offsets_ext4_free_inode {}; ++ ++struct trace_event_data_offsets_ext4_request_inode {}; ++ ++struct trace_event_data_offsets_ext4_allocate_inode {}; ++ ++struct trace_event_data_offsets_ext4_evict_inode {}; ++ ++struct trace_event_data_offsets_ext4_drop_inode {}; ++ ++struct trace_event_data_offsets_ext4_nfs_commit_metadata {}; ++ ++struct trace_event_data_offsets_ext4_mark_inode_dirty {}; ++ ++struct trace_event_data_offsets_ext4_begin_ordered_truncate {}; ++ ++struct trace_event_data_offsets_ext4__write_begin {}; ++ ++struct trace_event_data_offsets_ext4__write_end {}; ++ ++struct trace_event_data_offsets_ext4_writepages {}; ++ ++struct trace_event_data_offsets_ext4_da_write_pages {}; ++ ++struct trace_event_data_offsets_ext4_da_write_pages_extent {}; ++ ++struct trace_event_data_offsets_ext4_writepages_result {}; ++ ++struct trace_event_data_offsets_ext4__page_op {}; ++ ++struct trace_event_data_offsets_ext4_invalidate_folio_op {}; ++ ++struct trace_event_data_offsets_ext4_discard_blocks {}; ++ ++struct trace_event_data_offsets_ext4__mb_new_pa {}; ++ ++struct trace_event_data_offsets_ext4_mb_release_inode_pa {}; ++ ++struct trace_event_data_offsets_ext4_mb_release_group_pa {}; ++ ++struct trace_event_data_offsets_ext4_discard_preallocations {}; ++ ++struct trace_event_data_offsets_ext4_mb_discard_preallocations {}; ++ ++struct trace_event_data_offsets_ext4_request_blocks {}; ++ ++struct trace_event_data_offsets_ext4_allocate_blocks {}; ++ ++struct trace_event_data_offsets_ext4_free_blocks {}; ++ ++struct trace_event_data_offsets_ext4_sync_file_enter {}; ++ ++struct trace_event_data_offsets_ext4_sync_file_exit {}; ++ ++struct trace_event_data_offsets_ext4_sync_fs {}; ++ ++struct trace_event_data_offsets_ext4_alloc_da_blocks {}; ++ ++struct trace_event_data_offsets_ext4_mballoc_alloc {}; ++ ++struct trace_event_data_offsets_ext4_mballoc_prealloc {}; ++ ++struct trace_event_data_offsets_ext4__mballoc {}; ++ ++struct trace_event_data_offsets_ext4_forget {}; ++ ++struct trace_event_data_offsets_ext4_da_update_reserve_space {}; ++ ++struct trace_event_data_offsets_ext4_da_reserve_space {}; ++ ++struct trace_event_data_offsets_ext4_da_release_space {}; ++ ++struct trace_event_data_offsets_ext4__bitmap_load {}; ++ ++struct trace_event_data_offsets_ext4_read_block_bitmap_load {}; ++ ++struct trace_event_data_offsets_ext4__fallocate_mode {}; ++ ++struct trace_event_data_offsets_ext4_fallocate_exit {}; ++ ++struct trace_event_data_offsets_ext4_unlink_enter {}; ++ ++struct trace_event_data_offsets_ext4_unlink_exit {}; ++ ++struct trace_event_data_offsets_ext4__truncate {}; ++ ++struct trace_event_data_offsets_ext4_ext_convert_to_initialized_enter {}; ++ ++struct trace_event_data_offsets_ext4_ext_convert_to_initialized_fastpath {}; ++ ++struct trace_event_data_offsets_ext4__map_blocks_enter {}; ++ ++struct trace_event_data_offsets_ext4__map_blocks_exit {}; ++ ++struct trace_event_data_offsets_ext4_ext_load_extent {}; ++ ++struct trace_event_data_offsets_ext4_load_inode {}; ++ ++struct trace_event_data_offsets_ext4_journal_start_sb {}; ++ ++struct trace_event_data_offsets_ext4_journal_start_inode {}; ++ ++struct trace_event_data_offsets_ext4_journal_start_reserved {}; ++ ++struct trace_event_data_offsets_ext4__trim {}; ++ ++struct trace_event_data_offsets_ext4_ext_handle_unwritten_extents {}; ++ ++struct trace_event_data_offsets_ext4_get_implied_cluster_alloc_exit {}; ++ ++struct trace_event_data_offsets_ext4_ext_show_extent {}; ++ ++struct trace_event_data_offsets_ext4_remove_blocks {}; ++ ++struct trace_event_data_offsets_ext4_ext_rm_leaf {}; ++ ++struct trace_event_data_offsets_ext4_ext_rm_idx {}; ++ ++struct trace_event_data_offsets_ext4_ext_remove_space {}; ++ ++struct trace_event_data_offsets_ext4_ext_remove_space_done {}; ++ ++struct trace_event_data_offsets_ext4__es_extent {}; ++ ++struct trace_event_data_offsets_ext4_es_remove_extent {}; ++ ++struct trace_event_data_offsets_ext4_es_find_extent_range_enter {}; ++ ++struct trace_event_data_offsets_ext4_es_find_extent_range_exit {}; ++ ++struct trace_event_data_offsets_ext4_es_lookup_extent_enter {}; ++ ++struct trace_event_data_offsets_ext4_es_lookup_extent_exit {}; ++ ++struct trace_event_data_offsets_ext4__es_shrink_enter {}; ++ ++struct trace_event_data_offsets_ext4_es_shrink_scan_exit {}; ++ ++struct trace_event_data_offsets_ext4_collapse_range {}; ++ ++struct trace_event_data_offsets_ext4_insert_range {}; ++ ++struct trace_event_data_offsets_ext4_es_shrink {}; ++ ++struct trace_event_data_offsets_ext4_es_insert_delayed_block {}; ++ ++struct trace_event_data_offsets_ext4_fsmap_class {}; ++ ++struct trace_event_data_offsets_ext4_getfsmap_class {}; ++ ++struct trace_event_data_offsets_ext4_shutdown {}; ++ ++struct trace_event_data_offsets_ext4_error {}; ++ ++struct trace_event_data_offsets_ext4_prefetch_bitmaps {}; ++ ++struct trace_event_data_offsets_ext4_lazy_itable_init {}; ++ ++struct trace_event_data_offsets_ext4_fc_replay_scan {}; ++ ++struct trace_event_data_offsets_ext4_fc_replay {}; ++ ++struct trace_event_data_offsets_ext4_fc_commit_start {}; ++ ++struct trace_event_data_offsets_ext4_fc_commit_stop {}; ++ ++struct trace_event_data_offsets_ext4_fc_stats {}; ++ ++struct trace_event_data_offsets_ext4_fc_track_dentry {}; ++ ++struct trace_event_data_offsets_ext4_fc_track_inode {}; ++ ++struct trace_event_data_offsets_ext4_fc_track_range {}; ++ ++struct trace_event_data_offsets_ext4_fc_cleanup {}; ++ ++struct trace_event_data_offsets_ext4_update_sb {}; ++ ++typedef void (*btf_trace_ext4_other_inode_update_time)(void *, struct inode *, ino_t); ++ ++typedef void (*btf_trace_ext4_free_inode)(void *, struct inode *); ++ ++typedef void (*btf_trace_ext4_request_inode)(void *, struct inode *, int); ++ ++typedef void (*btf_trace_ext4_allocate_inode)(void *, struct inode *, struct inode *, int); ++ ++typedef void (*btf_trace_ext4_evict_inode)(void *, struct inode *); ++ ++typedef void (*btf_trace_ext4_drop_inode)(void *, struct inode *, int); ++ ++typedef void (*btf_trace_ext4_nfs_commit_metadata)(void *, struct inode *); ++ ++typedef void (*btf_trace_ext4_mark_inode_dirty)(void *, struct inode *, long unsigned int); ++ ++typedef void (*btf_trace_ext4_begin_ordered_truncate)(void *, struct inode *, loff_t); ++ ++typedef void (*btf_trace_ext4_write_begin)(void *, struct inode *, loff_t, unsigned int); ++ ++typedef void (*btf_trace_ext4_da_write_begin)(void *, struct inode *, loff_t, unsigned int); ++ ++typedef void (*btf_trace_ext4_write_end)(void *, struct inode *, loff_t, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_ext4_journalled_write_end)(void *, struct inode *, loff_t, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_ext4_da_write_end)(void *, struct inode *, loff_t, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_ext4_writepages)(void *, struct inode *, struct writeback_control *); ++ ++typedef void (*btf_trace_ext4_da_write_pages)(void *, struct inode *, long unsigned int, struct writeback_control *); ++ ++typedef void (*btf_trace_ext4_da_write_pages_extent)(void *, struct inode *, struct ext4_map_blocks *); ++ ++typedef void (*btf_trace_ext4_writepages_result)(void *, struct inode *, struct writeback_control *, int, int); ++ ++typedef void (*btf_trace_ext4_readpage)(void *, struct page *); ++ ++typedef void (*btf_trace_ext4_releasepage)(void *, struct page *); ++ ++typedef void (*btf_trace_ext4_invalidate_folio)(void *, struct folio *, size_t, size_t); ++ ++typedef void (*btf_trace_ext4_journalled_invalidate_folio)(void *, struct folio *, size_t, size_t); ++ ++typedef void (*btf_trace_ext4_discard_blocks)(void *, struct super_block *, long long unsigned int, long long unsigned int); ++ ++typedef void (*btf_trace_ext4_mb_new_inode_pa)(void *, struct ext4_allocation_context *, struct ext4_prealloc_space *); ++ ++typedef void (*btf_trace_ext4_mb_new_group_pa)(void *, struct ext4_allocation_context *, struct ext4_prealloc_space *); ++ ++typedef void (*btf_trace_ext4_mb_release_inode_pa)(void *, struct ext4_prealloc_space *, long long unsigned int, unsigned int); ++ ++typedef void (*btf_trace_ext4_mb_release_group_pa)(void *, struct super_block *, struct ext4_prealloc_space *); ++ ++typedef void (*btf_trace_ext4_discard_preallocations)(void *, struct inode *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_ext4_mb_discard_preallocations)(void *, struct super_block *, int); ++ ++typedef void (*btf_trace_ext4_request_blocks)(void *, struct ext4_allocation_request *); ++ ++typedef void (*btf_trace_ext4_allocate_blocks)(void *, struct ext4_allocation_request *, long long unsigned int); ++ ++typedef void (*btf_trace_ext4_free_blocks)(void *, struct inode *, __u64, long unsigned int, int); ++ ++typedef void (*btf_trace_ext4_sync_file_enter)(void *, struct file *, int); ++ ++typedef void (*btf_trace_ext4_sync_file_exit)(void *, struct inode *, int); ++ ++typedef void (*btf_trace_ext4_sync_fs)(void *, struct super_block *, int); ++ ++typedef void (*btf_trace_ext4_alloc_da_blocks)(void *, struct inode *); ++ ++typedef void (*btf_trace_ext4_mballoc_alloc)(void *, struct ext4_allocation_context *); ++ ++typedef void (*btf_trace_ext4_mballoc_prealloc)(void *, struct ext4_allocation_context *); ++ ++typedef void (*btf_trace_ext4_mballoc_discard)(void *, struct super_block *, struct inode *, ext4_group_t, ext4_grpblk_t, ext4_grpblk_t); ++ ++typedef void (*btf_trace_ext4_mballoc_free)(void *, struct super_block *, struct inode *, ext4_group_t, ext4_grpblk_t, ext4_grpblk_t); ++ ++typedef void (*btf_trace_ext4_forget)(void *, struct inode *, int, __u64); ++ ++typedef void (*btf_trace_ext4_da_update_reserve_space)(void *, struct inode *, int, int); ++ ++typedef void (*btf_trace_ext4_da_reserve_space)(void *, struct inode *); ++ ++typedef void (*btf_trace_ext4_da_release_space)(void *, struct inode *, int); ++ ++typedef void (*btf_trace_ext4_mb_bitmap_load)(void *, struct super_block *, long unsigned int); ++ ++typedef void (*btf_trace_ext4_mb_buddy_bitmap_load)(void *, struct super_block *, long unsigned int); ++ ++typedef void (*btf_trace_ext4_load_inode_bitmap)(void *, struct super_block *, long unsigned int); ++ ++typedef void (*btf_trace_ext4_read_block_bitmap_load)(void *, struct super_block *, long unsigned int, bool); ++ ++typedef void (*btf_trace_ext4_fallocate_enter)(void *, struct inode *, loff_t, loff_t, int); ++ ++typedef void (*btf_trace_ext4_punch_hole)(void *, struct inode *, loff_t, loff_t, int); ++ ++typedef void (*btf_trace_ext4_zero_range)(void *, struct inode *, loff_t, loff_t, int); ++ ++typedef void (*btf_trace_ext4_fallocate_exit)(void *, struct inode *, loff_t, unsigned int, int); ++ ++typedef void (*btf_trace_ext4_unlink_enter)(void *, struct inode *, struct dentry *); ++ ++typedef void (*btf_trace_ext4_unlink_exit)(void *, struct dentry *, int); ++ ++typedef void (*btf_trace_ext4_truncate_enter)(void *, struct inode *); ++ ++typedef void (*btf_trace_ext4_truncate_exit)(void *, struct inode *); ++ ++typedef void (*btf_trace_ext4_ext_convert_to_initialized_enter)(void *, struct inode *, struct ext4_map_blocks *, struct ext4_extent *); ++ ++typedef void (*btf_trace_ext4_ext_convert_to_initialized_fastpath)(void *, struct inode *, struct ext4_map_blocks *, struct ext4_extent *, struct ext4_extent *); ++ ++typedef void (*btf_trace_ext4_ext_map_blocks_enter)(void *, struct inode *, ext4_lblk_t, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_ext4_ind_map_blocks_enter)(void *, struct inode *, ext4_lblk_t, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_ext4_ext_map_blocks_exit)(void *, struct inode *, unsigned int, struct ext4_map_blocks *, int); ++ ++typedef void (*btf_trace_ext4_ind_map_blocks_exit)(void *, struct inode *, unsigned int, struct ext4_map_blocks *, int); ++ ++typedef void (*btf_trace_ext4_ext_load_extent)(void *, struct inode *, ext4_lblk_t, ext4_fsblk_t); ++ ++typedef void (*btf_trace_ext4_load_inode)(void *, struct super_block *, long unsigned int); ++ ++typedef void (*btf_trace_ext4_journal_start_sb)(void *, struct super_block *, int, int, int, int, long unsigned int); ++ ++typedef void (*btf_trace_ext4_journal_start_inode)(void *, struct inode *, int, int, int, int, long unsigned int); ++ ++typedef void (*btf_trace_ext4_journal_start_reserved)(void *, struct super_block *, int, long unsigned int); ++ ++typedef void (*btf_trace_ext4_trim_extent)(void *, struct super_block *, ext4_group_t, ext4_grpblk_t, ext4_grpblk_t); ++ ++typedef void (*btf_trace_ext4_trim_all_free)(void *, struct super_block *, ext4_group_t, ext4_grpblk_t, ext4_grpblk_t); ++ ++typedef void (*btf_trace_ext4_ext_handle_unwritten_extents)(void *, struct inode *, struct ext4_map_blocks *, int, unsigned int, ext4_fsblk_t); ++ ++typedef void (*btf_trace_ext4_get_implied_cluster_alloc_exit)(void *, struct super_block *, struct ext4_map_blocks *, int); ++ ++typedef void (*btf_trace_ext4_ext_show_extent)(void *, struct inode *, ext4_lblk_t, ext4_fsblk_t, short unsigned int); ++ ++typedef void (*btf_trace_ext4_remove_blocks)(void *, struct inode *, struct ext4_extent *, ext4_lblk_t, ext4_fsblk_t, struct partial_cluster *); ++ ++typedef void (*btf_trace_ext4_ext_rm_leaf)(void *, struct inode *, ext4_lblk_t, struct ext4_extent *, struct partial_cluster *); ++ ++typedef void (*btf_trace_ext4_ext_rm_idx)(void *, struct inode *, ext4_fsblk_t); ++ ++typedef void (*btf_trace_ext4_ext_remove_space)(void *, struct inode *, ext4_lblk_t, ext4_lblk_t, int); ++ ++typedef void (*btf_trace_ext4_ext_remove_space_done)(void *, struct inode *, ext4_lblk_t, ext4_lblk_t, int, struct partial_cluster *, __le16); ++ ++typedef void (*btf_trace_ext4_es_insert_extent)(void *, struct inode *, struct extent_status *); ++ ++typedef void (*btf_trace_ext4_es_cache_extent)(void *, struct inode *, struct extent_status *); ++ ++typedef void (*btf_trace_ext4_es_remove_extent)(void *, struct inode *, ext4_lblk_t, ext4_lblk_t); ++ ++typedef void (*btf_trace_ext4_es_find_extent_range_enter)(void *, struct inode *, ext4_lblk_t); ++ ++typedef void (*btf_trace_ext4_es_find_extent_range_exit)(void *, struct inode *, struct extent_status *); ++ ++typedef void (*btf_trace_ext4_es_lookup_extent_enter)(void *, struct inode *, ext4_lblk_t); ++ ++typedef void (*btf_trace_ext4_es_lookup_extent_exit)(void *, struct inode *, struct extent_status *, int); ++ ++typedef void (*btf_trace_ext4_es_shrink_count)(void *, struct super_block *, int, int); ++ ++typedef void (*btf_trace_ext4_es_shrink_scan_enter)(void *, struct super_block *, int, int); ++ ++typedef void (*btf_trace_ext4_es_shrink_scan_exit)(void *, struct super_block *, int, int); ++ ++typedef void (*btf_trace_ext4_collapse_range)(void *, struct inode *, loff_t, loff_t); ++ ++typedef void (*btf_trace_ext4_insert_range)(void *, struct inode *, loff_t, loff_t); ++ ++typedef void (*btf_trace_ext4_es_shrink)(void *, struct super_block *, int, u64, int, int); ++ ++typedef void (*btf_trace_ext4_es_insert_delayed_block)(void *, struct inode *, struct extent_status *, bool); ++ ++typedef void (*btf_trace_ext4_fsmap_low_key)(void *, struct super_block *, u32, u32, u64, u64, u64); ++ ++typedef void (*btf_trace_ext4_fsmap_high_key)(void *, struct super_block *, u32, u32, u64, u64, u64); ++ ++typedef void (*btf_trace_ext4_fsmap_mapping)(void *, struct super_block *, u32, u32, u64, u64, u64); ++ ++typedef void (*btf_trace_ext4_getfsmap_low_key)(void *, struct super_block *, struct ext4_fsmap *); ++ ++typedef void (*btf_trace_ext4_getfsmap_high_key)(void *, struct super_block *, struct ext4_fsmap *); ++ ++typedef void (*btf_trace_ext4_getfsmap_mapping)(void *, struct super_block *, struct ext4_fsmap *); ++ ++typedef void (*btf_trace_ext4_shutdown)(void *, struct super_block *, long unsigned int); ++ ++typedef void (*btf_trace_ext4_error)(void *, struct super_block *, const char *, unsigned int); ++ ++typedef void (*btf_trace_ext4_prefetch_bitmaps)(void *, struct super_block *, ext4_group_t, ext4_group_t, unsigned int); ++ ++typedef void (*btf_trace_ext4_lazy_itable_init)(void *, struct super_block *, ext4_group_t); ++ ++typedef void (*btf_trace_ext4_fc_replay_scan)(void *, struct super_block *, int, int); ++ ++typedef void (*btf_trace_ext4_fc_replay)(void *, struct super_block *, int, int, int, int); ++ ++typedef void (*btf_trace_ext4_fc_commit_start)(void *, struct super_block *, tid_t); ++ ++typedef void (*btf_trace_ext4_fc_commit_stop)(void *, struct super_block *, int, int, tid_t); ++ ++typedef void (*btf_trace_ext4_fc_stats)(void *, struct super_block *); ++ ++typedef void (*btf_trace_ext4_fc_track_create)(void *, handle_t *, struct inode *, struct dentry *, int); ++ ++typedef void (*btf_trace_ext4_fc_track_link)(void *, handle_t *, struct inode *, struct dentry *, int); ++ ++typedef void (*btf_trace_ext4_fc_track_unlink)(void *, handle_t *, struct inode *, struct dentry *, int); ++ ++typedef void (*btf_trace_ext4_fc_track_inode)(void *, handle_t *, struct inode *, int); ++ ++typedef void (*btf_trace_ext4_fc_track_range)(void *, handle_t *, struct inode *, long int, long int, int); ++ ++typedef void (*btf_trace_ext4_fc_cleanup)(void *, journal_t *, int, tid_t); ++ ++typedef void (*btf_trace_ext4_update_sb)(void *, struct super_block *, ext4_fsblk_t, unsigned int); ++ ++struct ext4_err_translation { ++ int code; ++ int errno; ++}; ++ ++enum { ++ Opt_bsd_df = 0, ++ Opt_minix_df = 1, ++ Opt_grpid = 2, ++ Opt_nogrpid = 3, ++ Opt_resgid = 4, ++ Opt_resuid = 5, ++ Opt_sb = 6, ++ Opt_nouid32 = 7, ++ Opt_debug = 8, ++ Opt_removed = 9, ++ Opt_user_xattr = 10, ++ Opt_acl___2 = 11, ++ Opt_auto_da_alloc = 12, ++ Opt_noauto_da_alloc = 13, ++ Opt_noload = 14, ++ Opt_commit = 15, ++ Opt_min_batch_time = 16, ++ Opt_max_batch_time = 17, ++ Opt_journal_dev = 18, ++ Opt_journal_path = 19, ++ Opt_journal_checksum = 20, ++ Opt_journal_async_commit = 21, ++ Opt_abort = 22, ++ Opt_data_journal = 23, ++ Opt_data_ordered = 24, ++ Opt_data_writeback = 25, ++ Opt_data_err_abort = 26, ++ Opt_data_err_ignore = 27, ++ Opt_test_dummy_encryption = 28, ++ Opt_inlinecrypt = 29, ++ Opt_usrjquota = 30, ++ Opt_grpjquota = 31, ++ Opt_quota = 32, ++ Opt_noquota = 33, ++ Opt_barrier___2 = 34, ++ Opt_nobarrier___2 = 35, ++ Opt_err___5 = 36, ++ Opt_usrquota = 37, ++ Opt_grpquota = 38, ++ Opt_prjquota = 39, ++ Opt_dax = 40, ++ Opt_dax_always = 41, ++ Opt_dax_inode = 42, ++ Opt_dax_never = 43, ++ Opt_stripe = 44, ++ Opt_delalloc = 45, ++ Opt_nodelalloc = 46, ++ Opt_warn_on_error = 47, ++ Opt_nowarn_on_error = 48, ++ Opt_mblk_io_submit = 49, ++ Opt_debug_want_extra_isize = 50, ++ Opt_nomblk_io_submit = 51, ++ Opt_block_validity = 52, ++ Opt_noblock_validity = 53, ++ Opt_inode_readahead_blks = 54, ++ Opt_journal_ioprio = 55, ++ Opt_dioread_nolock = 56, ++ Opt_dioread_lock = 57, ++ Opt_discard___2 = 58, ++ Opt_nodiscard___2 = 59, ++ Opt_init_itable = 60, ++ Opt_noinit_itable = 61, ++ Opt_max_dir_size_kb = 62, ++ Opt_nojournal_checksum = 63, ++ Opt_nombcache = 64, ++ Opt_no_prefetch_block_bitmaps = 65, ++ Opt_mb_optimize_scan = 66, ++ Opt_errors = 67, ++ Opt_data = 68, ++ Opt_data_err = 69, ++ Opt_jqfmt = 70, ++ Opt_dax_type = 71, ++}; ++ ++struct mount_opts { ++ int token; ++ int mount_opt; ++ int flags; ++}; ++ ++struct ext4_sb_encodings { ++ __u16 magic; ++ char *name; ++ unsigned int version; ++}; ++ ++struct ext4_fs_context { ++ char *s_qf_names[3]; ++ struct fscrypt_dummy_policy dummy_enc_policy; ++ int s_jquota_fmt; ++ short unsigned int qname_spec; ++ long unsigned int vals_s_flags; ++ long unsigned int mask_s_flags; ++ long unsigned int journal_devnum; ++ long unsigned int s_commit_interval; ++ long unsigned int s_stripe; ++ unsigned int s_inode_readahead_blks; ++ unsigned int s_want_extra_isize; ++ unsigned int s_li_wait_mult; ++ unsigned int s_max_dir_size_kb; ++ unsigned int journal_ioprio; ++ unsigned int vals_s_mount_opt; ++ unsigned int mask_s_mount_opt; ++ unsigned int vals_s_mount_opt2; ++ unsigned int mask_s_mount_opt2; ++ long unsigned int vals_s_mount_flags; ++ long unsigned int mask_s_mount_flags; ++ unsigned int opt_flags; ++ unsigned int spec; ++ u32 s_max_batch_time; ++ u32 s_min_batch_time; ++ kuid_t s_resuid; ++ kgid_t s_resgid; ++ ext4_fsblk_t s_sb_block; ++}; ++ ++struct ext4_mount_options { ++ long unsigned int s_mount_opt; ++ long unsigned int s_mount_opt2; ++ kuid_t s_resuid; ++ kgid_t s_resgid; ++ long unsigned int s_commit_interval; ++ u32 s_min_batch_time; ++ u32 s_max_batch_time; ++ int s_jquota_fmt; ++ char *s_qf_names[3]; ++}; ++ ++struct ramfs_mount_opts { ++ umode_t mode; ++}; ++ ++struct ramfs_fs_info { ++ struct ramfs_mount_opts mount_opts; ++}; ++ ++enum ramfs_param { ++ Opt_mode___3 = 0, ++}; ++ ++enum hugetlbfs_size_type { ++ NO_SIZE = 0, ++ SIZE_STD = 1, ++ SIZE_PERCENT = 2, ++}; ++ ++struct hugetlbfs_fs_context { ++ struct hstate *hstate; ++ long long unsigned int max_size_opt; ++ long long unsigned int min_size_opt; ++ long int max_hpages; ++ long int nr_inodes; ++ long int min_hpages; ++ enum hugetlbfs_size_type max_val_type; ++ enum hugetlbfs_size_type min_val_type; ++ kuid_t uid; ++ kgid_t gid; ++ umode_t mode; ++}; ++ ++enum hugetlb_param { ++ Opt_gid___6 = 0, ++ Opt_min_size = 1, ++ Opt_mode___4 = 2, ++ Opt_nr_inodes___2 = 3, ++ Opt_pagesize = 4, ++ Opt_size___2 = 5, ++ Opt_uid___5 = 6, ++}; ++ ++enum { ++ AUTOFS_IOC_READY_CMD = 96, ++ AUTOFS_IOC_FAIL_CMD = 97, ++ AUTOFS_IOC_CATATONIC_CMD = 98, ++ AUTOFS_IOC_PROTOVER_CMD = 99, ++ AUTOFS_IOC_SETTIMEOUT_CMD = 100, ++ AUTOFS_IOC_EXPIRE_CMD = 101, ++}; ++ ++enum { ++ AUTOFS_IOC_EXPIRE_MULTI_CMD = 102, ++ AUTOFS_IOC_PROTOSUBVER_CMD = 103, ++ AUTOFS_IOC_ASKUMOUNT_CMD = 112, ++}; ++ ++struct debugfs_mount_opts { ++ kuid_t uid; ++ kgid_t gid; ++ umode_t mode; ++ unsigned int opts; ++}; ++ ++enum { ++ Opt_uid___6 = 0, ++ Opt_gid___7 = 1, ++ Opt_mode___5 = 2, ++ Opt_err___6 = 3, ++}; ++ ++struct debugfs_fs_info { ++ struct debugfs_mount_opts mount_opts; ++}; ++ ++struct tracefs_dir_ops { ++ int (*mkdir)(const char *); ++ int (*rmdir)(const char *); ++}; ++ ++struct tracefs_mount_opts { ++ kuid_t uid; ++ kgid_t gid; ++ umode_t mode; ++ unsigned int opts; ++}; ++ ++struct tracefs_fs_info { ++ struct tracefs_mount_opts mount_opts; ++}; ++ ++enum btrfs_caching_type { ++ BTRFS_CACHE_NO = 0, ++ BTRFS_CACHE_STARTED = 1, ++ BTRFS_CACHE_FINISHED = 2, ++ BTRFS_CACHE_ERROR = 3, ++}; ++ ++enum btrfs_inline_ref_type { ++ BTRFS_REF_TYPE_INVALID = 0, ++ BTRFS_REF_TYPE_BLOCK = 1, ++ BTRFS_REF_TYPE_DATA = 2, ++ BTRFS_REF_TYPE_ANY = 3, ++}; ++ ++struct btrfs_discard_stripe { ++ struct btrfs_device *dev; ++ u64 physical; ++ u64 length; ++}; ++ ++enum btrfs_loop_type { ++ LOOP_CACHING_NOWAIT = 0, ++ LOOP_CACHING_WAIT = 1, ++ LOOP_UNSET_SIZE_CLASS = 2, ++ LOOP_ALLOC_CHUNK = 3, ++ LOOP_WRONG_SIZE_CLASS = 4, ++ LOOP_NO_EMPTY_SIZE = 5, ++}; ++ ++struct walk_control___2 { ++ u64 refs[8]; ++ u64 flags[8]; ++ struct btrfs_key update_progress; ++ struct btrfs_key drop_progress; ++ int drop_level; ++ int stage; ++ int level; ++ int shared_level; ++ int update_ref; ++ int keep_locks; ++ int reada_slot; ++ int reada_count; ++ int restarted; ++}; ++ ++struct btrfs_csum_item { ++ __u8 csum; ++}; ++ ++enum btrfs_csum_type { ++ BTRFS_CSUM_TYPE_CRC32 = 0, ++ BTRFS_CSUM_TYPE_XXHASH = 1, ++ BTRFS_CSUM_TYPE_SHA256 = 2, ++ BTRFS_CSUM_TYPE_BLAKE2 = 3, ++}; ++ ++enum btrfs_disk_cache_state { ++ BTRFS_DC_WRITTEN = 0, ++ BTRFS_DC_ERROR = 1, ++ BTRFS_DC_CLEAR = 2, ++ BTRFS_DC_SETUP = 3, ++}; ++ ++struct btrfs_ioctl_encoded_io_args { ++ const struct iovec *iov; ++ long unsigned int iovcnt; ++ __s64 offset; ++ __u64 flags; ++ __u64 len; ++ __u64 unencoded_len; ++ __u64 unencoded_offset; ++ __u32 compression; ++ __u32 encryption; ++ __u8 reserved[64]; ++}; ++ ++enum { ++ __PAGE_UNLOCK_BIT = 0, ++ PAGE_UNLOCK = 1, ++ __PAGE_UNLOCK_SEQ = 0, ++ __PAGE_START_WRITEBACK_BIT = 1, ++ PAGE_START_WRITEBACK = 2, ++ __PAGE_START_WRITEBACK_SEQ = 1, ++ __PAGE_END_WRITEBACK_BIT = 2, ++ PAGE_END_WRITEBACK = 4, ++ __PAGE_END_WRITEBACK_SEQ = 2, ++ __PAGE_SET_ORDERED_BIT = 3, ++ PAGE_SET_ORDERED = 8, ++ __PAGE_SET_ORDERED_SEQ = 3, ++ __PAGE_SET_ERROR_BIT = 4, ++ PAGE_SET_ERROR = 16, ++ __PAGE_SET_ERROR_SEQ = 4, ++ __PAGE_LOCK_BIT = 5, ++ PAGE_LOCK = 32, ++ __PAGE_LOCK_SEQ = 5, ++}; ++ ++struct btrfs_file_private { ++ void *filldir_buf; ++ u64 last_index; ++ struct extent_state *llseek_cached_state; ++}; ++ ++struct btrfs_new_inode_args { ++ struct inode *dir; ++ struct dentry *dentry; ++ struct inode *inode; ++ bool orphan; ++ bool subvol; ++ struct posix_acl *default_acl; ++ struct posix_acl *acl; ++ struct fscrypt_name fname; ++}; ++ ++struct btrfs_swapfile_pin { ++ struct rb_node node; ++ void *ptr; ++ struct inode *inode; ++ bool is_block_group; ++ int bg_extent_count; ++}; ++ ++struct btrfs_subpage { ++ spinlock_t lock; ++ atomic_t readers; ++ union { ++ atomic_t eb_refs; ++ atomic_t writers; ++ }; ++ long unsigned int bitmaps[0]; ++}; ++ ++struct btrfs_iget_args { ++ u64 ino; ++ struct btrfs_root *root; ++}; ++ ++struct btrfs_dio_data { ++ ssize_t submitted; ++ struct extent_changeset *data_reserved; ++ struct btrfs_ordered_extent *ordered; ++ bool data_space_reserved; ++ bool nocow_done; ++}; ++ ++struct btrfs_dio_private { ++ u64 file_offset; ++ u32 bytes; ++ struct btrfs_bio bbio; ++}; ++ ++struct btrfs_rename_ctx { ++ u64 index; ++}; ++ ++struct async_extent { ++ u64 start; ++ u64 ram_size; ++ u64 compressed_size; ++ struct page **pages; ++ long unsigned int nr_pages; ++ int compress_type; ++ struct list_head list; ++}; ++ ++struct async_cow; ++ ++struct async_chunk { ++ struct btrfs_inode *inode; ++ struct page *locked_page; ++ u64 start; ++ u64 end; ++ blk_opf_t write_flags; ++ struct list_head extents; ++ struct cgroup_subsys_state *blkcg_css; ++ struct btrfs_work work; ++ struct async_cow *async_cow; ++}; ++ ++struct async_cow { ++ atomic_t num_chunks; ++ struct async_chunk chunks[0]; ++}; ++ ++struct can_nocow_file_extent_args { ++ u64 start; ++ u64 end; ++ bool writeback_path; ++ bool strict; ++ bool free_path; ++ u64 disk_bytenr; ++ u64 disk_num_bytes; ++ u64 extent_offset; ++ u64 num_bytes; ++}; ++ ++struct btrfs_writepage_fixup { ++ struct page *page; ++ struct btrfs_inode *inode; ++ struct btrfs_work work; ++}; ++ ++struct dir_entry { ++ u64 ino; ++ u64 offset; ++ unsigned int type; ++ int name_len; ++}; ++ ++struct btrfs_delalloc_work { ++ struct inode *inode; ++ struct completion completion; ++ struct list_head list; ++ struct btrfs_work work; ++}; ++ ++struct btrfs_encoded_read_private { ++ wait_queue_head_t wait; ++ atomic_t pending; ++ blk_status_t status; ++}; ++ ++struct btrfs_swap_info { ++ u64 start; ++ u64 block_start; ++ u64 block_len; ++ u64 lowest_ppage; ++ u64 highest_ppage; ++ long unsigned int nr_pages; ++ int nr_extents; ++}; ++ ++enum btrfs_feature_set { ++ FEAT_COMPAT = 0, ++ FEAT_COMPAT_RO = 1, ++ FEAT_INCOMPAT = 2, ++ FEAT_MAX = 3, ++}; ++ ++struct btrfs_feature_attr { ++ struct kobj_attribute kobj_attr; ++ enum btrfs_feature_set feature_set; ++ u64 feature_bit; ++}; ++ ++struct raid_kobject { ++ u64 flags; ++ struct kobject kobj; ++}; ++ ++struct btrfs_ioctl_balance_args { ++ __u64 flags; ++ __u64 state; ++ struct btrfs_balance_args data; ++ struct btrfs_balance_args meta; ++ struct btrfs_balance_args sys; ++ struct btrfs_balance_progress stat; ++ __u64 unused[72]; ++}; ++ ++struct btrfs_ioctl_get_dev_stats { ++ __u64 devid; ++ __u64 nr_items; ++ __u64 flags; ++ __u64 values[5]; ++ __u64 unused[121]; ++}; ++ ++enum btrfs_err_code { ++ BTRFS_ERROR_DEV_RAID1_MIN_NOT_MET = 1, ++ BTRFS_ERROR_DEV_RAID10_MIN_NOT_MET = 2, ++ BTRFS_ERROR_DEV_RAID5_MIN_NOT_MET = 3, ++ BTRFS_ERROR_DEV_RAID6_MIN_NOT_MET = 4, ++ BTRFS_ERROR_DEV_TGT_REPLACE = 5, ++ BTRFS_ERROR_DEV_MISSING_NOT_FOUND = 6, ++ BTRFS_ERROR_DEV_ONLY_WRITABLE = 7, ++ BTRFS_ERROR_DEV_EXCL_RUN_IN_PROGRESS = 8, ++ BTRFS_ERROR_DEV_RAID1C3_MIN_NOT_MET = 9, ++ BTRFS_ERROR_DEV_RAID1C4_MIN_NOT_MET = 10, ++}; ++ ++struct btrfs_disk_balance_args { ++ __le64 profiles; ++ union { ++ __le64 usage; ++ struct { ++ __le32 usage_min; ++ __le32 usage_max; ++ }; ++ }; ++ __le64 devid; ++ __le64 pstart; ++ __le64 pend; ++ __le64 vstart; ++ __le64 vend; ++ __le64 target; ++ __le64 flags; ++ union { ++ __le64 limit; ++ struct { ++ __le32 limit_min; ++ __le32 limit_max; ++ }; ++ }; ++ __le32 stripes_min; ++ __le32 stripes_max; ++ __le64 unused[6]; ++}; ++ ++struct btrfs_balance_item { ++ __le64 flags; ++ struct btrfs_disk_balance_args data; ++ struct btrfs_disk_balance_args meta; ++ struct btrfs_disk_balance_args sys; ++ __le64 unused[4]; ++}; ++ ++struct btrfs_dev_stats_item { ++ __le64 values[5]; ++}; ++ ++struct alloc_chunk_ctl { ++ u64 start; ++ u64 type; ++ int num_stripes; ++ int sub_stripes; ++ int dev_stripes; ++ int devs_max; ++ int devs_min; ++ int devs_increment; ++ int ncopies; ++ int nparity; ++ u64 max_stripe_size; ++ u64 max_chunk_size; ++ u64 dev_extent_min; ++ u64 stripe_size; ++ u64 chunk_size; ++ int ndevs; ++}; ++ ++struct btrfs_fid { ++ u64 objectid; ++ u64 root_objectid; ++ u32 gen; ++ u64 parent_objectid; ++ u32 parent_gen; ++ u64 parent_root_objectid; ++} __attribute__((packed)); ++ ++struct workspace { ++ z_stream strm; ++ char *buf; ++ unsigned int buf_size; ++ struct list_head list; ++ int level; ++}; ++ ++struct workspace___2 { ++ void *mem; ++ void *buf; ++ void *cbuf; ++ struct list_head list; ++}; ++ ++struct rb_simple_node { ++ struct rb_node rb_node; ++ u64 bytenr; ++}; ++ ++struct ulist_iterator { ++ struct list_head *cur_list; ++}; ++ ++struct ulist_node { ++ u64 val; ++ u64 aux; ++ struct list_head list; ++ struct rb_node rb_node; ++}; ++ ++struct btrfs_backref_node; ++ ++struct btrfs_backref_cache { ++ struct rb_root rb_root; ++ struct btrfs_backref_node *path[8]; ++ struct list_head pending[8]; ++ struct list_head leaves; ++ struct list_head changed; ++ struct list_head detached; ++ u64 last_trans; ++ int nr_nodes; ++ int nr_edges; ++ struct list_head pending_edge; ++ struct list_head useless_node; ++ struct btrfs_fs_info *fs_info; ++ unsigned int is_reloc; ++}; ++ ++struct file_extent_cluster { ++ u64 start; ++ u64 end; ++ u64 boundary[128]; ++ unsigned int nr; ++}; ++ ++struct mapping_tree { ++ struct rb_root rb_root; ++ spinlock_t lock; ++}; ++ ++struct reloc_control { ++ struct btrfs_block_group *block_group; ++ struct btrfs_root *extent_root; ++ struct inode *data_inode; ++ struct btrfs_block_rsv *block_rsv; ++ struct btrfs_backref_cache backref_cache; ++ struct file_extent_cluster cluster; ++ struct extent_io_tree processed_blocks; ++ struct mapping_tree reloc_root_tree; ++ struct list_head reloc_roots; ++ struct list_head dirty_subvol_roots; ++ u64 merging_rsv_size; ++ u64 nodes_relocated; ++ u64 reserved_bytes; ++ u64 search_start; ++ u64 extents_found; ++ unsigned int stage: 8; ++ unsigned int create_reloc_tree: 1; ++ unsigned int merge_reloc_tree: 1; ++ unsigned int found_file_extent: 1; ++}; ++ ++struct btrfs_backref_iter { ++ u64 bytenr; ++ struct btrfs_path *path; ++ struct btrfs_fs_info *fs_info; ++ struct btrfs_key cur_key; ++ u32 item_ptr; ++ u32 cur_ptr; ++ u32 end_ptr; ++}; ++ ++struct btrfs_backref_node { ++ struct { ++ struct rb_node rb_node; ++ u64 bytenr; ++ }; ++ u64 new_bytenr; ++ u64 owner; ++ struct list_head list; ++ struct list_head upper; ++ struct list_head lower; ++ struct btrfs_root *root; ++ struct extent_buffer *eb; ++ unsigned int level: 8; ++ unsigned int cowonly: 1; ++ unsigned int lowest: 1; ++ unsigned int locked: 1; ++ unsigned int processed: 1; ++ unsigned int checked: 1; ++ unsigned int pending: 1; ++ unsigned int detached: 1; ++ unsigned int is_reloc_root: 1; ++}; ++ ++struct btrfs_backref_edge { ++ struct list_head list[2]; ++ struct btrfs_backref_node *node[2]; ++}; ++ ++struct mapping_node { ++ struct { ++ struct rb_node rb_node; ++ u64 bytenr; ++ }; ++ void *data; ++}; ++ ++struct tree_block { ++ struct { ++ struct rb_node rb_node; ++ u64 bytenr; ++ }; ++ u64 owner; ++ struct btrfs_key key; ++ unsigned int level: 8; ++ unsigned int key_ready: 1; ++}; ++ ++struct btrfs_backref_shared_cache_entry { ++ u64 bytenr; ++ u64 gen; ++ bool is_shared; ++}; ++ ++struct btrfs_backref_share_check_ctx { ++ struct ulist refs; ++ u64 curr_leaf_bytenr; ++ u64 prev_leaf_bytenr; ++ struct btrfs_backref_shared_cache_entry path_cache_entries[8]; ++ bool use_path_cache; ++ struct { ++ u64 bytenr; ++ bool is_shared; ++ } prev_extents_cache[8]; ++ int prev_extents_cache_slot; ++}; ++ ++struct extent_inode_elem { ++ u64 inum; ++ u64 offset; ++ u64 num_bytes; ++ struct extent_inode_elem *next; ++}; ++ ++struct preftree { ++ struct rb_root_cached root; ++ unsigned int count; ++}; ++ ++struct preftrees { ++ struct preftree direct; ++ struct preftree indirect; ++ struct preftree indirect_missing_keys; ++}; ++ ++struct share_check { ++ struct btrfs_backref_share_check_ctx *ctx; ++ struct btrfs_root *root; ++ u64 inum; ++ u64 data_bytenr; ++ u64 data_extent_gen; ++ int share_count; ++ int self_ref_count; ++ bool have_delayed_delete_refs; ++}; ++ ++struct btrfs_stream_header { ++ char magic[13]; ++ __le32 version; ++} __attribute__((packed)); ++ ++struct btrfs_cmd_header { ++ __le32 len; ++ __le16 cmd; ++ __le32 crc; ++} __attribute__((packed)); ++ ++struct btrfs_tlv_header { ++ __le16 tlv_type; ++ __le16 tlv_len; ++}; ++ ++enum btrfs_send_cmd { ++ BTRFS_SEND_C_UNSPEC = 0, ++ BTRFS_SEND_C_SUBVOL = 1, ++ BTRFS_SEND_C_SNAPSHOT = 2, ++ BTRFS_SEND_C_MKFILE = 3, ++ BTRFS_SEND_C_MKDIR = 4, ++ BTRFS_SEND_C_MKNOD = 5, ++ BTRFS_SEND_C_MKFIFO = 6, ++ BTRFS_SEND_C_MKSOCK = 7, ++ BTRFS_SEND_C_SYMLINK = 8, ++ BTRFS_SEND_C_RENAME = 9, ++ BTRFS_SEND_C_LINK = 10, ++ BTRFS_SEND_C_UNLINK = 11, ++ BTRFS_SEND_C_RMDIR = 12, ++ BTRFS_SEND_C_SET_XATTR = 13, ++ BTRFS_SEND_C_REMOVE_XATTR = 14, ++ BTRFS_SEND_C_WRITE = 15, ++ BTRFS_SEND_C_CLONE = 16, ++ BTRFS_SEND_C_TRUNCATE = 17, ++ BTRFS_SEND_C_CHMOD = 18, ++ BTRFS_SEND_C_CHOWN = 19, ++ BTRFS_SEND_C_UTIMES = 20, ++ BTRFS_SEND_C_END = 21, ++ BTRFS_SEND_C_UPDATE_EXTENT = 22, ++ BTRFS_SEND_C_MAX_V1 = 22, ++ BTRFS_SEND_C_FALLOCATE = 23, ++ BTRFS_SEND_C_FILEATTR = 24, ++ BTRFS_SEND_C_ENCODED_WRITE = 25, ++ BTRFS_SEND_C_MAX_V2 = 25, ++ BTRFS_SEND_C_ENABLE_VERITY = 26, ++ BTRFS_SEND_C_MAX_V3 = 26, ++ BTRFS_SEND_C_MAX = 26, ++}; ++ ++enum { ++ BTRFS_SEND_A_UNSPEC = 0, ++ BTRFS_SEND_A_UUID = 1, ++ BTRFS_SEND_A_CTRANSID = 2, ++ BTRFS_SEND_A_INO = 3, ++ BTRFS_SEND_A_SIZE = 4, ++ BTRFS_SEND_A_MODE = 5, ++ BTRFS_SEND_A_UID = 6, ++ BTRFS_SEND_A_GID = 7, ++ BTRFS_SEND_A_RDEV = 8, ++ BTRFS_SEND_A_CTIME = 9, ++ BTRFS_SEND_A_MTIME = 10, ++ BTRFS_SEND_A_ATIME = 11, ++ BTRFS_SEND_A_OTIME = 12, ++ BTRFS_SEND_A_XATTR_NAME = 13, ++ BTRFS_SEND_A_XATTR_DATA = 14, ++ BTRFS_SEND_A_PATH = 15, ++ BTRFS_SEND_A_PATH_TO = 16, ++ BTRFS_SEND_A_PATH_LINK = 17, ++ BTRFS_SEND_A_FILE_OFFSET = 18, ++ BTRFS_SEND_A_DATA = 19, ++ BTRFS_SEND_A_CLONE_UUID = 20, ++ BTRFS_SEND_A_CLONE_CTRANSID = 21, ++ BTRFS_SEND_A_CLONE_PATH = 22, ++ BTRFS_SEND_A_CLONE_OFFSET = 23, ++ BTRFS_SEND_A_CLONE_LEN = 24, ++ BTRFS_SEND_A_MAX_V1 = 24, ++ BTRFS_SEND_A_FALLOCATE_MODE = 25, ++ BTRFS_SEND_A_FILEATTR = 26, ++ BTRFS_SEND_A_UNENCODED_FILE_LEN = 27, ++ BTRFS_SEND_A_UNENCODED_LEN = 28, ++ BTRFS_SEND_A_UNENCODED_OFFSET = 29, ++ BTRFS_SEND_A_COMPRESSION = 30, ++ BTRFS_SEND_A_ENCRYPTION = 31, ++ BTRFS_SEND_A_MAX_V2 = 31, ++ BTRFS_SEND_A_VERITY_ALGORITHM = 32, ++ BTRFS_SEND_A_VERITY_BLOCK_SIZE = 33, ++ BTRFS_SEND_A_VERITY_SALT_DATA = 34, ++ BTRFS_SEND_A_VERITY_SIG_DATA = 35, ++ BTRFS_SEND_A_MAX_V3 = 35, ++ __BTRFS_SEND_A_MAX = 35, ++}; ++ ++struct btrfs_ioctl_send_args { ++ __s64 send_fd; ++ __u64 clone_sources_count; ++ __u64 *clone_sources; ++ __u64 parent_root; ++ __u64 flags; ++ __u32 version; ++ __u8 reserved[28]; ++}; ++ ++struct fs_path { ++ union { ++ struct { ++ char *start; ++ char *end; ++ char *buf; ++ short unsigned int buf_len: 15; ++ short unsigned int reversed: 1; ++ char inline_buf[0]; ++ }; ++ char pad[256]; ++ }; ++}; ++ ++struct clone_root { ++ struct btrfs_root *root; ++ u64 ino; ++ u64 offset; ++ u64 num_bytes; ++ bool found_ref; ++}; ++ ++struct backref_cache_entry { ++ struct btrfs_lru_cache_entry entry; ++ u64 root_ids[17]; ++ int num_roots; ++}; ++ ++struct send_ctx { ++ struct file *send_filp; ++ loff_t send_off; ++ char *send_buf; ++ u32 send_size; ++ u32 send_max_size; ++ bool put_data; ++ struct page **send_buf_pages; ++ u64 flags; ++ u32 proto; ++ struct btrfs_root *send_root; ++ struct btrfs_root *parent_root; ++ struct clone_root *clone_roots; ++ int clone_roots_cnt; ++ struct btrfs_path *left_path; ++ struct btrfs_path *right_path; ++ struct btrfs_key *cmp_key; ++ u64 last_reloc_trans; ++ u64 cur_ino; ++ u64 cur_inode_gen; ++ u64 cur_inode_size; ++ u64 cur_inode_mode; ++ u64 cur_inode_rdev; ++ u64 cur_inode_last_extent; ++ u64 cur_inode_next_write_offset; ++ bool cur_inode_new; ++ bool cur_inode_new_gen; ++ bool cur_inode_deleted; ++ bool ignore_cur_inode; ++ bool cur_inode_needs_verity; ++ void *verity_descriptor; ++ u64 send_progress; ++ struct list_head new_refs; ++ struct list_head deleted_refs; ++ struct btrfs_lru_cache name_cache; ++ struct inode *cur_inode; ++ struct file_ra_state ra; ++ u64 page_cache_clear_start; ++ bool clean_page_cache; ++ struct rb_root pending_dir_moves; ++ struct rb_root waiting_dir_moves; ++ struct rb_root orphan_dirs; ++ struct rb_root rbtree_new_refs; ++ struct rb_root rbtree_deleted_refs; ++ struct btrfs_lru_cache backref_cache; ++ u64 backref_cache_last_reloc_trans; ++ struct btrfs_lru_cache dir_created_cache; ++ struct btrfs_lru_cache dir_utimes_cache; ++}; ++ ++struct pending_dir_move { ++ struct rb_node node; ++ struct list_head list; ++ u64 parent_ino; ++ u64 ino; ++ u64 gen; ++ struct list_head update_refs; ++}; ++ ++struct waiting_dir_move { ++ struct rb_node node; ++ u64 ino; ++ u64 rmdir_ino; ++ u64 rmdir_gen; ++ bool orphanized; ++}; ++ ++struct orphan_dir_info { ++ struct rb_node node; ++ u64 ino; ++ u64 gen; ++ u64 last_dir_index_offset; ++ u64 dir_high_seq_ino; ++}; ++ ++struct name_cache_entry { ++ struct btrfs_lru_cache_entry entry; ++ u64 parent_ino; ++ u64 parent_gen; ++ int ret; ++ int need_later_update; ++ int name_len; ++ char name[0]; ++}; ++ ++enum btrfs_compare_tree_result { ++ BTRFS_COMPARE_TREE_NEW = 0, ++ BTRFS_COMPARE_TREE_DELETED = 1, ++ BTRFS_COMPARE_TREE_CHANGED = 2, ++ BTRFS_COMPARE_TREE_SAME = 3, ++}; ++ ++struct btrfs_inode_info { ++ u64 size; ++ u64 gen; ++ u64 mode; ++ u64 uid; ++ u64 gid; ++ u64 rdev; ++ u64 fileattr; ++ u64 nlink; ++}; ++ ++typedef int (*iterate_inode_ref_t)(int, u64, int, struct fs_path *, void *); ++ ++typedef int (*iterate_dir_item_t)(int, struct btrfs_key *, const char *, int, const char *, int, void *); ++ ++struct backref_ctx { ++ struct send_ctx *sctx; ++ u64 found; ++ u64 cur_objectid; ++ u64 cur_offset; ++ u64 extent_len; ++ u64 bytenr; ++ u64 backref_owner; ++ u64 backref_offset; ++}; ++ ++enum inode_state { ++ inode_state_no_change = 0, ++ inode_state_will_create = 1, ++ inode_state_did_create = 2, ++ inode_state_will_delete = 3, ++ inode_state_did_delete = 4, ++}; ++ ++struct recorded_ref { ++ struct list_head list; ++ char *name; ++ struct fs_path *full_path; ++ u64 dir; ++ u64 dir_gen; ++ int name_len; ++ struct rb_node node; ++ struct rb_root *root; ++}; ++ ++struct find_xattr_ctx { ++ const char *name; ++ int name_len; ++ int found_idx; ++ char *found_data; ++ int found_data_len; ++}; ++ ++struct prop_handler { ++ struct hlist_node node; ++ const char *xattr_name; ++ int (*validate)(const struct btrfs_inode *, const char *, size_t); ++ int (*apply)(struct inode *, const char *, size_t); ++ const char * (*extract)(struct inode *); ++ bool (*ignore)(const struct btrfs_inode *); ++ int inheritable; ++}; ++ ++enum btrfs_subpage_type { ++ BTRFS_SUBPAGE_METADATA = 0, ++ BTRFS_SUBPAGE_DATA = 1, ++}; ++ ++enum blk_zone_type { ++ BLK_ZONE_TYPE_CONVENTIONAL = 1, ++ BLK_ZONE_TYPE_SEQWRITE_REQ = 2, ++ BLK_ZONE_TYPE_SEQWRITE_PREF = 3, ++}; ++ ++enum blk_zone_cond { ++ BLK_ZONE_COND_NOT_WP = 0, ++ BLK_ZONE_COND_EMPTY = 1, ++ BLK_ZONE_COND_IMP_OPEN = 2, ++ BLK_ZONE_COND_EXP_OPEN = 3, ++ BLK_ZONE_COND_CLOSED = 4, ++ BLK_ZONE_COND_READONLY = 13, ++ BLK_ZONE_COND_FULL = 14, ++ BLK_ZONE_COND_OFFLINE = 15, ++}; ++ ++enum pstore_type_id { ++ PSTORE_TYPE_DMESG = 0, ++ PSTORE_TYPE_MCE = 1, ++ PSTORE_TYPE_CONSOLE = 2, ++ PSTORE_TYPE_FTRACE = 3, ++ PSTORE_TYPE_PPC_RTAS = 4, ++ PSTORE_TYPE_PPC_OF = 5, ++ PSTORE_TYPE_PPC_COMMON = 6, ++ PSTORE_TYPE_PMSG = 7, ++ PSTORE_TYPE_PPC_OPAL = 8, ++ PSTORE_TYPE_MAX = 9, ++}; ++ ++struct pstore_info; ++ ++struct pstore_record { ++ struct pstore_info *psi; ++ enum pstore_type_id type; ++ u64 id; ++ struct timespec64 time; ++ char *buf; ++ ssize_t size; ++ ssize_t ecc_notice_size; ++ void *priv; ++ int count; ++ enum kmsg_dump_reason reason; ++ unsigned int part; ++ bool compressed; ++}; ++ ++struct pstore_info { ++ struct module *owner; ++ const char *name; ++ spinlock_t buf_lock; ++ char *buf; ++ size_t bufsize; ++ struct mutex read_mutex; ++ int flags; ++ int max_reason; ++ void *data; ++ int (*open)(struct pstore_info *); ++ int (*close)(struct pstore_info *); ++ ssize_t (*read)(struct pstore_record *); ++ int (*write)(struct pstore_record *); ++ int (*write_user)(struct pstore_record *, const char *); ++ int (*erase)(struct pstore_record *); ++}; ++ ++struct pstore_zbackend { ++ int (*zbufsize)(size_t); ++ const char *name; ++}; ++ ++struct msgbuf { ++ __kernel_long_t mtype; ++ char mtext[1]; ++}; ++ ++struct msg; ++ ++struct msqid_ds { ++ struct ipc_perm msg_perm; ++ struct msg *msg_first; ++ struct msg *msg_last; ++ __kernel_old_time_t msg_stime; ++ __kernel_old_time_t msg_rtime; ++ __kernel_old_time_t msg_ctime; ++ long unsigned int msg_lcbytes; ++ long unsigned int msg_lqbytes; ++ short unsigned int msg_cbytes; ++ short unsigned int msg_qnum; ++ short unsigned int msg_qbytes; ++ __kernel_ipc_pid_t msg_lspid; ++ __kernel_ipc_pid_t msg_lrpid; ++}; ++ ++struct msqid64_ds { ++ struct ipc64_perm msg_perm; ++ long int msg_stime; ++ long int msg_rtime; ++ long int msg_ctime; ++ long unsigned int msg_cbytes; ++ long unsigned int msg_qnum; ++ long unsigned int msg_qbytes; ++ __kernel_pid_t msg_lspid; ++ __kernel_pid_t msg_lrpid; ++ long unsigned int __unused4; ++ long unsigned int __unused5; ++}; ++ ++struct msginfo { ++ int msgpool; ++ int msgmap; ++ int msgmax; ++ int msgmnb; ++ int msgmni; ++ int msgssz; ++ int msgtql; ++ short unsigned int msgseg; ++}; ++ ++struct compat_msqid64_ds { ++ struct compat_ipc64_perm msg_perm; ++ compat_ulong_t msg_stime; ++ compat_ulong_t msg_stime_high; ++ compat_ulong_t msg_rtime; ++ compat_ulong_t msg_rtime_high; ++ compat_ulong_t msg_ctime; ++ compat_ulong_t msg_ctime_high; ++ compat_ulong_t msg_cbytes; ++ compat_ulong_t msg_qnum; ++ compat_ulong_t msg_qbytes; ++ compat_pid_t msg_lspid; ++ compat_pid_t msg_lrpid; ++ compat_ulong_t __unused4; ++ compat_ulong_t __unused5; ++}; ++ ++struct msg_queue { ++ struct kern_ipc_perm q_perm; ++ time64_t q_stime; ++ time64_t q_rtime; ++ time64_t q_ctime; ++ long unsigned int q_cbytes; ++ long unsigned int q_qnum; ++ long unsigned int q_qbytes; ++ struct pid *q_lspid; ++ struct pid *q_lrpid; ++ struct list_head q_messages; ++ struct list_head q_receivers; ++ struct list_head q_senders; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct msg_receiver { ++ struct list_head r_list; ++ struct task_struct *r_tsk; ++ int r_mode; ++ long int r_msgtype; ++ long int r_maxsize; ++ struct msg_msg *r_msg; ++}; ++ ++struct msg_sender { ++ struct list_head list; ++ struct task_struct *tsk; ++ size_t msgsz; ++}; ++ ++struct compat_msqid_ds { ++ struct compat_ipc_perm msg_perm; ++ compat_uptr_t msg_first; ++ compat_uptr_t msg_last; ++ old_time32_t msg_stime; ++ old_time32_t msg_rtime; ++ old_time32_t msg_ctime; ++ compat_ulong_t msg_lcbytes; ++ compat_ulong_t msg_lqbytes; ++ short unsigned int msg_cbytes; ++ short unsigned int msg_qnum; ++ short unsigned int msg_qbytes; ++ compat_ipc_pid_t msg_lspid; ++ compat_ipc_pid_t msg_lrpid; ++}; ++ ++struct compat_msgbuf { ++ compat_long_t mtype; ++ char mtext[1]; ++}; ++ ++struct ipc_kludge { ++ struct msgbuf *msgp; ++ long int msgtyp; ++}; ++ ++struct compat_ipc_kludge { ++ compat_uptr_t msgp; ++ compat_long_t msgtyp; ++}; ++ ++enum key_notification_subtype { ++ NOTIFY_KEY_INSTANTIATED = 0, ++ NOTIFY_KEY_UPDATED = 1, ++ NOTIFY_KEY_LINKED = 2, ++ NOTIFY_KEY_UNLINKED = 3, ++ NOTIFY_KEY_CLEARED = 4, ++ NOTIFY_KEY_REVOKED = 5, ++ NOTIFY_KEY_INVALIDATED = 6, ++ NOTIFY_KEY_SETATTR = 7, ++}; ++ ++struct key_notification { ++ struct watch_notification watch; ++ __u32 key_id; ++ __u32 aux; ++}; ++ ++enum { ++ Opt_err___7 = 0, ++ Opt_enc = 1, ++ Opt_hash = 2, ++}; ++ ++struct osapsess { ++ uint32_t handle; ++ unsigned char secret[20]; ++ unsigned char enonce[20]; ++}; ++ ++enum { ++ SEAL_keytype = 1, ++ SRK_keytype = 4, ++}; ++ ++struct sdesc { ++ struct shash_desc shash; ++ char ctx[0]; ++}; ++ ++struct tpm_digests { ++ unsigned char encauth[20]; ++ unsigned char pubauth[20]; ++ unsigned char xorwork[40]; ++ unsigned char xorhash[20]; ++ unsigned char nonceodd[20]; ++}; ++ ++enum { ++ Opt_err___8 = 0, ++ Opt_keyhandle = 1, ++ Opt_keyauth = 2, ++ Opt_blobauth = 3, ++ Opt_pcrinfo = 4, ++ Opt_pcrlock = 5, ++ Opt_migratable = 6, ++ Opt_hash___2 = 7, ++ Opt_policydigest = 8, ++ Opt_policyhandle = 9, ++}; ++ ++struct ecryptfs_session_key { ++ u32 flags; ++ u32 encrypted_key_size; ++ u32 decrypted_key_size; ++ u8 encrypted_key[512]; ++ u8 decrypted_key[64]; ++}; ++ ++struct ecryptfs_password { ++ u32 password_bytes; ++ s32 hash_algo; ++ u32 hash_iterations; ++ u32 session_key_encryption_key_bytes; ++ u32 flags; ++ u8 session_key_encryption_key[64]; ++ u8 signature[17]; ++ u8 salt[8]; ++}; ++ ++enum ecryptfs_token_types { ++ ECRYPTFS_PASSWORD = 0, ++ ECRYPTFS_PRIVATE_KEY = 1, ++}; ++ ++struct ecryptfs_private_key { ++ u32 key_size; ++ u32 data_len; ++ u8 signature[17]; ++ char pki_type[17]; ++ u8 data[0]; ++}; ++ ++struct ecryptfs_auth_tok { ++ u16 version; ++ u16 token_type; ++ u32 flags; ++ struct ecryptfs_session_key session_key; ++ u8 reserved[32]; ++ union { ++ struct ecryptfs_password password; ++ struct ecryptfs_private_key private_key; ++ } token; ++}; ++ ++struct vfs_cap_data { ++ __le32 magic_etc; ++ struct { ++ __le32 permitted; ++ __le32 inheritable; ++ } data[2]; ++}; ++ ++struct vfs_ns_cap_data { ++ __le32 magic_etc; ++ struct { ++ __le32 permitted; ++ __le32 inheritable; ++ } data[2]; ++ __le32 rootid; ++}; ++ ++enum nf_ip_hook_priorities { ++ NF_IP_PRI_FIRST = -2147483648, ++ NF_IP_PRI_RAW_BEFORE_DEFRAG = -450, ++ NF_IP_PRI_CONNTRACK_DEFRAG = -400, ++ NF_IP_PRI_RAW = -300, ++ NF_IP_PRI_SELINUX_FIRST = -225, ++ NF_IP_PRI_CONNTRACK = -200, ++ NF_IP_PRI_MANGLE = -150, ++ NF_IP_PRI_NAT_DST = -100, ++ NF_IP_PRI_FILTER = 0, ++ NF_IP_PRI_SECURITY = 50, ++ NF_IP_PRI_NAT_SRC = 100, ++ NF_IP_PRI_SELINUX_LAST = 225, ++ NF_IP_PRI_CONNTRACK_HELPER = 300, ++ NF_IP_PRI_CONNTRACK_CONFIRM = 2147483647, ++ NF_IP_PRI_LAST = 2147483647, ++}; ++ ++enum nf_ip6_hook_priorities { ++ NF_IP6_PRI_FIRST = -2147483648, ++ NF_IP6_PRI_RAW_BEFORE_DEFRAG = -450, ++ NF_IP6_PRI_CONNTRACK_DEFRAG = -400, ++ NF_IP6_PRI_RAW = -300, ++ NF_IP6_PRI_SELINUX_FIRST = -225, ++ NF_IP6_PRI_CONNTRACK = -200, ++ NF_IP6_PRI_MANGLE = -150, ++ NF_IP6_PRI_NAT_DST = -100, ++ NF_IP6_PRI_FILTER = 0, ++ NF_IP6_PRI_SECURITY = 50, ++ NF_IP6_PRI_NAT_SRC = 100, ++ NF_IP6_PRI_SELINUX_LAST = 225, ++ NF_IP6_PRI_CONNTRACK_HELPER = 300, ++ NF_IP6_PRI_LAST = 2147483647, ++}; ++ ++enum bpf_cmd { ++ BPF_MAP_CREATE = 0, ++ BPF_MAP_LOOKUP_ELEM = 1, ++ BPF_MAP_UPDATE_ELEM = 2, ++ BPF_MAP_DELETE_ELEM = 3, ++ BPF_MAP_GET_NEXT_KEY = 4, ++ BPF_PROG_LOAD = 5, ++ BPF_OBJ_PIN = 6, ++ BPF_OBJ_GET = 7, ++ BPF_PROG_ATTACH = 8, ++ BPF_PROG_DETACH = 9, ++ BPF_PROG_TEST_RUN = 10, ++ BPF_PROG_RUN = 10, ++ BPF_PROG_GET_NEXT_ID = 11, ++ BPF_MAP_GET_NEXT_ID = 12, ++ BPF_PROG_GET_FD_BY_ID = 13, ++ BPF_MAP_GET_FD_BY_ID = 14, ++ BPF_OBJ_GET_INFO_BY_FD = 15, ++ BPF_PROG_QUERY = 16, ++ BPF_RAW_TRACEPOINT_OPEN = 17, ++ BPF_BTF_LOAD = 18, ++ BPF_BTF_GET_FD_BY_ID = 19, ++ BPF_TASK_FD_QUERY = 20, ++ BPF_MAP_LOOKUP_AND_DELETE_ELEM = 21, ++ BPF_MAP_FREEZE = 22, ++ BPF_BTF_GET_NEXT_ID = 23, ++ BPF_MAP_LOOKUP_BATCH = 24, ++ BPF_MAP_LOOKUP_AND_DELETE_BATCH = 25, ++ BPF_MAP_UPDATE_BATCH = 26, ++ BPF_MAP_DELETE_BATCH = 27, ++ BPF_LINK_CREATE = 28, ++ BPF_LINK_UPDATE = 29, ++ BPF_LINK_GET_FD_BY_ID = 30, ++ BPF_LINK_GET_NEXT_ID = 31, ++ BPF_ENABLE_STATS = 32, ++ BPF_ITER_CREATE = 33, ++ BPF_LINK_DETACH = 34, ++ BPF_PROG_BIND_MAP = 35, ++}; ++ ++struct tty_file_private { ++ struct tty_struct *tty; ++ struct file *file; ++ struct list_head list; ++}; ++ ++struct dccp_hdr { ++ __be16 dccph_sport; ++ __be16 dccph_dport; ++ __u8 dccph_doff; ++ __u8 dccph_ccval: 4; ++ __u8 dccph_cscov: 4; ++ __sum16 dccph_checksum; ++ __u8 dccph_reserved: 3; ++ __u8 dccph_type: 4; ++ __u8 dccph_x: 1; ++ __u8 dccph_seq2; ++ __be16 dccph_seq; ++}; ++ ++enum dccp_state { ++ DCCP_OPEN = 1, ++ DCCP_REQUESTING = 2, ++ DCCP_LISTEN = 10, ++ DCCP_RESPOND = 3, ++ DCCP_ACTIVE_CLOSEREQ = 4, ++ DCCP_PASSIVE_CLOSE = 8, ++ DCCP_CLOSING = 11, ++ DCCP_TIME_WAIT = 6, ++ DCCP_CLOSED = 7, ++ DCCP_NEW_SYN_RECV = 12, ++ DCCP_PARTOPEN = 13, ++ DCCP_PASSIVE_CLOSEREQ = 14, ++ DCCP_MAX_STATES = 15, ++}; ++ ++enum sctp_msg_flags { ++ MSG_NOTIFICATION = 32768, ++}; ++ ++enum sctp_cid { ++ SCTP_CID_DATA = 0, ++ SCTP_CID_INIT = 1, ++ SCTP_CID_INIT_ACK = 2, ++ SCTP_CID_SACK = 3, ++ SCTP_CID_HEARTBEAT = 4, ++ SCTP_CID_HEARTBEAT_ACK = 5, ++ SCTP_CID_ABORT = 6, ++ SCTP_CID_SHUTDOWN = 7, ++ SCTP_CID_SHUTDOWN_ACK = 8, ++ SCTP_CID_ERROR = 9, ++ SCTP_CID_COOKIE_ECHO = 10, ++ SCTP_CID_COOKIE_ACK = 11, ++ SCTP_CID_ECN_ECNE = 12, ++ SCTP_CID_ECN_CWR = 13, ++ SCTP_CID_SHUTDOWN_COMPLETE = 14, ++ SCTP_CID_AUTH = 15, ++ SCTP_CID_I_DATA = 64, ++ SCTP_CID_FWD_TSN = 192, ++ SCTP_CID_ASCONF = 193, ++ SCTP_CID_I_FWD_TSN = 194, ++ SCTP_CID_ASCONF_ACK = 128, ++ SCTP_CID_RECONF = 130, ++ SCTP_CID_PAD = 132, ++}; ++ ++enum sctp_param { ++ SCTP_PARAM_HEARTBEAT_INFO = 1, ++ SCTP_PARAM_IPV4_ADDRESS = 5, ++ SCTP_PARAM_IPV6_ADDRESS = 6, ++ SCTP_PARAM_STATE_COOKIE = 7, ++ SCTP_PARAM_UNRECOGNIZED_PARAMETERS = 8, ++ SCTP_PARAM_COOKIE_PRESERVATIVE = 9, ++ SCTP_PARAM_HOST_NAME_ADDRESS = 11, ++ SCTP_PARAM_SUPPORTED_ADDRESS_TYPES = 12, ++ SCTP_PARAM_ECN_CAPABLE = 32768, ++ SCTP_PARAM_RANDOM = 32770, ++ SCTP_PARAM_CHUNKS = 32771, ++ SCTP_PARAM_HMAC_ALGO = 32772, ++ SCTP_PARAM_SUPPORTED_EXT = 32776, ++ SCTP_PARAM_FWD_TSN_SUPPORT = 49152, ++ SCTP_PARAM_ADD_IP = 49153, ++ SCTP_PARAM_DEL_IP = 49154, ++ SCTP_PARAM_ERR_CAUSE = 49155, ++ SCTP_PARAM_SET_PRIMARY = 49156, ++ SCTP_PARAM_SUCCESS_REPORT = 49157, ++ SCTP_PARAM_ADAPTATION_LAYER_IND = 49158, ++ SCTP_PARAM_RESET_OUT_REQUEST = 13, ++ SCTP_PARAM_RESET_IN_REQUEST = 14, ++ SCTP_PARAM_RESET_TSN_REQUEST = 15, ++ SCTP_PARAM_RESET_RESPONSE = 16, ++ SCTP_PARAM_RESET_ADD_OUT_STREAMS = 17, ++ SCTP_PARAM_RESET_ADD_IN_STREAMS = 18, ++}; ++ ++enum { ++ SCTP_MAX_STREAM = 65535, ++}; ++ ++enum sctp_event_timeout { ++ SCTP_EVENT_TIMEOUT_NONE = 0, ++ SCTP_EVENT_TIMEOUT_T1_COOKIE = 1, ++ SCTP_EVENT_TIMEOUT_T1_INIT = 2, ++ SCTP_EVENT_TIMEOUT_T2_SHUTDOWN = 3, ++ SCTP_EVENT_TIMEOUT_T3_RTX = 4, ++ SCTP_EVENT_TIMEOUT_T4_RTO = 5, ++ SCTP_EVENT_TIMEOUT_T5_SHUTDOWN_GUARD = 6, ++ SCTP_EVENT_TIMEOUT_HEARTBEAT = 7, ++ SCTP_EVENT_TIMEOUT_RECONF = 8, ++ SCTP_EVENT_TIMEOUT_PROBE = 9, ++ SCTP_EVENT_TIMEOUT_SACK = 10, ++ SCTP_EVENT_TIMEOUT_AUTOCLOSE = 11, ++}; ++ ++enum { ++ SCTP_MAX_DUP_TSNS = 16, ++}; ++ ++enum { ++ SCTP_AUTH_HMAC_ID_RESERVED_0 = 0, ++ SCTP_AUTH_HMAC_ID_SHA1 = 1, ++ SCTP_AUTH_HMAC_ID_RESERVED_2 = 2, ++ SCTP_AUTH_HMAC_ID_SHA256 = 3, ++ __SCTP_AUTH_HMAC_MAX = 4, ++}; ++ ++struct file_security_struct { ++ u32 sid; ++ u32 fown_sid; ++ u32 isid; ++ u32 pseqno; ++}; ++ ++struct superblock_security_struct { ++ u32 sid; ++ u32 def_sid; ++ u32 mntpoint_sid; ++ short unsigned int behavior; ++ short unsigned int flags; ++ struct mutex lock; ++ struct list_head isec_head; ++ spinlock_t isec_lock; ++}; ++ ++struct msg_security_struct { ++ u32 sid; ++}; ++ ++struct ipc_security_struct { ++ u16 sclass; ++ u32 sid; ++}; ++ ++struct tun_security_struct { ++ u32 sid; ++}; ++ ++struct key_security_struct { ++ u32 sid; ++}; ++ ++struct ib_security_struct { ++ u32 sid; ++}; ++ ++struct bpf_security_struct { ++ u32 sid; ++}; ++ ++struct perf_event_security_struct { ++ u32 sid; ++}; ++ ++struct security_class_mapping { ++ const char *name; ++ const char *perms[33]; ++}; ++ ++struct selinux_mnt_opts { ++ u32 fscontext_sid; ++ u32 context_sid; ++ u32 rootcontext_sid; ++ u32 defcontext_sid; ++}; ++ ++enum { ++ Opt_error = -1, ++ Opt_context = 0, ++ Opt_defcontext = 1, ++ Opt_fscontext = 2, ++ Opt_rootcontext = 3, ++ Opt_seclabel = 4, ++}; ++ ++struct hashtab_node { ++ void *key; ++ void *datum; ++ struct hashtab_node *next; ++}; ++ ++struct hashtab { ++ struct hashtab_node **htable; ++ u32 size; ++ u32 nel; ++}; ++ ++struct hashtab_info { ++ u32 slots_used; ++ u32 max_chain_len; ++}; ++ ++struct hashtab_key_params { ++ u32 (*hash)(const void *); ++ int (*cmp)(const void *, const void *); ++}; ++ ++struct symtab { ++ struct hashtab table; ++ u32 nprim; ++}; ++ ++struct extended_perms_data { ++ u32 p[8]; ++}; ++ ++struct avtab_key { ++ u16 source_type; ++ u16 target_type; ++ u16 target_class; ++ u16 specified; ++}; ++ ++struct avtab_extended_perms { ++ u8 specified; ++ u8 driver; ++ struct extended_perms_data perms; ++}; ++ ++struct avtab_datum { ++ union { ++ u32 data; ++ struct avtab_extended_perms *xperms; ++ } u; ++}; ++ ++struct avtab_node { ++ struct avtab_key key; ++ struct avtab_datum datum; ++ struct avtab_node *next; ++}; ++ ++struct avtab { ++ struct avtab_node **htable; ++ u32 nel; ++ u32 nslot; ++ u32 mask; ++}; ++ ++struct ebitmap_node { ++ struct ebitmap_node *next; ++ long unsigned int maps[6]; ++ u32 startbit; ++}; ++ ++struct ebitmap { ++ struct ebitmap_node *node; ++ u32 highbit; ++}; ++ ++struct mls_level { ++ u32 sens; ++ struct ebitmap cat; ++}; ++ ++struct mls_range { ++ struct mls_level level[2]; ++}; ++ ++struct context { ++ u32 user; ++ u32 role; ++ u32 type; ++ u32 len; ++ struct mls_range range; ++ char *str; ++}; ++ ++struct type_set; ++ ++struct constraint_expr { ++ u32 expr_type; ++ u32 attr; ++ u32 op; ++ struct ebitmap names; ++ struct type_set *type_names; ++ struct constraint_expr *next; ++}; ++ ++struct type_set { ++ struct ebitmap types; ++ struct ebitmap negset; ++ u32 flags; ++}; ++ ++struct constraint_node { ++ u32 permissions; ++ struct constraint_expr *expr; ++ struct constraint_node *next; ++}; ++ ++struct common_datum { ++ u32 value; ++ struct symtab permissions; ++}; ++ ++struct class_datum { ++ u32 value; ++ char *comkey; ++ struct common_datum *comdatum; ++ struct symtab permissions; ++ struct constraint_node *constraints; ++ struct constraint_node *validatetrans; ++ char default_user; ++ char default_role; ++ char default_type; ++ char default_range; ++}; ++ ++struct role_datum { ++ u32 value; ++ u32 bounds; ++ struct ebitmap dominates; ++ struct ebitmap types; ++}; ++ ++struct role_allow { ++ u32 role; ++ u32 new_role; ++ struct role_allow *next; ++}; ++ ++struct type_datum { ++ u32 value; ++ u32 bounds; ++ unsigned char primary; ++ unsigned char attribute; ++}; ++ ++struct user_datum { ++ u32 value; ++ u32 bounds; ++ struct ebitmap roles; ++ struct mls_range range; ++ struct mls_level dfltlevel; ++}; ++ ++struct cond_bool_datum { ++ __u32 value; ++ int state; ++}; ++ ++struct ocontext { ++ union { ++ char *name; ++ struct { ++ u8 protocol; ++ u16 low_port; ++ u16 high_port; ++ } port; ++ struct { ++ u32 addr; ++ u32 mask; ++ } node; ++ struct { ++ u32 addr[4]; ++ u32 mask[4]; ++ } node6; ++ struct { ++ u64 subnet_prefix; ++ u16 low_pkey; ++ u16 high_pkey; ++ } ibpkey; ++ struct { ++ char *dev_name; ++ u8 port; ++ } ibendport; ++ } u; ++ union { ++ u32 sclass; ++ u32 behavior; ++ } v; ++ struct context context[2]; ++ u32 sid[2]; ++ struct ocontext *next; ++}; ++ ++struct genfs { ++ char *fstype; ++ struct ocontext *head; ++ struct genfs *next; ++}; ++ ++struct cond_node; ++ ++struct policydb { ++ int mls_enabled; ++ struct symtab symtab[8]; ++ char **sym_val_to_name[8]; ++ struct class_datum **class_val_to_struct; ++ struct role_datum **role_val_to_struct; ++ struct user_datum **user_val_to_struct; ++ struct type_datum **type_val_to_struct; ++ struct avtab te_avtab; ++ struct hashtab role_tr; ++ struct ebitmap filename_trans_ttypes; ++ struct hashtab filename_trans; ++ u32 compat_filename_trans_count; ++ struct cond_bool_datum **bool_val_to_struct; ++ struct avtab te_cond_avtab; ++ struct cond_node *cond_list; ++ u32 cond_list_len; ++ struct role_allow *role_allow; ++ struct ocontext *ocontexts[9]; ++ struct genfs *genfs; ++ struct hashtab range_tr; ++ struct ebitmap *type_attr_map_array; ++ struct ebitmap policycaps; ++ struct ebitmap permissive_map; ++ size_t len; ++ unsigned int policyvers; ++ unsigned int reject_unknown: 1; ++ unsigned int allow_unknown: 1; ++ u16 process_class; ++ u32 process_trans_perms; ++}; ++ ++struct policy_file { ++ char *data; ++ size_t len; ++}; ++ ++struct selinux_mapping; ++ ++struct selinux_map { ++ struct selinux_mapping *mapping; ++ u16 size; ++}; ++ ++struct sidtab; ++ ++struct selinux_policy { ++ struct sidtab *sidtab; ++ struct policydb policydb; ++ struct selinux_map map; ++ u32 latest_granting; ++}; ++ ++struct convert_context_args { ++ struct policydb *oldp; ++ struct policydb *newp; ++}; ++ ++struct sidtab_convert_params { ++ struct convert_context_args *args; ++ struct sidtab *target; ++}; ++ ++struct selinux_policy_convert_data { ++ struct convert_context_args args; ++ struct sidtab_convert_params sidtab_params; ++}; ++ ++struct extended_perms_decision { ++ u8 used; ++ u8 driver; ++ struct extended_perms_data *allowed; ++ struct extended_perms_data *auditallow; ++ struct extended_perms_data *dontaudit; ++}; ++ ++struct extended_perms { ++ u16 len; ++ struct extended_perms_data drivers; ++}; ++ ++struct sidtab_str_cache; ++ ++struct sidtab_entry { ++ u32 sid; ++ u32 hash; ++ struct context context; ++ struct sidtab_str_cache *cache; ++ struct hlist_node list; ++}; ++ ++struct sidtab_node_inner; ++ ++struct sidtab_node_leaf; ++ ++union sidtab_entry_inner { ++ struct sidtab_node_inner *ptr_inner; ++ struct sidtab_node_leaf *ptr_leaf; ++}; ++ ++struct sidtab_node_inner { ++ union sidtab_entry_inner entries[512]; ++}; ++ ++struct sidtab_node_leaf { ++ struct sidtab_entry entries[39]; ++}; ++ ++struct sidtab_isid_entry { ++ int set; ++ struct sidtab_entry entry; ++}; ++ ++struct sidtab { ++ union sidtab_entry_inner roots[4]; ++ u32 count; ++ struct sidtab_convert_params *convert; ++ bool frozen; ++ spinlock_t lock; ++ u32 cache_free_slots; ++ struct list_head cache_lru_list; ++ spinlock_t cache_lock; ++ struct sidtab_isid_entry isids[27]; ++ struct hlist_head context_to_sid[512]; ++}; ++ ++struct perm_datum { ++ u32 value; ++}; ++ ++struct role_trans_key { ++ u32 role; ++ u32 type; ++ u32 tclass; ++}; ++ ++struct role_trans_datum { ++ u32 new_role; ++}; ++ ++struct filename_trans_key { ++ u32 ttype; ++ u16 tclass; ++ const char *name; ++}; ++ ++struct filename_trans_datum { ++ struct ebitmap stypes; ++ u32 otype; ++ struct filename_trans_datum *next; ++}; ++ ++struct cond_expr_node; ++ ++struct cond_expr { ++ struct cond_expr_node *nodes; ++ u32 len; ++}; ++ ++struct cond_av_list { ++ struct avtab_node **nodes; ++ u32 len; ++}; ++ ++struct cond_node { ++ int cur_state; ++ struct cond_expr expr; ++ struct cond_av_list true_list; ++ struct cond_av_list false_list; ++}; ++ ++struct selinux_mapping { ++ u16 value; ++ unsigned int num_perms; ++ u32 perms[32]; ++}; ++ ++struct cond_expr_node { ++ u32 expr_type; ++ u32 bool; ++}; ++ ++struct selinux_audit_rule { ++ u32 au_seqno; ++ struct context au_ctxt; ++}; ++ ++struct landlock_ruleset_attr { ++ __u64 handled_access_fs; ++}; ++ ++enum landlock_rule_type { ++ LANDLOCK_RULE_PATH_BENEATH = 1, ++}; ++ ++struct landlock_path_beneath_attr { ++ __u64 allowed_access; ++ __s32 parent_fd; ++} __attribute__((packed)); ++ ++struct ima_max_digest_data { ++ struct ima_digest_data hdr; ++ u8 digest[64]; ++}; ++ ++enum lsm_rule_types { ++ LSM_OBJ_USER = 0, ++ LSM_OBJ_ROLE = 1, ++ LSM_OBJ_TYPE = 2, ++ LSM_SUBJ_USER = 3, ++ LSM_SUBJ_ROLE = 4, ++ LSM_SUBJ_TYPE = 5, ++}; ++ ++enum policy_types { ++ ORIGINAL_TCB = 1, ++ DEFAULT_TCB = 2, ++}; ++ ++enum policy_rule_list { ++ IMA_DEFAULT_POLICY = 1, ++ IMA_CUSTOM_POLICY = 2, ++}; ++ ++struct ima_rule_opt_list { ++ size_t count; ++ char *items[0]; ++}; ++ ++struct ima_rule_entry { ++ struct list_head list; ++ int action; ++ unsigned int flags; ++ enum ima_hooks func; ++ int mask; ++ long unsigned int fsmagic; ++ uuid_t fsuuid; ++ kuid_t uid; ++ kgid_t gid; ++ kuid_t fowner; ++ kgid_t fgroup; ++ bool (*uid_op)(kuid_t, kuid_t); ++ bool (*gid_op)(kgid_t, kgid_t); ++ bool (*fowner_op)(vfsuid_t, kuid_t); ++ bool (*fgroup_op)(vfsgid_t, kgid_t); ++ int pcr; ++ unsigned int allowed_algos; ++ struct { ++ void *rule; ++ char *args_p; ++ int type; ++ } lsm[6]; ++ char *fsname; ++ struct ima_rule_opt_list *keyrings; ++ struct ima_rule_opt_list *label; ++ struct ima_template_desc *template; ++}; ++ ++enum policy_opt { ++ Opt_measure = 0, ++ Opt_dont_measure = 1, ++ Opt_appraise = 2, ++ Opt_dont_appraise = 3, ++ Opt_audit = 4, ++ Opt_hash___3 = 5, ++ Opt_dont_hash = 6, ++ Opt_obj_user = 7, ++ Opt_obj_role = 8, ++ Opt_obj_type = 9, ++ Opt_subj_user = 10, ++ Opt_subj_role = 11, ++ Opt_subj_type = 12, ++ Opt_func = 13, ++ Opt_mask = 14, ++ Opt_fsmagic = 15, ++ Opt_fsname = 16, ++ Opt_fsuuid = 17, ++ Opt_uid_eq = 18, ++ Opt_euid_eq = 19, ++ Opt_gid_eq = 20, ++ Opt_egid_eq = 21, ++ Opt_fowner_eq = 22, ++ Opt_fgroup_eq = 23, ++ Opt_uid_gt = 24, ++ Opt_euid_gt = 25, ++ Opt_gid_gt = 26, ++ Opt_egid_gt = 27, ++ Opt_fowner_gt = 28, ++ Opt_fgroup_gt = 29, ++ Opt_uid_lt = 30, ++ Opt_euid_lt = 31, ++ Opt_gid_lt = 32, ++ Opt_egid_lt = 33, ++ Opt_fowner_lt = 34, ++ Opt_fgroup_lt = 35, ++ Opt_digest_type = 36, ++ Opt_appraise_type = 37, ++ Opt_appraise_flag = 38, ++ Opt_appraise_algos = 39, ++ Opt_permit_directio = 40, ++ Opt_pcr = 41, ++ Opt_template = 42, ++ Opt_keyrings = 43, ++ Opt_label = 44, ++ Opt_err___9 = 45, ++}; ++ ++enum { ++ mask_exec = 0, ++ mask_write = 1, ++ mask_read = 2, ++ mask_append = 3, ++}; ++ ++struct ima_file_id { ++ __u8 hash_type; ++ __u8 hash_algorithm; ++ __u8 hash[64]; ++}; ++ ++struct evm_xattr { ++ struct evm_ima_xattr_data data; ++ u8 digest[20]; ++}; ++ ++struct crypto_istat_aead { ++ atomic64_t encrypt_cnt; ++ atomic64_t encrypt_tlen; ++ atomic64_t decrypt_cnt; ++ atomic64_t decrypt_tlen; ++ atomic64_t err_cnt; ++}; ++ ++struct crypto_report_aead { ++ char type[64]; ++ char geniv[64]; ++ unsigned int blocksize; ++ unsigned int maxauthsize; ++ unsigned int ivsize; ++}; ++ ++struct crypto_istat_cipher { ++ atomic64_t encrypt_cnt; ++ atomic64_t encrypt_tlen; ++ atomic64_t decrypt_cnt; ++ atomic64_t decrypt_tlen; ++ atomic64_t err_cnt; ++}; ++ ++struct crypto_report_blkcipher { ++ char type[64]; ++ char geniv[64]; ++ unsigned int blocksize; ++ unsigned int min_keysize; ++ unsigned int max_keysize; ++ unsigned int ivsize; ++}; ++ ++enum { ++ SKCIPHER_WALK_PHYS = 1, ++ SKCIPHER_WALK_SLOW = 2, ++ SKCIPHER_WALK_COPY = 4, ++ SKCIPHER_WALK_DIFF = 8, ++ SKCIPHER_WALK_SLEEP = 16, ++}; ++ ++struct skcipher_walk_buffer { ++ struct list_head entry; ++ struct scatter_walk dst; ++ unsigned int len; ++ u8 *data; ++ u8 buffer[0]; ++}; ++ ++struct crypto_report_hash { ++ char type[64]; ++ unsigned int blocksize; ++ unsigned int digestsize; ++}; ++ ++struct crypto_istat_hash { ++ atomic64_t hash_cnt; ++ atomic64_t hash_tlen; ++ atomic64_t err_cnt; ++}; ++ ++struct ahash_alg { ++ int (*init)(struct ahash_request *); ++ int (*update)(struct ahash_request *); ++ int (*final)(struct ahash_request *); ++ int (*finup)(struct ahash_request *); ++ int (*digest)(struct ahash_request *); ++ int (*export)(struct ahash_request *, void *); ++ int (*import)(struct ahash_request *, const void *); ++ int (*setkey)(struct crypto_ahash *, const u8 *, unsigned int); ++ int (*init_tfm)(struct crypto_ahash *); ++ void (*exit_tfm)(struct crypto_ahash *); ++ int (*clone_tfm)(struct crypto_ahash *, struct crypto_ahash *); ++ struct hash_alg_common halg; ++}; ++ ++struct crypto_hash_walk { ++ char *data; ++ unsigned int offset; ++ unsigned int alignmask; ++ struct page *pg; ++ unsigned int entrylen; ++ unsigned int total; ++ struct scatterlist *sg; ++ unsigned int flags; ++}; ++ ++struct ahash_instance { ++ void (*free)(struct ahash_instance *); ++ union { ++ struct { ++ char head[96]; ++ struct crypto_instance base; ++ } s; ++ struct ahash_alg alg; ++ }; ++}; ++ ++struct crypto_ahash_spawn { ++ struct crypto_spawn base; ++}; ++ ++struct crypto_istat_kpp { ++ atomic64_t setsecret_cnt; ++ atomic64_t generate_public_key_cnt; ++ atomic64_t compute_shared_secret_cnt; ++ atomic64_t err_cnt; ++}; ++ ++struct crypto_report_kpp { ++ char type[64]; ++}; ++ ++enum rsapubkey_actions { ++ ACT_rsa_get_e = 0, ++ ACT_rsa_get_n = 1, ++ NR__rsapubkey_actions = 2, ++}; ++ ++enum rsaprivkey_actions { ++ ACT_rsa_get_d = 0, ++ ACT_rsa_get_dp = 1, ++ ACT_rsa_get_dq = 2, ++ ACT_rsa_get_e___2 = 3, ++ ACT_rsa_get_n___2 = 4, ++ ACT_rsa_get_p = 5, ++ ACT_rsa_get_q = 6, ++ ACT_rsa_get_qinv = 7, ++ NR__rsaprivkey_actions = 8, ++}; ++ ++struct rsa_key { ++ const u8 *n; ++ const u8 *e; ++ const u8 *d; ++ const u8 *p; ++ const u8 *q; ++ const u8 *dp; ++ const u8 *dq; ++ const u8 *qinv; ++ size_t n_sz; ++ size_t e_sz; ++ size_t d_sz; ++ size_t p_sz; ++ size_t q_sz; ++ size_t dp_sz; ++ size_t dq_sz; ++ size_t qinv_sz; ++}; ++ ++struct rsa_asn1_template { ++ const char *name; ++ const u8 *data; ++ size_t size; ++}; ++ ++struct pkcs1pad_ctx { ++ struct crypto_akcipher *child; ++ unsigned int key_size; ++}; ++ ++struct pkcs1pad_inst_ctx { ++ struct crypto_akcipher_spawn spawn; ++ const struct rsa_asn1_template *digest_info; ++}; ++ ++struct pkcs1pad_request { ++ struct scatterlist in_sg[2]; ++ struct scatterlist out_sg[1]; ++ uint8_t *in_buf; ++ uint8_t *out_buf; ++ struct akcipher_request child_req; ++}; ++ ++struct ecc_point { ++ u64 *x; ++ u64 *y; ++ u8 ndigits; ++}; ++ ++struct ecc_curve { ++ char *name; ++ struct ecc_point g; ++ u64 *p; ++ u64 *n; ++ u64 *a; ++ u64 *b; ++}; ++ ++struct ecc_ctx { ++ unsigned int curve_id; ++ const struct ecc_curve *curve; ++ bool pub_key_set; ++ u64 x[8]; ++ u64 y[8]; ++ struct ecc_point pub_key; ++}; ++ ++struct ecdsa_signature_ctx { ++ const struct ecc_curve *curve; ++ u64 r[8]; ++ u64 s[8]; ++}; ++ ++struct crypto_attr_alg { ++ char name[128]; ++}; ++ ++struct crypto_attr_type { ++ u32 type; ++ u32 mask; ++}; ++ ++struct crypto_larval { ++ struct crypto_alg alg; ++ struct crypto_alg *adult; ++ struct completion completion; ++ u32 mask; ++ bool test_started; ++}; ++ ++enum { ++ CRYPTOA_UNSPEC = 0, ++ CRYPTOA_ALG = 1, ++ CRYPTOA_TYPE = 2, ++ __CRYPTOA_MAX = 3, ++}; ++ ++struct cryptomgr_param { ++ struct rtattr *tb[34]; ++ struct { ++ struct rtattr attr; ++ struct crypto_attr_type data; ++ } type; ++ struct { ++ struct rtattr attr; ++ struct crypto_attr_alg data; ++ } attrs[32]; ++ char template[128]; ++ struct crypto_larval *larval; ++ u32 otype; ++ u32 omask; ++}; ++ ++struct crypto_test_param { ++ char driver[128]; ++ char alg[128]; ++ u32 type; ++}; ++ ++struct crypto_shash_spawn { ++ struct crypto_spawn base; ++}; ++ ++struct hmac_ctx { ++ struct crypto_shash *hash; ++}; ++ ++struct md5_state { ++ u32 hash[4]; ++ u32 block[16]; ++ u64 byte_count; ++}; ++ ++struct sha256_state { ++ u32 state[8]; ++ u64 count; ++ u8 buf[64]; ++}; ++ ++struct sha3_state { ++ u64 st[25]; ++ unsigned int rsiz; ++ unsigned int rsizw; ++ unsigned int partial; ++ u8 buf[144]; ++}; ++ ++struct xts_tfm_ctx { ++ struct crypto_skcipher *child; ++ struct crypto_cipher *tweak; ++}; ++ ++struct xts_instance_ctx { ++ struct crypto_skcipher_spawn spawn; ++ char name[128]; ++}; ++ ++struct xts_request_ctx { ++ le128 t; ++ struct scatterlist *tail; ++ struct scatterlist sg[2]; ++ struct skcipher_request subreq; ++}; ++ ++struct gcm_instance_ctx { ++ struct crypto_skcipher_spawn ctr; ++ struct crypto_ahash_spawn ghash; ++}; ++ ++struct crypto_gcm_ctx { ++ struct crypto_skcipher *ctr; ++ struct crypto_ahash *ghash; ++}; ++ ++struct crypto_rfc4106_ctx { ++ struct crypto_aead *child; ++ u8 nonce[4]; ++}; ++ ++struct crypto_rfc4106_req_ctx { ++ struct scatterlist src[3]; ++ struct scatterlist dst[3]; ++ struct aead_request subreq; ++}; ++ ++struct crypto_rfc4543_instance_ctx { ++ struct crypto_aead_spawn aead; ++}; ++ ++struct crypto_rfc4543_ctx { ++ struct crypto_aead *child; ++ struct crypto_sync_skcipher *null; ++ u8 nonce[4]; ++}; ++ ++struct crypto_rfc4543_req_ctx { ++ struct aead_request subreq; ++}; ++ ++struct crypto_gcm_ghash_ctx { ++ unsigned int cryptlen; ++ struct scatterlist *src; ++ int (*complete)(struct aead_request *, u32); ++}; ++ ++struct crypto_gcm_req_priv_ctx { ++ u8 iv[16]; ++ u8 auth_tag[16]; ++ u8 iauth_tag[16]; ++ struct scatterlist src[3]; ++ struct scatterlist dst[3]; ++ struct scatterlist sg; ++ struct crypto_gcm_ghash_ctx ghash_ctx; ++ union { ++ struct ahash_request ahreq; ++ struct skcipher_request skreq; ++ } u; ++}; ++ ++struct crypto_aes_ctx { ++ u32 key_enc[60]; ++ u32 key_dec[60]; ++ u32 key_length; ++}; ++ ++struct chksum_desc_ctx { ++ __u16 crc; ++}; ++ ++enum { ++ CRYPTO_AUTHENC_KEYA_UNSPEC = 0, ++ CRYPTO_AUTHENC_KEYA_PARAM = 1, ++}; ++ ++struct crypto_authenc_key_param { ++ __be32 enckeylen; ++}; ++ ++struct crypto_authenc_keys { ++ const u8 *authkey; ++ const u8 *enckey; ++ unsigned int authkeylen; ++ unsigned int enckeylen; ++}; ++ ++struct authenc_instance_ctx { ++ struct crypto_ahash_spawn auth; ++ struct crypto_skcipher_spawn enc; ++ unsigned int reqoff; ++}; ++ ++struct crypto_authenc_ctx { ++ struct crypto_ahash *auth; ++ struct crypto_skcipher *enc; ++ struct crypto_sync_skcipher *null; ++}; ++ ++struct authenc_request_ctx { ++ struct scatterlist src[2]; ++ struct scatterlist dst[2]; ++ char tail[0]; ++}; ++ ++struct lzorle_ctx { ++ void *lzorle_comp_mem; ++}; ++ ++struct crypto842_ctx { ++ void *wmem; ++}; ++ ++struct drbg_string { ++ const unsigned char *buf; ++ size_t len; ++ struct list_head list; ++}; ++ ++typedef uint32_t drbg_flag_t; ++ ++struct drbg_core { ++ drbg_flag_t flags; ++ __u8 statelen; ++ __u8 blocklen_bytes; ++ char cra_name[128]; ++ char backend_cra_name[128]; ++}; ++ ++struct drbg_state; ++ ++struct drbg_state_ops { ++ int (*update)(struct drbg_state *, struct list_head *, int); ++ int (*generate)(struct drbg_state *, unsigned char *, unsigned int, struct list_head *); ++ int (*crypto_init)(struct drbg_state *); ++ int (*crypto_fini)(struct drbg_state *); ++}; ++ ++enum drbg_seed_state { ++ DRBG_SEED_STATE_UNSEEDED = 0, ++ DRBG_SEED_STATE_PARTIAL = 1, ++ DRBG_SEED_STATE_FULL = 2, ++}; ++ ++struct drbg_state { ++ struct mutex drbg_mutex; ++ unsigned char *V; ++ unsigned char *Vbuf; ++ unsigned char *C; ++ unsigned char *Cbuf; ++ size_t reseed_ctr; ++ size_t reseed_threshold; ++ unsigned char *scratchpad; ++ unsigned char *scratchpadbuf; ++ void *priv_data; ++ struct crypto_skcipher *ctr_handle; ++ struct skcipher_request *ctr_req; ++ __u8 *outscratchpadbuf; ++ __u8 *outscratchpad; ++ struct crypto_wait ctr_wait; ++ struct scatterlist sg_in; ++ struct scatterlist sg_out; ++ enum drbg_seed_state seeded; ++ long unsigned int last_seed_time; ++ bool pr; ++ bool fips_primed; ++ unsigned char *prev; ++ struct crypto_rng *jent; ++ const struct drbg_state_ops *d_ops; ++ const struct drbg_core *core; ++ struct drbg_string test_data; ++}; ++ ++enum drbg_prefixes { ++ DRBG_PREFIX0 = 0, ++ DRBG_PREFIX1 = 1, ++ DRBG_PREFIX2 = 2, ++ DRBG_PREFIX3 = 3, ++}; ++ ++struct s { ++ __be32 conv; ++}; ++ ++struct ghash_ctx { ++ struct gf128mul_4k *gf128; ++}; ++ ++struct ghash_desc_ctx { ++ u8 buffer[16]; ++ u32 bytes; ++}; ++ ++struct rng_ctx { ++ unsigned int len; ++ struct crypto_rng *drng; ++ u8 *addtl; ++ size_t addtl_len; ++}; ++ ++struct rng_parent_ctx { ++ struct crypto_rng *drng; ++ u8 *entropy; ++}; ++ ++struct ecdh { ++ char *key; ++ short unsigned int key_size; ++}; ++ ++struct ecdh_ctx { ++ unsigned int curve_id; ++ unsigned int ndigits; ++ u64 private_key[8]; ++}; ++ ++enum { ++ CRYPTO_KPP_SECRET_TYPE_UNKNOWN = 0, ++ CRYPTO_KPP_SECRET_TYPE_DH = 1, ++ CRYPTO_KPP_SECRET_TYPE_ECDH = 2, ++}; ++ ++struct kpp_secret { ++ short unsigned int type; ++ short unsigned int len; ++}; ++ ++struct xor_block_template { ++ struct xor_block_template *next; ++ const char *name; ++ int speed; ++ void (*do_2)(long unsigned int, long unsigned int * restrict, const long unsigned int * restrict); ++ void (*do_3)(long unsigned int, long unsigned int * restrict, const long unsigned int * restrict, const long unsigned int * restrict); ++ void (*do_4)(long unsigned int, long unsigned int * restrict, const long unsigned int * restrict, const long unsigned int * restrict, const long unsigned int * restrict); ++ void (*do_5)(long unsigned int, long unsigned int * restrict, const long unsigned int * restrict, const long unsigned int * restrict, const long unsigned int * restrict, const long unsigned int * restrict); ++}; ++ ++struct asymmetric_key_ids { ++ void *id[3]; ++}; ++ ++struct asymmetric_key_subtype { ++ struct module *owner; ++ const char *name; ++ short unsigned int name_len; ++ void (*describe)(const struct key *, struct seq_file *); ++ void (*destroy)(void *, void *); ++ int (*query)(const struct kernel_pkey_params *, struct kernel_pkey_query *); ++ int (*eds_op)(struct kernel_pkey_params *, const void *, void *); ++ int (*verify_signature)(const struct key *, const struct public_key_signature *); ++}; ++ ++struct asymmetric_key_parser { ++ struct list_head link; ++ struct module *owner; ++ const char *name; ++ int (*parse)(struct key_preparsed_payload *); ++}; ++ ++enum mscode_actions { ++ ACT_mscode_note_content_type = 0, ++ ACT_mscode_note_digest = 1, ++ ACT_mscode_note_digest_algo = 2, ++ NR__mscode_actions = 3, ++}; ++ ++struct kdf_testvec { ++ unsigned char *key; ++ size_t keylen; ++ unsigned char *ikm; ++ size_t ikmlen; ++ struct kvec info; ++ unsigned char *expected; ++ size_t expectedlen; ++}; ++ ++enum { ++ DIO_SHOULD_DIRTY = 1, ++ DIO_IS_SYNC = 2, ++}; ++ ++struct blkdev_dio { ++ union { ++ struct kiocb *iocb; ++ struct task_struct *waiter; ++ }; ++ size_t size; ++ atomic_t ref; ++ unsigned int flags; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct bio bio; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct blk_plug_cb; ++ ++typedef void (*blk_plug_cb_fn)(struct blk_plug_cb *, bool); ++ ++struct blk_plug_cb { ++ struct list_head list; ++ blk_plug_cb_fn callback; ++ void *data; ++}; ++ ++struct trace_event_raw_block_buffer { ++ struct trace_entry ent; ++ dev_t dev; ++ sector_t sector; ++ size_t size; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_block_rq_requeue { ++ struct trace_entry ent; ++ dev_t dev; ++ sector_t sector; ++ unsigned int nr_sector; ++ char rwbs[8]; ++ u32 __data_loc_cmd; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_block_rq_completion { ++ struct trace_entry ent; ++ dev_t dev; ++ sector_t sector; ++ unsigned int nr_sector; ++ int error; ++ char rwbs[8]; ++ u32 __data_loc_cmd; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_block_rq { ++ struct trace_entry ent; ++ dev_t dev; ++ sector_t sector; ++ unsigned int nr_sector; ++ unsigned int bytes; ++ char rwbs[8]; ++ char comm[16]; ++ u32 __data_loc_cmd; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_block_bio_complete { ++ struct trace_entry ent; ++ dev_t dev; ++ sector_t sector; ++ unsigned int nr_sector; ++ int error; ++ char rwbs[8]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_block_bio { ++ struct trace_entry ent; ++ dev_t dev; ++ sector_t sector; ++ unsigned int nr_sector; ++ char rwbs[8]; ++ char comm[16]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_block_plug { ++ struct trace_entry ent; ++ char comm[16]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_block_unplug { ++ struct trace_entry ent; ++ int nr_rq; ++ char comm[16]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_block_split { ++ struct trace_entry ent; ++ dev_t dev; ++ sector_t sector; ++ sector_t new_sector; ++ char rwbs[8]; ++ char comm[16]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_block_bio_remap { ++ struct trace_entry ent; ++ dev_t dev; ++ sector_t sector; ++ unsigned int nr_sector; ++ dev_t old_dev; ++ sector_t old_sector; ++ char rwbs[8]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_block_rq_remap { ++ struct trace_entry ent; ++ dev_t dev; ++ sector_t sector; ++ unsigned int nr_sector; ++ dev_t old_dev; ++ sector_t old_sector; ++ unsigned int nr_bios; ++ char rwbs[8]; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_block_buffer {}; ++ ++struct trace_event_data_offsets_block_rq_requeue { ++ u32 cmd; ++}; ++ ++struct trace_event_data_offsets_block_rq_completion { ++ u32 cmd; ++}; ++ ++struct trace_event_data_offsets_block_rq { ++ u32 cmd; ++}; ++ ++struct trace_event_data_offsets_block_bio_complete {}; ++ ++struct trace_event_data_offsets_block_bio {}; ++ ++struct trace_event_data_offsets_block_plug {}; ++ ++struct trace_event_data_offsets_block_unplug {}; ++ ++struct trace_event_data_offsets_block_split {}; ++ ++struct trace_event_data_offsets_block_bio_remap {}; ++ ++struct trace_event_data_offsets_block_rq_remap {}; ++ ++typedef void (*btf_trace_block_touch_buffer)(void *, struct buffer_head *); ++ ++typedef void (*btf_trace_block_dirty_buffer)(void *, struct buffer_head *); ++ ++typedef void (*btf_trace_block_rq_requeue)(void *, struct request *); ++ ++typedef void (*btf_trace_block_rq_complete)(void *, struct request *, blk_status_t, unsigned int); ++ ++typedef void (*btf_trace_block_rq_error)(void *, struct request *, blk_status_t, unsigned int); ++ ++typedef void (*btf_trace_block_rq_insert)(void *, struct request *); ++ ++typedef void (*btf_trace_block_rq_issue)(void *, struct request *); ++ ++typedef void (*btf_trace_block_rq_merge)(void *, struct request *); ++ ++typedef void (*btf_trace_block_bio_complete)(void *, struct request_queue *, struct bio *); ++ ++typedef void (*btf_trace_block_bio_bounce)(void *, struct bio *); ++ ++typedef void (*btf_trace_block_bio_backmerge)(void *, struct bio *); ++ ++typedef void (*btf_trace_block_bio_frontmerge)(void *, struct bio *); ++ ++typedef void (*btf_trace_block_bio_queue)(void *, struct bio *); ++ ++typedef void (*btf_trace_block_getrq)(void *, struct bio *); ++ ++typedef void (*btf_trace_block_plug)(void *, struct request_queue *); ++ ++typedef void (*btf_trace_block_unplug)(void *, struct request_queue *, unsigned int, bool); ++ ++typedef void (*btf_trace_block_split)(void *, struct bio *, unsigned int); ++ ++typedef void (*btf_trace_block_bio_remap)(void *, struct bio *, dev_t, sector_t); ++ ++typedef void (*btf_trace_block_rq_remap)(void *, struct request *, dev_t, sector_t); ++ ++struct throtl_grp; ++ ++struct throtl_qnode { ++ struct list_head node; ++ struct bio_list bios; ++ struct throtl_grp *tg; ++}; ++ ++struct throtl_service_queue { ++ struct throtl_service_queue *parent_sq; ++ struct list_head queued[2]; ++ unsigned int nr_queued[2]; ++ struct rb_root_cached pending_tree; ++ unsigned int nr_pending; ++ long unsigned int first_pending_disptime; ++ struct timer_list pending_timer; ++}; ++ ++struct throtl_grp { ++ struct blkg_policy_data pd; ++ struct rb_node rb_node; ++ struct throtl_data *td; ++ struct throtl_service_queue service_queue; ++ struct throtl_qnode qnode_on_self[2]; ++ struct throtl_qnode qnode_on_parent[2]; ++ long unsigned int disptime; ++ unsigned int flags; ++ bool has_rules_bps[2]; ++ bool has_rules_iops[2]; ++ uint64_t bps[4]; ++ uint64_t bps_conf[4]; ++ unsigned int iops[4]; ++ unsigned int iops_conf[4]; ++ uint64_t bytes_disp[2]; ++ unsigned int io_disp[2]; ++ long unsigned int last_low_overflow_time[2]; ++ uint64_t last_bytes_disp[2]; ++ unsigned int last_io_disp[2]; ++ uint64_t carryover_bytes[2]; ++ unsigned int carryover_ios[2]; ++ long unsigned int last_check_time; ++ long unsigned int latency_target; ++ long unsigned int latency_target_conf; ++ long unsigned int slice_start[2]; ++ long unsigned int slice_end[2]; ++ long unsigned int last_finish_time; ++ long unsigned int checked_last_finish_time; ++ long unsigned int avg_idletime; ++ long unsigned int idletime_threshold; ++ long unsigned int idletime_threshold_conf; ++ unsigned int bio_cnt; ++ unsigned int bad_bio_cnt; ++ long unsigned int bio_cnt_reset_time; ++ struct blkg_rwstat stat_bytes; ++ struct blkg_rwstat stat_ios; ++}; ++ ++enum blk_default_limits { ++ BLK_MAX_SEGMENTS = 128, ++ BLK_SAFE_MAX_SECTORS = 255, ++ BLK_MAX_SEGMENT_SIZE = 65536, ++ BLK_SEG_BOUNDARY_MASK = 4294967295, ++}; ++ ++struct req_iterator { ++ struct bvec_iter iter; ++ struct bio *bio; ++}; ++ ++enum bio_merge_status { ++ BIO_MERGE_OK = 0, ++ BIO_MERGE_NONE = 1, ++ BIO_MERGE_FAILED = 2, ++}; ++ ++struct blk_queue_stats { ++ struct list_head callbacks; ++ spinlock_t lock; ++ int accounting; ++}; ++ ++struct blk_mq_hw_ctx_sysfs_entry { ++ struct attribute attr; ++ ssize_t (*show)(struct blk_mq_hw_ctx *, char *); ++}; ++ ++struct blkpg_ioctl_arg { ++ int op; ++ int flags; ++ int datalen; ++ void *data; ++}; ++ ++struct blkpg_partition { ++ long long int start; ++ long long int length; ++ int pno; ++ char devname[64]; ++ char volname[64]; ++}; ++ ++struct pr_reservation { ++ __u64 key; ++ __u32 type; ++ __u32 flags; ++}; ++ ++struct pr_registration { ++ __u64 old_key; ++ __u64 new_key; ++ __u32 flags; ++ __u32 __pad; ++}; ++ ++struct pr_preempt { ++ __u64 old_key; ++ __u64 new_key; ++ __u32 type; ++ __u32 flags; ++}; ++ ++struct pr_clear { ++ __u64 key; ++ __u32 flags; ++ __u32 __pad; ++}; ++ ++struct compat_blkpg_ioctl_arg { ++ compat_int_t op; ++ compat_int_t flags; ++ compat_int_t datalen; ++ compat_caddr_t data; ++}; ++ ++struct compat_hd_geometry { ++ unsigned char heads; ++ unsigned char sectors; ++ short unsigned int cylinders; ++ u32 start; ++}; ++ ++enum { ++ IOPRIO_WHO_PROCESS = 1, ++ IOPRIO_WHO_PGRP = 2, ++ IOPRIO_WHO_USER = 3, ++}; ++ ++enum { ++ GENHD_FL_REMOVABLE = 1, ++ GENHD_FL_HIDDEN = 2, ++ GENHD_FL_NO_PART = 4, ++}; ++ ++struct msdos_partition { ++ u8 boot_ind; ++ u8 head; ++ u8 sector; ++ u8 cyl; ++ u8 sys_ind; ++ u8 end_head; ++ u8 end_sector; ++ u8 end_cyl; ++ __le32 start_sect; ++ __le32 nr_sects; ++}; ++ ++struct frag { ++ struct list_head list; ++ u32 group; ++ u8 num; ++ u8 rec; ++ u8 map; ++ u8 data[0]; ++}; ++ ++struct privhead { ++ u16 ver_major; ++ u16 ver_minor; ++ u64 logical_disk_start; ++ u64 logical_disk_size; ++ u64 config_start; ++ u64 config_size; ++ uuid_t disk_id; ++}; ++ ++struct tocblock { ++ u8 bitmap1_name[16]; ++ u64 bitmap1_start; ++ u64 bitmap1_size; ++ u8 bitmap2_name[16]; ++ u64 bitmap2_start; ++ u64 bitmap2_size; ++}; ++ ++struct vmdb { ++ u16 ver_major; ++ u16 ver_minor; ++ u32 vblk_size; ++ u32 vblk_offset; ++ u32 last_vblk_seq; ++}; ++ ++struct vblk_comp { ++ u8 state[16]; ++ u64 parent_id; ++ u8 type; ++ u8 children; ++ u16 chunksize; ++}; ++ ++struct vblk_dgrp { ++ u8 disk_id[64]; ++}; ++ ++struct vblk_disk { ++ uuid_t disk_id; ++ u8 alt_name[128]; ++}; ++ ++struct vblk_part { ++ u64 start; ++ u64 size; ++ u64 volume_offset; ++ u64 parent_id; ++ u64 disk_id; ++ u8 partnum; ++}; ++ ++struct vblk_volu { ++ u8 volume_type[16]; ++ u8 volume_state[16]; ++ u8 guid[16]; ++ u8 drive_hint[4]; ++ u64 size; ++ u8 partition_type; ++}; ++ ++struct vblk { ++ u8 name[64]; ++ u64 obj_id; ++ u32 sequence; ++ u8 flags; ++ u8 type; ++ union { ++ struct vblk_comp comp; ++ struct vblk_dgrp dgrp; ++ struct vblk_disk disk; ++ struct vblk_part part; ++ struct vblk_volu volu; ++ } vblk; ++ struct list_head list; ++}; ++ ++struct ldmdb { ++ struct privhead ph; ++ struct tocblock toc; ++ struct vmdb vm; ++ struct list_head v_dgrp; ++ struct list_head v_disk; ++ struct list_head v_volu; ++ struct list_head v_comp; ++ struct list_head v_part; ++}; ++ ++struct dasd_information2_t { ++ unsigned int devno; ++ unsigned int real_devno; ++ unsigned int schid; ++ unsigned int cu_type: 16; ++ unsigned int cu_model: 8; ++ long: 8; ++ unsigned int dev_type: 16; ++ unsigned int dev_model: 8; ++ unsigned int open_count; ++ unsigned int req_queue_len; ++ unsigned int chanq_len; ++ char type[4]; ++ unsigned int status; ++ unsigned int label_block; ++ unsigned int FBA_layout; ++ unsigned int characteristics_size; ++ unsigned int confdata_size; ++ char characteristics[64]; ++ char configuration_data[256]; ++ unsigned int format; ++ unsigned int features; ++ unsigned int reserved0; ++ unsigned int reserved1; ++ unsigned int reserved2; ++ unsigned int reserved3; ++ unsigned int reserved4; ++ unsigned int reserved5; ++ unsigned int reserved6; ++ unsigned int reserved7; ++}; ++ ++typedef struct dasd_information2_t dasd_information2_t; ++ ++struct vtoc_ttr { ++ __u16 tt; ++ __u8 r; ++} __attribute__((packed)); ++ ++struct vtoc_cchhb { ++ __u16 cc; ++ __u16 hh; ++ __u8 b; ++} __attribute__((packed)); ++ ++struct vtoc_cchh { ++ __u16 cc; ++ __u16 hh; ++}; ++ ++struct vtoc_labeldate { ++ __u8 year; ++ __u16 day; ++} __attribute__((packed)); ++ ++struct vtoc_volume_label_cdl { ++ char volkey[4]; ++ char vollbl[4]; ++ char volid[6]; ++ __u8 security; ++ struct vtoc_cchhb vtoc; ++ char res1[5]; ++ char cisize[4]; ++ char blkperci[4]; ++ char labperci[4]; ++ char res2[4]; ++ char lvtoc[14]; ++ char res3[29]; ++}; ++ ++struct vtoc_volume_label_ldl { ++ char vollbl[4]; ++ char volid[6]; ++ char res3[69]; ++ char ldl_version; ++ __u64 formatted_blocks; ++}; ++ ++struct vtoc_extent { ++ __u8 typeind; ++ __u8 seqno; ++ struct vtoc_cchh llimit; ++ struct vtoc_cchh ulimit; ++}; ++ ++struct vtoc_format1_label { ++ char DS1DSNAM[44]; ++ __u8 DS1FMTID; ++ char DS1DSSN[6]; ++ __u16 DS1VOLSQ; ++ struct vtoc_labeldate DS1CREDT; ++ struct vtoc_labeldate DS1EXPDT; ++ __u8 DS1NOEPV; ++ __u8 DS1NOBDB; ++ __u8 DS1FLAG1; ++ char DS1SYSCD[13]; ++ struct vtoc_labeldate DS1REFD; ++ __u8 DS1SMSFG; ++ __u8 DS1SCXTF; ++ __u16 DS1SCXTV; ++ __u8 DS1DSRG1; ++ __u8 DS1DSRG2; ++ __u8 DS1RECFM; ++ __u8 DS1OPTCD; ++ __u16 DS1BLKL; ++ __u16 DS1LRECL; ++ __u8 DS1KEYL; ++ __u16 DS1RKP; ++ __u8 DS1DSIND; ++ __u8 DS1SCAL1; ++ char DS1SCAL3[3]; ++ struct vtoc_ttr DS1LSTAR; ++ __u16 DS1TRBAL; ++ __u16 res1; ++ struct vtoc_extent DS1EXT1; ++ struct vtoc_extent DS1EXT2; ++ struct vtoc_extent DS1EXT3; ++ struct vtoc_cchhb DS1PTRDS; ++} __attribute__((packed)); ++ ++struct vtoc_cms_label { ++ __u8 label_id[4]; ++ __u8 vol_id[6]; ++ __u16 version_id; ++ __u32 block_size; ++ __u32 origin_ptr; ++ __u32 usable_count; ++ __u32 formatted_count; ++ __u32 block_count; ++ __u32 used_count; ++ __u32 fst_size; ++ __u32 fst_count; ++ __u8 format_date[6]; ++ __u8 reserved1[2]; ++ __u32 disk_offset; ++ __u32 map_block; ++ __u32 hblk_disp; ++ __u32 user_disp; ++ __u8 reserved2[4]; ++ __u8 segment_name[8]; ++}; ++ ++union label_t { ++ struct vtoc_volume_label_cdl vol; ++ struct vtoc_volume_label_ldl lnx; ++ struct vtoc_cms_label cms; ++}; ++ ++enum { ++ DISK_EVENT_FLAG_POLL = 1, ++ DISK_EVENT_FLAG_UEVENT = 2, ++ DISK_EVENT_FLAG_BLOCK_ON_EXCL_WRITE = 4, ++}; ++ ++struct disk_events { ++ struct list_head node; ++ struct gendisk *disk; ++ spinlock_t lock; ++ struct mutex block_mutex; ++ int block; ++ unsigned int pending; ++ unsigned int clearing; ++ long int poll_msecs; ++ struct delayed_work dwork; ++}; ++ ++struct bsg_device { ++ struct request_queue *queue; ++ struct device device; ++ struct cdev cdev; ++ int max_queue; ++ unsigned int timeout; ++ unsigned int reserved_size; ++ bsg_sg_io_fn *sg_io_fn; ++}; ++ ++struct latency_bucket { ++ long unsigned int total_latency; ++ int samples; ++}; ++ ++struct avg_latency_bucket { ++ long unsigned int latency; ++ bool valid; ++}; ++ ++struct throtl_data { ++ struct throtl_service_queue service_queue; ++ struct request_queue *queue; ++ unsigned int nr_queued[2]; ++ unsigned int throtl_slice; ++ struct work_struct dispatch_work; ++ unsigned int limit_index; ++ bool limit_valid[2]; ++ long unsigned int low_upgrade_time; ++ long unsigned int low_downgrade_time; ++ unsigned int scale; ++ struct latency_bucket tmp_buckets[18]; ++ struct avg_latency_bucket avg_buckets[18]; ++ struct latency_bucket *latency_buckets[2]; ++ long unsigned int last_calculate_time; ++ long unsigned int filtered_latency; ++ bool track_bio_latency; ++}; ++ ++enum tg_state_flags { ++ THROTL_TG_PENDING = 1, ++ THROTL_TG_WAS_EMPTY = 2, ++ THROTL_TG_CANCELING = 4, ++}; ++ ++struct blk_iolatency { ++ struct rq_qos rqos; ++ struct timer_list timer; ++ bool enabled; ++ atomic_t enable_cnt; ++ struct work_struct enable_work; ++}; ++ ++struct iolatency_grp; ++ ++struct child_latency_info { ++ spinlock_t lock; ++ u64 last_scale_event; ++ u64 scale_lat; ++ u64 nr_samples; ++ struct iolatency_grp *scale_grp; ++ atomic_t scale_cookie; ++}; ++ ++struct percentile_stats { ++ u64 total; ++ u64 missed; ++}; ++ ++struct latency_stat { ++ union { ++ struct percentile_stats ps; ++ struct blk_rq_stat rqs; ++ }; ++}; ++ ++struct iolatency_grp { ++ struct blkg_policy_data pd; ++ struct latency_stat *stats; ++ struct latency_stat cur_stat; ++ struct blk_iolatency *blkiolat; ++ unsigned int max_depth; ++ struct rq_wait rq_wait; ++ atomic64_t window_start; ++ atomic_t scale_cookie; ++ u64 min_lat_nsec; ++ u64 cur_win_nsec; ++ u64 lat_avg; ++ u64 nr_samples; ++ bool ssd; ++ struct child_latency_info child_lat; ++}; ++ ++struct sbq_wait { ++ struct sbitmap_queue *sbq; ++ struct wait_queue_entry wait; ++}; ++ ++struct trace_event_raw_kyber_latency { ++ struct trace_entry ent; ++ dev_t dev; ++ char domain[16]; ++ char type[8]; ++ u8 percentile; ++ u8 numerator; ++ u8 denominator; ++ unsigned int samples; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_kyber_adjust { ++ struct trace_entry ent; ++ dev_t dev; ++ char domain[16]; ++ unsigned int depth; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_kyber_throttled { ++ struct trace_entry ent; ++ dev_t dev; ++ char domain[16]; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_kyber_latency {}; ++ ++struct trace_event_data_offsets_kyber_adjust {}; ++ ++struct trace_event_data_offsets_kyber_throttled {}; ++ ++typedef void (*btf_trace_kyber_latency)(void *, dev_t, const char *, const char *, unsigned int, unsigned int, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_kyber_adjust)(void *, dev_t, const char *, unsigned int); ++ ++typedef void (*btf_trace_kyber_throttled)(void *, dev_t, const char *); ++ ++enum { ++ KYBER_READ = 0, ++ KYBER_WRITE = 1, ++ KYBER_DISCARD = 2, ++ KYBER_OTHER = 3, ++ KYBER_NUM_DOMAINS = 4, ++}; ++ ++enum { ++ KYBER_ASYNC_PERCENT = 75, ++}; ++ ++enum { ++ KYBER_LATENCY_SHIFT = 2, ++ KYBER_GOOD_BUCKETS = 4, ++ KYBER_LATENCY_BUCKETS = 8, ++}; ++ ++enum { ++ KYBER_TOTAL_LATENCY = 0, ++ KYBER_IO_LATENCY = 1, ++}; ++ ++struct kyber_cpu_latency { ++ atomic_t buckets[48]; ++}; ++ ++struct kyber_ctx_queue { ++ spinlock_t lock; ++ struct list_head rq_list[4]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct kyber_queue_data { ++ struct request_queue *q; ++ dev_t dev; ++ struct sbitmap_queue domain_tokens[4]; ++ unsigned int async_depth; ++ struct kyber_cpu_latency *cpu_latency; ++ struct timer_list timer; ++ unsigned int latency_buckets[48]; ++ long unsigned int latency_timeout[3]; ++ int domain_p99[3]; ++ u64 latency_targets[3]; ++}; ++ ++struct kyber_hctx_data { ++ spinlock_t lock; ++ struct list_head rqs[4]; ++ unsigned int cur_domain; ++ unsigned int batching; ++ struct kyber_ctx_queue *kcqs; ++ struct sbitmap kcq_map[4]; ++ struct sbq_wait domain_wait[4]; ++ struct sbq_wait_state *domain_ws[4]; ++ atomic_t wait_index[4]; ++}; ++ ++struct flush_kcq_data { ++ struct kyber_hctx_data *khd; ++ unsigned int sched_domain; ++ struct list_head *list; ++}; ++ ++enum bip_flags { ++ BIP_BLOCK_INTEGRITY = 1, ++ BIP_MAPPED_INTEGRITY = 2, ++ BIP_CTRL_NOCHECK = 4, ++ BIP_DISK_NOCHECK = 8, ++ BIP_IP_CHECKSUM = 16, ++}; ++ ++struct show_busy_params { ++ struct seq_file *m; ++ struct blk_mq_hw_ctx *hctx; ++}; ++ ++enum opal_mbr { ++ OPAL_MBR_ENABLE = 0, ++ OPAL_MBR_DISABLE = 1, ++}; ++ ++enum opal_mbr_done_flag { ++ OPAL_MBR_NOT_DONE = 0, ++ OPAL_MBR_DONE = 1, ++}; ++ ++enum opal_user { ++ OPAL_ADMIN1 = 0, ++ OPAL_USER1 = 1, ++ OPAL_USER2 = 2, ++ OPAL_USER3 = 3, ++ OPAL_USER4 = 4, ++ OPAL_USER5 = 5, ++ OPAL_USER6 = 6, ++ OPAL_USER7 = 7, ++ OPAL_USER8 = 8, ++ OPAL_USER9 = 9, ++}; ++ ++enum opal_lock_state { ++ OPAL_RO = 1, ++ OPAL_RW = 2, ++ OPAL_LK = 4, ++}; ++ ++enum opal_lock_flags { ++ OPAL_SAVE_FOR_LOCK = 1, ++}; ++ ++struct opal_key { ++ __u8 lr; ++ __u8 key_len; ++ __u8 __align[6]; ++ __u8 key[256]; ++}; ++ ++struct opal_lr_act { ++ struct opal_key key; ++ __u32 sum; ++ __u8 num_lrs; ++ __u8 lr[9]; ++ __u8 align[2]; ++}; ++ ++struct opal_session_info { ++ __u32 sum; ++ __u32 who; ++ struct opal_key opal_key; ++}; ++ ++struct opal_user_lr_setup { ++ __u64 range_start; ++ __u64 range_length; ++ __u32 RLE; ++ __u32 WLE; ++ struct opal_session_info session; ++}; ++ ++struct opal_lr_status { ++ struct opal_session_info session; ++ __u64 range_start; ++ __u64 range_length; ++ __u32 RLE; ++ __u32 WLE; ++ __u32 l_state; ++ __u8 align[4]; ++}; ++ ++struct opal_lock_unlock { ++ struct opal_session_info session; ++ __u32 l_state; ++ __u16 flags; ++ __u8 __align[2]; ++}; ++ ++struct opal_new_pw { ++ struct opal_session_info session; ++ struct opal_session_info new_user_pw; ++}; ++ ++struct opal_mbr_data { ++ struct opal_key key; ++ __u8 enable_disable; ++ __u8 __align[7]; ++}; ++ ++struct opal_mbr_done { ++ struct opal_key key; ++ __u8 done_flag; ++ __u8 __align[7]; ++}; ++ ++struct opal_shadow_mbr { ++ struct opal_key key; ++ const __u64 data; ++ __u64 offset; ++ __u64 size; ++}; ++ ++enum opal_table_ops { ++ OPAL_READ_TABLE = 0, ++ OPAL_WRITE_TABLE = 1, ++}; ++ ++struct opal_read_write_table { ++ struct opal_key key; ++ const __u64 data; ++ const __u8 table_uid[8]; ++ __u64 offset; ++ __u64 size; ++ __u64 flags; ++ __u64 priv; ++}; ++ ++struct opal_status { ++ __u32 flags; ++ __u32 reserved; ++}; ++ ++struct opal_geometry { ++ __u8 align; ++ __u32 logical_block_size; ++ __u64 alignment_granularity; ++ __u64 lowest_aligned_lba; ++ __u8 __align[3]; ++}; ++ ++typedef int sec_send_recv(void *, u16, u8, void *, size_t, bool); ++ ++enum { ++ TCG_SECP_00 = 0, ++ TCG_SECP_01 = 1, ++}; ++ ++enum opal_response_token { ++ OPAL_DTA_TOKENID_BYTESTRING = 224, ++ OPAL_DTA_TOKENID_SINT = 225, ++ OPAL_DTA_TOKENID_UINT = 226, ++ OPAL_DTA_TOKENID_TOKEN = 227, ++ OPAL_DTA_TOKENID_INVALID = 0, ++}; ++ ++enum opal_uid { ++ OPAL_SMUID_UID = 0, ++ OPAL_THISSP_UID = 1, ++ OPAL_ADMINSP_UID = 2, ++ OPAL_LOCKINGSP_UID = 3, ++ OPAL_ENTERPRISE_LOCKINGSP_UID = 4, ++ OPAL_ANYBODY_UID = 5, ++ OPAL_SID_UID = 6, ++ OPAL_ADMIN1_UID = 7, ++ OPAL_USER1_UID = 8, ++ OPAL_USER2_UID = 9, ++ OPAL_PSID_UID = 10, ++ OPAL_ENTERPRISE_BANDMASTER0_UID = 11, ++ OPAL_ENTERPRISE_ERASEMASTER_UID = 12, ++ OPAL_TABLE_TABLE = 13, ++ OPAL_LOCKINGRANGE_GLOBAL = 14, ++ OPAL_LOCKINGRANGE_ACE_START_TO_KEY = 15, ++ OPAL_LOCKINGRANGE_ACE_RDLOCKED = 16, ++ OPAL_LOCKINGRANGE_ACE_WRLOCKED = 17, ++ OPAL_MBRCONTROL = 18, ++ OPAL_MBR = 19, ++ OPAL_AUTHORITY_TABLE = 20, ++ OPAL_C_PIN_TABLE = 21, ++ OPAL_LOCKING_INFO_TABLE = 22, ++ OPAL_ENTERPRISE_LOCKING_INFO_TABLE = 23, ++ OPAL_DATASTORE = 24, ++ OPAL_C_PIN_MSID = 25, ++ OPAL_C_PIN_SID = 26, ++ OPAL_C_PIN_ADMIN1 = 27, ++ OPAL_HALF_UID_AUTHORITY_OBJ_REF = 28, ++ OPAL_HALF_UID_BOOLEAN_ACE = 29, ++ OPAL_UID_HEXFF = 30, ++}; ++ ++enum opal_method { ++ OPAL_PROPERTIES = 0, ++ OPAL_STARTSESSION = 1, ++ OPAL_REVERT = 2, ++ OPAL_ACTIVATE = 3, ++ OPAL_EGET = 4, ++ OPAL_ESET = 5, ++ OPAL_NEXT = 6, ++ OPAL_EAUTHENTICATE = 7, ++ OPAL_GETACL = 8, ++ OPAL_GENKEY = 9, ++ OPAL_REVERTSP = 10, ++ OPAL_GET = 11, ++ OPAL_SET = 12, ++ OPAL_AUTHENTICATE = 13, ++ OPAL_RANDOM = 14, ++ OPAL_ERASE = 15, ++}; ++ ++enum opal_token { ++ OPAL_TRUE = 1, ++ OPAL_FALSE = 0, ++ OPAL_BOOLEAN_EXPR = 3, ++ OPAL_TABLE = 0, ++ OPAL_STARTROW = 1, ++ OPAL_ENDROW = 2, ++ OPAL_STARTCOLUMN = 3, ++ OPAL_ENDCOLUMN = 4, ++ OPAL_VALUES = 1, ++ OPAL_TABLE_UID = 0, ++ OPAL_TABLE_NAME = 1, ++ OPAL_TABLE_COMMON = 2, ++ OPAL_TABLE_TEMPLATE = 3, ++ OPAL_TABLE_KIND = 4, ++ OPAL_TABLE_COLUMN = 5, ++ OPAL_TABLE_COLUMNS = 6, ++ OPAL_TABLE_ROWS = 7, ++ OPAL_TABLE_ROWS_FREE = 8, ++ OPAL_TABLE_ROW_BYTES = 9, ++ OPAL_TABLE_LASTID = 10, ++ OPAL_TABLE_MIN = 11, ++ OPAL_TABLE_MAX = 12, ++ OPAL_PIN = 3, ++ OPAL_RANGESTART = 3, ++ OPAL_RANGELENGTH = 4, ++ OPAL_READLOCKENABLED = 5, ++ OPAL_WRITELOCKENABLED = 6, ++ OPAL_READLOCKED = 7, ++ OPAL_WRITELOCKED = 8, ++ OPAL_ACTIVEKEY = 10, ++ OPAL_LIFECYCLE = 6, ++ OPAL_MAXRANGES = 4, ++ OPAL_MBRENABLE = 1, ++ OPAL_MBRDONE = 2, ++ OPAL_HOSTPROPERTIES = 0, ++ OPAL_STARTLIST = 240, ++ OPAL_ENDLIST = 241, ++ OPAL_STARTNAME = 242, ++ OPAL_ENDNAME = 243, ++ OPAL_CALL = 248, ++ OPAL_ENDOFDATA = 249, ++ OPAL_ENDOFSESSION = 250, ++ OPAL_STARTTRANSACTON = 251, ++ OPAL_ENDTRANSACTON = 252, ++ OPAL_EMPTYATOM = 255, ++ OPAL_WHERE = 0, ++}; ++ ++enum opal_parameter { ++ OPAL_SUM_SET_LIST = 393216, ++}; ++ ++struct opal_compacket { ++ __be32 reserved0; ++ u8 extendedComID[4]; ++ __be32 outstandingData; ++ __be32 minTransfer; ++ __be32 length; ++}; ++ ++struct opal_packet { ++ __be32 tsn; ++ __be32 hsn; ++ __be32 seq_number; ++ __be16 reserved0; ++ __be16 ack_type; ++ __be32 acknowledgment; ++ __be32 length; ++}; ++ ++struct opal_data_subpacket { ++ u8 reserved0[6]; ++ __be16 kind; ++ __be32 length; ++}; ++ ++struct opal_header { ++ struct opal_compacket cp; ++ struct opal_packet pkt; ++ struct opal_data_subpacket subpkt; ++}; ++ ++struct d0_header { ++ __be32 length; ++ __be32 revision; ++ __be32 reserved01; ++ __be32 reserved02; ++ u8 ignored[32]; ++}; ++ ++struct d0_tper_features { ++ u8 supported_features; ++ u8 reserved01[3]; ++ __be32 reserved02; ++ __be32 reserved03; ++}; ++ ++struct d0_locking_features { ++ u8 supported_features; ++ u8 reserved01[3]; ++ __be32 reserved02; ++ __be32 reserved03; ++}; ++ ++struct d0_geometry_features { ++ u8 header[4]; ++ u8 reserved01; ++ u8 reserved02[7]; ++ __be32 logical_block_size; ++ __be64 alignment_granularity; ++ __be64 lowest_aligned_lba; ++}; ++ ++struct d0_opal_v100 { ++ __be16 baseComID; ++ __be16 numComIDs; ++}; ++ ++struct d0_single_user_mode { ++ __be32 num_locking_objects; ++ u8 reserved01; ++ u8 reserved02; ++ __be16 reserved03; ++ __be32 reserved04; ++}; ++ ++struct d0_opal_v200 { ++ __be16 baseComID; ++ __be16 numComIDs; ++ u8 range_crossing; ++ u8 num_locking_admin_auth[2]; ++ u8 num_locking_user_auth[2]; ++ u8 initialPIN; ++ u8 revertedPIN; ++ u8 reserved01; ++ __be32 reserved02; ++}; ++ ++struct d0_features { ++ __be16 code; ++ u8 r_version; ++ u8 length; ++ u8 features[0]; ++}; ++ ++struct opal_step { ++ int (*fn)(struct opal_dev *, void *); ++ void *data; ++}; ++ ++enum opal_atom_width { ++ OPAL_WIDTH_TINY = 0, ++ OPAL_WIDTH_SHORT = 1, ++ OPAL_WIDTH_MEDIUM = 2, ++ OPAL_WIDTH_LONG = 3, ++ OPAL_WIDTH_TOKEN = 4, ++}; ++ ++struct opal_resp_tok { ++ const u8 *pos; ++ size_t len; ++ enum opal_response_token type; ++ enum opal_atom_width width; ++ union { ++ u64 u; ++ s64 s; ++ } stored; ++}; ++ ++struct parsed_resp { ++ int num; ++ struct opal_resp_tok toks[64]; ++}; ++ ++struct opal_dev { ++ u32 flags; ++ void *data; ++ sec_send_recv *send_recv; ++ struct mutex dev_lock; ++ u16 comid; ++ u32 hsn; ++ u32 tsn; ++ u64 align; ++ u64 lowest_lba; ++ u32 logical_block_size; ++ u8 align_required; ++ size_t pos; ++ u8 *cmd; ++ u8 *resp; ++ struct parsed_resp parsed; ++ size_t prev_d_len; ++ void *prev_data; ++ struct list_head unlk_lst; ++}; ++ ++typedef int cont_fn(struct opal_dev *); ++ ++struct opal_suspend_data { ++ struct opal_lock_unlock unlk; ++ u8 lr; ++ struct list_head node; ++}; ++ ++struct blk_crypto_keyslot { ++ atomic_t slot_refs; ++ struct list_head idle_slot_node; ++ struct hlist_node hash_node; ++ const struct blk_crypto_key *key; ++ struct blk_crypto_profile *profile; ++}; ++ ++struct bd_holder_disk { ++ struct list_head list; ++ struct kobject *holder_dir; ++ int refcnt; ++}; ++ ++struct io_rename { ++ struct file *file; ++ int old_dfd; ++ int new_dfd; ++ struct filename *oldpath; ++ struct filename *newpath; ++ int flags; ++}; ++ ++struct io_unlink { ++ struct file *file; ++ int dfd; ++ int flags; ++ struct filename *filename; ++}; ++ ++struct io_mkdir { ++ struct file *file; ++ int dfd; ++ umode_t mode; ++ struct filename *filename; ++}; ++ ++struct io_link { ++ struct file *file; ++ int old_dfd; ++ int new_dfd; ++ struct filename *oldpath; ++ struct filename *newpath; ++ int flags; ++}; ++ ++struct io_sync { ++ struct file *file; ++ loff_t len; ++ loff_t off; ++ int flags; ++ int mode; ++}; ++ ++struct io_uring_file_index_range { ++ __u32 off; ++ __u32 len; ++ __u64 resv; ++}; ++ ++struct io_cancel_data { ++ struct io_ring_ctx *ctx; ++ union { ++ u64 data; ++ struct file *file; ++ }; ++ u32 flags; ++ int seq; ++}; ++ ++struct io_timeout_data { ++ struct io_kiocb *req; ++ struct hrtimer timer; ++ struct timespec64 ts; ++ enum hrtimer_mode mode; ++ u32 flags; ++}; ++ ++struct io_timeout { ++ struct file *file; ++ u32 off; ++ u32 target_seq; ++ u32 repeats; ++ struct list_head list; ++ struct io_kiocb *head; ++ struct io_kiocb *prev; ++}; ++ ++struct io_timeout_rem { ++ struct file *file; ++ u64 addr; ++ struct timespec64 ts; ++ u32 flags; ++ bool ltimeout; ++}; ++ ++struct io_overflow_cqe { ++ struct list_head list; ++ struct io_uring_cqe cqe; ++}; ++ ++struct io_uring_sync_cancel_reg { ++ __u64 addr; ++ __s32 fd; ++ __u32 flags; ++ struct __kernel_timespec timeout; ++ __u64 pad[4]; ++}; ++ ++enum io_wq_cancel { ++ IO_WQ_CANCEL_OK = 0, ++ IO_WQ_CANCEL_RUNNING = 1, ++ IO_WQ_CANCEL_NOTFOUND = 2, ++}; ++ ++typedef bool work_cancel_fn(struct io_wq_work *, void *); ++ ++struct io_cancel { ++ struct file *file; ++ u64 addr; ++ u32 flags; ++ s32 fd; ++}; ++ ++struct io_uring_rsrc_register { ++ __u32 nr; ++ __u32 flags; ++ __u64 resv2; ++ __u64 data; ++ __u64 tags; ++}; ++ ++struct io_uring_rsrc_update2 { ++ __u32 offset; ++ __u32 resv; ++ __u64 data; ++ __u64 tags; ++ __u32 nr; ++ __u32 resv2; ++}; ++ ++enum { ++ IORING_RSRC_FILE = 0, ++ IORING_RSRC_BUFFER = 1, ++}; ++ ++struct io_rsrc_update { ++ struct file *file; ++ u64 arg; ++ u32 nr_args; ++ u32 offset; ++}; ++ ++enum { ++ IOU_F_TWQ_FORCE_NORMAL = 1, ++ IOU_F_TWQ_LAZY_WAKE = 2, ++}; ++ ++struct io_notif_data { ++ struct file *file; ++ struct ubuf_info uarg; ++ long unsigned int account_pages; ++ bool zc_report; ++ bool zc_used; ++ bool zc_copied; ++}; ++ ++struct region { ++ unsigned int start; ++ unsigned int off; ++ unsigned int group_len; ++ unsigned int end; ++ unsigned int nbits; ++}; ++ ++enum { ++ REG_OP_ISFREE = 0, ++ REG_OP_ALLOC = 1, ++ REG_OP_RELEASE = 2, ++}; ++ ++struct once_work { ++ struct work_struct work; ++ struct static_key_true *key; ++ struct module *module; ++}; ++ ++struct strarray { ++ char **array; ++ size_t n; ++}; ++ ++enum devm_ioremap_type { ++ DEVM_IOREMAP = 0, ++ DEVM_IOREMAP_UC = 1, ++ DEVM_IOREMAP_WC = 2, ++ DEVM_IOREMAP_NP = 3, ++}; ++ ++struct pcim_iomap_devres { ++ void *table[6]; ++}; ++ ++struct arch_io_reserve_memtype_wc_devres { ++ resource_size_t start; ++ resource_size_t size; ++}; ++ ++struct xxh32_state { ++ uint32_t total_len_32; ++ uint32_t large_len; ++ uint32_t v1; ++ uint32_t v2; ++ uint32_t v3; ++ uint32_t v4; ++ uint32_t mem32[4]; ++ uint32_t memsize; ++}; ++ ++struct gen_pool_chunk { ++ struct list_head next_chunk; ++ atomic_long_t avail; ++ phys_addr_t phys_addr; ++ void *owner; ++ long unsigned int start_addr; ++ long unsigned int end_addr; ++ long unsigned int bits[0]; ++}; ++ ++struct genpool_data_align { ++ int align; ++}; ++ ++struct genpool_data_fixed { ++ long unsigned int offset; ++}; ++ ++typedef struct { ++ unsigned char op; ++ unsigned char bits; ++ short unsigned int val; ++} code; ++ ++typedef enum { ++ HEAD = 0, ++ FLAGS = 1, ++ TIME = 2, ++ OS = 3, ++ EXLEN = 4, ++ EXTRA = 5, ++ NAME = 6, ++ COMMENT = 7, ++ HCRC = 8, ++ DICTID = 9, ++ DICT = 10, ++ TYPE = 11, ++ TYPEDO = 12, ++ STORED = 13, ++ COPY = 14, ++ TABLE = 15, ++ LENLENS = 16, ++ CODELENS = 17, ++ LEN = 18, ++ LENEXT = 19, ++ DIST = 20, ++ DISTEXT = 21, ++ MATCH = 22, ++ LIT = 23, ++ CHECK = 24, ++ LENGTH = 25, ++ DONE = 26, ++ BAD = 27, ++ MEM = 28, ++ SYNC = 29, ++} inflate_mode; ++ ++struct inflate_state { ++ inflate_mode mode; ++ int last; ++ int wrap; ++ int havedict; ++ int flags; ++ unsigned int dmax; ++ long unsigned int check; ++ long unsigned int total; ++ unsigned int wbits; ++ unsigned int wsize; ++ unsigned int whave; ++ unsigned int write; ++ unsigned char *window; ++ long unsigned int hold; ++ unsigned int bits; ++ unsigned int length; ++ unsigned int offset; ++ unsigned int extra; ++ const code *lencode; ++ const code *distcode; ++ unsigned int lenbits; ++ unsigned int distbits; ++ unsigned int ncode; ++ unsigned int nlen; ++ unsigned int ndist; ++ unsigned int have; ++ code *next; ++ short unsigned int lens[320]; ++ short unsigned int work[288]; ++ code codes[2048]; ++}; ++ ++union uu { ++ short unsigned int us; ++ unsigned char b[2]; ++}; ++ ++typedef unsigned int uInt; ++ ++typedef enum { ++ CODES = 0, ++ LENS = 1, ++ DISTS = 2, ++} codetype; ++ ++struct dfltcc_qaf_param { ++ char fns[16]; ++ char reserved1[8]; ++ char fmts[2]; ++ char reserved2[6]; ++}; ++ ++struct dfltcc_param_v0 { ++ uint16_t pbvn; ++ uint8_t mvn; ++ uint8_t ribm; ++ unsigned int reserved32: 31; ++ unsigned int cf: 1; ++ uint8_t reserved64[8]; ++ unsigned int nt: 1; ++ unsigned int reserved129: 1; ++ unsigned int cvt: 1; ++ unsigned int reserved131: 1; ++ unsigned int htt: 1; ++ unsigned int bcf: 1; ++ unsigned int bcc: 1; ++ unsigned int bhf: 1; ++ unsigned int reserved136: 1; ++ unsigned int reserved137: 1; ++ unsigned int dhtgc: 1; ++ unsigned int reserved139: 5; ++ unsigned int reserved144: 5; ++ unsigned int sbb: 3; ++ uint8_t oesc; ++ unsigned int reserved160: 12; ++ unsigned int ifs: 4; ++ uint16_t ifl; ++ uint8_t reserved192[8]; ++ uint8_t reserved256[8]; ++ uint8_t reserved320[4]; ++ uint16_t hl; ++ unsigned int reserved368: 1; ++ uint16_t ho: 15; ++ uint32_t cv; ++ unsigned int eobs: 15; ++ unsigned int reserved431: 1; ++ uint8_t eobl: 4; ++ unsigned int reserved436: 12; ++ unsigned int reserved448: 4; ++ uint16_t cdhtl: 12; ++ uint8_t reserved464[6]; ++ uint8_t cdht[288]; ++ uint8_t reserved[32]; ++ uint8_t csb[1152]; ++}; ++ ++struct dfltcc_state { ++ struct dfltcc_param_v0 param; ++ struct dfltcc_qaf_param af; ++ char msg[64]; ++}; ++ ++struct inflate_workspace { ++ struct inflate_state inflate_state; ++ struct dfltcc_state dfltcc_state; ++ unsigned char working_window[36864]; ++}; ++ ++typedef enum { ++ DFLTCC_INFLATE_CONTINUE = 0, ++ DFLTCC_INFLATE_BREAK = 1, ++ DFLTCC_INFLATE_SOFTWARE = 2, ++} dfltcc_inflate_action; ++ ++typedef unsigned char uch; ++ ++typedef short unsigned int ush; ++ ++typedef long unsigned int ulg; ++ ++struct ct_data_s { ++ union { ++ ush freq; ++ ush code; ++ } fc; ++ union { ++ ush dad; ++ ush len; ++ } dl; ++}; ++ ++typedef struct ct_data_s ct_data; ++ ++struct static_tree_desc_s { ++ const ct_data *static_tree; ++ const int *extra_bits; ++ int extra_base; ++ int elems; ++ int max_length; ++}; ++ ++typedef struct static_tree_desc_s static_tree_desc; ++ ++struct tree_desc_s { ++ ct_data *dyn_tree; ++ int max_code; ++ static_tree_desc *stat_desc; ++}; ++ ++typedef ush Pos; ++ ++typedef unsigned int IPos; ++ ++struct deflate_state { ++ z_streamp strm; ++ int status; ++ Byte *pending_buf; ++ ulg pending_buf_size; ++ Byte *pending_out; ++ int pending; ++ int noheader; ++ Byte data_type; ++ Byte method; ++ int last_flush; ++ uInt w_size; ++ uInt w_bits; ++ uInt w_mask; ++ Byte *window; ++ ulg window_size; ++ Pos *prev; ++ Pos *head; ++ uInt ins_h; ++ uInt hash_size; ++ uInt hash_bits; ++ uInt hash_mask; ++ uInt hash_shift; ++ long int block_start; ++ uInt match_length; ++ IPos prev_match; ++ int match_available; ++ uInt strstart; ++ uInt match_start; ++ uInt lookahead; ++ uInt prev_length; ++ uInt max_chain_length; ++ uInt max_lazy_match; ++ int level; ++ int strategy; ++ uInt good_match; ++ int nice_match; ++ struct ct_data_s dyn_ltree[573]; ++ struct ct_data_s dyn_dtree[61]; ++ struct ct_data_s bl_tree[39]; ++ struct tree_desc_s l_desc; ++ struct tree_desc_s d_desc; ++ struct tree_desc_s bl_desc; ++ ush bl_count[16]; ++ int heap[573]; ++ int heap_len; ++ int heap_max; ++ uch depth[573]; ++ uch *l_buf; ++ uInt lit_bufsize; ++ uInt last_lit; ++ ush *d_buf; ++ ulg opt_len; ++ ulg static_len; ++ ulg compressed_len; ++ uInt matches; ++ int last_eob_len; ++ ush bi_buf; ++ int bi_valid; ++}; ++ ++typedef struct deflate_state deflate_state; ++ ++typedef enum { ++ need_more = 0, ++ block_done = 1, ++ finish_started = 2, ++ finish_done = 3, ++} block_state; ++ ++struct dfltcc_deflate_state { ++ struct dfltcc_state common; ++ uLong level_mask; ++ uLong block_size; ++ uLong block_threshold; ++ uLong dht_threshold; ++}; ++ ++typedef block_state (*compress_func)(deflate_state *, int); ++ ++struct deflate_workspace { ++ deflate_state deflate_memory; ++ struct dfltcc_deflate_state dfltcc_memory; ++ Byte *window_memory; ++ Pos *prev_memory; ++ Pos *head_memory; ++ char *overlay_memory; ++}; ++ ++typedef struct deflate_workspace deflate_workspace; ++ ++struct config_s { ++ ush good_length; ++ ush max_lazy; ++ ush nice_length; ++ ush max_chain; ++ compress_func func; ++}; ++ ++typedef struct config_s config; ++ ++typedef enum { ++ DFLTCC_CC_OK = 0, ++ DFLTCC_CC_OP1_TOO_SHORT = 1, ++ DFLTCC_CC_OP2_TOO_SHORT = 2, ++ DFLTCC_CC_OP2_CORRUPT = 2, ++ DFLTCC_CC_AGAIN = 3, ++} dfltcc_cc; ++ ++typedef ZSTD_compressionParameters zstd_compression_parameters; ++ ++typedef ZSTD_parameters zstd_parameters; ++ ++typedef ZSTD_CCtx zstd_cctx; ++ ++typedef ZSTD_CStream zstd_cstream; ++ ++typedef enum { ++ trustInput = 0, ++ checkMaxSymbolValue = 1, ++} HIST_checkInput_e; ++ ++typedef struct { ++ FSE_CTable CTable[59]; ++ U32 scratchBuffer[41]; ++ unsigned int count[13]; ++ S16 norm[13]; ++} HUF_CompressWeightsWksp; ++ ++typedef struct { ++ HUF_CompressWeightsWksp wksp; ++ BYTE bitsToWeight[13]; ++ BYTE huffWeight[255]; ++} HUF_WriteCTableWksp; ++ ++struct nodeElt_s { ++ U32 count; ++ U16 parent; ++ BYTE byte; ++ BYTE nbBits; ++}; ++ ++typedef struct nodeElt_s nodeElt; ++ ++typedef struct { ++ U16 base; ++ U16 curr; ++} rankPos; ++ ++typedef nodeElt huffNodeTable[512]; ++ ++typedef struct { ++ huffNodeTable huffNodeTbl; ++ rankPos rankPosition[192]; ++} HUF_buildCTable_wksp_tables; ++ ++typedef struct { ++ size_t bitContainer[2]; ++ size_t bitPos[2]; ++ BYTE *startPtr; ++ BYTE *ptr; ++ BYTE *endPtr; ++} HUF_CStream_t; ++ ++typedef enum { ++ HUF_singleStream = 0, ++ HUF_fourStreams = 1, ++} HUF_nbStreams_e; ++ ++typedef struct { ++ unsigned int count[256]; ++ HUF_CElt CTable[257]; ++ union { ++ HUF_buildCTable_wksp_tables buildCTable_wksp; ++ HUF_WriteCTableWksp writeCTable_wksp; ++ U32 hist_wksp[1024]; ++ } wksps; ++} HUF_compress_tables_t; ++ ++typedef struct { ++ U32 litLength; ++ U32 matchLength; ++} ZSTD_sequenceLength; ++ ++typedef U64 ZSTD_VecMask; ++ ++typedef enum { ++ search_hashChain = 0, ++ search_binaryTree = 1, ++ search_rowHash = 2, ++} searchMethod_e; ++ ++struct syscall_info { ++ __u64 sp; ++ struct seccomp_data data; ++}; ++ ++struct cpu_rmap { ++ struct kref refcount; ++ u16 size; ++ void **obj; ++ struct { ++ u16 index; ++ u16 dist; ++ } near[0]; ++}; ++ ++struct irq_glue { ++ struct irq_affinity_notify notify; ++ struct cpu_rmap *rmap; ++ u16 index; ++}; ++ ++typedef mpi_limb_t *mpi_ptr_t; ++ ++typedef int mpi_size_t; ++ ++typedef mpi_limb_t UWtype; ++ ++typedef unsigned int UHWtype; ++ ++enum gcry_mpi_constants { ++ MPI_C_ZERO = 0, ++ MPI_C_ONE = 1, ++ MPI_C_TWO = 2, ++ MPI_C_THREE = 3, ++ MPI_C_FOUR = 4, ++ MPI_C_EIGHT = 5, ++}; ++ ++struct barrett_ctx_s; ++ ++typedef struct barrett_ctx_s *mpi_barrett_t; ++ ++struct gcry_mpi_point { ++ MPI x; ++ MPI y; ++ MPI z; ++}; ++ ++typedef struct gcry_mpi_point *MPI_POINT; ++ ++enum gcry_mpi_ec_models { ++ MPI_EC_WEIERSTRASS = 0, ++ MPI_EC_MONTGOMERY = 1, ++ MPI_EC_EDWARDS = 2, ++}; ++ ++enum ecc_dialects { ++ ECC_DIALECT_STANDARD = 0, ++ ECC_DIALECT_ED25519 = 1, ++ ECC_DIALECT_SAFECURVE = 2, ++}; ++ ++struct mpi_ec_ctx { ++ enum gcry_mpi_ec_models model; ++ enum ecc_dialects dialect; ++ int flags; ++ unsigned int nbits; ++ MPI p; ++ MPI a; ++ MPI b; ++ MPI_POINT G; ++ MPI n; ++ unsigned int h; ++ MPI_POINT Q; ++ MPI d; ++ const char *name; ++ struct { ++ struct { ++ unsigned int a_is_pminus3: 1; ++ unsigned int two_inv_p: 1; ++ } valid; ++ int a_is_pminus3; ++ MPI two_inv_p; ++ mpi_barrett_t p_barrett; ++ MPI scratch[11]; ++ } t; ++ void (*addm)(MPI, MPI, MPI, struct mpi_ec_ctx *); ++ void (*subm)(MPI, MPI, MPI, struct mpi_ec_ctx *); ++ void (*mulm)(MPI, MPI, MPI, struct mpi_ec_ctx *); ++ void (*pow2)(MPI, const MPI, struct mpi_ec_ctx *); ++ void (*mul2)(MPI, MPI, struct mpi_ec_ctx *); ++}; ++ ++struct field_table { ++ const char *p; ++ void (*addm)(MPI, MPI, MPI, struct mpi_ec_ctx *); ++ void (*subm)(MPI, MPI, MPI, struct mpi_ec_ctx *); ++ void (*mulm)(MPI, MPI, MPI, struct mpi_ec_ctx *); ++ void (*mul2)(MPI, MPI, struct mpi_ec_ctx *); ++ void (*pow2)(MPI, const MPI, struct mpi_ec_ctx *); ++}; ++ ++struct barrett_ctx_s { ++ MPI m; ++ int m_copied; ++ int k; ++ MPI y; ++ MPI r1; ++ MPI r2; ++ MPI r3; ++}; ++ ++struct karatsuba_ctx { ++ struct karatsuba_ctx *next; ++ mpi_ptr_t tspace; ++ mpi_size_t tspace_size; ++ mpi_ptr_t tp; ++ mpi_size_t tp_size; ++}; ++ ++typedef long int mpi_limb_signed_t; ++ ++struct dim_sample { ++ ktime_t time; ++ u32 pkt_ctr; ++ u32 byte_ctr; ++ u16 event_ctr; ++ u32 comp_ctr; ++}; ++ ++struct dim_stats { ++ int ppms; ++ int bpms; ++ int epms; ++ int cpms; ++ int cpe_ratio; ++}; ++ ++struct dim { ++ u8 state; ++ struct dim_stats prev_stats; ++ struct dim_sample start_sample; ++ struct dim_sample measuring_sample; ++ struct work_struct work; ++ void *priv; ++ u8 profile_ix; ++ u8 mode; ++ u8 tune_state; ++ u8 steps_right; ++ u8 steps_left; ++ u8 tired; ++}; ++ ++enum dim_tune_state { ++ DIM_PARKING_ON_TOP = 0, ++ DIM_PARKING_TIRED = 1, ++ DIM_GOING_RIGHT = 2, ++ DIM_GOING_LEFT = 3, ++}; ++ ++struct dim_cq_moder { ++ u16 usec; ++ u16 pkts; ++ u16 comps; ++ u8 cq_period_mode; ++}; ++ ++enum dim_cq_period_mode { ++ DIM_CQ_PERIOD_MODE_START_FROM_EQE = 0, ++ DIM_CQ_PERIOD_MODE_START_FROM_CQE = 1, ++ DIM_CQ_PERIOD_NUM_MODES = 2, ++}; ++ ++enum dim_state { ++ DIM_START_MEASURE = 0, ++ DIM_MEASURE_IN_PROGRESS = 1, ++ DIM_APPLY_NEW_PROFILE = 2, ++}; ++ ++enum dim_stats_state { ++ DIM_STATS_WORSE = 0, ++ DIM_STATS_SAME = 1, ++ DIM_STATS_BETTER = 2, ++}; ++ ++enum dim_step_result { ++ DIM_STEPPED = 0, ++ DIM_TOO_TIRED = 1, ++ DIM_ON_EDGE = 2, ++}; ++ ++enum pubkey_algo { ++ PUBKEY_ALGO_RSA = 0, ++ PUBKEY_ALGO_MAX = 1, ++}; ++ ++struct pubkey_hdr { ++ uint8_t version; ++ uint32_t timestamp; ++ uint8_t algo; ++ uint8_t nmpi; ++ char mpi[0]; ++} __attribute__((packed)); ++ ++struct signature_hdr { ++ uint8_t version; ++ uint32_t timestamp; ++ uint8_t algo; ++ uint8_t hash; ++ uint8_t keyid[8]; ++ uint8_t nmpi; ++ char mpi[0]; ++} __attribute__((packed)); ++ ++struct font_desc { ++ int idx; ++ const char *name; ++ unsigned int width; ++ unsigned int height; ++ unsigned int charcount; ++ const void *data; ++ int pref; ++}; ++ ++struct font_data { ++ unsigned int extra[4]; ++ const unsigned char data[0]; ++}; ++ ++enum { ++ PCI_REASSIGN_ALL_RSRC = 1, ++ PCI_REASSIGN_ALL_BUS = 2, ++ PCI_PROBE_ONLY = 4, ++ PCI_CAN_SKIP_ISA_ALIGN = 8, ++ PCI_ENABLE_PROC_DOMAINS = 16, ++ PCI_COMPAT_DOMAIN_0 = 32, ++ PCI_SCAN_ALL_PCIE_DEVS = 64, ++}; ++ ++enum pcie_bus_config_types { ++ PCIE_BUS_TUNE_OFF = 0, ++ PCIE_BUS_DEFAULT = 1, ++ PCIE_BUS_SAFE = 2, ++ PCIE_BUS_PERFORMANCE = 3, ++ PCIE_BUS_PEER2PEER = 4, ++}; ++ ++enum pci_bar_type { ++ pci_bar_unknown = 0, ++ pci_bar_io = 1, ++ pci_bar_mem32 = 2, ++ pci_bar_mem64 = 3, ++}; ++ ++struct pci_domain_busn_res { ++ struct list_head list; ++ struct resource res; ++ int domain_nr; ++}; ++ ++struct pcie_port_service_driver { ++ const char *name; ++ int (*probe)(struct pcie_device *); ++ void (*remove)(struct pcie_device *); ++ int (*suspend)(struct pcie_device *); ++ int (*resume_noirq)(struct pcie_device *); ++ int (*resume)(struct pcie_device *); ++ int (*runtime_suspend)(struct pcie_device *); ++ int (*runtime_resume)(struct pcie_device *); ++ int (*slot_reset)(struct pcie_device *); ++ int port_type; ++ u32 service; ++ struct device_driver driver; ++}; ++ ++struct pci_dynid { ++ struct list_head node; ++ struct pci_device_id id; ++}; ++ ++struct drv_dev_and_id { ++ struct pci_driver *drv; ++ struct pci_dev *dev; ++ const struct pci_device_id *id; ++}; ++ ++enum pcie_link_width { ++ PCIE_LNK_WIDTH_RESRV = 0, ++ PCIE_LNK_X1 = 1, ++ PCIE_LNK_X2 = 2, ++ PCIE_LNK_X4 = 4, ++ PCIE_LNK_X8 = 8, ++ PCIE_LNK_X12 = 12, ++ PCIE_LNK_X16 = 16, ++ PCIE_LNK_X32 = 32, ++ PCIE_LNK_WIDTH_UNKNOWN = 255, ++}; ++ ++struct msi_domain_template { ++ char name[48]; ++ struct irq_chip chip; ++ struct msi_domain_ops ops; ++ struct msi_domain_info info; ++}; ++ ++struct pcie_link_state { ++ struct pci_dev *pdev; ++ struct pci_dev *downstream; ++ struct pcie_link_state *root; ++ struct pcie_link_state *parent; ++ struct list_head sibling; ++ u32 aspm_support: 7; ++ u32 aspm_enabled: 7; ++ u32 aspm_capable: 7; ++ u32 aspm_default: 7; ++ int: 4; ++ u32 aspm_disable: 7; ++ u32 clkpm_capable: 1; ++ u32 clkpm_enabled: 1; ++ u32 clkpm_default: 1; ++ u32 clkpm_disable: 1; ++}; ++ ++struct aer_header_log_regs { ++ unsigned int dw0; ++ unsigned int dw1; ++ unsigned int dw2; ++ unsigned int dw3; ++}; ++ ++struct aer_err_info { ++ struct pci_dev *dev[5]; ++ int error_dev_num; ++ unsigned int id: 16; ++ unsigned int severity: 2; ++ unsigned int __pad1: 5; ++ unsigned int multi_error_valid: 1; ++ unsigned int first_error: 5; ++ unsigned int __pad2: 2; ++ unsigned int tlp_header_valid: 1; ++ unsigned int status; ++ unsigned int mask; ++ struct aer_header_log_regs tlp; ++}; ++ ++struct pci_slot_attribute { ++ struct attribute attr; ++ ssize_t (*show)(struct pci_slot *, char *); ++ ssize_t (*store)(struct pci_slot *, const char *, size_t); ++}; ++ ++struct aperture_range { ++ struct device *dev; ++ resource_size_t base; ++ resource_size_t size; ++ struct list_head lh; ++ void (*detach)(struct device *); ++}; ++ ++struct linux_logo { ++ int type; ++ unsigned int width; ++ unsigned int height; ++ unsigned int clutsize; ++ const unsigned char *clut; ++ const unsigned char *data; ++}; ++ ++enum { ++ FB_BLANK_UNBLANK = 0, ++ FB_BLANK_NORMAL = 1, ++ FB_BLANK_VSYNC_SUSPEND = 2, ++ FB_BLANK_HSYNC_SUSPEND = 3, ++ FB_BLANK_POWERDOWN = 4, ++}; ++ ++struct fb_event { ++ struct fb_info *info; ++ void *data; ++}; ++ ++struct logo_data { ++ int depth; ++ int needs_directpalette; ++ int needs_truepalette; ++ int needs_cmapreset; ++ const struct linux_logo *logo; ++}; ++ ++struct fb_fix_screeninfo32 { ++ char id[16]; ++ compat_caddr_t smem_start; ++ u32 smem_len; ++ u32 type; ++ u32 type_aux; ++ u32 visual; ++ u16 xpanstep; ++ u16 ypanstep; ++ u16 ywrapstep; ++ u32 line_length; ++ compat_caddr_t mmio_start; ++ u32 mmio_len; ++ u32 accel; ++ u16 reserved[3]; ++}; ++ ++struct fb_cmap32 { ++ u32 start; ++ u32 len; ++ compat_caddr_t red; ++ compat_caddr_t green; ++ compat_caddr_t blue; ++ compat_caddr_t transp; ++}; ++ ++enum { ++ KERNEL_PARAM_FL_UNSAFE = 1, ++ KERNEL_PARAM_FL_HWPARAM = 2, ++}; ++ ++struct fb_con2fbmap { ++ __u32 console; ++ __u32 framebuffer; ++}; ++ ++enum { ++ FBCON_LOGO_CANSHOW = -1, ++ FBCON_LOGO_DRAW = -2, ++ FBCON_LOGO_DONTSHOW = -3, ++}; ++ ++enum ipmi_addr_src { ++ SI_INVALID = 0, ++ SI_HOTMOD = 1, ++ SI_HARDCODED = 2, ++ SI_SPMI = 3, ++ SI_ACPI = 4, ++ SI_SMBIOS = 5, ++ SI_PCI = 6, ++ SI_DEVICETREE = 7, ++ SI_PLATFORM = 8, ++ SI_LAST = 9, ++}; ++ ++enum ipmi_plat_interface_type { ++ IPMI_PLAT_IF_SI = 0, ++ IPMI_PLAT_IF_SSIF = 1, ++}; ++ ++struct ipmi_plat_data { ++ enum ipmi_plat_interface_type iftype; ++ unsigned int type; ++ unsigned int space; ++ long unsigned int addr; ++ unsigned int regspacing; ++ unsigned int regsize; ++ unsigned int regshift; ++ unsigned int irq; ++ unsigned int slave_addr; ++ enum ipmi_addr_src addr_source; ++}; ++ ++enum si_type { ++ SI_TYPE_INVALID = 0, ++ SI_KCS = 1, ++ SI_SMIC = 2, ++ SI_BT = 3, ++ SI_TYPE_MAX = 4, ++}; ++ ++enum ipmi_addr_space { ++ IPMI_IO_ADDR_SPACE = 0, ++ IPMI_MEM_ADDR_SPACE = 1, ++}; ++ ++struct clk_bulk_data { ++ const char *id; ++ struct clk *clk; ++}; ++ ++struct clk_notifier { ++ struct clk *clk; ++ struct srcu_notifier_head notifier_head; ++ struct list_head node; ++}; ++ ++struct clk { ++ struct clk_core *core; ++ struct device *dev; ++ const char *dev_id; ++ const char *con_id; ++ long unsigned int min_rate; ++ long unsigned int max_rate; ++ unsigned int exclusive_count; ++ struct hlist_node clks_node; ++}; ++ ++struct clk_notifier_data { ++ struct clk *clk; ++ long unsigned int old_rate; ++ long unsigned int new_rate; ++}; ++ ++struct clk_parent_map; ++ ++struct clk_core { ++ const char *name; ++ const struct clk_ops *ops; ++ struct clk_hw *hw; ++ struct module *owner; ++ struct device *dev; ++ struct device_node *of_node; ++ struct clk_core *parent; ++ struct clk_parent_map *parents; ++ u8 num_parents; ++ u8 new_parent_index; ++ long unsigned int rate; ++ long unsigned int req_rate; ++ long unsigned int new_rate; ++ struct clk_core *new_parent; ++ struct clk_core *new_child; ++ long unsigned int flags; ++ bool orphan; ++ bool rpm_enabled; ++ unsigned int enable_count; ++ unsigned int prepare_count; ++ unsigned int protect_count; ++ long unsigned int min_rate; ++ long unsigned int max_rate; ++ long unsigned int accuracy; ++ int phase; ++ struct clk_duty duty; ++ struct hlist_head children; ++ struct hlist_node child_node; ++ struct hlist_head clks; ++ unsigned int notifier_count; ++ struct dentry *dentry; ++ struct hlist_node debug_node; ++ struct kref ref; ++}; ++ ++struct clk_parent_map { ++ const struct clk_hw *hw; ++ struct clk_core *core; ++ const char *fw_name; ++ const char *name; ++ int index; ++}; ++ ++struct trace_event_raw_clk { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_clk_rate { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ long unsigned int rate; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_clk_rate_range { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ long unsigned int min; ++ long unsigned int max; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_clk_parent { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ u32 __data_loc_pname; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_clk_phase { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ int phase; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_clk_duty_cycle { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ unsigned int num; ++ unsigned int den; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_clk_rate_request { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ u32 __data_loc_pname; ++ long unsigned int min; ++ long unsigned int max; ++ long unsigned int prate; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_clk { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_clk_rate { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_clk_rate_range { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_clk_parent { ++ u32 name; ++ u32 pname; ++}; ++ ++struct trace_event_data_offsets_clk_phase { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_clk_duty_cycle { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_clk_rate_request { ++ u32 name; ++ u32 pname; ++}; ++ ++typedef void (*btf_trace_clk_enable)(void *, struct clk_core *); ++ ++typedef void (*btf_trace_clk_enable_complete)(void *, struct clk_core *); ++ ++typedef void (*btf_trace_clk_disable)(void *, struct clk_core *); ++ ++typedef void (*btf_trace_clk_disable_complete)(void *, struct clk_core *); ++ ++typedef void (*btf_trace_clk_prepare)(void *, struct clk_core *); ++ ++typedef void (*btf_trace_clk_prepare_complete)(void *, struct clk_core *); ++ ++typedef void (*btf_trace_clk_unprepare)(void *, struct clk_core *); ++ ++typedef void (*btf_trace_clk_unprepare_complete)(void *, struct clk_core *); ++ ++typedef void (*btf_trace_clk_set_rate)(void *, struct clk_core *, long unsigned int); ++ ++typedef void (*btf_trace_clk_set_rate_complete)(void *, struct clk_core *, long unsigned int); ++ ++typedef void (*btf_trace_clk_set_min_rate)(void *, struct clk_core *, long unsigned int); ++ ++typedef void (*btf_trace_clk_set_max_rate)(void *, struct clk_core *, long unsigned int); ++ ++typedef void (*btf_trace_clk_set_rate_range)(void *, struct clk_core *, long unsigned int, long unsigned int); ++ ++typedef void (*btf_trace_clk_set_parent)(void *, struct clk_core *, struct clk_core *); ++ ++typedef void (*btf_trace_clk_set_parent_complete)(void *, struct clk_core *, struct clk_core *); ++ ++typedef void (*btf_trace_clk_set_phase)(void *, struct clk_core *, int); ++ ++typedef void (*btf_trace_clk_set_phase_complete)(void *, struct clk_core *, int); ++ ++typedef void (*btf_trace_clk_set_duty_cycle)(void *, struct clk_core *, struct clk_duty *); ++ ++typedef void (*btf_trace_clk_set_duty_cycle_complete)(void *, struct clk_core *, struct clk_duty *); ++ ++typedef void (*btf_trace_clk_rate_request_start)(void *, struct clk_rate_request *); ++ ++typedef void (*btf_trace_clk_rate_request_done)(void *, struct clk_rate_request *); ++ ++struct clk_notifier_devres { ++ struct clk *clk; ++ struct notifier_block *nb; ++}; ++ ++struct clk_gate { ++ struct clk_hw hw; ++ void *reg; ++ u8 bit_idx; ++ u8 flags; ++ spinlock_t *lock; ++}; ++ ++struct u32_fract { ++ __u32 numerator; ++ __u32 denominator; ++}; ++ ++struct clk_fractional_divider { ++ struct clk_hw hw; ++ void *reg; ++ u8 mshift; ++ u8 mwidth; ++ u8 nshift; ++ u8 nwidth; ++ u8 flags; ++ void (*approximation)(struct clk_hw *, long unsigned int, long unsigned int *, long unsigned int *, long unsigned int *); ++ spinlock_t *lock; ++}; ++ ++struct clk_gpio { ++ struct clk_hw hw; ++ struct gpio_desc *gpiod; ++}; ++ ++struct pre_voltage_change_data { ++ long unsigned int old_uV; ++ long unsigned int min_uV; ++ long unsigned int max_uV; ++}; ++ ++struct regulator_coupler { ++ struct list_head list; ++ int (*attach_regulator)(struct regulator_coupler *, struct regulator_dev *); ++ int (*detach_regulator)(struct regulator_coupler *, struct regulator_dev *); ++ int (*balance_voltage)(struct regulator_coupler *, struct regulator_dev *, suspend_state_t); ++}; ++ ++enum regulator_status { ++ REGULATOR_STATUS_OFF = 0, ++ REGULATOR_STATUS_ON = 1, ++ REGULATOR_STATUS_ERROR = 2, ++ REGULATOR_STATUS_FAST = 3, ++ REGULATOR_STATUS_NORMAL = 4, ++ REGULATOR_STATUS_IDLE = 5, ++ REGULATOR_STATUS_STANDBY = 6, ++ REGULATOR_STATUS_BYPASS = 7, ++ REGULATOR_STATUS_UNDEFINED = 8, ++}; ++ ++enum regulator_detection_severity { ++ REGULATOR_SEVERITY_PROT = 0, ++ REGULATOR_SEVERITY_ERR = 1, ++ REGULATOR_SEVERITY_WARN = 2, ++}; ++ ++struct regulator_enable_gpio { ++ struct list_head list; ++ struct gpio_desc *gpiod; ++ u32 enable_count; ++ u32 request_count; ++}; ++ ++enum regulator_active_discharge { ++ REGULATOR_ACTIVE_DISCHARGE_DEFAULT = 0, ++ REGULATOR_ACTIVE_DISCHARGE_DISABLE = 1, ++ REGULATOR_ACTIVE_DISCHARGE_ENABLE = 2, ++}; ++ ++struct trace_event_raw_regulator_basic { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_regulator_range { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ int min; ++ int max; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_regulator_value { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ unsigned int val; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_regulator_basic { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_regulator_range { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_regulator_value { ++ u32 name; ++}; ++ ++typedef void (*btf_trace_regulator_enable)(void *, const char *); ++ ++typedef void (*btf_trace_regulator_enable_delay)(void *, const char *); ++ ++typedef void (*btf_trace_regulator_enable_complete)(void *, const char *); ++ ++typedef void (*btf_trace_regulator_disable)(void *, const char *); ++ ++typedef void (*btf_trace_regulator_disable_complete)(void *, const char *); ++ ++typedef void (*btf_trace_regulator_bypass_enable)(void *, const char *); ++ ++typedef void (*btf_trace_regulator_bypass_enable_complete)(void *, const char *); ++ ++typedef void (*btf_trace_regulator_bypass_disable)(void *, const char *); ++ ++typedef void (*btf_trace_regulator_bypass_disable_complete)(void *, const char *); ++ ++typedef void (*btf_trace_regulator_set_voltage)(void *, const char *, int, int); ++ ++typedef void (*btf_trace_regulator_set_voltage_complete)(void *, const char *, unsigned int); ++ ++struct regulator_map { ++ struct list_head list; ++ const char *dev_name; ++ const char *supply; ++ struct regulator_dev *regulator; ++}; ++ ++struct regulator_supply_alias { ++ struct list_head list; ++ struct device *src_dev; ++ const char *src_supply; ++ struct device *alias_dev; ++ const char *alias_supply; ++}; ++ ++struct summary_data { ++ struct seq_file *s; ++ struct regulator_dev *parent; ++ int level; ++}; ++ ++struct summary_lock_data { ++ struct ww_acquire_ctx *ww_ctx; ++ struct regulator_dev **new_contended_rdev; ++ struct regulator_dev **old_contended_rdev; ++}; ++ ++struct serial_struct32 { ++ compat_int_t type; ++ compat_int_t line; ++ compat_uint_t port; ++ compat_int_t irq; ++ compat_int_t flags; ++ compat_int_t xmit_fifo_size; ++ compat_int_t custom_divisor; ++ compat_int_t baud_base; ++ short unsigned int close_delay; ++ char io_type; ++ char reserved_char; ++ compat_int_t hub6; ++ short unsigned int closing_wait; ++ short unsigned int closing_wait2; ++ compat_uint_t iomem_base; ++ short unsigned int iomem_reg_shift; ++ unsigned int port_high; ++ compat_int_t reserved; ++}; ++ ++struct n_tty_data { ++ size_t read_head; ++ size_t commit_head; ++ size_t canon_head; ++ size_t echo_head; ++ size_t echo_commit; ++ size_t echo_mark; ++ long unsigned int char_map[4]; ++ long unsigned int overrun_time; ++ int num_overrun; ++ bool no_room; ++ unsigned char lnext: 1; ++ unsigned char erasing: 1; ++ unsigned char raw: 1; ++ unsigned char real_raw: 1; ++ unsigned char icanon: 1; ++ unsigned char push: 1; ++ char read_buf[4096]; ++ long unsigned int read_flags[64]; ++ unsigned char echo_buf[4096]; ++ size_t read_tail; ++ size_t line_start; ++ size_t lookahead_count; ++ unsigned int column; ++ unsigned int canon_column; ++ size_t echo_tail; ++ struct mutex atomic_read_lock; ++ struct mutex output_lock; ++}; ++ ++enum { ++ ERASE = 0, ++ WERASE = 1, ++ KILL = 2, ++}; ++ ++struct ldsem_waiter { ++ struct list_head list; ++ struct task_struct *task; ++}; ++ ++struct unipair { ++ short unsigned int unicode; ++ short unsigned int fontpos; ++}; ++ ++struct unimapdesc { ++ short unsigned int entry_ct; ++ struct unipair *entries; ++}; ++ ++struct console_font_op { ++ unsigned int op; ++ unsigned int flags; ++ unsigned int width; ++ unsigned int height; ++ unsigned int charcount; ++ unsigned char *data; ++}; ++ ++struct vt_stat { ++ short unsigned int v_active; ++ short unsigned int v_signal; ++ short unsigned int v_state; ++}; ++ ++struct vt_sizes { ++ short unsigned int v_rows; ++ short unsigned int v_cols; ++ short unsigned int v_scrollsize; ++}; ++ ++struct vt_consize { ++ short unsigned int v_rows; ++ short unsigned int v_cols; ++ short unsigned int v_vlin; ++ short unsigned int v_clin; ++ short unsigned int v_vcol; ++ short unsigned int v_ccol; ++}; ++ ++struct vt_event { ++ unsigned int event; ++ unsigned int oldev; ++ unsigned int newev; ++ unsigned int pad[4]; ++}; ++ ++struct vt_setactivate { ++ unsigned int console; ++ struct vt_mode mode; ++}; ++ ++struct vt_event_wait { ++ struct list_head list; ++ struct vt_event event; ++ int done; ++}; ++ ++struct compat_console_font_op { ++ compat_uint_t op; ++ compat_uint_t flags; ++ compat_uint_t width; ++ compat_uint_t height; ++ compat_uint_t charcount; ++ compat_caddr_t data; ++}; ++ ++struct compat_unimapdesc { ++ short unsigned int entry_ct; ++ compat_caddr_t entries; ++}; ++ ++struct tiocl_selection { ++ short unsigned int xs; ++ short unsigned int ys; ++ short unsigned int xe; ++ short unsigned int ye; ++ short unsigned int sel_mode; ++}; ++ ++struct vc_selection { ++ struct mutex lock; ++ struct vc_data *cons; ++ char *buffer; ++ unsigned int buf_len; ++ volatile int start; ++ int end; ++}; ++ ++enum translation_map { ++ LAT1_MAP = 0, ++ GRAF_MAP = 1, ++ IBMPC_MAP = 2, ++ USER_MAP = 3, ++ FIRST_MAP = 0, ++ LAST_MAP = 3, ++}; ++ ++struct con_driver { ++ const struct consw *con; ++ const char *desc; ++ struct device *dev; ++ int node; ++ int first; ++ int last; ++ int flag; ++}; ++ ++enum { ++ blank_off = 0, ++ blank_normal_wait = 1, ++ blank_vesa_wait = 2, ++}; ++ ++enum { ++ EPecma = 0, ++ EPdec = 1, ++ EPeq = 2, ++ EPgt = 3, ++ EPlt = 4, ++}; ++ ++struct rgb { ++ u8 r; ++ u8 g; ++ u8 b; ++}; ++ ++enum { ++ ESnormal = 0, ++ ESesc = 1, ++ ESsquare = 2, ++ ESgetpars = 3, ++ ESfunckey = 4, ++ EShash = 5, ++ ESsetG0 = 6, ++ ESsetG1 = 7, ++ ESpercent = 8, ++ EScsiignore = 9, ++ ESnonstd = 10, ++ ESpalette = 11, ++ ESosc = 12, ++ ESapc = 13, ++ ESpm = 14, ++ ESdcs = 15, ++}; ++ ++struct interval { ++ uint32_t first; ++ uint32_t last; ++}; ++ ++struct vc_draw_region { ++ long unsigned int from; ++ long unsigned int to; ++ int x; ++}; ++ ++struct iucv_tty_msg { ++ u8 version; ++ u8 type; ++ u16 datalen; ++ u8 data[0]; ++}; ++ ++enum iucv_state_t { ++ IUCV_DISCONN = 0, ++ IUCV_CONNECTED = 1, ++ IUCV_SEVERED = 2, ++}; ++ ++enum tty_state_t { ++ TTY_CLOSED = 0, ++ TTY_OPENED = 1, ++}; ++ ++struct hvc_iucv_private { ++ struct hvc_struct *hvc; ++ u8 srv_name[8]; ++ unsigned char is_console; ++ enum iucv_state_t iucv_state; ++ enum tty_state_t tty_state; ++ struct iucv_path *path; ++ spinlock_t lock; ++ void *sndbuf; ++ size_t sndbuf_len; ++ struct delayed_work sndbuf_work; ++ wait_queue_head_t sndbuf_waitq; ++ struct list_head tty_outqueue; ++ struct list_head tty_inqueue; ++ struct device *dev; ++ u8 info_path[16]; ++}; ++ ++struct iucv_tty_buffer { ++ struct list_head list; ++ struct iucv_message msg; ++ size_t offset; ++ struct iucv_tty_msg *mbuf; ++}; ++ ++struct serdev_device; ++ ++struct serdev_device_ops { ++ int (*receive_buf)(struct serdev_device *, const unsigned char *, size_t); ++ void (*write_wakeup)(struct serdev_device *); ++}; ++ ++struct serdev_controller; ++ ++struct serdev_device { ++ struct device dev; ++ int nr; ++ struct serdev_controller *ctrl; ++ const struct serdev_device_ops *ops; ++ struct completion write_comp; ++ struct mutex write_lock; ++}; ++ ++struct serdev_controller_ops; ++ ++struct serdev_controller { ++ struct device dev; ++ unsigned int nr; ++ struct serdev_device *serdev; ++ const struct serdev_controller_ops *ops; ++}; ++ ++enum serdev_parity { ++ SERDEV_PARITY_NONE = 0, ++ SERDEV_PARITY_EVEN = 1, ++ SERDEV_PARITY_ODD = 2, ++}; ++ ++struct serdev_controller_ops { ++ int (*write_buf)(struct serdev_controller *, const unsigned char *, size_t); ++ void (*write_flush)(struct serdev_controller *); ++ int (*write_room)(struct serdev_controller *); ++ int (*open)(struct serdev_controller *); ++ void (*close)(struct serdev_controller *); ++ void (*set_flow_control)(struct serdev_controller *, bool); ++ int (*set_parity)(struct serdev_controller *, enum serdev_parity); ++ unsigned int (*set_baudrate)(struct serdev_controller *, unsigned int); ++ void (*wait_until_sent)(struct serdev_controller *, long int); ++ int (*get_tiocm)(struct serdev_controller *); ++ int (*set_tiocm)(struct serdev_controller *, unsigned int, unsigned int); ++ int (*break_ctl)(struct serdev_controller *, unsigned int); ++}; ++ ++struct serport { ++ struct tty_port *port; ++ struct tty_struct *tty; ++ struct tty_driver *tty_drv; ++ int tty_idx; ++ long unsigned int flags; ++}; ++ ++struct timer_rand_state { ++ long unsigned int last_time; ++ long int last_delta; ++ long int last_delta2; ++}; ++ ++enum chacha_constants { ++ CHACHA_CONSTANT_EXPA = 1634760805, ++ CHACHA_CONSTANT_ND_3 = 857760878, ++ CHACHA_CONSTANT_2_BY = 2036477234, ++ CHACHA_CONSTANT_TE_K = 1797285236, ++}; ++ ++enum { ++ CRNG_EMPTY = 0, ++ CRNG_EARLY = 1, ++ CRNG_READY = 2, ++}; ++ ++enum { ++ CRNG_RESEED_START_INTERVAL = 100, ++ CRNG_RESEED_INTERVAL = 6000, ++}; ++ ++struct crng { ++ u8 key[32]; ++ long unsigned int generation; ++ local_lock_t lock; ++}; ++ ++struct batch_u8 { ++ u8 entropy[96]; ++ local_lock_t lock; ++ long unsigned int generation; ++ unsigned int position; ++}; ++ ++struct batch_u16 { ++ u16 entropy[48]; ++ local_lock_t lock; ++ long unsigned int generation; ++ unsigned int position; ++}; ++ ++struct batch_u32 { ++ u32 entropy[24]; ++ local_lock_t lock; ++ long unsigned int generation; ++ unsigned int position; ++}; ++ ++struct batch_u64 { ++ u64 entropy[12]; ++ local_lock_t lock; ++ long unsigned int generation; ++ unsigned int position; ++}; ++ ++enum { ++ POOL_BITS = 256, ++ POOL_READY_BITS = 256, ++ POOL_EARLY_BITS = 128, ++}; ++ ++struct fast_pool { ++ long unsigned int pool[4]; ++ long unsigned int last; ++ unsigned int count; ++ struct timer_list mix; ++}; ++ ++struct entropy_timer_state { ++ long unsigned int entropy; ++ struct timer_list timer; ++ atomic_t samples; ++ unsigned int samples_per_bit; ++}; ++ ++enum { ++ NUM_TRIAL_SAMPLES = 8192, ++ MAX_SAMPLES_PER_BIT = 6, ++}; ++ ++enum { ++ MIX_INFLIGHT = 2147483648, ++}; ++ ++enum TPM_OPS_FLAGS { ++ TPM_OPS_AUTO_STARTUP = 1, ++}; ++ ++enum tpm_timeout { ++ TPM_TIMEOUT = 5, ++ TPM_TIMEOUT_RETRY = 100, ++ TPM_TIMEOUT_RANGE_US = 300, ++ TPM_TIMEOUT_POLL = 1, ++ TPM_TIMEOUT_USECS_MIN = 100, ++ TPM_TIMEOUT_USECS_MAX = 500, ++}; ++ ++struct stclear_flags_t { ++ __be16 tag; ++ u8 deactivated; ++ u8 disableForceClear; ++ u8 physicalPresence; ++ u8 physicalPresenceLock; ++ u8 bGlobalLock; ++} __attribute__((packed)); ++ ++struct tpm1_version { ++ u8 major; ++ u8 minor; ++ u8 rev_major; ++ u8 rev_minor; ++}; ++ ++struct tpm1_version2 { ++ __be16 tag; ++ struct tpm1_version version; ++}; ++ ++struct timeout_t { ++ __be32 a; ++ __be32 b; ++ __be32 c; ++ __be32 d; ++}; ++ ++struct duration_t { ++ __be32 tpm_short; ++ __be32 tpm_medium; ++ __be32 tpm_long; ++}; ++ ++struct permanent_flags_t { ++ __be16 tag; ++ u8 disable; ++ u8 ownership; ++ u8 deactivated; ++ u8 readPubek; ++ u8 disableOwnerClear; ++ u8 allowMaintenance; ++ u8 physicalPresenceLifetimeLock; ++ u8 physicalPresenceHWEnable; ++ u8 physicalPresenceCMDEnable; ++ u8 CEKPUsed; ++ u8 TPMpost; ++ u8 TPMpostLock; ++ u8 FIPS; ++ u8 operator; ++ u8 enableRevokeEK; ++ u8 nvLocked; ++ u8 readSRKPub; ++ u8 tpmEstablished; ++ u8 maintenanceDone; ++ u8 disableFullDALogicInfo; ++}; ++ ++typedef union { ++ struct permanent_flags_t perm_flags; ++ struct stclear_flags_t stclear_flags; ++ __u8 owned; ++ __be32 num_pcrs; ++ struct tpm1_version version1; ++ struct tpm1_version2 version2; ++ __be32 manufacturer_id; ++ struct timeout_t timeout; ++ struct duration_t duration; ++} cap_t; ++ ++enum tpm_capabilities { ++ TPM_CAP_FLAG = 4, ++ TPM_CAP_PROP = 5, ++ TPM_CAP_VERSION_1_1 = 6, ++ TPM_CAP_VERSION_1_2 = 26, ++}; ++ ++enum tpm_sub_capabilities { ++ TPM_CAP_PROP_PCR = 257, ++ TPM_CAP_PROP_MANUFACTURER = 259, ++ TPM_CAP_FLAG_PERM = 264, ++ TPM_CAP_FLAG_VOL = 265, ++ TPM_CAP_PROP_OWNER = 273, ++ TPM_CAP_PROP_TIS_TIMEOUT = 277, ++ TPM_CAP_PROP_TIS_DURATION = 288, ++}; ++ ++struct tpm1_get_random_out { ++ __be32 rng_data_len; ++ u8 rng_data[128]; ++}; ++ ++enum tpm2_handle_types { ++ TPM2_HT_HMAC_SESSION = 33554432, ++ TPM2_HT_POLICY_SESSION = 50331648, ++ TPM2_HT_TRANSIENT = 2147483648, ++}; ++ ++struct tpm2_context { ++ __be64 sequence; ++ __be32 saved_handle; ++ __be32 hierarchy; ++ __be16 blob_size; ++} __attribute__((packed)); ++ ++struct tpm2_cap_handles { ++ u8 more_data; ++ __be32 capability; ++ __be32 count; ++ __be32 handles[0]; ++} __attribute__((packed)); ++ ++struct tpm_readpubek_out { ++ u8 algorithm[4]; ++ u8 encscheme[2]; ++ u8 sigscheme[2]; ++ __be32 paramsize; ++ u8 parameters[12]; ++ __be32 keysize; ++ u8 modulus[256]; ++ u8 checksum[20]; ++}; ++ ++struct tpm_pcr_attr { ++ int alg_id; ++ int pcr; ++ struct device_attribute attr; ++}; ++ ++struct tcg_efi_specid_event_algs { ++ u16 alg_id; ++ u16 digest_size; ++}; ++ ++struct tcg_efi_specid_event_head { ++ u8 signature[16]; ++ u32 platform_class; ++ u8 spec_version_minor; ++ u8 spec_version_major; ++ u8 spec_errata; ++ u8 uintnsize; ++ u32 num_algs; ++ struct tcg_efi_specid_event_algs digest_sizes[0]; ++}; ++ ++struct tcg_pcr_event { ++ u32 pcr_idx; ++ u32 event_type; ++ u8 digest[20]; ++ u32 event_size; ++ u8 event[0]; ++}; ++ ++struct tcg_event_field { ++ u32 event_size; ++ u8 event[0]; ++}; ++ ++struct tcg_pcr_event2_head { ++ u32 pcr_idx; ++ u32 event_type; ++ u32 count; ++ struct tpm_digest digests[0]; ++}; ++ ++enum iommu_resv_type { ++ IOMMU_RESV_DIRECT = 0, ++ IOMMU_RESV_DIRECT_RELAXABLE = 1, ++ IOMMU_RESV_RESERVED = 2, ++ IOMMU_RESV_MSI = 3, ++ IOMMU_RESV_SW_MSI = 4, ++}; ++ ++struct iommu_resv_region { ++ struct list_head list; ++ phys_addr_t start; ++ size_t length; ++ int prot; ++ enum iommu_resv_type type; ++ void (*free)(struct device *, struct iommu_resv_region *); ++}; ++ ++struct s390_domain { ++ struct iommu_domain domain; ++ struct list_head devices; ++ long unsigned int *dma_table; ++ spinlock_t list_lock; ++ struct callback_head rcu; ++}; ++ ++struct drm_mode_atomic { ++ __u32 flags; ++ __u32 count_objs; ++ __u64 objs_ptr; ++ __u64 count_props_ptr; ++ __u64 props_ptr; ++ __u64 prop_values_ptr; ++ __u64 reserved; ++ __u64 user_data; ++}; ++ ++struct drm_writeback_job { ++ struct drm_writeback_connector *connector; ++ bool prepared; ++ struct work_struct cleanup_work; ++ struct list_head list_entry; ++ struct drm_framebuffer *fb; ++ struct dma_fence *out_fence; ++ void *priv; ++}; ++ ++struct drm_writeback_connector { ++ struct drm_connector base; ++ struct drm_encoder encoder; ++ struct drm_property_blob *pixel_formats_blob_ptr; ++ spinlock_t job_lock; ++ struct list_head job_queue; ++ unsigned int fence_context; ++ spinlock_t fence_lock; ++ long unsigned int fence_seqno; ++ char timeline_name[32]; ++}; ++ ++struct drm_out_fence_state { ++ s32 *out_fence_ptr; ++ struct sync_file *sync_file; ++ int fd; ++}; ++ ++struct drm_mode_fb_cmd { ++ __u32 fb_id; ++ __u32 width; ++ __u32 height; ++ __u32 pitch; ++ __u32 bpp; ++ __u32 depth; ++ __u32 handle; ++}; ++ ++struct drm_mode_get_connector { ++ __u64 encoders_ptr; ++ __u64 modes_ptr; ++ __u64 props_ptr; ++ __u64 prop_values_ptr; ++ __u32 count_modes; ++ __u32 count_props; ++ __u32 count_encoders; ++ __u32 encoder_id; ++ __u32 connector_id; ++ __u32 connector_type; ++ __u32 connector_type_id; ++ __u32 connection; ++ __u32 mm_width; ++ __u32 mm_height; ++ __u32 subpixel; ++ __u32 pad; ++}; ++ ++struct drm_mode_connector_set_property { ++ __u64 value; ++ __u32 prop_id; ++ __u32 connector_id; ++}; ++ ++struct drm_mode_obj_set_property { ++ __u64 value; ++ __u32 prop_id; ++ __u32 obj_id; ++ __u32 obj_type; ++}; ++ ++struct display_timing; ++ ++struct drm_panel; ++ ++struct drm_panel_funcs { ++ int (*prepare)(struct drm_panel *); ++ int (*enable)(struct drm_panel *); ++ int (*disable)(struct drm_panel *); ++ int (*unprepare)(struct drm_panel *); ++ int (*get_modes)(struct drm_panel *, struct drm_connector *); ++ enum drm_panel_orientation (*get_orientation)(struct drm_panel *); ++ int (*get_timings)(struct drm_panel *, unsigned int, struct display_timing *); ++ void (*debugfs_init)(struct drm_panel *, struct dentry *); ++}; ++ ++struct backlight_device; ++ ++struct drm_panel { ++ struct device *dev; ++ struct backlight_device *backlight; ++ const struct drm_panel_funcs *funcs; ++ int connector_type; ++ struct list_head list; ++ bool prepare_prev_first; ++}; ++ ++struct drm_conn_prop_enum_list { ++ int type; ++ const char *name; ++ struct ida ida; ++}; ++ ++struct drm_mode_map_dumb { ++ __u32 handle; ++ __u32 pad; ++ __u64 offset; ++}; ++ ++struct drm_mode_destroy_dumb { ++ __u32 handle; ++}; ++ ++struct displayid_block { ++ u8 tag; ++ u8 rev; ++ u8 num_bytes; ++}; ++ ++struct displayid_tiled_block { ++ struct displayid_block base; ++ u8 tile_cap; ++ u8 topo[3]; ++ u8 tile_size[4]; ++ u8 tile_pixel_bezel[5]; ++ u8 topology_id[8]; ++}; ++ ++struct displayid_detailed_timings_1 { ++ u8 pixel_clock[3]; ++ u8 flags; ++ u8 hactive[2]; ++ u8 hblank[2]; ++ u8 hsync[2]; ++ u8 hsw[2]; ++ u8 vactive[2]; ++ u8 vblank[2]; ++ u8 vsync[2]; ++ u8 vsw[2]; ++}; ++ ++struct displayid_detailed_timing_block { ++ struct displayid_block base; ++ struct displayid_detailed_timings_1 timings[0]; ++}; ++ ++struct displayid_vesa_vendor_specific_block { ++ struct displayid_block base; ++ u8 oui[3]; ++ u8 data_structure_type; ++ u8 mso; ++}; ++ ++struct displayid_iter { ++ const struct drm_edid *drm_edid; ++ const u8 *section; ++ int length; ++ int idx; ++ int ext_index; ++ u8 version; ++ u8 primary_use; ++}; ++ ++struct cea_sad { ++ u8 format; ++ u8 channels; ++ u8 freq; ++ u8 byte2; ++}; ++ ++struct detailed_mode_closure { ++ struct drm_connector *connector; ++ const struct drm_edid *drm_edid; ++ bool preferred; ++ int modes; ++}; ++ ++struct edid_quirk { ++ u32 panel_id; ++ u32 quirks; ++}; ++ ++struct minimode { ++ short int w; ++ short int h; ++ short int r; ++ short int rb; ++}; ++ ++struct drm_edid_iter { ++ const struct drm_edid *drm_edid; ++ int index; ++}; ++ ++enum edid_block_status { ++ EDID_BLOCK_OK = 0, ++ EDID_BLOCK_READ_FAIL = 1, ++ EDID_BLOCK_NULL = 2, ++ EDID_BLOCK_ZERO = 3, ++ EDID_BLOCK_HEADER_CORRUPT = 4, ++ EDID_BLOCK_HEADER_REPAIR = 5, ++ EDID_BLOCK_HEADER_FIXED = 6, ++ EDID_BLOCK_CHECKSUM = 7, ++ EDID_BLOCK_VERSION = 8, ++}; ++ ++typedef int read_block_fn___2(void *, u8 *, unsigned int, size_t); ++ ++typedef void detailed_cb(const struct detailed_timing *, void *); ++ ++struct stereo_mandatory_mode { ++ int width; ++ int height; ++ int vrefresh; ++ unsigned int flags; ++}; ++ ++struct cea_db_iter { ++ struct drm_edid_iter edid_iter; ++ struct displayid_iter displayid_iter; ++ const u8 *collection; ++ int index; ++ int end; ++}; ++ ++struct cea_db { ++ u8 tag_length; ++ u8 data[0]; ++}; ++ ++struct drm_stats { ++ long unsigned int count; ++ struct { ++ long unsigned int value; ++ enum drm_stat_type type; ++ } data[15]; ++}; ++ ++struct drm_set_version { ++ int drm_di_major; ++ int drm_di_minor; ++ int drm_dd_major; ++ int drm_dd_minor; ++}; ++ ++struct drm_get_cap { ++ __u64 capability; ++ __u64 value; ++}; ++ ++struct drm_set_client_cap { ++ __u64 capability; ++ __u64 value; ++}; ++ ++enum drm_minor_type { ++ DRM_MINOR_PRIMARY = 0, ++ DRM_MINOR_CONTROL = 1, ++ DRM_MINOR_RENDER = 2, ++ DRM_MINOR_ACCEL = 32, ++}; ++ ++struct drm_mode_create_lease { ++ __u64 object_ids; ++ __u32 object_count; ++ __u32 flags; ++ __u32 lessee_id; ++ __u32 fd; ++}; ++ ++struct drm_mode_list_lessees { ++ __u32 count_lessees; ++ __u32 pad; ++ __u64 lessees_ptr; ++}; ++ ++struct drm_mode_get_lease { ++ __u32 count_objects; ++ __u32 pad; ++ __u64 objects_ptr; ++}; ++ ++struct drm_mode_revoke_lease { ++ __u32 lessee_id; ++}; ++ ++struct drm_mm_scan { ++ struct drm_mm *mm; ++ u64 size; ++ u64 alignment; ++ u64 remainder_mask; ++ u64 range_start; ++ u64 range_end; ++ u64 hit_start; ++ u64 hit_end; ++ long unsigned int color; ++ enum drm_mm_insert_mode mode; ++}; ++ ++struct drm_mode_card_res { ++ __u64 fb_id_ptr; ++ __u64 crtc_id_ptr; ++ __u64 connector_id_ptr; ++ __u64 encoder_id_ptr; ++ __u32 count_fbs; ++ __u32 count_crtcs; ++ __u32 count_connectors; ++ __u32 count_encoders; ++ __u32 min_width; ++ __u32 max_width; ++ __u32 min_height; ++ __u32 max_height; ++}; ++ ++struct drm_mode_set_plane { ++ __u32 plane_id; ++ __u32 crtc_id; ++ __u32 fb_id; ++ __u32 flags; ++ __s32 crtc_x; ++ __s32 crtc_y; ++ __u32 crtc_w; ++ __u32 crtc_h; ++ __u32 src_x; ++ __u32 src_y; ++ __u32 src_h; ++ __u32 src_w; ++}; ++ ++struct drm_mode_get_plane { ++ __u32 plane_id; ++ __u32 crtc_id; ++ __u32 fb_id; ++ __u32 possible_crtcs; ++ __u32 gamma_size; ++ __u32 count_format_types; ++ __u64 format_type_ptr; ++}; ++ ++struct drm_mode_get_plane_res { ++ __u64 plane_id_ptr; ++ __u32 count_planes; ++}; ++ ++struct drm_mode_cursor { ++ __u32 flags; ++ __u32 crtc_id; ++ __s32 x; ++ __s32 y; ++ __u32 width; ++ __u32 height; ++ __u32 handle; ++}; ++ ++struct drm_mode_cursor2 { ++ __u32 flags; ++ __u32 crtc_id; ++ __s32 x; ++ __s32 y; ++ __u32 width; ++ __u32 height; ++ __u32 handle; ++ __s32 hot_x; ++ __s32 hot_y; ++}; ++ ++struct drm_mode_crtc_page_flip_target { ++ __u32 crtc_id; ++ __u32 fb_id; ++ __u32 flags; ++ __u32 sequence; ++ __u64 user_data; ++}; ++ ++struct drm_format_modifier_blob { ++ __u32 version; ++ __u32 flags; ++ __u32 count_formats; ++ __u32 formats_offset; ++ __u32 count_modifiers; ++ __u32 modifiers_offset; ++}; ++ ++struct drm_format_modifier { ++ __u64 formats; ++ __u32 offset; ++ __u32 pad; ++ __u64 modifier; ++}; ++ ++struct drm_mode_rect { ++ __s32 x1; ++ __s32 y1; ++ __s32 x2; ++ __s32 y2; ++}; ++ ++struct drm_print_iterator { ++ void *data; ++ ssize_t start; ++ ssize_t remain; ++ ssize_t offset; ++}; ++ ++struct drm_mode_property_enum { ++ __u64 value; ++ char name[32]; ++}; ++ ++struct drm_mode_get_property { ++ __u64 values_ptr; ++ __u64 enum_blob_ptr; ++ __u32 prop_id; ++ __u32 flags; ++ char name[32]; ++ __u32 count_values; ++ __u32 count_enum_blobs; ++}; ++ ++struct drm_mode_get_blob { ++ __u32 blob_id; ++ __u32 length; ++ __u64 data; ++}; ++ ++struct drm_mode_create_blob { ++ __u64 data; ++ __u32 length; ++ __u32 blob_id; ++}; ++ ++struct drm_mode_destroy_blob { ++ __u32 blob_id; ++}; ++ ++struct drm_property_enum { ++ uint64_t value; ++ struct list_head head; ++ char name[32]; ++}; ++ ++struct drm_modeset_ctl { ++ __u32 crtc; ++ __u32 cmd; ++}; ++ ++struct drm_crtc_get_sequence { ++ __u32 crtc_id; ++ __u32 active; ++ __u64 sequence; ++ __s64 sequence_ns; ++}; ++ ++struct drm_crtc_queue_sequence { ++ __u32 crtc_id; ++ __u32 flags; ++ __u64 sequence; ++ __u64 user_data; ++}; ++ ++typedef bool (*drm_vblank_get_scanout_position_func)(struct drm_crtc *, bool, int *, int *, ktime_t *, ktime_t *, const struct drm_display_mode *); ++ ++struct drm_vblank_work { ++ struct kthread_work base; ++ struct drm_vblank_crtc *vblank; ++ u64 count; ++ int cancelling; ++ struct list_head node; ++}; ++ ++enum backlight_type { ++ BACKLIGHT_RAW = 1, ++ BACKLIGHT_PLATFORM = 2, ++ BACKLIGHT_FIRMWARE = 3, ++ BACKLIGHT_TYPE_MAX = 4, ++}; ++ ++enum backlight_scale { ++ BACKLIGHT_SCALE_UNKNOWN = 0, ++ BACKLIGHT_SCALE_LINEAR = 1, ++ BACKLIGHT_SCALE_NON_LINEAR = 2, ++}; ++ ++struct backlight_ops { ++ unsigned int options; ++ int (*update_status)(struct backlight_device *); ++ int (*get_brightness)(struct backlight_device *); ++ int (*check_fb)(struct backlight_device *, struct fb_info *); ++}; ++ ++struct backlight_properties { ++ int brightness; ++ int max_brightness; ++ int power; ++ int fb_blank; ++ enum backlight_type type; ++ unsigned int state; ++ enum backlight_scale scale; ++}; ++ ++struct backlight_device { ++ struct backlight_properties props; ++ struct mutex update_lock; ++ struct mutex ops_lock; ++ const struct backlight_ops *ops; ++ struct notifier_block fb_notif; ++ struct list_head entry; ++ struct device dev; ++ bool fb_bl_on[32]; ++ int use_count; ++}; ++ ++struct drm_encoder_slave_funcs { ++ void (*set_config)(struct drm_encoder *, void *); ++ void (*destroy)(struct drm_encoder *); ++ void (*dpms)(struct drm_encoder *, int); ++ void (*save)(struct drm_encoder *); ++ void (*restore)(struct drm_encoder *); ++ bool (*mode_fixup)(struct drm_encoder *, const struct drm_display_mode *, struct drm_display_mode *); ++ int (*mode_valid)(struct drm_encoder *, struct drm_display_mode *); ++ void (*mode_set)(struct drm_encoder *, struct drm_display_mode *, struct drm_display_mode *); ++ enum drm_connector_status (*detect)(struct drm_encoder *, struct drm_connector *); ++ int (*get_modes)(struct drm_encoder *, struct drm_connector *); ++ int (*create_resources)(struct drm_encoder *, struct drm_connector *); ++ int (*set_property)(struct drm_encoder *, struct drm_connector *, struct drm_property *, uint64_t); ++}; ++ ++struct drm_encoder_slave { ++ struct drm_encoder base; ++ const struct drm_encoder_slave_funcs *slave_funcs; ++ void *slave_priv; ++ void *bus_priv; ++}; ++ ++struct drm_i2c_encoder_driver { ++ struct i2c_driver i2c_driver; ++ int (*encoder_init)(struct i2c_client *, struct drm_device *, struct drm_encoder_slave *); ++}; ++ ++struct drm_afbc_framebuffer { ++ struct drm_framebuffer base; ++ u32 block_width; ++ u32 block_height; ++ u32 aligned_width; ++ u32 aligned_height; ++ u32 offset; ++ u32 afbc_size; ++}; ++ ++struct ewma_psr_time { ++ long unsigned int internal; ++}; ++ ++struct drm_self_refresh_data { ++ struct drm_crtc *crtc; ++ struct delayed_work entry_work; ++ struct mutex avg_mutex; ++ struct ewma_psr_time entry_avg_ms; ++ struct ewma_psr_time exit_avg_ms; ++}; ++ ++struct panel_bridge { ++ struct drm_bridge bridge; ++ struct drm_connector connector; ++ struct drm_panel *panel; ++ u32 connector_type; ++}; ++ ++struct simplefb_format { ++ const char *name; ++ u32 bits_per_pixel; ++ struct fb_bitfield red; ++ struct fb_bitfield green; ++ struct fb_bitfield blue; ++ struct fb_bitfield transp; ++ u32 fourcc; ++}; ++ ++struct simplefb_platform_data { ++ u32 width; ++ u32 height; ++ u32 stride; ++ const char *format; ++}; ++ ++struct drm_atomic_helper_damage_iter { ++ struct drm_rect plane_src; ++ const struct drm_rect *clips; ++ uint32_t num_clips; ++ uint32_t curr_clip; ++ bool full_update; ++}; ++ ++struct drm_shadow_plane_state { ++ struct drm_plane_state base; ++ struct iosys_map map[4]; ++ struct iosys_map data[4]; ++}; ++ ++struct simpledrm_device { ++ struct drm_device dev; ++ struct drm_display_mode mode; ++ const struct drm_format_info *format; ++ unsigned int pitch; ++ struct iosys_map screen_base; ++ uint32_t formats[8]; ++ size_t nformats; ++ struct drm_plane primary_plane; ++ struct drm_crtc crtc; ++ struct drm_encoder encoder; ++ struct drm_connector connector; ++}; ++ ++struct component_master_ops { ++ int (*bind)(struct device *); ++ void (*unbind)(struct device *); ++}; ++ ++struct component; ++ ++struct component_match_array { ++ void *data; ++ int (*compare)(struct device *, void *); ++ int (*compare_typed)(struct device *, int, void *); ++ void (*release)(struct device *, void *); ++ struct component *component; ++ bool duplicate; ++}; ++ ++struct aggregate_device; ++ ++struct component { ++ struct list_head node; ++ struct aggregate_device *adev; ++ bool bound; ++ const struct component_ops *ops; ++ int subcomponent; ++ struct device *dev; ++}; ++ ++struct component_match { ++ size_t alloc; ++ size_t num; ++ struct component_match_array *compare; ++}; ++ ++struct aggregate_device { ++ struct list_head node; ++ bool bound; ++ const struct component_master_ops *ops; ++ struct device *parent; ++ struct component_match *match; ++}; ++ ++struct subsys_interface { ++ const char *name; ++ struct bus_type *subsys; ++ struct list_head node; ++ int (*add_dev)(struct device *, struct subsys_interface *); ++ void (*remove_dev)(struct device *, struct subsys_interface *); ++}; ++ ++struct subsys_dev_iter { ++ struct klist_iter ki; ++ const struct device_type *type; ++}; ++ ++typedef void *acpi_handle; ++ ++struct irq_affinity_devres { ++ unsigned int count; ++ unsigned int irq[0]; ++}; ++ ++struct platform_object { ++ struct platform_device pdev; ++ char name[0]; ++}; ++ ++struct devres_node { ++ struct list_head entry; ++ dr_release_t release; ++ const char *name; ++ size_t size; ++}; ++ ++struct devres { ++ struct devres_node node; ++ u8 data[0]; ++}; ++ ++struct devres_group { ++ struct devres_node node[2]; ++ void *id; ++ int color; ++}; ++ ++struct action_devres { ++ void *data; ++ void (*action)(void *); ++}; ++ ++struct pages_devres { ++ long unsigned int addr; ++ unsigned int order; ++}; ++ ++struct transport_class { ++ struct class class; ++ int (*setup)(struct transport_container *, struct device *, struct device *); ++ int (*configure)(struct transport_container *, struct device *, struct device *); ++ int (*remove)(struct transport_container *, struct device *, struct device *); ++}; ++ ++struct anon_transport_class { ++ struct transport_class tclass; ++ struct attribute_container container; ++}; ++ ++typedef void * (*devcon_match_fn_t)(const struct fwnode_handle *, const char *, void *); ++ ++struct auxiliary_device_id { ++ char name[32]; ++ kernel_ulong_t driver_data; ++}; ++ ++struct auxiliary_device { ++ struct device dev; ++ const char *name; ++ u32 id; ++}; ++ ++struct auxiliary_driver { ++ int (*probe)(struct auxiliary_device *, const struct auxiliary_device_id *); ++ void (*remove)(struct auxiliary_device *); ++ void (*shutdown)(struct auxiliary_device *); ++ int (*suspend)(struct auxiliary_device *, pm_message_t); ++ int (*resume)(struct auxiliary_device *); ++ const char *name; ++ struct device_driver driver; ++ const struct auxiliary_device_id *id_table; ++}; ++ ++struct pm_clk_notifier_block { ++ struct notifier_block nb; ++ struct dev_pm_domain *pm_domain; ++ char *con_ids[0]; ++}; ++ ++struct firmware_cache { ++ spinlock_t lock; ++ struct list_head head; ++ int state; ++}; ++ ++struct firmware_work { ++ struct work_struct work; ++ struct module *module; ++ const char *name; ++ struct device *device; ++ void *context; ++ void (*cont)(const struct firmware *, void *); ++ u32 opt_flags; ++}; ++ ++struct node_access_nodes { ++ struct device dev; ++ struct list_head list_node; ++ unsigned int access; ++}; ++ ++struct node_attr { ++ struct device_attribute attr; ++ enum node_states state; ++}; ++ ++struct regmap_irq_type { ++ unsigned int type_reg_offset; ++ unsigned int type_reg_mask; ++ unsigned int type_rising_val; ++ unsigned int type_falling_val; ++ unsigned int type_level_low_val; ++ unsigned int type_level_high_val; ++ unsigned int types_supported; ++}; ++ ++struct regmap_irq { ++ unsigned int reg_offset; ++ unsigned int mask; ++ struct regmap_irq_type type; ++}; ++ ++struct regmap_irq_sub_irq_map { ++ unsigned int num_regs; ++ unsigned int *offset; ++}; ++ ++struct regmap_irq_chip_data; ++ ++struct regmap_irq_chip { ++ const char *name; ++ unsigned int main_status; ++ unsigned int num_main_status_bits; ++ struct regmap_irq_sub_irq_map *sub_reg_offsets; ++ int num_main_regs; ++ unsigned int status_base; ++ unsigned int mask_base; ++ unsigned int unmask_base; ++ unsigned int ack_base; ++ unsigned int wake_base; ++ unsigned int type_base; ++ unsigned int *virt_reg_base; ++ const unsigned int *config_base; ++ unsigned int irq_reg_stride; ++ unsigned int init_ack_masked: 1; ++ unsigned int mask_unmask_non_inverted: 1; ++ unsigned int use_ack: 1; ++ unsigned int ack_invert: 1; ++ unsigned int clear_ack: 1; ++ unsigned int status_invert: 1; ++ unsigned int wake_invert: 1; ++ unsigned int type_in_mask: 1; ++ unsigned int clear_on_unmask: 1; ++ unsigned int runtime_pm: 1; ++ unsigned int not_fixed_stride: 1; ++ unsigned int no_status: 1; ++ int num_regs; ++ const struct regmap_irq *irqs; ++ int num_irqs; ++ int num_type_reg; ++ int num_virt_regs; ++ int num_config_bases; ++ int num_config_regs; ++ int (*handle_pre_irq)(void *); ++ int (*handle_post_irq)(void *); ++ int (*handle_mask_sync)(struct regmap *, int, unsigned int, unsigned int, void *); ++ int (*set_type_virt)(unsigned int **, unsigned int, long unsigned int, int); ++ int (*set_type_config)(unsigned int **, unsigned int, const struct regmap_irq *, int, void *); ++ unsigned int (*get_irq_reg)(struct regmap_irq_chip_data *, unsigned int, int); ++ void *irq_drv_data; ++}; ++ ++struct regmap_irq_chip_data { ++ struct mutex lock; ++ struct irq_chip irq_chip; ++ struct regmap *map; ++ const struct regmap_irq_chip *chip; ++ int irq_base; ++ struct irq_domain *domain; ++ int irq; ++ int wake_count; ++ unsigned int mask_base; ++ unsigned int unmask_base; ++ void *status_reg_buf; ++ unsigned int *main_status_buf; ++ unsigned int *status_buf; ++ unsigned int *mask_buf; ++ unsigned int *mask_buf_def; ++ unsigned int *wake_buf; ++ unsigned int *type_buf; ++ unsigned int *type_buf_def; ++ unsigned int **virt_buf; ++ unsigned int **config_buf; ++ unsigned int irq_reg_stride; ++ unsigned int (*get_irq_reg)(struct regmap_irq_chip_data *, unsigned int, int); ++ unsigned int clear_status: 1; ++}; ++ ++typedef void (*irq_write_msi_msg_t)(struct msi_desc *, struct msi_msg *); ++ ++struct platform_msi_priv_data { ++ struct device *dev; ++ void *host_data; ++ msi_alloc_info_t arg; ++ irq_write_msi_msg_t write_msg; ++ int devid; ++}; ++ ++struct dax_device { ++ struct inode inode; ++ struct cdev cdev; ++ void *private; ++ long unsigned int flags; ++ const struct dax_operations *ops; ++ void *holder_data; ++ const struct dax_holder_operations *holder_ops; ++}; ++ ++enum dax_device_flags { ++ DAXDEV_ALIVE = 0, ++ DAXDEV_WRITE_CACHE = 1, ++ DAXDEV_SYNC = 2, ++ DAXDEV_NOCACHE = 3, ++ DAXDEV_NOMC = 4, ++}; ++ ++struct dma_buf_export_info { ++ const char *exp_name; ++ struct module *owner; ++ const struct dma_buf_ops *ops; ++ size_t size; ++ int flags; ++ struct dma_resv *resv; ++ void *priv; ++}; ++ ++struct dma_resv_iter { ++ struct dma_resv *obj; ++ enum dma_resv_usage usage; ++ struct dma_fence *fence; ++ enum dma_resv_usage fence_usage; ++ unsigned int index; ++ struct dma_resv_list *fences; ++ unsigned int num_fences; ++ bool is_restarted; ++}; ++ ++struct dma_buf_sync { ++ __u64 flags; ++}; ++ ++struct dma_buf_export_sync_file { ++ __u32 flags; ++ __s32 fd; ++}; ++ ++struct dma_buf_import_sync_file { ++ __u32 flags; ++ __s32 fd; ++}; ++ ++struct dma_buf_list { ++ struct list_head head; ++ struct mutex lock; ++}; ++ ++struct dma_fence_array; ++ ++struct dma_fence_array_cb { ++ struct dma_fence_cb cb; ++ struct dma_fence_array *array; ++}; ++ ++struct dma_fence_array { ++ struct dma_fence base; ++ spinlock_t lock; ++ unsigned int num_fences; ++ atomic_t num_pending; ++ struct dma_fence **fences; ++ struct irq_work work; ++}; ++ ++struct system_heap_buffer { ++ struct dma_heap *heap; ++ struct list_head attachments; ++ struct mutex lock; ++ long unsigned int len; ++ struct sg_table sg_table; ++ int vmap_cnt; ++ void *vaddr; ++}; ++ ++struct dma_heap_attachment { ++ struct device *dev; ++ struct sg_table *table; ++ struct list_head list; ++ bool mapped; ++}; ++ ++struct sync_merge_data { ++ char name[32]; ++ __s32 fd2; ++ __s32 fence; ++ __u32 flags; ++ __u32 pad; ++}; ++ ++struct sync_fence_info { ++ char obj_name[32]; ++ char driver_name[32]; ++ __s32 status; ++ __u32 flags; ++ __u64 timestamp_ns; ++}; ++ ++struct sync_file_info { ++ char name[32]; ++ __s32 status; ++ __u32 flags; ++ __u32 num_fences; ++ __u32 pad; ++ __u64 sync_fence_info; ++}; ++ ++struct scsi_host_busy_iter_data { ++ bool (*fn)(struct scsi_cmnd *, void *); ++ void *priv; ++}; ++ ++struct scsi_mode_data { ++ __u32 length; ++ __u16 block_descriptor_length; ++ __u8 medium_type; ++ __u8 device_specific; ++ __u8 header_length; ++ __u8 longlba: 1; ++}; ++ ++struct scsi_event { ++ enum scsi_device_event evt_type; ++ struct list_head node; ++}; ++ ++enum { ++ ACTION_FAIL = 0, ++ ACTION_REPREP = 1, ++ ACTION_DELAYED_REPREP = 2, ++ ACTION_RETRY = 3, ++ ACTION_DELAYED_RETRY = 4, ++}; ++ ++struct scsi_proc_entry { ++ struct list_head entry; ++ const struct scsi_host_template *sht; ++ struct proc_dir_entry *proc_dir; ++ unsigned int present; ++}; ++ ++struct scsi_varlen_cdb_hdr { ++ __u8 opcode; ++ __u8 control; ++ __u8 misc[5]; ++ __u8 additional_cdb_length; ++ __be16 service_action; ++}; ++ ++enum pr_status { ++ PR_STS_SUCCESS = 0, ++ PR_STS_IOERR = 2, ++ PR_STS_RESERVATION_CONFLICT = 24, ++ PR_STS_RETRY_PATH_FAILURE = 917504, ++ PR_STS_PATH_FAST_FAILED = 983040, ++ PR_STS_PATH_FAILED = 65536, ++}; ++ ++enum scsi_prot_flags { ++ SCSI_PROT_TRANSFER_PI = 1, ++ SCSI_PROT_GUARD_CHECK = 2, ++ SCSI_PROT_REF_CHECK = 4, ++ SCSI_PROT_REF_INCREMENT = 8, ++ SCSI_PROT_IP_CHECKSUM = 16, ++}; ++ ++enum { ++ SD_EXT_CDB_SIZE = 32, ++ SD_MEMPOOL_SIZE = 2, ++}; ++ ++enum { ++ SD_DEF_XFER_BLOCKS = 65535, ++ SD_MAX_XFER_BLOCKS = 4294967295, ++ SD_MAX_WS10_BLOCKS = 65535, ++ SD_MAX_WS16_BLOCKS = 8388607, ++}; ++ ++enum { ++ SD_LBP_FULL = 0, ++ SD_LBP_UNMAP = 1, ++ SD_LBP_WS16 = 2, ++ SD_LBP_WS10 = 3, ++ SD_LBP_ZERO = 4, ++ SD_LBP_DISABLE = 5, ++}; ++ ++enum { ++ SD_ZERO_WRITE = 0, ++ SD_ZERO_WS = 1, ++ SD_ZERO_WS16_UNMAP = 2, ++ SD_ZERO_WS10_UNMAP = 3, ++}; ++ ++enum { ++ mechtype_caddy = 0, ++ mechtype_tray = 1, ++ mechtype_popup = 2, ++ mechtype_individual_changer = 4, ++ mechtype_cartridge_changer = 5, ++}; ++ ++struct event_header { ++ __be16 data_len; ++ __u8 nea: 1; ++ __u8 reserved1: 4; ++ __u8 notification_class: 3; ++ __u8 supp_event_class; ++}; ++ ++struct ccs_modesel_head { ++ __u8 _r1; ++ __u8 medium; ++ __u8 _r2; ++ __u8 block_desc_length; ++ __u8 density; ++ __u8 number_blocks_hi; ++ __u8 number_blocks_med; ++ __u8 number_blocks_lo; ++ __u8 _r3; ++ __u8 block_length_hi; ++ __u8 block_length_med; ++ __u8 block_length_lo; ++}; ++ ++enum ata_xfer_mask { ++ ATA_MASK_PIO = 127, ++ ATA_MASK_MWDMA = 3968, ++ ATA_MASK_UDMA = 1044480, ++}; ++ ++struct trace_event_raw_ata_qc_issue_template { ++ struct trace_entry ent; ++ unsigned int ata_port; ++ unsigned int ata_dev; ++ unsigned int tag; ++ unsigned char cmd; ++ unsigned char dev; ++ unsigned char lbal; ++ unsigned char lbam; ++ unsigned char lbah; ++ unsigned char nsect; ++ unsigned char feature; ++ unsigned char hob_lbal; ++ unsigned char hob_lbam; ++ unsigned char hob_lbah; ++ unsigned char hob_nsect; ++ unsigned char hob_feature; ++ unsigned char ctl; ++ unsigned char proto; ++ long unsigned int flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ata_qc_complete_template { ++ struct trace_entry ent; ++ unsigned int ata_port; ++ unsigned int ata_dev; ++ unsigned int tag; ++ unsigned char status; ++ unsigned char dev; ++ unsigned char lbal; ++ unsigned char lbam; ++ unsigned char lbah; ++ unsigned char nsect; ++ unsigned char error; ++ unsigned char hob_lbal; ++ unsigned char hob_lbam; ++ unsigned char hob_lbah; ++ unsigned char hob_nsect; ++ unsigned char hob_feature; ++ unsigned char ctl; ++ long unsigned int flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ata_tf_load { ++ struct trace_entry ent; ++ unsigned int ata_port; ++ unsigned char cmd; ++ unsigned char dev; ++ unsigned char lbal; ++ unsigned char lbam; ++ unsigned char lbah; ++ unsigned char nsect; ++ unsigned char feature; ++ unsigned char hob_lbal; ++ unsigned char hob_lbam; ++ unsigned char hob_lbah; ++ unsigned char hob_nsect; ++ unsigned char hob_feature; ++ unsigned char proto; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ata_exec_command_template { ++ struct trace_entry ent; ++ unsigned int ata_port; ++ unsigned int tag; ++ unsigned char cmd; ++ unsigned char feature; ++ unsigned char hob_nsect; ++ unsigned char proto; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ata_bmdma_status { ++ struct trace_entry ent; ++ unsigned int ata_port; ++ unsigned int tag; ++ unsigned char host_stat; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ata_eh_link_autopsy { ++ struct trace_entry ent; ++ unsigned int ata_port; ++ unsigned int ata_dev; ++ unsigned int eh_action; ++ unsigned int eh_err_mask; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ata_eh_link_autopsy_qc { ++ struct trace_entry ent; ++ unsigned int ata_port; ++ unsigned int ata_dev; ++ unsigned int tag; ++ unsigned int qc_flags; ++ unsigned int eh_err_mask; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ata_eh_action_template { ++ struct trace_entry ent; ++ unsigned int ata_port; ++ unsigned int ata_dev; ++ unsigned int eh_action; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ata_link_reset_begin_template { ++ struct trace_entry ent; ++ unsigned int ata_port; ++ unsigned int class[2]; ++ long unsigned int deadline; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ata_link_reset_end_template { ++ struct trace_entry ent; ++ unsigned int ata_port; ++ unsigned int class[2]; ++ int rc; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ata_port_eh_begin_template { ++ struct trace_entry ent; ++ unsigned int ata_port; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ata_sff_hsm_template { ++ struct trace_entry ent; ++ unsigned int ata_port; ++ unsigned int ata_dev; ++ unsigned int tag; ++ unsigned int qc_flags; ++ unsigned int protocol; ++ unsigned int hsm_state; ++ unsigned char dev_state; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ata_transfer_data_template { ++ struct trace_entry ent; ++ unsigned int ata_port; ++ unsigned int ata_dev; ++ unsigned int tag; ++ unsigned int flags; ++ unsigned int offset; ++ unsigned int bytes; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ata_sff_template { ++ struct trace_entry ent; ++ unsigned int ata_port; ++ unsigned char hsm_state; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_ata_qc_issue_template {}; ++ ++struct trace_event_data_offsets_ata_qc_complete_template {}; ++ ++struct trace_event_data_offsets_ata_tf_load {}; ++ ++struct trace_event_data_offsets_ata_exec_command_template {}; ++ ++struct trace_event_data_offsets_ata_bmdma_status {}; ++ ++struct trace_event_data_offsets_ata_eh_link_autopsy {}; ++ ++struct trace_event_data_offsets_ata_eh_link_autopsy_qc {}; ++ ++struct trace_event_data_offsets_ata_eh_action_template {}; ++ ++struct trace_event_data_offsets_ata_link_reset_begin_template {}; ++ ++struct trace_event_data_offsets_ata_link_reset_end_template {}; ++ ++struct trace_event_data_offsets_ata_port_eh_begin_template {}; ++ ++struct trace_event_data_offsets_ata_sff_hsm_template {}; ++ ++struct trace_event_data_offsets_ata_transfer_data_template {}; ++ ++struct trace_event_data_offsets_ata_sff_template {}; ++ ++typedef void (*btf_trace_ata_qc_prep)(void *, struct ata_queued_cmd *); ++ ++typedef void (*btf_trace_ata_qc_issue)(void *, struct ata_queued_cmd *); ++ ++typedef void (*btf_trace_ata_qc_complete_internal)(void *, struct ata_queued_cmd *); ++ ++typedef void (*btf_trace_ata_qc_complete_failed)(void *, struct ata_queued_cmd *); ++ ++typedef void (*btf_trace_ata_qc_complete_done)(void *, struct ata_queued_cmd *); ++ ++typedef void (*btf_trace_ata_tf_load)(void *, struct ata_port *, const struct ata_taskfile *); ++ ++typedef void (*btf_trace_ata_exec_command)(void *, struct ata_port *, const struct ata_taskfile *, unsigned int); ++ ++typedef void (*btf_trace_ata_bmdma_setup)(void *, struct ata_port *, const struct ata_taskfile *, unsigned int); ++ ++typedef void (*btf_trace_ata_bmdma_start)(void *, struct ata_port *, const struct ata_taskfile *, unsigned int); ++ ++typedef void (*btf_trace_ata_bmdma_stop)(void *, struct ata_port *, const struct ata_taskfile *, unsigned int); ++ ++typedef void (*btf_trace_ata_bmdma_status)(void *, struct ata_port *, unsigned int); ++ ++typedef void (*btf_trace_ata_eh_link_autopsy)(void *, struct ata_device *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_ata_eh_link_autopsy_qc)(void *, struct ata_queued_cmd *); ++ ++typedef void (*btf_trace_ata_eh_about_to_do)(void *, struct ata_link *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_ata_eh_done)(void *, struct ata_link *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_ata_link_hardreset_begin)(void *, struct ata_link *, unsigned int *, long unsigned int); ++ ++typedef void (*btf_trace_ata_slave_hardreset_begin)(void *, struct ata_link *, unsigned int *, long unsigned int); ++ ++typedef void (*btf_trace_ata_link_softreset_begin)(void *, struct ata_link *, unsigned int *, long unsigned int); ++ ++typedef void (*btf_trace_ata_link_hardreset_end)(void *, struct ata_link *, unsigned int *, int); ++ ++typedef void (*btf_trace_ata_slave_hardreset_end)(void *, struct ata_link *, unsigned int *, int); ++ ++typedef void (*btf_trace_ata_link_softreset_end)(void *, struct ata_link *, unsigned int *, int); ++ ++typedef void (*btf_trace_ata_link_postreset)(void *, struct ata_link *, unsigned int *, int); ++ ++typedef void (*btf_trace_ata_slave_postreset)(void *, struct ata_link *, unsigned int *, int); ++ ++typedef void (*btf_trace_ata_std_sched_eh)(void *, struct ata_port *); ++ ++typedef void (*btf_trace_ata_port_freeze)(void *, struct ata_port *); ++ ++typedef void (*btf_trace_ata_port_thaw)(void *, struct ata_port *); ++ ++typedef void (*btf_trace_ata_sff_hsm_state)(void *, struct ata_queued_cmd *, unsigned char); ++ ++typedef void (*btf_trace_ata_sff_hsm_command_complete)(void *, struct ata_queued_cmd *, unsigned char); ++ ++typedef void (*btf_trace_ata_sff_port_intr)(void *, struct ata_queued_cmd *, unsigned char); ++ ++typedef void (*btf_trace_ata_sff_pio_transfer_data)(void *, struct ata_queued_cmd *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_atapi_pio_transfer_data)(void *, struct ata_queued_cmd *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_atapi_send_cdb)(void *, struct ata_queued_cmd *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_ata_sff_flush_pio_task)(void *, struct ata_port *); ++ ++struct ata_force_param { ++ const char *name; ++ u8 cbl; ++ u8 spd_limit; ++ unsigned int xfer_mask; ++ unsigned int horkage_on; ++ unsigned int horkage_off; ++ u16 lflags_on; ++ u16 lflags_off; ++}; ++ ++struct ata_force_ent { ++ int port; ++ int device; ++ struct ata_force_param param; ++}; ++ ++struct ata_xfer_ent { ++ int shift; ++ int bits; ++ u8 base; ++}; ++ ++struct ata_blacklist_entry { ++ const char *model_num; ++ const char *model_rev; ++ long unsigned int horkage; ++}; ++ ++struct ata_internal { ++ struct scsi_transport_template t; ++ struct device_attribute private_port_attrs[3]; ++ struct device_attribute private_link_attrs[3]; ++ struct device_attribute private_dev_attrs[9]; ++ struct transport_container link_attr_cont; ++ struct transport_container dev_attr_cont; ++ struct device_attribute *link_attrs[4]; ++ struct device_attribute *port_attrs[4]; ++ struct device_attribute *dev_attrs[10]; ++}; ++ ++struct ata_show_ering_arg { ++ char *buf; ++ int written; ++}; ++ ++struct ata_timing { ++ short unsigned int mode; ++ short unsigned int setup; ++ short unsigned int act8b; ++ short unsigned int rec8b; ++ short unsigned int cyc8b; ++ short unsigned int active; ++ short unsigned int recover; ++ short unsigned int dmack_hold; ++ short unsigned int cycle; ++ short unsigned int udma; ++}; ++ ++struct phy_configure_opts_dp { ++ unsigned int link_rate; ++ unsigned int lanes; ++ unsigned int voltage[4]; ++ unsigned int pre[4]; ++ u8 ssc: 1; ++ u8 set_rate: 1; ++ u8 set_lanes: 1; ++ u8 set_voltages: 1; ++}; ++ ++struct phy_configure_opts_lvds { ++ unsigned int bits_per_lane_and_dclk_cycle; ++ long unsigned int differential_clk_rate; ++ unsigned int lanes; ++ bool is_slave; ++}; ++ ++struct phy_configure_opts_mipi_dphy { ++ unsigned int clk_miss; ++ unsigned int clk_post; ++ unsigned int clk_pre; ++ unsigned int clk_prepare; ++ unsigned int clk_settle; ++ unsigned int clk_term_en; ++ unsigned int clk_trail; ++ unsigned int clk_zero; ++ unsigned int d_term_en; ++ unsigned int eot; ++ unsigned int hs_exit; ++ unsigned int hs_prepare; ++ unsigned int hs_settle; ++ unsigned int hs_skip; ++ unsigned int hs_trail; ++ unsigned int hs_zero; ++ unsigned int init; ++ unsigned int lpx; ++ unsigned int ta_get; ++ unsigned int ta_go; ++ unsigned int ta_sure; ++ unsigned int wakeup; ++ long unsigned int hs_clk_rate; ++ long unsigned int lp_clk_rate; ++ unsigned char lanes; ++}; ++ ++enum phy_mode { ++ PHY_MODE_INVALID = 0, ++ PHY_MODE_USB_HOST = 1, ++ PHY_MODE_USB_HOST_LS = 2, ++ PHY_MODE_USB_HOST_FS = 3, ++ PHY_MODE_USB_HOST_HS = 4, ++ PHY_MODE_USB_HOST_SS = 5, ++ PHY_MODE_USB_DEVICE = 6, ++ PHY_MODE_USB_DEVICE_LS = 7, ++ PHY_MODE_USB_DEVICE_FS = 8, ++ PHY_MODE_USB_DEVICE_HS = 9, ++ PHY_MODE_USB_DEVICE_SS = 10, ++ PHY_MODE_USB_OTG = 11, ++ PHY_MODE_UFS_HS_A = 12, ++ PHY_MODE_UFS_HS_B = 13, ++ PHY_MODE_PCIE = 14, ++ PHY_MODE_ETHERNET = 15, ++ PHY_MODE_MIPI_DPHY = 16, ++ PHY_MODE_SATA = 17, ++ PHY_MODE_LVDS = 18, ++ PHY_MODE_DP = 19, ++}; ++ ++enum phy_media { ++ PHY_MEDIA_DEFAULT = 0, ++ PHY_MEDIA_SR = 1, ++ PHY_MEDIA_DAC = 2, ++}; ++ ++union phy_configure_opts { ++ struct phy_configure_opts_mipi_dphy mipi_dphy; ++ struct phy_configure_opts_dp dp; ++ struct phy_configure_opts_lvds lvds; ++}; ++ ++struct phy; ++ ++struct phy_ops { ++ int (*init)(struct phy *); ++ int (*exit)(struct phy *); ++ int (*power_on)(struct phy *); ++ int (*power_off)(struct phy *); ++ int (*set_mode)(struct phy *, enum phy_mode, int); ++ int (*set_media)(struct phy *, enum phy_media); ++ int (*set_speed)(struct phy *, int); ++ int (*configure)(struct phy *, union phy_configure_opts *); ++ int (*validate)(struct phy *, enum phy_mode, int, union phy_configure_opts *); ++ int (*reset)(struct phy *); ++ int (*calibrate)(struct phy *); ++ void (*release)(struct phy *); ++ struct module *owner; ++}; ++ ++struct phy_attrs { ++ u32 bus_width; ++ u32 max_link_rate; ++ enum phy_mode mode; ++}; ++ ++struct phy { ++ struct device dev; ++ int id; ++ const struct phy_ops *ops; ++ struct mutex mutex; ++ int init_count; ++ int power_count; ++ struct phy_attrs attrs; ++ struct regulator *pwr; ++}; ++ ++enum { ++ AHCI_MAX_PORTS = 32, ++ AHCI_MAX_SG = 168, ++ AHCI_DMA_BOUNDARY = 4294967295, ++ AHCI_MAX_CMDS = 32, ++ AHCI_CMD_SZ = 32, ++ AHCI_CMD_SLOT_SZ = 1024, ++ AHCI_RX_FIS_SZ = 256, ++ AHCI_CMD_TBL_CDB = 64, ++ AHCI_CMD_TBL_HDR_SZ = 128, ++ AHCI_CMD_TBL_SZ = 2816, ++ AHCI_CMD_TBL_AR_SZ = 90112, ++ AHCI_PORT_PRIV_DMA_SZ = 91392, ++ AHCI_PORT_PRIV_FBS_DMA_SZ = 95232, ++ AHCI_IRQ_ON_SG = 2147483648, ++ AHCI_CMD_ATAPI = 32, ++ AHCI_CMD_WRITE = 64, ++ AHCI_CMD_PREFETCH = 128, ++ AHCI_CMD_RESET = 256, ++ AHCI_CMD_CLR_BUSY = 1024, ++ RX_FIS_PIO_SETUP = 32, ++ RX_FIS_D2H_REG = 64, ++ RX_FIS_SDB = 88, ++ RX_FIS_UNK = 96, ++ HOST_CAP = 0, ++ HOST_CTL = 4, ++ HOST_IRQ_STAT = 8, ++ HOST_PORTS_IMPL = 12, ++ HOST_VERSION = 16, ++ HOST_EM_LOC = 28, ++ HOST_EM_CTL = 32, ++ HOST_CAP2 = 36, ++ HOST_RESET = 1, ++ HOST_IRQ_EN = 2, ++ HOST_MRSM = 4, ++ HOST_AHCI_EN = 2147483648, ++ HOST_CAP_SXS = 32, ++ HOST_CAP_EMS = 64, ++ HOST_CAP_CCC = 128, ++ HOST_CAP_PART = 8192, ++ HOST_CAP_SSC = 16384, ++ HOST_CAP_PIO_MULTI = 32768, ++ HOST_CAP_FBS = 65536, ++ HOST_CAP_PMP = 131072, ++ HOST_CAP_ONLY = 262144, ++ HOST_CAP_CLO = 16777216, ++ HOST_CAP_LED = 33554432, ++ HOST_CAP_ALPM = 67108864, ++ HOST_CAP_SSS = 134217728, ++ HOST_CAP_MPS = 268435456, ++ HOST_CAP_SNTF = 536870912, ++ HOST_CAP_NCQ = 1073741824, ++ HOST_CAP_64 = 2147483648, ++ HOST_CAP2_BOH = 1, ++ HOST_CAP2_NVMHCI = 2, ++ HOST_CAP2_APST = 4, ++ HOST_CAP2_SDS = 8, ++ HOST_CAP2_SADM = 16, ++ HOST_CAP2_DESO = 32, ++ PORT_LST_ADDR = 0, ++ PORT_LST_ADDR_HI = 4, ++ PORT_FIS_ADDR = 8, ++ PORT_FIS_ADDR_HI = 12, ++ PORT_IRQ_STAT = 16, ++ PORT_IRQ_MASK = 20, ++ PORT_CMD = 24, ++ PORT_TFDATA = 32, ++ PORT_SIG = 36, ++ PORT_CMD_ISSUE = 56, ++ PORT_SCR_STAT = 40, ++ PORT_SCR_CTL = 44, ++ PORT_SCR_ERR = 48, ++ PORT_SCR_ACT = 52, ++ PORT_SCR_NTF = 60, ++ PORT_FBS = 64, ++ PORT_DEVSLP = 68, ++ PORT_IRQ_COLD_PRES = 2147483648, ++ PORT_IRQ_TF_ERR = 1073741824, ++ PORT_IRQ_HBUS_ERR = 536870912, ++ PORT_IRQ_HBUS_DATA_ERR = 268435456, ++ PORT_IRQ_IF_ERR = 134217728, ++ PORT_IRQ_IF_NONFATAL = 67108864, ++ PORT_IRQ_OVERFLOW = 16777216, ++ PORT_IRQ_BAD_PMP = 8388608, ++ PORT_IRQ_PHYRDY = 4194304, ++ PORT_IRQ_DMPS = 128, ++ PORT_IRQ_CONNECT = 64, ++ PORT_IRQ_SG_DONE = 32, ++ PORT_IRQ_UNK_FIS = 16, ++ PORT_IRQ_SDB_FIS = 8, ++ PORT_IRQ_DMAS_FIS = 4, ++ PORT_IRQ_PIOS_FIS = 2, ++ PORT_IRQ_D2H_REG_FIS = 1, ++ PORT_IRQ_FREEZE = 683671632, ++ PORT_IRQ_ERROR = 2025848912, ++ DEF_PORT_IRQ = 2025848959, ++ PORT_CMD_ASP = 134217728, ++ PORT_CMD_ALPE = 67108864, ++ PORT_CMD_ATAPI = 16777216, ++ PORT_CMD_FBSCP = 4194304, ++ PORT_CMD_ESP = 2097152, ++ PORT_CMD_CPD = 1048576, ++ PORT_CMD_MPSP = 524288, ++ PORT_CMD_HPCP = 262144, ++ PORT_CMD_PMP = 131072, ++ PORT_CMD_LIST_ON = 32768, ++ PORT_CMD_FIS_ON = 16384, ++ PORT_CMD_FIS_RX = 16, ++ PORT_CMD_CLO = 8, ++ PORT_CMD_POWER_ON = 4, ++ PORT_CMD_SPIN_UP = 2, ++ PORT_CMD_START = 1, ++ PORT_CMD_ICC_MASK = 4026531840, ++ PORT_CMD_ICC_ACTIVE = 268435456, ++ PORT_CMD_ICC_PARTIAL = 536870912, ++ PORT_CMD_ICC_SLUMBER = 1610612736, ++ PORT_CMD_CAP = 8126464, ++ PORT_FBS_DWE_OFFSET = 16, ++ PORT_FBS_ADO_OFFSET = 12, ++ PORT_FBS_DEV_OFFSET = 8, ++ PORT_FBS_DEV_MASK = 3840, ++ PORT_FBS_SDE = 4, ++ PORT_FBS_DEC = 2, ++ PORT_FBS_EN = 1, ++ PORT_DEVSLP_DM_OFFSET = 25, ++ PORT_DEVSLP_DM_MASK = 503316480, ++ PORT_DEVSLP_DITO_OFFSET = 15, ++ PORT_DEVSLP_MDAT_OFFSET = 10, ++ PORT_DEVSLP_DETO_OFFSET = 2, ++ PORT_DEVSLP_DSP = 2, ++ PORT_DEVSLP_ADSE = 1, ++ AHCI_HFLAG_NO_NCQ = 1, ++ AHCI_HFLAG_IGN_IRQ_IF_ERR = 2, ++ AHCI_HFLAG_IGN_SERR_INTERNAL = 4, ++ AHCI_HFLAG_32BIT_ONLY = 8, ++ AHCI_HFLAG_MV_PATA = 16, ++ AHCI_HFLAG_NO_MSI = 32, ++ AHCI_HFLAG_NO_PMP = 64, ++ AHCI_HFLAG_SECT255 = 256, ++ AHCI_HFLAG_YES_NCQ = 512, ++ AHCI_HFLAG_NO_SUSPEND = 1024, ++ AHCI_HFLAG_SRST_TOUT_IS_OFFLINE = 2048, ++ AHCI_HFLAG_NO_SNTF = 4096, ++ AHCI_HFLAG_NO_FPDMA_AA = 8192, ++ AHCI_HFLAG_YES_FBS = 16384, ++ AHCI_HFLAG_DELAY_ENGINE = 32768, ++ AHCI_HFLAG_NO_DEVSLP = 131072, ++ AHCI_HFLAG_NO_FBS = 262144, ++ AHCI_HFLAG_MULTI_MSI = 1048576, ++ AHCI_HFLAG_WAKE_BEFORE_STOP = 4194304, ++ AHCI_HFLAG_YES_ALPM = 8388608, ++ AHCI_HFLAG_NO_WRITE_TO_RO = 16777216, ++ AHCI_HFLAG_USE_LPM_POLICY = 33554432, ++ AHCI_HFLAG_SUSPEND_PHYS = 67108864, ++ AHCI_HFLAG_NO_SXS = 268435456, ++ AHCI_FLAG_COMMON = 393346, ++ ICH_MAP = 144, ++ PCS_6 = 146, ++ PCS_7 = 148, ++ EM_MAX_SLOTS = 15, ++ EM_MAX_RETRY = 5, ++ EM_CTL_RST = 512, ++ EM_CTL_TM = 256, ++ EM_CTL_MR = 1, ++ EM_CTL_ALHD = 67108864, ++ EM_CTL_XMT = 33554432, ++ EM_CTL_SMB = 16777216, ++ EM_CTL_SGPIO = 524288, ++ EM_CTL_SES = 262144, ++ EM_CTL_SAFTE = 131072, ++ EM_CTL_LED = 65536, ++ EM_MSG_TYPE_LED = 1, ++ EM_MSG_TYPE_SAFTE = 2, ++ EM_MSG_TYPE_SES2 = 4, ++ EM_MSG_TYPE_SGPIO = 8, ++}; ++ ++struct ahci_cmd_hdr { ++ __le32 opts; ++ __le32 status; ++ __le32 tbl_addr; ++ __le32 tbl_addr_hi; ++ __le32 reserved[4]; ++}; ++ ++struct ahci_sg { ++ __le32 addr; ++ __le32 addr_hi; ++ __le32 reserved; ++ __le32 flags_size; ++}; ++ ++struct ahci_em_priv { ++ enum sw_activity blink_policy; ++ struct timer_list timer; ++ long unsigned int saved_activity; ++ long unsigned int activity; ++ long unsigned int led_state; ++ struct ata_link *link; ++}; ++ ++struct ahci_port_priv { ++ struct ata_link *active_link; ++ struct ahci_cmd_hdr *cmd_slot; ++ dma_addr_t cmd_slot_dma; ++ void *cmd_tbl; ++ dma_addr_t cmd_tbl_dma; ++ void *rx_fis; ++ dma_addr_t rx_fis_dma; ++ unsigned int ncq_saw_d2h: 1; ++ unsigned int ncq_saw_dmas: 1; ++ unsigned int ncq_saw_sdb: 1; ++ spinlock_t lock; ++ u32 intr_mask; ++ bool fbs_supported; ++ bool fbs_enabled; ++ int fbs_last_dev; ++ struct ahci_em_priv em_priv[15]; ++ char *irq_desc; ++}; ++ ++struct ahci_host_priv { ++ unsigned int flags; ++ u32 mask_port_map; ++ void *mmio; ++ u32 cap; ++ u32 cap2; ++ u32 version; ++ u32 port_map; ++ u32 saved_cap; ++ u32 saved_cap2; ++ u32 saved_port_map; ++ u32 saved_port_cap[32]; ++ u32 em_loc; ++ u32 em_buf_sz; ++ u32 em_msg_type; ++ u32 remapped_nvme; ++ bool got_runtime_pm; ++ unsigned int n_clks; ++ struct clk_bulk_data *clks; ++ unsigned int f_rsts; ++ struct reset_control *rsts; ++ struct regulator **target_pwrs; ++ struct regulator *ahci_regulator; ++ struct regulator *phy_regulator; ++ struct phy **phys; ++ unsigned int nports; ++ void *plat_data; ++ unsigned int irq; ++ void (*start_engine)(struct ata_port *); ++ int (*stop_engine)(struct ata_port *); ++ irqreturn_t (*irq_handler)(int, void *); ++ int (*get_irq_vector)(struct ata_host *, int); ++}; ++ ++struct mdiobus_devres { ++ struct mii_bus *mii; ++}; ++ ++struct mii_ioctl_data { ++ __u16 phy_id; ++ __u16 reg_num; ++ __u16 val_in; ++ __u16 val_out; ++}; ++ ++struct dev_ext_attribute { ++ struct device_attribute attr; ++ void *var; ++}; ++ ++struct trace_event_raw_mdio_access { ++ struct trace_entry ent; ++ char busid[61]; ++ char read; ++ u8 addr; ++ u16 val; ++ unsigned int regnum; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_mdio_access {}; ++ ++typedef void (*btf_trace_mdio_access)(void *, struct mii_bus *, char, u8, unsigned int, u16, int); ++ ++struct mdio_bus_stat_attr { ++ int addr; ++ unsigned int field_offset; ++}; ++ ++struct sfp; ++ ++struct sfp_socket_ops; ++ ++struct sfp_quirk; ++ ++struct sfp_upstream_ops; ++ ++struct sfp_bus { ++ struct kref kref; ++ struct list_head node; ++ const struct fwnode_handle *fwnode; ++ const struct sfp_socket_ops *socket_ops; ++ struct device *sfp_dev; ++ struct sfp *sfp; ++ const struct sfp_quirk *sfp_quirk; ++ const struct sfp_upstream_ops *upstream_ops; ++ void *upstream; ++ struct phy_device *phydev; ++ bool registered; ++ bool started; ++}; ++ ++struct sfp_eeprom_base { ++ u8 phys_id; ++ u8 phys_ext_id; ++ u8 connector; ++ u8 e10g_base_er: 1; ++ u8 e10g_base_lrm: 1; ++ u8 e10g_base_lr: 1; ++ u8 e10g_base_sr: 1; ++ u8 if_1x_sx: 1; ++ u8 if_1x_lx: 1; ++ u8 if_1x_copper_active: 1; ++ u8 if_1x_copper_passive: 1; ++ u8 escon_mmf_1310_led: 1; ++ u8 escon_smf_1310_laser: 1; ++ u8 sonet_oc192_short_reach: 1; ++ u8 sonet_reach_bit1: 1; ++ u8 sonet_reach_bit2: 1; ++ u8 sonet_oc48_long_reach: 1; ++ u8 sonet_oc48_intermediate_reach: 1; ++ u8 sonet_oc48_short_reach: 1; ++ u8 unallocated_5_7: 1; ++ u8 sonet_oc12_smf_long_reach: 1; ++ u8 sonet_oc12_smf_intermediate_reach: 1; ++ u8 sonet_oc12_short_reach: 1; ++ u8 unallocated_5_3: 1; ++ u8 sonet_oc3_smf_long_reach: 1; ++ u8 sonet_oc3_smf_intermediate_reach: 1; ++ u8 sonet_oc3_short_reach: 1; ++ u8 e_base_px: 1; ++ u8 e_base_bx10: 1; ++ u8 e100_base_fx: 1; ++ u8 e100_base_lx: 1; ++ u8 e1000_base_t: 1; ++ u8 e1000_base_cx: 1; ++ u8 e1000_base_lx: 1; ++ u8 e1000_base_sx: 1; ++ u8 fc_ll_v: 1; ++ u8 fc_ll_s: 1; ++ u8 fc_ll_i: 1; ++ u8 fc_ll_l: 1; ++ u8 fc_ll_m: 1; ++ u8 fc_tech_sa: 1; ++ u8 fc_tech_lc: 1; ++ u8 fc_tech_electrical_inter_enclosure: 1; ++ u8 fc_tech_electrical_intra_enclosure: 1; ++ u8 fc_tech_sn: 1; ++ u8 fc_tech_sl: 1; ++ u8 fc_tech_ll: 1; ++ u8 sfp_ct_active: 1; ++ u8 sfp_ct_passive: 1; ++ u8 unallocated_8_1: 1; ++ u8 unallocated_8_0: 1; ++ u8 fc_media_tw: 1; ++ u8 fc_media_tp: 1; ++ u8 fc_media_mi: 1; ++ u8 fc_media_tv: 1; ++ u8 fc_media_m6: 1; ++ u8 fc_media_m5: 1; ++ u8 unallocated_9_1: 1; ++ u8 fc_media_sm: 1; ++ u8 fc_speed_1200: 1; ++ u8 fc_speed_800: 1; ++ u8 fc_speed_1600: 1; ++ u8 fc_speed_400: 1; ++ u8 fc_speed_3200: 1; ++ u8 fc_speed_200: 1; ++ u8 unallocated_10_1: 1; ++ u8 fc_speed_100: 1; ++ u8 encoding; ++ u8 br_nominal; ++ u8 rate_id; ++ u8 link_len[6]; ++ char vendor_name[16]; ++ u8 extended_cc; ++ char vendor_oui[3]; ++ char vendor_pn[16]; ++ char vendor_rev[4]; ++ union { ++ __be16 optical_wavelength; ++ __be16 cable_compliance; ++ struct { ++ u8 reserved60_2: 6; ++ u8 fc_pi_4_app_h: 1; ++ u8 sff8431_app_e: 1; ++ u8 reserved61: 8; ++ } passive; ++ struct { ++ u8 reserved60_4: 4; ++ u8 fc_pi_4_lim: 1; ++ u8 sff8431_lim: 1; ++ u8 fc_pi_4_app_h: 1; ++ u8 sff8431_app_e: 1; ++ u8 reserved61: 8; ++ } active; ++ }; ++ u8 reserved62; ++ u8 cc_base; ++}; ++ ++struct sfp_eeprom_ext { ++ __be16 options; ++ u8 br_max; ++ u8 br_min; ++ char vendor_sn[16]; ++ char datecode[8]; ++ u8 diagmon; ++ u8 enhopts; ++ u8 sff8472_compliance; ++ u8 cc_ext; ++}; ++ ++struct sfp_eeprom_id { ++ struct sfp_eeprom_base base; ++ struct sfp_eeprom_ext ext; ++}; ++ ++enum { ++ SFF8024_ID_UNK = 0, ++ SFF8024_ID_SFF_8472 = 2, ++ SFF8024_ID_SFP = 3, ++ SFF8024_ID_DWDM_SFP = 11, ++ SFF8024_ID_QSFP_8438 = 12, ++ SFF8024_ID_QSFP_8436_8636 = 13, ++ SFF8024_ID_QSFP28_8636 = 17, ++ SFF8024_ENCODING_UNSPEC = 0, ++ SFF8024_ENCODING_8B10B = 1, ++ SFF8024_ENCODING_4B5B = 2, ++ SFF8024_ENCODING_NRZ = 3, ++ SFF8024_ENCODING_8472_MANCHESTER = 4, ++ SFF8024_ENCODING_8472_SONET = 5, ++ SFF8024_ENCODING_8472_64B66B = 6, ++ SFF8024_ENCODING_8436_MANCHESTER = 6, ++ SFF8024_ENCODING_8436_SONET = 4, ++ SFF8024_ENCODING_8436_64B66B = 5, ++ SFF8024_ENCODING_256B257B = 7, ++ SFF8024_ENCODING_PAM4 = 8, ++ SFF8024_CONNECTOR_UNSPEC = 0, ++ SFF8024_CONNECTOR_SC = 1, ++ SFF8024_CONNECTOR_FIBERJACK = 6, ++ SFF8024_CONNECTOR_LC = 7, ++ SFF8024_CONNECTOR_MT_RJ = 8, ++ SFF8024_CONNECTOR_MU = 9, ++ SFF8024_CONNECTOR_SG = 10, ++ SFF8024_CONNECTOR_OPTICAL_PIGTAIL = 11, ++ SFF8024_CONNECTOR_MPO_1X12 = 12, ++ SFF8024_CONNECTOR_MPO_2X16 = 13, ++ SFF8024_CONNECTOR_HSSDC_II = 32, ++ SFF8024_CONNECTOR_COPPER_PIGTAIL = 33, ++ SFF8024_CONNECTOR_RJ45 = 34, ++ SFF8024_CONNECTOR_NOSEPARATE = 35, ++ SFF8024_CONNECTOR_MXC_2X16 = 36, ++ SFF8024_ECC_UNSPEC = 0, ++ SFF8024_ECC_100G_25GAUI_C2M_AOC = 1, ++ SFF8024_ECC_100GBASE_SR4_25GBASE_SR = 2, ++ SFF8024_ECC_100GBASE_LR4_25GBASE_LR = 3, ++ SFF8024_ECC_100GBASE_ER4_25GBASE_ER = 4, ++ SFF8024_ECC_100GBASE_SR10 = 5, ++ SFF8024_ECC_100GBASE_CR4 = 11, ++ SFF8024_ECC_25GBASE_CR_S = 12, ++ SFF8024_ECC_25GBASE_CR_N = 13, ++ SFF8024_ECC_10GBASE_T_SFI = 22, ++ SFF8024_ECC_10GBASE_T_SR = 28, ++ SFF8024_ECC_5GBASE_T = 29, ++ SFF8024_ECC_2_5GBASE_T = 30, ++}; ++ ++struct sfp_upstream_ops { ++ void (*attach)(void *, struct sfp_bus *); ++ void (*detach)(void *, struct sfp_bus *); ++ int (*module_insert)(void *, const struct sfp_eeprom_id *); ++ void (*module_remove)(void *); ++ int (*module_start)(void *); ++ void (*module_stop)(void *); ++ void (*link_down)(void *); ++ void (*link_up)(void *); ++ int (*connect_phy)(void *, struct phy_device *); ++ void (*disconnect_phy)(void *); ++}; ++ ++struct sfp_quirk { ++ const char *vendor; ++ const char *part; ++ void (*modes)(const struct sfp_eeprom_id *, long unsigned int *, long unsigned int *); ++ void (*fixup)(struct sfp *); ++}; ++ ++struct sfp_socket_ops { ++ void (*attach)(struct sfp *); ++ void (*detach)(struct sfp *); ++ void (*start)(struct sfp *); ++ void (*stop)(struct sfp *); ++ int (*module_info)(struct sfp *, struct ethtool_modinfo *); ++ int (*module_eeprom)(struct sfp *, struct ethtool_eeprom *, u8 *); ++ int (*module_eeprom_by_page)(struct sfp *, const struct ethtool_module_eeprom *, struct netlink_ext_ack *); ++}; ++ ++enum wwan_port_type { ++ WWAN_PORT_AT = 0, ++ WWAN_PORT_MBIM = 1, ++ WWAN_PORT_QMI = 2, ++ WWAN_PORT_QCDM = 3, ++ WWAN_PORT_FIREHOSE = 4, ++ WWAN_PORT_XMMRPC = 5, ++ __WWAN_PORT_MAX = 6, ++ WWAN_PORT_MAX = 5, ++ WWAN_PORT_UNKNOWN = 6, ++}; ++ ++struct wwan_port; ++ ++struct wwan_port_ops { ++ int (*start)(struct wwan_port *); ++ void (*stop)(struct wwan_port *); ++ int (*tx)(struct wwan_port *, struct sk_buff *); ++ int (*tx_blocking)(struct wwan_port *, struct sk_buff *); ++ __poll_t (*tx_poll)(struct wwan_port *, struct file *, poll_table *); ++}; ++ ++struct wwan_port { ++ enum wwan_port_type type; ++ unsigned int start_count; ++ long unsigned int flags; ++ const struct wwan_port_ops *ops; ++ struct mutex ops_lock; ++ struct device dev; ++ struct sk_buff_head rxq; ++ wait_queue_head_t waitqueue; ++ struct mutex data_lock; ++ size_t headroom_len; ++ size_t frag_len; ++ union { ++ struct { ++ struct ktermios termios; ++ int mdmbits; ++ } at_data; ++ }; ++}; ++ ++struct wwan_port_caps { ++ size_t frag_len; ++ unsigned int headroom_len; ++}; ++ ++struct wwan_netdev_priv { ++ u32 link_id; ++ long: 0; ++ u8 drv_priv[0]; ++}; ++ ++struct wwan_ops { ++ unsigned int priv_size; ++ void (*setup)(struct net_device *); ++ int (*newlink)(void *, struct net_device *, u32, struct netlink_ext_ack *); ++ void (*dellink)(void *, struct net_device *, struct list_head *); ++}; ++ ++enum { ++ IFLA_WWAN_UNSPEC = 0, ++ IFLA_WWAN_LINK_ID = 1, ++ __IFLA_WWAN_MAX = 2, ++}; ++ ++struct wwan_device { ++ unsigned int id; ++ struct device dev; ++ atomic_t port_id; ++ const struct wwan_ops *ops; ++ void *ops_ctxt; ++ struct dentry *debugfs_dir; ++}; ++ ++struct input_event { ++ __kernel_ulong_t __sec; ++ __kernel_ulong_t __usec; ++ __u16 type; ++ __u16 code; ++ __s32 value; ++}; ++ ++struct input_event_compat { ++ compat_ulong_t sec; ++ compat_ulong_t usec; ++ __u16 type; ++ __u16 code; ++ __s32 value; ++}; ++ ++struct ff_periodic_effect_compat { ++ __u16 waveform; ++ __u16 period; ++ __s16 magnitude; ++ __s16 offset; ++ __u16 phase; ++ struct ff_envelope envelope; ++ __u32 custom_len; ++ compat_uptr_t custom_data; ++}; ++ ++struct ff_effect_compat { ++ __u16 type; ++ __s16 id; ++ __u16 direction; ++ struct ff_trigger trigger; ++ struct ff_replay replay; ++ union { ++ struct ff_constant_effect constant; ++ struct ff_ramp_effect ramp; ++ struct ff_periodic_effect_compat periodic; ++ struct ff_condition_effect condition[2]; ++ struct ff_rumble_effect rumble; ++ } u; ++}; ++ ++struct lirc_scancode { ++ __u64 timestamp; ++ __u16 flags; ++ __u16 rc_proto; ++ __u32 keycode; ++ __u64 scancode; ++}; ++ ++enum rc_filter_type { ++ RC_FILTER_NORMAL = 0, ++ RC_FILTER_WAKEUP = 1, ++ RC_FILTER_MAX = 2, ++}; ++ ++struct led_trigger {}; ++ ++struct rc_filter_attribute { ++ struct device_attribute attr; ++ enum rc_filter_type type; ++ bool mask; ++}; ++ ++struct ir_raw_handler { ++ struct list_head list; ++ u64 protocols; ++ int (*decode)(struct rc_dev *, struct ir_raw_event); ++ int (*encode)(enum rc_proto, u32, struct ir_raw_event *, unsigned int); ++ u32 carrier; ++ u32 min_timeout; ++ int (*raw_register)(struct rc_dev *); ++ int (*raw_unregister)(struct rc_dev *); ++}; ++ ++struct ir_raw_timings_manchester { ++ unsigned int leader_pulse; ++ unsigned int leader_space; ++ unsigned int clock; ++ unsigned int invert: 1; ++ unsigned int trailer_space; ++}; ++ ++struct ir_raw_timings_pd { ++ unsigned int header_pulse; ++ unsigned int header_space; ++ unsigned int bit_pulse; ++ unsigned int bit_space[2]; ++ unsigned int trailer_pulse; ++ unsigned int trailer_space; ++ unsigned int msb_first: 1; ++}; ++ ++struct ir_raw_timings_pl { ++ unsigned int header_pulse; ++ unsigned int bit_space; ++ unsigned int bit_pulse[2]; ++ unsigned int trailer_space; ++ unsigned int msb_first: 1; ++}; ++ ++struct pps_ktime_compat { ++ __s64 sec; ++ __s32 nsec; ++ __u32 flags; ++}; ++ ++struct pps_kinfo { ++ __u32 assert_sequence; ++ __u32 clear_sequence; ++ struct pps_ktime assert_tu; ++ struct pps_ktime clear_tu; ++ int current_mode; ++}; ++ ++struct pps_kinfo_compat { ++ __u32 assert_sequence; ++ __u32 clear_sequence; ++ struct pps_ktime_compat assert_tu; ++ struct pps_ktime_compat clear_tu; ++ int current_mode; ++} __attribute__((packed)); ++ ++struct pps_fdata { ++ struct pps_kinfo info; ++ struct pps_ktime timeout; ++}; ++ ++struct pps_fdata_compat { ++ struct pps_kinfo_compat info; ++ struct pps_ktime_compat timeout; ++} __attribute__((packed)); ++ ++struct pps_bind_args { ++ int tsformat; ++ int edge; ++ int consumer; ++}; ++ ++struct ptp_clock_caps { ++ int max_adj; ++ int n_alarm; ++ int n_ext_ts; ++ int n_per_out; ++ int pps; ++ int n_pins; ++ int cross_timestamping; ++ int adjust_phase; ++ int rsv[12]; ++}; ++ ++struct ptp_sys_offset { ++ unsigned int n_samples; ++ unsigned int rsv[3]; ++ struct ptp_clock_time ts[51]; ++}; ++ ++struct ptp_sys_offset_extended { ++ unsigned int n_samples; ++ unsigned int rsv[3]; ++ struct ptp_clock_time ts[75]; ++}; ++ ++struct ptp_sys_offset_precise { ++ struct ptp_clock_time device; ++ struct ptp_clock_time sys_realtime; ++ struct ptp_clock_time sys_monoraw; ++ unsigned int rsv[4]; ++}; ++ ++enum { ++ POWER_SUPPLY_SCOPE_UNKNOWN = 0, ++ POWER_SUPPLY_SCOPE_SYSTEM = 1, ++ POWER_SUPPLY_SCOPE_DEVICE = 2, ++}; ++ ++enum power_supply_property { ++ POWER_SUPPLY_PROP_STATUS = 0, ++ POWER_SUPPLY_PROP_CHARGE_TYPE = 1, ++ POWER_SUPPLY_PROP_HEALTH = 2, ++ POWER_SUPPLY_PROP_PRESENT = 3, ++ POWER_SUPPLY_PROP_ONLINE = 4, ++ POWER_SUPPLY_PROP_AUTHENTIC = 5, ++ POWER_SUPPLY_PROP_TECHNOLOGY = 6, ++ POWER_SUPPLY_PROP_CYCLE_COUNT = 7, ++ POWER_SUPPLY_PROP_VOLTAGE_MAX = 8, ++ POWER_SUPPLY_PROP_VOLTAGE_MIN = 9, ++ POWER_SUPPLY_PROP_VOLTAGE_MAX_DESIGN = 10, ++ POWER_SUPPLY_PROP_VOLTAGE_MIN_DESIGN = 11, ++ POWER_SUPPLY_PROP_VOLTAGE_NOW = 12, ++ POWER_SUPPLY_PROP_VOLTAGE_AVG = 13, ++ POWER_SUPPLY_PROP_VOLTAGE_OCV = 14, ++ POWER_SUPPLY_PROP_VOLTAGE_BOOT = 15, ++ POWER_SUPPLY_PROP_CURRENT_MAX = 16, ++ POWER_SUPPLY_PROP_CURRENT_NOW = 17, ++ POWER_SUPPLY_PROP_CURRENT_AVG = 18, ++ POWER_SUPPLY_PROP_CURRENT_BOOT = 19, ++ POWER_SUPPLY_PROP_POWER_NOW = 20, ++ POWER_SUPPLY_PROP_POWER_AVG = 21, ++ POWER_SUPPLY_PROP_CHARGE_FULL_DESIGN = 22, ++ POWER_SUPPLY_PROP_CHARGE_EMPTY_DESIGN = 23, ++ POWER_SUPPLY_PROP_CHARGE_FULL = 24, ++ POWER_SUPPLY_PROP_CHARGE_EMPTY = 25, ++ POWER_SUPPLY_PROP_CHARGE_NOW = 26, ++ POWER_SUPPLY_PROP_CHARGE_AVG = 27, ++ POWER_SUPPLY_PROP_CHARGE_COUNTER = 28, ++ POWER_SUPPLY_PROP_CONSTANT_CHARGE_CURRENT = 29, ++ POWER_SUPPLY_PROP_CONSTANT_CHARGE_CURRENT_MAX = 30, ++ POWER_SUPPLY_PROP_CONSTANT_CHARGE_VOLTAGE = 31, ++ POWER_SUPPLY_PROP_CONSTANT_CHARGE_VOLTAGE_MAX = 32, ++ POWER_SUPPLY_PROP_CHARGE_CONTROL_LIMIT = 33, ++ POWER_SUPPLY_PROP_CHARGE_CONTROL_LIMIT_MAX = 34, ++ POWER_SUPPLY_PROP_CHARGE_CONTROL_START_THRESHOLD = 35, ++ POWER_SUPPLY_PROP_CHARGE_CONTROL_END_THRESHOLD = 36, ++ POWER_SUPPLY_PROP_CHARGE_BEHAVIOUR = 37, ++ POWER_SUPPLY_PROP_INPUT_CURRENT_LIMIT = 38, ++ POWER_SUPPLY_PROP_INPUT_VOLTAGE_LIMIT = 39, ++ POWER_SUPPLY_PROP_INPUT_POWER_LIMIT = 40, ++ POWER_SUPPLY_PROP_ENERGY_FULL_DESIGN = 41, ++ POWER_SUPPLY_PROP_ENERGY_EMPTY_DESIGN = 42, ++ POWER_SUPPLY_PROP_ENERGY_FULL = 43, ++ POWER_SUPPLY_PROP_ENERGY_EMPTY = 44, ++ POWER_SUPPLY_PROP_ENERGY_NOW = 45, ++ POWER_SUPPLY_PROP_ENERGY_AVG = 46, ++ POWER_SUPPLY_PROP_CAPACITY = 47, ++ POWER_SUPPLY_PROP_CAPACITY_ALERT_MIN = 48, ++ POWER_SUPPLY_PROP_CAPACITY_ALERT_MAX = 49, ++ POWER_SUPPLY_PROP_CAPACITY_ERROR_MARGIN = 50, ++ POWER_SUPPLY_PROP_CAPACITY_LEVEL = 51, ++ POWER_SUPPLY_PROP_TEMP = 52, ++ POWER_SUPPLY_PROP_TEMP_MAX = 53, ++ POWER_SUPPLY_PROP_TEMP_MIN = 54, ++ POWER_SUPPLY_PROP_TEMP_ALERT_MIN = 55, ++ POWER_SUPPLY_PROP_TEMP_ALERT_MAX = 56, ++ POWER_SUPPLY_PROP_TEMP_AMBIENT = 57, ++ POWER_SUPPLY_PROP_TEMP_AMBIENT_ALERT_MIN = 58, ++ POWER_SUPPLY_PROP_TEMP_AMBIENT_ALERT_MAX = 59, ++ POWER_SUPPLY_PROP_TIME_TO_EMPTY_NOW = 60, ++ POWER_SUPPLY_PROP_TIME_TO_EMPTY_AVG = 61, ++ POWER_SUPPLY_PROP_TIME_TO_FULL_NOW = 62, ++ POWER_SUPPLY_PROP_TIME_TO_FULL_AVG = 63, ++ POWER_SUPPLY_PROP_TYPE = 64, ++ POWER_SUPPLY_PROP_USB_TYPE = 65, ++ POWER_SUPPLY_PROP_SCOPE = 66, ++ POWER_SUPPLY_PROP_PRECHARGE_CURRENT = 67, ++ POWER_SUPPLY_PROP_CHARGE_TERM_CURRENT = 68, ++ POWER_SUPPLY_PROP_CALIBRATE = 69, ++ POWER_SUPPLY_PROP_MANUFACTURE_YEAR = 70, ++ POWER_SUPPLY_PROP_MANUFACTURE_MONTH = 71, ++ POWER_SUPPLY_PROP_MANUFACTURE_DAY = 72, ++ POWER_SUPPLY_PROP_MODEL_NAME = 73, ++ POWER_SUPPLY_PROP_MANUFACTURER = 74, ++ POWER_SUPPLY_PROP_SERIAL_NUMBER = 75, ++}; ++ ++enum power_supply_type { ++ POWER_SUPPLY_TYPE_UNKNOWN = 0, ++ POWER_SUPPLY_TYPE_BATTERY = 1, ++ POWER_SUPPLY_TYPE_UPS = 2, ++ POWER_SUPPLY_TYPE_MAINS = 3, ++ POWER_SUPPLY_TYPE_USB = 4, ++ POWER_SUPPLY_TYPE_USB_DCP = 5, ++ POWER_SUPPLY_TYPE_USB_CDP = 6, ++ POWER_SUPPLY_TYPE_USB_ACA = 7, ++ POWER_SUPPLY_TYPE_USB_TYPE_C = 8, ++ POWER_SUPPLY_TYPE_USB_PD = 9, ++ POWER_SUPPLY_TYPE_USB_PD_DRP = 10, ++ POWER_SUPPLY_TYPE_APPLE_BRICK_ID = 11, ++ POWER_SUPPLY_TYPE_WIRELESS = 12, ++}; ++ ++enum power_supply_usb_type { ++ POWER_SUPPLY_USB_TYPE_UNKNOWN = 0, ++ POWER_SUPPLY_USB_TYPE_SDP = 1, ++ POWER_SUPPLY_USB_TYPE_DCP = 2, ++ POWER_SUPPLY_USB_TYPE_CDP = 3, ++ POWER_SUPPLY_USB_TYPE_ACA = 4, ++ POWER_SUPPLY_USB_TYPE_C = 5, ++ POWER_SUPPLY_USB_TYPE_PD = 6, ++ POWER_SUPPLY_USB_TYPE_PD_DRP = 7, ++ POWER_SUPPLY_USB_TYPE_PD_PPS = 8, ++ POWER_SUPPLY_USB_TYPE_APPLE_BRICK_ID = 9, ++}; ++ ++enum power_supply_notifier_events { ++ PSY_EVENT_PROP_CHANGED = 0, ++}; ++ ++union power_supply_propval { ++ int intval; ++ const char *strval; ++}; ++ ++struct power_supply_config { ++ struct device_node *of_node; ++ struct fwnode_handle *fwnode; ++ void *drv_data; ++ const struct attribute_group **attr_grp; ++ char **supplied_to; ++ size_t num_supplicants; ++}; ++ ++struct power_supply; ++ ++struct power_supply_desc { ++ const char *name; ++ enum power_supply_type type; ++ const enum power_supply_usb_type *usb_types; ++ size_t num_usb_types; ++ const enum power_supply_property *properties; ++ size_t num_properties; ++ int (*get_property)(struct power_supply *, enum power_supply_property, union power_supply_propval *); ++ int (*set_property)(struct power_supply *, enum power_supply_property, const union power_supply_propval *); ++ int (*property_is_writeable)(struct power_supply *, enum power_supply_property); ++ void (*external_power_changed)(struct power_supply *); ++ void (*set_charged)(struct power_supply *); ++ bool no_thermal; ++ int use_for_apm; ++}; ++ ++struct power_supply { ++ const struct power_supply_desc *desc; ++ char **supplied_to; ++ size_t num_supplicants; ++ char **supplied_from; ++ size_t num_supplies; ++ struct device_node *of_node; ++ void *drv_data; ++ struct device dev; ++ struct work_struct changed_work; ++ struct delayed_work deferred_register_work; ++ spinlock_t changed_lock; ++ bool changed; ++ bool initialized; ++ bool removing; ++ atomic_t use_cnt; ++ struct power_supply_battery_info *battery_info; ++}; ++ ++struct psy_am_i_supplied_data { ++ struct power_supply *psy; ++ unsigned int count; ++}; ++ ++struct psy_get_supplier_prop_data { ++ struct power_supply *psy; ++ enum power_supply_property psp; ++ union power_supply_propval *val; ++}; ++ ++struct watchdog_info { ++ __u32 options; ++ __u32 firmware_version; ++ __u8 identity[32]; ++}; ++ ++struct watchdog_device; ++ ++struct watchdog_ops { ++ struct module *owner; ++ int (*start)(struct watchdog_device *); ++ int (*stop)(struct watchdog_device *); ++ int (*ping)(struct watchdog_device *); ++ unsigned int (*status)(struct watchdog_device *); ++ int (*set_timeout)(struct watchdog_device *, unsigned int); ++ int (*set_pretimeout)(struct watchdog_device *, unsigned int); ++ unsigned int (*get_timeleft)(struct watchdog_device *); ++ int (*restart)(struct watchdog_device *, long unsigned int, void *); ++ long int (*ioctl)(struct watchdog_device *, unsigned int, long unsigned int); ++}; ++ ++struct watchdog_governor; ++ ++struct watchdog_core_data; ++ ++struct watchdog_device { ++ int id; ++ struct device *parent; ++ const struct attribute_group **groups; ++ const struct watchdog_info *info; ++ const struct watchdog_ops *ops; ++ const struct watchdog_governor *gov; ++ unsigned int bootstatus; ++ unsigned int timeout; ++ unsigned int pretimeout; ++ unsigned int min_timeout; ++ unsigned int max_timeout; ++ unsigned int min_hw_heartbeat_ms; ++ unsigned int max_hw_heartbeat_ms; ++ struct notifier_block reboot_nb; ++ struct notifier_block restart_nb; ++ struct notifier_block pm_nb; ++ void *driver_data; ++ struct watchdog_core_data *wd_data; ++ long unsigned int status; ++ struct list_head deferred; ++}; ++ ++struct watchdog_governor { ++ const char name[20]; ++ void (*pretimeout)(struct watchdog_device *); ++}; ++ ++struct watchdog_core_data { ++ struct device dev; ++ struct cdev cdev; ++ struct watchdog_device *wdd; ++ struct mutex lock; ++ ktime_t last_keepalive; ++ ktime_t last_hw_keepalive; ++ ktime_t open_deadline; ++ struct hrtimer timer; ++ struct kthread_work work; ++ long unsigned int status; ++}; ++ ++struct trace_event_raw_watchdog_template { ++ struct trace_entry ent; ++ int id; ++ int err; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_watchdog_set_timeout { ++ struct trace_entry ent; ++ int id; ++ unsigned int timeout; ++ int err; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_watchdog_template {}; ++ ++struct trace_event_data_offsets_watchdog_set_timeout {}; ++ ++typedef void (*btf_trace_watchdog_start)(void *, struct watchdog_device *, int); ++ ++typedef void (*btf_trace_watchdog_ping)(void *, struct watchdog_device *, int); ++ ++typedef void (*btf_trace_watchdog_stop)(void *, struct watchdog_device *, int); ++ ++typedef void (*btf_trace_watchdog_set_timeout)(void *, struct watchdog_device *, unsigned int, int); ++ ++struct md_setup_args { ++ int minor; ++ int partitioned; ++ int level; ++ int chunk; ++ char *device_names; ++}; ++ ++enum dm_uevent_type { ++ DM_UEVENT_PATH_FAILED = 0, ++ DM_UEVENT_PATH_REINSTATED = 1, ++}; ++ ++struct dm_uevent { ++ struct mapped_device *md; ++ enum kobject_action action; ++ struct kobj_uevent_env ku_env; ++ struct list_head elist; ++ char name[128]; ++ char uuid[129]; ++}; ++ ++enum { ++ DM_TIO_INSIDE_DM_IO = 0, ++ DM_TIO_IS_DUPLICATE_BIO = 1, ++}; ++ ++enum { ++ DM_IO_ACCOUNTED = 0, ++ DM_IO_WAS_SPLIT = 1, ++}; ++ ++struct clone_info { ++ struct dm_table *map; ++ struct bio *bio; ++ struct dm_io *io; ++ sector_t sector; ++ unsigned int sector_count; ++ bool is_abnormal_io: 1; ++ bool submit_as_polled: 1; ++}; ++ ++struct table_device { ++ struct list_head list; ++ refcount_t count; ++ struct dm_dev dm_dev; ++}; ++ ++struct dm_pr { ++ u64 old_key; ++ u64 new_key; ++ u32 flags; ++ bool abort; ++ bool fail_early; ++ int ret; ++ enum pr_type type; ++}; ++ ++struct dm_io_client { ++ mempool_t pool; ++ struct bio_set bios; ++}; ++ ++struct io { ++ long unsigned int error_bits; ++ atomic_t count; ++ struct dm_io_client *client; ++ io_notify_fn callback; ++ void *context; ++ void *vma_invalidate_address; ++ long unsigned int vma_invalidate_size; ++ long: 64; ++}; ++ ++struct dpages { ++ void (*get_page)(struct dpages *, struct page **, long unsigned int *, unsigned int *); ++ void (*next_page)(struct dpages *); ++ union { ++ unsigned int context_u; ++ struct bvec_iter context_bi; ++ }; ++ void *context_ptr; ++ void *vma_invalidate_address; ++ long unsigned int vma_invalidate_size; ++}; ++ ++struct sync_io { ++ long unsigned int error_bits; ++ struct completion wait; ++}; ++ ++struct dm_stats_last_position { ++ sector_t last_sector; ++ unsigned int last_rw; ++}; ++ ++struct dm_stat_percpu { ++ long long unsigned int sectors[2]; ++ long long unsigned int ios[2]; ++ long long unsigned int merges[2]; ++ long long unsigned int ticks[2]; ++ long long unsigned int io_ticks[2]; ++ long long unsigned int io_ticks_total; ++ long long unsigned int time_in_queue; ++ long long unsigned int *histogram; ++}; ++ ++struct dm_stat_shared { ++ atomic_t in_flight[2]; ++ long long unsigned int stamp; ++ struct dm_stat_percpu tmp; ++}; ++ ++struct dm_stat { ++ struct list_head list_entry; ++ int id; ++ unsigned int stat_flags; ++ size_t n_entries; ++ sector_t start; ++ sector_t end; ++ sector_t step; ++ unsigned int n_histogram_entries; ++ long long unsigned int *histogram_boundaries; ++ const char *program_id; ++ const char *aux_data; ++ struct callback_head callback_head; ++ size_t shared_alloc_size; ++ size_t percpu_alloc_size; ++ size_t histogram_alloc_size; ++ struct dm_stat_percpu *stat_percpu[512]; ++ struct dm_stat_shared stat_shared[0]; ++}; ++ ++struct dm_bio_details { ++ struct block_device *bi_bdev; ++ int __bi_remaining; ++ long unsigned int bi_flags; ++ struct bvec_iter bi_iter; ++ bio_end_io_t *bi_end_io; ++ struct bio_integrity_payload *bi_integrity; ++}; ++ ++enum dm_raid1_error { ++ DM_RAID1_WRITE_ERROR = 0, ++ DM_RAID1_FLUSH_ERROR = 1, ++ DM_RAID1_SYNC_ERROR = 2, ++ DM_RAID1_READ_ERROR = 3, ++}; ++ ++struct mirror_set; ++ ++struct mirror { ++ struct mirror_set *ms; ++ atomic_t error_count; ++ long unsigned int error_type; ++ struct dm_dev *dev; ++ sector_t offset; ++}; ++ ++struct mirror_set { ++ struct dm_target *ti; ++ struct list_head list; ++ uint64_t features; ++ spinlock_t lock; ++ struct bio_list reads; ++ struct bio_list writes; ++ struct bio_list failures; ++ struct bio_list holds; ++ struct dm_region_hash *rh; ++ struct dm_kcopyd_client *kcopyd_client; ++ struct dm_io_client *io_client; ++ region_t nr_regions; ++ int in_sync; ++ int log_failure; ++ int leg_failure; ++ atomic_t suspend; ++ atomic_t default_mirror; ++ struct workqueue_struct *kmirrord_wq; ++ struct work_struct kmirrord_work; ++ struct timer_list timer; ++ long unsigned int timer_pending; ++ struct work_struct trigger_event; ++ unsigned int nr_mirrors; ++ struct mirror mirror[0]; ++}; ++ ++struct dm_raid1_bio_record { ++ struct mirror *m; ++ struct dm_bio_details details; ++ region_t write_region; ++}; ++ ++struct hwspinlock_ops { ++ int (*trylock)(struct hwspinlock *); ++ void (*unlock)(struct hwspinlock *); ++ void (*relax)(struct hwspinlock *); ++}; ++ ++struct hwspinlock_device; ++ ++struct hwspinlock { ++ struct hwspinlock_device *bank; ++ spinlock_t lock; ++ void *priv; ++}; ++ ++struct hwspinlock_device { ++ struct device *dev; ++ const struct hwspinlock_ops *ops; ++ int base_id; ++ int num_locks; ++ struct hwspinlock lock[0]; ++}; ++ ++struct powercap_control_type; ++ ++struct powercap_control_type_ops { ++ int (*set_enable)(struct powercap_control_type *, bool); ++ int (*get_enable)(struct powercap_control_type *, bool *); ++ int (*release)(struct powercap_control_type *); ++}; ++ ++struct powercap_control_type { ++ struct device dev; ++ struct idr idr; ++ int nr_zones; ++ const struct powercap_control_type_ops *ops; ++ struct mutex lock; ++ bool allocated; ++ struct list_head node; ++}; ++ ++struct powercap_zone; ++ ++struct powercap_zone_ops { ++ int (*get_max_energy_range_uj)(struct powercap_zone *, u64 *); ++ int (*get_energy_uj)(struct powercap_zone *, u64 *); ++ int (*reset_energy_uj)(struct powercap_zone *); ++ int (*get_max_power_range_uw)(struct powercap_zone *, u64 *); ++ int (*get_power_uw)(struct powercap_zone *, u64 *); ++ int (*set_enable)(struct powercap_zone *, bool); ++ int (*get_enable)(struct powercap_zone *, bool *); ++ int (*release)(struct powercap_zone *); ++}; ++ ++struct powercap_zone_constraint; ++ ++struct powercap_zone { ++ int id; ++ char *name; ++ void *control_type_inst; ++ const struct powercap_zone_ops *ops; ++ struct device dev; ++ int const_id_cnt; ++ struct idr idr; ++ struct idr *parent_idr; ++ void *private_data; ++ struct attribute **zone_dev_attrs; ++ int zone_attr_count; ++ struct attribute_group dev_zone_attr_group; ++ const struct attribute_group *dev_attr_groups[2]; ++ bool allocated; ++ struct powercap_zone_constraint *constraints; ++}; ++ ++struct powercap_zone_constraint_ops; ++ ++struct powercap_zone_constraint { ++ int id; ++ struct powercap_zone *power_zone; ++ const struct powercap_zone_constraint_ops *ops; ++}; ++ ++struct powercap_zone_constraint_ops { ++ int (*set_power_limit_uw)(struct powercap_zone *, int, u64); ++ int (*get_power_limit_uw)(struct powercap_zone *, int, u64 *); ++ int (*set_time_window_us)(struct powercap_zone *, int, u64); ++ int (*get_time_window_us)(struct powercap_zone *, int, u64 *); ++ int (*get_max_power_uw)(struct powercap_zone *, int, u64 *); ++ int (*get_min_power_uw)(struct powercap_zone *, int, u64 *); ++ int (*get_max_time_window_us)(struct powercap_zone *, int, u64 *); ++ int (*get_min_time_window_us)(struct powercap_zone *, int, u64 *); ++ const char * (*get_name)(struct powercap_zone *, int); ++}; ++ ++struct powercap_constraint_attr { ++ struct device_attribute power_limit_attr; ++ struct device_attribute time_window_attr; ++ struct device_attribute max_power_attr; ++ struct device_attribute min_power_attr; ++ struct device_attribute max_time_window_attr; ++ struct device_attribute min_time_window_attr; ++ struct device_attribute name_attr; ++}; ++ ++enum icl_lc_mailbox_cmd { ++ ICL_LC_GO2SX = 2, ++ ICL_LC_GO2SX_NO_WAKE = 3, ++ ICL_LC_PREPARE_FOR_RESET = 33, ++}; ++ ++struct cfg_error_pkg { ++ struct tb_cfg_header header; ++ enum tb_cfg_error error: 8; ++ u32 port: 6; ++ u32 reserved: 16; ++ u32 pg: 2; ++}; ++ ++struct cfg_event_pkg { ++ struct tb_cfg_header header; ++ u32 port: 6; ++ u32 zero: 25; ++ bool unplug: 1; ++}; ++ ++struct tb_cm { ++ struct list_head tunnel_list; ++ struct list_head dp_resources; ++ bool hotplug_active; ++ struct delayed_work remove_work; ++ struct tb_bandwidth_group groups[7]; ++}; ++ ++struct tb_hotplug_event { ++ struct work_struct work; ++ struct tb *tb; ++ u64 route; ++ u8 port; ++ bool unplug; ++}; ++ ++struct tb_regs_hop { ++ u32 next_hop: 11; ++ u32 out_port: 6; ++ u32 initial_credits: 8; ++ u32 unknown1: 6; ++ bool enable: 1; ++ u32 weight: 4; ++ u32 unknown2: 4; ++ u32 priority: 3; ++ bool drop_packages: 1; ++ u32 counter: 11; ++ bool counter_enable: 1; ++ bool ingress_fc: 1; ++ bool egress_fc: 1; ++ bool ingress_shared_buffer: 1; ++ bool egress_shared_buffer: 1; ++ bool pending: 1; ++ u32 unknown3: 3; ++}; ++ ++struct tb_service_id { ++ __u32 match_flags; ++ char protocol_key[9]; ++ __u32 protocol_id; ++ __u32 protocol_version; ++ __u32 protocol_revision; ++ kernel_ulong_t driver_data; ++}; ++ ++struct tb_service { ++ struct device dev; ++ int id; ++ const char *key; ++ u32 prtcid; ++ u32 prtcvers; ++ u32 prtcrevs; ++ u32 prtcstns; ++ struct dentry *debugfs_dir; ++}; ++ ++struct tb_service_driver { ++ struct device_driver driver; ++ int (*probe)(struct tb_service *, const struct tb_service_id *); ++ void (*remove)(struct tb_service *); ++ void (*shutdown)(struct tb_service *); ++ const struct tb_service_id *id_table; ++}; ++ ++typedef bool (*event_cb)(void *, enum tb_cfg_pkg_type, const void *, size_t); ++ ++enum tb_property_type { ++ TB_PROPERTY_TYPE_UNKNOWN = 0, ++ TB_PROPERTY_TYPE_DIRECTORY = 68, ++ TB_PROPERTY_TYPE_DATA = 100, ++ TB_PROPERTY_TYPE_TEXT = 116, ++ TB_PROPERTY_TYPE_VALUE = 118, ++}; ++ ++struct tb_property { ++ struct list_head list; ++ char key[9]; ++ enum tb_property_type type; ++ size_t length; ++ union { ++ struct tb_property_dir *dir; ++ u8 *data; ++ char *text; ++ u32 immediate; ++ } value; ++}; ++ ++struct tb_property_entry { ++ u32 key_hi; ++ u32 key_lo; ++ u16 length; ++ u8 reserved; ++ u8 type; ++ u32 value; ++}; ++ ++struct tb_property_rootdir_entry { ++ u32 magic; ++ u32 length; ++ struct tb_property_entry entries[0]; ++}; ++ ++struct tb_property_dir_entry { ++ u32 uuid[4]; ++ struct tb_property_entry entries[0]; ++}; ++ ++struct tb_protocol_handler { ++ const uuid_t *uuid; ++ int (*callback)(const void *, size_t, void *); ++ void *data; ++ struct list_head list; ++}; ++ ++struct tb_xdomain_header { ++ u32 route_hi; ++ u32 route_lo; ++ u32 length_sn; ++}; ++ ++enum tb_xdp_type { ++ UUID_REQUEST_OLD = 1, ++ UUID_RESPONSE = 2, ++ PROPERTIES_REQUEST = 3, ++ PROPERTIES_RESPONSE = 4, ++ PROPERTIES_CHANGED_REQUEST = 5, ++ PROPERTIES_CHANGED_RESPONSE = 6, ++ ERROR_RESPONSE = 7, ++ UUID_REQUEST = 12, ++ LINK_STATE_STATUS_REQUEST = 15, ++ LINK_STATE_STATUS_RESPONSE = 16, ++ LINK_STATE_CHANGE_REQUEST = 17, ++ LINK_STATE_CHANGE_RESPONSE = 18, ++}; ++ ++struct tb_xdp_header { ++ struct tb_xdomain_header xd_hdr; ++ uuid_t uuid; ++ u32 type; ++}; ++ ++struct tb_xdp_error_response { ++ struct tb_xdp_header hdr; ++ u32 error; ++}; ++ ++struct tb_xdp_link_state_status { ++ struct tb_xdp_header hdr; ++}; ++ ++struct tb_xdp_link_state_status_response { ++ union { ++ struct tb_xdp_error_response err; ++ struct { ++ struct tb_xdp_header hdr; ++ u32 status; ++ u8 slw; ++ u8 tlw; ++ u8 sls; ++ u8 tls; ++ }; ++ }; ++}; ++ ++struct tb_xdp_link_state_change { ++ struct tb_xdp_header hdr; ++ u8 tlw; ++ u8 tls; ++ u16 reserved; ++}; ++ ++struct tb_xdp_link_state_change_response { ++ union { ++ struct tb_xdp_error_response err; ++ struct { ++ struct tb_xdp_header hdr; ++ u32 status; ++ }; ++ }; ++}; ++ ++struct tb_xdp_uuid { ++ struct tb_xdp_header hdr; ++}; ++ ++struct tb_xdp_uuid_response { ++ union { ++ struct tb_xdp_error_response err; ++ struct { ++ struct tb_xdp_header hdr; ++ uuid_t src_uuid; ++ u32 src_route_hi; ++ u32 src_route_lo; ++ }; ++ }; ++}; ++ ++struct tb_xdp_properties { ++ struct tb_xdp_header hdr; ++ uuid_t src_uuid; ++ uuid_t dst_uuid; ++ u16 offset; ++ u16 reserved; ++}; ++ ++struct tb_xdp_properties_response { ++ union { ++ struct tb_xdp_error_response err; ++ struct { ++ struct tb_xdp_header hdr; ++ uuid_t src_uuid; ++ uuid_t dst_uuid; ++ u16 offset; ++ u16 data_length; ++ u32 generation; ++ u32 data[0]; ++ }; ++ }; ++}; ++ ++struct tb_xdp_properties_changed { ++ struct tb_xdp_header hdr; ++ uuid_t src_uuid; ++}; ++ ++struct tb_xdp_properties_changed_response { ++ union { ++ struct tb_xdp_error_response err; ++ struct tb_xdp_header hdr; ++ }; ++}; ++ ++enum tb_xdp_error { ++ ERROR_SUCCESS = 0, ++ ERROR_UNKNOWN_PACKET = 1, ++ ERROR_UNKNOWN_DOMAIN = 2, ++ ERROR_NOT_SUPPORTED = 3, ++ ERROR_NOT_READY = 4, ++}; ++ ++enum { ++ XDOMAIN_STATE_INIT = 0, ++ XDOMAIN_STATE_UUID = 1, ++ XDOMAIN_STATE_LINK_STATUS = 2, ++ XDOMAIN_STATE_LINK_STATE_CHANGE = 3, ++ XDOMAIN_STATE_LINK_STATUS2 = 4, ++ XDOMAIN_STATE_BONDING_UUID_LOW = 5, ++ XDOMAIN_STATE_BONDING_UUID_HIGH = 6, ++ XDOMAIN_STATE_PROPERTIES = 7, ++ XDOMAIN_STATE_ENUMERATED = 8, ++ XDOMAIN_STATE_ERROR = 9, ++}; ++ ++struct xdomain_request_work { ++ struct work_struct work; ++ struct tb_xdp_header *pkg; ++ struct tb *tb; ++}; ++ ++struct tb_xdomain_lookup { ++ const uuid_t *uuid; ++ u8 link; ++ u8 depth; ++ u64 route; ++}; ++ ++typedef int (*nvmem_reg_read_t)(void *, unsigned int, void *, size_t); ++ ++typedef int (*nvmem_reg_write_t)(void *, unsigned int, void *, size_t); ++ ++typedef int (*nvmem_cell_post_process_t)(void *, const char *, int, unsigned int, void *, size_t); ++ ++enum nvmem_type { ++ NVMEM_TYPE_UNKNOWN = 0, ++ NVMEM_TYPE_EEPROM = 1, ++ NVMEM_TYPE_OTP = 2, ++ NVMEM_TYPE_BATTERY_BACKED = 3, ++ NVMEM_TYPE_FRAM = 4, ++}; ++ ++struct nvmem_keepout { ++ unsigned int start; ++ unsigned int end; ++ unsigned char value; ++}; ++ ++struct nvmem_cell_info { ++ const char *name; ++ unsigned int offset; ++ size_t raw_len; ++ unsigned int bytes; ++ unsigned int bit_offset; ++ unsigned int nbits; ++ struct device_node *np; ++ nvmem_cell_post_process_t read_post_process; ++ void *priv; ++}; ++ ++struct nvmem_layout; ++ ++struct nvmem_config { ++ struct device *dev; ++ const char *name; ++ int id; ++ struct module *owner; ++ const struct nvmem_cell_info *cells; ++ int ncells; ++ const struct nvmem_keepout *keepout; ++ unsigned int nkeepout; ++ enum nvmem_type type; ++ bool read_only; ++ bool root_only; ++ bool ignore_wp; ++ struct nvmem_layout *layout; ++ struct device_node *of_node; ++ bool no_of_node; ++ nvmem_reg_read_t reg_read; ++ nvmem_reg_write_t reg_write; ++ int size; ++ int word_size; ++ int stride; ++ void *priv; ++ bool compat; ++ struct device *base_dev; ++}; ++ ++struct nvmem_layout { ++ const char *name; ++ const struct of_device_id *of_match_table; ++ int (*add_cells)(struct device *, struct nvmem_device *, struct nvmem_layout *); ++ void (*fixup_cell_info)(struct nvmem_device *, struct nvmem_layout *, struct nvmem_cell_info *); ++ struct module *owner; ++ struct list_head node; ++}; ++ ++struct tb_retimer { ++ struct device dev; ++ struct tb *tb; ++ u8 index; ++ u32 vendor; ++ u32 device; ++ struct tb_port *port; ++ struct tb_nvm *nvm; ++ bool no_nvm_upgrade; ++ u32 auth_status; ++}; ++ ++struct tb_nvm_vendor { ++ u16 vendor; ++ const struct tb_nvm_vendor_ops *vops; ++}; ++ ++struct icc_node; ++ ++struct icc_req { ++ struct hlist_node req_node; ++ struct icc_node *node; ++ struct device *dev; ++ bool enabled; ++ u32 tag; ++ u32 avg_bw; ++ u32 peak_bw; ++}; ++ ++struct icc_path { ++ const char *name; ++ size_t num_nodes; ++ struct icc_req reqs[0]; ++}; ++ ++struct icc_node_data { ++ struct icc_node *node; ++ u32 tag; ++}; ++ ++struct icc_provider; ++ ++struct icc_node { ++ int id; ++ const char *name; ++ struct icc_node **links; ++ size_t num_links; ++ struct icc_provider *provider; ++ struct list_head node_list; ++ struct list_head search_list; ++ struct icc_node *reverse; ++ u8 is_traversed: 1; ++ struct hlist_head req_list; ++ u32 avg_bw; ++ u32 peak_bw; ++ u32 init_avg; ++ u32 init_peak; ++ void *data; ++}; ++ ++struct icc_onecell_data { ++ unsigned int num_nodes; ++ struct icc_node *nodes[0]; ++}; ++ ++struct icc_provider { ++ struct list_head provider_list; ++ struct list_head nodes; ++ int (*set)(struct icc_node *, struct icc_node *); ++ int (*aggregate)(struct icc_node *, u32, u32, u32, u32 *, u32 *); ++ void (*pre_aggregate)(struct icc_node *); ++ int (*get_bw)(struct icc_node *, u32 *, u32 *); ++ struct icc_node * (*xlate)(struct of_phandle_args *, void *); ++ struct icc_node_data * (*xlate_extended)(struct of_phandle_args *, void *); ++ struct device *dev; ++ int users; ++ bool inter_set; ++ void *data; ++}; ++ ++struct trace_event_raw_icc_set_bw { ++ struct trace_entry ent; ++ u32 __data_loc_path_name; ++ u32 __data_loc_dev; ++ u32 __data_loc_node_name; ++ u32 avg_bw; ++ u32 peak_bw; ++ u32 node_avg_bw; ++ u32 node_peak_bw; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_icc_set_bw_end { ++ struct trace_entry ent; ++ u32 __data_loc_path_name; ++ u32 __data_loc_dev; ++ int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_icc_set_bw { ++ u32 path_name; ++ u32 dev; ++ u32 node_name; ++}; ++ ++struct trace_event_data_offsets_icc_set_bw_end { ++ u32 path_name; ++ u32 dev; ++}; ++ ++typedef void (*btf_trace_icc_set_bw)(void *, struct icc_path *, struct icc_node *, int, u32, u32); ++ ++typedef void (*btf_trace_icc_set_bw_end)(void *, struct icc_path *, int); ++ ++enum txtime_flags { ++ SOF_TXTIME_DEADLINE_MODE = 1, ++ SOF_TXTIME_REPORT_ERRORS = 2, ++ SOF_TXTIME_FLAGS_LAST = 2, ++ SOF_TXTIME_FLAGS_MASK = 3, ++}; ++ ++struct sock_txtime { ++ __kernel_clockid_t clockid; ++ __u32 flags; ++}; ++ ++enum sk_pacing { ++ SK_PACING_NONE = 0, ++ SK_PACING_NEEDED = 1, ++ SK_PACING_FQ = 2, ++}; ++ ++struct cgroup_cls_state { ++ struct cgroup_subsys_state css; ++ u32 classid; ++}; ++ ++enum { ++ SK_MEMINFO_RMEM_ALLOC = 0, ++ SK_MEMINFO_RCVBUF = 1, ++ SK_MEMINFO_WMEM_ALLOC = 2, ++ SK_MEMINFO_SNDBUF = 3, ++ SK_MEMINFO_FWD_ALLOC = 4, ++ SK_MEMINFO_WMEM_QUEUED = 5, ++ SK_MEMINFO_OPTMEM = 6, ++ SK_MEMINFO_BACKLOG = 7, ++ SK_MEMINFO_DROPS = 8, ++ SK_MEMINFO_VARS = 9, ++}; ++ ++enum sknetlink_groups { ++ SKNLGRP_NONE = 0, ++ SKNLGRP_INET_TCP_DESTROY = 1, ++ SKNLGRP_INET_UDP_DESTROY = 2, ++ SKNLGRP_INET6_TCP_DESTROY = 3, ++ SKNLGRP_INET6_UDP_DESTROY = 4, ++ __SKNLGRP_MAX = 5, ++}; ++ ++struct scm_timestamping { ++ struct __kernel_old_timespec ts[3]; ++}; ++ ++struct scm_timestamping64 { ++ struct __kernel_timespec ts[3]; ++}; ++ ++struct ifbond { ++ __s32 bond_mode; ++ __s32 num_slaves; ++ __s32 miimon; ++}; ++ ++typedef struct ifbond ifbond; ++ ++struct ifslave { ++ __s32 slave_id; ++ char slave_name[16]; ++ __s8 link; ++ __s8 state; ++ __u32 link_failure_count; ++}; ++ ++typedef struct ifslave ifslave; ++ ++enum { ++ NAPIF_STATE_SCHED = 1, ++ NAPIF_STATE_MISSED = 2, ++ NAPIF_STATE_DISABLE = 4, ++ NAPIF_STATE_NPSVC = 8, ++ NAPIF_STATE_LISTED = 16, ++ NAPIF_STATE_NO_BUSY_POLL = 32, ++ NAPIF_STATE_IN_BUSY_POLL = 64, ++ NAPIF_STATE_PREFER_BUSY_POLL = 128, ++ NAPIF_STATE_THREADED = 256, ++ NAPIF_STATE_SCHED_THREADED = 512, ++}; ++ ++struct net_device_path_stack { ++ int num_paths; ++ struct net_device_path path[5]; ++}; ++ ++struct bpf_xdp_link { ++ struct bpf_link link; ++ struct net_device *dev; ++ int flags; ++}; ++ ++struct netdev_net_notifier { ++ struct list_head list; ++ struct notifier_block *nb; ++}; ++ ++struct netdev_notifier_change_info { ++ struct netdev_notifier_info info; ++ unsigned int flags_changed; ++}; ++ ++struct netdev_notifier_changelowerstate_info { ++ struct netdev_notifier_info info; ++ void *lower_state_info; ++}; ++ ++struct netdev_notifier_pre_changeaddr_info { ++ struct netdev_notifier_info info; ++ const unsigned char *dev_addr; ++}; ++ ++struct netdev_notifier_offload_xstats_rd { ++ struct rtnl_hw_stats64 stats; ++ bool used; ++}; ++ ++struct netdev_notifier_offload_xstats_ru { ++ bool used; ++}; ++ ++struct netdev_notifier_offload_xstats_info { ++ struct netdev_notifier_info info; ++ enum netdev_offload_xstats_type type; ++ union { ++ struct netdev_notifier_offload_xstats_rd *report_delta; ++ struct netdev_notifier_offload_xstats_ru *report_used; ++ }; ++}; ++ ++enum { ++ NESTED_SYNC_IMM_BIT = 0, ++ NESTED_SYNC_TODO_BIT = 1, ++}; ++ ++struct netdev_bonding_info { ++ ifslave slave; ++ ifbond master; ++}; ++ ++struct netdev_notifier_bonding_info { ++ struct netdev_notifier_info info; ++ struct netdev_bonding_info bonding_info; ++}; ++ ++typedef int (*bpf_op_t)(struct net_device *, struct netdev_bpf *); ++ ++struct dev_kfree_skb_cb { ++ enum skb_drop_reason reason; ++}; ++ ++struct netdev_adjacent { ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ bool master; ++ bool ignore; ++ u16 ref_nr; ++ void *private; ++ struct list_head list; ++ struct callback_head rcu; ++}; ++ ++typedef u64 u_int64_t; ++ ++enum { ++ BPF_F_RECOMPUTE_CSUM = 1, ++ BPF_F_INVALIDATE_HASH = 2, ++}; ++ ++enum { ++ BPF_F_HDR_FIELD_MASK = 15, ++}; ++ ++enum { ++ BPF_F_PSEUDO_HDR = 16, ++ BPF_F_MARK_MANGLED_0 = 32, ++ BPF_F_MARK_ENFORCE = 64, ++}; ++ ++enum { ++ BPF_F_TUNINFO_IPV6 = 1, ++}; ++ ++enum { ++ BPF_F_ZERO_CSUM_TX = 2, ++ BPF_F_DONT_FRAGMENT = 4, ++ BPF_F_SEQ_NUMBER = 8, ++ BPF_F_NO_TUNNEL_KEY = 16, ++}; ++ ++enum { ++ BPF_F_TUNINFO_FLAGS = 16, ++}; ++ ++enum { ++ BPF_CSUM_LEVEL_QUERY = 0, ++ BPF_CSUM_LEVEL_INC = 1, ++ BPF_CSUM_LEVEL_DEC = 2, ++ BPF_CSUM_LEVEL_RESET = 3, ++}; ++ ++enum { ++ BPF_F_ADJ_ROOM_FIXED_GSO = 1, ++ BPF_F_ADJ_ROOM_ENCAP_L3_IPV4 = 2, ++ BPF_F_ADJ_ROOM_ENCAP_L3_IPV6 = 4, ++ BPF_F_ADJ_ROOM_ENCAP_L4_GRE = 8, ++ BPF_F_ADJ_ROOM_ENCAP_L4_UDP = 16, ++ BPF_F_ADJ_ROOM_NO_CSUM_RESET = 32, ++ BPF_F_ADJ_ROOM_ENCAP_L2_ETH = 64, ++ BPF_F_ADJ_ROOM_DECAP_L3_IPV4 = 128, ++ BPF_F_ADJ_ROOM_DECAP_L3_IPV6 = 256, ++}; ++ ++enum { ++ BPF_ADJ_ROOM_ENCAP_L2_MASK = 255, ++ BPF_ADJ_ROOM_ENCAP_L2_SHIFT = 56, ++}; ++ ++enum { ++ BPF_SK_LOOKUP_F_REPLACE = 1, ++ BPF_SK_LOOKUP_F_NO_REUSEPORT = 2, ++}; ++ ++enum bpf_adj_room_mode { ++ BPF_ADJ_ROOM_NET = 0, ++ BPF_ADJ_ROOM_MAC = 1, ++}; ++ ++enum bpf_hdr_start_off { ++ BPF_HDR_START_MAC = 0, ++ BPF_HDR_START_NET = 1, ++}; ++ ++enum bpf_lwt_encap_mode { ++ BPF_LWT_ENCAP_SEG6 = 0, ++ BPF_LWT_ENCAP_SEG6_INLINE = 1, ++ BPF_LWT_ENCAP_IP = 2, ++}; ++ ++enum { ++ BPF_SKB_TSTAMP_UNSPEC = 0, ++ BPF_SKB_TSTAMP_DELIVERY_MONO = 1, ++}; ++ ++struct bpf_tunnel_key { ++ __u32 tunnel_id; ++ union { ++ __u32 remote_ipv4; ++ __u32 remote_ipv6[4]; ++ }; ++ __u8 tunnel_tos; ++ __u8 tunnel_ttl; ++ union { ++ __u16 tunnel_ext; ++ __be16 tunnel_flags; ++ }; ++ __u32 tunnel_label; ++ union { ++ __u32 local_ipv4; ++ __u32 local_ipv6[4]; ++ }; ++}; ++ ++struct bpf_xfrm_state { ++ __u32 reqid; ++ __u32 spi; ++ __u16 family; ++ __u16 ext; ++ union { ++ __u32 remote_ipv4; ++ __u32 remote_ipv6[4]; ++ }; ++}; ++ ++struct bpf_tcp_sock { ++ __u32 snd_cwnd; ++ __u32 srtt_us; ++ __u32 rtt_min; ++ __u32 snd_ssthresh; ++ __u32 rcv_nxt; ++ __u32 snd_nxt; ++ __u32 snd_una; ++ __u32 mss_cache; ++ __u32 ecn_flags; ++ __u32 rate_delivered; ++ __u32 rate_interval_us; ++ __u32 packets_out; ++ __u32 retrans_out; ++ __u32 total_retrans; ++ __u32 segs_in; ++ __u32 data_segs_in; ++ __u32 segs_out; ++ __u32 data_segs_out; ++ __u32 lost_out; ++ __u32 sacked_out; ++ __u64 bytes_received; ++ __u64 bytes_acked; ++ __u32 dsack_dups; ++ __u32 delivered; ++ __u32 delivered_ce; ++ __u32 icsk_retransmits; ++}; ++ ++struct bpf_sock_tuple { ++ union { ++ struct { ++ __be32 saddr; ++ __be32 daddr; ++ __be16 sport; ++ __be16 dport; ++ } ipv4; ++ struct { ++ __be32 saddr[4]; ++ __be32 daddr[4]; ++ __be16 sport; ++ __be16 dport; ++ } ipv6; ++ }; ++}; ++ ++struct bpf_xdp_sock { ++ __u32 queue_id; ++}; ++ ++enum { ++ TCP_BPF_IW = 1001, ++ TCP_BPF_SNDCWND_CLAMP = 1002, ++ TCP_BPF_DELACK_MAX = 1003, ++ TCP_BPF_RTO_MIN = 1004, ++ TCP_BPF_SYN = 1005, ++ TCP_BPF_SYN_IP = 1006, ++ TCP_BPF_SYN_MAC = 1007, ++}; ++ ++enum { ++ BPF_LOAD_HDR_OPT_TCP_SYN = 1, ++}; ++ ++enum { ++ BPF_FIB_LOOKUP_DIRECT = 1, ++ BPF_FIB_LOOKUP_OUTPUT = 2, ++ BPF_FIB_LOOKUP_SKIP_NEIGH = 4, ++}; ++ ++enum { ++ BPF_FIB_LKUP_RET_SUCCESS = 0, ++ BPF_FIB_LKUP_RET_BLACKHOLE = 1, ++ BPF_FIB_LKUP_RET_UNREACHABLE = 2, ++ BPF_FIB_LKUP_RET_PROHIBIT = 3, ++ BPF_FIB_LKUP_RET_NOT_FWDED = 4, ++ BPF_FIB_LKUP_RET_FWD_DISABLED = 5, ++ BPF_FIB_LKUP_RET_UNSUPP_LWT = 6, ++ BPF_FIB_LKUP_RET_NO_NEIGH = 7, ++ BPF_FIB_LKUP_RET_FRAG_NEEDED = 8, ++}; ++ ++struct bpf_fib_lookup { ++ __u8 family; ++ __u8 l4_protocol; ++ __be16 sport; ++ __be16 dport; ++ union { ++ __u16 tot_len; ++ __u16 mtu_result; ++ }; ++ __u32 ifindex; ++ union { ++ __u8 tos; ++ __be32 flowinfo; ++ __u32 rt_metric; ++ }; ++ union { ++ __be32 ipv4_src; ++ __u32 ipv6_src[4]; ++ }; ++ union { ++ __be32 ipv4_dst; ++ __u32 ipv6_dst[4]; ++ }; ++ __be16 h_vlan_proto; ++ __be16 h_vlan_TCI; ++ __u8 smac[6]; ++ __u8 dmac[6]; ++}; ++ ++struct bpf_redir_neigh { ++ __u32 nh_family; ++ union { ++ __be32 ipv4_nh; ++ __u32 ipv6_nh[4]; ++ }; ++}; ++ ++enum bpf_check_mtu_flags { ++ BPF_MTU_CHK_SEGS = 1, ++}; ++ ++enum bpf_check_mtu_ret { ++ BPF_MTU_CHK_RET_SUCCESS = 0, ++ BPF_MTU_CHK_RET_FRAG_NEEDED = 1, ++ BPF_MTU_CHK_RET_SEGS_TOOBIG = 2, ++}; ++ ++union nf_inet_addr { ++ __u32 all[4]; ++ __be32 ip; ++ __be32 ip6[4]; ++ struct in_addr in; ++ struct in6_addr in6; ++}; ++ ++struct ip_ct_tcp_state { ++ u_int32_t td_end; ++ u_int32_t td_maxend; ++ u_int32_t td_maxwin; ++ u_int32_t td_maxack; ++ u_int8_t td_scale; ++ u_int8_t flags; ++}; ++ ++struct ip_ct_tcp { ++ struct ip_ct_tcp_state seen[2]; ++ u_int8_t state; ++ u_int8_t last_dir; ++ u_int8_t retrans; ++ u_int8_t last_index; ++ u_int32_t last_seq; ++ u_int32_t last_ack; ++ u_int32_t last_end; ++ u_int16_t last_win; ++ u_int8_t last_wscale; ++ u_int8_t last_flags; ++}; ++ ++union nf_conntrack_man_proto { ++ __be16 all; ++ struct { ++ __be16 port; ++ } tcp; ++ struct { ++ __be16 port; ++ } udp; ++ struct { ++ __be16 id; ++ } icmp; ++ struct { ++ __be16 port; ++ } dccp; ++ struct { ++ __be16 port; ++ } sctp; ++ struct { ++ __be16 key; ++ } gre; ++}; ++ ++struct nf_ct_dccp { ++ u_int8_t role[2]; ++ u_int8_t state; ++ u_int8_t last_pkt; ++ u_int8_t last_dir; ++ u_int64_t handshake_seq; ++}; ++ ++struct ip_ct_sctp { ++ enum sctp_conntrack state; ++ __be32 vtag[2]; ++ u8 last_dir; ++ u8 flags; ++}; ++ ++struct compat_sock_fprog { ++ u16 len; ++ compat_uptr_t filter; ++}; ++ ++struct udp6_sock { ++ struct udp_sock udp; ++ struct ipv6_pinfo inet6; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct tcp6_sock { ++ struct tcp_sock tcp; ++ struct ipv6_pinfo inet6; ++}; ++ ++struct strp_msg { ++ int full_len; ++ int offset; ++}; ++ ++struct _strp_msg { ++ struct strp_msg strp; ++ int accum_len; ++}; ++ ++struct tls_msg { ++ u8 control; ++}; ++ ++struct sk_skb_cb { ++ unsigned char data[20]; ++ unsigned char pad[4]; ++ struct _strp_msg strp; ++ struct tls_msg tls; ++ u64 temp_reg; ++}; ++ ++enum { ++ SEG6_LOCAL_ACTION_UNSPEC = 0, ++ SEG6_LOCAL_ACTION_END = 1, ++ SEG6_LOCAL_ACTION_END_X = 2, ++ SEG6_LOCAL_ACTION_END_T = 3, ++ SEG6_LOCAL_ACTION_END_DX2 = 4, ++ SEG6_LOCAL_ACTION_END_DX6 = 5, ++ SEG6_LOCAL_ACTION_END_DX4 = 6, ++ SEG6_LOCAL_ACTION_END_DT6 = 7, ++ SEG6_LOCAL_ACTION_END_DT4 = 8, ++ SEG6_LOCAL_ACTION_END_B6 = 9, ++ SEG6_LOCAL_ACTION_END_B6_ENCAP = 10, ++ SEG6_LOCAL_ACTION_END_BM = 11, ++ SEG6_LOCAL_ACTION_END_S = 12, ++ SEG6_LOCAL_ACTION_END_AS = 13, ++ SEG6_LOCAL_ACTION_END_AM = 14, ++ SEG6_LOCAL_ACTION_END_BPF = 15, ++ SEG6_LOCAL_ACTION_END_DT46 = 16, ++ __SEG6_LOCAL_ACTION_MAX = 17, ++}; ++ ++struct seg6_bpf_srh_state { ++ struct ipv6_sr_hdr *srh; ++ u16 hdrlen; ++ bool valid; ++}; ++ ++struct tls_strparser { ++ struct sock *sk; ++ u32 mark: 8; ++ u32 stopped: 1; ++ u32 copy_mode: 1; ++ u32 mixed_decrypted: 1; ++ u32 msg_ready: 1; ++ struct strp_msg stm; ++ struct sk_buff *anchor; ++ struct work_struct work; ++}; ++ ++struct tls_sw_context_rx { ++ struct crypto_aead *aead_recv; ++ struct crypto_wait async_wait; ++ struct sk_buff_head rx_list; ++ void (*saved_data_ready)(struct sock *); ++ u8 reader_present; ++ u8 async_capable: 1; ++ u8 zc_capable: 1; ++ u8 reader_contended: 1; ++ struct tls_strparser strp; ++ atomic_t decrypt_pending; ++ spinlock_t decrypt_compl_lock; ++ struct sk_buff_head async_hold; ++ struct wait_queue_head wq; ++}; ++ ++struct nf_ct_gre { ++ unsigned int stream_timeout; ++ unsigned int timeout; ++}; ++ ++struct nf_conntrack_zone { ++ u16 id; ++ u8 flags; ++ u8 dir; ++}; ++ ++struct nf_conntrack_man { ++ union nf_inet_addr u3; ++ union nf_conntrack_man_proto u; ++ u_int16_t l3num; ++}; ++ ++struct nf_conntrack_tuple { ++ struct nf_conntrack_man src; ++ struct { ++ union nf_inet_addr u3; ++ union { ++ __be16 all; ++ struct { ++ __be16 port; ++ } tcp; ++ struct { ++ __be16 port; ++ } udp; ++ struct { ++ u_int8_t type; ++ u_int8_t code; ++ } icmp; ++ struct { ++ __be16 port; ++ } dccp; ++ struct { ++ __be16 port; ++ } sctp; ++ struct { ++ __be16 key; ++ } gre; ++ } u; ++ u_int8_t protonum; ++ struct {} __nfct_hash_offsetend; ++ u_int8_t dir; ++ } dst; ++}; ++ ++struct nf_conntrack_tuple_hash { ++ struct hlist_nulls_node hnnode; ++ struct nf_conntrack_tuple tuple; ++}; ++ ++struct nf_ct_udp { ++ long unsigned int stream_ts; ++}; ++ ++union nf_conntrack_proto { ++ struct nf_ct_dccp dccp; ++ struct ip_ct_sctp sctp; ++ struct ip_ct_tcp tcp; ++ struct nf_ct_udp udp; ++ struct nf_ct_gre gre; ++ unsigned int tmpl_padto; ++}; ++ ++struct nf_ct_ext; ++ ++struct nf_conn { ++ struct nf_conntrack ct_general; ++ spinlock_t lock; ++ u32 timeout; ++ struct nf_conntrack_zone zone; ++ struct nf_conntrack_tuple_hash tuplehash[2]; ++ long unsigned int status; ++ possible_net_t ct_net; ++ struct hlist_node nat_bysource; ++ struct {} __nfct_init_offset; ++ struct nf_conn *master; ++ u_int32_t mark; ++ u_int32_t secmark; ++ struct nf_ct_ext *ext; ++ union nf_conntrack_proto proto; ++}; ++ ++struct nf_conn___init { ++ struct nf_conn ct; ++}; ++ ++typedef u64 (*btf_bpf_skb_get_pay_offset)(struct sk_buff *); ++ ++typedef u64 (*btf_bpf_skb_get_nlattr)(struct sk_buff *, u32, u32); ++ ++typedef u64 (*btf_bpf_skb_get_nlattr_nest)(struct sk_buff *, u32, u32); ++ ++typedef u64 (*btf_bpf_skb_load_helper_8)(const struct sk_buff *, const void *, int, int); ++ ++typedef u64 (*btf_bpf_skb_load_helper_8_no_cache)(const struct sk_buff *, int); ++ ++typedef u64 (*btf_bpf_skb_load_helper_16)(const struct sk_buff *, const void *, int, int); ++ ++typedef u64 (*btf_bpf_skb_load_helper_16_no_cache)(const struct sk_buff *, int); ++ ++typedef u64 (*btf_bpf_skb_load_helper_32)(const struct sk_buff *, const void *, int, int); ++ ++typedef u64 (*btf_bpf_skb_load_helper_32_no_cache)(const struct sk_buff *, int); ++ ++struct bpf_scratchpad { ++ union { ++ __be32 diff[128]; ++ u8 buff[512]; ++ }; ++}; ++ ++typedef u64 (*btf_bpf_skb_store_bytes)(struct sk_buff *, u32, const void *, u32, u64); ++ ++typedef u64 (*btf_bpf_skb_load_bytes)(const struct sk_buff *, u32, void *, u32); ++ ++typedef u64 (*btf_bpf_flow_dissector_load_bytes)(const struct bpf_flow_dissector *, u32, void *, u32); ++ ++typedef u64 (*btf_bpf_skb_load_bytes_relative)(const struct sk_buff *, u32, void *, u32, u32); ++ ++typedef u64 (*btf_bpf_skb_pull_data)(struct sk_buff *, u32); ++ ++typedef u64 (*btf_bpf_sk_fullsock)(struct sock *); ++ ++typedef u64 (*btf_sk_skb_pull_data)(struct sk_buff *, u32); ++ ++typedef u64 (*btf_bpf_l3_csum_replace)(struct sk_buff *, u32, u64, u64, u64); ++ ++typedef u64 (*btf_bpf_l4_csum_replace)(struct sk_buff *, u32, u64, u64, u64); ++ ++typedef u64 (*btf_bpf_csum_diff)(__be32 *, u32, __be32 *, u32, __wsum); ++ ++typedef u64 (*btf_bpf_csum_update)(struct sk_buff *, __wsum); ++ ++typedef u64 (*btf_bpf_csum_level)(struct sk_buff *, u64); ++ ++enum { ++ BPF_F_NEIGH = 2, ++ BPF_F_PEER = 4, ++ BPF_F_NEXTHOP = 8, ++}; ++ ++typedef u64 (*btf_bpf_clone_redirect)(struct sk_buff *, u32, u64); ++ ++typedef u64 (*btf_bpf_redirect)(u32, u64); ++ ++typedef u64 (*btf_bpf_redirect_peer)(u32, u64); ++ ++typedef u64 (*btf_bpf_redirect_neigh)(u32, struct bpf_redir_neigh *, int, u64); ++ ++typedef u64 (*btf_bpf_msg_apply_bytes)(struct sk_msg *, u32); ++ ++typedef u64 (*btf_bpf_msg_cork_bytes)(struct sk_msg *, u32); ++ ++typedef u64 (*btf_bpf_msg_pull_data)(struct sk_msg *, u32, u32, u64); ++ ++typedef u64 (*btf_bpf_msg_push_data)(struct sk_msg *, u32, u32, u64); ++ ++typedef u64 (*btf_bpf_msg_pop_data)(struct sk_msg *, u32, u32, u64); ++ ++typedef u64 (*btf_bpf_get_cgroup_classid_curr)(); ++ ++typedef u64 (*btf_bpf_skb_cgroup_classid)(const struct sk_buff *); ++ ++typedef u64 (*btf_bpf_get_cgroup_classid)(const struct sk_buff *); ++ ++typedef u64 (*btf_bpf_get_route_realm)(const struct sk_buff *); ++ ++typedef u64 (*btf_bpf_get_hash_recalc)(struct sk_buff *); ++ ++typedef u64 (*btf_bpf_set_hash_invalid)(struct sk_buff *); ++ ++typedef u64 (*btf_bpf_set_hash)(struct sk_buff *, u32); ++ ++typedef u64 (*btf_bpf_skb_vlan_push)(struct sk_buff *, __be16, u16); ++ ++typedef u64 (*btf_bpf_skb_vlan_pop)(struct sk_buff *); ++ ++typedef u64 (*btf_bpf_skb_change_proto)(struct sk_buff *, __be16, u64); ++ ++typedef u64 (*btf_bpf_skb_change_type)(struct sk_buff *, u32); ++ ++typedef u64 (*btf_sk_skb_adjust_room)(struct sk_buff *, s32, u32, u64); ++ ++typedef u64 (*btf_bpf_skb_adjust_room)(struct sk_buff *, s32, u32, u64); ++ ++typedef u64 (*btf_bpf_skb_change_tail)(struct sk_buff *, u32, u64); ++ ++typedef u64 (*btf_sk_skb_change_tail)(struct sk_buff *, u32, u64); ++ ++typedef u64 (*btf_bpf_skb_change_head)(struct sk_buff *, u32, u64); ++ ++typedef u64 (*btf_sk_skb_change_head)(struct sk_buff *, u32, u64); ++ ++typedef u64 (*btf_bpf_xdp_get_buff_len)(struct xdp_buff *); ++ ++typedef u64 (*btf_bpf_xdp_adjust_head)(struct xdp_buff *, int); ++ ++typedef u64 (*btf_bpf_xdp_load_bytes)(struct xdp_buff *, u32, void *, u32); ++ ++typedef u64 (*btf_bpf_xdp_store_bytes)(struct xdp_buff *, u32, void *, u32); ++ ++typedef u64 (*btf_bpf_xdp_adjust_tail)(struct xdp_buff *, int); ++ ++typedef u64 (*btf_bpf_xdp_adjust_meta)(struct xdp_buff *, int); ++ ++typedef u64 (*btf_bpf_xdp_redirect)(u32, u64); ++ ++typedef u64 (*btf_bpf_xdp_redirect_map)(struct bpf_map *, u64, u64); ++ ++typedef u64 (*btf_bpf_skb_event_output)(struct sk_buff *, struct bpf_map *, u64, void *, u64); ++ ++typedef u64 (*btf_bpf_skb_get_tunnel_key)(struct sk_buff *, struct bpf_tunnel_key *, u32, u64); ++ ++typedef u64 (*btf_bpf_skb_get_tunnel_opt)(struct sk_buff *, u8 *, u32); ++ ++typedef u64 (*btf_bpf_skb_set_tunnel_key)(struct sk_buff *, const struct bpf_tunnel_key *, u32, u64); ++ ++typedef u64 (*btf_bpf_skb_set_tunnel_opt)(struct sk_buff *, const u8 *, u32); ++ ++typedef u64 (*btf_bpf_skb_under_cgroup)(struct sk_buff *, struct bpf_map *, u32); ++ ++typedef u64 (*btf_bpf_skb_cgroup_id)(const struct sk_buff *); ++ ++typedef u64 (*btf_bpf_skb_ancestor_cgroup_id)(const struct sk_buff *, int); ++ ++typedef u64 (*btf_bpf_sk_cgroup_id)(struct sock *); ++ ++typedef u64 (*btf_bpf_sk_ancestor_cgroup_id)(struct sock *, int); ++ ++typedef u64 (*btf_bpf_xdp_event_output)(struct xdp_buff *, struct bpf_map *, u64, void *, u64); ++ ++typedef u64 (*btf_bpf_get_socket_cookie)(struct sk_buff *); ++ ++typedef u64 (*btf_bpf_get_socket_cookie_sock_addr)(struct bpf_sock_addr_kern *); ++ ++typedef u64 (*btf_bpf_get_socket_cookie_sock)(struct sock *); ++ ++typedef u64 (*btf_bpf_get_socket_ptr_cookie)(struct sock *); ++ ++typedef u64 (*btf_bpf_get_socket_cookie_sock_ops)(struct bpf_sock_ops_kern *); ++ ++typedef u64 (*btf_bpf_get_netns_cookie_sock)(struct sock *); ++ ++typedef u64 (*btf_bpf_get_netns_cookie_sock_addr)(struct bpf_sock_addr_kern *); ++ ++typedef u64 (*btf_bpf_get_netns_cookie_sock_ops)(struct bpf_sock_ops_kern *); ++ ++typedef u64 (*btf_bpf_get_netns_cookie_sk_msg)(struct sk_msg *); ++ ++typedef u64 (*btf_bpf_get_socket_uid)(struct sk_buff *); ++ ++typedef u64 (*btf_bpf_sk_setsockopt)(struct sock *, int, int, char *, int); ++ ++typedef u64 (*btf_bpf_sk_getsockopt)(struct sock *, int, int, char *, int); ++ ++typedef u64 (*btf_bpf_unlocked_sk_setsockopt)(struct sock *, int, int, char *, int); ++ ++typedef u64 (*btf_bpf_unlocked_sk_getsockopt)(struct sock *, int, int, char *, int); ++ ++typedef u64 (*btf_bpf_sock_addr_setsockopt)(struct bpf_sock_addr_kern *, int, int, char *, int); ++ ++typedef u64 (*btf_bpf_sock_addr_getsockopt)(struct bpf_sock_addr_kern *, int, int, char *, int); ++ ++typedef u64 (*btf_bpf_sock_ops_setsockopt)(struct bpf_sock_ops_kern *, int, int, char *, int); ++ ++typedef u64 (*btf_bpf_sock_ops_getsockopt)(struct bpf_sock_ops_kern *, int, int, char *, int); ++ ++typedef u64 (*btf_bpf_sock_ops_cb_flags_set)(struct bpf_sock_ops_kern *, int); ++ ++typedef u64 (*btf_bpf_bind)(struct bpf_sock_addr_kern *, struct sockaddr *, int); ++ ++typedef u64 (*btf_bpf_skb_get_xfrm_state)(struct sk_buff *, u32, struct bpf_xfrm_state *, u32, u64); ++ ++typedef u64 (*btf_bpf_xdp_fib_lookup)(struct xdp_buff *, struct bpf_fib_lookup *, int, u32); ++ ++typedef u64 (*btf_bpf_skb_fib_lookup)(struct sk_buff *, struct bpf_fib_lookup *, int, u32); ++ ++typedef u64 (*btf_bpf_skb_check_mtu)(struct sk_buff *, u32, u32 *, s32, u64); ++ ++typedef u64 (*btf_bpf_xdp_check_mtu)(struct xdp_buff *, u32, u32 *, s32, u64); ++ ++typedef u64 (*btf_bpf_lwt_in_push_encap)(struct sk_buff *, u32, void *, u32); ++ ++typedef u64 (*btf_bpf_lwt_xmit_push_encap)(struct sk_buff *, u32, void *, u32); ++ ++typedef u64 (*btf_bpf_lwt_seg6_store_bytes)(struct sk_buff *, u32, const void *, u32); ++ ++typedef u64 (*btf_bpf_lwt_seg6_action)(struct sk_buff *, u32, void *, u32); ++ ++typedef u64 (*btf_bpf_lwt_seg6_adjust_srh)(struct sk_buff *, u32, s32); ++ ++typedef u64 (*btf_bpf_skc_lookup_tcp)(struct sk_buff *, struct bpf_sock_tuple *, u32, u64, u64); ++ ++typedef u64 (*btf_bpf_sk_lookup_tcp)(struct sk_buff *, struct bpf_sock_tuple *, u32, u64, u64); ++ ++typedef u64 (*btf_bpf_sk_lookup_udp)(struct sk_buff *, struct bpf_sock_tuple *, u32, u64, u64); ++ ++typedef u64 (*btf_bpf_tc_skc_lookup_tcp)(struct sk_buff *, struct bpf_sock_tuple *, u32, u64, u64); ++ ++typedef u64 (*btf_bpf_tc_sk_lookup_tcp)(struct sk_buff *, struct bpf_sock_tuple *, u32, u64, u64); ++ ++typedef u64 (*btf_bpf_tc_sk_lookup_udp)(struct sk_buff *, struct bpf_sock_tuple *, u32, u64, u64); ++ ++typedef u64 (*btf_bpf_sk_release)(struct sock *); ++ ++typedef u64 (*btf_bpf_xdp_sk_lookup_udp)(struct xdp_buff *, struct bpf_sock_tuple *, u32, u32, u64); ++ ++typedef u64 (*btf_bpf_xdp_skc_lookup_tcp)(struct xdp_buff *, struct bpf_sock_tuple *, u32, u32, u64); ++ ++typedef u64 (*btf_bpf_xdp_sk_lookup_tcp)(struct xdp_buff *, struct bpf_sock_tuple *, u32, u32, u64); ++ ++typedef u64 (*btf_bpf_sock_addr_skc_lookup_tcp)(struct bpf_sock_addr_kern *, struct bpf_sock_tuple *, u32, u64, u64); ++ ++typedef u64 (*btf_bpf_sock_addr_sk_lookup_tcp)(struct bpf_sock_addr_kern *, struct bpf_sock_tuple *, u32, u64, u64); ++ ++typedef u64 (*btf_bpf_sock_addr_sk_lookup_udp)(struct bpf_sock_addr_kern *, struct bpf_sock_tuple *, u32, u64, u64); ++ ++typedef u64 (*btf_bpf_tcp_sock)(struct sock *); ++ ++typedef u64 (*btf_bpf_get_listener_sock)(struct sock *); ++ ++typedef u64 (*btf_bpf_skb_ecn_set_ce)(struct sk_buff *); ++ ++typedef u64 (*btf_bpf_tcp_check_syncookie)(struct sock *, void *, u32, struct tcphdr *, u32); ++ ++typedef u64 (*btf_bpf_tcp_gen_syncookie)(struct sock *, void *, u32, struct tcphdr *, u32); ++ ++typedef u64 (*btf_bpf_sk_assign)(struct sk_buff *, struct sock *, u64); ++ ++typedef u64 (*btf_bpf_sock_ops_load_hdr_opt)(struct bpf_sock_ops_kern *, void *, u32, u64); ++ ++typedef u64 (*btf_bpf_sock_ops_store_hdr_opt)(struct bpf_sock_ops_kern *, const void *, u32, u64); ++ ++typedef u64 (*btf_bpf_sock_ops_reserve_hdr_opt)(struct bpf_sock_ops_kern *, u32, u64); ++ ++typedef u64 (*btf_bpf_skb_set_tstamp)(struct sk_buff *, u64, u32); ++ ++typedef u64 (*btf_bpf_tcp_raw_gen_syncookie_ipv4)(struct iphdr *, struct tcphdr *, u32); ++ ++typedef u64 (*btf_bpf_tcp_raw_gen_syncookie_ipv6)(struct ipv6hdr *, struct tcphdr *, u32); ++ ++typedef u64 (*btf_bpf_tcp_raw_check_syncookie_ipv4)(struct iphdr *, struct tcphdr *); ++ ++typedef u64 (*btf_bpf_tcp_raw_check_syncookie_ipv6)(struct ipv6hdr *, struct tcphdr *); ++ ++typedef u64 (*btf_sk_select_reuseport)(struct sk_reuseport_kern *, struct bpf_map *, void *, u32); ++ ++typedef u64 (*btf_sk_reuseport_load_bytes)(const struct sk_reuseport_kern *, u32, void *, u32); ++ ++typedef u64 (*btf_sk_reuseport_load_bytes_relative)(const struct sk_reuseport_kern *, u32, void *, u32, u32); ++ ++typedef u64 (*btf_bpf_sk_lookup_assign)(struct bpf_sk_lookup_kern *, struct sock *, u64); ++ ++typedef u64 (*btf_bpf_skc_to_tcp6_sock)(struct sock *); ++ ++typedef u64 (*btf_bpf_skc_to_tcp_sock)(struct sock *); ++ ++typedef u64 (*btf_bpf_skc_to_tcp_timewait_sock)(struct sock *); ++ ++typedef u64 (*btf_bpf_skc_to_tcp_request_sock)(struct sock *); ++ ++typedef u64 (*btf_bpf_skc_to_udp6_sock)(struct sock *); ++ ++typedef u64 (*btf_bpf_skc_to_unix_sock)(struct sock *); ++ ++typedef u64 (*btf_bpf_skc_to_mptcp_sock)(struct sock *); ++ ++typedef u64 (*btf_bpf_sock_from_file)(struct file *); ++ ++struct flow_dissector_mpls_lse { ++ u32 mpls_ttl: 8; ++ u32 mpls_bos: 1; ++ u32 mpls_tc: 3; ++ u32 mpls_label: 20; ++}; ++ ++struct flow_dissector_key_mpls { ++ struct flow_dissector_mpls_lse ls[7]; ++ u8 used_lses; ++}; ++ ++struct flow_dissector_key_enc_opts { ++ u8 data[255]; ++ u8 len; ++ __be16 dst_opt_type; ++}; ++ ++struct flow_dissector_key_arp { ++ __u32 sip; ++ __u32 tip; ++ __u8 op; ++ unsigned char sha[6]; ++ unsigned char tha[6]; ++}; ++ ++struct flow_dissector_key_ports_range { ++ union { ++ struct flow_dissector_key_ports tp; ++ struct { ++ struct flow_dissector_key_ports tp_min; ++ struct flow_dissector_key_ports tp_max; ++ }; ++ }; ++}; ++ ++struct flow_dissector_key_eth_addrs { ++ unsigned char dst[6]; ++ unsigned char src[6]; ++}; ++ ++struct flow_dissector_key_tcp { ++ __be16 flags; ++}; ++ ++struct flow_dissector_key_ip { ++ __u8 tos; ++ __u8 ttl; ++}; ++ ++struct flow_dissector_key_meta { ++ int ingress_ifindex; ++ u16 ingress_iftype; ++}; ++ ++struct flow_dissector_key_ct { ++ u16 ct_state; ++ u16 ct_zone; ++ u32 ct_mark; ++ u32 ct_labels[4]; ++}; ++ ++struct flow_dissector_key_pppoe { ++ __be16 session_id; ++ __be16 ppp_proto; ++ __be16 type; ++}; ++ ++struct flow_dissector_key_l2tpv3 { ++ __be32 session_id; ++}; ++ ++struct flow_match_meta { ++ struct flow_dissector_key_meta *key; ++ struct flow_dissector_key_meta *mask; ++}; ++ ++struct flow_match_basic { ++ struct flow_dissector_key_basic *key; ++ struct flow_dissector_key_basic *mask; ++}; ++ ++struct flow_match_control { ++ struct flow_dissector_key_control *key; ++ struct flow_dissector_key_control *mask; ++}; ++ ++struct flow_match_eth_addrs { ++ struct flow_dissector_key_eth_addrs *key; ++ struct flow_dissector_key_eth_addrs *mask; ++}; ++ ++struct flow_match_vlan { ++ struct flow_dissector_key_vlan *key; ++ struct flow_dissector_key_vlan *mask; ++}; ++ ++struct flow_match_arp { ++ struct flow_dissector_key_arp *key; ++ struct flow_dissector_key_arp *mask; ++}; ++ ++struct flow_match_ipv4_addrs { ++ struct flow_dissector_key_ipv4_addrs *key; ++ struct flow_dissector_key_ipv4_addrs *mask; ++}; ++ ++struct flow_match_ipv6_addrs { ++ struct flow_dissector_key_ipv6_addrs *key; ++ struct flow_dissector_key_ipv6_addrs *mask; ++}; ++ ++struct flow_match_ip { ++ struct flow_dissector_key_ip *key; ++ struct flow_dissector_key_ip *mask; ++}; ++ ++struct flow_match_ports { ++ struct flow_dissector_key_ports *key; ++ struct flow_dissector_key_ports *mask; ++}; ++ ++struct flow_match_ports_range { ++ struct flow_dissector_key_ports_range *key; ++ struct flow_dissector_key_ports_range *mask; ++}; ++ ++struct flow_match_icmp { ++ struct flow_dissector_key_icmp *key; ++ struct flow_dissector_key_icmp *mask; ++}; ++ ++struct flow_match_tcp { ++ struct flow_dissector_key_tcp *key; ++ struct flow_dissector_key_tcp *mask; ++}; ++ ++struct flow_match_mpls { ++ struct flow_dissector_key_mpls *key; ++ struct flow_dissector_key_mpls *mask; ++}; ++ ++struct flow_match_enc_keyid { ++ struct flow_dissector_key_keyid *key; ++ struct flow_dissector_key_keyid *mask; ++}; ++ ++struct flow_match_enc_opts { ++ struct flow_dissector_key_enc_opts *key; ++ struct flow_dissector_key_enc_opts *mask; ++}; ++ ++struct flow_match_ct { ++ struct flow_dissector_key_ct *key; ++ struct flow_dissector_key_ct *mask; ++}; ++ ++struct flow_match_pppoe { ++ struct flow_dissector_key_pppoe *key; ++ struct flow_dissector_key_pppoe *mask; ++}; ++ ++struct flow_match_l2tpv3 { ++ struct flow_dissector_key_l2tpv3 *key; ++ struct flow_dissector_key_l2tpv3 *mask; ++}; ++ ++enum offload_act_command { ++ FLOW_ACT_REPLACE = 0, ++ FLOW_ACT_DESTROY = 1, ++ FLOW_ACT_STATS = 2, ++}; ++ ++struct flow_offload_action { ++ struct netlink_ext_ack *extack; ++ enum offload_act_command command; ++ enum flow_action_id id; ++ u32 index; ++ long unsigned int cookie; ++ struct flow_stats stats; ++ struct flow_action action; ++}; ++ ++typedef int flow_indr_block_bind_cb_t(struct net_device *, struct Qdisc *, void *, enum tc_setup_type, void *, void *, void (*)(struct flow_block_cb *)); ++ ++struct flow_indr_dev { ++ struct list_head list; ++ flow_indr_block_bind_cb_t *cb; ++ void *cb_priv; ++ refcount_t refcnt; ++}; ++ ++struct flow_indir_dev_info { ++ void *data; ++ struct net_device *dev; ++ struct Qdisc *sch; ++ enum tc_setup_type type; ++ void (*cleanup)(struct flow_block_cb *); ++ struct list_head list; ++ enum flow_block_command command; ++ enum flow_block_binder_type binder_type; ++ struct list_head *cb_list; ++}; ++ ++struct rx_queue_attribute { ++ struct attribute attr; ++ ssize_t (*show)(struct netdev_rx_queue *, char *); ++ ssize_t (*store)(struct netdev_rx_queue *, const char *, size_t); ++}; ++ ++struct netdev_queue_attribute { ++ struct attribute attr; ++ ssize_t (*show)(struct netdev_queue *, char *); ++ ssize_t (*store)(struct netdev_queue *, const char *, size_t); ++}; ++ ++struct dm_hw_stat_delta { ++ long unsigned int last_rx; ++ long unsigned int last_drop_val; ++ struct callback_head rcu; ++}; ++ ++struct net_dm_drop_point { ++ __u8 pc[8]; ++ __u32 count; ++}; ++ ++struct net_dm_alert_msg { ++ __u32 entries; ++ struct net_dm_drop_point points[0]; ++}; ++ ++enum { ++ NET_DM_CMD_UNSPEC = 0, ++ NET_DM_CMD_ALERT = 1, ++ NET_DM_CMD_CONFIG = 2, ++ NET_DM_CMD_START = 3, ++ NET_DM_CMD_STOP = 4, ++ NET_DM_CMD_PACKET_ALERT = 5, ++ NET_DM_CMD_CONFIG_GET = 6, ++ NET_DM_CMD_CONFIG_NEW = 7, ++ NET_DM_CMD_STATS_GET = 8, ++ NET_DM_CMD_STATS_NEW = 9, ++ _NET_DM_CMD_MAX = 10, ++}; ++ ++enum net_dm_attr { ++ NET_DM_ATTR_UNSPEC = 0, ++ NET_DM_ATTR_ALERT_MODE = 1, ++ NET_DM_ATTR_PC = 2, ++ NET_DM_ATTR_SYMBOL = 3, ++ NET_DM_ATTR_IN_PORT = 4, ++ NET_DM_ATTR_TIMESTAMP = 5, ++ NET_DM_ATTR_PROTO = 6, ++ NET_DM_ATTR_PAYLOAD = 7, ++ NET_DM_ATTR_PAD = 8, ++ NET_DM_ATTR_TRUNC_LEN = 9, ++ NET_DM_ATTR_ORIG_LEN = 10, ++ NET_DM_ATTR_QUEUE_LEN = 11, ++ NET_DM_ATTR_STATS = 12, ++ NET_DM_ATTR_HW_STATS = 13, ++ NET_DM_ATTR_ORIGIN = 14, ++ NET_DM_ATTR_HW_TRAP_GROUP_NAME = 15, ++ NET_DM_ATTR_HW_TRAP_NAME = 16, ++ NET_DM_ATTR_HW_ENTRIES = 17, ++ NET_DM_ATTR_HW_ENTRY = 18, ++ NET_DM_ATTR_HW_TRAP_COUNT = 19, ++ NET_DM_ATTR_SW_DROPS = 20, ++ NET_DM_ATTR_HW_DROPS = 21, ++ NET_DM_ATTR_FLOW_ACTION_COOKIE = 22, ++ NET_DM_ATTR_REASON = 23, ++ __NET_DM_ATTR_MAX = 24, ++ NET_DM_ATTR_MAX = 23, ++}; ++ ++enum net_dm_alert_mode { ++ NET_DM_ALERT_MODE_SUMMARY = 0, ++ NET_DM_ALERT_MODE_PACKET = 1, ++}; ++ ++enum { ++ NET_DM_ATTR_PORT_NETDEV_IFINDEX = 0, ++ NET_DM_ATTR_PORT_NETDEV_NAME = 1, ++ __NET_DM_ATTR_PORT_MAX = 2, ++ NET_DM_ATTR_PORT_MAX = 1, ++}; ++ ++enum { ++ NET_DM_ATTR_STATS_DROPPED = 0, ++ __NET_DM_ATTR_STATS_MAX = 1, ++ NET_DM_ATTR_STATS_MAX = 0, ++}; ++ ++enum net_dm_origin { ++ NET_DM_ORIGIN_SW = 0, ++ NET_DM_ORIGIN_HW = 1, ++}; ++ ++struct devlink_trap_metadata { ++ const char *trap_name; ++ const char *trap_group_name; ++ struct net_device *input_dev; ++ netdevice_tracker dev_tracker; ++ const struct flow_action_cookie *fa_cookie; ++ enum devlink_trap_type trap_type; ++}; ++ ++struct net_dm_stats { ++ u64_stats_t dropped; ++ struct u64_stats_sync syncp; ++}; ++ ++struct net_dm_hw_entry { ++ char trap_name[40]; ++ u32 count; ++}; ++ ++struct net_dm_hw_entries { ++ u32 num_entries; ++ struct net_dm_hw_entry entries[0]; ++}; ++ ++struct per_cpu_dm_data { ++ spinlock_t lock; ++ union { ++ struct sk_buff *skb; ++ struct net_dm_hw_entries *hw_entries; ++ }; ++ struct sk_buff_head drop_queue; ++ struct work_struct dm_alert_work; ++ struct timer_list send_timer; ++ struct net_dm_stats stats; ++}; ++ ++struct net_dm_alert_ops { ++ void (*kfree_skb_probe)(void *, struct sk_buff *, void *, enum skb_drop_reason); ++ void (*napi_poll_probe)(void *, struct napi_struct *, int, int); ++ void (*work_item_func)(struct work_struct *); ++ void (*hw_work_item_func)(struct work_struct *); ++ void (*hw_trap_probe)(void *, const struct devlink *, struct sk_buff *, const struct devlink_trap_metadata *); ++}; ++ ++struct net_dm_skb_cb { ++ union { ++ struct devlink_trap_metadata *hw_metadata; ++ void *pc; ++ }; ++ enum skb_drop_reason reason; ++}; ++ ++struct update_classid_context { ++ u32 classid; ++ unsigned int batch; ++}; ++ ++struct dst_cache_pcpu { ++ long unsigned int refresh_ts; ++ struct dst_entry *dst; ++ u32 cookie; ++ union { ++ struct in_addr in_saddr; ++ struct in6_addr in6_saddr; ++ }; ++}; ++ ++struct compat_cmsghdr { ++ compat_size_t cmsg_len; ++ compat_int_t cmsg_level; ++ compat_int_t cmsg_type; ++}; ++ ++struct sch_frag_data { ++ long unsigned int dst; ++ struct qdisc_skb_cb cb; ++ __be16 inner_protocol; ++ u16 vlan_tci; ++ __be16 vlan_proto; ++ unsigned int l2_len; ++ u8 l2_data[18]; ++ int (*xmit)(struct sk_buff *); ++}; ++ ++struct tcamsg { ++ unsigned char tca_family; ++ unsigned char tca__pad1; ++ short unsigned int tca__pad2; ++}; ++ ++enum { ++ TCA_ROOT_UNSPEC = 0, ++ TCA_ROOT_TAB = 1, ++ TCA_ROOT_FLAGS = 2, ++ TCA_ROOT_COUNT = 3, ++ TCA_ROOT_TIME_DELTA = 4, ++ TCA_ROOT_EXT_WARN_MSG = 5, ++ __TCA_ROOT_MAX = 6, ++}; ++ ++struct tc_action_net { ++ struct tcf_idrinfo *idrinfo; ++ const struct tc_action_ops *ops; ++}; ++ ++struct tc_act_pernet_id { ++ struct list_head list; ++ unsigned int id; ++}; ++ ++enum { ++ TCA_FQ_CODEL_UNSPEC = 0, ++ TCA_FQ_CODEL_TARGET = 1, ++ TCA_FQ_CODEL_LIMIT = 2, ++ TCA_FQ_CODEL_INTERVAL = 3, ++ TCA_FQ_CODEL_ECN = 4, ++ TCA_FQ_CODEL_FLOWS = 5, ++ TCA_FQ_CODEL_QUANTUM = 6, ++ TCA_FQ_CODEL_CE_THRESHOLD = 7, ++ TCA_FQ_CODEL_DROP_BATCH_SIZE = 8, ++ TCA_FQ_CODEL_MEMORY_LIMIT = 9, ++ TCA_FQ_CODEL_CE_THRESHOLD_SELECTOR = 10, ++ TCA_FQ_CODEL_CE_THRESHOLD_MASK = 11, ++ __TCA_FQ_CODEL_MAX = 12, ++}; ++ ++enum { ++ TCA_FQ_CODEL_XSTATS_QDISC = 0, ++ TCA_FQ_CODEL_XSTATS_CLASS = 1, ++}; ++ ++struct tc_fq_codel_qd_stats { ++ __u32 maxpacket; ++ __u32 drop_overlimit; ++ __u32 ecn_mark; ++ __u32 new_flow_count; ++ __u32 new_flows_len; ++ __u32 old_flows_len; ++ __u32 ce_mark; ++ __u32 memory_usage; ++ __u32 drop_overmemory; ++}; ++ ++struct tc_fq_codel_cl_stats { ++ __s32 deficit; ++ __u32 ldelay; ++ __u32 count; ++ __u32 lastcount; ++ __u32 dropping; ++ __s32 drop_next; ++}; ++ ++struct tc_fq_codel_xstats { ++ __u32 type; ++ union { ++ struct tc_fq_codel_qd_stats qdisc_stats; ++ struct tc_fq_codel_cl_stats class_stats; ++ }; ++}; ++ ++typedef u32 codel_time_t; ++ ++typedef s32 codel_tdiff_t; ++ ++struct codel_params { ++ codel_time_t target; ++ codel_time_t ce_threshold; ++ codel_time_t interval; ++ u32 mtu; ++ bool ecn; ++ u8 ce_threshold_selector; ++ u8 ce_threshold_mask; ++}; ++ ++struct codel_vars { ++ u32 count; ++ u32 lastcount; ++ bool dropping; ++ u16 rec_inv_sqrt; ++ codel_time_t first_above_time; ++ codel_time_t drop_next; ++ codel_time_t ldelay; ++}; ++ ++struct codel_stats { ++ u32 maxpacket; ++ u32 drop_count; ++ u32 drop_len; ++ u32 ecn_mark; ++ u32 ce_mark; ++}; ++ ++typedef u32 (*codel_skb_len_t)(const struct sk_buff *); ++ ++typedef codel_time_t (*codel_skb_time_t)(const struct sk_buff *); ++ ++typedef void (*codel_skb_drop_t)(struct sk_buff *, void *); ++ ++typedef struct sk_buff * (*codel_skb_dequeue_t)(struct codel_vars *, void *); ++ ++struct codel_skb_cb { ++ codel_time_t enqueue_time; ++ unsigned int mem_usage; ++}; ++ ++struct fq_codel_flow { ++ struct sk_buff *head; ++ struct sk_buff *tail; ++ struct list_head flowchain; ++ int deficit; ++ struct codel_vars cvars; ++}; ++ ++struct fq_codel_sched_data { ++ struct tcf_proto *filter_list; ++ struct tcf_block *block; ++ struct fq_codel_flow *flows; ++ u32 *backlogs; ++ u32 flows_cnt; ++ u32 quantum; ++ u32 drop_batch_size; ++ u32 memory_limit; ++ struct codel_params cparams; ++ struct codel_stats cstats; ++ u32 memory_usage; ++ u32 drop_overmemory; ++ u32 drop_overlimit; ++ u32 new_flow_count; ++ struct list_head new_flows; ++ struct list_head old_flows; ++}; ++ ++struct sockaddr_nl { ++ __kernel_sa_family_t nl_family; ++ short unsigned int nl_pad; ++ __u32 nl_pid; ++ __u32 nl_groups; ++}; ++ ++enum nlmsgerr_attrs { ++ NLMSGERR_ATTR_UNUSED = 0, ++ NLMSGERR_ATTR_MSG = 1, ++ NLMSGERR_ATTR_OFFS = 2, ++ NLMSGERR_ATTR_COOKIE = 3, ++ NLMSGERR_ATTR_POLICY = 4, ++ NLMSGERR_ATTR_MISS_TYPE = 5, ++ NLMSGERR_ATTR_MISS_NEST = 6, ++ __NLMSGERR_ATTR_MAX = 7, ++ NLMSGERR_ATTR_MAX = 6, ++}; ++ ++struct nl_pktinfo { ++ __u32 group; ++}; ++ ++enum { ++ NETLINK_UNCONNECTED = 0, ++ NETLINK_CONNECTED = 1, ++}; ++ ++enum netlink_skb_flags { ++ NETLINK_SKB_DST = 8, ++}; ++ ++struct netlink_notify { ++ struct net *net; ++ u32 portid; ++ int protocol; ++}; ++ ++struct netlink_tap { ++ struct net_device *dev; ++ struct module *module; ++ struct list_head list; ++}; ++ ++struct trace_event_raw_netlink_extack { ++ struct trace_entry ent; ++ u32 __data_loc_msg; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_netlink_extack { ++ u32 msg; ++}; ++ ++typedef void (*btf_trace_netlink_extack)(void *, const char *); ++ ++struct netlink_sock { ++ struct sock sk; ++ u32 portid; ++ u32 dst_portid; ++ u32 dst_group; ++ u32 flags; ++ u32 subscriptions; ++ u32 ngroups; ++ long unsigned int *groups; ++ long unsigned int state; ++ size_t max_recvmsg_len; ++ wait_queue_head_t wait; ++ bool bound; ++ bool cb_running; ++ int dump_done_errno; ++ struct netlink_callback cb; ++ struct mutex *cb_mutex; ++ struct mutex cb_def_mutex; ++ void (*netlink_rcv)(struct sk_buff *); ++ int (*netlink_bind)(struct net *, int); ++ void (*netlink_unbind)(struct net *, int); ++ struct module *module; ++ struct rhash_head node; ++ struct callback_head rcu; ++ struct work_struct work; ++}; ++ ++struct listeners; ++ ++struct netlink_table { ++ struct rhashtable hash; ++ struct hlist_head mc_list; ++ struct listeners *listeners; ++ unsigned int flags; ++ unsigned int groups; ++ struct mutex *cb_mutex; ++ struct module *module; ++ int (*bind)(struct net *, int); ++ void (*unbind)(struct net *, int); ++ int registered; ++}; ++ ++struct listeners { ++ struct callback_head rcu; ++ long unsigned int masks[0]; ++}; ++ ++struct netlink_tap_net { ++ struct list_head netlink_tap_all; ++ struct mutex netlink_tap_lock; ++}; ++ ++struct netlink_compare_arg { ++ possible_net_t pnet; ++ u32 portid; ++}; ++ ++struct netlink_broadcast_data { ++ struct sock *exclude_sk; ++ struct net *net; ++ u32 portid; ++ u32 group; ++ int failure; ++ int delivery_failure; ++ int congested; ++ int delivered; ++ gfp_t allocation; ++ struct sk_buff *skb; ++ struct sk_buff *skb2; ++}; ++ ++struct netlink_set_err_data { ++ struct sock *exclude_sk; ++ u32 portid; ++ u32 group; ++ int code; ++}; ++ ++struct nl_seq_iter { ++ struct seq_net_private p; ++ struct rhashtable_iter hti; ++ int link; ++}; ++ ++struct bpf_iter__netlink { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct netlink_sock *sk; ++ }; ++}; ++ ++struct ethtool_cmd { ++ __u32 cmd; ++ __u32 supported; ++ __u32 advertising; ++ __u16 speed; ++ __u8 duplex; ++ __u8 port; ++ __u8 phy_address; ++ __u8 transceiver; ++ __u8 autoneg; ++ __u8 mdio_support; ++ __u32 maxtxpkt; ++ __u32 maxrxpkt; ++ __u16 speed_hi; ++ __u8 eth_tp_mdix; ++ __u8 eth_tp_mdix_ctrl; ++ __u32 lp_advertising; ++ __u32 reserved[2]; ++}; ++ ++struct ethtool_value { ++ __u32 cmd; ++ __u32 data; ++}; ++ ++enum tunable_type_id { ++ ETHTOOL_TUNABLE_UNSPEC = 0, ++ ETHTOOL_TUNABLE_U8 = 1, ++ ETHTOOL_TUNABLE_U16 = 2, ++ ETHTOOL_TUNABLE_U32 = 3, ++ ETHTOOL_TUNABLE_U64 = 4, ++ ETHTOOL_TUNABLE_STRING = 5, ++ ETHTOOL_TUNABLE_S8 = 6, ++ ETHTOOL_TUNABLE_S16 = 7, ++ ETHTOOL_TUNABLE_S32 = 8, ++ ETHTOOL_TUNABLE_S64 = 9, ++}; ++ ++enum ethtool_stringset { ++ ETH_SS_TEST = 0, ++ ETH_SS_STATS = 1, ++ ETH_SS_PRIV_FLAGS = 2, ++ ETH_SS_NTUPLE_FILTERS = 3, ++ ETH_SS_FEATURES = 4, ++ ETH_SS_RSS_HASH_FUNCS = 5, ++ ETH_SS_TUNABLES = 6, ++ ETH_SS_PHY_STATS = 7, ++ ETH_SS_PHY_TUNABLES = 8, ++ ETH_SS_LINK_MODES = 9, ++ ETH_SS_MSG_CLASSES = 10, ++ ETH_SS_WOL_MODES = 11, ++ ETH_SS_SOF_TIMESTAMPING = 12, ++ ETH_SS_TS_TX_TYPES = 13, ++ ETH_SS_TS_RX_FILTERS = 14, ++ ETH_SS_UDP_TUNNEL_TYPES = 15, ++ ETH_SS_STATS_STD = 16, ++ ETH_SS_STATS_ETH_PHY = 17, ++ ETH_SS_STATS_ETH_MAC = 18, ++ ETH_SS_STATS_ETH_CTRL = 19, ++ ETH_SS_STATS_RMON = 20, ++ ETH_SS_COUNT = 21, ++}; ++ ++struct ethtool_gstrings { ++ __u32 cmd; ++ __u32 string_set; ++ __u32 len; ++ __u8 data[0]; ++}; ++ ++struct ethtool_sset_info { ++ __u32 cmd; ++ __u32 reserved; ++ __u64 sset_mask; ++ __u32 data[0]; ++}; ++ ++struct ethtool_perm_addr { ++ __u32 cmd; ++ __u32 size; ++ __u8 data[0]; ++}; ++ ++enum ethtool_flags { ++ ETH_FLAG_TXVLAN = 128, ++ ETH_FLAG_RXVLAN = 256, ++ ETH_FLAG_LRO = 32768, ++ ETH_FLAG_NTUPLE = 134217728, ++ ETH_FLAG_RXHASH = 268435456, ++}; ++ ++struct ethtool_rxfh { ++ __u32 cmd; ++ __u32 rss_context; ++ __u32 indir_size; ++ __u32 key_size; ++ __u8 hfunc; ++ __u8 rsvd8[3]; ++ __u32 rsvd32; ++ __u32 rss_config[0]; ++}; ++ ++struct ethtool_get_features_block { ++ __u32 available; ++ __u32 requested; ++ __u32 active; ++ __u32 never_changed; ++}; ++ ++struct ethtool_gfeatures { ++ __u32 cmd; ++ __u32 size; ++ struct ethtool_get_features_block features[0]; ++}; ++ ++struct ethtool_set_features_block { ++ __u32 valid; ++ __u32 requested; ++}; ++ ++struct ethtool_sfeatures { ++ __u32 cmd; ++ __u32 size; ++ struct ethtool_set_features_block features[0]; ++}; ++ ++enum ethtool_sfeatures_retval_bits { ++ ETHTOOL_F_UNSUPPORTED__BIT = 0, ++ ETHTOOL_F_WISH__BIT = 1, ++ ETHTOOL_F_COMPAT__BIT = 2, ++}; ++ ++struct ethtool_per_queue_op { ++ __u32 cmd; ++ __u32 sub_command; ++ __u32 queue_mask[128]; ++ char data[0]; ++}; ++ ++enum ethtool_fec_config_bits { ++ ETHTOOL_FEC_NONE_BIT = 0, ++ ETHTOOL_FEC_AUTO_BIT = 1, ++ ETHTOOL_FEC_OFF_BIT = 2, ++ ETHTOOL_FEC_RS_BIT = 3, ++ ETHTOOL_FEC_BASER_BIT = 4, ++ ETHTOOL_FEC_LLRS_BIT = 5, ++}; ++ ++struct ethtool_rx_flow_rule { ++ struct flow_rule *rule; ++ long unsigned int priv[0]; ++}; ++ ++struct ethtool_rx_flow_spec_input { ++ const struct ethtool_rx_flow_spec *fs; ++ u32 rss_ctx; ++}; ++ ++struct ethtool_devlink_compat { ++ struct devlink *devlink; ++ union { ++ struct ethtool_flash efl; ++ struct ethtool_drvinfo info; ++ }; ++}; ++ ++struct ethtool_link_usettings { ++ struct ethtool_link_settings base; ++ struct { ++ __u32 supported[4]; ++ __u32 advertising[4]; ++ __u32 lp_advertising[4]; ++ } link_modes; ++}; ++ ++struct ethtool_rx_flow_key { ++ struct flow_dissector_key_basic basic; ++ union { ++ struct flow_dissector_key_ipv4_addrs ipv4; ++ struct flow_dissector_key_ipv6_addrs ipv6; ++ }; ++ struct flow_dissector_key_ports tp; ++ struct flow_dissector_key_ip ip; ++ struct flow_dissector_key_vlan vlan; ++ struct flow_dissector_key_eth_addrs eth_addrs; ++}; ++ ++struct ethtool_rx_flow_match { ++ struct flow_dissector dissector; ++ long: 0; ++ struct ethtool_rx_flow_key key; ++ struct ethtool_rx_flow_key mask; ++}; ++ ++enum { ++ ETHTOOL_A_BITSET_BIT_UNSPEC = 0, ++ ETHTOOL_A_BITSET_BIT_INDEX = 1, ++ ETHTOOL_A_BITSET_BIT_NAME = 2, ++ ETHTOOL_A_BITSET_BIT_VALUE = 3, ++ __ETHTOOL_A_BITSET_BIT_CNT = 4, ++ ETHTOOL_A_BITSET_BIT_MAX = 3, ++}; ++ ++enum { ++ ETHTOOL_A_BITSET_BITS_UNSPEC = 0, ++ ETHTOOL_A_BITSET_BITS_BIT = 1, ++ __ETHTOOL_A_BITSET_BITS_CNT = 2, ++ ETHTOOL_A_BITSET_BITS_MAX = 1, ++}; ++ ++enum { ++ ETHTOOL_A_BITSET_UNSPEC = 0, ++ ETHTOOL_A_BITSET_NOMASK = 1, ++ ETHTOOL_A_BITSET_SIZE = 2, ++ ETHTOOL_A_BITSET_BITS = 3, ++ ETHTOOL_A_BITSET_VALUE = 4, ++ ETHTOOL_A_BITSET_MASK = 5, ++ __ETHTOOL_A_BITSET_CNT = 6, ++ ETHTOOL_A_BITSET_MAX = 5, ++}; ++ ++struct linkinfo_reply_data { ++ struct ethnl_reply_data base; ++ struct ethtool_link_ksettings ksettings; ++ struct ethtool_link_settings *lsettings; ++}; ++ ++struct linkstate_reply_data { ++ struct ethnl_reply_data base; ++ int link; ++ int sqi; ++ int sqi_max; ++ struct ethtool_link_ext_stats link_stats; ++ bool link_ext_state_provided; ++ struct ethtool_link_ext_state_info ethtool_link_ext_state_info; ++}; ++ ++struct features_reply_data { ++ struct ethnl_reply_data base; ++ u32 hw[2]; ++ u32 wanted[2]; ++ u32 active[2]; ++ u32 nochange[2]; ++ u32 all[2]; ++}; ++ ++enum ethtool_supported_ring_param { ++ ETHTOOL_RING_USE_RX_BUF_LEN = 1, ++ ETHTOOL_RING_USE_CQE_SIZE = 2, ++ ETHTOOL_RING_USE_TX_PUSH = 4, ++ ETHTOOL_RING_USE_RX_PUSH = 8, ++ ETHTOOL_RING_USE_TX_PUSH_BUF_LEN = 16, ++}; ++ ++enum { ++ ETHTOOL_TCP_DATA_SPLIT_UNKNOWN = 0, ++ ETHTOOL_TCP_DATA_SPLIT_DISABLED = 1, ++ ETHTOOL_TCP_DATA_SPLIT_ENABLED = 2, ++}; ++ ++struct rings_reply_data { ++ struct ethnl_reply_data base; ++ struct ethtool_ringparam ringparam; ++ struct kernel_ethtool_ringparam kernel_ringparam; ++ u32 supported_ring_params; ++}; ++ ++enum { ++ ETHTOOL_A_PAUSE_STAT_UNSPEC = 0, ++ ETHTOOL_A_PAUSE_STAT_PAD = 1, ++ ETHTOOL_A_PAUSE_STAT_TX_FRAMES = 2, ++ ETHTOOL_A_PAUSE_STAT_RX_FRAMES = 3, ++ __ETHTOOL_A_PAUSE_STAT_CNT = 4, ++ ETHTOOL_A_PAUSE_STAT_MAX = 3, ++}; ++ ++struct pause_req_info { ++ struct ethnl_req_info base; ++ enum ethtool_mac_stats_src src; ++}; ++ ++struct pause_reply_data { ++ struct ethnl_reply_data base; ++ struct ethtool_pauseparam pauseparam; ++ struct ethtool_pause_stats pausestat; ++}; ++ ++enum { ++ ETHTOOL_A_CABLE_PAIR_A = 0, ++ ETHTOOL_A_CABLE_PAIR_B = 1, ++ ETHTOOL_A_CABLE_PAIR_C = 2, ++ ETHTOOL_A_CABLE_PAIR_D = 3, ++}; ++ ++enum { ++ ETHTOOL_A_CABLE_RESULT_UNSPEC = 0, ++ ETHTOOL_A_CABLE_RESULT_PAIR = 1, ++ ETHTOOL_A_CABLE_RESULT_CODE = 2, ++ __ETHTOOL_A_CABLE_RESULT_CNT = 3, ++ ETHTOOL_A_CABLE_RESULT_MAX = 2, ++}; ++ ++enum { ++ ETHTOOL_A_CABLE_FAULT_LENGTH_UNSPEC = 0, ++ ETHTOOL_A_CABLE_FAULT_LENGTH_PAIR = 1, ++ ETHTOOL_A_CABLE_FAULT_LENGTH_CM = 2, ++ __ETHTOOL_A_CABLE_FAULT_LENGTH_CNT = 3, ++ ETHTOOL_A_CABLE_FAULT_LENGTH_MAX = 2, ++}; ++ ++enum { ++ ETHTOOL_A_CABLE_TEST_NTF_STATUS_UNSPEC = 0, ++ ETHTOOL_A_CABLE_TEST_NTF_STATUS_STARTED = 1, ++ ETHTOOL_A_CABLE_TEST_NTF_STATUS_COMPLETED = 2, ++}; ++ ++enum { ++ ETHTOOL_A_CABLE_NEST_UNSPEC = 0, ++ ETHTOOL_A_CABLE_NEST_RESULT = 1, ++ ETHTOOL_A_CABLE_NEST_FAULT_LENGTH = 2, ++ __ETHTOOL_A_CABLE_NEST_CNT = 3, ++ ETHTOOL_A_CABLE_NEST_MAX = 2, ++}; ++ ++enum { ++ ETHTOOL_A_CABLE_TEST_NTF_UNSPEC = 0, ++ ETHTOOL_A_CABLE_TEST_NTF_HEADER = 1, ++ ETHTOOL_A_CABLE_TEST_NTF_STATUS = 2, ++ ETHTOOL_A_CABLE_TEST_NTF_NEST = 3, ++ __ETHTOOL_A_CABLE_TEST_NTF_CNT = 4, ++ ETHTOOL_A_CABLE_TEST_NTF_MAX = 3, ++}; ++ ++enum { ++ ETHTOOL_A_CABLE_TEST_TDR_CFG_UNSPEC = 0, ++ ETHTOOL_A_CABLE_TEST_TDR_CFG_FIRST = 1, ++ ETHTOOL_A_CABLE_TEST_TDR_CFG_LAST = 2, ++ ETHTOOL_A_CABLE_TEST_TDR_CFG_STEP = 3, ++ ETHTOOL_A_CABLE_TEST_TDR_CFG_PAIR = 4, ++ __ETHTOOL_A_CABLE_TEST_TDR_CFG_CNT = 5, ++ ETHTOOL_A_CABLE_TEST_TDR_CFG_MAX = 4, ++}; ++ ++enum { ++ ETHTOOL_A_CABLE_AMPLITUDE_UNSPEC = 0, ++ ETHTOOL_A_CABLE_AMPLITUDE_PAIR = 1, ++ ETHTOOL_A_CABLE_AMPLITUDE_mV = 2, ++ __ETHTOOL_A_CABLE_AMPLITUDE_CNT = 3, ++ ETHTOOL_A_CABLE_AMPLITUDE_MAX = 2, ++}; ++ ++enum { ++ ETHTOOL_A_CABLE_PULSE_UNSPEC = 0, ++ ETHTOOL_A_CABLE_PULSE_mV = 1, ++ __ETHTOOL_A_CABLE_PULSE_CNT = 2, ++ ETHTOOL_A_CABLE_PULSE_MAX = 1, ++}; ++ ++enum { ++ ETHTOOL_A_CABLE_STEP_UNSPEC = 0, ++ ETHTOOL_A_CABLE_STEP_FIRST_DISTANCE = 1, ++ ETHTOOL_A_CABLE_STEP_LAST_DISTANCE = 2, ++ ETHTOOL_A_CABLE_STEP_STEP_DISTANCE = 3, ++ __ETHTOOL_A_CABLE_STEP_CNT = 4, ++ ETHTOOL_A_CABLE_STEP_MAX = 3, ++}; ++ ++enum { ++ ETHTOOL_A_CABLE_TDR_NEST_UNSPEC = 0, ++ ETHTOOL_A_CABLE_TDR_NEST_STEP = 1, ++ ETHTOOL_A_CABLE_TDR_NEST_AMPLITUDE = 2, ++ ETHTOOL_A_CABLE_TDR_NEST_PULSE = 3, ++ __ETHTOOL_A_CABLE_TDR_NEST_CNT = 4, ++ ETHTOOL_A_CABLE_TDR_NEST_MAX = 3, ++}; ++ ++struct eeprom_req_info { ++ struct ethnl_req_info base; ++ u32 offset; ++ u32 length; ++ u8 page; ++ u8 bank; ++ u8 i2c_address; ++}; ++ ++struct eeprom_reply_data { ++ struct ethnl_reply_data base; ++ u32 length; ++ u8 *data; ++}; ++ ++enum { ++ ETHTOOL_A_STATS_GRP_UNSPEC = 0, ++ ETHTOOL_A_STATS_GRP_PAD = 1, ++ ETHTOOL_A_STATS_GRP_ID = 2, ++ ETHTOOL_A_STATS_GRP_SS_ID = 3, ++ ETHTOOL_A_STATS_GRP_STAT = 4, ++ ETHTOOL_A_STATS_GRP_HIST_RX = 5, ++ ETHTOOL_A_STATS_GRP_HIST_TX = 6, ++ ETHTOOL_A_STATS_GRP_HIST_BKT_LOW = 7, ++ ETHTOOL_A_STATS_GRP_HIST_BKT_HI = 8, ++ ETHTOOL_A_STATS_GRP_HIST_VAL = 9, ++ __ETHTOOL_A_STATS_GRP_CNT = 10, ++ ETHTOOL_A_STATS_GRP_MAX = 9, ++}; ++ ++struct stats_req_info { ++ struct ethnl_req_info base; ++ long unsigned int stat_mask[1]; ++ enum ethtool_mac_stats_src src; ++}; ++ ++struct stats_reply_data { ++ struct ethnl_reply_data base; ++ union { ++ struct { ++ struct ethtool_eth_phy_stats phy_stats; ++ struct ethtool_eth_mac_stats mac_stats; ++ struct ethtool_eth_ctrl_stats ctrl_stats; ++ struct ethtool_rmon_stats rmon_stats; ++ }; ++ struct { ++ struct ethtool_eth_phy_stats phy_stats; ++ struct ethtool_eth_mac_stats mac_stats; ++ struct ethtool_eth_ctrl_stats ctrl_stats; ++ struct ethtool_rmon_stats rmon_stats; ++ } stats; ++ }; ++ const struct ethtool_rmon_hist_range *rmon_ranges; ++}; ++ ++struct module_reply_data { ++ struct ethnl_reply_data base; ++ struct ethtool_module_power_mode_params power; ++}; ++ ++enum ip_conntrack_info { ++ IP_CT_ESTABLISHED = 0, ++ IP_CT_RELATED = 1, ++ IP_CT_NEW = 2, ++ IP_CT_IS_REPLY = 3, ++ IP_CT_ESTABLISHED_REPLY = 3, ++ IP_CT_RELATED_REPLY = 4, ++ IP_CT_NUMBER = 5, ++ IP_CT_UNTRACKED = 7, ++}; ++ ++struct nf_hook_entries_rcu_head { ++ struct callback_head head; ++ void *allocation; ++}; ++ ++struct nf_ct_hook { ++ int (*update)(struct net *, struct sk_buff *); ++ void (*destroy)(struct nf_conntrack *); ++ bool (*get_tuple_skb)(struct nf_conntrack_tuple *, const struct sk_buff *); ++ void (*attach)(struct sk_buff *, const struct sk_buff *); ++ void (*set_closing)(struct nf_conntrack *); ++}; ++ ++struct nfnl_ct_hook { ++ size_t (*build_size)(const struct nf_conn *); ++ int (*build)(struct sk_buff *, struct nf_conn *, enum ip_conntrack_info, u_int16_t, u_int16_t); ++ int (*parse)(const struct nlattr *, struct nf_conn *); ++ int (*attach_expect)(const struct nlattr *, struct nf_conn *, u32, u32); ++ void (*seq_adjust)(struct sk_buff *, struct nf_conn *, enum ip_conntrack_info, s32); ++}; ++ ++struct nf_bridge_info { ++ enum { ++ BRNF_PROTO_UNCHANGED = 0, ++ BRNF_PROTO_8021Q = 1, ++ BRNF_PROTO_PPPOE = 2, ++ } orig_proto: 8; ++ u8 pkt_otherhost: 1; ++ u8 in_prerouting: 1; ++ u8 bridged_dnat: 1; ++ u8 sabotage_in_done: 1; ++ __u16 frag_max_size; ++ struct net_device *physindev; ++ struct net_device *physoutdev; ++ union { ++ __be32 ipv4_daddr; ++ struct in6_addr ipv6_daddr; ++ char neigh_header[8]; ++ }; ++}; ++ ++struct ip_rt_info { ++ __be32 daddr; ++ __be32 saddr; ++ u_int8_t tos; ++ u_int32_t mark; ++}; ++ ++enum l2tp_debug_flags { ++ L2TP_MSG_DEBUG = 1, ++ L2TP_MSG_CONTROL = 2, ++ L2TP_MSG_SEQ = 4, ++ L2TP_MSG_DATA = 8, ++}; ++ ++struct nf_queue_handler { ++ int (*outfn)(struct nf_queue_entry *, unsigned int); ++ void (*nf_hook_drop)(struct net *); ++}; ++ ++struct bpf_nf_link { ++ struct bpf_link link; ++ struct nf_hook_ops hook_ops; ++ struct net *net; ++ u32 dead; ++}; ++ ++struct xt_tcp { ++ __u16 spts[2]; ++ __u16 dpts[2]; ++ __u8 option; ++ __u8 flg_mask; ++ __u8 flg_cmp; ++ __u8 invflags; ++}; ++ ++struct xt_udp { ++ __u16 spts[2]; ++ __u16 dpts[2]; ++ __u8 invflags; ++}; ++ ++struct ipt_icmp { ++ __u8 type; ++ __u8 code[2]; ++ __u8 invflags; ++}; ++ ++struct ip6t_icmp { ++ __u8 type; ++ __u8 code[2]; ++ __u8 invflags; ++}; ++ ++enum { ++ INET_FRAG_FIRST_IN = 1, ++ INET_FRAG_LAST_IN = 2, ++ INET_FRAG_COMPLETE = 4, ++ INET_FRAG_HASH_DEAD = 8, ++ INET_FRAG_DROP = 16, ++}; ++ ++struct ipq { ++ struct inet_frag_queue q; ++ u8 ecn; ++ u16 max_df_size; ++ int iif; ++ unsigned int rid; ++ struct inet_peer *peer; ++}; ++ ++enum pkt_hash_types { ++ PKT_HASH_TYPE_NONE = 0, ++ PKT_HASH_TYPE_L2 = 1, ++ PKT_HASH_TYPE_L3 = 2, ++ PKT_HASH_TYPE_L4 = 3, ++}; ++ ++struct ip_fraglist_iter { ++ struct sk_buff *frag; ++ struct iphdr *iph; ++ int offset; ++ unsigned int hlen; ++}; ++ ++struct ip_frag_state { ++ bool DF; ++ unsigned int hlen; ++ unsigned int ll_rs; ++ unsigned int mtu; ++ unsigned int left; ++ int offset; ++ int ptr; ++ __be16 not_last_frag; ++}; ++ ++struct ip_reply_arg { ++ struct kvec iov[1]; ++ int flags; ++ __wsum csum; ++ int csumoffset; ++ int bound_dev_if; ++ u8 tos; ++ kuid_t uid; ++}; ++ ++struct tcp_repair_opt { ++ __u32 opt_code; ++ __u32 opt_val; ++}; ++ ++struct tcp_repair_window { ++ __u32 snd_wl1; ++ __u32 snd_wnd; ++ __u32 max_window; ++ __u32 rcv_wnd; ++ __u32 rcv_wup; ++}; ++ ++enum { ++ TCP_NO_QUEUE = 0, ++ TCP_RECV_QUEUE = 1, ++ TCP_SEND_QUEUE = 2, ++ TCP_QUEUES_NR = 3, ++}; ++ ++enum { ++ TCP_NLA_PAD = 0, ++ TCP_NLA_BUSY = 1, ++ TCP_NLA_RWND_LIMITED = 2, ++ TCP_NLA_SNDBUF_LIMITED = 3, ++ TCP_NLA_DATA_SEGS_OUT = 4, ++ TCP_NLA_TOTAL_RETRANS = 5, ++ TCP_NLA_PACING_RATE = 6, ++ TCP_NLA_DELIVERY_RATE = 7, ++ TCP_NLA_SND_CWND = 8, ++ TCP_NLA_REORDERING = 9, ++ TCP_NLA_MIN_RTT = 10, ++ TCP_NLA_RECUR_RETRANS = 11, ++ TCP_NLA_DELIVERY_RATE_APP_LMT = 12, ++ TCP_NLA_SNDQ_SIZE = 13, ++ TCP_NLA_CA_STATE = 14, ++ TCP_NLA_SND_SSTHRESH = 15, ++ TCP_NLA_DELIVERED = 16, ++ TCP_NLA_DELIVERED_CE = 17, ++ TCP_NLA_BYTES_SENT = 18, ++ TCP_NLA_BYTES_RETRANS = 19, ++ TCP_NLA_DSACK_DUPS = 20, ++ TCP_NLA_REORD_SEEN = 21, ++ TCP_NLA_SRTT = 22, ++ TCP_NLA_TIMEOUT_REHASH = 23, ++ TCP_NLA_BYTES_NOTSENT = 24, ++ TCP_NLA_EDT = 25, ++ TCP_NLA_TTL = 26, ++ TCP_NLA_REHASH = 27, ++}; ++ ++struct tcp_zerocopy_receive { ++ __u64 address; ++ __u32 length; ++ __u32 recv_skip_hint; ++ __u32 inq; ++ __s32 err; ++ __u64 copybuf_address; ++ __s32 copybuf_len; ++ __u32 flags; ++ __u64 msg_control; ++ __u64 msg_controllen; ++ __u32 msg_flags; ++ __u32 reserved; ++}; ++ ++enum { ++ BPF_TCP_ESTABLISHED = 1, ++ BPF_TCP_SYN_SENT = 2, ++ BPF_TCP_SYN_RECV = 3, ++ BPF_TCP_FIN_WAIT1 = 4, ++ BPF_TCP_FIN_WAIT2 = 5, ++ BPF_TCP_TIME_WAIT = 6, ++ BPF_TCP_CLOSE = 7, ++ BPF_TCP_CLOSE_WAIT = 8, ++ BPF_TCP_LAST_ACK = 9, ++ BPF_TCP_LISTEN = 10, ++ BPF_TCP_CLOSING = 11, ++ BPF_TCP_NEW_SYN_RECV = 12, ++ BPF_TCP_MAX_STATES = 13, ++}; ++ ++struct tcp_md5sig_pool { ++ struct ahash_request *md5_req; ++ void *scratch; ++}; ++ ++enum { ++ TCP_CMSG_INQ = 1, ++ TCP_CMSG_TS = 2, ++}; ++ ++struct tcp_splice_state { ++ struct pipe_inode_info *pipe; ++ size_t len; ++ unsigned int flags; ++}; ++ ++enum tsq_flags { ++ TSQF_THROTTLED = 1, ++ TSQF_QUEUED = 2, ++ TCPF_TSQ_DEFERRED = 4, ++ TCPF_WRITE_TIMER_DEFERRED = 8, ++ TCPF_DELACK_TIMER_DEFERRED = 16, ++ TCPF_MTU_REDUCED_DEFERRED = 32, ++}; ++ ++enum { ++ BPF_WRITE_HDR_TCP_CURRENT_MSS = 1, ++ BPF_WRITE_HDR_TCP_SYNACK_COOKIE = 2, ++}; ++ ++struct mptcp_out_options { ++ u16 suboptions; ++ struct mptcp_rm_list rm_list; ++ u8 join_id; ++ u8 backup; ++ u8 reset_reason: 4; ++ u8 reset_transient: 1; ++ u8 csum_reqd: 1; ++ u8 allow_join_id0: 1; ++ union { ++ struct { ++ u64 sndr_key; ++ u64 rcvr_key; ++ u64 data_seq; ++ u32 subflow_seq; ++ u16 data_len; ++ __sum16 csum; ++ }; ++ struct { ++ struct mptcp_addr_info addr; ++ u64 ahmac; ++ }; ++ struct { ++ struct mptcp_ext ext_copy; ++ u64 fail_seq; ++ }; ++ struct { ++ u32 nonce; ++ u32 token; ++ u64 thmac; ++ u8 hmac[20]; ++ }; ++ }; ++}; ++ ++struct tcp_out_options { ++ u16 options; ++ u16 mss; ++ u8 ws; ++ u8 num_sack_blocks; ++ u8 hash_size; ++ u8 bpf_opt_len; ++ __u8 *hash_location; ++ __u32 tsval; ++ __u32 tsecr; ++ struct tcp_fastopen_cookie *fastopen_cookie; ++ struct mptcp_out_options mptcp; ++}; ++ ++struct tsq_tasklet { ++ struct tasklet_struct tasklet; ++ struct list_head head; ++}; ++ ++struct tcp_md5sig { ++ struct __kernel_sockaddr_storage tcpm_addr; ++ __u8 tcpm_flags; ++ __u8 tcpm_prefixlen; ++ __u16 tcpm_keylen; ++ int tcpm_ifindex; ++ __u8 tcpm_key[80]; ++}; ++ ++struct tcp4_pseudohdr { ++ __be32 saddr; ++ __be32 daddr; ++ __u8 pad; ++ __u8 protocol; ++ __be16 len; ++}; ++ ++enum tcp_seq_states { ++ TCP_SEQ_STATE_LISTENING = 0, ++ TCP_SEQ_STATE_ESTABLISHED = 1, ++}; ++ ++struct tcp_seq_afinfo { ++ sa_family_t family; ++}; ++ ++struct tcp_iter_state { ++ struct seq_net_private p; ++ enum tcp_seq_states state; ++ struct sock *syn_wait_sk; ++ int bucket; ++ int offset; ++ int sbucket; ++ int num; ++ loff_t last_pos; ++}; ++ ++struct bpf_tcp_iter_state { ++ struct tcp_iter_state state; ++ unsigned int cur_sk; ++ unsigned int end_sk; ++ unsigned int max_sk; ++ struct sock **batch; ++ bool st_bucket_done; ++}; ++ ++struct bpf_iter__tcp { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct sock_common *sk_common; ++ }; ++ uid_t uid; ++}; ++ ++enum tcp_metric_index { ++ TCP_METRIC_RTT = 0, ++ TCP_METRIC_RTTVAR = 1, ++ TCP_METRIC_SSTHRESH = 2, ++ TCP_METRIC_CWND = 3, ++ TCP_METRIC_REORDERING = 4, ++ TCP_METRIC_RTT_US = 5, ++ TCP_METRIC_RTTVAR_US = 6, ++ __TCP_METRIC_MAX = 7, ++}; ++ ++enum { ++ TCP_METRICS_ATTR_UNSPEC = 0, ++ TCP_METRICS_ATTR_ADDR_IPV4 = 1, ++ TCP_METRICS_ATTR_ADDR_IPV6 = 2, ++ TCP_METRICS_ATTR_AGE = 3, ++ TCP_METRICS_ATTR_TW_TSVAL = 4, ++ TCP_METRICS_ATTR_TW_TS_STAMP = 5, ++ TCP_METRICS_ATTR_VALS = 6, ++ TCP_METRICS_ATTR_FOPEN_MSS = 7, ++ TCP_METRICS_ATTR_FOPEN_SYN_DROPS = 8, ++ TCP_METRICS_ATTR_FOPEN_SYN_DROP_TS = 9, ++ TCP_METRICS_ATTR_FOPEN_COOKIE = 10, ++ TCP_METRICS_ATTR_SADDR_IPV4 = 11, ++ TCP_METRICS_ATTR_SADDR_IPV6 = 12, ++ TCP_METRICS_ATTR_PAD = 13, ++ __TCP_METRICS_ATTR_MAX = 14, ++}; ++ ++enum { ++ TCP_METRICS_CMD_UNSPEC = 0, ++ TCP_METRICS_CMD_GET = 1, ++ TCP_METRICS_CMD_DEL = 2, ++ __TCP_METRICS_CMD_MAX = 3, ++}; ++ ++struct tcp_fastopen_metrics { ++ u16 mss; ++ u16 syn_loss: 10; ++ u16 try_exp: 2; ++ long unsigned int last_syn_loss; ++ struct tcp_fastopen_cookie cookie; ++}; ++ ++struct tcp_metrics_block { ++ struct tcp_metrics_block *tcpm_next; ++ struct net *tcpm_net; ++ struct inetpeer_addr tcpm_saddr; ++ struct inetpeer_addr tcpm_daddr; ++ long unsigned int tcpm_stamp; ++ u32 tcpm_lock; ++ u32 tcpm_vals[5]; ++ struct tcp_fastopen_metrics tcpm_fastopen; ++ struct callback_head callback_head; ++}; ++ ++struct tcpm_hash_bucket { ++ struct tcp_metrics_block *chain; ++}; ++ ++typedef struct sock * (*udp_lookup_t)(const struct sk_buff *, __be16, __be16); ++ ++typedef struct sk_buff * (*gro_receive_t)(struct list_head *, struct sk_buff *); ++ ++typedef struct sk_buff * (*gro_receive_sk_t)(struct sock *, struct list_head *, struct sk_buff *); ++ ++enum ip_conntrack_status { ++ IPS_EXPECTED_BIT = 0, ++ IPS_EXPECTED = 1, ++ IPS_SEEN_REPLY_BIT = 1, ++ IPS_SEEN_REPLY = 2, ++ IPS_ASSURED_BIT = 2, ++ IPS_ASSURED = 4, ++ IPS_CONFIRMED_BIT = 3, ++ IPS_CONFIRMED = 8, ++ IPS_SRC_NAT_BIT = 4, ++ IPS_SRC_NAT = 16, ++ IPS_DST_NAT_BIT = 5, ++ IPS_DST_NAT = 32, ++ IPS_NAT_MASK = 48, ++ IPS_SEQ_ADJUST_BIT = 6, ++ IPS_SEQ_ADJUST = 64, ++ IPS_SRC_NAT_DONE_BIT = 7, ++ IPS_SRC_NAT_DONE = 128, ++ IPS_DST_NAT_DONE_BIT = 8, ++ IPS_DST_NAT_DONE = 256, ++ IPS_NAT_DONE_MASK = 384, ++ IPS_DYING_BIT = 9, ++ IPS_DYING = 512, ++ IPS_FIXED_TIMEOUT_BIT = 10, ++ IPS_FIXED_TIMEOUT = 1024, ++ IPS_TEMPLATE_BIT = 11, ++ IPS_TEMPLATE = 2048, ++ IPS_UNTRACKED_BIT = 12, ++ IPS_UNTRACKED = 4096, ++ IPS_NAT_CLASH_BIT = 12, ++ IPS_NAT_CLASH = 4096, ++ IPS_HELPER_BIT = 13, ++ IPS_HELPER = 8192, ++ IPS_OFFLOAD_BIT = 14, ++ IPS_OFFLOAD = 16384, ++ IPS_HW_OFFLOAD_BIT = 15, ++ IPS_HW_OFFLOAD = 32768, ++ IPS_UNCHANGEABLE_MASK = 56313, ++ __IPS_MAX_BIT = 16, ++}; ++ ++struct icmp_ext_hdr { ++ __u8 version: 4; ++ __u8 reserved1: 4; ++ __u8 reserved2; ++ __sum16 checksum; ++}; ++ ++struct icmp_extobj_hdr { ++ __be16 length; ++ __u8 class_num; ++ __u8 class_type; ++}; ++ ++struct icmp_ext_echo_ctype3_hdr { ++ __be16 afi; ++ __u8 addrlen; ++ __u8 reserved; ++}; ++ ++struct icmp_ext_echo_iio { ++ struct icmp_extobj_hdr extobj_hdr; ++ union { ++ char name[16]; ++ __be32 ifindex; ++ struct { ++ struct icmp_ext_echo_ctype3_hdr ctype3_hdr; ++ union { ++ __be32 ipv4_addr; ++ struct in6_addr ipv6_addr; ++ } ip_addr; ++ } addr; ++ } ident; ++}; ++ ++struct icmp_bxm { ++ struct sk_buff *skb; ++ int offset; ++ int data_len; ++ struct { ++ struct icmphdr icmph; ++ __be32 times[3]; ++ } data; ++ int head_len; ++ struct ip_options_data replyopts; ++}; ++ ++struct icmp_control { ++ enum skb_drop_reason (*handler)(struct sk_buff *); ++ short int error; ++}; ++ ++struct compat_rtentry { ++ u32 rt_pad1; ++ struct sockaddr rt_dst; ++ struct sockaddr rt_gateway; ++ struct sockaddr rt_genmask; ++ short unsigned int rt_flags; ++ short int rt_pad2; ++ u32 rt_pad3; ++ unsigned char rt_tos; ++ unsigned char rt_class; ++ short int rt_pad4; ++ short int rt_metric; ++ compat_uptr_t rt_dev; ++ u32 rt_mtu; ++ u32 rt_window; ++ short unsigned int rt_irtt; ++}; ++ ++struct fib_rt_info { ++ struct fib_info *fi; ++ u32 tb_id; ++ __be32 dst; ++ int dst_len; ++ dscp_t dscp; ++ u8 type; ++ u8 offload: 1; ++ u8 trap: 1; ++ u8 offload_failed: 1; ++ u8 unused: 5; ++}; ++ ++struct fib_nh_notifier_info { ++ struct fib_notifier_info info; ++ struct fib_nh *fib_nh; ++}; ++ ++struct fib_alias { ++ struct hlist_node fa_list; ++ struct fib_info *fa_info; ++ dscp_t fa_dscp; ++ u8 fa_type; ++ u8 fa_state; ++ u8 fa_slen; ++ u32 tb_id; ++ s16 fa_default; ++ u8 offload; ++ u8 trap; ++ u8 offload_failed; ++ struct callback_head rcu; ++}; ++ ++struct fib_prop { ++ int error; ++ u8 scope; ++}; ++ ++struct ipfrag_skb_cb { ++ union { ++ struct inet_skb_parm h4; ++ struct inet6_skb_parm h6; ++ }; ++ struct sk_buff *next_frag; ++ int frag_run_len; ++}; ++ ++enum lwtunnel_ip_t { ++ LWTUNNEL_IP_UNSPEC = 0, ++ LWTUNNEL_IP_ID = 1, ++ LWTUNNEL_IP_DST = 2, ++ LWTUNNEL_IP_SRC = 3, ++ LWTUNNEL_IP_TTL = 4, ++ LWTUNNEL_IP_TOS = 5, ++ LWTUNNEL_IP_FLAGS = 6, ++ LWTUNNEL_IP_PAD = 7, ++ LWTUNNEL_IP_OPTS = 8, ++ __LWTUNNEL_IP_MAX = 9, ++}; ++ ++enum lwtunnel_ip6_t { ++ LWTUNNEL_IP6_UNSPEC = 0, ++ LWTUNNEL_IP6_ID = 1, ++ LWTUNNEL_IP6_DST = 2, ++ LWTUNNEL_IP6_SRC = 3, ++ LWTUNNEL_IP6_HOPLIMIT = 4, ++ LWTUNNEL_IP6_TC = 5, ++ LWTUNNEL_IP6_FLAGS = 6, ++ LWTUNNEL_IP6_PAD = 7, ++ LWTUNNEL_IP6_OPTS = 8, ++ __LWTUNNEL_IP6_MAX = 9, ++}; ++ ++enum { ++ LWTUNNEL_IP_OPTS_UNSPEC = 0, ++ LWTUNNEL_IP_OPTS_GENEVE = 1, ++ LWTUNNEL_IP_OPTS_VXLAN = 2, ++ LWTUNNEL_IP_OPTS_ERSPAN = 3, ++ __LWTUNNEL_IP_OPTS_MAX = 4, ++}; ++ ++enum { ++ LWTUNNEL_IP_OPT_GENEVE_UNSPEC = 0, ++ LWTUNNEL_IP_OPT_GENEVE_CLASS = 1, ++ LWTUNNEL_IP_OPT_GENEVE_TYPE = 2, ++ LWTUNNEL_IP_OPT_GENEVE_DATA = 3, ++ __LWTUNNEL_IP_OPT_GENEVE_MAX = 4, ++}; ++ ++enum { ++ LWTUNNEL_IP_OPT_VXLAN_UNSPEC = 0, ++ LWTUNNEL_IP_OPT_VXLAN_GBP = 1, ++ __LWTUNNEL_IP_OPT_VXLAN_MAX = 2, ++}; ++ ++enum { ++ LWTUNNEL_IP_OPT_ERSPAN_UNSPEC = 0, ++ LWTUNNEL_IP_OPT_ERSPAN_VER = 1, ++ LWTUNNEL_IP_OPT_ERSPAN_INDEX = 2, ++ LWTUNNEL_IP_OPT_ERSPAN_DIR = 3, ++ LWTUNNEL_IP_OPT_ERSPAN_HWID = 4, ++ __LWTUNNEL_IP_OPT_ERSPAN_MAX = 5, ++}; ++ ++enum { ++ IFLA_IPTUN_UNSPEC = 0, ++ IFLA_IPTUN_LINK = 1, ++ IFLA_IPTUN_LOCAL = 2, ++ IFLA_IPTUN_REMOTE = 3, ++ IFLA_IPTUN_TTL = 4, ++ IFLA_IPTUN_TOS = 5, ++ IFLA_IPTUN_ENCAP_LIMIT = 6, ++ IFLA_IPTUN_FLOWINFO = 7, ++ IFLA_IPTUN_FLAGS = 8, ++ IFLA_IPTUN_PROTO = 9, ++ IFLA_IPTUN_PMTUDISC = 10, ++ IFLA_IPTUN_6RD_PREFIX = 11, ++ IFLA_IPTUN_6RD_RELAY_PREFIX = 12, ++ IFLA_IPTUN_6RD_PREFIXLEN = 13, ++ IFLA_IPTUN_6RD_RELAY_PREFIXLEN = 14, ++ IFLA_IPTUN_ENCAP_TYPE = 15, ++ IFLA_IPTUN_ENCAP_FLAGS = 16, ++ IFLA_IPTUN_ENCAP_SPORT = 17, ++ IFLA_IPTUN_ENCAP_DPORT = 18, ++ IFLA_IPTUN_COLLECT_METADATA = 19, ++ IFLA_IPTUN_FWMARK = 20, ++ __IFLA_IPTUN_MAX = 21, ++}; ++ ++struct ip6_tnl_encap_ops { ++ size_t (*encap_hlen)(struct ip_tunnel_encap *); ++ int (*build_header)(struct sk_buff *, struct ip_tunnel_encap *, u8 *, struct flowi6 *); ++ int (*err_handler)(struct sk_buff *, struct inet6_skb_parm *, u8, u8, int, __be32); ++}; ++ ++struct geneve_opt { ++ __be16 opt_class; ++ u8 type; ++ u8 r1: 1; ++ u8 r2: 1; ++ u8 r3: 1; ++ u8 length: 5; ++ u8 opt_data[0]; ++}; ++ ++struct vxlan_metadata { ++ u32 gbp; ++}; ++ ++struct erspan_md2 { ++ __be32 timestamp; ++ __be16 sgt; ++ __u8 p: 1; ++ __u8 ft: 5; ++ __u8 hwid_upper: 2; ++ __u8 hwid: 4; ++ __u8 dir: 1; ++ __u8 gra: 2; ++ __u8 o: 1; ++}; ++ ++struct erspan_metadata { ++ int version; ++ union { ++ __be32 index; ++ struct erspan_md2 md2; ++ } u; ++}; ++ ++struct fib4_rule { ++ struct fib_rule common; ++ u8 dst_len; ++ u8 src_len; ++ dscp_t dscp; ++ __be32 src; ++ __be32 srcmask; ++ __be32 dst; ++ __be32 dstmask; ++ u32 tclassid; ++}; ++ ++struct rta_mfc_stats { ++ __u64 mfcs_packets; ++ __u64 mfcs_bytes; ++ __u64 mfcs_wrong_if; ++}; ++ ++enum { ++ INET_DIAG_NONE = 0, ++ INET_DIAG_MEMINFO = 1, ++ INET_DIAG_INFO = 2, ++ INET_DIAG_VEGASINFO = 3, ++ INET_DIAG_CONG = 4, ++ INET_DIAG_TOS = 5, ++ INET_DIAG_TCLASS = 6, ++ INET_DIAG_SKMEMINFO = 7, ++ INET_DIAG_SHUTDOWN = 8, ++ INET_DIAG_DCTCPINFO = 9, ++ INET_DIAG_PROTOCOL = 10, ++ INET_DIAG_SKV6ONLY = 11, ++ INET_DIAG_LOCALS = 12, ++ INET_DIAG_PEERS = 13, ++ INET_DIAG_PAD = 14, ++ INET_DIAG_MARK = 15, ++ INET_DIAG_BBRINFO = 16, ++ INET_DIAG_CLASS_ID = 17, ++ INET_DIAG_MD5SIG = 18, ++ INET_DIAG_ULP_INFO = 19, ++ INET_DIAG_SK_BPF_STORAGES = 20, ++ INET_DIAG_CGROUP_ID = 21, ++ INET_DIAG_SOCKOPT = 22, ++ __INET_DIAG_MAX = 23, ++}; ++ ++struct tcp_diag_md5sig { ++ __u8 tcpm_family; ++ __u8 tcpm_prefixlen; ++ __u16 tcpm_keylen; ++ __be32 tcpm_addr[4]; ++ __u8 tcpm_key[80]; ++}; ++ ++struct inet_diag_req_raw { ++ __u8 sdiag_family; ++ __u8 sdiag_protocol; ++ __u8 idiag_ext; ++ __u8 sdiag_raw_protocol; ++ __u32 idiag_states; ++ struct inet_diag_sockid id; ++}; ++ ++enum { ++ UDP_BPF_IPV4 = 0, ++ UDP_BPF_IPV6 = 1, ++ UDP_BPF_NUM_PROTS = 2, ++}; ++ ++struct xfrm_input_afinfo { ++ u8 family; ++ bool is_ipip; ++ int (*callback)(struct sk_buff *, u8, int); ++}; ++ ++struct xfrm4_protocol { ++ int (*handler)(struct sk_buff *); ++ int (*input_handler)(struct sk_buff *, int, __be32, int); ++ int (*cb_handler)(struct sk_buff *, int); ++ int (*err_handler)(struct sk_buff *, u32); ++ struct xfrm4_protocol *next; ++ int priority; ++}; ++ ++typedef u64 (*btf_bpf_tcp_send_ack)(struct tcp_sock *, u32); ++ ++struct ip_beet_phdr { ++ __u8 nexthdr; ++ __u8 hdrlen; ++ __u8 padlen; ++ __u8 reserved; ++}; ++ ++struct gro_cell; ++ ++struct gro_cells { ++ struct gro_cell *cells; ++}; ++ ++enum { ++ XFRM_MODE_FLAG_TUNNEL = 1, ++}; ++ ++struct ip_tunnel_6rd_parm { ++ struct in6_addr prefix; ++ __be32 relay_prefix; ++ u16 prefixlen; ++ u16 relay_prefixlen; ++}; ++ ++struct ip_tunnel_prl_entry; ++ ++struct ip_tunnel { ++ struct ip_tunnel *next; ++ struct hlist_node hash_node; ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ struct net *net; ++ long unsigned int err_time; ++ int err_count; ++ u32 i_seqno; ++ atomic_t o_seqno; ++ int tun_hlen; ++ u32 index; ++ u8 erspan_ver; ++ u8 dir; ++ u16 hwid; ++ struct dst_cache dst_cache; ++ struct ip_tunnel_parm parms; ++ int mlink; ++ int encap_hlen; ++ int hlen; ++ struct ip_tunnel_encap encap; ++ struct ip_tunnel_6rd_parm ip6rd; ++ struct ip_tunnel_prl_entry *prl; ++ unsigned int prl_count; ++ unsigned int ip_tnl_net_id; ++ struct gro_cells gro_cells; ++ __u32 fwmark; ++ bool collect_md; ++ bool ignore_df; ++}; ++ ++struct __ip6_tnl_parm { ++ char name[16]; ++ int link; ++ __u8 proto; ++ __u8 encap_limit; ++ __u8 hop_limit; ++ bool collect_md; ++ __be32 flowinfo; ++ __u32 flags; ++ struct in6_addr laddr; ++ struct in6_addr raddr; ++ __be16 i_flags; ++ __be16 o_flags; ++ __be32 i_key; ++ __be32 o_key; ++ __u32 fwmark; ++ __u32 index; ++ __u8 erspan_ver; ++ __u8 dir; ++ __u16 hwid; ++}; ++ ++struct ip6_tnl { ++ struct ip6_tnl *next; ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ struct net *net; ++ struct __ip6_tnl_parm parms; ++ struct flowi fl; ++ struct dst_cache dst_cache; ++ struct gro_cells gro_cells; ++ int err_count; ++ long unsigned int err_time; ++ __u32 i_seqno; ++ atomic_t o_seqno; ++ int hlen; ++ int tun_hlen; ++ int encap_hlen; ++ struct ip_tunnel_encap encap; ++ int mlink; ++}; ++ ++struct xfrm_skb_cb { ++ struct xfrm_tunnel_skb_cb header; ++ union { ++ struct { ++ __u32 low; ++ __u32 hi; ++ } output; ++ struct { ++ __be32 low; ++ __be32 hi; ++ } input; ++ } seq; ++}; ++ ++struct ip_tunnel_prl_entry { ++ struct ip_tunnel_prl_entry *next; ++ __be32 addr; ++ u16 flags; ++ struct callback_head callback_head; ++}; ++ ++struct xfrm_trans_tasklet { ++ struct work_struct work; ++ spinlock_t queue_lock; ++ struct sk_buff_head queue; ++}; ++ ++struct xfrm_trans_cb { ++ union { ++ struct inet_skb_parm h4; ++ struct inet6_skb_parm h6; ++ } header; ++ int (*finish)(struct net *, struct sock *, struct sk_buff *); ++ struct net *net; ++}; ++ ++struct xfrm_user_offload { ++ int ifindex; ++ __u8 flags; ++}; ++ ++enum { ++ XFRM_DEV_OFFLOAD_IN = 1, ++ XFRM_DEV_OFFLOAD_OUT = 2, ++ XFRM_DEV_OFFLOAD_FWD = 3, ++}; ++ ++struct sadb_alg { ++ __u8 sadb_alg_id; ++ __u8 sadb_alg_ivlen; ++ __u16 sadb_alg_minbits; ++ __u16 sadb_alg_maxbits; ++ __u16 sadb_alg_reserved; ++}; ++ ++struct xfrm_algo_aead_info { ++ char *geniv; ++ u16 icv_truncbits; ++}; ++ ++struct xfrm_algo_auth_info { ++ u16 icv_truncbits; ++ u16 icv_fullbits; ++}; ++ ++struct xfrm_algo_encr_info { ++ char *geniv; ++ u16 blockbits; ++ u16 defkeybits; ++}; ++ ++struct xfrm_algo_comp_info { ++ u16 threshold; ++}; ++ ++struct xfrm_algo_desc { ++ char *name; ++ char *compat; ++ u8 available: 1; ++ u8 pfkey_supported: 1; ++ union { ++ struct xfrm_algo_aead_info aead; ++ struct xfrm_algo_auth_info auth; ++ struct xfrm_algo_encr_info encr; ++ struct xfrm_algo_comp_info comp; ++ } uinfo; ++ struct sadb_alg desc; ++}; ++ ++struct xfrm_algo_list { ++ struct xfrm_algo_desc *algs; ++ int entries; ++ u32 type; ++ u32 mask; ++}; ++ ++struct xfrm_aead_name { ++ const char *name; ++ int icvbits; ++}; ++ ++enum { ++ XFRM_SHARE_ANY = 0, ++ XFRM_SHARE_SESSION = 1, ++ XFRM_SHARE_USER = 2, ++ XFRM_SHARE_UNIQUE = 3, ++}; ++ ++struct xfrm_user_tmpl { ++ struct xfrm_id id; ++ __u16 family; ++ xfrm_address_t saddr; ++ __u32 reqid; ++ __u8 mode; ++ __u8 share; ++ __u8 optional; ++ __u32 aalgos; ++ __u32 ealgos; ++ __u32 calgos; ++}; ++ ++enum xfrm_ae_ftype_t { ++ XFRM_AE_UNSPEC = 0, ++ XFRM_AE_RTHR = 1, ++ XFRM_AE_RVAL = 2, ++ XFRM_AE_LVAL = 4, ++ XFRM_AE_ETHR = 8, ++ XFRM_AE_CR = 16, ++ XFRM_AE_CE = 32, ++ XFRM_AE_CU = 64, ++ __XFRM_AE_MAX = 65, ++}; ++ ++struct xfrm_userpolicy_type { ++ __u8 type; ++ __u16 reserved1; ++ __u8 reserved2; ++}; ++ ++enum xfrm_sadattr_type_t { ++ XFRMA_SAD_UNSPEC = 0, ++ XFRMA_SAD_CNT = 1, ++ XFRMA_SAD_HINFO = 2, ++ __XFRMA_SAD_MAX = 3, ++}; ++ ++struct xfrmu_sadhinfo { ++ __u32 sadhcnt; ++ __u32 sadhmcnt; ++}; ++ ++enum xfrm_spdattr_type_t { ++ XFRMA_SPD_UNSPEC = 0, ++ XFRMA_SPD_INFO = 1, ++ XFRMA_SPD_HINFO = 2, ++ XFRMA_SPD_IPV4_HTHRESH = 3, ++ XFRMA_SPD_IPV6_HTHRESH = 4, ++ __XFRMA_SPD_MAX = 5, ++}; ++ ++struct xfrmu_spdinfo { ++ __u32 incnt; ++ __u32 outcnt; ++ __u32 fwdcnt; ++ __u32 inscnt; ++ __u32 outscnt; ++ __u32 fwdscnt; ++}; ++ ++struct xfrmu_spdhinfo { ++ __u32 spdhcnt; ++ __u32 spdhmcnt; ++}; ++ ++struct xfrmu_spdhthresh { ++ __u8 lbits; ++ __u8 rbits; ++}; ++ ++struct xfrm_usersa_info { ++ struct xfrm_selector sel; ++ struct xfrm_id id; ++ xfrm_address_t saddr; ++ struct xfrm_lifetime_cfg lft; ++ struct xfrm_lifetime_cur curlft; ++ struct xfrm_stats stats; ++ __u32 seq; ++ __u32 reqid; ++ __u16 family; ++ __u8 mode; ++ __u8 replay_window; ++ __u8 flags; ++}; ++ ++struct xfrm_usersa_id { ++ xfrm_address_t daddr; ++ __be32 spi; ++ __u16 family; ++ __u8 proto; ++}; ++ ++struct xfrm_aevent_id { ++ struct xfrm_usersa_id sa_id; ++ xfrm_address_t saddr; ++ __u32 flags; ++ __u32 reqid; ++}; ++ ++struct xfrm_userspi_info { ++ struct xfrm_usersa_info info; ++ __u32 min; ++ __u32 max; ++}; ++ ++struct xfrm_userpolicy_info { ++ struct xfrm_selector sel; ++ struct xfrm_lifetime_cfg lft; ++ struct xfrm_lifetime_cur curlft; ++ __u32 priority; ++ __u32 index; ++ __u8 dir; ++ __u8 action; ++ __u8 flags; ++ __u8 share; ++}; ++ ++struct xfrm_userpolicy_id { ++ struct xfrm_selector sel; ++ __u32 index; ++ __u8 dir; ++}; ++ ++struct xfrm_user_acquire { ++ struct xfrm_id id; ++ xfrm_address_t saddr; ++ struct xfrm_selector sel; ++ struct xfrm_userpolicy_info policy; ++ __u32 aalgos; ++ __u32 ealgos; ++ __u32 calgos; ++ __u32 seq; ++}; ++ ++struct xfrm_user_expire { ++ struct xfrm_usersa_info state; ++ __u8 hard; ++}; ++ ++struct xfrm_user_polexpire { ++ struct xfrm_userpolicy_info pol; ++ __u8 hard; ++}; ++ ++struct xfrm_usersa_flush { ++ __u8 proto; ++}; ++ ++struct xfrm_user_report { ++ __u8 proto; ++ struct xfrm_selector sel; ++}; ++ ++struct xfrm_user_kmaddress { ++ xfrm_address_t local; ++ xfrm_address_t remote; ++ __u32 reserved; ++ __u16 family; ++}; ++ ++struct xfrm_user_migrate { ++ xfrm_address_t old_daddr; ++ xfrm_address_t old_saddr; ++ xfrm_address_t new_daddr; ++ xfrm_address_t new_saddr; ++ __u8 proto; ++ __u8 mode; ++ __u16 reserved; ++ __u32 reqid; ++ __u16 old_family; ++ __u16 new_family; ++}; ++ ++struct xfrm_user_mapping { ++ struct xfrm_usersa_id id; ++ __u32 reqid; ++ xfrm_address_t old_saddr; ++ xfrm_address_t new_saddr; ++ __be16 old_sport; ++ __be16 new_sport; ++}; ++ ++struct xfrm_userpolicy_default { ++ __u8 in; ++ __u8 fwd; ++ __u8 out; ++}; ++ ++enum xfrm_nlgroups { ++ XFRMNLGRP_NONE = 0, ++ XFRMNLGRP_ACQUIRE = 1, ++ XFRMNLGRP_EXPIRE = 2, ++ XFRMNLGRP_SA = 3, ++ XFRMNLGRP_POLICY = 4, ++ XFRMNLGRP_AEVENTS = 5, ++ XFRMNLGRP_REPORT = 6, ++ XFRMNLGRP_MIGRATE = 7, ++ XFRMNLGRP_MAPPING = 8, ++ __XFRMNLGRP_MAX = 9, ++}; ++ ++struct km_event { ++ union { ++ u32 hard; ++ u32 proto; ++ u32 byid; ++ u32 aevent; ++ u32 type; ++ } data; ++ u32 seq; ++ u32 portid; ++ u32 event; ++ struct net *net; ++}; ++ ++struct xfrm_mgr { ++ struct list_head list; ++ int (*notify)(struct xfrm_state *, const struct km_event *); ++ int (*acquire)(struct xfrm_state *, struct xfrm_tmpl *, struct xfrm_policy *); ++ struct xfrm_policy * (*compile_policy)(struct sock *, int, u8 *, int, int *); ++ int (*new_mapping)(struct xfrm_state *, xfrm_address_t *, __be16); ++ int (*notify_policy)(struct xfrm_policy *, int, const struct km_event *); ++ int (*report)(struct net *, u8, struct xfrm_selector *, xfrm_address_t *); ++ int (*migrate)(const struct xfrm_selector *, u8, u8, const struct xfrm_migrate *, int, const struct xfrm_kmaddress *, const struct xfrm_encap_tmpl *); ++ bool (*is_alive)(const struct km_event *); ++}; ++ ++struct xfrmk_sadinfo { ++ u32 sadhcnt; ++ u32 sadhmcnt; ++ u32 sadcnt; ++}; ++ ++struct xfrm_translator { ++ int (*alloc_compat)(struct sk_buff *, const struct nlmsghdr *); ++ struct nlmsghdr * (*rcv_msg_compat)(const struct nlmsghdr *, int, const struct nla_policy *, struct netlink_ext_ack *); ++ int (*xlate_user_policy_sockptr)(u8 **, int); ++ struct module *owner; ++}; ++ ++struct xfrm_dump_info { ++ struct sk_buff *in_skb; ++ struct sk_buff *out_skb; ++ u32 nlmsg_seq; ++ u16 nlmsg_flags; ++}; ++ ++struct xfrm_link { ++ int (*doit)(struct sk_buff *, struct nlmsghdr *, struct nlattr **, struct netlink_ext_ack *); ++ int (*start)(struct netlink_callback *); ++ int (*dump)(struct sk_buff *, struct netlink_callback *); ++ int (*done)(struct netlink_callback *); ++ const struct nla_policy *nla_pol; ++ int nla_max; ++}; ++ ++struct ac6_iter_state { ++ struct seq_net_private p; ++ struct net_device *dev; ++ struct inet6_dev *idev; ++}; ++ ++enum { ++ IFLA_INET6_UNSPEC = 0, ++ IFLA_INET6_FLAGS = 1, ++ IFLA_INET6_CONF = 2, ++ IFLA_INET6_STATS = 3, ++ IFLA_INET6_MCAST = 4, ++ IFLA_INET6_CACHEINFO = 5, ++ IFLA_INET6_ICMP6STATS = 6, ++ IFLA_INET6_TOKEN = 7, ++ IFLA_INET6_ADDR_GEN_MODE = 8, ++ IFLA_INET6_RA_MTU = 9, ++ __IFLA_INET6_MAX = 10, ++}; ++ ++enum in6_addr_gen_mode { ++ IN6_ADDR_GEN_MODE_EUI64 = 0, ++ IN6_ADDR_GEN_MODE_NONE = 1, ++ IN6_ADDR_GEN_MODE_STABLE_PRIVACY = 2, ++ IN6_ADDR_GEN_MODE_RANDOM = 3, ++}; ++ ++struct ifla_cacheinfo { ++ __u32 max_reasm_len; ++ __u32 tstamp; ++ __u32 reachable_time; ++ __u32 retrans_time; ++}; ++ ++struct wpan_phy; ++ ++struct wpan_dev_header_ops; ++ ++struct wpan_dev { ++ struct wpan_phy *wpan_phy; ++ int iftype; ++ struct list_head list; ++ struct net_device *netdev; ++ const struct wpan_dev_header_ops *header_ops; ++ struct net_device *lowpan_dev; ++ u32 identifier; ++ __le16 pan_id; ++ __le16 short_addr; ++ __le64 extended_addr; ++ atomic_t bsn; ++ atomic_t dsn; ++ u8 min_be; ++ u8 max_be; ++ u8 csma_retries; ++ s8 frame_retries; ++ bool lbt; ++ bool ackreq; ++}; ++ ++struct prefixmsg { ++ unsigned char prefix_family; ++ unsigned char prefix_pad1; ++ short unsigned int prefix_pad2; ++ int prefix_ifindex; ++ unsigned char prefix_type; ++ unsigned char prefix_len; ++ unsigned char prefix_flags; ++ unsigned char prefix_pad3; ++}; ++ ++enum { ++ PREFIX_UNSPEC = 0, ++ PREFIX_ADDRESS = 1, ++ PREFIX_CACHEINFO = 2, ++ __PREFIX_MAX = 3, ++}; ++ ++struct prefix_cacheinfo { ++ __u32 preferred_time; ++ __u32 valid_time; ++}; ++ ++struct in6_ifreq { ++ struct in6_addr ifr6_addr; ++ __u32 ifr6_prefixlen; ++ int ifr6_ifindex; ++}; ++ ++enum { ++ DEVCONF_FORWARDING = 0, ++ DEVCONF_HOPLIMIT = 1, ++ DEVCONF_MTU6 = 2, ++ DEVCONF_ACCEPT_RA = 3, ++ DEVCONF_ACCEPT_REDIRECTS = 4, ++ DEVCONF_AUTOCONF = 5, ++ DEVCONF_DAD_TRANSMITS = 6, ++ DEVCONF_RTR_SOLICITS = 7, ++ DEVCONF_RTR_SOLICIT_INTERVAL = 8, ++ DEVCONF_RTR_SOLICIT_DELAY = 9, ++ DEVCONF_USE_TEMPADDR = 10, ++ DEVCONF_TEMP_VALID_LFT = 11, ++ DEVCONF_TEMP_PREFERED_LFT = 12, ++ DEVCONF_REGEN_MAX_RETRY = 13, ++ DEVCONF_MAX_DESYNC_FACTOR = 14, ++ DEVCONF_MAX_ADDRESSES = 15, ++ DEVCONF_FORCE_MLD_VERSION = 16, ++ DEVCONF_ACCEPT_RA_DEFRTR = 17, ++ DEVCONF_ACCEPT_RA_PINFO = 18, ++ DEVCONF_ACCEPT_RA_RTR_PREF = 19, ++ DEVCONF_RTR_PROBE_INTERVAL = 20, ++ DEVCONF_ACCEPT_RA_RT_INFO_MAX_PLEN = 21, ++ DEVCONF_PROXY_NDP = 22, ++ DEVCONF_OPTIMISTIC_DAD = 23, ++ DEVCONF_ACCEPT_SOURCE_ROUTE = 24, ++ DEVCONF_MC_FORWARDING = 25, ++ DEVCONF_DISABLE_IPV6 = 26, ++ DEVCONF_ACCEPT_DAD = 27, ++ DEVCONF_FORCE_TLLAO = 28, ++ DEVCONF_NDISC_NOTIFY = 29, ++ DEVCONF_MLDV1_UNSOLICITED_REPORT_INTERVAL = 30, ++ DEVCONF_MLDV2_UNSOLICITED_REPORT_INTERVAL = 31, ++ DEVCONF_SUPPRESS_FRAG_NDISC = 32, ++ DEVCONF_ACCEPT_RA_FROM_LOCAL = 33, ++ DEVCONF_USE_OPTIMISTIC = 34, ++ DEVCONF_ACCEPT_RA_MTU = 35, ++ DEVCONF_STABLE_SECRET = 36, ++ DEVCONF_USE_OIF_ADDRS_ONLY = 37, ++ DEVCONF_ACCEPT_RA_MIN_HOP_LIMIT = 38, ++ DEVCONF_IGNORE_ROUTES_WITH_LINKDOWN = 39, ++ DEVCONF_DROP_UNICAST_IN_L2_MULTICAST = 40, ++ DEVCONF_DROP_UNSOLICITED_NA = 41, ++ DEVCONF_KEEP_ADDR_ON_DOWN = 42, ++ DEVCONF_RTR_SOLICIT_MAX_INTERVAL = 43, ++ DEVCONF_SEG6_ENABLED = 44, ++ DEVCONF_SEG6_REQUIRE_HMAC = 45, ++ DEVCONF_ENHANCED_DAD = 46, ++ DEVCONF_ADDR_GEN_MODE = 47, ++ DEVCONF_DISABLE_POLICY = 48, ++ DEVCONF_ACCEPT_RA_RT_INFO_MIN_PLEN = 49, ++ DEVCONF_NDISC_TCLASS = 50, ++ DEVCONF_RPL_SEG_ENABLED = 51, ++ DEVCONF_RA_DEFRTR_METRIC = 52, ++ DEVCONF_IOAM6_ENABLED = 53, ++ DEVCONF_IOAM6_ID = 54, ++ DEVCONF_IOAM6_ID_WIDE = 55, ++ DEVCONF_NDISC_EVICT_NOCARRIER = 56, ++ DEVCONF_ACCEPT_UNTRACKED_NA = 57, ++ DEVCONF_MAX = 58, ++}; ++ ++enum { ++ INET6_IFADDR_STATE_PREDAD = 0, ++ INET6_IFADDR_STATE_DAD = 1, ++ INET6_IFADDR_STATE_POSTDAD = 2, ++ INET6_IFADDR_STATE_ERRDAD = 3, ++ INET6_IFADDR_STATE_DEAD = 4, ++}; ++ ++enum ieee802154_filtering_level { ++ IEEE802154_FILTERING_NONE = 0, ++ IEEE802154_FILTERING_1_FCS = 1, ++ IEEE802154_FILTERING_2_PROMISCUOUS = 2, ++ IEEE802154_FILTERING_3_SCAN = 3, ++ IEEE802154_FILTERING_4_FRAME_FIELDS = 4, ++}; ++ ++enum nl802154_cca_modes { ++ __NL802154_CCA_INVALID = 0, ++ NL802154_CCA_ENERGY = 1, ++ NL802154_CCA_CARRIER = 2, ++ NL802154_CCA_ENERGY_CARRIER = 3, ++ NL802154_CCA_ALOHA = 4, ++ NL802154_CCA_UWB_SHR = 5, ++ NL802154_CCA_UWB_MULTIPLEXED = 6, ++ __NL802154_CCA_ATTR_AFTER_LAST = 7, ++ NL802154_CCA_ATTR_MAX = 6, ++}; ++ ++enum nl802154_cca_opts { ++ NL802154_CCA_OPT_ENERGY_CARRIER_AND = 0, ++ NL802154_CCA_OPT_ENERGY_CARRIER_OR = 1, ++ __NL802154_CCA_OPT_ATTR_AFTER_LAST = 2, ++ NL802154_CCA_OPT_ATTR_MAX = 1, ++}; ++ ++enum nl802154_supported_bool_states { ++ NL802154_SUPPORTED_BOOL_FALSE = 0, ++ NL802154_SUPPORTED_BOOL_TRUE = 1, ++ __NL802154_SUPPORTED_BOOL_INVALD = 2, ++ NL802154_SUPPORTED_BOOL_BOTH = 3, ++ __NL802154_SUPPORTED_BOOL_AFTER_LAST = 4, ++ NL802154_SUPPORTED_BOOL_MAX = 3, ++}; ++ ++struct wpan_phy_supported { ++ u32 channels[32]; ++ u32 cca_modes; ++ u32 cca_opts; ++ u32 iftypes; ++ enum nl802154_supported_bool_states lbt; ++ u8 min_minbe; ++ u8 max_minbe; ++ u8 min_maxbe; ++ u8 max_maxbe; ++ u8 min_csma_backoffs; ++ u8 max_csma_backoffs; ++ s8 min_frame_retries; ++ s8 max_frame_retries; ++ size_t tx_powers_size; ++ size_t cca_ed_levels_size; ++ const s32 *tx_powers; ++ const s32 *cca_ed_levels; ++}; ++ ++struct wpan_phy_cca { ++ enum nl802154_cca_modes mode; ++ enum nl802154_cca_opts opt; ++}; ++ ++struct wpan_phy { ++ const void *privid; ++ long unsigned int flags; ++ u8 current_channel; ++ u8 current_page; ++ struct wpan_phy_supported supported; ++ s32 transmit_power; ++ struct wpan_phy_cca cca; ++ __le64 perm_extended_addr; ++ s32 cca_ed_level; ++ u32 symbol_duration; ++ u16 lifs_period; ++ u16 sifs_period; ++ struct device dev; ++ possible_net_t _net; ++ spinlock_t queue_lock; ++ atomic_t ongoing_txs; ++ atomic_t hold_txs; ++ wait_queue_head_t sync_txq; ++ enum ieee802154_filtering_level filtering; ++ long: 0; ++ char priv[0]; ++}; ++ ++struct ieee802154_addr { ++ u8 mode; ++ __le16 pan_id; ++ union { ++ __le16 short_addr; ++ __le64 extended_addr; ++ }; ++}; ++ ++struct wpan_dev_header_ops { ++ int (*create)(struct sk_buff *, struct net_device *, const struct ieee802154_addr *, const struct ieee802154_addr *, unsigned int); ++}; ++ ++union fwnet_hwaddr { ++ u8 u[16]; ++ struct { ++ __be64 uniq_id; ++ u8 max_rec; ++ u8 sspd; ++ u8 fifo[6]; ++ } uc; ++}; ++ ++struct in6_validator_info { ++ struct in6_addr i6vi_addr; ++ struct inet6_dev *i6vi_dev; ++ struct netlink_ext_ack *extack; ++}; ++ ++struct ifa6_config { ++ const struct in6_addr *pfx; ++ unsigned int plen; ++ u8 ifa_proto; ++ const struct in6_addr *peer_pfx; ++ u32 rt_priority; ++ u32 ifa_flags; ++ u32 preferred_lft; ++ u32 valid_lft; ++ u16 scope; ++}; ++ ++enum cleanup_prefix_rt_t { ++ CLEANUP_PREFIX_RT_NOP = 0, ++ CLEANUP_PREFIX_RT_DEL = 1, ++ CLEANUP_PREFIX_RT_EXPIRE = 2, ++}; ++ ++enum { ++ IPV6_SADDR_RULE_INIT = 0, ++ IPV6_SADDR_RULE_LOCAL = 1, ++ IPV6_SADDR_RULE_SCOPE = 2, ++ IPV6_SADDR_RULE_PREFERRED = 3, ++ IPV6_SADDR_RULE_HOA = 4, ++ IPV6_SADDR_RULE_OIF = 5, ++ IPV6_SADDR_RULE_LABEL = 6, ++ IPV6_SADDR_RULE_PRIVACY = 7, ++ IPV6_SADDR_RULE_ORCHID = 8, ++ IPV6_SADDR_RULE_PREFIX = 9, ++ IPV6_SADDR_RULE_NOT_OPTIMISTIC = 10, ++ IPV6_SADDR_RULE_MAX = 11, ++}; ++ ++struct ipv6_saddr_score { ++ int rule; ++ int addr_type; ++ struct inet6_ifaddr *ifa; ++ long unsigned int scorebits[1]; ++ int scopedist; ++ int matchlen; ++}; ++ ++struct ipv6_saddr_dst { ++ const struct in6_addr *addr; ++ int ifindex; ++ int scope; ++ int label; ++ unsigned int prefs; ++}; ++ ++struct if6_iter_state { ++ struct seq_net_private p; ++ int bucket; ++ int offset; ++}; ++ ++enum addr_type_t { ++ UNICAST_ADDR = 0, ++ MULTICAST_ADDR = 1, ++ ANYCAST_ADDR = 2, ++}; ++ ++struct inet6_fill_args { ++ u32 portid; ++ u32 seq; ++ int event; ++ unsigned int flags; ++ int netnsid; ++ int ifindex; ++ enum addr_type_t type; ++}; ++ ++enum { ++ DAD_PROCESS = 0, ++ DAD_BEGIN = 1, ++ DAD_ABORT = 2, ++}; ++ ++struct ipv6_mreq { ++ struct in6_addr ipv6mr_multiaddr; ++ int ipv6mr_ifindex; ++}; ++ ++struct group_req { ++ __u32 gr_interface; ++ struct __kernel_sockaddr_storage gr_group; ++}; ++ ++struct ip6_mtuinfo { ++ struct sockaddr_in6 ip6m_addr; ++ __u32 ip6m_mtu; ++}; ++ ++struct ip6_ra_chain { ++ struct ip6_ra_chain *next; ++ struct sock *sk; ++ int sel; ++ void (*destructor)(struct sock *); ++}; ++ ++struct compat_group_req { ++ __u32 gr_interface; ++ struct __kernel_sockaddr_storage gr_group; ++} __attribute__((packed)); ++ ++struct compat_group_source_req { ++ __u32 gsr_interface; ++ struct __kernel_sockaddr_storage gsr_group; ++ struct __kernel_sockaddr_storage gsr_source; ++} __attribute__((packed)); ++ ++struct compat_group_filter { ++ union { ++ struct { ++ __u32 gf_interface_aux; ++ struct __kernel_sockaddr_storage gf_group_aux; ++ __u32 gf_fmode_aux; ++ __u32 gf_numsrc_aux; ++ struct __kernel_sockaddr_storage gf_slist[1]; ++ } __attribute__((packed)); ++ struct { ++ __u32 gf_interface; ++ struct __kernel_sockaddr_storage gf_group; ++ __u32 gf_fmode; ++ __u32 gf_numsrc; ++ struct __kernel_sockaddr_storage gf_slist_flex[0]; ++ } __attribute__((packed)); ++ }; ++}; ++ ++struct raw_iter_state { ++ struct seq_net_private p; ++ int bucket; ++}; ++ ++typedef int mh_filter_t(struct sock *, struct sk_buff *); ++ ++struct raw6_frag_vec { ++ struct msghdr *msg; ++ int hlen; ++ char c[4]; ++}; ++ ++enum ip6_defrag_users { ++ IP6_DEFRAG_LOCAL_DELIVER = 0, ++ IP6_DEFRAG_CONNTRACK_IN = 1, ++ __IP6_DEFRAG_CONNTRACK_IN = 65536, ++ IP6_DEFRAG_CONNTRACK_OUT = 65537, ++ __IP6_DEFRAG_CONNTRACK_OUT = 131072, ++ IP6_DEFRAG_CONNTRACK_BRIDGE_IN = 131073, ++ __IP6_DEFRAG_CONNTRACK_BRIDGE_IN = 196608, ++}; ++ ++struct frag_queue { ++ struct inet_frag_queue q; ++ int iif; ++ __u16 nhoffset; ++ u8 ecn; ++}; ++ ++struct tcp6_pseudohdr { ++ struct in6_addr saddr; ++ struct in6_addr daddr; ++ __be32 len; ++ __be32 protocol; ++}; ++ ++struct in_pktinfo { ++ int ipi_ifindex; ++ struct in_addr ipi_spec_dst; ++ struct in_addr ipi_addr; ++}; ++ ++enum { ++ SEG6_ATTR_UNSPEC = 0, ++ SEG6_ATTR_DST = 1, ++ SEG6_ATTR_DSTLEN = 2, ++ SEG6_ATTR_HMACKEYID = 3, ++ SEG6_ATTR_SECRET = 4, ++ SEG6_ATTR_SECRETLEN = 5, ++ SEG6_ATTR_ALGID = 6, ++ SEG6_ATTR_HMACINFO = 7, ++ __SEG6_ATTR_MAX = 8, ++}; ++ ++enum { ++ SEG6_CMD_UNSPEC = 0, ++ SEG6_CMD_SETHMAC = 1, ++ SEG6_CMD_DUMPHMAC = 2, ++ SEG6_CMD_SET_TUNSRC = 3, ++ SEG6_CMD_GET_TUNSRC = 4, ++ __SEG6_CMD_MAX = 5, ++}; ++ ++enum { ++ IOAM6_ATTR_UNSPEC = 0, ++ IOAM6_ATTR_NS_ID = 1, ++ IOAM6_ATTR_NS_DATA = 2, ++ IOAM6_ATTR_NS_DATA_WIDE = 3, ++ IOAM6_ATTR_SC_ID = 4, ++ IOAM6_ATTR_SC_DATA = 5, ++ IOAM6_ATTR_SC_NONE = 6, ++ IOAM6_ATTR_PAD = 7, ++ __IOAM6_ATTR_MAX = 8, ++}; ++ ++enum { ++ IOAM6_CMD_UNSPEC = 0, ++ IOAM6_CMD_ADD_NAMESPACE = 1, ++ IOAM6_CMD_DEL_NAMESPACE = 2, ++ IOAM6_CMD_DUMP_NAMESPACES = 3, ++ IOAM6_CMD_ADD_SCHEMA = 4, ++ IOAM6_CMD_DEL_SCHEMA = 5, ++ IOAM6_CMD_DUMP_SCHEMAS = 6, ++ IOAM6_CMD_NS_SET_SCHEMA = 7, ++ __IOAM6_CMD_MAX = 8, ++}; ++ ++struct fib6_rule { ++ struct fib_rule common; ++ struct rt6key src; ++ struct rt6key dst; ++ dscp_t dscp; ++}; ++ ++enum l3mdev_type { ++ L3MDEV_TYPE_UNSPEC = 0, ++ L3MDEV_TYPE_VRF = 1, ++ __L3MDEV_TYPE_MAX = 2, ++}; ++ ++enum { ++ SEG6_LOCAL_UNSPEC = 0, ++ SEG6_LOCAL_ACTION = 1, ++ SEG6_LOCAL_SRH = 2, ++ SEG6_LOCAL_TABLE = 3, ++ SEG6_LOCAL_NH4 = 4, ++ SEG6_LOCAL_NH6 = 5, ++ SEG6_LOCAL_IIF = 6, ++ SEG6_LOCAL_OIF = 7, ++ SEG6_LOCAL_BPF = 8, ++ SEG6_LOCAL_VRFTABLE = 9, ++ SEG6_LOCAL_COUNTERS = 10, ++ SEG6_LOCAL_FLAVORS = 11, ++ __SEG6_LOCAL_MAX = 12, ++}; ++ ++enum { ++ SEG6_LOCAL_BPF_PROG_UNSPEC = 0, ++ SEG6_LOCAL_BPF_PROG = 1, ++ SEG6_LOCAL_BPF_PROG_NAME = 2, ++ __SEG6_LOCAL_BPF_PROG_MAX = 3, ++}; ++ ++enum { ++ SEG6_LOCAL_CNT_UNSPEC = 0, ++ SEG6_LOCAL_CNT_PAD = 1, ++ SEG6_LOCAL_CNT_PACKETS = 2, ++ SEG6_LOCAL_CNT_BYTES = 3, ++ SEG6_LOCAL_CNT_ERRORS = 4, ++ __SEG6_LOCAL_CNT_MAX = 5, ++}; ++ ++enum { ++ SEG6_LOCAL_FLV_UNSPEC = 0, ++ SEG6_LOCAL_FLV_OPERATION = 1, ++ SEG6_LOCAL_FLV_LCBLOCK_BITS = 2, ++ SEG6_LOCAL_FLV_LCNODE_FN_BITS = 3, ++ __SEG6_LOCAL_FLV_MAX = 4, ++}; ++ ++enum { ++ SEG6_LOCAL_FLV_OP_UNSPEC = 0, ++ SEG6_LOCAL_FLV_OP_PSP = 1, ++ SEG6_LOCAL_FLV_OP_USP = 2, ++ SEG6_LOCAL_FLV_OP_USD = 3, ++ SEG6_LOCAL_FLV_OP_NEXT_CSID = 4, ++ __SEG6_LOCAL_FLV_OP_MAX = 5, ++}; ++ ++struct seg6_local_lwt; ++ ++struct seg6_local_lwtunnel_ops { ++ int (*build_state)(struct seg6_local_lwt *, const void *, struct netlink_ext_ack *); ++ void (*destroy_state)(struct seg6_local_lwt *); ++}; ++ ++enum seg6_end_dt_mode { ++ DT_INVALID_MODE = -22, ++ DT_LEGACY_MODE = 0, ++ DT_VRF_MODE = 1, ++}; ++ ++struct seg6_end_dt_info { ++ enum seg6_end_dt_mode mode; ++ struct net *net; ++ int vrf_ifindex; ++ int vrf_table; ++ u16 family; ++}; ++ ++struct seg6_flavors_info { ++ __u32 flv_ops; ++ __u8 lcblock_bits; ++ __u8 lcnode_func_bits; ++}; ++ ++struct pcpu_seg6_local_counters; ++ ++struct seg6_action_desc; ++ ++struct seg6_local_lwt { ++ int action; ++ struct ipv6_sr_hdr *srh; ++ int table; ++ struct in_addr nh4; ++ struct in6_addr nh6; ++ int iif; ++ int oif; ++ struct bpf_lwt_prog bpf; ++ struct seg6_end_dt_info dt_info; ++ struct seg6_flavors_info flv_info; ++ struct pcpu_seg6_local_counters *pcpu_counters; ++ int headroom; ++ struct seg6_action_desc *desc; ++ long unsigned int parsed_optattrs; ++}; ++ ++struct seg6_action_desc { ++ int action; ++ long unsigned int attrs; ++ long unsigned int optattrs; ++ int (*input)(struct sk_buff *, struct seg6_local_lwt *); ++ int static_headroom; ++ struct seg6_local_lwtunnel_ops slwt_ops; ++}; ++ ++struct pcpu_seg6_local_counters { ++ u64_stats_t packets; ++ u64_stats_t bytes; ++ u64_stats_t errors; ++ struct u64_stats_sync syncp; ++}; ++ ++struct seg6_local_counters { ++ __u64 packets; ++ __u64 bytes; ++ __u64 errors; ++}; ++ ++enum seg6_local_pktinfo { ++ SEG6_LOCAL_PKTINFO_NOHDR = 0, ++ SEG6_LOCAL_PKTINFO_SL_ZERO = 1, ++ SEG6_LOCAL_PKTINFO_SL_ONE = 2, ++ SEG6_LOCAL_PKTINFO_SL_MORE = 3, ++ __SEG6_LOCAL_PKTINFO_MAX = 4, ++}; ++ ++enum seg6_local_flv_action { ++ SEG6_LOCAL_FLV_ACT_UNSPEC = 0, ++ SEG6_LOCAL_FLV_ACT_END = 1, ++ SEG6_LOCAL_FLV_ACT_PSP = 2, ++ SEG6_LOCAL_FLV_ACT_USP = 3, ++ SEG6_LOCAL_FLV_ACT_USD = 4, ++ __SEG6_LOCAL_FLV_ACT_MAX = 5, ++}; ++ ++struct seg6_action_param { ++ int (*parse)(struct nlattr **, struct seg6_local_lwt *, struct netlink_ext_ack *); ++ int (*put)(struct sk_buff *, struct seg6_local_lwt *); ++ int (*cmp)(struct seg6_local_lwt *, struct seg6_local_lwt *); ++ void (*destroy)(struct seg6_local_lwt *); ++}; ++ ++enum { ++ __IOAM6_IPTUNNEL_MODE_MIN = 0, ++ IOAM6_IPTUNNEL_MODE_INLINE = 1, ++ IOAM6_IPTUNNEL_MODE_ENCAP = 2, ++ IOAM6_IPTUNNEL_MODE_AUTO = 3, ++ __IOAM6_IPTUNNEL_MODE_MAX = 4, ++}; ++ ++enum { ++ IOAM6_IPTUNNEL_UNSPEC = 0, ++ IOAM6_IPTUNNEL_MODE = 1, ++ IOAM6_IPTUNNEL_DST = 2, ++ IOAM6_IPTUNNEL_TRACE = 3, ++ IOAM6_IPTUNNEL_FREQ_K = 4, ++ IOAM6_IPTUNNEL_FREQ_N = 5, ++ __IOAM6_IPTUNNEL_MAX = 6, ++}; ++ ++struct ioam6_lwt_encap { ++ struct ipv6_opt_hdr eh; ++ u8 pad[2]; ++ struct ioam6_hdr ioamh; ++ struct ioam6_trace_hdr traceh; ++}; ++ ++struct ioam6_lwt_freq { ++ u32 k; ++ u32 n; ++}; ++ ++struct ioam6_lwt { ++ struct dst_cache cache; ++ struct ioam6_lwt_freq freq; ++ atomic_t pkt_cnt; ++ u8 mode; ++ struct in6_addr tundst; ++ struct ioam6_lwt_encap tuninfo; ++}; ++ ++struct rt2_hdr { ++ struct ipv6_rt_hdr rt_hdr; ++ __u32 reserved; ++ struct in6_addr addr; ++}; ++ ++struct mip6_report_rate_limiter { ++ spinlock_t lock; ++ ktime_t stamp; ++ int iif; ++ struct in6_addr src; ++ struct in6_addr dst; ++}; ++ ++struct packet_diag_req { ++ __u8 sdiag_family; ++ __u8 sdiag_protocol; ++ __u16 pad; ++ __u32 pdiag_ino; ++ __u32 pdiag_show; ++ __u32 pdiag_cookie[2]; ++}; ++ ++struct packet_diag_msg { ++ __u8 pdiag_family; ++ __u8 pdiag_type; ++ __u16 pdiag_num; ++ __u32 pdiag_ino; ++ __u32 pdiag_cookie[2]; ++}; ++ ++enum { ++ PACKET_DIAG_INFO = 0, ++ PACKET_DIAG_MCLIST = 1, ++ PACKET_DIAG_RX_RING = 2, ++ PACKET_DIAG_TX_RING = 3, ++ PACKET_DIAG_FANOUT = 4, ++ PACKET_DIAG_UID = 5, ++ PACKET_DIAG_MEMINFO = 6, ++ PACKET_DIAG_FILTER = 7, ++ __PACKET_DIAG_MAX = 8, ++}; ++ ++struct packet_diag_info { ++ __u32 pdi_index; ++ __u32 pdi_version; ++ __u32 pdi_reserve; ++ __u32 pdi_copy_thresh; ++ __u32 pdi_tstamp; ++ __u32 pdi_flags; ++}; ++ ++struct packet_diag_mclist { ++ __u32 pdmc_index; ++ __u32 pdmc_count; ++ __u16 pdmc_type; ++ __u16 pdmc_alen; ++ __u8 pdmc_addr[32]; ++}; ++ ++struct packet_diag_ring { ++ __u32 pdr_block_size; ++ __u32 pdr_block_nr; ++ __u32 pdr_frame_size; ++ __u32 pdr_frame_nr; ++ __u32 pdr_retire_tmo; ++ __u32 pdr_sizeof_priv; ++ __u32 pdr_features; ++}; ++ ++enum devlink_eswitch_mode { ++ DEVLINK_ESWITCH_MODE_LEGACY = 0, ++ DEVLINK_ESWITCH_MODE_SWITCHDEV = 1, ++}; ++ ++enum { ++ DEVLINK_FLASH_OVERWRITE_SETTINGS_BIT = 0, ++ DEVLINK_FLASH_OVERWRITE_IDENTIFIERS_BIT = 1, ++ __DEVLINK_FLASH_OVERWRITE_MAX_BIT = 2, ++ DEVLINK_FLASH_OVERWRITE_MAX_BIT = 1, ++}; ++ ++struct devlink_nl_dump_state { ++ long unsigned int instance; ++ int idx; ++ union { ++ struct { ++ u64 start_offset; ++ }; ++ struct { ++ u64 dump_ts; ++ }; ++ }; ++}; ++ ++enum devlink_health_reporter_state { ++ DEVLINK_HEALTH_REPORTER_STATE_HEALTHY = 0, ++ DEVLINK_HEALTH_REPORTER_STATE_ERROR = 1, ++}; ++ ++struct devlink_health_reporter; ++ ++struct devlink_fmsg; ++ ++struct devlink_health_reporter_ops { ++ char *name; ++ int (*recover)(struct devlink_health_reporter *, void *, struct netlink_ext_ack *); ++ int (*dump)(struct devlink_health_reporter *, struct devlink_fmsg *, void *, struct netlink_ext_ack *); ++ int (*diagnose)(struct devlink_health_reporter *, struct devlink_fmsg *, struct netlink_ext_ack *); ++ int (*test)(struct devlink_health_reporter *, struct netlink_ext_ack *); ++}; ++ ++struct devlink_health_reporter { ++ struct list_head list; ++ void *priv; ++ const struct devlink_health_reporter_ops *ops; ++ struct devlink *devlink; ++ struct devlink_port *devlink_port; ++ struct devlink_fmsg *dump_fmsg; ++ struct mutex dump_lock; ++ u64 graceful_period; ++ bool auto_recover; ++ bool auto_dump; ++ u8 health_state; ++ u64 dump_ts; ++ u64 dump_real_ts; ++ u64 error_count; ++ u64 recovery_count; ++ u64 last_recovery_ts; ++}; ++ ++struct devlink_fmsg { ++ struct list_head item_list; ++ bool putting_binary; ++}; ++ ++struct devlink_fmsg_item { ++ struct list_head list; ++ int attrtype; ++ u8 nla_type; ++ u16 len; ++ int value[0]; ++}; ++ ++struct netlbl_domhsh_tbl { ++ struct list_head *tbl; ++ u32 size; ++}; ++ ++enum { ++ NLBL_UNLABEL_C_UNSPEC = 0, ++ NLBL_UNLABEL_C_ACCEPT = 1, ++ NLBL_UNLABEL_C_LIST = 2, ++ NLBL_UNLABEL_C_STATICADD = 3, ++ NLBL_UNLABEL_C_STATICREMOVE = 4, ++ NLBL_UNLABEL_C_STATICLIST = 5, ++ NLBL_UNLABEL_C_STATICADDDEF = 6, ++ NLBL_UNLABEL_C_STATICREMOVEDEF = 7, ++ NLBL_UNLABEL_C_STATICLISTDEF = 8, ++ __NLBL_UNLABEL_C_MAX = 9, ++}; ++ ++enum { ++ NLBL_UNLABEL_A_UNSPEC = 0, ++ NLBL_UNLABEL_A_ACPTFLG = 1, ++ NLBL_UNLABEL_A_IPV6ADDR = 2, ++ NLBL_UNLABEL_A_IPV6MASK = 3, ++ NLBL_UNLABEL_A_IPV4ADDR = 4, ++ NLBL_UNLABEL_A_IPV4MASK = 5, ++ NLBL_UNLABEL_A_IFACE = 6, ++ NLBL_UNLABEL_A_SECCTX = 7, ++ __NLBL_UNLABEL_A_MAX = 8, ++}; ++ ++struct netlbl_unlhsh_tbl { ++ struct list_head *tbl; ++ u32 size; ++}; ++ ++struct netlbl_unlhsh_addr4 { ++ u32 secid; ++ struct netlbl_af4list list; ++ struct callback_head rcu; ++}; ++ ++struct netlbl_unlhsh_addr6 { ++ u32 secid; ++ struct netlbl_af6list list; ++ struct callback_head rcu; ++}; ++ ++struct netlbl_unlhsh_iface { ++ int ifindex; ++ struct list_head addr4_list; ++ struct list_head addr6_list; ++ u32 valid; ++ struct list_head list; ++ struct callback_head rcu; ++}; ++ ++struct netlbl_unlhsh_walk_arg { ++ struct netlink_callback *nl_cb; ++ struct sk_buff *skb; ++ u32 seq; ++}; ++ ++enum { ++ NLBL_CALIPSO_C_UNSPEC = 0, ++ NLBL_CALIPSO_C_ADD = 1, ++ NLBL_CALIPSO_C_REMOVE = 2, ++ NLBL_CALIPSO_C_LIST = 3, ++ NLBL_CALIPSO_C_LISTALL = 4, ++ __NLBL_CALIPSO_C_MAX = 5, ++}; ++ ++enum { ++ NLBL_CALIPSO_A_UNSPEC = 0, ++ NLBL_CALIPSO_A_DOI = 1, ++ NLBL_CALIPSO_A_MTYPE = 2, ++ __NLBL_CALIPSO_A_MAX = 3, ++}; ++ ++struct netlbl_calipso_doiwalk_arg { ++ struct netlink_callback *nl_cb; ++ struct sk_buff *skb; ++ u32 seq; ++}; ++ ++struct netlbl_domhsh_walk_arg___2 { ++ struct netlbl_audit *audit_info; ++ u32 doi; ++}; ++ ++struct dcbmsg { ++ __u8 dcb_family; ++ __u8 cmd; ++ __u16 dcb_pad; ++}; ++ ++enum dcbnl_commands { ++ DCB_CMD_UNDEFINED = 0, ++ DCB_CMD_GSTATE = 1, ++ DCB_CMD_SSTATE = 2, ++ DCB_CMD_PGTX_GCFG = 3, ++ DCB_CMD_PGTX_SCFG = 4, ++ DCB_CMD_PGRX_GCFG = 5, ++ DCB_CMD_PGRX_SCFG = 6, ++ DCB_CMD_PFC_GCFG = 7, ++ DCB_CMD_PFC_SCFG = 8, ++ DCB_CMD_SET_ALL = 9, ++ DCB_CMD_GPERM_HWADDR = 10, ++ DCB_CMD_GCAP = 11, ++ DCB_CMD_GNUMTCS = 12, ++ DCB_CMD_SNUMTCS = 13, ++ DCB_CMD_PFC_GSTATE = 14, ++ DCB_CMD_PFC_SSTATE = 15, ++ DCB_CMD_BCN_GCFG = 16, ++ DCB_CMD_BCN_SCFG = 17, ++ DCB_CMD_GAPP = 18, ++ DCB_CMD_SAPP = 19, ++ DCB_CMD_IEEE_SET = 20, ++ DCB_CMD_IEEE_GET = 21, ++ DCB_CMD_GDCBX = 22, ++ DCB_CMD_SDCBX = 23, ++ DCB_CMD_GFEATCFG = 24, ++ DCB_CMD_SFEATCFG = 25, ++ DCB_CMD_CEE_GET = 26, ++ DCB_CMD_IEEE_DEL = 27, ++ __DCB_CMD_ENUM_MAX = 28, ++ DCB_CMD_MAX = 27, ++}; ++ ++enum dcbnl_attrs { ++ DCB_ATTR_UNDEFINED = 0, ++ DCB_ATTR_IFNAME = 1, ++ DCB_ATTR_STATE = 2, ++ DCB_ATTR_PFC_STATE = 3, ++ DCB_ATTR_PFC_CFG = 4, ++ DCB_ATTR_NUM_TC = 5, ++ DCB_ATTR_PG_CFG = 6, ++ DCB_ATTR_SET_ALL = 7, ++ DCB_ATTR_PERM_HWADDR = 8, ++ DCB_ATTR_CAP = 9, ++ DCB_ATTR_NUMTCS = 10, ++ DCB_ATTR_BCN = 11, ++ DCB_ATTR_APP = 12, ++ DCB_ATTR_IEEE = 13, ++ DCB_ATTR_DCBX = 14, ++ DCB_ATTR_FEATCFG = 15, ++ DCB_ATTR_CEE = 16, ++ __DCB_ATTR_ENUM_MAX = 17, ++ DCB_ATTR_MAX = 16, ++}; ++ ++enum ieee_attrs { ++ DCB_ATTR_IEEE_UNSPEC = 0, ++ DCB_ATTR_IEEE_ETS = 1, ++ DCB_ATTR_IEEE_PFC = 2, ++ DCB_ATTR_IEEE_APP_TABLE = 3, ++ DCB_ATTR_IEEE_PEER_ETS = 4, ++ DCB_ATTR_IEEE_PEER_PFC = 5, ++ DCB_ATTR_IEEE_PEER_APP = 6, ++ DCB_ATTR_IEEE_MAXRATE = 7, ++ DCB_ATTR_IEEE_QCN = 8, ++ DCB_ATTR_IEEE_QCN_STATS = 9, ++ DCB_ATTR_DCB_BUFFER = 10, ++ DCB_ATTR_DCB_APP_TRUST_TABLE = 11, ++ DCB_ATTR_DCB_REWR_TABLE = 12, ++ __DCB_ATTR_IEEE_MAX = 13, ++}; ++ ++enum ieee_attrs_app { ++ DCB_ATTR_IEEE_APP_UNSPEC = 0, ++ DCB_ATTR_IEEE_APP = 1, ++ DCB_ATTR_DCB_APP = 2, ++ __DCB_ATTR_IEEE_APP_MAX = 3, ++}; ++ ++enum cee_attrs { ++ DCB_ATTR_CEE_UNSPEC = 0, ++ DCB_ATTR_CEE_PEER_PG = 1, ++ DCB_ATTR_CEE_PEER_PFC = 2, ++ DCB_ATTR_CEE_PEER_APP_TABLE = 3, ++ DCB_ATTR_CEE_TX_PG = 4, ++ DCB_ATTR_CEE_RX_PG = 5, ++ DCB_ATTR_CEE_PFC = 6, ++ DCB_ATTR_CEE_APP_TABLE = 7, ++ DCB_ATTR_CEE_FEAT = 8, ++ __DCB_ATTR_CEE_MAX = 9, ++}; ++ ++enum peer_app_attr { ++ DCB_ATTR_CEE_PEER_APP_UNSPEC = 0, ++ DCB_ATTR_CEE_PEER_APP_INFO = 1, ++ DCB_ATTR_CEE_PEER_APP = 2, ++ __DCB_ATTR_CEE_PEER_APP_MAX = 3, ++}; ++ ++enum dcbnl_pfc_up_attrs { ++ DCB_PFC_UP_ATTR_UNDEFINED = 0, ++ DCB_PFC_UP_ATTR_0 = 1, ++ DCB_PFC_UP_ATTR_1 = 2, ++ DCB_PFC_UP_ATTR_2 = 3, ++ DCB_PFC_UP_ATTR_3 = 4, ++ DCB_PFC_UP_ATTR_4 = 5, ++ DCB_PFC_UP_ATTR_5 = 6, ++ DCB_PFC_UP_ATTR_6 = 7, ++ DCB_PFC_UP_ATTR_7 = 8, ++ DCB_PFC_UP_ATTR_ALL = 9, ++ __DCB_PFC_UP_ATTR_ENUM_MAX = 10, ++ DCB_PFC_UP_ATTR_MAX = 9, ++}; ++ ++enum dcbnl_pg_attrs { ++ DCB_PG_ATTR_UNDEFINED = 0, ++ DCB_PG_ATTR_TC_0 = 1, ++ DCB_PG_ATTR_TC_1 = 2, ++ DCB_PG_ATTR_TC_2 = 3, ++ DCB_PG_ATTR_TC_3 = 4, ++ DCB_PG_ATTR_TC_4 = 5, ++ DCB_PG_ATTR_TC_5 = 6, ++ DCB_PG_ATTR_TC_6 = 7, ++ DCB_PG_ATTR_TC_7 = 8, ++ DCB_PG_ATTR_TC_MAX = 9, ++ DCB_PG_ATTR_TC_ALL = 10, ++ DCB_PG_ATTR_BW_ID_0 = 11, ++ DCB_PG_ATTR_BW_ID_1 = 12, ++ DCB_PG_ATTR_BW_ID_2 = 13, ++ DCB_PG_ATTR_BW_ID_3 = 14, ++ DCB_PG_ATTR_BW_ID_4 = 15, ++ DCB_PG_ATTR_BW_ID_5 = 16, ++ DCB_PG_ATTR_BW_ID_6 = 17, ++ DCB_PG_ATTR_BW_ID_7 = 18, ++ DCB_PG_ATTR_BW_ID_MAX = 19, ++ DCB_PG_ATTR_BW_ID_ALL = 20, ++ __DCB_PG_ATTR_ENUM_MAX = 21, ++ DCB_PG_ATTR_MAX = 20, ++}; ++ ++enum dcbnl_tc_attrs { ++ DCB_TC_ATTR_PARAM_UNDEFINED = 0, ++ DCB_TC_ATTR_PARAM_PGID = 1, ++ DCB_TC_ATTR_PARAM_UP_MAPPING = 2, ++ DCB_TC_ATTR_PARAM_STRICT_PRIO = 3, ++ DCB_TC_ATTR_PARAM_BW_PCT = 4, ++ DCB_TC_ATTR_PARAM_ALL = 5, ++ __DCB_TC_ATTR_PARAM_ENUM_MAX = 6, ++ DCB_TC_ATTR_PARAM_MAX = 5, ++}; ++ ++enum dcbnl_cap_attrs { ++ DCB_CAP_ATTR_UNDEFINED = 0, ++ DCB_CAP_ATTR_ALL = 1, ++ DCB_CAP_ATTR_PG = 2, ++ DCB_CAP_ATTR_PFC = 3, ++ DCB_CAP_ATTR_UP2TC = 4, ++ DCB_CAP_ATTR_PG_TCS = 5, ++ DCB_CAP_ATTR_PFC_TCS = 6, ++ DCB_CAP_ATTR_GSP = 7, ++ DCB_CAP_ATTR_BCN = 8, ++ DCB_CAP_ATTR_DCBX = 9, ++ __DCB_CAP_ATTR_ENUM_MAX = 10, ++ DCB_CAP_ATTR_MAX = 9, ++}; ++ ++enum dcbnl_numtcs_attrs { ++ DCB_NUMTCS_ATTR_UNDEFINED = 0, ++ DCB_NUMTCS_ATTR_ALL = 1, ++ DCB_NUMTCS_ATTR_PG = 2, ++ DCB_NUMTCS_ATTR_PFC = 3, ++ __DCB_NUMTCS_ATTR_ENUM_MAX = 4, ++ DCB_NUMTCS_ATTR_MAX = 3, ++}; ++ ++enum dcbnl_bcn_attrs { ++ DCB_BCN_ATTR_UNDEFINED = 0, ++ DCB_BCN_ATTR_RP_0 = 1, ++ DCB_BCN_ATTR_RP_1 = 2, ++ DCB_BCN_ATTR_RP_2 = 3, ++ DCB_BCN_ATTR_RP_3 = 4, ++ DCB_BCN_ATTR_RP_4 = 5, ++ DCB_BCN_ATTR_RP_5 = 6, ++ DCB_BCN_ATTR_RP_6 = 7, ++ DCB_BCN_ATTR_RP_7 = 8, ++ DCB_BCN_ATTR_RP_ALL = 9, ++ DCB_BCN_ATTR_BCNA_0 = 10, ++ DCB_BCN_ATTR_BCNA_1 = 11, ++ DCB_BCN_ATTR_ALPHA = 12, ++ DCB_BCN_ATTR_BETA = 13, ++ DCB_BCN_ATTR_GD = 14, ++ DCB_BCN_ATTR_GI = 15, ++ DCB_BCN_ATTR_TMAX = 16, ++ DCB_BCN_ATTR_TD = 17, ++ DCB_BCN_ATTR_RMIN = 18, ++ DCB_BCN_ATTR_W = 19, ++ DCB_BCN_ATTR_RD = 20, ++ DCB_BCN_ATTR_RU = 21, ++ DCB_BCN_ATTR_WRTT = 22, ++ DCB_BCN_ATTR_RI = 23, ++ DCB_BCN_ATTR_C = 24, ++ DCB_BCN_ATTR_ALL = 25, ++ __DCB_BCN_ATTR_ENUM_MAX = 26, ++ DCB_BCN_ATTR_MAX = 25, ++}; ++ ++enum dcb_general_attr_values { ++ DCB_ATTR_VALUE_UNDEFINED = 255, ++}; ++ ++enum dcbnl_app_attrs { ++ DCB_APP_ATTR_UNDEFINED = 0, ++ DCB_APP_ATTR_IDTYPE = 1, ++ DCB_APP_ATTR_ID = 2, ++ DCB_APP_ATTR_PRIORITY = 3, ++ __DCB_APP_ATTR_ENUM_MAX = 4, ++ DCB_APP_ATTR_MAX = 3, ++}; ++ ++enum dcbnl_featcfg_attrs { ++ DCB_FEATCFG_ATTR_UNDEFINED = 0, ++ DCB_FEATCFG_ATTR_ALL = 1, ++ DCB_FEATCFG_ATTR_PG = 2, ++ DCB_FEATCFG_ATTR_PFC = 3, ++ DCB_FEATCFG_ATTR_APP = 4, ++ __DCB_FEATCFG_ATTR_ENUM_MAX = 5, ++ DCB_FEATCFG_ATTR_MAX = 4, ++}; ++ ++struct dcb_app_type { ++ int ifindex; ++ struct dcb_app app; ++ struct list_head list; ++ u8 dcbx; ++}; ++ ++struct dcb_rewr_prio_pcp_map { ++ u16 map[8]; ++}; ++ ++struct dcb_ieee_app_prio_map { ++ u64 map[8]; ++}; ++ ++struct dcb_ieee_app_dscp_map { ++ u8 map[64]; ++}; ++ ++enum dcbevent_notif_type { ++ DCB_APP_EVENT = 1, ++}; ++ ++struct reply_func { ++ int type; ++ int (*cb)(struct net_device *, struct nlmsghdr *, u32, struct nlattr **, struct sk_buff *); ++}; ++ ++typedef int (*lookup_by_table_id_t)(struct net *, u32); ++ ++struct l3mdev_handler { ++ lookup_by_table_id_t dev_lookup; ++}; ++ ++enum { ++ NCSI_CAP_GENERIC_HWA = 1, ++ NCSI_CAP_GENERIC_HDS = 2, ++ NCSI_CAP_GENERIC_FC = 4, ++ NCSI_CAP_GENERIC_FC1 = 8, ++ NCSI_CAP_GENERIC_MC = 16, ++ NCSI_CAP_GENERIC_HWA_UNKNOWN = 0, ++ NCSI_CAP_GENERIC_HWA_SUPPORT = 32, ++ NCSI_CAP_GENERIC_HWA_NOT_SUPPORT = 64, ++ NCSI_CAP_GENERIC_HWA_RESERVED = 96, ++ NCSI_CAP_GENERIC_HWA_MASK = 96, ++ NCSI_CAP_GENERIC_MASK = 127, ++ NCSI_CAP_BC_ARP = 1, ++ NCSI_CAP_BC_DHCPC = 2, ++ NCSI_CAP_BC_DHCPS = 4, ++ NCSI_CAP_BC_NETBIOS = 8, ++ NCSI_CAP_BC_MASK = 15, ++ NCSI_CAP_MC_IPV6_NEIGHBOR = 1, ++ NCSI_CAP_MC_IPV6_ROUTER = 2, ++ NCSI_CAP_MC_DHCPV6_RELAY = 4, ++ NCSI_CAP_MC_DHCPV6_WELL_KNOWN = 8, ++ NCSI_CAP_MC_IPV6_MLD = 16, ++ NCSI_CAP_MC_IPV6_NEIGHBOR_S = 32, ++ NCSI_CAP_MC_MASK = 63, ++ NCSI_CAP_AEN_LSC = 1, ++ NCSI_CAP_AEN_CR = 2, ++ NCSI_CAP_AEN_HDS = 4, ++ NCSI_CAP_AEN_MASK = 7, ++ NCSI_CAP_VLAN_ONLY = 1, ++ NCSI_CAP_VLAN_NO = 2, ++ NCSI_CAP_VLAN_ANY = 4, ++ NCSI_CAP_VLAN_MASK = 7, ++}; ++ ++struct ncsi_cmd_pkt_hdr { ++ struct ncsi_pkt_hdr common; ++}; ++ ++struct ncsi_rsp_pkt_hdr { ++ struct ncsi_pkt_hdr common; ++ __be16 code; ++ __be16 reason; ++}; ++ ++struct ncsi_rsp_pkt { ++ struct ncsi_rsp_pkt_hdr rsp; ++ __be32 checksum; ++ unsigned char pad[22]; ++}; ++ ++struct ncsi_cmd_ae_pkt { ++ struct ncsi_cmd_pkt_hdr cmd; ++ unsigned char reserved[3]; ++ unsigned char mc_id; ++ __be32 mode; ++ __be32 checksum; ++ unsigned char pad[18]; ++}; ++ ++struct ncsi_cmd_sl_pkt { ++ struct ncsi_cmd_pkt_hdr cmd; ++ __be32 mode; ++ __be32 oem_mode; ++ __be32 checksum; ++ unsigned char pad[18]; ++}; ++ ++struct ncsi_cmd_svf_pkt { ++ struct ncsi_cmd_pkt_hdr cmd; ++ __be16 reserved; ++ __be16 vlan; ++ __be16 reserved1; ++ unsigned char index; ++ unsigned char enable; ++ __be32 checksum; ++ unsigned char pad[18]; ++}; ++ ++struct ncsi_cmd_ev_pkt { ++ struct ncsi_cmd_pkt_hdr cmd; ++ unsigned char reserved[3]; ++ unsigned char mode; ++ __be32 checksum; ++ unsigned char pad[22]; ++}; ++ ++struct ncsi_cmd_sma_pkt { ++ struct ncsi_cmd_pkt_hdr cmd; ++ unsigned char mac[6]; ++ unsigned char index; ++ unsigned char at_e; ++ __be32 checksum; ++ unsigned char pad[18]; ++}; ++ ++struct ncsi_cmd_ebf_pkt { ++ struct ncsi_cmd_pkt_hdr cmd; ++ __be32 mode; ++ __be32 checksum; ++ unsigned char pad[22]; ++}; ++ ++struct ncsi_cmd_egmf_pkt { ++ struct ncsi_cmd_pkt_hdr cmd; ++ __be32 mode; ++ __be32 checksum; ++ unsigned char pad[22]; ++}; ++ ++struct ncsi_cmd_snfc_pkt { ++ struct ncsi_cmd_pkt_hdr cmd; ++ unsigned char reserved[3]; ++ unsigned char mode; ++ __be32 checksum; ++ unsigned char pad[22]; ++}; ++ ++struct ncsi_rsp_oem_pkt { ++ struct ncsi_rsp_pkt_hdr rsp; ++ __be32 mfr_id; ++ unsigned char data[0]; ++}; ++ ++struct ncsi_rsp_oem_mlx_pkt { ++ unsigned char cmd_rev; ++ unsigned char cmd; ++ unsigned char param; ++ unsigned char optional; ++ unsigned char data[0]; ++}; ++ ++struct ncsi_rsp_oem_bcm_pkt { ++ unsigned char ver; ++ unsigned char type; ++ __be16 len; ++ unsigned char data[0]; ++}; ++ ++struct ncsi_rsp_oem_intel_pkt { ++ unsigned char cmd; ++ unsigned char data[0]; ++}; ++ ++struct ncsi_rsp_gls_pkt { ++ struct ncsi_rsp_pkt_hdr rsp; ++ __be32 status; ++ __be32 other; ++ __be32 oem_status; ++ __be32 checksum; ++ unsigned char pad[10]; ++}; ++ ++struct ncsi_rsp_gvi_pkt { ++ struct ncsi_rsp_pkt_hdr rsp; ++ __be32 ncsi_version; ++ unsigned char reserved[3]; ++ unsigned char alpha2; ++ unsigned char fw_name[12]; ++ __be32 fw_version; ++ __be16 pci_ids[4]; ++ __be32 mf_id; ++ __be32 checksum; ++}; ++ ++struct ncsi_rsp_gc_pkt { ++ struct ncsi_rsp_pkt_hdr rsp; ++ __be32 cap; ++ __be32 bc_cap; ++ __be32 mc_cap; ++ __be32 buf_cap; ++ __be32 aen_cap; ++ unsigned char vlan_cnt; ++ unsigned char mixed_cnt; ++ unsigned char mc_cnt; ++ unsigned char uc_cnt; ++ unsigned char reserved[2]; ++ unsigned char vlan_mode; ++ unsigned char channel_cnt; ++ __be32 checksum; ++}; ++ ++struct ncsi_rsp_gp_pkt { ++ struct ncsi_rsp_pkt_hdr rsp; ++ unsigned char mac_cnt; ++ unsigned char reserved[2]; ++ unsigned char mac_enable; ++ unsigned char vlan_cnt; ++ unsigned char reserved1; ++ __be16 vlan_enable; ++ __be32 link_mode; ++ __be32 bc_mode; ++ __be32 valid_modes; ++ unsigned char vlan_mode; ++ unsigned char fc_mode; ++ unsigned char reserved2[2]; ++ __be32 aen_mode; ++ unsigned char mac[6]; ++ __be16 vlan; ++ __be32 checksum; ++}; ++ ++struct ncsi_rsp_gcps_pkt { ++ struct ncsi_rsp_pkt_hdr rsp; ++ __be32 cnt_hi; ++ __be32 cnt_lo; ++ __be32 rx_bytes; ++ __be32 tx_bytes; ++ __be32 rx_uc_pkts; ++ __be32 rx_mc_pkts; ++ __be32 rx_bc_pkts; ++ __be32 tx_uc_pkts; ++ __be32 tx_mc_pkts; ++ __be32 tx_bc_pkts; ++ __be32 fcs_err; ++ __be32 align_err; ++ __be32 false_carrier; ++ __be32 runt_pkts; ++ __be32 jabber_pkts; ++ __be32 rx_pause_xon; ++ __be32 rx_pause_xoff; ++ __be32 tx_pause_xon; ++ __be32 tx_pause_xoff; ++ __be32 tx_s_collision; ++ __be32 tx_m_collision; ++ __be32 l_collision; ++ __be32 e_collision; ++ __be32 rx_ctl_frames; ++ __be32 rx_64_frames; ++ __be32 rx_127_frames; ++ __be32 rx_255_frames; ++ __be32 rx_511_frames; ++ __be32 rx_1023_frames; ++ __be32 rx_1522_frames; ++ __be32 rx_9022_frames; ++ __be32 tx_64_frames; ++ __be32 tx_127_frames; ++ __be32 tx_255_frames; ++ __be32 tx_511_frames; ++ __be32 tx_1023_frames; ++ __be32 tx_1522_frames; ++ __be32 tx_9022_frames; ++ __be32 rx_valid_bytes; ++ __be32 rx_runt_pkts; ++ __be32 rx_jabber_pkts; ++ __be32 checksum; ++}; ++ ++struct ncsi_rsp_gns_pkt { ++ struct ncsi_rsp_pkt_hdr rsp; ++ __be32 rx_cmds; ++ __be32 dropped_cmds; ++ __be32 cmd_type_errs; ++ __be32 cmd_csum_errs; ++ __be32 rx_pkts; ++ __be32 tx_pkts; ++ __be32 tx_aen_pkts; ++ __be32 checksum; ++}; ++ ++struct ncsi_rsp_gnpts_pkt { ++ struct ncsi_rsp_pkt_hdr rsp; ++ __be32 tx_pkts; ++ __be32 tx_dropped; ++ __be32 tx_channel_err; ++ __be32 tx_us_err; ++ __be32 rx_pkts; ++ __be32 rx_dropped; ++ __be32 rx_channel_err; ++ __be32 rx_us_err; ++ __be32 rx_os_err; ++ __be32 checksum; ++}; ++ ++struct ncsi_rsp_gps_pkt { ++ struct ncsi_rsp_pkt_hdr rsp; ++ __be32 status; ++ __be32 checksum; ++}; ++ ++struct ncsi_rsp_gpuuid_pkt { ++ struct ncsi_rsp_pkt_hdr rsp; ++ unsigned char uuid[16]; ++ __be32 checksum; ++}; ++ ++struct ncsi_rsp_oem_handler { ++ unsigned int mfr_id; ++ int (*handler)(struct ncsi_request *); ++}; ++ ++struct ncsi_rsp_handler { ++ unsigned char type; ++ int payload; ++ int (*handler)(struct ncsi_request *); ++}; ++ ++enum ncsi_nl_commands { ++ NCSI_CMD_UNSPEC = 0, ++ NCSI_CMD_PKG_INFO = 1, ++ NCSI_CMD_SET_INTERFACE = 2, ++ NCSI_CMD_CLEAR_INTERFACE = 3, ++ NCSI_CMD_SEND_CMD = 4, ++ NCSI_CMD_SET_PACKAGE_MASK = 5, ++ NCSI_CMD_SET_CHANNEL_MASK = 6, ++ __NCSI_CMD_AFTER_LAST = 7, ++ NCSI_CMD_MAX = 6, ++}; ++ ++enum ncsi_nl_attrs { ++ NCSI_ATTR_UNSPEC = 0, ++ NCSI_ATTR_IFINDEX = 1, ++ NCSI_ATTR_PACKAGE_LIST = 2, ++ NCSI_ATTR_PACKAGE_ID = 3, ++ NCSI_ATTR_CHANNEL_ID = 4, ++ NCSI_ATTR_DATA = 5, ++ NCSI_ATTR_MULTI_FLAG = 6, ++ NCSI_ATTR_PACKAGE_MASK = 7, ++ NCSI_ATTR_CHANNEL_MASK = 8, ++ __NCSI_ATTR_AFTER_LAST = 9, ++ NCSI_ATTR_MAX = 8, ++}; ++ ++enum ncsi_nl_pkg_attrs { ++ NCSI_PKG_ATTR_UNSPEC = 0, ++ NCSI_PKG_ATTR = 1, ++ NCSI_PKG_ATTR_ID = 2, ++ NCSI_PKG_ATTR_FORCED = 3, ++ NCSI_PKG_ATTR_CHANNEL_LIST = 4, ++ __NCSI_PKG_ATTR_AFTER_LAST = 5, ++ NCSI_PKG_ATTR_MAX = 4, ++}; ++ ++enum ncsi_nl_channel_attrs { ++ NCSI_CHANNEL_ATTR_UNSPEC = 0, ++ NCSI_CHANNEL_ATTR = 1, ++ NCSI_CHANNEL_ATTR_ID = 2, ++ NCSI_CHANNEL_ATTR_VERSION_MAJOR = 3, ++ NCSI_CHANNEL_ATTR_VERSION_MINOR = 4, ++ NCSI_CHANNEL_ATTR_VERSION_STR = 5, ++ NCSI_CHANNEL_ATTR_LINK_STATE = 6, ++ NCSI_CHANNEL_ATTR_ACTIVE = 7, ++ NCSI_CHANNEL_ATTR_FORCED = 8, ++ NCSI_CHANNEL_ATTR_VLAN_LIST = 9, ++ NCSI_CHANNEL_ATTR_VLAN_ID = 10, ++ __NCSI_CHANNEL_ATTR_AFTER_LAST = 11, ++ NCSI_CHANNEL_ATTR_MAX = 10, ++}; ++ ++struct ncsi_cmd_arg { ++ struct ncsi_dev_priv *ndp; ++ unsigned char type; ++ unsigned char id; ++ unsigned char package; ++ unsigned char channel; ++ short unsigned int payload; ++ unsigned int req_flags; ++ union { ++ unsigned char bytes[16]; ++ short unsigned int words[8]; ++ unsigned int dwords[4]; ++ }; ++ unsigned char *data; ++ struct genl_info *info; ++}; ++ ++struct xsk_dma_map { ++ dma_addr_t *dma_pages; ++ struct device *dev; ++ struct net_device *netdev; ++ refcount_t users; ++ struct list_head list; ++ u32 dma_pages_cnt; ++ bool dma_need_sync; ++}; ++ ++enum mptcp_addr_signal_status { ++ MPTCP_ADD_ADDR_SIGNAL = 0, ++ MPTCP_ADD_ADDR_ECHO = 1, ++ MPTCP_RM_ADDR_SIGNAL = 2, ++}; ++ ++struct csum_pseudo_header { ++ __be64 data_seq; ++ __be32 subflow_seq; ++ __be16 data_len; ++ __sum16 csum; ++}; ++ ++struct token_bucket { ++ spinlock_t lock; ++ int chain_len; ++ struct hlist_nulls_head req_chain; ++ struct hlist_nulls_head msk_chain; ++}; ++ ++enum mptcp_pm_status { ++ MPTCP_PM_ADD_ADDR_RECEIVED = 0, ++ MPTCP_PM_ADD_ADDR_SEND_ACK = 1, ++ MPTCP_PM_RM_ADDR_RECEIVED = 2, ++ MPTCP_PM_ESTABLISHED = 3, ++ MPTCP_PM_SUBFLOW_ESTABLISHED = 4, ++ MPTCP_PM_ALREADY_ESTABLISHED = 5, ++ MPTCP_PM_MPC_ENDPOINT_ACCOUNTED = 6, ++}; ++ ++enum { ++ MPTCP_PM_ATTR_UNSPEC = 0, ++ MPTCP_PM_ATTR_ADDR = 1, ++ MPTCP_PM_ATTR_RCV_ADD_ADDRS = 2, ++ MPTCP_PM_ATTR_SUBFLOWS = 3, ++ MPTCP_PM_ATTR_TOKEN = 4, ++ MPTCP_PM_ATTR_LOC_ID = 5, ++ MPTCP_PM_ATTR_ADDR_REMOTE = 6, ++ __MPTCP_PM_ATTR_MAX = 7, ++}; ++ ++enum { ++ MPTCP_PM_ADDR_ATTR_UNSPEC = 0, ++ MPTCP_PM_ADDR_ATTR_FAMILY = 1, ++ MPTCP_PM_ADDR_ATTR_ID = 2, ++ MPTCP_PM_ADDR_ATTR_ADDR4 = 3, ++ MPTCP_PM_ADDR_ATTR_ADDR6 = 4, ++ MPTCP_PM_ADDR_ATTR_PORT = 5, ++ MPTCP_PM_ADDR_ATTR_FLAGS = 6, ++ MPTCP_PM_ADDR_ATTR_IF_IDX = 7, ++ __MPTCP_PM_ADDR_ATTR_MAX = 8, ++}; ++ ++enum { ++ MPTCP_PM_CMD_UNSPEC = 0, ++ MPTCP_PM_CMD_ADD_ADDR = 1, ++ MPTCP_PM_CMD_DEL_ADDR = 2, ++ MPTCP_PM_CMD_GET_ADDR = 3, ++ MPTCP_PM_CMD_FLUSH_ADDRS = 4, ++ MPTCP_PM_CMD_SET_LIMITS = 5, ++ MPTCP_PM_CMD_GET_LIMITS = 6, ++ MPTCP_PM_CMD_SET_FLAGS = 7, ++ MPTCP_PM_CMD_ANNOUNCE = 8, ++ MPTCP_PM_CMD_REMOVE = 9, ++ MPTCP_PM_CMD_SUBFLOW_CREATE = 10, ++ MPTCP_PM_CMD_SUBFLOW_DESTROY = 11, ++ __MPTCP_PM_CMD_AFTER_LAST = 12, ++}; ++ ++enum mptcp_event_attr { ++ MPTCP_ATTR_UNSPEC = 0, ++ MPTCP_ATTR_TOKEN = 1, ++ MPTCP_ATTR_FAMILY = 2, ++ MPTCP_ATTR_LOC_ID = 3, ++ MPTCP_ATTR_REM_ID = 4, ++ MPTCP_ATTR_SADDR4 = 5, ++ MPTCP_ATTR_SADDR6 = 6, ++ MPTCP_ATTR_DADDR4 = 7, ++ MPTCP_ATTR_DADDR6 = 8, ++ MPTCP_ATTR_SPORT = 9, ++ MPTCP_ATTR_DPORT = 10, ++ MPTCP_ATTR_BACKUP = 11, ++ MPTCP_ATTR_ERROR = 12, ++ MPTCP_ATTR_FLAGS = 13, ++ MPTCP_ATTR_TIMEOUT = 14, ++ MPTCP_ATTR_IF_IDX = 15, ++ MPTCP_ATTR_RESET_REASON = 16, ++ MPTCP_ATTR_RESET_FLAGS = 17, ++ MPTCP_ATTR_SERVER_SIDE = 18, ++ __MPTCP_ATTR_AFTER_LAST = 19, ++}; ++ ++struct mptcp_pm_addr_entry { ++ struct list_head list; ++ struct mptcp_addr_info addr; ++ u8 flags; ++ int ifindex; ++ struct socket *lsk; ++}; ++ ++struct mptcp_pm_add_entry { ++ struct list_head list; ++ struct mptcp_addr_info addr; ++ struct timer_list add_timer; ++ struct mptcp_sock *sock; ++ u8 retrans_times; ++}; ++ ++struct pm_nl_pernet { ++ spinlock_t lock; ++ struct list_head local_addr_list; ++ unsigned int addrs; ++ unsigned int stale_loss_cnt; ++ unsigned int add_addr_signal_max; ++ unsigned int add_addr_accept_max; ++ unsigned int local_addr_max; ++ unsigned int subflows_max; ++ unsigned int next_id; ++ long unsigned int id_bitmap[4]; ++}; ++ ++enum { ++ HANDSHAKE_CMD_READY = 1, ++ HANDSHAKE_CMD_ACCEPT = 2, ++ HANDSHAKE_CMD_DONE = 3, ++ __HANDSHAKE_CMD_MAX = 4, ++ HANDSHAKE_CMD_MAX = 3, ++}; ++ ++struct handshake_net { ++ spinlock_t hn_lock; ++ int hn_pending; ++ int hn_pending_max; ++ struct list_head hn_requests; ++ long unsigned int hn_flags; ++}; ++ ++enum hn_flags_bits { ++ HANDSHAKE_F_NET_DRAINING = 0, ++}; ++ ++struct trace_event_raw_handshake_event_class { ++ struct trace_entry ent; ++ const void *req; ++ const void *sk; ++ unsigned int netns_ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_handshake_fd_class { ++ struct trace_entry ent; ++ const void *req; ++ const void *sk; ++ int fd; ++ unsigned int netns_ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_handshake_error_class { ++ struct trace_entry ent; ++ const void *req; ++ const void *sk; ++ int err; ++ unsigned int netns_ino; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_handshake_complete { ++ struct trace_entry ent; ++ const void *req; ++ const void *sk; ++ int status; ++ unsigned int netns_ino; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_handshake_event_class {}; ++ ++struct trace_event_data_offsets_handshake_fd_class {}; ++ ++struct trace_event_data_offsets_handshake_error_class {}; ++ ++struct trace_event_data_offsets_handshake_complete {}; ++ ++typedef void (*btf_trace_handshake_submit)(void *, const struct net *, const struct handshake_req *, const struct sock *); ++ ++typedef void (*btf_trace_handshake_submit_err)(void *, const struct net *, const struct handshake_req *, const struct sock *, int); ++ ++typedef void (*btf_trace_handshake_cancel)(void *, const struct net *, const struct handshake_req *, const struct sock *); ++ ++typedef void (*btf_trace_handshake_cancel_none)(void *, const struct net *, const struct handshake_req *, const struct sock *); ++ ++typedef void (*btf_trace_handshake_cancel_busy)(void *, const struct net *, const struct handshake_req *, const struct sock *); ++ ++typedef void (*btf_trace_handshake_destruct)(void *, const struct net *, const struct handshake_req *, const struct sock *); ++ ++typedef void (*btf_trace_handshake_complete)(void *, const struct net *, const struct handshake_req *, const struct sock *, int); ++ ++typedef void (*btf_trace_handshake_notify_err)(void *, const struct net *, const struct handshake_req *, const struct sock *, int); ++ ++typedef void (*btf_trace_handshake_cmd_accept)(void *, const struct net *, const struct handshake_req *, const struct sock *, int); ++ ++typedef void (*btf_trace_handshake_cmd_accept_err)(void *, const struct net *, const struct handshake_req *, const struct sock *, int); ++ ++typedef void (*btf_trace_handshake_cmd_done)(void *, const struct net *, const struct handshake_req *, const struct sock *, int); ++ ++typedef void (*btf_trace_handshake_cmd_done_err)(void *, const struct net *, const struct handshake_req *, const struct sock *, int); ++ ++struct node_descriptor { ++ union { ++ struct { ++ u32 validity: 3; ++ u32 reserved: 5; ++ } __attribute__((packed)); ++ u8 byte0; ++ }; ++ u32 params: 24; ++ char type[6]; ++ char model[3]; ++ char manufacturer[3]; ++ char plant[2]; ++ char seq[12]; ++ u16 tag; ++}; ++ ++struct chsc_header { ++ __u16 length; ++ __u16 code; ++}; ++ ++struct chsc_pnso_naid_l2 { ++ u64 nit; ++ struct { ++ u8 mac[6]; ++ u16 lnid; ++ } addr_lnid; ++}; ++ ++struct chsc_pnso_resume_token { ++ u64 t1; ++ u64 t2; ++}; ++ ++struct chsc_pnso_naihdr { ++ struct chsc_pnso_resume_token resume_token; ++ int: 32; ++ u32 instance; ++ int: 24; ++ u8 naids; ++ u32 reserved[3]; ++}; ++ ++struct chsc_pnso_area { ++ struct chsc_header request; ++ char: 2; ++ u8 m: 1; ++ char: 5; ++ char: 2; ++ u8 ssid: 2; ++ u8 fmt: 4; ++ u16 sch; ++ char: 8; ++ u8 cssid; ++ int: 0; ++ u8 oc; ++ struct chsc_pnso_resume_token resume_token; ++ u32 n: 1; ++ u32 reserved[3]; ++ struct chsc_header response; ++ struct chsc_pnso_naihdr naihdr; ++ struct chsc_pnso_naid_l2 entries[0]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct crw { ++ __u32 res1: 1; ++ __u32 slct: 1; ++ __u32 oflw: 1; ++ __u32 chn: 1; ++ __u32 rsc: 4; ++ __u32 anc: 1; ++ __u32 res2: 1; ++ __u32 erc: 6; ++ __u32 rsid: 16; ++}; ++ ++typedef void (*crw_handler_t)(struct crw *, struct crw *, int); ++ ++struct qdio_ssqd_desc { ++ u8 flags; ++ u16 sch; ++ u8 qfmt; ++ u8 parm; ++ u8 qdioac1; ++ u8 sch_class; ++ u8 pcnt; ++ u8 icnt; ++ char: 8; ++ u8 ocnt; ++ char: 8; ++ u8 mbccnt; ++ u16 qdioac2; ++ u64 sch_token; ++ u8 mro; ++ u8 mri; ++ u16 qdioac3; ++ int: 24; ++ u8 mmwc; ++}; ++ ++struct css_chsc_char { ++ u64 res; ++ int: 20; ++ u32 secm: 1; ++ char: 1; ++ u32 scmc: 1; ++ int: 9; ++ short: 11; ++ u32 scssc: 1; ++ u32 scsscf: 1; ++ short: 3; ++ char: 4; ++ u32 pnso: 1; ++}; ++ ++struct chsc_ssqd_area { ++ struct chsc_header request; ++ short: 10; ++ u8 ssid: 2; ++ u8 fmt: 4; ++ u16 first_sch; ++ short: 16; ++ u16 last_sch; ++ long: 0; ++ struct chsc_header response; ++ struct qdio_ssqd_desc qdio_ssqd; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct chsc_scssc_area { ++ struct chsc_header request; ++ u16 operation_code; ++ long: 64; ++ u64 summary_indicator_addr; ++ u64 subchannel_indicator_addr; ++ u32 ks: 4; ++ u32 kc: 4; ++ short: 8; ++ short: 13; ++ u32 isc: 3; ++ u32 word_with_d_bit; ++ int: 32; ++ struct subchannel_id schid; ++ u32 reserved[1004]; ++ struct chsc_header response; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct chsc_scpd { ++ struct chsc_header request; ++ char: 2; ++ u32 m: 1; ++ u32 c: 1; ++ u32 fmt: 4; ++ u32 cssid: 8; ++ char: 4; ++ u32 rfmt: 4; ++ u32 first_chpid: 8; ++ int: 24; ++ u32 last_chpid: 8; ++ u32 zeroes1; ++ struct chsc_header response; ++ long: 0; ++ u8 data[0]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct chsc_sda_area { ++ struct chsc_header request; ++ char: 4; ++ u8 format: 4; ++ u16 operation_code; ++ long: 64; ++ u32 operation_data_area[252]; ++ struct chsc_header response; ++ char: 4; ++ u32 format2: 4; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct sale { ++ u64 sa; ++ u32 p: 4; ++ u32 op_state: 4; ++ u32 data_state: 4; ++ u32 rank: 4; ++ u32 r: 1; ++ char: 7; ++ u32 rid: 8; ++}; ++ ++struct chsc_scm_info { ++ struct chsc_header request; ++ u64 reqtok; ++ u32 reserved1[4]; ++ struct chsc_header response; ++ long: 0; ++ int: 24; ++ u8 rq; ++ u32 mbc; ++ u64 msa; ++ u16 is; ++ u16 mmc; ++ u32 mci; ++ u64 nr_scm_ini; ++ u64 nr_scm_unini; ++ u32 reserved2[10]; ++ u64 restok; ++ struct sale scmal[248]; ++}; ++ ++struct chsc_ssd_area { ++ struct chsc_header request; ++ short: 10; ++ u16 ssid: 2; ++ u16 f_sch; ++ short: 16; ++ u16 l_sch; ++ long: 0; ++ struct chsc_header response; ++ long: 0; ++ u8 sch_valid: 1; ++ u8 dev_valid: 1; ++ u8 st: 3; ++ u8 zeroes: 3; ++ u8 unit_addr; ++ u16 devno; ++ u8 path_mask; ++ u8 fla_valid_mask; ++ u16 sch; ++ u8 chpid[8]; ++ u16 fla[8]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct chsc_sei_nt0_area { ++ u8 flags; ++ u8 vf; ++ u8 rs; ++ u8 cc; ++ u16 fla; ++ u16 rsid; ++ u32 reserved1; ++ u32 reserved2; ++ u8 ccdf[4056]; ++}; ++ ++struct chsc_sei_nt2_area { ++ u8 flags; ++ u8 reserved1; ++ u8 reserved2; ++ u8 cc; ++ u32 reserved3[13]; ++ u8 ccdf[4016]; ++}; ++ ++struct chsc_sei { ++ struct chsc_header request; ++ u32 reserved1; ++ u64 ntsm; ++ struct chsc_header response; ++ int: 24; ++ u8 nt; ++ union { ++ struct chsc_sei_nt0_area nt0_area; ++ struct chsc_sei_nt2_area nt2_area; ++ u8 nt_area[4072]; ++ } u; ++}; ++ ++struct lir { ++ struct { ++ u32 null: 1; ++ u32 reserved: 3; ++ u32 class: 2; ++ u32 reserved2: 2; ++ } __attribute__((packed)) iq; ++ u32 ic: 8; ++ u32 reserved: 16; ++ struct node_descriptor incident_node; ++ struct node_descriptor attached_node; ++ u8 reserved2[32]; ++}; ++ ++struct chp_config_data { ++ u8 map[32]; ++ u8 op; ++ u8 pc; ++}; ++ ++struct chse_cudb { ++ u16 flags: 8; ++ u16 chp_valid: 8; ++ u16 cu; ++ u32 esm_valid: 8; ++ long: 0; ++ u8 chpid[8]; ++ long: 64; ++ u8 esm[8]; ++ u32 efla[8]; ++}; ++ ++struct chsc_scud { ++ struct chsc_header request; ++ char: 4; ++ u16 fmt: 4; ++ u16 cssid: 8; ++ u16 first_cu; ++ short: 16; ++ u16 last_cu; ++ long: 0; ++ struct chsc_header response; ++ char: 4; ++ u16 fmt_resp: 4; ++ struct chse_cudb cudb[0]; ++}; ++ ++struct cmg_entry { ++ u32 values[8]; ++}; ++ ++enum cfg_task_t { ++ cfg_none = 0, ++ cfg_configure = 1, ++ cfg_deconfigure = 2, ++}; ++ ++typedef struct { ++ char _[4096]; ++} addr_type; ++ ++struct arqb { ++ u64 data; ++ u16 fmt: 4; ++ u16 cmd_code; ++ short: 16; ++ u16 msb_count; ++ u32 reserved[12]; ++}; ++ ++struct arsb { ++ u16 fmt: 4; ++ int: 0; ++ u8 ef; ++ short: 0; ++ u8 ecbi; ++ long: 0; ++ u8 fvf; ++ short: 0; ++ char: 8; ++ u8 eqc; ++ u64 fail_msb; ++ u64 fail_aidaw; ++ u64 fail_ms; ++ u64 fail_scm; ++ u32 reserved[4]; ++}; ++ ++struct msb { ++ u8 fmt: 4; ++ u8 oc: 4; ++ u8 flags; ++ short: 12; ++ u16 bs: 4; ++ u32 blk_count; ++ u64 data_addr; ++ u64 scm_addr; ++ long: 64; ++}; ++ ++struct aob { ++ struct arqb request; ++ struct arsb response; ++ struct msb msb[124]; ++}; ++ ++struct scm_device; ++ ++struct aob_rq_header { ++ struct scm_device *scmdev; ++ char data[0]; ++}; ++ ++struct scm_device { ++ u64 address; ++ u64 size; ++ unsigned int nr_max_block; ++ struct device dev; ++ struct { ++ unsigned int persistence: 4; ++ unsigned int oper_state: 4; ++ unsigned int data_state: 4; ++ unsigned int rank: 4; ++ unsigned int release: 1; ++ unsigned int res_id: 8; ++ } attrs; ++}; ++ ++enum scm_event { ++ SCM_CHANGE = 0, ++ SCM_AVAIL = 1, ++}; ++ ++struct scm_driver { ++ struct device_driver drv; ++ int (*probe)(struct scm_device *); ++ void (*remove)(struct scm_device *); ++ void (*notify)(struct scm_device *, enum scm_event); ++ void (*handler)(struct scm_device *, void *, blk_status_t); ++}; ++ ++struct kbd_data; ++ ++typedef void fn_handler_fn___2(struct kbd_data *); ++ ++struct kbd_data { ++ struct tty_port *port; ++ short unsigned int **key_maps; ++ char **func_table; ++ fn_handler_fn___2 **fn_handler; ++ struct kbdiacruc *accent_table; ++ unsigned int accent_table_size; ++ unsigned int diacr; ++ short unsigned int sysrq; ++}; ++ ++typedef void k_handler_fn___2(struct kbd_data *, unsigned char); ++ ++struct conf_mgm_data { ++ u8 reserved; ++ u8 ev_qualifier; ++}; ++ ++struct sclp_ofb_evbuf { ++ struct evbuf_header header; ++ struct conf_mgm_data cm_data; ++ char ev_data[64]; ++}; ++ ++struct sclp_ofb_sccb { ++ struct sccb_header header; ++ struct sclp_ofb_evbuf ofb_evbuf; ++}; ++ ++struct cpi_evbuf { ++ struct evbuf_header header; ++ u8 id_format; ++ u8 reserved0; ++ u8 system_type[8]; ++ u64 reserved1; ++ u8 system_name[8]; ++ u64 reserved2; ++ u64 system_level; ++ u64 reserved3; ++ u8 sysplex_name[8]; ++ u8 reserved4[16]; ++}; ++ ++struct cpi_sccb { ++ struct sccb_header header; ++ struct cpi_evbuf cpi_evbuf; ++}; ++ ++struct sclp_ipl_info { ++ int is_valid; ++ int has_dump; ++ char loadparm[8]; ++}; ++ ++struct init_sccb { ++ struct sccb_header header; ++ u16 _reserved; ++ u16 mask_length; ++ u8 masks[4084]; ++}; ++ ++struct read_info_sccb { ++ struct sccb_header header; ++ u16 rnmax; ++ u8 rnsize; ++ u8 _pad_11[5]; ++ u16 ncpurl; ++ u16 cpuoff; ++ u8 _pad_20[4]; ++ u8 loadparm[8]; ++ u8 _pad_32[10]; ++ u8 fac42; ++ u8 fac43; ++ u8 _pad_44[4]; ++ u64 facilities; ++ u8 _pad_56[10]; ++ u8 fac66; ++ u8 _pad_67[9]; ++ u32 ibc; ++ u8 _pad80[4]; ++ u8 fac84; ++ u8 fac85; ++ u8 _pad_86[5]; ++ u8 fac91; ++ u8 _pad_92[6]; ++ u8 fac98; ++ u8 hamaxpow; ++ u32 rnsize2; ++ u64 rnmax2; ++ u32 hsa_size; ++ u8 fac116; ++ u8 fac117; ++ u8 fac118; ++ u8 fac119; ++ u16 hcpua; ++ u8 _pad_122[2]; ++ u32 hmfai; ++ u8 _pad_128[6]; ++ u8 byte_134; ++ u8 cpudirq; ++ u16 cbl; ++ u8 _pad_138[12150]; ++}; ++ ++struct physmem_range { ++ u64 start; ++ u64 end; ++}; ++ ++enum reserved_range_type { ++ RR_DECOMPRESSOR = 0, ++ RR_INITRD = 1, ++ RR_VMLINUX = 2, ++ RR_AMODE31 = 3, ++ RR_IPLREPORT = 4, ++ RR_CERT_COMP_LIST = 5, ++ RR_MEM_DETECT_EXTENDED = 6, ++ RR_VMEM = 7, ++ RR_MAX = 8, ++}; ++ ++struct reserved_range { ++ long unsigned int start; ++ long unsigned int end; ++ struct reserved_range *chain; ++}; ++ ++struct physmem_info { ++ u32 range_count; ++ u8 info_source; ++ long unsigned int usable; ++ struct reserved_range reserved[8]; ++ struct physmem_range online[255]; ++ struct physmem_range *online_extended; ++}; ++ ++enum { ++ sclp_init_state_uninitialized = 0, ++ sclp_init_state_initializing = 1, ++ sclp_init_state_initialized = 2, ++}; ++ ++struct write_sccb { ++ struct sccb_header header; ++ struct msg_buf msg; ++}; ++ ++struct vt220_sccb { ++ struct sccb_header header; ++ struct { ++ struct evbuf_header header; ++ char data[0]; ++ } msg; ++}; ++ ++struct idal_buffer { ++ size_t size; ++ size_t page_order; ++ void *data[0]; ++}; ++ ++struct raw3270_view; ++ ++struct raw3270_request { ++ struct list_head list; ++ struct raw3270_view *view; ++ struct ccw1 ccw; ++ void *buffer; ++ size_t size; ++ int rescnt; ++ int rc; ++ void (*callback)(struct raw3270_request *, void *); ++ void *callback_data; ++}; ++ ++struct raw3270; ++ ++struct raw3270_fn; ++ ++struct raw3270_view { ++ struct list_head list; ++ spinlock_t lock; ++ atomic_t ref_count; ++ struct raw3270 *dev; ++ struct raw3270_fn *fn; ++ unsigned int model; ++ unsigned int rows; ++ unsigned int cols; ++ unsigned char *ascebc; ++}; ++ ++struct raw3270_fn { ++ int (*activate)(struct raw3270_view *); ++ void (*deactivate)(struct raw3270_view *); ++ void (*intv)(struct raw3270_view *, struct raw3270_request *, struct irb *); ++ void (*release)(struct raw3270_view *); ++ void (*free)(struct raw3270_view *); ++ void (*resize)(struct raw3270_view *, int, int, int, int, int, int); ++}; ++ ++struct raw3270 { ++ struct list_head list; ++ struct ccw_device *cdev; ++ int minor; ++ int model; ++ int rows; ++ int cols; ++ int old_model; ++ int old_rows; ++ int old_cols; ++ unsigned int state; ++ long unsigned int flags; ++ struct list_head req_queue; ++ struct list_head view_list; ++ struct raw3270_view *view; ++ struct timer_list timer; ++ unsigned char *ascebc; ++ struct raw3270_view init_view; ++ struct raw3270_request init_reset; ++ struct raw3270_request init_readpart; ++ struct raw3270_request init_readmod; ++ unsigned char init_data[256]; ++ struct work_struct resize_work; ++}; ++ ++struct raw3270_notifier { ++ struct list_head list; ++ void (*create)(int); ++ void (*destroy)(int); ++}; ++ ++struct raw3270_ua { ++ struct { ++ short int l; ++ char sfid; ++ char qcode; ++ char flags0; ++ char flags1; ++ short int w; ++ short int h; ++ char units; ++ int xr; ++ int yr; ++ char aw; ++ char ah; ++ short int buffsz; ++ char xmin; ++ char ymin; ++ char xmax; ++ char ymax; ++ } __attribute__((packed)) uab; ++ struct { ++ char l; ++ char sdpid; ++ char res; ++ char auaid; ++ short int wauai; ++ short int hauai; ++ char auaunits; ++ int auaxr; ++ int auayr; ++ char awauai; ++ char ahauai; ++ } __attribute__((packed)) aua; ++}; ++ ++struct err_notify_evbuf { ++ struct evbuf_header header; ++ u8 action; ++ u8 atype; ++ u32 fh; ++ u32 fid; ++ u8 data[0]; ++}; ++ ++struct err_notify_sccb { ++ struct sccb_header header; ++ struct err_notify_evbuf evbuf; ++}; ++ ++struct pci_cfg_sccb { ++ struct sccb_header header; ++ u8 atype; ++ u8 reserved1; ++ u16 reserved2; ++ u32 aid; ++}; ++ ++struct ap_cfg_sccb { ++ struct sccb_header header; ++}; ++ ++struct vmcp_session { ++ char *response; ++ unsigned int bufsize; ++ unsigned int cma_alloc: 1; ++ int resp_size; ++ int resp_code; ++ struct mutex mutex; ++}; ++ ++enum arch_id { ++ ARCH_S390 = 0, ++ ARCH_S390X = 1, ++}; ++ ++struct ipib_info { ++ long unsigned int ipib; ++ u32 checksum; ++} __attribute__((packed)); ++ ++struct vq_config_block { ++ __u16 index; ++ __u16 num; ++}; ++ ++struct vcdev_dma_area { ++ long unsigned int indicators; ++ long unsigned int indicators2; ++ struct vq_config_block config_block; ++ __u8 status; ++}; ++ ++struct virtio_ccw_device { ++ struct virtio_device vdev; ++ __u8 config[256]; ++ struct ccw_device *cdev; ++ __u32 curr_io; ++ int err; ++ unsigned int revision; ++ wait_queue_head_t wait_q; ++ spinlock_t lock; ++ rwlock_t irq_lock; ++ struct mutex io_lock; ++ struct list_head virtqueues; ++ bool is_thinint; ++ bool going_away; ++ bool device_lost; ++ unsigned int config_ready; ++ void *airq_info; ++ struct vcdev_dma_area *dma_area; ++}; ++ ++struct vq_info_block_legacy { ++ __u64 queue; ++ __u32 align; ++ __u16 index; ++ __u16 num; ++}; ++ ++struct vq_info_block { ++ __u64 desc; ++ __u32 res0; ++ __u16 index; ++ __u16 num; ++ __u64 avail; ++ __u64 used; ++}; ++ ++struct virtio_feature_desc { ++ __le32 features; ++ __u8 index; ++} __attribute__((packed)); ++ ++struct virtio_thinint_area { ++ long unsigned int summary_indicator; ++ long unsigned int indicator; ++ u64 bit_nr; ++ u8 isc; ++} __attribute__((packed)); ++ ++struct virtio_rev_info { ++ __u16 revision; ++ __u16 length; ++ __u8 data[0]; ++}; ++ ++struct virtio_ccw_vq_info { ++ struct virtqueue *vq; ++ int num; ++ union { ++ struct vq_info_block s; ++ struct vq_info_block_legacy l; ++ } *info_block; ++ int bit_nr; ++ struct list_head node; ++ long int cookie; ++}; ++ ++struct airq_info { ++ rwlock_t lock; ++ u8 summary_indicator_idx; ++ struct airq_struct airq; ++ struct airq_iv *aiv; ++}; ++ ++struct fprop_local_single { ++ long unsigned int events; ++ unsigned int period; ++ raw_spinlock_t lock; ++}; ++ ++struct ida_bitmap { ++ long unsigned int bitmap[16]; ++}; ++ ++struct klist_waiter { ++ struct list_head list; ++ struct klist_node *node; ++ struct task_struct *process; ++ int woken; ++}; ++ ++enum { ++ LOGIC_PIO_INDIRECT = 0, ++ LOGIC_PIO_CPU_MMIO = 1, ++}; ++ ++struct logic_pio_host_ops; ++ ++struct logic_pio_hwaddr { ++ struct list_head list; ++ struct fwnode_handle *fwnode; ++ resource_size_t hw_start; ++ resource_size_t io_start; ++ resource_size_t size; ++ long unsigned int flags; ++ void *hostdata; ++ const struct logic_pio_host_ops *ops; ++}; ++ ++struct logic_pio_host_ops { ++ u32 (*in)(void *, long unsigned int, size_t); ++ void (*out)(void *, long unsigned int, u32, size_t); ++ u32 (*ins)(void *, long unsigned int, void *, size_t, unsigned int); ++ void (*outs)(void *, long unsigned int, const void *, size_t, unsigned int); ++}; ++ ++struct maple_metadata { ++ unsigned char end; ++ unsigned char gap; ++}; ++ ++struct maple_pnode; ++ ++struct maple_range_64 { ++ struct maple_pnode *parent; ++ long unsigned int pivot[15]; ++ union { ++ void *slot[16]; ++ struct { ++ void *pad[15]; ++ struct maple_metadata meta; ++ }; ++ }; ++}; ++ ++struct maple_arange_64 { ++ struct maple_pnode *parent; ++ long unsigned int pivot[9]; ++ void *slot[10]; ++ long unsigned int gap[10]; ++ struct maple_metadata meta; ++}; ++ ++struct maple_topiary { ++ struct maple_pnode *parent; ++ struct maple_enode *next; ++}; ++ ++enum maple_type { ++ maple_dense = 0, ++ maple_leaf_64 = 1, ++ maple_range_64 = 2, ++ maple_arange_64 = 3, ++}; ++ ++struct maple_node { ++ union { ++ struct { ++ struct maple_pnode *parent; ++ void *slot[31]; ++ }; ++ struct { ++ void *pad; ++ struct callback_head rcu; ++ struct maple_enode *piv_parent; ++ unsigned char parent_slot; ++ enum maple_type type; ++ unsigned char slot_len; ++ unsigned int ma_flags; ++ }; ++ struct maple_range_64 mr64; ++ struct maple_arange_64 ma64; ++ struct maple_alloc alloc; ++ }; ++}; ++ ++struct ma_topiary { ++ struct maple_enode *head; ++ struct maple_enode *tail; ++ struct maple_tree *mtree; ++}; ++ ++struct ma_wr_state { ++ struct ma_state *mas; ++ struct maple_node *node; ++ long unsigned int r_min; ++ long unsigned int r_max; ++ enum maple_type type; ++ unsigned char offset_end; ++ unsigned char node_end; ++ long unsigned int *pivots; ++ long unsigned int end_piv; ++ void **slots; ++ void *entry; ++ void *content; ++}; ++ ++struct trace_event_raw_ma_op { ++ struct trace_entry ent; ++ const char *fn; ++ long unsigned int min; ++ long unsigned int max; ++ long unsigned int index; ++ long unsigned int last; ++ void *node; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ma_read { ++ struct trace_entry ent; ++ const char *fn; ++ long unsigned int min; ++ long unsigned int max; ++ long unsigned int index; ++ long unsigned int last; ++ void *node; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ma_write { ++ struct trace_entry ent; ++ const char *fn; ++ long unsigned int min; ++ long unsigned int max; ++ long unsigned int index; ++ long unsigned int last; ++ long unsigned int piv; ++ void *val; ++ void *node; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_ma_op {}; ++ ++struct trace_event_data_offsets_ma_read {}; ++ ++struct trace_event_data_offsets_ma_write {}; ++ ++typedef void (*btf_trace_ma_op)(void *, const char *, struct ma_state *); ++ ++typedef void (*btf_trace_ma_read)(void *, const char *, struct ma_state *); ++ ++typedef void (*btf_trace_ma_write)(void *, const char *, struct ma_state *, long unsigned int, void *); ++ ++struct maple_big_node { ++ struct maple_pnode *parent; ++ long unsigned int pivot[33]; ++ union { ++ struct maple_enode *slot[34]; ++ struct { ++ long unsigned int padding[21]; ++ long unsigned int gap[21]; ++ }; ++ }; ++ unsigned char b_end; ++ enum maple_type type; ++}; ++ ++struct maple_subtree_state { ++ struct ma_state *orig_l; ++ struct ma_state *orig_r; ++ struct ma_state *l; ++ struct ma_state *m; ++ struct ma_state *r; ++ struct ma_topiary *free; ++ struct ma_topiary *destroy; ++ struct maple_big_node *bn; ++}; ++ ++typedef struct { ++ long unsigned int key[2]; ++} hsiphash_key_t; ++ ++struct tm { ++ int tm_sec; ++ int tm_min; ++ int tm_hour; ++ int tm_mday; ++ int tm_mon; ++ long int tm_year; ++ int tm_wday; ++ int tm_yday; ++}; ++ ++struct rtc_time { ++ int tm_sec; ++ int tm_min; ++ int tm_hour; ++ int tm_mday; ++ int tm_mon; ++ int tm_year; ++ int tm_wday; ++ int tm_yday; ++ int tm_isdst; ++}; ++ ++enum format_type { ++ FORMAT_TYPE_NONE = 0, ++ FORMAT_TYPE_WIDTH = 1, ++ FORMAT_TYPE_PRECISION = 2, ++ FORMAT_TYPE_CHAR = 3, ++ FORMAT_TYPE_STR = 4, ++ FORMAT_TYPE_PTR = 5, ++ FORMAT_TYPE_PERCENT_CHAR = 6, ++ FORMAT_TYPE_INVALID = 7, ++ FORMAT_TYPE_LONG_LONG = 8, ++ FORMAT_TYPE_ULONG = 9, ++ FORMAT_TYPE_LONG = 10, ++ FORMAT_TYPE_UBYTE = 11, ++ FORMAT_TYPE_BYTE = 12, ++ FORMAT_TYPE_USHORT = 13, ++ FORMAT_TYPE_SHORT = 14, ++ FORMAT_TYPE_UINT = 15, ++ FORMAT_TYPE_INT = 16, ++ FORMAT_TYPE_SIZE_T = 17, ++ FORMAT_TYPE_PTRDIFF = 18, ++}; ++ ++struct printf_spec { ++ unsigned int type: 8; ++ int field_width: 24; ++ unsigned int flags: 8; ++ unsigned int base: 8; ++ int precision: 16; ++}; ++ ++struct page_flags_fields { ++ int width; ++ int shift; ++ int mask; ++ const struct printf_spec *spec; ++ const char *name; ++}; ++ ++union ctlreg5 { ++ long unsigned int val; ++ struct { ++ long: 33; ++ long unsigned int pasteo: 25; ++ }; ++}; ++ ++union ctlreg15 { ++ long unsigned int val; ++ struct { ++ long unsigned int lsea: 61; ++ }; ++}; ++ ++typedef struct { ++ unsigned char bytes[16]; ++} cpacf_mask_t; ++ ++struct ipl_rb_component_entry { ++ __u64 addr; ++ __u64 len; ++ __u8 flags; ++ __u8 reserved1[5]; ++ __u16 certificate_index; ++ __u8 reserved2[8]; ++}; ++ ++enum physmem_info_source { ++ MEM_DETECT_NONE = 0, ++ MEM_DETECT_SCLP_STOR_INFO = 1, ++ MEM_DETECT_DIAG260 = 2, ++ MEM_DETECT_SCLP_READ_INFO = 3, ++ MEM_DETECT_BIN_SEARCH = 4, ++}; ++ ++typedef struct { ++ unsigned int len; ++ long unsigned int kernel_addr; ++ long unsigned int process_addr; ++} ptrace_area; ++ ++typedef struct { ++ __u32 len; ++ __u32 kernel_addr; ++ __u32 process_addr; ++} compat_ptrace_area; ++ ++struct file_private_info { ++ loff_t offset; ++ int act_area; ++ int act_page; ++ int act_entry; ++ size_t act_entry_offset; ++ char temp_buf[2048]; ++ debug_info_t *debug_info_org; ++ debug_info_t *debug_info_snap; ++ struct debug_view *view; ++}; ++ ++typedef struct file_private_info file_private_info_t; ++ ++typedef struct { ++ char *string; ++ long int args[0]; ++} debug_sprintf_entry_t; ++ ++enum ipl_pbt { ++ IPL_PBT_FCP = 0, ++ IPL_PBT_SCP_DATA = 1, ++ IPL_PBT_CCW = 2, ++ IPL_PBT_ECKD = 3, ++ IPL_PBT_NVME = 4, ++}; ++ ++struct ipl_rl_hdr { ++ __u32 len; ++ __u8 flags; ++ __u8 reserved1[2]; ++ __u8 version; ++ __u8 reserved2[8]; ++}; ++ ++enum ipl_rbt { ++ IPL_RBT_CERTIFICATES = 1, ++ IPL_RBT_COMPONENTS = 2, ++}; ++ ++struct ipl_rb_certificate_entry { ++ __u64 addr; ++ __u64 len; ++}; ++ ++struct ipl_rb_certificates { ++ __u32 len; ++ __u8 rbt; ++ __u8 reserved1[11]; ++ struct ipl_rb_certificate_entry entries[0]; ++}; ++ ++struct ipl_rb_components { ++ __u32 len; ++ __u8 rbt; ++ __u8 reserved1[11]; ++ struct ipl_rb_component_entry entries[0]; ++}; ++ ++struct ipl_report_component { ++ struct list_head list; ++ struct ipl_rb_component_entry entry; ++}; ++ ++struct ipl_report_certificate { ++ struct list_head list; ++ struct ipl_rb_certificate_entry entry; ++ void *key; ++}; ++ ++struct shutdown_action; ++ ++struct shutdown_trigger { ++ char *name; ++ struct shutdown_action *action; ++}; ++ ++struct shutdown_action { ++ char *name; ++ void (*fn)(struct shutdown_trigger *); ++ int (*init)(); ++ int init_rc; ++}; ++ ++enum dump_type { ++ DUMP_TYPE_NONE = 1, ++ DUMP_TYPE_CCW = 2, ++ DUMP_TYPE_FCP = 4, ++ DUMP_TYPE_NVME = 8, ++ DUMP_TYPE_ECKD = 16, ++}; ++ ++struct os_info_entry { ++ u64 addr; ++ u64 size; ++ u32 csum; ++} __attribute__((packed)); ++ ++struct os_info { ++ u64 magic; ++ u32 csum; ++ u16 version_major; ++ u16 version_minor; ++ u64 crashkernel_addr; ++ u64 crashkernel_size; ++ struct os_info_entry entry[2]; ++ u8 reserved[4024]; ++}; ++ ++enum { ++ BRCL_EXPOLINE = 0, ++ BRASL_EXPOLINE = 1, ++}; ++ ++struct stat64_emu31 { ++ long long unsigned int st_dev; ++ unsigned int __pad1; ++ u32 __st_ino; ++ unsigned int st_mode; ++ unsigned int st_nlink; ++ u32 st_uid; ++ u32 st_gid; ++ long long unsigned int st_rdev; ++ unsigned int __pad3; ++ long int st_size; ++ u32 st_blksize; ++ unsigned char __pad4[4]; ++ u32 __pad5; ++ u32 st_blocks; ++ u32 st_atime; ++ u32 __pad6; ++ u32 st_mtime; ++ u32 __pad7; ++ u32 st_ctime; ++ u32 __pad8; ++ long unsigned int st_ino; ++}; ++ ++struct mmap_arg_struct_emu31 { ++ compat_ulong_t addr; ++ compat_ulong_t len; ++ compat_ulong_t prot; ++ compat_ulong_t flags; ++ compat_ulong_t fd; ++ compat_ulong_t offset; ++}; ++ ++struct fadvise64_64_args { ++ int fd; ++ long long int offset; ++ long long int len; ++ int advice; ++}; ++ ++struct s390_ctrset_start { ++ __u64 version; ++ __u64 data_bytes; ++ __u64 cpumask_len; ++ __u64 *cpumask; ++ __u64 counter_sets; ++}; ++ ++struct s390_ctrset_setdata { ++ __u32 set; ++ __u32 no_cnts; ++ __u64 cv[0]; ++}; ++ ++struct s390_ctrset_cpudata { ++ __u32 cpu_nr; ++ __u32 no_sets; ++ struct s390_ctrset_setdata data[0]; ++}; ++ ++struct s390_ctrset_read { ++ __u64 no_cpus; ++ struct s390_ctrset_cpudata data[0]; ++}; ++ ++enum cpumf_ctr_set { ++ CPUMF_CTR_SET_BASIC = 0, ++ CPUMF_CTR_SET_USER = 1, ++ CPUMF_CTR_SET_CRYPTO = 2, ++ CPUMF_CTR_SET_EXT = 3, ++ CPUMF_CTR_SET_MT_DIAG = 4, ++ CPUMF_CTR_SET_MAX = 5, ++}; ++ ++struct cpu_cf_events { ++ atomic_t ctr_set[5]; ++ u64 state; ++ u64 dev_state; ++ unsigned int flags; ++ size_t used; ++ size_t usedss; ++ unsigned char start[4096]; ++ unsigned char stop[4096]; ++ unsigned char data[4096]; ++ unsigned int sets; ++}; ++ ++struct cf_ctrset_entry { ++ unsigned int def: 16; ++ unsigned int set: 16; ++ unsigned int ctr: 16; ++ unsigned int res1: 16; ++}; ++ ++struct cf_trailer_entry { ++ union { ++ struct { ++ unsigned int clock_base: 1; ++ unsigned int speed: 1; ++ unsigned int mtda: 1; ++ unsigned int caca: 1; ++ unsigned int lcda: 1; ++ }; ++ long unsigned int flags; ++ }; ++ unsigned int cfvn: 16; ++ unsigned int csvn: 16; ++ unsigned int cpu_speed: 32; ++ long unsigned int timestamp; ++ union { ++ struct { ++ long unsigned int progusage1; ++ long unsigned int progusage2; ++ long unsigned int progusage3; ++ long unsigned int tod_base; ++ }; ++ long unsigned int progusage[4]; ++ }; ++ unsigned int mach_type: 16; ++ unsigned int res1: 16; ++ unsigned int res2: 32; ++}; ++ ++struct cfset_call_on_cpu_parm { ++ unsigned int sets; ++ atomic_t cpus_ack; ++}; ++ ++struct cfset_session { ++ struct list_head head; ++}; ++ ++struct cfset_request { ++ long unsigned int ctrset; ++ cpumask_t mask; ++ struct list_head node; ++}; ++ ++enum perf_event_s390_regs { ++ PERF_REG_S390_R0 = 0, ++ PERF_REG_S390_R1 = 1, ++ PERF_REG_S390_R2 = 2, ++ PERF_REG_S390_R3 = 3, ++ PERF_REG_S390_R4 = 4, ++ PERF_REG_S390_R5 = 5, ++ PERF_REG_S390_R6 = 6, ++ PERF_REG_S390_R7 = 7, ++ PERF_REG_S390_R8 = 8, ++ PERF_REG_S390_R9 = 9, ++ PERF_REG_S390_R10 = 10, ++ PERF_REG_S390_R11 = 11, ++ PERF_REG_S390_R12 = 12, ++ PERF_REG_S390_R13 = 13, ++ PERF_REG_S390_R14 = 14, ++ PERF_REG_S390_R15 = 15, ++ PERF_REG_S390_FP0 = 16, ++ PERF_REG_S390_FP1 = 17, ++ PERF_REG_S390_FP2 = 18, ++ PERF_REG_S390_FP3 = 19, ++ PERF_REG_S390_FP4 = 20, ++ PERF_REG_S390_FP5 = 21, ++ PERF_REG_S390_FP6 = 22, ++ PERF_REG_S390_FP7 = 23, ++ PERF_REG_S390_FP8 = 24, ++ PERF_REG_S390_FP9 = 25, ++ PERF_REG_S390_FP10 = 26, ++ PERF_REG_S390_FP11 = 27, ++ PERF_REG_S390_FP12 = 28, ++ PERF_REG_S390_FP13 = 29, ++ PERF_REG_S390_FP14 = 30, ++ PERF_REG_S390_FP15 = 31, ++ PERF_REG_S390_MASK = 32, ++ PERF_REG_S390_PC = 33, ++ PERF_REG_S390_MAX = 34, ++}; ++ ++struct trace_event_raw_s390_diagnose { ++ struct trace_entry ent; ++ short unsigned int nr; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_s390_diagnose {}; ++ ++typedef void (*btf_trace_s390_diagnose)(void *, short unsigned int); ++ ++struct qrange { ++ long unsigned int start; ++ long unsigned int end; ++}; ++ ++struct qout64 { ++ long unsigned int segstart; ++ long unsigned int segend; ++ int segcnt; ++ int segrcnt; ++ struct qrange range[6]; ++}; ++ ++struct qin64 { ++ char qopcode; ++ char rsrv1[3]; ++ char qrcode; ++ char rsrv2[3]; ++ char qname[8]; ++ unsigned int qoutptr; ++ short int qoutlen; ++}; ++ ++struct dcss_segment { ++ struct list_head list; ++ char dcss_name[8]; ++ char res_name[16]; ++ long unsigned int start_addr; ++ long unsigned int end; ++ refcount_t ref_count; ++ int do_nonshared; ++ unsigned int vm_segtype; ++ struct qrange range[6]; ++ int segcnt; ++ struct resource *res; ++}; ++ ++struct chacha_ctx { ++ u32 key[8]; ++ int nrounds; ++}; ++ ++struct cpu_feature { ++ __u16 feature; ++}; ++ ++struct diag204_info_blk_hdr { ++ __u8 npar; ++ __u8 flags; ++ __u16 tslice; ++ __u16 phys_cpus; ++ __u16 this_part; ++ __u64 curtod; ++}; ++ ++struct diag204_part_hdr { ++ __u8 pn; ++ __u8 cpus; ++ char reserved[6]; ++ char part_name[8]; ++}; ++ ++struct diag204_cpu_info { ++ __u16 cpu_addr; ++ char reserved1[2]; ++ __u8 ctidx; ++ __u8 cflag; ++ __u16 weight; ++ __u64 acc_time; ++ __u64 lp_time; ++}; ++ ++struct diag204_phys_hdr { ++ char reserved1[1]; ++ __u8 cpus; ++ char reserved2[6]; ++ char mgm_name[8]; ++}; ++ ++struct diag204_phys_cpu { ++ __u16 cpu_addr; ++ char reserved1[2]; ++ __u8 ctidx; ++ char reserved2[3]; ++ __u64 mgm_time; ++ char reserved3[8]; ++}; ++ ++struct dbfs_d204_hdr { ++ u64 len; ++ u16 version; ++ u8 sc; ++ char reserved[53]; ++}; ++ ++struct dbfs_d204 { ++ struct dbfs_d204_hdr hdr; ++ char buf[0]; ++}; ++ ++struct x_info_blk_hdr; ++ ++struct hypfs_diag304 { ++ __u32 args[2]; ++ __u64 data; ++ __u64 rc; ++}; ++ ++enum zpci_ioat_dtype { ++ ZPCI_IOTA_STO = 0, ++ ZPCI_IOTA_RTTO = 1, ++ ZPCI_IOTA_RSTO = 2, ++ ZPCI_IOTA_RFTO = 3, ++ ZPCI_IOTA_PFAA = 4, ++ ZPCI_IOTA_IOPFAA = 5, ++ ZPCI_IOTA_IOPTO = 7, ++}; ++ ++struct clp_req_hdr { ++ u16 len; ++ u16 cmd; ++ u32 fmt: 4; ++ u32 reserved1: 28; ++ u64 reserved2; ++}; ++ ++struct clp_rsp_hdr { ++ u16 len; ++ u16 rsp; ++ u32 fmt: 4; ++ u32 reserved1: 28; ++ u64 reserved2; ++}; ++ ++struct clp_req_slpc { ++ struct clp_req_hdr hdr; ++}; ++ ++struct clp_rsp_slpc { ++ struct clp_rsp_hdr hdr; ++ u32 reserved2[4]; ++ u32 lpif[8]; ++ u32 reserved3[8]; ++ u32 lpic[8]; ++}; ++ ++struct clp_req_rsp_slpc { ++ struct clp_req_slpc request; ++ struct clp_rsp_slpc response; ++}; ++ ++struct clp_fh_list_entry { ++ u16 device_id; ++ u16 vendor_id; ++ u32 config_state: 1; ++ u32 fid; ++ u32 fh; ++}; ++ ++struct clp_rsp_slpc_pci { ++ struct clp_rsp_hdr hdr; ++ u32 reserved2[4]; ++ u32 lpif[8]; ++ u32 reserved3[4]; ++ u32 vwb: 1; ++ char: 1; ++ u32 mio_wb: 6; ++ u32 reserved5[3]; ++ u32 lpic[8]; ++}; ++ ++struct clp_req_list_pci { ++ struct clp_req_hdr hdr; ++ u64 resume_token; ++ u64 reserved2; ++}; ++ ++struct clp_rsp_list_pci { ++ struct clp_rsp_hdr hdr; ++ u64 resume_token; ++ u32 reserved2; ++ u16 max_fn; ++ char: 7; ++ u8 uid_checking: 1; ++ u8 entry_size; ++ struct clp_fh_list_entry fh_list[252]; ++}; ++ ++struct mio_info { ++ u32 valid: 6; ++ struct { ++ u64 wb; ++ u64 wt; ++ } addr[6]; ++ u32 reserved[6]; ++}; ++ ++struct clp_req_query_pci { ++ struct clp_req_hdr hdr; ++ u32 fh; ++ u32 reserved2; ++ u64 reserved3; ++}; ++ ++struct clp_rsp_query_pci { ++ struct clp_rsp_hdr hdr; ++ u16 vfn; ++ char: 3; ++ u16 rid_avail: 1; ++ u16 is_physfn: 1; ++ u16 reserved1: 1; ++ u16 mio_addr_avail: 1; ++ u16 util_str_avail: 1; ++ u16 pfgid: 8; ++ u32 fid; ++ u8 bar_size[6]; ++ u16 pchid; ++ __le32 bar[6]; ++ u8 pfip[4]; ++ short: 12; ++ u16 port: 4; ++ u8 fmb_len; ++ u8 pft; ++ u64 sdma; ++ u64 edma; ++ u16 rid; ++ u16 reserved0; ++ u32 reserved[10]; ++ u32 uid; ++ u8 util_str[64]; ++ u32 reserved2[16]; ++ struct mio_info mio; ++}; ++ ++struct clp_req_query_pci_grp { ++ struct clp_req_hdr hdr; ++ u32 reserved2: 24; ++ u32 pfgid: 8; ++ u32 reserved3; ++ u64 reserved4; ++}; ++ ++struct clp_rsp_query_pci_grp { ++ struct clp_rsp_hdr hdr; ++ char: 4; ++ u16 noi: 12; ++ u8 version; ++ char: 6; ++ u8 frame: 1; ++ u8 refresh: 1; ++ char: 3; ++ u16 maxstbl: 13; ++ u16 mui; ++ u8 dtsm; ++ u8 reserved3; ++ u16 maxfaal; ++ char: 4; ++ u16 dnoi: 12; ++ u16 maxcpu; ++ u64 dasm; ++ u64 msia; ++ u64 reserved4; ++ u64 reserved5; ++}; ++ ++struct clp_req_set_pci { ++ struct clp_req_hdr hdr; ++ u32 fh; ++ u16 reserved2; ++ u8 oc; ++ u8 ndas; ++ u32 reserved3; ++ u32 gisa; ++}; ++ ++struct clp_rsp_set_pci { ++ struct clp_rsp_hdr hdr; ++ u32 fh; ++ u32 reserved1; ++ u64 reserved2; ++ struct mio_info mio; ++}; ++ ++struct clp_req_rsp_slpc_pci { ++ struct clp_req_slpc request; ++ struct clp_rsp_slpc_pci response; ++}; ++ ++struct clp_req_rsp_list_pci { ++ struct clp_req_list_pci request; ++ struct clp_rsp_list_pci response; ++}; ++ ++struct clp_req_rsp_set_pci { ++ struct clp_req_set_pci request; ++ struct clp_rsp_set_pci response; ++}; ++ ++struct clp_req_rsp_query_pci { ++ struct clp_req_query_pci request; ++ struct clp_rsp_query_pci response; ++}; ++ ++struct clp_req_rsp_query_pci_grp { ++ struct clp_req_query_pci_grp request; ++ struct clp_rsp_query_pci_grp response; ++}; ++ ++struct clp_req { ++ unsigned int c: 1; ++ unsigned int r: 1; ++ unsigned int lps: 6; ++ unsigned int cmd: 8; ++ unsigned int reserved; ++ __u64 data_p; ++}; ++ ++struct clone_args { ++ __u64 flags; ++ __u64 pidfd; ++ __u64 child_tid; ++ __u64 parent_tid; ++ __u64 exit_signal; ++ __u64 stack; ++ __u64 stack_size; ++ __u64 tls; ++ __u64 set_tid; ++ __u64 set_tid_size; ++ __u64 cgroup; ++}; ++ ++struct multiprocess_signals { ++ sigset_t signal; ++ struct hlist_node node; ++}; ++ ++struct trace_event_raw_task_newtask { ++ struct trace_entry ent; ++ pid_t pid; ++ char comm[16]; ++ long unsigned int clone_flags; ++ short int oom_score_adj; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_task_rename { ++ struct trace_entry ent; ++ pid_t pid; ++ char oldcomm[16]; ++ char newcomm[16]; ++ short int oom_score_adj; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_task_newtask {}; ++ ++struct trace_event_data_offsets_task_rename {}; ++ ++typedef void (*btf_trace_task_newtask)(void *, struct task_struct *, long unsigned int); ++ ++typedef void (*btf_trace_task_rename)(void *, struct task_struct *, const char *); ++ ++struct vm_stack { ++ struct callback_head rcu; ++ struct vm_struct *stack_vm_area; ++}; ++ ++typedef struct vm_struct *pcp_op_T_____8; ++ ++struct resource_constraint { ++ resource_size_t min; ++ resource_size_t max; ++ resource_size_t align; ++ resource_size_t (*alignf)(void *, const struct resource *, resource_size_t, resource_size_t); ++ void *alignf_data; ++}; ++ ++enum { ++ MAX_IORES_LEVEL = 5, ++}; ++ ++struct region_devres { ++ struct resource *parent; ++ resource_size_t start; ++ resource_size_t n; ++}; ++ ++struct __user_cap_header_struct { ++ __u32 version; ++ int pid; ++}; ++ ++typedef struct __user_cap_header_struct *cap_user_header_t; ++ ++struct __user_cap_data_struct { ++ __u32 effective; ++ __u32 permitted; ++ __u32 inheritable; ++}; ++ ++typedef struct __user_cap_data_struct *cap_user_data_t; ++ ++typedef long unsigned int old_sigset_t; ++ ++struct old_sigaction { ++ __sighandler_t sa_handler; ++ old_sigset_t sa_mask; ++ long unsigned int sa_flags; ++ __sigrestore_t sa_restorer; ++}; ++ ++typedef u32 compat_old_sigset_t; ++ ++struct compat_sigaction { ++ compat_uptr_t sa_handler; ++ compat_ulong_t sa_flags; ++ compat_uptr_t sa_restorer; ++ compat_sigset_t sa_mask; ++}; ++ ++struct compat_old_sigaction { ++ compat_uptr_t sa_handler; ++ compat_old_sigset_t sa_mask; ++ compat_ulong_t sa_flags; ++ compat_uptr_t sa_restorer; ++}; ++ ++enum { ++ TRACE_SIGNAL_DELIVERED = 0, ++ TRACE_SIGNAL_IGNORED = 1, ++ TRACE_SIGNAL_ALREADY_PENDING = 2, ++ TRACE_SIGNAL_OVERFLOW_FAIL = 3, ++ TRACE_SIGNAL_LOSE_INFO = 4, ++}; ++ ++struct trace_event_raw_signal_generate { ++ struct trace_entry ent; ++ int sig; ++ int errno; ++ int code; ++ char comm[16]; ++ pid_t pid; ++ int group; ++ int result; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_signal_deliver { ++ struct trace_entry ent; ++ int sig; ++ int errno; ++ int code; ++ long unsigned int sa_handler; ++ long unsigned int sa_flags; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_signal_generate {}; ++ ++struct trace_event_data_offsets_signal_deliver {}; ++ ++typedef void (*btf_trace_signal_generate)(void *, int, struct kernel_siginfo *, struct task_struct *, int, int); ++ ++typedef void (*btf_trace_signal_deliver)(void *, int, struct kernel_siginfo *, struct k_sigaction *); ++ ++enum sig_handler { ++ HANDLER_CURRENT = 0, ++ HANDLER_SIG_DFL = 1, ++ HANDLER_EXIT = 2, ++}; ++ ++struct wq_flusher; ++ ++struct worker; ++ ++struct pool_workqueue; ++ ++struct wq_device; ++ ++struct workqueue_struct { ++ struct list_head pwqs; ++ struct list_head list; ++ struct mutex mutex; ++ int work_color; ++ int flush_color; ++ atomic_t nr_pwqs_to_flush; ++ struct wq_flusher *first_flusher; ++ struct list_head flusher_queue; ++ struct list_head flusher_overflow; ++ struct list_head maydays; ++ struct worker *rescuer; ++ int nr_drainers; ++ int saved_max_active; ++ struct workqueue_attrs *unbound_attrs; ++ struct pool_workqueue *dfl_pwq; ++ struct wq_device *wq_dev; ++ char name[24]; ++ struct callback_head rcu; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ unsigned int flags; ++ struct pool_workqueue *cpu_pwqs; ++ struct pool_workqueue *numa_pwq_tbl[0]; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct worker_pool; ++ ++struct worker { ++ union { ++ struct list_head entry; ++ struct hlist_node hentry; ++ }; ++ struct work_struct *current_work; ++ work_func_t current_func; ++ struct pool_workqueue *current_pwq; ++ unsigned int current_color; ++ struct list_head scheduled; ++ struct task_struct *task; ++ struct worker_pool *pool; ++ struct list_head node; ++ long unsigned int last_active; ++ unsigned int flags; ++ int id; ++ int sleeping; ++ char desc[24]; ++ struct workqueue_struct *rescue_wq; ++ work_func_t last_func; ++}; ++ ++struct pool_workqueue { ++ struct worker_pool *pool; ++ struct workqueue_struct *wq; ++ int work_color; ++ int flush_color; ++ int refcnt; ++ int nr_in_flight[16]; ++ int nr_active; ++ int max_active; ++ struct list_head inactive_works; ++ struct list_head pwqs_node; ++ struct list_head mayday_node; ++ struct work_struct unbound_release_work; ++ struct callback_head rcu; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct worker_pool { ++ raw_spinlock_t lock; ++ int cpu; ++ int node; ++ int id; ++ unsigned int flags; ++ long unsigned int watchdog_ts; ++ bool cpu_stall; ++ int nr_running; ++ struct list_head worklist; ++ int nr_workers; ++ int nr_idle; ++ struct list_head idle_list; ++ struct timer_list idle_timer; ++ struct work_struct idle_cull_work; ++ struct timer_list mayday_timer; ++ struct hlist_head busy_hash[64]; ++ struct worker *manager; ++ struct list_head workers; ++ struct list_head dying_workers; ++ struct completion *detach_completion; ++ struct ida worker_ida; ++ struct workqueue_attrs *attrs; ++ struct hlist_node hash_node; ++ int refcnt; ++ struct callback_head rcu; ++}; ++ ++enum { ++ POOL_MANAGER_ACTIVE = 1, ++ POOL_DISASSOCIATED = 4, ++ WORKER_DIE = 2, ++ WORKER_IDLE = 4, ++ WORKER_PREP = 8, ++ WORKER_CPU_INTENSIVE = 64, ++ WORKER_UNBOUND = 128, ++ WORKER_REBOUND = 256, ++ WORKER_NOT_RUNNING = 456, ++ NR_STD_WORKER_POOLS = 2, ++ UNBOUND_POOL_HASH_ORDER = 6, ++ BUSY_WORKER_HASH_ORDER = 6, ++ MAX_IDLE_WORKERS_RATIO = 4, ++ IDLE_WORKER_TIMEOUT = 30000, ++ MAYDAY_INITIAL_TIMEOUT = 2, ++ MAYDAY_INTERVAL = 10, ++ CREATE_COOLDOWN = 100, ++ RESCUER_NICE_LEVEL = -20, ++ HIGHPRI_NICE_LEVEL = -20, ++ WQ_NAME_LEN = 24, ++}; ++ ++struct wq_flusher { ++ struct list_head list; ++ int flush_color; ++ struct completion done; ++}; ++ ++struct wq_device { ++ struct workqueue_struct *wq; ++ struct device dev; ++}; ++ ++struct trace_event_raw_workqueue_queue_work { ++ struct trace_entry ent; ++ void *work; ++ void *function; ++ u32 __data_loc_workqueue; ++ int req_cpu; ++ int cpu; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_workqueue_activate_work { ++ struct trace_entry ent; ++ void *work; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_workqueue_execute_start { ++ struct trace_entry ent; ++ void *work; ++ void *function; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_workqueue_execute_end { ++ struct trace_entry ent; ++ void *work; ++ void *function; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_workqueue_queue_work { ++ u32 workqueue; ++}; ++ ++struct trace_event_data_offsets_workqueue_activate_work {}; ++ ++struct trace_event_data_offsets_workqueue_execute_start {}; ++ ++struct trace_event_data_offsets_workqueue_execute_end {}; ++ ++typedef void (*btf_trace_workqueue_queue_work)(void *, int, struct pool_workqueue *, struct work_struct *); ++ ++typedef void (*btf_trace_workqueue_activate_work)(void *, struct work_struct *); ++ ++typedef void (*btf_trace_workqueue_execute_start)(void *, struct work_struct *); ++ ++typedef void (*btf_trace_workqueue_execute_end)(void *, struct work_struct *, work_func_t); ++ ++struct wq_barrier { ++ struct work_struct work; ++ struct completion done; ++ struct task_struct *task; ++}; ++ ++struct cwt_wait { ++ wait_queue_entry_t wait; ++ struct work_struct *work; ++}; ++ ++struct apply_wqattrs_ctx { ++ struct workqueue_struct *wq; ++ struct workqueue_attrs *attrs; ++ struct list_head list; ++ struct pool_workqueue *dfl_pwq; ++ struct pool_workqueue *pwq_tbl[0]; ++}; ++ ++struct pr_cont_work_struct { ++ bool comma; ++ work_func_t func; ++ long int ctr; ++}; ++ ++struct work_for_cpu { ++ struct work_struct work; ++ long int (*fn)(void *); ++ void *arg; ++ long int ret; ++}; ++ ++enum { ++ KERNEL_PARAM_OPS_FL_NOARG = 1, ++}; ++ ++struct param_attribute { ++ struct module_attribute mattr; ++ const struct kernel_param *param; ++}; ++ ++struct module_param_attrs { ++ unsigned int num; ++ struct attribute_group grp; ++ struct param_attribute attrs[0]; ++}; ++ ++struct kmalloced_param { ++ struct list_head list; ++ char val[0]; ++}; ++ ++enum what { ++ PROC_EVENT_NONE = 0, ++ PROC_EVENT_FORK = 1, ++ PROC_EVENT_EXEC = 2, ++ PROC_EVENT_UID = 4, ++ PROC_EVENT_GID = 64, ++ PROC_EVENT_SID = 128, ++ PROC_EVENT_PTRACE = 256, ++ PROC_EVENT_COMM = 512, ++ PROC_EVENT_COREDUMP = 1073741824, ++ PROC_EVENT_EXIT = 2147483648, ++}; ++ ++struct async_entry { ++ struct list_head domain_list; ++ struct list_head global_list; ++ struct work_struct work; ++ async_cookie_t cookie; ++ async_func_t func; ++ void *data; ++ struct async_domain *domain; ++}; ++ ++enum uclamp_id { ++ UCLAMP_MIN = 0, ++ UCLAMP_MAX = 1, ++ UCLAMP_CNT = 2, ++}; ++ ++struct numa_group { ++ refcount_t refcount; ++ spinlock_t lock; ++ int nr_tasks; ++ pid_t gid; ++ int active_nodes; ++ struct callback_head rcu; ++ long unsigned int total_faults; ++ long unsigned int max_faults_cpu; ++ long unsigned int faults[0]; ++}; ++ ++struct cpuidle_state_usage { ++ long long unsigned int disable; ++ long long unsigned int usage; ++ u64 time_ns; ++ long long unsigned int above; ++ long long unsigned int below; ++ long long unsigned int rejected; ++}; ++ ++struct cpuidle_device; ++ ++struct cpuidle_driver; ++ ++struct cpuidle_state { ++ char name[16]; ++ char desc[32]; ++ s64 exit_latency_ns; ++ s64 target_residency_ns; ++ unsigned int flags; ++ unsigned int exit_latency; ++ int power_usage; ++ unsigned int target_residency; ++ int (*enter)(struct cpuidle_device *, struct cpuidle_driver *, int); ++ int (*enter_dead)(struct cpuidle_device *, int); ++ int (*enter_s2idle)(struct cpuidle_device *, struct cpuidle_driver *, int); ++}; ++ ++struct cpuidle_state_kobj; ++ ++struct cpuidle_driver_kobj; ++ ++struct cpuidle_device_kobj; ++ ++struct cpuidle_device { ++ unsigned int registered: 1; ++ unsigned int enabled: 1; ++ unsigned int poll_time_limit: 1; ++ unsigned int cpu; ++ ktime_t next_hrtimer; ++ int last_state_idx; ++ u64 last_residency_ns; ++ u64 poll_limit_ns; ++ u64 forced_idle_latency_limit_ns; ++ struct cpuidle_state_usage states_usage[10]; ++ struct cpuidle_state_kobj *kobjs[10]; ++ struct cpuidle_driver_kobj *kobj_driver; ++ struct cpuidle_device_kobj *kobj_dev; ++ struct list_head device_list; ++}; ++ ++struct cpuidle_driver { ++ const char *name; ++ struct module *owner; ++ unsigned int bctimer: 1; ++ struct cpuidle_state states[10]; ++ int state_count; ++ int safe_state_index; ++ struct cpumask *cpumask; ++ const char *governor; ++}; ++ ++enum numa_faults_stats { ++ NUMA_MEM = 0, ++ NUMA_CPU = 1, ++ NUMA_MEMBUF = 2, ++ NUMA_CPUBUF = 3, ++}; ++ ++enum numa_type { ++ node_has_spare = 0, ++ node_fully_busy = 1, ++ node_overloaded = 2, ++}; ++ ++struct numa_stats { ++ long unsigned int load; ++ long unsigned int runnable; ++ long unsigned int util; ++ long unsigned int compute_capacity; ++ unsigned int nr_running; ++ unsigned int weight; ++ enum numa_type node_type; ++ int idle_cpu; ++}; ++ ++struct task_numa_env { ++ struct task_struct *p; ++ int src_cpu; ++ int src_nid; ++ int dst_cpu; ++ int dst_nid; ++ int imb_numa_nr; ++ struct numa_stats src_stats; ++ struct numa_stats dst_stats; ++ int imbalance_pct; ++ int dist; ++ struct task_struct *best_task; ++ long int best_imp; ++ int best_cpu; ++}; ++ ++struct energy_env { ++ long unsigned int task_busy_time; ++ long unsigned int pd_busy_time; ++ long unsigned int cpu_cap; ++ long unsigned int pd_cap; ++}; ++ ++enum fbq_type { ++ regular = 0, ++ remote = 1, ++ all = 2, ++}; ++ ++enum group_type { ++ group_has_spare = 0, ++ group_fully_busy = 1, ++ group_misfit_task = 2, ++ group_asym_packing = 3, ++ group_imbalanced = 4, ++ group_overloaded = 5, ++}; ++ ++enum migration_type { ++ migrate_load = 0, ++ migrate_util = 1, ++ migrate_task = 2, ++ migrate_misfit = 3, ++}; ++ ++struct lb_env { ++ struct sched_domain *sd; ++ struct rq *src_rq; ++ int src_cpu; ++ int dst_cpu; ++ struct rq *dst_rq; ++ struct cpumask *dst_grpmask; ++ int new_dst_cpu; ++ enum cpu_idle_type idle; ++ long int imbalance; ++ struct cpumask *cpus; ++ unsigned int flags; ++ unsigned int loop; ++ unsigned int loop_break; ++ unsigned int loop_max; ++ enum fbq_type fbq_type; ++ enum migration_type migration_type; ++ struct list_head tasks; ++}; ++ ++struct sg_lb_stats { ++ long unsigned int avg_load; ++ long unsigned int group_load; ++ long unsigned int group_capacity; ++ long unsigned int group_util; ++ long unsigned int group_runnable; ++ unsigned int sum_nr_running; ++ unsigned int sum_h_nr_running; ++ unsigned int idle_cpus; ++ unsigned int group_weight; ++ enum group_type group_type; ++ unsigned int group_asym_packing; ++ long unsigned int group_misfit_task_load; ++ unsigned int nr_numa_running; ++ unsigned int nr_preferred_running; ++}; ++ ++struct sd_lb_stats { ++ struct sched_group *busiest; ++ struct sched_group *local; ++ long unsigned int total_load; ++ long unsigned int total_capacity; ++ long unsigned int avg_load; ++ unsigned int prefer_sibling; ++ struct sg_lb_stats busiest_stat; ++ struct sg_lb_stats local_stat; ++}; ++ ++struct dl_bandwidth { ++ raw_spinlock_t dl_runtime_lock; ++ u64 dl_runtime; ++ u64 dl_period; ++}; ++ ++struct idle_timer { ++ struct hrtimer timer; ++ int done; ++}; ++ ++typedef struct rt_rq *rt_rq_iter_t; ++ ++enum dl_bw_request { ++ dl_bw_req_check_overflow = 0, ++ dl_bw_req_alloc = 1, ++ dl_bw_req_free = 2, ++}; ++ ++struct trace_event_raw_contention_begin { ++ struct trace_entry ent; ++ void *lock_addr; ++ unsigned int flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_contention_end { ++ struct trace_entry ent; ++ void *lock_addr; ++ int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_contention_begin {}; ++ ++struct trace_event_data_offsets_contention_end {}; ++ ++typedef void (*btf_trace_contention_begin)(void *, void *, unsigned int); ++ ++typedef void (*btf_trace_contention_end)(void *, void *, int); ++ ++struct mutex_waiter { ++ struct list_head list; ++ struct task_struct *task; ++ struct ww_acquire_ctx *ww_ctx; ++}; ++ ++enum rwsem_waiter_type { ++ RWSEM_WAITING_FOR_WRITE = 0, ++ RWSEM_WAITING_FOR_READ = 1, ++}; ++ ++struct rwsem_waiter { ++ struct list_head list; ++ struct task_struct *task; ++ enum rwsem_waiter_type type; ++ long unsigned int timeout; ++ bool handoff_set; ++}; ++ ++enum rwsem_wake_type { ++ RWSEM_WAKE_ANY = 0, ++ RWSEM_WAKE_READERS = 1, ++ RWSEM_WAKE_READ_OWNED = 2, ++}; ++ ++enum owner_state { ++ OWNER_NULL = 1, ++ OWNER_WRITER = 2, ++ OWNER_READER = 4, ++ OWNER_NONSPINNABLE = 8, ++}; ++ ++struct optimistic_spin_node { ++ struct optimistic_spin_node *next; ++ struct optimistic_spin_node *prev; ++ int locked; ++ int cpu; ++}; ++ ++struct prb_data_block { ++ long unsigned int id; ++ char data[0]; ++}; ++ ++enum { ++ IRQ_STARTUP_NORMAL = 0, ++ IRQ_STARTUP_MANAGED = 1, ++ IRQ_STARTUP_ABORT = 2, ++}; ++ ++struct msi_map { ++ int index; ++ int virq; ++}; ++ ++struct msi_ctrl { ++ unsigned int domid; ++ unsigned int first; ++ unsigned int last; ++ unsigned int nirqs; ++}; ++ ++struct rcu_gp_oldstate { ++ long unsigned int rgos_norm; ++ long unsigned int rgos_exp; ++}; ++ ++struct rcu_exp_work { ++ long unsigned int rew_s; ++ struct work_struct rew_work; ++}; ++ ++struct rcu_node { ++ raw_spinlock_t lock; ++ long unsigned int gp_seq; ++ long unsigned int gp_seq_needed; ++ long unsigned int completedqs; ++ long unsigned int qsmask; ++ long unsigned int rcu_gp_init_mask; ++ long unsigned int qsmaskinit; ++ long unsigned int qsmaskinitnext; ++ long unsigned int expmask; ++ long unsigned int expmaskinit; ++ long unsigned int expmaskinitnext; ++ long unsigned int cbovldmask; ++ long unsigned int ffmask; ++ long unsigned int grpmask; ++ int grplo; ++ int grphi; ++ u8 grpnum; ++ u8 level; ++ bool wait_blkd_tasks; ++ struct rcu_node *parent; ++ struct list_head blkd_tasks; ++ struct list_head *gp_tasks; ++ struct list_head *exp_tasks; ++ struct list_head *boost_tasks; ++ struct rt_mutex boost_mtx; ++ long unsigned int boost_time; ++ struct mutex boost_kthread_mutex; ++ struct task_struct *boost_kthread_task; ++ unsigned int boost_kthread_status; ++ long unsigned int n_boosts; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ raw_spinlock_t fqslock; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ spinlock_t exp_lock; ++ long unsigned int exp_seq_rq; ++ wait_queue_head_t exp_wq[4]; ++ struct rcu_exp_work rew; ++ bool exp_need_flush; ++ raw_spinlock_t exp_poll_lock; ++ long unsigned int exp_seq_poll_rq; ++ struct work_struct exp_poll_wq; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++union rcu_noqs { ++ struct { ++ u8 norm; ++ u8 exp; ++ } b; ++ u16 s; ++}; ++ ++struct rcu_snap_record { ++ long unsigned int gp_seq; ++ u64 cputime_irq; ++ u64 cputime_softirq; ++ u64 cputime_system; ++ long unsigned int nr_hardirqs; ++ unsigned int nr_softirqs; ++ long long unsigned int nr_csw; ++ long unsigned int jiffies; ++}; ++ ++struct rcu_data { ++ long unsigned int gp_seq; ++ long unsigned int gp_seq_needed; ++ union rcu_noqs cpu_no_qs; ++ bool core_needs_qs; ++ bool beenonline; ++ bool gpwrap; ++ bool cpu_started; ++ struct rcu_node *mynode; ++ long unsigned int grpmask; ++ long unsigned int ticks_this_gp; ++ struct irq_work defer_qs_iw; ++ bool defer_qs_iw_pending; ++ struct work_struct strict_work; ++ struct rcu_segcblist cblist; ++ long int qlen_last_fqs_check; ++ long unsigned int n_cbs_invoked; ++ long unsigned int n_force_qs_snap; ++ long int blimit; ++ int dynticks_snap; ++ bool rcu_need_heavy_qs; ++ bool rcu_urgent_qs; ++ bool rcu_forced_tick; ++ bool rcu_forced_tick_exp; ++ long unsigned int barrier_seq_snap; ++ struct callback_head barrier_head; ++ int exp_dynticks_snap; ++ struct task_struct *rcu_cpu_kthread_task; ++ unsigned int rcu_cpu_kthread_status; ++ char rcu_cpu_has_work; ++ long unsigned int rcuc_activity; ++ unsigned int softirq_snap; ++ struct irq_work rcu_iw; ++ bool rcu_iw_pending; ++ long unsigned int rcu_iw_gp_seq; ++ long unsigned int rcu_ofl_gp_seq; ++ short int rcu_ofl_gp_flags; ++ long unsigned int rcu_onl_gp_seq; ++ short int rcu_onl_gp_flags; ++ long unsigned int last_fqs_resched; ++ long unsigned int last_sched_clock; ++ struct rcu_snap_record snap_record; ++ long int lazy_len; ++ int cpu; ++}; ++ ++struct rcu_state { ++ struct rcu_node node[33]; ++ struct rcu_node *level[3]; ++ int ncpus; ++ int n_online_cpus; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long unsigned int gp_seq; ++ long unsigned int gp_max; ++ struct task_struct *gp_kthread; ++ struct swait_queue_head gp_wq; ++ short int gp_flags; ++ short int gp_state; ++ long unsigned int gp_wake_time; ++ long unsigned int gp_wake_seq; ++ long unsigned int gp_seq_polled; ++ long unsigned int gp_seq_polled_snap; ++ long unsigned int gp_seq_polled_exp_snap; ++ struct mutex barrier_mutex; ++ atomic_t barrier_cpu_count; ++ struct completion barrier_completion; ++ long unsigned int barrier_sequence; ++ raw_spinlock_t barrier_lock; ++ struct mutex exp_mutex; ++ struct mutex exp_wake_mutex; ++ long unsigned int expedited_sequence; ++ atomic_t expedited_need_qs; ++ struct swait_queue_head expedited_wq; ++ int ncpus_snap; ++ u8 cbovld; ++ u8 cbovldnext; ++ long unsigned int jiffies_force_qs; ++ long unsigned int jiffies_kick_kthreads; ++ long unsigned int n_force_qs; ++ long unsigned int gp_start; ++ long unsigned int gp_end; ++ long unsigned int gp_activity; ++ long unsigned int gp_req_activity; ++ long unsigned int jiffies_stall; ++ long unsigned int jiffies_resched; ++ long unsigned int n_force_qs_gpstart; ++ const char *name; ++ char abbr; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ arch_spinlock_t ofl_lock; ++ int nocb_is_setup; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct kvfree_rcu_bulk_data { ++ struct list_head list; ++ long unsigned int gp_snap; ++ long unsigned int nr_records; ++ void *records[0]; ++}; ++ ++struct kfree_rcu_cpu; ++ ++struct kfree_rcu_cpu_work { ++ struct rcu_work rcu_work; ++ struct callback_head *head_free; ++ struct list_head bulk_head_free[2]; ++ struct kfree_rcu_cpu *krcp; ++}; ++ ++struct kfree_rcu_cpu { ++ struct callback_head *head; ++ long unsigned int head_gp_snap; ++ atomic_t head_count; ++ struct list_head bulk_head[2]; ++ atomic_t bulk_count[2]; ++ struct kfree_rcu_cpu_work krw_arr[2]; ++ raw_spinlock_t lock; ++ struct delayed_work monitor_work; ++ bool initialized; ++ struct delayed_work page_cache_work; ++ atomic_t backoff_page_cache_fill; ++ atomic_t work_in_progress; ++ struct hrtimer hrtimer; ++ struct llist_head bkvcache; ++ int nr_bkv_objs; ++}; ++ ++enum mod_license { ++ NOT_GPL_ONLY = 0, ++ GPL_ONLY = 1, ++}; ++ ++struct find_symbol_arg { ++ const char *name; ++ bool gplok; ++ bool warn; ++ struct module *owner; ++ const s32 *crc; ++ const struct kernel_symbol *sym; ++ enum mod_license license; ++}; ++ ++enum fail_dup_mod_reason { ++ FAIL_DUP_MOD_BECOMING = 0, ++ FAIL_DUP_MOD_LOAD = 1, ++}; ++ ++struct trace_event_raw_module_load { ++ struct trace_entry ent; ++ unsigned int taints; ++ u32 __data_loc_name; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_module_free { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_module_refcnt { ++ struct trace_entry ent; ++ long unsigned int ip; ++ int refcnt; ++ u32 __data_loc_name; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_module_request { ++ struct trace_entry ent; ++ long unsigned int ip; ++ bool wait; ++ u32 __data_loc_name; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_module_load { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_module_free { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_module_refcnt { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_module_request { ++ u32 name; ++}; ++ ++typedef void (*btf_trace_module_load)(void *, struct module *); ++ ++typedef void (*btf_trace_module_free)(void *, struct module *); ++ ++typedef void (*btf_trace_module_get)(void *, struct module *, long unsigned int); ++ ++typedef void (*btf_trace_module_put)(void *, struct module *, long unsigned int); ++ ++typedef void (*btf_trace_module_request)(void *, char *, bool, long unsigned int); ++ ++struct symsearch { ++ const struct kernel_symbol *start; ++ const struct kernel_symbol *stop; ++ const s32 *crcs; ++ enum mod_license license; ++}; ++ ++struct mod_initfree { ++ struct llist_node node; ++ void *init_text; ++ void *init_data; ++ void *init_rodata; ++}; ++ ++enum kcmp_type { ++ KCMP_FILE = 0, ++ KCMP_VM = 1, ++ KCMP_FILES = 2, ++ KCMP_FS = 3, ++ KCMP_SIGHAND = 4, ++ KCMP_IO = 5, ++ KCMP_SYSVSEM = 6, ++ KCMP_EPOLL_TFD = 7, ++ KCMP_TYPES = 8, ++}; ++ ++struct kcmp_epoll_slot { ++ __u32 efd; ++ __u32 tfd; ++ __u32 toff; ++}; ++ ++struct stacktrace_cookie { ++ long unsigned int *store; ++ unsigned int size; ++ unsigned int skip; ++ unsigned int len; ++}; ++ ++struct rtc_wkalrm { ++ unsigned char enabled; ++ unsigned char pending; ++ struct rtc_time time; ++}; ++ ++struct rtc_param { ++ __u64 param; ++ union { ++ __u64 uvalue; ++ __s64 svalue; ++ __u64 ptr; ++ }; ++ __u32 index; ++ __u32 __pad; ++}; ++ ++struct rtc_class_ops { ++ int (*ioctl)(struct device *, unsigned int, long unsigned int); ++ int (*read_time)(struct device *, struct rtc_time *); ++ int (*set_time)(struct device *, struct rtc_time *); ++ int (*read_alarm)(struct device *, struct rtc_wkalrm *); ++ int (*set_alarm)(struct device *, struct rtc_wkalrm *); ++ int (*proc)(struct device *, struct seq_file *); ++ int (*alarm_irq_enable)(struct device *, unsigned int); ++ int (*read_offset)(struct device *, long int *); ++ int (*set_offset)(struct device *, long int); ++ int (*param_get)(struct device *, struct rtc_param *); ++ int (*param_set)(struct device *, struct rtc_param *); ++}; ++ ++struct rtc_device; ++ ++struct rtc_timer { ++ struct timerqueue_node node; ++ ktime_t period; ++ void (*func)(struct rtc_device *); ++ struct rtc_device *rtc; ++ int enabled; ++}; ++ ++struct rtc_device { ++ struct device dev; ++ struct module *owner; ++ int id; ++ const struct rtc_class_ops *ops; ++ struct mutex ops_lock; ++ struct cdev char_dev; ++ long unsigned int flags; ++ long unsigned int irq_data; ++ spinlock_t irq_lock; ++ wait_queue_head_t irq_queue; ++ struct fasync_struct *async_queue; ++ int irq_freq; ++ int max_user_freq; ++ struct timerqueue_head timerqueue; ++ struct rtc_timer aie_timer; ++ struct rtc_timer uie_rtctimer; ++ struct hrtimer pie_timer; ++ int pie_enabled; ++ struct work_struct irqwork; ++ long unsigned int set_offset_nsec; ++ long unsigned int features[1]; ++ time64_t range_min; ++ timeu64_t range_max; ++ time64_t start_secs; ++ time64_t offset_secs; ++ bool set_start_time; ++}; ++ ++struct trace_event_raw_alarmtimer_suspend { ++ struct trace_entry ent; ++ s64 expires; ++ unsigned char alarm_type; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_alarm_class { ++ struct trace_entry ent; ++ void *alarm; ++ unsigned char alarm_type; ++ s64 expires; ++ s64 now; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_alarmtimer_suspend {}; ++ ++struct trace_event_data_offsets_alarm_class {}; ++ ++typedef void (*btf_trace_alarmtimer_suspend)(void *, ktime_t, int); ++ ++typedef void (*btf_trace_alarmtimer_fired)(void *, struct alarm *, ktime_t); ++ ++typedef void (*btf_trace_alarmtimer_start)(void *, struct alarm *, ktime_t); ++ ++typedef void (*btf_trace_alarmtimer_cancel)(void *, struct alarm *, ktime_t); ++ ++struct alarm_base { ++ spinlock_t lock; ++ struct timerqueue_head timerqueue; ++ ktime_t (*get_ktime)(); ++ void (*get_timespec)(struct timespec64 *); ++ clockid_t base_clockid; ++}; ++ ++struct __kernel_old_itimerval { ++ struct __kernel_old_timeval it_interval; ++ struct __kernel_old_timeval it_value; ++}; ++ ++struct old_itimerval32 { ++ struct old_timeval32 it_interval; ++ struct old_timeval32 it_value; ++}; ++ ++enum { ++ Q_REQUEUE_PI_NONE = 0, ++ Q_REQUEUE_PI_IGNORE = 1, ++ Q_REQUEUE_PI_IN_PROGRESS = 2, ++ Q_REQUEUE_PI_WAIT = 3, ++ Q_REQUEUE_PI_DONE = 4, ++ Q_REQUEUE_PI_LOCKED = 5, ++}; ++ ++struct kallsym_iter { ++ loff_t pos; ++ loff_t pos_arch_end; ++ loff_t pos_mod_end; ++ loff_t pos_ftrace_mod_end; ++ loff_t pos_bpf_end; ++ long unsigned int value; ++ unsigned int nameoff; ++ char type; ++ char name[512]; ++ char module_name[56]; ++ int exported; ++ int show_value; ++}; ++ ++struct bpf_iter__ksym { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct kallsym_iter *ksym; ++ }; ++}; ++ ++enum cgroup_filetype { ++ CGROUP_FILE_PROCS = 0, ++ CGROUP_FILE_TASKS = 1, ++}; ++ ++struct cgroup_pidlist { ++ struct { ++ enum cgroup_filetype type; ++ struct pid_namespace *ns; ++ } key; ++ pid_t *list; ++ int length; ++ struct list_head links; ++ struct cgroup *owner; ++ struct delayed_work destroy_dwork; ++}; ++ ++enum cgroup1_param { ++ Opt_all = 0, ++ Opt_clone_children = 1, ++ Opt_cpuset_v2_mode = 2, ++ Opt_name = 3, ++ Opt_none = 4, ++ Opt_noprefix = 5, ++ Opt_release_agent = 6, ++ Opt_xattr = 7, ++ Opt_favordynmods___2 = 8, ++ Opt_nofavordynmods = 9, ++}; ++ ++enum freezer_state_flags { ++ CGROUP_FREEZER_ONLINE = 1, ++ CGROUP_FREEZING_SELF = 2, ++ CGROUP_FREEZING_PARENT = 4, ++ CGROUP_FROZEN = 8, ++ CGROUP_FREEZING = 6, ++}; ++ ++struct freezer { ++ struct cgroup_subsys_state css; ++ unsigned int state; ++}; ++ ++enum misc_res_type { ++ MISC_CG_RES_TYPES = 0, ++}; ++ ++struct misc_res { ++ long unsigned int max; ++ atomic_long_t usage; ++ atomic_long_t events; ++}; ++ ++struct misc_cg { ++ struct cgroup_subsys_state css; ++ struct cgroup_file events_file; ++ struct misc_res res[0]; ++}; ++ ++struct idmap_key { ++ bool map_up; ++ u32 id; ++ u32 count; ++}; ++ ++enum audit_nlgrps { ++ AUDIT_NLGRP_NONE = 0, ++ AUDIT_NLGRP_READLOG = 1, ++ __AUDIT_NLGRP_MAX = 2, ++}; ++ ++struct audit_status { ++ __u32 mask; ++ __u32 enabled; ++ __u32 failure; ++ __u32 pid; ++ __u32 rate_limit; ++ __u32 backlog_limit; ++ __u32 lost; ++ __u32 backlog; ++ union { ++ __u32 version; ++ __u32 feature_bitmap; ++ }; ++ __u32 backlog_wait_time; ++ __u32 backlog_wait_time_actual; ++}; ++ ++struct audit_features { ++ __u32 vers; ++ __u32 mask; ++ __u32 features; ++ __u32 lock; ++}; ++ ++struct audit_tty_status { ++ __u32 enabled; ++ __u32 log_passwd; ++}; ++ ++struct audit_sig_info { ++ uid_t uid; ++ pid_t pid; ++ char ctx[0]; ++}; ++ ++struct audit_net { ++ struct sock *sk; ++}; ++ ++struct auditd_connection { ++ struct pid *pid; ++ u32 portid; ++ struct net *net; ++ struct callback_head rcu; ++}; ++ ++struct audit_ctl_mutex { ++ struct mutex lock; ++ void *owner; ++}; ++ ++struct audit_buffer { ++ struct sk_buff *skb; ++ struct audit_context *ctx; ++ gfp_t gfp_mask; ++}; ++ ++struct audit_reply { ++ __u32 portid; ++ struct net *net; ++ struct sk_buff *skb; ++}; ++ ++struct audit_fsnotify_mark { ++ dev_t dev; ++ long unsigned int ino; ++ char *path; ++ struct fsnotify_mark mark; ++ struct audit_krule *rule; ++}; ++ ++struct seccomp_notif_sizes { ++ __u16 seccomp_notif; ++ __u16 seccomp_notif_resp; ++ __u16 seccomp_data; ++}; ++ ++struct seccomp_notif { ++ __u64 id; ++ __u32 pid; ++ __u32 flags; ++ struct seccomp_data data; ++}; ++ ++struct seccomp_notif_resp { ++ __u64 id; ++ __s64 val; ++ __s32 error; ++ __u32 flags; ++}; ++ ++struct seccomp_notif_addfd { ++ __u64 id; ++ __u32 flags; ++ __u32 srcfd; ++ __u32 newfd; ++ __u32 newfd_flags; ++}; ++ ++struct action_cache { ++ long unsigned int allow_native[8]; ++ long unsigned int allow_compat[8]; ++}; ++ ++struct notification; ++ ++struct seccomp_filter { ++ refcount_t refs; ++ refcount_t users; ++ bool log; ++ bool wait_killable_recv; ++ struct action_cache cache; ++ struct seccomp_filter *prev; ++ struct bpf_prog *prog; ++ struct notification *notif; ++ struct mutex notify_lock; ++ wait_queue_head_t wqh; ++}; ++ ++struct seccomp_metadata { ++ __u64 filter_off; ++ __u64 flags; ++}; ++ ++enum notify_state { ++ SECCOMP_NOTIFY_INIT = 0, ++ SECCOMP_NOTIFY_SENT = 1, ++ SECCOMP_NOTIFY_REPLIED = 2, ++}; ++ ++struct seccomp_knotif { ++ struct task_struct *task; ++ u64 id; ++ const struct seccomp_data *data; ++ enum notify_state state; ++ int error; ++ long int val; ++ u32 flags; ++ struct completion ready; ++ struct list_head list; ++ struct list_head addfd; ++}; ++ ++struct seccomp_kaddfd { ++ struct file *file; ++ int fd; ++ unsigned int flags; ++ __u32 ioctl_flags; ++ union { ++ bool setfd; ++ int ret; ++ }; ++ struct completion completion; ++ struct list_head list; ++}; ++ ++struct notification { ++ struct semaphore request; ++ u64 next_id; ++ struct list_head notifications; ++}; ++ ++struct seccomp_log_name { ++ u32 log; ++ const char *name; ++}; ++ ++enum ring_buffer_type { ++ RINGBUF_TYPE_DATA_TYPE_LEN_MAX = 28, ++ RINGBUF_TYPE_PADDING = 29, ++ RINGBUF_TYPE_TIME_EXTEND = 30, ++ RINGBUF_TYPE_TIME_STAMP = 31, ++}; ++ ++struct ring_buffer_per_cpu; ++ ++struct buffer_page; ++ ++struct ring_buffer_iter { ++ struct ring_buffer_per_cpu *cpu_buffer; ++ long unsigned int head; ++ long unsigned int next_event; ++ struct buffer_page *head_page; ++ struct buffer_page *cache_reader_page; ++ long unsigned int cache_read; ++ long unsigned int cache_pages_removed; ++ u64 read_stamp; ++ u64 page_stamp; ++ struct ring_buffer_event *event; ++ int missed_events; ++}; ++ ++struct rb_irq_work { ++ struct irq_work work; ++ wait_queue_head_t waiters; ++ wait_queue_head_t full_waiters; ++ long int wait_index; ++ bool waiters_pending; ++ bool full_waiters_pending; ++ bool wakeup_full; ++}; ++ ++struct trace_buffer { ++ unsigned int flags; ++ int cpus; ++ atomic_t record_disabled; ++ atomic_t resizing; ++ cpumask_var_t cpumask; ++ struct lock_class_key *reader_lock_key; ++ struct mutex mutex; ++ struct ring_buffer_per_cpu **buffers; ++ struct hlist_node node; ++ u64 (*clock)(); ++ struct rb_irq_work irq_work; ++ bool time_stamp_abs; ++}; ++ ++enum { ++ RB_LEN_TIME_EXTEND = 8, ++ RB_LEN_TIME_STAMP = 8, ++}; ++ ++struct buffer_data_page { ++ u64 time_stamp; ++ local_t commit; ++ unsigned char data[0]; ++}; ++ ++struct buffer_page { ++ struct list_head list; ++ local_t write; ++ unsigned int read; ++ local_t entries; ++ long unsigned int real_end; ++ struct buffer_data_page *page; ++}; ++ ++struct rb_event_info { ++ u64 ts; ++ u64 delta; ++ u64 before; ++ u64 after; ++ long unsigned int length; ++ struct buffer_page *tail_page; ++ int add_timestamp; ++}; ++ ++enum { ++ RB_ADD_STAMP_NONE = 0, ++ RB_ADD_STAMP_EXTEND = 2, ++ RB_ADD_STAMP_ABSOLUTE = 4, ++ RB_ADD_STAMP_FORCE = 8, ++}; ++ ++enum { ++ RB_CTX_TRANSITION = 0, ++ RB_CTX_NMI = 1, ++ RB_CTX_IRQ = 2, ++ RB_CTX_SOFTIRQ = 3, ++ RB_CTX_NORMAL = 4, ++ RB_CTX_MAX = 5, ++}; ++ ++struct rb_time_struct { ++ local64_t time; ++}; ++ ++typedef struct rb_time_struct rb_time_t; ++ ++struct ring_buffer_per_cpu { ++ int cpu; ++ atomic_t record_disabled; ++ atomic_t resize_disabled; ++ struct trace_buffer *buffer; ++ raw_spinlock_t reader_lock; ++ arch_spinlock_t lock; ++ struct lock_class_key lock_key; ++ struct buffer_data_page *free_page; ++ long unsigned int nr_pages; ++ unsigned int current_context; ++ struct list_head *pages; ++ struct buffer_page *head_page; ++ struct buffer_page *tail_page; ++ struct buffer_page *commit_page; ++ struct buffer_page *reader_page; ++ long unsigned int lost_events; ++ long unsigned int last_overrun; ++ long unsigned int nest; ++ local_t entries_bytes; ++ local_t entries; ++ local_t overrun; ++ local_t commit_overrun; ++ local_t dropped_events; ++ local_t committing; ++ local_t commits; ++ local_t pages_touched; ++ local_t pages_lost; ++ local_t pages_read; ++ long int last_pages_touch; ++ size_t shortest_full; ++ long unsigned int read; ++ long unsigned int read_bytes; ++ rb_time_t write_stamp; ++ rb_time_t before_stamp; ++ u64 event_stamp[5]; ++ u64 read_stamp; ++ long unsigned int pages_removed; ++ long int nr_pages_to_update; ++ struct list_head new_pages; ++ struct work_struct update_pages_work; ++ struct completion update_done; ++ struct rb_irq_work irq_work; ++}; ++ ++struct trace_bprintk_fmt { ++ struct list_head list; ++ const char *fmt; ++}; ++ ++struct tracing_map_entry { ++ u32 key; ++ struct tracing_map_elt *val; ++}; ++ ++struct trace_event_raw_thread_noise { ++ struct trace_entry ent; ++ char comm[16]; ++ u64 start; ++ u64 duration; ++ pid_t pid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_softirq_noise { ++ struct trace_entry ent; ++ u64 start; ++ u64 duration; ++ int vector; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_irq_noise { ++ struct trace_entry ent; ++ u64 start; ++ u64 duration; ++ u32 __data_loc_desc; ++ int vector; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_nmi_noise { ++ struct trace_entry ent; ++ u64 start; ++ u64 duration; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sample_threshold { ++ struct trace_entry ent; ++ u64 start; ++ u64 duration; ++ u64 interference; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_thread_noise {}; ++ ++struct trace_event_data_offsets_softirq_noise {}; ++ ++struct trace_event_data_offsets_irq_noise { ++ u32 desc; ++}; ++ ++struct trace_event_data_offsets_nmi_noise {}; ++ ++struct trace_event_data_offsets_sample_threshold {}; ++ ++typedef void (*btf_trace_thread_noise)(void *, struct task_struct *, u64, u64); ++ ++typedef void (*btf_trace_softirq_noise)(void *, int, u64, u64); ++ ++typedef void (*btf_trace_irq_noise)(void *, int, const char *, u64, u64); ++ ++typedef void (*btf_trace_nmi_noise)(void *, u64, u64); ++ ++typedef void (*btf_trace_sample_threshold)(void *, u64, u64, u64); ++ ++enum osnoise_options_index { ++ OSN_DEFAULTS = 0, ++ OSN_WORKLOAD = 1, ++ OSN_PANIC_ON_STOP = 2, ++ OSN_PREEMPT_DISABLE = 3, ++ OSN_IRQ_DISABLE = 4, ++ OSN_MAX = 5, ++}; ++ ++struct osnoise_instance { ++ struct list_head list; ++ struct trace_array *tr; ++}; ++ ++struct osn_nmi { ++ u64 count; ++ u64 delta_start; ++}; ++ ++struct osn_irq { ++ u64 count; ++ u64 arrival_time; ++ u64 delta_start; ++}; ++ ++struct osn_softirq { ++ u64 count; ++ u64 arrival_time; ++ u64 delta_start; ++}; ++ ++struct osn_thread { ++ u64 count; ++ u64 arrival_time; ++ u64 delta_start; ++}; ++ ++struct osnoise_variables { ++ struct task_struct *kthread; ++ bool sampling; ++ pid_t pid; ++ struct osn_nmi nmi; ++ struct osn_irq irq; ++ struct osn_softirq softirq; ++ struct osn_thread thread; ++ local_t int_counter; ++}; ++ ++struct timerlat_variables { ++ struct task_struct *kthread; ++ struct hrtimer timer; ++ u64 rel_period; ++ u64 abs_period; ++ bool tracing_thread; ++ u64 count; ++}; ++ ++struct osnoise_sample { ++ u64 runtime; ++ u64 noise; ++ u64 max_sample; ++ int hw_count; ++ int nmi_count; ++ int irq_count; ++ int softirq_count; ++ int thread_count; ++}; ++ ++struct timerlat_sample { ++ u64 timer_latency; ++ unsigned int seqnum; ++ int context; ++}; ++ ++struct osnoise_data { ++ u64 sample_period; ++ u64 sample_runtime; ++ u64 stop_tracing; ++ u64 stop_tracing_total; ++ u64 timerlat_period; ++ u64 print_stack; ++ int timerlat_tracer; ++ bool tainted; ++}; ++ ++struct trace_stack { ++ int stack_size; ++ int nr_entries; ++ long unsigned int calls[256]; ++}; ++ ++struct module_string { ++ struct list_head next; ++ struct module *module; ++ char *str; ++}; ++ ++enum { ++ FORMAT_HEADER = 1, ++ FORMAT_FIELD_SEPERATOR = 2, ++ FORMAT_PRINTFMT = 3, ++}; ++ ++struct boot_triggers { ++ const char *event; ++ char *trigger; ++}; ++ ++struct event_probe_data { ++ struct trace_event_file *file; ++ long unsigned int count; ++ int ref; ++ bool enable; ++}; ++ ++struct trace_dynamic_info { ++ u16 offset; ++ u16 len; ++}; ++ ++struct synth_field_desc { ++ const char *type; ++ const char *name; ++}; ++ ++struct synth_trace_event; ++ ++struct synth_event_trace_state { ++ struct trace_event_buffer fbuffer; ++ struct synth_trace_event *entry; ++ struct trace_buffer *buffer; ++ struct synth_event *event; ++ unsigned int cur_field; ++ unsigned int n_u64; ++ bool disabled; ++ bool add_next; ++ bool add_name; ++}; ++ ++union trace_synth_field { ++ u8 as_u8; ++ u16 as_u16; ++ u32 as_u32; ++ u64 as_u64; ++ struct trace_dynamic_info as_dynamic; ++}; ++ ++struct synth_trace_event { ++ struct trace_entry ent; ++ union trace_synth_field fields[0]; ++}; ++ ++enum { ++ SYNTH_ERR_BAD_NAME = 0, ++ SYNTH_ERR_INVALID_CMD = 1, ++ SYNTH_ERR_INVALID_DYN_CMD = 2, ++ SYNTH_ERR_EVENT_EXISTS = 3, ++ SYNTH_ERR_TOO_MANY_FIELDS = 4, ++ SYNTH_ERR_INCOMPLETE_TYPE = 5, ++ SYNTH_ERR_INVALID_TYPE = 6, ++ SYNTH_ERR_INVALID_FIELD = 7, ++ SYNTH_ERR_INVALID_ARRAY_SPEC = 8, ++}; ++ ++struct kprobe_trace_entry_head { ++ struct trace_entry ent; ++ long unsigned int ip; ++}; ++ ++struct kretprobe_trace_entry_head { ++ struct trace_entry ent; ++ long unsigned int func; ++ long unsigned int ret_ip; ++}; ++ ++struct trace_kprobe { ++ struct dyn_event devent; ++ struct kretprobe rp; ++ long unsigned int *nhit; ++ const char *symbol; ++ struct trace_probe tp; ++}; ++ ++enum uprobe_filter_ctx { ++ UPROBE_FILTER_REGISTER = 0, ++ UPROBE_FILTER_UNREGISTER = 1, ++ UPROBE_FILTER_MMAP = 2, ++}; ++ ++struct uprobe_consumer { ++ int (*handler)(struct uprobe_consumer *, struct pt_regs *); ++ int (*ret_handler)(struct uprobe_consumer *, long unsigned int, struct pt_regs *); ++ bool (*filter)(struct uprobe_consumer *, enum uprobe_filter_ctx, struct mm_struct *); ++ struct uprobe_consumer *next; ++}; ++ ++struct uprobe_trace_entry_head { ++ struct trace_entry ent; ++ long unsigned int vaddr[0]; ++}; ++ ++struct trace_uprobe { ++ struct dyn_event devent; ++ struct uprobe_consumer consumer; ++ struct path path; ++ struct inode *inode; ++ char *filename; ++ long unsigned int offset; ++ long unsigned int ref_ctr_offset; ++ long unsigned int nhit; ++ struct trace_probe tp; ++}; ++ ++struct uprobe_dispatch_data { ++ struct trace_uprobe *tu; ++ long unsigned int bp_addr; ++}; ++ ++struct uprobe_cpu_buffer { ++ struct mutex mutex; ++ void *buf; ++}; ++ ++typedef bool (*filter_func_t)(struct uprobe_consumer *, enum uprobe_filter_ctx, struct mm_struct *); ++ ++enum states_wwnr { ++ not_running_wwnr = 0, ++ running_wwnr = 1, ++ state_max_wwnr = 2, ++}; ++ ++enum events_wwnr { ++ switch_in_wwnr = 0, ++ switch_out_wwnr = 1, ++ wakeup_wwnr = 2, ++ event_max_wwnr = 3, ++}; ++ ++struct automaton_wwnr { ++ char *state_names[2]; ++ char *event_names[3]; ++ unsigned char function[6]; ++ unsigned char initial_state; ++ bool final_states[2]; ++}; ++ ++enum bpf_stats_type { ++ BPF_STATS_RUN_TIME = 0, ++}; ++ ++struct bpf_tracing_link { ++ struct bpf_tramp_link link; ++ enum bpf_attach_type attach_type; ++ struct bpf_trampoline *trampoline; ++ struct bpf_prog *tgt_prog; ++}; ++ ++enum bpf_audit { ++ BPF_AUDIT_LOAD = 0, ++ BPF_AUDIT_UNLOAD = 1, ++ BPF_AUDIT_MAX = 2, ++}; ++ ++struct bpf_prog_kstats { ++ u64 nsecs; ++ u64 cnt; ++ u64 misses; ++}; ++ ++struct bpf_raw_tp_link { ++ struct bpf_link link; ++ struct bpf_raw_event_map *btp; ++}; ++ ++struct bpf_perf_link { ++ struct bpf_link link; ++ struct file *perf_file; ++}; ++ ++typedef u64 (*btf_bpf_sys_bpf)(int, union bpf_attr *, u32); ++ ++typedef u64 (*btf_bpf_sys_close)(u32); ++ ++typedef u64 (*btf_bpf_kallsyms_lookup_name)(const char *, int, int, u64 *); ++ ++struct bpf_iter_seq_task_common { ++ struct pid_namespace *ns; ++ enum bpf_iter_task_type type; ++ u32 pid; ++ u32 pid_visiting; ++}; ++ ++struct bpf_iter_seq_task_info { ++ struct bpf_iter_seq_task_common common; ++ u32 tid; ++}; ++ ++struct bpf_iter__task { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct task_struct *task; ++ }; ++}; ++ ++struct bpf_iter_seq_task_file_info { ++ struct bpf_iter_seq_task_common common; ++ struct task_struct *task; ++ u32 tid; ++ u32 fd; ++}; ++ ++struct bpf_iter__task_file { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct task_struct *task; ++ }; ++ u32 fd; ++ union { ++ struct file *file; ++ }; ++}; ++ ++struct bpf_iter_seq_task_vma_info { ++ struct bpf_iter_seq_task_common common; ++ struct task_struct *task; ++ struct mm_struct *mm; ++ struct vm_area_struct *vma; ++ u32 tid; ++ long unsigned int prev_vm_start; ++ long unsigned int prev_vm_end; ++}; ++ ++enum bpf_task_vma_iter_find_op { ++ task_vma_iter_first_vma = 0, ++ task_vma_iter_next_vma = 1, ++ task_vma_iter_find_vma = 2, ++}; ++ ++struct bpf_iter__task_vma { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct task_struct *task; ++ }; ++ union { ++ struct vm_area_struct *vma; ++ }; ++}; ++ ++typedef u64 (*btf_bpf_find_vma)(struct task_struct *, u64, bpf_callback_t, void *, u64); ++ ++struct bpf_iter_seq_array_map_info { ++ struct bpf_map *map; ++ void *percpu_value_buf; ++ u32 index; ++}; ++ ++struct prog_poke_elem { ++ struct list_head list; ++ struct bpf_prog_aux *aux; ++}; ++ ++struct bpf_cgroup_storage_map { ++ struct bpf_map map; ++ spinlock_t lock; ++ struct rb_root root; ++ struct list_head list; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++typedef u64 (*btf_bpf_task_storage_get_recur)(struct bpf_map *, struct task_struct *, void *, u64, gfp_t); ++ ++typedef u64 (*btf_bpf_task_storage_get)(struct bpf_map *, struct task_struct *, void *, u64, gfp_t); ++ ++typedef u64 (*btf_bpf_task_storage_delete_recur)(struct bpf_map *, struct task_struct *); ++ ++typedef u64 (*btf_bpf_task_storage_delete)(struct bpf_map *, struct task_struct *); ++ ++struct bpf_mem_cache { ++ struct llist_head free_llist; ++ local_t active; ++ struct llist_head free_llist_extra; ++ struct irq_work refill_work; ++ struct obj_cgroup *objcg; ++ int unit_size; ++ int free_cnt; ++ int low_watermark; ++ int high_watermark; ++ int batch; ++ int percpu_size; ++ struct callback_head rcu; ++ struct llist_head free_by_rcu; ++ struct llist_head waiting_for_gp; ++ atomic_t call_rcu_in_progress; ++}; ++ ++struct bpf_mem_caches { ++ struct bpf_mem_cache cache[11]; ++}; ++ ++struct bpf_devmap_val { ++ __u32 ifindex; ++ union { ++ int fd; ++ __u32 id; ++ } bpf_prog; ++}; ++ ++struct xdp_dev_bulk_queue { ++ struct xdp_frame *q[16]; ++ struct list_head flush_node; ++ struct net_device *dev; ++ struct net_device *dev_rx; ++ struct bpf_prog *xdp_prog; ++ unsigned int count; ++}; ++ ++struct bpf_dtab; ++ ++struct bpf_dtab_netdev { ++ struct net_device *dev; ++ struct hlist_node index_hlist; ++ struct bpf_dtab *dtab; ++ struct bpf_prog *xdp_prog; ++ struct callback_head rcu; ++ unsigned int idx; ++ struct bpf_devmap_val val; ++}; ++ ++struct bpf_dtab { ++ struct bpf_map map; ++ struct bpf_dtab_netdev **netdev_map; ++ struct list_head list; ++ struct hlist_head *dev_index_head; ++ spinlock_t index_lock; ++ unsigned int items; ++ u32 n_buckets; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_netns_link { ++ struct bpf_link link; ++ enum bpf_attach_type type; ++ enum netns_bpf_attach_type netns_type; ++ struct net *net; ++ struct list_head node; ++}; ++ ++typedef u64 (*btf_bpf_cgrp_storage_get)(struct bpf_map *, struct cgroup *, void *, u64, gfp_t); ++ ++typedef u64 (*btf_bpf_cgrp_storage_delete)(struct bpf_map *, struct cgroup *); ++ ++struct bpf_cpumask { ++ cpumask_t cpumask; ++ struct callback_head rcu; ++ refcount_t usage; ++}; ++ ++struct uprobe { ++ struct rb_node rb_node; ++ refcount_t ref; ++ struct rw_semaphore register_rwsem; ++ struct rw_semaphore consumer_rwsem; ++ struct list_head pending_list; ++ struct uprobe_consumer *consumers; ++ struct inode *inode; ++ loff_t offset; ++ loff_t ref_ctr_offset; ++ long unsigned int flags; ++ struct arch_uprobe arch; ++}; ++ ++struct xol_area { ++ wait_queue_head_t wq; ++ atomic_t slot_count; ++ long unsigned int *bitmap; ++ struct vm_special_mapping xol_mapping; ++ struct page *pages[2]; ++ long unsigned int vaddr; ++}; ++ ++struct delayed_uprobe { ++ struct list_head list; ++ struct uprobe *uprobe; ++ struct mm_struct *mm; ++}; ++ ++struct __uprobe_key { ++ struct inode *inode; ++ loff_t offset; ++}; ++ ++struct map_info___2 { ++ struct map_info___2 *next; ++ struct mm_struct *mm; ++ long unsigned int vaddr; ++}; ++ ++struct static_key_mod { ++ struct static_key_mod *next; ++ struct jump_entry *entries; ++ struct module *mod; ++}; ++ ++struct static_key_deferred { ++ struct static_key key; ++ long unsigned int timeout; ++ struct delayed_work work; ++}; ++ ++struct wait_page_key { ++ struct folio *folio; ++ int bit_nr; ++ int page_match; ++}; ++ ++struct trace_event_raw_mm_filemap_op_page_cache { ++ struct trace_entry ent; ++ long unsigned int pfn; ++ long unsigned int i_ino; ++ long unsigned int index; ++ dev_t s_dev; ++ unsigned char order; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_filemap_set_wb_err { ++ struct trace_entry ent; ++ long unsigned int i_ino; ++ dev_t s_dev; ++ errseq_t errseq; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_file_check_and_advance_wb_err { ++ struct trace_entry ent; ++ struct file *file; ++ long unsigned int i_ino; ++ dev_t s_dev; ++ errseq_t old; ++ errseq_t new; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_mm_filemap_op_page_cache {}; ++ ++struct trace_event_data_offsets_filemap_set_wb_err {}; ++ ++struct trace_event_data_offsets_file_check_and_advance_wb_err {}; ++ ++typedef void (*btf_trace_mm_filemap_delete_from_page_cache)(void *, struct folio *); ++ ++typedef void (*btf_trace_mm_filemap_add_to_page_cache)(void *, struct folio *); ++ ++typedef void (*btf_trace_filemap_set_wb_err)(void *, struct address_space *, errseq_t); ++ ++typedef void (*btf_trace_file_check_and_advance_wb_err)(void *, struct file *, errseq_t); ++ ++enum behavior { ++ EXCLUSIVE = 0, ++ SHARED = 1, ++ DROP = 2, ++}; ++ ++enum lruvec_flags { ++ LRUVEC_CONGESTED = 0, ++}; ++ ++enum pgdat_flags { ++ PGDAT_DIRTY = 0, ++ PGDAT_WRITEBACK = 1, ++ PGDAT_RECLAIM_LOCKED = 2, ++}; ++ ++struct reclaim_stat { ++ unsigned int nr_dirty; ++ unsigned int nr_unqueued_dirty; ++ unsigned int nr_congested; ++ unsigned int nr_writeback; ++ unsigned int nr_immediate; ++ unsigned int nr_pageout; ++ unsigned int nr_activate[2]; ++ unsigned int nr_ref_keep; ++ unsigned int nr_unmap_fail; ++ unsigned int nr_lazyfree_fail; ++}; ++ ++struct trace_event_raw_mm_vmscan_kswapd_sleep { ++ struct trace_entry ent; ++ int nid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_vmscan_kswapd_wake { ++ struct trace_entry ent; ++ int nid; ++ int zid; ++ int order; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_vmscan_wakeup_kswapd { ++ struct trace_entry ent; ++ int nid; ++ int zid; ++ int order; ++ long unsigned int gfp_flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_vmscan_direct_reclaim_begin_template { ++ struct trace_entry ent; ++ int order; ++ long unsigned int gfp_flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_vmscan_direct_reclaim_end_template { ++ struct trace_entry ent; ++ long unsigned int nr_reclaimed; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_shrink_slab_start { ++ struct trace_entry ent; ++ struct shrinker *shr; ++ void *shrink; ++ int nid; ++ long int nr_objects_to_shrink; ++ long unsigned int gfp_flags; ++ long unsigned int cache_items; ++ long long unsigned int delta; ++ long unsigned int total_scan; ++ int priority; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_shrink_slab_end { ++ struct trace_entry ent; ++ struct shrinker *shr; ++ int nid; ++ void *shrink; ++ long int unused_scan; ++ long int new_scan; ++ int retval; ++ long int total_scan; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_vmscan_lru_isolate { ++ struct trace_entry ent; ++ int highest_zoneidx; ++ int order; ++ long unsigned int nr_requested; ++ long unsigned int nr_scanned; ++ long unsigned int nr_skipped; ++ long unsigned int nr_taken; ++ unsigned int isolate_mode; ++ int lru; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_vmscan_write_folio { ++ struct trace_entry ent; ++ long unsigned int pfn; ++ int reclaim_flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_vmscan_lru_shrink_inactive { ++ struct trace_entry ent; ++ int nid; ++ long unsigned int nr_scanned; ++ long unsigned int nr_reclaimed; ++ long unsigned int nr_dirty; ++ long unsigned int nr_writeback; ++ long unsigned int nr_congested; ++ long unsigned int nr_immediate; ++ unsigned int nr_activate0; ++ unsigned int nr_activate1; ++ long unsigned int nr_ref_keep; ++ long unsigned int nr_unmap_fail; ++ int priority; ++ int reclaim_flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_vmscan_lru_shrink_active { ++ struct trace_entry ent; ++ int nid; ++ long unsigned int nr_taken; ++ long unsigned int nr_active; ++ long unsigned int nr_deactivated; ++ long unsigned int nr_referenced; ++ int priority; ++ int reclaim_flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_vmscan_node_reclaim_begin { ++ struct trace_entry ent; ++ int nid; ++ int order; ++ long unsigned int gfp_flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_vmscan_throttled { ++ struct trace_entry ent; ++ int nid; ++ int usec_timeout; ++ int usec_delayed; ++ int reason; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_mm_vmscan_kswapd_sleep {}; ++ ++struct trace_event_data_offsets_mm_vmscan_kswapd_wake {}; ++ ++struct trace_event_data_offsets_mm_vmscan_wakeup_kswapd {}; ++ ++struct trace_event_data_offsets_mm_vmscan_direct_reclaim_begin_template {}; ++ ++struct trace_event_data_offsets_mm_vmscan_direct_reclaim_end_template {}; ++ ++struct trace_event_data_offsets_mm_shrink_slab_start {}; ++ ++struct trace_event_data_offsets_mm_shrink_slab_end {}; ++ ++struct trace_event_data_offsets_mm_vmscan_lru_isolate {}; ++ ++struct trace_event_data_offsets_mm_vmscan_write_folio {}; ++ ++struct trace_event_data_offsets_mm_vmscan_lru_shrink_inactive {}; ++ ++struct trace_event_data_offsets_mm_vmscan_lru_shrink_active {}; ++ ++struct trace_event_data_offsets_mm_vmscan_node_reclaim_begin {}; ++ ++struct trace_event_data_offsets_mm_vmscan_throttled {}; ++ ++typedef void (*btf_trace_mm_vmscan_kswapd_sleep)(void *, int); ++ ++typedef void (*btf_trace_mm_vmscan_kswapd_wake)(void *, int, int, int); ++ ++typedef void (*btf_trace_mm_vmscan_wakeup_kswapd)(void *, int, int, int, gfp_t); ++ ++typedef void (*btf_trace_mm_vmscan_direct_reclaim_begin)(void *, int, gfp_t); ++ ++typedef void (*btf_trace_mm_vmscan_memcg_reclaim_begin)(void *, int, gfp_t); ++ ++typedef void (*btf_trace_mm_vmscan_memcg_softlimit_reclaim_begin)(void *, int, gfp_t); ++ ++typedef void (*btf_trace_mm_vmscan_direct_reclaim_end)(void *, long unsigned int); ++ ++typedef void (*btf_trace_mm_vmscan_memcg_reclaim_end)(void *, long unsigned int); ++ ++typedef void (*btf_trace_mm_vmscan_memcg_softlimit_reclaim_end)(void *, long unsigned int); ++ ++typedef void (*btf_trace_mm_shrink_slab_start)(void *, struct shrinker *, struct shrink_control *, long int, long unsigned int, long long unsigned int, long unsigned int, int); ++ ++typedef void (*btf_trace_mm_shrink_slab_end)(void *, struct shrinker *, int, int, long int, long int, long int); ++ ++typedef void (*btf_trace_mm_vmscan_lru_isolate)(void *, int, int, long unsigned int, long unsigned int, long unsigned int, long unsigned int, isolate_mode_t, int); ++ ++typedef void (*btf_trace_mm_vmscan_write_folio)(void *, struct folio *); ++ ++typedef void (*btf_trace_mm_vmscan_lru_shrink_inactive)(void *, int, long unsigned int, long unsigned int, struct reclaim_stat *, int, int); ++ ++typedef void (*btf_trace_mm_vmscan_lru_shrink_active)(void *, int, long unsigned int, long unsigned int, long unsigned int, long unsigned int, int, int); ++ ++typedef void (*btf_trace_mm_vmscan_node_reclaim_begin)(void *, int, int, gfp_t); ++ ++typedef void (*btf_trace_mm_vmscan_node_reclaim_end)(void *, long unsigned int); ++ ++typedef void (*btf_trace_mm_vmscan_throttled)(void *, int, int, int, int); ++ ++struct scan_control { ++ long unsigned int nr_to_reclaim; ++ nodemask_t *nodemask; ++ struct mem_cgroup *target_mem_cgroup; ++ long unsigned int anon_cost; ++ long unsigned int file_cost; ++ unsigned int may_deactivate: 2; ++ unsigned int force_deactivate: 1; ++ unsigned int skipped_deactivate: 1; ++ unsigned int may_writepage: 1; ++ unsigned int may_unmap: 1; ++ unsigned int may_swap: 1; ++ unsigned int proactive: 1; ++ unsigned int memcg_low_reclaim: 1; ++ unsigned int memcg_low_skipped: 1; ++ unsigned int hibernation_mode: 1; ++ unsigned int compaction_ready: 1; ++ unsigned int cache_trim_mode: 1; ++ unsigned int file_is_tiny: 1; ++ unsigned int no_demotion: 1; ++ s8 order; ++ s8 priority; ++ s8 reclaim_idx; ++ gfp_t gfp_mask; ++ long unsigned int nr_scanned; ++ long unsigned int nr_reclaimed; ++ struct { ++ unsigned int dirty; ++ unsigned int unqueued_dirty; ++ unsigned int congested; ++ unsigned int writeback; ++ unsigned int immediate; ++ unsigned int file_taken; ++ unsigned int taken; ++ } nr; ++ struct reclaim_state reclaim_state; ++}; ++ ++typedef enum { ++ PAGE_KEEP = 0, ++ PAGE_ACTIVATE = 1, ++ PAGE_SUCCESS = 2, ++ PAGE_CLEAN = 3, ++} pageout_t; ++ ++enum folio_references { ++ FOLIOREF_RECLAIM = 0, ++ FOLIOREF_RECLAIM_CLEAN = 1, ++ FOLIOREF_KEEP = 2, ++ FOLIOREF_ACTIVATE = 3, ++}; ++ ++enum scan_balance { ++ SCAN_EQUAL = 0, ++ SCAN_FRACT = 1, ++ SCAN_ANON = 2, ++ SCAN_FILE = 3, ++}; ++ ++struct ctrl_pos { ++ long unsigned int refaulted; ++ long unsigned int total; ++ int gain; ++}; ++ ++enum { ++ MEMCG_LRU_NOP = 0, ++ MEMCG_LRU_HEAD = 1, ++ MEMCG_LRU_TAIL = 2, ++ MEMCG_LRU_OLD = 3, ++ MEMCG_LRU_YOUNG = 4, ++}; ++ ++struct kmalloc_info_struct { ++ const char *name[4]; ++ unsigned int size; ++}; ++ ++struct trace_event_raw_kmem_cache_alloc { ++ struct trace_entry ent; ++ long unsigned int call_site; ++ const void *ptr; ++ size_t bytes_req; ++ size_t bytes_alloc; ++ long unsigned int gfp_flags; ++ int node; ++ bool accounted; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_kmalloc { ++ struct trace_entry ent; ++ long unsigned int call_site; ++ const void *ptr; ++ size_t bytes_req; ++ size_t bytes_alloc; ++ long unsigned int gfp_flags; ++ int node; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_kfree { ++ struct trace_entry ent; ++ long unsigned int call_site; ++ const void *ptr; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_kmem_cache_free { ++ struct trace_entry ent; ++ long unsigned int call_site; ++ const void *ptr; ++ u32 __data_loc_name; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_page_free { ++ struct trace_entry ent; ++ long unsigned int pfn; ++ unsigned int order; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_page_free_batched { ++ struct trace_entry ent; ++ long unsigned int pfn; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_page_alloc { ++ struct trace_entry ent; ++ long unsigned int pfn; ++ unsigned int order; ++ long unsigned int gfp_flags; ++ int migratetype; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_page { ++ struct trace_entry ent; ++ long unsigned int pfn; ++ unsigned int order; ++ int migratetype; ++ int percpu_refill; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_page_pcpu_drain { ++ struct trace_entry ent; ++ long unsigned int pfn; ++ unsigned int order; ++ int migratetype; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_mm_page_alloc_extfrag { ++ struct trace_entry ent; ++ long unsigned int pfn; ++ int alloc_order; ++ int fallback_order; ++ int alloc_migratetype; ++ int fallback_migratetype; ++ int change_ownership; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_rss_stat { ++ struct trace_entry ent; ++ unsigned int mm_id; ++ unsigned int curr; ++ int member; ++ long int size; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_kmem_cache_alloc {}; ++ ++struct trace_event_data_offsets_kmalloc {}; ++ ++struct trace_event_data_offsets_kfree {}; ++ ++struct trace_event_data_offsets_kmem_cache_free { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_mm_page_free {}; ++ ++struct trace_event_data_offsets_mm_page_free_batched {}; ++ ++struct trace_event_data_offsets_mm_page_alloc {}; ++ ++struct trace_event_data_offsets_mm_page {}; ++ ++struct trace_event_data_offsets_mm_page_pcpu_drain {}; ++ ++struct trace_event_data_offsets_mm_page_alloc_extfrag {}; ++ ++struct trace_event_data_offsets_rss_stat {}; ++ ++typedef void (*btf_trace_kmem_cache_alloc)(void *, long unsigned int, const void *, struct kmem_cache *, gfp_t, int); ++ ++typedef void (*btf_trace_kmalloc)(void *, long unsigned int, const void *, size_t, size_t, gfp_t, int); ++ ++typedef void (*btf_trace_kfree)(void *, long unsigned int, const void *); ++ ++typedef void (*btf_trace_kmem_cache_free)(void *, long unsigned int, const void *, const struct kmem_cache *); ++ ++typedef void (*btf_trace_mm_page_free)(void *, struct page *, unsigned int); ++ ++typedef void (*btf_trace_mm_page_free_batched)(void *, struct page *); ++ ++typedef void (*btf_trace_mm_page_alloc)(void *, struct page *, unsigned int, gfp_t, int); ++ ++typedef void (*btf_trace_mm_page_alloc_zone_locked)(void *, struct page *, unsigned int, int, int); ++ ++typedef void (*btf_trace_mm_page_pcpu_drain)(void *, struct page *, unsigned int, int); ++ ++typedef void (*btf_trace_mm_page_alloc_extfrag)(void *, struct page *, int, int, int, int); ++ ++typedef void (*btf_trace_rss_stat)(void *, struct mm_struct *, int); ++ ++struct list_lru_memcg { ++ struct callback_head rcu; ++ struct list_lru_one node[0]; ++}; ++ ++struct list_lru_memcg_table { ++ struct list_lru_memcg *mlru; ++ struct mem_cgroup *memcg; ++}; ++ ++typedef struct { ++ long unsigned int pd; ++} hugepd_t; ++ ++struct follow_page_context { ++ struct dev_pagemap *pgmap; ++ unsigned int page_mask; ++}; ++ ++struct mlock_fbatch { ++ local_lock_t lock; ++ struct folio_batch fbatch; ++}; ++ ++typedef void (*online_page_callback_t)(struct page *, unsigned int); ++ ++enum { ++ ONLINE_POLICY_CONTIG_ZONES = 0, ++ ONLINE_POLICY_AUTO_MOVABLE = 1, ++}; ++ ++struct auto_movable_stats { ++ long unsigned int kernel_early_pages; ++ long unsigned int movable_pages; ++}; ++ ++struct auto_movable_group_stats { ++ long unsigned int movable_pages; ++ long unsigned int req_kernel_early_pages; ++}; ++ ++union swap_header { ++ struct { ++ char reserved[4086]; ++ char magic[10]; ++ } magic; ++ struct { ++ char bootbits[1024]; ++ __u32 version; ++ __u32 last_page; ++ __u32 nr_badpages; ++ unsigned char sws_uuid[16]; ++ unsigned char sws_volume[16]; ++ __u32 padding[117]; ++ __u32 badpages[1]; ++ } info; ++}; ++ ++struct swap_extent { ++ struct rb_node rb_node; ++ long unsigned int start_page; ++ long unsigned int nr_pages; ++ sector_t start_block; ++}; ++ ++struct vmemmap_remap_walk { ++ void (*remap_pte)(pte_t *, long unsigned int, struct vmemmap_remap_walk *); ++ long unsigned int nr_walked; ++ struct page *reuse_page; ++ long unsigned int reuse_addr; ++ struct list_head *vmemmap_pages; ++}; ++ ++struct mm_slot { ++ struct hlist_node hash; ++ struct list_head mm_node; ++ struct mm_struct *mm; ++}; ++ ++struct trace_event_raw_ksm_scan_template { ++ struct trace_entry ent; ++ int seq; ++ u32 rmap_entries; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ksm_enter_exit_template { ++ struct trace_entry ent; ++ void *mm; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ksm_merge_one_page { ++ struct trace_entry ent; ++ long unsigned int pfn; ++ void *rmap_item; ++ void *mm; ++ int err; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ksm_merge_with_ksm_page { ++ struct trace_entry ent; ++ void *ksm_page; ++ long unsigned int pfn; ++ void *rmap_item; ++ void *mm; ++ int err; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ksm_remove_ksm_page { ++ struct trace_entry ent; ++ long unsigned int pfn; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_ksm_remove_rmap_item { ++ struct trace_entry ent; ++ long unsigned int pfn; ++ void *rmap_item; ++ void *mm; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_ksm_scan_template {}; ++ ++struct trace_event_data_offsets_ksm_enter_exit_template {}; ++ ++struct trace_event_data_offsets_ksm_merge_one_page {}; ++ ++struct trace_event_data_offsets_ksm_merge_with_ksm_page {}; ++ ++struct trace_event_data_offsets_ksm_remove_ksm_page {}; ++ ++struct trace_event_data_offsets_ksm_remove_rmap_item {}; ++ ++typedef void (*btf_trace_ksm_start_scan)(void *, int, u32); ++ ++typedef void (*btf_trace_ksm_stop_scan)(void *, int, u32); ++ ++typedef void (*btf_trace_ksm_enter)(void *, void *); ++ ++typedef void (*btf_trace_ksm_exit)(void *, void *); ++ ++typedef void (*btf_trace_ksm_merge_one_page)(void *, long unsigned int, void *, void *, int); ++ ++typedef void (*btf_trace_ksm_merge_with_ksm_page)(void *, void *, long unsigned int, void *, void *, int); ++ ++typedef void (*btf_trace_ksm_remove_ksm_page)(void *, long unsigned int); ++ ++typedef void (*btf_trace_ksm_remove_rmap_item)(void *, long unsigned int, void *, void *); ++ ++struct ksm_rmap_item; ++ ++struct ksm_mm_slot { ++ struct mm_slot slot; ++ struct ksm_rmap_item *rmap_list; ++}; ++ ++struct ksm_stable_node; ++ ++struct ksm_rmap_item { ++ struct ksm_rmap_item *rmap_list; ++ union { ++ struct anon_vma *anon_vma; ++ int nid; ++ }; ++ struct mm_struct *mm; ++ long unsigned int address; ++ unsigned int oldchecksum; ++ union { ++ struct rb_node node; ++ struct { ++ struct ksm_stable_node *head; ++ struct hlist_node hlist; ++ }; ++ }; ++}; ++ ++struct ksm_scan { ++ struct ksm_mm_slot *mm_slot; ++ long unsigned int address; ++ struct ksm_rmap_item **rmap_list; ++ long unsigned int seqnr; ++}; ++ ++struct ksm_stable_node { ++ union { ++ struct rb_node node; ++ struct { ++ struct list_head *head; ++ struct { ++ struct hlist_node hlist_dup; ++ struct list_head list; ++ }; ++ }; ++ }; ++ struct hlist_head hlist; ++ union { ++ long unsigned int kpfn; ++ long unsigned int chain_prune_time; ++ }; ++ int rmap_hlist_len; ++ int nid; ++}; ++ ++enum get_ksm_page_flags { ++ GET_KSM_PAGE_NOLOCK = 0, ++ GET_KSM_PAGE_LOCK = 1, ++ GET_KSM_PAGE_TRYLOCK = 2, ++}; ++ ++struct memory_tier { ++ struct list_head list; ++ struct list_head memory_types; ++ int adistance_start; ++ struct device dev; ++ nodemask_t lower_tier_mask; ++}; ++ ++struct memory_dev_type { ++ struct list_head tier_sibiling; ++ int adistance; ++ nodemask_t nodes; ++ struct kref kref; ++}; ++ ++struct demotion_nodes { ++ nodemask_t preferred; ++}; ++ ++struct node_memory_type_map { ++ struct memory_dev_type *memtype; ++ int map_count; ++}; ++ ++enum vmpressure_levels { ++ VMPRESSURE_LOW = 0, ++ VMPRESSURE_MEDIUM = 1, ++ VMPRESSURE_CRITICAL = 2, ++ VMPRESSURE_NUM_LEVELS = 3, ++}; ++ ++enum vmpressure_modes { ++ VMPRESSURE_NO_PASSTHROUGH = 0, ++ VMPRESSURE_HIERARCHY = 1, ++ VMPRESSURE_LOCAL = 2, ++ VMPRESSURE_NUM_MODES = 3, ++}; ++ ++struct vmpressure_event { ++ struct eventfd_ctx *efd; ++ enum vmpressure_levels level; ++ enum vmpressure_modes mode; ++ struct list_head node; ++}; ++ ++enum page_ext_flags { ++ PAGE_EXT_OWNER = 0, ++ PAGE_EXT_OWNER_ALLOCATED = 1, ++}; ++ ++struct page_owner { ++ short unsigned int order; ++ short int last_migrate_reason; ++ gfp_t gfp_mask; ++ depot_stack_handle_t handle; ++ depot_stack_handle_t free_handle; ++ u64 ts_nsec; ++ u64 free_ts_nsec; ++ char comm[16]; ++ pid_t pid; ++ pid_t tgid; ++}; ++ ++struct zbud_pool { ++ spinlock_t lock; ++ union { ++ struct list_head buddied; ++ struct list_head unbuddied[63]; ++ }; ++ struct list_head lru; ++ u64 pages_nr; ++ struct zpool *zpool; ++ const struct zpool_ops *zpool_ops; ++}; ++ ++struct zbud_header { ++ struct list_head buddy; ++ struct list_head lru; ++ unsigned int first_chunks; ++ unsigned int last_chunks; ++ bool under_reclaim; ++}; ++ ++enum buddy___2 { ++ FIRST___2 = 0, ++ LAST___2 = 1, ++}; ++ ++struct balloon_dev_info { ++ long unsigned int isolated_pages; ++ spinlock_t pages_lock; ++ struct list_head pages; ++ int (*migratepage)(struct balloon_dev_info *, struct page *, struct page *, enum migrate_mode); ++}; ++ ++struct damon_young_walk_private { ++ long unsigned int *folio_sz; ++ bool young; ++}; ++ ++struct damon_sysfs_region { ++ struct kobject kobj; ++ struct damon_addr_range ar; ++}; ++ ++struct damon_sysfs_regions { ++ struct kobject kobj; ++ struct damon_sysfs_region **regions_arr; ++ int nr; ++}; ++ ++struct damon_sysfs_target { ++ struct kobject kobj; ++ struct damon_sysfs_regions *regions; ++ int pid; ++}; ++ ++struct damon_sysfs_targets { ++ struct kobject kobj; ++ struct damon_sysfs_target **targets_arr; ++ int nr; ++}; ++ ++struct damon_sysfs_intervals { ++ struct kobject kobj; ++ long unsigned int sample_us; ++ long unsigned int aggr_us; ++ long unsigned int update_us; ++}; ++ ++struct damon_sysfs_attrs { ++ struct kobject kobj; ++ struct damon_sysfs_intervals *intervals; ++ struct damon_sysfs_ul_range *nr_regions_range; ++}; ++ ++struct damon_sysfs_context { ++ struct kobject kobj; ++ enum damon_ops_id ops_id; ++ struct damon_sysfs_attrs *attrs; ++ struct damon_sysfs_targets *targets; ++ struct damon_sysfs_schemes *schemes; ++}; ++ ++struct damon_sysfs_contexts { ++ struct kobject kobj; ++ struct damon_sysfs_context **contexts_arr; ++ int nr; ++}; ++ ++struct damon_sysfs_kdamond { ++ struct kobject kobj; ++ struct damon_sysfs_contexts *contexts; ++ struct damon_ctx *damon_ctx; ++}; ++ ++enum damon_sysfs_cmd { ++ DAMON_SYSFS_CMD_ON = 0, ++ DAMON_SYSFS_CMD_OFF = 1, ++ DAMON_SYSFS_CMD_COMMIT = 2, ++ DAMON_SYSFS_CMD_UPDATE_SCHEMES_STATS = 3, ++ DAMON_SYSFS_CMD_UPDATE_SCHEMES_TRIED_REGIONS = 4, ++ DAMON_SYSFS_CMD_CLEAR_SCHEMES_TRIED_REGIONS = 5, ++ NR_DAMON_SYSFS_CMDS = 6, ++}; ++ ++struct damon_sysfs_cmd_request { ++ enum damon_sysfs_cmd cmd; ++ struct damon_sysfs_kdamond *kdamond; ++}; ++ ++struct damon_sysfs_kdamonds { ++ struct kobject kobj; ++ struct damon_sysfs_kdamond **kdamonds_arr; ++ int nr; ++}; ++ ++struct damon_sysfs_ui_dir { ++ struct kobject kobj; ++ struct damon_sysfs_kdamonds *kdamonds; ++}; ++ ++enum { ++ BAD_STACK = -1, ++ NOT_STACK = 0, ++ GOOD_FRAME = 1, ++ GOOD_STACK = 2, ++}; ++ ++struct file_clone_range { ++ __s64 src_fd; ++ __u64 src_offset; ++ __u64 src_length; ++ __u64 dest_offset; ++}; ++ ++struct fsxattr { ++ __u32 fsx_xflags; ++ __u32 fsx_extsize; ++ __u32 fsx_nextents; ++ __u32 fsx_projid; ++ __u32 fsx_cowextsize; ++ unsigned char fsx_pad[8]; ++}; ++ ++struct space_resv { ++ __s16 l_type; ++ __s16 l_whence; ++ __s64 l_start; ++ __s64 l_len; ++ __s32 l_sysid; ++ __u32 l_pid; ++ __s32 l_pad[4]; ++}; ++ ++struct linux_dirent64 { ++ u64 d_ino; ++ s64 d_off; ++ short unsigned int d_reclen; ++ unsigned char d_type; ++ char d_name[0]; ++}; ++ ++struct old_linux_dirent { ++ long unsigned int d_ino; ++ long unsigned int d_offset; ++ short unsigned int d_namlen; ++ char d_name[1]; ++}; ++ ++struct readdir_callback { ++ struct dir_context ctx; ++ struct old_linux_dirent *dirent; ++ int result; ++}; ++ ++struct linux_dirent { ++ long unsigned int d_ino; ++ long unsigned int d_off; ++ short unsigned int d_reclen; ++ char d_name[1]; ++}; ++ ++struct getdents_callback { ++ struct dir_context ctx; ++ struct linux_dirent *current_dir; ++ int prev_reclen; ++ int count; ++ int error; ++}; ++ ++struct getdents_callback64 { ++ struct dir_context ctx; ++ struct linux_dirent64 *current_dir; ++ int prev_reclen; ++ int count; ++ int error; ++}; ++ ++struct compat_old_linux_dirent { ++ compat_ulong_t d_ino; ++ compat_ulong_t d_offset; ++ short unsigned int d_namlen; ++ char d_name[1]; ++}; ++ ++struct compat_readdir_callback { ++ struct dir_context ctx; ++ struct compat_old_linux_dirent *dirent; ++ int result; ++}; ++ ++struct compat_linux_dirent { ++ compat_ulong_t d_ino; ++ compat_ulong_t d_off; ++ short unsigned int d_reclen; ++ char d_name[1]; ++}; ++ ++struct compat_getdents_callback { ++ struct dir_context ctx; ++ struct compat_linux_dirent *current_dir; ++ int prev_reclen; ++ int count; ++ int error; ++}; ++ ++struct dentry_stat_t { ++ long int nr_dentry; ++ long int nr_unused; ++ long int age_limit; ++ long int want_pages; ++ long int nr_negative; ++ long int dummy; ++}; ++ ++struct external_name { ++ union { ++ atomic_t count; ++ struct callback_head head; ++ } u; ++ unsigned char name[0]; ++}; ++ ++enum d_walk_ret { ++ D_WALK_CONTINUE = 0, ++ D_WALK_QUIT = 1, ++ D_WALK_NORETRY = 2, ++ D_WALK_SKIP = 3, ++}; ++ ++struct check_mount { ++ struct vfsmount *mnt; ++ unsigned int mounted; ++}; ++ ++struct select_data { ++ struct dentry *start; ++ union { ++ long int found; ++ struct dentry *victim; ++ }; ++ struct list_head dispose; ++}; ++ ++typedef long int pcp_op_T_____9; ++ ++struct simple_transaction_argresp { ++ ssize_t size; ++ char data[0]; ++}; ++ ++struct simple_attr { ++ int (*get)(void *, u64 *); ++ int (*set)(void *, u64); ++ char get_buf[24]; ++ char set_buf[24]; ++ void *data; ++ const char *fmt; ++ struct mutex mutex; ++}; ++ ++typedef int splice_direct_actor(struct pipe_inode_info *, struct splice_desc *); ++ ++struct postprocess_bh_ctx { ++ struct work_struct work; ++ struct buffer_head *bh; ++}; ++ ++struct bh_lru { ++ struct buffer_head *bhs[16]; ++}; ++ ++struct bh_accounting { ++ int nr; ++ int ratelimit; ++}; ++ ++struct epoll_event { ++ __poll_t events; ++ __u64 data; ++}; ++ ++struct wake_irq; ++ ++struct wakeup_source { ++ const char *name; ++ int id; ++ struct list_head entry; ++ spinlock_t lock; ++ struct wake_irq *wakeirq; ++ struct timer_list timer; ++ long unsigned int timer_expires; ++ ktime_t total_time; ++ ktime_t max_time; ++ ktime_t last_time; ++ ktime_t start_prevent_time; ++ ktime_t prevent_sleep_time; ++ long unsigned int event_count; ++ long unsigned int active_count; ++ long unsigned int relax_count; ++ long unsigned int expire_count; ++ long unsigned int wakeup_count; ++ struct device *dev; ++ bool active: 1; ++ bool autosleep_enabled: 1; ++}; ++ ++struct epoll_filefd { ++ struct file *file; ++ int fd; ++} __attribute__((packed)); ++ ++struct epitem; ++ ++struct eppoll_entry { ++ struct eppoll_entry *next; ++ struct epitem *base; ++ wait_queue_entry_t wait; ++ wait_queue_head_t *whead; ++}; ++ ++struct eventpoll; ++ ++struct epitem { ++ union { ++ struct rb_node rbn; ++ struct callback_head rcu; ++ }; ++ struct list_head rdllink; ++ struct epitem *next; ++ struct epoll_filefd ffd; ++ bool dying; ++ struct eppoll_entry *pwqlist; ++ struct eventpoll *ep; ++ struct hlist_node fllink; ++ struct wakeup_source *ws; ++ struct epoll_event event; ++}; ++ ++struct eventpoll { ++ struct mutex mtx; ++ wait_queue_head_t wq; ++ wait_queue_head_t poll_wait; ++ struct list_head rdllist; ++ rwlock_t lock; ++ struct rb_root_cached rbr; ++ struct epitem *ovflist; ++ struct wakeup_source *ws; ++ struct user_struct *user; ++ struct file *file; ++ u64 gen; ++ struct hlist_head refs; ++ refcount_t refcount; ++ unsigned int napi_id; ++}; ++ ++struct ep_pqueue { ++ poll_table pt; ++ struct epitem *epi; ++}; ++ ++struct epitems_head { ++ struct hlist_head epitems; ++ struct epitems_head *next; ++}; ++ ++struct kioctx; ++ ++struct kioctx_table { ++ struct callback_head rcu; ++ unsigned int nr; ++ struct kioctx *table[0]; ++}; ++ ++typedef __kernel_ulong_t aio_context_t; ++ ++enum { ++ IOCB_CMD_PREAD = 0, ++ IOCB_CMD_PWRITE = 1, ++ IOCB_CMD_FSYNC = 2, ++ IOCB_CMD_FDSYNC = 3, ++ IOCB_CMD_POLL = 5, ++ IOCB_CMD_NOOP = 6, ++ IOCB_CMD_PREADV = 7, ++ IOCB_CMD_PWRITEV = 8, ++}; ++ ++struct io_event { ++ __u64 data; ++ __u64 obj; ++ __s64 res; ++ __s64 res2; ++}; ++ ++struct iocb { ++ __u64 aio_data; ++ __kernel_rwf_t aio_rw_flags; ++ __u32 aio_key; ++ __u16 aio_lio_opcode; ++ __s16 aio_reqprio; ++ __u32 aio_fildes; ++ __u64 aio_buf; ++ __u64 aio_nbytes; ++ __s64 aio_offset; ++ __u64 aio_reserved2; ++ __u32 aio_flags; ++ __u32 aio_resfd; ++}; ++ ++typedef u32 compat_aio_context_t; ++ ++typedef int kiocb_cancel_fn(struct kiocb *); ++ ++struct aio_ring { ++ unsigned int id; ++ unsigned int nr; ++ unsigned int head; ++ unsigned int tail; ++ unsigned int magic; ++ unsigned int compat_features; ++ unsigned int incompat_features; ++ unsigned int header_length; ++ struct io_event io_events[0]; ++}; ++ ++struct kioctx_cpu; ++ ++struct ctx_rq_wait; ++ ++struct kioctx { ++ struct percpu_ref users; ++ atomic_t dead; ++ struct percpu_ref reqs; ++ long unsigned int user_id; ++ struct kioctx_cpu *cpu; ++ unsigned int req_batch; ++ unsigned int max_reqs; ++ unsigned int nr_events; ++ long unsigned int mmap_base; ++ long unsigned int mmap_size; ++ struct page **ring_pages; ++ long int nr_pages; ++ struct rcu_work free_rwork; ++ struct ctx_rq_wait *rq_wait; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct { ++ atomic_t reqs_available; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ }; ++ struct { ++ spinlock_t ctx_lock; ++ struct list_head active_reqs; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ }; ++ struct { ++ struct mutex ring_lock; ++ wait_queue_head_t wait; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ }; ++ struct { ++ unsigned int tail; ++ unsigned int completed_events; ++ spinlock_t completion_lock; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ }; ++ struct page *internal_pages[8]; ++ struct file *aio_ring_file; ++ unsigned int id; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct kioctx_cpu { ++ unsigned int reqs_available; ++}; ++ ++struct ctx_rq_wait { ++ struct completion comp; ++ atomic_t count; ++}; ++ ++struct fsync_iocb { ++ struct file *file; ++ struct work_struct work; ++ bool datasync; ++ struct cred *creds; ++}; ++ ++struct poll_iocb { ++ struct file *file; ++ struct wait_queue_head *head; ++ __poll_t events; ++ bool cancelled; ++ bool work_scheduled; ++ bool work_need_resched; ++ struct wait_queue_entry wait; ++ struct work_struct work; ++}; ++ ++struct aio_kiocb { ++ union { ++ struct file *ki_filp; ++ struct kiocb rw; ++ struct fsync_iocb fsync; ++ struct poll_iocb poll; ++ }; ++ struct kioctx *ki_ctx; ++ kiocb_cancel_fn *ki_cancel; ++ struct io_event ki_res; ++ struct list_head ki_list; ++ refcount_t ki_refcnt; ++ struct eventfd_ctx *ki_eventfd; ++}; ++ ++struct aio_poll_table { ++ struct poll_table_struct pt; ++ struct aio_kiocb *iocb; ++ bool queued; ++ int error; ++}; ++ ++struct __aio_sigset { ++ const sigset_t *sigmask; ++ size_t sigsetsize; ++}; ++ ++struct __compat_aio_sigset { ++ compat_uptr_t sigmask; ++ compat_size_t sigsetsize; ++}; ++ ++struct fsverity_read_metadata_arg { ++ __u64 metadata_type; ++ __u64 offset; ++ __u64 length; ++ __u64 buf_ptr; ++ __u64 __reserved; ++}; ++ ++struct elf_thread_core_info___2 { ++ struct elf_thread_core_info___2 *next; ++ struct task_struct *task; ++ struct elf_prstatus prstatus; ++ struct memelfnote notes[0]; ++}; ++ ++struct elf_note_info___2 { ++ struct elf_thread_core_info___2 *thread; ++ struct memelfnote psinfo; ++ struct memelfnote signote; ++ struct memelfnote auxv; ++ struct memelfnote files; ++ siginfo_t csigdata; ++ size_t size; ++ int thread_notes; ++}; ++ ++struct core_name { ++ char *corename; ++ int used; ++ int size; ++}; ++ ++struct iomap_ioend { ++ struct list_head io_list; ++ u16 io_type; ++ u16 io_flags; ++ u32 io_folios; ++ struct inode *io_inode; ++ size_t io_size; ++ loff_t io_offset; ++ sector_t io_sector; ++ struct bio *io_bio; ++ struct bio io_inline_bio; ++}; ++ ++struct iomap_writepage_ctx; ++ ++struct iomap_writeback_ops { ++ int (*map_blocks)(struct iomap_writepage_ctx *, struct inode *, loff_t); ++ int (*prepare_ioend)(struct iomap_ioend *, int); ++ void (*discard_folio)(struct folio *, loff_t); ++}; ++ ++struct iomap_writepage_ctx { ++ struct iomap iomap; ++ struct iomap_ioend *ioend; ++ const struct iomap_writeback_ops *ops; ++}; ++ ++struct iomap_page { ++ atomic_t read_bytes_pending; ++ atomic_t write_bytes_pending; ++ spinlock_t uptodate_lock; ++ long unsigned int uptodate[0]; ++}; ++ ++struct iomap_readpage_ctx { ++ struct folio *cur_folio; ++ bool cur_folio_in_bio; ++ struct bio *bio; ++ struct readahead_control *rac; ++}; ++ ++enum { ++ QUOTA_NL_C_UNSPEC = 0, ++ QUOTA_NL_C_WARNING = 1, ++ __QUOTA_NL_C_MAX = 2, ++}; ++ ++enum { ++ QUOTA_NL_A_UNSPEC = 0, ++ QUOTA_NL_A_QTYPE = 1, ++ QUOTA_NL_A_EXCESS_ID = 2, ++ QUOTA_NL_A_WARNING = 3, ++ QUOTA_NL_A_DEV_MAJOR = 4, ++ QUOTA_NL_A_DEV_MINOR = 5, ++ QUOTA_NL_A_CAUSED_ID = 6, ++ QUOTA_NL_A_PAD = 7, ++ __QUOTA_NL_A_MAX = 8, ++}; ++ ++enum resctrl_conf_type { ++ CDP_NONE = 0, ++ CDP_CODE = 1, ++ CDP_DATA = 2, ++}; ++ ++struct pid_entry { ++ const char *name; ++ unsigned int len; ++ umode_t mode; ++ const struct inode_operations *iop; ++ const struct file_operations *fop; ++ union proc_op op; ++}; ++ ++struct limit_names { ++ const char *name; ++ const char *unit; ++}; ++ ++struct map_files_info { ++ long unsigned int start; ++ long unsigned int end; ++ fmode_t mode; ++}; ++ ++struct timers_private { ++ struct pid *pid; ++ struct task_struct *task; ++ struct sighand_struct *sighand; ++ struct pid_namespace *ns; ++ long unsigned int flags; ++}; ++ ++struct tgid_iter { ++ unsigned int tgid; ++ struct task_struct *task; ++}; ++ ++struct sysctl_alias { ++ const char *kernel_param; ++ const char *sysctl_param; ++}; ++ ++struct kernfs_global_locks { ++ struct mutex open_file_mutex[1024]; ++}; ++ ++struct kernfs_open_node { ++ struct callback_head callback_head; ++ atomic_t event; ++ wait_queue_head_t poll; ++ struct list_head files; ++ unsigned int nr_mmapped; ++ unsigned int nr_to_release; ++}; ++ ++struct configfs_buffer { ++ size_t count; ++ loff_t pos; ++ char *page; ++ struct configfs_item_operations *ops; ++ struct mutex mutex; ++ int needs_read_fill; ++ bool read_in_progress; ++ bool write_in_progress; ++ char *bin_buffer; ++ int bin_buffer_size; ++ int cb_max_size; ++ struct config_item *item; ++ struct module *owner; ++ union { ++ struct configfs_attribute *attr; ++ struct configfs_bin_attribute *bin_attr; ++ }; ++}; ++ ++typedef __kernel_mode_t mode_t; ++ ++struct orlov_stats { ++ __u64 free_clusters; ++ __u32 free_inodes; ++ __u32 used_dirs; ++}; ++ ++typedef short unsigned int __kernel_uid16_t; ++ ++typedef short unsigned int __kernel_gid16_t; ++ ++typedef __kernel_uid16_t uid16_t; ++ ++typedef __kernel_gid16_t gid16_t; ++ ++struct ext4_io_submit { ++ struct writeback_control *io_wbc; ++ struct bio *io_bio; ++ ext4_io_end_t *io_end; ++ sector_t io_next_block; ++}; ++ ++struct mpage_da_data { ++ struct inode *inode; ++ struct writeback_control *wbc; ++ unsigned int can_map: 1; ++ long unsigned int first_page; ++ long unsigned int next_page; ++ long unsigned int last_page; ++ struct ext4_map_blocks map; ++ struct ext4_io_submit io_submit; ++ unsigned int do_map: 1; ++ unsigned int scanned_until_end: 1; ++ unsigned int journalled_more_data: 1; ++}; ++ ++struct ext4_free_data { ++ struct list_head efd_list; ++ struct rb_node efd_node; ++ ext4_group_t efd_group; ++ ext4_grpblk_t efd_start_cluster; ++ ext4_grpblk_t efd_count; ++ tid_t efd_tid; ++}; ++ ++enum { ++ MB_INODE_PA = 0, ++ MB_GROUP_PA = 1, ++}; ++ ++struct ext4_buddy { ++ struct page *bd_buddy_page; ++ void *bd_buddy; ++ struct page *bd_bitmap_page; ++ void *bd_bitmap; ++ struct ext4_group_info *bd_info; ++ struct super_block *bd_sb; ++ __u16 bd_blkbits; ++ ext4_group_t bd_group; ++}; ++ ++struct sg { ++ struct ext4_group_info info; ++ ext4_grpblk_t counters[18]; ++}; ++ ++enum { ++ BLOCK_BITMAP = 0, ++ INODE_BITMAP = 1, ++ INODE_TABLE = 2, ++ GROUP_TABLE_COUNT = 3, ++}; ++ ++struct ext4_rcu_ptr { ++ struct callback_head rcu; ++ void *ptr; ++}; ++ ++struct ext4_new_flex_group_data { ++ struct ext4_new_group_data *groups; ++ __u16 *bg_flags; ++ ext4_group_t count; ++}; ++ ++enum { ++ attr_noop = 0, ++ attr_delayed_allocation_blocks = 1, ++ attr_session_write_kbytes = 2, ++ attr_lifetime_write_kbytes = 3, ++ attr_reserved_clusters = 4, ++ attr_sra_exceeded_retry_limit = 5, ++ attr_inode_readahead = 6, ++ attr_trigger_test_error = 7, ++ attr_first_error_time = 8, ++ attr_last_error_time = 9, ++ attr_feature = 10, ++ attr_pointer_ui = 11, ++ attr_pointer_ul = 12, ++ attr_pointer_u64 = 13, ++ attr_pointer_u8 = 14, ++ attr_pointer_string = 15, ++ attr_pointer_atomic = 16, ++ attr_journal_task = 17, ++}; ++ ++enum { ++ ptr_explicit = 0, ++ ptr_ext4_sb_info_offset = 1, ++ ptr_ext4_super_block_offset = 2, ++}; ++ ++struct ext4_attr { ++ struct attribute attr; ++ short int attr_id; ++ short int attr_ptr; ++ short unsigned int attr_size; ++ union { ++ int offset; ++ void *explicit_ptr; ++ } u; ++}; ++ ++struct ext4_fc_tl { ++ __le16 fc_tag; ++ __le16 fc_len; ++}; ++ ++struct ext4_fc_head { ++ __le32 fc_features; ++ __le32 fc_tid; ++}; ++ ++struct ext4_fc_add_range { ++ __le32 fc_ino; ++ __u8 fc_ex[12]; ++}; ++ ++struct ext4_fc_del_range { ++ __le32 fc_ino; ++ __le32 fc_lblk; ++ __le32 fc_len; ++}; ++ ++struct ext4_fc_dentry_info { ++ __le32 fc_parent_ino; ++ __le32 fc_ino; ++ __u8 fc_dname[0]; ++}; ++ ++struct ext4_fc_inode { ++ __le32 fc_ino; ++ __u8 fc_raw_inode[0]; ++}; ++ ++struct ext4_fc_tail { ++ __le32 fc_tid; ++ __le32 fc_crc; ++}; ++ ++enum { ++ EXT4_FC_STATUS_OK = 0, ++ EXT4_FC_STATUS_INELIGIBLE = 1, ++ EXT4_FC_STATUS_SKIPPED = 2, ++ EXT4_FC_STATUS_FAILED = 3, ++}; ++ ++struct ext4_fc_dentry_update { ++ int fcd_op; ++ int fcd_parent; ++ int fcd_ino; ++ struct qstr fcd_name; ++ unsigned char fcd_iname[32]; ++ struct list_head fcd_list; ++ struct list_head fcd_dilist; ++}; ++ ++struct __track_dentry_update_args { ++ struct dentry *dentry; ++ int op; ++}; ++ ++struct __track_range_args { ++ ext4_lblk_t start; ++ ext4_lblk_t end; ++}; ++ ++struct dentry_info_args { ++ int parent_ino; ++ int dname_len; ++ int ino; ++ int inode_len; ++ char *dname; ++}; ++ ++struct ext4_fc_tl_mem { ++ u16 fc_tag; ++ u16 fc_len; ++}; ++ ++typedef struct { ++ __le16 e_tag; ++ __le16 e_perm; ++ __le32 e_id; ++} ext4_acl_entry; ++ ++typedef struct { ++ __le32 a_version; ++} ext4_acl_header; ++ ++struct jbd2_journal_revoke_header_s { ++ journal_header_t r_header; ++ __be32 r_count; ++}; ++ ++typedef struct jbd2_journal_revoke_header_s jbd2_journal_revoke_header_t; ++ ++struct recovery_info { ++ tid_t start_transaction; ++ tid_t end_transaction; ++ int nr_replays; ++ int nr_revokes; ++ int nr_revoke_hits; ++}; ++ ++struct jbd2_revoke_table_s { ++ int hash_size; ++ int hash_shift; ++ struct list_head *hash_table; ++}; ++ ++struct jbd2_revoke_record_s { ++ struct list_head hash; ++ tid_t sequence; ++ long long unsigned int blocknr; ++}; ++ ++struct getdents_callback___2 { ++ struct dir_context ctx; ++ char *name; ++ u64 ino; ++ int found; ++ int sequence; ++}; ++ ++struct args_protover { ++ __u32 version; ++}; ++ ++struct args_protosubver { ++ __u32 sub_version; ++}; ++ ++struct args_openmount { ++ __u32 devid; ++}; ++ ++struct args_ready { ++ __u32 token; ++}; ++ ++struct args_fail { ++ __u32 token; ++ __s32 status; ++}; ++ ++struct args_setpipefd { ++ __s32 pipefd; ++}; ++ ++struct args_timeout { ++ __u64 timeout; ++}; ++ ++struct args_requester { ++ __u32 uid; ++ __u32 gid; ++}; ++ ++struct args_expire { ++ __u32 how; ++}; ++ ++struct args_askumount { ++ __u32 may_umount; ++}; ++ ++struct args_in { ++ __u32 type; ++}; ++ ++struct args_out { ++ __u32 devid; ++ __u32 magic; ++}; ++ ++struct args_ismountpoint { ++ union { ++ struct args_in in; ++ struct args_out out; ++ }; ++}; ++ ++struct autofs_dev_ioctl { ++ __u32 ver_major; ++ __u32 ver_minor; ++ __u32 size; ++ __s32 ioctlfd; ++ union { ++ struct args_protover protover; ++ struct args_protosubver protosubver; ++ struct args_openmount openmount; ++ struct args_ready ready; ++ struct args_fail fail; ++ struct args_setpipefd setpipefd; ++ struct args_timeout timeout; ++ struct args_requester requester; ++ struct args_expire expire; ++ struct args_askumount askumount; ++ struct args_ismountpoint ismountpoint; ++ }; ++ char path[0]; ++}; ++ ++enum { ++ AUTOFS_DEV_IOCTL_VERSION_CMD = 113, ++ AUTOFS_DEV_IOCTL_PROTOVER_CMD = 114, ++ AUTOFS_DEV_IOCTL_PROTOSUBVER_CMD = 115, ++ AUTOFS_DEV_IOCTL_OPENMOUNT_CMD = 116, ++ AUTOFS_DEV_IOCTL_CLOSEMOUNT_CMD = 117, ++ AUTOFS_DEV_IOCTL_READY_CMD = 118, ++ AUTOFS_DEV_IOCTL_FAIL_CMD = 119, ++ AUTOFS_DEV_IOCTL_SETPIPEFD_CMD = 120, ++ AUTOFS_DEV_IOCTL_CATATONIC_CMD = 121, ++ AUTOFS_DEV_IOCTL_TIMEOUT_CMD = 122, ++ AUTOFS_DEV_IOCTL_REQUESTER_CMD = 123, ++ AUTOFS_DEV_IOCTL_EXPIRE_CMD = 124, ++ AUTOFS_DEV_IOCTL_ASKUMOUNT_CMD = 125, ++ AUTOFS_DEV_IOCTL_ISMOUNTPOINT_CMD = 126, ++}; ++ ++typedef int (*ioctl_fn)(struct file *, struct autofs_sb_info *, struct autofs_dev_ioctl *); ++ ++struct btrfs_csums { ++ u16 size; ++ const char name[10]; ++ const char driver[12]; ++}; ++ ++struct root_name_map { ++ u64 id; ++ char name[16]; ++}; ++ ++struct falloc_range { ++ struct list_head list; ++ u64 start; ++ u64 len; ++}; ++ ++enum { ++ RANGE_BOUNDARY_WRITTEN_EXTENT = 0, ++ RANGE_BOUNDARY_PREALLOC_EXTENT = 1, ++ RANGE_BOUNDARY_HOLE = 2, ++}; ++ ++struct btrfs_bio_ctrl { ++ struct btrfs_bio *bbio; ++ enum btrfs_compression_type compress_type; ++ u32 len_to_oe_boundary; ++ blk_opf_t opf; ++ btrfs_bio_end_io_t end_io_func; ++ struct writeback_control *wbc; ++}; ++ ++struct processed_extent { ++ struct btrfs_inode *inode; ++ u64 start; ++ u64 end; ++ bool uptodate; ++}; ++ ++struct fiemap_cache { ++ u64 offset; ++ u64 phys; ++ u64 len; ++ u32 flags; ++ bool cached; ++}; ++ ++struct btrfs_ioctl_qgroup_limit_args { ++ __u64 qgroupid; ++ struct btrfs_qgroup_limit lim; ++}; ++ ++struct btrfs_ioctl_vol_args_v2 { ++ __s64 fd; ++ __u64 transid; ++ __u64 flags; ++ union { ++ struct { ++ __u64 size; ++ struct btrfs_qgroup_inherit *qgroup_inherit; ++ }; ++ __u64 unused[4]; ++ }; ++ union { ++ char name[4040]; ++ __u64 devid; ++ __u64 subvolid; ++ }; ++}; ++ ++struct btrfs_ioctl_scrub_args { ++ __u64 devid; ++ __u64 start; ++ __u64 end; ++ __u64 flags; ++ struct btrfs_scrub_progress progress; ++ __u64 unused[109]; ++}; ++ ++struct btrfs_ioctl_dev_info_args { ++ __u64 devid; ++ __u8 uuid[16]; ++ __u64 bytes_used; ++ __u64 total_bytes; ++ __u8 fsid[16]; ++ __u64 unused[377]; ++ __u8 path[1024]; ++}; ++ ++struct btrfs_ioctl_fs_info_args { ++ __u64 max_id; ++ __u64 num_devices; ++ __u8 fsid[16]; ++ __u32 nodesize; ++ __u32 sectorsize; ++ __u32 clone_alignment; ++ __u16 csum_type; ++ __u16 csum_size; ++ __u64 flags; ++ __u64 generation; ++ __u8 metadata_uuid[16]; ++ __u8 reserved[944]; ++}; ++ ++struct btrfs_ioctl_feature_flags { ++ __u64 compat_flags; ++ __u64 compat_ro_flags; ++ __u64 incompat_flags; ++}; ++ ++struct btrfs_ioctl_ino_lookup_args { ++ __u64 treeid; ++ __u64 objectid; ++ char name[4080]; ++}; ++ ++struct btrfs_ioctl_ino_lookup_user_args { ++ __u64 dirid; ++ __u64 treeid; ++ char name[256]; ++ char path[3824]; ++}; ++ ++struct btrfs_ioctl_search_key { ++ __u64 tree_id; ++ __u64 min_objectid; ++ __u64 max_objectid; ++ __u64 min_offset; ++ __u64 max_offset; ++ __u64 min_transid; ++ __u64 max_transid; ++ __u32 min_type; ++ __u32 max_type; ++ __u32 nr_items; ++ __u32 unused; ++ __u64 unused1; ++ __u64 unused2; ++ __u64 unused3; ++ __u64 unused4; ++}; ++ ++struct btrfs_ioctl_search_header { ++ __u64 transid; ++ __u64 objectid; ++ __u64 offset; ++ __u32 type; ++ __u32 len; ++}; ++ ++struct btrfs_ioctl_search_args { ++ struct btrfs_ioctl_search_key key; ++ char buf[3992]; ++}; ++ ++struct btrfs_ioctl_search_args_v2 { ++ struct btrfs_ioctl_search_key key; ++ __u64 buf_size; ++ __u64 buf[0]; ++}; ++ ++struct btrfs_ioctl_space_info { ++ __u64 flags; ++ __u64 total_bytes; ++ __u64 used_bytes; ++}; ++ ++struct btrfs_ioctl_space_args { ++ __u64 space_slots; ++ __u64 total_spaces; ++ struct btrfs_ioctl_space_info spaces[0]; ++}; ++ ++struct btrfs_ioctl_ino_path_args { ++ __u64 inum; ++ __u64 size; ++ __u64 reserved[4]; ++ __u64 fspath; ++}; ++ ++struct btrfs_ioctl_logical_ino_args { ++ __u64 logical; ++ __u64 size; ++ __u64 reserved[3]; ++ __u64 flags; ++ __u64 inodes; ++}; ++ ++struct btrfs_ioctl_quota_ctl_args { ++ __u64 cmd; ++ __u64 status; ++}; ++ ++struct btrfs_ioctl_quota_rescan_args { ++ __u64 flags; ++ __u64 progress; ++ __u64 reserved[6]; ++}; ++ ++struct btrfs_ioctl_qgroup_assign_args { ++ __u64 assign; ++ __u64 src; ++ __u64 dst; ++}; ++ ++struct btrfs_ioctl_qgroup_create_args { ++ __u64 create; ++ __u64 qgroupid; ++}; ++ ++struct btrfs_ioctl_timespec { ++ __u64 sec; ++ __u32 nsec; ++}; ++ ++struct btrfs_ioctl_received_subvol_args { ++ char uuid[16]; ++ __u64 stransid; ++ __u64 rtransid; ++ struct btrfs_ioctl_timespec stime; ++ struct btrfs_ioctl_timespec rtime; ++ __u64 flags; ++ __u64 reserved[16]; ++}; ++ ++struct btrfs_ioctl_get_subvol_info_args { ++ __u64 treeid; ++ char name[256]; ++ __u64 parent_id; ++ __u64 dirid; ++ __u64 generation; ++ __u64 flags; ++ __u8 uuid[16]; ++ __u8 parent_uuid[16]; ++ __u8 received_uuid[16]; ++ __u64 ctransid; ++ __u64 otransid; ++ __u64 stransid; ++ __u64 rtransid; ++ struct btrfs_ioctl_timespec ctime; ++ struct btrfs_ioctl_timespec otime; ++ struct btrfs_ioctl_timespec stime; ++ struct btrfs_ioctl_timespec rtime; ++ __u64 reserved[8]; ++}; ++ ++struct btrfs_ioctl_get_subvol_rootref_args { ++ __u64 min_treeid; ++ struct { ++ __u64 treeid; ++ __u64 dirid; ++ } rootref[255]; ++ __u8 num_items; ++ __u8 align[7]; ++}; ++ ++struct btrfs_ioctl_timespec_32 { ++ __u64 sec; ++ __u32 nsec; ++} __attribute__((packed)); ++ ++struct btrfs_ioctl_received_subvol_args_32 { ++ char uuid[16]; ++ __u64 stransid; ++ __u64 rtransid; ++ struct btrfs_ioctl_timespec_32 stime; ++ struct btrfs_ioctl_timespec_32 rtime; ++ __u64 flags; ++ __u64 reserved[16]; ++}; ++ ++struct btrfs_ioctl_send_args_32 { ++ __s64 send_fd; ++ __u64 clone_sources_count; ++ compat_uptr_t clone_sources; ++ __u64 parent_root; ++ __u64 flags; ++ __u32 version; ++ __u8 reserved[28]; ++} __attribute__((packed)); ++ ++struct btrfs_ioctl_encoded_io_args_32 { ++ compat_uptr_t iov; ++ compat_ulong_t iovcnt; ++ __s64 offset; ++ __u64 flags; ++ __u64 len; ++ __u64 unencoded_len; ++ __u64 unencoded_offset; ++ __u32 compression; ++ __u32 encryption; ++ __u8 reserved[64]; ++}; ++ ++struct btrfs_free_space_entry { ++ __le64 offset; ++ __le64 bytes; ++ __u8 type; ++} __attribute__((packed)); ++ ++struct btrfs_free_space_header { ++ struct btrfs_disk_key location; ++ __le64 generation; ++ __le64 num_entries; ++ __le64 num_bitmaps; ++} __attribute__((packed)); ++ ++struct btrfs_trim_range { ++ u64 start; ++ u64 bytes; ++ struct list_head list; ++}; ++ ++struct workspace___3 { ++ void *mem; ++ size_t size; ++ char *buf; ++ unsigned int level; ++ unsigned int req_level; ++ long unsigned int last_used; ++ struct list_head list; ++ struct list_head lru_list; ++ zstd_in_buffer in_buf; ++ zstd_out_buffer out_buf; ++}; ++ ++struct zstd_workspace_manager { ++ const struct btrfs_compress_op *ops; ++ spinlock_t lock; ++ struct list_head lru_list; ++ struct list_head idle_ws[15]; ++ long unsigned int active_map; ++ wait_queue_head_t wait; ++ struct timer_list timer; ++}; ++ ++struct btrfs_async_delayed_work { ++ struct btrfs_delayed_root *delayed_root; ++ int nr; ++ struct btrfs_work work; ++}; ++ ++struct btrfs_qgroup_status_item { ++ __le64 version; ++ __le64 generation; ++ __le64 flags; ++ __le64 rescan; ++}; ++ ++struct btrfs_qgroup_info_item { ++ __le64 generation; ++ __le64 rfer; ++ __le64 rfer_cmpr; ++ __le64 excl; ++ __le64 excl_cmpr; ++}; ++ ++struct btrfs_qgroup_limit_item { ++ __le64 flags; ++ __le64 max_rfer; ++ __le64 max_excl; ++ __le64 rsv_rfer; ++ __le64 rsv_excl; ++}; ++ ++struct btrfs_qgroup_swapped_block { ++ struct rb_node node; ++ int level; ++ bool trace_leaf; ++ u64 subvol_bytenr; ++ u64 subvol_generation; ++ u64 reloc_bytenr; ++ u64 reloc_generation; ++ u64 last_snapshot; ++ struct btrfs_key first_key; ++}; ++ ++struct btrfs_qgroup_list { ++ struct list_head next_group; ++ struct list_head next_member; ++ struct btrfs_qgroup *group; ++ struct btrfs_qgroup *member; ++}; ++ ++struct btrfs_stripe_hash { ++ struct list_head hash_list; ++ spinlock_t lock; ++}; ++ ++struct btrfs_stripe_hash_table { ++ struct list_head stripe_cache; ++ spinlock_t cache_lock; ++ int cache_size; ++ struct btrfs_stripe_hash table[0]; ++}; ++ ++struct sector_ptr { ++ struct page *page; ++ unsigned int pgoff: 24; ++ unsigned int uptodate: 8; ++}; ++ ++struct btrfs_plug_cb { ++ struct blk_plug_cb cb; ++ struct btrfs_fs_info *info; ++ struct list_head rbio_list; ++ struct work_struct work; ++}; ++ ++struct btrfs_free_space_info { ++ __le32 extent_count; ++ __le32 flags; ++}; ++ ++struct reserve_ticket { ++ u64 bytes; ++ int error; ++ bool steal; ++ struct list_head list; ++ wait_queue_head_t wait; ++}; ++ ++struct btrfs_failed_bio { ++ struct btrfs_bio *bbio; ++ int num_copies; ++ atomic_t repair_count; ++}; ++ ++struct async_submit_bio { ++ struct btrfs_bio *bbio; ++ struct btrfs_io_context *bioc; ++ struct btrfs_io_stripe smap; ++ int mirror_num; ++ struct btrfs_work work; ++}; ++ ++struct pstore_ftrace_record { ++ long unsigned int ip; ++ long unsigned int parent_ip; ++ u64 ts; ++}; ++ ++struct pstore_private { ++ struct list_head list; ++ struct dentry *dentry; ++ struct pstore_record *record; ++ size_t total_size; ++}; ++ ++struct pstore_ftrace_seq_data { ++ const void *ptr; ++ size_t off; ++ size_t size; ++}; ++ ++enum { ++ Opt_kmsg_bytes = 0, ++ Opt_err___10 = 1, ++}; ++ ++struct msg_msgseg { ++ struct msg_msgseg *next; ++}; ++ ++struct sem; ++ ++struct sem_queue; ++ ++struct sem_undo; ++ ++struct semid_ds { ++ struct ipc_perm sem_perm; ++ __kernel_old_time_t sem_otime; ++ __kernel_old_time_t sem_ctime; ++ struct sem *sem_base; ++ struct sem_queue *sem_pending; ++ struct sem_queue **sem_pending_last; ++ struct sem_undo *undo; ++ short unsigned int sem_nsems; ++}; ++ ++struct sem { ++ int semval; ++ struct pid *sempid; ++ spinlock_t lock; ++ struct list_head pending_alter; ++ struct list_head pending_const; ++ time64_t sem_otime; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct sem_queue { ++ struct list_head list; ++ struct task_struct *sleeper; ++ struct sem_undo *undo; ++ struct pid *pid; ++ int status; ++ struct sembuf *sops; ++ struct sembuf *blocking; ++ int nsops; ++ bool alter; ++ bool dupsop; ++}; ++ ++struct sem_undo { ++ struct list_head list_proc; ++ struct callback_head rcu; ++ struct sem_undo_list *ulp; ++ struct list_head list_id; ++ int semid; ++ short int *semadj; ++}; ++ ++struct semid64_ds { ++ struct ipc64_perm sem_perm; ++ long int sem_otime; ++ long int sem_ctime; ++ long unsigned int sem_nsems; ++ long unsigned int __unused3; ++ long unsigned int __unused4; ++}; ++ ++struct seminfo { ++ int semmap; ++ int semmni; ++ int semmns; ++ int semmnu; ++ int semmsl; ++ int semopm; ++ int semume; ++ int semusz; ++ int semvmx; ++ int semaem; ++}; ++ ++struct sem_undo_list { ++ refcount_t refcnt; ++ spinlock_t lock; ++ struct list_head list_proc; ++}; ++ ++struct compat_semid64_ds { ++ struct compat_ipc64_perm sem_perm; ++ compat_ulong_t sem_otime; ++ compat_ulong_t sem_otime_high; ++ compat_ulong_t sem_ctime; ++ compat_ulong_t sem_ctime_high; ++ compat_ulong_t sem_nsems; ++ compat_ulong_t __unused3; ++ compat_ulong_t __unused4; ++}; ++ ++struct sem_array { ++ struct kern_ipc_perm sem_perm; ++ time64_t sem_ctime; ++ struct list_head pending_alter; ++ struct list_head pending_const; ++ struct list_head list_id; ++ int sem_nsems; ++ int complex_count; ++ unsigned int use_global_lock; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct sem sems[0]; ++}; ++ ++struct compat_semid_ds { ++ struct compat_ipc_perm sem_perm; ++ old_time32_t sem_otime; ++ old_time32_t sem_ctime; ++ compat_uptr_t sem_base; ++ compat_uptr_t sem_pending; ++ compat_uptr_t sem_pending_last; ++ compat_uptr_t undo; ++ short unsigned int sem_nsems; ++}; ++ ++struct keyring_read_iterator_context { ++ size_t buflen; ++ size_t count; ++ key_serial_t *buffer; ++}; ++ ++struct big_key_payload { ++ u8 *data; ++ struct path path; ++ size_t length; ++}; ++ ++enum tpm2key_actions { ++ ACT_tpm2_key_parent = 0, ++ ACT_tpm2_key_priv = 1, ++ ACT_tpm2_key_pub = 2, ++ ACT_tpm2_key_type = 3, ++ NR__tpm2key_actions = 4, ++}; ++ ++enum { ++ Opt_new___2 = 0, ++ Opt_load___2 = 1, ++ Opt_update___2 = 2, ++ Opt_err___11 = 3, ++}; ++ ++enum { ++ Opt_default = 0, ++ Opt_ecryptfs = 1, ++ Opt_enc32 = 2, ++ Opt_error___2 = 3, ++}; ++ ++enum derived_key_type { ++ ENC_KEY = 0, ++ AUTH_KEY = 1, ++}; ++ ++struct trace_event_raw_selinux_audited { ++ struct trace_entry ent; ++ u32 requested; ++ u32 denied; ++ u32 audited; ++ int result; ++ u32 __data_loc_scontext; ++ u32 __data_loc_tcontext; ++ u32 __data_loc_tclass; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_selinux_audited { ++ u32 scontext; ++ u32 tcontext; ++ u32 tclass; ++}; ++ ++typedef void (*btf_trace_selinux_audited)(void *, struct selinux_audit_data *, char *, char *, const char *); ++ ++struct avc_xperms_node; ++ ++struct avc_entry { ++ u32 ssid; ++ u32 tsid; ++ u16 tclass; ++ struct av_decision avd; ++ struct avc_xperms_node *xp_node; ++}; ++ ++struct avc_xperms_node { ++ struct extended_perms xp; ++ struct list_head xpd_head; ++}; ++ ++struct avc_node { ++ struct avc_entry ae; ++ struct hlist_node list; ++ struct callback_head rhead; ++}; ++ ++struct avc_xperms_decision_node { ++ struct extended_perms_decision xpd; ++ struct list_head xpd_list; ++}; ++ ++struct avc_cache { ++ struct hlist_head slots[512]; ++ spinlock_t slots_lock[512]; ++ atomic_t lru_hint; ++ atomic_t active_nodes; ++ u32 latest_notif; ++}; ++ ++struct avc_callback_node { ++ int (*callback)(u32); ++ u32 events; ++ struct avc_callback_node *next; ++}; ++ ++struct selinux_avc { ++ unsigned int avc_cache_threshold; ++ struct avc_cache avc_cache; ++}; ++ ++enum { ++ SELNL_MSG_SETENFORCE = 16, ++ SELNL_MSG_POLICYLOAD = 17, ++ SELNL_MSG_MAX = 18, ++}; ++ ++enum selinux_nlgroups { ++ SELNLGRP_NONE = 0, ++ SELNLGRP_AVC = 1, ++ __SELNLGRP_MAX = 2, ++}; ++ ++struct selnl_msg_setenforce { ++ __s32 val; ++}; ++ ++struct selnl_msg_policyload { ++ __u32 seqno; ++}; ++ ++struct nlmsg_perm { ++ u16 nlmsg_type; ++ u32 perm; ++}; ++ ++struct netif_security_struct { ++ struct net *ns; ++ int ifindex; ++ u32 sid; ++}; ++ ++struct sel_netif { ++ struct list_head list; ++ struct netif_security_struct nsec; ++ struct callback_head callback_head; ++}; ++ ++struct netport_security_struct { ++ u32 sid; ++ u16 port; ++ u8 protocol; ++}; ++ ++struct sel_netport_bkt { ++ int size; ++ struct list_head list; ++}; ++ ++struct sel_netport { ++ struct netport_security_struct psec; ++ struct list_head list; ++ struct callback_head rcu; ++}; ++ ++struct sidtab_str_cache { ++ struct callback_head rcu_member; ++ struct list_head lru_member; ++ struct sidtab_entry *parent; ++ u32 len; ++ char str[0]; ++}; ++ ++struct level_datum { ++ struct mls_level *level; ++ unsigned char isalias; ++}; ++ ++struct cat_datum { ++ u32 value; ++ unsigned char isalias; ++}; ++ ++struct range_trans { ++ u32 source_type; ++ u32 target_type; ++ u32 target_class; ++}; ++ ++struct policy_data { ++ struct policydb *p; ++ void *fp; ++}; ++ ++struct policydb_compat_info { ++ int version; ++ int sym_num; ++ int ocon_num; ++}; ++ ++struct cond_insertf_data { ++ struct policydb *p; ++ struct avtab_node **dst; ++ struct cond_av_list *other; ++}; ++ ++struct pkey_security_struct { ++ u64 subnet_prefix; ++ u16 pkey; ++ u32 sid; ++}; ++ ++struct sel_ib_pkey_bkt { ++ int size; ++ struct list_head list; ++}; ++ ++struct sel_ib_pkey { ++ struct pkey_security_struct psec; ++ struct list_head list; ++ struct callback_head rcu; ++}; ++ ++struct ptrace_relation { ++ struct task_struct *tracer; ++ struct task_struct *tracee; ++ bool invalid; ++ struct list_head node; ++ struct callback_head rcu; ++}; ++ ++struct access_report_info { ++ struct callback_head work; ++ const char *access; ++ struct task_struct *target; ++ struct task_struct *agent; ++}; ++ ++enum data_formats { ++ DATA_FMT_DIGEST = 0, ++ DATA_FMT_DIGEST_WITH_ALGO = 1, ++ DATA_FMT_DIGEST_WITH_TYPE_AND_ALGO = 2, ++ DATA_FMT_STRING = 3, ++ DATA_FMT_HEX = 4, ++ DATA_FMT_UINT = 5, ++}; ++ ++enum digest_type { ++ DIGEST_TYPE_IMA = 0, ++ DIGEST_TYPE_VERITY = 1, ++ DIGEST_TYPE__LAST = 2, ++}; ++ ++struct ima_key_entry { ++ struct list_head list; ++ void *payload; ++ size_t payload_len; ++ char *keyring_name; ++}; ++ ++struct crypto_queue { ++ struct list_head list; ++ struct list_head *backlog; ++ unsigned int qlen; ++ unsigned int max_qlen; ++}; ++ ++struct rsa_mpi_key { ++ MPI n; ++ MPI e; ++ MPI d; ++ MPI p; ++ MPI q; ++ MPI dp; ++ MPI dq; ++ MPI qinv; ++}; ++ ++struct acomp_alg { ++ int (*compress)(struct acomp_req *); ++ int (*decompress)(struct acomp_req *); ++ void (*dst_free)(struct scatterlist *); ++ int (*init)(struct crypto_acomp *); ++ void (*exit)(struct crypto_acomp *); ++ unsigned int reqsize; ++ union { ++ struct { ++ struct crypto_alg base; ++ }; ++ struct comp_alg_common calg; ++ }; ++}; ++ ++struct crypto_report_acomp { ++ char type[64]; ++}; ++ ++struct drbg_test_data { ++ struct drbg_string *testentropy; ++}; ++ ++struct hash_testvec { ++ const char *key; ++ const char *plaintext; ++ const char *digest; ++ unsigned int psize; ++ short unsigned int ksize; ++ int setkey_error; ++ int digest_error; ++ bool fips_skip; ++}; ++ ++struct cipher_testvec { ++ const char *key; ++ const char *iv; ++ const char *iv_out; ++ const char *ptext; ++ const char *ctext; ++ unsigned char wk; ++ short unsigned int klen; ++ unsigned int len; ++ bool fips_skip; ++ bool generates_iv; ++ int setkey_error; ++ int crypt_error; ++}; ++ ++struct aead_testvec { ++ const char *key; ++ const char *iv; ++ const char *ptext; ++ const char *assoc; ++ const char *ctext; ++ unsigned char novrfy; ++ unsigned char wk; ++ unsigned char klen; ++ unsigned int plen; ++ unsigned int clen; ++ unsigned int alen; ++ int setkey_error; ++ int setauthsize_error; ++ int crypt_error; ++}; ++ ++struct cprng_testvec { ++ const char *key; ++ const char *dt; ++ const char *v; ++ const char *result; ++ unsigned char klen; ++ short unsigned int dtlen; ++ short unsigned int vlen; ++ short unsigned int rlen; ++ short unsigned int loops; ++}; ++ ++struct drbg_testvec { ++ const unsigned char *entropy; ++ size_t entropylen; ++ const unsigned char *entpra; ++ const unsigned char *entprb; ++ size_t entprlen; ++ const unsigned char *addtla; ++ const unsigned char *addtlb; ++ size_t addtllen; ++ const unsigned char *pers; ++ size_t perslen; ++ const unsigned char *expected; ++ size_t expectedlen; ++}; ++ ++struct akcipher_testvec { ++ const unsigned char *key; ++ const unsigned char *params; ++ const unsigned char *m; ++ const unsigned char *c; ++ unsigned int key_len; ++ unsigned int param_len; ++ unsigned int m_size; ++ unsigned int c_size; ++ bool public_key_vec; ++ bool siggen_sigver_test; ++ enum OID algo; ++}; ++ ++struct kpp_testvec { ++ const unsigned char *secret; ++ const unsigned char *b_secret; ++ const unsigned char *b_public; ++ const unsigned char *expected_a_public; ++ const unsigned char *expected_ss; ++ short unsigned int secret_size; ++ short unsigned int b_secret_size; ++ short unsigned int b_public_size; ++ short unsigned int expected_a_public_size; ++ short unsigned int expected_ss_size; ++ bool genkey; ++}; ++ ++struct comp_testvec { ++ int inlen; ++ int outlen; ++ char input[512]; ++ char output[512]; ++}; ++ ++struct aead_test_suite { ++ const struct aead_testvec *vecs; ++ unsigned int count; ++ unsigned int einval_allowed: 1; ++ unsigned int aad_iv: 1; ++}; ++ ++struct cipher_test_suite { ++ const struct cipher_testvec *vecs; ++ unsigned int count; ++}; ++ ++struct comp_test_suite { ++ struct { ++ const struct comp_testvec *vecs; ++ unsigned int count; ++ } comp; ++ struct { ++ const struct comp_testvec *vecs; ++ unsigned int count; ++ } decomp; ++}; ++ ++struct hash_test_suite { ++ const struct hash_testvec *vecs; ++ unsigned int count; ++}; ++ ++struct cprng_test_suite { ++ const struct cprng_testvec *vecs; ++ unsigned int count; ++}; ++ ++struct drbg_test_suite { ++ const struct drbg_testvec *vecs; ++ unsigned int count; ++}; ++ ++struct akcipher_test_suite { ++ const struct akcipher_testvec *vecs; ++ unsigned int count; ++}; ++ ++struct kpp_test_suite { ++ const struct kpp_testvec *vecs; ++ unsigned int count; ++}; ++ ++struct alg_test_desc { ++ const char *alg; ++ const char *generic_driver; ++ int (*test)(const struct alg_test_desc *, const char *, u32, u32); ++ int fips_allowed; ++ union { ++ struct aead_test_suite aead; ++ struct cipher_test_suite cipher; ++ struct comp_test_suite comp; ++ struct hash_test_suite hash; ++ struct cprng_test_suite cprng; ++ struct drbg_test_suite drbg; ++ struct akcipher_test_suite akcipher; ++ struct kpp_test_suite kpp; ++ } suite; ++}; ++ ++enum flush_type { ++ FLUSH_TYPE_NONE = 0, ++ FLUSH_TYPE_FLUSH = 1, ++ FLUSH_TYPE_REIMPORT = 2, ++}; ++ ++enum finalization_type { ++ FINALIZATION_TYPE_FINAL = 0, ++ FINALIZATION_TYPE_FINUP = 1, ++ FINALIZATION_TYPE_DIGEST = 2, ++}; ++ ++enum inplace_mode { ++ OUT_OF_PLACE = 0, ++ INPLACE_ONE_SGLIST = 1, ++ INPLACE_TWO_SGLISTS = 2, ++}; ++ ++struct test_sg_division { ++ unsigned int proportion_of_total; ++ unsigned int offset; ++ bool offset_relative_to_alignmask; ++ enum flush_type flush_type; ++ bool nosimd; ++}; ++ ++struct testvec_config { ++ const char *name; ++ enum inplace_mode inplace_mode; ++ u32 req_flags; ++ struct test_sg_division src_divs[8]; ++ struct test_sg_division dst_divs[8]; ++ unsigned int iv_offset; ++ unsigned int key_offset; ++ bool iv_offset_relative_to_alignmask; ++ bool key_offset_relative_to_alignmask; ++ enum finalization_type finalization_type; ++ bool nosimd; ++}; ++ ++struct test_sglist { ++ char *bufs[8]; ++ struct scatterlist sgl[8]; ++ struct scatterlist sgl_saved[8]; ++ struct scatterlist *sgl_ptr; ++ unsigned int nents; ++}; ++ ++struct cipher_test_sglists { ++ struct test_sglist src; ++ struct test_sglist dst; ++}; ++ ++struct sha512_state { ++ u64 state[8]; ++ u64 count[2]; ++ u8 buf[128]; ++}; ++ ++typedef void sha512_block_fn(struct sha512_state *, const u8 *, int); ++ ++struct crypto_cts_ctx { ++ struct crypto_skcipher *child; ++}; ++ ++struct crypto_cts_reqctx { ++ struct scatterlist sg[2]; ++ unsigned int offset; ++ struct skcipher_request subreq; ++}; ++ ++struct ccm_instance_ctx { ++ struct crypto_skcipher_spawn ctr; ++ struct crypto_ahash_spawn mac; ++}; ++ ++struct crypto_ccm_ctx { ++ struct crypto_ahash *mac; ++ struct crypto_skcipher *ctr; ++}; ++ ++struct crypto_rfc4309_ctx { ++ struct crypto_aead *child; ++ u8 nonce[3]; ++}; ++ ++struct crypto_rfc4309_req_ctx { ++ struct scatterlist src[3]; ++ struct scatterlist dst[3]; ++ struct aead_request subreq; ++}; ++ ++struct crypto_ccm_req_priv_ctx { ++ u8 odata[16]; ++ u8 idata[16]; ++ u8 auth_tag[16]; ++ u32 flags; ++ struct scatterlist src[3]; ++ struct scatterlist dst[3]; ++ union { ++ struct ahash_request ahreq; ++ struct skcipher_request skreq; ++ }; ++}; ++ ++struct cbcmac_tfm_ctx { ++ struct crypto_cipher *child; ++}; ++ ++struct cbcmac_desc_ctx { ++ unsigned int len; ++}; ++ ++struct chksum_ctx { ++ u32 key; ++}; ++ ++struct chksum_desc_ctx___2 { ++ u32 crc; ++}; ++ ++struct authenc_esn_instance_ctx { ++ struct crypto_ahash_spawn auth; ++ struct crypto_skcipher_spawn enc; ++}; ++ ++struct crypto_authenc_esn_ctx { ++ unsigned int reqoff; ++ struct crypto_ahash *auth; ++ struct crypto_skcipher *enc; ++ struct crypto_sync_skcipher *null; ++}; ++ ++struct authenc_esn_request_ctx { ++ struct scatterlist src[2]; ++ struct scatterlist dst[2]; ++ char tail[0]; ++}; ++ ++struct xxhash64_tfm_ctx { ++ u64 seed; ++}; ++ ++struct xxhash64_desc_ctx { ++ struct xxh64_state xxhstate; ++}; ++ ++struct rand_data { ++ __u64 data; ++ __u64 old_data; ++ __u64 prev_time; ++ __u64 last_delta; ++ __s64 last_delta2; ++ unsigned int osr; ++ unsigned char *mem; ++ unsigned int memlocation; ++ unsigned int memblocks; ++ unsigned int memblocksize; ++ unsigned int memaccessloops; ++ unsigned int rct_count; ++ unsigned int apt_observations; ++ unsigned int apt_count; ++ unsigned int apt_base; ++ unsigned int apt_base_set: 1; ++}; ++ ++struct jitterentropy { ++ spinlock_t jent_lock; ++ struct rand_data *entropy_collector; ++}; ++ ++struct sockaddr_alg_new { ++ __u16 salg_family; ++ __u8 salg_type[14]; ++ __u32 salg_feat; ++ __u32 salg_mask; ++ __u8 salg_name[0]; ++}; ++ ++struct af_alg_iv { ++ __u32 ivlen; ++ __u8 iv[0]; ++}; ++ ++struct af_alg_control { ++ struct af_alg_iv *iv; ++ int op; ++ unsigned int aead_assoclen; ++}; ++ ++struct alg_type_list { ++ const struct af_alg_type *type; ++ struct list_head list; ++}; ++ ++typedef struct { ++ u64 m_low; ++ u64 m_high; ++} uint128_t; ++ ++enum x509_akid_actions { ++ ACT_x509_akid_note_kid = 0, ++ ACT_x509_akid_note_name = 1, ++ ACT_x509_akid_note_serial = 2, ++ ACT_x509_extract_name_segment___2 = 3, ++ ACT_x509_note_OID___2 = 4, ++ NR__x509_akid_actions = 5, ++}; ++ ++struct bdev_inode { ++ struct block_device bdev; ++ struct inode vfs_inode; ++}; ++ ++struct queue_sysfs_entry { ++ struct attribute attr; ++ ssize_t (*show)(struct request_queue *, char *); ++ ssize_t (*store)(struct request_queue *, const char *, size_t); ++}; ++ ++enum { ++ ICQ_EXITED = 4, ++ ICQ_DESTROYED = 8, ++}; ++ ++struct bt_iter_data { ++ struct blk_mq_hw_ctx *hctx; ++ struct request_queue *q; ++ busy_tag_iter_fn *fn; ++ void *data; ++ bool reserved; ++}; ++ ++struct bt_tags_iter_data { ++ struct blk_mq_tags *tags; ++ busy_tag_iter_fn *fn; ++ void *data; ++ unsigned int flags; ++}; ++ ++struct blk_major_name { ++ struct blk_major_name *next; ++ int major; ++ char name[16]; ++ void (*probe)(dev_t); ++}; ++ ++struct fat_boot_sector { ++ __u8 ignored[3]; ++ __u8 system_id[8]; ++ __u8 sector_size[2]; ++ __u8 sec_per_clus; ++ __le16 reserved; ++ __u8 fats; ++ __u8 dir_entries[2]; ++ __u8 sectors[2]; ++ __u8 media; ++ __le16 fat_length; ++ __le16 secs_track; ++ __le16 heads; ++ __le32 hidden; ++ __le32 total_sect; ++ union { ++ struct { ++ __u8 drive_number; ++ __u8 state; ++ __u8 signature; ++ __u8 vol_id[4]; ++ __u8 vol_label[11]; ++ __u8 fs_type[8]; ++ } fat16; ++ struct { ++ __le32 length; ++ __le16 flags; ++ __u8 version[2]; ++ __le32 root_cluster; ++ __le16 info_sector; ++ __le16 backup_boot; ++ __le16 reserved2[6]; ++ __u8 drive_number; ++ __u8 state; ++ __u8 signature; ++ __u8 vol_id[4]; ++ __u8 vol_label[11]; ++ __u8 fs_type[8]; ++ } fat32; ++ }; ++}; ++ ++enum msdos_sys_ind { ++ DOS_EXTENDED_PARTITION = 5, ++ LINUX_EXTENDED_PARTITION = 133, ++ WIN98_EXTENDED_PARTITION = 15, ++ LINUX_DATA_PARTITION = 131, ++ LINUX_LVM_PARTITION = 142, ++ LINUX_RAID_PARTITION = 253, ++ SOLARIS_X86_PARTITION = 130, ++ NEW_SOLARIS_X86_PARTITION = 191, ++ DM6_AUX1PARTITION = 81, ++ DM6_AUX3PARTITION = 83, ++ DM6_PARTITION = 84, ++ EZD_PARTITION = 85, ++ FREEBSD_PARTITION = 165, ++ OPENBSD_PARTITION = 166, ++ NETBSD_PARTITION = 169, ++ BSDI_PARTITION = 183, ++ MINIX_PARTITION = 129, ++ UNIXWARE_PARTITION = 99, ++}; ++ ++struct solaris_x86_slice { ++ __le16 s_tag; ++ __le16 s_flag; ++ __le32 s_start; ++ __le32 s_size; ++}; ++ ++struct solaris_x86_vtoc { ++ unsigned int v_bootinfo[3]; ++ __le32 v_sanity; ++ __le32 v_version; ++ char v_volume[8]; ++ __le16 v_sectorsz; ++ __le16 v_nparts; ++ unsigned int v_reserved[10]; ++ struct solaris_x86_slice v_slice[16]; ++ unsigned int timestamp[16]; ++ char v_asciilabel[128]; ++}; ++ ++struct bsd_partition { ++ __le32 p_size; ++ __le32 p_offset; ++ __le32 p_fsize; ++ __u8 p_fstype; ++ __u8 p_frag; ++ __le16 p_cpg; ++}; ++ ++struct bsd_disklabel { ++ __le32 d_magic; ++ __s16 d_type; ++ __s16 d_subtype; ++ char d_typename[16]; ++ char d_packname[16]; ++ __u32 d_secsize; ++ __u32 d_nsectors; ++ __u32 d_ntracks; ++ __u32 d_ncylinders; ++ __u32 d_secpercyl; ++ __u32 d_secperunit; ++ __u16 d_sparespertrack; ++ __u16 d_sparespercyl; ++ __u32 d_acylinders; ++ __u16 d_rpm; ++ __u16 d_interleave; ++ __u16 d_trackskew; ++ __u16 d_cylskew; ++ __u32 d_headswitch; ++ __u32 d_trkseek; ++ __u32 d_flags; ++ __u32 d_drivedata[5]; ++ __u32 d_spare[5]; ++ __le32 d_magic2; ++ __le16 d_checksum; ++ __le16 d_npartitions; ++ __le32 d_bbsize; ++ __le32 d_sbsize; ++ struct bsd_partition d_partitions[16]; ++}; ++ ++struct unixware_slice { ++ __le16 s_label; ++ __le16 s_flags; ++ __le32 start_sect; ++ __le32 nr_sects; ++}; ++ ++struct unixware_vtoc { ++ __le32 v_magic; ++ __le32 v_version; ++ char v_name[8]; ++ __le16 v_nslices; ++ __le16 v_unknown1; ++ __le32 v_reserved[10]; ++ struct unixware_slice v_slice[16]; ++}; ++ ++struct unixware_disklabel { ++ __le32 d_type; ++ __le32 d_magic; ++ __le32 d_version; ++ char d_serial[12]; ++ __le32 d_ncylinders; ++ __le32 d_ntracks; ++ __le32 d_nsectors; ++ __le32 d_secsize; ++ __le32 d_part_start; ++ __le32 d_unknown1[12]; ++ __le32 d_alt_tbl; ++ __le32 d_alt_len; ++ __le32 d_phys_cyl; ++ __le32 d_phys_trk; ++ __le32 d_phys_sec; ++ __le32 d_phys_bytes; ++ __le32 d_unknown2; ++ __le32 d_unknown3; ++ __le32 d_pad[8]; ++ struct unixware_vtoc vtoc; ++}; ++ ++struct rq_qos_wait_data { ++ struct wait_queue_entry wq; ++ struct task_struct *task; ++ struct rq_wait *rqw; ++ acquire_inflight_cb_t *cb; ++ void *private_data; ++ bool got_token; ++}; ++ ++struct bsg_job; ++ ++typedef int bsg_job_fn(struct bsg_job *); ++ ++struct bsg_buffer { ++ unsigned int payload_len; ++ int sg_cnt; ++ struct scatterlist *sg_list; ++}; ++ ++struct bsg_job { ++ struct device *dev; ++ struct kref kref; ++ unsigned int timeout; ++ void *request; ++ void *reply; ++ unsigned int request_len; ++ unsigned int reply_len; ++ struct bsg_buffer request_payload; ++ struct bsg_buffer reply_payload; ++ int result; ++ unsigned int reply_payload_rcv_len; ++ struct request *bidi_rq; ++ struct bio *bidi_bio; ++ void *dd_data; ++}; ++ ++typedef enum blk_eh_timer_return bsg_timeout_fn(struct request *); ++ ++struct bsg_set { ++ struct blk_mq_tag_set tag_set; ++ struct bsg_device *bd; ++ bsg_job_fn *job_fn; ++ bsg_timeout_fn *timeout_fn; ++}; ++ ++enum prio_policy { ++ POLICY_NO_CHANGE = 0, ++ POLICY_NONE_TO_RT = 1, ++ POLICY_RESTRICT_TO_BE = 2, ++ POLICY_ALL_TO_IDLE = 3, ++}; ++ ++struct ioprio_blkg { ++ struct blkg_policy_data pd; ++}; ++ ++struct ioprio_blkcg { ++ struct blkcg_policy_data cpd; ++ enum prio_policy prio_policy; ++}; ++ ++enum dd_data_dir { ++ DD_READ = 0, ++ DD_WRITE = 1, ++}; ++ ++enum { ++ DD_DIR_COUNT = 2, ++}; ++ ++enum dd_prio { ++ DD_RT_PRIO = 0, ++ DD_BE_PRIO = 1, ++ DD_IDLE_PRIO = 2, ++ DD_PRIO_MAX = 2, ++}; ++ ++enum { ++ DD_PRIO_COUNT = 3, ++}; ++ ++struct io_stats_per_prio { ++ uint32_t inserted; ++ uint32_t merged; ++ uint32_t dispatched; ++ atomic_t completed; ++}; ++ ++struct dd_per_prio { ++ struct list_head dispatch; ++ struct rb_root sort_list[2]; ++ struct list_head fifo_list[2]; ++ struct request *next_rq[2]; ++ struct io_stats_per_prio stats; ++}; ++ ++struct deadline_data { ++ struct dd_per_prio per_prio[3]; ++ enum dd_data_dir last_dir; ++ unsigned int batching; ++ unsigned int starved; ++ int fifo_expire[2]; ++ int fifo_batch; ++ int writes_starved; ++ int front_merges; ++ u32 async_depth; ++ int prio_aging_expire; ++ spinlock_t lock; ++ spinlock_t zone_lock; ++}; ++ ++enum bfqq_state_flags { ++ BFQQF_just_created = 0, ++ BFQQF_busy = 1, ++ BFQQF_wait_request = 2, ++ BFQQF_non_blocking_wait_rq = 3, ++ BFQQF_fifo_expire = 4, ++ BFQQF_has_short_ttime = 5, ++ BFQQF_sync = 6, ++ BFQQF_IO_bound = 7, ++ BFQQF_in_large_burst = 8, ++ BFQQF_softrt_update = 9, ++ BFQQF_coop = 10, ++ BFQQF_split_coop = 11, ++}; ++ ++struct t10_pi_tuple { ++ __be16 guard_tag; ++ __be16 app_tag; ++ __be32 ref_tag; ++}; ++ ++struct crc64_pi_tuple { ++ __be64 guard_tag; ++ __be16 app_tag; ++ __u8 ref_tag[6]; ++}; ++ ++typedef __be16 csum_fn(void *, unsigned int); ++ ++enum blk_zone_report_flags { ++ BLK_ZONE_REP_CAPACITY = 1, ++}; ++ ++struct blk_zone_report { ++ __u64 sector; ++ __u32 nr_zones; ++ __u32 flags; ++ struct blk_zone zones[0]; ++}; ++ ++struct blk_zone_range { ++ __u64 sector; ++ __u64 nr_sectors; ++}; ++ ++struct zone_report_args { ++ struct blk_zone *zones; ++}; ++ ++struct blk_revalidate_zone_args { ++ struct gendisk *disk; ++ long unsigned int *conv_zones_bitmap; ++ long unsigned int *seq_zones_wlock; ++ unsigned int nr_zones; ++ sector_t zone_sectors; ++ sector_t sector; ++}; ++ ++struct blk_crypto_kobj { ++ struct kobject kobj; ++ struct blk_crypto_profile *profile; ++}; ++ ++struct blk_crypto_attr { ++ struct attribute attr; ++ ssize_t (*show)(struct blk_crypto_profile *, struct blk_crypto_attr *, char *); ++}; ++ ++struct io_sqring_offsets { ++ __u32 head; ++ __u32 tail; ++ __u32 ring_mask; ++ __u32 ring_entries; ++ __u32 flags; ++ __u32 dropped; ++ __u32 array; ++ __u32 resv1; ++ __u64 resv2; ++}; ++ ++struct io_cqring_offsets { ++ __u32 head; ++ __u32 tail; ++ __u32 ring_mask; ++ __u32 ring_entries; ++ __u32 overflow; ++ __u32 cqes; ++ __u32 flags; ++ __u32 resv1; ++ __u64 resv2; ++}; ++ ++struct io_uring_params { ++ __u32 sq_entries; ++ __u32 cq_entries; ++ __u32 flags; ++ __u32 sq_thread_cpu; ++ __u32 sq_thread_idle; ++ __u32 features; ++ __u32 wq_fd; ++ __u32 resv[3]; ++ struct io_sqring_offsets sq_off; ++ struct io_cqring_offsets cq_off; ++}; ++ ++struct io_uring_probe_op { ++ __u8 op; ++ __u8 resv; ++ __u16 flags; ++ __u32 resv2; ++}; ++ ++struct io_uring_probe { ++ __u8 last_op; ++ __u8 ops_len; ++ __u16 resv; ++ __u32 resv2[3]; ++ struct io_uring_probe_op ops[0]; ++}; ++ ++struct io_uring_restriction { ++ __u16 opcode; ++ union { ++ __u8 register_op; ++ __u8 sqe_op; ++ __u8 sqe_flags; ++ }; ++ __u8 resv; ++ __u32 resv2[3]; ++}; ++ ++enum { ++ IORING_RESTRICTION_REGISTER_OP = 0, ++ IORING_RESTRICTION_SQE_OP = 1, ++ IORING_RESTRICTION_SQE_FLAGS_ALLOWED = 2, ++ IORING_RESTRICTION_SQE_FLAGS_REQUIRED = 3, ++ IORING_RESTRICTION_LAST = 4, ++}; ++ ++struct io_uring_getevents_arg { ++ __u64 sigmask; ++ __u32 sigmask_sz; ++ __u32 pad; ++ __u64 ts; ++}; ++ ++struct trace_event_raw_io_uring_create { ++ struct trace_entry ent; ++ int fd; ++ void *ctx; ++ u32 sq_entries; ++ u32 cq_entries; ++ u32 flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_register { ++ struct trace_entry ent; ++ void *ctx; ++ unsigned int opcode; ++ unsigned int nr_files; ++ unsigned int nr_bufs; ++ long int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_file_get { ++ struct trace_entry ent; ++ void *ctx; ++ void *req; ++ u64 user_data; ++ int fd; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_queue_async_work { ++ struct trace_entry ent; ++ void *ctx; ++ void *req; ++ u64 user_data; ++ u8 opcode; ++ unsigned int flags; ++ struct io_wq_work *work; ++ int rw; ++ u32 __data_loc_op_str; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_defer { ++ struct trace_entry ent; ++ void *ctx; ++ void *req; ++ long long unsigned int data; ++ u8 opcode; ++ u32 __data_loc_op_str; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_link { ++ struct trace_entry ent; ++ void *ctx; ++ void *req; ++ void *target_req; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_cqring_wait { ++ struct trace_entry ent; ++ void *ctx; ++ int min_events; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_fail_link { ++ struct trace_entry ent; ++ void *ctx; ++ void *req; ++ long long unsigned int user_data; ++ u8 opcode; ++ void *link; ++ u32 __data_loc_op_str; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_complete { ++ struct trace_entry ent; ++ void *ctx; ++ void *req; ++ u64 user_data; ++ int res; ++ unsigned int cflags; ++ u64 extra1; ++ u64 extra2; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_submit_req { ++ struct trace_entry ent; ++ void *ctx; ++ void *req; ++ long long unsigned int user_data; ++ u8 opcode; ++ u32 flags; ++ bool sq_thread; ++ u32 __data_loc_op_str; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_poll_arm { ++ struct trace_entry ent; ++ void *ctx; ++ void *req; ++ long long unsigned int user_data; ++ u8 opcode; ++ int mask; ++ int events; ++ u32 __data_loc_op_str; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_task_add { ++ struct trace_entry ent; ++ void *ctx; ++ void *req; ++ long long unsigned int user_data; ++ u8 opcode; ++ int mask; ++ u32 __data_loc_op_str; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_req_failed { ++ struct trace_entry ent; ++ void *ctx; ++ void *req; ++ long long unsigned int user_data; ++ u8 opcode; ++ u8 flags; ++ u8 ioprio; ++ u64 off; ++ u64 addr; ++ u32 len; ++ u32 op_flags; ++ u16 buf_index; ++ u16 personality; ++ u32 file_index; ++ u64 pad1; ++ u64 addr3; ++ int error; ++ u32 __data_loc_op_str; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_cqe_overflow { ++ struct trace_entry ent; ++ void *ctx; ++ long long unsigned int user_data; ++ s32 res; ++ u32 cflags; ++ void *ocqe; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_task_work_run { ++ struct trace_entry ent; ++ void *tctx; ++ unsigned int count; ++ unsigned int loops; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_short_write { ++ struct trace_entry ent; ++ void *ctx; ++ u64 fpos; ++ u64 wanted; ++ u64 got; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_io_uring_local_work_run { ++ struct trace_entry ent; ++ void *ctx; ++ int count; ++ unsigned int loops; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_io_uring_create {}; ++ ++struct trace_event_data_offsets_io_uring_register {}; ++ ++struct trace_event_data_offsets_io_uring_file_get {}; ++ ++struct trace_event_data_offsets_io_uring_queue_async_work { ++ u32 op_str; ++}; ++ ++struct trace_event_data_offsets_io_uring_defer { ++ u32 op_str; ++}; ++ ++struct trace_event_data_offsets_io_uring_link {}; ++ ++struct trace_event_data_offsets_io_uring_cqring_wait {}; ++ ++struct trace_event_data_offsets_io_uring_fail_link { ++ u32 op_str; ++}; ++ ++struct trace_event_data_offsets_io_uring_complete {}; ++ ++struct trace_event_data_offsets_io_uring_submit_req { ++ u32 op_str; ++}; ++ ++struct trace_event_data_offsets_io_uring_poll_arm { ++ u32 op_str; ++}; ++ ++struct trace_event_data_offsets_io_uring_task_add { ++ u32 op_str; ++}; ++ ++struct trace_event_data_offsets_io_uring_req_failed { ++ u32 op_str; ++}; ++ ++struct trace_event_data_offsets_io_uring_cqe_overflow {}; ++ ++struct trace_event_data_offsets_io_uring_task_work_run {}; ++ ++struct trace_event_data_offsets_io_uring_short_write {}; ++ ++struct trace_event_data_offsets_io_uring_local_work_run {}; ++ ++typedef void (*btf_trace_io_uring_create)(void *, int, void *, u32, u32, u32); ++ ++typedef void (*btf_trace_io_uring_register)(void *, void *, unsigned int, unsigned int, unsigned int, long int); ++ ++typedef void (*btf_trace_io_uring_file_get)(void *, struct io_kiocb *, int); ++ ++typedef void (*btf_trace_io_uring_queue_async_work)(void *, struct io_kiocb *, int); ++ ++typedef void (*btf_trace_io_uring_defer)(void *, struct io_kiocb *); ++ ++typedef void (*btf_trace_io_uring_link)(void *, struct io_kiocb *, struct io_kiocb *); ++ ++typedef void (*btf_trace_io_uring_cqring_wait)(void *, void *, int); ++ ++typedef void (*btf_trace_io_uring_fail_link)(void *, struct io_kiocb *, struct io_kiocb *); ++ ++typedef void (*btf_trace_io_uring_complete)(void *, void *, void *, u64, int, unsigned int, u64, u64); ++ ++typedef void (*btf_trace_io_uring_submit_req)(void *, struct io_kiocb *); ++ ++typedef void (*btf_trace_io_uring_poll_arm)(void *, struct io_kiocb *, int, int); ++ ++typedef void (*btf_trace_io_uring_task_add)(void *, struct io_kiocb *, int); ++ ++typedef void (*btf_trace_io_uring_req_failed)(void *, const struct io_uring_sqe *, struct io_kiocb *, int); ++ ++typedef void (*btf_trace_io_uring_cqe_overflow)(void *, void *, long long unsigned int, s32, u32, void *); ++ ++typedef void (*btf_trace_io_uring_task_work_run)(void *, void *, unsigned int, unsigned int); ++ ++typedef void (*btf_trace_io_uring_short_write)(void *, void *, u64, u64, u64); ++ ++typedef void (*btf_trace_io_uring_local_work_run)(void *, void *, int, unsigned int); ++ ++enum { ++ IO_WQ_WORK_CANCEL = 1, ++ IO_WQ_WORK_HASHED = 2, ++ IO_WQ_WORK_UNBOUND = 4, ++ IO_WQ_WORK_CONCURRENT = 16, ++ IO_WQ_HASH_SHIFT = 24, ++}; ++ ++enum { ++ IO_APOLL_OK = 0, ++ IO_APOLL_ABORTED = 1, ++ IO_APOLL_READY = 2, ++}; ++ ++enum { ++ IO_CHECK_CQ_OVERFLOW_BIT = 0, ++ IO_CHECK_CQ_DROPPED_BIT = 1, ++}; ++ ++enum { ++ IO_EVENTFD_OP_SIGNAL_BIT = 0, ++ IO_EVENTFD_OP_FREE_BIT = 1, ++}; ++ ++struct io_defer_entry { ++ struct list_head list; ++ struct io_kiocb *req; ++ u32 seq; ++}; ++ ++struct io_wait_queue { ++ struct wait_queue_entry wq; ++ struct io_ring_ctx *ctx; ++ unsigned int cq_tail; ++ unsigned int nr_timeouts; ++ ktime_t timeout; ++}; ++ ++struct io_tctx_exit { ++ struct callback_head task_work; ++ struct completion completion; ++ struct io_ring_ctx *ctx; ++}; ++ ++struct io_task_cancel { ++ struct task_struct *task; ++ bool all; ++}; ++ ++struct creds; ++ ++struct io_epoll { ++ struct file *file; ++ int epfd; ++ int op; ++ int fd; ++ struct epoll_event event; ++}; ++ ++struct io_uring_recvmsg_out { ++ __u32 namelen; ++ __u32 controllen; ++ __u32 payloadlen; ++ __u32 flags; ++}; ++ ++struct io_async_msghdr { ++ union { ++ struct iovec fast_iov[8]; ++ struct { ++ struct iovec fast_iov_one; ++ __kernel_size_t controllen; ++ int namelen; ++ __kernel_size_t payloadlen; ++ }; ++ struct io_cache_entry cache; ++ }; ++ struct iovec *free_iov; ++ struct sockaddr *uaddr; ++ struct msghdr msg; ++ struct __kernel_sockaddr_storage addr; ++}; ++ ++struct io_async_connect { ++ struct __kernel_sockaddr_storage address; ++}; ++ ++struct io_shutdown { ++ struct file *file; ++ int how; ++}; ++ ++struct io_accept { ++ struct file *file; ++ struct sockaddr *addr; ++ int *addr_len; ++ int flags; ++ u32 file_slot; ++ long unsigned int nofile; ++}; ++ ++struct io_socket { ++ struct file *file; ++ int domain; ++ int type; ++ int protocol; ++ int flags; ++ u32 file_slot; ++ long unsigned int nofile; ++}; ++ ++struct io_connect { ++ struct file *file; ++ struct sockaddr *addr; ++ int addr_len; ++ bool in_progress; ++ bool seen_econnaborted; ++}; ++ ++struct io_sr_msg { ++ struct file *file; ++ union { ++ struct compat_msghdr *umsg_compat; ++ struct user_msghdr *umsg; ++ void *buf; ++ }; ++ unsigned int len; ++ unsigned int done_io; ++ unsigned int msg_flags; ++ u16 flags; ++ u16 addr_len; ++ u16 buf_group; ++ void *addr; ++ void *msg_control; ++ struct io_kiocb *notif; ++}; ++ ++struct io_recvmsg_multishot_hdr { ++ struct io_uring_recvmsg_out msg; ++ struct __kernel_sockaddr_storage addr; ++}; ++ ++enum { ++ IO_SQ_THREAD_SHOULD_STOP = 0, ++ IO_SQ_THREAD_SHOULD_PARK = 1, ++}; ++ ++struct io_poll_update { ++ struct file *file; ++ u64 old_user_data; ++ u64 new_user_data; ++ __poll_t events; ++ bool update_events; ++ bool update_user_data; ++}; ++ ++struct io_poll_table { ++ struct poll_table_struct pt; ++ struct io_kiocb *req; ++ int nr_entries; ++ int error; ++ bool owning; ++ __poll_t result_mask; ++}; ++ ++enum { ++ IOU_POLL_DONE = 0, ++ IOU_POLL_NO_ACTION = 1, ++ IOU_POLL_REMOVE_POLL_USE_RES = 2, ++ IOU_POLL_REISSUE = 3, ++}; ++ ++struct io_rw_state { ++ struct iov_iter iter; ++ struct iov_iter_state iter_state; ++ struct iovec fast_iov[8]; ++}; ++ ++struct io_async_rw { ++ struct io_rw_state s; ++ const struct iovec *free_iovec; ++ size_t bytes_done; ++ struct wait_page_queue wpq; ++}; ++ ++struct io_rw { ++ struct kiocb kiocb; ++ u64 addr; ++ u32 len; ++ rwf_t flags; ++}; ++ ++enum { ++ IO_WQ_BOUND = 0, ++ IO_WQ_UNBOUND = 1, ++}; ++ ++struct io_wq_acct { ++ unsigned int nr_workers; ++ unsigned int max_workers; ++ int index; ++ atomic_t nr_running; ++ raw_spinlock_t lock; ++ struct io_wq_work_list work_list; ++ long unsigned int flags; ++}; ++ ++struct io_wq { ++ long unsigned int state; ++ free_work_fn *free_work; ++ io_wq_work_fn *do_work; ++ struct io_wq_hash *hash; ++ atomic_t worker_refs; ++ struct completion worker_done; ++ struct hlist_node cpuhp_node; ++ struct task_struct *task; ++ struct io_wq_acct acct[2]; ++ raw_spinlock_t lock; ++ struct hlist_nulls_head free_list; ++ struct list_head all_list; ++ struct wait_queue_entry wait; ++ struct io_wq_work *hash_tail[64]; ++ cpumask_var_t cpu_mask; ++}; ++ ++enum { ++ IO_WORKER_F_UP = 1, ++ IO_WORKER_F_RUNNING = 2, ++ IO_WORKER_F_FREE = 4, ++ IO_WORKER_F_BOUND = 8, ++}; ++ ++enum { ++ IO_WQ_BIT_EXIT = 0, ++}; ++ ++enum { ++ IO_ACCT_STALLED_BIT = 0, ++}; ++ ++struct io_worker { ++ refcount_t ref; ++ unsigned int flags; ++ struct hlist_nulls_node nulls_node; ++ struct list_head all_list; ++ struct task_struct *task; ++ struct io_wq *wq; ++ struct io_wq_work *cur_work; ++ struct io_wq_work *next_work; ++ raw_spinlock_t lock; ++ struct completion ref_done; ++ long unsigned int create_state; ++ struct callback_head create_work; ++ int create_index; ++ union { ++ struct callback_head rcu; ++ struct work_struct work; ++ }; ++}; ++ ++enum { ++ IO_WQ_ACCT_BOUND = 0, ++ IO_WQ_ACCT_UNBOUND = 1, ++ IO_WQ_ACCT_NR = 2, ++}; ++ ++struct io_cb_cancel_data { ++ work_cancel_fn *fn; ++ void *data; ++ int nr_running; ++ int nr_pending; ++ bool cancel_all; ++}; ++ ++struct online_data { ++ unsigned int cpu; ++ bool online; ++}; ++ ++union nested_table { ++ union nested_table *table; ++ struct rhash_lock_head *bucket; ++}; ++ ++struct genradix_iter { ++ size_t offset; ++ size_t pos; ++}; ++ ++struct genradix_node { ++ union { ++ struct genradix_node *children[512]; ++ u8 data[4096]; ++ }; ++}; ++ ++struct kunit_hooks_table { ++ void (*fail_current_test)(const char *, int, const char *, ...); ++ void * (*get_static_stub_address)(struct kunit *, void *); ++}; ++ ++struct reciprocal_value_adv { ++ u32 m; ++ u8 sh; ++ u8 exp; ++ bool is_wide_m; ++}; ++ ++enum { ++ TEST_ALIGNMENT = 16, ++}; ++ ++struct interval_tree_span_iter { ++ struct interval_tree_node *nodes[2]; ++ long unsigned int first_index; ++ long unsigned int last_index; ++ union { ++ long unsigned int start_hole; ++ long unsigned int start_used; ++ }; ++ union { ++ long unsigned int last_hole; ++ long unsigned int last_used; ++ }; ++ int is_hole; ++}; ++ ++enum packing_op { ++ PACK = 0, ++ UNPACK = 1, ++}; ++ ++struct sw842_hlist_node8 { ++ struct hlist_node node; ++ u64 data; ++ u8 index; ++}; ++ ++struct sw842_hlist_node4 { ++ struct hlist_node node; ++ u32 data; ++ u16 index; ++}; ++ ++struct sw842_hlist_node2 { ++ struct hlist_node node; ++ u16 data; ++ u8 index; ++}; ++ ++struct sw842_param___2 { ++ u8 *in; ++ u8 *instart; ++ u64 ilen; ++ u8 *out; ++ u64 olen; ++ u8 bit; ++ u64 data8[1]; ++ u32 data4[2]; ++ u16 data2[4]; ++ int index8[1]; ++ int index4[2]; ++ int index2[4]; ++ struct hlist_head htable8[1024]; ++ struct hlist_head htable4[2048]; ++ struct hlist_head htable2[1024]; ++ struct sw842_hlist_node8 node8[256]; ++ struct sw842_hlist_node4 node4[512]; ++ struct sw842_hlist_node2 node2[256]; ++}; ++ ++typedef struct tree_desc_s tree_desc; ++ ++typedef struct { ++ const uint8_t *externalDict; ++ size_t extDictSize; ++ const uint8_t *prefixEnd; ++ size_t prefixSize; ++} LZ4_streamDecode_t_internal; ++ ++typedef union { ++ long long unsigned int table[4]; ++ LZ4_streamDecode_t_internal internal_donotuse; ++} LZ4_streamDecode_t; ++ ++typedef uintptr_t uptrval; ++ ++typedef enum { ++ noDict = 0, ++ withPrefix64k = 1, ++ usingExtDict = 2, ++} dict_directive; ++ ++typedef enum { ++ endOnOutputSize = 0, ++ endOnInputSize = 1, ++} endCondition_directive; ++ ++typedef enum { ++ decode_full_block = 0, ++ partial_decode = 1, ++} earlyEnd_directive; ++ ++typedef struct { ++ S16 norm[53]; ++ U32 wksp[285]; ++} ZSTD_BuildCTableWksp; ++ ++typedef struct { ++ U64 rolling; ++ U64 stopMask; ++} ldmRollingHashState_t; ++ ++typedef U32 (*ZSTD_getAllMatchesFn)(ZSTD_match_t *, ZSTD_matchState_t *, U32 *, const BYTE *, const BYTE *, const U32 *, const U32, const U32); ++ ++typedef struct { ++ rawSeqStore_t seqStore; ++ U32 startPosInBlock; ++ U32 endPosInBlock; ++ U32 offset; ++} ZSTD_optLdm_t; ++ ++typedef struct { ++ U32 fastMode; ++ U32 tableLog; ++} ZSTD_seqSymbol_header; ++ ++typedef struct { ++ size_t litLength; ++ size_t matchLength; ++ size_t offset; ++} seq_t; ++ ++typedef struct { ++ size_t state; ++ const ZSTD_seqSymbol *table; ++} ZSTD_fseState; ++ ++typedef struct { ++ BIT_DStream_t DStream; ++ ZSTD_fseState stateLL; ++ ZSTD_fseState stateOffb; ++ ZSTD_fseState stateML; ++ size_t prevOffset[3]; ++} seqState_t; ++ ++typedef enum { ++ ZSTD_lo_isRegularOffset = 0, ++ ZSTD_lo_isLongOffset = 1, ++} ZSTD_longOffset_e; ++ ++typedef struct { ++ u8 _[128]; ++} addrtype___5; ++ ++typedef s32 pcp_op_T_____10; ++ ++struct ddebug_class_param { ++ union { ++ long unsigned int *bits; ++ unsigned int *lvl; ++ }; ++ char flags[8]; ++ const struct ddebug_class_map *map; ++}; ++ ++struct irq_poll; ++ ++typedef int irq_poll_fn(struct irq_poll *, int); ++ ++struct irq_poll { ++ struct list_head list; ++ long unsigned int state; ++ int weight; ++ irq_poll_fn *poll; ++}; ++ ++enum rdma_nl_counter_mode { ++ RDMA_COUNTER_MODE_NONE = 0, ++ RDMA_COUNTER_MODE_AUTO = 1, ++ RDMA_COUNTER_MODE_MANUAL = 2, ++ RDMA_COUNTER_MODE_MAX = 3, ++}; ++ ++enum rdma_nl_counter_mask { ++ RDMA_COUNTER_MASK_QP_TYPE = 1, ++ RDMA_COUNTER_MASK_PID = 2, ++}; ++ ++enum rdma_restrack_type { ++ RDMA_RESTRACK_PD = 0, ++ RDMA_RESTRACK_CQ = 1, ++ RDMA_RESTRACK_QP = 2, ++ RDMA_RESTRACK_CM_ID = 3, ++ RDMA_RESTRACK_MR = 4, ++ RDMA_RESTRACK_CTX = 5, ++ RDMA_RESTRACK_COUNTER = 6, ++ RDMA_RESTRACK_SRQ = 7, ++ RDMA_RESTRACK_MAX = 8, ++}; ++ ++struct rdma_restrack_entry { ++ bool valid; ++ u8 no_track: 1; ++ struct kref kref; ++ struct completion comp; ++ struct task_struct *task; ++ const char *kern_name; ++ enum rdma_restrack_type type; ++ bool user; ++ u32 id; ++}; ++ ++struct rdma_link_ops { ++ struct list_head list; ++ const char *type; ++ int (*newlink)(const char *, struct net_device *); ++}; ++ ++struct auto_mode_param { ++ int qp_type; ++}; ++ ++struct rdma_counter_mode { ++ enum rdma_nl_counter_mode mode; ++ enum rdma_nl_counter_mask mask; ++ struct auto_mode_param param; ++}; ++ ++struct rdma_hw_stats; ++ ++struct rdma_port_counter { ++ struct rdma_counter_mode mode; ++ struct rdma_hw_stats *hstats; ++ unsigned int num_counters; ++ struct mutex lock; ++}; ++ ++struct rdma_stat_desc; ++ ++struct rdma_hw_stats { ++ struct mutex lock; ++ long unsigned int timestamp; ++ long unsigned int lifespan; ++ const struct rdma_stat_desc *descs; ++ long unsigned int *is_disabled; ++ int num_counters; ++ u64 value[0]; ++}; ++ ++struct rdma_counter { ++ struct rdma_restrack_entry res; ++ struct ib_device *device; ++ uint32_t id; ++ struct kref kref; ++ struct rdma_counter_mode mode; ++ struct mutex lock; ++ struct rdma_hw_stats *stats; ++ u32 port; ++}; ++ ++enum rdma_driver_id { ++ RDMA_DRIVER_UNKNOWN = 0, ++ RDMA_DRIVER_MLX5 = 1, ++ RDMA_DRIVER_MLX4 = 2, ++ RDMA_DRIVER_CXGB3 = 3, ++ RDMA_DRIVER_CXGB4 = 4, ++ RDMA_DRIVER_MTHCA = 5, ++ RDMA_DRIVER_BNXT_RE = 6, ++ RDMA_DRIVER_OCRDMA = 7, ++ RDMA_DRIVER_NES = 8, ++ RDMA_DRIVER_I40IW = 9, ++ RDMA_DRIVER_IRDMA = 9, ++ RDMA_DRIVER_VMW_PVRDMA = 10, ++ RDMA_DRIVER_QEDR = 11, ++ RDMA_DRIVER_HNS = 12, ++ RDMA_DRIVER_USNIC = 13, ++ RDMA_DRIVER_RXE = 14, ++ RDMA_DRIVER_HFI1 = 15, ++ RDMA_DRIVER_QIB = 16, ++ RDMA_DRIVER_EFA = 17, ++ RDMA_DRIVER_SIW = 18, ++ RDMA_DRIVER_ERDMA = 19, ++ RDMA_DRIVER_MANA = 20, ++}; ++ ++enum ib_cq_notify_flags { ++ IB_CQ_SOLICITED = 1, ++ IB_CQ_NEXT_COMP = 2, ++ IB_CQ_SOLICITED_MASK = 3, ++ IB_CQ_REPORT_MISSED_EVENTS = 4, ++}; ++ ++struct ib_mad; ++ ++enum rdma_link_layer { ++ IB_LINK_LAYER_UNSPECIFIED = 0, ++ IB_LINK_LAYER_INFINIBAND = 1, ++ IB_LINK_LAYER_ETHERNET = 2, ++}; ++ ++enum rdma_netdev_t { ++ RDMA_NETDEV_OPA_VNIC = 0, ++ RDMA_NETDEV_IPOIB = 1, ++}; ++ ++enum ib_srq_attr_mask { ++ IB_SRQ_MAX_WR = 1, ++ IB_SRQ_LIMIT = 2, ++}; ++ ++enum ib_mr_type { ++ IB_MR_TYPE_MEM_REG = 0, ++ IB_MR_TYPE_SG_GAPS = 1, ++ IB_MR_TYPE_DM = 2, ++ IB_MR_TYPE_USER = 3, ++ IB_MR_TYPE_DMA = 4, ++ IB_MR_TYPE_INTEGRITY = 5, ++}; ++ ++enum ib_uverbs_advise_mr_advice { ++ IB_UVERBS_ADVISE_MR_ADVICE_PREFETCH = 0, ++ IB_UVERBS_ADVISE_MR_ADVICE_PREFETCH_WRITE = 1, ++ IB_UVERBS_ADVISE_MR_ADVICE_PREFETCH_NO_FAULT = 2, ++}; ++ ++struct uverbs_attr_bundle; ++ ++struct rdma_cm_id; ++ ++struct iw_cm_id; ++ ++struct iw_cm_conn_param; ++ ++struct ib_qp; ++ ++struct ib_send_wr; ++ ++struct ib_recv_wr; ++ ++struct ib_cq; ++ ++struct ib_wc; ++ ++struct ib_srq; ++ ++struct ib_grh; ++ ++struct ib_device_attr; ++ ++struct ib_udata; ++ ++struct ib_device_modify; ++ ++struct ib_port_attr; ++ ++struct ib_port_modify; ++ ++struct ib_port_immutable; ++ ++struct rdma_netdev_alloc_params; ++ ++union ib_gid; ++ ++struct ib_gid_attr; ++ ++struct ib_ucontext; ++ ++struct rdma_user_mmap_entry; ++ ++struct ib_pd; ++ ++struct ib_ah; ++ ++struct rdma_ah_init_attr; ++ ++struct rdma_ah_attr; ++ ++struct ib_srq_init_attr; ++ ++struct ib_srq_attr; ++ ++struct ib_qp_init_attr; ++ ++struct ib_qp_attr; ++ ++struct ib_cq_init_attr; ++ ++struct ib_mr; ++ ++struct ib_sge; ++ ++struct ib_mr_status; ++ ++struct ib_mw; ++ ++struct ib_xrcd; ++ ++struct ib_flow; ++ ++struct ib_flow_attr; ++ ++struct ib_flow_action; ++ ++struct ib_wq; ++ ++struct ib_wq_init_attr; ++ ++struct ib_wq_attr; ++ ++struct ib_rwq_ind_table; ++ ++struct ib_rwq_ind_table_init_attr; ++ ++struct ib_dm; ++ ++struct ib_dm_alloc_attr; ++ ++struct ib_dm_mr_attr; ++ ++struct ib_counters; ++ ++struct ib_counters_read_attr; ++ ++struct ib_device_ops { ++ struct module *owner; ++ enum rdma_driver_id driver_id; ++ u32 uverbs_abi_ver; ++ unsigned int uverbs_no_driver_id_binding: 1; ++ const struct attribute_group *device_group; ++ const struct attribute_group **port_groups; ++ int (*post_send)(struct ib_qp *, const struct ib_send_wr *, const struct ib_send_wr **); ++ int (*post_recv)(struct ib_qp *, const struct ib_recv_wr *, const struct ib_recv_wr **); ++ void (*drain_rq)(struct ib_qp *); ++ void (*drain_sq)(struct ib_qp *); ++ int (*poll_cq)(struct ib_cq *, int, struct ib_wc *); ++ int (*peek_cq)(struct ib_cq *, int); ++ int (*req_notify_cq)(struct ib_cq *, enum ib_cq_notify_flags); ++ int (*post_srq_recv)(struct ib_srq *, const struct ib_recv_wr *, const struct ib_recv_wr **); ++ int (*process_mad)(struct ib_device *, int, u32, const struct ib_wc *, const struct ib_grh *, const struct ib_mad *, struct ib_mad *, size_t *, u16 *); ++ int (*query_device)(struct ib_device *, struct ib_device_attr *, struct ib_udata *); ++ int (*modify_device)(struct ib_device *, int, struct ib_device_modify *); ++ void (*get_dev_fw_str)(struct ib_device *, char *); ++ const struct cpumask * (*get_vector_affinity)(struct ib_device *, int); ++ int (*query_port)(struct ib_device *, u32, struct ib_port_attr *); ++ int (*modify_port)(struct ib_device *, u32, int, struct ib_port_modify *); ++ int (*get_port_immutable)(struct ib_device *, u32, struct ib_port_immutable *); ++ enum rdma_link_layer (*get_link_layer)(struct ib_device *, u32); ++ struct net_device * (*get_netdev)(struct ib_device *, u32); ++ struct net_device * (*alloc_rdma_netdev)(struct ib_device *, u32, enum rdma_netdev_t, const char *, unsigned char, void (*)(struct net_device *)); ++ int (*rdma_netdev_get_params)(struct ib_device *, u32, enum rdma_netdev_t, struct rdma_netdev_alloc_params *); ++ int (*query_gid)(struct ib_device *, u32, int, union ib_gid *); ++ int (*add_gid)(const struct ib_gid_attr *, void **); ++ int (*del_gid)(const struct ib_gid_attr *, void **); ++ int (*query_pkey)(struct ib_device *, u32, u16, u16 *); ++ int (*alloc_ucontext)(struct ib_ucontext *, struct ib_udata *); ++ void (*dealloc_ucontext)(struct ib_ucontext *); ++ int (*mmap)(struct ib_ucontext *, struct vm_area_struct *); ++ void (*mmap_free)(struct rdma_user_mmap_entry *); ++ void (*disassociate_ucontext)(struct ib_ucontext *); ++ int (*alloc_pd)(struct ib_pd *, struct ib_udata *); ++ int (*dealloc_pd)(struct ib_pd *, struct ib_udata *); ++ int (*create_ah)(struct ib_ah *, struct rdma_ah_init_attr *, struct ib_udata *); ++ int (*create_user_ah)(struct ib_ah *, struct rdma_ah_init_attr *, struct ib_udata *); ++ int (*modify_ah)(struct ib_ah *, struct rdma_ah_attr *); ++ int (*query_ah)(struct ib_ah *, struct rdma_ah_attr *); ++ int (*destroy_ah)(struct ib_ah *, u32); ++ int (*create_srq)(struct ib_srq *, struct ib_srq_init_attr *, struct ib_udata *); ++ int (*modify_srq)(struct ib_srq *, struct ib_srq_attr *, enum ib_srq_attr_mask, struct ib_udata *); ++ int (*query_srq)(struct ib_srq *, struct ib_srq_attr *); ++ int (*destroy_srq)(struct ib_srq *, struct ib_udata *); ++ int (*create_qp)(struct ib_qp *, struct ib_qp_init_attr *, struct ib_udata *); ++ int (*modify_qp)(struct ib_qp *, struct ib_qp_attr *, int, struct ib_udata *); ++ int (*query_qp)(struct ib_qp *, struct ib_qp_attr *, int, struct ib_qp_init_attr *); ++ int (*destroy_qp)(struct ib_qp *, struct ib_udata *); ++ int (*create_cq)(struct ib_cq *, const struct ib_cq_init_attr *, struct ib_udata *); ++ int (*modify_cq)(struct ib_cq *, u16, u16); ++ int (*destroy_cq)(struct ib_cq *, struct ib_udata *); ++ int (*resize_cq)(struct ib_cq *, int, struct ib_udata *); ++ struct ib_mr * (*get_dma_mr)(struct ib_pd *, int); ++ struct ib_mr * (*reg_user_mr)(struct ib_pd *, u64, u64, u64, int, struct ib_udata *); ++ struct ib_mr * (*reg_user_mr_dmabuf)(struct ib_pd *, u64, u64, u64, int, int, struct ib_udata *); ++ struct ib_mr * (*rereg_user_mr)(struct ib_mr *, int, u64, u64, u64, int, struct ib_pd *, struct ib_udata *); ++ int (*dereg_mr)(struct ib_mr *, struct ib_udata *); ++ struct ib_mr * (*alloc_mr)(struct ib_pd *, enum ib_mr_type, u32); ++ struct ib_mr * (*alloc_mr_integrity)(struct ib_pd *, u32, u32); ++ int (*advise_mr)(struct ib_pd *, enum ib_uverbs_advise_mr_advice, u32, struct ib_sge *, u32, struct uverbs_attr_bundle *); ++ int (*map_mr_sg)(struct ib_mr *, struct scatterlist *, int, unsigned int *); ++ int (*check_mr_status)(struct ib_mr *, u32, struct ib_mr_status *); ++ int (*alloc_mw)(struct ib_mw *, struct ib_udata *); ++ int (*dealloc_mw)(struct ib_mw *); ++ int (*attach_mcast)(struct ib_qp *, union ib_gid *, u16); ++ int (*detach_mcast)(struct ib_qp *, union ib_gid *, u16); ++ int (*alloc_xrcd)(struct ib_xrcd *, struct ib_udata *); ++ int (*dealloc_xrcd)(struct ib_xrcd *, struct ib_udata *); ++ struct ib_flow * (*create_flow)(struct ib_qp *, struct ib_flow_attr *, struct ib_udata *); ++ int (*destroy_flow)(struct ib_flow *); ++ int (*destroy_flow_action)(struct ib_flow_action *); ++ int (*set_vf_link_state)(struct ib_device *, int, u32, int); ++ int (*get_vf_config)(struct ib_device *, int, u32, struct ifla_vf_info *); ++ int (*get_vf_stats)(struct ib_device *, int, u32, struct ifla_vf_stats *); ++ int (*get_vf_guid)(struct ib_device *, int, u32, struct ifla_vf_guid *, struct ifla_vf_guid *); ++ int (*set_vf_guid)(struct ib_device *, int, u32, u64, int); ++ struct ib_wq * (*create_wq)(struct ib_pd *, struct ib_wq_init_attr *, struct ib_udata *); ++ int (*destroy_wq)(struct ib_wq *, struct ib_udata *); ++ int (*modify_wq)(struct ib_wq *, struct ib_wq_attr *, u32, struct ib_udata *); ++ int (*create_rwq_ind_table)(struct ib_rwq_ind_table *, struct ib_rwq_ind_table_init_attr *, struct ib_udata *); ++ int (*destroy_rwq_ind_table)(struct ib_rwq_ind_table *); ++ struct ib_dm * (*alloc_dm)(struct ib_device *, struct ib_ucontext *, struct ib_dm_alloc_attr *, struct uverbs_attr_bundle *); ++ int (*dealloc_dm)(struct ib_dm *, struct uverbs_attr_bundle *); ++ struct ib_mr * (*reg_dm_mr)(struct ib_pd *, struct ib_dm *, struct ib_dm_mr_attr *, struct uverbs_attr_bundle *); ++ int (*create_counters)(struct ib_counters *, struct uverbs_attr_bundle *); ++ int (*destroy_counters)(struct ib_counters *); ++ int (*read_counters)(struct ib_counters *, struct ib_counters_read_attr *, struct uverbs_attr_bundle *); ++ int (*map_mr_sg_pi)(struct ib_mr *, struct scatterlist *, int, unsigned int *, struct scatterlist *, int, unsigned int *); ++ struct rdma_hw_stats * (*alloc_hw_device_stats)(struct ib_device *); ++ struct rdma_hw_stats * (*alloc_hw_port_stats)(struct ib_device *, u32); ++ int (*get_hw_stats)(struct ib_device *, struct rdma_hw_stats *, u32, int); ++ int (*modify_hw_stat)(struct ib_device *, u32, unsigned int, bool); ++ int (*fill_res_mr_entry)(struct sk_buff *, struct ib_mr *); ++ int (*fill_res_mr_entry_raw)(struct sk_buff *, struct ib_mr *); ++ int (*fill_res_cq_entry)(struct sk_buff *, struct ib_cq *); ++ int (*fill_res_cq_entry_raw)(struct sk_buff *, struct ib_cq *); ++ int (*fill_res_qp_entry)(struct sk_buff *, struct ib_qp *); ++ int (*fill_res_qp_entry_raw)(struct sk_buff *, struct ib_qp *); ++ int (*fill_res_cm_id_entry)(struct sk_buff *, struct rdma_cm_id *); ++ int (*enable_driver)(struct ib_device *); ++ void (*dealloc_driver)(struct ib_device *); ++ void (*iw_add_ref)(struct ib_qp *); ++ void (*iw_rem_ref)(struct ib_qp *); ++ struct ib_qp * (*iw_get_qp)(struct ib_device *, int); ++ int (*iw_connect)(struct iw_cm_id *, struct iw_cm_conn_param *); ++ int (*iw_accept)(struct iw_cm_id *, struct iw_cm_conn_param *); ++ int (*iw_reject)(struct iw_cm_id *, const void *, u8); ++ int (*iw_create_listen)(struct iw_cm_id *, int); ++ int (*iw_destroy_listen)(struct iw_cm_id *); ++ int (*counter_bind_qp)(struct rdma_counter *, struct ib_qp *); ++ int (*counter_unbind_qp)(struct ib_qp *); ++ int (*counter_dealloc)(struct rdma_counter *); ++ struct rdma_hw_stats * (*counter_alloc_stats)(struct rdma_counter *); ++ int (*counter_update_stats)(struct rdma_counter *); ++ int (*fill_stat_mr_entry)(struct sk_buff *, struct ib_mr *); ++ int (*query_ucontext)(struct ib_ucontext *, struct uverbs_attr_bundle *); ++ int (*get_numa_node)(struct ib_device *); ++ size_t size_ib_ah; ++ size_t size_ib_counters; ++ size_t size_ib_cq; ++ size_t size_ib_mw; ++ size_t size_ib_pd; ++ size_t size_ib_qp; ++ size_t size_ib_rwq_ind_table; ++ size_t size_ib_srq; ++ size_t size_ib_ucontext; ++ size_t size_ib_xrcd; ++}; ++ ++struct ib_core_device { ++ struct device dev; ++ possible_net_t rdma_net; ++ struct kobject *ports_kobj; ++ struct list_head port_list; ++ struct ib_device *owner; ++}; ++ ++enum ib_atomic_cap { ++ IB_ATOMIC_NONE = 0, ++ IB_ATOMIC_HCA = 1, ++ IB_ATOMIC_GLOB = 2, ++}; ++ ++struct ib_odp_caps { ++ uint64_t general_caps; ++ struct { ++ uint32_t rc_odp_caps; ++ uint32_t uc_odp_caps; ++ uint32_t ud_odp_caps; ++ uint32_t xrc_odp_caps; ++ } per_transport_caps; ++}; ++ ++struct ib_rss_caps { ++ u32 supported_qpts; ++ u32 max_rwq_indirection_tables; ++ u32 max_rwq_indirection_table_size; ++}; ++ ++struct ib_tm_caps { ++ u32 max_rndv_hdr_size; ++ u32 max_num_tags; ++ u32 flags; ++ u32 max_ops; ++ u32 max_sge; ++}; ++ ++struct ib_cq_caps { ++ u16 max_cq_moderation_count; ++ u16 max_cq_moderation_period; ++}; ++ ++struct ib_device_attr { ++ u64 fw_ver; ++ __be64 sys_image_guid; ++ u64 max_mr_size; ++ u64 page_size_cap; ++ u32 vendor_id; ++ u32 vendor_part_id; ++ u32 hw_ver; ++ int max_qp; ++ int max_qp_wr; ++ u64 device_cap_flags; ++ u64 kernel_cap_flags; ++ int max_send_sge; ++ int max_recv_sge; ++ int max_sge_rd; ++ int max_cq; ++ int max_cqe; ++ int max_mr; ++ int max_pd; ++ int max_qp_rd_atom; ++ int max_ee_rd_atom; ++ int max_res_rd_atom; ++ int max_qp_init_rd_atom; ++ int max_ee_init_rd_atom; ++ enum ib_atomic_cap atomic_cap; ++ enum ib_atomic_cap masked_atomic_cap; ++ int max_ee; ++ int max_rdd; ++ int max_mw; ++ int max_raw_ipv6_qp; ++ int max_raw_ethy_qp; ++ int max_mcast_grp; ++ int max_mcast_qp_attach; ++ int max_total_mcast_qp_attach; ++ int max_ah; ++ int max_srq; ++ int max_srq_wr; ++ int max_srq_sge; ++ unsigned int max_fast_reg_page_list_len; ++ unsigned int max_pi_fast_reg_page_list_len; ++ u16 max_pkeys; ++ u8 local_ca_ack_delay; ++ int sig_prot_cap; ++ int sig_guard_cap; ++ struct ib_odp_caps odp_caps; ++ uint64_t timestamp_mask; ++ uint64_t hca_core_clock; ++ struct ib_rss_caps rss_caps; ++ u32 max_wq_type_rq; ++ u32 raw_packet_caps; ++ struct ib_tm_caps tm_caps; ++ struct ib_cq_caps cq_caps; ++ u64 max_dm_size; ++ u32 max_sgl_rd; ++}; ++ ++struct hw_stats_device_data; ++ ++struct rdma_restrack_root; ++ ++struct uapi_definition; ++ ++struct ib_port_data; ++ ++struct ib_device { ++ struct device *dma_device; ++ struct ib_device_ops ops; ++ char name[64]; ++ struct callback_head callback_head; ++ struct list_head event_handler_list; ++ struct rw_semaphore event_handler_rwsem; ++ spinlock_t qp_open_list_lock; ++ struct rw_semaphore client_data_rwsem; ++ struct xarray client_data; ++ struct mutex unregistration_lock; ++ rwlock_t cache_lock; ++ struct ib_port_data *port_data; ++ int num_comp_vectors; ++ union { ++ struct device dev; ++ struct ib_core_device coredev; ++ }; ++ const struct attribute_group *groups[4]; ++ u64 uverbs_cmd_mask; ++ char node_desc[64]; ++ __be64 node_guid; ++ u32 local_dma_lkey; ++ u16 is_switch: 1; ++ u16 kverbs_provider: 1; ++ u16 use_cq_dim: 1; ++ u8 node_type; ++ u32 phys_port_cnt; ++ struct ib_device_attr attrs; ++ struct hw_stats_device_data *hw_stats_data; ++ struct rdmacg_device cg_device; ++ u32 index; ++ spinlock_t cq_pools_lock; ++ struct list_head cq_pools[3]; ++ struct rdma_restrack_root *res; ++ const struct uapi_definition *driver_def; ++ refcount_t refcount; ++ struct completion unreg_completion; ++ struct work_struct unregistration_work; ++ const struct rdma_link_ops *link_ops; ++ struct mutex compat_devs_mutex; ++ struct xarray compat_devs; ++ char iw_ifname[16]; ++ u32 iw_driver_flags; ++ u32 lag_flags; ++}; ++ ++enum ib_signature_type { ++ IB_SIG_TYPE_NONE = 0, ++ IB_SIG_TYPE_T10_DIF = 1, ++}; ++ ++enum ib_t10_dif_bg_type { ++ IB_T10DIF_CRC = 0, ++ IB_T10DIF_CSUM = 1, ++}; ++ ++struct ib_t10_dif_domain { ++ enum ib_t10_dif_bg_type bg_type; ++ u16 pi_interval; ++ u16 bg; ++ u16 app_tag; ++ u32 ref_tag; ++ bool ref_remap; ++ bool app_escape; ++ bool ref_escape; ++ u16 apptag_check_mask; ++}; ++ ++struct ib_sig_domain { ++ enum ib_signature_type sig_type; ++ union { ++ struct ib_t10_dif_domain dif; ++ } sig; ++}; ++ ++struct ib_sig_attrs { ++ u8 check_mask; ++ struct ib_sig_domain mem; ++ struct ib_sig_domain wire; ++ int meta_length; ++}; ++ ++enum ib_sig_err_type { ++ IB_SIG_BAD_GUARD = 0, ++ IB_SIG_BAD_REFTAG = 1, ++ IB_SIG_BAD_APPTAG = 2, ++}; ++ ++struct ib_sig_err { ++ enum ib_sig_err_type err_type; ++ u32 expected; ++ u32 actual; ++ u64 sig_err_offset; ++ u32 key; ++}; ++ ++union ib_gid { ++ u8 raw[16]; ++ struct { ++ __be64 subnet_prefix; ++ __be64 interface_id; ++ } global; ++}; ++ ++enum ib_gid_type { ++ IB_GID_TYPE_IB = 0, ++ IB_GID_TYPE_ROCE = 1, ++ IB_GID_TYPE_ROCE_UDP_ENCAP = 2, ++ IB_GID_TYPE_SIZE = 3, ++}; ++ ++struct ib_gid_attr { ++ struct net_device *ndev; ++ struct ib_device *device; ++ union ib_gid gid; ++ enum ib_gid_type gid_type; ++ u16 index; ++ u32 port_num; ++}; ++ ++struct ib_cq_init_attr { ++ unsigned int cqe; ++ u32 comp_vector; ++ u32 flags; ++}; ++ ++struct ib_dm_mr_attr { ++ u64 length; ++ u64 offset; ++ u32 access_flags; ++}; ++ ++struct ib_dm_alloc_attr { ++ u64 length; ++ u32 alignment; ++ u32 flags; ++}; ++ ++enum ib_mtu { ++ IB_MTU_256 = 1, ++ IB_MTU_512 = 2, ++ IB_MTU_1024 = 3, ++ IB_MTU_2048 = 4, ++ IB_MTU_4096 = 5, ++}; ++ ++enum ib_port_state { ++ IB_PORT_NOP = 0, ++ IB_PORT_DOWN = 1, ++ IB_PORT_INIT = 2, ++ IB_PORT_ARMED = 3, ++ IB_PORT_ACTIVE = 4, ++ IB_PORT_ACTIVE_DEFER = 5, ++}; ++ ++struct rdma_stat_desc { ++ const char *name; ++ unsigned int flags; ++ const void *priv; ++}; ++ ++struct ib_port_attr { ++ u64 subnet_prefix; ++ enum ib_port_state state; ++ enum ib_mtu max_mtu; ++ enum ib_mtu active_mtu; ++ u32 phys_mtu; ++ int gid_tbl_len; ++ unsigned int ip_gids: 1; ++ u32 port_cap_flags; ++ u32 max_msg_sz; ++ u32 bad_pkey_cntr; ++ u32 qkey_viol_cntr; ++ u16 pkey_tbl_len; ++ u32 sm_lid; ++ u32 lid; ++ u8 lmc; ++ u8 max_vl_num; ++ u8 sm_sl; ++ u8 subnet_timeout; ++ u8 init_type_reply; ++ u8 active_width; ++ u16 active_speed; ++ u8 phys_state; ++ u16 port_cap_flags2; ++}; ++ ++struct ib_device_modify { ++ u64 sys_image_guid; ++ char node_desc[64]; ++}; ++ ++struct ib_port_modify { ++ u32 set_port_cap_mask; ++ u32 clr_port_cap_mask; ++ u8 init_type; ++}; ++ ++enum ib_event_type { ++ IB_EVENT_CQ_ERR = 0, ++ IB_EVENT_QP_FATAL = 1, ++ IB_EVENT_QP_REQ_ERR = 2, ++ IB_EVENT_QP_ACCESS_ERR = 3, ++ IB_EVENT_COMM_EST = 4, ++ IB_EVENT_SQ_DRAINED = 5, ++ IB_EVENT_PATH_MIG = 6, ++ IB_EVENT_PATH_MIG_ERR = 7, ++ IB_EVENT_DEVICE_FATAL = 8, ++ IB_EVENT_PORT_ACTIVE = 9, ++ IB_EVENT_PORT_ERR = 10, ++ IB_EVENT_LID_CHANGE = 11, ++ IB_EVENT_PKEY_CHANGE = 12, ++ IB_EVENT_SM_CHANGE = 13, ++ IB_EVENT_SRQ_ERR = 14, ++ IB_EVENT_SRQ_LIMIT_REACHED = 15, ++ IB_EVENT_QP_LAST_WQE_REACHED = 16, ++ IB_EVENT_CLIENT_REREGISTER = 17, ++ IB_EVENT_GID_CHANGE = 18, ++ IB_EVENT_WQ_FATAL = 19, ++}; ++ ++struct ib_ucq_object; ++ ++typedef void (*ib_comp_handler)(struct ib_cq *, void *); ++ ++struct ib_event; ++ ++struct ib_cq { ++ struct ib_device *device; ++ struct ib_ucq_object *uobject; ++ ib_comp_handler comp_handler; ++ void (*event_handler)(struct ib_event *, void *); ++ void *cq_context; ++ int cqe; ++ unsigned int cqe_used; ++ atomic_t usecnt; ++ enum ib_poll_context poll_ctx; ++ struct ib_wc *wc; ++ struct list_head pool_entry; ++ union { ++ struct irq_poll iop; ++ struct work_struct work; ++ }; ++ struct workqueue_struct *comp_wq; ++ struct dim *dim; ++ ktime_t timestamp; ++ u8 interrupt: 1; ++ u8 shared: 1; ++ unsigned int comp_vector; ++ struct rdma_restrack_entry res; ++}; ++ ++struct ib_uqp_object; ++ ++enum ib_qp_type { ++ IB_QPT_SMI = 0, ++ IB_QPT_GSI = 1, ++ IB_QPT_RC = 2, ++ IB_QPT_UC = 3, ++ IB_QPT_UD = 4, ++ IB_QPT_RAW_IPV6 = 5, ++ IB_QPT_RAW_ETHERTYPE = 6, ++ IB_QPT_RAW_PACKET = 8, ++ IB_QPT_XRC_INI = 9, ++ IB_QPT_XRC_TGT = 10, ++ IB_QPT_MAX = 11, ++ IB_QPT_DRIVER = 255, ++ IB_QPT_RESERVED1 = 4096, ++ IB_QPT_RESERVED2 = 4097, ++ IB_QPT_RESERVED3 = 4098, ++ IB_QPT_RESERVED4 = 4099, ++ IB_QPT_RESERVED5 = 4100, ++ IB_QPT_RESERVED6 = 4101, ++ IB_QPT_RESERVED7 = 4102, ++ IB_QPT_RESERVED8 = 4103, ++ IB_QPT_RESERVED9 = 4104, ++ IB_QPT_RESERVED10 = 4105, ++}; ++ ++struct ib_qp_security; ++ ++struct ib_qp { ++ struct ib_device *device; ++ struct ib_pd *pd; ++ struct ib_cq *send_cq; ++ struct ib_cq *recv_cq; ++ spinlock_t mr_lock; ++ int mrs_used; ++ struct list_head rdma_mrs; ++ struct list_head sig_mrs; ++ struct ib_srq *srq; ++ struct ib_xrcd *xrcd; ++ struct list_head xrcd_list; ++ atomic_t usecnt; ++ struct list_head open_list; ++ struct ib_qp *real_qp; ++ struct ib_uqp_object *uobject; ++ void (*event_handler)(struct ib_event *, void *); ++ void *qp_context; ++ const struct ib_gid_attr *av_sgid_attr; ++ const struct ib_gid_attr *alt_path_sgid_attr; ++ u32 qp_num; ++ u32 max_write_sge; ++ u32 max_read_sge; ++ enum ib_qp_type qp_type; ++ struct ib_rwq_ind_table *rwq_ind_tbl; ++ struct ib_qp_security *qp_sec; ++ u32 port; ++ bool integrity_en; ++ struct rdma_restrack_entry res; ++ struct rdma_counter *counter; ++}; ++ ++struct ib_usrq_object; ++ ++enum ib_srq_type { ++ IB_SRQT_BASIC = 0, ++ IB_SRQT_XRC = 1, ++ IB_SRQT_TM = 2, ++}; ++ ++struct ib_srq { ++ struct ib_device *device; ++ struct ib_pd *pd; ++ struct ib_usrq_object *uobject; ++ void (*event_handler)(struct ib_event *, void *); ++ void *srq_context; ++ enum ib_srq_type srq_type; ++ atomic_t usecnt; ++ struct { ++ struct ib_cq *cq; ++ union { ++ struct { ++ struct ib_xrcd *xrcd; ++ u32 srq_num; ++ } xrc; ++ }; ++ } ext; ++ struct rdma_restrack_entry res; ++}; ++ ++struct ib_uwq_object; ++ ++enum ib_wq_state { ++ IB_WQS_RESET = 0, ++ IB_WQS_RDY = 1, ++ IB_WQS_ERR = 2, ++}; ++ ++enum ib_wq_type { ++ IB_WQT_RQ = 0, ++}; ++ ++struct ib_wq { ++ struct ib_device *device; ++ struct ib_uwq_object *uobject; ++ void *wq_context; ++ void (*event_handler)(struct ib_event *, void *); ++ struct ib_pd *pd; ++ struct ib_cq *cq; ++ u32 wq_num; ++ enum ib_wq_state state; ++ enum ib_wq_type wq_type; ++ atomic_t usecnt; ++}; ++ ++struct ib_event { ++ struct ib_device *device; ++ union { ++ struct ib_cq *cq; ++ struct ib_qp *qp; ++ struct ib_srq *srq; ++ struct ib_wq *wq; ++ u32 port_num; ++ } element; ++ enum ib_event_type event; ++}; ++ ++struct ib_global_route { ++ const struct ib_gid_attr *sgid_attr; ++ union ib_gid dgid; ++ u32 flow_label; ++ u8 sgid_index; ++ u8 hop_limit; ++ u8 traffic_class; ++}; ++ ++struct ib_grh { ++ __be32 version_tclass_flow; ++ __be16 paylen; ++ u8 next_hdr; ++ u8 hop_limit; ++ union ib_gid sgid; ++ union ib_gid dgid; ++}; ++ ++struct ib_mr_status { ++ u32 fail_status; ++ struct ib_sig_err sig_err; ++}; ++ ++struct rdma_ah_init_attr { ++ struct rdma_ah_attr *ah_attr; ++ u32 flags; ++ struct net_device *xmit_slave; ++}; ++ ++enum rdma_ah_attr_type { ++ RDMA_AH_ATTR_TYPE_UNDEFINED = 0, ++ RDMA_AH_ATTR_TYPE_IB = 1, ++ RDMA_AH_ATTR_TYPE_ROCE = 2, ++ RDMA_AH_ATTR_TYPE_OPA = 3, ++}; ++ ++struct ib_ah_attr { ++ u16 dlid; ++ u8 src_path_bits; ++}; ++ ++struct roce_ah_attr { ++ u8 dmac[6]; ++}; ++ ++struct opa_ah_attr { ++ u32 dlid; ++ u8 src_path_bits; ++ bool make_grd; ++}; ++ ++struct rdma_ah_attr { ++ struct ib_global_route grh; ++ u8 sl; ++ u8 static_rate; ++ u32 port_num; ++ u8 ah_flags; ++ enum rdma_ah_attr_type type; ++ union { ++ struct ib_ah_attr ib; ++ struct roce_ah_attr roce; ++ struct opa_ah_attr opa; ++ }; ++}; ++ ++enum ib_wc_status { ++ IB_WC_SUCCESS = 0, ++ IB_WC_LOC_LEN_ERR = 1, ++ IB_WC_LOC_QP_OP_ERR = 2, ++ IB_WC_LOC_EEC_OP_ERR = 3, ++ IB_WC_LOC_PROT_ERR = 4, ++ IB_WC_WR_FLUSH_ERR = 5, ++ IB_WC_MW_BIND_ERR = 6, ++ IB_WC_BAD_RESP_ERR = 7, ++ IB_WC_LOC_ACCESS_ERR = 8, ++ IB_WC_REM_INV_REQ_ERR = 9, ++ IB_WC_REM_ACCESS_ERR = 10, ++ IB_WC_REM_OP_ERR = 11, ++ IB_WC_RETRY_EXC_ERR = 12, ++ IB_WC_RNR_RETRY_EXC_ERR = 13, ++ IB_WC_LOC_RDD_VIOL_ERR = 14, ++ IB_WC_REM_INV_RD_REQ_ERR = 15, ++ IB_WC_REM_ABORT_ERR = 16, ++ IB_WC_INV_EECN_ERR = 17, ++ IB_WC_INV_EEC_STATE_ERR = 18, ++ IB_WC_FATAL_ERR = 19, ++ IB_WC_RESP_TIMEOUT_ERR = 20, ++ IB_WC_GENERAL_ERR = 21, ++}; ++ ++enum ib_wc_opcode { ++ IB_WC_SEND = 0, ++ IB_WC_RDMA_WRITE = 1, ++ IB_WC_RDMA_READ = 2, ++ IB_WC_COMP_SWAP = 3, ++ IB_WC_FETCH_ADD = 4, ++ IB_WC_BIND_MW = 5, ++ IB_WC_LOCAL_INV = 6, ++ IB_WC_LSO = 7, ++ IB_WC_ATOMIC_WRITE = 9, ++ IB_WC_REG_MR = 10, ++ IB_WC_MASKED_COMP_SWAP = 11, ++ IB_WC_MASKED_FETCH_ADD = 12, ++ IB_WC_FLUSH = 8, ++ IB_WC_RECV = 128, ++ IB_WC_RECV_RDMA_WITH_IMM = 129, ++}; ++ ++struct ib_cqe { ++ void (*done)(struct ib_cq *, struct ib_wc *); ++}; ++ ++struct ib_wc { ++ union { ++ u64 wr_id; ++ struct ib_cqe *wr_cqe; ++ }; ++ enum ib_wc_status status; ++ enum ib_wc_opcode opcode; ++ u32 vendor_err; ++ u32 byte_len; ++ struct ib_qp *qp; ++ union { ++ __be32 imm_data; ++ u32 invalidate_rkey; ++ } ex; ++ u32 src_qp; ++ u32 slid; ++ int wc_flags; ++ u16 pkey_index; ++ u8 sl; ++ u8 dlid_path_bits; ++ u32 port_num; ++ u8 smac[6]; ++ u16 vlan_id; ++ u8 network_hdr_type; ++}; ++ ++struct ib_srq_attr { ++ u32 max_wr; ++ u32 max_sge; ++ u32 srq_limit; ++}; ++ ++struct ib_xrcd { ++ struct ib_device *device; ++ atomic_t usecnt; ++ struct inode *inode; ++ struct rw_semaphore tgt_qps_rwsem; ++ struct xarray tgt_qps; ++}; ++ ++struct ib_srq_init_attr { ++ void (*event_handler)(struct ib_event *, void *); ++ void *srq_context; ++ struct ib_srq_attr attr; ++ enum ib_srq_type srq_type; ++ struct { ++ struct ib_cq *cq; ++ union { ++ struct { ++ struct ib_xrcd *xrcd; ++ } xrc; ++ struct { ++ u32 max_num_tags; ++ } tag_matching; ++ }; ++ } ext; ++}; ++ ++struct ib_qp_cap { ++ u32 max_send_wr; ++ u32 max_recv_wr; ++ u32 max_send_sge; ++ u32 max_recv_sge; ++ u32 max_inline_data; ++ u32 max_rdma_ctxs; ++}; ++ ++enum ib_sig_type { ++ IB_SIGNAL_ALL_WR = 0, ++ IB_SIGNAL_REQ_WR = 1, ++}; ++ ++struct ib_qp_init_attr { ++ void (*event_handler)(struct ib_event *, void *); ++ void *qp_context; ++ struct ib_cq *send_cq; ++ struct ib_cq *recv_cq; ++ struct ib_srq *srq; ++ struct ib_xrcd *xrcd; ++ struct ib_qp_cap cap; ++ enum ib_sig_type sq_sig_type; ++ enum ib_qp_type qp_type; ++ u32 create_flags; ++ u32 port_num; ++ struct ib_rwq_ind_table *rwq_ind_tbl; ++ u32 source_qpn; ++}; ++ ++struct ib_uobject; ++ ++struct ib_rwq_ind_table { ++ struct ib_device *device; ++ struct ib_uobject *uobject; ++ atomic_t usecnt; ++ u32 ind_tbl_num; ++ u32 log_ind_tbl_size; ++ struct ib_wq **ind_tbl; ++}; ++ ++enum ib_qp_state { ++ IB_QPS_RESET = 0, ++ IB_QPS_INIT = 1, ++ IB_QPS_RTR = 2, ++ IB_QPS_RTS = 3, ++ IB_QPS_SQD = 4, ++ IB_QPS_SQE = 5, ++ IB_QPS_ERR = 6, ++}; ++ ++enum ib_mig_state { ++ IB_MIG_MIGRATED = 0, ++ IB_MIG_REARM = 1, ++ IB_MIG_ARMED = 2, ++}; ++ ++enum ib_mw_type { ++ IB_MW_TYPE_1 = 1, ++ IB_MW_TYPE_2 = 2, ++}; ++ ++struct ib_qp_attr { ++ enum ib_qp_state qp_state; ++ enum ib_qp_state cur_qp_state; ++ enum ib_mtu path_mtu; ++ enum ib_mig_state path_mig_state; ++ u32 qkey; ++ u32 rq_psn; ++ u32 sq_psn; ++ u32 dest_qp_num; ++ int qp_access_flags; ++ struct ib_qp_cap cap; ++ struct rdma_ah_attr ah_attr; ++ struct rdma_ah_attr alt_ah_attr; ++ u16 pkey_index; ++ u16 alt_pkey_index; ++ u8 en_sqd_async_notify; ++ u8 sq_draining; ++ u8 max_rd_atomic; ++ u8 max_dest_rd_atomic; ++ u8 min_rnr_timer; ++ u32 port_num; ++ u8 timeout; ++ u8 retry_cnt; ++ u8 rnr_retry; ++ u32 alt_port_num; ++ u8 alt_timeout; ++ u32 rate_limit; ++ struct net_device *xmit_slave; ++}; ++ ++enum ib_wr_opcode { ++ IB_WR_RDMA_WRITE = 0, ++ IB_WR_RDMA_WRITE_WITH_IMM = 1, ++ IB_WR_SEND = 2, ++ IB_WR_SEND_WITH_IMM = 3, ++ IB_WR_RDMA_READ = 4, ++ IB_WR_ATOMIC_CMP_AND_SWP = 5, ++ IB_WR_ATOMIC_FETCH_AND_ADD = 6, ++ IB_WR_BIND_MW = 8, ++ IB_WR_LSO = 10, ++ IB_WR_SEND_WITH_INV = 9, ++ IB_WR_RDMA_READ_WITH_INV = 11, ++ IB_WR_LOCAL_INV = 7, ++ IB_WR_MASKED_ATOMIC_CMP_AND_SWP = 12, ++ IB_WR_MASKED_ATOMIC_FETCH_AND_ADD = 13, ++ IB_WR_FLUSH = 14, ++ IB_WR_ATOMIC_WRITE = 15, ++ IB_WR_REG_MR = 32, ++ IB_WR_REG_MR_INTEGRITY = 33, ++ IB_WR_RESERVED1 = 240, ++ IB_WR_RESERVED2 = 241, ++ IB_WR_RESERVED3 = 242, ++ IB_WR_RESERVED4 = 243, ++ IB_WR_RESERVED5 = 244, ++ IB_WR_RESERVED6 = 245, ++ IB_WR_RESERVED7 = 246, ++ IB_WR_RESERVED8 = 247, ++ IB_WR_RESERVED9 = 248, ++ IB_WR_RESERVED10 = 249, ++}; ++ ++struct ib_sge { ++ u64 addr; ++ u32 length; ++ u32 lkey; ++}; ++ ++struct ib_send_wr { ++ struct ib_send_wr *next; ++ union { ++ u64 wr_id; ++ struct ib_cqe *wr_cqe; ++ }; ++ struct ib_sge *sg_list; ++ int num_sge; ++ enum ib_wr_opcode opcode; ++ int send_flags; ++ union { ++ __be32 imm_data; ++ u32 invalidate_rkey; ++ } ex; ++}; ++ ++struct ib_ah { ++ struct ib_device *device; ++ struct ib_pd *pd; ++ struct ib_uobject *uobject; ++ const struct ib_gid_attr *sgid_attr; ++ enum rdma_ah_attr_type type; ++}; ++ ++struct ib_mr { ++ struct ib_device *device; ++ struct ib_pd *pd; ++ u32 lkey; ++ u32 rkey; ++ u64 iova; ++ u64 length; ++ unsigned int page_size; ++ enum ib_mr_type type; ++ bool need_inval; ++ union { ++ struct ib_uobject *uobject; ++ struct list_head qp_entry; ++ }; ++ struct ib_dm *dm; ++ struct ib_sig_attrs *sig_attrs; ++ struct rdma_restrack_entry res; ++}; ++ ++struct ib_recv_wr { ++ struct ib_recv_wr *next; ++ union { ++ u64 wr_id; ++ struct ib_cqe *wr_cqe; ++ }; ++ struct ib_sge *sg_list; ++ int num_sge; ++}; ++ ++struct ib_rdmacg_object { ++ struct rdma_cgroup *cg; ++}; ++ ++struct ib_uverbs_file; ++ ++struct ib_ucontext { ++ struct ib_device *device; ++ struct ib_uverbs_file *ufile; ++ struct ib_rdmacg_object cg_obj; ++ struct rdma_restrack_entry res; ++ struct xarray mmap_xa; ++}; ++ ++struct uverbs_api_object; ++ ++struct ib_uobject { ++ u64 user_handle; ++ struct ib_uverbs_file *ufile; ++ struct ib_ucontext *context; ++ void *object; ++ struct list_head list; ++ struct ib_rdmacg_object cg_obj; ++ int id; ++ struct kref ref; ++ atomic_t usecnt; ++ struct callback_head rcu; ++ const struct uverbs_api_object *uapi_object; ++}; ++ ++struct ib_udata { ++ const void *inbuf; ++ void *outbuf; ++ size_t inlen; ++ size_t outlen; ++}; ++ ++struct ib_pd { ++ u32 local_dma_lkey; ++ u32 flags; ++ struct ib_device *device; ++ struct ib_uobject *uobject; ++ atomic_t usecnt; ++ u32 unsafe_global_rkey; ++ struct ib_mr *__internal_mr; ++ struct rdma_restrack_entry res; ++}; ++ ++struct ib_wq_init_attr { ++ void *wq_context; ++ enum ib_wq_type wq_type; ++ u32 max_wr; ++ u32 max_sge; ++ struct ib_cq *cq; ++ void (*event_handler)(struct ib_event *, void *); ++ u32 create_flags; ++}; ++ ++struct ib_wq_attr { ++ enum ib_wq_state wq_state; ++ enum ib_wq_state curr_wq_state; ++ u32 flags; ++ u32 flags_mask; ++}; ++ ++struct ib_rwq_ind_table_init_attr { ++ u32 log_ind_tbl_size; ++ struct ib_wq **ind_tbl; ++}; ++ ++enum port_pkey_state { ++ IB_PORT_PKEY_NOT_VALID = 0, ++ IB_PORT_PKEY_VALID = 1, ++ IB_PORT_PKEY_LISTED = 2, ++}; ++ ++struct ib_port_pkey { ++ enum port_pkey_state state; ++ u16 pkey_index; ++ u32 port_num; ++ struct list_head qp_list; ++ struct list_head to_error_list; ++ struct ib_qp_security *sec; ++}; ++ ++struct ib_ports_pkeys; ++ ++struct ib_qp_security { ++ struct ib_qp *qp; ++ struct ib_device *dev; ++ struct mutex mutex; ++ struct ib_ports_pkeys *ports_pkeys; ++ struct list_head shared_qp_list; ++ void *security; ++ bool destroying; ++ atomic_t error_list_count; ++ struct completion error_complete; ++ int error_comps_pending; ++}; ++ ++struct ib_ports_pkeys { ++ struct ib_port_pkey main; ++ struct ib_port_pkey alt; ++}; ++ ++struct ib_dm { ++ struct ib_device *device; ++ u32 length; ++ u32 flags; ++ struct ib_uobject *uobject; ++ atomic_t usecnt; ++}; ++ ++struct ib_mw { ++ struct ib_device *device; ++ struct ib_pd *pd; ++ struct ib_uobject *uobject; ++ u32 rkey; ++ enum ib_mw_type type; ++}; ++ ++enum ib_flow_attr_type { ++ IB_FLOW_ATTR_NORMAL = 0, ++ IB_FLOW_ATTR_ALL_DEFAULT = 1, ++ IB_FLOW_ATTR_MC_DEFAULT = 2, ++ IB_FLOW_ATTR_SNIFFER = 3, ++}; ++ ++enum ib_flow_spec_type { ++ IB_FLOW_SPEC_ETH = 32, ++ IB_FLOW_SPEC_IB = 34, ++ IB_FLOW_SPEC_IPV4 = 48, ++ IB_FLOW_SPEC_IPV6 = 49, ++ IB_FLOW_SPEC_ESP = 52, ++ IB_FLOW_SPEC_TCP = 64, ++ IB_FLOW_SPEC_UDP = 65, ++ IB_FLOW_SPEC_VXLAN_TUNNEL = 80, ++ IB_FLOW_SPEC_GRE = 81, ++ IB_FLOW_SPEC_MPLS = 96, ++ IB_FLOW_SPEC_INNER = 256, ++ IB_FLOW_SPEC_ACTION_TAG = 4096, ++ IB_FLOW_SPEC_ACTION_DROP = 4097, ++ IB_FLOW_SPEC_ACTION_HANDLE = 4098, ++ IB_FLOW_SPEC_ACTION_COUNT = 4099, ++}; ++ ++struct ib_flow_eth_filter { ++ u8 dst_mac[6]; ++ u8 src_mac[6]; ++ __be16 ether_type; ++ __be16 vlan_tag; ++ u8 real_sz[0]; ++}; ++ ++struct ib_flow_spec_eth { ++ u32 type; ++ u16 size; ++ struct ib_flow_eth_filter val; ++ struct ib_flow_eth_filter mask; ++}; ++ ++struct ib_flow_ib_filter { ++ __be16 dlid; ++ __u8 sl; ++ u8 real_sz[0]; ++}; ++ ++struct ib_flow_spec_ib { ++ u32 type; ++ u16 size; ++ struct ib_flow_ib_filter val; ++ struct ib_flow_ib_filter mask; ++}; ++ ++struct ib_flow_ipv4_filter { ++ __be32 src_ip; ++ __be32 dst_ip; ++ u8 proto; ++ u8 tos; ++ u8 ttl; ++ u8 flags; ++ u8 real_sz[0]; ++}; ++ ++struct ib_flow_spec_ipv4 { ++ u32 type; ++ u16 size; ++ struct ib_flow_ipv4_filter val; ++ struct ib_flow_ipv4_filter mask; ++}; ++ ++struct ib_flow_ipv6_filter { ++ u8 src_ip[16]; ++ u8 dst_ip[16]; ++ __be32 flow_label; ++ u8 next_hdr; ++ u8 traffic_class; ++ u8 hop_limit; ++ u8 real_sz[0]; ++}; ++ ++struct ib_flow_spec_ipv6 { ++ u32 type; ++ u16 size; ++ struct ib_flow_ipv6_filter val; ++ struct ib_flow_ipv6_filter mask; ++}; ++ ++struct ib_flow_tcp_udp_filter { ++ __be16 dst_port; ++ __be16 src_port; ++ u8 real_sz[0]; ++}; ++ ++struct ib_flow_spec_tcp_udp { ++ u32 type; ++ u16 size; ++ struct ib_flow_tcp_udp_filter val; ++ struct ib_flow_tcp_udp_filter mask; ++}; ++ ++struct ib_flow_tunnel_filter { ++ __be32 tunnel_id; ++ u8 real_sz[0]; ++}; ++ ++struct ib_flow_spec_tunnel { ++ u32 type; ++ u16 size; ++ struct ib_flow_tunnel_filter val; ++ struct ib_flow_tunnel_filter mask; ++}; ++ ++struct ib_flow_esp_filter { ++ __be32 spi; ++ __be32 seq; ++ u8 real_sz[0]; ++}; ++ ++struct ib_flow_spec_esp { ++ u32 type; ++ u16 size; ++ struct ib_flow_esp_filter val; ++ struct ib_flow_esp_filter mask; ++}; ++ ++struct ib_flow_gre_filter { ++ __be16 c_ks_res0_ver; ++ __be16 protocol; ++ __be32 key; ++ u8 real_sz[0]; ++}; ++ ++struct ib_flow_spec_gre { ++ u32 type; ++ u16 size; ++ struct ib_flow_gre_filter val; ++ struct ib_flow_gre_filter mask; ++}; ++ ++struct ib_flow_mpls_filter { ++ __be32 tag; ++ u8 real_sz[0]; ++}; ++ ++struct ib_flow_spec_mpls { ++ u32 type; ++ u16 size; ++ struct ib_flow_mpls_filter val; ++ struct ib_flow_mpls_filter mask; ++}; ++ ++struct ib_flow_spec_action_tag { ++ enum ib_flow_spec_type type; ++ u16 size; ++ u32 tag_id; ++}; ++ ++struct ib_flow_spec_action_drop { ++ enum ib_flow_spec_type type; ++ u16 size; ++}; ++ ++struct ib_flow_spec_action_handle { ++ enum ib_flow_spec_type type; ++ u16 size; ++ struct ib_flow_action *act; ++}; ++ ++enum ib_flow_action_type { ++ IB_FLOW_ACTION_UNSPECIFIED = 0, ++ IB_FLOW_ACTION_ESP = 1, ++}; ++ ++struct ib_flow_action { ++ struct ib_device *device; ++ struct ib_uobject *uobject; ++ enum ib_flow_action_type type; ++ atomic_t usecnt; ++}; ++ ++struct ib_flow_spec_action_count { ++ enum ib_flow_spec_type type; ++ u16 size; ++ struct ib_counters *counters; ++}; ++ ++struct ib_counters { ++ struct ib_device *device; ++ struct ib_uobject *uobject; ++ atomic_t usecnt; ++}; ++ ++union ib_flow_spec { ++ struct { ++ u32 type; ++ u16 size; ++ }; ++ struct ib_flow_spec_eth eth; ++ struct ib_flow_spec_ib ib; ++ struct ib_flow_spec_ipv4 ipv4; ++ struct ib_flow_spec_tcp_udp tcp_udp; ++ struct ib_flow_spec_ipv6 ipv6; ++ struct ib_flow_spec_tunnel tunnel; ++ struct ib_flow_spec_esp esp; ++ struct ib_flow_spec_gre gre; ++ struct ib_flow_spec_mpls mpls; ++ struct ib_flow_spec_action_tag flow_tag; ++ struct ib_flow_spec_action_drop drop; ++ struct ib_flow_spec_action_handle action; ++ struct ib_flow_spec_action_count flow_count; ++}; ++ ++struct ib_flow_attr { ++ enum ib_flow_attr_type type; ++ u16 size; ++ u16 priority; ++ u32 flags; ++ u8 num_of_specs; ++ u32 port; ++ union ib_flow_spec flows[0]; ++}; ++ ++struct ib_flow { ++ struct ib_qp *qp; ++ struct ib_device *device; ++ struct ib_uobject *uobject; ++}; ++ ++struct ib_pkey_cache; ++ ++struct ib_gid_table; ++ ++struct ib_port_cache { ++ u64 subnet_prefix; ++ struct ib_pkey_cache *pkey; ++ struct ib_gid_table *gid; ++ u8 lmc; ++ enum ib_port_state port_state; ++}; ++ ++struct ib_port_immutable { ++ int pkey_tbl_len; ++ int gid_tbl_len; ++ u32 core_cap_flags; ++ u32 max_mad_size; ++}; ++ ++struct ib_port; ++ ++struct ib_port_data { ++ struct ib_device *ib_dev; ++ struct ib_port_immutable immutable; ++ spinlock_t pkey_list_lock; ++ spinlock_t netdev_lock; ++ struct list_head pkey_list; ++ struct ib_port_cache cache; ++ struct net_device *netdev; ++ netdevice_tracker netdev_tracker; ++ struct hlist_node ndev_hash_link; ++ struct rdma_port_counter port_counter; ++ struct ib_port *sysfs; ++}; ++ ++struct rdma_netdev_alloc_params { ++ size_t sizeof_priv; ++ unsigned int txqs; ++ unsigned int rxqs; ++ void *param; ++ int (*initialize_rdma_netdev)(struct ib_device *, u32, struct net_device *, void *); ++}; ++ ++struct ib_counters_read_attr { ++ u64 *counters_buff; ++ u32 ncounters; ++ u32 flags; ++}; ++ ++struct rdma_user_mmap_entry { ++ struct kref ref; ++ struct ib_ucontext *ucontext; ++ long unsigned int start_pgoff; ++ size_t npages; ++ bool driver_removed; ++}; ++ ++struct ddebug_table { ++ struct list_head link; ++ struct list_head maps; ++ const char *mod_name; ++ unsigned int num_ddebugs; ++ struct _ddebug *ddebugs; ++}; ++ ++struct ddebug_query { ++ const char *filename; ++ const char *module; ++ const char *function; ++ const char *format; ++ const char *class_string; ++ unsigned int first_lineno; ++ unsigned int last_lineno; ++}; ++ ++struct ddebug_iter { ++ struct ddebug_table *table; ++ int idx; ++}; ++ ++struct flag_settings { ++ unsigned int flags; ++ unsigned int mask; ++}; ++ ++struct flagsbuf { ++ char buf[7]; ++}; ++ ++struct word_at_a_time { ++ const long unsigned int high_bits; ++ const long unsigned int low_bits; ++}; ++ ++struct sg_pool { ++ size_t size; ++ char *name; ++ struct kmem_cache *slab; ++ mempool_t *pool; ++}; ++ ++enum { ++ IRQ_POLL_F_SCHED = 0, ++ IRQ_POLL_F_DISABLE = 1, ++}; ++ ++struct node_groups { ++ unsigned int id; ++ union { ++ unsigned int ngroups; ++ unsigned int ncpus; ++ }; ++}; ++ ++struct phy_provider { ++ struct device *dev; ++ struct device_node *children; ++ struct module *owner; ++ struct list_head list; ++ struct phy * (*of_xlate)(struct device *, struct of_phandle_args *); ++}; ++ ++struct phy_lookup { ++ struct list_head node; ++ const char *dev_id; ++ const char *con_id; ++ struct phy *phy; ++}; ++ ++enum pcie_reset_state { ++ pcie_deassert_reset = 1, ++ pcie_warm_reset = 2, ++ pcie_hot_reset = 3, ++}; ++ ++typedef int (*arch_set_vga_state_t)(struct pci_dev *, bool, unsigned int, u32); ++ ++struct pci_reset_fn_method { ++ int (*reset_fn)(struct pci_dev *, bool); ++ char *name; ++}; ++ ++struct pci_pme_device { ++ struct list_head list; ++ struct pci_dev *dev; ++}; ++ ++struct pci_saved_state { ++ u32 config_space[16]; ++ struct pci_cap_saved_data cap[0]; ++}; ++ ++struct pci_devres { ++ unsigned int enabled: 1; ++ unsigned int pinned: 1; ++ unsigned int orig_intx: 1; ++ unsigned int restore_intx: 1; ++ unsigned int mwi: 1; ++ u32 region_mask; ++}; ++ ++struct pci_dev_resource { ++ struct list_head list; ++ struct resource *res; ++ struct pci_dev *dev; ++ resource_size_t start; ++ resource_size_t end; ++ resource_size_t add_size; ++ resource_size_t min_align; ++ long unsigned int flags; ++}; ++ ++enum release_type { ++ leaf_only = 0, ++ whole_subtree = 1, ++}; ++ ++enum enable_type { ++ undefined = -1, ++ user_disabled = 0, ++ auto_disabled = 1, ++ user_enabled = 2, ++ auto_enabled = 3, ++}; ++ ++struct portdrv_service_data { ++ struct pcie_port_service_driver *drv; ++ struct device *dev; ++ u32 service; ++}; ++ ++typedef int (*pcie_callback_t)(struct pcie_device *); ++ ++struct aer_stats { ++ u64 dev_cor_errs[16]; ++ u64 dev_fatal_errs[27]; ++ u64 dev_nonfatal_errs[27]; ++ u64 dev_total_cor_errs; ++ u64 dev_total_fatal_errs; ++ u64 dev_total_nonfatal_errs; ++ u64 rootport_total_cor_errs; ++ u64 rootport_total_fatal_errs; ++ u64 rootport_total_nonfatal_errs; ++}; ++ ++struct aer_err_source { ++ unsigned int status; ++ unsigned int id; ++}; ++ ++struct aer_rpc { ++ struct pci_dev *rpd; ++ struct { ++ union { ++ struct __kfifo kfifo; ++ struct aer_err_source *type; ++ const struct aer_err_source *const_type; ++ char (*rectype)[0]; ++ struct aer_err_source *ptr; ++ const struct aer_err_source *ptr_const; ++ }; ++ struct aer_err_source buf[128]; ++ } aer_fifo; ++}; ++ ++struct pci_fixup { ++ u16 vendor; ++ u16 device; ++ u32 class; ++ unsigned int class_shift; ++ void (*hook)(struct pci_dev *); ++}; ++ ++struct pci_doe_mb { ++ struct pci_dev *pdev; ++ u16 cap_offset; ++ struct xarray prots; ++ wait_queue_head_t wq; ++ struct workqueue_struct *work_queue; ++ long unsigned int flags; ++}; ++ ++struct pci_doe_protocol { ++ u16 vid; ++ u8 type; ++}; ++ ++struct pci_doe_task { ++ struct pci_doe_protocol prot; ++ const __le32 *request_pl; ++ size_t request_pl_sz; ++ __le32 *response_pl; ++ size_t response_pl_sz; ++ int rv; ++ void (*complete)(struct pci_doe_task *); ++ void *private; ++ struct work_struct work; ++ struct pci_doe_mb *doe_mb; ++}; ++ ++struct broken_edid { ++ u8 manufacturer[4]; ++ u32 model; ++ u32 fix; ++}; ++ ++struct __fb_timings { ++ u32 dclk; ++ u32 hfreq; ++ u32 vfreq; ++ u32 hactive; ++ u32 vactive; ++ u32 hblank; ++ u32 vblank; ++ u32 htotal; ++ u32 vtotal; ++}; ++ ++struct fb_cvt_data { ++ u32 xres; ++ u32 yres; ++ u32 refresh; ++ u32 f_refresh; ++ u32 pixclock; ++ u32 hperiod; ++ u32 hblank; ++ u32 hfreq; ++ u32 htotal; ++ u32 vtotal; ++ u32 vsync; ++ u32 hsync; ++ u32 h_front_porch; ++ u32 h_back_porch; ++ u32 v_front_porch; ++ u32 v_back_porch; ++ u32 h_margin; ++ u32 v_margin; ++ u32 interlace; ++ u32 aspect_ratio; ++ u32 active_pixels; ++ u32 flags; ++ u32 status; ++}; ++ ++struct devm_clk_state { ++ struct clk *clk; ++ void (*exit)(struct clk *); ++}; ++ ++struct clk_bulk_devres { ++ struct clk_bulk_data *clks; ++ int num_clks; ++}; ++ ++struct clk_div_table { ++ unsigned int val; ++ unsigned int div; ++}; ++ ++struct clk_divider { ++ struct clk_hw hw; ++ void *reg; ++ u8 shift; ++ u8 width; ++ u8 flags; ++ const struct clk_div_table *table; ++ spinlock_t *lock; ++}; ++ ++struct clk_fixed_factor { ++ struct clk_hw hw; ++ unsigned int mult; ++ unsigned int div; ++}; ++ ++struct clk_multiplier { ++ struct clk_hw hw; ++ void *reg; ++ u8 shift; ++ u8 width; ++ u8 flags; ++ spinlock_t *lock; ++}; ++ ++struct clk_composite { ++ struct clk_hw hw; ++ struct clk_ops ops; ++ struct clk_hw *mux_hw; ++ struct clk_hw *rate_hw; ++ struct clk_hw *gate_hw; ++ const struct clk_ops *mux_ops; ++ const struct clk_ops *rate_ops; ++ const struct clk_ops *gate_ops; ++}; ++ ++typedef __u32 __virtio32; ++ ++typedef __u64 __virtio64; ++ ++struct vring_desc { ++ __virtio64 addr; ++ __virtio32 len; ++ __virtio16 flags; ++ __virtio16 next; ++}; ++ ++struct vring_avail { ++ __virtio16 flags; ++ __virtio16 idx; ++ __virtio16 ring[0]; ++}; ++ ++struct vring_used_elem { ++ __virtio32 id; ++ __virtio32 len; ++}; ++ ++typedef struct vring_used_elem vring_used_elem_t; ++ ++struct vring_used { ++ __virtio16 flags; ++ __virtio16 idx; ++ vring_used_elem_t ring[0]; ++}; ++ ++typedef struct vring_desc vring_desc_t; ++ ++typedef struct vring_avail vring_avail_t; ++ ++typedef struct vring_used vring_used_t; ++ ++struct vring { ++ unsigned int num; ++ vring_desc_t *desc; ++ vring_avail_t *avail; ++ vring_used_t *used; ++}; ++ ++struct vring_packed_desc_event { ++ __le16 off_wrap; ++ __le16 flags; ++}; ++ ++struct vring_packed_desc { ++ __le64 addr; ++ __le32 len; ++ __le16 id; ++ __le16 flags; ++}; ++ ++struct vring_desc_state_split { ++ void *data; ++ struct vring_desc *indir_desc; ++}; ++ ++struct vring_desc_state_packed { ++ void *data; ++ struct vring_packed_desc *indir_desc; ++ u16 num; ++ u16 last; ++}; ++ ++struct vring_desc_extra { ++ dma_addr_t addr; ++ u32 len; ++ u16 flags; ++ u16 next; ++}; ++ ++struct vring_virtqueue_split { ++ struct vring vring; ++ u16 avail_flags_shadow; ++ u16 avail_idx_shadow; ++ struct vring_desc_state_split *desc_state; ++ struct vring_desc_extra *desc_extra; ++ dma_addr_t queue_dma_addr; ++ size_t queue_size_in_bytes; ++ u32 vring_align; ++ bool may_reduce_num; ++}; ++ ++struct vring_virtqueue_packed { ++ struct { ++ unsigned int num; ++ struct vring_packed_desc *desc; ++ struct vring_packed_desc_event *driver; ++ struct vring_packed_desc_event *device; ++ } vring; ++ bool avail_wrap_counter; ++ u16 avail_used_flags; ++ u16 next_avail_idx; ++ u16 event_flags_shadow; ++ struct vring_desc_state_packed *desc_state; ++ struct vring_desc_extra *desc_extra; ++ dma_addr_t ring_dma_addr; ++ dma_addr_t driver_event_dma_addr; ++ dma_addr_t device_event_dma_addr; ++ size_t ring_size_in_bytes; ++ size_t event_size_in_bytes; ++}; ++ ++struct vring_virtqueue { ++ struct virtqueue vq; ++ bool packed_ring; ++ bool use_dma_api; ++ bool weak_barriers; ++ bool broken; ++ bool indirect; ++ bool event; ++ unsigned int free_head; ++ unsigned int num_added; ++ u16 last_used_idx; ++ bool event_triggered; ++ union { ++ struct vring_virtqueue_split split; ++ struct vring_virtqueue_packed packed; ++ }; ++ bool (*notify)(struct virtqueue *); ++ bool we_own_ring; ++ struct device *dma_dev; ++}; ++ ++enum { ++ VP_MSIX_CONFIG_VECTOR = 0, ++ VP_MSIX_VQ_VECTOR = 1, ++}; ++ ++struct fixed_voltage_config { ++ const char *supply_name; ++ const char *input_supply; ++ int microvolts; ++ unsigned int startup_delay; ++ unsigned int off_on_delay; ++ unsigned int enabled_at_boot: 1; ++ struct regulator_init_data *init_data; ++}; ++ ++struct fixed_regulator_data { ++ struct fixed_voltage_config cfg; ++ struct regulator_init_data init_data; ++ struct platform_device pdev; ++}; ++ ++enum { ++ REGULATOR_ERROR_CLEARED = 0, ++ REGULATOR_FAILED_RETRY = 1, ++ REGULATOR_ERROR_ON = 2, ++}; ++ ++struct regulator_irq { ++ struct regulator_irq_data rdata; ++ struct regulator_irq_desc desc; ++ int irq; ++ int retry_cnt; ++ struct delayed_work isr_work; ++}; ++ ++struct sysrq_state { ++ struct input_handle handle; ++ struct work_struct reinject_work; ++ long unsigned int key_down[12]; ++ unsigned int alt; ++ unsigned int alt_use; ++ unsigned int shift; ++ unsigned int shift_use; ++ bool active; ++ bool need_reinject; ++ bool reinjecting; ++ bool reset_canceled; ++ bool reset_requested; ++ long unsigned int reset_keybit[12]; ++ int reset_seq_len; ++ int reset_seq_cnt; ++ int reset_seq_version; ++ struct timer_list keyreset_timer; ++}; ++ ++struct uni_pagedict { ++ u16 **uni_pgdir[32]; ++ long unsigned int refcount; ++ long unsigned int sum; ++ unsigned char *inverse_translations[4]; ++ u16 *inverse_trans_unicode; ++}; ++ ++struct serdev_device_driver { ++ struct device_driver driver; ++ int (*probe)(struct serdev_device *); ++ void (*remove)(struct serdev_device *); ++}; ++ ++struct tpmrm_priv { ++ struct file_priv priv; ++ struct tpm_space space; ++}; ++ ++struct iommu_group { ++ struct kobject kobj; ++ struct kobject *devices_kobj; ++ struct list_head devices; ++ struct xarray pasid_array; ++ struct mutex mutex; ++ void *iommu_data; ++ void (*iommu_data_release)(void *); ++ char *name; ++ int id; ++ struct iommu_domain *default_domain; ++ struct iommu_domain *blocking_domain; ++ struct iommu_domain *domain; ++ struct list_head entry; ++ unsigned int owner_cnt; ++ void *owner; ++}; ++ ++enum iommu_fault_type { ++ IOMMU_FAULT_DMA_UNRECOV = 1, ++ IOMMU_FAULT_PAGE_REQ = 2, ++}; ++ ++struct fsl_mc_obj_desc { ++ char type[16]; ++ int id; ++ u16 vendor; ++ u16 ver_major; ++ u16 ver_minor; ++ u8 irq_count; ++ u8 region_count; ++ u32 state; ++ char label[16]; ++ u16 flags; ++}; ++ ++struct fsl_mc_io; ++ ++struct fsl_mc_device_irq; ++ ++struct fsl_mc_resource; ++ ++struct fsl_mc_device { ++ struct device dev; ++ u64 dma_mask; ++ u16 flags; ++ u32 icid; ++ u16 mc_handle; ++ struct fsl_mc_io *mc_io; ++ struct fsl_mc_obj_desc obj_desc; ++ struct resource *regions; ++ struct fsl_mc_device_irq **irqs; ++ struct fsl_mc_resource *resource; ++ struct device_link *consumer_link; ++ const char *driver_override; ++}; ++ ++enum fsl_mc_pool_type { ++ FSL_MC_POOL_DPMCP = 0, ++ FSL_MC_POOL_DPBP = 1, ++ FSL_MC_POOL_DPCON = 2, ++ FSL_MC_POOL_IRQ = 3, ++ FSL_MC_NUM_POOL_TYPES = 4, ++}; ++ ++struct fsl_mc_resource_pool; ++ ++struct fsl_mc_resource { ++ enum fsl_mc_pool_type type; ++ s32 id; ++ void *data; ++ struct fsl_mc_resource_pool *parent_pool; ++ struct list_head node; ++}; ++ ++struct fsl_mc_device_irq { ++ unsigned int virq; ++ struct fsl_mc_device *mc_dev; ++ u8 dev_irq_index; ++ struct fsl_mc_resource resource; ++}; ++ ++struct fsl_mc_io { ++ struct device *dev; ++ u16 flags; ++ u32 portal_size; ++ phys_addr_t portal_phys_addr; ++ void *portal_virt_addr; ++ struct fsl_mc_device *dpmcp_dev; ++ union { ++ struct mutex mutex; ++ raw_spinlock_t spinlock; ++ }; ++}; ++ ++struct group_device { ++ struct list_head list; ++ struct device *dev; ++ char *name; ++}; ++ ++struct iommu_group_attribute { ++ struct attribute attr; ++ ssize_t (*show)(struct iommu_group *, char *); ++ ssize_t (*store)(struct iommu_group *, const char *, size_t); ++}; ++ ++struct group_for_pci_data { ++ struct pci_dev *pdev; ++ struct iommu_group *group; ++}; ++ ++struct __group_domain_type { ++ struct device *dev; ++ unsigned int type; ++}; ++ ++struct drm_client_offset { ++ int x; ++ int y; ++}; ++ ++struct displayid_header { ++ u8 rev; ++ u8 bytes; ++ u8 prod_id; ++ u8 ext_count; ++}; ++ ++struct drm_mode_fb_dirty_cmd { ++ __u32 fb_id; ++ __u32 flags; ++ __u32 color; ++ __u32 num_clips; ++ __u64 clips_ptr; ++}; ++ ++struct drm_mode_rmfb_work { ++ struct work_struct work; ++ struct list_head fbs; ++}; ++ ++struct drmres_node { ++ struct list_head entry; ++ drmres_release_t release; ++ const char *name; ++ size_t size; ++}; ++ ++struct drmres { ++ struct drmres_node node; ++ u8 data[0]; ++}; ++ ++struct drm_mode_obj_get_properties { ++ __u64 props_ptr; ++ __u64 prop_values_ptr; ++ __u32 count_props; ++ __u32 obj_id; ++ __u32 obj_type; ++}; ++ ++struct drm_prime_handle { ++ __u32 handle; ++ __u32 flags; ++ __s32 fd; ++}; ++ ++struct drm_prime_member { ++ struct dma_buf *dma_buf; ++ uint32_t handle; ++ struct rb_node dmabuf_rb; ++ struct rb_node handle_rb; ++}; ++ ++struct trace_event_raw_drm_vblank_event { ++ struct trace_entry ent; ++ int crtc; ++ unsigned int seq; ++ ktime_t time; ++ bool high_prec; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_drm_vblank_event_queued { ++ struct trace_entry ent; ++ struct drm_file *file; ++ int crtc; ++ unsigned int seq; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_drm_vblank_event_delivered { ++ struct trace_entry ent; ++ struct drm_file *file; ++ int crtc; ++ unsigned int seq; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_drm_vblank_event {}; ++ ++struct trace_event_data_offsets_drm_vblank_event_queued {}; ++ ++struct trace_event_data_offsets_drm_vblank_event_delivered {}; ++ ++typedef void (*btf_trace_drm_vblank_event)(void *, int, unsigned int, ktime_t, bool); ++ ++typedef void (*btf_trace_drm_vblank_event_queued)(void *, struct drm_file *, int, unsigned int); ++ ++typedef void (*btf_trace_drm_vblank_event_delivered)(void *, struct drm_file *, int, unsigned int); ++ ++struct drm_gem_shmem_object { ++ struct drm_gem_object base; ++ struct mutex pages_lock; ++ struct page **pages; ++ unsigned int pages_use_count; ++ int madv; ++ struct list_head madv_list; ++ struct sg_table *sgt; ++ struct mutex vmap_lock; ++ void *vaddr; ++ unsigned int vmap_use_count; ++ bool pages_mark_dirty_on_put: 1; ++ bool pages_mark_accessed_on_put: 1; ++ bool map_wc: 1; ++}; ++ ++struct drm_simple_display_pipe; ++ ++struct drm_simple_display_pipe_funcs { ++ enum drm_mode_status (*mode_valid)(struct drm_simple_display_pipe *, const struct drm_display_mode *); ++ void (*enable)(struct drm_simple_display_pipe *, struct drm_crtc_state *, struct drm_plane_state *); ++ void (*disable)(struct drm_simple_display_pipe *); ++ int (*check)(struct drm_simple_display_pipe *, struct drm_plane_state *, struct drm_crtc_state *); ++ void (*update)(struct drm_simple_display_pipe *, struct drm_plane_state *); ++ int (*prepare_fb)(struct drm_simple_display_pipe *, struct drm_plane_state *); ++ void (*cleanup_fb)(struct drm_simple_display_pipe *, struct drm_plane_state *); ++ int (*begin_fb_access)(struct drm_simple_display_pipe *, struct drm_plane_state *); ++ void (*end_fb_access)(struct drm_simple_display_pipe *, struct drm_plane_state *); ++ int (*enable_vblank)(struct drm_simple_display_pipe *); ++ void (*disable_vblank)(struct drm_simple_display_pipe *); ++ void (*reset_crtc)(struct drm_simple_display_pipe *); ++ struct drm_crtc_state * (*duplicate_crtc_state)(struct drm_simple_display_pipe *); ++ void (*destroy_crtc_state)(struct drm_simple_display_pipe *, struct drm_crtc_state *); ++ void (*reset_plane)(struct drm_simple_display_pipe *); ++ struct drm_plane_state * (*duplicate_plane_state)(struct drm_simple_display_pipe *); ++ void (*destroy_plane_state)(struct drm_simple_display_pipe *, struct drm_plane_state *); ++}; ++ ++struct drm_simple_display_pipe { ++ struct drm_crtc crtc; ++ struct drm_plane plane; ++ struct drm_encoder encoder; ++ struct drm_connector *connector; ++ const struct drm_simple_display_pipe_funcs *funcs; ++}; ++ ++typedef ssize_t (*drm_fb_helper_read_screen)(struct fb_info *, char *, size_t, loff_t); ++ ++typedef ssize_t (*drm_fb_helper_write_screen)(struct fb_info *, const char *, size_t, loff_t); ++ ++enum proc_cn_mcast_op { ++ PROC_CN_MCAST_LISTEN = 1, ++ PROC_CN_MCAST_IGNORE = 2, ++}; ++ ++struct fork_proc_event { ++ __kernel_pid_t parent_pid; ++ __kernel_pid_t parent_tgid; ++ __kernel_pid_t child_pid; ++ __kernel_pid_t child_tgid; ++}; ++ ++struct exec_proc_event { ++ __kernel_pid_t process_pid; ++ __kernel_pid_t process_tgid; ++}; ++ ++struct id_proc_event { ++ __kernel_pid_t process_pid; ++ __kernel_pid_t process_tgid; ++ union { ++ __u32 ruid; ++ __u32 rgid; ++ } r; ++ union { ++ __u32 euid; ++ __u32 egid; ++ } e; ++}; ++ ++struct sid_proc_event { ++ __kernel_pid_t process_pid; ++ __kernel_pid_t process_tgid; ++}; ++ ++struct ptrace_proc_event { ++ __kernel_pid_t process_pid; ++ __kernel_pid_t process_tgid; ++ __kernel_pid_t tracer_pid; ++ __kernel_pid_t tracer_tgid; ++}; ++ ++struct comm_proc_event { ++ __kernel_pid_t process_pid; ++ __kernel_pid_t process_tgid; ++ char comm[16]; ++}; ++ ++struct coredump_proc_event { ++ __kernel_pid_t process_pid; ++ __kernel_pid_t process_tgid; ++ __kernel_pid_t parent_pid; ++ __kernel_pid_t parent_tgid; ++}; ++ ++struct exit_proc_event { ++ __kernel_pid_t process_pid; ++ __kernel_pid_t process_tgid; ++ __u32 exit_code; ++ __u32 exit_signal; ++ __kernel_pid_t parent_pid; ++ __kernel_pid_t parent_tgid; ++}; ++ ++struct proc_event { ++ enum what what; ++ __u32 cpu; ++ __u64 timestamp_ns; ++ union { ++ struct { ++ __u32 err; ++ } ack; ++ struct fork_proc_event fork; ++ struct exec_proc_event exec; ++ struct id_proc_event id; ++ struct sid_proc_event sid; ++ struct ptrace_proc_event ptrace; ++ struct comm_proc_event comm; ++ struct coredump_proc_event coredump; ++ struct exit_proc_event exit; ++ } event_data; ++}; ++ ++struct local_event { ++ local_lock_t lock; ++ __u32 count; ++}; ++ ++struct fwnode_link { ++ struct fwnode_handle *supplier; ++ struct list_head s_hook; ++ struct fwnode_handle *consumer; ++ struct list_head c_hook; ++ u8 flags; ++}; ++ ++union device_attr_group_devres { ++ const struct attribute_group *group; ++ const struct attribute_group **groups; ++}; ++ ++struct class_dir { ++ struct kobject kobj; ++ const struct class *class; ++}; ++ ++struct root_device { ++ struct device dev; ++ struct module *owner; ++}; ++ ++struct probe { ++ struct probe *next; ++ dev_t dev; ++ long unsigned int range; ++ struct module *owner; ++ kobj_probe_t *get; ++ int (*lock)(dev_t, void *); ++ void *data; ++}; ++ ++struct kobj_map { ++ struct probe *probes[255]; ++ struct mutex *lock; ++}; ++ ++struct container_dev { ++ struct device dev; ++ int (*offline)(struct container_dev *); ++}; ++ ++struct software_node_ref_args { ++ const struct software_node *node; ++ unsigned int nargs; ++ u64 args[8]; ++}; ++ ++struct swnode { ++ struct kobject kobj; ++ struct fwnode_handle fwnode; ++ const struct software_node *node; ++ int id; ++ struct ida child_ids; ++ struct list_head entry; ++ struct list_head children; ++ struct swnode *parent; ++ unsigned int allocated: 1; ++ unsigned int managed: 1; ++}; ++ ++enum fw_upload_err { ++ FW_UPLOAD_ERR_NONE = 0, ++ FW_UPLOAD_ERR_HW_ERROR = 1, ++ FW_UPLOAD_ERR_TIMEOUT = 2, ++ FW_UPLOAD_ERR_CANCELED = 3, ++ FW_UPLOAD_ERR_BUSY = 4, ++ FW_UPLOAD_ERR_INVALID_SIZE = 5, ++ FW_UPLOAD_ERR_RW_ERROR = 6, ++ FW_UPLOAD_ERR_WEAROUT = 7, ++ FW_UPLOAD_ERR_MAX = 8, ++}; ++ ++struct fw_upload { ++ void *dd_handle; ++ void *priv; ++}; ++ ++struct fw_upload_ops { ++ enum fw_upload_err (*prepare)(struct fw_upload *, const u8 *, u32); ++ enum fw_upload_err (*write)(struct fw_upload *, const u8 *, u32, u32, u32 *); ++ enum fw_upload_err (*poll_complete)(struct fw_upload *); ++ void (*cancel)(struct fw_upload *); ++ void (*cleanup)(struct fw_upload *); ++}; ++ ++enum fw_upload_prog { ++ FW_UPLOAD_PROG_IDLE = 0, ++ FW_UPLOAD_PROG_RECEIVING = 1, ++ FW_UPLOAD_PROG_PREPARING = 2, ++ FW_UPLOAD_PROG_TRANSFERRING = 3, ++ FW_UPLOAD_PROG_PROGRAMMING = 4, ++ FW_UPLOAD_PROG_MAX = 5, ++}; ++ ++struct fw_upload_priv { ++ struct fw_upload *fw_upload; ++ struct module *module; ++ const char *name; ++ const struct fw_upload_ops *ops; ++ struct mutex lock; ++ struct work_struct work; ++ const u8 *data; ++ u32 remaining_size; ++ enum fw_upload_prog progress; ++ enum fw_upload_prog err_progress; ++ enum fw_upload_err err_code; ++}; ++ ++struct regcache_rbtree_node { ++ void *block; ++ long int *cache_present; ++ unsigned int base_reg; ++ unsigned int blklen; ++ struct rb_node node; ++}; ++ ++struct regcache_rbtree_ctx { ++ struct rb_root root; ++ struct regcache_rbtree_node *cached_rbnode; ++}; ++ ++struct regmap_mmio_context { ++ void *regs; ++ unsigned int val_bytes; ++ bool big_endian; ++ bool attached_clk; ++ struct clk *clk; ++ void (*reg_write)(struct regmap_mmio_context *, unsigned int, unsigned int); ++ unsigned int (*reg_read)(struct regmap_mmio_context *, unsigned int); ++}; ++ ++struct devcd_entry { ++ struct device devcd_dev; ++ void *data; ++ size_t datalen; ++ struct mutex mutex; ++ bool delete_work; ++ struct module *owner; ++ ssize_t (*read)(char *, loff_t, size_t, void *, size_t); ++ void (*free)(void *); ++ struct delayed_work del_wk; ++ struct device *failing_dev; ++}; ++ ++struct syscon_platform_data { ++ const char *label; ++}; ++ ++struct syscon { ++ struct device_node *np; ++ struct regmap *regmap; ++ struct reset_control *reset; ++ struct list_head list; ++}; ++ ++struct trace_event_raw_dma_fence { ++ struct trace_entry ent; ++ u32 __data_loc_driver; ++ u32 __data_loc_timeline; ++ unsigned int context; ++ unsigned int seqno; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_dma_fence { ++ u32 driver; ++ u32 timeline; ++}; ++ ++typedef void (*btf_trace_dma_fence_emit)(void *, struct dma_fence *); ++ ++typedef void (*btf_trace_dma_fence_init)(void *, struct dma_fence *); ++ ++typedef void (*btf_trace_dma_fence_destroy)(void *, struct dma_fence *); ++ ++typedef void (*btf_trace_dma_fence_enable_signal)(void *, struct dma_fence *); ++ ++typedef void (*btf_trace_dma_fence_signaled)(void *, struct dma_fence *); ++ ++typedef void (*btf_trace_dma_fence_wait_start)(void *, struct dma_fence *); ++ ++typedef void (*btf_trace_dma_fence_wait_end)(void *, struct dma_fence *); ++ ++struct default_wait_cb { ++ struct dma_fence_cb base; ++ struct task_struct *task; ++}; ++ ++struct dma_resv_list { ++ struct callback_head rcu; ++ u32 num_fences; ++ u32 max_fences; ++ struct dma_fence *table[0]; ++}; ++ ++struct udmabuf_create { ++ __u32 memfd; ++ __u32 flags; ++ __u64 offset; ++ __u64 size; ++}; ++ ++struct udmabuf_create_item { ++ __u32 memfd; ++ __u32 __pad; ++ __u64 offset; ++ __u64 size; ++}; ++ ++struct udmabuf_create_list { ++ __u32 flags; ++ __u32 count; ++ struct udmabuf_create_item list[0]; ++}; ++ ++struct udmabuf { ++ long unsigned int pagecount; ++ struct page **pages; ++ struct sg_table *sg; ++ struct miscdevice *device; ++}; ++ ++struct request_sense; ++ ++struct cdrom_generic_command { ++ unsigned char cmd[12]; ++ unsigned char *buffer; ++ unsigned int buflen; ++ int stat; ++ struct request_sense *sense; ++ unsigned char data_direction; ++ int quiet; ++ int timeout; ++ union { ++ void *reserved[1]; ++ void *unused; ++ }; ++}; ++ ++struct request_sense { ++ __u8 valid: 1; ++ __u8 error_code: 7; ++ __u8 segment_number; ++ __u8 reserved1: 2; ++ __u8 ili: 1; ++ __u8 reserved2: 1; ++ __u8 sense_key: 4; ++ __u8 information[4]; ++ __u8 add_sense_len; ++ __u8 command_info[4]; ++ __u8 asc; ++ __u8 ascq; ++ __u8 fruc; ++ __u8 sks[3]; ++ __u8 asb[46]; ++}; ++ ++struct scsi_ioctl_command { ++ unsigned int inlen; ++ unsigned int outlen; ++ unsigned char data[0]; ++}; ++ ++struct scsi_idlun { ++ __u32 dev_id; ++ __u32 host_unique_id; ++}; ++ ++struct compat_cdrom_generic_command { ++ unsigned char cmd[12]; ++ compat_caddr_t buffer; ++ compat_uint_t buflen; ++ compat_int_t stat; ++ compat_caddr_t sense; ++ unsigned char data_direction; ++ unsigned char pad[3]; ++ compat_int_t quiet; ++ compat_int_t timeout; ++ compat_caddr_t unused; ++}; ++ ++struct value_name_pair; ++ ++struct sa_name_list { ++ int opcode; ++ const struct value_name_pair *arr; ++ int arr_sz; ++}; ++ ++struct value_name_pair { ++ int value; ++ const char *name; ++}; ++ ++struct error_info { ++ short unsigned int code12; ++ short unsigned int size; ++}; ++ ++struct error_info2 { ++ unsigned char code1; ++ unsigned char code2_min; ++ unsigned char code2_max; ++ const char *str; ++ const char *fmt; ++}; ++ ++enum scsi_devinfo_key { ++ SCSI_DEVINFO_GLOBAL = 0, ++ SCSI_DEVINFO_SPI = 1, ++}; ++ ++struct scsi_dev_info_list { ++ struct list_head dev_info_list; ++ char vendor[8]; ++ char model[16]; ++ blist_flags_t flags; ++ unsigned int compatible; ++}; ++ ++struct scsi_dev_info_list_table { ++ struct list_head node; ++ struct list_head scsi_dev_info_list; ++ const char *name; ++ int key; ++}; ++ ++struct double_list { ++ struct list_head *top; ++ struct list_head *bottom; ++}; ++ ++enum { ++ SCSI_DH_OK = 0, ++ SCSI_DH_DEV_FAILED = 1, ++ SCSI_DH_DEV_TEMP_BUSY = 2, ++ SCSI_DH_DEV_UNSUPP = 3, ++ SCSI_DH_DEVICE_MAX = 4, ++ SCSI_DH_NOTCONN = 5, ++ SCSI_DH_CONN_FAILURE = 6, ++ SCSI_DH_TRANSPORT_MAX = 7, ++ SCSI_DH_IO = 8, ++ SCSI_DH_INVALID_IO = 9, ++ SCSI_DH_RETRY = 10, ++ SCSI_DH_IMM_RETRY = 11, ++ SCSI_DH_TIMED_OUT = 12, ++ SCSI_DH_RES_TEMP_UNAVAIL = 13, ++ SCSI_DH_DEV_OFFLINED = 14, ++ SCSI_DH_NOMEM = 15, ++ SCSI_DH_NOSYS = 16, ++ SCSI_DH_DRIVER_MAX = 17, ++}; ++ ++struct scsi_dh_blist { ++ const char *vendor; ++ const char *model; ++ const char *driver; ++}; ++ ++enum zbc_zone_type { ++ ZBC_ZONE_TYPE_CONV = 1, ++ ZBC_ZONE_TYPE_SEQWRITE_REQ = 2, ++ ZBC_ZONE_TYPE_SEQWRITE_PREF = 3, ++ ZBC_ZONE_TYPE_SEQ_OR_BEFORE_REQ = 4, ++ ZBC_ZONE_TYPE_GAP = 5, ++}; ++ ++enum zbc_zone_cond { ++ ZBC_ZONE_COND_NO_WP = 0, ++ ZBC_ZONE_COND_EMPTY = 1, ++ ZBC_ZONE_COND_IMP_OPEN = 2, ++ ZBC_ZONE_COND_EXP_OPEN = 3, ++ ZBC_ZONE_COND_CLOSED = 4, ++ ZBC_ZONE_COND_READONLY = 13, ++ ZBC_ZONE_COND_FULL = 14, ++ ZBC_ZONE_COND_OFFLINE = 15, ++}; ++ ++enum zbc_zone_alignment_method { ++ ZBC_CONSTANT_ZONE_LENGTH = 1, ++ ZBC_CONSTANT_ZONE_START_OFFSET = 8, ++}; ++ ++struct trace_event_raw_scsi_prepare_zone_append { ++ struct trace_entry ent; ++ unsigned int host_no; ++ unsigned int channel; ++ unsigned int id; ++ unsigned int lun; ++ sector_t lba; ++ unsigned int wp_offset; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_scsi_zone_wp_update { ++ struct trace_entry ent; ++ unsigned int host_no; ++ unsigned int channel; ++ unsigned int id; ++ unsigned int lun; ++ sector_t rq_sector; ++ unsigned int wp_offset; ++ unsigned int good_bytes; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_scsi_prepare_zone_append {}; ++ ++struct trace_event_data_offsets_scsi_zone_wp_update {}; ++ ++typedef void (*btf_trace_scsi_prepare_zone_append)(void *, struct scsi_cmnd *, sector_t, unsigned int); ++ ++typedef void (*btf_trace_scsi_zone_wp_update)(void *, struct scsi_cmnd *, sector_t, unsigned int, unsigned int); ++ ++typedef unsigned int (*ata_xlat_func_t)(struct ata_queued_cmd *); ++ ++struct ata_scsi_args { ++ struct ata_device *dev; ++ u16 *id; ++ struct scsi_cmnd *cmd; ++}; ++ ++enum { ++ AHCI_PCI_BAR_STA2X11 = 0, ++ AHCI_PCI_BAR_CAVIUM = 0, ++ AHCI_PCI_BAR_LOONGSON = 0, ++ AHCI_PCI_BAR_ENMOTUS = 2, ++ AHCI_PCI_BAR_CAVIUM_GEN5 = 4, ++ AHCI_PCI_BAR_STANDARD = 5, ++}; ++ ++enum board_ids { ++ board_ahci = 0, ++ board_ahci_ign_iferr = 1, ++ board_ahci_low_power = 2, ++ board_ahci_no_debounce_delay = 3, ++ board_ahci_nomsi = 4, ++ board_ahci_noncq = 5, ++ board_ahci_nosntf = 6, ++ board_ahci_yes_fbs = 7, ++ board_ahci_al = 8, ++ board_ahci_avn = 9, ++ board_ahci_mcp65 = 10, ++ board_ahci_mcp77 = 11, ++ board_ahci_mcp89 = 12, ++ board_ahci_mv = 13, ++ board_ahci_sb600 = 14, ++ board_ahci_sb700 = 15, ++ board_ahci_vt8251 = 16, ++ board_ahci_pcs7 = 17, ++ board_ahci_mcp_linux = 10, ++ board_ahci_mcp67 = 10, ++ board_ahci_mcp73 = 10, ++ board_ahci_mcp79 = 11, ++}; ++ ++struct phy_fixup { ++ struct list_head list; ++ char bus_id[64]; ++ u32 phy_uid; ++ u32 phy_uid_mask; ++ int (*run)(struct phy_device *); ++}; ++ ++struct mdio_device_id { ++ __u32 phy_id; ++ __u32 phy_id_mask; ++}; ++ ++struct rtl821x_priv { ++ u16 phycr1; ++ u16 phycr2; ++ bool has_phycr2; ++}; ++ ++struct cdrom_msf { ++ __u8 cdmsf_min0; ++ __u8 cdmsf_sec0; ++ __u8 cdmsf_frame0; ++ __u8 cdmsf_min1; ++ __u8 cdmsf_sec1; ++ __u8 cdmsf_frame1; ++}; ++ ++struct cdrom_volctrl { ++ __u8 channel0; ++ __u8 channel1; ++ __u8 channel2; ++ __u8 channel3; ++}; ++ ++struct cdrom_subchnl { ++ __u8 cdsc_format; ++ __u8 cdsc_audiostatus; ++ __u8 cdsc_adr: 4; ++ __u8 cdsc_ctrl: 4; ++ __u8 cdsc_trk; ++ __u8 cdsc_ind; ++ union cdrom_addr cdsc_absaddr; ++ union cdrom_addr cdsc_reladdr; ++}; ++ ++struct cdrom_read_audio { ++ union cdrom_addr addr; ++ __u8 addr_format; ++ int nframes; ++ __u8 *buf; ++}; ++ ++struct cdrom_blk { ++ unsigned int from; ++ short unsigned int len; ++}; ++ ++struct cdrom_timed_media_change_info { ++ __s64 last_media_change; ++ __u64 media_flags; ++}; ++ ++struct dvd_layer { ++ __u8 book_version: 4; ++ __u8 book_type: 4; ++ __u8 min_rate: 4; ++ __u8 disc_size: 4; ++ __u8 layer_type: 4; ++ __u8 track_path: 1; ++ __u8 nlayers: 2; ++ char: 1; ++ __u8 track_density: 4; ++ __u8 linear_density: 4; ++ __u8 bca: 1; ++ __u32 start_sector; ++ __u32 end_sector; ++ __u32 end_sector_l0; ++}; ++ ++struct dvd_physical { ++ __u8 type; ++ __u8 layer_num; ++ struct dvd_layer layer[4]; ++}; ++ ++struct dvd_copyright { ++ __u8 type; ++ __u8 layer_num; ++ __u8 cpst; ++ __u8 rmi; ++}; ++ ++struct dvd_disckey { ++ __u8 type; ++ unsigned int agid: 2; ++ __u8 value[2048]; ++}; ++ ++struct dvd_bca { ++ __u8 type; ++ int len; ++ __u8 value[188]; ++}; ++ ++struct dvd_manufact { ++ __u8 type; ++ __u8 layer_num; ++ int len; ++ __u8 value[2048]; ++}; ++ ++typedef union { ++ __u8 type; ++ struct dvd_physical physical; ++ struct dvd_copyright copyright; ++ struct dvd_disckey disckey; ++ struct dvd_bca bca; ++ struct dvd_manufact manufact; ++} dvd_struct; ++ ++typedef __u8 dvd_key[5]; ++ ++typedef __u8 dvd_challenge[10]; ++ ++struct dvd_lu_send_agid { ++ __u8 type; ++ unsigned int agid: 2; ++}; ++ ++struct dvd_host_send_challenge { ++ __u8 type; ++ unsigned int agid: 2; ++ dvd_challenge chal; ++}; ++ ++struct dvd_send_key { ++ __u8 type; ++ unsigned int agid: 2; ++ dvd_key key; ++}; ++ ++struct dvd_lu_send_challenge { ++ __u8 type; ++ unsigned int agid: 2; ++ dvd_challenge chal; ++}; ++ ++struct dvd_lu_send_title_key { ++ __u8 type; ++ unsigned int agid: 2; ++ dvd_key title_key; ++ int lba; ++ unsigned int cpm: 1; ++ unsigned int cp_sec: 1; ++ unsigned int cgms: 2; ++}; ++ ++struct dvd_lu_send_asf { ++ __u8 type; ++ unsigned int agid: 2; ++ unsigned int asf: 1; ++}; ++ ++struct dvd_host_send_rpcstate { ++ __u8 type; ++ __u8 pdrc; ++}; ++ ++struct dvd_lu_send_rpcstate { ++ __u8 type: 2; ++ __u8 vra: 3; ++ __u8 ucca: 3; ++ __u8 region_mask; ++ __u8 rpc_scheme; ++}; ++ ++typedef union { ++ __u8 type; ++ struct dvd_lu_send_agid lsa; ++ struct dvd_host_send_challenge hsc; ++ struct dvd_send_key lsk; ++ struct dvd_lu_send_challenge lsc; ++ struct dvd_send_key hsk; ++ struct dvd_lu_send_title_key lstk; ++ struct dvd_lu_send_asf lsasf; ++ struct dvd_host_send_rpcstate hrpcs; ++ struct dvd_lu_send_rpcstate lrpcs; ++} dvd_authinfo; ++ ++struct mrw_feature_desc { ++ __be16 feature_code; ++ __u8 reserved1: 2; ++ __u8 feature_version: 4; ++ __u8 persistent: 1; ++ __u8 curr: 1; ++ __u8 add_len; ++ __u8 reserved2: 7; ++ __u8 write: 1; ++ __u8 reserved3; ++ __u8 reserved4; ++ __u8 reserved5; ++}; ++ ++struct rwrt_feature_desc { ++ __be16 feature_code; ++ __u8 reserved1: 2; ++ __u8 feature_version: 4; ++ __u8 persistent: 1; ++ __u8 curr: 1; ++ __u8 add_len; ++ __u32 last_lba; ++ __u32 block_size; ++ __u16 blocking; ++ __u8 reserved2: 7; ++ __u8 page_present: 1; ++ __u8 reserved3; ++}; ++ ++typedef struct { ++ __be16 disc_information_length; ++ __u8 reserved1: 3; ++ __u8 erasable: 1; ++ __u8 border_status: 2; ++ __u8 disc_status: 2; ++ __u8 n_first_track; ++ __u8 n_sessions_lsb; ++ __u8 first_track_lsb; ++ __u8 last_track_lsb; ++ __u8 did_v: 1; ++ __u8 dbc_v: 1; ++ __u8 uru: 1; ++ __u8 reserved2: 2; ++ __u8 dbit: 1; ++ __u8 mrw_status: 2; ++ __u8 disc_type; ++ __u8 n_sessions_msb; ++ __u8 first_track_msb; ++ __u8 last_track_msb; ++ __u32 disc_id; ++ __u32 lead_in; ++ __u32 lead_out; ++ __u8 disc_bar_code[8]; ++ __u8 reserved3; ++ __u8 n_opc; ++} disc_information; ++ ++typedef struct { ++ __be16 track_information_length; ++ __u8 track_lsb; ++ __u8 session_lsb; ++ __u8 reserved1; ++ __u8 reserved2: 2; ++ __u8 damage: 1; ++ __u8 copy: 1; ++ __u8 track_mode: 4; ++ __u8 rt: 1; ++ __u8 blank: 1; ++ __u8 packet: 1; ++ __u8 fp: 1; ++ __u8 data_mode: 4; ++ __u8 reserved3: 6; ++ __u8 lra_v: 1; ++ __u8 nwa_v: 1; ++ __be32 track_start; ++ __be32 next_writable; ++ __be32 free_blocks; ++ __be32 fixed_packet_size; ++ __be32 track_size; ++ __be32 last_rec_address; ++} track_information; ++ ++struct mode_page_header { ++ __be16 mode_data_length; ++ __u8 medium_type; ++ __u8 reserved1; ++ __u8 reserved2; ++ __u8 reserved3; ++ __be16 desc_length; ++}; ++ ++typedef struct { ++ int data; ++ int audio; ++ int cdi; ++ int xa; ++ long int error; ++} tracktype; ++ ++struct cdrom_mechstat_header { ++ __u8 fault: 1; ++ __u8 changer_state: 2; ++ __u8 curslot: 5; ++ __u8 mech_state: 3; ++ __u8 door_open: 1; ++ __u8 reserved1: 4; ++ __u8 curlba[3]; ++ __u8 nslots; ++ __u16 slot_tablelen; ++}; ++ ++struct cdrom_slot { ++ __u8 disc_present: 1; ++ __u8 reserved1: 6; ++ __u8 change: 1; ++ __u8 reserved2[3]; ++}; ++ ++struct cdrom_changer_info { ++ struct cdrom_mechstat_header hdr; ++ struct cdrom_slot slots[256]; ++}; ++ ++struct modesel_head { ++ __u8 reserved1; ++ __u8 medium; ++ __u8 reserved2; ++ __u8 block_desc_length; ++ __u8 density; ++ __u8 number_of_blocks_hi; ++ __u8 number_of_blocks_med; ++ __u8 number_of_blocks_lo; ++ __u8 reserved3; ++ __u8 block_length_hi; ++ __u8 block_length_med; ++ __u8 block_length_lo; ++}; ++ ++typedef struct { ++ __u16 report_key_length; ++ __u8 reserved1; ++ __u8 reserved2; ++ __u8 type_code: 2; ++ __u8 vra: 3; ++ __u8 ucca: 3; ++ __u8 region_mask; ++ __u8 rpc_scheme; ++ __u8 reserved3; ++} rpc_state_t; ++ ++struct cdrom_sysctl_settings { ++ char info[1000]; ++ int autoclose; ++ int autoeject; ++ int debug; ++ int lock; ++ int check; ++}; ++ ++enum cdrom_print_option { ++ CTL_NAME = 0, ++ CTL_SPEED = 1, ++ CTL_SLOTS = 2, ++ CTL_CAPABILITY = 3, ++}; ++ ++struct compat_cdrom_read_audio { ++ union cdrom_addr addr; ++ u8 addr_format; ++ compat_int_t nframes; ++ compat_caddr_t buf; ++}; ++ ++struct input_dev_poller { ++ void (*poll)(struct input_dev *); ++ unsigned int poll_interval; ++ unsigned int poll_interval_max; ++ unsigned int poll_interval_min; ++ struct input_dev *input; ++ struct delayed_work work; ++}; ++ ++struct input_mask { ++ __u32 type; ++ __u32 codes_size; ++ __u64 codes_ptr; ++}; ++ ++struct evdev_client; ++ ++struct evdev { ++ int open; ++ struct input_handle handle; ++ struct evdev_client *grab; ++ struct list_head client_list; ++ spinlock_t client_lock; ++ struct mutex mutex; ++ struct device dev; ++ struct cdev cdev; ++ bool exist; ++}; ++ ++struct evdev_client { ++ unsigned int head; ++ unsigned int tail; ++ unsigned int packet_head; ++ spinlock_t buffer_lock; ++ wait_queue_head_t wait; ++ struct fasync_struct *fasync; ++ struct evdev *evdev; ++ struct list_head node; ++ enum input_clock_type clk_type; ++ bool revoked; ++ long unsigned int *evmasks[32]; ++ unsigned int bufsize; ++ struct input_event buffer[0]; ++}; ++ ++struct i2c_smbus_alert_setup { ++ int irq; ++}; ++ ++struct trace_event_raw_smbus_write { ++ struct trace_entry ent; ++ int adapter_nr; ++ __u16 addr; ++ __u16 flags; ++ __u8 command; ++ __u8 len; ++ __u32 protocol; ++ __u8 buf[34]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_smbus_read { ++ struct trace_entry ent; ++ int adapter_nr; ++ __u16 flags; ++ __u16 addr; ++ __u8 command; ++ __u32 protocol; ++ __u8 buf[34]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_smbus_reply { ++ struct trace_entry ent; ++ int adapter_nr; ++ __u16 addr; ++ __u16 flags; ++ __u8 command; ++ __u8 len; ++ __u32 protocol; ++ __u8 buf[34]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_smbus_result { ++ struct trace_entry ent; ++ int adapter_nr; ++ __u16 addr; ++ __u16 flags; ++ __u8 read_write; ++ __u8 command; ++ __s16 res; ++ __u32 protocol; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_smbus_write {}; ++ ++struct trace_event_data_offsets_smbus_read {}; ++ ++struct trace_event_data_offsets_smbus_reply {}; ++ ++struct trace_event_data_offsets_smbus_result {}; ++ ++typedef void (*btf_trace_smbus_write)(void *, const struct i2c_adapter *, u16, short unsigned int, char, u8, int, const union i2c_smbus_data *); ++ ++typedef void (*btf_trace_smbus_read)(void *, const struct i2c_adapter *, u16, short unsigned int, char, u8, int); ++ ++typedef void (*btf_trace_smbus_reply)(void *, const struct i2c_adapter *, u16, short unsigned int, char, u8, int, const union i2c_smbus_data *, int); ++ ++typedef void (*btf_trace_smbus_result)(void *, const struct i2c_adapter *, u16, short unsigned int, char, u8, int, int); ++ ++struct lirc_fh { ++ struct list_head list; ++ struct rc_dev *rc; ++ int carrier_low; ++ struct { ++ union { ++ struct __kfifo kfifo; ++ unsigned int *type; ++ const unsigned int *const_type; ++ char (*rectype)[0]; ++ unsigned int *ptr; ++ const unsigned int *ptr_const; ++ }; ++ unsigned int buf[0]; ++ } rawir; ++ struct { ++ union { ++ struct __kfifo kfifo; ++ struct lirc_scancode *type; ++ const struct lirc_scancode *const_type; ++ char (*rectype)[0]; ++ struct lirc_scancode *ptr; ++ const struct lirc_scancode *ptr_const; ++ }; ++ struct lirc_scancode buf[0]; ++ } scancodes; ++ wait_queue_head_t wait_poll; ++ u8 send_mode; ++ u8 rec_mode; ++}; ++ ++struct pps_event_time { ++ struct timespec64 ts_real; ++}; ++ ++enum ptp_clock_events { ++ PTP_CLOCK_ALARM = 0, ++ PTP_CLOCK_EXTTS = 1, ++ PTP_CLOCK_PPS = 2, ++ PTP_CLOCK_PPSUSR = 3, ++}; ++ ++struct ptp_clock_event { ++ int type; ++ int index; ++ union { ++ u64 timestamp; ++ struct pps_event_time pps_times; ++ }; ++}; ++ ++enum { ++ POWER_SUPPLY_STATUS_UNKNOWN = 0, ++ POWER_SUPPLY_STATUS_CHARGING = 1, ++ POWER_SUPPLY_STATUS_DISCHARGING = 2, ++ POWER_SUPPLY_STATUS_NOT_CHARGING = 3, ++ POWER_SUPPLY_STATUS_FULL = 4, ++}; ++ ++enum { ++ POWER_SUPPLY_CHARGE_TYPE_UNKNOWN = 0, ++ POWER_SUPPLY_CHARGE_TYPE_NONE = 1, ++ POWER_SUPPLY_CHARGE_TYPE_TRICKLE = 2, ++ POWER_SUPPLY_CHARGE_TYPE_FAST = 3, ++ POWER_SUPPLY_CHARGE_TYPE_STANDARD = 4, ++ POWER_SUPPLY_CHARGE_TYPE_ADAPTIVE = 5, ++ POWER_SUPPLY_CHARGE_TYPE_CUSTOM = 6, ++ POWER_SUPPLY_CHARGE_TYPE_LONGLIFE = 7, ++ POWER_SUPPLY_CHARGE_TYPE_BYPASS = 8, ++}; ++ ++enum { ++ POWER_SUPPLY_HEALTH_UNKNOWN = 0, ++ POWER_SUPPLY_HEALTH_GOOD = 1, ++ POWER_SUPPLY_HEALTH_OVERHEAT = 2, ++ POWER_SUPPLY_HEALTH_DEAD = 3, ++ POWER_SUPPLY_HEALTH_OVERVOLTAGE = 4, ++ POWER_SUPPLY_HEALTH_UNSPEC_FAILURE = 5, ++ POWER_SUPPLY_HEALTH_COLD = 6, ++ POWER_SUPPLY_HEALTH_WATCHDOG_TIMER_EXPIRE = 7, ++ POWER_SUPPLY_HEALTH_SAFETY_TIMER_EXPIRE = 8, ++ POWER_SUPPLY_HEALTH_OVERCURRENT = 9, ++ POWER_SUPPLY_HEALTH_CALIBRATION_REQUIRED = 10, ++ POWER_SUPPLY_HEALTH_WARM = 11, ++ POWER_SUPPLY_HEALTH_COOL = 12, ++ POWER_SUPPLY_HEALTH_HOT = 13, ++ POWER_SUPPLY_HEALTH_NO_BATTERY = 14, ++}; ++ ++enum { ++ POWER_SUPPLY_CAPACITY_LEVEL_UNKNOWN = 0, ++ POWER_SUPPLY_CAPACITY_LEVEL_CRITICAL = 1, ++ POWER_SUPPLY_CAPACITY_LEVEL_LOW = 2, ++ POWER_SUPPLY_CAPACITY_LEVEL_NORMAL = 3, ++ POWER_SUPPLY_CAPACITY_LEVEL_HIGH = 4, ++ POWER_SUPPLY_CAPACITY_LEVEL_FULL = 5, ++}; ++ ++enum power_supply_charge_behaviour { ++ POWER_SUPPLY_CHARGE_BEHAVIOUR_AUTO = 0, ++ POWER_SUPPLY_CHARGE_BEHAVIOUR_INHIBIT_CHARGE = 1, ++ POWER_SUPPLY_CHARGE_BEHAVIOUR_FORCE_DISCHARGE = 2, ++}; ++ ++struct power_supply_attr { ++ const char *prop_name; ++ char attr_name[31]; ++ struct device_attribute dev_attr; ++ const char * const *text_values; ++ int text_values_len; ++}; ++ ++typedef __u16 bitmap_counter_t; ++ ++enum bitmap_state { ++ BITMAP_STALE = 1, ++ BITMAP_WRITE_ERROR = 2, ++ BITMAP_HOSTENDIAN = 15, ++}; ++ ++struct bitmap_super_s { ++ __le32 magic; ++ __le32 version; ++ __u8 uuid[16]; ++ __le64 events; ++ __le64 events_cleared; ++ __le64 sync_size; ++ __le32 state; ++ __le32 chunksize; ++ __le32 daemon_sleep; ++ __le32 write_behind; ++ __le32 sectors_reserved; ++ __le32 nodes; ++ __u8 cluster_name[64]; ++ __u8 pad[120]; ++}; ++ ++typedef struct bitmap_super_s bitmap_super_t; ++ ++enum bitmap_page_attr { ++ BITMAP_PAGE_DIRTY = 0, ++ BITMAP_PAGE_PENDING = 1, ++ BITMAP_PAGE_NEEDWRITE = 2, ++}; ++ ++struct dm_ioctl { ++ __u32 version[3]; ++ __u32 data_size; ++ __u32 data_start; ++ __u32 target_count; ++ __s32 open_count; ++ __u32 flags; ++ __u32 event_nr; ++ __u32 padding; ++ __u64 dev; ++ char name[128]; ++ char uuid[129]; ++ char data[7]; ++}; ++ ++struct dm_target_spec { ++ __u64 sector_start; ++ __u64 length; ++ __s32 status; ++ __u32 next; ++ char target_type[16]; ++}; ++ ++struct dm_device { ++ struct dm_ioctl dmi; ++ struct dm_target_spec *table[256]; ++ char *target_args_array[256]; ++ struct list_head list; ++}; ++ ++struct orig_bio_details { ++ enum req_op op; ++ unsigned int nr_sectors; ++}; ++ ++struct dm_dev_internal { ++ struct list_head list; ++ refcount_t count; ++ struct dm_dev *dm_dev; ++}; ++ ++struct dm_crypto_profile { ++ struct blk_crypto_profile profile; ++ struct mapped_device *md; ++}; ++ ++enum suspend_mode { ++ PRESUSPEND = 0, ++ PRESUSPEND_UNDO = 1, ++ POSTSUSPEND = 2, ++}; ++ ++struct linear_c { ++ struct dm_dev *dev; ++ sector_t start; ++}; ++ ++struct dm_target_deps { ++ __u32 count; ++ __u32 padding; ++ __u64 dev[0]; ++}; ++ ++struct dm_name_list { ++ __u64 dev; ++ __u32 next; ++ char name[0]; ++}; ++ ++struct dm_target_versions { ++ __u32 next; ++ __u32 version[3]; ++ char name[0]; ++}; ++ ++struct dm_target_msg { ++ __u64 sector; ++ char message[0]; ++}; ++ ++enum { ++ DM_VERSION_CMD = 0, ++ DM_REMOVE_ALL_CMD = 1, ++ DM_LIST_DEVICES_CMD = 2, ++ DM_DEV_CREATE_CMD = 3, ++ DM_DEV_REMOVE_CMD = 4, ++ DM_DEV_RENAME_CMD = 5, ++ DM_DEV_SUSPEND_CMD = 6, ++ DM_DEV_STATUS_CMD = 7, ++ DM_DEV_WAIT_CMD = 8, ++ DM_TABLE_LOAD_CMD = 9, ++ DM_TABLE_CLEAR_CMD = 10, ++ DM_TABLE_DEPS_CMD = 11, ++ DM_TABLE_STATUS_CMD = 12, ++ DM_LIST_VERSIONS_CMD = 13, ++ DM_TARGET_MSG_CMD = 14, ++ DM_DEV_SET_GEOMETRY_CMD = 15, ++ DM_DEV_ARM_POLL_CMD = 16, ++ DM_GET_TARGET_VERSION_CMD = 17, ++}; ++ ++struct dm_file { ++ volatile unsigned int global_event_nr; ++}; ++ ++struct hash_cell { ++ struct rb_node name_node; ++ struct rb_node uuid_node; ++ bool name_set; ++ bool uuid_set; ++ char *name; ++ char *uuid; ++ struct mapped_device *md; ++ struct dm_table *new_map; ++}; ++ ++struct vers_iter { ++ size_t param_size; ++ struct dm_target_versions *vers; ++ struct dm_target_versions *old_vers; ++ char *end; ++ uint32_t flags; ++}; ++ ++typedef int (*ioctl_fn___2)(struct file *, struct dm_ioctl *, size_t); ++ ++struct dm_sysfs_attr { ++ struct attribute attr; ++ ssize_t (*show)(struct mapped_device *, char *); ++ ssize_t (*store)(struct mapped_device *, const char *, size_t); ++}; ++ ++struct lru_entry { ++ struct list_head list; ++ atomic_t referenced; ++}; ++ ++struct lru; ++ ++struct lru_iter { ++ struct lru *lru; ++ struct list_head list; ++ struct lru_entry *stop; ++ struct lru_entry *e; ++}; ++ ++struct lru { ++ struct list_head *cursor; ++ long unsigned int count; ++ struct list_head iterators; ++}; ++ ++typedef bool (*iter_predicate)(struct lru_entry *, void *); ++ ++enum evict_result { ++ ER_EVICT = 0, ++ ER_DONT_EVICT = 1, ++ ER_STOP = 2, ++}; ++ ++typedef enum evict_result (*le_predicate)(struct lru_entry *, void *); ++ ++enum data_mode { ++ DATA_MODE_SLAB = 0, ++ DATA_MODE_GET_FREE_PAGES = 1, ++ DATA_MODE_VMALLOC = 2, ++ DATA_MODE_LIMIT = 3, ++}; ++ ++struct dm_bufio_client; ++ ++struct dm_buffer { ++ struct rb_node node; ++ sector_t block; ++ void *data; ++ unsigned char data_mode; ++ atomic_t hold_count; ++ long unsigned int last_accessed; ++ long unsigned int state; ++ struct lru_entry lru; ++ unsigned char list_mode; ++ blk_status_t read_error; ++ blk_status_t write_error; ++ unsigned int dirty_start; ++ unsigned int dirty_end; ++ unsigned int write_start; ++ unsigned int write_end; ++ struct list_head write_list; ++ struct dm_bufio_client *c; ++ void (*end_io)(struct dm_buffer *, blk_status_t); ++}; ++ ++struct buffer_tree { ++ struct rw_semaphore lock; ++ struct rb_root root; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct dm_buffer_cache { ++ struct lru lru[2]; ++ unsigned int num_locks; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct buffer_tree trees[0]; ++}; ++ ++struct dm_bufio_client { ++ struct block_device *bdev; ++ unsigned int block_size; ++ s8 sectors_per_block_bits; ++ bool no_sleep; ++ struct mutex lock; ++ spinlock_t spinlock; ++ int async_write_error; ++ void (*alloc_callback)(struct dm_buffer *); ++ void (*write_callback)(struct dm_buffer *); ++ struct kmem_cache *slab_buffer; ++ struct kmem_cache *slab_cache; ++ struct dm_io_client *dm_io; ++ struct list_head reserved_buffers; ++ unsigned int need_reserved_buffers; ++ unsigned int minimum_buffers; ++ sector_t start; ++ struct shrinker shrinker; ++ struct work_struct shrink_work; ++ atomic_long_t need_shrink; ++ wait_queue_head_t free_buffer_wait; ++ struct list_head client_list; ++ long unsigned int oldest_buffer; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct dm_buffer_cache cache; ++}; ++ ++struct lock_history { ++ struct dm_buffer_cache *cache; ++ bool write; ++ unsigned int previous; ++ unsigned int no_previous; ++}; ++ ++typedef enum evict_result (*b_predicate)(struct dm_buffer *, void *); ++ ++struct evict_wrapper { ++ struct lock_history *lh; ++ b_predicate pred; ++ void *context; ++}; ++ ++enum it_action { ++ IT_NEXT = 0, ++ IT_COMPLETE = 1, ++}; ++ ++typedef enum it_action (*iter_fn)(struct dm_buffer *, void *); ++ ++typedef void (*b_release)(struct dm_buffer *); ++ ++enum new_flag { ++ NF_FRESH = 0, ++ NF_READ = 1, ++ NF_GET = 2, ++ NF_PREFETCH = 3, ++}; ++ ++struct write_context { ++ int no_wait; ++ struct list_head *write_list; ++}; ++ ++struct evict_params { ++ gfp_t gfp; ++ long unsigned int age_hz; ++ long unsigned int last_accessed; ++}; ++ ++struct transient_c { ++ sector_t next_free; ++}; ++ ++struct log_header_disk { ++ __le32 magic; ++ __le32 version; ++ __le64 nr_regions; ++}; ++ ++struct log_header_core { ++ uint32_t magic; ++ uint32_t version; ++ uint64_t nr_regions; ++}; ++ ++enum sync { ++ DEFAULTSYNC = 0, ++ NOSYNC = 1, ++ FORCESYNC = 2, ++}; ++ ++struct log_c { ++ struct dm_target *ti; ++ int touched_dirtied; ++ int touched_cleaned; ++ int flush_failed; ++ uint32_t region_size; ++ unsigned int region_count; ++ region_t sync_count; ++ unsigned int bitset_uint32_count; ++ uint32_t *clean_bits; ++ uint32_t *sync_bits; ++ uint32_t *recovering_bits; ++ int sync_search; ++ enum sync sync; ++ struct dm_io_request io_req; ++ int log_dev_failed; ++ int log_dev_flush_failed; ++ struct dm_dev *log_dev; ++ struct log_header_core header; ++ struct dm_io_region header_location; ++ struct log_header_disk *disk_header; ++}; ++ ++struct rproc_dump_segment { ++ struct list_head node; ++ dma_addr_t da; ++ size_t size; ++ void *priv; ++ void (*dump)(struct rproc *, struct rproc_dump_segment *, void *, size_t, size_t); ++ loff_t offset; ++}; ++ ++struct rproc_coredump_state { ++ struct rproc *rproc; ++ void *header; ++ struct completion dump_done; ++}; ++ ++union extcon_property_value { ++ int intval; ++}; ++ ++struct extcon_cable { ++ struct extcon_dev *edev; ++ int cable_index; ++ struct attribute_group attr_g; ++ struct device_attribute attr_name; ++ struct device_attribute attr_state; ++ struct attribute *attrs[3]; ++ union extcon_property_value usb_propval[3]; ++ union extcon_property_value chg_propval[1]; ++ union extcon_property_value jack_propval[1]; ++ union extcon_property_value disp_propval[2]; ++ long unsigned int usb_bits[1]; ++ long unsigned int chg_bits[1]; ++ long unsigned int jack_bits[1]; ++ long unsigned int disp_bits[1]; ++}; ++ ++struct __extcon_info { ++ unsigned int type; ++ unsigned int id; ++ const char *name; ++}; ++ ++struct cper_sec_proc_arm { ++ u32 validation_bits; ++ u16 err_info_num; ++ u16 context_info_num; ++ u32 section_length; ++ u8 affinity_level; ++ u8 reserved[3]; ++ u64 mpidr; ++ u64 midr; ++ u32 running_state; ++ u32 psci_state; ++}; ++ ++enum hw_event_mc_err_type { ++ HW_EVENT_ERR_CORRECTED = 0, ++ HW_EVENT_ERR_UNCORRECTED = 1, ++ HW_EVENT_ERR_DEFERRED = 2, ++ HW_EVENT_ERR_FATAL = 3, ++ HW_EVENT_ERR_INFO = 4, ++}; ++ ++struct trace_event_raw_mc_event { ++ struct trace_entry ent; ++ unsigned int error_type; ++ u32 __data_loc_msg; ++ u32 __data_loc_label; ++ u16 error_count; ++ u8 mc_index; ++ s8 top_layer; ++ s8 middle_layer; ++ s8 lower_layer; ++ long int address; ++ u8 grain_bits; ++ long int syndrome; ++ u32 __data_loc_driver_detail; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_arm_event { ++ struct trace_entry ent; ++ u64 mpidr; ++ u64 midr; ++ u32 running_state; ++ u32 psci_state; ++ u8 affinity; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_non_standard_event { ++ struct trace_entry ent; ++ char sec_type[16]; ++ char fru_id[16]; ++ u32 __data_loc_fru_text; ++ u8 sev; ++ u32 len; ++ u32 __data_loc_buf; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_aer_event { ++ struct trace_entry ent; ++ u32 __data_loc_dev_name; ++ u32 status; ++ u8 severity; ++ u8 tlp_header_valid; ++ u32 tlp_header[4]; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_mc_event { ++ u32 msg; ++ u32 label; ++ u32 driver_detail; ++}; ++ ++struct trace_event_data_offsets_arm_event {}; ++ ++struct trace_event_data_offsets_non_standard_event { ++ u32 fru_text; ++ u32 buf; ++}; ++ ++struct trace_event_data_offsets_aer_event { ++ u32 dev_name; ++}; ++ ++typedef void (*btf_trace_mc_event)(void *, const unsigned int, const char *, const char *, const int, const u8, const s8, const s8, const s8, long unsigned int, const u8, long unsigned int, const char *); ++ ++typedef void (*btf_trace_arm_event)(void *, const struct cper_sec_proc_arm *); ++ ++typedef void (*btf_trace_non_standard_event)(void *, const guid_t *, const guid_t *, const char *, const u8, const u8 *, const u32); ++ ++typedef void (*btf_trace_aer_event)(void *, const char *, const u32, const u8, const u8, struct aer_header_log_regs *); ++ ++struct tb_ctl { ++ struct tb_nhi *nhi; ++ struct tb_ring *tx; ++ struct tb_ring *rx; ++ struct dma_pool *frame_pool; ++ struct ctl_pkg *rx_packets[10]; ++ struct mutex request_queue_lock; ++ struct list_head request_queue; ++ bool running; ++ int timeout_msec; ++ event_cb callback; ++ void *callback_data; ++}; ++ ++struct cfg_ack_pkg { ++ struct tb_cfg_header header; ++}; ++ ++struct cfg_reset_pkg { ++ struct tb_cfg_header header; ++}; ++ ++enum tb_port_state { ++ TB_PORT_DISABLED = 0, ++ TB_PORT_CONNECTING = 1, ++ TB_PORT_UP = 2, ++ TB_PORT_TX_CL0S = 3, ++ TB_PORT_RX_CL0S = 4, ++ TB_PORT_CL1 = 5, ++ TB_PORT_CL2 = 6, ++ TB_PORT_UNPLUGGED = 7, ++}; ++ ++struct tb_cap_phy { ++ struct tb_cap_basic cap_header; ++ u32 unknown1: 16; ++ u32 unknown2: 14; ++ bool disable: 1; ++ u32 unknown3: 11; ++ enum tb_port_state state: 4; ++ u32 unknown4: 2; ++}; ++ ++enum tb_nvm_write_ops { ++ WRITE_AND_AUTHENTICATE = 1, ++ WRITE_ONLY = 2, ++ AUTHENTICATE_ONLY = 3, ++}; ++ ++struct nvm_auth_status { ++ struct list_head list; ++ uuid_t uuid; ++ u32 status; ++}; ++ ++struct tb_sw_lookup { ++ struct tb *tb; ++ u8 link; ++ u8 depth; ++ const uuid_t *uuid; ++ u64 route; ++}; ++ ++struct tb_eeprom_ctl { ++ bool fl_sk: 1; ++ bool fl_cs: 1; ++ bool fl_di: 1; ++ bool fl_do: 1; ++ bool bit_banging_enable: 1; ++ bool not_present: 1; ++ bool unknown1: 1; ++ bool present: 1; ++ u32 unknown2: 24; ++}; ++ ++struct tb_cap_plug_events { ++ struct tb_cap_extended_short cap_header; ++ u32 __unknown1: 2; ++ u32 plug_events: 5; ++ u32 __unknown2: 25; ++ u32 vsc_cs_2; ++ u32 vsc_cs_3; ++ struct tb_eeprom_ctl eeprom_ctl; ++ u32 __unknown5[7]; ++ u32 drom_offset; ++}; ++ ++enum tb_eeprom_transfer { ++ TB_EEPROM_IN = 0, ++ TB_EEPROM_OUT = 1, ++}; ++ ++struct tb_drom_header { ++ u8 uid_crc8; ++ u64 uid; ++ u32 data_crc32; ++ u8 device_rom_revision; ++ u16 data_len: 12; ++ u8 reserved: 4; ++ u16 vendor_id; ++ u16 model_id; ++ u8 model_rev; ++ u8 eeprom_rev; ++} __attribute__((packed)); ++ ++enum tb_drom_entry_type { ++ TB_DROM_ENTRY_GENERIC = 0, ++ TB_DROM_ENTRY_PORT = 1, ++}; ++ ++struct tb_drom_entry_header { ++ u8 len; ++ u8 index: 6; ++ bool port_disabled: 1; ++ enum tb_drom_entry_type type: 1; ++} __attribute__((packed)); ++ ++struct tb_drom_entry_generic { ++ struct tb_drom_entry_header header; ++ u8 data[0]; ++}; ++ ++struct tb_drom_entry_port { ++ struct tb_drom_entry_header header; ++ u8 dual_link_port_rid: 4; ++ u8 link_nr: 1; ++ u8 unknown1: 2; ++ bool has_dual_link_port: 1; ++ u8 dual_link_port_nr: 6; ++ u8 unknown2: 2; ++ u8 micro2: 4; ++ u8 micro1: 4; ++ u8 micro3; ++ u8 peer_port_rid: 4; ++ u8 unknown3: 3; ++ bool has_peer_port: 1; ++ u8 peer_port_nr: 6; ++ u8 unknown4: 2; ++}; ++ ++struct tb_drom_entry_desc { ++ struct tb_drom_entry_header header; ++ u16 bcdUSBSpec; ++ u16 idVendor; ++ u16 idProduct; ++ u16 bcdProductFWRevision; ++ u32 TID; ++ u8 productHWRevision; ++}; ++ ++enum icm_pkg_code { ++ ICM_GET_TOPOLOGY = 1, ++ ICM_DRIVER_READY = 3, ++ ICM_APPROVE_DEVICE = 4, ++ ICM_CHALLENGE_DEVICE = 5, ++ ICM_ADD_DEVICE_KEY = 6, ++ ICM_GET_ROUTE = 10, ++ ICM_APPROVE_XDOMAIN = 16, ++ ICM_DISCONNECT_XDOMAIN = 17, ++ ICM_PREBOOT_ACL = 24, ++ ICM_USB4_SWITCH_OP = 32, ++}; ++ ++enum icm_event_code { ++ ICM_EVENT_DEVICE_CONNECTED = 3, ++ ICM_EVENT_DEVICE_DISCONNECTED = 4, ++ ICM_EVENT_XDOMAIN_CONNECTED = 6, ++ ICM_EVENT_XDOMAIN_DISCONNECTED = 7, ++ ICM_EVENT_RTD3_VETO = 10, ++}; ++ ++struct icm_pkg_header { ++ u8 code; ++ u8 flags; ++ u8 packet_id; ++ u8 total_packets; ++}; ++ ++struct icm_pkg_driver_ready { ++ struct icm_pkg_header hdr; ++}; ++ ++struct icm_fr_pkg_driver_ready_response { ++ struct icm_pkg_header hdr; ++ u8 romver; ++ u8 ramver; ++ u16 security_level; ++}; ++ ++struct icm_fr_pkg_get_topology { ++ struct icm_pkg_header hdr; ++}; ++ ++struct icm_fr_pkg_get_topology_response { ++ struct icm_pkg_header hdr; ++ u32 route_lo; ++ u32 route_hi; ++ u8 first_data; ++ u8 second_data; ++ u8 drom_i2c_address_index; ++ u8 switch_index; ++ u32 reserved[2]; ++ u32 ports[16]; ++ u32 port_hop_info[16]; ++}; ++ ++struct icm_fr_event_device_connected { ++ struct icm_pkg_header hdr; ++ uuid_t ep_uuid; ++ u8 connection_key; ++ u8 connection_id; ++ u16 link_info; ++ u32 ep_name[55]; ++}; ++ ++struct icm_fr_pkg_approve_device { ++ struct icm_pkg_header hdr; ++ uuid_t ep_uuid; ++ u8 connection_key; ++ u8 connection_id; ++ u16 reserved; ++}; ++ ++struct icm_fr_event_device_disconnected { ++ struct icm_pkg_header hdr; ++ u16 reserved; ++ u16 link_info; ++}; ++ ++struct icm_fr_event_xdomain_connected { ++ struct icm_pkg_header hdr; ++ u16 reserved; ++ u16 link_info; ++ uuid_t remote_uuid; ++ uuid_t local_uuid; ++ u32 local_route_hi; ++ u32 local_route_lo; ++ u32 remote_route_hi; ++ u32 remote_route_lo; ++}; ++ ++struct icm_fr_event_xdomain_disconnected { ++ struct icm_pkg_header hdr; ++ u16 reserved; ++ u16 link_info; ++ uuid_t remote_uuid; ++}; ++ ++struct icm_fr_pkg_add_device_key { ++ struct icm_pkg_header hdr; ++ uuid_t ep_uuid; ++ u8 connection_key; ++ u8 connection_id; ++ u16 reserved; ++ u32 key[8]; ++}; ++ ++struct icm_fr_pkg_add_device_key_response { ++ struct icm_pkg_header hdr; ++ uuid_t ep_uuid; ++ u8 connection_key; ++ u8 connection_id; ++ u16 reserved; ++}; ++ ++struct icm_fr_pkg_challenge_device { ++ struct icm_pkg_header hdr; ++ uuid_t ep_uuid; ++ u8 connection_key; ++ u8 connection_id; ++ u16 reserved; ++ u32 challenge[8]; ++}; ++ ++struct icm_fr_pkg_challenge_device_response { ++ struct icm_pkg_header hdr; ++ uuid_t ep_uuid; ++ u8 connection_key; ++ u8 connection_id; ++ u16 reserved; ++ u32 challenge[8]; ++ u32 response[8]; ++}; ++ ++struct icm_fr_pkg_approve_xdomain { ++ struct icm_pkg_header hdr; ++ u16 reserved; ++ u16 link_info; ++ uuid_t remote_uuid; ++ u16 transmit_path; ++ u16 transmit_ring; ++ u16 receive_path; ++ u16 receive_ring; ++}; ++ ++struct icm_fr_pkg_approve_xdomain_response { ++ struct icm_pkg_header hdr; ++ u16 reserved; ++ u16 link_info; ++ uuid_t remote_uuid; ++ u16 transmit_path; ++ u16 transmit_ring; ++ u16 receive_path; ++ u16 receive_ring; ++}; ++ ++struct icm_ar_pkg_driver_ready_response { ++ struct icm_pkg_header hdr; ++ u8 romver; ++ u8 ramver; ++ u16 info; ++}; ++ ++struct icm_ar_pkg_get_route { ++ struct icm_pkg_header hdr; ++ u16 reserved; ++ u16 link_info; ++}; ++ ++struct icm_ar_pkg_get_route_response { ++ struct icm_pkg_header hdr; ++ u16 reserved; ++ u16 link_info; ++ u32 route_hi; ++ u32 route_lo; ++}; ++ ++struct icm_ar_boot_acl_entry { ++ u32 uuid_lo; ++ u32 uuid_hi; ++}; ++ ++struct icm_ar_pkg_preboot_acl { ++ struct icm_pkg_header hdr; ++ struct icm_ar_boot_acl_entry acl[16]; ++}; ++ ++struct icm_ar_pkg_preboot_acl_response { ++ struct icm_pkg_header hdr; ++ struct icm_ar_boot_acl_entry acl[16]; ++}; ++ ++struct icm_tr_pkg_driver_ready_response { ++ struct icm_pkg_header hdr; ++ u16 reserved1; ++ u16 info; ++ u32 nvm_version; ++ u16 device_id; ++ u16 reserved2; ++}; ++ ++struct icm_tr_event_device_connected { ++ struct icm_pkg_header hdr; ++ uuid_t ep_uuid; ++ u32 route_hi; ++ u32 route_lo; ++ u8 connection_id; ++ u8 reserved; ++ u16 link_info; ++ u32 ep_name[55]; ++}; ++ ++struct icm_tr_event_device_disconnected { ++ struct icm_pkg_header hdr; ++ u32 route_hi; ++ u32 route_lo; ++}; ++ ++struct icm_tr_event_xdomain_connected { ++ struct icm_pkg_header hdr; ++ u16 reserved; ++ u16 link_info; ++ uuid_t remote_uuid; ++ uuid_t local_uuid; ++ u32 local_route_hi; ++ u32 local_route_lo; ++ u32 remote_route_hi; ++ u32 remote_route_lo; ++}; ++ ++struct icm_tr_event_xdomain_disconnected { ++ struct icm_pkg_header hdr; ++ u32 route_hi; ++ u32 route_lo; ++ uuid_t remote_uuid; ++}; ++ ++struct icm_tr_pkg_approve_device { ++ struct icm_pkg_header hdr; ++ uuid_t ep_uuid; ++ u32 route_hi; ++ u32 route_lo; ++ u8 connection_id; ++ u8 reserved1[3]; ++}; ++ ++struct icm_tr_pkg_add_device_key { ++ struct icm_pkg_header hdr; ++ uuid_t ep_uuid; ++ u32 route_hi; ++ u32 route_lo; ++ u8 connection_id; ++ u8 reserved[3]; ++ u32 key[8]; ++}; ++ ++struct icm_tr_pkg_challenge_device { ++ struct icm_pkg_header hdr; ++ uuid_t ep_uuid; ++ u32 route_hi; ++ u32 route_lo; ++ u8 connection_id; ++ u8 reserved[3]; ++ u32 challenge[8]; ++}; ++ ++struct icm_tr_pkg_approve_xdomain { ++ struct icm_pkg_header hdr; ++ u32 route_hi; ++ u32 route_lo; ++ uuid_t remote_uuid; ++ u16 transmit_path; ++ u16 transmit_ring; ++ u16 receive_path; ++ u16 receive_ring; ++}; ++ ++struct icm_tr_pkg_disconnect_xdomain { ++ struct icm_pkg_header hdr; ++ u8 stage; ++ u8 reserved[3]; ++ u32 route_hi; ++ u32 route_lo; ++ uuid_t remote_uuid; ++}; ++ ++struct icm_tr_pkg_challenge_device_response { ++ struct icm_pkg_header hdr; ++ uuid_t ep_uuid; ++ u32 route_hi; ++ u32 route_lo; ++ u8 connection_id; ++ u8 reserved[3]; ++ u32 challenge[8]; ++ u32 response[8]; ++}; ++ ++struct icm_tr_pkg_add_device_key_response { ++ struct icm_pkg_header hdr; ++ uuid_t ep_uuid; ++ u32 route_hi; ++ u32 route_lo; ++ u8 connection_id; ++ u8 reserved[3]; ++}; ++ ++struct icm_tr_pkg_approve_xdomain_response { ++ struct icm_pkg_header hdr; ++ u32 route_hi; ++ u32 route_lo; ++ uuid_t remote_uuid; ++ u16 transmit_path; ++ u16 transmit_ring; ++ u16 receive_path; ++ u16 receive_ring; ++}; ++ ++struct icm_tr_pkg_disconnect_xdomain_response { ++ struct icm_pkg_header hdr; ++ u8 stage; ++ u8 reserved[3]; ++ u32 route_hi; ++ u32 route_lo; ++ uuid_t remote_uuid; ++}; ++ ++struct icm_icl_event_rtd3_veto { ++ struct icm_pkg_header hdr; ++ u32 veto_reason; ++}; ++ ++struct icm_usb4_switch_op { ++ struct icm_pkg_header hdr; ++ u32 route_hi; ++ u32 route_lo; ++ u32 metadata; ++ u16 opcode; ++ u16 data_len_valid; ++ u32 data[16]; ++}; ++ ++struct icm_usb4_switch_op_response { ++ struct icm_pkg_header hdr; ++ u32 route_hi; ++ u32 route_lo; ++ u32 metadata; ++ u16 opcode; ++ u16 status; ++ u32 data[16]; ++}; ++ ++enum usb4_switch_op { ++ USB4_SWITCH_OP_QUERY_DP_RESOURCE = 16, ++ USB4_SWITCH_OP_ALLOC_DP_RESOURCE = 17, ++ USB4_SWITCH_OP_DEALLOC_DP_RESOURCE = 18, ++ USB4_SWITCH_OP_NVM_WRITE = 32, ++ USB4_SWITCH_OP_NVM_AUTH = 33, ++ USB4_SWITCH_OP_NVM_READ = 34, ++ USB4_SWITCH_OP_NVM_SET_OFFSET = 35, ++ USB4_SWITCH_OP_DROM_READ = 36, ++ USB4_SWITCH_OP_NVM_SECTOR_SIZE = 37, ++ USB4_SWITCH_OP_BUFFER_ALLOC = 51, ++}; ++ ++struct icm; ++ ++struct usb4_switch_nvm_auth { ++ struct icm_usb4_switch_op_response reply; ++ struct icm_usb4_switch_op request; ++ struct icm *icm; ++}; ++ ++struct icm { ++ struct mutex request_lock; ++ struct delayed_work rescan_work; ++ struct pci_dev *upstream_port; ++ int vnd_cap; ++ bool safe_mode; ++ size_t max_boot_acl; ++ bool rpm; ++ bool can_upgrade_nvm; ++ u8 proto_version; ++ struct usb4_switch_nvm_auth *last_nvm_auth; ++ bool veto; ++ bool (*is_supported)(struct tb *); ++ int (*cio_reset)(struct tb *); ++ int (*get_mode)(struct tb *); ++ int (*get_route)(struct tb *, u8, u8, u64 *); ++ void (*save_devices)(struct tb *); ++ int (*driver_ready)(struct tb *, enum tb_security_level *, u8 *, size_t *, bool *); ++ void (*set_uuid)(struct tb *); ++ void (*device_connected)(struct tb *, const struct icm_pkg_header *); ++ void (*device_disconnected)(struct tb *, const struct icm_pkg_header *); ++ void (*xdomain_connected)(struct tb *, const struct icm_pkg_header *); ++ void (*xdomain_disconnected)(struct tb *, const struct icm_pkg_header *); ++ void (*rtd3_veto)(struct tb *, const struct icm_pkg_header *); ++}; ++ ++struct icm_notification { ++ struct work_struct work; ++ struct icm_pkg_header *pkg; ++ struct tb *tb; ++}; ++ ++struct ep_name_entry { ++ u8 len; ++ u8 type; ++ u8 data[0]; ++}; ++ ++struct intel_vss { ++ u16 vendor; ++ u16 model; ++ u8 mc; ++ u8 flags; ++ u16 pci_devid; ++ u32 nvm_version; ++}; ++ ++enum usb4_sb_opcode { ++ USB4_SB_OPCODE_ERR = 542265925, ++ USB4_SB_OPCODE_ONS = 1145914145, ++ USB4_SB_OPCODE_ROUTER_OFFLINE = 1313166156, ++ USB4_SB_OPCODE_ENUMERATE_RETIMERS = 1297436229, ++ USB4_SB_OPCODE_SET_INBOUND_SBTX = 1347769164, ++ USB4_SB_OPCODE_UNSET_INBOUND_SBTX = 1347769173, ++ USB4_SB_OPCODE_QUERY_LAST_RETIMER = 1414742348, ++ USB4_SB_OPCODE_GET_NVM_SECTOR_SIZE = 1397968455, ++ USB4_SB_OPCODE_NVM_SET_OFFSET = 1397772098, ++ USB4_SB_OPCODE_NVM_BLOCK_WRITE = 1464552514, ++ USB4_SB_OPCODE_NVM_AUTH_WRITE = 1213486401, ++ USB4_SB_OPCODE_NVM_READ = 1381123649, ++ USB4_SB_OPCODE_READ_LANE_MARGINING_CAP = 1346585682, ++ USB4_SB_OPCODE_RUN_HW_LANE_MARGINING = 1196247122, ++ USB4_SB_OPCODE_RUN_SW_LANE_MARGINING = 1196249938, ++ USB4_SB_OPCODE_READ_SW_MARGIN_ERR = 1465074770, ++}; ++ ++enum usb4_sb_target { ++ USB4_SB_TARGET_ROUTER = 0, ++ USB4_SB_TARGET_PARTNER = 1, ++ USB4_SB_TARGET_RETIMER = 2, ++}; ++ ++enum usb4_ba_index { ++ USB4_BA_MAX_USB3 = 1, ++ USB4_BA_MIN_DP_AUX = 2, ++ USB4_BA_MIN_DP_MAIN = 3, ++ USB4_BA_MAX_PCIE = 4, ++ USB4_BA_MAX_HI = 5, ++}; ++ ++struct retimer_info { ++ struct tb_port *port; ++ u8 index; ++}; ++ ++struct tb_retimer_lookup { ++ const struct tb_port *port; ++ u8 index; ++}; ++ ++struct nvmem_cell_lookup { ++ const char *nvmem_name; ++ const char *cell_name; ++ const char *dev_id; ++ const char *con_id; ++ struct list_head node; ++}; ++ ++enum { ++ NVMEM_ADD = 1, ++ NVMEM_REMOVE = 2, ++ NVMEM_CELL_ADD = 3, ++ NVMEM_CELL_REMOVE = 4, ++}; ++ ++struct nvmem_cell_table { ++ const char *nvmem_name; ++ const struct nvmem_cell_info *cells; ++ size_t ncells; ++ struct list_head node; ++}; ++ ++struct nvmem_device { ++ struct module *owner; ++ struct device dev; ++ int stride; ++ int word_size; ++ int id; ++ struct kref refcnt; ++ size_t size; ++ bool read_only; ++ bool root_only; ++ int flags; ++ enum nvmem_type type; ++ struct bin_attribute eeprom; ++ struct device *base_dev; ++ struct list_head cells; ++ const struct nvmem_keepout *keepout; ++ unsigned int nkeepout; ++ nvmem_reg_read_t reg_read; ++ nvmem_reg_write_t reg_write; ++ struct gpio_desc *wp_gpio; ++ struct nvmem_layout *layout; ++ void *priv; ++}; ++ ++struct nvmem_cell_entry { ++ const char *name; ++ int offset; ++ size_t raw_len; ++ int bytes; ++ int bit_offset; ++ int nbits; ++ nvmem_cell_post_process_t read_post_process; ++ void *priv; ++ struct device_node *np; ++ struct nvmem_device *nvmem; ++ struct list_head node; ++}; ++ ++struct nvmem_cell { ++ struct nvmem_cell_entry *entry; ++ const char *id; ++ int index; ++}; ++ ++struct net_device_devres { ++ struct net_device *ndev; ++}; ++ ++enum { ++ TCA_STATS_UNSPEC = 0, ++ TCA_STATS_BASIC = 1, ++ TCA_STATS_RATE_EST = 2, ++ TCA_STATS_QUEUE = 3, ++ TCA_STATS_APP = 4, ++ TCA_STATS_RATE_EST64 = 5, ++ TCA_STATS_PAD = 6, ++ TCA_STATS_BASIC_HW = 7, ++ TCA_STATS_PKT64 = 8, ++ __TCA_STATS_MAX = 9, ++}; ++ ++struct gnet_stats_basic { ++ __u64 bytes; ++ __u32 packets; ++}; ++ ++struct gnet_stats_rate_est { ++ __u32 bps; ++ __u32 pps; ++}; ++ ++enum flow_dissect_ret { ++ FLOW_DISSECT_RET_OUT_GOOD = 0, ++ FLOW_DISSECT_RET_OUT_BAD = 1, ++ FLOW_DISSECT_RET_PROTO_AGAIN = 2, ++ FLOW_DISSECT_RET_IPPROTO_AGAIN = 3, ++ FLOW_DISSECT_RET_CONTINUE = 4, ++}; ++ ++struct flow_dissector_key_hash { ++ u32 hash; ++}; ++ ++struct flow_dissector_key_num_of_vlans { ++ u8 num_of_vlans; ++}; ++ ++struct flow_dissector_key { ++ enum flow_dissector_key_id key_id; ++ size_t offset; ++}; ++ ++struct flow_keys_digest { ++ u8 data[16]; ++}; ++ ++struct nf_ct_event; ++ ++struct nf_exp_event; ++ ++struct nf_ct_event_notifier { ++ int (*ct_event)(unsigned int, const struct nf_ct_event *); ++ int (*exp_event)(unsigned int, const struct nf_exp_event *); ++}; ++ ++enum { ++ TCA_FLOWER_KEY_CT_FLAGS_NEW = 1, ++ TCA_FLOWER_KEY_CT_FLAGS_ESTABLISHED = 2, ++ TCA_FLOWER_KEY_CT_FLAGS_RELATED = 4, ++ TCA_FLOWER_KEY_CT_FLAGS_TRACKED = 8, ++ TCA_FLOWER_KEY_CT_FLAGS_INVALID = 16, ++ TCA_FLOWER_KEY_CT_FLAGS_REPLY = 32, ++ __TCA_FLOWER_KEY_CT_FLAGS_MAX = 33, ++}; ++ ++enum { ++ BPF_FLOW_DISSECTOR_F_PARSE_1ST_FRAG = 1, ++ BPF_FLOW_DISSECTOR_F_STOP_AT_FLOW_LABEL = 2, ++ BPF_FLOW_DISSECTOR_F_STOP_AT_ENCAP = 4, ++}; ++ ++struct arphdr { ++ __be16 ar_hrd; ++ __be16 ar_pro; ++ unsigned char ar_hln; ++ unsigned char ar_pln; ++ __be16 ar_op; ++}; ++ ++struct gre_full_hdr { ++ struct gre_base_hdr fixed_header; ++ __be16 csum; ++ __be16 reserved1; ++ __be32 key; ++ __be32 seq; ++}; ++ ++struct pptp_gre_header { ++ struct gre_base_hdr gre_hd; ++ __be16 payload_len; ++ __be16 call_id; ++ __be32 seq; ++ __be32 ack; ++}; ++ ++struct tipc_basic_hdr { ++ __be32 w[4]; ++}; ++ ++struct pppoe_tag { ++ __be16 tag_type; ++ __be16 tag_len; ++ char tag_data[0]; ++}; ++ ++struct pppoe_hdr { ++ __u8 ver: 4; ++ __u8 type: 4; ++ __u8 code; ++ __be16 sid; ++ __be16 length; ++ struct pppoe_tag tag[0]; ++}; ++ ++struct hsr_tag { ++ __be16 path_and_LSDU_size; ++ __be16 sequence_nr; ++ __be16 encap_proto; ++}; ++ ++struct mpls_label { ++ __be32 entry; ++}; ++ ++struct clock_identity { ++ u8 id[8]; ++}; ++ ++struct port_identity { ++ struct clock_identity clock_identity; ++ __be16 port_number; ++}; ++ ++struct ptp_header { ++ u8 tsmt; ++ u8 ver; ++ __be16 message_length; ++ u8 domain_number; ++ u8 reserved1; ++ u8 flag_field[2]; ++ __be64 correction; ++ __be32 reserved2; ++ struct port_identity source_port_identity; ++ __be16 sequence_id; ++ u8 control; ++ u8 log_message_interval; ++} __attribute__((packed)); ++ ++enum batadv_packettype { ++ BATADV_IV_OGM = 0, ++ BATADV_BCAST = 1, ++ BATADV_CODED = 2, ++ BATADV_ELP = 3, ++ BATADV_OGM2 = 4, ++ BATADV_MCAST = 5, ++ BATADV_UNICAST = 64, ++ BATADV_UNICAST_FRAG = 65, ++ BATADV_UNICAST_4ADDR = 66, ++ BATADV_ICMP = 67, ++ BATADV_UNICAST_TVLV = 68, ++}; ++ ++struct batadv_unicast_packet { ++ __u8 packet_type; ++ __u8 version; ++ __u8 ttl; ++ __u8 ttvn; ++ __u8 dest[6]; ++}; ++ ++struct nf_conntrack_tuple_mask { ++ struct { ++ union nf_inet_addr u3; ++ union nf_conntrack_man_proto u; ++ } src; ++}; ++ ++struct nf_ct_ext { ++ u8 offset[9]; ++ u8 len; ++ unsigned int gen_id; ++ char data[0]; ++}; ++ ++struct nf_conntrack_helper; ++ ++struct nf_conntrack_expect { ++ struct hlist_node lnode; ++ struct hlist_node hnode; ++ struct nf_conntrack_tuple tuple; ++ struct nf_conntrack_tuple_mask mask; ++ void (*expectfn)(struct nf_conn *, struct nf_conntrack_expect *); ++ struct nf_conntrack_helper *helper; ++ struct nf_conn *master; ++ struct timer_list timeout; ++ refcount_t use; ++ unsigned int flags; ++ unsigned int class; ++ union nf_inet_addr saved_addr; ++ union nf_conntrack_man_proto saved_proto; ++ enum ip_conntrack_dir dir; ++ struct callback_head rcu; ++}; ++ ++struct nf_ct_event { ++ struct nf_conn *ct; ++ u32 portid; ++ int report; ++}; ++ ++struct nf_exp_event { ++ struct nf_conntrack_expect *exp; ++ u32 portid; ++ int report; ++}; ++ ++struct nf_conn_labels { ++ long unsigned int bits[2]; ++}; ++ ++struct _flow_keys_digest_data { ++ __be16 n_proto; ++ u8 ip_proto; ++ u8 padding; ++ __be32 ports; ++ __be32 src; ++ __be32 dst; ++}; ++ ++struct nda_cacheinfo { ++ __u32 ndm_confirmed; ++ __u32 ndm_used; ++ __u32 ndm_updated; ++ __u32 ndm_refcnt; ++}; ++ ++struct ndt_stats { ++ __u64 ndts_allocs; ++ __u64 ndts_destroys; ++ __u64 ndts_hash_grows; ++ __u64 ndts_res_failed; ++ __u64 ndts_lookups; ++ __u64 ndts_hits; ++ __u64 ndts_rcv_probes_mcast; ++ __u64 ndts_rcv_probes_ucast; ++ __u64 ndts_periodic_gc_runs; ++ __u64 ndts_forced_gc_runs; ++ __u64 ndts_table_fulls; ++}; ++ ++enum { ++ NDTPA_UNSPEC = 0, ++ NDTPA_IFINDEX = 1, ++ NDTPA_REFCNT = 2, ++ NDTPA_REACHABLE_TIME = 3, ++ NDTPA_BASE_REACHABLE_TIME = 4, ++ NDTPA_RETRANS_TIME = 5, ++ NDTPA_GC_STALETIME = 6, ++ NDTPA_DELAY_PROBE_TIME = 7, ++ NDTPA_QUEUE_LEN = 8, ++ NDTPA_APP_PROBES = 9, ++ NDTPA_UCAST_PROBES = 10, ++ NDTPA_MCAST_PROBES = 11, ++ NDTPA_ANYCAST_DELAY = 12, ++ NDTPA_PROXY_DELAY = 13, ++ NDTPA_PROXY_QLEN = 14, ++ NDTPA_LOCKTIME = 15, ++ NDTPA_QUEUE_LENBYTES = 16, ++ NDTPA_MCAST_REPROBES = 17, ++ NDTPA_PAD = 18, ++ NDTPA_INTERVAL_PROBE_TIME_MS = 19, ++ __NDTPA_MAX = 20, ++}; ++ ++struct ndtmsg { ++ __u8 ndtm_family; ++ __u8 ndtm_pad1; ++ __u16 ndtm_pad2; ++}; ++ ++struct ndt_config { ++ __u16 ndtc_key_len; ++ __u16 ndtc_entry_size; ++ __u32 ndtc_entries; ++ __u32 ndtc_last_flush; ++ __u32 ndtc_last_rand; ++ __u32 ndtc_hash_rnd; ++ __u32 ndtc_hash_mask; ++ __u32 ndtc_hash_chain_gc; ++ __u32 ndtc_proxy_qlen; ++}; ++ ++enum { ++ NDTA_UNSPEC = 0, ++ NDTA_NAME = 1, ++ NDTA_THRESH1 = 2, ++ NDTA_THRESH2 = 3, ++ NDTA_THRESH3 = 4, ++ NDTA_CONFIG = 5, ++ NDTA_PARMS = 6, ++ NDTA_STATS = 7, ++ NDTA_GC_INTERVAL = 8, ++ NDTA_PAD = 9, ++ __NDTA_MAX = 10, ++}; ++ ++enum { ++ NEIGH_ARP_TABLE = 0, ++ NEIGH_ND_TABLE = 1, ++ NEIGH_DN_TABLE = 2, ++ NEIGH_NR_TABLES = 3, ++ NEIGH_LINK_TABLE = 3, ++}; ++ ++struct neigh_seq_state { ++ struct seq_net_private p; ++ struct neigh_table *tbl; ++ struct neigh_hash_table *nht; ++ void * (*neigh_sub_iter)(struct neigh_seq_state *, struct neighbour *, loff_t *); ++ unsigned int bucket; ++ unsigned int flags; ++}; ++ ++struct neighbour_cb { ++ long unsigned int sched_next; ++ unsigned int flags; ++}; ++ ++struct neigh_dump_filter { ++ int master_idx; ++ int dev_idx; ++}; ++ ++struct neigh_sysctl_table { ++ struct ctl_table_header *sysctl_header; ++ struct ctl_table neigh_vars[22]; ++}; ++ ++enum { ++ IF_LINK_MODE_DEFAULT = 0, ++ IF_LINK_MODE_DORMANT = 1, ++ IF_LINK_MODE_TESTING = 2, ++}; ++ ++enum lw_bits { ++ LW_URGENT = 0, ++}; ++ ++struct sock_diag_req { ++ __u8 sdiag_family; ++ __u8 sdiag_protocol; ++}; ++ ++struct broadcast_sk { ++ struct sock *sk; ++ struct work_struct work; ++}; ++ ++struct tso_t { ++ int next_frag_idx; ++ int size; ++ void *data; ++ u16 ip_id; ++ u8 tlen; ++ bool ipv6; ++ u32 tcp_seq; ++}; ++ ++struct fib_notifier_net { ++ struct list_head fib_notifier_ops; ++ struct atomic_notifier_head fib_chain; ++}; ++ ++enum gro_result { ++ GRO_MERGED = 0, ++ GRO_MERGED_FREE = 1, ++ GRO_HELD = 2, ++ GRO_NORMAL = 3, ++ GRO_CONSUMED = 4, ++}; ++ ++typedef enum gro_result gro_result_t; ++ ++struct trace_event_raw_kfree_skb { ++ struct trace_entry ent; ++ void *skbaddr; ++ void *location; ++ short unsigned int protocol; ++ enum skb_drop_reason reason; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_consume_skb { ++ struct trace_entry ent; ++ void *skbaddr; ++ void *location; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_skb_copy_datagram_iovec { ++ struct trace_entry ent; ++ const void *skbaddr; ++ int len; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_kfree_skb {}; ++ ++struct trace_event_data_offsets_consume_skb {}; ++ ++struct trace_event_data_offsets_skb_copy_datagram_iovec {}; ++ ++typedef void (*btf_trace_kfree_skb)(void *, struct sk_buff *, void *, enum skb_drop_reason); ++ ++typedef void (*btf_trace_consume_skb)(void *, struct sk_buff *, void *); ++ ++typedef void (*btf_trace_skb_copy_datagram_iovec)(void *, const struct sk_buff *, int); ++ ++struct trace_event_raw_net_dev_start_xmit { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ u16 queue_mapping; ++ const void *skbaddr; ++ bool vlan_tagged; ++ u16 vlan_proto; ++ u16 vlan_tci; ++ u16 protocol; ++ u8 ip_summed; ++ unsigned int len; ++ unsigned int data_len; ++ int network_offset; ++ bool transport_offset_valid; ++ int transport_offset; ++ u8 tx_flags; ++ u16 gso_size; ++ u16 gso_segs; ++ u16 gso_type; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_net_dev_xmit { ++ struct trace_entry ent; ++ void *skbaddr; ++ unsigned int len; ++ int rc; ++ u32 __data_loc_name; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_net_dev_xmit_timeout { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ u32 __data_loc_driver; ++ int queue_index; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_net_dev_template { ++ struct trace_entry ent; ++ void *skbaddr; ++ unsigned int len; ++ u32 __data_loc_name; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_net_dev_rx_verbose_template { ++ struct trace_entry ent; ++ u32 __data_loc_name; ++ unsigned int napi_id; ++ u16 queue_mapping; ++ const void *skbaddr; ++ bool vlan_tagged; ++ u16 vlan_proto; ++ u16 vlan_tci; ++ u16 protocol; ++ u8 ip_summed; ++ u32 hash; ++ bool l4_hash; ++ unsigned int len; ++ unsigned int data_len; ++ unsigned int truesize; ++ bool mac_header_valid; ++ int mac_header; ++ unsigned char nr_frags; ++ u16 gso_size; ++ u16 gso_type; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_net_dev_rx_exit_template { ++ struct trace_entry ent; ++ int ret; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_net_dev_start_xmit { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_net_dev_xmit { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_net_dev_xmit_timeout { ++ u32 name; ++ u32 driver; ++}; ++ ++struct trace_event_data_offsets_net_dev_template { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_net_dev_rx_verbose_template { ++ u32 name; ++}; ++ ++struct trace_event_data_offsets_net_dev_rx_exit_template {}; ++ ++typedef void (*btf_trace_net_dev_start_xmit)(void *, const struct sk_buff *, const struct net_device *); ++ ++typedef void (*btf_trace_net_dev_xmit)(void *, struct sk_buff *, int, struct net_device *, unsigned int); ++ ++typedef void (*btf_trace_net_dev_xmit_timeout)(void *, struct net_device *, int); ++ ++typedef void (*btf_trace_net_dev_queue)(void *, struct sk_buff *); ++ ++typedef void (*btf_trace_netif_receive_skb)(void *, struct sk_buff *); ++ ++typedef void (*btf_trace_netif_rx)(void *, struct sk_buff *); ++ ++typedef void (*btf_trace_napi_gro_frags_entry)(void *, const struct sk_buff *); ++ ++typedef void (*btf_trace_napi_gro_receive_entry)(void *, const struct sk_buff *); ++ ++typedef void (*btf_trace_netif_receive_skb_entry)(void *, const struct sk_buff *); ++ ++typedef void (*btf_trace_netif_receive_skb_list_entry)(void *, const struct sk_buff *); ++ ++typedef void (*btf_trace_netif_rx_entry)(void *, const struct sk_buff *); ++ ++typedef void (*btf_trace_napi_gro_frags_exit)(void *, int); ++ ++typedef void (*btf_trace_napi_gro_receive_exit)(void *, int); ++ ++typedef void (*btf_trace_netif_receive_skb_exit)(void *, int); ++ ++typedef void (*btf_trace_netif_rx_exit)(void *, int); ++ ++typedef void (*btf_trace_netif_receive_skb_list_exit)(void *, int); ++ ++struct trace_event_raw_napi_poll { ++ struct trace_entry ent; ++ struct napi_struct *napi; ++ u32 __data_loc_dev_name; ++ int work; ++ int budget; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_napi_poll { ++ u32 dev_name; ++}; ++ ++typedef void (*btf_trace_napi_poll)(void *, struct napi_struct *, int, int); ++ ++struct trace_event_raw_sock_rcvqueue_full { ++ struct trace_entry ent; ++ int rmem_alloc; ++ unsigned int truesize; ++ int sk_rcvbuf; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sock_exceed_buf_limit { ++ struct trace_entry ent; ++ char name[32]; ++ long int sysctl_mem[3]; ++ long int allocated; ++ int sysctl_rmem; ++ int rmem_alloc; ++ int sysctl_wmem; ++ int wmem_alloc; ++ int wmem_queued; ++ int kind; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_inet_sock_set_state { ++ struct trace_entry ent; ++ const void *skaddr; ++ int oldstate; ++ int newstate; ++ __u16 sport; ++ __u16 dport; ++ __u16 family; ++ __u16 protocol; ++ __u8 saddr[4]; ++ __u8 daddr[4]; ++ __u8 saddr_v6[16]; ++ __u8 daddr_v6[16]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_inet_sk_error_report { ++ struct trace_entry ent; ++ int error; ++ __u16 sport; ++ __u16 dport; ++ __u16 family; ++ __u16 protocol; ++ __u8 saddr[4]; ++ __u8 daddr[4]; ++ __u8 saddr_v6[16]; ++ __u8 daddr_v6[16]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sk_data_ready { ++ struct trace_entry ent; ++ const void *skaddr; ++ __u16 family; ++ __u16 protocol; ++ long unsigned int ip; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_sock_msg_length { ++ struct trace_entry ent; ++ void *sk; ++ __u16 family; ++ __u16 protocol; ++ int ret; ++ int flags; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_sock_rcvqueue_full {}; ++ ++struct trace_event_data_offsets_sock_exceed_buf_limit {}; ++ ++struct trace_event_data_offsets_inet_sock_set_state {}; ++ ++struct trace_event_data_offsets_inet_sk_error_report {}; ++ ++struct trace_event_data_offsets_sk_data_ready {}; ++ ++struct trace_event_data_offsets_sock_msg_length {}; ++ ++typedef void (*btf_trace_sock_rcvqueue_full)(void *, struct sock *, struct sk_buff *); ++ ++typedef void (*btf_trace_sock_exceed_buf_limit)(void *, struct sock *, struct proto *, long int, int); ++ ++typedef void (*btf_trace_inet_sock_set_state)(void *, const struct sock *, const int, const int); ++ ++typedef void (*btf_trace_inet_sk_error_report)(void *, const struct sock *); ++ ++typedef void (*btf_trace_sk_data_ready)(void *, const struct sock *); ++ ++typedef void (*btf_trace_sock_send_length)(void *, struct sock *, int, int); ++ ++typedef void (*btf_trace_sock_recv_length)(void *, struct sock *, int, int); ++ ++struct trace_event_raw_udp_fail_queue_rcv_skb { ++ struct trace_entry ent; ++ int rc; ++ __u16 lport; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_udp_fail_queue_rcv_skb {}; ++ ++typedef void (*btf_trace_udp_fail_queue_rcv_skb)(void *, int, struct sock *); ++ ++struct trace_event_raw_tcp_event_sk_skb { ++ struct trace_entry ent; ++ const void *skbaddr; ++ const void *skaddr; ++ int state; ++ __u16 sport; ++ __u16 dport; ++ __u16 family; ++ __u8 saddr[4]; ++ __u8 daddr[4]; ++ __u8 saddr_v6[16]; ++ __u8 daddr_v6[16]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_tcp_event_sk { ++ struct trace_entry ent; ++ const void *skaddr; ++ __u16 sport; ++ __u16 dport; ++ __u16 family; ++ __u8 saddr[4]; ++ __u8 daddr[4]; ++ __u8 saddr_v6[16]; ++ __u8 daddr_v6[16]; ++ __u64 sock_cookie; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_tcp_retransmit_synack { ++ struct trace_entry ent; ++ const void *skaddr; ++ const void *req; ++ __u16 sport; ++ __u16 dport; ++ __u16 family; ++ __u8 saddr[4]; ++ __u8 daddr[4]; ++ __u8 saddr_v6[16]; ++ __u8 daddr_v6[16]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_tcp_probe { ++ struct trace_entry ent; ++ __u8 saddr[28]; ++ __u8 daddr[28]; ++ __u16 sport; ++ __u16 dport; ++ __u16 family; ++ __u32 mark; ++ __u16 data_len; ++ __u32 snd_nxt; ++ __u32 snd_una; ++ __u32 snd_cwnd; ++ __u32 ssthresh; ++ __u32 snd_wnd; ++ __u32 srtt; ++ __u32 rcv_wnd; ++ __u64 sock_cookie; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_tcp_event_skb { ++ struct trace_entry ent; ++ const void *skbaddr; ++ __u8 saddr[28]; ++ __u8 daddr[28]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_tcp_cong_state_set { ++ struct trace_entry ent; ++ const void *skaddr; ++ __u16 sport; ++ __u16 dport; ++ __u16 family; ++ __u8 saddr[4]; ++ __u8 daddr[4]; ++ __u8 saddr_v6[16]; ++ __u8 daddr_v6[16]; ++ __u8 cong_state; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_tcp_event_sk_skb {}; ++ ++struct trace_event_data_offsets_tcp_event_sk {}; ++ ++struct trace_event_data_offsets_tcp_retransmit_synack {}; ++ ++struct trace_event_data_offsets_tcp_probe {}; ++ ++struct trace_event_data_offsets_tcp_event_skb {}; ++ ++struct trace_event_data_offsets_tcp_cong_state_set {}; ++ ++typedef void (*btf_trace_tcp_retransmit_skb)(void *, const struct sock *, const struct sk_buff *); ++ ++typedef void (*btf_trace_tcp_send_reset)(void *, const struct sock *, const struct sk_buff *); ++ ++typedef void (*btf_trace_tcp_receive_reset)(void *, struct sock *); ++ ++typedef void (*btf_trace_tcp_destroy_sock)(void *, struct sock *); ++ ++typedef void (*btf_trace_tcp_rcv_space_adjust)(void *, struct sock *); ++ ++typedef void (*btf_trace_tcp_retransmit_synack)(void *, const struct sock *, const struct request_sock *); ++ ++typedef void (*btf_trace_tcp_probe)(void *, struct sock *, struct sk_buff *); ++ ++typedef void (*btf_trace_tcp_bad_csum)(void *, const struct sk_buff *); ++ ++typedef void (*btf_trace_tcp_cong_state_set)(void *, struct sock *, const u8); ++ ++struct trace_event_raw_fib_table_lookup { ++ struct trace_entry ent; ++ u32 tb_id; ++ int err; ++ int oif; ++ int iif; ++ u8 proto; ++ __u8 tos; ++ __u8 scope; ++ __u8 flags; ++ __u8 src[4]; ++ __u8 dst[4]; ++ __u8 gw4[4]; ++ __u8 gw6[16]; ++ u16 sport; ++ u16 dport; ++ char name[16]; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_fib_table_lookup {}; ++ ++typedef void (*btf_trace_fib_table_lookup)(void *, u32, const struct flowi4 *, const struct fib_nh_common *, int); ++ ++struct trace_event_raw_qdisc_dequeue { ++ struct trace_entry ent; ++ struct Qdisc *qdisc; ++ const struct netdev_queue *txq; ++ int packets; ++ void *skbaddr; ++ int ifindex; ++ u32 handle; ++ u32 parent; ++ long unsigned int txq_state; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_qdisc_enqueue { ++ struct trace_entry ent; ++ struct Qdisc *qdisc; ++ const struct netdev_queue *txq; ++ void *skbaddr; ++ int ifindex; ++ u32 handle; ++ u32 parent; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_qdisc_reset { ++ struct trace_entry ent; ++ u32 __data_loc_dev; ++ u32 __data_loc_kind; ++ u32 parent; ++ u32 handle; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_qdisc_destroy { ++ struct trace_entry ent; ++ u32 __data_loc_dev; ++ u32 __data_loc_kind; ++ u32 parent; ++ u32 handle; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_qdisc_create { ++ struct trace_entry ent; ++ u32 __data_loc_dev; ++ u32 __data_loc_kind; ++ u32 parent; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_qdisc_dequeue {}; ++ ++struct trace_event_data_offsets_qdisc_enqueue {}; ++ ++struct trace_event_data_offsets_qdisc_reset { ++ u32 dev; ++ u32 kind; ++}; ++ ++struct trace_event_data_offsets_qdisc_destroy { ++ u32 dev; ++ u32 kind; ++}; ++ ++struct trace_event_data_offsets_qdisc_create { ++ u32 dev; ++ u32 kind; ++}; ++ ++typedef void (*btf_trace_qdisc_dequeue)(void *, struct Qdisc *, const struct netdev_queue *, int, struct sk_buff *); ++ ++typedef void (*btf_trace_qdisc_enqueue)(void *, struct Qdisc *, const struct netdev_queue *, struct sk_buff *); ++ ++typedef void (*btf_trace_qdisc_reset)(void *, struct Qdisc *); ++ ++typedef void (*btf_trace_qdisc_destroy)(void *, struct Qdisc *); ++ ++typedef void (*btf_trace_qdisc_create)(void *, const struct Qdisc_ops *, struct net_device *, u32); ++ ++struct bridge_stp_xstats { ++ __u64 transition_blk; ++ __u64 transition_fwd; ++ __u64 rx_bpdu; ++ __u64 tx_bpdu; ++ __u64 rx_tcn; ++ __u64 tx_tcn; ++}; ++ ++struct br_mcast_stats { ++ __u64 igmp_v1queries[2]; ++ __u64 igmp_v2queries[2]; ++ __u64 igmp_v3queries[2]; ++ __u64 igmp_leaves[2]; ++ __u64 igmp_v1reports[2]; ++ __u64 igmp_v2reports[2]; ++ __u64 igmp_v3reports[2]; ++ __u64 igmp_parse_errors; ++ __u64 mld_v1queries[2]; ++ __u64 mld_v2queries[2]; ++ __u64 mld_leaves[2]; ++ __u64 mld_v1reports[2]; ++ __u64 mld_v2reports[2]; ++ __u64 mld_parse_errors; ++ __u64 mcast_bytes[2]; ++ __u64 mcast_packets[2]; ++}; ++ ++struct br_ip { ++ union { ++ __be32 ip4; ++ struct in6_addr ip6; ++ } src; ++ union { ++ __be32 ip4; ++ struct in6_addr ip6; ++ unsigned char mac_addr[6]; ++ } dst; ++ __be16 proto; ++ __u16 vid; ++}; ++ ++struct bridge_id { ++ unsigned char prio[2]; ++ unsigned char addr[6]; ++}; ++ ++typedef struct bridge_id bridge_id; ++ ++struct mac_addr { ++ unsigned char addr[6]; ++}; ++ ++typedef struct mac_addr mac_addr; ++ ++typedef __u16 port_id; ++ ++struct bridge_mcast_own_query { ++ struct timer_list timer; ++ u32 startup_sent; ++}; ++ ++struct bridge_mcast_other_query { ++ struct timer_list timer; ++ long unsigned int delay_time; ++}; ++ ++struct bridge_mcast_querier { ++ struct br_ip addr; ++ int port_ifidx; ++ seqcount_spinlock_t seq; ++}; ++ ++struct bridge_mcast_stats { ++ struct br_mcast_stats mstats; ++ struct u64_stats_sync syncp; ++}; ++ ++struct net_bridge; ++ ++struct net_bridge_vlan; ++ ++struct net_bridge_mcast { ++ struct net_bridge *br; ++ struct net_bridge_vlan *vlan; ++ u32 multicast_last_member_count; ++ u32 multicast_startup_query_count; ++ u8 multicast_querier; ++ u8 multicast_igmp_version; ++ u8 multicast_router; ++ u8 multicast_mld_version; ++ long unsigned int multicast_last_member_interval; ++ long unsigned int multicast_membership_interval; ++ long unsigned int multicast_querier_interval; ++ long unsigned int multicast_query_interval; ++ long unsigned int multicast_query_response_interval; ++ long unsigned int multicast_startup_query_interval; ++ struct hlist_head ip4_mc_router_list; ++ struct timer_list ip4_mc_router_timer; ++ struct bridge_mcast_other_query ip4_other_query; ++ struct bridge_mcast_own_query ip4_own_query; ++ struct bridge_mcast_querier ip4_querier; ++ struct hlist_head ip6_mc_router_list; ++ struct timer_list ip6_mc_router_timer; ++ struct bridge_mcast_other_query ip6_other_query; ++ struct bridge_mcast_own_query ip6_own_query; ++ struct bridge_mcast_querier ip6_querier; ++}; ++ ++struct net_bridge_vlan_group; ++ ++struct net_bridge { ++ spinlock_t lock; ++ spinlock_t hash_lock; ++ struct hlist_head frame_type_list; ++ struct net_device *dev; ++ long unsigned int options; ++ __be16 vlan_proto; ++ u16 default_pvid; ++ struct net_bridge_vlan_group *vlgrp; ++ struct rhashtable fdb_hash_tbl; ++ struct list_head port_list; ++ union { ++ struct rtable fake_rtable; ++ struct rt6_info fake_rt6_info; ++ }; ++ u16 group_fwd_mask; ++ u16 group_fwd_mask_required; ++ bridge_id designated_root; ++ bridge_id bridge_id; ++ unsigned char topology_change; ++ unsigned char topology_change_detected; ++ u16 root_port; ++ long unsigned int max_age; ++ long unsigned int hello_time; ++ long unsigned int forward_delay; ++ long unsigned int ageing_time; ++ long unsigned int bridge_max_age; ++ long unsigned int bridge_hello_time; ++ long unsigned int bridge_forward_delay; ++ long unsigned int bridge_ageing_time; ++ u32 root_path_cost; ++ u8 group_addr[6]; ++ enum { ++ BR_NO_STP = 0, ++ BR_KERNEL_STP = 1, ++ BR_USER_STP = 2, ++ } stp_enabled; ++ struct net_bridge_mcast multicast_ctx; ++ struct bridge_mcast_stats *mcast_stats; ++ u32 hash_max; ++ spinlock_t multicast_lock; ++ struct rhashtable mdb_hash_tbl; ++ struct rhashtable sg_port_tbl; ++ struct hlist_head mcast_gc_list; ++ struct hlist_head mdb_list; ++ struct work_struct mcast_gc_work; ++ struct timer_list hello_timer; ++ struct timer_list tcn_timer; ++ struct timer_list topology_change_timer; ++ struct delayed_work gc_work; ++ struct kobject *ifobj; ++ u32 auto_cnt; ++ int last_hwdom; ++ long unsigned int busy_hwdoms; ++ struct hlist_head fdb_list; ++ struct hlist_head mrp_list; ++ struct hlist_head mep_list; ++}; ++ ++struct net_bridge_port; ++ ++struct net_bridge_mcast_port { ++ struct net_bridge_port *port; ++ struct net_bridge_vlan *vlan; ++ struct bridge_mcast_own_query ip4_own_query; ++ struct timer_list ip4_mc_router_timer; ++ struct hlist_node ip4_rlist; ++ struct bridge_mcast_own_query ip6_own_query; ++ struct timer_list ip6_mc_router_timer; ++ struct hlist_node ip6_rlist; ++ unsigned char multicast_router; ++ u32 mdb_n_entries; ++ u32 mdb_max_entries; ++}; ++ ++struct net_bridge_port { ++ struct net_bridge *br; ++ struct net_device *dev; ++ netdevice_tracker dev_tracker; ++ struct list_head list; ++ long unsigned int flags; ++ struct net_bridge_vlan_group *vlgrp; ++ struct net_bridge_port *backup_port; ++ u8 priority; ++ u8 state; ++ u16 port_no; ++ unsigned char topology_change_ack; ++ unsigned char config_pending; ++ port_id port_id; ++ port_id designated_port; ++ bridge_id designated_root; ++ bridge_id designated_bridge; ++ u32 path_cost; ++ u32 designated_cost; ++ long unsigned int designated_age; ++ struct timer_list forward_delay_timer; ++ struct timer_list hold_timer; ++ struct timer_list message_age_timer; ++ struct kobject kobj; ++ struct callback_head rcu; ++ struct net_bridge_mcast_port multicast_ctx; ++ struct bridge_mcast_stats *mcast_stats; ++ u32 multicast_eht_hosts_limit; ++ u32 multicast_eht_hosts_cnt; ++ struct hlist_head mglist; ++ char sysfs_name[16]; ++ struct netpoll *np; ++ int hwdom; ++ int offload_count; ++ struct netdev_phys_item_id ppid; ++ u16 group_fwd_mask; ++ u16 backup_redirected_cnt; ++ struct bridge_stp_xstats stp_xstats; ++}; ++ ++struct br_tunnel_info { ++ __be64 tunnel_id; ++ struct metadata_dst *tunnel_dst; ++}; ++ ++struct net_bridge_vlan { ++ struct rhash_head vnode; ++ struct rhash_head tnode; ++ u16 vid; ++ u16 flags; ++ u16 priv_flags; ++ u8 state; ++ struct pcpu_sw_netstats *stats; ++ union { ++ struct net_bridge *br; ++ struct net_bridge_port *port; ++ }; ++ union { ++ refcount_t refcnt; ++ struct net_bridge_vlan *brvlan; ++ }; ++ struct br_tunnel_info tinfo; ++ union { ++ struct net_bridge_mcast br_mcast_ctx; ++ struct net_bridge_mcast_port port_mcast_ctx; ++ }; ++ u16 msti; ++ struct list_head vlist; ++ struct callback_head rcu; ++}; ++ ++struct net_bridge_vlan_group { ++ struct rhashtable vlan_hash; ++ struct rhashtable tunnel_hash; ++ struct list_head vlan_list; ++ u16 num_vlans; ++ u16 pvid; ++ u8 pvid_state; ++}; ++ ++struct net_bridge_fdb_key { ++ mac_addr addr; ++ u16 vlan_id; ++}; ++ ++struct net_bridge_fdb_entry { ++ struct rhash_head rhnode; ++ struct net_bridge_port *dst; ++ struct net_bridge_fdb_key key; ++ struct hlist_node fdb_node; ++ long unsigned int flags; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long unsigned int updated; ++ long unsigned int used; ++ struct callback_head rcu; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct trace_event_raw_br_fdb_add { ++ struct trace_entry ent; ++ u8 ndm_flags; ++ u32 __data_loc_dev; ++ unsigned char addr[6]; ++ u16 vid; ++ u16 nlh_flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_br_fdb_external_learn_add { ++ struct trace_entry ent; ++ u32 __data_loc_br_dev; ++ u32 __data_loc_dev; ++ unsigned char addr[6]; ++ u16 vid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_fdb_delete { ++ struct trace_entry ent; ++ u32 __data_loc_br_dev; ++ u32 __data_loc_dev; ++ unsigned char addr[6]; ++ u16 vid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_br_fdb_update { ++ struct trace_entry ent; ++ u32 __data_loc_br_dev; ++ u32 __data_loc_dev; ++ unsigned char addr[6]; ++ u16 vid; ++ long unsigned int flags; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_br_mdb_full { ++ struct trace_entry ent; ++ u32 __data_loc_dev; ++ int af; ++ u16 vid; ++ __u8 src[16]; ++ __u8 grp[16]; ++ __u8 grpmac[6]; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_br_fdb_add { ++ u32 dev; ++}; ++ ++struct trace_event_data_offsets_br_fdb_external_learn_add { ++ u32 br_dev; ++ u32 dev; ++}; ++ ++struct trace_event_data_offsets_fdb_delete { ++ u32 br_dev; ++ u32 dev; ++}; ++ ++struct trace_event_data_offsets_br_fdb_update { ++ u32 br_dev; ++ u32 dev; ++}; ++ ++struct trace_event_data_offsets_br_mdb_full { ++ u32 dev; ++}; ++ ++typedef void (*btf_trace_br_fdb_add)(void *, struct ndmsg *, struct net_device *, const unsigned char *, u16, u16); ++ ++typedef void (*btf_trace_br_fdb_external_learn_add)(void *, struct net_bridge *, struct net_bridge_port *, const unsigned char *, u16); ++ ++typedef void (*btf_trace_fdb_delete)(void *, struct net_bridge *, struct net_bridge_fdb_entry *); ++ ++typedef void (*btf_trace_br_fdb_update)(void *, struct net_bridge *, struct net_bridge_port *, const unsigned char *, u16, long unsigned int); ++ ++typedef void (*btf_trace_br_mdb_full)(void *, const struct net_device *, const struct br_ip *); ++ ++struct trace_event_raw_page_pool_release { ++ struct trace_entry ent; ++ const struct page_pool *pool; ++ s32 inflight; ++ u32 hold; ++ u32 release; ++ u64 cnt; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_page_pool_state_release { ++ struct trace_entry ent; ++ const struct page_pool *pool; ++ const struct page *page; ++ u32 release; ++ long unsigned int pfn; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_page_pool_state_hold { ++ struct trace_entry ent; ++ const struct page_pool *pool; ++ const struct page *page; ++ u32 hold; ++ long unsigned int pfn; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_page_pool_update_nid { ++ struct trace_entry ent; ++ const struct page_pool *pool; ++ int pool_nid; ++ int new_nid; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_page_pool_release {}; ++ ++struct trace_event_data_offsets_page_pool_state_release {}; ++ ++struct trace_event_data_offsets_page_pool_state_hold {}; ++ ++struct trace_event_data_offsets_page_pool_update_nid {}; ++ ++typedef void (*btf_trace_page_pool_release)(void *, const struct page_pool *, s32, u32, u32); ++ ++typedef void (*btf_trace_page_pool_state_release)(void *, const struct page_pool *, const struct page *, u32); ++ ++typedef void (*btf_trace_page_pool_state_hold)(void *, const struct page_pool *, const struct page *, u32); ++ ++typedef void (*btf_trace_page_pool_update_nid)(void *, const struct page_pool *, int); ++ ++struct trace_event_raw_neigh_create { ++ struct trace_entry ent; ++ u32 family; ++ u32 __data_loc_dev; ++ int entries; ++ u8 created; ++ u8 gc_exempt; ++ u8 primary_key4[4]; ++ u8 primary_key6[16]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_neigh_update { ++ struct trace_entry ent; ++ u32 family; ++ u32 __data_loc_dev; ++ u8 lladdr[32]; ++ u8 lladdr_len; ++ u8 flags; ++ u8 nud_state; ++ u8 type; ++ u8 dead; ++ int refcnt; ++ __u8 primary_key4[4]; ++ __u8 primary_key6[16]; ++ long unsigned int confirmed; ++ long unsigned int updated; ++ long unsigned int used; ++ u8 new_lladdr[32]; ++ u8 new_state; ++ u32 update_flags; ++ u32 pid; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_neigh__update { ++ struct trace_entry ent; ++ u32 family; ++ u32 __data_loc_dev; ++ u8 lladdr[32]; ++ u8 lladdr_len; ++ u8 flags; ++ u8 nud_state; ++ u8 type; ++ u8 dead; ++ int refcnt; ++ __u8 primary_key4[4]; ++ __u8 primary_key6[16]; ++ long unsigned int confirmed; ++ long unsigned int updated; ++ long unsigned int used; ++ u32 err; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_neigh_create { ++ u32 dev; ++}; ++ ++struct trace_event_data_offsets_neigh_update { ++ u32 dev; ++}; ++ ++struct trace_event_data_offsets_neigh__update { ++ u32 dev; ++}; ++ ++typedef void (*btf_trace_neigh_create)(void *, struct neigh_table *, struct net_device *, const void *, const struct neighbour *, bool); ++ ++typedef void (*btf_trace_neigh_update)(void *, struct neighbour *, const u8 *, u8, u32, u32); ++ ++typedef void (*btf_trace_neigh_update_done)(void *, struct neighbour *, int); ++ ++typedef void (*btf_trace_neigh_timer_handler)(void *, struct neighbour *, int); ++ ++typedef void (*btf_trace_neigh_event_send_done)(void *, struct neighbour *, int); ++ ++typedef void (*btf_trace_neigh_event_send_dead)(void *, struct neighbour *, int); ++ ++typedef void (*btf_trace_neigh_cleanup_and_release)(void *, struct neighbour *, int); ++ ++struct gro_cell { ++ struct sk_buff_head napi_skbs; ++ struct napi_struct napi; ++}; ++ ++struct percpu_free_defer { ++ struct callback_head rcu; ++ void *ptr; ++}; ++ ++enum { ++ SK_DIAG_BPF_STORAGE_REQ_NONE = 0, ++ SK_DIAG_BPF_STORAGE_REQ_MAP_FD = 1, ++ __SK_DIAG_BPF_STORAGE_REQ_MAX = 2, ++}; ++ ++enum { ++ SK_DIAG_BPF_STORAGE_REP_NONE = 0, ++ SK_DIAG_BPF_STORAGE = 1, ++ __SK_DIAG_BPF_STORAGE_REP_MAX = 2, ++}; ++ ++enum { ++ SK_DIAG_BPF_STORAGE_NONE = 0, ++ SK_DIAG_BPF_STORAGE_PAD = 1, ++ SK_DIAG_BPF_STORAGE_MAP_ID = 2, ++ SK_DIAG_BPF_STORAGE_MAP_VALUE = 3, ++ __SK_DIAG_BPF_STORAGE_MAX = 4, ++}; ++ ++typedef u64 (*btf_bpf_sk_storage_get)(struct bpf_map *, struct sock *, void *, u64, gfp_t); ++ ++typedef u64 (*btf_bpf_sk_storage_delete)(struct bpf_map *, struct sock *); ++ ++typedef u64 (*btf_bpf_sk_storage_get_tracing)(struct bpf_map *, struct sock *, void *, u64, gfp_t); ++ ++typedef u64 (*btf_bpf_sk_storage_delete_tracing)(struct bpf_map *, struct sock *); ++ ++struct bpf_sk_storage_diag { ++ u32 nr_maps; ++ struct bpf_map *maps[0]; ++}; ++ ++struct bpf_iter_seq_sk_storage_map_info { ++ struct bpf_map *map; ++ unsigned int bucket_id; ++ unsigned int skip_elems; ++}; ++ ++struct bpf_iter__bpf_sk_storage_map { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct bpf_map *map; ++ }; ++ union { ++ struct sock *sk; ++ }; ++ union { ++ void *value; ++ }; ++}; ++ ++struct tc_qopt_offload_stats { ++ struct gnet_stats_basic_sync *bstats; ++ struct gnet_stats_queue *qstats; ++}; ++ ++enum tc_mq_command { ++ TC_MQ_CREATE = 0, ++ TC_MQ_DESTROY = 1, ++ TC_MQ_STATS = 2, ++ TC_MQ_GRAFT = 3, ++}; ++ ++struct tc_mq_opt_offload_graft_params { ++ long unsigned int queue; ++ u32 child_handle; ++}; ++ ++struct tc_mq_qopt_offload { ++ enum tc_mq_command command; ++ u32 handle; ++ union { ++ struct tc_qopt_offload_stats stats; ++ struct tc_mq_opt_offload_graft_params graft_params; ++ }; ++}; ++ ++struct mq_sched { ++ struct Qdisc **qdiscs; ++}; ++ ++enum tc_link_layer { ++ TC_LINKLAYER_UNAWARE = 0, ++ TC_LINKLAYER_ETHERNET = 1, ++ TC_LINKLAYER_ATM = 2, ++}; ++ ++enum { ++ TCA_STAB_UNSPEC = 0, ++ TCA_STAB_BASE = 1, ++ TCA_STAB_DATA = 2, ++ __TCA_STAB_MAX = 3, ++}; ++ ++struct qdisc_rate_table { ++ struct tc_ratespec rate; ++ u32 data[256]; ++ struct qdisc_rate_table *next; ++ int refcnt; ++}; ++ ++struct Qdisc_class_common { ++ u32 classid; ++ struct hlist_node hnode; ++}; ++ ++struct Qdisc_class_hash { ++ struct hlist_head *hash; ++ unsigned int hashsize; ++ unsigned int hashmask; ++ unsigned int hashelems; ++}; ++ ++struct qdisc_watchdog { ++ struct hrtimer timer; ++ struct Qdisc *qdisc; ++}; ++ ++struct tc_query_caps_base { ++ enum tc_setup_type type; ++ void *caps; ++}; ++ ++enum tc_root_command { ++ TC_ROOT_GRAFT = 0, ++}; ++ ++struct tc_root_qopt_offload { ++ enum tc_root_command command; ++ u32 handle; ++ bool ingress; ++}; ++ ++struct check_loop_arg { ++ struct qdisc_walker w; ++ struct Qdisc *p; ++ int depth; ++}; ++ ++struct tcf_bind_args { ++ struct tcf_walker w; ++ long unsigned int base; ++ long unsigned int cl; ++ u32 classid; ++}; ++ ++struct tc_bind_class_args { ++ struct qdisc_walker w; ++ long unsigned int new_cl; ++ u32 portid; ++ u32 clid; ++}; ++ ++struct qdisc_dump_args { ++ struct qdisc_walker w; ++ struct sk_buff *skb; ++ struct netlink_callback *cb; ++}; ++ ++struct tc_fifo_qopt { ++ __u32 limit; ++}; ++ ++enum tc_fifo_command { ++ TC_FIFO_REPLACE = 0, ++ TC_FIFO_DESTROY = 1, ++ TC_FIFO_STATS = 2, ++}; ++ ++struct tc_fifo_qopt_offload { ++ enum tc_fifo_command command; ++ u32 handle; ++ u32 parent; ++ union { ++ struct tc_qopt_offload_stats stats; ++ }; ++}; ++ ++enum { ++ TCA_CGROUP_UNSPEC = 0, ++ TCA_CGROUP_ACT = 1, ++ TCA_CGROUP_POLICE = 2, ++ TCA_CGROUP_EMATCHES = 3, ++ __TCA_CGROUP_MAX = 4, ++}; ++ ++struct cls_cgroup_head { ++ u32 handle; ++ struct tcf_exts exts; ++ struct tcf_ematch_tree ematches; ++ struct tcf_proto *tp; ++ struct rcu_work rwork; ++}; ++ ++enum netlink_attribute_type { ++ NL_ATTR_TYPE_INVALID = 0, ++ NL_ATTR_TYPE_FLAG = 1, ++ NL_ATTR_TYPE_U8 = 2, ++ NL_ATTR_TYPE_U16 = 3, ++ NL_ATTR_TYPE_U32 = 4, ++ NL_ATTR_TYPE_U64 = 5, ++ NL_ATTR_TYPE_S8 = 6, ++ NL_ATTR_TYPE_S16 = 7, ++ NL_ATTR_TYPE_S32 = 8, ++ NL_ATTR_TYPE_S64 = 9, ++ NL_ATTR_TYPE_BINARY = 10, ++ NL_ATTR_TYPE_STRING = 11, ++ NL_ATTR_TYPE_NUL_STRING = 12, ++ NL_ATTR_TYPE_NESTED = 13, ++ NL_ATTR_TYPE_NESTED_ARRAY = 14, ++ NL_ATTR_TYPE_BITFIELD32 = 15, ++}; ++ ++enum netlink_policy_type_attr { ++ NL_POLICY_TYPE_ATTR_UNSPEC = 0, ++ NL_POLICY_TYPE_ATTR_TYPE = 1, ++ NL_POLICY_TYPE_ATTR_MIN_VALUE_S = 2, ++ NL_POLICY_TYPE_ATTR_MAX_VALUE_S = 3, ++ NL_POLICY_TYPE_ATTR_MIN_VALUE_U = 4, ++ NL_POLICY_TYPE_ATTR_MAX_VALUE_U = 5, ++ NL_POLICY_TYPE_ATTR_MIN_LENGTH = 6, ++ NL_POLICY_TYPE_ATTR_MAX_LENGTH = 7, ++ NL_POLICY_TYPE_ATTR_POLICY_IDX = 8, ++ NL_POLICY_TYPE_ATTR_POLICY_MAXTYPE = 9, ++ NL_POLICY_TYPE_ATTR_BITFIELD32_MASK = 10, ++ NL_POLICY_TYPE_ATTR_PAD = 11, ++ NL_POLICY_TYPE_ATTR_MASK = 12, ++ __NL_POLICY_TYPE_ATTR_MAX = 13, ++ NL_POLICY_TYPE_ATTR_MAX = 12, ++}; ++ ++struct netlink_policy_dump_state { ++ unsigned int policy_idx; ++ unsigned int attr_idx; ++ unsigned int n_alloc; ++ struct { ++ const struct nla_policy *policy; ++ unsigned int maxtype; ++ } policies[0]; ++}; ++ ++struct netlink_diag_req { ++ __u8 sdiag_family; ++ __u8 sdiag_protocol; ++ __u16 pad; ++ __u32 ndiag_ino; ++ __u32 ndiag_show; ++ __u32 ndiag_cookie[2]; ++}; ++ ++struct netlink_diag_msg { ++ __u8 ndiag_family; ++ __u8 ndiag_type; ++ __u8 ndiag_protocol; ++ __u8 ndiag_state; ++ __u32 ndiag_portid; ++ __u32 ndiag_dst_portid; ++ __u32 ndiag_dst_group; ++ __u32 ndiag_ino; ++ __u32 ndiag_cookie[2]; ++}; ++ ++enum { ++ NETLINK_DIAG_MEMINFO = 0, ++ NETLINK_DIAG_GROUPS = 1, ++ NETLINK_DIAG_RX_RING = 2, ++ NETLINK_DIAG_TX_RING = 3, ++ NETLINK_DIAG_FLAGS = 4, ++ __NETLINK_DIAG_MAX = 5, ++}; ++ ++struct trace_event_raw_bpf_test_finish { ++ struct trace_entry ent; ++ int err; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_bpf_test_finish {}; ++ ++typedef void (*btf_trace_bpf_test_finish)(void *, int *); ++ ++struct bpf_test_timer { ++ enum { ++ NO_PREEMPT = 0, ++ NO_MIGRATE = 1, ++ } mode; ++ u32 i; ++ u64 time_start; ++ u64 time_spent; ++}; ++ ++struct xdp_page_head { ++ struct xdp_buff orig_ctx; ++ struct xdp_buff ctx; ++ union { ++ struct { ++ struct {} __empty_frame; ++ struct xdp_frame frame[0]; ++ }; ++ struct { ++ struct {} __empty_data; ++ u8 data[0]; ++ }; ++ }; ++}; ++ ++struct xdp_test_data { ++ struct xdp_buff *orig_ctx; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ struct xdp_rxq_info rxq; ++ struct net_device *dev; ++ struct page_pool *pp; ++ struct xdp_frame **frames; ++ struct sk_buff **skbs; ++ struct xdp_mem_info mem; ++ u32 batch_size; ++ u32 frame_cnt; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++ long: 64; ++}; ++ ++struct bpf_fentry_test_t { ++ struct bpf_fentry_test_t *a; ++}; ++ ++struct prog_test_member1 { ++ int a; ++}; ++ ++struct prog_test_member { ++ struct prog_test_member1 m; ++ int c; ++}; ++ ++struct prog_test_ref_kfunc { ++ int a; ++ int b; ++ struct prog_test_member memb; ++ struct prog_test_ref_kfunc *next; ++ refcount_t cnt; ++}; ++ ++struct prog_test_pass1 { ++ int x0; ++ struct { ++ int x1; ++ struct { ++ int x2; ++ struct { ++ int x3; ++ }; ++ }; ++ }; ++}; ++ ++struct prog_test_pass2 { ++ int len; ++ short int arr1[4]; ++ struct { ++ char arr2[4]; ++ long unsigned int arr3[8]; ++ } x; ++}; ++ ++struct prog_test_fail1 { ++ void *p; ++ int x; ++}; ++ ++struct prog_test_fail2 { ++ int x8; ++ struct prog_test_pass1 x; ++}; ++ ++struct prog_test_fail3 { ++ int len; ++ char arr1[2]; ++ char arr2[0]; ++}; ++ ++struct bpf_raw_tp_test_run_info { ++ struct bpf_prog *prog; ++ void *ctx; ++ u32 retval; ++}; ++ ++enum { ++ ETHTOOL_A_STRING_UNSPEC = 0, ++ ETHTOOL_A_STRING_INDEX = 1, ++ ETHTOOL_A_STRING_VALUE = 2, ++ __ETHTOOL_A_STRING_CNT = 3, ++ ETHTOOL_A_STRING_MAX = 2, ++}; ++ ++enum { ++ ETHTOOL_A_STRINGS_UNSPEC = 0, ++ ETHTOOL_A_STRINGS_STRING = 1, ++ __ETHTOOL_A_STRINGS_CNT = 2, ++ ETHTOOL_A_STRINGS_MAX = 1, ++}; ++ ++enum { ++ ETHTOOL_A_STRINGSET_UNSPEC = 0, ++ ETHTOOL_A_STRINGSET_ID = 1, ++ ETHTOOL_A_STRINGSET_COUNT = 2, ++ ETHTOOL_A_STRINGSET_STRINGS = 3, ++ __ETHTOOL_A_STRINGSET_CNT = 4, ++ ETHTOOL_A_STRINGSET_MAX = 3, ++}; ++ ++enum { ++ ETHTOOL_A_STRINGSETS_UNSPEC = 0, ++ ETHTOOL_A_STRINGSETS_STRINGSET = 1, ++ __ETHTOOL_A_STRINGSETS_CNT = 2, ++ ETHTOOL_A_STRINGSETS_MAX = 1, ++}; ++ ++struct strset_info { ++ bool per_dev; ++ bool free_strings; ++ unsigned int count; ++ const char (*strings)[32]; ++}; ++ ++struct strset_req_info { ++ struct ethnl_req_info base; ++ u32 req_ids; ++ bool counts_only; ++}; ++ ++struct strset_reply_data { ++ struct ethnl_reply_data base; ++ struct strset_info sets[21]; ++}; ++ ++struct rss_req_info { ++ struct ethnl_req_info base; ++ u32 rss_context; ++}; ++ ++struct rss_reply_data { ++ struct ethnl_reply_data base; ++ u32 indir_size; ++ u32 hkey_size; ++ u32 hfunc; ++ u32 *indir_table; ++ u8 *hkey; ++}; ++ ++struct debug_reply_data { ++ struct ethnl_reply_data base; ++ u32 msg_mask; ++}; ++ ++struct privflags_reply_data { ++ struct ethnl_reply_data base; ++ const char (*priv_flag_names)[32]; ++ unsigned int n_priv_flags; ++ u32 priv_flags; ++}; ++ ++struct coalesce_reply_data { ++ struct ethnl_reply_data base; ++ struct ethtool_coalesce coalesce; ++ struct kernel_ethtool_coalesce kernel_coalesce; ++ u32 supported_params; ++}; ++ ++struct tsinfo_reply_data { ++ struct ethnl_reply_data base; ++ struct ethtool_ts_info ts_info; ++}; ++ ++enum { ++ ETHTOOL_A_FEC_STAT_UNSPEC = 0, ++ ETHTOOL_A_FEC_STAT_PAD = 1, ++ ETHTOOL_A_FEC_STAT_CORRECTED = 2, ++ ETHTOOL_A_FEC_STAT_UNCORR = 3, ++ ETHTOOL_A_FEC_STAT_CORR_BITS = 4, ++ __ETHTOOL_A_FEC_STAT_CNT = 5, ++ ETHTOOL_A_FEC_STAT_MAX = 4, ++}; ++ ++struct fec_stat_grp { ++ u64 stats[9]; ++ u8 cnt; ++}; ++ ++struct fec_reply_data { ++ struct ethnl_reply_data base; ++ long unsigned int fec_link_modes[2]; ++ u32 active_fec; ++ u8 fec_auto; ++ struct fec_stat_grp corr; ++ struct fec_stat_grp uncorr; ++ struct fec_stat_grp corr_bits; ++}; ++ ++enum { ++ ETHTOOL_A_MM_STAT_UNSPEC = 0, ++ ETHTOOL_A_MM_STAT_PAD = 1, ++ ETHTOOL_A_MM_STAT_REASSEMBLY_ERRORS = 2, ++ ETHTOOL_A_MM_STAT_SMD_ERRORS = 3, ++ ETHTOOL_A_MM_STAT_REASSEMBLY_OK = 4, ++ ETHTOOL_A_MM_STAT_RX_FRAG_COUNT = 5, ++ ETHTOOL_A_MM_STAT_TX_FRAG_COUNT = 6, ++ ETHTOOL_A_MM_STAT_HOLD_COUNT = 7, ++ __ETHTOOL_A_MM_STAT_CNT = 8, ++ ETHTOOL_A_MM_STAT_MAX = 7, ++}; ++ ++struct mm_reply_data { ++ struct ethnl_reply_data base; ++ struct ethtool_mm_state state; ++ struct ethtool_mm_stats stats; ++}; ++ ++enum ethtool_podl_pse_admin_state { ++ ETHTOOL_PODL_PSE_ADMIN_STATE_UNKNOWN = 1, ++ ETHTOOL_PODL_PSE_ADMIN_STATE_DISABLED = 2, ++ ETHTOOL_PODL_PSE_ADMIN_STATE_ENABLED = 3, ++}; ++ ++enum ethtool_podl_pse_pw_d_status { ++ ETHTOOL_PODL_PSE_PW_D_STATUS_UNKNOWN = 1, ++ ETHTOOL_PODL_PSE_PW_D_STATUS_DISABLED = 2, ++ ETHTOOL_PODL_PSE_PW_D_STATUS_SEARCHING = 3, ++ ETHTOOL_PODL_PSE_PW_D_STATUS_DELIVERING = 4, ++ ETHTOOL_PODL_PSE_PW_D_STATUS_SLEEP = 5, ++ ETHTOOL_PODL_PSE_PW_D_STATUS_IDLE = 6, ++ ETHTOOL_PODL_PSE_PW_D_STATUS_ERROR = 7, ++}; ++ ++struct pse_control_config { ++ enum ethtool_podl_pse_admin_state admin_cotrol; ++}; ++ ++struct pse_control_status { ++ enum ethtool_podl_pse_admin_state podl_admin_state; ++ enum ethtool_podl_pse_pw_d_status podl_pw_status; ++}; ++ ++struct pse_reply_data { ++ struct ethnl_reply_data base; ++ struct pse_control_status status; ++}; ++ ++struct nf_loginfo { ++ u_int8_t type; ++ union { ++ struct { ++ u_int32_t copy_len; ++ u_int16_t group; ++ u_int16_t qthreshold; ++ u_int16_t flags; ++ } ulog; ++ struct { ++ u_int8_t level; ++ u_int8_t logflags; ++ } log; ++ } u; ++}; ++ ++struct nf_log_buf { ++ unsigned int count; ++ char buf[1020]; ++}; ++ ++struct rt_cache_stat { ++ unsigned int in_slow_tot; ++ unsigned int in_slow_mc; ++ unsigned int in_no_route; ++ unsigned int in_brd; ++ unsigned int in_martian_dst; ++ unsigned int in_martian_src; ++ unsigned int out_slow_tot; ++ unsigned int out_slow_mc; ++}; ++ ++struct ip_mreq_source { ++ __be32 imr_multiaddr; ++ __be32 imr_interface; ++ __be32 imr_sourceaddr; ++}; ++ ++struct ip_msfilter { ++ __be32 imsf_multiaddr; ++ __be32 imsf_interface; ++ __u32 imsf_fmode; ++ __u32 imsf_numsrc; ++ union { ++ __be32 imsf_slist[1]; ++ struct { ++ struct {} __empty_imsf_slist_flex; ++ __be32 imsf_slist_flex[0]; ++ }; ++ }; ++}; ++ ++struct icmp_filter { ++ __u32 data; ++}; ++ ++struct raw_sock { ++ struct inet_sock inet; ++ struct icmp_filter filter; ++ u32 ipmr_table; ++}; ++ ++struct raw_frag_vec { ++ struct msghdr *msg; ++ union { ++ struct icmphdr icmph; ++ char c[1]; ++ } hdr; ++ int hlen; ++}; ++ ++struct arpreq { ++ struct sockaddr arp_pa; ++ struct sockaddr arp_ha; ++ int arp_flags; ++ struct sockaddr arp_netmask; ++ char arp_dev[16]; ++}; ++ ++enum { ++ AX25_VALUES_IPDEFMODE = 0, ++ AX25_VALUES_AXDEFMODE = 1, ++ AX25_VALUES_BACKOFF = 2, ++ AX25_VALUES_CONMODE = 3, ++ AX25_VALUES_WINDOW = 4, ++ AX25_VALUES_EWINDOW = 5, ++ AX25_VALUES_T1 = 6, ++ AX25_VALUES_T2 = 7, ++ AX25_VALUES_T3 = 8, ++ AX25_VALUES_IDLE = 9, ++ AX25_VALUES_N2 = 10, ++ AX25_VALUES_PACLEN = 11, ++ AX25_VALUES_PROTOCOL = 12, ++ AX25_VALUES_DS_TIMEOUT = 13, ++ AX25_MAX_VALUES = 14, ++}; ++ ++struct igmphdr { ++ __u8 type; ++ __u8 code; ++ __sum16 csum; ++ __be32 group; ++}; ++ ++struct igmpv3_grec { ++ __u8 grec_type; ++ __u8 grec_auxwords; ++ __be16 grec_nsrcs; ++ __be32 grec_mca; ++ __be32 grec_src[0]; ++}; ++ ++struct igmpv3_report { ++ __u8 type; ++ __u8 resv1; ++ __sum16 csum; ++ __be16 resv2; ++ __be16 ngrec; ++ struct igmpv3_grec grec[0]; ++}; ++ ++struct igmpv3_query { ++ __u8 type; ++ __u8 code; ++ __sum16 csum; ++ __be32 group; ++ __u8 resv: 4; ++ __u8 suppress: 1; ++ __u8 qrv: 3; ++ __u8 qqic; ++ __be16 nsrcs; ++ __be32 srcs[0]; ++}; ++ ++struct igmp_mc_iter_state { ++ struct seq_net_private p; ++ struct net_device *dev; ++ struct in_device *in_dev; ++}; ++ ++struct igmp_mcf_iter_state { ++ struct seq_net_private p; ++ struct net_device *dev; ++ struct in_device *idev; ++ struct ip_mc_list *im; ++}; ++ ++struct fib_entry_notifier_info { ++ struct fib_notifier_info info; ++ u32 dst; ++ int dst_len; ++ struct fib_info *fi; ++ dscp_t dscp; ++ u8 type; ++ u32 tb_id; ++}; ++ ++typedef unsigned int t_key; ++ ++struct key_vector { ++ t_key key; ++ unsigned char pos; ++ unsigned char bits; ++ unsigned char slen; ++ union { ++ struct hlist_head leaf; ++ struct { ++ struct {} __empty_tnode; ++ struct key_vector *tnode[0]; ++ }; ++ }; ++}; ++ ++struct tnode { ++ struct callback_head rcu; ++ t_key empty_children; ++ t_key full_children; ++ struct key_vector *parent; ++ struct key_vector kv[1]; ++}; ++ ++struct trie_use_stats { ++ unsigned int gets; ++ unsigned int backtrack; ++ unsigned int semantic_match_passed; ++ unsigned int semantic_match_miss; ++ unsigned int null_node_hit; ++ unsigned int resize_node_skipped; ++}; ++ ++struct trie_stat { ++ unsigned int totdepth; ++ unsigned int maxdepth; ++ unsigned int tnodes; ++ unsigned int leaves; ++ unsigned int nullpointers; ++ unsigned int prefixes; ++ unsigned int nodesizes[32]; ++}; ++ ++struct trie { ++ struct key_vector kv[1]; ++ struct trie_use_stats *stats; ++}; ++ ++struct fib_trie_iter { ++ struct seq_net_private p; ++ struct fib_table *tb; ++ struct key_vector *tnode; ++ unsigned int index; ++ unsigned int depth; ++}; ++ ++struct fib_route_iter { ++ struct seq_net_private p; ++ struct fib_table *main_tb; ++ struct key_vector *tnode; ++ loff_t pos; ++ t_key key; ++}; ++ ++struct nhmsg { ++ unsigned char nh_family; ++ unsigned char nh_scope; ++ unsigned char nh_protocol; ++ unsigned char resvd; ++ unsigned int nh_flags; ++}; ++ ++struct nexthop_grp { ++ __u32 id; ++ __u8 weight; ++ __u8 resvd1; ++ __u16 resvd2; ++}; ++ ++enum { ++ NEXTHOP_GRP_TYPE_MPATH = 0, ++ NEXTHOP_GRP_TYPE_RES = 1, ++ __NEXTHOP_GRP_TYPE_MAX = 2, ++}; ++ ++enum { ++ NHA_UNSPEC = 0, ++ NHA_ID = 1, ++ NHA_GROUP = 2, ++ NHA_GROUP_TYPE = 3, ++ NHA_BLACKHOLE = 4, ++ NHA_OIF = 5, ++ NHA_GATEWAY = 6, ++ NHA_ENCAP_TYPE = 7, ++ NHA_ENCAP = 8, ++ NHA_GROUPS = 9, ++ NHA_MASTER = 10, ++ NHA_FDB = 11, ++ NHA_RES_GROUP = 12, ++ NHA_RES_BUCKET = 13, ++ __NHA_MAX = 14, ++}; ++ ++enum { ++ NHA_RES_GROUP_UNSPEC = 0, ++ NHA_RES_GROUP_PAD = 0, ++ NHA_RES_GROUP_BUCKETS = 1, ++ NHA_RES_GROUP_IDLE_TIMER = 2, ++ NHA_RES_GROUP_UNBALANCED_TIMER = 3, ++ NHA_RES_GROUP_UNBALANCED_TIME = 4, ++ __NHA_RES_GROUP_MAX = 5, ++}; ++ ++enum { ++ NHA_RES_BUCKET_UNSPEC = 0, ++ NHA_RES_BUCKET_PAD = 0, ++ NHA_RES_BUCKET_INDEX = 1, ++ NHA_RES_BUCKET_IDLE_TIME = 2, ++ NHA_RES_BUCKET_NH_ID = 3, ++ __NHA_RES_BUCKET_MAX = 4, ++}; ++ ++struct nh_config { ++ u32 nh_id; ++ u8 nh_family; ++ u8 nh_protocol; ++ u8 nh_blackhole; ++ u8 nh_fdb; ++ u32 nh_flags; ++ int nh_ifindex; ++ struct net_device *dev; ++ union { ++ __be32 ipv4; ++ struct in6_addr ipv6; ++ } gw; ++ struct nlattr *nh_grp; ++ u16 nh_grp_type; ++ u16 nh_grp_res_num_buckets; ++ long unsigned int nh_grp_res_idle_timer; ++ long unsigned int nh_grp_res_unbalanced_timer; ++ bool nh_grp_res_has_num_buckets; ++ bool nh_grp_res_has_idle_timer; ++ bool nh_grp_res_has_unbalanced_timer; ++ struct nlattr *nh_encap; ++ u16 nh_encap_type; ++ u32 nlflags; ++ struct nl_info nlinfo; ++}; ++ ++enum nexthop_event_type { ++ NEXTHOP_EVENT_DEL = 0, ++ NEXTHOP_EVENT_REPLACE = 1, ++ NEXTHOP_EVENT_RES_TABLE_PRE_REPLACE = 2, ++ NEXTHOP_EVENT_BUCKET_REPLACE = 3, ++}; ++ ++enum nh_notifier_info_type { ++ NH_NOTIFIER_INFO_TYPE_SINGLE = 0, ++ NH_NOTIFIER_INFO_TYPE_GRP = 1, ++ NH_NOTIFIER_INFO_TYPE_RES_TABLE = 2, ++ NH_NOTIFIER_INFO_TYPE_RES_BUCKET = 3, ++}; ++ ++struct nh_notifier_single_info { ++ struct net_device *dev; ++ u8 gw_family; ++ union { ++ __be32 ipv4; ++ struct in6_addr ipv6; ++ }; ++ u8 is_reject: 1; ++ u8 is_fdb: 1; ++ u8 has_encap: 1; ++}; ++ ++struct nh_notifier_grp_entry_info { ++ u8 weight; ++ u32 id; ++ struct nh_notifier_single_info nh; ++}; ++ ++struct nh_notifier_grp_info { ++ u16 num_nh; ++ bool is_fdb; ++ struct nh_notifier_grp_entry_info nh_entries[0]; ++}; ++ ++struct nh_notifier_res_bucket_info { ++ u16 bucket_index; ++ unsigned int idle_timer_ms; ++ bool force; ++ struct nh_notifier_single_info old_nh; ++ struct nh_notifier_single_info new_nh; ++}; ++ ++struct nh_notifier_res_table_info { ++ u16 num_nh_buckets; ++ struct nh_notifier_single_info nhs[0]; ++}; ++ ++struct nh_notifier_info { ++ struct net *net; ++ struct netlink_ext_ack *extack; ++ u32 id; ++ enum nh_notifier_info_type type; ++ union { ++ struct nh_notifier_single_info *nh; ++ struct nh_notifier_grp_info *nh_grp; ++ struct nh_notifier_res_table_info *nh_res_table; ++ struct nh_notifier_res_bucket_info *nh_res_bucket; ++ }; ++}; ++ ++struct nh_dump_filter { ++ u32 nh_id; ++ int dev_idx; ++ int master_idx; ++ bool group_filter; ++ bool fdb_filter; ++ u32 res_bucket_nh_id; ++}; ++ ++struct rtm_dump_nh_ctx { ++ u32 idx; ++}; ++ ++struct rtm_dump_res_bucket_ctx { ++ struct rtm_dump_nh_ctx nh; ++ u16 bucket_index; ++ u32 done_nh_idx; ++}; ++ ++struct rtm_dump_nexthop_bucket_data { ++ struct rtm_dump_res_bucket_ctx *ctx; ++ struct nh_dump_filter filter; ++}; ++ ++typedef short unsigned int vifi_t; ++ ++struct vifctl { ++ vifi_t vifc_vifi; ++ unsigned char vifc_flags; ++ unsigned char vifc_threshold; ++ unsigned int vifc_rate_limit; ++ union { ++ struct in_addr vifc_lcl_addr; ++ int vifc_lcl_ifindex; ++ }; ++ struct in_addr vifc_rmt_addr; ++}; ++ ++struct mfcctl { ++ struct in_addr mfcc_origin; ++ struct in_addr mfcc_mcastgrp; ++ vifi_t mfcc_parent; ++ unsigned char mfcc_ttls[32]; ++ unsigned int mfcc_pkt_cnt; ++ unsigned int mfcc_byte_cnt; ++ unsigned int mfcc_wrong_if; ++ int mfcc_expire; ++}; ++ ++struct sioc_sg_req { ++ struct in_addr src; ++ struct in_addr grp; ++ long unsigned int pktcnt; ++ long unsigned int bytecnt; ++ long unsigned int wrong_if; ++}; ++ ++struct sioc_vif_req { ++ vifi_t vifi; ++ long unsigned int icount; ++ long unsigned int ocount; ++ long unsigned int ibytes; ++ long unsigned int obytes; ++}; ++ ++struct igmpmsg { ++ __u32 unused1; ++ __u32 unused2; ++ unsigned char im_msgtype; ++ unsigned char im_mbz; ++ unsigned char im_vif; ++ unsigned char im_vif_hi; ++ struct in_addr im_src; ++ struct in_addr im_dst; ++}; ++ ++enum { ++ IPMRA_TABLE_UNSPEC = 0, ++ IPMRA_TABLE_ID = 1, ++ IPMRA_TABLE_CACHE_RES_QUEUE_LEN = 2, ++ IPMRA_TABLE_MROUTE_REG_VIF_NUM = 3, ++ IPMRA_TABLE_MROUTE_DO_ASSERT = 4, ++ IPMRA_TABLE_MROUTE_DO_PIM = 5, ++ IPMRA_TABLE_VIFS = 6, ++ IPMRA_TABLE_MROUTE_DO_WRVIFWHOLE = 7, ++ __IPMRA_TABLE_MAX = 8, ++}; ++ ++enum { ++ IPMRA_VIF_UNSPEC = 0, ++ IPMRA_VIF = 1, ++ __IPMRA_VIF_MAX = 2, ++}; ++ ++enum { ++ IPMRA_VIFA_UNSPEC = 0, ++ IPMRA_VIFA_IFINDEX = 1, ++ IPMRA_VIFA_VIF_ID = 2, ++ IPMRA_VIFA_FLAGS = 3, ++ IPMRA_VIFA_BYTES_IN = 4, ++ IPMRA_VIFA_BYTES_OUT = 5, ++ IPMRA_VIFA_PACKETS_IN = 6, ++ IPMRA_VIFA_PACKETS_OUT = 7, ++ IPMRA_VIFA_LOCAL_ADDR = 8, ++ IPMRA_VIFA_REMOTE_ADDR = 9, ++ IPMRA_VIFA_PAD = 10, ++ __IPMRA_VIFA_MAX = 11, ++}; ++ ++enum { ++ IPMRA_CREPORT_UNSPEC = 0, ++ IPMRA_CREPORT_MSGTYPE = 1, ++ IPMRA_CREPORT_VIF_ID = 2, ++ IPMRA_CREPORT_SRC_ADDR = 3, ++ IPMRA_CREPORT_DST_ADDR = 4, ++ IPMRA_CREPORT_PKT = 5, ++ IPMRA_CREPORT_TABLE = 6, ++ __IPMRA_CREPORT_MAX = 7, ++}; ++ ++struct mfc_cache_cmp_arg { ++ __be32 mfc_mcastgrp; ++ __be32 mfc_origin; ++}; ++ ++struct mfc_cache { ++ struct mr_mfc _c; ++ union { ++ struct { ++ __be32 mfc_mcastgrp; ++ __be32 mfc_origin; ++ }; ++ struct mfc_cache_cmp_arg cmparg; ++ }; ++}; ++ ++struct ipmr_result { ++ struct mr_table *mrt; ++}; ++ ++struct compat_sioc_sg_req { ++ struct in_addr src; ++ struct in_addr grp; ++ compat_ulong_t pktcnt; ++ compat_ulong_t bytecnt; ++ compat_ulong_t wrong_if; ++}; ++ ++struct compat_sioc_vif_req { ++ vifi_t vifi; ++ compat_ulong_t icount; ++ compat_ulong_t ocount; ++ compat_ulong_t ibytes; ++ compat_ulong_t obytes; ++}; ++ ++struct inet_diag_req { ++ __u8 idiag_family; ++ __u8 idiag_src_len; ++ __u8 idiag_dst_len; ++ __u8 idiag_ext; ++ struct inet_diag_sockid id; ++ __u32 idiag_states; ++ __u32 idiag_dbs; ++}; ++ ++struct inet_diag_bc_op { ++ unsigned char code; ++ unsigned char yes; ++ short unsigned int no; ++}; ++ ++enum { ++ INET_DIAG_BC_NOP = 0, ++ INET_DIAG_BC_JMP = 1, ++ INET_DIAG_BC_S_GE = 2, ++ INET_DIAG_BC_S_LE = 3, ++ INET_DIAG_BC_D_GE = 4, ++ INET_DIAG_BC_D_LE = 5, ++ INET_DIAG_BC_AUTO = 6, ++ INET_DIAG_BC_S_COND = 7, ++ INET_DIAG_BC_D_COND = 8, ++ INET_DIAG_BC_DEV_COND = 9, ++ INET_DIAG_BC_MARK_COND = 10, ++ INET_DIAG_BC_S_EQ = 11, ++ INET_DIAG_BC_D_EQ = 12, ++ INET_DIAG_BC_CGROUP_COND = 13, ++}; ++ ++struct inet_diag_hostcond { ++ __u8 family; ++ __u8 prefix_len; ++ int port; ++ __be32 addr[0]; ++}; ++ ++struct inet_diag_markcond { ++ __u32 mark; ++ __u32 mask; ++}; ++ ++struct inet_diag_meminfo { ++ __u32 idiag_rmem; ++ __u32 idiag_wmem; ++ __u32 idiag_fmem; ++ __u32 idiag_tmem; ++}; ++ ++struct inet_diag_sockopt { ++ __u8 recverr: 1; ++ __u8 is_icsk: 1; ++ __u8 freebind: 1; ++ __u8 hdrincl: 1; ++ __u8 mc_loop: 1; ++ __u8 transparent: 1; ++ __u8 mc_all: 1; ++ __u8 nodefrag: 1; ++ __u8 bind_address_no_port: 1; ++ __u8 recverr_rfc4884: 1; ++ __u8 defer_connect: 1; ++ __u8 unused: 5; ++}; ++ ++struct inet_diag_entry { ++ const __be32 *saddr; ++ const __be32 *daddr; ++ u16 sport; ++ u16 dport; ++ u16 family; ++ u16 userlocks; ++ u32 ifindex; ++ u32 mark; ++ u64 cgroup_id; ++}; ++ ++struct bictcp { ++ u32 cnt; ++ u32 last_max_cwnd; ++ u32 last_cwnd; ++ u32 last_time; ++ u32 bic_origin_point; ++ u32 bic_K; ++ u32 delay_min; ++ u32 epoch_start; ++ u32 ack_cnt; ++ u32 tcp_cwnd; ++ u16 unused; ++ u8 sample_cnt; ++ u8 found; ++ u32 round_start; ++ u32 end_seq; ++ u32 last_ack; ++ u32 curr_rtt; ++}; ++ ++struct cipso_v4_map_cache_bkt { ++ spinlock_t lock; ++ u32 size; ++ struct list_head list; ++}; ++ ++struct cipso_v4_map_cache_entry { ++ u32 hash; ++ unsigned char *key; ++ size_t key_len; ++ struct netlbl_lsm_cache *lsm_data; ++ u32 activity; ++ struct list_head list; ++}; ++ ++enum { ++ XFRM_DEV_OFFLOAD_FLAG_ACQ = 1, ++}; ++ ++struct espintcp_msg { ++ struct sk_buff *skb; ++ struct sk_msg skmsg; ++ int offset; ++ int len; ++}; ++ ++struct espintcp_ctx { ++ struct strparser strp; ++ struct sk_buff_head ike_queue; ++ struct sk_buff_head out_queue; ++ struct espintcp_msg partial; ++ void (*saved_data_ready)(struct sock *); ++ void (*saved_write_space)(struct sock *); ++ void (*saved_destruct)(struct sock *); ++ struct work_struct work; ++ bool tx_running; ++}; ++ ++struct hop_jumbo_hdr { ++ u8 nexthdr; ++ u8 hdrlen; ++ u8 tlv_type; ++ u8 tlv_len; ++ __be32 jumbo_payload_len; ++}; ++ ++enum fib6_walk_state { ++ FWS_S = 0, ++ FWS_L = 1, ++ FWS_R = 2, ++ FWS_C = 3, ++ FWS_U = 4, ++}; ++ ++struct fib6_walker { ++ struct list_head lh; ++ struct fib6_node *root; ++ struct fib6_node *node; ++ struct fib6_info *leaf; ++ enum fib6_walk_state state; ++ unsigned int skip; ++ unsigned int count; ++ unsigned int skip_in_node; ++ int (*func)(struct fib6_walker *); ++ void *args; ++}; ++ ++struct fib6_entry_notifier_info { ++ struct fib_notifier_info info; ++ struct fib6_info *rt; ++ unsigned int nsiblings; ++}; ++ ++struct ipv6_route_iter { ++ struct seq_net_private p; ++ struct fib6_walker w; ++ loff_t skip; ++ struct fib6_table *tbl; ++ int sernum; ++}; ++ ++struct bpf_iter__ipv6_route { ++ union { ++ struct bpf_iter_meta *meta; ++ }; ++ union { ++ struct fib6_info *rt; ++ }; ++}; ++ ++struct fib6_cleaner { ++ struct fib6_walker w; ++ struct net *net; ++ int (*func)(struct fib6_info *, void *); ++ int sernum; ++ void *arg; ++ bool skip_notify; ++}; ++ ++enum { ++ FIB6_NO_SERNUM_CHANGE = 0, ++}; ++ ++struct fib6_dump_arg { ++ struct net *net; ++ struct notifier_block *nb; ++ struct netlink_ext_ack *extack; ++}; ++ ++struct fib6_nh_pcpu_arg { ++ struct fib6_info *from; ++ const struct fib6_table *table; ++}; ++ ++struct lookup_args { ++ int offset; ++ const struct in6_addr *addr; ++}; ++ ++struct nduseroptmsg { ++ unsigned char nduseropt_family; ++ unsigned char nduseropt_pad1; ++ short unsigned int nduseropt_opts_len; ++ int nduseropt_ifindex; ++ __u8 nduseropt_icmp_type; ++ __u8 nduseropt_icmp_code; ++ short unsigned int nduseropt_pad2; ++ unsigned int nduseropt_pad3; ++}; ++ ++enum { ++ NDUSEROPT_UNSPEC = 0, ++ NDUSEROPT_SRCADDR = 1, ++ __NDUSEROPT_MAX = 2, ++}; ++ ++struct nd_msg { ++ struct icmp6hdr icmph; ++ struct in6_addr target; ++ __u8 opt[0]; ++}; ++ ++struct rs_msg { ++ struct icmp6hdr icmph; ++ __u8 opt[0]; ++}; ++ ++struct ra_msg { ++ struct icmp6hdr icmph; ++ __be32 reachable_time; ++ __be32 retrans_timer; ++}; ++ ++typedef void ip6_icmp_send_t(struct sk_buff *, u8, u8, __u32, const struct in6_addr *, const struct inet6_skb_parm *); ++ ++struct icmpv6_msg { ++ struct sk_buff *skb; ++ int offset; ++ uint8_t type; ++}; ++ ++struct icmp6_err { ++ int err; ++ int fatal; ++}; ++ ++struct xfrm6_protocol { ++ int (*handler)(struct sk_buff *); ++ int (*input_handler)(struct sk_buff *, int, __be32, int); ++ int (*cb_handler)(struct sk_buff *, int); ++ int (*err_handler)(struct sk_buff *, struct inet6_skb_parm *, u8, u8, int, __be32); ++ struct xfrm6_protocol *next; ++ int priority; ++}; ++ ++enum { ++ SEG6_IPTUNNEL_UNSPEC = 0, ++ SEG6_IPTUNNEL_SRH = 1, ++ __SEG6_IPTUNNEL_MAX = 2, ++}; ++ ++struct seg6_iptunnel_encap { ++ int mode; ++ struct ipv6_sr_hdr srh[0]; ++}; ++ ++enum { ++ SEG6_IPTUN_MODE_INLINE = 0, ++ SEG6_IPTUN_MODE_ENCAP = 1, ++ SEG6_IPTUN_MODE_L2ENCAP = 2, ++ SEG6_IPTUN_MODE_ENCAP_RED = 3, ++ SEG6_IPTUN_MODE_L2ENCAP_RED = 4, ++}; ++ ++struct seg6_lwt { ++ struct dst_cache cache; ++ struct seg6_iptunnel_encap tuninfo[0]; ++}; ++ ++enum { ++ RPL_IPTUNNEL_UNSPEC = 0, ++ RPL_IPTUNNEL_SRH = 1, ++ __RPL_IPTUNNEL_MAX = 2, ++}; ++ ++struct rpl_iptunnel_encap { ++ struct { ++ struct {} __empty_srh; ++ struct ipv6_rpl_sr_hdr srh[0]; ++ }; ++}; ++ ++struct rpl_lwt { ++ struct dst_cache cache; ++ struct rpl_iptunnel_encap tuninfo; ++}; ++ ++enum { ++ DEVLINK_ATTR_STATS_RX_PACKETS = 0, ++ DEVLINK_ATTR_STATS_RX_BYTES = 1, ++ DEVLINK_ATTR_STATS_RX_DROPPED = 2, ++ __DEVLINK_ATTR_STATS_MAX = 3, ++ DEVLINK_ATTR_STATS_MAX = 2, ++}; ++ ++enum { ++ DEVLINK_ATTR_TRAP_METADATA_TYPE_IN_PORT = 0, ++ DEVLINK_ATTR_TRAP_METADATA_TYPE_FA_COOKIE = 1, ++}; ++ ++enum devlink_linecard_state { ++ DEVLINK_LINECARD_STATE_UNSPEC = 0, ++ DEVLINK_LINECARD_STATE_UNPROVISIONED = 1, ++ DEVLINK_LINECARD_STATE_UNPROVISIONING = 2, ++ DEVLINK_LINECARD_STATE_PROVISIONING = 3, ++ DEVLINK_LINECARD_STATE_PROVISIONING_FAILED = 4, ++ DEVLINK_LINECARD_STATE_PROVISIONED = 5, ++ DEVLINK_LINECARD_STATE_ACTIVE = 6, ++ __DEVLINK_LINECARD_STATE_MAX = 7, ++ DEVLINK_LINECARD_STATE_MAX = 6, ++}; ++ ++enum devlink_dpipe_match_type { ++ DEVLINK_DPIPE_MATCH_TYPE_FIELD_EXACT = 0, ++}; ++ ++enum devlink_dpipe_action_type { ++ DEVLINK_DPIPE_ACTION_TYPE_FIELD_MODIFY = 0, ++}; ++ ++enum devlink_dpipe_field_ethernet_id { ++ DEVLINK_DPIPE_FIELD_ETHERNET_DST_MAC = 0, ++}; ++ ++enum devlink_dpipe_field_ipv4_id { ++ DEVLINK_DPIPE_FIELD_IPV4_DST_IP = 0, ++}; ++ ++enum devlink_dpipe_field_ipv6_id { ++ DEVLINK_DPIPE_FIELD_IPV6_DST_IP = 0, ++}; ++ ++enum devlink_dpipe_header_id { ++ DEVLINK_DPIPE_HEADER_ETHERNET = 0, ++ DEVLINK_DPIPE_HEADER_IPV4 = 1, ++ DEVLINK_DPIPE_HEADER_IPV6 = 2, ++}; ++ ++enum devlink_resource_unit { ++ DEVLINK_RESOURCE_UNIT_ENTRY = 0, ++}; ++ ++enum devlink_port_fn_attr_cap { ++ DEVLINK_PORT_FN_ATTR_CAP_ROCE_BIT = 0, ++ DEVLINK_PORT_FN_ATTR_CAP_MIGRATABLE_BIT = 1, ++ __DEVLINK_PORT_FN_ATTR_CAPS_MAX = 2, ++}; ++ ++enum devlink_port_function_attr { ++ DEVLINK_PORT_FUNCTION_ATTR_UNSPEC = 0, ++ DEVLINK_PORT_FUNCTION_ATTR_HW_ADDR = 1, ++ DEVLINK_PORT_FN_ATTR_STATE = 2, ++ DEVLINK_PORT_FN_ATTR_OPSTATE = 3, ++ DEVLINK_PORT_FN_ATTR_CAPS = 4, ++ __DEVLINK_PORT_FUNCTION_ATTR_MAX = 5, ++ DEVLINK_PORT_FUNCTION_ATTR_MAX = 4, ++}; ++ ++struct devlink_linecard_ops; ++ ++struct devlink_linecard_type; ++ ++struct devlink_linecard { ++ struct list_head list; ++ struct devlink *devlink; ++ unsigned int index; ++ const struct devlink_linecard_ops *ops; ++ void *priv; ++ enum devlink_linecard_state state; ++ struct mutex state_lock; ++ const char *type; ++ struct devlink_linecard_type *types; ++ unsigned int types_count; ++ struct devlink *nested_devlink; ++}; ++ ++struct devlink_linecard_ops { ++ int (*provision)(struct devlink_linecard *, void *, const char *, const void *, struct netlink_ext_ack *); ++ int (*unprovision)(struct devlink_linecard *, void *, struct netlink_ext_ack *); ++ bool (*same_provision)(struct devlink_linecard *, void *, const char *, const void *); ++ unsigned int (*types_count)(struct devlink_linecard *, void *); ++ void (*types_get)(struct devlink_linecard *, void *, unsigned int, const char **, const void **); ++}; ++ ++struct devlink_dpipe_match { ++ enum devlink_dpipe_match_type type; ++ unsigned int header_index; ++ struct devlink_dpipe_header *header; ++ unsigned int field_id; ++}; ++ ++struct devlink_dpipe_action { ++ enum devlink_dpipe_action_type type; ++ unsigned int header_index; ++ struct devlink_dpipe_header *header; ++ unsigned int field_id; ++}; ++ ++struct devlink_dpipe_value { ++ union { ++ struct devlink_dpipe_action *action; ++ struct devlink_dpipe_match *match; ++ }; ++ unsigned int mapping_value; ++ bool mapping_valid; ++ unsigned int value_size; ++ void *value; ++ void *mask; ++}; ++ ++struct devlink_dpipe_entry { ++ u64 index; ++ struct devlink_dpipe_value *match_values; ++ unsigned int match_values_count; ++ struct devlink_dpipe_value *action_values; ++ unsigned int action_values_count; ++ u64 counter; ++ bool counter_valid; ++}; ++ ++struct devlink_dpipe_dump_ctx { ++ struct genl_info *info; ++ enum devlink_command cmd; ++ struct sk_buff *skb; ++ struct nlattr *nest; ++ void *hdr; ++}; ++ ++struct devlink_dpipe_table_ops; ++ ++struct devlink_dpipe_table { ++ void *priv; ++ struct list_head list; ++ const char *name; ++ bool counters_enabled; ++ bool counter_control_extern; ++ bool resource_valid; ++ u64 resource_id; ++ u64 resource_units; ++ struct devlink_dpipe_table_ops *table_ops; ++ struct callback_head rcu; ++}; ++ ++struct devlink_dpipe_table_ops { ++ int (*actions_dump)(void *, struct sk_buff *); ++ int (*matches_dump)(void *, struct sk_buff *); ++ int (*entries_dump)(void *, bool, struct devlink_dpipe_dump_ctx *); ++ int (*counters_set_update)(void *, bool); ++ u64 (*size_get)(void *); ++}; ++ ++struct devlink_resource_size_params { ++ u64 size_min; ++ u64 size_max; ++ u64 size_granularity; ++ enum devlink_resource_unit unit; ++}; ++ ++typedef u64 devlink_resource_occ_get_t(void *); ++ ++enum devlink_param_type { ++ DEVLINK_PARAM_TYPE_U8 = 0, ++ DEVLINK_PARAM_TYPE_U16 = 1, ++ DEVLINK_PARAM_TYPE_U32 = 2, ++ DEVLINK_PARAM_TYPE_STRING = 3, ++ DEVLINK_PARAM_TYPE_BOOL = 4, ++}; ++ ++struct devlink_param { ++ u32 id; ++ const char *name; ++ bool generic; ++ enum devlink_param_type type; ++ long unsigned int supported_cmodes; ++ int (*get)(struct devlink *, u32, struct devlink_param_gset_ctx *); ++ int (*set)(struct devlink *, u32, struct devlink_param_gset_ctx *); ++ int (*validate)(struct devlink *, u32, union devlink_param_value, struct netlink_ext_ack *); ++}; ++ ++struct devlink_param_item { ++ struct list_head list; ++ const struct devlink_param *param; ++ union devlink_param_value driverinit_value; ++ bool driverinit_value_valid; ++ union devlink_param_value driverinit_value_new; ++ bool driverinit_value_new_valid; ++}; ++ ++enum devlink_param_generic_id { ++ DEVLINK_PARAM_GENERIC_ID_INT_ERR_RESET = 0, ++ DEVLINK_PARAM_GENERIC_ID_MAX_MACS = 1, ++ DEVLINK_PARAM_GENERIC_ID_ENABLE_SRIOV = 2, ++ DEVLINK_PARAM_GENERIC_ID_REGION_SNAPSHOT = 3, ++ DEVLINK_PARAM_GENERIC_ID_IGNORE_ARI = 4, ++ DEVLINK_PARAM_GENERIC_ID_MSIX_VEC_PER_PF_MAX = 5, ++ DEVLINK_PARAM_GENERIC_ID_MSIX_VEC_PER_PF_MIN = 6, ++ DEVLINK_PARAM_GENERIC_ID_FW_LOAD_POLICY = 7, ++ DEVLINK_PARAM_GENERIC_ID_RESET_DEV_ON_DRV_PROBE = 8, ++ DEVLINK_PARAM_GENERIC_ID_ENABLE_ROCE = 9, ++ DEVLINK_PARAM_GENERIC_ID_ENABLE_REMOTE_DEV_RESET = 10, ++ DEVLINK_PARAM_GENERIC_ID_ENABLE_ETH = 11, ++ DEVLINK_PARAM_GENERIC_ID_ENABLE_RDMA = 12, ++ DEVLINK_PARAM_GENERIC_ID_ENABLE_VNET = 13, ++ DEVLINK_PARAM_GENERIC_ID_ENABLE_IWARP = 14, ++ DEVLINK_PARAM_GENERIC_ID_IO_EQ_SIZE = 15, ++ DEVLINK_PARAM_GENERIC_ID_EVENT_EQ_SIZE = 16, ++ __DEVLINK_PARAM_GENERIC_ID_MAX = 17, ++ DEVLINK_PARAM_GENERIC_ID_MAX = 16, ++}; ++ ++struct devlink_region_ops { ++ const char *name; ++ void (*destructor)(const void *); ++ int (*snapshot)(struct devlink *, const struct devlink_region_ops *, struct netlink_ext_ack *, u8 **); ++ int (*read)(struct devlink *, const struct devlink_region_ops *, struct netlink_ext_ack *, u64, u32, u8 *); ++ void *priv; ++}; ++ ++struct devlink_port_region_ops { ++ const char *name; ++ void (*destructor)(const void *); ++ int (*snapshot)(struct devlink_port *, const struct devlink_port_region_ops *, struct netlink_ext_ack *, u8 **); ++ int (*read)(struct devlink_port *, const struct devlink_port_region_ops *, struct netlink_ext_ack *, u64, u32, u8 *); ++ void *priv; ++}; ++ ++enum devlink_trap_generic_id { ++ DEVLINK_TRAP_GENERIC_ID_SMAC_MC = 0, ++ DEVLINK_TRAP_GENERIC_ID_VLAN_TAG_MISMATCH = 1, ++ DEVLINK_TRAP_GENERIC_ID_INGRESS_VLAN_FILTER = 2, ++ DEVLINK_TRAP_GENERIC_ID_INGRESS_STP_FILTER = 3, ++ DEVLINK_TRAP_GENERIC_ID_EMPTY_TX_LIST = 4, ++ DEVLINK_TRAP_GENERIC_ID_PORT_LOOPBACK_FILTER = 5, ++ DEVLINK_TRAP_GENERIC_ID_BLACKHOLE_ROUTE = 6, ++ DEVLINK_TRAP_GENERIC_ID_TTL_ERROR = 7, ++ DEVLINK_TRAP_GENERIC_ID_TAIL_DROP = 8, ++ DEVLINK_TRAP_GENERIC_ID_NON_IP_PACKET = 9, ++ DEVLINK_TRAP_GENERIC_ID_UC_DIP_MC_DMAC = 10, ++ DEVLINK_TRAP_GENERIC_ID_DIP_LB = 11, ++ DEVLINK_TRAP_GENERIC_ID_SIP_MC = 12, ++ DEVLINK_TRAP_GENERIC_ID_SIP_LB = 13, ++ DEVLINK_TRAP_GENERIC_ID_CORRUPTED_IP_HDR = 14, ++ DEVLINK_TRAP_GENERIC_ID_IPV4_SIP_BC = 15, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_MC_DIP_RESERVED_SCOPE = 16, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_MC_DIP_INTERFACE_LOCAL_SCOPE = 17, ++ DEVLINK_TRAP_GENERIC_ID_MTU_ERROR = 18, ++ DEVLINK_TRAP_GENERIC_ID_UNRESOLVED_NEIGH = 19, ++ DEVLINK_TRAP_GENERIC_ID_RPF = 20, ++ DEVLINK_TRAP_GENERIC_ID_REJECT_ROUTE = 21, ++ DEVLINK_TRAP_GENERIC_ID_IPV4_LPM_UNICAST_MISS = 22, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_LPM_UNICAST_MISS = 23, ++ DEVLINK_TRAP_GENERIC_ID_NON_ROUTABLE = 24, ++ DEVLINK_TRAP_GENERIC_ID_DECAP_ERROR = 25, ++ DEVLINK_TRAP_GENERIC_ID_OVERLAY_SMAC_MC = 26, ++ DEVLINK_TRAP_GENERIC_ID_INGRESS_FLOW_ACTION_DROP = 27, ++ DEVLINK_TRAP_GENERIC_ID_EGRESS_FLOW_ACTION_DROP = 28, ++ DEVLINK_TRAP_GENERIC_ID_STP = 29, ++ DEVLINK_TRAP_GENERIC_ID_LACP = 30, ++ DEVLINK_TRAP_GENERIC_ID_LLDP = 31, ++ DEVLINK_TRAP_GENERIC_ID_IGMP_QUERY = 32, ++ DEVLINK_TRAP_GENERIC_ID_IGMP_V1_REPORT = 33, ++ DEVLINK_TRAP_GENERIC_ID_IGMP_V2_REPORT = 34, ++ DEVLINK_TRAP_GENERIC_ID_IGMP_V3_REPORT = 35, ++ DEVLINK_TRAP_GENERIC_ID_IGMP_V2_LEAVE = 36, ++ DEVLINK_TRAP_GENERIC_ID_MLD_QUERY = 37, ++ DEVLINK_TRAP_GENERIC_ID_MLD_V1_REPORT = 38, ++ DEVLINK_TRAP_GENERIC_ID_MLD_V2_REPORT = 39, ++ DEVLINK_TRAP_GENERIC_ID_MLD_V1_DONE = 40, ++ DEVLINK_TRAP_GENERIC_ID_IPV4_DHCP = 41, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_DHCP = 42, ++ DEVLINK_TRAP_GENERIC_ID_ARP_REQUEST = 43, ++ DEVLINK_TRAP_GENERIC_ID_ARP_RESPONSE = 44, ++ DEVLINK_TRAP_GENERIC_ID_ARP_OVERLAY = 45, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_NEIGH_SOLICIT = 46, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_NEIGH_ADVERT = 47, ++ DEVLINK_TRAP_GENERIC_ID_IPV4_BFD = 48, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_BFD = 49, ++ DEVLINK_TRAP_GENERIC_ID_IPV4_OSPF = 50, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_OSPF = 51, ++ DEVLINK_TRAP_GENERIC_ID_IPV4_BGP = 52, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_BGP = 53, ++ DEVLINK_TRAP_GENERIC_ID_IPV4_VRRP = 54, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_VRRP = 55, ++ DEVLINK_TRAP_GENERIC_ID_IPV4_PIM = 56, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_PIM = 57, ++ DEVLINK_TRAP_GENERIC_ID_UC_LB = 58, ++ DEVLINK_TRAP_GENERIC_ID_LOCAL_ROUTE = 59, ++ DEVLINK_TRAP_GENERIC_ID_EXTERNAL_ROUTE = 60, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_UC_DIP_LINK_LOCAL_SCOPE = 61, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_DIP_ALL_NODES = 62, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_DIP_ALL_ROUTERS = 63, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_ROUTER_SOLICIT = 64, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_ROUTER_ADVERT = 65, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_REDIRECT = 66, ++ DEVLINK_TRAP_GENERIC_ID_IPV4_ROUTER_ALERT = 67, ++ DEVLINK_TRAP_GENERIC_ID_IPV6_ROUTER_ALERT = 68, ++ DEVLINK_TRAP_GENERIC_ID_PTP_EVENT = 69, ++ DEVLINK_TRAP_GENERIC_ID_PTP_GENERAL = 70, ++ DEVLINK_TRAP_GENERIC_ID_FLOW_ACTION_SAMPLE = 71, ++ DEVLINK_TRAP_GENERIC_ID_FLOW_ACTION_TRAP = 72, ++ DEVLINK_TRAP_GENERIC_ID_EARLY_DROP = 73, ++ DEVLINK_TRAP_GENERIC_ID_VXLAN_PARSING = 74, ++ DEVLINK_TRAP_GENERIC_ID_LLC_SNAP_PARSING = 75, ++ DEVLINK_TRAP_GENERIC_ID_VLAN_PARSING = 76, ++ DEVLINK_TRAP_GENERIC_ID_PPPOE_PPP_PARSING = 77, ++ DEVLINK_TRAP_GENERIC_ID_MPLS_PARSING = 78, ++ DEVLINK_TRAP_GENERIC_ID_ARP_PARSING = 79, ++ DEVLINK_TRAP_GENERIC_ID_IP_1_PARSING = 80, ++ DEVLINK_TRAP_GENERIC_ID_IP_N_PARSING = 81, ++ DEVLINK_TRAP_GENERIC_ID_GRE_PARSING = 82, ++ DEVLINK_TRAP_GENERIC_ID_UDP_PARSING = 83, ++ DEVLINK_TRAP_GENERIC_ID_TCP_PARSING = 84, ++ DEVLINK_TRAP_GENERIC_ID_IPSEC_PARSING = 85, ++ DEVLINK_TRAP_GENERIC_ID_SCTP_PARSING = 86, ++ DEVLINK_TRAP_GENERIC_ID_DCCP_PARSING = 87, ++ DEVLINK_TRAP_GENERIC_ID_GTP_PARSING = 88, ++ DEVLINK_TRAP_GENERIC_ID_ESP_PARSING = 89, ++ DEVLINK_TRAP_GENERIC_ID_BLACKHOLE_NEXTHOP = 90, ++ DEVLINK_TRAP_GENERIC_ID_DMAC_FILTER = 91, ++ DEVLINK_TRAP_GENERIC_ID_EAPOL = 92, ++ DEVLINK_TRAP_GENERIC_ID_LOCKED_PORT = 93, ++ __DEVLINK_TRAP_GENERIC_ID_MAX = 94, ++ DEVLINK_TRAP_GENERIC_ID_MAX = 93, ++}; ++ ++enum devlink_trap_group_generic_id { ++ DEVLINK_TRAP_GROUP_GENERIC_ID_L2_DROPS = 0, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_L3_DROPS = 1, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_L3_EXCEPTIONS = 2, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_BUFFER_DROPS = 3, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_TUNNEL_DROPS = 4, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_ACL_DROPS = 5, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_STP = 6, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_LACP = 7, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_LLDP = 8, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_MC_SNOOPING = 9, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_DHCP = 10, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_NEIGH_DISCOVERY = 11, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_BFD = 12, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_OSPF = 13, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_BGP = 14, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_VRRP = 15, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_PIM = 16, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_UC_LB = 17, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_LOCAL_DELIVERY = 18, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_EXTERNAL_DELIVERY = 19, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_IPV6 = 20, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_PTP_EVENT = 21, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_PTP_GENERAL = 22, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_ACL_SAMPLE = 23, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_ACL_TRAP = 24, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_PARSER_ERROR_DROPS = 25, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_EAPOL = 26, ++ __DEVLINK_TRAP_GROUP_GENERIC_ID_MAX = 27, ++ DEVLINK_TRAP_GROUP_GENERIC_ID_MAX = 26, ++}; ++ ++struct trace_event_raw_devlink_hwmsg { ++ struct trace_entry ent; ++ u32 __data_loc_bus_name; ++ u32 __data_loc_dev_name; ++ u32 __data_loc_driver_name; ++ bool incoming; ++ long unsigned int type; ++ u32 __data_loc_buf; ++ size_t len; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_devlink_hwerr { ++ struct trace_entry ent; ++ u32 __data_loc_bus_name; ++ u32 __data_loc_dev_name; ++ u32 __data_loc_driver_name; ++ int err; ++ u32 __data_loc_msg; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_devlink_health_report { ++ struct trace_entry ent; ++ u32 __data_loc_bus_name; ++ u32 __data_loc_dev_name; ++ u32 __data_loc_driver_name; ++ u32 __data_loc_reporter_name; ++ u32 __data_loc_msg; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_devlink_health_recover_aborted { ++ struct trace_entry ent; ++ u32 __data_loc_bus_name; ++ u32 __data_loc_dev_name; ++ u32 __data_loc_driver_name; ++ u32 __data_loc_reporter_name; ++ bool health_state; ++ u64 time_since_last_recover; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_devlink_health_reporter_state_update { ++ struct trace_entry ent; ++ u32 __data_loc_bus_name; ++ u32 __data_loc_dev_name; ++ u32 __data_loc_driver_name; ++ u32 __data_loc_reporter_name; ++ u8 new_state; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_devlink_trap_report { ++ struct trace_entry ent; ++ u32 __data_loc_bus_name; ++ u32 __data_loc_dev_name; ++ u32 __data_loc_driver_name; ++ u32 __data_loc_trap_name; ++ u32 __data_loc_trap_group_name; ++ char input_dev_name[16]; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_devlink_hwmsg { ++ u32 bus_name; ++ u32 dev_name; ++ u32 driver_name; ++ u32 buf; ++}; ++ ++struct trace_event_data_offsets_devlink_hwerr { ++ u32 bus_name; ++ u32 dev_name; ++ u32 driver_name; ++ u32 msg; ++}; ++ ++struct trace_event_data_offsets_devlink_health_report { ++ u32 bus_name; ++ u32 dev_name; ++ u32 driver_name; ++ u32 reporter_name; ++ u32 msg; ++}; ++ ++struct trace_event_data_offsets_devlink_health_recover_aborted { ++ u32 bus_name; ++ u32 dev_name; ++ u32 driver_name; ++ u32 reporter_name; ++}; ++ ++struct trace_event_data_offsets_devlink_health_reporter_state_update { ++ u32 bus_name; ++ u32 dev_name; ++ u32 driver_name; ++ u32 reporter_name; ++}; ++ ++struct trace_event_data_offsets_devlink_trap_report { ++ u32 bus_name; ++ u32 dev_name; ++ u32 driver_name; ++ u32 trap_name; ++ u32 trap_group_name; ++}; ++ ++typedef void (*btf_trace_devlink_hwmsg)(void *, const struct devlink *, bool, long unsigned int, const u8 *, size_t); ++ ++typedef void (*btf_trace_devlink_hwerr)(void *, const struct devlink *, int, const char *); ++ ++typedef void (*btf_trace_devlink_health_report)(void *, const struct devlink *, const char *, const char *); ++ ++typedef void (*btf_trace_devlink_health_recover_aborted)(void *, const struct devlink *, const char *, bool, u64); ++ ++typedef void (*btf_trace_devlink_health_reporter_state_update)(void *, const struct devlink *, const char *, bool); ++ ++typedef void (*btf_trace_devlink_trap_report)(void *, const struct devlink *, struct sk_buff *, const struct devlink_trap_metadata *); ++ ++struct devlink_linecard_type { ++ const char *type; ++ const void *priv; ++}; ++ ++struct devlink_resource { ++ const char *name; ++ u64 id; ++ u64 size; ++ u64 size_new; ++ bool size_valid; ++ struct devlink_resource *parent; ++ struct devlink_resource_size_params size_params; ++ struct list_head list; ++ struct list_head resource_list; ++ devlink_resource_occ_get_t *occ_get; ++ void *occ_get_priv; ++}; ++ ++struct devlink_sb { ++ struct list_head list; ++ unsigned int index; ++ u32 size; ++ u16 ingress_pools_count; ++ u16 egress_pools_count; ++ u16 ingress_tc_count; ++ u16 egress_tc_count; ++}; ++ ++struct devlink_region { ++ struct devlink *devlink; ++ struct devlink_port *port; ++ struct list_head list; ++ union { ++ const struct devlink_region_ops *ops; ++ const struct devlink_port_region_ops *port_ops; ++ }; ++ struct mutex snapshot_lock; ++ struct list_head snapshot_list; ++ u32 max_snapshots; ++ u32 cur_snapshots; ++ u64 size; ++}; ++ ++struct devlink_snapshot { ++ struct list_head list; ++ struct devlink_region *region; ++ u8 *data; ++ u32 id; ++}; ++ ++typedef int devlink_chunk_fill_t(void *, u8 *, u32, u64, struct netlink_ext_ack *); ++ ++struct devlink_stats { ++ u64_stats_t rx_bytes; ++ u64_stats_t rx_packets; ++ struct u64_stats_sync syncp; ++}; ++ ++struct devlink_trap_policer_item { ++ const struct devlink_trap_policer *policer; ++ u64 rate; ++ u64 burst; ++ struct list_head list; ++}; ++ ++struct devlink_trap_group_item { ++ const struct devlink_trap_group *group; ++ struct devlink_trap_policer_item *policer_item; ++ struct list_head list; ++ struct devlink_stats *stats; ++}; ++ ++struct devlink_trap_item { ++ const struct devlink_trap *trap; ++ struct devlink_trap_group_item *group_item; ++ struct list_head list; ++ enum devlink_trap_action action; ++ struct devlink_stats *stats; ++ void *priv; ++}; ++ ++enum { ++ NLBL_CIPSOV4_C_UNSPEC = 0, ++ NLBL_CIPSOV4_C_ADD = 1, ++ NLBL_CIPSOV4_C_REMOVE = 2, ++ NLBL_CIPSOV4_C_LIST = 3, ++ NLBL_CIPSOV4_C_LISTALL = 4, ++ __NLBL_CIPSOV4_C_MAX = 5, ++}; ++ ++enum { ++ NLBL_CIPSOV4_A_UNSPEC = 0, ++ NLBL_CIPSOV4_A_DOI = 1, ++ NLBL_CIPSOV4_A_MTYPE = 2, ++ NLBL_CIPSOV4_A_TAG = 3, ++ NLBL_CIPSOV4_A_TAGLST = 4, ++ NLBL_CIPSOV4_A_MLSLVLLOC = 5, ++ NLBL_CIPSOV4_A_MLSLVLREM = 6, ++ NLBL_CIPSOV4_A_MLSLVL = 7, ++ NLBL_CIPSOV4_A_MLSLVLLST = 8, ++ NLBL_CIPSOV4_A_MLSCATLOC = 9, ++ NLBL_CIPSOV4_A_MLSCATREM = 10, ++ NLBL_CIPSOV4_A_MLSCAT = 11, ++ NLBL_CIPSOV4_A_MLSCATLST = 12, ++ __NLBL_CIPSOV4_A_MAX = 13, ++}; ++ ++struct netlbl_cipsov4_doiwalk_arg { ++ struct netlink_callback *nl_cb; ++ struct sk_buff *skb; ++ u32 seq; ++}; ++ ++struct ncsi_cmd_pkt { ++ struct ncsi_cmd_pkt_hdr cmd; ++ __be32 checksum; ++ unsigned char pad[26]; ++}; ++ ++struct ncsi_cmd_sp_pkt { ++ struct ncsi_cmd_pkt_hdr cmd; ++ unsigned char reserved[3]; ++ unsigned char hw_arbitration; ++ __be32 checksum; ++ unsigned char pad[22]; ++}; ++ ++struct ncsi_cmd_dc_pkt { ++ struct ncsi_cmd_pkt_hdr cmd; ++ unsigned char reserved[3]; ++ unsigned char ald; ++ __be32 checksum; ++ unsigned char pad[22]; ++}; ++ ++struct ncsi_cmd_rc_pkt { ++ struct ncsi_cmd_pkt_hdr cmd; ++ __be32 reserved; ++ __be32 checksum; ++ unsigned char pad[22]; ++}; ++ ++struct ncsi_cmd_oem_pkt { ++ struct ncsi_cmd_pkt_hdr cmd; ++ __be32 mfr_id; ++ unsigned char data[0]; ++}; ++ ++struct ncsi_cmd_handler { ++ unsigned char type; ++ int payload; ++ int (*handler)(struct sk_buff *, struct ncsi_cmd_arg *); ++}; ++ ++enum { ++ ncsi_dev_state_registered = 0, ++ ncsi_dev_state_functional = 256, ++ ncsi_dev_state_probe = 512, ++ ncsi_dev_state_config = 768, ++ ncsi_dev_state_suspend = 1024, ++}; ++ ++enum { ++ MLX_MC_RBT_SUPPORT = 1, ++ MLX_MC_RBT_AVL = 8, ++}; ++ ++enum { ++ ncsi_dev_state_major = 65280, ++ ncsi_dev_state_minor = 255, ++ ncsi_dev_state_probe_deselect = 513, ++ ncsi_dev_state_probe_package = 514, ++ ncsi_dev_state_probe_channel = 515, ++ ncsi_dev_state_probe_mlx_gma = 516, ++ ncsi_dev_state_probe_mlx_smaf = 517, ++ ncsi_dev_state_probe_cis = 518, ++ ncsi_dev_state_probe_keep_phy = 519, ++ ncsi_dev_state_probe_gvi = 520, ++ ncsi_dev_state_probe_gc = 521, ++ ncsi_dev_state_probe_gls = 522, ++ ncsi_dev_state_probe_dp = 523, ++ ncsi_dev_state_config_sp = 769, ++ ncsi_dev_state_config_cis = 770, ++ ncsi_dev_state_config_oem_gma = 771, ++ ncsi_dev_state_config_clear_vids = 772, ++ ncsi_dev_state_config_svf = 773, ++ ncsi_dev_state_config_ev = 774, ++ ncsi_dev_state_config_sma = 775, ++ ncsi_dev_state_config_ebf = 776, ++ ncsi_dev_state_config_dgmf = 777, ++ ncsi_dev_state_config_ecnt = 778, ++ ncsi_dev_state_config_ec = 779, ++ ncsi_dev_state_config_ae = 780, ++ ncsi_dev_state_config_gls = 781, ++ ncsi_dev_state_config_done = 782, ++ ncsi_dev_state_suspend_select = 1025, ++ ncsi_dev_state_suspend_gls = 1026, ++ ncsi_dev_state_suspend_dcnt = 1027, ++ ncsi_dev_state_suspend_dc = 1028, ++ ncsi_dev_state_suspend_deselect = 1029, ++ ncsi_dev_state_suspend_done = 1030, ++}; ++ ++struct vlan_vid { ++ struct list_head list; ++ __be16 proto; ++ u16 vid; ++}; ++ ++struct ncsi_oem_gma_handler { ++ unsigned int mfr_id; ++ int (*handler)(struct ncsi_cmd_arg *); ++}; ++ ++enum mapping_status { ++ MAPPING_OK = 0, ++ MAPPING_INVALID = 1, ++ MAPPING_EMPTY = 2, ++ MAPPING_DATA_FIN = 3, ++ MAPPING_DUMMY = 4, ++ MAPPING_BAD_CSUM = 5, ++}; ++ ++struct mptcp_pernet { ++ struct ctl_table_header *ctl_table_hdr; ++ unsigned int add_addr_timeout; ++ unsigned int stale_loss_cnt; ++ u8 mptcp_enabled; ++ u8 checksum_enabled; ++ u8 allow_join_initial_addr_port; ++ u8 pm_type; ++}; ++ ++struct mptcp_diag_ctx { ++ long int s_slot; ++ long int s_num; ++ unsigned int l_slot; ++ unsigned int l_num; ++}; ++ ++enum { ++ IFLA_MCTP_UNSPEC = 0, ++ IFLA_MCTP_NET = 1, ++ __IFLA_MCTP_MAX = 2, ++}; ++ ++struct mctp_dump_cb { ++ int h; ++ int idx; ++ size_t a_idx; ++}; ++ ++enum { ++ HANDSHAKE_NLGRP_NONE = 0, ++ HANDSHAKE_NLGRP_TLSHD = 1, ++}; ++ ++enum hr_flags_bits { ++ HANDSHAKE_F_REQ_COMPLETED = 0, ++}; ++ ++struct cb_data { ++ void *data; ++ struct idset *set; ++ int (*fn_known_sch)(struct subchannel *, void *); ++ int (*fn_unknown_sch)(struct subchannel_id, void *); ++}; ++ ++struct trace_event_raw_s390_class_schib { ++ struct trace_entry ent; ++ u8 cssid; ++ u8 ssid; ++ u16 schno; ++ u16 devno; ++ long: 0; ++ struct schib schib; ++ u8 pmcw_ena; ++ u8 pmcw_st; ++ u8 pmcw_dnv; ++ u16 pmcw_dev; ++ u8 pmcw_lpm; ++ u8 pmcw_pnom; ++ u8 pmcw_lpum; ++ u8 pmcw_pim; ++ u8 pmcw_pam; ++ u8 pmcw_pom; ++ u64 pmcw_chpid; ++ int cc; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_s390_cio_tsch { ++ struct trace_entry ent; ++ u8 cssid; ++ u8 ssid; ++ u16 schno; ++ struct irb irb; ++ u8 scsw_dcc; ++ u8 scsw_pno; ++ u8 scsw_fctl; ++ u8 scsw_actl; ++ u8 scsw_stctl; ++ u8 scsw_dstat; ++ u8 scsw_cstat; ++ int cc; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_s390_cio_tpi { ++ struct trace_entry ent; ++ int cc; ++ struct tpi_info tpi_info; ++ u8 cssid; ++ u8 ssid; ++ u16 schno; ++ u8 adapter_IO; ++ u8 isc; ++ u8 type; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_s390_cio_ssch { ++ struct trace_entry ent; ++ u8 cssid; ++ u8 ssid; ++ u16 schno; ++ union orb orb; ++ int cc; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_s390_class_schid { ++ struct trace_entry ent; ++ u8 cssid; ++ u8 ssid; ++ u16 schno; ++ int cc; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_s390_cio_chsc { ++ struct trace_entry ent; ++ int cc; ++ u16 code; ++ u16 rcode; ++ u8 request[64]; ++ u8 response[64]; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_s390_cio_interrupt { ++ struct trace_entry ent; ++ struct tpi_info tpi_info; ++ u8 cssid; ++ u8 ssid; ++ u16 schno; ++ u8 isc; ++ u8 type; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_s390_cio_adapter_int { ++ struct trace_entry ent; ++ struct tpi_info tpi_info; ++ u8 isc; ++ char __data[0]; ++}; ++ ++struct trace_event_raw_s390_cio_stcrw { ++ struct trace_entry ent; ++ struct crw crw; ++ int cc; ++ u8 slct; ++ u8 oflw; ++ u8 chn; ++ u8 rsc; ++ u8 anc; ++ u8 erc; ++ u16 rsid; ++ char __data[0]; ++}; ++ ++struct trace_event_data_offsets_s390_class_schib {}; ++ ++struct trace_event_data_offsets_s390_cio_tsch {}; ++ ++struct trace_event_data_offsets_s390_cio_tpi {}; ++ ++struct trace_event_data_offsets_s390_cio_ssch {}; ++ ++struct trace_event_data_offsets_s390_class_schid {}; ++ ++struct trace_event_data_offsets_s390_cio_chsc {}; ++ ++struct trace_event_data_offsets_s390_cio_interrupt {}; ++ ++struct trace_event_data_offsets_s390_cio_adapter_int {}; ++ ++struct trace_event_data_offsets_s390_cio_stcrw {}; ++ ++typedef void (*btf_trace_s390_cio_stsch)(void *, struct subchannel_id, struct schib *, int); ++ ++typedef void (*btf_trace_s390_cio_msch)(void *, struct subchannel_id, struct schib *, int); ++ ++typedef void (*btf_trace_s390_cio_tsch)(void *, struct subchannel_id, struct irb *, int); ++ ++typedef void (*btf_trace_s390_cio_tpi)(void *, struct tpi_info *, int); ++ ++typedef void (*btf_trace_s390_cio_ssch)(void *, struct subchannel_id, union orb *, int); ++ ++typedef void (*btf_trace_s390_cio_csch)(void *, struct subchannel_id, int); ++ ++typedef void (*btf_trace_s390_cio_hsch)(void *, struct subchannel_id, int); ++ ++typedef void (*btf_trace_s390_cio_xsch)(void *, struct subchannel_id, int); ++ ++typedef void (*btf_trace_s390_cio_rsch)(void *, struct subchannel_id, int); ++ ++typedef void (*btf_trace_s390_cio_chsc)(void *, struct chsc_header *, int); ++ ++typedef void (*btf_trace_s390_cio_interrupt)(void *, struct tpi_info *); ++ ++typedef void (*btf_trace_s390_cio_adapter_int)(void *, struct tpi_info *); ++ ++typedef void (*btf_trace_s390_cio_stcrw)(void *, struct crw *, int); ++ ++struct cmbdata { ++ __u64 size; ++ __u64 elapsed_time; ++ __u64 ssch_rsch_count; ++ __u64 sample_count; ++ __u64 device_connect_time; ++ __u64 function_pending_time; ++ __u64 device_disconnect_time; ++ __u64 control_unit_queuing_time; ++ __u64 device_active_only_time; ++ __u64 device_busy_time; ++ __u64 initial_command_response_time; ++}; ++ ++enum cmb_index { ++ avg_utilization = -1, ++ cmb_ssch_rsch_count = 0, ++ cmb_sample_count = 1, ++ cmb_device_connect_time = 2, ++ cmb_function_pending_time = 3, ++ cmb_device_disconnect_time = 4, ++ cmb_control_unit_queuing_time = 5, ++ cmb_device_active_only_time = 6, ++ cmb_device_busy_time = 7, ++ cmb_initial_command_response_time = 8, ++}; ++ ++enum cmb_format { ++ CMF_BASIC = 0, ++ CMF_EXTENDED = 1, ++ CMF_AUTODETECT = -1, ++}; ++ ++struct cmb_operations { ++ int (*alloc)(struct ccw_device *); ++ void (*free)(struct ccw_device *); ++ int (*set)(struct ccw_device *, u32); ++ u64 (*read)(struct ccw_device *, int); ++ int (*readall)(struct ccw_device *, struct cmbdata *); ++ void (*reset)(struct ccw_device *); ++ struct attribute_group *attr_group; ++}; ++ ++struct cmb_data { ++ void *hw_block; ++ void *last_block; ++ int size; ++ long long unsigned int last_update; ++}; ++ ++struct set_schib_struct { ++ u32 mme; ++ int mbfc; ++ long unsigned int address; ++ wait_queue_head_t wait; ++ int ret; ++}; ++ ++struct copy_block_struct { ++ wait_queue_head_t wait; ++ int ret; ++}; ++ ++struct cmb; ++ ++struct cmb_area { ++ struct cmb *mem; ++ struct list_head list; ++ int num_channels; ++ spinlock_t lock; ++}; ++ ++struct cmb { ++ u16 ssch_rsch_count; ++ u16 sample_count; ++ u32 device_connect_time; ++ u32 function_pending_time; ++ u32 device_disconnect_time; ++ u32 control_unit_queuing_time; ++ u32 device_active_only_time; ++ u32 reserved[2]; ++}; ++ ++struct cmbe { ++ u32 ssch_rsch_count; ++ u32 sample_count; ++ u32 device_connect_time; ++ u32 function_pending_time; ++ u32 device_disconnect_time; ++ u32 control_unit_queuing_time; ++ u32 device_active_only_time; ++ u32 device_busy_time; ++ u32 initial_command_response_time; ++ u32 reserved[7]; ++}; ++ ++struct sclp_trace_entry { ++ char id[4]; ++ u32 a; ++ u64 b; ++}; ++ ++enum sclp_running_state_t { ++ sclp_running_state_idle = 0, ++ sclp_running_state_running = 1, ++ sclp_running_state_reset_pending = 2, ++}; ++ ++enum sclp_reading_state_t { ++ sclp_reading_state_idle = 0, ++ sclp_reading_state_reading = 1, ++}; ++ ++enum sclp_activation_state_t { ++ sclp_activation_state_active = 0, ++ sclp_activation_state_deactivating = 1, ++ sclp_activation_state_inactive = 2, ++ sclp_activation_state_activating = 3, ++}; ++ ++enum sclp_mask_state_t { ++ sclp_mask_state_idle = 0, ++ sclp_mask_state_initializing = 1, ++}; ++ ++struct sclp_statechangebuf { ++ struct evbuf_header header; ++ u8 validity_sclp_active_facility_mask: 1; ++ u8 validity_sclp_receive_mask: 1; ++ u8 validity_sclp_send_mask: 1; ++ u8 validity_read_data_function_mask: 1; ++ u16 _zeros: 12; ++ u16 mask_length; ++ u64 sclp_active_facility_mask; ++ u8 masks[2046]; ++} __attribute__((packed)); ++ ++struct sclp_ctl_sccb { ++ __u32 cmdw; ++ __u64 sccb; ++} __attribute__((packed)); ++ ++struct tty3270_attribute { ++ unsigned char alternate_charset: 1; ++ unsigned char highlight: 3; ++ unsigned char f_color: 4; ++ unsigned char b_color: 4; ++}; ++ ++struct tty3270_cell { ++ unsigned char character; ++ struct tty3270_attribute attributes; ++}; ++ ++struct tty3270_line { ++ struct tty3270_cell *cells; ++ int len; ++ int dirty; ++}; ++ ++struct tty3270 { ++ struct raw3270_view view; ++ struct tty_port port; ++ unsigned char wcc; ++ int nr_up; ++ long unsigned int update_flags; ++ struct raw3270_request *write; ++ struct timer_list timer; ++ char *converted_line; ++ unsigned int line_view_start; ++ unsigned int line_write_start; ++ unsigned int oops_line; ++ unsigned int cx; ++ unsigned int cy; ++ struct tty3270_attribute attributes; ++ struct tty3270_attribute saved_attributes; ++ int allocated_lines; ++ struct tty3270_line *screen; ++ char *prompt; ++ char *input; ++ struct raw3270_request *read; ++ struct raw3270_request *kreset; ++ struct raw3270_request *readpartreq; ++ unsigned char inattr; ++ int throttle; ++ int attn; ++ struct tasklet_struct readlet; ++ struct tasklet_struct hanglet; ++ struct kbd_data *kbd; ++ int esc_state; ++ int esc_ques; ++ int esc_npar; ++ int esc_par[8]; ++ unsigned int saved_cx; ++ unsigned int saved_cy; ++ char **rcl_lines; ++ int rcl_write_index; ++ int rcl_read_index; ++ unsigned int char_count; ++ char char_buf[256]; ++}; ++ ++enum { ++ ES_NORMAL = 0, ++ ES_ESC = 1, ++ ES_SQUARE = 2, ++ ES_PAREN = 3, ++ ES_GETPARS = 4, ++}; ++ ++struct sdias_evbuf { ++ struct evbuf_header hdr; ++ u8 event_qual; ++ u8 data_id; ++ u64 reserved2; ++ u32 event_id; ++ u16 reserved3; ++ u8 asa_size; ++ u8 event_status; ++ u32 reserved4; ++ u32 blk_cnt; ++ u64 asa; ++ u32 reserved5; ++ u32 fbn; ++ u32 reserved6; ++ u32 lbn; ++ u16 reserved7; ++ u16 dbs; ++} __attribute__((packed)); ++ ++struct sdias_sccb { ++ struct sccb_header hdr; ++ struct sdias_evbuf evbuf; ++}; ++ ++union ap_qirq_ctrl { ++ long unsigned int value; ++ struct { ++ char: 8; ++ unsigned int zone: 8; ++ unsigned int ir: 1; ++ char: 4; ++ unsigned int gisc: 3; ++ char: 6; ++ unsigned int gf: 2; ++ char: 1; ++ unsigned int gisa: 27; ++ char: 1; ++ unsigned int isc: 3; ++ }; ++}; ++ ++typedef enum ap_sm_wait ap_func_t(struct ap_queue *); ++ ++struct uevent_sock { ++ struct list_head list; ++ struct sock *sk; ++}; ++ ++#ifndef BPF_NO_PRESERVE_ACCESS_INDEX ++#pragma clang attribute pop ++#endif ++ ++#endif /* __VMLINUX_H__ */ +-- +2.43.0 + diff --git a/SOURCES/bcc-0.28.0-libbpf-tools-add-block_io_-start-done-tracepoints-su.patch b/SOURCES/bcc-0.28.0-libbpf-tools-add-block_io_-start-done-tracepoints-su.patch new file mode 100644 index 0000000..7796e6a --- /dev/null +++ b/SOURCES/bcc-0.28.0-libbpf-tools-add-block_io_-start-done-tracepoints-su.patch @@ -0,0 +1,476 @@ +From 60860bf3a400dcf72b4026fb2973803cfb12ccf1 Mon Sep 17 00:00:00 2001 +From: mickey_zhu +Date: Tue, 27 Jun 2023 16:32:44 +0800 +Subject: [PATCH] libbpf-tools: add block_io_{start,done} tracepoints support + to bio tools + +Some bio tools fail to kprobe blk_account_io_{start,done} after v5.17, +because they become inlined, see [0]. To fix this issue, tracepoints +blick_io_{start,done} are introcuded in kernel, see[1]. + +Update related bio tools to support new tracepoints, and also simplify +attach. + +[0] Kernel commit 450b7879e345 (block: move blk_account_io_{start,done} to blk-mq.c) +[1] Kernel commit 5a80bd075f3b (block: introduce block_io_start/block_io_done tracepoints) + +Change-Id: I62b957abd7ce2901eb114bd57c78938e4f083e4d +Signed-off-by: Mickey Zhu +--- + libbpf-tools/biosnoop.bpf.c | 9 ++++ + libbpf-tools/biosnoop.c | 78 +++++++++++++-------------------- + libbpf-tools/biostacks.bpf.c | 46 +++++++++++++------ + libbpf-tools/biostacks.c | 85 +++++++++++++++++++++--------------- + libbpf-tools/biotop.bpf.c | 44 +++++++++++++++++-- + libbpf-tools/biotop.c | 59 ++++++++++++++++--------- + 6 files changed, 199 insertions(+), 122 deletions(-) + +diff --git a/libbpf-tools/biosnoop.bpf.c b/libbpf-tools/biosnoop.bpf.c +index b791555f..fcc5c5ce 100644 +--- a/libbpf-tools/biosnoop.bpf.c ++++ b/libbpf-tools/biosnoop.bpf.c +@@ -76,6 +76,15 @@ int BPF_PROG(blk_account_io_start, struct request *rq) + return trace_pid(rq); + } + ++SEC("tp_btf/block_io_start") ++int BPF_PROG(block_io_start, struct request *rq) ++{ ++ if (filter_cg && !bpf_current_task_under_cgroup(&cgroup_map, 0)) ++ return 0; ++ ++ return trace_pid(rq); ++} ++ + SEC("kprobe/blk_account_io_merge_bio") + int BPF_KPROBE(blk_account_io_merge_bio, struct request *rq) + { +diff --git a/libbpf-tools/biosnoop.c b/libbpf-tools/biosnoop.c +index 21773729..f9468900 100644 +--- a/libbpf-tools/biosnoop.c ++++ b/libbpf-tools/biosnoop.c +@@ -212,6 +212,16 @@ void handle_lost_events(void *ctx, int cpu, __u64 lost_cnt) + fprintf(stderr, "lost %llu events on CPU #%d\n", lost_cnt, cpu); + } + ++static void blk_account_io_set_attach_target(struct biosnoop_bpf *obj) ++{ ++ if (fentry_can_attach("blk_account_io_start", NULL)) ++ bpf_program__set_attach_target(obj->progs.blk_account_io_start, ++ 0, "blk_account_io_start"); ++ else ++ bpf_program__set_attach_target(obj->progs.blk_account_io_start, ++ 0, "__blk_account_io_start"); ++} ++ + int main(int argc, char **argv) + { + const struct partition *partition; +@@ -260,12 +270,23 @@ int main(int argc, char **argv) + obj->rodata->filter_cg = env.cg; + obj->rodata->min_ns = env.min_lat_ms * 1000000; + +- if (fentry_can_attach("blk_account_io_start", NULL)) +- bpf_program__set_attach_target(obj->progs.blk_account_io_start, 0, +- "blk_account_io_start"); +- else +- bpf_program__set_attach_target(obj->progs.blk_account_io_start, 0, +- "__blk_account_io_start"); ++ if (tracepoint_exists("block", "block_io_start")) ++ bpf_program__set_autoload(obj->progs.blk_account_io_start, false); ++ else { ++ bpf_program__set_autoload(obj->progs.block_io_start, false); ++ blk_account_io_set_attach_target(obj); ++ } ++ ++ ksyms = ksyms__load(); ++ if (!ksyms) { ++ fprintf(stderr, "failed to load kallsyms\n"); ++ goto cleanup; ++ } ++ if (!ksyms__get_symbol(ksyms, "blk_account_io_merge_bio")) ++ bpf_program__set_autoload(obj->progs.blk_account_io_merge_bio, false); ++ ++ if (!env.queued) ++ bpf_program__set_autoload(obj->progs.block_rq_insert, false); + + err = biosnoop_bpf__load(obj); + if (err) { +@@ -288,48 +309,9 @@ int main(int argc, char **argv) + } + } + +- obj->links.blk_account_io_start = bpf_program__attach(obj->progs.blk_account_io_start); +- if (!obj->links.blk_account_io_start) { +- err = -errno; +- fprintf(stderr, "failed to attach blk_account_io_start: %s\n", +- strerror(-err)); +- goto cleanup; +- } +- ksyms = ksyms__load(); +- if (!ksyms) { +- err = -ENOMEM; +- fprintf(stderr, "failed to load kallsyms\n"); +- goto cleanup; +- } +- if (ksyms__get_symbol(ksyms, "blk_account_io_merge_bio")) { +- obj->links.blk_account_io_merge_bio = +- bpf_program__attach(obj->progs.blk_account_io_merge_bio); +- if (!obj->links.blk_account_io_merge_bio) { +- err = -errno; +- fprintf(stderr, "failed to attach blk_account_io_merge_bio: %s\n", +- strerror(-err)); +- goto cleanup; +- } +- } +- if (env.queued) { +- obj->links.block_rq_insert = +- bpf_program__attach(obj->progs.block_rq_insert); +- if (!obj->links.block_rq_insert) { +- err = -errno; +- fprintf(stderr, "failed to attach block_rq_insert: %s\n", strerror(-err)); +- goto cleanup; +- } +- } +- obj->links.block_rq_issue = bpf_program__attach(obj->progs.block_rq_issue); +- if (!obj->links.block_rq_issue) { +- err = -errno; +- fprintf(stderr, "failed to attach block_rq_issue: %s\n", strerror(-err)); +- goto cleanup; +- } +- obj->links.block_rq_complete = bpf_program__attach(obj->progs.block_rq_complete); +- if (!obj->links.block_rq_complete) { +- err = -errno; +- fprintf(stderr, "failed to attach block_rq_complete: %s\n", strerror(-err)); ++ err = biosnoop_bpf__attach(obj); ++ if (err) { ++ fprintf(stderr, "failed to attach BPF programs: %d\n", err); + goto cleanup; + } + +diff --git a/libbpf-tools/biostacks.bpf.c b/libbpf-tools/biostacks.bpf.c +index c3950910..0ca69880 100644 +--- a/libbpf-tools/biostacks.bpf.c ++++ b/libbpf-tools/biostacks.bpf.c +@@ -67,20 +67,8 @@ int trace_start(void *ctx, struct request *rq, bool merge_bio) + return 0; + } + +-SEC("fentry/blk_account_io_start") +-int BPF_PROG(blk_account_io_start, struct request *rq) +-{ +- return trace_start(ctx, rq, false); +-} +- +-SEC("kprobe/blk_account_io_merge_bio") +-int BPF_KPROBE(blk_account_io_merge_bio, struct request *rq) +-{ +- return trace_start(ctx, rq, true); +-} +- +-SEC("fentry/blk_account_io_done") +-int BPF_PROG(blk_account_io_done, struct request *rq) ++static __always_inline ++int trace_done(void *ctx, struct request *rq) + { + u64 slot, ts = bpf_ktime_get_ns(); + struct internal_rqinfo *i_rqinfop; +@@ -110,4 +98,34 @@ int BPF_PROG(blk_account_io_done, struct request *rq) + return 0; + } + ++SEC("kprobe/blk_account_io_merge_bio") ++int BPF_KPROBE(blk_account_io_merge_bio, struct request *rq) ++{ ++ return trace_start(ctx, rq, true); ++} ++ ++SEC("fentry/blk_account_io_start") ++int BPF_PROG(blk_account_io_start, struct request *rq) ++{ ++ return trace_start(ctx, rq, false); ++} ++ ++SEC("fentry/blk_account_io_done") ++int BPF_PROG(blk_account_io_done, struct request *rq) ++{ ++ return trace_done(ctx, rq); ++} ++ ++SEC("tp_btf/block_io_start") ++int BPF_PROG(block_io_start, struct request *rq) ++{ ++ return trace_start(ctx, rq, false); ++} ++ ++SEC("tp_btf/block_io_done") ++int BPF_PROG(block_io_done, struct request *rq) ++{ ++ return trace_done(ctx, rq); ++} ++ + char LICENSE[] SEC("license") = "GPL"; +diff --git a/libbpf-tools/biostacks.c b/libbpf-tools/biostacks.c +index e1878d1f..e7875f76 100644 +--- a/libbpf-tools/biostacks.c ++++ b/libbpf-tools/biostacks.c +@@ -128,6 +128,39 @@ void print_map(struct ksyms *ksyms, struct partitions *partitions, int fd) + return; + } + ++static bool has_block_io_tracepoints(void) ++{ ++ return tracepoint_exists("block", "block_io_start") && ++ tracepoint_exists("block", "block_io_done"); ++} ++ ++static void disable_block_io_tracepoints(struct biostacks_bpf *obj) ++{ ++ bpf_program__set_autoload(obj->progs.block_io_start, false); ++ bpf_program__set_autoload(obj->progs.block_io_done, false); ++} ++ ++static void disable_blk_account_io_fentry(struct biostacks_bpf *obj) ++{ ++ bpf_program__set_autoload(obj->progs.blk_account_io_start, false); ++ bpf_program__set_autoload(obj->progs.blk_account_io_done, false); ++} ++ ++static void blk_account_io_set_attach_target(struct biostacks_bpf *obj) ++{ ++ if (fentry_can_attach("blk_account_io_start", NULL)) { ++ bpf_program__set_attach_target(obj->progs.blk_account_io_start, ++ 0, "blk_account_io_start"); ++ bpf_program__set_attach_target(obj->progs.blk_account_io_done, ++ 0, "blk_account_io_done"); ++ } else { ++ bpf_program__set_attach_target(obj->progs.blk_account_io_start, ++ 0, "__blk_account_io_start"); ++ bpf_program__set_attach_target(obj->progs.blk_account_io_done, ++ 0, "__blk_account_io_done"); ++ } ++} ++ + int main(int argc, char **argv) + { + struct partitions *partitions = NULL; +@@ -172,50 +205,30 @@ int main(int argc, char **argv) + + obj->rodata->targ_ms = env.milliseconds; + +- if (fentry_can_attach("blk_account_io_start", NULL)) { +- bpf_program__set_attach_target(obj->progs.blk_account_io_start, 0, +- "blk_account_io_start"); +- bpf_program__set_attach_target(obj->progs.blk_account_io_done, 0, +- "blk_account_io_done"); +- } else { +- bpf_program__set_attach_target(obj->progs.blk_account_io_start, 0, +- "__blk_account_io_start"); +- bpf_program__set_attach_target(obj->progs.blk_account_io_done, 0, +- "__blk_account_io_done"); +- } +- +- err = biostacks_bpf__load(obj); +- if (err) { +- fprintf(stderr, "failed to load BPF object: %d\n", err); +- goto cleanup; ++ if (has_block_io_tracepoints()) ++ disable_blk_account_io_fentry(obj); ++ else { ++ disable_block_io_tracepoints(obj); ++ blk_account_io_set_attach_target(obj); + } + +- obj->links.blk_account_io_start = bpf_program__attach(obj->progs.blk_account_io_start); +- if (!obj->links.blk_account_io_start) { +- err = -errno; +- fprintf(stderr, "failed to attach blk_account_io_start: %s\n", strerror(-err)); +- goto cleanup; +- } + ksyms = ksyms__load(); + if (!ksyms) { + fprintf(stderr, "failed to load kallsyms\n"); + goto cleanup; + } +- if (ksyms__get_symbol(ksyms, "blk_account_io_merge_bio")) { +- obj->links.blk_account_io_merge_bio = +- bpf_program__attach(obj->progs.blk_account_io_merge_bio); +- if (!obj->links.blk_account_io_merge_bio) { +- err = -errno; +- fprintf(stderr, "failed to attach blk_account_io_merge_bio: %s\n", +- strerror(-err)); +- goto cleanup; +- } ++ if (!ksyms__get_symbol(ksyms, "blk_account_io_merge_bio")) ++ bpf_program__set_autoload(obj->progs.blk_account_io_merge_bio, false); ++ ++ err = biostacks_bpf__load(obj); ++ if (err) { ++ fprintf(stderr, "failed to load BPF object: %d\n", err); ++ goto cleanup; + } +- obj->links.blk_account_io_done = bpf_program__attach(obj->progs.blk_account_io_done); +- if (!obj->links.blk_account_io_done) { +- err = -errno; +- fprintf(stderr, "failed to attach blk_account_io_done: %s\n", +- strerror(-err)); ++ ++ err = biostacks_bpf__attach(obj); ++ if (err) { ++ fprintf(stderr, "failed to attach BPF programs: %d\n", err); + goto cleanup; + } + +diff --git a/libbpf-tools/biotop.bpf.c b/libbpf-tools/biotop.bpf.c +index 226e32d3..07631378 100644 +--- a/libbpf-tools/biotop.bpf.c ++++ b/libbpf-tools/biotop.bpf.c +@@ -30,8 +30,8 @@ struct { + __type(value, struct val_t); + } counts SEC(".maps"); + +-SEC("kprobe") +-int BPF_KPROBE(blk_account_io_start, struct request *req) ++static __always_inline ++int trace_start(struct request *req) + { + struct who_t who = {}; + +@@ -56,8 +56,8 @@ int BPF_KPROBE(blk_mq_start_request, struct request *req) + return 0; + } + +-SEC("kprobe") +-int BPF_KPROBE(blk_account_io_done, struct request *req, u64 now) ++static __always_inline ++int trace_done(struct request *req) + { + struct val_t *valp, zero = {}; + struct info_t info = {}; +@@ -103,4 +103,40 @@ int BPF_KPROBE(blk_account_io_done, struct request *req, u64 now) + return 0; + } + ++SEC("kprobe/blk_account_io_start") ++int BPF_KPROBE(blk_account_io_start, struct request *req) ++{ ++ return trace_start(req); ++} ++ ++SEC("kprobe/blk_account_io_done") ++int BPF_KPROBE(blk_account_io_done, struct request *req) ++{ ++ return trace_done(req); ++} ++ ++SEC("kprobe/__blk_account_io_start") ++int BPF_KPROBE(__blk_account_io_start, struct request *req) ++{ ++ return trace_start(req); ++} ++ ++SEC("kprobe/__blk_account_io_done") ++int BPF_KPROBE(__blk_account_io_done, struct request *req) ++{ ++ return trace_done(req); ++} ++ ++SEC("tp_btf/block_io_start") ++int BPF_PROG(block_io_start, struct request *req) ++{ ++ return trace_start(req); ++} ++ ++SEC("tp_btf/block_io_done") ++int BPF_PROG(block_io_done, struct request *req) ++{ ++ return trace_done(req); ++} ++ + char LICENSE[] SEC("license") = "GPL"; +diff --git a/libbpf-tools/biotop.c b/libbpf-tools/biotop.c +index 75484281..5b3a7cf3 100644 +--- a/libbpf-tools/biotop.c ++++ b/libbpf-tools/biotop.c +@@ -354,6 +354,38 @@ static int print_stat(struct biotop_bpf *obj) + return err; + } + ++static bool has_block_io_tracepoints(void) ++{ ++ return tracepoint_exists("block", "block_io_start") && ++ tracepoint_exists("block", "block_io_done"); ++} ++ ++static void disable_block_io_tracepoints(struct biotop_bpf *obj) ++{ ++ bpf_program__set_autoload(obj->progs.block_io_start, false); ++ bpf_program__set_autoload(obj->progs.block_io_done, false); ++} ++ ++static void disable_blk_account_io_kprobes(struct biotop_bpf *obj) ++{ ++ bpf_program__set_autoload(obj->progs.blk_account_io_start, false); ++ bpf_program__set_autoload(obj->progs.blk_account_io_done, false); ++ bpf_program__set_autoload(obj->progs.__blk_account_io_start, false); ++ bpf_program__set_autoload(obj->progs.__blk_account_io_done, false); ++} ++ ++static void blk_account_io_set_autoload(struct biotop_bpf *obj, ++ struct ksyms *ksyms) ++{ ++ if (!ksyms__get_symbol(ksyms, "__blk_account_io_start")) { ++ bpf_program__set_autoload(obj->progs.__blk_account_io_start, false); ++ bpf_program__set_autoload(obj->progs.__blk_account_io_done, false); ++ } else { ++ bpf_program__set_autoload(obj->progs.blk_account_io_start, false); ++ bpf_program__set_autoload(obj->progs.blk_account_io_done, false); ++ } ++} ++ + int main(int argc, char **argv) + { + static const struct argp argp = { +@@ -386,32 +418,19 @@ int main(int argc, char **argv) + goto cleanup; + } + ++ if (has_block_io_tracepoints()) ++ disable_blk_account_io_kprobes(obj); ++ else { ++ disable_block_io_tracepoints(obj); ++ blk_account_io_set_autoload(obj, ksyms); ++ } ++ + err = biotop_bpf__load(obj); + if (err) { + warn("failed to load BPF object: %d\n", err); + goto cleanup; + } + +- if (ksyms__get_symbol(ksyms, "__blk_account_io_start")) +- obj->links.blk_account_io_start = bpf_program__attach_kprobe(obj->progs.blk_account_io_start, false, "__blk_account_io_start"); +- else +- obj->links.blk_account_io_start = bpf_program__attach_kprobe(obj->progs.blk_account_io_start, false, "blk_account_io_start"); +- +- if (!obj->links.blk_account_io_start) { +- warn("failed to load attach blk_account_io_start\n"); +- goto cleanup; +- } +- +- if (ksyms__get_symbol(ksyms, "__blk_account_io_done")) +- obj->links.blk_account_io_done = bpf_program__attach_kprobe(obj->progs.blk_account_io_done, false, "__blk_account_io_done"); +- else +- obj->links.blk_account_io_done = bpf_program__attach_kprobe(obj->progs.blk_account_io_done, false, "blk_account_io_done"); +- +- if (!obj->links.blk_account_io_done) { +- warn("failed to load attach blk_account_io_done\n"); +- goto cleanup; +- } +- + err = biotop_bpf__attach(obj); + if (err) { + warn("failed to attach BPF programs: %d\n", err); +-- +2.41.0 + diff --git a/SOURCES/bcc-0.28.0-tools-Add-support-for-the-new-block_io_-tracepoints.patch b/SOURCES/bcc-0.28.0-tools-Add-support-for-the-new-block_io_-tracepoints.patch new file mode 100644 index 0000000..e2b4da7 --- /dev/null +++ b/SOURCES/bcc-0.28.0-tools-Add-support-for-the-new-block_io_-tracepoints.patch @@ -0,0 +1,855 @@ +From 2e758b65231f976c67a0aad791aabc7927ea7086 Mon Sep 17 00:00:00 2001 +From: Jerome Marchand +Date: Thu, 27 Jul 2023 18:19:18 +0200 +Subject: [PATCH] tools: Add support for the new block_io_* tracepoints + +The bio tools currently depends on blk_account_io_done/start functions +that can be inlined. To fix that, a couple of tracepoints have been +added upstream (block:block_io_start/done). This patch add the support +for those tracepoints when they are available. + +Unfortunately, the bio tools relies on data that is not available to +the tracepoints (mostly the struct request). So the tracepoints can't +be used as drop in replacement for blk_account_io_*. Main difference, +is that we can't use the struct request as the hash key anymore, so it +now uses the couple (dev_t, sector) for that purpose. + +For the biolatency tool, the -F option is disabled when only the +tracepoints are available because the flags are not all accessible +from the tracepoints. Otherwise, all features of the tools should +remain. + +Closes #4261 + +Signed-off-by: Jerome Marchand +--- + tools/biolatency.py | 166 ++++++++++++++++++++++++++++-------- + tools/biosnoop.py | 200 +++++++++++++++++++++++++++++++++----------- + tools/biotop.py | 108 +++++++++++++++++++----- + 3 files changed, 371 insertions(+), 103 deletions(-) + +diff --git a/tools/biolatency.py b/tools/biolatency.py +index 8fe43a7c..03b48a4c 100755 +--- a/tools/biolatency.py ++++ b/tools/biolatency.py +@@ -11,6 +11,7 @@ + # + # 20-Sep-2015 Brendan Gregg Created this. + # 31-Mar-2022 Rocky Xing Added disk filter support. ++# 01-Aug-2023 Jerome Marchand Added support for block tracepoints + + from __future__ import print_function + from bcc import BPF +@@ -72,7 +73,7 @@ bpf_text = """ + #include + + typedef struct disk_key { +- char disk[DISK_NAME_LEN]; ++ dev_t dev; + u64 slot; + } disk_key_t; + +@@ -86,26 +87,70 @@ typedef struct ext_val { + u64 count; + } ext_val_t; + +-BPF_HASH(start, struct request *); ++struct tp_args { ++ u64 __unused__; ++ dev_t dev; ++ sector_t sector; ++ unsigned int nr_sector; ++ unsigned int bytes; ++ char rwbs[8]; ++ char comm[16]; ++ char cmd[]; ++}; ++ ++struct start_key { ++ dev_t dev; ++ u32 _pad; ++ sector_t sector; ++ CMD_FLAGS ++}; ++ ++BPF_HASH(start, struct start_key); + STORAGE + ++static dev_t ddevt(struct gendisk *disk) { ++ return (disk->major << 20) | disk->first_minor; ++} ++ + // time block I/O +-int trace_req_start(struct pt_regs *ctx, struct request *req) ++static int __trace_req_start(struct start_key key) + { + DISK_FILTER + + u64 ts = bpf_ktime_get_ns(); +- start.update(&req, &ts); ++ start.update(&key, &ts); + return 0; + } + ++int trace_req_start(struct pt_regs *ctx, struct request *req) ++{ ++ struct start_key key = { ++ .dev = ddevt(req->__RQ_DISK__), ++ .sector = req->__sector ++ }; ++ ++ SET_FLAGS ++ ++ return __trace_req_start(key); ++} ++ ++int trace_req_start_tp(struct tp_args *args) ++{ ++ struct start_key key = { ++ .dev = args->dev, ++ .sector = args->sector ++ }; ++ ++ return __trace_req_start(key); ++} ++ + // output +-int trace_req_done(struct pt_regs *ctx, struct request *req) ++static int __trace_req_done(struct start_key key) + { + u64 *tsp, delta; + + // fetch timestamp and calculate delta +- tsp = start.lookup(&req); ++ tsp = start.lookup(&key); + if (tsp == 0) { + return 0; // missed issue + } +@@ -116,9 +161,31 @@ int trace_req_done(struct pt_regs *ctx, struct request *req) + // store as histogram + STORE + +- start.delete(&req); ++ start.delete(&key); + return 0; + } ++ ++int trace_req_done(struct pt_regs *ctx, struct request *req) ++{ ++ struct start_key key = { ++ .dev = ddevt(req->__RQ_DISK__), ++ .sector = req->__sector ++ }; ++ ++ SET_FLAGS ++ ++ return __trace_req_done(key); ++} ++ ++int trace_req_done_tp(struct tp_args *args) ++{ ++ struct start_key key = { ++ .dev = args->dev, ++ .sector = args->sector ++ }; ++ ++ return __trace_req_done(key); ++} + """ + + # code substitutions +@@ -134,21 +201,18 @@ store_str = "" + if args.disks: + storage_str += "BPF_HISTOGRAM(dist, disk_key_t);" + disks_str = """ +- disk_key_t key = {.slot = bpf_log2l(delta)}; +- void *__tmp = (void *)req->__RQ_DISK__->disk_name; +- bpf_probe_read(&key.disk, sizeof(key.disk), __tmp); +- dist.atomic_increment(key); ++ disk_key_t dkey = {}; ++ dkey.dev = key.dev; ++ dkey.slot = bpf_log2l(delta); ++ dist.atomic_increment(dkey); + """ +- if BPF.kernel_struct_has_field(b'request', b'rq_disk') == 1: +- store_str += disks_str.replace('__RQ_DISK__', 'rq_disk') +- else: +- store_str += disks_str.replace('__RQ_DISK__', 'q->disk') ++ store_str += disks_str + elif args.flags: + storage_str += "BPF_HISTOGRAM(dist, flag_key_t);" + store_str += """ +- flag_key_t key = {.slot = bpf_log2l(delta)}; +- key.flags = req->cmd_flags; +- dist.atomic_increment(key); ++ flag_key_t fkey = {.slot = bpf_log2l(delta)}; ++ fkey.flags = key.flags; ++ dist.atomic_increment(fkey); + """ + else: + storage_str += "BPF_HISTOGRAM(dist);" +@@ -161,21 +225,13 @@ store_str = "" + exit(1) + + stat_info = os.stat(disk_path) +- major = os.major(stat_info.st_rdev) +- minor = os.minor(stat_info.st_rdev) +- +- disk_field_str = "" +- if BPF.kernel_struct_has_field(b'request', b'rq_disk') == 1: +- disk_field_str = 'req->rq_disk' +- else: +- disk_field_str = 'req->q->disk' ++ dev = os.major(stat_info.st_rdev) << 20 | os.minor(stat_info.st_rdev) + + disk_filter_str = """ +- struct gendisk *disk = %s; +- if (!(disk->major == %d && disk->first_minor == %d)) { ++ if(key.dev != %s) { + return 0; + } +- """ % (disk_field_str, major, minor) ++ """ % (dev) + + bpf_text = bpf_text.replace('DISK_FILTER', disk_filter_str) + else: +@@ -194,6 +250,16 @@ store_str = "" + + bpf_text = bpf_text.replace("STORAGE", storage_str) + bpf_text = bpf_text.replace("STORE", store_str) ++if BPF.kernel_struct_has_field(b'request', b'rq_disk') == 1: ++ bpf_text = bpf_text.replace('__RQ_DISK__', 'rq_disk') ++else: ++ bpf_text = bpf_text.replace('__RQ_DISK__', 'q->disk') ++if args.flags: ++ bpf_text = bpf_text.replace('CMD_FLAGS', 'u64 flags;') ++ bpf_text = bpf_text.replace('SET_FLAGS', 'key.flags = req->cmd_flags;') ++else: ++ bpf_text = bpf_text.replace('CMD_FLAGS', '') ++ bpf_text = bpf_text.replace('SET_FLAGS', '') + + if debug or args.ebpf: + print(bpf_text) +@@ -205,25 +271,53 @@ b = BPF(text=bpf_text) + if args.queued: + if BPF.get_kprobe_functions(b'__blk_account_io_start'): + b.attach_kprobe(event="__blk_account_io_start", fn_name="trace_req_start") +- else: ++ elif BPF.get_kprobe_functions(b'blk_account_io_start'): + b.attach_kprobe(event="blk_account_io_start", fn_name="trace_req_start") ++ else: ++ if args.flags: ++ # Some flags are accessible in the rwbs field (RAHEAD, SYNC and META) ++ # but other aren't. Disable the -F option for tracepoint for now. ++ print("ERROR: blk_account_io_start probe not available. Can't use -F.") ++ exit() ++ b.attach_tracepoint(tp="block:block_io_start", fn_name="trace_req_start_tp") + else: + if BPF.get_kprobe_functions(b'blk_start_request'): + b.attach_kprobe(event="blk_start_request", fn_name="trace_req_start") + b.attach_kprobe(event="blk_mq_start_request", fn_name="trace_req_start") ++ + if BPF.get_kprobe_functions(b'__blk_account_io_done'): + b.attach_kprobe(event="__blk_account_io_done", fn_name="trace_req_done") +-else: ++elif BPF.get_kprobe_functions(b'blk_account_io_done'): + b.attach_kprobe(event="blk_account_io_done", fn_name="trace_req_done") ++else: ++ if args.flags: ++ print("ERROR: blk_account_io_done probe not available. Can't use -F.") ++ exit() ++ b.attach_tracepoint(tp="block:block_io_done", fn_name="trace_req_done_tp") ++ + + if not args.json: + print("Tracing block device I/O... Hit Ctrl-C to end.") + +-def disk_print(s): +- disk = s.decode('utf-8', 'replace') +- if not disk: +- disk = "" +- return disk ++# cache disk major,minor -> diskname ++diskstats = "/proc/diskstats" ++disklookup = {} ++with open(diskstats) as stats: ++ for line in stats: ++ a = line.split() ++ disklookup[a[0] + "," + a[1]] = a[2] ++ ++def disk_print(d): ++ major = d >> 20 ++ minor = d & ((1 << 20) - 1) ++ ++ disk = str(major) + "," + str(minor) ++ if disk in disklookup: ++ diskname = disklookup[disk] ++ else: ++ diskname = "?" ++ ++ return diskname + + # see blk_fill_rwbs(): + req_opf = { +diff --git a/tools/biosnoop.py b/tools/biosnoop.py +index 33703233..f0fef98b 100755 +--- a/tools/biosnoop.py ++++ b/tools/biosnoop.py +@@ -14,6 +14,7 @@ + # 11-Feb-2016 Allan McAleavy updated for BPF_PERF_OUTPUT + # 21-Jun-2022 Rocky Xing Added disk filter support. + # 13-Oct-2022 Rocky Xing Added support for displaying block I/O pattern. ++# 01-Aug-2023 Jerome Marchand Added support for block tracepoints + + from __future__ import print_function + from bcc import BPF +@@ -64,6 +65,24 @@ struct val_t { + char name[TASK_COMM_LEN]; + }; + ++struct tp_args { ++ u64 __unused__; ++ dev_t dev; ++ sector_t sector; ++ unsigned int nr_sector; ++ unsigned int bytes; ++ char rwbs[8]; ++ char comm[16]; ++ char cmd[]; ++}; ++ ++struct hash_key { ++ dev_t dev; ++ u32 rwflag; ++ sector_t sector; ++}; ++ ++ + #ifdef INCLUDE_PATTERN + struct sector_key_t { + u32 dev_major; +@@ -79,6 +98,7 @@ enum bio_pattern { + + struct data_t { + u32 pid; ++ u32 dev; + u64 rwflag; + u64 delta; + u64 qdelta; +@@ -88,7 +108,6 @@ struct data_t { + enum bio_pattern pattern; + #endif + u64 ts; +- char disk_name[DISK_NAME_LEN]; + char name[TASK_COMM_LEN]; + }; + +@@ -96,12 +115,45 @@ struct data_t { + BPF_HASH(last_sectors, struct sector_key_t, u64); + #endif + +-BPF_HASH(start, struct request *, struct start_req_t); +-BPF_HASH(infobyreq, struct request *, struct val_t); ++BPF_HASH(start, struct hash_key, struct start_req_t); ++BPF_HASH(infobyreq, struct hash_key, struct val_t); + BPF_PERF_OUTPUT(events); + ++static dev_t ddevt(struct gendisk *disk) { ++ return (disk->major << 20) | disk->first_minor; ++} ++ ++/* ++ * The following deals with a kernel version change (in mainline 4.7, although ++ * it may be backported to earlier kernels) with how block request write flags ++ * are tested. We handle both pre- and post-change versions here. Please avoid ++ * kernel version tests like this as much as possible: they inflate the code, ++ * test, and maintenance burden. ++ */ ++static int get_rwflag(u32 cmd_flags) { ++#ifdef REQ_WRITE ++ return !!(cmd_flags & REQ_WRITE); ++#elif defined(REQ_OP_SHIFT) ++ return !!((cmd_flags >> REQ_OP_SHIFT) == REQ_OP_WRITE); ++#else ++ return !!((cmd_flags & REQ_OP_MASK) == REQ_OP_WRITE); ++#endif ++} ++ ++#define RWBS_LEN 8 ++ ++static int get_rwflag_tp(char *rwbs) { ++ for (int i = 0; i < RWBS_LEN; i++) { ++ if (rwbs[i] == 'W') ++ return 1; ++ if (rwbs[i] == '\\0') ++ return 0; ++ } ++ return 0; ++} ++ + // cache PID and comm by-req +-int trace_pid_start(struct pt_regs *ctx, struct request *req) ++static int __trace_pid_start(struct hash_key key) + { + DISK_FILTER + +@@ -113,47 +165,76 @@ int trace_pid_start(struct pt_regs *ctx, struct request *req) + if (##QUEUE##) { + val.ts = bpf_ktime_get_ns(); + } +- infobyreq.update(&req, &val); ++ infobyreq.update(&key, &val); + } + return 0; + } + ++ ++int trace_pid_start(struct pt_regs *ctx, struct request *req) ++{ ++ struct hash_key key = { ++ .dev = ddevt(req->__RQ_DISK__), ++ .rwflag = get_rwflag(req->cmd_flags), ++ .sector = req->__sector ++ }; ++ ++ return __trace_pid_start(key); ++} ++ ++int trace_pid_start_tp(struct tp_args *args) ++{ ++ struct hash_key key = { ++ .dev = args->dev, ++ .rwflag = get_rwflag_tp(args->rwbs), ++ .sector = args->sector ++ }; ++ ++ return __trace_pid_start(key); ++} ++ + // time block I/O + int trace_req_start(struct pt_regs *ctx, struct request *req) + { ++ struct hash_key key = { ++ .dev = ddevt(req->__RQ_DISK__), ++ .rwflag = get_rwflag(req->cmd_flags), ++ .sector = req->__sector ++ }; ++ + DISK_FILTER + + struct start_req_t start_req = { + .ts = bpf_ktime_get_ns(), + .data_len = req->__data_len + }; +- start.update(&req, &start_req); ++ start.update(&key, &start_req); + return 0; + } + + // output +-int trace_req_completion(struct pt_regs *ctx, struct request *req) ++static int __trace_req_completion(void *ctx, struct hash_key key) + { + struct start_req_t *startp; + struct val_t *valp; + struct data_t data = {}; +- struct gendisk *rq_disk; ++ //struct gendisk *rq_disk; + u64 ts; + + // fetch timestamp and calculate delta +- startp = start.lookup(&req); ++ startp = start.lookup(&key); + if (startp == 0) { + // missed tracing issue + return 0; + } + ts = bpf_ktime_get_ns(); +- rq_disk = req->__RQ_DISK__; ++ //rq_disk = req->__RQ_DISK__; + data.delta = ts - startp->ts; + data.ts = ts / 1000; + data.qdelta = 0; + data.len = startp->data_len; + +- valp = infobyreq.lookup(&req); ++ valp = infobyreq.lookup(&key); + if (valp == 0) { + data.name[0] = '?'; + data.name[1] = 0; +@@ -162,10 +243,9 @@ int trace_req_completion(struct pt_regs *ctx, struct request *req) + data.qdelta = startp->ts - valp->ts; + } + data.pid = valp->pid; +- data.sector = req->__sector; ++ data.sector = key.sector; ++ data.dev = key.dev; + bpf_probe_read_kernel(&data.name, sizeof(data.name), valp->name); +- bpf_probe_read_kernel(&data.disk_name, sizeof(data.disk_name), +- rq_disk->disk_name); + } + + #ifdef INCLUDE_PATTERN +@@ -174,8 +254,8 @@ int trace_req_completion(struct pt_regs *ctx, struct request *req) + u64 *sector, last_sector; + + struct sector_key_t sector_key = { +- .dev_major = rq_disk->major, +- .dev_minor = rq_disk->first_minor ++ .dev_major = key.dev >> 20, ++ .dev_minor = key.dev & ((1 << 20) - 1) + }; + + sector = last_sectors.lookup(§or_key); +@@ -187,27 +267,36 @@ int trace_req_completion(struct pt_regs *ctx, struct request *req) + last_sectors.update(§or_key, &last_sector); + #endif + +-/* +- * The following deals with a kernel version change (in mainline 4.7, although +- * it may be backported to earlier kernels) with how block request write flags +- * are tested. We handle both pre- and post-change versions here. Please avoid +- * kernel version tests like this as much as possible: they inflate the code, +- * test, and maintenance burden. +- */ +-#ifdef REQ_WRITE +- data.rwflag = !!(req->cmd_flags & REQ_WRITE); +-#elif defined(REQ_OP_SHIFT) +- data.rwflag = !!((req->cmd_flags >> REQ_OP_SHIFT) == REQ_OP_WRITE); +-#else +- data.rwflag = !!((req->cmd_flags & REQ_OP_MASK) == REQ_OP_WRITE); +-#endif ++ data.rwflag = key.rwflag; + + events.perf_submit(ctx, &data, sizeof(data)); +- start.delete(&req); +- infobyreq.delete(&req); ++ start.delete(&key); ++ infobyreq.delete(&key); + + return 0; + } ++ ++int trace_req_completion(struct pt_regs *ctx, struct request *req) ++{ ++ struct hash_key key = { ++ .dev = ddevt(req->__RQ_DISK__), ++ .rwflag = get_rwflag(req->cmd_flags), ++ .sector = req->__sector ++ }; ++ ++ return __trace_req_completion(ctx, key); ++} ++ ++int trace_req_completion_tp(struct tp_args *args) ++{ ++ struct hash_key key = { ++ .dev = args->dev, ++ .rwflag = get_rwflag_tp(args->rwbs), ++ .sector = args->sector ++ }; ++ ++ return __trace_req_completion(args, key); ++} + """ + if args.queue: + bpf_text = bpf_text.replace('##QUEUE##', '1') +@@ -225,21 +314,13 @@ int trace_req_completion(struct pt_regs *ctx, struct request *req) + exit(1) + + stat_info = os.stat(disk_path) +- major = os.major(stat_info.st_rdev) +- minor = os.minor(stat_info.st_rdev) +- +- disk_field_str = "" +- if BPF.kernel_struct_has_field(b'request', b'rq_disk') == 1: +- disk_field_str = 'req->rq_disk' +- else: +- disk_field_str = 'req->q->disk' ++ dev = os.major(stat_info.st_rdev) << 20 | os.minor(stat_info.st_rdev) + + disk_filter_str = """ +- struct gendisk *disk = %s; +- if (!(disk->major == %d && disk->first_minor == %d)) { ++ if(key.dev != %s) { + return 0; + } +- """ % (disk_field_str, major, minor) ++ """ % (dev) + + bpf_text = bpf_text.replace('DISK_FILTER', disk_filter_str) + else: +@@ -254,15 +335,19 @@ int trace_req_completion(struct pt_regs *ctx, struct request *req) + b = BPF(text=bpf_text) + if BPF.get_kprobe_functions(b'__blk_account_io_start'): + b.attach_kprobe(event="__blk_account_io_start", fn_name="trace_pid_start") +-else: ++elif BPF.get_kprobe_functions(b'blk_account_io_start'): + b.attach_kprobe(event="blk_account_io_start", fn_name="trace_pid_start") ++else: ++ b.attach_tracepoint(tp="block:block_io_start", fn_name="trace_pid_start_tp") + if BPF.get_kprobe_functions(b'blk_start_request'): + b.attach_kprobe(event="blk_start_request", fn_name="trace_req_start") + b.attach_kprobe(event="blk_mq_start_request", fn_name="trace_req_start") + if BPF.get_kprobe_functions(b'__blk_account_io_done'): + b.attach_kprobe(event="__blk_account_io_done", fn_name="trace_req_completion") +-else: ++elif BPF.get_kprobe_functions(b'blk_account_io_done'): + b.attach_kprobe(event="blk_account_io_done", fn_name="trace_req_completion") ++else: ++ b.attach_tracepoint(tp="block:block_io_done", fn_name="trace_req_completion_tp") + + # header + print("%-11s %-14s %-7s %-9s %-1s %-10s %-7s" % ("TIME(s)", "COMM", "PID", +@@ -273,6 +358,27 @@ print("%-11s %-14s %-7s %-9s %-1s %-10s %-7s" % ("TIME(s)", "COMM", "PID", + print("%7s " % ("QUE(ms)"), end="") + print("%7s" % "LAT(ms)") + ++ ++# cache disk major,minor -> diskname ++diskstats = "/proc/diskstats" ++disklookup = {} ++with open(diskstats) as stats: ++ for line in stats: ++ a = line.split() ++ disklookup[a[0] + "," + a[1]] = a[2] ++ ++def disk_print(d): ++ major = d >> 20 ++ minor = d & ((1 << 20) - 1) ++ ++ disk = str(major) + "," + str(minor) ++ if disk in disklookup: ++ diskname = disklookup[disk] ++ else: ++ diskname = "" ++ ++ return diskname ++ + rwflg = "" + pattern = "" + start_ts = 0 +@@ -297,9 +403,7 @@ P_RANDOM = 2 + + delta = float(event.ts) - start_ts + +- disk_name = event.disk_name.decode('utf-8', 'replace') +- if not disk_name: +- disk_name = '' ++ disk_name = disk_print(event.dev) + + print("%-11.6f %-14.14s %-7s %-9s %-1s %-10s %-7s" % ( + delta / 1000000, event.name.decode('utf-8', 'replace'), event.pid, +diff --git a/tools/biotop.py b/tools/biotop.py +index fcdd373f..2620983a 100755 +--- a/tools/biotop.py ++++ b/tools/biotop.py +@@ -14,6 +14,7 @@ + # + # 06-Feb-2016 Brendan Gregg Created this. + # 17-Mar-2022 Rocky Xing Added PID filter support. ++# 01-Aug-2023 Jerome Marchand Added support for block tracepoints + + from __future__ import print_function + from bcc import BPF +@@ -88,14 +89,35 @@ struct val_t { + u32 io; + }; + +-BPF_HASH(start, struct request *, struct start_req_t); +-BPF_HASH(whobyreq, struct request *, struct who_t); ++struct tp_args { ++ u64 __unused__; ++ dev_t dev; ++ sector_t sector; ++ unsigned int nr_sector; ++ unsigned int bytes; ++ char rwbs[8]; ++ char comm[16]; ++ char cmd[]; ++}; ++ ++struct hash_key { ++ dev_t dev; ++ u32 _pad; ++ sector_t sector; ++}; ++ ++BPF_HASH(start, struct hash_key, struct start_req_t); ++BPF_HASH(whobyreq, struct hash_key, struct who_t); + BPF_HASH(counts, struct info_t, struct val_t); + ++static dev_t ddevt(struct gendisk *disk) { ++ return (disk->major << 20) | disk->first_minor; ++} ++ + // cache PID and comm by-req +-int trace_pid_start(struct pt_regs *ctx, struct request *req) ++static int __trace_pid_start(struct hash_key key) + { +- struct who_t who = {}; ++ struct who_t who; + u32 pid; + + if (bpf_get_current_comm(&who.name, sizeof(who.name)) == 0) { +@@ -104,30 +126,54 @@ int trace_pid_start(struct pt_regs *ctx, struct request *req) + return 0; + + who.pid = pid; +- whobyreq.update(&req, &who); ++ whobyreq.update(&key, &who); + } + + return 0; + } + ++int trace_pid_start(struct pt_regs *ctx, struct request *req) ++{ ++ struct hash_key key = { ++ .dev = ddevt(req->__RQ_DISK__), ++ .sector = req->__sector ++ }; ++ ++ return __trace_pid_start(key); ++} ++ ++int trace_pid_start_tp(struct tp_args *args) ++{ ++ struct hash_key key = { ++ .dev = args->dev, ++ .sector = args->sector ++ }; ++ ++ return __trace_pid_start(key); ++} ++ + // time block I/O + int trace_req_start(struct pt_regs *ctx, struct request *req) + { ++ struct hash_key key = { ++ .dev = ddevt(req->__RQ_DISK__), ++ .sector = req->__sector ++ }; + struct start_req_t start_req = { + .ts = bpf_ktime_get_ns(), + .data_len = req->__data_len + }; +- start.update(&req, &start_req); ++ start.update(&key, &start_req); + return 0; + } + + // output +-int trace_req_completion(struct pt_regs *ctx, struct request *req) ++static int __trace_req_completion(struct hash_key key) + { + struct start_req_t *startp; + + // fetch timestamp and calculate delta +- startp = start.lookup(&req); ++ startp = start.lookup(&key); + if (startp == 0) { + return 0; // missed tracing issue + } +@@ -135,12 +181,12 @@ int trace_req_completion(struct pt_regs *ctx, struct request *req) + struct who_t *whop; + u32 pid; + +- whop = whobyreq.lookup(&req); ++ whop = whobyreq.lookup(&key); + pid = whop != 0 ? whop->pid : 0; + if (FILTER_PID) { +- start.delete(&req); ++ start.delete(&key); + if (whop != 0) { +- whobyreq.delete(&req); ++ whobyreq.delete(&key); + } + return 0; + } +@@ -150,8 +196,8 @@ int trace_req_completion(struct pt_regs *ctx, struct request *req) + + // setup info_t key + struct info_t info = {}; +- info.major = req->__RQ_DISK__->major; +- info.minor = req->__RQ_DISK__->first_minor; ++ info.major = key.dev >> 20; ++ info.minor = key.dev & ((1 << 20) - 1); + /* + * The following deals with a kernel version change (in mainline 4.7, although + * it may be backported to earlier kernels) with how block request write flags +@@ -159,13 +205,13 @@ int trace_req_completion(struct pt_regs *ctx, struct request *req) + * kernel version tests like this as much as possible: they inflate the code, + * test, and maintenance burden. + */ +-#ifdef REQ_WRITE ++/*#ifdef REQ_WRITE + info.rwflag = !!(req->cmd_flags & REQ_WRITE); + #elif defined(REQ_OP_SHIFT) + info.rwflag = !!((req->cmd_flags >> REQ_OP_SHIFT) == REQ_OP_WRITE); + #else + info.rwflag = !!((req->cmd_flags & REQ_OP_MASK) == REQ_OP_WRITE); +-#endif ++#endif*/ + + if (whop == 0) { + // missed pid who, save stats as pid 0 +@@ -183,11 +229,31 @@ int trace_req_completion(struct pt_regs *ctx, struct request *req) + valp->io++; + } + +- start.delete(&req); +- whobyreq.delete(&req); ++ start.delete(&key); ++ whobyreq.delete(&key); + + return 0; + } ++ ++int trace_req_completion(struct pt_regs *ctx, struct request *req) ++{ ++ struct hash_key key = { ++ .dev = ddevt(req->__RQ_DISK__), ++ .sector = req->__sector ++ }; ++ ++ return __trace_req_completion(key); ++} ++ ++int trace_req_completion_tp(struct tp_args *args) ++{ ++ struct hash_key key = { ++ .dev = args->dev, ++ .sector = args->sector ++ }; ++ ++ return __trace_req_completion(key); ++} + """ + + if args.ebpf: +@@ -207,15 +273,19 @@ int trace_req_completion(struct pt_regs *ctx, struct request *req) + b = BPF(text=bpf_text) + if BPF.get_kprobe_functions(b'__blk_account_io_start'): + b.attach_kprobe(event="__blk_account_io_start", fn_name="trace_pid_start") +-else: ++elif BPF.get_kprobe_functions(b'blk_account_io_start'): + b.attach_kprobe(event="blk_account_io_start", fn_name="trace_pid_start") ++else: ++ b.attach_tracepoint(tp="block:block_io_start", fn_name="trace_pid_start_tp") + if BPF.get_kprobe_functions(b'blk_start_request'): + b.attach_kprobe(event="blk_start_request", fn_name="trace_req_start") + b.attach_kprobe(event="blk_mq_start_request", fn_name="trace_req_start") + if BPF.get_kprobe_functions(b'__blk_account_io_done'): + b.attach_kprobe(event="__blk_account_io_done", fn_name="trace_req_completion") +-else: ++elif BPF.get_kprobe_functions(b'blk_account_io_done'): + b.attach_kprobe(event="blk_account_io_done", fn_name="trace_req_completion") ++else: ++ b.attach_tracepoint(tp="block:block_io_done", fn_name="trace_req_completion_tp") + + print('Tracing... Output every %d secs. Hit Ctrl-C to end' % interval) + +-- +2.41.0 + diff --git a/SOURCES/bcc-0.28.0-tools-tcpstates-fix-IPv6-journal.patch b/SOURCES/bcc-0.28.0-tools-tcpstates-fix-IPv6-journal.patch new file mode 100644 index 0000000..d62c794 --- /dev/null +++ b/SOURCES/bcc-0.28.0-tools-tcpstates-fix-IPv6-journal.patch @@ -0,0 +1,156 @@ +From 0d1a67ba9490aabbb874819d8d07b1868c8c2b1d Mon Sep 17 00:00:00 2001 +From: Jerome Marchand +Date: Wed, 1 Feb 2023 17:30:03 +0100 +Subject: [PATCH 2/2] tools/tcpstates: fix IPv6 journal + +When logging ipv6 state change, journal_fields tries to pack +event.addr and event.daddr, which is not an integer in this, to +present a bytes-like object to socket.inet_ntop. This can be fixed by +having a similar type for [sd]addr for IPv4 and IPv6. Making both an +array of u32 solves the issue by presenting a bytes-like object +directly to inet_ntop, without the need for the struct packing stage. + +Also now, the similar behavior, makes it easier to factor code for +IPv4 and IPv6. + +It solves the following error: +/usr/share/bcc/tools/tcpstates -Y +SKADDR C-PID C-COMM LADDR LPORT RADDR RPORT OLDSTATE -> NEWSTATE MS +ffff8b2e83e56180 0 swapper/9 :: 22 :: 0 LISTEN -> SYN_RECV 0.000 +Exception ignored on calling ctypes callback function: .raw_cb_ at 0x7f894c8d7f70> +Traceback (most recent call last): + File "/usr/lib/python3.9/site-packages/bcc/table.py", line 982, in raw_cb_ + callback(cpu, data, size) + File "/usr/share/bcc/tools/tcpstates", line 419, in print_ipv6_event + journal.send(**journal_fields(event, AF_INET6)) + File "/usr/share/bcc/tools/tcpstates", line 348, in journal_fields + 'OBJECT_' + addr_pfx + '_SOURCE_ADDRESS': inet_ntop(addr_family, pack("I", event.saddr)), +struct.error: required argument is not an integer +ffff8b2e83e56180 0 swapper/9 2620:52:0:2580:5054:ff:fe6b:6f1f 22 2620:52:0:2b11:2f5e:407d:b35d:4663 60396 SYN_RECV -> ESTABLISHED 0.010 +Exception ignored on calling ctypes callback function: .raw_cb_ at 0x7f894c8d7f70> +Traceback (most recent call last): + File "/usr/lib/python3.9/site-packages/bcc/table.py", line 982, in raw_cb_ + callback(cpu, data, size) + File "/usr/share/bcc/tools/tcpstates", line 419, in print_ipv6_event + journal.send(**journal_fields(event, AF_INET6)) + File "/usr/share/bcc/tools/tcpstates", line 348, in journal_fields + 'OBJECT_' + addr_pfx + '_SOURCE_ADDRESS': inet_ntop(addr_family, pack("I", event.saddr)), +struct.error: required argument is not an integer + +Signed-off-by: Jerome Marchand +--- + tools/tcpstates.py | 55 +++++++++++++++++----------------------------- + 1 file changed, 20 insertions(+), 35 deletions(-) + +diff --git a/tools/tcpstates.py b/tools/tcpstates.py +index 9b2ccfa4..6c845c9b 100755 +--- a/tools/tcpstates.py ++++ b/tools/tcpstates.py +@@ -19,7 +19,6 @@ from __future__ import print_function + from bcc import BPF + import argparse + from socket import inet_ntop, AF_INET, AF_INET6 +-from struct import pack + from time import strftime, time + from os import getuid + +@@ -78,8 +77,8 @@ BPF_HASH(last, struct sock *, u64); + struct ipv4_data_t { + u64 ts_us; + u64 skaddr; +- u32 saddr; +- u32 daddr; ++ u32 saddr[1]; ++ u32 daddr[1]; + u64 span_us; + u32 pid; + u16 lport; +@@ -93,8 +92,8 @@ BPF_PERF_OUTPUT(ipv4_events); + struct ipv6_data_t { + u64 ts_us; + u64 skaddr; +- unsigned __int128 saddr; +- unsigned __int128 daddr; ++ u32 saddr[4]; ++ u32 daddr[4]; + u64 span_us; + u32 pid; + u16 lport; +@@ -350,9 +349,9 @@ format_string = ("%-16x %-5d %-10.10s %s%-15s %-5d %-15s %-5d %-11s " + + 'OBJECT_PID': str(event.pid), + 'OBJECT_COMM': event.task.decode('utf-8', 'replace'), + # Custom fields, aka "stuff we sort of made up". +- 'OBJECT_' + addr_pfx + '_SOURCE_ADDRESS': inet_ntop(addr_family, pack("I", event.saddr)), ++ 'OBJECT_' + addr_pfx + '_SOURCE_ADDRESS': inet_ntop(addr_family, event.saddr), + 'OBJECT_TCP_SOURCE_PORT': str(event.lport), +- 'OBJECT_' + addr_pfx + '_DESTINATION_ADDRESS': inet_ntop(addr_family, pack("I", event.daddr)), ++ 'OBJECT_' + addr_pfx + '_DESTINATION_ADDRESS': inet_ntop(addr_family, event.daddr), + 'OBJECT_TCP_DESTINATION_PORT': str(event.dport), + 'OBJECT_TCP_OLD_STATE': tcpstate2str(event.oldstate), + 'OBJECT_TCP_NEW_STATE': tcpstate2str(event.newstate), +@@ -373,8 +372,7 @@ format_string = ("%-16x %-5d %-10.10s %s%-15s %-5d %-15s %-5d %-11s " + + return fields + + # process event +-def print_ipv4_event(cpu, data, size): +- event = b["ipv4_events"].event(data) ++def print_event(event, addr_family): + global start_ts + if args.time: + if args.csv: +@@ -389,39 +387,26 @@ format_string = ("%-16x %-5d %-10.10s %s%-15s %-5d %-15s %-5d %-11s " + + print("%.6f," % delta_s, end="") + else: + print("%-9.6f " % delta_s, end="") ++ if addr_family == AF_INET: ++ version = "4" ++ else: ++ version = "6" + print(format_string % (event.skaddr, event.pid, event.task.decode('utf-8', 'replace'), +- "4" if args.wide or args.csv else "", +- inet_ntop(AF_INET, pack("I", event.saddr)), event.lport, +- inet_ntop(AF_INET, pack("I", event.daddr)), event.dport, ++ version if args.wide or args.csv else "", ++ inet_ntop(addr_family, event.saddr), event.lport, ++ inet_ntop(addr_family, event.daddr), event.dport, + tcpstate2str(event.oldstate), tcpstate2str(event.newstate), + float(event.span_us) / 1000)) + if args.journal: +- journal.send(**journal_fields(event, AF_INET)) ++ journal.send(**journal_fields(event, addr_family)) ++ ++def print_ipv4_event(cpu, data, size): ++ event = b["ipv4_events"].event(data) ++ print_event(event, AF_INET) + + def print_ipv6_event(cpu, data, size): + event = b["ipv6_events"].event(data) +- global start_ts +- if args.time: +- if args.csv: +- print("%s," % strftime("%H:%M:%S"), end="") +- else: +- print("%-8s " % strftime("%H:%M:%S"), end="") +- if args.timestamp: +- if start_ts == 0: +- start_ts = event.ts_us +- delta_s = (float(event.ts_us) - start_ts) / 1000000 +- if args.csv: +- print("%.6f," % delta_s, end="") +- else: +- print("%-9.6f " % delta_s, end="") +- print(format_string % (event.skaddr, event.pid, event.task.decode('utf-8', 'replace'), +- "6" if args.wide or args.csv else "", +- inet_ntop(AF_INET6, event.saddr), event.lport, +- inet_ntop(AF_INET6, event.daddr), event.dport, +- tcpstate2str(event.oldstate), tcpstate2str(event.newstate), +- float(event.span_us) / 1000)) +- if args.journal: +- journal.send(**journal_fields(event, AF_INET6)) ++ print_event(event, AF_INET6) + + # initialize BPF + b = BPF(text=bpf_text) +-- +2.41.0 + diff --git a/SOURCES/bcc-0.28.0-tools-tcpstates-fix-context-ptr-modified-error.patch b/SOURCES/bcc-0.28.0-tools-tcpstates-fix-context-ptr-modified-error.patch new file mode 100644 index 0000000..a35d536 --- /dev/null +++ b/SOURCES/bcc-0.28.0-tools-tcpstates-fix-context-ptr-modified-error.patch @@ -0,0 +1,144 @@ +From 53b89f35e8970beef55046c1bf035264f110f06d Mon Sep 17 00:00:00 2001 +From: hejun01 +Date: Thu, 29 Jun 2023 20:24:07 +0800 +Subject: [PATCH 1/2] tools/tcpstates: fix context ptr modified error + +Introduce local variable tcp_new_state, +to avoid llvm optimization of args->newstate, +which will cause context ptr args modified. +spilt event.ports to lport and dport. +switch type of TCP state from unsigned int to int. +--- + tools/tcpstates.py | 47 +++++++++++++++++++++++++--------------------- + 1 file changed, 26 insertions(+), 21 deletions(-) + +diff --git a/tools/tcpstates.py b/tools/tcpstates.py +index 89f3638c..9b2ccfa4 100755 +--- a/tools/tcpstates.py ++++ b/tools/tcpstates.py +@@ -82,9 +82,10 @@ struct ipv4_data_t { + u32 daddr; + u64 span_us; + u32 pid; +- u32 ports; +- u32 oldstate; +- u32 newstate; ++ u16 lport; ++ u16 dport; ++ int oldstate; ++ int newstate; + char task[TASK_COMM_LEN]; + }; + BPF_PERF_OUTPUT(ipv4_events); +@@ -96,9 +97,10 @@ struct ipv6_data_t { + unsigned __int128 daddr; + u64 span_us; + u32 pid; +- u32 ports; +- u32 oldstate; +- u32 newstate; ++ u16 lport; ++ u16 dport; ++ int oldstate; ++ int newstate; + char task[TASK_COMM_LEN]; + }; + BPF_PERF_OUTPUT(ipv6_events); +@@ -132,6 +134,9 @@ TRACEPOINT_PROBE(sock, inet_sock_set_state) + u16 family = args->family; + FILTER_FAMILY + ++ // workaround to avoid llvm optimization which will cause context ptr args modified ++ int tcp_newstate = args->newstate; ++ + if (args->family == AF_INET) { + struct ipv4_data_t data4 = { + .span_us = delta_us, +@@ -141,8 +146,8 @@ TRACEPOINT_PROBE(sock, inet_sock_set_state) + data4.ts_us = bpf_ktime_get_ns() / 1000; + __builtin_memcpy(&data4.saddr, args->saddr, sizeof(data4.saddr)); + __builtin_memcpy(&data4.daddr, args->daddr, sizeof(data4.daddr)); +- // a workaround until data4 compiles with separate lport/dport +- data4.ports = dport + ((0ULL + lport) << 16); ++ data4.lport = lport; ++ data4.dport = dport; + data4.pid = pid; + + bpf_get_current_comm(&data4.task, sizeof(data4.task)); +@@ -157,14 +162,14 @@ TRACEPOINT_PROBE(sock, inet_sock_set_state) + data6.ts_us = bpf_ktime_get_ns() / 1000; + __builtin_memcpy(&data6.saddr, args->saddr_v6, sizeof(data6.saddr)); + __builtin_memcpy(&data6.daddr, args->daddr_v6, sizeof(data6.daddr)); +- // a workaround until data6 compiles with separate lport/dport +- data6.ports = dport + ((0ULL + lport) << 16); ++ data6.lport = lport; ++ data6.dport = dport; + data6.pid = pid; + bpf_get_current_comm(&data6.task, sizeof(data6.task)); + ipv6_events.perf_submit(args, &data6, sizeof(data6)); + } + +- if (args->newstate == TCP_CLOSE) { ++ if (tcp_newstate == TCP_CLOSE) { + last.delete(&sk); + } else { + u64 ts = bpf_ktime_get_ns(); +@@ -210,8 +215,8 @@ int kprobe__tcp_set_state(struct pt_regs *ctx, struct sock *sk, int state) + data4.ts_us = bpf_ktime_get_ns() / 1000; + data4.saddr = sk->__sk_common.skc_rcv_saddr; + data4.daddr = sk->__sk_common.skc_daddr; +- // a workaround until data4 compiles with separate lport/dport +- data4.ports = dport + ((0ULL + lport) << 16); ++ data4.lport = lport; ++ data4.dport = dport; + data4.pid = pid; + + bpf_get_current_comm(&data4.task, sizeof(data4.task)); +@@ -228,8 +233,8 @@ int kprobe__tcp_set_state(struct pt_regs *ctx, struct sock *sk, int state) + sk->__sk_common.skc_v6_rcv_saddr.in6_u.u6_addr32); + bpf_probe_read_kernel(&data6.daddr, sizeof(data6.daddr), + sk->__sk_common.skc_v6_daddr.in6_u.u6_addr32); +- // a workaround until data6 compiles with separate lport/dport +- data6.ports = dport + ((0ULL + lport) << 16); ++ data6.lport = lport; ++ data6.dport = dport; + data6.pid = pid; + bpf_get_current_comm(&data6.task, sizeof(data6.task)); + ipv6_events.perf_submit(ctx, &data6, sizeof(data6)); +@@ -346,9 +351,9 @@ format_string = ("%-16x %-5d %-10.10s %s%-15s %-5d %-15s %-5d %-11s " + + 'OBJECT_COMM': event.task.decode('utf-8', 'replace'), + # Custom fields, aka "stuff we sort of made up". + 'OBJECT_' + addr_pfx + '_SOURCE_ADDRESS': inet_ntop(addr_family, pack("I", event.saddr)), +- 'OBJECT_TCP_SOURCE_PORT': str(event.ports >> 16), ++ 'OBJECT_TCP_SOURCE_PORT': str(event.lport), + 'OBJECT_' + addr_pfx + '_DESTINATION_ADDRESS': inet_ntop(addr_family, pack("I", event.daddr)), +- 'OBJECT_TCP_DESTINATION_PORT': str(event.ports & 0xffff), ++ 'OBJECT_TCP_DESTINATION_PORT': str(event.dport), + 'OBJECT_TCP_OLD_STATE': tcpstate2str(event.oldstate), + 'OBJECT_TCP_NEW_STATE': tcpstate2str(event.newstate), + 'OBJECT_TCP_SPAN_TIME': str(event.span_us) +@@ -386,8 +391,8 @@ format_string = ("%-16x %-5d %-10.10s %s%-15s %-5d %-15s %-5d %-11s " + + print("%-9.6f " % delta_s, end="") + print(format_string % (event.skaddr, event.pid, event.task.decode('utf-8', 'replace'), + "4" if args.wide or args.csv else "", +- inet_ntop(AF_INET, pack("I", event.saddr)), event.ports >> 16, +- inet_ntop(AF_INET, pack("I", event.daddr)), event.ports & 0xffff, ++ inet_ntop(AF_INET, pack("I", event.saddr)), event.lport, ++ inet_ntop(AF_INET, pack("I", event.daddr)), event.dport, + tcpstate2str(event.oldstate), tcpstate2str(event.newstate), + float(event.span_us) / 1000)) + if args.journal: +@@ -411,8 +416,8 @@ format_string = ("%-16x %-5d %-10.10s %s%-15s %-5d %-15s %-5d %-11s " + + print("%-9.6f " % delta_s, end="") + print(format_string % (event.skaddr, event.pid, event.task.decode('utf-8', 'replace'), + "6" if args.wide or args.csv else "", +- inet_ntop(AF_INET6, event.saddr), event.ports >> 16, +- inet_ntop(AF_INET6, event.daddr), event.ports & 0xffff, ++ inet_ntop(AF_INET6, event.saddr), event.lport, ++ inet_ntop(AF_INET6, event.daddr), event.dport, + tcpstate2str(event.oldstate), tcpstate2str(event.newstate), + float(event.span_us) / 1000)) + if args.journal: +-- +2.41.0 + diff --git a/SOURCES/bcc-0.28.0-tools-trace-don-t-raise-an-exception-in-a-ctype-call.patch b/SOURCES/bcc-0.28.0-tools-trace-don-t-raise-an-exception-in-a-ctype-call.patch new file mode 100644 index 0000000..a96fb88 --- /dev/null +++ b/SOURCES/bcc-0.28.0-tools-trace-don-t-raise-an-exception-in-a-ctype-call.patch @@ -0,0 +1,53 @@ +From 88274e83ca1a61699741d5b1d5499beb64cac753 Mon Sep 17 00:00:00 2001 +From: Jerome Marchand +Date: Mon, 16 Oct 2023 19:41:29 +0200 +Subject: [PATCH] tools/trace: don't raise an exception in a ctype callback + +To exit the tool when the maximal number of event is reached (-M +option), the tool currently call exit(), which raise a SystemExit +exception. The handling of exception from ctype callback doesn't seem +straightforward and dependent on python version. + +This patch avoid the issue altogether by using a global variable +instead. + +Closes #3049 + +Signed-off-by: Jerome Marchand +--- + tools/trace.py | 5 +++-- + 1 file changed, 3 insertions(+), 2 deletions(-) + +diff --git a/tools/trace.py b/tools/trace.py +index 9c7cca71..2aa096fa 100755 +--- a/tools/trace.py ++++ b/tools/trace.py +@@ -43,6 +43,7 @@ import sys + build_id_enabled = False + aggregate = False + symcount = {} ++ done = False + + @classmethod + def configure(cls, args): +@@ -635,7 +636,7 @@ BPF_PERF_OUTPUT(%s); + if self.aggregate: + self.print_aggregate_events() + sys.stdout.flush() +- exit() ++ Probe.done = True; + + def attach(self, bpf, verbose): + if len(self.library) == 0: +@@ -895,7 +896,7 @@ trace -s /lib/x86_64-linux-gnu/libc.so.6,/bin/ping 'p:c:inet_pton' -U + "-" if not all_probes_trivial else "")) + sys.stdout.flush() + +- while True: ++ while not Probe.done: + self.bpf.perf_buffer_poll() + + def run(self): +-- +2.41.0 + diff --git a/SPECS/bcc.spec b/SPECS/bcc.spec index 5aeca4f..955dd13 100644 --- a/SPECS/bcc.spec +++ b/SPECS/bcc.spec @@ -10,7 +10,7 @@ %endif %endif -%ifarch x86_64 ppc64 ppc64le aarch64 +%ifarch x86_64 ppc64 ppc64le aarch64 s390x %bcond_without libbpf_tools %else %bcond_with libbpf_tools @@ -24,20 +24,21 @@ Name: bcc -Version: 0.26.0 -Release: 4%{?dist} +Version: 0.28.0 +Release: 5%{?dist} Summary: BPF Compiler Collection (BCC) License: ASL 2.0 URL: https://github.com/iovisor/bcc Source0: %{url}/archive/v%{version}/%{name}-%{version}.tar.gz -Patch0: %%{name}-%%{version}-tools-nfsslower.py-Fix-uninitialized-struct-pad-erro.patch -Patch1: %%{name}-%%{version}-tools-slabratetop-Fix-error-incomplete-definition-of.patch -Patch2: %%{name}-%%{version}-tools-readahead-Fix-Failed-to-attach-BPF-program-ent.patch -Patch3: %%{name}-%%{version}-tools-compactsnoop.py-Fix-raw_tracepoint-Invalid-arg.patch -Patch4: %%{name}-%%{version}-killsnoop-add-missing-s-and-T-options-to-the-synopsi.patch -Patch5: %%{name}-%%{version}-tools-funcslower-fix-printing-of-folded-stacks.patch -Patch6: %%{name}-%%{version}-tools-deadlock-Add-an-option-to-set-the-maximum-numb.patch -Patch7: %%{name}-%%{version}-sync-with-latest-libbpf-repo.patch +Patch0: %%{name}-%%{version}-tools-tcpstates-fix-context-ptr-modified-error.patch +Patch1: %%{name}-%%{version}-tools-tcpstates-fix-IPv6-journal.patch +Patch2: %%{name}-%%{version}-tools-Add-support-for-the-new-block_io_-tracepoints.patch +Patch3: %%{name}-%%{version}-tools-trace-don-t-raise-an-exception-in-a-ctype-call.patch +Patch4: %%{name}-%%{version}-libbpf-tools-add-block_io_-start-done-tracepoints-su.patch +Patch5: %%{name}-%%{version}-libbpf-tools-Add-s390x-support.patch +Patch6: %%{name}-%%{version}-Fixing-pvalloc-memleak-test.patch +Patch7: %%{name}-%%{version}-Skipping-USDT-tests-for-Power-processor.patch +Patch8: %%{name}-%%{version}-Adding-memory-zones-for-Power-server.patch # Arches will be included as upstream support is added and dependencies are # satisfied in the respective arches @@ -261,6 +262,29 @@ cp -a libbpf-tools/tmp-install/bin/* %{buildroot}/%{_sbindir}/ %endif %changelog +* Wed Dec 13 2023 Jerome Marchand - 0.28.0-5 +- Fix libbpf bio tools (RHEL-19368) +- Add S390x support to libbpf-tools (RHEL-16325) +- Power enhancements(RHEL-11477) + +* Tue Nov 21 2023 Jerome Marchand - 0.28.0-4 +- Rebuild with LLVM 17 in the side tag (RHEL-10591) + +* Tue Nov 21 2023 Jerome Marchand - 0.28.0-3 +- Rebuild with LLVM 17 (RHEL-10591) + +* Mon Nov 06 2023 Jerome Marchand - 0.28.0-2 +- Fix trace tool (RHEL-8605) + +* Mon Oct 23 2023 Jerome Marchand - 0.28.0-1 +- Rebase to v0.28.0 (RHEL-9976) +- Rebuild with LLVM 17 (RHEL-10591) +- Fix bpf-biosnoop out of bound access (RHEL-8664) +- Fix kvmexit missing VM exit reasons and statistics (RHEL-8702) +- Fix multi-word array type handling (RHEL-8674) +- Fix tcpstates -Y (RHEL-8490) +- Fix bio tools (RHEL-8553) + * Wed Aug 09 2023 Jerome Marchand - 0.26.0-4 - Fix tcpretrans (rhbz#2226967)