Home Home > GIT Browse > SLE12-SP5-AZURE
summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorKernel Build Daemon <kbuild@suse.de>2019-08-23 07:10:59 +0200
committerKernel Build Daemon <kbuild@suse.de>2019-08-23 07:10:59 +0200
commit6fae7304707ae5caddd980b62e0479a6f594a6dc (patch)
tree3bb60824c0d9737a39cde97af89a41a733fc3374
parenta26b9aefeb92ea06cb97d51e941ce1b9629a2689 (diff)
parenta902929b14542a54b2c6baccf8f046283084a846 (diff)
Merge branch 'SLE12-SP5' into SLE12-SP5-AZURE
-rw-r--r--blacklist.conf2
-rw-r--r--kabi/arm64/symtypes-default270188
-rw-r--r--kabi/arm64/symvers-default19383
-rw-r--r--kabi/ppc64le/symtypes-default205586
-rw-r--r--kabi/ppc64le/symvers-default15737
-rw-r--r--kabi/s390x/symtypes-default134463
-rw-r--r--kabi/s390x/symvers-default11872
-rw-r--r--kabi/severities26
-rw-r--r--kabi/x86_64/symtypes-default281434
-rw-r--r--kabi/x86_64/symvers-default20047
-rw-r--r--patches.arch/kvm-add-machine-check-counter-to-kvm_stat.patch59
-rw-r--r--patches.arch/kvm-s390-abstract-adapter-interruption-word-generation-from-isc.patch55
-rw-r--r--patches.arch/kvm-s390-activate-gisa-for-emulated-interrupts.patch41
-rw-r--r--patches.arch/kvm-s390-add-debug-tracing-for-cpu-features-of-cpu-model.patch79
-rw-r--r--patches.arch/kvm-s390-add-exit-io-request-stats-and-simplify-code.patch81
-rw-r--r--patches.arch/kvm-s390-add-gisa-interrupts-to-flic-ioctl-interface.patch71
-rw-r--r--patches.arch/kvm-s390-clarify-kvm-related-kernel-message.patch47
-rw-r--r--patches.arch/kvm-s390-cleanup-struct-kvm_s390_float_interrupt.patch222
-rw-r--r--patches.arch/kvm-s390-define-gisa-format-0-data-structure.patch94
-rw-r--r--patches.arch/kvm-s390-do-not-set-intervention-requests-for-gisa-interrupts.patch51
-rw-r--r--patches.arch/kvm-s390-drop-use-of-spin-lock-in-_floating_irq_kick.patch41
-rw-r--r--patches.arch/kvm-s390-exploit-gisa-and-aiv-for-emulated-interrupts.patch269
-rw-r--r--patches.arch/kvm-s390-fix-fallthrough-annotation.patch60
-rw-r--r--patches.arch/kvm-s390-fix-kmsg-component-kvm-s390.patch44
-rw-r--r--patches.arch/kvm-s390-fix-potential-spectre-warnings.patch59
-rw-r--r--patches.arch/kvm-s390-generalize-kvm_s390_get_tod_clock_ext.patch74
-rw-r--r--patches.arch/kvm-s390-implement-gisa-ipm-related-primitives.patch62
-rw-r--r--patches.arch/kvm-s390-introduce-the-format-1-gisa.patch127
-rw-r--r--patches.arch/kvm-s390-make-kvm_s390_get_io_int-aware-of-gisa.patch124
-rw-r--r--patches.arch/kvm-s390-no-need-to-inititalize-kvm-arch-members-to-0.patch59
-rw-r--r--patches.arch/kvm-s390-provide-counters-for-all-interrupt-injects-delivery.patch301
-rw-r--r--patches.arch/kvm-s390-reverse-bit-ordering-of-irqs-in-pending-mask.patch143
-rw-r--r--patches.arch/kvm-s390-set-host-program-identifier.patch70
-rw-r--r--patches.arch/kvm-s390-tracing-apcb-changes.patch96
-rw-r--r--patches.arch/kvm-s390-unregister-debug-feature-on-failing-arch-init.patch63
-rw-r--r--patches.arch/kvm-s390-use-switch-vs-jump-table-in-intercept-c.patch113
-rw-r--r--patches.arch/kvm-s390-use-switch-vs-jump-table-in-interrupt-c.patch138
-rw-r--r--patches.arch/kvm-s390-use-switch-vs-jump-table-in-priv-c.patch254
-rw-r--r--patches.arch/kvm-s390-vsie-avoid-sparse-warning.patch32
-rw-r--r--patches.arch/kvm-s390-vsie-simplify-8k-address-checks.patch69
-rw-r--r--patches.arch/kvm-s390-vsie-store-guest-addresses-of-satellite-blocks-in-vsie_page.patch139
-rw-r--r--patches.arch/powerpc-rtas-use-device-model-APIs-and-serialization.patch98
-rw-r--r--patches.arch/s390-bitops-add-test_and_clear_bit_inv.patch40
-rw-r--r--patches.arch/s390-css-indicate-the-availability-of-the-aiv-facility.patch40
-rw-r--r--patches.arch/s390-sclp-detect-dirq-facility6
-rw-r--r--patches.arch/s390-sclp-expose-the-gisa-format-facility.patch53
-rw-r--r--patches.arch/s390-setup-set-control-program-code-via-diag-3186
-rw-r--r--patches.arch/s390-sles15sp1-00-04-12-KVM-s390-Add-skey-emulation-fault-handling.patch60
-rw-r--r--patches.arch/s390-sles15sp1-00-07-03-KVM-s390-refactor-crypto-initialization.patch14
-rw-r--r--patches.arch/v2-powerpc-Allow-flush_-inval_-dcache_range-to-work-across-ranges-4GB.patch58
-rw-r--r--patches.arch/x86-unwind-handle-null-pointer-calls-better-in-frame-unwinder.patch126
-rw-r--r--patches.drivers/ALSA-info-Fix-racy-addition-deletion-of-nodes.patch2
-rw-r--r--patches.drivers/ALSA-line6-Fix-write-on-zero-sized-buffer.patch2
-rw-r--r--patches.drivers/ALSA-usb-audio-Fix-gpf-in-snd_usb_pipe_sanity_check.patch2
-rw-r--r--patches.drivers/media-usb-siano-Fix-general-protection-fault-in-smsu.patch2
-rw-r--r--patches.fixes/0001-HID-wacom-Correct-distance-scale-for-2nd-gen-Intuos-.patch39
-rw-r--r--patches.fixes/0001-HID-wacom-correct-misreported-EKR-ring-values.patch39
-rw-r--r--patches.fixes/0001-media-cpia2_usb-first-wake-up-then-free-in-disconnec.patch2
-rw-r--r--patches.fixes/kvm-s390-add-etoken-support-for-guests.patch18
-rw-r--r--patches.fixes/nvme-multipath-fix-ana-log-nsid-lookup-when-nsid-is-.patch71
-rw-r--r--patches.fixes/scsi-qedi-remove-memset-memcpy-to-nfunc-and-use-func-instead165
-rw-r--r--patches.kabi/KVM-Fix-kABI.patch31
-rw-r--r--patches.kabi/kABI-Fix-kABI-for-struct-amd_iommu.patch29
-rw-r--r--patches.kabi/qla2xxx-kABI-fixes-for-v10.01.00.18-k.patch160
-rw-r--r--patches.suse/libnvdimm-pfn-Store-correct-value-of-npfns-in-namespace.patch47
-rw-r--r--patches.suse/nvme-Return-BLK_STS_TARGET-if-the-DNR-bit-is-set.patch36
-rw-r--r--series.conf51
67 files changed, 963034 insertions, 38 deletions
diff --git a/blacklist.conf b/blacklist.conf
index edffe2e719..a2d3912725 100644
--- a/blacklist.conf
+++ b/blacklist.conf
@@ -1291,3 +1291,5 @@ d065ee93aab6ef4c2a5af5c455b5044bd5136547 # config-only fix
1e1c50a929bc9e49bc3f9935b92450d9e69f8158 # affects only single core-machines
c2d1b3aae33605a61cbab445d8ae1c708ccd2698 # effectively reverted in upstream
2170a0d53bee1a6c1a4ebd042f99d85aafc6c0ea # only affects libnvdimm unit test code
+ce02ef06fcf7a399a6276adb83f37373d10cbbe1 # fixed in gcc instead, see bsc#1131264
+a9d57ef15cbe327fe54416dd194ee0ea66ae53a4 # ditto
diff --git a/kabi/arm64/symtypes-default b/kabi/arm64/symtypes-default
new file mode 100644
index 0000000000..c652bb33dd
--- /dev/null
+++ b/kabi/arm64/symtypes-default
@@ -0,0 +1,270188 @@
+/* arch/arm64/crypto/aes-ce-cipher.o */
+t#u32 typedef unsigned int u32
+s#crypto_aes_ctx struct crypto_aes_ctx { t#u32 key_enc [ ( ( 15 * 16 ) / sizeof ( u32 ) ) ] ; t#u32 key_dec [ ( ( 15 * 16 ) / sizeof ( u32 ) ) ] ; t#u32 key_length ; }
+t#u8 typedef unsigned char u8
+ce_aes_expandkey int ce_aes_expandkey ( s#crypto_aes_ctx * , const t#u8 * , unsigned int )
+s#crypto_ablkcipher struct crypto_ablkcipher { s#crypto_tfm base ; }
+s#list_head struct list_head { s#list_head * next , * prev ; }
+t#crypto_completion_t typedef void ( * crypto_completion_t ) ( s#crypto_async_request * , int )
+s#crypto_async_request struct crypto_async_request { s#list_head list ; t#crypto_completion_t complete ; void * data ; s#crypto_tfm * tfm ; t#u32 flags ; }
+s##scatterlist
+s#ablkcipher_request struct ablkcipher_request { s#crypto_async_request base ; unsigned int nbytes ; void * info ; s#scatterlist * src ; s#scatterlist * dst ; void * __ctx [ ] __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; }
+s#ablkcipher_tfm struct ablkcipher_tfm { int ( * setkey ) ( s#crypto_ablkcipher * , const t#u8 * , unsigned int ) ; int ( * encrypt ) ( s#ablkcipher_request * ) ; int ( * decrypt ) ( s#ablkcipher_request * ) ; s#crypto_ablkcipher * base ; unsigned int ivsize ; unsigned int reqsize ; }
+s#crypto_blkcipher struct crypto_blkcipher { s#crypto_tfm base ; }
+s#blkcipher_desc struct blkcipher_desc { s#crypto_blkcipher * tfm ; void * info ; t#u32 flags ; }
+s#blkcipher_tfm struct blkcipher_tfm { void * iv ; int ( * setkey ) ( s#crypto_tfm * , const t#u8 * , unsigned int ) ; int ( * encrypt ) ( s#blkcipher_desc * , s#scatterlist * , s#scatterlist * , unsigned int ) ; int ( * decrypt ) ( s#blkcipher_desc * , s#scatterlist * , s#scatterlist * , unsigned int ) ; }
+s#cipher_tfm struct cipher_tfm { int ( * cit_setkey ) ( s#crypto_tfm * , const t#u8 * , unsigned int ) ; void ( * cit_encrypt_one ) ( s#crypto_tfm * , t#u8 * , const t#u8 * ) ; void ( * cit_decrypt_one ) ( s#crypto_tfm * , t#u8 * , const t#u8 * ) ; }
+s#compress_tfm struct compress_tfm { int ( * cot_compress ) ( s#crypto_tfm * , const t#u8 * , unsigned int , t#u8 * , unsigned int * ) ; int ( * cot_decompress ) ( s#crypto_tfm * , const t#u8 * , unsigned int , t#u8 * , unsigned int * ) ; }
+t#atomic_t typedef struct { int counter ; } atomic_t
+s##crypto_type
+s##skcipher_givcrypt_request
+s#ablkcipher_alg struct ablkcipher_alg { int ( * setkey ) ( s#crypto_ablkcipher * , const t#u8 * , unsigned int ) ; int ( * encrypt ) ( s#ablkcipher_request * ) ; int ( * decrypt ) ( s#ablkcipher_request * ) ; int ( * givencrypt ) ( s#skcipher_givcrypt_request * ) ; int ( * givdecrypt ) ( s#skcipher_givcrypt_request * ) ; const char * geniv ; unsigned int min_keysize ; unsigned int max_keysize ; unsigned int ivsize ; }
+s#blkcipher_alg struct blkcipher_alg { int ( * setkey ) ( s#crypto_tfm * , const t#u8 * , unsigned int ) ; int ( * encrypt ) ( s#blkcipher_desc * , s#scatterlist * , s#scatterlist * , unsigned int ) ; int ( * decrypt ) ( s#blkcipher_desc * , s#scatterlist * , s#scatterlist * , unsigned int ) ; const char * geniv ; unsigned int min_keysize ; unsigned int max_keysize ; unsigned int ivsize ; }
+s#cipher_alg struct cipher_alg { unsigned int cia_min_keysize ; unsigned int cia_max_keysize ; int ( * cia_setkey ) ( s#crypto_tfm * , const t#u8 * , unsigned int ) ; void ( * cia_encrypt ) ( s#crypto_tfm * , t#u8 * , const t#u8 * ) ; void ( * cia_decrypt ) ( s#crypto_tfm * , t#u8 * , const t#u8 * ) ; }
+s#compress_alg struct compress_alg { int ( * coa_compress ) ( s#crypto_tfm * , const t#u8 * , unsigned int , t#u8 * , unsigned int * ) ; int ( * coa_decompress ) ( s#crypto_tfm * , const t#u8 * , unsigned int , t#u8 * , unsigned int * ) ; }
+e#module_state enum module_state { MODULE_STATE_LIVE , MODULE_STATE_COMING , MODULE_STATE_GOING , MODULE_STATE_UNFORMED , }
+t#u16 typedef unsigned short u16
+t#arch_spinlock_t typedef struct { t#u16 owner ; t#u16 next ; } __attribute__ ( ( aligned ( 4 ) ) ) arch_spinlock_t
+s#raw_spinlock struct raw_spinlock { t#arch_spinlock_t raw_lock ; }
+s#spinlock struct spinlock { union { s#raw_spinlock rlock ; } ; }
+t#spinlock_t typedef s#spinlock spinlock_t
+s#kobj_uevent_env struct kobj_uevent_env { char * argv [ 3 ] ; char * envp [ 32 ] ; int envp_idx ; char buf [ 2048 ] ; int buflen ; }
+s#kset_uevent_ops struct kset_uevent_ops { int ( * const filter ) ( s#kset * , s#kobject * ) ; const char * ( * const name ) ( s#kset * , s#kobject * ) ; int ( * const uevent ) ( s#kset * , s#kobject * , s#kobj_uevent_env * ) ; }
+s#kset struct kset { s#list_head list ; t#spinlock_t list_lock ; s#kobject kobj ; const s#kset_uevent_ops * uevent_ops ; }
+t#__kernel_long_t typedef long __kernel_long_t
+t#__kernel_ssize_t typedef t#__kernel_long_t __kernel_ssize_t
+t#ssize_t typedef t#__kernel_ssize_t ssize_t
+t#umode_t typedef unsigned short umode_t
+s#attribute struct attribute { const char * name ; t#umode_t mode ; }
+t#__kernel_ulong_t typedef unsigned long __kernel_ulong_t
+t#__kernel_size_t typedef t#__kernel_ulong_t __kernel_size_t
+t#size_t typedef t#__kernel_size_t size_t
+s#sysfs_ops struct sysfs_ops { t#ssize_t ( * show ) ( s#kobject * , s#attribute * , char * ) ; t#ssize_t ( * store ) ( s#kobject * , s#attribute * , const char * , t#size_t ) ; }
+e#kobj_ns_type enum kobj_ns_type { KOBJ_NS_TYPE_NONE = 0 , KOBJ_NS_TYPE_NET , KOBJ_NS_TYPES }
+t#bool typedef _Bool bool
+s##sock
+s#kobj_ns_type_operations struct kobj_ns_type_operations { e#kobj_ns_type type ; t#bool ( * current_may_mount ) ( void ) ; void * ( * grab_current_ns ) ( void ) ; const void * ( * netlink_ns ) ( s#sock * ) ; const void * ( * initial_ns ) ( void ) ; void ( * drop_ns ) ( void * ) ; }
+s#kobj_type struct kobj_type { void ( * release ) ( s#kobject * ) ; const s#sysfs_ops * sysfs_ops ; s#attribute * * default_attrs ; const s#kobj_ns_type_operations * ( * child_ns_type ) ( s#kobject * ) ; const void * ( * namespace ) ( s#kobject * ) ; }
+s#rb_node struct rb_node { unsigned long __rb_parent_color ; s#rb_node * rb_right ; s#rb_node * rb_left ; }
+s#rb_root struct rb_root { s#rb_node * rb_node ; }
+t#gfp_t typedef unsigned gfp_t
+s#callback_head struct callback_head { s#callback_head * next ; void ( * func ) ( s#callback_head * ) ; }
+s#radix_tree_node struct radix_tree_node { unsigned char shift ; unsigned char offset ; unsigned char count ; unsigned char exceptional ; s#radix_tree_node * parent ; s#radix_tree_root * root ; union { s#list_head private_list ; s#callback_head callback_head ; } ; void * slots [ ( 1UL << ( 0 ? 4 : 6 ) ) ] ; unsigned long tags [ 3 ] [ ( ( ( 1UL << ( 0 ? 4 : 6 ) ) + 64 - 1 ) / 64 ) ] ; }
+s#radix_tree_root struct radix_tree_root { t#gfp_t gfp_mask ; s#radix_tree_node * rnode ; }
+s#idr struct idr { s#radix_tree_root idr_rt ; unsigned int idr_next ; }
+s##seq_file
+s#kernfs_syscall_ops struct kernfs_syscall_ops { int ( * remount_fs ) ( s#kernfs_root * , int * , char * ) ; int ( * show_options ) ( s#seq_file * , s#kernfs_root * ) ; int ( * mkdir ) ( s#kernfs_node * , const char * , t#umode_t ) ; int ( * rmdir ) ( s#kernfs_node * ) ; int ( * rename ) ( s#kernfs_node * , s#kernfs_node * , const char * ) ; int ( * show_path ) ( s#seq_file * , s#kernfs_node * , s#kernfs_root * ) ; }
+s#wait_queue_head struct wait_queue_head { t#spinlock_t lock ; s#list_head head ; }
+t#wait_queue_head_t typedef s#wait_queue_head wait_queue_head_t
+s#kernfs_root struct kernfs_root { s#kernfs_node * kn ; unsigned int flags ; s#idr ino_idr ; t#u32 next_generation ; s#kernfs_syscall_ops * syscall_ops ; s#list_head supers ; t#wait_queue_head_t deactivate_waitq ; }
+s#kernfs_elem_dir struct kernfs_elem_dir { unsigned long subdirs ; s#rb_root children ; s#kernfs_root * root ; }
+s#kernfs_elem_symlink struct kernfs_elem_symlink { s#kernfs_node * target_kn ; }
+s##file
+t#atomic64_t typedef struct { long counter ; } atomic64_t
+t#atomic_long_t typedef t#atomic64_t atomic_long_t
+s#optimistic_spin_queue struct optimistic_spin_queue { t#atomic_t tail ; }
+s#mutex struct mutex { t#atomic_long_t owner ; t#spinlock_t wait_lock ; s#optimistic_spin_queue osq ; s#list_head wait_list ; }
+s##vm_operations_struct
+s#kernfs_open_file struct kernfs_open_file { s#kernfs_node * kn ; s#file * file ; s#seq_file * seq_file ; void * priv ; s#mutex mutex ; s#mutex prealloc_mutex ; int event ; s#list_head list ; char * prealloc_buf ; t#size_t atomic_write_len ; t#bool mmapped : 1 ; t#bool released : 1 ; const s#vm_operations_struct * vm_ops ; }
+t#__kernel_loff_t typedef long long __kernel_loff_t
+t#loff_t typedef t#__kernel_loff_t loff_t
+t#u64 typedef unsigned long long u64
+t#pgdval_t typedef t#u64 pgdval_t
+t#pgd_t typedef struct { t#pgdval_t pgd ; } pgd_t
+t#raw_spinlock_t typedef s#raw_spinlock raw_spinlock_t
+t#mm_segment_t typedef unsigned long mm_segment_t
+s#thread_info struct thread_info { unsigned long flags ; t#mm_segment_t addr_limit ; int preempt_count ; }
+s#llist_node struct llist_node { s#llist_node * next ; }
+s##sched_class
+s#load_weight struct load_weight { unsigned long weight ; t#u32 inv_weight ; }
+t#s64 typedef signed long long s64
+s#sched_statistics struct sched_statistics { t#u64 wait_start ; t#u64 wait_max ; t#u64 wait_count ; t#u64 wait_sum ; t#u64 iowait_count ; t#u64 iowait_sum ; t#u64 sleep_start ; t#u64 sleep_max ; t#s64 sum_sleep_runtime ; t#u64 block_start ; t#u64 block_max ; t#u64 exec_max ; t#u64 slice_max ; t#u64 nr_migrations_cold ; t#u64 nr_failed_migrations_affine ; t#u64 nr_failed_migrations_running ; t#u64 nr_failed_migrations_hot ; t#u64 nr_forced_migrations ; t#u64 nr_wakeups ; t#u64 nr_wakeups_sync ; t#u64 nr_wakeups_migrate ; t#u64 nr_wakeups_local ; t#u64 nr_wakeups_remote ; t#u64 nr_wakeups_affine ; t#u64 nr_wakeups_affine_attempts ; t#u64 nr_wakeups_passive ; t#u64 nr_wakeups_idle ; }
+s##cfs_rq
+s#sched_avg struct sched_avg { t#u64 last_update_time ; t#u64 load_sum ; t#u64 runnable_load_sum ; t#u32 util_sum ; t#u32 period_contrib ; unsigned long load_avg ; unsigned long runnable_load_avg ; unsigned long util_avg ; }
+s#sched_entity struct sched_entity { s#load_weight load ; unsigned long runnable_weight ; s#rb_node run_node ; s#list_head group_node ; unsigned int on_rq ; t#u64 exec_start ; t#u64 sum_exec_runtime ; t#u64 vruntime ; t#u64 prev_sum_exec_runtime ; t#u64 nr_migrations ; s#sched_statistics statistics ; int depth ; s#sched_entity * parent ; s#cfs_rq * cfs_rq ; s#cfs_rq * my_q ; s#sched_avg avg __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; }
+s#sched_rt_entity struct sched_rt_entity { s#list_head run_list ; unsigned long timeout ; unsigned long watchdog_stamp ; unsigned int time_slice ; unsigned short on_rq ; unsigned short on_list ; s#sched_rt_entity * back ; void * suse_kabi_padding ; }
+s##task_group
+t#ktime_t typedef t#s64 ktime_t
+s#timerqueue_node struct timerqueue_node { s#rb_node node ; t#ktime_t expires ; }
+e#hrtimer_restart enum hrtimer_restart { HRTIMER_NORESTART , HRTIMER_RESTART , }
+s#seqcount struct seqcount { unsigned sequence ; }
+t#seqcount_t typedef s#seqcount seqcount_t
+E#HRTIMER_MAX_CLOCK_BASES 4
+s#hrtimer_cpu_base struct hrtimer_cpu_base { t#raw_spinlock_t lock ; t#seqcount_t seq ; s#hrtimer * running ; unsigned int cpu ; unsigned int active_bases ; unsigned int clock_was_set_seq ; t#bool migration_enabled ; t#bool nohz_active ; unsigned int in_hrtirq : 1 , hres_active : 1 , hang_detected : 1 ; t#ktime_t expires_next ; s#hrtimer * next_timer ; unsigned int nr_events ; unsigned int nr_retries ; unsigned int nr_hangs ; unsigned int max_hang_time ; s#hrtimer_clock_base clock_base [ E#HRTIMER_MAX_CLOCK_BASES ] ; }
+t#__kernel_clockid_t typedef int __kernel_clockid_t
+t#clockid_t typedef t#__kernel_clockid_t clockid_t
+s#timerqueue_head struct timerqueue_head { s#rb_root head ; s#timerqueue_node * next ; }
+s#hrtimer_clock_base struct hrtimer_clock_base { s#hrtimer_cpu_base * cpu_base ; int index ; t#clockid_t clockid ; s#timerqueue_head active ; t#ktime_t ( * get_time ) ( void ) ; t#ktime_t offset ; }
+s#hrtimer struct hrtimer { s#timerqueue_node node ; t#ktime_t _softexpires ; e#hrtimer_restart ( * function ) ( s#hrtimer * ) ; s#hrtimer_clock_base * base ; t#u8 state ; t#u8 is_rel ; }
+s#sched_dl_entity struct sched_dl_entity { s#rb_node rb_node ; t#u64 dl_runtime ; t#u64 dl_deadline ; t#u64 dl_period ; t#u64 dl_bw ; t#s64 runtime ; t#u64 deadline ; unsigned int flags ; int dl_throttled ; int dl_boosted ; int dl_yielded ; s#hrtimer dl_timer ; }
+s#hlist_node struct hlist_node { s#hlist_node * next , * * pprev ; }
+s#hlist_head struct hlist_head { s#hlist_node * first ; }
+s#cpumask struct cpumask { unsigned long bits [ ( ( ( 480 ) + ( 8 * sizeof ( long ) ) - 1 ) / ( 8 * sizeof ( long ) ) ) ] ; }
+t#cpumask_t typedef s#cpumask cpumask_t
+s#sched_info struct sched_info { unsigned long pcount ; unsigned long long run_delay ; unsigned long long last_arrival ; unsigned long long last_queued ; }
+s#plist_node struct plist_node { int prio ; s#list_head prio_list ; s#list_head node_list ; }
+s#vmacache struct vmacache { t#u32 seqnum ; s#vm_area_struct * vmas [ ( 1U << 2 ) ] ; }
+E#NR_MM_COUNTERS 4
+s#task_rss_stat struct task_rss_stat { int events ; int count [ E#NR_MM_COUNTERS ] ; }
+t#__kernel_time_t typedef t#__kernel_long_t __kernel_time_t
+s#timespec struct timespec { t#__kernel_time_t tv_sec ; long tv_nsec ; }
+t#s32 typedef signed int s32
+t#compat_time_t typedef t#s32 compat_time_t
+s#compat_timespec struct compat_timespec { t#compat_time_t tv_sec ; t#s32 tv_nsec ; }
+s##pollfd
+s#restart_block struct restart_block { long ( * fn ) ( s#restart_block * ) ; union { struct { t#u32 * uaddr ; t#u32 val ; t#u32 flags ; t#u32 bitset ; t#u64 time ; t#u32 * uaddr2 ; } futex ; struct { t#clockid_t clockid ; s#timespec * rmtp ; s#compat_timespec * compat_rmtp ; t#u64 expires ; } nanosleep ; struct { s#pollfd * ufds ; int nfds ; int has_timeout ; unsigned long tv_sec ; unsigned long tv_nsec ; } poll ; } ; }
+t#__kernel_pid_t typedef int __kernel_pid_t
+t#pid_t typedef t#__kernel_pid_t pid_t
+E#PIDTYPE_MAX 3
+s##pid_namespace
+s#upid struct upid { int nr ; s#pid_namespace * ns ; s#hlist_node pid_chain ; }
+s#pid struct pid { t#atomic_t count ; unsigned int level ; s#hlist_head tasks [ E#PIDTYPE_MAX ] ; s#callback_head rcu ; s#upid numbers [ 1 ] ; }
+s#pid_link struct pid_link { s#hlist_node node ; s#pid * pid ; }
+s#completion struct completion { unsigned int done ; t#wait_queue_head_t wait ; }
+s#prev_cputime struct prev_cputime { t#u64 utime ; t#u64 stime ; t#raw_spinlock_t lock ; }
+e#vtime_state enum vtime_state { VTIME_INACTIVE = 0 , VTIME_USER , VTIME_SYS , }
+s#vtime struct vtime { t#seqcount_t seqcount ; unsigned long long starttime ; e#vtime_state state ; t#u64 utime ; t#u64 stime ; t#u64 gtime ; }
+s#task_cputime struct task_cputime { t#u64 utime ; t#u64 stime ; unsigned long long sum_exec_runtime ; }
+s##cred
+s##nameidata
+s##sem_undo_list
+s#sysv_sem struct sysv_sem { s#sem_undo_list * undo_list ; }
+s#sysv_shm struct sysv_shm { s#list_head shm_clist ; }
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+t#sigset_t typedef struct { unsigned long sig [ ( 64 / 64 ) ] ; } sigset_t
+s#sigpending struct sigpending { s#list_head list ; t#sigset_t signal ; }
+s##audit_context
+t#__kernel_uid32_t typedef unsigned int __kernel_uid32_t
+t#uid_t typedef t#__kernel_uid32_t uid_t
+t#kuid_t typedef struct { t#uid_t val ; } kuid_t
+s##seccomp_filter
+s#seccomp struct seccomp { int mode ; s#seccomp_filter * filter ; }
+s#wake_q_node struct wake_q_node { s#wake_q_node * next ; }
+s#rb_root_cached struct rb_root_cached { s#rb_root rb_root ; s#rb_node * rb_leftmost ; }
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+t#__kernel_timer_t typedef int __kernel_timer_t
+u#sigval union sigval { int sival_int ; void * sival_ptr ; }
+t#sigval_t typedef u#sigval sigval_t
+t#__kernel_clock_t typedef t#__kernel_long_t __kernel_clock_t
+t#__u32 typedef unsigned int __u32
+s#siginfo struct siginfo { int si_signo ; int si_errno ; int si_code ; union { int _pad [ ( ( 128 - ( 4 * sizeof ( int ) ) ) / sizeof ( int ) ) ] ; struct { t#__kernel_pid_t _pid ; t#__kernel_uid32_t _uid ; } _kill ; struct { t#__kernel_timer_t _tid ; int _overrun ; char _pad [ sizeof ( __kernel_uid32_t ) - sizeof ( int ) ] ; t#sigval_t _sigval ; int _sys_private ; } _timer ; struct { t#__kernel_pid_t _pid ; t#__kernel_uid32_t _uid ; t#sigval_t _sigval ; } _rt ; struct { t#__kernel_pid_t _pid ; t#__kernel_uid32_t _uid ; int _status ; t#__kernel_clock_t _utime ; t#__kernel_clock_t _stime ; } _sigchld ; struct { void * _addr ; short _addr_lsb ; union { struct { void * _lower ; void * _upper ; } _addr_bnd ; t#__u32 _pkey ; } ; } _sigfault ; struct { long _band ; int _fd ; } _sigpoll ; struct { void * _call_addr ; int _syscall ; unsigned int _arch ; } _sigsys ; } _sifields ; }
+t#siginfo_t typedef s#siginfo siginfo_t
+s#task_io_accounting struct task_io_accounting { t#u64 rchar ; t#u64 wchar ; t#u64 syscr ; t#u64 syscw ; t#u64 read_bytes ; t#u64 write_bytes ; t#u64 cancelled_write_bytes ; }
+t#nodemask_t typedef struct { unsigned long bits [ ( ( ( ( 1 << 2 ) ) + ( 8 * sizeof ( long ) ) - 1 ) / ( 8 * sizeof ( long ) ) ) ] ; } nodemask_t
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+E#perf_nr_task_contexts ( 0 ) + 2
+s##mempolicy
+s##numa_group
+s#tlbflush_unmap_batch struct tlbflush_unmap_batch { }
+s##pipe_inode_info
+s##address_space
+s##dev_pagemap
+s##kmem_cache
+s##mem_cgroup
+s#page struct page { unsigned long flags ; union { s#address_space * mapping ; void * s_mem ; t#atomic_t compound_mapcount ; } ; union { t#atomic_t pt_frag_refcount ; unsigned long index ; void * freelist ; } ; union { unsigned counters ; struct { union { t#atomic_t _mapcount ; unsigned int active ; struct { unsigned inuse : 16 ; unsigned objects : 15 ; unsigned frozen : 1 ; } ; int units ; } ; t#atomic_t _refcount ; } ; } ; union { s#list_head lru ; s#dev_pagemap * pgmap ; struct { s#page * next ; int pages ; int pobjects ; } ; s#callback_head callback_head ; struct { unsigned long compound_head ; unsigned int compound_dtor ; unsigned int compound_order ; } ; } ; union { unsigned long private ; t#spinlock_t ptl ; s#kmem_cache * slab_cache ; } ; s#mem_cgroup * mem_cgroup ; }
+s#page_frag struct page_frag { s#page * page ; t#__u32 offset ; t#__u32 size ; }
+s##task_delay_info
+s#latency_record struct latency_record { unsigned long backtrace [ 12 ] ; unsigned int count ; unsigned long time ; unsigned long max ; }
+s##ftrace_ret_stack
+e#uprobe_task_state enum uprobe_task_state { UTASK_RUNNING , UTASK_SSTEP , UTASK_SSTEP_ACK , UTASK_SSTEP_TRAPPED , }
+s#arch_uprobe_task struct arch_uprobe_task { }
+s##uprobe
+s#return_instance struct return_instance { s#uprobe * uprobe ; unsigned long func ; unsigned long stack ; unsigned long orig_ret_vaddr ; t#bool chained ; s#return_instance * next ; }
+s#uprobe_task struct uprobe_task { e#uprobe_task_state state ; union { struct { s#arch_uprobe_task autask ; unsigned long vaddr ; } ; struct { s#callback_head dup_xol_work ; unsigned long dup_xol_addr ; } ; } ; s#uprobe * active_uprobe ; unsigned long xol_vaddr ; s#return_instance * return_instances ; unsigned int depth ; }
+s#cpu_context struct cpu_context { unsigned long x19 ; unsigned long x20 ; unsigned long x21 ; unsigned long x22 ; unsigned long x23 ; unsigned long x24 ; unsigned long x25 ; unsigned long x26 ; unsigned long x27 ; unsigned long x28 ; unsigned long fp ; unsigned long sp ; unsigned long pc ; }
+s#user_fpsimd_state struct user_fpsimd_state { __uint128_t vregs [ 32 ] ; t#__u32 fpsr ; t#__u32 fpcr ; t#__u32 __reserved [ 2 ] ; }
+s##perf_event
+s#debug_info struct debug_info { int suspended_step ; int bps_disabled ; int wps_disabled ; s#perf_event * hbp_break [ 16 ] ; s#perf_event * hbp_watch [ 16 ] ; }
+s#thread_struct struct thread_struct { s#cpu_context cpu_context ; unsigned long tp_value ; unsigned long tp2_value ; s#user_fpsimd_state fpsimd_state ; unsigned int fpsimd_cpu ; void * sve_state ; unsigned int sve_vl ; unsigned int sve_vl_onexec ; unsigned long fault_address ; unsigned long fault_code ; s#debug_info debug ; }
+s#task_struct struct task_struct { s#thread_info thread_info ; volatile long state ; void * stack ; t#atomic_t usage ; unsigned int flags ; unsigned int ptrace ; s#llist_node wake_entry ; int on_cpu ; unsigned int cpu ; unsigned int wakee_flips ; unsigned long wakee_flip_decay_ts ; s#task_struct * last_wakee ; int recent_used_cpu ; int wake_cpu ; int on_rq ; int prio ; int static_prio ; int normal_prio ; unsigned int rt_priority ; const s#sched_class * sched_class ; s#sched_entity se ; s#sched_rt_entity rt ; s#task_group * sched_task_group ; s#sched_dl_entity dl ; s#hlist_head preempt_notifiers ; unsigned int btrace_seq ; unsigned int policy ; int nr_cpus_allowed ; t#cpumask_t cpus_allowed ; unsigned long rcu_tasks_nvcsw ; t#bool rcu_tasks_holdout ; s#list_head rcu_tasks_holdout_list ; int rcu_tasks_idle_cpu ; s#sched_info sched_info ; s#list_head tasks ; s#plist_node pushable_tasks ; s#rb_node pushable_dl_tasks ; s#mm_struct * mm ; s#mm_struct * active_mm ; s#vmacache vmacache ; s#task_rss_stat rss_stat ; int exit_state ; int exit_code ; int exit_signal ; int pdeath_signal ; unsigned long jobctl ; unsigned int personality ; unsigned sched_reset_on_fork : 1 ; unsigned sched_contributes_to_load : 1 ; unsigned sched_migrated : 1 ; unsigned sched_remote_wakeup : 1 ; unsigned : 0 ; unsigned in_execve : 1 ; unsigned in_iowait : 1 ; unsigned memcg_may_oom : 1 ; unsigned memcg_kmem_skip_account : 1 ; unsigned no_cgroup_migration : 1 ; unsigned long atomic_flags ; s#restart_block restart_block ; t#pid_t pid ; t#pid_t tgid ; unsigned long stack_canary ; s#task_struct * real_parent ; s#task_struct * parent ; s#list_head children ; s#list_head sibling ; s#task_struct * group_leader ; s#list_head ptraced ; s#list_head ptrace_entry ; s#pid_link pids [ E#PIDTYPE_MAX ] ; s#list_head thread_group ; s#list_head thread_node ; s#completion * vfork_done ; int * set_child_tid ; int * clear_child_tid ; t#u64 utime ; t#u64 stime ; t#u64 gtime ; s#prev_cputime prev_cputime ; s#vtime vtime ; t#atomic_t tick_dep_mask ; unsigned long nvcsw ; unsigned long nivcsw ; t#u64 start_time ; t#u64 real_start_time ; unsigned long min_flt ; unsigned long maj_flt ; s#task_cputime cputime_expires ; s#list_head cpu_timers [ 3 ] ; const s#cred * ptracer_cred ; const s#cred * real_cred ; const s#cred * cred ; char comm [ 16 ] ; s#nameidata * nameidata ; s#sysv_sem sysvsem ; s#sysv_shm sysvshm ; unsigned long last_switch_count ; s#fs_struct * fs ; s#files_struct * files ; s#nsproxy * nsproxy ; s#signal_struct * signal ; s#sighand_struct * sighand ; t#sigset_t blocked ; t#sigset_t real_blocked ; t#sigset_t saved_sigmask ; s#sigpending pending ; unsigned long sas_ss_sp ; t#size_t sas_ss_size ; unsigned int sas_ss_flags ; s#callback_head * task_works ; s#audit_context * audit_context ; t#kuid_t loginuid ; unsigned int sessionid ; s#seccomp seccomp ; t#u32 parent_exec_id ; t#u32 self_exec_id ; t#spinlock_t alloc_lock ; t#raw_spinlock_t pi_lock ; s#wake_q_node wake_q ; s#rb_root_cached pi_waiters ; s#task_struct * pi_top_task ; s#rt_mutex_waiter * pi_blocked_on ; void * journal_info ; s#bio_list * bio_list ; s#blk_plug * plug ; s#reclaim_state * reclaim_state ; s#backing_dev_info * backing_dev_info ; s#io_context * io_context ; unsigned long ptrace_message ; t#siginfo_t * last_siginfo ; s#task_io_accounting ioac ; t#u64 acct_rss_mem1 ; t#u64 acct_vm_mem1 ; t#u64 acct_timexpd ; t#nodemask_t mems_allowed ; t#seqcount_t mems_allowed_seq ; int cpuset_mem_spread_rotor ; int cpuset_slab_spread_rotor ; s#css_set * cgroups ; s#list_head cg_list ; s#robust_list_head * robust_list ; s#compat_robust_list_head * compat_robust_list ; s#list_head pi_state_list ; s#futex_pi_state * pi_state_cache ; s#perf_event_context * perf_event_ctxp [ E#perf_nr_task_contexts ] ; s#mutex perf_event_mutex ; s#list_head perf_event_list ; s#mempolicy * mempolicy ; short il_next ; short pref_node_fork ; int numa_scan_seq ; unsigned int numa_scan_period ; unsigned int numa_scan_period_max ; int numa_preferred_nid ; unsigned long numa_migrate_retry ; t#u64 node_stamp ; t#u64 last_task_numa_placement ; t#u64 last_sum_exec_runtime ; s#callback_head numa_work ; s#numa_group * numa_group ; unsigned long * numa_faults ; unsigned long total_numa_faults ; unsigned long numa_faults_locality [ 3 ] ; unsigned long numa_pages_migrated ; s#tlbflush_unmap_batch tlb_ubc ; s#callback_head rcu ; s#pipe_inode_info * splice_pipe ; s#page_frag task_frag ; s#task_delay_info * delays ; int make_it_fail ; int nr_dirtied ; int nr_dirtied_pause ; unsigned long dirty_paused_when ; int latency_record_count ; s#latency_record latency_record [ 32 ] ; t#u64 timer_slack_ns ; t#u64 default_timer_slack_ns ; int curr_ret_stack ; s#ftrace_ret_stack * ret_stack ; unsigned long long ftrace_timestamp ; t#atomic_t trace_overrun ; t#atomic_t tracing_graph_pause ; unsigned long trace ; unsigned long trace_recursion ; s#mem_cgroup * memcg_in_oom ; t#gfp_t memcg_oom_gfp_mask ; int memcg_oom_order ; unsigned int memcg_nr_pages_over_high ; s#uprobe_task * utask ; unsigned int sequential_io ; unsigned int sequential_io_avg ; int pagefault_disabled ; s#task_struct * oom_reaper_list ; t#atomic_t stack_refcount ; void * security ; void * suse_kabi_padding ; s#thread_struct thread ; }
+s#rw_semaphore struct rw_semaphore { t#atomic_long_t count ; s#list_head wait_list ; t#raw_spinlock_t wait_lock ; s#optimistic_spin_queue osq ; s#task_struct * owner ; }
+s#mm_rss_stat struct mm_rss_stat { t#atomic_long_t count [ E#NR_MM_COUNTERS ] ; }
+s##linux_binfmt
+t#cpumask_var_t typedef s#cpumask cpumask_var_t [ 1 ]
+t#mm_context_t typedef struct { t#atomic64_t id ; void * vdso ; unsigned long flags ; } mm_context_t
+s#core_thread struct core_thread { s#task_struct * task ; s#core_thread * next ; }
+s#core_state struct core_state { t#atomic_t nr_threads ; s#core_thread dumper ; s#completion startup ; }
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+t#pgtable_t typedef s#page * pgtable_t
+s##xol_area
+s#uprobes_state struct uprobes_state { s#xol_area * xol_area ; }
+t#work_func_t typedef void ( * work_func_t ) ( s#work_struct * )
+s#work_struct struct work_struct { t#atomic_long_t data ; s#list_head entry ; t#work_func_t func ; }
+s#mm_struct struct mm_struct { s#vm_area_struct * mmap ; s#rb_root mm_rb ; t#u32 vmacache_seqnum ; unsigned long ( * get_unmapped_area ) ( s#file * , unsigned long , unsigned long , unsigned long , unsigned long ) ; unsigned long mmap_base ; unsigned long mmap_legacy_base ; unsigned long task_size ; unsigned long highest_vm_end ; t#pgd_t * pgd ; t#atomic_t mm_users ; t#atomic_t mm_count ; t#atomic_long_t nr_ptes ; t#atomic_long_t nr_pmds ; int map_count ; t#spinlock_t page_table_lock ; s#rw_semaphore mmap_sem ; s#list_head mmlist ; unsigned long hiwater_rss ; unsigned long hiwater_vm ; unsigned long total_vm ; unsigned long locked_vm ; unsigned long pinned_vm ; unsigned long data_vm ; unsigned long exec_vm ; unsigned long stack_vm ; unsigned long def_flags ; unsigned long start_code , end_code , start_data , end_data ; unsigned long start_brk , brk , start_stack ; unsigned long arg_start , arg_end , env_start , env_end ; unsigned long saved_auxv [ ( 2 * ( 2 + 20 + 1 ) ) ] ; s#mm_rss_stat rss_stat ; s#linux_binfmt * binfmt ; t#cpumask_var_t cpu_vm_mask_var ; t#mm_context_t context ; unsigned long flags ; s#core_state * core_state ; t#spinlock_t ioctx_lock ; s#kioctx_table * ioctx_table ; s#task_struct * owner ; s#user_namespace * user_ns ; s#file * exe_file ; s#mmu_notifier_mm * mmu_notifier_mm ; t#pgtable_t pmd_huge_pte ; unsigned long numa_next_scan ; unsigned long numa_scan_offset ; int numa_scan_seq ; t#atomic_t tlb_flush_pending ; s#uprobes_state uprobes_state ; t#atomic_long_t hugetlb_usage ; s#work_struct async_put_work ; }
+t#pteval_t typedef t#u64 pteval_t
+t#pgprot_t typedef struct { t#pteval_t pgprot ; } pgprot_t
+s##anon_vma
+s##userfaultfd_ctx
+s#vm_userfaultfd_ctx struct vm_userfaultfd_ctx { s#userfaultfd_ctx * ctx ; }
+s#vm_area_struct struct vm_area_struct { unsigned long vm_start ; unsigned long vm_end ; s#vm_area_struct * vm_next , * vm_prev ; s#rb_node vm_rb ; unsigned long rb_subtree_gap ; s#mm_struct * vm_mm ; t#pgprot_t vm_page_prot ; unsigned long vm_flags ; struct { s#rb_node rb ; unsigned long rb_subtree_last ; } shared ; s#list_head anon_vma_chain ; s#anon_vma * anon_vma ; const s#vm_operations_struct * vm_ops ; unsigned long vm_pgoff ; s#file * vm_file ; void * vm_private_data ; s#mempolicy * vm_policy ; s#vm_userfaultfd_ctx vm_userfaultfd_ctx ; void * suse_kabi_padding ; }
+s#kernfs_ops struct kernfs_ops { int ( * open ) ( s#kernfs_open_file * ) ; void ( * release ) ( s#kernfs_open_file * ) ; int ( * seq_show ) ( s#seq_file * , void * ) ; void * ( * seq_start ) ( s#seq_file * , t#loff_t * ) ; void * ( * seq_next ) ( s#seq_file * , void * , t#loff_t * ) ; void ( * seq_stop ) ( s#seq_file * , void * ) ; t#ssize_t ( * read ) ( s#kernfs_open_file * , char * , t#size_t , t#loff_t ) ; t#size_t atomic_write_len ; t#bool prealloc ; t#ssize_t ( * write ) ( s#kernfs_open_file * , char * , t#size_t , t#loff_t ) ; int ( * mmap ) ( s#kernfs_open_file * , s#vm_area_struct * ) ; }
+s##kernfs_open_node
+s#kernfs_elem_attr struct kernfs_elem_attr { const s#kernfs_ops * ops ; s#kernfs_open_node * open ; t#loff_t size ; s#kernfs_node * notify_next ; }
+u#kernfs_node_id union kernfs_node_id { struct { t#u32 ino ; t#u32 generation ; } ; t#u64 id ; }
+s##kernfs_iattrs
+s#kernfs_node struct kernfs_node { t#atomic_t count ; t#atomic_t active ; s#kernfs_node * parent ; const char * name ; s#rb_node rb ; const void * ns ; unsigned int hash ; union { s#kernfs_elem_dir dir ; s#kernfs_elem_symlink symlink ; s#kernfs_elem_attr attr ; } ; void * priv ; u#kernfs_node_id id ; unsigned short flags ; t#umode_t mode ; s#kernfs_iattrs * iattr ; }
+s#refcount_struct struct refcount_struct { t#atomic_t refs ; }
+t#refcount_t typedef s#refcount_struct refcount_t
+s#kref struct kref { t#refcount_t refcount ; }
+s#kobject struct kobject { const char * name ; s#list_head entry ; s#kobject * parent ; s#kset * kset ; s#kobj_type * ktype ; s#kernfs_node * sd ; s#kref kref ; unsigned int state_initialized : 1 ; unsigned int state_in_sysfs : 1 ; unsigned int state_add_uevent_sent : 1 ; unsigned int state_remove_uevent_sent : 1 ; unsigned int uevent_suppress : 1 ; }
+s##module_param_attrs
+s#module_kobject struct module_kobject { s#kobject kobj ; s#module * mod ; s#kobject * drivers_dir ; s#module_param_attrs * mp ; s#completion * kobj_completion ; }
+s#module_attribute struct module_attribute { s#attribute attr ; t#ssize_t ( * show ) ( s#module_attribute * , s#module_kobject * , char * ) ; t#ssize_t ( * store ) ( s#module_attribute * , s#module_kobject * , const char * , t#size_t ) ; void ( * setup ) ( s#module * , const char * ) ; int ( * test ) ( s#module * ) ; void ( * free ) ( s#module * ) ; }
+s#kernel_symbol struct kernel_symbol { unsigned long value ; const char * name ; }
+s#kernel_param_ops struct kernel_param_ops { unsigned int flags ; int ( * set ) ( const char * , const s#kernel_param * ) ; int ( * get ) ( char * , const s#kernel_param * ) ; void ( * free ) ( void * ) ; }
+t#s8 typedef signed char s8
+s#kparam_string struct kparam_string { unsigned int maxlen ; char * string ; }
+s#kparam_array struct kparam_array { unsigned int max ; unsigned int elemsize ; unsigned int * num ; const s#kernel_param_ops * ops ; void * elem ; }
+s#kernel_param struct kernel_param { const char * name ; s#module * mod ; const s#kernel_param_ops * ops ; const t#u16 perm ; t#s8 level ; t#u8 flags ; union { void * arg ; const s#kparam_string * str ; const s#kparam_array * arr ; } ; }
+s#exception_table_entry struct exception_table_entry { int insn , fixup ; }
+s#latch_tree_node struct latch_tree_node { s#rb_node node [ 2 ] ; }
+s#mod_tree_node struct mod_tree_node { s#module * mod ; s#latch_tree_node node ; }
+s#module_layout struct module_layout { void * base ; unsigned int size ; unsigned int text_size ; unsigned int ro_size ; unsigned int ro_after_init_size ; s#mod_tree_node mtn ; }
+t#Elf64_Word typedef t#__u32 Elf64_Word
+t#__u64 __extension__ typedef unsigned long long __u64
+t#Elf64_Xword typedef t#__u64 Elf64_Xword
+t#Elf64_Addr typedef t#__u64 Elf64_Addr
+t#Elf64_Off typedef t#__u64 Elf64_Off
+s#elf64_shdr struct elf64_shdr { t#Elf64_Word sh_name ; t#Elf64_Word sh_type ; t#Elf64_Xword sh_flags ; t#Elf64_Addr sh_addr ; t#Elf64_Off sh_offset ; t#Elf64_Xword sh_size ; t#Elf64_Word sh_link ; t#Elf64_Word sh_info ; t#Elf64_Xword sh_addralign ; t#Elf64_Xword sh_entsize ; }
+s#mod_plt_sec struct mod_plt_sec { s#elf64_shdr * plt ; int plt_num_entries ; int plt_max_entries ; }
+s#mod_arch_specific struct mod_arch_specific { s#mod_plt_sec core ; s#mod_plt_sec init ; }
+s#bug_entry struct bug_entry { signed int bug_addr_disp ; signed int file_disp ; unsigned short line ; unsigned short flags ; }
+t#__u16 typedef unsigned short __u16
+t#Elf64_Half typedef t#__u16 Elf64_Half
+s#elf64_sym struct elf64_sym { t#Elf64_Word st_name ; unsigned char st_info ; unsigned char st_other ; t#Elf64_Half st_shndx ; t#Elf64_Addr st_value ; t#Elf64_Xword st_size ; }
+t#Elf64_Sym typedef s#elf64_sym Elf64_Sym
+s#mod_kallsyms struct mod_kallsyms { t#Elf64_Sym * symtab ; unsigned int num_symtab ; char * strtab ; }
+s##module_sect_attrs
+s##module_notes_attrs
+s##tracepoint
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s#module struct module { e#module_state state ; s#list_head list ; char name [ ( 64 - sizeof ( unsigned long ) ) ] ; s#module_kobject mkobj ; s#module_attribute * modinfo_attrs ; const char * version ; const char * srcversion ; s#kobject * holders_dir ; const s#kernel_symbol * syms ; const t#s32 * crcs ; unsigned int num_syms ; s#mutex param_lock ; s#kernel_param * kp ; unsigned int num_kp ; unsigned int num_gpl_syms ; const s#kernel_symbol * gpl_syms ; const t#s32 * gpl_crcs ; const s#kernel_symbol * unused_syms ; const t#s32 * unused_crcs ; unsigned int num_unused_syms ; unsigned int num_unused_gpl_syms ; const s#kernel_symbol * unused_gpl_syms ; const t#s32 * unused_gpl_crcs ; t#bool sig_ok ; t#bool async_probe_requested ; const s#kernel_symbol * gpl_future_syms ; const t#s32 * gpl_future_crcs ; unsigned int num_gpl_future_syms ; unsigned int num_exentries ; s#exception_table_entry * extable ; int ( * init ) ( void ) ; s#module_layout core_layout __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; s#module_layout init_layout ; s#mod_arch_specific arch ; unsigned long taints ; unsigned num_bugs ; s#list_head bug_list ; s#bug_entry * bug_table ; s#mod_kallsyms * kallsyms ; s#mod_kallsyms core_kallsyms ; s#module_sect_attrs * sect_attrs ; s#module_notes_attrs * notes_attrs ; char * args ; void * percpu ; unsigned int percpu_size ; unsigned int num_tracepoints ; s#tracepoint * const * tracepoints_ptrs ; unsigned int num_trace_bprintk_fmt ; const char * * trace_bprintk_fmt_start ; s#trace_event_call * * trace_events ; unsigned int num_trace_events ; s#trace_enum_map * * trace_enums ; unsigned int num_trace_enums ; unsigned int num_ftrace_callsites ; unsigned long * ftrace_callsites ; t#bool klp ; t#bool klp_alive ; s#klp_modinfo * klp_info ; s#list_head source_list ; s#list_head target_list ; void ( * exit ) ( void ) ; t#atomic_t refcnt ; void * suse_kabi_padding ; }
+s#crypto_alg struct crypto_alg { s#list_head cra_list ; s#list_head cra_users ; t#u32 cra_flags ; unsigned int cra_blocksize ; unsigned int cra_ctxsize ; unsigned int cra_alignmask ; int cra_priority ; t#atomic_t cra_refcnt ; char cra_name [ 128 ] ; char cra_driver_name [ 128 ] ; const s#crypto_type * cra_type ; union { s#ablkcipher_alg ablkcipher ; s#blkcipher_alg blkcipher ; s#cipher_alg cipher ; s#compress_alg compress ; } cra_u ; int ( * cra_init ) ( s#crypto_tfm * ) ; void ( * cra_exit ) ( s#crypto_tfm * ) ; void ( * cra_destroy ) ( s#crypto_alg * ) ; s#module * cra_module ; }
+s#crypto_tfm struct crypto_tfm { t#u32 crt_flags ; union { s#ablkcipher_tfm ablkcipher ; s#blkcipher_tfm blkcipher ; s#cipher_tfm cipher ; s#compress_tfm compress ; } crt_u ; void ( * exit ) ( s#crypto_tfm * ) ; s#crypto_alg * __crt_alg ; void * __crt_ctx [ ] __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; }
+ce_aes_setkey int ce_aes_setkey ( s#crypto_tfm * , const t#u8 * , unsigned int )
+
+/* arch/arm64/crypto/aes-cipher-glue.o */
+__aes_arm64_encrypt void __aes_arm64_encrypt ( t#u32 * , t#u8 * , const t#u8 * , int )
+__aes_arm64_decrypt void __aes_arm64_decrypt ( t#u32 * , t#u8 * , const t#u8 * , int )
+
+/* arch/arm64/crypto/aes-glue-neon.o */
+neon_aes_ecb_encrypt void neon_aes_ecb_encrypt ( t#u8 [ ] , t#u8 const [ ] , t#u8 const [ ] , int , int , int )
+neon_aes_cbc_encrypt void neon_aes_cbc_encrypt ( t#u8 [ ] , t#u8 const [ ] , t#u8 const [ ] , int , int , t#u8 [ ] , int )
+
+/* arch/arm64/crypto/sha256-glue.o */
+sha256_block_data_order void sha256_block_data_order ( t#u32 * , const void * , unsigned int )
+
+/* arch/arm64/kernel/acpi.o */
+acpi_disabled extern int acpi_disabled
+acpi_pci_disabled extern int acpi_pci_disabled
+
+/* arch/arm64/kernel/arm64ksyms.o */
+copy_page extern void copy_page ( void * , const void * )
+clear_page extern void clear_page ( void * )
+__arch_copy_from_user extern unsigned long __attribute__ ( ( warn_unused_result ) ) __arch_copy_from_user ( void * , const void * , unsigned long )
+__arch_copy_to_user extern unsigned long __attribute__ ( ( warn_unused_result ) ) __arch_copy_to_user ( void * , const void * , unsigned long )
+__arch_clear_user extern unsigned long __attribute__ ( ( warn_unused_result ) ) __arch_clear_user ( void * , unsigned long )
+__arch_copy_in_user extern unsigned long __attribute__ ( ( warn_unused_result ) ) __arch_copy_in_user ( void * , const void * , unsigned long )
+memstart_addr extern t#s64 memstart_addr
+strchr extern char * strchr ( const char * , int )
+strrchr extern char * strrchr ( const char * , int )
+strcmp extern int strcmp ( const char * , const char * )
+strncmp extern int strncmp ( const char * , const char * , t#__kernel_size_t )
+strlen extern t#__kernel_size_t strlen ( const char * )
+strnlen extern t#__kernel_size_t strnlen ( const char * , t#__kernel_size_t )
+memset extern void * memset ( void * , int , t#__kernel_size_t )
+memcpy extern void * memcpy ( void * , const void * , t#__kernel_size_t )
+memmove extern void * memmove ( void * , const void * , t#__kernel_size_t )
+__memset extern void * __memset ( void * , int , t#__kernel_size_t )
+__memcpy extern void * __memcpy ( void * , const void * , t#__kernel_size_t )
+__memmove extern void * __memmove ( void * , const void * , t#__kernel_size_t )
+memchr extern void * memchr ( const void * , int , t#__kernel_size_t )
+memcmp extern int memcmp ( const void * , const void * , t#size_t )
+set_bit extern void set_bit ( int , volatile unsigned long * )
+test_and_set_bit extern int test_and_set_bit ( int , volatile unsigned long * )
+clear_bit extern void clear_bit ( int , volatile unsigned long * )
+test_and_clear_bit extern int test_and_clear_bit ( int , volatile unsigned long * )
+change_bit extern void change_bit ( int , volatile unsigned long * )
+test_and_change_bit extern int test_and_change_bit ( int , volatile unsigned long * )
+_mcount extern void _mcount ( unsigned long )
+s#arm_smccc_res struct arm_smccc_res { unsigned long a0 ; unsigned long a1 ; unsigned long a2 ; unsigned long a3 ; }
+s#arm_smccc_quirk struct arm_smccc_quirk { int id ; union { unsigned long a6 ; } state ; }
+__arm_smccc_smc void __arm_smccc_smc ( unsigned long , unsigned long , unsigned long , unsigned long , unsigned long , unsigned long , unsigned long , unsigned long , s#arm_smccc_res * , s#arm_smccc_quirk * )
+__arm_smccc_hvc void __arm_smccc_hvc ( unsigned long , unsigned long , unsigned long , unsigned long , unsigned long , unsigned long , unsigned long , unsigned long , s#arm_smccc_res * , s#arm_smccc_quirk * )
+
+/* arch/arm64/kernel/cpufeature.o */
+elf_hwcap extern unsigned long elf_hwcap
+cpu_hwcaps extern unsigned long cpu_hwcaps [ ( ( ( 40 ) + ( 8 * sizeof ( long ) ) - 1 ) / ( 8 * sizeof ( long ) ) ) ]
+s#static_key struct static_key { t#atomic_t enabled ; }
+s#static_key_false struct static_key_false { s#static_key key ; }
+cpu_hwcap_keys extern s#static_key_false cpu_hwcap_keys [ 40 ]
+arm64_const_caps_ready extern s#static_key_false arm64_const_caps_ready
+
+/* arch/arm64/kernel/fpsimd.o */
+kernel_neon_busy extern __attribute__ ( ( section ( ".data..percpu" "" ) ) ) __typeof__ ( t#bool ) kernel_neon_busy
+kernel_neon_begin void kernel_neon_begin ( void )
+kernel_neon_end void kernel_neon_end ( void )
+
+/* arch/arm64/kernel/hibernate.o */
+arch_hibernation_header_save int arch_hibernation_header_save ( void * , unsigned int )
+arch_hibernation_header_restore int arch_hibernation_header_restore ( void * )
+
+/* arch/arm64/kernel/io.o */
+__memcpy_fromio extern void __memcpy_fromio ( void * , const volatile void * , t#size_t )
+__memcpy_toio extern void __memcpy_toio ( volatile void * , const void * , t#size_t )
+__memset_io extern void __memset_io ( volatile void * , int , t#size_t )
+
+/* arch/arm64/kernel/pci.o */
+s##proc_dir_entry
+s##hotplug_slot
+s##sock
+s#seq_operations struct seq_operations { void * ( * start ) ( s#seq_file * , t#loff_t * ) ; void ( * stop ) ( s#seq_file * , void * ) ; void * ( * next ) ( s#seq_file * , void * , t#loff_t * ) ; int ( * show ) ( s#seq_file * , void * ) ; }
+s##vfsmount
+s#hlist_bl_node struct hlist_bl_node { s#hlist_bl_node * next , * * pprev ; }
+s#qstr struct qstr { union { struct { t#u32 hash ; t#u32 len ; } ; t#u64 hash_len ; } ; const unsigned char * name ; }
+t#__kernel_gid32_t typedef unsigned int __kernel_gid32_t
+t#gid_t typedef t#__kernel_gid32_t gid_t
+t#kgid_t typedef struct { t#gid_t val ; } kgid_t
+s##posix_acl
+s#delayed_call struct delayed_call { void ( * fn ) ( void * ) ; void * arg ; }
+t#__kernel_dev_t typedef t#__u32 __kernel_dev_t
+t#dev_t typedef t#__kernel_dev_t dev_t
+s#iattr struct iattr { unsigned int ia_valid ; t#umode_t ia_mode ; t#kuid_t ia_uid ; t#kgid_t ia_gid ; t#loff_t ia_size ; s#timespec ia_atime ; s#timespec ia_mtime ; s#timespec ia_ctime ; s#file * ia_file ; }
+t#uint32_t typedef t#__u32 uint32_t
+s#kstat struct kstat { t#u32 result_mask ; t#umode_t mode ; unsigned int nlink ; t#uint32_t blksize ; t#u64 attributes ; t#u64 attributes_mask ; t#u64 ino ; t#dev_t dev ; t#dev_t rdev ; t#kuid_t uid ; t#kgid_t gid ; t#loff_t size ; s#timespec atime ; s#timespec mtime ; s#timespec ctime ; s#timespec btime ; t#u64 blocks ; }
+s#fiemap_extent struct fiemap_extent { t#__u64 fe_logical ; t#__u64 fe_physical ; t#__u64 fe_length ; t#__u64 fe_reserved64 [ 2 ] ; t#__u32 fe_flags ; t#__u32 fe_reserved [ 3 ] ; }
+s#fiemap_extent_info struct fiemap_extent_info { unsigned int fi_flags ; unsigned int fi_extents_mapped ; unsigned int fi_extents_max ; s#fiemap_extent * fi_extents_start ; }
+s#inode_operations struct inode_operations { s#dentry * ( * lookup ) ( s#inode * , s#dentry * , unsigned int ) ; const char * ( * get_link ) ( s#dentry * , s#inode * , s#delayed_call * ) ; int ( * permission ) ( s#inode * , int ) ; s#posix_acl * ( * get_acl ) ( s#inode * , int ) ; int ( * readlink ) ( s#dentry * , char * , int ) ; int ( * create ) ( s#inode * , s#dentry * , t#umode_t , t#bool ) ; int ( * link ) ( s#dentry * , s#inode * , s#dentry * ) ; int ( * unlink ) ( s#inode * , s#dentry * ) ; int ( * symlink ) ( s#inode * , s#dentry * , const char * ) ; int ( * mkdir ) ( s#inode * , s#dentry * , t#umode_t ) ; int ( * rmdir ) ( s#inode * , s#dentry * ) ; int ( * mknod ) ( s#inode * , s#dentry * , t#umode_t , t#dev_t ) ; int ( * rename ) ( s#inode * , s#dentry * , s#inode * , s#dentry * , unsigned int ) ; int ( * setattr ) ( s#dentry * , s#iattr * ) ; int ( * getattr ) ( const s#path * , s#kstat * , t#u32 , unsigned int ) ; t#ssize_t ( * listxattr ) ( s#dentry * , char * , t#size_t ) ; int ( * fiemap ) ( s#inode * , s#fiemap_extent_info * , t#u64 , t#u64 ) ; int ( * update_time ) ( s#inode * , s#timespec * , int ) ; int ( * atomic_open ) ( s#inode * , s#dentry * , s#file * , unsigned , t#umode_t , int * ) ; int ( * tmpfile ) ( s#inode * , s#dentry * , t#umode_t ) ; int ( * set_acl ) ( s#inode * , s#posix_acl * , int ) ; }
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s#tracepoint_func struct tracepoint_func { void * func ; void * data ; int prio ; }
+s#tracepoint struct tracepoint { const char * name ; s#static_key key ; int ( * regfunc ) ( void ) ; void ( * unregfunc ) ( void ) ; s#tracepoint_func * funcs ; }
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s#lock_class_key struct lock_class_key { }
+E#SB_FREEZE_COMPLETE 4
+s#file_system_type struct file_system_type { const char * name ; int fs_flags ; s#dentry * ( * mount ) ( s#file_system_type * , int , const char * , void * ) ; void ( * kill_sb ) ( s#super_block * ) ; s#module * owner ; s#file_system_type * next ; s#hlist_head fs_supers ; s#lock_class_key s_lock_key ; s#lock_class_key s_umount_key ; s#lock_class_key s_vfs_rename_key ; s#lock_class_key s_writers_key [ ( E#SB_FREEZE_COMPLETE - 1 ) ] ; s#lock_class_key i_lock_key ; s#lock_class_key i_mutex_key ; s#lock_class_key i_mutex_dir_key ; }
+s##writeback_control
+s##kstatfs
+t#projid_t typedef t#__kernel_uid32_t projid_t
+t#kprojid_t typedef struct { t#projid_t val ; } kprojid_t
+e#quota_type enum quota_type { USRQUOTA = 0 , GRPQUOTA = 1 , PRJQUOTA = 2 , }
+s#kqid struct kqid { union { t#kuid_t uid ; t#kgid_t gid ; t#kprojid_t projid ; } ; e#quota_type type ; }
+t#qsize_t typedef long long qsize_t
+t#__s64 __extension__ typedef __signed__ long long __s64
+t#time64_t typedef t#__s64 time64_t
+s#mem_dqblk struct mem_dqblk { t#qsize_t dqb_bhardlimit ; t#qsize_t dqb_bsoftlimit ; t#qsize_t dqb_curspace ; t#qsize_t dqb_rsvspace ; t#qsize_t dqb_ihardlimit ; t#qsize_t dqb_isoftlimit ; t#qsize_t dqb_curinodes ; t#time64_t dqb_btime ; t#time64_t dqb_itime ; }
+s#dquot struct dquot { s#hlist_node dq_hash ; s#list_head dq_inuse ; s#list_head dq_free ; s#list_head dq_dirty ; s#mutex dq_lock ; t#atomic_t dq_count ; t#wait_queue_head_t dq_wait_unused ; s#super_block * dq_sb ; s#kqid dq_id ; t#loff_t dq_off ; unsigned long dq_flags ; s#mem_dqblk dq_dqb ; }
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+t#pmdval_t typedef t#u64 pmdval_t
+t#pmd_t typedef struct { t#pmdval_t pmd ; } pmd_t
+t#pudval_t typedef t#u64 pudval_t
+t#pud_t typedef struct { t#pudval_t pud ; } pud_t
+t#pte_t typedef struct { t#pteval_t pte ; } pte_t
+s##mem_cgroup
+s#vm_fault struct vm_fault { s#vm_area_struct * vma ; unsigned int flags ; t#gfp_t gfp_mask ; unsigned long pgoff ; unsigned long address ; t#pmd_t * pmd ; t#pud_t * pud ; t#pte_t orig_pte ; s#page * cow_page ; s#mem_cgroup * memcg ; s#page * page ; t#pte_t * pte ; t#spinlock_t * ptl ; t#pgtable_t prealloc_pte ; }
+e#page_entry_size enum page_entry_size { PE_SIZE_PTE = 0 , PE_SIZE_PMD , PE_SIZE_PUD , }
+s##mempolicy
+s#vm_operations_struct struct vm_operations_struct { void ( * open ) ( s#vm_area_struct * ) ; void ( * close ) ( s#vm_area_struct * ) ; int ( * split ) ( s#vm_area_struct * , unsigned long ) ; int ( * mremap ) ( s#vm_area_struct * ) ; int ( * fault ) ( s#vm_fault * ) ; int ( * huge_fault ) ( s#vm_fault * , e#page_entry_size ) ; void ( * map_pages ) ( s#vm_fault * , unsigned long , unsigned long ) ; unsigned long ( * pagesize ) ( s#vm_area_struct * ) ; int ( * page_mkwrite ) ( s#vm_fault * ) ; int ( * pfn_mkwrite ) ( s#vm_fault * ) ; int ( * access ) ( s#vm_area_struct * , unsigned long , void * , int , int ) ; const char * ( * name ) ( s#vm_area_struct * ) ; int ( * set_policy ) ( s#vm_area_struct * , s#mempolicy * ) ; s#mempolicy * ( * get_policy ) ( s#vm_area_struct * , unsigned long ) ; s#page * ( * find_special_page ) ( s#vm_area_struct * , unsigned long ) ; }
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s#pollfd struct pollfd { int fd ; short events ; short revents ; }
+s##pid_namespace
+s#kernel_cap_struct struct kernel_cap_struct { t#__u32 cap [ 2 ] ; }
+t#kernel_cap_t typedef s#kernel_cap_struct kernel_cap_t
+t#__s32 typedef __signed__ int __s32
+t#int32_t typedef t#__s32 int32_t
+t#key_serial_t typedef t#int32_t key_serial_t
+s##key_user
+t#time_t typedef t#__kernel_time_t time_t
+t#key_perm_t typedef t#uint32_t key_perm_t
+s##key_type
+s#keyring_index_key struct keyring_index_key { s#key_type * type ; const char * description ; t#size_t desc_len ; }
+u#key_payload union key_payload { void * rcu_data0 ; void * data [ 4 ] ; }
+s##assoc_array_ptr
+s#assoc_array struct assoc_array { s#assoc_array_ptr * root ; unsigned long nr_leaves_on_tree ; }
+t#key_restrict_link_func_t typedef int ( * key_restrict_link_func_t ) ( s#key * , const s#key_type * , const u#key_payload * , s#key * )
+s#key_restriction struct key_restriction { t#key_restrict_link_func_t check ; s#key * key ; s#key_type * keytype ; }
+s#key struct key { t#refcount_t usage ; t#key_serial_t serial ; union { s#list_head graveyard_link ; s#rb_node serial_node ; } ; s#rw_semaphore sem ; s#key_user * user ; void * security ; union { t#time_t expiry ; t#time_t revoked_at ; } ; t#time_t last_used_at ; t#kuid_t uid ; t#kgid_t gid ; t#key_perm_t perm ; unsigned short quotalen ; unsigned short datalen ; short state ; unsigned long flags ; union { s#keyring_index_key index_key ; struct { s#key_type * type ; char * description ; } ; } ; union { u#key_payload payload ; struct { s#list_head name_link ; s#assoc_array keys ; } ; } ; s#key_restriction * restrict_link ; }
+s#user_struct struct user_struct { t#atomic_t __count ; t#atomic_t processes ; t#atomic_t sigpending ; t#atomic_t fanotify_listeners ; t#atomic_long_t epoll_watches ; unsigned long mq_bytes ; unsigned long locked_shm ; unsigned long unix_inflight ; t#atomic_long_t pipe_bufs ; s#key * uid_keyring ; s#key * session_keyring ; s#hlist_node uidhash_node ; t#kuid_t uid ; t#atomic_long_t locked_vm ; }
+s#group_info struct group_info { t#atomic_t usage ; int ngroups ; t#kgid_t gid [ 0 ] ; }
+s#cred struct cred { t#atomic_t usage ; t#kuid_t uid ; t#kgid_t gid ; t#kuid_t suid ; t#kgid_t sgid ; t#kuid_t euid ; t#kgid_t egid ; t#kuid_t fsuid ; t#kgid_t fsgid ; unsigned securebits ; t#kernel_cap_t cap_inheritable ; t#kernel_cap_t cap_permitted ; t#kernel_cap_t cap_effective ; t#kernel_cap_t cap_bset ; t#kernel_cap_t cap_ambient ; unsigned char jit_keyring ; s#key * session_keyring ; s#key * process_keyring ; s#key * thread_keyring ; s#key * request_key_auth ; void * security ; s#user_struct * user ; s#user_namespace * user_ns ; s#group_info * group_info ; s#callback_head rcu ; }
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+t#sector_t typedef unsigned long sector_t
+e#rw_hint enum rw_hint { WRITE_LIFE_NOT_SET = 0 , WRITE_LIFE_NONE = 1 , WRITE_LIFE_SHORT = 2 , WRITE_LIFE_MEDIUM = 3 , WRITE_LIFE_LONG = 4 , WRITE_LIFE_EXTREME = 5 , }
+s#kiocb struct kiocb { s#file * ki_filp ; t#loff_t ki_pos ; void ( * ki_complete ) ( s#kiocb * , long , long ) ; void * private ; int ki_flags ; e#rw_hint ki_hint ; }
+s##iov_iter
+e#migrate_mode enum migrate_mode { MIGRATE_ASYNC , MIGRATE_SYNC_LIGHT , MIGRATE_SYNC , MIGRATE_SYNC_NO_COPY , }
+t#isolate_mode_t typedef unsigned isolate_mode_t
+s##swap_info_struct
+s#address_space_operations struct address_space_operations { int ( * writepage ) ( s#page * , s#writeback_control * ) ; int ( * readpage ) ( s#file * , s#page * ) ; int ( * writepages ) ( s#address_space * , s#writeback_control * ) ; int ( * set_page_dirty ) ( s#page * ) ; int ( * readpages ) ( s#file * , s#address_space * , s#list_head * , unsigned ) ; int ( * write_begin ) ( s#file * , s#address_space * , t#loff_t , unsigned , unsigned , s#page * * , void * * ) ; int ( * write_end ) ( s#file * , s#address_space * , t#loff_t , unsigned , unsigned , s#page * , void * ) ; t#sector_t ( * bmap ) ( s#address_space * , t#sector_t ) ; void ( * invalidatepage ) ( s#page * , unsigned int , unsigned int ) ; int ( * releasepage ) ( s#page * , t#gfp_t ) ; void ( * freepage ) ( s#page * ) ; t#ssize_t ( * direct_IO ) ( s#kiocb * , s#iov_iter * ) ; int ( * migratepage ) ( s#address_space * , s#page * , s#page * , e#migrate_mode ) ; t#bool ( * isolate_page ) ( s#page * , t#isolate_mode_t ) ; void ( * putback_page ) ( s#page * ) ; int ( * launder_page ) ( s#page * ) ; int ( * is_partially_uptodate ) ( s#page * , unsigned long , unsigned long ) ; void ( * is_dirty_writeback ) ( s#page * , t#bool * , t#bool * ) ; int ( * error_remove_page ) ( s#address_space * , s#page * ) ; int ( * swap_activate ) ( s#swap_info_struct * , s#file * , t#sector_t * ) ; void ( * swap_deactivate ) ( s#file * ) ; }
+t#errseq_t typedef t#u32 errseq_t
+s#address_space struct address_space { s#inode * host ; s#radix_tree_root page_tree ; t#spinlock_t tree_lock ; t#atomic_t i_mmap_writable ; s#rb_root i_mmap ; s#rw_semaphore i_mmap_rwsem ; unsigned long nrpages ; unsigned long nrexceptional ; unsigned long writeback_index ; const s#address_space_operations * a_ops ; unsigned long flags ; t#spinlock_t private_lock ; t#gfp_t gfp_mask ; s#list_head private_list ; void * private_data ; t#errseq_t wb_err ; }
+t#dev_page_fault_t typedef int ( * dev_page_fault_t ) ( s#vm_area_struct * , unsigned long , const s#page * , unsigned int , t#pmd_t * )
+t#dev_page_free_t typedef void ( * dev_page_free_t ) ( s#page * , void * )
+s#vmem_altmap struct vmem_altmap { const unsigned long base_pfn ; const unsigned long reserve ; unsigned long free ; unsigned long align ; unsigned long alloc ; }
+t#phys_addr_t typedef t#u64 phys_addr_t
+t#resource_size_t typedef t#phys_addr_t resource_size_t
+s#resource struct resource { t#resource_size_t start ; t#resource_size_t end ; const char * name ; unsigned long flags ; unsigned long desc ; s#resource * parent , * sibling , * child ; }
+t#percpu_ref_func_t typedef void ( percpu_ref_func_t ) ( s#percpu_ref * )
+s#percpu_ref struct percpu_ref { t#atomic_long_t count ; unsigned long percpu_count_ptr ; t#percpu_ref_func_t * release ; t#percpu_ref_func_t * confirm_switch ; t#bool force_atomic : 1 ; s#callback_head rcu ; }
+s##device_private
+s#bin_attribute struct bin_attribute { s#attribute attr ; t#size_t size ; void * private ; t#ssize_t ( * read ) ( s#file * , s#kobject * , s#bin_attribute * , char * , t#loff_t , t#size_t ) ; t#ssize_t ( * write ) ( s#file * , s#kobject * , s#bin_attribute * , char * , t#loff_t , t#size_t ) ; int ( * mmap ) ( s#file * , s#kobject * , s#bin_attribute * , s#vm_area_struct * ) ; }
+s#attribute_group struct attribute_group { const char * name ; t#umode_t ( * is_visible ) ( s#kobject * , s#attribute * , int ) ; t#umode_t ( * is_bin_visible ) ( s#kobject * , s#bin_attribute * , int ) ; s#attribute * * attrs ; s#bin_attribute * * bin_attrs ; }
+s#dev_pm_ops struct dev_pm_ops { int ( * prepare ) ( s#device * ) ; void ( * complete ) ( s#device * ) ; int ( * suspend ) ( s#device * ) ; int ( * resume ) ( s#device * ) ; int ( * freeze ) ( s#device * ) ; int ( * thaw ) ( s#device * ) ; int ( * poweroff ) ( s#device * ) ; int ( * restore ) ( s#device * ) ; int ( * suspend_late ) ( s#device * ) ; int ( * resume_early ) ( s#device * ) ; int ( * freeze_late ) ( s#device * ) ; int ( * thaw_early ) ( s#device * ) ; int ( * poweroff_late ) ( s#device * ) ; int ( * restore_early ) ( s#device * ) ; int ( * suspend_noirq ) ( s#device * ) ; int ( * resume_noirq ) ( s#device * ) ; int ( * freeze_noirq ) ( s#device * ) ; int ( * thaw_noirq ) ( s#device * ) ; int ( * poweroff_noirq ) ( s#device * ) ; int ( * restore_noirq ) ( s#device * ) ; int ( * runtime_suspend ) ( s#device * ) ; int ( * runtime_resume ) ( s#device * ) ; int ( * runtime_idle ) ( s#device * ) ; }
+s#device_type struct device_type { const char * name ; const s#attribute_group * * groups ; int ( * uevent ) ( s#device * , s#kobj_uevent_env * ) ; char * ( * devnode ) ( s#device * , t#umode_t * , t#kuid_t * , t#kgid_t * ) ; void ( * release ) ( s#device * ) ; const s#dev_pm_ops * pm ; }
+s#device_attribute struct device_attribute { s#attribute attr ; t#ssize_t ( * show ) ( s#device * , s#device_attribute * , char * ) ; t#ssize_t ( * store ) ( s#device * , s#device_attribute * , const char * , t#size_t ) ; }
+e#probe_type enum probe_type { PROBE_DEFAULT_STRATEGY , PROBE_PREFER_ASYNCHRONOUS , PROBE_FORCE_SYNCHRONOUS , }
+s#of_device_id struct of_device_id { char name [ 32 ] ; char type [ 32 ] ; char compatible [ 128 ] ; const void * data ; }
+t#__u8 typedef unsigned char __u8
+t#kernel_ulong_t typedef unsigned long kernel_ulong_t
+s#acpi_device_id struct acpi_device_id { t#__u8 id [ 9 ] ; t#kernel_ulong_t driver_data ; t#__u32 cls ; t#__u32 cls_msk ; }
+s#pm_message struct pm_message { int event ; }
+t#pm_message_t typedef s#pm_message pm_message_t
+s##driver_private
+s#device_driver struct device_driver { const char * name ; s#bus_type * bus ; s#module * owner ; const char * mod_name ; t#bool suppress_bind_attrs ; e#probe_type probe_type ; const s#of_device_id * of_match_table ; const s#acpi_device_id * acpi_match_table ; int ( * probe ) ( s#device * ) ; int ( * remove ) ( s#device * ) ; void ( * shutdown ) ( s#device * ) ; int ( * suspend ) ( s#device * , t#pm_message_t ) ; int ( * resume ) ( s#device * ) ; const s#attribute_group * * groups ; const s#dev_pm_ops * pm ; s#driver_private * p ; }
+s##iommu_ops
+s##subsys_private
+s#bus_type struct bus_type { const char * name ; const char * dev_name ; s#device * dev_root ; s#device_attribute * dev_attrs ; const s#attribute_group * * bus_groups ; const s#attribute_group * * dev_groups ; const s#attribute_group * * drv_groups ; int ( * match ) ( s#device * , s#device_driver * ) ; int ( * uevent ) ( s#device * , s#kobj_uevent_env * ) ; int ( * probe ) ( s#device * ) ; int ( * remove ) ( s#device * ) ; void ( * shutdown ) ( s#device * ) ; int ( * online ) ( s#device * ) ; int ( * offline ) ( s#device * ) ; int ( * suspend ) ( s#device * , t#pm_message_t ) ; int ( * resume ) ( s#device * ) ; int ( * num_vf ) ( s#device * ) ; int ( * dma_configure ) ( s#device * ) ; const s#dev_pm_ops * pm ; const s#iommu_ops * iommu_ops ; s#subsys_private * p ; s#lock_class_key lock_key ; }
+e#dl_dev_state enum dl_dev_state { DL_DEV_NO_DRIVER = 0 , DL_DEV_PROBING , DL_DEV_DRIVER_BOUND , DL_DEV_UNBINDING , }
+s#dev_links_info struct dev_links_info { s#list_head suppliers ; s#list_head consumers ; e#dl_dev_state status ; }
+s##wake_irq
+s#timer_list struct timer_list { s#hlist_node entry ; unsigned long expires ; void ( * function ) ( unsigned long ) ; unsigned long data ; t#u32 flags ; }
+s#wakeup_source struct wakeup_source { const char * name ; s#list_head entry ; t#spinlock_t lock ; s#wake_irq * wakeirq ; s#timer_list timer ; unsigned long timer_expires ; t#ktime_t total_time ; t#ktime_t max_time ; t#ktime_t last_time ; t#ktime_t start_prevent_time ; t#ktime_t prevent_sleep_time ; unsigned long event_count ; unsigned long active_count ; unsigned long relax_count ; unsigned long expire_count ; unsigned long wakeup_count ; t#bool active : 1 ; t#bool autosleep_enabled : 1 ; }
+e#rpm_request enum rpm_request { RPM_REQ_NONE = 0 , RPM_REQ_IDLE , RPM_REQ_SUSPEND , RPM_REQ_AUTOSUSPEND , RPM_REQ_RESUME , }
+e#rpm_status enum rpm_status { RPM_ACTIVE = 0 , RPM_RESUMING , RPM_SUSPENDED , RPM_SUSPENDING , }
+s##pm_domain_data
+s#pm_subsys_data struct pm_subsys_data { t#spinlock_t lock ; unsigned int refcount ; s#list_head clock_list ; s#pm_domain_data * domain_data ; }
+s##dev_pm_qos
+s#dev_pm_info struct dev_pm_info { t#pm_message_t power_state ; unsigned int can_wakeup : 1 ; unsigned int async_suspend : 1 ; t#bool in_dpm_list : 1 ; t#bool is_prepared : 1 ; t#bool is_suspended : 1 ; t#bool is_noirq_suspended : 1 ; t#bool is_late_suspended : 1 ; t#bool early_init : 1 ; t#bool direct_complete : 1 ; t#u32 driver_flags ; t#spinlock_t lock ; s#list_head entry ; s#completion completion ; s#wakeup_source * wakeup ; t#bool wakeup_path : 1 ; t#bool syscore : 1 ; t#bool no_pm_callbacks : 1 ; unsigned int must_resume : 1 ; unsigned int may_skip_resume : 1 ; s#timer_list suspend_timer ; unsigned long timer_expires ; s#work_struct work ; t#wait_queue_head_t wait_queue ; s#wake_irq * wakeirq ; t#atomic_t usage_count ; t#atomic_t child_count ; unsigned int disable_depth : 3 ; unsigned int idle_notification : 1 ; unsigned int request_pending : 1 ; unsigned int deferred_resume : 1 ; unsigned int runtime_auto : 1 ; t#bool ignore_children : 1 ; unsigned int no_callbacks : 1 ; unsigned int irq_safe : 1 ; unsigned int use_autosuspend : 1 ; unsigned int timer_autosuspends : 1 ; unsigned int memalloc_noio : 1 ; unsigned int links_count ; e#rpm_request request ; e#rpm_status runtime_status ; int runtime_error ; int autosuspend_delay ; unsigned long last_busy ; unsigned long active_jiffies ; unsigned long suspended_jiffies ; unsigned long accounting_timestamp ; s#pm_subsys_data * subsys_data ; void ( * set_latency_tolerance ) ( s#device * , t#s32 ) ; s#dev_pm_qos * qos ; }
+s#dev_pm_domain struct dev_pm_domain { s#dev_pm_ops ops ; void ( * detach ) ( s#device * , t#bool ) ; int ( * activate ) ( s#device * ) ; void ( * sync ) ( s#device * ) ; void ( * dismiss ) ( s#device * ) ; }
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s#dev_pin_info struct dev_pin_info { s#pinctrl * p ; s#pinctrl_state * default_state ; s#pinctrl_state * init_state ; s#pinctrl_state * sleep_state ; s#pinctrl_state * idle_state ; }
+t#dma_addr_t typedef t#u64 dma_addr_t
+s#scatterlist struct scatterlist { unsigned long page_link ; unsigned int offset ; unsigned int length ; t#dma_addr_t dma_address ; unsigned int dma_length ; }
+s#sg_table struct sg_table { s#scatterlist * sgl ; unsigned int nents ; unsigned int orig_nents ; }
+e#dma_data_direction enum dma_data_direction { DMA_BIDIRECTIONAL = 0 , DMA_TO_DEVICE = 1 , DMA_FROM_DEVICE = 2 , DMA_NONE = 3 , }
+s#dma_map_ops struct dma_map_ops { void * ( * alloc ) ( s#device * , t#size_t , t#dma_addr_t * , t#gfp_t , unsigned long ) ; void ( * free ) ( s#device * , t#size_t , void * , t#dma_addr_t , unsigned long ) ; int ( * mmap ) ( s#device * , s#vm_area_struct * , void * , t#dma_addr_t , t#size_t , unsigned long ) ; int ( * get_sgtable ) ( s#device * , s#sg_table * , void * , t#dma_addr_t , t#size_t , unsigned long ) ; t#dma_addr_t ( * map_page ) ( s#device * , s#page * , unsigned long , t#size_t , e#dma_data_direction , unsigned long ) ; void ( * unmap_page ) ( s#device * , t#dma_addr_t , t#size_t , e#dma_data_direction , unsigned long ) ; int ( * map_sg ) ( s#device * , s#scatterlist * , int , e#dma_data_direction , unsigned long ) ; void ( * unmap_sg ) ( s#device * , s#scatterlist * , int , e#dma_data_direction , unsigned long ) ; t#dma_addr_t ( * map_resource ) ( s#device * , t#phys_addr_t , t#size_t , e#dma_data_direction , unsigned long ) ; void ( * unmap_resource ) ( s#device * , t#dma_addr_t , t#size_t , e#dma_data_direction , unsigned long ) ; void ( * sync_single_for_cpu ) ( s#device * , t#dma_addr_t , t#size_t , e#dma_data_direction ) ; void ( * sync_single_for_device ) ( s#device * , t#dma_addr_t , t#size_t , e#dma_data_direction ) ; void ( * sync_sg_for_cpu ) ( s#device * , s#scatterlist * , int , e#dma_data_direction ) ; void ( * sync_sg_for_device ) ( s#device * , s#scatterlist * , int , e#dma_data_direction ) ; int ( * mapping_error ) ( s#device * , t#dma_addr_t ) ; int ( * dma_supported ) ( s#device * , t#u64 ) ; int ( * set_dma_mask ) ( s#device * , t#u64 ) ; t#size_t ( * max_mapping_size ) ( s#device * ) ; int is_phys ; }
+s#device_dma_parameters struct device_dma_parameters { unsigned int max_segment_size ; unsigned long segment_boundary_mask ; }
+s##dma_coherent_mem
+s##cma
+s#dev_archdata struct dev_archdata { void * iommu ; t#bool dma_coherent ; }
+t#phandle typedef t#u32 phandle
+s#fwnode_reference_args struct fwnode_reference_args { s#fwnode_handle * fwnode ; unsigned int nargs ; t#u64 args [ 8 ] ; }
+s#fwnode_endpoint struct fwnode_endpoint { unsigned int port ; unsigned int id ; const s#fwnode_handle * local_fwnode ; }
+s#fwnode_operations struct fwnode_operations { void ( * get ) ( s#fwnode_handle * ) ; void ( * put ) ( s#fwnode_handle * ) ; t#bool ( * device_is_available ) ( const s#fwnode_handle * ) ; t#bool ( * property_present ) ( const s#fwnode_handle * , const char * ) ; int ( * property_read_int_array ) ( const s#fwnode_handle * , const char * , unsigned int , void * , t#size_t ) ; int ( * property_read_string_array ) ( const s#fwnode_handle * , const char * , const char * * , t#size_t ) ; s#fwnode_handle * ( * get_parent ) ( const s#fwnode_handle * ) ; s#fwnode_handle * ( * get_next_child_node ) ( const s#fwnode_handle * , s#fwnode_handle * ) ; s#fwnode_handle * ( * get_named_child_node ) ( const s#fwnode_handle * , const char * ) ; int ( * get_reference_args ) ( const s#fwnode_handle * , const char * , const char * , unsigned int , unsigned int , s#fwnode_reference_args * ) ; s#fwnode_handle * ( * graph_get_next_endpoint ) ( const s#fwnode_handle * , s#fwnode_handle * ) ; s#fwnode_handle * ( * graph_get_remote_endpoint ) ( const s#fwnode_handle * ) ; s#fwnode_handle * ( * graph_get_port_parent ) ( s#fwnode_handle * ) ; int ( * graph_parse_endpoint ) ( const s#fwnode_handle * , s#fwnode_endpoint * ) ; }
+s#fwnode_handle struct fwnode_handle { s#fwnode_handle * secondary ; const s#fwnode_operations * ops ; }
+s#property struct property { char * name ; int length ; void * value ; s#property * next ; unsigned long _flags ; unsigned int unique_id ; s#bin_attribute attr ; }
+s#device_node struct device_node { const char * name ; const char * type ; t#phandle t#phandle ; const char * full_name ; s#fwnode_handle fwnode ; s#property * properties ; s#property * deadprops ; s#device_node * parent ; s#device_node * child ; s#device_node * sibling ; s#kobject kobj ; unsigned long _flags ; void * data ; }
+s#klist_node struct klist_node { void * n_klist ; s#list_head n_node ; s#kref n_ref ; }
+s#class_attribute struct class_attribute { s#attribute attr ; t#ssize_t ( * show ) ( s#class * , s#class_attribute * , char * ) ; t#ssize_t ( * store ) ( s#class * , s#class_attribute * , const char * , t#size_t ) ; }
+s#class struct class { const char * name ; s#module * owner ; s#class_attribute * class_attrs ; const s#attribute_group * * class_groups ; const s#attribute_group * * dev_groups ; s#kobject * dev_kobj ; int ( * dev_uevent ) ( s#device * , s#kobj_uevent_env * ) ; char * ( * devnode ) ( s#device * , t#umode_t * ) ; void ( * class_release ) ( s#class * ) ; void ( * dev_release ) ( s#device * ) ; int ( * suspend ) ( s#device * , t#pm_message_t ) ; int ( * resume ) ( s#device * ) ; int ( * shutdown ) ( s#device * ) ; const s#kobj_ns_type_operations * ns_type ; const void * ( * namespace ) ( s#device * ) ; const s#dev_pm_ops * pm ; s#subsys_private * p ; }
+s##iommu_group
+s##iommu_fwspec
+s#device struct device { s#device * parent ; s#device_private * p ; s#kobject kobj ; const char * init_name ; const s#device_type * type ; s#mutex mutex ; s#bus_type * bus ; s#device_driver * driver ; void * platform_data ; void * driver_data ; s#dev_links_info links ; s#dev_pm_info power ; s#dev_pm_domain * pm_domain ; s#irq_domain * msi_domain ; s#dev_pin_info * pins ; s#list_head msi_list ; int numa_node ; const s#dma_map_ops * dma_ops ; t#u64 * dma_mask ; t#u64 coherent_dma_mask ; unsigned long dma_pfn_offset ; s#device_dma_parameters * dma_parms ; s#list_head dma_pools ; s#dma_coherent_mem * dma_mem ; s#cma * cma_area ; s#dev_archdata archdata ; s#device_node * of_node ; s#fwnode_handle * fwnode ; t#dev_t devt ; t#u32 id ; t#spinlock_t devres_lock ; s#list_head devres_head ; s#klist_node knode_class ; s#class * class ; const s#attribute_group * * groups ; void ( * release ) ( s#device * ) ; s#iommu_group * iommu_group ; s#iommu_fwspec * iommu_fwspec ; t#bool offline_disabled : 1 ; t#bool offline : 1 ; }
+e#memory_type enum memory_type { MEMORY_DEVICE_PRIVATE = 1 , MEMORY_DEVICE_PUBLIC , MEMORY_DEVICE_FS_DAX , MEMORY_DEVICE_PCI_P2PDMA , }
+s#dev_pagemap struct dev_pagemap { t#dev_page_fault_t page_fault ; t#dev_page_free_t page_free ; s#vmem_altmap altmap ; t#bool altmap_valid ; s#resource res ; s#percpu_ref * ref ; void ( * kill ) ( s#percpu_ref * ) ; s#device * dev ; void * data ; e#memory_type type ; t#u64 pci_p2pdma_bus_offset ; }
+s##kmem_cache
+s#shrink_control struct shrink_control { t#gfp_t gfp_mask ; unsigned long nr_to_scan ; unsigned long nr_scanned ; int nid ; s#mem_cgroup * memcg ; }
+s#super_operations struct super_operations { s#inode * ( * alloc_inode ) ( s#super_block * ) ; void ( * destroy_inode ) ( s#inode * ) ; void ( * dirty_inode ) ( s#inode * , int ) ; int ( * write_inode ) ( s#inode * , s#writeback_control * ) ; int ( * drop_inode ) ( s#inode * ) ; void ( * evict_inode ) ( s#inode * ) ; void ( * put_super ) ( s#super_block * ) ; int ( * sync_fs ) ( s#super_block * , int ) ; int ( * freeze_super ) ( s#super_block * ) ; int ( * freeze_fs ) ( s#super_block * ) ; int ( * thaw_super ) ( s#super_block * ) ; int ( * unfreeze_fs ) ( s#super_block * ) ; int ( * statfs ) ( s#dentry * , s#kstatfs * ) ; int ( * remount_fs ) ( s#super_block * , int * , char * ) ; void ( * umount_begin ) ( s#super_block * ) ; int ( * show_options ) ( s#seq_file * , s#dentry * ) ; int ( * show_devname ) ( s#seq_file * , s#dentry * ) ; int ( * show_path ) ( s#seq_file * , s#dentry * ) ; int ( * show_stats ) ( s#seq_file * , s#dentry * ) ; t#ssize_t ( * quota_read ) ( s#super_block * , int , char * , t#size_t , t#loff_t ) ; t#ssize_t ( * quota_write ) ( s#super_block * , int , const char * , t#size_t , t#loff_t ) ; s#dquot * * ( * get_dquots ) ( s#inode * ) ; int ( * bdev_try_to_free_page ) ( s#super_block * , s#page * , t#gfp_t ) ; long ( * nr_cached_objects ) ( s#super_block * , s#shrink_control * ) ; long ( * free_cached_objects ) ( s#super_block * , s#shrink_control * ) ; t#dev_t ( * get_inode_dev ) ( const s#inode * ) ; }
+s#dquot_operations struct dquot_operations { int ( * write_dquot ) ( s#dquot * ) ; s#dquot * ( * alloc_dquot ) ( s#super_block * , int ) ; void ( * destroy_dquot ) ( s#dquot * ) ; int ( * acquire_dquot ) ( s#dquot * ) ; int ( * release_dquot ) ( s#dquot * ) ; int ( * mark_dirty ) ( s#dquot * ) ; int ( * write_info ) ( s#super_block * , int ) ; t#qsize_t * ( * get_reserved_space ) ( s#inode * ) ; int ( * get_projid ) ( s#inode * , t#kprojid_t * ) ; int ( * get_next_id ) ( s#super_block * , s#kqid * ) ; }
+s#qc_info struct qc_info { int i_fieldmask ; unsigned int i_flags ; unsigned int i_spc_timelimit ; unsigned int i_ino_timelimit ; unsigned int i_rt_spc_timelimit ; unsigned int i_spc_warnlimit ; unsigned int i_ino_warnlimit ; unsigned int i_rt_spc_warnlimit ; }
+s#qc_dqblk struct qc_dqblk { int d_fieldmask ; t#u64 d_spc_hardlimit ; t#u64 d_spc_softlimit ; t#u64 d_ino_hardlimit ; t#u64 d_ino_softlimit ; t#u64 d_space ; t#u64 d_ino_count ; t#s64 d_ino_timer ; t#s64 d_spc_timer ; int d_ino_warns ; int d_spc_warns ; t#u64 d_rt_spc_hardlimit ; t#u64 d_rt_spc_softlimit ; t#u64 d_rt_space ; t#s64 d_rt_spc_timer ; int d_rt_spc_warns ; }
+t#blkcnt_t typedef unsigned long blkcnt_t
+s#qc_type_state struct qc_type_state { unsigned int flags ; unsigned int spc_timelimit ; unsigned int ino_timelimit ; unsigned int rt_spc_timelimit ; unsigned int spc_warnlimit ; unsigned int ino_warnlimit ; unsigned int rt_spc_warnlimit ; unsigned long long ino ; t#blkcnt_t blocks ; t#blkcnt_t nextents ; }
+s#qc_state struct qc_state { unsigned int s_incoredqs ; s#qc_type_state s_state [ 3 ] ; }
+s#quotactl_ops struct quotactl_ops { int ( * quota_on ) ( s#super_block * , int , int , const s#path * ) ; int ( * quota_off ) ( s#super_block * , int ) ; int ( * quota_enable ) ( s#super_block * , unsigned int ) ; int ( * quota_disable ) ( s#super_block * , unsigned int ) ; int ( * quota_sync ) ( s#super_block * , int ) ; int ( * set_info ) ( s#super_block * , int , s#qc_info * ) ; int ( * get_dqblk ) ( s#super_block * , s#kqid , s#qc_dqblk * ) ; int ( * get_nextdqblk ) ( s#super_block * , s#kqid * , s#qc_dqblk * ) ; int ( * set_dqblk ) ( s#super_block * , s#kqid , s#qc_dqblk * ) ; int ( * get_state ) ( s#super_block * , s#qc_state * ) ; int ( * rm_xquota ) ( s#super_block * , unsigned int ) ; }
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s#hlist_bl_head struct hlist_bl_head { s#hlist_bl_node * first ; }
+s##hd_struct
+s##gendisk
+s##request_queue
+s#block_device struct block_device { t#dev_t bd_dev ; int bd_openers ; s#inode * bd_inode ; s#super_block * bd_super ; s#mutex bd_mutex ; void * bd_claiming ; void * bd_holder ; int bd_holders ; t#bool bd_write_holder ; s#list_head bd_holder_disks ; s#block_device * bd_contains ; unsigned bd_block_size ; t#u8 bd_partno ; s#hd_struct * bd_part ; unsigned bd_part_count ; int bd_invalidated ; s#gendisk * bd_disk ; s#request_queue * bd_queue ; s#backing_dev_info * bd_bdi ; s#list_head bd_list ; unsigned long bd_private ; int bd_fsfreeze_count ; s#mutex bd_fsfreeze_mutex ; }
+s##mtd_info
+s#quota_format_ops struct quota_format_ops { int ( * check_quota_file ) ( s#super_block * , int ) ; int ( * read_file_info ) ( s#super_block * , int ) ; int ( * write_file_info ) ( s#super_block * , int ) ; int ( * free_file_info ) ( s#super_block * , int ) ; int ( * read_dqblk ) ( s#dquot * ) ; int ( * commit_dqblk ) ( s#dquot * ) ; int ( * release_dqblk ) ( s#dquot * ) ; int ( * get_next_id ) ( s#super_block * , s#kqid * ) ; }
+s#quota_format_type struct quota_format_type { int qf_fmt_id ; const s#quota_format_ops * qf_ops ; s#module * qf_owner ; s#quota_format_type * qf_next ; }
+s#mem_dqinfo struct mem_dqinfo { s#quota_format_type * dqi_format ; int dqi_fmt_id ; s#list_head dqi_dirty_list ; unsigned long dqi_flags ; unsigned int dqi_bgrace ; unsigned int dqi_igrace ; t#qsize_t dqi_max_spc_limit ; t#qsize_t dqi_max_ino_limit ; void * dqi_priv ; }
+s#quota_info struct quota_info { unsigned int flags ; s#mutex dqio_mutex ; s#inode * files [ 3 ] ; s#mem_dqinfo info [ 3 ] ; const s#quota_format_ops * ops [ 3 ] ; }
+e#rcu_sync_type enum rcu_sync_type { RCU_SYNC , RCU_SCHED_SYNC , RCU_BH_SYNC }
+s#rcu_sync struct rcu_sync { int gp_state ; int gp_count ; t#wait_queue_head_t gp_wait ; int cb_state ; s#callback_head cb_head ; e#rcu_sync_type gp_type ; }
+s#rcuwait struct rcuwait { s#task_struct * task ; }
+s#percpu_rw_semaphore struct percpu_rw_semaphore { s#rcu_sync rss ; unsigned int * read_count ; s#rw_semaphore rw_sem ; s#rcuwait writer ; int readers_block ; }
+s#sb_writers struct sb_writers { int frozen ; t#wait_queue_head_t wait_unfrozen ; s#percpu_rw_semaphore rw_sem [ ( E#SB_FREEZE_COMPLETE - 1 ) ] ; }
+t#uuid_t typedef struct { t#__u8 b [ 16 ] ; } uuid_t
+t#fmode_t typedef unsigned fmode_t
+s#dentry_operations struct dentry_operations { int ( * d_revalidate ) ( s#dentry * , unsigned int ) ; int ( * d_weak_revalidate ) ( s#dentry * , unsigned int ) ; int ( * d_hash ) ( const s#dentry * , s#qstr * ) ; int ( * d_compare ) ( const s#dentry * , unsigned int , const char * , const s#qstr * ) ; int ( * d_delete ) ( const s#dentry * ) ; int ( * d_init ) ( s#dentry * ) ; void ( * d_release ) ( s#dentry * ) ; void ( * d_prune ) ( s#dentry * ) ; void ( * d_iput ) ( s#dentry * , s#inode * ) ; char * ( * d_dname ) ( s#dentry * , char * , int ) ; s#vfsmount * ( * d_automount ) ( s#path * ) ; int ( * d_manage ) ( const s#path * , t#bool ) ; s#dentry * ( * d_real ) ( s#dentry * , const s#inode * , unsigned int ) ; }
+s#shrinker struct shrinker { unsigned long ( * count_objects ) ( s#shrinker * , s#shrink_control * ) ; unsigned long ( * scan_objects ) ( s#shrinker * , s#shrink_control * ) ; int seeks ; long batch ; unsigned long flags ; s#list_head list ; t#atomic_long_t * nr_deferred ; }
+s##workqueue_struct
+s#list_lru_one struct list_lru_one { s#list_head list ; long nr_items ; }
+s#list_lru_memcg struct list_lru_memcg { s#list_lru_one * lru [ 0 ] ; }
+s#list_lru_node struct list_lru_node { t#spinlock_t lock ; s#list_lru_one lru ; s#list_lru_memcg * memcg_lrus ; long nr_items ; }
+s#list_lru struct list_lru { s#list_lru_node * node ; s#list_head list ; t#bool memcg_aware ; }
+s#super_block struct super_block { s#list_head s_list ; t#dev_t s_dev ; unsigned char s_blocksize_bits ; unsigned long s_blocksize ; t#loff_t s_maxbytes ; s#file_system_type * s_type ; const s#super_operations * s_op ; const s#dquot_operations * dq_op ; const s#quotactl_ops * s_qcop ; const s#export_operations * s_export_op ; unsigned long s_flags ; unsigned long s_iflags ; unsigned long s_magic ; s#dentry * s_root ; s#rw_semaphore s_umount ; int s_count ; t#atomic_t s_active ; void * s_security ; const s#xattr_handler * * s_xattr ; const s#fscrypt_operations * s_cop ; s#hlist_bl_head s_anon ; s#list_head s_sbdevs ; s#list_head s_mounts ; s#block_device * s_bdev ; s#backing_dev_info * s_bdi ; s#mtd_info * s_mtd ; s#hlist_node s_instances ; unsigned int s_quota_types ; s#quota_info s_dquot ; s#sb_writers s_writers ; char s_id [ 32 ] ; t#uuid_t s_uuid ; void * s_fs_info ; unsigned int s_max_links ; t#fmode_t s_mode ; t#u32 s_time_gran ; s#mutex s_vfs_rename_mutex ; char * s_subtype ; char * s_options ; const s#dentry_operations * s_d_op ; int cleancache_poolid ; s#shrinker s_shrink ; t#atomic_long_t s_remove_count ; t#atomic_long_t s_fsnotify_inode_refs ; int s_readonly_remount ; s#workqueue_struct * s_dio_done_wq ; s#hlist_head s_pins ; s#user_namespace * s_user_ns ; s#list_lru s_dentry_lru __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; s#list_lru s_inode_lru __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; s#callback_head rcu ; s#work_struct destroy_work ; s#mutex s_sync_lock ; int s_stack_depth ; t#spinlock_t s_inode_list_lock __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; s#list_head s_inodes ; t#spinlock_t s_inode_wblist_lock ; s#list_head s_inodes_wb ; }
+s##bdi_writeback
+t#filldir_t typedef int ( * filldir_t ) ( s#dir_context * , const char * , int , t#loff_t , t#u64 , unsigned )
+s#dir_context struct dir_context { const t#filldir_t actor ; t#loff_t pos ; }
+t#poll_queue_proc typedef void ( * poll_queue_proc ) ( s#file * , t#wait_queue_head_t * , s#poll_table_struct * )
+s#poll_table_struct struct poll_table_struct { t#poll_queue_proc _qproc ; unsigned long _key ; }
+t#fl_owner_t typedef void * fl_owner_t
+s#fasync_struct struct fasync_struct { t#spinlock_t fa_lock ; int magic ; int fa_fd ; s#fasync_struct * fa_next ; s#file * fa_file ; s#callback_head fa_rcu ; }
+s#file_lock_operations struct file_lock_operations { void ( * fl_copy_lock ) ( s#file_lock * , s#file_lock * ) ; void ( * fl_release_private ) ( s#file_lock * ) ; }
+s#lock_manager_operations struct lock_manager_operations { int ( * lm_compare_owner ) ( s#file_lock * , s#file_lock * ) ; unsigned long ( * lm_owner_key ) ( s#file_lock * ) ; t#fl_owner_t ( * lm_get_owner ) ( t#fl_owner_t ) ; void ( * lm_put_owner ) ( t#fl_owner_t ) ; void ( * lm_notify ) ( s#file_lock * ) ; int ( * lm_grant ) ( s#file_lock * , int ) ; t#bool ( * lm_break ) ( s#file_lock * ) ; int ( * lm_change ) ( s#file_lock * , int , s#list_head * ) ; void ( * lm_setup ) ( s#file_lock * , void * * ) ; }
+s##nlm_lockowner
+s#nfs_lock_info struct nfs_lock_info { t#u32 state ; s#nlm_lockowner * owner ; s#list_head list ; }
+s##nfs4_lock_state
+s#nfs4_lock_info struct nfs4_lock_info { s#nfs4_lock_state * owner ; }
+s#file_lock struct file_lock { s#file_lock * fl_next ; s#list_head fl_list ; s#hlist_node fl_link ; s#list_head fl_block ; t#fl_owner_t fl_owner ; unsigned int fl_flags ; unsigned char fl_type ; unsigned int fl_pid ; int fl_link_cpu ; s#pid * fl_nspid ; t#wait_queue_head_t fl_wait ; s#file * fl_file ; t#loff_t fl_start ; t#loff_t fl_end ; s#fasync_struct * fl_fasync ; unsigned long fl_break_time ; unsigned long fl_downgrade_time ; const s#file_lock_operations * fl_ops ; const s#lock_manager_operations * fl_lmops ; union { s#nfs_lock_info nfs_fl ; s#nfs4_lock_info nfs4_fl ; struct { s#list_head link ; int state ; } afs ; } fl_u ; }
+s#file_operations struct file_operations { s#module * owner ; t#loff_t ( * llseek ) ( s#file * , t#loff_t , int ) ; t#ssize_t ( * read ) ( s#file * , char * , t#size_t , t#loff_t * ) ; t#ssize_t ( * write ) ( s#file * , const char * , t#size_t , t#loff_t * ) ; t#ssize_t ( * read_iter ) ( s#kiocb * , s#iov_iter * ) ; t#ssize_t ( * write_iter ) ( s#kiocb * , s#iov_iter * ) ; int ( * iterate ) ( s#file * , s#dir_context * ) ; int ( * iterate_shared ) ( s#file * , s#dir_context * ) ; unsigned int ( * poll ) ( s#file * , s#poll_table_struct * ) ; long ( * unlocked_ioctl ) ( s#file * , unsigned int , unsigned long ) ; long ( * compat_ioctl ) ( s#file * , unsigned int , unsigned long ) ; int ( * mmap ) ( s#file * , s#vm_area_struct * ) ; unsigned long mmap_supported_flags ; int ( * open ) ( s#inode * , s#file * ) ; int ( * flush ) ( s#file * , t#fl_owner_t ) ; int ( * release ) ( s#inode * , s#file * ) ; int ( * fsync ) ( s#file * , t#loff_t , t#loff_t , int ) ; int ( * fasync ) ( int , s#file * , int ) ; int ( * lock ) ( s#file * , int , s#file_lock * ) ; t#ssize_t ( * sendpage ) ( s#file * , s#page * , int , t#size_t , t#loff_t * , int ) ; unsigned long ( * get_unmapped_area ) ( s#file * , unsigned long , unsigned long , unsigned long , unsigned long ) ; int ( * check_flags ) ( int ) ; int ( * flock ) ( s#file * , int , s#file_lock * ) ; t#ssize_t ( * splice_write ) ( s#pipe_inode_info * , s#file * , t#loff_t * , t#size_t , unsigned int ) ; t#ssize_t ( * splice_read ) ( s#file * , t#loff_t * , s#pipe_inode_info * , t#size_t , unsigned int ) ; int ( * setlease ) ( s#file * , long , s#file_lock * * , void * * ) ; long ( * fallocate ) ( s#file * , int , t#loff_t , t#loff_t ) ; void ( * show_fdinfo ) ( s#seq_file * , s#file * ) ; t#ssize_t ( * copy_file_range ) ( s#file * , t#loff_t , s#file * , t#loff_t , t#size_t , unsigned int ) ; int ( * clone_file_range ) ( s#file * , t#loff_t , s#file * , t#loff_t , t#u64 ) ; t#ssize_t ( * dedupe_file_range ) ( s#file * , t#u64 , t#u64 , s#file * , t#u64 ) ; }
+s#file_lock_context struct file_lock_context { t#spinlock_t flc_lock ; s#list_head flc_flock ; s#list_head flc_posix ; s#list_head flc_lease ; }
+s#cdev struct cdev { s#kobject kobj ; s#module * owner ; const s#file_operations * ops ; s#list_head list ; t#dev_t dev ; unsigned int count ; }
+s##fsnotify_mark_connector
+s##fscrypt_info
+s#inode struct inode { t#umode_t i_mode ; unsigned short i_opflags ; t#kuid_t i_uid ; t#kgid_t i_gid ; unsigned int i_flags ; s#posix_acl * i_acl ; s#posix_acl * i_default_acl ; const s#inode_operations * i_op ; s#super_block * i_sb ; s#address_space * i_mapping ; void * i_security ; unsigned long i_ino ; union { const unsigned int i_nlink ; unsigned int __i_nlink ; } ; t#dev_t i_rdev ; t#loff_t i_size ; s#timespec i_atime ; s#timespec i_mtime ; s#timespec i_ctime ; t#spinlock_t i_lock ; unsigned short i_bytes ; unsigned int i_blkbits ; e#rw_hint i_write_hint ; t#blkcnt_t i_blocks ; unsigned long i_state ; s#rw_semaphore i_rwsem ; unsigned long dirtied_when ; unsigned long dirtied_time_when ; s#hlist_node i_hash ; s#list_head i_io_list ; s#bdi_writeback * i_wb ; int i_wb_frn_winner ; t#u16 i_wb_frn_avg_time ; t#u16 i_wb_frn_history ; s#list_head i_lru ; s#list_head i_sb_list ; s#list_head i_wb_list ; union { s#hlist_head i_dentry ; s#callback_head i_rcu ; } ; t#atomic64_t i_version ; t#atomic_t i_count ; t#atomic_t i_dio_count ; t#atomic_t i_writecount ; t#atomic_t i_readcount ; const s#file_operations * i_fop ; s#file_lock_context * i_flctx ; s#address_space i_data ; s#list_head i_devices ; union { s#pipe_inode_info * i_pipe ; s#block_device * i_bdev ; s#cdev * i_cdev ; char * i_link ; unsigned i_dir_seq ; } ; t#__u32 i_generation ; t#__u32 i_fsnotify_mask ; s#fsnotify_mark_connector * i_fsnotify_marks ; s#fscrypt_info * i_crypt_info ; void * i_private ; }
+s#lockref struct lockref { union { t#__u64 __attribute__ ( ( aligned ( 8 ) ) ) lock_count ; struct { t#spinlock_t lock ; int count ; } ; } ; }
+s#dentry struct dentry { unsigned int d_flags ; t#seqcount_t d_seq ; s#hlist_bl_node d_hash ; s#dentry * d_parent ; s#qstr d_name ; s#inode * d_inode ; unsigned char d_iname [ 32 ] ; s#lockref d_lockref ; const s#dentry_operations * d_op ; s#super_block * d_sb ; unsigned long d_time ; void * d_fsdata ; union { s#list_head d_lru ; t#wait_queue_head_t * d_wait ; } ; s#list_head d_child ; s#list_head d_subdirs ; union { s#hlist_node d_alias ; s#hlist_bl_node d_in_lookup_hash ; s#callback_head d_rcu ; } d_u ; }
+s#path struct path { s#vfsmount * mnt ; s#dentry * dentry ; }
+t#arch_rwlock_t typedef struct { volatile unsigned int lock ; } arch_rwlock_t
+t#rwlock_t typedef struct { t#arch_rwlock_t raw_lock ; } rwlock_t
+e#pid_type enum pid_type { PIDTYPE_PID , PIDTYPE_PGID , PIDTYPE_SID , PIDTYPE_MAX , __PIDTYPE_TGID }
+s#fown_struct struct fown_struct { t#rwlock_t lock ; s#pid * pid ; e#pid_type pid_type ; t#kuid_t uid , euid ; int signum ; }
+s#file_ra_state struct file_ra_state { unsigned long start ; unsigned int size ; unsigned int async_size ; unsigned int ra_pages ; unsigned int mmap_miss ; t#loff_t prev_pos ; }
+s#file struct file { union { s#llist_node fu_llist ; s#callback_head fu_rcuhead ; } f_u ; s#path f_path ; s#inode * f_inode ; const s#file_operations * f_op ; t#spinlock_t f_lock ; e#rw_hint f_write_hint ; t#atomic_long_t f_count ; unsigned int f_flags ; t#fmode_t f_mode ; s#mutex f_pos_lock ; t#loff_t f_pos ; s#fown_struct f_owner ; const s#cred * f_cred ; s#file_ra_state f_ra ; t#u64 f_version ; void * f_security ; void * private_data ; s#list_head f_ep_links ; s#list_head f_tfile_llink ; s#address_space * f_mapping ; t#errseq_t f_wb_err ; }
+s#seq_file struct seq_file { char * buf ; t#size_t size ; t#size_t from ; t#size_t count ; t#size_t pad_until ; t#loff_t index ; t#loff_t read_pos ; t#u64 version ; s#mutex lock ; const s#seq_operations * op ; int poll_event ; const s#file * file ; void * private ; }
+s##kernfs_open_node
+s##kernfs_iattrs
+s#pci_slot struct pci_slot { s#pci_bus * bus ; s#list_head list ; s#hotplug_slot * hotplug ; unsigned char number ; s#kobject kobj ; }
+s#pci_device_id struct pci_device_id { t#__u32 vendor , device ; t#__u32 subvendor , subdevice ; t#__u32 class , class_mask ; t#kernel_ulong_t driver_data ; }
+t#pci_ers_result_t typedef unsigned int pci_ers_result_t
+e#pci_channel_state enum pci_channel_state { pci_channel_io_normal = ( pci_channel_state_t ) 1 , pci_channel_io_frozen = ( pci_channel_state_t ) 2 , pci_channel_io_perm_failure = ( pci_channel_state_t ) 3 , }
+s#pci_error_handlers struct pci_error_handlers { t#pci_ers_result_t ( * error_detected ) ( s#pci_dev * , e#pci_channel_state ) ; t#pci_ers_result_t ( * mmio_enabled ) ( s#pci_dev * ) ; t#pci_ers_result_t ( * slot_reset ) ( s#pci_dev * ) ; void ( * reset_prepare ) ( s#pci_dev * ) ; void ( * reset_done ) ( s#pci_dev * ) ; void ( * resume ) ( s#pci_dev * ) ; }
+s#pci_dynids struct pci_dynids { t#spinlock_t lock ; s#list_head list ; }
+s#pci_driver struct pci_driver { s#list_head node ; const char * name ; const s#pci_device_id * id_table ; int ( * probe ) ( s#pci_dev * , const s#pci_device_id * ) ; void ( * remove ) ( s#pci_dev * ) ; int ( * suspend ) ( s#pci_dev * , t#pm_message_t ) ; int ( * suspend_late ) ( s#pci_dev * , t#pm_message_t ) ; int ( * resume_early ) ( s#pci_dev * ) ; int ( * resume ) ( s#pci_dev * ) ; void ( * shutdown ) ( s#pci_dev * ) ; int ( * sriov_configure ) ( s#pci_dev * , int ) ; const s#pci_error_handlers * err_handler ; const s#attribute_group * * groups ; s#device_driver driver ; s#pci_dynids dynids ; }
+t#pci_power_t typedef int pci_power_t
+s##pcie_link_state
+t#pci_channel_state_t typedef unsigned int pci_channel_state_t
+E#PCI_IOV_RESOURCES ( 5 ) + 2
+E#PCI_BRIDGE_RESOURCES ( E#PCI_IOV_RESOURCES + 6 - 1 ) + 1
+E#PCI_NUM_RESOURCES ( E#PCI_BRIDGE_RESOURCES + 4 - 1 ) + 1
+E#DEVICE_COUNT_RESOURCE E#PCI_NUM_RESOURCES
+t#pci_dev_flags_t typedef unsigned short pci_dev_flags_t
+s##pci_vpd
+s##pci_sriov
+s#pci_dev struct pci_dev { s#list_head bus_list ; s#pci_bus * bus ; s#pci_bus * subordinate ; void * sysdata ; s#proc_dir_entry * procent ; s#pci_slot * slot ; unsigned int devfn ; unsigned short vendor ; unsigned short device ; unsigned short subsystem_vendor ; unsigned short subsystem_device ; unsigned int class ; t#u8 revision ; t#u8 hdr_type ; t#u16 aer_cap ; t#u8 pcie_cap ; t#u8 msi_cap ; t#u8 msix_cap ; t#u8 pcie_mpss : 3 ; t#u8 rom_base_reg ; t#u8 pin ; t#u16 pcie_flags_reg ; unsigned long * dma_alias_mask ; s#pci_driver * driver ; t#u64 dma_mask ; s#device_dma_parameters dma_parms ; t#pci_power_t current_state ; unsigned int imm_ready : 1 ; t#u8 pm_cap ; unsigned int pme_support : 5 ; unsigned int pme_poll : 1 ; unsigned int d1_support : 1 ; unsigned int d2_support : 1 ; unsigned int no_d1d2 : 1 ; unsigned int no_d3cold : 1 ; unsigned int bridge_d3 : 1 ; unsigned int d3cold_allowed : 1 ; unsigned int mmio_always_on : 1 ; unsigned int wakeup_prepared : 1 ; unsigned int runtime_d3cold : 1 ; unsigned int skip_bus_pm : 1 ; unsigned int ignore_hotplug : 1 ; unsigned int hotplug_user_indicators : 1 ; unsigned int clear_retrain_link : 1 ; unsigned int d3_delay ; unsigned int d3cold_delay ; s#pcie_link_state * link_state ; unsigned int ltr_path : 1 ; unsigned int eetlp_prefix_path : 1 ; t#pci_channel_state_t error_state ; s#device dev ; int cfg_size ; unsigned int irq ; s#resource resource [ E#DEVICE_COUNT_RESOURCE ] ; t#bool match_driver ; unsigned int transparent : 1 ; unsigned int multifunction : 1 ; unsigned int is_busmaster : 1 ; unsigned int no_msi : 1 ; unsigned int no_64bit_msi : 1 ; unsigned int block_cfg_access : 1 ; unsigned int broken_parity_status : 1 ; unsigned int irq_reroute_variant : 2 ; unsigned int msi_enabled : 1 ; unsigned int msix_enabled : 1 ; unsigned int ari_enabled : 1 ; unsigned int ats_enabled : 1 ; unsigned int pasid_enabled : 1 ; unsigned int pri_enabled : 1 ; unsigned int is_managed : 1 ; unsigned int needs_freset : 1 ; unsigned int state_saved : 1 ; unsigned int is_physfn : 1 ; unsigned int is_virtfn : 1 ; unsigned int reset_fn : 1 ; unsigned int is_hotplug_bridge : 1 ; unsigned int is_thunderbolt : 1 ; unsigned int untrusted : 1 ; unsigned int __aer_firmware_first_valid : 1 ; unsigned int __aer_firmware_first : 1 ; unsigned int broken_intx_masking : 1 ; unsigned int io_window_1k : 1 ; unsigned int irq_managed : 1 ; unsigned int has_secondary_link : 1 ; unsigned int non_compliant_bars : 1 ; unsigned int no_vf_scan : 1 ; t#pci_dev_flags_t dev_flags ; t#atomic_t enable_cnt ; t#u32 saved_config_space [ 16 ] ; s#hlist_head saved_cap_space ; s#bin_attribute * rom_attr ; int rom_attr_enabled ; s#bin_attribute * res_attr [ E#DEVICE_COUNT_RESOURCE ] ; s#bin_attribute * res_attr_wc [ E#DEVICE_COUNT_RESOURCE ] ; unsigned int broken_cmd_compl : 1 ; unsigned int ptm_root : 1 ; unsigned int ptm_enabled : 1 ; t#u8 ptm_granularity ; const s#attribute_group * * msi_irq_groups ; s#pci_vpd * vpd ; union { s#pci_sriov * sriov ; s#pci_dev * physfn ; } ; t#u16 ats_cap ; t#u8 ats_stu ; t#atomic_t ats_ref_cnt ; t#u32 pri_reqs_alloc ; t#u16 pasid_features ; t#phys_addr_t rom ; t#size_t romlen ; char * driver_override ; unsigned long priv_flags ; }
+s#pci_ops struct pci_ops { int ( * add_bus ) ( s#pci_bus * ) ; void ( * remove_bus ) ( s#pci_bus * ) ; void * ( * map_bus ) ( s#pci_bus * , unsigned int , int ) ; int ( * read ) ( s#pci_bus * , unsigned int , int , int , t#u32 * ) ; int ( * write ) ( s#pci_bus * , unsigned int , int , int , t#u32 ) ; }
+s#msi_msg struct msi_msg { t#u32 address_lo ; t#u32 address_hi ; t#u32 data ; }
+s##platform_msi_priv_data
+s#platform_msi_desc struct platform_msi_desc { s#platform_msi_priv_data * msi_priv_data ; t#u16 msi_index ; }
+s#fsl_mc_msi_desc struct fsl_mc_msi_desc { t#u16 msi_index ; }
+s#msi_desc struct msi_desc { s#list_head list ; unsigned int irq ; unsigned int nvec_used ; s#device * dev ; s#msi_msg msg ; s#cpumask * affinity ; union { struct { t#u32 masked ; struct { t#__u8 is_msix : 1 ; t#__u8 multiple : 3 ; t#__u8 multi_cap : 3 ; t#__u8 maskbit : 1 ; t#__u8 is_64 : 1 ; t#__u16 entry_nr ; unsigned default_irq ; } msi_attrib ; union { t#u8 mask_pos ; void * mask_base ; } ; } ; s#platform_msi_desc platform ; s#fsl_mc_msi_desc fsl_mc ; } ; }
+s#msi_controller struct msi_controller { s#module * owner ; s#device * dev ; s#device_node * of_node ; s#list_head list ; int ( * setup_irq ) ( s#msi_controller * , s#pci_dev * , s#msi_desc * ) ; int ( * setup_irqs ) ( s#msi_controller * , s#pci_dev * , int , int ) ; void ( * teardown_irq ) ( s#msi_controller * , unsigned int ) ; }
+t#pci_bus_flags_t typedef unsigned short pci_bus_flags_t
+s#pci_bus struct pci_bus { s#list_head node ; s#pci_bus * parent ; s#list_head children ; s#list_head devices ; s#pci_dev * self ; s#list_head slots ; s#resource * resource [ 4 ] ; s#list_head resources ; s#resource busn_res ; s#pci_ops * ops ; s#msi_controller * msi ; void * sysdata ; s#proc_dir_entry * procdir ; unsigned char number ; unsigned char primary ; unsigned char max_bus_speed ; unsigned char cur_bus_speed ; int domain_nr ; char name [ 48 ] ; unsigned short bridge_ctl ; t#pci_bus_flags_t bus_flags ; s#device * bridge ; s#device dev ; s#bin_attribute * legacy_io ; s#bin_attribute * legacy_mem ; unsigned int is_added : 1 ; }
+pcibus_to_node int pcibus_to_node ( s#pci_bus * )
+
+/* arch/arm64/kernel/process.o */
+__stack_chk_guard extern unsigned long __stack_chk_guard
+pm_power_off extern void ( * pm_power_off ) ( void )
+
+/* arch/arm64/kernel/return_address.o */
+return_address extern void * return_address ( unsigned int )
+
+/* arch/arm64/kernel/smp.o */
+cpu_number extern __attribute__ ( ( section ( ".data..percpu" "..read_mostly" ) ) ) __typeof__ ( int ) cpu_number
+
+/* arch/arm64/kernel/stacktrace.o */
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##posix_acl
+s##vfsmount
+s##module
+s##writeback_control
+s##kstatfs
+s##seq_file
+s##mem_cgroup
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##backing_dev_info
+s##mtd_info
+s##workqueue_struct
+s##user_namespace
+s#iovec struct iovec { void * iov_base ; t#__kernel_size_t iov_len ; }
+s#kvec struct kvec { void * iov_base ; t#size_t iov_len ; }
+s##bio_vec
+s##pipe_inode_info
+s#iov_iter struct iov_iter { int type ; t#size_t iov_offset ; t#size_t count ; union { const s#iovec * iov ; const s#kvec * kvec ; const s#bio_vec * bvec ; s#pipe_inode_info * pipe ; } ; union { unsigned long nr_segs ; struct { int idx ; int start_idx ; } ; } ; }
+s##poll_table_struct
+s#pidmap struct pidmap { t#atomic_t nr_free ; void * page ; }
+s##kmem_cache
+s##fs_pin
+s##ucounts
+s##proc_ns_operations
+s#ns_common struct ns_common { t#atomic_long_t stashed ; const s#proc_ns_operations * ops ; unsigned int inum ; }
+s#pid_namespace struct pid_namespace { s#kref kref ; s#pidmap pidmap [ ( ( ( 0 ? ( ( 1UL ) << 12 ) * 8 : ( sizeof ( long ) > 4 ? 4 * 1024 * 1024 : ( 0 ? 0x1000 : 0x8000 ) ) ) + ( ( ( 1UL ) << 12 ) * 8 ) - 1 ) / ( ( ( 1UL ) << 12 ) * 8 ) ) ] ; s#callback_head rcu ; int last_pid ; unsigned int nr_hashed ; s#task_struct * child_reaper ; s#kmem_cache * pid_cachep ; unsigned int level ; s#pid_namespace * parent ; s#vfsmount * proc_mnt ; s#dentry * proc_self ; s#dentry * proc_thread_self ; s#fs_pin * bacct ; s#user_namespace * user_ns ; s#ucounts * ucounts ; s#work_struct proc_work ; t#kgid_t pid_gid ; int hide_pid ; int reboot ; s#ns_common ns ; }
+s##nlm_lockowner
+s##nfs4_lock_state
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##bdi_writeback
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##swap_info_struct
+s##device
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##cgroup_namespace
+s#nsproxy struct nsproxy { t#atomic_t count ; s#uts_namespace * uts_ns ; s#ipc_namespace * ipc_ns ; s#mnt_namespace * mnt_ns ; s#pid_namespace * pid_ns_for_children ; s#net * net_ns ; s#cgroup_namespace * cgroup_ns ; }
+s#cpu_itimer struct cpu_itimer { t#u64 expires ; t#u64 incr ; }
+s#task_cputime_atomic struct task_cputime_atomic { t#atomic64_t utime ; t#atomic64_t stime ; t#atomic64_t sum_exec_runtime ; }
+s#thread_group_cputimer struct thread_group_cputimer { s#task_cputime_atomic cputime_atomic ; t#bool running ; t#bool checking_timer ; }
+s##tty_struct
+t#seqlock_t typedef struct { s#seqcount seqcount ; t#spinlock_t lock ; } seqlock_t
+s#rlimit struct rlimit { t#__kernel_ulong_t rlim_cur ; t#__kernel_ulong_t rlim_max ; }
+s#pacct_struct struct pacct_struct { int ac_flag ; long ac_exitcode ; unsigned long ac_mem ; t#u64 ac_utime , ac_stime ; unsigned long ac_minflt , ac_majflt ; }
+s##taskstats
+s##tty_audit_buf
+s#signal_struct struct signal_struct { t#atomic_t sigcnt ; t#atomic_t live ; int nr_threads ; s#list_head thread_head ; t#wait_queue_head_t wait_chldexit ; s#task_struct * curr_target ; s#sigpending shared_pending ; int group_exit_code ; int notify_count ; s#task_struct * group_exit_task ; int group_stop_count ; unsigned int flags ; unsigned int is_child_subreaper : 1 ; unsigned int has_child_subreaper : 1 ; int posix_timer_id ; s#list_head posix_timers ; s#hrtimer real_timer ; t#ktime_t it_real_incr ; s#cpu_itimer it [ 2 ] ; s#thread_group_cputimer cputimer ; s#task_cputime cputime_expires ; s#list_head cpu_timers [ 3 ] ; s#pid * leader_pid ; t#atomic_t tick_dep_mask ; s#pid * tty_old_pgrp ; int leader ; s#tty_struct * tty ; t#seqlock_t stats_lock ; t#u64 utime , stime , cutime , cstime ; t#u64 gtime ; t#u64 cgtime ; s#prev_cputime prev_cputime ; unsigned long nvcsw , nivcsw , cnvcsw , cnivcsw ; unsigned long min_flt , maj_flt , cmin_flt , cmaj_flt ; unsigned long inblock , oublock , cinblock , coublock ; unsigned long maxrss , cmaxrss ; s#task_io_accounting ioac ; unsigned long long sum_sched_runtime ; s#rlimit rlim [ 16 ] ; s#pacct_struct pacct ; s#taskstats * stats ; unsigned audit_tty ; s#tty_audit_buf * tty_audit_buf ; t#bool oom_flag_origin ; short oom_score_adj ; short oom_score_adj_min ; s#mm_struct * oom_mm ; s#mutex cred_guard_mutex ; void * suse_kabi_padding ; }
+t#__signalfn_t typedef void __signalfn_t ( int )
+t#__sighandler_t typedef t#__signalfn_t * __sighandler_t
+t#__restorefn_t typedef void __restorefn_t ( void )
+t#__sigrestore_t typedef t#__restorefn_t * __sigrestore_t
+s#sigaction struct sigaction { t#__sighandler_t sa_handler ; unsigned long sa_flags ; t#__sigrestore_t sa_restorer ; t#sigset_t sa_mask ; }
+s#k_sigaction struct k_sigaction { s#sigaction sa ; }
+s#sighand_struct struct sighand_struct { t#atomic_t count ; s#k_sigaction action [ 64 ] ; t#spinlock_t siglock ; t#wait_queue_head_t signalfd_wqh ; }
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##io_context
+s##css_set
+s##robust_list_head
+t#compat_uptr_t typedef t#u32 compat_uptr_t
+s#compat_robust_list struct compat_robust_list { t#compat_uptr_t next ; }
+t#compat_long_t typedef t#s32 compat_long_t
+s#compat_robust_list_head struct compat_robust_list_head { s#compat_robust_list list ; t#compat_long_t futex_offset ; t#compat_uptr_t list_op_pending ; }
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s#ftrace_ret_stack struct ftrace_ret_stack { unsigned long ret ; unsigned long func ; unsigned long long calltime ; unsigned long long subtime ; }
+s##uprobe
+s##perf_event
+s#stack_trace struct stack_trace { unsigned int nr_entries , max_entries ; unsigned long * entries ; int skip ; }
+save_stack_trace_tsk extern void save_stack_trace_tsk ( s#task_struct * , s#stack_trace * )
+save_stack_trace extern void save_stack_trace ( s#stack_trace * )
+
+/* arch/arm64/kernel/time.o */
+s#user_pt_regs struct user_pt_regs { t#__u64 regs [ 31 ] ; t#__u64 sp ; t#__u64 pc ; t#__u64 pstate ; }
+s#pt_regs struct pt_regs { union { s#user_pt_regs user_regs ; struct { t#u64 regs [ 31 ] ; t#u64 sp ; t#u64 pc ; t#u64 pstate ; } ; } ; t#u64 orig_x0 ; t#s32 syscallno ; t#u32 unused2 ; t#u64 orig_addr_limit ; t#u64 unused ; t#u64 stackframe [ 2 ] ; }
+profile_pc extern unsigned long profile_pc ( s#pt_regs * )
+
+/* arch/arm64/kernel/topology.o */
+s#cpu_topology struct cpu_topology { int thread_id ; int core_id ; int package_id ; int llc_id ; t#cpumask_t thread_sibling ; t#cpumask_t core_sibling ; t#cpumask_t llc_siblings ; }
+cpu_topology extern s#cpu_topology cpu_topology [ 480 ]
+
+/* arch/arm64/lib/atomic_ll_sc.o */
+__ll_sc_atomic_add __attribute__ ( ( no_instrument_function ) ) void __ll_sc_atomic_add ( int , t#atomic_t * )
+__ll_sc_atomic_add_return __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_add_return ( int , t#atomic_t * )
+__ll_sc_atomic_add_return_relaxed __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_add_return_relaxed ( int , t#atomic_t * )
+__ll_sc_atomic_add_return_acquire __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_add_return_acquire ( int , t#atomic_t * )
+__ll_sc_atomic_add_return_release __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_add_return_release ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_add __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_add ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_add_relaxed __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_add_relaxed ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_add_acquire __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_add_acquire ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_add_release __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_add_release ( int , t#atomic_t * )
+__ll_sc_atomic_sub __attribute__ ( ( no_instrument_function ) ) void __ll_sc_atomic_sub ( int , t#atomic_t * )
+__ll_sc_atomic_sub_return __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_sub_return ( int , t#atomic_t * )
+__ll_sc_atomic_sub_return_relaxed __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_sub_return_relaxed ( int , t#atomic_t * )
+__ll_sc_atomic_sub_return_acquire __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_sub_return_acquire ( int , t#atomic_t * )
+__ll_sc_atomic_sub_return_release __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_sub_return_release ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_sub __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_sub ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_sub_relaxed __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_sub_relaxed ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_sub_acquire __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_sub_acquire ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_sub_release __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_sub_release ( int , t#atomic_t * )
+__ll_sc_atomic_and __attribute__ ( ( no_instrument_function ) ) void __ll_sc_atomic_and ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_and __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_and ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_and_relaxed __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_and_relaxed ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_and_acquire __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_and_acquire ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_and_release __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_and_release ( int , t#atomic_t * )
+__ll_sc_atomic_andnot __attribute__ ( ( no_instrument_function ) ) void __ll_sc_atomic_andnot ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_andnot __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_andnot ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_andnot_relaxed __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_andnot_relaxed ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_andnot_acquire __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_andnot_acquire ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_andnot_release __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_andnot_release ( int , t#atomic_t * )
+__ll_sc_atomic_or __attribute__ ( ( no_instrument_function ) ) void __ll_sc_atomic_or ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_or __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_or ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_or_relaxed __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_or_relaxed ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_or_acquire __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_or_acquire ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_or_release __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_or_release ( int , t#atomic_t * )
+__ll_sc_atomic_xor __attribute__ ( ( no_instrument_function ) ) void __ll_sc_atomic_xor ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_xor __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_xor ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_xor_relaxed __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_xor_relaxed ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_xor_acquire __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_xor_acquire ( int , t#atomic_t * )
+__ll_sc_atomic_fetch_xor_release __attribute__ ( ( no_instrument_function ) ) int __ll_sc_atomic_fetch_xor_release ( int , t#atomic_t * )
+__ll_sc_atomic64_add __attribute__ ( ( no_instrument_function ) ) void __ll_sc_atomic64_add ( long , t#atomic64_t * )
+__ll_sc_atomic64_add_return __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_add_return ( long , t#atomic64_t * )
+__ll_sc_atomic64_add_return_relaxed __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_add_return_relaxed ( long , t#atomic64_t * )
+__ll_sc_atomic64_add_return_acquire __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_add_return_acquire ( long , t#atomic64_t * )
+__ll_sc_atomic64_add_return_release __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_add_return_release ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_add __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_add ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_add_relaxed __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_add_relaxed ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_add_acquire __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_add_acquire ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_add_release __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_add_release ( long , t#atomic64_t * )
+__ll_sc_atomic64_sub __attribute__ ( ( no_instrument_function ) ) void __ll_sc_atomic64_sub ( long , t#atomic64_t * )
+__ll_sc_atomic64_sub_return __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_sub_return ( long , t#atomic64_t * )
+__ll_sc_atomic64_sub_return_relaxed __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_sub_return_relaxed ( long , t#atomic64_t * )
+__ll_sc_atomic64_sub_return_acquire __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_sub_return_acquire ( long , t#atomic64_t * )
+__ll_sc_atomic64_sub_return_release __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_sub_return_release ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_sub __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_sub ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_sub_relaxed __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_sub_relaxed ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_sub_acquire __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_sub_acquire ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_sub_release __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_sub_release ( long , t#atomic64_t * )
+__ll_sc_atomic64_and __attribute__ ( ( no_instrument_function ) ) void __ll_sc_atomic64_and ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_and __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_and ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_and_relaxed __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_and_relaxed ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_and_acquire __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_and_acquire ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_and_release __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_and_release ( long , t#atomic64_t * )
+__ll_sc_atomic64_andnot __attribute__ ( ( no_instrument_function ) ) void __ll_sc_atomic64_andnot ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_andnot __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_andnot ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_andnot_relaxed __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_andnot_relaxed ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_andnot_acquire __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_andnot_acquire ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_andnot_release __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_andnot_release ( long , t#atomic64_t * )
+__ll_sc_atomic64_or __attribute__ ( ( no_instrument_function ) ) void __ll_sc_atomic64_or ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_or __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_or ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_or_relaxed __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_or_relaxed ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_or_acquire __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_or_acquire ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_or_release __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_or_release ( long , t#atomic64_t * )
+__ll_sc_atomic64_xor __attribute__ ( ( no_instrument_function ) ) void __ll_sc_atomic64_xor ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_xor __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_xor ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_xor_relaxed __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_xor_relaxed ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_xor_acquire __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_xor_acquire ( long , t#atomic64_t * )
+__ll_sc_atomic64_fetch_xor_release __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_fetch_xor_release ( long , t#atomic64_t * )
+__ll_sc_atomic64_dec_if_positive __attribute__ ( ( no_instrument_function ) ) long __ll_sc_atomic64_dec_if_positive ( t#atomic64_t * )
+__ll_sc___cmpxchg_case_1 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_1 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_case_2 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_2 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_case_4 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_4 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_case_8 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_8 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_case_acq_1 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_acq_1 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_case_acq_2 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_acq_2 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_case_acq_4 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_acq_4 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_case_acq_8 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_acq_8 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_case_rel_1 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_rel_1 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_case_rel_2 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_rel_2 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_case_rel_4 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_rel_4 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_case_rel_8 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_rel_8 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_case_mb_1 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_mb_1 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_case_mb_2 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_mb_2 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_case_mb_4 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_mb_4 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_case_mb_8 __attribute__ ( ( no_instrument_function ) ) unsigned long __ll_sc___cmpxchg_case_mb_8 ( volatile void * , unsigned long , unsigned long )
+__ll_sc___cmpxchg_double __attribute__ ( ( no_instrument_function ) ) long __ll_sc___cmpxchg_double ( unsigned long , unsigned long , unsigned long , unsigned long , volatile void * )
+__ll_sc___cmpxchg_double_mb __attribute__ ( ( no_instrument_function ) ) long __ll_sc___cmpxchg_double_mb ( unsigned long , unsigned long , unsigned long , unsigned long , volatile void * )
+
+/* arch/arm64/lib/delay.o */
+__delay extern void __delay ( unsigned long )
+__const_udelay extern void __const_udelay ( unsigned long )
+__udelay extern void __udelay ( unsigned long )
+__ndelay extern void __ndelay ( unsigned long )
+
+/* arch/arm64/mm/copypage.o */
+__cpu_copy_user_page extern void __cpu_copy_user_page ( void * , const void * , unsigned long )
+__cpu_clear_user_page extern void __cpu_clear_user_page ( void * , unsigned long )
+
+/* arch/arm64/mm/dma-mapping.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+dummy_dma_ops extern const s#dma_map_ops dummy_dma_ops
+
+/* arch/arm64/mm/flush.o */
+s##posix_acl
+s##vfsmount
+s##module
+e#writeback_sync_modes enum writeback_sync_modes { WB_SYNC_NONE , WB_SYNC_ALL , }
+t#congested_fn typedef int ( congested_fn ) ( void * , int )
+s##device_private
+s##sock
+s##iov_iter
+s##poll_table_struct
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s#uid_gid_extent struct uid_gid_extent { t#u32 first ; t#u32 lower_first ; t#u32 count ; }
+s#uid_gid_map struct uid_gid_map { t#u32 nr_extents ; s#uid_gid_extent extent [ 5 ] ; }
+s##proc_ns_operations
+t#proc_handler typedef int proc_handler ( s#ctl_table * , int , void * , t#size_t * , t#loff_t * )
+s#ctl_table_poll struct ctl_table_poll { t#atomic_t event ; t#wait_queue_head_t wait ; }
+s#ctl_table struct ctl_table { const char * procname ; void * data ; int maxlen ; t#umode_t mode ; s#ctl_table * child ; t#proc_handler * proc_handler ; s#ctl_table_poll * poll ; void * extra1 ; void * extra2 ; }
+s#ctl_table_root struct ctl_table_root { s#ctl_table_set default_set ; s#ctl_table_set * ( * lookup ) ( s#ctl_table_root * ) ; void ( * set_ownership ) ( s#ctl_table_header * , s#ctl_table * , t#kuid_t * , t#kgid_t * ) ; int ( * permissions ) ( s#ctl_table_header * , s#ctl_table * ) ; }
+s#ctl_node struct ctl_node { s#rb_node node ; s#ctl_table_header * header ; }
+s#ctl_table_header struct ctl_table_header { union { struct { s#ctl_table * ctl_table ; int used ; int count ; int nreg ; } ; s#callback_head rcu ; } ; s#completion * unregistering ; s#ctl_table * ctl_table_arg ; s#ctl_table_root * root ; s#ctl_table_set * set ; s#ctl_dir * parent ; s#ctl_node * node ; s#hlist_head inodes ; }
+s#ctl_dir struct ctl_dir { s#ctl_table_header header ; s#rb_root root ; }
+s#ctl_table_set struct ctl_table_set { int ( * is_seen ) ( s#ctl_table_set * ) ; s#ctl_dir dir ; }
+E#UCOUNT_COUNTS 9
+s#ucounts struct ucounts { s#hlist_node node ; s#user_namespace * ns ; t#kuid_t uid ; int count ; t#atomic_t ucount [ E#UCOUNT_COUNTS ] ; }
+s#user_namespace struct user_namespace { s#uid_gid_map uid_map ; s#uid_gid_map gid_map ; s#uid_gid_map projid_map ; t#atomic_t count ; s#user_namespace * parent ; int level ; t#kuid_t owner ; t#kgid_t group ; s#ns_common ns ; unsigned long flags ; s#key * persistent_keyring_register ; s#rw_semaphore persistent_keyring_register_sem ; s#work_struct work ; s#ctl_table_set set ; s#ctl_table_header * sysctls ; s#ucounts * ucounts ; int ucount_max [ E#UCOUNT_COUNTS ] ; }
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s#cgroup_file struct cgroup_file { s#kernfs_node * kn ; }
+E#CGROUP_SUBSYS_COUNT 13
+s#cgroup_root struct cgroup_root { s#kernfs_root * kf_root ; unsigned int subsys_mask ; int hierarchy_id ; s#cgroup cgrp ; int cgrp_ancestor_id_storage ; t#atomic_t nr_cgrps ; s#list_head root_list ; unsigned int flags ; s#idr cgroup_idr ; char release_agent_path [ 4096 ] ; char name [ 64 ] ; }
+s##bpf_prog_array
+E#__MAX_BPF_ATTACH_TYPE 16
+s#cgroup_bpf struct cgroup_bpf { s#bpf_prog_array * effective [ E#__MAX_BPF_ATTACH_TYPE ] ; s#list_head progs [ E#__MAX_BPF_ATTACH_TYPE ] ; t#u32 flags [ E#__MAX_BPF_ATTACH_TYPE ] ; s#bpf_prog_array * inactive ; }
+s#cgroup struct cgroup { s#cgroup_subsys_state self ; unsigned long flags ; int id ; int level ; int populated_cnt ; s#kernfs_node * kn ; s#cgroup_file procs_file ; s#cgroup_file events_file ; t#u16 subtree_control ; t#u16 subtree_ss_mask ; t#u16 old_subtree_control ; t#u16 old_subtree_ss_mask ; s#cgroup_subsys_state * subsys [ E#CGROUP_SUBSYS_COUNT ] ; s#cgroup_root * root ; s#list_head cset_links ; s#list_head e_csets [ E#CGROUP_SUBSYS_COUNT ] ; s#list_head pidlists ; s#mutex pidlist_mutex ; t#wait_queue_head_t offline_waitq ; s#work_struct release_agent_work ; s#cgroup_bpf bpf ; int ancestor_ids [ ] ; }
+s##cgroup_taskset
+s#mem_cgroup_id struct mem_cgroup_id { int id ; t#atomic_t ref ; }
+s#page_counter struct page_counter { t#atomic_long_t count ; unsigned long limit ; s#page_counter * parent ; unsigned long watermark ; unsigned long failcnt ; }
+s#vmpressure struct vmpressure { unsigned long scanned ; unsigned long reclaimed ; unsigned long tree_scanned ; unsigned long tree_reclaimed ; s#spinlock sr_lock ; s#list_head events ; s#mutex events_lock ; s#work_struct work ; }
+s##eventfd_ctx
+s#mem_cgroup_threshold struct mem_cgroup_threshold { s#eventfd_ctx * eventfd ; unsigned long threshold ; }
+s#mem_cgroup_threshold_ary struct mem_cgroup_threshold_ary { int current_threshold ; unsigned int size ; s#mem_cgroup_threshold entries [ 0 ] ; }
+s#mem_cgroup_thresholds struct mem_cgroup_thresholds { s#mem_cgroup_threshold_ary * primary ; s#mem_cgroup_threshold_ary * spare ; }
+E#NR_LRU_BASE 0
+E#NR_VM_NODE_STAT_ITEMS ( E#NR_LRU_BASE ) + 25
+E#MEMCG_NR_STAT ( E#NR_VM_NODE_STAT_ITEMS ) + 8
+E#NR_VM_EVENT_ITEMS 76
+E#MEMCG_NR_EVENTS ( E#NR_VM_EVENT_ITEMS ) + 4
+E#MEM_CGROUP_NTARGETS 3
+s#mem_cgroup_stat_cpu struct mem_cgroup_stat_cpu { long count [ E#MEMCG_NR_STAT ] ; unsigned long events [ E#MEMCG_NR_EVENTS ] ; unsigned long nr_page_events ; unsigned long targets [ E#MEM_CGROUP_NTARGETS ] ; }
+e#memcg_kmem_state enum memcg_kmem_state { KMEM_NONE , KMEM_ALLOCATED , KMEM_ONLINE , }
+s#percpu_counter struct percpu_counter { t#raw_spinlock_t lock ; t#s64 count ; s#list_head list ; t#s32 * counters ; }
+s#fprop_global struct fprop_global { s#percpu_counter events ; unsigned int period ; t#seqcount_t sequence ; }
+s#wb_domain struct wb_domain { t#spinlock_t lock ; s#fprop_global completions ; s#timer_list period_timer ; unsigned long period_time ; unsigned long dirty_limit_tstamp ; unsigned long dirty_limit ; }
+E#NR_LRU_LISTS ( 0 + 2 + 1 ) + 2
+s#zone_reclaim_stat struct zone_reclaim_stat { unsigned long recent_rotated [ 2 ] ; unsigned long recent_scanned [ 2 ] ; }
+E#NR_WMARK 3
+E#MIGRATE_PCPTYPES 3
+s#per_cpu_pages struct per_cpu_pages { int count ; int high ; int batch ; s#list_head lists [ E#MIGRATE_PCPTYPES ] ; }
+E#NR_VM_NUMA_STAT_ITEMS 6
+E#NR_ZONE_LRU_BASE 1
+E#NR_VM_ZONE_STAT_ITEMS ( E#NR_ZONE_LRU_BASE ) + 14
+s#per_cpu_pageset struct per_cpu_pageset { s#per_cpu_pages pcp ; t#s8 expire ; t#u16 vm_numa_stat_diff [ E#NR_VM_NUMA_STAT_ITEMS ] ; t#s8 stat_threshold ; t#s8 vm_stat_diff [ E#NR_VM_ZONE_STAT_ITEMS ] ; }
+s#zone_padding struct zone_padding { char x [ 0 ] ; }
+E#MIGRATE_TYPES ( E#MIGRATE_PCPTYPES ) + 3
+s#free_area struct free_area { s#list_head free_list [ E#MIGRATE_TYPES ] ; unsigned long nr_free ; }
+s#zone struct zone { unsigned long _watermark [ E#NR_WMARK ] ; unsigned long watermark_boost ; unsigned long nr_reserved_highatomic ; long lowmem_reserve [ 3 ] ; int node ; s#pglist_data * zone_pgdat ; s#per_cpu_pageset * pageset ; unsigned long zone_start_pfn ; unsigned long managed_pages ; unsigned long spanned_pages ; unsigned long present_pages ; const char * name ; unsigned long nr_isolate_pageblock ; int initialized ; s#zone_padding _pad1_ ; s#free_area free_area [ 13 ] ; unsigned long flags ; t#spinlock_t lock ; s#zone_padding _pad2_ ; unsigned long percpu_drift_mark ; unsigned long compact_cached_free_pfn ; unsigned long compact_cached_migrate_pfn [ 2 ] ; unsigned int compact_considered ; unsigned int compact_defer_shift ; int compact_order_failed ; t#bool compact_blockskip_flush ; t#bool contiguous ; s#zone_padding _pad3_ ; t#atomic_long_t vm_stat [ E#NR_VM_ZONE_STAT_ITEMS ] ; t#atomic_long_t vm_numa_stat [ E#NR_VM_NUMA_STAT_ITEMS ] ; void * suse_kabi_padding ; }
+s#zoneref struct zoneref { s#zone * zone ; int zone_idx ; }
+s#zonelist struct zonelist { s#zoneref _zonerefs [ ( ( 1 << 2 ) * 3 ) + 1 ] ; }
+E#MAX_ZONELISTS 2
+e#zone_type enum zone_type { ZONE_DMA , ZONE_NORMAL , ZONE_MOVABLE , __MAX_NR_ZONES }
+s#per_cpu_nodestat struct per_cpu_nodestat { t#s8 stat_threshold ; t#s8 vm_node_stat_diff [ E#NR_VM_NODE_STAT_ITEMS ] ; }
+s#pglist_data struct pglist_data { s#zone node_zones [ 3 ] ; s#zonelist node_zonelists [ E#MAX_ZONELISTS ] ; int nr_zones ; unsigned long node_start_pfn ; unsigned long node_present_pages ; unsigned long node_spanned_pages ; int node_id ; t#wait_queue_head_t kswapd_wait ; t#wait_queue_head_t pfmemalloc_wait ; s#task_struct * kswapd ; int kswapd_order ; e#zone_type kswapd_classzone_idx ; int kswapd_failures ; int kcompactd_max_order ; e#zone_type kcompactd_classzone_idx ; t#wait_queue_head_t kcompactd_wait ; s#task_struct * kcompactd ; unsigned long totalreserve_pages ; unsigned long min_unmapped_pages ; unsigned long min_slab_pages ; s#zone_padding _pad1_ ; t#spinlock_t lru_lock ; t#spinlock_t split_queue_lock ; s#list_head split_queue ; unsigned long split_queue_len ; s#lruvec lruvec ; unsigned int inactive_ratio ; unsigned long flags ; t#atomic_t pagecache_reclaim ; s#zone_padding _pad2_ ; s#per_cpu_nodestat * per_cpu_nodestats ; t#atomic_long_t vm_stat [ E#NR_VM_NODE_STAT_ITEMS ] ; }
+s#lruvec struct lruvec { s#list_head lists [ E#NR_LRU_LISTS ] ; s#zone_reclaim_stat reclaim_stat ; t#atomic_long_t inactive_age ; unsigned long refaults ; s#pglist_data * pgdat ; }
+s#mem_cgroup_reclaim_iter struct mem_cgroup_reclaim_iter { s#mem_cgroup * position ; unsigned int generation ; }
+s#mem_cgroup_per_node struct mem_cgroup_per_node { s#lruvec lruvec ; unsigned long lru_zone_size [ 3 ] [ E#NR_LRU_LISTS ] ; s#mem_cgroup_reclaim_iter iter [ 12 + 1 ] ; s#rb_node tree_node ; unsigned long usage_in_excess ; t#bool on_tree ; s#mem_cgroup * memcg ; }
+s#mem_cgroup struct mem_cgroup { s#cgroup_subsys_state css ; s#mem_cgroup_id id ; s#page_counter memory ; s#page_counter swap ; s#page_counter memsw ; s#page_counter kmem ; s#page_counter tcpmem ; unsigned long low ; unsigned long high ; s#work_struct high_work ; unsigned long soft_limit ; s#vmpressure vmpressure ; t#bool use_hierarchy ; t#bool oom_lock ; int under_oom ; int swappiness ; int oom_kill_disable ; s#cgroup_file events_file ; s#mutex thresholds_lock ; s#mem_cgroup_thresholds thresholds ; s#mem_cgroup_thresholds memsw_thresholds ; s#list_head oom_notify ; unsigned long move_charge_at_immigrate ; t#atomic_t moving_account ; t#spinlock_t move_lock ; s#task_struct * move_lock_task ; unsigned long move_lock_flags ; s#mem_cgroup_stat_cpu * stat ; unsigned long socket_pressure ; t#bool tcpmem_active ; int tcpmem_pressure ; int kmemcg_id ; e#memcg_kmem_state kmem_state ; s#list_head kmem_caches ; int last_scanned_node ; t#nodemask_t scan_nodes ; t#atomic_t numainfo_events ; t#atomic_t numainfo_updating ; s#list_head cgwb_list ; s#wb_domain cgwb_domain ; s#list_head event_list ; t#spinlock_t event_list_lock ; s#mem_cgroup_per_node * nodeinfo [ 0 ] ; }
+s##mempolicy
+s#cftype struct cftype { char name [ 64 ] ; unsigned long private ; t#size_t max_write_len ; unsigned int flags ; unsigned int file_offset ; s#cgroup_subsys * ss ; s#list_head node ; s#kernfs_ops * kf_ops ; int ( * open ) ( s#kernfs_open_file * ) ; void ( * release ) ( s#kernfs_open_file * ) ; t#u64 ( * read_u64 ) ( s#cgroup_subsys_state * , s#cftype * ) ; t#s64 ( * read_s64 ) ( s#cgroup_subsys_state * , s#cftype * ) ; int ( * seq_show ) ( s#seq_file * , void * ) ; void * ( * seq_start ) ( s#seq_file * , t#loff_t * ) ; void * ( * seq_next ) ( s#seq_file * , void * , t#loff_t * ) ; void ( * seq_stop ) ( s#seq_file * , void * ) ; int ( * write_u64 ) ( s#cgroup_subsys_state * , s#cftype * , t#u64 ) ; int ( * write_s64 ) ( s#cgroup_subsys_state * , s#cftype * , t#s64 ) ; t#ssize_t ( * write ) ( s#kernfs_open_file * , char * , t#size_t , t#loff_t ) ; }
+s#cgroup_subsys struct cgroup_subsys { s#cgroup_subsys_state * ( * css_alloc ) ( s#cgroup_subsys_state * ) ; int ( * css_online ) ( s#cgroup_subsys_state * ) ; void ( * css_offline ) ( s#cgroup_subsys_state * ) ; void ( * css_released ) ( s#cgroup_subsys_state * ) ; void ( * css_free ) ( s#cgroup_subsys_state * ) ; void ( * css_reset ) ( s#cgroup_subsys_state * ) ; int ( * can_attach ) ( s#cgroup_taskset * ) ; void ( * cancel_attach ) ( s#cgroup_taskset * ) ; void ( * attach ) ( s#cgroup_taskset * ) ; void ( * post_attach ) ( void ) ; int ( * can_fork ) ( s#task_struct * ) ; void ( * cancel_fork ) ( s#task_struct * ) ; void ( * fork ) ( s#task_struct * ) ; void ( * exit ) ( s#task_struct * ) ; void ( * free ) ( s#task_struct * ) ; void ( * bind ) ( s#cgroup_subsys_state * ) ; t#bool early_init : 1 ; t#bool implicit_on_dfl : 1 ; t#bool broken_hierarchy : 1 ; t#bool warned_broken_hierarchy : 1 ; int id ; const char * name ; const char * legacy_name ; s#cgroup_root * root ; s#idr css_idr ; s#list_head cfts ; s#cftype * dfl_cftypes ; s#cftype * legacy_cftypes ; unsigned int depends_on ; }
+s#cgroup_subsys_state struct cgroup_subsys_state { s#cgroup * cgroup ; s#cgroup_subsys * ss ; s#percpu_ref refcnt ; s#list_head sibling ; s#list_head children ; int id ; unsigned int flags ; t#u64 serial_nr ; t#atomic_t online_cnt ; s#callback_head callback_head ; s#work_struct destroy_work ; s#cgroup_subsys_state * parent ; }
+s#css_set struct css_set { s#cgroup_subsys_state * subsys [ E#CGROUP_SUBSYS_COUNT ] ; t#refcount_t refcount ; s#cgroup * dfl_cgrp ; s#list_head tasks ; s#list_head mg_tasks ; s#list_head task_iters ; s#list_head e_cset_node [ E#CGROUP_SUBSYS_COUNT ] ; s#hlist_node hlist ; s#list_head cgrp_links ; s#list_head mg_preload_node ; s#list_head mg_node ; s#cgroup * mg_src_cgrp ; s#cgroup * mg_dst_cgrp ; s#css_set * mg_dst_cset ; t#bool dead ; s#callback_head callback_head ; }
+s#cgroup_namespace struct cgroup_namespace { t#refcount_t count ; s#ns_common ns ; s#user_namespace * user_ns ; s#ucounts * ucounts ; s#css_set * root_cset ; }
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##gendisk
+t#blk_status_t typedef t#u8 blk_status_t
+s#bvec_iter struct bvec_iter { t#sector_t bi_sector ; unsigned int bi_size ; unsigned int bi_idx ; unsigned int bi_bvec_done ; }
+t#bio_end_io_t typedef void ( bio_end_io_t ) ( s#bio * )
+s##request_queue
+s##kmem_cache
+s#io_cq struct io_cq { s#request_queue * q ; s#io_context * ioc ; union { s#list_head q_node ; s#kmem_cache * __rcu_icq_cache ; } ; union { s#hlist_node ioc_node ; s#callback_head __rcu_head ; } ; unsigned int flags ; }
+s#io_context struct io_context { t#atomic_long_t refcount ; t#atomic_t active_ref ; t#atomic_t nr_tasks ; t#spinlock_t lock ; unsigned short ioprio ; int nr_batch_requests ; unsigned long last_waited ; s#radix_tree_root icq_tree ; s#io_cq * icq_hint ; s#hlist_head icq_list ; s#work_struct release_work ; }
+s#bio_vec struct bio_vec { s#page * bv_page ; unsigned int bv_len ; unsigned int bv_offset ; }
+s#bio_integrity_payload struct bio_integrity_payload { s#bio * bip_bio ; s#bvec_iter bip_iter ; unsigned short bip_slab ; unsigned short bip_vcnt ; unsigned short bip_max_vcnt ; unsigned short bip_flags ; s#bvec_iter bio_iter ; s#work_struct bip_work ; s#bio_vec * bip_vec ; s#bio_vec bip_inline_vecs [ 0 ] ; }
+t#mempool_alloc_t typedef void * ( mempool_alloc_t ) ( t#gfp_t , void * )
+t#mempool_free_t typedef void ( mempool_free_t ) ( void * , void * )
+s#mempool_s struct mempool_s { t#spinlock_t lock ; int min_nr ; int curr_nr ; void * * elements ; void * pool_data ; t#mempool_alloc_t * alloc ; t#mempool_free_t * free ; t#wait_queue_head_t wait ; }
+t#mempool_t typedef s#mempool_s mempool_t
+s##workqueue_struct
+s#bio_set struct bio_set { s#kmem_cache * bio_slab ; unsigned int front_pad ; t#mempool_t bio_pool ; t#mempool_t bvec_pool ; t#mempool_t bio_integrity_pool ; t#mempool_t bvec_integrity_pool ; t#spinlock_t rescue_lock ; s#bio_list rescue_list ; s#work_struct rescue_work ; s#workqueue_struct * rescue_workqueue ; }
+s#bio struct bio { s#bio * bi_next ; s#gendisk * bi_disk ; unsigned int bi_opf ; unsigned short bi_flags ; unsigned short bi_ioprio ; unsigned short bi_write_hint ; t#blk_status_t bi_status ; t#u8 bi_partno ; unsigned int bi_phys_segments ; unsigned int bi_seg_front_size ; unsigned int bi_seg_back_size ; s#bvec_iter bi_iter ; t#atomic_t __bi_remaining ; t#bio_end_io_t * bi_end_io ; void * bi_private ; s#io_context * bi_ioc ; s#cgroup_subsys_state * bi_css ; union { s#bio_integrity_payload * bi_integrity ; } ; unsigned short bi_vcnt ; unsigned short bi_max_vecs ; t#atomic_t __bi_cnt ; s#bio_vec * bi_io_vec ; s#bio_set * bi_pool ; s#bio_vec bi_inline_vecs [ 0 ] ; }
+s#bio_list struct bio_list { s#bio * head ; s#bio * tail ; }
+s##blk_plug
+s#reclaim_state struct reclaim_state { unsigned long reclaimed_slab ; }
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##nlm_lockowner
+s##nfs4_lock_state
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##fwnode_handle
+s##iommu_group
+s##iommu_fwspec
+s#backing_dev_info struct backing_dev_info { s#list_head bdi_list ; unsigned long ra_pages ; unsigned long io_pages ; t#congested_fn * congested_fn ; void * congested_data ; const char * name ; s#kref refcnt ; unsigned int capabilities ; unsigned int min_ratio ; unsigned int max_ratio , max_prop_frac ; t#atomic_long_t tot_write_bandwidth ; s#bdi_writeback wb ; s#list_head wb_list ; s#radix_tree_root cgwb_tree ; s#rb_root cgwb_congested_tree ; s#mutex cgwb_release_mutex ; t#wait_queue_head_t wb_waitq ; s#device * dev ; s#device * owner ; s#timer_list laptop_mode_wb_timer ; s#dentry * debug_dir ; s#dentry * debug_stats ; }
+E#NR_WB_STAT_ITEMS 4
+s#bdi_writeback_congested struct bdi_writeback_congested { unsigned long state ; t#atomic_t refcnt ; s#backing_dev_info * __bdi ; int blkcg_id ; s#rb_node rb_node ; }
+s#fprop_local_percpu struct fprop_local_percpu { s#percpu_counter events ; unsigned int period ; t#raw_spinlock_t lock ; }
+e#wb_reason enum wb_reason { WB_REASON_BACKGROUND , WB_REASON_VMSCAN , WB_REASON_SYNC , WB_REASON_PERIODIC , WB_REASON_LAPTOP_TIMER , WB_REASON_FREE_MORE_MEM , WB_REASON_FS_FREE_SPACE , WB_REASON_FORKER_THREAD , WB_REASON_MAX , }
+s#delayed_work struct delayed_work { s#work_struct work ; s#timer_list timer ; s#workqueue_struct * wq ; int cpu ; }
+s#bdi_writeback struct bdi_writeback { s#backing_dev_info * bdi ; unsigned long state ; unsigned long last_old_flush ; s#list_head b_dirty ; s#list_head b_io ; s#list_head b_more_io ; s#list_head b_dirty_time ; t#spinlock_t list_lock ; s#percpu_counter stat [ E#NR_WB_STAT_ITEMS ] ; s#bdi_writeback_congested * congested ; unsigned long bw_time_stamp ; unsigned long dirtied_stamp ; unsigned long written_stamp ; unsigned long write_bandwidth ; unsigned long avg_write_bandwidth ; unsigned long dirty_ratelimit ; unsigned long balanced_dirty_ratelimit ; s#fprop_local_percpu completions ; int dirty_exceeded ; e#wb_reason start_all_reason ; t#spinlock_t work_lock ; s#list_head work_list ; s#delayed_work dwork ; unsigned long dirty_sleep ; s#list_head bdi_node ; s#percpu_ref refcnt ; s#fprop_local_percpu memcg_completions ; s#cgroup_subsys_state * memcg_css ; s#cgroup_subsys_state * blkcg_css ; s#list_head memcg_node ; s#list_head blkcg_node ; union { s#work_struct release_work ; s#callback_head rcu ; } ; }
+s#writeback_control struct writeback_control { long nr_to_write ; long pages_skipped ; t#loff_t range_start ; t#loff_t range_end ; e#writeback_sync_modes sync_mode ; unsigned for_kupdate : 1 ; unsigned for_background : 1 ; unsigned tagged_writepages : 1 ; unsigned for_reclaim : 1 ; unsigned range_cyclic : 1 ; unsigned for_sync : 1 ; s#bdi_writeback * wb ; s#inode * inode ; int wb_id ; int wb_lcand_id ; int wb_tcand_id ; t#size_t wb_bytes ; t#size_t wb_lcand_bytes ; t#size_t wb_tcand_bytes ; }
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##mtd_info
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s#swap_cluster_info struct swap_cluster_info { t#spinlock_t lock ; unsigned int data : 24 ; unsigned int flags : 8 ; }
+s#swap_cluster_list struct swap_cluster_list { s#swap_cluster_info head ; s#swap_cluster_info tail ; }
+s#percpu_cluster struct percpu_cluster { s#swap_cluster_info index ; unsigned int next ; }
+s#swap_extent struct swap_extent { s#list_head list ; unsigned long start_page ; unsigned long nr_pages ; t#sector_t start_block ; }
+s#swap_info_struct struct swap_info_struct { unsigned long flags ; signed short prio ; s#plist_node list ; s#plist_node avail_list ; signed char type ; unsigned int max ; unsigned char * swap_map ; s#swap_cluster_info * cluster_info ; s#swap_cluster_list free_clusters ; unsigned int lowest_bit ; unsigned int highest_bit ; unsigned int pages ; unsigned int inuse_pages ; unsigned int cluster_next ; unsigned int cluster_nr ; s#percpu_cluster * percpu_cluster ; s#swap_extent * curr_swap_extent ; s#swap_extent first_swap_extent ; s#block_device * bdev ; s#file * swap_file ; unsigned int old_block_size ; unsigned long * frontswap_map ; t#atomic_t frontswap_pages ; t#spinlock_t lock ; t#spinlock_t cont_lock ; s#work_struct discard_work ; s#swap_cluster_list discard_clusters ; }
+flush_dcache_page extern void flush_dcache_page ( s#page * )
+flush_icache_range extern void flush_icache_range ( unsigned long , unsigned long )
+
+/* arch/arm64/mm/init.o */
+pfn_valid extern int pfn_valid ( unsigned long )
+
+/* arch/arm64/mm/ioremap.o */
+__ioremap extern void * __ioremap ( t#phys_addr_t , t#size_t , t#pgprot_t )
+__iounmap extern void __iounmap ( volatile void * )
+ioremap_cache extern void * ioremap_cache ( t#phys_addr_t , t#size_t )
+
+/* arch/arm64/mm/mmu.o */
+kimage_voffset extern t#u64 kimage_voffset
+empty_zero_page extern unsigned long empty_zero_page [ ( ( 1UL ) << 12 ) / sizeof ( unsigned long ) ]
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##kmem_cache
+s##fs_pin
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##tty_struct
+s#taskstats struct taskstats { t#__u16 version ; t#__u32 ac_exitcode ; t#__u8 ac_flag ; t#__u8 ac_nice ; t#__u64 cpu_count __attribute__ ( ( aligned ( 8 ) ) ) ; t#__u64 cpu_delay_total ; t#__u64 blkio_count ; t#__u64 blkio_delay_total ; t#__u64 swapin_count ; t#__u64 swapin_delay_total ; t#__u64 cpu_run_real_total ; t#__u64 cpu_run_virtual_total ; char ac_comm [ 32 ] ; t#__u8 ac_sched __attribute__ ( ( aligned ( 8 ) ) ) ; t#__u8 ac_pad [ 3 ] ; t#__u32 ac_uid __attribute__ ( ( aligned ( 8 ) ) ) ; t#__u32 ac_gid ; t#__u32 ac_pid ; t#__u32 ac_ppid ; t#__u32 ac_btime ; t#__u64 ac_etime __attribute__ ( ( aligned ( 8 ) ) ) ; t#__u64 ac_utime ; t#__u64 ac_stime ; t#__u64 ac_minflt ; t#__u64 ac_majflt ; t#__u64 coremem ; t#__u64 virtmem ; t#__u64 hiwater_rss ; t#__u64 hiwater_vm ; t#__u64 read_char ; t#__u64 write_char ; t#__u64 read_syscalls ; t#__u64 write_syscalls ; t#__u64 read_bytes ; t#__u64 write_bytes ; t#__u64 cancelled_write_bytes ; t#__u64 nvcsw ; t#__u64 nivcsw ; t#__u64 ac_utimescaled ; t#__u64 ac_stimescaled ; t#__u64 cpu_scaled_run_real_total ; t#__u64 freepages_count ; t#__u64 freepages_delay_total ; }
+s##tty_audit_buf
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##gendisk
+s##request_queue
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##workqueue_struct
+s##pipe_inode_info
+s##hd_struct
+s##eventfd_ctx
+s##blk_plug
+s##robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+phys_mem_access_prot extern t#pgprot_t phys_mem_access_prot ( s#file * , unsigned long , unsigned long , t#pgprot_t )
+
+/* arch/arm64/mm/numa.o */
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##posix_acl
+s##vfsmount
+s##sock
+s##iov_iter
+s##pid_namespace
+s##nlm_lockowner
+s##nfs4_lock_state
+s##pipe_inode_info
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##mem_cgroup
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##backing_dev_info
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##swap_info_struct
+s##device_private
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+node_data extern s#pglist_data * node_data [ ]
+node_to_cpumask_map extern t#cpumask_var_t node_to_cpumask_map [ ( 1 << 2 ) ]
+__per_cpu_offset extern unsigned long __per_cpu_offset [ 480 ]
+__node_distance int __node_distance ( int , int )
+
+/* arch/arm64/mm/pageattr.o */
+set_memory_nx int set_memory_nx ( unsigned long , int )
+set_memory_x int set_memory_x ( unsigned long , int )
+
+/* block/badblocks.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##exception_table_entry
+s##module_sect_attrs
+s##module_notes_attrs
+s##tracepoint
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##vm_operations_struct
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##mem_cgroup
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##dev_pagemap
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##fwnode_handle
+s##iommu_group
+s##iommu_fwspec
+s#badblocks struct badblocks { s#device * dev ; int count ; int unacked_exist ; int shift ; t#u64 * page ; int changed ; t#seqlock_t lock ; t#sector_t sector ; t#sector_t size ; }
+badblocks_check int badblocks_check ( s#badblocks * , t#sector_t , int , t#sector_t * , int * )
+badblocks_set int badblocks_set ( s#badblocks * , t#sector_t , int , int )
+badblocks_clear int badblocks_clear ( s#badblocks * , t#sector_t , int )
+ack_all_badblocks void ack_all_badblocks ( s#badblocks * )
+badblocks_show t#ssize_t badblocks_show ( s#badblocks * , char * , int )
+badblocks_store t#ssize_t badblocks_store ( s#badblocks * , const char * , t#size_t , int )
+badblocks_init int badblocks_init ( s#badblocks * , int )
+devm_init_badblocks int devm_init_badblocks ( s#device * , s#badblocks * )
+badblocks_exit void badblocks_exit ( s#badblocks * )
+
+/* block/bio.o */
+s##kmem_cache
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##pipe_inode_info
+s#blk_mq_ctx struct blk_mq_ctx { struct { t#spinlock_t lock ; s#list_head rq_list ; } __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; unsigned int cpu ; unsigned int index_hw ; unsigned long rq_dispatched [ 2 ] ; unsigned long rq_merged ; unsigned long __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) rq_completed [ 2 ] ; s#request_queue * queue ; s#kobject kobj ; }
+t#req_flags_t typedef t#__u32 req_flags_t
+t#rq_end_io_fn typedef void ( rq_end_io_fn ) ( s#request * , t#blk_status_t )
+e#mq_rq_state enum mq_rq_state { MQ_RQ_IDLE = 0 , MQ_RQ_IN_FLIGHT = 1 , MQ_RQ_COMPLETE = 2 , }
+t#smp_call_func_t typedef void ( * smp_call_func_t ) ( void * )
+s#__call_single_data struct __call_single_data { s#llist_node llist ; t#smp_call_func_t func ; void * info ; unsigned int flags ; }
+s##blkcg_gq
+s#request_list struct request_list { s#request_queue * q ; s#blkcg_gq * blkg ; int count [ 2 ] ; int starved [ 2 ] ; t#mempool_t * rq_pool ; t#wait_queue_head_t wait [ 2 ] ; unsigned int flags ; }
+s#request struct request { s#request_queue * q ; s#blk_mq_ctx * mq_ctx ; int cpu ; unsigned int cmd_flags ; t#req_flags_t rq_flags ; int internal_tag ; unsigned int __data_len ; int tag ; t#sector_t __sector ; s#bio * bio ; s#bio * biotail ; s#list_head queuelist ; union { s#hlist_node hash ; s#list_head ipi_list ; } ; union { s#rb_node rb_node ; s#bio_vec special_vec ; void * completion_data ; int error_count ; } ; union { struct { s#io_cq * icq ; void * priv [ 2 ] ; } elv ; struct { unsigned int seq ; s#list_head list ; t#rq_end_io_fn * saved_end_io ; } flush ; } ; s#gendisk * rq_disk ; s#hd_struct * part ; t#u64 start_time_ns ; t#u64 io_start_time_ns ; unsigned short wbt_flags ; unsigned short nr_phys_segments ; unsigned short nr_integrity_segments ; unsigned short write_hint ; unsigned short ioprio ; void * special ; unsigned int extra_len ; e#mq_rq_state state ; t#refcount_t ref ; unsigned int timeout ; unsigned long __deadline ; s#list_head timeout_list ; union { s#__call_single_data csd ; t#u64 fifo_time ; } ; t#rq_end_io_fn * end_io ; void * end_io_data ; s#request * next_rq ; s#request_list * rl ; }
+e#elv_merge enum elv_merge { ELEVATOR_NO_MERGE = 0 , ELEVATOR_FRONT_MERGE = 1 , ELEVATOR_BACK_MERGE = 2 , ELEVATOR_DISCARD_MERGE = 3 , }
+t#elevator_merge_fn typedef e#elv_merge ( elevator_merge_fn ) ( s#request_queue * , s#request * * , s#bio * )
+t#elevator_merged_fn typedef void ( elevator_merged_fn ) ( s#request_queue * , s#request * , e#elv_merge )
+t#elevator_merge_req_fn typedef void ( elevator_merge_req_fn ) ( s#request_queue * , s#request * , s#request * )
+t#elevator_allow_bio_merge_fn typedef int ( elevator_allow_bio_merge_fn ) ( s#request_queue * , s#request * , s#bio * )
+t#elevator_allow_rq_merge_fn typedef int ( elevator_allow_rq_merge_fn ) ( s#request_queue * , s#request * , s#request * )
+t#elevator_bio_merged_fn typedef void ( elevator_bio_merged_fn ) ( s#request_queue * , s#request * , s#bio * )
+t#elevator_dispatch_fn typedef int ( elevator_dispatch_fn ) ( s#request_queue * , int )
+t#elevator_add_req_fn typedef void ( elevator_add_req_fn ) ( s#request_queue * , s#request * )
+t#elevator_activate_req_fn typedef void ( elevator_activate_req_fn ) ( s#request_queue * , s#request * )
+t#elevator_deactivate_req_fn typedef void ( elevator_deactivate_req_fn ) ( s#request_queue * , s#request * )
+t#elevator_completed_req_fn typedef void ( elevator_completed_req_fn ) ( s#request_queue * , s#request * )
+t#elevator_request_list_fn typedef s#request * ( elevator_request_list_fn ) ( s#request_queue * , s#request * )
+t#elevator_init_icq_fn typedef void ( elevator_init_icq_fn ) ( s#io_cq * )
+t#elevator_exit_icq_fn typedef void ( elevator_exit_icq_fn ) ( s#io_cq * )
+t#elevator_set_req_fn typedef int ( elevator_set_req_fn ) ( s#request_queue * , s#request * , s#bio * , t#gfp_t )
+t#elevator_put_req_fn typedef void ( elevator_put_req_fn ) ( s#request * )
+t#elevator_may_queue_fn typedef int ( elevator_may_queue_fn ) ( s#request_queue * , unsigned int )
+t#elevator_init_fn typedef int ( elevator_init_fn ) ( s#request_queue * , s#elevator_type * )
+t#elevator_exit_fn typedef void ( elevator_exit_fn ) ( s#elevator_queue * )
+t#elevator_registered_fn typedef void ( elevator_registered_fn ) ( s#request_queue * )
+s#elevator_ops struct elevator_ops { t#elevator_merge_fn * elevator_merge_fn ; t#elevator_merged_fn * elevator_merged_fn ; t#elevator_merge_req_fn * elevator_merge_req_fn ; t#elevator_allow_bio_merge_fn * elevator_allow_bio_merge_fn ; t#elevator_allow_rq_merge_fn * elevator_allow_rq_merge_fn ; t#elevator_bio_merged_fn * elevator_bio_merged_fn ; t#elevator_dispatch_fn * elevator_dispatch_fn ; t#elevator_add_req_fn * elevator_add_req_fn ; t#elevator_activate_req_fn * elevator_activate_req_fn ; t#elevator_deactivate_req_fn * elevator_deactivate_req_fn ; t#elevator_completed_req_fn * elevator_completed_req_fn ; t#elevator_request_list_fn * elevator_former_req_fn ; t#elevator_request_list_fn * elevator_latter_req_fn ; t#elevator_init_icq_fn * elevator_init_icq_fn ; t#elevator_exit_icq_fn * elevator_exit_icq_fn ; t#elevator_set_req_fn * elevator_set_req_fn ; t#elevator_put_req_fn * elevator_put_req_fn ; t#elevator_may_queue_fn * elevator_may_queue_fn ; t#elevator_init_fn * elevator_init_fn ; t#elevator_exit_fn * elevator_exit_fn ; t#elevator_registered_fn * elevator_registered_fn ; }
+s#blk_flush_queue struct blk_flush_queue { unsigned int flush_queue_delayed : 1 ; unsigned int flush_pending_idx : 1 ; unsigned int flush_running_idx : 1 ; unsigned long flush_pending_since ; s#list_head flush_queue [ 2 ] ; s#list_head flush_data_in_flight ; s#request * flush_rq ; s#request * orig_rq ; t#spinlock_t mq_flush_lock ; }
+s#sbitmap_word struct sbitmap_word { unsigned long word ; unsigned long depth ; }
+s#sbitmap struct sbitmap { unsigned int depth ; unsigned int shift ; unsigned int map_nr ; s#sbitmap_word * map ; }
+t#wait_queue_func_t typedef int ( * wait_queue_func_t ) ( s#wait_queue_entry * , unsigned , int , void * )
+s#wait_queue_entry struct wait_queue_entry { unsigned int flags ; void * private ; t#wait_queue_func_t func ; s#list_head entry ; }
+t#wait_queue_entry_t typedef s#wait_queue_entry wait_queue_entry_t
+s#sbq_wait_state struct sbq_wait_state { t#atomic_t wait_cnt ; t#wait_queue_head_t wait ; }
+s#sbitmap_queue struct sbitmap_queue { s#sbitmap sb ; unsigned int * alloc_hint ; unsigned int wake_batch ; t#atomic_t wake_index ; s#sbq_wait_state * ws ; t#bool round_robin ; unsigned int min_shallow_depth ; }
+s#blk_mq_tags struct blk_mq_tags { unsigned int nr_tags ; unsigned int nr_reserved_tags ; t#atomic_t active_queues ; s#sbitmap_queue bitmap_tags ; s#sbitmap_queue breserved_tags ; s#request * * rqs ; s#request * * static_rqs ; s#list_head page_list ; }
+s#srcu_node struct srcu_node { t#spinlock_t lock ; unsigned long srcu_have_cbs [ 4 ] ; unsigned long srcu_data_have_cbs [ 4 ] ; unsigned long srcu_gp_seq_needed_exp ; s#srcu_node * srcu_parent ; int grplo ; int grphi ; }
+s#rcu_segcblist struct rcu_segcblist { s#callback_head * head ; s#callback_head * * tails [ 4 ] ; unsigned long gp_seq [ 4 ] ; long len ; long len_lazy ; }
+s#srcu_data struct srcu_data { unsigned long srcu_lock_count [ 2 ] ; unsigned long srcu_unlock_count [ 2 ] ; t#spinlock_t lock __attribute__ ( ( __aligned__ ( 1 << ( ( 6 ) ) ) ) ) ; s#rcu_segcblist srcu_cblist ; unsigned long srcu_gp_seq_needed ; unsigned long srcu_gp_seq_needed_exp ; t#bool srcu_cblist_invoking ; s#delayed_work work ; s#callback_head srcu_barrier_head ; s#srcu_node * mynode ; unsigned long grpmask ; int cpu ; s#srcu_struct * sp ; }
+s#srcu_struct struct srcu_struct { s#srcu_node node [ ( 1 + ( ( ( 480 ) + ( ( 16 ) ) - 1 ) / ( ( 16 ) ) ) ) ] ; s#srcu_node * level [ 2 + 1 ] ; s#mutex srcu_cb_mutex ; t#spinlock_t gp_lock ; s#mutex srcu_gp_mutex ; unsigned int srcu_idx ; unsigned long srcu_gp_seq ; unsigned long srcu_gp_seq_needed ; unsigned long srcu_gp_seq_needed_exp ; unsigned long srcu_last_gp_end ; s#srcu_data * sda ; unsigned long srcu_barrier_seq ; s#mutex srcu_barrier_mutex ; s#completion srcu_barrier_completion ; t#atomic_t srcu_barrier_cpu_cnt ; s#delayed_work work ; }
+s#blk_mq_hw_ctx struct blk_mq_hw_ctx { struct { t#spinlock_t lock ; s#list_head dispatch ; unsigned long state ; } __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; s#delayed_work run_work ; t#cpumask_var_t cpumask ; int next_cpu ; int next_cpu_batch ; unsigned long flags ; void * sched_data ; s#request_queue * queue ; s#blk_flush_queue * fq ; void * driver_data ; s#sbitmap ctx_map ; s#blk_mq_ctx * dispatch_from ; unsigned int dispatch_busy ; unsigned int nr_ctx ; s#blk_mq_ctx * * ctxs ; t#spinlock_t dispatch_wait_lock ; t#wait_queue_entry_t dispatch_wait ; t#atomic_t wait_index ; s#blk_mq_tags * tags ; s#blk_mq_tags * sched_tags ; unsigned long queued ; unsigned long run ; unsigned long dispatched [ 7 ] ; unsigned int numa_node ; unsigned int queue_num ; t#atomic_t nr_active ; unsigned int nr_expired ; s#hlist_node cpuhp_dead ; s#kobject kobj ; unsigned long poll_considered ; unsigned long poll_invoked ; unsigned long poll_success ; s#dentry * debugfs_dir ; s#dentry * sched_debugfs_dir ; s#srcu_struct srcu [ 0 ] ; }
+t#blk_mq_req_flags_t typedef t#__u32 blk_mq_req_flags_t
+s#blk_mq_alloc_data struct blk_mq_alloc_data { s#request_queue * q ; t#blk_mq_req_flags_t flags ; unsigned int shallow_depth ; s#blk_mq_ctx * ctx ; s#blk_mq_hw_ctx * hctx ; }
+s#elevator_mq_ops struct elevator_mq_ops { int ( * init_sched ) ( s#request_queue * , s#elevator_type * ) ; void ( * exit_sched ) ( s#elevator_queue * ) ; int ( * init_hctx ) ( s#blk_mq_hw_ctx * , unsigned int ) ; void ( * exit_hctx ) ( s#blk_mq_hw_ctx * , unsigned int ) ; t#bool ( * allow_merge ) ( s#request_queue * , s#request * , s#bio * ) ; t#bool ( * bio_merge ) ( s#blk_mq_hw_ctx * , s#bio * ) ; int ( * request_merge ) ( s#request_queue * , s#request * * , s#bio * ) ; void ( * request_merged ) ( s#request_queue * , s#request * , e#elv_merge ) ; void ( * requests_merged ) ( s#request_queue * , s#request * , s#request * ) ; void ( * limit_depth ) ( unsigned int , s#blk_mq_alloc_data * ) ; void ( * prepare_request ) ( s#request * , s#bio * ) ; void ( * finish_request ) ( s#request * ) ; void ( * insert_requests ) ( s#blk_mq_hw_ctx * , s#list_head * , t#bool ) ; s#request * ( * dispatch_request ) ( s#blk_mq_hw_ctx * ) ; t#bool ( * has_work ) ( s#blk_mq_hw_ctx * ) ; void ( * completed_request ) ( s#request * , t#u64 ) ; void ( * started_request ) ( s#request * ) ; void ( * requeue_request ) ( s#request * ) ; s#request * ( * former_request ) ( s#request_queue * , s#request * ) ; s#request * ( * next_request ) ( s#request_queue * , s#request * ) ; void ( * init_icq ) ( s#io_cq * ) ; void ( * exit_icq ) ( s#io_cq * ) ; }
+s#elv_fs_entry struct elv_fs_entry { s#attribute attr ; t#ssize_t ( * show ) ( s#elevator_queue * , char * ) ; t#ssize_t ( * store ) ( s#elevator_queue * , const char * , t#size_t ) ; }
+s##blk_mq_debugfs_attr
+s#elevator_type struct elevator_type { s#kmem_cache * icq_cache ; union { s#elevator_ops sq ; s#elevator_mq_ops mq ; } ops ; t#size_t icq_size ; t#size_t icq_align ; s#elv_fs_entry * elevator_attrs ; char elevator_name [ ( 16 ) ] ; const char * elevator_alias ; s#module * elevator_owner ; t#bool uses_mq ; const s#blk_mq_debugfs_attr * queue_debugfs_attrs ; const s#blk_mq_debugfs_attr * hctx_debugfs_attrs ; char icq_cache_name [ ( 16 ) + 6 ] ; s#list_head list ; }
+s#elevator_queue struct elevator_queue { s#elevator_type * type ; void * elevator_data ; s#kobject kobj ; s#mutex sysfs_lock ; unsigned int registered : 1 ; unsigned int uses_mq : 1 ; s#hlist_head hash [ 1 << ( 6 ) ] ; }
+s##blk_queue_stats
+s##rq_wb
+t#request_fn_proc typedef void ( request_fn_proc ) ( s#request_queue * )
+t#blk_qc_t typedef unsigned int blk_qc_t
+t#make_request_fn typedef t#blk_qc_t ( make_request_fn ) ( s#request_queue * , s#bio * )
+t#poll_q_fn typedef t#bool ( poll_q_fn ) ( s#request_queue * , t#blk_qc_t )
+t#prep_rq_fn typedef int ( prep_rq_fn ) ( s#request_queue * , s#request * )
+t#unprep_rq_fn typedef void ( unprep_rq_fn ) ( s#request_queue * , s#request * )
+t#softirq_done_fn typedef void ( softirq_done_fn ) ( s#request * )
+e#blk_eh_timer_return enum blk_eh_timer_return { BLK_EH_DONE , BLK_EH_RESET_TIMER , }
+t#rq_timed_out_fn typedef e#blk_eh_timer_return ( rq_timed_out_fn ) ( s#request * )
+t#dma_drain_needed_fn typedef int ( dma_drain_needed_fn ) ( s#request * )
+t#lld_busy_fn typedef int ( lld_busy_fn ) ( s#request_queue * )
+t#init_rq_fn typedef int ( init_rq_fn ) ( s#request_queue * , s#request * , t#gfp_t )
+t#exit_rq_fn typedef void ( exit_rq_fn ) ( s#request_queue * , s#request * )
+s#blk_mq_queue_data struct blk_mq_queue_data { s#request * rq ; t#bool last ; }
+t#queue_rq_fn typedef t#blk_status_t ( queue_rq_fn ) ( s#blk_mq_hw_ctx * , const s#blk_mq_queue_data * )
+t#get_budget_fn typedef t#bool ( get_budget_fn ) ( s#blk_mq_hw_ctx * )
+t#put_budget_fn typedef void ( put_budget_fn ) ( s#blk_mq_hw_ctx * )
+t#timeout_fn typedef e#blk_eh_timer_return ( timeout_fn ) ( s#request * , t#bool )
+t#poll_fn typedef int ( poll_fn ) ( s#blk_mq_hw_ctx * , unsigned int )
+t#init_hctx_fn typedef int ( init_hctx_fn ) ( s#blk_mq_hw_ctx * , void * , unsigned int )
+t#exit_hctx_fn typedef void ( exit_hctx_fn ) ( s#blk_mq_hw_ctx * , unsigned int )
+s#blk_mq_tag_set struct blk_mq_tag_set { unsigned int * mq_map ; const s#blk_mq_ops * ops ; unsigned int nr_hw_queues ; unsigned int queue_depth ; unsigned int reserved_tags ; unsigned int cmd_size ; int numa_node ; unsigned int timeout ; unsigned int flags ; void * driver_data ; s#blk_mq_tags * * tags ; s#mutex tag_list_lock ; s#list_head tag_list ; }
+t#init_request_fn typedef int ( init_request_fn ) ( s#blk_mq_tag_set * , s#request * , unsigned int , unsigned int )
+t#exit_request_fn typedef void ( exit_request_fn ) ( s#blk_mq_tag_set * , s#request * , unsigned int )
+t#map_queues_fn typedef int ( map_queues_fn ) ( s#blk_mq_tag_set * )
+s#blk_mq_ops struct blk_mq_ops { t#queue_rq_fn * queue_rq ; t#get_budget_fn * get_budget ; t#put_budget_fn * put_budget ; t#timeout_fn * timeout ; t#poll_fn * poll ; t#softirq_done_fn * complete ; t#init_hctx_fn * init_hctx ; t#exit_hctx_fn * exit_hctx ; t#init_request_fn * init_request ; t#exit_request_fn * exit_request ; void ( * initialize_rq_fn ) ( s#request * ) ; t#map_queues_fn * map_queues ; void ( * show_rq ) ( s#seq_file * , s#request * ) ; }
+s#blk_integrity_iter struct blk_integrity_iter { void * prot_buf ; void * data_buf ; t#sector_t seed ; unsigned int data_size ; unsigned short interval ; const char * disk_name ; }
+t#integrity_processing_fn typedef t#blk_status_t ( integrity_processing_fn ) ( s#blk_integrity_iter * )
+s#blk_integrity_profile struct blk_integrity_profile { t#integrity_processing_fn * generate_fn ; t#integrity_processing_fn * verify_fn ; const char * name ; }
+s#blk_integrity struct blk_integrity { const s#blk_integrity_profile * profile ; unsigned char flags ; unsigned char tuple_size ; unsigned char interval_exp ; unsigned char tag_size ; }
+s#blk_queue_tag struct blk_queue_tag { s#request * * tag_index ; unsigned long * tag_map ; int max_depth ; int real_max_depth ; t#atomic_t refcnt ; int alloc_policy ; int next_tag ; }
+s#blk_rq_stat struct blk_rq_stat { t#u64 mean ; t#u64 min ; t#u64 max ; t#u32 nr_samples ; t#u64 batch ; }
+s#blk_stat_callback struct blk_stat_callback { s#list_head list ; s#timer_list timer ; s#blk_rq_stat * cpu_stat ; int ( * bucket_fn ) ( const s#request * ) ; unsigned int buckets ; s#blk_rq_stat * stat ; void ( * timer_fn ) ( s#blk_stat_callback * ) ; void * data ; s#callback_head rcu ; }
+e#blk_zoned_model enum blk_zoned_model { BLK_ZONED_NONE , BLK_ZONED_HA , BLK_ZONED_HM , }
+s#queue_limits struct queue_limits { unsigned long bounce_pfn ; unsigned long seg_boundary_mask ; unsigned long virt_boundary_mask ; unsigned int max_hw_sectors ; unsigned int max_dev_sectors ; unsigned int chunk_sectors ; unsigned int max_sectors ; unsigned int max_segment_size ; unsigned int physical_block_size ; unsigned int alignment_offset ; unsigned int io_min ; unsigned int io_opt ; unsigned int max_discard_sectors ; unsigned int max_hw_discard_sectors ; unsigned int max_write_same_sectors ; unsigned int max_write_zeroes_sectors ; unsigned int discard_granularity ; unsigned int discard_alignment ; unsigned short logical_block_size ; unsigned short max_segments ; unsigned short max_integrity_segments ; unsigned short max_discard_segments ; unsigned char misaligned ; unsigned char discard_misaligned ; unsigned char cluster ; unsigned char raid_partial_stripes_expensive ; e#blk_zoned_model zoned ; }
+s#irq_work struct irq_work { unsigned long flags ; s#llist_node llnode ; void ( * func ) ( s#irq_work * ) ; }
+s#rchan_buf struct rchan_buf { void * start ; void * data ; t#size_t offset ; t#size_t subbufs_produced ; t#size_t subbufs_consumed ; s#rchan * chan ; t#wait_queue_head_t read_wait ; s#irq_work wakeup_work ; s#dentry * dentry ; s#kref kref ; s#page * * page_array ; unsigned int page_count ; unsigned int finalized ; t#size_t * padding ; t#size_t prev_padding ; t#size_t bytes_consumed ; t#size_t early_bytes ; unsigned int cpu ; }
+s#rchan_callbacks struct rchan_callbacks { int ( * subbuf_start ) ( s#rchan_buf * , void * , void * , t#size_t ) ; void ( * buf_mapped ) ( s#rchan_buf * , s#file * ) ; void ( * buf_unmapped ) ( s#rchan_buf * , s#file * ) ; s#dentry * ( * create_buf_file ) ( const char * , s#dentry * , t#umode_t , s#rchan_buf * , int * ) ; int ( * remove_buf_file ) ( s#dentry * ) ; }
+s#rchan struct rchan { t#u32 version ; t#size_t subbuf_size ; t#size_t n_subbufs ; t#size_t alloc_size ; s#rchan_callbacks * cb ; s#kref kref ; void * private_data ; t#size_t last_toobig ; s#rchan_buf * * buf ; int is_global ; s#list_head list ; s#dentry * parent ; int has_base_filename ; char base_filename [ 255 ] ; }
+s#blk_trace struct blk_trace { int trace_state ; s#rchan * rchan ; unsigned long * sequence ; unsigned char * msg_data ; t#u16 act_mask ; t#u64 start_lba ; t#u64 end_lba ; t#u32 pid ; t#u32 dev ; s#dentry * dir ; s#dentry * dropped_file ; s#dentry * msg_file ; s#list_head running_list ; t#atomic_t dropped ; }
+s##bsg_job
+t#bsg_job_fn typedef int ( bsg_job_fn ) ( s#bsg_job * )
+s#sg_io_v4 struct sg_io_v4 { t#__s32 guard ; t#__u32 protocol ; t#__u32 subprotocol ; t#__u32 request_len ; t#__u64 request ; t#__u64 request_tag ; t#__u32 request_attr ; t#__u32 request_priority ; t#__u32 request_extra ; t#__u32 max_response_len ; t#__u64 response ; t#__u32 dout_iovec_count ; t#__u32 dout_xfer_len ; t#__u32 din_iovec_count ; t#__u32 din_xfer_len ; t#__u64 dout_xferp ; t#__u64 din_xferp ; t#__u32 timeout ; t#__u32 flags ; t#__u64 usr_ptr ; t#__u32 spare_in ; t#__u32 driver_status ; t#__u32 transport_status ; t#__u32 device_status ; t#__u32 retry_delay ; t#__u32 info ; t#__u32 duration ; t#__u32 response_len ; t#__s32 din_resid ; t#__s32 dout_resid ; t#__u64 generated_tag ; t#__u32 spare_out ; t#__u32 padding ; }
+s#bsg_ops struct bsg_ops { int ( * check_proto ) ( s#sg_io_v4 * ) ; int ( * fill_hdr ) ( s#request * , s#sg_io_v4 * , t#fmode_t ) ; int ( * complete_rq ) ( s#request * , s#sg_io_v4 * ) ; void ( * free_rq ) ( s#request * ) ; }
+s#bsg_class_device struct bsg_class_device { s#device * class_dev ; int minor ; s#request_queue * queue ; const s#bsg_ops * ops ; }
+s##throtl_data
+s#request_queue struct request_queue { s#list_head queue_head ; s#request * last_merge ; s#elevator_queue * elevator ; int nr_rqs [ 2 ] ; int nr_rqs_elvpriv ; s#blk_queue_stats * stats ; s#rq_wb * rq_wb ; s#request_list root_rl ; t#request_fn_proc * request_fn ; t#make_request_fn * make_request_fn ; t#poll_q_fn * poll_fn ; t#prep_rq_fn * prep_rq_fn ; t#unprep_rq_fn * unprep_rq_fn ; t#softirq_done_fn * softirq_done_fn ; t#rq_timed_out_fn * rq_timed_out_fn ; t#dma_drain_needed_fn * dma_drain_needed ; t#lld_busy_fn * lld_busy_fn ; t#init_rq_fn * init_rq_fn ; t#exit_rq_fn * exit_rq_fn ; void ( * initialize_rq_fn ) ( s#request * ) ; const s#blk_mq_ops * mq_ops ; unsigned int * mq_map ; s#blk_mq_ctx * queue_ctx ; unsigned int nr_queues ; unsigned int queue_depth ; s#blk_mq_hw_ctx * * queue_hw_ctx ; unsigned int nr_hw_queues ; t#sector_t end_sector ; s#request * boundary_rq ; s#delayed_work delay_work ; s#backing_dev_info * backing_dev_info ; void * queuedata ; unsigned long queue_flags ; int id ; t#gfp_t bounce_gfp ; t#spinlock_t __queue_lock ; t#spinlock_t * queue_lock ; s#kobject kobj ; s#kobject mq_kobj ; s#blk_integrity integrity ; s#device * dev ; int rpm_status ; unsigned int nr_pending ; unsigned long nr_requests ; unsigned int nr_congestion_on ; unsigned int nr_congestion_off ; unsigned int nr_batching ; unsigned int dma_drain_size ; void * dma_drain_buffer ; unsigned int dma_pad_mask ; unsigned int dma_alignment ; s#blk_queue_tag * queue_tags ; unsigned int nr_sorted ; unsigned int in_flight [ 2 ] ; unsigned int request_fn_active ; unsigned int rq_timeout ; int poll_nsec ; s#blk_stat_callback * poll_cb ; s#blk_rq_stat poll_stat [ 16 ] ; s#timer_list timeout ; s#work_struct timeout_work ; s#list_head timeout_list ; s#list_head icq_list ; unsigned long blkcg_pols [ ( ( ( 5 ) + ( 8 * sizeof ( long ) ) - 1 ) / ( 8 * sizeof ( long ) ) ) ] ; s#blkcg_gq * root_blkg ; s#list_head blkg_list ; s#queue_limits limits ; unsigned int nr_zones ; unsigned long * seq_zones_bitmap ; unsigned long * seq_zones_wlock ; unsigned int sg_timeout ; unsigned int sg_reserved_size ; int node ; s#blk_trace * blk_trace ; s#mutex blk_trace_mutex ; s#blk_flush_queue * fq ; s#list_head requeue_list ; t#spinlock_t requeue_lock ; s#delayed_work requeue_work ; s#mutex sysfs_lock ; int bypass_depth ; t#atomic_t mq_freeze_depth ; t#bsg_job_fn * bsg_job_fn ; s#bsg_class_device bsg_dev ; s#throtl_data * td ; s#callback_head callback_head ; t#wait_queue_head_t mq_freeze_wq ; s#percpu_ref q_usage_counter ; s#list_head all_q_node ; s#blk_mq_tag_set * tag_set ; s#list_head tag_set_list ; s#bio_set bio_split ; s#dentry * debugfs_dir ; s#dentry * sched_debugfs_dir ; t#bool mq_sysfs_init_done ; t#size_t cmd_size ; void * rq_alloc_data ; s#work_struct release_work ; t#u64 write_hints [ 5 ] ; }
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s#blk_plug struct blk_plug { s#list_head list ; s#list_head mq_list ; s#list_head cb_list ; }
+s##robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s#partition_meta_info struct partition_meta_info { char uuid [ ( 36 + 1 ) ] ; t#u8 volname [ 64 ] ; }
+s#disk_stats struct disk_stats { unsigned long sectors [ 2 ] ; unsigned long ios [ 2 ] ; unsigned long merges [ 2 ] ; unsigned long ticks [ 2 ] ; unsigned long io_ticks ; unsigned long time_in_queue ; }
+s#hd_struct struct hd_struct { t#sector_t start_sect ; t#sector_t nr_sects ; t#seqcount_t nr_sects_seq ; t#sector_t alignment_offset ; unsigned int discard_alignment ; s#device __dev ; s#kobject * holder_dir ; int policy , partno ; s#partition_meta_info * info ; int make_it_fail ; unsigned long stamp ; t#atomic_t in_flight [ 2 ] ; s#disk_stats * dkstats ; s#percpu_ref ref ; s#callback_head callback_head ; }
+s#disk_part_tbl struct disk_part_tbl { s#callback_head callback_head ; int len ; s#hd_struct * last_lookup ; s#hd_struct * part [ ] ; }
+s##hd_geometry
+s##pr_ops
+s#block_device_operations struct block_device_operations { int ( * open ) ( s#block_device * , t#fmode_t ) ; void ( * release ) ( s#gendisk * , t#fmode_t ) ; int ( * rw_page ) ( s#block_device * , t#sector_t , s#page * , t#bool ) ; int ( * ioctl ) ( s#block_device * , t#fmode_t , unsigned , unsigned long ) ; int ( * compat_ioctl ) ( s#block_device * , t#fmode_t , unsigned , unsigned long ) ; unsigned int ( * check_events ) ( s#gendisk * , unsigned int ) ; int ( * media_changed ) ( s#gendisk * ) ; void ( * unlock_native_capacity ) ( s#gendisk * ) ; int ( * revalidate_disk ) ( s#gendisk * ) ; int ( * getgeo ) ( s#block_device * , s#hd_geometry * ) ; void ( * swap_slot_free_notify ) ( s#block_device * , unsigned long ) ; s#module * owner ; const s#pr_ops * pr_ops ; }
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s#gendisk struct gendisk { int major ; int first_minor ; int minors ; char disk_name [ 32 ] ; char * ( * devnode ) ( s#gendisk * , t#umode_t * ) ; unsigned int events ; unsigned int async_events ; s#disk_part_tbl * part_tbl ; s#hd_struct part0 ; const s#block_device_operations * fops ; s#request_queue * queue ; void * private_data ; int flags ; s#rw_semaphore lookup_sem ; s#kobject * slave_dir ; s#timer_rand_state * random ; t#atomic_t sync_io ; s#disk_events * ev ; s#kobject integrity_kobj ; int node_id ; s#badblocks * bb ; }
+fs_bio_set extern s#bio_set fs_bio_set
+bio_uninit extern void bio_uninit ( s#bio * )
+bio_init extern void bio_init ( s#bio * , s#bio_vec * , unsigned short )
+bio_reset extern void bio_reset ( s#bio * )
+bio_chain void bio_chain ( s#bio * , s#bio * )
+bio_alloc_bioset extern s#bio * bio_alloc_bioset ( t#gfp_t , unsigned int , s#bio_set * )
+zero_fill_bio_iter void zero_fill_bio_iter ( s#bio * , s#bvec_iter )
+bio_put extern void bio_put ( s#bio * )
+bio_phys_segments extern int bio_phys_segments ( s#request_queue * , s#bio * )
+__bio_clone_fast extern void __bio_clone_fast ( s#bio * , s#bio * )
+bio_clone_fast extern s#bio * bio_clone_fast ( s#bio * , t#gfp_t , s#bio_set * )
+bio_clone_bioset extern s#bio * bio_clone_bioset ( s#bio * , t#gfp_t , s#bio_set * )
+bio_add_pc_page extern int bio_add_pc_page ( s#request_queue * , s#bio * , s#page * , unsigned int , unsigned int )
+bio_add_page extern int bio_add_page ( s#bio * , s#page * , unsigned int , unsigned int )
+bio_iov_iter_get_pages int bio_iov_iter_get_pages ( s#bio * , s#iov_iter * )
+submit_bio_wait extern int submit_bio_wait ( s#bio * )
+bio_advance extern void bio_advance ( s#bio * , unsigned )
+bio_copy_data_iter extern void bio_copy_data_iter ( s#bio * , s#bvec_iter * , s#bio * , s#bvec_iter * )
+bio_copy_data extern void bio_copy_data ( s#bio * , s#bio * )
+bio_list_copy_data extern void bio_list_copy_data ( s#bio * , s#bio * )
+bio_free_pages extern void bio_free_pages ( s#bio * )
+bio_map_user_iov extern s#bio * bio_map_user_iov ( s#request_queue * , s#iov_iter * , t#gfp_t )
+bio_map_kern extern s#bio * bio_map_kern ( s#request_queue * , void * , unsigned int , t#gfp_t )
+bio_set_pages_dirty extern void bio_set_pages_dirty ( s#bio * )
+bio_check_pages_dirty extern void bio_check_pages_dirty ( s#bio * )
+generic_start_io_acct void generic_start_io_acct ( s#request_queue * , int , unsigned long , s#hd_struct * )
+generic_end_io_acct void generic_end_io_acct ( s#request_queue * , int , s#hd_struct * , unsigned long )
+bio_flush_dcache_pages extern void bio_flush_dcache_pages ( s#bio * )
+bio_endio extern void bio_endio ( s#bio * )
+bio_split extern s#bio * bio_split ( s#bio * , int , t#gfp_t , s#bio_set * )
+bio_trim extern void bio_trim ( s#bio * , int , int )
+bioset_exit extern void bioset_exit ( s#bio_set * )
+bioset_free extern void bioset_free ( s#bio_set * )
+bioset_init extern int bioset_init ( s#bio_set * , unsigned int , unsigned int , int )
+bioset_create extern s#bio_set * bioset_create ( unsigned int , unsigned int , int )
+bio_associate_blkcg int bio_associate_blkcg ( s#bio * , s#cgroup_subsys_state * )
+bio_clone_blkcg_association void bio_clone_blkcg_association ( s#bio * , s#bio * )
+
+/* block/bio-integrity.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##iov_iter
+s##kmem_cache
+s##blkcg_gq
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+bio_integrity_alloc extern s#bio_integrity_payload * bio_integrity_alloc ( s#bio * , t#gfp_t , unsigned int )
+bio_integrity_add_page extern int bio_integrity_add_page ( s#bio * , s#page * , unsigned int , unsigned int )
+bio_integrity_prep extern t#bool bio_integrity_prep ( s#bio * )
+bio_integrity_advance extern void bio_integrity_advance ( s#bio * , unsigned int )
+bio_integrity_trim extern void bio_integrity_trim ( s#bio * )
+bio_integrity_clone extern int bio_integrity_clone ( s#bio * , s#bio * , t#gfp_t )
+bioset_integrity_create extern int bioset_integrity_create ( s#bio_set * , int )
+bioset_integrity_free extern void bioset_integrity_free ( s#bio_set * )
+
+/* block/blk-cgroup.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##device_private
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##tty_struct
+s##tty_audit_buf
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##iov_iter
+s##kmem_cache
+s##eventfd_ctx
+E#BLKG_RWSTAT_NR 4
+s#blkg_rwstat struct blkg_rwstat { s#percpu_counter cpu_cnt [ E#BLKG_RWSTAT_NR ] ; t#atomic64_t aux_cnt [ E#BLKG_RWSTAT_NR ] ; }
+s#blkg_policy_data struct blkg_policy_data { s#blkcg_gq * blkg ; int plid ; }
+s#blkcg_gq struct blkcg_gq { s#request_queue * q ; s#list_head q_node ; s#hlist_node blkcg_node ; s#blkcg * blkcg ; s#bdi_writeback_congested * wb_congested ; s#blkcg_gq * parent ; s#request_list rl ; t#atomic_t refcnt ; t#bool online ; s#blkg_rwstat stat_bytes ; s#blkg_rwstat stat_ios ; s#blkg_policy_data * pd [ 5 ] ; s#callback_head callback_head ; }
+s##workqueue_struct
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##bpf_prog_array
+s##cgroup_taskset
+s#blkcg_policy_data struct blkcg_policy_data { s#blkcg * blkcg ; int plid ; }
+s#blkcg struct blkcg { s#cgroup_subsys_state css ; t#spinlock_t lock ; s#radix_tree_root blkg_tree ; s#blkcg_gq * blkg_hint ; s#hlist_head blkg_list ; s#blkcg_policy_data * cpd [ 5 ] ; s#list_head all_blkcgs_node ; s#list_head cgwb_list ; }
+blkcg_root extern s#blkcg blkcg_root
+blkg_lookup_slowpath s#blkcg_gq * blkg_lookup_slowpath ( s#blkcg * , s#request_queue * , t#bool )
+__blkg_release_rcu void __blkg_release_rcu ( s#callback_head * )
+blkg_dev_name const char * blkg_dev_name ( s#blkcg_gq * )
+t#blkcg_pol_alloc_cpd_fn typedef s#blkcg_policy_data * ( blkcg_pol_alloc_cpd_fn ) ( t#gfp_t )
+t#blkcg_pol_init_cpd_fn typedef void ( blkcg_pol_init_cpd_fn ) ( s#blkcg_policy_data * )
+t#blkcg_pol_free_cpd_fn typedef void ( blkcg_pol_free_cpd_fn ) ( s#blkcg_policy_data * )
+t#blkcg_pol_bind_cpd_fn typedef void ( blkcg_pol_bind_cpd_fn ) ( s#blkcg_policy_data * )
+t#blkcg_pol_alloc_pd_fn typedef s#blkg_policy_data * ( blkcg_pol_alloc_pd_fn ) ( t#gfp_t , int )
+t#blkcg_pol_init_pd_fn typedef void ( blkcg_pol_init_pd_fn ) ( s#blkg_policy_data * )
+t#blkcg_pol_online_pd_fn typedef void ( blkcg_pol_online_pd_fn ) ( s#blkg_policy_data * )
+t#blkcg_pol_offline_pd_fn typedef void ( blkcg_pol_offline_pd_fn ) ( s#blkg_policy_data * )
+t#blkcg_pol_free_pd_fn typedef void ( blkcg_pol_free_pd_fn ) ( s#blkg_policy_data * )
+t#blkcg_pol_reset_pd_stats_fn typedef void ( blkcg_pol_reset_pd_stats_fn ) ( s#blkg_policy_data * )
+s#blkcg_policy struct blkcg_policy { int plid ; s#cftype * dfl_cftypes ; s#cftype * legacy_cftypes ; t#blkcg_pol_alloc_cpd_fn * cpd_alloc_fn ; t#blkcg_pol_init_cpd_fn * cpd_init_fn ; t#blkcg_pol_free_cpd_fn * cpd_free_fn ; t#blkcg_pol_bind_cpd_fn * cpd_bind_fn ; t#blkcg_pol_alloc_pd_fn * pd_alloc_fn ; t#blkcg_pol_init_pd_fn * pd_init_fn ; t#blkcg_pol_online_pd_fn * pd_online_fn ; t#blkcg_pol_offline_pd_fn * pd_offline_fn ; t#blkcg_pol_free_pd_fn * pd_free_fn ; t#blkcg_pol_reset_pd_stats_fn * pd_reset_stats_fn ; }
+blkcg_print_blkgs void blkcg_print_blkgs ( s#seq_file * , s#blkcg * , t#u64 ( * ) ( s#seq_file * , s#blkg_policy_data * , int ) , const s#blkcg_policy * , int , t#bool )
+__blkg_prfill_u64 t#u64 __blkg_prfill_u64 ( s#seq_file * , s#blkg_policy_data * , t#u64 )
+__blkg_prfill_rwstat t#u64 __blkg_prfill_rwstat ( s#seq_file * , s#blkg_policy_data * , const s#blkg_rwstat * )
+blkg_prfill_stat t#u64 blkg_prfill_stat ( s#seq_file * , s#blkg_policy_data * , int )
+blkg_prfill_rwstat t#u64 blkg_prfill_rwstat ( s#seq_file * , s#blkg_policy_data * , int )
+blkg_print_stat_bytes int blkg_print_stat_bytes ( s#seq_file * , void * )
+blkg_print_stat_ios int blkg_print_stat_ios ( s#seq_file * , void * )
+blkg_print_stat_bytes_recursive int blkg_print_stat_bytes_recursive ( s#seq_file * , void * )
+blkg_print_stat_ios_recursive int blkg_print_stat_ios_recursive ( s#seq_file * , void * )
+blkg_stat_recursive_sum t#u64 blkg_stat_recursive_sum ( s#blkcg_gq * , s#blkcg_policy * , int )
+blkg_rwstat_recursive_sum s#blkg_rwstat blkg_rwstat_recursive_sum ( s#blkcg_gq * , s#blkcg_policy * , int )
+s#blkg_conf_ctx struct blkg_conf_ctx { s#gendisk * disk ; s#blkcg_gq * blkg ; char * body ; }
+blkg_conf_prep int blkg_conf_prep ( s#blkcg * , const s#blkcg_policy * , char * , s#blkg_conf_ctx * )
+blkg_conf_finish void blkg_conf_finish ( s#blkg_conf_ctx * )
+io_cgrp_subsys extern s#cgroup_subsys io_cgrp_subsys
+blkcg_activate_policy int blkcg_activate_policy ( s#request_queue * , const s#blkcg_policy * )
+blkcg_deactivate_policy void blkcg_deactivate_policy ( s#request_queue * , const s#blkcg_policy * )
+blkcg_policy_register int blkcg_policy_register ( s#blkcg_policy * )
+blkcg_policy_unregister void blkcg_policy_unregister ( s#blkcg_policy * )
+
+/* block/blk-core.o */
+__tracepoint_block_bio_remap extern s#tracepoint __tracepoint_block_bio_remap
+__tracepoint_block_rq_remap extern s#tracepoint __tracepoint_block_rq_remap
+__tracepoint_block_bio_complete extern s#tracepoint __tracepoint_block_bio_complete
+__tracepoint_block_split extern s#tracepoint __tracepoint_block_split
+__tracepoint_block_unplug extern s#tracepoint __tracepoint_block_unplug
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+e#trace_reg enum trace_reg { TRACE_REG_REGISTER , TRACE_REG_UNREGISTER , TRACE_REG_PERF_REGISTER , TRACE_REG_PERF_UNREGISTER , TRACE_REG_PERF_OPEN , TRACE_REG_PERF_CLOSE , TRACE_REG_PERF_ADD , TRACE_REG_PERF_DEL , }
+s#trace_event_class struct trace_event_class { const char * system ; void * probe ; void * perf_probe ; int ( * reg ) ( s#trace_event_call * , e#trace_reg , void * ) ; int ( * define_fields ) ( s#trace_event_call * ) ; s#list_head * ( * get_fields ) ( s#trace_event_call * ) ; s#list_head fields ; int ( * raw_init ) ( s#trace_event_call * ) ; }
+e#print_line_t enum print_line_t { TRACE_TYPE_PARTIAL_LINE = 0 , TRACE_TYPE_HANDLED = 1 , TRACE_TYPE_UNHANDLED = 2 , TRACE_TYPE_NO_CONSUME = 3 }
+s##trace_array
+s##tracer
+s##trace_buffer
+s##ring_buffer_iter
+s#seq_buf struct seq_buf { char * buffer ; t#size_t size ; t#size_t len ; t#loff_t readpos ; }
+s#trace_seq struct trace_seq { unsigned char buffer [ ( ( 1UL ) << 12 ) ] ; s#seq_buf seq ; int full ; }
+s#trace_entry struct trace_entry { unsigned short type ; unsigned char flags ; unsigned char preempt_count ; int pid ; }
+s#trace_iterator struct trace_iterator { s#trace_array * tr ; s#tracer * trace ; s#trace_buffer * trace_buffer ; void * private ; int cpu_file ; s#mutex mutex ; s#ring_buffer_iter * * buffer_iter ; unsigned long iter_flags ; s#trace_seq tmp_seq ; t#cpumask_var_t started ; t#bool snapshot ; s#trace_seq seq ; s#trace_entry * ent ; unsigned long lost_events ; int leftover ; int ent_size ; int cpu ; t#u64 ts ; t#loff_t pos ; long idx ; }
+t#trace_print_func typedef e#print_line_t ( * trace_print_func ) ( s#trace_iterator * , int , s#trace_event * )
+s#trace_event_functions struct trace_event_functions { t#trace_print_func trace ; t#trace_print_func raw ; t#trace_print_func hex ; t#trace_print_func binary ; }
+s#trace_event struct trace_event { s#hlist_node node ; s#list_head list ; int type ; s#trace_event_functions * funcs ; }
+s##event_filter
+s##bpf_prog_array
+s##device_private
+s##sched_class
+s##cfs_rq
+s##task_group
+s##kmem_cache
+s##fs_pin
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##cgroup_taskset
+s##workqueue_struct
+s##pipe_inode_info
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##eventfd_ctx
+s##mempolicy
+s##tty_struct
+s##tty_audit_buf
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##futex_pi_state
+s#perf_event_groups struct perf_event_groups { s#rb_root tree ; t#u64 index ; }
+s#perf_event_context struct perf_event_context { s#pmu * pmu ; t#raw_spinlock_t lock ; s#mutex mutex ; s#list_head active_ctx_list ; s#perf_event_groups pinned_groups ; s#perf_event_groups flexible_groups ; s#list_head event_list ; s#list_head pinned_active ; s#list_head flexible_active ; int nr_events ; int nr_active ; int is_active ; int nr_stat ; int nr_freq ; int rotate_disable ; t#atomic_t refcount ; s#task_struct * task ; t#u64 time ; t#u64 timestamp ; s#perf_event_context * parent_ctx ; t#u64 parent_gen ; t#u64 generation ; int pin_count ; int nr_cgroups ; void * task_ctx_data ; s#callback_head callback_head ; }
+s##numa_group
+s##task_delay_info
+s##uprobe
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s#perf_cgroup_info struct perf_cgroup_info { t#u64 time ; t#u64 timestamp ; }
+s#perf_cgroup struct perf_cgroup { s#cgroup_subsys_state css ; s#perf_cgroup_info * info ; }
+s#perf_cpu_context struct perf_cpu_context { s#perf_event_context ctx ; s#perf_event_context * task_ctx ; int active_oncpu ; int exclusive ; t#raw_spinlock_t hrtimer_lock ; s#hrtimer hrtimer ; t#ktime_t hrtimer_interval ; unsigned int hrtimer_active ; s#perf_cgroup * cgrp ; s#list_head cgrp_cpuctx_entry ; s#list_head sched_cb_entry ; int sched_cb_usage ; }
+s#pmu struct pmu { s#list_head entry ; s#module * module ; s#device * dev ; const s#attribute_group * * attr_groups ; const char * name ; int type ; int capabilities ; int * pmu_disable_count ; s#perf_cpu_context * pmu_cpu_context ; t#atomic_t exclusive_cnt ; int task_ctx_nr ; int hrtimer_interval_ms ; unsigned int nr_addr_filters ; void ( * pmu_enable ) ( s#pmu * ) ; void ( * pmu_disable ) ( s#pmu * ) ; int ( * event_init ) ( s#perf_event * ) ; void ( * event_mapped ) ( s#perf_event * , s#mm_struct * ) ; void ( * event_unmapped ) ( s#perf_event * , s#mm_struct * ) ; int ( * add ) ( s#perf_event * , int ) ; void ( * del ) ( s#perf_event * , int ) ; void ( * start ) ( s#perf_event * , int ) ; void ( * stop ) ( s#perf_event * , int ) ; void ( * read ) ( s#perf_event * ) ; void ( * start_txn ) ( s#pmu * , unsigned int ) ; int ( * commit_txn ) ( s#pmu * ) ; void ( * cancel_txn ) ( s#pmu * ) ; int ( * event_idx ) ( s#perf_event * ) ; void ( * sched_task ) ( s#perf_event_context * , t#bool ) ; t#size_t task_ctx_size ; void * ( * setup_aux ) ( int , void * * , int , t#bool ) ; void ( * free_aux ) ( void * ) ; int ( * addr_filters_validate ) ( s#list_head * ) ; void ( * addr_filters_sync ) ( s#perf_event * ) ; int ( * filter_match ) ( s#perf_event * ) ; }
+e#perf_event_state enum perf_event_state { PERF_EVENT_STATE_DEAD = - 4 , PERF_EVENT_STATE_EXIT = - 3 , PERF_EVENT_STATE_ERROR = - 2 , PERF_EVENT_STATE_OFF = - 1 , PERF_EVENT_STATE_INACTIVE = 0 , PERF_EVENT_STATE_ACTIVE = 1 , }
+t#local_t typedef struct { t#atomic_long_t a ; } local_t
+t#local64_t typedef struct { t#local_t a ; } local64_t
+s#perf_event_attr struct perf_event_attr { t#__u32 type ; t#__u32 size ; t#__u64 config ; union { t#__u64 sample_period ; t#__u64 sample_freq ; } ; t#__u64 sample_type ; t#__u64 read_format ; t#__u64 disabled : 1 , inherit : 1 , pinned : 1 , exclusive : 1 , exclude_user : 1 , exclude_kernel : 1 , exclude_hv : 1 , exclude_idle : 1 , mmap : 1 , comm : 1 , freq : 1 , inherit_stat : 1 , enable_on_exec : 1 , task : 1 , watermark : 1 , precise_ip : 2 , mmap_data : 1 , sample_id_all : 1 , exclude_host : 1 , exclude_guest : 1 , exclude_callchain_kernel : 1 , exclude_callchain_user : 1 , mmap2 : 1 , comm_exec : 1 , use_clockid : 1 , context_switch : 1 , write_backward : 1 , namespaces : 1 , __reserved_1 : 35 ; union { t#__u32 wakeup_events ; t#__u32 wakeup_watermark ; } ; t#__u32 bp_type ; union { t#__u64 bp_addr ; t#__u64 kprobe_func ; t#__u64 uprobe_path ; t#__u64 config1 ; } ; union { t#__u64 bp_len ; t#__u64 kprobe_addr ; t#__u64 probe_offset ; t#__u64 config2 ; } ; t#__u64 branch_sample_type ; t#__u64 sample_regs_user ; t#__u32 sample_stack_user ; t#__s32 clockid ; t#__u64 sample_regs_intr ; t#__u32 aux_watermark ; t#__u16 sample_max_stack ; t#__u16 __reserved_2 ; }
+s#hw_perf_event_extra struct hw_perf_event_extra { t#u64 config ; unsigned int reg ; int alloc ; int idx ; }
+s#arch_hw_breakpoint_ctrl struct arch_hw_breakpoint_ctrl { t#u32 __reserved : 19 , len : 8 , type : 2 , privilege : 2 , enabled : 1 ; }
+s#arch_hw_breakpoint struct arch_hw_breakpoint { t#u64 address ; t#u64 trigger ; s#arch_hw_breakpoint_ctrl ctrl ; }
+s#hw_perf_event struct hw_perf_event { union { struct { t#u64 config ; t#u64 last_tag ; unsigned long config_base ; unsigned long event_base ; int event_base_rdpmc ; int idx ; int last_cpu ; int flags ; s#hw_perf_event_extra extra_reg ; s#hw_perf_event_extra branch_reg ; } ; struct { s#hrtimer hrtimer ; } ; struct { s#list_head tp_list ; } ; struct { int itrace_started ; } ; struct { t#u64 pwr_acc ; t#u64 ptsc ; } ; struct { s#arch_hw_breakpoint info ; s#list_head bp_list ; } ; struct { t#u8 iommu_bank ; t#u8 iommu_cntr ; t#u16 padding ; t#u64 conf ; t#u64 conf1 ; } ; } ; s#task_struct * target ; void * addr_filters ; unsigned long addr_filters_gen ; int state ; t#local64_t prev_count ; t#u64 sample_period ; t#u64 last_period ; t#local64_t period_left ; t#u64 interrupts_seq ; t#u64 interrupts ; t#u64 freq_time_stamp ; t#u64 freq_count_stamp ; }
+s##ring_buffer
+s#perf_addr_filters_head struct perf_addr_filters_head { s#list_head list ; t#raw_spinlock_t lock ; unsigned int nr_file_filters ; }
+t#perf_copy_f typedef unsigned long ( * perf_copy_f ) ( void * , const void * , unsigned long , unsigned long )
+s#perf_raw_frag struct perf_raw_frag { union { s#perf_raw_frag * next ; unsigned long pad ; } ; t#perf_copy_f copy ; void * data ; t#u32 size ; }
+s#perf_raw_record struct perf_raw_record { s#perf_raw_frag frag ; t#u32 size ; }
+s#perf_branch_entry struct perf_branch_entry { t#__u64 from ; t#__u64 to ; t#__u64 mispred : 1 , predicted : 1 , in_tx : 1 , abort : 1 , cycles : 16 , reserved : 44 ; }
+s#perf_branch_stack struct perf_branch_stack { t#__u64 nr ; s#perf_branch_entry entries [ 0 ] ; }
+u#perf_mem_data_src union perf_mem_data_src { t#__u64 val ; struct { t#__u64 mem_op : 5 , mem_lvl : 14 , mem_snoop : 5 , mem_lock : 2 , mem_dtlb : 7 , mem_lvl_num : 4 , mem_remote : 1 , mem_snoopx : 2 , mem_rsvd : 24 ; } ; }
+s#perf_callchain_entry struct perf_callchain_entry { t#__u64 nr ; t#__u64 ip [ 0 ] ; }
+s#perf_regs struct perf_regs { t#__u64 abi ; s#pt_regs * regs ; }
+s#perf_sample_data struct perf_sample_data { t#u64 addr ; s#perf_raw_record * raw ; s#perf_branch_stack * br_stack ; t#u64 period ; t#u64 weight ; t#u64 txn ; u#perf_mem_data_src data_src ; t#u64 type ; t#u64 ip ; struct { t#u32 pid ; t#u32 tid ; } tid_entry ; t#u64 time ; t#u64 id ; t#u64 stream_id ; struct { t#u32 cpu ; t#u32 reserved ; } cpu_entry ; s#perf_callchain_entry * callchain ; s#perf_regs regs_user ; s#pt_regs regs_user_copy ; s#perf_regs regs_intr ; t#u64 stack_user_size ; t#u64 phys_addr ; }
+t#perf_overflow_handler_t typedef void ( * perf_overflow_handler_t ) ( s#perf_event * , s#perf_sample_data * , s#pt_regs * )
+s##bpf_prog
+t#ftrace_func_t typedef void ( * ftrace_func_t ) ( unsigned long , unsigned long , s#ftrace_ops * , s#pt_regs * )
+s##ftrace_hash
+s#ftrace_ops_hash struct ftrace_ops_hash { s#ftrace_hash * notrace_hash ; s#ftrace_hash * filter_hash ; s#mutex regex_lock ; }
+s#ftrace_ops struct ftrace_ops { t#ftrace_func_t func ; s#ftrace_ops * next ; unsigned long flags ; void * private ; t#ftrace_func_t saved_func ; int * disabled ; s#ftrace_ops_hash local_hash ; s#ftrace_ops_hash * func_hash ; s#ftrace_ops_hash old_hash ; unsigned long trampoline ; unsigned long trampoline_size ; }
+s#perf_event struct perf_event { s#list_head event_entry ; s#list_head sibling_list ; s#list_head active_list ; s#rb_node group_node ; t#u64 group_index ; s#list_head migrate_entry ; s#hlist_node hlist_entry ; s#list_head active_entry ; int nr_siblings ; int event_caps ; int group_caps ; s#perf_event * group_leader ; s#pmu * pmu ; void * pmu_private ; e#perf_event_state state ; unsigned int attach_state ; t#local64_t count ; t#atomic64_t child_count ; t#u64 total_time_enabled ; t#u64 total_time_running ; t#u64 tstamp ; t#u64 shadow_ctx_time ; s#perf_event_attr attr ; t#u16 header_size ; t#u16 id_header_size ; t#u16 read_size ; s#hw_perf_event hw ; s#perf_event_context * ctx ; t#atomic_long_t refcount ; t#atomic64_t child_total_time_enabled ; t#atomic64_t child_total_time_running ; s#mutex child_mutex ; s#list_head child_list ; s#perf_event * parent ; int oncpu ; int cpu ; s#list_head owner_entry ; s#task_struct * owner ; s#mutex mmap_mutex ; t#atomic_t mmap_count ; s#ring_buffer * rb ; s#list_head rb_entry ; unsigned long rcu_batches ; int rcu_pending ; t#wait_queue_head_t waitq ; s#fasync_struct * fasync ; int pending_wakeup ; int pending_kill ; int pending_disable ; s#irq_work pending ; t#atomic_t event_limit ; s#perf_addr_filters_head addr_filters ; unsigned long * addr_filters_offs ; unsigned long addr_filters_gen ; void ( * destroy ) ( s#perf_event * ) ; s#callback_head callback_head ; s#pid_namespace * ns ; t#u64 id ; t#u64 ( * clock ) ( void ) ; t#perf_overflow_handler_t overflow_handler ; void * overflow_handler_context ; t#perf_overflow_handler_t orig_overflow_handler ; s#bpf_prog * prog ; s#trace_event_call * tp_event ; s#event_filter * filter ; s#ftrace_ops ftrace_ops ; s#perf_cgroup * cgrp ; s#list_head sb_list ; }
+s#trace_event_call struct trace_event_call { s#list_head list ; s#trace_event_class * class ; union { char * name ; s#tracepoint * tp ; } ; s#trace_event event ; char * print_fmt ; s#event_filter * filter ; void * mod ; void * data ; int flags ; int perf_refcount ; s#hlist_head * perf_events ; s#bpf_prog_array * prog_array ; int ( * perf_perm ) ( s#trace_event_call * , s#perf_event * ) ; }
+s#trace_enum_map struct trace_enum_map { const char * system ; const char * enum_string ; unsigned long enum_value ; }
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s#rq_wait struct rq_wait { t#wait_queue_head_t wait ; t#atomic_t inflight ; }
+E#WBT_NUM_RWQ ( 0 ) + 3
+s#rq_wb struct rq_wb { unsigned int wb_background ; unsigned int wb_normal ; unsigned int wb_max ; int scale_step ; t#bool scaled_max ; short enable_state ; unsigned int unknown_cnt ; t#u64 win_nsec ; t#u64 cur_win_nsec ; s#blk_stat_callback * cb ; t#u64 sync_issue ; void * sync_cookie ; unsigned int wc ; unsigned int queue_depth ; unsigned long last_issue ; unsigned long last_comp ; unsigned long min_lat_nsec ; s#request_queue * queue ; s#rq_wait rq_wait [ E#WBT_NUM_RWQ ] ; }
+s##bsg_job
+s##throtl_data
+blk_queue_flag_set void blk_queue_flag_set ( unsigned int , s#request_queue * )
+blk_queue_flag_clear void blk_queue_flag_clear ( unsigned int , s#request_queue * )
+blk_queue_flag_test_and_set t#bool blk_queue_flag_test_and_set ( unsigned int , s#request_queue * )
+blk_queue_flag_test_and_clear t#bool blk_queue_flag_test_and_clear ( unsigned int , s#request_queue * )
+blk_rq_init extern void blk_rq_init ( s#request_queue * , s#request * )
+errno_to_blk_status t#blk_status_t errno_to_blk_status ( int )
+blk_status_to_errno int blk_status_to_errno ( t#blk_status_t )
+blk_dump_rq_flags extern void blk_dump_rq_flags ( s#request * , char * )
+blk_delay_queue extern void blk_delay_queue ( s#request_queue * , unsigned long )
+blk_start_queue_async extern void blk_start_queue_async ( s#request_queue * )
+blk_start_queue extern void blk_start_queue ( s#request_queue * )
+blk_stop_queue extern void blk_stop_queue ( s#request_queue * )
+blk_sync_queue extern void blk_sync_queue ( s#request_queue * )
+blk_set_preempt_only extern int blk_set_preempt_only ( s#request_queue * )
+blk_clear_preempt_only extern void blk_clear_preempt_only ( s#request_queue * )
+__blk_run_queue_uncond extern void __blk_run_queue_uncond ( s#request_queue * )
+__blk_run_queue extern void __blk_run_queue ( s#request_queue * )
+blk_run_queue_async extern void blk_run_queue_async ( s#request_queue * )
+blk_run_queue extern void blk_run_queue ( s#request_queue * )
+blk_put_queue extern void blk_put_queue ( s#request_queue * )
+blk_queue_bypass_start void blk_queue_bypass_start ( s#request_queue * )
+blk_queue_bypass_end void blk_queue_bypass_end ( s#request_queue * )
+blk_set_queue_dying extern void blk_set_queue_dying ( s#request_queue * )
+blk_cleanup_queue extern void blk_cleanup_queue ( s#request_queue * )
+blk_alloc_queue s#request_queue * blk_alloc_queue ( t#gfp_t )
+blk_alloc_queue_node s#request_queue * blk_alloc_queue_node ( t#gfp_t , int , t#spinlock_t * )
+blk_init_queue extern s#request_queue * blk_init_queue ( t#request_fn_proc * , t#spinlock_t * )
+blk_init_queue_node extern s#request_queue * blk_init_queue_node ( t#request_fn_proc * , t#spinlock_t * , int )
+blk_init_allocated_queue extern int blk_init_allocated_queue ( s#request_queue * )
+blk_get_queue t#bool __attribute__ ( ( warn_unused_result ) ) blk_get_queue ( s#request_queue * )
+blk_get_request extern s#request * blk_get_request ( s#request_queue * , unsigned int , t#blk_mq_req_flags_t )
+blk_requeue_request extern void blk_requeue_request ( s#request_queue * , s#request * )
+part_round_stats extern void part_round_stats ( s#request_queue * , int , s#hd_struct * )
+__blk_put_request extern void __blk_put_request ( s#request_queue * , s#request * )
+blk_put_request extern void blk_put_request ( s#request * )
+blk_init_request_from_bio extern void blk_init_request_from_bio ( s#request * , s#bio * )
+generic_make_request extern t#blk_qc_t generic_make_request ( s#bio * )
+direct_make_request extern t#blk_qc_t direct_make_request ( s#bio * )
+submit_bio extern t#blk_qc_t submit_bio ( s#bio * )
+blk_poll t#bool blk_poll ( s#request_queue * , t#blk_qc_t )
+blk_insert_cloned_request extern t#blk_status_t blk_insert_cloned_request ( s#request_queue * , s#request * )
+blk_rq_err_bytes extern unsigned int blk_rq_err_bytes ( const s#request * )
+blk_peek_request extern s#request * blk_peek_request ( s#request_queue * )
+blk_start_request extern void blk_start_request ( s#request * )
+blk_fetch_request extern s#request * blk_fetch_request ( s#request_queue * )
+blk_steal_bios void blk_steal_bios ( s#bio_list * , s#request * )
+blk_update_request extern t#bool blk_update_request ( s#request * , t#blk_status_t , unsigned int )
+blk_unprep_request extern void blk_unprep_request ( s#request * )
+blk_finish_request extern void blk_finish_request ( s#request * , t#blk_status_t )
+blk_end_request extern t#bool blk_end_request ( s#request * , t#blk_status_t , unsigned int )
+blk_end_request_all extern void blk_end_request_all ( s#request * , t#blk_status_t )
+__blk_end_request extern t#bool __blk_end_request ( s#request * , t#blk_status_t , unsigned int )
+__blk_end_request_all extern void __blk_end_request_all ( s#request * , t#blk_status_t )
+__blk_end_request_cur extern t#bool __blk_end_request_cur ( s#request * , t#blk_status_t )
+rq_flush_dcache_pages extern void rq_flush_dcache_pages ( s#request * )
+blk_lld_busy extern int blk_lld_busy ( s#request_queue * )
+blk_rq_unprep_clone extern void blk_rq_unprep_clone ( s#request * )
+blk_rq_prep_clone extern int blk_rq_prep_clone ( s#request * , s#request * , s#bio_set * , t#gfp_t , int ( * ) ( s#bio * , s#bio * , void * ) , void * )
+kblockd_schedule_work int kblockd_schedule_work ( s#work_struct * )
+kblockd_schedule_work_on int kblockd_schedule_work_on ( int , s#work_struct * )
+kblockd_mod_delayed_work_on int kblockd_mod_delayed_work_on ( int , s#delayed_work * , unsigned long )
+blk_start_plug extern void blk_start_plug ( s#blk_plug * )
+t#blk_plug_cb_fn typedef void ( * blk_plug_cb_fn ) ( s#blk_plug_cb * , t#bool )
+s#blk_plug_cb struct blk_plug_cb { s#list_head list ; t#blk_plug_cb_fn callback ; void * data ; }
+blk_check_plugged extern s#blk_plug_cb * blk_check_plugged ( t#blk_plug_cb_fn , void * , int )
+blk_finish_plug extern void blk_finish_plug ( s#blk_plug * )
+blk_pm_runtime_init extern void blk_pm_runtime_init ( s#request_queue * , s#device * )
+blk_pre_runtime_suspend extern int blk_pre_runtime_suspend ( s#request_queue * )
+blk_post_runtime_suspend extern void blk_post_runtime_suspend ( s#request_queue * , int )
+blk_pre_runtime_resume extern void blk_pre_runtime_resume ( s#request_queue * )
+blk_post_runtime_resume extern void blk_post_runtime_resume ( s#request_queue * , int )
+blk_set_runtime_active extern void blk_set_runtime_active ( s#request_queue * )
+
+/* block/blk-exec.o */
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##device_private
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##iov_iter
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##workqueue_struct
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##blkcg_gq
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_trace
+s##bsg_job
+s##throtl_data
+blk_execute_rq_nowait extern void blk_execute_rq_nowait ( s#request_queue * , s#gendisk * , s#request * , int , t#rq_end_io_fn * )
+blk_execute_rq extern void blk_execute_rq ( s#request_queue * , s#gendisk * , s#request * , int )
+
+/* block/blk-flush.o */
+s##posix_acl
+s##vfsmount
+s##sock
+s##iov_iter
+s##poll_table_struct
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##workqueue_struct
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##hd_geometry
+s##pr_ops
+s##blkcg_gq
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##nlm_lockowner
+s##nfs4_lock_state
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+blkdev_issue_flush extern int blkdev_issue_flush ( s#block_device * , t#gfp_t , t#sector_t * )
+
+/* block/blk-integrity.o */
+s##sock
+s##vfsmount
+s##posix_acl
+s##module
+s##device_private
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##iov_iter
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##workqueue_struct
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_trace
+s##bsg_job
+s##throtl_data
+blk_rq_count_integrity_sg extern int blk_rq_count_integrity_sg ( s#request_queue * , s#bio * )
+blk_rq_map_integrity_sg extern int blk_rq_map_integrity_sg ( s#request_queue * , s#bio * , s#scatterlist * )
+blk_integrity_compare extern int blk_integrity_compare ( s#gendisk * , s#gendisk * )
+blk_integrity_merge_rq extern t#bool blk_integrity_merge_rq ( s#request_queue * , s#request * , s#request * )
+blk_integrity_merge_bio extern t#bool blk_integrity_merge_bio ( s#request_queue * , s#request * , s#bio * )
+blk_integrity_register extern void blk_integrity_register ( s#gendisk * , s#blk_integrity * )
+blk_integrity_unregister extern void blk_integrity_unregister ( s#gendisk * )
+
+/* block/blk-ioc.o */
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##device_private
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##iov_iter
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##workqueue_struct
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##blkcg_gq
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_trace
+s##bsg_job
+s##throtl_data
+get_io_context void get_io_context ( s#io_context * )
+put_io_context void put_io_context ( s#io_context * )
+get_task_io_context s#io_context * get_task_io_context ( s#task_struct * , t#gfp_t , int )
+ioc_lookup_icq s#io_cq * ioc_lookup_icq ( s#io_context * , s#request_queue * )
+
+/* block/blk-lib.o */
+s##posix_acl
+s##vfsmount
+s##sock
+s##iov_iter
+s##poll_table_struct
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##workqueue_struct
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##hd_geometry
+s##pr_ops
+s##blkcg_gq
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##nlm_lockowner
+s##nfs4_lock_state
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+__blkdev_issue_discard extern int __blkdev_issue_discard ( s#block_device * , t#sector_t , t#sector_t , t#gfp_t , int , s#bio * * )
+blkdev_issue_discard extern int blkdev_issue_discard ( s#block_device * , t#sector_t , t#sector_t , t#gfp_t , unsigned long )
+blkdev_issue_write_same extern int blkdev_issue_write_same ( s#block_device * , t#sector_t , t#sector_t , t#gfp_t , s#page * )
+__blkdev_issue_zeroout extern int __blkdev_issue_zeroout ( s#block_device * , t#sector_t , t#sector_t , t#gfp_t , s#bio * * , unsigned )
+blkdev_issue_zeroout extern int blkdev_issue_zeroout ( s#block_device * , t#sector_t , t#sector_t , t#gfp_t , unsigned )
+
+/* block/blk-map.o */
+s##kmem_cache
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##pipe_inode_info
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blkcg_gq
+s##blk_trace
+s##bsg_job
+s##throtl_data
+blk_rq_append_bio extern int blk_rq_append_bio ( s#request * , s#bio * * )
+s#rq_map_data struct rq_map_data { s#page * * pages ; int page_order ; int nr_entries ; unsigned long offset ; int null_mapped ; int from_user ; }
+blk_rq_map_user_iov extern int blk_rq_map_user_iov ( s#request_queue * , s#request * , s#rq_map_data * , const s#iov_iter * , t#gfp_t )
+blk_rq_map_user extern int blk_rq_map_user ( s#request_queue * , s#request * , s#rq_map_data * , void * , unsigned long , t#gfp_t )
+blk_rq_unmap_user extern int blk_rq_unmap_user ( s#bio * )
+blk_rq_map_kern extern int blk_rq_map_kern ( s#request_queue * , s#request * , void * , unsigned int , t#gfp_t )
+
+/* block/blk-merge.o */
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##klp_modinfo
+s##device_private
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##pipe_inode_info
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##workqueue_struct
+s##robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##blkcg_gq
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##bsg_job
+s##throtl_data
+blk_queue_split extern void blk_queue_split ( s#request_queue * , s#bio * * )
+blk_recount_segments extern void blk_recount_segments ( s#request_queue * , s#bio * )
+blk_rq_map_sg extern int blk_rq_map_sg ( s#request_queue * , s#request * , s#scatterlist * )
+
+/* block/blk-mq.o */
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##klp_modinfo
+s##device_private
+s##sched_class
+s##cfs_rq
+s##task_group
+s##kmem_cache
+s##fs_pin
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##pipe_inode_info
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##eventfd_ctx
+s##mempolicy
+s##tty_struct
+s##tty_audit_buf
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##workqueue_struct
+s##robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s#blk_mq_debugfs_attr struct blk_mq_debugfs_attr { const char * name ; t#umode_t mode ; int ( * show ) ( void * , s#seq_file * ) ; t#ssize_t ( * write ) ( void * , const char * , t#size_t , t#loff_t * ) ; const s#seq_operations * seq_ops ; }
+s##blk_queue_stats
+s##bsg_job
+s##throtl_data
+blk_freeze_queue_start void blk_freeze_queue_start ( s#request_queue * )
+blk_mq_freeze_queue_wait void blk_mq_freeze_queue_wait ( s#request_queue * )
+blk_mq_freeze_queue_wait_timeout int blk_mq_freeze_queue_wait_timeout ( s#request_queue * , unsigned long )
+blk_mq_freeze_queue void blk_mq_freeze_queue ( s#request_queue * )
+blk_mq_unfreeze_queue void blk_mq_unfreeze_queue ( s#request_queue * )
+blk_mq_quiesce_queue_nowait void blk_mq_quiesce_queue_nowait ( s#request_queue * )
+blk_mq_quiesce_queue void blk_mq_quiesce_queue ( s#request_queue * )
+blk_mq_unquiesce_queue void blk_mq_unquiesce_queue ( s#request_queue * )
+blk_mq_can_queue t#bool blk_mq_can_queue ( s#blk_mq_hw_ctx * )
+blk_mq_alloc_request s#request * blk_mq_alloc_request ( s#request_queue * , unsigned int , t#blk_mq_req_flags_t )
+blk_mq_alloc_request_hctx s#request * blk_mq_alloc_request_hctx ( s#request_queue * , unsigned int , t#blk_mq_req_flags_t , unsigned int )
+blk_mq_free_request void blk_mq_free_request ( s#request * )
+__blk_mq_end_request void __blk_mq_end_request ( s#request * , t#blk_status_t )
+blk_mq_end_request void blk_mq_end_request ( s#request * , t#blk_status_t )
+blk_mq_complete_request void blk_mq_complete_request ( s#request * )
+blk_mq_request_started int blk_mq_request_started ( s#request * )
+blk_mq_start_request void blk_mq_start_request ( s#request * )
+blk_mq_requeue_request void blk_mq_requeue_request ( s#request * , t#bool )
+blk_mq_add_to_requeue_list void blk_mq_add_to_requeue_list ( s#request * , t#bool , t#bool )
+blk_mq_kick_requeue_list void blk_mq_kick_requeue_list ( s#request_queue * )
+blk_mq_delay_kick_requeue_list void blk_mq_delay_kick_requeue_list ( s#request_queue * , unsigned long )
+blk_mq_tag_to_rq s#request * blk_mq_tag_to_rq ( s#blk_mq_tags * , unsigned int )
+blk_mq_flush_busy_ctxs void blk_mq_flush_busy_ctxs ( s#blk_mq_hw_ctx * , s#list_head * )
+blk_mq_delay_run_hw_queue void blk_mq_delay_run_hw_queue ( s#blk_mq_hw_ctx * , unsigned long )
+blk_mq_run_hw_queue t#bool blk_mq_run_hw_queue ( s#blk_mq_hw_ctx * , t#bool )
+blk_mq_run_hw_queues void blk_mq_run_hw_queues ( s#request_queue * , t#bool )
+blk_mq_queue_stopped t#bool blk_mq_queue_stopped ( s#request_queue * )
+blk_mq_stop_hw_queue void blk_mq_stop_hw_queue ( s#blk_mq_hw_ctx * )
+blk_mq_stop_hw_queues void blk_mq_stop_hw_queues ( s#request_queue * )
+blk_mq_start_hw_queue void blk_mq_start_hw_queue ( s#blk_mq_hw_ctx * )
+blk_mq_start_hw_queues void blk_mq_start_hw_queues ( s#request_queue * )
+blk_mq_start_stopped_hw_queue void blk_mq_start_stopped_hw_queue ( s#blk_mq_hw_ctx * , t#bool )
+blk_mq_start_stopped_hw_queues void blk_mq_start_stopped_hw_queues ( s#request_queue * , t#bool )
+blk_mq_init_queue s#request_queue * blk_mq_init_queue ( s#blk_mq_tag_set * )
+blk_mq_init_allocated_queue s#request_queue * blk_mq_init_allocated_queue ( s#blk_mq_tag_set * , s#request_queue * )
+blk_mq_alloc_tag_set int blk_mq_alloc_tag_set ( s#blk_mq_tag_set * )
+blk_mq_free_tag_set void blk_mq_free_tag_set ( s#blk_mq_tag_set * )
+blk_mq_update_nr_hw_queues void blk_mq_update_nr_hw_queues ( s#blk_mq_tag_set * , int )
+
+/* block/blk-mq-cpumap.o */
+s##workqueue_struct
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##device_private
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##iov_iter
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##blkcg_gq
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_trace
+s##bsg_job
+s##throtl_data
+blk_mq_map_queues int blk_mq_map_queues ( s#blk_mq_tag_set * )
+
+/* block/blk-mq-debugfs.o */
+s##vfsmount
+s##posix_acl
+s##module
+s##device_private
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##iov_iter
+s##kmem_cache
+s##eventfd_ctx
+s##blkcg_gq
+s##workqueue_struct
+s##blk_queue_stats
+s##rq_wb
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+__blk_mq_debugfs_rq_show int __blk_mq_debugfs_rq_show ( s#seq_file * , s#request * )
+blk_mq_debugfs_rq_show int blk_mq_debugfs_rq_show ( s#seq_file * , void * )
+
+/* block/blk-mq-pci.o */
+s##workqueue_struct
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##device_private
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##iov_iter
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##blkcg_gq
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_trace
+s##blk_flush_queue
+s##bsg_job
+s##throtl_data
+s##msi_controller
+s##proc_dir_entry
+s##hotplug_slot
+s##pcie_link_state
+s##pci_vpd
+s##pci_sriov
+blk_mq_pci_map_queues int blk_mq_pci_map_queues ( s#blk_mq_tag_set * , s#pci_dev * , int )
+
+/* block/blk-mq-rdma.o */
+s##workqueue_struct
+s##blk_mq_ctx
+s##device_private
+t#__addrpair typedef t#__u64 __addrpair
+t#__be32 typedef t#__u32 __be32
+t#__portpair typedef t#__u32 __portpair
+t#__be16 typedef t#__u16 __be16
+t#__kernel_sa_family_t typedef unsigned short __kernel_sa_family_t
+t#sa_family_t typedef t#__kernel_sa_family_t sa_family_t
+s#sockaddr struct sockaddr { t#sa_family_t sa_family ; char sa_data [ 14 ] ; }
+s##posix_acl
+s##vfsmount
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##proc_dir_entry
+s##uevent_sock
+t#netdev_features_t typedef t#u64 netdev_features_t
+s#net_device_stats struct net_device_stats { unsigned long rx_packets ; unsigned long tx_packets ; unsigned long rx_bytes ; unsigned long tx_bytes ; unsigned long rx_errors ; unsigned long tx_errors ; unsigned long rx_dropped ; unsigned long tx_dropped ; unsigned long multicast ; unsigned long collisions ; unsigned long rx_length_errors ; unsigned long rx_over_errors ; unsigned long rx_crc_errors ; unsigned long rx_frame_errors ; unsigned long rx_fifo_errors ; unsigned long rx_missed_errors ; unsigned long tx_aborted_errors ; unsigned long tx_carrier_errors ; unsigned long tx_fifo_errors ; unsigned long tx_heartbeat_errors ; unsigned long tx_window_errors ; unsigned long rx_compressed ; unsigned long tx_compressed ; }
+s##iw_handler_def
+s##iw_public_data
+e#netdev_tx enum netdev_tx { __NETDEV_TX_MIN = ( - ( ( int ) ( ~ 0U >> 1 ) ) - 1 ) , NETDEV_TX_OK = 0x00 , NETDEV_TX_BUSY = 0x10 , }
+t#netdev_tx_t typedef e#netdev_tx netdev_tx_t
+s##sec_path
+s#in6_addr struct in6_addr { union { t#__u8 u6_addr8 [ 16 ] ; t#__be16 u6_addr16 [ 8 ] ; t#__be32 u6_addr32 [ 4 ] ; } in6_u ; }
+s#nf_bridge_info struct nf_bridge_info { t#atomic_t use ; enum { BRNF_PROTO_UNCHANGED , BRNF_PROTO_8021Q , BRNF_PROTO_PPPOE } orig_proto : 8 ; t#u8 pkt_otherhost : 1 ; t#u8 in_prerouting : 1 ; t#u8 bridged_dnat : 1 ; t#__u16 frag_max_size ; s#net_device * physindev ; s#net_device * physoutdev ; union { t#__be32 ipv4_daddr ; s#in6_addr ipv6_daddr ; char neigh_header [ 8 ] ; } ; }
+t#__wsum typedef t#__u32 __wsum
+t#sk_buff_data_t typedef unsigned int sk_buff_data_t
+s#sk_buff struct sk_buff { union { struct { s#sk_buff * next ; s#sk_buff * prev ; union { s#net_device * dev ; unsigned long dev_scratch ; } ; } ; s#rb_node rbnode ; s#list_head list ; } ; union { s#sock * sk ; int ip_defrag_offset ; } ; union { t#ktime_t tstamp ; t#u64 skb_mstamp ; } ; char cb [ 48 ] __attribute__ ( ( aligned ( 8 ) ) ) ; unsigned long _skb_refdst ; void ( * destructor ) ( s#sk_buff * ) ; s#sec_path * sp ; unsigned long _nfct ; s#nf_bridge_info * nf_bridge ; unsigned int len , data_len ; t#__u16 mac_len , hdr_len ; ; t#__u16 queue_mapping ; t#__u8 __cloned_offset [ 0 ] ; t#__u8 cloned : 1 , nohdr : 1 , fclone : 2 , peeked : 1 , head_frag : 1 , xmit_more : 1 , __unused : 1 ; ; t#__u32 headers_start [ 0 ] ; t#__u8 __pkt_type_offset [ 0 ] ; t#__u8 pkt_type : 3 ; t#__u8 pfmemalloc : 1 ; t#__u8 ignore_df : 1 ; t#__u8 nf_trace : 1 ; t#__u8 ip_summed : 2 ; t#__u8 ooo_okay : 1 ; t#__u8 l4_hash : 1 ; t#__u8 sw_hash : 1 ; t#__u8 wifi_acked_valid : 1 ; t#__u8 wifi_acked : 1 ; t#__u8 no_fcs : 1 ; t#__u8 encapsulation : 1 ; t#__u8 encap_hdr_csum : 1 ; t#__u8 csum_valid : 1 ; t#__u8 csum_complete_sw : 1 ; t#__u8 csum_level : 2 ; t#__u8 csum_not_inet : 1 ; t#__u8 dst_pending_confirm : 1 ; t#__u8 ndisc_nodetype : 2 ; t#__u8 ipvs_property : 1 ; t#__u8 inner_protocol_type : 1 ; t#__u8 remcsum_offload : 1 ; t#__u8 offload_fwd_mark : 1 ; t#__u8 offload_mr_fwd_mark : 1 ; t#__u8 tc_skip_classify : 1 ; t#__u8 tc_at_ingress : 1 ; t#__u8 tc_redirected : 1 ; t#__u8 tc_from_ingress : 1 ; t#__u8 decrypted : 1 ; t#__u16 tc_index ; union { t#__wsum csum ; struct { t#__u16 csum_start ; t#__u16 csum_offset ; } ; } ; t#__u32 priority ; int skb_iif ; t#__u32 hash ; t#__be16 vlan_proto ; t#__u16 vlan_tci ; union { unsigned int napi_id ; unsigned int sender_cpu ; } ; t#__u32 secmark ; union { t#__u32 mark ; t#__u32 reserved_tailroom ; } ; union { t#__be16 inner_protocol ; t#__u8 inner_ipproto ; } ; t#__u16 inner_transport_header ; t#__u16 inner_network_header ; t#__u16 inner_mac_header ; t#__be16 protocol ; t#__u16 transport_header ; t#__u16 network_header ; t#__u16 mac_header ; t#__u32 headers_end [ 0 ] ; t#sk_buff_data_t tail ; t#sk_buff_data_t end ; unsigned char * head , * data ; unsigned int truesize ; t#refcount_t users ; }
+t#select_queue_fallback_t typedef t#u16 ( * select_queue_fallback_t ) ( s#net_device * , s#sk_buff * , s#net_device * )
+s#ifmap struct ifmap { unsigned long mem_start ; unsigned long mem_end ; unsigned short base_addr ; unsigned char irq ; unsigned char dma ; unsigned char port ; }
+t#raw_hdlc_proto typedef struct { unsigned short encoding ; unsigned short parity ; } raw_hdlc_proto
+t#cisco_proto typedef struct { unsigned int interval ; unsigned int timeout ; } cisco_proto
+t#fr_proto typedef struct { unsigned int t391 ; unsigned int t392 ; unsigned int n391 ; unsigned int n392 ; unsigned int n393 ; unsigned short lmi ; unsigned short dce ; } fr_proto
+t#fr_proto_pvc typedef struct { unsigned int dlci ; } fr_proto_pvc
+t#fr_proto_pvc_info typedef struct { unsigned int dlci ; char master [ 16 ] ; } fr_proto_pvc_info
+t#sync_serial_settings typedef struct { unsigned int clock_rate ; unsigned int clock_type ; unsigned short loopback ; } sync_serial_settings
+t#te1_settings typedef struct { unsigned int clock_rate ; unsigned int clock_type ; unsigned short loopback ; unsigned int slot_map ; } te1_settings
+s#if_settings struct if_settings { unsigned int type ; unsigned int size ; union { t#raw_hdlc_proto * raw_hdlc ; t#cisco_proto * cisco ; t#fr_proto * fr ; t#fr_proto_pvc * fr_pvc ; t#fr_proto_pvc_info * fr_pvc_info ; t#sync_serial_settings * sync ; t#te1_settings * te1 ; } ifs_ifsu ; }
+s#ifreq struct ifreq { union { char ifrn_name [ 16 ] ; } ifr_ifrn ; union { s#sockaddr ifru_addr ; s#sockaddr ifru_dstaddr ; s#sockaddr ifru_broadaddr ; s#sockaddr ifru_netmask ; s#sockaddr ifru_hwaddr ; short ifru_flags ; int ifru_ivalue ; int ifru_mtu ; s#ifmap ifru_map ; char ifru_slave [ 16 ] ; char ifru_newname [ 16 ] ; void * ifru_data ; s#if_settings ifru_settings ; } ifr_ifru ; }
+t#possible_net_t typedef struct { s#net * net ; } possible_net_t
+s#pneigh_entry struct pneigh_entry { s#pneigh_entry * next ; t#possible_net_t net ; s#net_device * dev ; t#u8 flags ; t#u8 key [ 0 ] ; }
+s#sk_buff_head struct sk_buff_head { s#sk_buff * next ; s#sk_buff * prev ; t#__u32 qlen ; t#spinlock_t lock ; }
+s#neigh_statistics struct neigh_statistics { unsigned long allocs ; unsigned long destroys ; unsigned long hash_grows ; unsigned long res_failed ; unsigned long lookups ; unsigned long hits ; unsigned long rcv_probes_mcast ; unsigned long rcv_probes_ucast ; unsigned long periodic_gc_runs ; unsigned long forced_gc_runs ; unsigned long unres_discards ; unsigned long table_fulls ; }
+s#neigh_hash_table struct neigh_hash_table { s#neighbour * * hash_buckets ; unsigned int hash_shift ; t#__u32 hash_rnd [ 4 ] ; s#callback_head rcu ; }
+s#neigh_table struct neigh_table { int family ; int entry_size ; int key_len ; t#__be16 protocol ; t#__u32 ( * hash ) ( const void * , const s#net_device * , t#__u32 * ) ; t#bool ( * key_eq ) ( const s#neighbour * , const void * ) ; int ( * constructor ) ( s#neighbour * ) ; int ( * pconstructor ) ( s#pneigh_entry * ) ; void ( * pdestructor ) ( s#pneigh_entry * ) ; void ( * proxy_redo ) ( s#sk_buff * ) ; char * id ; s#neigh_parms parms ; s#list_head parms_list ; int gc_interval ; int gc_thresh1 ; int gc_thresh2 ; int gc_thresh3 ; unsigned long last_flush ; s#delayed_work gc_work ; s#timer_list proxy_timer ; s#sk_buff_head proxy_queue ; t#atomic_t entries ; t#rwlock_t lock ; unsigned long last_rand ; s#neigh_statistics * stats ; s#neigh_hash_table * nht ; s#pneigh_entry * * phash_buckets ; }
+s#hh_cache struct hh_cache { unsigned int hh_len ; t#seqlock_t hh_lock ; unsigned long hh_data [ ( ( ( 128 ) + ( 16 - 1 ) ) & ~ ( 16 - 1 ) ) / sizeof ( long ) ] ; }
+s#neigh_ops struct neigh_ops { int family ; void ( * solicit ) ( s#neighbour * , s#sk_buff * ) ; void ( * error_report ) ( s#neighbour * , s#sk_buff * ) ; int ( * output ) ( s#neighbour * , s#sk_buff * ) ; int ( * connected_output ) ( s#neighbour * , s#sk_buff * ) ; }
+s#neighbour struct neighbour { s#neighbour * next ; s#neigh_table * tbl ; s#neigh_parms * parms ; unsigned long confirmed ; unsigned long updated ; t#rwlock_t lock ; t#atomic_t refcnt ; s#sk_buff_head arp_queue ; unsigned int arp_queue_len_bytes ; s#timer_list timer ; unsigned long used ; t#atomic_t probes ; t#__u8 flags ; t#__u8 nud_state ; t#__u8 type ; t#__u8 dead ; t#seqlock_t ha_lock ; unsigned char ha [ ( ( ( ( 32 ) ) + ( ( typeof ( ( 32 ) ) ) ( ( sizeof ( unsigned long ) ) ) - 1 ) ) & ~ ( ( typeof ( ( 32 ) ) ) ( ( sizeof ( unsigned long ) ) ) - 1 ) ) ] ; s#hh_cache hh ; int ( * output ) ( s#neighbour * , s#sk_buff * ) ; const s#neigh_ops * ops ; s#callback_head rcu ; s#net_device * dev ; t#u8 primary_key [ 0 ] ; }
+E#NEIGH_VAR_LOCKTIME 12
+s#neigh_parms struct neigh_parms { t#possible_net_t net ; s#net_device * dev ; s#list_head list ; int ( * neigh_setup ) ( s#neighbour * ) ; void ( * neigh_cleanup ) ( s#neighbour * ) ; s#neigh_table * tbl ; void * sysctl_table ; int dead ; t#atomic_t refcnt ; s#callback_head callback_head ; int reachable_time ; int data [ ( E#NEIGH_VAR_LOCKTIME + 1 ) ] ; unsigned long data_state [ ( ( ( ( E#NEIGH_VAR_LOCKTIME + 1 ) ) + ( 8 * sizeof ( long ) ) - 1 ) / ( 8 * sizeof ( long ) ) ) ] ; }
+s#rtnl_link_stats64 struct rtnl_link_stats64 { t#__u64 rx_packets ; t#__u64 tx_packets ; t#__u64 rx_bytes ; t#__u64 tx_bytes ; t#__u64 rx_errors ; t#__u64 tx_errors ; t#__u64 rx_dropped ; t#__u64 tx_dropped ; t#__u64 multicast ; t#__u64 collisions ; t#__u64 rx_length_errors ; t#__u64 rx_over_errors ; t#__u64 rx_crc_errors ; t#__u64 rx_frame_errors ; t#__u64 rx_fifo_errors ; t#__u64 rx_missed_errors ; t#__u64 tx_aborted_errors ; t#__u64 tx_carrier_errors ; t#__u64 tx_fifo_errors ; t#__u64 tx_heartbeat_errors ; t#__u64 tx_window_errors ; t#__u64 rx_compressed ; t#__u64 tx_compressed ; t#__u64 rx_nohandler ; }
+s##netpoll_info
+s#ifla_vf_info struct ifla_vf_info { t#__u32 vf ; t#__u8 mac [ 32 ] ; t#__u32 vlan ; t#__u32 qos ; t#__u32 spoofchk ; t#__u32 linkstate ; t#__u32 min_tx_rate ; t#__u32 max_tx_rate ; t#__u32 rss_query_en ; t#__u32 trusted ; t#__be16 vlan_proto ; }
+s#ifla_vf_stats struct ifla_vf_stats { t#__u64 rx_packets ; t#__u64 tx_packets ; t#__u64 rx_bytes ; t#__u64 tx_bytes ; t#__u64 broadcast ; t#__u64 multicast ; t#__u64 rx_dropped ; t#__u64 tx_dropped ; }
+s#nlattr struct nlattr { t#__u16 nla_len ; t#__u16 nla_type ; }
+e#tc_setup_type enum tc_setup_type { TC_SETUP_QDISC_MQPRIO , TC_SETUP_CLSU32 , TC_SETUP_CLSFLOWER , TC_SETUP_CLSMATCHALL , TC_SETUP_CLSBPF , TC_SETUP_BLOCK , TC_SETUP_QDISC_CBS , TC_SETUP_QDISC_RED , TC_SETUP_QDISC_PRIO , TC_SETUP_QDISC_MQ , }
+s#netdev_fcoe_hbainfo struct netdev_fcoe_hbainfo { char manufacturer [ 64 ] ; char serial_number [ 64 ] ; char hardware_version [ 64 ] ; char driver_version [ 64 ] ; char optionrom_version [ 64 ] ; char firmware_version [ 64 ] ; char model [ 256 ] ; char model_description [ 256 ] ; }
+s#netlink_ext_ack struct netlink_ext_ack { const char * _msg ; const s#nlattr * bad_attr ; t#u8 cookie [ 20 ] ; t#u8 cookie_len ; }
+s#ndmsg struct ndmsg { t#__u8 ndm_family ; t#__u8 ndm_pad1 ; t#__u16 ndm_pad2 ; t#__s32 ndm_ifindex ; t#__u16 ndm_state ; t#__u8 ndm_flags ; t#__u8 ndm_type ; }
+s#nlmsghdr struct nlmsghdr { t#__u32 nlmsg_len ; t#__u16 nlmsg_type ; t#__u16 nlmsg_flags ; t#__u32 nlmsg_seq ; t#__u32 nlmsg_pid ; }
+s#netlink_callback struct netlink_callback { s#sk_buff * skb ; const s#nlmsghdr * nlh ; int ( * start ) ( s#netlink_callback * ) ; int ( * dump ) ( s#sk_buff * , s#netlink_callback * ) ; int ( * done ) ( s#netlink_callback * ) ; void * data ; s#module * module ; t#u16 family ; t#u16 min_dump_alloc ; unsigned int prev_seq , seq ; long args [ 6 ] ; }
+s#netdev_phys_item_id struct netdev_phys_item_id { unsigned char id [ 32 ] ; unsigned char id_len ; }
+s##udp_tunnel_info
+e#bpf_netdev_command enum bpf_netdev_command { XDP_SETUP_PROG , XDP_SETUP_PROG_HW , XDP_QUERY_PROG , XDP_QUERY_PROG_HW , BPF_OFFLOAD_VERIFIER_PREP , BPF_OFFLOAD_TRANSLATE , BPF_OFFLOAD_DESTROY , BPF_OFFLOAD_MAP_ALLOC , BPF_OFFLOAD_MAP_FREE , XDP_QUERY_XSK_UMEM , XDP_SETUP_XSK_UMEM , }
+e#bpf_prog_type enum bpf_prog_type { BPF_PROG_TYPE_UNSPEC , BPF_PROG_TYPE_SOCKET_FILTER , BPF_PROG_TYPE_KPROBE , BPF_PROG_TYPE_SCHED_CLS , BPF_PROG_TYPE_SCHED_ACT , BPF_PROG_TYPE_TRACEPOINT , BPF_PROG_TYPE_XDP , BPF_PROG_TYPE_PERF_EVENT , BPF_PROG_TYPE_CGROUP_SKB , BPF_PROG_TYPE_CGROUP_SOCK , BPF_PROG_TYPE_LWT_IN , BPF_PROG_TYPE_LWT_OUT , BPF_PROG_TYPE_LWT_XMIT , BPF_PROG_TYPE_SOCK_OPS , BPF_PROG_TYPE_SK_SKB , BPF_PROG_TYPE_CGROUP_DEVICE , BPF_PROG_TYPE_SK_MSG , BPF_PROG_TYPE_RAW_TRACEPOINT , BPF_PROG_TYPE_CGROUP_SOCK_ADDR , }
+e#bpf_attach_type enum bpf_attach_type { BPF_CGROUP_INET_INGRESS , BPF_CGROUP_INET_EGRESS , BPF_CGROUP_INET_SOCK_CREATE , BPF_CGROUP_SOCK_OPS , BPF_SK_SKB_STREAM_PARSER , BPF_SK_SKB_STREAM_VERDICT , BPF_CGROUP_DEVICE , BPF_SK_MSG_VERDICT , BPF_CGROUP_INET4_BIND , BPF_CGROUP_INET6_BIND , BPF_CGROUP_INET4_CONNECT , BPF_CGROUP_INET6_CONNECT , BPF_CGROUP_INET4_POST_BIND , BPF_CGROUP_INET6_POST_BIND , BPF_CGROUP_UDP4_SENDMSG , BPF_CGROUP_UDP6_SENDMSG , __MAX_BPF_ATTACH_TYPE }
+s##bpf_prog_aux
+s#sock_filter struct sock_filter { t#__u16 code ; t#__u8 jt ; t#__u8 jf ; t#__u32 k ; }
+s#sock_fprog_kern struct sock_fprog_kern { t#u16 len ; s#sock_filter * filter ; }
+t#__s16 typedef __signed__ short __s16
+s#bpf_insn struct bpf_insn { t#__u8 code ; t#__u8 dst_reg : 4 ; t#__u8 src_reg : 4 ; t#__s16 off ; t#__s32 imm ; }
+s#bpf_prog struct bpf_prog { t#u16 pages ; ; t#u16 jited : 1 , jit_requested : 1 , undo_set_mem : 1 , gpl_compatible : 1 , cb_access : 1 , dst_needed : 1 , blinded : 1 , is_func : 1 , has_callchain_buf : 1 ; e#bpf_prog_type type ; e#bpf_attach_type expected_attach_type ; t#u32 len ; t#u32 jited_len ; t#u8 tag [ 8 ] ; s#bpf_prog_aux * aux ; s#sock_fprog_kern * orig_prog ; unsigned int ( * bpf_func ) ( const void * , const s#bpf_insn * ) ; union { s#sock_filter insns [ 0 ] ; s#bpf_insn insnsi [ 0 ] ; } ; }
+s##bpf_prog_offload_ops
+s##bpf_offloaded_map
+s##xdp_umem
+s#netdev_bpf struct netdev_bpf { e#bpf_netdev_command command ; union { struct { t#u32 flags ; s#bpf_prog * prog ; s#netlink_ext_ack * extack ; } ; struct { t#u32 prog_id ; t#u32 prog_flags ; } ; struct { s#bpf_prog * prog ; const s#bpf_prog_offload_ops * ops ; } verifier ; struct { s#bpf_prog * prog ; } offload ; struct { s#bpf_offloaded_map * offmap ; } ; struct { s#xdp_umem * umem ; t#u16 queue_id ; } xsk ; } ; }
+s#xdp_mem_info struct xdp_mem_info { t#u32 type ; t#u32 id ; }
+s#xdp_frame struct xdp_frame { void * data ; t#u16 len ; t#u16 headroom ; t#u16 metasize ; s#xdp_mem_info mem ; s#net_device * dev_rx ; }
+s#net_device_ops struct net_device_ops { int ( * ndo_init ) ( s#net_device * ) ; void ( * ndo_uninit ) ( s#net_device * ) ; int ( * ndo_open ) ( s#net_device * ) ; int ( * ndo_stop ) ( s#net_device * ) ; t#netdev_tx_t ( * ndo_start_xmit ) ( s#sk_buff * , s#net_device * ) ; t#netdev_features_t ( * ndo_features_check ) ( s#sk_buff * , s#net_device * , t#netdev_features_t ) ; t#u16 ( * ndo_select_queue ) ( s#net_device * , s#sk_buff * , s#net_device * , t#select_queue_fallback_t ) ; void ( * ndo_change_rx_flags ) ( s#net_device * , int ) ; void ( * ndo_set_rx_mode ) ( s#net_device * ) ; int ( * ndo_set_mac_address ) ( s#net_device * , void * ) ; int ( * ndo_validate_addr ) ( s#net_device * ) ; int ( * ndo_do_ioctl ) ( s#net_device * , s#ifreq * , int ) ; int ( * ndo_set_config ) ( s#net_device * , s#ifmap * ) ; int ( * ndo_change_mtu ) ( s#net_device * , int ) ; int ( * ndo_neigh_setup ) ( s#net_device * , s#neigh_parms * ) ; void ( * ndo_tx_timeout ) ( s#net_device * ) ; void ( * ndo_get_stats64 ) ( s#net_device * , s#rtnl_link_stats64 * ) ; t#bool ( * ndo_has_offload_stats ) ( const s#net_device * , int ) ; int ( * ndo_get_offload_stats ) ( int , const s#net_device * , void * ) ; s#net_device_stats * ( * ndo_get_stats ) ( s#net_device * ) ; int ( * ndo_vlan_rx_add_vid ) ( s#net_device * , t#__be16 , t#u16 ) ; int ( * ndo_vlan_rx_kill_vid ) ( s#net_device * , t#__be16 , t#u16 ) ; void ( * ndo_poll_controller ) ( s#net_device * ) ; int ( * ndo_netpoll_setup ) ( s#net_device * , s#netpoll_info * ) ; void ( * ndo_netpoll_cleanup ) ( s#net_device * ) ; int ( * ndo_set_vf_mac ) ( s#net_device * , int , t#u8 * ) ; int ( * ndo_set_vf_vlan ) ( s#net_device * , int , t#u16 , t#u8 , t#__be16 ) ; int ( * ndo_set_vf_rate ) ( s#net_device * , int , int , int ) ; int ( * ndo_set_vf_spoofchk ) ( s#net_device * , int , t#bool ) ; int ( * ndo_set_vf_trust ) ( s#net_device * , int , t#bool ) ; int ( * ndo_get_vf_config ) ( s#net_device * , int , s#ifla_vf_info * ) ; int ( * ndo_set_vf_link_state ) ( s#net_device * , int , int ) ; int ( * ndo_get_vf_stats ) ( s#net_device * , int , s#ifla_vf_stats * ) ; int ( * ndo_set_vf_port ) ( s#net_device * , int , s#nlattr * [ ] ) ; int ( * ndo_get_vf_port ) ( s#net_device * , int , s#sk_buff * ) ; int ( * ndo_set_vf_guid ) ( s#net_device * , int , t#u64 , int ) ; int ( * ndo_set_vf_rss_query_en ) ( s#net_device * , int , t#bool ) ; int ( * ndo_setup_tc ) ( s#net_device * , e#tc_setup_type , void * ) ; int ( * ndo_fcoe_enable ) ( s#net_device * ) ; int ( * ndo_fcoe_disable ) ( s#net_device * ) ; int ( * ndo_fcoe_ddp_setup ) ( s#net_device * , t#u16 , s#scatterlist * , unsigned int ) ; int ( * ndo_fcoe_ddp_done ) ( s#net_device * , t#u16 ) ; int ( * ndo_fcoe_ddp_target ) ( s#net_device * , t#u16 , s#scatterlist * , unsigned int ) ; int ( * ndo_fcoe_get_hbainfo ) ( s#net_device * , s#netdev_fcoe_hbainfo * ) ; int ( * ndo_fcoe_get_wwn ) ( s#net_device * , t#u64 * , int ) ; int ( * ndo_rx_flow_steer ) ( s#net_device * , const s#sk_buff * , t#u16 , t#u32 ) ; int ( * ndo_add_slave ) ( s#net_device * , s#net_device * , s#netlink_ext_ack * ) ; int ( * ndo_del_slave ) ( s#net_device * , s#net_device * ) ; t#netdev_features_t ( * ndo_fix_features ) ( s#net_device * , t#netdev_features_t ) ; int ( * ndo_set_features ) ( s#net_device * , t#netdev_features_t ) ; int ( * ndo_neigh_construct ) ( s#net_device * , s#neighbour * ) ; void ( * ndo_neigh_destroy ) ( s#net_device * , s#neighbour * ) ; int ( * ndo_fdb_add ) ( s#ndmsg * , s#nlattr * [ ] , s#net_device * , const unsigned char * , t#u16 , t#u16 ) ; int ( * ndo_fdb_del ) ( s#ndmsg * , s#nlattr * [ ] , s#net_device * , const unsigned char * , t#u16 ) ; int ( * ndo_fdb_dump ) ( s#sk_buff * , s#netlink_callback * , s#net_device * , s#net_device * , int * ) ; int ( * ndo_bridge_setlink ) ( s#net_device * , s#nlmsghdr * , t#u16 ) ; int ( * ndo_bridge_getlink ) ( s#sk_buff * , t#u32 , t#u32 , s#net_device * , t#u32 , int ) ; int ( * ndo_bridge_dellink ) ( s#net_device * , s#nlmsghdr * , t#u16 ) ; int ( * ndo_change_carrier ) ( s#net_device * , t#bool ) ; int ( * ndo_get_phys_port_id ) ( s#net_device * , s#netdev_phys_item_id * ) ; int ( * ndo_get_phys_port_name ) ( s#net_device * , char * , t#size_t ) ; void ( * ndo_udp_tunnel_add ) ( s#net_device * , s#udp_tunnel_info * ) ; void ( * ndo_udp_tunnel_del ) ( s#net_device * , s#udp_tunnel_info * ) ; void * ( * ndo_dfwd_add_station ) ( s#net_device * , s#net_device * ) ; void ( * ndo_dfwd_del_station ) ( s#net_device * , void * ) ; int ( * ndo_get_lock_subclass ) ( s#net_device * ) ; int ( * ndo_set_tx_maxrate ) ( s#net_device * , int , t#u32 ) ; int ( * ndo_get_iflink ) ( const s#net_device * ) ; int ( * ndo_change_proto_down ) ( s#net_device * , t#bool ) ; int ( * ndo_fill_metadata_dst ) ( s#net_device * , s#sk_buff * ) ; void ( * ndo_set_rx_headroom ) ( s#net_device * , int ) ; int ( * ndo_bpf ) ( s#net_device * , s#netdev_bpf * ) ; int ( * ndo_xdp_xmit ) ( s#net_device * , int , s#xdp_frame * * , t#u32 ) ; int ( * ndo_xsk_async_xmit ) ( s#net_device * , t#u32 ) ; }
+s#ethtool_cmd struct ethtool_cmd { t#__u32 cmd ; t#__u32 supported ; t#__u32 advertising ; t#__u16 speed ; t#__u8 duplex ; t#__u8 port ; t#__u8 phy_address ; t#__u8 transceiver ; t#__u8 autoneg ; t#__u8 mdio_support ; t#__u32 maxtxpkt ; t#__u32 maxrxpkt ; t#__u16 speed_hi ; t#__u8 eth_tp_mdix ; t#__u8 eth_tp_mdix_ctrl ; t#__u32 lp_advertising ; t#__u32 reserved [ 2 ] ; }
+s#ethtool_drvinfo struct ethtool_drvinfo { t#__u32 cmd ; char driver [ 32 ] ; char version [ 32 ] ; char fw_version [ 32 ] ; char bus_info [ 32 ] ; char erom_version [ 32 ] ; char reserved2 [ 12 ] ; t#__u32 n_priv_flags ; t#__u32 n_stats ; t#__u32 testinfo_len ; t#__u32 eedump_len ; t#__u32 regdump_len ; }
+s#ethtool_regs struct ethtool_regs { t#__u32 cmd ; t#__u32 version ; t#__u32 len ; t#__u8 data [ 0 ] ; }
+s#ethtool_wolinfo struct ethtool_wolinfo { t#__u32 cmd ; t#__u32 supported ; t#__u32 wolopts ; t#__u8 sopass [ 6 ] ; }
+s#ethtool_eeprom struct ethtool_eeprom { t#__u32 cmd ; t#__u32 magic ; t#__u32 offset ; t#__u32 len ; t#__u8 data [ 0 ] ; }
+s#ethtool_coalesce struct ethtool_coalesce { t#__u32 cmd ; t#__u32 rx_coalesce_usecs ; t#__u32 rx_max_coalesced_frames ; t#__u32 rx_coalesce_usecs_irq ; t#__u32 rx_max_coalesced_frames_irq ; t#__u32 tx_coalesce_usecs ; t#__u32 tx_max_coalesced_frames ; t#__u32 tx_coalesce_usecs_irq ; t#__u32 tx_max_coalesced_frames_irq ; t#__u32 stats_block_coalesce_usecs ; t#__u32 use_adaptive_rx_coalesce ; t#__u32 use_adaptive_tx_coalesce ; t#__u32 pkt_rate_low ; t#__u32 rx_coalesce_usecs_low ; t#__u32 rx_max_coalesced_frames_low ; t#__u32 tx_coalesce_usecs_low ; t#__u32 tx_max_coalesced_frames_low ; t#__u32 pkt_rate_high ; t#__u32 rx_coalesce_usecs_high ; t#__u32 rx_max_coalesced_frames_high ; t#__u32 tx_coalesce_usecs_high ; t#__u32 tx_max_coalesced_frames_high ; t#__u32 rate_sample_interval ; }
+s#ethtool_ringparam struct ethtool_ringparam { t#__u32 cmd ; t#__u32 rx_max_pending ; t#__u32 rx_mini_max_pending ; t#__u32 rx_jumbo_max_pending ; t#__u32 tx_max_pending ; t#__u32 rx_pending ; t#__u32 rx_mini_pending ; t#__u32 rx_jumbo_pending ; t#__u32 tx_pending ; }
+s#ethtool_pauseparam struct ethtool_pauseparam { t#__u32 cmd ; t#__u32 autoneg ; t#__u32 rx_pause ; t#__u32 tx_pause ; }
+s#ethtool_test struct ethtool_test { t#__u32 cmd ; t#__u32 flags ; t#__u32 reserved ; t#__u32 len ; t#__u64 data [ 0 ] ; }
+e#ethtool_phys_id_state enum ethtool_phys_id_state { ETHTOOL_ID_INACTIVE , ETHTOOL_ID_ACTIVE , ETHTOOL_ID_ON , ETHTOOL_ID_OFF }
+s#ethtool_stats struct ethtool_stats { t#__u32 cmd ; t#__u32 n_stats ; t#__u64 data [ 0 ] ; }
+s#ethtool_tcpip4_spec struct ethtool_tcpip4_spec { t#__be32 ip4src ; t#__be32 ip4dst ; t#__be16 psrc ; t#__be16 pdst ; t#__u8 tos ; }
+s#ethtool_ah_espip4_spec struct ethtool_ah_espip4_spec { t#__be32 ip4src ; t#__be32 ip4dst ; t#__be32 spi ; t#__u8 tos ; }
+s#ethtool_usrip4_spec struct ethtool_usrip4_spec { t#__be32 ip4src ; t#__be32 ip4dst ; t#__be32 l4_4_bytes ; t#__u8 tos ; t#__u8 ip_ver ; t#__u8 proto ; }
+s#ethtool_tcpip6_spec struct ethtool_tcpip6_spec { t#__be32 ip6src [ 4 ] ; t#__be32 ip6dst [ 4 ] ; t#__be16 psrc ; t#__be16 pdst ; t#__u8 tclass ; }
+s#ethtool_ah_espip6_spec struct ethtool_ah_espip6_spec { t#__be32 ip6src [ 4 ] ; t#__be32 ip6dst [ 4 ] ; t#__be32 spi ; t#__u8 tclass ; }
+s#ethtool_usrip6_spec struct ethtool_usrip6_spec { t#__be32 ip6src [ 4 ] ; t#__be32 ip6dst [ 4 ] ; t#__be32 l4_4_bytes ; t#__u8 tclass ; t#__u8 l4_proto ; }
+s#ethhdr struct ethhdr { unsigned char h_dest [ 6 ] ; unsigned char h_source [ 6 ] ; t#__be16 h_proto ; }
+u#ethtool_flow_union union ethtool_flow_union { s#ethtool_tcpip4_spec tcp_ip4_spec ; s#ethtool_tcpip4_spec udp_ip4_spec ; s#ethtool_tcpip4_spec sctp_ip4_spec ; s#ethtool_ah_espip4_spec ah_ip4_spec ; s#ethtool_ah_espip4_spec esp_ip4_spec ; s#ethtool_usrip4_spec usr_ip4_spec ; s#ethtool_tcpip6_spec tcp_ip6_spec ; s#ethtool_tcpip6_spec udp_ip6_spec ; s#ethtool_tcpip6_spec sctp_ip6_spec ; s#ethtool_ah_espip6_spec ah_ip6_spec ; s#ethtool_ah_espip6_spec esp_ip6_spec ; s#ethtool_usrip6_spec usr_ip6_spec ; s#ethhdr ether_spec ; t#__u8 hdata [ 52 ] ; }
+s#ethtool_flow_ext struct ethtool_flow_ext { t#__u8 padding [ 2 ] ; unsigned char h_dest [ 6 ] ; t#__be16 vlan_etype ; t#__be16 vlan_tci ; t#__be32 data [ 2 ] ; }
+s#ethtool_rx_flow_spec struct ethtool_rx_flow_spec { t#__u32 flow_type ; u#ethtool_flow_union h_u ; s#ethtool_flow_ext h_ext ; u#ethtool_flow_union m_u ; s#ethtool_flow_ext m_ext ; t#__u64 ring_cookie ; t#__u32 location ; }
+s#ethtool_rxnfc struct ethtool_rxnfc { t#__u32 cmd ; t#__u32 flow_type ; t#__u64 data ; s#ethtool_rx_flow_spec fs ; union { t#__u32 rule_cnt ; t#__u32 rss_context ; } ; t#__u32 rule_locs [ 0 ] ; }
+s#ethtool_flash struct ethtool_flash { t#__u32 cmd ; t#__u32 region ; char data [ 128 ] ; }
+s#ethtool_channels struct ethtool_channels { t#__u32 cmd ; t#__u32 max_rx ; t#__u32 max_tx ; t#__u32 max_other ; t#__u32 max_combined ; t#__u32 rx_count ; t#__u32 tx_count ; t#__u32 other_count ; t#__u32 combined_count ; }
+s#ethtool_dump struct ethtool_dump { t#__u32 cmd ; t#__u32 version ; t#__u32 flag ; t#__u32 len ; t#__u8 data [ 0 ] ; }
+s#ethtool_ts_info struct ethtool_ts_info { t#__u32 cmd ; t#__u32 so_timestamping ; t#__s32 phc_index ; t#__u32 tx_types ; t#__u32 tx_reserved [ 3 ] ; t#__u32 rx_filters ; t#__u32 rx_reserved [ 3 ] ; }
+s#ethtool_modinfo struct ethtool_modinfo { t#__u32 cmd ; t#__u32 type ; t#__u32 eeprom_len ; t#__u32 reserved [ 8 ] ; }
+s#ethtool_eee struct ethtool_eee { t#__u32 cmd ; t#__u32 supported ; t#__u32 advertised ; t#__u32 lp_advertised ; t#__u32 eee_active ; t#__u32 eee_enabled ; t#__u32 tx_lpi_enabled ; t#__u32 tx_lpi_timer ; t#__u32 reserved [ 2 ] ; }
+s#ethtool_tunable struct ethtool_tunable { t#__u32 cmd ; t#__u32 id ; t#__u32 type_id ; t#__u32 len ; void * data [ 0 ] ; }
+t#__s8 typedef __signed__ char __s8
+s#ethtool_link_settings struct ethtool_link_settings { t#__u32 cmd ; t#__u32 speed ; t#__u8 duplex ; t#__u8 port ; t#__u8 phy_address ; t#__u8 autoneg ; t#__u8 mdio_support ; t#__u8 eth_tp_mdix ; t#__u8 eth_tp_mdix_ctrl ; t#__s8 link_mode_masks_nwords ; t#__u8 transceiver ; t#__u8 reserved1 [ 3 ] ; t#__u32 reserved [ 7 ] ; t#__u32 link_mode_masks [ 0 ] ; }
+E#ETHTOOL_LINK_MODE_FEC_BASER_BIT 51
+E#__ETHTOOL_LINK_MODE_LAST E#ETHTOOL_LINK_MODE_FEC_BASER_BIT
+s#ethtool_link_ksettings struct ethtool_link_ksettings { s#ethtool_link_settings base ; struct { unsigned long supported [ ( ( ( ( E#__ETHTOOL_LINK_MODE_LAST + 1 ) ) + ( 8 * sizeof ( long ) ) - 1 ) / ( 8 * sizeof ( long ) ) ) ] ; unsigned long advertising [ ( ( ( ( E#__ETHTOOL_LINK_MODE_LAST + 1 ) ) + ( 8 * sizeof ( long ) ) - 1 ) / ( 8 * sizeof ( long ) ) ) ] ; unsigned long lp_advertising [ ( ( ( ( E#__ETHTOOL_LINK_MODE_LAST + 1 ) ) + ( 8 * sizeof ( long ) ) - 1 ) / ( 8 * sizeof ( long ) ) ) ] ; } link_modes ; }
+s#ethtool_fecparam struct ethtool_fecparam { t#__u32 cmd ; t#__u32 active_fec ; t#__u32 fec ; t#__u32 reserved ; }
+s#ethtool_ops struct ethtool_ops { int ( * get_settings ) ( s#net_device * , s#ethtool_cmd * ) ; int ( * set_settings ) ( s#net_device * , s#ethtool_cmd * ) ; void ( * get_drvinfo ) ( s#net_device * , s#ethtool_drvinfo * ) ; int ( * get_regs_len ) ( s#net_device * ) ; void ( * get_regs ) ( s#net_device * , s#ethtool_regs * , void * ) ; void ( * get_wol ) ( s#net_device * , s#ethtool_wolinfo * ) ; int ( * set_wol ) ( s#net_device * , s#ethtool_wolinfo * ) ; t#u32 ( * get_msglevel ) ( s#net_device * ) ; void ( * set_msglevel ) ( s#net_device * , t#u32 ) ; int ( * nway_reset ) ( s#net_device * ) ; t#u32 ( * get_link ) ( s#net_device * ) ; int ( * get_eeprom_len ) ( s#net_device * ) ; int ( * get_eeprom ) ( s#net_device * , s#ethtool_eeprom * , t#u8 * ) ; int ( * set_eeprom ) ( s#net_device * , s#ethtool_eeprom * , t#u8 * ) ; int ( * get_coalesce ) ( s#net_device * , s#ethtool_coalesce * ) ; int ( * set_coalesce ) ( s#net_device * , s#ethtool_coalesce * ) ; void ( * get_ringparam ) ( s#net_device * , s#ethtool_ringparam * ) ; int ( * set_ringparam ) ( s#net_device * , s#ethtool_ringparam * ) ; void ( * get_pauseparam ) ( s#net_device * , s#ethtool_pauseparam * ) ; int ( * set_pauseparam ) ( s#net_device * , s#ethtool_pauseparam * ) ; void ( * self_test ) ( s#net_device * , s#ethtool_test * , t#u64 * ) ; void ( * get_strings ) ( s#net_device * , t#u32 , t#u8 * ) ; int ( * set_phys_id ) ( s#net_device * , e#ethtool_phys_id_state ) ; void ( * get_ethtool_stats ) ( s#net_device * , s#ethtool_stats * , t#u64 * ) ; int ( * begin ) ( s#net_device * ) ; void ( * complete ) ( s#net_device * ) ; t#u32 ( * get_priv_flags ) ( s#net_device * ) ; int ( * set_priv_flags ) ( s#net_device * , t#u32 ) ; int ( * get_sset_count ) ( s#net_device * , int ) ; int ( * get_rxnfc ) ( s#net_device * , s#ethtool_rxnfc * , t#u32 * ) ; int ( * set_rxnfc ) ( s#net_device * , s#ethtool_rxnfc * ) ; int ( * flash_device ) ( s#net_device * , s#ethtool_flash * ) ; int ( * reset ) ( s#net_device * , t#u32 * ) ; t#u32 ( * get_rxfh_key_size ) ( s#net_device * ) ; t#u32 ( * get_rxfh_indir_size ) ( s#net_device * ) ; int ( * get_rxfh ) ( s#net_device * , t#u32 * , t#u8 * , t#u8 * ) ; int ( * set_rxfh ) ( s#net_device * , const t#u32 * , const t#u8 * , const t#u8 ) ; int ( * get_rxfh_context ) ( s#net_device * , t#u32 * , t#u8 * , t#u8 * , t#u32 ) ; int ( * set_rxfh_context ) ( s#net_device * , const t#u32 * , const t#u8 * , const t#u8 , t#u32 * , t#bool ) ; void ( * get_channels ) ( s#net_device * , s#ethtool_channels * ) ; int ( * set_channels ) ( s#net_device * , s#ethtool_channels * ) ; int ( * get_dump_flag ) ( s#net_device * , s#ethtool_dump * ) ; int ( * get_dump_data ) ( s#net_device * , s#ethtool_dump * , void * ) ; int ( * set_dump ) ( s#net_device * , s#ethtool_dump * ) ; int ( * get_ts_info ) ( s#net_device * , s#ethtool_ts_info * ) ; int ( * get_module_info ) ( s#net_device * , s#ethtool_modinfo * ) ; int ( * get_module_eeprom ) ( s#net_device * , s#ethtool_eeprom * , t#u8 * ) ; int ( * get_eee ) ( s#net_device * , s#ethtool_eee * ) ; int ( * set_eee ) ( s#net_device * , s#ethtool_eee * ) ; int ( * get_tunable ) ( s#net_device * , const s#ethtool_tunable * , void * ) ; int ( * set_tunable ) ( s#net_device * , const s#ethtool_tunable * , const void * ) ; int ( * get_per_queue_coalesce ) ( s#net_device * , t#u32 , s#ethtool_coalesce * ) ; int ( * set_per_queue_coalesce ) ( s#net_device * , t#u32 , s#ethtool_coalesce * ) ; int ( * get_link_ksettings ) ( s#net_device * , s#ethtool_link_ksettings * ) ; int ( * set_link_ksettings ) ( s#net_device * , const s#ethtool_link_ksettings * ) ; int ( * get_fecparam ) ( s#net_device * , s#ethtool_fecparam * ) ; int ( * set_fecparam ) ( s#net_device * , s#ethtool_fecparam * ) ; }
+s##switchdev_ops
+s##kmem_cache
+s#dst_ops struct dst_ops { unsigned short family ; unsigned int gc_thresh ; int ( * gc ) ( s#dst_ops * ) ; s#dst_entry * ( * check ) ( s#dst_entry * , t#__u32 ) ; unsigned int ( * default_advmss ) ( const s#dst_entry * ) ; unsigned int ( * mtu ) ( const s#dst_entry * ) ; t#u32 * ( * cow_metrics ) ( s#dst_entry * , unsigned long ) ; void ( * destroy ) ( s#dst_entry * ) ; void ( * ifdown ) ( s#dst_entry * , s#net_device * , int ) ; s#dst_entry * ( * negative_advice ) ( s#dst_entry * ) ; void ( * link_failure ) ( s#sk_buff * ) ; void ( * update_pmtu ) ( s#dst_entry * , s#sock * , s#sk_buff * , t#u32 ) ; void ( * redirect ) ( s#dst_entry * , s#sock * , s#sk_buff * ) ; int ( * local_out ) ( s#net * , s#sock * , s#sk_buff * ) ; s#neighbour * ( * neigh_lookup ) ( const s#dst_entry * , s#sk_buff * , const void * ) ; void ( * confirm_neigh ) ( const s#dst_entry * , const void * ) ; s#kmem_cache * kmem_cachep ; s#percpu_counter pcpuc_entries __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; }
+s##xfrm_state
+s##lwtunnel_state
+s##uncached_list
+s#rtable struct rtable { s#dst_entry dst ; int rt_genid ; unsigned int rt_flags ; t#__u16 rt_type ; t#__u8 rt_is_input ; t#__u8 rt_uses_gateway ; int rt_iif ; t#__be32 rt_gateway ; t#u32 rt_mtu_locked : 1 , rt_pmtu : 31 ; t#u32 rt_table_id ; s#list_head rt_uncached ; s#uncached_list * rt_uncached_list ; }
+s##rt6_info
+s##dn_route
+s#dst_entry struct dst_entry { s#callback_head callback_head ; s#dst_entry * child ; s#net_device * dev ; s#dst_ops * ops ; unsigned long _metrics ; unsigned long expires ; s#dst_entry * path ; s#dst_entry * from ; s#xfrm_state * xfrm ; int ( * input ) ( s#sk_buff * ) ; int ( * output ) ( s#net * , s#sock * , s#sk_buff * ) ; unsigned short flags ; short error ; short obsolete ; unsigned short header_len ; unsigned short trailer_len ; unsigned short __pad3 ; t#__u32 tclassid ; long __pad_to_align_refcnt [ 2 ] ; t#atomic_t __refcnt ; int __use ; unsigned long lastuse ; s#lwtunnel_state * lwtstate ; union { s#dst_entry * next ; s#rtable * rt_next ; s#rt6_info * rt6_next ; s#dn_route * dn_next ; } ; }
+t#__be64 typedef t#__u64 __be64
+s#flowi_tunnel struct flowi_tunnel { t#__be64 tun_id ; }
+s#flowi_common struct flowi_common { int flowic_oif ; int flowic_iif ; t#__u32 flowic_mark ; t#__u8 flowic_tos ; t#__u8 flowic_scope ; t#__u8 flowic_proto ; t#__u8 flowic_flags ; t#__u32 flowic_secid ; s#flowi_tunnel flowic_tun_key ; t#kuid_t flowic_uid ; }
+t#__le16 typedef t#__u16 __le16
+u#flowi_uli union flowi_uli { struct { t#__be16 dport ; t#__be16 sport ; } ports ; struct { t#__u8 type ; t#__u8 code ; } icmpt ; struct { t#__le16 dport ; t#__le16 sport ; } dnports ; t#__be32 spi ; t#__be32 gre_key ; struct { t#__u8 type ; } mht ; }
+s#flowi6 struct flowi6 { s#flowi_common __fl_common ; s#in6_addr daddr ; s#in6_addr saddr ; t#__be32 flowlabel ; u#flowi_uli uli ; }
+s#l3mdev_ops struct l3mdev_ops { t#u32 ( * l3mdev_fib_table ) ( const s#net_device * ) ; s#sk_buff * ( * l3mdev_l3_rcv ) ( s#net_device * , s#sk_buff * , t#u16 ) ; s#sk_buff * ( * l3mdev_l3_out ) ( s#net_device * , s#sock * , s#sk_buff * , t#u16 ) ; s#dst_entry * ( * l3mdev_link_scope_lookup ) ( const s#net_device * , s#flowi6 * ) ; }
+s#nd_opt_hdr struct nd_opt_hdr { t#__u8 nd_opt_type ; t#__u8 nd_opt_len ; }
+E#__ND_OPT_ARRAY_MAX ( 14 ) + 1
+s#ndisc_options struct ndisc_options { s#nd_opt_hdr * nd_opt_array [ E#__ND_OPT_ARRAY_MAX ] ; s#nd_opt_hdr * nd_opts_ri ; s#nd_opt_hdr * nd_opts_ri_end ; s#nd_opt_hdr * nd_useropts ; s#nd_opt_hdr * nd_useropts_end ; }
+s##prefix_info
+s#ip6_sf_list struct ip6_sf_list { s#ip6_sf_list * sf_next ; s#in6_addr sf_addr ; unsigned long sf_count [ 2 ] ; unsigned char sf_gsresp ; unsigned char sf_oldin ; unsigned char sf_crcount ; }
+s#ifmcaddr6 struct ifmcaddr6 { s#in6_addr mca_addr ; s#inet6_dev * idev ; s#ifmcaddr6 * next ; s#ip6_sf_list * mca_sources ; s#ip6_sf_list * mca_tomb ; unsigned int mca_sfmode ; unsigned char mca_crcount ; unsigned long mca_sfcount [ 2 ] ; s#timer_list mca_timer ; unsigned int mca_flags ; int mca_users ; t#atomic_t mca_refcnt ; t#spinlock_t mca_lock ; unsigned long mca_cstamp ; unsigned long mca_tstamp ; }
+s#ifacaddr6 struct ifacaddr6 { s#in6_addr aca_addr ; s#inet6_dev * aca_idev ; s#rt6_info * aca_rt ; s#ifacaddr6 * aca_next ; int aca_users ; t#atomic_t aca_refcnt ; unsigned long aca_cstamp ; unsigned long aca_tstamp ; }
+s#ipv6_stable_secret struct ipv6_stable_secret { t#bool initialized ; s#in6_addr secret ; }
+s#ipv6_devconf struct ipv6_devconf { t#__s32 forwarding ; t#__s32 hop_limit ; t#__s32 mtu6 ; t#__s32 accept_ra ; t#__s32 accept_redirects ; t#__s32 autoconf ; t#__s32 dad_transmits ; t#__s32 rtr_solicits ; t#__s32 rtr_solicit_interval ; t#__s32 rtr_solicit_max_interval ; t#__s32 rtr_solicit_delay ; t#__s32 force_mld_version ; t#__s32 mldv1_unsolicited_report_interval ; t#__s32 mldv2_unsolicited_report_interval ; t#__s32 use_tempaddr ; t#__s32 temp_valid_lft ; t#__s32 temp_prefered_lft ; t#__s32 regen_max_retry ; t#__s32 max_desync_factor ; t#__s32 max_addresses ; t#__s32 accept_ra_defrtr ; t#__s32 accept_ra_min_hop_limit ; t#__s32 accept_ra_pinfo ; t#__s32 ignore_routes_with_linkdown ; t#__s32 accept_ra_rtr_pref ; t#__s32 rtr_probe_interval ; t#__s32 accept_ra_rt_info_min_plen ; t#__s32 accept_ra_rt_info_max_plen ; t#__s32 proxy_ndp ; t#__s32 accept_source_route ; t#__s32 accept_ra_from_local ; t#__s32 mc_forwarding ; t#__s32 disable_ipv6 ; t#__s32 drop_unicast_in_l2_multicast ; t#__s32 accept_dad ; t#__s32 force_tllao ; t#__s32 ndisc_notify ; t#__s32 suppress_frag_ndisc ; t#__s32 accept_ra_mtu ; t#__s32 drop_unsolicited_na ; s#ipv6_stable_secret stable_secret ; t#__s32 use_oif_addrs_only ; t#__s32 keep_addr_on_down ; t#__s32 seg6_enabled ; t#__s32 seg6_require_hmac ; t#__u32 enhanced_dad ; t#__u32 addr_gen_mode ; t#__s32 disable_policy ; s#ctl_table_header * sysctl_header ; }
+E#__IPSTATS_MIB_MAX ( 0 ) + 37
+s#u64_stats_sync struct u64_stats_sync { }
+s#ipstats_mib struct ipstats_mib { t#u64 mibs [ E#__IPSTATS_MIB_MAX ] ; s#u64_stats_sync syncp ; }
+E#__ICMP6_MIB_MAX ( 0 ) + 6
+s#icmpv6_mib_device struct icmpv6_mib_device { t#atomic_long_t mibs [ E#__ICMP6_MIB_MAX ] ; }
+s#icmpv6msg_mib_device struct icmpv6msg_mib_device { t#atomic_long_t mibs [ 512 ] ; }
+s#ipv6_devstat struct ipv6_devstat { s#proc_dir_entry * proc_dir_entry ; __typeof__ ( s#ipstats_mib ) * ipv6 ; __typeof__ ( s#icmpv6_mib_device ) * icmpv6dev ; __typeof__ ( s#icmpv6msg_mib_device ) * icmpv6msgdev ; }
+s#inet6_dev struct inet6_dev { s#net_device * dev ; s#list_head addr_list ; s#ifmcaddr6 * mc_list ; s#ifmcaddr6 * mc_tomb ; t#spinlock_t mc_lock ; unsigned char mc_qrv ; unsigned char mc_gq_running ; unsigned char mc_ifc_count ; unsigned char mc_dad_count ; unsigned long mc_v1_seen ; unsigned long mc_qi ; unsigned long mc_qri ; unsigned long mc_maxdelay ; s#timer_list mc_gq_timer ; s#timer_list mc_ifc_timer ; s#timer_list mc_dad_timer ; s#ifacaddr6 * ac_list ; t#rwlock_t lock ; t#atomic_t refcnt ; t#__u32 if_flags ; int dead ; t#u32 desync_factor ; t#u8 rndid [ 8 ] ; s#list_head tempaddr_list ; s#in6_addr token ; s#neigh_parms * nd_parms ; s#ipv6_devconf cnf ; s#ipv6_devstat stats ; s#timer_list rs_timer ; t#__s32 rs_interval ; t#__u8 rs_probes ; unsigned long tstamp ; s#callback_head rcu ; }
+s#ndisc_ops struct ndisc_ops { int ( * is_useropt ) ( t#u8 ) ; int ( * parse_options ) ( const s#net_device * , s#nd_opt_hdr * , s#ndisc_options * ) ; void ( * update ) ( const s#net_device * , s#neighbour * , t#u32 , t#u8 , const s#ndisc_options * ) ; int ( * opt_addr_space ) ( const s#net_device * , t#u8 , s#neighbour * , t#u8 * , t#u8 * * ) ; void ( * fill_addr_option ) ( const s#net_device * , s#sk_buff * , t#u8 , const t#u8 * ) ; void ( * prefix_rcv_add_addr ) ( s#net * , s#net_device * , const s#prefix_info * , s#inet6_dev * , s#in6_addr * , int , t#u32 , t#bool , t#bool , t#__u32 , t#u32 , t#bool ) ; }
+s#xfrmdev_ops struct xfrmdev_ops { int ( * xdo_dev_state_add ) ( s#xfrm_state * ) ; void ( * xdo_dev_state_delete ) ( s#xfrm_state * ) ; void ( * xdo_dev_state_free ) ( s#xfrm_state * ) ; t#bool ( * xdo_dev_offload_ok ) ( s#sk_buff * , s#xfrm_state * ) ; void ( * xdo_dev_state_advance_esn ) ( s#xfrm_state * ) ; }
+e#tls_offload_ctx_dir enum tls_offload_ctx_dir { TLS_OFFLOAD_CTX_DIR_RX , TLS_OFFLOAD_CTX_DIR_TX , }
+s##tls_crypto_info
+s##tls_context
+s#tlsdev_ops struct tlsdev_ops { int ( * tls_dev_add ) ( s#net_device * , s#sock * , e#tls_offload_ctx_dir , s#tls_crypto_info * , t#u32 ) ; void ( * tls_dev_del ) ( s#net_device * , s#tls_context * , e#tls_offload_ctx_dir ) ; void ( * tls_dev_resync_rx ) ( s#net_device * , s#sock * , t#u32 , t#u64 ) ; }
+s#header_ops struct header_ops { int ( * create ) ( s#sk_buff * , s#net_device * , unsigned short , const void * , const void * , unsigned int ) ; int ( * parse ) ( const s#sk_buff * , unsigned char * ) ; int ( * cache ) ( const s#neighbour * , s#hh_cache * , t#__be16 ) ; void ( * cache_update ) ( s#hh_cache * , const s#net_device * , const unsigned char * ) ; t#bool ( * validate ) ( const char * , unsigned int ) ; }
+s#netdev_hw_addr_list struct netdev_hw_addr_list { s#list_head list ; int count ; }
+s##vlan_info
+s##dsa_switch_tree
+s##in_device
+s##dn_dev
+s##wireless_dev
+s##wpan_dev
+s##mpls_dev
+s#rps_map struct rps_map { unsigned int len ; s#callback_head rcu ; t#u16 cpus [ 0 ] ; }
+s#rps_dev_flow struct rps_dev_flow { t#u16 cpu ; t#u16 filter ; unsigned int last_qtail ; }
+s#rps_dev_flow_table struct rps_dev_flow_table { unsigned int mask ; s#callback_head rcu ; s#rps_dev_flow flows [ 0 ] ; }
+s#xdp_rxq_info struct xdp_rxq_info { s#net_device * dev ; t#u32 queue_index ; t#u32 reg_state ; s#xdp_mem_info mem ; }
+s#netdev_rx_queue struct netdev_rx_queue { s#rps_map * rps_map ; s#rps_dev_flow_table * rps_flow_table ; s#kobject kobj ; s#net_device * dev ; s#xdp_rxq_info xdp_rxq ; s#xdp_umem * umem ; }
+e#rx_handler_result enum rx_handler_result { RX_HANDLER_CONSUMED , RX_HANDLER_ANOTHER , RX_HANDLER_EXACT , RX_HANDLER_PASS , }
+t#rx_handler_result_t typedef e#rx_handler_result rx_handler_result_t
+t#rx_handler_func_t typedef t#rx_handler_result_t rx_handler_func_t ( s#sk_buff * * )
+s#gnet_stats_queue struct gnet_stats_queue { t#__u32 qlen ; t#__u32 backlog ; t#__u32 drops ; t#__u32 requeues ; t#__u32 overlimits ; }
+s#tcf_result struct tcf_result { union { struct { unsigned long class ; t#u32 classid ; } ; const s#tcf_proto * goto_tp ; struct { t#bool ingress ; s#gnet_stats_queue * qstats ; } ; } ; }
+s##tcf_walker
+t#tc_setup_cb_t typedef int tc_setup_cb_t ( e#tc_setup_type , void * , void * )
+s#dql struct dql { unsigned int num_queued ; unsigned int adj_limit ; unsigned int last_obj_cnt ; unsigned int limit __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; unsigned int num_completed ; unsigned int prev_ovlimit ; unsigned int prev_num_queued ; unsigned int prev_last_obj_cnt ; unsigned int lowest_slack ; unsigned long slack_start_time ; unsigned int max_limit ; unsigned int min_limit ; unsigned int slack_hold_time ; }
+s#netdev_queue struct netdev_queue { s#net_device * dev ; s#Qdisc * qdisc ; s#Qdisc * qdisc_sleeping ; s#kobject kobj ; int numa_node ; unsigned long tx_maxrate ; unsigned long trans_timeout ; s#net_device * sb_dev ; s#xdp_umem * umem ; t#spinlock_t _xmit_lock __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; int xmit_lock_owner ; unsigned long trans_start ; unsigned long state ; s#dql dql ; }
+s#tcmsg struct tcmsg { unsigned char tcm_family ; unsigned char tcm__pad1 ; unsigned short tcm__pad2 ; int tcm_ifindex ; t#__u32 tcm_handle ; t#__u32 tcm_parent ; t#__u32 tcm_info ; }
+s##qdisc_walker
+s#tc_stats struct tc_stats { t#__u64 bytes ; t#__u32 packets ; t#__u32 drops ; t#__u32 overlimits ; t#__u32 bps ; t#__u32 pps ; t#__u32 qlen ; t#__u32 backlog ; }
+s#gnet_dump struct gnet_dump { t#spinlock_t * lock ; s#sk_buff * skb ; s#nlattr * tail ; int compat_tc_stats ; int compat_xstats ; int padattr ; void * xstats ; int xstats_len ; s#tc_stats tc_stats ; }
+s#Qdisc_class_ops struct Qdisc_class_ops { s#netdev_queue * ( * select_queue ) ( s#Qdisc * , s#tcmsg * ) ; int ( * graft ) ( s#Qdisc * , unsigned long , s#Qdisc * , s#Qdisc * * , s#netlink_ext_ack * ) ; s#Qdisc * ( * leaf ) ( s#Qdisc * , unsigned long ) ; void ( * qlen_notify ) ( s#Qdisc * , unsigned long ) ; unsigned long ( * find ) ( s#Qdisc * , t#u32 ) ; int ( * change ) ( s#Qdisc * , t#u32 , t#u32 , s#nlattr * * , unsigned long * , s#netlink_ext_ack * ) ; int ( * delete ) ( s#Qdisc * , unsigned long ) ; void ( * walk ) ( s#Qdisc * , s#qdisc_walker * ) ; s#tcf_block * ( * tcf_block ) ( s#Qdisc * , unsigned long , s#netlink_ext_ack * ) ; unsigned long ( * bind_tcf ) ( s#Qdisc * , unsigned long , t#u32 ) ; void ( * unbind_tcf ) ( s#Qdisc * , unsigned long ) ; int ( * dump ) ( s#Qdisc * , unsigned long , s#sk_buff * , s#tcmsg * ) ; int ( * dump_stats ) ( s#Qdisc * , unsigned long , s#gnet_dump * ) ; }
+s#Qdisc_ops struct Qdisc_ops { s#Qdisc_ops * next ; const s#Qdisc_class_ops * cl_ops ; char id [ 16 ] ; int priv_size ; unsigned int static_flags ; int ( * enqueue ) ( s#sk_buff * , s#Qdisc * , s#sk_buff * * ) ; s#sk_buff * ( * dequeue ) ( s#Qdisc * ) ; s#sk_buff * ( * peek ) ( s#Qdisc * ) ; int ( * init ) ( s#Qdisc * , s#nlattr * , s#netlink_ext_ack * ) ; void ( * reset ) ( s#Qdisc * ) ; void ( * destroy ) ( s#Qdisc * ) ; int ( * change ) ( s#Qdisc * , s#nlattr * , s#netlink_ext_ack * ) ; void ( * attach ) ( s#Qdisc * ) ; int ( * change_tx_queue_len ) ( s#Qdisc * , unsigned int ) ; int ( * dump ) ( s#Qdisc * , s#sk_buff * ) ; int ( * dump_stats ) ( s#Qdisc * , s#gnet_dump * ) ; void ( * ingress_block_set ) ( s#Qdisc * , t#u32 ) ; void ( * egress_block_set ) ( s#Qdisc * , t#u32 ) ; t#u32 ( * ingress_block_get ) ( s#Qdisc * ) ; t#u32 ( * egress_block_get ) ( s#Qdisc * ) ; s#module * owner ; }
+s#tc_sizespec struct tc_sizespec { unsigned char cell_log ; unsigned char size_log ; short cell_align ; int overhead ; unsigned int linklayer ; unsigned int mpu ; unsigned int mtu ; unsigned int tsize ; }
+s#qdisc_size_table struct qdisc_size_table { s#callback_head rcu ; s#list_head list ; s#tc_sizespec szopts ; int refcnt ; t#u16 data [ ] ; }
+s##net_rate_estimator
+s#gnet_stats_basic_packed struct gnet_stats_basic_packed { t#__u64 bytes ; t#__u32 packets ; }
+s#gnet_stats_basic_cpu struct gnet_stats_basic_cpu { s#gnet_stats_basic_packed bstats ; s#u64_stats_sync syncp ; }
+s#qdisc_skb_head struct qdisc_skb_head { s#sk_buff * head ; s#sk_buff * tail ; union { t#u32 qlen ; t#atomic_t atomic_qlen ; } ; t#spinlock_t lock ; }
+s#Qdisc struct Qdisc { int ( * enqueue ) ( s#sk_buff * , s#Qdisc * , s#sk_buff * * ) ; s#sk_buff * ( * dequeue ) ( s#Qdisc * ) ; unsigned int flags ; t#u32 limit ; const s#Qdisc_ops * ops ; s#qdisc_size_table * stab ; s#hlist_node hash ; t#u32 handle ; t#u32 parent ; s#netdev_queue * dev_queue ; s#net_rate_estimator * rate_est ; s#gnet_stats_basic_cpu * cpu_bstats ; s#gnet_stats_queue * cpu_qstats ; s#sk_buff_head gso_skb __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; s#qdisc_skb_head q ; s#gnet_stats_basic_packed bstats ; t#seqcount_t running ; s#gnet_stats_queue qstats ; unsigned long state ; s#Qdisc * next_sched ; s#sk_buff_head skb_bad_txq ; int padded ; t#refcount_t refcnt ; t#spinlock_t busylock __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; t#spinlock_t seqlock ; s#callback_head rcu ; }
+s#tcf_block struct tcf_block { s#list_head chain_list ; t#u32 index ; t#refcount_t refcnt ; s#net * net ; s#Qdisc * q ; s#list_head cb_list ; s#list_head owner_list ; t#bool keep_dst ; unsigned int offloadcnt ; unsigned int nooffloaddevcnt ; struct { s#tcf_chain * chain ; s#list_head filter_chain_list ; } chain0 ; s#callback_head rcu ; }
+s#tcf_chain struct tcf_chain { s#tcf_proto * filter_chain ; s#list_head list ; s#tcf_block * block ; t#u32 index ; unsigned int refcnt ; unsigned int action_refcnt ; t#bool explicitly_created ; const s#tcf_proto_ops * tmplt_ops ; void * tmplt_priv ; }
+s#tcf_proto_ops struct tcf_proto_ops { s#list_head head ; char kind [ 16 ] ; int ( * classify ) ( s#sk_buff * , const s#tcf_proto * , s#tcf_result * ) ; int ( * init ) ( s#tcf_proto * ) ; void ( * destroy ) ( s#tcf_proto * , s#netlink_ext_ack * ) ; void * ( * get ) ( s#tcf_proto * , t#u32 ) ; int ( * change ) ( s#net * , s#sk_buff * , s#tcf_proto * , unsigned long , t#u32 , s#nlattr * * , void * * , t#bool , s#netlink_ext_ack * ) ; int ( * delete ) ( s#tcf_proto * , void * , t#bool * , s#netlink_ext_ack * ) ; void ( * walk ) ( s#tcf_proto * , s#tcf_walker * ) ; int ( * reoffload ) ( s#tcf_proto * , t#bool , t#tc_setup_cb_t * , void * , s#netlink_ext_ack * ) ; void ( * bind_class ) ( void * , t#u32 , unsigned long ) ; void * ( * tmplt_create ) ( s#net * , s#tcf_chain * , s#nlattr * * , s#netlink_ext_ack * ) ; void ( * tmplt_destroy ) ( void * ) ; int ( * dump ) ( s#net * , s#tcf_proto * , void * , s#sk_buff * , s#tcmsg * ) ; int ( * tmplt_dump ) ( s#sk_buff * , s#net * , void * ) ; s#module * owner ; }
+s#tcf_proto struct tcf_proto { s#tcf_proto * next ; void * root ; int ( * classify ) ( s#sk_buff * , const s#tcf_proto * , s#tcf_result * ) ; t#__be16 protocol ; t#u32 prio ; void * data ; const s#tcf_proto_ops * ops ; s#tcf_chain * chain ; s#callback_head rcu ; }
+s#mini_Qdisc struct mini_Qdisc { s#tcf_proto * filter_list ; s#gnet_stats_basic_cpu * cpu_bstats ; s#gnet_stats_queue * cpu_qstats ; s#callback_head rcu ; }
+s##nf_hook_entry
+s##cpu_rmap
+s#xps_map struct xps_map { unsigned int len ; unsigned int alloc_len ; s#callback_head rcu ; t#u16 queues [ 0 ] ; }
+s#xps_dev_maps struct xps_dev_maps { s#callback_head rcu ; s#xps_map * attr_map [ 0 ] ; }
+s#pcpu_lstats struct pcpu_lstats { t#u64 packets ; t#u64 bytes ; s#u64_stats_sync syncp ; }
+s#pcpu_sw_netstats struct pcpu_sw_netstats { t#u64 rx_packets ; t#u64 rx_bytes ; t#u64 tx_packets ; t#u64 tx_bytes ; s#u64_stats_sync syncp ; }
+s##pcpu_dstats
+s##garp_port
+s##mrp_port
+s#nla_policy struct nla_policy { t#u16 type ; t#u16 len ; void * validation_data ; }
+s#rtnl_link_ops struct rtnl_link_ops { s#list_head list ; const char * kind ; t#size_t priv_size ; void ( * setup ) ( s#net_device * ) ; int maxtype ; const s#nla_policy * policy ; int ( * validate ) ( s#nlattr * [ ] , s#nlattr * [ ] , s#netlink_ext_ack * ) ; int ( * newlink ) ( s#net * , s#net_device * , s#nlattr * [ ] , s#nlattr * [ ] , s#netlink_ext_ack * ) ; int ( * changelink ) ( s#net_device * , s#nlattr * [ ] , s#nlattr * [ ] , s#netlink_ext_ack * ) ; void ( * dellink ) ( s#net_device * , s#list_head * ) ; t#size_t ( * get_size ) ( const s#net_device * ) ; int ( * fill_info ) ( s#sk_buff * , const s#net_device * ) ; t#size_t ( * get_xstats_size ) ( const s#net_device * ) ; int ( * fill_xstats ) ( s#sk_buff * , const s#net_device * ) ; unsigned int ( * get_num_tx_queues ) ( void ) ; unsigned int ( * get_num_rx_queues ) ( void ) ; int slave_maxtype ; const s#nla_policy * slave_policy ; int ( * slave_validate ) ( s#nlattr * [ ] , s#nlattr * [ ] ) ; int ( * slave_changelink ) ( s#net_device * , s#net_device * , s#nlattr * [ ] , s#nlattr * [ ] ) ; t#size_t ( * get_slave_size ) ( const s#net_device * , const s#net_device * ) ; int ( * fill_slave_info ) ( s#sk_buff * , const s#net_device * , const s#net_device * ) ; s#net * ( * get_link_net ) ( const s#net_device * ) ; t#size_t ( * get_linkxstats_size ) ( const s#net_device * , int ) ; int ( * fill_linkxstats ) ( s#sk_buff * , const s#net_device * , int * , int ) ; }
+s#ieee_ets struct ieee_ets { t#__u8 willing ; t#__u8 ets_cap ; t#__u8 cbs ; t#__u8 tc_tx_bw [ 8 ] ; t#__u8 tc_rx_bw [ 8 ] ; t#__u8 tc_tsa [ 8 ] ; t#__u8 prio_tc [ 8 ] ; t#__u8 tc_reco_bw [ 8 ] ; t#__u8 tc_reco_tsa [ 8 ] ; t#__u8 reco_prio_tc [ 8 ] ; }
+s#ieee_maxrate struct ieee_maxrate { t#__u64 tc_maxrate [ 8 ] ; }
+s#ieee_qcn struct ieee_qcn { t#__u8 rpg_enable [ 8 ] ; t#__u32 rppp_max_rps [ 8 ] ; t#__u32 rpg_time_reset [ 8 ] ; t#__u32 rpg_byte_reset [ 8 ] ; t#__u32 rpg_threshold [ 8 ] ; t#__u32 rpg_max_rate [ 8 ] ; t#__u32 rpg_ai_rate [ 8 ] ; t#__u32 rpg_hai_rate [ 8 ] ; t#__u32 rpg_gd [ 8 ] ; t#__u32 rpg_min_dec_fac [ 8 ] ; t#__u32 rpg_min_rate [ 8 ] ; t#__u32 cndd_state_machine [ 8 ] ; }
+s#ieee_qcn_stats struct ieee_qcn_stats { t#__u64 rppp_rp_centiseconds [ 8 ] ; t#__u32 rppp_created_rps [ 8 ] ; }
+s#ieee_pfc struct ieee_pfc { t#__u8 pfc_cap ; t#__u8 pfc_en ; t#__u8 mbc ; t#__u16 delay ; t#__u64 requests [ 8 ] ; t#__u64 indications [ 8 ] ; }
+s#dcb_app struct dcb_app { t#__u8 selector ; t#__u8 priority ; t#__u16 protocol ; }
+s#dcb_peer_app_info struct dcb_peer_app_info { t#__u8 willing ; t#__u8 error ; }
+s#cee_pg struct cee_pg { t#__u8 willing ; t#__u8 error ; t#__u8 pg_en ; t#__u8 tcs_supported ; t#__u8 pg_bw [ 8 ] ; t#__u8 prio_pg [ 8 ] ; }
+s#cee_pfc struct cee_pfc { t#__u8 willing ; t#__u8 error ; t#__u8 pfc_en ; t#__u8 tcs_supported ; }
+s#dcbnl_buffer struct dcbnl_buffer { t#__u8 prio2buffer [ 8 ] ; t#__u32 buffer_size [ 8 ] ; t#__u32 total_size ; }
+s#dcbnl_rtnl_ops struct dcbnl_rtnl_ops { int ( * ieee_getets ) ( s#net_device * , s#ieee_ets * ) ; int ( * ieee_setets ) ( s#net_device * , s#ieee_ets * ) ; int ( * ieee_getmaxrate ) ( s#net_device * , s#ieee_maxrate * ) ; int ( * ieee_setmaxrate ) ( s#net_device * , s#ieee_maxrate * ) ; int ( * ieee_getqcn ) ( s#net_device * , s#ieee_qcn * ) ; int ( * ieee_setqcn ) ( s#net_device * , s#ieee_qcn * ) ; int ( * ieee_getqcnstats ) ( s#net_device * , s#ieee_qcn_stats * ) ; int ( * ieee_getpfc ) ( s#net_device * , s#ieee_pfc * ) ; int ( * ieee_setpfc ) ( s#net_device * , s#ieee_pfc * ) ; int ( * ieee_getapp ) ( s#net_device * , s#dcb_app * ) ; int ( * ieee_setapp ) ( s#net_device * , s#dcb_app * ) ; int ( * ieee_delapp ) ( s#net_device * , s#dcb_app * ) ; int ( * ieee_peer_getets ) ( s#net_device * , s#ieee_ets * ) ; int ( * ieee_peer_getpfc ) ( s#net_device * , s#ieee_pfc * ) ; t#u8 ( * getstate ) ( s#net_device * ) ; t#u8 ( * setstate ) ( s#net_device * , t#u8 ) ; void ( * getpermhwaddr ) ( s#net_device * , t#u8 * ) ; void ( * setpgtccfgtx ) ( s#net_device * , int , t#u8 , t#u8 , t#u8 , t#u8 ) ; void ( * setpgbwgcfgtx ) ( s#net_device * , int , t#u8 ) ; void ( * setpgtccfgrx ) ( s#net_device * , int , t#u8 , t#u8 , t#u8 , t#u8 ) ; void ( * setpgbwgcfgrx ) ( s#net_device * , int , t#u8 ) ; void ( * getpgtccfgtx ) ( s#net_device * , int , t#u8 * , t#u8 * , t#u8 * , t#u8 * ) ; void ( * getpgbwgcfgtx ) ( s#net_device * , int , t#u8 * ) ; void ( * getpgtccfgrx ) ( s#net_device * , int , t#u8 * , t#u8 * , t#u8 * , t#u8 * ) ; void ( * getpgbwgcfgrx ) ( s#net_device * , int , t#u8 * ) ; void ( * setpfccfg ) ( s#net_device * , int , t#u8 ) ; void ( * getpfccfg ) ( s#net_device * , int , t#u8 * ) ; t#u8 ( * setall ) ( s#net_device * ) ; t#u8 ( * getcap ) ( s#net_device * , int , t#u8 * ) ; int ( * getnumtcs ) ( s#net_device * , int , t#u8 * ) ; int ( * setnumtcs ) ( s#net_device * , int , t#u8 ) ; t#u8 ( * getpfcstate ) ( s#net_device * ) ; void ( * setpfcstate ) ( s#net_device * , t#u8 ) ; void ( * getbcncfg ) ( s#net_device * , int , t#u32 * ) ; void ( * setbcncfg ) ( s#net_device * , int , t#u32 ) ; void ( * getbcnrp ) ( s#net_device * , int , t#u8 * ) ; void ( * setbcnrp ) ( s#net_device * , int , t#u8 ) ; int ( * setapp ) ( s#net_device * , t#u8 , t#u16 , t#u8 ) ; int ( * getapp ) ( s#net_device * , t#u8 , t#u16 ) ; t#u8 ( * getfeatcfg ) ( s#net_device * , int , t#u8 * ) ; t#u8 ( * setfeatcfg ) ( s#net_device * , int , t#u8 ) ; t#u8 ( * getdcbx ) ( s#net_device * ) ; t#u8 ( * setdcbx ) ( s#net_device * , t#u8 ) ; int ( * peer_getappinfo ) ( s#net_device * , s#dcb_peer_app_info * , t#u16 * ) ; int ( * peer_getapptable ) ( s#net_device * , s#dcb_app * ) ; int ( * cee_peer_getpg ) ( s#net_device * , s#cee_pg * ) ; int ( * cee_peer_getpfc ) ( s#net_device * , s#cee_pfc * ) ; int ( * dcbnl_getbuffer ) ( s#net_device * , s#dcbnl_buffer * ) ; int ( * dcbnl_setbuffer ) ( s#net_device * , s#dcbnl_buffer * ) ; }
+t#s16 typedef signed short s16
+s#netdev_tc_txq struct netdev_tc_txq { t#u16 count ; t#u16 offset ; }
+s#netprio_map struct netprio_map { s#callback_head rcu ; t#u32 priomap_len ; t#u32 priomap [ ] ; }
+s##phy_device
+s##sfp_bus
+s#net_device struct net_device { char name [ 16 ] ; s#hlist_node name_hlist ; char * ifalias ; unsigned long mem_end ; unsigned long mem_start ; unsigned long base_addr ; int irq ; t#atomic_t carrier_changes ; unsigned long state ; s#list_head dev_list ; s#list_head napi_list ; s#list_head unreg_list ; s#list_head close_list ; s#list_head ptype_all ; s#list_head ptype_specific ; struct { s#list_head upper ; s#list_head lower ; } adj_list ; t#netdev_features_t features ; t#netdev_features_t hw_features ; t#netdev_features_t wanted_features ; t#netdev_features_t vlan_features ; t#netdev_features_t hw_enc_features ; t#netdev_features_t mpls_features ; t#netdev_features_t gso_partial_features ; int ifindex ; int group ; s#net_device_stats stats ; t#atomic_long_t rx_dropped ; t#atomic_long_t tx_dropped ; t#atomic_long_t rx_nohandler ; const s#iw_handler_def * wireless_handlers ; s#iw_public_data * wireless_data ; const s#net_device_ops * netdev_ops ; const s#ethtool_ops * ethtool_ops ; const s#switchdev_ops * switchdev_ops ; const s#l3mdev_ops * l3mdev_ops ; const s#ndisc_ops * ndisc_ops ; const s#xfrmdev_ops * xfrmdev_ops ; const s#tlsdev_ops * tlsdev_ops ; const s#header_ops * header_ops ; unsigned int flags ; unsigned int priv_flags ; unsigned short gflags ; unsigned short padded ; unsigned char operstate ; unsigned char link_mode ; unsigned char if_port ; unsigned char dma ; unsigned int mtu ; unsigned int min_mtu ; unsigned int max_mtu ; unsigned short type ; unsigned short hard_header_len ; unsigned char min_header_len ; unsigned short needed_headroom ; unsigned short needed_tailroom ; unsigned char perm_addr [ 32 ] ; unsigned char addr_assign_type ; unsigned char addr_len ; unsigned short neigh_priv_len ; unsigned short dev_id ; unsigned short dev_port ; t#spinlock_t addr_list_lock ; unsigned char name_assign_type ; t#bool uc_promisc ; s#netdev_hw_addr_list uc ; s#netdev_hw_addr_list mc ; s#netdev_hw_addr_list dev_addrs ; s#kset * queues_kset ; unsigned int promiscuity ; unsigned int allmulti ; s#vlan_info * vlan_info ; s#dsa_switch_tree * dsa_ptr ; void * atalk_ptr ; s#in_device * ip_ptr ; s#dn_dev * dn_ptr ; s#inet6_dev * ip6_ptr ; void * ax25_ptr ; s#wireless_dev * ieee80211_ptr ; s#wpan_dev * ieee802154_ptr ; s#mpls_dev * mpls_ptr ; unsigned char * dev_addr ; s#netdev_rx_queue * _rx ; unsigned int num_rx_queues ; unsigned int real_num_rx_queues ; s#bpf_prog * xdp_prog ; unsigned long gro_flush_timeout ; t#rx_handler_func_t * rx_handler ; void * rx_handler_data ; s#mini_Qdisc * miniq_ingress ; s#netdev_queue * ingress_queue ; s#nf_hook_entry * nf_hooks_ingress ; unsigned char broadcast [ 32 ] ; s#cpu_rmap * rx_cpu_rmap ; s#hlist_node index_hlist ; s#netdev_queue * _tx __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; unsigned int num_tx_queues ; unsigned int real_num_tx_queues ; s#Qdisc * qdisc ; s#hlist_head qdisc_hash [ 1 << ( 4 ) ] ; unsigned int tx_queue_len ; t#spinlock_t tx_global_lock ; int watchdog_timeo ; s#xps_dev_maps * xps_cpus_map ; s#xps_dev_maps * xps_rxqs_map ; s#mini_Qdisc * miniq_egress ; s#timer_list watchdog_timer ; int * pcpu_refcnt ; s#list_head todo_list ; s#list_head link_watch_list ; enum { NETREG_UNINITIALIZED = 0 , NETREG_REGISTERED , NETREG_UNREGISTERING , NETREG_UNREGISTERED , NETREG_RELEASED , NETREG_DUMMY , } reg_state : 8 ; t#bool dismantle ; enum { RTNL_LINK_INITIALIZED , RTNL_LINK_INITIALIZING , } rtnl_link_state : 16 ; t#bool needs_free_netdev ; void ( * priv_destructor ) ( s#net_device * ) ; s#netpoll_info * npinfo ; t#possible_net_t nd_net ; union { void * ml_priv ; s#pcpu_lstats * lstats ; s#pcpu_sw_netstats * tstats ; s#pcpu_dstats * dstats ; } ; s#garp_port * garp_port ; s#mrp_port * mrp_port ; s#device dev ; const s#attribute_group * sysfs_groups [ 4 ] ; const s#attribute_group * sysfs_rx_queue_group ; const s#rtnl_link_ops * rtnl_link_ops ; unsigned int gso_max_size ; t#u16 gso_max_segs ; const s#dcbnl_rtnl_ops * dcbnl_ops ; t#s16 num_tc ; s#netdev_tc_txq tc_to_txq [ 16 ] ; t#u8 prio_tc_map [ 15 + 1 ] ; unsigned int fcoe_ddp_xid ; s#netprio_map * priomap ; s#phy_device * phydev ; s#sfp_bus * sfp_bus ; s#lock_class_key * qdisc_tx_busylock ; s#lock_class_key * qdisc_running_key ; t#bool proto_down ; unsigned wol_enabled : 1 ; }
+s##prot_inuse
+s#netns_core struct netns_core { s#ctl_table_header * sysctl_hdr ; int sysctl_somaxconn ; s#prot_inuse * inuse ; }
+E#__TCP_MIB_MAX ( 0 ) + 16
+s#tcp_mib struct tcp_mib { unsigned long mibs [ E#__TCP_MIB_MAX ] ; }
+E#__LINUX_MIB_MAX ( 0 ) + 113
+s#linux_mib struct linux_mib { unsigned long mibs [ E#__LINUX_MIB_MAX ] ; }
+E#__UDP_MIB_MAX ( 0 ) + 9
+s#udp_mib struct udp_mib { unsigned long mibs [ E#__UDP_MIB_MAX ] ; }
+E#__ICMP_MIB_MAX ( 0 ) + 28
+s#icmp_mib struct icmp_mib { unsigned long mibs [ E#__ICMP_MIB_MAX ] ; }
+s#icmpmsg_mib struct icmpmsg_mib { t#atomic_long_t mibs [ 512 ] ; }
+s#icmpv6_mib struct icmpv6_mib { unsigned long mibs [ E#__ICMP6_MIB_MAX ] ; }
+s#icmpv6msg_mib struct icmpv6msg_mib { t#atomic_long_t mibs [ 512 ] ; }
+s#netns_mib struct netns_mib { __typeof__ ( s#tcp_mib ) * tcp_statistics ; __typeof__ ( s#ipstats_mib ) * ip_statistics ; __typeof__ ( s#linux_mib ) * net_statistics ; __typeof__ ( s#udp_mib ) * udp_statistics ; __typeof__ ( s#udp_mib ) * udplite_statistics ; __typeof__ ( s#icmp_mib ) * icmp_statistics ; __typeof__ ( s#icmpmsg_mib ) * icmpmsg_statistics ; s#proc_dir_entry * proc_net_devsnmp6 ; __typeof__ ( s#udp_mib ) * udp_stats_in6 ; __typeof__ ( s#udp_mib ) * udplite_stats_in6 ; __typeof__ ( s#ipstats_mib ) * ipv6_statistics ; __typeof__ ( s#icmpv6_mib ) * icmpv6_statistics ; __typeof__ ( s#icmpv6msg_mib ) * icmpv6msg_statistics ; }
+s#netns_packet struct netns_packet { s#mutex sklist_lock ; s#hlist_head sklist ; }
+s#netns_unix struct netns_unix { int sysctl_max_dgram_qlen ; s#ctl_table_header * ctl ; }
+s##ipv4_devconf
+s#fib_kuid_range struct fib_kuid_range { t#kuid_t start ; t#kuid_t end ; }
+s#fib_rule struct fib_rule { s#list_head list ; int iifindex ; int oifindex ; t#u32 mark ; t#u32 mark_mask ; t#u32 flags ; t#u32 table ; t#u8 action ; t#u8 l3mdev ; t#u32 target ; t#__be64 tun_id ; s#fib_rule * ctarget ; s#net * fr_net ; t#atomic_t refcnt ; t#u32 pref ; int suppress_ifgroup ; int suppress_prefixlen ; char iifname [ 16 ] ; char oifname [ 16 ] ; s#fib_kuid_range uid_range ; s#callback_head rcu ; }
+s#flowi4 struct flowi4 { s#flowi_common __fl_common ; t#__be32 saddr ; t#__be32 daddr ; u#flowi_uli uli ; }
+s#flowidn struct flowidn { s#flowi_common __fl_common ; t#__le16 daddr ; t#__le16 saddr ; u#flowi_uli uli ; }
+s#flowi struct flowi { union { s#flowi_common __fl_common ; s#flowi4 ip4 ; s#flowi6 ip6 ; s#flowidn dn ; } u ; }
+s#fib_lookup_arg struct fib_lookup_arg { void * lookup_ptr ; void * result ; s#fib_rule * rule ; t#u32 table ; int flags ; }
+s#fib_rule_hdr struct fib_rule_hdr { t#__u8 family ; t#__u8 dst_len ; t#__u8 src_len ; t#__u8 tos ; t#__u8 table ; t#__u8 res1 ; t#__u8 res2 ; t#__u8 action ; t#__u32 flags ; }
+s#fib_rules_ops struct fib_rules_ops { int family ; s#list_head list ; int rule_size ; int addr_size ; int unresolved_rules ; int nr_goto_rules ; unsigned int fib_rules_seq ; int ( * action ) ( s#fib_rule * , s#flowi * , int , s#fib_lookup_arg * ) ; t#bool ( * suppress ) ( s#fib_rule * , s#fib_lookup_arg * ) ; int ( * match ) ( s#fib_rule * , s#flowi * , int ) ; int ( * configure ) ( s#fib_rule * , s#sk_buff * , s#fib_rule_hdr * , s#nlattr * * ) ; int ( * delete ) ( s#fib_rule * ) ; int ( * compare ) ( s#fib_rule * , s#fib_rule_hdr * , s#nlattr * * ) ; int ( * fill ) ( s#fib_rule * , s#sk_buff * , s#fib_rule_hdr * ) ; t#size_t ( * nlmsg_payload ) ( s#fib_rule * ) ; void ( * flush_cache ) ( s#fib_rules_ops * ) ; int nlgroup ; const s#nla_policy * policy ; s#list_head rules_list ; s#module * owner ; s#net * fro_net ; s#callback_head rcu ; }
+s#fib_table struct fib_table { s#hlist_node tb_hlist ; t#u32 tb_id ; int tb_num_default ; s#callback_head rcu ; unsigned long * tb_data ; unsigned long __data [ 0 ] ; }
+s#inet_peer_base struct inet_peer_base { s#rb_root rb_root ; t#seqlock_t lock ; int total ; }
+s#rhash_head struct rhash_head { s#rhash_head * next ; }
+s#frag_v4_compare_key struct frag_v4_compare_key { t#__be32 saddr ; t#__be32 daddr ; t#u32 user ; t#u32 vif ; t#__be16 id ; t#u16 protocol ; }
+s#frag_v6_compare_key struct frag_v6_compare_key { s#in6_addr saddr ; s#in6_addr daddr ; t#u32 user ; t#__be32 id ; t#u32 iif ; }
+s#inet_frag_queue struct inet_frag_queue { s#rhash_head node ; union { s#frag_v4_compare_key v4 ; s#frag_v6_compare_key v6 ; } key ; s#timer_list timer ; t#spinlock_t lock ; t#refcount_t refcnt ; s#rb_root rb_fragments ; s#sk_buff * fragments_tail ; s#sk_buff * last_run_head ; t#ktime_t stamp ; int len ; int meat ; t#__u8 flags ; t#u16 max_size ; s#netns_frags * net ; s#callback_head rcu ; }
+t#rht_hashfn_t typedef t#u32 ( * rht_hashfn_t ) ( const void * , t#u32 , t#u32 )
+t#rht_obj_hashfn_t typedef t#u32 ( * rht_obj_hashfn_t ) ( const void * , t#u32 , t#u32 )
+s#bucket_table struct bucket_table { unsigned int size ; unsigned int nest ; unsigned int rehash ; t#u32 hash_rnd ; unsigned int locks_mask ; t#spinlock_t * locks ; s#list_head walkers ; s#callback_head rcu ; s#bucket_table * future_tbl ; s#rhash_head * buckets [ ] __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; }
+s#rhashtable struct rhashtable { s#bucket_table * tbl ; unsigned int key_len ; unsigned int max_elems ; s#rhashtable_params p ; t#bool rhlist ; s#work_struct run_work ; s#mutex mutex ; t#spinlock_t lock ; t#atomic_t nelems ; }
+s#rhashtable_compare_arg struct rhashtable_compare_arg { s#rhashtable * ht ; const void * key ; }
+t#rht_obj_cmpfn_t typedef int ( * rht_obj_cmpfn_t ) ( s#rhashtable_compare_arg * , const void * )
+s#rhashtable_params struct rhashtable_params { t#u16 nelem_hint ; t#u16 key_len ; t#u16 key_offset ; t#u16 head_offset ; unsigned int max_size ; t#u16 min_size ; t#bool automatic_shrinking ; t#u8 locks_mul ; t#u32 nulls_base ; t#rht_hashfn_t hashfn ; t#rht_obj_hashfn_t obj_hashfn ; t#rht_obj_cmpfn_t obj_cmpfn ; }
+s#inet_frags struct inet_frags { unsigned int qsize ; void ( * constructor ) ( s#inet_frag_queue * , const void * ) ; void ( * destructor ) ( s#inet_frag_queue * ) ; void ( * frag_expire ) ( unsigned long ) ; s#kmem_cache * frags_cachep ; const char * frags_cache_name ; s#rhashtable_params rhash_params ; }
+s#netns_frags struct netns_frags { long high_thresh ; long low_thresh ; int timeout ; int max_dist ; s#inet_frags * f ; s#rhashtable rhashtable __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; t#atomic_long_t mem __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; }
+s##xt_table
+s#local_ports struct local_ports { t#seqlock_t lock ; int range [ 2 ] ; t#bool warned ; }
+s##inet_hashinfo
+s#inet_timewait_death_row struct inet_timewait_death_row { t#atomic_t tw_count ; s#inet_hashinfo * hashinfo __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; int sysctl_max_tw_buckets ; }
+s#ping_group_range struct ping_group_range { t#seqlock_t lock ; t#kgid_t range [ 2 ] ; }
+t#notifier_fn_t typedef int ( * notifier_fn_t ) ( s#notifier_block * , unsigned long , void * )
+s#notifier_block struct notifier_block { t#notifier_fn_t notifier_call ; s#notifier_block * next ; int priority ; }
+s#fib_notifier_ops struct fib_notifier_ops { int family ; s#list_head list ; unsigned int ( * fib_seq_read ) ( s#net * ) ; int ( * fib_dump ) ( s#net * , s#notifier_block * ) ; s#module * owner ; s#callback_head rcu ; }
+t#siphash_key_t typedef struct { t#u64 key [ 2 ] ; } siphash_key_t
+s#netns_ipv4 struct netns_ipv4 { s#ctl_table_header * forw_hdr ; s#ctl_table_header * frags_hdr ; s#ctl_table_header * ipv4_hdr ; s#ctl_table_header * route_hdr ; s#ctl_table_header * xfrm4_hdr ; s#ipv4_devconf * devconf_all ; s#ipv4_devconf * devconf_dflt ; s#fib_rules_ops * rules_ops ; t#bool fib_has_custom_rules ; s#fib_table * fib_main ; s#fib_table * fib_default ; int fib_num_tclassid_users ; s#hlist_head * fib_table_hash ; t#bool fib_offload_disabled ; s#sock * fibnl ; s#sock * * icmp_sk ; s#sock * mc_autojoin_sk ; s#inet_peer_base * peers ; s#sock * * tcp_sk ; s#netns_frags frags ; s#xt_table * iptable_filter ; s#xt_table * iptable_mangle ; s#xt_table * iptable_raw ; s#xt_table * arptable_filter ; s#xt_table * iptable_security ; s#xt_table * nat_table ; int sysctl_icmp_echo_ignore_all ; int sysctl_icmp_echo_ignore_broadcasts ; int sysctl_icmp_ignore_bogus_error_responses ; int sysctl_icmp_ratelimit ; int sysctl_icmp_ratemask ; int sysctl_icmp_errors_use_inbound_ifaddr ; s#local_ports ip_local_ports ; int sysctl_tcp_ecn ; int sysctl_tcp_ecn_fallback ; int sysctl_ip_default_ttl ; int sysctl_ip_no_pmtu_disc ; int sysctl_ip_fwd_use_pmtu ; int sysctl_ip_nonlocal_bind ; int sysctl_ip_dynaddr ; int sysctl_ip_early_demux ; int sysctl_tcp_early_demux ; int sysctl_udp_early_demux ; int sysctl_fwmark_reflect ; int sysctl_tcp_fwmark_accept ; int sysctl_tcp_l3mdev_accept ; int sysctl_tcp_mtu_probing ; int sysctl_tcp_base_mss ; int sysctl_tcp_min_snd_mss ; int sysctl_tcp_probe_threshold ; t#u32 sysctl_tcp_probe_interval ; int sysctl_tcp_keepalive_time ; int sysctl_tcp_keepalive_probes ; int sysctl_tcp_keepalive_intvl ; int sysctl_tcp_syn_retries ; int sysctl_tcp_synack_retries ; int sysctl_tcp_syncookies ; int sysctl_tcp_reordering ; int sysctl_tcp_retries1 ; int sysctl_tcp_retries2 ; int sysctl_tcp_orphan_retries ; int sysctl_tcp_fin_timeout ; unsigned int sysctl_tcp_notsent_lowat ; int sysctl_tcp_tw_reuse ; s#inet_timewait_death_row tcp_death_row ; int sysctl_max_syn_backlog ; int sysctl_udp_l3mdev_accept ; int sysctl_igmp_max_memberships ; int sysctl_igmp_max_msf ; int sysctl_igmp_llm_reports ; int sysctl_igmp_qrv ; s#ping_group_range ping_group_range ; t#atomic_t dev_addr_genid ; unsigned long * sysctl_local_reserved_ports ; int sysctl_ip_prot_sock ; s#list_head mr_tables ; s#fib_rules_ops * mr_rules_ops ; int sysctl_fib_multipath_use_neigh ; int sysctl_fib_multipath_hash_policy ; s#fib_notifier_ops * notifier_ops ; unsigned int fib_seq ; s#fib_notifier_ops * ipmr_notifier_ops ; unsigned int ipmr_seq ; t#atomic_t rt_genid ; t#siphash_key_t ip_id_key ; }
+s#netns_sysctl_ipv6 struct netns_sysctl_ipv6 { s#ctl_table_header * hdr ; s#ctl_table_header * route_hdr ; s#ctl_table_header * icmp_hdr ; s#ctl_table_header * frags_hdr ; s#ctl_table_header * xfrm6_hdr ; int bindv6only ; int flush_delay ; int ip6_rt_max_size ; int ip6_rt_gc_min_interval ; int ip6_rt_gc_timeout ; int ip6_rt_gc_interval ; int ip6_rt_gc_elasticity ; int ip6_rt_mtu_expires ; int ip6_rt_min_advmss ; int flowlabel_consistency ; int auto_flowlabels ; int icmpv6_time ; int anycast_src_echo_reply ; int ip_nonlocal_bind ; int fwmark_reflect ; int idgen_retries ; int idgen_delay ; int flowlabel_state_ranges ; }
+s##rt6_statistics
+s##fib6_table
+s##seg6_pernet_data
+s#netns_ipv6 struct netns_ipv6 { s#netns_sysctl_ipv6 sysctl ; s#ipv6_devconf * devconf_all ; s#ipv6_devconf * devconf_dflt ; s#inet_peer_base * peers ; s#netns_frags frags ; s#xt_table * ip6table_filter ; s#xt_table * ip6table_mangle ; s#xt_table * ip6table_raw ; s#xt_table * ip6table_security ; s#xt_table * ip6table_nat ; s#rt6_info * ip6_null_entry ; s#rt6_statistics * rt6_stats ; s#timer_list ip6_fib_timer ; s#hlist_head * fib_table_hash ; s#fib6_table * fib6_main_tbl ; s#list_head fib6_walkers ; s#dst_ops ip6_dst_ops ; t#rwlock_t fib6_walker_lock ; t#spinlock_t fib6_gc_lock ; unsigned int ip6_rt_gc_expire ; unsigned long ip6_rt_last_gc ; s#rt6_info * ip6_prohibit_entry ; s#rt6_info * ip6_blk_hole_entry ; s#fib6_table * fib6_local_tbl ; s#fib_rules_ops * fib6_rules_ops ; s#sock * * icmp_sk ; s#sock * ndisc_sk ; s#sock * tcp_sk ; s#sock * igmp_sk ; s#sock * mc_autojoin_sk ; s#list_head mr6_tables ; s#fib_rules_ops * mr6_rules_ops ; t#atomic_t dev_addr_genid ; t#atomic_t fib6_sernum ; s#seg6_pernet_data * seg6_data ; s#fib_notifier_ops * notifier_ops ; struct { s#hlist_head head ; t#spinlock_t lock ; t#u32 seq ; } ip6addrlbl_table ; }
+s##sctp_mib
+s#netns_sctp struct netns_sctp { __typeof__ ( s#sctp_mib ) * sctp_statistics ; s#proc_dir_entry * proc_net_sctp ; s#ctl_table_header * sysctl_header ; s#sock * ctl_sock ; s#list_head local_addr_list ; s#list_head addr_waitq ; s#timer_list addr_wq_timer ; s#list_head auto_asconf_splist ; t#spinlock_t addr_wq_lock ; t#spinlock_t local_addr_lock ; unsigned int rto_initial ; unsigned int rto_min ; unsigned int rto_max ; int rto_alpha ; int rto_beta ; int max_burst ; int cookie_preserve_enable ; char * sctp_hmac_alg ; unsigned int valid_cookie_life ; unsigned int sack_timeout ; unsigned int hb_interval ; int max_retrans_association ; int max_retrans_path ; int max_retrans_init ; int pf_retrans ; int pf_enable ; int sndbuf_policy ; int rcvbuf_policy ; int default_auto_asconf ; int addip_enable ; int addip_noauth ; int prsctp_enable ; int reconf_enable ; int auth_enable ; int scope_policy ; int rwnd_upd_shift ; unsigned long max_autoclose ; }
+s##nf_queue_handler
+s##nf_logger
+E#NFPROTO_NUMPROTO ( 12 ) + 1
+s#netns_nf struct netns_nf { s#proc_dir_entry * proc_netfilter ; const s#nf_queue_handler * queue_handler ; const s#nf_logger * nf_loggers [ E#NFPROTO_NUMPROTO ] ; s#ctl_table_header * nf_log_dir_header ; s#nf_hook_entry * hooks [ E#NFPROTO_NUMPROTO ] [ 8 ] ; t#bool defrag_ipv4 ; t#bool defrag_ipv6 ; }
+s##ebt_table
+s#netns_xt struct netns_xt { s#list_head tables [ E#NFPROTO_NUMPROTO ] ; t#bool notrack_deprecated_warning ; t#bool clusterip_deprecated_warning ; s#ebt_table * broute_table ; s#ebt_table * frame_filter ; s#ebt_table * frame_nat ; }
+s#hlist_nulls_node struct hlist_nulls_node { s#hlist_nulls_node * next , * * pprev ; }
+s#hlist_nulls_head struct hlist_nulls_head { s#hlist_nulls_node * first ; }
+s#ct_pcpu struct ct_pcpu { t#spinlock_t lock ; s#hlist_nulls_head unconfirmed ; s#hlist_nulls_head dying ; }
+s#ip_conntrack_stat struct ip_conntrack_stat { unsigned int found ; unsigned int invalid ; unsigned int ignore ; unsigned int insert ; unsigned int insert_failed ; unsigned int drop ; unsigned int early_drop ; unsigned int error ; unsigned int expect_new ; unsigned int expect_create ; unsigned int expect_delete ; unsigned int search_restart ; }
+s##nf_ct_event_notifier
+s##nf_exp_event_notifier
+s#nf_proto_net struct nf_proto_net { s#ctl_table_header * ctl_table_header ; s#ctl_table * ctl_table ; unsigned int users ; }
+s#nf_generic_net struct nf_generic_net { s#nf_proto_net pn ; unsigned int timeout ; }
+E#TCP_CONNTRACK_TIMEOUT_MAX 14
+s#nf_tcp_net struct nf_tcp_net { s#nf_proto_net pn ; unsigned int timeouts [ E#TCP_CONNTRACK_TIMEOUT_MAX ] ; unsigned int tcp_loose ; unsigned int tcp_be_liberal ; unsigned int tcp_max_retrans ; }
+E#UDP_CT_MAX 2
+s#nf_udp_net struct nf_udp_net { s#nf_proto_net pn ; unsigned int timeouts [ E#UDP_CT_MAX ] ; }
+s#nf_icmp_net struct nf_icmp_net { s#nf_proto_net pn ; unsigned int timeout ; }
+E#__CT_DCCP_MAX 10
+s#nf_dccp_net struct nf_dccp_net { s#nf_proto_net pn ; int dccp_loose ; unsigned int dccp_timeout [ ( E#__CT_DCCP_MAX - 1 ) + 1 ] ; }
+E#SCTP_CONNTRACK_MAX 10
+s#nf_sctp_net struct nf_sctp_net { s#nf_proto_net pn ; unsigned int timeouts [ E#SCTP_CONNTRACK_MAX ] ; }
+s#nf_ip_net struct nf_ip_net { s#nf_generic_net generic ; s#nf_tcp_net tcp ; s#nf_udp_net udp ; s#nf_icmp_net icmp ; s#nf_icmp_net icmpv6 ; s#nf_dccp_net dccp ; s#nf_sctp_net sctp ; }
+s#netns_ct struct netns_ct { t#atomic_t count ; unsigned int expect_count ; s#delayed_work ecache_dwork ; t#bool ecache_dwork_pending ; s#ctl_table_header * sysctl_header ; s#ctl_table_header * acct_sysctl_header ; s#ctl_table_header * tstamp_sysctl_header ; s#ctl_table_header * event_sysctl_header ; s#ctl_table_header * helper_sysctl_header ; unsigned int sysctl_log_invalid ; int sysctl_events ; int sysctl_acct ; int sysctl_auto_assign_helper ; t#bool auto_assign_helper_warned ; int sysctl_tstamp ; int sysctl_checksum ; s#ct_pcpu * pcpu_lists ; s#ip_conntrack_stat * stat ; s#nf_ct_event_notifier * nf_conntrack_event_cb ; s#nf_exp_event_notifier * nf_expect_event_cb ; s#nf_ip_net nf_ct_proto ; unsigned int labels_used ; }
+s##nft_af_info
+s#netns_nftables struct netns_nftables { s#list_head af_info ; s#list_head commit_list ; s#nft_af_info * ipv4 ; s#nft_af_info * ipv6 ; s#nft_af_info * inet ; s#nft_af_info * arp ; s#nft_af_info * bridge ; s#nft_af_info * netdev ; unsigned int base_seq ; t#u8 gencursor ; }
+s#netns_nf_frag struct netns_nf_frag { s#netns_frags frags ; }
+s##net_generic
+E#XFRM_POLICY_MAX 3
+s#xfrm_policy_hash struct xfrm_policy_hash { s#hlist_head * table ; unsigned int hmask ; t#u8 dbits4 ; t#u8 sbits4 ; t#u8 dbits6 ; t#u8 sbits6 ; }
+s#xfrm_policy_hthresh struct xfrm_policy_hthresh { s#work_struct work ; t#seqlock_t lock ; t#u8 lbits4 ; t#u8 rbits4 ; t#u8 lbits6 ; t#u8 rbits6 ; }
+s#tasklet_struct struct tasklet_struct { s#tasklet_struct * next ; unsigned long state ; t#atomic_t count ; void ( * func ) ( unsigned long ) ; unsigned long data ; }
+s#flow_cache_percpu struct flow_cache_percpu { s#hlist_head * hash_table ; unsigned int hash_count ; t#u32 hash_rnd ; int hash_rnd_recalc ; s#tasklet_struct flush_tasklet ; }
+s#flow_cache struct flow_cache { t#u32 hash_shift ; s#flow_cache_percpu * percpu ; s#hlist_node node ; unsigned int low_watermark ; unsigned int high_watermark ; s#timer_list rnd_timer ; }
+s#netns_xfrm struct netns_xfrm { s#list_head state_all ; s#hlist_head * state_bydst ; s#hlist_head * state_bysrc ; s#hlist_head * state_byspi ; unsigned int state_hmask ; unsigned int state_num ; s#work_struct state_hash_work ; s#list_head policy_all ; s#hlist_head * policy_byidx ; unsigned int policy_idx_hmask ; s#hlist_head policy_inexact [ E#XFRM_POLICY_MAX ] ; s#xfrm_policy_hash policy_bydst [ E#XFRM_POLICY_MAX ] ; unsigned int policy_count [ E#XFRM_POLICY_MAX * 2 ] ; s#work_struct policy_hash_work ; s#xfrm_policy_hthresh policy_hthresh ; s#sock * nlsk ; s#sock * nlsk_stash ; t#u32 sysctl_aevent_etime ; t#u32 sysctl_aevent_rseqth ; int sysctl_larval_drop ; t#u32 sysctl_acq_expires ; s#ctl_table_header * sysctl_hdr ; s#dst_ops xfrm4_dst_ops ; s#dst_ops xfrm6_dst_ops ; t#spinlock_t xfrm_state_lock ; t#spinlock_t xfrm_policy_lock ; s#mutex xfrm_cfg_mutex ; s#flow_cache flow_cache_global ; t#atomic_t flow_cache_genid ; s#list_head flow_cache_gc_list ; t#atomic_t flow_cache_gc_count ; t#spinlock_t flow_cache_gc_lock ; s#work_struct flow_cache_gc_work ; s#work_struct flow_cache_flush_work ; s#mutex flow_flush_sem ; }
+s##netns_ipvs
+s##mpls_route
+s#netns_mpls struct netns_mpls { int ip_ttl_propagate ; int default_ttl ; t#size_t platform_labels ; s#mpls_route * * platform_label ; s#ctl_table_header * ctl ; }
+s##dev_rcv_lists
+s##s_stats
+s##s_pstats
+s#netns_can struct netns_can { s#proc_dir_entry * proc_dir ; s#proc_dir_entry * pde_version ; s#proc_dir_entry * pde_stats ; s#proc_dir_entry * pde_reset_stats ; s#proc_dir_entry * pde_rcvlist_all ; s#proc_dir_entry * pde_rcvlist_fil ; s#proc_dir_entry * pde_rcvlist_inv ; s#proc_dir_entry * pde_rcvlist_sff ; s#proc_dir_entry * pde_rcvlist_eff ; s#proc_dir_entry * pde_rcvlist_err ; s#proc_dir_entry * bcmproc_dir ; s#dev_rcv_lists * can_rx_alldev_list ; t#spinlock_t can_rcvlists_lock ; s#timer_list can_stattimer ; s#s_stats * can_stats ; s#s_pstats * can_pstats ; s#hlist_head cgw_list ; }
+s#net struct net { t#atomic_t passive ; t#atomic_t count ; t#spinlock_t rules_mod_lock ; t#u32 hash_mix ; t#atomic64_t cookie_gen ; s#list_head list ; s#list_head cleanup_list ; s#list_head exit_list ; s#user_namespace * user_ns ; s#ucounts * ucounts ; t#spinlock_t nsid_lock ; s#idr netns_ids ; s#ns_common ns ; s#proc_dir_entry * proc_net ; s#proc_dir_entry * proc_net_stat ; s#ctl_table_set sysctls ; s#sock * rtnl ; s#sock * genl_sock ; s#uevent_sock * uevent_sock ; s#list_head dev_base_head ; s#hlist_head * dev_name_head ; s#hlist_head * dev_index_head ; unsigned int dev_base_seq ; int ifindex ; unsigned int dev_unreg_count ; s#list_head rules_ops ; s#list_head fib_notifier_ops ; s#net_device * loopback_dev ; s#netns_core core ; s#netns_mib mib ; s#netns_packet packet ; s#netns_unix unx ; s#netns_ipv4 ipv4 ; s#netns_ipv6 ipv6 ; s#netns_sctp sctp ; s#netns_nf nf ; s#netns_xt xt ; s#netns_ct ct ; s#netns_nftables nft ; s#netns_nf_frag nf_frag ; s#ctl_table_header * nf_frag_frags_hdr ; s#sock * nfnl ; s#sock * nfnl_stash ; s#list_head nfnl_acct_list ; s#list_head nfct_timeout_list ; s#sk_buff_head wext_nlevents ; s#net_generic * gen ; s#netns_xfrm xfrm ; s#netns_ipvs * ipvs ; s#netns_mpls mpls ; s#netns_can can ; s#sock * diag_nlsk ; t#atomic_t fnhe_genid ; }
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s#pipe_buf_operations struct pipe_buf_operations { int can_merge ; int ( * confirm ) ( s#pipe_inode_info * , s#pipe_buffer * ) ; void ( * release ) ( s#pipe_inode_info * , s#pipe_buffer * ) ; int ( * steal ) ( s#pipe_inode_info * , s#pipe_buffer * ) ; t#bool ( * get ) ( s#pipe_inode_info * , s#pipe_buffer * ) ; }
+s#pipe_buffer struct pipe_buffer { s#page * page ; unsigned int offset , len ; const s#pipe_buf_operations * ops ; unsigned int flags ; unsigned long private ; }
+s#pipe_inode_info struct pipe_inode_info { s#mutex mutex ; t#wait_queue_head_t wait ; unsigned int nrbufs , curbuf , buffers ; unsigned int readers ; unsigned int writers ; unsigned int files ; unsigned int waiting_writers ; unsigned int r_counter ; unsigned int w_counter ; s#page * tmp_page ; s#fasync_struct * fasync_readers ; s#fasync_struct * fasync_writers ; s#pipe_buffer * bufs ; s#user_struct * user ; }
+s##task_delay_info
+s##ftrace_ret_stack
+s##eventfd_ctx
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s#mmu_notifier_mm struct mmu_notifier_mm { s#hlist_head list ; t#spinlock_t lock ; }
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##nlm_lockowner
+s##nfs4_lock_state
+s##kernfs_open_node
+s##kernfs_iattrs
+s##bpf_prog_array
+s##cgroup_taskset
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s#msghdr struct msghdr { void * msg_name ; int msg_namelen ; s#iov_iter msg_iter ; void * msg_control ; t#__kernel_size_t msg_controllen ; unsigned int msg_flags ; s#kiocb * msg_iocb ; }
+s#request_sock struct request_sock { s#sock_common __req_common ; s#request_sock * dl_next ; t#u16 mss ; t#u8 num_retrans ; t#u8 cookie_ts : 1 ; t#u8 num_timeout : 7 ; t#u32 ts_recent ; s#timer_list rsk_timer ; const s#request_sock_ops * rsk_ops ; s#sock * sk ; t#u32 * saved_syn ; t#u32 secid ; t#u32 peer_secid ; }
+s#request_sock_ops struct request_sock_ops { int family ; int obj_size ; s#kmem_cache * slab ; char * slab_name ; int ( * rtx_syn_ack ) ( const s#sock * , s#request_sock * ) ; void ( * send_ack ) ( const s#sock * , s#sk_buff * , s#request_sock * ) ; void ( * send_reset ) ( const s#sock * , s#sk_buff * ) ; void ( * destructor ) ( s#request_sock * ) ; void ( * syn_ack_timeout ) ( const s#request_sock * ) ; }
+s#timewait_sock_ops struct timewait_sock_ops { s#kmem_cache * twsk_slab ; char * twsk_slab_name ; unsigned int twsk_obj_size ; int ( * twsk_unique ) ( s#sock * , s#sock * , void * ) ; void ( * twsk_destructor ) ( s#sock * ) ; }
+s##udp_table
+s##raw_hashinfo
+s#smc_hashinfo struct smc_hashinfo { t#rwlock_t lock ; s#hlist_head ht ; }
+s#proto struct proto { void ( * close ) ( s#sock * , long ) ; int ( * pre_connect ) ( s#sock * , s#sockaddr * , int ) ; int ( * connect ) ( s#sock * , s#sockaddr * , int ) ; int ( * disconnect ) ( s#sock * , int ) ; s#sock * ( * accept ) ( s#sock * , int , int * , t#bool ) ; int ( * ioctl ) ( s#sock * , int , unsigned long ) ; int ( * init ) ( s#sock * ) ; void ( * destroy ) ( s#sock * ) ; void ( * shutdown ) ( s#sock * , int ) ; int ( * setsockopt ) ( s#sock * , int , int , char * , unsigned int ) ; int ( * getsockopt ) ( s#sock * , int , int , char * , int * ) ; void ( * keepalive ) ( s#sock * , int ) ; int ( * compat_setsockopt ) ( s#sock * , int , int , char * , unsigned int ) ; int ( * compat_getsockopt ) ( s#sock * , int , int , char * , int * ) ; int ( * compat_ioctl ) ( s#sock * , unsigned int , unsigned long ) ; int ( * sendmsg ) ( s#sock * , s#msghdr * , t#size_t ) ; int ( * recvmsg ) ( s#sock * , s#msghdr * , t#size_t , int , int , int * ) ; int ( * sendpage ) ( s#sock * , s#page * , int , t#size_t , int ) ; int ( * bind ) ( s#sock * , s#sockaddr * , int ) ; int ( * backlog_rcv ) ( s#sock * , s#sk_buff * ) ; void ( * release_cb ) ( s#sock * ) ; int ( * hash ) ( s#sock * ) ; void ( * unhash ) ( s#sock * ) ; void ( * rehash ) ( s#sock * ) ; int ( * get_port ) ( s#sock * , unsigned short ) ; unsigned int inuse_idx ; t#bool ( * stream_memory_free ) ( const s#sock * ) ; t#bool ( * stream_memory_read ) ( const s#sock * ) ; void ( * enter_memory_pressure ) ( s#sock * ) ; t#atomic_long_t * memory_allocated ; s#percpu_counter * sockets_allocated ; int * memory_pressure ; long * sysctl_mem ; int * sysctl_wmem ; int * sysctl_rmem ; int max_header ; t#bool no_autobind ; s#kmem_cache * slab ; unsigned int obj_size ; int slab_flags ; s#percpu_counter * orphan_count ; s#request_sock_ops * rsk_prot ; s#timewait_sock_ops * twsk_prot ; union { s#inet_hashinfo * hashinfo ; s#udp_table * udp_table ; s#raw_hashinfo * raw_hash ; s#smc_hashinfo * smc_hash ; } h ; s#module * owner ; char name [ 32 ] ; s#list_head node ; int ( * diag_destroy ) ( s#sock * , int ) ; }
+s#sock_common struct sock_common { union { t#__addrpair skc_addrpair ; struct { t#__be32 skc_daddr ; t#__be32 skc_rcv_saddr ; } ; } ; union { unsigned int skc_hash ; t#__u16 skc_u16hashes [ 2 ] ; } ; union { t#__portpair skc_portpair ; struct { t#__be16 skc_dport ; t#__u16 skc_num ; } ; } ; unsigned short skc_family ; volatile unsigned char skc_state ; unsigned char skc_reuse : 4 ; unsigned char skc_reuseport : 1 ; unsigned char skc_ipv6only : 1 ; unsigned char skc_net_refcnt : 1 ; int skc_bound_dev_if ; union { s#hlist_node skc_bind_node ; s#hlist_node skc_portaddr_node ; } ; s#proto * skc_prot ; t#possible_net_t skc_net ; s#in6_addr skc_v6_daddr ; s#in6_addr skc_v6_rcv_saddr ; t#atomic64_t skc_cookie ; union { unsigned long skc_flags ; s#sock * skc_listener ; s#inet_timewait_death_row * skc_tw_dr ; } ; int skc_dontcopy_begin [ 0 ] ; union { s#hlist_node skc_node ; s#hlist_nulls_node skc_nulls_node ; } ; unsigned short skc_tx_queue_mapping ; unsigned short skc_rx_queue_mapping ; union { int skc_incoming_cpu ; t#u32 skc_rcv_wnd ; t#u32 skc_tw_rcv_nxt ; } ; t#atomic_t skc_refcnt ; int skc_dontcopy_end [ 0 ] ; union { t#u32 skc_rxhash ; t#u32 skc_window_clamp ; t#u32 skc_tw_snd_nxt ; } ; }
+t#socket_lock_t typedef struct { t#spinlock_t slock ; int owned ; t#wait_queue_head_t wq ; } socket_lock_t
+s#sk_filter struct sk_filter { t#refcount_t refcnt ; s#callback_head rcu ; s#bpf_prog * prog ; }
+s#socket_wq struct socket_wq { t#wait_queue_head_t wait ; s#fasync_struct * fasync_list ; unsigned long flags ; s#callback_head rcu ; }
+s##xfrm_policy
+t#socket_state typedef enum { SS_FREE = 0 , SS_UNCONNECTED , SS_CONNECTING , SS_CONNECTED , SS_DISCONNECTING } socket_state
+t#read_descriptor_t typedef struct { t#size_t written ; t#size_t count ; union { char * buf ; void * data ; } arg ; int error ; } read_descriptor_t
+t#sk_read_actor_t typedef int ( * sk_read_actor_t ) ( t#read_descriptor_t * , s#sk_buff * , unsigned int , t#size_t )
+s#proto_ops struct proto_ops { int family ; s#module * owner ; int ( * release ) ( s#socket * ) ; int ( * bind ) ( s#socket * , s#sockaddr * , int ) ; int ( * connect ) ( s#socket * , s#sockaddr * , int , int ) ; int ( * socketpair ) ( s#socket * , s#socket * ) ; int ( * accept ) ( s#socket * , s#socket * , int , t#bool ) ; int ( * getname ) ( s#socket * , s#sockaddr * , int * , int ) ; unsigned int ( * poll ) ( s#file * , s#socket * , s#poll_table_struct * ) ; int ( * ioctl ) ( s#socket * , unsigned int , unsigned long ) ; int ( * compat_ioctl ) ( s#socket * , unsigned int , unsigned long ) ; int ( * listen ) ( s#socket * , int ) ; int ( * shutdown ) ( s#socket * , int ) ; int ( * setsockopt ) ( s#socket * , int , int , char * , unsigned int ) ; int ( * getsockopt ) ( s#socket * , int , int , char * , int * ) ; int ( * compat_setsockopt ) ( s#socket * , int , int , char * , unsigned int ) ; int ( * compat_getsockopt ) ( s#socket * , int , int , char * , int * ) ; int ( * sendmsg ) ( s#socket * , s#msghdr * , t#size_t ) ; int ( * recvmsg ) ( s#socket * , s#msghdr * , t#size_t , int ) ; int ( * mmap ) ( s#file * , s#socket * , s#vm_area_struct * ) ; t#ssize_t ( * sendpage ) ( s#socket * , s#page * , int , t#size_t , int ) ; t#ssize_t ( * splice_read ) ( s#socket * , t#loff_t * , s#pipe_inode_info * , t#size_t , unsigned int ) ; int ( * set_peek_off ) ( s#sock * , int ) ; int ( * peek_len ) ( s#socket * ) ; int ( * read_sock ) ( s#sock * , t#read_descriptor_t * , t#sk_read_actor_t ) ; int ( * sendpage_locked ) ( s#sock * , s#page * , int , t#size_t , int ) ; int ( * sendmsg_locked ) ( s#sock * , s#msghdr * , t#size_t ) ; }
+s#socket struct socket { t#socket_state state ; ; short type ; ; unsigned long flags ; s#socket_wq * wq ; s#file * file ; s#sock * sk ; const s#proto_ops * ops ; }
+s#sock_cgroup_data struct sock_cgroup_data { union { struct { t#u8 is_data ; t#u8 padding ; t#u16 prioidx ; t#u32 classid ; } __attribute__ ( ( packed ) ) ; t#u64 val ; } ; }
+s##sock_reuseport
+s#sock struct sock { s#sock_common __sk_common ; t#socket_lock_t sk_lock ; t#atomic_t sk_drops ; int sk_rcvlowat ; s#sk_buff_head sk_error_queue ; s#sk_buff_head sk_receive_queue ; struct { t#atomic_t rmem_alloc ; int len ; s#sk_buff * head ; s#sk_buff * tail ; } sk_backlog ; int sk_forward_alloc ; unsigned int sk_ll_usec ; unsigned int sk_napi_id ; int sk_rcvbuf ; s#sk_filter * sk_filter ; union { s#socket_wq * sk_wq ; s#socket_wq * sk_wq_raw ; } ; s#xfrm_policy * sk_policy [ 2 ] ; s#dst_entry * sk_rx_dst ; s#dst_entry * sk_dst_cache ; t#atomic_t sk_omem_alloc ; int sk_sndbuf ; int sk_wmem_queued ; t#atomic_t sk_wmem_alloc ; unsigned long sk_tsq_flags ; s#sk_buff * sk_send_head ; s#sk_buff_head sk_write_queue ; t#__s32 sk_peek_off ; int sk_write_pending ; t#__u32 sk_dst_pending_confirm ; t#u32 sk_pacing_status ; long sk_sndtimeo ; s#timer_list sk_timer ; t#__u32 sk_priority ; t#__u32 sk_mark ; t#u32 sk_pacing_rate ; t#u32 sk_max_pacing_rate ; s#page_frag sk_frag ; t#netdev_features_t sk_route_caps ; t#netdev_features_t sk_route_nocaps ; int sk_gso_type ; unsigned int sk_gso_max_size ; t#gfp_t sk_allocation ; t#__u32 sk_txhash ; unsigned int __sk_flags_offset [ 0 ] ; ; unsigned int sk_padding : 1 , sk_kern_sock : 1 , sk_no_check_tx : 1 , sk_no_check_rx : 1 , sk_userlocks : 4 , sk_protocol : 8 , sk_type : 16 ; ; t#u16 sk_gso_max_segs ; unsigned long sk_lingertime ; s#proto * sk_prot_creator ; t#rwlock_t sk_callback_lock ; int sk_err , sk_err_soft ; t#u32 sk_ack_backlog ; t#u32 sk_max_ack_backlog ; t#kuid_t sk_uid ; s#pid * sk_peer_pid ; const s#cred * sk_peer_cred ; long sk_rcvtimeo ; t#ktime_t sk_stamp ; t#u16 sk_tsflags ; t#u8 sk_shutdown ; t#u32 sk_tskey ; s#socket * sk_socket ; void * sk_user_data ; void * sk_security ; s#sock_cgroup_data sk_cgrp_data ; s#mem_cgroup * sk_memcg ; void ( * sk_state_change ) ( s#sock * ) ; void ( * sk_data_ready ) ( s#sock * ) ; void ( * sk_write_space ) ( s#sock * ) ; void ( * sk_error_report ) ( s#sock * ) ; int ( * sk_backlog_rcv ) ( s#sock * , s#sk_buff * ) ; s#sk_buff * ( * sk_validate_xmit_skb ) ( s#sock * , s#net_device * , s#sk_buff * ) ; void ( * sk_destruct ) ( s#sock * ) ; void * suse_kabi_padding ; s#sock_reuseport * sk_reuseport_cb ; s#callback_head sk_rcu ; }
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##blkcg_gq
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_stat_callback
+s##blk_trace
+s##blk_flush_queue
+s##bsg_job
+s##throtl_data
+s##blk_mq_tags
+s##ib_uverbs_file
+s##ib_umem_odp
+s#rdma_cgroup struct rdma_cgroup { s#cgroup_subsys_state css ; s#list_head rpools ; }
+s#ib_rdmacg_object struct ib_rdmacg_object { s#rdma_cgroup * cg ; }
+s#ib_ucontext struct ib_ucontext { s#ib_device * device ; s#ib_uverbs_file * ufile ; t#bool closing ; t#bool cleanup_retryable ; void ( * invalidate_range ) ( s#ib_umem_odp * , unsigned long , unsigned long ) ; s#mutex per_mm_list_lock ; s#list_head per_mm_list ; s#ib_rdmacg_object cg_obj ; }
+s##uverbs_api_object
+s#ib_uobject struct ib_uobject { t#u64 user_handle ; s#ib_uverbs_file * ufile ; s#ib_ucontext * context ; void * object ; s#list_head list ; s#ib_rdmacg_object cg_obj ; int id ; s#kref ref ; t#atomic_t usecnt ; s#callback_head rcu ; const s#uverbs_api_object * uapi_object ; }
+t#ib_comp_handler typedef void ( * ib_comp_handler ) ( s#ib_cq * , void * )
+e#ib_poll_context enum ib_poll_context { IB_POLL_DIRECT , IB_POLL_SOFTIRQ , IB_POLL_WORKQUEUE , IB_POLL_UNBOUND_WORKQUEUE , }
+s#ib_cqe struct ib_cqe { void ( * done ) ( s#ib_cq * , s#ib_wc * ) ; }
+e#ib_wc_status enum ib_wc_status { IB_WC_SUCCESS , IB_WC_LOC_LEN_ERR , IB_WC_LOC_QP_OP_ERR , IB_WC_LOC_EEC_OP_ERR , IB_WC_LOC_PROT_ERR , IB_WC_WR_FLUSH_ERR , IB_WC_MW_BIND_ERR , IB_WC_BAD_RESP_ERR , IB_WC_LOC_ACCESS_ERR , IB_WC_REM_INV_REQ_ERR , IB_WC_REM_ACCESS_ERR , IB_WC_REM_OP_ERR , IB_WC_RETRY_EXC_ERR , IB_WC_RNR_RETRY_EXC_ERR , IB_WC_LOC_RDD_VIOL_ERR , IB_WC_REM_INV_RD_REQ_ERR , IB_WC_REM_ABORT_ERR , IB_WC_INV_EECN_ERR , IB_WC_INV_EEC_STATE_ERR , IB_WC_FATAL_ERR , IB_WC_RESP_TIMEOUT_ERR , IB_WC_GENERAL_ERR }
+e#ib_wc_opcode enum ib_wc_opcode { IB_WC_SEND , IB_WC_RDMA_WRITE , IB_WC_RDMA_READ , IB_WC_COMP_SWAP , IB_WC_FETCH_ADD , IB_WC_LSO , IB_WC_LOCAL_INV , IB_WC_REG_MR , IB_WC_MASKED_COMP_SWAP , IB_WC_MASKED_FETCH_ADD , IB_WC_RECV = 1 << 7 , IB_WC_RECV_RDMA_WITH_IMM }
+s#ib_dm struct ib_dm { s#ib_device * device ; t#u32 length ; t#u32 flags ; s#ib_uobject * uobject ; t#atomic_t usecnt ; }
+e#rdma_restrack_type enum rdma_restrack_type { RDMA_RESTRACK_PD , RDMA_RESTRACK_CQ , RDMA_RESTRACK_QP , RDMA_RESTRACK_CM_ID , RDMA_RESTRACK_MR , RDMA_RESTRACK_MAX }
+s#rdma_restrack_entry struct rdma_restrack_entry { t#bool valid ; s#kref kref ; s#completion comp ; s#task_struct * task ; const char * kern_name ; s#hlist_node node ; e#rdma_restrack_type type ; }
+s#ib_mr struct ib_mr { s#ib_device * device ; s#ib_pd * pd ; t#u32 lkey ; t#u32 rkey ; t#u64 iova ; t#u64 length ; unsigned int page_size ; t#bool need_inval ; union { s#ib_uobject * uobject ; s#list_head qp_entry ; } ; s#ib_dm * dm ; s#rdma_restrack_entry res ; }
+s#ib_pd struct ib_pd { t#u32 local_dma_lkey ; t#u32 flags ; s#ib_device * device ; s#ib_uobject * uobject ; t#atomic_t usecnt ; t#u32 unsafe_global_rkey ; s#ib_mr * __internal_mr ; s#rdma_restrack_entry res ; }
+e#ib_srq_type enum ib_srq_type { IB_SRQT_BASIC , IB_SRQT_XRC , IB_SRQT_TM , }
+s#ib_xrcd struct ib_xrcd { s#ib_device * device ; t#atomic_t usecnt ; s#inode * inode ; s#mutex tgt_qp_mutex ; s#list_head tgt_qp_list ; }
+s#ib_srq struct ib_srq { s#ib_device * device ; s#ib_pd * pd ; s#ib_uobject * uobject ; void ( * event_handler ) ( s#ib_event * , void * ) ; void * srq_context ; e#ib_srq_type srq_type ; t#atomic_t usecnt ; struct { s#ib_cq * cq ; union { struct { s#ib_xrcd * xrcd ; t#u32 srq_num ; } xrc ; } ; } ext ; }
+u#ib_gid union ib_gid { t#u8 raw [ 16 ] ; struct { t#__be64 subnet_prefix ; t#__be64 interface_id ; } global ; }
+e#ib_gid_type enum ib_gid_type { IB_GID_TYPE_IB = 0 , IB_GID_TYPE_ROCE = 0 , IB_GID_TYPE_ROCE_UDP_ENCAP = 1 , IB_GID_TYPE_SIZE }
+s#ib_gid_attr struct ib_gid_attr { s#net_device * ndev ; s#ib_device * device ; u#ib_gid gid ; e#ib_gid_type gid_type ; t#u16 index ; t#u8 port_num ; }
+e#ib_qp_type enum ib_qp_type { IB_QPT_SMI , IB_QPT_GSI , IB_QPT_RC , IB_QPT_UC , IB_QPT_UD , IB_QPT_RAW_IPV6 , IB_QPT_RAW_ETHERTYPE , IB_QPT_RAW_PACKET = 8 , IB_QPT_XRC_INI = 9 , IB_QPT_XRC_TGT , IB_QPT_MAX , IB_QPT_DRIVER = 0xFF , IB_QPT_RESERVED1 = 0x1000 , IB_QPT_RESERVED2 , IB_QPT_RESERVED3 , IB_QPT_RESERVED4 , IB_QPT_RESERVED5 , IB_QPT_RESERVED6 , IB_QPT_RESERVED7 , IB_QPT_RESERVED8 , IB_QPT_RESERVED9 , IB_QPT_RESERVED10 , }
+e#ib_wq_state enum ib_wq_state { IB_WQS_RESET , IB_WQS_RDY , IB_WQS_ERR }
+e#ib_wq_type enum ib_wq_type { IB_WQT_RQ }
+s#ib_wq struct ib_wq { s#ib_device * device ; s#ib_uobject * uobject ; void * wq_context ; void ( * event_handler ) ( s#ib_event * , void * ) ; s#ib_pd * pd ; s#ib_cq * cq ; t#u32 wq_num ; e#ib_wq_state state ; e#ib_wq_type wq_type ; t#atomic_t usecnt ; }
+s#ib_rwq_ind_table struct ib_rwq_ind_table { s#ib_device * device ; s#ib_uobject * uobject ; t#atomic_t usecnt ; t#u32 ind_tbl_num ; t#u32 log_ind_tbl_size ; s#ib_wq * * ind_tbl ; }
+e#port_pkey_state enum port_pkey_state { IB_PORT_PKEY_NOT_VALID = 0 , IB_PORT_PKEY_VALID = 1 , IB_PORT_PKEY_LISTED = 2 , }
+s#ib_port_pkey struct ib_port_pkey { e#port_pkey_state state ; t#u16 pkey_index ; t#u8 port_num ; s#list_head qp_list ; s#list_head to_error_list ; s#ib_qp_security * sec ; }
+s#ib_ports_pkeys struct ib_ports_pkeys { s#ib_port_pkey main ; s#ib_port_pkey alt ; }
+s#ib_qp_security struct ib_qp_security { s#ib_qp * qp ; s#ib_device * dev ; s#mutex mutex ; s#ib_ports_pkeys * ports_pkeys ; s#list_head shared_qp_list ; void * security ; t#bool destroying ; t#atomic_t error_list_count ; s#completion error_complete ; int error_comps_pending ; }
+s#ib_qp struct ib_qp { s#ib_device * device ; s#ib_pd * pd ; s#ib_cq * send_cq ; s#ib_cq * recv_cq ; t#spinlock_t mr_lock ; int mrs_used ; s#list_head rdma_mrs ; s#list_head sig_mrs ; s#ib_srq * srq ; s#ib_xrcd * xrcd ; s#list_head xrcd_list ; t#atomic_t usecnt ; s#list_head open_list ; s#ib_qp * real_qp ; s#ib_uobject * uobject ; void ( * event_handler ) ( s#ib_event * , void * ) ; void * qp_context ; const s#ib_gid_attr * av_sgid_attr ; const s#ib_gid_attr * alt_path_sgid_attr ; t#u32 qp_num ; t#u32 max_write_sge ; t#u32 max_read_sge ; e#ib_qp_type qp_type ; s#ib_rwq_ind_table * rwq_ind_tbl ; s#ib_qp_security * qp_sec ; t#u8 port ; s#rdma_restrack_entry res ; }
+s#ib_wc struct ib_wc { union { t#u64 wr_id ; s#ib_cqe * wr_cqe ; } ; e#ib_wc_status status ; e#ib_wc_opcode opcode ; t#u32 vendor_err ; t#u32 byte_len ; s#ib_qp * qp ; union { t#__be32 imm_data ; t#u32 invalidate_rkey ; } ex ; t#u32 src_qp ; t#u32 slid ; int wc_flags ; t#u16 pkey_index ; t#u8 sl ; t#u8 dlid_path_bits ; t#u8 port_num ; t#u8 smac [ 6 ] ; t#u16 vlan_id ; t#u8 network_hdr_type ; }
+t#irq_poll_fn typedef int ( irq_poll_fn ) ( s#irq_poll * , int )
+s#irq_poll struct irq_poll { s#list_head list ; unsigned long state ; int weight ; t#irq_poll_fn * poll ; }
+s#ib_cq struct ib_cq { s#ib_device * device ; s#ib_uobject * uobject ; t#ib_comp_handler comp_handler ; void ( * event_handler ) ( s#ib_event * , void * ) ; void * cq_context ; int cqe ; t#atomic_t usecnt ; e#ib_poll_context poll_ctx ; s#ib_wc * wc ; union { s#irq_poll iop ; s#work_struct work ; } ; s#workqueue_struct * comp_wq ; s#rdma_restrack_entry res ; }
+e#ib_event_type enum ib_event_type { IB_EVENT_CQ_ERR , IB_EVENT_QP_FATAL , IB_EVENT_QP_REQ_ERR , IB_EVENT_QP_ACCESS_ERR , IB_EVENT_COMM_EST , IB_EVENT_SQ_DRAINED , IB_EVENT_PATH_MIG , IB_EVENT_PATH_MIG_ERR , IB_EVENT_DEVICE_FATAL , IB_EVENT_PORT_ACTIVE , IB_EVENT_PORT_ERR , IB_EVENT_LID_CHANGE , IB_EVENT_PKEY_CHANGE , IB_EVENT_SM_CHANGE , IB_EVENT_SRQ_ERR , IB_EVENT_SRQ_LIMIT_REACHED , IB_EVENT_QP_LAST_WQE_REACHED , IB_EVENT_CLIENT_REREGISTER , IB_EVENT_GID_CHANGE , IB_EVENT_WQ_FATAL , }
+s#ib_event struct ib_event { s#ib_device * device ; union { s#ib_cq * cq ; s#ib_qp * qp ; s#ib_srq * srq ; s#ib_wq * wq ; t#u8 port_num ; } element ; e#ib_event_type event ; }
+s#ib_event_handler struct ib_event_handler { s#ib_device * device ; void ( * handler ) ( s#ib_event_handler * , s#ib_event * ) ; s#list_head list ; }
+s##ib_pkey_cache
+s##ib_gid_table
+e#ib_port_state enum ib_port_state { IB_PORT_NOP = 0 , IB_PORT_DOWN = 1 , IB_PORT_INIT = 2 , IB_PORT_ARMED = 3 , IB_PORT_ACTIVE = 4 , IB_PORT_ACTIVE_DEFER = 5 }
+s#ib_port_cache struct ib_port_cache { t#u64 subnet_prefix ; s#ib_pkey_cache * pkey ; s#ib_gid_table * gid ; t#u8 lmc ; e#ib_port_state port_state ; }
+s#ib_cache struct ib_cache { t#rwlock_t lock ; s#ib_event_handler event_handler ; s#ib_port_cache * ports ; }
+s#ib_port_immutable struct ib_port_immutable { int pkey_tbl_len ; int gid_tbl_len ; t#u32 core_cap_flags ; t#u32 max_mad_size ; }
+s#ib_port_pkey_list struct ib_port_pkey_list { t#spinlock_t list_lock ; s#list_head pkey_list ; }
+s##iw_cm_verbs
+s#rdma_hw_stats struct rdma_hw_stats { s#mutex lock ; unsigned long timestamp ; unsigned long lifespan ; const char * const * names ; int num_counters ; t#u64 value [ ] ; }
+e#ib_atomic_cap enum ib_atomic_cap { IB_ATOMIC_NONE , IB_ATOMIC_HCA , IB_ATOMIC_GLOB }
+t#uint64_t typedef t#__u64 uint64_t
+s#ib_odp_caps struct ib_odp_caps { t#uint64_t general_caps ; struct { t#uint32_t rc_odp_caps ; t#uint32_t uc_odp_caps ; t#uint32_t ud_odp_caps ; } per_transport_caps ; }
+s#ib_rss_caps struct ib_rss_caps { t#u32 supported_qpts ; t#u32 max_rwq_indirection_tables ; t#u32 max_rwq_indirection_table_size ; }
+s#ib_tm_caps struct ib_tm_caps { t#u32 max_rndv_hdr_size ; t#u32 max_num_tags ; t#u32 flags ; t#u32 max_ops ; t#u32 max_sge ; }
+s#ib_cq_caps struct ib_cq_caps { t#u16 max_cq_moderation_count ; t#u16 max_cq_moderation_period ; }
+s#ib_device_attr struct ib_device_attr { t#u64 fw_ver ; t#__be64 sys_image_guid ; t#u64 max_mr_size ; t#u64 page_size_cap ; t#u32 vendor_id ; t#u32 vendor_part_id ; t#u32 hw_ver ; int max_qp ; int max_qp_wr ; t#u64 device_cap_flags ; int max_send_sge ; int max_recv_sge ; int max_sge_rd ; int max_cq ; int max_cqe ; int max_mr ; int max_pd ; int max_qp_rd_atom ; int max_ee_rd_atom ; int max_res_rd_atom ; int max_qp_init_rd_atom ; int max_ee_init_rd_atom ; e#ib_atomic_cap atomic_cap ; e#ib_atomic_cap masked_atomic_cap ; int max_ee ; int max_rdd ; int max_mw ; int max_raw_ipv6_qp ; int max_raw_ethy_qp ; int max_mcast_grp ; int max_mcast_qp_attach ; int max_total_mcast_qp_attach ; int max_ah ; int max_fmr ; int max_map_per_fmr ; int max_srq ; int max_srq_wr ; int max_srq_sge ; unsigned int max_fast_reg_page_list_len ; t#u16 max_pkeys ; t#u8 local_ca_ack_delay ; int sig_prot_cap ; int sig_guard_cap ; s#ib_odp_caps odp_caps ; t#uint64_t timestamp_mask ; t#uint64_t hca_core_clock ; s#ib_rss_caps rss_caps ; t#u32 max_wq_type_rq ; t#u32 raw_packet_caps ; s#ib_tm_caps tm_caps ; s#ib_cq_caps cq_caps ; t#u64 max_dm_size ; }
+s#ib_udata struct ib_udata { const void * inbuf ; void * outbuf ; t#size_t inlen ; t#size_t outlen ; }
+e#ib_mtu enum ib_mtu { IB_MTU_256 = 1 , IB_MTU_512 = 2 , IB_MTU_1024 = 3 , IB_MTU_2048 = 4 , IB_MTU_4096 = 5 }
+s#ib_port_attr struct ib_port_attr { t#u64 subnet_prefix ; e#ib_port_state state ; e#ib_mtu max_mtu ; e#ib_mtu active_mtu ; int gid_tbl_len ; unsigned int ip_gids : 1 ; t#u32 port_cap_flags ; t#u32 max_msg_sz ; t#u32 bad_pkey_cntr ; t#u32 qkey_viol_cntr ; t#u16 pkey_tbl_len ; t#u32 sm_lid ; t#u32 lid ; t#u8 lmc ; t#u8 max_vl_num ; t#u8 sm_sl ; t#u8 subnet_timeout ; t#u8 init_type_reply ; t#u8 active_width ; t#u8 active_speed ; t#u8 phys_state ; }
+e#rdma_link_layer enum rdma_link_layer { IB_LINK_LAYER_UNSPECIFIED , IB_LINK_LAYER_INFINIBAND , IB_LINK_LAYER_ETHERNET , }
+s#ib_device_modify struct ib_device_modify { t#u64 sys_image_guid ; char node_desc [ 64 ] ; }
+s#ib_port_modify struct ib_port_modify { t#u32 set_port_cap_mask ; t#u32 clr_port_cap_mask ; t#u8 init_type ; }
+e#rdma_ah_attr_type enum rdma_ah_attr_type { RDMA_AH_ATTR_TYPE_UNDEFINED , RDMA_AH_ATTR_TYPE_IB , RDMA_AH_ATTR_TYPE_ROCE , RDMA_AH_ATTR_TYPE_OPA , }
+s#ib_ah struct ib_ah { s#ib_device * device ; s#ib_pd * pd ; s#ib_uobject * uobject ; const s#ib_gid_attr * sgid_attr ; e#rdma_ah_attr_type type ; }
+s#ib_global_route struct ib_global_route { const s#ib_gid_attr * sgid_attr ; u#ib_gid dgid ; t#u32 flow_label ; t#u8 sgid_index ; t#u8 hop_limit ; t#u8 traffic_class ; }
+s#ib_ah_attr struct ib_ah_attr { t#u16 dlid ; t#u8 src_path_bits ; }
+s#roce_ah_attr struct roce_ah_attr { t#u8 dmac [ 6 ] ; }
+s#opa_ah_attr struct opa_ah_attr { t#u32 dlid ; t#u8 src_path_bits ; t#bool make_grd ; }
+s#rdma_ah_attr struct rdma_ah_attr { s#ib_global_route grh ; t#u8 sl ; t#u8 static_rate ; t#u8 port_num ; t#u8 ah_flags ; e#rdma_ah_attr_type type ; union { s#ib_ah_attr ib ; s#roce_ah_attr roce ; s#opa_ah_attr opa ; } ; }
+s#ib_srq_attr struct ib_srq_attr { t#u32 max_wr ; t#u32 max_sge ; t#u32 srq_limit ; }
+s#ib_srq_init_attr struct ib_srq_init_attr { void ( * event_handler ) ( s#ib_event * , void * ) ; void * srq_context ; s#ib_srq_attr attr ; e#ib_srq_type srq_type ; struct { s#ib_cq * cq ; union { struct { s#ib_xrcd * xrcd ; } xrc ; struct { t#u32 max_num_tags ; } tag_matching ; } ; } ext ; }
+e#ib_srq_attr_mask enum ib_srq_attr_mask { IB_SRQ_MAX_WR = 1 << 0 , IB_SRQ_LIMIT = 1 << 1 , }
+s#ib_sge struct ib_sge { t#u64 addr ; t#u32 length ; t#u32 lkey ; }
+s#ib_recv_wr struct ib_recv_wr { s#ib_recv_wr * next ; union { t#u64 wr_id ; s#ib_cqe * wr_cqe ; } ; s#ib_sge * sg_list ; int num_sge ; }
+s#ib_qp_cap struct ib_qp_cap { t#u32 max_send_wr ; t#u32 max_recv_wr ; t#u32 max_send_sge ; t#u32 max_recv_sge ; t#u32 max_inline_data ; t#u32 max_rdma_ctxs ; }
+e#ib_sig_type enum ib_sig_type { IB_SIGNAL_ALL_WR , IB_SIGNAL_REQ_WR }
+s#ib_qp_init_attr struct ib_qp_init_attr { void ( * event_handler ) ( s#ib_event * , void * ) ; void * qp_context ; s#ib_cq * send_cq ; s#ib_cq * recv_cq ; s#ib_srq * srq ; s#ib_xrcd * xrcd ; s#ib_qp_cap cap ; e#ib_sig_type sq_sig_type ; e#ib_qp_type qp_type ; t#u32 create_flags ; t#u8 port_num ; s#ib_rwq_ind_table * rwq_ind_tbl ; t#u32 source_qpn ; }
+e#ib_qp_state enum ib_qp_state { IB_QPS_RESET , IB_QPS_INIT , IB_QPS_RTR , IB_QPS_RTS , IB_QPS_SQD , IB_QPS_SQE , IB_QPS_ERR }
+e#ib_mig_state enum ib_mig_state { IB_MIG_MIGRATED , IB_MIG_REARM , IB_MIG_ARMED }
+s#ib_qp_attr struct ib_qp_attr { e#ib_qp_state qp_state ; e#ib_qp_state cur_qp_state ; e#ib_mtu path_mtu ; e#ib_mig_state path_mig_state ; t#u32 qkey ; t#u32 rq_psn ; t#u32 sq_psn ; t#u32 dest_qp_num ; int qp_access_flags ; s#ib_qp_cap cap ; s#rdma_ah_attr ah_attr ; s#rdma_ah_attr alt_ah_attr ; t#u16 pkey_index ; t#u16 alt_pkey_index ; t#u8 en_sqd_async_notify ; t#u8 sq_draining ; t#u8 max_rd_atomic ; t#u8 max_dest_rd_atomic ; t#u8 min_rnr_timer ; t#u8 port_num ; t#u8 timeout ; t#u8 retry_cnt ; t#u8 rnr_retry ; t#u8 alt_port_num ; t#u8 alt_timeout ; t#u32 rate_limit ; }
+E#IB_UVERBS_WR_RDMA_WRITE 0
+E#IB_UVERBS_WR_RDMA_WRITE_WITH_IMM 1
+E#IB_UVERBS_WR_SEND 2
+E#IB_UVERBS_WR_SEND_WITH_IMM 3
+E#IB_UVERBS_WR_RDMA_READ 4
+E#IB_UVERBS_WR_ATOMIC_CMP_AND_SWP 5
+E#IB_UVERBS_WR_ATOMIC_FETCH_AND_ADD 6
+E#IB_UVERBS_WR_TSO 10
+E#IB_UVERBS_WR_SEND_WITH_INV 9
+E#IB_UVERBS_WR_RDMA_READ_WITH_INV 11
+E#IB_UVERBS_WR_LOCAL_INV 7
+E#IB_UVERBS_WR_MASKED_ATOMIC_CMP_AND_SWP 12
+E#IB_UVERBS_WR_MASKED_ATOMIC_FETCH_AND_ADD 13
+e#ib_wr_opcode enum ib_wr_opcode { IB_WR_RDMA_WRITE = E#IB_UVERBS_WR_RDMA_WRITE , IB_WR_RDMA_WRITE_WITH_IMM = E#IB_UVERBS_WR_RDMA_WRITE_WITH_IMM , IB_WR_SEND = E#IB_UVERBS_WR_SEND , IB_WR_SEND_WITH_IMM = E#IB_UVERBS_WR_SEND_WITH_IMM , IB_WR_RDMA_READ = E#IB_UVERBS_WR_RDMA_READ , IB_WR_ATOMIC_CMP_AND_SWP = E#IB_UVERBS_WR_ATOMIC_CMP_AND_SWP , IB_WR_ATOMIC_FETCH_AND_ADD = E#IB_UVERBS_WR_ATOMIC_FETCH_AND_ADD , IB_WR_LSO = E#IB_UVERBS_WR_TSO , IB_WR_SEND_WITH_INV = E#IB_UVERBS_WR_SEND_WITH_INV , IB_WR_RDMA_READ_WITH_INV = E#IB_UVERBS_WR_RDMA_READ_WITH_INV , IB_WR_LOCAL_INV = E#IB_UVERBS_WR_LOCAL_INV , IB_WR_MASKED_ATOMIC_CMP_AND_SWP = E#IB_UVERBS_WR_MASKED_ATOMIC_CMP_AND_SWP , IB_WR_MASKED_ATOMIC_FETCH_AND_ADD = E#IB_UVERBS_WR_MASKED_ATOMIC_FETCH_AND_ADD , IB_WR_REG_MR = 0x20 , IB_WR_REG_SIG_MR , IB_WR_RESERVED1 = 0xf0 , IB_WR_RESERVED2 , IB_WR_RESERVED3 , IB_WR_RESERVED4 , IB_WR_RESERVED5 , IB_WR_RESERVED6 , IB_WR_RESERVED7 , IB_WR_RESERVED8 , IB_WR_RESERVED9 , IB_WR_RESERVED10 , }
+s#ib_send_wr struct ib_send_wr { s#ib_send_wr * next ; union { t#u64 wr_id ; s#ib_cqe * wr_cqe ; } ; s#ib_sge * sg_list ; int num_sge ; e#ib_wr_opcode opcode ; int send_flags ; union { t#__be32 imm_data ; t#u32 invalidate_rkey ; } ex ; }
+s#ib_cq_init_attr struct ib_cq_init_attr { unsigned int cqe ; int comp_vector ; t#u32 flags ; }
+E#IB_CQ_SOLICITED 1 << 0
+E#IB_CQ_NEXT_COMP 1 << 1
+e#ib_cq_notify_flags enum ib_cq_notify_flags { IB_CQ_SOLICITED = 1 << 0 , IB_CQ_NEXT_COMP = 1 << 1 , IB_CQ_SOLICITED_MASK = E#IB_CQ_SOLICITED | E#IB_CQ_NEXT_COMP , IB_CQ_REPORT_MISSED_EVENTS = 1 << 2 , }
+e#ib_mr_type enum ib_mr_type { IB_MR_TYPE_MEM_REG , IB_MR_TYPE_SIGNATURE , IB_MR_TYPE_SG_GAPS , }
+e#ib_mw_type enum ib_mw_type { IB_MW_TYPE_1 = 1 , IB_MW_TYPE_2 = 2 }
+s#ib_mw struct ib_mw { s#ib_device * device ; s#ib_pd * pd ; s#ib_uobject * uobject ; t#u32 rkey ; e#ib_mw_type type ; }
+s#ib_fmr struct ib_fmr { s#ib_device * device ; s#ib_pd * pd ; s#list_head list ; t#u32 lkey ; t#u32 rkey ; }
+s#ib_fmr_attr struct ib_fmr_attr { int max_pages ; int max_maps ; t#u8 page_shift ; }
+s#ib_grh struct ib_grh { t#__be32 version_tclass_flow ; t#__be16 paylen ; t#u8 next_hdr ; t#u8 hop_limit ; u#ib_gid sgid ; u#ib_gid dgid ; }
+s##ib_mad_hdr
+s#ib_flow struct ib_flow { s#ib_qp * qp ; s#ib_device * device ; s#ib_uobject * uobject ; }
+e#ib_flow_attr_type enum ib_flow_attr_type { IB_FLOW_ATTR_NORMAL = 0x0 , IB_FLOW_ATTR_ALL_DEFAULT = 0x1 , IB_FLOW_ATTR_MC_DEFAULT = 0x2 , IB_FLOW_ATTR_SNIFFER = 0x3 }
+s#ib_flow_attr struct ib_flow_attr { e#ib_flow_attr_type type ; t#u16 size ; t#u16 priority ; t#u32 flags ; t#u8 num_of_specs ; t#u8 port ; }
+e#ib_sig_err_type enum ib_sig_err_type { IB_SIG_BAD_GUARD , IB_SIG_BAD_REFTAG , IB_SIG_BAD_APPTAG , }
+s#ib_sig_err struct ib_sig_err { e#ib_sig_err_type err_type ; t#u32 expected ; t#u32 actual ; t#u64 sig_err_offset ; t#u32 key ; }
+s#ib_mr_status struct ib_mr_status { t#u32 fail_status ; s#ib_sig_err sig_err ; }
+s#ib_wq_init_attr struct ib_wq_init_attr { void * wq_context ; e#ib_wq_type wq_type ; t#u32 max_wr ; t#u32 max_sge ; s#ib_cq * cq ; void ( * event_handler ) ( s#ib_event * , void * ) ; t#u32 create_flags ; }
+s#ib_wq_attr struct ib_wq_attr { e#ib_wq_state wq_state ; e#ib_wq_state curr_wq_state ; t#u32 flags ; t#u32 flags_mask ; }
+s#ib_rwq_ind_table_init_attr struct ib_rwq_ind_table_init_attr { t#u32 log_ind_tbl_size ; s#ib_wq * * ind_tbl ; }
+e#ib_flow_action_type enum ib_flow_action_type { IB_FLOW_ACTION_UNSPECIFIED , IB_FLOW_ACTION_ESP = 1 , }
+s#ib_flow_action struct ib_flow_action { s#ib_device * device ; s#ib_uobject * uobject ; e#ib_flow_action_type type ; t#atomic_t usecnt ; }
+e#ib_uverbs_flow_action_esp_keymat enum ib_uverbs_flow_action_esp_keymat { IB_UVERBS_FLOW_ACTION_ESP_KEYMAT_AES_GCM , }
+s#ib_uverbs_flow_action_esp_keymat_aes_gcm struct ib_uverbs_flow_action_esp_keymat_aes_gcm { t#__u64 __attribute__ ( ( aligned ( 8 ) ) ) iv ; t#__u32 iv_algo ; t#__u32 salt ; t#__u32 icv_len ; t#__u32 key_len ; t#__u32 aes_key [ 256 / 32 ] ; }
+s#ib_flow_action_attrs_esp_keymats struct ib_flow_action_attrs_esp_keymats { e#ib_uverbs_flow_action_esp_keymat protocol ; union { s#ib_uverbs_flow_action_esp_keymat_aes_gcm aes_gcm ; } keymat ; }
+e#ib_uverbs_flow_action_esp_replay enum ib_uverbs_flow_action_esp_replay { IB_UVERBS_FLOW_ACTION_ESP_REPLAY_NONE , IB_UVERBS_FLOW_ACTION_ESP_REPLAY_BMP , }
+s#ib_uverbs_flow_action_esp_replay_bmp struct ib_uverbs_flow_action_esp_replay_bmp { t#__u32 size ; }
+s#ib_flow_action_attrs_esp_replays struct ib_flow_action_attrs_esp_replays { e#ib_uverbs_flow_action_esp_replay protocol ; union { s#ib_uverbs_flow_action_esp_replay_bmp bmp ; } replay ; }
+s#ib_flow_eth_filter struct ib_flow_eth_filter { t#u8 dst_mac [ 6 ] ; t#u8 src_mac [ 6 ] ; t#__be16 ether_type ; t#__be16 vlan_tag ; t#u8 real_sz [ 0 ] ; }
+s#ib_flow_spec_eth struct ib_flow_spec_eth { t#u32 type ; t#u16 size ; s#ib_flow_eth_filter val ; s#ib_flow_eth_filter mask ; }
+s#ib_flow_ib_filter struct ib_flow_ib_filter { t#__be16 dlid ; t#__u8 sl ; t#u8 real_sz [ 0 ] ; }
+s#ib_flow_spec_ib struct ib_flow_spec_ib { t#u32 type ; t#u16 size ; s#ib_flow_ib_filter val ; s#ib_flow_ib_filter mask ; }
+s#ib_flow_ipv4_filter struct ib_flow_ipv4_filter { t#__be32 src_ip ; t#__be32 dst_ip ; t#u8 proto ; t#u8 tos ; t#u8 ttl ; t#u8 flags ; t#u8 real_sz [ 0 ] ; }
+s#ib_flow_spec_ipv4 struct ib_flow_spec_ipv4 { t#u32 type ; t#u16 size ; s#ib_flow_ipv4_filter val ; s#ib_flow_ipv4_filter mask ; }
+s#ib_flow_tcp_udp_filter struct ib_flow_tcp_udp_filter { t#__be16 dst_port ; t#__be16 src_port ; t#u8 real_sz [ 0 ] ; }
+s#ib_flow_spec_tcp_udp struct ib_flow_spec_tcp_udp { t#u32 type ; t#u16 size ; s#ib_flow_tcp_udp_filter val ; s#ib_flow_tcp_udp_filter mask ; }
+s#ib_flow_ipv6_filter struct ib_flow_ipv6_filter { t#u8 src_ip [ 16 ] ; t#u8 dst_ip [ 16 ] ; t#__be32 flow_label ; t#u8 next_hdr ; t#u8 traffic_class ; t#u8 hop_limit ; t#u8 real_sz [ 0 ] ; }
+s#ib_flow_spec_ipv6 struct ib_flow_spec_ipv6 { t#u32 type ; t#u16 size ; s#ib_flow_ipv6_filter val ; s#ib_flow_ipv6_filter mask ; }
+s#ib_flow_tunnel_filter struct ib_flow_tunnel_filter { t#__be32 tunnel_id ; t#u8 real_sz [ 0 ] ; }
+s#ib_flow_spec_tunnel struct ib_flow_spec_tunnel { t#u32 type ; t#u16 size ; s#ib_flow_tunnel_filter val ; s#ib_flow_tunnel_filter mask ; }
+s#ib_flow_esp_filter struct ib_flow_esp_filter { t#__be32 spi ; t#__be32 seq ; t#u8 real_sz [ 0 ] ; }
+s#ib_flow_spec_esp struct ib_flow_spec_esp { t#u32 type ; t#u16 size ; s#ib_flow_esp_filter val ; s#ib_flow_esp_filter mask ; }
+s#ib_flow_gre_filter struct ib_flow_gre_filter { t#__be16 c_ks_res0_ver ; t#__be16 protocol ; t#__be32 key ; t#u8 real_sz [ 0 ] ; }
+s#ib_flow_spec_gre struct ib_flow_spec_gre { t#u32 type ; t#u16 size ; s#ib_flow_gre_filter val ; s#ib_flow_gre_filter mask ; }
+s#ib_flow_mpls_filter struct ib_flow_mpls_filter { t#__be32 tag ; t#u8 real_sz [ 0 ] ; }
+s#ib_flow_spec_mpls struct ib_flow_spec_mpls { t#u32 type ; t#u16 size ; s#ib_flow_mpls_filter val ; s#ib_flow_mpls_filter mask ; }
+e#ib_flow_spec_type enum ib_flow_spec_type { IB_FLOW_SPEC_ETH = 0x20 , IB_FLOW_SPEC_IB = 0x22 , IB_FLOW_SPEC_IPV4 = 0x30 , IB_FLOW_SPEC_IPV6 = 0x31 , IB_FLOW_SPEC_ESP = 0x34 , IB_FLOW_SPEC_TCP = 0x40 , IB_FLOW_SPEC_UDP = 0x41 , IB_FLOW_SPEC_VXLAN_TUNNEL = 0x50 , IB_FLOW_SPEC_GRE = 0x51 , IB_FLOW_SPEC_MPLS = 0x60 , IB_FLOW_SPEC_INNER = 0x100 , IB_FLOW_SPEC_ACTION_TAG = 0x1000 , IB_FLOW_SPEC_ACTION_DROP = 0x1001 , IB_FLOW_SPEC_ACTION_HANDLE = 0x1002 , IB_FLOW_SPEC_ACTION_COUNT = 0x1003 , }
+s#ib_flow_spec_action_tag struct ib_flow_spec_action_tag { e#ib_flow_spec_type type ; t#u16 size ; t#u32 tag_id ; }
+s#ib_flow_spec_action_drop struct ib_flow_spec_action_drop { e#ib_flow_spec_type type ; t#u16 size ; }
+s#ib_flow_spec_action_handle struct ib_flow_spec_action_handle { e#ib_flow_spec_type type ; t#u16 size ; s#ib_flow_action * act ; }
+s#ib_counters struct ib_counters { s#ib_device * device ; s#ib_uobject * uobject ; t#atomic_t usecnt ; }
+s#ib_flow_spec_action_count struct ib_flow_spec_action_count { e#ib_flow_spec_type type ; t#u16 size ; s#ib_counters * counters ; }
+u#ib_flow_spec union ib_flow_spec { struct { t#u32 type ; t#u16 size ; } ; s#ib_flow_spec_eth eth ; s#ib_flow_spec_ib ib ; s#ib_flow_spec_ipv4 ipv4 ; s#ib_flow_spec_tcp_udp tcp_udp ; s#ib_flow_spec_ipv6 ipv6 ; s#ib_flow_spec_tunnel tunnel ; s#ib_flow_spec_esp esp ; s#ib_flow_spec_gre gre ; s#ib_flow_spec_mpls mpls ; s#ib_flow_spec_action_tag flow_tag ; s#ib_flow_spec_action_drop drop ; s#ib_flow_spec_action_handle action ; s#ib_flow_spec_action_count flow_count ; }
+s#ib_flow_spec_list struct ib_flow_spec_list { s#ib_flow_spec_list * next ; u#ib_flow_spec spec ; }
+s#ib_flow_action_attrs_esp struct ib_flow_action_attrs_esp { s#ib_flow_action_attrs_esp_keymats * keymat ; s#ib_flow_action_attrs_esp_replays * replay ; s#ib_flow_spec_list * encap ; t#u32 esn ; t#u32 spi ; t#u32 seq ; t#u32 tfc_pad ; t#u64 flags ; t#u64 hard_limit_pkts ; }
+s##uverbs_attr_bundle
+s#ib_dm_alloc_attr struct ib_dm_alloc_attr { t#u64 length ; t#u32 alignment ; t#u32 flags ; }
+s#ib_dm_mr_attr struct ib_dm_mr_attr { t#u64 length ; t#u64 offset ; t#u32 access_flags ; }
+s#ib_counters_read_attr struct ib_counters_read_attr { t#u64 * counters_buff ; t#u32 ncounters ; t#u32 flags ; }
+e#rdma_netdev_t enum rdma_netdev_t { RDMA_NETDEV_OPA_VNIC , RDMA_NETDEV_IPOIB , }
+s#rdma_netdev_alloc_params struct rdma_netdev_alloc_params { t#size_t sizeof_priv ; unsigned int txqs ; unsigned int rxqs ; void * param ; int ( * initialize_rdma_netdev ) ( s#ib_device * , t#u8 , s#net_device * , void * ) ; }
+s#rdmacg_device struct rdmacg_device { s#list_head dev_node ; s#list_head rpools ; char * name ; }
+s#rdma_restrack_root struct rdma_restrack_root { s#rw_semaphore rwsem ; s#hlist_head hash [ 1 << ( 8 ) ] ; int ( * fill_res_entry ) ( s#sk_buff * , s#rdma_restrack_entry * ) ; }
+s##uverbs_object_tree_def
+e#rdma_driver_id enum rdma_driver_id { RDMA_DRIVER_UNKNOWN , RDMA_DRIVER_MLX5 , RDMA_DRIVER_MLX4 , RDMA_DRIVER_CXGB3 , RDMA_DRIVER_CXGB4 , RDMA_DRIVER_MTHCA , RDMA_DRIVER_BNXT_RE , RDMA_DRIVER_OCRDMA , RDMA_DRIVER_NES , RDMA_DRIVER_I40IW , RDMA_DRIVER_VMW_PVRDMA , RDMA_DRIVER_QEDR , RDMA_DRIVER_HNS , RDMA_DRIVER_USNIC , RDMA_DRIVER_RXE , RDMA_DRIVER_HFI1 , RDMA_DRIVER_QIB , }
+s#ib_device struct ib_device { s#device * dma_device ; char name [ 64 ] ; s#list_head event_handler_list ; t#spinlock_t event_handler_lock ; t#rwlock_t client_data_lock ; s#list_head core_list ; s#list_head client_data_list ; s#ib_cache cache ; s#ib_port_immutable * port_immutable ; int num_comp_vectors ; s#ib_port_pkey_list * port_pkey_list ; s#iw_cm_verbs * iwcm ; s#rdma_hw_stats * ( * alloc_hw_stats ) ( s#ib_device * , t#u8 ) ; int ( * get_hw_stats ) ( s#ib_device * , s#rdma_hw_stats * , t#u8 , int ) ; int ( * query_device ) ( s#ib_device * , s#ib_device_attr * , s#ib_udata * ) ; int ( * query_port ) ( s#ib_device * , t#u8 , s#ib_port_attr * ) ; e#rdma_link_layer ( * get_link_layer ) ( s#ib_device * , t#u8 ) ; s#net_device * ( * get_netdev ) ( s#ib_device * , t#u8 ) ; int ( * query_gid ) ( s#ib_device * , t#u8 , int , u#ib_gid * ) ; int ( * add_gid ) ( const s#ib_gid_attr * , void * * ) ; int ( * del_gid ) ( const s#ib_gid_attr * , void * * ) ; int ( * query_pkey ) ( s#ib_device * , t#u8 , t#u16 , t#u16 * ) ; int ( * modify_device ) ( s#ib_device * , int , s#ib_device_modify * ) ; int ( * modify_port ) ( s#ib_device * , t#u8 , int , s#ib_port_modify * ) ; s#ib_ucontext * ( * alloc_ucontext ) ( s#ib_device * , s#ib_udata * ) ; int ( * dealloc_ucontext ) ( s#ib_ucontext * ) ; int ( * mmap ) ( s#ib_ucontext * , s#vm_area_struct * ) ; s#ib_pd * ( * alloc_pd ) ( s#ib_device * , s#ib_ucontext * , s#ib_udata * ) ; int ( * dealloc_pd ) ( s#ib_pd * ) ; s#ib_ah * ( * create_ah ) ( s#ib_pd * , s#rdma_ah_attr * , s#ib_udata * ) ; int ( * modify_ah ) ( s#ib_ah * , s#rdma_ah_attr * ) ; int ( * query_ah ) ( s#ib_ah * , s#rdma_ah_attr * ) ; int ( * destroy_ah ) ( s#ib_ah * ) ; s#ib_srq * ( * create_srq ) ( s#ib_pd * , s#ib_srq_init_attr * , s#ib_udata * ) ; int ( * modify_srq ) ( s#ib_srq * , s#ib_srq_attr * , e#ib_srq_attr_mask , s#ib_udata * ) ; int ( * query_srq ) ( s#ib_srq * , s#ib_srq_attr * ) ; int ( * destroy_srq ) ( s#ib_srq * ) ; int ( * post_srq_recv ) ( s#ib_srq * , const s#ib_recv_wr * , const s#ib_recv_wr * * ) ; s#ib_qp * ( * create_qp ) ( s#ib_pd * , s#ib_qp_init_attr * , s#ib_udata * ) ; int ( * modify_qp ) ( s#ib_qp * , s#ib_qp_attr * , int , s#ib_udata * ) ; int ( * query_qp ) ( s#ib_qp * , s#ib_qp_attr * , int , s#ib_qp_init_attr * ) ; int ( * destroy_qp ) ( s#ib_qp * ) ; int ( * post_send ) ( s#ib_qp * , const s#ib_send_wr * , const s#ib_send_wr * * ) ; int ( * post_recv ) ( s#ib_qp * , const s#ib_recv_wr * , const s#ib_recv_wr * * ) ; s#ib_cq * ( * create_cq ) ( s#ib_device * , const s#ib_cq_init_attr * , s#ib_ucontext * , s#ib_udata * ) ; int ( * modify_cq ) ( s#ib_cq * , t#u16 , t#u16 ) ; int ( * destroy_cq ) ( s#ib_cq * ) ; int ( * resize_cq ) ( s#ib_cq * , int , s#ib_udata * ) ; int ( * poll_cq ) ( s#ib_cq * , int , s#ib_wc * ) ; int ( * peek_cq ) ( s#ib_cq * , int ) ; int ( * req_notify_cq ) ( s#ib_cq * , e#ib_cq_notify_flags ) ; int ( * req_ncomp_notif ) ( s#ib_cq * , int ) ; s#ib_mr * ( * get_dma_mr ) ( s#ib_pd * , int ) ; s#ib_mr * ( * reg_user_mr ) ( s#ib_pd * , t#u64 , t#u64 , t#u64 , int , s#ib_udata * ) ; int ( * rereg_user_mr ) ( s#ib_mr * , int , t#u64 , t#u64 , t#u64 , int , s#ib_pd * , s#ib_udata * ) ; int ( * dereg_mr ) ( s#ib_mr * ) ; s#ib_mr * ( * alloc_mr ) ( s#ib_pd * , e#ib_mr_type , t#u32 ) ; int ( * map_mr_sg ) ( s#ib_mr * , s#scatterlist * , int , unsigned int * ) ; s#ib_mw * ( * alloc_mw ) ( s#ib_pd * , e#ib_mw_type , s#ib_udata * ) ; int ( * dealloc_mw ) ( s#ib_mw * ) ; s#ib_fmr * ( * alloc_fmr ) ( s#ib_pd * , int , s#ib_fmr_attr * ) ; int ( * map_phys_fmr ) ( s#ib_fmr * , t#u64 * , int , t#u64 ) ; int ( * unmap_fmr ) ( s#list_head * ) ; int ( * dealloc_fmr ) ( s#ib_fmr * ) ; int ( * attach_mcast ) ( s#ib_qp * , u#ib_gid * , t#u16 ) ; int ( * detach_mcast ) ( s#ib_qp * , u#ib_gid * , t#u16 ) ; int ( * process_mad ) ( s#ib_device * , int , t#u8 , const s#ib_wc * , const s#ib_grh * , const s#ib_mad_hdr * , t#size_t , s#ib_mad_hdr * , t#size_t * , t#u16 * ) ; s#ib_xrcd * ( * alloc_xrcd ) ( s#ib_device * , s#ib_ucontext * , s#ib_udata * ) ; int ( * dealloc_xrcd ) ( s#ib_xrcd * ) ; s#ib_flow * ( * create_flow ) ( s#ib_qp * , s#ib_flow_attr * , int , s#ib_udata * ) ; int ( * destroy_flow ) ( s#ib_flow * ) ; int ( * check_mr_status ) ( s#ib_mr * , t#u32 , s#ib_mr_status * ) ; void ( * disassociate_ucontext ) ( s#ib_ucontext * ) ; void ( * drain_rq ) ( s#ib_qp * ) ; void ( * drain_sq ) ( s#ib_qp * ) ; int ( * set_vf_link_state ) ( s#ib_device * , int , t#u8 , int ) ; int ( * get_vf_config ) ( s#ib_device * , int , t#u8 , s#ifla_vf_info * ) ; int ( * get_vf_stats ) ( s#ib_device * , int , t#u8 , s#ifla_vf_stats * ) ; int ( * set_vf_guid ) ( s#ib_device * , int , t#u8 , t#u64 , int ) ; s#ib_wq * ( * create_wq ) ( s#ib_pd * , s#ib_wq_init_attr * , s#ib_udata * ) ; int ( * destroy_wq ) ( s#ib_wq * ) ; int ( * modify_wq ) ( s#ib_wq * , s#ib_wq_attr * , t#u32 , s#ib_udata * ) ; s#ib_rwq_ind_table * ( * create_rwq_ind_table ) ( s#ib_device * , s#ib_rwq_ind_table_init_attr * , s#ib_udata * ) ; int ( * destroy_rwq_ind_table ) ( s#ib_rwq_ind_table * ) ; s#ib_flow_action * ( * create_flow_action_esp ) ( s#ib_device * , const s#ib_flow_action_attrs_esp * , s#uverbs_attr_bundle * ) ; int ( * destroy_flow_action ) ( s#ib_flow_action * ) ; int ( * modify_flow_action_esp ) ( s#ib_flow_action * , const s#ib_flow_action_attrs_esp * , s#uverbs_attr_bundle * ) ; s#ib_dm * ( * alloc_dm ) ( s#ib_device * , s#ib_ucontext * , s#ib_dm_alloc_attr * , s#uverbs_attr_bundle * ) ; int ( * dealloc_dm ) ( s#ib_dm * ) ; s#ib_mr * ( * reg_dm_mr ) ( s#ib_pd * , s#ib_dm * , s#ib_dm_mr_attr * , s#uverbs_attr_bundle * ) ; s#ib_counters * ( * create_counters ) ( s#ib_device * , s#uverbs_attr_bundle * ) ; int ( * destroy_counters ) ( s#ib_counters * ) ; int ( * read_counters ) ( s#ib_counters * , s#ib_counters_read_attr * , s#uverbs_attr_bundle * ) ; s#net_device * ( * alloc_rdma_netdev ) ( s#ib_device * , t#u8 , e#rdma_netdev_t , const char * , unsigned char , void ( * ) ( s#net_device * ) ) ; int ( * rdma_netdev_get_params ) ( s#ib_device * , t#u8 , e#rdma_netdev_t , s#rdma_netdev_alloc_params * ) ; s#module * owner ; s#device dev ; const s#attribute_group * groups [ 3 ] ; s#kobject * ports_kobj ; s#list_head port_list ; enum { IB_DEV_UNINITIALIZED , IB_DEV_REGISTERED , IB_DEV_UNREGISTERED } reg_state ; int uverbs_abi_ver ; t#u64 uverbs_cmd_mask ; t#u64 uverbs_ex_cmd_mask ; char node_desc [ 64 ] ; t#__be64 node_guid ; t#u32 local_dma_lkey ; t#u16 is_switch : 1 ; t#u8 node_type ; t#u8 phys_port_cnt ; s#ib_device_attr attrs ; s#attribute_group * hw_stats_ag ; s#rdma_hw_stats * hw_stats ; s#rdmacg_device cg_device ; t#u32 index ; s#rdma_restrack_root res ; int ( * get_port_immutable ) ( s#ib_device * , t#u8 , s#ib_port_immutable * ) ; void ( * get_dev_fw_str ) ( s#ib_device * , char * ) ; const s#cpumask * ( * get_vector_affinity ) ( s#ib_device * , int ) ; const s#uverbs_object_tree_def * const * driver_specs ; e#rdma_driver_id driver_id ; t#refcount_t refcount ; s#completion unreg_completion ; }
+blk_mq_rdma_map_queues int blk_mq_rdma_map_queues ( s#blk_mq_tag_set * , s#ib_device * , int )
+
+/* block/blk-mq-sched.o */
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##klp_modinfo
+s##device_private
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##pipe_inode_info
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##workqueue_struct
+s##robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##blkcg_gq
+s##blk_queue_stats
+s##bsg_job
+s##throtl_data
+blk_mq_sched_free_hctx_data void blk_mq_sched_free_hctx_data ( s#request_queue * , void ( * ) ( s#blk_mq_hw_ctx * ) )
+blk_mq_sched_mark_restart_hctx void blk_mq_sched_mark_restart_hctx ( s#blk_mq_hw_ctx * )
+blk_mq_sched_try_merge t#bool blk_mq_sched_try_merge ( s#request_queue * , s#bio * , s#request * * )
+blk_mq_bio_list_merge t#bool blk_mq_bio_list_merge ( s#request_queue * , s#list_head * , s#bio * )
+blk_mq_sched_try_insert_merge t#bool blk_mq_sched_try_insert_merge ( s#request_queue * , s#request * )
+blk_mq_sched_request_inserted void blk_mq_sched_request_inserted ( s#request * )
+
+/* block/blk-mq-sysfs.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##iov_iter
+s##hd_geometry
+s##pr_ops
+s##kmem_cache
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+blk_mq_register_dev int blk_mq_register_dev ( s#device * , s#request_queue * )
+
+/* block/blk-mq-tag.o */
+s##workqueue_struct
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##device_private
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##iov_iter
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##blkcg_gq
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_trace
+s##bsg_job
+s##throtl_data
+t#busy_tag_iter_fn typedef void ( busy_tag_iter_fn ) ( s#request * , void * , t#bool )
+blk_mq_tagset_busy_iter void blk_mq_tagset_busy_iter ( s#blk_mq_tag_set * , t#busy_tag_iter_fn * , void * )
+blk_mq_unique_tag t#u32 blk_mq_unique_tag ( s#request * )
+
+/* block/blk-mq-virtio.o */
+s##workqueue_struct
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##device_private
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##pipe_inode_info
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##blkcg_gq
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_trace
+s##blk_flush_queue
+s##bsg_job
+s##throtl_data
+s#virtio_device_id struct virtio_device_id { t#__u32 device ; t#__u32 vendor ; }
+s#virtqueue struct virtqueue { s#list_head list ; void ( * callback ) ( s#virtqueue * ) ; const char * name ; s#virtio_device * vdev ; unsigned int index ; unsigned int num_free ; void * priv ; }
+t#vq_callback_t typedef void vq_callback_t ( s#virtqueue * )
+s##irq_affinity
+s#virtio_config_ops struct virtio_config_ops { void ( * get ) ( s#virtio_device * , unsigned , void * , unsigned ) ; void ( * set ) ( s#virtio_device * , unsigned , const void * , unsigned ) ; t#u32 ( * generation ) ( s#virtio_device * ) ; t#u8 ( * get_status ) ( s#virtio_device * ) ; void ( * set_status ) ( s#virtio_device * , t#u8 ) ; void ( * reset ) ( s#virtio_device * ) ; int ( * find_vqs ) ( s#virtio_device * , unsigned , s#virtqueue * [ ] , t#vq_callback_t * [ ] , const char * const [ ] , const t#bool * , s#irq_affinity * ) ; void ( * del_vqs ) ( s#virtio_device * ) ; t#u64 ( * get_features ) ( s#virtio_device * ) ; int ( * finalize_features ) ( s#virtio_device * ) ; const char * ( * bus_name ) ( s#virtio_device * ) ; int ( * set_vq_affinity ) ( s#virtqueue * , const s#cpumask * ) ; const s#cpumask * ( * get_vq_affinity ) ( s#virtio_device * , int ) ; }
+t#__virtio64 typedef t#__u64 __virtio64
+t#__virtio32 typedef t#__u32 __virtio32
+t#__virtio16 typedef t#__u16 __virtio16
+s#vring_desc struct vring_desc { t#__virtio64 addr ; t#__virtio32 len ; t#__virtio16 flags ; t#__virtio16 next ; }
+s#vring_avail struct vring_avail { t#__virtio16 flags ; t#__virtio16 idx ; t#__virtio16 ring [ ] ; }
+s#vring_used_elem struct vring_used_elem { t#__virtio32 id ; t#__virtio32 len ; }
+s#vring_used struct vring_used { t#__virtio16 flags ; t#__virtio16 idx ; s#vring_used_elem ring [ ] ; }
+s#vring struct vring { unsigned int num ; s#vring_desc * desc ; s#vring_avail * avail ; s#vring_used * used ; }
+s#vringh struct vringh { t#bool little_endian ; t#bool event_indices ; t#bool weak_barriers ; t#u16 last_avail_idx ; t#u16 last_used_idx ; t#u32 completed ; s#vring vring ; void ( * notify ) ( s#vringh * ) ; }
+t#vrh_callback_t typedef void vrh_callback_t ( s#virtio_device * , s#vringh * )
+s#vringh_config_ops struct vringh_config_ops { int ( * find_vrhs ) ( s#virtio_device * , unsigned , s#vringh * [ ] , t#vrh_callback_t * [ ] ) ; void ( * del_vrhs ) ( s#virtio_device * ) ; }
+s#virtio_device struct virtio_device { int index ; t#bool failed ; t#bool config_enabled ; t#bool config_change_pending ; t#spinlock_t config_lock ; s#device dev ; s#virtio_device_id id ; const s#virtio_config_ops * config ; const s#vringh_config_ops * vringh_config ; s#list_head vqs ; t#u64 features ; void * priv ; }
+blk_mq_virtio_map_queues int blk_mq_virtio_map_queues ( s#blk_mq_tag_set * , s#virtio_device * , int )
+
+/* block/blk-settings.o */
+blk_max_low_pfn extern unsigned long blk_max_low_pfn
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##device_private
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##iov_iter
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##workqueue_struct
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##blkcg_gq
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##blk_trace
+s##bsg_job
+s##throtl_data
+blk_queue_prep_rq extern void blk_queue_prep_rq ( s#request_queue * , t#prep_rq_fn * )
+blk_queue_unprep_rq extern void blk_queue_unprep_rq ( s#request_queue * , t#unprep_rq_fn * )
+blk_queue_softirq_done extern void blk_queue_softirq_done ( s#request_queue * , t#softirq_done_fn * )
+blk_queue_rq_timeout extern void blk_queue_rq_timeout ( s#request_queue * , unsigned int )
+blk_queue_rq_timed_out extern void blk_queue_rq_timed_out ( s#request_queue * , t#rq_timed_out_fn * )
+blk_queue_lld_busy extern void blk_queue_lld_busy ( s#request_queue * , t#lld_busy_fn * )
+blk_set_default_limits extern void blk_set_default_limits ( s#queue_limits * )
+blk_set_stacking_limits extern void blk_set_stacking_limits ( s#queue_limits * )
+blk_queue_make_request extern void blk_queue_make_request ( s#request_queue * , t#make_request_fn * )
+blk_queue_bounce_limit extern void blk_queue_bounce_limit ( s#request_queue * , t#u64 )
+blk_queue_max_hw_sectors extern void blk_queue_max_hw_sectors ( s#request_queue * , unsigned int )
+blk_queue_chunk_sectors extern void blk_queue_chunk_sectors ( s#request_queue * , unsigned int )
+blk_queue_max_discard_sectors extern void blk_queue_max_discard_sectors ( s#request_queue * , unsigned int )
+blk_queue_max_write_same_sectors extern void blk_queue_max_write_same_sectors ( s#request_queue * , unsigned int )
+blk_queue_max_write_zeroes_sectors extern void blk_queue_max_write_zeroes_sectors ( s#request_queue * , unsigned int )
+blk_queue_max_segments extern void blk_queue_max_segments ( s#request_queue * , unsigned short )
+blk_queue_max_discard_segments extern void blk_queue_max_discard_segments ( s#request_queue * , unsigned short )
+blk_queue_max_segment_size extern void blk_queue_max_segment_size ( s#request_queue * , unsigned int )
+blk_queue_logical_block_size extern void blk_queue_logical_block_size ( s#request_queue * , unsigned short )
+blk_queue_physical_block_size extern void blk_queue_physical_block_size ( s#request_queue * , unsigned int )
+blk_queue_alignment_offset extern void blk_queue_alignment_offset ( s#request_queue * , unsigned int )
+blk_limits_io_min extern void blk_limits_io_min ( s#queue_limits * , unsigned int )
+blk_queue_io_min extern void blk_queue_io_min ( s#request_queue * , unsigned int )
+blk_limits_io_opt extern void blk_limits_io_opt ( s#queue_limits * , unsigned int )
+blk_queue_io_opt extern void blk_queue_io_opt ( s#request_queue * , unsigned int )
+blk_queue_stack_limits extern void blk_queue_stack_limits ( s#request_queue * , s#request_queue * )
+blk_stack_limits extern int blk_stack_limits ( s#queue_limits * , s#queue_limits * , t#sector_t )
+bdev_stack_limits extern int bdev_stack_limits ( s#queue_limits * , s#block_device * , t#sector_t )
+disk_stack_limits extern void disk_stack_limits ( s#gendisk * , s#block_device * , t#sector_t )
+blk_queue_dma_pad extern void blk_queue_dma_pad ( s#request_queue * , unsigned int )
+blk_queue_update_dma_pad extern void blk_queue_update_dma_pad ( s#request_queue * , unsigned int )
+blk_queue_dma_drain extern int blk_queue_dma_drain ( s#request_queue * , t#dma_drain_needed_fn * , void * , unsigned int )
+blk_queue_segment_boundary extern void blk_queue_segment_boundary ( s#request_queue * , unsigned long )
+blk_queue_virt_boundary extern void blk_queue_virt_boundary ( s#request_queue * , unsigned long )
+blk_queue_dma_alignment extern void blk_queue_dma_alignment ( s#request_queue * , int )
+blk_queue_update_dma_alignment extern void blk_queue_update_dma_alignment ( s#request_queue * , int )
+blk_queue_flush_queueable extern void blk_queue_flush_queueable ( s#request_queue * , t#bool )
+blk_set_queue_depth extern void blk_set_queue_depth ( s#request_queue * , unsigned int )
+blk_queue_write_cache extern void blk_queue_write_cache ( s#request_queue * , t#bool , t#bool )
+
+/* block/blk-softirq.o */
+s##kmem_cache
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##iov_iter
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blkcg_gq
+s##blk_trace
+s##bsg_job
+s##throtl_data
+__blk_complete_request extern void __blk_complete_request ( s#request * )
+blk_complete_request extern void blk_complete_request ( s#request * )
+
+/* block/blk-stat.o */
+s##kmem_cache
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##iov_iter
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blkcg_gq
+s##blk_trace
+s##bsg_job
+s##throtl_data
+blk_stat_alloc_callback s#blk_stat_callback * blk_stat_alloc_callback ( void ( * ) ( s#blk_stat_callback * ) , int ( * ) ( const s#request * ) , unsigned int , void * )
+blk_stat_add_callback void blk_stat_add_callback ( s#request_queue * , s#blk_stat_callback * )
+blk_stat_remove_callback void blk_stat_remove_callback ( s#request_queue * , s#blk_stat_callback * )
+blk_stat_free_callback void blk_stat_free_callback ( s#blk_stat_callback * )
+blk_stat_enable_accounting void blk_stat_enable_accounting ( s#request_queue * )
+
+/* block/blk-sysfs.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##pipe_inode_info
+s##kmem_cache
+s##blk_queue_stats
+s##bsg_job
+s##throtl_data
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+blk_register_queue extern int blk_register_queue ( s#gendisk * )
+
+/* block/blk-tag.o */
+s##kmem_cache
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##iov_iter
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blkcg_gq
+s##blk_trace
+s##bsg_job
+s##throtl_data
+blk_queue_find_tag extern s#request * blk_queue_find_tag ( s#request_queue * , int )
+blk_free_tags extern void blk_free_tags ( s#blk_queue_tag * )
+blk_queue_free_tags extern void blk_queue_free_tags ( s#request_queue * )
+blk_init_tags extern s#blk_queue_tag * blk_init_tags ( int , int )
+blk_queue_init_tags extern int blk_queue_init_tags ( s#request_queue * , int , s#blk_queue_tag * , int )
+blk_queue_resize_tags extern int blk_queue_resize_tags ( s#request_queue * , int )
+blk_queue_start_tag extern int blk_queue_start_tag ( s#request_queue * , s#request * )
+
+/* block/blk-timeout.o */
+s##kmem_cache
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##iov_iter
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blkcg_gq
+s##blk_trace
+s##bsg_job
+s##throtl_data
+blk_abort_request extern void blk_abort_request ( s#request * )
+
+/* block/blk-wbt.o */
+s##blk_mq_ctx
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##pipe_inode_info
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##fs_pin
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##tty_struct
+s##tty_audit_buf
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##futex_pi_state
+s##ring_buffer
+s##bpf_prog
+s##trace_array
+s##tracer
+s##trace_buffer
+s##ring_buffer_iter
+s##event_filter
+s##ftrace_hash
+s##numa_group
+s##task_delay_info
+s##uprobe
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##blk_mq_hw_ctx
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##blk_mq_ops
+s##blk_trace
+s##blk_flush_queue
+s##bsg_job
+s##throtl_data
+s##blk_mq_tag_set
+wbt_disable_default void wbt_disable_default ( s#request_queue * )
+wbt_enable_default void wbt_enable_default ( s#request_queue * )
+
+/* block/blk-zoned.o */
+s##kmem_cache
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##iov_iter
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blkcg_gq
+s##blk_trace
+s##bsg_job
+s##throtl_data
+blk_req_needs_zone_write_lock t#bool blk_req_needs_zone_write_lock ( s#request * )
+__blk_req_zone_write_lock void __blk_req_zone_write_lock ( s#request * )
+__blk_req_zone_write_unlock void __blk_req_zone_write_unlock ( s#request * )
+blkdev_nr_zones extern unsigned int blkdev_nr_zones ( s#block_device * )
+s#blk_zone struct blk_zone { t#__u64 start ; t#__u64 len ; t#__u64 wp ; t#__u8 type ; t#__u8 cond ; t#__u8 non_seq ; t#__u8 reset ; t#__u8 reserved [ 36 ] ; }
+blkdev_report_zones extern int blkdev_report_zones ( s#block_device * , t#sector_t , s#blk_zone * , unsigned int * , t#gfp_t )
+blkdev_reset_zones extern int blkdev_reset_zones ( s#block_device * , t#sector_t , t#sector_t , t#gfp_t )
+
+/* block/bsg.o */
+s##blk_mq_ctx
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##iov_iter
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##blkcg_gq
+s##blk_flush_queue
+s##blk_mq_tags
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_stat_callback
+s##blk_trace
+s##bsg_job
+s##throtl_data
+bsg_unregister_queue void bsg_unregister_queue ( s#request_queue * )
+bsg_scsi_register_queue int bsg_scsi_register_queue ( s#request_queue * , s#device * )
+
+/* block/bsg-lib.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##iov_iter
+s##hd_geometry
+s##pr_ops
+s##blk_mq_ctx
+s##kmem_cache
+s##blkcg_gq
+s##blk_flush_queue
+s##blk_mq_tags
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_stat_callback
+s##blk_trace
+s##throtl_data
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s#bsg_buffer struct bsg_buffer { unsigned int payload_len ; int sg_cnt ; s#scatterlist * sg_list ; }
+s#bsg_job struct bsg_job { s#device * dev ; s#kref kref ; unsigned int timeout ; void * request ; void * reply ; unsigned int request_len ; unsigned int reply_len ; s#bsg_buffer request_payload ; s#bsg_buffer reply_payload ; int result ; unsigned int reply_payload_rcv_len ; void * dd_data ; }
+bsg_job_put void bsg_job_put ( s#bsg_job * )
+bsg_job_get int __attribute__ ( ( warn_unused_result ) ) bsg_job_get ( s#bsg_job * )
+bsg_job_done void bsg_job_done ( s#bsg_job * , int , unsigned int )
+bsg_setup_queue s#request_queue * bsg_setup_queue ( s#device * , const char * , t#bsg_job_fn * , int )
+
+/* block/elevator.o */
+s##kmem_cache
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##pipe_inode_info
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##bsg_job
+s##throtl_data
+elv_bio_merge_ok extern t#bool elv_bio_merge_ok ( s#request * , s#bio * )
+elevator_alloc extern s#elevator_queue * elevator_alloc ( s#request_queue * , s#elevator_type * )
+elv_rqhash_del void elv_rqhash_del ( s#request_queue * , s#request * )
+elv_rqhash_add void elv_rqhash_add ( s#request_queue * , s#request * )
+elv_rb_add extern void elv_rb_add ( s#rb_root * , s#request * )
+elv_rb_del extern void elv_rb_del ( s#rb_root * , s#request * )
+elv_rb_find extern s#request * elv_rb_find ( s#rb_root * , t#sector_t )
+elv_dispatch_sort extern void elv_dispatch_sort ( s#request_queue * , s#request * )
+elv_dispatch_add_tail extern void elv_dispatch_add_tail ( s#request_queue * , s#request * )
+__elv_add_request extern void __elv_add_request ( s#request_queue * , s#request * , int )
+elv_add_request extern void elv_add_request ( s#request_queue * , s#request * , int )
+elv_register extern int elv_register ( s#elevator_type * )
+elv_unregister extern void elv_unregister ( s#elevator_type * )
+elv_rb_former_request extern s#request * elv_rb_former_request ( s#request_queue * , s#request * )
+elv_rb_latter_request extern s#request * elv_rb_latter_request ( s#request_queue * , s#request * )
+
+/* block/genhd.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##iov_iter
+s##hd_geometry
+s##pr_ops
+s##kmem_cache
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##timer_rand_state
+s##disk_events
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+disk_get_part extern s#hd_struct * disk_get_part ( s#gendisk * , int )
+s#disk_part_iter struct disk_part_iter { s#gendisk * disk ; s#hd_struct * part ; int idx ; unsigned int flags ; }
+disk_part_iter_init extern void disk_part_iter_init ( s#disk_part_iter * , s#gendisk * , unsigned int )
+disk_part_iter_next extern s#hd_struct * disk_part_iter_next ( s#disk_part_iter * )
+disk_part_iter_exit extern void disk_part_iter_exit ( s#disk_part_iter * )
+disk_map_sector_rcu extern s#hd_struct * disk_map_sector_rcu ( s#gendisk * , t#sector_t )
+register_blkdev extern int register_blkdev ( unsigned int , const char * )
+unregister_blkdev extern void unregister_blkdev ( unsigned int , const char * )
+blk_register_region extern void blk_register_region ( t#dev_t , unsigned long , s#module * , s#kobject * ( * ) ( t#dev_t , int * , void * ) , int ( * ) ( t#dev_t , void * ) , void * )
+blk_unregister_region extern void blk_unregister_region ( t#dev_t , unsigned long )
+device_add_disk extern void device_add_disk ( s#device * , s#gendisk * , const s#attribute_group * * )
+device_add_disk_no_queue_reg extern void device_add_disk_no_queue_reg ( s#device * , s#gendisk * )
+del_gendisk extern void del_gendisk ( s#gendisk * )
+get_gendisk extern s#gendisk * get_gendisk ( t#dev_t , int * )
+bdget_disk extern s#block_device * bdget_disk ( s#gendisk * , int )
+blk_lookup_devt extern t#dev_t blk_lookup_devt ( const char * , int )
+alloc_disk extern s#gendisk * alloc_disk ( int )
+alloc_disk_node extern s#gendisk * alloc_disk_node ( int , int )
+get_disk_and_module extern s#kobject * get_disk_and_module ( s#gendisk * )
+put_disk extern void put_disk ( s#gendisk * )
+put_disk_and_module extern void put_disk_and_module ( s#gendisk * )
+set_device_ro extern void set_device_ro ( s#block_device * , int )
+set_disk_ro extern void set_disk_ro ( s#gendisk * , int )
+bdev_read_only extern int bdev_read_only ( s#block_device * )
+invalidate_partition extern int invalidate_partition ( s#gendisk * , int )
+
+/* block/ioctl.o */
+s##posix_acl
+s##vfsmount
+s##module
+s##device_private
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##bpf_prog_array
+s##cgroup_taskset
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s#hd_geometry struct hd_geometry { unsigned char heads ; unsigned char sectors ; unsigned short cylinders ; unsigned long start ; }
+e#pr_type enum pr_type { PR_WRITE_EXCLUSIVE = 1 , PR_EXCLUSIVE_ACCESS = 2 , PR_WRITE_EXCLUSIVE_REG_ONLY = 3 , PR_EXCLUSIVE_ACCESS_REG_ONLY = 4 , PR_WRITE_EXCLUSIVE_ALL_REGS = 5 , PR_EXCLUSIVE_ACCESS_ALL_REGS = 6 , }
+s#pr_ops struct pr_ops { int ( * pr_register ) ( s#block_device * , t#u64 , t#u64 , t#u32 ) ; int ( * pr_reserve ) ( s#block_device * , t#u64 , e#pr_type , t#u32 ) ; int ( * pr_release ) ( s#block_device * , t#u64 , e#pr_type ) ; int ( * pr_preempt ) ( s#block_device * , t#u64 , t#u64 , e#pr_type , t#bool ) ; int ( * pr_clear ) ( s#block_device * , t#u64 ) ; }
+s##blk_mq_ctx
+s##kmem_cache
+s##blk_mq_hw_ctx
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_mq_ops
+s##workqueue_struct
+s##blk_stat_callback
+s##blk_flush_queue
+s##bsg_job
+s##throtl_data
+s##blk_mq_tag_set
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##nlm_lockowner
+s##nfs4_lock_state
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+__blkdev_reread_part extern int __blkdev_reread_part ( s#block_device * )
+blkdev_reread_part extern int blkdev_reread_part ( s#block_device * )
+__blkdev_driver_ioctl extern int __blkdev_driver_ioctl ( s#block_device * , t#fmode_t , unsigned int , unsigned long )
+blkdev_ioctl extern int blkdev_ioctl ( s#block_device * , t#fmode_t , unsigned , unsigned long )
+
+/* block/ioprio.o */
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##posix_acl
+s##vfsmount
+s##module
+s##device_private
+s##sock
+s##pipe_inode_info
+s##kmem_cache
+s##fs_pin
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nlm_lockowner
+s##nfs4_lock_state
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##kstatfs
+s##eventfd_ctx
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_geometry
+s##pr_ops
+s##blk_mq_ctx
+s##blkcg_gq
+s##blk_mq_hw_ctx
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_mq_ops
+s##blk_stat_callback
+s##blk_trace
+s##blk_flush_queue
+s##bsg_job
+s##throtl_data
+s##blk_mq_tag_set
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##mtd_info
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##tty_struct
+s##tty_audit_buf
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##futex_pi_state
+s##ring_buffer
+s##bpf_prog
+s##trace_array
+s##tracer
+s##trace_buffer
+s##ring_buffer_iter
+s##event_filter
+s##ftrace_hash
+s##numa_group
+s##task_delay_info
+s##uprobe
+set_task_ioprio extern int set_task_ioprio ( s#task_struct * , int )
+
+/* block/partition-generic.o */
+s##posix_acl
+s##vfsmount
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##bpf_prog_array
+s##cgroup_taskset
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##hd_geometry
+s##pr_ops
+s##blk_mq_ctx
+s##kmem_cache
+s##blkcg_gq
+s##blk_mq_hw_ctx
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_mq_ops
+s##workqueue_struct
+s##blk_stat_callback
+s##blk_flush_queue
+s##bsg_job
+s##throtl_data
+s##blk_mq_tag_set
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##nlm_lockowner
+s##nfs4_lock_state
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+bdevname extern const char * bdevname ( s#block_device * , char * )
+bio_devname extern const char * bio_devname ( s#bio * , char * )
+__bdevname extern const char * __bdevname ( t#dev_t , char * )
+t#Sector typedef struct { s#page * v ; } Sector
+read_dev_sector unsigned char * read_dev_sector ( s#block_device * , t#sector_t , t#Sector * )
+
+/* block/scsi_ioctl.o */
+scsi_command_size_tbl extern const unsigned char scsi_command_size_tbl [ 8 ]
+blk_verify_command extern int blk_verify_command ( unsigned char * , t#fmode_t )
+s##blk_mq_ctx
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##pipe_inode_info
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##blkcg_gq
+s##blk_flush_queue
+s##blk_mq_tags
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_stat_callback
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s#scsi_ioctl_command struct scsi_ioctl_command { unsigned int inlen ; unsigned int outlen ; unsigned char data [ 0 ] ; }
+sg_scsi_ioctl extern int sg_scsi_ioctl ( s#request_queue * , s#gendisk * , t#fmode_t , s#scsi_ioctl_command * )
+scsi_cmd_ioctl extern int scsi_cmd_ioctl ( s#request_queue * , s#gendisk * , t#fmode_t , unsigned int , void * )
+scsi_verify_blk_ioctl extern int scsi_verify_blk_ioctl ( s#block_device * , unsigned int )
+scsi_cmd_blk_ioctl extern int scsi_cmd_blk_ioctl ( s#block_device * , t#fmode_t , unsigned int , void * )
+s#scsi_request struct scsi_request { unsigned char __cmd [ 16 ] ; unsigned char * cmd ; unsigned short cmd_len ; int result ; unsigned int sense_len ; unsigned int resid_len ; int retries ; void * sense ; }
+scsi_req_init void scsi_req_init ( s#scsi_request * )
+
+/* block/t10-pi.o */
+t10_pi_type1_crc extern const s#blk_integrity_profile t10_pi_type1_crc
+t10_pi_type1_ip extern const s#blk_integrity_profile t10_pi_type1_ip
+t10_pi_type3_crc extern const s#blk_integrity_profile t10_pi_type3_crc
+t10_pi_type3_ip extern const s#blk_integrity_profile t10_pi_type3_ip
+s##kmem_cache
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##iov_iter
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##hd_geometry
+s##pr_ops
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##blk_mq_hw_ctx
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blkcg_gq
+s##blk_mq_ops
+s##blk_mq_ctx
+s##blk_stat_callback
+s##blk_trace
+s##blk_flush_queue
+s##bsg_job
+s##throtl_data
+s##blk_mq_tag_set
+t10_pi_prepare extern void t10_pi_prepare ( s#request * , t#u8 )
+t10_pi_complete extern void t10_pi_complete ( s#request * , t#u8 , unsigned int )
+
+/* certs/blacklist.o */
+is_hash_blacklisted extern int is_hash_blacklisted ( const t#u8 * , t#size_t , const char * )
+
+/* certs/system_keyring.o */
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##vm_operations_struct
+s##file
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##address_space
+s##dev_pagemap
+s##kmem_cache
+s##mem_cgroup
+s##xol_area
+s##compat_timespec
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##key_user
+s#key_preparsed_payload struct key_preparsed_payload { char * description ; u#key_payload payload ; const void * data ; t#size_t datalen ; t#size_t quotalen ; t#time_t expiry ; }
+s#key_match_data struct key_match_data { t#bool ( * cmp ) ( const s#key * , const s#key_match_data * ) ; const void * raw_data ; void * preparsed ; unsigned lookup_type ; }
+s##seq_file
+s#key_construction struct key_construction { s#key * key ; s#key * authkey ; }
+t#request_key_actor_t typedef int ( * request_key_actor_t ) ( s#key_construction * , const char * , void * )
+s#key_type struct key_type { const char * name ; t#size_t def_datalen ; int ( * vet_description ) ( const char * ) ; int ( * preparse ) ( s#key_preparsed_payload * ) ; void ( * free_preparse ) ( s#key_preparsed_payload * ) ; int ( * instantiate ) ( s#key * , s#key_preparsed_payload * ) ; int ( * update ) ( s#key * , s#key_preparsed_payload * ) ; int ( * match_preparse ) ( s#key_match_data * ) ; void ( * match_free ) ( s#key_match_data * ) ; void ( * revoke ) ( s#key * ) ; void ( * destroy ) ( s#key * ) ; void ( * describe ) ( const s#key * , s#seq_file * ) ; long ( * read ) ( const s#key * , char * , t#size_t ) ; t#request_key_actor_t request_key ; s#key_restriction * ( * lookup_restriction ) ( const char * ) ; s#list_head link ; s#lock_class_key lock_class ; }
+s##assoc_array_ptr
+e#key_being_used_for enum key_being_used_for { VERIFYING_MODULE_SIGNATURE , VERIFYING_FIRMWARE_SIGNATURE , VERIFYING_KEXEC_PE_SIGNATURE , VERIFYING_KEY_SIGNATURE , VERIFYING_KEY_SELF_SIGNATURE , VERIFYING_UNSPECIFIED_SIGNATURE , NR__KEY_BEING_USED_FOR }
+verify_pkcs7_signature extern int verify_pkcs7_signature ( const void * , t#size_t , const void * , t#size_t , s#key * , e#key_being_used_for , int ( * ) ( void * , const void * , t#size_t , t#size_t ) , void * )
+
+/* crypto/ablkcipher.o */
+s##posix_acl
+s##vfsmount
+s##sock
+s##bio_vec
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##cgroup_namespace
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s#scatter_walk struct scatter_walk { s#scatterlist * sg ; unsigned int offset ; }
+s#ablkcipher_walk struct ablkcipher_walk { struct { s#page * page ; unsigned int offset ; } src , dst ; s#scatter_walk in ; unsigned int nbytes ; s#scatter_walk out ; unsigned int total ; s#list_head buffers ; t#u8 * iv_buffer ; t#u8 * iv ; int flags ; unsigned int blocksize ; }
+__ablkcipher_walk_complete void __ablkcipher_walk_complete ( s#ablkcipher_walk * )
+s##net_device
+s##sec_path
+s##rtattr
+s#crypto_template struct crypto_template { s#list_head list ; s#hlist_head instances ; s#module * module ; s#crypto_instance * ( * alloc ) ( s#rtattr * * ) ; void ( * free ) ( s#crypto_instance * ) ; int ( * create ) ( s#crypto_template * , s#rtattr * * ) ; char name [ 128 ] ; }
+s#crypto_instance struct crypto_instance { s#crypto_alg alg ; s#crypto_template * tmpl ; s#hlist_node list ; void * __ctx [ ] __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; }
+s#crypto_type struct crypto_type { unsigned int ( * ctxsize ) ( s#crypto_alg * , t#u32 , t#u32 ) ; unsigned int ( * extsize ) ( s#crypto_alg * ) ; int ( * init ) ( s#crypto_tfm * , t#u32 , t#u32 ) ; int ( * init_tfm ) ( s#crypto_tfm * ) ; void ( * show ) ( s#seq_file * , s#crypto_alg * ) ; int ( * report ) ( s#sk_buff * , s#crypto_alg * ) ; s#crypto_alg * ( * lookup ) ( const char * , t#u32 , t#u32 ) ; void ( * free ) ( s#crypto_instance * ) ; unsigned int type ; unsigned int maskclear ; unsigned int maskset ; unsigned int tfmsize ; }
+s#skcipher_givcrypt_request struct skcipher_givcrypt_request { t#u64 seq ; t#u8 * giv ; s#ablkcipher_request creq ; }
+ablkcipher_walk_done int ablkcipher_walk_done ( s#ablkcipher_request * , s#ablkcipher_walk * , int )
+ablkcipher_walk_phys int ablkcipher_walk_phys ( s#ablkcipher_request * , s#ablkcipher_walk * )
+crypto_ablkcipher_type extern const s#crypto_type crypto_ablkcipher_type
+crypto_givcipher_type extern const s#crypto_type crypto_givcipher_type
+
+/* crypto/acompress.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##cgroup_namespace
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s##skcipher_givcrypt_request
+s#acomp_req struct acomp_req { s#crypto_async_request base ; s#scatterlist * src ; s#scatterlist * dst ; unsigned int slen ; unsigned int dlen ; t#u32 flags ; void * __ctx [ ] __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; }
+s#crypto_acomp struct crypto_acomp { int ( * compress ) ( s#acomp_req * ) ; int ( * decompress ) ( s#acomp_req * ) ; void ( * dst_free ) ( s#scatterlist * ) ; unsigned int reqsize ; s#crypto_tfm base ; }
+crypto_alloc_acomp s#crypto_acomp * crypto_alloc_acomp ( const char * , t#u32 , t#u32 )
+acomp_request_alloc s#acomp_req * acomp_request_alloc ( s#crypto_acomp * )
+acomp_request_free void acomp_request_free ( s#acomp_req * )
+s#acomp_alg struct acomp_alg { int ( * compress ) ( s#acomp_req * ) ; int ( * decompress ) ( s#acomp_req * ) ; void ( * dst_free ) ( s#scatterlist * ) ; int ( * init ) ( s#crypto_acomp * ) ; void ( * exit ) ( s#crypto_acomp * ) ; unsigned int reqsize ; s#crypto_alg base ; }
+crypto_register_acomp int crypto_register_acomp ( s#acomp_alg * )
+crypto_unregister_acomp int crypto_unregister_acomp ( s#acomp_alg * )
+crypto_register_acomps int crypto_register_acomps ( s#acomp_alg * , int )
+crypto_unregister_acomps void crypto_unregister_acomps ( s#acomp_alg * , int )
+
+/* crypto/aead.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##proc_dir_entry
+s##uevent_sock
+s##iw_handler_def
+s##iw_public_data
+s##sec_path
+s##neigh_parms
+s##netpoll_info
+s##neighbour
+s##udp_tunnel_info
+s##bpf_prog
+s##bpf_prog_offload_ops
+s##bpf_offloaded_map
+s##xdp_umem
+s##switchdev_ops
+s##l3mdev_ops
+s##ndisc_ops
+s##xfrm_state
+s##tls_crypto_info
+s##tls_context
+s##vlan_info
+s##dsa_switch_tree
+s##in_device
+s##dn_dev
+s##inet6_dev
+s##wireless_dev
+s##wpan_dev
+s##mpls_dev
+s##mini_Qdisc
+s##Qdisc
+s##nf_hook_entry
+s##cpu_rmap
+s##pcpu_dstats
+s##garp_port
+s##mrp_port
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##kmem_cache
+s##mem_cgroup
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##rtnl_link_ops
+s##phy_device
+s##sfp_bus
+s##prot_inuse
+s##ipv4_devconf
+s##fib_rules_ops
+s##fib_table
+s##inet_peer_base
+s##xt_table
+s##inet_hashinfo
+s##fib_notifier_ops
+s##ipv6_devconf
+s##rt6_info
+s##rt6_statistics
+s##fib6_table
+s##dst_entry
+s##seg6_pernet_data
+s##sctp_mib
+s##nf_queue_handler
+s##nf_logger
+s##ebt_table
+s##workqueue_struct
+s##nf_ct_event_notifier
+s##nf_exp_event_notifier
+s##nft_af_info
+s##net_generic
+s##netns_ipvs
+s##mpls_route
+s##dev_rcv_lists
+s##s_stats
+s##s_pstats
+s##bpf_prog_array
+s##cgroup_taskset
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s#rtattr struct rtattr { unsigned short rta_len ; unsigned short rta_type ; }
+s##skcipher_givcrypt_request
+s#crypto_aead struct crypto_aead { unsigned int authsize ; unsigned int reqsize ; s#crypto_tfm base ; }
+crypto_aead_setkey int crypto_aead_setkey ( s#crypto_aead * , const t#u8 * , unsigned int )
+crypto_aead_setauthsize int crypto_aead_setauthsize ( s#crypto_aead * , unsigned int )
+s#aead_request struct aead_request { s#crypto_async_request base ; unsigned int assoclen ; unsigned int cryptlen ; t#u8 * iv ; s#scatterlist * src ; s#scatterlist * dst ; void * __ctx [ ] __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; }
+s#aead_alg struct aead_alg { int ( * setkey ) ( s#crypto_aead * , const t#u8 * , unsigned int ) ; int ( * setauthsize ) ( s#crypto_aead * , unsigned int ) ; int ( * encrypt ) ( s#aead_request * ) ; int ( * decrypt ) ( s#aead_request * ) ; int ( * init ) ( s#crypto_aead * ) ; void ( * exit ) ( s#crypto_aead * ) ; const char * geniv ; unsigned int ivsize ; unsigned int maxauthsize ; unsigned int chunksize ; s#crypto_alg base ; }
+s#aead_instance struct aead_instance { void ( * free ) ( s#aead_instance * ) ; union { struct { char head [ __builtin_offsetof ( struct aead_alg , base ) ] ; s#crypto_instance base ; } s ; s#aead_alg alg ; } ; }
+aead_geniv_alloc s#aead_instance * aead_geniv_alloc ( s#crypto_template * , s#rtattr * * , t#u32 , t#u32 )
+aead_geniv_free void aead_geniv_free ( s#aead_instance * )
+aead_init_geniv int aead_init_geniv ( s#crypto_aead * )
+aead_exit_geniv void aead_exit_geniv ( s#crypto_aead * )
+s#crypto_spawn struct crypto_spawn { s#list_head list ; s#crypto_alg * alg ; s#crypto_instance * inst ; const s#crypto_type * frontend ; t#u32 mask ; }
+s#crypto_aead_spawn struct crypto_aead_spawn { s#crypto_spawn base ; }
+crypto_grab_aead int crypto_grab_aead ( s#crypto_aead_spawn * , const char * , t#u32 , t#u32 )
+crypto_alloc_aead s#crypto_aead * crypto_alloc_aead ( const char * , t#u32 , t#u32 )
+crypto_register_aead int crypto_register_aead ( s#aead_alg * )
+crypto_unregister_aead void crypto_unregister_aead ( s#aead_alg * )
+crypto_register_aeads int crypto_register_aeads ( s#aead_alg * , int )
+crypto_unregister_aeads void crypto_unregister_aeads ( s#aead_alg * , int )
+aead_register_instance int aead_register_instance ( s#crypto_template * , s#aead_instance * )
+
+/* crypto/aes_generic.o */
+crypto_ft_tab extern const t#u32 crypto_ft_tab [ 4 ] [ 256 ]
+crypto_fl_tab extern const t#u32 crypto_fl_tab [ 4 ] [ 256 ]
+crypto_it_tab extern const t#u32 crypto_it_tab [ 4 ] [ 256 ]
+crypto_il_tab extern const t#u32 crypto_il_tab [ 4 ] [ 256 ]
+crypto_aes_expand_key int crypto_aes_expand_key ( s#crypto_aes_ctx * , const t#u8 * , unsigned int )
+s##scatterlist
+s##crypto_type
+s##skcipher_givcrypt_request
+s##sock
+s##seq_file
+s##file
+s##vm_operations_struct
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##cred
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##pipe_inode_info
+s##address_space
+s##dev_pagemap
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##tracepoint
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+crypto_aes_set_key int crypto_aes_set_key ( s#crypto_tfm * , const t#u8 * , unsigned int )
+
+/* crypto/af_alg.o */
+s##posix_acl
+s##vfsmount
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##proc_dir_entry
+s##uevent_sock
+s##iw_handler_def
+s##iw_public_data
+s##sec_path
+s##workqueue_struct
+s##netpoll_info
+s##udp_tunnel_info
+s##bpf_prog_aux
+s##bpf_prog_offload_ops
+s##bpf_offloaded_map
+s##xdp_umem
+s##switchdev_ops
+s##l3mdev_ops
+s##ndisc_ops
+s##xfrm_state
+s##tls_crypto_info
+s##tls_context
+s##vlan_info
+s##dsa_switch_tree
+s##in_device
+s##dn_dev
+s##inet6_dev
+s##wireless_dev
+s##wpan_dev
+s##mpls_dev
+s##tcf_walker
+s##qdisc_walker
+s##net_rate_estimator
+s##nf_hook_entry
+s##cpu_rmap
+s##pcpu_dstats
+s##garp_port
+s##mrp_port
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##phy_device
+s##sfp_bus
+s##prot_inuse
+s##ipv4_devconf
+s##fib_rules_ops
+s##fib_table
+s##inet_peer_base
+s##kmem_cache
+s##xt_table
+s##inet_hashinfo
+s##fib_notifier_ops
+s##ipv6_devconf
+s##rt6_info
+s##rt6_statistics
+s##fib6_table
+s##lwtunnel_state
+s##rtable
+s##dn_route
+s##seg6_pernet_data
+s##sctp_mib
+s##nf_queue_handler
+s##nf_logger
+s##ebt_table
+s##nf_ct_event_notifier
+s##nf_exp_event_notifier
+s##nft_af_info
+s##net_generic
+s##netns_ipvs
+s##mpls_route
+s##dev_rcv_lists
+s##s_stats
+s##s_pstats
+s##bpf_prog_array
+s##cgroup_taskset
+s##eventfd_ctx
+s##mempolicy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##gendisk
+s##request_queue
+s##blk_plug
+s##reclaim_state
+s##robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##nlm_lockowner
+s##nfs4_lock_state
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##mtd_info
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##swap_info_struct
+s##request_sock_ops
+s##timewait_sock_ops
+s##udp_table
+s##raw_hashinfo
+s##xfrm_policy
+s##sock_reuseport
+s#af_alg_type struct af_alg_type { void * ( * bind ) ( const char * , t#u32 , t#u32 ) ; void ( * release ) ( void * ) ; int ( * setkey ) ( void * , const t#u8 * , unsigned int ) ; int ( * accept ) ( void * , s#sock * ) ; int ( * accept_nokey ) ( void * , s#sock * ) ; int ( * setauthsize ) ( void * , unsigned int ) ; s#proto_ops * ops ; s#proto_ops * ops_nokey ; s#module * owner ; char name [ 14 ] ; }
+af_alg_register_type int af_alg_register_type ( const s#af_alg_type * )
+af_alg_unregister_type int af_alg_unregister_type ( const s#af_alg_type * )
+af_alg_release int af_alg_release ( s#socket * )
+af_alg_release_parent void af_alg_release_parent ( s#sock * )
+af_alg_accept int af_alg_accept ( s#sock * , s#socket * , t#bool )
+s#af_alg_sgl struct af_alg_sgl { s#scatterlist sg [ 16 + 1 ] ; s#page * pages [ 16 ] ; unsigned int npages ; }
+af_alg_make_sg int af_alg_make_sg ( s#af_alg_sgl * , s#iov_iter * , int )
+af_alg_link_sg void af_alg_link_sg ( s#af_alg_sgl * , s#af_alg_sgl * )
+af_alg_free_sg void af_alg_free_sg ( s#af_alg_sgl * )
+s#af_alg_iv struct af_alg_iv { t#__u32 ivlen ; t#__u8 iv [ 0 ] ; }
+s#af_alg_control struct af_alg_control { s#af_alg_iv * iv ; int op ; unsigned int aead_assoclen ; }
+af_alg_cmsg_send int af_alg_cmsg_send ( s#msghdr * , s#af_alg_control * )
+s#af_alg_completion struct af_alg_completion { s#completion completion ; int err ; }
+af_alg_wait_for_completion int af_alg_wait_for_completion ( int , s#af_alg_completion * )
+s##crypto_type
+s##skcipher_givcrypt_request
+af_alg_complete void af_alg_complete ( s#crypto_async_request * , int )
+
+/* crypto/ahash.o */
+s##posix_acl
+s##vfsmount
+s##sock
+s##bio_vec
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##cgroup_namespace
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s#crypto_hash_walk struct crypto_hash_walk { char * data ; unsigned int offset ; unsigned int alignmask ; s#page * pg ; unsigned int entrylen ; unsigned int total ; s#scatterlist * sg ; unsigned int flags ; }
+crypto_hash_walk_done int crypto_hash_walk_done ( s#crypto_hash_walk * , int )
+s##net_device
+s##sec_path
+s##rtattr
+s##skcipher_givcrypt_request
+s#ahash_request struct ahash_request { s#crypto_async_request base ; unsigned int nbytes ; s#scatterlist * src ; t#u8 * result ; void * priv ; void * __ctx [ ] __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; }
+crypto_hash_walk_first int crypto_hash_walk_first ( s#ahash_request * , s#crypto_hash_walk * )
+crypto_ahash_walk_first int crypto_ahash_walk_first ( s#ahash_request * , s#crypto_hash_walk * )
+s#crypto_ahash struct crypto_ahash { int ( * init ) ( s#ahash_request * ) ; int ( * update ) ( s#ahash_request * ) ; int ( * final ) ( s#ahash_request * ) ; int ( * finup ) ( s#ahash_request * ) ; int ( * digest ) ( s#ahash_request * ) ; int ( * export ) ( s#ahash_request * , void * ) ; int ( * import ) ( s#ahash_request * , const void * ) ; int ( * setkey ) ( s#crypto_ahash * , const t#u8 * , unsigned int ) ; unsigned int reqsize ; s#crypto_tfm base ; }
+crypto_ahash_setkey int crypto_ahash_setkey ( s#crypto_ahash * , const t#u8 * , unsigned int )
+crypto_ahash_final int crypto_ahash_final ( s#ahash_request * )
+crypto_ahash_finup int crypto_ahash_finup ( s#ahash_request * )
+crypto_ahash_digest int crypto_ahash_digest ( s#ahash_request * )
+crypto_ahash_type extern const s#crypto_type crypto_ahash_type
+crypto_alloc_ahash s#crypto_ahash * crypto_alloc_ahash ( const char * , t#u32 , t#u32 )
+crypto_has_ahash int crypto_has_ahash ( const char * , t#u32 , t#u32 )
+s#hash_alg_common struct hash_alg_common { unsigned int digestsize ; unsigned int statesize ; s#crypto_alg base ; }
+s#ahash_alg struct ahash_alg { int ( * init ) ( s#ahash_request * ) ; int ( * update ) ( s#ahash_request * ) ; int ( * final ) ( s#ahash_request * ) ; int ( * finup ) ( s#ahash_request * ) ; int ( * digest ) ( s#ahash_request * ) ; int ( * export ) ( s#ahash_request * , void * ) ; int ( * import ) ( s#ahash_request * , const void * ) ; int ( * setkey ) ( s#crypto_ahash * , const t#u8 * , unsigned int ) ; s#hash_alg_common halg ; }
+crypto_register_ahash int crypto_register_ahash ( s#ahash_alg * )
+crypto_unregister_ahash int crypto_unregister_ahash ( s#ahash_alg * )
+s#ahash_instance struct ahash_instance { s#ahash_alg alg ; }
+ahash_register_instance int ahash_register_instance ( s#crypto_template * , s#ahash_instance * )
+ahash_free_instance void ahash_free_instance ( s#crypto_instance * )
+s#crypto_ahash_spawn struct crypto_ahash_spawn { s#crypto_spawn base ; }
+crypto_init_ahash_spawn int crypto_init_ahash_spawn ( s#crypto_ahash_spawn * , s#hash_alg_common * , s#crypto_instance * )
+ahash_attr_alg s#hash_alg_common * ahash_attr_alg ( s#rtattr * , t#u32 , t#u32 )
+crypto_hash_alg_has_setkey t#bool crypto_hash_alg_has_setkey ( s#hash_alg_common * )
+
+/* crypto/akcipher.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##cgroup_namespace
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s##skcipher_givcrypt_request
+s#crypto_akcipher_spawn struct crypto_akcipher_spawn { s#crypto_spawn base ; }
+crypto_grab_akcipher int crypto_grab_akcipher ( s#crypto_akcipher_spawn * , const char * , t#u32 , t#u32 )
+s#crypto_akcipher struct crypto_akcipher { s#crypto_tfm base ; }
+crypto_alloc_akcipher s#crypto_akcipher * crypto_alloc_akcipher ( const char * , t#u32 , t#u32 )
+s#akcipher_request struct akcipher_request { s#crypto_async_request base ; s#scatterlist * src ; s#scatterlist * dst ; unsigned int src_len ; unsigned int dst_len ; void * __ctx [ ] __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; }
+s#akcipher_alg struct akcipher_alg { int ( * sign ) ( s#akcipher_request * ) ; int ( * verify ) ( s#akcipher_request * ) ; int ( * encrypt ) ( s#akcipher_request * ) ; int ( * decrypt ) ( s#akcipher_request * ) ; int ( * set_pub_key ) ( s#crypto_akcipher * , const void * , unsigned int ) ; int ( * set_priv_key ) ( s#crypto_akcipher * , const void * , unsigned int ) ; unsigned int ( * max_size ) ( s#crypto_akcipher * ) ; int ( * init ) ( s#crypto_akcipher * ) ; void ( * exit ) ( s#crypto_akcipher * ) ; unsigned int reqsize ; s#crypto_alg base ; }
+crypto_register_akcipher int crypto_register_akcipher ( s#akcipher_alg * )
+crypto_unregister_akcipher void crypto_unregister_akcipher ( s#akcipher_alg * )
+s#akcipher_instance struct akcipher_instance { void ( * free ) ( s#akcipher_instance * ) ; union { struct { char head [ __builtin_offsetof ( struct akcipher_alg , base ) ] ; s#crypto_instance base ; } s ; s#akcipher_alg alg ; } ; }
+akcipher_register_instance int akcipher_register_instance ( s#crypto_template * , s#akcipher_instance * )
+
+/* crypto/algapi.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##proc_dir_entry
+s##uevent_sock
+s##iw_handler_def
+s##iw_public_data
+s##sec_path
+s##neigh_parms
+s##netpoll_info
+s##neighbour
+s##udp_tunnel_info
+s##bpf_prog
+s##bpf_prog_offload_ops
+s##bpf_offloaded_map
+s##xdp_umem
+s##switchdev_ops
+s##l3mdev_ops
+s##ndisc_ops
+s##xfrm_state
+s##tls_crypto_info
+s##tls_context
+s##vlan_info
+s##dsa_switch_tree
+s##in_device
+s##dn_dev
+s##inet6_dev
+s##wireless_dev
+s##wpan_dev
+s##mpls_dev
+s##mini_Qdisc
+s##Qdisc
+s##nf_hook_entry
+s##cpu_rmap
+s##pcpu_dstats
+s##garp_port
+s##mrp_port
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##kmem_cache
+s##mem_cgroup
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##rtnl_link_ops
+s##phy_device
+s##sfp_bus
+s##prot_inuse
+s##ipv4_devconf
+s##fib_rules_ops
+s##fib_table
+s##inet_peer_base
+s##xt_table
+s##inet_hashinfo
+s##fib_notifier_ops
+s##ipv6_devconf
+s##rt6_info
+s##rt6_statistics
+s##fib6_table
+s##dst_entry
+s##seg6_pernet_data
+s##sctp_mib
+s##nf_queue_handler
+s##nf_logger
+s##ebt_table
+s##workqueue_struct
+s##nf_ct_event_notifier
+s##nf_exp_event_notifier
+s##nft_af_info
+s##net_generic
+s##netns_ipvs
+s##mpls_route
+s##dev_rcv_lists
+s##s_stats
+s##s_pstats
+s##bpf_prog_array
+s##cgroup_taskset
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##skcipher_givcrypt_request
+crypto_remove_spawns void crypto_remove_spawns ( s#crypto_alg * , s#list_head * , s#crypto_alg * )
+crypto_alg_tested void crypto_alg_tested ( const char * , int )
+crypto_remove_final void crypto_remove_final ( s#list_head * )
+crypto_register_alg int crypto_register_alg ( s#crypto_alg * )
+crypto_unregister_alg int crypto_unregister_alg ( s#crypto_alg * )
+crypto_register_algs int crypto_register_algs ( s#crypto_alg * , int )
+crypto_unregister_algs int crypto_unregister_algs ( s#crypto_alg * , int )
+crypto_register_template int crypto_register_template ( s#crypto_template * )
+crypto_unregister_template void crypto_unregister_template ( s#crypto_template * )
+crypto_lookup_template s#crypto_template * crypto_lookup_template ( const char * )
+crypto_register_instance int crypto_register_instance ( s#crypto_template * , s#crypto_instance * )
+crypto_unregister_instance int crypto_unregister_instance ( s#crypto_instance * )
+crypto_init_spawn int crypto_init_spawn ( s#crypto_spawn * , s#crypto_alg * , s#crypto_instance * , t#u32 )
+crypto_init_spawn2 int crypto_init_spawn2 ( s#crypto_spawn * , s#crypto_alg * , s#crypto_instance * , const s#crypto_type * )
+crypto_grab_spawn int crypto_grab_spawn ( s#crypto_spawn * , const char * , t#u32 , t#u32 )
+crypto_drop_spawn void crypto_drop_spawn ( s#crypto_spawn * )
+crypto_spawn_tfm s#crypto_tfm * crypto_spawn_tfm ( s#crypto_spawn * , t#u32 , t#u32 )
+crypto_spawn_tfm2 void * crypto_spawn_tfm2 ( s#crypto_spawn * )
+crypto_register_notifier int crypto_register_notifier ( s#notifier_block * )
+crypto_unregister_notifier int crypto_unregister_notifier ( s#notifier_block * )
+s#crypto_attr_type struct crypto_attr_type { t#u32 type ; t#u32 mask ; }
+crypto_get_attr_type s#crypto_attr_type * crypto_get_attr_type ( s#rtattr * * )
+crypto_check_attr_type int crypto_check_attr_type ( s#rtattr * * , t#u32 )
+crypto_attr_alg_name const char * crypto_attr_alg_name ( s#rtattr * )
+crypto_attr_alg2 s#crypto_alg * crypto_attr_alg2 ( s#rtattr * , const s#crypto_type * , t#u32 , t#u32 )
+crypto_attr_u32 int crypto_attr_u32 ( s#rtattr * , t#u32 * )
+crypto_inst_setname int crypto_inst_setname ( s#crypto_instance * , const char * , s#crypto_alg * )
+crypto_alloc_instance2 void * crypto_alloc_instance2 ( const char * , s#crypto_alg * , unsigned int )
+crypto_alloc_instance s#crypto_instance * crypto_alloc_instance ( const char * , s#crypto_alg * )
+s#crypto_queue struct crypto_queue { s#list_head list ; s#list_head * backlog ; unsigned int qlen ; unsigned int max_qlen ; }
+crypto_init_queue void crypto_init_queue ( s#crypto_queue * , unsigned int )
+crypto_enqueue_request int crypto_enqueue_request ( s#crypto_queue * , s#crypto_async_request * )
+crypto_dequeue_request s#crypto_async_request * crypto_dequeue_request ( s#crypto_queue * )
+crypto_tfm_in_queue int crypto_tfm_in_queue ( s#crypto_queue * , s#crypto_tfm * )
+crypto_inc void crypto_inc ( t#u8 * , unsigned int )
+__crypto_xor void __crypto_xor ( t#u8 * , const t#u8 * , const t#u8 * , unsigned int )
+crypto_alg_extsize unsigned int crypto_alg_extsize ( s#crypto_alg * )
+crypto_type_has_alg int crypto_type_has_alg ( const char * , const s#crypto_type * , t#u32 , t#u32 )
+
+/* crypto/api.o */
+crypto_alg_list extern s#list_head crypto_alg_list
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##posix_acl
+s##vfsmount
+s##sock
+s##bio_vec
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##poll_table_struct
+s##pid_namespace
+s##nlm_lockowner
+s##nfs4_lock_state
+s##user_namespace
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##mem_cgroup
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##backing_dev_info
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##tty_struct
+s##taskstats
+s##tty_audit_buf
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+crypto_alg_sem extern s#rw_semaphore crypto_alg_sem
+s#blocking_notifier_head struct blocking_notifier_head { s#rw_semaphore rwsem ; s#notifier_block * head ; }
+crypto_chain extern s#blocking_notifier_head crypto_chain
+s##net_device
+s##sec_path
+s##rtattr
+s##skcipher_givcrypt_request
+crypto_mod_get s#crypto_alg * crypto_mod_get ( s#crypto_alg * )
+crypto_mod_put void crypto_mod_put ( s#crypto_alg * )
+s#crypto_larval struct crypto_larval { s#crypto_alg alg ; s#crypto_alg * adult ; s#completion completion ; t#u32 mask ; }
+crypto_larval_alloc s#crypto_larval * crypto_larval_alloc ( const char * , t#u32 , t#u32 )
+crypto_larval_kill void crypto_larval_kill ( s#crypto_alg * )
+crypto_alg_lookup s#crypto_alg * crypto_alg_lookup ( const char * , t#u32 , t#u32 )
+crypto_larval_lookup s#crypto_alg * crypto_larval_lookup ( const char * , t#u32 , t#u32 )
+crypto_probing_notify int crypto_probing_notify ( unsigned long , void * )
+crypto_alg_mod_lookup s#crypto_alg * crypto_alg_mod_lookup ( const char * , t#u32 , t#u32 )
+crypto_shoot_alg void crypto_shoot_alg ( s#crypto_alg * )
+__crypto_alloc_tfm s#crypto_tfm * __crypto_alloc_tfm ( s#crypto_alg * , t#u32 , t#u32 )
+crypto_alloc_base s#crypto_tfm * crypto_alloc_base ( const char * , t#u32 , t#u32 )
+crypto_create_tfm void * crypto_create_tfm ( s#crypto_alg * , const s#crypto_type * )
+crypto_find_alg s#crypto_alg * crypto_find_alg ( const char * , const s#crypto_type * , t#u32 , t#u32 )
+crypto_alloc_tfm void * crypto_alloc_tfm ( const char * , const s#crypto_type * , t#u32 , t#u32 )
+crypto_destroy_tfm void crypto_destroy_tfm ( void * , s#crypto_tfm * )
+crypto_has_alg int crypto_has_alg ( const char * , t#u32 , t#u32 )
+crypto_req_done void crypto_req_done ( s#crypto_async_request * , int )
+
+/* crypto/asymmetric_keys/asymmetric_type.o */
+E#NR__KEY_BEING_USED_FOR 6
+key_being_used_for extern const char * const key_being_used_for [ E#NR__KEY_BEING_USED_FOR ]
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##vm_operations_struct
+s##vfsmount
+s##posix_acl
+s##sock
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##exception_table_entry
+s##module_sect_attrs
+s##module_notes_attrs
+s##tracepoint
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##iov_iter
+s##swap_info_struct
+s##dev_pagemap
+s##kmem_cache
+s##mem_cgroup
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##backing_dev_info
+s##mtd_info
+s##workqueue_struct
+s##user_namespace
+s##bdi_writeback
+s##poll_table_struct
+s##pid_namespace
+s##nlm_lockowner
+s##nfs4_lock_state
+s##pipe_inode_info
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##key_user
+s##assoc_array_ptr
+s#asymmetric_key_id struct asymmetric_key_id { unsigned short len ; unsigned char data [ ] ; }
+find_asymmetric_key extern s#key * find_asymmetric_key ( s#key * , const s#asymmetric_key_id * , const s#asymmetric_key_id * , t#bool )
+asymmetric_key_generate_id extern s#asymmetric_key_id * asymmetric_key_generate_id ( const void * , t#size_t , const void * , t#size_t )
+asymmetric_key_id_same extern t#bool asymmetric_key_id_same ( const s#asymmetric_key_id * , const s#asymmetric_key_id * )
+asymmetric_key_id_partial extern t#bool asymmetric_key_id_partial ( const s#asymmetric_key_id * , const s#asymmetric_key_id * )
+key_type_asymmetric extern s#key_type key_type_asymmetric
+s#asymmetric_key_parser struct asymmetric_key_parser { s#list_head link ; s#module * owner ; const char * name ; int ( * parse ) ( s#key_preparsed_payload * ) ; }
+register_asymmetric_key_parser extern int register_asymmetric_key_parser ( s#asymmetric_key_parser * )
+unregister_asymmetric_key_parser extern void unregister_asymmetric_key_parser ( s#asymmetric_key_parser * )
+
+/* crypto/asymmetric_keys/pkcs7_parser.o */
+s#public_key struct public_key { void * key ; t#u32 keylen ; const char * id_type ; const char * pkey_algo ; }
+s#public_key_signature struct public_key_signature { s#asymmetric_key_id * auth_ids [ 2 ] ; t#u8 * s ; t#u32 s_size ; t#u8 * digest ; t#u8 digest_size ; const char * pkey_algo ; const char * hash_algo ; }
+s#x509_certificate struct x509_certificate { s#x509_certificate * next ; s#x509_certificate * signer ; s#public_key * pub ; s#public_key_signature * sig ; char * issuer ; char * subject ; s#asymmetric_key_id * id ; s#asymmetric_key_id * skid ; t#time64_t valid_from ; t#time64_t valid_to ; const void * tbs ; unsigned tbs_size ; unsigned raw_sig_size ; const void * raw_sig ; const void * raw_serial ; unsigned raw_serial_size ; unsigned raw_issuer_size ; const void * raw_issuer ; const void * raw_subject ; unsigned raw_subject_size ; unsigned raw_skid_size ; const void * raw_skid ; unsigned index ; t#bool seen ; t#bool verified ; t#bool self_signed ; t#bool unsupported_key ; t#bool unsupported_sig ; t#bool blacklisted ; }
+s#pkcs7_signed_info struct pkcs7_signed_info { s#pkcs7_signed_info * next ; s#x509_certificate * signer ; unsigned index ; t#bool unsupported_crypto ; t#bool blacklisted ; const void * msgdigest ; unsigned msgdigest_len ; unsigned authattrs_len ; const void * authattrs ; unsigned long aa_set ; t#time64_t signing_time ; s#public_key_signature * sig ; }
+e#OID enum OID { OID_id_dsa_with_sha1 , OID_id_dsa , OID_id_ecdsa_with_sha1 , OID_id_ecPublicKey , OID_rsaEncryption , OID_md2WithRSAEncryption , OID_md3WithRSAEncryption , OID_md4WithRSAEncryption , OID_sha1WithRSAEncryption , OID_sha256WithRSAEncryption , OID_sha384WithRSAEncryption , OID_sha512WithRSAEncryption , OID_sha224WithRSAEncryption , OID_data , OID_signed_data , OID_email_address , OID_contentType , OID_messageDigest , OID_signingTime , OID_smimeCapabilites , OID_smimeAuthenticatedAttrs , OID_md2 , OID_md4 , OID_md5 , OID_msIndirectData , OID_msStatementType , OID_msSpOpusInfo , OID_msPeImageDataObjId , OID_msIndividualSPKeyPurpose , OID_msOutlookExpress , OID_certAuthInfoAccess , OID_sha1 , OID_sha256 , OID_sha384 , OID_sha512 , OID_sha224 , OID_commonName , OID_surname , OID_countryName , OID_locality , OID_stateOrProvinceName , OID_organizationName , OID_organizationUnitName , OID_title , OID_description , OID_name , OID_givenName , OID_initials , OID_generationalQualifier , OID_subjectKeyIdentifier , OID_keyUsage , OID_subjectAltName , OID_issuerAltName , OID_basicConstraints , OID_crlDistributionPoints , OID_certPolicies , OID_authorityKeyIdentifier , OID_extKeyUsage , OID__NR }
+s#pkcs7_message struct pkcs7_message { s#x509_certificate * certs ; s#x509_certificate * crl ; s#pkcs7_signed_info * signed_infos ; t#u8 version ; t#bool have_authattrs ; e#OID data_type ; t#size_t data_len ; t#size_t data_hdrlen ; const void * data ; }
+pkcs7_free_message extern void pkcs7_free_message ( s#pkcs7_message * )
+pkcs7_parse_message extern s#pkcs7_message * pkcs7_parse_message ( const void * , t#size_t )
+pkcs7_get_content_data extern int pkcs7_get_content_data ( const s#pkcs7_message * , const void * * , t#size_t * , t#size_t * )
+
+/* crypto/asymmetric_keys/pkcs7_trust.o */
+s##task_struct
+s##key_user
+s##seq_file
+s##assoc_array_ptr
+pkcs7_validate_trust extern int pkcs7_validate_trust ( s#pkcs7_message * , s#key * )
+
+/* crypto/asymmetric_keys/pkcs7_verify.o */
+pkcs7_verify extern int pkcs7_verify ( s#pkcs7_message * , e#key_being_used_for )
+
+/* crypto/asymmetric_keys/public_key.o */
+public_key_free extern void public_key_free ( s#public_key * )
+public_key_verify_signature int public_key_verify_signature ( const s#public_key * , const s#public_key_signature * )
+s##sock
+s##vfsmount
+s##posix_acl
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##key_user
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##device
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s#asymmetric_key_subtype struct asymmetric_key_subtype { s#module * owner ; const char * name ; unsigned short name_len ; void ( * describe ) ( const s#key * , s#seq_file * ) ; void ( * destroy ) ( void * , void * ) ; int ( * verify_signature ) ( const s#key * , const s#public_key_signature * ) ; }
+public_key_subtype extern s#asymmetric_key_subtype public_key_subtype
+
+/* crypto/asymmetric_keys/signature.o */
+public_key_signature_free extern void public_key_signature_free ( s#public_key_signature * )
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##vm_operations_struct
+s##vfsmount
+s##posix_acl
+s##module
+s##writeback_control
+s##kstatfs
+s##iov_iter
+s##swap_info_struct
+s##dev_pagemap
+s##kmem_cache
+s##mem_cgroup
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##backing_dev_info
+s##mtd_info
+s##workqueue_struct
+s##user_namespace
+s##bdi_writeback
+s##poll_table_struct
+s##pid_namespace
+s##nlm_lockowner
+s##nfs4_lock_state
+s##pipe_inode_info
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##key_user
+s##assoc_array_ptr
+verify_signature extern int verify_signature ( const s#key * , const s#public_key_signature * )
+
+/* crypto/asymmetric_keys/x509_cert_parser.o */
+x509_free_certificate extern void x509_free_certificate ( s#x509_certificate * )
+x509_cert_parse extern s#x509_certificate * x509_cert_parse ( const void * , t#size_t )
+x509_decode_time extern int x509_decode_time ( t#time64_t * , t#size_t , unsigned char , const unsigned char * , t#size_t )
+
+/* crypto/async_tx/async_memcpy.o */
+t#dma_cookie_t typedef t#s32 dma_cookie_t
+e#dma_ctrl_flags enum dma_ctrl_flags { DMA_PREP_INTERRUPT = ( 1 << 0 ) , DMA_CTRL_ACK = ( 1 << 1 ) , DMA_PREP_PQ_DISABLE_P = ( 1 << 2 ) , DMA_PREP_PQ_DISABLE_Q = ( 1 << 3 ) , DMA_PREP_CONTINUE = ( 1 << 4 ) , DMA_PREP_FENCE = ( 1 << 5 ) , DMA_CTRL_REUSE = ( 1 << 6 ) , }
+t#dma_filter_fn typedef t#bool ( * dma_filter_fn ) ( s#dma_chan * , void * )
+s#dma_slave_map struct dma_slave_map { const char * devname ; const char * slave ; void * param ; }
+s#dma_filter struct dma_filter { t#dma_filter_fn fn ; int mapcnt ; const s#dma_slave_map * map ; }
+E#DMA_TX_TYPE_END 14
+t#dma_cap_mask_t typedef struct { unsigned long bits [ ( ( ( E#DMA_TX_TYPE_END ) + ( 8 * sizeof ( long ) ) - 1 ) / ( 8 * sizeof ( long ) ) ) ] ; } dma_cap_mask_t
+e#dmaengine_alignment enum dmaengine_alignment { DMAENGINE_ALIGN_1_BYTE = 0 , DMAENGINE_ALIGN_2_BYTES = 1 , DMAENGINE_ALIGN_4_BYTES = 2 , DMAENGINE_ALIGN_8_BYTES = 3 , DMAENGINE_ALIGN_16_BYTES = 4 , DMAENGINE_ALIGN_32_BYTES = 5 , DMAENGINE_ALIGN_64_BYTES = 6 , }
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##bio_vec
+s##swap_info_struct
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+e#dma_residue_granularity enum dma_residue_granularity { DMA_RESIDUE_GRANULARITY_DESCRIPTOR = 0 , DMA_RESIDUE_GRANULARITY_SEGMENT = 1 , DMA_RESIDUE_GRANULARITY_BURST = 2 , }
+E#SUM_CHECK_P 0
+E#SUM_CHECK_Q 1
+e#sum_check_flags enum sum_check_flags { SUM_CHECK_P_RESULT = ( 1 << E#SUM_CHECK_P ) , SUM_CHECK_Q_RESULT = ( 1 << E#SUM_CHECK_Q ) , }
+e#dma_transfer_direction enum dma_transfer_direction { DMA_MEM_TO_MEM , DMA_MEM_TO_DEV , DMA_DEV_TO_MEM , DMA_DEV_TO_DEV , DMA_TRANS_NONE , }
+s#data_chunk struct data_chunk { t#size_t size ; t#size_t icg ; t#size_t dst_icg ; t#size_t src_icg ; }
+s#dma_interleaved_template struct dma_interleaved_template { t#dma_addr_t src_start ; t#dma_addr_t dst_start ; e#dma_transfer_direction dir ; t#bool src_inc ; t#bool dst_inc ; t#bool src_sgl ; t#bool dst_sgl ; t#size_t numf ; t#size_t frame_size ; s#data_chunk sgl [ 0 ] ; }
+e#dma_slave_buswidth enum dma_slave_buswidth { DMA_SLAVE_BUSWIDTH_UNDEFINED = 0 , DMA_SLAVE_BUSWIDTH_1_BYTE = 1 , DMA_SLAVE_BUSWIDTH_2_BYTES = 2 , DMA_SLAVE_BUSWIDTH_3_BYTES = 3 , DMA_SLAVE_BUSWIDTH_4_BYTES = 4 , DMA_SLAVE_BUSWIDTH_8_BYTES = 8 , DMA_SLAVE_BUSWIDTH_16_BYTES = 16 , DMA_SLAVE_BUSWIDTH_32_BYTES = 32 , DMA_SLAVE_BUSWIDTH_64_BYTES = 64 , }
+s#dma_slave_config struct dma_slave_config { e#dma_transfer_direction direction ; t#phys_addr_t src_addr ; t#phys_addr_t dst_addr ; e#dma_slave_buswidth src_addr_width ; e#dma_slave_buswidth dst_addr_width ; t#u32 src_maxburst ; t#u32 dst_maxburst ; t#u32 src_port_window_size ; t#u32 dst_port_window_size ; t#bool device_fc ; unsigned int slave_id ; }
+e#dma_status enum dma_status { DMA_COMPLETE , DMA_IN_PROGRESS , DMA_PAUSED , DMA_ERROR , }
+s#dma_tx_state struct dma_tx_state { t#dma_cookie_t last ; t#dma_cookie_t used ; t#u32 residue ; }
+s#dma_device struct dma_device { unsigned int chancnt ; unsigned int privatecnt ; s#list_head channels ; s#list_head global_node ; s#dma_filter filter ; t#dma_cap_mask_t cap_mask ; unsigned short max_xor ; unsigned short max_pq ; e#dmaengine_alignment copy_align ; e#dmaengine_alignment xor_align ; e#dmaengine_alignment pq_align ; e#dmaengine_alignment fill_align ; int dev_id ; s#device * dev ; t#u32 src_addr_widths ; t#u32 dst_addr_widths ; t#u32 directions ; t#u32 max_burst ; t#bool descriptor_reuse ; e#dma_residue_granularity residue_granularity ; int ( * device_alloc_chan_resources ) ( s#dma_chan * ) ; void ( * device_free_chan_resources ) ( s#dma_chan * ) ; s#dma_async_tx_descriptor * ( * device_prep_dma_memcpy ) ( s#dma_chan * , t#dma_addr_t , t#dma_addr_t , t#size_t , unsigned long ) ; s#dma_async_tx_descriptor * ( * device_prep_dma_xor ) ( s#dma_chan * , t#dma_addr_t , t#dma_addr_t * , unsigned int , t#size_t , unsigned long ) ; s#dma_async_tx_descriptor * ( * device_prep_dma_xor_val ) ( s#dma_chan * , t#dma_addr_t * , unsigned int , t#size_t , e#sum_check_flags * , unsigned long ) ; s#dma_async_tx_descriptor * ( * device_prep_dma_pq ) ( s#dma_chan * , t#dma_addr_t * , t#dma_addr_t * , unsigned int , const unsigned char * , t#size_t , unsigned long ) ; s#dma_async_tx_descriptor * ( * device_prep_dma_pq_val ) ( s#dma_chan * , t#dma_addr_t * , t#dma_addr_t * , unsigned int , const unsigned char * , t#size_t , e#sum_check_flags * , unsigned long ) ; s#dma_async_tx_descriptor * ( * device_prep_dma_memset ) ( s#dma_chan * , t#dma_addr_t , int , t#size_t , unsigned long ) ; s#dma_async_tx_descriptor * ( * device_prep_dma_memset_sg ) ( s#dma_chan * , s#scatterlist * , unsigned int , int , unsigned long ) ; s#dma_async_tx_descriptor * ( * device_prep_dma_interrupt ) ( s#dma_chan * , unsigned long ) ; s#dma_async_tx_descriptor * ( * device_prep_dma_sg ) ( s#dma_chan * , s#scatterlist * , unsigned int , s#scatterlist * , unsigned int , unsigned long ) ; s#dma_async_tx_descriptor * ( * device_prep_slave_sg ) ( s#dma_chan * , s#scatterlist * , unsigned int , e#dma_transfer_direction , unsigned long , void * ) ; s#dma_async_tx_descriptor * ( * device_prep_dma_cyclic ) ( s#dma_chan * , t#dma_addr_t , t#size_t , t#size_t , e#dma_transfer_direction , unsigned long ) ; s#dma_async_tx_descriptor * ( * device_prep_interleaved_dma ) ( s#dma_chan * , s#dma_interleaved_template * , unsigned long ) ; s#dma_async_tx_descriptor * ( * device_prep_dma_imm_data ) ( s#dma_chan * , t#dma_addr_t , t#u64 , unsigned long ) ; int ( * device_config ) ( s#dma_chan * , s#dma_slave_config * ) ; int ( * device_pause ) ( s#dma_chan * ) ; int ( * device_resume ) ( s#dma_chan * ) ; int ( * device_terminate_all ) ( s#dma_chan * ) ; void ( * device_synchronize ) ( s#dma_chan * ) ; e#dma_status ( * device_tx_status ) ( s#dma_chan * , t#dma_cookie_t , s#dma_tx_state * ) ; void ( * device_issue_pending ) ( s#dma_chan * ) ; }
+s#dma_chan_dev struct dma_chan_dev { s#dma_chan * chan ; s#device device ; int dev_id ; t#atomic_t * idr_ref ; }
+s#dma_chan_percpu struct dma_chan_percpu { unsigned long memcpy_count ; unsigned long bytes_transferred ; }
+s#dma_router struct dma_router { s#device * dev ; void ( * route_free ) ( s#device * , void * ) ; }
+s#dma_chan struct dma_chan { s#dma_device * device ; t#dma_cookie_t cookie ; t#dma_cookie_t completed_cookie ; int chan_id ; s#dma_chan_dev * dev ; s#list_head device_node ; s#dma_chan_percpu * local ; int client_count ; int table_count ; s#dma_router * router ; void * route_data ; void * private ; }
+t#dma_async_tx_callback typedef void ( * dma_async_tx_callback ) ( void * )
+e#dmaengine_tx_result enum dmaengine_tx_result { DMA_TRANS_NOERROR = 0 , DMA_TRANS_READ_FAILED , DMA_TRANS_WRITE_FAILED , DMA_TRANS_ABORTED , }
+s#dmaengine_result struct dmaengine_result { e#dmaengine_tx_result result ; t#u32 residue ; }
+t#dma_async_tx_callback_result typedef void ( * dma_async_tx_callback_result ) ( void * , const s#dmaengine_result * )
+s#dmaengine_unmap_data struct dmaengine_unmap_data { t#u8 map_cnt ; t#u8 to_cnt ; t#u8 from_cnt ; t#u8 bidi_cnt ; s#device * dev ; s#kref kref ; t#size_t len ; t#dma_addr_t addr [ 0 ] ; }
+s#dma_async_tx_descriptor struct dma_async_tx_descriptor { t#dma_cookie_t cookie ; e#dma_ctrl_flags flags ; t#dma_addr_t phys ; s#dma_chan * chan ; t#dma_cookie_t ( * tx_submit ) ( s#dma_async_tx_descriptor * ) ; int ( * desc_free ) ( s#dma_async_tx_descriptor * ) ; t#dma_async_tx_callback callback ; t#dma_async_tx_callback_result callback_result ; void * callback_param ; s#dmaengine_unmap_data * unmap ; s#dma_async_tx_descriptor * next ; s#dma_async_tx_descriptor * parent ; t#spinlock_t lock ; }
+e#async_tx_flags enum async_tx_flags { ASYNC_TX_XOR_ZERO_DST = ( 1 << 0 ) , ASYNC_TX_XOR_DROP_DST = ( 1 << 1 ) , ASYNC_TX_ACK = ( 1 << 2 ) , ASYNC_TX_FENCE = ( 1 << 3 ) , ASYNC_TX_PQ_XOR_DST = ( 1 << 4 ) , }
+s#async_submit_ctl struct async_submit_ctl { e#async_tx_flags flags ; s#dma_async_tx_descriptor * depend_tx ; t#dma_async_tx_callback cb_fn ; void * cb_param ; void * scribble ; }
+async_memcpy s#dma_async_tx_descriptor * async_memcpy ( s#page * , s#page * , unsigned int , unsigned int , t#size_t , s#async_submit_ctl * )
+
+/* crypto/async_tx/async_pq.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##pipe_inode_info
+s##hd_geometry
+s##pr_ops
+s##blk_mq_ctx
+s##kmem_cache
+s##blkcg_gq
+s##blk_mq_hw_ctx
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_mq_ops
+s##blk_stat_callback
+s##blk_trace
+s##blk_flush_queue
+s##bsg_job
+s##throtl_data
+s##blk_mq_tag_set
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+async_gen_syndrome s#dma_async_tx_descriptor * async_gen_syndrome ( s#page * * , unsigned int , int , t#size_t , s#async_submit_ctl * )
+async_syndrome_val s#dma_async_tx_descriptor * async_syndrome_val ( s#page * * , unsigned int , int , t#size_t , e#sum_check_flags * , s#page * , s#async_submit_ctl * )
+
+/* crypto/async_tx/async_raid6_recov.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##pipe_inode_info
+s##hd_geometry
+s##pr_ops
+s##blk_mq_ctx
+s##kmem_cache
+s##blkcg_gq
+s##blk_mq_hw_ctx
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_mq_ops
+s##blk_stat_callback
+s##blk_trace
+s##blk_flush_queue
+s##bsg_job
+s##throtl_data
+s##blk_mq_tag_set
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+async_raid6_2data_recov s#dma_async_tx_descriptor * async_raid6_2data_recov ( int , t#size_t , int , int , s#page * * , s#async_submit_ctl * )
+async_raid6_datap_recov s#dma_async_tx_descriptor * async_raid6_datap_recov ( int , t#size_t , int , s#page * * , s#async_submit_ctl * )
+
+/* crypto/async_tx/async_tx.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##bio_vec
+s##swap_info_struct
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+e#dma_transaction_type enum dma_transaction_type { DMA_MEMCPY , DMA_XOR , DMA_PQ , DMA_XOR_VAL , DMA_PQ_VAL , DMA_MEMSET , DMA_MEMSET_SG , DMA_INTERRUPT , DMA_SG , DMA_PRIVATE , DMA_ASYNC_TX , DMA_SLAVE , DMA_CYCLIC , DMA_INTERLEAVE , DMA_TX_TYPE_END , }
+__async_tx_find_channel s#dma_chan * __async_tx_find_channel ( s#async_submit_ctl * , e#dma_transaction_type )
+async_tx_submit void async_tx_submit ( s#dma_chan * , s#dma_async_tx_descriptor * , s#async_submit_ctl * )
+async_trigger_callback s#dma_async_tx_descriptor * async_trigger_callback ( s#async_submit_ctl * )
+async_tx_quiesce void async_tx_quiesce ( s#dma_async_tx_descriptor * * )
+
+/* crypto/async_tx/async_xor.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##bio_vec
+s##swap_info_struct
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+async_xor s#dma_async_tx_descriptor * async_xor ( s#page * , s#page * * , unsigned int , int , t#size_t , s#async_submit_ctl * )
+async_xor_val s#dma_async_tx_descriptor * async_xor_val ( s#page * , s#page * * , unsigned int , int , t#size_t , e#sum_check_flags * , s#async_submit_ctl * )
+
+/* crypto/authenc.o */
+s#crypto_authenc_keys struct crypto_authenc_keys { const t#u8 * authkey ; const t#u8 * enckey ; unsigned int authkeylen ; unsigned int enckeylen ; }
+crypto_authenc_extractkeys int crypto_authenc_extractkeys ( s#crypto_authenc_keys * , const t#u8 * , unsigned int )
+
+/* crypto/blkcipher.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##cgroup_namespace
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s#blkcipher_walk struct blkcipher_walk { union { struct { s#page * page ; unsigned long offset ; } phys ; struct { t#u8 * page ; t#u8 * addr ; } virt ; } src , dst ; s#scatter_walk in ; unsigned int nbytes ; s#scatter_walk out ; unsigned int total ; void * page ; t#u8 * buffer ; t#u8 * iv ; unsigned int ivsize ; int flags ; unsigned int walk_blocksize ; unsigned int cipher_blocksize ; unsigned int alignmask ; }
+blkcipher_walk_done int blkcipher_walk_done ( s#blkcipher_desc * , s#blkcipher_walk * , int )
+blkcipher_walk_virt int blkcipher_walk_virt ( s#blkcipher_desc * , s#blkcipher_walk * )
+blkcipher_walk_phys int blkcipher_walk_phys ( s#blkcipher_desc * , s#blkcipher_walk * )
+blkcipher_walk_virt_block int blkcipher_walk_virt_block ( s#blkcipher_desc * , s#blkcipher_walk * , unsigned int )
+blkcipher_aead_walk_virt_block int blkcipher_aead_walk_virt_block ( s#blkcipher_desc * , s#blkcipher_walk * , s#crypto_aead * , unsigned int )
+crypto_blkcipher_type extern const s#crypto_type crypto_blkcipher_type
+
+/* crypto/blowfish_common.o */
+s##scatterlist
+s##crypto_type
+s##skcipher_givcrypt_request
+s##sock
+s##seq_file
+s##vfsmount
+s##posix_acl
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##cred
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##device
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+blowfish_setkey int blowfish_setkey ( s#crypto_tfm * , const t#u8 * , unsigned int )
+
+/* crypto/cast5_generic.o */
+s#cast5_ctx struct cast5_ctx { t#u32 Km [ 16 ] ; t#u8 Kr [ 16 ] ; int rr ; }
+__cast5_encrypt void __cast5_encrypt ( s#cast5_ctx * , t#u8 * , const t#u8 * )
+__cast5_decrypt void __cast5_decrypt ( s#cast5_ctx * , t#u8 * , const t#u8 * )
+s##scatterlist
+s##crypto_type
+s##skcipher_givcrypt_request
+s##sock
+s##seq_file
+s##file
+s##vm_operations_struct
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##cred
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##pipe_inode_info
+s##address_space
+s##dev_pagemap
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##tracepoint
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+cast5_setkey int cast5_setkey ( s#crypto_tfm * , const t#u8 * , unsigned int )
+
+/* crypto/cast6_generic.o */
+s#cast6_ctx struct cast6_ctx { t#u32 Km [ 12 ] [ 4 ] ; t#u8 Kr [ 12 ] [ 4 ] ; }
+__cast6_setkey int __cast6_setkey ( s#cast6_ctx * , const t#u8 * , unsigned int , t#u32 * )
+s##scatterlist
+s##crypto_type
+s##skcipher_givcrypt_request
+s##sock
+s##seq_file
+s##file
+s##vm_operations_struct
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##cred
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##pipe_inode_info
+s##address_space
+s##dev_pagemap
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##tracepoint
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+cast6_setkey int cast6_setkey ( s#crypto_tfm * , const t#u8 * , unsigned int )
+__cast6_encrypt void __cast6_encrypt ( s#cast6_ctx * , t#u8 * , const t#u8 * )
+__cast6_decrypt void __cast6_decrypt ( s#cast6_ctx * , t#u8 * , const t#u8 * )
+
+/* crypto/cast_common.o */
+cast_s1 extern const t#u32 cast_s1 [ 256 ]
+cast_s2 extern const t#u32 cast_s2 [ 256 ]
+cast_s3 extern const t#u32 cast_s3 [ 256 ]
+cast_s4 extern const t#u32 cast_s4 [ 256 ]
+
+/* crypto/chacha20_generic.o */
+s#chacha20_ctx struct chacha20_ctx { t#u32 key [ 8 ] ; }
+crypto_chacha20_init void crypto_chacha20_init ( t#u32 * , s#chacha20_ctx * , t#u8 * )
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s#skcipher_request struct skcipher_request { unsigned int cryptlen ; t#u8 * iv ; s#scatterlist * src ; s#scatterlist * dst ; s#crypto_async_request base ; void * __ctx [ ] __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; }
+s#crypto_skcipher struct crypto_skcipher { int ( * setkey ) ( s#crypto_skcipher * , const t#u8 * , unsigned int ) ; int ( * encrypt ) ( s#skcipher_request * ) ; int ( * decrypt ) ( s#skcipher_request * ) ; unsigned int ivsize ; unsigned int reqsize ; unsigned int keysize ; s#crypto_tfm base ; }
+crypto_chacha20_setkey int crypto_chacha20_setkey ( s#crypto_skcipher * , const t#u8 * , unsigned int )
+crypto_chacha20_crypt int crypto_chacha20_crypt ( s#skcipher_request * )
+
+/* crypto/crct10dif_common.o */
+crc_t10dif_generic extern t#__u16 crc_t10dif_generic ( t#__u16 , const unsigned char * , t#size_t )
+
+/* crypto/cryptd.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s#cryptd_ablkcipher struct cryptd_ablkcipher { s#crypto_ablkcipher base ; }
+cryptd_alloc_ablkcipher s#cryptd_ablkcipher * cryptd_alloc_ablkcipher ( const char * , t#u32 , t#u32 )
+cryptd_ablkcipher_child s#crypto_blkcipher * cryptd_ablkcipher_child ( s#cryptd_ablkcipher * )
+cryptd_ablkcipher_queued t#bool cryptd_ablkcipher_queued ( s#cryptd_ablkcipher * )
+cryptd_free_ablkcipher void cryptd_free_ablkcipher ( s#cryptd_ablkcipher * )
+s#cryptd_skcipher struct cryptd_skcipher { s#crypto_skcipher base ; }
+cryptd_alloc_skcipher s#cryptd_skcipher * cryptd_alloc_skcipher ( const char * , t#u32 , t#u32 )
+cryptd_skcipher_child s#crypto_skcipher * cryptd_skcipher_child ( s#cryptd_skcipher * )
+cryptd_skcipher_queued t#bool cryptd_skcipher_queued ( s#cryptd_skcipher * )
+cryptd_free_skcipher void cryptd_free_skcipher ( s#cryptd_skcipher * )
+s#cryptd_ahash struct cryptd_ahash { s#crypto_ahash base ; }
+cryptd_alloc_ahash s#cryptd_ahash * cryptd_alloc_ahash ( const char * , t#u32 , t#u32 )
+s#crypto_shash struct crypto_shash { unsigned int descsize ; s#crypto_tfm base ; }
+cryptd_ahash_child s#crypto_shash * cryptd_ahash_child ( s#cryptd_ahash * )
+s#shash_desc struct shash_desc { s#crypto_shash * tfm ; t#u32 flags ; void * __ctx [ ] __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; }
+cryptd_shash_desc s#shash_desc * cryptd_shash_desc ( s#ahash_request * )
+cryptd_ahash_queued t#bool cryptd_ahash_queued ( s#cryptd_ahash * )
+cryptd_free_ahash void cryptd_free_ahash ( s#cryptd_ahash * )
+s#cryptd_aead struct cryptd_aead { s#crypto_aead base ; }
+cryptd_alloc_aead s#cryptd_aead * cryptd_alloc_aead ( const char * , t#u32 , t#u32 )
+cryptd_aead_child s#crypto_aead * cryptd_aead_child ( s#cryptd_aead * )
+cryptd_aead_queued t#bool cryptd_aead_queued ( s#cryptd_aead * )
+cryptd_free_aead void cryptd_free_aead ( s#cryptd_aead * )
+
+/* crypto/crypto_engine.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s##skcipher_givcrypt_request
+t#kthread_work_func_t typedef void ( * kthread_work_func_t ) ( s#kthread_work * )
+s#kthread_work struct kthread_work { s#list_head node ; t#kthread_work_func_t func ; s#kthread_worker * worker ; int canceling ; }
+s#kthread_worker struct kthread_worker { unsigned int flags ; t#spinlock_t lock ; s#list_head work_list ; s#list_head delayed_work_list ; s#task_struct * task ; s#kthread_work * current_work ; }
+s#crypto_engine struct crypto_engine { char name [ 30 ] ; t#bool idling ; t#bool busy ; t#bool running ; t#bool cur_req_prepared ; s#list_head list ; t#spinlock_t queue_lock ; s#crypto_queue queue ; t#bool rt ; int ( * prepare_crypt_hardware ) ( s#crypto_engine * ) ; int ( * unprepare_crypt_hardware ) ( s#crypto_engine * ) ; int ( * prepare_cipher_request ) ( s#crypto_engine * , s#ablkcipher_request * ) ; int ( * unprepare_cipher_request ) ( s#crypto_engine * , s#ablkcipher_request * ) ; int ( * prepare_hash_request ) ( s#crypto_engine * , s#ahash_request * ) ; int ( * unprepare_hash_request ) ( s#crypto_engine * , s#ahash_request * ) ; int ( * cipher_one_request ) ( s#crypto_engine * , s#ablkcipher_request * ) ; int ( * hash_one_request ) ( s#crypto_engine * , s#ahash_request * ) ; s#kthread_worker * kworker ; s#kthread_work pump_requests ; void * priv_data ; s#crypto_async_request * cur_req ; }
+crypto_transfer_cipher_request int crypto_transfer_cipher_request ( s#crypto_engine * , s#ablkcipher_request * , t#bool )
+crypto_transfer_cipher_request_to_engine int crypto_transfer_cipher_request_to_engine ( s#crypto_engine * , s#ablkcipher_request * )
+crypto_transfer_hash_request int crypto_transfer_hash_request ( s#crypto_engine * , s#ahash_request * , t#bool )
+crypto_transfer_hash_request_to_engine int crypto_transfer_hash_request_to_engine ( s#crypto_engine * , s#ahash_request * )
+crypto_finalize_cipher_request void crypto_finalize_cipher_request ( s#crypto_engine * , s#ablkcipher_request * , int )
+crypto_finalize_hash_request void crypto_finalize_hash_request ( s#crypto_engine * , s#ahash_request * , int )
+crypto_engine_start int crypto_engine_start ( s#crypto_engine * )
+crypto_engine_stop int crypto_engine_stop ( s#crypto_engine * )
+crypto_engine_alloc_init s#crypto_engine * crypto_engine_alloc_init ( s#device * , t#bool )
+crypto_engine_exit int crypto_engine_exit ( s#crypto_engine * )
+
+/* crypto/crypto_null.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+crypto_get_default_null_skcipher s#crypto_skcipher * crypto_get_default_null_skcipher ( void )
+crypto_put_default_null_skcipher void crypto_put_default_null_skcipher ( void )
+
+/* crypto/crypto_wq.o */
+s##workqueue_struct
+kcrypto_wq extern s#workqueue_struct * kcrypto_wq
+
+/* crypto/des_generic.o */
+des_ekey extern unsigned long des_ekey ( t#u32 * , const t#u8 * )
+__des3_ede_setkey extern int __des3_ede_setkey ( t#u32 * , t#u32 * , const t#u8 * , unsigned int )
+
+/* crypto/dh_helper.o */
+s#dh struct dh { void * key ; void * p ; void * g ; unsigned int key_size ; unsigned int p_size ; unsigned int g_size ; }
+crypto_dh_key_len int crypto_dh_key_len ( const s#dh * )
+crypto_dh_encode_key int crypto_dh_encode_key ( char * , unsigned int , const s#dh * )
+crypto_dh_decode_key int crypto_dh_decode_key ( const char * , unsigned int , s#dh * )
+
+/* crypto/ecdh_helper.o */
+s#ecdh struct ecdh { unsigned short curve_id ; char * key ; unsigned short key_size ; }
+crypto_ecdh_key_len int crypto_ecdh_key_len ( const s#ecdh * )
+crypto_ecdh_encode_key int crypto_ecdh_encode_key ( char * , unsigned int , const s#ecdh * )
+crypto_ecdh_decode_key int crypto_ecdh_decode_key ( const char * , unsigned int , s#ecdh * )
+
+/* crypto/fips.o */
+fips_enabled extern int fips_enabled
+
+/* crypto/gf128mul.o */
+t#__le64 typedef t#__u64 __le64
+t#le128 typedef struct { t#__le64 b , a ; } le128
+gf128mul_x8_ble void gf128mul_x8_ble ( t#le128 * , const t#le128 * )
+t#be128 typedef struct { t#__be64 a , b ; } be128
+gf128mul_lle void gf128mul_lle ( t#be128 * , const t#be128 * )
+gf128mul_bbe void gf128mul_bbe ( t#be128 * , const t#be128 * )
+s#gf128mul_4k struct gf128mul_4k { t#be128 t [ 256 ] ; }
+s#gf128mul_64k struct gf128mul_64k { s#gf128mul_4k * t [ 16 ] ; }
+gf128mul_init_64k_bbe s#gf128mul_64k * gf128mul_init_64k_bbe ( const t#be128 * )
+gf128mul_free_64k void gf128mul_free_64k ( s#gf128mul_64k * )
+gf128mul_64k_bbe void gf128mul_64k_bbe ( t#be128 * , const s#gf128mul_64k * )
+gf128mul_init_4k_lle s#gf128mul_4k * gf128mul_init_4k_lle ( const t#be128 * )
+gf128mul_init_4k_bbe s#gf128mul_4k * gf128mul_init_4k_bbe ( const t#be128 * )
+gf128mul_4k_lle void gf128mul_4k_lle ( t#be128 * , const s#gf128mul_4k * )
+gf128mul_4k_bbe void gf128mul_4k_bbe ( t#be128 * , const s#gf128mul_4k * )
+
+/* crypto/hash_info.o */
+E#HASH_ALGO__LAST 18
+hash_algo_name extern const char * const hash_algo_name [ E#HASH_ALGO__LAST ]
+hash_digest_size extern const int hash_digest_size [ E#HASH_ALGO__LAST ]
+
+/* crypto/kpp.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##cgroup_namespace
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s##skcipher_givcrypt_request
+s#crypto_kpp struct crypto_kpp { s#crypto_tfm base ; }
+crypto_alloc_kpp s#crypto_kpp * crypto_alloc_kpp ( const char * , t#u32 , t#u32 )
+s#kpp_request struct kpp_request { s#crypto_async_request base ; s#scatterlist * src ; s#scatterlist * dst ; unsigned int src_len ; unsigned int dst_len ; void * __ctx [ ] __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; }
+s#kpp_alg struct kpp_alg { int ( * set_secret ) ( s#crypto_kpp * , const void * , unsigned int ) ; int ( * generate_public_key ) ( s#kpp_request * ) ; int ( * compute_shared_secret ) ( s#kpp_request * ) ; unsigned int ( * max_size ) ( s#crypto_kpp * ) ; int ( * init ) ( s#crypto_kpp * ) ; void ( * exit ) ( s#crypto_kpp * ) ; unsigned int reqsize ; s#crypto_alg base ; }
+crypto_register_kpp int crypto_register_kpp ( s#kpp_alg * )
+crypto_unregister_kpp void crypto_unregister_kpp ( s#kpp_alg * )
+
+/* crypto/lrw.o */
+s#lrw_table_ctx struct lrw_table_ctx { s#gf128mul_64k * table ; t#be128 mulinc [ 128 ] ; }
+lrw_init_table int lrw_init_table ( s#lrw_table_ctx * , const t#u8 * )
+lrw_free_table void lrw_free_table ( s#lrw_table_ctx * )
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s#lrw_crypt_req struct lrw_crypt_req { t#be128 * tbuf ; unsigned int tbuflen ; s#lrw_table_ctx * table_ctx ; void * crypt_ctx ; void ( * crypt_fn ) ( void * , t#u8 * , unsigned int ) ; }
+lrw_crypt int lrw_crypt ( s#blkcipher_desc * , s#scatterlist * , s#scatterlist * , unsigned int , s#lrw_crypt_req * )
+
+/* crypto/mcryptd.o */
+s##workqueue_struct
+s#mcryptd_alg_state struct mcryptd_alg_state { s#mcryptd_alg_cstate * alg_cstate ; unsigned long ( * flusher ) ( s#mcryptd_alg_cstate * ) ; }
+s#mcryptd_alg_cstate struct mcryptd_alg_cstate { unsigned long next_flush ; unsigned next_seq_num ; t#bool flusher_engaged ; s#delayed_work flush ; int cpu ; s#mcryptd_alg_state * alg_state ; void * mgr ; t#spinlock_t work_lock ; s#list_head work_list ; s#list_head flush_list ; }
+mcryptd_arm_flusher void mcryptd_arm_flusher ( s#mcryptd_alg_cstate * , unsigned long )
+mcryptd_flusher void mcryptd_flusher ( s#work_struct * )
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s##skcipher_givcrypt_request
+s#mcryptd_ahash struct mcryptd_ahash { s#crypto_ahash base ; }
+mcryptd_alloc_ahash s#mcryptd_ahash * mcryptd_alloc_ahash ( const char * , t#u32 , t#u32 )
+mcryptd_ahash_child s#crypto_ahash * mcryptd_ahash_child ( s#mcryptd_ahash * )
+mcryptd_ahash_desc s#ahash_request * mcryptd_ahash_desc ( s#ahash_request * )
+mcryptd_free_ahash void mcryptd_free_ahash ( s#mcryptd_ahash * )
+
+/* crypto/md5.o */
+md5_zero_message_hash extern const t#u8 md5_zero_message_hash [ 16 ]
+
+/* crypto/memneq.o */
+__crypto_memneq __attribute__ ( ( noinline ) ) unsigned long __crypto_memneq ( const void * , const void * , t#size_t )
+
+/* crypto/poly1305_generic.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s##skcipher_givcrypt_request
+crypto_poly1305_init int crypto_poly1305_init ( s#shash_desc * )
+s#poly1305_desc_ctx struct poly1305_desc_ctx { t#u32 r [ 5 ] ; t#u32 s [ 4 ] ; t#u32 h [ 5 ] ; t#u8 buf [ 16 ] ; unsigned int buflen ; t#bool rset ; t#bool sset ; }
+crypto_poly1305_setdesckey unsigned int crypto_poly1305_setdesckey ( s#poly1305_desc_ctx * , const t#u8 * , unsigned int )
+crypto_poly1305_update int crypto_poly1305_update ( s#shash_desc * , const t#u8 * , unsigned int )
+crypto_poly1305_final int crypto_poly1305_final ( s#shash_desc * , t#u8 * )
+
+/* crypto/rng.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##cgroup_namespace
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s##skcipher_givcrypt_request
+s#crypto_rng struct crypto_rng { s#crypto_tfm base ; }
+crypto_default_rng extern s#crypto_rng * crypto_default_rng
+crypto_rng_reset int crypto_rng_reset ( s#crypto_rng * , const t#u8 * , unsigned int )
+crypto_alloc_rng s#crypto_rng * crypto_alloc_rng ( const char * , t#u32 , t#u32 )
+crypto_get_default_rng int crypto_get_default_rng ( void )
+crypto_put_default_rng void crypto_put_default_rng ( void )
+crypto_del_default_rng int crypto_del_default_rng ( void )
+s#rng_alg struct rng_alg { int ( * generate ) ( s#crypto_rng * , const t#u8 * , unsigned int , t#u8 * , unsigned int ) ; int ( * seed ) ( s#crypto_rng * , const t#u8 * , unsigned int ) ; void ( * set_ent ) ( s#crypto_rng * , const t#u8 * , unsigned int ) ; unsigned int seedsize ; s#crypto_alg base ; }
+crypto_register_rng int crypto_register_rng ( s#rng_alg * )
+crypto_unregister_rng void crypto_unregister_rng ( s#rng_alg * )
+crypto_register_rngs int crypto_register_rngs ( s#rng_alg * , int )
+crypto_unregister_rngs void crypto_unregister_rngs ( s#rng_alg * , int )
+
+/* crypto/rsa_helper.o */
+s#rsa_key struct rsa_key { const t#u8 * n ; const t#u8 * e ; const t#u8 * d ; const t#u8 * p ; const t#u8 * q ; const t#u8 * dp ; const t#u8 * dq ; const t#u8 * qinv ; t#size_t n_sz ; t#size_t e_sz ; t#size_t d_sz ; t#size_t p_sz ; t#size_t q_sz ; t#size_t dp_sz ; t#size_t dq_sz ; t#size_t qinv_sz ; }
+rsa_parse_pub_key int rsa_parse_pub_key ( s#rsa_key * , const void * , unsigned int )
+rsa_parse_priv_key int rsa_parse_priv_key ( s#rsa_key * , const void * , unsigned int )
+
+/* crypto/scatterwalk.o */
+scatterwalk_copychunks void scatterwalk_copychunks ( void * , s#scatter_walk * , t#size_t , int )
+scatterwalk_map_and_copy void scatterwalk_map_and_copy ( void * , s#scatterlist * , unsigned int , unsigned int , int )
+scatterwalk_ffwd s#scatterlist * scatterwalk_ffwd ( s#scatterlist [ 2 ] , s#scatterlist * , unsigned int )
+
+/* crypto/scompress.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##cgroup_namespace
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s##skcipher_givcrypt_request
+s#crypto_scomp struct crypto_scomp { s#crypto_tfm base ; }
+s#scomp_alg struct scomp_alg { void * ( * alloc_ctx ) ( s#crypto_scomp * ) ; void ( * free_ctx ) ( s#crypto_scomp * , void * ) ; int ( * compress ) ( s#crypto_scomp * , const t#u8 * , unsigned int , t#u8 * , unsigned int * , void * ) ; int ( * decompress ) ( s#crypto_scomp * , const t#u8 * , unsigned int , t#u8 * , unsigned int * , void * ) ; s#crypto_alg base ; }
+crypto_register_scomp int crypto_register_scomp ( s#scomp_alg * )
+crypto_unregister_scomp int crypto_unregister_scomp ( s#scomp_alg * )
+crypto_register_scomps int crypto_register_scomps ( s#scomp_alg * , int )
+crypto_unregister_scomps void crypto_unregister_scomps ( s#scomp_alg * , int )
+
+/* crypto/serpent_generic.o */
+s#serpent_ctx struct serpent_ctx { t#u32 expkey [ 132 ] ; }
+__serpent_setkey int __serpent_setkey ( s#serpent_ctx * , const t#u8 * , unsigned int )
+s##scatterlist
+s##crypto_type
+s##skcipher_givcrypt_request
+s##sock
+s##seq_file
+s##file
+s##vm_operations_struct
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##cred
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##pipe_inode_info
+s##address_space
+s##dev_pagemap
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##tracepoint
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+serpent_setkey int serpent_setkey ( s#crypto_tfm * , const t#u8 * , unsigned int )
+__serpent_encrypt void __serpent_encrypt ( s#serpent_ctx * , t#u8 * , const t#u8 * )
+__serpent_decrypt void __serpent_decrypt ( s#serpent_ctx * , t#u8 * , const t#u8 * )
+
+/* crypto/sha1_generic.o */
+sha1_zero_message_hash extern const t#u8 sha1_zero_message_hash [ 20 ]
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s##skcipher_givcrypt_request
+crypto_sha1_update extern int crypto_sha1_update ( s#shash_desc * , const t#u8 * , unsigned int )
+crypto_sha1_finup extern int crypto_sha1_finup ( s#shash_desc * , const t#u8 * , unsigned int , t#u8 * )
+
+/* crypto/sha256_generic.o */
+sha224_zero_message_hash extern const t#u8 sha224_zero_message_hash [ 28 ]
+sha256_zero_message_hash extern const t#u8 sha256_zero_message_hash [ 32 ]
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s##skcipher_givcrypt_request
+crypto_sha256_update extern int crypto_sha256_update ( s#shash_desc * , const t#u8 * , unsigned int )
+crypto_sha256_finup extern int crypto_sha256_finup ( s#shash_desc * , const t#u8 * , unsigned int , t#u8 * )
+
+/* crypto/sha512_generic.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s##skcipher_givcrypt_request
+crypto_sha512_update extern int crypto_sha512_update ( s#shash_desc * , const t#u8 * , unsigned int )
+crypto_sha512_finup extern int crypto_sha512_finup ( s#shash_desc * , const t#u8 * , unsigned int , t#u8 * )
+
+/* crypto/shash.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##cgroup_namespace
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s##skcipher_givcrypt_request
+shash_no_setkey int shash_no_setkey ( s#crypto_shash * , const t#u8 * , unsigned int )
+crypto_shash_setkey int crypto_shash_setkey ( s#crypto_shash * , const t#u8 * , unsigned int )
+crypto_shash_update int crypto_shash_update ( s#shash_desc * , const t#u8 * , unsigned int )
+crypto_shash_final int crypto_shash_final ( s#shash_desc * , t#u8 * )
+crypto_shash_finup int crypto_shash_finup ( s#shash_desc * , const t#u8 * , unsigned int , t#u8 * )
+crypto_shash_digest int crypto_shash_digest ( s#shash_desc * , const t#u8 * , unsigned int , t#u8 * )
+shash_ahash_update int shash_ahash_update ( s#ahash_request * , s#shash_desc * )
+shash_ahash_finup int shash_ahash_finup ( s#ahash_request * , s#shash_desc * )
+shash_ahash_digest int shash_ahash_digest ( s#ahash_request * , s#shash_desc * )
+crypto_alloc_shash s#crypto_shash * crypto_alloc_shash ( const char * , t#u32 , t#u32 )
+s#shash_alg struct shash_alg { int ( * init ) ( s#shash_desc * ) ; int ( * update ) ( s#shash_desc * , const t#u8 * , unsigned int ) ; int ( * final ) ( s#shash_desc * , t#u8 * ) ; int ( * finup ) ( s#shash_desc * , const t#u8 * , unsigned int , t#u8 * ) ; int ( * digest ) ( s#shash_desc * , const t#u8 * , unsigned int , t#u8 * ) ; int ( * export ) ( s#shash_desc * , void * ) ; int ( * import ) ( s#shash_desc * , const void * ) ; int ( * setkey ) ( s#crypto_shash * , const t#u8 * , unsigned int ) ; unsigned int descsize ; unsigned int digestsize __attribute__ ( ( aligned ( __alignof__ ( struct hash_alg_common ) ) ) ) ; unsigned int statesize ; s#crypto_alg base ; }
+crypto_register_shash int crypto_register_shash ( s#shash_alg * )
+crypto_unregister_shash int crypto_unregister_shash ( s#shash_alg * )
+crypto_register_shashes int crypto_register_shashes ( s#shash_alg * , int )
+crypto_unregister_shashes int crypto_unregister_shashes ( s#shash_alg * , int )
+s#shash_instance struct shash_instance { s#shash_alg alg ; }
+shash_register_instance int shash_register_instance ( s#crypto_template * , s#shash_instance * )
+shash_free_instance void shash_free_instance ( s#crypto_instance * )
+s#crypto_shash_spawn struct crypto_shash_spawn { s#crypto_spawn base ; }
+crypto_init_shash_spawn int crypto_init_shash_spawn ( s#crypto_shash_spawn * , s#shash_alg * , s#crypto_instance * )
+shash_attr_alg s#shash_alg * shash_attr_alg ( s#rtattr * , t#u32 , t#u32 )
+
+/* crypto/simd.o */
+s##simd_skcipher_alg
+simd_skcipher_create_compat s#simd_skcipher_alg * simd_skcipher_create_compat ( const char * , const char * , const char * )
+simd_skcipher_create s#simd_skcipher_alg * simd_skcipher_create ( const char * , const char * )
+simd_skcipher_free void simd_skcipher_free ( s#simd_skcipher_alg * )
+
+/* crypto/skcipher.o */
+s##posix_acl
+s##vfsmount
+s##sock
+s##bio_vec
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##proc_dir_entry
+s##uevent_sock
+s##iw_handler_def
+s##iw_public_data
+s##sec_path
+s##neigh_parms
+s##netpoll_info
+s##neighbour
+s##udp_tunnel_info
+s##bpf_prog
+s##bpf_prog_offload_ops
+s##bpf_offloaded_map
+s##xdp_umem
+s##switchdev_ops
+s##l3mdev_ops
+s##ndisc_ops
+s##xfrm_state
+s##tls_crypto_info
+s##tls_context
+s##vlan_info
+s##dsa_switch_tree
+s##in_device
+s##dn_dev
+s##inet6_dev
+s##wireless_dev
+s##wpan_dev
+s##mpls_dev
+s##mini_Qdisc
+s##Qdisc
+s##nf_hook_entry
+s##cpu_rmap
+s##pcpu_dstats
+s##garp_port
+s##mrp_port
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##rtnl_link_ops
+s##phy_device
+s##sfp_bus
+s##prot_inuse
+s##ipv4_devconf
+s##fib_rules_ops
+s##fib_table
+s##inet_peer_base
+s##kmem_cache
+s##xt_table
+s##inet_hashinfo
+s##fib_notifier_ops
+s##ipv6_devconf
+s##rt6_info
+s##rt6_statistics
+s##fib6_table
+s##dst_entry
+s##seg6_pernet_data
+s##sctp_mib
+s##nf_queue_handler
+s##nf_logger
+s##ebt_table
+s##workqueue_struct
+s##nf_ct_event_notifier
+s##nf_exp_event_notifier
+s##nft_af_info
+s##net_generic
+s##netns_ipvs
+s##mpls_route
+s##dev_rcv_lists
+s##s_stats
+s##s_pstats
+s##bpf_prog_array
+s##cgroup_taskset
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##bdi_writeback
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##swap_info_struct
+s#skcipher_walk struct skcipher_walk { union { struct { s#page * page ; unsigned long offset ; } phys ; struct { t#u8 * page ; void * addr ; } virt ; } src , dst ; s#scatter_walk in ; unsigned int nbytes ; s#scatter_walk out ; unsigned int total ; s#list_head buffers ; t#u8 * page ; t#u8 * buffer ; t#u8 * oiv ; void * iv ; unsigned int ivsize ; int flags ; unsigned int blocksize ; unsigned int stride ; unsigned int alignmask ; }
+skcipher_walk_done int skcipher_walk_done ( s#skcipher_walk * , int )
+skcipher_walk_complete void skcipher_walk_complete ( s#skcipher_walk * , int )
+skcipher_walk_next int skcipher_walk_next ( s#skcipher_walk * )
+skcipher_walk_virt int skcipher_walk_virt ( s#skcipher_walk * , s#skcipher_request * , t#bool )
+skcipher_walk_atomise void skcipher_walk_atomise ( s#skcipher_walk * )
+skcipher_walk_async int skcipher_walk_async ( s#skcipher_walk * , s#skcipher_request * )
+skcipher_walk_aead int skcipher_walk_aead ( s#skcipher_walk * , s#aead_request * , t#bool )
+skcipher_walk_aead_encrypt int skcipher_walk_aead_encrypt ( s#skcipher_walk * , s#aead_request * , t#bool )
+skcipher_walk_aead_decrypt int skcipher_walk_aead_decrypt ( s#skcipher_walk * , s#aead_request * , t#bool )
+s#crypto_skcipher_spawn struct crypto_skcipher_spawn { s#crypto_spawn base ; }
+crypto_grab_skcipher int crypto_grab_skcipher ( s#crypto_skcipher_spawn * , const char * , t#u32 , t#u32 )
+crypto_alloc_skcipher s#crypto_skcipher * crypto_alloc_skcipher ( const char * , t#u32 , t#u32 )
+s#crypto_sync_skcipher struct crypto_sync_skcipher { s#crypto_skcipher base ; }
+crypto_alloc_sync_skcipher s#crypto_sync_skcipher * crypto_alloc_sync_skcipher ( const char * , t#u32 , t#u32 )
+crypto_has_skcipher2 int crypto_has_skcipher2 ( const char * , t#u32 , t#u32 )
+s#skcipher_alg struct skcipher_alg { int ( * setkey ) ( s#crypto_skcipher * , const t#u8 * , unsigned int ) ; int ( * encrypt ) ( s#skcipher_request * ) ; int ( * decrypt ) ( s#skcipher_request * ) ; int ( * init ) ( s#crypto_skcipher * ) ; void ( * exit ) ( s#crypto_skcipher * ) ; unsigned int min_keysize ; unsigned int max_keysize ; unsigned int ivsize ; unsigned int chunksize ; unsigned int walksize ; s#crypto_alg base ; }
+crypto_register_skcipher int crypto_register_skcipher ( s#skcipher_alg * )
+crypto_unregister_skcipher void crypto_unregister_skcipher ( s#skcipher_alg * )
+crypto_register_skciphers int crypto_register_skciphers ( s#skcipher_alg * , int )
+crypto_unregister_skciphers void crypto_unregister_skciphers ( s#skcipher_alg * , int )
+s#skcipher_instance struct skcipher_instance { void ( * free ) ( s#skcipher_instance * ) ; union { struct { char head [ __builtin_offsetof ( struct skcipher_alg , base ) ] ; s#crypto_instance base ; } s ; s#skcipher_alg alg ; } ; }
+skcipher_register_instance int skcipher_register_instance ( s#crypto_template * , s#skcipher_instance * )
+
+/* crypto/testmgr.o */
+alg_test int alg_test ( const char * , const char * , t#u32 , t#u32 )
+
+/* crypto/twofish_common.o */
+s#twofish_ctx struct twofish_ctx { t#u32 s [ 4 ] [ 256 ] , w [ 8 ] , k [ 32 ] ; }
+__twofish_setkey int __twofish_setkey ( s#twofish_ctx * , const t#u8 * , unsigned int , t#u32 * )
+s##scatterlist
+s##crypto_type
+s##skcipher_givcrypt_request
+s##sock
+s##seq_file
+s##file
+s##vm_operations_struct
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##cred
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##pipe_inode_info
+s##address_space
+s##dev_pagemap
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##tracepoint
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+twofish_setkey int twofish_setkey ( s#crypto_tfm * , const t#u8 * , unsigned int )
+
+/* crypto/xor.o */
+xor_blocks extern void xor_blocks ( unsigned int , unsigned int , void * , void * * )
+
+/* crypto/xts.o */
+s##vfsmount
+s##posix_acl
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##writeback_control
+s##bio_vec
+s##swap_info_struct
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##net_device
+s##sec_path
+s##rtattr
+s#xts_crypt_req struct xts_crypt_req { t#le128 * tbuf ; unsigned int tbuflen ; void * tweak_ctx ; void ( * tweak_fn ) ( void * , t#u8 * , const t#u8 * ) ; void * crypt_ctx ; void ( * crypt_fn ) ( void * , t#u8 * , unsigned int ) ; }
+xts_crypt int xts_crypt ( s#blkcipher_desc * , s#scatterlist * , s#scatterlist * , unsigned int , s#xts_crypt_req * )
+
+/* drivers/acpi/acpi_lpat.o */
+s#acpi_lpat struct acpi_lpat { int temp ; int raw ; }
+s#acpi_lpat_conversion_table struct acpi_lpat_conversion_table { s#acpi_lpat * lpat ; int lpat_count ; }
+acpi_lpat_raw_to_temp int acpi_lpat_raw_to_temp ( s#acpi_lpat_conversion_table * , int )
+acpi_lpat_temp_to_raw int acpi_lpat_temp_to_raw ( s#acpi_lpat_conversion_table * , int )
+t#acpi_handle typedef void * acpi_handle
+acpi_lpat_get_conversion_table s#acpi_lpat_conversion_table * acpi_lpat_get_conversion_table ( t#acpi_handle )
+acpi_lpat_free_conversion_table void acpi_lpat_free_conversion_table ( s#acpi_lpat_conversion_table * )
+
+/* drivers/acpi/acpi_platform.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s#platform_device_id struct platform_device_id { char name [ 20 ] ; t#kernel_ulong_t driver_data ; }
+s##mfd_cell
+s#pdev_archdata struct pdev_archdata { }
+s#platform_device struct platform_device { const char * name ; int id ; t#bool id_auto ; s#device dev ; t#u32 num_resources ; s#resource * resource ; const s#platform_device_id * id_entry ; char * driver_override ; s#mfd_cell * mfd_cell ; s#pdev_archdata archdata ; }
+s#acpi_device_status struct acpi_device_status { t#u32 present : 1 ; t#u32 enabled : 1 ; t#u32 show_in_ui : 1 ; t#u32 functional : 1 ; t#u32 battery_present : 1 ; t#u32 reserved : 27 ; }
+s#acpi_device_flags struct acpi_device_flags { t#u32 dynamic_status : 1 ; t#u32 removable : 1 ; t#u32 ejectable : 1 ; t#u32 power_manageable : 1 ; t#u32 match_driver : 1 ; t#u32 initialized : 1 ; t#u32 visited : 1 ; t#u32 hotplug_notify : 1 ; t#u32 is_dock_station : 1 ; t#u32 of_compatible_ok : 1 ; t#u32 coherent_dma : 1 ; t#u32 cca_seen : 1 ; t#u32 enumeration_by_parent : 1 ; t#u32 reserved : 19 ; }
+t#acpi_bus_id typedef char acpi_bus_id [ 8 ]
+s#acpi_pnp_type struct acpi_pnp_type { t#u32 hardware_id : 1 ; t#u32 bus_address : 1 ; t#u32 platform_id : 1 ; t#u32 reserved : 29 ; }
+t#acpi_bus_address typedef unsigned long acpi_bus_address
+t#acpi_device_name typedef char acpi_device_name [ 40 ]
+t#acpi_device_class typedef char acpi_device_class [ 20 ]
+t#acpi_object_type typedef t#u32 acpi_object_type
+t#acpi_io_address typedef t#u64 acpi_io_address
+u#acpi_object union acpi_object { t#acpi_object_type type ; struct { t#acpi_object_type type ; t#u64 value ; } integer ; struct { t#acpi_object_type type ; t#u32 length ; char * pointer ; } string ; struct { t#acpi_object_type type ; t#u32 length ; t#u8 * pointer ; } buffer ; struct { t#acpi_object_type type ; t#u32 count ; u#acpi_object * elements ; } package ; struct { t#acpi_object_type type ; t#acpi_object_type actual_type ; t#acpi_handle handle ; } reference ; struct { t#acpi_object_type type ; t#u32 proc_id ; t#acpi_io_address pblk_address ; t#u32 pblk_length ; } processor ; struct { t#acpi_object_type type ; t#u32 system_level ; t#u32 resource_order ; } power_resource ; }
+s#acpi_device_pnp struct acpi_device_pnp { t#acpi_bus_id bus_id ; s#acpi_pnp_type type ; t#acpi_bus_address bus_address ; char * unique_id ; s#list_head ids ; t#acpi_device_name device_name ; t#acpi_device_class device_class ; u#acpi_object * str_obj ; }
+s#acpi_device_power_flags struct acpi_device_power_flags { t#u32 explicit_get : 1 ; t#u32 power_resources : 1 ; t#u32 inrush_current : 1 ; t#u32 power_removed : 1 ; t#u32 ignore_parent : 1 ; t#u32 dsw_present : 1 ; t#u32 reserved : 26 ; }
+s#acpi_device_power_state struct acpi_device_power_state { struct { t#u8 valid : 1 ; t#u8 explicit_set : 1 ; t#u8 reserved : 6 ; } flags ; int power ; int latency ; s#list_head resources ; }
+s#acpi_device_power struct acpi_device_power { int state ; s#acpi_device_power_flags flags ; s#acpi_device_power_state states [ 5 ] ; }
+s#acpi_device_wakeup_flags struct acpi_device_wakeup_flags { t#u8 valid : 1 ; t#u8 notifier_present : 1 ; }
+s#acpi_device_wakeup_context struct acpi_device_wakeup_context { void ( * func ) ( s#acpi_device_wakeup_context * ) ; s#device * dev ; }
+s#acpi_device_wakeup struct acpi_device_wakeup { t#acpi_handle gpe_device ; t#u64 gpe_number ; t#u64 sleep_state ; s#list_head resources ; s#acpi_device_wakeup_flags flags ; s#acpi_device_wakeup_context context ; s#wakeup_source * ws ; int prepare_count ; int enable_count ; }
+s#acpi_device_perf_flags struct acpi_device_perf_flags { t#u8 reserved : 8 ; }
+s#acpi_device_perf_state struct acpi_device_perf_state { struct { t#u8 valid : 1 ; t#u8 reserved : 7 ; } flags ; t#u8 power ; t#u8 performance ; int latency ; }
+s#acpi_device_perf struct acpi_device_perf { int state ; s#acpi_device_perf_flags flags ; int state_count ; s#acpi_device_perf_state * states ; }
+s##proc_dir_entry
+s#acpi_device_dir struct acpi_device_dir { s#proc_dir_entry * entry ; }
+s#acpi_device_data struct acpi_device_data { const u#acpi_object * pointer ; s#list_head properties ; const u#acpi_object * of_compatible ; s#list_head subnodes ; }
+s#acpi_hotplug_profile struct acpi_hotplug_profile { s#kobject kobj ; int ( * scan_dependent ) ( s#acpi_device * ) ; void ( * notify_online ) ( s#acpi_device * ) ; t#bool enabled : 1 ; t#bool demand_offline : 1 ; }
+s#acpi_scan_handler struct acpi_scan_handler { const s#acpi_device_id * ids ; s#list_head list_node ; t#bool ( * match ) ( const char * , const s#acpi_device_id * * ) ; int ( * attach ) ( s#acpi_device * , const s#acpi_device_id * ) ; void ( * detach ) ( s#acpi_device * ) ; void ( * bind ) ( s#device * ) ; void ( * unbind ) ( s#device * ) ; s#acpi_hotplug_profile hotplug ; }
+s#acpi_hotplug_context struct acpi_hotplug_context { s#acpi_device * self ; int ( * notify ) ( s#acpi_device * , t#u32 ) ; void ( * uevent ) ( s#acpi_device * , t#u32 ) ; void ( * fixup ) ( s#acpi_device * ) ; }
+t#acpi_op_add typedef int ( * acpi_op_add ) ( s#acpi_device * )
+t#acpi_op_remove typedef int ( * acpi_op_remove ) ( s#acpi_device * )
+t#acpi_op_notify typedef void ( * acpi_op_notify ) ( s#acpi_device * , t#u32 )
+s#acpi_device_ops struct acpi_device_ops { t#acpi_op_add add ; t#acpi_op_remove remove ; t#acpi_op_notify notify ; }
+s#acpi_driver struct acpi_driver { char name [ 80 ] ; char class [ 80 ] ; const s#acpi_device_id * ids ; unsigned int flags ; s#acpi_device_ops ops ; s#device_driver drv ; s#module * owner ; }
+s#acpi_gpio_params struct acpi_gpio_params { unsigned int crs_entry_index ; unsigned int line_index ; t#bool active_low ; }
+s#acpi_gpio_mapping struct acpi_gpio_mapping { const char * name ; const s#acpi_gpio_params * data ; unsigned int size ; }
+s#acpi_device struct acpi_device { int device_type ; t#acpi_handle handle ; s#fwnode_handle fwnode ; s#acpi_device * parent ; s#list_head children ; s#list_head node ; s#list_head wakeup_list ; s#list_head del_list ; s#acpi_device_status status ; s#acpi_device_flags flags ; s#acpi_device_pnp pnp ; s#acpi_device_power power ; s#acpi_device_wakeup wakeup ; s#acpi_device_perf performance ; s#acpi_device_dir dir ; s#acpi_device_data data ; s#acpi_scan_handler * handler ; s#acpi_hotplug_context * hp ; s#acpi_driver * driver ; const s#acpi_gpio_mapping * driver_gpios ; void * driver_data ; s#device dev ; unsigned int physical_node_count ; unsigned int dep_unmet ; s#list_head physical_node_list ; s#mutex physical_node_lock ; void ( * remove ) ( s#acpi_device * ) ; }
+s#property_entry struct property_entry { const char * name ; t#size_t length ; t#bool is_array ; t#bool is_string ; union { union { const void * raw_data ; const t#u8 * u8_data ; const t#u16 * u16_data ; const t#u32 * u32_data ; const t#u64 * u64_data ; const char * const * str ; } pointer ; union { unsigned long long raw_data ; t#u8 u8_data ; t#u16 u16_data ; t#u32 u32_data ; t#u64 u64_data ; const char * str ; } value ; } ; }
+acpi_create_platform_device s#platform_device * acpi_create_platform_device ( s#acpi_device * , s#property_entry * )
+
+/* drivers/acpi/acpi_pnp.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##proc_dir_entry
+acpi_is_pnp_device extern t#bool acpi_is_pnp_device ( s#acpi_device * )
+
+/* drivers/acpi/acpi_processor.o */
+t#phys_cpuid_t typedef t#u64 phys_cpuid_t
+s#acpi_processor_flags struct acpi_processor_flags { t#u8 power : 1 ; t#u8 performance : 1 ; t#u8 throttling : 1 ; t#u8 limit : 1 ; t#u8 bm_control : 1 ; t#u8 bm_check : 1 ; t#u8 has_cst : 1 ; t#u8 has_lpi : 1 ; t#u8 power_setup_done : 1 ; t#u8 bm_rld_set : 1 ; t#u8 need_hotplug_init : 1 ; }
+s#acpi_processor_cx struct acpi_processor_cx { t#u8 valid ; t#u8 type ; t#u32 address ; t#u8 entry_method ; t#u8 index ; t#u32 latency ; t#u8 bm_sts_skip ; char desc [ 32 ] ; }
+s#acpi_lpi_state struct acpi_lpi_state { t#u32 min_residency ; t#u32 wake_latency ; t#u32 flags ; t#u32 arch_flags ; t#u32 res_cnt_freq ; t#u32 enable_parent_state ; t#u64 address ; t#u8 index ; t#u8 entry_method ; char desc [ 32 ] ; }
+s#acpi_processor_power struct acpi_processor_power { int count ; union { s#acpi_processor_cx states [ 8 ] ; s#acpi_lpi_state lpi_states [ 8 ] ; } ; int timer_broadcast_on_state ; }
+s#acpi_pct_register struct acpi_pct_register { t#u8 descriptor ; t#u16 length ; t#u8 space_id ; t#u8 bit_width ; t#u8 bit_offset ; t#u8 reserved ; t#u64 address ; }
+s#acpi_processor_px struct acpi_processor_px { t#u64 core_frequency ; t#u64 power ; t#u64 transition_latency ; t#u64 bus_master_latency ; t#u64 control ; t#u64 status ; }
+s#acpi_psd_package struct acpi_psd_package { t#u64 num_entries ; t#u64 revision ; t#u64 domain ; t#u64 coord_type ; t#u64 num_processors ; }
+s#acpi_processor_performance struct acpi_processor_performance { unsigned int state ; unsigned int platform_limit ; s#acpi_pct_register control_register ; s#acpi_pct_register status_register ; unsigned int state_count ; s#acpi_processor_px * states ; s#acpi_psd_package domain_info ; t#cpumask_var_t shared_cpu_map ; unsigned int shared_type ; }
+s#acpi_processor_tx_tss struct acpi_processor_tx_tss { t#u64 freqpercentage ; t#u64 power ; t#u64 transition_latency ; t#u64 control ; t#u64 status ; }
+s#acpi_tsd_package struct acpi_tsd_package { t#u64 num_entries ; t#u64 revision ; t#u64 domain ; t#u64 coord_type ; t#u64 num_processors ; }
+s#acpi_processor_tx struct acpi_processor_tx { t#u16 power ; t#u16 performance ; }
+s#acpi_processor_throttling struct acpi_processor_throttling { unsigned int state ; unsigned int platform_limit ; s#acpi_pct_register control_register ; s#acpi_pct_register status_register ; unsigned int state_count ; s#acpi_processor_tx_tss * states_tss ; s#acpi_tsd_package domain_info ; t#cpumask_var_t shared_cpu_map ; int ( * acpi_processor_get_throttling ) ( s#acpi_processor * ) ; int ( * acpi_processor_set_throttling ) ( s#acpi_processor * , int , t#bool ) ; t#u32 address ; t#u8 duty_offset ; t#u8 duty_width ; t#u8 tsd_valid_flag ; unsigned int shared_type ; s#acpi_processor_tx states [ 16 ] ; }
+s#acpi_processor_lx struct acpi_processor_lx { int px ; int tx ; }
+s#acpi_processor_limit struct acpi_processor_limit { s#acpi_processor_lx state ; s#acpi_processor_lx thermal ; s#acpi_processor_lx user ; }
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##iommu_group
+s##iommu_fwspec
+s#thermal_attr struct thermal_attr { s#device_attribute attr ; char name [ 20 ] ; }
+e#thermal_device_mode enum thermal_device_mode { THERMAL_DEVICE_DISABLED = 0 , THERMAL_DEVICE_ENABLED , }
+e#thermal_trip_type enum thermal_trip_type { THERMAL_TRIP_ACTIVE = 0 , THERMAL_TRIP_PASSIVE , THERMAL_TRIP_HOT , THERMAL_TRIP_CRITICAL , }
+e#thermal_trend enum thermal_trend { THERMAL_TREND_STABLE , THERMAL_TREND_RAISING , THERMAL_TREND_DROPPING , THERMAL_TREND_RAISE_FULL , THERMAL_TREND_DROP_FULL , }
+s#thermal_zone_device_ops struct thermal_zone_device_ops { int ( * bind ) ( s#thermal_zone_device * , s#thermal_cooling_device * ) ; int ( * unbind ) ( s#thermal_zone_device * , s#thermal_cooling_device * ) ; int ( * get_temp ) ( s#thermal_zone_device * , int * ) ; int ( * set_trips ) ( s#thermal_zone_device * , int , int ) ; int ( * get_mode ) ( s#thermal_zone_device * , e#thermal_device_mode * ) ; int ( * set_mode ) ( s#thermal_zone_device * , e#thermal_device_mode ) ; int ( * get_trip_type ) ( s#thermal_zone_device * , int , e#thermal_trip_type * ) ; int ( * get_trip_temp ) ( s#thermal_zone_device * , int , int * ) ; int ( * set_trip_temp ) ( s#thermal_zone_device * , int , int ) ; int ( * get_trip_hyst ) ( s#thermal_zone_device * , int , int * ) ; int ( * set_trip_hyst ) ( s#thermal_zone_device * , int , int ) ; int ( * get_crit_temp ) ( s#thermal_zone_device * , int * ) ; int ( * set_emul_temp ) ( s#thermal_zone_device * , int ) ; int ( * get_trend ) ( s#thermal_zone_device * , int , e#thermal_trend * ) ; int ( * notify ) ( s#thermal_zone_device * , int , e#thermal_trip_type ) ; }
+s#thermal_bind_params struct thermal_bind_params { s#thermal_cooling_device * cdev ; int weight ; int trip_mask ; unsigned long * binding_limits ; int ( * match ) ( s#thermal_zone_device * , s#thermal_cooling_device * ) ; }
+s#thermal_zone_params struct thermal_zone_params { char governor_name [ 20 ] ; t#bool no_hwmon ; int num_tbps ; s#thermal_bind_params * tbp ; t#u32 sustainable_power ; t#s32 k_po ; t#s32 k_pu ; t#s32 k_i ; t#s32 k_d ; t#s32 integral_cutoff ; int slope ; int offset ; }
+s#thermal_governor struct thermal_governor { char name [ 20 ] ; int ( * bind_to_tz ) ( s#thermal_zone_device * ) ; void ( * unbind_from_tz ) ( s#thermal_zone_device * ) ; int ( * throttle ) ( s#thermal_zone_device * , int ) ; s#list_head governor_list ; }
+s#ida struct ida { s#radix_tree_root ida_rt ; }
+e#thermal_notify_event enum thermal_notify_event { THERMAL_EVENT_UNSPECIFIED , THERMAL_EVENT_TEMP_SAMPLE , THERMAL_TRIP_VIOLATED , THERMAL_TRIP_CHANGED , THERMAL_DEVICE_DOWN , THERMAL_DEVICE_UP , THERMAL_DEVICE_POWER_CAPABILITY_CHANGED , }
+s#thermal_zone_device struct thermal_zone_device { int id ; char type [ 20 ] ; s#device device ; s#attribute_group trips_attribute_group ; s#thermal_attr * trip_temp_attrs ; s#thermal_attr * trip_type_attrs ; s#thermal_attr * trip_hyst_attrs ; void * devdata ; int trips ; unsigned long trips_disabled ; int passive_delay ; int polling_delay ; int temperature ; int last_temperature ; int emul_temperature ; int passive ; int prev_low_trip ; int prev_high_trip ; unsigned int forced_passive ; t#atomic_t need_update ; s#thermal_zone_device_ops * ops ; s#thermal_zone_params * tzp ; s#thermal_governor * governor ; void * governor_data ; s#list_head thermal_instances ; s#ida ida ; s#mutex lock ; s#list_head node ; s#delayed_work poll_queue ; e#thermal_notify_event notify_event ; }
+s#thermal_cooling_device_ops struct thermal_cooling_device_ops { int ( * get_max_state ) ( s#thermal_cooling_device * , unsigned long * ) ; int ( * get_cur_state ) ( s#thermal_cooling_device * , unsigned long * ) ; int ( * set_cur_state ) ( s#thermal_cooling_device * , unsigned long ) ; int ( * get_requested_power ) ( s#thermal_cooling_device * , s#thermal_zone_device * , t#u32 * ) ; int ( * state2power ) ( s#thermal_cooling_device * , s#thermal_zone_device * , unsigned long , t#u32 * ) ; int ( * power2state ) ( s#thermal_cooling_device * , s#thermal_zone_device * , t#u32 , unsigned long * ) ; }
+s#thermal_cooling_device struct thermal_cooling_device { int id ; char type [ 20 ] ; s#device device ; s#device_node * np ; void * devdata ; const s#thermal_cooling_device_ops * ops ; t#bool updated ; s#mutex lock ; s#list_head thermal_instances ; s#list_head node ; }
+s#acpi_processor struct acpi_processor { t#acpi_handle handle ; t#u32 acpi_id ; t#phys_cpuid_t phys_id ; t#u32 id ; t#u32 pblk ; int performance_platform_limit ; int throttling_platform_limit ; s#acpi_processor_flags flags ; s#acpi_processor_power power ; s#acpi_processor_performance * performance ; s#acpi_processor_throttling throttling ; s#acpi_processor_limit limit ; s#thermal_cooling_device * cdev ; s#device * dev ; }
+processors extern __attribute__ ( ( section ( ".data..percpu" "" ) ) ) __typeof__ ( s#acpi_processor * ) processors
+s#acpi_processor_errata struct acpi_processor_errata { t#u8 smp ; struct { t#u8 throttle : 1 ; t#u8 fdma : 1 ; t#u8 reserved : 6 ; t#u32 bmisx ; } piix4 ; }
+errata extern s#acpi_processor_errata errata
+
+/* drivers/acpi/acpi_watchdog.o */
+acpi_has_watchdog extern t#bool acpi_has_watchdog ( void )
+
+/* drivers/acpi/acpica/evxface.o */
+t#acpi_status typedef t#u32 acpi_status
+t#acpi_notify_handler typedef void ( * acpi_notify_handler ) ( t#acpi_handle , t#u32 , void * )
+acpi_install_notify_handler t#acpi_status acpi_install_notify_handler ( t#acpi_handle , t#u32 , t#acpi_notify_handler , void * )
+acpi_remove_notify_handler t#acpi_status acpi_remove_notify_handler ( t#acpi_handle , t#u32 , t#acpi_notify_handler )
+
+/* drivers/acpi/acpica/evxfregn.o */
+t#acpi_adr_space_type typedef t#u8 acpi_adr_space_type
+t#acpi_physical_address typedef t#u64 acpi_physical_address
+t#acpi_adr_space_handler typedef t#acpi_status ( * acpi_adr_space_handler ) ( t#u32 , t#acpi_physical_address , t#u32 , t#u64 * , void * , void * )
+t#acpi_adr_space_setup typedef t#acpi_status ( * acpi_adr_space_setup ) ( t#acpi_handle , t#u32 , void * , void * * )
+acpi_install_address_space_handler t#acpi_status acpi_install_address_space_handler ( t#acpi_handle , t#acpi_adr_space_type , t#acpi_adr_space_handler , t#acpi_adr_space_setup , void * )
+acpi_remove_address_space_handler t#acpi_status acpi_remove_address_space_handler ( t#acpi_handle , t#acpi_adr_space_type , t#acpi_adr_space_handler )
+
+/* drivers/acpi/acpica/hwxface.o */
+acpi_reset t#acpi_status acpi_reset ( void )
+s#acpi_generic_address struct acpi_generic_address { t#u8 space_id ; t#u8 bit_width ; t#u8 bit_offset ; t#u8 access_width ; t#u64 address ; }
+acpi_read t#acpi_status acpi_read ( t#u64 * , s#acpi_generic_address * )
+acpi_write t#acpi_status acpi_write ( t#u64 , s#acpi_generic_address * )
+acpi_get_sleep_type_data t#acpi_status acpi_get_sleep_type_data ( t#u8 , t#u8 * , t#u8 * )
+
+/* drivers/acpi/acpica/hwxfsleep.o */
+acpi_enter_sleep_state_prep t#acpi_status acpi_enter_sleep_state_prep ( t#u8 )
+acpi_enter_sleep_state t#acpi_status acpi_enter_sleep_state ( t#u8 )
+acpi_leave_sleep_state_prep t#acpi_status acpi_leave_sleep_state_prep ( t#u8 )
+acpi_leave_sleep_state t#acpi_status acpi_leave_sleep_state ( t#u8 )
+
+/* drivers/acpi/acpica/nsxfeval.o */
+t#acpi_string typedef char * acpi_string
+s#acpi_object_list struct acpi_object_list { t#u32 count ; u#acpi_object * pointer ; }
+t#acpi_size typedef t#u64 acpi_size
+s#acpi_buffer struct acpi_buffer { t#acpi_size length ; void * pointer ; }
+acpi_evaluate_object_typed t#acpi_status acpi_evaluate_object_typed ( t#acpi_handle , t#acpi_string , s#acpi_object_list * , s#acpi_buffer * , t#acpi_object_type )
+acpi_evaluate_object t#acpi_status acpi_evaluate_object ( t#acpi_handle , t#acpi_string , s#acpi_object_list * , s#acpi_buffer * )
+t#acpi_walk_callback typedef t#acpi_status ( * acpi_walk_callback ) ( t#acpi_handle , t#u32 , void * , void * * )
+acpi_walk_namespace t#acpi_status acpi_walk_namespace ( t#acpi_object_type , t#acpi_handle , t#u32 , t#acpi_walk_callback , t#acpi_walk_callback , void * , void * * )
+acpi_get_devices t#acpi_status acpi_get_devices ( const char * , t#acpi_walk_callback , void * , void * * )
+t#acpi_object_handler typedef void ( * acpi_object_handler ) ( t#acpi_handle , void * )
+acpi_attach_data t#acpi_status acpi_attach_data ( t#acpi_handle , t#acpi_object_handler , void * )
+acpi_detach_data t#acpi_status acpi_detach_data ( t#acpi_handle , t#acpi_object_handler )
+acpi_get_data_full t#acpi_status acpi_get_data_full ( t#acpi_handle , t#acpi_object_handler , void * * , void ( * ) ( void * ) )
+acpi_get_data t#acpi_status acpi_get_data ( t#acpi_handle , t#acpi_object_handler , void * * )
+
+/* drivers/acpi/acpica/nsxfname.o */
+acpi_get_handle t#acpi_status acpi_get_handle ( t#acpi_handle , t#acpi_string , t#acpi_handle * )
+acpi_get_name t#acpi_status acpi_get_name ( t#acpi_handle , t#u32 , s#acpi_buffer * )
+s#acpi_pnp_device_id struct acpi_pnp_device_id { t#u32 length ; char * string ; }
+s#acpi_pnp_device_id_list struct acpi_pnp_device_id_list { t#u32 count ; t#u32 list_size ; s#acpi_pnp_device_id ids [ 1 ] ; }
+s#acpi_device_info struct acpi_device_info { t#u32 info_size ; t#u32 name ; t#acpi_object_type type ; t#u8 param_count ; t#u16 valid ; t#u8 flags ; t#u8 highest_dstates [ 4 ] ; t#u8 lowest_dstates [ 5 ] ; t#u64 address ; s#acpi_pnp_device_id hardware_id ; s#acpi_pnp_device_id unique_id ; s#acpi_pnp_device_id class_code ; s#acpi_pnp_device_id_list compatible_id_list ; }
+acpi_get_object_info t#acpi_status acpi_get_object_info ( t#acpi_handle , s#acpi_device_info * * )
+acpi_install_method t#acpi_status acpi_install_method ( t#u8 * )
+
+/* drivers/acpi/acpica/nsxfobj.o */
+acpi_get_type t#acpi_status acpi_get_type ( t#acpi_handle , t#acpi_object_type * )
+acpi_get_parent t#acpi_status acpi_get_parent ( t#acpi_handle , t#acpi_handle * )
+acpi_get_next_object t#acpi_status acpi_get_next_object ( t#acpi_object_type , t#acpi_handle , t#acpi_handle , t#acpi_handle * )
+
+/* drivers/acpi/acpica/rscreate.o */
+s#acpi_resource_irq struct acpi_resource_irq { t#u8 descriptor_length ; t#u8 triggering ; t#u8 polarity ; t#u8 sharable ; t#u8 wake_capable ; t#u8 interrupt_count ; t#u8 interrupts [ 1 ] ; }
+s#acpi_resource_dma struct acpi_resource_dma { t#u8 type ; t#u8 bus_master ; t#u8 transfer ; t#u8 channel_count ; t#u8 channels [ 1 ] ; }
+s#acpi_resource_start_dependent struct acpi_resource_start_dependent { t#u8 descriptor_length ; t#u8 compatibility_priority ; t#u8 performance_robustness ; }
+s#acpi_resource_io struct acpi_resource_io { t#u8 io_decode ; t#u8 alignment ; t#u8 address_length ; t#u16 minimum ; t#u16 maximum ; }
+s#acpi_resource_fixed_io struct acpi_resource_fixed_io { t#u16 address ; t#u8 address_length ; }
+s#acpi_resource_fixed_dma struct acpi_resource_fixed_dma { t#u16 request_lines ; t#u16 channels ; t#u8 width ; }
+s#acpi_resource_vendor struct acpi_resource_vendor { t#u16 byte_length ; t#u8 byte_data [ 1 ] ; }
+s#acpi_resource_vendor_typed struct acpi_resource_vendor_typed { t#u16 byte_length ; t#u8 uuid_subtype ; t#u8 uuid [ 16 ] ; t#u8 byte_data [ 1 ] ; }
+s#acpi_resource_end_tag struct acpi_resource_end_tag { t#u8 checksum ; }
+s#acpi_resource_memory24 struct acpi_resource_memory24 { t#u8 write_protect ; t#u16 minimum ; t#u16 maximum ; t#u16 alignment ; t#u16 address_length ; }
+s#acpi_resource_memory32 struct acpi_resource_memory32 { t#u8 write_protect ; t#u32 minimum ; t#u32 maximum ; t#u32 alignment ; t#u32 address_length ; }
+s#acpi_resource_fixed_memory32 struct acpi_resource_fixed_memory32 { t#u8 write_protect ; t#u32 address ; t#u32 address_length ; }
+s#acpi_memory_attribute struct acpi_memory_attribute { t#u8 write_protect ; t#u8 caching ; t#u8 range_type ; t#u8 translation ; }
+s#acpi_io_attribute struct acpi_io_attribute { t#u8 range_type ; t#u8 translation ; t#u8 translation_type ; t#u8 reserved1 ; }
+u#acpi_resource_attribute union acpi_resource_attribute { s#acpi_memory_attribute mem ; s#acpi_io_attribute io ; t#u8 type_specific ; }
+s#acpi_address16_attribute struct acpi_address16_attribute { t#u16 granularity ; t#u16 minimum ; t#u16 maximum ; t#u16 translation_offset ; t#u16 address_length ; }
+s#acpi_resource_source struct acpi_resource_source { t#u8 index ; t#u16 string_length ; char * string_ptr ; }
+s#acpi_resource_address16 struct acpi_resource_address16 { t#u8 resource_type ; t#u8 producer_consumer ; t#u8 decode ; t#u8 min_address_fixed ; t#u8 max_address_fixed ; u#acpi_resource_attribute info ; s#acpi_address16_attribute address ; s#acpi_resource_source resource_source ; }
+s#acpi_address32_attribute struct acpi_address32_attribute { t#u32 granularity ; t#u32 minimum ; t#u32 maximum ; t#u32 translation_offset ; t#u32 address_length ; }
+s#acpi_resource_address32 struct acpi_resource_address32 { t#u8 resource_type ; t#u8 producer_consumer ; t#u8 decode ; t#u8 min_address_fixed ; t#u8 max_address_fixed ; u#acpi_resource_attribute info ; s#acpi_address32_attribute address ; s#acpi_resource_source resource_source ; }
+s#acpi_address64_attribute struct acpi_address64_attribute { t#u64 granularity ; t#u64 minimum ; t#u64 maximum ; t#u64 translation_offset ; t#u64 address_length ; }
+s#acpi_resource_address64 struct acpi_resource_address64 { t#u8 resource_type ; t#u8 producer_consumer ; t#u8 decode ; t#u8 min_address_fixed ; t#u8 max_address_fixed ; u#acpi_resource_attribute info ; s#acpi_address64_attribute address ; s#acpi_resource_source resource_source ; }
+s#acpi_resource_extended_address64 struct acpi_resource_extended_address64 { t#u8 resource_type ; t#u8 producer_consumer ; t#u8 decode ; t#u8 min_address_fixed ; t#u8 max_address_fixed ; u#acpi_resource_attribute info ; t#u8 revision_ID ; s#acpi_address64_attribute address ; t#u64 type_specific ; }
+s#acpi_resource_extended_irq struct acpi_resource_extended_irq { t#u8 producer_consumer ; t#u8 triggering ; t#u8 polarity ; t#u8 sharable ; t#u8 wake_capable ; t#u8 interrupt_count ; s#acpi_resource_source resource_source ; t#u32 interrupts [ 1 ] ; }
+s#acpi_resource_generic_register struct acpi_resource_generic_register { t#u8 space_id ; t#u8 bit_width ; t#u8 bit_offset ; t#u8 access_size ; t#u64 address ; }
+s#acpi_resource_gpio struct acpi_resource_gpio { t#u8 revision_id ; t#u8 connection_type ; t#u8 producer_consumer ; t#u8 pin_config ; t#u8 sharable ; t#u8 wake_capable ; t#u8 io_restriction ; t#u8 triggering ; t#u8 polarity ; t#u16 drive_strength ; t#u16 debounce_timeout ; t#u16 pin_table_length ; t#u16 vendor_length ; s#acpi_resource_source resource_source ; t#u16 * pin_table ; t#u8 * vendor_data ; }
+s#acpi_resource_i2c_serialbus struct acpi_resource_i2c_serialbus { t#u8 revision_id ; t#u8 type ; t#u8 producer_consumer ; t#u8 slave_mode ; t#u8 connection_sharing ; t#u8 type_revision_id ; t#u16 type_data_length ; t#u16 vendor_length ; s#acpi_resource_source resource_source ; t#u8 * vendor_data ; t#u8 access_mode ; t#u16 slave_address ; t#u32 connection_speed ; }
+s#acpi_resource_spi_serialbus struct acpi_resource_spi_serialbus { t#u8 revision_id ; t#u8 type ; t#u8 producer_consumer ; t#u8 slave_mode ; t#u8 connection_sharing ; t#u8 type_revision_id ; t#u16 type_data_length ; t#u16 vendor_length ; s#acpi_resource_source resource_source ; t#u8 * vendor_data ; t#u8 wire_mode ; t#u8 device_polarity ; t#u8 data_bit_length ; t#u8 clock_phase ; t#u8 clock_polarity ; t#u16 device_selection ; t#u32 connection_speed ; }
+s#acpi_resource_uart_serialbus struct acpi_resource_uart_serialbus { t#u8 revision_id ; t#u8 type ; t#u8 producer_consumer ; t#u8 slave_mode ; t#u8 connection_sharing ; t#u8 type_revision_id ; t#u16 type_data_length ; t#u16 vendor_length ; s#acpi_resource_source resource_source ; t#u8 * vendor_data ; t#u8 endian ; t#u8 data_bits ; t#u8 stop_bits ; t#u8 flow_control ; t#u8 parity ; t#u8 lines_enabled ; t#u16 rx_fifo_size ; t#u16 tx_fifo_size ; t#u32 default_baud_rate ; }
+s#acpi_resource_common_serialbus struct acpi_resource_common_serialbus { t#u8 revision_id ; t#u8 type ; t#u8 producer_consumer ; t#u8 slave_mode ; t#u8 connection_sharing ; t#u8 type_revision_id ; t#u16 type_data_length ; t#u16 vendor_length ; s#acpi_resource_source resource_source ; t#u8 * vendor_data ; }
+s#acpi_resource_pin_function struct acpi_resource_pin_function { t#u8 revision_id ; t#u8 pin_config ; t#u8 sharable ; t#u16 function_number ; t#u16 pin_table_length ; t#u16 vendor_length ; s#acpi_resource_source resource_source ; t#u16 * pin_table ; t#u8 * vendor_data ; }
+s#acpi_resource_pin_config struct acpi_resource_pin_config { t#u8 revision_id ; t#u8 producer_consumer ; t#u8 sharable ; t#u8 pin_config_type ; t#u32 pin_config_value ; t#u16 pin_table_length ; t#u16 vendor_length ; s#acpi_resource_source resource_source ; t#u16 * pin_table ; t#u8 * vendor_data ; }
+s#acpi_resource_label struct acpi_resource_label { t#u16 string_length ; char * string_ptr ; }
+s#acpi_resource_pin_group struct acpi_resource_pin_group { t#u8 revision_id ; t#u8 producer_consumer ; t#u16 pin_table_length ; t#u16 vendor_length ; t#u16 * pin_table ; s#acpi_resource_label resource_label ; t#u8 * vendor_data ; }
+s#acpi_resource_pin_group_function struct acpi_resource_pin_group_function { t#u8 revision_id ; t#u8 producer_consumer ; t#u8 sharable ; t#u16 function_number ; t#u16 vendor_length ; s#acpi_resource_source resource_source ; s#acpi_resource_label resource_source_label ; t#u8 * vendor_data ; }
+s#acpi_resource_pin_group_config struct acpi_resource_pin_group_config { t#u8 revision_id ; t#u8 producer_consumer ; t#u8 sharable ; t#u8 pin_config_type ; t#u32 pin_config_value ; t#u16 vendor_length ; s#acpi_resource_source resource_source ; s#acpi_resource_label resource_source_label ; t#u8 * vendor_data ; }
+s#acpi_resource_address struct acpi_resource_address { t#u8 resource_type ; t#u8 producer_consumer ; t#u8 decode ; t#u8 min_address_fixed ; t#u8 max_address_fixed ; u#acpi_resource_attribute info ; }
+u#acpi_resource_data union acpi_resource_data { s#acpi_resource_irq irq ; s#acpi_resource_dma dma ; s#acpi_resource_start_dependent start_dpf ; s#acpi_resource_io io ; s#acpi_resource_fixed_io fixed_io ; s#acpi_resource_fixed_dma fixed_dma ; s#acpi_resource_vendor vendor ; s#acpi_resource_vendor_typed vendor_typed ; s#acpi_resource_end_tag end_tag ; s#acpi_resource_memory24 memory24 ; s#acpi_resource_memory32 memory32 ; s#acpi_resource_fixed_memory32 fixed_memory32 ; s#acpi_resource_address16 address16 ; s#acpi_resource_address32 address32 ; s#acpi_resource_address64 address64 ; s#acpi_resource_extended_address64 ext_address64 ; s#acpi_resource_extended_irq extended_irq ; s#acpi_resource_generic_register generic_reg ; s#acpi_resource_gpio gpio ; s#acpi_resource_i2c_serialbus i2c_serial_bus ; s#acpi_resource_spi_serialbus spi_serial_bus ; s#acpi_resource_uart_serialbus uart_serial_bus ; s#acpi_resource_common_serialbus common_serial_bus ; s#acpi_resource_pin_function pin_function ; s#acpi_resource_pin_config pin_config ; s#acpi_resource_pin_group pin_group ; s#acpi_resource_pin_group_function pin_group_function ; s#acpi_resource_pin_group_config pin_group_config ; s#acpi_resource_address address ; }
+s#acpi_resource struct acpi_resource { t#u32 type ; t#u32 length ; u#acpi_resource_data data ; }
+acpi_buffer_to_resource t#acpi_status acpi_buffer_to_resource ( t#u8 * , t#u16 , s#acpi_resource * * )
+
+/* drivers/acpi/acpica/rsxface.o */
+acpi_get_irq_routing_table t#acpi_status acpi_get_irq_routing_table ( t#acpi_handle , s#acpi_buffer * )
+acpi_get_current_resources t#acpi_status acpi_get_current_resources ( t#acpi_handle , s#acpi_buffer * )
+acpi_get_possible_resources t#acpi_status acpi_get_possible_resources ( t#acpi_handle , s#acpi_buffer * )
+acpi_set_current_resources t#acpi_status acpi_set_current_resources ( t#acpi_handle , s#acpi_buffer * )
+acpi_get_event_resources t#acpi_status acpi_get_event_resources ( t#acpi_handle , s#acpi_buffer * )
+acpi_resource_to_address64 t#acpi_status acpi_resource_to_address64 ( s#acpi_resource * , s#acpi_resource_address64 * )
+s#acpi_vendor_uuid struct acpi_vendor_uuid { t#u8 subtype ; t#u8 data [ 16 ] ; }
+acpi_get_vendor_resource t#acpi_status acpi_get_vendor_resource ( t#acpi_handle , char * , s#acpi_vendor_uuid * , s#acpi_buffer * )
+t#acpi_walk_resource_callback typedef t#acpi_status ( * acpi_walk_resource_callback ) ( s#acpi_resource * , void * )
+acpi_walk_resource_buffer t#acpi_status acpi_walk_resource_buffer ( s#acpi_buffer * , t#acpi_walk_resource_callback , void * )
+acpi_walk_resources t#acpi_status acpi_walk_resources ( t#acpi_handle , char * , t#acpi_walk_resource_callback , void * )
+
+/* drivers/acpi/acpica/tbdata.o */
+acpi_tb_install_and_load_table t#acpi_status acpi_tb_install_and_load_table ( t#acpi_physical_address , t#u8 , t#u8 , t#u32 * )
+acpi_tb_unload_table t#acpi_status acpi_tb_unload_table ( t#u32 )
+
+/* drivers/acpi/acpica/tbxface.o */
+s#acpi_table_header struct acpi_table_header { char signature [ 4 ] ; t#u32 length ; t#u8 revision ; t#u8 checksum ; char oem_id [ 6 ] ; char oem_table_id [ 8 ] ; t#u32 oem_revision ; char asl_compiler_id [ 4 ] ; t#u32 asl_compiler_revision ; }
+acpi_get_table_header t#acpi_status acpi_get_table_header ( t#acpi_string , t#u32 , s#acpi_table_header * )
+acpi_get_table t#acpi_status acpi_get_table ( t#acpi_string , t#u32 , s#acpi_table_header * * )
+acpi_put_table void acpi_put_table ( s#acpi_table_header * )
+acpi_get_table_by_index t#acpi_status acpi_get_table_by_index ( t#u32 , s#acpi_table_header * * )
+t#acpi_table_handler typedef t#acpi_status ( * acpi_table_handler ) ( t#u32 , void * , void * )
+acpi_install_table_handler t#acpi_status acpi_install_table_handler ( t#acpi_table_handler , void * )
+acpi_remove_table_handler t#acpi_status acpi_remove_table_handler ( t#acpi_table_handler )
+
+/* drivers/acpi/acpica/tbxfload.o */
+acpi_load_table t#acpi_status acpi_load_table ( s#acpi_table_header * )
+acpi_unload_parent_table t#acpi_status acpi_unload_parent_table ( t#acpi_handle )
+
+/* drivers/acpi/acpica/utexcep.o */
+acpi_format_exception const char * acpi_format_exception ( t#acpi_status )
+
+/* drivers/acpi/acpica/utglobal.o */
+s#acpi_table_fadt struct acpi_table_fadt { s#acpi_table_header header ; t#u32 facs ; t#u32 dsdt ; t#u8 model ; t#u8 preferred_profile ; t#u16 sci_interrupt ; t#u32 smi_command ; t#u8 acpi_enable ; t#u8 acpi_disable ; t#u8 s4_bios_request ; t#u8 pstate_control ; t#u32 pm1a_event_block ; t#u32 pm1b_event_block ; t#u32 pm1a_control_block ; t#u32 pm1b_control_block ; t#u32 pm2_control_block ; t#u32 pm_timer_block ; t#u32 gpe0_block ; t#u32 gpe1_block ; t#u8 pm1_event_length ; t#u8 pm1_control_length ; t#u8 pm2_control_length ; t#u8 pm_timer_length ; t#u8 gpe0_block_length ; t#u8 gpe1_block_length ; t#u8 gpe1_base ; t#u8 cst_control ; t#u16 c2_latency ; t#u16 c3_latency ; t#u16 flush_size ; t#u16 flush_stride ; t#u8 duty_offset ; t#u8 duty_width ; t#u8 day_alarm ; t#u8 month_alarm ; t#u8 century ; t#u16 boot_flags ; t#u8 reserved ; t#u32 flags ; s#acpi_generic_address reset_register ; t#u8 reset_value ; t#u16 arm_boot_flags ; t#u8 minor_revision ; t#u64 Xfacs ; t#u64 Xdsdt ; s#acpi_generic_address xpm1a_event_block ; s#acpi_generic_address xpm1b_event_block ; s#acpi_generic_address xpm1a_control_block ; s#acpi_generic_address xpm1b_control_block ; s#acpi_generic_address xpm2_control_block ; s#acpi_generic_address xpm_timer_block ; s#acpi_generic_address xgpe0_block ; s#acpi_generic_address xgpe1_block ; s#acpi_generic_address sleep_control ; s#acpi_generic_address sleep_status ; t#u64 hypervisor_id ; }
+acpi_gbl_FADT extern s#acpi_table_fadt acpi_gbl_FADT
+acpi_dbg_level t#u32 acpi_dbg_level
+acpi_dbg_layer t#u32 acpi_dbg_layer
+acpi_gpe_count extern t#u32 acpi_gpe_count
+acpi_current_gpe_count extern t#u32 acpi_current_gpe_count
+
+/* drivers/acpi/acpica/utxface.o */
+acpi_purge_cached_objects t#acpi_status acpi_purge_cached_objects ( void )
+acpi_install_interface t#acpi_status acpi_install_interface ( t#acpi_string )
+acpi_remove_interface t#acpi_status acpi_remove_interface ( t#acpi_string )
+t#acpi_interface_handler typedef t#u32 ( * acpi_interface_handler ) ( t#acpi_string , t#u32 )
+acpi_install_interface_handler t#acpi_status acpi_install_interface_handler ( t#acpi_interface_handler )
+acpi_check_address_range t#u32 acpi_check_address_range ( t#acpi_adr_space_type , t#acpi_physical_address , t#acpi_size , t#u8 )
+s#acpi_pld_info struct acpi_pld_info { t#u8 revision ; t#u8 ignore_color ; t#u8 red ; t#u8 green ; t#u8 blue ; t#u16 width ; t#u16 height ; t#u8 user_visible ; t#u8 dock ; t#u8 lid ; t#u8 panel ; t#u8 vertical_position ; t#u8 horizontal_position ; t#u8 shape ; t#u8 group_orientation ; t#u8 group_token ; t#u8 group_position ; t#u8 bay ; t#u8 ejectable ; t#u8 ospm_eject_required ; t#u8 cabinet_number ; t#u8 card_cage_number ; t#u8 reference ; t#u8 rotation ; t#u8 order ; t#u8 reserved ; t#u16 vertical_offset ; t#u16 horizontal_offset ; }
+acpi_decode_pld_buffer t#acpi_status acpi_decode_pld_buffer ( t#u8 * , t#acpi_size , s#acpi_pld_info * * )
+
+/* drivers/acpi/acpica/utxferror.o */
+acpi_error __attribute__ ( ( __format__ ( __printf__ , 3 , 3 + 1 ) ) ) void acpi_error ( const char * , t#u32 , const char * , ... )
+acpi_exception __attribute__ ( ( __format__ ( __printf__ , 4 , 4 + 1 ) ) ) void acpi_exception ( const char * , t#u32 , t#acpi_status , const char * , ... )
+acpi_warning __attribute__ ( ( __format__ ( __printf__ , 3 , 3 + 1 ) ) ) void acpi_warning ( const char * , t#u32 , const char * , ... )
+acpi_info __attribute__ ( ( __format__ ( __printf__ , 1 , 1 + 1 ) ) ) void acpi_info ( const char * , ... )
+acpi_bios_error __attribute__ ( ( __format__ ( __printf__ , 3 , 3 + 1 ) ) ) void acpi_bios_error ( const char * , t#u32 , const char * , ... )
+acpi_bios_warning __attribute__ ( ( __format__ ( __printf__ , 3 , 3 + 1 ) ) ) void acpi_bios_warning ( const char * , t#u32 , const char * , ... )
+
+/* drivers/acpi/acpica/utxfmutex.o */
+acpi_acquire_mutex t#acpi_status acpi_acquire_mutex ( t#acpi_handle , t#acpi_string , t#u16 )
+acpi_release_mutex t#acpi_status acpi_release_mutex ( t#acpi_handle , t#acpi_string )
+
+/* drivers/acpi/apei/apei-base.o */
+s#acpi_whea_header struct acpi_whea_header { t#u8 action ; t#u8 instruction ; t#u8 flags ; t#u8 reserved ; s#acpi_generic_address register_region ; t#u64 value ; t#u64 mask ; }
+t#apei_exec_ins_func_t typedef int ( * apei_exec_ins_func_t ) ( s#apei_exec_context * , s#acpi_whea_header * )
+s#apei_exec_ins_type struct apei_exec_ins_type { t#u32 flags ; t#apei_exec_ins_func_t run ; }
+s#apei_exec_context struct apei_exec_context { t#u32 ip ; t#u64 value ; t#u64 var1 ; t#u64 var2 ; t#u64 src_base ; t#u64 dst_base ; s#apei_exec_ins_type * ins_table ; t#u32 instructions ; s#acpi_whea_header * action_table ; t#u32 entries ; }
+apei_exec_ctx_init void apei_exec_ctx_init ( s#apei_exec_context * , s#apei_exec_ins_type * , t#u32 , s#acpi_whea_header * , t#u32 )
+apei_exec_read_register int apei_exec_read_register ( s#apei_exec_context * , s#acpi_whea_header * )
+apei_exec_read_register_value int apei_exec_read_register_value ( s#apei_exec_context * , s#acpi_whea_header * )
+apei_exec_write_register int apei_exec_write_register ( s#apei_exec_context * , s#acpi_whea_header * )
+apei_exec_write_register_value int apei_exec_write_register_value ( s#apei_exec_context * , s#acpi_whea_header * )
+apei_exec_noop int apei_exec_noop ( s#apei_exec_context * , s#acpi_whea_header * )
+__apei_exec_run int __apei_exec_run ( s#apei_exec_context * , t#u8 , t#bool )
+apei_exec_pre_map_gars int apei_exec_pre_map_gars ( s#apei_exec_context * )
+apei_exec_post_unmap_gars int apei_exec_post_unmap_gars ( s#apei_exec_context * )
+s#apei_resources struct apei_resources { s#list_head iomem ; s#list_head ioport ; }
+apei_resources_fini void apei_resources_fini ( s#apei_resources * )
+apei_resources_add int apei_resources_add ( s#apei_resources * , unsigned long , unsigned long , t#bool )
+apei_resources_sub int apei_resources_sub ( s#apei_resources * , s#apei_resources * )
+apei_resources_request int apei_resources_request ( s#apei_resources * , const char * )
+apei_resources_release void apei_resources_release ( s#apei_resources * )
+apei_map_generic_address int apei_map_generic_address ( s#acpi_generic_address * )
+apei_read int apei_read ( t#u64 * , s#acpi_generic_address * )
+apei_write int apei_write ( t#u64 , s#acpi_generic_address * )
+apei_exec_collect_resources int apei_exec_collect_resources ( s#apei_exec_context * , s#apei_resources * )
+s##posix_acl
+s##vfsmount
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##pipe_inode_info
+s##writeback_control
+s##iov_iter
+s##swap_info_struct
+s##device_private
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##nlm_lockowner
+s##nfs4_lock_state
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##fsnotify_mark_connector
+s##fscrypt_info
+apei_get_debugfs_dir s#dentry * apei_get_debugfs_dir ( void )
+s#acpi_hest_header struct acpi_hest_header { t#u16 type ; t#u16 source_id ; }
+arch_apei_enable_cmcff int __attribute__ ( ( weak ) ) arch_apei_enable_cmcff ( s#acpi_hest_header * , void * )
+s#cper_sec_mem_err struct cper_sec_mem_err { t#__u64 validation_bits ; t#__u64 error_status ; t#__u64 physical_addr ; t#__u64 physical_addr_mask ; t#__u16 node ; t#__u16 card ; t#__u16 module ; t#__u16 bank ; t#__u16 device ; t#__u16 row ; t#__u16 column ; t#__u16 bit_pos ; t#__u64 requestor_id ; t#__u64 responder_id ; t#__u64 target_id ; t#__u8 error_type ; t#__u8 reserved ; t#__u16 rank ; t#__u16 mem_array_handle ; t#__u16 mem_dev_handle ; }
+arch_apei_report_mem_error void __attribute__ ( ( weak ) ) arch_apei_report_mem_error ( int , s#cper_sec_mem_err * )
+apei_osc_setup int apei_osc_setup ( void )
+
+/* drivers/acpi/apei/erst.o */
+erst_disable extern int erst_disable
+erst_get_record_count t#ssize_t erst_get_record_count ( void )
+erst_get_record_id_begin int erst_get_record_id_begin ( int * )
+erst_get_record_id_next int erst_get_record_id_next ( int * , t#u64 * )
+erst_get_record_id_end void erst_get_record_id_end ( void )
+t#guid_t typedef struct { t#__u8 b [ 16 ] ; } guid_t
+t#uuid_le typedef t#guid_t uuid_le
+s#cper_record_header struct cper_record_header { char signature [ 4 ] ; t#__u16 revision ; t#__u32 signature_end ; t#__u16 section_count ; t#__u32 error_severity ; t#__u32 validation_bits ; t#__u32 record_length ; t#__u64 timestamp ; t#uuid_le platform_id ; t#uuid_le partition_id ; t#uuid_le creator_id ; t#uuid_le notification_type ; t#__u64 record_id ; t#__u32 flags ; t#__u64 persistence_information ; t#__u8 reserved [ 12 ] ; }
+erst_write int erst_write ( const s#cper_record_header * )
+erst_read t#ssize_t erst_read ( t#u64 , s#cper_record_header * , t#size_t )
+erst_clear int erst_clear ( t#u64 )
+
+/* drivers/acpi/apei/hest.o */
+hest_disable extern t#bool hest_disable
+t#apei_hest_func_t typedef int ( * apei_hest_func_t ) ( s#acpi_hest_header * , void * )
+apei_hest_parse int apei_hest_parse ( t#apei_hest_func_t , void * )
+
+/* drivers/acpi/bus.o */
+s##proc_dir_entry
+acpi_root_dir extern s#proc_dir_entry * acpi_root_dir
+acpi_bus_get_status_handle t#acpi_status acpi_bus_get_status_handle ( t#acpi_handle , unsigned long long * )
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##iov_iter
+s##hd_struct
+s##gendisk
+s##request_queue
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##blk_plug
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+e#irq_domain_bus_token enum irq_domain_bus_token { DOMAIN_BUS_ANY = 0 , DOMAIN_BUS_WIRED , DOMAIN_BUS_PCI_MSI , DOMAIN_BUS_PLATFORM_MSI , DOMAIN_BUS_NEXUS , DOMAIN_BUS_IPI , DOMAIN_BUS_FSL_MC_MSI , }
+s#irq_fwspec struct irq_fwspec { s#fwnode_handle * fwnode ; int param_count ; t#u32 param [ 16 ] ; }
+t#irq_hw_number_t typedef unsigned long irq_hw_number_t
+s##irq_data
+s#irq_domain_ops struct irq_domain_ops { int ( * match ) ( s#irq_domain * , s#device_node * , e#irq_domain_bus_token ) ; int ( * select ) ( s#irq_domain * , s#irq_fwspec * , e#irq_domain_bus_token ) ; int ( * map ) ( s#irq_domain * , unsigned int , t#irq_hw_number_t ) ; void ( * unmap ) ( s#irq_domain * , unsigned int ) ; int ( * xlate ) ( s#irq_domain * , s#device_node * , const t#u32 * , unsigned int , unsigned long * , unsigned int * ) ; int ( * alloc ) ( s#irq_domain * , unsigned int , unsigned int , void * ) ; void ( * free ) ( s#irq_domain * , unsigned int , unsigned int ) ; void ( * activate ) ( s#irq_domain * , s#irq_data * ) ; void ( * deactivate ) ( s#irq_domain * , s#irq_data * ) ; int ( * translate ) ( s#irq_domain * , s#irq_fwspec * , unsigned long * , unsigned int * ) ; }
+s##irq_domain_chip_generic
+s#irq_domain struct irq_domain { s#list_head link ; const char * name ; const s#irq_domain_ops * ops ; void * host_data ; unsigned int flags ; s#fwnode_handle * fwnode ; e#irq_domain_bus_token bus_token ; s#irq_domain_chip_generic * gc ; s#irq_domain * parent ; t#irq_hw_number_t hwirq_max ; unsigned int revmap_direct_max_irq ; unsigned int revmap_size ; s#radix_tree_root revmap_tree ; unsigned int linear_revmap [ ] ; }
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##iommu_group
+s##iommu_fwspec
+acpi_bus_get_status int acpi_bus_get_status ( s#acpi_device * )
+acpi_bus_private_data_handler void acpi_bus_private_data_handler ( t#acpi_handle , void * )
+acpi_bus_attach_private_data int acpi_bus_attach_private_data ( t#acpi_handle , void * )
+acpi_bus_get_private_data int acpi_bus_get_private_data ( t#acpi_handle , void * * )
+acpi_bus_detach_private_data void acpi_bus_detach_private_data ( t#acpi_handle )
+s#acpi_osc_context struct acpi_osc_context { char * uuid_str ; int rev ; s#acpi_buffer cap ; s#acpi_buffer ret ; }
+acpi_run_osc t#acpi_status acpi_run_osc ( t#acpi_handle , s#acpi_osc_context * )
+osc_pc_lpi_support_confirmed extern t#bool osc_pc_lpi_support_confirmed
+acpi_set_modalias void acpi_set_modalias ( s#acpi_device * , const char * , char * , t#size_t )
+acpi_match_device const s#acpi_device_id * acpi_match_device ( const s#acpi_device_id * , const s#device * )
+acpi_match_device_ids int acpi_match_device_ids ( s#acpi_device * , const s#acpi_device_id * )
+acpi_driver_match_device extern t#bool acpi_driver_match_device ( s#device * , const s#device_driver * )
+acpi_bus_register_driver int acpi_bus_register_driver ( s#acpi_driver * )
+acpi_bus_unregister_driver void acpi_bus_unregister_driver ( s#acpi_driver * )
+acpi_kobj extern s#kobject * acpi_kobj
+
+/* drivers/acpi/button.o */
+acpi_lid_notifier_register extern int acpi_lid_notifier_register ( s#notifier_block * )
+acpi_lid_notifier_unregister extern int acpi_lid_notifier_unregister ( s#notifier_block * )
+acpi_lid_open extern int acpi_lid_open ( void )
+
+/* drivers/acpi/cppc_acpi.o */
+s#cppc_perf_caps struct cppc_perf_caps { t#u32 guaranteed_perf ; t#u32 highest_perf ; t#u32 nominal_perf ; t#u32 lowest_perf ; t#u32 lowest_nonlinear_perf ; t#u32 lowest_freq ; t#u32 nominal_freq ; }
+s#cppc_perf_ctrls struct cppc_perf_ctrls { t#u32 max_perf ; t#u32 min_perf ; t#u32 desired_perf ; }
+s#cppc_perf_fb_ctrs struct cppc_perf_fb_ctrs { t#u64 reference ; t#u64 delivered ; t#u64 reference_perf ; t#u64 wraparound_time ; }
+s##clk
+s#cpufreq_cpuinfo struct cpufreq_cpuinfo { unsigned int max_freq ; unsigned int min_freq ; unsigned int transition_latency ; }
+s##sock
+s##vfsmount
+s##posix_acl
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##device_private
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s#cpufreq_governor struct cpufreq_governor { char name [ 16 ] ; int ( * init ) ( s#cpufreq_policy * ) ; void ( * exit ) ( s#cpufreq_policy * ) ; int ( * start ) ( s#cpufreq_policy * ) ; void ( * stop ) ( s#cpufreq_policy * ) ; void ( * limits ) ( s#cpufreq_policy * ) ; t#ssize_t ( * show_setspeed ) ( s#cpufreq_policy * , char * ) ; int ( * store_setspeed ) ( s#cpufreq_policy * , unsigned int ) ; unsigned int max_transition_latency ; s#list_head governor_list ; s#module * owner ; }
+s#cpufreq_user_policy struct cpufreq_user_policy { unsigned int min ; unsigned int max ; }
+s#cpufreq_frequency_table struct cpufreq_frequency_table { unsigned int flags ; unsigned int driver_data ; unsigned int frequency ; }
+e#cpufreq_table_sorting enum cpufreq_table_sorting { CPUFREQ_TABLE_UNSORTED , CPUFREQ_TABLE_SORTED_ASCENDING , CPUFREQ_TABLE_SORTED_DESCENDING }
+s##cpufreq_stats
+s#cpufreq_policy struct cpufreq_policy { t#cpumask_var_t cpus ; t#cpumask_var_t related_cpus ; t#cpumask_var_t real_cpus ; unsigned int shared_type ; unsigned int cpu ; s#clk * clk ; s#cpufreq_cpuinfo cpuinfo ; unsigned int min ; unsigned int max ; unsigned int cur ; unsigned int restore_freq ; unsigned int suspend_freq ; unsigned int policy ; unsigned int last_policy ; s#cpufreq_governor * governor ; void * governor_data ; char last_governor [ 16 ] ; s#work_struct update ; s#cpufreq_user_policy user_policy ; s#cpufreq_frequency_table * freq_table ; e#cpufreq_table_sorting freq_table_sorted ; s#list_head policy_list ; s#kobject kobj ; s#completion kobj_unregister ; s#rw_semaphore rwsem ; t#bool fast_switch_possible ; t#bool fast_switch_enabled ; unsigned int transition_delay_us ; unsigned int cached_target_freq ; int cached_resolved_idx ; t#bool transition_ongoing ; t#spinlock_t transition_lock ; t#wait_queue_head_t transition_wait ; s#task_struct * transition_task ; s#cpufreq_stats * stats ; void * driver_data ; }
+s#cppc_cpudata struct cppc_cpudata { int cpu ; s#cppc_perf_caps perf_caps ; s#cppc_perf_ctrls perf_ctrls ; s#cppc_perf_fb_ctrs perf_fb_ctrs ; s#cpufreq_policy * cur_policy ; unsigned int shared_type ; t#cpumask_var_t shared_cpu_map ; }
+acpi_get_psd_map extern int acpi_get_psd_map ( s#cppc_cpudata * * )
+acpi_cppc_processor_probe extern int acpi_cppc_processor_probe ( s#acpi_processor * )
+acpi_cppc_processor_exit extern void acpi_cppc_processor_exit ( s#acpi_processor * )
+cppc_get_desired_perf extern int cppc_get_desired_perf ( int , t#u64 * )
+cppc_get_perf_caps extern int cppc_get_perf_caps ( int , s#cppc_perf_caps * )
+cppc_get_perf_ctrs extern int cppc_get_perf_ctrs ( int , s#cppc_perf_fb_ctrs * )
+cppc_set_perf extern int cppc_set_perf ( int , s#cppc_perf_ctrls * )
+cppc_get_transition_latency extern unsigned int cppc_get_transition_latency ( int )
+
+/* drivers/acpi/debugfs.o */
+s##posix_acl
+s##vfsmount
+s##sock
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##user_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##mempolicy
+s##numa_group
+s##pipe_inode_info
+s##writeback_control
+s##iov_iter
+s##swap_info_struct
+s##device_private
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##kmem_cache
+s##mem_cgroup
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##nlm_lockowner
+s##nfs4_lock_state
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##fsnotify_mark_connector
+s##fscrypt_info
+acpi_debugfs_dir extern s#dentry * acpi_debugfs_dir
+
+/* drivers/acpi/device_pm.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##iov_iter
+s##hd_struct
+s##gendisk
+s##request_queue
+s##kmem_cache
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##blk_plug
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s#pm_domain_data struct pm_domain_data { s#list_head list_node ; s#device * dev ; }
+s#plist_head struct plist_head { s#list_head node_list ; }
+e#pm_qos_type enum pm_qos_type { PM_QOS_UNITIALIZED , PM_QOS_MAX , PM_QOS_MIN , PM_QOS_SUM }
+s#pm_qos_constraints struct pm_qos_constraints { s#plist_head list ; t#s32 target_value ; t#s32 default_value ; t#s32 no_constraint_value ; e#pm_qos_type type ; s#blocking_notifier_head * notifiers ; }
+s#pm_qos_flags struct pm_qos_flags { s#list_head list ; t#s32 effective_flags ; }
+e#dev_pm_qos_req_type enum dev_pm_qos_req_type { DEV_PM_QOS_RESUME_LATENCY = 1 , DEV_PM_QOS_LATENCY_TOLERANCE , DEV_PM_QOS_FLAGS , }
+s#pm_qos_flags_request struct pm_qos_flags_request { s#list_head node ; t#s32 flags ; }
+s#dev_pm_qos_request struct dev_pm_qos_request { e#dev_pm_qos_req_type type ; union { s#plist_node pnode ; s#pm_qos_flags_request flr ; } data ; s#device * dev ; }
+s#dev_pm_qos struct dev_pm_qos { s#pm_qos_constraints resume_latency ; s#pm_qos_constraints latency_tolerance ; s#pm_qos_flags flags ; s#dev_pm_qos_request * resume_latency_req ; s#dev_pm_qos_request * latency_tolerance_req ; s#dev_pm_qos_request * flags_req ; }
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##iommu_group
+s##iommu_fwspec
+s##proc_dir_entry
+acpi_device_set_power int acpi_device_set_power ( s#acpi_device * , int )
+acpi_bus_set_power int acpi_bus_set_power ( t#acpi_handle , int )
+acpi_device_fix_up_power int acpi_device_fix_up_power ( s#acpi_device * )
+acpi_device_update_power int acpi_device_update_power ( s#acpi_device * , int * )
+acpi_bus_update_power int acpi_bus_update_power ( t#acpi_handle , int * )
+acpi_bus_power_manageable t#bool acpi_bus_power_manageable ( t#acpi_handle )
+acpi_pm_wakeup_event void acpi_pm_wakeup_event ( s#device * )
+acpi_bus_can_wakeup t#bool acpi_bus_can_wakeup ( t#acpi_handle )
+acpi_pm_device_sleep_state int acpi_pm_device_sleep_state ( s#device * , int * , int )
+acpi_pm_set_device_wakeup int acpi_pm_set_device_wakeup ( s#device * , t#bool )
+acpi_pm_set_bridge_wakeup int acpi_pm_set_bridge_wakeup ( s#device * , t#bool )
+acpi_dev_suspend int acpi_dev_suspend ( s#device * , t#bool )
+acpi_dev_resume int acpi_dev_resume ( s#device * )
+acpi_subsys_runtime_suspend int acpi_subsys_runtime_suspend ( s#device * )
+acpi_subsys_runtime_resume int acpi_subsys_runtime_resume ( s#device * )
+acpi_subsys_prepare int acpi_subsys_prepare ( s#device * )
+acpi_subsys_complete void acpi_subsys_complete ( s#device * )
+acpi_subsys_suspend int acpi_subsys_suspend ( s#device * )
+acpi_subsys_suspend_late int acpi_subsys_suspend_late ( s#device * )
+acpi_subsys_suspend_noirq int acpi_subsys_suspend_noirq ( s#device * )
+acpi_subsys_resume_noirq int acpi_subsys_resume_noirq ( s#device * )
+acpi_subsys_resume_early int acpi_subsys_resume_early ( s#device * )
+acpi_subsys_freeze int acpi_subsys_freeze ( s#device * )
+acpi_subsys_freeze_late int acpi_subsys_freeze_late ( s#device * )
+acpi_subsys_freeze_noirq int acpi_subsys_freeze_noirq ( s#device * )
+acpi_subsys_thaw_noirq int acpi_subsys_thaw_noirq ( s#device * )
+acpi_dev_pm_attach int acpi_dev_pm_attach ( s#device * , t#bool )
+
+/* drivers/acpi/device_sysfs.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+acpi_device_uevent_modalias int acpi_device_uevent_modalias ( s#device * , s#kobj_uevent_env * )
+acpi_device_modalias int acpi_device_modalias ( s#device * , char * , int )
+
+/* drivers/acpi/ec.o */
+s##transaction
+s#acpi_ec struct acpi_ec { t#acpi_handle handle ; t#u32 gpe ; unsigned long command_addr ; unsigned long data_addr ; t#bool global_lock ; unsigned long flags ; unsigned long reference_count ; s#mutex mutex ; t#wait_queue_head_t wait ; s#list_head list ; s#transaction * curr ; t#spinlock_t lock ; s#work_struct work ; unsigned long timestamp ; unsigned long nr_pending_queries ; t#bool busy_polling ; unsigned int polling_guard ; }
+first_ec extern s#acpi_ec * first_ec
+ec_read extern int ec_read ( t#u8 , t#u8 * )
+ec_write extern int ec_write ( t#u8 , t#u8 )
+ec_transaction extern int ec_transaction ( t#u8 , const t#u8 * , unsigned , t#u8 * , unsigned )
+ec_get_handle extern t#acpi_handle ec_get_handle ( void )
+t#acpi_ec_query_func typedef int ( * acpi_ec_query_func ) ( void * )
+acpi_ec_add_query_handler int acpi_ec_add_query_handler ( s#acpi_ec * , t#u8 , t#acpi_handle , t#acpi_ec_query_func , void * )
+acpi_ec_remove_query_handler void acpi_ec_remove_query_handler ( s#acpi_ec * , t#u8 )
+
+/* drivers/acpi/event.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##ucounts
+s##proc_ns_operations
+s##proc_dir_entry
+s##uevent_sock
+s##net_device
+s##prot_inuse
+s##ipv4_devconf
+s##fib_rules_ops
+s##fib_table
+s##inet_peer_base
+s##sec_path
+s##kmem_cache
+s##xt_table
+s##inet_hashinfo
+s##fib_notifier_ops
+s##ipv6_devconf
+s##rt6_info
+s##rt6_statistics
+s##fib6_table
+s##dst_entry
+s##neighbour
+s##seg6_pernet_data
+s##sctp_mib
+s##nf_queue_handler
+s##nf_logger
+s##nf_hook_entry
+s##ebt_table
+s##workqueue_struct
+s##nf_ct_event_notifier
+s##nf_exp_event_notifier
+s##nft_af_info
+s##net_generic
+s##netns_ipvs
+s##mpls_route
+s##dev_rcv_lists
+s##s_stats
+s##s_pstats
+s##cgroup_namespace
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##bio_vec
+s##swap_info_struct
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##bdi_writeback
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+acpi_notifier_call_chain extern int acpi_notifier_call_chain ( s#acpi_device * , t#u32 , t#u32 )
+register_acpi_notifier extern int register_acpi_notifier ( s#notifier_block * )
+unregister_acpi_notifier extern int unregister_acpi_notifier ( s#notifier_block * )
+acpi_bus_generate_netlink_event extern int acpi_bus_generate_netlink_event ( const char * , const char * , t#u8 , int )
+
+/* drivers/acpi/glue.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_data
+s##irq_domain_chip_generic
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##iommu_group
+s##iommu_fwspec
+s##proc_dir_entry
+s#acpi_bus_type struct acpi_bus_type { s#list_head list ; const char * name ; t#bool ( * match ) ( s#device * ) ; s#acpi_device * ( * find_companion ) ( s#device * ) ; void ( * setup ) ( s#device * ) ; void ( * cleanup ) ( s#device * ) ; }
+register_acpi_bus_type int register_acpi_bus_type ( s#acpi_bus_type * )
+unregister_acpi_bus_type int unregister_acpi_bus_type ( s#acpi_bus_type * )
+acpi_find_child_device s#acpi_device * acpi_find_child_device ( s#acpi_device * , t#u64 , t#bool )
+acpi_bind_one int acpi_bind_one ( s#device * , s#acpi_device * )
+acpi_unbind_one int acpi_unbind_one ( s#device * )
+
+/* drivers/acpi/hed.o */
+register_acpi_hed_notifier int register_acpi_hed_notifier ( s#notifier_block * )
+unregister_acpi_hed_notifier void unregister_acpi_hed_notifier ( s#notifier_block * )
+
+/* drivers/acpi/irq.o */
+acpi_gsi_to_irq int acpi_gsi_to_irq ( t#u32 , unsigned int * )
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##msi_desc
+s#irq_common_data struct irq_common_data { unsigned int state_use_accessors ; unsigned int node ; void * handler_data ; s#msi_desc * msi_desc ; t#cpumask_var_t affinity ; }
+s##msi_msg
+e#irqchip_irq_state enum irqchip_irq_state { IRQCHIP_STATE_PENDING , IRQCHIP_STATE_ACTIVE , IRQCHIP_STATE_MASKED , IRQCHIP_STATE_LINE_LEVEL , }
+s#irq_chip struct irq_chip { s#device * parent_device ; const char * name ; unsigned int ( * irq_startup ) ( s#irq_data * ) ; void ( * irq_shutdown ) ( s#irq_data * ) ; void ( * irq_enable ) ( s#irq_data * ) ; void ( * irq_disable ) ( s#irq_data * ) ; void ( * irq_ack ) ( s#irq_data * ) ; void ( * irq_mask ) ( s#irq_data * ) ; void ( * irq_mask_ack ) ( s#irq_data * ) ; void ( * irq_unmask ) ( s#irq_data * ) ; void ( * irq_eoi ) ( s#irq_data * ) ; int ( * irq_set_affinity ) ( s#irq_data * , const s#cpumask * , t#bool ) ; int ( * irq_retrigger ) ( s#irq_data * ) ; int ( * irq_set_type ) ( s#irq_data * , unsigned int ) ; int ( * irq_set_wake ) ( s#irq_data * , unsigned int ) ; void ( * irq_bus_lock ) ( s#irq_data * ) ; void ( * irq_bus_sync_unlock ) ( s#irq_data * ) ; void ( * irq_cpu_online ) ( s#irq_data * ) ; void ( * irq_cpu_offline ) ( s#irq_data * ) ; void ( * irq_suspend ) ( s#irq_data * ) ; void ( * irq_resume ) ( s#irq_data * ) ; void ( * irq_pm_shutdown ) ( s#irq_data * ) ; void ( * irq_calc_mask ) ( s#irq_data * ) ; void ( * irq_print_chip ) ( s#irq_data * , s#seq_file * ) ; int ( * irq_request_resources ) ( s#irq_data * ) ; void ( * irq_release_resources ) ( s#irq_data * ) ; void ( * irq_compose_msi_msg ) ( s#irq_data * , s#msi_msg * ) ; void ( * irq_write_msi_msg ) ( s#irq_data * , s#msi_msg * ) ; int ( * irq_get_irqchip_state ) ( s#irq_data * , e#irqchip_irq_state , t#bool * ) ; int ( * irq_set_irqchip_state ) ( s#irq_data * , e#irqchip_irq_state , t#bool ) ; int ( * irq_set_vcpu_affinity ) ( s#irq_data * , void * ) ; void ( * ipi_send_single ) ( s#irq_data * , unsigned int ) ; void ( * ipi_send_mask ) ( s#irq_data * , const s#cpumask * ) ; unsigned long flags ; }
+s#irq_data struct irq_data { t#u32 mask ; unsigned int irq ; unsigned long hwirq ; s#irq_common_data * common ; s#irq_chip * chip ; s#irq_domain * domain ; s#irq_data * parent_data ; void * chip_data ; }
+e#irq_gc_flags enum irq_gc_flags { IRQ_GC_INIT_MASK_CACHE = 1 << 0 , IRQ_GC_INIT_NESTED_LOCK = 1 << 1 , IRQ_GC_MASK_CACHE_PER_TYPE = 1 << 2 , IRQ_GC_NO_MASK = 1 << 3 , IRQ_GC_BE_IO = 1 << 4 , }
+s#irq_chip_regs struct irq_chip_regs { unsigned long enable ; unsigned long disable ; unsigned long mask ; unsigned long ack ; unsigned long eoi ; unsigned long type ; unsigned long polarity ; }
+e#irqreturn enum irqreturn { IRQ_NONE = ( 0 << 0 ) , IRQ_HANDLED = ( 1 << 0 ) , IRQ_WAKE_THREAD = ( 1 << 1 ) , }
+t#irqreturn_t typedef e#irqreturn irqreturn_t
+t#irq_handler_t typedef t#irqreturn_t ( * irq_handler_t ) ( int , void * )
+s##proc_dir_entry
+s#irqaction struct irqaction { t#irq_handler_t handler ; void * dev_id ; void * percpu_dev_id ; s#irqaction * next ; t#irq_handler_t thread_fn ; s#task_struct * thread ; s#irqaction * secondary ; unsigned int irq ; unsigned int flags ; unsigned long thread_flags ; unsigned long thread_mask ; const char * name ; s#proc_dir_entry * dir ; }
+s#irq_affinity_notify struct irq_affinity_notify { unsigned int irq ; s#kref kref ; s#work_struct work ; void ( * notify ) ( s#irq_affinity_notify * , const t#cpumask_t * ) ; void ( * release ) ( s#kref * ) ; }
+s#irq_desc struct irq_desc { s#irq_common_data irq_common_data ; s#irq_data irq_data ; unsigned int * kstat_irqs ; t#irq_flow_handler_t handle_irq ; s#irqaction * action ; unsigned int status_use_accessors ; unsigned int core_internal_state__do_not_mess_with_it ; unsigned int depth ; unsigned int wake_depth ; unsigned int irq_count ; unsigned long last_unhandled ; unsigned int irqs_unhandled ; t#atomic_t threads_handled ; int threads_handled_last ; t#raw_spinlock_t lock ; s#cpumask * percpu_enabled ; const s#cpumask * percpu_affinity ; const s#cpumask * affinity_hint ; s#irq_affinity_notify * affinity_notify ; unsigned long threads_oneshot ; t#atomic_t threads_active ; t#wait_queue_head_t wait_for_threads ; unsigned int nr_actions ; unsigned int no_suspend_depth ; unsigned int cond_suspend_depth ; unsigned int force_resume_depth ; s#proc_dir_entry * dir ; s#callback_head rcu ; s#kobject kobj ; int parent_irq ; s#module * owner ; const char * name ; }
+t#irq_flow_handler_t typedef void ( * irq_flow_handler_t ) ( s#irq_desc * )
+s#irq_chip_type struct irq_chip_type { s#irq_chip chip ; s#irq_chip_regs regs ; t#irq_flow_handler_t handler ; t#u32 type ; t#u32 mask_cache_priv ; t#u32 * mask_cache ; }
+s#irq_chip_generic struct irq_chip_generic { t#raw_spinlock_t lock ; void * reg_base ; t#u32 ( * reg_readl ) ( void * ) ; void ( * reg_writel ) ( t#u32 , void * ) ; void ( * suspend ) ( s#irq_chip_generic * ) ; void ( * resume ) ( s#irq_chip_generic * ) ; unsigned int irq_base ; unsigned int irq_cnt ; t#u32 mask_cache ; t#u32 type_cache ; t#u32 polarity_cache ; t#u32 wake_enabled ; t#u32 wake_active ; unsigned int num_ct ; void * private ; unsigned long installed ; unsigned long unused ; s#irq_domain * domain ; s#list_head list ; s#irq_chip_type chip_types [ 0 ] ; }
+s#irq_domain_chip_generic struct irq_domain_chip_generic { unsigned int irqs_per_chip ; unsigned int num_chips ; unsigned int irq_flags_to_clear ; unsigned int irq_flags_to_set ; e#irq_gc_flags gc_flags ; s#irq_chip_generic * gc [ 0 ] ; }
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##iommu_group
+s##iommu_fwspec
+acpi_register_gsi int acpi_register_gsi ( s#device * , t#u32 , int , int )
+acpi_unregister_gsi void acpi_unregister_gsi ( t#u32 )
+acpi_irq_get int acpi_irq_get ( t#acpi_handle , unsigned int , s#resource * )
+
+/* drivers/acpi/numa.o */
+acpi_map_pxm_to_node extern int acpi_map_pxm_to_node ( int )
+acpi_map_pxm_to_online_node int acpi_map_pxm_to_online_node ( int )
+acpi_get_node int acpi_get_node ( t#acpi_handle )
+
+/* drivers/acpi/osi.o */
+acpi_osi_is_win8 extern t#bool acpi_osi_is_win8 ( void )
+
+/* drivers/acpi/osl.o */
+acpi_os_printf void acpi_os_printf ( const char * , ... )
+acpi_os_get_iomem void * acpi_os_get_iomem ( t#acpi_physical_address , unsigned int )
+acpi_os_map_iomem void * __attribute__ ( ( __section__ ( ".ref.text" ) ) ) __attribute__ ( ( noinline ) ) acpi_os_map_iomem ( t#acpi_physical_address , t#acpi_size )
+acpi_os_map_memory void * acpi_os_map_memory ( t#acpi_physical_address , t#acpi_size )
+acpi_os_unmap_iomem void __attribute__ ( ( __section__ ( ".ref.text" ) ) ) __attribute__ ( ( noinline ) ) acpi_os_unmap_iomem ( void * , t#acpi_size )
+acpi_os_unmap_memory void acpi_os_unmap_memory ( void * , t#acpi_size )
+acpi_os_map_generic_address int acpi_os_map_generic_address ( s#acpi_generic_address * )
+acpi_os_unmap_generic_address void acpi_os_unmap_generic_address ( s#acpi_generic_address * )
+acpi_os_read_port t#acpi_status acpi_os_read_port ( t#acpi_io_address , t#u32 * , t#u32 )
+acpi_os_write_port t#acpi_status acpi_os_write_port ( t#acpi_io_address , t#u32 , t#u32 )
+t#acpi_execute_type typedef enum { OSL_GLOBAL_LOCK_HANDLER , OSL_NOTIFY_HANDLER , OSL_GPE_HANDLER , OSL_DEBUGGER_MAIN_THREAD , OSL_DEBUGGER_EXEC_THREAD , OSL_EC_POLL_HANDLER , OSL_EC_BURST_HANDLER } acpi_execute_type
+t#acpi_osd_exec_callback typedef void ( * acpi_osd_exec_callback ) ( void * )
+acpi_os_execute t#acpi_status acpi_os_execute ( t#acpi_execute_type , t#acpi_osd_exec_callback , void * )
+acpi_os_get_line t#acpi_status acpi_os_get_line ( char * , t#u32 , t#u32 * )
+acpi_check_resource_conflict int acpi_check_resource_conflict ( const s#resource * )
+acpi_check_region int acpi_check_region ( t#resource_size_t , t#resource_size_t , const char * )
+acpi_release_memory t#acpi_status acpi_release_memory ( t#acpi_handle , s#resource * , t#u32 )
+acpi_resources_are_enforced int acpi_resources_are_enforced ( void )
+
+/* drivers/acpi/pci_root.o */
+acpi_is_root_bridge int acpi_is_root_bridge ( t#acpi_handle )
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##proc_dir_entry
+s##hotplug_slot
+s##pcie_link_state
+s##pci_vpd
+s##pci_sriov
+s##platform_msi_priv_data
+s#acpi_pci_root struct acpi_pci_root { s#acpi_device * device ; s#pci_bus * bus ; t#u16 segment ; s#resource secondary ; t#u32 osc_support_set ; t#u32 osc_control_set ; t#phys_addr_t mcfg_addr ; }
+acpi_pci_find_root s#acpi_pci_root * acpi_pci_find_root ( t#acpi_handle )
+acpi_get_pci_dev s#pci_dev * acpi_get_pci_dev ( t#acpi_handle )
+acpi_pci_osc_control_set extern t#acpi_status acpi_pci_osc_control_set ( t#acpi_handle , t#u32 * , t#u32 )
+
+/* drivers/acpi/processor_core.o */
+acpi_get_cpuid int acpi_get_cpuid ( t#acpi_handle , int , t#u32 )
+
+/* drivers/acpi/processor_perflib.o */
+acpi_processor_get_bios_limit extern int acpi_processor_get_bios_limit ( int , unsigned int * )
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##iommu_group
+s##iommu_fwspec
+acpi_processor_get_performance_info extern int acpi_processor_get_performance_info ( s#acpi_processor * )
+acpi_processor_notify_smm int acpi_processor_notify_smm ( s#module * )
+acpi_processor_preregister_performance extern int acpi_processor_preregister_performance ( s#acpi_processor_performance * )
+acpi_processor_register_performance extern int acpi_processor_register_performance ( s#acpi_processor_performance * , unsigned int )
+acpi_processor_unregister_performance extern void acpi_processor_unregister_performance ( unsigned int )
+
+/* drivers/acpi/property.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##proc_dir_entry
+acpi_dev_get_property int acpi_dev_get_property ( const s#acpi_device * , const char * , t#acpi_object_type , const u#acpi_object * * )
+__acpi_node_get_property_reference int __acpi_node_get_property_reference ( const s#fwnode_handle * , const char * , t#size_t , t#size_t , s#fwnode_reference_args * )
+acpi_device_fwnode_ops extern const s#fwnode_operations acpi_device_fwnode_ops
+acpi_data_fwnode_ops extern const s#fwnode_operations acpi_data_fwnode_ops
+is_acpi_device_node t#bool is_acpi_device_node ( const s#fwnode_handle * )
+is_acpi_data_node t#bool is_acpi_data_node ( const s#fwnode_handle * )
+
+/* drivers/acpi/resource.o */
+acpi_dev_resource_memory t#bool acpi_dev_resource_memory ( s#acpi_resource * , s#resource * )
+acpi_dev_resource_io t#bool acpi_dev_resource_io ( s#acpi_resource * , s#resource * )
+s#resource_win struct resource_win { s#resource res ; t#resource_size_t offset ; }
+acpi_dev_resource_address_space t#bool acpi_dev_resource_address_space ( s#acpi_resource * , s#resource_win * )
+acpi_dev_resource_ext_address_space t#bool acpi_dev_resource_ext_address_space ( s#acpi_resource * , s#resource_win * )
+acpi_dev_irq_flags unsigned long acpi_dev_irq_flags ( t#u8 , t#u8 , t#u8 )
+acpi_dev_get_irq_type unsigned int acpi_dev_get_irq_type ( int , int )
+acpi_dev_resource_interrupt t#bool acpi_dev_resource_interrupt ( s#acpi_resource * , int , s#resource * )
+acpi_dev_free_resource_list void acpi_dev_free_resource_list ( s#list_head * )
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##proc_dir_entry
+acpi_dev_get_resources int acpi_dev_get_resources ( s#acpi_device * , s#list_head * , int ( * ) ( s#acpi_resource * , void * ) , void * )
+acpi_dev_filter_resource_type int acpi_dev_filter_resource_type ( s#acpi_resource * , unsigned long )
+
+/* drivers/acpi/scan.o */
+acpi_scan_lock_acquire void acpi_scan_lock_acquire ( void )
+acpi_scan_lock_release void acpi_scan_lock_release ( void )
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##mem_cgroup
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##bpf_prog_array
+s##cgroup_taskset
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_data
+s##irq_domain_chip_generic
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##iommu_group
+s##iommu_fwspec
+s##proc_dir_entry
+acpi_initialize_hp_context void acpi_initialize_hp_context ( s#acpi_device * , s#acpi_hotplug_context * , int ( * ) ( s#acpi_device * , t#u32 ) , void ( * ) ( s#acpi_device * , t#u32 ) )
+acpi_bus_get_device int acpi_bus_get_device ( t#acpi_handle , s#acpi_device * * )
+acpi_bus_get_ejd t#acpi_status acpi_bus_get_ejd ( t#acpi_handle , t#acpi_handle * )
+acpi_is_video_device extern long acpi_is_video_device ( t#acpi_handle )
+acpi_device_hid const char * acpi_device_hid ( s#acpi_device * )
+e#dev_dma_attr enum dev_dma_attr { DEV_DMA_NOT_SUPPORTED , DEV_DMA_NON_COHERENT , DEV_DMA_COHERENT , }
+acpi_dma_configure int acpi_dma_configure ( s#device * , e#dev_dma_attr )
+acpi_dma_deconfigure void acpi_dma_deconfigure ( s#device * )
+acpi_walk_dep_device_list void acpi_walk_dep_device_list ( t#acpi_handle )
+acpi_bus_scan int acpi_bus_scan ( t#acpi_handle )
+acpi_bus_trim void acpi_bus_trim ( s#acpi_device * )
+acpi_bus_register_early_device int acpi_bus_register_early_device ( int )
+acpi_reconfig_notifier_register int acpi_reconfig_notifier_register ( s#notifier_block * )
+acpi_reconfig_notifier_unregister int acpi_reconfig_notifier_unregister ( s#notifier_block * )
+
+/* drivers/acpi/spcr.o */
+qdf2400_e44_present extern t#bool qdf2400_e44_present
+
+/* drivers/acpi/utils.o */
+acpi_extract_package t#acpi_status acpi_extract_package ( u#acpi_object * , s#acpi_buffer * , s#acpi_buffer * )
+acpi_evaluate_integer t#acpi_status acpi_evaluate_integer ( t#acpi_handle , t#acpi_string , s#acpi_object_list * , unsigned long long * )
+s#acpi_handle_list struct acpi_handle_list { t#u32 count ; t#acpi_handle handles [ 10 ] ; }
+acpi_evaluate_reference t#acpi_status acpi_evaluate_reference ( t#acpi_handle , t#acpi_string , s#acpi_object_list * , s#acpi_handle_list * )
+acpi_get_physical_device_location t#acpi_status acpi_get_physical_device_location ( t#acpi_handle , s#acpi_pld_info * * )
+acpi_evaluate_ost t#acpi_status acpi_evaluate_ost ( t#acpi_handle , t#u32 , t#u32 , s#acpi_buffer * )
+acpi_handle_printk __attribute__ ( ( format ( printf , 3 , 4 ) ) ) void acpi_handle_printk ( const char * , t#acpi_handle , const char * , ... )
+s#_ddebug struct _ddebug { const char * modname ; const char * function ; const char * filename ; const char * format ; unsigned int lineno : 18 ; unsigned int flags : 8 ; }
+__acpi_handle_debug __attribute__ ( ( format ( printf , 3 , 4 ) ) ) void __acpi_handle_debug ( s#_ddebug * , t#acpi_handle , const char * , ... )
+acpi_has_method t#bool acpi_has_method ( t#acpi_handle , char * )
+acpi_execute_simple_method t#acpi_status acpi_execute_simple_method ( t#acpi_handle , char * , t#u64 )
+acpi_evaluate_dsm u#acpi_object * acpi_evaluate_dsm ( t#acpi_handle , const t#guid_t * , t#u64 , t#u64 , u#acpi_object * )
+acpi_check_dsm t#bool acpi_check_dsm ( t#acpi_handle , const t#guid_t * , t#u64 , t#u64 )
+acpi_dev_found t#bool acpi_dev_found ( const char * )
+acpi_dev_present t#bool acpi_dev_present ( const char * , const char * , t#s64 )
+acpi_video_backlight_string extern char acpi_video_backlight_string [ ]
+e#acpi_predicate enum acpi_predicate { all_versions , less_than_or_equal , equal , greater_than_or_equal , }
+s#acpi_platform_list struct acpi_platform_list { char oem_id [ 6 + 1 ] ; char oem_table_id [ 8 + 1 ] ; t#u32 oem_revision ; char * table ; e#acpi_predicate pred ; char * reason ; t#u32 data ; }
+acpi_match_platform_list int acpi_match_platform_list ( const s#acpi_platform_list * )
+
+/* drivers/amba/bus.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##exception_table_entry
+s##module_sect_attrs
+s##module_notes_attrs
+s##tracepoint
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##writeback_control
+s##kstatfs
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##vm_operations_struct
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##user_namespace
+s##mmu_notifier_mm
+s##xol_area
+s##pollfd
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##nsproxy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##bio_list
+s##blk_plug
+s##reclaim_state
+s##backing_dev_info
+s##io_context
+s##css_set
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##mem_cgroup
+s##uprobe
+s##perf_event
+s##iov_iter
+s##swap_info_struct
+s##dev_pagemap
+s##kmem_cache
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##gendisk
+s##request_queue
+s##mtd_info
+s##workqueue_struct
+s##bdi_writeback
+s##poll_table_struct
+s##nlm_lockowner
+s##nfs4_lock_state
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_map_ops
+s##dma_coherent_mem
+s##cma
+s##iommu_group
+s##iommu_fwspec
+s##clk
+s#amba_device struct amba_device { s#device dev ; s#resource res ; s#clk * pclk ; unsigned int periphid ; unsigned int irq [ 9 ] ; char * driver_override ; }
+amba_device_add int amba_device_add ( s#amba_device * , s#resource * )
+amba_apb_device_add s#amba_device * amba_apb_device_add ( s#device * , const char * , t#resource_size_t , t#size_t , int , int , void * , unsigned int )
+amba_ahb_device_add s#amba_device * amba_ahb_device_add ( s#device * , const char * , t#resource_size_t , t#size_t , int , int , void * , unsigned int )
+amba_apb_device_add_res s#amba_device * amba_apb_device_add_res ( s#device * , const char * , t#resource_size_t , t#size_t , int , int , void * , unsigned int , s#resource * )
+amba_ahb_device_add_res s#amba_device * amba_ahb_device_add_res ( s#device * , const char * , t#resource_size_t , t#size_t , int , int , void * , unsigned int , s#resource * )
+amba_device_alloc s#amba_device * amba_device_alloc ( const char * , t#resource_size_t , t#size_t )
+amba_device_put void amba_device_put ( s#amba_device * )
+s#amba_id struct amba_id { unsigned int id ; unsigned int mask ; void * data ; }
+s#amba_driver struct amba_driver { s#device_driver drv ; int ( * probe ) ( s#amba_device * , const s#amba_id * ) ; int ( * remove ) ( s#amba_device * ) ; void ( * shutdown ) ( s#amba_device * ) ; const s#amba_id * id_table ; }
+amba_driver_register int amba_driver_register ( s#amba_driver * )
+amba_driver_unregister void amba_driver_unregister ( s#amba_driver * )
+amba_device_register int amba_device_register ( s#amba_device * , s#resource * )
+amba_device_unregister void amba_device_unregister ( s#amba_device * )
+amba_find_device s#amba_device * amba_find_device ( const char * , s#device * , unsigned int , unsigned int )
+amba_request_regions int amba_request_regions ( s#amba_device * , const char * )
+amba_release_regions void amba_release_regions ( s#amba_device * )
+
+/* drivers/ata/libahci.o */
+ahci_ignore_sss extern int ahci_ignore_sss
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##iov_iter
+s##hd_geometry
+s##pr_ops
+s##blk_mq_ctx
+s##kmem_cache
+s##blkcg_gq
+s##blk_flush_queue
+s##blk_mq_tags
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_stat_callback
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##iommu_group
+s##iommu_fwspec
+ahci_shost_attrs extern s#device_attribute * ahci_shost_attrs [ ]
+ahci_sdev_attrs extern s#device_attribute * ahci_sdev_attrs [ ]
+s#scsi_vpd struct scsi_vpd { s#callback_head rcu ; int len ; unsigned char data [ ] ; }
+e#scsi_target_state enum scsi_target_state { STARGET_CREATED = 1 , STARGET_RUNNING , STARGET_REMOVE , STARGET_CREATED_REMOVE , STARGET_DEL , }
+s#scsi_target struct scsi_target { s#scsi_device * starget_sdev_user ; s#list_head siblings ; s#list_head devices ; s#device dev ; s#kref reap_ref ; unsigned int channel ; unsigned int id ; unsigned int create : 1 ; unsigned int single_lun : 1 ; unsigned int pdt_1f_for_no_lun : 1 ; unsigned int no_report_luns : 1 ; unsigned int expecting_lun_change : 1 ; t#atomic_t target_busy ; t#atomic_t target_blocked ; unsigned int can_queue ; unsigned int max_target_blocked ; char scsi_level ; e#scsi_target_state state ; void * hostdata ; unsigned long starget_data [ 0 ] ; }
+t#blist_flags_t typedef t#__u64 blist_flags_t
+E#SDEV_EVT_POWER_ON_RESET_OCCURRED ( 1 ) + 7
+E#SDEV_EVT_LAST E#SDEV_EVT_POWER_ON_RESET_OCCURRED
+E#SDEV_EVT_MAXBITS E#SDEV_EVT_LAST + 1
+s#execute_work struct execute_work { s#work_struct work ; }
+s##scsi_device_handler
+e#scsi_device_state enum scsi_device_state { SDEV_CREATED = 1 , SDEV_RUNNING , SDEV_CANCEL , SDEV_DEL , SDEV_QUIESCE , SDEV_OFFLINE , SDEV_TRANSPORT_OFFLINE , SDEV_BLOCK , SDEV_CREATED_BLOCK , }
+s#scsi_device struct scsi_device { s#Scsi_Host * host ; s#request_queue * request_queue ; s#list_head siblings ; s#list_head same_target_siblings ; t#atomic_t device_busy ; t#atomic_t device_blocked ; t#spinlock_t list_lock ; s#list_head cmd_list ; s#list_head starved_entry ; unsigned short queue_depth ; unsigned short max_queue_depth ; unsigned short last_queue_full_depth ; unsigned short last_queue_full_count ; unsigned long last_queue_full_time ; unsigned long queue_ramp_up_period ; unsigned long last_queue_ramp_up ; unsigned int id , channel ; t#u64 lun ; unsigned int manufacturer ; unsigned sector_size ; void * hostdata ; unsigned char type ; char scsi_level ; char inq_periph_qual ; s#mutex inquiry_mutex ; unsigned char inquiry_len ; unsigned char * inquiry ; const char * vendor ; const char * model ; const char * rev ; s#scsi_vpd * vpd_pg83 ; s#scsi_vpd * vpd_pg80 ; unsigned char current_tag ; s#scsi_target * sdev_target ; t#blist_flags_t sdev_bflags ; unsigned int eh_timeout ; unsigned removable : 1 ; unsigned changed : 1 ; unsigned busy : 1 ; unsigned lockable : 1 ; unsigned locked : 1 ; unsigned borken : 1 ; unsigned disconnect : 1 ; unsigned soft_reset : 1 ; unsigned sdtr : 1 ; unsigned wdtr : 1 ; unsigned ppr : 1 ; unsigned tagged_supported : 1 ; unsigned simple_tags : 1 ; unsigned was_reset : 1 ; unsigned expecting_cc_ua : 1 ; unsigned use_10_for_rw : 1 ; unsigned use_10_for_ms : 1 ; unsigned no_report_opcodes : 1 ; unsigned no_write_same : 1 ; unsigned use_16_for_rw : 1 ; unsigned skip_ms_page_8 : 1 ; unsigned skip_ms_page_3f : 1 ; unsigned skip_vpd_pages : 1 ; unsigned try_vpd_pages : 1 ; unsigned use_192_bytes_for_3f : 1 ; unsigned no_start_on_add : 1 ; unsigned allow_restart : 1 ; unsigned manage_start_stop : 1 ; unsigned start_stop_pwr_cond : 1 ; unsigned no_uld_attach : 1 ; unsigned select_no_atn : 1 ; unsigned fix_capacity : 1 ; unsigned guess_capacity : 1 ; unsigned retry_hwerror : 1 ; unsigned last_sector_bug : 1 ; unsigned no_read_disc_info : 1 ; unsigned no_read_capacity_16 : 1 ; unsigned try_rc_10_first : 1 ; unsigned security_supported : 1 ; unsigned is_visible : 1 ; unsigned wce_default_on : 1 ; unsigned no_dif : 1 ; unsigned broken_fua : 1 ; unsigned lun_in_cdb : 1 ; unsigned unmap_limit_for_ws : 1 ; t#atomic_t disk_events_disable_depth ; unsigned long supported_events [ ( ( ( E#SDEV_EVT_MAXBITS ) + ( 8 * sizeof ( long ) ) - 1 ) / ( 8 * sizeof ( long ) ) ) ] ; unsigned long pending_events [ ( ( ( E#SDEV_EVT_MAXBITS ) + ( 8 * sizeof ( long ) ) - 1 ) / ( 8 * sizeof ( long ) ) ) ] ; s#list_head event_list ; s#work_struct event_work ; unsigned int max_device_blocked ; t#atomic_t iorequest_cnt ; t#atomic_t iodone_cnt ; t#atomic_t ioerr_cnt ; s#device sdev_gendev , sdev_dev ; s#execute_work ew ; s#work_struct requeue_work ; s#scsi_device_handler * handler ; void * handler_data ; unsigned char access_state ; s#mutex state_mutex ; e#scsi_device_state sdev_state ; s#task_struct * quiesced_by ; unsigned long sdev_data [ 0 ] ; }
+s#scsi_data_buffer struct scsi_data_buffer { s#sg_table table ; unsigned length ; int resid ; }
+s#scsi_pointer struct scsi_pointer { char * ptr ; int this_residual ; s#scatterlist * buffer ; int buffers_residual ; t#dma_addr_t dma_handle ; volatile int Status ; volatile int Message ; volatile int have_data_in ; volatile int sent_command ; volatile int phase ; }
+s#scsi_cmnd struct scsi_cmnd { s#scsi_request req ; s#scsi_device * device ; s#list_head list ; s#list_head eh_entry ; s#delayed_work abort_work ; s#callback_head rcu ; int eh_eflags ; unsigned long serial_number ; unsigned long jiffies_at_alloc ; int retries ; int allowed ; unsigned char prot_op ; unsigned char prot_type ; unsigned char prot_flags ; unsigned short cmd_len ; e#dma_data_direction sc_data_direction ; unsigned char * cmnd ; s#scsi_data_buffer sdb ; s#scsi_data_buffer * prot_sdb ; unsigned underflow ; unsigned transfersize ; s#request * request ; unsigned char * sense_buffer ; void ( * scsi_done ) ( s#scsi_cmnd * ) ; s#scsi_pointer SCp ; unsigned char * host_scribble ; int result ; int flags ; unsigned char tag ; }
+s##proc_dir_entry
+s##scsi_host_cmd_pool
+s#scsi_host_template struct scsi_host_template { s#module * module ; const char * name ; int ( * detect ) ( s#scsi_host_template * ) ; int ( * release ) ( s#Scsi_Host * ) ; const char * ( * info ) ( s#Scsi_Host * ) ; int ( * ioctl ) ( s#scsi_device * , int , void * ) ; int ( * compat_ioctl ) ( s#scsi_device * , int , void * ) ; int ( * queuecommand ) ( s#Scsi_Host * , s#scsi_cmnd * ) ; int ( * eh_abort_handler ) ( s#scsi_cmnd * ) ; int ( * eh_device_reset_handler ) ( s#scsi_cmnd * ) ; int ( * eh_target_reset_handler ) ( s#scsi_cmnd * ) ; int ( * eh_bus_reset_handler ) ( s#scsi_cmnd * ) ; int ( * eh_host_reset_handler ) ( s#scsi_cmnd * ) ; int ( * slave_alloc ) ( s#scsi_device * ) ; int ( * slave_configure ) ( s#scsi_device * ) ; void ( * slave_destroy ) ( s#scsi_device * ) ; int ( * target_alloc ) ( s#scsi_target * ) ; void ( * target_destroy ) ( s#scsi_target * ) ; int ( * scan_finished ) ( s#Scsi_Host * , unsigned long ) ; void ( * scan_start ) ( s#Scsi_Host * ) ; int ( * change_queue_depth ) ( s#scsi_device * , int ) ; int ( * map_queues ) ( s#Scsi_Host * ) ; int ( * bios_param ) ( s#scsi_device * , s#block_device * , t#sector_t , int [ ] ) ; void ( * unlock_native_capacity ) ( s#scsi_device * ) ; int ( * show_info ) ( s#seq_file * , s#Scsi_Host * ) ; int ( * write_info ) ( s#Scsi_Host * , char * , int ) ; e#blk_eh_timer_return ( * eh_timed_out ) ( s#scsi_cmnd * ) ; int ( * host_reset ) ( s#Scsi_Host * , int ) ; const char * proc_name ; s#proc_dir_entry * proc_dir ; int can_queue ; int this_id ; unsigned short sg_tablesize ; unsigned short sg_prot_tablesize ; unsigned int max_sectors ; unsigned long dma_boundary ; short cmd_per_lun ; unsigned char present ; int tag_alloc_policy ; unsigned track_queue_depth : 1 ; unsigned supported_mode : 2 ; unsigned unchecked_isa_dma : 1 ; unsigned use_clustering : 1 ; unsigned emulated : 1 ; unsigned skip_settle_delay : 1 ; unsigned no_write_same : 1 ; unsigned force_blk_mq : 1 ; unsigned int max_host_blocked ; s#device_attribute * * shost_attrs ; s#device_attribute * * sdev_attrs ; const s#attribute_group * * sdev_groups ; s#list_head legacy_hosts ; t#u64 vendor_id ; unsigned int cmd_size ; s#scsi_host_cmd_pool * cmd_pool ; }
+s##scsi_transport_template
+e#scsi_host_state enum scsi_host_state { SHOST_CREATED = 1 , SHOST_RUNNING , SHOST_CANCEL , SHOST_DEL , SHOST_RECOVERY , SHOST_CANCEL_RECOVERY , SHOST_DEL_RECOVERY , }
+s#Scsi_Host struct Scsi_Host { s#list_head __devices ; s#list_head __targets ; s#list_head starved_list ; t#spinlock_t default_lock ; t#spinlock_t * host_lock ; s#mutex scan_mutex ; s#list_head eh_cmd_q ; s#task_struct * ehandler ; s#completion * eh_action ; t#wait_queue_head_t host_wait ; s#scsi_host_template * hostt ; s#scsi_transport_template * transportt ; union { s#blk_queue_tag * bqt ; s#blk_mq_tag_set tag_set ; } ; t#atomic_t host_busy ; t#atomic_t host_blocked ; unsigned int host_failed ; unsigned int host_eh_scheduled ; unsigned int host_no ; int eh_deadline ; unsigned long last_reset ; unsigned int max_channel ; unsigned int max_id ; t#u64 max_lun ; unsigned int unique_id ; unsigned short max_cmd_len ; int this_id ; int can_queue ; short cmd_per_lun ; short unsigned int sg_tablesize ; short unsigned int sg_prot_tablesize ; unsigned int max_sectors ; unsigned long dma_boundary ; unsigned nr_hw_queues ; unsigned long cmd_serial_number ; unsigned active_mode : 2 ; unsigned unchecked_isa_dma : 1 ; unsigned use_clustering : 1 ; unsigned host_self_blocked : 1 ; unsigned reverse_ordering : 1 ; unsigned tmf_in_progress : 1 ; unsigned async_scan : 1 ; unsigned eh_noresume : 1 ; unsigned no_write_same : 1 ; unsigned use_blk_mq : 1 ; unsigned use_cmd_list : 1 ; unsigned short_inquiry : 1 ; char work_q_name [ 20 ] ; s#workqueue_struct * work_q ; s#workqueue_struct * tmf_work_q ; unsigned no_scsi2_lun_in_cdb : 1 ; unsigned int max_host_blocked ; unsigned int prot_capabilities ; unsigned char prot_guard_type ; unsigned long base ; unsigned long io_port ; unsigned char n_io_port ; unsigned char dma_channel ; unsigned int irq ; e#scsi_host_state shost_state ; s#device shost_gendev , shost_dev ; s#list_head sht_legacy_list ; void * shost_data ; s#device * dma_dev ; unsigned long hostdata [ 0 ] __attribute__ ( ( aligned ( sizeof ( unsigned long ) ) ) ) ; }
+s#ata_ioports struct ata_ioports { void * cmd_addr ; void * data_addr ; void * error_addr ; void * feature_addr ; void * nsect_addr ; void * lbal_addr ; void * lbam_addr ; void * lbah_addr ; void * device_addr ; void * status_addr ; void * command_addr ; void * altstatus_addr ; void * ctl_addr ; void * bmdma_addr ; void * scr_addr ; }
+e#ata_lpm_policy enum ata_lpm_policy { ATA_LPM_UNKNOWN , ATA_LPM_MAX_POWER , ATA_LPM_MED_POWER , ATA_LPM_MIN_POWER , }
+E#ATA_ID_WORDS 256
+E#SATA_PMP_GSCR_DWORDS 128
+E#ATA_LOG_DEVSLP_SIZE 0x08
+E#ATA_LOG_NCQ_SEND_RECV_SIZE 0x14
+E#ATA_LOG_NCQ_NON_DATA_SIZE 0x40
+s#ata_ering_entry struct ata_ering_entry { unsigned int eflags ; unsigned int err_mask ; t#u64 timestamp ; }
+E#ATA_ERING_SIZE 32
+s#ata_ering struct ata_ering { int cursor ; s#ata_ering_entry ring [ E#ATA_ERING_SIZE ] ; }
+s#ata_device struct ata_device { s#ata_link * link ; unsigned int devno ; unsigned int horkage ; unsigned long flags ; s#scsi_device * sdev ; void * private_data ; u#acpi_object * gtf_cache ; unsigned int gtf_filter ; s#device tdev ; t#u64 n_sectors ; t#u64 n_native_sectors ; unsigned int class ; unsigned long unpark_deadline ; t#u8 pio_mode ; t#u8 dma_mode ; t#u8 xfer_mode ; unsigned int xfer_shift ; unsigned int multi_count ; unsigned int max_sectors ; unsigned int cdb_len ; unsigned long pio_mask ; unsigned long mwdma_mask ; unsigned long udma_mask ; t#u16 cylinders ; t#u16 heads ; t#u16 sectors ; union { t#u16 id [ E#ATA_ID_WORDS ] ; t#u32 gscr [ E#SATA_PMP_GSCR_DWORDS ] ; } __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; t#u8 devslp_timing [ E#ATA_LOG_DEVSLP_SIZE ] ; t#u8 ncq_send_recv_cmds [ E#ATA_LOG_NCQ_SEND_RECV_SIZE ] ; t#u8 ncq_non_data_cmds [ E#ATA_LOG_NCQ_NON_DATA_SIZE ] ; t#u32 zac_zoned_cap ; t#u32 zac_zones_optimal_open ; t#u32 zac_zones_optimal_nonseq ; t#u32 zac_zones_max_open ; int spdn_cnt ; s#ata_ering ering ; }
+E#ATA_MAX_DEVICES 2
+E#ATA_EH_DESC_LEN 80
+s#ata_eh_info struct ata_eh_info { s#ata_device * dev ; t#u32 serror ; unsigned int err_mask ; unsigned int action ; unsigned int dev_action [ E#ATA_MAX_DEVICES ] ; unsigned int flags ; unsigned int probe_mask ; char desc [ E#ATA_EH_DESC_LEN ] ; int desc_len ; }
+E#ATA_EH_CMD_TIMEOUT_TABLE_SIZE 6
+s#ata_eh_context struct ata_eh_context { s#ata_eh_info i ; int tries [ E#ATA_MAX_DEVICES ] ; int cmd_timeout_idx [ E#ATA_MAX_DEVICES ] [ E#ATA_EH_CMD_TIMEOUT_TABLE_SIZE ] ; unsigned int classes [ E#ATA_MAX_DEVICES ] ; unsigned int did_probe_mask ; unsigned int unloaded_mask ; unsigned int saved_ncq_enabled ; t#u8 saved_xfer_mode [ E#ATA_MAX_DEVICES ] ; unsigned long last_reset ; }
+s#ata_link struct ata_link { s#ata_port * ap ; int pmp ; s#device tdev ; unsigned int active_tag ; t#u32 sactive ; unsigned int flags ; t#u32 saved_scontrol ; unsigned int hw_sata_spd_limit ; unsigned int sata_spd_limit ; unsigned int sata_spd ; e#ata_lpm_policy lpm_policy ; s#ata_eh_info eh_info ; s#ata_eh_context eh_context ; s#ata_device device [ E#ATA_MAX_DEVICES ] ; unsigned long last_lpm_change ; }
+t#__le32 typedef t#__u32 __le32
+s#ata_bmdma_prd struct ata_bmdma_prd { t#__le32 addr ; t#__le32 flags_len ; }
+E#ATA_MAX_QUEUE 32
+s#ata_port_stats struct ata_port_stats { unsigned long unhandled_irq ; unsigned long idle_irq ; unsigned long rw_reqbuf ; }
+s#ata_host struct ata_host { t#spinlock_t lock ; s#device * dev ; void * const * iomap ; unsigned int n_ports ; unsigned int n_tags ; void * private_data ; s#ata_port_operations * ops ; unsigned long flags ; s#kref kref ; s#mutex eh_mutex ; s#task_struct * eh_owner ; s#ata_port * simplex_claimed ; s#ata_port * ports [ 0 ] ; }
+s#ata_acpi_drive struct ata_acpi_drive { t#u32 pio ; t#u32 dma ; }
+s#ata_acpi_gtm struct ata_acpi_gtm { s#ata_acpi_drive drive [ 2 ] ; t#u32 flags ; }
+E#ATA_SECT_SIZE 512
+s#ata_port struct ata_port { s#Scsi_Host * scsi_host ; s#ata_port_operations * ops ; t#spinlock_t * lock ; unsigned long flags ; unsigned int pflags ; unsigned int print_id ; unsigned int local_port_no ; unsigned int port_no ; s#ata_ioports ioaddr ; t#u8 ctl ; t#u8 last_ctl ; s#ata_link * sff_pio_task_link ; s#delayed_work sff_pio_task ; s#ata_bmdma_prd * bmdma_prd ; t#dma_addr_t bmdma_prd_dma ; unsigned int pio_mask ; unsigned int mwdma_mask ; unsigned int udma_mask ; unsigned int cbl ; s#ata_queued_cmd qcmd [ E#ATA_MAX_QUEUE ] ; unsigned long sas_tag_allocated ; unsigned int qc_active ; int nr_active_links ; unsigned int sas_last_tag ; s#ata_link link ; s#ata_link * slave_link ; int nr_pmp_links ; s#ata_link * pmp_link ; s#ata_link * excl_link ; s#ata_port_stats stats ; s#ata_host * host ; s#device * dev ; s#device tdev ; s#mutex scsi_scan_mutex ; s#delayed_work hotplug_task ; s#work_struct scsi_rescan_task ; unsigned int hsm_task_state ; t#u32 msg_enable ; s#list_head eh_done_q ; t#wait_queue_head_t eh_wait_q ; int eh_tries ; s#completion park_req_pending ; t#pm_message_t pm_mesg ; e#ata_lpm_policy target_lpm_policy ; s#timer_list fastdrain_timer ; unsigned long fastdrain_cnt ; int em_message_type ; void * private_data ; s#ata_acpi_gtm __acpi_init_gtm ; t#u8 sector_buf [ E#ATA_SECT_SIZE ] __attribute__ ( ( __aligned__ ( ( 1 << ( 6 ) ) ) ) ) ; }
+s#ata_taskfile struct ata_taskfile { unsigned long flags ; t#u8 protocol ; t#u8 ctl ; t#u8 hob_feature ; t#u8 hob_nsect ; t#u8 hob_lbal ; t#u8 hob_lbam ; t#u8 hob_lbah ; t#u8 feature ; t#u8 nsect ; t#u8 lbal ; t#u8 lbam ; t#u8 lbah ; t#u8 device ; t#u8 command ; t#u32 auxiliary ; }
+E#ATAPI_CDB_LEN 16
+t#ata_qc_cb_t typedef void ( * ata_qc_cb_t ) ( s#ata_queued_cmd * )
+s#ata_queued_cmd struct ata_queued_cmd { s#ata_port * ap ; s#ata_device * dev ; s#scsi_cmnd * scsicmd ; void ( * scsidone ) ( s#scsi_cmnd * ) ; s#ata_taskfile tf ; t#u8 cdb [ E#ATAPI_CDB_LEN ] ; unsigned long flags ; unsigned int tag ; unsigned int n_elem ; unsigned int orig_n_elem ; int dma_dir ; unsigned int sect_size ; unsigned int nbytes ; unsigned int extrabytes ; unsigned int curbytes ; s#scatterlist sgent ; s#scatterlist * sg ; s#scatterlist * cursg ; unsigned int cursg_ofs ; unsigned int err_mask ; s#ata_taskfile result_tf ; t#ata_qc_cb_t complete_fn ; void * private_data ; void * lldd_task ; }
+t#ata_prereset_fn_t typedef int ( * ata_prereset_fn_t ) ( s#ata_link * , unsigned long )
+t#ata_reset_fn_t typedef int ( * ata_reset_fn_t ) ( s#ata_link * , unsigned int * , unsigned long )
+t#ata_postreset_fn_t typedef void ( * ata_postreset_fn_t ) ( s#ata_link * , unsigned int * )
+e#sw_activity enum sw_activity { OFF , BLINK_ON , BLINK_OFF , }
+s#ata_port_operations struct ata_port_operations { int ( * qc_defer ) ( s#ata_queued_cmd * ) ; int ( * check_atapi_dma ) ( s#ata_queued_cmd * ) ; void ( * qc_prep ) ( s#ata_queued_cmd * ) ; unsigned int ( * qc_issue ) ( s#ata_queued_cmd * ) ; t#bool ( * qc_fill_rtf ) ( s#ata_queued_cmd * ) ; int ( * cable_detect ) ( s#ata_port * ) ; unsigned long ( * mode_filter ) ( s#ata_device * , unsigned long ) ; void ( * set_piomode ) ( s#ata_port * , s#ata_device * ) ; void ( * set_dmamode ) ( s#ata_port * , s#ata_device * ) ; int ( * set_mode ) ( s#ata_link * , s#ata_device * * ) ; unsigned int ( * read_id ) ( s#ata_device * , s#ata_taskfile * , t#u16 * ) ; void ( * dev_config ) ( s#ata_device * ) ; void ( * freeze ) ( s#ata_port * ) ; void ( * thaw ) ( s#ata_port * ) ; t#ata_prereset_fn_t prereset ; t#ata_reset_fn_t softreset ; t#ata_reset_fn_t hardreset ; t#ata_postreset_fn_t postreset ; t#ata_prereset_fn_t pmp_prereset ; t#ata_reset_fn_t pmp_softreset ; t#ata_reset_fn_t pmp_hardreset ; t#ata_postreset_fn_t pmp_postreset ; void ( * error_handler ) ( s#ata_port * ) ; void ( * lost_interrupt ) ( s#ata_port * ) ; void ( * post_internal_cmd ) ( s#ata_queued_cmd * ) ; void ( * sched_eh ) ( s#ata_port * ) ; void ( * end_eh ) ( s#ata_port * ) ; int ( * scr_read ) ( s#ata_link * , unsigned int , t#u32 * ) ; int ( * scr_write ) ( s#ata_link * , unsigned int , t#u32 ) ; void ( * pmp_attach ) ( s#ata_port * ) ; void ( * pmp_detach ) ( s#ata_port * ) ; int ( * set_lpm ) ( s#ata_link * , e#ata_lpm_policy , unsigned ) ; int ( * port_suspend ) ( s#ata_port * , t#pm_message_t ) ; int ( * port_resume ) ( s#ata_port * ) ; int ( * port_start ) ( s#ata_port * ) ; void ( * port_stop ) ( s#ata_port * ) ; void ( * host_stop ) ( s#ata_host * ) ; void ( * sff_dev_select ) ( s#ata_port * , unsigned int ) ; void ( * sff_set_devctl ) ( s#ata_port * , t#u8 ) ; t#u8 ( * sff_check_status ) ( s#ata_port * ) ; t#u8 ( * sff_check_altstatus ) ( s#ata_port * ) ; void ( * sff_tf_load ) ( s#ata_port * , const s#ata_taskfile * ) ; void ( * sff_tf_read ) ( s#ata_port * , s#ata_taskfile * ) ; void ( * sff_exec_command ) ( s#ata_port * , const s#ata_taskfile * ) ; unsigned int ( * sff_data_xfer ) ( s#ata_queued_cmd * , unsigned char * , unsigned int , int ) ; void ( * sff_irq_on ) ( s#ata_port * ) ; t#bool ( * sff_irq_check ) ( s#ata_port * ) ; void ( * sff_irq_clear ) ( s#ata_port * ) ; void ( * sff_drain_fifo ) ( s#ata_queued_cmd * ) ; void ( * bmdma_setup ) ( s#ata_queued_cmd * ) ; void ( * bmdma_start ) ( s#ata_queued_cmd * ) ; void ( * bmdma_stop ) ( s#ata_queued_cmd * ) ; t#u8 ( * bmdma_status ) ( s#ata_port * ) ; t#ssize_t ( * em_show ) ( s#ata_port * , char * ) ; t#ssize_t ( * em_store ) ( s#ata_port * , const char * , t#size_t ) ; t#ssize_t ( * sw_activity_show ) ( s#ata_device * , char * ) ; t#ssize_t ( * sw_activity_store ) ( s#ata_device * , e#sw_activity ) ; t#ssize_t ( * transmit_led_message ) ( s#ata_port * , t#u32 , t#ssize_t ) ; void ( * phy_reset ) ( s#ata_port * ) ; void ( * eng_timeout ) ( s#ata_port * ) ; const s#ata_port_operations * inherits ; }
+ahci_ops extern s#ata_port_operations ahci_ops
+ahci_pmp_retry_srst_ops extern s#ata_port_operations ahci_pmp_retry_srst_ops
+ahci_em_messages t#bool ahci_em_messages __attribute__ ( ( __section__ ( ".data..read_mostly" ) ) )
+s##clk
+E#AHCI_MAX_CLKS 5
+s##regulator
+e#phy_mode enum phy_mode { PHY_MODE_INVALID , PHY_MODE_USB_HOST , PHY_MODE_USB_DEVICE , PHY_MODE_USB_OTG , PHY_MODE_SGMII , PHY_MODE_2500SGMII , PHY_MODE_10GKR , }
+s#phy_ops struct phy_ops { int ( * init ) ( s#phy * ) ; int ( * exit ) ( s#phy * ) ; int ( * power_on ) ( s#phy * ) ; int ( * power_off ) ( s#phy * ) ; int ( * set_mode ) ( s#phy * , e#phy_mode ) ; int ( * reset ) ( s#phy * ) ; s#module * owner ; }
+s#phy_attrs struct phy_attrs { t#u32 bus_width ; }
+s#phy struct phy { s#device dev ; int id ; const s#phy_ops * ops ; s#mutex mutex ; int init_count ; int power_count ; s#phy_attrs attrs ; s#regulator * pwr ; }
+s#ahci_host_priv struct ahci_host_priv { unsigned int flags ; t#u32 force_port_map ; t#u32 mask_port_map ; void * mmio ; t#u32 cap ; t#u32 cap2 ; t#u32 version ; t#u32 port_map ; t#u32 saved_cap ; t#u32 saved_cap2 ; t#u32 saved_port_map ; t#u32 em_loc ; t#u32 em_buf_sz ; t#u32 em_msg_type ; t#bool got_runtime_pm ; s#clk * clks [ E#AHCI_MAX_CLKS ] ; s#regulator * * target_pwrs ; s#phy * * phys ; unsigned nports ; void * plat_data ; unsigned int irq ; void ( * start_engine ) ( s#ata_port * ) ; int ( * stop_engine ) ( s#ata_port * ) ; t#irqreturn_t ( * irq_handler ) ( int , void * ) ; int ( * get_irq_vector ) ( s#ata_host * , int ) ; }
+ahci_save_initial_config void ahci_save_initial_config ( s#device * , s#ahci_host_priv * )
+ahci_start_engine void ahci_start_engine ( s#ata_port * )
+ahci_stop_engine int ahci_stop_engine ( s#ata_port * )
+ahci_start_fis_rx void ahci_start_fis_rx ( s#ata_port * )
+ahci_reset_controller int ahci_reset_controller ( s#ata_host * )
+ahci_reset_em int ahci_reset_em ( s#ata_host * )
+ahci_init_controller void ahci_init_controller ( s#ata_host * )
+ahci_dev_classify unsigned int ahci_dev_classify ( s#ata_port * )
+s#ahci_cmd_hdr struct ahci_cmd_hdr { t#__le32 opts ; t#__le32 status ; t#__le32 tbl_addr ; t#__le32 tbl_addr_hi ; t#__le32 reserved [ 4 ] ; }
+s#ahci_em_priv struct ahci_em_priv { e#sw_activity blink_policy ; s#timer_list timer ; unsigned long saved_activity ; unsigned long activity ; unsigned long led_state ; }
+E#EM_MAX_SLOTS 8
+s#ahci_port_priv struct ahci_port_priv { s#ata_link * active_link ; s#ahci_cmd_hdr * cmd_slot ; t#dma_addr_t cmd_slot_dma ; void * cmd_tbl ; t#dma_addr_t cmd_tbl_dma ; void * rx_fis ; t#dma_addr_t rx_fis_dma ; unsigned int ncq_saw_d2h : 1 ; unsigned int ncq_saw_dmas : 1 ; unsigned int ncq_saw_sdb : 1 ; t#spinlock_t lock ; t#u32 intr_mask ; t#bool fbs_supported ; t#bool fbs_enabled ; int fbs_last_dev ; s#ahci_em_priv em_priv [ E#EM_MAX_SLOTS ] ; char * irq_desc ; }
+ahci_fill_cmd_slot void ahci_fill_cmd_slot ( s#ahci_port_priv * , unsigned int , t#u32 )
+ahci_kick_engine int ahci_kick_engine ( s#ata_port * )
+ahci_check_ready int ahci_check_ready ( s#ata_link * )
+ahci_do_softreset int ahci_do_softreset ( s#ata_link * , unsigned int * , int , unsigned long , int ( * ) ( s#ata_link * ) )
+ahci_do_hardreset int ahci_do_hardreset ( s#ata_link * , unsigned int * , unsigned long , t#bool * )
+ahci_handle_port_intr t#u32 ahci_handle_port_intr ( s#ata_host * , t#u32 )
+ahci_qc_issue unsigned int ahci_qc_issue ( s#ata_queued_cmd * )
+ahci_error_handler void ahci_error_handler ( s#ata_port * )
+ahci_port_resume int ahci_port_resume ( s#ata_port * )
+ahci_print_info void ahci_print_info ( s#ata_host * , const char * )
+s#ata_port_info struct ata_port_info { unsigned long flags ; unsigned long link_flags ; unsigned long pio_mask ; unsigned long mwdma_mask ; unsigned long udma_mask ; s#ata_port_operations * port_ops ; void * private_data ; }
+ahci_set_em_messages void ahci_set_em_messages ( s#ahci_host_priv * , s#ata_port_info * )
+ahci_host_activate int ahci_host_activate ( s#ata_host * , s#scsi_host_template * )
+
+/* drivers/ata/libahci_platform.o */
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##posix_acl
+s##vfsmount
+s##sock
+s##iov_iter
+s##pid_namespace
+s##nlm_lockowner
+s##nfs4_lock_state
+s##pipe_inode_info
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##device_private
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##iommu_group
+s##iommu_fwspec
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##kstatfs
+s##eventfd_ctx
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_geometry
+s##pr_ops
+s##blk_mq_ctx
+s##kmem_cache
+s##blkcg_gq
+s##blk_flush_queue
+s##blk_mq_tags
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_stat_callback
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##mtd_info
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##scsi_device
+s##scsi_cmnd
+s##scsi_target
+s##proc_dir_entry
+s##scsi_host_cmd_pool
+s##scsi_transport_template
+ahci_platform_ops extern s#ata_port_operations ahci_platform_ops
+s##clk
+s##regulator
+ahci_platform_enable_clks int ahci_platform_enable_clks ( s#ahci_host_priv * )
+ahci_platform_disable_clks void ahci_platform_disable_clks ( s#ahci_host_priv * )
+ahci_platform_enable_regulators int ahci_platform_enable_regulators ( s#ahci_host_priv * )
+ahci_platform_disable_regulators void ahci_platform_disable_regulators ( s#ahci_host_priv * )
+ahci_platform_enable_resources int ahci_platform_enable_resources ( s#ahci_host_priv * )
+ahci_platform_disable_resources void ahci_platform_disable_resources ( s#ahci_host_priv * )
+s##mfd_cell
+ahci_platform_get_resources s#ahci_host_priv * ahci_platform_get_resources ( s#platform_device * )
+ahci_platform_init_host int ahci_platform_init_host ( s#platform_device * , s#ahci_host_priv * , const s#ata_port_info * , s#scsi_host_template * )
+ahci_platform_suspend_host int ahci_platform_suspend_host ( s#device * )
+ahci_platform_resume_host int ahci_platform_resume_host ( s#device * )
+ahci_platform_suspend int ahci_platform_suspend ( s#device * )
+ahci_platform_resume int ahci_platform_resume ( s#device * )
+
+/* drivers/ata/libata-acpi.o */
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##posix_acl
+s##vfsmount
+s##sock
+s##iov_iter
+s##pid_namespace
+s##nlm_lockowner
+s##nfs4_lock_state
+s##pipe_inode_info
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##device_private
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##kstatfs
+s##eventfd_ctx
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_geometry
+s##pr_ops
+s##blk_mq_ctx
+s##kmem_cache
+s##blkcg_gq
+s##blk_flush_queue
+s##blk_mq_tags
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_stat_callback
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##mtd_info
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##scsi_device_handler
+s##scsi_cmnd
+s##proc_dir_entry
+s##scsi_host_cmd_pool
+s##scsi_transport_template
+ata_acpi_gtm int ata_acpi_gtm ( s#ata_port * , s#ata_acpi_gtm * )
+ata_acpi_stm int ata_acpi_stm ( s#ata_port * , const s#ata_acpi_gtm * )
+ata_acpi_gtm_xfermask unsigned long ata_acpi_gtm_xfermask ( s#ata_device * , const s#ata_acpi_gtm * )
+ata_acpi_cbl_80wire int ata_acpi_cbl_80wire ( s#ata_port * , const s#ata_acpi_gtm * )
+
+/* drivers/ata/libata-core.o */
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##posix_acl
+s##vfsmount
+s##sock
+s##pipe_inode_info
+s##kmem_cache
+s##fs_pin
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##nlm_lockowner
+s##nfs4_lock_state
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_array
+s##tracer
+s##trace_buffer
+s##ring_buffer_iter
+s##event_filter
+s##bpf_prog_array
+s##device_private
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##cgroup_taskset
+s##ring_buffer
+s##bpf_prog
+s##ftrace_hash
+s##klp_modinfo
+s##workqueue_struct
+s##kstatfs
+s##eventfd_ctx
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_geometry
+s##pr_ops
+s##blk_mq_ctx
+s##blkcg_gq
+s##blk_flush_queue
+s##blk_mq_tags
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_stat_callback
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##mtd_info
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##tty_struct
+s##tty_audit_buf
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##futex_pi_state
+s##numa_group
+s##task_delay_info
+s##uprobe
+s##scsi_device_handler
+s##proc_dir_entry
+s##scsi_host_cmd_pool
+s##scsi_transport_template
+ata_dev_set_feature extern unsigned int ata_dev_set_feature ( s#ata_device * , t#u8 , t#u8 )
+ata_sas_port_suspend extern void ata_sas_port_suspend ( s#ata_port * )
+ata_sas_port_resume extern void ata_sas_port_resume ( s#ata_port * )
+sata_lpm_ignore_phy_events extern t#bool sata_lpm_ignore_phy_events ( s#ata_link * )
+ata_port_printk __attribute__ ( ( format ( printf , 3 , 4 ) ) ) void ata_port_printk ( const s#ata_port * , const char * , const char * , ... )
+ata_link_printk __attribute__ ( ( format ( printf , 3 , 4 ) ) ) void ata_link_printk ( const s#ata_link * , const char * , const char * , ... )
+ata_dev_printk __attribute__ ( ( format ( printf , 3 , 4 ) ) ) void ata_dev_printk ( const s#ata_device * , const char * , const char * , ... )
+ata_print_version void ata_print_version ( const s#device * , const char * )
+sata_deb_timing_normal extern const unsigned long sata_deb_timing_normal [ ]
+sata_deb_timing_hotplug extern const unsigned long sata_deb_timing_hotplug [ ]
+sata_deb_timing_long extern const unsigned long sata_deb_timing_long [ ]
+ata_base_port_ops extern const s#ata_port_operations ata_base_port_ops
+sata_port_ops extern const s#ata_port_operations sata_port_ops
+ata_dummy_port_ops extern s#ata_port_operations ata_dummy_port_ops
+ata_dummy_port_info extern const s#ata_port_info ata_dummy_port_info
+e#ata_link_iter_mode enum ata_link_iter_mode { ATA_LITER_EDGE , ATA_LITER_HOST_FIRST , ATA_LITER_PMP_FIRST , }
+ata_link_next extern s#ata_link * ata_link_next ( s#ata_link * , s#ata_port * , e#ata_link_iter_mode )
+e#ata_dev_iter_mode enum ata_dev_iter_mode { ATA_DITER_ENABLED , ATA_DITER_ENABLED_REVERSE , ATA_DITER_ALL , ATA_DITER_ALL_REVERSE , }
+ata_dev_next extern s#ata_device * ata_dev_next ( s#ata_device * , s#ata_link * , e#ata_dev_iter_mode )
+ata_std_bios_param extern int ata_std_bios_param ( s#scsi_device * , s#block_device * , t#sector_t , int [ ] )
+ata_scsi_unlock_native_capacity extern void ata_scsi_unlock_native_capacity ( s#scsi_device * )
+ata_host_init extern void ata_host_init ( s#ata_host * , s#device * , s#ata_port_operations * )
+ata_host_alloc extern s#ata_host * ata_host_alloc ( s#device * , int )
+ata_host_alloc_pinfo extern s#ata_host * ata_host_alloc_pinfo ( s#device * , const s#ata_port_info * const * , int )
+ata_slave_link_init extern int ata_slave_link_init ( s#ata_port * )
+ata_host_start extern int ata_host_start ( s#ata_host * )
+ata_host_register extern int ata_host_register ( s#ata_host * , s#scsi_host_template * )
+ata_host_activate extern int ata_host_activate ( s#ata_host * , int , t#irq_handler_t , unsigned long , s#scsi_host_template * )
+ata_host_detach extern void ata_host_detach ( s#ata_host * )
+ata_sg_init extern void ata_sg_init ( s#ata_queued_cmd * , s#scatterlist * , unsigned int )
+ata_qc_complete extern void ata_qc_complete ( s#ata_queued_cmd * )
+ata_qc_complete_multiple extern int ata_qc_complete_multiple ( s#ata_port * , t#u32 )
+atapi_cmd_type extern int atapi_cmd_type ( t#u8 )
+ata_tf_to_fis extern void ata_tf_to_fis ( const s#ata_taskfile * , t#u8 , int , t#u8 * )
+ata_tf_from_fis extern void ata_tf_from_fis ( const t#u8 * , s#ata_taskfile * )
+ata_pack_xfermask extern unsigned long ata_pack_xfermask ( unsigned long , unsigned long , unsigned long )
+ata_unpack_xfermask extern void ata_unpack_xfermask ( unsigned long , unsigned long * , unsigned long * , unsigned long * )
+ata_xfer_mask2mode extern t#u8 ata_xfer_mask2mode ( unsigned long )
+ata_xfer_mode2mask extern unsigned long ata_xfer_mode2mask ( t#u8 )
+ata_xfer_mode2shift extern int ata_xfer_mode2shift ( unsigned long )
+ata_mode_string extern const char * ata_mode_string ( unsigned long )
+ata_id_xfermask extern unsigned long ata_id_xfermask ( const t#u16 * )
+ata_do_set_mode extern int ata_do_set_mode ( s#ata_link * , s#ata_device * * )
+ata_std_qc_defer extern int ata_std_qc_defer ( s#ata_queued_cmd * )
+ata_noop_qc_prep extern void ata_noop_qc_prep ( s#ata_queued_cmd * )
+ata_dev_disable extern void ata_dev_disable ( s#ata_device * )
+sata_set_spd extern int sata_set_spd ( s#ata_link * )
+ata_wait_after_reset extern int ata_wait_after_reset ( s#ata_link * , unsigned long , int ( * ) ( s#ata_link * ) )
+sata_link_debounce extern int sata_link_debounce ( s#ata_link * , const unsigned long * , unsigned long )
+sata_link_resume extern int sata_link_resume ( s#ata_link * , const unsigned long * , unsigned long )
+sata_link_scr_lpm extern int sata_link_scr_lpm ( s#ata_link * , e#ata_lpm_policy , t#bool )
+ata_std_prereset extern int ata_std_prereset ( s#ata_link * , unsigned long )
+sata_link_hardreset extern int sata_link_hardreset ( s#ata_link * , const unsigned long * , unsigned long , t#bool * , int ( * ) ( s#ata_link * ) )
+sata_std_hardreset extern int sata_std_hardreset ( s#ata_link * , unsigned int * , unsigned long )
+ata_std_postreset extern void ata_std_postreset ( s#ata_link * , unsigned int * )
+ata_dev_classify extern unsigned int ata_dev_classify ( const s#ata_taskfile * )
+ata_dev_pair extern s#ata_device * ata_dev_pair ( s#ata_device * )
+ata_ratelimit extern int ata_ratelimit ( void )
+ata_msleep extern void ata_msleep ( s#ata_port * , unsigned int )
+ata_wait_register extern t#u32 ata_wait_register ( s#ata_port * , void * , t#u32 , t#u32 , unsigned long , unsigned long )
+ata_scsi_queuecmd extern int ata_scsi_queuecmd ( s#Scsi_Host * , s#scsi_cmnd * )
+ata_scsi_slave_config extern int ata_scsi_slave_config ( s#scsi_device * )
+ata_scsi_slave_destroy extern void ata_scsi_slave_destroy ( s#scsi_device * )
+ata_scsi_change_queue_depth extern int ata_scsi_change_queue_depth ( s#scsi_device * , int )
+__ata_change_queue_depth extern int __ata_change_queue_depth ( s#ata_port * , s#scsi_device * , int )
+sata_scr_valid extern int sata_scr_valid ( s#ata_link * )
+sata_scr_read extern int sata_scr_read ( s#ata_link * , int , t#u32 * )
+sata_scr_write extern int sata_scr_write ( s#ata_link * , int , t#u32 )
+sata_scr_write_flush extern int sata_scr_write_flush ( s#ata_link * , int , t#u32 )
+ata_link_online extern t#bool ata_link_online ( s#ata_link * )
+ata_link_offline extern t#bool ata_link_offline ( s#ata_link * )
+ata_host_suspend extern int ata_host_suspend ( s#ata_host * , t#pm_message_t )
+ata_host_resume extern void ata_host_resume ( s#ata_host * )
+ata_id_string extern void ata_id_string ( const t#u16 * , unsigned char * , unsigned int , unsigned int )
+ata_id_c_string extern void ata_id_c_string ( const t#u16 * , unsigned char * , unsigned int , unsigned int )
+ata_do_dev_read_id extern unsigned int ata_do_dev_read_id ( s#ata_device * , s#ata_taskfile * , t#u16 * )
+ata_scsi_simulate extern void ata_scsi_simulate ( s#ata_device * , s#scsi_cmnd * )
+ata_pio_need_iordy extern unsigned int ata_pio_need_iordy ( const s#ata_device * )
+s#ata_timing struct ata_timing { unsigned short mode ; unsigned short setup ; unsigned short act8b ; unsigned short rec8b ; unsigned short cyc8b ; unsigned short active ; unsigned short recover ; unsigned short dmack_hold ; unsigned short cycle ; unsigned short udma ; }
+ata_timing_find_mode extern const s#ata_timing * ata_timing_find_mode ( t#u8 )
+ata_timing_compute extern int ata_timing_compute ( s#ata_device * , unsigned short , s#ata_timing * , int , int )
+ata_timing_merge extern void ata_timing_merge ( const s#ata_timing * , const s#ata_timing * , s#ata_timing * , unsigned int )
+ata_timing_cycle2mode extern t#u8 ata_timing_cycle2mode ( unsigned int , int )
+s##msi_controller
+s##hotplug_slot
+s##pcie_link_state
+s##pci_vpd
+s##pci_sriov
+s#pci_bits struct pci_bits { unsigned int reg ; unsigned int width ; unsigned long mask ; unsigned long val ; }
+pci_test_config_bits extern int pci_test_config_bits ( s#pci_dev * , const s#pci_bits * )
+ata_pci_remove_one extern void ata_pci_remove_one ( s#pci_dev * )
+ata_pci_device_do_suspend extern void ata_pci_device_do_suspend ( s#pci_dev * , t#pm_message_t )
+ata_pci_device_do_resume extern int __attribute__ ( ( warn_unused_result ) ) ata_pci_device_do_resume ( s#pci_dev * )
+ata_pci_device_suspend extern int ata_pci_device_suspend ( s#pci_dev * , t#pm_message_t )
+ata_pci_device_resume extern int ata_pci_device_resume ( s#pci_dev * )
+s##mfd_cell
+ata_platform_remove_one extern int ata_platform_remove_one ( s#platform_device * )
+__ata_ehi_push_desc extern __attribute__ ( ( format ( printf , 2 , 3 ) ) ) void __ata_ehi_push_desc ( s#ata_eh_info * , const char * , ... )
+ata_ehi_push_desc extern __attribute__ ( ( format ( printf , 2 , 3 ) ) ) void ata_ehi_push_desc ( s#ata_eh_info * , const char * , ... )
+ata_ehi_clear_desc extern void ata_ehi_clear_desc ( s#ata_eh_info * )
+ata_port_desc extern __attribute__ ( ( format ( printf , 2 , 3 ) ) ) void ata_port_desc ( s#ata_port * , const char * , ... )
+ata_port_pbar_desc extern void ata_port_pbar_desc ( s#ata_port * , int , t#ssize_t , const char * )
+ata_port_schedule_eh extern void ata_port_schedule_eh ( s#ata_port * )
+ata_link_abort extern int ata_link_abort ( s#ata_link * )
+ata_port_abort extern int ata_port_abort ( s#ata_port * )
+ata_port_freeze extern int ata_port_freeze ( s#ata_port * )
+sata_async_notification extern int sata_async_notification ( s#ata_port * )
+ata_eh_freeze_port extern void ata_eh_freeze_port ( s#ata_port * )
+ata_eh_thaw_port extern void ata_eh_thaw_port ( s#ata_port * )
+ata_eh_qc_complete extern void ata_eh_qc_complete ( s#ata_queued_cmd * )
+ata_eh_qc_retry extern void ata_eh_qc_retry ( s#ata_queued_cmd * )
+ata_eh_analyze_ncq_error extern void ata_eh_analyze_ncq_error ( s#ata_link * )
+ata_do_eh extern void ata_do_eh ( s#ata_port * , t#ata_prereset_fn_t , t#ata_reset_fn_t , t#ata_reset_fn_t , t#ata_postreset_fn_t )
+ata_std_error_handler extern void ata_std_error_handler ( s#ata_port * )
+ata_cable_40wire extern int ata_cable_40wire ( s#ata_port * )
+ata_cable_80wire extern int ata_cable_80wire ( s#ata_port * )
+ata_cable_unknown extern int ata_cable_unknown ( s#ata_port * )
+ata_cable_ignore extern int ata_cable_ignore ( s#ata_port * )
+ata_cable_sata extern int ata_cable_sata ( s#ata_port * )
+ata_host_get extern void ata_host_get ( s#ata_host * )
+ata_host_put extern void ata_host_put ( s#ata_host * )
+
+/* drivers/ata/libata-eh.o */
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##posix_acl
+s##vfsmount
+s##sock
+s##iov_iter
+s##pid_namespace
+s##nlm_lockowner
+s##nfs4_lock_state
+s##pipe_inode_info
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##klp_modinfo
+s##device_private
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##kstatfs
+s##eventfd_ctx
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_geometry
+s##pr_ops
+s##blk_mq_ctx
+s##kmem_cache
+s##blkcg_gq
+s##blk_flush_queue
+s##blk_mq_tags
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_stat_callback
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##mtd_info
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##scsi_device_handler
+s##proc_dir_entry
+s##scsi_host_cmd_pool
+s##scsi_transport_template
+ata_scsi_cmd_error_handler extern void ata_scsi_cmd_error_handler ( s#Scsi_Host * , s#ata_port * , s#list_head * )
+ata_scsi_port_error_handler extern void ata_scsi_port_error_handler ( s#Scsi_Host * , s#ata_port * )
+ata_port_wait_eh extern void ata_port_wait_eh ( s#ata_port * )
+ata_std_sched_eh extern void ata_std_sched_eh ( s#ata_port * )
+ata_std_end_eh extern void ata_std_end_eh ( s#ata_port * )
+ata_get_cmd_descript const char * ata_get_cmd_descript ( t#u8 )
+
+/* drivers/ata/libata-pmp.o */
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##posix_acl
+s##vfsmount
+s##sock
+s##iov_iter
+s##pid_namespace
+s##nlm_lockowner
+s##nfs4_lock_state
+s##pipe_inode_info
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##device_private
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##kstatfs
+s##eventfd_ctx
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_geometry
+s##pr_ops
+s##blk_mq_ctx
+s##kmem_cache
+s##blkcg_gq
+s##blk_flush_queue
+s##blk_mq_tags
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_stat_callback
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##mtd_info
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##scsi_device
+s##scsi_cmnd
+s##scsi_target
+s##proc_dir_entry
+s##scsi_host_cmd_pool
+s##scsi_transport_template
+sata_pmp_port_ops extern const s#ata_port_operations sata_pmp_port_ops
+sata_pmp_qc_defer_cmd_switch extern int sata_pmp_qc_defer_cmd_switch ( s#ata_queued_cmd * )
+sata_pmp_error_handler extern void sata_pmp_error_handler ( s#ata_port * )
+
+/* drivers/ata/libata-scsi.o */
+s##device_private
+s##sock
+s##vfsmount
+s##posix_acl
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##iov_iter
+s##pr_ops
+s##blk_mq_ctx
+s##kmem_cache
+s##blkcg_gq
+s##blk_flush_queue
+s##blk_mq_tags
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_stat_callback
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##eventfd_ctx
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##proc_ns_operations
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##mmu_notifier_mm
+s##xol_area
+s##pid_namespace
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##pipe_inode_info
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##mtd_info
+s##nlm_lockowner
+s##nfs4_lock_state
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##kernfs_open_node
+s##kernfs_iattrs
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+dev_attr_link_power_management_policy extern s#device_attribute dev_attr_link_power_management_policy
+dev_attr_unload_heads extern s#device_attribute dev_attr_unload_heads
+dev_attr_ncq_prio_enable extern s#device_attribute dev_attr_ncq_prio_enable
+dev_attr_em_message extern s#device_attribute dev_attr_em_message
+dev_attr_em_message_type extern s#device_attribute dev_attr_em_message_type
+dev_attr_sw_activity extern s#device_attribute dev_attr_sw_activity
+ata_common_sdev_attrs extern s#device_attribute * ata_common_sdev_attrs [ ]
+s##scsi_device_handler
+s##proc_dir_entry
+s##scsi_host_cmd_pool
+s#klist struct klist { t#spinlock_t k_lock ; s#list_head k_list ; void ( * get ) ( s#klist_node * ) ; void ( * put ) ( s#klist_node * ) ; }
+s#attribute_container struct attribute_container { s#list_head node ; s#klist containers ; s#class * class ; const s#attribute_group * grp ; s#device_attribute * * attrs ; int ( * match ) ( s#attribute_container * , s#device * ) ; unsigned long flags ; }
+s#transport_container struct transport_container { s#attribute_container ac ; const s#attribute_group * statistics ; }
+t#uint typedef unsigned int uint
+s#scsi_transport_template struct scsi_transport_template { s#transport_container host_attrs ; s#transport_container target_attrs ; s#transport_container device_attrs ; int ( * user_scan ) ( s#Scsi_Host * , t#uint , t#uint , t#u64 ) ; int device_size ; int device_private_offset ; int target_size ; int target_private_offset ; int host_size ; unsigned int create_work_queue : 1 ; void ( * eh_strategy_handler ) ( s#Scsi_Host * ) ; }
+ata_sas_scsi_ioctl extern int ata_sas_scsi_ioctl ( s#ata_port * , s#scsi_device * , int , void * )
+ata_scsi_ioctl extern int ata_scsi_ioctl ( s#scsi_device * , int , void * )
+ata_sas_port_alloc extern s#ata_port * ata_sas_port_alloc ( s#ata_host * , s#ata_port_info * , s#Scsi_Host * )
+ata_sas_port_start extern int ata_sas_port_start ( s#ata_port * )
+ata_sas_port_stop extern void ata_sas_port_stop ( s#ata_port * )
+ata_sas_async_probe extern void ata_sas_async_probe ( s#ata_port * )
+ata_sas_sync_probe extern int ata_sas_sync_probe ( s#ata_port * )
+ata_sas_port_init extern int ata_sas_port_init ( s#ata_port * )
+ata_sas_tport_add extern int ata_sas_tport_add ( s#device * , s#ata_port * )
+ata_sas_tport_delete extern void ata_sas_tport_delete ( s#ata_port * )
+ata_sas_port_destroy extern void ata_sas_port_destroy ( s#ata_port * )
+ata_sas_slave_configure extern int ata_sas_slave_configure ( s#scsi_device * , s#ata_port * )
+ata_sas_queuecmd extern int ata_sas_queuecmd ( s#scsi_cmnd * , s#ata_port * )
+
+/* drivers/ata/libata-sff.o */
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##posix_acl
+s##vfsmount
+s##sock
+s##iov_iter
+s##pid_namespace
+s##nlm_lockowner
+s##nfs4_lock_state
+s##pipe_inode_info
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##device_private
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##kstatfs
+s##eventfd_ctx
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_geometry
+s##pr_ops
+s##blk_mq_ctx
+s##kmem_cache
+s##blkcg_gq
+s##blk_flush_queue
+s##blk_mq_tags
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_stat_callback
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##mtd_info
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##scsi_device
+s##scsi_cmnd
+s##scsi_target
+s##proc_dir_entry
+s##scsi_host_cmd_pool
+s##scsi_transport_template
+ata_sff_port_ops extern const s#ata_port_operations ata_sff_port_ops
+ata_sff_check_status extern t#u8 ata_sff_check_status ( s#ata_port * )
+ata_sff_pause extern void ata_sff_pause ( s#ata_port * )
+ata_sff_dma_pause extern void ata_sff_dma_pause ( s#ata_port * )
+ata_sff_busy_sleep extern int ata_sff_busy_sleep ( s#ata_port * , unsigned long , unsigned long )
+ata_sff_wait_ready extern int ata_sff_wait_ready ( s#ata_link * , unsigned long )
+ata_sff_dev_select extern void ata_sff_dev_select ( s#ata_port * , unsigned int )
+ata_sff_irq_on extern void ata_sff_irq_on ( s#ata_port * )
+ata_sff_tf_load extern void ata_sff_tf_load ( s#ata_port * , const s#ata_taskfile * )
+ata_sff_tf_read extern void ata_sff_tf_read ( s#ata_port * , s#ata_taskfile * )
+ata_sff_exec_command extern void ata_sff_exec_command ( s#ata_port * , const s#ata_taskfile * )
+ata_sff_data_xfer extern unsigned int ata_sff_data_xfer ( s#ata_queued_cmd * , unsigned char * , unsigned int , int )
+ata_sff_data_xfer32 extern unsigned int ata_sff_data_xfer32 ( s#ata_queued_cmd * , unsigned char * , unsigned int , int )
+ata_sff_data_xfer_noirq extern unsigned int ata_sff_data_xfer_noirq ( s#ata_queued_cmd * , unsigned char * , unsigned int , int )
+ata_sff_hsm_move extern int ata_sff_hsm_move ( s#ata_port * , s#ata_queued_cmd * , t#u8 , int )
+ata_sff_queue_work extern void ata_sff_queue_work ( s#work_struct * )
+ata_sff_queue_delayed_work extern void ata_sff_queue_delayed_work ( s#delayed_work * , unsigned long )
+ata_sff_queue_pio_task extern void ata_sff_queue_pio_task ( s#ata_link * , unsigned long )
+ata_sff_qc_issue extern unsigned int ata_sff_qc_issue ( s#ata_queued_cmd * )
+ata_sff_qc_fill_rtf extern t#bool ata_sff_qc_fill_rtf ( s#ata_queued_cmd * )
+ata_sff_port_intr extern unsigned int ata_sff_port_intr ( s#ata_port * , s#ata_queued_cmd * )
+ata_sff_interrupt extern t#irqreturn_t ata_sff_interrupt ( int , void * )
+ata_sff_lost_interrupt extern void ata_sff_lost_interrupt ( s#ata_port * )
+ata_sff_freeze extern void ata_sff_freeze ( s#ata_port * )
+ata_sff_thaw extern void ata_sff_thaw ( s#ata_port * )
+ata_sff_prereset extern int ata_sff_prereset ( s#ata_link * , unsigned long )
+ata_sff_dev_classify extern unsigned int ata_sff_dev_classify ( s#ata_device * , int , t#u8 * )
+ata_sff_wait_after_reset extern int ata_sff_wait_after_reset ( s#ata_link * , unsigned int , unsigned long )
+ata_sff_softreset extern int ata_sff_softreset ( s#ata_link * , unsigned int * , unsigned long )
+sata_sff_hardreset extern int sata_sff_hardreset ( s#ata_link * , unsigned int * , unsigned long )
+ata_sff_postreset extern void ata_sff_postreset ( s#ata_link * , unsigned int * )
+ata_sff_drain_fifo extern void ata_sff_drain_fifo ( s#ata_queued_cmd * )
+ata_sff_error_handler extern void ata_sff_error_handler ( s#ata_port * )
+ata_sff_std_ports extern void ata_sff_std_ports ( s#ata_ioports * )
+ata_pci_sff_init_host extern int ata_pci_sff_init_host ( s#ata_host * )
+s##msi_controller
+s##hotplug_slot
+s##pcie_link_state
+s##pci_vpd
+s##pci_sriov
+ata_pci_sff_prepare_host extern int ata_pci_sff_prepare_host ( s#pci_dev * , const s#ata_port_info * const * , s#ata_host * * )
+ata_pci_sff_activate_host extern int ata_pci_sff_activate_host ( s#ata_host * , t#irq_handler_t , s#scsi_host_template * )
+ata_pci_sff_init_one extern int ata_pci_sff_init_one ( s#pci_dev * , const s#ata_port_info * const * , s#scsi_host_template * , void * , int )
+ata_bmdma_port_ops extern const s#ata_port_operations ata_bmdma_port_ops
+ata_bmdma32_port_ops extern const s#ata_port_operations ata_bmdma32_port_ops
+ata_bmdma_qc_prep extern void ata_bmdma_qc_prep ( s#ata_queued_cmd * )
+ata_bmdma_dumb_qc_prep extern void ata_bmdma_dumb_qc_prep ( s#ata_queued_cmd * )
+ata_bmdma_qc_issue extern unsigned int ata_bmdma_qc_issue ( s#ata_queued_cmd * )
+ata_bmdma_port_intr extern unsigned int ata_bmdma_port_intr ( s#ata_port * , s#ata_queued_cmd * )
+ata_bmdma_interrupt extern t#irqreturn_t ata_bmdma_interrupt ( int , void * )
+ata_bmdma_error_handler extern void ata_bmdma_error_handler ( s#ata_port * )
+ata_bmdma_post_internal_cmd extern void ata_bmdma_post_internal_cmd ( s#ata_queued_cmd * )
+ata_bmdma_irq_clear extern void ata_bmdma_irq_clear ( s#ata_port * )
+ata_bmdma_setup extern void ata_bmdma_setup ( s#ata_queued_cmd * )
+ata_bmdma_start extern void ata_bmdma_start ( s#ata_queued_cmd * )
+ata_bmdma_stop extern void ata_bmdma_stop ( s#ata_queued_cmd * )
+ata_bmdma_status extern t#u8 ata_bmdma_status ( s#ata_port * )
+ata_bmdma_port_start extern int ata_bmdma_port_start ( s#ata_port * )
+ata_bmdma_port_start32 extern int ata_bmdma_port_start32 ( s#ata_port * )
+ata_pci_bmdma_clear_simplex extern int ata_pci_bmdma_clear_simplex ( s#pci_dev * )
+ata_pci_bmdma_init extern void ata_pci_bmdma_init ( s#ata_host * )
+ata_pci_bmdma_prepare_host extern int ata_pci_bmdma_prepare_host ( s#pci_dev * , const s#ata_port_info * const * , s#ata_host * * )
+ata_pci_bmdma_init_one extern int ata_pci_bmdma_init_one ( s#pci_dev * , const s#ata_port_info * const * , s#scsi_host_template * , void * , int )
+
+/* drivers/ata/pata_sis.o */
+s##sched_class
+s##cfs_rq
+s##task_group
+s##anon_vma
+s##posix_acl
+s##vfsmount
+s##sock
+s##iov_iter
+s##pid_namespace
+s##nlm_lockowner
+s##nfs4_lock_state
+s##pipe_inode_info
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##device_private
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##workqueue_struct
+s##bpf_prog_array
+s##cgroup_taskset
+s##kstatfs
+s##eventfd_ctx
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_geometry
+s##pr_ops
+s##blk_mq_ctx
+s##kmem_cache
+s##blkcg_gq
+s##blk_flush_queue
+s##blk_mq_tags
+s##blk_mq_alloc_data
+s##blk_mq_debugfs_attr
+s##blk_queue_stats
+s##rq_wb
+s##blk_stat_callback
+s##blk_trace
+s##bsg_job
+s##throtl_data
+s##timer_rand_state
+s##disk_events
+s##badblocks
+s##mtd_info
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##mempolicy
+s##userfaultfd_ctx
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##net
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##robust_list_head
+s##compat_robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##scsi_device
+s##scsi_cmnd
+s##scsi_target
+s##proc_dir_entry
+s##scsi_host_cmd_pool
+s##scsi_transport_template
+sis_info133_for_sata extern const s#ata_port_info sis_info133_for_sata
+
+/* drivers/atm/idt77105.o */
+s##posix_acl
+s##vfsmount
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##proc_dir_entry
+s##uevent_sock
+s##iw_handler_def
+s##iw_public_data
+s##sec_path
+s##workqueue_struct
+s##netpoll_info
+s##udp_tunnel_info
+s##bpf_prog_aux
+s##bpf_prog_offload_ops
+s##bpf_offloaded_map
+s##xdp_umem
+s##switchdev_ops
+s##l3mdev_ops
+s##ndisc_ops
+s##xfrm_state
+s##tls_crypto_info
+s##tls_context
+s##vlan_info
+s##dsa_switch_tree
+s##in_device
+s##dn_dev
+s##inet6_dev
+s##wireless_dev
+s##wpan_dev
+s##mpls_dev
+s##tcf_walker
+s##qdisc_walker
+s##net_rate_estimator
+s##nf_hook_entry
+s##cpu_rmap
+s##pcpu_dstats
+s##garp_port
+s##mrp_port
+s##device_private
+s##of_device_id
+s##acpi_device_id
+s##driver_private
+s##iommu_ops
+s##subsys_private
+s##wake_irq
+s##pm_domain_data
+s##dev_pm_qos
+s##irq_domain
+s##pinctrl
+s##pinctrl_state
+s##dma_coherent_mem
+s##cma
+s##device_node
+s##iommu_group
+s##iommu_fwspec
+s##phy_device
+s##sfp_bus
+s##prot_inuse
+s##ipv4_devconf
+s##fib_rules_ops
+s##fib_table
+s##inet_peer_base
+s##kmem_cache
+s##xt_table
+s##inet_hashinfo
+s##fib_notifier_ops
+s##ipv6_devconf
+s##rt6_info
+s##rt6_statistics
+s##fib6_table
+s##lwtunnel_state
+s##rtable
+s##dn_route
+s##seg6_pernet_data
+s##sctp_mib
+s##nf_queue_handler
+s##nf_logger
+s##ebt_table
+s##nf_ct_event_notifier
+s##nf_exp_event_notifier
+s##nft_af_info
+s##net_generic
+s##netns_ipvs
+s##mpls_route
+s##dev_rcv_lists
+s##s_stats
+s##s_pstats
+s##bpf_prog_array
+s##cgroup_taskset
+s##eventfd_ctx
+s##mempolicy
+s##signal_struct
+s##sighand_struct
+s##audit_context
+s##seccomp_filter
+s##rt_mutex_waiter
+s##gendisk
+s##request_queue
+s##blk_plug
+s##reclaim_state
+s##robust_list_head
+s##futex_pi_state
+s##perf_event_context
+s##numa_group
+s##task_delay_info
+s##ftrace_ret_stack
+s##uprobe
+s##perf_event
+s##linux_binfmt
+s##kioctx_table
+s##mmu_notifier_mm
+s##xol_area
+s##anon_vma
+s##userfaultfd_ctx
+s##nlm_lockowner
+s##nfs4_lock_state
+s##kernfs_open_node
+s##kernfs_iattrs
+s##module_param_attrs
+s##module_sect_attrs
+s##module_notes_attrs
+s##trace_event_call
+s##trace_enum_map
+s##klp_modinfo
+s##kstatfs
+s##export_operations
+s##xattr_handler
+s##fscrypt_operations
+s##hd_struct
+s##mtd_info
+s##cdev
+s##fsnotify_mark_connector
+s##fscrypt_info
+s##swap_info_struct
+s##request_sock_ops
+s##timewait_sock_ops
+s##udp_table
+s##raw_hashinfo
+s##xfrm_policy
+s##sock_reuseport
+s#atm_trafprm struct atm_trafprm { unsigned char traffic_class ; int max_pcr ; int pcr ; int min_pcr ; int max_cdv ; int max_sdu ; unsigned int icr ; unsigned int tbe ; unsigned int frtt : 24 ; unsigned int rif : 4 ; unsigned int rdf : 4 ; unsigned int nrm_pres : 1 ; unsigned int trm_pres : 1 ; unsigned int adtf_pres : 1 ; unsigned int cdf_pres : 1 ; unsigned int nrm : 3 ; unsigned int trm : 3 ; unsigned int adtf : 10 ; unsigned int cdf : 3 ; unsigned int spare : 9 ; }
+s#atm_qos struct atm_qos { s#atm_trafprm txtp ; s#atm_trafprm rxtp ; unsigned char aal ; }
+s#atm_bhli struct atm_bhli { unsigned char hl_type ; unsigned char hl_length ; unsigned char hl_info [ 8 ] ; }
+s#atm_blli struct atm_blli { unsigned char l2_proto ; union { struct { unsigned char mode ; unsigned char window ; } itu ; unsigned char user ; } l2 ; unsigned char l3_proto ; union { struct { unsigned char mode ; unsigned char def_size ; unsigned char window ; } itu ; unsigned char user ; struct { unsigned char term_type ; unsigned char fw_mpx_cap ; unsigned char bw_mpx_cap ; } h310 ; struct { unsigned char ipi ; unsigned char snap [ 5 ] ; } tr9577 ; } l3 ; }
+s#atm_sap struct atm_sap { s#atm_bhli bhli ; s#atm_blli blli [ 3 ] ; }
+s#k_atm_aal_stats struct k_atm_aal_stats { t#atomic_t tx ; t#atomic_t tx_err ; t#atomic_t rx ; t#atomic_t rx_err ; t#atomic_t rx_drop ; }
+s#sockaddr_atmsvc struct sockaddr_atmsvc { unsigned short sas_family ; struct { unsigned char prv [ 20 ] ; char pub [ 12 + 1 ] ; char lij_type ; t#__u32 lij_id ; } sas_addr ; }
+s#atm_vcc struct atm_vcc { s#sock sk ; unsigned long flags ; short vpi ; int vci ; unsigned long aal_options ; unsigned long atm_options ; s#atm_dev * dev ; s#atm_qos qos ; s#atm_sap sap ; void ( * release_cb ) ( s#atm_vcc * ) ; void ( * push ) ( s#atm_vcc * , s#sk_buff * ) ; void ( * pop ) ( s#atm_vcc * , s#sk_buff * ) ; int ( * push_oam ) ( s#atm_vcc * , void * ) ; int ( * send ) ( s#atm_vcc * , s#sk_buff * ) ; void * dev_data ; void * proto_data ; s#k_atm_aal_stats * stats ; s#module * owner ; short itf ; s#sockaddr_atmsvc local ; s#sockaddr_atmsvc remote ; s#atm_vcc * session ; void * user_back ; }
+s#atmdev_ops struct atmdev_ops { void ( * dev_close ) ( s#atm_dev * ) ; int ( * open ) ( s#atm_vcc * ) ; void ( * close ) ( s#atm_vcc * ) ; int ( * ioctl ) ( s#atm_dev * , unsigned int , void * ) ; int ( * compat_ioctl ) ( s#atm_dev * , unsigned int , void * ) ; int ( * getsockopt ) ( s#atm_vcc * , int , int , void * , int ) ; int ( * setsockopt ) ( s#atm_vcc * , int , int , void * , unsigned int ) ; int ( * send ) ( s#atm_vcc * , s#sk_buff * ) ; int ( * send_oam ) ( s#atm_vcc * , void * , int ) ; void ( * phy_put ) ( s#atm_dev * , unsigned char , unsigned long ) ; unsigned char ( * phy_get ) ( s#atm_dev * , unsigned long ) ; int ( * change_qos ) ( s#atm_vcc * , s#atm_qos * , int ) ; int ( * proc_read ) ( s#atm_dev * , t#loff_t * , char * ) ; s#module * owner ; }
+s#atmphy_ops struct atmphy_ops { int ( * start ) ( s#atm_dev * ) ; int ( * ioctl ) ( s#atm_dev * , unsigned int , void * ) ; void ( * interrupt ) ( s#atm_dev * ) ; int ( * stop ) ( s#atm_dev * ) ; }
+s#atm_cirange struct atm_cirange { signed char vpi_bits ; signed char vci_bits ; }
+s#k_atm_dev_stats struct k_atm_dev_stats { s#k_atm_aal_stats aal0 ; s#k_atm_aal_stats aal34 ; s#k_atm_aal_stats aal5 ; }
+s#atm_dev struct atm_dev { const s#atmdev_ops * ops ; const s#atmphy_ops * phy ; const char * type ; int number ; void * dev_data ; void * phy_data ; unsigned long flags ; s#list_head local ; s#list_head lecs ; unsigned char esi [ 6 ] ; s#atm_cirange ci_range ; s#k_atm_dev_stats stats ; char signal ; int link_rate ; t#atomic_t refcnt ; t#spinlock_t lock ; s#proc_dir_entry * proc_entry ; char * proc_name ; s#device class_dev ; s#list_head dev_list ; }
+idt77105_init int idt77105_init ( s#atm_dev * )
+
+/* drivers/atm/suni.o */
+s##posix_acl
+s##vfsmount
+s##sched_class
+s##cfs_rq
+s##task_group
+s##pid_namespace
+s##key_user
+s##key_type
+s##assoc_array_ptr
+s##proc_ns_operations
+s##nameidata
+s##sem_undo_list
+s##fs_struct
+s##files_struct
+s##uts_namespace
+s##ipc_namespace
+s##mnt_namespace
+s##proc_dir_entry
+s##uevent_sock
+s##iw_handler_def
+s##iw_public_data
+s##sec_path
+s##workqueue_struct
+s##netpoll_info
+s##udp_tunnel_info
+s##bpf_prog_aux
+s##bpf_prog_offload_ops
+s##bpf_offloaded_map
+s##xdp_umem
+s##switchdev_ops
+s##l3mdev_ops
+s##ndisc_ops
+s##xfrm_state
+s##tls_crypto_info
+s##tls_context
+s##vlan_info
+s##dsa_switch_tree
+s##in_device
<