// SPDX-License-Identifier: GPL-2.0-only /* Copyright (c) 2024 Google */ #include #include #include #include #include #include "../../mm/slab.h" /* kmem_cache, slab_caches and slab_mutex */ /* open-coded version */ struct bpf_iter_kmem_cache { __u64 __opaque[1]; } __attribute__((aligned(8))); struct bpf_iter_kmem_cache_kern { struct kmem_cache *pos; } __attribute__((aligned(8))); #define KMEM_CACHE_POS_START ((void *)1L) __bpf_kfunc_start_defs(); __bpf_kfunc int bpf_iter_kmem_cache_new(struct bpf_iter_kmem_cache *it) { struct bpf_iter_kmem_cache_kern *kit = (void *)it; BUILD_BUG_ON(sizeof(*kit) > sizeof(*it)); BUILD_BUG_ON(__alignof__(*kit) != __alignof__(*it)); kit->pos = KMEM_CACHE_POS_START; return 0; } __bpf_kfunc struct kmem_cache *bpf_iter_kmem_cache_next(struct bpf_iter_kmem_cache *it) { struct bpf_iter_kmem_cache_kern *kit = (void *)it; struct kmem_cache *prev = kit->pos; struct kmem_cache *next; bool destroy = false; if (!prev) return NULL; mutex_lock(&slab_mutex); if (list_empty(&slab_caches)) { mutex_unlock(&slab_mutex); return NULL; } if (prev == KMEM_CACHE_POS_START) next = list_first_entry(&slab_caches, struct kmem_cache, list); else if (list_last_entry(&slab_caches, struct kmem_cache, list) == prev) next = NULL; else next = list_next_entry(prev, list); /* boot_caches have negative refcount, don't touch them */ if (next && next->refcount > 0) next->refcount++; /* Skip kmem_cache_destroy() for active entries */ if (prev && prev != KMEM_CACHE_POS_START) { if (prev->refcount > 1) prev->refcount--; else if (prev->refcount == 1) destroy = true; } mutex_unlock(&slab_mutex); if (destroy) kmem_cache_destroy(prev); kit->pos = next; return next; } __bpf_kfunc void bpf_iter_kmem_cache_destroy(struct bpf_iter_kmem_cache *it) { struct bpf_iter_kmem_cache_kern *kit = (void *)it; struct kmem_cache *s = kit->pos; bool destroy = false; if (s == NULL || s == KMEM_CACHE_POS_START) return; mutex_lock(&slab_mutex); /* Skip kmem_cache_destroy() for active entries */ if (s->refcount > 1) s->refcount--; else if (s->refcount == 1) destroy = true; mutex_unlock(&slab_mutex); if (destroy) kmem_cache_destroy(s); } __bpf_kfunc_end_defs(); struct bpf_iter__kmem_cache { __bpf_md_ptr(struct bpf_iter_meta *, meta); __bpf_md_ptr(struct kmem_cache *, s); }; union kmem_cache_iter_priv { struct bpf_iter_kmem_cache it; struct bpf_iter_kmem_cache_kern kit; }; static void *kmem_cache_iter_seq_start(struct seq_file *seq, loff_t *pos) { loff_t cnt = 0; bool found = false; struct kmem_cache *s; union kmem_cache_iter_priv *p = seq->private; mutex_lock(&slab_mutex); /* Find an entry at the given position in the slab_caches list instead * of keeping a reference (of the last visited entry, if any) out of * slab_mutex. It might miss something if one is deleted in the middle * while it releases the lock. But it should be rare and there's not * much we can do about it. */ list_for_each_entry(s, &slab_caches, list) { if (cnt == *pos) { /* Make sure this entry remains in the list by getting * a new reference count. Note that boot_cache entries * have a negative refcount, so don't touch them. */ if (s->refcount > 0) s->refcount++; found = true; break; } cnt++; } mutex_unlock(&slab_mutex); if (!found) s = NULL; p->kit.pos = s; return s; } static void kmem_cache_iter_seq_stop(struct seq_file *seq, void *v) { struct bpf_iter_meta meta; struct bpf_iter__kmem_cache ctx = { .meta = &meta, .s = v, }; union kmem_cache_iter_priv *p = seq->private; struct bpf_prog *prog; meta.seq = seq; prog = bpf_iter_get_info(&meta, true); if (prog && !ctx.s) bpf_iter_run_prog(prog, &ctx); bpf_iter_kmem_cache_destroy(&p->it); } static void *kmem_cache_iter_seq_next(struct seq_file *seq, void *v, loff_t *pos) { union kmem_cache_iter_priv *p = seq->private; ++*pos; return bpf_iter_kmem_cache_next(&p->it); } static int kmem_cache_iter_seq_show(struct seq_file *seq, void *v) { struct bpf_iter_meta meta; struct bpf_iter__kmem_cache ctx = { .meta = &meta, .s = v, }; struct bpf_prog *prog; int ret = 0; meta.seq = seq; prog = bpf_iter_get_info(&meta, false); if (prog) ret = bpf_iter_run_prog(prog, &ctx); return ret; } static const struct seq_operations kmem_cache_iter_seq_ops = { .start = kmem_cache_iter_seq_start, .next = kmem_cache_iter_seq_next, .stop = kmem_cache_iter_seq_stop, .show = kmem_cache_iter_seq_show, }; BTF_ID_LIST_GLOBAL_SINGLE(bpf_kmem_cache_btf_id, struct, kmem_cache) static const struct bpf_iter_seq_info kmem_cache_iter_seq_info = { .seq_ops = &kmem_cache_iter_seq_ops, .seq_priv_size = sizeof(union kmem_cache_iter_priv), }; static void bpf_iter_kmem_cache_show_fdinfo(const struct bpf_iter_aux_info *aux, struct seq_file *seq) { seq_puts(seq, "kmem_cache iter\n"); } DEFINE_BPF_ITER_FUNC(kmem_cache, struct bpf_iter_meta *meta, struct kmem_cache *s) static struct bpf_iter_reg bpf_kmem_cache_reg_info = { .target = "kmem_cache", .feature = BPF_ITER_RESCHED, .show_fdinfo = bpf_iter_kmem_cache_show_fdinfo, .ctx_arg_info_size = 1, .ctx_arg_info = { { offsetof(struct bpf_iter__kmem_cache, s), PTR_TO_BTF_ID_OR_NULL | PTR_TRUSTED }, }, .seq_info = &kmem_cache_iter_seq_info, }; static int __init bpf_kmem_cache_iter_init(void) { bpf_kmem_cache_reg_info.ctx_arg_info[0].btf_id = bpf_kmem_cache_btf_id[0]; return bpf_iter_reg_target(&bpf_kmem_cache_reg_info); } late_initcall(bpf_kmem_cache_iter_init);