kfence: add sysfs interface to disable kfence for selected slabs.
authorImran Khan <imran.f.khan@oracle.com>
Sun, 14 Aug 2022 19:53:53 +0000 (05:53 +1000)
committerAndrew Morton <akpm@linux-foundation.org>
Mon, 12 Sep 2022 03:25:52 +0000 (20:25 -0700)
By default kfence allocation can happen for any slab object, whose size is
up to PAGE_SIZE, as long as that allocation is the first allocation after
expiration of kfence sample interval.  But in certain debugging scenarios
we may be interested in debugging corruptions involving some specific slub
objects like dentry or ext4_* etc.  In such cases limiting kfence for
allocations involving only specific slub objects will increase the
probablity of catching the issue since kfence pool will not be consumed by
other slab objects.

This patch introduces a sysfs interface
'/sys/kernel/slab/<name>/skip_kfence' to disable kfence for specific
slabs.  Having the interface work in this way does not impact
current/default behavior of kfence and allows us to use kfence for
specific slabs (when needed) as well.  The decision to skip/use kfence is
taken depending on whether kmem_cache.flags has (newly introduced)
SLAB_SKIP_KFENCE flag set or not.

Link: https://lkml.kernel.org/r/20220814195353.2540848-1-imran.f.khan@oracle.com
Signed-off-by: Imran Khan <imran.f.khan@oracle.com>
Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
Reviewed-by: Marco Elver <elver@google.com>
Reviewed-by: Hyeonggon Yoo <42.hyeyoo@gmail.com>
Cc: Alexander Potapenko <glider@google.com>
Cc: Dmitry Vyukov <dvyukov@google.com>
Cc: Christoph Lameter <cl@linux.com>
Cc: Pekka Enberg <penberg@kernel.org>
Cc: David Rientjes <rientjes@google.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Roman Gushchin <roman.gushchin@linux.dev>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
include/linux/slab.h
mm/kfence/core.c
mm/slub.c

index 0fefdf528e0d2e08287fe349dec0aeeed3e5a78b..352e3f082acce01706d23bf66a46760cb1267c6f 100644 (file)
  */
 #define SLAB_NO_USER_FLAGS     ((slab_flags_t __force)0x10000000U)
 
+#ifdef CONFIG_KFENCE
+#define SLAB_SKIP_KFENCE       ((slab_flags_t __force)0x20000000U)
+#else
+#define SLAB_SKIP_KFENCE       0
+#endif
+
 /* The following flags affect the page allocator grouping pages by mobility */
 /* Objects are reclaimable */
 #define SLAB_RECLAIM_ACCOUNT   ((slab_flags_t __force)0x00020000U)
index c252081b11dfe57f5d608f385785787c06b72347..8c08ae2101d7a254f162bd7e95c217afba10d2c4 100644 (file)
@@ -1003,6 +1003,13 @@ void *__kfence_alloc(struct kmem_cache *s, size_t size, gfp_t flags)
                return NULL;
        }
 
+       /*
+        * Skip allocations for this slab, if KFENCE has been disabled for
+        * this slab.
+        */
+       if (s->flags & SLAB_SKIP_KFENCE)
+               return NULL;
+
        if (atomic_inc_return(&kfence_allocation_gate) > 1)
                return NULL;
 #ifdef CONFIG_KFENCE_STATIC_KEYS
index 862dbd9af4f521fe820a9b5bdb69e7a726ba3d5d..6953c3367bc2003f2f06a31e4c5bd0c0055e058e 100644 (file)
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -5745,6 +5745,29 @@ STAT_ATTR(CPU_PARTIAL_NODE, cpu_partial_node);
 STAT_ATTR(CPU_PARTIAL_DRAIN, cpu_partial_drain);
 #endif /* CONFIG_SLUB_STATS */
 
+#ifdef CONFIG_KFENCE
+static ssize_t skip_kfence_show(struct kmem_cache *s, char *buf)
+{
+       return sysfs_emit(buf, "%d\n", !!(s->flags & SLAB_SKIP_KFENCE));
+}
+
+static ssize_t skip_kfence_store(struct kmem_cache *s,
+                       const char *buf, size_t length)
+{
+       int ret = length;
+
+       if (buf[0] == '0')
+               s->flags &= ~SLAB_SKIP_KFENCE;
+       else if (buf[0] == '1')
+               s->flags |= SLAB_SKIP_KFENCE;
+       else
+               ret = -EINVAL;
+
+       return ret;
+}
+SLAB_ATTR(skip_kfence);
+#endif
+
 static struct attribute *slab_attrs[] = {
        &slab_size_attr.attr,
        &object_size_attr.attr,
@@ -5812,6 +5835,9 @@ static struct attribute *slab_attrs[] = {
        &failslab_attr.attr,
 #endif
        &usersize_attr.attr,
+#ifdef CONFIG_KFENCE
+       &skip_kfence_attr.attr,
+#endif
 
        NULL
 };