]> granicus.if.org Git - zfs/commitdiff
Update skc_obj_alloc for spl kmem caches that are backed by Linux
authorSerapheim Dimitropoulos <serapheim@delphix.com>
Fri, 18 Oct 2019 17:24:28 +0000 (10:24 -0700)
committerBrian Behlendorf <behlendorf1@llnl.gov>
Fri, 18 Oct 2019 17:24:28 +0000 (13:24 -0400)
Currently, for certain sizes and classes of allocations we use
SPL caches that are backed by caches in the Linux Slab allocator
to reduce fragmentation and increase utilization of memory. The
way things are implemented for these caches as of now though is
that we don't keep any statistics of the allocations that we
make from these caches.

This patch enables the tracking of allocated objects in those
SPL caches by making the trade-off of grabbing the cache lock
at every object allocation and free to update the respective
counter.

Additionally, this patch makes those caches visible in the
/proc/spl/kmem/slab special file.

As a side note, enabling the specific counter for those caches
enables SDB to create a more user-friendly interface than
/proc/spl/kmem/slab that can also cross-reference data from
slabinfo. Here is for example the output of one of those
caches in SDB that outputs the name of the underlying Linux
cache, the memory of SPL objects allocated in that cache,
and the percentage of those objects compared to all the
objects in it:
```
> spl_kmem_caches | filter obj.skc_name == "zio_buf_512" | pp
name        ...            source total_memory util
----------- ... ----------------- ------------ ----
zio_buf_512 ... kmalloc-512[SLUB]       16.9MB    8
```

Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Serapheim Dimitropoulos <serapheim@delphix.com>
Closes #9474

module/os/linux/spl/spl-kmem-cache.c
module/os/linux/spl/spl-proc.c

index 853cf33f0100a6c6e292a4adc194b6a4a28c8934..865f2e2f8e877425ffb368597caa45e1e035eec3 100644 (file)
@@ -1467,6 +1467,17 @@ spl_kmem_cache_alloc(spl_kmem_cache_t *skc, int flags)
                        obj = kmem_cache_alloc(slc, kmem_flags_convert(flags));
                } while ((obj == NULL) && !(flags & KM_NOSLEEP));
 
+               if (obj != NULL) {
+                       /*
+                        * Even though we leave everything up to the
+                        * underlying cache we still keep track of
+                        * how many objects we've allocated in it for
+                        * better debuggability.
+                        */
+                       spin_lock(&skc->skc_lock);
+                       skc->skc_obj_alloc++;
+                       spin_unlock(&skc->skc_lock);
+               }
                goto ret;
        }
 
@@ -1540,6 +1551,9 @@ spl_kmem_cache_free(spl_kmem_cache_t *skc, void *obj)
         */
        if (skc->skc_flags & KMC_SLAB) {
                kmem_cache_free(skc->skc_linux_cache, obj);
+               spin_lock(&skc->skc_lock);
+               skc->skc_obj_alloc--;
+               spin_unlock(&skc->skc_lock);
                return;
        }
 
index a75bcc2145bd4a4cb797080508669854c31f1f78..13eaa6301d7213d9c9dd112d8ef7d7ef5c96b562 100644 (file)
@@ -437,11 +437,29 @@ slab_seq_show(struct seq_file *f, void *p)
 
        ASSERT(skc->skc_magic == SKC_MAGIC);
 
-       /*
-        * Backed by Linux slab see /proc/slabinfo.
-        */
-       if (skc->skc_flags & KMC_SLAB)
+       if (skc->skc_flags & KMC_SLAB) {
+               /*
+                * This cache is backed by a generic Linux kmem cache which
+                * has its own accounting. For these caches we only track
+                * the number of active allocated objects that exist within
+                * the underlying Linux slabs. For the overall statistics of
+                * the underlying Linux cache please refer to /proc/slabinfo.
+                */
+               spin_lock(&skc->skc_lock);
+               seq_printf(f, "%-36s  ", skc->skc_name);
+               seq_printf(f, "0x%05lx %9s %9lu %8s %8u  "
+                   "%5s %5s %5s  %5s %5lu %5s  %5s %5s %5s\n",
+                   (long unsigned)skc->skc_flags,
+                   "-",
+                   (long unsigned)(skc->skc_obj_size * skc->skc_obj_alloc),
+                   "-",
+                   (unsigned)skc->skc_obj_size,
+                   "-", "-", "-", "-",
+                   (long unsigned)skc->skc_obj_alloc,
+                   "-", "-", "-", "-");
+               spin_unlock(&skc->skc_lock);
                return (0);
+       }
 
        spin_lock(&skc->skc_lock);
        seq_printf(f, "%-36s  ", skc->skc_name);
@@ -461,9 +479,7 @@ slab_seq_show(struct seq_file *f, void *p)
            (long unsigned)skc->skc_obj_deadlock,
            (long unsigned)skc->skc_obj_emergency,
            (long unsigned)skc->skc_obj_emergency_max);
-
        spin_unlock(&skc->skc_lock);
-
        return (0);
 }