-
Notifications
You must be signed in to change notification settings - Fork 1.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Update skc_obj_alloc for spl kmem caches that are backed by Linux #9474
Merged
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
sdimitro
force-pushed
the
zol-spl-kmem-counters
branch
2 times, most recently
from
October 16, 2019 20:50
fbc475c
to
3a7308f
Compare
ahrens
approved these changes
Oct 16, 2019
behlendorf
requested changes
Oct 17, 2019
Codecov Report
@@ Coverage Diff @@
## master #9474 +/- ##
==========================================
- Coverage 79.28% 79% -0.29%
==========================================
Files 415 415
Lines 123632 123640 +8
==========================================
- Hits 98019 97678 -341
- Misses 25613 25962 +349
Continue to review full report at Codecov.
|
sdimitro
force-pushed
the
zol-spl-kmem-counters
branch
from
October 17, 2019 16:42
3a7308f
to
dd9049a
Compare
Currently, for certain sizes and classes of allocations we use SPL caches that are backed by caches in the Linux Slab allocator to reduce fragmentation and increase utilization of memory. The way things are implemented for these caches as of now though is that we don't keep any statistics of the allocations that we make from these caches. This patch enables the tracking of allocated objects in those SPL caches by making the trade-off of grabbing the cache lock at every object allocation and free to update the respective counter. Additionally, this patch makes those caches visible in the /proc/spl/kmem/slab special file. As a side note, enabling the specific counter for those caches enables SDB to create a more user-friendly interface than /proc/spl/kmem/slab that can also cross-reference data from slabinfo. Here is for example the output of one of those caches in SDB that outputs the name of the underlying Linux cache, the memory of SPL objects allocated in that cache, and the percentage of those objects compared to all the objects in it: ``` > spl_kmem_caches | filter obj.skc_name == "zio_buf_512" | pp name ... source total_memory util ----------- ... ----------------- ------------ ---- zio_buf_512 ... kmalloc-512[SLUB] 16.9MB 8 ``` Signed-off-by: Serapheim Dimitropoulos <serapheim@delphix.com>
sdimitro
force-pushed
the
zol-spl-kmem-counters
branch
from
October 17, 2019 17:01
dd9049a
to
9a170fc
Compare
behlendorf
approved these changes
Oct 18, 2019
behlendorf
added
Status: Accepted
Ready to integrate (reviewed, tested)
and removed
Status: Code Review Needed
Ready for review and testing
labels
Oct 18, 2019
sdimitro
added a commit
to delphix/sdb
that referenced
this pull request
Oct 21, 2019
= Motivation `/proc/spl/kmem/slab` is a useful tool for debugging issues in ZoL that are memory-related. Unfortunately, it is not applicable to crash dumps and not very user friendly overall (e.g. you get almost no information for spl caches that are backed by Linux Slab caches). Besides the above, the specific use-case within Delphix is a replacement for `::kmastat` that we had in MDB in the illumos version of the product. The `spl_kmem_caches` command of this commit is implemented to complement the `slabs` command and cover the functionality of `::kmastat` that we miss further. = Patch This commit introduces the `spl` directory in our code base and implements `spl_kmem_caches`. By default the command iterates over all the SPL caches and prints the following statistics in human-readable form sorted by active_memory used: ``` name entry_size active_objs active_memory source total_memory util ------------------------ ---------- ----------- ------------- ------------------------------ ------------ ---- zio_data_buf_131072 135680 760 98.3MB zio_data_buf_131072[SPL ] 119.0MB 82 dnode_t 1168 25533 28.4MB dnode_t[SLUB] 28.6MB 99 zfs_znode_cache 1104 23039 24.3MB zfs_znode_cache[SLUB] 24.4MB 99 ... ``` All the properties can be seen in the help message and the user may specify the properties they are looking for with the `-o` option (and also sort them by a specific field using `-s`): ``` > spl_kmem_caches -h usage: spl_kmem_caches [-h] [-H] [-o FIELDS] [-p] [-r] [-s FIELD] [-v] optional arguments: -h, --help show this help message and exit -H do not display headers and separate fields by a single tab (scripted mode) -o FIELDS comma-separated list of fields to display -p display numbers in parseable (exact) values -r, --recursive recurse down children caches and print statistics -s FIELD sort rows by FIELD -v Print all statistics FIELDS := address, name, flags, object_size, entry_size, slab_size, objects_per_slab, entries_per_slab, slabs, active_slabs, active_memory, total_memory, objs, active_objs, inactive_objs, source, util If -o is not specified the default fields used are name, entry_size, active_objs, active_memory, source, total_memory, util. If the -s option is not specified and the command's output is not piped anywhere then we sort by the following fields in order: active_memory, address, name. If none of those exists in the field-set we sort by the first field specified in the set. > spl_kmem_caches -o "name,flags,slabs" -s slabs name flags slabs ------------------------ ----------------------- ----- zio_data_buf_131072 KMC_NODEBUG|KMC_VMEM 115 zio_buf_131072 KMC_NODEBUG|KMC_VMEM 25 zio_data_buf_40960 KMC_NODEBUG|KMC_VMEM 19 ... ``` Besides a PrettyPrinter, the command is also a Locator which means that the user can do things like this: ``` > spl_kmem_caches -s total_memory | head 1 | pp name entry_size active_objs active_memory source total_memory util ------------------- ---------- ----------- ------------- ------------------------- ------------ ---- zio_data_buf_131072 135680 760 98.3MB zio_data_buf_131072[SPL ] 119.0MB 82 ``` A couple more options mimicking Illumos conventions are the `-p` and `-H` options that output numbers to raw form and skip printing the headers respectively. This is generally useful for scripts or output that is to be processed later: ``` > spl_kmem_caches -H -p zio_data_buf_131072 135680 760 103116800 zio_data_buf_131072[SPL ] 124825600 82 dnode_t 1168 25573 29869264 dnode_t[SLUB] 30005920 99 zfs_znode_cache 1104 23039 25435056 zfs_znode_cache[SLUB] 25548768 99 ``` Note that for these command to fully work for SPL caches that are backed by Linux Slab caches we need the appropriate ZoL changes that enables us to track allocations for those cases (see PR: openzfs/zfs#9474). If the commit from that PR is not in the running system then all of these caches will incorectly be reported empty. Side-change: As the `spl_kmem_caches` command is very close both in terms of implementation and functionality with the `slabs` command this patch also updates some of the columns and calculations in the output of `slabs`, to make the commands consistent with each other. = Testing/Verification I verified the values of all their fields in the following ways: [1] For the fields that are also visible in /proc/slabinfo, I ensured that the values matched. [2] For anything else I did the math and cross-referenced all the values ensuring that they are consistent with each other. Besides the manual testing done shown in the output above I also ensured the following error cases: * ask for invalid field name ``` > spl_kmem_caches -o bogus sdb: spl_kmem_caches: 'bogus' is not a valid field ``` * sort by invalid field name ``` sdb: spl_kmem_caches: invalid input: 'bogus' is not in field set (name, entry_size, active_objs, active_memory, source, total_memory, util) ``` * attempt to sort by field that is not in the requested field set ``` > spl_kmem_caches -o "name,address" -s objs sdb: spl_kmem_caches: invalid input: 'objs' is not in field set (name, address) ``` = Other Notes The command only works with SLUB and until we have proper lexing we still pass fields like this: `> spl_kmem_caches -o "name,address"`
sdimitro
added a commit
to delphix/sdb
that referenced
this pull request
Oct 25, 2019
= Motivation `/proc/spl/kmem/slab` is a useful tool for debugging issues in ZoL that are memory-related. Unfortunately, it is not applicable to crash dumps and not very user friendly overall (e.g. you get almost no information for spl caches that are backed by Linux Slab caches). Besides the above, the specific use-case within Delphix is a replacement for `::kmastat` that we had in MDB in the illumos version of the product. The `spl_kmem_caches` command of this commit is implemented to complement the `slabs` command and cover the functionality of `::kmastat` that we miss further. = Patch This commit introduces the `spl` directory in our code base and implements `spl_kmem_caches`. By default the command iterates over all the SPL caches and prints the following statistics in human-readable form sorted by active_memory used: ``` name entry_size active_objs active_memory source total_memory util ------------------------ ---------- ----------- ------------- ------------------------------ ------------ ---- zio_data_buf_131072 135680 760 98.3MB zio_data_buf_131072[SPL ] 119.0MB 82 dnode_t 1168 25533 28.4MB dnode_t[SLUB] 28.6MB 99 zfs_znode_cache 1104 23039 24.3MB zfs_znode_cache[SLUB] 24.4MB 99 ... ``` All the properties can be seen in the help message and the user may specify the properties they are looking for with the `-o` option (and also sort them by a specific field using `-s`): ``` > spl_kmem_caches -h usage: spl_kmem_caches [-h] [-H] [-o FIELDS] [-p] [-r] [-s FIELD] [-v] optional arguments: -h, --help show this help message and exit -H do not display headers and separate fields by a single tab (scripted mode) -o FIELDS comma-separated list of fields to display -p display numbers in parseable (exact) values -r, --recursive recurse down children caches and print statistics -s FIELD sort rows by FIELD -v Print all statistics FIELDS := address, name, flags, object_size, entry_size, slab_size, objects_per_slab, entries_per_slab, slabs, active_slabs, active_memory, total_memory, objs, active_objs, inactive_objs, source, util If -o is not specified the default fields used are name, entry_size, active_objs, active_memory, source, total_memory, util. If the -s option is not specified and the command's output is not piped anywhere then we sort by the following fields in order: active_memory, address, name. If none of those exists in the field-set we sort by the first field specified in the set. > spl_kmem_caches -o "name,flags,slabs" -s slabs name flags slabs ------------------------ ----------------------- ----- zio_data_buf_131072 KMC_NODEBUG|KMC_VMEM 115 zio_buf_131072 KMC_NODEBUG|KMC_VMEM 25 zio_data_buf_40960 KMC_NODEBUG|KMC_VMEM 19 ... ``` Besides a PrettyPrinter, the command is also a Locator which means that the user can do things like this: ``` > spl_kmem_caches -s total_memory | head 1 | pp name entry_size active_objs active_memory source total_memory util ------------------- ---------- ----------- ------------- ------------------------- ------------ ---- zio_data_buf_131072 135680 760 98.3MB zio_data_buf_131072[SPL ] 119.0MB 82 ``` A couple more options mimicking Illumos conventions are the `-p` and `-H` options that output numbers to raw form and skip printing the headers respectively. This is generally useful for scripts or output that is to be processed later: ``` > spl_kmem_caches -H -p zio_data_buf_131072 135680 760 103116800 zio_data_buf_131072[SPL ] 124825600 82 dnode_t 1168 25573 29869264 dnode_t[SLUB] 30005920 99 zfs_znode_cache 1104 23039 25435056 zfs_znode_cache[SLUB] 25548768 99 ``` Note that for these command to fully work for SPL caches that are backed by Linux Slab caches we need the appropriate ZoL changes that enables us to track allocations for those cases (see PR: openzfs/zfs#9474). If the commit from that PR is not in the running system then all of these caches will incorectly be reported empty. Side-change: As the `spl_kmem_caches` command is very close both in terms of implementation and functionality with the `slabs` command this patch also updates some of the columns and calculations in the output of `slabs`, to make the commands consistent with each other. = Testing/Verification I verified the values of all their fields in the following ways: [1] For the fields that are also visible in /proc/slabinfo, I ensured that the values matched. [2] For anything else I did the math and cross-referenced all the values ensuring that they are consistent with each other. Besides the manual testing done shown in the output above I also ensured the following error cases: * ask for invalid field name ``` > spl_kmem_caches -o bogus sdb: spl_kmem_caches: 'bogus' is not a valid field ``` * sort by invalid field name ``` sdb: spl_kmem_caches: invalid input: 'bogus' is not in field set (name, entry_size, active_objs, active_memory, source, total_memory, util) ``` * attempt to sort by field that is not in the requested field set ``` > spl_kmem_caches -o "name,address" -s objs sdb: spl_kmem_caches: invalid input: 'objs' is not in field set (name, address) ``` = Other Notes The command only works with SLUB and until we have proper lexing we still pass fields like this: `> spl_kmem_caches -o "name,address"`
tonyhutter
pushed a commit
to tonyhutter/zfs
that referenced
this pull request
Dec 26, 2019
Currently, for certain sizes and classes of allocations we use SPL caches that are backed by caches in the Linux Slab allocator to reduce fragmentation and increase utilization of memory. The way things are implemented for these caches as of now though is that we don't keep any statistics of the allocations that we make from these caches. This patch enables the tracking of allocated objects in those SPL caches by making the trade-off of grabbing the cache lock at every object allocation and free to update the respective counter. Additionally, this patch makes those caches visible in the /proc/spl/kmem/slab special file. As a side note, enabling the specific counter for those caches enables SDB to create a more user-friendly interface than /proc/spl/kmem/slab that can also cross-reference data from slabinfo. Here is for example the output of one of those caches in SDB that outputs the name of the underlying Linux cache, the memory of SPL objects allocated in that cache, and the percentage of those objects compared to all the objects in it: ``` > spl_kmem_caches | filter obj.skc_name == "zio_buf_512" | pp name ... source total_memory util ----------- ... ----------------- ------------ ---- zio_buf_512 ... kmalloc-512[SLUB] 16.9MB 8 ``` Reviewed-by: Matt Ahrens <matt@delphix.com> Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov> Signed-off-by: Serapheim Dimitropoulos <serapheim@delphix.com> Closes openzfs#9474
tonyhutter
pushed a commit
to tonyhutter/zfs
that referenced
this pull request
Dec 27, 2019
Currently, for certain sizes and classes of allocations we use SPL caches that are backed by caches in the Linux Slab allocator to reduce fragmentation and increase utilization of memory. The way things are implemented for these caches as of now though is that we don't keep any statistics of the allocations that we make from these caches. This patch enables the tracking of allocated objects in those SPL caches by making the trade-off of grabbing the cache lock at every object allocation and free to update the respective counter. Additionally, this patch makes those caches visible in the /proc/spl/kmem/slab special file. As a side note, enabling the specific counter for those caches enables SDB to create a more user-friendly interface than /proc/spl/kmem/slab that can also cross-reference data from slabinfo. Here is for example the output of one of those caches in SDB that outputs the name of the underlying Linux cache, the memory of SPL objects allocated in that cache, and the percentage of those objects compared to all the objects in it: ``` > spl_kmem_caches | filter obj.skc_name == "zio_buf_512" | pp name ... source total_memory util ----------- ... ----------------- ------------ ---- zio_buf_512 ... kmalloc-512[SLUB] 16.9MB 8 ``` Reviewed-by: Matt Ahrens <matt@delphix.com> Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov> Signed-off-by: Serapheim Dimitropoulos <serapheim@delphix.com> Closes openzfs#9474
tonyhutter
pushed a commit
that referenced
this pull request
Jan 23, 2020
Currently, for certain sizes and classes of allocations we use SPL caches that are backed by caches in the Linux Slab allocator to reduce fragmentation and increase utilization of memory. The way things are implemented for these caches as of now though is that we don't keep any statistics of the allocations that we make from these caches. This patch enables the tracking of allocated objects in those SPL caches by making the trade-off of grabbing the cache lock at every object allocation and free to update the respective counter. Additionally, this patch makes those caches visible in the /proc/spl/kmem/slab special file. As a side note, enabling the specific counter for those caches enables SDB to create a more user-friendly interface than /proc/spl/kmem/slab that can also cross-reference data from slabinfo. Here is for example the output of one of those caches in SDB that outputs the name of the underlying Linux cache, the memory of SPL objects allocated in that cache, and the percentage of those objects compared to all the objects in it: ``` > spl_kmem_caches | filter obj.skc_name == "zio_buf_512" | pp name ... source total_memory util ----------- ... ----------------- ------------ ---- zio_buf_512 ... kmalloc-512[SLUB] 16.9MB 8 ``` Reviewed-by: Matt Ahrens <matt@delphix.com> Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov> Signed-off-by: Serapheim Dimitropoulos <serapheim@delphix.com> Closes #9474
sdimitro
added a commit
to sdimitro/zfs
that referenced
this pull request
Mar 13, 2020
Currently, for certain sizes and classes of allocations we use SPL caches that are backed by caches in the Linux Slab allocator to reduce fragmentation and increase utilization of memory. The way things are implemented for these caches as of now though is that we don't keep any statistics of the allocations that we make from these caches. This patch enables the tracking of allocated objects in those SPL caches by making the trade-off of grabbing the cache lock at every object allocation and free to update the respective counter. Additionally, this patch makes those caches visible in the /proc/spl/kmem/slab special file. As a side note, enabling the specific counter for those caches enables SDB to create a more user-friendly interface than /proc/spl/kmem/slab that can also cross-reference data from slabinfo. Here is for example the output of one of those caches in SDB that outputs the name of the underlying Linux cache, the memory of SPL objects allocated in that cache, and the percentage of those objects compared to all the objects in it: ``` > spl_kmem_caches | filter obj.skc_name == "zio_buf_512" | pp name ... source total_memory util ----------- ... ----------------- ------------ ---- zio_buf_512 ... kmalloc-512[SLUB] 16.9MB 8 ``` Reviewed-by: Matt Ahrens <matt@delphix.com> Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov> Signed-off-by: Serapheim Dimitropoulos <serapheim@delphix.com> Closes openzfs#9474
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Currently, for certain sizes and classes of allocations we use
SPL caches that are backed by caches in the Linux Slab allocator
to reduce fragmentation and increase utilization of memory. The
way things are implemented for these caches as of now though is
that we don't keep any statistics of the allocations that we
make from these caches.
This patch enables the tracking of allocated objects in those
SPL caches by making the trade-off of grabbing the cache lock
at every object allocation and free to update the respective
counter.
Additionally, this patch makes those caches visible in the
/proc/spl/kmem/slab special file.
As a side note, enabling the specific counter for those caches
enables SDB to create a more user-friendly interface than
/proc/spl/kmem/slab that can also cross-reference data from
slabinfo. Here is for example the output of one of those
caches in SDB that outputs the name of the underlying Linux
cache, the memory of SPL objects allocated in that cache,
and the percentage of those objects compared to all the
objects in it:
Signed-off-by: Serapheim Dimitropoulos serapheim@delphix.com
How Has This Been Tested?
Besides running the test-suite to ensure no regression, I did some manual testing by double checking the values in the
/proc
filesystem and withsdb
. I'm open to feedback on how to test this further.Types of changes
Checklist:
Signed-off-by
.