diff options
author | Dennis Zhou (Facebook) <dennisszhou@gmail.com> | 2017-07-15 22:23:08 -0400 |
---|---|---|
committer | Tejun Heo <tj@kernel.org> | 2017-07-17 10:46:58 -0400 |
commit | 6b9b6f39946c4a7a98fac31ee91bac43cd2b73a9 (patch) | |
tree | f6405a03492b80c36dec85a784430106d85ce82a /mm | |
parent | 02459164a27e6cf9577fd12fc3ebed78f8fa767f (diff) | |
download | op-kernel-dev-6b9b6f39946c4a7a98fac31ee91bac43cd2b73a9.zip op-kernel-dev-6b9b6f39946c4a7a98fac31ee91bac43cd2b73a9.tar.gz |
percpu: expose pcpu_nr_empty_pop_pages in pcpu_stats
Percpu memory holds a minimum threshold of pages that are populated
in order to serve atomic percpu memory requests. This change makes it
easier to verify that there are a minimum number of populated pages
lying around.
Signed-off-by: Dennis Zhou <dennisszhou@gmail.com>
Signed-off-by: Tejun Heo <tj@kernel.org>
Diffstat (limited to 'mm')
-rw-r--r-- | mm/percpu-internal.h | 1 | ||||
-rw-r--r-- | mm/percpu-stats.c | 1 | ||||
-rw-r--r-- | mm/percpu.c | 2 |
3 files changed, 3 insertions, 1 deletions
diff --git a/mm/percpu-internal.h b/mm/percpu-internal.h index cd2442e..c9158a4 100644 --- a/mm/percpu-internal.h +++ b/mm/percpu-internal.h @@ -36,6 +36,7 @@ extern spinlock_t pcpu_lock; extern struct list_head *pcpu_slot; extern int pcpu_nr_slots; +extern int pcpu_nr_empty_pop_pages; extern struct pcpu_chunk *pcpu_first_chunk; extern struct pcpu_chunk *pcpu_reserved_chunk; diff --git a/mm/percpu-stats.c b/mm/percpu-stats.c index fa0f5de..44e561d 100644 --- a/mm/percpu-stats.c +++ b/mm/percpu-stats.c @@ -164,6 +164,7 @@ alloc_buffer: PU(nr_max_chunks); PU(min_alloc_size); PU(max_alloc_size); + P("empty_pop_pages", pcpu_nr_empty_pop_pages); seq_putc(m, '\n'); #undef PU diff --git a/mm/percpu.c b/mm/percpu.c index bd4130a..9ec5fd4 100644 --- a/mm/percpu.c +++ b/mm/percpu.c @@ -160,7 +160,7 @@ static LIST_HEAD(pcpu_map_extend_chunks); * The number of empty populated pages, protected by pcpu_lock. The * reserved chunk doesn't contribute to the count. */ -static int pcpu_nr_empty_pop_pages; +int pcpu_nr_empty_pop_pages; /* * Balance work is used to populate or destroy chunks asynchronously. We |