mm: oom analysis: add shmem vmstat
authorKOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
Tue, 22 Sep 2009 00:01:33 +0000 (17:01 -0700)
committerLinus Torvalds <torvalds@linux-foundation.org>
Tue, 22 Sep 2009 14:17:27 +0000 (07:17 -0700)
Recently we encountered OOM problems due to memory use of the GEM cache.
Generally a large amuont of Shmem/Tmpfs pages tend to create a memory
shortage problem.

We often use the following calculation to determine the amount of shmem
pages:

shmem = NR_ACTIVE_ANON + NR_INACTIVE_ANON - NR_ANON_PAGES

however the expression does not consider isolated and mlocked pages.

This patch adds explicit accounting for pages used by shmem and tmpfs.

Signed-off-by: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
Acked-by: Rik van Riel <riel@redhat.com>
Reviewed-by: Christoph Lameter <cl@linux-foundation.org>
Acked-by: Wu Fengguang <fengguang.wu@intel.com>
Cc: David Rientjes <rientjes@google.com>
Cc: Hugh Dickins <hugh.dickins@tiscali.co.uk>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
drivers/base/node.c
fs/proc/meminfo.c
include/linux/mmzone.h
mm/filemap.c
mm/migrate.c
mm/page_alloc.c
mm/vmstat.c

index b560c17f6d4e2a79b080fafaac75b61558e29b4a..1fe5536d404f134e53b61324f9350c11080337dc 100644 (file)
@@ -85,6 +85,7 @@ static ssize_t node_read_meminfo(struct sys_device * dev,
                       "Node %d FilePages:      %8lu kB\n"
                       "Node %d Mapped:         %8lu kB\n"
                       "Node %d AnonPages:      %8lu kB\n"
+                      "Node %d Shmem:          %8lu kB\n"
                       "Node %d KernelStack:    %8lu kB\n"
                       "Node %d PageTables:     %8lu kB\n"
                       "Node %d NFS_Unstable:   %8lu kB\n"
@@ -117,6 +118,7 @@ static ssize_t node_read_meminfo(struct sys_device * dev,
                       nid, K(node_page_state(nid, NR_FILE_PAGES)),
                       nid, K(node_page_state(nid, NR_FILE_MAPPED)),
                       nid, K(node_page_state(nid, NR_ANON_PAGES)),
+                      nid, K(node_page_state(nid, NR_SHMEM)),
                       nid, node_page_state(nid, NR_KERNEL_STACK) *
                                THREAD_SIZE / 1024,
                       nid, K(node_page_state(nid, NR_PAGETABLE)),
index 1fc588f430e4b94c0f40a210fb05bc3bdf186650..171e052c07b3684f4a4264ebb44e802affc20f6e 100644 (file)
@@ -81,6 +81,7 @@ static int meminfo_proc_show(struct seq_file *m, void *v)
                "Writeback:      %8lu kB\n"
                "AnonPages:      %8lu kB\n"
                "Mapped:         %8lu kB\n"
+               "Shmem:          %8lu kB\n"
                "Slab:           %8lu kB\n"
                "SReclaimable:   %8lu kB\n"
                "SUnreclaim:     %8lu kB\n"
@@ -125,6 +126,7 @@ static int meminfo_proc_show(struct seq_file *m, void *v)
                K(global_page_state(NR_WRITEBACK)),
                K(global_page_state(NR_ANON_PAGES)),
                K(global_page_state(NR_FILE_MAPPED)),
+               K(global_page_state(NR_SHMEM)),
                K(global_page_state(NR_SLAB_RECLAIMABLE) +
                                global_page_state(NR_SLAB_UNRECLAIMABLE)),
                K(global_page_state(NR_SLAB_RECLAIMABLE)),
index d9335b8de84af682850910fa399acfeb3159e89b..b3583b93b77e5a04265e7751c187efc8765f12de 100644 (file)
@@ -100,6 +100,7 @@ enum zone_stat_item {
        NR_BOUNCE,
        NR_VMSCAN_WRITE,
        NR_WRITEBACK_TEMP,      /* Writeback using temporary buffers */
+       NR_SHMEM,               /* shmem pages (included tmpfs/GEM pages) */
 #ifdef CONFIG_NUMA
        NUMA_HIT,               /* allocated in intended node */
        NUMA_MISS,              /* allocated in non intended node */
index dd51c68e2b868adbb403b5967c4e7c6be357e722..bcc7372aebbc4375d0763e4f3acd8d096bcb612d 100644 (file)
@@ -119,6 +119,8 @@ void __remove_from_page_cache(struct page *page)
        page->mapping = NULL;
        mapping->nrpages--;
        __dec_zone_page_state(page, NR_FILE_PAGES);
+       if (PageSwapBacked(page))
+               __dec_zone_page_state(page, NR_SHMEM);
        BUG_ON(page_mapped(page));
 
        /*
@@ -431,6 +433,8 @@ int add_to_page_cache_locked(struct page *page, struct address_space *mapping,
                if (likely(!error)) {
                        mapping->nrpages++;
                        __inc_zone_page_state(page, NR_FILE_PAGES);
+                       if (PageSwapBacked(page))
+                               __inc_zone_page_state(page, NR_SHMEM);
                        spin_unlock_irq(&mapping->tree_lock);
                } else {
                        page->mapping = NULL;
index 0edeac91348d98f51f5f89a173268b917707fda7..37143b924484438b82af273e6b62cf4f587d74be 100644 (file)
@@ -312,7 +312,10 @@ static int migrate_page_move_mapping(struct address_space *mapping,
         */
        __dec_zone_page_state(page, NR_FILE_PAGES);
        __inc_zone_page_state(newpage, NR_FILE_PAGES);
-
+       if (PageSwapBacked(page)) {
+               __dec_zone_page_state(page, NR_SHMEM);
+               __inc_zone_page_state(newpage, NR_SHMEM);
+       }
        spin_unlock_irq(&mapping->tree_lock);
 
        return 0;
index 4e050f325ebd8ca1ed49362d390e747ea02bd6f0..e50c22545b8f9c6412d137675cc9c45b133d6a79 100644 (file)
@@ -2139,7 +2139,7 @@ void show_free_areas(void)
                " unevictable:%lu"
                " dirty:%lu writeback:%lu unstable:%lu buffer:%lu\n"
                " free:%lu slab_reclaimable:%lu slab_unreclaimable:%lu\n"
-               " mapped:%lu pagetables:%lu bounce:%lu\n",
+               " mapped:%lu shmem:%lu pagetables:%lu bounce:%lu\n",
                global_page_state(NR_ACTIVE_ANON),
                global_page_state(NR_ACTIVE_FILE),
                global_page_state(NR_INACTIVE_ANON),
@@ -2153,6 +2153,7 @@ void show_free_areas(void)
                global_page_state(NR_SLAB_RECLAIMABLE),
                global_page_state(NR_SLAB_UNRECLAIMABLE),
                global_page_state(NR_FILE_MAPPED),
+               global_page_state(NR_SHMEM),
                global_page_state(NR_PAGETABLE),
                global_page_state(NR_BOUNCE));
 
@@ -2175,6 +2176,7 @@ void show_free_areas(void)
                        " dirty:%lukB"
                        " writeback:%lukB"
                        " mapped:%lukB"
+                       " shmem:%lukB"
                        " slab_reclaimable:%lukB"
                        " slab_unreclaimable:%lukB"
                        " kernel_stack:%lukB"
@@ -2200,6 +2202,7 @@ void show_free_areas(void)
                        K(zone_page_state(zone, NR_FILE_DIRTY)),
                        K(zone_page_state(zone, NR_WRITEBACK)),
                        K(zone_page_state(zone, NR_FILE_MAPPED)),
+                       K(zone_page_state(zone, NR_SHMEM)),
                        K(zone_page_state(zone, NR_SLAB_RECLAIMABLE)),
                        K(zone_page_state(zone, NR_SLAB_UNRECLAIMABLE)),
                        zone_page_state(zone, NR_KERNEL_STACK) *
index ceda39b63d7ea37c6e0e3fa3bf7fef3457627684..7214a45112579f5175e5c5e09b65f353fe6e333c 100644 (file)
@@ -644,7 +644,7 @@ static const char * const vmstat_text[] = {
        "nr_bounce",
        "nr_vmscan_write",
        "nr_writeback_temp",
-
+       "nr_shmem",
 #ifdef CONFIG_NUMA
        "numa_hit",
        "numa_miss",