mm/sparse: introduce alloc_usemap_and_memmap
authorWanpeng Li <liwanp@linux.vnet.ibm.com>
Wed, 11 Sep 2013 21:22:38 +0000 (14:22 -0700)
committerLinus Torvalds <torvalds@linux-foundation.org>
Wed, 11 Sep 2013 22:58:01 +0000 (15:58 -0700)
After commit 9bdac9142407 ("sparsemem: Put mem map for one node
together."), vmemmap for one node will be allocated together, its logic
is similar as memory allocation for pageblock flags.  This patch
introduces alloc_usemap_and_memmap to extract the same logic of memory
alloction for pageblock flags and vmemmap.

Signed-off-by: Wanpeng Li <liwanp@linux.vnet.ibm.com>
Cc: Dave Hansen <dave.hansen@linux.intel.com>
Cc: Rik van Riel <riel@redhat.com>
Cc: Fengguang Wu <fengguang.wu@intel.com>
Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Tejun Heo <tj@kernel.org>
Cc: Yasuaki Ishimatsu <isimatu.yasuaki@jp.fujitsu.com>
Cc: David Rientjes <rientjes@google.com>
Cc: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
Cc: Jiri Kosina <jkosina@suse.cz>
Cc: Yinghai Lu <yinghai@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
mm/sparse.c

index 308d50331bc353e69f00b8ef7ea78555b1426224..4ac1d7ef548f8ce75b65ae64057c300c8b0cc48c 100644 (file)
@@ -339,13 +339,14 @@ static void __init check_usemap_section_nr(int nid, unsigned long *usemap)
 }
 #endif /* CONFIG_MEMORY_HOTREMOVE */
 
-static void __init sparse_early_usemaps_alloc_node(unsigned long**usemap_map,
+static void __init sparse_early_usemaps_alloc_node(void *data,
                                 unsigned long pnum_begin,
                                 unsigned long pnum_end,
                                 unsigned long usemap_count, int nodeid)
 {
        void *usemap;
        unsigned long pnum;
+       unsigned long **usemap_map = (unsigned long **)data;
        int size = usemap_size();
 
        usemap = sparse_early_usemaps_alloc_pgdat_section(NODE_DATA(nodeid),
@@ -430,11 +431,12 @@ void __init sparse_mem_maps_populate_node(struct page **map_map,
 #endif /* !CONFIG_SPARSEMEM_VMEMMAP */
 
 #ifdef CONFIG_SPARSEMEM_ALLOC_MEM_MAP_TOGETHER
-static void __init sparse_early_mem_maps_alloc_node(struct page **map_map,
+static void __init sparse_early_mem_maps_alloc_node(void *data,
                                 unsigned long pnum_begin,
                                 unsigned long pnum_end,
                                 unsigned long map_count, int nodeid)
 {
+       struct page **map_map = (struct page **)data;
        sparse_mem_maps_populate_node(map_map, pnum_begin, pnum_end,
                                         map_count, nodeid);
 }
@@ -460,6 +462,55 @@ void __attribute__((weak)) __meminit vmemmap_populate_print_last(void)
 {
 }
 
+/**
+ *  alloc_usemap_and_memmap - memory alloction for pageblock flags and vmemmap
+ *  @map: usemap_map for pageblock flags or mmap_map for vmemmap
+ */
+static void __init alloc_usemap_and_memmap(void (*alloc_func)
+                                       (void *, unsigned long, unsigned long,
+                                       unsigned long, int), void *data)
+{
+       unsigned long pnum;
+       unsigned long map_count;
+       int nodeid_begin = 0;
+       unsigned long pnum_begin = 0;
+
+       for (pnum = 0; pnum < NR_MEM_SECTIONS; pnum++) {
+               struct mem_section *ms;
+
+               if (!present_section_nr(pnum))
+                       continue;
+               ms = __nr_to_section(pnum);
+               nodeid_begin = sparse_early_nid(ms);
+               pnum_begin = pnum;
+               break;
+       }
+       map_count = 1;
+       for (pnum = pnum_begin + 1; pnum < NR_MEM_SECTIONS; pnum++) {
+               struct mem_section *ms;
+               int nodeid;
+
+               if (!present_section_nr(pnum))
+                       continue;
+               ms = __nr_to_section(pnum);
+               nodeid = sparse_early_nid(ms);
+               if (nodeid == nodeid_begin) {
+                       map_count++;
+                       continue;
+               }
+               /* ok, we need to take cake of from pnum_begin to pnum - 1*/
+               alloc_func(data, pnum_begin, pnum,
+                                               map_count, nodeid_begin);
+               /* new start, update count etc*/
+               nodeid_begin = nodeid;
+               pnum_begin = pnum;
+               map_count = 1;
+       }
+       /* ok, last chunk */
+       alloc_func(data, pnum_begin, NR_MEM_SECTIONS,
+                                               map_count, nodeid_begin);
+}
+
 /*
  * Allocate the accumulated non-linear sections, allocate a mem_map
  * for each and record the physical to section mapping.
@@ -471,11 +522,7 @@ void __init sparse_init(void)
        unsigned long *usemap;
        unsigned long **usemap_map;
        int size;
-       int nodeid_begin = 0;
-       unsigned long pnum_begin = 0;
-       unsigned long usemap_count;
 #ifdef CONFIG_SPARSEMEM_ALLOC_MEM_MAP_TOGETHER
-       unsigned long map_count;
        int size2;
        struct page **map_map;
 #endif
@@ -501,82 +548,16 @@ void __init sparse_init(void)
        usemap_map = alloc_bootmem(size);
        if (!usemap_map)
                panic("can not allocate usemap_map\n");
-
-       for (pnum = 0; pnum < NR_MEM_SECTIONS; pnum++) {
-               struct mem_section *ms;
-
-               if (!present_section_nr(pnum))
-                       continue;
-               ms = __nr_to_section(pnum);
-               nodeid_begin = sparse_early_nid(ms);
-               pnum_begin = pnum;
-               break;
-       }
-       usemap_count = 1;
-       for (pnum = pnum_begin + 1; pnum < NR_MEM_SECTIONS; pnum++) {
-               struct mem_section *ms;
-               int nodeid;
-
-               if (!present_section_nr(pnum))
-                       continue;
-               ms = __nr_to_section(pnum);
-               nodeid = sparse_early_nid(ms);
-               if (nodeid == nodeid_begin) {
-                       usemap_count++;
-                       continue;
-               }
-               /* ok, we need to take cake of from pnum_begin to pnum - 1*/
-               sparse_early_usemaps_alloc_node(usemap_map, pnum_begin, pnum,
-                                                usemap_count, nodeid_begin);
-               /* new start, update count etc*/
-               nodeid_begin = nodeid;
-               pnum_begin = pnum;
-               usemap_count = 1;
-       }
-       /* ok, last chunk */
-       sparse_early_usemaps_alloc_node(usemap_map, pnum_begin, NR_MEM_SECTIONS,
-                                        usemap_count, nodeid_begin);
+       alloc_usemap_and_memmap(sparse_early_usemaps_alloc_node,
+                                                       (void *)usemap_map);
 
 #ifdef CONFIG_SPARSEMEM_ALLOC_MEM_MAP_TOGETHER
        size2 = sizeof(struct page *) * NR_MEM_SECTIONS;
        map_map = alloc_bootmem(size2);
        if (!map_map)
                panic("can not allocate map_map\n");
-
-       for (pnum = 0; pnum < NR_MEM_SECTIONS; pnum++) {
-               struct mem_section *ms;
-
-               if (!present_section_nr(pnum))
-                       continue;
-               ms = __nr_to_section(pnum);
-               nodeid_begin = sparse_early_nid(ms);
-               pnum_begin = pnum;
-               break;
-       }
-       map_count = 1;
-       for (pnum = pnum_begin + 1; pnum < NR_MEM_SECTIONS; pnum++) {
-               struct mem_section *ms;
-               int nodeid;
-
-               if (!present_section_nr(pnum))
-                       continue;
-               ms = __nr_to_section(pnum);
-               nodeid = sparse_early_nid(ms);
-               if (nodeid == nodeid_begin) {
-                       map_count++;
-                       continue;
-               }
-               /* ok, we need to take cake of from pnum_begin to pnum - 1*/
-               sparse_early_mem_maps_alloc_node(map_map, pnum_begin, pnum,
-                                                map_count, nodeid_begin);
-               /* new start, update count etc*/
-               nodeid_begin = nodeid;
-               pnum_begin = pnum;
-               map_count = 1;
-       }
-       /* ok, last chunk */
-       sparse_early_mem_maps_alloc_node(map_map, pnum_begin, NR_MEM_SECTIONS,
-                                        map_count, nodeid_begin);
+       alloc_usemap_and_memmap(sparse_early_mem_maps_alloc_node,
+                                                       (void *)map_map);
 #endif
 
        for (pnum = 0; pnum < NR_MEM_SECTIONS; pnum++) {