cpumask: Use mm_cpumask() wrapper instead of cpu_vm_mask
authorRusty Russell <rusty@rustcorp.com.au>
Sun, 15 Mar 2009 18:16:43 +0000 (18:16 +0000)
committerBenjamin Herrenschmidt <benh@kernel.crashing.org>
Tue, 24 Mar 2009 02:47:29 +0000 (13:47 +1100)
Makes code futureproof against the impending change to mm->cpu_vm_mask.

It's also a chance to use the new cpumask_ ops which take a pointer
(the older ones are deprecated, but there's no hurry for arch code).

Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org>
arch/powerpc/include/asm/mmu_context.h
arch/powerpc/mm/hash_utils_64.c
arch/powerpc/mm/mmu_context_nohash.c
arch/powerpc/mm/pgtable.c
arch/powerpc/mm/tlb_hash64.c
arch/powerpc/mm/tlb_nohash.c
arch/powerpc/platforms/cell/spu_base.c

index ab4f19263c4286b199a084ea5dff32259d50820b..b7063669f972b04238c16c945c34987a67e81524 100644 (file)
@@ -31,7 +31,7 @@ static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
                             struct task_struct *tsk)
 {
        /* Mark this context has been used on the new CPU */
-       cpu_set(smp_processor_id(), next->cpu_vm_mask);
+       cpumask_set_cpu(smp_processor_id(), mm_cpumask(next));
 
        /* 32-bit keeps track of the current PGDIR in the thread struct */
 #ifdef CONFIG_PPC32
index f5bc1b213f2478d986c91ad5e55c54cdaa7b6849..86c00c885e682e31591592673b8e36f716b7fde3 100644 (file)
@@ -859,7 +859,7 @@ int hash_page(unsigned long ea, unsigned long access, unsigned long trap)
        unsigned long vsid;
        struct mm_struct *mm;
        pte_t *ptep;
-       cpumask_t tmp;
+       const struct cpumask *tmp;
        int rc, user_region = 0, local = 0;
        int psize, ssize;
 
@@ -907,8 +907,8 @@ int hash_page(unsigned long ea, unsigned long access, unsigned long trap)
                return 1;
 
        /* Check CPU locality */
-       tmp = cpumask_of_cpu(smp_processor_id());
-       if (user_region && cpus_equal(mm->cpu_vm_mask, tmp))
+       tmp = cpumask_of(smp_processor_id());
+       if (user_region && cpumask_equal(mm_cpumask(mm), tmp))
                local = 1;
 
 #ifdef CONFIG_HUGETLB_PAGE
@@ -1024,7 +1024,6 @@ void hash_preload(struct mm_struct *mm, unsigned long ea,
        unsigned long vsid;
        void *pgdir;
        pte_t *ptep;
-       cpumask_t mask;
        unsigned long flags;
        int local = 0;
        int ssize;
@@ -1067,8 +1066,7 @@ void hash_preload(struct mm_struct *mm, unsigned long ea,
        local_irq_save(flags);
 
        /* Is that local to this CPU ? */
-       mask = cpumask_of_cpu(smp_processor_id());
-       if (cpus_equal(mm->cpu_vm_mask, mask))
+       if (cpumask_equal(mm_cpumask(mm), cpumask_of(smp_processor_id())))
                local = 1;
 
        /* Hash it in */
index 52a0cfc38b6488eecd50374d56271f1fc3a83553..ac4cb04ceb69edc6a1e4c2379125642016aee377 100644 (file)
@@ -97,7 +97,7 @@ static unsigned int steal_context_smp(unsigned int id)
                mm->context.id = MMU_NO_CONTEXT;
 
                /* Mark it stale on all CPUs that used this mm */
-               for_each_cpu_mask_nr(cpu, mm->cpu_vm_mask)
+               for_each_cpu(cpu, mm_cpumask(mm))
                        __set_bit(id, stale_map[cpu]);
                return id;
        }
index a27ded3adac5aa8316213d987a3e3443d052d25f..f5c6fd42265c795a00e6244a9b694d282832bb06 100644 (file)
@@ -82,11 +82,10 @@ static void pte_free_submit(struct pte_freelist_batch *batch)
 void pgtable_free_tlb(struct mmu_gather *tlb, pgtable_free_t pgf)
 {
        /* This is safe since tlb_gather_mmu has disabled preemption */
-        cpumask_t local_cpumask = cpumask_of_cpu(smp_processor_id());
        struct pte_freelist_batch **batchp = &__get_cpu_var(pte_freelist_cur);
 
        if (atomic_read(&tlb->mm->mm_users) < 2 ||
-           cpus_equal(tlb->mm->cpu_vm_mask, local_cpumask)) {
+           cpumask_equal(mm_cpumask(tlb->mm), cpumask_of(smp_processor_id()))){
                pgtable_free(pgf);
                return;
        }
index c931bc7d1079bec0e9241620781aa5eb5251aba6..1be1b5e597962727bf41607684d6767b98548238 100644 (file)
@@ -139,12 +139,12 @@ void hpte_need_flush(struct mm_struct *mm, unsigned long addr,
  */
 void __flush_tlb_pending(struct ppc64_tlb_batch *batch)
 {
-       cpumask_t tmp;
+       const struct cpumask *tmp;
        int i, local = 0;
 
        i = batch->index;
-       tmp = cpumask_of_cpu(smp_processor_id());
-       if (cpus_equal(batch->mm->cpu_vm_mask, tmp))
+       tmp = cpumask_of(smp_processor_id());
+       if (cpumask_equal(mm_cpumask(batch->mm), tmp))
                local = 1;
        if (i == 1)
                flush_hash_page(batch->vaddr[0], batch->pte[0],
index 39ac22b13c73ddaa10a0379129e8b2c279f97c95..7af72970faed2c8b33c3e4d63e2b61b297621f36 100644 (file)
@@ -132,11 +132,11 @@ void flush_tlb_mm(struct mm_struct *mm)
        pid = mm->context.id;
        if (unlikely(pid == MMU_NO_CONTEXT))
                goto no_context;
-       cpu_mask = mm->cpu_vm_mask;
-       cpu_clear(smp_processor_id(), cpu_mask);
-       if (!cpus_empty(cpu_mask)) {
+       if (!cpumask_equal(mm_cpumask(mm), cpumask_of(smp_processor_id()))) {
                struct tlb_flush_param p = { .pid = pid };
-               smp_call_function_mask(cpu_mask, do_flush_tlb_mm_ipi, &p, 1);
+               /* Ignores smp_processor_id() even if set. */
+               smp_call_function_many(mm_cpumask(mm),
+                                      do_flush_tlb_mm_ipi, &p, 1);
        }
        _tlbil_pid(pid);
  no_context:
@@ -146,16 +146,15 @@ EXPORT_SYMBOL(flush_tlb_mm);
 
 void flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr)
 {
-       cpumask_t cpu_mask;
+       struct cpumask *cpu_mask;
        unsigned int pid;
 
        preempt_disable();
        pid = vma ? vma->vm_mm->context.id : 0;
        if (unlikely(pid == MMU_NO_CONTEXT))
                goto bail;
-       cpu_mask = vma->vm_mm->cpu_vm_mask;
-       cpu_clear(smp_processor_id(), cpu_mask);
-       if (!cpus_empty(cpu_mask)) {
+       cpu_mask = mm_cpumask(vma->vm_mm);
+       if (!cpumask_equal(cpu_mask, cpumask_of(smp_processor_id()))) {
                /* If broadcast tlbivax is supported, use it */
                if (mmu_has_feature(MMU_FTR_USE_TLBIVAX_BCAST)) {
                        int lock = mmu_has_feature(MMU_FTR_LOCK_BCAST_INVAL);
@@ -167,7 +166,8 @@ void flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr)
                        goto bail;
                } else {
                        struct tlb_flush_param p = { .pid = pid, .addr = vmaddr };
-                       smp_call_function_mask(cpu_mask,
+                       /* Ignores smp_processor_id() even if set in cpu_mask */
+                       smp_call_function_many(cpu_mask,
                                               do_flush_tlb_page_ipi, &p, 1);
                }
        }
index e487ad68ac1113882db05b88142dd1542b67cdd4..9abd210d87c1362f831b7a50bae27bb78d6c31f7 100644 (file)
@@ -114,7 +114,7 @@ static inline void mm_needs_global_tlbie(struct mm_struct *mm)
        int nr = (NR_CPUS > 1) ? NR_CPUS : NR_CPUS + 1;
 
        /* Global TLBIE broadcast required with SPEs. */
-       __cpus_setall(&mm->cpu_vm_mask, nr);
+       bitmap_fill(cpumask_bits(mm_cpumask(mm)), nr);
 }
 
 void spu_associate_mm(struct spu *spu, struct mm_struct *mm)