drm/nouveau: add instmem flush() hook
authorBen Skeggs <bskeggs@redhat.com>
Thu, 8 Jul 2010 01:29:10 +0000 (11:29 +1000)
committerBen Skeggs <bskeggs@redhat.com>
Tue, 13 Jul 2010 00:13:40 +0000 (10:13 +1000)
This removes the previous prepare_access() and finish_access() hooks, and
replaces it with a much simpler flush() hook.

All the chipset-specific code before nv50 has its use removed completely,
as it's not required there at all.

Signed-off-by: Ben Skeggs <bskeggs@redhat.com>
15 files changed:
drivers/gpu/drm/nouveau/nouveau_drv.h
drivers/gpu/drm/nouveau/nouveau_mem.c
drivers/gpu/drm/nouveau/nouveau_object.c
drivers/gpu/drm/nouveau/nouveau_sgdma.c
drivers/gpu/drm/nouveau/nouveau_state.c
drivers/gpu/drm/nouveau/nv04_fifo.c
drivers/gpu/drm/nouveau/nv04_instmem.c
drivers/gpu/drm/nouveau/nv10_fifo.c
drivers/gpu/drm/nouveau/nv20_graph.c
drivers/gpu/drm/nouveau/nv40_fifo.c
drivers/gpu/drm/nouveau/nv40_graph.c
drivers/gpu/drm/nouveau/nv50_display.c
drivers/gpu/drm/nouveau/nv50_fifo.c
drivers/gpu/drm/nouveau/nv50_graph.c
drivers/gpu/drm/nouveau/nv50_instmem.c

index afebd32af2032a196a56b51ac3edea1c8361ee23..e21eacc47290daed4db3b4cd8411c7708279cddd 100644 (file)
@@ -269,8 +269,7 @@ struct nouveau_instmem_engine {
        void    (*clear)(struct drm_device *, struct nouveau_gpuobj *);
        int     (*bind)(struct drm_device *, struct nouveau_gpuobj *);
        int     (*unbind)(struct drm_device *, struct nouveau_gpuobj *);
-       void    (*prepare_access)(struct drm_device *, bool write);
-       void    (*finish_access)(struct drm_device *);
+       void    (*flush)(struct drm_device *);
 };
 
 struct nouveau_mc_engine {
@@ -1027,8 +1026,7 @@ extern int  nv04_instmem_populate(struct drm_device *, struct nouveau_gpuobj *,
 extern void nv04_instmem_clear(struct drm_device *, struct nouveau_gpuobj *);
 extern int  nv04_instmem_bind(struct drm_device *, struct nouveau_gpuobj *);
 extern int  nv04_instmem_unbind(struct drm_device *, struct nouveau_gpuobj *);
-extern void nv04_instmem_prepare_access(struct drm_device *, bool write);
-extern void nv04_instmem_finish_access(struct drm_device *);
+extern void nv04_instmem_flush(struct drm_device *);
 
 /* nv50_instmem.c */
 extern int  nv50_instmem_init(struct drm_device *);
@@ -1040,8 +1038,7 @@ extern int  nv50_instmem_populate(struct drm_device *, struct nouveau_gpuobj *,
 extern void nv50_instmem_clear(struct drm_device *, struct nouveau_gpuobj *);
 extern int  nv50_instmem_bind(struct drm_device *, struct nouveau_gpuobj *);
 extern int  nv50_instmem_unbind(struct drm_device *, struct nouveau_gpuobj *);
-extern void nv50_instmem_prepare_access(struct drm_device *, bool write);
-extern void nv50_instmem_finish_access(struct drm_device *);
+extern void nv50_instmem_flush(struct drm_device *);
 
 /* nv04_mc.c */
 extern int  nv04_mc_init(struct drm_device *);
index 4b42bf218f61d31b2a023d2600b8a6e087f633c6..5152c0a7e6f70b4c4ce17baca9faf831893ee77f 100644 (file)
@@ -143,7 +143,6 @@ nv50_mem_vm_bind_linear(struct drm_device *dev, uint64_t virt, uint32_t size,
                phys |= 0x30;
        }
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
        while (size) {
                unsigned offset_h = upper_32_bits(phys);
                unsigned offset_l = lower_32_bits(phys);
@@ -175,7 +174,7 @@ nv50_mem_vm_bind_linear(struct drm_device *dev, uint64_t virt, uint32_t size,
                        }
                }
        }
-       dev_priv->engine.instmem.finish_access(dev);
+       dev_priv->engine.instmem.flush(dev);
 
        nv_wr32(dev, 0x100c80, 0x00050001);
        if (!nv_wait(0x100c80, 0x00000001, 0x00000000)) {
@@ -218,7 +217,6 @@ nv50_mem_vm_unbind(struct drm_device *dev, uint64_t virt, uint32_t size)
        virt -= dev_priv->vm_vram_base;
        pages = (size >> 16) << 1;
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
        while (pages) {
                pgt = dev_priv->vm_vram_pt[virt >> 29];
                pte = (virt & 0x1ffe0000ULL) >> 15;
@@ -232,7 +230,7 @@ nv50_mem_vm_unbind(struct drm_device *dev, uint64_t virt, uint32_t size)
                while (pte < end)
                        nv_wo32(dev, pgt, pte++, 0);
        }
-       dev_priv->engine.instmem.finish_access(dev);
+       dev_priv->engine.instmem.flush(dev);
 
        nv_wr32(dev, 0x100c80, 0x00050001);
        if (!nv_wait(0x100c80, 0x00000001, 0x00000000)) {
index 5624f37e4c75bc1fcfd0b6eb543535b675605847..7d86e05ac883d7cb943dcb4dc2337bd42d7198b8 100644 (file)
@@ -132,7 +132,6 @@ nouveau_ramht_insert(struct drm_device *dev, struct nouveau_gpuobj_ref *ref)
                }
        }
 
-       instmem->prepare_access(dev, true);
        co = ho = nouveau_ramht_hash_handle(dev, chan->id, ref->handle);
        do {
                if (!nouveau_ramht_entry_valid(dev, ramht, co)) {
@@ -143,7 +142,7 @@ nouveau_ramht_insert(struct drm_device *dev, struct nouveau_gpuobj_ref *ref)
                        nv_wo32(dev, ramht, (co + 4)/4, ctx);
 
                        list_add_tail(&ref->list, &chan->ramht_refs);
-                       instmem->finish_access(dev);
+                       instmem->flush(dev);
                        return 0;
                }
                NV_DEBUG(dev, "collision ch%d 0x%08x: h=0x%08x\n",
@@ -153,7 +152,6 @@ nouveau_ramht_insert(struct drm_device *dev, struct nouveau_gpuobj_ref *ref)
                if (co >= dev_priv->ramht_size)
                        co = 0;
        } while (co != ho);
-       instmem->finish_access(dev);
 
        NV_ERROR(dev, "RAMHT space exhausted. ch=%d\n", chan->id);
        return -ENOMEM;
@@ -173,7 +171,6 @@ nouveau_ramht_remove(struct drm_device *dev, struct nouveau_gpuobj_ref *ref)
                return;
        }
 
-       instmem->prepare_access(dev, true);
        co = ho = nouveau_ramht_hash_handle(dev, chan->id, ref->handle);
        do {
                if (nouveau_ramht_entry_valid(dev, ramht, co) &&
@@ -186,7 +183,7 @@ nouveau_ramht_remove(struct drm_device *dev, struct nouveau_gpuobj_ref *ref)
                        nv_wo32(dev, ramht, (co + 4)/4, 0x00000000);
 
                        list_del(&ref->list);
-                       instmem->finish_access(dev);
+                       instmem->flush(dev);
                        return;
                }
 
@@ -195,7 +192,6 @@ nouveau_ramht_remove(struct drm_device *dev, struct nouveau_gpuobj_ref *ref)
                        co = 0;
        } while (co != ho);
        list_del(&ref->list);
-       instmem->finish_access(dev);
 
        NV_ERROR(dev, "RAMHT entry not found. ch=%d, handle=0x%08x\n",
                 chan->id, ref->handle);
@@ -280,10 +276,9 @@ nouveau_gpuobj_new(struct drm_device *dev, struct nouveau_channel *chan,
        if (gpuobj->flags & NVOBJ_FLAG_ZERO_ALLOC) {
                int i;
 
-               engine->instmem.prepare_access(dev, true);
                for (i = 0; i < gpuobj->im_pramin->size; i += 4)
                        nv_wo32(dev, gpuobj, i/4, 0);
-               engine->instmem.finish_access(dev);
+               engine->instmem.flush(dev);
        }
 
        *gpuobj_ret = gpuobj;
@@ -371,10 +366,9 @@ nouveau_gpuobj_del(struct drm_device *dev, struct nouveau_gpuobj **pgpuobj)
        }
 
        if (gpuobj->im_pramin && (gpuobj->flags & NVOBJ_FLAG_ZERO_FREE)) {
-               engine->instmem.prepare_access(dev, true);
                for (i = 0; i < gpuobj->im_pramin->size; i += 4)
                        nv_wo32(dev, gpuobj, i/4, 0);
-               engine->instmem.finish_access(dev);
+               engine->instmem.flush(dev);
        }
 
        if (gpuobj->dtor)
@@ -606,10 +600,9 @@ nouveau_gpuobj_new_fake(struct drm_device *dev, uint32_t p_offset,
        }
 
        if (gpuobj->flags & NVOBJ_FLAG_ZERO_ALLOC) {
-               dev_priv->engine.instmem.prepare_access(dev, true);
                for (i = 0; i < gpuobj->im_pramin->size; i += 4)
                        nv_wo32(dev, gpuobj, i/4, 0);
-               dev_priv->engine.instmem.finish_access(dev);
+               dev_priv->engine.instmem.flush(dev);
        }
 
        if (pref) {
@@ -697,8 +690,6 @@ nouveau_gpuobj_dma_new(struct nouveau_channel *chan, int class,
                return ret;
        }
 
-       instmem->prepare_access(dev, true);
-
        if (dev_priv->card_type < NV_50) {
                uint32_t frame, adjust, pte_flags = 0;
 
@@ -735,7 +726,7 @@ nouveau_gpuobj_dma_new(struct nouveau_channel *chan, int class,
                nv_wo32(dev, *gpuobj, 5, flags5);
        }
 
-       instmem->finish_access(dev);
+       instmem->flush(dev);
 
        (*gpuobj)->engine = NVOBJ_ENGINE_SW;
        (*gpuobj)->class  = class;
@@ -850,7 +841,6 @@ nouveau_gpuobj_gr_new(struct nouveau_channel *chan, int class,
                return ret;
        }
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
        if (dev_priv->card_type >= NV_50) {
                nv_wo32(dev, *gpuobj, 0, class);
                nv_wo32(dev, *gpuobj, 5, 0x00010000);
@@ -875,7 +865,7 @@ nouveau_gpuobj_gr_new(struct nouveau_channel *chan, int class,
                        }
                }
        }
-       dev_priv->engine.instmem.finish_access(dev);
+       dev_priv->engine.instmem.flush(dev);
 
        (*gpuobj)->engine = NVOBJ_ENGINE_GR;
        (*gpuobj)->class  = class;
@@ -988,17 +978,13 @@ nouveau_gpuobj_channel_init(struct nouveau_channel *chan,
        if (dev_priv->card_type >= NV_50) {
                uint32_t vm_offset, pde;
 
-               instmem->prepare_access(dev, true);
-
                vm_offset = (dev_priv->chipset & 0xf0) == 0x50 ? 0x1400 : 0x200;
                vm_offset += chan->ramin->gpuobj->im_pramin->start;
 
                ret = nouveau_gpuobj_new_fake(dev, vm_offset, ~0, 0x4000,
                                                        0, &chan->vm_pd, NULL);
-               if (ret) {
-                       instmem->finish_access(dev);
+               if (ret)
                        return ret;
-               }
                for (i = 0; i < 0x4000; i += 8) {
                        nv_wo32(dev, chan->vm_pd, (i+0)/4, 0x00000000);
                        nv_wo32(dev, chan->vm_pd, (i+4)/4, 0xdeadcafe);
@@ -1008,10 +994,8 @@ nouveau_gpuobj_channel_init(struct nouveau_channel *chan,
                ret = nouveau_gpuobj_ref_add(dev, NULL, 0,
                                             dev_priv->gart_info.sg_ctxdma,
                                             &chan->vm_gart_pt);
-               if (ret) {
-                       instmem->finish_access(dev);
+               if (ret)
                        return ret;
-               }
                nv_wo32(dev, chan->vm_pd, pde++,
                            chan->vm_gart_pt->instance | 0x03);
                nv_wo32(dev, chan->vm_pd, pde++, 0x00000000);
@@ -1021,17 +1005,15 @@ nouveau_gpuobj_channel_init(struct nouveau_channel *chan,
                        ret = nouveau_gpuobj_ref_add(dev, NULL, 0,
                                                     dev_priv->vm_vram_pt[i],
                                                     &chan->vm_vram_pt[i]);
-                       if (ret) {
-                               instmem->finish_access(dev);
+                       if (ret)
                                return ret;
-                       }
 
                        nv_wo32(dev, chan->vm_pd, pde++,
                                    chan->vm_vram_pt[i]->instance | 0x61);
                        nv_wo32(dev, chan->vm_pd, pde++, 0x00000000);
                }
 
-               instmem->finish_access(dev);
+               instmem->flush(dev);
        }
 
        /* RAMHT */
@@ -1164,10 +1146,8 @@ nouveau_gpuobj_suspend(struct drm_device *dev)
                        return -ENOMEM;
                }
 
-               dev_priv->engine.instmem.prepare_access(dev, false);
                for (i = 0; i < gpuobj->im_pramin->size / 4; i++)
                        gpuobj->im_backing_suspend[i] = nv_ro32(dev, gpuobj, i);
-               dev_priv->engine.instmem.finish_access(dev);
        }
 
        return 0;
@@ -1212,10 +1192,9 @@ nouveau_gpuobj_resume(struct drm_device *dev)
                if (!gpuobj->im_backing_suspend)
                        continue;
 
-               dev_priv->engine.instmem.prepare_access(dev, true);
                for (i = 0; i < gpuobj->im_pramin->size / 4; i++)
                        nv_wo32(dev, gpuobj, i, gpuobj->im_backing_suspend[i]);
-               dev_priv->engine.instmem.finish_access(dev);
+               dev_priv->engine.instmem.flush(dev);
        }
 
        nouveau_gpuobj_suspend_cleanup(dev);
index 1d6ee8b55154c95aa763ed278c31b14da32fa9a7..1b2ab5a714ceb5e4877e33c4c777a4c161a9c97f 100644 (file)
@@ -97,7 +97,6 @@ nouveau_sgdma_bind(struct ttm_backend *be, struct ttm_mem_reg *mem)
 
        NV_DEBUG(dev, "pg=0x%lx\n", mem->mm_node->start);
 
-       dev_priv->engine.instmem.prepare_access(nvbe->dev, true);
        pte = nouveau_sgdma_pte(nvbe->dev, mem->mm_node->start << PAGE_SHIFT);
        nvbe->pte_start = pte;
        for (i = 0; i < nvbe->nr_pages; i++) {
@@ -116,7 +115,7 @@ nouveau_sgdma_bind(struct ttm_backend *be, struct ttm_mem_reg *mem)
                        dma_offset += NV_CTXDMA_PAGE_SIZE;
                }
        }
-       dev_priv->engine.instmem.finish_access(nvbe->dev);
+       dev_priv->engine.instmem.flush(nvbe->dev);
 
        if (dev_priv->card_type == NV_50) {
                nv_wr32(dev, 0x100c80, 0x00050001);
@@ -154,7 +153,6 @@ nouveau_sgdma_unbind(struct ttm_backend *be)
        if (!nvbe->bound)
                return 0;
 
-       dev_priv->engine.instmem.prepare_access(nvbe->dev, true);
        pte = nvbe->pte_start;
        for (i = 0; i < nvbe->nr_pages; i++) {
                dma_addr_t dma_offset = dev_priv->gart_info.sg_dummy_bus;
@@ -170,7 +168,7 @@ nouveau_sgdma_unbind(struct ttm_backend *be)
                        dma_offset += NV_CTXDMA_PAGE_SIZE;
                }
        }
-       dev_priv->engine.instmem.finish_access(nvbe->dev);
+       dev_priv->engine.instmem.flush(nvbe->dev);
 
        if (dev_priv->card_type == NV_50) {
                nv_wr32(dev, 0x100c80, 0x00050001);
@@ -272,7 +270,6 @@ nouveau_sgdma_init(struct drm_device *dev)
                pci_map_page(dev->pdev, dev_priv->gart_info.sg_dummy_page, 0,
                             PAGE_SIZE, PCI_DMA_BIDIRECTIONAL);
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
        if (dev_priv->card_type < NV_50) {
                /* Maybe use NV_DMA_TARGET_AGP for PCIE? NVIDIA do this, and
                 * confirmed to work on c51.  Perhaps means NV_DMA_TARGET_PCIE
@@ -294,7 +291,7 @@ nouveau_sgdma_init(struct drm_device *dev)
                        nv_wo32(dev, gpuobj, (i+4)/4, 0);
                }
        }
-       dev_priv->engine.instmem.finish_access(dev);
+       dev_priv->engine.instmem.flush(dev);
 
        dev_priv->gart_info.type      = NOUVEAU_GART_SGDMA;
        dev_priv->gart_info.aper_base = 0;
@@ -325,14 +322,11 @@ nouveau_sgdma_get_page(struct drm_device *dev, uint32_t offset, uint32_t *page)
 {
        struct drm_nouveau_private *dev_priv = dev->dev_private;
        struct nouveau_gpuobj *gpuobj = dev_priv->gart_info.sg_ctxdma;
-       struct nouveau_instmem_engine *instmem = &dev_priv->engine.instmem;
        int pte;
 
        pte = (offset >> NV_CTXDMA_PAGE_SHIFT);
        if (dev_priv->card_type < NV_50) {
-               instmem->prepare_access(dev, false);
                *page = nv_ro32(dev, gpuobj, (pte + 2)) & ~NV_CTXDMA_PAGE_MASK;
-               instmem->finish_access(dev);
                return 0;
        }
 
index 6fd99f10eed647ea63251508f4d3553875c17b17..67ee32fbba9d48a43fe5f6889871ab22424c1fd6 100644 (file)
@@ -54,8 +54,7 @@ static int nouveau_init_engine_ptrs(struct drm_device *dev)
                engine->instmem.clear           = nv04_instmem_clear;
                engine->instmem.bind            = nv04_instmem_bind;
                engine->instmem.unbind          = nv04_instmem_unbind;
-               engine->instmem.prepare_access  = nv04_instmem_prepare_access;
-               engine->instmem.finish_access   = nv04_instmem_finish_access;
+               engine->instmem.flush           = nv04_instmem_flush;
                engine->mc.init                 = nv04_mc_init;
                engine->mc.takedown             = nv04_mc_takedown;
                engine->timer.init              = nv04_timer_init;
@@ -95,8 +94,7 @@ static int nouveau_init_engine_ptrs(struct drm_device *dev)
                engine->instmem.clear           = nv04_instmem_clear;
                engine->instmem.bind            = nv04_instmem_bind;
                engine->instmem.unbind          = nv04_instmem_unbind;
-               engine->instmem.prepare_access  = nv04_instmem_prepare_access;
-               engine->instmem.finish_access   = nv04_instmem_finish_access;
+               engine->instmem.flush           = nv04_instmem_flush;
                engine->mc.init                 = nv04_mc_init;
                engine->mc.takedown             = nv04_mc_takedown;
                engine->timer.init              = nv04_timer_init;
@@ -138,8 +136,7 @@ static int nouveau_init_engine_ptrs(struct drm_device *dev)
                engine->instmem.clear           = nv04_instmem_clear;
                engine->instmem.bind            = nv04_instmem_bind;
                engine->instmem.unbind          = nv04_instmem_unbind;
-               engine->instmem.prepare_access  = nv04_instmem_prepare_access;
-               engine->instmem.finish_access   = nv04_instmem_finish_access;
+               engine->instmem.flush           = nv04_instmem_flush;
                engine->mc.init                 = nv04_mc_init;
                engine->mc.takedown             = nv04_mc_takedown;
                engine->timer.init              = nv04_timer_init;
@@ -181,8 +178,7 @@ static int nouveau_init_engine_ptrs(struct drm_device *dev)
                engine->instmem.clear           = nv04_instmem_clear;
                engine->instmem.bind            = nv04_instmem_bind;
                engine->instmem.unbind          = nv04_instmem_unbind;
-               engine->instmem.prepare_access  = nv04_instmem_prepare_access;
-               engine->instmem.finish_access   = nv04_instmem_finish_access;
+               engine->instmem.flush           = nv04_instmem_flush;
                engine->mc.init                 = nv04_mc_init;
                engine->mc.takedown             = nv04_mc_takedown;
                engine->timer.init              = nv04_timer_init;
@@ -225,8 +221,7 @@ static int nouveau_init_engine_ptrs(struct drm_device *dev)
                engine->instmem.clear           = nv04_instmem_clear;
                engine->instmem.bind            = nv04_instmem_bind;
                engine->instmem.unbind          = nv04_instmem_unbind;
-               engine->instmem.prepare_access  = nv04_instmem_prepare_access;
-               engine->instmem.finish_access   = nv04_instmem_finish_access;
+               engine->instmem.flush           = nv04_instmem_flush;
                engine->mc.init                 = nv40_mc_init;
                engine->mc.takedown             = nv40_mc_takedown;
                engine->timer.init              = nv04_timer_init;
@@ -271,8 +266,7 @@ static int nouveau_init_engine_ptrs(struct drm_device *dev)
                engine->instmem.clear           = nv50_instmem_clear;
                engine->instmem.bind            = nv50_instmem_bind;
                engine->instmem.unbind          = nv50_instmem_unbind;
-               engine->instmem.prepare_access  = nv50_instmem_prepare_access;
-               engine->instmem.finish_access   = nv50_instmem_finish_access;
+               engine->instmem.flush           = nv50_instmem_flush;
                engine->mc.init                 = nv50_mc_init;
                engine->mc.takedown             = nv50_mc_takedown;
                engine->timer.init              = nv04_timer_init;
index 611c83e6d9f466c4d23ff683ada472d13c30e8b9..b2c01fe899e9795fb8f6a8d765d8a7ad8d1f3e9c 100644 (file)
@@ -137,7 +137,6 @@ nv04_fifo_create_context(struct nouveau_channel *chan)
        spin_lock_irqsave(&dev_priv->context_switch_lock, flags);
 
        /* Setup initial state */
-       dev_priv->engine.instmem.prepare_access(dev, true);
        RAMFC_WR(DMA_PUT, chan->pushbuf_base);
        RAMFC_WR(DMA_GET, chan->pushbuf_base);
        RAMFC_WR(DMA_INSTANCE, chan->pushbuf->instance >> 4);
@@ -145,7 +144,6 @@ nv04_fifo_create_context(struct nouveau_channel *chan)
                             NV_PFIFO_CACHE1_DMA_FETCH_SIZE_128_BYTES |
                             NV_PFIFO_CACHE1_DMA_FETCH_MAX_REQS_8 |
                             DMA_FETCH_ENDIANNESS));
-       dev_priv->engine.instmem.finish_access(dev);
 
        /* enable the fifo dma operation */
        nv_wr32(dev, NV04_PFIFO_MODE,
@@ -172,8 +170,6 @@ nv04_fifo_do_load_context(struct drm_device *dev, int chid)
        struct drm_nouveau_private *dev_priv = dev->dev_private;
        uint32_t fc = NV04_RAMFC(chid), tmp;
 
-       dev_priv->engine.instmem.prepare_access(dev, false);
-
        nv_wr32(dev, NV04_PFIFO_CACHE1_DMA_PUT, nv_ri32(dev, fc + 0));
        nv_wr32(dev, NV04_PFIFO_CACHE1_DMA_GET, nv_ri32(dev, fc + 4));
        tmp = nv_ri32(dev, fc + 8);
@@ -184,8 +180,6 @@ nv04_fifo_do_load_context(struct drm_device *dev, int chid)
        nv_wr32(dev, NV04_PFIFO_CACHE1_ENGINE, nv_ri32(dev, fc + 20));
        nv_wr32(dev, NV04_PFIFO_CACHE1_PULL1, nv_ri32(dev, fc + 24));
 
-       dev_priv->engine.instmem.finish_access(dev);
-
        nv_wr32(dev, NV03_PFIFO_CACHE1_GET, 0);
        nv_wr32(dev, NV03_PFIFO_CACHE1_PUT, 0);
 }
@@ -226,7 +220,6 @@ nv04_fifo_unload_context(struct drm_device *dev)
                return -EINVAL;
        }
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
        RAMFC_WR(DMA_PUT, nv_rd32(dev, NV04_PFIFO_CACHE1_DMA_PUT));
        RAMFC_WR(DMA_GET, nv_rd32(dev, NV04_PFIFO_CACHE1_DMA_GET));
        tmp  = nv_rd32(dev, NV04_PFIFO_CACHE1_DMA_DCOUNT) << 16;
@@ -236,7 +229,6 @@ nv04_fifo_unload_context(struct drm_device *dev)
        RAMFC_WR(DMA_FETCH, nv_rd32(dev, NV04_PFIFO_CACHE1_DMA_FETCH));
        RAMFC_WR(ENGINE, nv_rd32(dev, NV04_PFIFO_CACHE1_ENGINE));
        RAMFC_WR(PULL1_ENGINE, nv_rd32(dev, NV04_PFIFO_CACHE1_PULL1));
-       dev_priv->engine.instmem.finish_access(dev);
 
        nv04_fifo_do_load_context(dev, pfifo->channels - 1);
        nv_wr32(dev, NV03_PFIFO_CACHE1_PUSH1, pfifo->channels - 1);
index 17af702d6ddcbc434c1c823f3907356c29c2a358..4408232d33f179805205763974beede957ddd6f1 100644 (file)
@@ -49,10 +49,8 @@ nv04_instmem_determine_amount(struct drm_device *dev)
        NV_DEBUG(dev, "RAMIN size: %dKiB\n", dev_priv->ramin_rsvd_vram >> 10);
 
        /* Clear all of it, except the BIOS image that's in the first 64KiB */
-       dev_priv->engine.instmem.prepare_access(dev, true);
        for (i = 64 * 1024; i < dev_priv->ramin_rsvd_vram; i += 4)
                nv_wi32(dev, i, 0x00000000);
-       dev_priv->engine.instmem.finish_access(dev);
 }
 
 static void
@@ -186,12 +184,7 @@ nv04_instmem_unbind(struct drm_device *dev, struct nouveau_gpuobj *gpuobj)
 }
 
 void
-nv04_instmem_prepare_access(struct drm_device *dev, bool write)
-{
-}
-
-void
-nv04_instmem_finish_access(struct drm_device *dev)
+nv04_instmem_flush(struct drm_device *dev)
 {
 }
 
index 7aeabf262bc0a43affe3c1fdad393e9301d95263..7a4069cf5d0b835caccd983722bb640df7f6e351 100644 (file)
@@ -55,7 +55,6 @@ nv10_fifo_create_context(struct nouveau_channel *chan)
        /* Fill entries that are seen filled in dumps of nvidia driver just
         * after channel's is put into DMA mode
         */
-       dev_priv->engine.instmem.prepare_access(dev, true);
        nv_wi32(dev, fc +  0, chan->pushbuf_base);
        nv_wi32(dev, fc +  4, chan->pushbuf_base);
        nv_wi32(dev, fc + 12, chan->pushbuf->instance >> 4);
@@ -66,7 +65,6 @@ nv10_fifo_create_context(struct nouveau_channel *chan)
                              NV_PFIFO_CACHE1_BIG_ENDIAN |
 #endif
                              0);
-       dev_priv->engine.instmem.finish_access(dev);
 
        /* enable the fifo dma operation */
        nv_wr32(dev, NV04_PFIFO_MODE,
@@ -91,8 +89,6 @@ nv10_fifo_do_load_context(struct drm_device *dev, int chid)
        struct drm_nouveau_private *dev_priv = dev->dev_private;
        uint32_t fc = NV10_RAMFC(chid), tmp;
 
-       dev_priv->engine.instmem.prepare_access(dev, false);
-
        nv_wr32(dev, NV04_PFIFO_CACHE1_DMA_PUT, nv_ri32(dev, fc + 0));
        nv_wr32(dev, NV04_PFIFO_CACHE1_DMA_GET, nv_ri32(dev, fc + 4));
        nv_wr32(dev, NV10_PFIFO_CACHE1_REF_CNT, nv_ri32(dev, fc + 8));
@@ -117,8 +113,6 @@ nv10_fifo_do_load_context(struct drm_device *dev, int chid)
        nv_wr32(dev, NV10_PFIFO_CACHE1_DMA_SUBROUTINE, nv_ri32(dev, fc + 48));
 
 out:
-       dev_priv->engine.instmem.finish_access(dev);
-
        nv_wr32(dev, NV03_PFIFO_CACHE1_GET, 0);
        nv_wr32(dev, NV03_PFIFO_CACHE1_PUT, 0);
 }
@@ -155,8 +149,6 @@ nv10_fifo_unload_context(struct drm_device *dev)
                return 0;
        fc = NV10_RAMFC(chid);
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
-
        nv_wi32(dev, fc +  0, nv_rd32(dev, NV04_PFIFO_CACHE1_DMA_PUT));
        nv_wi32(dev, fc +  4, nv_rd32(dev, NV04_PFIFO_CACHE1_DMA_GET));
        nv_wi32(dev, fc +  8, nv_rd32(dev, NV10_PFIFO_CACHE1_REF_CNT));
@@ -179,8 +171,6 @@ nv10_fifo_unload_context(struct drm_device *dev)
        nv_wi32(dev, fc + 48, nv_rd32(dev, NV04_PFIFO_CACHE1_DMA_GET));
 
 out:
-       dev_priv->engine.instmem.finish_access(dev);
-
        nv10_fifo_do_load_context(dev, pfifo->channels - 1);
        nv_wr32(dev, NV03_PFIFO_CACHE1_PUSH1, pfifo->channels - 1);
        return 0;
index fe2349b115f0f3f8906c42baaae6774a7a8270ce..f3e6dd70d22ef571a4fb185efc2c29ea8ebacff1 100644 (file)
@@ -421,7 +421,6 @@ nv20_graph_create_context(struct nouveau_channel *chan)
                return ret;
 
        /* Initialise default context values */
-       dev_priv->engine.instmem.prepare_access(dev, true);
        ctx_init(dev, chan->ramin_grctx->gpuobj);
 
        /* nv20: nv_wo32(dev, chan->ramin_grctx->gpuobj, 10, chan->id<<24); */
@@ -430,8 +429,6 @@ nv20_graph_create_context(struct nouveau_channel *chan)
 
        nv_wo32(dev, dev_priv->ctx_table->gpuobj, chan->id,
                        chan->ramin_grctx->instance >> 4);
-
-       dev_priv->engine.instmem.finish_access(dev);
        return 0;
 }
 
@@ -444,9 +441,7 @@ nv20_graph_destroy_context(struct nouveau_channel *chan)
        if (chan->ramin_grctx)
                nouveau_gpuobj_ref_del(dev, &chan->ramin_grctx);
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
        nv_wo32(dev, dev_priv->ctx_table->gpuobj, chan->id, 0);
-       dev_priv->engine.instmem.finish_access(dev);
 }
 
 int
index 500ccfd3a0b8c664b62cdc308ebaff0cc95d3042..2b67f1835c3942c4fddbe63b2b84a9050df81134 100644 (file)
@@ -48,7 +48,6 @@ nv40_fifo_create_context(struct nouveau_channel *chan)
 
        spin_lock_irqsave(&dev_priv->context_switch_lock, flags);
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
        nv_wi32(dev, fc +  0, chan->pushbuf_base);
        nv_wi32(dev, fc +  4, chan->pushbuf_base);
        nv_wi32(dev, fc + 12, chan->pushbuf->instance >> 4);
@@ -61,7 +60,6 @@ nv40_fifo_create_context(struct nouveau_channel *chan)
                              0x30000000 /* no idea.. */);
        nv_wi32(dev, fc + 56, chan->ramin_grctx->instance >> 4);
        nv_wi32(dev, fc + 60, 0x0001FFFF);
-       dev_priv->engine.instmem.finish_access(dev);
 
        /* enable the fifo dma operation */
        nv_wr32(dev, NV04_PFIFO_MODE,
@@ -89,8 +87,6 @@ nv40_fifo_do_load_context(struct drm_device *dev, int chid)
        struct drm_nouveau_private *dev_priv = dev->dev_private;
        uint32_t fc = NV40_RAMFC(chid), tmp, tmp2;
 
-       dev_priv->engine.instmem.prepare_access(dev, false);
-
        nv_wr32(dev, NV04_PFIFO_CACHE1_DMA_PUT, nv_ri32(dev, fc + 0));
        nv_wr32(dev, NV04_PFIFO_CACHE1_DMA_GET, nv_ri32(dev, fc + 4));
        nv_wr32(dev, NV10_PFIFO_CACHE1_REF_CNT, nv_ri32(dev, fc + 8));
@@ -127,8 +123,6 @@ nv40_fifo_do_load_context(struct drm_device *dev, int chid)
        nv_wr32(dev, 0x2088, nv_ri32(dev, fc + 76));
        nv_wr32(dev, 0x3300, nv_ri32(dev, fc + 80));
 
-       dev_priv->engine.instmem.finish_access(dev);
-
        nv_wr32(dev, NV03_PFIFO_CACHE1_GET, 0);
        nv_wr32(dev, NV03_PFIFO_CACHE1_PUT, 0);
 }
@@ -166,7 +160,6 @@ nv40_fifo_unload_context(struct drm_device *dev)
                return 0;
        fc = NV40_RAMFC(chid);
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
        nv_wi32(dev, fc + 0, nv_rd32(dev, NV04_PFIFO_CACHE1_DMA_PUT));
        nv_wi32(dev, fc + 4, nv_rd32(dev, NV04_PFIFO_CACHE1_DMA_GET));
        nv_wi32(dev, fc + 8, nv_rd32(dev, NV10_PFIFO_CACHE1_REF_CNT));
@@ -200,7 +193,6 @@ nv40_fifo_unload_context(struct drm_device *dev)
        tmp |= (nv_rd32(dev, NV04_PFIFO_CACHE1_PUT) << 16);
        nv_wi32(dev, fc + 72, tmp);
 #endif
-       dev_priv->engine.instmem.finish_access(dev);
 
        nv40_fifo_do_load_context(dev, pfifo->channels - 1);
        nv_wr32(dev, NV03_PFIFO_CACHE1_PUSH1,
index 65b13b54c5ae0e478fc3dac2aa0deb2274775b4b..2608c34eca8279953bdaeab8c49f4a43b9e9c7c8 100644 (file)
@@ -67,7 +67,6 @@ nv40_graph_create_context(struct nouveau_channel *chan)
                return ret;
 
        /* Initialise default context values */
-       dev_priv->engine.instmem.prepare_access(dev, true);
        if (!pgraph->ctxprog) {
                struct nouveau_grctx ctx = {};
 
@@ -80,7 +79,6 @@ nv40_graph_create_context(struct nouveau_channel *chan)
        }
        nv_wo32(dev, chan->ramin_grctx->gpuobj, 0,
                     chan->ramin_grctx->gpuobj->im_pramin->start);
-       dev_priv->engine.instmem.finish_access(dev);
        return 0;
 }
 
index 711128c42de88518d8b90f430bc0cb4b2645b3c0..6a293c818b61dcf4683fbcecb550b90865a003fe 100644 (file)
@@ -71,14 +71,13 @@ nv50_evo_dmaobj_new(struct nouveau_channel *evo, uint32_t class, uint32_t name,
                return ret;
        }
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
        nv_wo32(dev, obj, 0, (tile_flags << 22) | (magic_flags << 16) | class);
        nv_wo32(dev, obj, 1, limit);
        nv_wo32(dev, obj, 2, offset);
        nv_wo32(dev, obj, 3, 0x00000000);
        nv_wo32(dev, obj, 4, 0x00000000);
        nv_wo32(dev, obj, 5, 0x00010000);
-       dev_priv->engine.instmem.finish_access(dev);
+       dev_priv->engine.instmem.flush(dev);
 
        return 0;
 }
index e20c0e2474f3b56035b3c6391b86a17e4c693f0f..d2d4fd0044f86e027b0a938a074383bd5eebcb6d 100644 (file)
@@ -49,12 +49,11 @@ nv50_fifo_init_thingo(struct drm_device *dev)
        priv->cur_thingo = !priv->cur_thingo;
 
        /* We never schedule channel 0 or 127 */
-       dev_priv->engine.instmem.prepare_access(dev, true);
        for (i = 1, nr = 0; i < 127; i++) {
                if (dev_priv->fifos[i] && dev_priv->fifos[i]->ramfc)
                        nv_wo32(dev, cur->gpuobj, nr++, i);
        }
-       dev_priv->engine.instmem.finish_access(dev);
+       dev_priv->engine.instmem.flush(dev);
 
        nv_wr32(dev, 0x32f4, cur->instance >> 12);
        nv_wr32(dev, 0x32ec, nr);
@@ -281,8 +280,6 @@ nv50_fifo_create_context(struct nouveau_channel *chan)
 
        spin_lock_irqsave(&dev_priv->context_switch_lock, flags);
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
-
        nv_wo32(dev, ramfc, 0x48/4, chan->pushbuf->instance >> 4);
        nv_wo32(dev, ramfc, 0x80/4, (0xc << 24) | (chan->ramht->instance >> 4));
        nv_wo32(dev, ramfc, 0x44/4, 0x2101ffff);
@@ -304,7 +301,7 @@ nv50_fifo_create_context(struct nouveau_channel *chan)
                nv_wo32(dev, ramfc, 0x98/4, chan->ramin->instance >> 12);
        }
 
-       dev_priv->engine.instmem.finish_access(dev);
+       dev_priv->engine.instmem.flush(dev);
 
        ret = nv50_fifo_channel_enable(dev, chan->id, false);
        if (ret) {
@@ -349,8 +346,6 @@ nv50_fifo_load_context(struct nouveau_channel *chan)
 
        NV_DEBUG(dev, "ch%d\n", chan->id);
 
-       dev_priv->engine.instmem.prepare_access(dev, false);
-
        nv_wr32(dev, 0x3330, nv_ro32(dev, ramfc, 0x00/4));
        nv_wr32(dev, 0x3334, nv_ro32(dev, ramfc, 0x04/4));
        nv_wr32(dev, 0x3240, nv_ro32(dev, ramfc, 0x08/4));
@@ -404,8 +399,6 @@ nv50_fifo_load_context(struct nouveau_channel *chan)
                nv_wr32(dev, 0x3410, nv_ro32(dev, ramfc, 0x98/4));
        }
 
-       dev_priv->engine.instmem.finish_access(dev);
-
        nv_wr32(dev, NV03_PFIFO_CACHE1_PUSH1, chan->id | (1<<16));
        return 0;
 }
@@ -434,8 +427,6 @@ nv50_fifo_unload_context(struct drm_device *dev)
        ramfc = chan->ramfc->gpuobj;
        cache = chan->cache->gpuobj;
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
-
        nv_wo32(dev, ramfc, 0x00/4, nv_rd32(dev, 0x3330));
        nv_wo32(dev, ramfc, 0x04/4, nv_rd32(dev, 0x3334));
        nv_wo32(dev, ramfc, 0x08/4, nv_rd32(dev, 0x3240));
@@ -491,7 +482,7 @@ nv50_fifo_unload_context(struct drm_device *dev)
                nv_wo32(dev, ramfc, 0x98/4, nv_rd32(dev, 0x3410));
        }
 
-       dev_priv->engine.instmem.finish_access(dev);
+       dev_priv->engine.instmem.flush(dev);
 
        /*XXX: probably reload ch127 (NULL) state back too */
        nv_wr32(dev, NV03_PFIFO_CACHE1_PUSH1, 127);
index b04e7c8449a551f794b5eb92a0f6650028d2f58d..5dc3be5696a961f37d67e327e886ef25b4c7ec36 100644 (file)
@@ -226,7 +226,6 @@ nv50_graph_create_context(struct nouveau_channel *chan)
        obj = chan->ramin_grctx->gpuobj;
 
        hdr = IS_G80 ? 0x200 : 0x20;
-       dev_priv->engine.instmem.prepare_access(dev, true);
        nv_wo32(dev, ramin, (hdr + 0x00)/4, 0x00190002);
        nv_wo32(dev, ramin, (hdr + 0x04)/4, chan->ramin_grctx->instance +
                                           pgraph->grctx_size - 1);
@@ -234,9 +233,7 @@ nv50_graph_create_context(struct nouveau_channel *chan)
        nv_wo32(dev, ramin, (hdr + 0x0c)/4, 0);
        nv_wo32(dev, ramin, (hdr + 0x10)/4, 0);
        nv_wo32(dev, ramin, (hdr + 0x14)/4, 0x00010000);
-       dev_priv->engine.instmem.finish_access(dev);
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
        if (!pgraph->ctxprog) {
                struct nouveau_grctx ctx = {};
                ctx.dev = chan->dev;
@@ -247,8 +244,8 @@ nv50_graph_create_context(struct nouveau_channel *chan)
                nouveau_grctx_vals_load(dev, obj);
        }
        nv_wo32(dev, obj, 0x00000/4, chan->ramin->instance >> 12);
-       dev_priv->engine.instmem.finish_access(dev);
 
+       dev_priv->engine.instmem.flush(dev);
        return 0;
 }
 
@@ -264,10 +261,9 @@ nv50_graph_destroy_context(struct nouveau_channel *chan)
        if (!chan->ramin || !chan->ramin->gpuobj)
                return;
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
        for (i = hdr; i < hdr + 24; i += 4)
                nv_wo32(dev, chan->ramin->gpuobj, i/4, 0);
-       dev_priv->engine.instmem.finish_access(dev);
+       dev_priv->engine.instmem.flush(dev);
 
        nouveau_gpuobj_ref_del(dev, &chan->ramin_grctx);
 }
index a361d1612bd74162fdfa4686163b8e41cd221ac4..d9feee3b9f58a11cbc4bf6196066bb098794238f 100644 (file)
@@ -35,8 +35,6 @@ struct nv50_instmem_priv {
        struct nouveau_gpuobj_ref *pramin_pt;
        struct nouveau_gpuobj_ref *pramin_bar;
        struct nouveau_gpuobj_ref *fb_bar;
-
-       bool last_access_wr;
 };
 
 #define NV50_INSTMEM_PAGE_SHIFT 12
@@ -262,16 +260,13 @@ nv50_instmem_init(struct drm_device *dev)
 
        /* Assume that praying isn't enough, check that we can re-read the
         * entire fake channel back from the PRAMIN BAR */
-       dev_priv->engine.instmem.prepare_access(dev, false);
        for (i = 0; i < c_size; i += 4) {
                if (nv_rd32(dev, NV_RAMIN + i) != nv_ri32(dev, i)) {
                        NV_ERROR(dev, "Error reading back PRAMIN at 0x%08x\n",
                                                                        i);
-                       dev_priv->engine.instmem.finish_access(dev);
                        return -EINVAL;
                }
        }
-       dev_priv->engine.instmem.finish_access(dev);
 
        nv_wr32(dev, NV50_PUNK_BAR0_PRAMIN, save_nv001700);
 
@@ -451,13 +446,12 @@ nv50_instmem_bind(struct drm_device *dev, struct nouveau_gpuobj *gpuobj)
                vram |= 0x30;
        }
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
        while (pte < pte_end) {
                nv_wo32(dev, pramin_pt, pte++, lower_32_bits(vram));
                nv_wo32(dev, pramin_pt, pte++, upper_32_bits(vram));
                vram += NV50_INSTMEM_PAGE_SIZE;
        }
-       dev_priv->engine.instmem.finish_access(dev);
+       dev_priv->engine.instmem.flush(dev);
 
        nv_wr32(dev, 0x100c80, 0x00040001);
        if (!nv_wait(0x100c80, 0x00000001, 0x00000000)) {
@@ -490,36 +484,21 @@ nv50_instmem_unbind(struct drm_device *dev, struct nouveau_gpuobj *gpuobj)
        pte     = (gpuobj->im_pramin->start >> 12) << 1;
        pte_end = ((gpuobj->im_pramin->size >> 12) << 1) + pte;
 
-       dev_priv->engine.instmem.prepare_access(dev, true);
        while (pte < pte_end) {
                nv_wo32(dev, priv->pramin_pt->gpuobj, pte++, 0x00000000);
                nv_wo32(dev, priv->pramin_pt->gpuobj, pte++, 0x00000000);
        }
-       dev_priv->engine.instmem.finish_access(dev);
+       dev_priv->engine.instmem.flush(dev);
 
        gpuobj->im_bound = 0;
        return 0;
 }
 
 void
-nv50_instmem_prepare_access(struct drm_device *dev, bool write)
-{
-       struct drm_nouveau_private *dev_priv = dev->dev_private;
-       struct nv50_instmem_priv *priv = dev_priv->engine.instmem.priv;
-
-       priv->last_access_wr = write;
-}
-
-void
-nv50_instmem_finish_access(struct drm_device *dev)
+nv50_instmem_flush(struct drm_device *dev)
 {
-       struct drm_nouveau_private *dev_priv = dev->dev_private;
-       struct nv50_instmem_priv *priv = dev_priv->engine.instmem.priv;
-
-       if (priv->last_access_wr) {
-               nv_wr32(dev, 0x070000, 0x00000001);
-               if (!nv_wait(0x070000, 0x00000001, 0x00000000))
-                       NV_ERROR(dev, "PRAMIN flush timeout\n");
-       }
+       nv_wr32(dev, 0x070000, 0x00000001);
+       if (!nv_wait(0x070000, 0x00000001, 0x00000000))
+               NV_ERROR(dev, "PRAMIN flush timeout\n");
 }