igb: Add support for DMA_ATTR_WEAK_ORDERING
authorAlexander Duyck <alexander.h.duyck@intel.com>
Tue, 7 Feb 2017 02:25:26 +0000 (18:25 -0800)
committerJeff Kirsher <jeffrey.t.kirsher@intel.com>
Fri, 17 Mar 2017 19:11:43 +0000 (12:11 -0700)
Since we are already using DMA attributes in igb for Rx there is no reason
why we can't also apply DMA_ATTR_WEAK_ORDERING which is needed on some
platforms to improve performance.

Signed-off-by: Alexander Duyck <alexander.h.duyck@intel.com>
Tested-by: Aaron Brown <aaron.f.brown@intel.com>
Signed-off-by: Jeff Kirsher <jeffrey.t.kirsher@intel.com>
drivers/net/ethernet/intel/igb/igb.h
drivers/net/ethernet/intel/igb/igb_main.c

index acbc3abe2dddfc7bdf87a3724a95b136237a519e..87c9fe9d6f18f05c800e25d4349f61917ebe213d 100644 (file)
@@ -148,6 +148,9 @@ struct vf_data_storage {
 /* How many Rx Buffers do we bundle into one write to the hardware ? */
 #define IGB_RX_BUFFER_WRITE    16 /* Must be power of 2 */
 
+#define IGB_RX_DMA_ATTR \
+       (DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING)
+
 #define AUTO_ALL_MODES         0
 #define IGB_EEPROM_APME                0x0400
 
index be456bae816906e24338006a8b3597b539f86959..cf7ee9cdac6f96df318fd7dcf3cdb2fc12b49b00 100644 (file)
@@ -3963,7 +3963,7 @@ static void igb_clean_rx_ring(struct igb_ring *rx_ring)
                                     buffer_info->dma,
                                     PAGE_SIZE,
                                     DMA_FROM_DEVICE,
-                                    DMA_ATTR_SKIP_CPU_SYNC);
+                                    IGB_RX_DMA_ATTR);
                __page_frag_cache_drain(buffer_info->page,
                                        buffer_info->pagecnt_bias);
 
@@ -6990,7 +6990,7 @@ static struct sk_buff *igb_fetch_rx_buffer(struct igb_ring *rx_ring,
                 */
                dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma,
                                     PAGE_SIZE, DMA_FROM_DEVICE,
-                                    DMA_ATTR_SKIP_CPU_SYNC);
+                                    IGB_RX_DMA_ATTR);
                __page_frag_cache_drain(page, rx_buffer->pagecnt_bias);
        }
 
@@ -7250,7 +7250,7 @@ static bool igb_alloc_mapped_page(struct igb_ring *rx_ring,
 
        /* map page for use */
        dma = dma_map_page_attrs(rx_ring->dev, page, 0, PAGE_SIZE,
-                                DMA_FROM_DEVICE, DMA_ATTR_SKIP_CPU_SYNC);
+                                DMA_FROM_DEVICE, IGB_RX_DMA_ATTR);
 
        /* if mapping failed free memory back to system since
         * there isn't much point in holding memory we can't use