if (ring->descriptors) {
dma_free_coherent(ring->hw_dev,
ring->n_links *
- sizeof(struct mite_dma_descriptor),
+ sizeof(struct mite_dma_desc),
ring->descriptors,
ring->descriptors_dma_addr);
}
if (ring->descriptors) {
dma_free_coherent(ring->hw_dev,
ring->n_links *
- sizeof(struct mite_dma_descriptor),
+ sizeof(struct mite_dma_desc),
ring->descriptors,
ring->descriptors_dma_addr);
}
ring->descriptors =
dma_alloc_coherent(ring->hw_dev,
- n_links * sizeof(struct mite_dma_descriptor),
+ n_links * sizeof(struct mite_dma_desc),
&ring->descriptors_dma_addr, GFP_KERNEL);
if (!ring->descriptors) {
dev_err(s->device->class_dev,
cpu_to_le32(async->buf_map->page_list[i].dma_addr);
ring->descriptors[i].next =
cpu_to_le32(ring->descriptors_dma_addr +
- (i + 1) * sizeof(struct mite_dma_descriptor));
+ (i + 1) * sizeof(struct mite_dma_desc));
}
/* the last link is either a remainder or was a full link. */
struct device;
struct pci_dev;
-struct mite_dma_descriptor {
+struct mite_dma_desc {
__le32 count;
__le32 addr;
__le32 next;
struct mite_dma_descriptor_ring {
struct device *hw_dev;
unsigned int n_links;
- struct mite_dma_descriptor *descriptors;
+ struct mite_dma_desc *descriptors;
dma_addr_t descriptors_dma_addr;
};