Lines Matching defs:gmi

215     struct gdma_mem_info *gmi)
221 if (!gc || !gmi)
237 &gmi->dma_tag);
249 err = bus_dmamem_alloc(gmi->dma_tag, &buf,
250 BUS_DMA_NOWAIT | BUS_DMA_COHERENT | BUS_DMA_ZERO, &gmi->dma_map);
254 bus_dma_tag_destroy(gmi->dma_tag);
258 err = bus_dmamap_load(gmi->dma_tag, gmi->dma_map, buf,
263 bus_dmamem_free(gmi->dma_tag, buf, gmi->dma_map);
264 bus_dma_tag_destroy(gmi->dma_tag);
268 gmi->dev = gc->dev;
269 gmi->dma_handle = dma_handle;
270 gmi->virt_addr = buf;
271 gmi->length = length;
277 mana_gd_free_memory(struct gdma_mem_info *gmi)
279 bus_dmamap_unload(gmi->dma_tag, gmi->dma_map);
280 bus_dmamem_free(gmi->dma_tag, gmi->virt_addr, gmi->dma_map);
281 bus_dma_tag_destroy(gmi->dma_tag);
866 struct gdma_mem_info *gmi;
871 gmi = &queue->mem_info;
872 err = mana_gd_alloc_memory(gc, spec->queue_size, gmi);
878 queue->queue_mem_ptr = gmi->virt_addr;
895 mana_gd_free_memory(gmi);
930 struct gdma_mem_info *gmi)
932 unsigned int num_page = gmi->length / PAGE_SIZE;
937 uint32_t length = gmi->length;
943 mana_err(NULL, "gmi size incorrect: %u\n", length);
947 if (offset_in_page((uintptr_t)gmi->virt_addr) != 0) {
948 mana_err(NULL, "gmi not page aligned: %p\n",
949 gmi->virt_addr);
971 req->page_addr_list[i] = gmi->dma_handle + i * PAGE_SIZE;
985 gmi->dma_region_handle = resp.dma_region_handle;
997 struct gdma_mem_info *gmi;
1005 gmi = &queue->mem_info;
1006 err = mana_gd_alloc_memory(gc, spec->queue_size, gmi);
1010 err = mana_gd_create_dma_region(gd, gmi);
1016 queue->queue_mem_ptr = gmi->virt_addr;
1030 mana_gd_free_memory(gmi);
1041 struct gdma_mem_info *gmi;
1050 gmi = &queue->mem_info;
1051 err = mana_gd_alloc_memory(gc, spec->queue_size, gmi);
1055 err = mana_gd_create_dma_region(gd, gmi);
1061 queue->queue_mem_ptr = gmi->virt_addr;
1074 mana_gd_free_memory(gmi);
1083 struct gdma_mem_info *gmi = &queue->mem_info;
1107 mana_gd_destroy_dma_region(gc, gmi->dma_region_handle);
1108 mana_gd_free_memory(gmi);