Lines Matching defs:dma
145 dma_addr_t dma;
151 dma = tx_data->dma;
153 dma_unmap_single(&pdev->dev, dma, skb->len, DMA_TO_DEVICE);
169 dma_free_coherent(&pdev->dev, ring_sz, head, tx_ring->r.dma);
179 dma_addr_t dma;
187 dma = *((dma_addr_t *)skb->cb);
188 dma_unmap_single(&pdev->dev, dma, buf_sz, DMA_FROM_DEVICE);
203 dma_free_coherent(&pdev->dev, ring_sz, head, rx_ring->r.dma);
230 dma_addr_t dma;
238 head = dma_alloc_coherent(&pdev->dev, ring_sz, &dma, GFP_KERNEL);
246 tx_ring->r.dma = dma;
262 dma_addr_t dma;
267 dma = dma_map_single(&pdev->dev, skb->data, buf_sz, DMA_FROM_DEVICE);
268 if (dma_mapping_error(&pdev->dev, dma))
271 *((dma_addr_t *)skb->cb) = dma;
276 buf_desc->dma = cpu_to_le32(dma);
281 static void rtw_pci_sync_rx_desc_device(struct rtw_dev *rtwdev, dma_addr_t dma,
289 dma_sync_single_for_device(dev, dma, buf_sz, DMA_FROM_DEVICE);
295 buf_desc->dma = cpu_to_le32(dma);
304 dma_addr_t dma;
311 head = dma_alloc_coherent(&pdev->dev, ring_sz, &dma, GFP_KERNEL);
336 rx_ring->r.dma = dma;
349 dma = *((dma_addr_t *)skb->cb);
350 dma_unmap_single(&pdev->dev, dma, buf_sz, DMA_FROM_DEVICE);
354 dma_free_coherent(&pdev->dev, ring_sz, head, dma);
446 dma_addr_t dma;
451 dma = rtwpci->tx_rings[RTW_TX_QUEUE_BCN].r.dma;
452 rtw_write32(rtwdev, RTK_PCI_TXBD_DESA_BCNQ, dma);
456 dma = rtwpci->tx_rings[RTW_TX_QUEUE_H2C].r.dma;
460 rtw_write32(rtwdev, RTK_PCI_TXBD_DESA_H2CQ, dma);
464 dma = rtwpci->tx_rings[RTW_TX_QUEUE_BK].r.dma;
468 rtw_write32(rtwdev, RTK_PCI_TXBD_DESA_BKQ, dma);
471 dma = rtwpci->tx_rings[RTW_TX_QUEUE_BE].r.dma;
475 rtw_write32(rtwdev, RTK_PCI_TXBD_DESA_BEQ, dma);
478 dma = rtwpci->tx_rings[RTW_TX_QUEUE_VO].r.dma;
482 rtw_write32(rtwdev, RTK_PCI_TXBD_DESA_VOQ, dma);
485 dma = rtwpci->tx_rings[RTW_TX_QUEUE_VI].r.dma;
489 rtw_write32(rtwdev, RTK_PCI_TXBD_DESA_VIQ, dma);
492 dma = rtwpci->tx_rings[RTW_TX_QUEUE_MGMT].r.dma;
496 rtw_write32(rtwdev, RTK_PCI_TXBD_DESA_MGMTQ, dma);
499 dma = rtwpci->tx_rings[RTW_TX_QUEUE_HI0].r.dma;
503 rtw_write32(rtwdev, RTK_PCI_TXBD_DESA_HI0Q, dma);
506 dma = rtwpci->rx_rings[RTW_RX_QUEUE_MPDU].r.dma;
510 rtw_write32(rtwdev, RTK_PCI_RXBD_DESA_MPDUQ, dma);
567 /* reset dma and rx tag */
722 dma_addr_t dma;
728 dma = tx_data->dma;
729 dma_unmap_single(&rtwpci->pdev->dev, dma, prev->len, DMA_TO_DEVICE);
749 rtw_warn(rtwdev, "pci bus timeout, check dma status\n");
858 dma_addr_t dma;
879 dma = dma_map_single(&rtwpci->pdev->dev, skb->data, skb->len,
881 if (dma_mapping_error(&rtwpci->pdev->dev, dma))
884 /* after this we got dma mapped, there is no way back */
893 buf_desc[0].dma = cpu_to_le32(dma);
895 buf_desc[1].dma = cpu_to_le32(dma + tx_pkt_desc_sz);
898 tx_data->dma = dma;
1029 dma_unmap_single(&rtwpci->pdev->dev, tx_data->dma, skb->len,
1113 dma_addr_t dma;
1121 dma = *((dma_addr_t *)skb->cb);
1122 dma_sync_single_for_cpu(rtwdev->dev, dma, RTK_PCI_RX_BUF_SIZE,
1156 rtw_pci_sync_rx_desc_device(rtwdev, dma, ring, cur_rp,
1272 rtw_err(rtwdev, "failed to set dma mask to 32-bit\n");
1278 rtw_err(rtwdev, "failed to set consistent dma mask to 32-bit\n");
1752 * not be processed immediately. Check whether dma ring is