1 /* $NetBSD: drm_memory.c,v 1.17 2021/12/19 10:47:13 riastradh Exp $ */ 2 3 /*- 4 * Copyright (c) 2013 The NetBSD Foundation, Inc. 5 * All rights reserved. 6 * 7 * This code is derived from software contributed to The NetBSD Foundation 8 * by Taylor R. Campbell. 9 * 10 * Redistribution and use in source and binary forms, with or without 11 * modification, are permitted provided that the following conditions 12 * are met: 13 * 1. Redistributions of source code must retain the above copyright 14 * notice, this list of conditions and the following disclaimer. 15 * 2. Redistributions in binary form must reproduce the above copyright 16 * notice, this list of conditions and the following disclaimer in the 17 * documentation and/or other materials provided with the distribution. 18 * 19 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS 20 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED 21 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 22 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS 23 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 24 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 25 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 26 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 27 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 28 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 29 * POSSIBILITY OF SUCH DAMAGE. 30 */ 31 32 #include <sys/cdefs.h> 33 __KERNEL_RCSID(0, "$NetBSD: drm_memory.c,v 1.17 2021/12/19 10:47:13 riastradh Exp $"); 34 35 #if defined(__i386__) || defined(__x86_64__) 36 37 # ifdef _KERNEL_OPT 38 # include "agp.h" 39 # if NAGP > 0 40 # include "agp_i810.h" 41 # else 42 # define NAGP_I810 0 43 # endif 44 # include "genfb.h" 45 # else 46 # define NAGP_I810 1 47 # define NGENFB 0 48 # endif 49 50 #else 51 52 # ifdef _KERNEL_OPT 53 # define NAGP_I810 0 54 # include "genfb.h" 55 # else 56 # define NAGP_I810 0 57 # define NGENFB 0 58 # endif 59 60 #endif 61 62 #include <sys/bus.h> 63 64 #if NAGP_I810 > 0 65 /* XXX include order botch -- shouldn't need to include pcivar.h */ 66 #include <dev/pci/pcivar.h> 67 #include <dev/pci/agpvar.h> 68 #endif 69 70 #if NGENFB > 0 71 #include <dev/wsfb/genfbvar.h> 72 #endif 73 74 #include <drm/drm_device.h> 75 #include <drm/drm_drv.h> 76 #include <drm/drm_cache.h> 77 #include <drm/drm_legacy.h> 78 #include <drm/drm_pci.h> 79 #include <drm/drm_print.h> 80 81 /* 82 * XXX drm_bus_borrow is a horrible kludge! 83 */ 84 static bool 85 drm_bus_borrow(bus_addr_t base, bus_size_t size, bus_space_handle_t *handlep) 86 { 87 88 #if NAGP_I810 > 0 89 if (agp_i810_borrow(base, size, handlep)) 90 return true; 91 #endif 92 93 #if NGENFB > 0 94 if (genfb_borrow(base, handlep)) 95 return true; 96 #endif 97 98 return false; 99 } 100 101 void 102 drm_legacy_ioremap(struct drm_local_map *map, struct drm_device *dev) 103 { 104 const bus_space_tag_t bst = dev->bst; 105 unsigned int unit; 106 107 /* 108 * Search dev's bus maps for a match. 109 */ 110 for (unit = 0; unit < dev->bus_nmaps; unit++) { 111 struct drm_bus_map *const bm = &dev->bus_maps[unit]; 112 int flags = bm->bm_flags; 113 114 /* Reject maps starting after the request. */ 115 if (map->offset < bm->bm_base) 116 continue; 117 118 /* Reject maps smaller than the request. */ 119 if (bm->bm_size < map->size) 120 continue; 121 122 /* Reject maps that the request doesn't fit in. */ 123 if ((bm->bm_size - map->size) < 124 (map->offset - bm->bm_base)) 125 continue; 126 127 /* Ensure we can map the space into virtual memory. */ 128 if (!ISSET(flags, BUS_SPACE_MAP_LINEAR)) 129 continue; 130 131 /* Reflect requested flags in the bus_space map. */ 132 if (ISSET(map->flags, _DRM_WRITE_COMBINING)) 133 flags |= BUS_SPACE_MAP_PREFETCHABLE; 134 135 /* Map it. */ 136 if (bus_space_map(bst, map->offset, map->size, flags, 137 &map->lm_data.bus_space.bsh)) 138 break; 139 140 map->lm_data.bus_space.bus_map = bm; 141 goto win; 142 } 143 144 /* Couldn't map it. Try borrowing from someone else. */ 145 if (drm_bus_borrow(map->offset, map->size, 146 &map->lm_data.bus_space.bsh)) { 147 map->lm_data.bus_space.bus_map = NULL; 148 goto win; 149 } 150 151 /* Failure! */ 152 return; 153 154 win: map->lm_data.bus_space.bst = bst; 155 map->handle = bus_space_vaddr(bst, map->lm_data.bus_space.bsh); 156 } 157 158 void 159 drm_legacy_ioremapfree(struct drm_local_map *map, struct drm_device *dev) 160 { 161 if (map->lm_data.bus_space.bus_map != NULL) { 162 bus_space_unmap(map->lm_data.bus_space.bst, 163 map->lm_data.bus_space.bsh, map->size); 164 map->lm_data.bus_space.bus_map = NULL; 165 map->handle = NULL; 166 } 167 } 168 169 /* 170 * Allocate a drm dma handle, allocate memory fit for DMA, and map it. 171 * 172 * XXX This is called drm_pci_alloc for hysterical raisins; it is not 173 * specific to PCI. 174 * 175 * XXX For now, we use non-blocking allocations because this is called 176 * by ioctls with the drm global mutex held. 177 * 178 * XXX Error information is lost because this returns NULL on failure, 179 * not even an error embedded in a pointer. 180 */ 181 struct drm_dma_handle * 182 drm_pci_alloc(struct drm_device *dev, size_t size, size_t align) 183 { 184 int nsegs; 185 int error; 186 187 /* 188 * Allocate a drm_dma_handle record. 189 */ 190 struct drm_dma_handle *const dmah = kmem_alloc(sizeof(*dmah), 191 KM_NOSLEEP); 192 if (dmah == NULL) { 193 error = -ENOMEM; 194 goto out; 195 } 196 dmah->dmah_tag = dev->dmat; 197 198 /* 199 * Allocate the requested amount of DMA-safe memory. 200 */ 201 /* XXX errno NetBSD->Linux */ 202 error = -bus_dmamem_alloc(dmah->dmah_tag, size, align, 0, 203 &dmah->dmah_seg, 1, &nsegs, BUS_DMA_NOWAIT); 204 if (error) 205 goto fail0; 206 KASSERT(nsegs == 1); 207 208 /* 209 * Map the DMA-safe memory into kernel virtual address space. 210 */ 211 /* XXX errno NetBSD->Linux */ 212 error = -bus_dmamem_map(dmah->dmah_tag, &dmah->dmah_seg, 1, size, 213 &dmah->vaddr, 214 (BUS_DMA_NOWAIT | BUS_DMA_COHERENT | BUS_DMA_NOCACHE)); 215 if (error) 216 goto fail1; 217 dmah->size = size; 218 219 /* 220 * Create a map for DMA transfers. 221 */ 222 /* XXX errno NetBSD->Linux */ 223 error = -bus_dmamap_create(dmah->dmah_tag, size, 1, size, 0, 224 BUS_DMA_NOWAIT, &dmah->dmah_map); 225 if (error) 226 goto fail2; 227 228 /* 229 * Load the kva buffer into the map for DMA transfers. 230 */ 231 /* XXX errno NetBSD->Linux */ 232 error = -bus_dmamap_load(dmah->dmah_tag, dmah->dmah_map, dmah->vaddr, 233 size, NULL, (BUS_DMA_NOWAIT | BUS_DMA_NOCACHE)); 234 if (error) 235 goto fail3; 236 237 /* Record the bus address for convenient reference. */ 238 dmah->busaddr = dmah->dmah_map->dm_segs[0].ds_addr; 239 240 /* Zero the DMA buffer. XXX Yikes! Is this necessary? */ 241 memset(dmah->vaddr, 0, size); 242 243 /* Success! */ 244 return dmah; 245 246 fail3: bus_dmamap_destroy(dmah->dmah_tag, dmah->dmah_map); 247 fail2: bus_dmamem_unmap(dmah->dmah_tag, dmah->vaddr, dmah->size); 248 fail1: bus_dmamem_free(dmah->dmah_tag, &dmah->dmah_seg, 1); 249 fail0: dmah->dmah_tag = NULL; /* XXX paranoia */ 250 kmem_free(dmah, sizeof(*dmah)); 251 out: DRM_DEBUG("drm_pci_alloc failed: %d\n", error); 252 return NULL; 253 } 254 255 /* 256 * Release the bus DMA mappings and memory in dmah, and deallocate it. 257 */ 258 void 259 drm_pci_free(struct drm_device *dev, struct drm_dma_handle *dmah) 260 { 261 262 bus_dmamap_unload(dmah->dmah_tag, dmah->dmah_map); 263 bus_dmamap_destroy(dmah->dmah_tag, dmah->dmah_map); 264 bus_dmamem_unmap(dmah->dmah_tag, dmah->vaddr, dmah->size); 265 bus_dmamem_free(dmah->dmah_tag, &dmah->dmah_seg, 1); 266 dmah->dmah_tag = NULL; /* XXX paranoia */ 267 kmem_free(dmah, sizeof(*dmah)); 268 } 269 270 /* 271 * Make sure the DMA-safe memory allocated for dev lies between 272 * min_addr and max_addr. Can be used multiple times to restrict the 273 * bounds further, but never to expand the bounds again. 274 * 275 * XXX Caller must guarantee nobody has used the tag yet, 276 * i.e. allocated any DMA memory. 277 */ 278 int 279 drm_limit_dma_space(struct drm_device *dev, resource_size_t min_addr, 280 resource_size_t max_addr) 281 { 282 int ret; 283 284 KASSERT(min_addr <= max_addr); 285 286 /* 287 * Limit it further if we have already limited it, and destroy 288 * the old subregion DMA tag. 289 */ 290 if (dev->dmat_subregion_p) { 291 min_addr = MAX(min_addr, dev->dmat_subregion_min); 292 max_addr = MIN(max_addr, dev->dmat_subregion_max); 293 bus_dmatag_destroy(dev->dmat); 294 } 295 296 /* 297 * If our limit contains the 32-bit space but for some reason 298 * we can't use a subregion, either because the bus doesn't 299 * support >32-bit DMA or because bus_dma(9) on this platform 300 * lacks bus_dmatag_subregion, just use the 32-bit space. 301 */ 302 if (min_addr == 0 && max_addr >= UINT32_C(0xffffffff) && 303 dev->bus_dmat == dev->bus_dmat32) { 304 dma32: dev->dmat = dev->bus_dmat32; 305 dev->dmat_subregion_p = false; 306 dev->dmat_subregion_min = 0; 307 dev->dmat_subregion_max = UINT32_C(0xffffffff); 308 return 0; 309 } 310 311 /* 312 * Create a DMA tag for a subregion from the bus's DMA tag. If 313 * that fails, restore dev->dmat to the whole region so that we 314 * need not worry about dev->dmat being uninitialized (not that 315 * the caller should try to allocate DMA-safe memory on failure 316 * anyway, but...paranoia). 317 */ 318 /* XXX errno NetBSD->Linux */ 319 ret = -bus_dmatag_subregion(dev->bus_dmat, min_addr, max_addr, 320 &dev->dmat, BUS_DMA_WAITOK); 321 if (ret) { 322 /* 323 * bus_dmatag_subregion may fail. If so, and if the 324 * subregion contains the 32-bit space, just use the 325 * 32-bit DMA tag. 326 */ 327 if (ret == -EOPNOTSUPP && dev->bus_dmat32 && 328 min_addr == 0 && max_addr >= UINT32_C(0xffffffff)) 329 goto dma32; 330 /* XXX Back out? */ 331 dev->dmat = dev->bus_dmat; 332 dev->dmat_subregion_p = false; 333 dev->dmat_subregion_min = 0; 334 dev->dmat_subregion_max = __type_max(bus_addr_t); 335 return ret; 336 } 337 338 /* 339 * Remember that we have a subregion tag so that we know to 340 * destroy it later, and record the bounds in case we need to 341 * limit them again. 342 */ 343 dev->dmat_subregion_p = true; 344 dev->dmat_subregion_min = min_addr; 345 dev->dmat_subregion_max = max_addr; 346 347 /* Success! */ 348 return 0; 349 } 350 351 bool 352 drm_need_swiotlb(int dma_bits) 353 { 354 355 return false; 356 } 357