xref: /netbsd-src/sys/external/bsd/drm2/dist/drm/drm_pci.c (revision b7b7574d3bf8eeb51a1fa3977b59142ec6434a55)
1 /* drm_pci.h -- PCI DMA memory management wrappers for DRM -*- linux-c -*- */
2 /**
3  * \file drm_pci.c
4  * \brief Functions and ioctls to manage PCI memory
5  *
6  * \warning These interfaces aren't stable yet.
7  *
8  * \todo Implement the remaining ioctl's for the PCI pools.
9  * \todo The wrappers here are so thin that they would be better off inlined..
10  *
11  * \author José Fonseca <jrfonseca@tungstengraphics.com>
12  * \author Leif Delgass <ldelgass@retinalburn.net>
13  */
14 
15 /*
16  * Copyright 2003 José Fonseca.
17  * Copyright 2003 Leif Delgass.
18  * All Rights Reserved.
19  *
20  * Permission is hereby granted, free of charge, to any person obtaining a
21  * copy of this software and associated documentation files (the "Software"),
22  * to deal in the Software without restriction, including without limitation
23  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
24  * and/or sell copies of the Software, and to permit persons to whom the
25  * Software is furnished to do so, subject to the following conditions:
26  *
27  * The above copyright notice and this permission notice (including the next
28  * paragraph) shall be included in all copies or substantial portions of the
29  * Software.
30  *
31  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
32  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
33  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL THE
34  * AUTHORS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
35  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
36  * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
37  */
38 
39 #include <linux/pci.h>
40 #include <linux/slab.h>
41 #include <linux/dma-mapping.h>
42 #include <linux/export.h>
43 #include <drm/drmP.h>
44 
45 /**********************************************************************/
46 /** \name PCI memory */
47 /*@{*/
48 
49 /**
50  * \brief Allocate a PCI consistent memory block, for DMA.
51  */
52 drm_dma_handle_t *drm_pci_alloc(struct drm_device * dev, size_t size, size_t align)
53 {
54 	drm_dma_handle_t *dmah;
55 #if 1
56 	unsigned long addr;
57 	size_t sz;
58 #endif
59 
60 	/* pci_alloc_consistent only guarantees alignment to the smallest
61 	 * PAGE_SIZE order which is greater than or equal to the requested size.
62 	 * Return NULL here for now to make sure nobody tries for larger alignment
63 	 */
64 	if (align > size)
65 		return NULL;
66 
67 	dmah = kmalloc(sizeof(drm_dma_handle_t), GFP_KERNEL);
68 	if (!dmah)
69 		return NULL;
70 
71 	dmah->size = size;
72 	dmah->vaddr = dma_alloc_coherent(&dev->pdev->dev, size, &dmah->busaddr, GFP_KERNEL | __GFP_COMP);
73 
74 	if (dmah->vaddr == NULL) {
75 		kfree(dmah);
76 		return NULL;
77 	}
78 
79 	memset(dmah->vaddr, 0, size);
80 
81 	/* XXX - Is virt_to_page() legal for consistent mem? */
82 	/* Reserve */
83 	for (addr = (unsigned long)dmah->vaddr, sz = size;
84 	     sz > 0; addr += PAGE_SIZE, sz -= PAGE_SIZE) {
85 		SetPageReserved(virt_to_page(addr));
86 	}
87 
88 	return dmah;
89 }
90 
91 EXPORT_SYMBOL(drm_pci_alloc);
92 
93 /**
94  * \brief Free a PCI consistent memory block without freeing its descriptor.
95  *
96  * This function is for internal use in the Linux-specific DRM core code.
97  */
98 void __drm_pci_free(struct drm_device * dev, drm_dma_handle_t * dmah)
99 {
100 #if 1
101 	unsigned long addr;
102 	size_t sz;
103 #endif
104 
105 	if (dmah->vaddr) {
106 		/* XXX - Is virt_to_page() legal for consistent mem? */
107 		/* Unreserve */
108 		for (addr = (unsigned long)dmah->vaddr, sz = dmah->size;
109 		     sz > 0; addr += PAGE_SIZE, sz -= PAGE_SIZE) {
110 			ClearPageReserved(virt_to_page(addr));
111 		}
112 		dma_free_coherent(&dev->pdev->dev, dmah->size, dmah->vaddr,
113 				  dmah->busaddr);
114 	}
115 }
116 
117 /**
118  * \brief Free a PCI consistent memory block
119  */
120 void drm_pci_free(struct drm_device * dev, drm_dma_handle_t * dmah)
121 {
122 	__drm_pci_free(dev, dmah);
123 	kfree(dmah);
124 }
125 
126 EXPORT_SYMBOL(drm_pci_free);
127 
128 #ifdef CONFIG_PCI
129 
130 static int drm_get_pci_domain(struct drm_device *dev)
131 {
132 #ifndef __alpha__
133 	/* For historical reasons, drm_get_pci_domain() is busticated
134 	 * on most archs and has to remain so for userspace interface
135 	 * < 1.4, except on alpha which was right from the beginning
136 	 */
137 	if (dev->if_version < 0x10004)
138 		return 0;
139 #endif /* __alpha__ */
140 
141 	return pci_domain_nr(dev->pdev->bus);
142 }
143 
144 static int drm_pci_get_irq(struct drm_device *dev)
145 {
146 	return dev->pdev->irq;
147 }
148 
149 static const char *drm_pci_get_name(struct drm_device *dev)
150 {
151 	struct pci_driver *pdriver = dev->driver->kdriver.pci;
152 	return pdriver->name;
153 }
154 
155 int drm_pci_set_busid(struct drm_device *dev, struct drm_master *master)
156 {
157 	int len, ret;
158 	struct pci_driver *pdriver = dev->driver->kdriver.pci;
159 	master->unique_len = 40;
160 	master->unique_size = master->unique_len;
161 	master->unique = kmalloc(master->unique_size, GFP_KERNEL);
162 	if (master->unique == NULL)
163 		return -ENOMEM;
164 
165 
166 	len = snprintf(master->unique, master->unique_len,
167 		       "pci:%04x:%02x:%02x.%d",
168 		       drm_get_pci_domain(dev),
169 		       dev->pdev->bus->number,
170 		       PCI_SLOT(dev->pdev->devfn),
171 		       PCI_FUNC(dev->pdev->devfn));
172 
173 	if (len >= master->unique_len) {
174 		DRM_ERROR("buffer overflow");
175 		ret = -EINVAL;
176 		goto err;
177 	} else
178 		master->unique_len = len;
179 
180 	size_t devlen = strlen(pdriver->name) + master->unique_len + 2;
181 	dev->devname = kmalloc(dev->devlen, GFP_KERNEL);
182 
183 	if (dev->devname == NULL) {
184 		ret = -ENOMEM;
185 		goto err;
186 	}
187 
188 	snprintf(dev->devname, devlen, "%s@%s", pdriver->name,
189 		master->unique);
190 
191 	return 0;
192 err:
193 	return ret;
194 }
195 
196 int drm_pci_set_unique(struct drm_device *dev,
197 		       struct drm_master *master,
198 		       struct drm_unique *u)
199 {
200 	int domain, bus, slot, func, ret;
201 	const char *bus_name;
202 
203 	master->unique_len = u->unique_len;
204 	master->unique_size = u->unique_len + 1;
205 	master->unique = kmalloc(master->unique_size, GFP_KERNEL);
206 	if (!master->unique) {
207 		ret = -ENOMEM;
208 		goto err;
209 	}
210 
211 	if (copy_from_user(master->unique, u->unique, master->unique_len)) {
212 		ret = -EFAULT;
213 		goto err;
214 	}
215 
216 	master->unique[master->unique_len] = '\0';
217 
218 	bus_name = dev->driver->bus->get_name(dev);
219 	size_t devlen = strlen(bus_name) + strlen(master->unique) + 2;
220 	dev->devname = kmalloc(devlen, GFP_KERNEL);
221 	if (!dev->devname) {
222 		ret = -ENOMEM;
223 		goto err;
224 	}
225 
226 	snprintf(dev->devname, devlen, "%s@%s", bus_name,
227 		master->unique);
228 
229 	/* Return error if the busid submitted doesn't match the device's actual
230 	 * busid.
231 	 */
232 	ret = sscanf(master->unique, "PCI:%d:%d:%d", &bus, &slot, &func);
233 	if (ret != 3) {
234 		ret = -EINVAL;
235 		goto err;
236 	}
237 
238 	domain = bus >> 8;
239 	bus &= 0xff;
240 
241 	if ((domain != drm_get_pci_domain(dev)) ||
242 	    (bus != dev->pdev->bus->number) ||
243 	    (slot != PCI_SLOT(dev->pdev->devfn)) ||
244 	    (func != PCI_FUNC(dev->pdev->devfn))) {
245 		ret = -EINVAL;
246 		goto err;
247 	}
248 	return 0;
249 err:
250 	return ret;
251 }
252 
253 
254 static int drm_pci_irq_by_busid(struct drm_device *dev, struct drm_irq_busid *p)
255 {
256 	if ((p->busnum >> 8) != drm_get_pci_domain(dev) ||
257 	    (p->busnum & 0xff) != dev->pdev->bus->number ||
258 	    p->devnum != PCI_SLOT(dev->pdev->devfn) || p->funcnum != PCI_FUNC(dev->pdev->devfn))
259 		return -EINVAL;
260 
261 	p->irq = dev->pdev->irq;
262 
263 	DRM_DEBUG("%d:%d:%d => IRQ %d\n", p->busnum, p->devnum, p->funcnum,
264 		  p->irq);
265 	return 0;
266 }
267 
268 int drm_pci_agp_init(struct drm_device *dev)
269 {
270 	if (drm_core_has_AGP(dev)) {
271 		if (drm_pci_device_is_agp(dev))
272 			dev->agp = drm_agp_init(dev);
273 		if (drm_core_check_feature(dev, DRIVER_REQUIRE_AGP)
274 		    && (dev->agp == NULL)) {
275 			DRM_ERROR("Cannot initialize the agpgart module.\n");
276 			return -EINVAL;
277 		}
278 		if (drm_core_has_MTRR(dev)) {
279 			if (dev->agp)
280 				dev->agp->agp_mtrr =
281 					mtrr_add(dev->agp->agp_info.aper_base,
282 						 dev->agp->agp_info.aper_size *
283 						 1024 * 1024, MTRR_TYPE_WRCOMB, 1);
284 		}
285 	}
286 	return 0;
287 }
288 
289 static struct drm_bus drm_pci_bus = {
290 	.bus_type = DRIVER_BUS_PCI,
291 	.get_irq = drm_pci_get_irq,
292 	.get_name = drm_pci_get_name,
293 	.set_busid = drm_pci_set_busid,
294 	.set_unique = drm_pci_set_unique,
295 	.irq_by_busid = drm_pci_irq_by_busid,
296 	.agp_init = drm_pci_agp_init,
297 };
298 
299 /**
300  * Register.
301  *
302  * \param pdev - PCI device structure
303  * \param ent entry from the PCI ID table with device type flags
304  * \return zero on success or a negative number on failure.
305  *
306  * Attempt to gets inter module "drm" information. If we are first
307  * then register the character device and inter module information.
308  * Try and register, if we fail to register, backout previous work.
309  */
310 int drm_get_pci_dev(struct pci_dev *pdev, const struct pci_device_id *ent,
311 		    struct drm_driver *driver)
312 {
313 	struct drm_device *dev;
314 	int ret;
315 
316 	DRM_DEBUG("\n");
317 
318 	dev = kzalloc(sizeof(*dev), GFP_KERNEL);
319 	if (!dev)
320 		return -ENOMEM;
321 
322 	ret = pci_enable_device(pdev);
323 	if (ret)
324 		goto err_g1;
325 
326 	dev->pdev = pdev;
327 	dev->dev = &pdev->dev;
328 
329 	dev->pci_device = pdev->device;
330 	dev->pci_vendor = pdev->vendor;
331 
332 #ifdef __alpha__
333 	dev->hose = pdev->sysdata;
334 #endif
335 
336 	mutex_lock(&drm_global_mutex);
337 
338 	if ((ret = drm_fill_in_dev(dev, ent, driver))) {
339 		printk(KERN_ERR "DRM: Fill_in_dev failed.\n");
340 		goto err_g2;
341 	}
342 
343 	if (drm_core_check_feature(dev, DRIVER_MODESET)) {
344 		pci_set_drvdata(pdev, dev);
345 		ret = drm_get_minor(dev, &dev->control, DRM_MINOR_CONTROL);
346 		if (ret)
347 			goto err_g2;
348 	}
349 
350 	if ((ret = drm_get_minor(dev, &dev->primary, DRM_MINOR_LEGACY)))
351 		goto err_g3;
352 
353 	if (dev->driver->load) {
354 		ret = dev->driver->load(dev, ent->driver_data);
355 		if (ret)
356 			goto err_g4;
357 	}
358 
359 	/* setup the grouping for the legacy output */
360 	if (drm_core_check_feature(dev, DRIVER_MODESET)) {
361 		ret = drm_mode_group_init_legacy_group(dev,
362 						&dev->primary->mode_group);
363 		if (ret)
364 			goto err_g4;
365 	}
366 
367 	list_add_tail(&dev->driver_item, &driver->device_list);
368 
369 	DRM_INFO("Initialized %s %d.%d.%d %s for %s on minor %d\n",
370 		 driver->name, driver->major, driver->minor, driver->patchlevel,
371 		 driver->date, pci_name(pdev), dev->primary->index);
372 
373 	mutex_unlock(&drm_global_mutex);
374 	return 0;
375 
376 err_g4:
377 	drm_put_minor(&dev->primary);
378 err_g3:
379 	if (drm_core_check_feature(dev, DRIVER_MODESET))
380 		drm_put_minor(&dev->control);
381 err_g2:
382 	pci_disable_device(pdev);
383 err_g1:
384 	kfree(dev);
385 	mutex_unlock(&drm_global_mutex);
386 	return ret;
387 }
388 EXPORT_SYMBOL(drm_get_pci_dev);
389 
390 /**
391  * PCI device initialization. Called direct from modules at load time.
392  *
393  * \return zero on success or a negative number on failure.
394  *
395  * Initializes a drm_device structures,registering the
396  * stubs and initializing the AGP device.
397  *
398  * Expands the \c DRIVER_PREINIT and \c DRIVER_POST_INIT macros before and
399  * after the initialization for driver customization.
400  */
401 int drm_pci_init(struct drm_driver *driver, struct pci_driver *pdriver)
402 {
403 	struct pci_dev *pdev = NULL;
404 	const struct pci_device_id *pid;
405 	int i;
406 
407 	DRM_DEBUG("\n");
408 
409 	INIT_LIST_HEAD(&driver->device_list);
410 	driver->kdriver.pci = pdriver;
411 	driver->bus = &drm_pci_bus;
412 
413 	if (driver->driver_features & DRIVER_MODESET)
414 		return pci_register_driver(pdriver);
415 
416 	/* If not using KMS, fall back to stealth mode manual scanning. */
417 	for (i = 0; pdriver->id_table[i].vendor != 0; i++) {
418 		pid = &pdriver->id_table[i];
419 
420 		/* Loop around setting up a DRM device for each PCI device
421 		 * matching our ID and device class.  If we had the internal
422 		 * function that pci_get_subsys and pci_get_class used, we'd
423 		 * be able to just pass pid in instead of doing a two-stage
424 		 * thing.
425 		 */
426 		pdev = NULL;
427 		while ((pdev =
428 			pci_get_subsys(pid->vendor, pid->device, pid->subvendor,
429 				       pid->subdevice, pdev)) != NULL) {
430 			if ((pdev->class & pid->class_mask) != pid->class)
431 				continue;
432 
433 			/* stealth mode requires a manual probe */
434 			pci_dev_get(pdev);
435 			drm_get_pci_dev(pdev, pid, driver);
436 		}
437 	}
438 	return 0;
439 }
440 
441 #else
442 
443 int drm_pci_init(struct drm_driver *driver, struct pci_driver *pdriver)
444 {
445 	return -1;
446 }
447 
448 #endif
449 
450 EXPORT_SYMBOL(drm_pci_init);
451 
452 /*@}*/
453 void drm_pci_exit(struct drm_driver *driver, struct pci_driver *pdriver)
454 {
455 	struct drm_device *dev, *tmp;
456 	DRM_DEBUG("\n");
457 
458 	if (driver->driver_features & DRIVER_MODESET) {
459 		pci_unregister_driver(pdriver);
460 	} else {
461 		list_for_each_entry_safe(dev, tmp, &driver->device_list, driver_item)
462 			drm_put_dev(dev);
463 	}
464 	DRM_INFO("Module unloaded\n");
465 }
466 EXPORT_SYMBOL(drm_pci_exit);
467 
468 int drm_pcie_get_speed_cap_mask(struct drm_device *dev, u32 *mask)
469 {
470 	struct pci_dev *root;
471 	int pos;
472 	u32 lnkcap = 0, lnkcap2 = 0;
473 
474 	*mask = 0;
475 	if (!dev->pdev)
476 		return -EINVAL;
477 
478 	if (!pci_is_pcie(dev->pdev))
479 		return -EINVAL;
480 
481 	root = dev->pdev->bus->self;
482 
483 	pos = pci_pcie_cap(root);
484 	if (!pos)
485 		return -EINVAL;
486 
487 	/* we've been informed via and serverworks don't make the cut */
488 	if (root->vendor == PCI_VENDOR_ID_VIA ||
489 	    root->vendor == PCI_VENDOR_ID_SERVERWORKS)
490 		return -EINVAL;
491 
492 	pci_read_config_dword(root, pos + PCI_EXP_LNKCAP, &lnkcap);
493 	pci_read_config_dword(root, pos + PCI_EXP_LNKCAP2, &lnkcap2);
494 
495 	lnkcap &= PCI_EXP_LNKCAP_SLS;
496 	lnkcap2 &= 0xfe;
497 
498 	if (lnkcap2) { /* PCIE GEN 3.0 */
499 		if (lnkcap2 & PCI_EXP_LNKCAP2_SLS_2_5GB)
500 			*mask |= DRM_PCIE_SPEED_25;
501 		if (lnkcap2 & PCI_EXP_LNKCAP2_SLS_5_0GB)
502 			*mask |= DRM_PCIE_SPEED_50;
503 		if (lnkcap2 & PCI_EXP_LNKCAP2_SLS_8_0GB)
504 			*mask |= DRM_PCIE_SPEED_80;
505 	} else {
506 		if (lnkcap & 1)
507 			*mask |= DRM_PCIE_SPEED_25;
508 		if (lnkcap & 2)
509 			*mask |= DRM_PCIE_SPEED_50;
510 	}
511 
512 	DRM_INFO("probing gen 2 caps for device %x:%x = %x/%x\n", root->vendor, root->device, lnkcap, lnkcap2);
513 	return 0;
514 }
515 EXPORT_SYMBOL(drm_pcie_get_speed_cap_mask);
516