1 /*
2 * CDDL HEADER START
3 *
4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
7 *
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or http://www.opensolaris.org/os/licensing.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
12 *
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
18 *
19 * CDDL HEADER END
20 */
21 /*
22 * Copyright (c) 2005, 2010, Oracle and/or its affiliates. All rights reserved.
23 */
24
25 /*
26 * Common x86 and SPARC PCI-E to PCI bus bridge nexus driver
27 */
28
29 #include <sys/sysmacros.h>
30 #include <sys/conf.h>
31 #include <sys/kmem.h>
32 #include <sys/debug.h>
33 #include <sys/modctl.h>
34 #include <sys/autoconf.h>
35 #include <sys/ddi_impldefs.h>
36 #include <sys/pci.h>
37 #include <sys/ddi.h>
38 #include <sys/sunddi.h>
39 #include <sys/sunndi.h>
40 #include <sys/fm/util.h>
41 #include <sys/pci_cap.h>
42 #include <sys/pci_impl.h>
43 #include <sys/pcie_impl.h>
44 #include <sys/open.h>
45 #include <sys/stat.h>
46 #include <sys/file.h>
47 #include <sys/promif.h> /* prom_printf */
48 #include <sys/disp.h>
49 #include <sys/pcie_pwr.h>
50 #include <sys/hotplug/pci/pcie_hp.h>
51 #include "pcieb.h"
52 #ifdef PX_PLX
53 #include <io/pciex/pcieb_plx.h>
54 #endif /* PX_PLX */
55
56 /*LINTLIBRARY*/
57
58 /* panic flag */
59 int pcieb_die = PF_ERR_FATAL_FLAGS;
60 int pcieb_disable_41210_wkarnd = 0;
61
62 /* flag to turn on MSI support */
63 int pcieb_enable_msi = 1;
64
65 #if defined(DEBUG)
66 uint_t pcieb_dbg_print = 0;
67
68 static char *pcieb_debug_sym [] = { /* same sequence as pcieb_debug_bit */
69 /* 0 */ "attach",
70 /* 1 */ "pwr",
71 /* 2 */ "intr"
72 };
73 #endif /* DEBUG */
74
75 static int pcieb_bus_map(dev_info_t *, dev_info_t *, ddi_map_req_t *, off_t,
76 off_t, caddr_t *);
77 static int pcieb_ctlops(dev_info_t *, dev_info_t *, ddi_ctl_enum_t, void *,
78 void *);
79 static int pcieb_fm_init(pcieb_devstate_t *pcieb_p);
80 static void pcieb_fm_fini(pcieb_devstate_t *pcieb_p);
81 static int pcieb_fm_init_child(dev_info_t *dip, dev_info_t *cdip, int cap,
82 ddi_iblock_cookie_t *ibc_p);
83 static int pcieb_dma_allochdl(dev_info_t *dip, dev_info_t *rdip,
84 ddi_dma_attr_t *attr_p, int (*waitfp)(caddr_t), caddr_t arg,
85 ddi_dma_handle_t *handlep);
86 static int pcieb_dma_mctl(dev_info_t *dip, dev_info_t *rdip,
87 ddi_dma_handle_t handle, enum ddi_dma_ctlops cmd, off_t *offp,
88 size_t *lenp, caddr_t *objp, uint_t cache_flags);
89 static int pcieb_intr_ops(dev_info_t *dip, dev_info_t *rdip,
90 ddi_intr_op_t intr_op, ddi_intr_handle_impl_t *hdlp, void *result);
91
92 static struct bus_ops pcieb_bus_ops = {
93 BUSO_REV,
94 pcieb_bus_map,
95 0,
96 0,
97 0,
98 i_ddi_map_fault,
99 ddi_dma_map,
100 pcieb_dma_allochdl,
101 ddi_dma_freehdl,
102 ddi_dma_bindhdl,
103 ddi_dma_unbindhdl,
104 ddi_dma_flush,
105 ddi_dma_win,
106 pcieb_dma_mctl,
107 pcieb_ctlops,
108 ddi_bus_prop_op,
109 ndi_busop_get_eventcookie, /* (*bus_get_eventcookie)(); */
110 ndi_busop_add_eventcall, /* (*bus_add_eventcall)(); */
111 ndi_busop_remove_eventcall, /* (*bus_remove_eventcall)(); */
112 ndi_post_event, /* (*bus_post_event)(); */
113 NULL, /* (*bus_intr_ctl)(); */
114 NULL, /* (*bus_config)(); */
115 NULL, /* (*bus_unconfig)(); */
116 pcieb_fm_init_child, /* (*bus_fm_init)(); */
117 NULL, /* (*bus_fm_fini)(); */
118 i_ndi_busop_access_enter, /* (*bus_fm_access_enter)(); */
119 i_ndi_busop_access_exit, /* (*bus_fm_access_exit)(); */
120 pcie_bus_power, /* (*bus_power)(); */
121 pcieb_intr_ops, /* (*bus_intr_op)(); */
122 pcie_hp_common_ops /* (*bus_hp_op)(); */
123 };
124
125 static int pcieb_open(dev_t *, int, int, cred_t *);
126 static int pcieb_close(dev_t, int, int, cred_t *);
127 static int pcieb_ioctl(dev_t, int, intptr_t, int, cred_t *, int *);
128 static int pcieb_info(dev_info_t *, ddi_info_cmd_t, void *, void **);
129 static uint_t pcieb_intr_handler(caddr_t arg1, caddr_t arg2);
130
131 /* PM related functions */
132 static int pcieb_pwr_setup(dev_info_t *dip);
133 static int pcieb_pwr_init_and_raise(dev_info_t *dip, pcie_pwr_t *pwr_p);
134 static void pcieb_pwr_teardown(dev_info_t *dip);
135 static int pcieb_pwr_disable(dev_info_t *dip);
136
137 /* Hotplug related functions */
138 static void pcieb_id_props(pcieb_devstate_t *pcieb);
139
140 /*
141 * soft state pointer
142 */
143 void *pcieb_state;
144
145 static struct cb_ops pcieb_cb_ops = {
146 pcieb_open, /* open */
147 pcieb_close, /* close */
148 nodev, /* strategy */
149 nodev, /* print */
150 nodev, /* dump */
151 nodev, /* read */
152 nodev, /* write */
153 pcieb_ioctl, /* ioctl */
154 nodev, /* devmap */
155 nodev, /* mmap */
156 nodev, /* segmap */
157 nochpoll, /* poll */
158 pcie_prop_op, /* cb_prop_op */
159 NULL, /* streamtab */
160 D_NEW | D_MP | D_HOTPLUG, /* Driver compatibility flag */
161 CB_REV, /* rev */
162 nodev, /* int (*cb_aread)() */
163 nodev /* int (*cb_awrite)() */
164 };
165
166 static int pcieb_probe(dev_info_t *);
167 static int pcieb_attach(dev_info_t *devi, ddi_attach_cmd_t cmd);
168 static int pcieb_detach(dev_info_t *devi, ddi_detach_cmd_t cmd);
169
170 static struct dev_ops pcieb_ops = {
171 DEVO_REV, /* devo_rev */
172 0, /* refcnt */
173 pcieb_info, /* info */
174 nulldev, /* identify */
175 pcieb_probe, /* probe */
176 pcieb_attach, /* attach */
177 pcieb_detach, /* detach */
178 nulldev, /* reset */
179 &pcieb_cb_ops, /* driver operations */
180 &pcieb_bus_ops, /* bus operations */
181 pcie_power, /* power */
182 ddi_quiesce_not_needed, /* quiesce */
183 };
184
185 /*
186 * Module linkage information for the kernel.
187 */
188
189 static struct modldrv modldrv = {
190 &mod_driverops, /* Type of module */
191 "PCIe bridge/switch driver",
192 &pcieb_ops, /* driver ops */
193 };
194
195 static struct modlinkage modlinkage = {
196 MODREV_1,
197 (void *)&modldrv,
198 NULL
199 };
200
201 /*
202 * forward function declarations:
203 */
204 static void pcieb_uninitchild(dev_info_t *);
205 static int pcieb_initchild(dev_info_t *child);
206 static void pcieb_create_ranges_prop(dev_info_t *, ddi_acc_handle_t);
207 static boolean_t pcieb_is_pcie_device_type(dev_info_t *dip);
208
209 /* interrupt related declarations */
210 static int pcieb_msi_supported(dev_info_t *);
211 static int pcieb_intr_attach(pcieb_devstate_t *pcieb);
212 static int pcieb_intr_init(pcieb_devstate_t *pcieb_p, int intr_type);
213 static void pcieb_intr_fini(pcieb_devstate_t *pcieb_p);
214
215 int
_init(void)216 _init(void)
217 {
218 int e;
219
220 if ((e = ddi_soft_state_init(&pcieb_state, sizeof (pcieb_devstate_t),
221 1)) == 0 && (e = mod_install(&modlinkage)) != 0)
222 ddi_soft_state_fini(&pcieb_state);
223 return (e);
224 }
225
226 int
_fini(void)227 _fini(void)
228 {
229 int e;
230
231 if ((e = mod_remove(&modlinkage)) == 0) {
232 ddi_soft_state_fini(&pcieb_state);
233 }
234 return (e);
235 }
236
237 int
_info(struct modinfo * modinfop)238 _info(struct modinfo *modinfop)
239 {
240 return (mod_info(&modlinkage, modinfop));
241 }
242
243 /* ARGSUSED */
244 static int
pcieb_info(dev_info_t * dip,ddi_info_cmd_t infocmd,void * arg,void ** result)245 pcieb_info(dev_info_t *dip, ddi_info_cmd_t infocmd, void *arg, void **result)
246 {
247 minor_t minor = getminor((dev_t)arg);
248 int instance = PCI_MINOR_NUM_TO_INSTANCE(minor);
249 pcieb_devstate_t *pcieb = ddi_get_soft_state(pcieb_state, instance);
250 int ret = DDI_SUCCESS;
251
252 switch (infocmd) {
253 case DDI_INFO_DEVT2INSTANCE:
254 *result = (void *)(intptr_t)instance;
255 break;
256 case DDI_INFO_DEVT2DEVINFO:
257 if (pcieb == NULL) {
258 ret = DDI_FAILURE;
259 break;
260 }
261
262 *result = (void *)pcieb->pcieb_dip;
263 break;
264 default:
265 ret = DDI_FAILURE;
266 break;
267 }
268
269 return (ret);
270 }
271
272
273 /*ARGSUSED*/
274 static int
pcieb_probe(dev_info_t * devi)275 pcieb_probe(dev_info_t *devi)
276 {
277 return (DDI_PROBE_SUCCESS);
278 }
279
280 /*
281 * This is a workaround for an undocumented HW erratum with the
282 * multi-function, F0 and F2, Intel 41210 PCIe-to-PCI bridge. When
283 * Fn (cdip) attaches, this workaround is called to initialize Fn's
284 * sibling (sdip) with MPS/MRRS if it isn't already configured.
285 * Doing so prevents a malformed TLP panic.
286 */
287 static void
pcieb_41210_mps_wkrnd(dev_info_t * cdip)288 pcieb_41210_mps_wkrnd(dev_info_t *cdip)
289 {
290 dev_info_t *sdip;
291 ddi_acc_handle_t cfg_hdl;
292 uint16_t cdip_dev_ctrl, cdip_mrrs_mps;
293 pcie_bus_t *cdip_bus_p = PCIE_DIP2BUS(cdip);
294
295 /* Get cdip's MPS/MRRS already setup by pcie_initchild_mps() */
296 ASSERT(cdip_bus_p);
297 cdip_dev_ctrl = PCIE_CAP_GET(16, cdip_bus_p, PCIE_DEVCTL);
298 cdip_mrrs_mps = cdip_dev_ctrl &
299 (PCIE_DEVCTL_MAX_READ_REQ_MASK | PCIE_DEVCTL_MAX_PAYLOAD_MASK);
300
301 /* Locate sdip and set its MPS/MRRS when applicable */
302 for (sdip = ddi_get_child(ddi_get_parent(cdip)); sdip;
303 sdip = ddi_get_next_sibling(sdip)) {
304 uint16_t sdip_dev_ctrl, sdip_mrrs_mps, cap_ptr;
305 uint32_t bus_dev_ven_id;
306
307 if (sdip == cdip || pci_config_setup(sdip, &cfg_hdl)
308 != DDI_SUCCESS)
309 continue;
310
311 /* must be an Intel 41210 bridge */
312 bus_dev_ven_id = pci_config_get32(cfg_hdl, PCI_CONF_VENID);
313 if (!PCIEB_IS_41210_BRIDGE(bus_dev_ven_id)) {
314 pci_config_teardown(&cfg_hdl);
315 continue;
316 }
317
318 if (PCI_CAP_LOCATE(cfg_hdl, PCI_CAP_ID_PCI_E, &cap_ptr)
319 != DDI_SUCCESS) {
320 pci_config_teardown(&cfg_hdl);
321 continue;
322 }
323
324 /* get sdip's MPS/MRRS to compare to cdip's */
325 sdip_dev_ctrl = PCI_CAP_GET16(cfg_hdl, NULL, cap_ptr,
326 PCIE_DEVCTL);
327 sdip_mrrs_mps = sdip_dev_ctrl &
328 (PCIE_DEVCTL_MAX_READ_REQ_MASK |
329 PCIE_DEVCTL_MAX_PAYLOAD_MASK);
330
331 /* if sdip already attached then its MPS/MRRS is configured */
332 if (i_ddi_devi_attached(sdip)) {
333 ASSERT(sdip_mrrs_mps == cdip_mrrs_mps);
334 pci_config_teardown(&cfg_hdl);
335 continue;
336 }
337
338 /* otherwise, update sdip's MPS/MRRS if different from cdip's */
339 if (sdip_mrrs_mps != cdip_mrrs_mps) {
340 sdip_dev_ctrl = (sdip_dev_ctrl &
341 ~(PCIE_DEVCTL_MAX_READ_REQ_MASK |
342 PCIE_DEVCTL_MAX_PAYLOAD_MASK)) | cdip_mrrs_mps;
343
344 PCI_CAP_PUT16(cfg_hdl, NULL, cap_ptr, PCIE_DEVCTL,
345 sdip_dev_ctrl);
346 }
347
348 /*
349 * note: sdip's bus_mps will be updated by
350 * pcie_initchild_mps()
351 */
352
353 pci_config_teardown(&cfg_hdl);
354
355 break;
356 }
357 }
358
359 static int
pcieb_attach(dev_info_t * devi,ddi_attach_cmd_t cmd)360 pcieb_attach(dev_info_t *devi, ddi_attach_cmd_t cmd)
361 {
362 int instance;
363 char device_type[8];
364 pcieb_devstate_t *pcieb;
365 pcie_bus_t *bus_p = PCIE_DIP2UPBUS(devi);
366 ddi_acc_handle_t config_handle = bus_p->bus_cfg_hdl;
367
368 switch (cmd) {
369 case DDI_RESUME:
370 (void) pcie_pwr_resume(devi);
371 return (DDI_SUCCESS);
372
373 default:
374 return (DDI_FAILURE);
375
376 case DDI_ATTACH:
377 break;
378 }
379
380 if (!(PCIE_IS_BDG(bus_p))) {
381 PCIEB_DEBUG(DBG_ATTACH, devi, "This is not a switch or"
382 " bridge\n");
383 return (DDI_FAILURE);
384 }
385
386 /*
387 * If PCIE_LINKCTL_LINK_DISABLE bit in the PCIe Config
388 * Space (PCIe Capability Link Control Register) is set,
389 * then do not bind the driver.
390 */
391 if (PCIE_CAP_GET(16, bus_p, PCIE_LINKCTL) & PCIE_LINKCTL_LINK_DISABLE)
392 return (DDI_FAILURE);
393
394 /*
395 * Allocate and get soft state structure.
396 */
397 instance = ddi_get_instance(devi);
398 if (ddi_soft_state_zalloc(pcieb_state, instance) != DDI_SUCCESS)
399 return (DDI_FAILURE);
400 pcieb = ddi_get_soft_state(pcieb_state, instance);
401 pcieb->pcieb_dip = devi;
402
403 if ((pcieb_fm_init(pcieb)) != DDI_SUCCESS) {
404 PCIEB_DEBUG(DBG_ATTACH, devi, "Failed in pcieb_fm_init\n");
405 goto fail;
406 }
407 pcieb->pcieb_init_flags |= PCIEB_INIT_FM;
408
409 mutex_init(&pcieb->pcieb_mutex, NULL, MUTEX_DRIVER, NULL);
410 mutex_init(&pcieb->pcieb_err_mutex, NULL, MUTEX_DRIVER,
411 (void *)pcieb->pcieb_fm_ibc);
412 mutex_init(&pcieb->pcieb_peek_poke_mutex, NULL, MUTEX_DRIVER,
413 (void *)pcieb->pcieb_fm_ibc);
414
415 /* create special properties for device identification */
416 pcieb_id_props(pcieb);
417
418 /*
419 * Power management setup. This also makes sure that switch/bridge
420 * is at D0 during attach.
421 */
422 if (pwr_common_setup(devi) != DDI_SUCCESS) {
423 PCIEB_DEBUG(DBG_PWR, devi, "pwr_common_setup failed\n");
424 goto fail;
425 }
426
427 if (pcieb_pwr_setup(devi) != DDI_SUCCESS) {
428 PCIEB_DEBUG(DBG_PWR, devi, "pxb_pwr_setup failed \n");
429 goto fail;
430 }
431
432 /*
433 * Make sure the "device_type" property exists.
434 */
435 if (pcieb_is_pcie_device_type(devi))
436 (void) strcpy(device_type, "pciex");
437 else
438 (void) strcpy(device_type, "pci");
439
440 (void) ddi_prop_update_string(DDI_DEV_T_NONE, devi,
441 "device_type", device_type);
442
443 /*
444 * Check whether the "ranges" property is present.
445 * Otherwise create the ranges property by reading
446 * the configuration registers
447 */
448 if (ddi_prop_exists(DDI_DEV_T_ANY, devi, DDI_PROP_DONTPASS,
449 "ranges") == 0) {
450 pcieb_create_ranges_prop(devi, config_handle);
451 }
452
453 if (PCIE_IS_PCI_BDG(bus_p))
454 pcieb_set_pci_perf_parameters(devi, config_handle);
455
456 #ifdef PX_PLX
457 pcieb_attach_plx_workarounds(pcieb);
458 #endif /* PX_PLX */
459
460 if (pcie_init(devi, NULL) != DDI_SUCCESS)
461 goto fail;
462
463 /* Intel PCIe-to-PCI 41210 bridge workaround -- if applicable */
464 if (pcieb_disable_41210_wkarnd == 0 &&
465 PCIEB_IS_41210_BRIDGE(bus_p->bus_dev_ven_id))
466 pcieb_41210_mps_wkrnd(devi);
467
468 /*
469 * Initialize interrupt handlers. Ignore return value.
470 */
471 (void) pcieb_intr_attach(pcieb);
472
473 (void) pcie_hpintr_enable(devi);
474
475 /* Do any platform specific workarounds needed at this time */
476 pcieb_plat_attach_workaround(devi);
477
478 /*
479 * If this is a root port, determine and set the max payload size.
480 * Since this will involve scanning the fabric, all error enabling
481 * and sw workarounds should be in place before doing this.
482 */
483 if (PCIE_IS_RP(bus_p))
484 pcie_init_root_port_mps(devi);
485
486 ddi_report_dev(devi);
487 return (DDI_SUCCESS);
488
489 fail:
490 (void) pcieb_detach(devi, DDI_DETACH);
491 return (DDI_FAILURE);
492 }
493
494 static int
pcieb_detach(dev_info_t * devi,ddi_detach_cmd_t cmd)495 pcieb_detach(dev_info_t *devi, ddi_detach_cmd_t cmd)
496 {
497 pcieb_devstate_t *pcieb;
498 int error = DDI_SUCCESS;
499
500 switch (cmd) {
501 case DDI_SUSPEND:
502 error = pcie_pwr_suspend(devi);
503 return (error);
504
505 case DDI_DETACH:
506 break;
507
508 default:
509 return (DDI_FAILURE);
510 }
511
512 pcieb = ddi_get_soft_state(pcieb_state, ddi_get_instance(devi));
513
514 /* disable hotplug interrupt */
515 (void) pcie_hpintr_disable(devi);
516
517 /* remove interrupt handlers */
518 pcieb_intr_fini(pcieb);
519
520 /* uninitialize inband PCI-E HPC if present */
521 (void) pcie_uninit(devi);
522
523 (void) ddi_prop_remove(DDI_DEV_T_NONE, devi, "device_type");
524
525 (void) ndi_prop_remove(DDI_DEV_T_NONE, pcieb->pcieb_dip,
526 "pcie_ce_mask");
527
528 if (pcieb->pcieb_init_flags & PCIEB_INIT_FM)
529 pcieb_fm_fini(pcieb);
530
531 pcieb_pwr_teardown(devi);
532 pwr_common_teardown(devi);
533
534 mutex_destroy(&pcieb->pcieb_peek_poke_mutex);
535 mutex_destroy(&pcieb->pcieb_err_mutex);
536 mutex_destroy(&pcieb->pcieb_mutex);
537
538 /*
539 * And finally free the per-pci soft state.
540 */
541 ddi_soft_state_free(pcieb_state, ddi_get_instance(devi));
542
543 return (DDI_SUCCESS);
544 }
545
546 static int
pcieb_bus_map(dev_info_t * dip,dev_info_t * rdip,ddi_map_req_t * mp,off_t offset,off_t len,caddr_t * vaddrp)547 pcieb_bus_map(dev_info_t *dip, dev_info_t *rdip, ddi_map_req_t *mp,
548 off_t offset, off_t len, caddr_t *vaddrp)
549 {
550 dev_info_t *pdip;
551
552 if (PCIE_IS_RP(PCIE_DIP2BUS(dip)) && mp->map_handlep != NULL) {
553 ddi_acc_impl_t *hdlp =
554 (ddi_acc_impl_t *)(mp->map_handlep)->ah_platform_private;
555
556 pcieb_set_prot_scan(dip, hdlp);
557 }
558 pdip = (dev_info_t *)DEVI(dip)->devi_parent;
559 return ((DEVI(pdip)->devi_ops->devo_bus_ops->bus_map)(pdip, rdip, mp,
560 offset, len, vaddrp));
561 }
562
563 static int
pcieb_ctlops(dev_info_t * dip,dev_info_t * rdip,ddi_ctl_enum_t ctlop,void * arg,void * result)564 pcieb_ctlops(dev_info_t *dip, dev_info_t *rdip, ddi_ctl_enum_t ctlop,
565 void *arg, void *result)
566 {
567 pci_regspec_t *drv_regp;
568 int reglen;
569 int rn;
570 int totreg;
571 pcieb_devstate_t *pcieb = ddi_get_soft_state(pcieb_state,
572 ddi_get_instance(dip));
573 struct detachspec *ds;
574 struct attachspec *as;
575
576 switch (ctlop) {
577 case DDI_CTLOPS_REPORTDEV:
578 if (rdip == (dev_info_t *)0)
579 return (DDI_FAILURE);
580
581 if (ddi_get_parent(rdip) == dip) {
582 cmn_err(CE_CONT, "?PCIE-device: %s@%s, %s%d\n",
583 ddi_node_name(rdip), ddi_get_name_addr(rdip),
584 ddi_driver_name(rdip), ddi_get_instance(rdip));
585 }
586
587 /* Pass it up for fabric sync */
588 (void) ddi_ctlops(dip, rdip, ctlop, arg, result);
589 return (DDI_SUCCESS);
590
591 case DDI_CTLOPS_INITCHILD:
592 return (pcieb_initchild((dev_info_t *)arg));
593
594 case DDI_CTLOPS_UNINITCHILD:
595 pcieb_uninitchild((dev_info_t *)arg);
596 return (DDI_SUCCESS);
597
598 case DDI_CTLOPS_SIDDEV:
599 return (DDI_SUCCESS);
600
601 case DDI_CTLOPS_REGSIZE:
602 case DDI_CTLOPS_NREGS:
603 if (rdip == (dev_info_t *)0)
604 return (DDI_FAILURE);
605 break;
606
607 case DDI_CTLOPS_PEEK:
608 case DDI_CTLOPS_POKE:
609 return (pcieb_plat_peekpoke(dip, rdip, ctlop, arg, result));
610 case DDI_CTLOPS_ATTACH:
611 if (!pcie_is_child(dip, rdip))
612 return (DDI_SUCCESS);
613
614 as = (struct attachspec *)arg;
615 switch (as->when) {
616 case DDI_PRE:
617 if (as->cmd == DDI_RESUME) {
618 pcie_clear_errors(rdip);
619 if (pcieb_plat_ctlops(rdip, ctlop, arg) !=
620 DDI_SUCCESS)
621 return (DDI_FAILURE);
622 }
623
624 if (as->cmd == DDI_ATTACH)
625 return (pcie_pm_hold(dip));
626
627 return (DDI_SUCCESS);
628
629 case DDI_POST:
630 if (as->cmd == DDI_ATTACH &&
631 as->result != DDI_SUCCESS) {
632 /*
633 * Attach failed for the child device. The child
634 * driver may have made PM calls before the
635 * attach failed. pcie_pm_remove_child() should
636 * cleanup PM state and holds (if any)
637 * associated with the child device.
638 */
639 return (pcie_pm_remove_child(dip, rdip));
640 }
641
642 if (as->result == DDI_SUCCESS) {
643 pf_init(rdip, (void *)pcieb->pcieb_fm_ibc,
644 as->cmd);
645
646 (void) pcieb_plat_ctlops(rdip, ctlop, arg);
647 }
648
649 /*
650 * For empty hotplug-capable slots, we should explicitly
651 * disable the errors, so that we won't panic upon
652 * unsupported hotplug messages.
653 */
654 if ((!ddi_prop_exists(DDI_DEV_T_ANY, rdip,
655 DDI_PROP_DONTPASS, "hotplug-capable")) ||
656 ddi_get_child(rdip)) {
657 (void) pcie_postattach_child(rdip);
658 return (DDI_SUCCESS);
659 }
660
661 pcie_disable_errors(rdip);
662
663 return (DDI_SUCCESS);
664 default:
665 break;
666 }
667 return (DDI_SUCCESS);
668
669 case DDI_CTLOPS_DETACH:
670 if (!pcie_is_child(dip, rdip))
671 return (DDI_SUCCESS);
672
673 ds = (struct detachspec *)arg;
674 switch (ds->when) {
675 case DDI_PRE:
676 pf_fini(rdip, ds->cmd);
677 return (DDI_SUCCESS);
678
679 case DDI_POST:
680 if (pcieb_plat_ctlops(rdip, ctlop, arg) != DDI_SUCCESS)
681 return (DDI_FAILURE);
682 if (ds->cmd == DDI_DETACH &&
683 ds->result == DDI_SUCCESS) {
684 return (pcie_pm_remove_child(dip, rdip));
685 }
686 return (DDI_SUCCESS);
687 default:
688 break;
689 }
690 return (DDI_SUCCESS);
691 default:
692 return (ddi_ctlops(dip, rdip, ctlop, arg, result));
693 }
694
695 *(int *)result = 0;
696 if (ddi_getlongprop(DDI_DEV_T_ANY, rdip,
697 DDI_PROP_DONTPASS | DDI_PROP_CANSLEEP, "reg", (caddr_t)&drv_regp,
698 ®len) != DDI_SUCCESS)
699 return (DDI_FAILURE);
700
701 totreg = reglen / sizeof (pci_regspec_t);
702 if (ctlop == DDI_CTLOPS_NREGS)
703 *(int *)result = totreg;
704 else if (ctlop == DDI_CTLOPS_REGSIZE) {
705 rn = *(int *)arg;
706 if (rn >= totreg) {
707 kmem_free(drv_regp, reglen);
708 return (DDI_FAILURE);
709 }
710
711 *(off_t *)result = drv_regp[rn].pci_size_low |
712 ((uint64_t)drv_regp[rn].pci_size_hi << 32);
713 }
714
715 kmem_free(drv_regp, reglen);
716 return (DDI_SUCCESS);
717 }
718
719 /*
720 * name_child
721 *
722 * This function is called from init_child to name a node. It is
723 * also passed as a callback for node merging functions.
724 *
725 * return value: DDI_SUCCESS, DDI_FAILURE
726 */
727 static int
pcieb_name_child(dev_info_t * child,char * name,int namelen)728 pcieb_name_child(dev_info_t *child, char *name, int namelen)
729 {
730 pci_regspec_t *pci_rp;
731 uint_t device, func;
732 char **unit_addr;
733 uint_t n;
734
735 /*
736 * For .conf nodes, use unit-address property as name
737 */
738 if (ndi_dev_is_persistent_node(child) == 0) {
739 if (ddi_prop_lookup_string_array(DDI_DEV_T_ANY, child,
740 DDI_PROP_DONTPASS, "unit-address", &unit_addr, &n) !=
741 DDI_PROP_SUCCESS) {
742 cmn_err(CE_WARN,
743 "cannot find unit-address in %s.conf",
744 ddi_driver_name(child));
745 return (DDI_FAILURE);
746 }
747 if (n != 1 || *unit_addr == NULL || **unit_addr == 0) {
748 cmn_err(CE_WARN, "unit-address property in %s.conf"
749 " not well-formed", ddi_driver_name(child));
750 ddi_prop_free(unit_addr);
751 return (DDI_FAILURE);
752 }
753 (void) snprintf(name, namelen, "%s", *unit_addr);
754 ddi_prop_free(unit_addr);
755 return (DDI_SUCCESS);
756 }
757
758 /*
759 * Get the address portion of the node name based on
760 * the function and device number.
761 */
762 if (ddi_prop_lookup_int_array(DDI_DEV_T_ANY, child,
763 DDI_PROP_DONTPASS, "reg", (int **)&pci_rp, &n) != DDI_SUCCESS) {
764 return (DDI_FAILURE);
765 }
766
767 /* copy the device identifications */
768 device = PCI_REG_DEV_G(pci_rp[0].pci_phys_hi);
769 func = PCI_REG_FUNC_G(pci_rp[0].pci_phys_hi);
770
771 if (pcie_ari_is_enabled(ddi_get_parent(child))
772 == PCIE_ARI_FORW_ENABLED) {
773 func = (device << 3) | func;
774 device = 0;
775 }
776
777 if (func != 0)
778 (void) snprintf(name, namelen, "%x,%x", device, func);
779 else
780 (void) snprintf(name, namelen, "%x", device);
781
782 ddi_prop_free(pci_rp);
783 return (DDI_SUCCESS);
784 }
785
786 static int
pcieb_initchild(dev_info_t * child)787 pcieb_initchild(dev_info_t *child)
788 {
789 char name[MAXNAMELEN];
790 int result = DDI_FAILURE;
791 pcieb_devstate_t *pcieb =
792 (pcieb_devstate_t *)ddi_get_soft_state(pcieb_state,
793 ddi_get_instance(ddi_get_parent(child)));
794
795 /*
796 * Name the child
797 */
798 if (pcieb_name_child(child, name, MAXNAMELEN) != DDI_SUCCESS) {
799 result = DDI_FAILURE;
800 goto done;
801 }
802 ddi_set_name_addr(child, name);
803
804 /*
805 * Pseudo nodes indicate a prototype node with per-instance
806 * properties to be merged into the real h/w device node.
807 * The interpretation of the unit-address is DD[,F]
808 * where DD is the device id and F is the function.
809 */
810 if (ndi_dev_is_persistent_node(child) == 0) {
811 extern int pci_allow_pseudo_children;
812
813 /*
814 * Try to merge the properties from this prototype
815 * node into real h/w nodes.
816 */
817 if (ndi_merge_node(child, pcieb_name_child) == DDI_SUCCESS) {
818 /*
819 * Merged ok - return failure to remove the node.
820 */
821 ddi_set_name_addr(child, NULL);
822 result = DDI_FAILURE;
823 goto done;
824 }
825
826 /* workaround for ddivs to run under PCI-E */
827 if (pci_allow_pseudo_children) {
828 result = DDI_SUCCESS;
829 goto done;
830 }
831
832 /*
833 * The child was not merged into a h/w node,
834 * but there's not much we can do with it other
835 * than return failure to cause the node to be removed.
836 */
837 cmn_err(CE_WARN, "!%s@%s: %s.conf properties not merged",
838 ddi_driver_name(child), ddi_get_name_addr(child),
839 ddi_driver_name(child));
840 ddi_set_name_addr(child, NULL);
841 result = DDI_NOT_WELL_FORMED;
842 goto done;
843 }
844
845 /* platform specific initchild */
846 pcieb_plat_initchild(child);
847
848 if (pcie_pm_hold(pcieb->pcieb_dip) != DDI_SUCCESS) {
849 PCIEB_DEBUG(DBG_PWR, pcieb->pcieb_dip,
850 "INITCHILD: px_pm_hold failed\n");
851 result = DDI_FAILURE;
852 goto done;
853 }
854 /* Any return from here must call pcie_pm_release */
855
856 /*
857 * If configuration registers were previously saved by
858 * child (before it entered D3), then let the child do the
859 * restore to set up the config regs as it'll first need to
860 * power the device out of D3.
861 */
862 if (ddi_prop_exists(DDI_DEV_T_ANY, child, DDI_PROP_DONTPASS,
863 "config-regs-saved-by-child") == 1) {
864 PCIEB_DEBUG(DBG_PWR, ddi_get_parent(child),
865 "INITCHILD: config regs to be restored by child"
866 " for %s@%s\n", ddi_node_name(child),
867 ddi_get_name_addr(child));
868
869 result = DDI_SUCCESS;
870 goto cleanup;
871 }
872
873 PCIEB_DEBUG(DBG_PWR, ddi_get_parent(child),
874 "INITCHILD: config regs setup for %s@%s\n",
875 ddi_node_name(child), ddi_get_name_addr(child));
876
877 pcie_init_dom(child);
878
879 if (pcie_initchild(child) != DDI_SUCCESS) {
880 result = DDI_FAILURE;
881 pcie_fini_dom(child);
882 goto cleanup;
883 }
884
885 #ifdef PX_PLX
886 if (pcieb_init_plx_workarounds(pcieb, child) == DDI_FAILURE) {
887 result = DDI_FAILURE;
888 pcie_fini_dom(child);
889 goto cleanup;
890 }
891 #endif /* PX_PLX */
892
893 result = DDI_SUCCESS;
894 cleanup:
895 pcie_pm_release(pcieb->pcieb_dip);
896 done:
897 return (result);
898 }
899
900 static void
pcieb_uninitchild(dev_info_t * dip)901 pcieb_uninitchild(dev_info_t *dip)
902 {
903
904 pcie_uninitchild(dip);
905
906 pcieb_plat_uninitchild(dip);
907
908 ddi_set_name_addr(dip, NULL);
909
910 /*
911 * Strip the node to properly convert it back to prototype form
912 */
913 ddi_remove_minor_node(dip, NULL);
914
915 ddi_prop_remove_all(dip);
916 }
917
918 static boolean_t
pcieb_is_pcie_device_type(dev_info_t * dip)919 pcieb_is_pcie_device_type(dev_info_t *dip)
920 {
921 pcie_bus_t *bus_p = PCIE_DIP2BUS(dip);
922
923 if (PCIE_IS_SW(bus_p) || PCIE_IS_RP(bus_p) || PCIE_IS_PCI2PCIE(bus_p))
924 return (B_TRUE);
925
926 return (B_FALSE);
927 }
928
929 static int
pcieb_intr_attach(pcieb_devstate_t * pcieb)930 pcieb_intr_attach(pcieb_devstate_t *pcieb)
931 {
932 int intr_types;
933 dev_info_t *dip = pcieb->pcieb_dip;
934
935 /* Allow platform specific code to do any initialization first */
936 pcieb_plat_intr_attach(pcieb);
937
938 /*
939 * Initialize interrupt handlers.
940 * If both MSI and FIXED are supported, try to attach MSI first.
941 * If MSI fails for any reason, then try FIXED, but only allow one
942 * type to be attached.
943 */
944 if (ddi_intr_get_supported_types(dip, &intr_types) != DDI_SUCCESS) {
945 PCIEB_DEBUG(DBG_ATTACH, dip, "ddi_intr_get_supported_types"
946 " failed\n");
947 goto FAIL;
948 }
949
950 if ((intr_types & DDI_INTR_TYPE_MSI) &&
951 (pcieb_msi_supported(dip) == DDI_SUCCESS)) {
952 if (pcieb_intr_init(pcieb, DDI_INTR_TYPE_MSI) == DDI_SUCCESS)
953 intr_types = DDI_INTR_TYPE_MSI;
954 else {
955 PCIEB_DEBUG(DBG_ATTACH, dip, "Unable to attach MSI"
956 " handler\n");
957 }
958 }
959
960 if (intr_types != DDI_INTR_TYPE_MSI) {
961 /*
962 * MSIs are not supported or MSI initialization failed. For Root
963 * Ports mark this so error handling might try to fallback to
964 * some other mechanism if available (machinecheck etc.).
965 */
966 if (PCIE_IS_RP(PCIE_DIP2UPBUS(dip)))
967 pcieb->pcieb_no_aer_msi = B_TRUE;
968 }
969
970 if (intr_types & DDI_INTR_TYPE_FIXED) {
971 if (pcieb_intr_init(pcieb, DDI_INTR_TYPE_FIXED) !=
972 DDI_SUCCESS) {
973 PCIEB_DEBUG(DBG_ATTACH, dip,
974 "Unable to attach INTx handler\n");
975 goto FAIL;
976 }
977 }
978 return (DDI_SUCCESS);
979
980 FAIL:
981 return (DDI_FAILURE);
982 }
983
984 /*
985 * This function initializes internally generated interrupts only.
986 * It does not affect any interrupts generated by downstream devices
987 * or the forwarding of them.
988 *
989 * Enable Device Specific Interrupts or Hotplug features here.
990 * Enabling features may change how many interrupts are requested
991 * by the device. If features are not enabled first, the
992 * device might not ask for any interrupts.
993 */
994
995 static int
pcieb_intr_init(pcieb_devstate_t * pcieb,int intr_type)996 pcieb_intr_init(pcieb_devstate_t *pcieb, int intr_type)
997 {
998 dev_info_t *dip = pcieb->pcieb_dip;
999 int nintrs, request, count, x;
1000 int intr_cap = 0;
1001 int inum = 0;
1002 int ret, hp_msi_off;
1003 pcie_bus_t *bus_p = PCIE_DIP2UPBUS(dip);
1004 uint16_t vendorid = bus_p->bus_dev_ven_id & 0xFFFF;
1005 boolean_t is_hp = B_FALSE;
1006 boolean_t is_pme = B_FALSE;
1007
1008 PCIEB_DEBUG(DBG_ATTACH, dip, "pcieb_intr_init: Attaching %s handler\n",
1009 (intr_type == DDI_INTR_TYPE_MSI) ? "MSI" : "INTx");
1010
1011 request = 0;
1012 if (PCIE_IS_HOTPLUG_ENABLED(dip)) {
1013 request++;
1014 is_hp = B_TRUE;
1015 }
1016
1017 /*
1018 * Hotplug and PME share the same MSI vector. If hotplug is not
1019 * supported check if MSI is needed for PME.
1020 */
1021 if ((intr_type == DDI_INTR_TYPE_MSI) && PCIE_IS_RP(bus_p) &&
1022 (vendorid == NVIDIA_VENDOR_ID)) {
1023 is_pme = B_TRUE;
1024 if (!is_hp)
1025 request++;
1026 }
1027
1028 /*
1029 * Setup MSI if this device is a Rootport and has AER. Currently no
1030 * SPARC Root Port supports fabric errors being reported through it.
1031 */
1032 if (intr_type == DDI_INTR_TYPE_MSI) {
1033 if (PCIE_IS_RP(bus_p) && PCIE_HAS_AER(bus_p))
1034 request++;
1035 }
1036
1037 if (request == 0)
1038 return (DDI_SUCCESS);
1039
1040 /*
1041 * Get number of supported interrupts.
1042 *
1043 * Several Bridges/Switches will not have this property set, resulting
1044 * in a FAILURE, if the device is not configured in a way that
1045 * interrupts are needed. (eg. hotplugging)
1046 */
1047 ret = ddi_intr_get_nintrs(dip, intr_type, &nintrs);
1048 if ((ret != DDI_SUCCESS) || (nintrs == 0)) {
1049 PCIEB_DEBUG(DBG_ATTACH, dip, "ddi_intr_get_nintrs ret:%d"
1050 " req:%d\n", ret, nintrs);
1051 return (DDI_FAILURE);
1052 }
1053
1054 PCIEB_DEBUG(DBG_ATTACH, dip, "bdf 0x%x: ddi_intr_get_nintrs: nintrs %d",
1055 " request %d\n", bus_p->bus_bdf, nintrs, request);
1056
1057 if (request > nintrs)
1058 request = nintrs;
1059
1060 /* Allocate an array of interrupt handlers */
1061 pcieb->pcieb_htable_size = sizeof (ddi_intr_handle_t) * request;
1062 pcieb->pcieb_htable = kmem_zalloc(pcieb->pcieb_htable_size,
1063 KM_SLEEP);
1064 pcieb->pcieb_init_flags |= PCIEB_INIT_HTABLE;
1065
1066 ret = ddi_intr_alloc(dip, pcieb->pcieb_htable, intr_type, inum,
1067 request, &count, DDI_INTR_ALLOC_NORMAL);
1068 if ((ret != DDI_SUCCESS) || (count == 0)) {
1069 PCIEB_DEBUG(DBG_ATTACH, dip, "ddi_intr_alloc() ret: %d ask: %d"
1070 " actual: %d\n", ret, request, count);
1071 goto FAIL;
1072 }
1073 pcieb->pcieb_init_flags |= PCIEB_INIT_ALLOC;
1074
1075 /* Save the actual number of interrupts allocated */
1076 pcieb->pcieb_intr_count = count;
1077 if (count < request) {
1078 PCIEB_DEBUG(DBG_ATTACH, dip, "bdf 0%x: Requested Intr: %d"
1079 " Received: %d\n", bus_p->bus_bdf, request, count);
1080 }
1081
1082 /*
1083 * NVidia (MCP55 and other) chipsets have a errata that if the number
1084 * of requested MSI intrs is not allocated we have to fall back to INTx.
1085 */
1086 if (intr_type == DDI_INTR_TYPE_MSI) {
1087 if (PCIE_IS_RP(bus_p) && (vendorid == NVIDIA_VENDOR_ID)) {
1088 if (request != count)
1089 goto FAIL;
1090 }
1091 }
1092
1093 /* Get interrupt priority */
1094 ret = ddi_intr_get_pri(pcieb->pcieb_htable[0],
1095 &pcieb->pcieb_intr_priority);
1096 if (ret != DDI_SUCCESS) {
1097 PCIEB_DEBUG(DBG_ATTACH, dip, "ddi_intr_get_pri() ret: %d\n",
1098 ret);
1099 goto FAIL;
1100 }
1101
1102 if (pcieb->pcieb_intr_priority >= LOCK_LEVEL) {
1103 pcieb->pcieb_intr_priority = LOCK_LEVEL - 1;
1104 ret = ddi_intr_set_pri(pcieb->pcieb_htable[0],
1105 pcieb->pcieb_intr_priority);
1106 if (ret != DDI_SUCCESS) {
1107 PCIEB_DEBUG(DBG_ATTACH, dip, "ddi_intr_set_pri() ret:"
1108 " %d\n", ret);
1109
1110 goto FAIL;
1111 }
1112 }
1113
1114 mutex_init(&pcieb->pcieb_intr_mutex, NULL, MUTEX_DRIVER, NULL);
1115
1116 pcieb->pcieb_init_flags |= PCIEB_INIT_MUTEX;
1117
1118 for (count = 0; count < pcieb->pcieb_intr_count; count++) {
1119 ret = ddi_intr_add_handler(pcieb->pcieb_htable[count],
1120 pcieb_intr_handler, (caddr_t)pcieb,
1121 (caddr_t)(uintptr_t)(inum + count));
1122
1123 if (ret != DDI_SUCCESS) {
1124 PCIEB_DEBUG(DBG_ATTACH, dip, "Cannot add "
1125 "interrupt(%d)\n", ret);
1126 break;
1127 }
1128 }
1129
1130 /* If unsucessful, remove the added handlers */
1131 if (ret != DDI_SUCCESS) {
1132 for (x = 0; x < count; x++) {
1133 (void) ddi_intr_remove_handler(pcieb->pcieb_htable[x]);
1134 }
1135 goto FAIL;
1136 }
1137
1138 pcieb->pcieb_init_flags |= PCIEB_INIT_HANDLER;
1139
1140 (void) ddi_intr_get_cap(pcieb->pcieb_htable[0], &intr_cap);
1141
1142 /*
1143 * Get this intr lock because we are not quite ready to handle
1144 * interrupts immediately after enabling it. The MSI multi register
1145 * gets programmed in ddi_intr_enable after which we need to get the
1146 * MSI offsets for Hotplug/AER.
1147 */
1148 mutex_enter(&pcieb->pcieb_intr_mutex);
1149
1150 if (intr_cap & DDI_INTR_FLAG_BLOCK) {
1151 (void) ddi_intr_block_enable(pcieb->pcieb_htable,
1152 pcieb->pcieb_intr_count);
1153 pcieb->pcieb_init_flags |= PCIEB_INIT_BLOCK;
1154 } else {
1155 for (count = 0; count < pcieb->pcieb_intr_count; count++) {
1156 (void) ddi_intr_enable(pcieb->pcieb_htable[count]);
1157 }
1158 }
1159 pcieb->pcieb_init_flags |= PCIEB_INIT_ENABLE;
1160
1161 /* Save the interrupt type */
1162 pcieb->pcieb_intr_type = intr_type;
1163
1164 /* Get the MSI offset for hotplug/PME from the PCIe cap reg */
1165 if (intr_type == DDI_INTR_TYPE_MSI) {
1166 hp_msi_off = PCI_CAP_GET16(bus_p->bus_cfg_hdl, NULL,
1167 bus_p->bus_pcie_off, PCIE_PCIECAP) &
1168 PCIE_PCIECAP_INT_MSG_NUM;
1169
1170 if (hp_msi_off >= count) {
1171 PCIEB_DEBUG(DBG_ATTACH, dip, "MSI number %d in PCIe "
1172 "cap > max allocated %d\n", hp_msi_off, count);
1173 mutex_exit(&pcieb->pcieb_intr_mutex);
1174 goto FAIL;
1175 }
1176
1177 if (is_hp)
1178 pcieb->pcieb_isr_tab[hp_msi_off] |= PCIEB_INTR_SRC_HP;
1179
1180 if (is_pme)
1181 pcieb->pcieb_isr_tab[hp_msi_off] |= PCIEB_INTR_SRC_PME;
1182 } else {
1183 /* INTx handles only Hotplug interrupts */
1184 if (is_hp)
1185 pcieb->pcieb_isr_tab[0] |= PCIEB_INTR_SRC_HP;
1186 }
1187
1188
1189 /*
1190 * Get the MSI offset for errors from the AER Root Error status
1191 * register.
1192 */
1193 if ((intr_type == DDI_INTR_TYPE_MSI) && PCIE_IS_RP(bus_p)) {
1194 if (PCIE_HAS_AER(bus_p)) {
1195 int aer_msi_off;
1196 aer_msi_off = (PCI_XCAP_GET32(bus_p->bus_cfg_hdl, NULL,
1197 bus_p->bus_aer_off, PCIE_AER_RE_STS) >>
1198 PCIE_AER_RE_STS_MSG_NUM_SHIFT) &
1199 PCIE_AER_RE_STS_MSG_NUM_MASK;
1200
1201 if (aer_msi_off >= count) {
1202 PCIEB_DEBUG(DBG_ATTACH, dip, "MSI number %d in"
1203 " AER cap > max allocated %d\n",
1204 aer_msi_off, count);
1205 mutex_exit(&pcieb->pcieb_intr_mutex);
1206 goto FAIL;
1207 }
1208 pcieb->pcieb_isr_tab[aer_msi_off] |= PCIEB_INTR_SRC_AER;
1209 } else {
1210 /*
1211 * This RP does not have AER. Fallback to the
1212 * SERR+Machinecheck approach if available.
1213 */
1214 pcieb->pcieb_no_aer_msi = B_TRUE;
1215 }
1216 }
1217
1218 mutex_exit(&pcieb->pcieb_intr_mutex);
1219 return (DDI_SUCCESS);
1220
1221 FAIL:
1222 pcieb_intr_fini(pcieb);
1223 return (DDI_FAILURE);
1224 }
1225
1226 static void
pcieb_intr_fini(pcieb_devstate_t * pcieb)1227 pcieb_intr_fini(pcieb_devstate_t *pcieb)
1228 {
1229 int x;
1230 int count = pcieb->pcieb_intr_count;
1231 int flags = pcieb->pcieb_init_flags;
1232
1233 if ((flags & PCIEB_INIT_ENABLE) &&
1234 (flags & PCIEB_INIT_BLOCK)) {
1235 (void) ddi_intr_block_disable(pcieb->pcieb_htable, count);
1236 flags &= ~(PCIEB_INIT_ENABLE |
1237 PCIEB_INIT_BLOCK);
1238 }
1239
1240 if (flags & PCIEB_INIT_MUTEX)
1241 mutex_destroy(&pcieb->pcieb_intr_mutex);
1242
1243 for (x = 0; x < count; x++) {
1244 if (flags & PCIEB_INIT_ENABLE)
1245 (void) ddi_intr_disable(pcieb->pcieb_htable[x]);
1246
1247 if (flags & PCIEB_INIT_HANDLER)
1248 (void) ddi_intr_remove_handler(pcieb->pcieb_htable[x]);
1249
1250 if (flags & PCIEB_INIT_ALLOC)
1251 (void) ddi_intr_free(pcieb->pcieb_htable[x]);
1252 }
1253
1254 flags &= ~(PCIEB_INIT_ENABLE | PCIEB_INIT_HANDLER | PCIEB_INIT_ALLOC |
1255 PCIEB_INIT_MUTEX);
1256
1257 if (flags & PCIEB_INIT_HTABLE)
1258 kmem_free(pcieb->pcieb_htable, pcieb->pcieb_htable_size);
1259
1260 flags &= ~PCIEB_INIT_HTABLE;
1261
1262 pcieb->pcieb_init_flags &= flags;
1263 }
1264
1265 /*
1266 * Checks if this device needs MSIs enabled or not.
1267 */
1268 /*ARGSUSED*/
1269 static int
pcieb_msi_supported(dev_info_t * dip)1270 pcieb_msi_supported(dev_info_t *dip)
1271 {
1272 return ((pcieb_enable_msi && pcieb_plat_msi_supported(dip)) ?
1273 DDI_SUCCESS: DDI_FAILURE);
1274 }
1275
1276 /*ARGSUSED*/
1277 static int
pcieb_fm_init_child(dev_info_t * dip,dev_info_t * tdip,int cap,ddi_iblock_cookie_t * ibc)1278 pcieb_fm_init_child(dev_info_t *dip, dev_info_t *tdip, int cap,
1279 ddi_iblock_cookie_t *ibc)
1280 {
1281 pcieb_devstate_t *pcieb = ddi_get_soft_state(pcieb_state,
1282 ddi_get_instance(dip));
1283
1284 ASSERT(ibc != NULL);
1285 *ibc = pcieb->pcieb_fm_ibc;
1286
1287 return (DEVI(dip)->devi_fmhdl->fh_cap | DDI_FM_ACCCHK_CAPABLE |
1288 DDI_FM_DMACHK_CAPABLE);
1289 }
1290
1291 static int
pcieb_fm_init(pcieb_devstate_t * pcieb_p)1292 pcieb_fm_init(pcieb_devstate_t *pcieb_p)
1293 {
1294 dev_info_t *dip = pcieb_p->pcieb_dip;
1295 int fm_cap = DDI_FM_EREPORT_CAPABLE;
1296
1297 /*
1298 * Request our capability level and get our parents capability
1299 * and ibc.
1300 */
1301 ddi_fm_init(dip, &fm_cap, &pcieb_p->pcieb_fm_ibc);
1302
1303 return (DDI_SUCCESS);
1304 }
1305
1306 /*
1307 * Breakdown our FMA resources
1308 */
1309 static void
pcieb_fm_fini(pcieb_devstate_t * pcieb_p)1310 pcieb_fm_fini(pcieb_devstate_t *pcieb_p)
1311 {
1312 /*
1313 * Clean up allocated fm structures
1314 */
1315 ddi_fm_fini(pcieb_p->pcieb_dip);
1316 }
1317
1318 static int
pcieb_open(dev_t * devp,int flags,int otyp,cred_t * credp)1319 pcieb_open(dev_t *devp, int flags, int otyp, cred_t *credp)
1320 {
1321 int inst = PCI_MINOR_NUM_TO_INSTANCE(getminor(*devp));
1322 pcieb_devstate_t *pcieb = ddi_get_soft_state(pcieb_state, inst);
1323 int rv;
1324
1325 if (pcieb == NULL)
1326 return (ENXIO);
1327
1328 mutex_enter(&pcieb->pcieb_mutex);
1329 rv = pcie_open(pcieb->pcieb_dip, devp, flags, otyp, credp);
1330 mutex_exit(&pcieb->pcieb_mutex);
1331
1332 return (rv);
1333 }
1334
1335 static int
pcieb_close(dev_t dev,int flags,int otyp,cred_t * credp)1336 pcieb_close(dev_t dev, int flags, int otyp, cred_t *credp)
1337 {
1338 int inst = PCI_MINOR_NUM_TO_INSTANCE(getminor(dev));
1339 pcieb_devstate_t *pcieb = ddi_get_soft_state(pcieb_state, inst);
1340 int rv;
1341
1342 if (pcieb == NULL)
1343 return (ENXIO);
1344
1345 mutex_enter(&pcieb->pcieb_mutex);
1346 rv = pcie_close(pcieb->pcieb_dip, dev, flags, otyp, credp);
1347 mutex_exit(&pcieb->pcieb_mutex);
1348
1349 return (rv);
1350 }
1351
1352 static int
pcieb_ioctl(dev_t dev,int cmd,intptr_t arg,int mode,cred_t * credp,int * rvalp)1353 pcieb_ioctl(dev_t dev, int cmd, intptr_t arg, int mode, cred_t *credp,
1354 int *rvalp)
1355 {
1356 int inst = PCI_MINOR_NUM_TO_INSTANCE(getminor(dev));
1357 pcieb_devstate_t *pcieb = ddi_get_soft_state(pcieb_state, inst);
1358 int rv;
1359
1360 if (pcieb == NULL)
1361 return (ENXIO);
1362
1363 /* To handle devctl and hotplug related ioctls */
1364 rv = pcie_ioctl(pcieb->pcieb_dip, dev, cmd, arg, mode, credp, rvalp);
1365
1366 return (rv);
1367 }
1368
1369 /*
1370 * Common interrupt handler for hotplug, PME and errors.
1371 */
1372 static uint_t
pcieb_intr_handler(caddr_t arg1,caddr_t arg2)1373 pcieb_intr_handler(caddr_t arg1, caddr_t arg2)
1374 {
1375 pcieb_devstate_t *pcieb_p = (pcieb_devstate_t *)arg1;
1376 dev_info_t *dip = pcieb_p->pcieb_dip;
1377 ddi_fm_error_t derr;
1378 int sts = 0;
1379 int ret = DDI_INTR_UNCLAIMED;
1380 int isrc;
1381
1382 if (!(pcieb_p->pcieb_init_flags & PCIEB_INIT_ENABLE))
1383 goto FAIL;
1384
1385 mutex_enter(&pcieb_p->pcieb_intr_mutex);
1386 isrc = pcieb_p->pcieb_isr_tab[(int)(uintptr_t)arg2];
1387 mutex_exit(&pcieb_p->pcieb_intr_mutex);
1388
1389 PCIEB_DEBUG(DBG_INTR, dip, "Received intr number %d\n",
1390 (int)(uintptr_t)arg2);
1391
1392 if (isrc == PCIEB_INTR_SRC_UNKNOWN)
1393 goto FAIL;
1394
1395 if (isrc & PCIEB_INTR_SRC_HP)
1396 ret = pcie_intr(dip);
1397
1398 if (isrc & PCIEB_INTR_SRC_PME)
1399 ret = DDI_INTR_CLAIMED;
1400
1401 /* AER Error */
1402 if (isrc & PCIEB_INTR_SRC_AER) {
1403 /*
1404 * If MSI is shared with PME/hotplug then check Root Error
1405 * Status Reg before claiming it. For now it's ok since
1406 * we know we get 2 MSIs.
1407 */
1408 ret = DDI_INTR_CLAIMED;
1409 bzero(&derr, sizeof (ddi_fm_error_t));
1410 derr.fme_version = DDI_FME_VERSION;
1411 mutex_enter(&pcieb_p->pcieb_peek_poke_mutex);
1412 mutex_enter(&pcieb_p->pcieb_err_mutex);
1413
1414 pf_eh_enter(PCIE_DIP2BUS(dip));
1415 PCIE_ROOT_EH_SRC(PCIE_DIP2PFD(dip))->intr_type =
1416 PF_INTR_TYPE_AER;
1417
1418 if ((DEVI(dip)->devi_fmhdl->fh_cap) & DDI_FM_EREPORT_CAPABLE)
1419 sts = pf_scan_fabric(dip, &derr, NULL);
1420 pf_eh_exit(PCIE_DIP2BUS(dip));
1421
1422 mutex_exit(&pcieb_p->pcieb_err_mutex);
1423 mutex_exit(&pcieb_p->pcieb_peek_poke_mutex);
1424 if (pcieb_die & sts)
1425 fm_panic("%s-%d: PCI(-X) Express Fatal Error. (0x%x)",
1426 ddi_driver_name(dip), ddi_get_instance(dip), sts);
1427 }
1428 FAIL:
1429 return (ret);
1430 }
1431
1432 /*
1433 * Some PCI-X to PCI-E bridges do not support full 64-bit addressing on the
1434 * PCI-X side of the bridge. We build a special version of this driver for
1435 * those bridges, which uses PCIEB_ADDR_LIMIT_LO and/or PCIEB_ADDR_LIMIT_HI
1436 * to define the range of values which the chip can handle. The code below
1437 * then clamps the DMA address range supplied by the driver, preventing the
1438 * PCI-E nexus driver from allocating any memory the bridge can't deal
1439 * with.
1440 */
1441 static int
pcieb_dma_allochdl(dev_info_t * dip,dev_info_t * rdip,ddi_dma_attr_t * attr_p,int (* waitfp)(caddr_t),caddr_t arg,ddi_dma_handle_t * handlep)1442 pcieb_dma_allochdl(dev_info_t *dip, dev_info_t *rdip,
1443 ddi_dma_attr_t *attr_p, int (*waitfp)(caddr_t), caddr_t arg,
1444 ddi_dma_handle_t *handlep)
1445 {
1446 int ret;
1447 #ifdef PCIEB_BCM
1448 uint64_t lim;
1449
1450 /*
1451 * If the leaf device's limits are outside than what the Broadcom
1452 * bridge can handle, we need to clip the values passed up the chain.
1453 */
1454 lim = attr_p->dma_attr_addr_lo;
1455 attr_p->dma_attr_addr_lo = MAX(lim, PCIEB_ADDR_LIMIT_LO);
1456
1457 lim = attr_p->dma_attr_addr_hi;
1458 attr_p->dma_attr_addr_hi = MIN(lim, PCIEB_ADDR_LIMIT_HI);
1459
1460 #endif /* PCIEB_BCM */
1461
1462 /*
1463 * This is a software workaround to fix the Broadcom 5714/5715 PCIe-PCI
1464 * bridge prefetch bug. Intercept the DMA alloc handle request and set
1465 * PX_DMAI_FLAGS_MAP_BUFZONE flag in the handle. If this flag is set,
1466 * the px nexus driver will allocate an extra page & make it valid one,
1467 * for any DVMA request that comes from any of the Broadcom bridge child
1468 * devices.
1469 */
1470 if ((ret = ddi_dma_allochdl(dip, rdip, attr_p, waitfp, arg,
1471 handlep)) == DDI_SUCCESS) {
1472 ddi_dma_impl_t *mp = (ddi_dma_impl_t *)*handlep;
1473 #ifdef PCIEB_BCM
1474 mp->dmai_inuse |= PX_DMAI_FLAGS_MAP_BUFZONE;
1475 #endif /* PCIEB_BCM */
1476 /*
1477 * For a given rdip, update mp->dmai_bdf with the bdf value
1478 * of pcieb's immediate child or secondary bus-id of the
1479 * PCIe2PCI bridge.
1480 */
1481 mp->dmai_minxfer = pcie_get_bdf_for_dma_xfer(dip, rdip);
1482 }
1483
1484 return (ret);
1485 }
1486
1487 /*
1488 * FDVMA feature is not supported for any child device of Broadcom 5714/5715
1489 * PCIe-PCI bridge due to prefetch bug. Return failure immediately, so that
1490 * these drivers will switch to regular DVMA path.
1491 */
1492 /*ARGSUSED*/
1493 static int
pcieb_dma_mctl(dev_info_t * dip,dev_info_t * rdip,ddi_dma_handle_t handle,enum ddi_dma_ctlops cmd,off_t * offp,size_t * lenp,caddr_t * objp,uint_t cache_flags)1494 pcieb_dma_mctl(dev_info_t *dip, dev_info_t *rdip, ddi_dma_handle_t handle,
1495 enum ddi_dma_ctlops cmd, off_t *offp, size_t *lenp, caddr_t *objp,
1496 uint_t cache_flags)
1497 {
1498 int ret;
1499
1500 #ifdef PCIEB_BCM
1501 if (cmd == DDI_DMA_RESERVE)
1502 return (DDI_FAILURE);
1503 #endif /* PCIEB_BCM */
1504
1505 if (((ret = ddi_dma_mctl(dip, rdip, handle, cmd, offp, lenp, objp,
1506 cache_flags)) == DDI_SUCCESS) && (cmd == DDI_DMA_RESERVE)) {
1507 ddi_dma_impl_t *mp = (ddi_dma_impl_t *)*objp;
1508
1509 /*
1510 * For a given rdip, update mp->dmai_bdf with the bdf value
1511 * of pcieb's immediate child or secondary bus-id of the
1512 * PCIe2PCI bridge.
1513 */
1514 mp->dmai_minxfer = pcie_get_bdf_for_dma_xfer(dip, rdip);
1515 }
1516
1517 return (ret);
1518 }
1519
1520 static int
pcieb_intr_ops(dev_info_t * dip,dev_info_t * rdip,ddi_intr_op_t intr_op,ddi_intr_handle_impl_t * hdlp,void * result)1521 pcieb_intr_ops(dev_info_t *dip, dev_info_t *rdip, ddi_intr_op_t intr_op,
1522 ddi_intr_handle_impl_t *hdlp, void *result)
1523 {
1524 return (pcieb_plat_intr_ops(dip, rdip, intr_op, hdlp, result));
1525
1526 }
1527
1528 /*
1529 * Power management related initialization specific to pcieb.
1530 * Called by pcieb_attach()
1531 */
1532 static int
pcieb_pwr_setup(dev_info_t * dip)1533 pcieb_pwr_setup(dev_info_t *dip)
1534 {
1535 char *comp_array[5];
1536 int i;
1537 ddi_acc_handle_t conf_hdl;
1538 uint16_t pmcap, cap_ptr;
1539 pcie_pwr_t *pwr_p;
1540
1541 /* Some platforms/devices may choose to disable PM */
1542 if (pcieb_plat_pwr_disable(dip)) {
1543 (void) pcieb_pwr_disable(dip);
1544 return (DDI_SUCCESS);
1545 }
1546
1547 ASSERT(PCIE_PMINFO(dip));
1548 pwr_p = PCIE_NEXUS_PMINFO(dip);
1549 ASSERT(pwr_p);
1550
1551 /* Code taken from pci_pci driver */
1552 if (pci_config_setup(dip, &pwr_p->pwr_conf_hdl) != DDI_SUCCESS) {
1553 PCIEB_DEBUG(DBG_PWR, dip, "pcieb_pwr_setup: pci_config_setup "
1554 "failed\n");
1555 return (DDI_FAILURE);
1556 }
1557 conf_hdl = pwr_p->pwr_conf_hdl;
1558
1559 /*
1560 * Walk the capabilities searching for a PM entry.
1561 */
1562 if ((PCI_CAP_LOCATE(conf_hdl, PCI_CAP_ID_PM, &cap_ptr)) ==
1563 DDI_FAILURE) {
1564 PCIEB_DEBUG(DBG_PWR, dip, "switch/bridge does not support PM. "
1565 " PCI PM data structure not found in config header\n");
1566 pci_config_teardown(&conf_hdl);
1567 return (DDI_SUCCESS);
1568 }
1569 /*
1570 * Save offset to pmcsr for future references.
1571 */
1572 pwr_p->pwr_pmcsr_offset = cap_ptr + PCI_PMCSR;
1573 pmcap = PCI_CAP_GET16(conf_hdl, NULL, cap_ptr, PCI_PMCAP);
1574 if (pmcap & PCI_PMCAP_D1) {
1575 PCIEB_DEBUG(DBG_PWR, dip, "D1 state supported\n");
1576 pwr_p->pwr_pmcaps |= PCIE_SUPPORTS_D1;
1577 }
1578 if (pmcap & PCI_PMCAP_D2) {
1579 PCIEB_DEBUG(DBG_PWR, dip, "D2 state supported\n");
1580 pwr_p->pwr_pmcaps |= PCIE_SUPPORTS_D2;
1581 }
1582
1583 i = 0;
1584 comp_array[i++] = "NAME=PCIe switch/bridge PM";
1585 comp_array[i++] = "0=Power Off (D3)";
1586 if (pwr_p->pwr_pmcaps & PCIE_SUPPORTS_D2)
1587 comp_array[i++] = "1=D2";
1588 if (pwr_p->pwr_pmcaps & PCIE_SUPPORTS_D1)
1589 comp_array[i++] = "2=D1";
1590 comp_array[i++] = "3=Full Power D0";
1591
1592 /*
1593 * Create pm-components property, if it does not exist already.
1594 */
1595 if (ddi_prop_update_string_array(DDI_DEV_T_NONE, dip,
1596 "pm-components", comp_array, i) != DDI_PROP_SUCCESS) {
1597 PCIEB_DEBUG(DBG_PWR, dip, "could not create pm-components "
1598 " prop\n");
1599 pci_config_teardown(&conf_hdl);
1600 return (DDI_FAILURE);
1601 }
1602 return (pcieb_pwr_init_and_raise(dip, pwr_p));
1603 }
1604
1605 /*
1606 * undo whatever is done in pcieb_pwr_setup. called by pcieb_detach()
1607 */
1608 static void
pcieb_pwr_teardown(dev_info_t * dip)1609 pcieb_pwr_teardown(dev_info_t *dip)
1610 {
1611 pcie_pwr_t *pwr_p;
1612
1613 if (!PCIE_PMINFO(dip) || !(pwr_p = PCIE_NEXUS_PMINFO(dip)))
1614 return;
1615
1616 (void) ddi_prop_remove(DDI_DEV_T_NONE, dip, "pm-components");
1617 if (pwr_p->pwr_conf_hdl)
1618 pci_config_teardown(&pwr_p->pwr_conf_hdl);
1619 }
1620
1621 /*
1622 * Initializes the power level and raise the power to D0, if it is
1623 * not at D0.
1624 */
1625 static int
pcieb_pwr_init_and_raise(dev_info_t * dip,pcie_pwr_t * pwr_p)1626 pcieb_pwr_init_and_raise(dev_info_t *dip, pcie_pwr_t *pwr_p)
1627 {
1628 uint16_t pmcsr;
1629 int ret = DDI_SUCCESS;
1630
1631 /*
1632 * Intialize our power level from PMCSR. The common code initializes
1633 * this to UNKNOWN. There is no guarantee that we will be at full
1634 * power at attach. If we are not at D0, raise the power.
1635 */
1636 pmcsr = pci_config_get16(pwr_p->pwr_conf_hdl, pwr_p->pwr_pmcsr_offset);
1637 pmcsr &= PCI_PMCSR_STATE_MASK;
1638 switch (pmcsr) {
1639 case PCI_PMCSR_D0:
1640 pwr_p->pwr_func_lvl = PM_LEVEL_D0;
1641 break;
1642
1643 case PCI_PMCSR_D1:
1644 pwr_p->pwr_func_lvl = PM_LEVEL_D1;
1645 break;
1646
1647 case PCI_PMCSR_D2:
1648 pwr_p->pwr_func_lvl = PM_LEVEL_D2;
1649 break;
1650
1651 case PCI_PMCSR_D3HOT:
1652 pwr_p->pwr_func_lvl = PM_LEVEL_D3;
1653 break;
1654
1655 default:
1656 break;
1657 }
1658
1659 /* Raise the power to D0. */
1660 if (pwr_p->pwr_func_lvl != PM_LEVEL_D0 &&
1661 ((ret = pm_raise_power(dip, 0, PM_LEVEL_D0)) != DDI_SUCCESS)) {
1662 /*
1663 * Read PMCSR again. If it is at D0, ignore the return
1664 * value from pm_raise_power.
1665 */
1666 pmcsr = pci_config_get16(pwr_p->pwr_conf_hdl,
1667 pwr_p->pwr_pmcsr_offset);
1668 if ((pmcsr & PCI_PMCSR_STATE_MASK) == PCI_PMCSR_D0)
1669 ret = DDI_SUCCESS;
1670 else {
1671 PCIEB_DEBUG(DBG_PWR, dip, "pcieb_pwr_setup: could not "
1672 "raise power to D0 \n");
1673 }
1674 }
1675 if (ret == DDI_SUCCESS)
1676 pwr_p->pwr_func_lvl = PM_LEVEL_D0;
1677 return (ret);
1678 }
1679
1680 /*
1681 * Disable PM for x86 and PLX 8532 switch.
1682 * For PLX Transitioning one port on this switch to low power causes links
1683 * on other ports on the same station to die. Due to PLX erratum #34, we
1684 * can't allow the downstream device go to non-D0 state.
1685 */
1686 static int
pcieb_pwr_disable(dev_info_t * dip)1687 pcieb_pwr_disable(dev_info_t *dip)
1688 {
1689 pcie_pwr_t *pwr_p;
1690
1691 ASSERT(PCIE_PMINFO(dip));
1692 pwr_p = PCIE_NEXUS_PMINFO(dip);
1693 ASSERT(pwr_p);
1694 PCIEB_DEBUG(DBG_PWR, dip, "pcieb_pwr_disable: disabling PM\n");
1695 pwr_p->pwr_func_lvl = PM_LEVEL_D0;
1696 pwr_p->pwr_flags = PCIE_NO_CHILD_PM;
1697 return (DDI_SUCCESS);
1698 }
1699
1700 #ifdef DEBUG
1701 int pcieb_dbg_intr_print = 0;
1702 void
pcieb_dbg(uint_t bit,dev_info_t * dip,char * fmt,...)1703 pcieb_dbg(uint_t bit, dev_info_t *dip, char *fmt, ...)
1704 {
1705 va_list ap;
1706
1707 if (!pcieb_dbg_print)
1708 return;
1709
1710 if (dip)
1711 prom_printf("%s(%d): %s", ddi_driver_name(dip),
1712 ddi_get_instance(dip), pcieb_debug_sym[bit]);
1713
1714 va_start(ap, fmt);
1715 if (servicing_interrupt()) {
1716 if (pcieb_dbg_intr_print)
1717 prom_vprintf(fmt, ap);
1718 } else {
1719 prom_vprintf(fmt, ap);
1720 }
1721
1722 va_end(ap);
1723 }
1724 #endif
1725
1726 static void
pcieb_id_props(pcieb_devstate_t * pcieb)1727 pcieb_id_props(pcieb_devstate_t *pcieb)
1728 {
1729 uint64_t serialid = 0; /* 40b field of EUI-64 serial no. register */
1730 uint16_t cap_ptr;
1731 uint8_t fic = 0; /* 1 = first in chassis device */
1732 pcie_bus_t *bus_p = PCIE_DIP2BUS(pcieb->pcieb_dip);
1733 ddi_acc_handle_t config_handle = bus_p->bus_cfg_hdl;
1734
1735 /*
1736 * Identify first in chassis. In the special case of a Sun branded
1737 * PLX device, it obviously is first in chassis. Otherwise, in the
1738 * general case, look for an Expansion Slot Register and check its
1739 * first-in-chassis bit.
1740 */
1741 #ifdef PX_PLX
1742 uint16_t vendor_id = bus_p->bus_dev_ven_id & 0xFFFF;
1743 uint16_t device_id = bus_p->bus_dev_ven_id >> 16;
1744 if ((vendor_id == PXB_VENDOR_SUN) &&
1745 ((device_id == PXB_DEVICE_PLX_PCIX) ||
1746 (device_id == PXB_DEVICE_PLX_PCIE))) {
1747 fic = 1;
1748 }
1749 #endif /* PX_PLX */
1750 if ((fic == 0) && ((PCI_CAP_LOCATE(config_handle,
1751 PCI_CAP_ID_SLOT_ID, &cap_ptr)) != DDI_FAILURE)) {
1752 uint8_t esr = PCI_CAP_GET8(config_handle, NULL,
1753 cap_ptr, PCI_CAP_ID_REGS_OFF);
1754 if (PCI_CAPSLOT_FIC(esr))
1755 fic = 1;
1756 }
1757
1758 if ((PCI_CAP_LOCATE(config_handle,
1759 PCI_CAP_XCFG_SPC(PCIE_EXT_CAP_ID_SER), &cap_ptr)) != DDI_FAILURE) {
1760 /* Serialid can be 0 thru a full 40b number */
1761 serialid = PCI_XCAP_GET32(config_handle, NULL,
1762 cap_ptr, PCIE_SER_SID_UPPER_DW);
1763 serialid <<= 32;
1764 serialid |= PCI_XCAP_GET32(config_handle, NULL,
1765 cap_ptr, PCIE_SER_SID_LOWER_DW);
1766 }
1767
1768 if (fic)
1769 (void) ndi_prop_create_boolean(DDI_DEV_T_NONE, pcieb->pcieb_dip,
1770 "first-in-chassis");
1771 if (serialid)
1772 (void) ddi_prop_update_int64(DDI_DEV_T_NONE, pcieb->pcieb_dip,
1773 "serialid#", serialid);
1774 }
1775
1776 static void
pcieb_create_ranges_prop(dev_info_t * dip,ddi_acc_handle_t config_handle)1777 pcieb_create_ranges_prop(dev_info_t *dip,
1778 ddi_acc_handle_t config_handle)
1779 {
1780 uint32_t base, limit;
1781 ppb_ranges_t ranges[PCIEB_RANGE_LEN];
1782 uint8_t io_base_lo, io_limit_lo;
1783 uint16_t io_base_hi, io_limit_hi, mem_base, mem_limit;
1784 int i = 0, rangelen = sizeof (ppb_ranges_t)/sizeof (int);
1785
1786 io_base_lo = pci_config_get8(config_handle, PCI_BCNF_IO_BASE_LOW);
1787 io_limit_lo = pci_config_get8(config_handle, PCI_BCNF_IO_LIMIT_LOW);
1788 io_base_hi = pci_config_get16(config_handle, PCI_BCNF_IO_BASE_HI);
1789 io_limit_hi = pci_config_get16(config_handle, PCI_BCNF_IO_LIMIT_HI);
1790 mem_base = pci_config_get16(config_handle, PCI_BCNF_MEM_BASE);
1791 mem_limit = pci_config_get16(config_handle, PCI_BCNF_MEM_LIMIT);
1792
1793 /*
1794 * Create ranges for IO space
1795 */
1796 ranges[i].size_low = ranges[i].size_high = 0;
1797 ranges[i].parent_mid = ranges[i].child_mid = ranges[i].parent_high = 0;
1798 ranges[i].child_high = ranges[i].parent_high |=
1799 (PCI_REG_REL_M | PCI_ADDR_IO);
1800 base = PCIEB_16bit_IOADDR(io_base_lo);
1801 limit = PCIEB_16bit_IOADDR(io_limit_lo);
1802
1803 if ((io_base_lo & 0xf) == PCIEB_32BIT_IO) {
1804 base = PCIEB_LADDR(base, io_base_hi);
1805 }
1806 if ((io_limit_lo & 0xf) == PCIEB_32BIT_IO) {
1807 limit = PCIEB_LADDR(limit, io_limit_hi);
1808 }
1809
1810 if ((io_base_lo & PCIEB_32BIT_IO) && (io_limit_hi > 0)) {
1811 base = PCIEB_LADDR(base, io_base_hi);
1812 limit = PCIEB_LADDR(limit, io_limit_hi);
1813 }
1814
1815 /*
1816 * Create ranges for 32bit memory space
1817 */
1818 base = PCIEB_32bit_MEMADDR(mem_base);
1819 limit = PCIEB_32bit_MEMADDR(mem_limit);
1820 ranges[i].size_low = ranges[i].size_high = 0;
1821 ranges[i].parent_mid = ranges[i].child_mid = ranges[i].parent_high = 0;
1822 ranges[i].child_high = ranges[i].parent_high |=
1823 (PCI_REG_REL_M | PCI_ADDR_MEM32);
1824 ranges[i].child_low = ranges[i].parent_low = base;
1825 if (limit >= base) {
1826 ranges[i].size_low = limit - base + PCIEB_MEMGRAIN;
1827 i++;
1828 }
1829
1830 if (i) {
1831 (void) ndi_prop_update_int_array(DDI_DEV_T_NONE, dip, "ranges",
1832 (int *)ranges, i * rangelen);
1833 }
1834 }
1835
1836 /*
1837 * For PCI and PCI-X devices including PCIe2PCI bridge, initialize
1838 * cache-line-size and latency timer configuration registers.
1839 */
1840 void
pcieb_set_pci_perf_parameters(dev_info_t * dip,ddi_acc_handle_t cfg_hdl)1841 pcieb_set_pci_perf_parameters(dev_info_t *dip, ddi_acc_handle_t cfg_hdl)
1842 {
1843 uint_t n;
1844
1845 /* Initialize cache-line-size configuration register if needed */
1846 if (ddi_getprop(DDI_DEV_T_ANY, dip, DDI_PROP_DONTPASS,
1847 "cache-line-size", 0) == 0) {
1848 pci_config_put8(cfg_hdl, PCI_CONF_CACHE_LINESZ,
1849 PCIEB_CACHE_LINE_SIZE);
1850 n = pci_config_get8(cfg_hdl, PCI_CONF_CACHE_LINESZ);
1851 if (n != 0) {
1852 (void) ndi_prop_update_int(DDI_DEV_T_NONE, dip,
1853 "cache-line-size", n);
1854 }
1855 }
1856
1857 /* Initialize latency timer configuration registers if needed */
1858 if (ddi_getprop(DDI_DEV_T_ANY, dip, DDI_PROP_DONTPASS,
1859 "latency-timer", 0) == 0) {
1860 uchar_t min_gnt, latency_timer;
1861 uchar_t header_type;
1862
1863 /* Determine the configuration header type */
1864 header_type = pci_config_get8(cfg_hdl, PCI_CONF_HEADER);
1865
1866 if ((header_type & PCI_HEADER_TYPE_M) == PCI_HEADER_ONE) {
1867 latency_timer = PCIEB_LATENCY_TIMER;
1868 pci_config_put8(cfg_hdl, PCI_BCNF_LATENCY_TIMER,
1869 latency_timer);
1870 } else {
1871 min_gnt = pci_config_get8(cfg_hdl, PCI_CONF_MIN_G);
1872 latency_timer = min_gnt * 8;
1873 }
1874
1875 pci_config_put8(cfg_hdl, PCI_CONF_LATENCY_TIMER,
1876 latency_timer);
1877 n = pci_config_get8(cfg_hdl, PCI_CONF_LATENCY_TIMER);
1878 if (n != 0) {
1879 (void) ndi_prop_update_int(DDI_DEV_T_NONE, dip,
1880 "latency-timer", n);
1881 }
1882 }
1883 }
1884