1 /* $NetBSD: acpi_cpu.c,v 1.44 2011/06/22 08:49:54 jruoho Exp $ */ 2 3 /*- 4 * Copyright (c) 2010, 2011 Jukka Ruohonen <jruohonen@iki.fi> 5 * All rights reserved. 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 11 * 1. Redistributions of source code must retain the above copyright 12 * notice, this list of conditions and the following disclaimer. 13 * 2. Redistributions in binary form must reproduce the above copyright 14 * notice, this list of conditions and the following disclaimer in the 15 * documentation and/or other materials provided with the distribution. 16 * 17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 18 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 19 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 20 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 21 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 22 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 23 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 24 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 25 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 26 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 27 * SUCH DAMAGE. 28 */ 29 #include <sys/cdefs.h> 30 __KERNEL_RCSID(0, "$NetBSD: acpi_cpu.c,v 1.44 2011/06/22 08:49:54 jruoho Exp $"); 31 32 #include <sys/param.h> 33 #include <sys/cpu.h> 34 #include <sys/evcnt.h> 35 #include <sys/kernel.h> 36 #include <sys/kmem.h> 37 #include <sys/module.h> 38 #include <sys/mutex.h> 39 #include <sys/sysctl.h> 40 41 #include <dev/acpi/acpireg.h> 42 #include <dev/acpi/acpivar.h> 43 #include <dev/acpi/acpi_cpu.h> 44 45 #include <machine/acpi_machdep.h> 46 #include <machine/cpuvar.h> 47 48 #define _COMPONENT ACPI_BUS_COMPONENT 49 ACPI_MODULE_NAME ("acpi_cpu") 50 51 static int acpicpu_match(device_t, cfdata_t, void *); 52 static void acpicpu_attach(device_t, device_t, void *); 53 static int acpicpu_detach(device_t, int); 54 static int acpicpu_once_attach(void); 55 static int acpicpu_once_detach(void); 56 static void acpicpu_start(device_t); 57 static void acpicpu_sysctl(device_t); 58 59 static ACPI_STATUS acpicpu_object(ACPI_HANDLE, struct acpicpu_object *); 60 static uint32_t acpicpu_cap(struct acpicpu_softc *); 61 static ACPI_STATUS acpicpu_cap_osc(struct acpicpu_softc *, 62 uint32_t, uint32_t *); 63 static void acpicpu_notify(ACPI_HANDLE, uint32_t, void *); 64 static bool acpicpu_suspend(device_t, const pmf_qual_t *); 65 static bool acpicpu_resume(device_t, const pmf_qual_t *); 66 static void acpicpu_evcnt_attach(device_t); 67 static void acpicpu_evcnt_detach(device_t); 68 static void acpicpu_debug_print(device_t); 69 static const char *acpicpu_debug_print_method(uint8_t); 70 static const char *acpicpu_debug_print_dep(uint32_t); 71 72 static uint32_t acpicpu_count = 0; 73 struct acpicpu_softc **acpicpu_sc = NULL; 74 static struct sysctllog *acpicpu_log = NULL; 75 static bool acpicpu_dynamic = true; 76 static bool acpicpu_passive = true; 77 78 static const struct { 79 const char *manu; 80 const char *prod; 81 const char *vers; 82 } acpicpu_quirks[] = { 83 { "Supermicro", "PDSMi-LN4", "0123456789" }, 84 }; 85 86 CFATTACH_DECL_NEW(acpicpu, sizeof(struct acpicpu_softc), 87 acpicpu_match, acpicpu_attach, acpicpu_detach, NULL); 88 89 static int 90 acpicpu_match(device_t parent, cfdata_t match, void *aux) 91 { 92 const char *manu, *prod, *vers; 93 struct cpu_info *ci; 94 size_t i; 95 96 if (acpi_softc == NULL) 97 return 0; 98 99 manu = pmf_get_platform("system-manufacturer"); 100 prod = pmf_get_platform("system-product-name"); 101 vers = pmf_get_platform("system-version"); 102 103 if (manu != NULL && prod != NULL && vers != NULL) { 104 105 for (i = 0; i < __arraycount(acpicpu_quirks); i++) { 106 107 if (strcasecmp(acpicpu_quirks[i].manu, manu) == 0 && 108 strcasecmp(acpicpu_quirks[i].prod, prod) == 0 && 109 strcasecmp(acpicpu_quirks[i].vers, vers) == 0) 110 return 0; 111 } 112 } 113 114 ci = acpicpu_md_match(parent, match, aux); 115 116 if (ci == NULL) 117 return 0; 118 119 if (acpi_match_cpu_info(ci) == NULL) 120 return 0; 121 122 return 10; 123 } 124 125 static void 126 acpicpu_attach(device_t parent, device_t self, void *aux) 127 { 128 struct acpicpu_softc *sc = device_private(self); 129 struct cpu_info *ci; 130 ACPI_HANDLE hdl; 131 cpuid_t id; 132 int rv; 133 134 ci = acpicpu_md_attach(parent, self, aux); 135 136 if (ci == NULL) 137 return; 138 139 sc->sc_ci = ci; 140 sc->sc_dev = self; 141 sc->sc_cold = true; 142 143 hdl = acpi_match_cpu_info(ci); 144 145 if (hdl == NULL) { 146 aprint_normal(": failed to match processor\n"); 147 return; 148 } 149 150 sc->sc_node = acpi_match_node(hdl); 151 152 if (acpicpu_once_attach() != 0) { 153 aprint_normal(": failed to initialize\n"); 154 return; 155 } 156 157 KASSERT(acpi_softc != NULL); 158 KASSERT(acpicpu_sc != NULL); 159 KASSERT(sc->sc_node != NULL); 160 161 id = sc->sc_ci->ci_acpiid; 162 163 if (acpicpu_sc[id] != NULL) { 164 aprint_normal(": already attached\n"); 165 return; 166 } 167 168 aprint_naive("\n"); 169 aprint_normal(": ACPI CPU\n"); 170 171 rv = acpicpu_object(sc->sc_node->ad_handle, &sc->sc_object); 172 173 if (ACPI_FAILURE(rv)) 174 aprint_verbose_dev(self, "failed to obtain CPU object\n"); 175 176 acpicpu_count++; 177 acpicpu_sc[id] = sc; 178 179 sc->sc_cap = acpicpu_cap(sc); 180 sc->sc_ncpus = acpi_md_ncpus(); 181 sc->sc_flags = acpicpu_md_flags(); 182 183 KASSERT(acpicpu_count <= sc->sc_ncpus); 184 KASSERT(sc->sc_node->ad_device == NULL); 185 186 sc->sc_node->ad_device = self; 187 mutex_init(&sc->sc_mtx, MUTEX_DEFAULT, IPL_NONE); 188 189 acpicpu_cstate_attach(self); 190 acpicpu_pstate_attach(self); 191 acpicpu_tstate_attach(self); 192 193 acpicpu_debug_print(self); 194 acpicpu_evcnt_attach(self); 195 196 (void)config_interrupts(self, acpicpu_start); 197 (void)acpi_register_notify(sc->sc_node, acpicpu_notify); 198 (void)pmf_device_register(self, acpicpu_suspend, acpicpu_resume); 199 } 200 201 static int 202 acpicpu_detach(device_t self, int flags) 203 { 204 struct acpicpu_softc *sc = device_private(self); 205 206 sc->sc_cold = true; 207 208 acpicpu_evcnt_detach(self); 209 acpi_deregister_notify(sc->sc_node); 210 211 acpicpu_cstate_detach(self); 212 acpicpu_pstate_detach(self); 213 acpicpu_tstate_detach(self); 214 215 mutex_destroy(&sc->sc_mtx); 216 sc->sc_node->ad_device = NULL; 217 218 acpicpu_count--; 219 acpicpu_once_detach(); 220 221 return 0; 222 } 223 224 static int 225 acpicpu_once_attach(void) 226 { 227 struct acpicpu_softc *sc; 228 unsigned int i; 229 230 if (acpicpu_count != 0) 231 return 0; 232 233 KASSERT(acpicpu_sc == NULL); 234 KASSERT(acpicpu_log == NULL); 235 236 acpicpu_sc = kmem_zalloc(maxcpus * sizeof(*sc), KM_SLEEP); 237 238 if (acpicpu_sc == NULL) 239 return ENOMEM; 240 241 for (i = 0; i < maxcpus; i++) 242 acpicpu_sc[i] = NULL; 243 244 return 0; 245 } 246 247 static int 248 acpicpu_once_detach(void) 249 { 250 struct acpicpu_softc *sc; 251 252 if (acpicpu_count != 0) 253 return EDEADLK; 254 255 if (acpicpu_log != NULL) 256 sysctl_teardown(&acpicpu_log); 257 258 if (acpicpu_sc != NULL) 259 kmem_free(acpicpu_sc, maxcpus * sizeof(*sc)); 260 261 return 0; 262 } 263 264 static void 265 acpicpu_start(device_t self) 266 { 267 struct acpicpu_softc *sc = device_private(self); 268 static uint32_t count = 0; 269 270 /* 271 * Run the state-specific initialization routines. These 272 * must run only once, after interrupts have been enabled, 273 * all CPUs are running, and all ACPI CPUs have attached. 274 */ 275 if (++count != acpicpu_count || acpicpu_count != sc->sc_ncpus) { 276 sc->sc_cold = false; 277 return; 278 } 279 280 /* 281 * Set the last ACPI CPU as non-cold 282 * only after C-states are enabled. 283 */ 284 if ((sc->sc_flags & ACPICPU_FLAG_C) != 0) 285 acpicpu_cstate_start(self); 286 287 sc->sc_cold = false; 288 289 if ((sc->sc_flags & ACPICPU_FLAG_P) != 0) 290 acpicpu_pstate_start(self); 291 292 if ((sc->sc_flags & ACPICPU_FLAG_T) != 0) 293 acpicpu_tstate_start(self); 294 295 acpicpu_sysctl(self); 296 aprint_debug_dev(self, "ACPI CPUs started\n"); 297 } 298 299 static void 300 acpicpu_sysctl(device_t self) 301 { 302 const struct sysctlnode *node; 303 int err; 304 305 KASSERT(acpicpu_log == NULL); 306 307 err = sysctl_createv(&acpicpu_log, 0, NULL, &node, 308 CTLFLAG_PERMANENT, CTLTYPE_NODE, "hw", NULL, 309 NULL, 0, NULL, 0, CTL_HW, CTL_EOL); 310 311 if (err != 0) 312 goto fail; 313 314 err = sysctl_createv(&acpicpu_log, 0, &node, &node, 315 CTLFLAG_PERMANENT, CTLTYPE_NODE, "acpi", NULL, 316 NULL, 0, NULL, 0, CTL_CREATE, CTL_EOL); 317 318 if (err != 0) 319 goto fail; 320 321 err = sysctl_createv(&acpicpu_log, 0, &node, &node, 322 0, CTLTYPE_NODE, "cpu", SYSCTL_DESCR("ACPI CPU"), 323 NULL, 0, NULL, 0, CTL_CREATE, CTL_EOL); 324 325 if (err != 0) 326 goto fail; 327 328 err = sysctl_createv(&acpicpu_log, 0, &node, NULL, 329 CTLFLAG_READWRITE, CTLTYPE_BOOL, "dynamic", 330 SYSCTL_DESCR("Dynamic states"), NULL, 0, 331 &acpicpu_dynamic, 0, CTL_CREATE, CTL_EOL); 332 333 if (err != 0) 334 goto fail; 335 336 err = sysctl_createv(&acpicpu_log, 0, &node, NULL, 337 CTLFLAG_READWRITE, CTLTYPE_BOOL, "passive", 338 SYSCTL_DESCR("Passive cooling"), NULL, 0, 339 &acpicpu_passive, 0, CTL_CREATE, CTL_EOL); 340 341 if (err != 0) 342 goto fail; 343 344 return; 345 346 fail: 347 aprint_error_dev(self, "failed to initialize sysctl (err %d)\n", err); 348 } 349 350 static ACPI_STATUS 351 acpicpu_object(ACPI_HANDLE hdl, struct acpicpu_object *ao) 352 { 353 ACPI_OBJECT *obj; 354 ACPI_BUFFER buf; 355 ACPI_STATUS rv; 356 357 rv = acpi_eval_struct(hdl, NULL, &buf); 358 359 if (ACPI_FAILURE(rv)) 360 goto out; 361 362 obj = buf.Pointer; 363 364 if (obj->Type != ACPI_TYPE_PROCESSOR) { 365 rv = AE_TYPE; 366 goto out; 367 } 368 369 if (obj->Processor.ProcId > (uint32_t)maxcpus) { 370 rv = AE_LIMIT; 371 goto out; 372 } 373 374 KDASSERT((uint64_t)obj->Processor.PblkAddress < UINT32_MAX); 375 376 if (ao != NULL) { 377 ao->ao_procid = obj->Processor.ProcId; 378 ao->ao_pblklen = obj->Processor.PblkLength; 379 ao->ao_pblkaddr = obj->Processor.PblkAddress; 380 } 381 382 out: 383 if (buf.Pointer != NULL) 384 ACPI_FREE(buf.Pointer); 385 386 return rv; 387 } 388 389 static uint32_t 390 acpicpu_cap(struct acpicpu_softc *sc) 391 { 392 uint32_t flags, cap = 0; 393 ACPI_STATUS rv; 394 395 /* 396 * Query and set machine-dependent capabilities. 397 * Note that the Intel-specific _PDC method has 398 * already been evaluated. It was furthermore 399 * deprecated in the ACPI 3.0 in favor of _OSC. 400 */ 401 flags = acpi_md_pdc(); 402 rv = acpicpu_cap_osc(sc, flags, &cap); 403 404 if (ACPI_FAILURE(rv) && rv != AE_NOT_FOUND) { 405 406 aprint_error_dev(sc->sc_dev, "failed to evaluate " 407 "_OSC: %s\n", AcpiFormatException(rv)); 408 } 409 410 return (cap != 0) ? cap : flags; 411 } 412 413 static ACPI_STATUS 414 acpicpu_cap_osc(struct acpicpu_softc *sc, uint32_t flags, uint32_t *val) 415 { 416 ACPI_OBJECT_LIST arg; 417 ACPI_OBJECT obj[4]; 418 ACPI_OBJECT *osc; 419 ACPI_BUFFER buf; 420 ACPI_STATUS rv; 421 uint32_t cap[2]; 422 uint32_t *ptr; 423 int i = 5; 424 425 static uint8_t intel_uuid[16] = { 426 0x16, 0xA6, 0x77, 0x40, 0x0C, 0x29, 0xBE, 0x47, 427 0x9E, 0xBD, 0xD8, 0x70, 0x58, 0x71, 0x39, 0x53 428 }; 429 430 cap[0] = ACPI_OSC_QUERY; 431 cap[1] = flags; 432 433 again: 434 arg.Count = 4; 435 arg.Pointer = obj; 436 437 obj[0].Type = ACPI_TYPE_BUFFER; 438 obj[0].Buffer.Length = sizeof(intel_uuid); 439 obj[0].Buffer.Pointer = intel_uuid; 440 441 obj[1].Type = ACPI_TYPE_INTEGER; 442 obj[1].Integer.Value = ACPICPU_PDC_REVID; 443 444 obj[2].Type = ACPI_TYPE_INTEGER; 445 obj[2].Integer.Value = __arraycount(cap); 446 447 obj[3].Type = ACPI_TYPE_BUFFER; 448 obj[3].Buffer.Length = sizeof(cap); 449 obj[3].Buffer.Pointer = (void *)cap; 450 451 buf.Pointer = NULL; 452 buf.Length = ACPI_ALLOCATE_LOCAL_BUFFER; 453 454 rv = AcpiEvaluateObject(sc->sc_node->ad_handle, "_OSC", &arg, &buf); 455 456 if (ACPI_FAILURE(rv)) 457 goto out; 458 459 osc = buf.Pointer; 460 461 if (osc->Type != ACPI_TYPE_BUFFER) { 462 rv = AE_TYPE; 463 goto out; 464 } 465 466 if (osc->Buffer.Length != sizeof(cap)) { 467 rv = AE_BUFFER_OVERFLOW; 468 goto out; 469 } 470 471 ptr = (uint32_t *)osc->Buffer.Pointer; 472 473 if ((ptr[0] & ACPI_OSC_ERROR) != 0) { 474 rv = AE_ERROR; 475 goto out; 476 } 477 478 if ((ptr[0] & (ACPI_OSC_ERROR_REV | ACPI_OSC_ERROR_UUID)) != 0) { 479 rv = AE_BAD_PARAMETER; 480 goto out; 481 } 482 483 /* 484 * "It is strongly recommended that the OS evaluate 485 * _OSC with the Query Support Flag set until _OSC 486 * returns the Capabilities Masked bit clear, to 487 * negotiate the set of features to be granted to 488 * the OS for native support (ACPI 4.0, 6.2.10)." 489 */ 490 if ((ptr[0] & ACPI_OSC_ERROR_MASKED) != 0 && i >= 0) { 491 492 ACPI_FREE(buf.Pointer); 493 i--; 494 495 goto again; 496 } 497 498 if ((cap[0] & ACPI_OSC_QUERY) != 0) { 499 500 ACPI_FREE(buf.Pointer); 501 cap[0] &= ~ACPI_OSC_QUERY; 502 503 goto again; 504 } 505 506 /* 507 * It is permitted for _OSC to return all 508 * bits cleared, but this is specified to 509 * vary on per-device basis. Assume that 510 * everything rather than nothing will be 511 * supported in this case; we do not need 512 * the firmware to know the CPU features. 513 */ 514 *val = (ptr[1] != 0) ? ptr[1] : cap[1]; 515 516 out: 517 if (buf.Pointer != NULL) 518 ACPI_FREE(buf.Pointer); 519 520 return rv; 521 } 522 523 static void 524 acpicpu_notify(ACPI_HANDLE hdl, uint32_t evt, void *aux) 525 { 526 ACPI_OSD_EXEC_CALLBACK func; 527 struct acpicpu_softc *sc; 528 device_t self = aux; 529 530 sc = device_private(self); 531 532 if (sc->sc_cold != false) 533 return; 534 535 if (acpicpu_dynamic != true) 536 return; 537 538 switch (evt) { 539 540 case ACPICPU_C_NOTIFY: 541 542 if ((sc->sc_flags & ACPICPU_FLAG_C) == 0) 543 return; 544 545 func = acpicpu_cstate_callback; 546 break; 547 548 case ACPICPU_P_NOTIFY: 549 550 if ((sc->sc_flags & ACPICPU_FLAG_P) == 0) 551 return; 552 553 func = acpicpu_pstate_callback; 554 break; 555 556 case ACPICPU_T_NOTIFY: 557 558 if ((sc->sc_flags & ACPICPU_FLAG_T) == 0) 559 return; 560 561 func = acpicpu_tstate_callback; 562 break; 563 564 default: 565 aprint_error_dev(sc->sc_dev, "unknown notify: 0x%02X\n", evt); 566 return; 567 } 568 569 (void)AcpiOsExecute(OSL_NOTIFY_HANDLER, func, sc->sc_dev); 570 } 571 572 static bool 573 acpicpu_suspend(device_t self, const pmf_qual_t *qual) 574 { 575 struct acpicpu_softc *sc = device_private(self); 576 577 if ((sc->sc_flags & ACPICPU_FLAG_C) != 0) 578 (void)acpicpu_cstate_suspend(self); 579 580 if ((sc->sc_flags & ACPICPU_FLAG_P) != 0) 581 (void)acpicpu_pstate_suspend(self); 582 583 if ((sc->sc_flags & ACPICPU_FLAG_T) != 0) 584 (void)acpicpu_tstate_suspend(self); 585 586 sc->sc_cold = true; 587 588 return true; 589 } 590 591 static bool 592 acpicpu_resume(device_t self, const pmf_qual_t *qual) 593 { 594 struct acpicpu_softc *sc = device_private(self); 595 static const int handler = OSL_NOTIFY_HANDLER; 596 597 sc->sc_cold = false; 598 599 if ((sc->sc_flags & ACPICPU_FLAG_C) != 0) 600 (void)AcpiOsExecute(handler, acpicpu_cstate_resume, self); 601 602 if ((sc->sc_flags & ACPICPU_FLAG_P) != 0) 603 (void)AcpiOsExecute(handler, acpicpu_pstate_resume, self); 604 605 if ((sc->sc_flags & ACPICPU_FLAG_T) != 0) 606 (void)AcpiOsExecute(handler, acpicpu_tstate_resume, self); 607 608 return true; 609 } 610 611 static void 612 acpicpu_evcnt_attach(device_t self) 613 { 614 struct acpicpu_softc *sc = device_private(self); 615 struct acpicpu_cstate *cs; 616 struct acpicpu_pstate *ps; 617 struct acpicpu_tstate *ts; 618 const char *str; 619 uint32_t i; 620 621 for (i = 0; i < __arraycount(sc->sc_cstate); i++) { 622 623 cs = &sc->sc_cstate[i]; 624 625 if (cs->cs_method == 0) 626 continue; 627 628 str = "HALT"; 629 630 if (cs->cs_method == ACPICPU_C_STATE_FFH) 631 str = "MWAIT"; 632 633 if (cs->cs_method == ACPICPU_C_STATE_SYSIO) 634 str = "I/O"; 635 636 (void)snprintf(cs->cs_name, sizeof(cs->cs_name), 637 "C%d (%s)", i, str); 638 639 evcnt_attach_dynamic(&cs->cs_evcnt, EVCNT_TYPE_MISC, 640 NULL, device_xname(sc->sc_dev), cs->cs_name); 641 } 642 643 for (i = 0; i < sc->sc_pstate_count; i++) { 644 645 ps = &sc->sc_pstate[i]; 646 647 if (ps->ps_freq == 0) 648 continue; 649 650 (void)snprintf(ps->ps_name, sizeof(ps->ps_name), 651 "P%u (%u MHz)", i, ps->ps_freq); 652 653 evcnt_attach_dynamic(&ps->ps_evcnt, EVCNT_TYPE_MISC, 654 NULL, device_xname(sc->sc_dev), ps->ps_name); 655 } 656 657 for (i = 0; i < sc->sc_tstate_count; i++) { 658 659 ts = &sc->sc_tstate[i]; 660 661 if (ts->ts_percent == 0) 662 continue; 663 664 (void)snprintf(ts->ts_name, sizeof(ts->ts_name), 665 "T%u (%u %%)", i, ts->ts_percent); 666 667 evcnt_attach_dynamic(&ts->ts_evcnt, EVCNT_TYPE_MISC, 668 NULL, device_xname(sc->sc_dev), ts->ts_name); 669 } 670 } 671 672 static void 673 acpicpu_evcnt_detach(device_t self) 674 { 675 struct acpicpu_softc *sc = device_private(self); 676 struct acpicpu_cstate *cs; 677 struct acpicpu_pstate *ps; 678 struct acpicpu_tstate *ts; 679 uint32_t i; 680 681 for (i = 0; i < __arraycount(sc->sc_cstate); i++) { 682 683 cs = &sc->sc_cstate[i]; 684 685 if (cs->cs_method != 0) 686 evcnt_detach(&cs->cs_evcnt); 687 } 688 689 for (i = 0; i < sc->sc_pstate_count; i++) { 690 691 ps = &sc->sc_pstate[i]; 692 693 if (ps->ps_freq != 0) 694 evcnt_detach(&ps->ps_evcnt); 695 } 696 697 for (i = 0; i < sc->sc_tstate_count; i++) { 698 699 ts = &sc->sc_tstate[i]; 700 701 if (ts->ts_percent != 0) 702 evcnt_detach(&ts->ts_evcnt); 703 } 704 } 705 706 static void 707 acpicpu_debug_print(device_t self) 708 { 709 struct acpicpu_softc *sc = device_private(self); 710 struct cpu_info *ci = sc->sc_ci; 711 struct acpicpu_cstate *cs; 712 struct acpicpu_pstate *ps; 713 struct acpicpu_tstate *ts; 714 static bool once = false; 715 struct acpicpu_dep *dep; 716 uint32_t i, method; 717 718 if (once != true) { 719 720 for (i = 0; i < __arraycount(sc->sc_cstate); i++) { 721 722 cs = &sc->sc_cstate[i]; 723 724 if (cs->cs_method == 0) 725 continue; 726 727 aprint_verbose_dev(sc->sc_dev, "C%d: %3s, " 728 "lat %3u us, pow %5u mW%s\n", i, 729 acpicpu_debug_print_method(cs->cs_method), 730 cs->cs_latency, cs->cs_power, 731 (cs->cs_flags != 0) ? ", bus master check" : ""); 732 } 733 734 method = sc->sc_pstate_control.reg_spaceid; 735 736 for (i = 0; i < sc->sc_pstate_count; i++) { 737 738 ps = &sc->sc_pstate[i]; 739 740 if (ps->ps_freq == 0) 741 continue; 742 743 aprint_verbose_dev(sc->sc_dev, "P%d: %3s, " 744 "lat %3u us, pow %5u mW, %4u MHz%s\n", i, 745 acpicpu_debug_print_method(method), 746 ps->ps_latency, ps->ps_power, ps->ps_freq, 747 (ps->ps_flags & ACPICPU_FLAG_P_TURBO) != 0 ? 748 ", turbo boost" : ""); 749 } 750 751 method = sc->sc_tstate_control.reg_spaceid; 752 753 for (i = 0; i < sc->sc_tstate_count; i++) { 754 755 ts = &sc->sc_tstate[i]; 756 757 if (ts->ts_percent == 0) 758 continue; 759 760 aprint_verbose_dev(sc->sc_dev, "T%u: %3s, " 761 "lat %3u us, pow %5u mW, %3u %%\n", i, 762 acpicpu_debug_print_method(method), 763 ts->ts_latency, ts->ts_power, ts->ts_percent); 764 } 765 766 once = true; 767 } 768 769 aprint_debug_dev(sc->sc_dev, "id %u, lapic id %u, " 770 "cap 0x%04x, flags 0x%08x\n", ci->ci_acpiid, 771 (uint32_t)ci->ci_cpuid, sc->sc_cap, sc->sc_flags); 772 773 if ((sc->sc_flags & ACPICPU_FLAG_C_DEP) != 0) { 774 775 dep = &sc->sc_cstate_dep; 776 777 aprint_debug_dev(sc->sc_dev, "C-state coordination: " 778 "%u CPUs, domain %u, type %s\n", dep->dep_ncpus, 779 dep->dep_domain, acpicpu_debug_print_dep(dep->dep_type)); 780 } 781 782 if ((sc->sc_flags & ACPICPU_FLAG_P_DEP) != 0) { 783 784 dep = &sc->sc_pstate_dep; 785 786 aprint_debug_dev(sc->sc_dev, "P-state coordination: " 787 "%u CPUs, domain %u, type %s\n", dep->dep_ncpus, 788 dep->dep_domain, acpicpu_debug_print_dep(dep->dep_type)); 789 } 790 791 if ((sc->sc_flags & ACPICPU_FLAG_T_DEP) != 0) { 792 793 dep = &sc->sc_tstate_dep; 794 795 aprint_debug_dev(sc->sc_dev, "T-state coordination: " 796 "%u CPUs, domain %u, type %s\n", dep->dep_ncpus, 797 dep->dep_domain, acpicpu_debug_print_dep(dep->dep_type)); 798 } 799 } 800 801 static const char * 802 acpicpu_debug_print_method(uint8_t val) 803 { 804 805 if (val == ACPICPU_C_STATE_FFH) 806 return "FFH"; 807 808 if (val == ACPICPU_C_STATE_HALT) 809 return "HLT"; 810 811 if (val == ACPICPU_C_STATE_SYSIO) 812 return "I/O"; 813 814 if (val == ACPI_ADR_SPACE_SYSTEM_IO) 815 return "I/O"; 816 817 if (val == ACPI_ADR_SPACE_FIXED_HARDWARE) 818 return "FFH"; 819 820 return "???"; 821 } 822 823 static const char * 824 acpicpu_debug_print_dep(uint32_t val) 825 { 826 827 switch (val) { 828 829 case ACPICPU_DEP_SW_ALL: 830 return "SW_ALL"; 831 832 case ACPICPU_DEP_SW_ANY: 833 return "SW_ANY"; 834 835 case ACPICPU_DEP_HW_ALL: 836 return "HW_ALL"; 837 838 default: 839 return "unknown"; 840 } 841 } 842 843 MODULE(MODULE_CLASS_DRIVER, acpicpu, NULL); 844 845 #ifdef _MODULE 846 #include "ioconf.c" 847 #endif 848 849 static int 850 acpicpu_modcmd(modcmd_t cmd, void *aux) 851 { 852 int rv = 0; 853 854 switch (cmd) { 855 856 case MODULE_CMD_INIT: 857 858 #ifdef _MODULE 859 rv = config_init_component(cfdriver_ioconf_acpicpu, 860 cfattach_ioconf_acpicpu, cfdata_ioconf_acpicpu); 861 #endif 862 break; 863 864 case MODULE_CMD_FINI: 865 866 #ifdef _MODULE 867 rv = config_fini_component(cfdriver_ioconf_acpicpu, 868 cfattach_ioconf_acpicpu, cfdata_ioconf_acpicpu); 869 #endif 870 break; 871 872 default: 873 rv = ENOTTY; 874 } 875 876 return rv; 877 } 878