15db2f26eSSascha Wildner /*- 25db2f26eSSascha Wildner * Copyright (c) 2003-2005 Nate Lawson (SDG) 35db2f26eSSascha Wildner * Copyright (c) 2001 Michael Smith 45db2f26eSSascha Wildner * All rights reserved. 55db2f26eSSascha Wildner * 65db2f26eSSascha Wildner * Redistribution and use in source and binary forms, with or without 75db2f26eSSascha Wildner * modification, are permitted provided that the following conditions 85db2f26eSSascha Wildner * are met: 95db2f26eSSascha Wildner * 1. Redistributions of source code must retain the above copyright 105db2f26eSSascha Wildner * notice, this list of conditions and the following disclaimer. 115db2f26eSSascha Wildner * 2. Redistributions in binary form must reproduce the above copyright 125db2f26eSSascha Wildner * notice, this list of conditions and the following disclaimer in the 135db2f26eSSascha Wildner * documentation and/or other materials provided with the distribution. 145db2f26eSSascha Wildner * 155db2f26eSSascha Wildner * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 165db2f26eSSascha Wildner * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 175db2f26eSSascha Wildner * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 185db2f26eSSascha Wildner * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 195db2f26eSSascha Wildner * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 205db2f26eSSascha Wildner * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 215db2f26eSSascha Wildner * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 225db2f26eSSascha Wildner * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 235db2f26eSSascha Wildner * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 245db2f26eSSascha Wildner * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 255db2f26eSSascha Wildner * SUCH DAMAGE. 265db2f26eSSascha Wildner * 275db2f26eSSascha Wildner * $FreeBSD: src/sys/dev/acpica/acpi_cpu.c,v 1.72 2008/04/12 12:06:00 rpaulo Exp $ 285db2f26eSSascha Wildner */ 295db2f26eSSascha Wildner 305db2f26eSSascha Wildner #include "opt_acpi.h" 315db2f26eSSascha Wildner #include <sys/param.h> 325db2f26eSSascha Wildner #include <sys/bus.h> 335db2f26eSSascha Wildner #include <sys/kernel.h> 345db2f26eSSascha Wildner #include <sys/malloc.h> 355db2f26eSSascha Wildner #include <sys/globaldata.h> 365db2f26eSSascha Wildner #include <sys/power.h> 375db2f26eSSascha Wildner #include <sys/proc.h> 385db2f26eSSascha Wildner #include <sys/sbuf.h> 395db2f26eSSascha Wildner #include <sys/thread2.h> 405db2f26eSSascha Wildner #include <sys/mplock2.h> 415db2f26eSSascha Wildner 425db2f26eSSascha Wildner #include <bus/pci/pcivar.h> 435db2f26eSSascha Wildner #include <machine/atomic.h> 445db2f26eSSascha Wildner #include <machine/globaldata.h> 455db2f26eSSascha Wildner #include <machine/md_var.h> 465db2f26eSSascha Wildner #include <machine/smp.h> 475db2f26eSSascha Wildner #include <sys/rman.h> 485db2f26eSSascha Wildner 495db2f26eSSascha Wildner #include "acpi.h" 505db2f26eSSascha Wildner #include "acpivar.h" 515db2f26eSSascha Wildner #include "acpi_cpu.h" 525db2f26eSSascha Wildner 535db2f26eSSascha Wildner /* 545db2f26eSSascha Wildner * Support for ACPI Processor devices, including C[1-3] sleep states. 555db2f26eSSascha Wildner */ 565db2f26eSSascha Wildner 575db2f26eSSascha Wildner /* Hooks for the ACPI CA debugging infrastructure */ 585db2f26eSSascha Wildner #define _COMPONENT ACPI_PROCESSOR 595db2f26eSSascha Wildner ACPI_MODULE_NAME("PROCESSOR") 605db2f26eSSascha Wildner 615db2f26eSSascha Wildner struct acpi_cx { 625db2f26eSSascha Wildner struct resource *p_lvlx; /* Register to read to enter state. */ 635db2f26eSSascha Wildner int rid; /* rid of p_lvlx */ 645db2f26eSSascha Wildner uint32_t type; /* C1-3 (C4 and up treated as C3). */ 655db2f26eSSascha Wildner uint32_t trans_lat; /* Transition latency (usec). */ 665db2f26eSSascha Wildner uint32_t power; /* Power consumed (mW). */ 675db2f26eSSascha Wildner int res_type; /* Resource type for p_lvlx. */ 685db2f26eSSascha Wildner }; 695db2f26eSSascha Wildner #define MAX_CX_STATES 8 705db2f26eSSascha Wildner 715db2f26eSSascha Wildner struct acpi_cpu_softc { 725db2f26eSSascha Wildner device_t cpu_dev; 735db2f26eSSascha Wildner struct acpi_cpux_softc *cpu_parent; 745db2f26eSSascha Wildner ACPI_HANDLE cpu_handle; 755db2f26eSSascha Wildner struct mdglobaldata *md; 765db2f26eSSascha Wildner uint32_t cpu_acpi_id; /* ACPI processor id */ 775db2f26eSSascha Wildner uint32_t cpu_p_blk; /* ACPI P_BLK location */ 785db2f26eSSascha Wildner uint32_t cpu_p_blk_len; /* P_BLK length (must be 6). */ 795db2f26eSSascha Wildner struct acpi_cx cpu_cx_states[MAX_CX_STATES]; 805db2f26eSSascha Wildner int cpu_cx_count; /* Number of valid Cx states. */ 815db2f26eSSascha Wildner int cpu_prev_sleep;/* Last idle sleep duration. */ 825db2f26eSSascha Wildner /* Runtime state. */ 835db2f26eSSascha Wildner int cpu_non_c3; /* Index of lowest non-C3 state. */ 845db2f26eSSascha Wildner u_int cpu_cx_stats[MAX_CX_STATES];/* Cx usage history. */ 855db2f26eSSascha Wildner /* Values for sysctl. */ 865db2f26eSSascha Wildner int cpu_cx_lowest; 875db2f26eSSascha Wildner char cpu_cx_supported[64]; 885db2f26eSSascha Wildner }; 895db2f26eSSascha Wildner 905db2f26eSSascha Wildner struct acpi_cpu_device { 915db2f26eSSascha Wildner struct resource_list ad_rl; 925db2f26eSSascha Wildner }; 935db2f26eSSascha Wildner 945db2f26eSSascha Wildner #define CPU_GET_REG(reg, width) \ 955db2f26eSSascha Wildner (bus_space_read_ ## width(rman_get_bustag((reg)), \ 965db2f26eSSascha Wildner rman_get_bushandle((reg)), 0)) 975db2f26eSSascha Wildner #define CPU_SET_REG(reg, width, val) \ 985db2f26eSSascha Wildner (bus_space_write_ ## width(rman_get_bustag((reg)), \ 995db2f26eSSascha Wildner rman_get_bushandle((reg)), 0, (val))) 1005db2f26eSSascha Wildner 1015db2f26eSSascha Wildner #define PM_USEC(x) ((x) >> 2) /* ~4 clocks per usec (3.57955 Mhz) */ 1025db2f26eSSascha Wildner 1035db2f26eSSascha Wildner #define ACPI_NOTIFY_CX_STATES 0x81 /* _CST changed. */ 1045db2f26eSSascha Wildner 1055db2f26eSSascha Wildner #define CPU_QUIRK_NO_C3 (1<<0) /* C3-type states are not usable. */ 1065db2f26eSSascha Wildner #define CPU_QUIRK_NO_BM_CTRL (1<<2) /* No bus mastering control. */ 1075db2f26eSSascha Wildner 1085db2f26eSSascha Wildner #define PCI_VENDOR_INTEL 0x8086 1095db2f26eSSascha Wildner #define PCI_DEVICE_82371AB_3 0x7113 /* PIIX4 chipset for quirks. */ 1105db2f26eSSascha Wildner #define PCI_REVISION_A_STEP 0 1115db2f26eSSascha Wildner #define PCI_REVISION_B_STEP 1 1125db2f26eSSascha Wildner #define PCI_REVISION_4E 2 1135db2f26eSSascha Wildner #define PCI_REVISION_4M 3 1145db2f26eSSascha Wildner #define PIIX4_DEVACTB_REG 0x58 1155db2f26eSSascha Wildner #define PIIX4_BRLD_EN_IRQ0 (1<<0) 1165db2f26eSSascha Wildner #define PIIX4_BRLD_EN_IRQ (1<<1) 1175db2f26eSSascha Wildner #define PIIX4_BRLD_EN_IRQ8 (1<<5) 1185db2f26eSSascha Wildner #define PIIX4_STOP_BREAK_MASK (PIIX4_BRLD_EN_IRQ0 | PIIX4_BRLD_EN_IRQ | PIIX4_BRLD_EN_IRQ8) 1195db2f26eSSascha Wildner #define PIIX4_PCNTRL_BST_EN (1<<10) 1205db2f26eSSascha Wildner 1215db2f26eSSascha Wildner /* Platform hardware resource information. */ 1225db2f26eSSascha Wildner static uint32_t cpu_smi_cmd; /* Value to write to SMI_CMD. */ 1235db2f26eSSascha Wildner static uint8_t cpu_cst_cnt; /* Indicate we are _CST aware. */ 1245db2f26eSSascha Wildner static int cpu_quirks; /* Indicate any hardware bugs. */ 1255db2f26eSSascha Wildner 1265db2f26eSSascha Wildner /* Runtime state. */ 1275db2f26eSSascha Wildner static int cpu_disable_idle; /* Disable entry to idle function */ 1285db2f26eSSascha Wildner static int cpu_cx_count; /* Number of valid Cx states */ 1295db2f26eSSascha Wildner 1305db2f26eSSascha Wildner /* Values for sysctl. */ 1315db2f26eSSascha Wildner static int cpu_cx_generic; 1325db2f26eSSascha Wildner static int cpu_cx_lowest; 1335db2f26eSSascha Wildner 1345db2f26eSSascha Wildner /* C3 state transition */ 1355db2f26eSSascha Wildner static int cpu_c3_ncpus; 1365db2f26eSSascha Wildner 1375db2f26eSSascha Wildner static device_t *cpu_devices; 1385db2f26eSSascha Wildner static int cpu_ndevices; 1395db2f26eSSascha Wildner static struct acpi_cpu_softc **cpu_softc; 1405db2f26eSSascha Wildner 1415db2f26eSSascha Wildner static int acpi_cpu_cst_probe(device_t dev); 1425db2f26eSSascha Wildner static int acpi_cpu_cst_attach(device_t dev); 1435db2f26eSSascha Wildner static int acpi_cpu_cst_suspend(device_t dev); 1445db2f26eSSascha Wildner static int acpi_cpu_cst_resume(device_t dev); 1455db2f26eSSascha Wildner static struct resource_list *acpi_cpu_cst_get_rlist(device_t dev, 1465db2f26eSSascha Wildner device_t child); 1475db2f26eSSascha Wildner static device_t acpi_cpu_cst_add_child(device_t bus, device_t parent, 1485db2f26eSSascha Wildner int order, const char *name, int unit); 1495db2f26eSSascha Wildner static int acpi_cpu_cst_read_ivar(device_t dev, device_t child, 1505db2f26eSSascha Wildner int index, uintptr_t *result); 1515db2f26eSSascha Wildner static int acpi_cpu_cst_shutdown(device_t dev); 1525db2f26eSSascha Wildner static void acpi_cpu_cx_probe(struct acpi_cpu_softc *sc); 1535db2f26eSSascha Wildner static void acpi_cpu_generic_cx_probe(struct acpi_cpu_softc *sc); 1545db2f26eSSascha Wildner static int acpi_cpu_cx_cst(struct acpi_cpu_softc *sc); 1555db2f26eSSascha Wildner static void acpi_cpu_startup(void *arg); 1565db2f26eSSascha Wildner static void acpi_cpu_startup_cx(struct acpi_cpu_softc *sc); 1575db2f26eSSascha Wildner static void acpi_cpu_cx_list(struct acpi_cpu_softc *sc); 1585db2f26eSSascha Wildner static void acpi_cpu_idle(void); 1595db2f26eSSascha Wildner static void acpi_cpu_cst_notify(device_t); 1605db2f26eSSascha Wildner static int acpi_cpu_quirks(void); 1615db2f26eSSascha Wildner static int acpi_cpu_usage_sysctl(SYSCTL_HANDLER_ARGS); 1625db2f26eSSascha Wildner static int acpi_cpu_set_cx_lowest(struct acpi_cpu_softc *sc, int val); 1635db2f26eSSascha Wildner static int acpi_cpu_cx_lowest_sysctl(SYSCTL_HANDLER_ARGS); 1645db2f26eSSascha Wildner static int acpi_cpu_global_cx_lowest_sysctl(SYSCTL_HANDLER_ARGS); 1655db2f26eSSascha Wildner 1665db2f26eSSascha Wildner static void acpi_cpu_c1(void); /* XXX */ 1675db2f26eSSascha Wildner 1685db2f26eSSascha Wildner static device_method_t acpi_cpu_cst_methods[] = { 1695db2f26eSSascha Wildner /* Device interface */ 1705db2f26eSSascha Wildner DEVMETHOD(device_probe, acpi_cpu_cst_probe), 1715db2f26eSSascha Wildner DEVMETHOD(device_attach, acpi_cpu_cst_attach), 1725db2f26eSSascha Wildner DEVMETHOD(device_detach, bus_generic_detach), 1735db2f26eSSascha Wildner DEVMETHOD(device_shutdown, acpi_cpu_cst_shutdown), 1745db2f26eSSascha Wildner DEVMETHOD(device_suspend, acpi_cpu_cst_suspend), 1755db2f26eSSascha Wildner DEVMETHOD(device_resume, acpi_cpu_cst_resume), 1765db2f26eSSascha Wildner 1775db2f26eSSascha Wildner /* Bus interface */ 1785db2f26eSSascha Wildner DEVMETHOD(bus_add_child, acpi_cpu_cst_add_child), 1795db2f26eSSascha Wildner DEVMETHOD(bus_read_ivar, acpi_cpu_cst_read_ivar), 1805db2f26eSSascha Wildner DEVMETHOD(bus_get_resource_list, acpi_cpu_cst_get_rlist), 1815db2f26eSSascha Wildner DEVMETHOD(bus_get_resource, bus_generic_rl_get_resource), 1825db2f26eSSascha Wildner DEVMETHOD(bus_set_resource, bus_generic_rl_set_resource), 1835db2f26eSSascha Wildner DEVMETHOD(bus_alloc_resource, bus_generic_rl_alloc_resource), 1845db2f26eSSascha Wildner DEVMETHOD(bus_release_resource, bus_generic_rl_release_resource), 1855db2f26eSSascha Wildner DEVMETHOD(bus_driver_added, bus_generic_driver_added), 1865db2f26eSSascha Wildner DEVMETHOD(bus_activate_resource, bus_generic_activate_resource), 1875db2f26eSSascha Wildner DEVMETHOD(bus_deactivate_resource, bus_generic_deactivate_resource), 1885db2f26eSSascha Wildner DEVMETHOD(bus_setup_intr, bus_generic_setup_intr), 1895db2f26eSSascha Wildner DEVMETHOD(bus_teardown_intr, bus_generic_teardown_intr), 190*d3c9c58eSSascha Wildner DEVMETHOD_END 1915db2f26eSSascha Wildner }; 1925db2f26eSSascha Wildner 1935db2f26eSSascha Wildner static driver_t acpi_cpu_cst_driver = { 1945db2f26eSSascha Wildner "cpu_cst", 1955db2f26eSSascha Wildner acpi_cpu_cst_methods, 1965db2f26eSSascha Wildner sizeof(struct acpi_cpu_softc), 1975db2f26eSSascha Wildner }; 1985db2f26eSSascha Wildner 1995db2f26eSSascha Wildner static devclass_t acpi_cpu_cst_devclass; 2005db2f26eSSascha Wildner DRIVER_MODULE(cpu_cst, cpu, acpi_cpu_cst_driver, acpi_cpu_cst_devclass, NULL, NULL); 2015db2f26eSSascha Wildner MODULE_DEPEND(cpu_cst, acpi, 1, 1, 1); 2025db2f26eSSascha Wildner 2035db2f26eSSascha Wildner static int 2045db2f26eSSascha Wildner acpi_cpu_cst_probe(device_t dev) 2055db2f26eSSascha Wildner { 2065db2f26eSSascha Wildner int cpu_id; 2075db2f26eSSascha Wildner 2085db2f26eSSascha Wildner if (acpi_disabled("cpu_cst") || acpi_get_type(dev) != ACPI_TYPE_PROCESSOR) 2095db2f26eSSascha Wildner return (ENXIO); 2105db2f26eSSascha Wildner 2115db2f26eSSascha Wildner cpu_id = acpi_get_magic(dev); 2125db2f26eSSascha Wildner 2135db2f26eSSascha Wildner if (cpu_softc == NULL) 2145db2f26eSSascha Wildner cpu_softc = kmalloc(sizeof(struct acpi_cpu_softc *) * 2155db2f26eSSascha Wildner SMP_MAXCPU, M_TEMP /* XXX */, M_INTWAIT | M_ZERO); 2165db2f26eSSascha Wildner 2175db2f26eSSascha Wildner /* 2185db2f26eSSascha Wildner * Check if we already probed this processor. We scan the bus twice 2195db2f26eSSascha Wildner * so it's possible we've already seen this one. 2205db2f26eSSascha Wildner */ 2215db2f26eSSascha Wildner if (cpu_softc[cpu_id] != NULL) { 2225db2f26eSSascha Wildner device_printf(dev, "CPU%d cstate already exist\n", cpu_id); 2235db2f26eSSascha Wildner return (ENXIO); 2245db2f26eSSascha Wildner } 2255db2f26eSSascha Wildner 2265db2f26eSSascha Wildner /* Mark this processor as in-use and save our derived id for attach. */ 2275db2f26eSSascha Wildner cpu_softc[cpu_id] = (void *)1; 2285db2f26eSSascha Wildner device_set_desc(dev, "ACPI CPU C-State"); 2295db2f26eSSascha Wildner 2305db2f26eSSascha Wildner return (0); 2315db2f26eSSascha Wildner } 2325db2f26eSSascha Wildner 2335db2f26eSSascha Wildner static int 2345db2f26eSSascha Wildner acpi_cpu_cst_attach(device_t dev) 2355db2f26eSSascha Wildner { 2365db2f26eSSascha Wildner ACPI_BUFFER buf; 2375db2f26eSSascha Wildner ACPI_OBJECT *obj; 2385db2f26eSSascha Wildner struct mdglobaldata *md; 2395db2f26eSSascha Wildner struct acpi_cpu_softc *sc; 2405db2f26eSSascha Wildner ACPI_STATUS status; 2415db2f26eSSascha Wildner int cpu_id; 2425db2f26eSSascha Wildner 2435db2f26eSSascha Wildner ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__); 2445db2f26eSSascha Wildner 2455db2f26eSSascha Wildner sc = device_get_softc(dev); 2465db2f26eSSascha Wildner sc->cpu_dev = dev; 2475db2f26eSSascha Wildner sc->cpu_parent = device_get_softc(device_get_parent(dev)); 2485db2f26eSSascha Wildner sc->cpu_handle = acpi_get_handle(dev); 2495db2f26eSSascha Wildner cpu_id = acpi_get_magic(dev); 2505db2f26eSSascha Wildner cpu_softc[cpu_id] = sc; 2515db2f26eSSascha Wildner md = (struct mdglobaldata *)globaldata_find(device_get_unit(dev)); 2525db2f26eSSascha Wildner sc->md = md; 2535db2f26eSSascha Wildner cpu_smi_cmd = AcpiGbl_FADT.SmiCommand; 2545db2f26eSSascha Wildner cpu_cst_cnt = AcpiGbl_FADT.CstControl; 2555db2f26eSSascha Wildner 2565db2f26eSSascha Wildner buf.Pointer = NULL; 2575db2f26eSSascha Wildner buf.Length = ACPI_ALLOCATE_BUFFER; 2585db2f26eSSascha Wildner status = AcpiEvaluateObject(sc->cpu_handle, NULL, NULL, &buf); 2595db2f26eSSascha Wildner if (ACPI_FAILURE(status)) { 2605db2f26eSSascha Wildner device_printf(dev, "attach failed to get Processor obj - %s\n", 2615db2f26eSSascha Wildner AcpiFormatException(status)); 2625db2f26eSSascha Wildner return (ENXIO); 2635db2f26eSSascha Wildner } 2645db2f26eSSascha Wildner obj = (ACPI_OBJECT *)buf.Pointer; 2655db2f26eSSascha Wildner sc->cpu_p_blk = obj->Processor.PblkAddress; 2665db2f26eSSascha Wildner sc->cpu_p_blk_len = obj->Processor.PblkLength; 2675db2f26eSSascha Wildner sc->cpu_acpi_id = obj->Processor.ProcId; 2685db2f26eSSascha Wildner AcpiOsFree(obj); 2695db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, "acpi_cpu%d: P_BLK at %#x/%d\n", 2705db2f26eSSascha Wildner device_get_unit(dev), sc->cpu_p_blk, sc->cpu_p_blk_len)); 2715db2f26eSSascha Wildner 2725db2f26eSSascha Wildner /* 2735db2f26eSSascha Wildner * If this is the first cpu we attach, create and initialize the generic 2745db2f26eSSascha Wildner * resources that will be used by all acpi cpu devices. 2755db2f26eSSascha Wildner */ 2765db2f26eSSascha Wildner if (device_get_unit(dev) == 0) { 2775db2f26eSSascha Wildner /* Assume we won't be using generic Cx mode by default */ 2785db2f26eSSascha Wildner cpu_cx_generic = FALSE; 2795db2f26eSSascha Wildner 2805db2f26eSSascha Wildner /* Queue post cpu-probing task handler */ 2815db2f26eSSascha Wildner AcpiOsExecute(OSL_NOTIFY_HANDLER, acpi_cpu_startup, NULL); 2825db2f26eSSascha Wildner } 2835db2f26eSSascha Wildner 2845db2f26eSSascha Wildner /* Probe for Cx state support. */ 2855db2f26eSSascha Wildner acpi_cpu_cx_probe(sc); 2865db2f26eSSascha Wildner 2875db2f26eSSascha Wildner /* Finally, call identify and probe/attach for child devices. */ 2885db2f26eSSascha Wildner bus_generic_probe(dev); 2895db2f26eSSascha Wildner bus_generic_attach(dev); 2905db2f26eSSascha Wildner 2915db2f26eSSascha Wildner return (0); 2925db2f26eSSascha Wildner } 2935db2f26eSSascha Wildner 2945db2f26eSSascha Wildner /* 2955db2f26eSSascha Wildner * Disable any entry to the idle function during suspend and re-enable it 2965db2f26eSSascha Wildner * during resume. 2975db2f26eSSascha Wildner */ 2985db2f26eSSascha Wildner static int 2995db2f26eSSascha Wildner acpi_cpu_cst_suspend(device_t dev) 3005db2f26eSSascha Wildner { 3015db2f26eSSascha Wildner int error; 3025db2f26eSSascha Wildner 3035db2f26eSSascha Wildner error = bus_generic_suspend(dev); 3045db2f26eSSascha Wildner if (error) 3055db2f26eSSascha Wildner return (error); 3065db2f26eSSascha Wildner cpu_disable_idle = TRUE; 3075db2f26eSSascha Wildner return (0); 3085db2f26eSSascha Wildner } 3095db2f26eSSascha Wildner 3105db2f26eSSascha Wildner static int 3115db2f26eSSascha Wildner acpi_cpu_cst_resume(device_t dev) 3125db2f26eSSascha Wildner { 3135db2f26eSSascha Wildner 3145db2f26eSSascha Wildner cpu_disable_idle = FALSE; 3155db2f26eSSascha Wildner return (bus_generic_resume(dev)); 3165db2f26eSSascha Wildner } 3175db2f26eSSascha Wildner 3185db2f26eSSascha Wildner static struct resource_list * 3195db2f26eSSascha Wildner acpi_cpu_cst_get_rlist(device_t dev, device_t child) 3205db2f26eSSascha Wildner { 3215db2f26eSSascha Wildner struct acpi_cpu_device *ad; 3225db2f26eSSascha Wildner 3235db2f26eSSascha Wildner ad = device_get_ivars(child); 3245db2f26eSSascha Wildner if (ad == NULL) 3255db2f26eSSascha Wildner return (NULL); 3265db2f26eSSascha Wildner return (&ad->ad_rl); 3275db2f26eSSascha Wildner } 3285db2f26eSSascha Wildner 3295db2f26eSSascha Wildner static device_t 3305db2f26eSSascha Wildner acpi_cpu_cst_add_child(device_t bus, device_t parent, int order, 3315db2f26eSSascha Wildner const char *name, int unit) 3325db2f26eSSascha Wildner { 3335db2f26eSSascha Wildner struct acpi_cpu_device *ad; 3345db2f26eSSascha Wildner device_t child; 3355db2f26eSSascha Wildner 3365db2f26eSSascha Wildner if ((ad = kmalloc(sizeof(*ad), M_TEMP, M_NOWAIT | M_ZERO)) == NULL) 3375db2f26eSSascha Wildner return (NULL); 3385db2f26eSSascha Wildner 3395db2f26eSSascha Wildner resource_list_init(&ad->ad_rl); 3405db2f26eSSascha Wildner 3415db2f26eSSascha Wildner child = device_add_child_ordered(parent, order, name, unit); 3425db2f26eSSascha Wildner if (child != NULL) 3435db2f26eSSascha Wildner device_set_ivars(child, ad); 3445db2f26eSSascha Wildner else 3455db2f26eSSascha Wildner kfree(ad, M_TEMP); 3465db2f26eSSascha Wildner return (child); 3475db2f26eSSascha Wildner } 3485db2f26eSSascha Wildner 3495db2f26eSSascha Wildner static int 3505db2f26eSSascha Wildner acpi_cpu_cst_read_ivar(device_t dev, device_t child, int index, 3515db2f26eSSascha Wildner uintptr_t *result) 3525db2f26eSSascha Wildner { 3535db2f26eSSascha Wildner struct acpi_cpu_softc *sc; 3545db2f26eSSascha Wildner 3555db2f26eSSascha Wildner sc = device_get_softc(dev); 3565db2f26eSSascha Wildner switch (index) { 3575db2f26eSSascha Wildner case ACPI_IVAR_HANDLE: 3585db2f26eSSascha Wildner *result = (uintptr_t)sc->cpu_handle; 3595db2f26eSSascha Wildner break; 3605db2f26eSSascha Wildner #if 0 3615db2f26eSSascha Wildner case CPU_IVAR_PCPU: 3625db2f26eSSascha Wildner *result = (uintptr_t)sc->cpu_pcpu; 3635db2f26eSSascha Wildner break; 3645db2f26eSSascha Wildner #endif 3655db2f26eSSascha Wildner default: 3665db2f26eSSascha Wildner return (ENOENT); 3675db2f26eSSascha Wildner } 3685db2f26eSSascha Wildner return (0); 3695db2f26eSSascha Wildner } 3705db2f26eSSascha Wildner 3715db2f26eSSascha Wildner static int 3725db2f26eSSascha Wildner acpi_cpu_cst_shutdown(device_t dev) 3735db2f26eSSascha Wildner { 3745db2f26eSSascha Wildner ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__); 3755db2f26eSSascha Wildner 3765db2f26eSSascha Wildner /* Allow children to shutdown first. */ 3775db2f26eSSascha Wildner bus_generic_shutdown(dev); 3785db2f26eSSascha Wildner 3795db2f26eSSascha Wildner /* 3805db2f26eSSascha Wildner * Disable any entry to the idle function. There is a small race where 3815db2f26eSSascha Wildner * an idle thread have passed this check but not gone to sleep. This 3825db2f26eSSascha Wildner * is ok since device_shutdown() does not free the softc, otherwise 3835db2f26eSSascha Wildner * we'd have to be sure all threads were evicted before returning. 3845db2f26eSSascha Wildner */ 3855db2f26eSSascha Wildner cpu_disable_idle = TRUE; 3865db2f26eSSascha Wildner 3875db2f26eSSascha Wildner return_VALUE (0); 3885db2f26eSSascha Wildner } 3895db2f26eSSascha Wildner 3905db2f26eSSascha Wildner static void 3915db2f26eSSascha Wildner acpi_cpu_cx_probe(struct acpi_cpu_softc *sc) 3925db2f26eSSascha Wildner { 3935db2f26eSSascha Wildner ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__); 3945db2f26eSSascha Wildner 3955db2f26eSSascha Wildner /* Use initial sleep value of 1 sec. to start with lowest idle state. */ 3965db2f26eSSascha Wildner sc->cpu_prev_sleep = 1000000; 3975db2f26eSSascha Wildner sc->cpu_cx_lowest = 0; 3985db2f26eSSascha Wildner 3995db2f26eSSascha Wildner /* 4005db2f26eSSascha Wildner * Check for the ACPI 2.0 _CST sleep states object. If we can't find 4015db2f26eSSascha Wildner * any, we'll revert to generic FADT/P_BLK Cx control method which will 4025db2f26eSSascha Wildner * be handled by acpi_cpu_startup. We need to defer to after having 4035db2f26eSSascha Wildner * probed all the cpus in the system before probing for generic Cx 4045db2f26eSSascha Wildner * states as we may already have found cpus with valid _CST packages 4055db2f26eSSascha Wildner */ 4065db2f26eSSascha Wildner if (!cpu_cx_generic && acpi_cpu_cx_cst(sc) != 0) { 4075db2f26eSSascha Wildner /* 4085db2f26eSSascha Wildner * We were unable to find a _CST package for this cpu or there 4095db2f26eSSascha Wildner * was an error parsing it. Switch back to generic mode. 4105db2f26eSSascha Wildner */ 4115db2f26eSSascha Wildner cpu_cx_generic = TRUE; 4125db2f26eSSascha Wildner if (bootverbose) 4135db2f26eSSascha Wildner device_printf(sc->cpu_dev, "switching to generic Cx mode\n"); 4145db2f26eSSascha Wildner } 4155db2f26eSSascha Wildner 4165db2f26eSSascha Wildner /* 4175db2f26eSSascha Wildner * TODO: _CSD Package should be checked here. 4185db2f26eSSascha Wildner */ 4195db2f26eSSascha Wildner } 4205db2f26eSSascha Wildner 4215db2f26eSSascha Wildner static void 4225db2f26eSSascha Wildner acpi_cpu_generic_cx_probe(struct acpi_cpu_softc *sc) 4235db2f26eSSascha Wildner { 4245db2f26eSSascha Wildner ACPI_GENERIC_ADDRESS gas; 4255db2f26eSSascha Wildner struct acpi_cx *cx_ptr; 4265db2f26eSSascha Wildner 4275db2f26eSSascha Wildner sc->cpu_cx_count = 0; 4285db2f26eSSascha Wildner cx_ptr = sc->cpu_cx_states; 4295db2f26eSSascha Wildner 4305db2f26eSSascha Wildner /* Use initial sleep value of 1 sec. to start with lowest idle state. */ 4315db2f26eSSascha Wildner sc->cpu_prev_sleep = 1000000; 4325db2f26eSSascha Wildner 4335db2f26eSSascha Wildner /* C1 has been required since just after ACPI 1.0 */ 4345db2f26eSSascha Wildner cx_ptr->type = ACPI_STATE_C1; 4355db2f26eSSascha Wildner cx_ptr->trans_lat = 0; 4365db2f26eSSascha Wildner cx_ptr++; 4375db2f26eSSascha Wildner sc->cpu_cx_count++; 4385db2f26eSSascha Wildner 4395db2f26eSSascha Wildner /* 4405db2f26eSSascha Wildner * The spec says P_BLK must be 6 bytes long. However, some systems 4415db2f26eSSascha Wildner * use it to indicate a fractional set of features present so we 4425db2f26eSSascha Wildner * take 5 as C2. Some may also have a value of 7 to indicate 4435db2f26eSSascha Wildner * another C3 but most use _CST for this (as required) and having 4445db2f26eSSascha Wildner * "only" C1-C3 is not a hardship. 4455db2f26eSSascha Wildner */ 4465db2f26eSSascha Wildner if (sc->cpu_p_blk_len < 5) 4475db2f26eSSascha Wildner return; 4485db2f26eSSascha Wildner 4495db2f26eSSascha Wildner /* Validate and allocate resources for C2 (P_LVL2). */ 4505db2f26eSSascha Wildner gas.SpaceId = ACPI_ADR_SPACE_SYSTEM_IO; 4515db2f26eSSascha Wildner gas.BitWidth = 8; 4525db2f26eSSascha Wildner if (AcpiGbl_FADT.C2Latency <= 100) { 4535db2f26eSSascha Wildner gas.Address = sc->cpu_p_blk + 4; 4545db2f26eSSascha Wildner 4555db2f26eSSascha Wildner cx_ptr->rid = sc->cpu_parent->cpux_next_rid; 4565db2f26eSSascha Wildner acpi_bus_alloc_gas(sc->cpu_dev, &cx_ptr->type, &cx_ptr->rid, &gas, &cx_ptr->p_lvlx, 4575db2f26eSSascha Wildner RF_SHAREABLE); 4585db2f26eSSascha Wildner if (cx_ptr->p_lvlx != NULL) { 4595db2f26eSSascha Wildner sc->cpu_parent->cpux_next_rid++; 4605db2f26eSSascha Wildner cx_ptr->type = ACPI_STATE_C2; 4615db2f26eSSascha Wildner cx_ptr->trans_lat = AcpiGbl_FADT.C2Latency; 4625db2f26eSSascha Wildner cx_ptr++; 4635db2f26eSSascha Wildner sc->cpu_cx_count++; 4645db2f26eSSascha Wildner } 4655db2f26eSSascha Wildner } 4665db2f26eSSascha Wildner if (sc->cpu_p_blk_len < 6) 4675db2f26eSSascha Wildner return; 4685db2f26eSSascha Wildner 4695db2f26eSSascha Wildner /* Validate and allocate resources for C3 (P_LVL3). */ 4705db2f26eSSascha Wildner if (AcpiGbl_FADT.C3Latency <= 1000 && !(cpu_quirks & CPU_QUIRK_NO_C3)) { 4715db2f26eSSascha Wildner gas.Address = sc->cpu_p_blk + 5; 4725db2f26eSSascha Wildner 4735db2f26eSSascha Wildner cx_ptr->rid = sc->cpu_parent->cpux_next_rid; 4745db2f26eSSascha Wildner acpi_bus_alloc_gas(sc->cpu_dev, &cx_ptr->type, &cx_ptr->rid, &gas, 4755db2f26eSSascha Wildner &cx_ptr->p_lvlx, RF_SHAREABLE); 4765db2f26eSSascha Wildner if (cx_ptr->p_lvlx != NULL) { 4775db2f26eSSascha Wildner sc->cpu_parent->cpux_next_rid++; 4785db2f26eSSascha Wildner cx_ptr->type = ACPI_STATE_C3; 4795db2f26eSSascha Wildner cx_ptr->trans_lat = AcpiGbl_FADT.C3Latency; 4805db2f26eSSascha Wildner cx_ptr++; 4815db2f26eSSascha Wildner sc->cpu_cx_count++; 4825db2f26eSSascha Wildner } 4835db2f26eSSascha Wildner } 4845db2f26eSSascha Wildner } 4855db2f26eSSascha Wildner 4865db2f26eSSascha Wildner /* 4875db2f26eSSascha Wildner * Parse a _CST package and set up its Cx states. Since the _CST object 4885db2f26eSSascha Wildner * can change dynamically, our notify handler may call this function 4895db2f26eSSascha Wildner * to clean up and probe the new _CST package. 4905db2f26eSSascha Wildner */ 4915db2f26eSSascha Wildner static int 4925db2f26eSSascha Wildner acpi_cpu_cx_cst(struct acpi_cpu_softc *sc) 4935db2f26eSSascha Wildner { 4945db2f26eSSascha Wildner struct acpi_cx *cx_ptr; 4955db2f26eSSascha Wildner ACPI_STATUS status; 4965db2f26eSSascha Wildner ACPI_BUFFER buf; 4975db2f26eSSascha Wildner ACPI_OBJECT *top; 4985db2f26eSSascha Wildner ACPI_OBJECT *pkg; 4995db2f26eSSascha Wildner uint32_t count; 5005db2f26eSSascha Wildner int i; 5015db2f26eSSascha Wildner 5025db2f26eSSascha Wildner ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__); 5035db2f26eSSascha Wildner 5045db2f26eSSascha Wildner buf.Pointer = NULL; 5055db2f26eSSascha Wildner buf.Length = ACPI_ALLOCATE_BUFFER; 5065db2f26eSSascha Wildner status = AcpiEvaluateObject(sc->cpu_handle, "_CST", NULL, &buf); 5075db2f26eSSascha Wildner if (ACPI_FAILURE(status)) 5085db2f26eSSascha Wildner return (ENXIO); 5095db2f26eSSascha Wildner 5105db2f26eSSascha Wildner /* _CST is a package with a count and at least one Cx package. */ 5115db2f26eSSascha Wildner top = (ACPI_OBJECT *)buf.Pointer; 5125db2f26eSSascha Wildner if (!ACPI_PKG_VALID(top, 2) || acpi_PkgInt32(top, 0, &count) != 0) { 5135db2f26eSSascha Wildner device_printf(sc->cpu_dev, "invalid _CST package\n"); 5145db2f26eSSascha Wildner AcpiOsFree(buf.Pointer); 5155db2f26eSSascha Wildner return (ENXIO); 5165db2f26eSSascha Wildner } 5175db2f26eSSascha Wildner if (count != top->Package.Count - 1) { 5185db2f26eSSascha Wildner device_printf(sc->cpu_dev, "invalid _CST state count (%d != %d)\n", 5195db2f26eSSascha Wildner count, top->Package.Count - 1); 5205db2f26eSSascha Wildner count = top->Package.Count - 1; 5215db2f26eSSascha Wildner } 5225db2f26eSSascha Wildner if (count > MAX_CX_STATES) { 5235db2f26eSSascha Wildner device_printf(sc->cpu_dev, "_CST has too many states (%d)\n", count); 5245db2f26eSSascha Wildner count = MAX_CX_STATES; 5255db2f26eSSascha Wildner } 5265db2f26eSSascha Wildner 5275db2f26eSSascha Wildner /* Set up all valid states. */ 5285db2f26eSSascha Wildner sc->cpu_cx_count = 0; 5295db2f26eSSascha Wildner cx_ptr = sc->cpu_cx_states; 5305db2f26eSSascha Wildner for (i = 0; i < count; i++) { 5315db2f26eSSascha Wildner pkg = &top->Package.Elements[i + 1]; 5325db2f26eSSascha Wildner if (!ACPI_PKG_VALID(pkg, 4) || 5335db2f26eSSascha Wildner acpi_PkgInt32(pkg, 1, &cx_ptr->type) != 0 || 5345db2f26eSSascha Wildner acpi_PkgInt32(pkg, 2, &cx_ptr->trans_lat) != 0 || 5355db2f26eSSascha Wildner acpi_PkgInt32(pkg, 3, &cx_ptr->power) != 0) { 5365db2f26eSSascha Wildner 5375db2f26eSSascha Wildner device_printf(sc->cpu_dev, "skipping invalid Cx state package\n"); 5385db2f26eSSascha Wildner continue; 5395db2f26eSSascha Wildner } 5405db2f26eSSascha Wildner 5415db2f26eSSascha Wildner /* Validate the state to see if we should use it. */ 5425db2f26eSSascha Wildner switch (cx_ptr->type) { 5435db2f26eSSascha Wildner case ACPI_STATE_C1: 5445db2f26eSSascha Wildner sc->cpu_non_c3 = i; 5455db2f26eSSascha Wildner cx_ptr++; 5465db2f26eSSascha Wildner sc->cpu_cx_count++; 5475db2f26eSSascha Wildner continue; 5485db2f26eSSascha Wildner case ACPI_STATE_C2: 5495db2f26eSSascha Wildner sc->cpu_non_c3 = i; 5505db2f26eSSascha Wildner break; 5515db2f26eSSascha Wildner case ACPI_STATE_C3: 5525db2f26eSSascha Wildner default: 5535db2f26eSSascha Wildner if ((cpu_quirks & CPU_QUIRK_NO_C3) != 0) { 5545db2f26eSSascha Wildner 5555db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 5565db2f26eSSascha Wildner "acpi_cpu%d: C3[%d] not available.\n", 5575db2f26eSSascha Wildner device_get_unit(sc->cpu_dev), i)); 5585db2f26eSSascha Wildner continue; 5595db2f26eSSascha Wildner } 5605db2f26eSSascha Wildner break; 5615db2f26eSSascha Wildner } 5625db2f26eSSascha Wildner 5635db2f26eSSascha Wildner #ifdef notyet 5645db2f26eSSascha Wildner /* Free up any previous register. */ 5655db2f26eSSascha Wildner if (cx_ptr->p_lvlx != NULL) { 5665db2f26eSSascha Wildner bus_release_resource(sc->cpu_dev, 0, 0, cx_ptr->p_lvlx); 5675db2f26eSSascha Wildner cx_ptr->p_lvlx = NULL; 5685db2f26eSSascha Wildner } 5695db2f26eSSascha Wildner #endif 5705db2f26eSSascha Wildner 5715db2f26eSSascha Wildner /* Allocate the control register for C2 or C3. */ 5725db2f26eSSascha Wildner cx_ptr->rid = sc->cpu_parent->cpux_next_rid; 5735db2f26eSSascha Wildner acpi_PkgGas(sc->cpu_dev, pkg, 0, &cx_ptr->res_type, &cx_ptr->rid, &cx_ptr->p_lvlx, 5745db2f26eSSascha Wildner RF_SHAREABLE); 5755db2f26eSSascha Wildner if (cx_ptr->p_lvlx) { 5765db2f26eSSascha Wildner sc->cpu_parent->cpux_next_rid++; 5775db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 5785db2f26eSSascha Wildner "acpi_cpu%d: Got C%d - %d latency\n", 5795db2f26eSSascha Wildner device_get_unit(sc->cpu_dev), cx_ptr->type, 5805db2f26eSSascha Wildner cx_ptr->trans_lat)); 5815db2f26eSSascha Wildner cx_ptr++; 5825db2f26eSSascha Wildner sc->cpu_cx_count++; 5835db2f26eSSascha Wildner } 5845db2f26eSSascha Wildner } 5855db2f26eSSascha Wildner AcpiOsFree(buf.Pointer); 5865db2f26eSSascha Wildner 5875db2f26eSSascha Wildner return (0); 5885db2f26eSSascha Wildner } 5895db2f26eSSascha Wildner 5905db2f26eSSascha Wildner /* 5915db2f26eSSascha Wildner * Call this *after* all CPUs have been attached. 5925db2f26eSSascha Wildner */ 5935db2f26eSSascha Wildner static void 5945db2f26eSSascha Wildner acpi_cpu_startup(void *arg) 5955db2f26eSSascha Wildner { 5965db2f26eSSascha Wildner struct acpi_cpu_softc *sc; 5975db2f26eSSascha Wildner int i; 5985db2f26eSSascha Wildner 5995db2f26eSSascha Wildner /* Get set of CPU devices */ 6005db2f26eSSascha Wildner devclass_get_devices(acpi_cpu_cst_devclass, &cpu_devices, &cpu_ndevices); 6015db2f26eSSascha Wildner 6025db2f26eSSascha Wildner /* 6035db2f26eSSascha Wildner * Setup any quirks that might necessary now that we have probed 6045db2f26eSSascha Wildner * all the CPUs 6055db2f26eSSascha Wildner */ 6065db2f26eSSascha Wildner acpi_cpu_quirks(); 6075db2f26eSSascha Wildner 6085db2f26eSSascha Wildner cpu_cx_count = 0; 6095db2f26eSSascha Wildner if (cpu_cx_generic) { 6105db2f26eSSascha Wildner /* 6115db2f26eSSascha Wildner * We are using generic Cx mode, probe for available Cx states 6125db2f26eSSascha Wildner * for all processors. 6135db2f26eSSascha Wildner */ 6145db2f26eSSascha Wildner for (i = 0; i < cpu_ndevices; i++) { 6155db2f26eSSascha Wildner sc = device_get_softc(cpu_devices[i]); 6165db2f26eSSascha Wildner acpi_cpu_generic_cx_probe(sc); 6175db2f26eSSascha Wildner if (sc->cpu_cx_count > cpu_cx_count) 6185db2f26eSSascha Wildner cpu_cx_count = sc->cpu_cx_count; 6195db2f26eSSascha Wildner } 6205db2f26eSSascha Wildner 6215db2f26eSSascha Wildner /* 6225db2f26eSSascha Wildner * Find the highest Cx state common to all CPUs 6235db2f26eSSascha Wildner * in the system, taking quirks into account. 6245db2f26eSSascha Wildner */ 6255db2f26eSSascha Wildner for (i = 0; i < cpu_ndevices; i++) { 6265db2f26eSSascha Wildner sc = device_get_softc(cpu_devices[i]); 6275db2f26eSSascha Wildner if (sc->cpu_cx_count < cpu_cx_count) 6285db2f26eSSascha Wildner cpu_cx_count = sc->cpu_cx_count; 6295db2f26eSSascha Wildner } 6305db2f26eSSascha Wildner } else { 6315db2f26eSSascha Wildner /* 6325db2f26eSSascha Wildner * We are using _CST mode, remove C3 state if necessary. 6335db2f26eSSascha Wildner * Update the largest Cx state supported in the global cpu_cx_count. 6345db2f26eSSascha Wildner * It will be used in the global Cx sysctl handler. 6355db2f26eSSascha Wildner * As we now know for sure that we will be using _CST mode 6365db2f26eSSascha Wildner * install our notify handler. 6375db2f26eSSascha Wildner */ 6385db2f26eSSascha Wildner for (i = 0; i < cpu_ndevices; i++) { 6395db2f26eSSascha Wildner sc = device_get_softc(cpu_devices[i]); 6405db2f26eSSascha Wildner if (cpu_quirks & CPU_QUIRK_NO_C3) { 6415db2f26eSSascha Wildner sc->cpu_cx_count = sc->cpu_non_c3 + 1; 6425db2f26eSSascha Wildner } 6435db2f26eSSascha Wildner if (sc->cpu_cx_count > cpu_cx_count) 6445db2f26eSSascha Wildner cpu_cx_count = sc->cpu_cx_count; 6455db2f26eSSascha Wildner sc->cpu_parent->cpux_cst_notify = acpi_cpu_cst_notify; 6465db2f26eSSascha Wildner } 6475db2f26eSSascha Wildner } 6485db2f26eSSascha Wildner 6495db2f26eSSascha Wildner /* Perform Cx final initialization. */ 6505db2f26eSSascha Wildner for (i = 0; i < cpu_ndevices; i++) { 6515db2f26eSSascha Wildner sc = device_get_softc(cpu_devices[i]); 6525db2f26eSSascha Wildner acpi_cpu_startup_cx(sc); 6535db2f26eSSascha Wildner 6545db2f26eSSascha Wildner if (sc->cpu_parent->glob_sysctl_tree != NULL) { 6555db2f26eSSascha Wildner struct acpi_cpux_softc *cpux = sc->cpu_parent; 6565db2f26eSSascha Wildner 6575db2f26eSSascha Wildner /* Add a sysctl handler to handle global Cx lowest setting */ 6585db2f26eSSascha Wildner SYSCTL_ADD_PROC(&cpux->glob_sysctl_ctx, 6595db2f26eSSascha Wildner SYSCTL_CHILDREN(cpux->glob_sysctl_tree), 6605db2f26eSSascha Wildner OID_AUTO, "cx_lowest", 6615db2f26eSSascha Wildner CTLTYPE_STRING | CTLFLAG_RW, NULL, 0, 6625db2f26eSSascha Wildner acpi_cpu_global_cx_lowest_sysctl, "A", 6635db2f26eSSascha Wildner "Global lowest Cx sleep state to use"); 6645db2f26eSSascha Wildner } 6655db2f26eSSascha Wildner } 6665db2f26eSSascha Wildner 6675db2f26eSSascha Wildner /* Take over idling from cpu_idle_default(). */ 6685db2f26eSSascha Wildner cpu_cx_lowest = 0; 6695db2f26eSSascha Wildner cpu_disable_idle = FALSE; 6705db2f26eSSascha Wildner cpu_idle_hook = acpi_cpu_idle; 6715db2f26eSSascha Wildner } 6725db2f26eSSascha Wildner 6735db2f26eSSascha Wildner static void 6745db2f26eSSascha Wildner acpi_cpu_cx_list(struct acpi_cpu_softc *sc) 6755db2f26eSSascha Wildner { 6765db2f26eSSascha Wildner struct sbuf sb; 6775db2f26eSSascha Wildner int i; 6785db2f26eSSascha Wildner 6795db2f26eSSascha Wildner /* 6805db2f26eSSascha Wildner * Set up the list of Cx states 6815db2f26eSSascha Wildner */ 6825db2f26eSSascha Wildner sc->cpu_non_c3 = 0; 6835db2f26eSSascha Wildner sbuf_new(&sb, sc->cpu_cx_supported, sizeof(sc->cpu_cx_supported), 6845db2f26eSSascha Wildner SBUF_FIXEDLEN); 6855db2f26eSSascha Wildner for (i = 0; i < sc->cpu_cx_count; i++) { 6865db2f26eSSascha Wildner sbuf_printf(&sb, "C%d/%d ", i + 1, sc->cpu_cx_states[i].trans_lat); 6875db2f26eSSascha Wildner if (sc->cpu_cx_states[i].type < ACPI_STATE_C3) 6885db2f26eSSascha Wildner sc->cpu_non_c3 = i; 6895db2f26eSSascha Wildner } 6905db2f26eSSascha Wildner sbuf_trim(&sb); 6915db2f26eSSascha Wildner sbuf_finish(&sb); 6925db2f26eSSascha Wildner } 6935db2f26eSSascha Wildner 6945db2f26eSSascha Wildner static void 6955db2f26eSSascha Wildner acpi_cpu_startup_cx(struct acpi_cpu_softc *sc) 6965db2f26eSSascha Wildner { 6975db2f26eSSascha Wildner struct acpi_cpux_softc *cpux = sc->cpu_parent; 6985db2f26eSSascha Wildner 6995db2f26eSSascha Wildner acpi_cpu_cx_list(sc); 7005db2f26eSSascha Wildner 7015db2f26eSSascha Wildner SYSCTL_ADD_STRING(&cpux->pcpu_sysctl_ctx, 7025db2f26eSSascha Wildner SYSCTL_CHILDREN(cpux->pcpu_sysctl_tree), 7035db2f26eSSascha Wildner OID_AUTO, "cx_supported", CTLFLAG_RD, 7045db2f26eSSascha Wildner sc->cpu_cx_supported, 0, 7055db2f26eSSascha Wildner "Cx/microsecond values for supported Cx states"); 7065db2f26eSSascha Wildner SYSCTL_ADD_PROC(&cpux->pcpu_sysctl_ctx, 7075db2f26eSSascha Wildner SYSCTL_CHILDREN(cpux->pcpu_sysctl_tree), 7085db2f26eSSascha Wildner OID_AUTO, "cx_lowest", CTLTYPE_STRING | CTLFLAG_RW, 7095db2f26eSSascha Wildner (void *)sc, 0, acpi_cpu_cx_lowest_sysctl, "A", 7105db2f26eSSascha Wildner "lowest Cx sleep state to use"); 7115db2f26eSSascha Wildner SYSCTL_ADD_PROC(&cpux->pcpu_sysctl_ctx, 7125db2f26eSSascha Wildner SYSCTL_CHILDREN(cpux->pcpu_sysctl_tree), 7135db2f26eSSascha Wildner OID_AUTO, "cx_usage", CTLTYPE_STRING | CTLFLAG_RD, 7145db2f26eSSascha Wildner (void *)sc, 0, acpi_cpu_usage_sysctl, "A", 7155db2f26eSSascha Wildner "percent usage for each Cx state"); 7165db2f26eSSascha Wildner 7175db2f26eSSascha Wildner #ifdef notyet 7185db2f26eSSascha Wildner /* Signal platform that we can handle _CST notification. */ 7195db2f26eSSascha Wildner if (!cpu_cx_generic && cpu_cst_cnt != 0) { 7205db2f26eSSascha Wildner ACPI_LOCK(acpi); 7215db2f26eSSascha Wildner AcpiOsWritePort(cpu_smi_cmd, cpu_cst_cnt, 8); 7225db2f26eSSascha Wildner ACPI_UNLOCK(acpi); 7235db2f26eSSascha Wildner } 7245db2f26eSSascha Wildner #endif 7255db2f26eSSascha Wildner } 7265db2f26eSSascha Wildner 7275db2f26eSSascha Wildner /* 7285db2f26eSSascha Wildner * Idle the CPU in the lowest state possible. This function is called with 7295db2f26eSSascha Wildner * interrupts disabled. Note that once it re-enables interrupts, a task 7305db2f26eSSascha Wildner * switch can occur so do not access shared data (i.e. the softc) after 7315db2f26eSSascha Wildner * interrupts are re-enabled. 7325db2f26eSSascha Wildner */ 7335db2f26eSSascha Wildner static void 7345db2f26eSSascha Wildner acpi_cpu_idle(void) 7355db2f26eSSascha Wildner { 7365db2f26eSSascha Wildner struct acpi_cpu_softc *sc; 7375db2f26eSSascha Wildner struct acpi_cx *cx_next; 7385db2f26eSSascha Wildner uint64_t start_time, end_time; 7395db2f26eSSascha Wildner int bm_active, cx_next_idx, i; 7405db2f26eSSascha Wildner 7415db2f26eSSascha Wildner /* If disabled, return immediately. */ 7425db2f26eSSascha Wildner if (cpu_disable_idle) { 7435db2f26eSSascha Wildner ACPI_ENABLE_IRQS(); 7445db2f26eSSascha Wildner return; 7455db2f26eSSascha Wildner } 7465db2f26eSSascha Wildner 7475db2f26eSSascha Wildner /* 7485db2f26eSSascha Wildner * Look up our CPU id to get our softc. If it's NULL, we'll use C1 7495db2f26eSSascha Wildner * since there is no ACPI processor object for this CPU. This occurs 7505db2f26eSSascha Wildner * for logical CPUs in the HTT case. 7515db2f26eSSascha Wildner */ 7525db2f26eSSascha Wildner sc = cpu_softc[mdcpu->mi.gd_cpuid]; 7535db2f26eSSascha Wildner if (sc == NULL) { 7545db2f26eSSascha Wildner acpi_cpu_c1(); 7555db2f26eSSascha Wildner return; 7565db2f26eSSascha Wildner } 7575db2f26eSSascha Wildner 7585db2f26eSSascha Wildner /* Find the lowest state that has small enough latency. */ 7595db2f26eSSascha Wildner cx_next_idx = 0; 7605db2f26eSSascha Wildner for (i = sc->cpu_cx_lowest; i >= 0; i--) { 7615db2f26eSSascha Wildner if (sc->cpu_cx_states[i].trans_lat * 3 <= sc->cpu_prev_sleep) { 7625db2f26eSSascha Wildner cx_next_idx = i; 7635db2f26eSSascha Wildner break; 7645db2f26eSSascha Wildner } 7655db2f26eSSascha Wildner } 7665db2f26eSSascha Wildner 7675db2f26eSSascha Wildner /* 7685db2f26eSSascha Wildner * Check for bus master activity. If there was activity, clear 7695db2f26eSSascha Wildner * the bit and use the lowest non-C3 state. Note that the USB 7705db2f26eSSascha Wildner * driver polling for new devices keeps this bit set all the 7715db2f26eSSascha Wildner * time if USB is loaded. 7725db2f26eSSascha Wildner */ 7735db2f26eSSascha Wildner if ((cpu_quirks & CPU_QUIRK_NO_BM_CTRL) == 0) { 7745db2f26eSSascha Wildner AcpiReadBitRegister(ACPI_BITREG_BUS_MASTER_STATUS, &bm_active); 7755db2f26eSSascha Wildner if (bm_active != 0) { 7765db2f26eSSascha Wildner AcpiWriteBitRegister(ACPI_BITREG_BUS_MASTER_STATUS, 1); 7775db2f26eSSascha Wildner cx_next_idx = min(cx_next_idx, sc->cpu_non_c3); 7785db2f26eSSascha Wildner } 7795db2f26eSSascha Wildner } 7805db2f26eSSascha Wildner 7815db2f26eSSascha Wildner /* Select the next state and update statistics. */ 7825db2f26eSSascha Wildner cx_next = &sc->cpu_cx_states[cx_next_idx]; 7835db2f26eSSascha Wildner sc->cpu_cx_stats[cx_next_idx]++; 7845db2f26eSSascha Wildner KASSERT(cx_next->type != ACPI_STATE_C0, ("acpi_cpu_idle: C0 sleep")); 7855db2f26eSSascha Wildner 7865db2f26eSSascha Wildner /* 7875db2f26eSSascha Wildner * Execute HLT (or equivalent) and wait for an interrupt. We can't 7885db2f26eSSascha Wildner * calculate the time spent in C1 since the place we wake up is an 7895db2f26eSSascha Wildner * ISR. Assume we slept half of quantum and return. 7905db2f26eSSascha Wildner */ 7915db2f26eSSascha Wildner if (cx_next->type == ACPI_STATE_C1) { 7925db2f26eSSascha Wildner sc->cpu_prev_sleep = (sc->cpu_prev_sleep * 3 + 500000 / hz) / 4; 7935db2f26eSSascha Wildner acpi_cpu_c1(); 7945db2f26eSSascha Wildner return; 7955db2f26eSSascha Wildner } 7965db2f26eSSascha Wildner 7975db2f26eSSascha Wildner /* 7985db2f26eSSascha Wildner * For C3, disable bus master arbitration and enable bus master wake 7995db2f26eSSascha Wildner * if BM control is available, otherwise flush the CPU cache. 8005db2f26eSSascha Wildner */ 8015db2f26eSSascha Wildner if (cx_next->type == ACPI_STATE_C3) { 8025db2f26eSSascha Wildner if ((cpu_quirks & CPU_QUIRK_NO_BM_CTRL) == 0) { 8035db2f26eSSascha Wildner AcpiWriteBitRegister(ACPI_BITREG_ARB_DISABLE, 1); 8045db2f26eSSascha Wildner AcpiWriteBitRegister(ACPI_BITREG_BUS_MASTER_RLD, 1); 8055db2f26eSSascha Wildner } else 8065db2f26eSSascha Wildner ACPI_FLUSH_CPU_CACHE(); 8075db2f26eSSascha Wildner } 8085db2f26eSSascha Wildner 8095db2f26eSSascha Wildner /* 8105db2f26eSSascha Wildner * Read from P_LVLx to enter C2(+), checking time spent asleep. 8115db2f26eSSascha Wildner * Use the ACPI timer for measuring sleep time. Since we need to 8125db2f26eSSascha Wildner * get the time very close to the CPU start/stop clock logic, this 8135db2f26eSSascha Wildner * is the only reliable time source. 8145db2f26eSSascha Wildner */ 8155db2f26eSSascha Wildner AcpiRead(&start_time, &AcpiGbl_FADT.XPmTimerBlock); 8165db2f26eSSascha Wildner CPU_GET_REG(cx_next->p_lvlx, 1); 8175db2f26eSSascha Wildner 8185db2f26eSSascha Wildner /* 8195db2f26eSSascha Wildner * Read the end time twice. Since it may take an arbitrary time 8205db2f26eSSascha Wildner * to enter the idle state, the first read may be executed before 8215db2f26eSSascha Wildner * the processor has stopped. Doing it again provides enough 8225db2f26eSSascha Wildner * margin that we are certain to have a correct value. 8235db2f26eSSascha Wildner */ 8245db2f26eSSascha Wildner AcpiRead(&end_time, &AcpiGbl_FADT.XPmTimerBlock); 8255db2f26eSSascha Wildner AcpiRead(&end_time, &AcpiGbl_FADT.XPmTimerBlock); 8265db2f26eSSascha Wildner 8275db2f26eSSascha Wildner /* Enable bus master arbitration and disable bus master wakeup. */ 8285db2f26eSSascha Wildner if (cx_next->type == ACPI_STATE_C3) { 8295db2f26eSSascha Wildner if ((cpu_quirks & CPU_QUIRK_NO_BM_CTRL) == 0) { 8305db2f26eSSascha Wildner AcpiWriteBitRegister(ACPI_BITREG_ARB_DISABLE, 0); 8315db2f26eSSascha Wildner AcpiWriteBitRegister(ACPI_BITREG_BUS_MASTER_RLD, 0); 8325db2f26eSSascha Wildner } 8335db2f26eSSascha Wildner } 8345db2f26eSSascha Wildner ACPI_ENABLE_IRQS(); 8355db2f26eSSascha Wildner 8365db2f26eSSascha Wildner /* Find the actual time asleep in microseconds. */ 8375db2f26eSSascha Wildner end_time = acpi_TimerDelta(end_time, start_time); 8385db2f26eSSascha Wildner sc->cpu_prev_sleep = (sc->cpu_prev_sleep * 3 + PM_USEC(end_time)) / 4; 8395db2f26eSSascha Wildner } 8405db2f26eSSascha Wildner 8415db2f26eSSascha Wildner /* 8425db2f26eSSascha Wildner * Re-evaluate the _CST object when we are notified that it changed. 8435db2f26eSSascha Wildner * 8445db2f26eSSascha Wildner * XXX Re-evaluation disabled until locking is done. 8455db2f26eSSascha Wildner */ 8465db2f26eSSascha Wildner static void 8475db2f26eSSascha Wildner acpi_cpu_cst_notify(device_t dev) 8485db2f26eSSascha Wildner { 8495db2f26eSSascha Wildner struct acpi_cpu_softc *sc = device_get_softc(dev); 8505db2f26eSSascha Wildner struct acpi_cpu_softc *isc; 8515db2f26eSSascha Wildner int i; 8525db2f26eSSascha Wildner 8535db2f26eSSascha Wildner /* Update the list of Cx states. */ 8545db2f26eSSascha Wildner acpi_cpu_cx_cst(sc); 8555db2f26eSSascha Wildner acpi_cpu_cx_list(sc); 8565db2f26eSSascha Wildner 8575db2f26eSSascha Wildner /* Update the new lowest useable Cx state for all CPUs. */ 8585db2f26eSSascha Wildner crit_enter(); 8595db2f26eSSascha Wildner cpu_cx_count = 0; 8605db2f26eSSascha Wildner for (i = 0; i < cpu_ndevices; i++) { 8615db2f26eSSascha Wildner isc = device_get_softc(cpu_devices[i]); 8625db2f26eSSascha Wildner if (isc->cpu_cx_count > cpu_cx_count) 8635db2f26eSSascha Wildner cpu_cx_count = isc->cpu_cx_count; 8645db2f26eSSascha Wildner } 8655db2f26eSSascha Wildner crit_exit(); 8665db2f26eSSascha Wildner } 8675db2f26eSSascha Wildner 8685db2f26eSSascha Wildner static int 8695db2f26eSSascha Wildner acpi_cpu_quirks(void) 8705db2f26eSSascha Wildner { 8715db2f26eSSascha Wildner device_t acpi_dev; 8725db2f26eSSascha Wildner uint32_t val; 8735db2f26eSSascha Wildner 8745db2f26eSSascha Wildner ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__); 8755db2f26eSSascha Wildner 8765db2f26eSSascha Wildner /* 8775db2f26eSSascha Wildner * Bus mastering arbitration control is needed to keep caches coherent 8785db2f26eSSascha Wildner * while sleeping in C3. If it's not present but a working flush cache 8795db2f26eSSascha Wildner * instruction is present, flush the caches before entering C3 instead. 8805db2f26eSSascha Wildner * Otherwise, just disable C3 completely. 8815db2f26eSSascha Wildner */ 8825db2f26eSSascha Wildner if (AcpiGbl_FADT.Pm2ControlBlock == 0 || 8835db2f26eSSascha Wildner AcpiGbl_FADT.Pm2ControlLength == 0) { 8845db2f26eSSascha Wildner if ((AcpiGbl_FADT.Flags & ACPI_FADT_WBINVD) && 8855db2f26eSSascha Wildner (AcpiGbl_FADT.Flags & ACPI_FADT_WBINVD_FLUSH) == 0) { 8865db2f26eSSascha Wildner cpu_quirks |= CPU_QUIRK_NO_BM_CTRL; 8875db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 8885db2f26eSSascha Wildner "acpi_cpu: no BM control, using flush cache method\n")); 8895db2f26eSSascha Wildner } else { 8905db2f26eSSascha Wildner cpu_quirks |= CPU_QUIRK_NO_C3; 8915db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 8925db2f26eSSascha Wildner "acpi_cpu: no BM control, C3 not available\n")); 8935db2f26eSSascha Wildner } 8945db2f26eSSascha Wildner } 8955db2f26eSSascha Wildner 8965db2f26eSSascha Wildner /* 8975db2f26eSSascha Wildner * If we are using generic Cx mode, C3 on multiple CPUs requires using 8985db2f26eSSascha Wildner * the expensive flush cache instruction. 8995db2f26eSSascha Wildner */ 9005db2f26eSSascha Wildner if (cpu_cx_generic && ncpus > 1) { 9015db2f26eSSascha Wildner cpu_quirks |= CPU_QUIRK_NO_BM_CTRL; 9025db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 9035db2f26eSSascha Wildner "acpi_cpu: SMP, using flush cache mode for C3\n")); 9045db2f26eSSascha Wildner } 9055db2f26eSSascha Wildner 9065db2f26eSSascha Wildner /* Look for various quirks of the PIIX4 part. */ 9075db2f26eSSascha Wildner acpi_dev = pci_find_device(PCI_VENDOR_INTEL, PCI_DEVICE_82371AB_3); 9085db2f26eSSascha Wildner if (acpi_dev != NULL) { 9095db2f26eSSascha Wildner switch (pci_get_revid(acpi_dev)) { 9105db2f26eSSascha Wildner /* 9115db2f26eSSascha Wildner * Disable C3 support for all PIIX4 chipsets. Some of these parts 9125db2f26eSSascha Wildner * do not report the BMIDE status to the BM status register and 9135db2f26eSSascha Wildner * others have a livelock bug if Type-F DMA is enabled. Linux 9145db2f26eSSascha Wildner * works around the BMIDE bug by reading the BM status directly 9155db2f26eSSascha Wildner * but we take the simpler approach of disabling C3 for these 9165db2f26eSSascha Wildner * parts. 9175db2f26eSSascha Wildner * 9185db2f26eSSascha Wildner * See erratum #18 ("C3 Power State/BMIDE and Type-F DMA 9195db2f26eSSascha Wildner * Livelock") from the January 2002 PIIX4 specification update. 9205db2f26eSSascha Wildner * Applies to all PIIX4 models. 9215db2f26eSSascha Wildner * 9225db2f26eSSascha Wildner * Also, make sure that all interrupts cause a "Stop Break" 9235db2f26eSSascha Wildner * event to exit from C2 state. 9245db2f26eSSascha Wildner * Also, BRLD_EN_BM (ACPI_BITREG_BUS_MASTER_RLD in ACPI-speak) 9255db2f26eSSascha Wildner * should be set to zero, otherwise it causes C2 to short-sleep. 9265db2f26eSSascha Wildner * PIIX4 doesn't properly support C3 and bus master activity 9275db2f26eSSascha Wildner * need not break out of C2. 9285db2f26eSSascha Wildner */ 9295db2f26eSSascha Wildner case PCI_REVISION_A_STEP: 9305db2f26eSSascha Wildner case PCI_REVISION_B_STEP: 9315db2f26eSSascha Wildner case PCI_REVISION_4E: 9325db2f26eSSascha Wildner case PCI_REVISION_4M: 9335db2f26eSSascha Wildner cpu_quirks |= CPU_QUIRK_NO_C3; 9345db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 9355db2f26eSSascha Wildner "acpi_cpu: working around PIIX4 bug, disabling C3\n")); 9365db2f26eSSascha Wildner 9375db2f26eSSascha Wildner val = pci_read_config(acpi_dev, PIIX4_DEVACTB_REG, 4); 9385db2f26eSSascha Wildner if ((val & PIIX4_STOP_BREAK_MASK) != PIIX4_STOP_BREAK_MASK) { 9395db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 9405db2f26eSSascha Wildner "acpi_cpu: PIIX4: enabling IRQs to generate Stop Break\n")); 9415db2f26eSSascha Wildner val |= PIIX4_STOP_BREAK_MASK; 9425db2f26eSSascha Wildner pci_write_config(acpi_dev, PIIX4_DEVACTB_REG, val, 4); 9435db2f26eSSascha Wildner } 9445db2f26eSSascha Wildner AcpiReadBitRegister(ACPI_BITREG_BUS_MASTER_RLD, &val); 9455db2f26eSSascha Wildner if (val) { 9465db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 9475db2f26eSSascha Wildner "acpi_cpu: PIIX4: reset BRLD_EN_BM\n")); 9485db2f26eSSascha Wildner AcpiWriteBitRegister(ACPI_BITREG_BUS_MASTER_RLD, 0); 9495db2f26eSSascha Wildner } 9505db2f26eSSascha Wildner break; 9515db2f26eSSascha Wildner default: 9525db2f26eSSascha Wildner break; 9535db2f26eSSascha Wildner } 9545db2f26eSSascha Wildner } 9555db2f26eSSascha Wildner 9565db2f26eSSascha Wildner return (0); 9575db2f26eSSascha Wildner } 9585db2f26eSSascha Wildner 9595db2f26eSSascha Wildner static int 9605db2f26eSSascha Wildner acpi_cpu_usage_sysctl(SYSCTL_HANDLER_ARGS) 9615db2f26eSSascha Wildner { 9625db2f26eSSascha Wildner struct acpi_cpu_softc *sc; 9635db2f26eSSascha Wildner struct sbuf sb; 9645db2f26eSSascha Wildner char buf[128]; 9655db2f26eSSascha Wildner int i; 9665db2f26eSSascha Wildner uintmax_t fract, sum, whole; 9675db2f26eSSascha Wildner 9685db2f26eSSascha Wildner sc = (struct acpi_cpu_softc *) arg1; 9695db2f26eSSascha Wildner sum = 0; 9705db2f26eSSascha Wildner for (i = 0; i < sc->cpu_cx_count; i++) 9715db2f26eSSascha Wildner sum += sc->cpu_cx_stats[i]; 9725db2f26eSSascha Wildner sbuf_new(&sb, buf, sizeof(buf), SBUF_FIXEDLEN); 9735db2f26eSSascha Wildner for (i = 0; i < sc->cpu_cx_count; i++) { 9745db2f26eSSascha Wildner if (sum > 0) { 9755db2f26eSSascha Wildner whole = (uintmax_t)sc->cpu_cx_stats[i] * 100; 9765db2f26eSSascha Wildner fract = (whole % sum) * 100; 9775db2f26eSSascha Wildner sbuf_printf(&sb, "%u.%02u%% ", (u_int)(whole / sum), 9785db2f26eSSascha Wildner (u_int)(fract / sum)); 9795db2f26eSSascha Wildner } else 9805db2f26eSSascha Wildner sbuf_printf(&sb, "0.00%% "); 9815db2f26eSSascha Wildner } 9825db2f26eSSascha Wildner sbuf_printf(&sb, "last %dus", sc->cpu_prev_sleep); 9835db2f26eSSascha Wildner sbuf_trim(&sb); 9845db2f26eSSascha Wildner sbuf_finish(&sb); 9855db2f26eSSascha Wildner sysctl_handle_string(oidp, sbuf_data(&sb), sbuf_len(&sb), req); 9865db2f26eSSascha Wildner sbuf_delete(&sb); 9875db2f26eSSascha Wildner 9885db2f26eSSascha Wildner return (0); 9895db2f26eSSascha Wildner } 9905db2f26eSSascha Wildner 9915db2f26eSSascha Wildner static int 9925db2f26eSSascha Wildner acpi_cpu_set_cx_lowest(struct acpi_cpu_softc *sc, int val) 9935db2f26eSSascha Wildner { 9945db2f26eSSascha Wildner int i, old_lowest, error = 0; 9955db2f26eSSascha Wildner uint32_t old_type, type; 9965db2f26eSSascha Wildner 9975db2f26eSSascha Wildner get_mplock(); 9985db2f26eSSascha Wildner 9995db2f26eSSascha Wildner old_lowest = atomic_swap_int(&sc->cpu_cx_lowest, val); 10005db2f26eSSascha Wildner 10015db2f26eSSascha Wildner old_type = sc->cpu_cx_states[old_lowest].type; 10025db2f26eSSascha Wildner type = sc->cpu_cx_states[val].type; 10035db2f26eSSascha Wildner if (old_type == ACPI_STATE_C3 && type != ACPI_STATE_C3) { 10045db2f26eSSascha Wildner KKASSERT(cpu_c3_ncpus > 0); 10055db2f26eSSascha Wildner if (atomic_fetchadd_int(&cpu_c3_ncpus, -1) == 1) { 10065db2f26eSSascha Wildner /* 10075db2f26eSSascha Wildner * All of the CPUs exit C3 state, use a better 10085db2f26eSSascha Wildner * one shot timer. 10095db2f26eSSascha Wildner */ 10105db2f26eSSascha Wildner error = cputimer_intr_select_caps(CPUTIMER_INTR_CAP_NONE); 10115db2f26eSSascha Wildner KKASSERT(!error); 10125db2f26eSSascha Wildner cputimer_intr_restart(); 10135db2f26eSSascha Wildner } 10145db2f26eSSascha Wildner } else if (type == ACPI_STATE_C3 && old_type != ACPI_STATE_C3) { 10155db2f26eSSascha Wildner if (atomic_fetchadd_int(&cpu_c3_ncpus, 1) == 0) { 10165db2f26eSSascha Wildner /* 10175db2f26eSSascha Wildner * When the first CPU enters C3 state, switch 10185db2f26eSSascha Wildner * to an one shot timer, which could handle 10195db2f26eSSascha Wildner * C3 state, i.e. the timer will not hang. 10205db2f26eSSascha Wildner */ 10215db2f26eSSascha Wildner error = cputimer_intr_select_caps(CPUTIMER_INTR_CAP_PS); 10225db2f26eSSascha Wildner if (!error) { 10235db2f26eSSascha Wildner cputimer_intr_restart(); 10245db2f26eSSascha Wildner } else { 10255db2f26eSSascha Wildner kprintf("no suitable intr cputimer found\n"); 10265db2f26eSSascha Wildner 10275db2f26eSSascha Wildner /* Restore */ 10285db2f26eSSascha Wildner sc->cpu_cx_lowest = old_lowest; 10295db2f26eSSascha Wildner atomic_fetchadd_int(&cpu_c3_ncpus, -1); 10305db2f26eSSascha Wildner } 10315db2f26eSSascha Wildner } 10325db2f26eSSascha Wildner } 10335db2f26eSSascha Wildner 10345db2f26eSSascha Wildner rel_mplock(); 10355db2f26eSSascha Wildner 10365db2f26eSSascha Wildner if (error) 10375db2f26eSSascha Wildner return error; 10385db2f26eSSascha Wildner 10395db2f26eSSascha Wildner /* If not disabling, cache the new lowest non-C3 state. */ 10405db2f26eSSascha Wildner sc->cpu_non_c3 = 0; 10415db2f26eSSascha Wildner for (i = sc->cpu_cx_lowest; i >= 0; i--) { 10425db2f26eSSascha Wildner if (sc->cpu_cx_states[i].type < ACPI_STATE_C3) { 10435db2f26eSSascha Wildner sc->cpu_non_c3 = i; 10445db2f26eSSascha Wildner break; 10455db2f26eSSascha Wildner } 10465db2f26eSSascha Wildner } 10475db2f26eSSascha Wildner 10485db2f26eSSascha Wildner /* Reset the statistics counters. */ 10495db2f26eSSascha Wildner bzero(sc->cpu_cx_stats, sizeof(sc->cpu_cx_stats)); 10505db2f26eSSascha Wildner return (0); 10515db2f26eSSascha Wildner } 10525db2f26eSSascha Wildner 10535db2f26eSSascha Wildner static int 10545db2f26eSSascha Wildner acpi_cpu_cx_lowest_sysctl(SYSCTL_HANDLER_ARGS) 10555db2f26eSSascha Wildner { 10565db2f26eSSascha Wildner struct acpi_cpu_softc *sc; 10575db2f26eSSascha Wildner char state[8]; 10585db2f26eSSascha Wildner int val, error; 10595db2f26eSSascha Wildner 10605db2f26eSSascha Wildner sc = (struct acpi_cpu_softc *) arg1; 10615db2f26eSSascha Wildner ksnprintf(state, sizeof(state), "C%d", sc->cpu_cx_lowest + 1); 10625db2f26eSSascha Wildner error = sysctl_handle_string(oidp, state, sizeof(state), req); 10635db2f26eSSascha Wildner if (error != 0 || req->newptr == NULL) 10645db2f26eSSascha Wildner return (error); 10655db2f26eSSascha Wildner if (strlen(state) < 2 || toupper(state[0]) != 'C') 10665db2f26eSSascha Wildner return (EINVAL); 10675db2f26eSSascha Wildner val = (int) strtol(state + 1, NULL, 10) - 1; 10685db2f26eSSascha Wildner if (val < 0 || val > sc->cpu_cx_count - 1) 10695db2f26eSSascha Wildner return (EINVAL); 10705db2f26eSSascha Wildner 10715db2f26eSSascha Wildner crit_enter(); 10725db2f26eSSascha Wildner error = acpi_cpu_set_cx_lowest(sc, val); 10735db2f26eSSascha Wildner crit_exit(); 10745db2f26eSSascha Wildner 10755db2f26eSSascha Wildner return error; 10765db2f26eSSascha Wildner } 10775db2f26eSSascha Wildner 10785db2f26eSSascha Wildner static int 10795db2f26eSSascha Wildner acpi_cpu_global_cx_lowest_sysctl(SYSCTL_HANDLER_ARGS) 10805db2f26eSSascha Wildner { 10815db2f26eSSascha Wildner struct acpi_cpu_softc *sc; 10825db2f26eSSascha Wildner char state[8]; 10835db2f26eSSascha Wildner int val, error, i; 10845db2f26eSSascha Wildner 10855db2f26eSSascha Wildner ksnprintf(state, sizeof(state), "C%d", cpu_cx_lowest + 1); 10865db2f26eSSascha Wildner error = sysctl_handle_string(oidp, state, sizeof(state), req); 10875db2f26eSSascha Wildner if (error != 0 || req->newptr == NULL) 10885db2f26eSSascha Wildner return (error); 10895db2f26eSSascha Wildner if (strlen(state) < 2 || toupper(state[0]) != 'C') 10905db2f26eSSascha Wildner return (EINVAL); 10915db2f26eSSascha Wildner val = (int) strtol(state + 1, NULL, 10) - 1; 10925db2f26eSSascha Wildner if (val < 0 || val > cpu_cx_count - 1) 10935db2f26eSSascha Wildner return (EINVAL); 10945db2f26eSSascha Wildner cpu_cx_lowest = val; 10955db2f26eSSascha Wildner 10965db2f26eSSascha Wildner /* Update the new lowest useable Cx state for all CPUs. */ 10975db2f26eSSascha Wildner crit_enter(); 10985db2f26eSSascha Wildner for (i = 0; i < cpu_ndevices; i++) { 10995db2f26eSSascha Wildner sc = device_get_softc(cpu_devices[i]); 11005db2f26eSSascha Wildner error = acpi_cpu_set_cx_lowest(sc, val); 11015db2f26eSSascha Wildner if (error) { 11025db2f26eSSascha Wildner KKASSERT(i == 0); 11035db2f26eSSascha Wildner break; 11045db2f26eSSascha Wildner } 11055db2f26eSSascha Wildner } 11065db2f26eSSascha Wildner crit_exit(); 11075db2f26eSSascha Wildner 11085db2f26eSSascha Wildner return error; 11095db2f26eSSascha Wildner } 11105db2f26eSSascha Wildner 11115db2f26eSSascha Wildner /* 11125db2f26eSSascha Wildner * Put the CPU in C1 in a machine-dependant way. 11135db2f26eSSascha Wildner * XXX: shouldn't be here! 11145db2f26eSSascha Wildner */ 11155db2f26eSSascha Wildner static void 11165db2f26eSSascha Wildner acpi_cpu_c1(void) 11175db2f26eSSascha Wildner { 11185db2f26eSSascha Wildner #ifdef __ia64__ 11195db2f26eSSascha Wildner ia64_call_pal_static(PAL_HALT_LIGHT, 0, 0, 0); 11205db2f26eSSascha Wildner #else 11215db2f26eSSascha Wildner splz(); 11225db2f26eSSascha Wildner if ((mycpu->gd_reqflags & RQF_IDLECHECK_WK_MASK) == 0) 11235db2f26eSSascha Wildner __asm __volatile("sti; hlt"); 11245db2f26eSSascha Wildner else 11255db2f26eSSascha Wildner __asm __volatile("sti; pause"); 11265db2f26eSSascha Wildner #endif /* !__ia64__ */ 11275db2f26eSSascha Wildner } 1128