15db2f26eSSascha Wildner /*- 25db2f26eSSascha Wildner * Copyright (c) 2003-2005 Nate Lawson (SDG) 35db2f26eSSascha Wildner * Copyright (c) 2001 Michael Smith 45db2f26eSSascha Wildner * All rights reserved. 55db2f26eSSascha Wildner * 65db2f26eSSascha Wildner * Redistribution and use in source and binary forms, with or without 75db2f26eSSascha Wildner * modification, are permitted provided that the following conditions 85db2f26eSSascha Wildner * are met: 95db2f26eSSascha Wildner * 1. Redistributions of source code must retain the above copyright 105db2f26eSSascha Wildner * notice, this list of conditions and the following disclaimer. 115db2f26eSSascha Wildner * 2. Redistributions in binary form must reproduce the above copyright 125db2f26eSSascha Wildner * notice, this list of conditions and the following disclaimer in the 135db2f26eSSascha Wildner * documentation and/or other materials provided with the distribution. 145db2f26eSSascha Wildner * 155db2f26eSSascha Wildner * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 165db2f26eSSascha Wildner * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 175db2f26eSSascha Wildner * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 185db2f26eSSascha Wildner * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 195db2f26eSSascha Wildner * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 205db2f26eSSascha Wildner * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 215db2f26eSSascha Wildner * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 225db2f26eSSascha Wildner * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 235db2f26eSSascha Wildner * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 245db2f26eSSascha Wildner * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 255db2f26eSSascha Wildner * SUCH DAMAGE. 265db2f26eSSascha Wildner * 275db2f26eSSascha Wildner * $FreeBSD: src/sys/dev/acpica/acpi_cpu.c,v 1.72 2008/04/12 12:06:00 rpaulo Exp $ 285db2f26eSSascha Wildner */ 295db2f26eSSascha Wildner 305db2f26eSSascha Wildner #include "opt_acpi.h" 315db2f26eSSascha Wildner #include <sys/param.h> 325db2f26eSSascha Wildner #include <sys/bus.h> 335db2f26eSSascha Wildner #include <sys/kernel.h> 345db2f26eSSascha Wildner #include <sys/malloc.h> 355db2f26eSSascha Wildner #include <sys/globaldata.h> 365db2f26eSSascha Wildner #include <sys/power.h> 375db2f26eSSascha Wildner #include <sys/proc.h> 385db2f26eSSascha Wildner #include <sys/sbuf.h> 395db2f26eSSascha Wildner #include <sys/thread2.h> 40*b45624acSSepherosa Ziehau #include <sys/serialize.h> 415db2f26eSSascha Wildner 425db2f26eSSascha Wildner #include <bus/pci/pcivar.h> 435db2f26eSSascha Wildner #include <machine/atomic.h> 445db2f26eSSascha Wildner #include <machine/globaldata.h> 455db2f26eSSascha Wildner #include <machine/md_var.h> 465db2f26eSSascha Wildner #include <machine/smp.h> 475db2f26eSSascha Wildner #include <sys/rman.h> 485db2f26eSSascha Wildner 495db2f26eSSascha Wildner #include "acpi.h" 505db2f26eSSascha Wildner #include "acpivar.h" 515db2f26eSSascha Wildner #include "acpi_cpu.h" 525db2f26eSSascha Wildner 535db2f26eSSascha Wildner /* 545db2f26eSSascha Wildner * Support for ACPI Processor devices, including C[1-3] sleep states. 555db2f26eSSascha Wildner */ 565db2f26eSSascha Wildner 575db2f26eSSascha Wildner /* Hooks for the ACPI CA debugging infrastructure */ 585db2f26eSSascha Wildner #define _COMPONENT ACPI_PROCESSOR 595db2f26eSSascha Wildner ACPI_MODULE_NAME("PROCESSOR") 605db2f26eSSascha Wildner 615db2f26eSSascha Wildner struct acpi_cx { 625db2f26eSSascha Wildner struct resource *p_lvlx; /* Register to read to enter state. */ 635db2f26eSSascha Wildner int rid; /* rid of p_lvlx */ 645db2f26eSSascha Wildner uint32_t type; /* C1-3 (C4 and up treated as C3). */ 655db2f26eSSascha Wildner uint32_t trans_lat; /* Transition latency (usec). */ 665db2f26eSSascha Wildner uint32_t power; /* Power consumed (mW). */ 675db2f26eSSascha Wildner int res_type; /* Resource type for p_lvlx. */ 685db2f26eSSascha Wildner }; 695db2f26eSSascha Wildner #define MAX_CX_STATES 8 705db2f26eSSascha Wildner 715db2f26eSSascha Wildner struct acpi_cpu_softc { 725db2f26eSSascha Wildner device_t cpu_dev; 735db2f26eSSascha Wildner struct acpi_cpux_softc *cpu_parent; 745db2f26eSSascha Wildner ACPI_HANDLE cpu_handle; 7572e8a1d8SSepherosa Ziehau int cpu_id; 765db2f26eSSascha Wildner uint32_t cpu_p_blk; /* ACPI P_BLK location */ 775db2f26eSSascha Wildner uint32_t cpu_p_blk_len; /* P_BLK length (must be 6). */ 785db2f26eSSascha Wildner struct acpi_cx cpu_cx_states[MAX_CX_STATES]; 795db2f26eSSascha Wildner int cpu_cx_count; /* Number of valid Cx states. */ 805db2f26eSSascha Wildner int cpu_prev_sleep;/* Last idle sleep duration. */ 815db2f26eSSascha Wildner /* Runtime state. */ 825db2f26eSSascha Wildner int cpu_non_c3; /* Index of lowest non-C3 state. */ 835db2f26eSSascha Wildner u_int cpu_cx_stats[MAX_CX_STATES];/* Cx usage history. */ 845db2f26eSSascha Wildner /* Values for sysctl. */ 851d730338SSepherosa Ziehau int cpu_cx_lowest; /* Current Cx lowest */ 861d730338SSepherosa Ziehau int cpu_cx_lowest_req; /* Requested Cx lowest */ 875db2f26eSSascha Wildner char cpu_cx_supported[64]; 885db2f26eSSascha Wildner }; 895db2f26eSSascha Wildner 905db2f26eSSascha Wildner struct acpi_cpu_device { 915db2f26eSSascha Wildner struct resource_list ad_rl; 925db2f26eSSascha Wildner }; 935db2f26eSSascha Wildner 945db2f26eSSascha Wildner #define CPU_GET_REG(reg, width) \ 955db2f26eSSascha Wildner (bus_space_read_ ## width(rman_get_bustag((reg)), \ 965db2f26eSSascha Wildner rman_get_bushandle((reg)), 0)) 975db2f26eSSascha Wildner #define CPU_SET_REG(reg, width, val) \ 985db2f26eSSascha Wildner (bus_space_write_ ## width(rman_get_bustag((reg)), \ 995db2f26eSSascha Wildner rman_get_bushandle((reg)), 0, (val))) 1005db2f26eSSascha Wildner 1015db2f26eSSascha Wildner #define PM_USEC(x) ((x) >> 2) /* ~4 clocks per usec (3.57955 Mhz) */ 1025db2f26eSSascha Wildner 1035db2f26eSSascha Wildner #define ACPI_NOTIFY_CX_STATES 0x81 /* _CST changed. */ 1045db2f26eSSascha Wildner 1055db2f26eSSascha Wildner #define CPU_QUIRK_NO_C3 (1<<0) /* C3-type states are not usable. */ 1065db2f26eSSascha Wildner #define CPU_QUIRK_NO_BM_CTRL (1<<2) /* No bus mastering control. */ 1075db2f26eSSascha Wildner 1085db2f26eSSascha Wildner #define PCI_VENDOR_INTEL 0x8086 1095db2f26eSSascha Wildner #define PCI_DEVICE_82371AB_3 0x7113 /* PIIX4 chipset for quirks. */ 1105db2f26eSSascha Wildner #define PCI_REVISION_A_STEP 0 1115db2f26eSSascha Wildner #define PCI_REVISION_B_STEP 1 1125db2f26eSSascha Wildner #define PCI_REVISION_4E 2 1135db2f26eSSascha Wildner #define PCI_REVISION_4M 3 1145db2f26eSSascha Wildner #define PIIX4_DEVACTB_REG 0x58 1155db2f26eSSascha Wildner #define PIIX4_BRLD_EN_IRQ0 (1<<0) 1165db2f26eSSascha Wildner #define PIIX4_BRLD_EN_IRQ (1<<1) 1175db2f26eSSascha Wildner #define PIIX4_BRLD_EN_IRQ8 (1<<5) 1185db2f26eSSascha Wildner #define PIIX4_STOP_BREAK_MASK (PIIX4_BRLD_EN_IRQ0 | PIIX4_BRLD_EN_IRQ | PIIX4_BRLD_EN_IRQ8) 1195db2f26eSSascha Wildner #define PIIX4_PCNTRL_BST_EN (1<<10) 1205db2f26eSSascha Wildner 1215db2f26eSSascha Wildner /* Platform hardware resource information. */ 1225db2f26eSSascha Wildner static uint32_t cpu_smi_cmd; /* Value to write to SMI_CMD. */ 1235db2f26eSSascha Wildner static uint8_t cpu_cst_cnt; /* Indicate we are _CST aware. */ 1245db2f26eSSascha Wildner static int cpu_quirks; /* Indicate any hardware bugs. */ 1255db2f26eSSascha Wildner 1265db2f26eSSascha Wildner /* Runtime state. */ 1275db2f26eSSascha Wildner static int cpu_disable_idle; /* Disable entry to idle function */ 1285db2f26eSSascha Wildner static int cpu_cx_count; /* Number of valid Cx states */ 1295db2f26eSSascha Wildner 1305db2f26eSSascha Wildner /* Values for sysctl. */ 1315db2f26eSSascha Wildner static int cpu_cx_generic; 1321d730338SSepherosa Ziehau static int cpu_cx_lowest; /* Current Cx lowest */ 1331d730338SSepherosa Ziehau static int cpu_cx_lowest_req; /* Requested Cx lowest */ 134*b45624acSSepherosa Ziehau static struct lwkt_serialize cpu_cx_slize = LWKT_SERIALIZE_INITIALIZER; 1355db2f26eSSascha Wildner 1365db2f26eSSascha Wildner /* C3 state transition */ 1375db2f26eSSascha Wildner static int cpu_c3_ncpus; 1385db2f26eSSascha Wildner 1395db2f26eSSascha Wildner static device_t *cpu_devices; 1405db2f26eSSascha Wildner static int cpu_ndevices; 1415db2f26eSSascha Wildner static struct acpi_cpu_softc **cpu_softc; 1425db2f26eSSascha Wildner 1435db2f26eSSascha Wildner static int acpi_cpu_cst_probe(device_t dev); 1445db2f26eSSascha Wildner static int acpi_cpu_cst_attach(device_t dev); 1455db2f26eSSascha Wildner static int acpi_cpu_cst_suspend(device_t dev); 1465db2f26eSSascha Wildner static int acpi_cpu_cst_resume(device_t dev); 1475db2f26eSSascha Wildner static struct resource_list *acpi_cpu_cst_get_rlist(device_t dev, 1485db2f26eSSascha Wildner device_t child); 1495db2f26eSSascha Wildner static device_t acpi_cpu_cst_add_child(device_t bus, device_t parent, 1505db2f26eSSascha Wildner int order, const char *name, int unit); 1515db2f26eSSascha Wildner static int acpi_cpu_cst_read_ivar(device_t dev, device_t child, 1525db2f26eSSascha Wildner int index, uintptr_t *result); 1535db2f26eSSascha Wildner static int acpi_cpu_cst_shutdown(device_t dev); 1545db2f26eSSascha Wildner static void acpi_cpu_cx_probe(struct acpi_cpu_softc *sc); 1555db2f26eSSascha Wildner static void acpi_cpu_generic_cx_probe(struct acpi_cpu_softc *sc); 1565db2f26eSSascha Wildner static int acpi_cpu_cx_cst(struct acpi_cpu_softc *sc); 1575db2f26eSSascha Wildner static void acpi_cpu_startup(void *arg); 1585db2f26eSSascha Wildner static void acpi_cpu_startup_cx(struct acpi_cpu_softc *sc); 1595db2f26eSSascha Wildner static void acpi_cpu_cx_list(struct acpi_cpu_softc *sc); 1605db2f26eSSascha Wildner static void acpi_cpu_idle(void); 1615db2f26eSSascha Wildner static void acpi_cpu_cst_notify(device_t); 1625db2f26eSSascha Wildner static int acpi_cpu_quirks(void); 1635db2f26eSSascha Wildner static int acpi_cpu_usage_sysctl(SYSCTL_HANDLER_ARGS); 1645db2f26eSSascha Wildner static int acpi_cpu_set_cx_lowest(struct acpi_cpu_softc *sc, int val); 1655db2f26eSSascha Wildner static int acpi_cpu_cx_lowest_sysctl(SYSCTL_HANDLER_ARGS); 1661d730338SSepherosa Ziehau static int acpi_cpu_cx_lowest_use_sysctl(SYSCTL_HANDLER_ARGS); 1675db2f26eSSascha Wildner static int acpi_cpu_global_cx_lowest_sysctl(SYSCTL_HANDLER_ARGS); 1681d730338SSepherosa Ziehau static int acpi_cpu_global_cx_lowest_use_sysctl(SYSCTL_HANDLER_ARGS); 1695db2f26eSSascha Wildner 1705db2f26eSSascha Wildner static void acpi_cpu_c1(void); /* XXX */ 1715db2f26eSSascha Wildner 1725db2f26eSSascha Wildner static device_method_t acpi_cpu_cst_methods[] = { 1735db2f26eSSascha Wildner /* Device interface */ 1745db2f26eSSascha Wildner DEVMETHOD(device_probe, acpi_cpu_cst_probe), 1755db2f26eSSascha Wildner DEVMETHOD(device_attach, acpi_cpu_cst_attach), 1765db2f26eSSascha Wildner DEVMETHOD(device_detach, bus_generic_detach), 1775db2f26eSSascha Wildner DEVMETHOD(device_shutdown, acpi_cpu_cst_shutdown), 1785db2f26eSSascha Wildner DEVMETHOD(device_suspend, acpi_cpu_cst_suspend), 1795db2f26eSSascha Wildner DEVMETHOD(device_resume, acpi_cpu_cst_resume), 1805db2f26eSSascha Wildner 1815db2f26eSSascha Wildner /* Bus interface */ 1825db2f26eSSascha Wildner DEVMETHOD(bus_add_child, acpi_cpu_cst_add_child), 1835db2f26eSSascha Wildner DEVMETHOD(bus_read_ivar, acpi_cpu_cst_read_ivar), 1845db2f26eSSascha Wildner DEVMETHOD(bus_get_resource_list, acpi_cpu_cst_get_rlist), 1855db2f26eSSascha Wildner DEVMETHOD(bus_get_resource, bus_generic_rl_get_resource), 1865db2f26eSSascha Wildner DEVMETHOD(bus_set_resource, bus_generic_rl_set_resource), 1875db2f26eSSascha Wildner DEVMETHOD(bus_alloc_resource, bus_generic_rl_alloc_resource), 1885db2f26eSSascha Wildner DEVMETHOD(bus_release_resource, bus_generic_rl_release_resource), 1895db2f26eSSascha Wildner DEVMETHOD(bus_driver_added, bus_generic_driver_added), 1905db2f26eSSascha Wildner DEVMETHOD(bus_activate_resource, bus_generic_activate_resource), 1915db2f26eSSascha Wildner DEVMETHOD(bus_deactivate_resource, bus_generic_deactivate_resource), 1925db2f26eSSascha Wildner DEVMETHOD(bus_setup_intr, bus_generic_setup_intr), 1935db2f26eSSascha Wildner DEVMETHOD(bus_teardown_intr, bus_generic_teardown_intr), 194d3c9c58eSSascha Wildner DEVMETHOD_END 1955db2f26eSSascha Wildner }; 1965db2f26eSSascha Wildner 1975db2f26eSSascha Wildner static driver_t acpi_cpu_cst_driver = { 1985db2f26eSSascha Wildner "cpu_cst", 1995db2f26eSSascha Wildner acpi_cpu_cst_methods, 2005db2f26eSSascha Wildner sizeof(struct acpi_cpu_softc), 2015db2f26eSSascha Wildner }; 2025db2f26eSSascha Wildner 2035db2f26eSSascha Wildner static devclass_t acpi_cpu_cst_devclass; 2045db2f26eSSascha Wildner DRIVER_MODULE(cpu_cst, cpu, acpi_cpu_cst_driver, acpi_cpu_cst_devclass, NULL, NULL); 2055db2f26eSSascha Wildner MODULE_DEPEND(cpu_cst, acpi, 1, 1, 1); 2065db2f26eSSascha Wildner 2075db2f26eSSascha Wildner static int 2085db2f26eSSascha Wildner acpi_cpu_cst_probe(device_t dev) 2095db2f26eSSascha Wildner { 2105db2f26eSSascha Wildner int cpu_id; 2115db2f26eSSascha Wildner 2125db2f26eSSascha Wildner if (acpi_disabled("cpu_cst") || acpi_get_type(dev) != ACPI_TYPE_PROCESSOR) 2135db2f26eSSascha Wildner return (ENXIO); 2145db2f26eSSascha Wildner 2155db2f26eSSascha Wildner cpu_id = acpi_get_magic(dev); 2165db2f26eSSascha Wildner 2175db2f26eSSascha Wildner if (cpu_softc == NULL) 2185db2f26eSSascha Wildner cpu_softc = kmalloc(sizeof(struct acpi_cpu_softc *) * 2195db2f26eSSascha Wildner SMP_MAXCPU, M_TEMP /* XXX */, M_INTWAIT | M_ZERO); 2205db2f26eSSascha Wildner 2215db2f26eSSascha Wildner /* 2225db2f26eSSascha Wildner * Check if we already probed this processor. We scan the bus twice 2235db2f26eSSascha Wildner * so it's possible we've already seen this one. 2245db2f26eSSascha Wildner */ 2255db2f26eSSascha Wildner if (cpu_softc[cpu_id] != NULL) { 2265db2f26eSSascha Wildner device_printf(dev, "CPU%d cstate already exist\n", cpu_id); 2275db2f26eSSascha Wildner return (ENXIO); 2285db2f26eSSascha Wildner } 2295db2f26eSSascha Wildner 2305db2f26eSSascha Wildner /* Mark this processor as in-use and save our derived id for attach. */ 2315db2f26eSSascha Wildner cpu_softc[cpu_id] = (void *)1; 2325db2f26eSSascha Wildner device_set_desc(dev, "ACPI CPU C-State"); 2335db2f26eSSascha Wildner 2345db2f26eSSascha Wildner return (0); 2355db2f26eSSascha Wildner } 2365db2f26eSSascha Wildner 2375db2f26eSSascha Wildner static int 2385db2f26eSSascha Wildner acpi_cpu_cst_attach(device_t dev) 2395db2f26eSSascha Wildner { 2405db2f26eSSascha Wildner ACPI_BUFFER buf; 2415db2f26eSSascha Wildner ACPI_OBJECT *obj; 2425db2f26eSSascha Wildner struct acpi_cpu_softc *sc; 2435db2f26eSSascha Wildner ACPI_STATUS status; 2445db2f26eSSascha Wildner 2455db2f26eSSascha Wildner ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__); 2465db2f26eSSascha Wildner 2475db2f26eSSascha Wildner sc = device_get_softc(dev); 2485db2f26eSSascha Wildner sc->cpu_dev = dev; 2495db2f26eSSascha Wildner sc->cpu_parent = device_get_softc(device_get_parent(dev)); 2505db2f26eSSascha Wildner sc->cpu_handle = acpi_get_handle(dev); 25172e8a1d8SSepherosa Ziehau sc->cpu_id = acpi_get_magic(dev); 25272e8a1d8SSepherosa Ziehau cpu_softc[sc->cpu_id] = sc; 2535db2f26eSSascha Wildner cpu_smi_cmd = AcpiGbl_FADT.SmiCommand; 2545db2f26eSSascha Wildner cpu_cst_cnt = AcpiGbl_FADT.CstControl; 2555db2f26eSSascha Wildner 2565db2f26eSSascha Wildner buf.Pointer = NULL; 2575db2f26eSSascha Wildner buf.Length = ACPI_ALLOCATE_BUFFER; 2585db2f26eSSascha Wildner status = AcpiEvaluateObject(sc->cpu_handle, NULL, NULL, &buf); 2595db2f26eSSascha Wildner if (ACPI_FAILURE(status)) { 2605db2f26eSSascha Wildner device_printf(dev, "attach failed to get Processor obj - %s\n", 2615db2f26eSSascha Wildner AcpiFormatException(status)); 2625db2f26eSSascha Wildner return (ENXIO); 2635db2f26eSSascha Wildner } 2645db2f26eSSascha Wildner obj = (ACPI_OBJECT *)buf.Pointer; 2655db2f26eSSascha Wildner sc->cpu_p_blk = obj->Processor.PblkAddress; 2665db2f26eSSascha Wildner sc->cpu_p_blk_len = obj->Processor.PblkLength; 2675db2f26eSSascha Wildner AcpiOsFree(obj); 2685db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, "acpi_cpu%d: P_BLK at %#x/%d\n", 2695db2f26eSSascha Wildner device_get_unit(dev), sc->cpu_p_blk, sc->cpu_p_blk_len)); 2705db2f26eSSascha Wildner 2715db2f26eSSascha Wildner /* 2725db2f26eSSascha Wildner * If this is the first cpu we attach, create and initialize the generic 2735db2f26eSSascha Wildner * resources that will be used by all acpi cpu devices. 2745db2f26eSSascha Wildner */ 2755db2f26eSSascha Wildner if (device_get_unit(dev) == 0) { 2765db2f26eSSascha Wildner /* Assume we won't be using generic Cx mode by default */ 2775db2f26eSSascha Wildner cpu_cx_generic = FALSE; 2785db2f26eSSascha Wildner 2795db2f26eSSascha Wildner /* Queue post cpu-probing task handler */ 2805db2f26eSSascha Wildner AcpiOsExecute(OSL_NOTIFY_HANDLER, acpi_cpu_startup, NULL); 2815db2f26eSSascha Wildner } 2825db2f26eSSascha Wildner 2835db2f26eSSascha Wildner /* Probe for Cx state support. */ 2845db2f26eSSascha Wildner acpi_cpu_cx_probe(sc); 2855db2f26eSSascha Wildner 2865db2f26eSSascha Wildner /* Finally, call identify and probe/attach for child devices. */ 2875db2f26eSSascha Wildner bus_generic_probe(dev); 2885db2f26eSSascha Wildner bus_generic_attach(dev); 2895db2f26eSSascha Wildner 2905db2f26eSSascha Wildner return (0); 2915db2f26eSSascha Wildner } 2925db2f26eSSascha Wildner 2935db2f26eSSascha Wildner /* 2945db2f26eSSascha Wildner * Disable any entry to the idle function during suspend and re-enable it 2955db2f26eSSascha Wildner * during resume. 2965db2f26eSSascha Wildner */ 2975db2f26eSSascha Wildner static int 2985db2f26eSSascha Wildner acpi_cpu_cst_suspend(device_t dev) 2995db2f26eSSascha Wildner { 3005db2f26eSSascha Wildner int error; 3015db2f26eSSascha Wildner 3025db2f26eSSascha Wildner error = bus_generic_suspend(dev); 3035db2f26eSSascha Wildner if (error) 3045db2f26eSSascha Wildner return (error); 3055db2f26eSSascha Wildner cpu_disable_idle = TRUE; 3065db2f26eSSascha Wildner return (0); 3075db2f26eSSascha Wildner } 3085db2f26eSSascha Wildner 3095db2f26eSSascha Wildner static int 3105db2f26eSSascha Wildner acpi_cpu_cst_resume(device_t dev) 3115db2f26eSSascha Wildner { 3125db2f26eSSascha Wildner 3135db2f26eSSascha Wildner cpu_disable_idle = FALSE; 3145db2f26eSSascha Wildner return (bus_generic_resume(dev)); 3155db2f26eSSascha Wildner } 3165db2f26eSSascha Wildner 3175db2f26eSSascha Wildner static struct resource_list * 3185db2f26eSSascha Wildner acpi_cpu_cst_get_rlist(device_t dev, device_t child) 3195db2f26eSSascha Wildner { 3205db2f26eSSascha Wildner struct acpi_cpu_device *ad; 3215db2f26eSSascha Wildner 3225db2f26eSSascha Wildner ad = device_get_ivars(child); 3235db2f26eSSascha Wildner if (ad == NULL) 3245db2f26eSSascha Wildner return (NULL); 3255db2f26eSSascha Wildner return (&ad->ad_rl); 3265db2f26eSSascha Wildner } 3275db2f26eSSascha Wildner 3285db2f26eSSascha Wildner static device_t 3295db2f26eSSascha Wildner acpi_cpu_cst_add_child(device_t bus, device_t parent, int order, 3305db2f26eSSascha Wildner const char *name, int unit) 3315db2f26eSSascha Wildner { 3325db2f26eSSascha Wildner struct acpi_cpu_device *ad; 3335db2f26eSSascha Wildner device_t child; 3345db2f26eSSascha Wildner 3355db2f26eSSascha Wildner if ((ad = kmalloc(sizeof(*ad), M_TEMP, M_NOWAIT | M_ZERO)) == NULL) 3365db2f26eSSascha Wildner return (NULL); 3375db2f26eSSascha Wildner 3385db2f26eSSascha Wildner resource_list_init(&ad->ad_rl); 3395db2f26eSSascha Wildner 3405db2f26eSSascha Wildner child = device_add_child_ordered(parent, order, name, unit); 3415db2f26eSSascha Wildner if (child != NULL) 3425db2f26eSSascha Wildner device_set_ivars(child, ad); 3435db2f26eSSascha Wildner else 3445db2f26eSSascha Wildner kfree(ad, M_TEMP); 3455db2f26eSSascha Wildner return (child); 3465db2f26eSSascha Wildner } 3475db2f26eSSascha Wildner 3485db2f26eSSascha Wildner static int 3495db2f26eSSascha Wildner acpi_cpu_cst_read_ivar(device_t dev, device_t child, int index, 3505db2f26eSSascha Wildner uintptr_t *result) 3515db2f26eSSascha Wildner { 3525db2f26eSSascha Wildner struct acpi_cpu_softc *sc; 3535db2f26eSSascha Wildner 3545db2f26eSSascha Wildner sc = device_get_softc(dev); 3555db2f26eSSascha Wildner switch (index) { 3565db2f26eSSascha Wildner case ACPI_IVAR_HANDLE: 3575db2f26eSSascha Wildner *result = (uintptr_t)sc->cpu_handle; 3585db2f26eSSascha Wildner break; 3595db2f26eSSascha Wildner #if 0 3605db2f26eSSascha Wildner case CPU_IVAR_PCPU: 3615db2f26eSSascha Wildner *result = (uintptr_t)sc->cpu_pcpu; 3625db2f26eSSascha Wildner break; 3635db2f26eSSascha Wildner #endif 3645db2f26eSSascha Wildner default: 3655db2f26eSSascha Wildner return (ENOENT); 3665db2f26eSSascha Wildner } 3675db2f26eSSascha Wildner return (0); 3685db2f26eSSascha Wildner } 3695db2f26eSSascha Wildner 3705db2f26eSSascha Wildner static int 3715db2f26eSSascha Wildner acpi_cpu_cst_shutdown(device_t dev) 3725db2f26eSSascha Wildner { 3735db2f26eSSascha Wildner ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__); 3745db2f26eSSascha Wildner 3755db2f26eSSascha Wildner /* Allow children to shutdown first. */ 3765db2f26eSSascha Wildner bus_generic_shutdown(dev); 3775db2f26eSSascha Wildner 3785db2f26eSSascha Wildner /* 3795db2f26eSSascha Wildner * Disable any entry to the idle function. There is a small race where 3805db2f26eSSascha Wildner * an idle thread have passed this check but not gone to sleep. This 3815db2f26eSSascha Wildner * is ok since device_shutdown() does not free the softc, otherwise 3825db2f26eSSascha Wildner * we'd have to be sure all threads were evicted before returning. 3835db2f26eSSascha Wildner */ 3845db2f26eSSascha Wildner cpu_disable_idle = TRUE; 3855db2f26eSSascha Wildner 3865db2f26eSSascha Wildner return_VALUE (0); 3875db2f26eSSascha Wildner } 3885db2f26eSSascha Wildner 3895db2f26eSSascha Wildner static void 3905db2f26eSSascha Wildner acpi_cpu_cx_probe(struct acpi_cpu_softc *sc) 3915db2f26eSSascha Wildner { 3925db2f26eSSascha Wildner ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__); 3935db2f26eSSascha Wildner 3945db2f26eSSascha Wildner /* Use initial sleep value of 1 sec. to start with lowest idle state. */ 3955db2f26eSSascha Wildner sc->cpu_prev_sleep = 1000000; 3965db2f26eSSascha Wildner sc->cpu_cx_lowest = 0; 3971d730338SSepherosa Ziehau sc->cpu_cx_lowest_req = 0; 3985db2f26eSSascha Wildner 3995db2f26eSSascha Wildner /* 4005db2f26eSSascha Wildner * Check for the ACPI 2.0 _CST sleep states object. If we can't find 4015db2f26eSSascha Wildner * any, we'll revert to generic FADT/P_BLK Cx control method which will 4025db2f26eSSascha Wildner * be handled by acpi_cpu_startup. We need to defer to after having 4035db2f26eSSascha Wildner * probed all the cpus in the system before probing for generic Cx 4045db2f26eSSascha Wildner * states as we may already have found cpus with valid _CST packages 4055db2f26eSSascha Wildner */ 4065db2f26eSSascha Wildner if (!cpu_cx_generic && acpi_cpu_cx_cst(sc) != 0) { 4075db2f26eSSascha Wildner /* 4085db2f26eSSascha Wildner * We were unable to find a _CST package for this cpu or there 4095db2f26eSSascha Wildner * was an error parsing it. Switch back to generic mode. 4105db2f26eSSascha Wildner */ 4115db2f26eSSascha Wildner cpu_cx_generic = TRUE; 4125db2f26eSSascha Wildner if (bootverbose) 4135db2f26eSSascha Wildner device_printf(sc->cpu_dev, "switching to generic Cx mode\n"); 4145db2f26eSSascha Wildner } 4155db2f26eSSascha Wildner 4165db2f26eSSascha Wildner /* 4175db2f26eSSascha Wildner * TODO: _CSD Package should be checked here. 4185db2f26eSSascha Wildner */ 4195db2f26eSSascha Wildner } 4205db2f26eSSascha Wildner 4215db2f26eSSascha Wildner static void 4225db2f26eSSascha Wildner acpi_cpu_generic_cx_probe(struct acpi_cpu_softc *sc) 4235db2f26eSSascha Wildner { 4245db2f26eSSascha Wildner ACPI_GENERIC_ADDRESS gas; 4255db2f26eSSascha Wildner struct acpi_cx *cx_ptr; 4265db2f26eSSascha Wildner 4275db2f26eSSascha Wildner sc->cpu_cx_count = 0; 4285db2f26eSSascha Wildner cx_ptr = sc->cpu_cx_states; 4295db2f26eSSascha Wildner 4305db2f26eSSascha Wildner /* Use initial sleep value of 1 sec. to start with lowest idle state. */ 4315db2f26eSSascha Wildner sc->cpu_prev_sleep = 1000000; 4325db2f26eSSascha Wildner 4335db2f26eSSascha Wildner /* C1 has been required since just after ACPI 1.0 */ 4345db2f26eSSascha Wildner cx_ptr->type = ACPI_STATE_C1; 4355db2f26eSSascha Wildner cx_ptr->trans_lat = 0; 4365db2f26eSSascha Wildner cx_ptr++; 4375db2f26eSSascha Wildner sc->cpu_cx_count++; 4385db2f26eSSascha Wildner 4395db2f26eSSascha Wildner /* 4405db2f26eSSascha Wildner * The spec says P_BLK must be 6 bytes long. However, some systems 4415db2f26eSSascha Wildner * use it to indicate a fractional set of features present so we 4425db2f26eSSascha Wildner * take 5 as C2. Some may also have a value of 7 to indicate 4435db2f26eSSascha Wildner * another C3 but most use _CST for this (as required) and having 4445db2f26eSSascha Wildner * "only" C1-C3 is not a hardship. 4455db2f26eSSascha Wildner */ 4465db2f26eSSascha Wildner if (sc->cpu_p_blk_len < 5) 4475db2f26eSSascha Wildner return; 4485db2f26eSSascha Wildner 4495db2f26eSSascha Wildner /* Validate and allocate resources for C2 (P_LVL2). */ 4505db2f26eSSascha Wildner gas.SpaceId = ACPI_ADR_SPACE_SYSTEM_IO; 4515db2f26eSSascha Wildner gas.BitWidth = 8; 4525db2f26eSSascha Wildner if (AcpiGbl_FADT.C2Latency <= 100) { 4535db2f26eSSascha Wildner gas.Address = sc->cpu_p_blk + 4; 4545db2f26eSSascha Wildner 4555db2f26eSSascha Wildner cx_ptr->rid = sc->cpu_parent->cpux_next_rid; 4565db2f26eSSascha Wildner acpi_bus_alloc_gas(sc->cpu_dev, &cx_ptr->type, &cx_ptr->rid, &gas, &cx_ptr->p_lvlx, 4575db2f26eSSascha Wildner RF_SHAREABLE); 4585db2f26eSSascha Wildner if (cx_ptr->p_lvlx != NULL) { 4595db2f26eSSascha Wildner sc->cpu_parent->cpux_next_rid++; 4605db2f26eSSascha Wildner cx_ptr->type = ACPI_STATE_C2; 4615db2f26eSSascha Wildner cx_ptr->trans_lat = AcpiGbl_FADT.C2Latency; 4625db2f26eSSascha Wildner cx_ptr++; 4635db2f26eSSascha Wildner sc->cpu_cx_count++; 4645db2f26eSSascha Wildner } 4655db2f26eSSascha Wildner } 4665db2f26eSSascha Wildner if (sc->cpu_p_blk_len < 6) 4675db2f26eSSascha Wildner return; 4685db2f26eSSascha Wildner 4695db2f26eSSascha Wildner /* Validate and allocate resources for C3 (P_LVL3). */ 4705db2f26eSSascha Wildner if (AcpiGbl_FADT.C3Latency <= 1000 && !(cpu_quirks & CPU_QUIRK_NO_C3)) { 4715db2f26eSSascha Wildner gas.Address = sc->cpu_p_blk + 5; 4725db2f26eSSascha Wildner 4735db2f26eSSascha Wildner cx_ptr->rid = sc->cpu_parent->cpux_next_rid; 4745db2f26eSSascha Wildner acpi_bus_alloc_gas(sc->cpu_dev, &cx_ptr->type, &cx_ptr->rid, &gas, 4755db2f26eSSascha Wildner &cx_ptr->p_lvlx, RF_SHAREABLE); 4765db2f26eSSascha Wildner if (cx_ptr->p_lvlx != NULL) { 4775db2f26eSSascha Wildner sc->cpu_parent->cpux_next_rid++; 4785db2f26eSSascha Wildner cx_ptr->type = ACPI_STATE_C3; 4795db2f26eSSascha Wildner cx_ptr->trans_lat = AcpiGbl_FADT.C3Latency; 4805db2f26eSSascha Wildner cx_ptr++; 4815db2f26eSSascha Wildner sc->cpu_cx_count++; 4825db2f26eSSascha Wildner } 4835db2f26eSSascha Wildner } 4845db2f26eSSascha Wildner } 4855db2f26eSSascha Wildner 4865db2f26eSSascha Wildner /* 4875db2f26eSSascha Wildner * Parse a _CST package and set up its Cx states. Since the _CST object 4885db2f26eSSascha Wildner * can change dynamically, our notify handler may call this function 4895db2f26eSSascha Wildner * to clean up and probe the new _CST package. 4905db2f26eSSascha Wildner */ 4915db2f26eSSascha Wildner static int 4925db2f26eSSascha Wildner acpi_cpu_cx_cst(struct acpi_cpu_softc *sc) 4935db2f26eSSascha Wildner { 4945db2f26eSSascha Wildner struct acpi_cx *cx_ptr; 4955db2f26eSSascha Wildner ACPI_STATUS status; 4965db2f26eSSascha Wildner ACPI_BUFFER buf; 4975db2f26eSSascha Wildner ACPI_OBJECT *top; 4985db2f26eSSascha Wildner ACPI_OBJECT *pkg; 4995db2f26eSSascha Wildner uint32_t count; 5005db2f26eSSascha Wildner int i; 5015db2f26eSSascha Wildner 5025db2f26eSSascha Wildner ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__); 5035db2f26eSSascha Wildner 5045db2f26eSSascha Wildner buf.Pointer = NULL; 5055db2f26eSSascha Wildner buf.Length = ACPI_ALLOCATE_BUFFER; 5065db2f26eSSascha Wildner status = AcpiEvaluateObject(sc->cpu_handle, "_CST", NULL, &buf); 5075db2f26eSSascha Wildner if (ACPI_FAILURE(status)) 5085db2f26eSSascha Wildner return (ENXIO); 5095db2f26eSSascha Wildner 5105db2f26eSSascha Wildner /* _CST is a package with a count and at least one Cx package. */ 5115db2f26eSSascha Wildner top = (ACPI_OBJECT *)buf.Pointer; 5125db2f26eSSascha Wildner if (!ACPI_PKG_VALID(top, 2) || acpi_PkgInt32(top, 0, &count) != 0) { 5135db2f26eSSascha Wildner device_printf(sc->cpu_dev, "invalid _CST package\n"); 5145db2f26eSSascha Wildner AcpiOsFree(buf.Pointer); 5155db2f26eSSascha Wildner return (ENXIO); 5165db2f26eSSascha Wildner } 5175db2f26eSSascha Wildner if (count != top->Package.Count - 1) { 5185db2f26eSSascha Wildner device_printf(sc->cpu_dev, "invalid _CST state count (%d != %d)\n", 5195db2f26eSSascha Wildner count, top->Package.Count - 1); 5205db2f26eSSascha Wildner count = top->Package.Count - 1; 5215db2f26eSSascha Wildner } 5225db2f26eSSascha Wildner if (count > MAX_CX_STATES) { 5235db2f26eSSascha Wildner device_printf(sc->cpu_dev, "_CST has too many states (%d)\n", count); 5245db2f26eSSascha Wildner count = MAX_CX_STATES; 5255db2f26eSSascha Wildner } 5265db2f26eSSascha Wildner 5275db2f26eSSascha Wildner /* Set up all valid states. */ 5285db2f26eSSascha Wildner sc->cpu_cx_count = 0; 5295db2f26eSSascha Wildner cx_ptr = sc->cpu_cx_states; 5305db2f26eSSascha Wildner for (i = 0; i < count; i++) { 5315db2f26eSSascha Wildner pkg = &top->Package.Elements[i + 1]; 5325db2f26eSSascha Wildner if (!ACPI_PKG_VALID(pkg, 4) || 5335db2f26eSSascha Wildner acpi_PkgInt32(pkg, 1, &cx_ptr->type) != 0 || 5345db2f26eSSascha Wildner acpi_PkgInt32(pkg, 2, &cx_ptr->trans_lat) != 0 || 5355db2f26eSSascha Wildner acpi_PkgInt32(pkg, 3, &cx_ptr->power) != 0) { 5365db2f26eSSascha Wildner 5375db2f26eSSascha Wildner device_printf(sc->cpu_dev, "skipping invalid Cx state package\n"); 5385db2f26eSSascha Wildner continue; 5395db2f26eSSascha Wildner } 5405db2f26eSSascha Wildner 5415db2f26eSSascha Wildner /* Validate the state to see if we should use it. */ 5425db2f26eSSascha Wildner switch (cx_ptr->type) { 5435db2f26eSSascha Wildner case ACPI_STATE_C1: 5445db2f26eSSascha Wildner sc->cpu_non_c3 = i; 5455db2f26eSSascha Wildner cx_ptr++; 5465db2f26eSSascha Wildner sc->cpu_cx_count++; 5475db2f26eSSascha Wildner continue; 5485db2f26eSSascha Wildner case ACPI_STATE_C2: 5495db2f26eSSascha Wildner sc->cpu_non_c3 = i; 5505db2f26eSSascha Wildner break; 5515db2f26eSSascha Wildner case ACPI_STATE_C3: 5525db2f26eSSascha Wildner default: 5535db2f26eSSascha Wildner if ((cpu_quirks & CPU_QUIRK_NO_C3) != 0) { 5545db2f26eSSascha Wildner 5555db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 5565db2f26eSSascha Wildner "acpi_cpu%d: C3[%d] not available.\n", 5575db2f26eSSascha Wildner device_get_unit(sc->cpu_dev), i)); 5585db2f26eSSascha Wildner continue; 5595db2f26eSSascha Wildner } 5605db2f26eSSascha Wildner break; 5615db2f26eSSascha Wildner } 5625db2f26eSSascha Wildner 5635db2f26eSSascha Wildner #ifdef notyet 5645db2f26eSSascha Wildner /* Free up any previous register. */ 5655db2f26eSSascha Wildner if (cx_ptr->p_lvlx != NULL) { 5665db2f26eSSascha Wildner bus_release_resource(sc->cpu_dev, 0, 0, cx_ptr->p_lvlx); 5675db2f26eSSascha Wildner cx_ptr->p_lvlx = NULL; 5685db2f26eSSascha Wildner } 5695db2f26eSSascha Wildner #endif 5705db2f26eSSascha Wildner 5715db2f26eSSascha Wildner /* Allocate the control register for C2 or C3. */ 5725db2f26eSSascha Wildner cx_ptr->rid = sc->cpu_parent->cpux_next_rid; 5735db2f26eSSascha Wildner acpi_PkgGas(sc->cpu_dev, pkg, 0, &cx_ptr->res_type, &cx_ptr->rid, &cx_ptr->p_lvlx, 5745db2f26eSSascha Wildner RF_SHAREABLE); 5755db2f26eSSascha Wildner if (cx_ptr->p_lvlx) { 5765db2f26eSSascha Wildner sc->cpu_parent->cpux_next_rid++; 5775db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 5785db2f26eSSascha Wildner "acpi_cpu%d: Got C%d - %d latency\n", 5795db2f26eSSascha Wildner device_get_unit(sc->cpu_dev), cx_ptr->type, 5805db2f26eSSascha Wildner cx_ptr->trans_lat)); 5815db2f26eSSascha Wildner cx_ptr++; 5825db2f26eSSascha Wildner sc->cpu_cx_count++; 5835db2f26eSSascha Wildner } 5845db2f26eSSascha Wildner } 5855db2f26eSSascha Wildner AcpiOsFree(buf.Pointer); 5865db2f26eSSascha Wildner 5871d730338SSepherosa Ziehau /* 5881d730338SSepherosa Ziehau * Fix up the lowest Cx being used 5891d730338SSepherosa Ziehau */ 5901d730338SSepherosa Ziehau if (sc->cpu_cx_lowest_req < sc->cpu_cx_count) 5911d730338SSepherosa Ziehau sc->cpu_cx_lowest = sc->cpu_cx_lowest_req; 5921d730338SSepherosa Ziehau if (sc->cpu_cx_lowest > sc->cpu_cx_count - 1) 5931d730338SSepherosa Ziehau sc->cpu_cx_lowest = sc->cpu_cx_count - 1; 5941d730338SSepherosa Ziehau 5955db2f26eSSascha Wildner return (0); 5965db2f26eSSascha Wildner } 5975db2f26eSSascha Wildner 5985db2f26eSSascha Wildner /* 5995db2f26eSSascha Wildner * Call this *after* all CPUs have been attached. 6005db2f26eSSascha Wildner */ 6015db2f26eSSascha Wildner static void 6025db2f26eSSascha Wildner acpi_cpu_startup(void *arg) 6035db2f26eSSascha Wildner { 6045db2f26eSSascha Wildner struct acpi_cpu_softc *sc; 6055db2f26eSSascha Wildner int i; 6065db2f26eSSascha Wildner 6075db2f26eSSascha Wildner /* Get set of CPU devices */ 6085db2f26eSSascha Wildner devclass_get_devices(acpi_cpu_cst_devclass, &cpu_devices, &cpu_ndevices); 6095db2f26eSSascha Wildner 6105db2f26eSSascha Wildner /* 6115db2f26eSSascha Wildner * Setup any quirks that might necessary now that we have probed 6125db2f26eSSascha Wildner * all the CPUs 6135db2f26eSSascha Wildner */ 6145db2f26eSSascha Wildner acpi_cpu_quirks(); 6155db2f26eSSascha Wildner 6165db2f26eSSascha Wildner cpu_cx_count = 0; 6175db2f26eSSascha Wildner if (cpu_cx_generic) { 6185db2f26eSSascha Wildner /* 6195db2f26eSSascha Wildner * We are using generic Cx mode, probe for available Cx states 6205db2f26eSSascha Wildner * for all processors. 6215db2f26eSSascha Wildner */ 6225db2f26eSSascha Wildner for (i = 0; i < cpu_ndevices; i++) { 6235db2f26eSSascha Wildner sc = device_get_softc(cpu_devices[i]); 6245db2f26eSSascha Wildner acpi_cpu_generic_cx_probe(sc); 6255db2f26eSSascha Wildner if (sc->cpu_cx_count > cpu_cx_count) 6265db2f26eSSascha Wildner cpu_cx_count = sc->cpu_cx_count; 6275db2f26eSSascha Wildner } 6285db2f26eSSascha Wildner 6295db2f26eSSascha Wildner /* 6305db2f26eSSascha Wildner * Find the highest Cx state common to all CPUs 6315db2f26eSSascha Wildner * in the system, taking quirks into account. 6325db2f26eSSascha Wildner */ 6335db2f26eSSascha Wildner for (i = 0; i < cpu_ndevices; i++) { 6345db2f26eSSascha Wildner sc = device_get_softc(cpu_devices[i]); 6355db2f26eSSascha Wildner if (sc->cpu_cx_count < cpu_cx_count) 6365db2f26eSSascha Wildner cpu_cx_count = sc->cpu_cx_count; 6375db2f26eSSascha Wildner } 6385db2f26eSSascha Wildner } else { 6395db2f26eSSascha Wildner /* 6405db2f26eSSascha Wildner * We are using _CST mode, remove C3 state if necessary. 6415db2f26eSSascha Wildner * Update the largest Cx state supported in the global cpu_cx_count. 6425db2f26eSSascha Wildner * It will be used in the global Cx sysctl handler. 6435db2f26eSSascha Wildner * As we now know for sure that we will be using _CST mode 6445db2f26eSSascha Wildner * install our notify handler. 6455db2f26eSSascha Wildner */ 6465db2f26eSSascha Wildner for (i = 0; i < cpu_ndevices; i++) { 6475db2f26eSSascha Wildner sc = device_get_softc(cpu_devices[i]); 6485db2f26eSSascha Wildner if (cpu_quirks & CPU_QUIRK_NO_C3) { 6495db2f26eSSascha Wildner sc->cpu_cx_count = sc->cpu_non_c3 + 1; 6505db2f26eSSascha Wildner } 6515db2f26eSSascha Wildner if (sc->cpu_cx_count > cpu_cx_count) 6525db2f26eSSascha Wildner cpu_cx_count = sc->cpu_cx_count; 6535db2f26eSSascha Wildner sc->cpu_parent->cpux_cst_notify = acpi_cpu_cst_notify; 6545db2f26eSSascha Wildner } 6555db2f26eSSascha Wildner } 6565db2f26eSSascha Wildner 6575db2f26eSSascha Wildner /* Perform Cx final initialization. */ 6585db2f26eSSascha Wildner for (i = 0; i < cpu_ndevices; i++) { 6595db2f26eSSascha Wildner sc = device_get_softc(cpu_devices[i]); 6605db2f26eSSascha Wildner acpi_cpu_startup_cx(sc); 6615db2f26eSSascha Wildner 6625db2f26eSSascha Wildner if (sc->cpu_parent->glob_sysctl_tree != NULL) { 6635db2f26eSSascha Wildner struct acpi_cpux_softc *cpux = sc->cpu_parent; 6645db2f26eSSascha Wildner 6655db2f26eSSascha Wildner /* Add a sysctl handler to handle global Cx lowest setting */ 6665db2f26eSSascha Wildner SYSCTL_ADD_PROC(&cpux->glob_sysctl_ctx, 6675db2f26eSSascha Wildner SYSCTL_CHILDREN(cpux->glob_sysctl_tree), 6685db2f26eSSascha Wildner OID_AUTO, "cx_lowest", 6695db2f26eSSascha Wildner CTLTYPE_STRING | CTLFLAG_RW, NULL, 0, 6705db2f26eSSascha Wildner acpi_cpu_global_cx_lowest_sysctl, "A", 6711d730338SSepherosa Ziehau "Requested global lowest Cx sleep state"); 6721d730338SSepherosa Ziehau SYSCTL_ADD_PROC(&cpux->glob_sysctl_ctx, 6731d730338SSepherosa Ziehau SYSCTL_CHILDREN(cpux->glob_sysctl_tree), 6741d730338SSepherosa Ziehau OID_AUTO, "cx_lowest_use", 6751d730338SSepherosa Ziehau CTLTYPE_STRING | CTLFLAG_RD, NULL, 0, 6761d730338SSepherosa Ziehau acpi_cpu_global_cx_lowest_use_sysctl, "A", 6775db2f26eSSascha Wildner "Global lowest Cx sleep state to use"); 6785db2f26eSSascha Wildner } 6795db2f26eSSascha Wildner } 6805db2f26eSSascha Wildner 6815db2f26eSSascha Wildner /* Take over idling from cpu_idle_default(). */ 6825db2f26eSSascha Wildner cpu_cx_lowest = 0; 6831d730338SSepherosa Ziehau cpu_cx_lowest_req = 0; 6845db2f26eSSascha Wildner cpu_disable_idle = FALSE; 6855db2f26eSSascha Wildner cpu_idle_hook = acpi_cpu_idle; 6865db2f26eSSascha Wildner } 6875db2f26eSSascha Wildner 6885db2f26eSSascha Wildner static void 6895db2f26eSSascha Wildner acpi_cpu_cx_list(struct acpi_cpu_softc *sc) 6905db2f26eSSascha Wildner { 6915db2f26eSSascha Wildner struct sbuf sb; 6925db2f26eSSascha Wildner int i; 6935db2f26eSSascha Wildner 6945db2f26eSSascha Wildner /* 6955db2f26eSSascha Wildner * Set up the list of Cx states 6965db2f26eSSascha Wildner */ 6975db2f26eSSascha Wildner sc->cpu_non_c3 = 0; 6985db2f26eSSascha Wildner sbuf_new(&sb, sc->cpu_cx_supported, sizeof(sc->cpu_cx_supported), 6995db2f26eSSascha Wildner SBUF_FIXEDLEN); 7005db2f26eSSascha Wildner for (i = 0; i < sc->cpu_cx_count; i++) { 7015db2f26eSSascha Wildner sbuf_printf(&sb, "C%d/%d ", i + 1, sc->cpu_cx_states[i].trans_lat); 7025db2f26eSSascha Wildner if (sc->cpu_cx_states[i].type < ACPI_STATE_C3) 7035db2f26eSSascha Wildner sc->cpu_non_c3 = i; 7045db2f26eSSascha Wildner } 7055db2f26eSSascha Wildner sbuf_trim(&sb); 7065db2f26eSSascha Wildner sbuf_finish(&sb); 7075db2f26eSSascha Wildner } 7085db2f26eSSascha Wildner 7095db2f26eSSascha Wildner static void 7105db2f26eSSascha Wildner acpi_cpu_startup_cx(struct acpi_cpu_softc *sc) 7115db2f26eSSascha Wildner { 7125db2f26eSSascha Wildner struct acpi_cpux_softc *cpux = sc->cpu_parent; 7135db2f26eSSascha Wildner 7145db2f26eSSascha Wildner acpi_cpu_cx_list(sc); 7155db2f26eSSascha Wildner 7165db2f26eSSascha Wildner SYSCTL_ADD_STRING(&cpux->pcpu_sysctl_ctx, 7175db2f26eSSascha Wildner SYSCTL_CHILDREN(cpux->pcpu_sysctl_tree), 7185db2f26eSSascha Wildner OID_AUTO, "cx_supported", CTLFLAG_RD, 7195db2f26eSSascha Wildner sc->cpu_cx_supported, 0, 7205db2f26eSSascha Wildner "Cx/microsecond values for supported Cx states"); 7215db2f26eSSascha Wildner SYSCTL_ADD_PROC(&cpux->pcpu_sysctl_ctx, 7225db2f26eSSascha Wildner SYSCTL_CHILDREN(cpux->pcpu_sysctl_tree), 7235db2f26eSSascha Wildner OID_AUTO, "cx_lowest", CTLTYPE_STRING | CTLFLAG_RW, 7245db2f26eSSascha Wildner (void *)sc, 0, acpi_cpu_cx_lowest_sysctl, "A", 7251d730338SSepherosa Ziehau "requested lowest Cx sleep state"); 7261d730338SSepherosa Ziehau SYSCTL_ADD_PROC(&cpux->pcpu_sysctl_ctx, 7271d730338SSepherosa Ziehau SYSCTL_CHILDREN(cpux->pcpu_sysctl_tree), 7281d730338SSepherosa Ziehau OID_AUTO, "cx_lowest_use", CTLTYPE_STRING | CTLFLAG_RD, 7291d730338SSepherosa Ziehau (void *)sc, 0, acpi_cpu_cx_lowest_use_sysctl, "A", 7305db2f26eSSascha Wildner "lowest Cx sleep state to use"); 7315db2f26eSSascha Wildner SYSCTL_ADD_PROC(&cpux->pcpu_sysctl_ctx, 7325db2f26eSSascha Wildner SYSCTL_CHILDREN(cpux->pcpu_sysctl_tree), 7335db2f26eSSascha Wildner OID_AUTO, "cx_usage", CTLTYPE_STRING | CTLFLAG_RD, 7345db2f26eSSascha Wildner (void *)sc, 0, acpi_cpu_usage_sysctl, "A", 7355db2f26eSSascha Wildner "percent usage for each Cx state"); 7365db2f26eSSascha Wildner 7375db2f26eSSascha Wildner #ifdef notyet 7385db2f26eSSascha Wildner /* Signal platform that we can handle _CST notification. */ 7395db2f26eSSascha Wildner if (!cpu_cx_generic && cpu_cst_cnt != 0) { 7405db2f26eSSascha Wildner ACPI_LOCK(acpi); 7415db2f26eSSascha Wildner AcpiOsWritePort(cpu_smi_cmd, cpu_cst_cnt, 8); 7425db2f26eSSascha Wildner ACPI_UNLOCK(acpi); 7435db2f26eSSascha Wildner } 7445db2f26eSSascha Wildner #endif 7455db2f26eSSascha Wildner } 7465db2f26eSSascha Wildner 7475db2f26eSSascha Wildner /* 7485db2f26eSSascha Wildner * Idle the CPU in the lowest state possible. This function is called with 7495db2f26eSSascha Wildner * interrupts disabled. Note that once it re-enables interrupts, a task 7505db2f26eSSascha Wildner * switch can occur so do not access shared data (i.e. the softc) after 7515db2f26eSSascha Wildner * interrupts are re-enabled. 7525db2f26eSSascha Wildner */ 7535db2f26eSSascha Wildner static void 7545db2f26eSSascha Wildner acpi_cpu_idle(void) 7555db2f26eSSascha Wildner { 7565db2f26eSSascha Wildner struct acpi_cpu_softc *sc; 7575db2f26eSSascha Wildner struct acpi_cx *cx_next; 7585db2f26eSSascha Wildner uint64_t start_time, end_time; 7595db2f26eSSascha Wildner int bm_active, cx_next_idx, i; 7605db2f26eSSascha Wildner 7615db2f26eSSascha Wildner /* If disabled, return immediately. */ 7625db2f26eSSascha Wildner if (cpu_disable_idle) { 7635db2f26eSSascha Wildner ACPI_ENABLE_IRQS(); 7645db2f26eSSascha Wildner return; 7655db2f26eSSascha Wildner } 7665db2f26eSSascha Wildner 7675db2f26eSSascha Wildner /* 7685db2f26eSSascha Wildner * Look up our CPU id to get our softc. If it's NULL, we'll use C1 7695db2f26eSSascha Wildner * since there is no ACPI processor object for this CPU. This occurs 7705db2f26eSSascha Wildner * for logical CPUs in the HTT case. 7715db2f26eSSascha Wildner */ 7725db2f26eSSascha Wildner sc = cpu_softc[mdcpu->mi.gd_cpuid]; 7735db2f26eSSascha Wildner if (sc == NULL) { 7745db2f26eSSascha Wildner acpi_cpu_c1(); 7755db2f26eSSascha Wildner return; 7765db2f26eSSascha Wildner } 7775db2f26eSSascha Wildner 7785db2f26eSSascha Wildner /* Find the lowest state that has small enough latency. */ 7795db2f26eSSascha Wildner cx_next_idx = 0; 7805db2f26eSSascha Wildner for (i = sc->cpu_cx_lowest; i >= 0; i--) { 7815db2f26eSSascha Wildner if (sc->cpu_cx_states[i].trans_lat * 3 <= sc->cpu_prev_sleep) { 7825db2f26eSSascha Wildner cx_next_idx = i; 7835db2f26eSSascha Wildner break; 7845db2f26eSSascha Wildner } 7855db2f26eSSascha Wildner } 7865db2f26eSSascha Wildner 7875db2f26eSSascha Wildner /* 7885db2f26eSSascha Wildner * Check for bus master activity. If there was activity, clear 7895db2f26eSSascha Wildner * the bit and use the lowest non-C3 state. Note that the USB 7905db2f26eSSascha Wildner * driver polling for new devices keeps this bit set all the 7915db2f26eSSascha Wildner * time if USB is loaded. 7925db2f26eSSascha Wildner */ 7935db2f26eSSascha Wildner if ((cpu_quirks & CPU_QUIRK_NO_BM_CTRL) == 0) { 7945db2f26eSSascha Wildner AcpiReadBitRegister(ACPI_BITREG_BUS_MASTER_STATUS, &bm_active); 7955db2f26eSSascha Wildner if (bm_active != 0) { 7965db2f26eSSascha Wildner AcpiWriteBitRegister(ACPI_BITREG_BUS_MASTER_STATUS, 1); 7975db2f26eSSascha Wildner cx_next_idx = min(cx_next_idx, sc->cpu_non_c3); 7985db2f26eSSascha Wildner } 7995db2f26eSSascha Wildner } 8005db2f26eSSascha Wildner 8015db2f26eSSascha Wildner /* Select the next state and update statistics. */ 8025db2f26eSSascha Wildner cx_next = &sc->cpu_cx_states[cx_next_idx]; 8035db2f26eSSascha Wildner sc->cpu_cx_stats[cx_next_idx]++; 8045db2f26eSSascha Wildner KASSERT(cx_next->type != ACPI_STATE_C0, ("acpi_cpu_idle: C0 sleep")); 8055db2f26eSSascha Wildner 8065db2f26eSSascha Wildner /* 8075db2f26eSSascha Wildner * Execute HLT (or equivalent) and wait for an interrupt. We can't 8085db2f26eSSascha Wildner * calculate the time spent in C1 since the place we wake up is an 8095db2f26eSSascha Wildner * ISR. Assume we slept half of quantum and return. 8105db2f26eSSascha Wildner */ 8115db2f26eSSascha Wildner if (cx_next->type == ACPI_STATE_C1) { 8125db2f26eSSascha Wildner sc->cpu_prev_sleep = (sc->cpu_prev_sleep * 3 + 500000 / hz) / 4; 8135db2f26eSSascha Wildner acpi_cpu_c1(); 8145db2f26eSSascha Wildner return; 8155db2f26eSSascha Wildner } 8165db2f26eSSascha Wildner 8175db2f26eSSascha Wildner /* 818b42fff25SSepherosa Ziehau * For C3(+), disable bus master arbitration and enable bus master wake 8195db2f26eSSascha Wildner * if BM control is available, otherwise flush the CPU cache. 8205db2f26eSSascha Wildner */ 821b42fff25SSepherosa Ziehau if (cx_next->type >= ACPI_STATE_C3) { 8225db2f26eSSascha Wildner if ((cpu_quirks & CPU_QUIRK_NO_BM_CTRL) == 0) { 8235db2f26eSSascha Wildner AcpiWriteBitRegister(ACPI_BITREG_ARB_DISABLE, 1); 8245db2f26eSSascha Wildner AcpiWriteBitRegister(ACPI_BITREG_BUS_MASTER_RLD, 1); 8255db2f26eSSascha Wildner } else 8265db2f26eSSascha Wildner ACPI_FLUSH_CPU_CACHE(); 8275db2f26eSSascha Wildner } 8285db2f26eSSascha Wildner 8295db2f26eSSascha Wildner /* 8305db2f26eSSascha Wildner * Read from P_LVLx to enter C2(+), checking time spent asleep. 8315db2f26eSSascha Wildner * Use the ACPI timer for measuring sleep time. Since we need to 8325db2f26eSSascha Wildner * get the time very close to the CPU start/stop clock logic, this 8335db2f26eSSascha Wildner * is the only reliable time source. 8345db2f26eSSascha Wildner */ 8355db2f26eSSascha Wildner AcpiRead(&start_time, &AcpiGbl_FADT.XPmTimerBlock); 8365db2f26eSSascha Wildner CPU_GET_REG(cx_next->p_lvlx, 1); 8375db2f26eSSascha Wildner 8385db2f26eSSascha Wildner /* 8395db2f26eSSascha Wildner * Read the end time twice. Since it may take an arbitrary time 8405db2f26eSSascha Wildner * to enter the idle state, the first read may be executed before 8415db2f26eSSascha Wildner * the processor has stopped. Doing it again provides enough 8425db2f26eSSascha Wildner * margin that we are certain to have a correct value. 8435db2f26eSSascha Wildner */ 8445db2f26eSSascha Wildner AcpiRead(&end_time, &AcpiGbl_FADT.XPmTimerBlock); 8455db2f26eSSascha Wildner AcpiRead(&end_time, &AcpiGbl_FADT.XPmTimerBlock); 8465db2f26eSSascha Wildner 8475db2f26eSSascha Wildner /* Enable bus master arbitration and disable bus master wakeup. */ 848b42fff25SSepherosa Ziehau if (cx_next->type >= ACPI_STATE_C3) { 8495db2f26eSSascha Wildner if ((cpu_quirks & CPU_QUIRK_NO_BM_CTRL) == 0) { 8505db2f26eSSascha Wildner AcpiWriteBitRegister(ACPI_BITREG_ARB_DISABLE, 0); 8515db2f26eSSascha Wildner AcpiWriteBitRegister(ACPI_BITREG_BUS_MASTER_RLD, 0); 8525db2f26eSSascha Wildner } 8535db2f26eSSascha Wildner } 8545db2f26eSSascha Wildner ACPI_ENABLE_IRQS(); 8555db2f26eSSascha Wildner 8565db2f26eSSascha Wildner /* Find the actual time asleep in microseconds. */ 8575db2f26eSSascha Wildner end_time = acpi_TimerDelta(end_time, start_time); 8585db2f26eSSascha Wildner sc->cpu_prev_sleep = (sc->cpu_prev_sleep * 3 + PM_USEC(end_time)) / 4; 8595db2f26eSSascha Wildner } 8605db2f26eSSascha Wildner 8615db2f26eSSascha Wildner /* 8625db2f26eSSascha Wildner * Re-evaluate the _CST object when we are notified that it changed. 8635db2f26eSSascha Wildner */ 8645db2f26eSSascha Wildner static void 8655db2f26eSSascha Wildner acpi_cpu_cst_notify(device_t dev) 8665db2f26eSSascha Wildner { 8675db2f26eSSascha Wildner struct acpi_cpu_softc *sc = device_get_softc(dev); 8685db2f26eSSascha Wildner struct acpi_cpu_softc *isc; 8695db2f26eSSascha Wildner int i; 8705db2f26eSSascha Wildner 871*b45624acSSepherosa Ziehau KASSERT(curthread->td_type != TD_TYPE_NETISR, 872*b45624acSSepherosa Ziehau ("notify in netisr%d", mycpuid)); 873*b45624acSSepherosa Ziehau 874*b45624acSSepherosa Ziehau lwkt_serialize_enter(&cpu_cx_slize); 875*b45624acSSepherosa Ziehau 8765db2f26eSSascha Wildner /* Update the list of Cx states. */ 8775db2f26eSSascha Wildner acpi_cpu_cx_cst(sc); 8785db2f26eSSascha Wildner acpi_cpu_cx_list(sc); 8795db2f26eSSascha Wildner 8805db2f26eSSascha Wildner /* Update the new lowest useable Cx state for all CPUs. */ 8815db2f26eSSascha Wildner cpu_cx_count = 0; 8825db2f26eSSascha Wildner for (i = 0; i < cpu_ndevices; i++) { 8835db2f26eSSascha Wildner isc = device_get_softc(cpu_devices[i]); 8845db2f26eSSascha Wildner if (isc->cpu_cx_count > cpu_cx_count) 8855db2f26eSSascha Wildner cpu_cx_count = isc->cpu_cx_count; 8865db2f26eSSascha Wildner } 8871d730338SSepherosa Ziehau 8881d730338SSepherosa Ziehau /* 8891d730338SSepherosa Ziehau * Fix up the lowest Cx being used 8901d730338SSepherosa Ziehau */ 8911d730338SSepherosa Ziehau if (cpu_cx_lowest_req < cpu_cx_count) 8921d730338SSepherosa Ziehau cpu_cx_lowest = cpu_cx_lowest_req; 8931d730338SSepherosa Ziehau if (cpu_cx_lowest > cpu_cx_count - 1) 8941d730338SSepherosa Ziehau cpu_cx_lowest = cpu_cx_count - 1; 8951d730338SSepherosa Ziehau 896*b45624acSSepherosa Ziehau lwkt_serialize_exit(&cpu_cx_slize); 8975db2f26eSSascha Wildner } 8985db2f26eSSascha Wildner 8995db2f26eSSascha Wildner static int 9005db2f26eSSascha Wildner acpi_cpu_quirks(void) 9015db2f26eSSascha Wildner { 9025db2f26eSSascha Wildner device_t acpi_dev; 9035db2f26eSSascha Wildner uint32_t val; 9045db2f26eSSascha Wildner 9055db2f26eSSascha Wildner ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__); 9065db2f26eSSascha Wildner 9075db2f26eSSascha Wildner /* 9085db2f26eSSascha Wildner * Bus mastering arbitration control is needed to keep caches coherent 9095db2f26eSSascha Wildner * while sleeping in C3. If it's not present but a working flush cache 9105db2f26eSSascha Wildner * instruction is present, flush the caches before entering C3 instead. 9115db2f26eSSascha Wildner * Otherwise, just disable C3 completely. 9125db2f26eSSascha Wildner */ 9135db2f26eSSascha Wildner if (AcpiGbl_FADT.Pm2ControlBlock == 0 || 9145db2f26eSSascha Wildner AcpiGbl_FADT.Pm2ControlLength == 0) { 9155db2f26eSSascha Wildner if ((AcpiGbl_FADT.Flags & ACPI_FADT_WBINVD) && 9165db2f26eSSascha Wildner (AcpiGbl_FADT.Flags & ACPI_FADT_WBINVD_FLUSH) == 0) { 9175db2f26eSSascha Wildner cpu_quirks |= CPU_QUIRK_NO_BM_CTRL; 9185db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 9195db2f26eSSascha Wildner "acpi_cpu: no BM control, using flush cache method\n")); 9205db2f26eSSascha Wildner } else { 9215db2f26eSSascha Wildner cpu_quirks |= CPU_QUIRK_NO_C3; 9225db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 9235db2f26eSSascha Wildner "acpi_cpu: no BM control, C3 not available\n")); 9245db2f26eSSascha Wildner } 9255db2f26eSSascha Wildner } 9265db2f26eSSascha Wildner 9275db2f26eSSascha Wildner /* 9285db2f26eSSascha Wildner * If we are using generic Cx mode, C3 on multiple CPUs requires using 9295db2f26eSSascha Wildner * the expensive flush cache instruction. 9305db2f26eSSascha Wildner */ 9315db2f26eSSascha Wildner if (cpu_cx_generic && ncpus > 1) { 9325db2f26eSSascha Wildner cpu_quirks |= CPU_QUIRK_NO_BM_CTRL; 9335db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 9345db2f26eSSascha Wildner "acpi_cpu: SMP, using flush cache mode for C3\n")); 9355db2f26eSSascha Wildner } 9365db2f26eSSascha Wildner 9375db2f26eSSascha Wildner /* Look for various quirks of the PIIX4 part. */ 9385db2f26eSSascha Wildner acpi_dev = pci_find_device(PCI_VENDOR_INTEL, PCI_DEVICE_82371AB_3); 9395db2f26eSSascha Wildner if (acpi_dev != NULL) { 9405db2f26eSSascha Wildner switch (pci_get_revid(acpi_dev)) { 9415db2f26eSSascha Wildner /* 9425db2f26eSSascha Wildner * Disable C3 support for all PIIX4 chipsets. Some of these parts 9435db2f26eSSascha Wildner * do not report the BMIDE status to the BM status register and 9445db2f26eSSascha Wildner * others have a livelock bug if Type-F DMA is enabled. Linux 9455db2f26eSSascha Wildner * works around the BMIDE bug by reading the BM status directly 9465db2f26eSSascha Wildner * but we take the simpler approach of disabling C3 for these 9475db2f26eSSascha Wildner * parts. 9485db2f26eSSascha Wildner * 9495db2f26eSSascha Wildner * See erratum #18 ("C3 Power State/BMIDE and Type-F DMA 9505db2f26eSSascha Wildner * Livelock") from the January 2002 PIIX4 specification update. 9515db2f26eSSascha Wildner * Applies to all PIIX4 models. 9525db2f26eSSascha Wildner * 9535db2f26eSSascha Wildner * Also, make sure that all interrupts cause a "Stop Break" 9545db2f26eSSascha Wildner * event to exit from C2 state. 9555db2f26eSSascha Wildner * Also, BRLD_EN_BM (ACPI_BITREG_BUS_MASTER_RLD in ACPI-speak) 9565db2f26eSSascha Wildner * should be set to zero, otherwise it causes C2 to short-sleep. 9575db2f26eSSascha Wildner * PIIX4 doesn't properly support C3 and bus master activity 9585db2f26eSSascha Wildner * need not break out of C2. 9595db2f26eSSascha Wildner */ 9605db2f26eSSascha Wildner case PCI_REVISION_A_STEP: 9615db2f26eSSascha Wildner case PCI_REVISION_B_STEP: 9625db2f26eSSascha Wildner case PCI_REVISION_4E: 9635db2f26eSSascha Wildner case PCI_REVISION_4M: 9645db2f26eSSascha Wildner cpu_quirks |= CPU_QUIRK_NO_C3; 9655db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 9665db2f26eSSascha Wildner "acpi_cpu: working around PIIX4 bug, disabling C3\n")); 9675db2f26eSSascha Wildner 9685db2f26eSSascha Wildner val = pci_read_config(acpi_dev, PIIX4_DEVACTB_REG, 4); 9695db2f26eSSascha Wildner if ((val & PIIX4_STOP_BREAK_MASK) != PIIX4_STOP_BREAK_MASK) { 9705db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 9715db2f26eSSascha Wildner "acpi_cpu: PIIX4: enabling IRQs to generate Stop Break\n")); 9725db2f26eSSascha Wildner val |= PIIX4_STOP_BREAK_MASK; 9735db2f26eSSascha Wildner pci_write_config(acpi_dev, PIIX4_DEVACTB_REG, val, 4); 9745db2f26eSSascha Wildner } 9755db2f26eSSascha Wildner AcpiReadBitRegister(ACPI_BITREG_BUS_MASTER_RLD, &val); 9765db2f26eSSascha Wildner if (val) { 9775db2f26eSSascha Wildner ACPI_DEBUG_PRINT((ACPI_DB_INFO, 9785db2f26eSSascha Wildner "acpi_cpu: PIIX4: reset BRLD_EN_BM\n")); 9795db2f26eSSascha Wildner AcpiWriteBitRegister(ACPI_BITREG_BUS_MASTER_RLD, 0); 9805db2f26eSSascha Wildner } 9815db2f26eSSascha Wildner break; 9825db2f26eSSascha Wildner default: 9835db2f26eSSascha Wildner break; 9845db2f26eSSascha Wildner } 9855db2f26eSSascha Wildner } 9865db2f26eSSascha Wildner 9875db2f26eSSascha Wildner return (0); 9885db2f26eSSascha Wildner } 9895db2f26eSSascha Wildner 9905db2f26eSSascha Wildner static int 9915db2f26eSSascha Wildner acpi_cpu_usage_sysctl(SYSCTL_HANDLER_ARGS) 9925db2f26eSSascha Wildner { 9935db2f26eSSascha Wildner struct acpi_cpu_softc *sc; 9945db2f26eSSascha Wildner struct sbuf sb; 9955db2f26eSSascha Wildner char buf[128]; 9965db2f26eSSascha Wildner int i; 9975db2f26eSSascha Wildner uintmax_t fract, sum, whole; 9985db2f26eSSascha Wildner 9995db2f26eSSascha Wildner sc = (struct acpi_cpu_softc *) arg1; 10005db2f26eSSascha Wildner sum = 0; 10015db2f26eSSascha Wildner for (i = 0; i < sc->cpu_cx_count; i++) 10025db2f26eSSascha Wildner sum += sc->cpu_cx_stats[i]; 10035db2f26eSSascha Wildner sbuf_new(&sb, buf, sizeof(buf), SBUF_FIXEDLEN); 10045db2f26eSSascha Wildner for (i = 0; i < sc->cpu_cx_count; i++) { 10055db2f26eSSascha Wildner if (sum > 0) { 10065db2f26eSSascha Wildner whole = (uintmax_t)sc->cpu_cx_stats[i] * 100; 10075db2f26eSSascha Wildner fract = (whole % sum) * 100; 10085db2f26eSSascha Wildner sbuf_printf(&sb, "%u.%02u%% ", (u_int)(whole / sum), 10095db2f26eSSascha Wildner (u_int)(fract / sum)); 10105db2f26eSSascha Wildner } else 10115db2f26eSSascha Wildner sbuf_printf(&sb, "0.00%% "); 10125db2f26eSSascha Wildner } 10135db2f26eSSascha Wildner sbuf_printf(&sb, "last %dus", sc->cpu_prev_sleep); 10145db2f26eSSascha Wildner sbuf_trim(&sb); 10155db2f26eSSascha Wildner sbuf_finish(&sb); 10165db2f26eSSascha Wildner sysctl_handle_string(oidp, sbuf_data(&sb), sbuf_len(&sb), req); 10175db2f26eSSascha Wildner sbuf_delete(&sb); 10185db2f26eSSascha Wildner 10195db2f26eSSascha Wildner return (0); 10205db2f26eSSascha Wildner } 10215db2f26eSSascha Wildner 10225db2f26eSSascha Wildner static int 10235db2f26eSSascha Wildner acpi_cpu_set_cx_lowest(struct acpi_cpu_softc *sc, int val) 10245db2f26eSSascha Wildner { 10255db2f26eSSascha Wildner int i, old_lowest, error = 0; 10265db2f26eSSascha Wildner uint32_t old_type, type; 10275db2f26eSSascha Wildner 10281d730338SSepherosa Ziehau sc->cpu_cx_lowest_req = val; 10291d730338SSepherosa Ziehau if (val > sc->cpu_cx_count - 1) 10301d730338SSepherosa Ziehau val = sc->cpu_cx_count - 1; 10315db2f26eSSascha Wildner old_lowest = atomic_swap_int(&sc->cpu_cx_lowest, val); 10325db2f26eSSascha Wildner 10335db2f26eSSascha Wildner old_type = sc->cpu_cx_states[old_lowest].type; 10345db2f26eSSascha Wildner type = sc->cpu_cx_states[val].type; 1035b42fff25SSepherosa Ziehau if (old_type >= ACPI_STATE_C3 && type < ACPI_STATE_C3) { 10365db2f26eSSascha Wildner KKASSERT(cpu_c3_ncpus > 0); 10375db2f26eSSascha Wildner if (atomic_fetchadd_int(&cpu_c3_ncpus, -1) == 1) { 10385db2f26eSSascha Wildner /* 10395db2f26eSSascha Wildner * All of the CPUs exit C3 state, use a better 10405db2f26eSSascha Wildner * one shot timer. 10415db2f26eSSascha Wildner */ 10425db2f26eSSascha Wildner error = cputimer_intr_select_caps(CPUTIMER_INTR_CAP_NONE); 10433b24650bSSepherosa Ziehau KKASSERT(!error || error == ERESTART); 10443b24650bSSepherosa Ziehau if (error == ERESTART) { 10453b24650bSSepherosa Ziehau if (bootverbose) 10463b24650bSSepherosa Ziehau kprintf("exit C3, restart intr cputimer\n"); 10475db2f26eSSascha Wildner cputimer_intr_restart(); 10485db2f26eSSascha Wildner } 10493b24650bSSepherosa Ziehau } 1050b42fff25SSepherosa Ziehau } else if (type >= ACPI_STATE_C3 && old_type < ACPI_STATE_C3) { 10515db2f26eSSascha Wildner if (atomic_fetchadd_int(&cpu_c3_ncpus, 1) == 0) { 10525db2f26eSSascha Wildner /* 1053b42fff25SSepherosa Ziehau * When the first CPU enters C3(+) state, switch 10545db2f26eSSascha Wildner * to an one shot timer, which could handle 1055b42fff25SSepherosa Ziehau * C3(+) state, i.e. the timer will not hang. 10565db2f26eSSascha Wildner */ 10575db2f26eSSascha Wildner error = cputimer_intr_select_caps(CPUTIMER_INTR_CAP_PS); 10583b24650bSSepherosa Ziehau if (error == ERESTART) { 10593b24650bSSepherosa Ziehau if (bootverbose) 10603b24650bSSepherosa Ziehau kprintf("enter C3, restart intr cputimer\n"); 10615db2f26eSSascha Wildner cputimer_intr_restart(); 10623b24650bSSepherosa Ziehau } else if (error) { 10635db2f26eSSascha Wildner kprintf("no suitable intr cputimer found\n"); 10645db2f26eSSascha Wildner 10655db2f26eSSascha Wildner /* Restore */ 10665db2f26eSSascha Wildner sc->cpu_cx_lowest = old_lowest; 10675db2f26eSSascha Wildner atomic_fetchadd_int(&cpu_c3_ncpus, -1); 10685db2f26eSSascha Wildner } 10695db2f26eSSascha Wildner } 10705db2f26eSSascha Wildner } 10715db2f26eSSascha Wildner 10725db2f26eSSascha Wildner if (error) 10735db2f26eSSascha Wildner return error; 10745db2f26eSSascha Wildner 10755db2f26eSSascha Wildner /* If not disabling, cache the new lowest non-C3 state. */ 10765db2f26eSSascha Wildner sc->cpu_non_c3 = 0; 10775db2f26eSSascha Wildner for (i = sc->cpu_cx_lowest; i >= 0; i--) { 10785db2f26eSSascha Wildner if (sc->cpu_cx_states[i].type < ACPI_STATE_C3) { 10795db2f26eSSascha Wildner sc->cpu_non_c3 = i; 10805db2f26eSSascha Wildner break; 10815db2f26eSSascha Wildner } 10825db2f26eSSascha Wildner } 10835db2f26eSSascha Wildner 10845db2f26eSSascha Wildner /* Reset the statistics counters. */ 10855db2f26eSSascha Wildner bzero(sc->cpu_cx_stats, sizeof(sc->cpu_cx_stats)); 10865db2f26eSSascha Wildner return (0); 10875db2f26eSSascha Wildner } 10885db2f26eSSascha Wildner 10895db2f26eSSascha Wildner static int 10905db2f26eSSascha Wildner acpi_cpu_cx_lowest_sysctl(SYSCTL_HANDLER_ARGS) 10915db2f26eSSascha Wildner { 10925db2f26eSSascha Wildner struct acpi_cpu_softc *sc; 10935db2f26eSSascha Wildner char state[8]; 10945db2f26eSSascha Wildner int val, error; 10955db2f26eSSascha Wildner 10965db2f26eSSascha Wildner sc = (struct acpi_cpu_softc *)arg1; 10971d730338SSepherosa Ziehau ksnprintf(state, sizeof(state), "C%d", sc->cpu_cx_lowest_req + 1); 10985db2f26eSSascha Wildner error = sysctl_handle_string(oidp, state, sizeof(state), req); 10995db2f26eSSascha Wildner if (error != 0 || req->newptr == NULL) 11005db2f26eSSascha Wildner return (error); 11015db2f26eSSascha Wildner if (strlen(state) < 2 || toupper(state[0]) != 'C') 11025db2f26eSSascha Wildner return (EINVAL); 11035db2f26eSSascha Wildner val = (int) strtol(state + 1, NULL, 10) - 1; 11041d730338SSepherosa Ziehau if (val < 0) 11055db2f26eSSascha Wildner return (EINVAL); 11065db2f26eSSascha Wildner 1107*b45624acSSepherosa Ziehau lwkt_serialize_enter(&cpu_cx_slize); 11085db2f26eSSascha Wildner error = acpi_cpu_set_cx_lowest(sc, val); 1109*b45624acSSepherosa Ziehau lwkt_serialize_exit(&cpu_cx_slize); 11105db2f26eSSascha Wildner 11115db2f26eSSascha Wildner return error; 11125db2f26eSSascha Wildner } 11135db2f26eSSascha Wildner 11145db2f26eSSascha Wildner static int 11151d730338SSepherosa Ziehau acpi_cpu_cx_lowest_use_sysctl(SYSCTL_HANDLER_ARGS) 11161d730338SSepherosa Ziehau { 11171d730338SSepherosa Ziehau struct acpi_cpu_softc *sc; 11181d730338SSepherosa Ziehau char state[8]; 11191d730338SSepherosa Ziehau 11201d730338SSepherosa Ziehau sc = (struct acpi_cpu_softc *)arg1; 11211d730338SSepherosa Ziehau ksnprintf(state, sizeof(state), "C%d", sc->cpu_cx_lowest + 1); 11221d730338SSepherosa Ziehau return sysctl_handle_string(oidp, state, sizeof(state), req); 11231d730338SSepherosa Ziehau } 11241d730338SSepherosa Ziehau 11251d730338SSepherosa Ziehau static int 11265db2f26eSSascha Wildner acpi_cpu_global_cx_lowest_sysctl(SYSCTL_HANDLER_ARGS) 11275db2f26eSSascha Wildner { 11285db2f26eSSascha Wildner struct acpi_cpu_softc *sc; 11295db2f26eSSascha Wildner char state[8]; 11305db2f26eSSascha Wildner int val, error, i; 11315db2f26eSSascha Wildner 11321d730338SSepherosa Ziehau ksnprintf(state, sizeof(state), "C%d", cpu_cx_lowest_req + 1); 11335db2f26eSSascha Wildner error = sysctl_handle_string(oidp, state, sizeof(state), req); 11345db2f26eSSascha Wildner if (error != 0 || req->newptr == NULL) 11355db2f26eSSascha Wildner return (error); 11365db2f26eSSascha Wildner if (strlen(state) < 2 || toupper(state[0]) != 'C') 11375db2f26eSSascha Wildner return (EINVAL); 11385db2f26eSSascha Wildner val = (int) strtol(state + 1, NULL, 10) - 1; 11391d730338SSepherosa Ziehau if (val < 0) 11405db2f26eSSascha Wildner return (EINVAL); 11411d730338SSepherosa Ziehau 1142*b45624acSSepherosa Ziehau lwkt_serialize_enter(&cpu_cx_slize); 1143*b45624acSSepherosa Ziehau 11441d730338SSepherosa Ziehau cpu_cx_lowest_req = val; 11455db2f26eSSascha Wildner cpu_cx_lowest = val; 11461d730338SSepherosa Ziehau if (cpu_cx_lowest > cpu_cx_count - 1) 11471d730338SSepherosa Ziehau cpu_cx_lowest = cpu_cx_count - 1; 11485db2f26eSSascha Wildner 11495db2f26eSSascha Wildner /* Update the new lowest useable Cx state for all CPUs. */ 11505db2f26eSSascha Wildner for (i = 0; i < cpu_ndevices; i++) { 11515db2f26eSSascha Wildner sc = device_get_softc(cpu_devices[i]); 11525db2f26eSSascha Wildner error = acpi_cpu_set_cx_lowest(sc, val); 11535db2f26eSSascha Wildner if (error) { 11545db2f26eSSascha Wildner KKASSERT(i == 0); 11555db2f26eSSascha Wildner break; 11565db2f26eSSascha Wildner } 11575db2f26eSSascha Wildner } 1158*b45624acSSepherosa Ziehau 1159*b45624acSSepherosa Ziehau lwkt_serialize_exit(&cpu_cx_slize); 11605db2f26eSSascha Wildner 11615db2f26eSSascha Wildner return error; 11625db2f26eSSascha Wildner } 11635db2f26eSSascha Wildner 11641d730338SSepherosa Ziehau static int 11651d730338SSepherosa Ziehau acpi_cpu_global_cx_lowest_use_sysctl(SYSCTL_HANDLER_ARGS) 11661d730338SSepherosa Ziehau { 11671d730338SSepherosa Ziehau char state[8]; 11681d730338SSepherosa Ziehau 11691d730338SSepherosa Ziehau ksnprintf(state, sizeof(state), "C%d", cpu_cx_lowest + 1); 11701d730338SSepherosa Ziehau return sysctl_handle_string(oidp, state, sizeof(state), req); 11711d730338SSepherosa Ziehau } 11721d730338SSepherosa Ziehau 11735db2f26eSSascha Wildner /* 11745db2f26eSSascha Wildner * Put the CPU in C1 in a machine-dependant way. 11755db2f26eSSascha Wildner * XXX: shouldn't be here! 11765db2f26eSSascha Wildner */ 11775db2f26eSSascha Wildner static void 11785db2f26eSSascha Wildner acpi_cpu_c1(void) 11795db2f26eSSascha Wildner { 11805db2f26eSSascha Wildner #ifdef __ia64__ 11815db2f26eSSascha Wildner ia64_call_pal_static(PAL_HALT_LIGHT, 0, 0, 0); 11825db2f26eSSascha Wildner #else 11835db2f26eSSascha Wildner splz(); 11845db2f26eSSascha Wildner if ((mycpu->gd_reqflags & RQF_IDLECHECK_WK_MASK) == 0) 11855db2f26eSSascha Wildner __asm __volatile("sti; hlt"); 11865db2f26eSSascha Wildner else 11875db2f26eSSascha Wildner __asm __volatile("sti; pause"); 11885db2f26eSSascha Wildner #endif /* !__ia64__ */ 11895db2f26eSSascha Wildner } 1190