11772Sjl139090 /* 21772Sjl139090 * CDDL HEADER START 31772Sjl139090 * 41772Sjl139090 * The contents of this file are subject to the terms of the 51772Sjl139090 * Common Development and Distribution License (the "License"). 61772Sjl139090 * You may not use this file except in compliance with the License. 71772Sjl139090 * 81772Sjl139090 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 91772Sjl139090 * or http://www.opensolaris.org/os/licensing. 101772Sjl139090 * See the License for the specific language governing permissions 111772Sjl139090 * and limitations under the License. 121772Sjl139090 * 131772Sjl139090 * When distributing Covered Code, include this CDDL HEADER in each 141772Sjl139090 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 151772Sjl139090 * If applicable, add the following below this CDDL HEADER, with the 161772Sjl139090 * fields enclosed by brackets "[]" replaced with your own identifying 171772Sjl139090 * information: Portions Copyright [yyyy] [name of copyright owner] 181772Sjl139090 * 191772Sjl139090 * CDDL HEADER END 201772Sjl139090 */ 211772Sjl139090 /* 223354Sjl139090 * Copyright 2007 Sun Microsystems, Inc. All rights reserved. 231772Sjl139090 * Use is subject to license terms. 241772Sjl139090 */ 251772Sjl139090 261772Sjl139090 #pragma ident "%Z%%M% %I% %E% SMI" 271772Sjl139090 281772Sjl139090 #include <sys/cpuvar.h> 291772Sjl139090 #include <sys/systm.h> 301772Sjl139090 #include <sys/sysmacros.h> 311772Sjl139090 #include <sys/promif.h> 321772Sjl139090 #include <sys/platform_module.h> 331772Sjl139090 #include <sys/cmn_err.h> 341772Sjl139090 #include <sys/errno.h> 351772Sjl139090 #include <sys/machsystm.h> 361772Sjl139090 #include <sys/bootconf.h> 371772Sjl139090 #include <sys/nvpair.h> 381772Sjl139090 #include <sys/kobj.h> 391772Sjl139090 #include <sys/mem_cage.h> 401772Sjl139090 #include <sys/opl.h> 411772Sjl139090 #include <sys/scfd/scfostoescf.h> 421772Sjl139090 #include <sys/cpu_sgnblk_defs.h> 431772Sjl139090 #include <sys/utsname.h> 441772Sjl139090 #include <sys/ddi.h> 451772Sjl139090 #include <sys/sunndi.h> 461772Sjl139090 #include <sys/lgrp.h> 471772Sjl139090 #include <sys/memnode.h> 481772Sjl139090 #include <sys/sysmacros.h> 493914Spm145316 #include <sys/time.h> 503914Spm145316 #include <sys/cpu.h> 511772Sjl139090 #include <vm/vm_dep.h> 521772Sjl139090 531772Sjl139090 int (*opl_get_mem_unum)(int, uint64_t, char *, int, int *); 542214Sav145390 int (*opl_get_mem_sid)(char *unum, char *buf, int buflen, int *lenp); 552214Sav145390 int (*opl_get_mem_offset)(uint64_t paddr, uint64_t *offp); 562214Sav145390 int (*opl_get_mem_addr)(char *unum, char *sid, 572214Sav145390 uint64_t offset, uint64_t *paddr); 581772Sjl139090 591772Sjl139090 /* Memory for fcode claims. 16k times # maximum possible IO units */ 601772Sjl139090 #define EFCODE_SIZE (OPL_MAX_BOARDS * OPL_MAX_IO_UNITS_PER_BOARD * 0x4000) 611772Sjl139090 int efcode_size = EFCODE_SIZE; 621772Sjl139090 631772Sjl139090 #define OPL_MC_MEMBOARD_SHIFT 38 /* Boards on 256BG boundary */ 641772Sjl139090 651772Sjl139090 /* Set the maximum number of boards for DR */ 661772Sjl139090 int opl_boards = OPL_MAX_BOARDS; 671772Sjl139090 681772Sjl139090 void sgn_update_all_cpus(ushort_t, uchar_t, uchar_t); 691772Sjl139090 701772Sjl139090 extern int tsb_lgrp_affinity; 711772Sjl139090 721772Sjl139090 int opl_tsb_spares = (OPL_MAX_BOARDS) * (OPL_MAX_PCICH_UNITS_PER_BOARD) * 731772Sjl139090 (OPL_MAX_TSBS_PER_PCICH); 741772Sjl139090 751772Sjl139090 pgcnt_t opl_startup_cage_size = 0; 761772Sjl139090 775347Sjfrank /* 785347Sjfrank * The length of the delay in seconds in communication with XSCF after 795347Sjfrank * which the warning message will be logged. 805347Sjfrank */ 815347Sjfrank uint_t xscf_connect_delay = 60 * 15; 825347Sjfrank 832241Shuah static opl_model_info_t opl_models[] = { 843123Ssubhan { "FF1", OPL_MAX_BOARDS_FF1, FF1, STD_DISPATCH_TABLE }, 853123Ssubhan { "FF2", OPL_MAX_BOARDS_FF2, FF2, STD_DISPATCH_TABLE }, 863123Ssubhan { "DC1", OPL_MAX_BOARDS_DC1, DC1, STD_DISPATCH_TABLE }, 873123Ssubhan { "DC2", OPL_MAX_BOARDS_DC2, DC2, EXT_DISPATCH_TABLE }, 883123Ssubhan { "DC3", OPL_MAX_BOARDS_DC3, DC3, EXT_DISPATCH_TABLE }, 892241Shuah }; 902241Shuah static int opl_num_models = sizeof (opl_models)/sizeof (opl_model_info_t); 912241Shuah 923123Ssubhan /* 933627Ssubhan * opl_cur_model 943123Ssubhan */ 953627Ssubhan static opl_model_info_t *opl_cur_model = NULL; 962241Shuah 971772Sjl139090 static struct memlist *opl_memlist_per_board(struct memlist *ml); 985347Sjfrank static void post_xscf_msg(char *, int); 995347Sjfrank static void pass2xscf_thread(); 1001772Sjl139090 1013914Spm145316 /* 1023914Spm145316 * Note FF/DC out-of-order instruction engine takes only a 1033914Spm145316 * single cycle to execute each spin loop 1043914Spm145316 * for comparison, Panther takes 6 cycles for same loop 1053914Spm145316 * 1500 approx nsec for OPL sleep instruction 1063914Spm145316 * if spin count = OPL_BOFF_SLEEP*OPL_BOFF_SPIN then 1073914Spm145316 * spin time should be equal to OPL_BOFF_TM nsecs 1083914Spm145316 * Listed values tuned for 2.15GHz to 2.4GHz systems 1093914Spm145316 * Value may change for future systems 1103914Spm145316 */ 1113914Spm145316 #define OPL_BOFF_SPIN 720 1123914Spm145316 #define OPL_BOFF_BASE 1 1133914Spm145316 #define OPL_BOFF_SLEEP 5 1143914Spm145316 #define OPL_BOFF_CAP1 20 1153914Spm145316 #define OPL_BOFF_CAP2 60 1163914Spm145316 #define OPL_BOFF_MAX (40 * OPL_BOFF_SLEEP) 1173914Spm145316 #define OPL_BOFF_TM 1500 1183914Spm145316 1191772Sjl139090 int 1201772Sjl139090 set_platform_max_ncpus(void) 1211772Sjl139090 { 1221772Sjl139090 return (OPL_MAX_CPU_PER_BOARD * OPL_MAX_BOARDS); 1231772Sjl139090 } 1241772Sjl139090 1251772Sjl139090 int 1261772Sjl139090 set_platform_tsb_spares(void) 1271772Sjl139090 { 1281772Sjl139090 return (MIN(opl_tsb_spares, MAX_UPA)); 1291772Sjl139090 } 1301772Sjl139090 1312241Shuah static void 1322241Shuah set_model_info() 1332241Shuah { 1343123Ssubhan extern int ts_dispatch_extended; 1352241Shuah char name[MAXSYSNAME]; 1362241Shuah int i; 1372241Shuah 1382241Shuah /* 1392241Shuah * Get model name from the root node. 1402241Shuah * 1412241Shuah * We are using the prom device tree since, at this point, 1422241Shuah * the Solaris device tree is not yet setup. 1432241Shuah */ 1442241Shuah (void) prom_getprop(prom_rootnode(), "model", (caddr_t)name); 1452241Shuah 1462241Shuah for (i = 0; i < opl_num_models; i++) { 1472241Shuah if (strncmp(name, opl_models[i].model_name, MAXSYSNAME) == 0) { 1482241Shuah opl_cur_model = &opl_models[i]; 1492241Shuah break; 1502241Shuah } 1512241Shuah } 1523123Ssubhan 1535539Swh31274 /* 1545539Swh31274 * If model not matched, it's an unknown model. 1555539Swh31274 * just return. 1565539Swh31274 */ 1572241Shuah if (i == opl_num_models) 1585539Swh31274 return; 1593123Ssubhan 1603123Ssubhan if ((opl_cur_model->model_cmds & EXT_DISPATCH_TABLE) && 1615037Sjl139090 (ts_dispatch_extended == -1)) { 1623123Ssubhan /* 1633123Ssubhan * Based on a platform model, select a dispatch table. 1643123Ssubhan * Only DC2 and DC3 systems uses the alternate/extended 1653123Ssubhan * TS dispatch table. 1663123Ssubhan * FF1, FF2 and DC1 systems used standard dispatch tables. 1673123Ssubhan */ 1683123Ssubhan ts_dispatch_extended = 1; 1693123Ssubhan } 1703123Ssubhan 1712241Shuah } 1722241Shuah 1732241Shuah static void 1742241Shuah set_max_mmu_ctxdoms() 1752241Shuah { 1762241Shuah extern uint_t max_mmu_ctxdoms; 1772241Shuah int max_boards; 1782241Shuah 1792241Shuah /* 1802241Shuah * From the model, get the maximum number of boards 1812241Shuah * supported and set the value accordingly. If the model 1822241Shuah * could not be determined or recognized, we assume the max value. 1832241Shuah */ 1842241Shuah if (opl_cur_model == NULL) 1852241Shuah max_boards = OPL_MAX_BOARDS; 1862241Shuah else 1872241Shuah max_boards = opl_cur_model->model_max_boards; 1882241Shuah 1892241Shuah /* 1902241Shuah * On OPL, cores and MMUs are one-to-one. 1912241Shuah */ 1922241Shuah max_mmu_ctxdoms = OPL_MAX_CORE_UNITS_PER_BOARD * max_boards; 1932241Shuah } 1942241Shuah 1951772Sjl139090 #pragma weak mmu_init_large_pages 1961772Sjl139090 1971772Sjl139090 void 1981772Sjl139090 set_platform_defaults(void) 1991772Sjl139090 { 2001772Sjl139090 extern char *tod_module_name; 2011772Sjl139090 extern void cpu_sgn_update(ushort_t, uchar_t, uchar_t, int); 2021772Sjl139090 extern void mmu_init_large_pages(size_t); 2031772Sjl139090 2041772Sjl139090 /* Set the CPU signature function pointer */ 2051772Sjl139090 cpu_sgn_func = cpu_sgn_update; 2061772Sjl139090 2071772Sjl139090 /* Set appropriate tod module for OPL platform */ 2081772Sjl139090 ASSERT(tod_module_name == NULL); 2091772Sjl139090 tod_module_name = "todopl"; 2101772Sjl139090 2111772Sjl139090 if ((mmu_page_sizes == max_mmu_page_sizes) && 2122659Ssusans (mmu_ism_pagesize != DEFAULT_ISM_PAGESIZE)) { 2131772Sjl139090 if (&mmu_init_large_pages) 2141772Sjl139090 mmu_init_large_pages(mmu_ism_pagesize); 2151772Sjl139090 } 2161772Sjl139090 2171772Sjl139090 tsb_lgrp_affinity = 1; 2182241Shuah 2192241Shuah set_max_mmu_ctxdoms(); 2201772Sjl139090 } 2211772Sjl139090 2221772Sjl139090 /* 2231772Sjl139090 * Convert logical a board number to a physical one. 2241772Sjl139090 */ 2251772Sjl139090 2261772Sjl139090 #define LSBPROP "board#" 2271772Sjl139090 #define PSBPROP "physical-board#" 2281772Sjl139090 2291772Sjl139090 int 2301772Sjl139090 opl_get_physical_board(int id) 2311772Sjl139090 { 2321772Sjl139090 dev_info_t *root_dip, *dip = NULL; 2331772Sjl139090 char *dname = NULL; 2341772Sjl139090 int circ; 2351772Sjl139090 2361772Sjl139090 pnode_t pnode; 2371772Sjl139090 char pname[MAXSYSNAME] = {0}; 2381772Sjl139090 2391772Sjl139090 int lsb_id; /* Logical System Board ID */ 2401772Sjl139090 int psb_id; /* Physical System Board ID */ 2411772Sjl139090 2421772Sjl139090 2431772Sjl139090 /* 2441772Sjl139090 * This function is called on early stage of bootup when the 2451772Sjl139090 * kernel device tree is not initialized yet, and also 2461772Sjl139090 * later on when the device tree is up. We want to try 2471772Sjl139090 * the fast track first. 2481772Sjl139090 */ 2491772Sjl139090 root_dip = ddi_root_node(); 2501772Sjl139090 if (root_dip) { 2511772Sjl139090 /* Get from devinfo node */ 2521772Sjl139090 ndi_devi_enter(root_dip, &circ); 2531772Sjl139090 for (dip = ddi_get_child(root_dip); dip; 2541772Sjl139090 dip = ddi_get_next_sibling(dip)) { 2551772Sjl139090 2561772Sjl139090 dname = ddi_node_name(dip); 2571772Sjl139090 if (strncmp(dname, "pseudo-mc", 9) != 0) 2581772Sjl139090 continue; 2591772Sjl139090 2601772Sjl139090 if ((lsb_id = (int)ddi_getprop(DDI_DEV_T_ANY, dip, 2611772Sjl139090 DDI_PROP_DONTPASS, LSBPROP, -1)) == -1) 2621772Sjl139090 continue; 2631772Sjl139090 2641772Sjl139090 if (id == lsb_id) { 2651772Sjl139090 if ((psb_id = (int)ddi_getprop(DDI_DEV_T_ANY, 2661772Sjl139090 dip, DDI_PROP_DONTPASS, PSBPROP, -1)) 2671772Sjl139090 == -1) { 2681772Sjl139090 ndi_devi_exit(root_dip, circ); 2691772Sjl139090 return (-1); 2701772Sjl139090 } else { 2711772Sjl139090 ndi_devi_exit(root_dip, circ); 2721772Sjl139090 return (psb_id); 2731772Sjl139090 } 2741772Sjl139090 } 2751772Sjl139090 } 2761772Sjl139090 ndi_devi_exit(root_dip, circ); 2771772Sjl139090 } 2781772Sjl139090 2791772Sjl139090 /* 2801772Sjl139090 * We do not have the kernel device tree, or we did not 2811772Sjl139090 * find the node for some reason (let's say the kernel 2821772Sjl139090 * device tree was modified), let's try the OBP tree. 2831772Sjl139090 */ 2841772Sjl139090 pnode = prom_rootnode(); 2851772Sjl139090 for (pnode = prom_childnode(pnode); pnode; 2861772Sjl139090 pnode = prom_nextnode(pnode)) { 2871772Sjl139090 2881772Sjl139090 if ((prom_getprop(pnode, "name", (caddr_t)pname) == -1) || 2891772Sjl139090 (strncmp(pname, "pseudo-mc", 9) != 0)) 2901772Sjl139090 continue; 2911772Sjl139090 2921772Sjl139090 if (prom_getprop(pnode, LSBPROP, (caddr_t)&lsb_id) == -1) 2931772Sjl139090 continue; 2941772Sjl139090 2951772Sjl139090 if (id == lsb_id) { 2961772Sjl139090 if (prom_getprop(pnode, PSBPROP, 2971772Sjl139090 (caddr_t)&psb_id) == -1) { 2981772Sjl139090 return (-1); 2991772Sjl139090 } else { 3001772Sjl139090 return (psb_id); 3011772Sjl139090 } 3021772Sjl139090 } 3031772Sjl139090 } 3041772Sjl139090 3051772Sjl139090 return (-1); 3061772Sjl139090 } 3071772Sjl139090 3081772Sjl139090 /* 3091772Sjl139090 * For OPL it's possible that memory from two or more successive boards 3101772Sjl139090 * will be contiguous across the boards, and therefore represented as a 3111772Sjl139090 * single chunk. 3121772Sjl139090 * This function splits such chunks down the board boundaries. 3131772Sjl139090 */ 3141772Sjl139090 static struct memlist * 3151772Sjl139090 opl_memlist_per_board(struct memlist *ml) 3161772Sjl139090 { 3171772Sjl139090 uint64_t ssize, low, high, boundary; 3181772Sjl139090 struct memlist *head, *tail, *new; 3191772Sjl139090 3201772Sjl139090 ssize = (1ull << OPL_MC_MEMBOARD_SHIFT); 3211772Sjl139090 3221772Sjl139090 head = tail = NULL; 3231772Sjl139090 3241772Sjl139090 for (; ml; ml = ml->next) { 3251772Sjl139090 low = (uint64_t)ml->address; 3261772Sjl139090 high = low+(uint64_t)(ml->size); 3271772Sjl139090 while (low < high) { 3281772Sjl139090 boundary = roundup(low+1, ssize); 3291772Sjl139090 boundary = MIN(high, boundary); 3301772Sjl139090 new = kmem_zalloc(sizeof (struct memlist), KM_SLEEP); 3311772Sjl139090 new->address = low; 3321772Sjl139090 new->size = boundary - low; 3331772Sjl139090 if (head == NULL) 3341772Sjl139090 head = new; 3351772Sjl139090 if (tail) { 3361772Sjl139090 tail->next = new; 3371772Sjl139090 new->prev = tail; 3381772Sjl139090 } 3391772Sjl139090 tail = new; 3401772Sjl139090 low = boundary; 3411772Sjl139090 } 3421772Sjl139090 } 3431772Sjl139090 return (head); 3441772Sjl139090 } 3451772Sjl139090 3461772Sjl139090 void 3471772Sjl139090 set_platform_cage_params(void) 3481772Sjl139090 { 3491772Sjl139090 extern pgcnt_t total_pages; 3501772Sjl139090 extern struct memlist *phys_avail; 3511772Sjl139090 struct memlist *ml, *tml; 3521772Sjl139090 3531772Sjl139090 if (kernel_cage_enable) { 3541772Sjl139090 pgcnt_t preferred_cage_size; 3551772Sjl139090 3565037Sjl139090 preferred_cage_size = MAX(opl_startup_cage_size, 3575037Sjl139090 total_pages / 256); 3581772Sjl139090 3591772Sjl139090 ml = opl_memlist_per_board(phys_avail); 3601772Sjl139090 3611772Sjl139090 /* 3621772Sjl139090 * Note: we are assuming that post has load the 3631772Sjl139090 * whole show in to the high end of memory. Having 3641772Sjl139090 * taken this leap, we copy the whole of phys_avail 3651772Sjl139090 * the glist and arrange for the cage to grow 3661772Sjl139090 * downward (descending pfns). 3671772Sjl139090 */ 3684266Sdp78419 kcage_range_init(ml, KCAGE_DOWN, preferred_cage_size); 3691772Sjl139090 3701772Sjl139090 /* free the memlist */ 3711772Sjl139090 do { 3721772Sjl139090 tml = ml->next; 3731772Sjl139090 kmem_free(ml, sizeof (struct memlist)); 3741772Sjl139090 ml = tml; 3751772Sjl139090 } while (ml != NULL); 3761772Sjl139090 } 3771772Sjl139090 3781772Sjl139090 if (kcage_on) 3791772Sjl139090 cmn_err(CE_NOTE, "!DR Kernel Cage is ENABLED"); 3801772Sjl139090 else 3811772Sjl139090 cmn_err(CE_NOTE, "!DR Kernel Cage is DISABLED"); 3821772Sjl139090 } 3831772Sjl139090 3841772Sjl139090 /*ARGSUSED*/ 3851772Sjl139090 int 3861772Sjl139090 plat_cpu_poweron(struct cpu *cp) 3871772Sjl139090 { 3881772Sjl139090 int (*opl_cpu_poweron)(struct cpu *) = NULL; 3891772Sjl139090 3901772Sjl139090 opl_cpu_poweron = 3911772Sjl139090 (int (*)(struct cpu *))kobj_getsymvalue("drmach_cpu_poweron", 0); 3921772Sjl139090 3931772Sjl139090 if (opl_cpu_poweron == NULL) 3941772Sjl139090 return (ENOTSUP); 3951772Sjl139090 else 3961772Sjl139090 return ((opl_cpu_poweron)(cp)); 3971772Sjl139090 3981772Sjl139090 } 3991772Sjl139090 4001772Sjl139090 /*ARGSUSED*/ 4011772Sjl139090 int 4021772Sjl139090 plat_cpu_poweroff(struct cpu *cp) 4031772Sjl139090 { 4041772Sjl139090 int (*opl_cpu_poweroff)(struct cpu *) = NULL; 4051772Sjl139090 4061772Sjl139090 opl_cpu_poweroff = 4071772Sjl139090 (int (*)(struct cpu *))kobj_getsymvalue("drmach_cpu_poweroff", 0); 4081772Sjl139090 4091772Sjl139090 if (opl_cpu_poweroff == NULL) 4101772Sjl139090 return (ENOTSUP); 4111772Sjl139090 else 4121772Sjl139090 return ((opl_cpu_poweroff)(cp)); 4131772Sjl139090 4141772Sjl139090 } 4151772Sjl139090 4161772Sjl139090 int 4171772Sjl139090 plat_max_boards(void) 4181772Sjl139090 { 4191772Sjl139090 return (OPL_MAX_BOARDS); 4201772Sjl139090 } 4211772Sjl139090 4221772Sjl139090 int 4231772Sjl139090 plat_max_cpu_units_per_board(void) 4241772Sjl139090 { 4251772Sjl139090 return (OPL_MAX_CPU_PER_BOARD); 4261772Sjl139090 } 4271772Sjl139090 4281772Sjl139090 int 4291772Sjl139090 plat_max_mem_units_per_board(void) 4301772Sjl139090 { 4311772Sjl139090 return (OPL_MAX_MEM_UNITS_PER_BOARD); 4321772Sjl139090 } 4331772Sjl139090 4341772Sjl139090 int 4351772Sjl139090 plat_max_io_units_per_board(void) 4361772Sjl139090 { 4371772Sjl139090 return (OPL_MAX_IO_UNITS_PER_BOARD); 4381772Sjl139090 } 4391772Sjl139090 4401772Sjl139090 int 4411772Sjl139090 plat_max_cmp_units_per_board(void) 4421772Sjl139090 { 4431772Sjl139090 return (OPL_MAX_CMP_UNITS_PER_BOARD); 4441772Sjl139090 } 4451772Sjl139090 4461772Sjl139090 int 4471772Sjl139090 plat_max_core_units_per_board(void) 4481772Sjl139090 { 4491772Sjl139090 return (OPL_MAX_CORE_UNITS_PER_BOARD); 4501772Sjl139090 } 4511772Sjl139090 4521772Sjl139090 int 4531772Sjl139090 plat_pfn_to_mem_node(pfn_t pfn) 4541772Sjl139090 { 4551772Sjl139090 return (pfn >> mem_node_pfn_shift); 4561772Sjl139090 } 4571772Sjl139090 4581772Sjl139090 /* ARGSUSED */ 4591772Sjl139090 void 460*5648Ssetje plat_build_mem_nodes(prom_memlist_t *list, size_t nelems) 4611772Sjl139090 { 4621772Sjl139090 size_t elem; 4631772Sjl139090 pfn_t basepfn; 4641772Sjl139090 pgcnt_t npgs; 4651772Sjl139090 uint64_t boundary, ssize; 4661772Sjl139090 uint64_t low, high; 4671772Sjl139090 4681772Sjl139090 /* 4691772Sjl139090 * OPL mem slices are always aligned on a 256GB boundary. 4701772Sjl139090 */ 4711772Sjl139090 mem_node_pfn_shift = OPL_MC_MEMBOARD_SHIFT - MMU_PAGESHIFT; 4721772Sjl139090 mem_node_physalign = 0; 4731772Sjl139090 4741772Sjl139090 /* 4751772Sjl139090 * Boot install lists are arranged <addr, len>, <addr, len>, ... 4761772Sjl139090 */ 4771772Sjl139090 ssize = (1ull << OPL_MC_MEMBOARD_SHIFT); 478*5648Ssetje for (elem = 0; elem < nelems; list++, elem++) { 479*5648Ssetje low = list->addr; 480*5648Ssetje high = low + list->size; 4811772Sjl139090 while (low < high) { 4821772Sjl139090 boundary = roundup(low+1, ssize); 4831772Sjl139090 boundary = MIN(high, boundary); 4841772Sjl139090 basepfn = btop(low); 4851772Sjl139090 npgs = btop(boundary - low); 4861772Sjl139090 mem_node_add_slice(basepfn, basepfn + npgs - 1); 4871772Sjl139090 low = boundary; 4881772Sjl139090 } 4891772Sjl139090 } 4901772Sjl139090 } 4911772Sjl139090 4921772Sjl139090 /* 4931772Sjl139090 * Find the CPU associated with a slice at boot-time. 4941772Sjl139090 */ 4951772Sjl139090 void 4961772Sjl139090 plat_fill_mc(pnode_t nodeid) 4971772Sjl139090 { 4981772Sjl139090 int board; 4991772Sjl139090 int memnode; 5001772Sjl139090 struct { 5011772Sjl139090 uint64_t addr; 5021772Sjl139090 uint64_t size; 5031772Sjl139090 } mem_range; 5041772Sjl139090 5051772Sjl139090 if (prom_getprop(nodeid, "board#", (caddr_t)&board) < 0) { 5061772Sjl139090 panic("Can not find board# property in mc node %x", nodeid); 5071772Sjl139090 } 5081772Sjl139090 if (prom_getprop(nodeid, "sb-mem-ranges", (caddr_t)&mem_range) < 0) { 5091772Sjl139090 panic("Can not find sb-mem-ranges property in mc node %x", 5105037Sjl139090 nodeid); 5111772Sjl139090 } 5121772Sjl139090 memnode = mem_range.addr >> OPL_MC_MEMBOARD_SHIFT; 5131772Sjl139090 plat_assign_lgrphand_to_mem_node(board, memnode); 5141772Sjl139090 } 5151772Sjl139090 5161772Sjl139090 /* 5171772Sjl139090 * Return the platform handle for the lgroup containing the given CPU 5181772Sjl139090 * 5191772Sjl139090 * For OPL, lgroup platform handle == board #. 5201772Sjl139090 */ 5211772Sjl139090 5221772Sjl139090 extern int mpo_disabled; 5231772Sjl139090 extern lgrp_handle_t lgrp_default_handle; 5241772Sjl139090 5251772Sjl139090 lgrp_handle_t 5261772Sjl139090 plat_lgrp_cpu_to_hand(processorid_t id) 5271772Sjl139090 { 5281772Sjl139090 lgrp_handle_t plathand; 5291772Sjl139090 5301772Sjl139090 /* 5311772Sjl139090 * Return the real platform handle for the CPU until 5321772Sjl139090 * such time as we know that MPO should be disabled. 5331772Sjl139090 * At that point, we set the "mpo_disabled" flag to true, 5341772Sjl139090 * and from that point on, return the default handle. 5351772Sjl139090 * 5361772Sjl139090 * By the time we know that MPO should be disabled, the 5371772Sjl139090 * first CPU will have already been added to a leaf 5381772Sjl139090 * lgroup, but that's ok. The common lgroup code will 5391772Sjl139090 * double check that the boot CPU is in the correct place, 5401772Sjl139090 * and in the case where mpo should be disabled, will move 5411772Sjl139090 * it to the root if necessary. 5421772Sjl139090 */ 5431772Sjl139090 if (mpo_disabled) { 5441772Sjl139090 /* If MPO is disabled, return the default (UMA) handle */ 5451772Sjl139090 plathand = lgrp_default_handle; 5461772Sjl139090 } else 5471772Sjl139090 plathand = (lgrp_handle_t)LSB_ID(id); 5481772Sjl139090 return (plathand); 5491772Sjl139090 } 5501772Sjl139090 5511772Sjl139090 /* 5521772Sjl139090 * Platform specific lgroup initialization 5531772Sjl139090 */ 5541772Sjl139090 void 5551772Sjl139090 plat_lgrp_init(void) 5561772Sjl139090 { 5571772Sjl139090 extern uint32_t lgrp_expand_proc_thresh; 5581772Sjl139090 extern uint32_t lgrp_expand_proc_diff; 5591772Sjl139090 5601772Sjl139090 /* 5611772Sjl139090 * Set tuneables for the OPL architecture 5621772Sjl139090 * 5631772Sjl139090 * lgrp_expand_proc_thresh is the minimum load on the lgroups 5641772Sjl139090 * this process is currently running on before considering 5651772Sjl139090 * expanding threads to another lgroup. 5661772Sjl139090 * 5671772Sjl139090 * lgrp_expand_proc_diff determines how much less the remote lgroup 5681772Sjl139090 * must be loaded before expanding to it. 5691772Sjl139090 * 5701772Sjl139090 * Since remote latencies can be costly, attempt to keep 3 threads 5711772Sjl139090 * within the same lgroup before expanding to the next lgroup. 5721772Sjl139090 */ 5731772Sjl139090 lgrp_expand_proc_thresh = LGRP_LOADAVG_THREAD_MAX * 3; 5741772Sjl139090 lgrp_expand_proc_diff = LGRP_LOADAVG_THREAD_MAX; 5751772Sjl139090 } 5761772Sjl139090 5771772Sjl139090 /* 5781772Sjl139090 * Platform notification of lgroup (re)configuration changes 5791772Sjl139090 */ 5801772Sjl139090 /*ARGSUSED*/ 5811772Sjl139090 void 5821772Sjl139090 plat_lgrp_config(lgrp_config_flag_t evt, uintptr_t arg) 5831772Sjl139090 { 5841772Sjl139090 update_membounds_t *umb; 5851772Sjl139090 lgrp_config_mem_rename_t lmr; 5861772Sjl139090 int sbd, tbd; 5871772Sjl139090 lgrp_handle_t hand, shand, thand; 5881772Sjl139090 int mnode, snode, tnode; 5891772Sjl139090 pfn_t start, end; 5901772Sjl139090 5911772Sjl139090 if (mpo_disabled) 5921772Sjl139090 return; 5931772Sjl139090 5941772Sjl139090 switch (evt) { 5951772Sjl139090 5961772Sjl139090 case LGRP_CONFIG_MEM_ADD: 5971772Sjl139090 /* 5981772Sjl139090 * Establish the lgroup handle to memnode translation. 5991772Sjl139090 */ 6001772Sjl139090 umb = (update_membounds_t *)arg; 6011772Sjl139090 6021772Sjl139090 hand = umb->u_board; 6031772Sjl139090 mnode = plat_pfn_to_mem_node(umb->u_base >> MMU_PAGESHIFT); 6041772Sjl139090 plat_assign_lgrphand_to_mem_node(hand, mnode); 6051772Sjl139090 6061772Sjl139090 break; 6071772Sjl139090 6081772Sjl139090 case LGRP_CONFIG_MEM_DEL: 6091772Sjl139090 /* 6101772Sjl139090 * Special handling for possible memory holes. 6111772Sjl139090 */ 6121772Sjl139090 umb = (update_membounds_t *)arg; 6131772Sjl139090 hand = umb->u_board; 6141772Sjl139090 if ((mnode = plat_lgrphand_to_mem_node(hand)) != -1) { 6151772Sjl139090 if (mem_node_config[mnode].exists) { 6161772Sjl139090 start = mem_node_config[mnode].physbase; 6171772Sjl139090 end = mem_node_config[mnode].physmax; 6181772Sjl139090 mem_node_pre_del_slice(start, end); 6191772Sjl139090 mem_node_post_del_slice(start, end, 0); 6201772Sjl139090 } 6211772Sjl139090 } 6221772Sjl139090 6231772Sjl139090 break; 6241772Sjl139090 6251772Sjl139090 case LGRP_CONFIG_MEM_RENAME: 6261772Sjl139090 /* 6271772Sjl139090 * During a DR copy-rename operation, all of the memory 6281772Sjl139090 * on one board is moved to another board -- but the 6291772Sjl139090 * addresses/pfns and memnodes don't change. This means 6301772Sjl139090 * the memory has changed locations without changing identity. 6311772Sjl139090 * 6321772Sjl139090 * Source is where we are copying from and target is where we 6331772Sjl139090 * are copying to. After source memnode is copied to target 6341772Sjl139090 * memnode, the physical addresses of the target memnode are 6351772Sjl139090 * renamed to match what the source memnode had. Then target 6361772Sjl139090 * memnode can be removed and source memnode can take its 6371772Sjl139090 * place. 6381772Sjl139090 * 6391772Sjl139090 * To do this, swap the lgroup handle to memnode mappings for 6401772Sjl139090 * the boards, so target lgroup will have source memnode and 6411772Sjl139090 * source lgroup will have empty target memnode which is where 6421772Sjl139090 * its memory will go (if any is added to it later). 6431772Sjl139090 * 6441772Sjl139090 * Then source memnode needs to be removed from its lgroup 6451772Sjl139090 * and added to the target lgroup where the memory was living 6461772Sjl139090 * but under a different name/memnode. The memory was in the 6471772Sjl139090 * target memnode and now lives in the source memnode with 6481772Sjl139090 * different physical addresses even though it is the same 6491772Sjl139090 * memory. 6501772Sjl139090 */ 6511772Sjl139090 sbd = arg & 0xffff; 6521772Sjl139090 tbd = (arg & 0xffff0000) >> 16; 6531772Sjl139090 shand = sbd; 6541772Sjl139090 thand = tbd; 6551772Sjl139090 snode = plat_lgrphand_to_mem_node(shand); 6561772Sjl139090 tnode = plat_lgrphand_to_mem_node(thand); 6571772Sjl139090 6581772Sjl139090 /* 6591772Sjl139090 * Special handling for possible memory holes. 6601772Sjl139090 */ 6611772Sjl139090 if (tnode != -1 && mem_node_config[tnode].exists) { 6623354Sjl139090 start = mem_node_config[tnode].physbase; 6633354Sjl139090 end = mem_node_config[tnode].physmax; 6641772Sjl139090 mem_node_pre_del_slice(start, end); 6651772Sjl139090 mem_node_post_del_slice(start, end, 0); 6661772Sjl139090 } 6671772Sjl139090 6681772Sjl139090 plat_assign_lgrphand_to_mem_node(thand, snode); 6691772Sjl139090 plat_assign_lgrphand_to_mem_node(shand, tnode); 6701772Sjl139090 6711772Sjl139090 lmr.lmem_rename_from = shand; 6721772Sjl139090 lmr.lmem_rename_to = thand; 6731772Sjl139090 6741772Sjl139090 /* 6751772Sjl139090 * Remove source memnode of copy rename from its lgroup 6761772Sjl139090 * and add it to its new target lgroup 6771772Sjl139090 */ 6781772Sjl139090 lgrp_config(LGRP_CONFIG_MEM_RENAME, (uintptr_t)snode, 6791772Sjl139090 (uintptr_t)&lmr); 6801772Sjl139090 6811772Sjl139090 break; 6821772Sjl139090 6831772Sjl139090 default: 6841772Sjl139090 break; 6851772Sjl139090 } 6861772Sjl139090 } 6871772Sjl139090 6881772Sjl139090 /* 6891772Sjl139090 * Return latency between "from" and "to" lgroups 6901772Sjl139090 * 6911772Sjl139090 * This latency number can only be used for relative comparison 6921772Sjl139090 * between lgroups on the running system, cannot be used across platforms, 6931772Sjl139090 * and may not reflect the actual latency. It is platform and implementation 6941772Sjl139090 * specific, so platform gets to decide its value. It would be nice if the 6951772Sjl139090 * number was at least proportional to make comparisons more meaningful though. 6961772Sjl139090 * NOTE: The numbers below are supposed to be load latencies for uncached 6971772Sjl139090 * memory divided by 10. 6981772Sjl139090 * 6991772Sjl139090 */ 7001772Sjl139090 int 7011772Sjl139090 plat_lgrp_latency(lgrp_handle_t from, lgrp_handle_t to) 7021772Sjl139090 { 7031772Sjl139090 /* 7041772Sjl139090 * Return min remote latency when there are more than two lgroups 7051772Sjl139090 * (root and child) and getting latency between two different lgroups 7061772Sjl139090 * or root is involved 7071772Sjl139090 */ 7081772Sjl139090 if (lgrp_optimizations() && (from != to || 7091772Sjl139090 from == LGRP_DEFAULT_HANDLE || to == LGRP_DEFAULT_HANDLE)) 7102491Shyw return (42); 7111772Sjl139090 else 7122491Shyw return (35); 7131772Sjl139090 } 7141772Sjl139090 7151772Sjl139090 /* 7161772Sjl139090 * Return platform handle for root lgroup 7171772Sjl139090 */ 7181772Sjl139090 lgrp_handle_t 7191772Sjl139090 plat_lgrp_root_hand(void) 7201772Sjl139090 { 7211772Sjl139090 if (mpo_disabled) 7221772Sjl139090 return (lgrp_default_handle); 7231772Sjl139090 7241772Sjl139090 return (LGRP_DEFAULT_HANDLE); 7251772Sjl139090 } 7261772Sjl139090 7271772Sjl139090 /*ARGSUSED*/ 7281772Sjl139090 void 7291772Sjl139090 plat_freelist_process(int mnode) 7301772Sjl139090 { 7311772Sjl139090 } 7321772Sjl139090 7331772Sjl139090 void 7341772Sjl139090 load_platform_drivers(void) 7351772Sjl139090 { 7361772Sjl139090 (void) i_ddi_attach_pseudo_node("dr"); 7371772Sjl139090 } 7381772Sjl139090 7391772Sjl139090 /* 7401772Sjl139090 * No platform drivers on this platform 7411772Sjl139090 */ 7421772Sjl139090 char *platform_module_list[] = { 7431772Sjl139090 (char *)0 7441772Sjl139090 }; 7451772Sjl139090 7461772Sjl139090 /*ARGSUSED*/ 7471772Sjl139090 void 7481772Sjl139090 plat_tod_fault(enum tod_fault_type tod_bad) 7491772Sjl139090 { 7501772Sjl139090 } 7511772Sjl139090 7521772Sjl139090 /*ARGSUSED*/ 7531772Sjl139090 void 7541772Sjl139090 cpu_sgn_update(ushort_t sgn, uchar_t state, uchar_t sub_state, int cpuid) 7551772Sjl139090 { 7561772Sjl139090 static void (*scf_panic_callback)(int); 7571772Sjl139090 static void (*scf_shutdown_callback)(int); 7581772Sjl139090 7591772Sjl139090 /* 7601772Sjl139090 * This is for notifing system panic/shutdown to SCF. 7611772Sjl139090 * In case of shutdown and panic, SCF call back 7621772Sjl139090 * function should be called. 7631772Sjl139090 * <SCF call back functions> 7641772Sjl139090 * scf_panic_callb() : panicsys()->panic_quiesce_hw() 7651772Sjl139090 * scf_shutdown_callb(): halt() or power_down() or reboot_machine() 7661772Sjl139090 * cpuid should be -1 and state should be SIGST_EXIT. 7671772Sjl139090 */ 7681772Sjl139090 if (state == SIGST_EXIT && cpuid == -1) { 7691772Sjl139090 7701772Sjl139090 /* 7711772Sjl139090 * find the symbol for the SCF panic callback routine in driver 7721772Sjl139090 */ 7731772Sjl139090 if (scf_panic_callback == NULL) 7741772Sjl139090 scf_panic_callback = (void (*)(int)) 7755037Sjl139090 modgetsymvalue("scf_panic_callb", 0); 7761772Sjl139090 if (scf_shutdown_callback == NULL) 7771772Sjl139090 scf_shutdown_callback = (void (*)(int)) 7785037Sjl139090 modgetsymvalue("scf_shutdown_callb", 0); 7791772Sjl139090 7801772Sjl139090 switch (sub_state) { 7811772Sjl139090 case SIGSUBST_PANIC: 7821772Sjl139090 if (scf_panic_callback == NULL) { 7831772Sjl139090 cmn_err(CE_NOTE, "!cpu_sgn_update: " 7841772Sjl139090 "scf_panic_callb not found\n"); 7851772Sjl139090 return; 7861772Sjl139090 } 7871772Sjl139090 scf_panic_callback(SIGSUBST_PANIC); 7881772Sjl139090 break; 7891772Sjl139090 7901772Sjl139090 case SIGSUBST_HALT: 7911772Sjl139090 if (scf_shutdown_callback == NULL) { 7921772Sjl139090 cmn_err(CE_NOTE, "!cpu_sgn_update: " 7931772Sjl139090 "scf_shutdown_callb not found\n"); 7941772Sjl139090 return; 7951772Sjl139090 } 7961772Sjl139090 scf_shutdown_callback(SIGSUBST_HALT); 7971772Sjl139090 break; 7981772Sjl139090 7991772Sjl139090 case SIGSUBST_ENVIRON: 8001772Sjl139090 if (scf_shutdown_callback == NULL) { 8011772Sjl139090 cmn_err(CE_NOTE, "!cpu_sgn_update: " 8021772Sjl139090 "scf_shutdown_callb not found\n"); 8031772Sjl139090 return; 8041772Sjl139090 } 8051772Sjl139090 scf_shutdown_callback(SIGSUBST_ENVIRON); 8061772Sjl139090 break; 8071772Sjl139090 8081772Sjl139090 case SIGSUBST_REBOOT: 8091772Sjl139090 if (scf_shutdown_callback == NULL) { 8101772Sjl139090 cmn_err(CE_NOTE, "!cpu_sgn_update: " 8111772Sjl139090 "scf_shutdown_callb not found\n"); 8121772Sjl139090 return; 8131772Sjl139090 } 8141772Sjl139090 scf_shutdown_callback(SIGSUBST_REBOOT); 8151772Sjl139090 break; 8161772Sjl139090 } 8171772Sjl139090 } 8181772Sjl139090 } 8191772Sjl139090 8201772Sjl139090 /*ARGSUSED*/ 8211772Sjl139090 int 8221772Sjl139090 plat_get_mem_unum(int synd_code, uint64_t flt_addr, int flt_bus_id, 8231772Sjl139090 int flt_in_memory, ushort_t flt_status, 8241772Sjl139090 char *buf, int buflen, int *lenp) 8251772Sjl139090 { 8261772Sjl139090 /* 8271772Sjl139090 * check if it's a Memory error. 8281772Sjl139090 */ 8291772Sjl139090 if (flt_in_memory) { 8301772Sjl139090 if (opl_get_mem_unum != NULL) { 8315037Sjl139090 return (opl_get_mem_unum(synd_code, flt_addr, buf, 8325037Sjl139090 buflen, lenp)); 8331772Sjl139090 } else { 8341772Sjl139090 return (ENOTSUP); 8351772Sjl139090 } 8361772Sjl139090 } else { 8371772Sjl139090 return (ENOTSUP); 8381772Sjl139090 } 8391772Sjl139090 } 8401772Sjl139090 8411772Sjl139090 /*ARGSUSED*/ 8421772Sjl139090 int 8431772Sjl139090 plat_get_cpu_unum(int cpuid, char *buf, int buflen, int *lenp) 8441772Sjl139090 { 8452214Sav145390 int ret = 0; 8465347Sjfrank int sb; 8473123Ssubhan int plen; 8481772Sjl139090 8491772Sjl139090 sb = opl_get_physical_board(LSB_ID(cpuid)); 8501772Sjl139090 if (sb == -1) { 8511772Sjl139090 return (ENXIO); 8521772Sjl139090 } 8531772Sjl139090 8543627Ssubhan /* 8553627Ssubhan * opl_cur_model is assigned here 8563627Ssubhan */ 8573627Ssubhan if (opl_cur_model == NULL) { 8583627Ssubhan set_model_info(); 8595539Swh31274 8605539Swh31274 /* 8615539Swh31274 * if not matched, return 8625539Swh31274 */ 8635539Swh31274 if (opl_cur_model == NULL) 8645539Swh31274 return (ENODEV); 8653627Ssubhan } 8663627Ssubhan 8673123Ssubhan ASSERT((opl_cur_model - opl_models) == (opl_cur_model->model_type)); 8683123Ssubhan 8693123Ssubhan switch (opl_cur_model->model_type) { 8703123Ssubhan case FF1: 8712214Sav145390 plen = snprintf(buf, buflen, "/%s/CPUM%d", "MBU_A", 8722214Sav145390 CHIP_ID(cpuid) / 2); 8732214Sav145390 break; 8742214Sav145390 8753123Ssubhan case FF2: 8762214Sav145390 plen = snprintf(buf, buflen, "/%s/CPUM%d", "MBU_B", 8772808Sav145390 (CHIP_ID(cpuid) / 2) + (sb * 2)); 8782214Sav145390 break; 8792214Sav145390 8803123Ssubhan case DC1: 8813123Ssubhan case DC2: 8823123Ssubhan case DC3: 8832214Sav145390 plen = snprintf(buf, buflen, "/%s%02d/CPUM%d", "CMU", sb, 8842214Sav145390 CHIP_ID(cpuid)); 8852214Sav145390 break; 8862214Sav145390 8872214Sav145390 default: 8882214Sav145390 /* This should never happen */ 8892214Sav145390 return (ENODEV); 8902214Sav145390 } 8912214Sav145390 8922214Sav145390 if (plen >= buflen) { 8932214Sav145390 ret = ENOSPC; 8941772Sjl139090 } else { 8951772Sjl139090 if (lenp) 8961772Sjl139090 *lenp = strlen(buf); 8971772Sjl139090 } 8982214Sav145390 return (ret); 8991772Sjl139090 } 9001772Sjl139090 9011772Sjl139090 void 9021772Sjl139090 plat_nodename_set(void) 9031772Sjl139090 { 9045347Sjfrank post_xscf_msg((char *)&utsname, sizeof (struct utsname)); 9051772Sjl139090 } 9061772Sjl139090 9071772Sjl139090 caddr_t efcode_vaddr = NULL; 9081772Sjl139090 9091772Sjl139090 /* 9101772Sjl139090 * Preallocate enough memory for fcode claims. 9111772Sjl139090 */ 9121772Sjl139090 9131772Sjl139090 caddr_t 9141772Sjl139090 efcode_alloc(caddr_t alloc_base) 9151772Sjl139090 { 9161772Sjl139090 caddr_t efcode_alloc_base = (caddr_t)roundup((uintptr_t)alloc_base, 9171772Sjl139090 MMU_PAGESIZE); 9181772Sjl139090 caddr_t vaddr; 9191772Sjl139090 9201772Sjl139090 /* 9211772Sjl139090 * allocate the physical memory for the Oberon fcode. 9221772Sjl139090 */ 9231772Sjl139090 if ((vaddr = (caddr_t)BOP_ALLOC(bootops, efcode_alloc_base, 9241772Sjl139090 efcode_size, MMU_PAGESIZE)) == NULL) 9251772Sjl139090 cmn_err(CE_PANIC, "Cannot allocate Efcode Memory"); 9261772Sjl139090 9271772Sjl139090 efcode_vaddr = vaddr; 9281772Sjl139090 9291772Sjl139090 return (efcode_alloc_base + efcode_size); 9301772Sjl139090 } 9311772Sjl139090 9321772Sjl139090 caddr_t 9331772Sjl139090 plat_startup_memlist(caddr_t alloc_base) 9341772Sjl139090 { 9351772Sjl139090 caddr_t tmp_alloc_base; 9361772Sjl139090 9371772Sjl139090 tmp_alloc_base = efcode_alloc(alloc_base); 9381772Sjl139090 tmp_alloc_base = 9391772Sjl139090 (caddr_t)roundup((uintptr_t)tmp_alloc_base, ecache_alignsize); 9401772Sjl139090 return (tmp_alloc_base); 9411772Sjl139090 } 9421772Sjl139090 9431772Sjl139090 void 9441772Sjl139090 startup_platform(void) 9451772Sjl139090 { 9461772Sjl139090 } 9472214Sav145390 9482241Shuah void 9492241Shuah plat_cpuid_to_mmu_ctx_info(processorid_t cpuid, mmu_ctx_info_t *info) 9502241Shuah { 9512241Shuah int impl; 9522241Shuah 9532241Shuah impl = cpunodes[cpuid].implementation; 9545037Sjl139090 if (IS_OLYMPUS_C(impl) || IS_JUPITER(impl)) { 9552457Smv143129 info->mmu_idx = MMU_ID(cpuid); 9562241Shuah info->mmu_nctxs = 8192; 9572241Shuah } else { 9582241Shuah cmn_err(CE_PANIC, "Unknown processor %d", impl); 9592241Shuah } 9602241Shuah } 9612241Shuah 9622214Sav145390 int 9632214Sav145390 plat_get_mem_sid(char *unum, char *buf, int buflen, int *lenp) 9642214Sav145390 { 9652214Sav145390 if (opl_get_mem_sid == NULL) { 9662214Sav145390 return (ENOTSUP); 9672214Sav145390 } 9682214Sav145390 return (opl_get_mem_sid(unum, buf, buflen, lenp)); 9692214Sav145390 } 9702214Sav145390 9712214Sav145390 int 9722214Sav145390 plat_get_mem_offset(uint64_t paddr, uint64_t *offp) 9732214Sav145390 { 9742214Sav145390 if (opl_get_mem_offset == NULL) { 9752214Sav145390 return (ENOTSUP); 9762214Sav145390 } 9772214Sav145390 return (opl_get_mem_offset(paddr, offp)); 9782214Sav145390 } 9792214Sav145390 9802214Sav145390 int 9812214Sav145390 plat_get_mem_addr(char *unum, char *sid, uint64_t offset, uint64_t *addrp) 9822214Sav145390 { 9832214Sav145390 if (opl_get_mem_addr == NULL) { 9842214Sav145390 return (ENOTSUP); 9852214Sav145390 } 9862214Sav145390 return (opl_get_mem_addr(unum, sid, offset, addrp)); 9872214Sav145390 } 9883914Spm145316 9893914Spm145316 void 9903914Spm145316 plat_lock_delay(int *backoff) 9913914Spm145316 { 9923914Spm145316 int i; 9933914Spm145316 int cnt; 9943914Spm145316 int flag; 9953914Spm145316 int ctr; 9963914Spm145316 hrtime_t delay_start; 9973914Spm145316 /* 9983914Spm145316 * Platform specific lock delay code for OPL 9993914Spm145316 * 10003914Spm145316 * Using staged linear increases in the delay. 10013914Spm145316 * The sleep instruction is the preferred method of delay, 10023914Spm145316 * but is too large of granularity for the initial backoff. 10033914Spm145316 */ 10043914Spm145316 10053914Spm145316 if (*backoff == 0) *backoff = OPL_BOFF_BASE; 10063914Spm145316 10073914Spm145316 flag = !*backoff; 10083914Spm145316 10093914Spm145316 if (*backoff < OPL_BOFF_CAP1) { 10103914Spm145316 /* 10113914Spm145316 * If desired backoff is long enough, 10123914Spm145316 * use sleep for most of it 10133914Spm145316 */ 10145037Sjl139090 for (cnt = *backoff; cnt >= OPL_BOFF_SLEEP; 10155037Sjl139090 cnt -= OPL_BOFF_SLEEP) { 10163914Spm145316 cpu_smt_pause(); 10173914Spm145316 } 10183914Spm145316 /* 10193914Spm145316 * spin for small remainder of backoff 10203914Spm145316 * 10213914Spm145316 * fake call to nulldev included to prevent 10223914Spm145316 * compiler from optimizing out the spin loop 10233914Spm145316 */ 10243914Spm145316 for (ctr = cnt * OPL_BOFF_SPIN; ctr; ctr--) { 10253914Spm145316 if (flag) (void) nulldev(); 10263914Spm145316 } 10273914Spm145316 } else { 10283914Spm145316 /* backoff is very large. Fill it by sleeping */ 10293914Spm145316 delay_start = gethrtime(); 10303914Spm145316 cnt = *backoff/OPL_BOFF_SLEEP; 10313914Spm145316 /* 10323914Spm145316 * use sleep instructions for delay 10333914Spm145316 */ 10343914Spm145316 for (i = 0; i < cnt; i++) { 10353914Spm145316 cpu_smt_pause(); 10363914Spm145316 } 10373914Spm145316 10383914Spm145316 /* 10393914Spm145316 * Note: if the other strand executes a sleep instruction, 10403914Spm145316 * then the sleep ends immediately with a minimum time of 10413914Spm145316 * 42 clocks. We check gethrtime to insure we have 10423914Spm145316 * waited long enough. And we include both a short 10433914Spm145316 * spin loop and a sleep for any final delay time. 10443914Spm145316 */ 10453914Spm145316 10463914Spm145316 while ((gethrtime() - delay_start) < cnt * OPL_BOFF_TM) { 10473914Spm145316 cpu_smt_pause(); 10483914Spm145316 for (ctr = OPL_BOFF_SPIN; ctr; ctr--) { 10493914Spm145316 if (flag) (void) nulldev(); 10503914Spm145316 } 10513914Spm145316 } 10523914Spm145316 } 10533914Spm145316 10543914Spm145316 /* 10553914Spm145316 * We adjust the backoff in three linear stages 10563914Spm145316 * The initial stage has small increases as this phase is 10573914Spm145316 * usually handle locks with light contention. We don't want 10583914Spm145316 * to have a long backoff on a lock that is available. 10593914Spm145316 * 10603914Spm145316 * In the second stage, we are in transition, unsure whether 10613914Spm145316 * the lock is under heavy contention. As the failures to 10623914Spm145316 * obtain the lock increase, we back off further. 10633914Spm145316 * 10643914Spm145316 * For the final stage, we are in a heavily contended or 10653914Spm145316 * long held long so we want to reduce the number of tries. 10663914Spm145316 */ 10673914Spm145316 if (*backoff < OPL_BOFF_CAP1) { 10683914Spm145316 *backoff += 1; 10693914Spm145316 } else { 10703914Spm145316 if (*backoff < OPL_BOFF_CAP2) { 10713914Spm145316 *backoff += OPL_BOFF_SLEEP; 10723914Spm145316 } else { 10733914Spm145316 *backoff += 2 * OPL_BOFF_SLEEP; 10743914Spm145316 } 10753914Spm145316 if (*backoff > OPL_BOFF_MAX) { 10763914Spm145316 *backoff = OPL_BOFF_MAX; 10773914Spm145316 } 10783914Spm145316 } 10793914Spm145316 } 10805347Sjfrank 10815347Sjfrank /* 10825347Sjfrank * The following code implements asynchronous call to XSCF to setup the 10835347Sjfrank * domain node name. 10845347Sjfrank */ 10855347Sjfrank 10865347Sjfrank #define FREE_MSG(m) kmem_free((m), NM_LEN((m)->len)) 10875347Sjfrank 10885347Sjfrank /* 10895347Sjfrank * The following three macros define the all operations on the request 10905347Sjfrank * list we are using here, and hide the details of the list 10915347Sjfrank * implementation from the code. 10925347Sjfrank */ 10935347Sjfrank #define PUSH(m) \ 10945347Sjfrank { \ 10955347Sjfrank (m)->next = ctl_msg.head; \ 10965347Sjfrank (m)->prev = NULL; \ 10975347Sjfrank if ((m)->next != NULL) \ 10985347Sjfrank (m)->next->prev = (m); \ 10995347Sjfrank ctl_msg.head = (m); \ 11005347Sjfrank } 11015347Sjfrank 11025347Sjfrank #define REMOVE(m) \ 11035347Sjfrank { \ 11045347Sjfrank if ((m)->prev != NULL) \ 11055347Sjfrank (m)->prev->next = (m)->next; \ 11065347Sjfrank else \ 11075347Sjfrank ctl_msg.head = (m)->next; \ 11085347Sjfrank if ((m)->next != NULL) \ 11095347Sjfrank (m)->next->prev = (m)->prev; \ 11105347Sjfrank } 11115347Sjfrank 11125347Sjfrank #define FREE_THE_TAIL(head) \ 11135347Sjfrank { \ 11145347Sjfrank nm_msg_t *n_msg, *m; \ 11155347Sjfrank m = (head)->next; \ 11165347Sjfrank (head)->next = NULL; \ 11175347Sjfrank while (m != NULL) { \ 11185347Sjfrank n_msg = m->next; \ 11195347Sjfrank FREE_MSG(m); \ 11205347Sjfrank m = n_msg; \ 11215347Sjfrank } \ 11225347Sjfrank } 11235347Sjfrank 11245347Sjfrank #define SCF_PUTINFO(f, s, p) \ 11255347Sjfrank f(KEY_ESCF, 0x01, 0, s, p) 11265347Sjfrank 11275347Sjfrank #define PASS2XSCF(m, r) ((r = SCF_PUTINFO(ctl_msg.scf_service_function, \ 11285347Sjfrank (m)->len, (m)->data)) == 0) 11295347Sjfrank 11305347Sjfrank /* 11315347Sjfrank * The value of the following macro loosely depends on the 11325347Sjfrank * value of the "device busy" timeout used in the SCF driver. 11335347Sjfrank * (See pass2xscf_thread()). 11345347Sjfrank */ 11355347Sjfrank #define SCF_DEVBUSY_DELAY 10 11365347Sjfrank 11375347Sjfrank /* 11385347Sjfrank * The default number of attempts to contact the scf driver 11395347Sjfrank * if we cannot fetch any information about the timeout value 11405347Sjfrank * it uses. 11415347Sjfrank */ 11425347Sjfrank 11435347Sjfrank #define REPEATS 4 11445347Sjfrank 11455347Sjfrank typedef struct nm_msg { 11465347Sjfrank struct nm_msg *next; 11475347Sjfrank struct nm_msg *prev; 11485347Sjfrank int len; 11495347Sjfrank char data[1]; 11505347Sjfrank } nm_msg_t; 11515347Sjfrank 11525347Sjfrank #define NM_LEN(len) (sizeof (nm_msg_t) + (len) - 1) 11535347Sjfrank 11545347Sjfrank static struct ctlmsg { 11555347Sjfrank nm_msg_t *head; 11565347Sjfrank nm_msg_t *now_serving; 11575347Sjfrank kmutex_t nm_lock; 11585347Sjfrank kthread_t *nmt; 11595347Sjfrank int cnt; 11605347Sjfrank int (*scf_service_function)(uint32_t, uint8_t, 11615347Sjfrank uint32_t, uint32_t, void *); 11625347Sjfrank } ctl_msg; 11635347Sjfrank 11645347Sjfrank static void 11655347Sjfrank post_xscf_msg(char *dp, int len) 11665347Sjfrank { 11675347Sjfrank nm_msg_t *msg; 11685347Sjfrank 11695347Sjfrank msg = (nm_msg_t *)kmem_zalloc(NM_LEN(len), KM_SLEEP); 11705347Sjfrank 11715347Sjfrank bcopy(dp, msg->data, len); 11725347Sjfrank msg->len = len; 11735347Sjfrank 11745347Sjfrank mutex_enter(&ctl_msg.nm_lock); 11755347Sjfrank if (ctl_msg.nmt == NULL) { 11765347Sjfrank ctl_msg.nmt = thread_create(NULL, 0, pass2xscf_thread, 11775347Sjfrank NULL, 0, &p0, TS_RUN, minclsyspri); 11785347Sjfrank } 11795347Sjfrank 11805347Sjfrank PUSH(msg); 11815347Sjfrank ctl_msg.cnt++; 11825347Sjfrank mutex_exit(&ctl_msg.nm_lock); 11835347Sjfrank } 11845347Sjfrank 11855347Sjfrank static void 11865347Sjfrank pass2xscf_thread() 11875347Sjfrank { 11885347Sjfrank nm_msg_t *msg; 11895347Sjfrank int ret; 11905347Sjfrank uint_t i, msg_sent, xscf_driver_delay; 11915347Sjfrank static uint_t repeat_cnt; 11925347Sjfrank uint_t *scf_wait_cnt; 11935347Sjfrank 11945347Sjfrank mutex_enter(&ctl_msg.nm_lock); 11955347Sjfrank 11965347Sjfrank /* 11975347Sjfrank * Find the address of the SCF put routine if it's not done yet. 11985347Sjfrank */ 11995347Sjfrank if (ctl_msg.scf_service_function == NULL) { 12005347Sjfrank if ((ctl_msg.scf_service_function = 12015347Sjfrank (int (*)(uint32_t, uint8_t, uint32_t, uint32_t, void *)) 12025347Sjfrank modgetsymvalue("scf_service_putinfo", 0)) == NULL) { 12035347Sjfrank cmn_err(CE_NOTE, "pass2xscf_thread: " 12045347Sjfrank "scf_service_putinfo not found\n"); 12055347Sjfrank ctl_msg.nmt = NULL; 12065347Sjfrank mutex_exit(&ctl_msg.nm_lock); 12075347Sjfrank return; 12085347Sjfrank } 12095347Sjfrank } 12105347Sjfrank 12115347Sjfrank /* 12125347Sjfrank * Calculate the number of attempts to connect XSCF based on the 12135347Sjfrank * scf driver delay (which is 12145347Sjfrank * SCF_DEVBUSY_DELAY*scf_online_wait_rcnt seconds) and the value 12155347Sjfrank * of xscf_connect_delay (the total number of seconds to wait 12165347Sjfrank * till xscf get ready.) 12175347Sjfrank */ 12185347Sjfrank if (repeat_cnt == 0) { 12195347Sjfrank if ((scf_wait_cnt = 12205347Sjfrank (uint_t *) 12215347Sjfrank modgetsymvalue("scf_online_wait_rcnt", 0)) == NULL) { 12225347Sjfrank repeat_cnt = REPEATS; 12235347Sjfrank } else { 12245347Sjfrank 12255347Sjfrank xscf_driver_delay = *scf_wait_cnt * 12265347Sjfrank SCF_DEVBUSY_DELAY; 12275347Sjfrank repeat_cnt = (xscf_connect_delay/xscf_driver_delay) + 1; 12285347Sjfrank } 12295347Sjfrank } 12305347Sjfrank 12315347Sjfrank while (ctl_msg.cnt != 0) { 12325347Sjfrank 12335347Sjfrank /* 12345347Sjfrank * Take the very last request from the queue, 12355347Sjfrank */ 12365347Sjfrank ctl_msg.now_serving = ctl_msg.head; 12375347Sjfrank ASSERT(ctl_msg.now_serving != NULL); 12385347Sjfrank 12395347Sjfrank /* 12405347Sjfrank * and discard all the others if any. 12415347Sjfrank */ 12425347Sjfrank FREE_THE_TAIL(ctl_msg.now_serving); 12435347Sjfrank ctl_msg.cnt = 1; 12445347Sjfrank mutex_exit(&ctl_msg.nm_lock); 12455347Sjfrank 12465347Sjfrank /* 12475347Sjfrank * Pass the name to XSCF. Note please, we do not hold the 12485347Sjfrank * mutex while we are doing this. 12495347Sjfrank */ 12505347Sjfrank msg_sent = 0; 12515347Sjfrank for (i = 0; i < repeat_cnt; i++) { 12525347Sjfrank if (PASS2XSCF(ctl_msg.now_serving, ret)) { 12535347Sjfrank msg_sent = 1; 12545347Sjfrank break; 12555347Sjfrank } else { 12565347Sjfrank if (ret != EBUSY) { 12575347Sjfrank cmn_err(CE_NOTE, "pass2xscf_thread:" 12585347Sjfrank " unexpected return code" 12595347Sjfrank " from scf_service_putinfo():" 12605347Sjfrank " %d\n", ret); 12615347Sjfrank } 12625347Sjfrank } 12635347Sjfrank } 12645347Sjfrank 12655347Sjfrank if (msg_sent) { 12665347Sjfrank 12675347Sjfrank /* 12685347Sjfrank * Remove the request from the list 12695347Sjfrank */ 12705347Sjfrank mutex_enter(&ctl_msg.nm_lock); 12715347Sjfrank msg = ctl_msg.now_serving; 12725347Sjfrank ctl_msg.now_serving = NULL; 12735347Sjfrank REMOVE(msg); 12745347Sjfrank ctl_msg.cnt--; 12755347Sjfrank mutex_exit(&ctl_msg.nm_lock); 12765347Sjfrank FREE_MSG(msg); 12775347Sjfrank } else { 12785347Sjfrank 12795347Sjfrank /* 12805347Sjfrank * If while we have tried to communicate with 12815347Sjfrank * XSCF there were any other requests we are 12825347Sjfrank * going to drop this one and take the latest 12835347Sjfrank * one. Otherwise we will try to pass this one 12845347Sjfrank * again. 12855347Sjfrank */ 12865347Sjfrank cmn_err(CE_NOTE, 12875347Sjfrank "pass2xscf_thread: " 12885347Sjfrank "scf_service_putinfo " 12895347Sjfrank "not responding\n"); 12905347Sjfrank } 12915347Sjfrank mutex_enter(&ctl_msg.nm_lock); 12925347Sjfrank } 12935347Sjfrank 12945347Sjfrank /* 12955347Sjfrank * The request queue is empty, exit. 12965347Sjfrank */ 12975347Sjfrank ctl_msg.nmt = NULL; 12985347Sjfrank mutex_exit(&ctl_msg.nm_lock); 12995347Sjfrank } 1300