11772Sjl139090 /* 21772Sjl139090 * CDDL HEADER START 31772Sjl139090 * 41772Sjl139090 * The contents of this file are subject to the terms of the 51772Sjl139090 * Common Development and Distribution License (the "License"). 61772Sjl139090 * You may not use this file except in compliance with the License. 71772Sjl139090 * 81772Sjl139090 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 91772Sjl139090 * or http://www.opensolaris.org/os/licensing. 101772Sjl139090 * See the License for the specific language governing permissions 111772Sjl139090 * and limitations under the License. 121772Sjl139090 * 131772Sjl139090 * When distributing Covered Code, include this CDDL HEADER in each 141772Sjl139090 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 151772Sjl139090 * If applicable, add the following below this CDDL HEADER, with the 161772Sjl139090 * fields enclosed by brackets "[]" replaced with your own identifying 171772Sjl139090 * information: Portions Copyright [yyyy] [name of copyright owner] 181772Sjl139090 * 191772Sjl139090 * CDDL HEADER END 201772Sjl139090 */ 211772Sjl139090 /* 225788Smv143129 * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 231772Sjl139090 * Use is subject to license terms. 241772Sjl139090 */ 251772Sjl139090 261772Sjl139090 #pragma ident "%Z%%M% %I% %E% SMI" 271772Sjl139090 281772Sjl139090 #include <sys/cpuvar.h> 291772Sjl139090 #include <sys/systm.h> 301772Sjl139090 #include <sys/sysmacros.h> 311772Sjl139090 #include <sys/promif.h> 321772Sjl139090 #include <sys/platform_module.h> 331772Sjl139090 #include <sys/cmn_err.h> 341772Sjl139090 #include <sys/errno.h> 351772Sjl139090 #include <sys/machsystm.h> 361772Sjl139090 #include <sys/bootconf.h> 371772Sjl139090 #include <sys/nvpair.h> 381772Sjl139090 #include <sys/kobj.h> 391772Sjl139090 #include <sys/mem_cage.h> 401772Sjl139090 #include <sys/opl.h> 411772Sjl139090 #include <sys/scfd/scfostoescf.h> 421772Sjl139090 #include <sys/cpu_sgnblk_defs.h> 431772Sjl139090 #include <sys/utsname.h> 441772Sjl139090 #include <sys/ddi.h> 451772Sjl139090 #include <sys/sunndi.h> 461772Sjl139090 #include <sys/lgrp.h> 471772Sjl139090 #include <sys/memnode.h> 481772Sjl139090 #include <sys/sysmacros.h> 493914Spm145316 #include <sys/time.h> 503914Spm145316 #include <sys/cpu.h> 511772Sjl139090 #include <vm/vm_dep.h> 521772Sjl139090 531772Sjl139090 int (*opl_get_mem_unum)(int, uint64_t, char *, int, int *); 542214Sav145390 int (*opl_get_mem_sid)(char *unum, char *buf, int buflen, int *lenp); 552214Sav145390 int (*opl_get_mem_offset)(uint64_t paddr, uint64_t *offp); 562214Sav145390 int (*opl_get_mem_addr)(char *unum, char *sid, 572214Sav145390 uint64_t offset, uint64_t *paddr); 581772Sjl139090 591772Sjl139090 /* Memory for fcode claims. 16k times # maximum possible IO units */ 601772Sjl139090 #define EFCODE_SIZE (OPL_MAX_BOARDS * OPL_MAX_IO_UNITS_PER_BOARD * 0x4000) 611772Sjl139090 int efcode_size = EFCODE_SIZE; 621772Sjl139090 631772Sjl139090 #define OPL_MC_MEMBOARD_SHIFT 38 /* Boards on 256BG boundary */ 641772Sjl139090 651772Sjl139090 /* Set the maximum number of boards for DR */ 661772Sjl139090 int opl_boards = OPL_MAX_BOARDS; 671772Sjl139090 681772Sjl139090 void sgn_update_all_cpus(ushort_t, uchar_t, uchar_t); 691772Sjl139090 701772Sjl139090 extern int tsb_lgrp_affinity; 711772Sjl139090 721772Sjl139090 int opl_tsb_spares = (OPL_MAX_BOARDS) * (OPL_MAX_PCICH_UNITS_PER_BOARD) * 731772Sjl139090 (OPL_MAX_TSBS_PER_PCICH); 741772Sjl139090 751772Sjl139090 pgcnt_t opl_startup_cage_size = 0; 761772Sjl139090 775347Sjfrank /* 785347Sjfrank * The length of the delay in seconds in communication with XSCF after 795347Sjfrank * which the warning message will be logged. 805347Sjfrank */ 815347Sjfrank uint_t xscf_connect_delay = 60 * 15; 825347Sjfrank 832241Shuah static opl_model_info_t opl_models[] = { 843123Ssubhan { "FF1", OPL_MAX_BOARDS_FF1, FF1, STD_DISPATCH_TABLE }, 853123Ssubhan { "FF2", OPL_MAX_BOARDS_FF2, FF2, STD_DISPATCH_TABLE }, 863123Ssubhan { "DC1", OPL_MAX_BOARDS_DC1, DC1, STD_DISPATCH_TABLE }, 873123Ssubhan { "DC2", OPL_MAX_BOARDS_DC2, DC2, EXT_DISPATCH_TABLE }, 883123Ssubhan { "DC3", OPL_MAX_BOARDS_DC3, DC3, EXT_DISPATCH_TABLE }, 892241Shuah }; 902241Shuah static int opl_num_models = sizeof (opl_models)/sizeof (opl_model_info_t); 912241Shuah 923123Ssubhan /* 933627Ssubhan * opl_cur_model 943123Ssubhan */ 953627Ssubhan static opl_model_info_t *opl_cur_model = NULL; 962241Shuah 971772Sjl139090 static struct memlist *opl_memlist_per_board(struct memlist *ml); 985347Sjfrank static void post_xscf_msg(char *, int); 995347Sjfrank static void pass2xscf_thread(); 1001772Sjl139090 1013914Spm145316 /* 1023914Spm145316 * Note FF/DC out-of-order instruction engine takes only a 1033914Spm145316 * single cycle to execute each spin loop 1043914Spm145316 * for comparison, Panther takes 6 cycles for same loop 1055834Spt157919 * OPL_BOFF_SPIN = base spin loop, roughly one memory reference time 1065834Spt157919 * OPL_BOFF_TM = approx nsec for OPL sleep instruction (1600 for OPL-C) 1075834Spt157919 * OPL_BOFF_SLEEP = approx number of SPIN iterations to equal one sleep 1085834Spt157919 * OPL_BOFF_MAX_SCALE - scaling factor for max backoff based on active cpus 1095834Spt157919 * Listed values tuned for 2.15GHz to 2.64GHz systems 1103914Spm145316 * Value may change for future systems 1113914Spm145316 */ 1125834Spt157919 #define OPL_BOFF_SPIN 7 1135834Spt157919 #define OPL_BOFF_SLEEP 4 1145834Spt157919 #define OPL_BOFF_TM 1600 1155834Spt157919 #define OPL_BOFF_MAX_SCALE 8 1163914Spm145316 1175788Smv143129 #define OPL_CLOCK_TICK_THRESHOLD 128 1185788Smv143129 #define OPL_CLOCK_TICK_NCPUS 64 1195788Smv143129 1205788Smv143129 extern int clock_tick_threshold; 1215788Smv143129 extern int clock_tick_ncpus; 1225788Smv143129 1231772Sjl139090 int 1241772Sjl139090 set_platform_max_ncpus(void) 1251772Sjl139090 { 1261772Sjl139090 return (OPL_MAX_CPU_PER_BOARD * OPL_MAX_BOARDS); 1271772Sjl139090 } 1281772Sjl139090 1291772Sjl139090 int 1301772Sjl139090 set_platform_tsb_spares(void) 1311772Sjl139090 { 1321772Sjl139090 return (MIN(opl_tsb_spares, MAX_UPA)); 1331772Sjl139090 } 1341772Sjl139090 1352241Shuah static void 1362241Shuah set_model_info() 1372241Shuah { 1383123Ssubhan extern int ts_dispatch_extended; 1392241Shuah char name[MAXSYSNAME]; 1402241Shuah int i; 1412241Shuah 1422241Shuah /* 1432241Shuah * Get model name from the root node. 1442241Shuah * 1452241Shuah * We are using the prom device tree since, at this point, 1462241Shuah * the Solaris device tree is not yet setup. 1472241Shuah */ 1482241Shuah (void) prom_getprop(prom_rootnode(), "model", (caddr_t)name); 1492241Shuah 1502241Shuah for (i = 0; i < opl_num_models; i++) { 1512241Shuah if (strncmp(name, opl_models[i].model_name, MAXSYSNAME) == 0) { 1522241Shuah opl_cur_model = &opl_models[i]; 1532241Shuah break; 1542241Shuah } 1552241Shuah } 1563123Ssubhan 1575539Swh31274 /* 1585539Swh31274 * If model not matched, it's an unknown model. 1595539Swh31274 * just return. 1605539Swh31274 */ 1612241Shuah if (i == opl_num_models) 1625539Swh31274 return; 1633123Ssubhan 1643123Ssubhan if ((opl_cur_model->model_cmds & EXT_DISPATCH_TABLE) && 1655037Sjl139090 (ts_dispatch_extended == -1)) { 1663123Ssubhan /* 1673123Ssubhan * Based on a platform model, select a dispatch table. 1683123Ssubhan * Only DC2 and DC3 systems uses the alternate/extended 1693123Ssubhan * TS dispatch table. 1703123Ssubhan * FF1, FF2 and DC1 systems used standard dispatch tables. 1713123Ssubhan */ 1723123Ssubhan ts_dispatch_extended = 1; 1733123Ssubhan } 1743123Ssubhan 1752241Shuah } 1762241Shuah 1772241Shuah static void 1782241Shuah set_max_mmu_ctxdoms() 1792241Shuah { 1802241Shuah extern uint_t max_mmu_ctxdoms; 1812241Shuah int max_boards; 1822241Shuah 1832241Shuah /* 1842241Shuah * From the model, get the maximum number of boards 1852241Shuah * supported and set the value accordingly. If the model 1862241Shuah * could not be determined or recognized, we assume the max value. 1872241Shuah */ 1882241Shuah if (opl_cur_model == NULL) 1892241Shuah max_boards = OPL_MAX_BOARDS; 1902241Shuah else 1912241Shuah max_boards = opl_cur_model->model_max_boards; 1922241Shuah 1932241Shuah /* 1942241Shuah * On OPL, cores and MMUs are one-to-one. 1952241Shuah */ 1962241Shuah max_mmu_ctxdoms = OPL_MAX_CORE_UNITS_PER_BOARD * max_boards; 1972241Shuah } 1982241Shuah 1991772Sjl139090 #pragma weak mmu_init_large_pages 2001772Sjl139090 2011772Sjl139090 void 2021772Sjl139090 set_platform_defaults(void) 2031772Sjl139090 { 2041772Sjl139090 extern char *tod_module_name; 2051772Sjl139090 extern void cpu_sgn_update(ushort_t, uchar_t, uchar_t, int); 2061772Sjl139090 extern void mmu_init_large_pages(size_t); 2071772Sjl139090 2081772Sjl139090 /* Set the CPU signature function pointer */ 2091772Sjl139090 cpu_sgn_func = cpu_sgn_update; 2101772Sjl139090 2111772Sjl139090 /* Set appropriate tod module for OPL platform */ 2121772Sjl139090 ASSERT(tod_module_name == NULL); 2131772Sjl139090 tod_module_name = "todopl"; 2141772Sjl139090 2151772Sjl139090 if ((mmu_page_sizes == max_mmu_page_sizes) && 2162659Ssusans (mmu_ism_pagesize != DEFAULT_ISM_PAGESIZE)) { 2171772Sjl139090 if (&mmu_init_large_pages) 2181772Sjl139090 mmu_init_large_pages(mmu_ism_pagesize); 2191772Sjl139090 } 2201772Sjl139090 2211772Sjl139090 tsb_lgrp_affinity = 1; 2222241Shuah 2232241Shuah set_max_mmu_ctxdoms(); 2241772Sjl139090 } 2251772Sjl139090 2261772Sjl139090 /* 2271772Sjl139090 * Convert logical a board number to a physical one. 2281772Sjl139090 */ 2291772Sjl139090 2301772Sjl139090 #define LSBPROP "board#" 2311772Sjl139090 #define PSBPROP "physical-board#" 2321772Sjl139090 2331772Sjl139090 int 2341772Sjl139090 opl_get_physical_board(int id) 2351772Sjl139090 { 2361772Sjl139090 dev_info_t *root_dip, *dip = NULL; 2371772Sjl139090 char *dname = NULL; 2381772Sjl139090 int circ; 2391772Sjl139090 2401772Sjl139090 pnode_t pnode; 2411772Sjl139090 char pname[MAXSYSNAME] = {0}; 2421772Sjl139090 2431772Sjl139090 int lsb_id; /* Logical System Board ID */ 2441772Sjl139090 int psb_id; /* Physical System Board ID */ 2451772Sjl139090 2461772Sjl139090 2471772Sjl139090 /* 2481772Sjl139090 * This function is called on early stage of bootup when the 2491772Sjl139090 * kernel device tree is not initialized yet, and also 2501772Sjl139090 * later on when the device tree is up. We want to try 2511772Sjl139090 * the fast track first. 2521772Sjl139090 */ 2531772Sjl139090 root_dip = ddi_root_node(); 2541772Sjl139090 if (root_dip) { 2551772Sjl139090 /* Get from devinfo node */ 2561772Sjl139090 ndi_devi_enter(root_dip, &circ); 2571772Sjl139090 for (dip = ddi_get_child(root_dip); dip; 2581772Sjl139090 dip = ddi_get_next_sibling(dip)) { 2591772Sjl139090 2601772Sjl139090 dname = ddi_node_name(dip); 2611772Sjl139090 if (strncmp(dname, "pseudo-mc", 9) != 0) 2621772Sjl139090 continue; 2631772Sjl139090 2641772Sjl139090 if ((lsb_id = (int)ddi_getprop(DDI_DEV_T_ANY, dip, 2651772Sjl139090 DDI_PROP_DONTPASS, LSBPROP, -1)) == -1) 2661772Sjl139090 continue; 2671772Sjl139090 2681772Sjl139090 if (id == lsb_id) { 2691772Sjl139090 if ((psb_id = (int)ddi_getprop(DDI_DEV_T_ANY, 2701772Sjl139090 dip, DDI_PROP_DONTPASS, PSBPROP, -1)) 2711772Sjl139090 == -1) { 2721772Sjl139090 ndi_devi_exit(root_dip, circ); 2731772Sjl139090 return (-1); 2741772Sjl139090 } else { 2751772Sjl139090 ndi_devi_exit(root_dip, circ); 2761772Sjl139090 return (psb_id); 2771772Sjl139090 } 2781772Sjl139090 } 2791772Sjl139090 } 2801772Sjl139090 ndi_devi_exit(root_dip, circ); 2811772Sjl139090 } 2821772Sjl139090 2831772Sjl139090 /* 2841772Sjl139090 * We do not have the kernel device tree, or we did not 2851772Sjl139090 * find the node for some reason (let's say the kernel 2861772Sjl139090 * device tree was modified), let's try the OBP tree. 2871772Sjl139090 */ 2881772Sjl139090 pnode = prom_rootnode(); 2891772Sjl139090 for (pnode = prom_childnode(pnode); pnode; 2901772Sjl139090 pnode = prom_nextnode(pnode)) { 2911772Sjl139090 2921772Sjl139090 if ((prom_getprop(pnode, "name", (caddr_t)pname) == -1) || 2931772Sjl139090 (strncmp(pname, "pseudo-mc", 9) != 0)) 2941772Sjl139090 continue; 2951772Sjl139090 2961772Sjl139090 if (prom_getprop(pnode, LSBPROP, (caddr_t)&lsb_id) == -1) 2971772Sjl139090 continue; 2981772Sjl139090 2991772Sjl139090 if (id == lsb_id) { 3001772Sjl139090 if (prom_getprop(pnode, PSBPROP, 3011772Sjl139090 (caddr_t)&psb_id) == -1) { 3021772Sjl139090 return (-1); 3031772Sjl139090 } else { 3041772Sjl139090 return (psb_id); 3051772Sjl139090 } 3061772Sjl139090 } 3071772Sjl139090 } 3081772Sjl139090 3091772Sjl139090 return (-1); 3101772Sjl139090 } 3111772Sjl139090 3121772Sjl139090 /* 3131772Sjl139090 * For OPL it's possible that memory from two or more successive boards 3141772Sjl139090 * will be contiguous across the boards, and therefore represented as a 3151772Sjl139090 * single chunk. 3161772Sjl139090 * This function splits such chunks down the board boundaries. 3171772Sjl139090 */ 3181772Sjl139090 static struct memlist * 3191772Sjl139090 opl_memlist_per_board(struct memlist *ml) 3201772Sjl139090 { 3211772Sjl139090 uint64_t ssize, low, high, boundary; 3221772Sjl139090 struct memlist *head, *tail, *new; 3231772Sjl139090 3241772Sjl139090 ssize = (1ull << OPL_MC_MEMBOARD_SHIFT); 3251772Sjl139090 3261772Sjl139090 head = tail = NULL; 3271772Sjl139090 3281772Sjl139090 for (; ml; ml = ml->next) { 3291772Sjl139090 low = (uint64_t)ml->address; 3301772Sjl139090 high = low+(uint64_t)(ml->size); 3311772Sjl139090 while (low < high) { 3321772Sjl139090 boundary = roundup(low+1, ssize); 3331772Sjl139090 boundary = MIN(high, boundary); 3341772Sjl139090 new = kmem_zalloc(sizeof (struct memlist), KM_SLEEP); 3351772Sjl139090 new->address = low; 3361772Sjl139090 new->size = boundary - low; 3371772Sjl139090 if (head == NULL) 3381772Sjl139090 head = new; 3391772Sjl139090 if (tail) { 3401772Sjl139090 tail->next = new; 3411772Sjl139090 new->prev = tail; 3421772Sjl139090 } 3431772Sjl139090 tail = new; 3441772Sjl139090 low = boundary; 3451772Sjl139090 } 3461772Sjl139090 } 3471772Sjl139090 return (head); 3481772Sjl139090 } 3491772Sjl139090 3501772Sjl139090 void 3511772Sjl139090 set_platform_cage_params(void) 3521772Sjl139090 { 3531772Sjl139090 extern pgcnt_t total_pages; 3541772Sjl139090 extern struct memlist *phys_avail; 3551772Sjl139090 struct memlist *ml, *tml; 3561772Sjl139090 3571772Sjl139090 if (kernel_cage_enable) { 3581772Sjl139090 pgcnt_t preferred_cage_size; 3591772Sjl139090 3605037Sjl139090 preferred_cage_size = MAX(opl_startup_cage_size, 3615037Sjl139090 total_pages / 256); 3621772Sjl139090 3631772Sjl139090 ml = opl_memlist_per_board(phys_avail); 3641772Sjl139090 3651772Sjl139090 /* 3661772Sjl139090 * Note: we are assuming that post has load the 3671772Sjl139090 * whole show in to the high end of memory. Having 3681772Sjl139090 * taken this leap, we copy the whole of phys_avail 3691772Sjl139090 * the glist and arrange for the cage to grow 3701772Sjl139090 * downward (descending pfns). 3711772Sjl139090 */ 3724266Sdp78419 kcage_range_init(ml, KCAGE_DOWN, preferred_cage_size); 3731772Sjl139090 3741772Sjl139090 /* free the memlist */ 3751772Sjl139090 do { 3761772Sjl139090 tml = ml->next; 3771772Sjl139090 kmem_free(ml, sizeof (struct memlist)); 3781772Sjl139090 ml = tml; 3791772Sjl139090 } while (ml != NULL); 3801772Sjl139090 } 3811772Sjl139090 3821772Sjl139090 if (kcage_on) 3831772Sjl139090 cmn_err(CE_NOTE, "!DR Kernel Cage is ENABLED"); 3841772Sjl139090 else 3851772Sjl139090 cmn_err(CE_NOTE, "!DR Kernel Cage is DISABLED"); 3861772Sjl139090 } 3871772Sjl139090 3881772Sjl139090 /*ARGSUSED*/ 3891772Sjl139090 int 3901772Sjl139090 plat_cpu_poweron(struct cpu *cp) 3911772Sjl139090 { 3921772Sjl139090 int (*opl_cpu_poweron)(struct cpu *) = NULL; 3931772Sjl139090 3941772Sjl139090 opl_cpu_poweron = 3951772Sjl139090 (int (*)(struct cpu *))kobj_getsymvalue("drmach_cpu_poweron", 0); 3961772Sjl139090 3971772Sjl139090 if (opl_cpu_poweron == NULL) 3981772Sjl139090 return (ENOTSUP); 3991772Sjl139090 else 4001772Sjl139090 return ((opl_cpu_poweron)(cp)); 4011772Sjl139090 4021772Sjl139090 } 4031772Sjl139090 4041772Sjl139090 /*ARGSUSED*/ 4051772Sjl139090 int 4061772Sjl139090 plat_cpu_poweroff(struct cpu *cp) 4071772Sjl139090 { 4081772Sjl139090 int (*opl_cpu_poweroff)(struct cpu *) = NULL; 4091772Sjl139090 4101772Sjl139090 opl_cpu_poweroff = 4111772Sjl139090 (int (*)(struct cpu *))kobj_getsymvalue("drmach_cpu_poweroff", 0); 4121772Sjl139090 4131772Sjl139090 if (opl_cpu_poweroff == NULL) 4141772Sjl139090 return (ENOTSUP); 4151772Sjl139090 else 4161772Sjl139090 return ((opl_cpu_poweroff)(cp)); 4171772Sjl139090 4181772Sjl139090 } 4191772Sjl139090 4201772Sjl139090 int 4211772Sjl139090 plat_max_boards(void) 4221772Sjl139090 { 4231772Sjl139090 return (OPL_MAX_BOARDS); 4241772Sjl139090 } 4251772Sjl139090 4261772Sjl139090 int 4271772Sjl139090 plat_max_cpu_units_per_board(void) 4281772Sjl139090 { 4291772Sjl139090 return (OPL_MAX_CPU_PER_BOARD); 4301772Sjl139090 } 4311772Sjl139090 4321772Sjl139090 int 4331772Sjl139090 plat_max_mem_units_per_board(void) 4341772Sjl139090 { 4351772Sjl139090 return (OPL_MAX_MEM_UNITS_PER_BOARD); 4361772Sjl139090 } 4371772Sjl139090 4381772Sjl139090 int 4391772Sjl139090 plat_max_io_units_per_board(void) 4401772Sjl139090 { 4411772Sjl139090 return (OPL_MAX_IO_UNITS_PER_BOARD); 4421772Sjl139090 } 4431772Sjl139090 4441772Sjl139090 int 4451772Sjl139090 plat_max_cmp_units_per_board(void) 4461772Sjl139090 { 4471772Sjl139090 return (OPL_MAX_CMP_UNITS_PER_BOARD); 4481772Sjl139090 } 4491772Sjl139090 4501772Sjl139090 int 4511772Sjl139090 plat_max_core_units_per_board(void) 4521772Sjl139090 { 4531772Sjl139090 return (OPL_MAX_CORE_UNITS_PER_BOARD); 4541772Sjl139090 } 4551772Sjl139090 4561772Sjl139090 int 4571772Sjl139090 plat_pfn_to_mem_node(pfn_t pfn) 4581772Sjl139090 { 4591772Sjl139090 return (pfn >> mem_node_pfn_shift); 4601772Sjl139090 } 4611772Sjl139090 4621772Sjl139090 /* ARGSUSED */ 4631772Sjl139090 void 4645648Ssetje plat_build_mem_nodes(prom_memlist_t *list, size_t nelems) 4651772Sjl139090 { 4661772Sjl139090 size_t elem; 4671772Sjl139090 pfn_t basepfn; 4681772Sjl139090 pgcnt_t npgs; 4691772Sjl139090 uint64_t boundary, ssize; 4701772Sjl139090 uint64_t low, high; 4711772Sjl139090 4721772Sjl139090 /* 4731772Sjl139090 * OPL mem slices are always aligned on a 256GB boundary. 4741772Sjl139090 */ 4751772Sjl139090 mem_node_pfn_shift = OPL_MC_MEMBOARD_SHIFT - MMU_PAGESHIFT; 4761772Sjl139090 mem_node_physalign = 0; 4771772Sjl139090 4781772Sjl139090 /* 4791772Sjl139090 * Boot install lists are arranged <addr, len>, <addr, len>, ... 4801772Sjl139090 */ 4811772Sjl139090 ssize = (1ull << OPL_MC_MEMBOARD_SHIFT); 4825648Ssetje for (elem = 0; elem < nelems; list++, elem++) { 4835648Ssetje low = list->addr; 4845648Ssetje high = low + list->size; 4851772Sjl139090 while (low < high) { 4861772Sjl139090 boundary = roundup(low+1, ssize); 4871772Sjl139090 boundary = MIN(high, boundary); 4881772Sjl139090 basepfn = btop(low); 4891772Sjl139090 npgs = btop(boundary - low); 4901772Sjl139090 mem_node_add_slice(basepfn, basepfn + npgs - 1); 4911772Sjl139090 low = boundary; 4921772Sjl139090 } 4931772Sjl139090 } 4941772Sjl139090 } 4951772Sjl139090 4961772Sjl139090 /* 4971772Sjl139090 * Find the CPU associated with a slice at boot-time. 4981772Sjl139090 */ 4991772Sjl139090 void 5001772Sjl139090 plat_fill_mc(pnode_t nodeid) 5011772Sjl139090 { 5021772Sjl139090 int board; 5031772Sjl139090 int memnode; 5041772Sjl139090 struct { 5051772Sjl139090 uint64_t addr; 5061772Sjl139090 uint64_t size; 5071772Sjl139090 } mem_range; 5081772Sjl139090 5091772Sjl139090 if (prom_getprop(nodeid, "board#", (caddr_t)&board) < 0) { 5101772Sjl139090 panic("Can not find board# property in mc node %x", nodeid); 5111772Sjl139090 } 5121772Sjl139090 if (prom_getprop(nodeid, "sb-mem-ranges", (caddr_t)&mem_range) < 0) { 5131772Sjl139090 panic("Can not find sb-mem-ranges property in mc node %x", 5145037Sjl139090 nodeid); 5151772Sjl139090 } 5161772Sjl139090 memnode = mem_range.addr >> OPL_MC_MEMBOARD_SHIFT; 5171772Sjl139090 plat_assign_lgrphand_to_mem_node(board, memnode); 5181772Sjl139090 } 5191772Sjl139090 5201772Sjl139090 /* 5211772Sjl139090 * Return the platform handle for the lgroup containing the given CPU 5221772Sjl139090 * 5231772Sjl139090 * For OPL, lgroup platform handle == board #. 5241772Sjl139090 */ 5251772Sjl139090 5261772Sjl139090 extern int mpo_disabled; 5271772Sjl139090 extern lgrp_handle_t lgrp_default_handle; 5281772Sjl139090 5291772Sjl139090 lgrp_handle_t 5301772Sjl139090 plat_lgrp_cpu_to_hand(processorid_t id) 5311772Sjl139090 { 5321772Sjl139090 lgrp_handle_t plathand; 5331772Sjl139090 5341772Sjl139090 /* 5351772Sjl139090 * Return the real platform handle for the CPU until 5361772Sjl139090 * such time as we know that MPO should be disabled. 5371772Sjl139090 * At that point, we set the "mpo_disabled" flag to true, 5381772Sjl139090 * and from that point on, return the default handle. 5391772Sjl139090 * 5401772Sjl139090 * By the time we know that MPO should be disabled, the 5411772Sjl139090 * first CPU will have already been added to a leaf 5421772Sjl139090 * lgroup, but that's ok. The common lgroup code will 5431772Sjl139090 * double check that the boot CPU is in the correct place, 5441772Sjl139090 * and in the case where mpo should be disabled, will move 5451772Sjl139090 * it to the root if necessary. 5461772Sjl139090 */ 5471772Sjl139090 if (mpo_disabled) { 5481772Sjl139090 /* If MPO is disabled, return the default (UMA) handle */ 5491772Sjl139090 plathand = lgrp_default_handle; 5501772Sjl139090 } else 5511772Sjl139090 plathand = (lgrp_handle_t)LSB_ID(id); 5521772Sjl139090 return (plathand); 5531772Sjl139090 } 5541772Sjl139090 5551772Sjl139090 /* 5561772Sjl139090 * Platform specific lgroup initialization 5571772Sjl139090 */ 5581772Sjl139090 void 5591772Sjl139090 plat_lgrp_init(void) 5601772Sjl139090 { 5611772Sjl139090 extern uint32_t lgrp_expand_proc_thresh; 5621772Sjl139090 extern uint32_t lgrp_expand_proc_diff; 5631772Sjl139090 5641772Sjl139090 /* 5651772Sjl139090 * Set tuneables for the OPL architecture 5661772Sjl139090 * 5671772Sjl139090 * lgrp_expand_proc_thresh is the minimum load on the lgroups 5681772Sjl139090 * this process is currently running on before considering 5691772Sjl139090 * expanding threads to another lgroup. 5701772Sjl139090 * 5711772Sjl139090 * lgrp_expand_proc_diff determines how much less the remote lgroup 5721772Sjl139090 * must be loaded before expanding to it. 5731772Sjl139090 * 5741772Sjl139090 * Since remote latencies can be costly, attempt to keep 3 threads 5751772Sjl139090 * within the same lgroup before expanding to the next lgroup. 5761772Sjl139090 */ 5771772Sjl139090 lgrp_expand_proc_thresh = LGRP_LOADAVG_THREAD_MAX * 3; 5781772Sjl139090 lgrp_expand_proc_diff = LGRP_LOADAVG_THREAD_MAX; 5791772Sjl139090 } 5801772Sjl139090 5811772Sjl139090 /* 5821772Sjl139090 * Platform notification of lgroup (re)configuration changes 5831772Sjl139090 */ 5841772Sjl139090 /*ARGSUSED*/ 5851772Sjl139090 void 5861772Sjl139090 plat_lgrp_config(lgrp_config_flag_t evt, uintptr_t arg) 5871772Sjl139090 { 5881772Sjl139090 update_membounds_t *umb; 5891772Sjl139090 lgrp_config_mem_rename_t lmr; 5901772Sjl139090 int sbd, tbd; 5911772Sjl139090 lgrp_handle_t hand, shand, thand; 5921772Sjl139090 int mnode, snode, tnode; 5931772Sjl139090 pfn_t start, end; 5941772Sjl139090 5951772Sjl139090 if (mpo_disabled) 5961772Sjl139090 return; 5971772Sjl139090 5981772Sjl139090 switch (evt) { 5991772Sjl139090 6001772Sjl139090 case LGRP_CONFIG_MEM_ADD: 6011772Sjl139090 /* 6021772Sjl139090 * Establish the lgroup handle to memnode translation. 6031772Sjl139090 */ 6041772Sjl139090 umb = (update_membounds_t *)arg; 6051772Sjl139090 6061772Sjl139090 hand = umb->u_board; 6071772Sjl139090 mnode = plat_pfn_to_mem_node(umb->u_base >> MMU_PAGESHIFT); 6081772Sjl139090 plat_assign_lgrphand_to_mem_node(hand, mnode); 6091772Sjl139090 6101772Sjl139090 break; 6111772Sjl139090 6121772Sjl139090 case LGRP_CONFIG_MEM_DEL: 6131772Sjl139090 /* 6141772Sjl139090 * Special handling for possible memory holes. 6151772Sjl139090 */ 6161772Sjl139090 umb = (update_membounds_t *)arg; 6171772Sjl139090 hand = umb->u_board; 6181772Sjl139090 if ((mnode = plat_lgrphand_to_mem_node(hand)) != -1) { 6191772Sjl139090 if (mem_node_config[mnode].exists) { 6201772Sjl139090 start = mem_node_config[mnode].physbase; 6211772Sjl139090 end = mem_node_config[mnode].physmax; 6221772Sjl139090 mem_node_pre_del_slice(start, end); 6231772Sjl139090 mem_node_post_del_slice(start, end, 0); 6241772Sjl139090 } 6251772Sjl139090 } 6261772Sjl139090 6271772Sjl139090 break; 6281772Sjl139090 6291772Sjl139090 case LGRP_CONFIG_MEM_RENAME: 6301772Sjl139090 /* 6311772Sjl139090 * During a DR copy-rename operation, all of the memory 6321772Sjl139090 * on one board is moved to another board -- but the 6331772Sjl139090 * addresses/pfns and memnodes don't change. This means 6341772Sjl139090 * the memory has changed locations without changing identity. 6351772Sjl139090 * 6361772Sjl139090 * Source is where we are copying from and target is where we 6371772Sjl139090 * are copying to. After source memnode is copied to target 6381772Sjl139090 * memnode, the physical addresses of the target memnode are 6391772Sjl139090 * renamed to match what the source memnode had. Then target 6401772Sjl139090 * memnode can be removed and source memnode can take its 6411772Sjl139090 * place. 6421772Sjl139090 * 6431772Sjl139090 * To do this, swap the lgroup handle to memnode mappings for 6441772Sjl139090 * the boards, so target lgroup will have source memnode and 6451772Sjl139090 * source lgroup will have empty target memnode which is where 6461772Sjl139090 * its memory will go (if any is added to it later). 6471772Sjl139090 * 6481772Sjl139090 * Then source memnode needs to be removed from its lgroup 6491772Sjl139090 * and added to the target lgroup where the memory was living 6501772Sjl139090 * but under a different name/memnode. The memory was in the 6511772Sjl139090 * target memnode and now lives in the source memnode with 6521772Sjl139090 * different physical addresses even though it is the same 6531772Sjl139090 * memory. 6541772Sjl139090 */ 6551772Sjl139090 sbd = arg & 0xffff; 6561772Sjl139090 tbd = (arg & 0xffff0000) >> 16; 6571772Sjl139090 shand = sbd; 6581772Sjl139090 thand = tbd; 6591772Sjl139090 snode = plat_lgrphand_to_mem_node(shand); 6601772Sjl139090 tnode = plat_lgrphand_to_mem_node(thand); 6611772Sjl139090 6621772Sjl139090 /* 6631772Sjl139090 * Special handling for possible memory holes. 6641772Sjl139090 */ 6651772Sjl139090 if (tnode != -1 && mem_node_config[tnode].exists) { 6663354Sjl139090 start = mem_node_config[tnode].physbase; 6673354Sjl139090 end = mem_node_config[tnode].physmax; 6681772Sjl139090 mem_node_pre_del_slice(start, end); 6691772Sjl139090 mem_node_post_del_slice(start, end, 0); 6701772Sjl139090 } 6711772Sjl139090 6721772Sjl139090 plat_assign_lgrphand_to_mem_node(thand, snode); 6731772Sjl139090 plat_assign_lgrphand_to_mem_node(shand, tnode); 6741772Sjl139090 6751772Sjl139090 lmr.lmem_rename_from = shand; 6761772Sjl139090 lmr.lmem_rename_to = thand; 6771772Sjl139090 6781772Sjl139090 /* 6791772Sjl139090 * Remove source memnode of copy rename from its lgroup 6801772Sjl139090 * and add it to its new target lgroup 6811772Sjl139090 */ 6821772Sjl139090 lgrp_config(LGRP_CONFIG_MEM_RENAME, (uintptr_t)snode, 6831772Sjl139090 (uintptr_t)&lmr); 6841772Sjl139090 6851772Sjl139090 break; 6861772Sjl139090 6871772Sjl139090 default: 6881772Sjl139090 break; 6891772Sjl139090 } 6901772Sjl139090 } 6911772Sjl139090 6921772Sjl139090 /* 6931772Sjl139090 * Return latency between "from" and "to" lgroups 6941772Sjl139090 * 6951772Sjl139090 * This latency number can only be used for relative comparison 6961772Sjl139090 * between lgroups on the running system, cannot be used across platforms, 6971772Sjl139090 * and may not reflect the actual latency. It is platform and implementation 6981772Sjl139090 * specific, so platform gets to decide its value. It would be nice if the 6991772Sjl139090 * number was at least proportional to make comparisons more meaningful though. 7001772Sjl139090 * NOTE: The numbers below are supposed to be load latencies for uncached 7011772Sjl139090 * memory divided by 10. 7021772Sjl139090 * 7031772Sjl139090 */ 7041772Sjl139090 int 7051772Sjl139090 plat_lgrp_latency(lgrp_handle_t from, lgrp_handle_t to) 7061772Sjl139090 { 7071772Sjl139090 /* 7081772Sjl139090 * Return min remote latency when there are more than two lgroups 7091772Sjl139090 * (root and child) and getting latency between two different lgroups 7101772Sjl139090 * or root is involved 7111772Sjl139090 */ 7121772Sjl139090 if (lgrp_optimizations() && (from != to || 7131772Sjl139090 from == LGRP_DEFAULT_HANDLE || to == LGRP_DEFAULT_HANDLE)) 7142491Shyw return (42); 7151772Sjl139090 else 7162491Shyw return (35); 7171772Sjl139090 } 7181772Sjl139090 7191772Sjl139090 /* 7201772Sjl139090 * Return platform handle for root lgroup 7211772Sjl139090 */ 7221772Sjl139090 lgrp_handle_t 7231772Sjl139090 plat_lgrp_root_hand(void) 7241772Sjl139090 { 7251772Sjl139090 if (mpo_disabled) 7261772Sjl139090 return (lgrp_default_handle); 7271772Sjl139090 7281772Sjl139090 return (LGRP_DEFAULT_HANDLE); 7291772Sjl139090 } 7301772Sjl139090 7311772Sjl139090 /*ARGSUSED*/ 7321772Sjl139090 void 7331772Sjl139090 plat_freelist_process(int mnode) 7341772Sjl139090 { 7351772Sjl139090 } 7361772Sjl139090 7371772Sjl139090 void 7381772Sjl139090 load_platform_drivers(void) 7391772Sjl139090 { 7401772Sjl139090 (void) i_ddi_attach_pseudo_node("dr"); 7411772Sjl139090 } 7421772Sjl139090 7431772Sjl139090 /* 7441772Sjl139090 * No platform drivers on this platform 7451772Sjl139090 */ 7461772Sjl139090 char *platform_module_list[] = { 7471772Sjl139090 (char *)0 7481772Sjl139090 }; 7491772Sjl139090 7501772Sjl139090 /*ARGSUSED*/ 7511772Sjl139090 void 7521772Sjl139090 plat_tod_fault(enum tod_fault_type tod_bad) 7531772Sjl139090 { 7541772Sjl139090 } 7551772Sjl139090 7561772Sjl139090 /*ARGSUSED*/ 7571772Sjl139090 void 7581772Sjl139090 cpu_sgn_update(ushort_t sgn, uchar_t state, uchar_t sub_state, int cpuid) 7591772Sjl139090 { 7601772Sjl139090 static void (*scf_panic_callback)(int); 7611772Sjl139090 static void (*scf_shutdown_callback)(int); 7621772Sjl139090 7631772Sjl139090 /* 7641772Sjl139090 * This is for notifing system panic/shutdown to SCF. 7651772Sjl139090 * In case of shutdown and panic, SCF call back 7661772Sjl139090 * function should be called. 7671772Sjl139090 * <SCF call back functions> 7681772Sjl139090 * scf_panic_callb() : panicsys()->panic_quiesce_hw() 7691772Sjl139090 * scf_shutdown_callb(): halt() or power_down() or reboot_machine() 7701772Sjl139090 * cpuid should be -1 and state should be SIGST_EXIT. 7711772Sjl139090 */ 7721772Sjl139090 if (state == SIGST_EXIT && cpuid == -1) { 7731772Sjl139090 7741772Sjl139090 /* 7751772Sjl139090 * find the symbol for the SCF panic callback routine in driver 7761772Sjl139090 */ 7771772Sjl139090 if (scf_panic_callback == NULL) 7781772Sjl139090 scf_panic_callback = (void (*)(int)) 7795037Sjl139090 modgetsymvalue("scf_panic_callb", 0); 7801772Sjl139090 if (scf_shutdown_callback == NULL) 7811772Sjl139090 scf_shutdown_callback = (void (*)(int)) 7825037Sjl139090 modgetsymvalue("scf_shutdown_callb", 0); 7831772Sjl139090 7841772Sjl139090 switch (sub_state) { 7851772Sjl139090 case SIGSUBST_PANIC: 7861772Sjl139090 if (scf_panic_callback == NULL) { 7871772Sjl139090 cmn_err(CE_NOTE, "!cpu_sgn_update: " 7881772Sjl139090 "scf_panic_callb not found\n"); 7891772Sjl139090 return; 7901772Sjl139090 } 7911772Sjl139090 scf_panic_callback(SIGSUBST_PANIC); 7921772Sjl139090 break; 7931772Sjl139090 7941772Sjl139090 case SIGSUBST_HALT: 7951772Sjl139090 if (scf_shutdown_callback == NULL) { 7961772Sjl139090 cmn_err(CE_NOTE, "!cpu_sgn_update: " 7971772Sjl139090 "scf_shutdown_callb not found\n"); 7981772Sjl139090 return; 7991772Sjl139090 } 8001772Sjl139090 scf_shutdown_callback(SIGSUBST_HALT); 8011772Sjl139090 break; 8021772Sjl139090 8031772Sjl139090 case SIGSUBST_ENVIRON: 8041772Sjl139090 if (scf_shutdown_callback == NULL) { 8051772Sjl139090 cmn_err(CE_NOTE, "!cpu_sgn_update: " 8061772Sjl139090 "scf_shutdown_callb not found\n"); 8071772Sjl139090 return; 8081772Sjl139090 } 8091772Sjl139090 scf_shutdown_callback(SIGSUBST_ENVIRON); 8101772Sjl139090 break; 8111772Sjl139090 8121772Sjl139090 case SIGSUBST_REBOOT: 8131772Sjl139090 if (scf_shutdown_callback == NULL) { 8141772Sjl139090 cmn_err(CE_NOTE, "!cpu_sgn_update: " 8151772Sjl139090 "scf_shutdown_callb not found\n"); 8161772Sjl139090 return; 8171772Sjl139090 } 8181772Sjl139090 scf_shutdown_callback(SIGSUBST_REBOOT); 8191772Sjl139090 break; 8201772Sjl139090 } 8211772Sjl139090 } 8221772Sjl139090 } 8231772Sjl139090 8241772Sjl139090 /*ARGSUSED*/ 8251772Sjl139090 int 8261772Sjl139090 plat_get_mem_unum(int synd_code, uint64_t flt_addr, int flt_bus_id, 8271772Sjl139090 int flt_in_memory, ushort_t flt_status, 8281772Sjl139090 char *buf, int buflen, int *lenp) 8291772Sjl139090 { 8301772Sjl139090 /* 8311772Sjl139090 * check if it's a Memory error. 8321772Sjl139090 */ 8331772Sjl139090 if (flt_in_memory) { 8341772Sjl139090 if (opl_get_mem_unum != NULL) { 8355037Sjl139090 return (opl_get_mem_unum(synd_code, flt_addr, buf, 8365037Sjl139090 buflen, lenp)); 8371772Sjl139090 } else { 8381772Sjl139090 return (ENOTSUP); 8391772Sjl139090 } 8401772Sjl139090 } else { 8411772Sjl139090 return (ENOTSUP); 8421772Sjl139090 } 8431772Sjl139090 } 8441772Sjl139090 8451772Sjl139090 /*ARGSUSED*/ 8461772Sjl139090 int 8471772Sjl139090 plat_get_cpu_unum(int cpuid, char *buf, int buflen, int *lenp) 8481772Sjl139090 { 8492214Sav145390 int ret = 0; 8505347Sjfrank int sb; 8513123Ssubhan int plen; 8521772Sjl139090 8531772Sjl139090 sb = opl_get_physical_board(LSB_ID(cpuid)); 8541772Sjl139090 if (sb == -1) { 8551772Sjl139090 return (ENXIO); 8561772Sjl139090 } 8571772Sjl139090 8583627Ssubhan /* 8593627Ssubhan * opl_cur_model is assigned here 8603627Ssubhan */ 8613627Ssubhan if (opl_cur_model == NULL) { 8623627Ssubhan set_model_info(); 8635539Swh31274 8645539Swh31274 /* 8655539Swh31274 * if not matched, return 8665539Swh31274 */ 8675539Swh31274 if (opl_cur_model == NULL) 8685539Swh31274 return (ENODEV); 8693627Ssubhan } 8703627Ssubhan 8713123Ssubhan ASSERT((opl_cur_model - opl_models) == (opl_cur_model->model_type)); 8723123Ssubhan 8733123Ssubhan switch (opl_cur_model->model_type) { 8743123Ssubhan case FF1: 8752214Sav145390 plen = snprintf(buf, buflen, "/%s/CPUM%d", "MBU_A", 8762214Sav145390 CHIP_ID(cpuid) / 2); 8772214Sav145390 break; 8782214Sav145390 8793123Ssubhan case FF2: 8802214Sav145390 plen = snprintf(buf, buflen, "/%s/CPUM%d", "MBU_B", 8812808Sav145390 (CHIP_ID(cpuid) / 2) + (sb * 2)); 8822214Sav145390 break; 8832214Sav145390 8843123Ssubhan case DC1: 8853123Ssubhan case DC2: 8863123Ssubhan case DC3: 8872214Sav145390 plen = snprintf(buf, buflen, "/%s%02d/CPUM%d", "CMU", sb, 8882214Sav145390 CHIP_ID(cpuid)); 8892214Sav145390 break; 8902214Sav145390 8912214Sav145390 default: 8922214Sav145390 /* This should never happen */ 8932214Sav145390 return (ENODEV); 8942214Sav145390 } 8952214Sav145390 8962214Sav145390 if (plen >= buflen) { 8972214Sav145390 ret = ENOSPC; 8981772Sjl139090 } else { 8991772Sjl139090 if (lenp) 9001772Sjl139090 *lenp = strlen(buf); 9011772Sjl139090 } 9022214Sav145390 return (ret); 9031772Sjl139090 } 9041772Sjl139090 9051772Sjl139090 void 9061772Sjl139090 plat_nodename_set(void) 9071772Sjl139090 { 9085347Sjfrank post_xscf_msg((char *)&utsname, sizeof (struct utsname)); 9091772Sjl139090 } 9101772Sjl139090 9111772Sjl139090 caddr_t efcode_vaddr = NULL; 9121772Sjl139090 9131772Sjl139090 /* 9141772Sjl139090 * Preallocate enough memory for fcode claims. 9151772Sjl139090 */ 9161772Sjl139090 9171772Sjl139090 caddr_t 9181772Sjl139090 efcode_alloc(caddr_t alloc_base) 9191772Sjl139090 { 9201772Sjl139090 caddr_t efcode_alloc_base = (caddr_t)roundup((uintptr_t)alloc_base, 9211772Sjl139090 MMU_PAGESIZE); 9221772Sjl139090 caddr_t vaddr; 9231772Sjl139090 9241772Sjl139090 /* 9251772Sjl139090 * allocate the physical memory for the Oberon fcode. 9261772Sjl139090 */ 9271772Sjl139090 if ((vaddr = (caddr_t)BOP_ALLOC(bootops, efcode_alloc_base, 9281772Sjl139090 efcode_size, MMU_PAGESIZE)) == NULL) 9291772Sjl139090 cmn_err(CE_PANIC, "Cannot allocate Efcode Memory"); 9301772Sjl139090 9311772Sjl139090 efcode_vaddr = vaddr; 9321772Sjl139090 9331772Sjl139090 return (efcode_alloc_base + efcode_size); 9341772Sjl139090 } 9351772Sjl139090 9361772Sjl139090 caddr_t 9371772Sjl139090 plat_startup_memlist(caddr_t alloc_base) 9381772Sjl139090 { 9391772Sjl139090 caddr_t tmp_alloc_base; 9401772Sjl139090 9411772Sjl139090 tmp_alloc_base = efcode_alloc(alloc_base); 9421772Sjl139090 tmp_alloc_base = 9431772Sjl139090 (caddr_t)roundup((uintptr_t)tmp_alloc_base, ecache_alignsize); 9441772Sjl139090 return (tmp_alloc_base); 9451772Sjl139090 } 9461772Sjl139090 9475834Spt157919 /* need to forward declare these */ 9485834Spt157919 static void plat_lock_delay(uint_t); 9495834Spt157919 9501772Sjl139090 void 9511772Sjl139090 startup_platform(void) 9521772Sjl139090 { 9535788Smv143129 if (clock_tick_threshold == 0) 9545788Smv143129 clock_tick_threshold = OPL_CLOCK_TICK_THRESHOLD; 9555788Smv143129 if (clock_tick_ncpus == 0) 9565788Smv143129 clock_tick_ncpus = OPL_CLOCK_TICK_NCPUS; 9575834Spt157919 mutex_lock_delay = plat_lock_delay; 9585834Spt157919 mutex_cap_factor = OPL_BOFF_MAX_SCALE; 9591772Sjl139090 } 9602214Sav145390 961*5923Sjfrank static uint_t 962*5923Sjfrank get_mmu_id(processorid_t cpuid) 963*5923Sjfrank { 964*5923Sjfrank int pb = opl_get_physical_board(LSB_ID(cpuid)); 965*5923Sjfrank 966*5923Sjfrank if (pb == -1) { 967*5923Sjfrank cmn_err(CE_PANIC, 968*5923Sjfrank "opl_get_physical_board failed (cpu %d LSB %u)", 969*5923Sjfrank cpuid, LSB_ID(cpuid)); 970*5923Sjfrank } 971*5923Sjfrank return (pb * OPL_MAX_COREID_PER_BOARD) + (CHIP_ID(cpuid) * 972*5923Sjfrank OPL_MAX_COREID_PER_CMP) + CORE_ID(cpuid); 973*5923Sjfrank } 974*5923Sjfrank 9752241Shuah void 9762241Shuah plat_cpuid_to_mmu_ctx_info(processorid_t cpuid, mmu_ctx_info_t *info) 9772241Shuah { 9782241Shuah int impl; 9792241Shuah 9802241Shuah impl = cpunodes[cpuid].implementation; 9815037Sjl139090 if (IS_OLYMPUS_C(impl) || IS_JUPITER(impl)) { 982*5923Sjfrank info->mmu_idx = get_mmu_id(cpuid); 9832241Shuah info->mmu_nctxs = 8192; 9842241Shuah } else { 9852241Shuah cmn_err(CE_PANIC, "Unknown processor %d", impl); 9862241Shuah } 9872241Shuah } 9882241Shuah 9892214Sav145390 int 9902214Sav145390 plat_get_mem_sid(char *unum, char *buf, int buflen, int *lenp) 9912214Sav145390 { 9922214Sav145390 if (opl_get_mem_sid == NULL) { 9932214Sav145390 return (ENOTSUP); 9942214Sav145390 } 9952214Sav145390 return (opl_get_mem_sid(unum, buf, buflen, lenp)); 9962214Sav145390 } 9972214Sav145390 9982214Sav145390 int 9992214Sav145390 plat_get_mem_offset(uint64_t paddr, uint64_t *offp) 10002214Sav145390 { 10012214Sav145390 if (opl_get_mem_offset == NULL) { 10022214Sav145390 return (ENOTSUP); 10032214Sav145390 } 10042214Sav145390 return (opl_get_mem_offset(paddr, offp)); 10052214Sav145390 } 10062214Sav145390 10072214Sav145390 int 10082214Sav145390 plat_get_mem_addr(char *unum, char *sid, uint64_t offset, uint64_t *addrp) 10092214Sav145390 { 10102214Sav145390 if (opl_get_mem_addr == NULL) { 10112214Sav145390 return (ENOTSUP); 10122214Sav145390 } 10132214Sav145390 return (opl_get_mem_addr(unum, sid, offset, addrp)); 10142214Sav145390 } 10153914Spm145316 10163914Spm145316 void 10175834Spt157919 plat_lock_delay(uint_t backoff) 10183914Spm145316 { 10193914Spm145316 int i; 10205834Spt157919 uint_t cnt, remcnt; 10213914Spm145316 int ctr; 10225834Spt157919 hrtime_t delay_start, rem_delay; 10233914Spm145316 /* 10243914Spm145316 * Platform specific lock delay code for OPL 10253914Spm145316 * 10263914Spm145316 * Using staged linear increases in the delay. 10273914Spm145316 * The sleep instruction is the preferred method of delay, 10283914Spm145316 * but is too large of granularity for the initial backoff. 10293914Spm145316 */ 10303914Spm145316 10315834Spt157919 if (backoff < 100) { 10323914Spm145316 /* 10333914Spm145316 * If desired backoff is long enough, 10343914Spm145316 * use sleep for most of it 10353914Spm145316 */ 10365834Spt157919 for (cnt = backoff; 10375834Spt157919 cnt >= OPL_BOFF_SLEEP; 10385037Sjl139090 cnt -= OPL_BOFF_SLEEP) { 10393914Spm145316 cpu_smt_pause(); 10403914Spm145316 } 10413914Spm145316 /* 10423914Spm145316 * spin for small remainder of backoff 10433914Spm145316 */ 10443914Spm145316 for (ctr = cnt * OPL_BOFF_SPIN; ctr; ctr--) { 10455834Spt157919 mutex_delay_default(); 10463914Spm145316 } 10473914Spm145316 } else { 10485834Spt157919 /* backoff is large. Fill it by sleeping */ 10493914Spm145316 delay_start = gethrtime(); 10505834Spt157919 cnt = backoff / OPL_BOFF_SLEEP; 10513914Spm145316 /* 10523914Spm145316 * use sleep instructions for delay 10533914Spm145316 */ 10543914Spm145316 for (i = 0; i < cnt; i++) { 10553914Spm145316 cpu_smt_pause(); 10563914Spm145316 } 10573914Spm145316 10583914Spm145316 /* 10593914Spm145316 * Note: if the other strand executes a sleep instruction, 10603914Spm145316 * then the sleep ends immediately with a minimum time of 10613914Spm145316 * 42 clocks. We check gethrtime to insure we have 10623914Spm145316 * waited long enough. And we include both a short 10635834Spt157919 * spin loop and a sleep for repeated delay times. 10643914Spm145316 */ 10653914Spm145316 10665834Spt157919 rem_delay = gethrtime() - delay_start; 10675834Spt157919 while (rem_delay < cnt * OPL_BOFF_TM) { 10685834Spt157919 remcnt = cnt - (rem_delay / OPL_BOFF_TM); 10695834Spt157919 for (i = 0; i < remcnt; i++) { 10705834Spt157919 cpu_smt_pause(); 10715834Spt157919 for (ctr = OPL_BOFF_SPIN; ctr; ctr--) { 10725834Spt157919 mutex_delay_default(); 10735834Spt157919 } 10743914Spm145316 } 10755834Spt157919 rem_delay = gethrtime() - delay_start; 10763914Spm145316 } 10773914Spm145316 } 10783914Spm145316 } 10795347Sjfrank 10805347Sjfrank /* 10815347Sjfrank * The following code implements asynchronous call to XSCF to setup the 10825347Sjfrank * domain node name. 10835347Sjfrank */ 10845347Sjfrank 10855347Sjfrank #define FREE_MSG(m) kmem_free((m), NM_LEN((m)->len)) 10865347Sjfrank 10875347Sjfrank /* 10885347Sjfrank * The following three macros define the all operations on the request 10895347Sjfrank * list we are using here, and hide the details of the list 10905347Sjfrank * implementation from the code. 10915347Sjfrank */ 10925347Sjfrank #define PUSH(m) \ 10935347Sjfrank { \ 10945347Sjfrank (m)->next = ctl_msg.head; \ 10955347Sjfrank (m)->prev = NULL; \ 10965347Sjfrank if ((m)->next != NULL) \ 10975347Sjfrank (m)->next->prev = (m); \ 10985347Sjfrank ctl_msg.head = (m); \ 10995347Sjfrank } 11005347Sjfrank 11015347Sjfrank #define REMOVE(m) \ 11025347Sjfrank { \ 11035347Sjfrank if ((m)->prev != NULL) \ 11045347Sjfrank (m)->prev->next = (m)->next; \ 11055347Sjfrank else \ 11065347Sjfrank ctl_msg.head = (m)->next; \ 11075347Sjfrank if ((m)->next != NULL) \ 11085347Sjfrank (m)->next->prev = (m)->prev; \ 11095347Sjfrank } 11105347Sjfrank 11115347Sjfrank #define FREE_THE_TAIL(head) \ 11125347Sjfrank { \ 11135347Sjfrank nm_msg_t *n_msg, *m; \ 11145347Sjfrank m = (head)->next; \ 11155347Sjfrank (head)->next = NULL; \ 11165347Sjfrank while (m != NULL) { \ 11175347Sjfrank n_msg = m->next; \ 11185347Sjfrank FREE_MSG(m); \ 11195347Sjfrank m = n_msg; \ 11205347Sjfrank } \ 11215347Sjfrank } 11225347Sjfrank 11235347Sjfrank #define SCF_PUTINFO(f, s, p) \ 11245347Sjfrank f(KEY_ESCF, 0x01, 0, s, p) 11255347Sjfrank 11265347Sjfrank #define PASS2XSCF(m, r) ((r = SCF_PUTINFO(ctl_msg.scf_service_function, \ 11275347Sjfrank (m)->len, (m)->data)) == 0) 11285347Sjfrank 11295347Sjfrank /* 11305347Sjfrank * The value of the following macro loosely depends on the 11315347Sjfrank * value of the "device busy" timeout used in the SCF driver. 11325347Sjfrank * (See pass2xscf_thread()). 11335347Sjfrank */ 11345347Sjfrank #define SCF_DEVBUSY_DELAY 10 11355347Sjfrank 11365347Sjfrank /* 11375347Sjfrank * The default number of attempts to contact the scf driver 11385347Sjfrank * if we cannot fetch any information about the timeout value 11395347Sjfrank * it uses. 11405347Sjfrank */ 11415347Sjfrank 11425347Sjfrank #define REPEATS 4 11435347Sjfrank 11445347Sjfrank typedef struct nm_msg { 11455347Sjfrank struct nm_msg *next; 11465347Sjfrank struct nm_msg *prev; 11475347Sjfrank int len; 11485347Sjfrank char data[1]; 11495347Sjfrank } nm_msg_t; 11505347Sjfrank 11515347Sjfrank #define NM_LEN(len) (sizeof (nm_msg_t) + (len) - 1) 11525347Sjfrank 11535347Sjfrank static struct ctlmsg { 11545347Sjfrank nm_msg_t *head; 11555347Sjfrank nm_msg_t *now_serving; 11565347Sjfrank kmutex_t nm_lock; 11575347Sjfrank kthread_t *nmt; 11585347Sjfrank int cnt; 11595347Sjfrank int (*scf_service_function)(uint32_t, uint8_t, 11605347Sjfrank uint32_t, uint32_t, void *); 11615347Sjfrank } ctl_msg; 11625347Sjfrank 11635347Sjfrank static void 11645347Sjfrank post_xscf_msg(char *dp, int len) 11655347Sjfrank { 11665347Sjfrank nm_msg_t *msg; 11675347Sjfrank 11685347Sjfrank msg = (nm_msg_t *)kmem_zalloc(NM_LEN(len), KM_SLEEP); 11695347Sjfrank 11705347Sjfrank bcopy(dp, msg->data, len); 11715347Sjfrank msg->len = len; 11725347Sjfrank 11735347Sjfrank mutex_enter(&ctl_msg.nm_lock); 11745347Sjfrank if (ctl_msg.nmt == NULL) { 11755347Sjfrank ctl_msg.nmt = thread_create(NULL, 0, pass2xscf_thread, 11765347Sjfrank NULL, 0, &p0, TS_RUN, minclsyspri); 11775347Sjfrank } 11785347Sjfrank 11795347Sjfrank PUSH(msg); 11805347Sjfrank ctl_msg.cnt++; 11815347Sjfrank mutex_exit(&ctl_msg.nm_lock); 11825347Sjfrank } 11835347Sjfrank 11845347Sjfrank static void 11855347Sjfrank pass2xscf_thread() 11865347Sjfrank { 11875347Sjfrank nm_msg_t *msg; 11885347Sjfrank int ret; 11895347Sjfrank uint_t i, msg_sent, xscf_driver_delay; 11905347Sjfrank static uint_t repeat_cnt; 11915347Sjfrank uint_t *scf_wait_cnt; 11925347Sjfrank 11935347Sjfrank mutex_enter(&ctl_msg.nm_lock); 11945347Sjfrank 11955347Sjfrank /* 11965347Sjfrank * Find the address of the SCF put routine if it's not done yet. 11975347Sjfrank */ 11985347Sjfrank if (ctl_msg.scf_service_function == NULL) { 11995347Sjfrank if ((ctl_msg.scf_service_function = 12005347Sjfrank (int (*)(uint32_t, uint8_t, uint32_t, uint32_t, void *)) 12015347Sjfrank modgetsymvalue("scf_service_putinfo", 0)) == NULL) { 12025347Sjfrank cmn_err(CE_NOTE, "pass2xscf_thread: " 12035347Sjfrank "scf_service_putinfo not found\n"); 12045347Sjfrank ctl_msg.nmt = NULL; 12055347Sjfrank mutex_exit(&ctl_msg.nm_lock); 12065347Sjfrank return; 12075347Sjfrank } 12085347Sjfrank } 12095347Sjfrank 12105347Sjfrank /* 12115347Sjfrank * Calculate the number of attempts to connect XSCF based on the 12125347Sjfrank * scf driver delay (which is 12135347Sjfrank * SCF_DEVBUSY_DELAY*scf_online_wait_rcnt seconds) and the value 12145347Sjfrank * of xscf_connect_delay (the total number of seconds to wait 12155347Sjfrank * till xscf get ready.) 12165347Sjfrank */ 12175347Sjfrank if (repeat_cnt == 0) { 12185347Sjfrank if ((scf_wait_cnt = 12195347Sjfrank (uint_t *) 12205347Sjfrank modgetsymvalue("scf_online_wait_rcnt", 0)) == NULL) { 12215347Sjfrank repeat_cnt = REPEATS; 12225347Sjfrank } else { 12235347Sjfrank 12245347Sjfrank xscf_driver_delay = *scf_wait_cnt * 12255347Sjfrank SCF_DEVBUSY_DELAY; 12265347Sjfrank repeat_cnt = (xscf_connect_delay/xscf_driver_delay) + 1; 12275347Sjfrank } 12285347Sjfrank } 12295347Sjfrank 12305347Sjfrank while (ctl_msg.cnt != 0) { 12315347Sjfrank 12325347Sjfrank /* 12335347Sjfrank * Take the very last request from the queue, 12345347Sjfrank */ 12355347Sjfrank ctl_msg.now_serving = ctl_msg.head; 12365347Sjfrank ASSERT(ctl_msg.now_serving != NULL); 12375347Sjfrank 12385347Sjfrank /* 12395347Sjfrank * and discard all the others if any. 12405347Sjfrank */ 12415347Sjfrank FREE_THE_TAIL(ctl_msg.now_serving); 12425347Sjfrank ctl_msg.cnt = 1; 12435347Sjfrank mutex_exit(&ctl_msg.nm_lock); 12445347Sjfrank 12455347Sjfrank /* 12465347Sjfrank * Pass the name to XSCF. Note please, we do not hold the 12475347Sjfrank * mutex while we are doing this. 12485347Sjfrank */ 12495347Sjfrank msg_sent = 0; 12505347Sjfrank for (i = 0; i < repeat_cnt; i++) { 12515347Sjfrank if (PASS2XSCF(ctl_msg.now_serving, ret)) { 12525347Sjfrank msg_sent = 1; 12535347Sjfrank break; 12545347Sjfrank } else { 12555347Sjfrank if (ret != EBUSY) { 12565347Sjfrank cmn_err(CE_NOTE, "pass2xscf_thread:" 12575347Sjfrank " unexpected return code" 12585347Sjfrank " from scf_service_putinfo():" 12595347Sjfrank " %d\n", ret); 12605347Sjfrank } 12615347Sjfrank } 12625347Sjfrank } 12635347Sjfrank 12645347Sjfrank if (msg_sent) { 12655347Sjfrank 12665347Sjfrank /* 12675347Sjfrank * Remove the request from the list 12685347Sjfrank */ 12695347Sjfrank mutex_enter(&ctl_msg.nm_lock); 12705347Sjfrank msg = ctl_msg.now_serving; 12715347Sjfrank ctl_msg.now_serving = NULL; 12725347Sjfrank REMOVE(msg); 12735347Sjfrank ctl_msg.cnt--; 12745347Sjfrank mutex_exit(&ctl_msg.nm_lock); 12755347Sjfrank FREE_MSG(msg); 12765347Sjfrank } else { 12775347Sjfrank 12785347Sjfrank /* 12795347Sjfrank * If while we have tried to communicate with 12805347Sjfrank * XSCF there were any other requests we are 12815347Sjfrank * going to drop this one and take the latest 12825347Sjfrank * one. Otherwise we will try to pass this one 12835347Sjfrank * again. 12845347Sjfrank */ 12855347Sjfrank cmn_err(CE_NOTE, 12865347Sjfrank "pass2xscf_thread: " 12875347Sjfrank "scf_service_putinfo " 12885347Sjfrank "not responding\n"); 12895347Sjfrank } 12905347Sjfrank mutex_enter(&ctl_msg.nm_lock); 12915347Sjfrank } 12925347Sjfrank 12935347Sjfrank /* 12945347Sjfrank * The request queue is empty, exit. 12955347Sjfrank */ 12965347Sjfrank ctl_msg.nmt = NULL; 12975347Sjfrank mutex_exit(&ctl_msg.nm_lock); 12985347Sjfrank } 1299