10Sstevel@tonic-gate /* 20Sstevel@tonic-gate * CDDL HEADER START 30Sstevel@tonic-gate * 40Sstevel@tonic-gate * The contents of this file are subject to the terms of the 51070Skais * Common Development and Distribution License (the "License"). 61070Skais * You may not use this file except in compliance with the License. 70Sstevel@tonic-gate * 80Sstevel@tonic-gate * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 90Sstevel@tonic-gate * or http://www.opensolaris.org/os/licensing. 100Sstevel@tonic-gate * See the License for the specific language governing permissions 110Sstevel@tonic-gate * and limitations under the License. 120Sstevel@tonic-gate * 130Sstevel@tonic-gate * When distributing Covered Code, include this CDDL HEADER in each 140Sstevel@tonic-gate * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 150Sstevel@tonic-gate * If applicable, add the following below this CDDL HEADER, with the 160Sstevel@tonic-gate * fields enclosed by brackets "[]" replaced with your own identifying 170Sstevel@tonic-gate * information: Portions Copyright [yyyy] [name of copyright owner] 180Sstevel@tonic-gate * 190Sstevel@tonic-gate * CDDL HEADER END 200Sstevel@tonic-gate */ 210Sstevel@tonic-gate /* 221920Smcpowers * Copyright 2006 Sun Microsystems, Inc. All rights reserved. 230Sstevel@tonic-gate * Use is subject to license terms. 240Sstevel@tonic-gate */ 250Sstevel@tonic-gate 260Sstevel@tonic-gate #pragma ident "%Z%%M% %I% %E% SMI" 270Sstevel@tonic-gate 280Sstevel@tonic-gate /* 290Sstevel@tonic-gate * This file implements the interfaces that the /dev/random 300Sstevel@tonic-gate * driver uses for read(2), write(2) and poll(2) on /dev/random or 310Sstevel@tonic-gate * /dev/urandom. It also implements the kernel API - random_add_entropy(), 320Sstevel@tonic-gate * random_get_pseudo_bytes() and random_get_bytes(). 330Sstevel@tonic-gate * 340Sstevel@tonic-gate * We periodically collect random bits from providers which are registered 350Sstevel@tonic-gate * with the Kernel Cryptographic Framework (kCF) as capable of random 360Sstevel@tonic-gate * number generation. The random bits are maintained in a cache and 370Sstevel@tonic-gate * it is used for high quality random numbers (/dev/random) requests. 380Sstevel@tonic-gate * We pick a provider and call its SPI routine, if the cache does not have 390Sstevel@tonic-gate * enough bytes to satisfy a request. 400Sstevel@tonic-gate * 410Sstevel@tonic-gate * /dev/urandom requests use a software-based generator algorithm that uses the 420Sstevel@tonic-gate * random bits in the cache as a seed. We create one pseudo-random generator 430Sstevel@tonic-gate * (for /dev/urandom) per possible CPU on the system, and use it, 440Sstevel@tonic-gate * kmem-magazine-style, to avoid cache line contention. 450Sstevel@tonic-gate * 460Sstevel@tonic-gate * LOCKING HIERARCHY: 470Sstevel@tonic-gate * 1) rmp->rm_lock protects the per-cpu pseudo-random generators. 480Sstevel@tonic-gate * 2) rndpool_lock protects the high-quality randomness pool. 490Sstevel@tonic-gate * It may be locked while a rmp->rm_lock is held. 500Sstevel@tonic-gate * 510Sstevel@tonic-gate * A history note: The kernel API and the software-based algorithms in this 520Sstevel@tonic-gate * file used to be part of the /dev/random driver. 530Sstevel@tonic-gate */ 540Sstevel@tonic-gate 550Sstevel@tonic-gate #include <sys/types.h> 560Sstevel@tonic-gate #include <sys/conf.h> 570Sstevel@tonic-gate #include <sys/sunddi.h> 580Sstevel@tonic-gate #include <sys/disp.h> 590Sstevel@tonic-gate #include <sys/modctl.h> 600Sstevel@tonic-gate #include <sys/ddi.h> 610Sstevel@tonic-gate #include <sys/crypto/common.h> 620Sstevel@tonic-gate #include <sys/crypto/api.h> 630Sstevel@tonic-gate #include <sys/crypto/impl.h> 640Sstevel@tonic-gate #include <sys/crypto/sched_impl.h> 650Sstevel@tonic-gate #include <sys/random.h> 660Sstevel@tonic-gate #include <sys/sha1.h> 670Sstevel@tonic-gate #include <sys/time.h> 680Sstevel@tonic-gate #include <sys/sysmacros.h> 690Sstevel@tonic-gate #include <sys/cpuvar.h> 700Sstevel@tonic-gate #include <sys/taskq.h> 710Sstevel@tonic-gate 720Sstevel@tonic-gate #define RNDPOOLSIZE 1024 /* Pool size in bytes */ 730Sstevel@tonic-gate #define MINEXTRACTBYTES 20 740Sstevel@tonic-gate #define MAXEXTRACTBYTES 1024 750Sstevel@tonic-gate #define PRNG_MAXOBLOCKS 1310720 /* Max output block per prng key */ 760Sstevel@tonic-gate #define TIMEOUT_INTERVAL 5 /* Periodic mixing interval in secs */ 770Sstevel@tonic-gate 780Sstevel@tonic-gate typedef enum extract_type { 790Sstevel@tonic-gate NONBLOCK_EXTRACT, 800Sstevel@tonic-gate BLOCKING_EXTRACT, 810Sstevel@tonic-gate ALWAYS_EXTRACT 820Sstevel@tonic-gate } extract_type_t; 830Sstevel@tonic-gate 840Sstevel@tonic-gate /* 850Sstevel@tonic-gate * Hash-algo generic definitions. For now, they are SHA1's. We use SHA1 860Sstevel@tonic-gate * routines directly instead of using k-API because we can't return any 870Sstevel@tonic-gate * error code in /dev/urandom case and we can get an error using k-API 880Sstevel@tonic-gate * if a mechanism is disabled. 890Sstevel@tonic-gate */ 900Sstevel@tonic-gate #define HASHSIZE 20 910Sstevel@tonic-gate #define HASH_CTX SHA1_CTX 920Sstevel@tonic-gate #define HashInit(ctx) SHA1Init((ctx)) 930Sstevel@tonic-gate #define HashUpdate(ctx, p, s) SHA1Update((ctx), (p), (s)) 940Sstevel@tonic-gate #define HashFinal(d, ctx) SHA1Final((d), (ctx)) 950Sstevel@tonic-gate 960Sstevel@tonic-gate /* HMAC-SHA1 */ 970Sstevel@tonic-gate #define HMAC_KEYSIZE 20 980Sstevel@tonic-gate #define HMAC_BLOCK_SIZE 64 990Sstevel@tonic-gate #define HMAC_KEYSCHED sha1keysched_t 1000Sstevel@tonic-gate #define SET_ENCRYPT_KEY(k, s, ks) hmac_key((k), (s), (ks)) 1010Sstevel@tonic-gate #define HMAC_ENCRYPT(ks, p, s, d) hmac_encr((ks), (uint8_t *)(p), s, d) 1020Sstevel@tonic-gate 1030Sstevel@tonic-gate /* HMAC-SHA1 "keyschedule" */ 1040Sstevel@tonic-gate typedef struct sha1keysched_s { 1050Sstevel@tonic-gate SHA1_CTX ictx; 1060Sstevel@tonic-gate SHA1_CTX octx; 1070Sstevel@tonic-gate } sha1keysched_t; 1080Sstevel@tonic-gate 1090Sstevel@tonic-gate /* 1100Sstevel@tonic-gate * Cache of random bytes implemented as a circular buffer. findex and rindex 1110Sstevel@tonic-gate * track the front and back of the circular buffer. 1120Sstevel@tonic-gate */ 1130Sstevel@tonic-gate uint8_t rndpool[RNDPOOLSIZE]; 1140Sstevel@tonic-gate static int findex, rindex; 1150Sstevel@tonic-gate static int rnbyte_cnt; /* Number of bytes in the cache */ 1160Sstevel@tonic-gate 1170Sstevel@tonic-gate static kmutex_t rndpool_lock; /* protects r/w accesses to the cache, */ 1180Sstevel@tonic-gate /* and the global variables */ 1190Sstevel@tonic-gate static kcondvar_t rndpool_read_cv; /* serializes poll/read syscalls */ 1200Sstevel@tonic-gate static int num_waiters; /* #threads waiting to read from /dev/random */ 1210Sstevel@tonic-gate 1220Sstevel@tonic-gate static struct pollhead rnd_pollhead; 1230Sstevel@tonic-gate static timeout_id_t kcf_rndtimeout_id; 1240Sstevel@tonic-gate static crypto_mech_type_t rngmech_type = CRYPTO_MECH_INVALID; 1250Sstevel@tonic-gate rnd_stats_t rnd_stats; 126*3096Skrishna static boolean_t rng_prov_found = B_TRUE; 127*3096Skrishna static boolean_t rng_ok_to_log = B_TRUE; 1280Sstevel@tonic-gate 1290Sstevel@tonic-gate static void rndc_addbytes(uint8_t *, size_t); 1300Sstevel@tonic-gate static void rndc_getbytes(uint8_t *ptr, size_t len); 1310Sstevel@tonic-gate static void rnd_handler(void *); 1320Sstevel@tonic-gate static void rnd_alloc_magazines(); 1330Sstevel@tonic-gate static void hmac_key(uint8_t *, size_t, void *); 1340Sstevel@tonic-gate static void hmac_encr(void *, uint8_t *, size_t, uint8_t *); 1350Sstevel@tonic-gate 1360Sstevel@tonic-gate 1370Sstevel@tonic-gate void 1380Sstevel@tonic-gate kcf_rnd_init() 1390Sstevel@tonic-gate { 1400Sstevel@tonic-gate hrtime_t ts; 1410Sstevel@tonic-gate time_t now; 1420Sstevel@tonic-gate 1430Sstevel@tonic-gate mutex_init(&rndpool_lock, NULL, MUTEX_DEFAULT, NULL); 1440Sstevel@tonic-gate cv_init(&rndpool_read_cv, NULL, CV_DEFAULT, NULL); 1450Sstevel@tonic-gate 1460Sstevel@tonic-gate /* 1470Sstevel@tonic-gate * Add bytes to the cache using 1480Sstevel@tonic-gate * . 2 unpredictable times: high resolution time since the boot-time, 1490Sstevel@tonic-gate * and the current time-of-the day. 1500Sstevel@tonic-gate * This is used only to make the timeout value in the timer 1510Sstevel@tonic-gate * unpredictable. 1520Sstevel@tonic-gate */ 1530Sstevel@tonic-gate ts = gethrtime(); 1540Sstevel@tonic-gate rndc_addbytes((uint8_t *)&ts, sizeof (ts)); 1550Sstevel@tonic-gate 1560Sstevel@tonic-gate (void) drv_getparm(TIME, &now); 1570Sstevel@tonic-gate rndc_addbytes((uint8_t *)&now, sizeof (now)); 1580Sstevel@tonic-gate 1590Sstevel@tonic-gate rnbyte_cnt = 0; 1600Sstevel@tonic-gate findex = rindex = 0; 1610Sstevel@tonic-gate num_waiters = 0; 1620Sstevel@tonic-gate rngmech_type = KCF_MECHID(KCF_MISC_CLASS, 0); 1630Sstevel@tonic-gate 1640Sstevel@tonic-gate rnd_alloc_magazines(); 1650Sstevel@tonic-gate } 1660Sstevel@tonic-gate 1670Sstevel@tonic-gate /* 1680Sstevel@tonic-gate * Return TRUE if at least one provider exists that can 1690Sstevel@tonic-gate * supply random numbers. 1700Sstevel@tonic-gate */ 1710Sstevel@tonic-gate boolean_t 1720Sstevel@tonic-gate kcf_rngprov_check(void) 1730Sstevel@tonic-gate { 1740Sstevel@tonic-gate int rv; 1750Sstevel@tonic-gate kcf_provider_desc_t *pd; 1760Sstevel@tonic-gate 1770Sstevel@tonic-gate if ((pd = kcf_get_mech_provider(rngmech_type, NULL, &rv, 1780Sstevel@tonic-gate NULL, CRYPTO_FG_RANDOM, B_FALSE, 0)) != NULL) { 1790Sstevel@tonic-gate KCF_PROV_REFRELE(pd); 180*3096Skrishna /* 181*3096Skrishna * We logged a warning once about no provider being available 182*3096Skrishna * and now a provider became available. So, set the flag so 183*3096Skrishna * that we can log again if the problem recurs. 184*3096Skrishna */ 185*3096Skrishna rng_ok_to_log = B_TRUE; 186*3096Skrishna rng_prov_found = B_TRUE; 1870Sstevel@tonic-gate return (B_TRUE); 188*3096Skrishna } else { 189*3096Skrishna rng_prov_found = B_FALSE; 1900Sstevel@tonic-gate return (B_FALSE); 191*3096Skrishna } 1920Sstevel@tonic-gate } 1930Sstevel@tonic-gate 1940Sstevel@tonic-gate /* 1950Sstevel@tonic-gate * Pick a software-based provider and submit a request to seed 1960Sstevel@tonic-gate * its random number generator. 1970Sstevel@tonic-gate */ 1980Sstevel@tonic-gate static void 1991920Smcpowers rngprov_seed(uint8_t *buf, int len, uint_t entropy_est, uint32_t flags) 2000Sstevel@tonic-gate { 2010Sstevel@tonic-gate kcf_provider_desc_t *pd = NULL; 2020Sstevel@tonic-gate 2030Sstevel@tonic-gate if (kcf_get_sw_prov(rngmech_type, &pd, B_FALSE) == CRYPTO_SUCCESS) { 2041920Smcpowers (void) KCF_PROV_SEED_RANDOM(pd, pd->pd_sid, buf, len, 2051920Smcpowers entropy_est, flags, NULL); 2060Sstevel@tonic-gate KCF_PROV_REFRELE(pd); 2070Sstevel@tonic-gate } 2080Sstevel@tonic-gate } 2090Sstevel@tonic-gate 2100Sstevel@tonic-gate /* Boot-time tunable for experimentation. */ 2110Sstevel@tonic-gate int kcf_limit_hwrng = 1; 2120Sstevel@tonic-gate 2130Sstevel@tonic-gate 2140Sstevel@tonic-gate /* 2150Sstevel@tonic-gate * This routine is called for blocking reads. 2160Sstevel@tonic-gate * 2170Sstevel@tonic-gate * The argument from_user_api indicates whether the caller is 2180Sstevel@tonic-gate * from userland coming via the /dev/random driver. 2190Sstevel@tonic-gate * 2200Sstevel@tonic-gate * The argument is_taskq_thr indicates whether the caller is 2210Sstevel@tonic-gate * the taskq thread dispatched by the timeout handler routine. 2220Sstevel@tonic-gate * In this case, we cycle through all the providers 2230Sstevel@tonic-gate * submitting a request to each provider to generate random numbers. 2240Sstevel@tonic-gate * 2250Sstevel@tonic-gate * For other cases, we pick a provider and submit a request to generate 2260Sstevel@tonic-gate * random numbers. We retry using another provider if we get an error. 2270Sstevel@tonic-gate * 2280Sstevel@tonic-gate * Returns the number of bytes that are written to 'ptr'. Returns -1 2290Sstevel@tonic-gate * if no provider is found. ptr and need are unchanged. 2300Sstevel@tonic-gate */ 2310Sstevel@tonic-gate static int 2320Sstevel@tonic-gate rngprov_getbytes(uint8_t *ptr, size_t need, boolean_t from_user_api, 2330Sstevel@tonic-gate boolean_t is_taskq_thr) 2340Sstevel@tonic-gate { 2350Sstevel@tonic-gate int rv; 2360Sstevel@tonic-gate int prov_cnt = 0; 2370Sstevel@tonic-gate int total_bytes = 0; 2380Sstevel@tonic-gate kcf_provider_desc_t *pd; 2390Sstevel@tonic-gate kcf_req_params_t params; 2400Sstevel@tonic-gate kcf_prov_tried_t *list = NULL; 2410Sstevel@tonic-gate 2420Sstevel@tonic-gate while ((pd = kcf_get_mech_provider(rngmech_type, NULL, &rv, 2430Sstevel@tonic-gate list, CRYPTO_FG_RANDOM, B_FALSE, 0)) != NULL) { 2440Sstevel@tonic-gate 2450Sstevel@tonic-gate prov_cnt++; 2460Sstevel@tonic-gate /* 2470Sstevel@tonic-gate * Typically a hardware RNG is a multi-purpose 2480Sstevel@tonic-gate * crypto card and hence we do not want to overload the card 2490Sstevel@tonic-gate * just for random numbers. The following check is to prevent 2500Sstevel@tonic-gate * a user process from hogging the hardware RNG. Note that we 2510Sstevel@tonic-gate * still use the hardware RNG from the periodically run 2520Sstevel@tonic-gate * taskq thread. 2530Sstevel@tonic-gate */ 2540Sstevel@tonic-gate if (pd->pd_prov_type == CRYPTO_HW_PROVIDER && from_user_api && 2550Sstevel@tonic-gate kcf_limit_hwrng == 1) { 2560Sstevel@tonic-gate ASSERT(is_taskq_thr == B_FALSE); 2570Sstevel@tonic-gate goto try_next; 2580Sstevel@tonic-gate } 2590Sstevel@tonic-gate 2600Sstevel@tonic-gate KCF_WRAP_RANDOM_OPS_PARAMS(¶ms, KCF_OP_RANDOM_GENERATE, 2611920Smcpowers pd->pd_sid, ptr, need, 0, 0); 2620Sstevel@tonic-gate rv = kcf_submit_request(pd, NULL, NULL, ¶ms, B_FALSE); 2630Sstevel@tonic-gate ASSERT(rv != CRYPTO_QUEUED); 2640Sstevel@tonic-gate 2650Sstevel@tonic-gate if (rv == CRYPTO_SUCCESS) { 2660Sstevel@tonic-gate total_bytes += need; 2670Sstevel@tonic-gate if (is_taskq_thr) 2680Sstevel@tonic-gate rndc_addbytes(ptr, need); 2690Sstevel@tonic-gate else { 2700Sstevel@tonic-gate KCF_PROV_REFRELE(pd); 2710Sstevel@tonic-gate break; 2720Sstevel@tonic-gate } 2730Sstevel@tonic-gate } 2740Sstevel@tonic-gate 2750Sstevel@tonic-gate if (is_taskq_thr || rv != CRYPTO_SUCCESS) { 2760Sstevel@tonic-gate try_next: 2770Sstevel@tonic-gate /* Add pd to the linked list of providers tried. */ 2780Sstevel@tonic-gate if (kcf_insert_triedlist(&list, pd, KM_SLEEP) == NULL) { 2790Sstevel@tonic-gate KCF_PROV_REFRELE(pd); 2800Sstevel@tonic-gate break; 2810Sstevel@tonic-gate } 2820Sstevel@tonic-gate } 2830Sstevel@tonic-gate 2840Sstevel@tonic-gate } 2850Sstevel@tonic-gate 2860Sstevel@tonic-gate if (list != NULL) 2870Sstevel@tonic-gate kcf_free_triedlist(list); 2880Sstevel@tonic-gate 2890Sstevel@tonic-gate if (prov_cnt == 0) { /* no provider could be found. */ 290*3096Skrishna rng_prov_found = B_FALSE; 2910Sstevel@tonic-gate return (-1); 292*3096Skrishna } else { 293*3096Skrishna rng_prov_found = B_TRUE; 294*3096Skrishna /* See comments in kcf_rngprov_check() */ 295*3096Skrishna rng_ok_to_log = B_TRUE; 2960Sstevel@tonic-gate } 2970Sstevel@tonic-gate 2980Sstevel@tonic-gate return (total_bytes); 2990Sstevel@tonic-gate } 3000Sstevel@tonic-gate 3010Sstevel@tonic-gate static void 3020Sstevel@tonic-gate notify_done(void *arg, int rv) 3030Sstevel@tonic-gate { 3040Sstevel@tonic-gate uchar_t *rndbuf = arg; 3050Sstevel@tonic-gate 3060Sstevel@tonic-gate if (rv == CRYPTO_SUCCESS) 3070Sstevel@tonic-gate rndc_addbytes(rndbuf, MINEXTRACTBYTES); 3080Sstevel@tonic-gate 3090Sstevel@tonic-gate bzero(rndbuf, MINEXTRACTBYTES); 3100Sstevel@tonic-gate kmem_free(rndbuf, MINEXTRACTBYTES); 3110Sstevel@tonic-gate } 3120Sstevel@tonic-gate 3130Sstevel@tonic-gate /* 3140Sstevel@tonic-gate * Cycle through all the providers submitting a request to each provider 3150Sstevel@tonic-gate * to generate random numbers. This is called for the modes - NONBLOCK_EXTRACT 3160Sstevel@tonic-gate * and ALWAYS_EXTRACT. 3170Sstevel@tonic-gate * 3180Sstevel@tonic-gate * Returns the number of bytes that are written to 'ptr'. Returns -1 3190Sstevel@tonic-gate * if no provider is found. ptr and len are unchanged. 3200Sstevel@tonic-gate */ 3210Sstevel@tonic-gate static int 3220Sstevel@tonic-gate rngprov_getbytes_nblk(uint8_t *ptr, size_t len, boolean_t from_user_api) 3230Sstevel@tonic-gate { 3240Sstevel@tonic-gate int rv, blen, total_bytes; 3250Sstevel@tonic-gate uchar_t *rndbuf; 3260Sstevel@tonic-gate kcf_provider_desc_t *pd; 3270Sstevel@tonic-gate kcf_req_params_t params; 3280Sstevel@tonic-gate crypto_call_req_t req; 3290Sstevel@tonic-gate kcf_prov_tried_t *list = NULL; 3300Sstevel@tonic-gate int prov_cnt = 0; 3310Sstevel@tonic-gate 3320Sstevel@tonic-gate blen = 0; 3330Sstevel@tonic-gate total_bytes = 0; 3340Sstevel@tonic-gate req.cr_flag = CRYPTO_SKIP_REQID; 3350Sstevel@tonic-gate req.cr_callback_func = notify_done; 3360Sstevel@tonic-gate 3370Sstevel@tonic-gate while ((pd = kcf_get_mech_provider(rngmech_type, NULL, &rv, 3380Sstevel@tonic-gate list, CRYPTO_FG_RANDOM, CHECK_RESTRICT(&req), 0)) != NULL) { 3390Sstevel@tonic-gate 3400Sstevel@tonic-gate prov_cnt ++; 3410Sstevel@tonic-gate switch (pd->pd_prov_type) { 3420Sstevel@tonic-gate case CRYPTO_HW_PROVIDER: 3430Sstevel@tonic-gate /* See comments in rngprov_getbytes() */ 3440Sstevel@tonic-gate if (from_user_api && kcf_limit_hwrng == 1) 3450Sstevel@tonic-gate goto try_next; 3460Sstevel@tonic-gate 3470Sstevel@tonic-gate /* 3480Sstevel@tonic-gate * We have to allocate a buffer here as we can not 3490Sstevel@tonic-gate * assume that the input buffer will remain valid 3500Sstevel@tonic-gate * when the callback comes. We use a fixed size buffer 3510Sstevel@tonic-gate * to simplify the book keeping. 3520Sstevel@tonic-gate */ 3530Sstevel@tonic-gate rndbuf = kmem_alloc(MINEXTRACTBYTES, KM_NOSLEEP); 3540Sstevel@tonic-gate if (rndbuf == NULL) { 3550Sstevel@tonic-gate KCF_PROV_REFRELE(pd); 3560Sstevel@tonic-gate if (list != NULL) 3570Sstevel@tonic-gate kcf_free_triedlist(list); 3580Sstevel@tonic-gate return (total_bytes); 3590Sstevel@tonic-gate } 3600Sstevel@tonic-gate req.cr_callback_arg = rndbuf; 3610Sstevel@tonic-gate KCF_WRAP_RANDOM_OPS_PARAMS(¶ms, 3620Sstevel@tonic-gate KCF_OP_RANDOM_GENERATE, 3631920Smcpowers pd->pd_sid, rndbuf, MINEXTRACTBYTES, 0, 0); 3640Sstevel@tonic-gate break; 3650Sstevel@tonic-gate 3660Sstevel@tonic-gate case CRYPTO_SW_PROVIDER: 3670Sstevel@tonic-gate /* 3680Sstevel@tonic-gate * We do not need to allocate a buffer in the software 3690Sstevel@tonic-gate * provider case as there is no callback involved. We 3700Sstevel@tonic-gate * avoid any extra data copy by directly passing 'ptr'. 3710Sstevel@tonic-gate */ 3720Sstevel@tonic-gate KCF_WRAP_RANDOM_OPS_PARAMS(¶ms, 3730Sstevel@tonic-gate KCF_OP_RANDOM_GENERATE, 3741920Smcpowers pd->pd_sid, ptr, len, 0, 0); 3750Sstevel@tonic-gate break; 3760Sstevel@tonic-gate } 3770Sstevel@tonic-gate 3780Sstevel@tonic-gate rv = kcf_submit_request(pd, NULL, &req, ¶ms, B_FALSE); 3790Sstevel@tonic-gate if (rv == CRYPTO_SUCCESS) { 3800Sstevel@tonic-gate switch (pd->pd_prov_type) { 3810Sstevel@tonic-gate case CRYPTO_HW_PROVIDER: 3820Sstevel@tonic-gate /* 3830Sstevel@tonic-gate * Since we have the input buffer handy, 3840Sstevel@tonic-gate * we directly copy to it rather than 3850Sstevel@tonic-gate * adding to the pool. 3860Sstevel@tonic-gate */ 3870Sstevel@tonic-gate blen = min(MINEXTRACTBYTES, len); 3880Sstevel@tonic-gate bcopy(rndbuf, ptr, blen); 3890Sstevel@tonic-gate if (len < MINEXTRACTBYTES) 3900Sstevel@tonic-gate rndc_addbytes(rndbuf + len, 3910Sstevel@tonic-gate MINEXTRACTBYTES - len); 3920Sstevel@tonic-gate ptr += blen; 3930Sstevel@tonic-gate len -= blen; 3940Sstevel@tonic-gate total_bytes += blen; 3950Sstevel@tonic-gate break; 3960Sstevel@tonic-gate 3970Sstevel@tonic-gate case CRYPTO_SW_PROVIDER: 3980Sstevel@tonic-gate total_bytes += len; 3990Sstevel@tonic-gate len = 0; 4000Sstevel@tonic-gate break; 4010Sstevel@tonic-gate } 4020Sstevel@tonic-gate } 4030Sstevel@tonic-gate 4040Sstevel@tonic-gate /* 4050Sstevel@tonic-gate * We free the buffer in the callback routine 4060Sstevel@tonic-gate * for the CRYPTO_QUEUED case. 4070Sstevel@tonic-gate */ 4080Sstevel@tonic-gate if (pd->pd_prov_type == CRYPTO_HW_PROVIDER && 4090Sstevel@tonic-gate rv != CRYPTO_QUEUED) { 4100Sstevel@tonic-gate bzero(rndbuf, MINEXTRACTBYTES); 4110Sstevel@tonic-gate kmem_free(rndbuf, MINEXTRACTBYTES); 4120Sstevel@tonic-gate } 4130Sstevel@tonic-gate 4140Sstevel@tonic-gate if (len == 0) { 4150Sstevel@tonic-gate KCF_PROV_REFRELE(pd); 4160Sstevel@tonic-gate break; 4170Sstevel@tonic-gate } 4180Sstevel@tonic-gate 4190Sstevel@tonic-gate if (rv != CRYPTO_SUCCESS) { 4200Sstevel@tonic-gate try_next: 4210Sstevel@tonic-gate /* Add pd to the linked list of providers tried. */ 4220Sstevel@tonic-gate if (kcf_insert_triedlist(&list, pd, KM_NOSLEEP) == 4230Sstevel@tonic-gate NULL) { 4240Sstevel@tonic-gate KCF_PROV_REFRELE(pd); 4250Sstevel@tonic-gate break; 4260Sstevel@tonic-gate } 4270Sstevel@tonic-gate } 4280Sstevel@tonic-gate } 4290Sstevel@tonic-gate 4300Sstevel@tonic-gate if (list != NULL) { 4310Sstevel@tonic-gate kcf_free_triedlist(list); 4320Sstevel@tonic-gate } 4330Sstevel@tonic-gate 4340Sstevel@tonic-gate if (prov_cnt == 0) { /* no provider could be found. */ 435*3096Skrishna rng_prov_found = B_FALSE; 4360Sstevel@tonic-gate return (-1); 437*3096Skrishna } else { 438*3096Skrishna rng_prov_found = B_TRUE; 439*3096Skrishna /* See comments in kcf_rngprov_check() */ 440*3096Skrishna rng_ok_to_log = B_TRUE; 4410Sstevel@tonic-gate } 4420Sstevel@tonic-gate 4430Sstevel@tonic-gate return (total_bytes); 4440Sstevel@tonic-gate } 4450Sstevel@tonic-gate 4460Sstevel@tonic-gate static void 4470Sstevel@tonic-gate rngprov_task(void *arg) 4480Sstevel@tonic-gate { 4490Sstevel@tonic-gate int len = (int)(uintptr_t)arg; 4500Sstevel@tonic-gate uchar_t tbuf[MAXEXTRACTBYTES]; 4510Sstevel@tonic-gate 4520Sstevel@tonic-gate ASSERT(len <= MAXEXTRACTBYTES); 453*3096Skrishna (void) rngprov_getbytes(tbuf, len, B_FALSE, B_TRUE); 4540Sstevel@tonic-gate } 4550Sstevel@tonic-gate 4560Sstevel@tonic-gate /* 4570Sstevel@tonic-gate * Returns "len" random or pseudo-random bytes in *ptr. 4580Sstevel@tonic-gate * Will block if not enough random bytes are available and the 4590Sstevel@tonic-gate * call is blocking. 4600Sstevel@tonic-gate * 4610Sstevel@tonic-gate * Called with rndpool_lock held (allowing caller to do optimistic locking; 4620Sstevel@tonic-gate * releases the lock before return). 4630Sstevel@tonic-gate */ 4640Sstevel@tonic-gate static int 4650Sstevel@tonic-gate rnd_get_bytes(uint8_t *ptr, size_t len, extract_type_t how, 4660Sstevel@tonic-gate boolean_t from_user_api) 4670Sstevel@tonic-gate { 4680Sstevel@tonic-gate int bytes; 4690Sstevel@tonic-gate size_t got; 4700Sstevel@tonic-gate 4710Sstevel@tonic-gate ASSERT(mutex_owned(&rndpool_lock)); 4720Sstevel@tonic-gate /* 4730Sstevel@tonic-gate * Check if the request can be satisfied from the cache 4740Sstevel@tonic-gate * of random bytes. 4750Sstevel@tonic-gate */ 4760Sstevel@tonic-gate if (len <= rnbyte_cnt) { 4770Sstevel@tonic-gate rndc_getbytes(ptr, len); 4780Sstevel@tonic-gate mutex_exit(&rndpool_lock); 4790Sstevel@tonic-gate return (0); 4800Sstevel@tonic-gate } 4810Sstevel@tonic-gate mutex_exit(&rndpool_lock); 4820Sstevel@tonic-gate 4830Sstevel@tonic-gate switch (how) { 4840Sstevel@tonic-gate case BLOCKING_EXTRACT: 4850Sstevel@tonic-gate if ((got = rngprov_getbytes(ptr, len, from_user_api, 4860Sstevel@tonic-gate B_FALSE)) == -1) 4870Sstevel@tonic-gate break; /* No provider found */ 4880Sstevel@tonic-gate 4890Sstevel@tonic-gate if (got == len) 4900Sstevel@tonic-gate return (0); 4910Sstevel@tonic-gate len -= got; 4920Sstevel@tonic-gate ptr += got; 4930Sstevel@tonic-gate break; 4940Sstevel@tonic-gate 4950Sstevel@tonic-gate case NONBLOCK_EXTRACT: 4960Sstevel@tonic-gate case ALWAYS_EXTRACT: 4970Sstevel@tonic-gate if ((got = rngprov_getbytes_nblk(ptr, len, 4980Sstevel@tonic-gate from_user_api)) == -1) { 4990Sstevel@tonic-gate /* No provider found */ 5000Sstevel@tonic-gate if (how == NONBLOCK_EXTRACT) { 5010Sstevel@tonic-gate return (EAGAIN); 5020Sstevel@tonic-gate } 5030Sstevel@tonic-gate } else { 5040Sstevel@tonic-gate if (got == len) 5050Sstevel@tonic-gate return (0); 5060Sstevel@tonic-gate len -= got; 5070Sstevel@tonic-gate ptr += got; 5080Sstevel@tonic-gate } 5090Sstevel@tonic-gate if (how == NONBLOCK_EXTRACT && (rnbyte_cnt < len)) 5100Sstevel@tonic-gate return (EAGAIN); 5110Sstevel@tonic-gate break; 5120Sstevel@tonic-gate } 5130Sstevel@tonic-gate 5140Sstevel@tonic-gate mutex_enter(&rndpool_lock); 5150Sstevel@tonic-gate while (len > 0) { 5160Sstevel@tonic-gate if (how == BLOCKING_EXTRACT) { 5170Sstevel@tonic-gate /* Check if there is enough */ 5180Sstevel@tonic-gate while (rnbyte_cnt < MINEXTRACTBYTES) { 5190Sstevel@tonic-gate num_waiters++; 5200Sstevel@tonic-gate if (cv_wait_sig(&rndpool_read_cv, 5210Sstevel@tonic-gate &rndpool_lock) == 0) { 5220Sstevel@tonic-gate num_waiters--; 5230Sstevel@tonic-gate mutex_exit(&rndpool_lock); 5240Sstevel@tonic-gate return (EINTR); 5250Sstevel@tonic-gate } 5260Sstevel@tonic-gate num_waiters--; 5270Sstevel@tonic-gate } 5280Sstevel@tonic-gate } 5290Sstevel@tonic-gate 5300Sstevel@tonic-gate /* Figure out how many bytes to extract */ 5310Sstevel@tonic-gate bytes = min(len, rnbyte_cnt); 5320Sstevel@tonic-gate rndc_getbytes(ptr, bytes); 5330Sstevel@tonic-gate 5340Sstevel@tonic-gate len -= bytes; 5350Sstevel@tonic-gate ptr += bytes; 5360Sstevel@tonic-gate 5370Sstevel@tonic-gate if (len > 0 && how == ALWAYS_EXTRACT) { 5380Sstevel@tonic-gate /* 5390Sstevel@tonic-gate * There are not enough bytes, but we can not block. 5400Sstevel@tonic-gate * This only happens in the case of /dev/urandom which 5410Sstevel@tonic-gate * runs an additional generation algorithm. So, there 5420Sstevel@tonic-gate * is no problem. 5430Sstevel@tonic-gate */ 5440Sstevel@tonic-gate while (len > 0) { 5450Sstevel@tonic-gate *ptr = rndpool[findex]; 5460Sstevel@tonic-gate ptr++; len--; 5470Sstevel@tonic-gate rindex = findex = (findex + 1) & 5480Sstevel@tonic-gate (RNDPOOLSIZE - 1); 5490Sstevel@tonic-gate } 5500Sstevel@tonic-gate break; 5510Sstevel@tonic-gate } 5520Sstevel@tonic-gate } 5530Sstevel@tonic-gate 5540Sstevel@tonic-gate mutex_exit(&rndpool_lock); 5550Sstevel@tonic-gate return (0); 5560Sstevel@tonic-gate } 5570Sstevel@tonic-gate 5580Sstevel@tonic-gate int 5590Sstevel@tonic-gate kcf_rnd_get_bytes(uint8_t *ptr, size_t len, boolean_t noblock, 5600Sstevel@tonic-gate boolean_t from_user_api) 5610Sstevel@tonic-gate { 5620Sstevel@tonic-gate extract_type_t how; 5630Sstevel@tonic-gate int error; 5640Sstevel@tonic-gate 5650Sstevel@tonic-gate how = noblock ? NONBLOCK_EXTRACT : BLOCKING_EXTRACT; 5660Sstevel@tonic-gate mutex_enter(&rndpool_lock); 5670Sstevel@tonic-gate if ((error = rnd_get_bytes(ptr, len, how, from_user_api)) != 0) 5680Sstevel@tonic-gate return (error); 5690Sstevel@tonic-gate 5700Sstevel@tonic-gate BUMP_RND_STATS(rs_rndOut, len); 5710Sstevel@tonic-gate return (0); 5720Sstevel@tonic-gate } 5730Sstevel@tonic-gate 5740Sstevel@tonic-gate /* 5750Sstevel@tonic-gate * Revisit this if the structs grow or we come up with a better way 5760Sstevel@tonic-gate * of cache-line-padding structures. 5770Sstevel@tonic-gate */ 5780Sstevel@tonic-gate #define RND_CPU_CACHE_SIZE 64 5790Sstevel@tonic-gate #define RND_CPU_PAD_SIZE RND_CPU_CACHE_SIZE*5 5800Sstevel@tonic-gate #define RND_CPU_PAD (RND_CPU_PAD_SIZE - \ 5810Sstevel@tonic-gate (sizeof (kmutex_t) + 3*sizeof (uint8_t *) + sizeof (HMAC_KEYSCHED) + \ 5820Sstevel@tonic-gate sizeof (uint64_t) + 3*sizeof (uint32_t) + sizeof (rnd_stats_t))) 5830Sstevel@tonic-gate 5840Sstevel@tonic-gate /* 5850Sstevel@tonic-gate * Per-CPU random state. Somewhat like like kmem's magazines, this provides 5860Sstevel@tonic-gate * a per-CPU instance of the pseudo-random generator. We have it much easier 5870Sstevel@tonic-gate * than kmem, as we can afford to "leak" random bits if a CPU is DR'ed out. 5880Sstevel@tonic-gate * 5890Sstevel@tonic-gate * Note that this usage is preemption-safe; a thread 5900Sstevel@tonic-gate * entering a critical section remembers which generator it locked 5910Sstevel@tonic-gate * and unlocks the same one; should it be preempted and wind up running on 5920Sstevel@tonic-gate * a different CPU, there will be a brief period of increased contention 5930Sstevel@tonic-gate * before it exits the critical section but nothing will melt. 5940Sstevel@tonic-gate */ 5950Sstevel@tonic-gate typedef struct rndmag_s 5960Sstevel@tonic-gate { 5970Sstevel@tonic-gate kmutex_t rm_lock; 5980Sstevel@tonic-gate uint8_t *rm_buffer; /* Start of buffer */ 5990Sstevel@tonic-gate uint8_t *rm_eptr; /* End of buffer */ 6000Sstevel@tonic-gate uint8_t *rm_rptr; /* Current read pointer */ 6010Sstevel@tonic-gate HMAC_KEYSCHED rm_ks; /* seed */ 6020Sstevel@tonic-gate uint64_t rm_counter; /* rotating counter for extracting */ 6030Sstevel@tonic-gate uint32_t rm_oblocks; /* time to rekey? */ 6040Sstevel@tonic-gate uint32_t rm_ofuzz; /* Rekey backoff state */ 6050Sstevel@tonic-gate uint32_t rm_olimit; /* Hard rekey limit */ 6060Sstevel@tonic-gate rnd_stats_t rm_stats; /* Per-CPU Statistics */ 6070Sstevel@tonic-gate uint8_t rm_pad[RND_CPU_PAD]; 6080Sstevel@tonic-gate } rndmag_t; 6090Sstevel@tonic-gate 6100Sstevel@tonic-gate /* 6110Sstevel@tonic-gate * Generate random bytes for /dev/urandom by encrypting a 6120Sstevel@tonic-gate * rotating counter with a key created from bytes extracted 6130Sstevel@tonic-gate * from the pool. A maximum of PRNG_MAXOBLOCKS output blocks 6140Sstevel@tonic-gate * is generated before a new key is obtained. 6150Sstevel@tonic-gate * 6160Sstevel@tonic-gate * Note that callers to this routine are likely to assume it can't fail. 6170Sstevel@tonic-gate * 6180Sstevel@tonic-gate * Called with rmp locked; releases lock. 6190Sstevel@tonic-gate */ 6200Sstevel@tonic-gate static int 6210Sstevel@tonic-gate rnd_generate_pseudo_bytes(rndmag_t *rmp, uint8_t *ptr, size_t len) 6220Sstevel@tonic-gate { 6230Sstevel@tonic-gate size_t bytes = len; 6240Sstevel@tonic-gate int nblock, size; 6250Sstevel@tonic-gate uint32_t oblocks; 6260Sstevel@tonic-gate uint8_t digest[HASHSIZE]; 6270Sstevel@tonic-gate 6280Sstevel@tonic-gate ASSERT(mutex_owned(&rmp->rm_lock)); 6290Sstevel@tonic-gate 6300Sstevel@tonic-gate /* Nothing is being asked */ 6310Sstevel@tonic-gate if (len == 0) { 6320Sstevel@tonic-gate mutex_exit(&rmp->rm_lock); 6330Sstevel@tonic-gate return (0); 6340Sstevel@tonic-gate } 6350Sstevel@tonic-gate 6360Sstevel@tonic-gate nblock = howmany(len, HASHSIZE); 6370Sstevel@tonic-gate 6380Sstevel@tonic-gate rmp->rm_oblocks += nblock; 6390Sstevel@tonic-gate oblocks = rmp->rm_oblocks; 6400Sstevel@tonic-gate 6410Sstevel@tonic-gate do { 6420Sstevel@tonic-gate if (oblocks >= rmp->rm_olimit) { 6430Sstevel@tonic-gate hrtime_t timestamp; 6440Sstevel@tonic-gate uint8_t key[HMAC_KEYSIZE]; 6450Sstevel@tonic-gate 6460Sstevel@tonic-gate /* 6470Sstevel@tonic-gate * Contention-avoiding rekey: see if 6480Sstevel@tonic-gate * the pool is locked, and if so, wait a bit. 6490Sstevel@tonic-gate * Do an 'exponential back-in' to ensure we don't 6500Sstevel@tonic-gate * run too long without rekey. 6510Sstevel@tonic-gate */ 6520Sstevel@tonic-gate if (rmp->rm_ofuzz) { 6530Sstevel@tonic-gate /* 6540Sstevel@tonic-gate * Decaying exponential back-in for rekey. 6550Sstevel@tonic-gate */ 6560Sstevel@tonic-gate if ((rnbyte_cnt < MINEXTRACTBYTES) || 6570Sstevel@tonic-gate (!mutex_tryenter(&rndpool_lock))) { 6580Sstevel@tonic-gate rmp->rm_olimit += rmp->rm_ofuzz; 6590Sstevel@tonic-gate rmp->rm_ofuzz >>= 1; 6600Sstevel@tonic-gate goto punt; 6610Sstevel@tonic-gate } 6620Sstevel@tonic-gate } else { 6630Sstevel@tonic-gate mutex_enter(&rndpool_lock); 6640Sstevel@tonic-gate } 6650Sstevel@tonic-gate 6660Sstevel@tonic-gate /* Get a new chunk of entropy */ 6670Sstevel@tonic-gate (void) rnd_get_bytes(key, HMAC_KEYSIZE, 6680Sstevel@tonic-gate ALWAYS_EXTRACT, B_FALSE); 6690Sstevel@tonic-gate 6700Sstevel@tonic-gate /* Set up key */ 6710Sstevel@tonic-gate SET_ENCRYPT_KEY(key, HMAC_KEYSIZE, &rmp->rm_ks); 6720Sstevel@tonic-gate 6730Sstevel@tonic-gate /* Get new counter value by encrypting timestamp */ 6740Sstevel@tonic-gate timestamp = gethrtime(); 6750Sstevel@tonic-gate HMAC_ENCRYPT(&rmp->rm_ks, ×tamp, 6760Sstevel@tonic-gate sizeof (timestamp), digest); 6770Sstevel@tonic-gate rmp->rm_olimit = PRNG_MAXOBLOCKS/2; 6780Sstevel@tonic-gate rmp->rm_ofuzz = PRNG_MAXOBLOCKS/4; 6790Sstevel@tonic-gate bcopy(digest, &rmp->rm_counter, sizeof (uint64_t)); 6800Sstevel@tonic-gate oblocks = 0; 6810Sstevel@tonic-gate rmp->rm_oblocks = nblock; 6820Sstevel@tonic-gate } 6830Sstevel@tonic-gate punt: 6840Sstevel@tonic-gate /* Hash counter to produce prn stream */ 6850Sstevel@tonic-gate if (bytes >= HASHSIZE) { 6860Sstevel@tonic-gate size = HASHSIZE; 6870Sstevel@tonic-gate HMAC_ENCRYPT(&rmp->rm_ks, &rmp->rm_counter, 6880Sstevel@tonic-gate sizeof (rmp->rm_counter), ptr); 6890Sstevel@tonic-gate } else { 6900Sstevel@tonic-gate size = min(bytes, HASHSIZE); 6910Sstevel@tonic-gate HMAC_ENCRYPT(&rmp->rm_ks, &rmp->rm_counter, 6920Sstevel@tonic-gate sizeof (rmp->rm_counter), digest); 6930Sstevel@tonic-gate bcopy(digest, ptr, size); 6940Sstevel@tonic-gate } 6950Sstevel@tonic-gate ptr += size; 6960Sstevel@tonic-gate bytes -= size; 6970Sstevel@tonic-gate rmp->rm_counter++; 6980Sstevel@tonic-gate oblocks++; 6990Sstevel@tonic-gate nblock--; 7000Sstevel@tonic-gate } while (bytes > 0); 7010Sstevel@tonic-gate 7020Sstevel@tonic-gate mutex_exit(&rmp->rm_lock); 7030Sstevel@tonic-gate return (0); 7040Sstevel@tonic-gate } 7050Sstevel@tonic-gate 7060Sstevel@tonic-gate /* 7070Sstevel@tonic-gate * Per-CPU Random magazines. 7080Sstevel@tonic-gate */ 7090Sstevel@tonic-gate static rndmag_t *rndmag; 7100Sstevel@tonic-gate static uint8_t *rndbuf; 7110Sstevel@tonic-gate static size_t rndmag_total; 7120Sstevel@tonic-gate /* 7130Sstevel@tonic-gate * common/os/cpu.c says that platform support code can shrinkwrap 7140Sstevel@tonic-gate * max_ncpus. On the off chance that we get loaded very early, we 7150Sstevel@tonic-gate * read it exactly once, to copy it here. 7160Sstevel@tonic-gate */ 7170Sstevel@tonic-gate static uint32_t random_max_ncpus = 0; 7180Sstevel@tonic-gate 7190Sstevel@tonic-gate /* 7200Sstevel@tonic-gate * Boot-time tunables, for experimentation. 7210Sstevel@tonic-gate */ 7221070Skais size_t rndmag_threshold = 2560; 7231070Skais size_t rndbuf_len = 5120; 724445Skrishna size_t rndmag_size = 1280; 7250Sstevel@tonic-gate 7260Sstevel@tonic-gate 7270Sstevel@tonic-gate int 7280Sstevel@tonic-gate kcf_rnd_get_pseudo_bytes(uint8_t *ptr, size_t len) 7290Sstevel@tonic-gate { 7300Sstevel@tonic-gate rndmag_t *rmp; 7310Sstevel@tonic-gate uint8_t *cptr, *eptr; 7320Sstevel@tonic-gate 7330Sstevel@tonic-gate /* 7340Sstevel@tonic-gate * Anyone who asks for zero bytes of randomness should get slapped. 7350Sstevel@tonic-gate */ 7360Sstevel@tonic-gate ASSERT(len > 0); 7370Sstevel@tonic-gate 7380Sstevel@tonic-gate /* 7390Sstevel@tonic-gate * Fast path. 7400Sstevel@tonic-gate */ 7410Sstevel@tonic-gate for (;;) { 7420Sstevel@tonic-gate rmp = &rndmag[CPU->cpu_seqid]; 7430Sstevel@tonic-gate mutex_enter(&rmp->rm_lock); 7440Sstevel@tonic-gate 7450Sstevel@tonic-gate /* 7460Sstevel@tonic-gate * Big requests bypass buffer and tail-call the 7470Sstevel@tonic-gate * generate routine directly. 7480Sstevel@tonic-gate */ 7490Sstevel@tonic-gate if (len > rndmag_threshold) { 7500Sstevel@tonic-gate BUMP_CPU_RND_STATS(rmp, rs_urndOut, len); 7510Sstevel@tonic-gate return (rnd_generate_pseudo_bytes(rmp, ptr, len)); 7520Sstevel@tonic-gate } 7530Sstevel@tonic-gate 7540Sstevel@tonic-gate cptr = rmp->rm_rptr; 7550Sstevel@tonic-gate eptr = cptr + len; 7560Sstevel@tonic-gate 7570Sstevel@tonic-gate if (eptr <= rmp->rm_eptr) { 7580Sstevel@tonic-gate rmp->rm_rptr = eptr; 7590Sstevel@tonic-gate bcopy(cptr, ptr, len); 7600Sstevel@tonic-gate BUMP_CPU_RND_STATS(rmp, rs_urndOut, len); 7610Sstevel@tonic-gate mutex_exit(&rmp->rm_lock); 7620Sstevel@tonic-gate 7630Sstevel@tonic-gate return (0); 7640Sstevel@tonic-gate } 7650Sstevel@tonic-gate /* 7660Sstevel@tonic-gate * End fast path. 7670Sstevel@tonic-gate */ 7680Sstevel@tonic-gate rmp->rm_rptr = rmp->rm_buffer; 7690Sstevel@tonic-gate /* 7700Sstevel@tonic-gate * Note: We assume the generate routine always succeeds 7710Sstevel@tonic-gate * in this case (because it does at present..) 7720Sstevel@tonic-gate * It also always releases rm_lock. 7730Sstevel@tonic-gate */ 7740Sstevel@tonic-gate (void) rnd_generate_pseudo_bytes(rmp, rmp->rm_buffer, 7750Sstevel@tonic-gate rndbuf_len); 7760Sstevel@tonic-gate } 7770Sstevel@tonic-gate } 7780Sstevel@tonic-gate 7790Sstevel@tonic-gate /* 7800Sstevel@tonic-gate * We set up (empty) magazines for all of max_ncpus, possibly wasting a 7810Sstevel@tonic-gate * little memory on big systems that don't have the full set installed. 7820Sstevel@tonic-gate * See above; "empty" means "rptr equal to eptr"; this will trigger the 7830Sstevel@tonic-gate * refill path in rnd_get_pseudo_bytes above on the first call for each CPU. 7840Sstevel@tonic-gate * 7850Sstevel@tonic-gate * TODO: make rndmag_size tunable at run time! 7860Sstevel@tonic-gate */ 7870Sstevel@tonic-gate static void 7880Sstevel@tonic-gate rnd_alloc_magazines() 7890Sstevel@tonic-gate { 7900Sstevel@tonic-gate rndmag_t *rmp; 7910Sstevel@tonic-gate int i; 7920Sstevel@tonic-gate 7930Sstevel@tonic-gate rndbuf_len = roundup(rndbuf_len, HASHSIZE); 7940Sstevel@tonic-gate if (rndmag_size < rndbuf_len) 7950Sstevel@tonic-gate rndmag_size = rndbuf_len; 7960Sstevel@tonic-gate rndmag_size = roundup(rndmag_size, RND_CPU_CACHE_SIZE); 7970Sstevel@tonic-gate 7980Sstevel@tonic-gate random_max_ncpus = max_ncpus; 7990Sstevel@tonic-gate rndmag_total = rndmag_size * random_max_ncpus; 8000Sstevel@tonic-gate 8010Sstevel@tonic-gate rndbuf = kmem_alloc(rndmag_total, KM_SLEEP); 8020Sstevel@tonic-gate rndmag = kmem_zalloc(sizeof (rndmag_t) * random_max_ncpus, KM_SLEEP); 8030Sstevel@tonic-gate 8040Sstevel@tonic-gate for (i = 0; i < random_max_ncpus; i++) { 8050Sstevel@tonic-gate uint8_t *buf; 8060Sstevel@tonic-gate 8070Sstevel@tonic-gate rmp = &rndmag[i]; 8080Sstevel@tonic-gate mutex_init(&rmp->rm_lock, NULL, MUTEX_DRIVER, NULL); 8090Sstevel@tonic-gate 8100Sstevel@tonic-gate buf = rndbuf + i * rndmag_size; 8110Sstevel@tonic-gate 8120Sstevel@tonic-gate rmp->rm_buffer = buf; 8130Sstevel@tonic-gate rmp->rm_eptr = buf + rndbuf_len; 8140Sstevel@tonic-gate rmp->rm_rptr = buf + rndbuf_len; 8150Sstevel@tonic-gate rmp->rm_oblocks = 1; 8160Sstevel@tonic-gate } 8170Sstevel@tonic-gate } 8180Sstevel@tonic-gate 8190Sstevel@tonic-gate void 8200Sstevel@tonic-gate kcf_rnd_schedule_timeout(boolean_t do_mech2id) 8210Sstevel@tonic-gate { 8220Sstevel@tonic-gate clock_t ut; /* time in microseconds */ 8230Sstevel@tonic-gate 8240Sstevel@tonic-gate if (do_mech2id) 8250Sstevel@tonic-gate rngmech_type = crypto_mech2id(SUN_RANDOM); 8260Sstevel@tonic-gate 8270Sstevel@tonic-gate /* 8280Sstevel@tonic-gate * The new timeout value is taken from the buffer of random bytes. 8290Sstevel@tonic-gate * We're merely reading the first 32 bits from the buffer here, not 8300Sstevel@tonic-gate * consuming any random bytes. 8310Sstevel@tonic-gate * The timeout multiplier value is a random value between 0.5 sec and 8320Sstevel@tonic-gate * 1.544480 sec (0.5 sec + 0xFF000 microseconds). 8330Sstevel@tonic-gate * The new timeout is TIMEOUT_INTERVAL times that multiplier. 8340Sstevel@tonic-gate */ 8350Sstevel@tonic-gate ut = 500000 + (clock_t)((((uint32_t)rndpool[findex]) << 12) & 0xFF000); 8360Sstevel@tonic-gate kcf_rndtimeout_id = timeout(rnd_handler, NULL, 8370Sstevel@tonic-gate TIMEOUT_INTERVAL * drv_usectohz(ut)); 8380Sstevel@tonic-gate } 8390Sstevel@tonic-gate 8400Sstevel@tonic-gate /* 8410Sstevel@tonic-gate * &rnd_pollhead is passed in *phpp in order to indicate the calling thread 8420Sstevel@tonic-gate * will block. When enough random bytes are available, later, the timeout 8430Sstevel@tonic-gate * handler routine will issue the pollwakeup() calls. 8440Sstevel@tonic-gate */ 8450Sstevel@tonic-gate void 8460Sstevel@tonic-gate kcf_rnd_chpoll(int anyyet, short *reventsp, struct pollhead **phpp) 8470Sstevel@tonic-gate { 8480Sstevel@tonic-gate /* 8490Sstevel@tonic-gate * Sampling of rnbyte_cnt is an atomic 8500Sstevel@tonic-gate * operation. Hence we do not need any locking. 8510Sstevel@tonic-gate */ 8520Sstevel@tonic-gate if (rnbyte_cnt >= MINEXTRACTBYTES) { 8530Sstevel@tonic-gate *reventsp |= (POLLIN | POLLRDNORM); 8540Sstevel@tonic-gate } else { 8550Sstevel@tonic-gate *reventsp = 0; 8560Sstevel@tonic-gate if (!anyyet) 8570Sstevel@tonic-gate *phpp = &rnd_pollhead; 8580Sstevel@tonic-gate } 8590Sstevel@tonic-gate } 8600Sstevel@tonic-gate 8610Sstevel@tonic-gate /*ARGSUSED*/ 8620Sstevel@tonic-gate static void 8630Sstevel@tonic-gate rnd_handler(void *arg) 8640Sstevel@tonic-gate { 8650Sstevel@tonic-gate int len = 0; 8660Sstevel@tonic-gate 867*3096Skrishna if (!rng_prov_found && rng_ok_to_log) { 868*3096Skrishna cmn_err(CE_WARN, "No randomness provider enabled for " 869*3096Skrishna "/dev/random. Use cryptoadm(1M) to enable a provider."); 870*3096Skrishna rng_ok_to_log = B_FALSE; 871*3096Skrishna } 872*3096Skrishna 8730Sstevel@tonic-gate if (num_waiters > 0) 8740Sstevel@tonic-gate len = MAXEXTRACTBYTES; 8750Sstevel@tonic-gate else if (rnbyte_cnt < RNDPOOLSIZE) 8760Sstevel@tonic-gate len = MINEXTRACTBYTES; 8770Sstevel@tonic-gate 8780Sstevel@tonic-gate if (len > 0) { 8790Sstevel@tonic-gate (void) taskq_dispatch(system_taskq, rngprov_task, 8800Sstevel@tonic-gate (void *)(uintptr_t)len, TQ_NOSLEEP); 8810Sstevel@tonic-gate } 8820Sstevel@tonic-gate 8830Sstevel@tonic-gate mutex_enter(&rndpool_lock); 8840Sstevel@tonic-gate /* 8850Sstevel@tonic-gate * Wake up threads waiting in poll() or for enough accumulated 8860Sstevel@tonic-gate * random bytes to read from /dev/random. In case a poll() is 8870Sstevel@tonic-gate * concurrent with a read(), the polling process may be woken up 8880Sstevel@tonic-gate * indicating that enough randomness is now available for reading, 8890Sstevel@tonic-gate * and another process *steals* the bits from the pool, causing the 8900Sstevel@tonic-gate * subsequent read() from the first process to block. It is acceptable 8910Sstevel@tonic-gate * since the blocking will eventually end, after the timeout 8920Sstevel@tonic-gate * has expired enough times to honor the read. 8930Sstevel@tonic-gate * 8940Sstevel@tonic-gate * Note - Since we hold the rndpool_lock across the pollwakeup() call 8950Sstevel@tonic-gate * we MUST NOT grab the rndpool_lock in kcf_rndchpoll(). 8960Sstevel@tonic-gate */ 8970Sstevel@tonic-gate if (rnbyte_cnt >= MINEXTRACTBYTES) 8980Sstevel@tonic-gate pollwakeup(&rnd_pollhead, POLLIN | POLLRDNORM); 8990Sstevel@tonic-gate 9000Sstevel@tonic-gate if (num_waiters > 0) 9010Sstevel@tonic-gate cv_broadcast(&rndpool_read_cv); 9020Sstevel@tonic-gate mutex_exit(&rndpool_lock); 9030Sstevel@tonic-gate 9040Sstevel@tonic-gate kcf_rnd_schedule_timeout(B_FALSE); 9050Sstevel@tonic-gate } 9060Sstevel@tonic-gate 9070Sstevel@tonic-gate /* Hashing functions */ 9080Sstevel@tonic-gate 9090Sstevel@tonic-gate static void 9100Sstevel@tonic-gate hmac_key(uint8_t *key, size_t keylen, void *buf) 9110Sstevel@tonic-gate { 9120Sstevel@tonic-gate uint32_t *ip, *op; 9130Sstevel@tonic-gate uint32_t ipad[HMAC_BLOCK_SIZE/sizeof (uint32_t)]; 9140Sstevel@tonic-gate uint32_t opad[HMAC_BLOCK_SIZE/sizeof (uint32_t)]; 9150Sstevel@tonic-gate HASH_CTX *icontext, *ocontext; 9160Sstevel@tonic-gate int i; 9170Sstevel@tonic-gate int nints; 9180Sstevel@tonic-gate 9190Sstevel@tonic-gate icontext = buf; 9200Sstevel@tonic-gate ocontext = (SHA1_CTX *)((uint8_t *)buf + sizeof (HASH_CTX)); 9210Sstevel@tonic-gate 9220Sstevel@tonic-gate bzero((uchar_t *)ipad, HMAC_BLOCK_SIZE); 9230Sstevel@tonic-gate bzero((uchar_t *)opad, HMAC_BLOCK_SIZE); 9240Sstevel@tonic-gate bcopy(key, (uchar_t *)ipad, keylen); 9250Sstevel@tonic-gate bcopy(key, (uchar_t *)opad, keylen); 9260Sstevel@tonic-gate 9270Sstevel@tonic-gate /* 9280Sstevel@tonic-gate * XOR key with ipad (0x36) and opad (0x5c) as defined 9290Sstevel@tonic-gate * in RFC 2104. 9300Sstevel@tonic-gate */ 9310Sstevel@tonic-gate ip = ipad; 9320Sstevel@tonic-gate op = opad; 9330Sstevel@tonic-gate nints = HMAC_BLOCK_SIZE/sizeof (uint32_t); 9340Sstevel@tonic-gate 9350Sstevel@tonic-gate for (i = 0; i < nints; i++) { 9360Sstevel@tonic-gate ip[i] ^= 0x36363636; 9370Sstevel@tonic-gate op[i] ^= 0x5c5c5c5c; 9380Sstevel@tonic-gate } 9390Sstevel@tonic-gate 9400Sstevel@tonic-gate /* Perform hash with ipad */ 9410Sstevel@tonic-gate HashInit(icontext); 9420Sstevel@tonic-gate HashUpdate(icontext, (uchar_t *)ipad, HMAC_BLOCK_SIZE); 9430Sstevel@tonic-gate 9440Sstevel@tonic-gate /* Perform hash with opad */ 9450Sstevel@tonic-gate HashInit(ocontext); 9460Sstevel@tonic-gate HashUpdate(ocontext, (uchar_t *)opad, HMAC_BLOCK_SIZE); 9470Sstevel@tonic-gate } 9480Sstevel@tonic-gate 9490Sstevel@tonic-gate static void 9500Sstevel@tonic-gate hmac_encr(void *ctx, uint8_t *ptr, size_t len, uint8_t *digest) 9510Sstevel@tonic-gate { 9520Sstevel@tonic-gate HASH_CTX *saved_contexts; 9530Sstevel@tonic-gate HASH_CTX icontext; 9540Sstevel@tonic-gate HASH_CTX ocontext; 9550Sstevel@tonic-gate 9560Sstevel@tonic-gate saved_contexts = (HASH_CTX *)ctx; 9570Sstevel@tonic-gate icontext = saved_contexts[0]; 9580Sstevel@tonic-gate ocontext = saved_contexts[1]; 9590Sstevel@tonic-gate 9600Sstevel@tonic-gate HashUpdate(&icontext, ptr, len); 9610Sstevel@tonic-gate HashFinal(digest, &icontext); 9620Sstevel@tonic-gate 9630Sstevel@tonic-gate /* 9640Sstevel@tonic-gate * Perform Hash(K XOR OPAD, DIGEST), where DIGEST is the 9650Sstevel@tonic-gate * Hash(K XOR IPAD, DATA). 9660Sstevel@tonic-gate */ 9670Sstevel@tonic-gate HashUpdate(&ocontext, digest, HASHSIZE); 9680Sstevel@tonic-gate HashFinal(digest, &ocontext); 9690Sstevel@tonic-gate } 9700Sstevel@tonic-gate 9710Sstevel@tonic-gate 9720Sstevel@tonic-gate static void 9730Sstevel@tonic-gate rndc_addbytes(uint8_t *ptr, size_t len) 9740Sstevel@tonic-gate { 9750Sstevel@tonic-gate ASSERT(ptr != NULL && len > 0); 9760Sstevel@tonic-gate ASSERT(rnbyte_cnt <= RNDPOOLSIZE); 9770Sstevel@tonic-gate 9780Sstevel@tonic-gate mutex_enter(&rndpool_lock); 9790Sstevel@tonic-gate while ((len > 0) && (rnbyte_cnt < RNDPOOLSIZE)) { 9800Sstevel@tonic-gate rndpool[rindex] ^= *ptr; 9810Sstevel@tonic-gate ptr++; len--; 9820Sstevel@tonic-gate rindex = (rindex + 1) & (RNDPOOLSIZE - 1); 9830Sstevel@tonic-gate rnbyte_cnt++; 9840Sstevel@tonic-gate } 9850Sstevel@tonic-gate 9860Sstevel@tonic-gate /* Handle buffer full case */ 9870Sstevel@tonic-gate while (len > 0) { 9880Sstevel@tonic-gate rndpool[rindex] ^= *ptr; 9890Sstevel@tonic-gate ptr++; len--; 9900Sstevel@tonic-gate findex = rindex = (rindex + 1) & (RNDPOOLSIZE - 1); 9910Sstevel@tonic-gate } 9920Sstevel@tonic-gate mutex_exit(&rndpool_lock); 9930Sstevel@tonic-gate } 9940Sstevel@tonic-gate 9950Sstevel@tonic-gate /* 9960Sstevel@tonic-gate * Caller should check len <= rnbyte_cnt under the 9970Sstevel@tonic-gate * rndpool_lock before calling. 9980Sstevel@tonic-gate */ 9990Sstevel@tonic-gate static void 10000Sstevel@tonic-gate rndc_getbytes(uint8_t *ptr, size_t len) 10010Sstevel@tonic-gate { 10020Sstevel@tonic-gate ASSERT(MUTEX_HELD(&rndpool_lock)); 10030Sstevel@tonic-gate ASSERT(len <= rnbyte_cnt && rnbyte_cnt <= RNDPOOLSIZE); 10040Sstevel@tonic-gate 10050Sstevel@tonic-gate BUMP_RND_STATS(rs_rndcOut, len); 10060Sstevel@tonic-gate 10070Sstevel@tonic-gate while (len > 0) { 10080Sstevel@tonic-gate *ptr = rndpool[findex]; 10090Sstevel@tonic-gate ptr++; len--; 10100Sstevel@tonic-gate findex = (findex + 1) & (RNDPOOLSIZE - 1); 10110Sstevel@tonic-gate rnbyte_cnt--; 10120Sstevel@tonic-gate } 10130Sstevel@tonic-gate } 10140Sstevel@tonic-gate 10150Sstevel@tonic-gate /* Random number exported entry points */ 10160Sstevel@tonic-gate 10170Sstevel@tonic-gate /* 10180Sstevel@tonic-gate * Mix the supplied bytes into the entropy pool of a kCF 10190Sstevel@tonic-gate * RNG provider. 10200Sstevel@tonic-gate */ 10210Sstevel@tonic-gate int 10221920Smcpowers random_add_pseudo_entropy(uint8_t *ptr, size_t len, uint_t entropy_est) 10230Sstevel@tonic-gate { 10240Sstevel@tonic-gate if (len < 1) 10250Sstevel@tonic-gate return (-1); 10260Sstevel@tonic-gate 10271920Smcpowers rngprov_seed(ptr, len, entropy_est, 0); 10281920Smcpowers 10291920Smcpowers return (0); 10301920Smcpowers } 10311920Smcpowers 10321920Smcpowers /* 10331920Smcpowers * Mix the supplied bytes into the entropy pool of a kCF 10341920Smcpowers * RNG provider. Mix immediately. 10351920Smcpowers */ 10361920Smcpowers int 10371920Smcpowers random_add_entropy(uint8_t *ptr, size_t len, uint_t entropy_est) 10381920Smcpowers { 10391920Smcpowers if (len < 1) 10401920Smcpowers return (-1); 10411920Smcpowers 10421920Smcpowers rngprov_seed(ptr, len, entropy_est, CRYPTO_SEED_NOW); 10430Sstevel@tonic-gate 10440Sstevel@tonic-gate return (0); 10450Sstevel@tonic-gate } 10460Sstevel@tonic-gate 10470Sstevel@tonic-gate /* 10480Sstevel@tonic-gate * Get bytes from the /dev/urandom generator. This function 10490Sstevel@tonic-gate * always succeeds. Returns 0. 10500Sstevel@tonic-gate */ 10510Sstevel@tonic-gate int 10520Sstevel@tonic-gate random_get_pseudo_bytes(uint8_t *ptr, size_t len) 10530Sstevel@tonic-gate { 10540Sstevel@tonic-gate ASSERT(!mutex_owned(&rndpool_lock)); 10550Sstevel@tonic-gate 10560Sstevel@tonic-gate if (len < 1) 10570Sstevel@tonic-gate return (0); 10580Sstevel@tonic-gate return (kcf_rnd_get_pseudo_bytes(ptr, len)); 10590Sstevel@tonic-gate } 10600Sstevel@tonic-gate 10610Sstevel@tonic-gate /* 10620Sstevel@tonic-gate * Get bytes from the /dev/random generator. Returns 0 10630Sstevel@tonic-gate * on success. Returns EAGAIN if there is insufficient entropy. 10640Sstevel@tonic-gate */ 10650Sstevel@tonic-gate int 10660Sstevel@tonic-gate random_get_bytes(uint8_t *ptr, size_t len) 10670Sstevel@tonic-gate { 10680Sstevel@tonic-gate ASSERT(!mutex_owned(&rndpool_lock)); 10690Sstevel@tonic-gate 10700Sstevel@tonic-gate if (len < 1) 10710Sstevel@tonic-gate return (0); 10720Sstevel@tonic-gate return (kcf_rnd_get_bytes(ptr, len, B_TRUE, B_FALSE)); 10730Sstevel@tonic-gate } 1074