xref: /onnv-gate/usr/src/uts/common/crypto/api/kcf_random.c (revision 8928:25403608cf1e)
10Sstevel@tonic-gate /*
20Sstevel@tonic-gate  * CDDL HEADER START
30Sstevel@tonic-gate  *
40Sstevel@tonic-gate  * The contents of this file are subject to the terms of the
51070Skais  * Common Development and Distribution License (the "License").
61070Skais  * You may not use this file except in compliance with the License.
70Sstevel@tonic-gate  *
80Sstevel@tonic-gate  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
90Sstevel@tonic-gate  * or http://www.opensolaris.org/os/licensing.
100Sstevel@tonic-gate  * See the License for the specific language governing permissions
110Sstevel@tonic-gate  * and limitations under the License.
120Sstevel@tonic-gate  *
130Sstevel@tonic-gate  * When distributing Covered Code, include this CDDL HEADER in each
140Sstevel@tonic-gate  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
150Sstevel@tonic-gate  * If applicable, add the following below this CDDL HEADER, with the
160Sstevel@tonic-gate  * fields enclosed by brackets "[]" replaced with your own identifying
170Sstevel@tonic-gate  * information: Portions Copyright [yyyy] [name of copyright owner]
180Sstevel@tonic-gate  *
190Sstevel@tonic-gate  * CDDL HEADER END
200Sstevel@tonic-gate  */
210Sstevel@tonic-gate /*
228513SVladimir.Kotal@Sun.COM  * Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
230Sstevel@tonic-gate  * Use is subject to license terms.
240Sstevel@tonic-gate  */
250Sstevel@tonic-gate 
260Sstevel@tonic-gate /*
270Sstevel@tonic-gate  * This file implements the interfaces that the /dev/random
280Sstevel@tonic-gate  * driver uses for read(2), write(2) and poll(2) on /dev/random or
290Sstevel@tonic-gate  * /dev/urandom. It also implements the kernel API - random_add_entropy(),
308513SVladimir.Kotal@Sun.COM  * random_add_pseudo_entropy(), random_get_pseudo_bytes()
318513SVladimir.Kotal@Sun.COM  * and random_get_bytes().
320Sstevel@tonic-gate  *
330Sstevel@tonic-gate  * We periodically collect random bits from providers which are registered
340Sstevel@tonic-gate  * with the Kernel Cryptographic Framework (kCF) as capable of random
350Sstevel@tonic-gate  * number generation. The random bits are maintained in a cache and
360Sstevel@tonic-gate  * it is used for high quality random numbers (/dev/random) requests.
370Sstevel@tonic-gate  * We pick a provider and call its SPI routine, if the cache does not have
380Sstevel@tonic-gate  * enough bytes to satisfy a request.
390Sstevel@tonic-gate  *
400Sstevel@tonic-gate  * /dev/urandom requests use a software-based generator algorithm that uses the
410Sstevel@tonic-gate  * random bits in the cache as a seed. We create one pseudo-random generator
420Sstevel@tonic-gate  * (for /dev/urandom) per possible CPU on the system, and use it,
430Sstevel@tonic-gate  * kmem-magazine-style, to avoid cache line contention.
440Sstevel@tonic-gate  *
450Sstevel@tonic-gate  * LOCKING HIERARCHY:
468029SHai-May.Chao@Sun.COM  *	1) rmp->rm_mag.rm_lock protects the per-cpu pseudo-random generators.
470Sstevel@tonic-gate  * 	2) rndpool_lock protects the high-quality randomness pool.
488029SHai-May.Chao@Sun.COM  *		It may be locked while a rmp->rm_mag.rm_lock is held.
490Sstevel@tonic-gate  *
500Sstevel@tonic-gate  * A history note: The kernel API and the software-based algorithms in this
510Sstevel@tonic-gate  * file used to be part of the /dev/random driver.
520Sstevel@tonic-gate  */
530Sstevel@tonic-gate 
540Sstevel@tonic-gate #include <sys/types.h>
550Sstevel@tonic-gate #include <sys/conf.h>
560Sstevel@tonic-gate #include <sys/sunddi.h>
570Sstevel@tonic-gate #include <sys/disp.h>
580Sstevel@tonic-gate #include <sys/modctl.h>
590Sstevel@tonic-gate #include <sys/ddi.h>
600Sstevel@tonic-gate #include <sys/crypto/common.h>
610Sstevel@tonic-gate #include <sys/crypto/api.h>
620Sstevel@tonic-gate #include <sys/crypto/impl.h>
630Sstevel@tonic-gate #include <sys/crypto/sched_impl.h>
640Sstevel@tonic-gate #include <sys/random.h>
650Sstevel@tonic-gate #include <sys/sha1.h>
660Sstevel@tonic-gate #include <sys/time.h>
670Sstevel@tonic-gate #include <sys/sysmacros.h>
680Sstevel@tonic-gate #include <sys/cpuvar.h>
690Sstevel@tonic-gate #include <sys/taskq.h>
708029SHai-May.Chao@Sun.COM #include <rng/fips_random.h>
710Sstevel@tonic-gate 
720Sstevel@tonic-gate #define	RNDPOOLSIZE		1024	/* Pool size in bytes */
730Sstevel@tonic-gate #define	MINEXTRACTBYTES		20
740Sstevel@tonic-gate #define	MAXEXTRACTBYTES		1024
750Sstevel@tonic-gate #define	PRNG_MAXOBLOCKS		1310720	/* Max output block per prng key */
760Sstevel@tonic-gate #define	TIMEOUT_INTERVAL	5	/* Periodic mixing interval in secs */
770Sstevel@tonic-gate 
780Sstevel@tonic-gate typedef enum    extract_type {
790Sstevel@tonic-gate 	NONBLOCK_EXTRACT,
800Sstevel@tonic-gate 	BLOCKING_EXTRACT,
810Sstevel@tonic-gate 	ALWAYS_EXTRACT
820Sstevel@tonic-gate } extract_type_t;
830Sstevel@tonic-gate 
840Sstevel@tonic-gate /*
850Sstevel@tonic-gate  * Hash-algo generic definitions. For now, they are SHA1's. We use SHA1
860Sstevel@tonic-gate  * routines directly instead of using k-API because we can't return any
870Sstevel@tonic-gate  * error code in /dev/urandom case and we can get an error using k-API
880Sstevel@tonic-gate  * if a mechanism is disabled.
890Sstevel@tonic-gate  */
900Sstevel@tonic-gate #define	HASHSIZE		20
910Sstevel@tonic-gate #define	HASH_CTX		SHA1_CTX
920Sstevel@tonic-gate #define	HashInit(ctx)		SHA1Init((ctx))
930Sstevel@tonic-gate #define	HashUpdate(ctx, p, s)	SHA1Update((ctx), (p), (s))
940Sstevel@tonic-gate #define	HashFinal(d, ctx)	SHA1Final((d), (ctx))
950Sstevel@tonic-gate 
960Sstevel@tonic-gate /* HMAC-SHA1 */
970Sstevel@tonic-gate #define	HMAC_KEYSIZE			20
980Sstevel@tonic-gate 
990Sstevel@tonic-gate /*
1000Sstevel@tonic-gate  * Cache of random bytes implemented as a circular buffer. findex and rindex
1010Sstevel@tonic-gate  * track the front and back of the circular buffer.
1020Sstevel@tonic-gate  */
1030Sstevel@tonic-gate uint8_t rndpool[RNDPOOLSIZE];
1040Sstevel@tonic-gate static int findex, rindex;
1050Sstevel@tonic-gate static int rnbyte_cnt;		/* Number of bytes in the cache */
1060Sstevel@tonic-gate 
1070Sstevel@tonic-gate static kmutex_t rndpool_lock;	/* protects r/w accesses to the cache, */
1080Sstevel@tonic-gate 				/* and the global variables */
1090Sstevel@tonic-gate static kcondvar_t rndpool_read_cv; /* serializes poll/read syscalls */
1100Sstevel@tonic-gate static int num_waiters;		/* #threads waiting to read from /dev/random */
1110Sstevel@tonic-gate 
1120Sstevel@tonic-gate static struct pollhead rnd_pollhead;
1130Sstevel@tonic-gate static timeout_id_t kcf_rndtimeout_id;
1140Sstevel@tonic-gate static crypto_mech_type_t rngmech_type = CRYPTO_MECH_INVALID;
1150Sstevel@tonic-gate rnd_stats_t rnd_stats;
1163096Skrishna static boolean_t rng_prov_found = B_TRUE;
1173096Skrishna static boolean_t rng_ok_to_log = B_TRUE;
1180Sstevel@tonic-gate 
1190Sstevel@tonic-gate static void rndc_addbytes(uint8_t *, size_t);
1200Sstevel@tonic-gate static void rndc_getbytes(uint8_t *ptr, size_t len);
1210Sstevel@tonic-gate static void rnd_handler(void *);
1220Sstevel@tonic-gate static void rnd_alloc_magazines();
1230Sstevel@tonic-gate 
1240Sstevel@tonic-gate void
1250Sstevel@tonic-gate kcf_rnd_init()
1260Sstevel@tonic-gate {
1270Sstevel@tonic-gate 	hrtime_t ts;
1280Sstevel@tonic-gate 	time_t now;
1290Sstevel@tonic-gate 
1300Sstevel@tonic-gate 	mutex_init(&rndpool_lock, NULL, MUTEX_DEFAULT, NULL);
1310Sstevel@tonic-gate 	cv_init(&rndpool_read_cv, NULL, CV_DEFAULT, NULL);
1320Sstevel@tonic-gate 
1330Sstevel@tonic-gate 	/*
1340Sstevel@tonic-gate 	 * Add bytes to the cache using
1350Sstevel@tonic-gate 	 * . 2 unpredictable times: high resolution time since the boot-time,
1360Sstevel@tonic-gate 	 *   and the current time-of-the day.
1370Sstevel@tonic-gate 	 * This is used only to make the timeout value in the timer
1380Sstevel@tonic-gate 	 * unpredictable.
1390Sstevel@tonic-gate 	 */
1400Sstevel@tonic-gate 	ts = gethrtime();
1410Sstevel@tonic-gate 	rndc_addbytes((uint8_t *)&ts, sizeof (ts));
1420Sstevel@tonic-gate 
1430Sstevel@tonic-gate 	(void) drv_getparm(TIME, &now);
1440Sstevel@tonic-gate 	rndc_addbytes((uint8_t *)&now, sizeof (now));
1450Sstevel@tonic-gate 
1460Sstevel@tonic-gate 	rnbyte_cnt = 0;
1470Sstevel@tonic-gate 	findex = rindex = 0;
1480Sstevel@tonic-gate 	num_waiters = 0;
1490Sstevel@tonic-gate 	rngmech_type = KCF_MECHID(KCF_MISC_CLASS, 0);
1500Sstevel@tonic-gate 
1510Sstevel@tonic-gate 	rnd_alloc_magazines();
1520Sstevel@tonic-gate }
1530Sstevel@tonic-gate 
1540Sstevel@tonic-gate /*
1550Sstevel@tonic-gate  * Return TRUE if at least one provider exists that can
1560Sstevel@tonic-gate  * supply random numbers.
1570Sstevel@tonic-gate  */
1580Sstevel@tonic-gate boolean_t
1590Sstevel@tonic-gate kcf_rngprov_check(void)
1600Sstevel@tonic-gate {
1610Sstevel@tonic-gate 	int rv;
1620Sstevel@tonic-gate 	kcf_provider_desc_t *pd;
1630Sstevel@tonic-gate 
1640Sstevel@tonic-gate 	if ((pd = kcf_get_mech_provider(rngmech_type, NULL, &rv,
1650Sstevel@tonic-gate 	    NULL, CRYPTO_FG_RANDOM, B_FALSE, 0)) != NULL) {
1660Sstevel@tonic-gate 		KCF_PROV_REFRELE(pd);
1673096Skrishna 		/*
1683096Skrishna 		 * We logged a warning once about no provider being available
1693096Skrishna 		 * and now a provider became available. So, set the flag so
1703096Skrishna 		 * that we can log again if the problem recurs.
1713096Skrishna 		 */
1723096Skrishna 		rng_ok_to_log = B_TRUE;
1733096Skrishna 		rng_prov_found = B_TRUE;
1740Sstevel@tonic-gate 		return (B_TRUE);
1753096Skrishna 	} else {
1763096Skrishna 		rng_prov_found = B_FALSE;
1770Sstevel@tonic-gate 		return (B_FALSE);
1783096Skrishna 	}
1790Sstevel@tonic-gate }
1800Sstevel@tonic-gate 
1810Sstevel@tonic-gate /*
1820Sstevel@tonic-gate  * Pick a software-based provider and submit a request to seed
1830Sstevel@tonic-gate  * its random number generator.
1840Sstevel@tonic-gate  */
1850Sstevel@tonic-gate static void
1861920Smcpowers rngprov_seed(uint8_t *buf, int len, uint_t entropy_est, uint32_t flags)
1870Sstevel@tonic-gate {
1880Sstevel@tonic-gate 	kcf_provider_desc_t *pd = NULL;
1890Sstevel@tonic-gate 
1903708Skrishna 	if (kcf_get_sw_prov(rngmech_type, &pd, NULL, B_FALSE) ==
1913708Skrishna 	    CRYPTO_SUCCESS) {
1921920Smcpowers 		(void) KCF_PROV_SEED_RANDOM(pd, pd->pd_sid, buf, len,
1931920Smcpowers 		    entropy_est, flags, NULL);
1940Sstevel@tonic-gate 		KCF_PROV_REFRELE(pd);
1950Sstevel@tonic-gate 	}
1960Sstevel@tonic-gate }
1970Sstevel@tonic-gate 
1980Sstevel@tonic-gate /* Boot-time tunable for experimentation. */
1990Sstevel@tonic-gate int kcf_limit_hwrng = 1;
2000Sstevel@tonic-gate 
2010Sstevel@tonic-gate 
2020Sstevel@tonic-gate /*
2030Sstevel@tonic-gate  * This routine is called for blocking reads.
2040Sstevel@tonic-gate  *
2050Sstevel@tonic-gate  * The argument from_user_api indicates whether the caller is
2060Sstevel@tonic-gate  * from userland coming via the /dev/random driver.
2070Sstevel@tonic-gate  *
2080Sstevel@tonic-gate  * The argument is_taskq_thr indicates whether the caller is
2090Sstevel@tonic-gate  * the taskq thread dispatched by the timeout handler routine.
2100Sstevel@tonic-gate  * In this case, we cycle through all the providers
2110Sstevel@tonic-gate  * submitting a request to each provider to generate random numbers.
2120Sstevel@tonic-gate  *
2130Sstevel@tonic-gate  * For other cases, we pick a provider and submit a request to generate
2140Sstevel@tonic-gate  * random numbers. We retry using another provider if we get an error.
2150Sstevel@tonic-gate  *
2160Sstevel@tonic-gate  * Returns the number of bytes that are written to 'ptr'. Returns -1
2170Sstevel@tonic-gate  * if no provider is found. ptr and need are unchanged.
2180Sstevel@tonic-gate  */
2190Sstevel@tonic-gate static int
2200Sstevel@tonic-gate rngprov_getbytes(uint8_t *ptr, size_t need, boolean_t from_user_api,
2210Sstevel@tonic-gate     boolean_t is_taskq_thr)
2220Sstevel@tonic-gate {
2230Sstevel@tonic-gate 	int rv;
2240Sstevel@tonic-gate 	int prov_cnt = 0;
2250Sstevel@tonic-gate 	int total_bytes = 0;
2260Sstevel@tonic-gate 	kcf_provider_desc_t *pd;
2270Sstevel@tonic-gate 	kcf_req_params_t params;
2280Sstevel@tonic-gate 	kcf_prov_tried_t *list = NULL;
2290Sstevel@tonic-gate 
2300Sstevel@tonic-gate 	while ((pd = kcf_get_mech_provider(rngmech_type, NULL, &rv,
2310Sstevel@tonic-gate 	    list, CRYPTO_FG_RANDOM, B_FALSE, 0)) != NULL) {
2320Sstevel@tonic-gate 
2330Sstevel@tonic-gate 		prov_cnt++;
2340Sstevel@tonic-gate 		/*
2350Sstevel@tonic-gate 		 * Typically a hardware RNG is a multi-purpose
2360Sstevel@tonic-gate 		 * crypto card and hence we do not want to overload the card
2370Sstevel@tonic-gate 		 * just for random numbers. The following check is to prevent
2380Sstevel@tonic-gate 		 * a user process from hogging the hardware RNG. Note that we
2390Sstevel@tonic-gate 		 * still use the hardware RNG from the periodically run
2400Sstevel@tonic-gate 		 * taskq thread.
2410Sstevel@tonic-gate 		 */
2420Sstevel@tonic-gate 		if (pd->pd_prov_type == CRYPTO_HW_PROVIDER && from_user_api &&
2430Sstevel@tonic-gate 		    kcf_limit_hwrng == 1) {
2440Sstevel@tonic-gate 			ASSERT(is_taskq_thr == B_FALSE);
2450Sstevel@tonic-gate 			goto try_next;
2460Sstevel@tonic-gate 		}
2470Sstevel@tonic-gate 
2480Sstevel@tonic-gate 		KCF_WRAP_RANDOM_OPS_PARAMS(&params, KCF_OP_RANDOM_GENERATE,
2491920Smcpowers 		    pd->pd_sid, ptr, need, 0, 0);
2500Sstevel@tonic-gate 		rv = kcf_submit_request(pd, NULL, NULL, &params, B_FALSE);
2510Sstevel@tonic-gate 		ASSERT(rv != CRYPTO_QUEUED);
2520Sstevel@tonic-gate 
2530Sstevel@tonic-gate 		if (rv == CRYPTO_SUCCESS) {
2540Sstevel@tonic-gate 			total_bytes += need;
2550Sstevel@tonic-gate 			if (is_taskq_thr)
2560Sstevel@tonic-gate 				rndc_addbytes(ptr, need);
2570Sstevel@tonic-gate 			else {
2580Sstevel@tonic-gate 				KCF_PROV_REFRELE(pd);
2590Sstevel@tonic-gate 				break;
2600Sstevel@tonic-gate 			}
2610Sstevel@tonic-gate 		}
2620Sstevel@tonic-gate 
2630Sstevel@tonic-gate 		if (is_taskq_thr || rv != CRYPTO_SUCCESS) {
2640Sstevel@tonic-gate try_next:
2650Sstevel@tonic-gate 			/* Add pd to the linked list of providers tried. */
2660Sstevel@tonic-gate 			if (kcf_insert_triedlist(&list, pd, KM_SLEEP) == NULL) {
2670Sstevel@tonic-gate 				KCF_PROV_REFRELE(pd);
2680Sstevel@tonic-gate 				break;
2690Sstevel@tonic-gate 			}
2700Sstevel@tonic-gate 		}
2710Sstevel@tonic-gate 
2720Sstevel@tonic-gate 	}
2730Sstevel@tonic-gate 
2740Sstevel@tonic-gate 	if (list != NULL)
2750Sstevel@tonic-gate 		kcf_free_triedlist(list);
2760Sstevel@tonic-gate 
2770Sstevel@tonic-gate 	if (prov_cnt == 0) { /* no provider could be found. */
2783096Skrishna 		rng_prov_found = B_FALSE;
2790Sstevel@tonic-gate 		return (-1);
2803096Skrishna 	} else {
2813096Skrishna 		rng_prov_found = B_TRUE;
2823096Skrishna 		/* See comments in kcf_rngprov_check() */
2833096Skrishna 		rng_ok_to_log = B_TRUE;
2840Sstevel@tonic-gate 	}
2850Sstevel@tonic-gate 
2860Sstevel@tonic-gate 	return (total_bytes);
2870Sstevel@tonic-gate }
2880Sstevel@tonic-gate 
2890Sstevel@tonic-gate static void
2900Sstevel@tonic-gate notify_done(void *arg, int rv)
2910Sstevel@tonic-gate {
2920Sstevel@tonic-gate 	uchar_t *rndbuf = arg;
2930Sstevel@tonic-gate 
2940Sstevel@tonic-gate 	if (rv == CRYPTO_SUCCESS)
2950Sstevel@tonic-gate 		rndc_addbytes(rndbuf, MINEXTRACTBYTES);
2960Sstevel@tonic-gate 
2970Sstevel@tonic-gate 	bzero(rndbuf, MINEXTRACTBYTES);
2980Sstevel@tonic-gate 	kmem_free(rndbuf, MINEXTRACTBYTES);
2990Sstevel@tonic-gate }
3000Sstevel@tonic-gate 
3010Sstevel@tonic-gate /*
3020Sstevel@tonic-gate  * Cycle through all the providers submitting a request to each provider
3030Sstevel@tonic-gate  * to generate random numbers. This is called for the modes - NONBLOCK_EXTRACT
3040Sstevel@tonic-gate  * and ALWAYS_EXTRACT.
3050Sstevel@tonic-gate  *
3060Sstevel@tonic-gate  * Returns the number of bytes that are written to 'ptr'. Returns -1
3070Sstevel@tonic-gate  * if no provider is found. ptr and len are unchanged.
3080Sstevel@tonic-gate  */
3090Sstevel@tonic-gate static int
3100Sstevel@tonic-gate rngprov_getbytes_nblk(uint8_t *ptr, size_t len, boolean_t from_user_api)
3110Sstevel@tonic-gate {
3120Sstevel@tonic-gate 	int rv, blen, total_bytes;
3130Sstevel@tonic-gate 	uchar_t *rndbuf;
3140Sstevel@tonic-gate 	kcf_provider_desc_t *pd;
3150Sstevel@tonic-gate 	kcf_req_params_t params;
3160Sstevel@tonic-gate 	crypto_call_req_t req;
3170Sstevel@tonic-gate 	kcf_prov_tried_t *list = NULL;
3180Sstevel@tonic-gate 	int prov_cnt = 0;
3190Sstevel@tonic-gate 
3200Sstevel@tonic-gate 	blen = 0;
3210Sstevel@tonic-gate 	total_bytes = 0;
3220Sstevel@tonic-gate 	req.cr_flag = CRYPTO_SKIP_REQID;
3230Sstevel@tonic-gate 	req.cr_callback_func = notify_done;
3240Sstevel@tonic-gate 
3250Sstevel@tonic-gate 	while ((pd = kcf_get_mech_provider(rngmech_type, NULL, &rv,
3260Sstevel@tonic-gate 	    list, CRYPTO_FG_RANDOM, CHECK_RESTRICT(&req), 0)) != NULL) {
3270Sstevel@tonic-gate 
3280Sstevel@tonic-gate 		prov_cnt ++;
3290Sstevel@tonic-gate 		switch (pd->pd_prov_type) {
3300Sstevel@tonic-gate 		case CRYPTO_HW_PROVIDER:
3310Sstevel@tonic-gate 			/* See comments in rngprov_getbytes() */
3320Sstevel@tonic-gate 			if (from_user_api && kcf_limit_hwrng == 1)
3330Sstevel@tonic-gate 				goto try_next;
3340Sstevel@tonic-gate 
3350Sstevel@tonic-gate 			/*
3360Sstevel@tonic-gate 			 * We have to allocate a buffer here as we can not
3370Sstevel@tonic-gate 			 * assume that the input buffer will remain valid
3380Sstevel@tonic-gate 			 * when the callback comes. We use a fixed size buffer
3390Sstevel@tonic-gate 			 * to simplify the book keeping.
3400Sstevel@tonic-gate 			 */
3410Sstevel@tonic-gate 			rndbuf = kmem_alloc(MINEXTRACTBYTES, KM_NOSLEEP);
3420Sstevel@tonic-gate 			if (rndbuf == NULL) {
3430Sstevel@tonic-gate 				KCF_PROV_REFRELE(pd);
3440Sstevel@tonic-gate 				if (list != NULL)
3450Sstevel@tonic-gate 					kcf_free_triedlist(list);
3460Sstevel@tonic-gate 				return (total_bytes);
3470Sstevel@tonic-gate 			}
3480Sstevel@tonic-gate 			req.cr_callback_arg = rndbuf;
3490Sstevel@tonic-gate 			KCF_WRAP_RANDOM_OPS_PARAMS(&params,
3500Sstevel@tonic-gate 			    KCF_OP_RANDOM_GENERATE,
3511920Smcpowers 			    pd->pd_sid, rndbuf, MINEXTRACTBYTES, 0, 0);
3520Sstevel@tonic-gate 			break;
3530Sstevel@tonic-gate 
3540Sstevel@tonic-gate 		case CRYPTO_SW_PROVIDER:
3550Sstevel@tonic-gate 			/*
3560Sstevel@tonic-gate 			 * We do not need to allocate a buffer in the software
3570Sstevel@tonic-gate 			 * provider case as there is no callback involved. We
3580Sstevel@tonic-gate 			 * avoid any extra data copy by directly passing 'ptr'.
3590Sstevel@tonic-gate 			 */
3600Sstevel@tonic-gate 			KCF_WRAP_RANDOM_OPS_PARAMS(&params,
3610Sstevel@tonic-gate 			    KCF_OP_RANDOM_GENERATE,
3621920Smcpowers 			    pd->pd_sid, ptr, len, 0, 0);
3630Sstevel@tonic-gate 			break;
3640Sstevel@tonic-gate 		}
3650Sstevel@tonic-gate 
3660Sstevel@tonic-gate 		rv = kcf_submit_request(pd, NULL, &req, &params, B_FALSE);
3670Sstevel@tonic-gate 		if (rv == CRYPTO_SUCCESS) {
3680Sstevel@tonic-gate 			switch (pd->pd_prov_type) {
3690Sstevel@tonic-gate 			case CRYPTO_HW_PROVIDER:
3700Sstevel@tonic-gate 				/*
3710Sstevel@tonic-gate 				 * Since we have the input buffer handy,
3720Sstevel@tonic-gate 				 * we directly copy to it rather than
3730Sstevel@tonic-gate 				 * adding to the pool.
3740Sstevel@tonic-gate 				 */
3750Sstevel@tonic-gate 				blen = min(MINEXTRACTBYTES, len);
3760Sstevel@tonic-gate 				bcopy(rndbuf, ptr, blen);
3770Sstevel@tonic-gate 				if (len < MINEXTRACTBYTES)
3780Sstevel@tonic-gate 					rndc_addbytes(rndbuf + len,
3790Sstevel@tonic-gate 					    MINEXTRACTBYTES - len);
3800Sstevel@tonic-gate 				ptr += blen;
3810Sstevel@tonic-gate 				len -= blen;
3820Sstevel@tonic-gate 				total_bytes += blen;
3830Sstevel@tonic-gate 				break;
3840Sstevel@tonic-gate 
3850Sstevel@tonic-gate 			case CRYPTO_SW_PROVIDER:
3860Sstevel@tonic-gate 				total_bytes += len;
3870Sstevel@tonic-gate 				len = 0;
3880Sstevel@tonic-gate 				break;
3890Sstevel@tonic-gate 			}
3900Sstevel@tonic-gate 		}
3910Sstevel@tonic-gate 
3920Sstevel@tonic-gate 		/*
3930Sstevel@tonic-gate 		 * We free the buffer in the callback routine
3940Sstevel@tonic-gate 		 * for the CRYPTO_QUEUED case.
3950Sstevel@tonic-gate 		 */
3960Sstevel@tonic-gate 		if (pd->pd_prov_type == CRYPTO_HW_PROVIDER &&
3970Sstevel@tonic-gate 		    rv != CRYPTO_QUEUED) {
3980Sstevel@tonic-gate 			bzero(rndbuf, MINEXTRACTBYTES);
3990Sstevel@tonic-gate 			kmem_free(rndbuf, MINEXTRACTBYTES);
4000Sstevel@tonic-gate 		}
4010Sstevel@tonic-gate 
4020Sstevel@tonic-gate 		if (len == 0) {
4030Sstevel@tonic-gate 			KCF_PROV_REFRELE(pd);
4040Sstevel@tonic-gate 			break;
4050Sstevel@tonic-gate 		}
4060Sstevel@tonic-gate 
4070Sstevel@tonic-gate 		if (rv != CRYPTO_SUCCESS) {
4080Sstevel@tonic-gate try_next:
4090Sstevel@tonic-gate 			/* Add pd to the linked list of providers tried. */
4100Sstevel@tonic-gate 			if (kcf_insert_triedlist(&list, pd, KM_NOSLEEP) ==
4110Sstevel@tonic-gate 			    NULL) {
4120Sstevel@tonic-gate 				KCF_PROV_REFRELE(pd);
4130Sstevel@tonic-gate 				break;
4140Sstevel@tonic-gate 			}
4150Sstevel@tonic-gate 		}
4160Sstevel@tonic-gate 	}
4170Sstevel@tonic-gate 
4180Sstevel@tonic-gate 	if (list != NULL) {
4190Sstevel@tonic-gate 		kcf_free_triedlist(list);
4200Sstevel@tonic-gate 	}
4210Sstevel@tonic-gate 
4220Sstevel@tonic-gate 	if (prov_cnt == 0) { /* no provider could be found. */
4233096Skrishna 		rng_prov_found = B_FALSE;
4240Sstevel@tonic-gate 		return (-1);
4253096Skrishna 	} else {
4263096Skrishna 		rng_prov_found = B_TRUE;
4273096Skrishna 		/* See comments in kcf_rngprov_check() */
4283096Skrishna 		rng_ok_to_log = B_TRUE;
4290Sstevel@tonic-gate 	}
4300Sstevel@tonic-gate 
4310Sstevel@tonic-gate 	return (total_bytes);
4320Sstevel@tonic-gate }
4330Sstevel@tonic-gate 
4340Sstevel@tonic-gate static void
4350Sstevel@tonic-gate rngprov_task(void *arg)
4360Sstevel@tonic-gate {
4370Sstevel@tonic-gate 	int len = (int)(uintptr_t)arg;
4380Sstevel@tonic-gate 	uchar_t tbuf[MAXEXTRACTBYTES];
4390Sstevel@tonic-gate 
4400Sstevel@tonic-gate 	ASSERT(len <= MAXEXTRACTBYTES);
4413096Skrishna 	(void) rngprov_getbytes(tbuf, len, B_FALSE, B_TRUE);
4420Sstevel@tonic-gate }
4430Sstevel@tonic-gate 
4440Sstevel@tonic-gate /*
4450Sstevel@tonic-gate  * Returns "len" random or pseudo-random bytes in *ptr.
4460Sstevel@tonic-gate  * Will block if not enough random bytes are available and the
4470Sstevel@tonic-gate  * call is blocking.
4480Sstevel@tonic-gate  *
4490Sstevel@tonic-gate  * Called with rndpool_lock held (allowing caller to do optimistic locking;
4500Sstevel@tonic-gate  * releases the lock before return).
4510Sstevel@tonic-gate  */
4520Sstevel@tonic-gate static int
4530Sstevel@tonic-gate rnd_get_bytes(uint8_t *ptr, size_t len, extract_type_t how,
4540Sstevel@tonic-gate     boolean_t from_user_api)
4550Sstevel@tonic-gate {
4560Sstevel@tonic-gate 	int bytes;
4570Sstevel@tonic-gate 	size_t got;
4580Sstevel@tonic-gate 
4590Sstevel@tonic-gate 	ASSERT(mutex_owned(&rndpool_lock));
4600Sstevel@tonic-gate 	/*
4610Sstevel@tonic-gate 	 * Check if the request can be satisfied from the cache
4620Sstevel@tonic-gate 	 * of random bytes.
4630Sstevel@tonic-gate 	 */
4640Sstevel@tonic-gate 	if (len <= rnbyte_cnt) {
4650Sstevel@tonic-gate 		rndc_getbytes(ptr, len);
4660Sstevel@tonic-gate 		mutex_exit(&rndpool_lock);
4670Sstevel@tonic-gate 		return (0);
4680Sstevel@tonic-gate 	}
4690Sstevel@tonic-gate 	mutex_exit(&rndpool_lock);
4700Sstevel@tonic-gate 
4710Sstevel@tonic-gate 	switch (how) {
4720Sstevel@tonic-gate 	case BLOCKING_EXTRACT:
4730Sstevel@tonic-gate 		if ((got = rngprov_getbytes(ptr, len, from_user_api,
4740Sstevel@tonic-gate 		    B_FALSE)) == -1)
4750Sstevel@tonic-gate 			break;	/* No provider found */
4760Sstevel@tonic-gate 
4770Sstevel@tonic-gate 		if (got == len)
4780Sstevel@tonic-gate 			return (0);
4790Sstevel@tonic-gate 		len -= got;
4800Sstevel@tonic-gate 		ptr += got;
4810Sstevel@tonic-gate 		break;
4820Sstevel@tonic-gate 
4830Sstevel@tonic-gate 	case NONBLOCK_EXTRACT:
4840Sstevel@tonic-gate 	case ALWAYS_EXTRACT:
4850Sstevel@tonic-gate 		if ((got = rngprov_getbytes_nblk(ptr, len,
4860Sstevel@tonic-gate 		    from_user_api)) == -1) {
4870Sstevel@tonic-gate 			/* No provider found */
4880Sstevel@tonic-gate 			if (how == NONBLOCK_EXTRACT) {
4890Sstevel@tonic-gate 				return (EAGAIN);
4900Sstevel@tonic-gate 			}
4910Sstevel@tonic-gate 		} else {
4920Sstevel@tonic-gate 			if (got == len)
4930Sstevel@tonic-gate 				return (0);
4940Sstevel@tonic-gate 			len -= got;
4950Sstevel@tonic-gate 			ptr += got;
4960Sstevel@tonic-gate 		}
4970Sstevel@tonic-gate 		if (how == NONBLOCK_EXTRACT && (rnbyte_cnt < len))
4980Sstevel@tonic-gate 			return (EAGAIN);
4990Sstevel@tonic-gate 		break;
5000Sstevel@tonic-gate 	}
5010Sstevel@tonic-gate 
5020Sstevel@tonic-gate 	mutex_enter(&rndpool_lock);
5030Sstevel@tonic-gate 	while (len > 0) {
5040Sstevel@tonic-gate 		if (how == BLOCKING_EXTRACT) {
5050Sstevel@tonic-gate 			/* Check if there is enough */
5060Sstevel@tonic-gate 			while (rnbyte_cnt < MINEXTRACTBYTES) {
5070Sstevel@tonic-gate 				num_waiters++;
5080Sstevel@tonic-gate 				if (cv_wait_sig(&rndpool_read_cv,
5090Sstevel@tonic-gate 				    &rndpool_lock) == 0) {
5100Sstevel@tonic-gate 					num_waiters--;
5110Sstevel@tonic-gate 					mutex_exit(&rndpool_lock);
5120Sstevel@tonic-gate 					return (EINTR);
5130Sstevel@tonic-gate 				}
5140Sstevel@tonic-gate 				num_waiters--;
5150Sstevel@tonic-gate 			}
5160Sstevel@tonic-gate 		}
5170Sstevel@tonic-gate 
5180Sstevel@tonic-gate 		/* Figure out how many bytes to extract */
5190Sstevel@tonic-gate 		bytes = min(len, rnbyte_cnt);
5200Sstevel@tonic-gate 		rndc_getbytes(ptr, bytes);
5210Sstevel@tonic-gate 
5220Sstevel@tonic-gate 		len -= bytes;
5230Sstevel@tonic-gate 		ptr += bytes;
5240Sstevel@tonic-gate 
5250Sstevel@tonic-gate 		if (len > 0 && how == ALWAYS_EXTRACT) {
5260Sstevel@tonic-gate 			/*
5270Sstevel@tonic-gate 			 * There are not enough bytes, but we can not block.
5280Sstevel@tonic-gate 			 * This only happens in the case of /dev/urandom which
5290Sstevel@tonic-gate 			 * runs an additional generation algorithm. So, there
5300Sstevel@tonic-gate 			 * is no problem.
5310Sstevel@tonic-gate 			 */
5320Sstevel@tonic-gate 			while (len > 0) {
5330Sstevel@tonic-gate 				*ptr = rndpool[findex];
5340Sstevel@tonic-gate 				ptr++; len--;
5350Sstevel@tonic-gate 				rindex = findex = (findex + 1) &
5360Sstevel@tonic-gate 				    (RNDPOOLSIZE - 1);
5370Sstevel@tonic-gate 			}
5380Sstevel@tonic-gate 			break;
5390Sstevel@tonic-gate 		}
5400Sstevel@tonic-gate 	}
5410Sstevel@tonic-gate 
5420Sstevel@tonic-gate 	mutex_exit(&rndpool_lock);
5430Sstevel@tonic-gate 	return (0);
5440Sstevel@tonic-gate }
5450Sstevel@tonic-gate 
5460Sstevel@tonic-gate int
5470Sstevel@tonic-gate kcf_rnd_get_bytes(uint8_t *ptr, size_t len, boolean_t noblock,
5480Sstevel@tonic-gate     boolean_t from_user_api)
5490Sstevel@tonic-gate {
5500Sstevel@tonic-gate 	extract_type_t how;
5510Sstevel@tonic-gate 	int error;
5520Sstevel@tonic-gate 
5530Sstevel@tonic-gate 	how = noblock ? NONBLOCK_EXTRACT : BLOCKING_EXTRACT;
5540Sstevel@tonic-gate 	mutex_enter(&rndpool_lock);
5550Sstevel@tonic-gate 	if ((error = rnd_get_bytes(ptr, len, how, from_user_api)) != 0)
5560Sstevel@tonic-gate 		return (error);
5570Sstevel@tonic-gate 
5580Sstevel@tonic-gate 	BUMP_RND_STATS(rs_rndOut, len);
5590Sstevel@tonic-gate 	return (0);
5600Sstevel@tonic-gate }
5610Sstevel@tonic-gate 
5620Sstevel@tonic-gate /*
5630Sstevel@tonic-gate  * Revisit this if the structs grow or we come up with a better way
5640Sstevel@tonic-gate  * of cache-line-padding structures.
5650Sstevel@tonic-gate  */
5660Sstevel@tonic-gate #define	RND_CPU_CACHE_SIZE	64
5678029SHai-May.Chao@Sun.COM #define	RND_CPU_PAD_SIZE	RND_CPU_CACHE_SIZE*6
5680Sstevel@tonic-gate #define	RND_CPU_PAD (RND_CPU_PAD_SIZE - \
5698029SHai-May.Chao@Sun.COM 	sizeof (rndmag_t))
5700Sstevel@tonic-gate /*
5710Sstevel@tonic-gate  * Per-CPU random state.  Somewhat like like kmem's magazines, this provides
5720Sstevel@tonic-gate  * a per-CPU instance of the pseudo-random generator.  We have it much easier
5730Sstevel@tonic-gate  * than kmem, as we can afford to "leak" random bits if a CPU is DR'ed out.
5740Sstevel@tonic-gate  *
5750Sstevel@tonic-gate  * Note that this usage is preemption-safe; a thread
5760Sstevel@tonic-gate  * entering a critical section remembers which generator it locked
5770Sstevel@tonic-gate  * and unlocks the same one; should it be preempted and wind up running on
5780Sstevel@tonic-gate  * a different CPU, there will be a brief period of increased contention
5790Sstevel@tonic-gate  * before it exits the critical section but nothing will melt.
5800Sstevel@tonic-gate  */
5810Sstevel@tonic-gate typedef struct rndmag_s
5820Sstevel@tonic-gate {
5830Sstevel@tonic-gate 	kmutex_t	rm_lock;
5848733SHai-May.Chao@Sun.COM 	uint8_t		*rm_buffer;	/* Start of buffer */
5850Sstevel@tonic-gate 	uint8_t		*rm_eptr;	/* End of buffer */
5860Sstevel@tonic-gate 	uint8_t		*rm_rptr;	/* Current read pointer */
5870Sstevel@tonic-gate 	uint32_t	rm_oblocks;	/* time to rekey? */
5880Sstevel@tonic-gate 	uint32_t	rm_ofuzz;	/* Rekey backoff state */
5890Sstevel@tonic-gate 	uint32_t	rm_olimit;	/* Hard rekey limit */
5900Sstevel@tonic-gate 	rnd_stats_t	rm_stats;	/* Per-CPU Statistics */
5918733SHai-May.Chao@Sun.COM 	uint32_t	rm_key[HASHSIZE/BYTES_IN_WORD];	/* FIPS XKEY */
5928733SHai-May.Chao@Sun.COM 	uint32_t	rm_seed[HASHSIZE/BYTES_IN_WORD]; /* seed for rekey */
5938733SHai-May.Chao@Sun.COM 	uint32_t	rm_previous[HASHSIZE/BYTES_IN_WORD]; /* prev random */
5940Sstevel@tonic-gate } rndmag_t;
5950Sstevel@tonic-gate 
5968029SHai-May.Chao@Sun.COM typedef struct rndmag_pad_s
5978029SHai-May.Chao@Sun.COM {
5988029SHai-May.Chao@Sun.COM 	rndmag_t	rm_mag;
5998029SHai-May.Chao@Sun.COM 	uint8_t		rm_pad[RND_CPU_PAD];
6008029SHai-May.Chao@Sun.COM } rndmag_pad_t;
6018029SHai-May.Chao@Sun.COM 
6020Sstevel@tonic-gate /*
6038029SHai-May.Chao@Sun.COM  * Generate random bytes for /dev/urandom by applying the
6048029SHai-May.Chao@Sun.COM  * FIPS 186-2 algorithm with a key created from bytes extracted
6050Sstevel@tonic-gate  * from the pool.  A maximum of PRNG_MAXOBLOCKS output blocks
6060Sstevel@tonic-gate  * is generated before a new key is obtained.
6070Sstevel@tonic-gate  *
6080Sstevel@tonic-gate  * Note that callers to this routine are likely to assume it can't fail.
6090Sstevel@tonic-gate  *
6100Sstevel@tonic-gate  * Called with rmp locked; releases lock.
6110Sstevel@tonic-gate  */
6120Sstevel@tonic-gate static int
6138029SHai-May.Chao@Sun.COM rnd_generate_pseudo_bytes(rndmag_pad_t *rmp, uint8_t *ptr, size_t len)
6140Sstevel@tonic-gate {
6150Sstevel@tonic-gate 	size_t bytes = len;
6160Sstevel@tonic-gate 	int nblock, size;
6170Sstevel@tonic-gate 	uint32_t oblocks;
6188733SHai-May.Chao@Sun.COM 	uint32_t tempout[HASHSIZE/BYTES_IN_WORD];
6198733SHai-May.Chao@Sun.COM 	uint32_t seed[HASHSIZE/BYTES_IN_WORD];
6208029SHai-May.Chao@Sun.COM 	int i;
6218029SHai-May.Chao@Sun.COM 	hrtime_t timestamp;
6228029SHai-May.Chao@Sun.COM 	uint8_t *src, *dst;
6230Sstevel@tonic-gate 
6248029SHai-May.Chao@Sun.COM 	ASSERT(mutex_owned(&rmp->rm_mag.rm_lock));
6250Sstevel@tonic-gate 
6260Sstevel@tonic-gate 	/* Nothing is being asked */
6270Sstevel@tonic-gate 	if (len == 0) {
6288029SHai-May.Chao@Sun.COM 		mutex_exit(&rmp->rm_mag.rm_lock);
6290Sstevel@tonic-gate 		return (0);
6300Sstevel@tonic-gate 	}
6310Sstevel@tonic-gate 
6320Sstevel@tonic-gate 	nblock = howmany(len, HASHSIZE);
6330Sstevel@tonic-gate 
6348029SHai-May.Chao@Sun.COM 	rmp->rm_mag.rm_oblocks += nblock;
6358029SHai-May.Chao@Sun.COM 	oblocks = rmp->rm_mag.rm_oblocks;
6360Sstevel@tonic-gate 
6370Sstevel@tonic-gate 	do {
6388029SHai-May.Chao@Sun.COM 		if (oblocks >= rmp->rm_mag.rm_olimit) {
6390Sstevel@tonic-gate 
6400Sstevel@tonic-gate 			/*
6410Sstevel@tonic-gate 			 * Contention-avoiding rekey: see if
6420Sstevel@tonic-gate 			 * the pool is locked, and if so, wait a bit.
6430Sstevel@tonic-gate 			 * Do an 'exponential back-in' to ensure we don't
6440Sstevel@tonic-gate 			 * run too long without rekey.
6450Sstevel@tonic-gate 			 */
6468029SHai-May.Chao@Sun.COM 			if (rmp->rm_mag.rm_ofuzz) {
6470Sstevel@tonic-gate 				/*
6480Sstevel@tonic-gate 				 * Decaying exponential back-in for rekey.
6490Sstevel@tonic-gate 				 */
6500Sstevel@tonic-gate 				if ((rnbyte_cnt < MINEXTRACTBYTES) ||
6510Sstevel@tonic-gate 				    (!mutex_tryenter(&rndpool_lock))) {
6528029SHai-May.Chao@Sun.COM 					rmp->rm_mag.rm_olimit +=
6538029SHai-May.Chao@Sun.COM 					    rmp->rm_mag.rm_ofuzz;
6548029SHai-May.Chao@Sun.COM 					rmp->rm_mag.rm_ofuzz >>= 1;
6550Sstevel@tonic-gate 					goto punt;
6560Sstevel@tonic-gate 				}
6570Sstevel@tonic-gate 			} else {
6580Sstevel@tonic-gate 				mutex_enter(&rndpool_lock);
6590Sstevel@tonic-gate 			}
6600Sstevel@tonic-gate 
6610Sstevel@tonic-gate 			/* Get a new chunk of entropy */
6628029SHai-May.Chao@Sun.COM 			(void) rnd_get_bytes((uint8_t *)rmp->rm_mag.rm_key,
6638029SHai-May.Chao@Sun.COM 			    HMAC_KEYSIZE, ALWAYS_EXTRACT, B_FALSE);
6640Sstevel@tonic-gate 
6658029SHai-May.Chao@Sun.COM 			rmp->rm_mag.rm_olimit = PRNG_MAXOBLOCKS/2;
6668029SHai-May.Chao@Sun.COM 			rmp->rm_mag.rm_ofuzz = PRNG_MAXOBLOCKS/4;
6670Sstevel@tonic-gate 			oblocks = 0;
6688029SHai-May.Chao@Sun.COM 			rmp->rm_mag.rm_oblocks = nblock;
6690Sstevel@tonic-gate 		}
6700Sstevel@tonic-gate punt:
6718029SHai-May.Chao@Sun.COM 		timestamp = gethrtime();
6728029SHai-May.Chao@Sun.COM 
6738029SHai-May.Chao@Sun.COM 		src = (uint8_t *)&timestamp;
6748029SHai-May.Chao@Sun.COM 		dst = (uint8_t *)rmp->rm_mag.rm_seed;
6758029SHai-May.Chao@Sun.COM 
6768029SHai-May.Chao@Sun.COM 		for (i = 0; i < HASHSIZE; i++) {
6778029SHai-May.Chao@Sun.COM 			dst[i] ^= src[i % sizeof (timestamp)];
6788029SHai-May.Chao@Sun.COM 		}
6798029SHai-May.Chao@Sun.COM 
6808029SHai-May.Chao@Sun.COM 		bcopy(rmp->rm_mag.rm_seed, seed, HASHSIZE);
6818029SHai-May.Chao@Sun.COM 
6828029SHai-May.Chao@Sun.COM 		fips_random_inner(rmp->rm_mag.rm_key, tempout,
6838029SHai-May.Chao@Sun.COM 		    seed);
6848029SHai-May.Chao@Sun.COM 
6850Sstevel@tonic-gate 		if (bytes >= HASHSIZE) {
6860Sstevel@tonic-gate 			size = HASHSIZE;
6870Sstevel@tonic-gate 		} else {
6880Sstevel@tonic-gate 			size = min(bytes, HASHSIZE);
6890Sstevel@tonic-gate 		}
6908029SHai-May.Chao@Sun.COM 
6918029SHai-May.Chao@Sun.COM 		/*
6928029SHai-May.Chao@Sun.COM 		 * FIPS 140-2: Continuous RNG test - each generation
6938029SHai-May.Chao@Sun.COM 		 * of an n-bit block shall be compared with the previously
6948029SHai-May.Chao@Sun.COM 		 * generated block. Test shall fail if any two compared
6958029SHai-May.Chao@Sun.COM 		 * n-bit blocks are equal.
6968029SHai-May.Chao@Sun.COM 		 */
6978733SHai-May.Chao@Sun.COM 		for (i = 0; i < HASHSIZE/BYTES_IN_WORD; i++) {
6988029SHai-May.Chao@Sun.COM 			if (tempout[i] != rmp->rm_mag.rm_previous[i])
6998029SHai-May.Chao@Sun.COM 				break;
7008029SHai-May.Chao@Sun.COM 		}
7018733SHai-May.Chao@Sun.COM 		if (i == HASHSIZE/BYTES_IN_WORD)
7028029SHai-May.Chao@Sun.COM 			cmn_err(CE_WARN, "kcf_random: The value of 160-bit "
7038029SHai-May.Chao@Sun.COM 			    "block random bytes are same as the previous "
7048029SHai-May.Chao@Sun.COM 			    "one.\n");
7058029SHai-May.Chao@Sun.COM 
7068029SHai-May.Chao@Sun.COM 		bcopy(tempout, rmp->rm_mag.rm_previous,
7078029SHai-May.Chao@Sun.COM 		    HASHSIZE);
7088029SHai-May.Chao@Sun.COM 
7098029SHai-May.Chao@Sun.COM 		bcopy(tempout, ptr, size);
7100Sstevel@tonic-gate 		ptr += size;
7110Sstevel@tonic-gate 		bytes -= size;
7120Sstevel@tonic-gate 		oblocks++;
7130Sstevel@tonic-gate 		nblock--;
7140Sstevel@tonic-gate 	} while (bytes > 0);
7150Sstevel@tonic-gate 
7168029SHai-May.Chao@Sun.COM 	/* Zero out sensitive information */
7178029SHai-May.Chao@Sun.COM 	bzero(seed, HASHSIZE);
7188029SHai-May.Chao@Sun.COM 	bzero(tempout, HASHSIZE);
7198029SHai-May.Chao@Sun.COM 	mutex_exit(&rmp->rm_mag.rm_lock);
7200Sstevel@tonic-gate 	return (0);
7210Sstevel@tonic-gate }
7220Sstevel@tonic-gate 
7230Sstevel@tonic-gate /*
7240Sstevel@tonic-gate  * Per-CPU Random magazines.
7250Sstevel@tonic-gate  */
7268029SHai-May.Chao@Sun.COM static rndmag_pad_t *rndmag;
7270Sstevel@tonic-gate static uint8_t	*rndbuf;
7280Sstevel@tonic-gate static size_t 	rndmag_total;
7290Sstevel@tonic-gate /*
7300Sstevel@tonic-gate  * common/os/cpu.c says that platform support code can shrinkwrap
7310Sstevel@tonic-gate  * max_ncpus.  On the off chance that we get loaded very early, we
7320Sstevel@tonic-gate  * read it exactly once, to copy it here.
7330Sstevel@tonic-gate  */
7340Sstevel@tonic-gate static uint32_t	random_max_ncpus = 0;
7350Sstevel@tonic-gate 
7360Sstevel@tonic-gate /*
7370Sstevel@tonic-gate  * Boot-time tunables, for experimentation.
7380Sstevel@tonic-gate  */
7391070Skais size_t	rndmag_threshold = 2560;
7401070Skais size_t	rndbuf_len = 5120;
741445Skrishna size_t	rndmag_size = 1280;
7420Sstevel@tonic-gate 
7430Sstevel@tonic-gate 
7440Sstevel@tonic-gate int
7450Sstevel@tonic-gate kcf_rnd_get_pseudo_bytes(uint8_t *ptr, size_t len)
7460Sstevel@tonic-gate {
7478029SHai-May.Chao@Sun.COM 	rndmag_pad_t *rmp;
7480Sstevel@tonic-gate 	uint8_t *cptr, *eptr;
7490Sstevel@tonic-gate 
7500Sstevel@tonic-gate 	/*
7510Sstevel@tonic-gate 	 * Anyone who asks for zero bytes of randomness should get slapped.
7520Sstevel@tonic-gate 	 */
7530Sstevel@tonic-gate 	ASSERT(len > 0);
7540Sstevel@tonic-gate 
7550Sstevel@tonic-gate 	/*
7560Sstevel@tonic-gate 	 * Fast path.
7570Sstevel@tonic-gate 	 */
7580Sstevel@tonic-gate 	for (;;) {
7590Sstevel@tonic-gate 		rmp = &rndmag[CPU->cpu_seqid];
7608029SHai-May.Chao@Sun.COM 		mutex_enter(&rmp->rm_mag.rm_lock);
7610Sstevel@tonic-gate 
7620Sstevel@tonic-gate 		/*
7630Sstevel@tonic-gate 		 * Big requests bypass buffer and tail-call the
7640Sstevel@tonic-gate 		 * generate routine directly.
7650Sstevel@tonic-gate 		 */
7660Sstevel@tonic-gate 		if (len > rndmag_threshold) {
7670Sstevel@tonic-gate 			BUMP_CPU_RND_STATS(rmp, rs_urndOut, len);
7680Sstevel@tonic-gate 			return (rnd_generate_pseudo_bytes(rmp, ptr, len));
7690Sstevel@tonic-gate 		}
7700Sstevel@tonic-gate 
7718029SHai-May.Chao@Sun.COM 		cptr = rmp->rm_mag.rm_rptr;
7720Sstevel@tonic-gate 		eptr = cptr + len;
7730Sstevel@tonic-gate 
7748029SHai-May.Chao@Sun.COM 		if (eptr <= rmp->rm_mag.rm_eptr) {
7758029SHai-May.Chao@Sun.COM 			rmp->rm_mag.rm_rptr = eptr;
7760Sstevel@tonic-gate 			bcopy(cptr, ptr, len);
7770Sstevel@tonic-gate 			BUMP_CPU_RND_STATS(rmp, rs_urndOut, len);
7788029SHai-May.Chao@Sun.COM 			mutex_exit(&rmp->rm_mag.rm_lock);
7790Sstevel@tonic-gate 
7800Sstevel@tonic-gate 			return (0);
7810Sstevel@tonic-gate 		}
7820Sstevel@tonic-gate 		/*
7830Sstevel@tonic-gate 		 * End fast path.
7840Sstevel@tonic-gate 		 */
7858029SHai-May.Chao@Sun.COM 		rmp->rm_mag.rm_rptr = rmp->rm_mag.rm_buffer;
7860Sstevel@tonic-gate 		/*
7870Sstevel@tonic-gate 		 * Note:  We assume the generate routine always succeeds
7880Sstevel@tonic-gate 		 * in this case (because it does at present..)
7890Sstevel@tonic-gate 		 * It also always releases rm_lock.
7900Sstevel@tonic-gate 		 */
7918029SHai-May.Chao@Sun.COM 		(void) rnd_generate_pseudo_bytes(rmp, rmp->rm_mag.rm_buffer,
7920Sstevel@tonic-gate 		    rndbuf_len);
7930Sstevel@tonic-gate 	}
7940Sstevel@tonic-gate }
7950Sstevel@tonic-gate 
7960Sstevel@tonic-gate /*
7970Sstevel@tonic-gate  * We set up (empty) magazines for all of max_ncpus, possibly wasting a
7980Sstevel@tonic-gate  * little memory on big systems that don't have the full set installed.
7990Sstevel@tonic-gate  * See above;  "empty" means "rptr equal to eptr"; this will trigger the
8000Sstevel@tonic-gate  * refill path in rnd_get_pseudo_bytes above on the first call for each CPU.
8010Sstevel@tonic-gate  *
8020Sstevel@tonic-gate  * TODO: make rndmag_size tunable at run time!
8030Sstevel@tonic-gate  */
8040Sstevel@tonic-gate static void
8050Sstevel@tonic-gate rnd_alloc_magazines()
8060Sstevel@tonic-gate {
8078029SHai-May.Chao@Sun.COM 	rndmag_pad_t *rmp;
8080Sstevel@tonic-gate 	int i;
8098029SHai-May.Chao@Sun.COM 	uint8_t discard_buf[HASHSIZE];
8100Sstevel@tonic-gate 
8110Sstevel@tonic-gate 	rndbuf_len = roundup(rndbuf_len, HASHSIZE);
8120Sstevel@tonic-gate 	if (rndmag_size < rndbuf_len)
8130Sstevel@tonic-gate 		rndmag_size = rndbuf_len;
8140Sstevel@tonic-gate 	rndmag_size = roundup(rndmag_size, RND_CPU_CACHE_SIZE);
8150Sstevel@tonic-gate 
8160Sstevel@tonic-gate 	random_max_ncpus = max_ncpus;
8170Sstevel@tonic-gate 	rndmag_total = rndmag_size * random_max_ncpus;
8180Sstevel@tonic-gate 
8190Sstevel@tonic-gate 	rndbuf = kmem_alloc(rndmag_total, KM_SLEEP);
8208029SHai-May.Chao@Sun.COM 	rndmag = kmem_zalloc(sizeof (rndmag_pad_t) * random_max_ncpus,
8218029SHai-May.Chao@Sun.COM 	    KM_SLEEP);
8220Sstevel@tonic-gate 
8230Sstevel@tonic-gate 	for (i = 0; i < random_max_ncpus; i++) {
8240Sstevel@tonic-gate 		uint8_t *buf;
8250Sstevel@tonic-gate 
8260Sstevel@tonic-gate 		rmp = &rndmag[i];
8278029SHai-May.Chao@Sun.COM 		mutex_init(&rmp->rm_mag.rm_lock, NULL, MUTEX_DRIVER, NULL);
8280Sstevel@tonic-gate 
8290Sstevel@tonic-gate 		buf = rndbuf + i * rndmag_size;
8300Sstevel@tonic-gate 
8318029SHai-May.Chao@Sun.COM 		rmp->rm_mag.rm_buffer = buf;
8328029SHai-May.Chao@Sun.COM 		rmp->rm_mag.rm_eptr = buf + rndbuf_len;
8338029SHai-May.Chao@Sun.COM 		rmp->rm_mag.rm_rptr = buf + rndbuf_len;
8348029SHai-May.Chao@Sun.COM 		rmp->rm_mag.rm_oblocks = 1;
8358029SHai-May.Chao@Sun.COM 
8368029SHai-May.Chao@Sun.COM 		mutex_enter(&rndpool_lock);
8378029SHai-May.Chao@Sun.COM 		/*
8388029SHai-May.Chao@Sun.COM 		 * FIPS 140-2: the first n-bit (n > 15) block generated
8398029SHai-May.Chao@Sun.COM 		 * after power-up, initialization, or reset shall not
8408029SHai-May.Chao@Sun.COM 		 * be used, but shall be saved for comparison.
8418029SHai-May.Chao@Sun.COM 		 */
8428029SHai-May.Chao@Sun.COM 		(void) rnd_get_bytes(discard_buf,
8438029SHai-May.Chao@Sun.COM 		    HMAC_KEYSIZE, ALWAYS_EXTRACT, B_FALSE);
8448029SHai-May.Chao@Sun.COM 		bcopy(discard_buf, rmp->rm_mag.rm_previous,
8458029SHai-May.Chao@Sun.COM 		    HMAC_KEYSIZE);
8468029SHai-May.Chao@Sun.COM 		/* rnd_get_bytes() will call mutex_exit(&rndpool_lock) */
8478029SHai-May.Chao@Sun.COM 		mutex_enter(&rndpool_lock);
8488029SHai-May.Chao@Sun.COM 		(void) rnd_get_bytes((uint8_t *)rmp->rm_mag.rm_key,
8498029SHai-May.Chao@Sun.COM 		    HMAC_KEYSIZE, ALWAYS_EXTRACT, B_FALSE);
8508029SHai-May.Chao@Sun.COM 		/* rnd_get_bytes() will call mutex_exit(&rndpool_lock) */
8518029SHai-May.Chao@Sun.COM 		mutex_enter(&rndpool_lock);
8528029SHai-May.Chao@Sun.COM 		(void) rnd_get_bytes((uint8_t *)rmp->rm_mag.rm_seed,
8538029SHai-May.Chao@Sun.COM 		    HMAC_KEYSIZE, ALWAYS_EXTRACT, B_FALSE);
8540Sstevel@tonic-gate 	}
8550Sstevel@tonic-gate }
8560Sstevel@tonic-gate 
8570Sstevel@tonic-gate void
8580Sstevel@tonic-gate kcf_rnd_schedule_timeout(boolean_t do_mech2id)
8590Sstevel@tonic-gate {
8600Sstevel@tonic-gate 	clock_t ut;	/* time in microseconds */
8610Sstevel@tonic-gate 
8620Sstevel@tonic-gate 	if (do_mech2id)
8630Sstevel@tonic-gate 		rngmech_type = crypto_mech2id(SUN_RANDOM);
8640Sstevel@tonic-gate 
8650Sstevel@tonic-gate 	/*
8660Sstevel@tonic-gate 	 * The new timeout value is taken from the buffer of random bytes.
8670Sstevel@tonic-gate 	 * We're merely reading the first 32 bits from the buffer here, not
8680Sstevel@tonic-gate 	 * consuming any random bytes.
8690Sstevel@tonic-gate 	 * The timeout multiplier value is a random value between 0.5 sec and
8700Sstevel@tonic-gate 	 * 1.544480 sec (0.5 sec + 0xFF000 microseconds).
8710Sstevel@tonic-gate 	 * The new timeout is TIMEOUT_INTERVAL times that multiplier.
8720Sstevel@tonic-gate 	 */
8730Sstevel@tonic-gate 	ut = 500000 + (clock_t)((((uint32_t)rndpool[findex]) << 12) & 0xFF000);
8740Sstevel@tonic-gate 	kcf_rndtimeout_id = timeout(rnd_handler, NULL,
8750Sstevel@tonic-gate 	    TIMEOUT_INTERVAL * drv_usectohz(ut));
8760Sstevel@tonic-gate }
8770Sstevel@tonic-gate 
8780Sstevel@tonic-gate /*
879*8928SBhargava.Yenduri@Sun.COM  * Called from the driver for a poll on /dev/random
880*8928SBhargava.Yenduri@Sun.COM  * . POLLOUT always succeeds.
881*8928SBhargava.Yenduri@Sun.COM  * . POLLIN and POLLRDNORM will block until a
882*8928SBhargava.Yenduri@Sun.COM  *   minimum amount of entropy is available.
883*8928SBhargava.Yenduri@Sun.COM  *
8840Sstevel@tonic-gate  * &rnd_pollhead is passed in *phpp in order to indicate the calling thread
8850Sstevel@tonic-gate  * will block. When enough random bytes are available, later, the timeout
8860Sstevel@tonic-gate  * handler routine will issue the pollwakeup() calls.
8870Sstevel@tonic-gate  */
8880Sstevel@tonic-gate void
889*8928SBhargava.Yenduri@Sun.COM kcf_rnd_chpoll(short events, int anyyet, short *reventsp,
890*8928SBhargava.Yenduri@Sun.COM     struct pollhead **phpp)
8910Sstevel@tonic-gate {
892*8928SBhargava.Yenduri@Sun.COM 	*reventsp = events & POLLOUT;
893*8928SBhargava.Yenduri@Sun.COM 
894*8928SBhargava.Yenduri@Sun.COM 	if (events & (POLLIN | POLLRDNORM)) {
895*8928SBhargava.Yenduri@Sun.COM 		/*
896*8928SBhargava.Yenduri@Sun.COM 		 * Sampling of rnbyte_cnt is an atomic
897*8928SBhargava.Yenduri@Sun.COM 		 * operation. Hence we do not need any locking.
898*8928SBhargava.Yenduri@Sun.COM 		 */
899*8928SBhargava.Yenduri@Sun.COM 		if (rnbyte_cnt >= MINEXTRACTBYTES)
900*8928SBhargava.Yenduri@Sun.COM 			*reventsp |= (events & (POLLIN | POLLRDNORM));
9010Sstevel@tonic-gate 	}
902*8928SBhargava.Yenduri@Sun.COM 
903*8928SBhargava.Yenduri@Sun.COM 	if (*reventsp == 0 && !anyyet)
904*8928SBhargava.Yenduri@Sun.COM 		*phpp = &rnd_pollhead;
9050Sstevel@tonic-gate }
9060Sstevel@tonic-gate 
9070Sstevel@tonic-gate /*ARGSUSED*/
9080Sstevel@tonic-gate static void
9090Sstevel@tonic-gate rnd_handler(void *arg)
9100Sstevel@tonic-gate {
9110Sstevel@tonic-gate 	int len = 0;
9120Sstevel@tonic-gate 
9133096Skrishna 	if (!rng_prov_found && rng_ok_to_log) {
9143096Skrishna 		cmn_err(CE_WARN, "No randomness provider enabled for "
9153096Skrishna 		    "/dev/random. Use cryptoadm(1M) to enable a provider.");
9163096Skrishna 		rng_ok_to_log = B_FALSE;
9173096Skrishna 	}
9183096Skrishna 
9190Sstevel@tonic-gate 	if (num_waiters > 0)
9200Sstevel@tonic-gate 		len = MAXEXTRACTBYTES;
9210Sstevel@tonic-gate 	else if (rnbyte_cnt < RNDPOOLSIZE)
9220Sstevel@tonic-gate 		len = MINEXTRACTBYTES;
9230Sstevel@tonic-gate 
9240Sstevel@tonic-gate 	if (len > 0) {
9250Sstevel@tonic-gate 		(void) taskq_dispatch(system_taskq, rngprov_task,
9260Sstevel@tonic-gate 		    (void *)(uintptr_t)len, TQ_NOSLEEP);
9270Sstevel@tonic-gate 	}
9280Sstevel@tonic-gate 
9290Sstevel@tonic-gate 	mutex_enter(&rndpool_lock);
9300Sstevel@tonic-gate 	/*
9310Sstevel@tonic-gate 	 * Wake up threads waiting in poll() or for enough accumulated
9320Sstevel@tonic-gate 	 * random bytes to read from /dev/random. In case a poll() is
9330Sstevel@tonic-gate 	 * concurrent with a read(), the polling process may be woken up
9340Sstevel@tonic-gate 	 * indicating that enough randomness is now available for reading,
9350Sstevel@tonic-gate 	 * and another process *steals* the bits from the pool, causing the
9360Sstevel@tonic-gate 	 * subsequent read() from the first process to block. It is acceptable
9370Sstevel@tonic-gate 	 * since the blocking will eventually end, after the timeout
9380Sstevel@tonic-gate 	 * has expired enough times to honor the read.
9390Sstevel@tonic-gate 	 *
9400Sstevel@tonic-gate 	 * Note - Since we hold the rndpool_lock across the pollwakeup() call
9410Sstevel@tonic-gate 	 * we MUST NOT grab the rndpool_lock in kcf_rndchpoll().
9420Sstevel@tonic-gate 	 */
9430Sstevel@tonic-gate 	if (rnbyte_cnt >= MINEXTRACTBYTES)
9440Sstevel@tonic-gate 		pollwakeup(&rnd_pollhead, POLLIN | POLLRDNORM);
9450Sstevel@tonic-gate 
9460Sstevel@tonic-gate 	if (num_waiters > 0)
9470Sstevel@tonic-gate 		cv_broadcast(&rndpool_read_cv);
9480Sstevel@tonic-gate 	mutex_exit(&rndpool_lock);
9490Sstevel@tonic-gate 
9500Sstevel@tonic-gate 	kcf_rnd_schedule_timeout(B_FALSE);
9510Sstevel@tonic-gate }
9520Sstevel@tonic-gate 
9530Sstevel@tonic-gate static void
9540Sstevel@tonic-gate rndc_addbytes(uint8_t *ptr, size_t len)
9550Sstevel@tonic-gate {
9560Sstevel@tonic-gate 	ASSERT(ptr != NULL && len > 0);
9570Sstevel@tonic-gate 	ASSERT(rnbyte_cnt <= RNDPOOLSIZE);
9580Sstevel@tonic-gate 
9590Sstevel@tonic-gate 	mutex_enter(&rndpool_lock);
9600Sstevel@tonic-gate 	while ((len > 0) && (rnbyte_cnt < RNDPOOLSIZE)) {
9610Sstevel@tonic-gate 		rndpool[rindex] ^= *ptr;
9620Sstevel@tonic-gate 		ptr++; len--;
9630Sstevel@tonic-gate 		rindex = (rindex + 1) & (RNDPOOLSIZE - 1);
9640Sstevel@tonic-gate 		rnbyte_cnt++;
9650Sstevel@tonic-gate 	}
9660Sstevel@tonic-gate 
9670Sstevel@tonic-gate 	/* Handle buffer full case */
9680Sstevel@tonic-gate 	while (len > 0) {
9690Sstevel@tonic-gate 		rndpool[rindex] ^= *ptr;
9700Sstevel@tonic-gate 		ptr++; len--;
9710Sstevel@tonic-gate 		findex = rindex = (rindex + 1) & (RNDPOOLSIZE - 1);
9720Sstevel@tonic-gate 	}
9730Sstevel@tonic-gate 	mutex_exit(&rndpool_lock);
9740Sstevel@tonic-gate }
9750Sstevel@tonic-gate 
9760Sstevel@tonic-gate /*
9770Sstevel@tonic-gate  * Caller should check len <= rnbyte_cnt under the
9780Sstevel@tonic-gate  * rndpool_lock before calling.
9790Sstevel@tonic-gate  */
9800Sstevel@tonic-gate static void
9810Sstevel@tonic-gate rndc_getbytes(uint8_t *ptr, size_t len)
9820Sstevel@tonic-gate {
9830Sstevel@tonic-gate 	ASSERT(MUTEX_HELD(&rndpool_lock));
9840Sstevel@tonic-gate 	ASSERT(len <= rnbyte_cnt && rnbyte_cnt <= RNDPOOLSIZE);
9850Sstevel@tonic-gate 
9860Sstevel@tonic-gate 	BUMP_RND_STATS(rs_rndcOut, len);
9870Sstevel@tonic-gate 
9880Sstevel@tonic-gate 	while (len > 0) {
9890Sstevel@tonic-gate 		*ptr = rndpool[findex];
9900Sstevel@tonic-gate 		ptr++; len--;
9910Sstevel@tonic-gate 		findex = (findex + 1) & (RNDPOOLSIZE - 1);
9920Sstevel@tonic-gate 		rnbyte_cnt--;
9930Sstevel@tonic-gate 	}
9940Sstevel@tonic-gate }
9950Sstevel@tonic-gate 
9960Sstevel@tonic-gate /* Random number exported entry points */
9970Sstevel@tonic-gate 
9980Sstevel@tonic-gate /*
9990Sstevel@tonic-gate  * Mix the supplied bytes into the entropy pool of a kCF
10000Sstevel@tonic-gate  * RNG provider.
10010Sstevel@tonic-gate  */
10020Sstevel@tonic-gate int
10031920Smcpowers random_add_pseudo_entropy(uint8_t *ptr, size_t len, uint_t entropy_est)
10040Sstevel@tonic-gate {
10050Sstevel@tonic-gate 	if (len < 1)
10060Sstevel@tonic-gate 		return (-1);
10070Sstevel@tonic-gate 
10081920Smcpowers 	rngprov_seed(ptr, len, entropy_est, 0);
10091920Smcpowers 
10101920Smcpowers 	return (0);
10111920Smcpowers }
10121920Smcpowers 
10131920Smcpowers /*
10141920Smcpowers  * Mix the supplied bytes into the entropy pool of a kCF
10151920Smcpowers  * RNG provider. Mix immediately.
10161920Smcpowers  */
10171920Smcpowers int
10181920Smcpowers random_add_entropy(uint8_t *ptr, size_t len, uint_t entropy_est)
10191920Smcpowers {
10201920Smcpowers 	if (len < 1)
10211920Smcpowers 		return (-1);
10221920Smcpowers 
10231920Smcpowers 	rngprov_seed(ptr, len, entropy_est, CRYPTO_SEED_NOW);
10240Sstevel@tonic-gate 
10250Sstevel@tonic-gate 	return (0);
10260Sstevel@tonic-gate }
10270Sstevel@tonic-gate 
10280Sstevel@tonic-gate /*
10290Sstevel@tonic-gate  * Get bytes from the /dev/urandom generator. This function
10300Sstevel@tonic-gate  * always succeeds. Returns 0.
10310Sstevel@tonic-gate  */
10320Sstevel@tonic-gate int
10330Sstevel@tonic-gate random_get_pseudo_bytes(uint8_t *ptr, size_t len)
10340Sstevel@tonic-gate {
10350Sstevel@tonic-gate 	ASSERT(!mutex_owned(&rndpool_lock));
10360Sstevel@tonic-gate 
10370Sstevel@tonic-gate 	if (len < 1)
10380Sstevel@tonic-gate 		return (0);
10390Sstevel@tonic-gate 	return (kcf_rnd_get_pseudo_bytes(ptr, len));
10400Sstevel@tonic-gate }
10410Sstevel@tonic-gate 
10420Sstevel@tonic-gate /*
10430Sstevel@tonic-gate  * Get bytes from the /dev/random generator. Returns 0
10440Sstevel@tonic-gate  * on success. Returns EAGAIN if there is insufficient entropy.
10450Sstevel@tonic-gate  */
10460Sstevel@tonic-gate int
10470Sstevel@tonic-gate random_get_bytes(uint8_t *ptr, size_t len)
10480Sstevel@tonic-gate {
10490Sstevel@tonic-gate 	ASSERT(!mutex_owned(&rndpool_lock));
10500Sstevel@tonic-gate 
10510Sstevel@tonic-gate 	if (len < 1)
10520Sstevel@tonic-gate 		return (0);
10530Sstevel@tonic-gate 	return (kcf_rnd_get_bytes(ptr, len, B_TRUE, B_FALSE));
10540Sstevel@tonic-gate }
1055