Lines Matching refs:rndpool_lock
48 * 2) rndpool_lock protects the high-quality randomness pool.
109 static kmutex_t rndpool_lock; /* protects r/w accesses to the cache, */
140 mutex_init(&rndpool_lock, NULL, MUTEX_DEFAULT, NULL);
462 * Called with rndpool_lock held (allowing caller to do optimistic locking;
471 ASSERT(mutex_owned(&rndpool_lock));
478 mutex_exit(&rndpool_lock);
481 mutex_exit(&rndpool_lock);
512 mutex_enter(&rndpool_lock);
519 &rndpool_lock) == 0) {
521 mutex_exit(&rndpool_lock);
552 mutex_exit(&rndpool_lock);
563 mutex_enter(&rndpool_lock);
660 (!mutex_tryenter(&rndpool_lock))) {
667 mutex_enter(&rndpool_lock);
865 /* rnd_get_bytes() will call mutex_exit(&rndpool_lock) */
866 mutex_enter(&rndpool_lock);
871 /* rnd_get_bytes() will call mutex_exit(&rndpool_lock) */
872 mutex_enter(&rndpool_lock);
875 /* rnd_get_bytes() will call mutex_exit(&rndpool_lock) */
876 mutex_enter(&rndpool_lock);
972 mutex_enter(&rndpool_lock);
983 * Note - Since we hold the rndpool_lock across the pollwakeup() call
984 * we MUST NOT grab the rndpool_lock in kcf_rndchpoll().
991 mutex_exit(&rndpool_lock);
1002 mutex_enter(&rndpool_lock);
1016 mutex_exit(&rndpool_lock);
1021 * rndpool_lock before calling.
1026 ASSERT(MUTEX_HELD(&rndpool_lock));
1078 ASSERT(!mutex_owned(&rndpool_lock));
1092 ASSERT(!mutex_owned(&rndpool_lock));
1102 ASSERT(!mutex_owned(&rndpool_lock));