semeventwait-r0drv-solaris.h revision 9da340c9fdb54065e988a1d6a8bc3f8a80834ece
/* $Id$ */
/** @file
* IPRT - Solaris Ring-0 Driver Helpers for Event Semaphore Waits.
*/
/*
* Copyright (C) 2006-2010 Oracle Corporation
*
* This file is part of VirtualBox Open Source Edition (OSE), as
* available from http://www.virtualbox.org. This file is free software;
* you can redistribute it and/or modify it under the terms of the GNU
* General Public License (GPL) as published by the Free Software
* Foundation, in version 2 as it comes in the "COPYING" file of the
* VirtualBox OSE distribution. VirtualBox OSE is distributed in the
* hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
*
* The contents of this file may alternatively be used under the terms
* of the Common Development and Distribution License Version 1.0
* (CDDL) only, as it comes in the "COPYING.CDDL" file of the
* VirtualBox OSE distribution, in which case the provisions of the
* CDDL are applicable instead of those of the GPL.
*
* You may elect to license modified versions of this file under the
* terms and conditions of either the GPL or the CDDL or both.
*/
#ifndef ___r0drv_solaris_semeventwait_r0drv_solaris_h
#define ___r0drv_solaris_semeventwait_r0drv_solaris_h
#include "the-solaris-kernel.h"
#include <iprt/err.h>
#include <iprt/string.h>
#include <iprt/time.h>
/** The resolution (nanoseconds) specified when using timeout_generic. */
#define RTR0SEMSOLWAIT_RESOLUTION 50000
/**
* Solaris semaphore wait structure.
*/
typedef struct RTR0SEMSOLWAIT
{
/** The absolute timeout given as nanoseconds since the start of the
* monotonic clock. */
uint64_t uNsAbsTimeout;
/** The timeout in nanoseconds relative to the start of the wait. */
uint64_t cNsRelTimeout;
/** The native timeout value. */
union
{
/** The timeout (abs lbolt) when fHighRes is false. */
clock_t lTimeout;
} u;
/** Set if we use high resolution timeouts. */
bool fHighRes;
/** Set if it's an indefinite wait. */
bool fIndefinite;
/** Set if we've already timed out.
* Set by rtR0SemSolWaitDoIt or rtR0SemSolWaitHighResTimeout, read by
* rtR0SemSolWaitHasTimedOut. */
bool volatile fTimedOut;
/** Whether the wait was interrupted. */
bool fInterrupted;
/** Interruptible or uninterruptible wait. */
bool fInterruptible;
/** The thread to wake up. */
kthread_t *pThread;
#if 0 /* @bugref{5342} */
/** Cylic timer ID (used by the timeout callback). */
cyclic_id_t idCy;
#endif
/** The mutex associated with the condition variable wait. */
void volatile *pvMtx;
} RTR0SEMSOLWAIT;
/** Pointer to a solaris semaphore wait structure. */
typedef RTR0SEMSOLWAIT *PRTR0SEMSOLWAIT;
/**
* Initializes a wait.
*
* The caller MUST check the wait condition BEFORE calling this function or the
* timeout logic will be flawed.
*
* @returns VINF_SUCCESS or VERR_TIMEOUT.
* @param pWait The wait structure.
* @param fFlags The wait flags.
* @param uTimeout The timeout.
* @param pWaitQueue The wait queue head.
*/
DECLINLINE(int) rtR0SemSolWaitInit(PRTR0SEMSOLWAIT pWait, uint32_t fFlags, uint64_t uTimeout)
{
/*
* Process the flags and timeout.
*/
if (!(fFlags & RTSEMWAIT_FLAGS_INDEFINITE))
{
if (fFlags & RTSEMWAIT_FLAGS_MILLISECS)
uTimeout = uTimeout < UINT64_MAX / UINT32_C(1000000) * UINT32_C(1000000)
? uTimeout * UINT32_C(1000000)
: UINT64_MAX;
if (uTimeout == UINT64_MAX)
fFlags |= RTSEMWAIT_FLAGS_INDEFINITE;
else
{
uint64_t u64Now;
if (fFlags & RTSEMWAIT_FLAGS_RELATIVE)
{
if (uTimeout == 0)
return VERR_TIMEOUT;
u64Now = RTTimeSystemNanoTS();
pWait->cNsRelTimeout = uTimeout;
pWait->uNsAbsTimeout = u64Now + uTimeout;
if (pWait->uNsAbsTimeout < u64Now) /* overflow */
fFlags |= RTSEMWAIT_FLAGS_INDEFINITE;
}
else
{
u64Now = RTTimeSystemNanoTS();
if (u64Now >= uTimeout)
return VERR_TIMEOUT;
pWait->cNsRelTimeout = uTimeout - u64Now;
pWait->uNsAbsTimeout = uTimeout;
}
}
}
if (!(fFlags & RTSEMWAIT_FLAGS_INDEFINITE))
{
pWait->fIndefinite = false;
if ( ( (fFlags & (RTSEMWAIT_FLAGS_NANOSECS | RTSEMWAIT_FLAGS_ABSOLUTE))
|| pWait->cNsRelTimeout < UINT32_C(1000000000) / 100 /*Hz*/ * 4)
&& g_pfnrtR0Sol_timeout_generic != NULL /* See @bugref{5342} */)
pWait->fHighRes = true;
else
{
uint64_t cTicks = NSEC_TO_TICK_ROUNDUP(uTimeout);
if (cTicks >= LONG_MAX)
fFlags |= RTSEMWAIT_FLAGS_INDEFINITE;
else
{
pWait->u.lTimeout = ddi_get_lbolt() + cTicks;
pWait->fHighRes = false;
}
}
}
if (fFlags & RTSEMWAIT_FLAGS_INDEFINITE)
{
pWait->fIndefinite = true;
pWait->fHighRes = false;
pWait->uNsAbsTimeout = UINT64_MAX;
pWait->cNsRelTimeout = UINT64_MAX;
pWait->u.lTimeout = LONG_MAX;
}
pWait->fTimedOut = false;
pWait->fInterrupted = false;
pWait->fInterruptible = !!(fFlags & RTSEMWAIT_FLAGS_INTERRUPTIBLE);
pWait->pThread = curthread;
pWait->pvMtx = NULL;
#if 0 /* @bugref{5342} */
pWait->idCy = CYCLIC_NONE;
#endif
return VINF_SUCCESS;
}
#if 0 /* @bugref{5342} */
/**
* Cyclic timeout callback that sets the timeout indicator and wakes up the
* waiting thread.
*
* @param pvUser The wait structure.
*/
static void rtR0SemSolWaitHighResTimeout(void *pvUser)
{
PRTR0SEMSOLWAIT pWait = (PRTR0SEMSOLWAIT)pvUser;
kthread_t *pThread = pWait->pThread;
kmutex_t *pMtx = (kmutex_t *)ASMAtomicReadPtr(&pWait->pvMtx);
if (VALID_PTR(pMtx))
{
/* Enter the mutex here to make sure the thread has gone to sleep
before we wake it up.
Note: Trying to take the cpu_lock here doesn't work. */
/** @todo LOCK ORDER INVERSION (pMtx & cpu_lock when arming the timer, here it's inverted).
* Possible fix: Use the thread lock for sleep/wakeup race prevention
* instead of the mutex associated with the cv/event. */
mutex_enter(pMtx);
if (mutex_owner(&cpu_lock) == curthread)
{
cyclic_remove(pWait->idCy);
pWait->idCy = CYCLIC_NONE;
}
ASMAtomicWriteBool(&pWait->fTimedOut, true);
mutex_exit(pMtx);
setrun(pThread);
}
}
#endif
/**
* Timeout callback that sets the timeout indicator and wakes up the waiting
* thread.
*
* @param pvUser The wait structure.
*/
static void rtR0SemSolWaitTimeout(void *pvUser)
{
PRTR0SEMSOLWAIT pWait = (PRTR0SEMSOLWAIT)pvUser;
kthread_t *pThread = pWait->pThread;
kmutex_t *pMtx = (kmutex_t *)ASMAtomicReadPtr(&pWait->pvMtx);
if (VALID_PTR(pMtx))
{
/* Enter the mutex here to make sure the thread has gone to sleep
before we wake it up. */
mutex_enter(pMtx);
ASMAtomicWriteBool(&pWait->fTimedOut, true);
mutex_exit(pMtx);
setrun(pThread);
}
}
/**
* Do the actual wait.
*
* @param pWait The wait structure.
* @param pCnd The condition variable to wait on.
* @param pMtx The mutex related to the condition variable.
* The caller has entered this.
*/
DECLINLINE(void) rtR0SemSolWaitDoIt(PRTR0SEMSOLWAIT pWait, kcondvar_t *pCnd, kmutex_t *pMtx)
{
union
{
callout_id_t idCo;
timeout_id_t idTom;
} u;
/*
* Arm the timeout callback.
*/
bool const fHasTimeout = !pWait->fIndefinite;
if (fHasTimeout)
{
ASMAtomicWritePtr(&pWait->pvMtx, pMtx); /* atomic is paranoia */
if (pWait->fHighRes)
{
#if 0 /* @bugref{5342} */
if (g_pfnrtR0Sol_timeout_generic != NULL)
#endif
{
/*
* High resolution timeout - arm a high resolution timeout callback
* for waking up the thread at the desired time.
*/
int OldPrioLevel = getpil();
u.idCo = g_pfnrtR0Sol_timeout_generic(CALLOUT_REALTIME, rtR0SemSolWaitTimeout, pWait,
pWait->uNsAbsTimeout, RTR0SEMSOLWAIT_RESOLUTION,
CALLOUT_FLAG_ABSOLUTE);
int NewPrioLevel = getpil();
AssertReleaseMsg(NewPrioLevel >= OldPrioLevel, ("Unexpected lowering of PIL (Old=%d New=%d)\n", OldPrioLevel, NewPrioLevel));
}
#if 0 /* @bugref{5342} */
else
{
/*
* High resolution timeout - arm a one-shot cyclic for waking up
* the thread at the desired time.
*/
cyc_handler_t Cyh;
Cyh.cyh_arg = pWait;
Cyh.cyh_func = rtR0SemSolWaitHighResTimeout;
Cyh.cyh_level = CY_LOW_LEVEL; /// @todo try CY_LOCK_LEVEL and CY_HIGH_LEVEL?
cyc_time_t Cyt;
Cyt.cyt_when = pWait->uNsAbsTimeout;
Cyt.cyt_interval = UINT64_C(1000000000) * 60;
mutex_enter(&cpu_lock);
pWait->idCy = cyclic_add(&Cyh, &Cyt);
mutex_exit(&cpu_lock);
}
#endif
}
else
{
/*
* Normal timeout.
* We're better off with our own callback like on the timeout man page,
* than calling cv_timedwait[_sig]().
*/
u.idTom = realtime_timeout(rtR0SemSolWaitTimeout, pWait, pWait->u.lTimeout);
}
}
/*
* Do the waiting.
* (rc > 0 - normal wake-up; rc == 0 - interruption; rc == -1 - timeout)
*/
if (pWait->fInterruptible)
{
int rc = cv_wait_sig(pCnd, pMtx);
if (RT_UNLIKELY(rc <= 0))
{
if (RT_LIKELY(rc == 0))
pWait->fInterrupted = true;
else
AssertMsgFailed(("rc=%d\n", rc)); /* no timeouts, see above! */
}
}
else
cv_wait(pCnd, pMtx);
/*
* Remove the timeout callback. Drop the lock while we're doing that
* to reduce lock contention / deadlocks. (Too bad we are stuck with the
* cv_* API here, it's doing a little bit too much.)
*/
if (fHasTimeout)
{
ASMAtomicWritePtr(&pWait->pvMtx, NULL);
mutex_exit(pMtx);
if (pWait->fHighRes)
{
#if 0 /* @bugref{5342} */
if (g_pfnrtR0Sol_timeout_generic != NULL)
#endif
g_pfnrtR0Sol_untimeout_generic(u.idCo, 0 /*nowait*/);
#if 0 /* @bugref{5342} */
else
{
mutex_enter(&cpu_lock);
if (pWait->idCy != CYCLIC_NONE)
{
cyclic_remove(pWait->idCy);
pWait->idCy = CYCLIC_NONE;
}
mutex_exit(&cpu_lock);
}
#endif
}
else
untimeout(u.idTom);
mutex_enter(pMtx);
}
}
/**
* Checks if a solaris wait was interrupted.
*
* @returns true / false
* @param pWait The wait structure.
* @remarks This shall be called before the first rtR0SemSolWaitDoIt().
*/
DECLINLINE(bool) rtR0SemSolWaitWasInterrupted(PRTR0SEMSOLWAIT pWait)
{
return pWait->fInterrupted;
}
/**
* Checks if a solaris wait has timed out.
*
* @returns true / false
* @param pWait The wait structure.
*/
DECLINLINE(bool) rtR0SemSolWaitHasTimedOut(PRTR0SEMSOLWAIT pWait)
{
return pWait->fTimedOut;
}
/**
* Deletes a solaris wait.
*
* @param pWait The wait structure.
*/
DECLINLINE(void) rtR0SemSolWaitDelete(PRTR0SEMSOLWAIT pWait)
{
pWait->pThread = NULL;
}
/**
* Enters the mutex, unpinning the underlying current thread if contended and
* we're on an interrupt thread.
*
* The unpinning is done to prevent a deadlock, see s this could lead to a
* deadlock (see #4259 for the full explanation)
*
* @param pMtx The mutex to enter.
*/
DECLINLINE(void) rtR0SemSolWaitEnterMutexWithUnpinningHack(kmutex_t *pMtx)
{
int fAcquired = mutex_tryenter(pMtx);
if (!fAcquired)
{
/*
* Note! This assumes nobody is using the RTThreadPreemptDisable in an
* interrupt context and expects it to work right. The swtch will
* result in a voluntary preemption. To fix this, we would have to
* do our own counting in RTThreadPreemptDisable/Restore like we do
* on systems which doesn't do preemption (OS/2, linux, ...) and
* check whether preemption was disabled via RTThreadPreemptDisable
* or not and only call swtch if RTThreadPreemptDisable wasn't called.
*/
if (curthread->t_intr && getpil() < DISP_LEVEL)
{
RTTHREADPREEMPTSTATE PreemptState = RTTHREADPREEMPTSTATE_INITIALIZER;
RTThreadPreemptDisable(&PreemptState);
preempt();
RTThreadPreemptRestore(&PreemptState);
}
mutex_enter(pMtx);
}
}
/**
* Gets the max resolution of the timeout machinery.
*
* @returns Resolution specified in nanoseconds.
*/
DECLINLINE(uint32_t) rtR0SemSolWaitGetResolution(void)
{
return g_pfnrtR0Sol_timeout_generic != NULL
? RTR0SEMSOLWAIT_RESOLUTION
: cyclic_getres();
}
#endif