summaryrefslogtreecommitdiffstats
path: root/sys/kern
diff options
context:
space:
mode:
authorbmilekic <bmilekic@FreeBSD.org>2001-01-09 23:58:56 +0000
committerbmilekic <bmilekic@FreeBSD.org>2001-01-09 23:58:56 +0000
commit3726db774df1f59602c4e4d8a310a2a197bb000f (patch)
tree00890872a21fe95579cac062e374c418968161c7 /sys/kern
parentf2dd7e71f7ccc84c1b9e0b2181b6a9c7fb528e2c (diff)
downloadFreeBSD-src-3726db774df1f59602c4e4d8a310a2a197bb000f.zip
FreeBSD-src-3726db774df1f59602c4e4d8a310a2a197bb000f.tar.gz
In m_mballoc_wait(), drop the mmbfree mutex lock prior to calling
m_reclaim() and re-acquire it when m_reclaim() returns. This means that we now call the drain routines without holding the mutex lock and recursing into it. This was done for mainly two reasons: (i) Avoid the long recursion; long recursions are typically bad and this is the case here because we block all other code from freeing mbufs if they need to. Doing that is kind of counter-productive, since we're really hoping that someone will free. (ii) More importantly, avoid a potential lock order reversal. Right now, not all the locks have been added to our networking code; but without this change, we're introducing the possibility for deadlock. Consider for example ip_drain(). We will likely eventually introduce a lock for ipq there, and so ip_freef() will be called with ipq lock held. But, ip_freef() calls m_freem() which in turn acquires the mmbfree lock. Since we were previously calling ip_drain() with mmbfree held, our lock order would be: mmbfree->ipq->mmbfree. Some other code may very well lock ipq first and then call ip_freef(). This would result in the regular lock order, ipq->mmbfree. Clearly, we have deadlock if one thread acquires the ipq lock and sits waiting for mmbfree while another thread calling m_reclaim() acquires mmbfree and sits waiting for the ipq lock. Also, make sure to add a comment above m_reclaim()'s definition briefly explaining this. Also document this above the call to m_reclaim() in m_mballoc_wait(). Suggested and reviewed by: alfred
Diffstat (limited to 'sys/kern')
-rw-r--r--sys/kern/uipc_mbuf.c27
1 files changed, 15 insertions, 12 deletions
diff --git a/sys/kern/uipc_mbuf.c b/sys/kern/uipc_mbuf.c
index d8cc445..4afacb2 100644
--- a/sys/kern/uipc_mbuf.c
+++ b/sys/kern/uipc_mbuf.c
@@ -309,10 +309,7 @@ m_mballoc(nmb, how)
* still cannot get anything, then we wait for an mbuf to be freed for a
* designated (mbuf_wait) time.
*
- * Must be called with the mmbfree mutex held, and we will probably end
- * up recursing into that lock from some of the drain routines, but
- * this should be okay, as long as we don't block there, or attempt
- * to allocate from them (theoretically impossible).
+ * Must be called with the mmbfree mutex held.
*/
struct mbuf *
m_mballoc_wait(void)
@@ -321,8 +318,18 @@ m_mballoc_wait(void)
/*
* See if we can drain some resources out of the protocols.
+ * We drop the mmbfree mutex to avoid recursing into it in some of
+ * the drain routines. Clearly, we're faced with a race here because
+ * once something is freed during the drain, it may be grabbed right
+ * from under us by some other thread. But we accept this possibility
+ * in order to avoid a potentially large lock recursion and, more
+ * importantly, to avoid a potential lock order reversal which may
+ * result in deadlock (See comment above m_reclaim()).
*/
+ mtx_exit(&mmbfree.m_mtx, MTX_DEF);
m_reclaim();
+
+ mtx_enter(&mmbfree.m_mtx, MTX_DEF);
_MGET(p, M_DONTWAIT);
if (p == NULL) {
@@ -449,14 +456,10 @@ m_clalloc_wait(void)
/*
* m_reclaim: drain protocols in hopes to free up some resources...
*
- * Should be called with mmbfree.m_mtx mutex held. We will most likely
- * recursively grab it from within some drain routines, but that's okay,
- * as the mutex will never be completely released until we let go of it
- * after our m_reclaim() is over.
- *
- * Note: Drain routines are only allowed to free mbufs (and mclusters,
- * as a consequence, if need be). They are not allowed to allocate
- * new ones (that would defeat the purpose, anyway).
+ * XXX: No locks should be held going in here. The drain routines have
+ * to presently acquire some locks which raises the possibility of lock
+ * order violation if we're holding any mutex if that mutex is acquired in
+ * reverse order relative to one of the locks in the drain routines.
*/
static void
m_reclaim()
OpenPOWER on IntegriCloud