cvs commit: src/sys/kern kern_proc.c
Bosko Milekic
bmilekic at FreeBSD.org
Wed Jun 9 16:18:26 GMT 2004
Warner wrote:
>Poul-Henning wrote:
>: Not to pick on anybody, but this is a perfect example of getting locking
>: almost right:
>:
>: BAD:
>:
>: LOCK(foo->lock)
>: foo->refcount--;
>: UNLOCK(foo->lock)
>: if (foo->refcount == 0)
>: destroy(foo);
>:
>: GOOD:
>:
>: LOCK(foo->lock)
>: i = --foo->refcount;
>: UNLOCK(foo->lock)
>: if (i == 0)
>: destroy(foo);
>:
>
>Can you provide a couple of lines about why BAD is BAD and why GOOD
>fixes that flaw? That should help others from making this mistake in
>the future.
>
>Warner
Frankly, I think it is obvious.
In the BAD case, you decrement safely but you compare for the refcount
having hit zero without the lock protecting the count being taken,
so what happens is that there is a race window between where you drop
the lock and check for the zero refcount where another racing thread
could drop the refcount finally to zero and go into the check as
well. Briefly, this can happen:
- refcount of 'foo' is 2.
- thread 1 enters BAD code, lowers refcount to 1, releases lock.
- thread 2 enters BAD code, lowers refcount to 0, releases lock.
- thread 1 checks against refcount being zero, decides it is now,
and proceeds to destroy(foo).
- thread 2 checks against refcount being zero, decides it is now,
and proceeds to destroy(foo) as well.
Conclusion: foo is destroyed twice.
The GOOD code does not suffer from this problem. Here is a way to
handle this sort of race if your reference counter is instead
manipulated atomically (as opposed to protected by a mutex):
[From Mbuf-related code]
MEXT_REM_REF(m); /* Atomic decrement of m->m_ext.ref_cnt */
if (atomic_cmpset_int(m->m_ext.ref_cnt, 0, 1)) {
/* Do the free here... */
}
return;
-Bosko
More information about the cvs-src
mailing list