diff options
| author | Jason Evans <jasone@FreeBSD.org> | 2000-10-12 22:37:28 +0000 |
|---|---|---|
| committer | Jason Evans <jasone@FreeBSD.org> | 2000-10-12 22:37:28 +0000 |
| commit | 9722d88fba2c4016b66c281a16a1be84a32af95c (patch) | |
| tree | 7ce9eab3c738ff1f9586ceac1b646763bab57396 /sys | |
| parent | 1df2031c1502b456210934bb3838bc23c64bb5d7 (diff) | |
Notes
Diffstat (limited to 'sys')
| -rw-r--r-- | sys/conf/param.c | 8 | ||||
| -rw-r--r-- | sys/kern/kern_lock.c | 101 | ||||
| -rw-r--r-- | sys/kern/subr_param.c | 8 | ||||
| -rw-r--r-- | sys/sys/kernel.h | 1 | ||||
| -rw-r--r-- | sys/sys/lock.h | 8 | ||||
| -rw-r--r-- | sys/sys/lockmgr.h | 8 | ||||
| -rw-r--r-- | sys/vm/vm_map.h | 8 |
7 files changed, 102 insertions, 40 deletions
diff --git a/sys/conf/param.c b/sys/conf/param.c index 42b124d751b7..a7841e45f192 100644 --- a/sys/conf/param.c +++ b/sys/conf/param.c @@ -96,3 +96,11 @@ int nswbuf = 0; * (if they've been externed everywhere else; hah!). */ struct buf *swbuf; + +/* + * Total number of shared mutexes to protect all lockmgr locks. + */ +#ifndef LOCKMUTEX +#define LOCKMUTEX 10 +#endif +int lock_nmtx = LOCKMUTEX; diff --git a/sys/kern/kern_lock.c b/sys/kern/kern_lock.c index 12c687080345..9265763d09f1 100644 --- a/sys/kern/kern_lock.c +++ b/sys/kern/kern_lock.c @@ -43,7 +43,9 @@ #include <sys/param.h> #include <sys/proc.h> +#include <sys/kernel.h> #include <sys/lock.h> +#include <sys/malloc.h> #include <sys/systm.h> #include <machine/mutex.h> @@ -71,10 +73,48 @@ #define LK_ALL (LK_HAVE_EXCL | LK_WANT_EXCL | LK_WANT_UPGRADE | \ LK_SHARE_NONZERO | LK_WAIT_NONZERO) +/* + * Mutex array variables. Rather than each lockmgr lock having its own mutex, + * share a fixed (at boot time) number of mutexes across all lockmgr locks in + * order to keep sizeof(struct lock) down. + */ +extern int lock_nmtx; +int lock_mtx_selector; +struct mtx *lock_mtx_array; +struct mtx lock_mtx; + static int acquire(struct lock *lkp, int extflags, int wanted); static int apause(struct lock *lkp, int flags); static int acquiredrain(struct lock *lkp, int extflags) ; +static void +lockmgr_init(void *dummy __unused) +{ + int i; + + /* + * Initialize the lockmgr protection mutex if it hasn't already been + * done. Unless something changes about kernel startup order, VM + * initialization will always cause this mutex to already be + * initialized in a call to lockinit(). + */ + if (lock_mtx_selector == 0) + mtx_init(&lock_mtx, "lockmgr", MTX_DEF); + else { + /* + * This is necessary if (lock_nmtx == 1) and doesn't hurt + * otherwise. + */ + lock_mtx_selector = 0; + } + + lock_mtx_array = (struct mtx *)malloc(sizeof(struct mtx) * lock_nmtx, + M_CACHE, M_WAITOK); + for (i = 0; i < lock_nmtx; i++) + mtx_init(&lock_mtx_array[i], "lockmgr interlock", MTX_DEF); +} +SYSINIT(lmgrinit, SI_SUB_LOCK, SI_ORDER_FIRST, lockmgr_init, NULL) + static LOCK_INLINE void sharelock(struct lock *lkp, int incr) { lkp->lk_flags |= LK_SHARE_NONZERO; @@ -113,11 +153,11 @@ apause(struct lock *lkp, int flags) return 0; #ifdef SMP for (lock_wait = LOCK_WAIT_TIME; lock_wait > 0; lock_wait--) { - mtx_exit(&lkp->lk_interlock, MTX_DEF); + mtx_exit(lkp->lk_interlock, MTX_DEF); for (i = LOCK_SAMPLE_WAIT; i > 0; i--) if ((lkp->lk_flags & flags) == 0) break; - mtx_enter(&lkp->lk_interlock, MTX_DEF); + mtx_enter(lkp->lk_interlock, MTX_DEF); if ((lkp->lk_flags & flags) == 0) return 0; } @@ -147,9 +187,9 @@ acquire(struct lock *lkp, int extflags, int wanted) { while ((lkp->lk_flags & wanted) != 0) { lkp->lk_flags |= LK_WAIT_NONZERO; lkp->lk_waitcount++; - mtx_exit(&lkp->lk_interlock, MTX_DEF); + mtx_exit(lkp->lk_interlock, MTX_DEF); error = tsleep(lkp, lkp->lk_prio, lkp->lk_wmesg, lkp->lk_timo); - mtx_enter(&lkp->lk_interlock, MTX_DEF); + mtx_enter(lkp->lk_interlock, MTX_DEF); if (lkp->lk_waitcount == 1) { lkp->lk_flags &= ~LK_WAIT_NONZERO; lkp->lk_waitcount = 0; @@ -206,7 +246,7 @@ debuglockmgr(lkp, flags, interlkp, p, name, file, line) else pid = p->p_pid; - mtx_enter(&lkp->lk_interlock, MTX_DEF); + mtx_enter(lkp->lk_interlock, MTX_DEF); if (flags & LK_INTERLOCK) mtx_exit(interlkp, MTX_DEF); @@ -434,7 +474,7 @@ debuglockmgr(lkp, flags, interlkp, p, name, file, line) break; default: - mtx_exit(&lkp->lk_interlock, MTX_DEF); + mtx_exit(lkp->lk_interlock, MTX_DEF); panic("lockmgr: unknown locktype request %d", flags & LK_TYPE_MASK); /* NOTREACHED */ @@ -445,7 +485,7 @@ debuglockmgr(lkp, flags, interlkp, p, name, file, line) lkp->lk_flags &= ~LK_WAITDRAIN; wakeup((void *)&lkp->lk_flags); } - mtx_exit(&lkp->lk_interlock, MTX_DEF); + mtx_exit(lkp->lk_interlock, MTX_DEF); return (error); } @@ -463,10 +503,10 @@ acquiredrain(struct lock *lkp, int extflags) { while (lkp->lk_flags & LK_ALL) { lkp->lk_flags |= LK_WAITDRAIN; - mtx_exit(&lkp->lk_interlock, MTX_DEF); + mtx_exit(lkp->lk_interlock, MTX_DEF); error = tsleep(&lkp->lk_flags, lkp->lk_prio, lkp->lk_wmesg, lkp->lk_timo); - mtx_enter(&lkp->lk_interlock, MTX_DEF); + mtx_enter(lkp->lk_interlock, MTX_DEF); if (error) return error; if (extflags & LK_SLEEPFAIL) { @@ -490,11 +530,32 @@ lockinit(lkp, prio, wmesg, timo, flags) CTR5(KTR_LOCKMGR, "lockinit(): lkp == %p, prio == %d, wmesg == \"%s\", " "timo == %d, flags = 0x%x\n", lkp, prio, wmesg, timo, flags); - if (lkp->lk_flags & LK_VALID) - lockdestroy(lkp); - - mtx_init(&lkp->lk_interlock, "lockmgr interlock", MTX_DEF); - lkp->lk_flags = (flags & LK_EXTFLG_MASK) | LK_VALID; + if (lock_mtx_array != NULL) { + mtx_enter(&lock_mtx, MTX_DEF); + lkp->lk_interlock = &lock_mtx_array[lock_mtx_selector]; + lock_mtx_selector++; + if (lock_mtx_selector == lock_nmtx) + lock_mtx_selector = 0; + mtx_exit(&lock_mtx, MTX_DEF); + } else { + /* + * Giving lockmgr locks that are initialized during boot a + * pointer to the internal lockmgr mutex is safe, since the + * lockmgr code itself doesn't call lockinit() (which could + * cause mutex recursion). + */ + if (lock_mtx_selector == 0) { + /* + * This case only happens during kernel bootstrapping, + * so there's no reason to protect modification of + * lock_mtx_selector or lock_mtx. + */ + mtx_init(&lock_mtx, "lockmgr", MTX_DEF); + lock_mtx_selector = 1; + } + lkp->lk_interlock = &lock_mtx; + } + lkp->lk_flags = (flags & LK_EXTFLG_MASK); lkp->lk_sharecount = 0; lkp->lk_waitcount = 0; lkp->lk_exclusivecount = 0; @@ -513,10 +574,6 @@ lockdestroy(lkp) { CTR2(KTR_LOCKMGR, "lockdestroy(): lkp == %p (lk_wmesg == \"%s\")", lkp, lkp->lk_wmesg); - if (lkp->lk_flags & LK_VALID) { - lkp->lk_flags &= ~LK_VALID; - mtx_destroy(&lkp->lk_interlock); - } } /* @@ -529,7 +586,7 @@ lockstatus(lkp, p) { int lock_type = 0; - mtx_enter(&lkp->lk_interlock, MTX_DEF); + mtx_enter(lkp->lk_interlock, MTX_DEF); if (lkp->lk_exclusivecount != 0) { if (p == NULL || lkp->lk_lockholder == p->p_pid) lock_type = LK_EXCLUSIVE; @@ -537,7 +594,7 @@ lockstatus(lkp, p) lock_type = LK_EXCLOTHER; } else if (lkp->lk_sharecount != 0) lock_type = LK_SHARED; - mtx_exit(&lkp->lk_interlock, MTX_DEF); + mtx_exit(lkp->lk_interlock, MTX_DEF); return (lock_type); } @@ -550,9 +607,9 @@ lockcount(lkp) { int count; - mtx_enter(&lkp->lk_interlock, MTX_DEF); + mtx_enter(lkp->lk_interlock, MTX_DEF); count = lkp->lk_exclusivecount + lkp->lk_sharecount; - mtx_exit(&lkp->lk_interlock, MTX_DEF); + mtx_exit(lkp->lk_interlock, MTX_DEF); return (count); } diff --git a/sys/kern/subr_param.c b/sys/kern/subr_param.c index 42b124d751b7..a7841e45f192 100644 --- a/sys/kern/subr_param.c +++ b/sys/kern/subr_param.c @@ -96,3 +96,11 @@ int nswbuf = 0; * (if they've been externed everywhere else; hah!). */ struct buf *swbuf; + +/* + * Total number of shared mutexes to protect all lockmgr locks. + */ +#ifndef LOCKMUTEX +#define LOCKMUTEX 10 +#endif +int lock_nmtx = LOCKMUTEX; diff --git a/sys/sys/kernel.h b/sys/sys/kernel.h index b30250840054..cc50426ef133 100644 --- a/sys/sys/kernel.h +++ b/sys/sys/kernel.h @@ -113,6 +113,7 @@ enum sysinit_sub_id { SI_SUB_VM = 0x1000000, /* virtual memory system init*/ SI_SUB_KMEM = 0x1800000, /* kernel memory*/ SI_SUB_KVM_RSRC = 0x1A00000, /* kvm operational limits*/ + SI_SUB_LOCK = 0x1B00000, /* lockmgr locks */ SI_SUB_CPU = 0x2000000, /* CPU resource(s)*/ SI_SUB_KLD = 0x2100000, /* KLD and module setup */ SI_SUB_INTRINSIC = 0x2200000, /* proc 0*/ diff --git a/sys/sys/lock.h b/sys/sys/lock.h index 1e126b13ea0a..d11b485f75b7 100644 --- a/sys/sys/lock.h +++ b/sys/sys/lock.h @@ -50,8 +50,7 @@ * can be gained. The simple locks are defined in <machine/param.h>. */ struct lock { - struct mtx lk_interlock; /* lock on remaining fields */ - struct mtxf lk_pad; /* padding to keep sizeof constant */ + struct mtx *lk_interlock; /* lock on remaining fields */ u_int lk_flags; /* see below */ int lk_sharecount; /* # of accepted shared locks */ int lk_waitcount; /* # of processes sleeping for lock */ @@ -133,11 +132,6 @@ struct lock { #define LK_HAVE_EXCL 0x00000400 /* exclusive lock obtained */ #define LK_WAITDRAIN 0x00000800 /* process waiting for lock to drain */ #define LK_DRAINING 0x00004000 /* lock is being drained */ -#define LK_VALID 0x00008000 /* - * Lock is initialized. This is a - * temporary hack to support vfs - * layering. - */ /* * Control flags * diff --git a/sys/sys/lockmgr.h b/sys/sys/lockmgr.h index 1e126b13ea0a..d11b485f75b7 100644 --- a/sys/sys/lockmgr.h +++ b/sys/sys/lockmgr.h @@ -50,8 +50,7 @@ * can be gained. The simple locks are defined in <machine/param.h>. */ struct lock { - struct mtx lk_interlock; /* lock on remaining fields */ - struct mtxf lk_pad; /* padding to keep sizeof constant */ + struct mtx *lk_interlock; /* lock on remaining fields */ u_int lk_flags; /* see below */ int lk_sharecount; /* # of accepted shared locks */ int lk_waitcount; /* # of processes sleeping for lock */ @@ -133,11 +132,6 @@ struct lock { #define LK_HAVE_EXCL 0x00000400 /* exclusive lock obtained */ #define LK_WAITDRAIN 0x00000800 /* process waiting for lock to drain */ #define LK_DRAINING 0x00004000 /* lock is being drained */ -#define LK_VALID 0x00008000 /* - * Lock is initialized. This is a - * temporary hack to support vfs - * layering. - */ /* * Control flags * diff --git a/sys/vm/vm_map.h b/sys/vm/vm_map.h index d23848805569..ef48af20d5df 100644 --- a/sys/vm/vm_map.h +++ b/sys/vm/vm_map.h @@ -291,15 +291,15 @@ _vm_map_lock_upgrade(vm_map_t map, struct proc *p) { #define vm_map_set_recursive(map) \ do { \ - mtx_enter(&(map)->lock.lk_interlock, MTX_DEF); \ + mtx_enter((map)->lock.lk_interlock, MTX_DEF); \ (map)->lock.lk_flags |= LK_CANRECURSE; \ - mtx_exit(&(map)->lock.lk_interlock, MTX_DEF); \ + mtx_exit((map)->lock.lk_interlock, MTX_DEF); \ } while(0) #define vm_map_clear_recursive(map) \ do { \ - mtx_enter(&(map)->lock.lk_interlock, MTX_DEF); \ + mtx_enter((map)->lock.lk_interlock, MTX_DEF); \ (map)->lock.lk_flags &= ~LK_CANRECURSE; \ - mtx_exit(&(map)->lock.lk_interlock, MTX_DEF); \ + mtx_exit((map)->lock.lk_interlock, MTX_DEF); \ } while(0) /* |
