diff options
author | Alfred Perlstein <alfred@FreeBSD.org> | 2001-11-29 17:36:45 +0000 |
---|---|---|
committer | Alfred Perlstein <alfred@FreeBSD.org> | 2001-11-29 17:36:45 +0000 |
commit | 8c7daee73581b8141f868e0fde09f5aab6112fe4 (patch) | |
tree | fb05d693361e60fc16dc8d0cc21d8ff39e4d80bd /usr.sbin/rpc.lockd | |
parent | a240291a2aa5de5f442eebf31068c6de7d4ad883 (diff) | |
parent | 5f07c7b294e285993b00990b3dd11a5390972168 (diff) | |
download | src-test2-8c7daee73581b8141f868e0fde09f5aab6112fe4.tar.gz src-test2-8c7daee73581b8141f868e0fde09f5aab6112fe4.zip |
Notes
Diffstat (limited to 'usr.sbin/rpc.lockd')
-rw-r--r-- | usr.sbin/rpc.lockd/lockd.c | 10 | ||||
-rw-r--r-- | usr.sbin/rpc.lockd/lockd.h | 1 | ||||
-rw-r--r-- | usr.sbin/rpc.lockd/lockd_lock.c | 2520 |
3 files changed, 1965 insertions, 566 deletions
diff --git a/usr.sbin/rpc.lockd/lockd.c b/usr.sbin/rpc.lockd/lockd.c index 0397514e5dcd..e78c69d8b37d 100644 --- a/usr.sbin/rpc.lockd/lockd.c +++ b/usr.sbin/rpc.lockd/lockd.c @@ -93,7 +93,7 @@ main(argc, argv) { SVCXPRT *transp; int ch, i, maxindex, s; - struct sigaction sigchild, sigalarm; + struct sigaction sigalarm; int grace_period = 30; struct netconfig *nconf; @@ -188,14 +188,6 @@ main(argc, argv) else syslog(LOG_INFO, "Starting"); - sigchild.sa_handler = sigchild_handler; - sigemptyset(&sigchild.sa_mask); - sigchild.sa_flags = SA_RESTART; - if (sigaction(SIGCHLD, &sigchild, NULL) != 0) { - syslog(LOG_WARNING, "sigaction(SIGCHLD) failed: %s", - strerror(errno)); - exit(1); - } sigalarm.sa_handler = (sig_t) sigalarm_handler; sigemptyset(&sigalarm.sa_mask); sigalarm.sa_flags = SA_RESETHAND; /* should only happen once */ diff --git a/usr.sbin/rpc.lockd/lockd.h b/usr.sbin/rpc.lockd/lockd.h index 464ac3a99f9e..0a8e2e60b616 100644 --- a/usr.sbin/rpc.lockd/lockd.h +++ b/usr.sbin/rpc.lockd/lockd.h @@ -36,7 +36,6 @@ extern int debug_level; extern int grace_expired; -void sigchild_handler __P((int)); pid_t client_request(void); extern int nsm_state; extern pid_t client_pid; diff --git a/usr.sbin/rpc.lockd/lockd_lock.c b/usr.sbin/rpc.lockd/lockd_lock.c index 70b511599fa6..a75bdd1c54cb 100644 --- a/usr.sbin/rpc.lockd/lockd_lock.c +++ b/usr.sbin/rpc.lockd/lockd_lock.c @@ -2,6 +2,7 @@ /* $FreeBSD$ */ /* + * Copyright (c) 2001 Andrew P. Lentvorski, Jr. * Copyright (c) 2000 Manuel Bouyer. * * Redistribution and use in source and binary forms, with or without @@ -34,7 +35,12 @@ * */ +#define LOCKD_DEBUG + #include <stdio.h> +#ifdef LOCKD_DEBUG +#include <stdarg.h> +#endif #include <stdlib.h> #include <unistd.h> #include <fcntl.h> @@ -54,30 +60,44 @@ #include "lockd_lock.h" #include "lockd.h" +#define MAXOBJECTSIZE 64 +#define MAXBUFFERSIZE 1024 + +/* + * SM_MAXSTRLEN is usually 1024. This means that lock requests and + * host name monitoring entries are *MUCH* larger than they should be + */ + /* * A set of utilities for managing file locking * * XXX: All locks are in a linked list, a better structure should be used * to improve search/access effeciency. */ -LIST_HEAD(lcklst_head, file_lock); -struct lcklst_head lcklst_head = LIST_HEAD_INITIALIZER(lcklst_head); /* struct describing a lock */ struct file_lock { - LIST_ENTRY(file_lock) lcklst; + LIST_ENTRY(file_lock) nfslocklist; fhandle_t filehandle; /* NFS filehandle */ struct sockaddr *addr; struct nlm4_holder client; /* lock holder */ + /* XXX: client_cookie used *only* in send_granted */ netobj client_cookie; /* cookie sent by the client */ - char client_name[128]; + char client_name[SM_MAXSTRLEN]; int nsm_status; /* status from the remote lock manager */ int status; /* lock status, see below */ int flags; /* lock flags, see lockd_lock.h */ + int blocking; /* blocking lock or not */ pid_t locker; /* pid of the child process trying to get the lock */ int fd; /* file descriptor for this lock */ }; +LIST_HEAD(nfslocklist_head, file_lock); +struct nfslocklist_head nfslocklist_head = LIST_HEAD_INITIALIZER(nfslocklist_head); + +LIST_HEAD(blockedlocklist_head, file_lock); +struct blockedlocklist_head blockedlocklist_head = LIST_HEAD_INITIALIZER(blockedlocklist_head); + /* lock status */ #define LKST_LOCKED 1 /* lock is locked */ /* XXX: Is this flag file specific or lock specific? */ @@ -85,494 +105,2005 @@ struct file_lock { #define LKST_PROCESSING 3 /* child is trying to aquire the lock */ #define LKST_DYING 4 /* must dies when we get news from the child */ -void lfree __P((struct file_lock *)); -enum nlm_stats do_lock __P((struct file_lock *, int)); -enum nlm_stats do_unlock __P((struct file_lock *)); -void send_granted __P((struct file_lock *, int)); -void siglock __P((void)); -void sigunlock __P((void)); -int regions_overlap __P((u_int64_t start1, u_int64_t len1, u_int64_t start2, - u_int64_t len2)); - -/* list of hosts we monitor */ -LIST_HEAD(hostlst_head, host); -struct hostlst_head hostlst_head = LIST_HEAD_INITIALIZER(hostlst_head); - -/* struct describing a lock */ +/* struct describing a monitored host */ struct host { LIST_ENTRY(host) hostlst; char name[SM_MAXSTRLEN]; int refcnt; }; +/* list of hosts we monitor */ +LIST_HEAD(hostlst_head, host); +struct hostlst_head hostlst_head = LIST_HEAD_INITIALIZER(hostlst_head); + +/* + * File monitoring handlers + * XXX: These might be able to be removed when kevent support + * is placed into the hardware lock/unlock routines. (ie. + * let the kernel do all the file monitoring) + */ + +/* Struct describing a monitored file */ +struct monfile { + LIST_ENTRY(monfile) monfilelist; + fhandle_t filehandle; /* Local access filehandle */ + int fd; /* file descriptor: remains open until unlock! */ + int refcount; + int exclusive; +}; + +/* List of files we monitor */ +LIST_HEAD(monfilelist_head, monfile); +struct monfilelist_head monfilelist_head = LIST_HEAD_INITIALIZER(monfilelist_head); + +static int debugdelay = 0; + +enum nfslock_status { NFS_GRANTED = 0, NFS_GRANTED_DUPLICATE, + NFS_DENIED, NFS_DENIED_NOLOCK, + NFS_RESERR }; + +enum hwlock_status { HW_GRANTED = 0, HW_GRANTED_DUPLICATE, + HW_DENIED, HW_DENIED_NOLOCK, + HW_STALEFH, HW_READONLY, HW_RESERR }; + +enum partialfilelock_status { PFL_GRANTED=0, PFL_GRANTED_DUPLICATE, PFL_DENIED, + PFL_NFSDENIED, PFL_NFSBLOCKED, PFL_NFSDENIED_NOLOCK, PFL_NFSRESERR, + PFL_HWDENIED, PFL_HWBLOCKED, PFL_HWDENIED_NOLOCK, PFL_HWRESERR}; + +enum LFLAGS {LEDGE_LEFT, LEDGE_LBOUNDARY, LEDGE_INSIDE, LEDGE_RBOUNDARY, LEDGE_RIGHT}; +enum RFLAGS {REDGE_LEFT, REDGE_LBOUNDARY, REDGE_INSIDE, REDGE_RBOUNDARY, REDGE_RIGHT}; +/* XXX: WARNING! I HAVE OVERLOADED THIS STATUS ENUM! SPLIT IT APART INTO TWO */ +enum split_status {SPL_DISJOINT=0, SPL_LOCK1=1, SPL_LOCK2=2, SPL_CONTAINED=4, SPL_RESERR=8}; + +enum partialfilelock_status lock_partialfilelock(struct file_lock *fl); + +void send_granted(struct file_lock *fl, int opcode); +void siglock(void); +void sigunlock(void); +void monitor_lock_host(const char *hostname); +void unmonitor_lock_host(const char *hostname); + +void copy_nlm4_lock_to_nlm4_holder(const struct nlm4_lock *src, + const bool_t exclusive, struct nlm4_holder *dest); +void deallocate_file_lock(struct file_lock *fl); +int regions_overlap(const u_int64_t start1, const u_int64_t len1, + const u_int64_t start2, const u_int64_t len2);; +enum split_status region_compare(const u_int64_t starte, const u_int64_t lene, + const u_int64_t startu, const u_int64_t lenu, + u_int64_t *start1, u_int64_t *len1, u_int64_t *start2, u_int64_t *len2); +int same_netobj(const netobj *n0, const netobj *n1); +int same_filelock_identity(const struct file_lock *fl0, + const struct file_lock *fl2); + +static void debuglog(char const *fmt, ...); +void dump_static_object(const unsigned char* object, const int sizeof_object, + unsigned char* hbuff, const int sizeof_hbuff, + unsigned char* cbuff, const int sizeof_cbuff); +void dump_netobj(const struct netobj *nobj); +void dump_filelock(const struct file_lock *fl); +struct file_lock * get_lock_matching_unlock(const struct file_lock *fl); +enum nfslock_status test_nfslock(const struct file_lock *fl, + struct file_lock **conflicting_fl); +enum nfslock_status lock_nfslock(struct file_lock *fl); +enum nfslock_status delete_nfslock(struct file_lock *fl); +enum nfslock_status unlock_nfslock(const struct file_lock *fl, + struct file_lock **released_lock, struct file_lock **left_lock, + struct file_lock **right_lock); +enum hwlock_status lock_hwlock(struct file_lock *fl); +enum split_status split_nfslock(const struct file_lock *exist_lock, + const struct file_lock *unlock_lock, struct file_lock **left_lock, + struct file_lock **right_lock); +void add_blockingfilelock(struct file_lock *fl); +enum hwlock_status unlock_hwlock(const struct file_lock *fl); +enum hwlock_status test_hwlock(const struct file_lock *fl, + struct file_lock **conflicting_fl); +void remove_blockingfilelock(struct file_lock *fl); +void clear_blockingfilelock(const char *hostname); +void retry_blockingfilelocklist(void); +enum partialfilelock_status unlock_partialfilelock( + const struct file_lock *fl); +void clear_partialfilelock(const char *hostname); +enum partialfilelock_status test_partialfilelock( + const struct file_lock *fl, struct file_lock **conflicting_fl); +enum nlm_stats do_test(struct file_lock *fl, + struct file_lock **conflicting_fl); +enum nlm_stats do_unlock(struct file_lock *fl); +enum nlm_stats do_lock(struct file_lock *fl); +void do_clear(const char *hostname); + + +void +debuglog(char const *fmt, ...) +{ + va_list ap; + + if (debug_level < 1) { + return; + } + + sleep(debugdelay); + + va_start(ap, fmt); + vsyslog(LOG_DEBUG, fmt, ap); + va_end(ap); +} + +void +dump_static_object(object, size_object, hbuff, size_hbuff, cbuff, size_cbuff) + const unsigned char *object; + const int size_object; + unsigned char *hbuff; + const int size_hbuff; + unsigned char *cbuff; + const int size_cbuff; +{ + int i, objectsize; + + if (debug_level < 2) { + return; + } + + objectsize = size_object; + + if (objectsize == 0) { + debuglog("object is size 0\n"); + } else { + if (objectsize > MAXOBJECTSIZE) { + debuglog("Object of size %d being clamped" + "to size %d\n", objectsize, MAXOBJECTSIZE); + objectsize = MAXOBJECTSIZE; + } + + if (hbuff != NULL) { + if (size_hbuff < objectsize*2+1) { + debuglog("Hbuff not large enough." + " Increase size\n"); + } else { + for(i=0;i<objectsize;i++) { + sprintf(hbuff+i*2,"%02x",*(object+i)); + } + *(hbuff+i*2) = '\0'; + } + } + + if (cbuff != NULL) { + if (size_cbuff < objectsize+1) { + debuglog("Cbuff not large enough." + " Increase Size\n"); + } + + for(i=0;i<objectsize;i++) { + if (*(object+i) >= 32 && *(object+i) <= 127) { + *(cbuff+i) = *(object+i); + } else { + *(cbuff+i) = '.'; + } + } + *(cbuff+i) = '\0'; + } + } +} + +void +dump_netobj(const struct netobj *nobj) +{ + char hbuff[MAXBUFFERSIZE*2]; + char cbuff[MAXBUFFERSIZE]; + + if (debug_level < 2) { + return; + } + + if (nobj == NULL) { + debuglog("Null netobj pointer\n"); + } + else if (nobj->n_len == 0) { + debuglog("Size zero netobj\n"); + } else { + dump_static_object(nobj->n_bytes, nobj->n_len, + hbuff, sizeof(hbuff), cbuff, sizeof(cbuff)); + debuglog("netobj: len: %d data: %s ::: %s\n", + nobj->n_len, hbuff, cbuff); + } +} + +void +dump_filelock(const struct file_lock *fl) +{ + char hbuff[MAXBUFFERSIZE*2]; + char cbuff[MAXBUFFERSIZE]; + + if (debug_level < 2) { + return; + } + + if (fl != NULL) { + debuglog("Dumping file lock structure @ %p\n", fl); + + /* + dump_static_object((unsigned char *)&fl->filehandle, + sizeof(fl->filehandle), hbuff, sizeof(hbuff), + cbuff, sizeof(cbuff)); + debuglog("Filehandle: %8s ::: %8s\n", hbuff, cbuff); + */ + + debuglog("Dumping nlm4_holder:\n" + "exc: %x svid: %x offset:len %llx:%llx\n", + fl->client.exclusive, fl->client.svid, + fl->client.l_offset, fl->client.l_len); + + /* + debuglog("Dumping client identity:\n"); + dump_netobj(&fl->client.oh); + + debuglog("Dumping client cookie:\n"); + dump_netobj(&fl->client_cookie); + + debuglog("nsm: %d status: %d flags: %d locker: %d" + " fd: %d\n", fl->nsm_status, fl->status, + fl->flags, fl->locker, fl->fd); + */ + } else { + debuglog("NULL file lock structure\n"); + } +} + +void +copy_nlm4_lock_to_nlm4_holder(src, exclusive, dest) + const struct nlm4_lock *src; + const bool_t exclusive; + struct nlm4_holder *dest; +{ + + dest->exclusive = exclusive; + dest->oh.n_len = src->oh.n_len; + dest->oh.n_bytes = src->oh.n_bytes; + dest->svid = src->svid; + dest->l_offset = src->l_offset; + dest->l_len = src->l_len; +} + + +/* + * allocate_file_lock: Create a lock with the given parameters + */ + +struct file_lock * +allocate_file_lock(const netobj *lockowner, const netobj *matchcookie) +{ + struct file_lock *newfl; + + newfl = malloc(sizeof(struct file_lock)); + if (newfl == NULL) { + return NULL; + } + bzero(newfl, sizeof(newfl)); + + newfl->client.oh.n_bytes = malloc(lockowner->n_len); + if (newfl->client.oh.n_bytes == NULL) { + free(newfl); + return NULL; + } + newfl->client.oh.n_len = lockowner->n_len; + bcopy(lockowner->n_bytes, newfl->client.oh.n_bytes, lockowner->n_len); + + newfl->client_cookie.n_bytes = malloc(matchcookie->n_len); + if (newfl->client_cookie.n_bytes == NULL) { + free(newfl->client.oh.n_bytes); + free(newfl); + return NULL; + } + newfl->client_cookie.n_len = matchcookie->n_len; + bcopy(matchcookie->n_bytes, newfl->client_cookie.n_bytes, matchcookie->n_len); + + return newfl; +} -void do_mon __P((char *)); +/* + * file_file_lock: Force creation of a valid file lock + */ +void +fill_file_lock(struct file_lock *fl, const fhandle_t *fh, struct sockaddr *addr, + const bool_t exclusive, const int32_t svid, const u_int64_t offset, const u_int64_t len, + const char* caller_name, const int state, const int status, const int flags, const int blocking) +{ + bcopy(fh, &fl->filehandle, sizeof(fhandle_t)); + fl->addr = addr; + + fl->client.exclusive = exclusive; + fl->client.svid = svid; + fl->client.l_offset = offset; + fl->client.l_len = len; + + strncpy(fl->client_name, caller_name, SM_MAXSTRLEN); + + fl->nsm_status = state; + fl->status = status; + fl->flags = flags; + fl->blocking = blocking; +} + +/* + * deallocate_file_lock: Free all storage associated with a file lock + */ +void +deallocate_file_lock(struct file_lock *fl) +{ + free(fl->client.oh.n_bytes); + free(fl->client_cookie.n_bytes); + free(fl); +} /* - * regions_overlap(): This function examines the two provided regions for overlap. - * It is non-trivial because start+len *CAN* overflow a 64-bit unsigned integer - * and NFS semantics are unspecified on this account. + * regions_overlap(): This function examines the two provided regions for + * overlap. */ int regions_overlap(start1, len1, start2, len2) - u_int64_t start1, len1, start2, len2; + const u_int64_t start1, len1, start2, len2; { - int result; - - /* XXX: Need to adjust checks to account for integer overflow */ - if (len1 == 0 && len2 == 0) { - /* Regions *must* overlap if they both extend to the end */ - result = TRUE; - } else if (len1 == 0 && start2+len2 < start1) { - /* Region 2 is completely to the left of Region 1 */ - result = FALSE; - } else if (start1+len1 < start2 && len2 == 0) { - /* Region 1 is completely to the left of region 2 */ - result = FALSE; - } else if (start1 + len1 <= start2 || start2+len2 <= start1) { - /* 1 is completely left of 2 or 2 is completely left of 1 */ - result = FALSE; + u_int64_t d1,d2,d3,d4; + enum split_status result; + + debuglog("Entering region overlap with vals: %llu:%llu--%llu:%llu\n", + start1, len1, start2, len2); + + result = region_compare(start1, len1, start2, len2, + &d1, &d2, &d3, &d4); + + debuglog("Exiting region overlap with val: %d\n",result); + + if (result == SPL_DISJOINT) { + return 0; } else { - result = TRUE; + return 1; } + return (result); } + /* - * testlock(): inform the caller if the requested lock would be granted or not - * returns NULL if lock would granted, or pointer to the current nlm4_holder - * otherwise. + * region_compare(): Examine lock regions and split appropriately + * + * XXX: Fix 64 bit overflow problems + * XXX: Check to make sure I got *ALL* the cases. + * XXX: This DESPERATELY needs a regression test. */ +enum split_status +region_compare(starte, lene, startu, lenu, + start1, len1, start2, len2) + const u_int64_t starte, lene, startu, lenu; + u_int64_t *start1, *len1, *start2, *len2; +{ + /* + * Please pay attention to the sequential exclusions + * of the if statements!!! + */ + enum LFLAGS lflags; + enum RFLAGS rflags; + enum split_status retval; -struct nlm4_holder * -testlock(lock, exclusive, flags) - struct nlm4_lock *lock; - bool_t exclusive; - int flags; + retval = SPL_DISJOINT; + + if (lene == 0 && lenu == 0) { + /* Examine left edge of locker */ + if (startu < starte) { + lflags = LEDGE_LEFT; + } else if (startu == starte) { + lflags = LEDGE_LBOUNDARY; + } else { + lflags = LEDGE_INSIDE; + } + + rflags = REDGE_RBOUNDARY; /* Both are infiinite */ + + if (lflags == LEDGE_INSIDE) { + *start1 = starte; + *len1 = startu - starte; + } + + if (lflags == LEDGE_LEFT || lflags == LEDGE_LBOUNDARY) { + retval = SPL_CONTAINED; + } else { + retval = SPL_LOCK1; + } + } else if (lene == 0 && lenu != 0) { + /* Established lock is infinite */ + /* Examine left edge of unlocker */ + if (startu < starte) { + lflags = LEDGE_LEFT; + } else if (startu == starte) { + lflags = LEDGE_LBOUNDARY; + } else if (startu > starte) { + lflags = LEDGE_INSIDE; + } + + /* Examine right edge of unlocker */ + if (startu + lenu < starte) { + /* Right edge of unlocker left of established lock */ + rflags = REDGE_LEFT; + return SPL_DISJOINT; + } else if (startu + lenu == starte) { + /* Right edge of unlocker on start of established lock */ + rflags = REDGE_LBOUNDARY; + return SPL_DISJOINT; + } else { /* Infinifty is right of finity */ + /* Right edge of unlocker inside established lock */ + rflags = REDGE_INSIDE; + } + + if (lflags == LEDGE_INSIDE) { + *start1 = starte; + *len1 = startu - starte; + retval |= SPL_LOCK1; + } + + if (rflags == REDGE_INSIDE) { + /* Create right lock */ + *start2 = startu+lenu; + *len2 = 0; + retval |= SPL_LOCK2; + } + } else if (lene != 0 && lenu == 0) { + /* Unlocker is infinite */ + /* Examine left edge of unlocker */ + if (startu < starte) { + lflags = LEDGE_LEFT; + retval = SPL_CONTAINED; + return retval; + } else if (startu == starte) { + lflags = LEDGE_LBOUNDARY; + retval = SPL_CONTAINED; + return retval; + } else if ((startu > starte) && (startu < starte + lene - 1)) { + lflags = LEDGE_INSIDE; + } else if (startu == starte + lene - 1) { + lflags = LEDGE_RBOUNDARY; + } else { /* startu > starte + lene -1 */ + lflags = LEDGE_RIGHT; + return SPL_DISJOINT; + } + + rflags = REDGE_RIGHT; /* Infinity is right of finity */ + + if (lflags == LEDGE_INSIDE || lflags == LEDGE_RBOUNDARY) { + *start1 = starte; + *len1 = startu - starte; + retval |= SPL_LOCK1; + return retval; + } + + } else { + /* Both locks are finite */ + + /* Examine left edge of unlocker */ + if (startu < starte) { + lflags = LEDGE_LEFT; + } else if (startu == starte) { + lflags = LEDGE_LBOUNDARY; + } else if ((startu > starte) && (startu < starte + lene - 1)) { + lflags = LEDGE_INSIDE; + } else if (startu == starte + lene - 1) { + lflags = LEDGE_RBOUNDARY; + } else { /* startu > starte + lene -1 */ + lflags = LEDGE_RIGHT; + return SPL_DISJOINT; + } + + /* Examine right edge of unlocker */ + if (startu + lenu < starte) { + /* Right edge of unlocker left of established lock */ + rflags = REDGE_LEFT; + return SPL_DISJOINT; + } else if (startu + lenu == starte) { + /* Right edge of unlocker on start of established lock */ + rflags = REDGE_LBOUNDARY; + return SPL_DISJOINT; + } else if (startu + lenu < starte + lene) { + /* Right edge of unlocker inside established lock */ + rflags = REDGE_INSIDE; + } else if (startu + lenu == starte + lene) { + /* Right edge of unlocker on right edge of established lock */ + rflags = REDGE_RBOUNDARY; + } else { /* startu + lenu > starte + lene */ + /* Right edge of unlocker is right of established lock */ + rflags = REDGE_RIGHT; + } + + if (lflags == LEDGE_INSIDE || lflags == LEDGE_RBOUNDARY) { + /* Create left lock */ + *start1 = starte; + *len1 = (startu - starte); + retval |= SPL_LOCK1; + } + + if (rflags == REDGE_INSIDE) { + /* Create right lock */ + *start2 = startu+lenu; + *len2 = starte+lene-(startu+lenu); + retval |= SPL_LOCK2; + } + + if ((lflags == LEDGE_LEFT || lflags == LEDGE_LBOUNDARY) && + (rflags == REDGE_RBOUNDARY || rflags == REDGE_RIGHT)) { + retval = SPL_CONTAINED; + } + } + + return retval; +} + +/* + * same_netobj: Compares the apprpriate bits of a netobj for identity + */ +int +same_netobj(const netobj *n0, const netobj *n1) { - struct file_lock *fl; - fhandle_t filehandle; + int retval; - /* convert lock to a local filehandle */ - memcpy(&filehandle, lock->fh.n_bytes, sizeof(filehandle)); + retval = 0; - siglock(); - /* search through the list for lock holder */ - for (fl = LIST_FIRST(&lcklst_head); fl != NULL; - fl = LIST_NEXT(fl, lcklst)) { - if (fl->status != LKST_LOCKED) + debuglog("Entering netobj identity check\n"); + + if (n0->n_len == n1->n_len) { + debuglog("Preliminary length check passed\n"); + retval = !bcmp(n0->n_bytes, n1->n_bytes, n0->n_len); + debuglog("netobj %smatch\n", retval ? "" : "mis"); + } + + return (retval); +} + +/* + * same_filelock_identity: Compares the appropriate bits of a file_lock + */ +int +same_filelock_identity(fl0, fl1) + const struct file_lock *fl0, *fl1; +{ + int retval; + + retval = 0; + + debuglog("Checking filelock identity\n"); + + /* + * Check process ids and host information. + */ + retval = (fl0->client.svid == fl1->client.svid && + same_netobj(&(fl0->client.oh), &(fl1->client.oh))); + + debuglog("Exiting checking filelock identity: retval: %d\n",retval); + + return (retval); +} + +/* + * Below here are routines associated with manipulating the NFS + * lock list. + */ + +/* + * get_lock_matching_unlock: Return a lock which matches the given unlock lock + * or NULL otehrwise + * XXX: It is a shame that this duplicates so much code from test_nfslock. + */ +struct file_lock * +get_lock_matching_unlock(const struct file_lock *fl) +{ + struct file_lock *ifl; /* Iterator */ + + debuglog("Entering lock_matching_unlock\n"); + debuglog("********Dump of fl*****************\n"); + dump_filelock(fl); + + LIST_FOREACH(ifl, &nfslocklist_head, nfslocklist) { + debuglog("Pointer to file lock: %p\n",ifl); + + debuglog("****Dump of ifl****\n"); + dump_filelock(ifl); + debuglog("*******************\n"); + + /* + * XXX: It is conceivable that someone could use the NLM RPC + * system to directly access filehandles. This may be a + * security hazard as the filehandle code may bypass normal + * file access controls + */ + if (bcmp(&fl->filehandle, &ifl->filehandle, sizeof(fhandle_t))) + continue; + + debuglog("matching_unlock: Filehandles match, " + "checking regions\n"); + + /* Filehandles match, check for region overlap */ + if (!regions_overlap(fl->client.l_offset, fl->client.l_len, + ifl->client.l_offset, ifl->client.l_len)) continue; + + debuglog("matching_unlock: Region overlap" + " found %llu : %llu -- %llu : %llu\n", + fl->client.l_offset,fl->client.l_len, + ifl->client.l_offset,ifl->client.l_len); + + /* Regions overlap, check the identity */ + if (!same_filelock_identity(fl,ifl)) + continue; + + debuglog("matching_unlock: Duplicate lock id. Granting\n"); + return (ifl); + } + + debuglog("Exiting lock_matching_unlock\n"); + + return (NULL); +} + +/* + * test_nfslock: check for NFS lock in lock list + * + * This routine makes the following assumptions: + * 1) Nothing will adjust the lock list during a lookup + * + * This routine has an intersting quirk which bit me hard. + * The conflicting_fl is the pointer to the conflicting lock. + * However, to modify the "*pointer* to the conflicting lock" rather + * that the "conflicting lock itself" one must pass in a "pointer to + * the pointer of the conflicting lock". Gross. + */ + +enum nfslock_status +test_nfslock(const struct file_lock *fl, struct file_lock **conflicting_fl) +{ + struct file_lock *ifl; /* Iterator */ + enum nfslock_status retval; + + debuglog("Entering test_nfslock\n"); + + retval = NFS_GRANTED; + (*conflicting_fl) = NULL; + + debuglog("Entering lock search loop\n"); + + debuglog("***********************************\n"); + debuglog("Dumping match filelock\n"); + debuglog("***********************************\n"); + dump_filelock(fl); + debuglog("***********************************\n"); + + LIST_FOREACH(ifl, &nfslocklist_head, nfslocklist) { + if (retval == NFS_DENIED) + break; + + debuglog("Top of lock loop\n"); + debuglog("Pointer to file lock: %p\n",ifl); + + debuglog("***********************************\n"); + debuglog("Dumping test filelock\n"); + debuglog("***********************************\n"); + dump_filelock(ifl); + debuglog("***********************************\n"); + /* - * XXX: Could we possibly have identical filehandles - * on different systems? - * ie. Do we need to check more than just the filehandle? - * ie. Could someone artificially create requests which are - * security violations? + * XXX: It is conceivable that someone could use the NLM RPC + * system to directly access filehandles. This may be a + * security hazard as the filehandle code may bypass normal + * file access controls */ - if (memcmp(&fl->filehandle, &filehandle, sizeof(filehandle))) + if (bcmp(&fl->filehandle, &ifl->filehandle, sizeof(fhandle_t))) continue; - /* File handles match, look for lock region overlap */ - if (regions_overlap(lock->l_offset, lock->l_len, - fl->client.l_offset, fl->client.l_len)) { - syslog(LOG_DEBUG, - "Region overlap found %llu : %llu -- %llu : %llu\n", - lock->l_offset, lock->l_len, - fl->client.l_offset,fl->client.l_len); - /* Regions overlap. Now check for exclusivity. */ - if (exclusive || fl->client.exclusive) { - /* Lock test must fail, regions are exclusive */ - break; - } + + debuglog("test_nfslock: filehandle match found\n"); + + /* Filehandles match, check for region overlap */ + if (!regions_overlap(fl->client.l_offset, fl->client.l_len, + ifl->client.l_offset, ifl->client.l_len)) + continue; + + debuglog("test_nfslock: Region overlap found" + " %llu : %llu -- %llu : %llu\n", + fl->client.l_offset,fl->client.l_len, + ifl->client.l_offset,ifl->client.l_len); + + /* Regions overlap, check the exclusivity */ + if (!(fl->client.exclusive || ifl->client.exclusive)) + continue; + + debuglog("test_nfslock: Exclusivity failure: %d %d\n", + fl->client.exclusive, + ifl->client.exclusive); + + if (same_filelock_identity(fl,ifl)) { + debuglog("test_nfslock: Duplicate id. Granting\n"); + (*conflicting_fl) = ifl; + retval = NFS_GRANTED_DUPLICATE; + } else { + /* locking attempt fails */ + debuglog("test_nfslock: Lock attempt failed\n"); + debuglog("Desired lock\n"); + dump_filelock(fl); + debuglog("Conflicting lock\n"); + dump_filelock(ifl); + (*conflicting_fl) = ifl; + retval = NFS_DENIED; } - /* Continue looping through all locks */ - } - sigunlock(); - if (fl == NULL) { - syslog(LOG_DEBUG, "test for %s: no lock found", - lock->caller_name); - return NULL; - } else { - syslog(LOG_DEBUG, "test for %s: found lock held by %s", - lock->caller_name, fl->client_name); - return (&fl->client); } + + debuglog("Dumping file locks\n"); + debuglog("Exiting test_nfslock\n"); + + return (retval); } /* - * getlock: try to aquire the lock. - * If file is already locked and we can sleep, put the lock in the list with - * status LKST_WAITING; it'll be processed later. - * Otherwise try to lock. If we're allowed to block, fork a child which - * will do the blocking lock. + * lock_nfslock: attempt to create a lock in the NFS lock list + * + * This routine tests whether the lock will be granted and then adds + * the entry to the lock list if so. + * + * Argument fl gets modified as its list housekeeping entries get modified + * upon insertion into the NFS lock list + * + * This routine makes several assumptions: + * 1) It is perfectly happy to grant a duplicate lock from the same pid. + * While this seems to be intuitively wrong, it is required for proper + * Posix semantics during unlock. It is absolutely imperative to not + * unlock the main lock before the two child locks are established. Thus, + * one has be be able to create duplicate locks over an existing lock + * 2) It currently accepts duplicate locks from the same id,pid */ -enum nlm_stats -getlock(lckarg, rqstp, flags) - nlm4_lockargs * lckarg; - struct svc_req *rqstp; - int flags; + +enum nfslock_status +lock_nfslock(struct file_lock *fl) { - struct file_lock *fl, *newfl; - enum nlm_stats retval; + enum nfslock_status retval; + struct file_lock *dummy_fl; - if (grace_expired == 0 && lckarg->reclaim == 0) - return (flags & LOCK_V4) ? - nlm4_denied_grace_period : nlm_denied_grace_period; - - /* allocate new file_lock for this request */ - newfl = malloc(sizeof(struct file_lock)); - if (newfl == NULL) { - syslog(LOG_NOTICE, "malloc failed: %s", strerror(errno)); - /* failed */ - return (flags & LOCK_V4) ? - nlm4_denied_nolocks : nlm_denied_nolocks; - } - if (lckarg->alock.fh.n_len != sizeof(fhandle_t)) { - syslog(LOG_DEBUG, "recieved fhandle size %d, local size %d", - lckarg->alock.fh.n_len, (int)sizeof(fhandle_t)); + dummy_fl = NULL; + + debuglog("Entering lock_nfslock...\n"); + + retval = test_nfslock(fl,&dummy_fl); + + if (retval == NFS_GRANTED || retval == NFS_GRANTED_DUPLICATE) { + debuglog("Inserting lock...\n"); + dump_filelock(fl); + LIST_INSERT_HEAD(&nfslocklist_head, fl, nfslocklist); } - memcpy(&newfl->filehandle, lckarg->alock.fh.n_bytes, sizeof(fhandle_t)); - newfl->addr = (struct sockaddr *)svc_getrpccaller(rqstp->rq_xprt)->buf; - newfl->client.exclusive = lckarg->exclusive; - newfl->client.svid = lckarg->alock.svid; - newfl->client.oh.n_bytes = malloc(lckarg->alock.oh.n_len); - if (newfl->client.oh.n_bytes == NULL) { - syslog(LOG_NOTICE, "malloc failed: %s", strerror(errno)); - free(newfl); - return (flags & LOCK_V4) ? - nlm4_denied_nolocks : nlm_denied_nolocks; + + debuglog("Exiting lock_nfslock...\n"); + + return (retval); +} + +/* + * delete_nfslock: delete an NFS lock list entry + * + * This routine is used to delete a lock out of the NFS lock list + * without regard to status, underlying locks, regions or anything else + * + * Note that this routine *does not deallocate memory* of the lock. + * It just disconnects it from the list. The lock can then be used + * by other routines without fear of trashing the list. + */ + +enum nfslock_status +delete_nfslock(struct file_lock *fl) +{ + + LIST_REMOVE(fl, nfslocklist); + + return (NFS_GRANTED); +} + +enum split_status +split_nfslock(exist_lock, unlock_lock, left_lock, right_lock) + const struct file_lock *exist_lock, *unlock_lock; + struct file_lock **left_lock, **right_lock; +{ + u_int64_t start1, len1, start2, len2; + enum split_status spstatus; + + spstatus = region_compare(exist_lock->client.l_offset, exist_lock->client.l_len, + unlock_lock->client.l_offset, unlock_lock->client.l_len, + &start1, &len1, &start2, &len2); + + if ((spstatus & SPL_LOCK1) != 0) { + *left_lock = allocate_file_lock(&exist_lock->client.oh, &exist_lock->client_cookie); + if (*left_lock == NULL) { + debuglog("Unable to allocate resource for split 1\n"); + return SPL_RESERR; + } + + fill_file_lock(*left_lock, &exist_lock->filehandle, exist_lock->addr, + exist_lock->client.exclusive, exist_lock->client.svid, + start1, len1, + exist_lock->client_name, exist_lock->nsm_status, + exist_lock->status, exist_lock->flags, exist_lock->blocking); } - newfl->client.oh.n_len = lckarg->alock.oh.n_len; - memcpy(newfl->client.oh.n_bytes, lckarg->alock.oh.n_bytes, - lckarg->alock.oh.n_len); - newfl->client.l_offset = lckarg->alock.l_offset; - newfl->client.l_len = lckarg->alock.l_len; - newfl->client_cookie.n_len = lckarg->cookie.n_len; - newfl->client_cookie.n_bytes = malloc(lckarg->cookie.n_len); - if (newfl->client_cookie.n_bytes == NULL) { - syslog(LOG_NOTICE, "malloc failed: %s", strerror(errno)); - free(newfl->client.oh.n_bytes); - free(newfl); - return (flags & LOCK_V4) ? - nlm4_denied_nolocks : nlm_denied_nolocks; + + if ((spstatus & SPL_LOCK2) != 0) { + *right_lock = allocate_file_lock(&exist_lock->client.oh, &exist_lock->client_cookie); + if (*right_lock == NULL) { + debuglog("Unable to allocate resource for split 1\n"); + if (*left_lock != NULL) { + deallocate_file_lock(*left_lock); + } + return SPL_RESERR; + } + + fill_file_lock(*right_lock, &exist_lock->filehandle, exist_lock->addr, + exist_lock->client.exclusive, exist_lock->client.svid, + start2, len2, + exist_lock->client_name, exist_lock->nsm_status, + exist_lock->status, exist_lock->flags, exist_lock->blocking); } - memcpy(newfl->client_cookie.n_bytes, lckarg->cookie.n_bytes, - lckarg->cookie.n_len); - strncpy(newfl->client_name, lckarg->alock.caller_name, 128); - newfl->nsm_status = lckarg->state; - newfl->status = 0; - newfl->flags = flags; - siglock(); - /* look for a lock rq from this host for this fh */ - for (fl = LIST_FIRST(&lcklst_head); fl != NULL; - fl = LIST_NEXT(fl, lcklst)) { - if (memcmp(&newfl->filehandle, &fl->filehandle, - sizeof(fhandle_t)) == 0) { - if (strcmp(newfl->client_name, fl->client_name) == 0 && - newfl->client.svid == fl->client.svid) { - /* already locked by this host ??? */ - sigunlock(); - syslog(LOG_NOTICE, "duplicate lock from %s", - newfl->client_name); - lfree(newfl); - switch(fl->status) { - case LKST_LOCKED: - return (flags & LOCK_V4) ? - nlm4_granted : nlm_granted; - case LKST_WAITING: - case LKST_PROCESSING: - return (flags & LOCK_V4) ? - nlm4_blocked : nlm_blocked; - case LKST_DYING: - return (flags & LOCK_V4) ? - nlm4_denied : nlm_denied; - default: - syslog(LOG_NOTICE, "bad status %d", - fl->status); - return (flags & LOCK_V4) ? - nlm4_failed : nlm_denied; - } + + return spstatus; +} + +enum nfslock_status +unlock_nfslock(fl, released_lock, left_lock, right_lock) + const struct file_lock *fl; + struct file_lock **released_lock; + struct file_lock **left_lock; + struct file_lock **right_lock; +{ + struct file_lock *mfl; /* Matching file lock */ + enum nfslock_status retval; + enum split_status spstatus; + + debuglog("Entering unlock_nfslock\n"); + + *released_lock = NULL; + *left_lock = NULL; + *right_lock = NULL; + + retval = NFS_DENIED_NOLOCK; + + printf("Attempting to match lock...\n"); + mfl = get_lock_matching_unlock(fl); + + if (mfl != NULL) { + debuglog("Unlock matched. Querying for split\n"); + + spstatus = split_nfslock(mfl, fl, left_lock, right_lock); + + debuglog("Split returned %d %p %p %p %p\n",spstatus,mfl,fl,*left_lock,*right_lock); + debuglog("********Split dumps********"); + dump_filelock(mfl); + dump_filelock(fl); + dump_filelock(*left_lock); + dump_filelock(*right_lock); + debuglog("********End Split dumps********"); + + if (spstatus == SPL_RESERR) { + if (*left_lock != NULL) { + deallocate_file_lock(*left_lock); + *left_lock = NULL; } - /* - * We already have a lock for this file. Put this one - * in waiting state if allowed to block - */ - if (lckarg->block) { - syslog(LOG_DEBUG, "lock from %s: already " - "locked, waiting", - lckarg->alock.caller_name); - newfl->status = LKST_WAITING; - LIST_INSERT_HEAD(&lcklst_head, newfl, lcklst); - do_mon(lckarg->alock.caller_name); - sigunlock(); - return (flags & LOCK_V4) ? - nlm4_blocked : nlm_blocked; - } else { - sigunlock(); - syslog(LOG_DEBUG, "lock from %s: already " - "locked, failed", - lckarg->alock.caller_name); - lfree(newfl); - return (flags & LOCK_V4) ? - nlm4_denied : nlm_denied; + + if (*right_lock != NULL) { + deallocate_file_lock(*right_lock); + *right_lock = NULL; } + + return NFS_RESERR; } + + /* Insert new locks from split if required */ + if (*left_lock != NULL) { + debuglog("Split left activated\n"); + LIST_INSERT_HEAD(&nfslocklist_head, *left_lock, nfslocklist); + } + + if (*right_lock != NULL) { + debuglog("Split right activated\n"); + LIST_INSERT_HEAD(&nfslocklist_head, *right_lock, nfslocklist); + } + + /* Unlock the lock since it matches identity */ + LIST_REMOVE(mfl, nfslocklist); + *released_lock = mfl; + retval = NFS_GRANTED; } - /* no entry for this file yet; add to list */ - LIST_INSERT_HEAD(&lcklst_head, newfl, lcklst); - /* do the lock */ - retval = do_lock(newfl, lckarg->block); - switch (retval) { - case nlm4_granted: - /* case nlm_granted: is the same as nlm4_granted */ - case nlm4_blocked: - /* case nlm_blocked: is the same as nlm4_blocked */ - do_mon(lckarg->alock.caller_name); - break; - default: - lfree(newfl); - break; - } - sigunlock(); + + debuglog("Exiting unlock_nfslock\n"); + return retval; } -/* unlock a filehandle */ -enum nlm_stats -unlock(lck, flags) - nlm4_lock *lck; - int flags; +/* + * Below here are the routines for manipulating the file lock directly + * on the disk hardware itself + */ +enum hwlock_status +lock_hwlock(struct file_lock *fl) { - struct file_lock *fl; - fhandle_t filehandle; - int err = (flags & LOCK_V4) ? nlm4_granted : nlm_granted; + struct monfile *imf,*nmf; + int lflags, flerror; - memcpy(&filehandle, lck->fh.n_bytes, sizeof(fhandle_t)); - siglock(); - for (fl = LIST_FIRST(&lcklst_head); fl != NULL; - fl = LIST_NEXT(fl, lcklst)) { - if (strcmp(fl->client_name, lck->caller_name) || - memcmp(&filehandle, &fl->filehandle, sizeof(fhandle_t)) || - fl->client.oh.n_len != lck->oh.n_len || - memcmp(fl->client.oh.n_bytes, lck->oh.n_bytes, - fl->client.oh.n_len) != 0 || - fl->client.svid != lck->svid) - continue; - /* Got it, unlock and remove from the queue */ - syslog(LOG_DEBUG, "unlock from %s: found struct, status %d", - lck->caller_name, fl->status); - switch (fl->status) { - case LKST_LOCKED: - err = do_unlock(fl); + /* Scan to see if filehandle already present */ + LIST_FOREACH(imf, &monfilelist_head, monfilelist) { + if (bcmp(&fl->filehandle, &imf->filehandle, + sizeof(fl->filehandle)) == 0) { + /* imf is the correct filehandle */ break; - case LKST_WAITING: - /* remove from the list */ - LIST_REMOVE(fl, lcklst); - lfree(fl); - break; - case LKST_PROCESSING: - /* - * being handled by a child; will clean up - * when the child exits - */ - fl->status = LKST_DYING; + } + } + + /* + * Filehandle already exists (we control the file) + * *AND* NFS has already cleared the lock for availability + * Grant it and bump the refcount. + */ + if (imf != NULL) { + ++(imf->refcount); + return (HW_GRANTED); + } + + /* No filehandle found, create and go */ + nmf = malloc(sizeof(struct monfile)); + if (nmf == NULL) { + debuglog("hwlock resource allocation failure\n"); + return (HW_RESERR); + } + + /* XXX: Is O_RDWR always the correct mode? */ + nmf->fd = fhopen(&fl->filehandle, O_RDWR); + if (nmf->fd < 0) { + debuglog("fhopen failed (from %16s): %32s\n", + fl->client_name, strerror(errno)); + free(nmf); + switch (errno) { + case ESTALE: + return (HW_STALEFH); + case EROFS: + return (HW_READONLY); + default: + return (HW_RESERR); + } + } + + /* File opened correctly, fill the monitor struct */ + bcopy(&fl->filehandle, &nmf->filehandle, sizeof(fl->filehandle)); + nmf->refcount = 1; + nmf->exclusive = fl->client.exclusive; + + lflags = (nmf->exclusive == 1) ? + (LOCK_EX | LOCK_NB) : (LOCK_SH | LOCK_NB); + + flerror = flock(nmf->fd, lflags); + + if (flerror != 0) { + debuglog("flock failed (from %16s): %32s\n", + fl->client_name, strerror(errno)); + close(nmf->fd); + free(nmf); + switch (errno) { + case EAGAIN: + return (HW_DENIED); + case ESTALE: + return (HW_STALEFH); + case EROFS: + return (HW_READONLY); + default: + return (HW_RESERR); break; - case LKST_DYING: - /* nothing to do */ + } + } + + /* File opened and locked */ + LIST_INSERT_HEAD(&monfilelist_head, nmf, monfilelist); + + debuglog("flock succeeded (from %16s)\n", fl->client_name); + return (HW_GRANTED); +} + +enum hwlock_status +unlock_hwlock(const struct file_lock *fl) +{ + struct monfile *imf; + + debuglog("Entering unlock_hwlock\n"); + debuglog("Entering loop interation\n"); + + /* Scan to see if filehandle already present */ + LIST_FOREACH(imf, &monfilelist_head, monfilelist) { + if (bcmp(&fl->filehandle, &imf->filehandle, + sizeof(fl->filehandle)) == 0) { + /* imf is the correct filehandle */ break; - default: - syslog(LOG_NOTICE, "unknow status %d for %s", - fl->status, fl->client_name); } - sigunlock(); - return err; } - sigunlock(); - /* didn't find a matching entry; log anyway */ - syslog(LOG_NOTICE, "no matching entry for %s", - lck->caller_name); - return (flags & LOCK_V4) ? nlm4_granted : nlm_granted; + + debuglog("Completed iteration. Proceeding\n"); + + if (imf == NULL) { + /* No lock found */ + debuglog("Exiting unlock_hwlock (HW_DENIED_NOLOCK)\n"); + return (HW_DENIED_NOLOCK); + } + + /* Lock found */ + --imf->refcount; + + if (imf->refcount < 0) { + debuglog("Negative hardware reference count\n"); + } + + if (imf->refcount <= 0) { + close(imf->fd); + LIST_REMOVE(imf, monfilelist); + free(imf); + } + debuglog("Exiting unlock_hwlock (HW_GRANTED)\n"); + return (HW_GRANTED); } +enum hwlock_status +test_hwlock(const struct file_lock *fl, struct file_lock **conflicting_fl) +{ + + /* + * XXX: lock tests on hardware are not required until + * true partial file testing is done on the underlying file + */ + return (HW_RESERR); +} + + + +/* + * Below here are routines for manipulating blocked lock requests + * They should only be called from the XXX_partialfilelock routines + * if at all possible + */ + void -lfree(fl) - struct file_lock *fl; +add_blockingfilelock(struct file_lock *fl) { - free(fl->client.oh.n_bytes); - free(fl->client_cookie.n_bytes); - free(fl); + + debuglog("Entering add_blockingfilelock\n"); + + /* + * Clear the blocking flag so that it can be reused without + * adding it to the blocking queue a second time + */ + + fl->blocking = 0; + LIST_INSERT_HEAD(&blockedlocklist_head, fl, nfslocklist); + + debuglog("Exiting add_blockingfilelock\n"); } void -sigchild_handler(sig) - int sig; +remove_blockingfilelock(struct file_lock *fl) { - int status; - pid_t pid; - struct file_lock *fl; - while (1) { - pid = wait4(-1, &status, WNOHANG, NULL); - if (pid == -1) { - if (errno != ECHILD) - syslog(LOG_NOTICE, "wait failed: %s", - strerror(errno)); - else - syslog(LOG_DEBUG, "wait failed: %s", - strerror(errno)); - return; + debuglog("Entering remove_blockingfilelock\n"); + + LIST_REMOVE(fl, nfslocklist); + + debuglog("Exiting remove_blockingfilelock\n"); +} + +void +clear_blockingfilelock(const char *hostname) +{ + struct file_lock *ifl,*nfl; + + /* + * Normally, LIST_FOREACH is called for, but since + * the current element *is* the iterator, deleting it + * would mess up the iteration. Thus, a next element + * must be used explicitly + */ + + ifl = LIST_FIRST(&blockedlocklist_head); + + while (ifl != NULL) { + nfl = LIST_NEXT(ifl, nfslocklist); + + if (strncmp(hostname, ifl->client_name, SM_MAXSTRLEN) == 0) { + remove_blockingfilelock(ifl); + deallocate_file_lock(ifl); } - if (pid == 0) { - /* no more child to handle yet */ - return; + + ifl = nfl; + } +} + +void +retry_blockingfilelocklist(void) +{ + /* Retry all locks in the blocked list */ + struct file_lock *ifl, *nfl; /* Iterator */ + enum partialfilelock_status pflstatus; + + debuglog("Entering retry_blockingfilelocklist\n"); + + ifl = LIST_FIRST(&blockedlocklist_head); + debuglog("Iterator choice %p\n",ifl); + + while (ifl != NULL) { + /* + * SUBTLE BUG: The next element must be worked out before the + * current element has been moved + */ + nfl = LIST_NEXT(ifl, nfslocklist); + debuglog("Iterator choice %p\n",ifl); + debuglog("Next iterator choice %p\n",nfl); + + /* + * SUBTLE BUG: The file_lock must be removed from the + * old list so that it's list pointers get disconnected + * before being allowed to participate in the new list + * which will automatically add it in if necessary. + */ + + LIST_REMOVE(ifl, nfslocklist); + pflstatus = lock_partialfilelock(ifl); + + if (pflstatus == PFL_GRANTED || pflstatus == PFL_GRANTED_DUPLICATE) { + debuglog("Granted blocked lock\n"); + /* lock granted and is now being used */ + send_granted(ifl,0); + } else { + /* Reinsert lock back into same place in blocked list */ + debuglog("Replacing blocked lock\n"); + LIST_INSERT_BEFORE(nfl, ifl, nfslocklist); } + + /* Valid increment behavior regardless of state of ifl */ + ifl = nfl; + } + + debuglog("Exiting retry_blockingfilelocklist\n"); +} + +/* + * Below here are routines associated with manipulating all + * aspects of the partial file locking system (list, hardware, etc.) + */ + +/* + * Please note that lock monitoring must be done at this level which + * keeps track of *individual* lock requests on lock and unlock + * + * XXX: Split unlocking is going to make the unlock code miserable + */ + +/* + * lock_partialfilelock: + * + * Argument fl gets modified as its list housekeeping entries get modified + * upon insertion into the NFS lock list + * + * This routine makes several assumptions: + * 1) It (will) pass locks through to flock to lock the entire underlying file + * and then parcel out NFS locks if it gets control of the file. + * This matches the old rpc.lockd file semantics (except where it + * is now more correct). It is the safe solution, but will cause + * overly restrictive blocking if someone is trying to use the + * underlying files without using NFS. This appears to be an + * acceptable tradeoff since most people use standalone NFS servers. + * XXX: The right solution is probably kevent combined with fcntl + * + * 2) Nothing modifies the lock lists between testing and granting + * I have no idea whether this is a useful assumption or not + */ + +enum partialfilelock_status +lock_partialfilelock(struct file_lock *fl) +{ + enum partialfilelock_status retval; + enum nfslock_status lnlstatus; + enum hwlock_status hwstatus; + + debuglog("Entering lock_partialfilelock\n"); + + retval = PFL_DENIED; + + /* + * Execute the NFS lock first, if possible, as it is significantly + * easier and less expensive to undo than the filesystem lock + */ + + lnlstatus = lock_nfslock(fl); + + switch (lnlstatus) { + case NFS_GRANTED: + case NFS_GRANTED_DUPLICATE: /* - * if we're here we have a child that exited - * Find the associated file_lock. + * At this point, the NFS lock is allocated and active. + * Remember to clean it up if the hardware lock fails */ - for (fl = LIST_FIRST(&lcklst_head); fl != NULL; - fl = LIST_NEXT(fl, lcklst)) { - if (pid == fl->locker) - break; + hwstatus = lock_hwlock(fl); + + switch (hwstatus) { + case HW_GRANTED: + case HW_GRANTED_DUPLICATE: + debuglog("HW GRANTED\n"); + /* + * XXX: Fixme: Check hwstatus for duplicate when + * true partial file locking and accounting is + * done on the hardware + */ + if (lnlstatus == NFS_GRANTED_DUPLICATE) { + retval = PFL_GRANTED_DUPLICATE; + } else { + retval = PFL_GRANTED; + } + monitor_lock_host(fl->client_name); + break; + case HW_RESERR: + debuglog("HW RESERR\n"); + retval = PFL_HWRESERR; + break; + case HW_DENIED: + debuglog("HW DENIED\n"); + retval = PFL_HWDENIED; + break; + default: + debuglog("Unmatched hwstatus %d\n",hwstatus); + break; } - if (pid != fl->locker) { - syslog(LOG_NOTICE, "unknow child %d", pid); + + if (retval != PFL_GRANTED && + retval != PFL_GRANTED_DUPLICATE) { + /* Clean up the NFS lock */ + debuglog("Deleting trial NFS lock\n"); + delete_nfslock(fl); + } + break; + case NFS_DENIED: + retval = PFL_NFSDENIED; + break; + case NFS_RESERR: + retval = PFL_NFSRESERR; + default: + debuglog("Unmatched lnlstatus %d\n"); + retval = PFL_NFSDENIED_NOLOCK; + break; + } + + /* + * By the time fl reaches here, it is completely free again on + * failure. The NFS lock done before attempting the + * hardware lock has been backed out + */ + + if (retval == PFL_NFSDENIED || retval == PFL_HWDENIED) { + /* Once last chance to check the lock */ + if (fl->blocking == 1) { + /* Queue the lock */ + debuglog("BLOCKING LOCK RECEIVED\n"); + retval = (retval == PFL_NFSDENIED ? + PFL_NFSBLOCKED : PFL_HWBLOCKED); + add_blockingfilelock(fl); + dump_filelock(fl); } else { - if (!WIFEXITED(status) || WEXITSTATUS(status) != 0) { - syslog(LOG_NOTICE, "child %d failed", pid); - /* - * can't do much here; we can't reply - * anything but OK for blocked locks - * Eventually the client will time out - * and retry. - */ - do_unlock(fl); - return; + /* Leave retval alone, it's already correct */ + debuglog("Lock denied. Non-blocking failure\n"); + dump_filelock(fl); + } + } + + debuglog("Exiting lock_partialfilelock\n"); + + return retval; +} + +/* + * unlock_partialfilelock: + * + * Given a file_lock, unlock all locks which match. + * + * Note that a given lock might have to unlock ITSELF! See + * clear_partialfilelock for example. + */ + +enum partialfilelock_status +unlock_partialfilelock(const struct file_lock *fl) +{ + struct file_lock *lfl,*rfl,*releasedfl,*selffl; + enum partialfilelock_status retval; + enum nfslock_status unlstatus; + enum hwlock_status unlhwstatus, lhwstatus; + + debuglog("Entering unlock_partialfilelock\n"); + + selffl = NULL; + lfl = NULL; + rfl = NULL; + releasedfl = NULL; + retval = PFL_DENIED; + + /* + * There are significant overlap and atomicity issues + * with partially releasing a lock. For example, releasing + * part of an NFS shared lock does *not* always release the + * corresponding part of the file since there is only one + * rpc.lockd UID but multiple users could be requesting it + * from NFS. Also, an unlock request should never allow + * another process to gain a lock on the remaining parts. + * ie. Always apply the new locks before releasing the + * old one + */ + + /* + * Loop is required since multiple little locks + * can be allocated and then deallocated with one + * big unlock. + * + * The loop is required to be here so that the nfs & + * hw subsystems do not need to communicate with one + * one another + */ + + do { + debuglog("Value of releasedfl: %p\n",releasedfl); + /* lfl&rfl are created *AND* placed into the NFS lock list if required */ + unlstatus = unlock_nfslock(fl, &releasedfl, &lfl, &rfl); + debuglog("Value of releasedfl: %p\n",releasedfl); + + + /* XXX: This is grungy. It should be refactored to be cleaner */ + if (lfl != NULL) { + lhwstatus = lock_hwlock(lfl); + if (lhwstatus != HW_GRANTED && + lhwstatus != HW_GRANTED_DUPLICATE) { + debuglog("HW duplicate lock failure for left split\n"); + } + monitor_lock_host(lfl->client_name); + } + + if (rfl != NULL) { + lhwstatus = lock_hwlock(rfl); + if (lhwstatus != HW_GRANTED && + lhwstatus != HW_GRANTED_DUPLICATE) { + debuglog("HW duplicate lock failure for right split\n"); } - - /* check lock status */ - syslog(LOG_DEBUG, "processing child %d, status %d", - pid, fl->status); - switch(fl->status) { - case LKST_PROCESSING: - fl->status = LKST_LOCKED; - send_granted(fl, (fl->flags & LOCK_V4) ? - nlm4_granted : nlm_granted); + monitor_lock_host(rfl->client_name); + } + + switch (unlstatus) { + case NFS_GRANTED: + /* Attempt to unlock on the hardware */ + debuglog("NFS unlock granted. Attempting hardware unlock\n"); + + /* This call *MUST NOT* unlock the two newly allocated locks */ + unlhwstatus = unlock_hwlock(fl); + debuglog("HW unlock returned with code %d\n",unlhwstatus); + + switch (unlhwstatus) { + case HW_GRANTED: + debuglog("HW unlock granted\n"); + unmonitor_lock_host(releasedfl->client_name); + retval = PFL_GRANTED; break; - case LKST_DYING: - do_unlock(fl); + case HW_DENIED_NOLOCK: + /* Huh?!?! This shouldn't happen */ + debuglog("HW unlock denied no lock\n"); + retval = PFL_HWRESERR; + /* Break out of do-while */ + unlstatus = NFS_RESERR; break; default: - syslog(LOG_NOTICE, "bad lock status (%d) for" - " child %d", fl->status, pid); + debuglog("HW unlock failed\n"); + retval = PFL_HWRESERR; + /* Break out of do-while */ + unlstatus = NFS_RESERR; + break; } + + debuglog("Exiting with status retval: %d\n",retval); + + retry_blockingfilelocklist(); + break; + case NFS_DENIED_NOLOCK: + retval = PFL_GRANTED; + debuglog("All locks cleaned out\n"); + break; + default: + retval = PFL_NFSRESERR; + debuglog("NFS unlock failure\n"); + dump_filelock(fl); + break; } + + if (releasedfl != NULL) { + if (fl == releasedfl) { + /* + * XXX: YECHHH!!! Attempt to unlock self succeeded + * but we can't deallocate the space yet. This is what + * happens when you don't write malloc and free together + */ + debuglog("Attempt to unlock self\n"); + selffl = releasedfl; + } else { + /* + * XXX: this deallocation *still* needs to migrate closer + * to the allocation code way up in get_lock or the allocation + * code needs to migrate down (violation of "When you write + * malloc you must write free") + */ + + deallocate_file_lock(releasedfl); + } + } + + } while (unlstatus == NFS_GRANTED); + + if (selffl != NULL) { + /* + * This statement wipes out the incoming file lock (fl) + * in spite of the fact that it is declared const + */ + debuglog("WARNING! Destroying incoming lock pointer\n"); + deallocate_file_lock(selffl); } + + debuglog("Exiting unlock_partialfilelock\n"); + + return retval; } /* + * clear_partialfilelock * - * try to aquire the lock described by fl. Eventually fock a child to do a - * blocking lock if allowed and required. + * Normally called in response to statd state number change. + * Wipe out all locks held by a host. As a bonus, the act of + * doing so should automatically clear their statd entries and + * unmonitor the host. + */ + +void +clear_partialfilelock(const char *hostname) +{ + struct file_lock *ifl, *nfl; + + /* Clear blocking file lock list */ + clear_blockingfilelock(hostname); + + /* do all required unlocks */ + /* Note that unlock can smash the current pointer to a lock */ + + /* + * Normally, LIST_FOREACH is called for, but since + * the current element *is* the iterator, deleting it + * would mess up the iteration. Thus, a next element + * must be used explicitly + */ + + ifl = LIST_FIRST(&nfslocklist_head); + + while (ifl != NULL) { + nfl = LIST_NEXT(ifl, nfslocklist); + + if (strncmp(hostname, ifl->client_name, SM_MAXSTRLEN) == 0) { + /* Unlock destroys ifl out from underneath */ + unlock_partialfilelock(ifl); + /* ifl is NO LONGER VALID AT THIS POINT */ + } + ifl = nfl; + } +} + +/* + * test_partialfilelock: + */ +enum partialfilelock_status +test_partialfilelock(const struct file_lock *fl, + struct file_lock **conflicting_fl) +{ + enum partialfilelock_status retval; + enum nfslock_status teststatus; + + debuglog("Entering testpartialfilelock...\n"); + + retval = PFL_DENIED; + + teststatus = test_nfslock(fl, conflicting_fl); + debuglog("test_partialfilelock: teststatus %d\n",teststatus); + + if (teststatus == NFS_GRANTED || teststatus == NFS_GRANTED_DUPLICATE) { + /* XXX: Add the underlying filesystem locking code */ + retval = (teststatus == NFS_GRANTED) ? + PFL_GRANTED : PFL_GRANTED_DUPLICATE; + debuglog("Dumping locks...\n"); + dump_filelock(fl); + dump_filelock(*conflicting_fl); + debuglog("Done dumping locks...\n"); + } else { + retval = PFL_NFSDENIED; + debuglog("NFS test denied.\n"); + dump_filelock(fl); + debuglog("Conflicting.\n"); + dump_filelock(*conflicting_fl); + } + + debuglog("Exiting testpartialfilelock...\n"); + + return retval; +} + +/* + * Below here are routines associated with translating the partial file locking + * codes into useful codes to send back to the NFS RPC messaging system + */ + +/* + * These routines translate the (relatively) useful return codes back onto + * the few return codes which the nlm subsystems wishes to trasmit */ enum nlm_stats -do_lock(fl, block) - struct file_lock *fl; - int block; +do_test(struct file_lock *fl, struct file_lock **conflicting_fl) { - int lflags, error; - struct stat st; + enum partialfilelock_status pfsret; + enum nlm_stats retval; - fl->fd = fhopen(&fl->filehandle, O_RDWR); - if (fl->fd < 0) { - switch (errno) { - case ESTALE: - error = nlm4_stale_fh; + debuglog("Entering do_test...\n"); + + pfsret = test_partialfilelock(fl,conflicting_fl); + + switch (pfsret) { + case PFL_GRANTED: + debuglog("PFL test lock granted\n"); + dump_filelock(fl); + dump_filelock(*conflicting_fl); + retval = (fl->flags & LOCK_V4) ? nlm4_granted : nlm_granted; + break; + case PFL_GRANTED_DUPLICATE: + debuglog("PFL test lock granted--duplicate id detected\n"); + dump_filelock(fl); + dump_filelock(*conflicting_fl); + debuglog("Clearing conflicting_fl for call semantics\n"); + *conflicting_fl = NULL; + retval = (fl->flags & LOCK_V4) ? nlm4_granted : nlm_granted; + break; + case PFL_NFSDENIED: + case PFL_HWDENIED: + debuglog("PFL test lock denied\n"); + dump_filelock(fl); + dump_filelock(*conflicting_fl); + retval = (fl->flags & LOCK_V4) ? nlm4_denied : nlm_denied; + break; + case PFL_NFSRESERR: + case PFL_HWRESERR: + debuglog("PFL test lock resource fail\n"); + dump_filelock(fl); + dump_filelock(*conflicting_fl); + retval = (fl->flags & LOCK_V4) ? nlm4_denied_nolocks : nlm_denied_nolocks; + break; + default: + debuglog("PFL test lock *FAILED*\n"); + dump_filelock(fl); + dump_filelock(*conflicting_fl); + retval = (fl->flags & LOCK_V4) ? nlm4_failed : nlm_denied; + break; + } + + debuglog("Exiting do_test...\n"); + + return retval; +} + +/* + * do_lock: Try to acquire a lock + * + * This routine makes a distinction between NLM versions. I am pretty + * convinced that this should be abstracted out and bounced up a level + */ + +enum nlm_stats +do_lock(struct file_lock *fl) +{ + enum partialfilelock_status pfsret; + enum nlm_stats retval; + + debuglog("Entering do_lock...\n"); + + pfsret = lock_partialfilelock(fl); + + switch (pfsret) { + case PFL_GRANTED: + debuglog("PFL lock granted"); + dump_filelock(fl); + retval = (fl->flags & LOCK_V4) ? nlm4_granted : nlm_granted; + break; + case PFL_GRANTED_DUPLICATE: + debuglog("PFL lock granted--duplicate id detected"); + dump_filelock(fl); + retval = (fl->flags & LOCK_V4) ? nlm4_granted : nlm_granted; + break; + case PFL_NFSDENIED: + case PFL_HWDENIED: + debuglog("PFL_NFS lock denied"); + dump_filelock(fl); + retval = (fl->flags & LOCK_V4) ? nlm4_denied : nlm_denied; + break; + case PFL_NFSBLOCKED: + case PFL_HWBLOCKED: + debuglog("PFL_NFS blocking lock denied. Queued.\n"); + dump_filelock(fl); + retval = (fl->flags & LOCK_V4) ? nlm4_blocked : nlm_blocked; + break; + case PFL_NFSRESERR: + case PFL_HWRESERR: + debuglog("PFL lock resource alocation fail\n"); + dump_filelock(fl); + retval = (fl->flags & LOCK_V4) ? nlm4_denied_nolocks : nlm_denied_nolocks; + break; + default: + debuglog("PFL lock *FAILED*"); + dump_filelock(fl); + retval = (fl->flags & LOCK_V4) ? nlm4_failed : nlm_denied; + break; + } + + debuglog("Exiting do_lock...\n"); + + return retval; +} + +enum nlm_stats +do_unlock(struct file_lock *fl) +{ + enum partialfilelock_status pfsret; + enum nlm_stats retval; + + debuglog("Entering do_unlock...\n"); + pfsret = unlock_partialfilelock(fl); + + switch (pfsret) { + case PFL_GRANTED: + debuglog("PFL unlock granted"); + dump_filelock(fl); + retval = (fl->flags & LOCK_V4) ? nlm4_granted : nlm_granted; + break; + case PFL_NFSDENIED: + case PFL_HWDENIED: + debuglog("PFL_NFS unlock denied"); + dump_filelock(fl); + retval = (fl->flags & LOCK_V4) ? nlm4_denied : nlm_denied; + break; + case PFL_NFSDENIED_NOLOCK: + case PFL_HWDENIED_NOLOCK: + debuglog("PFL_NFS no lock found\n"); + retval = (fl->flags & LOCK_V4) ? nlm4_granted : nlm_granted; + break; + case PFL_NFSRESERR: + case PFL_HWRESERR: + debuglog("PFL unlock resource failure"); + dump_filelock(fl); + retval = (fl->flags & LOCK_V4) ? nlm4_denied_nolocks : nlm_denied_nolocks; + break; + default: + debuglog("PFL unlock *FAILED*"); + dump_filelock(fl); + retval = (fl->flags & LOCK_V4) ? nlm4_failed : nlm_denied; + break; + } + + debuglog("Exiting do_unlock...\n"); + + return retval; +} + +/* + * do_clear + * + * This routine is non-existent because it doesn't have a return code. + * It is here for completeness in case someone *does* need to do return + * codes later. A decent compiler should optimize this away. + */ + +void +do_clear(const char *hostname) +{ + + clear_partialfilelock(hostname); +} + +/* + * The following routines are all called from the code which the + * RPC layer invokes + */ + +/* + * testlock(): inform the caller if the requested lock would be granted + * + * returns NULL if lock would granted + * returns pointer to a conflicting nlm4_holder if not + */ + +struct nlm4_holder * +testlock(struct nlm4_lock *lock, bool_t exclusive, int flags) +{ + struct file_lock test_fl, *conflicting_fl; + + bzero(&test_fl, sizeof(test_fl)); + + bcopy(lock->fh.n_bytes, &(test_fl.filehandle), sizeof(fhandle_t)); + copy_nlm4_lock_to_nlm4_holder(lock, exclusive, &test_fl.client); + + siglock(); + do_test(&test_fl, &conflicting_fl); + + if (conflicting_fl == NULL) { + debuglog("No conflicting lock found\n"); + sigunlock(); + return NULL; + } else { + debuglog("Found conflicting lock\n"); + dump_filelock(conflicting_fl); + sigunlock(); + return (&conflicting_fl->client); + } +} + +/* + * getlock: try to aquire the lock. + * If file is already locked and we can sleep, put the lock in the list with + * status LKST_WAITING; it'll be processed later. + * Otherwise try to lock. If we're allowed to block, fork a child which + * will do the blocking lock. + */ + +enum nlm_stats +getlock(nlm4_lockargs *lckarg, struct svc_req *rqstp, const int flags) +{ + struct file_lock *newfl; + enum nlm_stats retval; + + debuglog("Entering getlock...\n"); + + if (grace_expired == 0 && lckarg->reclaim == 0) + return (flags & LOCK_V4) ? + nlm4_denied_grace_period : nlm_denied_grace_period; + + /* allocate new file_lock for this request */ + newfl = allocate_file_lock(&lckarg->alock.oh, &lckarg->cookie); + if (newfl == NULL) { + syslog(LOG_NOTICE, "lock allocate failed: %s", strerror(errno)); + /* failed */ + return (flags & LOCK_V4) ? + nlm4_denied_nolocks : nlm_denied_nolocks; + } + + if (lckarg->alock.fh.n_len != sizeof(fhandle_t)) { + debuglog("recieved fhandle size %d, local size %d", + lckarg->alock.fh.n_len, (int)sizeof(fhandle_t)); + } + + fill_file_lock(newfl, (fhandle_t *)lckarg->alock.fh.n_bytes, + (struct sockaddr *)svc_getrpccaller(rqstp->rq_xprt)->buf, + lckarg->exclusive, lckarg->alock.svid, lckarg->alock.l_offset, lckarg->alock.l_len, + lckarg->alock.caller_name, lckarg->state, 0, flags, lckarg->block); + + /* + * newfl is now fully constructed and deallocate_file_lock + * can now be used to delete it + */ + + siglock(); + debuglog("Pointer to new lock is %p\n",newfl); + + retval = do_lock(newfl); + + debuglog("Pointer to new lock is %p\n",newfl); + sigunlock(); + + switch (retval) + { + case nlm4_granted: + /* case nlm_granted: is the same as nlm4_granted */ + /* do_mon(lckarg->alock.caller_name); */ break; - case EROFS: - error = nlm4_rofs; + case nlm4_blocked: + /* case nlm_blocked: is the same as nlm4_blocked */ + /* do_mon(lckarg->alock.caller_name); */ break; default: - error = nlm4_failed; + deallocate_file_lock(newfl); + break; + } + + debuglog("Exiting getlock...\n"); + + return retval; +} + + +/* unlock a filehandle */ +enum nlm_stats +unlock(nlm4_lock *lock, const int flags) +{ + struct file_lock fl; + enum nlm_stats err; + + siglock(); + + debuglog("Entering unlock...\n"); + + bzero(&fl,sizeof(struct file_lock)); + bcopy(lock->fh.n_bytes, &fl.filehandle, sizeof(fhandle_t)); + + copy_nlm4_lock_to_nlm4_holder(lock, 0, &fl.client); + + err = do_unlock(&fl); + + sigunlock(); + + debuglog("Exiting unlock...\n"); + + return err; +} + +/* + * XXX: The following monitor/unmonitor routines + * have not been extensively tested (ie. no regression + * script exists like for the locking sections + */ + +/* + * monitor_lock_host: monitor lock hosts locally with a ref count and + * inform statd + */ +void +monitor_lock_host(const char *hostname) +{ + struct host *ihp, *nhp; + struct mon smon; + struct sm_stat_res sres; + int rpcret, statflag; + + rpcret = 0; + statflag = 0; + + LIST_FOREACH(ihp, &hostlst_head, hostlst) { + if (strncmp(hostname, ihp->name, SM_MAXSTRLEN) == 0) { + /* Host is already monitored, bump refcount */ + ++ihp->refcnt; + /* Host should only be in the monitor list once */ + return; } - if ((fl->flags & LOCK_V4) == 0) - error = nlm_denied; - syslog(LOG_NOTICE, "fhopen failed (from %s): %s", - fl->client_name, strerror(errno)); - LIST_REMOVE(fl, lcklst); - return error;; } - if (fstat(fl->fd, &st) < 0) { - syslog(LOG_NOTICE, "fstat failed (from %s): %s", - fl->client_name, strerror(errno)); + + /* Host is not yet monitored, add it */ + nhp = malloc(sizeof(struct host)); + + if (nhp == NULL) { + debuglog("Unable to allocate entry for statd mon\n"); + return; } - syslog(LOG_DEBUG, "lock from %s for file%s%s: dev %d ino %d (uid %d), " - "flags %d", - fl->client_name, fl->client.exclusive ? " (exclusive)":"", - block ? " (block)":"", - st.st_dev, st.st_ino, st.st_uid, fl->flags); - lflags = LOCK_NB; - if (fl->client.exclusive == 0) - lflags |= LOCK_SH; - else - lflags |= LOCK_EX; - error = flock(fl->fd, lflags); - if (error != 0 && errno == EAGAIN && block) { - switch (fl->locker = fork()) { - case -1: /* fork failed */ - syslog(LOG_NOTICE, "fork failed: %s", strerror(errno)); - LIST_REMOVE(fl, lcklst); - close(fl->fd); - return (fl->flags & LOCK_V4) ? - nlm4_denied_nolocks : nlm_denied_nolocks; - case 0: - /* - * Attempt a blocking lock. Will have to call - * NLM_GRANTED later. - */ - setproctitle("%s", fl->client_name); - lflags &= ~LOCK_NB; - if(flock(fl->fd, lflags) != 0) { - syslog(LOG_NOTICE, "flock failed: %s", - strerror(errno)); - exit(-1); - } - /* lock granted */ - exit(0); - default: - syslog(LOG_DEBUG, "lock request from %s: forked %d", - fl->client_name, fl->locker); - fl->status = LKST_PROCESSING; - return (fl->flags & LOCK_V4) ? - nlm4_blocked : nlm_blocked; + + /* Allocated new host entry, now fill the fields */ + strncpy(nhp->name, hostname, SM_MAXSTRLEN); + nhp->refcnt = 1; + debuglog("Locally Monitoring host %16s\n",hostname); + + debuglog("Attempting to tell statd\n"); + + bzero(&smon,sizeof(smon)); + + smon.mon_id.mon_name = nhp->name; + smon.mon_id.my_id.my_name = "localhost\0"; + + smon.mon_id.my_id.my_prog = NLM_PROG; + smon.mon_id.my_id.my_vers = NLM_SM; + smon.mon_id.my_id.my_proc = NLM_SM_NOTIFY; + + rpcret = callrpc("localhost", SM_PROG, SM_VERS, SM_MON, xdr_mon, + &smon, xdr_sm_stat_res, &sres); + + if (rpcret == 0) { + if (sres.res_stat == stat_fail) { + debuglog("Statd call failed\n"); + statflag = 0; + } else { + statflag = 1; } + } else { + debuglog("Rpc call to statd failed with return value: %d\n", + rpcret); + statflag = 0; } - /* non block case */ - if (error != 0) { - switch (errno) { - case EAGAIN: - error = nlm4_denied; - break; - case ESTALE: - error = nlm4_stale_fh; - break; - case EROFS: - error = nlm4_rofs; + + if (statflag == 1) { + LIST_INSERT_HEAD(&hostlst_head, nhp, hostlst); + } else { + free(nhp); + } + +} + +/* + * unmonitor_lock_host: clear monitor ref counts and inform statd when gone + */ +void +unmonitor_lock_host(const char *hostname) +{ + struct host *ihp; + struct mon_id smon_id; + struct sm_stat smstat; + int rpcret; + + rpcret = 0; + + for( ihp=LIST_FIRST(&hostlst_head); ihp != NULL; + ihp=LIST_NEXT(ihp, hostlst)) { + if (strncmp(hostname, ihp->name, SM_MAXSTRLEN) == 0) { + /* Host is monitored, bump refcount */ + --ihp->refcnt; + /* Host should only be in the monitor list once */ break; - default: - error = nlm4_failed; } - if ((fl->flags & LOCK_V4) == 0) - error = nlm_denied; - if (errno != EAGAIN) - syslog(LOG_NOTICE, "flock for %s failed: %s", - fl->client_name, strerror(errno)); - else syslog(LOG_DEBUG, "flock for %s failed: %s", - fl->client_name, strerror(errno)); - LIST_REMOVE(fl, lcklst); - close(fl->fd); - return error; } - fl->status = LKST_LOCKED; - return (fl->flags & LOCK_V4) ? nlm4_granted : nlm_granted; + + if (ihp == NULL) { + debuglog("Could not find host %16s in mon list\n", hostname); + return; + } + + if (ihp->refcnt > 0) + return; + + if (ihp->refcnt < 0) { + debuglog("Negative refcount!: %d\n", + ihp->refcnt); + } + + debuglog("Attempting to unmonitor host %16s\n", hostname); + + bzero(&smon_id,sizeof(smon_id)); + + smon_id.mon_name = (char *)hostname; + smon_id.my_id.my_name = "localhost"; + smon_id.my_id.my_prog = NLM_PROG; + smon_id.my_id.my_vers = NLM_SM; + smon_id.my_id.my_proc = NLM_SM_NOTIFY; + + rpcret = callrpc("localhost", SM_PROG, SM_VERS, SM_UNMON, xdr_mon, + &smon_id, xdr_sm_stat_res, &smstat); + + if (rpcret != 0) { + debuglog("Rpc call to unmonitor statd failed with " + " return value: %d\n", rpcret); + } + + LIST_REMOVE(ihp, hostlst); + free(ihp); +} + +/* + * notify: Clear all locks from a host if statd complains + * + * XXX: This routine has not been thoroughly tested. However, neither + * had the old one been. It used to compare the statd crash state counter + * to the current lock state. The upshot of this was that it basically + * cleared all locks from the specified host 99% of the time (with the + * other 1% being a bug). Consequently, the assumption is that clearing + * all locks from a host when notified by statd is acceptable. + * + * Please note that this routine skips the usual level of redirection + * through a do_* type routine. This introduces a possible level of + * error and might better be written as do_notify and take this one out. + + */ + +void +notify(const char *hostname, const int state) +{ + debuglog("notify from %s, new state %d", hostname, state); + + siglock(); + do_clear(hostname); + sigunlock(); + + debuglog("Leaving notify\n"); } void @@ -587,6 +2118,10 @@ send_granted(fl, opcode) static struct nlm_res retval; static struct nlm4_res retval4; + debuglog("About to send granted on blocked lock\n"); + sleep(1); + debuglog("Blowing off return send\n"); + cli = get_client(fl->addr, (fl->flags & LOCK_V4) ? NLM_VERS4 : NLM_VERS); if (cli == NULL) { @@ -613,8 +2148,8 @@ send_granted(fl, opcode) res.alock.svid = fl->client.svid; res.alock.l_offset = fl->client.l_offset; res.alock.l_len = fl->client.l_len; - syslog(LOG_DEBUG, "sending v4 reply%s", - (fl->flags & LOCK_ASYNC) ? " (async)":""); + debuglog("sending v4 reply%s", + (fl->flags & LOCK_ASYNC) ? " (async)":""); if (fl->flags & LOCK_ASYNC) { success = clnt_call(cli, NLM4_GRANTED_MSG, xdr_nlm4_testargs, &res, xdr_void, &dummy, timeo); @@ -635,8 +2170,8 @@ send_granted(fl, opcode) res.alock.svid = fl->client.svid; res.alock.l_offset = fl->client.l_offset; res.alock.l_len = fl->client.l_len; - syslog(LOG_DEBUG, "sending v1 reply%s", - (fl->flags & LOCK_ASYNC) ? " (async)":""); + debuglog("sending v1 reply%s", + (fl->flags & LOCK_ASYNC) ? " (async)":""); if (fl->flags & LOCK_ASYNC) { success = clnt_call(cli, NLM_GRANTED_MSG, xdr_nlm_testargs, &res, xdr_void, &dummy, timeo); @@ -647,173 +2182,46 @@ send_granted(fl, opcode) } } if (debug_level > 2) - syslog(LOG_DEBUG, "clnt_call returns %d(%s) for granted", - success, clnt_sperrno(success)); + debuglog("clnt_call returns %d(%s) for granted", + success, clnt_sperrno(success)); } -enum nlm_stats -do_unlock(rfl) - struct file_lock *rfl; -{ - struct file_lock *fl; - int error; - int lockst; - - /* unlock the file: closing is enouth ! */ - if (close(rfl->fd) < 0) { - if (errno == ESTALE) - error = nlm4_stale_fh; - else - error = nlm4_failed; - if ((fl->flags & LOCK_V4) == 0) - error = nlm_denied; - syslog(LOG_NOTICE, - "close failed (from %s): %s", - rfl->client_name, strerror(errno)); - } else { - error = (fl->flags & LOCK_V4) ? - nlm4_granted : nlm_granted; - } - LIST_REMOVE(rfl, lcklst); - - /* process the next LKST_WAITING lock request for this fh */ - for (fl = LIST_FIRST(&lcklst_head); fl != NULL; - fl = LIST_NEXT(fl, lcklst)) { - if (fl->status != LKST_WAITING || - memcmp(&rfl->filehandle, &fl->filehandle, - sizeof(fhandle_t)) != 0) - continue; +/* + * Routines below here have not been modified in the overhaul + */ - lockst = do_lock(fl, 1); /* If it's LKST_WAITING we can block */ - switch (lockst) { - case nlm4_granted: - /* case nlm_granted: same as nlm4_granted */ - send_granted(fl, (fl->flags & LOCK_V4) ? - nlm4_granted : nlm_granted); - break; - case nlm4_blocked: - /* case nlm_blocked: same as nlm4_blocked */ - break; - default: - lfree(fl); - break; - } - break; - } - return error; -} +/* + * Are these two routines still required since lockd is not spawning off + * children to service locks anymore? Presumably they were originally + * put in place to prevent a one child from changing the lock list out + * from under another one. + */ void -siglock() +siglock(void) { - sigset_t block; + sigset_t block; - sigemptyset(&block); - sigaddset(&block, SIGCHLD); + sigemptyset(&block); + sigaddset(&block, SIGCHLD); - if (sigprocmask(SIG_BLOCK, &block, NULL) < 0) { - syslog(LOG_WARNING, "siglock failed: %s", strerror(errno)); - } + if (sigprocmask(SIG_BLOCK, &block, NULL) < 0) { + syslog(LOG_WARNING, "siglock failed: %s", strerror(errno)); + } } void -sigunlock() +sigunlock(void) { - sigset_t block; + sigset_t block; - sigemptyset(&block); - sigaddset(&block, SIGCHLD); - - if (sigprocmask(SIG_UNBLOCK, &block, NULL) < 0) { - syslog(LOG_WARNING, "sigunlock failed: %s", strerror(errno)); - } + sigemptyset(&block); + sigaddset(&block, SIGCHLD); + + if (sigprocmask(SIG_UNBLOCK, &block, NULL) < 0) { + syslog(LOG_WARNING, "sigunlock failed: %s", strerror(errno)); + } } -/* monitor a host through rpc.statd, and keep a ref count */ -void -do_mon(hostname) - char *hostname; -{ - struct host *hp; - struct mon my_mon; - struct sm_stat_res res; - int retval; - - for (hp = LIST_FIRST(&hostlst_head); hp != NULL; - hp = LIST_NEXT(hp, hostlst)) { - if (strcmp(hostname, hp->name) == 0) { - /* already monitored, just bump refcnt */ - hp->refcnt++; - return; - } - } - /* not found, have to create an entry for it */ - hp = malloc(sizeof(struct host)); - strncpy(hp->name, hostname, SM_MAXSTRLEN); - hp->refcnt = 1; - syslog(LOG_DEBUG, "monitoring host %s", - hostname); - memset(&my_mon, 0, sizeof(my_mon)); - my_mon.mon_id.mon_name = hp->name; - my_mon.mon_id.my_id.my_name = "localhost"; - my_mon.mon_id.my_id.my_prog = NLM_PROG; - my_mon.mon_id.my_id.my_vers = NLM_SM; - my_mon.mon_id.my_id.my_proc = NLM_SM_NOTIFY; - if ((retval = - callrpc("localhost", SM_PROG, SM_VERS, SM_MON, xdr_mon, - (char*)&my_mon, xdr_sm_stat_res, (char*)&res)) != 0) { - syslog(LOG_WARNING, "rpc to statd failed: %s", - clnt_sperrno((enum clnt_stat)retval)); - free(hp); - return; - } - if (res.res_stat == stat_fail) { - syslog(LOG_WARNING, "statd failed"); - free(hp); - return; - } - LIST_INSERT_HEAD(&hostlst_head, hp, hostlst); -} -void -notify(hostname, state) - const char *hostname; - int state; -{ - struct file_lock *fl, *next_fl; - int err; - syslog(LOG_DEBUG, "notify from %s, new state %d", hostname, state); - /* search all lock for this host; if status changed, release the lock */ - siglock(); - for (fl = LIST_FIRST(&lcklst_head); fl != NULL; fl = next_fl) { - next_fl = LIST_NEXT(fl, lcklst); - if (strcmp(hostname, fl->client_name) == 0 && - fl->nsm_status != state) { - syslog(LOG_DEBUG, "state %d, nsm_state %d, unlocking", - fl->status, fl->nsm_status); - switch(fl->status) { - case LKST_LOCKED: - err = do_unlock(fl); - if (err != nlm_granted) - syslog(LOG_DEBUG, - "notify: unlock failed for %s (%d)", - hostname, err); - break; - case LKST_WAITING: - LIST_REMOVE(fl, lcklst); - lfree(fl); - break; - case LKST_PROCESSING: - fl->status = LKST_DYING; - break; - case LKST_DYING: - break; - default: - syslog(LOG_NOTICE, "unknow status %d for %s", - fl->status, fl->client_name); - } - } - } - sigunlock(); -} |