]>
git.saurik.com Git - apple/xnu.git/blob - bsd/ufs/ufs/ufs_lockf.c
ec0a50957d1fe20723920b04f5fe0b7ae589cd21
2 * Copyright (c) 2000 Apple Computer, Inc. All rights reserved.
4 * @APPLE_LICENSE_HEADER_START@
6 * The contents of this file constitute Original Code as defined in and
7 * are subject to the Apple Public Source License Version 1.1 (the
8 * "License"). You may not use this file except in compliance with the
9 * License. Please obtain a copy of the License at
10 * http://www.apple.com/publicsource and read it before using this file.
12 * This Original Code and all software distributed under the License are
13 * distributed on an "AS IS" basis, WITHOUT WARRANTY OF ANY KIND, EITHER
14 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
15 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
16 * FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT. Please see the
17 * License for the specific language governing rights and limitations
20 * @APPLE_LICENSE_HEADER_END@
22 /* Copyright (c) 1995 NeXT Computer, Inc. All Rights Reserved */
24 * Copyright (c) 1982, 1986, 1989, 1993
25 * The Regents of the University of California. All rights reserved.
27 * This code is derived from software contributed to Berkeley by
28 * Scooter Morris at Genentech Inc.
30 * Redistribution and use in source and binary forms, with or without
31 * modification, are permitted provided that the following conditions
33 * 1. Redistributions of source code must retain the above copyright
34 * notice, this list of conditions and the following disclaimer.
35 * 2. Redistributions in binary form must reproduce the above copyright
36 * notice, this list of conditions and the following disclaimer in the
37 * documentation and/or other materials provided with the distribution.
38 * 3. All advertising materials mentioning features or use of this software
39 * must display the following acknowledgement:
40 * This product includes software developed by the University of
41 * California, Berkeley and its contributors.
42 * 4. Neither the name of the University nor the names of its contributors
43 * may be used to endorse or promote products derived from this software
44 * without specific prior written permission.
46 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
47 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
48 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
49 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
50 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
51 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
52 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
53 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
54 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
55 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
58 * @(#)ufs_lockf.c 8.4 (Berkeley) 10/26/94
61 #include <sys/param.h>
62 #include <sys/systm.h>
63 #include <sys/kernel.h>
66 #include <sys/vnode.h>
67 #include <sys/malloc.h>
68 #include <sys/fcntl.h>
70 #include <ufs/ufs/lockf.h>
71 #include <ufs/ufs/quota.h>
72 #include <ufs/ufs/inode.h>
73 #include <ufs/ufs/ufs_extern.h>
76 * This variable controls the maximum number of processes that will
77 * be checked in doing deadlock detection.
79 int maxlockdepth
= MAXDEPTH
;
83 #include <sys/sysctl.h>
85 struct ctldebug debug4
= { "lockf_debug", &lockf_debug
};
88 #define NOLOCKF (struct lockf *)0
93 * Set a byte-range lock.
97 register struct lockf
*lock
;
99 register struct lockf
*block
;
100 struct inode
*ip
= lock
->lf_inode
;
101 struct lockf
**prev
, *overlap
, *ltmp
;
102 static char lockstr
[] = "lockf";
103 int ovcase
, priority
, needtolink
, error
;
107 lf_print("lf_setlock", lock
);
108 #endif /* LOCKF_DEBUG */
114 if (lock
->lf_type
== F_WRLCK
)
118 * Scan lock list for this file looking for locks that would block us.
120 while (block
= lf_getblock(lock
)) {
122 * Free the structure and return if nonblocking.
124 if ((lock
->lf_flags
& F_WAIT
) == 0) {
129 * We are blocked. Since flock style locks cover
130 * the whole file, there is no chance for deadlock.
131 * For byte-range locks we must check for deadlock.
133 * Deadlock detection is done by looking through the
134 * wait channels to see if there are any cycles that
135 * involve us. MAXDEPTH is set just to make sure we
136 * do not go off into neverland.
138 if ((lock
->lf_flags
& F_POSIX
) &&
139 (block
->lf_flags
& F_POSIX
)) {
140 register struct proc
*wproc
;
141 register struct lockf
*waitblock
;
144 /* The block is waiting on something */
145 wproc
= (struct proc
*)block
->lf_id
;
146 while (wproc
->p_wchan
&&
147 (wproc
->p_wmesg
== lockstr
) &&
148 (i
++ < maxlockdepth
)) {
149 waitblock
= (struct lockf
*)wproc
->p_wchan
;
150 /* Get the owner of the blocking lock */
151 waitblock
= waitblock
->lf_next
;
152 if ((waitblock
->lf_flags
& F_POSIX
) == 0)
154 wproc
= (struct proc
*)waitblock
->lf_id
;
155 if (wproc
== (struct proc
*)lock
->lf_id
) {
156 _FREE(lock
, M_LOCKF
);
162 * For flock type locks, we must first remove
163 * any shared locks that we hold before we sleep
164 * waiting for an exclusive lock.
166 if ((lock
->lf_flags
& F_FLOCK
) &&
167 lock
->lf_type
== F_WRLCK
) {
168 lock
->lf_type
= F_UNLCK
;
169 (void) lf_clearlock(lock
);
170 lock
->lf_type
= F_WRLCK
;
173 * Add our lock to the blocked list and sleep until we're free.
174 * Remember who blocked us (for deadlock detection).
176 lock
->lf_next
= block
;
177 TAILQ_INSERT_TAIL(&block
->lf_blkhd
, lock
, lf_block
);
179 if (lockf_debug
& 1) {
180 lf_print("lf_setlock: blocking on", block
);
181 lf_printlist("lf_setlock", block
);
183 #endif /* LOCKF_DEBUG */
184 if (error
= tsleep((caddr_t
)lock
, priority
, lockstr
, 0)) {
186 * We may have been awakened by a signal (in
187 * which case we must remove ourselves from the
188 * blocked list) and/or by another process
189 * releasing a lock (in which case we have already
190 * been removed from the blocked list and our
191 * lf_next field set to NOLOCKF).
194 TAILQ_REMOVE(&lock
->lf_next
->lf_blkhd
, lock
,
196 _FREE(lock
, M_LOCKF
);
201 * No blocks!! Add the lock. Note that we will
202 * downgrade or upgrade any overlapping locks this
203 * process already owns.
205 * Skip over locks owned by other processes.
206 * Handle any locks that overlap and are owned by ourselves.
212 if (ovcase
= lf_findoverlap(block
, lock
, SELF
, &prev
, &overlap
))
213 block
= overlap
->lf_next
;
218 * 2) overlap contains lock
219 * 3) lock contains overlap
220 * 4) overlap starts before lock
221 * 5) overlap ends after lock
224 case 0: /* no overlap */
227 lock
->lf_next
= overlap
;
231 case 1: /* overlap == lock */
233 * If downgrading lock, others may be
234 * able to acquire it.
236 if (lock
->lf_type
== F_RDLCK
&&
237 overlap
->lf_type
== F_WRLCK
)
238 lf_wakelock(overlap
);
239 overlap
->lf_type
= lock
->lf_type
;
241 lock
= overlap
; /* for debug output below */
244 case 2: /* overlap contains lock */
246 * Check for common starting point and different types.
248 if (overlap
->lf_type
== lock
->lf_type
) {
249 _FREE(lock
, M_LOCKF
);
250 lock
= overlap
; /* for debug output below */
253 if (overlap
->lf_start
== lock
->lf_start
) {
255 lock
->lf_next
= overlap
;
256 overlap
->lf_start
= lock
->lf_end
+ 1;
258 lf_split(overlap
, lock
);
259 lf_wakelock(overlap
);
262 case 3: /* lock contains overlap */
264 * If downgrading lock, others may be able to
265 * acquire it, otherwise take the list.
267 if (lock
->lf_type
== F_RDLCK
&&
268 overlap
->lf_type
== F_WRLCK
) {
269 lf_wakelock(overlap
);
271 while (ltmp
= overlap
->lf_blkhd
.tqh_first
) {
272 TAILQ_REMOVE(&overlap
->lf_blkhd
, ltmp
,
274 TAILQ_INSERT_TAIL(&lock
->lf_blkhd
,
279 * Add the new lock if necessary and delete the overlap.
283 lock
->lf_next
= overlap
->lf_next
;
284 prev
= &lock
->lf_next
;
287 *prev
= overlap
->lf_next
;
288 _FREE(overlap
, M_LOCKF
);
291 case 4: /* overlap starts before lock */
293 * Add lock after overlap on the list.
295 lock
->lf_next
= overlap
->lf_next
;
296 overlap
->lf_next
= lock
;
297 overlap
->lf_end
= lock
->lf_start
- 1;
298 prev
= &lock
->lf_next
;
299 lf_wakelock(overlap
);
303 case 5: /* overlap ends after lock */
305 * Add the new lock before overlap.
309 lock
->lf_next
= overlap
;
311 overlap
->lf_start
= lock
->lf_end
+ 1;
312 lf_wakelock(overlap
);
318 if (lockf_debug
& 1) {
319 lf_print("lf_setlock: got the lock", lock
);
320 lf_printlist("lf_setlock", lock
);
322 #endif /* LOCKF_DEBUG */
327 * Remove a byte-range lock on an inode.
329 * Generally, find the lock (or an overlap to that lock)
330 * and remove it (or shrink it), then wakeup anyone we can.
334 register struct lockf
*unlock
;
336 struct inode
*ip
= unlock
->lf_inode
;
337 register struct lockf
*lf
= ip
->i_lockf
;
338 struct lockf
*overlap
, **prev
;
344 if (unlock
->lf_type
!= F_UNLCK
)
345 panic("lf_clearlock: bad type");
347 lf_print("lf_clearlock", unlock
);
348 #endif /* LOCKF_DEBUG */
350 while (ovcase
= lf_findoverlap(lf
, unlock
, SELF
, &prev
, &overlap
)) {
352 * Wakeup the list of locks to be retried.
354 lf_wakelock(overlap
);
358 case 1: /* overlap == lock */
359 *prev
= overlap
->lf_next
;
360 FREE(overlap
, M_LOCKF
);
363 case 2: /* overlap contains lock: split it */
364 if (overlap
->lf_start
== unlock
->lf_start
) {
365 overlap
->lf_start
= unlock
->lf_end
+ 1;
368 lf_split(overlap
, unlock
);
369 overlap
->lf_next
= unlock
->lf_next
;
372 case 3: /* lock contains overlap */
373 *prev
= overlap
->lf_next
;
374 lf
= overlap
->lf_next
;
375 _FREE(overlap
, M_LOCKF
);
378 case 4: /* overlap starts before lock */
379 overlap
->lf_end
= unlock
->lf_start
- 1;
380 prev
= &overlap
->lf_next
;
381 lf
= overlap
->lf_next
;
384 case 5: /* overlap ends after lock */
385 overlap
->lf_start
= unlock
->lf_end
+ 1;
392 lf_printlist("lf_clearlock", unlock
);
393 #endif /* LOCKF_DEBUG */
398 * Check whether there is a blocking lock,
399 * and if so return its process identifier.
403 register struct lockf
*lock
;
404 register struct flock
*fl
;
406 register struct lockf
*block
;
410 lf_print("lf_getlock", lock
);
411 #endif /* LOCKF_DEBUG */
413 if (block
= lf_getblock(lock
)) {
414 fl
->l_type
= block
->lf_type
;
415 fl
->l_whence
= SEEK_SET
;
416 fl
->l_start
= block
->lf_start
;
417 if (block
->lf_end
== -1)
420 fl
->l_len
= block
->lf_end
- block
->lf_start
+ 1;
421 if (block
->lf_flags
& F_POSIX
)
422 fl
->l_pid
= ((struct proc
*)(block
->lf_id
))->p_pid
;
426 fl
->l_type
= F_UNLCK
;
432 * Walk the list of locks for an inode and
433 * return the first blocking lock.
437 register struct lockf
*lock
;
439 struct lockf
**prev
, *overlap
, *lf
= lock
->lf_inode
->i_lockf
;
442 prev
= &lock
->lf_inode
->i_lockf
;
443 while (ovcase
= lf_findoverlap(lf
, lock
, OTHERS
, &prev
, &overlap
)) {
445 * We've found an overlap, see if it blocks us
447 if ((lock
->lf_type
== F_WRLCK
|| overlap
->lf_type
== F_WRLCK
))
450 * Nope, point to the next one on the list and
451 * see if it blocks us
453 lf
= overlap
->lf_next
;
459 * Walk the list of locks for an inode to
460 * find an overlapping lock (if any).
462 * NOTE: this returns only the FIRST overlapping lock. There
463 * may be more than one.
466 lf_findoverlap(lf
, lock
, type
, prev
, overlap
)
467 register struct lockf
*lf
;
470 struct lockf
***prev
;
471 struct lockf
**overlap
;
480 lf_print("lf_findoverlap: looking for overlap in", lock
);
481 #endif /* LOCKF_DEBUG */
482 start
= lock
->lf_start
;
484 while (lf
!= NOLOCKF
) {
485 if (((type
& SELF
) && lf
->lf_id
!= lock
->lf_id
) ||
486 ((type
& OTHERS
) && lf
->lf_id
== lock
->lf_id
)) {
487 *prev
= &lf
->lf_next
;
488 *overlap
= lf
= lf
->lf_next
;
493 lf_print("\tchecking", lf
);
494 #endif /* LOCKF_DEBUG */
496 * OK, check for overlap
501 * 2) overlap contains lock
502 * 3) lock contains overlap
503 * 4) overlap starts before lock
504 * 5) overlap ends after lock
506 if ((lf
->lf_end
!= -1 && start
> lf
->lf_end
) ||
507 (end
!= -1 && lf
->lf_start
> end
)) {
511 printf("no overlap\n");
512 #endif /* LOCKF_DEBUG */
513 if ((type
& SELF
) && end
!= -1 && lf
->lf_start
> end
)
515 *prev
= &lf
->lf_next
;
516 *overlap
= lf
= lf
->lf_next
;
519 if ((lf
->lf_start
== start
) && (lf
->lf_end
== end
)) {
523 printf("overlap == lock\n");
524 #endif /* LOCKF_DEBUG */
527 if ((lf
->lf_start
<= start
) &&
529 ((lf
->lf_end
>= end
) || (lf
->lf_end
== -1))) {
533 printf("overlap contains lock\n");
534 #endif /* LOCKF_DEBUG */
537 if (start
<= lf
->lf_start
&&
539 (lf
->lf_end
!= -1 && end
>= lf
->lf_end
))) {
543 printf("lock contains overlap\n");
544 #endif /* LOCKF_DEBUG */
547 if ((lf
->lf_start
< start
) &&
548 ((lf
->lf_end
>= start
) || (lf
->lf_end
== -1))) {
552 printf("overlap starts before lock\n");
553 #endif /* LOCKF_DEBUG */
556 if ((lf
->lf_start
> start
) &&
558 ((lf
->lf_end
> end
) || (lf
->lf_end
== -1))) {
562 printf("overlap ends after lock\n");
563 #endif /* LOCKF_DEBUG */
566 panic("lf_findoverlap: default");
572 * Split a lock and a contained region into
573 * two or three locks as necessary.
576 lf_split(lock1
, lock2
)
577 register struct lockf
*lock1
;
578 register struct lockf
*lock2
;
580 register struct lockf
*splitlock
;
583 if (lockf_debug
& 2) {
584 lf_print("lf_split", lock1
);
585 lf_print("splitting from", lock2
);
587 #endif /* LOCKF_DEBUG */
589 * Check to see if spliting into only two pieces.
591 if (lock1
->lf_start
== lock2
->lf_start
) {
592 lock1
->lf_start
= lock2
->lf_end
+ 1;
593 lock2
->lf_next
= lock1
;
596 if (lock1
->lf_end
== lock2
->lf_end
) {
597 lock1
->lf_end
= lock2
->lf_start
- 1;
598 lock2
->lf_next
= lock1
->lf_next
;
599 lock1
->lf_next
= lock2
;
603 * Make a new lock consisting of the last part of
604 * the encompassing lock
606 MALLOC(splitlock
, struct lockf
*, sizeof *splitlock
, M_LOCKF
, M_WAITOK
);
607 bcopy((caddr_t
)lock1
, (caddr_t
)splitlock
, sizeof *splitlock
);
608 splitlock
->lf_start
= lock2
->lf_end
+ 1;
609 TAILQ_INIT(&splitlock
->lf_blkhd
);
610 lock1
->lf_end
= lock2
->lf_start
- 1;
614 splitlock
->lf_next
= lock1
->lf_next
;
615 lock2
->lf_next
= splitlock
;
616 lock1
->lf_next
= lock2
;
623 lf_wakelock(listhead
)
624 struct lockf
*listhead
;
626 register struct lockf
*wakelock
;
628 while (wakelock
= listhead
->lf_blkhd
.tqh_first
) {
629 TAILQ_REMOVE(&listhead
->lf_blkhd
, wakelock
, lf_block
);
630 wakelock
->lf_next
= NOLOCKF
;
633 lf_print("lf_wakelock: awakening", wakelock
);
634 #endif /* LOCKF_DEBUG */
635 wakeup((caddr_t
)wakelock
);
645 register struct lockf
*lock
;
648 printf("%s: lock 0x%lx for ", tag
, lock
);
649 if (lock
->lf_flags
& F_POSIX
)
650 printf("proc %d", ((struct proc
*)(lock
->lf_id
))->p_pid
);
652 printf("id 0x%x", lock
->lf_id
);
653 printf(" in ino %d on dev <%d, %d>, %s, start %d, end %d",
654 lock
->lf_inode
->i_number
,
655 major(lock
->lf_inode
->i_dev
),
656 minor(lock
->lf_inode
->i_dev
),
657 lock
->lf_type
== F_RDLCK
? "shared" :
658 lock
->lf_type
== F_WRLCK
? "exclusive" :
659 lock
->lf_type
== F_UNLCK
? "unlock" :
660 "unknown", lock
->lf_start
, lock
->lf_end
);
661 if (lock
->lf_blkhd
.tqh_first
)
662 printf(" block 0x%x\n", lock
->lf_blkhd
.tqh_first
);
667 lf_printlist(tag
, lock
)
671 register struct lockf
*lf
, *blk
;
673 printf("%s: Lock list for ino %d on dev <%d, %d>:\n",
674 tag
, lock
->lf_inode
->i_number
,
675 major(lock
->lf_inode
->i_dev
),
676 minor(lock
->lf_inode
->i_dev
));
677 for (lf
= lock
->lf_inode
->i_lockf
; lf
; lf
= lf
->lf_next
) {
678 printf("\tlock 0x%lx for ", lf
);
679 if (lf
->lf_flags
& F_POSIX
)
680 printf("proc %d", ((struct proc
*)(lf
->lf_id
))->p_pid
);
682 printf("id 0x%x", lf
->lf_id
);
683 printf(", %s, start %d, end %d",
684 lf
->lf_type
== F_RDLCK
? "shared" :
685 lf
->lf_type
== F_WRLCK
? "exclusive" :
686 lf
->lf_type
== F_UNLCK
? "unlock" :
687 "unknown", lf
->lf_start
, lf
->lf_end
);
688 for (blk
= lf
->lf_blkhd
.tqh_first
; blk
;
689 blk
= blk
->lf_block
.tqe_next
) {
690 printf("\n\t\tlock request 0x%lx for ", blk
);
691 if (blk
->lf_flags
& F_POSIX
)
693 ((struct proc
*)(blk
->lf_id
))->p_pid
);
695 printf("id 0x%x", blk
->lf_id
);
696 printf(", %s, start %d, end %d",
697 blk
->lf_type
== F_RDLCK
? "shared" :
698 blk
->lf_type
== F_WRLCK
? "exclusive" :
699 blk
->lf_type
== F_UNLCK
? "unlock" :
700 "unknown", blk
->lf_start
, blk
->lf_end
);
701 if (blk
->lf_blkhd
.tqh_first
)
702 panic("lf_printlist: bad list");
707 #endif /* LOCKF_DEBUG */