Add VV_CROSSLOCK vnode flag to avoid cross-mount lookup LOR

When a lookup operation crosses into a new mountpoint, the mountpoint
must first be busied before the root vnode can be locked. When a
filesystem is unmounted, the vnode covered by the mountpoint must
first be locked, and then the busy count for the mountpoint drained.
Ordinarily, these two operations work fine if executed concurrently,
but with a stacked filesystem the root vnode may in fact use the
same lock as the covered vnode. By design, this will always be
the case for unionfs (with either the upper or lower root vnode
depending on mount options), and can also be the case for nullfs
if the target and mount point are the same (which admittedly is
very unlikely in practice).

In this case, we have LOR. The lookup path holds the mountpoint
busy while waiting on what is effectively the covered vnode lock,
while a concurrent unmount holds the covered vnode lock and waits
for the mountpoint's busy count to drain.

Attempt to resolve this LOR by allowing the stacked filesystem
to specify a new flag, VV_CROSSLOCK, on a covered vnode as necessary.
Upon observing this flag, the vfs_lookup() will leave the covered
vnode lock held while crossing into the mountpoint. Employ this flag
for unionfs with the caveat that it can't be used for '-o below' mounts
until other unionfs locking issues are resolved.

Reported by:	pho
Tested by:	pho
Reviewed by:	kib
Differential Revision:	https://reviews.freebsd.org/D35054
This commit is contained in:
Jason A. Harmening 2022-08-05 00:32:49 -05:00
parent 769b884e2e
commit 080ef8a418
4 changed files with 62 additions and 13 deletions

View File

@ -310,6 +310,30 @@ unionfs_domount(struct mount *mp)
return (ENOENT);
}
/*
* Specify that the covered vnode lock should remain held while
* lookup() performs the cross-mount walk. This prevents a lock-order
* reversal between the covered vnode lock (which is also locked by
* unionfs_lock()) and the mountpoint's busy count. Without this,
* unmount will lock the covered vnode lock (directly through the
* covered vnode) and wait for the busy count to drain, while a
* concurrent lookup will increment the busy count and then lock
* the covered vnode lock (indirectly through unionfs_lock()).
*
* Note that we can't yet use this facility for the 'below' case
* in which the upper vnode is the covered vnode, because that would
* introduce a different LOR in which the cross-mount lookup would
* effectively hold the upper vnode lock before acquiring the lower
* vnode lock, while an unrelated lock operation would still acquire
* the lower vnode lock before the upper vnode lock, which is the
* order unionfs currently requires.
*/
if (!below) {
vn_lock(mp->mnt_vnodecovered, LK_EXCLUSIVE | LK_RETRY);
mp->mnt_vnodecovered->v_vflag |= VV_CROSSLOCK;
VOP_UNLOCK(mp->mnt_vnodecovered);
}
MNT_ILOCK(mp);
if ((lowermp->mnt_flag & MNT_LOCAL) != 0 &&
(uppermp->mnt_flag & MNT_LOCAL) != 0)
@ -362,6 +386,9 @@ unionfs_unmount(struct mount *mp, int mntflags)
if (error)
return (error);
vn_lock(mp->mnt_vnodecovered, LK_EXCLUSIVE | LK_RETRY);
mp->mnt_vnodecovered->v_vflag &= ~VV_CROSSLOCK;
VOP_UNLOCK(mp->mnt_vnodecovered);
vfs_unregister_upper(ump->um_lowervp->v_mount, &ump->um_lower_link);
vfs_unregister_upper(ump->um_uppervp->v_mount, &ump->um_upper_link);
free(ump, M_UNIONFSMNT);

View File

@ -901,6 +901,8 @@ vfs_lookup(struct nameidata *ndp)
struct componentname *cnp = &ndp->ni_cnd;
int lkflags_save;
int ni_dvp_unlocked;
int crosslkflags;
bool crosslock;
/*
* Setup: break out flag bits into variables.
@ -1232,18 +1234,32 @@ vfs_lookup(struct nameidata *ndp)
*/
while (dp->v_type == VDIR && (mp = dp->v_mountedhere) &&
(cnp->cn_flags & NOCROSSMOUNT) == 0) {
crosslock = (dp->v_vflag & VV_CROSSLOCK) != 0;
crosslkflags = compute_cn_lkflags(mp, cnp->cn_lkflags,
cnp->cn_flags);
if (__predict_false(crosslock) &&
(crosslkflags & LK_EXCLUSIVE) != 0 &&
VOP_ISLOCKED(dp) != LK_EXCLUSIVE) {
vn_lock(dp, LK_UPGRADE | LK_RETRY);
if (VN_IS_DOOMED(dp)) {
error = ENOENT;
goto bad2;
}
}
if (vfs_busy(mp, 0))
continue;
vput(dp);
if (__predict_true(!crosslock))
vput(dp);
if (dp != ndp->ni_dvp)
vput(ndp->ni_dvp);
else
vrele(ndp->ni_dvp);
vrefact(vp_crossmp);
ndp->ni_dvp = vp_crossmp;
error = VFS_ROOT(mp, compute_cn_lkflags(mp, cnp->cn_lkflags,
cnp->cn_flags), &tdp);
error = VFS_ROOT(mp, crosslkflags, &tdp);
vfs_unbusy(mp);
if (__predict_false(crosslock))
vput(dp);
if (vn_lock(vp_crossmp, LK_SHARED | LK_NOWAIT))
panic("vp_crossmp exclusively locked or reclaimed");
if (error) {

View File

@ -770,17 +770,22 @@ SYSINIT(vfs, SI_SUB_VFS, SI_ORDER_FIRST, vntblinit, NULL);
* +->F->D->E
*
* The lookup() process for namei("/var") illustrates the process:
* VOP_LOOKUP() obtains B while A is held
* vfs_busy() obtains a shared lock on F while A and B are held
* vput() releases lock on B
* vput() releases lock on A
* VFS_ROOT() obtains lock on D while shared lock on F is held
* vfs_unbusy() releases shared lock on F
* vn_lock() obtains lock on deadfs vnode vp_crossmp instead of A.
* Attempt to lock A (instead of vp_crossmp) while D is held would
* violate the global order, causing deadlocks.
* 1. VOP_LOOKUP() obtains B while A is held
* 2. vfs_busy() obtains a shared lock on F while A and B are held
* 3. vput() releases lock on B
* 4. vput() releases lock on A
* 5. VFS_ROOT() obtains lock on D while shared lock on F is held
* 6. vfs_unbusy() releases shared lock on F
* 7. vn_lock() obtains lock on deadfs vnode vp_crossmp instead of A.
* Attempt to lock A (instead of vp_crossmp) while D is held would
* violate the global order, causing deadlocks.
*
* dounmount() locks B while F is drained.
* dounmount() locks B while F is drained. Note that for stacked
* filesystems, D and B in the example above may be the same lock,
* which introdues potential lock order reversal deadlock between
* dounmount() and step 5 above. These filesystems may avoid the LOR
* by setting VV_CROSSLOCK on the covered vnode so that lock B will
* remain held until after step 5.
*/
int
vfs_busy(struct mount *mp, int flags)

View File

@ -276,6 +276,7 @@ struct xvnode {
#define VV_FORCEINSMQ 0x1000 /* force the insmntque to succeed */
#define VV_READLINK 0x2000 /* fdescfs linux vnode */
#define VV_UNREF 0x4000 /* vunref, do not drop lock in inactive() */
#define VV_CROSSLOCK 0x8000 /* vnode lock is shared w/ root mounted here */
#define VMP_LAZYLIST 0x0001 /* Vnode is on mnt's lazy list */