version 1.86.4.1, 2007/03/13 17:51:21 |
version 1.99.2.1, 2008/09/18 04:37:05 |
|
|
/* $NetBSD$ */ |
/* $NetBSD$ */ |
|
|
|
/*- |
|
* Copyright (c) 2008 The NetBSD Foundation, Inc. |
|
* All rights reserved. |
|
* |
|
* This code is derived from software contributed to The NetBSD Foundation |
|
* by Wasabi Systems, Inc. |
|
* |
|
* Redistribution and use in source and binary forms, with or without |
|
* modification, are permitted provided that the following conditions |
|
* are met: |
|
* 1. Redistributions of source code must retain the above copyright |
|
* notice, this list of conditions and the following disclaimer. |
|
* 2. Redistributions in binary form must reproduce the above copyright |
|
* notice, this list of conditions and the following disclaimer in the |
|
* documentation and/or other materials provided with the distribution. |
|
* |
|
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS |
|
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED |
|
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR |
|
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS |
|
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
|
* POSSIBILITY OF SUCH DAMAGE. |
|
*/ |
|
|
/* |
/* |
* Copyright (c) 1982, 1986, 1989, 1993 |
* Copyright (c) 1982, 1986, 1989, 1993 |
* The Regents of the University of California. All rights reserved. |
* The Regents of the University of California. All rights reserved. |
|
|
#include <sys/cdefs.h> |
#include <sys/cdefs.h> |
__KERNEL_RCSID(0, "$NetBSD$"); |
__KERNEL_RCSID(0, "$NetBSD$"); |
|
|
|
#if defined(_KERNEL_OPT) |
|
#include "opt_ffs.h" |
|
#include "opt_wapbl.h" |
|
#endif |
|
|
#include <sys/param.h> |
#include <sys/param.h> |
#include <sys/systm.h> |
#include <sys/systm.h> |
#include <sys/resourcevar.h> |
#include <sys/resourcevar.h> |
Line 48 __KERNEL_RCSID(0, "$NetBSD$"); |
|
Line 82 __KERNEL_RCSID(0, "$NetBSD$"); |
|
#include <sys/pool.h> |
#include <sys/pool.h> |
#include <sys/signalvar.h> |
#include <sys/signalvar.h> |
#include <sys/kauth.h> |
#include <sys/kauth.h> |
|
#include <sys/wapbl.h> |
#include <sys/fstrans.h> |
#include <sys/fstrans.h> |
|
|
#include <miscfs/fifofs/fifo.h> |
#include <miscfs/fifofs/fifo.h> |
Line 58 __KERNEL_RCSID(0, "$NetBSD$"); |
|
Line 93 __KERNEL_RCSID(0, "$NetBSD$"); |
|
#include <ufs/ufs/dir.h> |
#include <ufs/ufs/dir.h> |
#include <ufs/ufs/ufs_extern.h> |
#include <ufs/ufs/ufs_extern.h> |
#include <ufs/ufs/ufsmount.h> |
#include <ufs/ufs/ufsmount.h> |
|
#include <ufs/ufs/ufs_wapbl.h> |
|
|
#include <ufs/ffs/fs.h> |
#include <ufs/ffs/fs.h> |
#include <ufs/ffs/ffs_extern.h> |
#include <ufs/ffs/ffs_extern.h> |
|
|
#include <uvm/uvm.h> |
#include <uvm/uvm.h> |
|
|
static int ffs_full_fsync(void *); |
|
|
|
/* Global vfs data structures for ufs. */ |
/* Global vfs data structures for ufs. */ |
int (**ffs_vnodeop_p)(void *); |
int (**ffs_vnodeop_p)(void *); |
const struct vnodeopv_entry_desc ffs_vnodeop_entries[] = { |
const struct vnodeopv_entry_desc ffs_vnodeop_entries[] = { |
Line 81 const struct vnodeopv_entry_desc ffs_vno |
|
Line 115 const struct vnodeopv_entry_desc ffs_vno |
|
{ &vop_setattr_desc, ufs_setattr }, /* setattr */ |
{ &vop_setattr_desc, ufs_setattr }, /* setattr */ |
{ &vop_read_desc, ffs_read }, /* read */ |
{ &vop_read_desc, ffs_read }, /* read */ |
{ &vop_write_desc, ffs_write }, /* write */ |
{ &vop_write_desc, ffs_write }, /* write */ |
{ &vop_lease_desc, ufs_lease_check }, /* lease */ |
|
{ &vop_ioctl_desc, ufs_ioctl }, /* ioctl */ |
{ &vop_ioctl_desc, ufs_ioctl }, /* ioctl */ |
{ &vop_fcntl_desc, ufs_fcntl }, /* fcntl */ |
{ &vop_fcntl_desc, ufs_fcntl }, /* fcntl */ |
{ &vop_poll_desc, ufs_poll }, /* poll */ |
{ &vop_poll_desc, ufs_poll }, /* poll */ |
Line 101 const struct vnodeopv_entry_desc ffs_vno |
|
Line 134 const struct vnodeopv_entry_desc ffs_vno |
|
{ &vop_abortop_desc, ufs_abortop }, /* abortop */ |
{ &vop_abortop_desc, ufs_abortop }, /* abortop */ |
{ &vop_inactive_desc, ufs_inactive }, /* inactive */ |
{ &vop_inactive_desc, ufs_inactive }, /* inactive */ |
{ &vop_reclaim_desc, ffs_reclaim }, /* reclaim */ |
{ &vop_reclaim_desc, ffs_reclaim }, /* reclaim */ |
{ &vop_lock_desc, ufs_lock }, /* lock */ |
{ &vop_lock_desc, ffs_lock }, /* lock */ |
{ &vop_unlock_desc, ufs_unlock }, /* unlock */ |
{ &vop_unlock_desc, ffs_unlock }, /* unlock */ |
{ &vop_bmap_desc, ufs_bmap }, /* bmap */ |
{ &vop_bmap_desc, ufs_bmap }, /* bmap */ |
{ &vop_strategy_desc, ufs_strategy }, /* strategy */ |
{ &vop_strategy_desc, ufs_strategy }, /* strategy */ |
{ &vop_print_desc, ufs_print }, /* print */ |
{ &vop_print_desc, ufs_print }, /* print */ |
{ &vop_islocked_desc, ufs_islocked }, /* islocked */ |
{ &vop_islocked_desc, ffs_islocked }, /* islocked */ |
{ &vop_pathconf_desc, ufs_pathconf }, /* pathconf */ |
{ &vop_pathconf_desc, ufs_pathconf }, /* pathconf */ |
{ &vop_advlock_desc, ufs_advlock }, /* advlock */ |
{ &vop_advlock_desc, ufs_advlock }, /* advlock */ |
{ &vop_bwrite_desc, vn_bwrite }, /* bwrite */ |
{ &vop_bwrite_desc, vn_bwrite }, /* bwrite */ |
{ &vop_getpages_desc, ffs_getpages }, /* getpages */ |
{ &vop_getpages_desc, genfs_getpages }, /* getpages */ |
{ &vop_putpages_desc, genfs_putpages }, /* putpages */ |
{ &vop_putpages_desc, genfs_putpages }, /* putpages */ |
{ &vop_openextattr_desc, ffs_openextattr }, /* openextattr */ |
{ &vop_openextattr_desc, ffs_openextattr }, /* openextattr */ |
{ &vop_closeextattr_desc, ffs_closeextattr }, /* closeextattr */ |
{ &vop_closeextattr_desc, ffs_closeextattr }, /* closeextattr */ |
Line 136 const struct vnodeopv_entry_desc ffs_spe |
|
Line 169 const struct vnodeopv_entry_desc ffs_spe |
|
{ &vop_setattr_desc, ufs_setattr }, /* setattr */ |
{ &vop_setattr_desc, ufs_setattr }, /* setattr */ |
{ &vop_read_desc, ufsspec_read }, /* read */ |
{ &vop_read_desc, ufsspec_read }, /* read */ |
{ &vop_write_desc, ufsspec_write }, /* write */ |
{ &vop_write_desc, ufsspec_write }, /* write */ |
{ &vop_lease_desc, spec_lease_check }, /* lease */ |
|
{ &vop_ioctl_desc, spec_ioctl }, /* ioctl */ |
{ &vop_ioctl_desc, spec_ioctl }, /* ioctl */ |
{ &vop_fcntl_desc, ufs_fcntl }, /* fcntl */ |
{ &vop_fcntl_desc, ufs_fcntl }, /* fcntl */ |
{ &vop_poll_desc, spec_poll }, /* poll */ |
{ &vop_poll_desc, spec_poll }, /* poll */ |
Line 156 const struct vnodeopv_entry_desc ffs_spe |
|
Line 188 const struct vnodeopv_entry_desc ffs_spe |
|
{ &vop_abortop_desc, spec_abortop }, /* abortop */ |
{ &vop_abortop_desc, spec_abortop }, /* abortop */ |
{ &vop_inactive_desc, ufs_inactive }, /* inactive */ |
{ &vop_inactive_desc, ufs_inactive }, /* inactive */ |
{ &vop_reclaim_desc, ffs_reclaim }, /* reclaim */ |
{ &vop_reclaim_desc, ffs_reclaim }, /* reclaim */ |
{ &vop_lock_desc, ufs_lock }, /* lock */ |
{ &vop_lock_desc, ffs_lock }, /* lock */ |
{ &vop_unlock_desc, ufs_unlock }, /* unlock */ |
{ &vop_unlock_desc, ffs_unlock }, /* unlock */ |
{ &vop_bmap_desc, spec_bmap }, /* bmap */ |
{ &vop_bmap_desc, spec_bmap }, /* bmap */ |
{ &vop_strategy_desc, spec_strategy }, /* strategy */ |
{ &vop_strategy_desc, spec_strategy }, /* strategy */ |
{ &vop_print_desc, ufs_print }, /* print */ |
{ &vop_print_desc, ufs_print }, /* print */ |
{ &vop_islocked_desc, ufs_islocked }, /* islocked */ |
{ &vop_islocked_desc, ffs_islocked }, /* islocked */ |
{ &vop_pathconf_desc, spec_pathconf }, /* pathconf */ |
{ &vop_pathconf_desc, spec_pathconf }, /* pathconf */ |
{ &vop_advlock_desc, spec_advlock }, /* advlock */ |
{ &vop_advlock_desc, spec_advlock }, /* advlock */ |
{ &vop_bwrite_desc, vn_bwrite }, /* bwrite */ |
{ &vop_bwrite_desc, vn_bwrite }, /* bwrite */ |
Line 191 const struct vnodeopv_entry_desc ffs_fif |
|
Line 223 const struct vnodeopv_entry_desc ffs_fif |
|
{ &vop_setattr_desc, ufs_setattr }, /* setattr */ |
{ &vop_setattr_desc, ufs_setattr }, /* setattr */ |
{ &vop_read_desc, ufsfifo_read }, /* read */ |
{ &vop_read_desc, ufsfifo_read }, /* read */ |
{ &vop_write_desc, ufsfifo_write }, /* write */ |
{ &vop_write_desc, ufsfifo_write }, /* write */ |
{ &vop_lease_desc, fifo_lease_check }, /* lease */ |
|
{ &vop_ioctl_desc, fifo_ioctl }, /* ioctl */ |
{ &vop_ioctl_desc, fifo_ioctl }, /* ioctl */ |
{ &vop_fcntl_desc, ufs_fcntl }, /* fcntl */ |
{ &vop_fcntl_desc, ufs_fcntl }, /* fcntl */ |
{ &vop_poll_desc, fifo_poll }, /* poll */ |
{ &vop_poll_desc, fifo_poll }, /* poll */ |
Line 211 const struct vnodeopv_entry_desc ffs_fif |
|
Line 242 const struct vnodeopv_entry_desc ffs_fif |
|
{ &vop_abortop_desc, fifo_abortop }, /* abortop */ |
{ &vop_abortop_desc, fifo_abortop }, /* abortop */ |
{ &vop_inactive_desc, ufs_inactive }, /* inactive */ |
{ &vop_inactive_desc, ufs_inactive }, /* inactive */ |
{ &vop_reclaim_desc, ffs_reclaim }, /* reclaim */ |
{ &vop_reclaim_desc, ffs_reclaim }, /* reclaim */ |
{ &vop_lock_desc, ufs_lock }, /* lock */ |
{ &vop_lock_desc, ffs_lock }, /* lock */ |
{ &vop_unlock_desc, ufs_unlock }, /* unlock */ |
{ &vop_unlock_desc, ffs_unlock }, /* unlock */ |
{ &vop_bmap_desc, fifo_bmap }, /* bmap */ |
{ &vop_bmap_desc, fifo_bmap }, /* bmap */ |
{ &vop_strategy_desc, fifo_strategy }, /* strategy */ |
{ &vop_strategy_desc, fifo_strategy }, /* strategy */ |
{ &vop_print_desc, ufs_print }, /* print */ |
{ &vop_print_desc, ufs_print }, /* print */ |
{ &vop_islocked_desc, ufs_islocked }, /* islocked */ |
{ &vop_islocked_desc, ffs_islocked }, /* islocked */ |
{ &vop_pathconf_desc, fifo_pathconf }, /* pathconf */ |
{ &vop_pathconf_desc, fifo_pathconf }, /* pathconf */ |
{ &vop_advlock_desc, fifo_advlock }, /* advlock */ |
{ &vop_advlock_desc, fifo_advlock }, /* advlock */ |
{ &vop_bwrite_desc, vn_bwrite }, /* bwrite */ |
{ &vop_bwrite_desc, vn_bwrite }, /* bwrite */ |
Line 246 ffs_fsync(void *v) |
|
Line 277 ffs_fsync(void *v) |
|
struct lwp *a_l; |
struct lwp *a_l; |
} */ *ap = v; |
} */ *ap = v; |
struct buf *bp; |
struct buf *bp; |
int s, num, error, i; |
int num, error, i; |
struct indir ia[NIADDR + 1]; |
struct indir ia[NIADDR + 1]; |
int bsize; |
int bsize; |
daddr_t blk_high; |
daddr_t blk_high; |
struct vnode *vp; |
struct vnode *vp; |
|
#ifdef WAPBL |
|
struct mount *mp; |
|
#endif |
|
|
vp = ap->a_vp; |
vp = ap->a_vp; |
|
|
if ((error = fstrans_start(vp->v_mount, FSTRANS_LAZY)) != 0) |
fstrans_start(vp->v_mount, FSTRANS_LAZY); |
return error; |
|
/* |
/* |
* XXX no easy way to sync a range in a file with softdep. |
* XXX no easy way to sync a range in a file with softdep. |
*/ |
*/ |
if ((ap->a_offlo == 0 && ap->a_offhi == 0) || DOINGSOFTDEP(vp) || |
if ((ap->a_offlo == 0 && ap->a_offhi == 0) || DOINGSOFTDEP(vp) || |
(vp->v_type != VREG)) { |
(vp->v_type != VREG)) { |
error = ffs_full_fsync(v); |
int flags = ap->a_flags; |
|
|
|
if (vp->v_type == VBLK) |
|
flags |= FSYNC_VFS; |
|
error = ffs_full_fsync(vp, flags); |
goto out; |
goto out; |
} |
} |
|
|
Line 282 ffs_fsync(void *v) |
|
Line 319 ffs_fsync(void *v) |
|
goto out; |
goto out; |
} |
} |
|
|
|
#ifdef WAPBL |
|
mp = wapbl_vptomp(vp); |
|
if (mp->mnt_wapbl) { |
|
if (ap->a_flags & FSYNC_DATAONLY) { |
|
fstrans_done(vp->v_mount); |
|
return 0; |
|
} |
|
error = 0; |
|
if (vp->v_tag == VT_UFS && VTOI(vp)->i_flag & |
|
(IN_ACCESS | IN_CHANGE | IN_UPDATE | IN_MODIFY | |
|
IN_MODIFIED | IN_ACCESSED)) { |
|
error = UFS_WAPBL_BEGIN(mp); |
|
if (error) { |
|
fstrans_done(vp->v_mount); |
|
return error; |
|
} |
|
error = ffs_update(vp, NULL, NULL, |
|
(ap->a_flags & FSYNC_WAIT) ? UPDATE_WAIT : 0); |
|
UFS_WAPBL_END(mp); |
|
} |
|
if (error || (ap->a_flags & FSYNC_NOLOG)) { |
|
fstrans_done(vp->v_mount); |
|
return error; |
|
} |
|
error = wapbl_flush(mp->mnt_wapbl, 0); |
|
fstrans_done(vp->v_mount); |
|
return error; |
|
} |
|
#endif /* WAPBL */ |
|
|
/* |
/* |
* Then, flush indirect blocks. |
* Then, flush indirect blocks. |
*/ |
*/ |
|
|
s = splbio(); |
|
if (blk_high >= NDADDR) { |
if (blk_high >= NDADDR) { |
error = ufs_getlbns(vp, blk_high, ia, &num); |
error = ufs_getlbns(vp, blk_high, ia, &num); |
if (error) { |
if (error) |
splx(s); |
|
goto out; |
goto out; |
} |
|
|
mutex_enter(&bufcache_lock); |
for (i = 0; i < num; i++) { |
for (i = 0; i < num; i++) { |
bp = incore(vp, ia[i].in_lbn); |
if ((bp = incore(vp, ia[i].in_lbn)) == NULL) |
if (bp != NULL) { |
continue; |
mutex_enter(&bp->b_interlock); |
if ((bp->b_cflags & BC_BUSY) != 0 || |
if (!(bp->b_flags & B_BUSY) && (bp->b_flags & B_DELWRI)) { |
(bp->b_oflags & BO_DELWRI) == 0) |
bp->b_flags |= B_BUSY | B_VFLUSH; |
continue; |
mutex_exit(&bp->b_interlock); |
bp->b_cflags |= BC_BUSY | BC_VFLUSH; |
splx(s); |
mutex_exit(&bufcache_lock); |
bawrite(bp); |
bawrite(bp); |
s = splbio(); |
mutex_enter(&bufcache_lock); |
} else { |
|
mutex_exit(&bp->b_interlock); |
|
} |
|
} |
|
} |
} |
|
mutex_exit(&bufcache_lock); |
} |
} |
|
|
if (ap->a_flags & FSYNC_WAIT) { |
if (ap->a_flags & FSYNC_WAIT) { |
mutex_enter(&global_v_numoutput_lock); |
mutex_enter(&vp->v_interlock); |
while (vp->v_numoutput > 0) { |
while (vp->v_numoutput > 0) |
vp->v_flag |= VBWAIT; |
cv_wait(&vp->v_cv, &vp->v_interlock); |
mtsleep(&vp->v_numoutput, PRIBIO + 1, "fsync_range", 0, |
mutex_exit(&vp->v_interlock); |
&global_v_numoutput_lock); |
|
} |
|
mutex_exit(&global_v_numoutput_lock); |
|
} |
} |
splx(s); |
|
|
|
error = ffs_update(vp, NULL, NULL, |
error = ffs_update(vp, NULL, NULL, |
((ap->a_flags & (FSYNC_WAIT | FSYNC_DATAONLY)) == FSYNC_WAIT) |
((ap->a_flags & (FSYNC_WAIT | FSYNC_DATAONLY)) == FSYNC_WAIT) |
Line 328 ffs_fsync(void *v) |
|
Line 387 ffs_fsync(void *v) |
|
if (error == 0 && ap->a_flags & FSYNC_CACHE) { |
if (error == 0 && ap->a_flags & FSYNC_CACHE) { |
int l = 0; |
int l = 0; |
VOP_IOCTL(VTOI(vp)->i_devvp, DIOCCACHESYNC, &l, FWRITE, |
VOP_IOCTL(VTOI(vp)->i_devvp, DIOCCACHESYNC, &l, FWRITE, |
ap->a_l->l_cred, ap->a_l); |
curlwp->l_cred); |
} |
} |
|
|
out: |
out: |
|
|
* Synch an open file. |
* Synch an open file. |
*/ |
*/ |
/* ARGSUSED */ |
/* ARGSUSED */ |
static int |
int |
ffs_full_fsync(void *v) |
ffs_full_fsync(struct vnode *vp, int flags) |
{ |
{ |
struct vop_fsync_args /* { |
|
struct vnode *a_vp; |
|
kauth_cred_t a_cred; |
|
int a_flags; |
|
off_t a_offlo; |
|
off_t a_offhi; |
|
struct lwp *a_l; |
|
} */ *ap = v; |
|
struct vnode *vp = ap->a_vp; |
|
struct buf *bp, *nbp; |
struct buf *bp, *nbp; |
int s, error, passes, skipmeta, inodedeps_only, waitfor; |
int error, passes, skipmeta, inodedeps_only, waitfor; |
|
struct mount *mp; |
|
|
|
error = 0; |
|
|
if (vp->v_type == VBLK && |
if (vp->v_type == VBLK && |
vp->v_specmountpoint != NULL && |
vp->v_specmountpoint != NULL && |
(vp->v_specmountpoint->mnt_flag & MNT_SOFTDEP)) |
(vp->v_specmountpoint->mnt_flag & MNT_SOFTDEP)) |
softdep_fsync_mountdev(vp); |
softdep_fsync_mountdev(vp); |
|
|
inodedeps_only = DOINGSOFTDEP(vp) && (ap->a_flags & FSYNC_RECLAIM) |
mutex_enter(&vp->v_interlock); |
&& vp->v_uobj.uo_npages == 0 && LIST_EMPTY(&vp->v_dirtyblkhd); |
|
|
inodedeps_only = DOINGSOFTDEP(vp) && (flags & FSYNC_RECLAIM) |
|
&& UVM_OBJ_IS_CLEAN(&vp->v_uobj) && LIST_EMPTY(&vp->v_dirtyblkhd); |
|
|
/* |
/* |
* Flush all dirty data associated with a vnode. |
* Flush all dirty data associated with a vnode. |
*/ |
*/ |
|
|
if (vp->v_type == VREG || vp->v_type == VBLK) { |
if (vp->v_type == VREG || vp->v_type == VBLK) { |
mutex_enter(&vp->v_interlock); |
int pflags = PGO_ALLPAGES | PGO_CLEANIT; |
error = VOP_PUTPAGES(vp, 0, 0, PGO_ALLPAGES | PGO_CLEANIT | |
|
((ap->a_flags & FSYNC_WAIT) ? PGO_SYNCIO : 0) | |
if ((flags & FSYNC_VFS) != 0 && vp->v_specmountpoint != NULL) |
(fstrans_getstate(vp->v_mount) == FSTRANS_SUSPENDING ? |
mp = vp->v_specmountpoint; |
PGO_FREE : 0)); |
else |
if (error) { |
mp = vp->v_mount; |
|
|
|
if ((flags & FSYNC_WAIT)) |
|
pflags |= PGO_SYNCIO; |
|
if (vp->v_type == VREG && |
|
fstrans_getstate(mp) == FSTRANS_SUSPENDING) |
|
pflags |= PGO_FREE; |
|
error = VOP_PUTPAGES(vp, 0, 0, pflags); |
|
if (error) |
|
return error; |
|
} else { |
|
mp = vp->v_mount; |
|
mutex_exit(&vp->v_interlock); |
|
} |
|
|
|
#ifdef WAPBL |
|
if (mp && mp->mnt_wapbl) { |
|
error = 0; |
|
if (flags & FSYNC_DATAONLY) |
return error; |
return error; |
|
|
|
if (VTOI(vp) && (VTOI(vp)->i_flag & |
|
(IN_ACCESS | IN_CHANGE | IN_UPDATE | IN_MODIFY | |
|
IN_MODIFIED | IN_ACCESSED))) { |
|
error = UFS_WAPBL_BEGIN(mp); |
|
if (error) |
|
return error; |
|
error = ffs_update(vp, NULL, NULL, |
|
(flags & FSYNC_WAIT) ? UPDATE_WAIT : 0); |
|
UFS_WAPBL_END(mp); |
|
} |
|
if (error || (flags & FSYNC_NOLOG)) |
|
return error; |
|
/* |
|
* Don't flush the log if the vnode being flushed |
|
* contains no dirty buffers that could be in the log. |
|
*/ |
|
if (!((flags & FSYNC_RECLAIM) && |
|
LIST_EMPTY(&vp->v_dirtyblkhd))) { |
|
error = wapbl_flush(mp->mnt_wapbl, 0); |
|
if (error) |
|
return error; |
} |
} |
|
|
|
/* |
|
* XXX temporary workaround for "dirty bufs" panic in |
|
* vinvalbuf. need a full fix for the v_numoutput |
|
* waiters issues. |
|
*/ |
|
if (flags & FSYNC_WAIT) { |
|
mutex_enter(&vp->v_interlock); |
|
while (vp->v_numoutput) |
|
cv_wait(&vp->v_cv, &vp->v_interlock); |
|
mutex_exit(&vp->v_interlock); |
|
} |
|
|
|
return error; |
} |
} |
|
#endif /* WAPBL */ |
|
|
passes = NIADDR + 1; |
passes = NIADDR + 1; |
skipmeta = 0; |
skipmeta = 0; |
if (ap->a_flags & FSYNC_WAIT) |
if (flags & FSYNC_WAIT) |
skipmeta = 1; |
skipmeta = 1; |
s = splbio(); |
|
|
|
loop: |
loop: |
LIST_FOREACH(bp, &vp->v_dirtyblkhd, b_vnbufs) |
mutex_enter(&bufcache_lock); |
bp->b_flags &= ~B_SCANNED; |
LIST_FOREACH(bp, &vp->v_dirtyblkhd, b_vnbufs) { |
|
bp->b_cflags &= ~BC_SCANNED; |
|
} |
for (bp = LIST_FIRST(&vp->v_dirtyblkhd); bp; bp = nbp) { |
for (bp = LIST_FIRST(&vp->v_dirtyblkhd); bp; bp = nbp) { |
nbp = LIST_NEXT(bp, b_vnbufs); |
nbp = LIST_NEXT(bp, b_vnbufs); |
mutex_enter(&bp->b_interlock); |
if (bp->b_cflags & (BC_BUSY | BC_SCANNED)) |
if (bp->b_flags & (B_BUSY | B_SCANNED)) { |
|
mutex_exit(&bp->b_interlock); |
|
continue; |
continue; |
} |
if ((bp->b_oflags & BO_DELWRI) == 0) |
if ((bp->b_flags & B_DELWRI) == 0) |
|
panic("ffs_fsync: not dirty"); |
panic("ffs_fsync: not dirty"); |
if (skipmeta && bp->b_lblkno < 0) { |
if (skipmeta && bp->b_lblkno < 0) |
mutex_exit(&bp->b_interlock); |
|
continue; |
continue; |
} |
bp->b_cflags |= BC_BUSY | BC_VFLUSH | BC_SCANNED; |
mutex_exit(&bp->b_interlock); |
mutex_exit(&bufcache_lock); |
bp->b_flags |= B_BUSY | B_VFLUSH | B_SCANNED; |
|
splx(s); |
|
/* |
/* |
* On our final pass through, do all I/O synchronously |
* On our final pass through, do all I/O synchronously |
* so that we can find out if our flush is failing |
* so that we can find out if our flush is failing |
* because of write errors. |
* because of write errors. |
*/ |
*/ |
if (passes > 0 || !(ap->a_flags & FSYNC_WAIT)) |
if (passes > 0 || !(flags & FSYNC_WAIT)) |
(void) bawrite(bp); |
(void) bawrite(bp); |
else if ((error = bwrite(bp)) != 0) |
else if ((error = bwrite(bp)) != 0) |
return (error); |
return (error); |
s = splbio(); |
|
/* |
/* |
* Since we may have slept during the I/O, we need |
* Since we unlocked during the I/O, we need |
* to start from a known point. |
* to start from a known point. |
*/ |
*/ |
|
mutex_enter(&bufcache_lock); |
nbp = LIST_FIRST(&vp->v_dirtyblkhd); |
nbp = LIST_FIRST(&vp->v_dirtyblkhd); |
} |
} |
|
mutex_exit(&bufcache_lock); |
if (skipmeta) { |
if (skipmeta) { |
skipmeta = 0; |
skipmeta = 0; |
goto loop; |
goto loop; |
} |
} |
if (ap->a_flags & FSYNC_WAIT) { |
|
mutex_enter(&global_v_numoutput_lock); |
if (flags & FSYNC_WAIT) { |
|
mutex_enter(&vp->v_interlock); |
while (vp->v_numoutput) { |
while (vp->v_numoutput) { |
vp->v_flag |= VBWAIT; |
cv_wait(&vp->v_cv, &vp->v_interlock); |
(void) mtsleep(&vp->v_numoutput, PRIBIO + 1, |
|
"ffsfsync", 0, &global_v_numoutput_lock); |
|
} |
} |
mutex_exit(&global_v_numoutput_lock); |
mutex_exit(&vp->v_interlock); |
splx(s); |
|
|
|
/* |
/* |
* Ensure that any filesystem metadata associated |
* Ensure that any filesystem metadata associated |
* with the vnode has been written. |
* with the vnode has been written. |
*/ |
*/ |
if ((error = softdep_sync_metadata(ap)) != 0) |
if ((error = softdep_sync_metadata(vp)) != 0) |
return (error); |
return (error); |
|
|
s = splbio(); |
|
if (!LIST_EMPTY(&vp->v_dirtyblkhd)) { |
if (!LIST_EMPTY(&vp->v_dirtyblkhd)) { |
/* |
/* |
* Block devices associated with filesystems may |
* Block devices associated with filesystems may |
|
|
#endif |
#endif |
} |
} |
} |
} |
splx(s); |
|
|
|
if (inodedeps_only) |
if (inodedeps_only) |
waitfor = 0; |
waitfor = 0; |
else |
else |
waitfor = (ap->a_flags & FSYNC_WAIT) ? UPDATE_WAIT : 0; |
waitfor = (flags & FSYNC_WAIT) ? UPDATE_WAIT : 0; |
error = ffs_update(vp, NULL, NULL, waitfor); |
|
|
|
if (error == 0 && ap->a_flags & FSYNC_CACHE) { |
if (vp->v_tag == VT_UFS) |
|
error = ffs_update(vp, NULL, NULL, waitfor); |
|
else { |
|
KASSERT(vp->v_type == VBLK); |
|
KASSERT((flags & FSYNC_VFS) != 0); |
|
} |
|
|
|
if (error == 0 && flags & FSYNC_CACHE) { |
int i = 0; |
int i = 0; |
VOP_IOCTL(VTOI(vp)->i_devvp, DIOCCACHESYNC, &i, FWRITE, |
if ((flags & FSYNC_VFS) == 0) { |
ap->a_l->l_cred, ap->a_l); |
KASSERT(VTOI(vp) != NULL); |
|
vp = VTOI(vp)->i_devvp; |
|
} |
|
VOP_IOCTL(vp, DIOCCACHESYNC, &i, FWRITE, curlwp->l_cred); |
} |
} |
|
|
return error; |
return error; |
Line 491 ffs_reclaim(void *v) |
|
Line 602 ffs_reclaim(void *v) |
|
struct inode *ip = VTOI(vp); |
struct inode *ip = VTOI(vp); |
struct mount *mp = vp->v_mount; |
struct mount *mp = vp->v_mount; |
struct ufsmount *ump = ip->i_ump; |
struct ufsmount *ump = ip->i_ump; |
|
void *data; |
int error; |
int error; |
|
|
if ((error = fstrans_start(mp, FSTRANS_LAZY)) != 0) |
fstrans_start(mp, FSTRANS_LAZY); |
return error; |
if ((error = ufs_reclaim(vp)) != 0) { |
if ((error = ufs_reclaim(vp, ap->a_l)) != 0) { |
|
fstrans_done(mp); |
fstrans_done(mp); |
return (error); |
return (error); |
} |
} |
if (ip->i_din.ffs1_din != NULL) { |
if (ip->i_din.ffs1_din != NULL) { |
if (ump->um_fstype == UFS1) |
if (ump->um_fstype == UFS1) |
pool_put(&ffs_dinode1_pool, ip->i_din.ffs1_din); |
pool_cache_put(ffs_dinode1_cache, ip->i_din.ffs1_din); |
else |
else |
pool_put(&ffs_dinode2_pool, ip->i_din.ffs2_din); |
pool_cache_put(ffs_dinode2_cache, ip->i_din.ffs2_din); |
} |
} |
/* |
/* |
* XXX MFS ends up here, too, to free an inode. Should we create |
* To interlock with ffs_sync(). |
* XXX a separate pool for MFS inodes? |
|
*/ |
*/ |
genfs_node_destroy(vp); |
genfs_node_destroy(vp); |
pool_put(&ffs_inode_pool, vp->v_data); |
mutex_enter(&vp->v_interlock); |
|
data = vp->v_data; |
vp->v_data = NULL; |
vp->v_data = NULL; |
|
mutex_exit(&vp->v_interlock); |
|
|
|
/* |
|
* XXX MFS ends up here, too, to free an inode. Should we create |
|
* XXX a separate pool for MFS inodes? |
|
*/ |
|
pool_cache_put(ffs_inode_cache, data); |
fstrans_done(mp); |
fstrans_done(mp); |
return (0); |
return (0); |
} |
} |
|
|
|
#if 0 |
int |
int |
ffs_getpages(void *v) |
ffs_getpages(void *v) |
{ |
{ |
Line 550 ffs_getpages(void *v) |
|
Line 669 ffs_getpages(void *v) |
|
} |
} |
return genfs_getpages(v); |
return genfs_getpages(v); |
} |
} |
|
#endif |
|
|
/* |
/* |
* Return the last logical file offset that should be written for this file |
* Return the last logical file offset that should be written for this file |
Line 573 ffs_gop_size(struct vnode *vp, off_t siz |
|
Line 693 ffs_gop_size(struct vnode *vp, off_t siz |
|
} |
} |
|
|
int |
int |
|
ffs_gop_write(struct vnode *vp, struct vm_page **pgs, int npages, int flags) |
|
{ |
|
int error; |
|
const bool need_wapbl = (curlwp != uvm.pagedaemon_lwp && |
|
vp->v_mount->mnt_wapbl && (flags & PGO_JOURNALLOCKED) == 0); |
|
|
|
if (need_wapbl) { |
|
error = UFS_WAPBL_BEGIN(vp->v_mount); |
|
if (error) |
|
return error; |
|
} |
|
error = genfs_gop_write(vp, pgs, npages, flags); |
|
if (need_wapbl) |
|
UFS_WAPBL_END(vp->v_mount); |
|
return error; |
|
} |
|
|
|
int |
ffs_openextattr(void *v) |
ffs_openextattr(void *v) |
{ |
{ |
struct vop_openextattr_args /* { |
struct vop_openextattr_args /* { |
Line 631 ffs_getextattr(void *v) |
|
Line 769 ffs_getextattr(void *v) |
|
#ifdef UFS_EXTATTR |
#ifdef UFS_EXTATTR |
int error; |
int error; |
|
|
if ((error = fstrans_start(vp->v_mount, FSTRANS_SHARED)) != 0) |
fstrans_start(vp->v_mount, FSTRANS_SHARED); |
return error; |
|
error = ufs_getextattr(ap); |
error = ufs_getextattr(ap); |
fstrans_done(vp->v_mount); |
fstrans_done(vp->v_mount); |
return error; |
return error; |
Line 664 ffs_setextattr(void *v) |
|
Line 801 ffs_setextattr(void *v) |
|
#ifdef UFS_EXTATTR |
#ifdef UFS_EXTATTR |
int error; |
int error; |
|
|
if ((error = fstrans_start(vp->v_mount, FSTRANS_SHARED)) != 0) |
fstrans_start(vp->v_mount, FSTRANS_SHARED); |
return error; |
|
error = ufs_setextattr(ap); |
error = ufs_setextattr(ap); |
fstrans_done(vp->v_mount); |
fstrans_done(vp->v_mount); |
return error; |
return error; |
Line 717 ffs_deleteextattr(void *v) |
|
Line 853 ffs_deleteextattr(void *v) |
|
#ifdef UFS_EXTATTR |
#ifdef UFS_EXTATTR |
int error; |
int error; |
|
|
if ((error = fstrans_start(vp->v_mount, FSTRANS_SHARED)) != 0) |
fstrans_start(vp->v_mount, FSTRANS_SHARED); |
return error; |
|
error = ufs_deleteextattr(ap); |
error = ufs_deleteextattr(ap); |
fstrans_done(vp->v_mount); |
fstrans_done(vp->v_mount); |
return error; |
return error; |
Line 730 ffs_deleteextattr(void *v) |
|
Line 865 ffs_deleteextattr(void *v) |
|
/* XXX Not implemented for UFS2 file systems. */ |
/* XXX Not implemented for UFS2 file systems. */ |
return (EOPNOTSUPP); |
return (EOPNOTSUPP); |
} |
} |
|
|
|
/* |
|
* Lock the node. |
|
*/ |
|
int |
|
ffs_lock(void *v) |
|
{ |
|
struct vop_lock_args /* { |
|
struct vnode *a_vp; |
|
int a_flags; |
|
} */ *ap = v; |
|
struct vnode *vp = ap->a_vp; |
|
struct mount *mp = vp->v_mount; |
|
int flags = ap->a_flags; |
|
|
|
if ((flags & LK_INTERLOCK) != 0) { |
|
mutex_exit(&vp->v_interlock); |
|
flags &= ~LK_INTERLOCK; |
|
} |
|
|
|
/* |
|
* Fake lock during file system suspension. |
|
*/ |
|
if ((vp->v_type == VREG || vp->v_type == VDIR) && |
|
fstrans_is_owner(mp) && |
|
fstrans_getstate(mp) == FSTRANS_SUSPENDING) { |
|
return 0; |
|
} |
|
|
|
return (vlockmgr(vp->v_vnlock, flags)); |
|
} |
|
|
|
/* |
|
* Unlock the node. |
|
*/ |
|
int |
|
ffs_unlock(void *v) |
|
{ |
|
struct vop_unlock_args /* { |
|
struct vnode *a_vp; |
|
int a_flags; |
|
} */ *ap = v; |
|
struct vnode *vp = ap->a_vp; |
|
struct mount *mp = vp->v_mount; |
|
|
|
KASSERT(ap->a_flags == 0); |
|
|
|
/* |
|
* Fake unlock during file system suspension. |
|
*/ |
|
if ((vp->v_type == VREG || vp->v_type == VDIR) && |
|
fstrans_is_owner(mp) && |
|
fstrans_getstate(mp) == FSTRANS_SUSPENDING) { |
|
return 0; |
|
} |
|
return (vlockmgr(vp->v_vnlock, LK_RELEASE)); |
|
} |
|
|
|
/* |
|
* Return whether or not the node is locked. |
|
*/ |
|
int |
|
ffs_islocked(void *v) |
|
{ |
|
struct vop_islocked_args /* { |
|
struct vnode *a_vp; |
|
} */ *ap = v; |
|
struct vnode *vp = ap->a_vp; |
|
|
|
return (vlockstatus(vp->v_vnlock)); |
|
} |