xref: /netbsd/sys/ufs/ffs/ffs_vnops.c (revision bf9ec67e)
1 /*	$NetBSD: ffs_vnops.c,v 1.49 2002/05/05 17:00:06 chs Exp $	*/
2 
3 /*
4  * Copyright (c) 1982, 1986, 1989, 1993
5  *	The Regents of the University of California.  All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions
9  * are met:
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  * 3. All advertising materials mentioning features or use of this software
16  *    must display the following acknowledgement:
17  *	This product includes software developed by the University of
18  *	California, Berkeley and its contributors.
19  * 4. Neither the name of the University nor the names of its contributors
20  *    may be used to endorse or promote products derived from this software
21  *    without specific prior written permission.
22  *
23  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
24  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
25  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
26  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
27  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
28  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
29  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
30  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
31  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
32  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
33  * SUCH DAMAGE.
34  *
35  *	@(#)ffs_vnops.c	8.15 (Berkeley) 5/14/95
36  */
37 
38 #include <sys/cdefs.h>
39 __KERNEL_RCSID(0, "$NetBSD: ffs_vnops.c,v 1.49 2002/05/05 17:00:06 chs Exp $");
40 
41 #include <sys/param.h>
42 #include <sys/systm.h>
43 #include <sys/resourcevar.h>
44 #include <sys/kernel.h>
45 #include <sys/file.h>
46 #include <sys/stat.h>
47 #include <sys/buf.h>
48 #include <sys/proc.h>
49 #include <sys/mount.h>
50 #include <sys/vnode.h>
51 #include <sys/pool.h>
52 #include <sys/signalvar.h>
53 
54 #include <miscfs/fifofs/fifo.h>
55 #include <miscfs/genfs/genfs.h>
56 #include <miscfs/specfs/specdev.h>
57 
58 #include <ufs/ufs/inode.h>
59 #include <ufs/ufs/dir.h>
60 #include <ufs/ufs/ufs_extern.h>
61 #include <ufs/ufs/ufsmount.h>
62 
63 #include <ufs/ffs/fs.h>
64 #include <ufs/ffs/ffs_extern.h>
65 
66 #include <uvm/uvm.h>
67 
68 static int ffs_full_fsync __P((void *));
69 
70 /* Global vfs data structures for ufs. */
71 int (**ffs_vnodeop_p) __P((void *));
72 const struct vnodeopv_entry_desc ffs_vnodeop_entries[] = {
73 	{ &vop_default_desc, vn_default_error },
74 	{ &vop_lookup_desc, ufs_lookup },		/* lookup */
75 	{ &vop_create_desc, ufs_create },		/* create */
76 	{ &vop_whiteout_desc, ufs_whiteout },		/* whiteout */
77 	{ &vop_mknod_desc, ufs_mknod },			/* mknod */
78 	{ &vop_open_desc, ufs_open },			/* open */
79 	{ &vop_close_desc, ufs_close },			/* close */
80 	{ &vop_access_desc, ufs_access },		/* access */
81 	{ &vop_getattr_desc, ufs_getattr },		/* getattr */
82 	{ &vop_setattr_desc, ufs_setattr },		/* setattr */
83 	{ &vop_read_desc, ffs_read },			/* read */
84 	{ &vop_write_desc, ffs_write },			/* write */
85 	{ &vop_lease_desc, ufs_lease_check },		/* lease */
86 	{ &vop_ioctl_desc, ufs_ioctl },			/* ioctl */
87 	{ &vop_fcntl_desc, ufs_fcntl },			/* fcntl */
88 	{ &vop_poll_desc, ufs_poll },			/* poll */
89 	{ &vop_revoke_desc, ufs_revoke },		/* revoke */
90 	{ &vop_mmap_desc, ufs_mmap },			/* mmap */
91 	{ &vop_fsync_desc, ffs_fsync },			/* fsync */
92 	{ &vop_seek_desc, ufs_seek },			/* seek */
93 	{ &vop_remove_desc, ufs_remove },		/* remove */
94 	{ &vop_link_desc, ufs_link },			/* link */
95 	{ &vop_rename_desc, ufs_rename },		/* rename */
96 	{ &vop_mkdir_desc, ufs_mkdir },			/* mkdir */
97 	{ &vop_rmdir_desc, ufs_rmdir },			/* rmdir */
98 	{ &vop_symlink_desc, ufs_symlink },		/* symlink */
99 	{ &vop_readdir_desc, ufs_readdir },		/* readdir */
100 	{ &vop_readlink_desc, ufs_readlink },		/* readlink */
101 	{ &vop_abortop_desc, ufs_abortop },		/* abortop */
102 	{ &vop_inactive_desc, ufs_inactive },		/* inactive */
103 	{ &vop_reclaim_desc, ffs_reclaim },		/* reclaim */
104 	{ &vop_lock_desc, ufs_lock },			/* lock */
105 	{ &vop_unlock_desc, ufs_unlock },		/* unlock */
106 	{ &vop_bmap_desc, ufs_bmap },			/* bmap */
107 	{ &vop_strategy_desc, ufs_strategy },		/* strategy */
108 	{ &vop_print_desc, ufs_print },			/* print */
109 	{ &vop_islocked_desc, ufs_islocked },		/* islocked */
110 	{ &vop_pathconf_desc, ufs_pathconf },		/* pathconf */
111 	{ &vop_advlock_desc, ufs_advlock },		/* advlock */
112 	{ &vop_blkatoff_desc, ffs_blkatoff },		/* blkatoff */
113 	{ &vop_valloc_desc, ffs_valloc },		/* valloc */
114 	{ &vop_balloc_desc, ffs_balloc },		/* balloc */
115 	{ &vop_reallocblks_desc, ffs_reallocblks },	/* reallocblks */
116 	{ &vop_vfree_desc, ffs_vfree },			/* vfree */
117 	{ &vop_truncate_desc, ffs_truncate },		/* truncate */
118 	{ &vop_update_desc, ffs_update },		/* update */
119 	{ &vop_bwrite_desc, vn_bwrite },		/* bwrite */
120 	{ &vop_getpages_desc, ffs_getpages },		/* getpages */
121 	{ &vop_putpages_desc, ffs_putpages },		/* putpages */
122 	{ NULL, NULL }
123 };
124 const struct vnodeopv_desc ffs_vnodeop_opv_desc =
125 	{ &ffs_vnodeop_p, ffs_vnodeop_entries };
126 
127 int (**ffs_specop_p) __P((void *));
128 const struct vnodeopv_entry_desc ffs_specop_entries[] = {
129 	{ &vop_default_desc, vn_default_error },
130 	{ &vop_lookup_desc, spec_lookup },		/* lookup */
131 	{ &vop_create_desc, spec_create },		/* create */
132 	{ &vop_mknod_desc, spec_mknod },		/* mknod */
133 	{ &vop_open_desc, spec_open },			/* open */
134 	{ &vop_close_desc, ufsspec_close },		/* close */
135 	{ &vop_access_desc, ufs_access },		/* access */
136 	{ &vop_getattr_desc, ufs_getattr },		/* getattr */
137 	{ &vop_setattr_desc, ufs_setattr },		/* setattr */
138 	{ &vop_read_desc, ufsspec_read },		/* read */
139 	{ &vop_write_desc, ufsspec_write },		/* write */
140 	{ &vop_lease_desc, spec_lease_check },		/* lease */
141 	{ &vop_ioctl_desc, spec_ioctl },		/* ioctl */
142 	{ &vop_fcntl_desc, ufs_fcntl },			/* fcntl */
143 	{ &vop_poll_desc, spec_poll },			/* poll */
144 	{ &vop_revoke_desc, spec_revoke },		/* revoke */
145 	{ &vop_mmap_desc, spec_mmap },			/* mmap */
146 	{ &vop_fsync_desc, ffs_fsync },			/* fsync */
147 	{ &vop_seek_desc, spec_seek },			/* seek */
148 	{ &vop_remove_desc, spec_remove },		/* remove */
149 	{ &vop_link_desc, spec_link },			/* link */
150 	{ &vop_rename_desc, spec_rename },		/* rename */
151 	{ &vop_mkdir_desc, spec_mkdir },		/* mkdir */
152 	{ &vop_rmdir_desc, spec_rmdir },		/* rmdir */
153 	{ &vop_symlink_desc, spec_symlink },		/* symlink */
154 	{ &vop_readdir_desc, spec_readdir },		/* readdir */
155 	{ &vop_readlink_desc, spec_readlink },		/* readlink */
156 	{ &vop_abortop_desc, spec_abortop },		/* abortop */
157 	{ &vop_inactive_desc, ufs_inactive },		/* inactive */
158 	{ &vop_reclaim_desc, ffs_reclaim },		/* reclaim */
159 	{ &vop_lock_desc, ufs_lock },			/* lock */
160 	{ &vop_unlock_desc, ufs_unlock },		/* unlock */
161 	{ &vop_bmap_desc, spec_bmap },			/* bmap */
162 	{ &vop_strategy_desc, spec_strategy },		/* strategy */
163 	{ &vop_print_desc, ufs_print },			/* print */
164 	{ &vop_islocked_desc, ufs_islocked },		/* islocked */
165 	{ &vop_pathconf_desc, spec_pathconf },		/* pathconf */
166 	{ &vop_advlock_desc, spec_advlock },		/* advlock */
167 	{ &vop_blkatoff_desc, spec_blkatoff },		/* blkatoff */
168 	{ &vop_valloc_desc, spec_valloc },		/* valloc */
169 	{ &vop_reallocblks_desc, spec_reallocblks },	/* reallocblks */
170 	{ &vop_vfree_desc, ffs_vfree },			/* vfree */
171 	{ &vop_truncate_desc, spec_truncate },		/* truncate */
172 	{ &vop_update_desc, ffs_update },		/* update */
173 	{ &vop_bwrite_desc, vn_bwrite },		/* bwrite */
174 	{ &vop_getpages_desc, spec_getpages },		/* getpages */
175 	{ &vop_putpages_desc, spec_putpages },		/* putpages */
176 	{ NULL, NULL }
177 };
178 const struct vnodeopv_desc ffs_specop_opv_desc =
179 	{ &ffs_specop_p, ffs_specop_entries };
180 
181 int (**ffs_fifoop_p) __P((void *));
182 const struct vnodeopv_entry_desc ffs_fifoop_entries[] = {
183 	{ &vop_default_desc, vn_default_error },
184 	{ &vop_lookup_desc, fifo_lookup },		/* lookup */
185 	{ &vop_create_desc, fifo_create },		/* create */
186 	{ &vop_mknod_desc, fifo_mknod },		/* mknod */
187 	{ &vop_open_desc, fifo_open },			/* open */
188 	{ &vop_close_desc, ufsfifo_close },		/* close */
189 	{ &vop_access_desc, ufs_access },		/* access */
190 	{ &vop_getattr_desc, ufs_getattr },		/* getattr */
191 	{ &vop_setattr_desc, ufs_setattr },		/* setattr */
192 	{ &vop_read_desc, ufsfifo_read },		/* read */
193 	{ &vop_write_desc, ufsfifo_write },		/* write */
194 	{ &vop_lease_desc, fifo_lease_check },		/* lease */
195 	{ &vop_ioctl_desc, fifo_ioctl },		/* ioctl */
196 	{ &vop_fcntl_desc, ufs_fcntl },			/* fcntl */
197 	{ &vop_poll_desc, fifo_poll },			/* poll */
198 	{ &vop_revoke_desc, fifo_revoke },		/* revoke */
199 	{ &vop_mmap_desc, fifo_mmap },			/* mmap */
200 	{ &vop_fsync_desc, ffs_fsync },			/* fsync */
201 	{ &vop_seek_desc, fifo_seek },			/* seek */
202 	{ &vop_remove_desc, fifo_remove },		/* remove */
203 	{ &vop_link_desc, fifo_link },			/* link */
204 	{ &vop_rename_desc, fifo_rename },		/* rename */
205 	{ &vop_mkdir_desc, fifo_mkdir },		/* mkdir */
206 	{ &vop_rmdir_desc, fifo_rmdir },		/* rmdir */
207 	{ &vop_symlink_desc, fifo_symlink },		/* symlink */
208 	{ &vop_readdir_desc, fifo_readdir },		/* readdir */
209 	{ &vop_readlink_desc, fifo_readlink },		/* readlink */
210 	{ &vop_abortop_desc, fifo_abortop },		/* abortop */
211 	{ &vop_inactive_desc, ufs_inactive },		/* inactive */
212 	{ &vop_reclaim_desc, ffs_reclaim },		/* reclaim */
213 	{ &vop_lock_desc, ufs_lock },			/* lock */
214 	{ &vop_unlock_desc, ufs_unlock },		/* unlock */
215 	{ &vop_bmap_desc, fifo_bmap },			/* bmap */
216 	{ &vop_strategy_desc, fifo_strategy },		/* strategy */
217 	{ &vop_print_desc, ufs_print },			/* print */
218 	{ &vop_islocked_desc, ufs_islocked },		/* islocked */
219 	{ &vop_pathconf_desc, fifo_pathconf },		/* pathconf */
220 	{ &vop_advlock_desc, fifo_advlock },		/* advlock */
221 	{ &vop_blkatoff_desc, fifo_blkatoff },		/* blkatoff */
222 	{ &vop_valloc_desc, fifo_valloc },		/* valloc */
223 	{ &vop_reallocblks_desc, fifo_reallocblks },	/* reallocblks */
224 	{ &vop_vfree_desc, ffs_vfree },			/* vfree */
225 	{ &vop_truncate_desc, fifo_truncate },		/* truncate */
226 	{ &vop_update_desc, ffs_update },		/* update */
227 	{ &vop_bwrite_desc, vn_bwrite },		/* bwrite */
228 	{ &vop_putpages_desc, fifo_putpages }, 		/* putpages */
229 	{ NULL, NULL }
230 };
231 const struct vnodeopv_desc ffs_fifoop_opv_desc =
232 	{ &ffs_fifoop_p, ffs_fifoop_entries };
233 
234 int doclusterread = 1;
235 int doclusterwrite = 1;
236 
237 #include <ufs/ufs/ufs_readwrite.c>
238 
239 int
240 ffs_fsync(v)
241 	void *v;
242 {
243 	struct vop_fsync_args /* {
244 		struct vnode *a_vp;
245 		struct ucred *a_cred;
246 		int a_flags;
247 		off_t a_offlo;
248 		off_t a_offhi;
249 		struct proc *a_p;
250 	} */ *ap = v;
251 	struct buf *bp;
252 	int s, num, error, i;
253 	struct indir ia[NIADDR + 1];
254 	int bsize;
255 	daddr_t blk_high;
256 	struct vnode *vp;
257 
258 	/*
259 	 * XXX no easy way to sync a range in a file with softdep.
260 	 */
261 	if ((ap->a_offlo == 0 && ap->a_offhi == 0) || DOINGSOFTDEP(ap->a_vp))
262 		return ffs_full_fsync(v);
263 
264 	vp = ap->a_vp;
265 
266 	bsize = ap->a_vp->v_mount->mnt_stat.f_iosize;
267 	blk_high = ap->a_offhi / bsize;
268 	if (ap->a_offhi % bsize != 0)
269 		blk_high++;
270 
271 	/*
272 	 * First, flush all pages in range.
273 	 */
274 
275 	if (vp->v_type == VREG) {
276 		simple_lock(&vp->v_interlock);
277 		error = VOP_PUTPAGES(vp, trunc_page(ap->a_offlo),
278 		    round_page(ap->a_offhi), PGO_CLEANIT|PGO_SYNCIO);
279 		if (error) {
280 			return error;
281 		}
282 	}
283 
284 	/*
285 	 * Then, flush indirect blocks.
286 	 */
287 
288 	s = splbio();
289 	if (!(ap->a_flags & FSYNC_DATAONLY) && blk_high >= NDADDR) {
290 		error = ufs_getlbns(vp, blk_high, ia, &num);
291 		if (error) {
292 			splx(s);
293 			return error;
294 		}
295 		for (i = 0; i < num; i++) {
296 			bp = incore(vp, ia[i].in_lbn);
297 			if (bp != NULL && !(bp->b_flags & B_BUSY) &&
298 			    (bp->b_flags & B_DELWRI)) {
299 				bp->b_flags |= B_BUSY | B_VFLUSH;
300 				splx(s);
301 				bawrite(bp);
302 				s = splbio();
303 			}
304 		}
305 	}
306 
307 	if (ap->a_flags & FSYNC_WAIT) {
308 		while (vp->v_numoutput > 0) {
309 			vp->v_flag |= VBWAIT;
310 			tsleep(&vp->v_numoutput, PRIBIO + 1, "fsync_range", 0);
311 		}
312 	}
313 	splx(s);
314 
315 	return (VOP_UPDATE(vp, NULL, NULL,
316 	    (ap->a_flags & FSYNC_WAIT) ? UPDATE_WAIT : 0));
317 }
318 
319 /*
320  * Synch an open file.
321  */
322 /* ARGSUSED */
323 static int
324 ffs_full_fsync(v)
325 	void *v;
326 {
327 	struct vop_fsync_args /* {
328 		struct vnode *a_vp;
329 		struct ucred *a_cred;
330 		int a_flags;
331 		off_t a_offlo;
332 		off_t a_offhi;
333 		struct proc *a_p;
334 	} */ *ap = v;
335 	struct vnode *vp = ap->a_vp;
336 	struct buf *bp, *nbp;
337 	int s, error, passes, skipmeta, inodedeps_only, waitfor;
338 
339 	if (vp->v_type == VBLK &&
340 	    vp->v_specmountpoint != NULL &&
341 	    (vp->v_specmountpoint->mnt_flag & MNT_SOFTDEP))
342 		softdep_fsync_mountdev(vp);
343 
344 	inodedeps_only = DOINGSOFTDEP(vp) && (ap->a_flags & FSYNC_RECLAIM)
345 	    && vp->v_uobj.uo_npages == 0 && LIST_EMPTY(&vp->v_dirtyblkhd);
346 
347 	/*
348 	 * Flush all dirty data associated with a vnode.
349 	 */
350 
351 	if (vp->v_type == VREG) {
352 		simple_lock(&vp->v_interlock);
353 		error = VOP_PUTPAGES(vp, 0, 0, PGO_ALLPAGES | PGO_CLEANIT |
354 		    ((ap->a_flags & FSYNC_WAIT) ? PGO_SYNCIO : 0));
355 		if (error) {
356 			return error;
357 		}
358 	}
359 
360 	passes = NIADDR + 1;
361 	skipmeta = 0;
362 	if (ap->a_flags & (FSYNC_DATAONLY|FSYNC_WAIT))
363 		skipmeta = 1;
364 	s = splbio();
365 
366 loop:
367 	LIST_FOREACH(bp, &vp->v_dirtyblkhd, b_vnbufs)
368 		bp->b_flags &= ~B_SCANNED;
369 	for (bp = LIST_FIRST(&vp->v_dirtyblkhd); bp; bp = nbp) {
370 		nbp = LIST_NEXT(bp, b_vnbufs);
371 		if (bp->b_flags & (B_BUSY | B_SCANNED))
372 			continue;
373 		if ((bp->b_flags & B_DELWRI) == 0)
374 			panic("ffs_fsync: not dirty");
375 		if (skipmeta && bp->b_lblkno < 0)
376 			continue;
377 		bp->b_flags |= B_BUSY | B_VFLUSH | B_SCANNED;
378 		splx(s);
379 		/*
380 		 * On our final pass through, do all I/O synchronously
381 		 * so that we can find out if our flush is failing
382 		 * because of write errors.
383 		 */
384 		if (passes > 0 || !(ap->a_flags & FSYNC_WAIT))
385 			(void) bawrite(bp);
386 		else if ((error = bwrite(bp)) != 0)
387 			return (error);
388 		s = splbio();
389 		/*
390 		 * Since we may have slept during the I/O, we need
391 		 * to start from a known point.
392 		 */
393 		nbp = LIST_FIRST(&vp->v_dirtyblkhd);
394 	}
395 	if (skipmeta && !(ap->a_flags & FSYNC_DATAONLY)) {
396 		skipmeta = 0;
397 		goto loop;
398 	}
399 	if (ap->a_flags & FSYNC_WAIT) {
400 		while (vp->v_numoutput) {
401 			vp->v_flag |= VBWAIT;
402 			(void) tsleep(&vp->v_numoutput, PRIBIO + 1,
403 			    "ffsfsync", 0);
404 		}
405 		splx(s);
406 
407 		if (ap->a_flags & FSYNC_DATAONLY)
408 			return (0);
409 
410 		/*
411 		 * Ensure that any filesystem metadata associated
412 		 * with the vnode has been written.
413 		 */
414 		if ((error = softdep_sync_metadata(ap)) != 0)
415 			return (error);
416 
417 		s = splbio();
418 		if (!LIST_EMPTY(&vp->v_dirtyblkhd)) {
419 			/*
420 			* Block devices associated with filesystems may
421 			* have new I/O requests posted for them even if
422 			* the vnode is locked, so no amount of trying will
423 			* get them clean. Thus we give block devices a
424 			* good effort, then just give up. For all other file
425 			* types, go around and try again until it is clean.
426 			*/
427 			if (passes > 0) {
428 				passes--;
429 				goto loop;
430 			}
431 #ifdef DIAGNOSTIC
432 			if (vp->v_type != VBLK)
433 				vprint("ffs_fsync: dirty", vp);
434 #endif
435 		}
436 	}
437 	splx(s);
438 
439 	if (inodedeps_only)
440 		waitfor = 0;
441 	else
442 		waitfor = (ap->a_flags & FSYNC_WAIT) ? UPDATE_WAIT : 0;
443 	return (VOP_UPDATE(vp, NULL, NULL, waitfor));
444 }
445 
446 /*
447  * Reclaim an inode so that it can be used for other purposes.
448  */
449 int
450 ffs_reclaim(v)
451 	void *v;
452 {
453 	struct vop_reclaim_args /* {
454 		struct vnode *a_vp;
455 		struct proc *a_p;
456 	} */ *ap = v;
457 	struct vnode *vp = ap->a_vp;
458 	int error;
459 
460 	if ((error = ufs_reclaim(vp, ap->a_p)) != 0)
461 		return (error);
462 	/*
463 	 * XXX MFS ends up here, too, to free an inode.  Should we create
464 	 * XXX a separate pool for MFS inodes?
465 	 */
466 	pool_put(&ffs_inode_pool, vp->v_data);
467 	vp->v_data = NULL;
468 	return (0);
469 }
470 
471 int
472 ffs_getpages(void *v)
473 {
474 	struct vop_getpages_args /* {
475 		struct vnode *a_vp;
476 		voff_t a_offset;
477 		struct vm_page **a_m;
478 		int *a_count;
479 		int a_centeridx;
480 		vm_prot_t a_access_type;
481 		int a_advice;
482 		int a_flags;
483 	} */ *ap = v;
484 	struct vnode *vp = ap->a_vp;
485 	struct inode *ip = VTOI(vp);
486 	struct fs *fs = ip->i_fs;
487 
488 	/*
489 	 * don't allow a softdep write to create pages for only part of a block.
490 	 * the dependency tracking requires that all pages be in memory for
491 	 * a block involved in a dependency.
492 	 */
493 
494 	if (ap->a_flags & PGO_OVERWRITE &&
495 	    (blkoff(fs, ap->a_offset) != 0 ||
496 	     blkoff(fs, *ap->a_count << PAGE_SHIFT) != 0) &&
497 	    DOINGSOFTDEP(ap->a_vp)) {
498 		if ((ap->a_flags & PGO_LOCKED) == 0) {
499 			simple_unlock(&vp->v_interlock);
500 		}
501 		return EINVAL;
502 	}
503 	return genfs_getpages(v);
504 }
505 
506 int
507 ffs_putpages(void *v)
508 {
509 	struct vop_putpages_args /* {
510 		struct vnode *a_vp;
511 		voff_t a_offlo;
512 		voff_t a_offhi;
513 		int a_flags;
514 	} */ *ap = v;
515 	struct vnode *vp = ap->a_vp;
516 	struct uvm_object *uobj = &vp->v_uobj;
517 	struct inode *ip = VTOI(vp);
518 	struct fs *fs = ip->i_fs;
519 	struct vm_page *pg;
520 	off_t off;
521 	ufs_lbn_t lbn;
522 
523 	if (!DOINGSOFTDEP(vp) || (ap->a_flags & PGO_CLEANIT) == 0) {
524 		return genfs_putpages(v);
525 	}
526 
527 	/*
528 	 * for softdep files, force the pages in a block to be written together.
529 	 * if we're the pagedaemon and we would have to wait for other pages,
530 	 * just fail the request.  the pagedaemon will pick a different page.
531 	 */
532 
533 	ap->a_offlo &= ~fs->fs_qbmask;
534 	lbn = lblkno(fs, ap->a_offhi);
535 	ap->a_offhi = blkroundup(fs, ap->a_offhi);
536 	if (curproc == uvm.pagedaemon_proc) {
537 		for (off = ap->a_offlo; off < ap->a_offhi; off += PAGE_SIZE) {
538 			pg = uvm_pagelookup(uobj, off);
539 
540 			/*
541 			 * we only have missing pages here because the
542 			 * calculation of offhi above doesn't account for
543 			 * fragments.  so once we see one missing page,
544 			 * the rest should be missing as well, but we'll
545 			 * check for the rest just to be paranoid.
546 			 */
547 
548 			if (pg == NULL) {
549 				continue;
550 			}
551 			if (pg->flags & PG_BUSY) {
552 				simple_unlock(&uobj->vmobjlock);
553 				return EBUSY;
554 			}
555 		}
556 	}
557 	return genfs_putpages(v);
558 }
559 
560 /*
561  * Return the last logical file offset that should be written for this file
562  * if we're doing a write that ends at "size".
563  */
564 
565 void
566 ffs_gop_size(struct vnode *vp, off_t size, off_t *eobp)
567 {
568 	struct inode *ip = VTOI(vp);
569 	struct fs *fs = ip->i_fs;
570 	ufs_lbn_t olbn, nlbn;
571 
572 	olbn = lblkno(fs, ip->i_ffs_size);
573 	nlbn = lblkno(fs, size);
574 	if (nlbn < NDADDR && olbn <= nlbn) {
575 		*eobp = fragroundup(fs, size);
576 	} else {
577 		*eobp = blkroundup(fs, size);
578 	}
579 }
580