1 /* 2 * Copyright (c) 1989, 1993 3 * The Regents of the University of California. All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 3. All advertising materials mentioning features or use of this software 14 * must display the following acknowledgement: 15 * This product includes software developed by the University of 16 * California, Berkeley and its contributors. 17 * 4. Neither the name of the University nor the names of its contributors 18 * may be used to endorse or promote products derived from this software 19 * without specific prior written permission. 20 * 21 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 22 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 23 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 24 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 25 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 26 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 27 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 28 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 29 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 30 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 31 * SUCH DAMAGE. 32 * 33 * @(#)vnode.h 8.7 (Berkeley) 2/4/94 34 * $FreeBSD: src/sys/sys/vnode.h,v 1.111.2.19 2002/12/29 18:19:53 dillon Exp $ 35 * $DragonFly: src/sys/sys/vnode.h,v 1.40 2006/02/17 19:18:07 dillon Exp $ 36 */ 37 38 #ifndef _SYS_VNODE_H_ 39 #define _SYS_VNODE_H_ 40 41 #include <sys/queue.h> 42 #include <sys/lock.h> 43 #include <sys/select.h> 44 #include <sys/biotrack.h> 45 #include <sys/uio.h> 46 #include <sys/acl.h> 47 #include <sys/namecache.h> 48 #if defined(_KERNEL) || defined(_KERNEL_STRUCTURES) 49 #include <sys/thread.h> 50 #endif 51 #include <sys/vfsops.h> 52 #include <sys/vfscache.h> 53 #include <sys/tree.h> 54 55 #include <machine/lock.h> 56 57 /* 58 * The vnode is the focus of all file activity in UNIX. There is a 59 * unique vnode allocated for each active file, each current directory, 60 * each mounted-on file, text file, and the root. 61 */ 62 63 /* 64 * Each underlying filesystem allocates its own private area and hangs 65 * it from v_data. If non-null, this area is freed in getnewvnode(). 66 */ 67 TAILQ_HEAD(buflists, buf); 68 69 /* 70 * Range locks protect offset ranges in files and directories at a high 71 * level, allowing the actual I/O to be broken down into smaller pieces. 72 * Range locks will eventually be integrated into the clustered cache 73 * coherency infrastructure. 74 * 75 * We use a simple data structure for now, but eventually this should 76 * probably be a btree or red-black tree. 77 */ 78 struct vrangelock; 79 80 TAILQ_HEAD(vrangelock_list, vrangelock); 81 82 struct vrangehead { 83 struct vrangelock_list vh_list; 84 }; 85 86 struct vrangelock { 87 TAILQ_ENTRY(vrangelock) vr_node; 88 int vr_flags; 89 off_t vr_offset; 90 off_t vr_length; 91 }; 92 93 #define RNGL_WAITING 0x0001 /* waiting for lock, else has lock */ 94 #define RNGL_CHECK 0x0002 /* check for work on unlock */ 95 #define RNGL_SHARED 0x0004 /* shared lock, else exclusive */ 96 #define RNGL_ONLIST 0x0008 /* sanity check */ 97 98 static __inline 99 void 100 vrange_init(struct vrangelock *vr, int flags, off_t offset, off_t length) 101 { 102 vr->vr_flags = flags; 103 vr->vr_offset = offset; 104 vr->vr_length = length; 105 } 106 107 #ifdef _KERNEL 108 109 void vrange_lock(struct vnode *vp, struct vrangelock *vr); 110 void vrange_unlock(struct vnode *vp, struct vrangelock *vr); 111 112 static __inline 113 void 114 vrange_lock_shared(struct vnode *vp, struct vrangelock *vr, 115 off_t offset, off_t length) 116 { 117 vrange_init(vr, RNGL_SHARED, offset, length); 118 vrange_lock(vp, vr); 119 } 120 121 static __inline 122 void 123 vrange_lock_excl(struct vnode *vp, struct vrangelock *vr, 124 off_t offset, off_t length) 125 { 126 vrange_init(vr, 0, offset, length); 127 vrange_lock(vp, vr); 128 } 129 130 #endif 131 132 /* 133 * The vnode infrastructure is being reorgranized. Most reference-related 134 * fields are locked by the BGL, and most file I/O related operations and 135 * vnode teardown functions are locked by the vnode lock. 136 * 137 * File read operations require a shared lock, file write operations require 138 * an exclusive lock. Most directory operations (read or write) currently 139 * require an exclusive lock due to the side effects stored in the directory 140 * inode (which we intend to fix). 141 * 142 * File reads and writes are further protected by a range lock. The intention 143 * is to be able to break I/O operations down into more easily managed pieces 144 * so vm_page arrays can be passed through rather then UIOs. This work will 145 * occur in multiple stages. The range locks will also eventually be used to 146 * deal with clustered cache coherency issues and, more immediately, to 147 * protect operations associated with the kernel-managed journaling module. 148 * 149 * NOTE: XXX v_opencount currently only used by specfs. It should be used 150 * universally. 151 * 152 * NOTE: The vnode operations vector, v_ops, is a double-indirect that 153 * typically points to &v_mount->mnt_vn_use_ops. We use a double 154 * pointer because mnt_vn_use_ops may change dynamically when e.g. 155 * journaling is turned on or off. 156 */ 157 RB_HEAD(buf_rb_tree, buf); 158 159 struct vnode { 160 u_long v_flag; /* vnode flags (see below) */ 161 int v_usecount; /* reference count of users */ 162 int v_writecount; 163 int v_holdcnt; /* page & buffer references */ 164 int v_opencount; /* number of explicit opens */ 165 struct bio_track v_track_read; /* track I/O's in progress */ 166 struct bio_track v_track_write; /* track I/O's in progress */ 167 u_long v_id; /* capability identifier */ 168 struct mount *v_mount; /* ptr to vfs we are in */ 169 struct vop_ops **v_ops; /* vnode operations vector */ 170 TAILQ_ENTRY(vnode) v_freelist; /* vnode freelist */ 171 TAILQ_ENTRY(vnode) v_nmntvnodes; /* vnodes for mount point */ 172 struct buf_rb_tree v_rbclean_tree; /* RB tree of clean bufs */ 173 struct buf_rb_tree v_rbdirty_tree; /* RB tree of dirty bufs */ 174 LIST_ENTRY(vnode) v_synclist; /* vnodes with dirty buffers */ 175 enum vtype v_type; /* vnode type */ 176 union { 177 struct mount *vu_mountedhere;/* ptr to mounted vfs (VDIR) */ 178 struct socket *vu_socket; /* unix ipc (VSOCK) */ 179 struct { 180 udev_t vu_udev; /* device number for attach */ 181 struct specinfo *vu_specinfo; /* device (VCHR, VBLK) */ 182 SLIST_ENTRY(vnode) vu_specnext; 183 } vu_spec; 184 struct fifoinfo *vu_fifoinfo; /* fifo (VFIFO) */ 185 } v_un; 186 struct nqlease *v_lease; /* Soft reference to lease */ 187 daddr_t v_lazyw; /* lazy write iterator */ 188 daddr_t v_lastw; /* last write (write cluster) */ 189 daddr_t v_cstart; /* start block of cluster */ 190 daddr_t v_lasta; /* last allocation */ 191 int v_clen; /* length of current cluster */ 192 struct vm_object *v_object; /* Place to store VM object */ 193 struct lock v_lock; /* file/dir ops lock */ 194 enum vtagtype v_tag; /* type of underlying data */ 195 void *v_data; /* private data for fs */ 196 struct namecache_list v_namecache; /* associated nc entries */ 197 struct { 198 struct lwkt_token vpi_token; /* lock to protect below */ 199 struct selinfo vpi_selinfo; /* identity of poller(s) */ 200 short vpi_events; /* what they are looking for */ 201 short vpi_revents; /* what has happened */ 202 } v_pollinfo; 203 struct vmresident *v_resident; /* optional vmresident */ 204 struct vrangehead v_range; /* range lock */ 205 #ifdef DEBUG_LOCKS 206 const char *filename; /* Source file doing locking */ 207 int line; /* Line number doing locking */ 208 #endif 209 void *v_xaddr; 210 }; 211 #define v_mountedhere v_un.vu_mountedhere 212 #define v_socket v_un.vu_socket 213 #define v_udev v_un.vu_spec.vu_udev 214 #define v_rdev v_un.vu_spec.vu_specinfo 215 #define v_specnext v_un.vu_spec.vu_specnext 216 #define v_fifoinfo v_un.vu_fifoinfo 217 218 #define VN_POLLEVENT(vp, events) \ 219 do { \ 220 if ((vp)->v_pollinfo.vpi_events & (events)) \ 221 vn_pollevent((vp), (events)); \ 222 } while (0) 223 224 /* 225 * Vnode flags. 226 */ 227 #define VROOT 0x00001 /* root of its file system */ 228 #define VTEXT 0x00002 /* vnode is a pure text prototype */ 229 #define VSYSTEM 0x00004 /* vnode being used by kernel */ 230 #define VISTTY 0x00008 /* vnode represents a tty */ 231 #define VCTTYISOPEN 0x00010 /* controlling terminal tty is open */ 232 #define VCKPT 0x00020 /* checkpoint-restored vnode */ 233 /* open for business 0x00040 */ 234 /* open for business 0x00080 */ 235 /* open for business 0x00100 */ 236 /* open for business 0x00200 */ 237 /* open for business 0x00400 */ 238 /* open for business 0x00800 */ 239 /* open for business 0x01000 */ 240 #define VOBJBUF 0x02000 /* Allocate buffers in VM object */ 241 #define VINACTIVE 0x04000 /* The vnode is inactive */ 242 #define VAGE 0x08000 /* Insert vnode at head of free list */ 243 #define VOLOCK 0x10000 /* vnode is locked waiting for an object */ 244 #define VOWANT 0x20000 /* a process is waiting for VOLOCK */ 245 #define VRECLAIMED 0x40000 /* This vnode has been destroyed */ 246 #define VFREE 0x80000 /* This vnode is on the freelist */ 247 /* open for business 0x100000 */ 248 #define VONWORKLST 0x200000 /* On syncer work-list */ 249 #define VMOUNT 0x400000 /* Mount in progress */ 250 #define VOBJDIRTY 0x800000 /* object might be dirty */ 251 252 /* 253 * vmntvnodescan() flags 254 */ 255 #define VMSC_GETVP 1 256 #define VMSC_GETVX 2 257 #define VMSC_REFVP 3 258 #define VMSC_NOWAIT 0x10 259 260 /* 261 * Flags for ioflag. (high 16 bits used to ask for read-ahead and 262 * help with write clustering) 263 */ 264 #define IO_UNIT 0x0001 /* do I/O as atomic unit */ 265 #define IO_APPEND 0x0002 /* append write to end */ 266 #define IO_SYNC 0x0004 /* do I/O synchronously */ 267 #define IO_NODELOCKED 0x0008 /* underlying node already locked */ 268 #define IO_NDELAY 0x0010 /* FNDELAY flag set in file table */ 269 #define IO_VMIO 0x0020 /* data already in VMIO space */ 270 #define IO_INVAL 0x0040 /* invalidate after I/O */ 271 #define IO_ASYNC 0x0080 /* bawrite rather then bdwrite */ 272 #define IO_DIRECT 0x0100 /* attempt to bypass buffer cache */ 273 #define IO_NOWDRAIN 0x0200 /* do not block on wdrain */ 274 #define IO_CORE 0x0400 /* I/O is part of core dump */ 275 276 #define IO_SEQMAX 0x7F /* seq heuristic max value */ 277 #define IO_SEQSHIFT 16 /* seq heuristic in upper 16 bits */ 278 279 /* 280 * Modes. Note that these V-modes must match file S_I*USR, SUID, SGID, 281 * and SVTX flag bits. 282 * 283 * VCREATE, VDELETE, and VEXCL may only be used in naccess() calls. 284 */ 285 #define VDELETE 040000 /* delete if the file/dir exists */ 286 #define VCREATE 020000 /* create if the file/dir does not exist */ 287 #define VEXCL 010000 /* error if the file/dir already exists */ 288 289 #define VSUID 04000 /* set user id on execution */ 290 #define VSGID 02000 /* set group id on execution */ 291 #define VSVTX 01000 /* save swapped text even after use */ 292 #define VREAD 00400 /* read, write, execute permissions */ 293 #define VWRITE 00200 294 #define VEXEC 00100 295 296 /* 297 * Token indicating no attribute value yet assigned. 298 */ 299 #define VNOVAL (-1) 300 301 /* 302 * LK_TIMELOCK timeout for vnode locks (used mainly by the pageout daemon) 303 */ 304 #define VLKTIMEOUT (hz / 20 + 1) 305 306 #ifdef _KERNEL 307 308 /* 309 * Convert between vnode types and inode formats (since POSIX.1 310 * defines mode word of stat structure in terms of inode formats). 311 */ 312 extern enum vtype iftovt_tab[]; 313 extern int vttoif_tab[]; 314 #define IFTOVT(mode) (iftovt_tab[((mode) & S_IFMT) >> 12]) 315 #define VTTOIF(indx) (vttoif_tab[(int)(indx)]) 316 #define MAKEIMODE(indx, mode) (int)(VTTOIF(indx) | (mode)) 317 318 /* 319 * Flags to various vnode functions. 320 */ 321 #define SKIPSYSTEM 0x0001 /* vflush: skip vnodes marked VSYSTEM */ 322 #define FORCECLOSE 0x0002 /* vflush: force file closure */ 323 #define WRITECLOSE 0x0004 /* vflush: only close writable files */ 324 #define DOCLOSE 0x0008 /* vclean: close active files */ 325 #define V_SAVE 0x0001 /* vinvalbuf: sync file first */ 326 #define REVOKEALL 0x0001 /* vop_revoke: revoke all aliases */ 327 328 #ifdef DIAGNOSTIC 329 #define VATTR_NULL(vap) vattr_null(vap) 330 #else 331 #define VATTR_NULL(vap) (*(vap) = va_null) /* initialize a vattr */ 332 #endif /* DIAGNOSTIC */ 333 334 #define NULLVP ((struct vnode *)NULL) 335 336 #define VNODEOP_SET(f) \ 337 C_SYSINIT(f##init, SI_SUB_VFS, SI_ORDER_SECOND, vfs_add_vnodeops_sysinit, &f); \ 338 C_SYSUNINIT(f##uninit, SI_SUB_VFS, SI_ORDER_SECOND,vfs_rm_vnodeops_sysinit, &f); 339 340 /* 341 * Global vnode data. 342 */ 343 extern struct vnode *rootvnode; /* root (i.e. "/") vnode */ 344 extern struct namecache *rootncp; /* root (i.e. "/") namecache */ 345 extern int desiredvnodes; /* number of vnodes desired */ 346 extern time_t syncdelay; /* max time to delay syncing data */ 347 extern time_t filedelay; /* time to delay syncing files */ 348 extern time_t dirdelay; /* time to delay syncing directories */ 349 extern time_t metadelay; /* time to delay syncing metadata */ 350 extern struct vm_zone *namei_zone; 351 extern int prtactive; /* nonzero to call vprint() */ 352 extern struct vattr va_null; /* predefined null vattr structure */ 353 extern int vfs_ioopt; 354 extern int numvnodes; 355 extern int freevnodes; 356 extern int vfs_fastdev; /* fast specfs device access */ 357 358 /* 359 * Macro/function to check for client cache inconsistency w.r.t. leasing. 360 */ 361 #define LEASE_READ 0x1 /* Check lease for readers */ 362 #define LEASE_WRITE 0x2 /* Check lease for modifiers */ 363 364 365 extern void (*lease_updatetime) (int deltat); 366 367 #endif /* _KERNEL */ 368 369 /* 370 * Mods for extensibility. 371 */ 372 373 /* 374 * Flags for vdesc_flags: 375 */ 376 #define VDESC_MAX_VPS 8 377 /* Low order 8 flag bits are reserved for willrele flags for vp arguments. */ 378 #define VDESC_VP0_WILLRELE 0x00000001 379 #define VDESC_VP1_WILLRELE 0x00000002 380 #define VDESC_VP2_WILLRELE 0x00000004 381 #define VDESC_VP3_WILLRELE 0x00000008 382 #define VDESC_VP4_WILLRELE 0x00000010 383 #define VDESC_VP5_WILLRELE 0x00000020 384 #define VDESC_VP6_WILLRELE 0x00000040 385 #define VDESC_VP7_WILLRELE 0x00000080 386 #define VDESC_NOMAP_VPP 0x00000100 387 #define VDESC_VPP_WILLRELE 0x00000200 388 #define VDESC_VP0_WILLUNLOCK 0x00010000 389 #define VDESC_VP1_WILLUNLOCK 0x00020000 390 #define VDESC_VP2_WILLUNLOCK 0x00040000 391 #define VDESC_VP3_WILLUNLOCK 0x00080000 392 #define VDESC_VP4_WILLUNLOCK 0x00100000 393 #define VDESC_VP5_WILLUNLOCK 0x00200000 394 #define VDESC_VP6_WILLUNLOCK 0x00400000 395 #define VDESC_VP7_WILLUNLOCK 0x00800000 396 397 /* 398 * VDESC_NO_OFFSET is used to identify the end of the offset list 399 * and in places where no such field exists. 400 */ 401 #define VDESC_NO_OFFSET -1 402 403 /* 404 * This structure describes the vnode operation taking place. 405 */ 406 struct vnodeop_desc { 407 int vdesc_offset; /* offset in vector--first for speed */ 408 char *vdesc_name; /* a readable name for debugging */ 409 int vdesc_flags; /* VDESC_* flags */ 410 411 /* 412 * These ops are used by bypass routines to map and locate arguments. 413 * Creds and procs are not needed in bypass routines, but sometimes 414 * they are useful to (for example) transport layers. 415 * Nameidata is useful because it has a cred in it. 416 */ 417 int *vdesc_vp_offsets; /* list ended by VDESC_NO_OFFSET */ 418 int vdesc_vpp_offset; /* return vpp location */ 419 int vdesc_cred_offset; /* cred location, if any */ 420 int vdesc_proc_offset; /* proc location, if any */ 421 int vdesc_componentname_offset; /* if any */ 422 }; 423 424 #ifdef _KERNEL 425 /* 426 * A list of all the operation descs. 427 */ 428 extern struct vnodeop_desc *vnodeop_descs[]; 429 430 /* 431 * Interlock for scanning list of vnodes attached to a mountpoint 432 */ 433 extern struct lwkt_token mntvnode_token; 434 435 /* 436 * This macro is very helpful in defining those offsets in the vdesc struct. 437 * 438 * This is stolen from X11R4. I ignored all the fancy stuff for 439 * Crays, so if you decide to port this to such a serious machine, 440 * you might want to consult Intrinsic.h's XtOffset{,Of,To}. 441 */ 442 #define VOPARG_OFFSET(p_type,field) \ 443 ((int) (((char *) (&(((p_type)NULL)->field))) - ((char *) NULL))) 444 #define VOPARG_OFFSETOF(s_type,field) \ 445 VOPARG_OFFSET(s_type*,field) 446 #define VOPARG_OFFSETTO(S_TYPE,S_OFFSET,STRUCT_P) \ 447 ((S_TYPE)(((char*)(STRUCT_P))+(S_OFFSET))) 448 449 typedef int (*vnodeopv_entry_t)(struct vop_generic_args *); 450 451 /* 452 * This structure is used to configure the new vnodeops vector. The entry 453 * descriptor describes a patricular VOP function while the operations 454 * vector descriptor recursively describes arrays of entry descriptors. 455 */ 456 struct vnodeopv_entry_desc { 457 struct vnodeop_desc *opve_op; 458 vnodeopv_entry_t opve_func; 459 }; 460 461 struct vnodeopv_desc { 462 struct vop_ops **opv_desc_vector; /* vect to allocate/fill*/ 463 struct vnodeopv_entry_desc *opv_desc_ops; /* null terminated list */ 464 int opv_flags; 465 }; 466 467 struct vnodeopv_node { 468 TAILQ_ENTRY(vnodeopv_node) entry; 469 struct vop_ops *ops; /* allocated vector */ 470 struct vnodeopv_entry_desc *descs; /* null terminated list */ 471 }; 472 473 #ifdef DEBUG_VFS_LOCKS 474 /* 475 * Macros to aid in tracing VFS locking problems. Not totally 476 * reliable since if the process sleeps between changing the lock 477 * state and checking it with the assert, some other process could 478 * change the state. They are good enough for debugging a single 479 * filesystem using a single-threaded test. I find that 'cvs co src' 480 * is a pretty good test. 481 */ 482 483 /* 484 * [dfr] Kludge until I get around to fixing all the vfs locking. 485 */ 486 #define IS_LOCKING_VFS(vp) ((vp)->v_tag == VT_UFS \ 487 || (vp)->v_tag == VT_MFS \ 488 || (vp)->v_tag == VT_NFS \ 489 || (vp)->v_tag == VT_LFS \ 490 || (vp)->v_tag == VT_ISOFS \ 491 || (vp)->v_tag == VT_MSDOSFS) 492 493 #define ASSERT_VOP_LOCKED(vp, str) assert_vop_locked(vp, str) 494 #define ASSERT_VOP_UNLOCKED(vp, str) assert_vop_unlocked(vp, str); 495 496 #define ASSERT_VOP_ELOCKED(vp, str) \ 497 do { \ 498 struct vnode *_vp = (vp); \ 499 \ 500 if (_vp && IS_LOCKING_VFS(_vp) && \ 501 VOP_ISLOCKED(_vp, curthread) != LK_EXCLUSIVE) \ 502 panic("%s: %p is not exclusive locked but should be", \ 503 str, _vp); \ 504 } while (0) 505 506 #define ASSERT_VOP_ELOCKED_OTHER(vp, str) \ 507 do { \ 508 struct vnode *_vp = (vp); \ 509 \ 510 if (_vp && IS_LOCKING_VFS(_vp) && \ 511 VOP_ISLOCKED(_vp, curthread) != LK_EXCLOTHER) \ 512 panic("%s: %p is not exclusive locked by another proc", \ 513 str, _vp); \ 514 } while (0) 515 516 #define ASSERT_VOP_SLOCKED(vp, str) \ 517 do { \ 518 struct vnode *_vp = (vp); \ 519 \ 520 if (_vp && IS_LOCKING_VFS(_vp) && \ 521 VOP_ISLOCKED(_vp, NULL) != LK_SHARED) \ 522 panic("%s: %p is not locked shared but should be", \ 523 str, _vp); \ 524 } while (0) 525 526 void assert_vop_locked(struct vnode *vp, const char *str); 527 void assert_vop_unlocked(struct vnode *vp, const char *str); 528 529 #else 530 531 #define ASSERT_VOP_LOCKED(vp, str) 532 #define ASSERT_VOP_UNLOCKED(vp, str) 533 534 #endif /* DEBUG_VFS_LOCKS */ 535 536 /* 537 * VOCALL calls an op given an ops vector. We break it out because BSD's 538 * vclean changes the ops vector and then wants to call ops with the old 539 * vector. 540 */ 541 542 typedef int (*vocall_func_t)(struct vop_generic_args *); 543 544 /* 545 * This call executes the vops vector for the offset stored in the ap's 546 * descriptor of the passed vops rather then the one related to the 547 * ap's vop_ops structure. It is used to chain VOPS calls on behalf of 548 * filesystems from a VFS's context ONLY (that is, from a VFS's own vops 549 * vector function). 550 */ 551 #define VOCALL(vops, ap) \ 552 (*(vocall_func_t *)((char *)(vops)+((ap)->a_desc->vdesc_offset)))(ap) 553 554 #define VDESC(OP) (& __CONCAT(OP,_desc)) 555 #define VOFFSET(OP) (VDESC(OP)->vdesc_offset) 556 557 /* 558 * VMIO support inline 559 */ 560 561 extern int vmiodirenable; 562 563 static __inline int 564 vn_canvmio(struct vnode *vp) 565 { 566 if (vp && (vp->v_type == VREG || (vmiodirenable && vp->v_type == VDIR))) 567 return(TRUE); 568 return(FALSE); 569 } 570 571 /* 572 * Public vnode manipulation functions. 573 */ 574 struct file; 575 struct mount; 576 struct nlookupdata; 577 struct proc; 578 struct thread; 579 struct stat; 580 struct ucred; 581 struct uio; 582 struct vattr; 583 struct vnode; 584 struct vop_bwrite_args; 585 586 extern int (*lease_check_hook) (struct vop_lease_args *); 587 588 void addaliasu (struct vnode *vp, udev_t nvp_udev); 589 int v_associate_rdev(struct vnode *vp, dev_t dev); 590 void v_release_rdev(struct vnode *vp); 591 int bdevvp (dev_t dev, struct vnode **vpp); 592 struct vnode *allocvnode(int lktimeout, int lkflags); 593 int getnewvnode (enum vtagtype tag, struct mount *mp, 594 struct vnode **vpp, int timo, int lkflags); 595 int getspecialvnode (enum vtagtype tag, struct mount *mp, 596 struct vop_ops **ops, struct vnode **vpp, int timo, 597 int lkflags); 598 int lease_check (struct vop_lease_args *ap); 599 int spec_vnoperate (struct vop_generic_args *); 600 int speedup_syncer (void); 601 void vattr_null (struct vattr *vap); 602 int vcount (struct vnode *vp); 603 int vfinddev (dev_t dev, enum vtype type, struct vnode **vpp); 604 void vfs_add_vnodeops_sysinit (const void *); 605 void vfs_rm_vnodeops_sysinit (const void *); 606 void vfs_add_vnodeops(struct mount *, struct vop_ops **, 607 struct vnodeopv_entry_desc *, int); 608 void vfs_rm_vnodeops(struct vop_ops **); 609 int vflush (struct mount *mp, int rootrefs, int flags); 610 int vmntvnodescan(struct mount *mp, int flags, 611 int (*fastfunc)(struct mount *mp, struct vnode *vp, void *data), 612 int (*slowfunc)(struct mount *mp, struct vnode *vp, void *data), 613 void *data); 614 void insmntque(struct vnode *vp, struct mount *mp); 615 616 void vclean (struct vnode *vp, int flags, struct thread *td); 617 void vgone (struct vnode *vp); 618 int vinvalbuf (struct vnode *vp, int save, 619 struct thread *td, int slpflag, int slptimeo); 620 int vtruncbuf (struct vnode *vp, struct thread *td, 621 off_t length, int blksize); 622 int vfsync(struct vnode *vp, int waitfor, int passes, daddr_t lbn, 623 int (*checkdef)(struct buf *), 624 int (*waitoutput)(struct vnode *, struct thread *)); 625 void vprint (char *label, struct vnode *vp); 626 int vrecycle (struct vnode *vp, struct thread *td); 627 void vn_strategy(struct vnode *vp, struct bio *bio); 628 int vn_close (struct vnode *vp, int flags, struct thread *td); 629 int vn_isdisk (struct vnode *vp, int *errp); 630 int vn_lock (struct vnode *vp, int flags, struct thread *td); 631 #ifdef DEBUG_LOCKS 632 int debug_vn_lock (struct vnode *vp, int flags, struct thread *td, 633 const char *filename, int line); 634 #define vn_lock(vp,flags,p) debug_vn_lock(vp,flags,p,__FILE__,__LINE__) 635 #endif 636 637 int vn_get_namelen(struct vnode *, int *); 638 void vn_setspecops (struct file *fp); 639 int vn_fullpath (struct proc *p, struct vnode *vn, char **retbuf, char **freebuf); 640 int vn_open (struct nlookupdata *ndp, struct file *fp, int fmode, int cmode); 641 void vn_pollevent (struct vnode *vp, int events); 642 void vn_pollgone (struct vnode *vp); 643 int vn_pollrecord (struct vnode *vp, struct thread *td, int events); 644 int vn_rdwr (enum uio_rw rw, struct vnode *vp, caddr_t base, 645 int len, off_t offset, enum uio_seg segflg, int ioflg, 646 struct ucred *cred, int *aresid, struct thread *td); 647 int vn_rdwr_inchunks (enum uio_rw rw, struct vnode *vp, caddr_t base, 648 int len, off_t offset, enum uio_seg segflg, int ioflg, 649 struct ucred *cred, int *aresid, struct thread *td); 650 int vn_stat (struct vnode *vp, struct stat *sb, struct thread *td); 651 dev_t vn_todev (struct vnode *vp); 652 int vfs_object_create (struct vnode *vp, struct thread *td); 653 void vfs_timestamp (struct timespec *); 654 int vn_writechk (struct vnode *vp); 655 int vop_stdbwrite (struct vop_bwrite_args *ap); 656 int vop_stdislocked (struct vop_islocked_args *ap); 657 int vop_stdlock (struct vop_lock_args *ap); 658 int vop_stdrlock (struct vop_lock_args *ap); 659 int vop_stdunlock (struct vop_unlock_args *ap); 660 int vop_nopoll (struct vop_poll_args *ap); 661 int vop_stdpathconf (struct vop_pathconf_args *ap); 662 int vop_stdpoll (struct vop_poll_args *ap); 663 int vop_stdrevoke (struct vop_revoke_args *ap); 664 int vop_eopnotsupp (struct vop_generic_args *ap); 665 int vop_ebadf (struct vop_generic_args *ap); 666 int vop_einval (struct vop_generic_args *ap); 667 int vop_enotty (struct vop_generic_args *ap); 668 int vop_defaultop (struct vop_generic_args *ap); 669 int vop_null (struct vop_generic_args *ap); 670 int vop_panic (struct vop_generic_args *ap); 671 int vop_stdcreatevobject (struct vop_createvobject_args *ap); 672 int vop_stddestroyvobject (struct vop_destroyvobject_args *ap); 673 int vop_stdgetvobject (struct vop_getvobject_args *ap); 674 int vop_write_dirent(int *, struct uio *, ino_t, uint8_t, uint16_t, 675 const char *); 676 677 int vop_compat_nresolve(struct vop_nresolve_args *ap); 678 int vop_compat_nlookupdotdot(struct vop_nlookupdotdot_args *ap); 679 int vop_compat_ncreate(struct vop_ncreate_args *ap); 680 int vop_compat_nmkdir(struct vop_nmkdir_args *ap); 681 int vop_compat_nmknod(struct vop_nmknod_args *ap); 682 int vop_compat_nlink(struct vop_nlink_args *ap); 683 int vop_compat_nsymlink(struct vop_nsymlink_args *ap); 684 int vop_compat_nwhiteout(struct vop_nwhiteout_args *ap); 685 int vop_compat_nremove(struct vop_nremove_args *ap); 686 int vop_compat_nrmdir(struct vop_nrmdir_args *ap); 687 int vop_compat_nrename(struct vop_nrename_args *ap); 688 689 int vx_lock (struct vnode *vp); 690 void vx_unlock (struct vnode *vp); 691 int vx_get (struct vnode *vp); 692 int vx_get_nonblock (struct vnode *vp); 693 void vx_put (struct vnode *vp); 694 int vget (struct vnode *vp, int lockflag, struct thread *td); 695 void vput (struct vnode *vp); 696 void vhold (struct vnode *); 697 void vdrop (struct vnode *); 698 void vref (struct vnode *vp); 699 void vrele (struct vnode *vp); 700 void vsetflags (struct vnode *vp, int flags); 701 void vclrflags (struct vnode *vp, int flags); 702 703 void vfs_subr_init(void); 704 void vfs_mount_init(void); 705 void vfs_lock_init(void); 706 void vfs_sync_init(void); 707 708 void vn_syncer_add_to_worklist(struct vnode *, int); 709 void vnlru_proc_wait(void); 710 711 extern struct vop_ops *default_vnode_vops; 712 extern struct vop_ops *spec_vnode_vops; 713 extern struct vop_ops *dead_vnode_vops; 714 715 #endif /* _KERNEL */ 716 717 #endif /* !_SYS_VNODE_H_ */ 718