1 /* 2 * Copyright (c) 1989 The Regents of the University of California. 3 * All rights reserved. 4 * 5 * %sccs.include.redist.c% 6 * 7 * @(#)mfs_vnops.c 7.34 (Berkeley) 06/04/92 8 */ 9 10 #include <sys/param.h> 11 #include <sys/systm.h> 12 #include <sys/time.h> 13 #include <sys/kernel.h> 14 #include <sys/proc.h> 15 #include <sys/buf.h> 16 #include <sys/map.h> 17 #include <sys/vnode.h> 18 #include <sys/malloc.h> 19 20 #include <machine/vmparam.h> 21 #if defined(tahoe) 22 #include <machine/mtpr.h> 23 #endif 24 25 #include <ufs/mfs/mfsnode.h> 26 #include <ufs/mfs/mfsiom.h> 27 #include <ufs/mfs/mfs_extern.h> 28 29 #if !defined(hp300) && !defined(i386) && !defined(mips) && !defined(sparc) && !defined(luna68k) 30 static int mfsmap_want; /* 1 => need kernel I/O resources */ 31 struct map mfsmap[MFS_MAPSIZE]; 32 extern char mfsiobuf[]; 33 #endif 34 35 /* 36 * mfs vnode operations. 37 */ 38 int (**mfs_vnodeop_p)(); 39 struct vnodeopv_entry_desc mfs_vnodeop_entries[] = { 40 { &vop_default_desc, vn_default_error }, 41 { &vop_lookup_desc, mfs_lookup }, /* lookup */ 42 { &vop_create_desc, mfs_create }, /* create */ 43 { &vop_mknod_desc, mfs_mknod }, /* mknod */ 44 { &vop_open_desc, mfs_open }, /* open */ 45 { &vop_close_desc, mfs_close }, /* close */ 46 { &vop_access_desc, mfs_access }, /* access */ 47 { &vop_getattr_desc, mfs_getattr }, /* getattr */ 48 { &vop_setattr_desc, mfs_setattr }, /* setattr */ 49 { &vop_read_desc, mfs_read }, /* read */ 50 { &vop_write_desc, mfs_write }, /* write */ 51 { &vop_ioctl_desc, mfs_ioctl }, /* ioctl */ 52 { &vop_select_desc, mfs_select }, /* select */ 53 { &vop_mmap_desc, mfs_mmap }, /* mmap */ 54 { &vop_fsync_desc, mfs_fsync }, /* fsync */ 55 { &vop_seek_desc, mfs_seek }, /* seek */ 56 { &vop_remove_desc, mfs_remove }, /* remove */ 57 { &vop_link_desc, mfs_link }, /* link */ 58 { &vop_rename_desc, mfs_rename }, /* rename */ 59 { &vop_mkdir_desc, mfs_mkdir }, /* mkdir */ 60 { &vop_rmdir_desc, mfs_rmdir }, /* rmdir */ 61 { &vop_symlink_desc, mfs_symlink }, /* symlink */ 62 { &vop_readdir_desc, mfs_readdir }, /* readdir */ 63 { &vop_readlink_desc, mfs_readlink }, /* readlink */ 64 { &vop_abortop_desc, mfs_abortop }, /* abortop */ 65 { &vop_inactive_desc, mfs_inactive }, /* inactive */ 66 { &vop_reclaim_desc, mfs_reclaim }, /* reclaim */ 67 { &vop_lock_desc, mfs_lock }, /* lock */ 68 { &vop_unlock_desc, mfs_unlock }, /* unlock */ 69 { &vop_bmap_desc, mfs_bmap }, /* bmap */ 70 { &vop_strategy_desc, mfs_strategy }, /* strategy */ 71 { &vop_print_desc, mfs_print }, /* print */ 72 { &vop_islocked_desc, mfs_islocked }, /* islocked */ 73 { &vop_advlock_desc, mfs_advlock }, /* advlock */ 74 { &vop_blkatoff_desc, mfs_blkatoff }, /* blkatoff */ 75 { &vop_vget_desc, mfs_vget }, /* vget */ 76 { &vop_valloc_desc, mfs_valloc }, /* valloc */ 77 { &vop_vfree_desc, mfs_vfree }, /* vfree */ 78 { &vop_truncate_desc, mfs_truncate }, /* truncate */ 79 { &vop_update_desc, mfs_update }, /* update */ 80 { &vop_bwrite_desc, mfs_bwrite }, /* bwrite */ 81 { (struct vnodeop_desc*)NULL, (int(*)())NULL } 82 }; 83 struct vnodeopv_desc mfs_vnodeop_opv_desc = 84 { &mfs_vnodeop_p, mfs_vnodeop_entries }; 85 86 /* 87 * Vnode Operations. 88 * 89 * Open called to allow memory filesystem to initialize and 90 * validate before actual IO. Record our process identifier 91 * so we can tell when we are doing I/O to ourself. 92 */ 93 /* ARGSUSED */ 94 int 95 mfs_open (ap) 96 struct vop_open_args *ap; 97 { 98 99 if (ap->a_vp->v_type != VBLK) { 100 panic("mfs_ioctl not VBLK"); 101 /* NOTREACHED */ 102 } 103 return (0); 104 } 105 106 /* 107 * Ioctl operation. 108 */ 109 /* ARGSUSED */ 110 int 111 mfs_ioctl (ap) 112 struct vop_ioctl_args *ap; 113 { 114 115 return (ENOTTY); 116 } 117 118 /* 119 * Pass I/O requests to the memory filesystem process. 120 */ 121 int 122 mfs_strategy (ap) 123 struct vop_strategy_args *ap; 124 { 125 register struct buf *bp = ap->a_bp; 126 register struct mfsnode *mfsp; 127 struct vnode *vp; 128 struct proc *p = curproc; /* XXX */ 129 130 if (vfinddev(bp->b_dev, VBLK, &vp) || vp->v_usecount == 0) 131 panic("mfs_strategy: bad dev"); 132 mfsp = VTOMFS(vp); 133 /* check for mini-root access */ 134 if (mfsp->mfs_pid == 0) { 135 caddr_t base; 136 137 base = mfsp->mfs_baseoff + (bp->b_blkno << DEV_BSHIFT); 138 if (bp->b_flags & B_READ) 139 bcopy(base, bp->b_un.b_addr, bp->b_bcount); 140 else 141 bcopy(bp->b_un.b_addr, base, bp->b_bcount); 142 biodone(bp); 143 } else if (mfsp->mfs_pid == p->p_pid) { 144 mfs_doio(bp, mfsp->mfs_baseoff); 145 } else { 146 bp->av_forw = mfsp->mfs_buflist; 147 mfsp->mfs_buflist = bp; 148 wakeup((caddr_t)vp); 149 } 150 return (0); 151 } 152 153 #if defined(vax) || defined(tahoe) 154 /* 155 * Memory file system I/O. 156 * 157 * Essentially play ubasetup() and disk interrupt service routine by 158 * doing the copies to or from the memfs process. If doing physio 159 * (i.e. pagein), we must map the I/O through the kernel virtual 160 * address space. 161 */ 162 void 163 mfs_doio(bp, base) 164 register struct buf *bp; 165 caddr_t base; 166 { 167 register struct pte *pte, *ppte; 168 register caddr_t vaddr; 169 int off, npf, npf2, reg; 170 caddr_t kernaddr, offset; 171 172 /* 173 * For phys I/O, map the b_addr into kernel virtual space using 174 * the Mfsiomap pte's. 175 */ 176 if ((bp->b_flags & B_PHYS) == 0) { 177 kernaddr = bp->b_un.b_addr; 178 } else { 179 if (bp->b_flags & (B_PAGET | B_UAREA | B_DIRTY)) 180 panic("swap on memfs?"); 181 off = (int)bp->b_un.b_addr & PGOFSET; 182 npf = btoc(bp->b_bcount + off); 183 /* 184 * Get some mapping page table entries 185 */ 186 while ((reg = rmalloc(mfsmap, (long)npf)) == 0) { 187 mfsmap_want++; 188 sleep((caddr_t)&mfsmap_want, PZERO-1); 189 } 190 reg--; 191 pte = vtopte(bp->b_proc, btop(bp->b_un.b_addr)); 192 /* 193 * Do vmaccess() but with the Mfsiomap page table. 194 */ 195 ppte = &Mfsiomap[reg]; 196 vaddr = &mfsiobuf[reg * NBPG]; 197 kernaddr = vaddr + off; 198 for (npf2 = npf; npf2; npf2--) { 199 mapin(ppte, (u_int)vaddr, pte->pg_pfnum, 200 (int)(PG_V|PG_KW)); 201 #if defined(tahoe) 202 if ((bp->b_flags & B_READ) == 0) 203 mtpr(P1DC, vaddr); 204 #endif 205 ppte++; 206 pte++; 207 vaddr += NBPG; 208 } 209 } 210 offset = base + (bp->b_blkno << DEV_BSHIFT); 211 if (bp->b_flags & B_READ) 212 bp->b_error = copyin(offset, kernaddr, bp->b_bcount); 213 else 214 bp->b_error = copyout(kernaddr, offset, bp->b_bcount); 215 if (bp->b_error) 216 bp->b_flags |= B_ERROR; 217 /* 218 * Release pte's used by physical I/O. 219 */ 220 if (bp->b_flags & B_PHYS) { 221 rmfree(mfsmap, (long)npf, (long)++reg); 222 if (mfsmap_want) { 223 mfsmap_want = 0; 224 wakeup((caddr_t)&mfsmap_want); 225 } 226 } 227 biodone(bp); 228 } 229 #endif /* vax || tahoe */ 230 231 #if defined(hp300) || defined(i386) || defined(mips) || defined(sparc) || defined(luna68k) 232 /* 233 * Memory file system I/O. 234 * 235 * Trivial on the HP since buffer has already been mapping into KVA space. 236 */ 237 void 238 mfs_doio(bp, base) 239 register struct buf *bp; 240 caddr_t base; 241 { 242 243 base += (bp->b_blkno << DEV_BSHIFT); 244 if (bp->b_flags & B_READ) 245 bp->b_error = copyin(base, bp->b_un.b_addr, bp->b_bcount); 246 else 247 bp->b_error = copyout(bp->b_un.b_addr, base, bp->b_bcount); 248 if (bp->b_error) 249 bp->b_flags |= B_ERROR; 250 biodone(bp); 251 } 252 #endif 253 254 /* 255 * This is a noop, simply returning what one has been given. 256 */ 257 int 258 mfs_bmap (ap) 259 struct vop_bmap_args *ap; 260 { 261 262 if (ap->a_vpp != NULL) 263 *ap->a_vpp = ap->a_vp; 264 if (ap->a_bnp != NULL) 265 *ap->a_bnp = ap->a_bn; 266 return (0); 267 } 268 269 /* 270 * Memory filesystem close routine 271 */ 272 /* ARGSUSED */ 273 int 274 mfs_close (ap) 275 struct vop_close_args *ap; 276 { 277 register struct vnode *vp = ap->a_vp; 278 register struct mfsnode *mfsp = VTOMFS(vp); 279 register struct buf *bp; 280 281 /* 282 * Finish any pending I/O requests. 283 */ 284 while (bp = mfsp->mfs_buflist) { 285 mfsp->mfs_buflist = bp->av_forw; 286 mfs_doio(bp, mfsp->mfs_baseoff); 287 wakeup((caddr_t)bp); 288 } 289 /* 290 * On last close of a memory filesystem 291 * we must invalidate any in core blocks, so that 292 * we can, free up its vnode. 293 */ 294 vflushbuf(vp, 0); 295 if (vinvalbuf(vp, 1)) 296 return (0); 297 /* 298 * There should be no way to have any more uses of this 299 * vnode, so if we find any other uses, it is a panic. 300 */ 301 if (vp->v_usecount > 1) 302 printf("mfs_close: ref count %d > 1\n", vp->v_usecount); 303 if (vp->v_usecount > 1 || mfsp->mfs_buflist) 304 panic("mfs_close"); 305 /* 306 * Send a request to the filesystem server to exit. 307 */ 308 mfsp->mfs_buflist = (struct buf *)(-1); 309 wakeup((caddr_t)vp); 310 return (0); 311 } 312 313 /* 314 * Memory filesystem inactive routine 315 */ 316 /* ARGSUSED */ 317 int 318 mfs_inactive (ap) 319 struct vop_inactive_args *ap; 320 { 321 register struct mfsnode *mfsp = VTOMFS(ap->a_vp); 322 323 if (mfsp->mfs_buflist && mfsp->mfs_buflist != (struct buf *)(-1)) 324 panic("mfs_inactive: not inactive (mfs_buflist %x)", 325 mfsp->mfs_buflist); 326 return (0); 327 } 328 329 /* 330 * Reclaim a memory filesystem devvp so that it can be reused. 331 */ 332 int 333 mfs_reclaim (ap) 334 struct vop_reclaim_args *ap; 335 { 336 337 FREE(ap->a_vp->v_data, M_MFSNODE); 338 ap->a_vp->v_data = NULL; 339 return (0); 340 } 341 342 /* 343 * Print out the contents of an mfsnode. 344 */ 345 int 346 mfs_print (ap) 347 struct vop_print_args *ap; 348 { 349 register struct mfsnode *mfsp = VTOMFS(ap->a_vp); 350 351 printf("tag VT_MFS, pid %d, base %d, size %d\n", mfsp->mfs_pid, 352 mfsp->mfs_baseoff, mfsp->mfs_size); 353 return (0); 354 } 355 356 /* 357 * Block device bad operation 358 */ 359 int 360 mfs_badop() 361 { 362 363 panic("mfs_badop called\n"); 364 /* NOTREACHED */ 365 } 366 367 /* 368 * Memory based filesystem initialization. 369 */ 370 mfs_init() 371 { 372 373 #if !defined(hp300) && !defined(i386) && !defined(mips) && !defined(sparc) && !defined(luna68k) 374 rminit(mfsmap, (long)MFS_MAPREG, (long)1, "mfs mapreg", MFS_MAPSIZE); 375 #endif 376 } 377