1 /*	$NetBSD: coda_psdev.c,v 1.57 2016/07/07 06:55:40 msaitoh Exp $	*/
2 
3 /*
4  *
5  *             Coda: an Experimental Distributed File System
6  *                              Release 3.1
7  *
8  *           Copyright (c) 1987-1998 Carnegie Mellon University
9  *                          All Rights Reserved
10  *
11  * Permission  to  use, copy, modify and distribute this software and its
12  * documentation is hereby granted,  provided  that  both  the  copyright
13  * notice  and  this  permission  notice  appear  in  all  copies  of the
14  * software, derivative works or  modified  versions,  and  any  portions
15  * thereof, and that both notices appear in supporting documentation, and
16  * that credit is given to Carnegie Mellon University  in  all  documents
17  * and publicity pertaining to direct or indirect use of this code or its
18  * derivatives.
19  *
20  * CODA IS AN EXPERIMENTAL SOFTWARE SYSTEM AND IS  KNOWN  TO  HAVE  BUGS,
21  * SOME  OF  WHICH MAY HAVE SERIOUS CONSEQUENCES.  CARNEGIE MELLON ALLOWS
22  * FREE USE OF THIS SOFTWARE IN ITS "AS IS" CONDITION.   CARNEGIE  MELLON
23  * DISCLAIMS  ANY  LIABILITY  OF  ANY  KIND  FOR  ANY  DAMAGES WHATSOEVER
24  * RESULTING DIRECTLY OR INDIRECTLY FROM THE USE OF THIS SOFTWARE  OR  OF
25  * ANY DERIVATIVE WORK.
26  *
27  * Carnegie  Mellon  encourages  users  of  this  software  to return any
28  * improvements or extensions that  they  make,  and  to  grant  Carnegie
29  * Mellon the rights to redistribute these changes without encumbrance.
30  *
31  * 	@(#) coda/coda_psdev.c,v 1.1.1.1 1998/08/29 21:26:45 rvb Exp $
32  */
33 
34 /*
35  * Mach Operating System
36  * Copyright (c) 1989 Carnegie-Mellon University
37  * All rights reserved.  The CMU software License Agreement specifies
38  * the terms and conditions for use and redistribution.
39  */
40 
41 /*
42  * This code was written for the Coda file system at Carnegie Mellon
43  * University.  Contributers include David Steere, James Kistler, and
44  * M. Satyanarayanan.  */
45 
46 /* These routines define the pseudo device for communication between
47  * Coda's Venus and Minicache in Mach 2.6. They used to be in cfs_subr.c,
48  * but I moved them to make it easier to port the Minicache without
49  * porting coda. -- DCS 10/12/94
50  *
51  * Following code depends on file-system CODA.
52  */
53 
54 /* These routines are the device entry points for Venus. */
55 
56 #include <sys/cdefs.h>
57 __KERNEL_RCSID(0, "$NetBSD: coda_psdev.c,v 1.57 2016/07/07 06:55:40 msaitoh Exp $");
58 
59 extern int coda_nc_initialized;    /* Set if cache has been initialized */
60 
61 #include <sys/param.h>
62 #include <sys/systm.h>
63 #include <sys/kernel.h>
64 #include <sys/malloc.h>
65 #include <sys/proc.h>
66 #include <sys/mount.h>
67 #include <sys/file.h>
68 #include <sys/ioctl.h>
69 #include <sys/poll.h>
70 #include <sys/select.h>
71 #include <sys/conf.h>
72 #include <sys/atomic.h>
73 #include <sys/module.h>
74 
75 #include <coda/coda.h>
76 #include <coda/cnode.h>
77 #include <coda/coda_namecache.h>
78 #include <coda/coda_io.h>
79 
80 #include "ioconf.h"
81 
82 #define CTL_C
83 
84 int coda_psdev_print_entry = 0;
85 static
86 int outstanding_upcalls = 0;
87 int coda_call_sleep = PZERO - 1;
88 #ifdef	CTL_C
89 int coda_pcatch = PCATCH;
90 #else
91 #endif
92 
93 int coda_kernel_version = CODA_KERNEL_VERSION;
94 
95 #define ENTRY if(coda_psdev_print_entry) myprintf(("Entered %s\n",__func__))
96 
97 dev_type_open(vc_nb_open);
98 dev_type_close(vc_nb_close);
99 dev_type_read(vc_nb_read);
100 dev_type_write(vc_nb_write);
101 dev_type_ioctl(vc_nb_ioctl);
102 dev_type_poll(vc_nb_poll);
103 dev_type_kqfilter(vc_nb_kqfilter);
104 
105 const struct cdevsw vcoda_cdevsw = {
106 	.d_open = vc_nb_open,
107 	.d_close = vc_nb_close,
108 	.d_read = vc_nb_read,
109 	.d_write = vc_nb_write,
110 	.d_ioctl = vc_nb_ioctl,
111 	.d_stop = nostop,
112 	.d_tty = notty,
113 	.d_poll = vc_nb_poll,
114 	.d_mmap = nommap,
115 	.d_kqfilter = vc_nb_kqfilter,
116 	.d_discard = nodiscard,
117 	.d_flag = D_OTHER,
118 };
119 
120 struct vmsg {
121     TAILQ_ENTRY(vmsg) vm_chain;
122     void *	 vm_data;
123     u_short	 vm_flags;
124     u_short      vm_inSize;	/* Size is at most 5000 bytes */
125     u_short	 vm_outSize;
126     u_short	 vm_opcode; 	/* copied from data to save ptr lookup */
127     int		 vm_unique;
128     void *	 vm_sleep;	/* Not used by Mach. */
129 };
130 
131 struct coda_mntinfo coda_mnttbl[NVCODA];
132 
133 #define	VM_READ	    1
134 #define	VM_WRITE    2
135 #define	VM_INTR	    4
136 
137 /* vcodaattach: do nothing */
138 void
vcodaattach(int n)139 vcodaattach(int n)
140 {
141 }
142 
143 /*
144  * These functions are written for NetBSD.
145  */
146 int
vc_nb_open(dev_t dev,int flag,int mode,struct lwp * l)147 vc_nb_open(dev_t dev, int flag, int mode,
148     struct lwp *l)
149 {
150     struct vcomm *vcp;
151 
152     ENTRY;
153 
154     if (minor(dev) >= NVCODA)
155 	return(ENXIO);
156 
157     if (!coda_nc_initialized)
158 	coda_nc_init();
159 
160     vcp = &coda_mnttbl[minor(dev)].mi_vcomm;
161     if (VC_OPEN(vcp))
162 	return(EBUSY);
163 
164     selinit(&vcp->vc_selproc);
165     TAILQ_INIT(&vcp->vc_requests);
166     TAILQ_INIT(&vcp->vc_replies);
167     MARK_VC_OPEN(vcp);
168 
169     coda_mnttbl[minor(dev)].mi_vfsp = NULL;
170     coda_mnttbl[minor(dev)].mi_rootvp = NULL;
171 
172     return(0);
173 }
174 
175 int
vc_nb_close(dev_t dev,int flag,int mode,struct lwp * l)176 vc_nb_close(dev_t dev, int flag, int mode, struct lwp *l)
177 {
178     struct vcomm *vcp;
179     struct vmsg *vmp;
180     struct coda_mntinfo *mi;
181     int                 err;
182 
183     ENTRY;
184 
185     if (minor(dev) >= NVCODA)
186 	return(ENXIO);
187 
188     mi = &coda_mnttbl[minor(dev)];
189     vcp = &(mi->mi_vcomm);
190 
191     if (!VC_OPEN(vcp))
192 	panic("vcclose: not open");
193 
194     /* prevent future operations on this vfs from succeeding by auto-
195      * unmounting any vfs mounted via this device. This frees user or
196      * sysadm from having to remember where all mount points are located.
197      * Put this before WAKEUPs to avoid queuing new messages between
198      * the WAKEUP and the unmount (which can happen if we're unlucky)
199      */
200     if (!mi->mi_rootvp) {
201 	/* just a simple open/close w no mount */
202 	MARK_VC_CLOSED(vcp);
203 	return 0;
204     }
205 
206     /* Let unmount know this is for real */
207     VTOC(mi->mi_rootvp)->c_flags |= C_UNMOUNTING;
208     coda_unmounting(mi->mi_vfsp);
209 
210     /* Wakeup clients so they can return. */
211     while ((vmp = TAILQ_FIRST(&vcp->vc_requests)) != NULL) {
212 	TAILQ_REMOVE(&vcp->vc_requests, vmp, vm_chain);
213 
214 	/* Free signal request messages and don't wakeup cause
215 	   no one is waiting. */
216 	if (vmp->vm_opcode == CODA_SIGNAL) {
217 	    CODA_FREE(vmp->vm_data, VC_IN_NO_DATA);
218 	    CODA_FREE(vmp, sizeof(struct vmsg));
219 	    continue;
220 	}
221 	outstanding_upcalls++;
222 	wakeup(&vmp->vm_sleep);
223     }
224 
225     while ((vmp = TAILQ_FIRST(&vcp->vc_replies)) != NULL) {
226 	TAILQ_REMOVE(&vcp->vc_replies, vmp, vm_chain);
227 
228 	outstanding_upcalls++;
229 	wakeup(&vmp->vm_sleep);
230     }
231 
232     MARK_VC_CLOSED(vcp);
233 
234     if (outstanding_upcalls) {
235 #ifdef	CODA_VERBOSE
236 	printf("presleep: outstanding_upcalls = %d\n", outstanding_upcalls);
237     	(void) tsleep(&outstanding_upcalls, coda_call_sleep, "coda_umount", 0);
238 	printf("postsleep: outstanding_upcalls = %d\n", outstanding_upcalls);
239 #else
240     	(void) tsleep(&outstanding_upcalls, coda_call_sleep, "coda_umount", 0);
241 #endif
242     }
243 
244     err = dounmount(mi->mi_vfsp, flag, l);
245     if (err)
246 	myprintf(("Error %d unmounting vfs in vcclose(%llu)\n",
247 	           err, (unsigned long long)minor(dev)));
248     seldestroy(&vcp->vc_selproc);
249     return 0;
250 }
251 
252 int
vc_nb_read(dev_t dev,struct uio * uiop,int flag)253 vc_nb_read(dev_t dev, struct uio *uiop, int flag)
254 {
255     struct vcomm *	vcp;
256     struct vmsg *vmp;
257     int error = 0;
258 
259     ENTRY;
260 
261     if (minor(dev) >= NVCODA)
262 	return(ENXIO);
263 
264     vcp = &coda_mnttbl[minor(dev)].mi_vcomm;
265 
266     /* Get message at head of request queue. */
267     vmp = TAILQ_FIRST(&vcp->vc_requests);
268     if (vmp == NULL)
269 	return(0);	/* Nothing to read */
270 
271     /* Move the input args into userspace */
272     uiop->uio_rw = UIO_READ;
273     error = uiomove(vmp->vm_data, vmp->vm_inSize, uiop);
274     if (error) {
275 	myprintf(("vcread: error (%d) on uiomove\n", error));
276 	error = EINVAL;
277     }
278 
279     TAILQ_REMOVE(&vcp->vc_requests, vmp, vm_chain);
280 
281     /* If request was a signal, free up the message and don't
282        enqueue it in the reply queue. */
283     if (vmp->vm_opcode == CODA_SIGNAL) {
284 	if (codadebug)
285 	    myprintf(("vcread: signal msg (%d, %d)\n",
286 		      vmp->vm_opcode, vmp->vm_unique));
287 	CODA_FREE(vmp->vm_data, VC_IN_NO_DATA);
288 	CODA_FREE(vmp, sizeof(struct vmsg));
289 	return(error);
290     }
291 
292     vmp->vm_flags |= VM_READ;
293     TAILQ_INSERT_TAIL(&vcp->vc_replies, vmp, vm_chain);
294 
295     return(error);
296 }
297 
298 int
vc_nb_write(dev_t dev,struct uio * uiop,int flag)299 vc_nb_write(dev_t dev, struct uio *uiop, int flag)
300 {
301     struct vcomm *	vcp;
302     struct vmsg *vmp;
303     struct coda_out_hdr *out;
304     u_long seq;
305     u_long opcode;
306     int tbuf[2];
307     int error = 0;
308 
309     ENTRY;
310 
311     if (minor(dev) >= NVCODA)
312 	return(ENXIO);
313 
314     vcp = &coda_mnttbl[minor(dev)].mi_vcomm;
315 
316     /* Peek at the opcode, unique without transfering the data. */
317     uiop->uio_rw = UIO_WRITE;
318     error = uiomove(tbuf, sizeof(int) * 2, uiop);
319     if (error) {
320 	myprintf(("vcwrite: error (%d) on uiomove\n", error));
321 	return(EINVAL);
322     }
323 
324     opcode = tbuf[0];
325     seq = tbuf[1];
326 
327     if (codadebug)
328 	myprintf(("vcwrite got a call for %ld.%ld\n", opcode, seq));
329 
330     if (DOWNCALL(opcode)) {
331 	union outputArgs pbuf;
332 
333 	/* get the rest of the data. */
334 	uiop->uio_rw = UIO_WRITE;
335 	error = uiomove(&pbuf.coda_purgeuser.oh.result, sizeof(pbuf) - (sizeof(int)*2), uiop);
336 	if (error) {
337 	    myprintf(("vcwrite: error (%d) on uiomove (Op %ld seq %ld)\n",
338 		      error, opcode, seq));
339 	    return(EINVAL);
340 	    }
341 
342 	return handleDownCall(opcode, &pbuf);
343     }
344 
345     /* Look for the message on the (waiting for) reply queue. */
346     TAILQ_FOREACH(vmp, &vcp->vc_replies, vm_chain) {
347 	if (vmp->vm_unique == seq) break;
348     }
349 
350     if (vmp == NULL) {
351 	if (codadebug)
352 	    myprintf(("vcwrite: msg (%ld, %ld) not found\n", opcode, seq));
353 
354 	return(ESRCH);
355     }
356 
357     /* Remove the message from the reply queue */
358     TAILQ_REMOVE(&vcp->vc_replies, vmp, vm_chain);
359 
360     /* move data into response buffer. */
361     out = (struct coda_out_hdr *)vmp->vm_data;
362     /* Don't need to copy opcode and uniquifier. */
363 
364     /* get the rest of the data. */
365     if (vmp->vm_outSize < uiop->uio_resid) {
366 	myprintf(("vcwrite: more data than asked for (%d < %lu)\n",
367 		  vmp->vm_outSize, (unsigned long) uiop->uio_resid));
368 	wakeup(&vmp->vm_sleep); 	/* Notify caller of the error. */
369 	return(EINVAL);
370     }
371 
372     tbuf[0] = uiop->uio_resid; 	/* Save this value. */
373     uiop->uio_rw = UIO_WRITE;
374     error = uiomove(&out->result, vmp->vm_outSize - (sizeof(int) * 2), uiop);
375     if (error) {
376 	myprintf(("vcwrite: error (%d) on uiomove (op %ld seq %ld)\n",
377 		  error, opcode, seq));
378 	return(EINVAL);
379     }
380 
381     /* I don't think these are used, but just in case. */
382     /* XXX - aren't these two already correct? -bnoble */
383     out->opcode = opcode;
384     out->unique = seq;
385     vmp->vm_outSize	= tbuf[0];	/* Amount of data transferred? */
386     vmp->vm_flags |= VM_WRITE;
387     wakeup(&vmp->vm_sleep);
388 
389     return(0);
390 }
391 
392 int
vc_nb_ioctl(dev_t dev,u_long cmd,void * addr,int flag,struct lwp * l)393 vc_nb_ioctl(dev_t dev, u_long cmd, void *addr, int flag,
394     struct lwp *l)
395 {
396     ENTRY;
397 
398     switch (cmd) {
399     case CODARESIZE: {
400 	struct coda_resize *data = (struct coda_resize *)addr;
401 	return(coda_nc_resize(data->hashsize, data->heapsize, IS_DOWNCALL));
402 	break;
403     }
404     case CODASTATS:
405 	if (coda_nc_use) {
406 	    coda_nc_gather_stats();
407 	    return(0);
408 	} else {
409 	    return(ENODEV);
410 	}
411 	break;
412     case CODAPRINT:
413 	if (coda_nc_use) {
414 	    print_coda_nc();
415 	    return(0);
416 	} else {
417 	    return(ENODEV);
418 	}
419 	break;
420     case CIOC_KERNEL_VERSION:
421 	switch (*(u_int *)addr) {
422 	case 0:
423 		*(u_int *)addr = coda_kernel_version;
424 		return 0;
425 		break;
426 	case 1:
427 	case 2:
428 		if (coda_kernel_version != *(u_int *)addr)
429 		    return ENOENT;
430 		else
431 		    return 0;
432 	default:
433 		return ENOENT;
434 	}
435     	break;
436     default :
437 	return(EINVAL);
438 	break;
439     }
440 }
441 
442 int
vc_nb_poll(dev_t dev,int events,struct lwp * l)443 vc_nb_poll(dev_t dev, int events, struct lwp *l)
444 {
445     struct vcomm *vcp;
446     int event_msk = 0;
447 
448     ENTRY;
449 
450     if (minor(dev) >= NVCODA)
451 	return(ENXIO);
452 
453     vcp = &coda_mnttbl[minor(dev)].mi_vcomm;
454 
455     event_msk = events & (POLLIN|POLLRDNORM);
456     if (!event_msk)
457 	return(0);
458 
459     if (!TAILQ_EMPTY(&vcp->vc_requests))
460 	return(events & (POLLIN|POLLRDNORM));
461 
462     selrecord(l, &(vcp->vc_selproc));
463 
464     return(0);
465 }
466 
467 static void
filt_vc_nb_detach(struct knote * kn)468 filt_vc_nb_detach(struct knote *kn)
469 {
470 	struct vcomm *vcp = kn->kn_hook;
471 
472 	SLIST_REMOVE(&vcp->vc_selproc.sel_klist, kn, knote, kn_selnext);
473 }
474 
475 static int
filt_vc_nb_read(struct knote * kn,long hint)476 filt_vc_nb_read(struct knote *kn, long hint)
477 {
478 	struct vcomm *vcp = kn->kn_hook;
479 	struct vmsg *vmp;
480 
481 	vmp = TAILQ_FIRST(&vcp->vc_requests);
482 	if (vmp == NULL)
483 		return (0);
484 
485 	kn->kn_data = vmp->vm_inSize;
486 	return (1);
487 }
488 
489 static const struct filterops vc_nb_read_filtops =
490 	{ 1, NULL, filt_vc_nb_detach, filt_vc_nb_read };
491 
492 int
vc_nb_kqfilter(dev_t dev,struct knote * kn)493 vc_nb_kqfilter(dev_t dev, struct knote *kn)
494 {
495 	struct vcomm *vcp;
496 	struct klist *klist;
497 
498 	ENTRY;
499 
500 	if (minor(dev) >= NVCODA)
501 		return(ENXIO);
502 
503 	vcp = &coda_mnttbl[minor(dev)].mi_vcomm;
504 
505 	switch (kn->kn_filter) {
506 	case EVFILT_READ:
507 		klist = &vcp->vc_selproc.sel_klist;
508 		kn->kn_fop = &vc_nb_read_filtops;
509 		break;
510 
511 	default:
512 		return (EINVAL);
513 	}
514 
515 	kn->kn_hook = vcp;
516 
517 	SLIST_INSERT_HEAD(klist, kn, kn_selnext);
518 
519 	return (0);
520 }
521 
522 /*
523  * Statistics
524  */
525 struct coda_clstat coda_clstat;
526 
527 /*
528  * Key question: whether to sleep interruptably or uninterruptably when
529  * waiting for Venus.  The former seems better (cause you can ^C a
530  * job), but then GNU-EMACS completion breaks. Use tsleep with no
531  * timeout, and no longjmp happens. But, when sleeping
532  * "uninterruptibly", we don't get told if it returns abnormally
533  * (e.g. kill -9).
534  */
535 
536 int
coda_call(struct coda_mntinfo * mntinfo,int inSize,int * outSize,void * buffer)537 coda_call(struct coda_mntinfo *mntinfo, int inSize, int *outSize,
538 	void *buffer)
539 {
540 	struct vcomm *vcp;
541 	struct vmsg *vmp;
542 	int error;
543 #ifdef	CTL_C
544 	struct lwp *l = curlwp;
545 	struct proc *p = l->l_proc;
546 	sigset_t psig_omask;
547 	int i;
548 	psig_omask = l->l_sigmask;	/* XXXSA */
549 #endif
550 	if (mntinfo == NULL) {
551 	    /* Unlikely, but could be a race condition with a dying warden */
552 	    return ENODEV;
553 	}
554 
555 	vcp = &(mntinfo->mi_vcomm);
556 
557 	coda_clstat.ncalls++;
558 	coda_clstat.reqs[((struct coda_in_hdr *)buffer)->opcode]++;
559 
560 	if (!VC_OPEN(vcp))
561 	    return(ENODEV);
562 
563 	CODA_ALLOC(vmp,struct vmsg *,sizeof(struct vmsg));
564 	/* Format the request message. */
565 	vmp->vm_data = buffer;
566 	vmp->vm_flags = 0;
567 	vmp->vm_inSize = inSize;
568 	vmp->vm_outSize
569 	    = *outSize ? *outSize : inSize; /* |buffer| >= inSize */
570 	vmp->vm_opcode = ((struct coda_in_hdr *)buffer)->opcode;
571 	vmp->vm_unique = ++vcp->vc_seq;
572 	if (codadebug)
573 	    myprintf(("Doing a call for %d.%d\n",
574 		      vmp->vm_opcode, vmp->vm_unique));
575 
576 	/* Fill in the common input args. */
577 	((struct coda_in_hdr *)buffer)->unique = vmp->vm_unique;
578 
579 	/* Append msg to request queue and poke Venus. */
580 	TAILQ_INSERT_TAIL(&vcp->vc_requests, vmp, vm_chain);
581 	selnotify(&(vcp->vc_selproc), 0, 0);
582 
583 	/* We can be interrupted while we wait for Venus to process
584 	 * our request.  If the interrupt occurs before Venus has read
585 	 * the request, we dequeue and return. If it occurs after the
586 	 * read but before the reply, we dequeue, send a signal
587 	 * message, and return. If it occurs after the reply we ignore
588 	 * it. In no case do we want to restart the syscall.  If it
589 	 * was interrupted by a venus shutdown (vcclose), return
590 	 * ENODEV.  */
591 
592 	/* Ignore return, We have to check anyway */
593 #ifdef	CTL_C
594 	/* This is work in progress.  Setting coda_pcatch lets tsleep reawaken
595 	   on a ^c or ^z.  The problem is that emacs sets certain interrupts
596 	   as SA_RESTART.  This means that we should exit sleep handle the
597 	   "signal" and then go to sleep again.  Mostly this is done by letting
598 	   the syscall complete and be restarted.  We are not idempotent and
599 	   can not do this.  A better solution is necessary.
600 	 */
601 	i = 0;
602 	do {
603 	    error = tsleep(&vmp->vm_sleep, (coda_call_sleep|coda_pcatch), "coda_call", hz*2);
604 	    if (error == 0)
605 	    	break;
606 	    mutex_enter(p->p_lock);
607 	    if (error == EWOULDBLOCK) {
608 #ifdef	CODA_VERBOSE
609 		    printf("coda_call: tsleep TIMEOUT %d sec\n", 2+2*i);
610 #endif
611     	    } else if (sigispending(l, SIGIO)) {
612 		    sigaddset(&l->l_sigmask, SIGIO);
613 #ifdef	CODA_VERBOSE
614 		    printf("coda_call: tsleep returns %d SIGIO, cnt %d\n", error, i);
615 #endif
616     	    } else if (sigispending(l, SIGALRM)) {
617 		    sigaddset(&l->l_sigmask, SIGALRM);
618 #ifdef	CODA_VERBOSE
619 		    printf("coda_call: tsleep returns %d SIGALRM, cnt %d\n", error, i);
620 #endif
621 	    } else {
622 		    sigset_t tmp;
623 		    tmp = p->p_sigpend.sp_set;	/* array assignment */
624 		    sigminusset(&l->l_sigmask, &tmp);
625 
626 #ifdef	CODA_VERBOSE
627 		    printf("coda_call: tsleep returns %d, cnt %d\n", error, i);
628 		    printf("coda_call: siglist = %x.%x.%x.%x, sigmask = %x.%x.%x.%x, mask %x.%x.%x.%x\n",
629 			    p->p_sigpend.sp_set.__bits[0], p->p_sigpend.sp_set.__bits[1],
630 			    p->p_sigpend.sp_set.__bits[2], p->p_sigpend.sp_set.__bits[3],
631 			    l->l_sigmask.__bits[0], l->l_sigmask.__bits[1],
632 			    l->l_sigmask.__bits[2], l->l_sigmask.__bits[3],
633 			    tmp.__bits[0], tmp.__bits[1], tmp.__bits[2], tmp.__bits[3]);
634 #endif
635 		    mutex_exit(p->p_lock);
636 		    break;
637 #ifdef	notyet
638 		    sigminusset(&l->l_sigmask, &p->p_sigpend.sp_set);
639 		    printf("coda_call: siglist = %x.%x.%x.%x, sigmask = %x.%x.%x.%x\n",
640 			    p->p_sigpend.sp_set.__bits[0], p->p_sigpend.sp_set.__bits[1],
641 			    p->p_sigpend.sp_set.__bits[2], p->p_sigpend.sp_set.__bits[3],
642 			    l->l_sigmask.__bits[0], l->l_sigmask.__bits[1],
643 			    l->l_sigmask.__bits[2], l->l_sigmask.__bits[3]);
644 #endif
645 	    }
646 	    mutex_exit(p->p_lock);
647 	} while (error && i++ < 128 && VC_OPEN(vcp));
648 	l->l_sigmask = psig_omask;	/* XXXSA */
649 #else
650 	(void) tsleep(&vmp->vm_sleep, coda_call_sleep, "coda_call", 0);
651 #endif
652 	if (VC_OPEN(vcp)) {	/* Venus is still alive */
653  	/* Op went through, interrupt or not... */
654 	    if (vmp->vm_flags & VM_WRITE) {
655 		error = 0;
656 		*outSize = vmp->vm_outSize;
657 	    }
658 
659 	    else if (!(vmp->vm_flags & VM_READ)) {
660 		/* Interrupted before venus read it. */
661 #ifdef	CODA_VERBOSE
662 		if (1)
663 #else
664 		if (codadebug)
665 #endif
666 		    myprintf(("interrupted before read: op = %d.%d, flags = %x\n",
667 			   vmp->vm_opcode, vmp->vm_unique, vmp->vm_flags));
668 
669 		TAILQ_REMOVE(&vcp->vc_requests, vmp, vm_chain);
670 		error = EINTR;
671 	    }
672 
673 	    else {
674 		/* (!(vmp->vm_flags & VM_WRITE)) means interrupted after
675                    upcall started */
676 		/* Interrupted after start of upcall, send venus a signal */
677 		struct coda_in_hdr *dog;
678 		struct vmsg *svmp;
679 
680 #ifdef	CODA_VERBOSE
681 		if (1)
682 #else
683 		if (codadebug)
684 #endif
685 		    myprintf(("Sending Venus a signal: op = %d.%d, flags = %x\n",
686 			   vmp->vm_opcode, vmp->vm_unique, vmp->vm_flags));
687 
688 		TAILQ_REMOVE(&vcp->vc_replies, vmp, vm_chain);
689 		error = EINTR;
690 
691 		CODA_ALLOC(svmp, struct vmsg *, sizeof (struct vmsg));
692 
693 		CODA_ALLOC((svmp->vm_data), char *, sizeof (struct coda_in_hdr));
694 		dog = (struct coda_in_hdr *)svmp->vm_data;
695 
696 		svmp->vm_flags = 0;
697 		dog->opcode = svmp->vm_opcode = CODA_SIGNAL;
698 		dog->unique = svmp->vm_unique = vmp->vm_unique;
699 		svmp->vm_inSize = sizeof (struct coda_in_hdr);
700 /*??? rvb */	svmp->vm_outSize = sizeof (struct coda_in_hdr);
701 
702 		if (codadebug)
703 		    myprintf(("coda_call: enqueing signal msg (%d, %d)\n",
704 			   svmp->vm_opcode, svmp->vm_unique));
705 
706 		/* insert at head of queue */
707 		TAILQ_INSERT_HEAD(&vcp->vc_requests, svmp, vm_chain);
708 		selnotify(&(vcp->vc_selproc), 0, 0);
709 	    }
710 	}
711 
712 	else {	/* If venus died (!VC_OPEN(vcp)) */
713 	    if (codadebug)
714 		myprintf(("vcclose woke op %d.%d flags %d\n",
715 		       vmp->vm_opcode, vmp->vm_unique, vmp->vm_flags));
716 
717 		error = ENODEV;
718 	}
719 
720 	CODA_FREE(vmp, sizeof(struct vmsg));
721 
722 	if (outstanding_upcalls > 0 && (--outstanding_upcalls == 0))
723 		wakeup(&outstanding_upcalls);
724 
725 	if (!error)
726 		error = ((struct coda_out_hdr *)buffer)->result;
727 	return(error);
728 }
729 
730 MODULE(MODULE_CLASS_DRIVER, vcoda, NULL);
731 
732 static int
vcoda_modcmd(modcmd_t cmd,void * arg)733 vcoda_modcmd(modcmd_t cmd, void *arg)
734 {
735 	int error = 0;
736 
737 	switch (cmd) {
738 	case MODULE_CMD_INIT:
739 #ifdef _MODULE
740 	{
741 		int cmajor, dmajor;
742 		vcodaattach(NVCODA);
743 
744 		dmajor = cmajor = -1;
745 		return devsw_attach("vcoda", NULL, &dmajor,
746 		    &vcoda_cdevsw, &cmajor);
747 	}
748 #endif
749 		break;
750 
751 	case MODULE_CMD_FINI:
752 #ifdef _MODULE
753 		{
754 			for  (size_t i = 0; i < NVCODA; i++) {
755 				struct vcomm *vcp = &coda_mnttbl[i].mi_vcomm;
756 				if (VC_OPEN(vcp))
757 					return EBUSY;
758 			}
759 			return devsw_detach(NULL, &vcoda_cdevsw);
760 		}
761 #endif
762 		break;
763 
764 	case MODULE_CMD_STAT:
765 		return ENOTTY;
766 
767 	default:
768 		return ENOTTY;
769 	}
770 	return error;
771 }
772