1 /*	$NetBSD: sys_mqueue.c,v 1.39 2015/06/29 15:44:45 christos Exp $	*/
2 
3 /*
4  * Copyright (c) 2007-2011 Mindaugas Rasiukevicius <rmind at NetBSD org>
5  * All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions
9  * are met:
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  *
16  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
17  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
18  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
19  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
20  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
21  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
22  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
23  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
24  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
25  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
26  * SUCH DAMAGE.
27  */
28 
29 /*
30  * Implementation of POSIX message queues.
31  * Defined in the Base Definitions volume of IEEE Std 1003.1-2001.
32  *
33  * Locking
34  *
35  * Global list of message queues (mqueue_head) is protected by mqlist_lock.
36  * Each message queue and its members are protected by mqueue::mq_mtx.
37  * Note that proc_t::p_mqueue_cnt is updated atomically.
38  *
39  * Lock order:
40  *
41  *	mqlist_lock ->
42  *		mqueue::mq_mtx
43  */
44 
45 #include <sys/cdefs.h>
46 __KERNEL_RCSID(0, "$NetBSD: sys_mqueue.c,v 1.39 2015/06/29 15:44:45 christos Exp $");
47 
48 #include <sys/param.h>
49 #include <sys/types.h>
50 #include <sys/atomic.h>
51 
52 #include <sys/file.h>
53 #include <sys/filedesc.h>
54 #include <sys/kauth.h>
55 #include <sys/lwp.h>
56 #include <sys/mqueue.h>
57 #include <sys/module.h>
58 #include <sys/poll.h>
59 #include <sys/select.h>
60 #include <sys/signal.h>
61 #include <sys/signalvar.h>
62 #include <sys/stat.h>
63 #include <sys/sysctl.h>
64 #include <sys/syscall.h>
65 #include <sys/syscallvar.h>
66 #include <sys/syscallargs.h>
67 
68 #include <miscfs/genfs/genfs.h>
69 
70 MODULE(MODULE_CLASS_MISC, mqueue, NULL);
71 
72 /* System-wide limits. */
73 static u_int			mq_open_max = MQ_OPEN_MAX;
74 static u_int			mq_prio_max = MQ_PRIO_MAX;
75 static u_int			mq_max_msgsize = 16 * MQ_DEF_MSGSIZE;
76 static u_int			mq_def_maxmsg = 32;
77 static u_int			mq_max_maxmsg = 16 * 32;
78 
79 static pool_cache_t		mqmsg_cache	__read_mostly;
80 static kmutex_t			mqlist_lock	__cacheline_aligned;
81 static LIST_HEAD(, mqueue)	mqueue_head	__cacheline_aligned;
82 static struct sysctllog *	mqsysctl_log;
83 
84 static kauth_listener_t		mq_listener;
85 
86 static int	mqueue_sysinit(void);
87 static int	mqueue_sysfini(bool);
88 static int	mqueue_sysctl_init(void);
89 static int	mq_poll_fop(file_t *, int);
90 static int	mq_stat_fop(file_t *, struct stat *);
91 static int	mq_close_fop(file_t *);
92 
93 static const struct fileops mqops = {
94 	.fo_read = fbadop_read,
95 	.fo_write = fbadop_write,
96 	.fo_ioctl = fbadop_ioctl,
97 	.fo_fcntl = fnullop_fcntl,
98 	.fo_poll = mq_poll_fop,
99 	.fo_stat = mq_stat_fop,
100 	.fo_close = mq_close_fop,
101 	.fo_kqfilter = fnullop_kqfilter,
102 	.fo_restart = fnullop_restart,
103 };
104 
105 static const struct syscall_package mqueue_syscalls[] = {
106 	{ SYS_mq_open, 0, (sy_call_t *)sys_mq_open },
107 	{ SYS_mq_close, 0, (sy_call_t *)sys_mq_close },
108 	{ SYS_mq_unlink, 0, (sy_call_t *)sys_mq_unlink },
109 	{ SYS_mq_getattr, 0, (sy_call_t *)sys_mq_getattr },
110 	{ SYS_mq_setattr, 0, (sy_call_t *)sys_mq_setattr },
111 	{ SYS_mq_notify, 0, (sy_call_t *)sys_mq_notify },
112 	{ SYS_mq_send, 0, (sy_call_t *)sys_mq_send },
113 	{ SYS_mq_receive, 0, (sy_call_t *)sys_mq_receive },
114 	{ SYS___mq_timedsend50, 0, (sy_call_t *)sys___mq_timedsend50 },
115 	{ SYS___mq_timedreceive50, 0, (sy_call_t *)sys___mq_timedreceive50 },
116 	{ 0, 0, NULL }
117 };
118 
119 static int
mq_listener_cb(kauth_cred_t cred,kauth_action_t action,void * cookie,void * arg0,void * arg1,void * arg2,void * arg3)120 mq_listener_cb(kauth_cred_t cred, kauth_action_t action, void *cookie,
121     void *arg0, void *arg1, void *arg2, void *arg3)
122 {
123 	mqueue_t *mq;
124 	int result;
125 
126 	if (action != KAUTH_SYSTEM_MQUEUE)
127 		return KAUTH_RESULT_DEFER;
128 
129 	result = KAUTH_RESULT_DEFER;
130 
131 	mq = arg1;
132 
133 	if (kauth_cred_geteuid(cred) == mq->mq_euid)
134 		result = KAUTH_RESULT_ALLOW;
135 
136 	return result;
137 }
138 
139 /*
140  * Initialisation and unloading of POSIX message queue subsystem.
141  */
142 
143 static int
mqueue_sysinit(void)144 mqueue_sysinit(void)
145 {
146 	int error;
147 
148 	mqmsg_cache = pool_cache_init(MQ_DEF_MSGSIZE, coherency_unit,
149 	    0, 0, "mqmsgpl", NULL, IPL_NONE, NULL, NULL, NULL);
150 	mutex_init(&mqlist_lock, MUTEX_DEFAULT, IPL_NONE);
151 	LIST_INIT(&mqueue_head);
152 
153 	error = mqueue_sysctl_init();
154 	if (error) {
155 		(void)mqueue_sysfini(false);
156 		return error;
157 	}
158 	error = syscall_establish(NULL, mqueue_syscalls);
159 	if (error) {
160 		(void)mqueue_sysfini(false);
161 	}
162 	mq_listener = kauth_listen_scope(KAUTH_SCOPE_SYSTEM,
163 	    mq_listener_cb, NULL);
164 	return error;
165 }
166 
167 static int
mqueue_sysfini(bool interface)168 mqueue_sysfini(bool interface)
169 {
170 
171 	if (interface) {
172 		int error;
173 		bool inuse;
174 
175 		/* Stop syscall activity. */
176 		error = syscall_disestablish(NULL, mqueue_syscalls);
177 		if (error)
178 			return error;
179 		/* Check if there are any message queues in use. */
180 		mutex_enter(&mqlist_lock);
181 		inuse = !LIST_EMPTY(&mqueue_head);
182 		mutex_exit(&mqlist_lock);
183 		if (inuse) {
184 			error = syscall_establish(NULL, mqueue_syscalls);
185 			KASSERT(error == 0);
186 			return EBUSY;
187 		}
188 	}
189 
190 	if (mqsysctl_log != NULL)
191 		sysctl_teardown(&mqsysctl_log);
192 
193 	kauth_unlisten_scope(mq_listener);
194 
195 	mutex_destroy(&mqlist_lock);
196 	pool_cache_destroy(mqmsg_cache);
197 	return 0;
198 }
199 
200 /*
201  * Module interface.
202  */
203 static int
mqueue_modcmd(modcmd_t cmd,void * arg)204 mqueue_modcmd(modcmd_t cmd, void *arg)
205 {
206 
207 	switch (cmd) {
208 	case MODULE_CMD_INIT:
209 		return mqueue_sysinit();
210 	case MODULE_CMD_FINI:
211 		return mqueue_sysfini(true);
212 	default:
213 		return ENOTTY;
214 	}
215 }
216 
217 /*
218  * Free the message.
219  */
220 static void
mqueue_freemsg(struct mq_msg * msg,const size_t size)221 mqueue_freemsg(struct mq_msg *msg, const size_t size)
222 {
223 
224 	if (size > MQ_DEF_MSGSIZE) {
225 		kmem_free(msg, size);
226 	} else {
227 		pool_cache_put(mqmsg_cache, msg);
228 	}
229 }
230 
231 /*
232  * Destroy the message queue.
233  */
234 static void
mqueue_destroy(struct mqueue * mq)235 mqueue_destroy(struct mqueue *mq)
236 {
237 	struct mq_msg *msg;
238 	size_t msz;
239 	u_int i;
240 
241 	/* Note MQ_PQSIZE + 1. */
242 	for (i = 0; i <= MQ_PQSIZE; i++) {
243 		while ((msg = TAILQ_FIRST(&mq->mq_head[i])) != NULL) {
244 			TAILQ_REMOVE(&mq->mq_head[i], msg, msg_queue);
245 			msz = sizeof(struct mq_msg) + msg->msg_len;
246 			mqueue_freemsg(msg, msz);
247 		}
248 	}
249 	if (mq->mq_name) {
250 		kmem_free(mq->mq_name, MQ_NAMELEN);
251 	}
252 	seldestroy(&mq->mq_rsel);
253 	seldestroy(&mq->mq_wsel);
254 	cv_destroy(&mq->mq_send_cv);
255 	cv_destroy(&mq->mq_recv_cv);
256 	mutex_destroy(&mq->mq_mtx);
257 	kmem_free(mq, sizeof(struct mqueue));
258 }
259 
260 /*
261  * mqueue_lookup: lookup for file name in general list of message queues.
262  *
263  * => locks the message queue on success
264  */
265 static mqueue_t *
mqueue_lookup(const char * name)266 mqueue_lookup(const char *name)
267 {
268 	mqueue_t *mq;
269 
270 	KASSERT(mutex_owned(&mqlist_lock));
271 
272 	LIST_FOREACH(mq, &mqueue_head, mq_list) {
273 		if (strncmp(mq->mq_name, name, MQ_NAMELEN) == 0) {
274 			mutex_enter(&mq->mq_mtx);
275 			return mq;
276 		}
277 	}
278 	return NULL;
279 }
280 
281 /*
282  * mqueue_get: get the mqueue from the descriptor.
283  *
284  * => locks the message queue, if found.
285  * => holds a reference on the file descriptor.
286  */
287 int
mqueue_get(mqd_t mqd,int fflag,mqueue_t ** mqret)288 mqueue_get(mqd_t mqd, int fflag, mqueue_t **mqret)
289 {
290 	const int fd = (int)mqd;
291 	mqueue_t *mq;
292 	file_t *fp;
293 
294 	fp = fd_getfile(fd);
295 	if (__predict_false(fp == NULL)) {
296 		return EBADF;
297 	}
298 	if (__predict_false(fp->f_type != DTYPE_MQUEUE)) {
299 		fd_putfile(fd);
300 		return EBADF;
301 	}
302 	if (fflag && (fp->f_flag & fflag) == 0) {
303 		fd_putfile(fd);
304 		return EBADF;
305 	}
306 	mq = fp->f_mqueue;
307 	mutex_enter(&mq->mq_mtx);
308 
309 	*mqret = mq;
310 	return 0;
311 }
312 
313 /*
314  * mqueue_linear_insert: perform linear insert according to the message
315  * priority into the reserved queue (MQ_PQRESQ).  Reserved queue is a
316  * sorted list used only when mq_prio_max is increased via sysctl.
317  */
318 static inline void
mqueue_linear_insert(struct mqueue * mq,struct mq_msg * msg)319 mqueue_linear_insert(struct mqueue *mq, struct mq_msg *msg)
320 {
321 	struct mq_msg *mit;
322 
323 	TAILQ_FOREACH(mit, &mq->mq_head[MQ_PQRESQ], msg_queue) {
324 		if (msg->msg_prio > mit->msg_prio)
325 			break;
326 	}
327 	if (mit == NULL) {
328 		TAILQ_INSERT_TAIL(&mq->mq_head[MQ_PQRESQ], msg, msg_queue);
329 	} else {
330 		TAILQ_INSERT_BEFORE(mit, msg, msg_queue);
331 	}
332 }
333 
334 static int
mq_stat_fop(file_t * fp,struct stat * st)335 mq_stat_fop(file_t *fp, struct stat *st)
336 {
337 	struct mqueue *mq = fp->f_mqueue;
338 
339 	memset(st, 0, sizeof(*st));
340 
341 	mutex_enter(&mq->mq_mtx);
342 	st->st_mode = mq->mq_mode;
343 	st->st_uid = mq->mq_euid;
344 	st->st_gid = mq->mq_egid;
345 	st->st_atimespec = mq->mq_atime;
346 	st->st_mtimespec = mq->mq_mtime;
347 	st->st_ctimespec = st->st_birthtimespec = mq->mq_btime;
348 	st->st_uid = kauth_cred_geteuid(fp->f_cred);
349 	st->st_gid = kauth_cred_getegid(fp->f_cred);
350 	mutex_exit(&mq->mq_mtx);
351 
352 	return 0;
353 }
354 
355 static int
mq_poll_fop(file_t * fp,int events)356 mq_poll_fop(file_t *fp, int events)
357 {
358 	struct mqueue *mq = fp->f_mqueue;
359 	struct mq_attr *mqattr;
360 	int revents = 0;
361 
362 	mutex_enter(&mq->mq_mtx);
363 	mqattr = &mq->mq_attrib;
364 	if (events & (POLLIN | POLLRDNORM)) {
365 		/* Ready for receiving, if there are messages in the queue. */
366 		if (mqattr->mq_curmsgs)
367 			revents |= events & (POLLIN | POLLRDNORM);
368 		else
369 			selrecord(curlwp, &mq->mq_rsel);
370 	}
371 	if (events & (POLLOUT | POLLWRNORM)) {
372 		/* Ready for sending, if the message queue is not full. */
373 		if (mqattr->mq_curmsgs < mqattr->mq_maxmsg)
374 			revents |= events & (POLLOUT | POLLWRNORM);
375 		else
376 			selrecord(curlwp, &mq->mq_wsel);
377 	}
378 	mutex_exit(&mq->mq_mtx);
379 
380 	return revents;
381 }
382 
383 static int
mq_close_fop(file_t * fp)384 mq_close_fop(file_t *fp)
385 {
386 	proc_t *p = curproc;
387 	mqueue_t *mq = fp->f_mqueue;
388 	bool destroy = false;
389 
390 	mutex_enter(&mq->mq_mtx);
391 	KASSERT(mq->mq_refcnt > 0);
392 	if (--mq->mq_refcnt == 0) {
393 		/* Destroy if the last reference and unlinked. */
394 		destroy = (mq->mq_attrib.mq_flags & MQ_UNLINKED) != 0;
395 	}
396 	mutex_exit(&mq->mq_mtx);
397 
398 	if (destroy) {
399 		mqueue_destroy(mq);
400 	}
401 	atomic_dec_uint(&p->p_mqueue_cnt);
402 	return 0;
403 }
404 
405 static int
mqueue_access(mqueue_t * mq,int access,kauth_cred_t cred)406 mqueue_access(mqueue_t *mq, int access, kauth_cred_t cred)
407 {
408 	mode_t acc_mode = 0;
409 
410 	/* Note the difference between VREAD/VWRITE and FREAD/FWRITE. */
411 	if (access & FREAD) {
412 		acc_mode |= VREAD;
413 	}
414 	if (access & FWRITE) {
415 		acc_mode |= VWRITE;
416 	}
417 	if (genfs_can_access(VNON, mq->mq_mode, mq->mq_euid,
418 	    mq->mq_egid, acc_mode, cred)) {
419 		return EACCES;
420 	}
421 	return 0;
422 }
423 
424 static int
mqueue_create(lwp_t * l,char * name,struct mq_attr * attr,mode_t mode,int oflag,mqueue_t ** mqret)425 mqueue_create(lwp_t *l, char *name, struct mq_attr *attr, mode_t mode,
426     int oflag, mqueue_t **mqret)
427 {
428 	proc_t *p = l->l_proc;
429 	struct cwdinfo *cwdi = p->p_cwdi;
430 	mqueue_t *mq;
431 	u_int i;
432 
433 	/* Pre-check the limit. */
434 	if (p->p_mqueue_cnt >= mq_open_max) {
435 		return EMFILE;
436 	}
437 
438 	/* Empty name is invalid. */
439 	if (name[0] == '\0') {
440 		return EINVAL;
441 	}
442 
443 	/* Check for mqueue attributes. */
444 	if (attr) {
445 		if (attr->mq_maxmsg <= 0 || attr->mq_maxmsg > mq_max_maxmsg ||
446 		    attr->mq_msgsize <= 0 ||
447 		    attr->mq_msgsize > mq_max_msgsize) {
448 			return EINVAL;
449 		}
450 		attr->mq_curmsgs = 0;
451 	}
452 
453 	/*
454 	 * Allocate new message queue, initialize data structures, copy the
455 	 * name attributes.  Note that the initial reference is set here.
456 	 */
457 	mq = kmem_zalloc(sizeof(mqueue_t), KM_SLEEP);
458 
459 	mutex_init(&mq->mq_mtx, MUTEX_DEFAULT, IPL_NONE);
460 	cv_init(&mq->mq_send_cv, "mqsendcv");
461 	cv_init(&mq->mq_recv_cv, "mqrecvcv");
462 	for (i = 0; i < (MQ_PQSIZE + 1); i++) {
463 		TAILQ_INIT(&mq->mq_head[i]);
464 	}
465 	selinit(&mq->mq_rsel);
466 	selinit(&mq->mq_wsel);
467 	mq->mq_name = name;
468 	mq->mq_refcnt = 1;
469 
470 	if (attr != NULL) {
471 		memcpy(&mq->mq_attrib, attr, sizeof(struct mq_attr));
472 	} else {
473 		memset(&mq->mq_attrib, 0, sizeof(struct mq_attr));
474 		mq->mq_attrib.mq_maxmsg = mq_def_maxmsg;
475 		mq->mq_attrib.mq_msgsize = MQ_DEF_MSGSIZE - sizeof(struct mq_msg);
476 	}
477 
478 	CTASSERT((O_MASK & (MQ_UNLINKED | MQ_RECEIVE)) == 0);
479 	mq->mq_attrib.mq_flags = (O_MASK & oflag);
480 
481 	/* Store mode and effective UID with GID. */
482 	mq->mq_mode = ((mode & ~cwdi->cwdi_cmask) & ALLPERMS) & ~S_ISTXT;
483 	mq->mq_euid = kauth_cred_geteuid(l->l_cred);
484 	mq->mq_egid = kauth_cred_getegid(l->l_cred);
485 
486 	*mqret = mq;
487 	return 0;
488 }
489 
490 /*
491  * Helper function for mq_open() - note that "u_name" is a userland pointer,
492  * while "attr" is a kernel pointer!
493  */
494 int
mq_handle_open(struct lwp * l,const char * u_name,int oflag,mode_t mode,struct mq_attr * attr,register_t * retval)495 mq_handle_open(struct lwp *l, const char *u_name, int oflag, mode_t mode,
496     struct mq_attr *attr, register_t *retval)
497 {
498 	struct proc *p = l->l_proc;
499 	struct mqueue *mq, *mq_new = NULL;
500 	int mqd, error;
501 	file_t *fp;
502 	char *name;
503 
504 	/* Get the name from the user-space. */
505 	name = kmem_alloc(MQ_NAMELEN, KM_SLEEP);
506 	error = copyinstr(u_name, name, MQ_NAMELEN - 1, NULL);
507 	if (error) {
508 		kmem_free(name, MQ_NAMELEN);
509 		return error;
510 	}
511 
512 	/* Allocate file structure and descriptor. */
513 	error = fd_allocfile(&fp, &mqd);
514 	if (error) {
515 		kmem_free(name, MQ_NAMELEN);
516 		return error;
517 	}
518 	fp->f_type = DTYPE_MQUEUE;
519 	fp->f_flag = FFLAGS(oflag) & (FREAD | FWRITE);
520 	fp->f_ops = &mqops;
521 
522 	if (oflag & O_CREAT) {
523 		/* Create a new message queue. */
524 		error = mqueue_create(l, name, attr, mode, oflag, &mq_new);
525 		if (error) {
526 			goto err;
527 		}
528 		KASSERT(mq_new != NULL);
529 	}
530 
531 	/* Lookup for a message queue with such name. */
532 	mutex_enter(&mqlist_lock);
533 	mq = mqueue_lookup(name);
534 	if (mq) {
535 		KASSERT(mutex_owned(&mq->mq_mtx));
536 		mutex_exit(&mqlist_lock);
537 
538 		/* Check for exclusive create. */
539 		if (oflag & O_EXCL) {
540 			mutex_exit(&mq->mq_mtx);
541 			error = EEXIST;
542 			goto err;
543 		}
544 
545 		/* Verify permissions. */
546 		if (mqueue_access(mq, fp->f_flag, l->l_cred) != 0) {
547 			mutex_exit(&mq->mq_mtx);
548 			error = EACCES;
549 			goto err;
550 		}
551 
552 		/* If we have the access, add a new reference. */
553 		mq->mq_refcnt++;
554 		mutex_exit(&mq->mq_mtx);
555 	} else {
556 		/* Fail if not found and not creating. */
557 		if ((oflag & O_CREAT) == 0) {
558 			mutex_exit(&mqlist_lock);
559 			KASSERT(mq_new == NULL);
560 			error = ENOENT;
561 			goto err;
562 		}
563 
564 		/* Account and check for the limit. */
565 		if (atomic_inc_uint_nv(&p->p_mqueue_cnt) > mq_open_max) {
566 			mutex_exit(&mqlist_lock);
567 			atomic_dec_uint(&p->p_mqueue_cnt);
568 			error = EMFILE;
569 			goto err;
570 		}
571 
572 		/* Initial timestamps. */
573 		mq = mq_new;
574 		getnanotime(&mq->mq_btime);
575 		mq->mq_atime = mq->mq_mtime = mq->mq_btime;
576 
577 		/*
578 		 * Finally, insert message queue into the list.
579 		 * Note: it already has the initial reference.
580 		 */
581 		LIST_INSERT_HEAD(&mqueue_head, mq, mq_list);
582 		mutex_exit(&mqlist_lock);
583 
584 		mq_new = NULL;
585 		name = NULL;
586 	}
587 	KASSERT(mq != NULL);
588 	fp->f_mqueue = mq;
589 	fd_affix(p, fp, mqd);
590 	*retval = mqd;
591 err:
592 	if (error) {
593 		fd_abort(p, fp, mqd);
594 	}
595 	if (mq_new) {
596 		/* Note: will free the 'name'. */
597 		mqueue_destroy(mq_new);
598 	} else if (name) {
599 		kmem_free(name, MQ_NAMELEN);
600 	}
601 	return error;
602 }
603 
604 /*
605  * General mqueue system calls.
606  */
607 
608 int
sys_mq_open(struct lwp * l,const struct sys_mq_open_args * uap,register_t * retval)609 sys_mq_open(struct lwp *l, const struct sys_mq_open_args *uap,
610     register_t *retval)
611 {
612 	/* {
613 		syscallarg(const char *) name;
614 		syscallarg(int) oflag;
615 		syscallarg(mode_t) mode;
616 		syscallarg(struct mq_attr) attr;
617 	} */
618 	struct mq_attr *attr = NULL, a;
619 	int error;
620 
621 	if ((SCARG(uap, oflag) & O_CREAT) != 0 && SCARG(uap, attr) != NULL) {
622 		error = copyin(SCARG(uap, attr), &a, sizeof(a));
623 		if (error)
624 			return error;
625 		attr = &a;
626 	}
627 
628 	return mq_handle_open(l, SCARG(uap, name), SCARG(uap, oflag),
629 	    SCARG(uap, mode), attr, retval);
630 }
631 
632 int
sys_mq_close(struct lwp * l,const struct sys_mq_close_args * uap,register_t * retval)633 sys_mq_close(struct lwp *l, const struct sys_mq_close_args *uap,
634     register_t *retval)
635 {
636 
637 	return sys_close(l, (const void *)uap, retval);
638 }
639 
640 /*
641  * Primary mq_recv1() function.
642  */
643 int
mq_recv1(mqd_t mqdes,void * msg_ptr,size_t msg_len,u_int * msg_prio,struct timespec * ts,ssize_t * mlen)644 mq_recv1(mqd_t mqdes, void *msg_ptr, size_t msg_len, u_int *msg_prio,
645     struct timespec *ts, ssize_t *mlen)
646 {
647 	struct mqueue *mq;
648 	struct mq_msg *msg = NULL;
649 	struct mq_attr *mqattr;
650 	u_int idx;
651 	int error;
652 
653 	error = mqueue_get(mqdes, FREAD, &mq);
654 	if (error) {
655 		return error;
656 	}
657 	getnanotime(&mq->mq_atime);
658 	mqattr = &mq->mq_attrib;
659 
660 	/* Check the message size limits */
661 	if (msg_len < mqattr->mq_msgsize) {
662 		error = EMSGSIZE;
663 		goto error;
664 	}
665 
666 	/* Check if queue is empty */
667 	while (mqattr->mq_curmsgs == 0) {
668 		int t;
669 
670 		if (mqattr->mq_flags & O_NONBLOCK) {
671 			error = EAGAIN;
672 			goto error;
673 		}
674 		if (ts) {
675 			error = ts2timo(CLOCK_REALTIME, TIMER_ABSTIME, ts, &t,
676 			    NULL);
677 			if (error)
678 				goto error;
679 		} else
680 			t = 0;
681 		/*
682 		 * Block until someone sends the message.
683 		 * While doing this, notification should not be sent.
684 		 */
685 		mqattr->mq_flags |= MQ_RECEIVE;
686 		error = cv_timedwait_sig(&mq->mq_send_cv, &mq->mq_mtx, t);
687 		mqattr->mq_flags &= ~MQ_RECEIVE;
688 		if (error || (mqattr->mq_flags & MQ_UNLINKED)) {
689 			error = (error == EWOULDBLOCK) ? ETIMEDOUT : EINTR;
690 			goto error;
691 		}
692 	}
693 
694 	/*
695 	 * Find the highest priority message, and remove it from the queue.
696 	 * At first, reserved queue is checked, bitmap is next.
697 	 */
698 	msg = TAILQ_FIRST(&mq->mq_head[MQ_PQRESQ]);
699 	if (__predict_true(msg == NULL)) {
700 		idx = ffs(mq->mq_bitmap);
701 		msg = TAILQ_FIRST(&mq->mq_head[idx]);
702 		KASSERT(msg != NULL);
703 	} else {
704 		idx = MQ_PQRESQ;
705 	}
706 	TAILQ_REMOVE(&mq->mq_head[idx], msg, msg_queue);
707 
708 	/* Unmark the bit, if last message. */
709 	if (__predict_true(idx) && TAILQ_EMPTY(&mq->mq_head[idx])) {
710 		KASSERT((MQ_PQSIZE - idx) == msg->msg_prio);
711 		mq->mq_bitmap &= ~(1 << --idx);
712 	}
713 
714 	/* Decrement the counter and signal waiter, if any */
715 	mqattr->mq_curmsgs--;
716 	cv_signal(&mq->mq_recv_cv);
717 
718 	/* Ready for sending now */
719 	selnotify(&mq->mq_wsel, POLLOUT | POLLWRNORM, 0);
720 error:
721 	mutex_exit(&mq->mq_mtx);
722 	fd_putfile((int)mqdes);
723 	if (error)
724 		return error;
725 
726 	/*
727 	 * Copy the data to the user-space.
728 	 * Note: According to POSIX, no message should be removed from the
729 	 * queue in case of fail - this would be violated.
730 	 */
731 	*mlen = msg->msg_len;
732 	error = copyout(msg->msg_ptr, msg_ptr, msg->msg_len);
733 	if (error == 0 && msg_prio)
734 		error = copyout(&msg->msg_prio, msg_prio, sizeof(unsigned));
735 	mqueue_freemsg(msg, sizeof(struct mq_msg) + msg->msg_len);
736 
737 	return error;
738 }
739 
740 int
sys_mq_receive(struct lwp * l,const struct sys_mq_receive_args * uap,register_t * retval)741 sys_mq_receive(struct lwp *l, const struct sys_mq_receive_args *uap,
742     register_t *retval)
743 {
744 	/* {
745 		syscallarg(mqd_t) mqdes;
746 		syscallarg(char *) msg_ptr;
747 		syscallarg(size_t) msg_len;
748 		syscallarg(unsigned *) msg_prio;
749 	} */
750 	ssize_t mlen;
751 	int error;
752 
753 	error = mq_recv1(SCARG(uap, mqdes), SCARG(uap, msg_ptr),
754 	    SCARG(uap, msg_len), SCARG(uap, msg_prio), NULL, &mlen);
755 	if (error == 0)
756 		*retval = mlen;
757 
758 	return error;
759 }
760 
761 int
sys___mq_timedreceive50(struct lwp * l,const struct sys___mq_timedreceive50_args * uap,register_t * retval)762 sys___mq_timedreceive50(struct lwp *l,
763     const struct sys___mq_timedreceive50_args *uap, register_t *retval)
764 {
765 	/* {
766 		syscallarg(mqd_t) mqdes;
767 		syscallarg(char *) msg_ptr;
768 		syscallarg(size_t) msg_len;
769 		syscallarg(unsigned *) msg_prio;
770 		syscallarg(const struct timespec *) abs_timeout;
771 	} */
772 	struct timespec ts, *tsp;
773 	ssize_t mlen;
774 	int error;
775 
776 	/* Get and convert time value */
777 	if (SCARG(uap, abs_timeout)) {
778 		error = copyin(SCARG(uap, abs_timeout), &ts, sizeof(ts));
779 		if (error)
780 			return error;
781 		tsp = &ts;
782 	} else {
783 		tsp = NULL;
784 	}
785 
786 	error = mq_recv1(SCARG(uap, mqdes), SCARG(uap, msg_ptr),
787 	    SCARG(uap, msg_len), SCARG(uap, msg_prio), tsp, &mlen);
788 	if (error == 0)
789 		*retval = mlen;
790 
791 	return error;
792 }
793 
794 /*
795  * Primary mq_send1() function.
796  */
797 int
mq_send1(mqd_t mqdes,const char * msg_ptr,size_t msg_len,u_int msg_prio,struct timespec * ts)798 mq_send1(mqd_t mqdes, const char *msg_ptr, size_t msg_len, u_int msg_prio,
799     struct timespec *ts)
800 {
801 	struct mqueue *mq;
802 	struct mq_msg *msg;
803 	struct mq_attr *mqattr;
804 	struct proc *notify = NULL;
805 	ksiginfo_t ksi;
806 	size_t size;
807 	int error;
808 
809 	/* Check the priority range */
810 	if (msg_prio >= mq_prio_max)
811 		return EINVAL;
812 
813 	/* Allocate a new message */
814 	size = sizeof(struct mq_msg) + msg_len;
815 	if (size > mq_max_msgsize)
816 		return EMSGSIZE;
817 
818 	if (size > MQ_DEF_MSGSIZE) {
819 		msg = kmem_alloc(size, KM_SLEEP);
820 	} else {
821 		msg = pool_cache_get(mqmsg_cache, PR_WAITOK);
822 	}
823 
824 	/* Get the data from user-space */
825 	error = copyin(msg_ptr, msg->msg_ptr, msg_len);
826 	if (error) {
827 		mqueue_freemsg(msg, size);
828 		return error;
829 	}
830 	msg->msg_len = msg_len;
831 	msg->msg_prio = msg_prio;
832 
833 	error = mqueue_get(mqdes, FWRITE, &mq);
834 	if (error) {
835 		mqueue_freemsg(msg, size);
836 		return error;
837 	}
838 	getnanotime(&mq->mq_mtime);
839 	mqattr = &mq->mq_attrib;
840 
841 	/* Check the message size limit */
842 	if (msg_len <= 0 || msg_len > mqattr->mq_msgsize) {
843 		error = EMSGSIZE;
844 		goto error;
845 	}
846 
847 	/* Check if queue is full */
848 	while (mqattr->mq_curmsgs >= mqattr->mq_maxmsg) {
849 		int t;
850 
851 		if (mqattr->mq_flags & O_NONBLOCK) {
852 			error = EAGAIN;
853 			goto error;
854 		}
855 		if (ts) {
856 			error = ts2timo(CLOCK_REALTIME, TIMER_ABSTIME, ts, &t,
857 			    NULL);
858 			if (error)
859 				goto error;
860 		} else
861 			t = 0;
862 		/* Block until queue becomes available */
863 		error = cv_timedwait_sig(&mq->mq_recv_cv, &mq->mq_mtx, t);
864 		if (error || (mqattr->mq_flags & MQ_UNLINKED)) {
865 			error = (error == EWOULDBLOCK) ? ETIMEDOUT : error;
866 			goto error;
867 		}
868 	}
869 	KASSERT(mqattr->mq_curmsgs < mqattr->mq_maxmsg);
870 
871 	/*
872 	 * Insert message into the queue, according to the priority.
873 	 * Note the difference between index and priority.
874 	 */
875 	if (__predict_true(msg_prio < MQ_PQSIZE)) {
876 		u_int idx = MQ_PQSIZE - msg_prio;
877 
878 		KASSERT(idx != MQ_PQRESQ);
879 		TAILQ_INSERT_TAIL(&mq->mq_head[idx], msg, msg_queue);
880 		mq->mq_bitmap |= (1 << --idx);
881 	} else {
882 		mqueue_linear_insert(mq, msg);
883 	}
884 
885 	/* Check for the notify */
886 	if (mqattr->mq_curmsgs == 0 && mq->mq_notify_proc &&
887 	    (mqattr->mq_flags & MQ_RECEIVE) == 0 &&
888 	    mq->mq_sig_notify.sigev_notify == SIGEV_SIGNAL) {
889 		/* Initialize the signal */
890 		KSI_INIT(&ksi);
891 		ksi.ksi_signo = mq->mq_sig_notify.sigev_signo;
892 		ksi.ksi_code = SI_MESGQ;
893 		ksi.ksi_value = mq->mq_sig_notify.sigev_value;
894 		/* Unregister the process */
895 		notify = mq->mq_notify_proc;
896 		mq->mq_notify_proc = NULL;
897 	}
898 
899 	/* Increment the counter and signal waiter, if any */
900 	mqattr->mq_curmsgs++;
901 	cv_signal(&mq->mq_send_cv);
902 
903 	/* Ready for receiving now */
904 	selnotify(&mq->mq_rsel, POLLIN | POLLRDNORM, 0);
905 error:
906 	mutex_exit(&mq->mq_mtx);
907 	fd_putfile((int)mqdes);
908 
909 	if (error) {
910 		mqueue_freemsg(msg, size);
911 	} else if (notify) {
912 		/* Send the notify, if needed */
913 		mutex_enter(proc_lock);
914 		kpsignal(notify, &ksi, NULL);
915 		mutex_exit(proc_lock);
916 	}
917 	return error;
918 }
919 
920 int
sys_mq_send(struct lwp * l,const struct sys_mq_send_args * uap,register_t * retval)921 sys_mq_send(struct lwp *l, const struct sys_mq_send_args *uap,
922     register_t *retval)
923 {
924 	/* {
925 		syscallarg(mqd_t) mqdes;
926 		syscallarg(const char *) msg_ptr;
927 		syscallarg(size_t) msg_len;
928 		syscallarg(unsigned) msg_prio;
929 	} */
930 
931 	return mq_send1(SCARG(uap, mqdes), SCARG(uap, msg_ptr),
932 	    SCARG(uap, msg_len), SCARG(uap, msg_prio), NULL);
933 }
934 
935 int
sys___mq_timedsend50(struct lwp * l,const struct sys___mq_timedsend50_args * uap,register_t * retval)936 sys___mq_timedsend50(struct lwp *l, const struct sys___mq_timedsend50_args *uap,
937     register_t *retval)
938 {
939 	/* {
940 		syscallarg(mqd_t) mqdes;
941 		syscallarg(const char *) msg_ptr;
942 		syscallarg(size_t) msg_len;
943 		syscallarg(unsigned) msg_prio;
944 		syscallarg(const struct timespec *) abs_timeout;
945 	} */
946 	struct timespec ts, *tsp;
947 	int error;
948 
949 	/* Get and convert time value */
950 	if (SCARG(uap, abs_timeout)) {
951 		error = copyin(SCARG(uap, abs_timeout), &ts, sizeof(ts));
952 		if (error)
953 			return error;
954 		tsp = &ts;
955 	} else {
956 		tsp = NULL;
957 	}
958 
959 	return mq_send1(SCARG(uap, mqdes), SCARG(uap, msg_ptr),
960 	    SCARG(uap, msg_len), SCARG(uap, msg_prio), tsp);
961 }
962 
963 int
sys_mq_notify(struct lwp * l,const struct sys_mq_notify_args * uap,register_t * retval)964 sys_mq_notify(struct lwp *l, const struct sys_mq_notify_args *uap,
965     register_t *retval)
966 {
967 	/* {
968 		syscallarg(mqd_t) mqdes;
969 		syscallarg(const struct sigevent *) notification;
970 	} */
971 	struct mqueue *mq;
972 	struct sigevent sig;
973 	int error;
974 
975 	if (SCARG(uap, notification)) {
976 		/* Get the signal from user-space */
977 		error = copyin(SCARG(uap, notification), &sig,
978 		    sizeof(struct sigevent));
979 		if (error)
980 			return error;
981 		if (sig.sigev_notify == SIGEV_SIGNAL &&
982 		    (sig.sigev_signo <=0 || sig.sigev_signo >= NSIG))
983 			return EINVAL;
984 	}
985 
986 	error = mqueue_get(SCARG(uap, mqdes), 0, &mq);
987 	if (error) {
988 		return error;
989 	}
990 	if (SCARG(uap, notification)) {
991 		/* Register notification: set the signal and target process */
992 		if (mq->mq_notify_proc == NULL) {
993 			memcpy(&mq->mq_sig_notify, &sig,
994 			    sizeof(struct sigevent));
995 			mq->mq_notify_proc = l->l_proc;
996 		} else {
997 			/* Fail if someone else already registered */
998 			error = EBUSY;
999 		}
1000 	} else {
1001 		/* Unregister the notification */
1002 		mq->mq_notify_proc = NULL;
1003 	}
1004 	mutex_exit(&mq->mq_mtx);
1005 	fd_putfile((int)SCARG(uap, mqdes));
1006 
1007 	return error;
1008 }
1009 
1010 int
sys_mq_getattr(struct lwp * l,const struct sys_mq_getattr_args * uap,register_t * retval)1011 sys_mq_getattr(struct lwp *l, const struct sys_mq_getattr_args *uap,
1012     register_t *retval)
1013 {
1014 	/* {
1015 		syscallarg(mqd_t) mqdes;
1016 		syscallarg(struct mq_attr *) mqstat;
1017 	} */
1018 	struct mqueue *mq;
1019 	struct mq_attr attr;
1020 	int error;
1021 
1022 	error = mqueue_get(SCARG(uap, mqdes), 0, &mq);
1023 	if (error) {
1024 		return error;
1025 	}
1026 	memcpy(&attr, &mq->mq_attrib, sizeof(struct mq_attr));
1027 	mutex_exit(&mq->mq_mtx);
1028 	fd_putfile((int)SCARG(uap, mqdes));
1029 
1030 	return copyout(&attr, SCARG(uap, mqstat), sizeof(struct mq_attr));
1031 }
1032 
1033 int
sys_mq_setattr(struct lwp * l,const struct sys_mq_setattr_args * uap,register_t * retval)1034 sys_mq_setattr(struct lwp *l, const struct sys_mq_setattr_args *uap,
1035     register_t *retval)
1036 {
1037 	/* {
1038 		syscallarg(mqd_t) mqdes;
1039 		syscallarg(const struct mq_attr *) mqstat;
1040 		syscallarg(struct mq_attr *) omqstat;
1041 	} */
1042 	struct mqueue *mq;
1043 	struct mq_attr attr;
1044 	int error, nonblock;
1045 
1046 	error = copyin(SCARG(uap, mqstat), &attr, sizeof(struct mq_attr));
1047 	if (error)
1048 		return error;
1049 	nonblock = (attr.mq_flags & O_NONBLOCK);
1050 
1051 	error = mqueue_get(SCARG(uap, mqdes), 0, &mq);
1052 	if (error) {
1053 		return error;
1054 	}
1055 
1056 	/* Copy the old attributes, if needed */
1057 	if (SCARG(uap, omqstat)) {
1058 		memcpy(&attr, &mq->mq_attrib, sizeof(struct mq_attr));
1059 	}
1060 
1061 	/* Ignore everything, except O_NONBLOCK */
1062 	if (nonblock)
1063 		mq->mq_attrib.mq_flags |= O_NONBLOCK;
1064 	else
1065 		mq->mq_attrib.mq_flags &= ~O_NONBLOCK;
1066 
1067 	mutex_exit(&mq->mq_mtx);
1068 	fd_putfile((int)SCARG(uap, mqdes));
1069 
1070 	/*
1071 	 * Copy the data to the user-space.
1072 	 * Note: According to POSIX, the new attributes should not be set in
1073 	 * case of fail - this would be violated.
1074 	 */
1075 	if (SCARG(uap, omqstat))
1076 		error = copyout(&attr, SCARG(uap, omqstat),
1077 		    sizeof(struct mq_attr));
1078 
1079 	return error;
1080 }
1081 
1082 int
sys_mq_unlink(struct lwp * l,const struct sys_mq_unlink_args * uap,register_t * retval)1083 sys_mq_unlink(struct lwp *l, const struct sys_mq_unlink_args *uap,
1084     register_t *retval)
1085 {
1086 	/* {
1087 		syscallarg(const char *) name;
1088 	} */
1089 	mqueue_t *mq;
1090 	char *name;
1091 	int error, refcnt = 0;
1092 
1093 	/* Get the name from the user-space */
1094 	name = kmem_alloc(MQ_NAMELEN, KM_SLEEP);
1095 	error = copyinstr(SCARG(uap, name), name, MQ_NAMELEN - 1, NULL);
1096 	if (error) {
1097 		kmem_free(name, MQ_NAMELEN);
1098 		return error;
1099 	}
1100 
1101 	mutex_enter(&mqlist_lock);
1102 	mq = mqueue_lookup(name);
1103 	if (mq == NULL) {
1104 		error = ENOENT;
1105 		goto err;
1106 	}
1107 	KASSERT(mutex_owned(&mq->mq_mtx));
1108 
1109 	/* Verify permissions. */
1110 	if (kauth_authorize_system(l->l_cred, KAUTH_SYSTEM_MQUEUE, 0, mq,
1111 	    NULL, NULL)) {
1112 		mutex_exit(&mq->mq_mtx);
1113 		error = EACCES;
1114 		goto err;
1115 	}
1116 
1117 	/* Remove and destroy if no references. */
1118 	LIST_REMOVE(mq, mq_list);
1119 	refcnt = mq->mq_refcnt;
1120 	if (refcnt) {
1121 		/* Mark as unlinked, if there are references. */
1122 		mq->mq_attrib.mq_flags |= MQ_UNLINKED;
1123 	}
1124 
1125 	/* Wake up waiters, if there are any. */
1126 	cv_broadcast(&mq->mq_send_cv);
1127 	cv_broadcast(&mq->mq_recv_cv);
1128 
1129 	selnotify(&mq->mq_rsel, POLLHUP, 0);
1130 	selnotify(&mq->mq_wsel, POLLHUP, 0);
1131 
1132 	mutex_exit(&mq->mq_mtx);
1133 err:
1134 	mutex_exit(&mqlist_lock);
1135 	/*
1136 	 * If last reference - destroy the message queue.  Otherwise,
1137 	 * the last mq_close() call will do that.
1138 	 */
1139 	if (!error && refcnt == 0) {
1140 		mqueue_destroy(mq);
1141 	}
1142 	kmem_free(name, MQ_NAMELEN);
1143 
1144 	return error;
1145 }
1146 
1147 /*
1148  * System control nodes.
1149  */
1150 static int
mqueue_sysctl_init(void)1151 mqueue_sysctl_init(void)
1152 {
1153 	const struct sysctlnode *node = NULL;
1154 
1155 	mqsysctl_log = NULL;
1156 
1157 	sysctl_createv(&mqsysctl_log, 0, NULL, NULL,
1158 		CTLFLAG_PERMANENT|CTLFLAG_IMMEDIATE,
1159 		CTLTYPE_INT, "posix_msg",
1160 		SYSCTL_DESCR("Version of IEEE Std 1003.1 and its "
1161 			     "Message Passing option to which the "
1162 			     "system attempts to conform"),
1163 		NULL, _POSIX_MESSAGE_PASSING, NULL, 0,
1164 		CTL_KERN, CTL_CREATE, CTL_EOL);
1165 	sysctl_createv(&mqsysctl_log, 0, NULL, &node,
1166 		CTLFLAG_PERMANENT,
1167 		CTLTYPE_NODE, "mqueue",
1168 		SYSCTL_DESCR("Message queue options"),
1169 		NULL, 0, NULL, 0,
1170 		CTL_KERN, CTL_CREATE, CTL_EOL);
1171 
1172 	if (node == NULL)
1173 		return ENXIO;
1174 
1175 	sysctl_createv(&mqsysctl_log, 0, &node, NULL,
1176 		CTLFLAG_PERMANENT | CTLFLAG_READWRITE,
1177 		CTLTYPE_INT, "mq_open_max",
1178 		SYSCTL_DESCR("Maximal number of message queue descriptors "
1179 			     "that process could open"),
1180 		NULL, 0, &mq_open_max, 0,
1181 		CTL_CREATE, CTL_EOL);
1182 	sysctl_createv(&mqsysctl_log, 0, &node, NULL,
1183 		CTLFLAG_PERMANENT | CTLFLAG_READWRITE,
1184 		CTLTYPE_INT, "mq_prio_max",
1185 		SYSCTL_DESCR("Maximal priority of the message"),
1186 		NULL, 0, &mq_prio_max, 0,
1187 		CTL_CREATE, CTL_EOL);
1188 	sysctl_createv(&mqsysctl_log, 0, &node, NULL,
1189 		CTLFLAG_PERMANENT | CTLFLAG_READWRITE,
1190 		CTLTYPE_INT, "mq_max_msgsize",
1191 		SYSCTL_DESCR("Maximal allowed size of the message"),
1192 		NULL, 0, &mq_max_msgsize, 0,
1193 		CTL_CREATE, CTL_EOL);
1194 	sysctl_createv(&mqsysctl_log, 0, &node, NULL,
1195 		CTLFLAG_PERMANENT | CTLFLAG_READWRITE,
1196 		CTLTYPE_INT, "mq_def_maxmsg",
1197 		SYSCTL_DESCR("Default maximal message count"),
1198 		NULL, 0, &mq_def_maxmsg, 0,
1199 		CTL_CREATE, CTL_EOL);
1200 	sysctl_createv(&mqsysctl_log, 0, &node, NULL,
1201 		CTLFLAG_PERMANENT | CTLFLAG_READWRITE,
1202 		CTLTYPE_INT, "mq_max_maxmsg",
1203 		SYSCTL_DESCR("Maximal allowed message count"),
1204 		NULL, 0, &mq_max_maxmsg, 0,
1205 		CTL_CREATE, CTL_EOL);
1206 
1207 	return 0;
1208 }
1209 
1210 /*
1211  * Debugging.
1212  */
1213 #if defined(DDB)
1214 
1215 void
mqueue_print_list(void (* pr)(const char *,...))1216 mqueue_print_list(void (*pr)(const char *, ...))
1217 {
1218 	struct mqueue *mq;
1219 
1220 	(*pr)("Global list of the message queues:\n");
1221 	(*pr)("%20s %10s %8s %8s %3s %4s %4s %4s\n",
1222 	    "Name", "Ptr", "Mode", "Flags",  "Ref",
1223 	    "MaxMsg", "MsgSze", "CurMsg");
1224 	LIST_FOREACH(mq, &mqueue_head, mq_list) {
1225 		(*pr)("%20s %10p %8x %8x %3u %6lu %6lu %6lu\n",
1226 		    mq->mq_name, mq, mq->mq_mode,
1227 		    mq->mq_attrib.mq_flags, mq->mq_refcnt,
1228 		    mq->mq_attrib.mq_maxmsg, mq->mq_attrib.mq_msgsize,
1229 		    mq->mq_attrib.mq_curmsgs);
1230 	}
1231 }
1232 
1233 #endif /* defined(DDB) */
1234