1 /* 2 * Sun RPC is a product of Sun Microsystems, Inc. and is provided for 3 * unrestricted use provided that this legend is included on all tape 4 * media and as a part of the software program in whole or part. Users 5 * may copy or modify Sun RPC without charge, but are not authorized 6 * to license or distribute it to anyone else except as part of a product or 7 * program developed by the user. 8 * 9 * SUN RPC IS PROVIDED AS IS WITH NO WARRANTIES OF ANY KIND INCLUDING THE 10 * WARRANTIES OF DESIGN, MERCHANTIBILITY AND FITNESS FOR A PARTICULAR 11 * PURPOSE, OR ARISING FROM A COURSE OF DEALING, USAGE OR TRADE PRACTICE. 12 * 13 * Sun RPC is provided with no support and without any obligation on the 14 * part of Sun Microsystems, Inc. to assist in its use, correction, 15 * modification or enhancement. 16 * 17 * SUN MICROSYSTEMS, INC. SHALL HAVE NO LIABILITY WITH RESPECT TO THE 18 * INFRINGEMENT OF COPYRIGHTS, TRADE SECRETS OR ANY PATENTS BY SUN RPC 19 * OR ANY PART THEREOF. 20 * 21 * In no event will Sun Microsystems, Inc. be liable for any lost revenue 22 * or profits or other special, indirect and consequential damages, even if 23 * Sun has been advised of the possibility of such damages. 24 * 25 * Sun Microsystems, Inc. 26 * 2550 Garcia Avenue 27 * Mountain View, California 94043 28 * 29 * @(#)svc_dg.c 1.17 94/04/24 SMI 30 * $NetBSD: svc_dg.c,v 1.4 2000/07/06 03:10:35 christos Exp $ 31 * $FreeBSD: src/lib/libc/rpc/svc_dg.c,v 1.8 2006/02/27 22:10:59 deischen Exp $ 32 */ 33 34 /* 35 * Copyright (c) 1986-1991 by Sun Microsystems Inc. 36 */ 37 38 /* 39 * svc_dg.c, Server side for connectionless RPC. 40 * 41 * Does some caching in the hopes of achieving execute-at-most-once semantics. 42 */ 43 44 #include "namespace.h" 45 #include "reentrant.h" 46 #include <sys/types.h> 47 #include <sys/socket.h> 48 #include <rpc/rpc.h> 49 #include <rpc/svc_dg.h> 50 #include <errno.h> 51 #include <unistd.h> 52 #include <stdio.h> 53 #include <stdlib.h> 54 #include <string.h> 55 #ifdef RPC_CACHE_DEBUG 56 #include <netconfig.h> 57 #include <netdir.h> 58 #endif 59 #include <err.h> 60 #include "un-namespace.h" 61 62 #include "rpc_com.h" 63 #include "mt_misc.h" 64 65 #define su_data(xprt) ((struct svc_dg_data *)(xprt->xp_p2)) 66 #define rpc_buffer(xprt) ((xprt)->xp_p1) 67 68 #ifndef MAX 69 #define MAX(a, b) (((a) > (b)) ? (a) : (b)) 70 #endif 71 72 static void svc_dg_ops(SVCXPRT *); 73 static enum xprt_stat svc_dg_stat(SVCXPRT *); 74 static bool_t svc_dg_recv(SVCXPRT *, struct rpc_msg *); 75 static bool_t svc_dg_reply(SVCXPRT *, struct rpc_msg *); 76 static bool_t svc_dg_getargs(SVCXPRT *, xdrproc_t, void *); 77 static bool_t svc_dg_freeargs(SVCXPRT *, xdrproc_t, void *); 78 static void svc_dg_destroy(SVCXPRT *); 79 static bool_t svc_dg_control(SVCXPRT *, const u_int, void *); 80 static int cache_get(SVCXPRT *, struct rpc_msg *, char **, size_t *); 81 static void cache_set(SVCXPRT *, size_t); 82 int svc_dg_enablecache(SVCXPRT *, u_int); 83 84 /* 85 * Usage: 86 * xprt = svc_dg_create(sock, sendsize, recvsize); 87 * Does other connectionless specific initializations. 88 * Once *xprt is initialized, it is registered. 89 * see (svc.h, xprt_register). If recvsize or sendsize are 0 suitable 90 * system defaults are chosen. 91 * The routines returns NULL if a problem occurred. 92 */ 93 static const char svc_dg_str[] = "svc_dg_create: %s"; 94 static const char svc_dg_err1[] = "could not get transport information"; 95 static const char svc_dg_err2[] = " transport does not support data transfer"; 96 static const char __no_mem_str[] = "out of memory"; 97 98 SVCXPRT * 99 svc_dg_create(int fd, u_int sendsize, u_int recvsize) 100 { 101 SVCXPRT *xprt; 102 struct svc_dg_data *su = NULL; 103 struct __rpc_sockinfo si; 104 struct sockaddr_storage ss; 105 socklen_t slen; 106 107 if (!__rpc_fd2sockinfo(fd, &si)) { 108 warnx(svc_dg_str, svc_dg_err1); 109 return (NULL); 110 } 111 /* 112 * Find the receive and the send size 113 */ 114 sendsize = __rpc_get_t_size(si.si_af, si.si_proto, (int)sendsize); 115 recvsize = __rpc_get_t_size(si.si_af, si.si_proto, (int)recvsize); 116 if ((sendsize == 0) || (recvsize == 0)) { 117 warnx(svc_dg_str, svc_dg_err2); 118 return (NULL); 119 } 120 121 xprt = mem_alloc(sizeof (SVCXPRT)); 122 if (xprt == NULL) 123 goto freedata; 124 memset(xprt, 0, sizeof (SVCXPRT)); 125 126 su = mem_alloc(sizeof (*su)); 127 if (su == NULL) 128 goto freedata; 129 su->su_iosz = ((MAX(sendsize, recvsize) + 3) / 4) * 4; 130 if ((rpc_buffer(xprt) = mem_alloc(su->su_iosz)) == NULL) 131 goto freedata; 132 xdrmem_create(&(su->su_xdrs), rpc_buffer(xprt), su->su_iosz, 133 XDR_DECODE); 134 su->su_cache = NULL; 135 xprt->xp_fd = fd; 136 xprt->xp_p2 = su; 137 xprt->xp_verf.oa_base = su->su_verfbody; 138 svc_dg_ops(xprt); 139 xprt->xp_rtaddr.maxlen = sizeof (struct sockaddr_storage); 140 141 slen = sizeof ss; 142 if (_getsockname(fd, (struct sockaddr *)(void *)&ss, &slen) < 0) 143 goto freedata; 144 xprt->xp_ltaddr.buf = mem_alloc(sizeof (struct sockaddr_storage)); 145 xprt->xp_ltaddr.maxlen = sizeof (struct sockaddr_storage); 146 xprt->xp_ltaddr.len = slen; 147 memcpy(xprt->xp_ltaddr.buf, &ss, slen); 148 149 xprt_register(xprt); 150 return (xprt); 151 freedata: 152 warnx(svc_dg_str, __no_mem_str); 153 if (xprt) { 154 if (su) 155 mem_free(su, sizeof (*su)); 156 mem_free(xprt, sizeof (SVCXPRT)); 157 } 158 return (NULL); 159 } 160 161 /*ARGSUSED*/ 162 static enum xprt_stat 163 svc_dg_stat(SVCXPRT *xprt __unused) 164 { 165 return (XPRT_IDLE); 166 } 167 168 static bool_t 169 svc_dg_recv(SVCXPRT *xprt, struct rpc_msg *msg) 170 { 171 struct svc_dg_data *su = su_data(xprt); 172 XDR *xdrs = &(su->su_xdrs); 173 char *reply; 174 struct sockaddr_storage ss; 175 socklen_t alen; 176 size_t replylen; 177 ssize_t rlen; 178 179 again: 180 alen = sizeof (struct sockaddr_storage); 181 rlen = _recvfrom(xprt->xp_fd, rpc_buffer(xprt), su->su_iosz, 0, 182 (struct sockaddr *)(void *)&ss, &alen); 183 if (rlen == -1 && errno == EINTR) 184 goto again; 185 if (rlen == -1 || (rlen < (ssize_t)(4 * sizeof (u_int32_t)))) 186 return (FALSE); 187 if (xprt->xp_rtaddr.len < alen) { 188 if (xprt->xp_rtaddr.len != 0) 189 mem_free(xprt->xp_rtaddr.buf, xprt->xp_rtaddr.len); 190 xprt->xp_rtaddr.buf = mem_alloc(alen); 191 xprt->xp_rtaddr.len = alen; 192 } 193 memcpy(xprt->xp_rtaddr.buf, &ss, alen); 194 #ifdef PORTMAP 195 if (ss.ss_family == AF_INET) { 196 xprt->xp_raddr = *(struct sockaddr_in *)xprt->xp_rtaddr.buf; 197 xprt->xp_addrlen = sizeof (struct sockaddr_in); 198 } 199 #endif /* PORTMAP */ 200 xdrs->x_op = XDR_DECODE; 201 XDR_SETPOS(xdrs, 0); 202 if (! xdr_callmsg(xdrs, msg)) { 203 return (FALSE); 204 } 205 su->su_xid = msg->rm_xid; 206 if (su->su_cache != NULL) { 207 if (cache_get(xprt, msg, &reply, &replylen)) { 208 _sendto(xprt->xp_fd, reply, replylen, 0, 209 (struct sockaddr *)(void *)&ss, alen); 210 return (FALSE); 211 } 212 } 213 return (TRUE); 214 } 215 216 static bool_t 217 svc_dg_reply(SVCXPRT *xprt, struct rpc_msg *msg) 218 { 219 struct svc_dg_data *su = su_data(xprt); 220 XDR *xdrs = &(su->su_xdrs); 221 bool_t stat = FALSE; 222 size_t slen; 223 224 xdrs->x_op = XDR_ENCODE; 225 XDR_SETPOS(xdrs, 0); 226 msg->rm_xid = su->su_xid; 227 if (xdr_replymsg(xdrs, msg)) { 228 slen = XDR_GETPOS(xdrs); 229 if (_sendto(xprt->xp_fd, rpc_buffer(xprt), slen, 0, 230 (struct sockaddr *)xprt->xp_rtaddr.buf, 231 (socklen_t)xprt->xp_rtaddr.len) == (ssize_t) slen) { 232 stat = TRUE; 233 if (su->su_cache) 234 cache_set(xprt, slen); 235 } 236 } 237 return (stat); 238 } 239 240 static bool_t 241 svc_dg_getargs(SVCXPRT *xprt, xdrproc_t xdr_args, void *args_ptr) 242 { 243 return (*xdr_args)(&(su_data(xprt)->su_xdrs), args_ptr); 244 } 245 246 static bool_t 247 svc_dg_freeargs(SVCXPRT *xprt, xdrproc_t xdr_args, void *args_ptr) 248 { 249 XDR *xdrs = &(su_data(xprt)->su_xdrs); 250 251 xdrs->x_op = XDR_FREE; 252 return (*xdr_args)(xdrs, args_ptr); 253 } 254 255 static void 256 svc_dg_destroy(SVCXPRT *xprt) 257 { 258 struct svc_dg_data *su = su_data(xprt); 259 260 xprt_unregister(xprt); 261 if (xprt->xp_fd != -1) 262 _close(xprt->xp_fd); 263 XDR_DESTROY(&(su->su_xdrs)); 264 mem_free(rpc_buffer(xprt), su->su_iosz); 265 mem_free(su, sizeof (*su)); 266 if (xprt->xp_rtaddr.buf) 267 mem_free(xprt->xp_rtaddr.buf, xprt->xp_rtaddr.maxlen); 268 if (xprt->xp_ltaddr.buf) 269 mem_free(xprt->xp_ltaddr.buf, xprt->xp_ltaddr.maxlen); 270 if (xprt->xp_tp) 271 free(xprt->xp_tp); 272 mem_free(xprt, sizeof (SVCXPRT)); 273 } 274 275 static bool_t 276 /*ARGSUSED*/ 277 svc_dg_control(SVCXPRT *xprt __unused, const u_int rq __unused, 278 void *in __unused) 279 { 280 return (FALSE); 281 } 282 283 static void 284 svc_dg_ops(SVCXPRT *xprt) 285 { 286 static struct xp_ops ops; 287 static struct xp_ops2 ops2; 288 289 /* VARIABLES PROTECTED BY ops_lock: ops */ 290 291 mutex_lock(&ops_lock); 292 if (ops.xp_recv == NULL) { 293 ops.xp_recv = svc_dg_recv; 294 ops.xp_stat = svc_dg_stat; 295 ops.xp_getargs = svc_dg_getargs; 296 ops.xp_reply = svc_dg_reply; 297 ops.xp_freeargs = svc_dg_freeargs; 298 ops.xp_destroy = svc_dg_destroy; 299 ops2.xp_control = svc_dg_control; 300 } 301 xprt->xp_ops = &ops; 302 xprt->xp_ops2 = &ops2; 303 mutex_unlock(&ops_lock); 304 } 305 306 /* The CACHING COMPONENT */ 307 308 /* 309 * Could have been a separate file, but some part of it depends upon the 310 * private structure of the client handle. 311 * 312 * Fifo cache for cl server 313 * Copies pointers to reply buffers into fifo cache 314 * Buffers are sent again if retransmissions are detected. 315 */ 316 317 #define SPARSENESS 4 /* 75% sparse */ 318 319 #define ALLOC(type, size) \ 320 (type *) mem_alloc((sizeof (type) * (size))) 321 322 #define MEMZERO(addr, type, size) \ 323 memset((void *) (addr), 0, sizeof (type) * (int) (size)) 324 325 #define FREE(addr, type, size) \ 326 mem_free((addr), (sizeof (type) * (size))) 327 328 /* 329 * An entry in the cache 330 */ 331 typedef struct cache_node *cache_ptr; 332 struct cache_node { 333 /* 334 * Index into cache is xid, proc, vers, prog and address 335 */ 336 u_int32_t cache_xid; 337 rpcproc_t cache_proc; 338 rpcvers_t cache_vers; 339 rpcprog_t cache_prog; 340 struct netbuf cache_addr; 341 /* 342 * The cached reply and length 343 */ 344 char *cache_reply; 345 size_t cache_replylen; 346 /* 347 * Next node on the list, if there is a collision 348 */ 349 cache_ptr cache_next; 350 }; 351 352 /* 353 * The entire cache 354 */ 355 struct cl_cache { 356 u_int uc_size; /* size of cache */ 357 cache_ptr *uc_entries; /* hash table of entries in cache */ 358 cache_ptr *uc_fifo; /* fifo list of entries in cache */ 359 u_int uc_nextvictim; /* points to next victim in fifo list */ 360 rpcprog_t uc_prog; /* saved program number */ 361 rpcvers_t uc_vers; /* saved version number */ 362 rpcproc_t uc_proc; /* saved procedure number */ 363 }; 364 365 366 /* 367 * the hashing function 368 */ 369 #define CACHE_LOC(transp, xid) \ 370 (xid % (SPARSENESS * ((struct cl_cache *) \ 371 su_data(transp)->su_cache)->uc_size)) 372 373 /* 374 * Enable use of the cache. Returns 1 on success, 0 on failure. 375 * Note: there is no disable. 376 */ 377 static const char cache_enable_str[] = "svc_enablecache: %s %s"; 378 static const char alloc_err[] = "could not allocate cache "; 379 static const char enable_err[] = "cache already enabled"; 380 381 int 382 svc_dg_enablecache(SVCXPRT *transp, u_int size) 383 { 384 struct svc_dg_data *su = su_data(transp); 385 struct cl_cache *uc; 386 387 mutex_lock(&dupreq_lock); 388 if (su->su_cache != NULL) { 389 warnx(cache_enable_str, enable_err, " "); 390 mutex_unlock(&dupreq_lock); 391 return (0); 392 } 393 uc = ALLOC(struct cl_cache, 1); 394 if (uc == NULL) { 395 warnx(cache_enable_str, alloc_err, " "); 396 mutex_unlock(&dupreq_lock); 397 return (0); 398 } 399 uc->uc_size = size; 400 uc->uc_nextvictim = 0; 401 uc->uc_entries = ALLOC(cache_ptr, size * SPARSENESS); 402 if (uc->uc_entries == NULL) { 403 warnx(cache_enable_str, alloc_err, "data"); 404 FREE(uc, struct cl_cache, 1); 405 mutex_unlock(&dupreq_lock); 406 return (0); 407 } 408 MEMZERO(uc->uc_entries, cache_ptr, size * SPARSENESS); 409 uc->uc_fifo = ALLOC(cache_ptr, size); 410 if (uc->uc_fifo == NULL) { 411 warnx(cache_enable_str, alloc_err, "fifo"); 412 FREE(uc->uc_entries, cache_ptr, size * SPARSENESS); 413 FREE(uc, struct cl_cache, 1); 414 mutex_unlock(&dupreq_lock); 415 return (0); 416 } 417 MEMZERO(uc->uc_fifo, cache_ptr, size); 418 su->su_cache = (char *)(void *)uc; 419 mutex_unlock(&dupreq_lock); 420 return (1); 421 } 422 423 /* 424 * Set an entry in the cache. It assumes that the uc entry is set from 425 * the earlier call to cache_get() for the same procedure. This will always 426 * happen because cache_get() is calle by svc_dg_recv and cache_set() is called 427 * by svc_dg_reply(). All this hoopla because the right RPC parameters are 428 * not available at svc_dg_reply time. 429 */ 430 431 static const char cache_set_str[] = "cache_set: %s"; 432 static const char cache_set_err1[] = "victim not found"; 433 static const char cache_set_err2[] = "victim alloc failed"; 434 static const char cache_set_err3[] = "could not allocate new rpc buffer"; 435 436 static void 437 cache_set(SVCXPRT *xprt, size_t replylen) 438 { 439 cache_ptr victim; 440 cache_ptr *vicp; 441 struct svc_dg_data *su = su_data(xprt); 442 struct cl_cache *uc = (struct cl_cache *) su->su_cache; 443 u_int loc; 444 char *newbuf; 445 #ifdef RPC_CACHE_DEBUG 446 struct netconfig *nconf; 447 char *uaddr; 448 #endif 449 450 mutex_lock(&dupreq_lock); 451 /* 452 * Find space for the new entry, either by 453 * reusing an old entry, or by mallocing a new one 454 */ 455 victim = uc->uc_fifo[uc->uc_nextvictim]; 456 if (victim != NULL) { 457 loc = CACHE_LOC(xprt, victim->cache_xid); 458 for (vicp = &uc->uc_entries[loc]; 459 *vicp != NULL && *vicp != victim; 460 vicp = &(*vicp)->cache_next) 461 ; 462 if (*vicp == NULL) { 463 warnx(cache_set_str, cache_set_err1); 464 mutex_unlock(&dupreq_lock); 465 return; 466 } 467 *vicp = victim->cache_next; /* remove from cache */ 468 newbuf = victim->cache_reply; 469 } else { 470 victim = ALLOC(struct cache_node, 1); 471 if (victim == NULL) { 472 warnx(cache_set_str, cache_set_err2); 473 mutex_unlock(&dupreq_lock); 474 return; 475 } 476 newbuf = mem_alloc(su->su_iosz); 477 if (newbuf == NULL) { 478 warnx(cache_set_str, cache_set_err3); 479 FREE(victim, struct cache_node, 1); 480 mutex_unlock(&dupreq_lock); 481 return; 482 } 483 } 484 485 /* 486 * Store it away 487 */ 488 #ifdef RPC_CACHE_DEBUG 489 if (nconf = getnetconfigent(xprt->xp_netid)) { 490 uaddr = taddr2uaddr(nconf, &xprt->xp_rtaddr); 491 freenetconfigent(nconf); 492 printf( 493 "cache set for xid= %x prog=%d vers=%d proc=%d for rmtaddr=%s\n", 494 su->su_xid, uc->uc_prog, uc->uc_vers, 495 uc->uc_proc, uaddr); 496 free(uaddr); 497 } 498 #endif 499 victim->cache_replylen = replylen; 500 victim->cache_reply = rpc_buffer(xprt); 501 rpc_buffer(xprt) = newbuf; 502 xdrmem_create(&(su->su_xdrs), rpc_buffer(xprt), 503 su->su_iosz, XDR_ENCODE); 504 victim->cache_xid = su->su_xid; 505 victim->cache_proc = uc->uc_proc; 506 victim->cache_vers = uc->uc_vers; 507 victim->cache_prog = uc->uc_prog; 508 victim->cache_addr = xprt->xp_rtaddr; 509 victim->cache_addr.buf = ALLOC(char, xprt->xp_rtaddr.len); 510 memcpy(victim->cache_addr.buf, xprt->xp_rtaddr.buf, 511 (size_t)xprt->xp_rtaddr.len); 512 loc = CACHE_LOC(xprt, victim->cache_xid); 513 victim->cache_next = uc->uc_entries[loc]; 514 uc->uc_entries[loc] = victim; 515 uc->uc_fifo[uc->uc_nextvictim++] = victim; 516 uc->uc_nextvictim %= uc->uc_size; 517 mutex_unlock(&dupreq_lock); 518 } 519 520 /* 521 * Try to get an entry from the cache 522 * return 1 if found, 0 if not found and set the stage for cache_set() 523 */ 524 static int 525 cache_get(SVCXPRT *xprt, struct rpc_msg *msg, char **replyp, size_t *replylenp) 526 { 527 u_int loc; 528 cache_ptr ent; 529 struct svc_dg_data *su = su_data(xprt); 530 struct cl_cache *uc = (struct cl_cache *) su->su_cache; 531 #ifdef RPC_CACHE_DEBUG 532 struct netconfig *nconf; 533 char *uaddr; 534 #endif 535 536 mutex_lock(&dupreq_lock); 537 loc = CACHE_LOC(xprt, su->su_xid); 538 for (ent = uc->uc_entries[loc]; ent != NULL; ent = ent->cache_next) { 539 if (ent->cache_xid == su->su_xid && 540 ent->cache_proc == msg->rm_call.cb_proc && 541 ent->cache_vers == msg->rm_call.cb_vers && 542 ent->cache_prog == msg->rm_call.cb_prog && 543 ent->cache_addr.len == xprt->xp_rtaddr.len && 544 (memcmp(ent->cache_addr.buf, xprt->xp_rtaddr.buf, 545 xprt->xp_rtaddr.len) == 0)) { 546 #ifdef RPC_CACHE_DEBUG 547 if (nconf = getnetconfigent(xprt->xp_netid)) { 548 uaddr = taddr2uaddr(nconf, &xprt->xp_rtaddr); 549 freenetconfigent(nconf); 550 printf( 551 "cache entry found for xid=%x prog=%d vers=%d proc=%d for rmtaddr=%s\n", 552 su->su_xid, msg->rm_call.cb_prog, 553 msg->rm_call.cb_vers, 554 msg->rm_call.cb_proc, uaddr); 555 free(uaddr); 556 } 557 #endif 558 *replyp = ent->cache_reply; 559 *replylenp = ent->cache_replylen; 560 mutex_unlock(&dupreq_lock); 561 return (1); 562 } 563 } 564 /* 565 * Failed to find entry 566 * Remember a few things so we can do a set later 567 */ 568 uc->uc_proc = msg->rm_call.cb_proc; 569 uc->uc_vers = msg->rm_call.cb_vers; 570 uc->uc_prog = msg->rm_call.cb_prog; 571 mutex_unlock(&dupreq_lock); 572 return (0); 573 } 574