1 /* 2 * Copyright (c) 2003, 2004 Jeffrey M. Hsu. All rights reserved. 3 * Copyright (c) 2003, 2004 The DragonFly Project. All rights reserved. 4 * 5 * This code is derived from software contributed to The DragonFly Project 6 * by Jeffrey M. Hsu. 7 * 8 * Redistribution and use in source and binary forms, with or without 9 * modification, are permitted provided that the following conditions 10 * are met: 11 * 1. Redistributions of source code must retain the above copyright 12 * notice, this list of conditions and the following disclaimer. 13 * 2. Redistributions in binary form must reproduce the above copyright 14 * notice, this list of conditions and the following disclaimer in the 15 * documentation and/or other materials provided with the distribution. 16 * 3. Neither the name of The DragonFly Project nor the names of its 17 * contributors may be used to endorse or promote products derived 18 * from this software without specific, prior written permission. 19 * 20 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 21 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 22 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS 23 * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE 24 * COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, 25 * INCIDENTAL, SPECIAL, EXEMPLARY OR CONSEQUENTIAL DAMAGES (INCLUDING, 26 * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; 27 * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED 28 * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, 29 * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT 30 * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 31 * SUCH DAMAGE. 32 */ 33 34 /* 35 * Copyright (c) 1982, 1986, 1988, 1993 36 * The Regents of the University of California. All rights reserved. 37 * 38 * Redistribution and use in source and binary forms, with or without 39 * modification, are permitted provided that the following conditions 40 * are met: 41 * 1. Redistributions of source code must retain the above copyright 42 * notice, this list of conditions and the following disclaimer. 43 * 2. Redistributions in binary form must reproduce the above copyright 44 * notice, this list of conditions and the following disclaimer in the 45 * documentation and/or other materials provided with the distribution. 46 * 3. Neither the name of the University nor the names of its contributors 47 * may be used to endorse or promote products derived from this software 48 * without specific prior written permission. 49 * 50 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 51 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 52 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 53 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 54 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 55 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 56 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 57 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 58 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 59 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 60 * SUCH DAMAGE. 61 * 62 * From: @(#)tcp_usrreq.c 8.2 (Berkeley) 1/3/94 63 * $FreeBSD: src/sys/netinet/tcp_usrreq.c,v 1.51.2.17 2002/10/11 11:46:44 ume Exp $ 64 */ 65 66 #include "opt_inet.h" 67 #include "opt_inet6.h" 68 #include "opt_tcpdebug.h" 69 70 #include <sys/param.h> 71 #include <sys/systm.h> 72 #include <sys/kernel.h> 73 #include <sys/malloc.h> 74 #include <sys/sysctl.h> 75 #include <sys/globaldata.h> 76 #include <sys/thread.h> 77 78 #include <sys/mbuf.h> 79 #ifdef INET6 80 #include <sys/domain.h> 81 #endif /* INET6 */ 82 #include <sys/socket.h> 83 #include <sys/socketvar.h> 84 #include <sys/socketops.h> 85 #include <sys/protosw.h> 86 #include <sys/jail.h> 87 88 #include <sys/thread2.h> 89 #include <sys/msgport2.h> 90 #include <sys/socketvar2.h> 91 92 #include <net/if.h> 93 #include <net/netisr.h> 94 #include <net/route.h> 95 96 #include <net/netmsg2.h> 97 #include <net/netisr2.h> 98 99 #include <netinet/in.h> 100 #include <netinet/in_systm.h> 101 #ifdef INET6 102 #include <netinet/ip6.h> 103 #endif 104 #include <netinet/in_pcb.h> 105 #ifdef INET6 106 #include <netinet6/in6_pcb.h> 107 #endif 108 #include <netinet/in_var.h> 109 #include <netinet/ip_var.h> 110 #ifdef INET6 111 #include <netinet6/ip6_var.h> 112 #include <netinet6/tcp6_var.h> 113 #endif 114 #include <netinet/tcp.h> 115 #include <netinet/tcp_fsm.h> 116 #include <netinet/tcp_seq.h> 117 #include <netinet/tcp_timer.h> 118 #include <netinet/tcp_timer2.h> 119 #include <netinet/tcp_var.h> 120 #include <netinet/tcpip.h> 121 #ifdef TCPDEBUG 122 #include <netinet/tcp_debug.h> 123 #endif 124 125 /* 126 * TCP protocol interface to socket abstraction. 127 */ 128 extern char *tcpstates[]; /* XXX ??? */ 129 130 static int tcp_attach (struct socket *, struct pru_attach_info *); 131 static void tcp_connect (netmsg_t msg); 132 #ifdef INET6 133 static void tcp6_connect (netmsg_t msg); 134 static int tcp6_connect_oncpu(struct tcpcb *tp, int flags, 135 struct mbuf **mp, 136 struct sockaddr_in6 *sin6, 137 struct in6_addr *addr6); 138 #endif /* INET6 */ 139 static struct tcpcb * 140 tcp_disconnect (struct tcpcb *); 141 static struct tcpcb * 142 tcp_usrclosed (struct tcpcb *); 143 144 #ifdef TCPDEBUG 145 #define TCPDEBUG0 int ostate = 0 146 #define TCPDEBUG1() ostate = tp ? tp->t_state : 0 147 #define TCPDEBUG2(req) if (tp && (so->so_options & SO_DEBUG)) \ 148 tcp_trace(TA_USER, ostate, tp, 0, 0, req) 149 #else 150 #define TCPDEBUG0 151 #define TCPDEBUG1() 152 #define TCPDEBUG2(req) 153 #endif 154 155 /* 156 * For some ill optimized programs, which try to use TCP_NOPUSH 157 * to improve performance, will have small amount of data sits 158 * in the sending buffer. These small amount of data will _not_ 159 * be pushed into the network until more data are written into 160 * the socket or the socket write side is shutdown. 161 */ 162 static int tcp_disable_nopush = 1; 163 SYSCTL_INT(_net_inet_tcp, OID_AUTO, disable_nopush, CTLFLAG_RW, 164 &tcp_disable_nopush, 0, "TCP_NOPUSH socket option will have no effect"); 165 166 /* 167 * Allocate socket buffer space. 168 */ 169 static int 170 tcp_usr_preattach(struct socket *so, int proto __unused, 171 struct pru_attach_info *ai) 172 { 173 int error; 174 175 if (so->so_snd.ssb_hiwat == 0 || so->so_rcv.ssb_hiwat == 0) { 176 error = soreserve(so, tcp_sendspace, tcp_recvspace, 177 ai->sb_rlimit); 178 if (error) 179 return (error); 180 } 181 atomic_set_int(&so->so_rcv.ssb_flags, SSB_AUTOSIZE); 182 atomic_set_int(&so->so_snd.ssb_flags, SSB_AUTOSIZE | SSB_PREALLOC); 183 184 return 0; 185 } 186 187 /* 188 * TCP attaches to socket via pru_attach(), reserving space, 189 * and an internet control block. This socket may move to 190 * other CPU later when we bind/connect. 191 */ 192 static void 193 tcp_usr_attach(netmsg_t msg) 194 { 195 struct socket *so = msg->base.nm_so; 196 struct pru_attach_info *ai = msg->attach.nm_ai; 197 int error; 198 struct inpcb *inp; 199 struct tcpcb *tp = NULL; 200 TCPDEBUG0; 201 202 inp = so->so_pcb; 203 KASSERT(inp == NULL, ("tcp socket attached")); 204 TCPDEBUG1(); 205 206 error = tcp_attach(so, ai); 207 if (error) 208 goto out; 209 210 if ((so->so_options & SO_LINGER) && so->so_linger == 0) 211 so->so_linger = TCP_LINGERTIME; 212 tp = sototcpcb(so); 213 out: 214 TCPDEBUG2(PRU_ATTACH); 215 lwkt_replymsg(&msg->lmsg, error); 216 } 217 218 /* 219 * pru_detach() detaches the TCP protocol from the socket. 220 * If the protocol state is non-embryonic, then can't 221 * do this directly: have to initiate a pru_disconnect(), 222 * which may finish later; embryonic TCB's can just 223 * be discarded here. 224 */ 225 static void 226 tcp_usr_detach(netmsg_t msg) 227 { 228 struct socket *so = msg->base.nm_so; 229 int error = 0; 230 struct inpcb *inp; 231 struct tcpcb *tp; 232 TCPDEBUG0; 233 234 inp = so->so_pcb; 235 236 /* 237 * If the inp is already detached or never attached, it may have 238 * been due to an async close or async attach failure. Just return 239 * as if no error occured. 240 */ 241 if (inp) { 242 tp = intotcpcb(inp); 243 KASSERT(tp != NULL, ("tcp_usr_detach: tp is NULL")); 244 TCPDEBUG1(); 245 tp = tcp_disconnect(tp); 246 TCPDEBUG2(PRU_DETACH); 247 } 248 lwkt_replymsg(&msg->lmsg, error); 249 } 250 251 /* 252 * NOTE: ignore_error is non-zero for certain disconnection races 253 * which we want to silently allow, otherwise close() may return 254 * an unexpected error. 255 * 256 * NOTE: The variables (msg) and (tp) are assumed. 257 */ 258 #define COMMON_START(so, inp, ignore_error) \ 259 TCPDEBUG0; \ 260 \ 261 inp = so->so_pcb; \ 262 do { \ 263 if (inp == NULL) { \ 264 error = ignore_error ? 0 : EINVAL; \ 265 tp = NULL; \ 266 goto out; \ 267 } \ 268 tp = intotcpcb(inp); \ 269 TCPDEBUG1(); \ 270 } while(0) 271 272 #define COMMON_END1(req, noreply) \ 273 out: do { \ 274 TCPDEBUG2(req); \ 275 if (!(noreply)) \ 276 lwkt_replymsg(&msg->lmsg, error); \ 277 return; \ 278 } while(0) 279 280 #define COMMON_END(req) COMMON_END1((req), 0) 281 282 static void 283 tcp_sosetport(struct lwkt_msg *msg, lwkt_port_t port) 284 { 285 sosetport(((struct netmsg_base *)msg)->nm_so, port); 286 } 287 288 /* 289 * Give the socket an address. 290 */ 291 static void 292 tcp_usr_bind(netmsg_t msg) 293 { 294 struct socket *so = msg->bind.base.nm_so; 295 struct sockaddr *nam = msg->bind.nm_nam; 296 struct thread *td = msg->bind.nm_td; 297 int error = 0; 298 struct inpcb *inp; 299 struct tcpcb *tp; 300 struct sockaddr_in *sinp; 301 lwkt_port_t port0 = netisr_cpuport(0); 302 303 COMMON_START(so, inp, 0); 304 305 /* 306 * Must check for multicast addresses and disallow binding 307 * to them. 308 */ 309 sinp = (struct sockaddr_in *)nam; 310 if (sinp->sin_family == AF_INET && 311 IN_MULTICAST(ntohl(sinp->sin_addr.s_addr))) { 312 error = EAFNOSUPPORT; 313 goto out; 314 } 315 316 /* 317 * Check "already bound" here (in_pcbbind() does the same check 318 * though), so we don't forward a connected socket to netisr0, 319 * which would panic in the following in_pcbunlink(). 320 */ 321 if (inp->inp_lport != 0 || inp->inp_laddr.s_addr != INADDR_ANY) { 322 error = EINVAL; /* already bound */ 323 goto out; 324 } 325 326 /* 327 * Use netisr0 to serialize in_pcbbind(), so that pru_detach and 328 * pru_bind for different sockets on the same local port could be 329 * properly ordered. The original race is illustrated here for 330 * reference. 331 * 332 * s1 = socket(); 333 * bind(s1, *.PORT); 334 * close(s1); <----- asynchronous 335 * s2 = socket(); 336 * bind(s2, *.PORT); 337 * 338 * All will expect bind(s2, *.PORT) to succeed. However, it will 339 * fail, if following sequence happens due to random socket initial 340 * msgport and asynchronous close(2): 341 * 342 * netisrN netisrM 343 * : : 344 * : pru_bind(s2) [*.PORT is used by s1] 345 * pru_detach(s1) : 346 */ 347 if (&curthread->td_msgport != port0) { 348 lwkt_msg_t lmsg = &msg->bind.base.lmsg; 349 350 KASSERT((msg->bind.nm_flags & PRUB_RELINK) == 0, 351 ("already asked to relink")); 352 353 in_pcbunlink(so->so_pcb, &tcbinfo[mycpuid]); 354 msg->bind.nm_flags |= PRUB_RELINK; 355 356 TCP_STATE_MIGRATE_START(tp); 357 358 /* See the related comment in tcp_connect() */ 359 lwkt_setmsg_receipt(lmsg, tcp_sosetport); 360 lwkt_forwardmsg(port0, lmsg); 361 /* msg invalid now */ 362 return; 363 } 364 KASSERT(so->so_port == port0, ("so_port is not netisr0")); 365 366 if (msg->bind.nm_flags & PRUB_RELINK) { 367 msg->bind.nm_flags &= ~PRUB_RELINK; 368 TCP_STATE_MIGRATE_END(tp); 369 in_pcblink(so->so_pcb, &tcbinfo[mycpuid]); 370 } 371 KASSERT(inp->inp_pcbinfo == &tcbinfo[0], ("pcbinfo is not tcbinfo0")); 372 373 error = in_pcbbind(inp, nam, td); 374 if (error) 375 goto out; 376 377 COMMON_END(PRU_BIND); 378 } 379 380 #ifdef INET6 381 382 static void 383 tcp6_usr_bind(netmsg_t msg) 384 { 385 struct socket *so = msg->bind.base.nm_so; 386 struct sockaddr *nam = msg->bind.nm_nam; 387 struct thread *td = msg->bind.nm_td; 388 int error = 0; 389 struct inpcb *inp; 390 struct tcpcb *tp; 391 struct sockaddr_in6 *sin6p; 392 393 COMMON_START(so, inp, 0); 394 395 /* 396 * Must check for multicast addresses and disallow binding 397 * to them. 398 */ 399 sin6p = (struct sockaddr_in6 *)nam; 400 if (sin6p->sin6_family == AF_INET6 && 401 IN6_IS_ADDR_MULTICAST(&sin6p->sin6_addr)) { 402 error = EAFNOSUPPORT; 403 goto out; 404 } 405 error = in6_pcbbind(inp, nam, td); 406 if (error) 407 goto out; 408 COMMON_END(PRU_BIND); 409 } 410 #endif /* INET6 */ 411 412 struct netmsg_inswildcard { 413 struct netmsg_base base; 414 struct inpcb *nm_inp; 415 }; 416 417 static void 418 in_pcbinswildcardhash_handler(netmsg_t msg) 419 { 420 struct netmsg_inswildcard *nm = (struct netmsg_inswildcard *)msg; 421 int cpu = mycpuid, nextcpu; 422 423 in_pcbinswildcardhash_oncpu(nm->nm_inp, &tcbinfo[cpu]); 424 425 nextcpu = cpu + 1; 426 if (nextcpu < netisr_ncpus) 427 lwkt_forwardmsg(netisr_cpuport(nextcpu), &nm->base.lmsg); 428 else 429 lwkt_replymsg(&nm->base.lmsg, 0); 430 } 431 432 /* 433 * Prepare to accept connections. 434 */ 435 static void 436 tcp_usr_listen(netmsg_t msg) 437 { 438 struct socket *so = msg->listen.base.nm_so; 439 struct thread *td = msg->listen.nm_td; 440 int error = 0; 441 struct inpcb *inp; 442 struct tcpcb *tp; 443 struct netmsg_inswildcard nm; 444 lwkt_port_t port0 = netisr_cpuport(0); 445 446 COMMON_START(so, inp, 0); 447 448 if (&curthread->td_msgport != port0) { 449 lwkt_msg_t lmsg = &msg->listen.base.lmsg; 450 451 KASSERT((msg->listen.nm_flags & PRUL_RELINK) == 0, 452 ("already asked to relink")); 453 454 in_pcbunlink(so->so_pcb, &tcbinfo[mycpuid]); 455 msg->listen.nm_flags |= PRUL_RELINK; 456 457 TCP_STATE_MIGRATE_START(tp); 458 459 /* See the related comment in tcp_connect() */ 460 lwkt_setmsg_receipt(lmsg, tcp_sosetport); 461 lwkt_forwardmsg(port0, lmsg); 462 /* msg invalid now */ 463 return; 464 } 465 KASSERT(so->so_port == port0, ("so_port is not netisr0")); 466 467 if (msg->listen.nm_flags & PRUL_RELINK) { 468 msg->listen.nm_flags &= ~PRUL_RELINK; 469 TCP_STATE_MIGRATE_END(tp); 470 in_pcblink(so->so_pcb, &tcbinfo[mycpuid]); 471 } 472 KASSERT(inp->inp_pcbinfo == &tcbinfo[0], ("pcbinfo is not tcbinfo0")); 473 474 if (tp->t_flags & TF_LISTEN) 475 goto out; 476 477 if (inp->inp_lport == 0) { 478 error = in_pcbbind(inp, NULL, td); 479 if (error) 480 goto out; 481 } 482 483 TCP_STATE_CHANGE(tp, TCPS_LISTEN); 484 tp->t_flags |= TF_LISTEN; 485 tp->tt_msg = NULL; /* Catch any invalid timer usage */ 486 487 /* 488 * Create tcpcb per-cpu port cache 489 * 490 * NOTE: 491 * This _must_ be done before installing this inpcb into 492 * wildcard hash. 493 */ 494 tcp_pcbport_create(tp); 495 496 if (netisr_ncpus > 1) { 497 /* 498 * Put this inpcb into wildcard hash on other cpus. 499 */ 500 ASSERT_INP_NOTINHASH(inp); 501 netmsg_init(&nm.base, NULL, &curthread->td_msgport, 502 MSGF_PRIORITY, in_pcbinswildcardhash_handler); 503 nm.nm_inp = inp; 504 lwkt_domsg(netisr_cpuport(1), &nm.base.lmsg, 0); 505 } 506 in_pcbinswildcardhash(inp); 507 COMMON_END(PRU_LISTEN); 508 } 509 510 #ifdef INET6 511 512 static void 513 tcp6_usr_listen(netmsg_t msg) 514 { 515 struct socket *so = msg->listen.base.nm_so; 516 struct thread *td = msg->listen.nm_td; 517 int error = 0; 518 struct inpcb *inp; 519 struct tcpcb *tp; 520 struct netmsg_inswildcard nm; 521 522 COMMON_START(so, inp, 0); 523 524 if (tp->t_flags & TF_LISTEN) 525 goto out; 526 527 if (inp->inp_lport == 0) { 528 error = in6_pcbbind(inp, NULL, td); 529 if (error) 530 goto out; 531 } 532 533 TCP_STATE_CHANGE(tp, TCPS_LISTEN); 534 tp->t_flags |= TF_LISTEN; 535 tp->tt_msg = NULL; /* Catch any invalid timer usage */ 536 537 /* 538 * Create tcpcb per-cpu port cache 539 * 540 * NOTE: 541 * This _must_ be done before installing this inpcb into 542 * wildcard hash. 543 */ 544 tcp_pcbport_create(tp); 545 546 if (netisr_ncpus > 1) { 547 /* 548 * Put this inpcb into wildcard hash on other cpus. 549 */ 550 KKASSERT(so->so_port == netisr_cpuport(0)); 551 ASSERT_NETISR0; 552 KKASSERT(inp->inp_pcbinfo == &tcbinfo[0]); 553 ASSERT_INP_NOTINHASH(inp); 554 555 netmsg_init(&nm.base, NULL, &curthread->td_msgport, 556 MSGF_PRIORITY, in_pcbinswildcardhash_handler); 557 nm.nm_inp = inp; 558 lwkt_domsg(netisr_cpuport(1), &nm.base.lmsg, 0); 559 } 560 in_pcbinswildcardhash(inp); 561 COMMON_END(PRU_LISTEN); 562 } 563 #endif /* INET6 */ 564 565 /* 566 * Initiate connection to peer. 567 * Create a template for use in transmissions on this connection. 568 * Enter SYN_SENT state, and mark socket as connecting. 569 * Start keep-alive timer, and seed output sequence space. 570 * Send initial segment on connection. 571 */ 572 static void 573 tcp_usr_connect(netmsg_t msg) 574 { 575 struct socket *so = msg->connect.base.nm_so; 576 struct sockaddr *nam = msg->connect.nm_nam; 577 struct thread *td = msg->connect.nm_td; 578 int error = 0; 579 struct inpcb *inp; 580 struct tcpcb *tp; 581 struct sockaddr_in *sinp; 582 583 ASSERT_NETISR_NCPUS(mycpuid); 584 585 COMMON_START(so, inp, 0); 586 587 /* 588 * Must disallow TCP ``connections'' to multicast addresses. 589 */ 590 sinp = (struct sockaddr_in *)nam; 591 if (sinp->sin_family == AF_INET 592 && IN_MULTICAST(ntohl(sinp->sin_addr.s_addr))) { 593 error = EAFNOSUPPORT; 594 goto out; 595 } 596 597 if (!prison_remote_ip(td, (struct sockaddr*)sinp)) { 598 error = EAFNOSUPPORT; /* IPv6 only jail */ 599 goto out; 600 } 601 602 tcp_connect(msg); 603 /* msg is invalid now */ 604 return; 605 out: 606 if (msg->connect.nm_m) { 607 m_freem(msg->connect.nm_m); 608 msg->connect.nm_m = NULL; 609 } 610 if (msg->connect.nm_flags & PRUC_HELDTD) 611 lwkt_rele(td); 612 if (error && (msg->connect.nm_flags & PRUC_ASYNC)) { 613 so->so_error = error; 614 soisdisconnected(so); 615 } 616 lwkt_replymsg(&msg->lmsg, error); 617 } 618 619 #ifdef INET6 620 621 static void 622 tcp6_usr_connect(netmsg_t msg) 623 { 624 struct socket *so = msg->connect.base.nm_so; 625 struct sockaddr *nam = msg->connect.nm_nam; 626 struct thread *td = msg->connect.nm_td; 627 int error = 0; 628 struct inpcb *inp; 629 struct tcpcb *tp; 630 struct sockaddr_in6 *sin6p; 631 632 ASSERT_NETISR_NCPUS(mycpuid); 633 634 COMMON_START(so, inp, 0); 635 636 /* 637 * Must disallow TCP ``connections'' to multicast addresses. 638 */ 639 sin6p = (struct sockaddr_in6 *)nam; 640 if (sin6p->sin6_family == AF_INET6 641 && IN6_IS_ADDR_MULTICAST(&sin6p->sin6_addr)) { 642 error = EAFNOSUPPORT; 643 goto out; 644 } 645 646 if (!prison_remote_ip(td, nam)) { 647 error = EAFNOSUPPORT; /* IPv4 only jail */ 648 goto out; 649 } 650 651 /* Reject v4-mapped address */ 652 if (IN6_IS_ADDR_V4MAPPED(&sin6p->sin6_addr)) { 653 error = EADDRNOTAVAIL; 654 goto out; 655 } 656 657 inp->inp_inc.inc_isipv6 = 1; 658 tcp6_connect(msg); 659 /* msg is invalid now */ 660 return; 661 out: 662 if (msg->connect.nm_m) { 663 m_freem(msg->connect.nm_m); 664 msg->connect.nm_m = NULL; 665 } 666 lwkt_replymsg(&msg->lmsg, error); 667 } 668 669 #endif /* INET6 */ 670 671 /* 672 * Initiate disconnect from peer. 673 * If connection never passed embryonic stage, just drop; 674 * else if don't need to let data drain, then can just drop anyways, 675 * else have to begin TCP shutdown process: mark socket disconnecting, 676 * drain unread data, state switch to reflect user close, and 677 * send segment (e.g. FIN) to peer. Socket will be really disconnected 678 * when peer sends FIN and acks ours. 679 * 680 * SHOULD IMPLEMENT LATER PRU_CONNECT VIA REALLOC TCPCB. 681 */ 682 static void 683 tcp_usr_disconnect(netmsg_t msg) 684 { 685 struct socket *so = msg->disconnect.base.nm_so; 686 int error = 0; 687 struct inpcb *inp; 688 struct tcpcb *tp; 689 690 COMMON_START(so, inp, 1); 691 tp = tcp_disconnect(tp); 692 COMMON_END(PRU_DISCONNECT); 693 } 694 695 /* 696 * Accept a connection. Essentially all the work is 697 * done at higher levels; just return the address 698 * of the peer, storing through addr. 699 */ 700 static void 701 tcp_usr_accept(netmsg_t msg) 702 { 703 struct socket *so = msg->accept.base.nm_so; 704 struct sockaddr **nam = msg->accept.nm_nam; 705 int error = 0; 706 struct inpcb *inp; 707 struct tcpcb *tp = NULL; 708 TCPDEBUG0; 709 710 inp = so->so_pcb; 711 if (so->so_state & SS_ISDISCONNECTED) { 712 error = ECONNABORTED; 713 goto out; 714 } 715 if (inp == NULL) { 716 error = EINVAL; 717 goto out; 718 } 719 720 tp = intotcpcb(inp); 721 TCPDEBUG1(); 722 in_setpeeraddr(so, nam); 723 COMMON_END(PRU_ACCEPT); 724 } 725 726 #ifdef INET6 727 static void 728 tcp6_usr_accept(netmsg_t msg) 729 { 730 struct socket *so = msg->accept.base.nm_so; 731 struct sockaddr **nam = msg->accept.nm_nam; 732 int error = 0; 733 struct inpcb *inp; 734 struct tcpcb *tp = NULL; 735 TCPDEBUG0; 736 737 inp = so->so_pcb; 738 739 if (so->so_state & SS_ISDISCONNECTED) { 740 error = ECONNABORTED; 741 goto out; 742 } 743 if (inp == NULL) { 744 error = EINVAL; 745 goto out; 746 } 747 tp = intotcpcb(inp); 748 TCPDEBUG1(); 749 in6_setpeeraddr(so, nam); 750 COMMON_END(PRU_ACCEPT); 751 } 752 #endif /* INET6 */ 753 754 /* 755 * Mark the connection as being incapable of further output. 756 */ 757 static void 758 tcp_usr_shutdown(netmsg_t msg) 759 { 760 struct socket *so = msg->shutdown.base.nm_so; 761 int error = 0; 762 struct inpcb *inp; 763 struct tcpcb *tp; 764 765 COMMON_START(so, inp, 0); 766 socantsendmore(so); 767 tp = tcp_usrclosed(tp); 768 if (tp) 769 error = tcp_output(tp); 770 COMMON_END(PRU_SHUTDOWN); 771 } 772 773 /* 774 * After a receive, possibly send window update to peer. 775 */ 776 static void 777 tcp_usr_rcvd(netmsg_t msg) 778 { 779 struct socket *so = msg->rcvd.base.nm_so; 780 int error = 0, noreply = 0; 781 struct inpcb *inp; 782 struct tcpcb *tp; 783 784 COMMON_START(so, inp, 0); 785 786 if (msg->rcvd.nm_pru_flags & PRUR_ASYNC) { 787 noreply = 1; 788 so_async_rcvd_reply(so); 789 } 790 tcp_output(tp); 791 792 COMMON_END1(PRU_RCVD, noreply); 793 } 794 795 /* 796 * Do a send by putting data in output queue and updating urgent 797 * marker if URG set. Possibly send more data. Unlike the other 798 * pru_*() routines, the mbuf chains are our responsibility. We 799 * must either enqueue them or free them. The other pru_* routines 800 * generally are caller-frees. 801 */ 802 static void 803 tcp_usr_send(netmsg_t msg) 804 { 805 struct socket *so = msg->send.base.nm_so; 806 int flags = msg->send.nm_flags; 807 struct mbuf *m = msg->send.nm_m; 808 int error = 0; 809 struct inpcb *inp; 810 struct tcpcb *tp; 811 TCPDEBUG0; 812 813 KKASSERT(msg->send.nm_control == NULL); 814 KKASSERT(msg->send.nm_addr == NULL); 815 KKASSERT((flags & PRUS_FREEADDR) == 0); 816 817 inp = so->so_pcb; 818 819 if (inp == NULL) { 820 /* 821 * OOPS! we lost a race, the TCP session got reset after 822 * we checked SS_CANTSENDMORE, eg: while doing uiomove or a 823 * network interrupt in the non-critical section of sosend(). 824 */ 825 m_freem(m); 826 error = ECONNRESET; /* XXX EPIPE? */ 827 tp = NULL; 828 TCPDEBUG1(); 829 goto out; 830 } 831 tp = intotcpcb(inp); 832 TCPDEBUG1(); 833 834 #ifdef foo 835 /* 836 * This is no longer necessary, since: 837 * - sosendtcp() has already checked it for us 838 * - It does not work with asynchronized send 839 */ 840 841 /* 842 * Don't let too much OOB data build up 843 */ 844 if (flags & PRUS_OOB) { 845 if (ssb_space(&so->so_snd) < -512) { 846 m_freem(m); 847 error = ENOBUFS; 848 goto out; 849 } 850 } 851 #endif 852 853 /* 854 * Pump the data into the socket. 855 */ 856 if (m) { 857 ssb_appendstream(&so->so_snd, m); 858 sowwakeup(so); 859 } 860 if (flags & PRUS_OOB) { 861 /* 862 * According to RFC961 (Assigned Protocols), 863 * the urgent pointer points to the last octet 864 * of urgent data. We continue, however, 865 * to consider it to indicate the first octet 866 * of data past the urgent section. 867 * Otherwise, snd_up should be one lower. 868 */ 869 tp->snd_up = tp->snd_una + so->so_snd.ssb_cc; 870 tp->t_flags |= TF_FORCE; 871 error = tcp_output(tp); 872 tp->t_flags &= ~TF_FORCE; 873 } else { 874 if (flags & PRUS_EOF) { 875 /* 876 * Close the send side of the connection after 877 * the data is sent. 878 */ 879 socantsendmore(so); 880 tp = tcp_usrclosed(tp); 881 } 882 if (tp != NULL && !tcp_output_pending(tp)) { 883 if (flags & PRUS_MORETOCOME) 884 tp->t_flags |= TF_MORETOCOME; 885 error = tcp_output_fair(tp); 886 if (flags & PRUS_MORETOCOME) 887 tp->t_flags &= ~TF_MORETOCOME; 888 } 889 } 890 COMMON_END1((flags & PRUS_OOB) ? PRU_SENDOOB : 891 ((flags & PRUS_EOF) ? PRU_SEND_EOF : PRU_SEND), 892 (flags & PRUS_NOREPLY)); 893 } 894 895 /* 896 * NOTE: (so) is referenced from soabort*() and netmsg_pru_abort() 897 * will sofree() it when we return. 898 */ 899 static void 900 tcp_usr_abort(netmsg_t msg) 901 { 902 struct socket *so = msg->abort.base.nm_so; 903 int error = 0; 904 struct inpcb *inp; 905 struct tcpcb *tp; 906 907 COMMON_START(so, inp, 1); 908 tp = tcp_drop(tp, ECONNABORTED); 909 COMMON_END(PRU_ABORT); 910 } 911 912 /* 913 * Receive out-of-band data. 914 */ 915 static void 916 tcp_usr_rcvoob(netmsg_t msg) 917 { 918 struct socket *so = msg->rcvoob.base.nm_so; 919 struct mbuf *m = msg->rcvoob.nm_m; 920 int flags = msg->rcvoob.nm_flags; 921 int error = 0; 922 struct inpcb *inp; 923 struct tcpcb *tp; 924 925 COMMON_START(so, inp, 0); 926 if ((so->so_oobmark == 0 && 927 (so->so_state & SS_RCVATMARK) == 0) || 928 so->so_options & SO_OOBINLINE || 929 tp->t_oobflags & TCPOOB_HADDATA) { 930 error = EINVAL; 931 goto out; 932 } 933 if ((tp->t_oobflags & TCPOOB_HAVEDATA) == 0) { 934 error = EWOULDBLOCK; 935 goto out; 936 } 937 m->m_len = 1; 938 *mtod(m, caddr_t) = tp->t_iobc; 939 if ((flags & MSG_PEEK) == 0) 940 tp->t_oobflags ^= (TCPOOB_HAVEDATA | TCPOOB_HADDATA); 941 COMMON_END(PRU_RCVOOB); 942 } 943 944 static void 945 tcp_usr_savefaddr(struct socket *so, const struct sockaddr *faddr) 946 { 947 in_savefaddr(so, faddr); 948 } 949 950 #ifdef INET6 951 static void 952 tcp6_usr_savefaddr(struct socket *so, const struct sockaddr *faddr) 953 { 954 in6_savefaddr(so, faddr); 955 } 956 #endif 957 958 static int 959 tcp_usr_preconnect(struct socket *so, const struct sockaddr *nam, 960 struct thread *td __unused) 961 { 962 const struct sockaddr_in *sinp; 963 964 sinp = (const struct sockaddr_in *)nam; 965 if (sinp->sin_family == AF_INET && 966 IN_MULTICAST(ntohl(sinp->sin_addr.s_addr))) 967 return EAFNOSUPPORT; 968 969 soisconnecting(so); 970 return 0; 971 } 972 973 /* xxx - should be const */ 974 struct pr_usrreqs tcp_usrreqs = { 975 .pru_abort = tcp_usr_abort, 976 .pru_accept = tcp_usr_accept, 977 .pru_attach = tcp_usr_attach, 978 .pru_bind = tcp_usr_bind, 979 .pru_connect = tcp_usr_connect, 980 .pru_connect2 = pr_generic_notsupp, 981 .pru_control = in_control_dispatch, 982 .pru_detach = tcp_usr_detach, 983 .pru_disconnect = tcp_usr_disconnect, 984 .pru_listen = tcp_usr_listen, 985 .pru_peeraddr = in_setpeeraddr_dispatch, 986 .pru_rcvd = tcp_usr_rcvd, 987 .pru_rcvoob = tcp_usr_rcvoob, 988 .pru_send = tcp_usr_send, 989 .pru_sense = pru_sense_null, 990 .pru_shutdown = tcp_usr_shutdown, 991 .pru_sockaddr = in_setsockaddr_dispatch, 992 .pru_sosend = sosendtcp, 993 .pru_soreceive = sorecvtcp, 994 .pru_savefaddr = tcp_usr_savefaddr, 995 .pru_preconnect = tcp_usr_preconnect, 996 .pru_preattach = tcp_usr_preattach 997 }; 998 999 #ifdef INET6 1000 struct pr_usrreqs tcp6_usrreqs = { 1001 .pru_abort = tcp_usr_abort, 1002 .pru_accept = tcp6_usr_accept, 1003 .pru_attach = tcp_usr_attach, 1004 .pru_bind = tcp6_usr_bind, 1005 .pru_connect = tcp6_usr_connect, 1006 .pru_connect2 = pr_generic_notsupp, 1007 .pru_control = in6_control_dispatch, 1008 .pru_detach = tcp_usr_detach, 1009 .pru_disconnect = tcp_usr_disconnect, 1010 .pru_listen = tcp6_usr_listen, 1011 .pru_peeraddr = in6_setpeeraddr_dispatch, 1012 .pru_rcvd = tcp_usr_rcvd, 1013 .pru_rcvoob = tcp_usr_rcvoob, 1014 .pru_send = tcp_usr_send, 1015 .pru_sense = pru_sense_null, 1016 .pru_shutdown = tcp_usr_shutdown, 1017 .pru_sockaddr = in6_setsockaddr_dispatch, 1018 .pru_sosend = sosendtcp, 1019 .pru_soreceive = sorecvtcp, 1020 .pru_savefaddr = tcp6_usr_savefaddr 1021 }; 1022 #endif /* INET6 */ 1023 1024 static int 1025 tcp_connect_oncpu(struct tcpcb *tp, int flags, struct mbuf *m, 1026 const struct sockaddr_in *sin, struct sockaddr_in *if_sin, 1027 uint16_t hash) 1028 { 1029 struct inpcb *inp = tp->t_inpcb, *oinp; 1030 struct socket *so = inp->inp_socket; 1031 struct route *ro = &inp->inp_route; 1032 1033 KASSERT(inp->inp_pcbinfo == &tcbinfo[mycpu->gd_cpuid], 1034 ("pcbinfo mismatch")); 1035 1036 oinp = in_pcblookup_hash(inp->inp_pcbinfo, 1037 sin->sin_addr, sin->sin_port, 1038 (inp->inp_laddr.s_addr != INADDR_ANY ? 1039 inp->inp_laddr : if_sin->sin_addr), 1040 inp->inp_lport, 0, NULL); 1041 if (oinp != NULL) { 1042 m_freem(m); 1043 return (EADDRINUSE); 1044 } 1045 if (inp->inp_laddr.s_addr == INADDR_ANY) 1046 inp->inp_laddr = if_sin->sin_addr; 1047 KASSERT(inp->inp_faddr.s_addr == sin->sin_addr.s_addr, 1048 ("faddr mismatch for reconnect")); 1049 KASSERT(inp->inp_fport == sin->sin_port, 1050 ("fport mismatch for reconnect")); 1051 in_pcbinsconnhash(inp); 1052 1053 inp->inp_flags |= INP_HASH; 1054 inp->inp_hashval = hash; 1055 1056 /* 1057 * We are now on the inpcb's owner CPU, if the cached route was 1058 * freed because the rtentry's owner CPU is not the current CPU 1059 * (e.g. in tcp_connect()), then we try to reallocate it here with 1060 * the hope that a rtentry may be cloned from a RTF_PRCLONING 1061 * rtentry. 1062 */ 1063 if (!(inp->inp_socket->so_options & SO_DONTROUTE) && /*XXX*/ 1064 ro->ro_rt == NULL) { 1065 bzero(&ro->ro_dst, sizeof(struct sockaddr_in)); 1066 ro->ro_dst.sa_family = AF_INET; 1067 ro->ro_dst.sa_len = sizeof(struct sockaddr_in); 1068 ((struct sockaddr_in *)&ro->ro_dst)->sin_addr = 1069 sin->sin_addr; 1070 rtalloc(ro); 1071 } 1072 1073 /* 1074 * Now that no more errors can occur, change the protocol processing 1075 * port to the current thread (which is the correct thread). 1076 * 1077 * Create TCP timer message now; we are on the tcpcb's owner 1078 * CPU/thread. 1079 */ 1080 tcp_create_timermsg(tp, &curthread->td_msgport); 1081 1082 /* 1083 * Compute window scaling to request. Use a larger scaling then 1084 * needed for the initial receive buffer in case the receive buffer 1085 * gets expanded. 1086 */ 1087 if (tp->request_r_scale < TCP_MIN_WINSHIFT) 1088 tp->request_r_scale = TCP_MIN_WINSHIFT; 1089 while (tp->request_r_scale < TCP_MAX_WINSHIFT && 1090 (TCP_MAXWIN << tp->request_r_scale) < so->so_rcv.ssb_hiwat 1091 ) { 1092 tp->request_r_scale++; 1093 } 1094 1095 soisconnecting(so); 1096 tcpstat.tcps_connattempt++; 1097 TCP_STATE_CHANGE(tp, TCPS_SYN_SENT); 1098 tcp_callout_reset(tp, tp->tt_keep, tp->t_keepinit, tcp_timer_keep); 1099 tp->iss = tcp_new_isn(tp); 1100 tcp_sendseqinit(tp); 1101 if (m) { 1102 ssb_appendstream(&so->so_snd, m); 1103 m = NULL; 1104 if (flags & PRUS_OOB) 1105 tp->snd_up = tp->snd_una + so->so_snd.ssb_cc; 1106 } 1107 1108 /* 1109 * Close the send side of the connection after 1110 * the data is sent if flagged. 1111 */ 1112 if ((flags & (PRUS_OOB|PRUS_EOF)) == PRUS_EOF) { 1113 socantsendmore(so); 1114 tp = tcp_usrclosed(tp); 1115 } 1116 return (tcp_output(tp)); 1117 } 1118 1119 /* 1120 * Common subroutine to open a TCP connection to remote host specified 1121 * by struct sockaddr_in in mbuf *nam. Call in_pcbbind to assign a local 1122 * port number if needed. Call in_pcbladdr to do the routing and to choose 1123 * a local host address (interface). 1124 * Initialize connection parameters and enter SYN-SENT state. 1125 */ 1126 static void 1127 tcp_connect(netmsg_t msg) 1128 { 1129 struct socket *so = msg->connect.base.nm_so; 1130 struct sockaddr *nam = msg->connect.nm_nam; 1131 struct thread *td = msg->connect.nm_td; 1132 struct sockaddr_in *sin = (struct sockaddr_in *)nam; 1133 struct sockaddr_in *if_sin = NULL; 1134 struct inpcb *inp; 1135 struct tcpcb *tp; 1136 int error; 1137 uint16_t hash; 1138 lwkt_port_t port; 1139 1140 COMMON_START(so, inp, 0); 1141 1142 /* 1143 * Reconnect our pcb if we have to 1144 */ 1145 if (msg->connect.nm_flags & PRUC_RECONNECT) { 1146 msg->connect.nm_flags &= ~PRUC_RECONNECT; 1147 TCP_STATE_MIGRATE_END(tp); 1148 in_pcblink(so->so_pcb, &tcbinfo[mycpu->gd_cpuid]); 1149 } else { 1150 if (inp->inp_faddr.s_addr != INADDR_ANY) { 1151 kprintf("inpcb %p, double-connect race\n", inp); 1152 error = EISCONN; 1153 if (so->so_state & SS_ISCONNECTING) 1154 error = EALREADY; 1155 goto out; 1156 } 1157 KASSERT(inp->inp_fport == 0, ("invalid fport")); 1158 } 1159 1160 /* 1161 * Select local port, if it is not yet selected. 1162 */ 1163 if (inp->inp_lport == 0) { 1164 KKASSERT(inp->inp_laddr.s_addr == INADDR_ANY); 1165 1166 error = in_pcbladdr(inp, nam, &if_sin, td); 1167 if (error) 1168 goto out; 1169 inp->inp_laddr.s_addr = if_sin->sin_addr.s_addr; 1170 msg->connect.nm_flags |= PRUC_HASLADDR; 1171 1172 /* 1173 * Install faddr/fport earlier, so that when this 1174 * inpcb is installed on to the lport hash, the 1175 * 4-tuple contains correct value. 1176 * 1177 * NOTE: The faddr/fport will have to be installed 1178 * after the in_pcbladdr(), which may change them. 1179 */ 1180 inp->inp_faddr = sin->sin_addr; 1181 inp->inp_fport = sin->sin_port; 1182 1183 error = in_pcbbind_remote(inp, nam, td); 1184 if (error) 1185 goto out; 1186 } 1187 1188 if ((msg->connect.nm_flags & PRUC_HASLADDR) == 0) { 1189 /* 1190 * Rarely used path: 1191 * This inpcb was bound before this connect. 1192 */ 1193 error = in_pcbladdr(inp, nam, &if_sin, td); 1194 if (error) 1195 goto out; 1196 1197 /* 1198 * Save or refresh the faddr/fport, since they may 1199 * be changed by in_pcbladdr(). 1200 */ 1201 inp->inp_faddr = sin->sin_addr; 1202 inp->inp_fport = sin->sin_port; 1203 } 1204 #ifdef INVARIANTS 1205 else { 1206 KASSERT(inp->inp_faddr.s_addr == sin->sin_addr.s_addr, 1207 ("faddr mismatch for reconnect")); 1208 KASSERT(inp->inp_fport == sin->sin_port, 1209 ("fport mismatch for reconnect")); 1210 } 1211 #endif 1212 KKASSERT(inp->inp_socket == so); 1213 1214 hash = tcp_addrhash(sin->sin_addr.s_addr, sin->sin_port, 1215 (inp->inp_laddr.s_addr != INADDR_ANY ? 1216 inp->inp_laddr.s_addr : if_sin->sin_addr.s_addr), 1217 inp->inp_lport); 1218 port = netisr_hashport(hash); 1219 1220 if (port != &curthread->td_msgport) { 1221 lwkt_msg_t lmsg = &msg->connect.base.lmsg; 1222 1223 /* 1224 * in_pcbladdr() may have allocated a route entry for us 1225 * on the current CPU, but we need a route entry on the 1226 * inpcb's owner CPU, so free it here. 1227 */ 1228 in_pcbresetroute(inp); 1229 1230 /* 1231 * We are moving the protocol processing port the socket 1232 * is on, we have to unlink here and re-link on the 1233 * target cpu. 1234 */ 1235 in_pcbunlink(so->so_pcb, &tcbinfo[mycpu->gd_cpuid]); 1236 msg->connect.nm_flags |= PRUC_RECONNECT; 1237 msg->connect.base.nm_dispatch = tcp_connect; 1238 1239 TCP_STATE_MIGRATE_START(tp); 1240 1241 /* 1242 * Use message put done receipt to change this socket's 1243 * so_port, i.e. _after_ this message was put onto the 1244 * target netisr's msgport but _before_ the message could 1245 * be pulled from the target netisr's msgport, so that: 1246 * - The upper half (socket code) will not see the new 1247 * msgport before this message reaches the new msgport 1248 * and messages for this socket will be ordered. 1249 * - This message will see the new msgport, when its 1250 * handler is called in the target netisr. 1251 * 1252 * NOTE: 1253 * We MUST use messege put done receipt to change this 1254 * socket's so_port: 1255 * If we changed the so_port in this netisr after the 1256 * lwkt_forwardmsg (so messages for this socket will be 1257 * ordered) and changed the so_port in the target netisr 1258 * at the very beginning of this message's handler, we 1259 * would suffer so_port overwritten race, given this 1260 * message might be forwarded again. 1261 * 1262 * NOTE: 1263 * This mechanism depends on that the netisr's msgport 1264 * is spin msgport (currently it is :). 1265 * 1266 * If the upper half saw the new msgport before this 1267 * message reached the target netisr's msgport, the 1268 * messages sent from the upper half could reach the new 1269 * msgport before this message, thus there would be 1270 * message reordering. The worst case could be soclose() 1271 * saw the new msgport and the detach message could reach 1272 * the new msgport before this message, i.e. the inpcb 1273 * could have been destroyed when this message was still 1274 * pending on or on its way to the new msgport. Other 1275 * weird cases could also happen, e.g. inpcb->inp_pcbinfo, 1276 * since we have unlinked this inpcb from the current 1277 * pcbinfo first. 1278 */ 1279 lwkt_setmsg_receipt(lmsg, tcp_sosetport); 1280 lwkt_forwardmsg(port, lmsg); 1281 /* msg invalid now */ 1282 return; 1283 } else if (msg->connect.nm_flags & PRUC_HELDTD) { 1284 /* 1285 * The original thread is no longer needed; release it. 1286 */ 1287 lwkt_rele(td); 1288 msg->connect.nm_flags &= ~PRUC_HELDTD; 1289 } 1290 error = tcp_connect_oncpu(tp, msg->connect.nm_sndflags, 1291 msg->connect.nm_m, sin, if_sin, hash); 1292 msg->connect.nm_m = NULL; 1293 out: 1294 if (msg->connect.nm_m) { 1295 m_freem(msg->connect.nm_m); 1296 msg->connect.nm_m = NULL; 1297 } 1298 if (msg->connect.nm_flags & PRUC_HELDTD) 1299 lwkt_rele(td); 1300 if (error && (msg->connect.nm_flags & PRUC_ASYNC)) { 1301 so->so_error = error; 1302 soisdisconnected(so); 1303 } 1304 lwkt_replymsg(&msg->connect.base.lmsg, error); 1305 /* msg invalid now */ 1306 } 1307 1308 #ifdef INET6 1309 1310 static void 1311 tcp6_connect(netmsg_t msg) 1312 { 1313 struct tcpcb *tp; 1314 struct socket *so = msg->connect.base.nm_so; 1315 struct sockaddr *nam = msg->connect.nm_nam; 1316 struct thread *td = msg->connect.nm_td; 1317 struct inpcb *inp; 1318 struct sockaddr_in6 *sin6 = (struct sockaddr_in6 *)nam; 1319 struct in6_addr *addr6; 1320 lwkt_port_t port; 1321 int error; 1322 1323 COMMON_START(so, inp, 0); 1324 1325 /* 1326 * Reconnect our pcb if we have to 1327 */ 1328 if (msg->connect.nm_flags & PRUC_RECONNECT) { 1329 msg->connect.nm_flags &= ~PRUC_RECONNECT; 1330 TCP_STATE_MIGRATE_END(tp); 1331 in_pcblink(so->so_pcb, &tcbinfo[mycpu->gd_cpuid]); 1332 } 1333 1334 /* 1335 * Bind if we have to 1336 */ 1337 if (inp->inp_lport == 0) { 1338 error = in6_pcbbind(inp, NULL, td); 1339 if (error) 1340 goto out; 1341 } 1342 1343 /* 1344 * Cannot simply call in_pcbconnect, because there might be an 1345 * earlier incarnation of this same connection still in 1346 * TIME_WAIT state, creating an ADDRINUSE error. 1347 */ 1348 error = in6_pcbladdr(inp, nam, &addr6, td); 1349 if (error) 1350 goto out; 1351 1352 port = tcp6_addrport(); /* XXX hack for now, always cpu0 */ 1353 1354 if (port != &curthread->td_msgport) { 1355 lwkt_msg_t lmsg = &msg->connect.base.lmsg; 1356 1357 /* 1358 * in_pcbladdr() may have allocated a route entry for us 1359 * on the current CPU, but we need a route entry on the 1360 * inpcb's owner CPU, so free it here. 1361 */ 1362 in_pcbresetroute(inp); 1363 1364 in_pcbunlink(so->so_pcb, &tcbinfo[mycpu->gd_cpuid]); 1365 msg->connect.nm_flags |= PRUC_RECONNECT; 1366 msg->connect.base.nm_dispatch = tcp6_connect; 1367 1368 TCP_STATE_MIGRATE_START(tp); 1369 1370 /* See the related comment in tcp_connect() */ 1371 lwkt_setmsg_receipt(lmsg, tcp_sosetport); 1372 lwkt_forwardmsg(port, lmsg); 1373 /* msg invalid now */ 1374 return; 1375 } 1376 error = tcp6_connect_oncpu(tp, msg->connect.nm_sndflags, 1377 &msg->connect.nm_m, sin6, addr6); 1378 /* nm_m may still be intact */ 1379 out: 1380 if (msg->connect.nm_m) { 1381 m_freem(msg->connect.nm_m); 1382 msg->connect.nm_m = NULL; 1383 } 1384 lwkt_replymsg(&msg->connect.base.lmsg, error); 1385 /* msg invalid now */ 1386 } 1387 1388 static int 1389 tcp6_connect_oncpu(struct tcpcb *tp, int flags, struct mbuf **mp, 1390 struct sockaddr_in6 *sin6, struct in6_addr *addr6) 1391 { 1392 struct mbuf *m = *mp; 1393 struct inpcb *inp = tp->t_inpcb; 1394 struct socket *so = inp->inp_socket; 1395 struct inpcb *oinp; 1396 1397 /* 1398 * Cannot simply call in_pcbconnect, because there might be an 1399 * earlier incarnation of this same connection still in 1400 * TIME_WAIT state, creating an ADDRINUSE error. 1401 */ 1402 oinp = in6_pcblookup_hash(inp->inp_pcbinfo, 1403 &sin6->sin6_addr, sin6->sin6_port, 1404 (IN6_IS_ADDR_UNSPECIFIED(&inp->in6p_laddr) ? 1405 addr6 : &inp->in6p_laddr), 1406 inp->inp_lport, 0, NULL); 1407 if (oinp) 1408 return (EADDRINUSE); 1409 1410 if (IN6_IS_ADDR_UNSPECIFIED(&inp->in6p_laddr)) 1411 inp->in6p_laddr = *addr6; 1412 inp->in6p_faddr = sin6->sin6_addr; 1413 inp->inp_fport = sin6->sin6_port; 1414 if ((sin6->sin6_flowinfo & IPV6_FLOWINFO_MASK) != 0) 1415 inp->in6p_flowinfo = sin6->sin6_flowinfo; 1416 in_pcbinsconnhash(inp); 1417 1418 /* 1419 * Now that no more errors can occur, change the protocol processing 1420 * port to the current thread (which is the correct thread). 1421 * 1422 * Create TCP timer message now; we are on the tcpcb's owner 1423 * CPU/thread. 1424 */ 1425 tcp_create_timermsg(tp, &curthread->td_msgport); 1426 1427 /* Compute window scaling to request. */ 1428 if (tp->request_r_scale < TCP_MIN_WINSHIFT) 1429 tp->request_r_scale = TCP_MIN_WINSHIFT; 1430 while (tp->request_r_scale < TCP_MAX_WINSHIFT && 1431 (TCP_MAXWIN << tp->request_r_scale) < so->so_rcv.ssb_hiwat) { 1432 tp->request_r_scale++; 1433 } 1434 1435 soisconnecting(so); 1436 tcpstat.tcps_connattempt++; 1437 TCP_STATE_CHANGE(tp, TCPS_SYN_SENT); 1438 tcp_callout_reset(tp, tp->tt_keep, tp->t_keepinit, tcp_timer_keep); 1439 tp->iss = tcp_new_isn(tp); 1440 tcp_sendseqinit(tp); 1441 if (m) { 1442 ssb_appendstream(&so->so_snd, m); 1443 *mp = NULL; 1444 if (flags & PRUS_OOB) 1445 tp->snd_up = tp->snd_una + so->so_snd.ssb_cc; 1446 } 1447 1448 /* 1449 * Close the send side of the connection after 1450 * the data is sent if flagged. 1451 */ 1452 if ((flags & (PRUS_OOB|PRUS_EOF)) == PRUS_EOF) { 1453 socantsendmore(so); 1454 tp = tcp_usrclosed(tp); 1455 } 1456 return (tcp_output(tp)); 1457 } 1458 1459 #endif /* INET6 */ 1460 1461 /* 1462 * The new sockopt interface makes it possible for us to block in the 1463 * copyin/out step (if we take a page fault). Taking a page fault while 1464 * in a critical section is probably a Bad Thing. (Since sockets and pcbs 1465 * both now use TSM, there probably isn't any need for this function to 1466 * run in a critical section any more. This needs more examination.) 1467 */ 1468 void 1469 tcp_ctloutput(netmsg_t msg) 1470 { 1471 struct socket *so = msg->base.nm_so; 1472 struct sockopt *sopt = msg->ctloutput.nm_sopt; 1473 struct thread *td = NULL; 1474 int error, opt, optval, opthz; 1475 struct inpcb *inp; 1476 struct tcpcb *tp; 1477 1478 if (msg->ctloutput.nm_flags & PRCO_HELDTD) 1479 td = sopt->sopt_td; 1480 1481 error = 0; 1482 inp = so->so_pcb; 1483 if (inp == NULL) { 1484 error = ECONNRESET; 1485 goto done; 1486 } 1487 tp = intotcpcb(inp); 1488 1489 /* Get socket's owner cpuid hint */ 1490 if (sopt->sopt_level == SOL_SOCKET && 1491 sopt->sopt_dir == SOPT_GET && 1492 sopt->sopt_name == SO_CPUHINT) { 1493 if (tp->t_flags & TF_LISTEN) { 1494 /* 1495 * Listen sockets owner cpuid is always 0, 1496 * which does not make sense if SO_REUSEPORT 1497 * is not set. 1498 * 1499 * NOTE: inp_lgrpindex is _not_ assigned in jail. 1500 */ 1501 if ((so->so_options & SO_REUSEPORT) && 1502 inp->inp_lgrpindex >= 0) 1503 optval = inp->inp_lgrpindex % netisr_ncpus; 1504 else 1505 optval = -1; /* no hint */ 1506 } else { 1507 optval = mycpuid; 1508 } 1509 soopt_from_kbuf(sopt, &optval, sizeof(optval)); 1510 goto done; 1511 } 1512 1513 if (sopt->sopt_level != IPPROTO_TCP) { 1514 if (sopt->sopt_level == IPPROTO_IP) { 1515 switch (sopt->sopt_name) { 1516 case IP_MULTICAST_IF: 1517 case IP_MULTICAST_VIF: 1518 case IP_MULTICAST_TTL: 1519 case IP_MULTICAST_LOOP: 1520 case IP_ADD_MEMBERSHIP: 1521 case IP_DROP_MEMBERSHIP: 1522 /* 1523 * Multicast does not make sense on 1524 * TCP sockets. 1525 */ 1526 error = EOPNOTSUPP; 1527 goto done; 1528 } 1529 } 1530 #ifdef INET6 1531 if (INP_CHECK_SOCKAF(so, AF_INET6)) 1532 ip6_ctloutput_dispatch(msg); 1533 else 1534 #endif /* INET6 */ 1535 ip_ctloutput(msg); 1536 /* msg invalid now */ 1537 if (td != NULL) 1538 lwkt_rele(td); 1539 return; 1540 } 1541 1542 switch (sopt->sopt_dir) { 1543 case SOPT_SET: 1544 error = soopt_to_kbuf(sopt, &optval, sizeof optval, 1545 sizeof optval); 1546 if (error) 1547 break; 1548 switch (sopt->sopt_name) { 1549 case TCP_FASTKEEP: 1550 if (optval > 0) 1551 tp->t_keepidle = tp->t_keepintvl; 1552 else 1553 tp->t_keepidle = tcp_keepidle; 1554 tcp_timer_keep_activity(tp, 0); 1555 break; 1556 #ifdef TCP_SIGNATURE 1557 case TCP_SIGNATURE_ENABLE: 1558 if (tp->t_state == TCPS_CLOSED) { 1559 /* 1560 * This is the only safe state that this 1561 * option could be changed. Some segments 1562 * could already have been sent in other 1563 * states. 1564 */ 1565 if (optval > 0) 1566 tp->t_flags |= TF_SIGNATURE; 1567 else 1568 tp->t_flags &= ~TF_SIGNATURE; 1569 } else { 1570 error = EOPNOTSUPP; 1571 } 1572 break; 1573 #endif /* TCP_SIGNATURE */ 1574 case TCP_NODELAY: 1575 case TCP_NOOPT: 1576 switch (sopt->sopt_name) { 1577 case TCP_NODELAY: 1578 opt = TF_NODELAY; 1579 break; 1580 case TCP_NOOPT: 1581 opt = TF_NOOPT; 1582 break; 1583 default: 1584 opt = 0; /* dead code to fool gcc */ 1585 break; 1586 } 1587 1588 if (optval) 1589 tp->t_flags |= opt; 1590 else 1591 tp->t_flags &= ~opt; 1592 break; 1593 1594 case TCP_NOPUSH: 1595 if (tcp_disable_nopush) 1596 break; 1597 if (optval) 1598 tp->t_flags |= TF_NOPUSH; 1599 else { 1600 tp->t_flags &= ~TF_NOPUSH; 1601 error = tcp_output(tp); 1602 } 1603 break; 1604 1605 case TCP_MAXSEG: 1606 /* 1607 * Must be between 0 and maxseg. If the requested 1608 * maxseg is too small to satisfy the desired minmss, 1609 * pump it up (silently so sysctl modifications of 1610 * minmss do not create unexpected program failures). 1611 * Handle degenerate cases. 1612 */ 1613 if (optval > 0 && optval <= tp->t_maxseg) { 1614 if (optval + 40 < tcp_minmss) { 1615 optval = tcp_minmss - 40; 1616 if (optval < 0) 1617 optval = 1; 1618 } 1619 tp->t_maxseg = optval; 1620 } else { 1621 error = EINVAL; 1622 } 1623 break; 1624 1625 case TCP_KEEPINIT: 1626 opthz = ((int64_t)optval * hz) / 1000; 1627 if (opthz >= 1) 1628 tp->t_keepinit = opthz; 1629 else 1630 error = EINVAL; 1631 break; 1632 1633 case TCP_KEEPIDLE: 1634 opthz = ((int64_t)optval * hz) / 1000; 1635 if (opthz >= 1) { 1636 tp->t_keepidle = opthz; 1637 tcp_timer_keep_activity(tp, 0); 1638 } else { 1639 error = EINVAL; 1640 } 1641 break; 1642 1643 case TCP_KEEPINTVL: 1644 opthz = ((int64_t)optval * hz) / 1000; 1645 if (opthz >= 1) { 1646 tp->t_keepintvl = opthz; 1647 tp->t_maxidle = tp->t_keepintvl * tp->t_keepcnt; 1648 } else { 1649 error = EINVAL; 1650 } 1651 break; 1652 1653 case TCP_KEEPCNT: 1654 if (optval > 0) { 1655 tp->t_keepcnt = optval; 1656 tp->t_maxidle = tp->t_keepintvl * tp->t_keepcnt; 1657 } else { 1658 error = EINVAL; 1659 } 1660 break; 1661 1662 default: 1663 error = ENOPROTOOPT; 1664 break; 1665 } 1666 break; 1667 1668 case SOPT_GET: 1669 switch (sopt->sopt_name) { 1670 #ifdef TCP_SIGNATURE 1671 case TCP_SIGNATURE_ENABLE: 1672 optval = (tp->t_flags & TF_SIGNATURE) ? 1 : 0; 1673 break; 1674 #endif /* TCP_SIGNATURE */ 1675 case TCP_NODELAY: 1676 optval = tp->t_flags & TF_NODELAY; 1677 break; 1678 case TCP_MAXSEG: 1679 optval = tp->t_maxseg; 1680 break; 1681 case TCP_NOOPT: 1682 optval = tp->t_flags & TF_NOOPT; 1683 break; 1684 case TCP_NOPUSH: 1685 optval = tp->t_flags & TF_NOPUSH; 1686 break; 1687 case TCP_KEEPINIT: 1688 optval = ((int64_t)tp->t_keepinit * 1000) / hz; 1689 break; 1690 case TCP_KEEPIDLE: 1691 optval = ((int64_t)tp->t_keepidle * 1000) / hz; 1692 break; 1693 case TCP_KEEPINTVL: 1694 optval = ((int64_t)tp->t_keepintvl * 1000) / hz; 1695 break; 1696 case TCP_KEEPCNT: 1697 optval = tp->t_keepcnt; 1698 break; 1699 default: 1700 error = ENOPROTOOPT; 1701 break; 1702 } 1703 if (error == 0) 1704 soopt_from_kbuf(sopt, &optval, sizeof optval); 1705 break; 1706 } 1707 done: 1708 if (td != NULL) 1709 lwkt_rele(td); 1710 lwkt_replymsg(&msg->lmsg, error); 1711 } 1712 1713 struct netmsg_tcp_ctloutput { 1714 struct netmsg_pr_ctloutput ctloutput; 1715 struct sockopt sopt; 1716 int sopt_val; 1717 }; 1718 1719 /* 1720 * Allocate netmsg_pr_ctloutput for asynchronous tcp_ctloutput. 1721 */ 1722 struct netmsg_pr_ctloutput * 1723 tcp_ctloutmsg(struct sockopt *sopt) 1724 { 1725 struct netmsg_tcp_ctloutput *msg; 1726 int flags = 0, error; 1727 1728 KASSERT(sopt->sopt_dir == SOPT_SET, ("not from ctloutput")); 1729 1730 /* Only small set of options allows asynchronous setting. */ 1731 if (sopt->sopt_level != IPPROTO_TCP) 1732 return NULL; 1733 switch (sopt->sopt_name) { 1734 case TCP_NODELAY: 1735 case TCP_NOOPT: 1736 case TCP_NOPUSH: 1737 case TCP_FASTKEEP: 1738 break; 1739 default: 1740 return NULL; 1741 } 1742 1743 msg = kmalloc(sizeof(*msg), M_LWKTMSG, M_WAITOK | M_NULLOK); 1744 if (msg == NULL) { 1745 /* Fallback to synchronous tcp_ctloutput */ 1746 return NULL; 1747 } 1748 1749 /* Save the sockopt */ 1750 msg->sopt = *sopt; 1751 1752 /* Fixup the sopt.sopt_val ptr */ 1753 error = sooptcopyin(sopt, &msg->sopt_val, 1754 sizeof(msg->sopt_val), sizeof(msg->sopt_val)); 1755 if (error) { 1756 kfree(msg, M_LWKTMSG); 1757 return NULL; 1758 } 1759 msg->sopt.sopt_val = &msg->sopt_val; 1760 1761 /* Hold the current thread */ 1762 if (msg->sopt.sopt_td != NULL) { 1763 flags |= PRCO_HELDTD; 1764 lwkt_hold(msg->sopt.sopt_td); 1765 } 1766 1767 msg->ctloutput.nm_flags = flags; 1768 msg->ctloutput.nm_sopt = &msg->sopt; 1769 1770 return &msg->ctloutput; 1771 } 1772 1773 /* 1774 * tcp_sendspace and tcp_recvspace are the default send and receive window 1775 * sizes, respectively. These are obsolescent (this information should 1776 * be set by the route). 1777 * 1778 * Use a default that does not require tcp window scaling to be turned 1779 * on. Individual programs or the administrator can increase the default. 1780 */ 1781 u_long tcp_sendspace = 57344; /* largest multiple of PAGE_SIZE < 64k */ 1782 SYSCTL_INT(_net_inet_tcp, TCPCTL_SENDSPACE, sendspace, CTLFLAG_RW, 1783 &tcp_sendspace , 0, "Maximum outgoing TCP datagram size"); 1784 u_long tcp_recvspace = 57344; /* largest multiple of PAGE_SIZE < 64k */ 1785 SYSCTL_INT(_net_inet_tcp, TCPCTL_RECVSPACE, recvspace, CTLFLAG_RW, 1786 &tcp_recvspace , 0, "Maximum incoming TCP datagram size"); 1787 1788 /* 1789 * Attach TCP protocol to socket, allocating internet protocol control 1790 * block, tcp control block, buffer space, and entering CLOSED state. 1791 */ 1792 static int 1793 tcp_attach(struct socket *so, struct pru_attach_info *ai) 1794 { 1795 struct inpcb *inp; 1796 int error; 1797 int cpu; 1798 #ifdef INET6 1799 boolean_t isipv6 = INP_CHECK_SOCKAF(so, AF_INET6); 1800 #endif 1801 1802 if (ai != NULL) { 1803 error = tcp_usr_preattach(so, 0 /* don't care */, ai); 1804 if (error) 1805 return (error); 1806 } else { 1807 /* Post attach; do nothing */ 1808 } 1809 1810 cpu = mycpu->gd_cpuid; 1811 1812 /* 1813 * Set the default pcbinfo. This will likely change when we 1814 * bind/connect. 1815 */ 1816 error = in_pcballoc(so, &tcbinfo[cpu]); 1817 if (error) 1818 return (error); 1819 inp = so->so_pcb; 1820 #ifdef INET6 1821 if (isipv6) 1822 inp->in6p_hops = -1; /* use kernel default */ 1823 #endif 1824 tcp_newtcpcb(inp); 1825 /* Keep a reference for asynchronized pru_rcvd */ 1826 soreference(so); 1827 return (0); 1828 } 1829 1830 /* 1831 * Initiate (or continue) disconnect. 1832 * If embryonic state, just send reset (once). 1833 * If in ``let data drain'' option and linger null, just drop. 1834 * Otherwise (hard), mark socket disconnecting and drop 1835 * current input data; switch states based on user close, and 1836 * send segment to peer (with FIN). 1837 */ 1838 static struct tcpcb * 1839 tcp_disconnect(struct tcpcb *tp) 1840 { 1841 struct socket *so = tp->t_inpcb->inp_socket; 1842 1843 if (tp->t_state < TCPS_ESTABLISHED) { 1844 tp = tcp_close(tp); 1845 } else if ((so->so_options & SO_LINGER) && so->so_linger == 0) { 1846 tp = tcp_drop(tp, 0); 1847 } else { 1848 lwkt_gettoken(&so->so_rcv.ssb_token); 1849 soisdisconnecting(so); 1850 sbflush(&so->so_rcv.sb); 1851 tp = tcp_usrclosed(tp); 1852 if (tp) 1853 tcp_output(tp); 1854 lwkt_reltoken(&so->so_rcv.ssb_token); 1855 } 1856 return (tp); 1857 } 1858 1859 /* 1860 * User issued close, and wish to trail through shutdown states: 1861 * if never received SYN, just forget it. If got a SYN from peer, 1862 * but haven't sent FIN, then go to FIN_WAIT_1 state to send peer a FIN. 1863 * If already got a FIN from peer, then almost done; go to LAST_ACK 1864 * state. In all other cases, have already sent FIN to peer (e.g. 1865 * after PRU_SHUTDOWN), and just have to play tedious game waiting 1866 * for peer to send FIN or not respond to keep-alives, etc. 1867 * We can let the user exit from the close as soon as the FIN is acked. 1868 */ 1869 static struct tcpcb * 1870 tcp_usrclosed(struct tcpcb *tp) 1871 { 1872 1873 switch (tp->t_state) { 1874 1875 case TCPS_CLOSED: 1876 case TCPS_LISTEN: 1877 TCP_STATE_CHANGE(tp, TCPS_CLOSED); 1878 tp = tcp_close(tp); 1879 break; 1880 1881 case TCPS_SYN_SENT: 1882 case TCPS_SYN_RECEIVED: 1883 tp->t_flags |= TF_NEEDFIN; 1884 break; 1885 1886 case TCPS_ESTABLISHED: 1887 TCP_STATE_CHANGE(tp, TCPS_FIN_WAIT_1); 1888 break; 1889 1890 case TCPS_CLOSE_WAIT: 1891 TCP_STATE_CHANGE(tp, TCPS_LAST_ACK); 1892 break; 1893 } 1894 if (tp && tp->t_state >= TCPS_FIN_WAIT_2) { 1895 soisdisconnected(tp->t_inpcb->inp_socket); 1896 /* To prevent the connection hanging in FIN_WAIT_2 forever. */ 1897 if (tp->t_state == TCPS_FIN_WAIT_2) { 1898 tcp_callout_reset(tp, tp->tt_2msl, tp->t_maxidle, 1899 tcp_timer_2msl); 1900 } 1901 } 1902 return (tp); 1903 } 1904