1 /*- 2 * SPDX-License-Identifier: BSD-2-Clause 3 * 4 * Copyright (c) 2021 Ng Peng Nam Sean 5 * Copyright (c) 2022 Alexander V. Chernikov <melifaro@FreeBSD.org> 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 1. Redistributions of source code must retain the above copyright 11 * notice, this list of conditions and the following disclaimer. 12 * 2. Redistributions in binary form must reproduce the above copyright 13 * notice, this list of conditions and the following disclaimer in the 14 * documentation and/or other materials provided with the distribution. 15 * 16 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 17 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 18 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 19 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 20 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 21 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 22 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 23 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 24 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 25 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 26 * SUCH DAMAGE. 27 */ 28 29 #include <sys/param.h> 30 #include <sys/ck.h> 31 #include <sys/lock.h> 32 #include <sys/malloc.h> 33 #include <sys/mbuf.h> 34 #include <sys/mutex.h> 35 #include <sys/socket.h> 36 #include <sys/socketvar.h> 37 #include <sys/syslog.h> 38 39 #include <netlink/netlink.h> 40 #include <netlink/netlink_ctl.h> 41 #include <netlink/netlink_linux.h> 42 #include <netlink/netlink_var.h> 43 44 #define DEBUG_MOD_NAME nl_io 45 #define DEBUG_MAX_LEVEL LOG_DEBUG3 46 #include <netlink/netlink_debug.h> 47 _DECLARE_DEBUG(LOG_INFO); 48 49 /* 50 * The logic below provide a p2p interface for receiving and 51 * sending netlink data between the kernel and userland. 52 */ 53 54 static bool nl_process_nbuf(struct nl_buf *nb, struct nlpcb *nlp); 55 56 struct nl_buf * 57 nl_buf_alloc(size_t len, int mflag) 58 { 59 struct nl_buf *nb; 60 61 nb = malloc(sizeof(struct nl_buf) + len, M_NETLINK, mflag); 62 if (__predict_true(nb != NULL)) { 63 nb->buflen = len; 64 nb->datalen = nb->offset = 0; 65 } 66 67 return (nb); 68 } 69 70 void 71 nl_buf_free(struct nl_buf *nb) 72 { 73 74 free(nb, M_NETLINK); 75 } 76 77 void 78 nl_schedule_taskqueue(struct nlpcb *nlp) 79 { 80 if (!nlp->nl_task_pending) { 81 nlp->nl_task_pending = true; 82 taskqueue_enqueue(nlp->nl_taskqueue, &nlp->nl_task); 83 NL_LOG(LOG_DEBUG3, "taskqueue scheduled"); 84 } else { 85 NL_LOG(LOG_DEBUG3, "taskqueue schedule skipped"); 86 } 87 } 88 89 static bool 90 nl_process_received_one(struct nlpcb *nlp) 91 { 92 struct socket *so = nlp->nl_socket; 93 struct sockbuf *sb; 94 struct nl_buf *nb; 95 bool reschedule = false; 96 97 NLP_LOCK(nlp); 98 nlp->nl_task_pending = false; 99 NLP_UNLOCK(nlp); 100 101 /* 102 * Do not process queued up requests if there is no space to queue 103 * replies. 104 */ 105 sb = &so->so_rcv; 106 SOCK_RECVBUF_LOCK(so); 107 if (sb->sb_hiwat <= sb->sb_ccc) { 108 SOCK_RECVBUF_UNLOCK(so); 109 return (false); 110 } 111 SOCK_RECVBUF_UNLOCK(so); 112 113 sb = &so->so_snd; 114 SOCK_SENDBUF_LOCK(so); 115 while ((nb = TAILQ_FIRST(&sb->nl_queue)) != NULL) { 116 TAILQ_REMOVE(&sb->nl_queue, nb, tailq); 117 SOCK_SENDBUF_UNLOCK(so); 118 reschedule = nl_process_nbuf(nb, nlp); 119 SOCK_SENDBUF_LOCK(so); 120 if (reschedule) { 121 sb->sb_acc -= nb->datalen; 122 sb->sb_ccc -= nb->datalen; 123 /* XXXGL: potentially can reduce lock&unlock count. */ 124 sowwakeup_locked(so); 125 nl_buf_free(nb); 126 SOCK_SENDBUF_LOCK(so); 127 } else { 128 TAILQ_INSERT_HEAD(&sb->nl_queue, nb, tailq); 129 break; 130 } 131 } 132 SOCK_SENDBUF_UNLOCK(so); 133 134 return (reschedule); 135 } 136 137 static void 138 nl_process_received(struct nlpcb *nlp) 139 { 140 NL_LOG(LOG_DEBUG3, "taskqueue called"); 141 142 if (__predict_false(nlp->nl_need_thread_setup)) { 143 nl_set_thread_nlp(curthread, nlp); 144 NLP_LOCK(nlp); 145 nlp->nl_need_thread_setup = false; 146 NLP_UNLOCK(nlp); 147 } 148 149 while (nl_process_received_one(nlp)) 150 ; 151 } 152 153 /* 154 * Called after some data have been read from the socket. 155 */ 156 void 157 nl_on_transmit(struct nlpcb *nlp) 158 { 159 NLP_LOCK(nlp); 160 161 struct socket *so = nlp->nl_socket; 162 if (__predict_false(nlp->nl_dropped_bytes > 0 && so != NULL)) { 163 unsigned long dropped_bytes = nlp->nl_dropped_bytes; 164 unsigned long dropped_messages = nlp->nl_dropped_messages; 165 nlp->nl_dropped_bytes = 0; 166 nlp->nl_dropped_messages = 0; 167 168 struct sockbuf *sb = &so->so_rcv; 169 NLP_LOG(LOG_DEBUG, nlp, 170 "socket RX overflowed, %lu messages (%lu bytes) dropped. " 171 "bytes: [%u/%u]", dropped_messages, dropped_bytes, 172 sb->sb_ccc, sb->sb_hiwat); 173 /* TODO: send netlink message */ 174 } 175 176 nl_schedule_taskqueue(nlp); 177 NLP_UNLOCK(nlp); 178 } 179 180 void 181 nl_taskqueue_handler(void *_arg, int pending) 182 { 183 struct nlpcb *nlp = (struct nlpcb *)_arg; 184 185 CURVNET_SET(nlp->nl_socket->so_vnet); 186 nl_process_received(nlp); 187 CURVNET_RESTORE(); 188 } 189 190 /* 191 * Tries to send current data buffer from writer. 192 * 193 * Returns true on success. 194 * If no queue overrunes happened, wakes up socket owner. 195 */ 196 bool 197 nl_send(struct nl_writer *nw, struct nlpcb *nlp) 198 { 199 struct socket *so = nlp->nl_socket; 200 struct sockbuf *sb = &so->so_rcv; 201 struct nl_buf *nb; 202 203 MPASS(nw->hdr == NULL); 204 205 IF_DEBUG_LEVEL(LOG_DEBUG2) { 206 struct nlmsghdr *hdr = (struct nlmsghdr *)nw->buf->data; 207 NLP_LOG(LOG_DEBUG2, nlp, 208 "TX len %u msgs %u msg type %d first hdrlen %u", 209 nw->buf->datalen, nw->num_messages, hdr->nlmsg_type, 210 hdr->nlmsg_len); 211 } 212 213 if (nlp->nl_linux && linux_netlink_p != NULL && 214 __predict_false(!linux_netlink_p->msgs_to_linux(nw, nlp))) { 215 nl_buf_free(nw->buf); 216 nw->buf = NULL; 217 return (false); 218 } 219 220 nb = nw->buf; 221 nw->buf = NULL; 222 223 SOCK_RECVBUF_LOCK(so); 224 if (!nw->ignore_limit && __predict_false(sb->sb_hiwat <= sb->sb_ccc)) { 225 SOCK_RECVBUF_UNLOCK(so); 226 NLP_LOCK(nlp); 227 nlp->nl_dropped_bytes += nb->datalen; 228 nlp->nl_dropped_messages += nw->num_messages; 229 NLP_LOG(LOG_DEBUG2, nlp, "RX oveflow: %lu m (+%d), %lu b (+%d)", 230 (unsigned long)nlp->nl_dropped_messages, nw->num_messages, 231 (unsigned long)nlp->nl_dropped_bytes, nb->datalen); 232 NLP_UNLOCK(nlp); 233 nl_buf_free(nb); 234 return (false); 235 } else { 236 bool full; 237 238 TAILQ_INSERT_TAIL(&sb->nl_queue, nb, tailq); 239 sb->sb_acc += nb->datalen; 240 sb->sb_ccc += nb->datalen; 241 full = sb->sb_hiwat <= sb->sb_ccc; 242 sorwakeup_locked(so); 243 if (full) { 244 NLP_LOCK(nlp); 245 nlp->nl_tx_blocked = true; 246 NLP_UNLOCK(nlp); 247 } 248 return (true); 249 } 250 } 251 252 static int 253 nl_receive_message(struct nlmsghdr *hdr, int remaining_length, 254 struct nlpcb *nlp, struct nl_pstate *npt) 255 { 256 nl_handler_f handler = nl_handlers[nlp->nl_proto].cb; 257 int error = 0; 258 259 NLP_LOG(LOG_DEBUG2, nlp, "msg len: %u type: %d: flags: 0x%X seq: %u pid: %u", 260 hdr->nlmsg_len, hdr->nlmsg_type, hdr->nlmsg_flags, hdr->nlmsg_seq, 261 hdr->nlmsg_pid); 262 263 if (__predict_false(hdr->nlmsg_len > remaining_length)) { 264 NLP_LOG(LOG_DEBUG, nlp, "message is not entirely present: want %d got %d", 265 hdr->nlmsg_len, remaining_length); 266 return (EINVAL); 267 } else if (__predict_false(hdr->nlmsg_len < sizeof(*hdr))) { 268 NL_LOG(LOG_DEBUG, "message too short: %d", hdr->nlmsg_len); 269 return (EINVAL); 270 } 271 /* Stamp each message with sender pid */ 272 hdr->nlmsg_pid = nlp->nl_port; 273 274 npt->hdr = hdr; 275 276 if (hdr->nlmsg_flags & NLM_F_REQUEST && hdr->nlmsg_type >= NLMSG_MIN_TYPE) { 277 NL_LOG(LOG_DEBUG2, "handling message with msg type: %d", 278 hdr->nlmsg_type); 279 280 if (nlp->nl_linux && linux_netlink_p != NULL) { 281 struct nlmsghdr *hdr_orig = hdr; 282 hdr = linux_netlink_p->msg_from_linux(nlp->nl_proto, hdr, npt); 283 if (hdr == NULL) { 284 /* Failed to translate to kernel format. Report an error back */ 285 hdr = hdr_orig; 286 npt->hdr = hdr; 287 if (hdr->nlmsg_flags & NLM_F_ACK) 288 nlmsg_ack(nlp, EOPNOTSUPP, hdr, npt); 289 return (0); 290 } 291 } 292 error = handler(hdr, npt); 293 NL_LOG(LOG_DEBUG2, "retcode: %d", error); 294 } 295 if ((hdr->nlmsg_flags & NLM_F_ACK) || (error != 0 && error != EINTR)) { 296 if (!npt->nw->suppress_ack) { 297 NL_LOG(LOG_DEBUG3, "ack"); 298 nlmsg_ack(nlp, error, hdr, npt); 299 } 300 } 301 302 return (0); 303 } 304 305 static void 306 npt_clear(struct nl_pstate *npt) 307 { 308 lb_clear(&npt->lb); 309 npt->error = 0; 310 npt->err_msg = NULL; 311 npt->err_off = 0; 312 npt->hdr = NULL; 313 npt->nw->suppress_ack = false; 314 } 315 316 /* 317 * Processes an incoming packet, which can contain multiple netlink messages 318 */ 319 static bool 320 nl_process_nbuf(struct nl_buf *nb, struct nlpcb *nlp) 321 { 322 struct nlmsghdr *hdr; 323 int error; 324 325 NL_LOG(LOG_DEBUG3, "RX netlink buf %p on %p", nb, nlp->nl_socket); 326 327 struct nl_writer nw = {}; 328 if (!nlmsg_get_unicast_writer(&nw, NLMSG_SMALL, nlp)) { 329 NL_LOG(LOG_DEBUG, "error allocating socket writer"); 330 return (true); 331 } 332 333 nlmsg_ignore_limit(&nw); 334 335 struct nl_pstate npt = { 336 .nlp = nlp, 337 .lb.base = &nb->data[roundup2(nb->datalen, 8)], 338 .lb.size = nb->buflen - roundup2(nb->datalen, 8), 339 .nw = &nw, 340 .strict = nlp->nl_flags & NLF_STRICT, 341 }; 342 343 for (; nb->offset + sizeof(struct nlmsghdr) <= nb->datalen;) { 344 hdr = (struct nlmsghdr *)&nb->data[nb->offset]; 345 /* Save length prior to calling handler */ 346 int msglen = NLMSG_ALIGN(hdr->nlmsg_len); 347 NL_LOG(LOG_DEBUG3, "parsing offset %d/%d", 348 nb->offset, nb->datalen); 349 npt_clear(&npt); 350 error = nl_receive_message(hdr, nb->datalen - nb->offset, nlp, 351 &npt); 352 nb->offset += msglen; 353 if (__predict_false(error != 0 || nlp->nl_tx_blocked)) 354 break; 355 } 356 NL_LOG(LOG_DEBUG3, "packet parsing done"); 357 nlmsg_flush(&nw); 358 359 if (nlp->nl_tx_blocked) { 360 NLP_LOCK(nlp); 361 nlp->nl_tx_blocked = false; 362 NLP_UNLOCK(nlp); 363 return (false); 364 } else 365 return (true); 366 } 367