xref: /freebsd/sys/dev/liquidio/lio_ioctl.c (revision 06c3fb27)
1 /*
2  *   BSD LICENSE
3  *
4  *   Copyright(c) 2017 Cavium, Inc.. All rights reserved.
5  *   All rights reserved.
6  *
7  *   Redistribution and use in source and binary forms, with or without
8  *   modification, are permitted provided that the following conditions
9  *   are met:
10  *
11  *     * Redistributions of source code must retain the above copyright
12  *       notice, this list of conditions and the following disclaimer.
13  *     * Redistributions in binary form must reproduce the above copyright
14  *       notice, this list of conditions and the following disclaimer in
15  *       the documentation and/or other materials provided with the
16  *       distribution.
17  *     * Neither the name of Cavium, Inc. nor the names of its
18  *       contributors may be used to endorse or promote products derived
19  *       from this software without specific prior written permission.
20  *
21  *   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
22  *   "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
23  *   LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
24  *   A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
25  *   OWNER(S) OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
26  *   SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
27  *   LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
28  *   DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
29  *   THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
30  *   (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
31  *   OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
32  */
33 
34 #include "lio_bsd.h"
35 #include "lio_common.h"
36 #include "lio_droq.h"
37 #include "lio_iq.h"
38 #include "lio_response_manager.h"
39 #include "lio_device.h"
40 #include "lio_network.h"
41 #include "lio_ctrl.h"
42 #include "cn23xx_pf_device.h"
43 #include "lio_image.h"
44 #include "lio_ioctl.h"
45 #include "lio_main.h"
46 #include "lio_rxtx.h"
47 
48 static int	lio_set_rx_csum(if_t ifp, uint32_t data);
49 static int	lio_set_tso4(if_t ifp);
50 static int	lio_set_tso6(if_t ifp);
51 static int	lio_set_lro(if_t ifp);
52 static int	lio_change_mtu(if_t ifp, int new_mtu);
53 static int	lio_set_mcast_list(if_t ifp);
54 static inline enum	lio_ifflags lio_get_new_flags(if_t ifp);
55 
56 static inline bool
57 lio_is_valid_ether_addr(const uint8_t *addr)
58 {
59 
60 	return (!(0x01 & addr[0]) && !((addr[0] + addr[1] + addr[2] + addr[3] +
61 					addr[4] + addr[5]) == 0x00));
62 }
63 
64 static int
65 lio_change_dev_flags(if_t ifp)
66 {
67 	struct lio_ctrl_pkt	nctrl;
68 	struct lio		*lio = if_getsoftc(ifp);
69 	struct octeon_device	*oct = lio->oct_dev;
70 	int ret = 0;
71 
72 	bzero(&nctrl, sizeof(struct lio_ctrl_pkt));
73 
74 	/* Create a ctrl pkt command to be sent to core app. */
75 	nctrl.ncmd.cmd64 = 0;
76 	nctrl.ncmd.s.cmd = LIO_CMD_CHANGE_DEVFLAGS;
77 	nctrl.ncmd.s.param1 = lio_get_new_flags(ifp);
78 	nctrl.iq_no = lio->linfo.txpciq[0].s.q_no;
79 	nctrl.lio = lio;
80 	nctrl.cb_fn = lio_ctrl_cmd_completion;
81 
82 	ret = lio_send_ctrl_pkt(oct, &nctrl);
83 	if (ret)
84 		lio_dev_err(oct, "Failed to change flags ret %d\n", ret);
85 
86 	return (ret);
87 }
88 
89 /*
90  * lio_ioctl : User calls this routine for configuring
91  * the interface.
92  *
93  * return 0 on success, positive on failure
94  */
95 int
96 lio_ioctl(if_t ifp, u_long cmd, caddr_t data)
97 {
98 	struct lio	*lio = if_getsoftc(ifp);
99 	struct ifreq	*ifrequest = (struct ifreq *)data;
100 	int	error = 0;
101 
102 	switch (cmd) {
103 	case SIOCSIFADDR:
104 		lio_dev_dbg(lio->oct_dev, "ioctl: SIOCSIFADDR\n");
105 		if_setflagbits(ifp, IFF_UP, 0);
106 		error = ether_ioctl(ifp, cmd, data);
107 		break;
108 	case SIOCSIFMTU:
109 		lio_dev_dbg(lio->oct_dev, "ioctl: SIOCSIFMTU\n");
110 		error = lio_change_mtu(ifp, ifrequest->ifr_mtu);
111 		break;
112 	case SIOCSIFFLAGS:
113 		lio_dev_dbg(lio->oct_dev, "ioctl: SIOCSIFFLAGS\n");
114 		if (if_getflags(ifp) & IFF_UP) {
115 			if (if_getdrvflags(ifp) & IFF_DRV_RUNNING) {
116 				if ((if_getflags(ifp) ^ lio->if_flags) &
117 				    (IFF_PROMISC | IFF_ALLMULTI))
118 					error = lio_change_dev_flags(ifp);
119 			} else {
120 				if (!(atomic_load_acq_int(&lio->ifstate) &
121 				      LIO_IFSTATE_DETACH))
122 					lio_open(lio);
123 			}
124 		} else {
125 			if (if_getdrvflags(ifp) & IFF_DRV_RUNNING)
126 				lio_stop(ifp);
127 		}
128 		lio->if_flags = if_getflags(ifp);
129 		break;
130 	case SIOCADDMULTI:
131 		lio_dev_dbg(lio->oct_dev, "ioctl: SIOCADDMULTI\n");
132 		if (if_getdrvflags(ifp) & IFF_DRV_RUNNING)
133 			error = lio_set_mcast_list(ifp);
134 		break;
135 	case SIOCDELMULTI:
136 		lio_dev_dbg(lio->oct_dev, "ioctl: SIOCSIFMULTI\n");
137 		break;
138 	case SIOCSIFMEDIA:
139 		lio_dev_dbg(lio->oct_dev, "ioctl: SIOCSIFMEDIA\n");
140 	case SIOCGIFMEDIA:
141 		lio_dev_dbg(lio->oct_dev, "ioctl: SIOCGIFMEDIA\n");
142 	case SIOCGIFXMEDIA:
143 		lio_dev_dbg(lio->oct_dev, "ioctl: SIOCGIFXMEDIA\n");
144 		error = ifmedia_ioctl(ifp, ifrequest, &lio->ifmedia, cmd);
145 		break;
146 	case SIOCSIFCAP:
147 		{
148 			int	features = ifrequest->ifr_reqcap ^
149 					if_getcapenable(ifp);
150 
151 			lio_dev_dbg(lio->oct_dev, "ioctl: SIOCSIFCAP (Set Capabilities)\n");
152 
153 			if (!features)
154 				break;
155 
156 			if (features & IFCAP_TXCSUM) {
157 				if_togglecapenable(ifp, IFCAP_TXCSUM);
158 				if (if_getcapenable(ifp) & IFCAP_TXCSUM)
159 					if_sethwassistbits(ifp, (CSUM_TCP |
160 								 CSUM_UDP |
161 								 CSUM_IP), 0);
162 				else
163 					if_sethwassistbits(ifp, 0,
164 							(CSUM_TCP | CSUM_UDP |
165 							 CSUM_IP));
166 			}
167 			if (features & IFCAP_TXCSUM_IPV6) {
168 				if_togglecapenable(ifp, IFCAP_TXCSUM_IPV6);
169 				if (if_getcapenable(ifp) & IFCAP_TXCSUM_IPV6)
170 					if_sethwassistbits(ifp, (CSUM_UDP_IPV6 |
171 							   CSUM_TCP_IPV6), 0);
172 				else
173 					if_sethwassistbits(ifp, 0,
174 							   (CSUM_UDP_IPV6 |
175 							    CSUM_TCP_IPV6));
176 			}
177 			if (features & (IFCAP_RXCSUM | IFCAP_RXCSUM_IPV6))
178 				error |= lio_set_rx_csum(ifp, (features &
179 							       (IFCAP_RXCSUM |
180 							 IFCAP_RXCSUM_IPV6)));
181 
182 			if (features & IFCAP_TSO4)
183 				error |= lio_set_tso4(ifp);
184 
185 			if (features & IFCAP_TSO6)
186 				error |= lio_set_tso6(ifp);
187 
188 			if (features & IFCAP_LRO)
189 				error |= lio_set_lro(ifp);
190 
191 			if (features & IFCAP_VLAN_HWTAGGING)
192 				if_togglecapenable(ifp, IFCAP_VLAN_HWTAGGING);
193 
194 			if (features & IFCAP_VLAN_HWFILTER)
195 				if_togglecapenable(ifp, IFCAP_VLAN_HWFILTER);
196 
197 			if (features & IFCAP_VLAN_HWTSO)
198 				if_togglecapenable(ifp, IFCAP_VLAN_HWTSO);
199 
200 			VLAN_CAPABILITIES(ifp);
201 			break;
202 		}
203 	default:
204 		lio_dev_dbg(lio->oct_dev, "ioctl: UNKNOWN (0x%X)\n", (int)cmd);
205 		error = ether_ioctl(ifp, cmd, data);
206 		break;
207 	}
208 
209 	return (error);
210 }
211 
212 static int
213 lio_set_tso4(if_t ifp)
214 {
215 	struct lio	*lio = if_getsoftc(ifp);
216 
217 	if (if_getcapabilities(ifp) & IFCAP_TSO4) {
218 		if_togglecapenable(ifp, IFCAP_TSO4);
219 		if (if_getcapenable(ifp) & IFCAP_TSO4)
220 			if_sethwassistbits(ifp, CSUM_IP_TSO, 0);
221 		else
222 			if_sethwassistbits(ifp, 0, CSUM_IP_TSO);
223 	} else {
224 		lio_dev_info(lio->oct_dev, "TSO4 capability not supported\n");
225 		return (EINVAL);
226 	}
227 
228 	return (0);
229 }
230 
231 static int
232 lio_set_tso6(if_t ifp)
233 {
234 	struct lio	*lio = if_getsoftc(ifp);
235 
236 	if (if_getcapabilities(ifp) & IFCAP_TSO6) {
237 		if_togglecapenable(ifp, IFCAP_TSO6);
238 		if (if_getcapenable(ifp) & IFCAP_TSO6)
239 			if_sethwassistbits(ifp, CSUM_IP6_TSO, 0);
240 		else
241 			if_sethwassistbits(ifp, 0, CSUM_IP6_TSO);
242 	} else {
243 		lio_dev_info(lio->oct_dev, "TSO6 capability not supported\n");
244 		return (EINVAL);
245 	}
246 
247 	return (0);
248 }
249 
250 static int
251 lio_set_rx_csum(if_t ifp, uint32_t data)
252 {
253 	struct lio	*lio = if_getsoftc(ifp);
254 	int	ret = 0;
255 
256 	if (if_getcapabilities(ifp) & (IFCAP_RXCSUM | IFCAP_RXCSUM_IPV6)) {
257 		if_togglecapenable(ifp, (IFCAP_RXCSUM | IFCAP_RXCSUM_IPV6));
258 
259 		if (data) {
260 			/* LRO requires RXCSUM */
261 			if ((if_getcapabilities(ifp) & IFCAP_LRO) &&
262 			    (if_getcapenable(ifp) & IFCAP_LRO)) {
263 				ret = lio_set_feature(ifp, LIO_CMD_LRO_DISABLE,
264 						      LIO_LROIPV4 |
265 						      LIO_LROIPV6);
266 				if_togglecapenable(ifp, IFCAP_LRO);
267 			}
268 		}
269 	} else {
270 		lio_dev_info(lio->oct_dev, "Rx checksum offload capability not supported\n");
271 		return (ENODEV);
272 	}
273 
274 	return ((ret) ? EINVAL : 0);
275 }
276 
277 static int
278 lio_set_lro(if_t ifp)
279 {
280 	struct lio	*lio = if_getsoftc(ifp);
281 	int	ret = 0;
282 
283 	if (!(if_getcapabilities(ifp) & IFCAP_LRO)) {
284 		lio_dev_info(lio->oct_dev, "LRO capability not supported\n");
285 		return (ENODEV);
286 	}
287 
288 	if ((!(if_getcapenable(ifp) & IFCAP_LRO)) &&
289 	    (if_getcapenable(ifp) & IFCAP_RXCSUM) &&
290 	    (if_getcapenable(ifp) & IFCAP_RXCSUM_IPV6)) {
291 		if_togglecapenable(ifp, IFCAP_LRO);
292 
293 		if (lio_hwlro)
294 			ret = lio_set_feature(ifp, LIO_CMD_LRO_ENABLE, LIO_LROIPV4 |
295 					      LIO_LROIPV6);
296 
297 	} else if (if_getcapenable(ifp) & IFCAP_LRO) {
298 		if_togglecapenable(ifp, IFCAP_LRO);
299 
300 		if (lio_hwlro)
301 			ret = lio_set_feature(ifp, LIO_CMD_LRO_DISABLE, LIO_LROIPV4 |
302 					      LIO_LROIPV6);
303 	} else
304 		lio_dev_info(lio->oct_dev, "LRO requires RXCSUM");
305 
306 	return ((ret) ? EINVAL : 0);
307 }
308 
309 static void
310 lio_mtu_ctl_callback(struct octeon_device *oct, uint32_t status, void *buf)
311 {
312 	struct lio_soft_command	*sc = buf;
313 	volatile int		*mtu_sc_ctx;
314 
315 	mtu_sc_ctx = sc->ctxptr;
316 
317 	if (status) {
318 		lio_dev_err(oct, "MTU updation ctl instruction failed. Status: %llx\n",
319 			    LIO_CAST64(status));
320 		*mtu_sc_ctx = -1;
321 		/*
322 		 * This barrier is required to be sure that the
323 		 * response has been written fully.
324 		 */
325 		wmb();
326 		return;
327 	}
328 
329 	*mtu_sc_ctx = 1;
330 
331 	/*
332 	 * This barrier is required to be sure that the response has been
333 	 * written fully.
334 	 */
335 	wmb();
336 }
337 
338 /* @param ifp is network device */
339 static int
340 lio_change_mtu(if_t ifp, int new_mtu)
341 {
342 	struct lio		*lio = if_getsoftc(ifp);
343 	struct octeon_device	*oct = lio->oct_dev;
344 	struct lio_soft_command	*sc;
345 	union octeon_cmd	*ncmd;
346 	volatile int		*mtu_sc_ctx;
347 	int	retval = 0;
348 
349 	if (lio->mtu == new_mtu)
350 		return (0);
351 
352 	/*
353 	 * Limit the MTU to make sure the ethernet packets are between
354 	 * LIO_MIN_MTU_SIZE bytes and LIO_MAX_MTU_SIZE bytes
355 	 */
356 	if ((new_mtu < LIO_MIN_MTU_SIZE) || (new_mtu > LIO_MAX_MTU_SIZE)) {
357 		lio_dev_err(oct, "Invalid MTU: %d\n", new_mtu);
358 		lio_dev_err(oct, "Valid range %d and %d\n",
359 			    LIO_MIN_MTU_SIZE, LIO_MAX_MTU_SIZE);
360 		return (EINVAL);
361 	}
362 
363 	sc = lio_alloc_soft_command(oct, OCTEON_CMD_SIZE, 16,
364 				    sizeof(*mtu_sc_ctx));
365 	if (sc == NULL)
366 		return (ENOMEM);
367 
368 	ncmd = (union octeon_cmd *)sc->virtdptr;
369 	mtu_sc_ctx = sc->ctxptr;
370 
371 	*mtu_sc_ctx = 0;
372 
373 	ncmd->cmd64 = 0;
374 	ncmd->s.cmd = LIO_CMD_CHANGE_MTU;
375 	ncmd->s.param1 = new_mtu;
376 
377 	lio_swap_8B_data((uint64_t *)ncmd, (OCTEON_CMD_SIZE >> 3));
378 
379 	sc->iq_no = lio->linfo.txpciq[0].s.q_no;
380 
381 	lio_prepare_soft_command(oct, sc, LIO_OPCODE_NIC,
382 				 LIO_OPCODE_NIC_CMD, 0, 0, 0);
383 
384 	sc->callback = lio_mtu_ctl_callback;
385 	sc->callback_arg = sc;
386 	sc->wait_time = 5000;
387 
388 	retval = lio_send_soft_command(oct, sc);
389 	if (retval == LIO_IQ_SEND_FAILED) {
390 		lio_dev_info(oct,
391 			     "Failed to send MTU update Control message\n");
392 		retval = EBUSY;
393 		goto mtu_updation_failed;
394 	}
395 
396 	/*
397 	 * Sleep on a wait queue till the cond flag indicates that the
398 	 * response arrived or timed-out.
399 	 */
400 	lio_sleep_cond(oct, mtu_sc_ctx);
401 
402 	if (*mtu_sc_ctx < 0) {
403 		retval = EBUSY;
404 		goto mtu_updation_failed;
405 	}
406 	lio_dev_info(oct, "MTU Changed from %d to %d\n", if_getmtu(ifp),
407 		     new_mtu);
408 	if_setmtu(ifp, new_mtu);
409 	lio->mtu = new_mtu;
410 	retval = 0;			/*
411 				         * this updation is make sure that LIO_IQ_SEND_STOP case
412 				         * also success
413 				         */
414 
415 mtu_updation_failed:
416 	lio_free_soft_command(oct, sc);
417 
418 	return (retval);
419 }
420 
421 /* @param ifp network device */
422 int
423 lio_set_mac(if_t ifp, uint8_t *p)
424 {
425 	struct lio_ctrl_pkt	nctrl;
426 	struct lio		*lio = if_getsoftc(ifp);
427 	struct octeon_device	*oct = lio->oct_dev;
428 	int	ret = 0;
429 
430 	if (!lio_is_valid_ether_addr(p))
431 		return (EADDRNOTAVAIL);
432 
433 	bzero(&nctrl, sizeof(struct lio_ctrl_pkt));
434 
435 	nctrl.ncmd.cmd64 = 0;
436 	nctrl.ncmd.s.cmd = LIO_CMD_CHANGE_MACADDR;
437 	nctrl.ncmd.s.param1 = 0;
438 	nctrl.ncmd.s.more = 1;
439 	nctrl.iq_no = lio->linfo.txpciq[0].s.q_no;
440 	nctrl.lio = lio;
441 	nctrl.cb_fn = lio_ctrl_cmd_completion;
442 	nctrl.wait_time = 100;
443 
444 	nctrl.udd[0] = 0;
445 	/* The MAC Address is presented in network byte order. */
446 	memcpy((uint8_t *)&nctrl.udd[0] + 2, p, ETHER_HDR_LEN);
447 
448 	ret = lio_send_ctrl_pkt(lio->oct_dev, &nctrl);
449 	if (ret < 0) {
450 		lio_dev_err(oct, "MAC Address change failed\n");
451 		return (ENOMEM);
452 	}
453 
454 	memcpy(((uint8_t *)&lio->linfo.hw_addr) + 2, p, ETHER_HDR_LEN);
455 
456 	return (0);
457 }
458 
459 /*
460  * \brief Converts a mask based on ifp flags
461  * @param ifp network device
462  *
463  * This routine generates a lio_ifflags mask from the ifp flags
464  * received from the OS.
465  */
466 static inline enum lio_ifflags
467 lio_get_new_flags(if_t ifp)
468 {
469 	enum lio_ifflags f = LIO_IFFLAG_UNICAST;
470 
471 	if (if_getflags(ifp) & IFF_PROMISC)
472 		f |= LIO_IFFLAG_PROMISC;
473 
474 	if (if_getflags(ifp) & IFF_ALLMULTI)
475 		f |= LIO_IFFLAG_ALLMULTI;
476 
477 	if (if_getflags(ifp) & IFF_MULTICAST) {
478 		f |= LIO_IFFLAG_MULTICAST;
479 
480 		/*
481 		 * Accept all multicast addresses if there are more than we
482 		 * can handle
483 		 */
484 		if (if_getamcount(ifp) > LIO_MAX_MULTICAST_ADDR)
485 			f |= LIO_IFFLAG_ALLMULTI;
486 	}
487 	if (if_getflags(ifp) & IFF_BROADCAST)
488 		f |= LIO_IFFLAG_BROADCAST;
489 
490 	return (f);
491 }
492 
493 static u_int
494 lio_copy_maddr(void *arg, struct sockaddr_dl *sdl, u_int cnt)
495 {
496 	uint64_t *mc = arg;
497 
498 	if (cnt == LIO_MAX_MULTICAST_ADDR)
499 		return (0);
500 
501 	mc += cnt;
502 	*mc = 0;
503 	memcpy(((uint8_t *)mc) + 2, LLADDR(sdl), ETHER_ADDR_LEN);
504 	/* no need to swap bytes */
505 
506 	return (1);
507 }
508 
509 /* @param ifp network device */
510 static int
511 lio_set_mcast_list(if_t ifp)
512 {
513 	struct lio		*lio = if_getsoftc(ifp);
514 	struct octeon_device	*oct = lio->oct_dev;
515 	struct lio_ctrl_pkt	nctrl;
516 	int	mc_count;
517 	int	ret;
518 
519 	bzero(&nctrl, sizeof(struct lio_ctrl_pkt));
520 
521 	/* Create a ctrl pkt command to be sent to core app. */
522 	nctrl.ncmd.cmd64 = 0;
523 	nctrl.ncmd.s.cmd = LIO_CMD_SET_MULTI_LIST;
524 	nctrl.ncmd.s.param1 = lio_get_new_flags(ifp);
525 	nctrl.iq_no = lio->linfo.txpciq[0].s.q_no;
526 	nctrl.lio = lio;
527 	nctrl.cb_fn = lio_ctrl_cmd_completion;
528 
529 	/* copy all the addresses into the udd */
530 	mc_count = if_foreach_llmaddr(ifp, lio_copy_maddr, &nctrl.udd[0]);
531 
532 	/*
533 	 * Apparently, any activity in this call from the kernel has to
534 	 * be atomic. So we won't wait for response.
535 	 */
536 	nctrl.wait_time = 0;
537 	nctrl.ncmd.s.param2 = mc_count;
538 	nctrl.ncmd.s.more = mc_count;
539 
540 	ret = lio_send_ctrl_pkt(lio->oct_dev, &nctrl);
541 	if (ret < 0) {
542 		lio_dev_err(oct, "DEVFLAGS change failed in core (ret: 0x%x)\n",
543 			    ret);
544 	}
545 
546 	return ((ret) ? EINVAL : 0);
547 }
548