xref: /netbsd/sys/net/npf/npf.c (revision 753f03b9)
1 /*-
2  * Copyright (c) 2009-2013 The NetBSD Foundation, Inc.
3  * All rights reserved.
4  *
5  * This material is based upon work partially supported by The
6  * NetBSD Foundation under a contract with Mindaugas Rasiukevicius.
7  *
8  * Redistribution and use in source and binary forms, with or without
9  * modification, are permitted provided that the following conditions
10  * are met:
11  * 1. Redistributions of source code must retain the above copyright
12  *    notice, this list of conditions and the following disclaimer.
13  * 2. Redistributions in binary form must reproduce the above copyright
14  *    notice, this list of conditions and the following disclaimer in the
15  *    documentation and/or other materials provided with the distribution.
16  *
17  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
18  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
19  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
20  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
21  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
22  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
23  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
24  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
25  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
26  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
27  * POSSIBILITY OF SUCH DAMAGE.
28  */
29 
30 /*
31  * NPF main: dynamic load/initialisation and unload routines.
32  */
33 
34 #ifdef _KERNEL
35 #include <sys/cdefs.h>
36 __KERNEL_RCSID(0, "$NetBSD: npf.c,v 1.44 2020/08/27 18:50:25 riastradh Exp $");
37 
38 #include <sys/param.h>
39 #include <sys/types.h>
40 
41 #include <sys/conf.h>
42 #include <sys/kmem.h>
43 #include <sys/percpu.h>
44 #include <sys/xcall.h>
45 #endif
46 
47 #include "npf_impl.h"
48 #include "npf_conn.h"
49 
50 static __read_mostly npf_t *	npf_kernel_ctx = NULL;
51 
52 __dso_public int
npfk_sysinit(unsigned nworkers)53 npfk_sysinit(unsigned nworkers)
54 {
55 
56 	npf_bpf_sysinit();
57 	npf_tableset_sysinit();
58 	npf_nat_sysinit();
59 	npf_portmap_sysinit();
60 	return npf_worker_sysinit(nworkers);
61 }
62 
63 __dso_public void
npfk_sysfini(void)64 npfk_sysfini(void)
65 {
66 
67 	npf_worker_sysfini();
68 	npf_portmap_sysfini();
69 	npf_nat_sysfini();
70 	npf_tableset_sysfini();
71 	npf_bpf_sysfini();
72 }
73 
74 __dso_public npf_t *
npfk_create(int flags,const npf_mbufops_t * mbufops,const npf_ifops_t * ifops,void * arg)75 npfk_create(int flags, const npf_mbufops_t *mbufops,
76     const npf_ifops_t *ifops, void *arg)
77 {
78 	npf_t *npf;
79 
80 	npf = kmem_zalloc(sizeof(npf_t), KM_SLEEP);
81 	npf->ebr = npf_ebr_create();
82 	npf->stats_percpu = percpu_alloc(NPF_STATS_SIZE);
83 	npf->mbufops = mbufops;
84 	npf->arg = arg;
85 
86 	npf_param_init(npf);
87 	npf_state_sysinit(npf);
88 	npf_ifmap_init(npf, ifops);
89 	npf_conn_init(npf);
90 	npf_portmap_init(npf);
91 	npf_alg_init(npf);
92 	npf_ext_init(npf);
93 
94 	/* Load an empty configuration. */
95 	npf_config_init(npf);
96 
97 	if ((flags & NPF_NO_GC) == 0) {
98 		npf_worker_enlist(npf);
99 	}
100 	return npf;
101 }
102 
103 __dso_public void
npfk_destroy(npf_t * npf)104 npfk_destroy(npf_t *npf)
105 {
106 	npf_worker_discharge(npf);
107 
108 	/*
109 	 * Destroy the current configuration.  Note: at this point all
110 	 * handlers must be deactivated; we will drain any processing.
111 	 */
112 	npf_config_fini(npf);
113 
114 	/* Finally, safe to destroy the subsystems. */
115 	npf_ext_fini(npf);
116 	npf_alg_fini(npf);
117 	npf_portmap_fini(npf);
118 	npf_conn_fini(npf);
119 	npf_ifmap_fini(npf);
120 	npf_state_sysfini(npf);
121 	npf_param_fini(npf);
122 
123 	npf_ebr_destroy(npf->ebr);
124 	percpu_free(npf->stats_percpu, NPF_STATS_SIZE);
125 	kmem_free(npf, sizeof(npf_t));
126 }
127 
128 
129 /*
130  * npfk_load: (re)load the configuration.
131  *
132  * => Will not modify the configuration reference.
133  */
134 __dso_public int
npfk_load(npf_t * npf,const void * config_ref,npf_error_t * err)135 npfk_load(npf_t *npf, const void *config_ref, npf_error_t *err)
136 {
137 	const nvlist_t *req = (const nvlist_t *)config_ref;
138 	nvlist_t *resp;
139 	int error;
140 
141 	resp = nvlist_create(0);
142 	error = npfctl_run_op(npf, IOC_NPF_LOAD, req, resp);
143 	nvlist_destroy(resp);
144 
145 	return error;
146 }
147 
148 __dso_public void
npfk_gc(npf_t * npf)149 npfk_gc(npf_t *npf)
150 {
151 	npf_conn_worker(npf);
152 }
153 
154 __dso_public void
npfk_thread_register(npf_t * npf)155 npfk_thread_register(npf_t *npf)
156 {
157 	npf_ebr_register(npf->ebr);
158 }
159 
160 __dso_public void
npfk_thread_unregister(npf_t * npf)161 npfk_thread_unregister(npf_t *npf)
162 {
163 	npf_ebr_full_sync(npf->ebr);
164 	npf_ebr_unregister(npf->ebr);
165 }
166 
167 __dso_public void *
npfk_getarg(npf_t * npf)168 npfk_getarg(npf_t *npf)
169 {
170 	return npf->arg;
171 }
172 
173 void
npf_setkernctx(npf_t * npf)174 npf_setkernctx(npf_t *npf)
175 {
176 	npf_kernel_ctx = npf;
177 }
178 
179 npf_t *
npf_getkernctx(void)180 npf_getkernctx(void)
181 {
182 	return npf_kernel_ctx;
183 }
184 
185 /*
186  * NPF statistics interface.
187  */
188 
189 void
npf_stats_inc(npf_t * npf,npf_stats_t st)190 npf_stats_inc(npf_t *npf, npf_stats_t st)
191 {
192 	uint64_t *stats = percpu_getref(npf->stats_percpu);
193 	stats[st]++;
194 	percpu_putref(npf->stats_percpu);
195 }
196 
197 void
npf_stats_dec(npf_t * npf,npf_stats_t st)198 npf_stats_dec(npf_t *npf, npf_stats_t st)
199 {
200 	uint64_t *stats = percpu_getref(npf->stats_percpu);
201 	stats[st]--;
202 	percpu_putref(npf->stats_percpu);
203 }
204 
205 static void
npf_stats_collect(void * mem,void * arg,struct cpu_info * ci)206 npf_stats_collect(void *mem, void *arg, struct cpu_info *ci)
207 {
208 	uint64_t *percpu_stats = mem, *full_stats = arg;
209 
210 	for (unsigned i = 0; i < NPF_STATS_COUNT; i++) {
211 		full_stats[i] += percpu_stats[i];
212 	}
213 }
214 
215 static void
npf_stats_clear_cb(void * mem,void * arg,struct cpu_info * ci)216 npf_stats_clear_cb(void *mem, void *arg, struct cpu_info *ci)
217 {
218 	uint64_t *percpu_stats = mem;
219 
220 	for (unsigned i = 0; i < NPF_STATS_COUNT; i++) {
221 		percpu_stats[i] = 0;
222 	}
223 }
224 
225 /*
226  * npf_stats: export collected statistics.
227  */
228 
229 __dso_public void
npfk_stats(npf_t * npf,uint64_t * buf)230 npfk_stats(npf_t *npf, uint64_t *buf)
231 {
232 	memset(buf, 0, NPF_STATS_SIZE);
233 	percpu_foreach_xcall(npf->stats_percpu, XC_HIGHPRI_IPL(IPL_SOFTNET),
234 	    npf_stats_collect, buf);
235 }
236 
237 __dso_public void
npfk_stats_clear(npf_t * npf)238 npfk_stats_clear(npf_t *npf)
239 {
240 	percpu_foreach_xcall(npf->stats_percpu, XC_HIGHPRI_IPL(IPL_SOFTNET),
241 	    npf_stats_clear_cb, NULL);
242 }
243