xref: /linux/tools/testing/nvdimm/test/iomap.c (revision 0be3ff0c)
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Copyright(c) 2013-2015 Intel Corporation. All rights reserved.
4  */
5 #include <linux/memremap.h>
6 #include <linux/rculist.h>
7 #include <linux/export.h>
8 #include <linux/ioport.h>
9 #include <linux/module.h>
10 #include <linux/types.h>
11 #include <linux/pfn_t.h>
12 #include <linux/acpi.h>
13 #include <linux/io.h>
14 #include <linux/mm.h>
15 #include "nfit_test.h"
16 
17 static LIST_HEAD(iomap_head);
18 
19 static struct iomap_ops {
20 	nfit_test_lookup_fn nfit_test_lookup;
21 	nfit_test_evaluate_dsm_fn evaluate_dsm;
22 	struct list_head list;
23 } iomap_ops = {
24 	.list = LIST_HEAD_INIT(iomap_ops.list),
25 };
26 
27 void nfit_test_setup(nfit_test_lookup_fn lookup,
28 		nfit_test_evaluate_dsm_fn evaluate)
29 {
30 	iomap_ops.nfit_test_lookup = lookup;
31 	iomap_ops.evaluate_dsm = evaluate;
32 	list_add_rcu(&iomap_ops.list, &iomap_head);
33 }
34 EXPORT_SYMBOL(nfit_test_setup);
35 
36 void nfit_test_teardown(void)
37 {
38 	list_del_rcu(&iomap_ops.list);
39 	synchronize_rcu();
40 }
41 EXPORT_SYMBOL(nfit_test_teardown);
42 
43 static struct nfit_test_resource *__get_nfit_res(resource_size_t resource)
44 {
45 	struct iomap_ops *ops;
46 
47 	ops = list_first_or_null_rcu(&iomap_head, typeof(*ops), list);
48 	if (ops)
49 		return ops->nfit_test_lookup(resource);
50 	return NULL;
51 }
52 
53 struct nfit_test_resource *get_nfit_res(resource_size_t resource)
54 {
55 	struct nfit_test_resource *res;
56 
57 	rcu_read_lock();
58 	res = __get_nfit_res(resource);
59 	rcu_read_unlock();
60 
61 	return res;
62 }
63 EXPORT_SYMBOL(get_nfit_res);
64 
65 static void __iomem *__nfit_test_ioremap(resource_size_t offset, unsigned long size,
66 		void __iomem *(*fallback_fn)(resource_size_t, unsigned long))
67 {
68 	struct nfit_test_resource *nfit_res = get_nfit_res(offset);
69 
70 	if (nfit_res)
71 		return (void __iomem *) nfit_res->buf + offset
72 			- nfit_res->res.start;
73 	return fallback_fn(offset, size);
74 }
75 
76 void __iomem *__wrap_devm_ioremap(struct device *dev,
77 		resource_size_t offset, unsigned long size)
78 {
79 	struct nfit_test_resource *nfit_res = get_nfit_res(offset);
80 
81 	if (nfit_res)
82 		return (void __iomem *) nfit_res->buf + offset
83 			- nfit_res->res.start;
84 	return devm_ioremap(dev, offset, size);
85 }
86 EXPORT_SYMBOL(__wrap_devm_ioremap);
87 
88 void *__wrap_devm_memremap(struct device *dev, resource_size_t offset,
89 		size_t size, unsigned long flags)
90 {
91 	struct nfit_test_resource *nfit_res = get_nfit_res(offset);
92 
93 	if (nfit_res)
94 		return nfit_res->buf + offset - nfit_res->res.start;
95 	return devm_memremap(dev, offset, size, flags);
96 }
97 EXPORT_SYMBOL(__wrap_devm_memremap);
98 
99 static void nfit_test_kill(void *_pgmap)
100 {
101 	struct dev_pagemap *pgmap = _pgmap;
102 
103 	WARN_ON(!pgmap);
104 
105 	percpu_ref_kill(&pgmap->ref);
106 
107 	wait_for_completion(&pgmap->done);
108 	percpu_ref_exit(&pgmap->ref);
109 }
110 
111 static void dev_pagemap_percpu_release(struct percpu_ref *ref)
112 {
113 	struct dev_pagemap *pgmap = container_of(ref, struct dev_pagemap, ref);
114 
115 	complete(&pgmap->done);
116 }
117 
118 void *__wrap_devm_memremap_pages(struct device *dev, struct dev_pagemap *pgmap)
119 {
120 	int error;
121 	resource_size_t offset = pgmap->range.start;
122 	struct nfit_test_resource *nfit_res = get_nfit_res(offset);
123 
124 	if (!nfit_res)
125 		return devm_memremap_pages(dev, pgmap);
126 
127 	init_completion(&pgmap->done);
128 	error = percpu_ref_init(&pgmap->ref, dev_pagemap_percpu_release, 0,
129 				GFP_KERNEL);
130 	if (error)
131 		return ERR_PTR(error);
132 
133 	error = devm_add_action_or_reset(dev, nfit_test_kill, pgmap);
134 	if (error)
135 		return ERR_PTR(error);
136 	return nfit_res->buf + offset - nfit_res->res.start;
137 }
138 EXPORT_SYMBOL_GPL(__wrap_devm_memremap_pages);
139 
140 pfn_t __wrap_phys_to_pfn_t(phys_addr_t addr, unsigned long flags)
141 {
142 	struct nfit_test_resource *nfit_res = get_nfit_res(addr);
143 
144 	if (nfit_res)
145 		flags &= ~PFN_MAP;
146         return phys_to_pfn_t(addr, flags);
147 }
148 EXPORT_SYMBOL(__wrap_phys_to_pfn_t);
149 
150 void *__wrap_memremap(resource_size_t offset, size_t size,
151 		unsigned long flags)
152 {
153 	struct nfit_test_resource *nfit_res = get_nfit_res(offset);
154 
155 	if (nfit_res)
156 		return nfit_res->buf + offset - nfit_res->res.start;
157 	return memremap(offset, size, flags);
158 }
159 EXPORT_SYMBOL(__wrap_memremap);
160 
161 void __wrap_devm_memunmap(struct device *dev, void *addr)
162 {
163 	struct nfit_test_resource *nfit_res = get_nfit_res((long) addr);
164 
165 	if (nfit_res)
166 		return;
167 	return devm_memunmap(dev, addr);
168 }
169 EXPORT_SYMBOL(__wrap_devm_memunmap);
170 
171 void __iomem *__wrap_ioremap(resource_size_t offset, unsigned long size)
172 {
173 	return __nfit_test_ioremap(offset, size, ioremap);
174 }
175 EXPORT_SYMBOL(__wrap_ioremap);
176 
177 void __iomem *__wrap_ioremap_wc(resource_size_t offset, unsigned long size)
178 {
179 	return __nfit_test_ioremap(offset, size, ioremap_wc);
180 }
181 EXPORT_SYMBOL(__wrap_ioremap_wc);
182 
183 void __wrap_iounmap(volatile void __iomem *addr)
184 {
185 	struct nfit_test_resource *nfit_res = get_nfit_res((long) addr);
186 	if (nfit_res)
187 		return;
188 	return iounmap(addr);
189 }
190 EXPORT_SYMBOL(__wrap_iounmap);
191 
192 void __wrap_memunmap(void *addr)
193 {
194 	struct nfit_test_resource *nfit_res = get_nfit_res((long) addr);
195 
196 	if (nfit_res)
197 		return;
198 	return memunmap(addr);
199 }
200 EXPORT_SYMBOL(__wrap_memunmap);
201 
202 static bool nfit_test_release_region(struct device *dev,
203 		struct resource *parent, resource_size_t start,
204 		resource_size_t n);
205 
206 static void nfit_devres_release(struct device *dev, void *data)
207 {
208 	struct resource *res = *((struct resource **) data);
209 
210 	WARN_ON(!nfit_test_release_region(NULL, &iomem_resource, res->start,
211 			resource_size(res)));
212 }
213 
214 static int match(struct device *dev, void *__res, void *match_data)
215 {
216 	struct resource *res = *((struct resource **) __res);
217 	resource_size_t start = *((resource_size_t *) match_data);
218 
219 	return res->start == start;
220 }
221 
222 static bool nfit_test_release_region(struct device *dev,
223 		struct resource *parent, resource_size_t start,
224 		resource_size_t n)
225 {
226 	if (parent == &iomem_resource) {
227 		struct nfit_test_resource *nfit_res = get_nfit_res(start);
228 
229 		if (nfit_res) {
230 			struct nfit_test_request *req;
231 			struct resource *res = NULL;
232 
233 			if (dev) {
234 				devres_release(dev, nfit_devres_release, match,
235 						&start);
236 				return true;
237 			}
238 
239 			spin_lock(&nfit_res->lock);
240 			list_for_each_entry(req, &nfit_res->requests, list)
241 				if (req->res.start == start) {
242 					res = &req->res;
243 					list_del(&req->list);
244 					break;
245 				}
246 			spin_unlock(&nfit_res->lock);
247 
248 			WARN(!res || resource_size(res) != n,
249 					"%s: start: %llx n: %llx mismatch: %pr\n",
250 						__func__, start, n, res);
251 			if (res)
252 				kfree(req);
253 			return true;
254 		}
255 	}
256 	return false;
257 }
258 
259 static struct resource *nfit_test_request_region(struct device *dev,
260 		struct resource *parent, resource_size_t start,
261 		resource_size_t n, const char *name, int flags)
262 {
263 	struct nfit_test_resource *nfit_res;
264 
265 	if (parent == &iomem_resource) {
266 		nfit_res = get_nfit_res(start);
267 		if (nfit_res) {
268 			struct nfit_test_request *req;
269 			struct resource *res = NULL;
270 
271 			if (start + n > nfit_res->res.start
272 					+ resource_size(&nfit_res->res)) {
273 				pr_debug("%s: start: %llx n: %llx overflow: %pr\n",
274 						__func__, start, n,
275 						&nfit_res->res);
276 				return NULL;
277 			}
278 
279 			spin_lock(&nfit_res->lock);
280 			list_for_each_entry(req, &nfit_res->requests, list)
281 				if (start == req->res.start) {
282 					res = &req->res;
283 					break;
284 				}
285 			spin_unlock(&nfit_res->lock);
286 
287 			if (res) {
288 				WARN(1, "%pr already busy\n", res);
289 				return NULL;
290 			}
291 
292 			req = kzalloc(sizeof(*req), GFP_KERNEL);
293 			if (!req)
294 				return NULL;
295 			INIT_LIST_HEAD(&req->list);
296 			res = &req->res;
297 
298 			res->start = start;
299 			res->end = start + n - 1;
300 			res->name = name;
301 			res->flags = resource_type(parent);
302 			res->flags |= IORESOURCE_BUSY | flags;
303 			spin_lock(&nfit_res->lock);
304 			list_add(&req->list, &nfit_res->requests);
305 			spin_unlock(&nfit_res->lock);
306 
307 			if (dev) {
308 				struct resource **d;
309 
310 				d = devres_alloc(nfit_devres_release,
311 						sizeof(struct resource *),
312 						GFP_KERNEL);
313 				if (!d)
314 					return NULL;
315 				*d = res;
316 				devres_add(dev, d);
317 			}
318 
319 			pr_debug("%s: %pr\n", __func__, res);
320 			return res;
321 		}
322 	}
323 	if (dev)
324 		return __devm_request_region(dev, parent, start, n, name);
325 	return __request_region(parent, start, n, name, flags);
326 }
327 
328 struct resource *__wrap___request_region(struct resource *parent,
329 		resource_size_t start, resource_size_t n, const char *name,
330 		int flags)
331 {
332 	return nfit_test_request_region(NULL, parent, start, n, name, flags);
333 }
334 EXPORT_SYMBOL(__wrap___request_region);
335 
336 int __wrap_insert_resource(struct resource *parent, struct resource *res)
337 {
338 	if (get_nfit_res(res->start))
339 		return 0;
340 	return insert_resource(parent, res);
341 }
342 EXPORT_SYMBOL(__wrap_insert_resource);
343 
344 int __wrap_remove_resource(struct resource *res)
345 {
346 	if (get_nfit_res(res->start))
347 		return 0;
348 	return remove_resource(res);
349 }
350 EXPORT_SYMBOL(__wrap_remove_resource);
351 
352 struct resource *__wrap___devm_request_region(struct device *dev,
353 		struct resource *parent, resource_size_t start,
354 		resource_size_t n, const char *name)
355 {
356 	if (!dev)
357 		return NULL;
358 	return nfit_test_request_region(dev, parent, start, n, name, 0);
359 }
360 EXPORT_SYMBOL(__wrap___devm_request_region);
361 
362 void __wrap___release_region(struct resource *parent, resource_size_t start,
363 		resource_size_t n)
364 {
365 	if (!nfit_test_release_region(NULL, parent, start, n))
366 		__release_region(parent, start, n);
367 }
368 EXPORT_SYMBOL(__wrap___release_region);
369 
370 void __wrap___devm_release_region(struct device *dev, struct resource *parent,
371 		resource_size_t start, resource_size_t n)
372 {
373 	if (!nfit_test_release_region(dev, parent, start, n))
374 		__devm_release_region(dev, parent, start, n);
375 }
376 EXPORT_SYMBOL(__wrap___devm_release_region);
377 
378 acpi_status __wrap_acpi_evaluate_object(acpi_handle handle, acpi_string path,
379 		struct acpi_object_list *p, struct acpi_buffer *buf)
380 {
381 	struct nfit_test_resource *nfit_res = get_nfit_res((long) handle);
382 	union acpi_object **obj;
383 
384 	if (!nfit_res || strcmp(path, "_FIT") || !buf)
385 		return acpi_evaluate_object(handle, path, p, buf);
386 
387 	obj = nfit_res->buf;
388 	buf->length = sizeof(union acpi_object);
389 	buf->pointer = *obj;
390 	return AE_OK;
391 }
392 EXPORT_SYMBOL(__wrap_acpi_evaluate_object);
393 
394 union acpi_object * __wrap_acpi_evaluate_dsm(acpi_handle handle, const guid_t *guid,
395 		u64 rev, u64 func, union acpi_object *argv4)
396 {
397 	union acpi_object *obj = ERR_PTR(-ENXIO);
398 	struct iomap_ops *ops;
399 
400 	rcu_read_lock();
401 	ops = list_first_or_null_rcu(&iomap_head, typeof(*ops), list);
402 	if (ops)
403 		obj = ops->evaluate_dsm(handle, guid, rev, func, argv4);
404 	rcu_read_unlock();
405 
406 	if (IS_ERR(obj))
407 		return acpi_evaluate_dsm(handle, guid, rev, func, argv4);
408 	return obj;
409 }
410 EXPORT_SYMBOL(__wrap_acpi_evaluate_dsm);
411 
412 MODULE_LICENSE("GPL v2");
413