1 /*-
2  * Copyright (c) 2010 Isilon Systems, Inc.
3  * Copyright (c) 2010 iX Systems, Inc.
4  * Copyright (c) 2010 Panasas, Inc.
5  * Copyright (c) 2013-2017 Mellanox Technologies, Ltd.
6  * Copyright (c) 2015 Matthew Dillon <dillon@backplane.com>
7  * All rights reserved.
8  *
9  * Redistribution and use in source and binary forms, with or without
10  * modification, are permitted provided that the following conditions
11  * are met:
12  * 1. Redistributions of source code must retain the above copyright
13  *    notice unmodified, this list of conditions, and the following
14  *    disclaimer.
15  * 2. Redistributions in binary form must reproduce the above copyright
16  *    notice, this list of conditions and the following disclaimer in the
17  *    documentation and/or other materials provided with the distribution.
18  *
19  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
20  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
21  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
22  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
23  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
24  * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
25  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
26  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
27  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
28  * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
29  *
30  * $FreeBSD$
31  */
32 #ifndef	_LINUX_SCATTERLIST_H_
33 #define	_LINUX_SCATTERLIST_H_
34 
35 #include <sys/types.h>
36 #include <sys/sf_buf.h>
37 
38 #include <linux/page.h>
39 #include <linux/slab.h>
40 #include <linux/mm.h>
41 
42 struct bus_dmamap;
43 struct scatterlist {
44 	unsigned long page_link;
45 #define	SG_PAGE_LINK_CHAIN	0x1UL
46 #define	SG_PAGE_LINK_LAST	0x2UL
47 #define	SG_PAGE_LINK_MASK	0x3UL
48 	unsigned int offset;
49 	unsigned int length;
50 	dma_addr_t dma_address;
51 	struct bus_dmamap *dma_map;	/* FreeBSD specific */
52 };
53 
54 CTASSERT((sizeof(struct scatterlist) & SG_PAGE_LINK_MASK) == 0);
55 
56 struct sg_table {
57 	struct scatterlist *sgl;
58 	unsigned int nents;
59 	unsigned int orig_nents;
60 };
61 
62 struct sg_page_iter {
63 	struct scatterlist *sg;
64 	unsigned int sg_pgoffset;
65 	unsigned int maxents;
66 	struct {
67 		unsigned int nents;
68 		int	pg_advance;
69 	} internal;
70 };
71 
72 struct sg_dma_page_iter {
73 	struct sg_page_iter base;
74 };
75 
76 #define	SCATTERLIST_MAX_SEGMENT	(-1U & ~(PAGE_SIZE - 1))
77 
78 #define	SG_MAX_SINGLE_ALLOC	(PAGE_SIZE / sizeof(struct scatterlist))
79 
80 #define	SG_MAGIC		0x87654321UL
81 #define	SG_CHAIN		SG_PAGE_LINK_CHAIN
82 #define	SG_END			SG_PAGE_LINK_LAST
83 
84 #define	sg_is_chain(sg)		((sg)->page_link & SG_PAGE_LINK_CHAIN)
85 #define	sg_is_last(sg)		((sg)->page_link & SG_PAGE_LINK_LAST)
86 #define	sg_chain_ptr(sg)	\
87 	((struct scatterlist *) ((sg)->page_link & ~SG_PAGE_LINK_MASK))
88 
89 #define	sg_dma_address(sg)	(sg)->dma_address
90 #define	sg_dma_len(sg)		(sg)->length
91 
92 #define	for_each_sg_page(sgl, iter, nents, pgoffset)			\
93 	for (_sg_iter_init(sgl, iter, nents, pgoffset);			\
94 	     (iter)->sg; _sg_iter_next(iter))
95 #define	for_each_sg_dma_page(sgl, iter, nents, pgoffset) 		\
96 	for_each_sg_page(sgl, &(iter)->base, nents, pgoffset)
97 
98 #define	for_each_sg(sglist, sg, sgmax, iter)				\
99 	for (iter = 0, sg = (sglist); iter < (sgmax); iter++, sg = sg_next(sg))
100 
101 typedef struct scatterlist *(sg_alloc_fn) (unsigned int, gfp_t);
102 typedef void (sg_free_fn) (struct scatterlist *, unsigned int);
103 
104 static inline void
105 sg_assign_page(struct scatterlist *sg, struct page *page)
106 {
107 	unsigned long page_link = sg->page_link & SG_PAGE_LINK_MASK;
108 
109 	sg->page_link = page_link | (unsigned long)page;
110 }
111 
112 static inline void
113 sg_set_page(struct scatterlist *sg, struct page *page, unsigned int len,
114     unsigned int offset)
115 {
116 	sg_assign_page(sg, page);
117 	sg->offset = offset;
118 	sg->length = len;
119 }
120 
121 static inline struct page *
122 sg_page(struct scatterlist *sg)
123 {
124 	return ((struct page *)((sg)->page_link & ~SG_PAGE_LINK_MASK));
125 }
126 
127 static inline void
128 sg_set_buf(struct scatterlist *sg, const void *buf, unsigned int buflen)
129 {
130 	sg_set_page(sg, virt_to_page(buf), buflen,
131 	    ((uintptr_t)buf) & (PAGE_SIZE - 1));
132 }
133 
134 static inline struct scatterlist *
135 sg_next(struct scatterlist *sg)
136 {
137 	if (sg_is_last(sg))
138 		return (NULL);
139 	sg++;
140 	if (sg_is_chain(sg))
141 		sg = sg_chain_ptr(sg);
142 	return (sg);
143 }
144 
145 static inline vm_paddr_t
146 sg_phys(struct scatterlist *sg)
147 {
148 	return (VM_PAGE_TO_PHYS(sg_page(sg)) + sg->offset);
149 }
150 
151 static inline void *
152 sg_virt(struct scatterlist *sg)
153 {
154 
155 	return ((void *)((unsigned long)page_address(sg_page(sg)) + sg->offset));
156 }
157 
158 static inline void
159 sg_chain(struct scatterlist *prv, unsigned int prv_nents,
160     struct scatterlist *sgl)
161 {
162 	struct scatterlist *sg = &prv[prv_nents - 1];
163 
164 	sg->offset = 0;
165 	sg->length = 0;
166 	sg->page_link = ((unsigned long)sgl |
167 	    SG_PAGE_LINK_CHAIN) & ~SG_PAGE_LINK_LAST;
168 }
169 
170 static inline void
171 sg_mark_end(struct scatterlist *sg)
172 {
173 	sg->page_link |= SG_PAGE_LINK_LAST;
174 	sg->page_link &= ~SG_PAGE_LINK_CHAIN;
175 }
176 
177 static inline void
178 sg_init_table(struct scatterlist *sg, unsigned int nents)
179 {
180 	bzero(sg, sizeof(*sg) * nents);
181 	sg_mark_end(&sg[nents - 1]);
182 }
183 
184 static struct scatterlist *
185 sg_kmalloc(unsigned int nents, gfp_t gfp_mask)
186 {
187 	if (nents == SG_MAX_SINGLE_ALLOC) {
188 		return ((void *)__get_free_page(gfp_mask));
189 	} else
190 		return (kmalloc(nents * sizeof(struct scatterlist), gfp_mask));
191 }
192 
193 static inline void
194 sg_kfree(struct scatterlist *sg, unsigned int nents)
195 {
196 	if (nents == SG_MAX_SINGLE_ALLOC) {
197 		free_page((unsigned long)sg);
198 	} else
199 		kfree(sg);
200 }
201 
202 static inline void
203 __sg_free_table(struct sg_table *table, unsigned int max_ents,
204     bool skip_first_chunk, sg_free_fn * free_fn)
205 {
206 	struct scatterlist *sgl, *next;
207 
208 	if (unlikely(!table->sgl))
209 		return;
210 
211 	sgl = table->sgl;
212 	while (table->orig_nents) {
213 		unsigned int alloc_size = table->orig_nents;
214 		unsigned int sg_size;
215 
216 		if (alloc_size > max_ents) {
217 			next = sg_chain_ptr(&sgl[max_ents - 1]);
218 			alloc_size = max_ents;
219 			sg_size = alloc_size - 1;
220 		} else {
221 			sg_size = alloc_size;
222 			next = NULL;
223 		}
224 
225 		table->orig_nents -= sg_size;
226 		if (skip_first_chunk)
227 			skip_first_chunk = 0;
228 		else
229 			free_fn(sgl, alloc_size);
230 		sgl = next;
231 	}
232 
233 	table->sgl = NULL;
234 }
235 
236 static inline void
237 sg_free_table(struct sg_table *table)
238 {
239 	__sg_free_table(table, SG_MAX_SINGLE_ALLOC, 0, sg_kfree);
240 }
241 
242 static inline int
243 __sg_alloc_table(struct sg_table *table, unsigned int nents,
244     unsigned int max_ents, struct scatterlist *first_chunk,
245     gfp_t gfp_mask, sg_alloc_fn *alloc_fn)
246 {
247 	struct scatterlist *sg, *prv;
248 	unsigned int left;
249 
250 	memset(table, 0, sizeof(*table));
251 
252 	if (nents == 0)
253 		return (-EINVAL);
254 	left = nents;
255 	prv = NULL;
256 	do {
257 		unsigned int sg_size;
258 		unsigned int alloc_size = left;
259 
260 		if (alloc_size > max_ents) {
261 			alloc_size = max_ents;
262 			sg_size = alloc_size - 1;
263 		} else
264 			sg_size = alloc_size;
265 
266 		left -= sg_size;
267 
268 		if (first_chunk) {
269 			sg = first_chunk;
270 			first_chunk = NULL;
271 		} else {
272 			sg = alloc_fn(alloc_size, gfp_mask);
273 		}
274 		if (unlikely(!sg)) {
275 			if (prv)
276 				table->nents = ++table->orig_nents;
277 
278 			return (-ENOMEM);
279 		}
280 		sg_init_table(sg, alloc_size);
281 		table->nents = table->orig_nents += sg_size;
282 
283 		if (prv)
284 			sg_chain(prv, max_ents, sg);
285 		else
286 			table->sgl = sg;
287 
288 		if (!left)
289 			sg_mark_end(&sg[sg_size - 1]);
290 
291 		prv = sg;
292 	} while (left);
293 
294 	return (0);
295 }
296 
297 static inline int
298 sg_alloc_table(struct sg_table *table, unsigned int nents, gfp_t gfp_mask)
299 {
300 	int ret;
301 
302 	ret = __sg_alloc_table(table, nents, SG_MAX_SINGLE_ALLOC,
303 	    NULL, gfp_mask, sg_kmalloc);
304 	if (unlikely(ret))
305 		__sg_free_table(table, SG_MAX_SINGLE_ALLOC, 0, sg_kfree);
306 
307 	return (ret);
308 }
309 
310 static inline int
311 __sg_alloc_table_from_pages(struct sg_table *sgt,
312     struct page **pages, unsigned int count,
313     unsigned long off, unsigned long size,
314     unsigned int max_segment, gfp_t gfp_mask)
315 {
316 	unsigned int i, segs, cur, len;
317 	int rc;
318 	struct scatterlist *s;
319 
320 	if (__predict_false(!max_segment || offset_in_page(max_segment)))
321 		return (-EINVAL);
322 
323 	len = 0;
324 	for (segs = i = 1; i < count; ++i) {
325 		len += PAGE_SIZE;
326 		if (len >= max_segment ||
327 		    page_to_pfn(pages[i]) != page_to_pfn(pages[i - 1]) + 1) {
328 			++segs;
329 			len = 0;
330 		}
331 	}
332 	if (__predict_false((rc = sg_alloc_table(sgt, segs, gfp_mask))))
333 		return (rc);
334 
335 	cur = 0;
336 	for_each_sg(sgt->sgl, s, sgt->orig_nents, i) {
337 		unsigned long seg_size;
338 		unsigned int j;
339 
340 		len = 0;
341 		for (j = cur + 1; j < count; ++j) {
342 			len += PAGE_SIZE;
343 			if (len >= max_segment || page_to_pfn(pages[j]) !=
344 			    page_to_pfn(pages[j - 1]) + 1)
345 				break;
346 		}
347 
348 		seg_size = ((j - cur) << PAGE_SHIFT) - off;
349 		sg_set_page(s, pages[cur], MIN(size, seg_size), off);
350 		size -= seg_size;
351 		off = 0;
352 		cur = j;
353 	}
354 	return (0);
355 }
356 
357 static inline int
358 sg_alloc_table_from_pages(struct sg_table *sgt,
359     struct page **pages, unsigned int count,
360     unsigned long off, unsigned long size,
361     gfp_t gfp_mask)
362 {
363 
364 	return (__sg_alloc_table_from_pages(sgt, pages, count, off, size,
365 	    SCATTERLIST_MAX_SEGMENT, gfp_mask));
366 }
367 
368 static inline int
369 sg_nents(struct scatterlist *sg)
370 {
371 	int nents;
372 
373 	for (nents = 0; sg; sg = sg_next(sg))
374 		nents++;
375 	return (nents);
376 }
377 
378 static inline void
379 __sg_page_iter_start(struct sg_page_iter *piter,
380     struct scatterlist *sglist, unsigned int nents,
381     unsigned long pgoffset)
382 {
383 	piter->internal.pg_advance = 0;
384 	piter->internal.nents = nents;
385 
386 	piter->sg = sglist;
387 	piter->sg_pgoffset = pgoffset;
388 }
389 
390 static inline void
391 _sg_iter_next(struct sg_page_iter *iter)
392 {
393 	struct scatterlist *sg;
394 	unsigned int pgcount;
395 
396 	sg = iter->sg;
397 	pgcount = (sg->offset + sg->length + PAGE_SIZE - 1) >> PAGE_SHIFT;
398 
399 	++iter->sg_pgoffset;
400 	while (iter->sg_pgoffset >= pgcount) {
401 		iter->sg_pgoffset -= pgcount;
402 		sg = sg_next(sg);
403 		--iter->maxents;
404 		if (sg == NULL || iter->maxents == 0)
405 			break;
406 		pgcount = (sg->offset + sg->length + PAGE_SIZE - 1) >> PAGE_SHIFT;
407 	}
408 	iter->sg = sg;
409 }
410 
411 static inline int
412 sg_page_count(struct scatterlist *sg)
413 {
414 	return (PAGE_ALIGN(sg->offset + sg->length) >> PAGE_SHIFT);
415 }
416 #define	sg_dma_page_count(sg) \
417 	sg_page_count(sg)
418 
419 static inline bool
420 __sg_page_iter_next(struct sg_page_iter *piter)
421 {
422 	unsigned int pgcount;
423 
424 	if (piter->internal.nents == 0)
425 		return (0);
426 	if (piter->sg == NULL)
427 		return (0);
428 
429 	piter->sg_pgoffset += piter->internal.pg_advance;
430 	piter->internal.pg_advance = 1;
431 
432 	while (1) {
433 		pgcount = sg_page_count(piter->sg);
434 		if (likely(piter->sg_pgoffset < pgcount))
435 			break;
436 		piter->sg_pgoffset -= pgcount;
437 		piter->sg = sg_next(piter->sg);
438 		if (--piter->internal.nents == 0)
439 			return (0);
440 		if (piter->sg == NULL)
441 			return (0);
442 	}
443 	return (1);
444 }
445 #define	__sg_page_iter_dma_next(itr) \
446 	__sg_page_iter_next(&(itr)->base)
447 
448 static inline void
449 _sg_iter_init(struct scatterlist *sgl, struct sg_page_iter *iter,
450     unsigned int nents, unsigned long pgoffset)
451 {
452 	if (nents) {
453 		iter->sg = sgl;
454 		iter->sg_pgoffset = pgoffset - 1;
455 		iter->maxents = nents;
456 		_sg_iter_next(iter);
457 	} else {
458 		iter->sg = NULL;
459 		iter->sg_pgoffset = 0;
460 		iter->maxents = 0;
461 	}
462 }
463 
464 /*
465  * sg_page_iter_dma_address() is implemented as a macro because it
466  * needs to accept two different and identical structure types. This
467  * allows both old and new code to co-exist. The compile time assert
468  * adds some safety, that the structure sizes match.
469  */
470 #define	sg_page_iter_dma_address(spi) ({		\
471 	struct sg_page_iter *__spi = (void *)(spi);	\
472 	dma_addr_t __dma_address;			\
473 	CTASSERT(sizeof(*(spi)) == sizeof(*__spi));	\
474 	__dma_address = __spi->sg->dma_address +	\
475 	    (__spi->sg_pgoffset << PAGE_SHIFT);		\
476 	__dma_address;					\
477 })
478 
479 static inline struct page *
480 sg_page_iter_page(struct sg_page_iter *piter)
481 {
482 	return (nth_page(sg_page(piter->sg), piter->sg_pgoffset));
483 }
484 
485 static __inline size_t
486 sg_pcopy_from_buffer(struct scatterlist *sgl, unsigned int nents,
487     const void *buf, size_t buflen, off_t skip)
488 {
489 	struct sg_page_iter piter;
490 	struct page *page;
491 	struct sf_buf *sf;
492 	size_t len, copied;
493 	char *p, *b;
494 
495 	if (buflen == 0)
496 		return (0);
497 
498 	b = __DECONST(char *, buf);
499 	copied = 0;
500 	sched_pin();
501 	for_each_sg_page(sgl, &piter, nents, 0) {
502 
503 		/* Skip to the start. */
504 		if (piter.sg->length <= skip) {
505 			skip -= piter.sg->length;
506 			continue;
507 		}
508 
509 		/* See how much to copy. */
510 		KASSERT(((piter.sg->length - skip) != 0 && (buflen != 0)),
511 		    ("%s: sg len %u - skip %ju || buflen %zu is 0\n",
512 		    __func__, piter.sg->length, (uintmax_t)skip, buflen));
513 		len = min(piter.sg->length - skip, buflen);
514 
515 		page = sg_page_iter_page(&piter);
516 		sf = sf_buf_alloc(page, SFB_CPUPRIVATE | SFB_NOWAIT);
517 		if (sf == NULL)
518 			break;
519 		p = (char *)sf_buf_kva(sf) + piter.sg_pgoffset + skip;
520 		memcpy(p, b, len);
521 		sf_buf_free(sf);
522 
523 		/* We copied so nothing more to skip. */
524 		skip = 0;
525 		copied += len;
526 		/* Either we exactly filled the page, or we are done. */
527 		buflen -= len;
528 		if (buflen == 0)
529 			break;
530 		b += len;
531 	}
532 	sched_unpin();
533 
534 	return (copied);
535 }
536 
537 #endif					/* _LINUX_SCATTERLIST_H_ */
538