xref: /linux/drivers/net/ethernet/mellanox/mlx5/core/wq.c (revision f86fd32d)
1 /*
2  * Copyright (c) 2013-2015, Mellanox Technologies, Ltd.  All rights reserved.
3  *
4  * This software is available to you under a choice of one of two
5  * licenses.  You may choose to be licensed under the terms of the GNU
6  * General Public License (GPL) Version 2, available from the file
7  * COPYING in the main directory of this source tree, or the
8  * OpenIB.org BSD license below:
9  *
10  *     Redistribution and use in source and binary forms, with or
11  *     without modification, are permitted provided that the following
12  *     conditions are met:
13  *
14  *      - Redistributions of source code must retain the above
15  *        copyright notice, this list of conditions and the following
16  *        disclaimer.
17  *
18  *      - Redistributions in binary form must reproduce the above
19  *        copyright notice, this list of conditions and the following
20  *        disclaimer in the documentation and/or other materials
21  *        provided with the distribution.
22  *
23  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
24  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
25  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
26  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
27  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
28  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
29  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
30  * SOFTWARE.
31  */
32 
33 #include <linux/mlx5/driver.h>
34 #include "wq.h"
35 #include "mlx5_core.h"
36 
37 static u32 wq_get_byte_sz(u8 log_sz, u8 log_stride)
38 {
39 	return ((u32)1 << log_sz) << log_stride;
40 }
41 
42 int mlx5_wq_cyc_create(struct mlx5_core_dev *mdev, struct mlx5_wq_param *param,
43 		       void *wqc, struct mlx5_wq_cyc *wq,
44 		       struct mlx5_wq_ctrl *wq_ctrl)
45 {
46 	u8 log_wq_stride = MLX5_GET(wq, wqc, log_wq_stride);
47 	u8 log_wq_sz     = MLX5_GET(wq, wqc, log_wq_sz);
48 	struct mlx5_frag_buf_ctrl *fbc = &wq->fbc;
49 	int err;
50 
51 	err = mlx5_db_alloc_node(mdev, &wq_ctrl->db, param->db_numa_node);
52 	if (err) {
53 		mlx5_core_warn(mdev, "mlx5_db_alloc_node() failed, %d\n", err);
54 		return err;
55 	}
56 
57 	wq->db  = wq_ctrl->db.db;
58 
59 	err = mlx5_frag_buf_alloc_node(mdev, wq_get_byte_sz(log_wq_sz, log_wq_stride),
60 				       &wq_ctrl->buf, param->buf_numa_node);
61 	if (err) {
62 		mlx5_core_warn(mdev, "mlx5_frag_buf_alloc_node() failed, %d\n", err);
63 		goto err_db_free;
64 	}
65 
66 	mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, fbc);
67 	wq->sz = mlx5_wq_cyc_get_size(wq);
68 
69 	wq_ctrl->mdev = mdev;
70 
71 	return 0;
72 
73 err_db_free:
74 	mlx5_db_free(mdev, &wq_ctrl->db);
75 
76 	return err;
77 }
78 
79 void mlx5_wq_cyc_wqe_dump(struct mlx5_wq_cyc *wq, u16 ix, u8 nstrides)
80 {
81 	size_t len;
82 	void *wqe;
83 
84 	if (!net_ratelimit())
85 		return;
86 
87 	nstrides = max_t(u8, nstrides, 1);
88 
89 	len = nstrides << wq->fbc.log_stride;
90 	wqe = mlx5_wq_cyc_get_wqe(wq, ix);
91 
92 	pr_info("WQE DUMP: WQ size %d WQ cur size %d, WQE index 0x%x, len: %zu\n",
93 		mlx5_wq_cyc_get_size(wq), wq->cur_sz, ix, len);
94 	print_hex_dump(KERN_WARNING, "", DUMP_PREFIX_OFFSET, 16, 1, wqe, len, false);
95 }
96 
97 int mlx5_wq_qp_create(struct mlx5_core_dev *mdev, struct mlx5_wq_param *param,
98 		      void *qpc, struct mlx5_wq_qp *wq,
99 		      struct mlx5_wq_ctrl *wq_ctrl)
100 {
101 	u8 log_rq_stride = MLX5_GET(qpc, qpc, log_rq_stride) + 4;
102 	u8 log_rq_sz     = MLX5_GET(qpc, qpc, log_rq_size);
103 	u8 log_sq_stride = ilog2(MLX5_SEND_WQE_BB);
104 	u8 log_sq_sz     = MLX5_GET(qpc, qpc, log_sq_size);
105 
106 	u32 rq_byte_size;
107 	int err;
108 
109 
110 
111 	err = mlx5_db_alloc_node(mdev, &wq_ctrl->db, param->db_numa_node);
112 	if (err) {
113 		mlx5_core_warn(mdev, "mlx5_db_alloc_node() failed, %d\n", err);
114 		return err;
115 	}
116 
117 	err = mlx5_frag_buf_alloc_node(mdev,
118 				       wq_get_byte_sz(log_rq_sz, log_rq_stride) +
119 				       wq_get_byte_sz(log_sq_sz, log_sq_stride),
120 				       &wq_ctrl->buf, param->buf_numa_node);
121 	if (err) {
122 		mlx5_core_warn(mdev, "mlx5_frag_buf_alloc_node() failed, %d\n", err);
123 		goto err_db_free;
124 	}
125 
126 	mlx5_init_fbc(wq_ctrl->buf.frags, log_rq_stride, log_rq_sz, &wq->rq.fbc);
127 
128 	rq_byte_size = wq_get_byte_sz(log_rq_sz, log_rq_stride);
129 
130 	if (rq_byte_size < PAGE_SIZE) {
131 		/* SQ starts within the same page of the RQ */
132 		u16 sq_strides_offset = rq_byte_size / MLX5_SEND_WQE_BB;
133 
134 		mlx5_init_fbc_offset(wq_ctrl->buf.frags,
135 				     log_sq_stride, log_sq_sz, sq_strides_offset,
136 				     &wq->sq.fbc);
137 	} else {
138 		u16 rq_npages = rq_byte_size >> PAGE_SHIFT;
139 
140 		mlx5_init_fbc(wq_ctrl->buf.frags + rq_npages,
141 			      log_sq_stride, log_sq_sz, &wq->sq.fbc);
142 	}
143 
144 	wq->rq.db  = &wq_ctrl->db.db[MLX5_RCV_DBR];
145 	wq->sq.db  = &wq_ctrl->db.db[MLX5_SND_DBR];
146 
147 	wq_ctrl->mdev = mdev;
148 
149 	return 0;
150 
151 err_db_free:
152 	mlx5_db_free(mdev, &wq_ctrl->db);
153 
154 	return err;
155 }
156 
157 int mlx5_cqwq_create(struct mlx5_core_dev *mdev, struct mlx5_wq_param *param,
158 		     void *cqc, struct mlx5_cqwq *wq,
159 		     struct mlx5_wq_ctrl *wq_ctrl)
160 {
161 	/* CQE_STRIDE_128 and CQE_STRIDE_128_PAD both mean 128B stride */
162 	u8 log_wq_stride = MLX5_GET(cqc, cqc, cqe_sz) == CQE_STRIDE_64 ? 6 : 7;
163 	u8 log_wq_sz     = MLX5_GET(cqc, cqc, log_cq_size);
164 	int err;
165 
166 	err = mlx5_db_alloc_node(mdev, &wq_ctrl->db, param->db_numa_node);
167 	if (err) {
168 		mlx5_core_warn(mdev, "mlx5_db_alloc_node() failed, %d\n", err);
169 		return err;
170 	}
171 
172 	wq->db  = wq_ctrl->db.db;
173 
174 	err = mlx5_frag_buf_alloc_node(mdev, wq_get_byte_sz(log_wq_sz, log_wq_stride),
175 				       &wq_ctrl->buf,
176 				       param->buf_numa_node);
177 	if (err) {
178 		mlx5_core_warn(mdev, "mlx5_frag_buf_alloc_node() failed, %d\n",
179 			       err);
180 		goto err_db_free;
181 	}
182 
183 	mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, &wq->fbc);
184 
185 	wq_ctrl->mdev = mdev;
186 
187 	return 0;
188 
189 err_db_free:
190 	mlx5_db_free(mdev, &wq_ctrl->db);
191 
192 	return err;
193 }
194 
195 int mlx5_wq_ll_create(struct mlx5_core_dev *mdev, struct mlx5_wq_param *param,
196 		      void *wqc, struct mlx5_wq_ll *wq,
197 		      struct mlx5_wq_ctrl *wq_ctrl)
198 {
199 	u8 log_wq_stride = MLX5_GET(wq, wqc, log_wq_stride);
200 	u8 log_wq_sz     = MLX5_GET(wq, wqc, log_wq_sz);
201 	struct mlx5_frag_buf_ctrl *fbc = &wq->fbc;
202 	struct mlx5_wqe_srq_next_seg *next_seg;
203 	int err;
204 	int i;
205 
206 	err = mlx5_db_alloc_node(mdev, &wq_ctrl->db, param->db_numa_node);
207 	if (err) {
208 		mlx5_core_warn(mdev, "mlx5_db_alloc_node() failed, %d\n", err);
209 		return err;
210 	}
211 
212 	wq->db  = wq_ctrl->db.db;
213 
214 	err = mlx5_frag_buf_alloc_node(mdev, wq_get_byte_sz(log_wq_sz, log_wq_stride),
215 				       &wq_ctrl->buf, param->buf_numa_node);
216 	if (err) {
217 		mlx5_core_warn(mdev, "mlx5_frag_buf_alloc_node() failed, %d\n", err);
218 		goto err_db_free;
219 	}
220 
221 	mlx5_init_fbc(wq_ctrl->buf.frags, log_wq_stride, log_wq_sz, fbc);
222 
223 	for (i = 0; i < fbc->sz_m1; i++) {
224 		next_seg = mlx5_wq_ll_get_wqe(wq, i);
225 		next_seg->next_wqe_index = cpu_to_be16(i + 1);
226 	}
227 	next_seg = mlx5_wq_ll_get_wqe(wq, i);
228 	wq->tail_next = &next_seg->next_wqe_index;
229 
230 	wq_ctrl->mdev = mdev;
231 
232 	return 0;
233 
234 err_db_free:
235 	mlx5_db_free(mdev, &wq_ctrl->db);
236 
237 	return err;
238 }
239 
240 void mlx5_wq_destroy(struct mlx5_wq_ctrl *wq_ctrl)
241 {
242 	mlx5_frag_buf_free(wq_ctrl->mdev, &wq_ctrl->buf);
243 	mlx5_db_free(wq_ctrl->mdev, &wq_ctrl->db);
244 }
245 
246