include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit...
[GitHub/mt8127/android_kernel_alcatel_ttab.git] / drivers / infiniband / hw / amso1100 / c2_mm.c
CommitLineData
f94b533d
TT
1/*
2 * Copyright (c) 2005 Ammasso, Inc. All rights reserved.
3 * Copyright (c) 2005 Open Grid Computing, Inc. All rights reserved.
4 *
5 * This software is available to you under a choice of one of two
6 * licenses. You may choose to be licensed under the terms of the GNU
7 * General Public License (GPL) Version 2, available from the file
8 * COPYING in the main directory of this source tree, or the
9 * OpenIB.org BSD license below:
10 *
11 * Redistribution and use in source and binary forms, with or
12 * without modification, are permitted provided that the following
13 * conditions are met:
14 *
15 * - Redistributions of source code must retain the above
16 * copyright notice, this list of conditions and the following
17 * disclaimer.
18 *
19 * - Redistributions in binary form must reproduce the above
20 * copyright notice, this list of conditions and the following
21 * disclaimer in the documentation and/or other materials
22 * provided with the distribution.
23 *
24 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
25 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
26 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
27 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
28 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
29 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
30 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
31 * SOFTWARE.
32 */
5a0e3ad6
TH
33#include <linux/slab.h>
34
f94b533d
TT
35#include "c2.h"
36#include "c2_vq.h"
37
38#define PBL_VIRT 1
39#define PBL_PHYS 2
40
41/*
42 * Send all the PBL messages to convey the remainder of the PBL
43 * Wait for the adapter's reply on the last one.
44 * This is indicated by setting the MEM_PBL_COMPLETE in the flags.
45 *
46 * NOTE: vq_req is _not_ freed by this function. The VQ Host
47 * Reply buffer _is_ freed by this function.
48 */
49static int
dc544bc9 50send_pbl_messages(struct c2_dev *c2dev, __be32 stag_index,
f94b533d
TT
51 unsigned long va, u32 pbl_depth,
52 struct c2_vq_req *vq_req, int pbl_type)
53{
54 u32 pbe_count; /* amt that fits in a PBL msg */
55 u32 count; /* amt in this PBL MSG. */
56 struct c2wr_nsmr_pbl_req *wr; /* PBL WR ptr */
57 struct c2wr_nsmr_pbl_rep *reply; /* reply ptr */
58 int err, pbl_virt, pbl_index, i;
59
60 switch (pbl_type) {
61 case PBL_VIRT:
62 pbl_virt = 1;
63 break;
64 case PBL_PHYS:
65 pbl_virt = 0;
66 break;
67 default:
68 return -EINVAL;
69 break;
70 }
71
72 pbe_count = (c2dev->req_vq.msg_size -
73 sizeof(struct c2wr_nsmr_pbl_req)) / sizeof(u64);
74 wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL);
75 if (!wr) {
76 return -ENOMEM;
77 }
78 c2_wr_set_id(wr, CCWR_NSMR_PBL);
79
80 /*
81 * Only the last PBL message will generate a reply from the verbs,
82 * so we set the context to 0 indicating there is no kernel verbs
83 * handler blocked awaiting this reply.
84 */
85 wr->hdr.context = 0;
86 wr->rnic_handle = c2dev->adapter_handle;
87 wr->stag_index = stag_index; /* already swapped */
88 wr->flags = 0;
89 pbl_index = 0;
90 while (pbl_depth) {
91 count = min(pbe_count, pbl_depth);
92 wr->addrs_length = cpu_to_be32(count);
93
94 /*
95 * If this is the last message, then reference the
96 * vq request struct cuz we're gonna wait for a reply.
97 * also make this PBL msg as the last one.
98 */
99 if (count == pbl_depth) {
100 /*
101 * reference the request struct. dereferenced in the
102 * int handler.
103 */
104 vq_req_get(c2dev, vq_req);
105 wr->flags = cpu_to_be32(MEM_PBL_COMPLETE);
106
107 /*
108 * This is the last PBL message.
109 * Set the context to our VQ Request Object so we can
110 * wait for the reply.
111 */
112 wr->hdr.context = (unsigned long) vq_req;
113 }
114
115 /*
116 * If pbl_virt is set then va is a virtual address
117 * that describes a virtually contiguous memory
118 * allocation. The wr needs the start of each virtual page
119 * to be converted to the corresponding physical address
120 * of the page. If pbl_virt is not set then va is an array
121 * of physical addresses and there is no conversion to do.
122 * Just fill in the wr with what is in the array.
123 */
124 for (i = 0; i < count; i++) {
125 if (pbl_virt) {
126 va += PAGE_SIZE;
127 } else {
128 wr->paddrs[i] =
129 cpu_to_be64(((u64 *)va)[pbl_index + i]);
130 }
131 }
132
133 /*
134 * Send WR to adapter
135 */
136 err = vq_send_wr(c2dev, (union c2wr *) wr);
137 if (err) {
138 if (count <= pbe_count) {
139 vq_req_put(c2dev, vq_req);
140 }
141 goto bail0;
142 }
143 pbl_depth -= count;
144 pbl_index += count;
145 }
146
147 /*
148 * Now wait for the reply...
149 */
150 err = vq_wait_for_reply(c2dev, vq_req);
151 if (err) {
152 goto bail0;
153 }
154
155 /*
156 * Process reply
157 */
158 reply = (struct c2wr_nsmr_pbl_rep *) (unsigned long) vq_req->reply_msg;
159 if (!reply) {
160 err = -ENOMEM;
161 goto bail0;
162 }
163
164 err = c2_errno(reply);
165
166 vq_repbuf_free(c2dev, reply);
167 bail0:
168 kfree(wr);
169 return err;
170}
171
172#define C2_PBL_MAX_DEPTH 131072
173int
174c2_nsmr_register_phys_kern(struct c2_dev *c2dev, u64 *addr_list,
175 int page_size, int pbl_depth, u32 length,
176 u32 offset, u64 *va, enum c2_acf acf,
177 struct c2_mr *mr)
178{
179 struct c2_vq_req *vq_req;
180 struct c2wr_nsmr_register_req *wr;
181 struct c2wr_nsmr_register_rep *reply;
182 u16 flags;
183 int i, pbe_count, count;
184 int err;
185
186 if (!va || !length || !addr_list || !pbl_depth)
187 return -EINTR;
188
189 /*
190 * Verify PBL depth is within rnic max
191 */
192 if (pbl_depth > C2_PBL_MAX_DEPTH) {
193 return -EINTR;
194 }
195
196 /*
197 * allocate verbs request object
198 */
199 vq_req = vq_req_alloc(c2dev);
200 if (!vq_req)
201 return -ENOMEM;
202
203 wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL);
204 if (!wr) {
205 err = -ENOMEM;
206 goto bail0;
207 }
208
209 /*
210 * build the WR
211 */
212 c2_wr_set_id(wr, CCWR_NSMR_REGISTER);
213 wr->hdr.context = (unsigned long) vq_req;
214 wr->rnic_handle = c2dev->adapter_handle;
215
216 flags = (acf | MEM_VA_BASED | MEM_REMOTE);
217
218 /*
219 * compute how many pbes can fit in the message
220 */
221 pbe_count = (c2dev->req_vq.msg_size -
222 sizeof(struct c2wr_nsmr_register_req)) / sizeof(u64);
223
224 if (pbl_depth <= pbe_count) {
225 flags |= MEM_PBL_COMPLETE;
226 }
227 wr->flags = cpu_to_be16(flags);
228 wr->stag_key = 0; //stag_key;
229 wr->va = cpu_to_be64(*va);
230 wr->pd_id = mr->pd->pd_id;
231 wr->pbe_size = cpu_to_be32(page_size);
232 wr->length = cpu_to_be32(length);
233 wr->pbl_depth = cpu_to_be32(pbl_depth);
234 wr->fbo = cpu_to_be32(offset);
235 count = min(pbl_depth, pbe_count);
236 wr->addrs_length = cpu_to_be32(count);
237
238 /*
239 * fill out the PBL for this message
240 */
241 for (i = 0; i < count; i++) {
242 wr->paddrs[i] = cpu_to_be64(addr_list[i]);
243 }
244
245 /*
246 * regerence the request struct
247 */
248 vq_req_get(c2dev, vq_req);
249
250 /*
251 * send the WR to the adapter
252 */
253 err = vq_send_wr(c2dev, (union c2wr *) wr);
254 if (err) {
255 vq_req_put(c2dev, vq_req);
256 goto bail1;
257 }
258
259 /*
260 * wait for reply from adapter
261 */
262 err = vq_wait_for_reply(c2dev, vq_req);
263 if (err) {
264 goto bail1;
265 }
266
267 /*
268 * process reply
269 */
270 reply =
271 (struct c2wr_nsmr_register_rep *) (unsigned long) (vq_req->reply_msg);
272 if (!reply) {
273 err = -ENOMEM;
274 goto bail1;
275 }
276 if ((err = c2_errno(reply))) {
277 goto bail2;
278 }
279 //*p_pb_entries = be32_to_cpu(reply->pbl_depth);
280 mr->ibmr.lkey = mr->ibmr.rkey = be32_to_cpu(reply->stag_index);
281 vq_repbuf_free(c2dev, reply);
282
283 /*
284 * if there are still more PBEs we need to send them to
285 * the adapter and wait for a reply on the final one.
286 * reuse vq_req for this purpose.
287 */
288 pbl_depth -= count;
289 if (pbl_depth) {
290
291 vq_req->reply_msg = (unsigned long) NULL;
292 atomic_set(&vq_req->reply_ready, 0);
293 err = send_pbl_messages(c2dev,
294 cpu_to_be32(mr->ibmr.lkey),
295 (unsigned long) &addr_list[i],
296 pbl_depth, vq_req, PBL_PHYS);
297 if (err) {
298 goto bail1;
299 }
300 }
301
302 vq_req_free(c2dev, vq_req);
303 kfree(wr);
304
305 return err;
306
307 bail2:
308 vq_repbuf_free(c2dev, reply);
309 bail1:
310 kfree(wr);
311 bail0:
312 vq_req_free(c2dev, vq_req);
313 return err;
314}
315
316int c2_stag_dealloc(struct c2_dev *c2dev, u32 stag_index)
317{
318 struct c2_vq_req *vq_req; /* verbs request object */
319 struct c2wr_stag_dealloc_req wr; /* work request */
320 struct c2wr_stag_dealloc_rep *reply; /* WR reply */
321 int err;
322
323
324 /*
325 * allocate verbs request object
326 */
327 vq_req = vq_req_alloc(c2dev);
328 if (!vq_req) {
329 return -ENOMEM;
330 }
331
332 /*
333 * Build the WR
334 */
335 c2_wr_set_id(&wr, CCWR_STAG_DEALLOC);
336 wr.hdr.context = (u64) (unsigned long) vq_req;
337 wr.rnic_handle = c2dev->adapter_handle;
338 wr.stag_index = cpu_to_be32(stag_index);
339
340 /*
341 * reference the request struct. dereferenced in the int handler.
342 */
343 vq_req_get(c2dev, vq_req);
344
345 /*
346 * Send WR to adapter
347 */
348 err = vq_send_wr(c2dev, (union c2wr *) & wr);
349 if (err) {
350 vq_req_put(c2dev, vq_req);
351 goto bail0;
352 }
353
354 /*
355 * Wait for reply from adapter
356 */
357 err = vq_wait_for_reply(c2dev, vq_req);
358 if (err) {
359 goto bail0;
360 }
361
362 /*
363 * Process reply
364 */
365 reply = (struct c2wr_stag_dealloc_rep *) (unsigned long) vq_req->reply_msg;
366 if (!reply) {
367 err = -ENOMEM;
368 goto bail0;
369 }
370
371 err = c2_errno(reply);
372
373 vq_repbuf_free(c2dev, reply);
374 bail0:
375 vq_req_free(c2dev, vq_req);
376 return err;
377}