Commit | Line | Data |
---|---|---|
bc38a6ab RD |
1 | /* |
2 | * Copyright (c) 2005 Topspin Communications. All rights reserved. | |
3 | * Copyright (c) 2005 Cisco Systems. All rights reserved. | |
2a1d9b7f RD |
4 | * Copyright (c) 2005 Mellanox Technologies. All rights reserved. |
5 | * Copyright (c) 2005 Voltaire, Inc. All rights reserved. | |
67cdb40c | 6 | * Copyright (c) 2005 PathScale, Inc. All rights reserved. |
bc38a6ab RD |
7 | * |
8 | * This software is available to you under a choice of one of two | |
9 | * licenses. You may choose to be licensed under the terms of the GNU | |
10 | * General Public License (GPL) Version 2, available from the file | |
11 | * COPYING in the main directory of this source tree, or the | |
12 | * OpenIB.org BSD license below: | |
13 | * | |
14 | * Redistribution and use in source and binary forms, with or | |
15 | * without modification, are permitted provided that the following | |
16 | * conditions are met: | |
17 | * | |
18 | * - Redistributions of source code must retain the above | |
19 | * copyright notice, this list of conditions and the following | |
20 | * disclaimer. | |
21 | * | |
22 | * - Redistributions in binary form must reproduce the above | |
23 | * copyright notice, this list of conditions and the following | |
24 | * disclaimer in the documentation and/or other materials | |
25 | * provided with the distribution. | |
26 | * | |
27 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, | |
28 | * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF | |
29 | * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND | |
30 | * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS | |
31 | * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN | |
32 | * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN | |
33 | * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | |
34 | * SOFTWARE. | |
35 | * | |
36 | * $Id: uverbs_main.c 2733 2005-06-28 19:14:34Z roland $ | |
37 | */ | |
38 | ||
39 | #include <linux/module.h> | |
40 | #include <linux/init.h> | |
41 | #include <linux/device.h> | |
42 | #include <linux/err.h> | |
43 | #include <linux/fs.h> | |
44 | #include <linux/poll.h> | |
45 | #include <linux/file.h> | |
46 | #include <linux/mount.h> | |
47 | ||
48 | #include <asm/uaccess.h> | |
49 | ||
50 | #include "uverbs.h" | |
51 | ||
52 | MODULE_AUTHOR("Roland Dreier"); | |
53 | MODULE_DESCRIPTION("InfiniBand userspace verbs access"); | |
54 | MODULE_LICENSE("Dual BSD/GPL"); | |
55 | ||
56 | #define INFINIBANDEVENTFS_MAGIC 0x49426576 /* "IBev" */ | |
57 | ||
58 | enum { | |
59 | IB_UVERBS_MAJOR = 231, | |
60 | IB_UVERBS_BASE_MINOR = 192, | |
61 | IB_UVERBS_MAX_DEVICES = 32 | |
62 | }; | |
63 | ||
64 | #define IB_UVERBS_BASE_DEV MKDEV(IB_UVERBS_MAJOR, IB_UVERBS_BASE_MINOR) | |
65 | ||
66 | DECLARE_MUTEX(ib_uverbs_idr_mutex); | |
67 | DEFINE_IDR(ib_uverbs_pd_idr); | |
68 | DEFINE_IDR(ib_uverbs_mr_idr); | |
69 | DEFINE_IDR(ib_uverbs_mw_idr); | |
70 | DEFINE_IDR(ib_uverbs_ah_idr); | |
71 | DEFINE_IDR(ib_uverbs_cq_idr); | |
72 | DEFINE_IDR(ib_uverbs_qp_idr); | |
f520ba5a | 73 | DEFINE_IDR(ib_uverbs_srq_idr); |
bc38a6ab RD |
74 | |
75 | static spinlock_t map_lock; | |
76 | static DECLARE_BITMAP(dev_map, IB_UVERBS_MAX_DEVICES); | |
77 | ||
78 | static ssize_t (*uverbs_cmd_table[])(struct ib_uverbs_file *file, | |
79 | const char __user *buf, int in_len, | |
80 | int out_len) = { | |
6b73597e RD |
81 | [IB_USER_VERBS_CMD_GET_CONTEXT] = ib_uverbs_get_context, |
82 | [IB_USER_VERBS_CMD_QUERY_DEVICE] = ib_uverbs_query_device, | |
83 | [IB_USER_VERBS_CMD_QUERY_PORT] = ib_uverbs_query_port, | |
84 | [IB_USER_VERBS_CMD_ALLOC_PD] = ib_uverbs_alloc_pd, | |
85 | [IB_USER_VERBS_CMD_DEALLOC_PD] = ib_uverbs_dealloc_pd, | |
86 | [IB_USER_VERBS_CMD_REG_MR] = ib_uverbs_reg_mr, | |
87 | [IB_USER_VERBS_CMD_DEREG_MR] = ib_uverbs_dereg_mr, | |
88 | [IB_USER_VERBS_CMD_CREATE_COMP_CHANNEL] = ib_uverbs_create_comp_channel, | |
89 | [IB_USER_VERBS_CMD_CREATE_CQ] = ib_uverbs_create_cq, | |
67cdb40c RD |
90 | [IB_USER_VERBS_CMD_POLL_CQ] = ib_uverbs_poll_cq, |
91 | [IB_USER_VERBS_CMD_REQ_NOTIFY_CQ] = ib_uverbs_req_notify_cq, | |
6b73597e RD |
92 | [IB_USER_VERBS_CMD_DESTROY_CQ] = ib_uverbs_destroy_cq, |
93 | [IB_USER_VERBS_CMD_CREATE_QP] = ib_uverbs_create_qp, | |
94 | [IB_USER_VERBS_CMD_MODIFY_QP] = ib_uverbs_modify_qp, | |
95 | [IB_USER_VERBS_CMD_DESTROY_QP] = ib_uverbs_destroy_qp, | |
67cdb40c RD |
96 | [IB_USER_VERBS_CMD_POST_SEND] = ib_uverbs_post_send, |
97 | [IB_USER_VERBS_CMD_POST_RECV] = ib_uverbs_post_recv, | |
98 | [IB_USER_VERBS_CMD_POST_SRQ_RECV] = ib_uverbs_post_srq_recv, | |
99 | [IB_USER_VERBS_CMD_CREATE_AH] = ib_uverbs_create_ah, | |
100 | [IB_USER_VERBS_CMD_DESTROY_AH] = ib_uverbs_destroy_ah, | |
6b73597e RD |
101 | [IB_USER_VERBS_CMD_ATTACH_MCAST] = ib_uverbs_attach_mcast, |
102 | [IB_USER_VERBS_CMD_DETACH_MCAST] = ib_uverbs_detach_mcast, | |
103 | [IB_USER_VERBS_CMD_CREATE_SRQ] = ib_uverbs_create_srq, | |
104 | [IB_USER_VERBS_CMD_MODIFY_SRQ] = ib_uverbs_modify_srq, | |
105 | [IB_USER_VERBS_CMD_DESTROY_SRQ] = ib_uverbs_destroy_srq, | |
bc38a6ab RD |
106 | }; |
107 | ||
108 | static struct vfsmount *uverbs_event_mnt; | |
109 | ||
110 | static void ib_uverbs_add_one(struct ib_device *device); | |
111 | static void ib_uverbs_remove_one(struct ib_device *device); | |
112 | ||
113 | static int ib_dealloc_ucontext(struct ib_ucontext *context) | |
114 | { | |
115 | struct ib_uobject *uobj, *tmp; | |
116 | ||
117 | if (!context) | |
118 | return 0; | |
119 | ||
120 | down(&ib_uverbs_idr_mutex); | |
121 | ||
67cdb40c RD |
122 | list_for_each_entry_safe(uobj, tmp, &context->ah_list, list) { |
123 | struct ib_ah *ah = idr_find(&ib_uverbs_ah_idr, uobj->id); | |
124 | idr_remove(&ib_uverbs_ah_idr, uobj->id); | |
125 | ib_destroy_ah(ah); | |
126 | list_del(&uobj->list); | |
127 | kfree(uobj); | |
128 | } | |
bc38a6ab RD |
129 | |
130 | list_for_each_entry_safe(uobj, tmp, &context->qp_list, list) { | |
131 | struct ib_qp *qp = idr_find(&ib_uverbs_qp_idr, uobj->id); | |
132 | idr_remove(&ib_uverbs_qp_idr, uobj->id); | |
133 | ib_destroy_qp(qp); | |
134 | list_del(&uobj->list); | |
63aaf647 | 135 | kfree(container_of(uobj, struct ib_uevent_object, uobject)); |
bc38a6ab RD |
136 | } |
137 | ||
138 | list_for_each_entry_safe(uobj, tmp, &context->cq_list, list) { | |
139 | struct ib_cq *cq = idr_find(&ib_uverbs_cq_idr, uobj->id); | |
140 | idr_remove(&ib_uverbs_cq_idr, uobj->id); | |
141 | ib_destroy_cq(cq); | |
142 | list_del(&uobj->list); | |
63aaf647 | 143 | kfree(container_of(uobj, struct ib_ucq_object, uobject)); |
bc38a6ab RD |
144 | } |
145 | ||
f520ba5a RD |
146 | list_for_each_entry_safe(uobj, tmp, &context->srq_list, list) { |
147 | struct ib_srq *srq = idr_find(&ib_uverbs_srq_idr, uobj->id); | |
148 | idr_remove(&ib_uverbs_srq_idr, uobj->id); | |
149 | ib_destroy_srq(srq); | |
150 | list_del(&uobj->list); | |
63aaf647 | 151 | kfree(container_of(uobj, struct ib_uevent_object, uobject)); |
f520ba5a RD |
152 | } |
153 | ||
bc38a6ab RD |
154 | /* XXX Free MWs */ |
155 | ||
156 | list_for_each_entry_safe(uobj, tmp, &context->mr_list, list) { | |
157 | struct ib_mr *mr = idr_find(&ib_uverbs_mr_idr, uobj->id); | |
e1bcfcaa | 158 | struct ib_device *mrdev = mr->device; |
bc38a6ab RD |
159 | struct ib_umem_object *memobj; |
160 | ||
161 | idr_remove(&ib_uverbs_mr_idr, uobj->id); | |
162 | ib_dereg_mr(mr); | |
163 | ||
164 | memobj = container_of(uobj, struct ib_umem_object, uobject); | |
e1bcfcaa | 165 | ib_umem_release_on_close(mrdev, &memobj->umem); |
bc38a6ab RD |
166 | |
167 | list_del(&uobj->list); | |
168 | kfree(memobj); | |
169 | } | |
170 | ||
171 | list_for_each_entry_safe(uobj, tmp, &context->pd_list, list) { | |
172 | struct ib_pd *pd = idr_find(&ib_uverbs_pd_idr, uobj->id); | |
173 | idr_remove(&ib_uverbs_pd_idr, uobj->id); | |
174 | ib_dealloc_pd(pd); | |
175 | list_del(&uobj->list); | |
176 | kfree(uobj); | |
177 | } | |
178 | ||
179 | up(&ib_uverbs_idr_mutex); | |
180 | ||
181 | return context->device->dealloc_ucontext(context); | |
182 | } | |
183 | ||
184 | static void ib_uverbs_release_file(struct kref *ref) | |
185 | { | |
186 | struct ib_uverbs_file *file = | |
187 | container_of(ref, struct ib_uverbs_file, ref); | |
188 | ||
189 | module_put(file->device->ib_dev->owner); | |
190 | kfree(file); | |
191 | } | |
192 | ||
193 | static ssize_t ib_uverbs_event_read(struct file *filp, char __user *buf, | |
194 | size_t count, loff_t *pos) | |
195 | { | |
196 | struct ib_uverbs_event_file *file = filp->private_data; | |
63aaf647 | 197 | struct ib_uverbs_event *event; |
bc38a6ab RD |
198 | int eventsz; |
199 | int ret = 0; | |
200 | ||
201 | spin_lock_irq(&file->lock); | |
202 | ||
6b73597e | 203 | while (list_empty(&file->event_list)) { |
bc38a6ab RD |
204 | spin_unlock_irq(&file->lock); |
205 | ||
206 | if (filp->f_flags & O_NONBLOCK) | |
207 | return -EAGAIN; | |
208 | ||
209 | if (wait_event_interruptible(file->poll_wait, | |
6b73597e | 210 | !list_empty(&file->event_list))) |
bc38a6ab RD |
211 | return -ERESTARTSYS; |
212 | ||
213 | spin_lock_irq(&file->lock); | |
214 | } | |
215 | ||
63aaf647 RD |
216 | event = list_entry(file->event_list.next, struct ib_uverbs_event, list); |
217 | ||
218 | if (file->is_async) | |
bc38a6ab | 219 | eventsz = sizeof (struct ib_uverbs_async_event_desc); |
63aaf647 | 220 | else |
bc38a6ab | 221 | eventsz = sizeof (struct ib_uverbs_comp_event_desc); |
bc38a6ab RD |
222 | |
223 | if (eventsz > count) { | |
224 | ret = -EINVAL; | |
225 | event = NULL; | |
63aaf647 | 226 | } else { |
bc38a6ab | 227 | list_del(file->event_list.next); |
63aaf647 RD |
228 | if (event->counter) { |
229 | ++(*event->counter); | |
230 | list_del(&event->obj_list); | |
231 | } | |
232 | } | |
bc38a6ab RD |
233 | |
234 | spin_unlock_irq(&file->lock); | |
235 | ||
236 | if (event) { | |
237 | if (copy_to_user(buf, event, eventsz)) | |
238 | ret = -EFAULT; | |
239 | else | |
240 | ret = eventsz; | |
241 | } | |
242 | ||
243 | kfree(event); | |
244 | ||
245 | return ret; | |
246 | } | |
247 | ||
248 | static unsigned int ib_uverbs_event_poll(struct file *filp, | |
249 | struct poll_table_struct *wait) | |
250 | { | |
251 | unsigned int pollflags = 0; | |
252 | struct ib_uverbs_event_file *file = filp->private_data; | |
253 | ||
254 | poll_wait(filp, &file->poll_wait, wait); | |
255 | ||
256 | spin_lock_irq(&file->lock); | |
6b73597e | 257 | if (!list_empty(&file->event_list)) |
bc38a6ab RD |
258 | pollflags = POLLIN | POLLRDNORM; |
259 | spin_unlock_irq(&file->lock); | |
260 | ||
261 | return pollflags; | |
262 | } | |
263 | ||
6b73597e | 264 | void ib_uverbs_release_event_file(struct kref *ref) |
bc38a6ab | 265 | { |
6b73597e RD |
266 | struct ib_uverbs_event_file *file = |
267 | container_of(ref, struct ib_uverbs_event_file, ref); | |
bc38a6ab | 268 | |
6b73597e | 269 | kfree(file); |
bc38a6ab RD |
270 | } |
271 | ||
abdf119b GN |
272 | static int ib_uverbs_event_fasync(int fd, struct file *filp, int on) |
273 | { | |
274 | struct ib_uverbs_event_file *file = filp->private_data; | |
275 | ||
276 | return fasync_helper(fd, filp, on, &file->async_queue); | |
277 | } | |
278 | ||
bc38a6ab RD |
279 | static int ib_uverbs_event_close(struct inode *inode, struct file *filp) |
280 | { | |
281 | struct ib_uverbs_event_file *file = filp->private_data; | |
6b73597e RD |
282 | struct ib_uverbs_event *entry, *tmp; |
283 | ||
284 | spin_lock_irq(&file->lock); | |
285 | file->file = NULL; | |
286 | list_for_each_entry_safe(entry, tmp, &file->event_list, list) { | |
287 | if (entry->counter) | |
288 | list_del(&entry->obj_list); | |
289 | kfree(entry); | |
290 | } | |
291 | spin_unlock_irq(&file->lock); | |
bc38a6ab | 292 | |
abdf119b | 293 | ib_uverbs_event_fasync(-1, filp, 0); |
6b73597e RD |
294 | |
295 | if (file->is_async) { | |
296 | ib_unregister_event_handler(&file->uverbs_file->event_handler); | |
297 | kref_put(&file->uverbs_file->ref, ib_uverbs_release_file); | |
298 | } | |
299 | kref_put(&file->ref, ib_uverbs_release_event_file); | |
bc38a6ab RD |
300 | |
301 | return 0; | |
302 | } | |
303 | ||
304 | static struct file_operations uverbs_event_fops = { | |
6b73597e | 305 | .owner = THIS_MODULE, |
bc38a6ab RD |
306 | .read = ib_uverbs_event_read, |
307 | .poll = ib_uverbs_event_poll, | |
abdf119b GN |
308 | .release = ib_uverbs_event_close, |
309 | .fasync = ib_uverbs_event_fasync | |
bc38a6ab RD |
310 | }; |
311 | ||
312 | void ib_uverbs_comp_handler(struct ib_cq *cq, void *cq_context) | |
313 | { | |
6b73597e RD |
314 | struct ib_uverbs_event_file *file = cq_context; |
315 | struct ib_ucq_object *uobj; | |
316 | struct ib_uverbs_event *entry; | |
317 | unsigned long flags; | |
318 | ||
319 | if (!file) | |
320 | return; | |
321 | ||
322 | spin_lock_irqsave(&file->lock, flags); | |
323 | if (!file->file) { | |
324 | spin_unlock_irqrestore(&file->lock, flags); | |
325 | return; | |
326 | } | |
bc38a6ab RD |
327 | |
328 | entry = kmalloc(sizeof *entry, GFP_ATOMIC); | |
305a7e87 RD |
329 | if (!entry) { |
330 | spin_unlock_irqrestore(&file->lock, flags); | |
bc38a6ab | 331 | return; |
305a7e87 | 332 | } |
bc38a6ab | 333 | |
63aaf647 RD |
334 | uobj = container_of(cq->uobject, struct ib_ucq_object, uobject); |
335 | ||
336 | entry->desc.comp.cq_handle = cq->uobject->user_handle; | |
337 | entry->counter = &uobj->comp_events_reported; | |
bc38a6ab | 338 | |
6b73597e | 339 | list_add_tail(&entry->list, &file->event_list); |
63aaf647 | 340 | list_add_tail(&entry->obj_list, &uobj->comp_list); |
6b73597e | 341 | spin_unlock_irqrestore(&file->lock, flags); |
bc38a6ab | 342 | |
6b73597e RD |
343 | wake_up_interruptible(&file->poll_wait); |
344 | kill_fasync(&file->async_queue, SIGIO, POLL_IN); | |
bc38a6ab RD |
345 | } |
346 | ||
347 | static void ib_uverbs_async_handler(struct ib_uverbs_file *file, | |
63aaf647 RD |
348 | __u64 element, __u64 event, |
349 | struct list_head *obj_list, | |
350 | u32 *counter) | |
bc38a6ab | 351 | { |
63aaf647 | 352 | struct ib_uverbs_event *entry; |
bc38a6ab RD |
353 | unsigned long flags; |
354 | ||
6b73597e RD |
355 | spin_lock_irqsave(&file->async_file->lock, flags); |
356 | if (!file->async_file->file) { | |
357 | spin_unlock_irqrestore(&file->async_file->lock, flags); | |
358 | return; | |
359 | } | |
360 | ||
bc38a6ab | 361 | entry = kmalloc(sizeof *entry, GFP_ATOMIC); |
305a7e87 RD |
362 | if (!entry) { |
363 | spin_unlock_irqrestore(&file->async_file->lock, flags); | |
bc38a6ab | 364 | return; |
305a7e87 | 365 | } |
bc38a6ab | 366 | |
63aaf647 RD |
367 | entry->desc.async.element = element; |
368 | entry->desc.async.event_type = event; | |
369 | entry->counter = counter; | |
bc38a6ab | 370 | |
6b73597e | 371 | list_add_tail(&entry->list, &file->async_file->event_list); |
63aaf647 RD |
372 | if (obj_list) |
373 | list_add_tail(&entry->obj_list, obj_list); | |
6b73597e | 374 | spin_unlock_irqrestore(&file->async_file->lock, flags); |
bc38a6ab | 375 | |
6b73597e RD |
376 | wake_up_interruptible(&file->async_file->poll_wait); |
377 | kill_fasync(&file->async_file->async_queue, SIGIO, POLL_IN); | |
bc38a6ab RD |
378 | } |
379 | ||
380 | void ib_uverbs_cq_event_handler(struct ib_event *event, void *context_ptr) | |
381 | { | |
6b73597e | 382 | struct ib_uverbs_event_file *ev_file = context_ptr; |
63aaf647 RD |
383 | struct ib_ucq_object *uobj; |
384 | ||
385 | uobj = container_of(event->element.cq->uobject, | |
386 | struct ib_ucq_object, uobject); | |
387 | ||
6b73597e | 388 | ib_uverbs_async_handler(ev_file->uverbs_file, uobj->uobject.user_handle, |
63aaf647 RD |
389 | event->event, &uobj->async_list, |
390 | &uobj->async_events_reported); | |
391 | ||
bc38a6ab RD |
392 | } |
393 | ||
394 | void ib_uverbs_qp_event_handler(struct ib_event *event, void *context_ptr) | |
395 | { | |
63aaf647 RD |
396 | struct ib_uevent_object *uobj; |
397 | ||
398 | uobj = container_of(event->element.qp->uobject, | |
399 | struct ib_uevent_object, uobject); | |
400 | ||
401 | ib_uverbs_async_handler(context_ptr, uobj->uobject.user_handle, | |
402 | event->event, &uobj->event_list, | |
403 | &uobj->events_reported); | |
bc38a6ab RD |
404 | } |
405 | ||
f520ba5a RD |
406 | void ib_uverbs_srq_event_handler(struct ib_event *event, void *context_ptr) |
407 | { | |
63aaf647 RD |
408 | struct ib_uevent_object *uobj; |
409 | ||
410 | uobj = container_of(event->element.srq->uobject, | |
411 | struct ib_uevent_object, uobject); | |
412 | ||
413 | ib_uverbs_async_handler(context_ptr, uobj->uobject.user_handle, | |
414 | event->event, &uobj->event_list, | |
415 | &uobj->events_reported); | |
f520ba5a RD |
416 | } |
417 | ||
6b73597e RD |
418 | void ib_uverbs_event_handler(struct ib_event_handler *handler, |
419 | struct ib_event *event) | |
bc38a6ab RD |
420 | { |
421 | struct ib_uverbs_file *file = | |
422 | container_of(handler, struct ib_uverbs_file, event_handler); | |
423 | ||
63aaf647 RD |
424 | ib_uverbs_async_handler(file, event->element.port_num, event->event, |
425 | NULL, NULL); | |
bc38a6ab RD |
426 | } |
427 | ||
6b73597e RD |
428 | struct file *ib_uverbs_alloc_event_file(struct ib_uverbs_file *uverbs_file, |
429 | int is_async, int *fd) | |
bc38a6ab | 430 | { |
6b73597e | 431 | struct ib_uverbs_event_file *ev_file; |
bc38a6ab | 432 | struct file *filp; |
6b73597e | 433 | int ret; |
bc38a6ab | 434 | |
6b73597e RD |
435 | ev_file = kmalloc(sizeof *ev_file, GFP_KERNEL); |
436 | if (!ev_file) | |
437 | return ERR_PTR(-ENOMEM); | |
438 | ||
439 | kref_init(&ev_file->ref); | |
440 | spin_lock_init(&ev_file->lock); | |
441 | INIT_LIST_HEAD(&ev_file->event_list); | |
442 | init_waitqueue_head(&ev_file->poll_wait); | |
443 | ev_file->uverbs_file = uverbs_file; | |
444 | ev_file->async_queue = NULL; | |
445 | ev_file->is_async = is_async; | |
446 | ||
447 | *fd = get_unused_fd(); | |
448 | if (*fd < 0) { | |
449 | ret = *fd; | |
450 | goto err; | |
451 | } | |
bc38a6ab RD |
452 | |
453 | filp = get_empty_filp(); | |
454 | if (!filp) { | |
6b73597e RD |
455 | ret = -ENFILE; |
456 | goto err_fd; | |
bc38a6ab RD |
457 | } |
458 | ||
6b73597e RD |
459 | ev_file->file = filp; |
460 | ||
461 | /* | |
462 | * fops_get() can't fail here, because we're coming from a | |
463 | * system call on a uverbs file, which will already have a | |
464 | * module reference. | |
465 | */ | |
466 | filp->f_op = fops_get(&uverbs_event_fops); | |
bc38a6ab RD |
467 | filp->f_vfsmnt = mntget(uverbs_event_mnt); |
468 | filp->f_dentry = dget(uverbs_event_mnt->mnt_root); | |
469 | filp->f_mapping = filp->f_dentry->d_inode->i_mapping; | |
470 | filp->f_flags = O_RDONLY; | |
471 | filp->f_mode = FMODE_READ; | |
6b73597e | 472 | filp->private_data = ev_file; |
bc38a6ab | 473 | |
6b73597e | 474 | return filp; |
bc38a6ab | 475 | |
6b73597e RD |
476 | err_fd: |
477 | put_unused_fd(*fd); | |
478 | ||
479 | err: | |
480 | kfree(ev_file); | |
481 | return ERR_PTR(ret); | |
482 | } | |
483 | ||
484 | /* | |
485 | * Look up a completion event file by FD. If lookup is successful, | |
486 | * takes a ref to the event file struct that it returns; if | |
487 | * unsuccessful, returns NULL. | |
488 | */ | |
489 | struct ib_uverbs_event_file *ib_uverbs_lookup_comp_file(int fd) | |
490 | { | |
491 | struct ib_uverbs_event_file *ev_file = NULL; | |
492 | struct file *filp; | |
493 | ||
494 | filp = fget(fd); | |
495 | if (!filp) | |
496 | return NULL; | |
497 | ||
498 | if (filp->f_op != &uverbs_event_fops) | |
499 | goto out; | |
500 | ||
501 | ev_file = filp->private_data; | |
502 | if (ev_file->is_async) { | |
503 | ev_file = NULL; | |
504 | goto out; | |
505 | } | |
506 | ||
507 | kref_get(&ev_file->ref); | |
508 | ||
509 | out: | |
510 | fput(filp); | |
511 | return ev_file; | |
bc38a6ab RD |
512 | } |
513 | ||
514 | static ssize_t ib_uverbs_write(struct file *filp, const char __user *buf, | |
515 | size_t count, loff_t *pos) | |
516 | { | |
517 | struct ib_uverbs_file *file = filp->private_data; | |
518 | struct ib_uverbs_cmd_hdr hdr; | |
519 | ||
520 | if (count < sizeof hdr) | |
521 | return -EINVAL; | |
522 | ||
523 | if (copy_from_user(&hdr, buf, sizeof hdr)) | |
524 | return -EFAULT; | |
525 | ||
526 | if (hdr.in_words * 4 != count) | |
527 | return -EINVAL; | |
528 | ||
63c47c28 RD |
529 | if (hdr.command < 0 || |
530 | hdr.command >= ARRAY_SIZE(uverbs_cmd_table) || | |
883a99c7 RD |
531 | !uverbs_cmd_table[hdr.command] || |
532 | !(file->device->ib_dev->uverbs_cmd_mask & (1ull << hdr.command))) | |
bc38a6ab RD |
533 | return -EINVAL; |
534 | ||
6b73597e | 535 | if (!file->ucontext && |
bc38a6ab RD |
536 | hdr.command != IB_USER_VERBS_CMD_GET_CONTEXT) |
537 | return -EINVAL; | |
538 | ||
539 | return uverbs_cmd_table[hdr.command](file, buf + sizeof hdr, | |
540 | hdr.in_words * 4, hdr.out_words * 4); | |
541 | } | |
542 | ||
543 | static int ib_uverbs_mmap(struct file *filp, struct vm_area_struct *vma) | |
544 | { | |
545 | struct ib_uverbs_file *file = filp->private_data; | |
546 | ||
547 | if (!file->ucontext) | |
548 | return -ENODEV; | |
549 | else | |
550 | return file->device->ib_dev->mmap(file->ucontext, vma); | |
551 | } | |
552 | ||
553 | static int ib_uverbs_open(struct inode *inode, struct file *filp) | |
554 | { | |
555 | struct ib_uverbs_device *dev = | |
556 | container_of(inode->i_cdev, struct ib_uverbs_device, dev); | |
557 | struct ib_uverbs_file *file; | |
bc38a6ab RD |
558 | |
559 | if (!try_module_get(dev->ib_dev->owner)) | |
560 | return -ENODEV; | |
561 | ||
6b73597e | 562 | file = kmalloc(sizeof *file, GFP_KERNEL); |
63c47c28 | 563 | if (!file) { |
6b73597e RD |
564 | module_put(dev->ib_dev->owner); |
565 | return -ENOMEM; | |
63c47c28 | 566 | } |
bc38a6ab | 567 | |
6b73597e RD |
568 | file->device = dev; |
569 | file->ucontext = NULL; | |
bc38a6ab | 570 | kref_init(&file->ref); |
63c47c28 | 571 | init_MUTEX(&file->mutex); |
bc38a6ab | 572 | |
bc38a6ab RD |
573 | filp->private_data = file; |
574 | ||
bc38a6ab | 575 | return 0; |
bc38a6ab RD |
576 | } |
577 | ||
578 | static int ib_uverbs_close(struct inode *inode, struct file *filp) | |
579 | { | |
580 | struct ib_uverbs_file *file = filp->private_data; | |
bc38a6ab | 581 | |
bc38a6ab RD |
582 | ib_dealloc_ucontext(file->ucontext); |
583 | ||
6b73597e | 584 | kref_put(&file->async_file->ref, ib_uverbs_release_event_file); |
bc38a6ab RD |
585 | kref_put(&file->ref, ib_uverbs_release_file); |
586 | ||
587 | return 0; | |
588 | } | |
589 | ||
590 | static struct file_operations uverbs_fops = { | |
591 | .owner = THIS_MODULE, | |
592 | .write = ib_uverbs_write, | |
593 | .open = ib_uverbs_open, | |
594 | .release = ib_uverbs_close | |
595 | }; | |
596 | ||
597 | static struct file_operations uverbs_mmap_fops = { | |
598 | .owner = THIS_MODULE, | |
599 | .write = ib_uverbs_write, | |
600 | .mmap = ib_uverbs_mmap, | |
601 | .open = ib_uverbs_open, | |
602 | .release = ib_uverbs_close | |
603 | }; | |
604 | ||
605 | static struct ib_client uverbs_client = { | |
606 | .name = "uverbs", | |
607 | .add = ib_uverbs_add_one, | |
608 | .remove = ib_uverbs_remove_one | |
609 | }; | |
610 | ||
611 | static ssize_t show_ibdev(struct class_device *class_dev, char *buf) | |
612 | { | |
613 | struct ib_uverbs_device *dev = | |
614 | container_of(class_dev, struct ib_uverbs_device, class_dev); | |
615 | ||
616 | return sprintf(buf, "%s\n", dev->ib_dev->name); | |
617 | } | |
618 | static CLASS_DEVICE_ATTR(ibdev, S_IRUGO, show_ibdev, NULL); | |
619 | ||
274c0891 RD |
620 | static ssize_t show_dev_abi_version(struct class_device *class_dev, char *buf) |
621 | { | |
622 | struct ib_uverbs_device *dev = | |
623 | container_of(class_dev, struct ib_uverbs_device, class_dev); | |
624 | ||
625 | return sprintf(buf, "%d\n", dev->ib_dev->uverbs_abi_ver); | |
626 | } | |
627 | static CLASS_DEVICE_ATTR(abi_version, S_IRUGO, show_dev_abi_version, NULL); | |
628 | ||
bc38a6ab RD |
629 | static void ib_uverbs_release_class_dev(struct class_device *class_dev) |
630 | { | |
631 | struct ib_uverbs_device *dev = | |
632 | container_of(class_dev, struct ib_uverbs_device, class_dev); | |
633 | ||
634 | cdev_del(&dev->dev); | |
635 | clear_bit(dev->devnum, dev_map); | |
636 | kfree(dev); | |
637 | } | |
638 | ||
639 | static struct class uverbs_class = { | |
640 | .name = "infiniband_verbs", | |
641 | .release = ib_uverbs_release_class_dev | |
642 | }; | |
643 | ||
644 | static ssize_t show_abi_version(struct class *class, char *buf) | |
645 | { | |
646 | return sprintf(buf, "%d\n", IB_USER_VERBS_ABI_VERSION); | |
647 | } | |
648 | static CLASS_ATTR(abi_version, S_IRUGO, show_abi_version, NULL); | |
649 | ||
650 | static void ib_uverbs_add_one(struct ib_device *device) | |
651 | { | |
652 | struct ib_uverbs_device *uverbs_dev; | |
653 | ||
654 | if (!device->alloc_ucontext) | |
655 | return; | |
656 | ||
657 | uverbs_dev = kmalloc(sizeof *uverbs_dev, GFP_KERNEL); | |
658 | if (!uverbs_dev) | |
659 | return; | |
660 | ||
661 | memset(uverbs_dev, 0, sizeof *uverbs_dev); | |
662 | ||
663 | spin_lock(&map_lock); | |
664 | uverbs_dev->devnum = find_first_zero_bit(dev_map, IB_UVERBS_MAX_DEVICES); | |
665 | if (uverbs_dev->devnum >= IB_UVERBS_MAX_DEVICES) { | |
666 | spin_unlock(&map_lock); | |
667 | goto err; | |
668 | } | |
669 | set_bit(uverbs_dev->devnum, dev_map); | |
670 | spin_unlock(&map_lock); | |
671 | ||
6b73597e RD |
672 | uverbs_dev->ib_dev = device; |
673 | uverbs_dev->num_comp_vectors = 1; | |
bc38a6ab RD |
674 | |
675 | if (device->mmap) | |
676 | cdev_init(&uverbs_dev->dev, &uverbs_mmap_fops); | |
677 | else | |
678 | cdev_init(&uverbs_dev->dev, &uverbs_fops); | |
679 | uverbs_dev->dev.owner = THIS_MODULE; | |
680 | kobject_set_name(&uverbs_dev->dev.kobj, "uverbs%d", uverbs_dev->devnum); | |
681 | if (cdev_add(&uverbs_dev->dev, IB_UVERBS_BASE_DEV + uverbs_dev->devnum, 1)) | |
682 | goto err; | |
683 | ||
684 | uverbs_dev->class_dev.class = &uverbs_class; | |
685 | uverbs_dev->class_dev.dev = device->dma_device; | |
686 | uverbs_dev->class_dev.devt = uverbs_dev->dev.dev; | |
687 | snprintf(uverbs_dev->class_dev.class_id, BUS_ID_SIZE, "uverbs%d", uverbs_dev->devnum); | |
688 | if (class_device_register(&uverbs_dev->class_dev)) | |
689 | goto err_cdev; | |
690 | ||
691 | if (class_device_create_file(&uverbs_dev->class_dev, &class_device_attr_ibdev)) | |
692 | goto err_class; | |
274c0891 RD |
693 | if (class_device_create_file(&uverbs_dev->class_dev, &class_device_attr_abi_version)) |
694 | goto err_class; | |
bc38a6ab RD |
695 | |
696 | ib_set_client_data(device, &uverbs_client, uverbs_dev); | |
697 | ||
698 | return; | |
699 | ||
700 | err_class: | |
701 | class_device_unregister(&uverbs_dev->class_dev); | |
702 | ||
703 | err_cdev: | |
704 | cdev_del(&uverbs_dev->dev); | |
705 | clear_bit(uverbs_dev->devnum, dev_map); | |
706 | ||
707 | err: | |
708 | kfree(uverbs_dev); | |
709 | return; | |
710 | } | |
711 | ||
712 | static void ib_uverbs_remove_one(struct ib_device *device) | |
713 | { | |
714 | struct ib_uverbs_device *uverbs_dev = ib_get_client_data(device, &uverbs_client); | |
715 | ||
716 | if (!uverbs_dev) | |
717 | return; | |
718 | ||
719 | class_device_unregister(&uverbs_dev->class_dev); | |
720 | } | |
721 | ||
722 | static struct super_block *uverbs_event_get_sb(struct file_system_type *fs_type, int flags, | |
723 | const char *dev_name, void *data) | |
724 | { | |
725 | return get_sb_pseudo(fs_type, "infinibandevent:", NULL, | |
726 | INFINIBANDEVENTFS_MAGIC); | |
727 | } | |
728 | ||
729 | static struct file_system_type uverbs_event_fs = { | |
730 | /* No owner field so module can be unloaded */ | |
731 | .name = "infinibandeventfs", | |
732 | .get_sb = uverbs_event_get_sb, | |
733 | .kill_sb = kill_litter_super | |
734 | }; | |
735 | ||
736 | static int __init ib_uverbs_init(void) | |
737 | { | |
738 | int ret; | |
739 | ||
740 | spin_lock_init(&map_lock); | |
741 | ||
742 | ret = register_chrdev_region(IB_UVERBS_BASE_DEV, IB_UVERBS_MAX_DEVICES, | |
743 | "infiniband_verbs"); | |
744 | if (ret) { | |
745 | printk(KERN_ERR "user_verbs: couldn't register device number\n"); | |
746 | goto out; | |
747 | } | |
748 | ||
749 | ret = class_register(&uverbs_class); | |
750 | if (ret) { | |
751 | printk(KERN_ERR "user_verbs: couldn't create class infiniband_verbs\n"); | |
752 | goto out_chrdev; | |
753 | } | |
754 | ||
755 | ret = class_create_file(&uverbs_class, &class_attr_abi_version); | |
756 | if (ret) { | |
757 | printk(KERN_ERR "user_verbs: couldn't create abi_version attribute\n"); | |
758 | goto out_class; | |
759 | } | |
760 | ||
761 | ret = register_filesystem(&uverbs_event_fs); | |
762 | if (ret) { | |
763 | printk(KERN_ERR "user_verbs: couldn't register infinibandeventfs\n"); | |
764 | goto out_class; | |
765 | } | |
766 | ||
767 | uverbs_event_mnt = kern_mount(&uverbs_event_fs); | |
768 | if (IS_ERR(uverbs_event_mnt)) { | |
769 | ret = PTR_ERR(uverbs_event_mnt); | |
770 | printk(KERN_ERR "user_verbs: couldn't mount infinibandeventfs\n"); | |
771 | goto out_fs; | |
772 | } | |
773 | ||
774 | ret = ib_register_client(&uverbs_client); | |
775 | if (ret) { | |
776 | printk(KERN_ERR "user_verbs: couldn't register client\n"); | |
777 | goto out_mnt; | |
778 | } | |
779 | ||
780 | return 0; | |
781 | ||
782 | out_mnt: | |
783 | mntput(uverbs_event_mnt); | |
784 | ||
785 | out_fs: | |
786 | unregister_filesystem(&uverbs_event_fs); | |
787 | ||
788 | out_class: | |
789 | class_unregister(&uverbs_class); | |
790 | ||
791 | out_chrdev: | |
792 | unregister_chrdev_region(IB_UVERBS_BASE_DEV, IB_UVERBS_MAX_DEVICES); | |
793 | ||
794 | out: | |
795 | return ret; | |
796 | } | |
797 | ||
798 | static void __exit ib_uverbs_cleanup(void) | |
799 | { | |
800 | ib_unregister_client(&uverbs_client); | |
801 | mntput(uverbs_event_mnt); | |
802 | unregister_filesystem(&uverbs_event_fs); | |
803 | class_unregister(&uverbs_class); | |
804 | unregister_chrdev_region(IB_UVERBS_BASE_DEV, IB_UVERBS_MAX_DEVICES); | |
5d7edb3c RD |
805 | idr_destroy(&ib_uverbs_pd_idr); |
806 | idr_destroy(&ib_uverbs_mr_idr); | |
807 | idr_destroy(&ib_uverbs_mw_idr); | |
808 | idr_destroy(&ib_uverbs_ah_idr); | |
809 | idr_destroy(&ib_uverbs_cq_idr); | |
810 | idr_destroy(&ib_uverbs_qp_idr); | |
811 | idr_destroy(&ib_uverbs_srq_idr); | |
bc38a6ab RD |
812 | } |
813 | ||
814 | module_init(ib_uverbs_init); | |
815 | module_exit(ib_uverbs_cleanup); |