2 * Copyright (c) 2006, 2007 Cisco Systems, Inc. All rights reserved.
3 * Copyright (c) 2005 Mellanox Technologies. All rights reserved.
5 * This software is available to you under a choice of one of two
6 * licenses. You may choose to be licensed under the terms of the GNU
7 * General Public License (GPL) Version 2, available from the file
8 * COPYING in the main directory of this source tree, or the
9 * OpenIB.org BSD license below:
11 * Redistribution and use in source and binary forms, with or
12 * without modification, are permitted provided that the following
15 * - Redistributions of source code must retain the above
16 * copyright notice, this list of conditions and the following
19 * - Redistributions in binary form must reproduce the above
20 * copyright notice, this list of conditions and the following
21 * disclaimer in the documentation and/or other materials
22 * provided with the distribution.
24 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
25 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
26 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
27 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
28 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
29 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
30 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
34 #include <linux/init.h>
35 #include <linux/errno.h>
36 #include <linux/export.h>
37 #include <linux/io-mapping.h>
45 MLX4_NUM_RESERVED_UARS
= 8
48 int mlx4_pd_alloc(struct mlx4_dev
*dev
, u32
*pdn
)
50 struct mlx4_priv
*priv
= mlx4_priv(dev
);
52 *pdn
= mlx4_bitmap_alloc(&priv
->pd_bitmap
);
55 if (mlx4_is_mfunc(dev
))
56 *pdn
|= (dev
->caps
.function
+ 1) << NOT_MASKED_PD_BITS
;
59 EXPORT_SYMBOL_GPL(mlx4_pd_alloc
);
61 void mlx4_pd_free(struct mlx4_dev
*dev
, u32 pdn
)
63 mlx4_bitmap_free(&mlx4_priv(dev
)->pd_bitmap
, pdn
);
65 EXPORT_SYMBOL_GPL(mlx4_pd_free
);
67 int mlx4_xrcd_alloc(struct mlx4_dev
*dev
, u32
*xrcdn
)
69 struct mlx4_priv
*priv
= mlx4_priv(dev
);
71 *xrcdn
= mlx4_bitmap_alloc(&priv
->xrcd_bitmap
);
77 EXPORT_SYMBOL_GPL(mlx4_xrcd_alloc
);
79 void mlx4_xrcd_free(struct mlx4_dev
*dev
, u32 xrcdn
)
81 mlx4_bitmap_free(&mlx4_priv(dev
)->xrcd_bitmap
, xrcdn
);
83 EXPORT_SYMBOL_GPL(mlx4_xrcd_free
);
85 int mlx4_init_pd_table(struct mlx4_dev
*dev
)
87 struct mlx4_priv
*priv
= mlx4_priv(dev
);
89 return mlx4_bitmap_init(&priv
->pd_bitmap
, dev
->caps
.num_pds
,
90 (1 << NOT_MASKED_PD_BITS
) - 1,
91 dev
->caps
.reserved_pds
, 0);
94 void mlx4_cleanup_pd_table(struct mlx4_dev
*dev
)
96 mlx4_bitmap_cleanup(&mlx4_priv(dev
)->pd_bitmap
);
99 int mlx4_init_xrcd_table(struct mlx4_dev
*dev
)
101 struct mlx4_priv
*priv
= mlx4_priv(dev
);
103 return mlx4_bitmap_init(&priv
->xrcd_bitmap
, (1 << 16),
104 (1 << 16) - 1, dev
->caps
.reserved_xrcds
+ 1, 0);
107 void mlx4_cleanup_xrcd_table(struct mlx4_dev
*dev
)
109 mlx4_bitmap_cleanup(&mlx4_priv(dev
)->xrcd_bitmap
);
112 int mlx4_uar_alloc(struct mlx4_dev
*dev
, struct mlx4_uar
*uar
)
116 uar
->index
= mlx4_bitmap_alloc(&mlx4_priv(dev
)->uar_table
.bitmap
);
117 if (uar
->index
== -1)
120 if (mlx4_is_slave(dev
))
121 offset
= uar
->index
% ((int) pci_resource_len(dev
->pdev
, 2) /
122 dev
->caps
.uar_page_size
);
125 uar
->pfn
= (pci_resource_start(dev
->pdev
, 2) >> PAGE_SHIFT
) + offset
;
129 EXPORT_SYMBOL_GPL(mlx4_uar_alloc
);
131 void mlx4_uar_free(struct mlx4_dev
*dev
, struct mlx4_uar
*uar
)
133 mlx4_bitmap_free(&mlx4_priv(dev
)->uar_table
.bitmap
, uar
->index
);
135 EXPORT_SYMBOL_GPL(mlx4_uar_free
);
137 int mlx4_bf_alloc(struct mlx4_dev
*dev
, struct mlx4_bf
*bf
)
139 struct mlx4_priv
*priv
= mlx4_priv(dev
);
140 struct mlx4_uar
*uar
;
144 if (!priv
->bf_mapping
)
147 mutex_lock(&priv
->bf_mutex
);
148 if (!list_empty(&priv
->bf_list
))
149 uar
= list_entry(priv
->bf_list
.next
, struct mlx4_uar
, bf_list
);
151 if (mlx4_bitmap_avail(&priv
->uar_table
.bitmap
) < MLX4_NUM_RESERVED_UARS
) {
155 uar
= kmalloc(sizeof *uar
, GFP_KERNEL
);
160 err
= mlx4_uar_alloc(dev
, uar
);
164 uar
->map
= ioremap(uar
->pfn
<< PAGE_SHIFT
, PAGE_SIZE
);
170 uar
->bf_map
= io_mapping_map_wc(priv
->bf_mapping
, uar
->index
<< PAGE_SHIFT
);
175 uar
->free_bf_bmap
= 0;
176 list_add(&uar
->bf_list
, &priv
->bf_list
);
180 idx
= ffz(uar
->free_bf_bmap
);
181 uar
->free_bf_bmap
|= 1 << idx
;
184 bf
->buf_size
= dev
->caps
.bf_reg_size
/ 2;
185 bf
->reg
= uar
->bf_map
+ idx
* dev
->caps
.bf_reg_size
;
186 if (uar
->free_bf_bmap
== (1 << dev
->caps
.bf_regs_per_page
) - 1)
187 list_del_init(&uar
->bf_list
);
196 mlx4_uar_free(dev
, uar
);
202 mutex_unlock(&priv
->bf_mutex
);
205 EXPORT_SYMBOL_GPL(mlx4_bf_alloc
);
207 void mlx4_bf_free(struct mlx4_dev
*dev
, struct mlx4_bf
*bf
)
209 struct mlx4_priv
*priv
= mlx4_priv(dev
);
212 if (!bf
->uar
|| !bf
->uar
->bf_map
)
215 mutex_lock(&priv
->bf_mutex
);
216 idx
= (bf
->reg
- bf
->uar
->bf_map
) / dev
->caps
.bf_reg_size
;
217 bf
->uar
->free_bf_bmap
&= ~(1 << idx
);
218 if (!bf
->uar
->free_bf_bmap
) {
219 if (!list_empty(&bf
->uar
->bf_list
))
220 list_del(&bf
->uar
->bf_list
);
222 io_mapping_unmap(bf
->uar
->bf_map
);
223 iounmap(bf
->uar
->map
);
224 mlx4_uar_free(dev
, bf
->uar
);
226 } else if (list_empty(&bf
->uar
->bf_list
))
227 list_add(&bf
->uar
->bf_list
, &priv
->bf_list
);
229 mutex_unlock(&priv
->bf_mutex
);
231 EXPORT_SYMBOL_GPL(mlx4_bf_free
);
233 int mlx4_init_uar_table(struct mlx4_dev
*dev
)
235 if (dev
->caps
.num_uars
<= 128) {
236 mlx4_err(dev
, "Only %d UAR pages (need more than 128)\n",
238 mlx4_err(dev
, "Increase firmware log2_uar_bar_megabytes?\n");
242 return mlx4_bitmap_init(&mlx4_priv(dev
)->uar_table
.bitmap
,
243 dev
->caps
.num_uars
, dev
->caps
.num_uars
- 1,
244 dev
->caps
.reserved_uars
, 0);
247 void mlx4_cleanup_uar_table(struct mlx4_dev
*dev
)
249 mlx4_bitmap_cleanup(&mlx4_priv(dev
)->uar_table
.bitmap
);