i40iw: add hmc resource files
authorFaisal Latif <faisal.latif@intel.com>
Wed, 20 Jan 2016 19:40:07 +0000 (13:40 -0600)
committerDoug Ledford <dledford@redhat.com>
Mon, 29 Feb 2016 22:10:53 +0000 (17:10 -0500)
i40iw_hmc.[ch] are to manage hmc for the device.

Acked-by: Anjali Singhai Jain <anjali.singhai@intel.com>
Acked-by: Shannon Nelson <shannon.nelson@intel.com>
Signed-off-by: Faisal Latif <faisal.latif@intel.com>
Signed-off-by: Doug Ledford <dledford@redhat.com>
drivers/infiniband/hw/i40iw/i40iw_hmc.c [new file with mode: 0644]
drivers/infiniband/hw/i40iw/i40iw_hmc.h [new file with mode: 0644]

diff --git a/drivers/infiniband/hw/i40iw/i40iw_hmc.c b/drivers/infiniband/hw/i40iw/i40iw_hmc.c
new file mode 100644 (file)
index 0000000..5484cbf
--- /dev/null
@@ -0,0 +1,821 @@
+/*******************************************************************************
+*
+* Copyright (c) 2015-2016 Intel Corporation.  All rights reserved.
+*
+* This software is available to you under a choice of one of two
+* licenses.  You may choose to be licensed under the terms of the GNU
+* General Public License (GPL) Version 2, available from the file
+* COPYING in the main directory of this source tree, or the
+* OpenFabrics.org BSD license below:
+*
+*   Redistribution and use in source and binary forms, with or
+*   without modification, are permitted provided that the following
+*   conditions are met:
+*
+*    - Redistributions of source code must retain the above
+*      copyright notice, this list of conditions and the following
+*      disclaimer.
+*
+*    - Redistributions in binary form must reproduce the above
+*      copyright notice, this list of conditions and the following
+*      disclaimer in the documentation and/or other materials
+*      provided with the distribution.
+*
+* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
+* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
+* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
+* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
+* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
+* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
+* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
+* SOFTWARE.
+*
+*******************************************************************************/
+
+#include "i40iw_osdep.h"
+#include "i40iw_register.h"
+#include "i40iw_status.h"
+#include "i40iw_hmc.h"
+#include "i40iw_d.h"
+#include "i40iw_type.h"
+#include "i40iw_p.h"
+#include "i40iw_vf.h"
+#include "i40iw_virtchnl.h"
+
+/**
+ * i40iw_find_sd_index_limit - finds segment descriptor index limit
+ * @hmc_info: pointer to the HMC configuration information structure
+ * @type: type of HMC resources we're searching
+ * @index: starting index for the object
+ * @cnt: number of objects we're trying to create
+ * @sd_idx: pointer to return index of the segment descriptor in question
+ * @sd_limit: pointer to return the maximum number of segment descriptors
+ *
+ * This function calculates the segment descriptor index and index limit
+ * for the resource defined by i40iw_hmc_rsrc_type.
+ */
+
+static inline void i40iw_find_sd_index_limit(struct i40iw_hmc_info *hmc_info,
+                                            u32 type,
+                                            u32 idx,
+                                            u32 cnt,
+                                            u32 *sd_idx,
+                                            u32 *sd_limit)
+{
+       u64 fpm_addr, fpm_limit;
+
+       fpm_addr = hmc_info->hmc_obj[(type)].base +
+                       hmc_info->hmc_obj[type].size * idx;
+       fpm_limit = fpm_addr + hmc_info->hmc_obj[type].size * cnt;
+       *sd_idx = (u32)(fpm_addr / I40IW_HMC_DIRECT_BP_SIZE);
+       *sd_limit = (u32)((fpm_limit - 1) / I40IW_HMC_DIRECT_BP_SIZE);
+       *sd_limit += 1;
+}
+
+/**
+ * i40iw_find_pd_index_limit - finds page descriptor index limit
+ * @hmc_info: pointer to the HMC configuration information struct
+ * @type: HMC resource type we're examining
+ * @idx: starting index for the object
+ * @cnt: number of objects we're trying to create
+ * @pd_index: pointer to return page descriptor index
+ * @pd_limit: pointer to return page descriptor index limit
+ *
+ * Calculates the page descriptor index and index limit for the resource
+ * defined by i40iw_hmc_rsrc_type.
+ */
+
+static inline void i40iw_find_pd_index_limit(struct i40iw_hmc_info *hmc_info,
+                                            u32 type,
+                                            u32 idx,
+                                            u32 cnt,
+                                            u32 *pd_idx,
+                                            u32 *pd_limit)
+{
+       u64 fpm_adr, fpm_limit;
+
+       fpm_adr = hmc_info->hmc_obj[type].base +
+                       hmc_info->hmc_obj[type].size * idx;
+       fpm_limit = fpm_adr + (hmc_info)->hmc_obj[(type)].size * (cnt);
+       *(pd_idx) = (u32)(fpm_adr / I40IW_HMC_PAGED_BP_SIZE);
+       *(pd_limit) = (u32)((fpm_limit - 1) / I40IW_HMC_PAGED_BP_SIZE);
+       *(pd_limit) += 1;
+}
+
+/**
+ * i40iw_set_sd_entry - setup entry for sd programming
+ * @pa: physical addr
+ * @idx: sd index
+ * @type: paged or direct sd
+ * @entry: sd entry ptr
+ */
+static inline void i40iw_set_sd_entry(u64 pa,
+                                     u32 idx,
+                                     enum i40iw_sd_entry_type type,
+                                     struct update_sd_entry *entry)
+{
+       entry->data = pa | (I40IW_HMC_MAX_BP_COUNT << I40E_PFHMC_SDDATALOW_PMSDBPCOUNT_SHIFT) |
+                       (((type == I40IW_SD_TYPE_PAGED) ? 0 : 1) <<
+                               I40E_PFHMC_SDDATALOW_PMSDTYPE_SHIFT) |
+                       (1 << I40E_PFHMC_SDDATALOW_PMSDVALID_SHIFT);
+       entry->cmd = (idx | (1 << I40E_PFHMC_SDCMD_PMSDWR_SHIFT) | (1 << 15));
+}
+
+/**
+ * i40iw_clr_sd_entry - setup entry for sd clear
+ * @idx: sd index
+ * @type: paged or direct sd
+ * @entry: sd entry ptr
+ */
+static inline void i40iw_clr_sd_entry(u32 idx, enum i40iw_sd_entry_type type,
+                                     struct update_sd_entry *entry)
+{
+       entry->data = (I40IW_HMC_MAX_BP_COUNT <<
+                       I40E_PFHMC_SDDATALOW_PMSDBPCOUNT_SHIFT) |
+                       (((type == I40IW_SD_TYPE_PAGED) ? 0 : 1) <<
+                               I40E_PFHMC_SDDATALOW_PMSDTYPE_SHIFT);
+       entry->cmd = (idx | (1 << I40E_PFHMC_SDCMD_PMSDWR_SHIFT) | (1 << 15));
+}
+
+/**
+ * i40iw_hmc_sd_one - setup 1 sd entry for cqp
+ * @dev: pointer to the device structure
+ * @hmc_fn_id: hmc's function id
+ * @pa: physical addr
+ * @sd_idx: sd index
+ * @type: paged or direct sd
+ * @setsd: flag to set or clear sd
+ */
+enum i40iw_status_code i40iw_hmc_sd_one(struct i40iw_sc_dev *dev,
+                                       u8 hmc_fn_id,
+                                       u64 pa, u32 sd_idx,
+                                       enum i40iw_sd_entry_type type,
+                                       bool setsd)
+{
+       struct i40iw_update_sds_info sdinfo;
+
+       sdinfo.cnt = 1;
+       sdinfo.hmc_fn_id = hmc_fn_id;
+       if (setsd)
+               i40iw_set_sd_entry(pa, sd_idx, type, sdinfo.entry);
+       else
+               i40iw_clr_sd_entry(sd_idx, type, sdinfo.entry);
+
+       return dev->cqp->process_cqp_sds(dev, &sdinfo);
+}
+
+/**
+ * i40iw_hmc_sd_grp - setup group od sd entries for cqp
+ * @dev: pointer to the device structure
+ * @hmc_info: pointer to the HMC configuration information struct
+ * @sd_index: sd index
+ * @sd_cnt: number of sd entries
+ * @setsd: flag to set or clear sd
+ */
+static enum i40iw_status_code i40iw_hmc_sd_grp(struct i40iw_sc_dev *dev,
+                                              struct i40iw_hmc_info *hmc_info,
+                                              u32 sd_index,
+                                              u32 sd_cnt,
+                                              bool setsd)
+{
+       struct i40iw_hmc_sd_entry *sd_entry;
+       struct i40iw_update_sds_info sdinfo;
+       u64 pa;
+       u32 i;
+       enum i40iw_status_code ret_code = 0;
+
+       memset(&sdinfo, 0, sizeof(sdinfo));
+       sdinfo.hmc_fn_id = hmc_info->hmc_fn_id;
+       for (i = sd_index; i < sd_index + sd_cnt; i++) {
+               sd_entry = &hmc_info->sd_table.sd_entry[i];
+               if (!sd_entry ||
+                   (!sd_entry->valid && setsd) ||
+                   (sd_entry->valid && !setsd))
+                       continue;
+               if (setsd) {
+                       pa = (sd_entry->entry_type == I40IW_SD_TYPE_PAGED) ?
+                           sd_entry->u.pd_table.pd_page_addr.pa :
+                           sd_entry->u.bp.addr.pa;
+                       i40iw_set_sd_entry(pa, i, sd_entry->entry_type,
+                                          &sdinfo.entry[sdinfo.cnt]);
+               } else {
+                       i40iw_clr_sd_entry(i, sd_entry->entry_type,
+                                          &sdinfo.entry[sdinfo.cnt]);
+               }
+               sdinfo.cnt++;
+               if (sdinfo.cnt == I40IW_MAX_SD_ENTRIES) {
+                       ret_code = dev->cqp->process_cqp_sds(dev, &sdinfo);
+                       if (ret_code) {
+                               i40iw_debug(dev, I40IW_DEBUG_HMC,
+                                           "i40iw_hmc_sd_grp: sd_programming failed err=%d\n",
+                                           ret_code);
+                               return ret_code;
+                       }
+                       sdinfo.cnt = 0;
+               }
+       }
+       if (sdinfo.cnt)
+               ret_code = dev->cqp->process_cqp_sds(dev, &sdinfo);
+
+       return ret_code;
+}
+
+/**
+ * i40iw_vfdev_from_fpm - return vf dev ptr for hmc function id
+ * @dev: pointer to the device structure
+ * @hmc_fn_id: hmc's function id
+ */
+struct i40iw_vfdev *i40iw_vfdev_from_fpm(struct i40iw_sc_dev *dev, u8 hmc_fn_id)
+{
+       struct i40iw_vfdev *vf_dev = NULL;
+       u16 idx;
+
+       for (idx = 0; idx < I40IW_MAX_PE_ENABLED_VF_COUNT; idx++) {
+               if (dev->vf_dev[idx] &&
+                   ((u8)dev->vf_dev[idx]->pmf_index == hmc_fn_id)) {
+                       vf_dev = dev->vf_dev[idx];
+                       break;
+               }
+       }
+       return vf_dev;
+}
+
+/**
+ * i40iw_vf_hmcinfo_from_fpm - get ptr to hmc for func_id
+ * @dev: pointer to the device structure
+ * @hmc_fn_id: hmc's function id
+ */
+struct i40iw_hmc_info *i40iw_vf_hmcinfo_from_fpm(struct i40iw_sc_dev *dev,
+                                                u8 hmc_fn_id)
+{
+       struct i40iw_hmc_info *hmc_info = NULL;
+       u16 idx;
+
+       for (idx = 0; idx < I40IW_MAX_PE_ENABLED_VF_COUNT; idx++) {
+               if (dev->vf_dev[idx] &&
+                   ((u8)dev->vf_dev[idx]->pmf_index == hmc_fn_id)) {
+                       hmc_info = &dev->vf_dev[idx]->hmc_info;
+                       break;
+               }
+       }
+       return hmc_info;
+}
+
+/**
+ * i40iw_hmc_finish_add_sd_reg - program sd entries for objects
+ * @dev: pointer to the device structure
+ * @info: create obj info
+ */
+static enum i40iw_status_code i40iw_hmc_finish_add_sd_reg(struct i40iw_sc_dev *dev,
+                                                         struct i40iw_hmc_create_obj_info *info)
+{
+       if (info->start_idx >= info->hmc_info->hmc_obj[info->rsrc_type].cnt)
+               return I40IW_ERR_INVALID_HMC_OBJ_INDEX;
+
+       if ((info->start_idx + info->count) >
+                       info->hmc_info->hmc_obj[info->rsrc_type].cnt)
+               return I40IW_ERR_INVALID_HMC_OBJ_COUNT;
+
+       if (!info->add_sd_cnt)
+               return 0;
+
+       return i40iw_hmc_sd_grp(dev, info->hmc_info,
+                               info->hmc_info->sd_indexes[0],
+                               info->add_sd_cnt, true);
+}
+
+/**
+ * i40iw_create_iw_hmc_obj - allocate backing store for hmc objects
+ * @dev: pointer to the device structure
+ * @info: pointer to i40iw_hmc_iw_create_obj_info struct
+ *
+ * This will allocate memory for PDs and backing pages and populate
+ * the sd and pd entries.
+ */
+enum i40iw_status_code i40iw_sc_create_hmc_obj(struct i40iw_sc_dev *dev,
+                                              struct i40iw_hmc_create_obj_info *info)
+{
+       struct i40iw_hmc_sd_entry *sd_entry;
+       u32 sd_idx, sd_lmt;
+       u32 pd_idx = 0, pd_lmt = 0;
+       u32 pd_idx1 = 0, pd_lmt1 = 0;
+       u32 i, j;
+       bool pd_error = false;
+       enum i40iw_status_code ret_code = 0;
+
+       if (info->start_idx >= info->hmc_info->hmc_obj[info->rsrc_type].cnt)
+               return I40IW_ERR_INVALID_HMC_OBJ_INDEX;
+
+       if ((info->start_idx + info->count) >
+           info->hmc_info->hmc_obj[info->rsrc_type].cnt) {
+               i40iw_debug(dev, I40IW_DEBUG_HMC,
+                           "%s: error type %u, start = %u, req cnt %u, cnt = %u\n",
+                           __func__, info->rsrc_type, info->start_idx, info->count,
+                           info->hmc_info->hmc_obj[info->rsrc_type].cnt);
+               return I40IW_ERR_INVALID_HMC_OBJ_COUNT;
+       }
+
+       if (!dev->is_pf)
+               return i40iw_vchnl_vf_add_hmc_objs(dev, info->rsrc_type, 0, info->count);
+
+       i40iw_find_sd_index_limit(info->hmc_info, info->rsrc_type,
+                                 info->start_idx, info->count,
+                                 &sd_idx, &sd_lmt);
+       if (sd_idx >= info->hmc_info->sd_table.sd_cnt ||
+           sd_lmt > info->hmc_info->sd_table.sd_cnt) {
+               return I40IW_ERR_INVALID_SD_INDEX;
+       }
+       i40iw_find_pd_index_limit(info->hmc_info, info->rsrc_type,
+                                 info->start_idx, info->count, &pd_idx, &pd_lmt);
+
+       for (j = sd_idx; j < sd_lmt; j++) {
+               ret_code = i40iw_add_sd_table_entry(dev->hw, info->hmc_info,
+                                                   j,
+                                                   info->entry_type,
+                                                   I40IW_HMC_DIRECT_BP_SIZE);
+               if (ret_code)
+                       goto exit_sd_error;
+               sd_entry = &info->hmc_info->sd_table.sd_entry[j];
+
+               if ((sd_entry->entry_type == I40IW_SD_TYPE_PAGED) &&
+                   ((dev->hmc_info == info->hmc_info) &&
+                    (info->rsrc_type != I40IW_HMC_IW_PBLE))) {
+                       pd_idx1 = max(pd_idx, (j * I40IW_HMC_MAX_BP_COUNT));
+                       pd_lmt1 = min(pd_lmt,
+                                     (j + 1) * I40IW_HMC_MAX_BP_COUNT);
+                       for (i = pd_idx1; i < pd_lmt1; i++) {
+                               /* update the pd table entry */
+                               ret_code = i40iw_add_pd_table_entry(dev->hw, info->hmc_info,
+                                                                   i, NULL);
+                               if (ret_code) {
+                                       pd_error = true;
+                                       break;
+                               }
+                       }
+                       if (pd_error) {
+                               while (i && (i > pd_idx1)) {
+                                       i40iw_remove_pd_bp(dev->hw, info->hmc_info, (i - 1),
+                                                          info->is_pf);
+                                       i--;
+                               }
+                       }
+               }
+               if (sd_entry->valid)
+                       continue;
+
+               info->hmc_info->sd_indexes[info->add_sd_cnt] = (u16)j;
+               info->add_sd_cnt++;
+               sd_entry->valid = true;
+       }
+       return i40iw_hmc_finish_add_sd_reg(dev, info);
+
+exit_sd_error:
+       while (j && (j > sd_idx)) {
+               sd_entry = &info->hmc_info->sd_table.sd_entry[j - 1];
+               switch (sd_entry->entry_type) {
+               case I40IW_SD_TYPE_PAGED:
+                       pd_idx1 = max(pd_idx,
+                                     (j - 1) * I40IW_HMC_MAX_BP_COUNT);
+                       pd_lmt1 = min(pd_lmt, (j * I40IW_HMC_MAX_BP_COUNT));
+                       for (i = pd_idx1; i < pd_lmt1; i++)
+                               i40iw_prep_remove_pd_page(info->hmc_info, i);
+                       break;
+               case I40IW_SD_TYPE_DIRECT:
+                       i40iw_prep_remove_pd_page(info->hmc_info, (j - 1));
+                       break;
+               default:
+                       ret_code = I40IW_ERR_INVALID_SD_TYPE;
+                       break;
+               }
+               j--;
+       }
+
+       return ret_code;
+}
+
+/**
+ * i40iw_finish_del_sd_reg - delete sd entries for objects
+ * @dev: pointer to the device structure
+ * @info: dele obj info
+ * @reset: true if called before reset
+ */
+static enum i40iw_status_code i40iw_finish_del_sd_reg(struct i40iw_sc_dev *dev,
+                                                     struct i40iw_hmc_del_obj_info *info,
+                                                     bool reset)
+{
+       struct i40iw_hmc_sd_entry *sd_entry;
+       enum i40iw_status_code ret_code = 0;
+       u32 i, sd_idx;
+       struct i40iw_dma_mem *mem;
+
+       if (dev->is_pf && !reset)
+               ret_code = i40iw_hmc_sd_grp(dev, info->hmc_info,
+                                           info->hmc_info->sd_indexes[0],
+                                           info->del_sd_cnt, false);
+
+       if (ret_code)
+               i40iw_debug(dev, I40IW_DEBUG_HMC, "%s: error cqp sd sd_grp\n", __func__);
+
+       for (i = 0; i < info->del_sd_cnt; i++) {
+               sd_idx = info->hmc_info->sd_indexes[i];
+               sd_entry = &info->hmc_info->sd_table.sd_entry[sd_idx];
+               if (!sd_entry)
+                       continue;
+               mem = (sd_entry->entry_type == I40IW_SD_TYPE_PAGED) ?
+                       &sd_entry->u.pd_table.pd_page_addr :
+                       &sd_entry->u.bp.addr;
+
+               if (!mem || !mem->va)
+                       i40iw_debug(dev, I40IW_DEBUG_HMC, "%s: error cqp sd mem\n", __func__);
+               else
+                       i40iw_free_dma_mem(dev->hw, mem);
+       }
+       return ret_code;
+}
+
+/**
+ * i40iw_del_iw_hmc_obj - remove pe hmc objects
+ * @dev: pointer to the device structure
+ * @info: pointer to i40iw_hmc_del_obj_info struct
+ * @reset: true if called before reset
+ *
+ * This will de-populate the SDs and PDs.  It frees
+ * the memory for PDS and backing storage.  After this function is returned,
+ * caller should deallocate memory allocated previously for
+ * book-keeping information about PDs and backing storage.
+ */
+enum i40iw_status_code i40iw_sc_del_hmc_obj(struct i40iw_sc_dev *dev,
+                                           struct i40iw_hmc_del_obj_info *info,
+                                           bool reset)
+{
+       struct i40iw_hmc_pd_table *pd_table;
+       u32 sd_idx, sd_lmt;
+       u32 pd_idx, pd_lmt, rel_pd_idx;
+       u32 i, j;
+       enum i40iw_status_code ret_code = 0;
+
+       if (info->start_idx >= info->hmc_info->hmc_obj[info->rsrc_type].cnt) {
+               i40iw_debug(dev, I40IW_DEBUG_HMC,
+                           "%s: error start_idx[%04d]  >= [type %04d].cnt[%04d]\n",
+                           __func__, info->start_idx, info->rsrc_type,
+                           info->hmc_info->hmc_obj[info->rsrc_type].cnt);
+               return I40IW_ERR_INVALID_HMC_OBJ_INDEX;
+       }
+
+       if ((info->start_idx + info->count) >
+           info->hmc_info->hmc_obj[info->rsrc_type].cnt) {
+               i40iw_debug(dev, I40IW_DEBUG_HMC,
+                           "%s: error start_idx[%04d] + count %04d  >= [type %04d].cnt[%04d]\n",
+                           __func__, info->start_idx, info->count,
+                           info->rsrc_type,
+                           info->hmc_info->hmc_obj[info->rsrc_type].cnt);
+               return I40IW_ERR_INVALID_HMC_OBJ_COUNT;
+       }
+       if (!dev->is_pf) {
+               ret_code = i40iw_vchnl_vf_del_hmc_obj(dev, info->rsrc_type, 0,
+                                                     info->count);
+               if (info->rsrc_type != I40IW_HMC_IW_PBLE)
+                       return ret_code;
+       }
+
+       i40iw_find_pd_index_limit(info->hmc_info, info->rsrc_type,
+                                 info->start_idx, info->count, &pd_idx, &pd_lmt);
+
+       for (j = pd_idx; j < pd_lmt; j++) {
+               sd_idx = j / I40IW_HMC_PD_CNT_IN_SD;
+
+               if (info->hmc_info->sd_table.sd_entry[sd_idx].entry_type !=
+                   I40IW_SD_TYPE_PAGED)
+                       continue;
+
+               rel_pd_idx = j % I40IW_HMC_PD_CNT_IN_SD;
+               pd_table = &info->hmc_info->sd_table.sd_entry[sd_idx].u.pd_table;
+               if (pd_table->pd_entry[rel_pd_idx].valid) {
+                       ret_code = i40iw_remove_pd_bp(dev->hw, info->hmc_info, j,
+                                                     info->is_pf);
+                       if (ret_code) {
+                               i40iw_debug(dev, I40IW_DEBUG_HMC, "%s: error\n", __func__);
+                               return ret_code;
+                       }
+               }
+       }
+
+       i40iw_find_sd_index_limit(info->hmc_info, info->rsrc_type,
+                                 info->start_idx, info->count, &sd_idx, &sd_lmt);
+       if (sd_idx >= info->hmc_info->sd_table.sd_cnt ||
+           sd_lmt > info->hmc_info->sd_table.sd_cnt) {
+               i40iw_debug(dev, I40IW_DEBUG_HMC, "%s: error invalid sd_idx\n", __func__);
+               return I40IW_ERR_INVALID_SD_INDEX;
+       }
+
+       for (i = sd_idx; i < sd_lmt; i++) {
+               if (!info->hmc_info->sd_table.sd_entry[i].valid)
+                       continue;
+               switch (info->hmc_info->sd_table.sd_entry[i].entry_type) {
+               case I40IW_SD_TYPE_DIRECT:
+                       ret_code = i40iw_prep_remove_sd_bp(info->hmc_info, i);
+                       if (!ret_code) {
+                               info->hmc_info->sd_indexes[info->del_sd_cnt] = (u16)i;
+                               info->del_sd_cnt++;
+                       }
+                       break;
+               case I40IW_SD_TYPE_PAGED:
+                       ret_code = i40iw_prep_remove_pd_page(info->hmc_info, i);
+                       if (!ret_code) {
+                               info->hmc_info->sd_indexes[info->del_sd_cnt] = (u16)i;
+                               info->del_sd_cnt++;
+                       }
+                       break;
+               default:
+                       break;
+               }
+       }
+       return i40iw_finish_del_sd_reg(dev, info, reset);
+}
+
+/**
+ * i40iw_add_sd_table_entry - Adds a segment descriptor to the table
+ * @hw: pointer to our hw struct
+ * @hmc_info: pointer to the HMC configuration information struct
+ * @sd_index: segment descriptor index to manipulate
+ * @type: what type of segment descriptor we're manipulating
+ * @direct_mode_sz: size to alloc in direct mode
+ */
+enum i40iw_status_code i40iw_add_sd_table_entry(struct i40iw_hw *hw,
+                                               struct i40iw_hmc_info *hmc_info,
+                                               u32 sd_index,
+                                               enum i40iw_sd_entry_type type,
+                                               u64 direct_mode_sz)
+{
+       enum i40iw_status_code ret_code = 0;
+       struct i40iw_hmc_sd_entry *sd_entry;
+       bool dma_mem_alloc_done = false;
+       struct i40iw_dma_mem mem;
+       u64 alloc_len;
+
+       sd_entry = &hmc_info->sd_table.sd_entry[sd_index];
+       if (!sd_entry->valid) {
+               if (type == I40IW_SD_TYPE_PAGED)
+                       alloc_len = I40IW_HMC_PAGED_BP_SIZE;
+               else
+                       alloc_len = direct_mode_sz;
+
+               /* allocate a 4K pd page or 2M backing page */
+               ret_code = i40iw_allocate_dma_mem(hw, &mem, alloc_len,
+                                                 I40IW_HMC_PD_BP_BUF_ALIGNMENT);
+               if (ret_code)
+                       goto exit;
+               dma_mem_alloc_done = true;
+               if (type == I40IW_SD_TYPE_PAGED) {
+                       ret_code = i40iw_allocate_virt_mem(hw,
+                                                          &sd_entry->u.pd_table.pd_entry_virt_mem,
+                                                          sizeof(struct i40iw_hmc_pd_entry) * 512);
+                       if (ret_code)
+                               goto exit;
+                       sd_entry->u.pd_table.pd_entry = (struct i40iw_hmc_pd_entry *)
+                                                        sd_entry->u.pd_table.pd_entry_virt_mem.va;
+
+                       memcpy(&sd_entry->u.pd_table.pd_page_addr, &mem, sizeof(struct i40iw_dma_mem));
+               } else {
+                       memcpy(&sd_entry->u.bp.addr, &mem, sizeof(struct i40iw_dma_mem));
+                       sd_entry->u.bp.sd_pd_index = sd_index;
+               }
+
+               hmc_info->sd_table.sd_entry[sd_index].entry_type = type;
+
+               I40IW_INC_SD_REFCNT(&hmc_info->sd_table);
+       }
+       if (sd_entry->entry_type == I40IW_SD_TYPE_DIRECT)
+               I40IW_INC_BP_REFCNT(&sd_entry->u.bp);
+exit:
+       if (ret_code)
+               if (dma_mem_alloc_done)
+                       i40iw_free_dma_mem(hw, &mem);
+
+       return ret_code;
+}
+
+/**
+ * i40iw_add_pd_table_entry - Adds page descriptor to the specified table
+ * @hw: pointer to our HW structure
+ * @hmc_info: pointer to the HMC configuration information structure
+ * @pd_index: which page descriptor index to manipulate
+ * @rsrc_pg: if not NULL, use preallocated page instead of allocating new one.
+ *
+ * This function:
+ *     1. Initializes the pd entry
+ *     2. Adds pd_entry in the pd_table
+ *     3. Mark the entry valid in i40iw_hmc_pd_entry structure
+ *     4. Initializes the pd_entry's ref count to 1
+ * assumptions:
+ *     1. The memory for pd should be pinned down, physically contiguous and
+ *        aligned on 4K boundary and zeroed memory.
+ *     2. It should be 4K in size.
+ */
+enum i40iw_status_code i40iw_add_pd_table_entry(struct i40iw_hw *hw,
+                                               struct i40iw_hmc_info *hmc_info,
+                                               u32 pd_index,
+                                               struct i40iw_dma_mem *rsrc_pg)
+{
+       enum i40iw_status_code ret_code = 0;
+       struct i40iw_hmc_pd_table *pd_table;
+       struct i40iw_hmc_pd_entry *pd_entry;
+       struct i40iw_dma_mem mem;
+       struct i40iw_dma_mem *page = &mem;
+       u32 sd_idx, rel_pd_idx;
+       u64 *pd_addr;
+       u64 page_desc;
+
+       if (pd_index / I40IW_HMC_PD_CNT_IN_SD >= hmc_info->sd_table.sd_cnt)
+               return I40IW_ERR_INVALID_PAGE_DESC_INDEX;
+
+       sd_idx = (pd_index / I40IW_HMC_PD_CNT_IN_SD);
+       if (hmc_info->sd_table.sd_entry[sd_idx].entry_type != I40IW_SD_TYPE_PAGED)
+               return 0;
+
+       rel_pd_idx = (pd_index % I40IW_HMC_PD_CNT_IN_SD);
+       pd_table = &hmc_info->sd_table.sd_entry[sd_idx].u.pd_table;
+       pd_entry = &pd_table->pd_entry[rel_pd_idx];
+       if (!pd_entry->valid) {
+               if (rsrc_pg) {
+                       pd_entry->rsrc_pg = true;
+                       page = rsrc_pg;
+               } else {
+                       ret_code = i40iw_allocate_dma_mem(hw, page,
+                                                         I40IW_HMC_PAGED_BP_SIZE,
+                                                         I40IW_HMC_PD_BP_BUF_ALIGNMENT);
+                       if (ret_code)
+                               return ret_code;
+                       pd_entry->rsrc_pg = false;
+               }
+
+               memcpy(&pd_entry->bp.addr, page, sizeof(struct i40iw_dma_mem));
+               pd_entry->bp.sd_pd_index = pd_index;
+               pd_entry->bp.entry_type = I40IW_SD_TYPE_PAGED;
+               page_desc = page->pa | 0x1;
+
+               pd_addr = (u64 *)pd_table->pd_page_addr.va;
+               pd_addr += rel_pd_idx;
+
+               memcpy(pd_addr, &page_desc, sizeof(*pd_addr));
+
+               pd_entry->sd_index = sd_idx;
+               pd_entry->valid = true;
+               I40IW_INC_PD_REFCNT(pd_table);
+               if (hmc_info->hmc_fn_id < I40IW_FIRST_VF_FPM_ID)
+                       I40IW_INVALIDATE_PF_HMC_PD(hw, sd_idx, rel_pd_idx);
+               else if (hw->hmc.hmc_fn_id != hmc_info->hmc_fn_id)
+                       I40IW_INVALIDATE_VF_HMC_PD(hw, sd_idx, rel_pd_idx,
+                                                  hmc_info->hmc_fn_id);
+       }
+       I40IW_INC_BP_REFCNT(&pd_entry->bp);
+
+       return 0;
+}
+
+/**
+ * i40iw_remove_pd_bp - remove a backing page from a page descriptor
+ * @hw: pointer to our HW structure
+ * @hmc_info: pointer to the HMC configuration information structure
+ * @idx: the page index
+ * @is_pf: distinguishes a VF from a PF
+ *
+ * This function:
+ *     1. Marks the entry in pd table (for paged address mode) or in sd table
+ *        (for direct address mode) invalid.
+ *     2. Write to register PMPDINV to invalidate the backing page in FV cache
+ *     3. Decrement the ref count for the pd _entry
+ * assumptions:
+ *     1. Caller can deallocate the memory used by backing storage after this
+ *        function returns.
+ */
+enum i40iw_status_code i40iw_remove_pd_bp(struct i40iw_hw *hw,
+                                         struct i40iw_hmc_info *hmc_info,
+                                         u32 idx,
+                                         bool is_pf)
+{
+       struct i40iw_hmc_pd_entry *pd_entry;
+       struct i40iw_hmc_pd_table *pd_table;
+       struct i40iw_hmc_sd_entry *sd_entry;
+       u32 sd_idx, rel_pd_idx;
+       struct i40iw_dma_mem *mem;
+       u64 *pd_addr;
+
+       sd_idx = idx / I40IW_HMC_PD_CNT_IN_SD;
+       rel_pd_idx = idx % I40IW_HMC_PD_CNT_IN_SD;
+       if (sd_idx >= hmc_info->sd_table.sd_cnt)
+               return I40IW_ERR_INVALID_PAGE_DESC_INDEX;
+
+       sd_entry = &hmc_info->sd_table.sd_entry[sd_idx];
+       if (sd_entry->entry_type != I40IW_SD_TYPE_PAGED)
+               return I40IW_ERR_INVALID_SD_TYPE;
+
+       pd_table = &hmc_info->sd_table.sd_entry[sd_idx].u.pd_table;
+       pd_entry = &pd_table->pd_entry[rel_pd_idx];
+       I40IW_DEC_BP_REFCNT(&pd_entry->bp);
+       if (pd_entry->bp.ref_cnt)
+               return 0;
+
+       pd_entry->valid = false;
+       I40IW_DEC_PD_REFCNT(pd_table);
+       pd_addr = (u64 *)pd_table->pd_page_addr.va;
+       pd_addr += rel_pd_idx;
+       memset(pd_addr, 0, sizeof(u64));
+       if (is_pf)
+               I40IW_INVALIDATE_PF_HMC_PD(hw, sd_idx, idx);
+       else
+               I40IW_INVALIDATE_VF_HMC_PD(hw, sd_idx, idx,
+                                          hmc_info->hmc_fn_id);
+
+       if (!pd_entry->rsrc_pg) {
+               mem = &pd_entry->bp.addr;
+               if (!mem || !mem->va)
+                       return I40IW_ERR_PARAM;
+               i40iw_free_dma_mem(hw, mem);
+       }
+       if (!pd_table->ref_cnt)
+               i40iw_free_virt_mem(hw, &pd_table->pd_entry_virt_mem);
+
+       return 0;
+}
+
+/**
+ * i40iw_prep_remove_sd_bp - Prepares to remove a backing page from a sd entry
+ * @hmc_info: pointer to the HMC configuration information structure
+ * @idx: the page index
+ */
+enum i40iw_status_code i40iw_prep_remove_sd_bp(struct i40iw_hmc_info *hmc_info, u32 idx)
+{
+       struct i40iw_hmc_sd_entry *sd_entry;
+
+       sd_entry = &hmc_info->sd_table.sd_entry[idx];
+       I40IW_DEC_BP_REFCNT(&sd_entry->u.bp);
+       if (sd_entry->u.bp.ref_cnt)
+               return I40IW_ERR_NOT_READY;
+
+       I40IW_DEC_SD_REFCNT(&hmc_info->sd_table);
+       sd_entry->valid = false;
+
+       return 0;
+}
+
+/**
+ * i40iw_prep_remove_pd_page - Prepares to remove a PD page from sd entry.
+ * @hmc_info: pointer to the HMC configuration information structure
+ * @idx: segment descriptor index to find the relevant page descriptor
+ */
+enum i40iw_status_code i40iw_prep_remove_pd_page(struct i40iw_hmc_info *hmc_info,
+                                                u32 idx)
+{
+       struct i40iw_hmc_sd_entry *sd_entry;
+
+       sd_entry = &hmc_info->sd_table.sd_entry[idx];
+
+       if (sd_entry->u.pd_table.ref_cnt)
+               return I40IW_ERR_NOT_READY;
+
+       sd_entry->valid = false;
+       I40IW_DEC_SD_REFCNT(&hmc_info->sd_table);
+
+       return 0;
+}
+
+/**
+ * i40iw_pf_init_vfhmc -
+ * @vf_cnt_array: array of cnt values of iwarp hmc objects
+ * @vf_hmc_fn_id: hmc function id ofr vf driver
+ * @dev: pointer to i40iw_dev struct
+ *
+ * Called by pf driver to initialize hmc_info for vf driver instance.
+ */
+enum i40iw_status_code i40iw_pf_init_vfhmc(struct i40iw_sc_dev *dev,
+                                          u8 vf_hmc_fn_id,
+                                          u32 *vf_cnt_array)
+{
+       struct i40iw_hmc_info *hmc_info;
+       enum i40iw_status_code ret_code = 0;
+       u32 i;
+
+       if ((vf_hmc_fn_id < I40IW_FIRST_VF_FPM_ID) ||
+           (vf_hmc_fn_id >= I40IW_FIRST_VF_FPM_ID +
+            I40IW_MAX_PE_ENABLED_VF_COUNT)) {
+               i40iw_debug(dev, I40IW_DEBUG_HMC, "%s: invalid vf_hmc_fn_id  0x%x\n",
+                           __func__, vf_hmc_fn_id);
+               return I40IW_ERR_INVALID_HMCFN_ID;
+       }
+
+       ret_code = i40iw_sc_init_iw_hmc(dev, vf_hmc_fn_id);
+       if (ret_code)
+               return ret_code;
+
+       hmc_info = i40iw_vf_hmcinfo_from_fpm(dev, vf_hmc_fn_id);
+
+       for (i = I40IW_HMC_IW_QP; i < I40IW_HMC_IW_MAX; i++)
+               if (vf_cnt_array)
+                       hmc_info->hmc_obj[i].cnt =
+                           vf_cnt_array[i - I40IW_HMC_IW_QP];
+               else
+                       hmc_info->hmc_obj[i].cnt = hmc_info->hmc_obj[i].max_cnt;
+
+       return 0;
+}
diff --git a/drivers/infiniband/hw/i40iw/i40iw_hmc.h b/drivers/infiniband/hw/i40iw/i40iw_hmc.h
new file mode 100644 (file)
index 0000000..4c3fdd8
--- /dev/null
@@ -0,0 +1,241 @@
+/*******************************************************************************
+*
+* Copyright (c) 2015-2016 Intel Corporation.  All rights reserved.
+*
+* This software is available to you under a choice of one of two
+* licenses.  You may choose to be licensed under the terms of the GNU
+* General Public License (GPL) Version 2, available from the file
+* COPYING in the main directory of this source tree, or the
+* OpenFabrics.org BSD license below:
+*
+*   Redistribution and use in source and binary forms, with or
+*   without modification, are permitted provided that the following
+*   conditions are met:
+*
+*    - Redistributions of source code must retain the above
+*      copyright notice, this list of conditions and the following
+*      disclaimer.
+*
+*    - Redistributions in binary form must reproduce the above
+*      copyright notice, this list of conditions and the following
+*      disclaimer in the documentation and/or other materials
+*      provided with the distribution.
+*
+* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
+* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
+* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
+* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
+* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
+* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
+* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
+* SOFTWARE.
+*
+*******************************************************************************/
+
+#ifndef I40IW_HMC_H
+#define I40IW_HMC_H
+
+#include "i40iw_d.h"
+
+struct i40iw_hw;
+enum i40iw_status_code;
+
+#define I40IW_HMC_MAX_BP_COUNT 512
+#define I40IW_MAX_SD_ENTRIES 11
+#define I40IW_HW_DBG_HMC_INVALID_BP_MARK     0xCA
+
+#define I40IW_HMC_INFO_SIGNATURE       0x484D5347
+#define I40IW_HMC_PD_CNT_IN_SD         512
+#define I40IW_HMC_DIRECT_BP_SIZE       0x200000
+#define I40IW_HMC_MAX_SD_COUNT         4096
+#define I40IW_HMC_PAGED_BP_SIZE                4096
+#define I40IW_HMC_PD_BP_BUF_ALIGNMENT  4096
+#define I40IW_FIRST_VF_FPM_ID          16
+#define FPM_MULTIPLIER                 1024
+
+#define I40IW_INC_SD_REFCNT(sd_table)   ((sd_table)->ref_cnt++)
+#define I40IW_INC_PD_REFCNT(pd_table)   ((pd_table)->ref_cnt++)
+#define I40IW_INC_BP_REFCNT(bp)         ((bp)->ref_cnt++)
+
+#define I40IW_DEC_SD_REFCNT(sd_table)   ((sd_table)->ref_cnt--)
+#define I40IW_DEC_PD_REFCNT(pd_table)   ((pd_table)->ref_cnt--)
+#define I40IW_DEC_BP_REFCNT(bp)         ((bp)->ref_cnt--)
+
+/**
+ * I40IW_INVALIDATE_PF_HMC_PD - Invalidates the pd cache in the hardware
+ * @hw: pointer to our hw struct
+ * @sd_idx: segment descriptor index
+ * @pd_idx: page descriptor index
+ */
+#define I40IW_INVALIDATE_PF_HMC_PD(hw, sd_idx, pd_idx)                  \
+       i40iw_wr32((hw), I40E_PFHMC_PDINV,                                    \
+               (((sd_idx) << I40E_PFHMC_PDINV_PMSDIDX_SHIFT) |             \
+               (0x1 << I40E_PFHMC_PDINV_PMSDPARTSEL_SHIFT) | \
+               ((pd_idx) << I40E_PFHMC_PDINV_PMPDIDX_SHIFT)))
+
+/**
+ * I40IW_INVALIDATE_VF_HMC_PD - Invalidates the pd cache in the hardware
+ * @hw: pointer to our hw struct
+ * @sd_idx: segment descriptor index
+ * @pd_idx: page descriptor index
+ * @hmc_fn_id: VF's function id
+ */
+#define I40IW_INVALIDATE_VF_HMC_PD(hw, sd_idx, pd_idx, hmc_fn_id)        \
+       i40iw_wr32(hw, I40E_GLHMC_VFPDINV(hmc_fn_id - I40IW_FIRST_VF_FPM_ID),  \
+            ((sd_idx << I40E_PFHMC_PDINV_PMSDIDX_SHIFT) |              \
+             (pd_idx << I40E_PFHMC_PDINV_PMPDIDX_SHIFT)))
+
+struct i40iw_hmc_obj_info {
+       u64 base;
+       u32 max_cnt;
+       u32 cnt;
+       u64 size;
+};
+
+enum i40iw_sd_entry_type {
+       I40IW_SD_TYPE_INVALID = 0,
+       I40IW_SD_TYPE_PAGED = 1,
+       I40IW_SD_TYPE_DIRECT = 2
+};
+
+struct i40iw_hmc_bp {
+       enum i40iw_sd_entry_type entry_type;
+       struct i40iw_dma_mem addr;
+       u32 sd_pd_index;
+       u32 ref_cnt;
+};
+
+struct i40iw_hmc_pd_entry {
+       struct i40iw_hmc_bp bp;
+       u32 sd_index;
+       bool rsrc_pg;
+       bool valid;
+};
+
+struct i40iw_hmc_pd_table {
+       struct i40iw_dma_mem pd_page_addr;
+       struct i40iw_hmc_pd_entry *pd_entry;
+       struct i40iw_virt_mem pd_entry_virt_mem;
+       u32 ref_cnt;
+       u32 sd_index;
+};
+
+struct i40iw_hmc_sd_entry {
+       enum i40iw_sd_entry_type entry_type;
+       bool valid;
+
+       union {
+               struct i40iw_hmc_pd_table pd_table;
+               struct i40iw_hmc_bp bp;
+       } u;
+};
+
+struct i40iw_hmc_sd_table {
+       struct i40iw_virt_mem addr;
+       u32 sd_cnt;
+       u32 ref_cnt;
+       struct i40iw_hmc_sd_entry *sd_entry;
+};
+
+struct i40iw_hmc_info {
+       u32 signature;
+       u8 hmc_fn_id;
+       u16 first_sd_index;
+
+       struct i40iw_hmc_obj_info *hmc_obj;
+       struct i40iw_virt_mem hmc_obj_virt_mem;
+       struct i40iw_hmc_sd_table sd_table;
+       u16 sd_indexes[I40IW_HMC_MAX_SD_COUNT];
+};
+
+struct update_sd_entry {
+       u64 cmd;
+       u64 data;
+};
+
+struct i40iw_update_sds_info {
+       u32 cnt;
+       u8 hmc_fn_id;
+       struct update_sd_entry entry[I40IW_MAX_SD_ENTRIES];
+};
+
+struct i40iw_ccq_cqe_info;
+struct i40iw_hmc_fcn_info {
+       void (*callback_fcn)(struct i40iw_sc_dev *, void *,
+                            struct i40iw_ccq_cqe_info *);
+       void *cqp_callback_param;
+       u32 vf_id;
+       u16 iw_vf_idx;
+       bool free_fcn;
+};
+
+enum i40iw_hmc_rsrc_type {
+       I40IW_HMC_IW_QP = 0,
+       I40IW_HMC_IW_CQ = 1,
+       I40IW_HMC_IW_SRQ = 2,
+       I40IW_HMC_IW_HTE = 3,
+       I40IW_HMC_IW_ARP = 4,
+       I40IW_HMC_IW_APBVT_ENTRY = 5,
+       I40IW_HMC_IW_MR = 6,
+       I40IW_HMC_IW_XF = 7,
+       I40IW_HMC_IW_XFFL = 8,
+       I40IW_HMC_IW_Q1 = 9,
+       I40IW_HMC_IW_Q1FL = 10,
+       I40IW_HMC_IW_TIMER = 11,
+       I40IW_HMC_IW_FSIMC = 12,
+       I40IW_HMC_IW_FSIAV = 13,
+       I40IW_HMC_IW_PBLE = 14,
+       I40IW_HMC_IW_MAX = 15,
+};
+
+struct i40iw_hmc_create_obj_info {
+       struct i40iw_hmc_info *hmc_info;
+       struct i40iw_virt_mem add_sd_virt_mem;
+       u32 rsrc_type;
+       u32 start_idx;
+       u32 count;
+       u32 add_sd_cnt;
+       enum i40iw_sd_entry_type entry_type;
+       bool is_pf;
+};
+
+struct i40iw_hmc_del_obj_info {
+       struct i40iw_hmc_info *hmc_info;
+       struct i40iw_virt_mem del_sd_virt_mem;
+       u32 rsrc_type;
+       u32 start_idx;
+       u32 count;
+       u32 del_sd_cnt;
+       bool is_pf;
+};
+
+enum i40iw_status_code i40iw_copy_dma_mem(struct i40iw_hw *hw, void *dest_buf,
+                                         struct i40iw_dma_mem *src_mem, u64 src_offset, u64 size);
+enum i40iw_status_code i40iw_sc_create_hmc_obj(struct i40iw_sc_dev *dev,
+                                              struct i40iw_hmc_create_obj_info *info);
+enum i40iw_status_code i40iw_sc_del_hmc_obj(struct i40iw_sc_dev *dev,
+                                           struct i40iw_hmc_del_obj_info *info,
+                                           bool reset);
+enum i40iw_status_code i40iw_hmc_sd_one(struct i40iw_sc_dev *dev, u8 hmc_fn_id,
+                                       u64 pa, u32 sd_idx, enum i40iw_sd_entry_type type,
+                                       bool setsd);
+enum i40iw_status_code i40iw_update_sds_noccq(struct i40iw_sc_dev *dev,
+                                             struct i40iw_update_sds_info *info);
+struct i40iw_vfdev *i40iw_vfdev_from_fpm(struct i40iw_sc_dev *dev, u8 hmc_fn_id);
+struct i40iw_hmc_info *i40iw_vf_hmcinfo_from_fpm(struct i40iw_sc_dev *dev,
+                                                u8 hmc_fn_id);
+enum i40iw_status_code i40iw_add_sd_table_entry(struct i40iw_hw *hw,
+                                               struct i40iw_hmc_info *hmc_info, u32 sd_index,
+                                               enum i40iw_sd_entry_type type, u64 direct_mode_sz);
+enum i40iw_status_code i40iw_add_pd_table_entry(struct i40iw_hw *hw,
+                                               struct i40iw_hmc_info *hmc_info, u32 pd_index,
+                                               struct i40iw_dma_mem *rsrc_pg);
+enum i40iw_status_code i40iw_remove_pd_bp(struct i40iw_hw *hw,
+                                         struct i40iw_hmc_info *hmc_info, u32 idx, bool is_pf);
+enum i40iw_status_code i40iw_prep_remove_sd_bp(struct i40iw_hmc_info *hmc_info, u32 idx);
+enum i40iw_status_code i40iw_prep_remove_pd_page(struct i40iw_hmc_info *hmc_info, u32 idx);
+
+#define     ENTER_SHARED_FUNCTION()
+#define     EXIT_SHARED_FUNCTION()
+
+#endif                         /* I40IW_HMC_H */