SET_DEVICE_OP(dev_ops, reg_user_mr_dmabuf);
SET_DEVICE_OP(dev_ops, req_notify_cq);
SET_DEVICE_OP(dev_ops, rereg_user_mr);
- SET_DEVICE_OP(dev_ops, resize_cq);
+ SET_DEVICE_OP(dev_ops, resize_user_cq);
SET_DEVICE_OP(dev_ops, set_vf_guid);
SET_DEVICE_OP(dev_ops, set_vf_link_state);
SET_DEVICE_OP(dev_ops, ufile_hw_cleanup);
if (IS_ERR(cq))
return PTR_ERR(cq);
- ret = cq->device->ops.resize_cq(cq, cmd.cqe, &attrs->driver_udata);
+ ret = cq->device->ops.resize_user_cq(cq, cmd.cqe, &attrs->driver_udata);
if (ret)
goto out;
UAPI_DEF_WRITE_UDATA_IO(
struct ib_uverbs_resize_cq,
struct ib_uverbs_resize_cq_resp),
- UAPI_DEF_METHOD_NEEDS_FN(resize_cq)),
+ UAPI_DEF_METHOD_NEEDS_FN(resize_user_cq)),
DECLARE_UVERBS_WRITE_EX(
IB_USER_VERBS_EX_CMD_CREATE_CQ,
ib_uverbs_ex_create_cq,
.reg_user_mr = bnxt_re_reg_user_mr,
.reg_user_mr_dmabuf = bnxt_re_reg_user_mr_dmabuf,
.req_notify_cq = bnxt_re_req_notify_cq,
- .resize_cq = bnxt_re_resize_cq,
+ .resize_user_cq = bnxt_re_resize_cq,
.create_flow = bnxt_re_create_flow,
.destroy_flow = bnxt_re_destroy_flow,
INIT_RDMA_OBJ_SIZE(ib_ah, bnxt_re_ah, ib_ah),
.reg_user_mr_dmabuf = irdma_reg_user_mr_dmabuf,
.rereg_user_mr = irdma_rereg_user_mr,
.req_notify_cq = irdma_req_notify_cq,
- .resize_cq = irdma_resize_cq,
+ .resize_user_cq = irdma_resize_cq,
INIT_RDMA_OBJ_SIZE(ib_pd, irdma_pd, ibpd),
INIT_RDMA_OBJ_SIZE(ib_ucontext, irdma_ucontext, ibucontext),
INIT_RDMA_OBJ_SIZE(ib_ah, irdma_ah, ibah),
.reg_user_mr = mlx4_ib_reg_user_mr,
.req_notify_cq = mlx4_ib_arm_cq,
.rereg_user_mr = mlx4_ib_rereg_user_mr,
- .resize_cq = mlx4_ib_resize_cq,
+ .resize_user_cq = mlx4_ib_resize_cq,
.report_port_event = mlx4_ib_port_event,
INIT_RDMA_OBJ_SIZE(ib_ah, mlx4_ib_ah, ibah),
.reg_user_mr_dmabuf = mlx5_ib_reg_user_mr_dmabuf,
.req_notify_cq = mlx5_ib_arm_cq,
.rereg_user_mr = mlx5_ib_rereg_user_mr,
- .resize_cq = mlx5_ib_resize_cq,
+ .resize_user_cq = mlx5_ib_resize_cq,
.ufile_hw_cleanup = mlx5_ib_ufile_hw_cleanup,
INIT_RDMA_OBJ_SIZE(ib_ah, mlx5_ib_ah, ibah),
.query_port = mthca_query_port,
.query_qp = mthca_query_qp,
.reg_user_mr = mthca_reg_user_mr,
- .resize_cq = mthca_resize_cq,
+ .resize_user_cq = mthca_resize_cq,
INIT_RDMA_OBJ_SIZE(ib_ah, mthca_ah, ibah),
INIT_RDMA_OBJ_SIZE(ib_cq, mthca_cq, ibcq),
.query_qp = ocrdma_query_qp,
.reg_user_mr = ocrdma_reg_user_mr,
.req_notify_cq = ocrdma_arm_cq,
- .resize_cq = ocrdma_resize_cq,
+ .resize_user_cq = ocrdma_resize_cq,
INIT_RDMA_OBJ_SIZE(ib_ah, ocrdma_ah, ibah),
INIT_RDMA_OBJ_SIZE(ib_cq, ocrdma_cq, ibcq),
.query_srq = rvt_query_srq,
.reg_user_mr = rvt_reg_user_mr,
.req_notify_cq = rvt_req_notify_cq,
- .resize_cq = rvt_resize_cq,
+ .resize_user_cq = rvt_resize_cq,
INIT_RDMA_OBJ_SIZE(ib_ah, rvt_ah, ibah),
INIT_RDMA_OBJ_SIZE(ib_cq, rvt_cq, ibcq),
.reg_user_mr = rxe_reg_user_mr,
.req_notify_cq = rxe_req_notify_cq,
.rereg_user_mr = rxe_rereg_user_mr,
- .resize_cq = rxe_resize_cq,
+ .resize_user_cq = rxe_resize_cq,
INIT_RDMA_OBJ_SIZE(ib_ah, rxe_ah, ibah),
INIT_RDMA_OBJ_SIZE(ib_cq, rxe_cq, ibcq),
struct uverbs_attr_bundle *attrs);
int (*modify_cq)(struct ib_cq *cq, u16 cq_count, u16 cq_period);
int (*destroy_cq)(struct ib_cq *cq, struct ib_udata *udata);
- int (*resize_cq)(struct ib_cq *cq, int cqe, struct ib_udata *udata);
+ int (*resize_user_cq)(struct ib_cq *cq, int cqe,
+ struct ib_udata *udata);
/*
* pre_destroy_cq - Prevent a cq from generating any new work
* completions, but not free any kernel resources