&pd->qplib_pd))
atomic_dec(&rdev->stats.res.pd_count);
}
- return 0;
+ return ib_respond_empty_udata(udata);
}
int bnxt_re_alloc_pd(struct ib_pd *ibpd, struct ib_udata *udata)
if (active_ahs > rdev->stats.res.ah_watermark)
rdev->stats.res.ah_watermark = active_ahs;
- return 0;
+ return ib_respond_empty_udata(udata);
}
int bnxt_re_query_ah(struct ib_ah *ib_ah, struct rdma_ah_attr *ah_attr)
if (scq_nq != rcq_nq)
bnxt_re_synchronize_nq(rcq_nq);
- return 0;
+ return ib_respond_empty_udata(udata);
}
static u8 __from_ib_qp_type(enum ib_qp_type type)
bnxt_qplib_destroy_srq(&rdev->qplib_res, qplib_srq);
ib_umem_release(srq->umem);
atomic_dec(&rdev->stats.res.srq_count);
- return 0;
+ return ib_respond_empty_udata(udata);
}
static int bnxt_re_init_user_srq(struct bnxt_re_dev *rdev,
/* On success, update the shadow */
srq->srq_limit = srq_attr->srq_limit;
/* No need to Build and send response back to udata */
- return 0;
+ return ib_respond_empty_udata(udata);
default:
ibdev_err(&rdev->ibdev,
"Unsupported srq_attr_mask 0x%x", srq_attr_mask);
ibdev_err(&rdev->ibdev, "Failed to modify HW QP");
return rc;
}
- if (ib_qp->qp_type == IB_QPT_GSI && rdev->gsi_ctx.gsi_sqp)
+ if (ib_qp->qp_type == IB_QPT_GSI && rdev->gsi_ctx.gsi_sqp) {
rc = bnxt_re_modify_shadow_qp(rdev, qp, qp_attr_mask);
- return rc;
+ if (rc)
+ return rc;
+ }
+ return ib_respond_empty_udata(udata);
}
int bnxt_re_query_qp(struct ib_qp *ib_qp, struct ib_qp_attr *qp_attr,
atomic_dec(&rdev->stats.res.cq_count);
kfree(cq->cql);
- return 0;
+ return ib_respond_empty_udata(udata);
}
int bnxt_re_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr,
cq->ib_cq.cqe = cq->resize_cqe;
atomic_inc(&rdev->stats.res.resize_count);
- return 0;
+ return ib_respond_empty_udata(udata);
fail:
if (cq->resize_umem) {
kfree(mr);
atomic_dec(&rdev->stats.res.mr_count);
- return rc;
+ if (rc)
+ return rc;
+ return ib_respond_empty_udata(udata);
}
static int bnxt_re_set_page(struct ib_mr *ib_mr, u64 addr)