1 /* SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) */
3 * Copyright(c) 2018 Intel Corporation.
6 #ifndef HFI1_TID_RDMA_H
7 #define HFI1_TID_RDMA_H
9 #include <linux/circ_buf.h>
12 /* Add a convenience helper */
13 #define CIRC_ADD(val, add, size) (((val) + (add)) & ((size) - 1))
14 #define CIRC_NEXT(val, size) CIRC_ADD(val, 1, size)
15 #define CIRC_PREV(val, size) CIRC_ADD(val, -1, size)
17 #define TID_RDMA_MIN_SEGMENT_SIZE BIT(18) /* 256 KiB (for now) */
18 #define TID_RDMA_MAX_SEGMENT_SIZE BIT(18) /* 256 KiB (for now) */
19 #define TID_RDMA_MAX_PAGES (BIT(18) >> PAGE_SHIFT)
22 * Bit definitions for priv->s_flags.
23 * These bit flags overload the bit flags defined for the QP's s_flags.
24 * Due to the fact that these bit fields are used only for the QP priv
25 * s_flags, there are no collisions.
27 * HFI1_S_TID_WAIT_INTERLCK - QP is waiting for requester interlock
29 #define HFI1_R_TID_RSC_TIMER BIT(2)
30 /* BIT(4) reserved for RVT_S_ACK_PENDING. */
31 #define HFI1_S_TID_WAIT_INTERLCK BIT(5)
32 #define HFI1_R_TID_SW_PSN BIT(19)
35 * Unlike regular IB RDMA VERBS, which do not require an entry
36 * in the s_ack_queue, TID RDMA WRITE requests do because they
38 * Therefore, the s_ack_queue needs to be extended by a certain
39 * amount. The key point is that the queue needs to be extended
40 * without letting the "user" know so they user doesn't end up
41 * using these extra entries.
43 #define HFI1_TID_RDMA_WRITE_CNT 8
45 struct tid_rdma_params
{
46 struct rcu_head rcu_head
;
57 struct tid_rdma_qp_params
{
58 struct work_struct trigger_work
;
59 struct tid_rdma_params local
;
60 struct tid_rdma_params __rcu
*remote
;
63 /* Track state for each hardware flow */
64 struct tid_flow_state
{
67 u32 r_next_psn
; /* next PSN to be received (in TID space) */
73 enum tid_rdma_req_state
{
74 TID_REQUEST_INACTIVE
= 0,
76 TID_REQUEST_INIT_RESEND
,
79 TID_REQUEST_RESEND_ACTIVE
,
86 struct tid_rdma_request
{
88 struct hfi1_ctxtdata
*rcd
;
90 struct rvt_swqe
*swqe
;
91 struct rvt_ack_entry
*ack
;
94 struct tid_rdma_flow
*flows
; /* array of tid flows */
95 struct rvt_sge_state ss
; /* SGE state for TID RDMA requests */
96 u16 n_flows
; /* size of the flow buffer window */
97 u16 setup_head
; /* flow index we are setting up */
98 u16 clear_tail
; /* flow index we are clearing */
99 u16 flow_idx
; /* flow index most recently set up */
104 u32 r_ack_psn
; /* next expected ack PSN */
105 u32 r_flow_psn
; /* IB PSN of next segment start */
106 u32 r_last_acked
; /* IB PSN of last ACK'ed packet */
107 u32 s_next_psn
; /* IB PSN of next segment start for read */
109 u32 total_segs
; /* segments required to complete a request */
110 u32 cur_seg
; /* index of current segment */
111 u32 comp_seg
; /* index of last completed segment */
112 u32 ack_seg
; /* index of last ack'ed segment */
113 u32 alloc_seg
; /* index of next segment to be allocated */
114 u32 isge
; /* index of "current" sge */
115 u32 ack_pending
; /* num acks pending for this request */
117 enum tid_rdma_req_state state
;
121 * When header suppression is used, PSNs associated with a "flow" are
122 * relevant (and not the PSNs maintained by verbs). Track per-flow
123 * PSNs here for a TID RDMA segment.
128 u32 resp_ib_psn
; /* The IB PSN of the response for this flow */
129 u32 generation
; /* generation of flow */
130 u32 spsn
; /* starting PSN in TID space */
131 u32 lpsn
; /* last PSN in TID space */
132 u32 r_next_psn
; /* next PSN to be received (in TID space) */
134 /* For tid rdma read */
135 u32 ib_spsn
; /* starting PSN in Verbs space */
136 u32 ib_lpsn
; /* last PSn in Verbs space */
139 struct tid_rdma_pageset
{
140 dma_addr_t addr
: 48; /* Only needed for the first page */
147 * kern_tid_node - used for managing TID's in TID groups
149 * @grp_idx: rcd relative index to tid_group
150 * @map: grp->map captured prior to programming this TID group in HW
151 * @cnt: Only @cnt of available group entries are actually programmed
153 struct kern_tid_node
{
154 struct tid_group
*grp
;
159 /* Overall info for a TID RDMA segment */
160 struct tid_rdma_flow
{
162 * While a TID RDMA segment is being transferred, it uses a QP number
163 * from the "KDETH section of QP numbers" (which is different from the
164 * QP number that originated the request). Bits 11-15 of these QP
165 * numbers identify the "TID flow" for the segment.
167 struct flow_state flow_state
;
168 struct tid_rdma_request
*req
;
181 struct kern_tid_node tnode
[TID_RDMA_MAX_PAGES
];
182 struct tid_rdma_pageset pagesets
[TID_RDMA_MAX_PAGES
];
183 u32 tid_entry
[TID_RDMA_MAX_PAGES
];
186 enum tid_rnr_nak_state
{
187 TID_RNR_NAK_INIT
= 0,
192 bool tid_rdma_conn_req(struct rvt_qp
*qp
, u64
*data
);
193 bool tid_rdma_conn_reply(struct rvt_qp
*qp
, u64 data
);
194 bool tid_rdma_conn_resp(struct rvt_qp
*qp
, u64
*data
);
195 void tid_rdma_conn_error(struct rvt_qp
*qp
);
196 void tid_rdma_opfn_init(struct rvt_qp
*qp
, struct tid_rdma_params
*p
);
198 int hfi1_kern_exp_rcv_init(struct hfi1_ctxtdata
*rcd
, int reinit
);
199 int hfi1_kern_exp_rcv_setup(struct tid_rdma_request
*req
,
200 struct rvt_sge_state
*ss
, bool *last
);
201 int hfi1_kern_exp_rcv_clear(struct tid_rdma_request
*req
);
202 void hfi1_kern_exp_rcv_clear_all(struct tid_rdma_request
*req
);
203 void __trdma_clean_swqe(struct rvt_qp
*qp
, struct rvt_swqe
*wqe
);
206 * trdma_clean_swqe - clean flows for swqe if large send queue
210 static inline void trdma_clean_swqe(struct rvt_qp
*qp
, struct rvt_swqe
*wqe
)
214 __trdma_clean_swqe(qp
, wqe
);
217 void hfi1_kern_read_tid_flow_free(struct rvt_qp
*qp
);
219 int hfi1_qp_priv_init(struct rvt_dev_info
*rdi
, struct rvt_qp
*qp
,
220 struct ib_qp_init_attr
*init_attr
);
221 void hfi1_qp_priv_tid_free(struct rvt_dev_info
*rdi
, struct rvt_qp
*qp
);
223 void hfi1_tid_rdma_flush_wait(struct rvt_qp
*qp
);
225 int hfi1_kern_setup_hw_flow(struct hfi1_ctxtdata
*rcd
, struct rvt_qp
*qp
);
226 void hfi1_kern_clear_hw_flow(struct hfi1_ctxtdata
*rcd
, struct rvt_qp
*qp
);
227 void hfi1_kern_init_ctxt_generations(struct hfi1_ctxtdata
*rcd
);
230 u64
hfi1_access_sw_tid_wait(const struct cntr_entry
*entry
,
231 void *context
, int vl
, int mode
, u64 data
);
233 u32
hfi1_build_tid_rdma_read_packet(struct rvt_swqe
*wqe
,
234 struct ib_other_headers
*ohdr
,
235 u32
*bth1
, u32
*bth2
, u32
*len
);
236 u32
hfi1_build_tid_rdma_read_req(struct rvt_qp
*qp
, struct rvt_swqe
*wqe
,
237 struct ib_other_headers
*ohdr
, u32
*bth1
,
238 u32
*bth2
, u32
*len
);
239 void hfi1_rc_rcv_tid_rdma_read_req(struct hfi1_packet
*packet
);
240 u32
hfi1_build_tid_rdma_read_resp(struct rvt_qp
*qp
, struct rvt_ack_entry
*e
,
241 struct ib_other_headers
*ohdr
, u32
*bth0
,
242 u32
*bth1
, u32
*bth2
, u32
*len
, bool *last
);
243 void hfi1_rc_rcv_tid_rdma_read_resp(struct hfi1_packet
*packet
);
244 bool hfi1_handle_kdeth_eflags(struct hfi1_ctxtdata
*rcd
,
245 struct hfi1_pportdata
*ppd
,
246 struct hfi1_packet
*packet
);
247 void hfi1_tid_rdma_restart_req(struct rvt_qp
*qp
, struct rvt_swqe
*wqe
,
249 void hfi1_qp_kern_exp_rcv_clear_all(struct rvt_qp
*qp
);
250 bool hfi1_tid_rdma_wqe_interlock(struct rvt_qp
*qp
, struct rvt_swqe
*wqe
);
252 void setup_tid_rdma_wqe(struct rvt_qp
*qp
, struct rvt_swqe
*wqe
);
253 static inline void hfi1_setup_tid_rdma_wqe(struct rvt_qp
*qp
,
254 struct rvt_swqe
*wqe
)
257 wqe
->wr
.opcode
== IB_WR_RDMA_READ
&&
258 wqe
->length
>= TID_RDMA_MIN_SEGMENT_SIZE
)
259 setup_tid_rdma_wqe(qp
, wqe
);
262 u32
hfi1_build_tid_rdma_write_req(struct rvt_qp
*qp
, struct rvt_swqe
*wqe
,
263 struct ib_other_headers
*ohdr
,
264 u32
*bth1
, u32
*bth2
, u32
*len
);
266 void hfi1_compute_tid_rdma_flow_wt(void);
268 void hfi1_rc_rcv_tid_rdma_write_req(struct hfi1_packet
*packet
);
270 u32
hfi1_build_tid_rdma_write_resp(struct rvt_qp
*qp
, struct rvt_ack_entry
*e
,
271 struct ib_other_headers
*ohdr
, u32
*bth1
,
273 struct rvt_sge_state
**ss
);
275 void hfi1_del_tid_reap_timer(struct rvt_qp
*qp
);
277 void hfi1_rc_rcv_tid_rdma_write_resp(struct hfi1_packet
*packet
);
279 bool hfi1_build_tid_rdma_packet(struct rvt_swqe
*wqe
,
280 struct ib_other_headers
*ohdr
,
281 u32
*bth1
, u32
*bth2
, u32
*len
);
283 void hfi1_rc_rcv_tid_rdma_write_data(struct hfi1_packet
*packet
);
285 u32
hfi1_build_tid_rdma_write_ack(struct rvt_qp
*qp
, struct rvt_ack_entry
*e
,
286 struct ib_other_headers
*ohdr
, u16 iflow
,
287 u32
*bth1
, u32
*bth2
);
289 void hfi1_rc_rcv_tid_rdma_ack(struct hfi1_packet
*packet
);
291 #endif /* HFI1_TID_RDMA_H */