1 | /* SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB */ |
2 | /* |
3 | * Copyright (c) 2016 Mellanox Technologies Ltd. All rights reserved. |
4 | * Copyright (c) 2015 System Fabric Works, Inc. All rights reserved. |
5 | */ |
6 | |
7 | #ifndef RXE_H |
8 | #define RXE_H |
9 | |
10 | #ifdef pr_fmt |
11 | #undef pr_fmt |
12 | #endif |
13 | #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt |
14 | |
15 | #include <linux/skbuff.h> |
16 | |
17 | #include <rdma/ib_verbs.h> |
18 | #include <rdma/ib_user_verbs.h> |
19 | #include <rdma/ib_pack.h> |
20 | #include <rdma/ib_smi.h> |
21 | #include <rdma/ib_umem.h> |
22 | #include <rdma/ib_cache.h> |
23 | #include <rdma/ib_addr.h> |
24 | #include <crypto/hash.h> |
25 | |
26 | #include "rxe_net.h" |
27 | #include "rxe_opcode.h" |
28 | #include "rxe_hdr.h" |
29 | #include "rxe_param.h" |
30 | #include "rxe_verbs.h" |
31 | #include "rxe_loc.h" |
32 | |
33 | /* |
34 | * Version 1 and Version 2 are identical on 64 bit machines, but on 32 bit |
35 | * machines Version 2 has a different struct layout. |
36 | */ |
37 | #define RXE_UVERBS_ABI_VERSION 2 |
38 | |
39 | #define RXE_ROCE_V2_SPORT (0xc000) |
40 | |
41 | #define rxe_dbg(fmt, ...) pr_debug("%s: " fmt "\n", __func__, ##__VA_ARGS__) |
42 | #define rxe_dbg_dev(rxe, fmt, ...) ibdev_dbg(&(rxe)->ib_dev, \ |
43 | "%s: " fmt, __func__, ##__VA_ARGS__) |
44 | #define rxe_dbg_uc(uc, fmt, ...) ibdev_dbg((uc)->ibuc.device, \ |
45 | "uc#%d %s: " fmt, (uc)->elem.index, __func__, ##__VA_ARGS__) |
46 | #define rxe_dbg_pd(pd, fmt, ...) ibdev_dbg((pd)->ibpd.device, \ |
47 | "pd#%d %s: " fmt, (pd)->elem.index, __func__, ##__VA_ARGS__) |
48 | #define rxe_dbg_ah(ah, fmt, ...) ibdev_dbg((ah)->ibah.device, \ |
49 | "ah#%d %s: " fmt, (ah)->elem.index, __func__, ##__VA_ARGS__) |
50 | #define rxe_dbg_srq(srq, fmt, ...) ibdev_dbg((srq)->ibsrq.device, \ |
51 | "srq#%d %s: " fmt, (srq)->elem.index, __func__, ##__VA_ARGS__) |
52 | #define rxe_dbg_qp(qp, fmt, ...) ibdev_dbg((qp)->ibqp.device, \ |
53 | "qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__) |
54 | #define rxe_dbg_cq(cq, fmt, ...) ibdev_dbg((cq)->ibcq.device, \ |
55 | "cq#%d %s: " fmt, (cq)->elem.index, __func__, ##__VA_ARGS__) |
56 | #define rxe_dbg_mr(mr, fmt, ...) ibdev_dbg((mr)->ibmr.device, \ |
57 | "mr#%d %s: " fmt, (mr)->elem.index, __func__, ##__VA_ARGS__) |
58 | #define rxe_dbg_mw(mw, fmt, ...) ibdev_dbg((mw)->ibmw.device, \ |
59 | "mw#%d %s: " fmt, (mw)->elem.index, __func__, ##__VA_ARGS__) |
60 | |
61 | #define rxe_err(fmt, ...) pr_err_ratelimited("%s: " fmt "\n", __func__, \ |
62 | ##__VA_ARGS__) |
63 | #define rxe_err_dev(rxe, fmt, ...) ibdev_err_ratelimited(&(rxe)->ib_dev, \ |
64 | "%s: " fmt, __func__, ##__VA_ARGS__) |
65 | #define rxe_err_uc(uc, fmt, ...) ibdev_err_ratelimited((uc)->ibuc.device, \ |
66 | "uc#%d %s: " fmt, (uc)->elem.index, __func__, ##__VA_ARGS__) |
67 | #define rxe_err_pd(pd, fmt, ...) ibdev_err_ratelimited((pd)->ibpd.device, \ |
68 | "pd#%d %s: " fmt, (pd)->elem.index, __func__, ##__VA_ARGS__) |
69 | #define rxe_err_ah(ah, fmt, ...) ibdev_err_ratelimited((ah)->ibah.device, \ |
70 | "ah#%d %s: " fmt, (ah)->elem.index, __func__, ##__VA_ARGS__) |
71 | #define rxe_err_srq(srq, fmt, ...) ibdev_err_ratelimited((srq)->ibsrq.device, \ |
72 | "srq#%d %s: " fmt, (srq)->elem.index, __func__, ##__VA_ARGS__) |
73 | #define rxe_err_qp(qp, fmt, ...) ibdev_err_ratelimited((qp)->ibqp.device, \ |
74 | "qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__) |
75 | #define rxe_err_cq(cq, fmt, ...) ibdev_err_ratelimited((cq)->ibcq.device, \ |
76 | "cq#%d %s: " fmt, (cq)->elem.index, __func__, ##__VA_ARGS__) |
77 | #define rxe_err_mr(mr, fmt, ...) ibdev_err_ratelimited((mr)->ibmr.device, \ |
78 | "mr#%d %s: " fmt, (mr)->elem.index, __func__, ##__VA_ARGS__) |
79 | #define rxe_err_mw(mw, fmt, ...) ibdev_err_ratelimited((mw)->ibmw.device, \ |
80 | "mw#%d %s: " fmt, (mw)->elem.index, __func__, ##__VA_ARGS__) |
81 | |
82 | #define rxe_info(fmt, ...) pr_info_ratelimited("%s: " fmt "\n", __func__, \ |
83 | ##__VA_ARGS__) |
84 | #define rxe_info_dev(rxe, fmt, ...) ibdev_info_ratelimited(&(rxe)->ib_dev, \ |
85 | "%s: " fmt, __func__, ##__VA_ARGS__) |
86 | #define rxe_info_uc(uc, fmt, ...) ibdev_info_ratelimited((uc)->ibuc.device, \ |
87 | "uc#%d %s: " fmt, (uc)->elem.index, __func__, ##__VA_ARGS__) |
88 | #define rxe_info_pd(pd, fmt, ...) ibdev_info_ratelimited((pd)->ibpd.device, \ |
89 | "pd#%d %s: " fmt, (pd)->elem.index, __func__, ##__VA_ARGS__) |
90 | #define rxe_info_ah(ah, fmt, ...) ibdev_info_ratelimited((ah)->ibah.device, \ |
91 | "ah#%d %s: " fmt, (ah)->elem.index, __func__, ##__VA_ARGS__) |
92 | #define rxe_info_srq(srq, fmt, ...) ibdev_info_ratelimited((srq)->ibsrq.device, \ |
93 | "srq#%d %s: " fmt, (srq)->elem.index, __func__, ##__VA_ARGS__) |
94 | #define rxe_info_qp(qp, fmt, ...) ibdev_info_ratelimited((qp)->ibqp.device, \ |
95 | "qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__) |
96 | #define rxe_info_cq(cq, fmt, ...) ibdev_info_ratelimited((cq)->ibcq.device, \ |
97 | "cq#%d %s: " fmt, (cq)->elem.index, __func__, ##__VA_ARGS__) |
98 | #define rxe_info_mr(mr, fmt, ...) ibdev_info_ratelimited((mr)->ibmr.device, \ |
99 | "mr#%d %s: " fmt, (mr)->elem.index, __func__, ##__VA_ARGS__) |
100 | #define rxe_info_mw(mw, fmt, ...) ibdev_info_ratelimited((mw)->ibmw.device, \ |
101 | "mw#%d %s: " fmt, (mw)->elem.index, __func__, ##__VA_ARGS__) |
102 | |
103 | /* responder states */ |
104 | enum resp_states { |
105 | RESPST_NONE, |
106 | RESPST_GET_REQ, |
107 | RESPST_CHK_PSN, |
108 | RESPST_CHK_OP_SEQ, |
109 | RESPST_CHK_OP_VALID, |
110 | RESPST_CHK_RESOURCE, |
111 | RESPST_CHK_LENGTH, |
112 | RESPST_CHK_RKEY, |
113 | RESPST_EXECUTE, |
114 | RESPST_READ_REPLY, |
115 | RESPST_ATOMIC_REPLY, |
116 | RESPST_ATOMIC_WRITE_REPLY, |
117 | RESPST_PROCESS_FLUSH, |
118 | RESPST_COMPLETE, |
119 | RESPST_ACKNOWLEDGE, |
120 | RESPST_CLEANUP, |
121 | RESPST_DUPLICATE_REQUEST, |
122 | RESPST_ERR_MALFORMED_WQE, |
123 | RESPST_ERR_UNSUPPORTED_OPCODE, |
124 | RESPST_ERR_MISALIGNED_ATOMIC, |
125 | RESPST_ERR_PSN_OUT_OF_SEQ, |
126 | RESPST_ERR_MISSING_OPCODE_FIRST, |
127 | RESPST_ERR_MISSING_OPCODE_LAST_C, |
128 | RESPST_ERR_MISSING_OPCODE_LAST_D1E, |
129 | RESPST_ERR_TOO_MANY_RDMA_ATM_REQ, |
130 | RESPST_ERR_RNR, |
131 | RESPST_ERR_RKEY_VIOLATION, |
132 | RESPST_ERR_INVALIDATE_RKEY, |
133 | RESPST_ERR_LENGTH, |
134 | RESPST_ERR_CQ_OVERFLOW, |
135 | RESPST_ERROR, |
136 | RESPST_DONE, |
137 | RESPST_EXIT, |
138 | }; |
139 | |
140 | void rxe_set_mtu(struct rxe_dev *rxe, unsigned int dev_mtu); |
141 | |
142 | int rxe_add(struct rxe_dev *rxe, unsigned int mtu, const char *ibdev_name); |
143 | |
144 | void rxe_rcv(struct sk_buff *skb); |
145 | |
146 | /* The caller must do a matching ib_device_put(&dev->ib_dev) */ |
147 | static inline struct rxe_dev *rxe_get_dev_from_net(struct net_device *ndev) |
148 | { |
149 | struct ib_device *ibdev = |
150 | ib_device_get_by_netdev(ndev, driver_id: RDMA_DRIVER_RXE); |
151 | |
152 | if (!ibdev) |
153 | return NULL; |
154 | return container_of(ibdev, struct rxe_dev, ib_dev); |
155 | } |
156 | |
157 | void rxe_port_up(struct rxe_dev *rxe); |
158 | void rxe_port_down(struct rxe_dev *rxe); |
159 | void rxe_set_port_state(struct rxe_dev *rxe); |
160 | |
161 | #endif /* RXE_H */ |
162 | |