1 /* SPDX-License-Identifier: GPL-2.0 */ 2 /* 3 * Shared Memory Communications over RDMA (SMC-R) and RoCE 4 * 5 * Connection Data Control (CDC) 6 * 7 * Copyright IBM Corp. 2016 8 * 9 * Author(s): Ursula Braun <ubraun@linux.vnet.ibm.com> 10 */ 11 12 #ifndef SMC_CDC_H 13 #define SMC_CDC_H 14 15 #include <linux/kernel.h> /* max_t */ 16 #include <linux/atomic.h> 17 #include <linux/in.h> 18 #include <linux/compiler.h> 19 20 #include "smc.h" 21 #include "smc_core.h" 22 #include "smc_wr.h" 23 24 #define SMC_CDC_MSG_TYPE 0xFE 25 26 /* in network byte order */ 27 union smc_cdc_cursor { /* SMC cursor */ 28 struct { 29 __be16 reserved; 30 __be16 wrap; 31 __be32 count; 32 }; 33 #ifdef KERNEL_HAS_ATOMIC64 34 atomic64_t acurs; /* for atomic processing */ 35 #else 36 u64 acurs; /* for atomic processing */ 37 #endif 38 } __aligned(8); 39 40 /* in network byte order */ 41 struct smc_cdc_msg { 42 struct smc_wr_rx_hdr common; /* .type = 0xFE */ 43 u8 len; /* 44 */ 44 __be16 seqno; 45 __be32 token; 46 union smc_cdc_cursor prod; 47 union smc_cdc_cursor cons; /* piggy backed "ack" */ 48 struct smc_cdc_producer_flags prod_flags; 49 struct smc_cdc_conn_state_flags conn_state_flags; 50 u8 reserved[18]; 51 } __packed; /* format defined in RFC7609 */ 52 53 /* CDC message for SMC-D */ 54 struct smcd_cdc_msg { 55 struct smc_wr_rx_hdr common; /* Type = 0xFE */ 56 u8 res1[7]; 57 u16 prod_wrap; 58 u32 prod_count; 59 u8 res2[2]; 60 u16 cons_wrap; 61 u32 cons_count; 62 struct smc_cdc_producer_flags prod_flags; 63 struct smc_cdc_conn_state_flags conn_state_flags; 64 u8 res3[8]; 65 } __packed; 66 67 static inline bool smc_cdc_rxed_any_close(struct smc_connection *conn) 68 { 69 return conn->local_rx_ctrl.conn_state_flags.peer_conn_abort || 70 conn->local_rx_ctrl.conn_state_flags.peer_conn_closed; 71 } 72 73 static inline bool smc_cdc_rxed_any_close_or_senddone( 74 struct smc_connection *conn) 75 { 76 return smc_cdc_rxed_any_close(conn) || 77 conn->local_rx_ctrl.conn_state_flags.peer_done_writing; 78 } 79 80 static inline void smc_curs_add(int size, union smc_host_cursor *curs, 81 int value) 82 { 83 curs->count += value; 84 if (curs->count >= size) { 85 curs->wrap++; 86 curs->count -= size; 87 } 88 } 89 90 /* SMC cursors are 8 bytes long and require atomic reading and writing */ 91 static inline u64 smc_curs_read(union smc_host_cursor *curs, 92 struct smc_connection *conn) 93 { 94 #ifndef KERNEL_HAS_ATOMIC64 95 unsigned long flags; 96 u64 ret; 97 98 spin_lock_irqsave(&conn->acurs_lock, flags); 99 ret = curs->acurs; 100 spin_unlock_irqrestore(&conn->acurs_lock, flags); 101 return ret; 102 #else 103 return atomic64_read(&curs->acurs); 104 #endif 105 } 106 107 static inline u64 smc_curs_read_net(union smc_cdc_cursor *curs, 108 struct smc_connection *conn) 109 { 110 #ifndef KERNEL_HAS_ATOMIC64 111 unsigned long flags; 112 u64 ret; 113 114 spin_lock_irqsave(&conn->acurs_lock, flags); 115 ret = curs->acurs; 116 spin_unlock_irqrestore(&conn->acurs_lock, flags); 117 return ret; 118 #else 119 return atomic64_read(&curs->acurs); 120 #endif 121 } 122 123 static inline void smc_curs_write(union smc_host_cursor *curs, u64 val, 124 struct smc_connection *conn) 125 { 126 #ifndef KERNEL_HAS_ATOMIC64 127 unsigned long flags; 128 129 spin_lock_irqsave(&conn->acurs_lock, flags); 130 curs->acurs = val; 131 spin_unlock_irqrestore(&conn->acurs_lock, flags); 132 #else 133 atomic64_set(&curs->acurs, val); 134 #endif 135 } 136 137 static inline void smc_curs_write_net(union smc_cdc_cursor *curs, u64 val, 138 struct smc_connection *conn) 139 { 140 #ifndef KERNEL_HAS_ATOMIC64 141 unsigned long flags; 142 143 spin_lock_irqsave(&conn->acurs_lock, flags); 144 curs->acurs = val; 145 spin_unlock_irqrestore(&conn->acurs_lock, flags); 146 #else 147 atomic64_set(&curs->acurs, val); 148 #endif 149 } 150 151 /* calculate cursor difference between old and new, where old <= new */ 152 static inline int smc_curs_diff(unsigned int size, 153 union smc_host_cursor *old, 154 union smc_host_cursor *new) 155 { 156 if (old->wrap != new->wrap) 157 return max_t(int, 0, 158 ((size - old->count) + new->count)); 159 160 return max_t(int, 0, (new->count - old->count)); 161 } 162 163 /* calculate cursor difference between old and new - returns negative 164 * value in case old > new 165 */ 166 static inline int smc_curs_comp(unsigned int size, 167 union smc_host_cursor *old, 168 union smc_host_cursor *new) 169 { 170 if (old->wrap > new->wrap || 171 (old->wrap == new->wrap && old->count > new->count)) 172 return -smc_curs_diff(size, new, old); 173 return smc_curs_diff(size, old, new); 174 } 175 176 static inline void smc_host_cursor_to_cdc(union smc_cdc_cursor *peer, 177 union smc_host_cursor *local, 178 struct smc_connection *conn) 179 { 180 union smc_host_cursor temp; 181 182 smc_curs_write(&temp, smc_curs_read(local, conn), conn); 183 peer->count = htonl(temp.count); 184 peer->wrap = htons(temp.wrap); 185 /* peer->reserved = htons(0); must be ensured by caller */ 186 } 187 188 static inline void smc_host_msg_to_cdc(struct smc_cdc_msg *peer, 189 struct smc_host_cdc_msg *local, 190 struct smc_connection *conn) 191 { 192 peer->common.type = local->common.type; 193 peer->len = local->len; 194 peer->seqno = htons(local->seqno); 195 peer->token = htonl(local->token); 196 smc_host_cursor_to_cdc(&peer->prod, &local->prod, conn); 197 smc_host_cursor_to_cdc(&peer->cons, &local->cons, conn); 198 peer->prod_flags = local->prod_flags; 199 peer->conn_state_flags = local->conn_state_flags; 200 } 201 202 static inline void smc_cdc_cursor_to_host(union smc_host_cursor *local, 203 union smc_cdc_cursor *peer, 204 struct smc_connection *conn) 205 { 206 union smc_host_cursor temp, old; 207 union smc_cdc_cursor net; 208 209 smc_curs_write(&old, smc_curs_read(local, conn), conn); 210 smc_curs_write_net(&net, smc_curs_read_net(peer, conn), conn); 211 temp.count = ntohl(net.count); 212 temp.wrap = ntohs(net.wrap); 213 if ((old.wrap > temp.wrap) && temp.wrap) 214 return; 215 if ((old.wrap == temp.wrap) && 216 (old.count > temp.count)) 217 return; 218 smc_curs_write(local, smc_curs_read(&temp, conn), conn); 219 } 220 221 static inline void smcr_cdc_msg_to_host(struct smc_host_cdc_msg *local, 222 struct smc_cdc_msg *peer, 223 struct smc_connection *conn) 224 { 225 local->common.type = peer->common.type; 226 local->len = peer->len; 227 local->seqno = ntohs(peer->seqno); 228 local->token = ntohl(peer->token); 229 smc_cdc_cursor_to_host(&local->prod, &peer->prod, conn); 230 smc_cdc_cursor_to_host(&local->cons, &peer->cons, conn); 231 local->prod_flags = peer->prod_flags; 232 local->conn_state_flags = peer->conn_state_flags; 233 } 234 235 static inline void smcd_cdc_msg_to_host(struct smc_host_cdc_msg *local, 236 struct smcd_cdc_msg *peer) 237 { 238 local->prod.wrap = peer->prod_wrap; 239 local->prod.count = peer->prod_count; 240 local->cons.wrap = peer->cons_wrap; 241 local->cons.count = peer->cons_count; 242 local->prod_flags = peer->prod_flags; 243 local->conn_state_flags = peer->conn_state_flags; 244 } 245 246 static inline void smc_cdc_msg_to_host(struct smc_host_cdc_msg *local, 247 struct smc_cdc_msg *peer, 248 struct smc_connection *conn) 249 { 250 if (conn->lgr->is_smcd) 251 smcd_cdc_msg_to_host(local, (struct smcd_cdc_msg *)peer); 252 else 253 smcr_cdc_msg_to_host(local, peer, conn); 254 } 255 256 struct smc_cdc_tx_pend; 257 258 int smc_cdc_get_free_slot(struct smc_connection *conn, 259 struct smc_wr_buf **wr_buf, 260 struct smc_cdc_tx_pend **pend); 261 void smc_cdc_tx_dismiss_slots(struct smc_connection *conn); 262 int smc_cdc_msg_send(struct smc_connection *conn, struct smc_wr_buf *wr_buf, 263 struct smc_cdc_tx_pend *pend); 264 int smc_cdc_get_slot_and_msg_send(struct smc_connection *conn); 265 int smcd_cdc_msg_send(struct smc_connection *conn); 266 int smc_cdc_init(void) __init; 267 void smcd_cdc_rx_init(struct smc_connection *conn); 268 269 #endif /* SMC_CDC_H */ 270