1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * Shared Memory Communications over RDMA (SMC-R) and RoCE 4 * 5 * Monitoring SMC transport protocol sockets 6 * 7 * Copyright IBM Corp. 2016 8 * 9 * Author(s): Ursula Braun <ubraun@linux.vnet.ibm.com> 10 */ 11 12 #include <linux/kernel.h> 13 #include <linux/module.h> 14 #include <linux/types.h> 15 #include <linux/init.h> 16 #include <linux/sock_diag.h> 17 #include <linux/inet_diag.h> 18 #include <linux/smc_diag.h> 19 #include <net/netlink.h> 20 #include <net/smc.h> 21 22 #include "smc.h" 23 #include "smc_core.h" 24 25 struct smc_diag_dump_ctx { 26 int pos[2]; 27 }; 28 29 static struct smc_diag_dump_ctx *smc_dump_context(struct netlink_callback *cb) 30 { 31 return (struct smc_diag_dump_ctx *)cb->ctx; 32 } 33 34 static void smc_diag_msg_common_fill(struct smc_diag_msg *r, struct sock *sk) 35 { 36 struct smc_sock *smc = smc_sk(sk); 37 38 memset(r, 0, sizeof(*r)); 39 r->diag_family = sk->sk_family; 40 sock_diag_save_cookie(sk, r->id.idiag_cookie); 41 if (!smc->clcsock) 42 return; 43 r->id.idiag_sport = htons(smc->clcsock->sk->sk_num); 44 r->id.idiag_dport = smc->clcsock->sk->sk_dport; 45 r->id.idiag_if = smc->clcsock->sk->sk_bound_dev_if; 46 if (sk->sk_protocol == SMCPROTO_SMC) { 47 r->id.idiag_src[0] = smc->clcsock->sk->sk_rcv_saddr; 48 r->id.idiag_dst[0] = smc->clcsock->sk->sk_daddr; 49 #if IS_ENABLED(CONFIG_IPV6) 50 } else if (sk->sk_protocol == SMCPROTO_SMC6) { 51 memcpy(&r->id.idiag_src, &smc->clcsock->sk->sk_v6_rcv_saddr, 52 sizeof(smc->clcsock->sk->sk_v6_rcv_saddr)); 53 memcpy(&r->id.idiag_dst, &smc->clcsock->sk->sk_v6_daddr, 54 sizeof(smc->clcsock->sk->sk_v6_daddr)); 55 #endif 56 } 57 } 58 59 static int smc_diag_msg_attrs_fill(struct sock *sk, struct sk_buff *skb, 60 struct smc_diag_msg *r, 61 struct user_namespace *user_ns) 62 { 63 if (nla_put_u8(skb, SMC_DIAG_SHUTDOWN, sk->sk_shutdown)) 64 return 1; 65 66 r->diag_uid = from_kuid_munged(user_ns, sock_i_uid(sk)); 67 r->diag_inode = sock_i_ino(sk); 68 return 0; 69 } 70 71 static int __smc_diag_dump(struct sock *sk, struct sk_buff *skb, 72 struct netlink_callback *cb, 73 const struct smc_diag_req *req, 74 struct nlattr *bc) 75 { 76 struct smc_sock *smc = smc_sk(sk); 77 struct smc_diag_fallback fallback; 78 struct user_namespace *user_ns; 79 struct smc_diag_msg *r; 80 struct nlmsghdr *nlh; 81 82 nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq, 83 cb->nlh->nlmsg_type, sizeof(*r), NLM_F_MULTI); 84 if (!nlh) 85 return -EMSGSIZE; 86 87 r = nlmsg_data(nlh); 88 smc_diag_msg_common_fill(r, sk); 89 r->diag_state = sk->sk_state; 90 if (smc->use_fallback) 91 r->diag_mode = SMC_DIAG_MODE_FALLBACK_TCP; 92 else if (smc_conn_lgr_valid(&smc->conn) && smc->conn.lgr->is_smcd) 93 r->diag_mode = SMC_DIAG_MODE_SMCD; 94 else 95 r->diag_mode = SMC_DIAG_MODE_SMCR; 96 user_ns = sk_user_ns(NETLINK_CB(cb->skb).sk); 97 if (smc_diag_msg_attrs_fill(sk, skb, r, user_ns)) 98 goto errout; 99 100 fallback.reason = smc->fallback_rsn; 101 fallback.peer_diagnosis = smc->peer_diagnosis; 102 if (nla_put(skb, SMC_DIAG_FALLBACK, sizeof(fallback), &fallback) < 0) 103 goto errout; 104 105 if ((req->diag_ext & (1 << (SMC_DIAG_CONNINFO - 1))) && 106 smc->conn.alert_token_local) { 107 struct smc_connection *conn = &smc->conn; 108 struct smc_diag_conninfo cinfo = { 109 .token = conn->alert_token_local, 110 .sndbuf_size = conn->sndbuf_desc ? 111 conn->sndbuf_desc->len : 0, 112 .rmbe_size = conn->rmb_desc ? conn->rmb_desc->len : 0, 113 .peer_rmbe_size = conn->peer_rmbe_size, 114 115 .rx_prod.wrap = conn->local_rx_ctrl.prod.wrap, 116 .rx_prod.count = conn->local_rx_ctrl.prod.count, 117 .rx_cons.wrap = conn->local_rx_ctrl.cons.wrap, 118 .rx_cons.count = conn->local_rx_ctrl.cons.count, 119 120 .tx_prod.wrap = conn->local_tx_ctrl.prod.wrap, 121 .tx_prod.count = conn->local_tx_ctrl.prod.count, 122 .tx_cons.wrap = conn->local_tx_ctrl.cons.wrap, 123 .tx_cons.count = conn->local_tx_ctrl.cons.count, 124 125 .tx_prod_flags = 126 *(u8 *)&conn->local_tx_ctrl.prod_flags, 127 .tx_conn_state_flags = 128 *(u8 *)&conn->local_tx_ctrl.conn_state_flags, 129 .rx_prod_flags = *(u8 *)&conn->local_rx_ctrl.prod_flags, 130 .rx_conn_state_flags = 131 *(u8 *)&conn->local_rx_ctrl.conn_state_flags, 132 133 .tx_prep.wrap = conn->tx_curs_prep.wrap, 134 .tx_prep.count = conn->tx_curs_prep.count, 135 .tx_sent.wrap = conn->tx_curs_sent.wrap, 136 .tx_sent.count = conn->tx_curs_sent.count, 137 .tx_fin.wrap = conn->tx_curs_fin.wrap, 138 .tx_fin.count = conn->tx_curs_fin.count, 139 }; 140 141 if (nla_put(skb, SMC_DIAG_CONNINFO, sizeof(cinfo), &cinfo) < 0) 142 goto errout; 143 } 144 145 if (smc_conn_lgr_valid(&smc->conn) && !smc->conn.lgr->is_smcd && 146 (req->diag_ext & (1 << (SMC_DIAG_LGRINFO - 1))) && 147 !list_empty(&smc->conn.lgr->list)) { 148 struct smc_link *link = smc->conn.lnk; 149 150 struct smc_diag_lgrinfo linfo = { 151 .role = smc->conn.lgr->role, 152 .lnk[0].ibport = link->ibport, 153 .lnk[0].link_id = link->link_id, 154 }; 155 156 memcpy(linfo.lnk[0].ibname, 157 smc->conn.lgr->lnk[0].smcibdev->ibdev->name, 158 sizeof(link->smcibdev->ibdev->name)); 159 smc_gid_be16_convert(linfo.lnk[0].gid, link->gid); 160 smc_gid_be16_convert(linfo.lnk[0].peer_gid, link->peer_gid); 161 162 if (nla_put(skb, SMC_DIAG_LGRINFO, sizeof(linfo), &linfo) < 0) 163 goto errout; 164 } 165 if (smc_conn_lgr_valid(&smc->conn) && smc->conn.lgr->is_smcd && 166 (req->diag_ext & (1 << (SMC_DIAG_DMBINFO - 1))) && 167 !list_empty(&smc->conn.lgr->list)) { 168 struct smc_connection *conn = &smc->conn; 169 struct smcd_diag_dmbinfo dinfo; 170 struct smcd_dev *smcd = conn->lgr->smcd; 171 172 memset(&dinfo, 0, sizeof(dinfo)); 173 174 dinfo.linkid = *((u32 *)conn->lgr->id); 175 dinfo.peer_gid = conn->lgr->peer_gid; 176 dinfo.my_gid = smcd->ops->get_local_gid(smcd); 177 dinfo.token = conn->rmb_desc->token; 178 dinfo.peer_token = conn->peer_token; 179 180 if (nla_put(skb, SMC_DIAG_DMBINFO, sizeof(dinfo), &dinfo) < 0) 181 goto errout; 182 } 183 184 nlmsg_end(skb, nlh); 185 return 0; 186 187 errout: 188 nlmsg_cancel(skb, nlh); 189 return -EMSGSIZE; 190 } 191 192 static int smc_diag_dump_proto(struct proto *prot, struct sk_buff *skb, 193 struct netlink_callback *cb, int p_type) 194 { 195 struct smc_diag_dump_ctx *cb_ctx = smc_dump_context(cb); 196 struct net *net = sock_net(skb->sk); 197 int snum = cb_ctx->pos[p_type]; 198 struct nlattr *bc = NULL; 199 struct hlist_head *head; 200 int rc = 0, num = 0; 201 struct sock *sk; 202 203 read_lock(&prot->h.smc_hash->lock); 204 head = &prot->h.smc_hash->ht; 205 if (hlist_empty(head)) 206 goto out; 207 208 sk_for_each(sk, head) { 209 if (!net_eq(sock_net(sk), net)) 210 continue; 211 if (num < snum) 212 goto next; 213 rc = __smc_diag_dump(sk, skb, cb, nlmsg_data(cb->nlh), bc); 214 if (rc < 0) 215 goto out; 216 next: 217 num++; 218 } 219 220 out: 221 read_unlock(&prot->h.smc_hash->lock); 222 cb_ctx->pos[p_type] = num; 223 return rc; 224 } 225 226 static int smc_diag_dump(struct sk_buff *skb, struct netlink_callback *cb) 227 { 228 int rc = 0; 229 230 rc = smc_diag_dump_proto(&smc_proto, skb, cb, SMCPROTO_SMC); 231 if (!rc) 232 smc_diag_dump_proto(&smc_proto6, skb, cb, SMCPROTO_SMC6); 233 return skb->len; 234 } 235 236 static int smc_diag_handler_dump(struct sk_buff *skb, struct nlmsghdr *h) 237 { 238 struct net *net = sock_net(skb->sk); 239 240 if (h->nlmsg_type == SOCK_DIAG_BY_FAMILY && 241 h->nlmsg_flags & NLM_F_DUMP) { 242 { 243 struct netlink_dump_control c = { 244 .dump = smc_diag_dump, 245 .min_dump_alloc = SKB_WITH_OVERHEAD(32768), 246 }; 247 return netlink_dump_start(net->diag_nlsk, skb, h, &c); 248 } 249 } 250 return 0; 251 } 252 253 static const struct sock_diag_handler smc_diag_handler = { 254 .family = AF_SMC, 255 .dump = smc_diag_handler_dump, 256 }; 257 258 static int __init smc_diag_init(void) 259 { 260 return sock_diag_register(&smc_diag_handler); 261 } 262 263 static void __exit smc_diag_exit(void) 264 { 265 sock_diag_unregister(&smc_diag_handler); 266 } 267 268 module_init(smc_diag_init); 269 module_exit(smc_diag_exit); 270 MODULE_LICENSE("GPL"); 271 MODULE_ALIAS_NET_PF_PROTO_TYPE(PF_NETLINK, NETLINK_SOCK_DIAG, 43 /* AF_SMC */); 272 MODULE_ALIAS_GENL_FAMILY(SMCR_GENL_FAMILY_NAME); 273