100e0f34cSAndy Grover /* 200e0f34cSAndy Grover * Copyright (c) 2006 Oracle. All rights reserved. 300e0f34cSAndy Grover * 400e0f34cSAndy Grover * This software is available to you under a choice of one of two 500e0f34cSAndy Grover * licenses. You may choose to be licensed under the terms of the GNU 600e0f34cSAndy Grover * General Public License (GPL) Version 2, available from the file 700e0f34cSAndy Grover * COPYING in the main directory of this source tree, or the 800e0f34cSAndy Grover * OpenIB.org BSD license below: 900e0f34cSAndy Grover * 1000e0f34cSAndy Grover * Redistribution and use in source and binary forms, with or 1100e0f34cSAndy Grover * without modification, are permitted provided that the following 1200e0f34cSAndy Grover * conditions are met: 1300e0f34cSAndy Grover * 1400e0f34cSAndy Grover * - Redistributions of source code must retain the above 1500e0f34cSAndy Grover * copyright notice, this list of conditions and the following 1600e0f34cSAndy Grover * disclaimer. 1700e0f34cSAndy Grover * 1800e0f34cSAndy Grover * - Redistributions in binary form must reproduce the above 1900e0f34cSAndy Grover * copyright notice, this list of conditions and the following 2000e0f34cSAndy Grover * disclaimer in the documentation and/or other materials 2100e0f34cSAndy Grover * provided with the distribution. 2200e0f34cSAndy Grover * 2300e0f34cSAndy Grover * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 2400e0f34cSAndy Grover * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 2500e0f34cSAndy Grover * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 2600e0f34cSAndy Grover * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 2700e0f34cSAndy Grover * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 2800e0f34cSAndy Grover * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 2900e0f34cSAndy Grover * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 3000e0f34cSAndy Grover * SOFTWARE. 3100e0f34cSAndy Grover * 3200e0f34cSAndy Grover */ 3300e0f34cSAndy Grover #include <linux/kernel.h> 3400e0f34cSAndy Grover #include <linux/list.h> 355a0e3ad6STejun Heo #include <linux/slab.h> 36bc3b2d7fSPaul Gortmaker #include <linux/export.h> 3700e0f34cSAndy Grover #include <net/inet_hashtables.h> 3800e0f34cSAndy Grover 3900e0f34cSAndy Grover #include "rds.h" 4000e0f34cSAndy Grover #include "loop.h" 4100e0f34cSAndy Grover 4200e0f34cSAndy Grover #define RDS_CONNECTION_HASH_BITS 12 4300e0f34cSAndy Grover #define RDS_CONNECTION_HASH_ENTRIES (1 << RDS_CONNECTION_HASH_BITS) 4400e0f34cSAndy Grover #define RDS_CONNECTION_HASH_MASK (RDS_CONNECTION_HASH_ENTRIES - 1) 4500e0f34cSAndy Grover 4600e0f34cSAndy Grover /* converting this to RCU is a chore for another day.. */ 4700e0f34cSAndy Grover static DEFINE_SPINLOCK(rds_conn_lock); 4800e0f34cSAndy Grover static unsigned long rds_conn_count; 4900e0f34cSAndy Grover static struct hlist_head rds_conn_hash[RDS_CONNECTION_HASH_ENTRIES]; 5000e0f34cSAndy Grover static struct kmem_cache *rds_conn_slab; 5100e0f34cSAndy Grover 5200e0f34cSAndy Grover static struct hlist_head *rds_conn_bucket(__be32 laddr, __be32 faddr) 5300e0f34cSAndy Grover { 541bbdceefSHannes Frederic Sowa static u32 rds_hash_secret __read_mostly; 551bbdceefSHannes Frederic Sowa 561bbdceefSHannes Frederic Sowa unsigned long hash; 571bbdceefSHannes Frederic Sowa 581bbdceefSHannes Frederic Sowa net_get_random_once(&rds_hash_secret, sizeof(rds_hash_secret)); 591bbdceefSHannes Frederic Sowa 6000e0f34cSAndy Grover /* Pass NULL, don't need struct net for hash */ 611bbdceefSHannes Frederic Sowa hash = __inet_ehashfn(be32_to_cpu(laddr), 0, 6265cd8033SHannes Frederic Sowa be32_to_cpu(faddr), 0, 631bbdceefSHannes Frederic Sowa rds_hash_secret); 6400e0f34cSAndy Grover return &rds_conn_hash[hash & RDS_CONNECTION_HASH_MASK]; 6500e0f34cSAndy Grover } 6600e0f34cSAndy Grover 6700e0f34cSAndy Grover #define rds_conn_info_set(var, test, suffix) do { \ 6800e0f34cSAndy Grover if (test) \ 6900e0f34cSAndy Grover var |= RDS_INFO_CONNECTION_FLAG_##suffix; \ 7000e0f34cSAndy Grover } while (0) 7100e0f34cSAndy Grover 72bcf50ef2SChris Mason /* rcu read lock must be held or the connection spinlock */ 738f384c01SSowmini Varadhan static struct rds_connection *rds_conn_lookup(struct net *net, 748f384c01SSowmini Varadhan struct hlist_head *head, 7500e0f34cSAndy Grover __be32 laddr, __be32 faddr, 7600e0f34cSAndy Grover struct rds_transport *trans) 7700e0f34cSAndy Grover { 7800e0f34cSAndy Grover struct rds_connection *conn, *ret = NULL; 7900e0f34cSAndy Grover 80b67bfe0dSSasha Levin hlist_for_each_entry_rcu(conn, head, c_hash_node) { 8100e0f34cSAndy Grover if (conn->c_faddr == faddr && conn->c_laddr == laddr && 828f384c01SSowmini Varadhan conn->c_trans == trans && net == rds_conn_net(conn)) { 8300e0f34cSAndy Grover ret = conn; 8400e0f34cSAndy Grover break; 8500e0f34cSAndy Grover } 8600e0f34cSAndy Grover } 8700e0f34cSAndy Grover rdsdebug("returning conn %p for %pI4 -> %pI4\n", ret, 8800e0f34cSAndy Grover &laddr, &faddr); 8900e0f34cSAndy Grover return ret; 9000e0f34cSAndy Grover } 9100e0f34cSAndy Grover 9200e0f34cSAndy Grover /* 9300e0f34cSAndy Grover * This is called by transports as they're bringing down a connection. 9400e0f34cSAndy Grover * It clears partial message state so that the transport can start sending 9500e0f34cSAndy Grover * and receiving over this connection again in the future. It is up to 9600e0f34cSAndy Grover * the transport to have serialized this call with its send and recv. 9700e0f34cSAndy Grover */ 98d769ef81SSowmini Varadhan static void rds_conn_path_reset(struct rds_conn_path *cp) 9900e0f34cSAndy Grover { 100d769ef81SSowmini Varadhan struct rds_connection *conn = cp->cp_conn; 101d769ef81SSowmini Varadhan 10200e0f34cSAndy Grover rdsdebug("connection %pI4 to %pI4 reset\n", 10300e0f34cSAndy Grover &conn->c_laddr, &conn->c_faddr); 10400e0f34cSAndy Grover 10500e0f34cSAndy Grover rds_stats_inc(s_conn_reset); 106d769ef81SSowmini Varadhan rds_send_path_reset(cp); 107d769ef81SSowmini Varadhan cp->cp_flags = 0; 10800e0f34cSAndy Grover 10900e0f34cSAndy Grover /* Do not clear next_rx_seq here, else we cannot distinguish 11000e0f34cSAndy Grover * retransmitted packets from new packets, and will hand all 11100e0f34cSAndy Grover * of them to the application. That is not consistent with the 11200e0f34cSAndy Grover * reliability guarantees of RDS. */ 11300e0f34cSAndy Grover } 11400e0f34cSAndy Grover 1151c5113cfSSowmini Varadhan static void __rds_conn_path_init(struct rds_connection *conn, 1161c5113cfSSowmini Varadhan struct rds_conn_path *cp, bool is_outgoing) 1171c5113cfSSowmini Varadhan { 1181c5113cfSSowmini Varadhan spin_lock_init(&cp->cp_lock); 1191c5113cfSSowmini Varadhan cp->cp_next_tx_seq = 1; 1201c5113cfSSowmini Varadhan init_waitqueue_head(&cp->cp_waitq); 1211c5113cfSSowmini Varadhan INIT_LIST_HEAD(&cp->cp_send_queue); 1221c5113cfSSowmini Varadhan INIT_LIST_HEAD(&cp->cp_retrans); 1231c5113cfSSowmini Varadhan 1241c5113cfSSowmini Varadhan cp->cp_conn = conn; 1251c5113cfSSowmini Varadhan atomic_set(&cp->cp_state, RDS_CONN_DOWN); 1261c5113cfSSowmini Varadhan cp->cp_send_gen = 0; 1271c5113cfSSowmini Varadhan cp->cp_reconnect_jiffies = 0; 1281c5113cfSSowmini Varadhan INIT_DELAYED_WORK(&cp->cp_send_w, rds_send_worker); 1291c5113cfSSowmini Varadhan INIT_DELAYED_WORK(&cp->cp_recv_w, rds_recv_worker); 1301c5113cfSSowmini Varadhan INIT_DELAYED_WORK(&cp->cp_conn_w, rds_connect_worker); 1311c5113cfSSowmini Varadhan INIT_WORK(&cp->cp_down_w, rds_shutdown_worker); 1321c5113cfSSowmini Varadhan mutex_init(&cp->cp_cm_lock); 1331c5113cfSSowmini Varadhan cp->cp_flags = 0; 1341c5113cfSSowmini Varadhan } 1351c5113cfSSowmini Varadhan 13600e0f34cSAndy Grover /* 13700e0f34cSAndy Grover * There is only every one 'conn' for a given pair of addresses in the 13800e0f34cSAndy Grover * system at a time. They contain messages to be retransmitted and so 13900e0f34cSAndy Grover * span the lifetime of the actual underlying transport connections. 14000e0f34cSAndy Grover * 14100e0f34cSAndy Grover * For now they are not garbage collected once they're created. They 14200e0f34cSAndy Grover * are torn down as the module is removed, if ever. 14300e0f34cSAndy Grover */ 144d5a8ac28SSowmini Varadhan static struct rds_connection *__rds_conn_create(struct net *net, 145d5a8ac28SSowmini Varadhan __be32 laddr, __be32 faddr, 14600e0f34cSAndy Grover struct rds_transport *trans, gfp_t gfp, 14700e0f34cSAndy Grover int is_outgoing) 14800e0f34cSAndy Grover { 149cb24405eSAndy Grover struct rds_connection *conn, *parent = NULL; 15000e0f34cSAndy Grover struct hlist_head *head = rds_conn_bucket(laddr, faddr); 1515adb5bc6SZach Brown struct rds_transport *loop_trans; 15200e0f34cSAndy Grover unsigned long flags; 1535916e2c1SSowmini Varadhan int ret, i; 154840df162SSowmini Varadhan int npaths = (trans->t_mp_capable ? RDS_MPATH_WORKERS : 1); 15500e0f34cSAndy Grover 156bcf50ef2SChris Mason rcu_read_lock(); 1578f384c01SSowmini Varadhan conn = rds_conn_lookup(net, head, laddr, faddr, trans); 158f64f9e71SJoe Perches if (conn && conn->c_loopback && conn->c_trans != &rds_loop_transport && 1591789b2c0SSowmini Varadhan laddr == faddr && !is_outgoing) { 16000e0f34cSAndy Grover /* This is a looped back IB connection, and we're 16100e0f34cSAndy Grover * called by the code handling the incoming connect. 16200e0f34cSAndy Grover * We need a second connection object into which we 16300e0f34cSAndy Grover * can stick the other QP. */ 16400e0f34cSAndy Grover parent = conn; 16500e0f34cSAndy Grover conn = parent->c_passive; 16600e0f34cSAndy Grover } 167bcf50ef2SChris Mason rcu_read_unlock(); 16800e0f34cSAndy Grover if (conn) 16900e0f34cSAndy Grover goto out; 17000e0f34cSAndy Grover 17105a178ecSWei Yongjun conn = kmem_cache_zalloc(rds_conn_slab, gfp); 1728690bfa1SAndy Grover if (!conn) { 17300e0f34cSAndy Grover conn = ERR_PTR(-ENOMEM); 17400e0f34cSAndy Grover goto out; 17500e0f34cSAndy Grover } 176840df162SSowmini Varadhan conn->c_path = kcalloc(npaths, sizeof(struct rds_conn_path), gfp); 177840df162SSowmini Varadhan if (!conn->c_path) { 178840df162SSowmini Varadhan kmem_cache_free(rds_conn_slab, conn); 179840df162SSowmini Varadhan conn = ERR_PTR(-ENOMEM); 180840df162SSowmini Varadhan goto out; 181840df162SSowmini Varadhan } 18200e0f34cSAndy Grover 18300e0f34cSAndy Grover INIT_HLIST_NODE(&conn->c_hash_node); 18400e0f34cSAndy Grover conn->c_laddr = laddr; 18500e0f34cSAndy Grover conn->c_faddr = faddr; 18600e0f34cSAndy Grover 1871c5113cfSSowmini Varadhan rds_conn_net_set(conn, net); 18800e0f34cSAndy Grover 18900e0f34cSAndy Grover ret = rds_cong_get_maps(conn); 19000e0f34cSAndy Grover if (ret) { 191840df162SSowmini Varadhan kfree(conn->c_path); 19200e0f34cSAndy Grover kmem_cache_free(rds_conn_slab, conn); 19300e0f34cSAndy Grover conn = ERR_PTR(ret); 19400e0f34cSAndy Grover goto out; 19500e0f34cSAndy Grover } 19600e0f34cSAndy Grover 19700e0f34cSAndy Grover /* 19800e0f34cSAndy Grover * This is where a connection becomes loopback. If *any* RDS sockets 19900e0f34cSAndy Grover * can bind to the destination address then we'd rather the messages 20000e0f34cSAndy Grover * flow through loopback rather than either transport. 20100e0f34cSAndy Grover */ 202d5a8ac28SSowmini Varadhan loop_trans = rds_trans_get_preferred(net, faddr); 2035adb5bc6SZach Brown if (loop_trans) { 2045adb5bc6SZach Brown rds_trans_put(loop_trans); 20500e0f34cSAndy Grover conn->c_loopback = 1; 20600e0f34cSAndy Grover if (is_outgoing && trans->t_prefer_loopback) { 20700e0f34cSAndy Grover /* "outgoing" connection - and the transport 20800e0f34cSAndy Grover * says it wants the connection handled by the 20900e0f34cSAndy Grover * loopback transport. This is what TCP does. 21000e0f34cSAndy Grover */ 21100e0f34cSAndy Grover trans = &rds_loop_transport; 21200e0f34cSAndy Grover } 21300e0f34cSAndy Grover } 21400e0f34cSAndy Grover 21500e0f34cSAndy Grover conn->c_trans = trans; 21600e0f34cSAndy Grover 2175916e2c1SSowmini Varadhan init_waitqueue_head(&conn->c_hs_waitq); 218840df162SSowmini Varadhan for (i = 0; i < npaths; i++) { 2195916e2c1SSowmini Varadhan __rds_conn_path_init(conn, &conn->c_path[i], 2205916e2c1SSowmini Varadhan is_outgoing); 2215916e2c1SSowmini Varadhan conn->c_path[i].cp_index = i; 2225916e2c1SSowmini Varadhan } 223ebeeb1adSSowmini Varadhan rcu_read_lock(); 224ebeeb1adSSowmini Varadhan if (rds_destroy_pending(conn)) 225ebeeb1adSSowmini Varadhan ret = -ENETDOWN; 226ebeeb1adSSowmini Varadhan else 227*d4014d8cSSowmini Varadhan ret = trans->conn_alloc(conn, GFP_ATOMIC); 22800e0f34cSAndy Grover if (ret) { 229ebeeb1adSSowmini Varadhan rcu_read_unlock(); 230840df162SSowmini Varadhan kfree(conn->c_path); 23100e0f34cSAndy Grover kmem_cache_free(rds_conn_slab, conn); 23200e0f34cSAndy Grover conn = ERR_PTR(ret); 23300e0f34cSAndy Grover goto out; 23400e0f34cSAndy Grover } 23500e0f34cSAndy Grover 23600e0f34cSAndy Grover rdsdebug("allocated conn %p for %pI4 -> %pI4 over %s %s\n", 23700e0f34cSAndy Grover conn, &laddr, &faddr, 238d9b86937SPrashant Bhole strnlen(trans->t_name, sizeof(trans->t_name)) ? trans->t_name : 239d9b86937SPrashant Bhole "[unknown]", is_outgoing ? "(outgoing)" : ""); 24000e0f34cSAndy Grover 241cb24405eSAndy Grover /* 242cb24405eSAndy Grover * Since we ran without holding the conn lock, someone could 243cb24405eSAndy Grover * have created the same conn (either normal or passive) in the 244cb24405eSAndy Grover * interim. We check while holding the lock. If we won, we complete 245cb24405eSAndy Grover * init and return our conn. If we lost, we rollback and return the 246cb24405eSAndy Grover * other one. 247cb24405eSAndy Grover */ 24800e0f34cSAndy Grover spin_lock_irqsave(&rds_conn_lock, flags); 249cb24405eSAndy Grover if (parent) { 250cb24405eSAndy Grover /* Creating passive conn */ 251cb24405eSAndy Grover if (parent->c_passive) { 2521c5113cfSSowmini Varadhan trans->conn_free(conn->c_path[0].cp_transport_data); 253840df162SSowmini Varadhan kfree(conn->c_path); 25400e0f34cSAndy Grover kmem_cache_free(rds_conn_slab, conn); 255cb24405eSAndy Grover conn = parent->c_passive; 25600e0f34cSAndy Grover } else { 257cb24405eSAndy Grover parent->c_passive = conn; 25800e0f34cSAndy Grover rds_cong_add_conn(conn); 25900e0f34cSAndy Grover rds_conn_count++; 26000e0f34cSAndy Grover } 261cb24405eSAndy Grover } else { 262cb24405eSAndy Grover /* Creating normal conn */ 263cb24405eSAndy Grover struct rds_connection *found; 26400e0f34cSAndy Grover 2658f384c01SSowmini Varadhan found = rds_conn_lookup(net, head, laddr, faddr, trans); 266cb24405eSAndy Grover if (found) { 2671c5113cfSSowmini Varadhan struct rds_conn_path *cp; 2681c5113cfSSowmini Varadhan int i; 2691c5113cfSSowmini Varadhan 270840df162SSowmini Varadhan for (i = 0; i < npaths; i++) { 2711c5113cfSSowmini Varadhan cp = &conn->c_path[i]; 27202105b2cSSowmini Varadhan /* The ->conn_alloc invocation may have 27302105b2cSSowmini Varadhan * allocated resource for all paths, so all 27402105b2cSSowmini Varadhan * of them may have to be freed here. 27502105b2cSSowmini Varadhan */ 27602105b2cSSowmini Varadhan if (cp->cp_transport_data) 2771c5113cfSSowmini Varadhan trans->conn_free(cp->cp_transport_data); 2781c5113cfSSowmini Varadhan } 279840df162SSowmini Varadhan kfree(conn->c_path); 280cb24405eSAndy Grover kmem_cache_free(rds_conn_slab, conn); 281cb24405eSAndy Grover conn = found; 282cb24405eSAndy Grover } else { 283905dd418SSowmini Varadhan conn->c_my_gen_num = rds_gen_num; 284905dd418SSowmini Varadhan conn->c_peer_gen_num = 0; 285bcf50ef2SChris Mason hlist_add_head_rcu(&conn->c_hash_node, head); 286cb24405eSAndy Grover rds_cong_add_conn(conn); 287cb24405eSAndy Grover rds_conn_count++; 288cb24405eSAndy Grover } 289cb24405eSAndy Grover } 29000e0f34cSAndy Grover spin_unlock_irqrestore(&rds_conn_lock, flags); 291ebeeb1adSSowmini Varadhan rcu_read_unlock(); 29200e0f34cSAndy Grover 29300e0f34cSAndy Grover out: 29400e0f34cSAndy Grover return conn; 29500e0f34cSAndy Grover } 29600e0f34cSAndy Grover 297d5a8ac28SSowmini Varadhan struct rds_connection *rds_conn_create(struct net *net, 298d5a8ac28SSowmini Varadhan __be32 laddr, __be32 faddr, 29900e0f34cSAndy Grover struct rds_transport *trans, gfp_t gfp) 30000e0f34cSAndy Grover { 301d5a8ac28SSowmini Varadhan return __rds_conn_create(net, laddr, faddr, trans, gfp, 0); 30200e0f34cSAndy Grover } 303616b757aSAndy Grover EXPORT_SYMBOL_GPL(rds_conn_create); 30400e0f34cSAndy Grover 305d5a8ac28SSowmini Varadhan struct rds_connection *rds_conn_create_outgoing(struct net *net, 306d5a8ac28SSowmini Varadhan __be32 laddr, __be32 faddr, 30700e0f34cSAndy Grover struct rds_transport *trans, gfp_t gfp) 30800e0f34cSAndy Grover { 309d5a8ac28SSowmini Varadhan return __rds_conn_create(net, laddr, faddr, trans, gfp, 1); 31000e0f34cSAndy Grover } 311616b757aSAndy Grover EXPORT_SYMBOL_GPL(rds_conn_create_outgoing); 31200e0f34cSAndy Grover 313d769ef81SSowmini Varadhan void rds_conn_shutdown(struct rds_conn_path *cp) 3142dc39357SAndy Grover { 315d769ef81SSowmini Varadhan struct rds_connection *conn = cp->cp_conn; 316d769ef81SSowmini Varadhan 3172dc39357SAndy Grover /* shut it down unless it's down already */ 318d769ef81SSowmini Varadhan if (!rds_conn_path_transition(cp, RDS_CONN_DOWN, RDS_CONN_DOWN)) { 3192dc39357SAndy Grover /* 3202dc39357SAndy Grover * Quiesce the connection mgmt handlers before we start tearing 3212dc39357SAndy Grover * things down. We don't hold the mutex for the entire 3222dc39357SAndy Grover * duration of the shutdown operation, else we may be 3232dc39357SAndy Grover * deadlocking with the CM handler. Instead, the CM event 3242dc39357SAndy Grover * handler is supposed to check for state DISCONNECTING 3252dc39357SAndy Grover */ 326d769ef81SSowmini Varadhan mutex_lock(&cp->cp_cm_lock); 327d769ef81SSowmini Varadhan if (!rds_conn_path_transition(cp, RDS_CONN_UP, 328d769ef81SSowmini Varadhan RDS_CONN_DISCONNECTING) && 329d769ef81SSowmini Varadhan !rds_conn_path_transition(cp, RDS_CONN_ERROR, 330d769ef81SSowmini Varadhan RDS_CONN_DISCONNECTING)) { 331d769ef81SSowmini Varadhan rds_conn_path_error(cp, 332d769ef81SSowmini Varadhan "shutdown called in state %d\n", 333d769ef81SSowmini Varadhan atomic_read(&cp->cp_state)); 334d769ef81SSowmini Varadhan mutex_unlock(&cp->cp_cm_lock); 3352dc39357SAndy Grover return; 3362dc39357SAndy Grover } 337d769ef81SSowmini Varadhan mutex_unlock(&cp->cp_cm_lock); 3382dc39357SAndy Grover 339d769ef81SSowmini Varadhan wait_event(cp->cp_waitq, 340d769ef81SSowmini Varadhan !test_bit(RDS_IN_XMIT, &cp->cp_flags)); 341d769ef81SSowmini Varadhan wait_event(cp->cp_waitq, 342d769ef81SSowmini Varadhan !test_bit(RDS_RECV_REFILL, &cp->cp_flags)); 3437e3f2952SChris Mason 344d769ef81SSowmini Varadhan conn->c_trans->conn_path_shutdown(cp); 345d769ef81SSowmini Varadhan rds_conn_path_reset(cp); 3462dc39357SAndy Grover 347d769ef81SSowmini Varadhan if (!rds_conn_path_transition(cp, RDS_CONN_DISCONNECTING, 348e97656d0SSowmini Varadhan RDS_CONN_DOWN) && 349e97656d0SSowmini Varadhan !rds_conn_path_transition(cp, RDS_CONN_ERROR, 350d769ef81SSowmini Varadhan RDS_CONN_DOWN)) { 3512dc39357SAndy Grover /* This can happen - eg when we're in the middle of tearing 3522dc39357SAndy Grover * down the connection, and someone unloads the rds module. 353e97656d0SSowmini Varadhan * Quite reproducible with loopback connections. 3542dc39357SAndy Grover * Mostly harmless. 355e97656d0SSowmini Varadhan * 356e97656d0SSowmini Varadhan * Note that this also happens with rds-tcp because 357e97656d0SSowmini Varadhan * we could have triggered rds_conn_path_drop in irq 358e97656d0SSowmini Varadhan * mode from rds_tcp_state change on the receipt of 359e97656d0SSowmini Varadhan * a FIN, thus we need to recheck for RDS_CONN_ERROR 360e97656d0SSowmini Varadhan * here. 3612dc39357SAndy Grover */ 362d769ef81SSowmini Varadhan rds_conn_path_error(cp, "%s: failed to transition " 363d769ef81SSowmini Varadhan "to state DOWN, current state " 364d769ef81SSowmini Varadhan "is %d\n", __func__, 365d769ef81SSowmini Varadhan atomic_read(&cp->cp_state)); 3662dc39357SAndy Grover return; 3672dc39357SAndy Grover } 3682dc39357SAndy Grover } 3692dc39357SAndy Grover 3702dc39357SAndy Grover /* Then reconnect if it's still live. 3712dc39357SAndy Grover * The passive side of an IB loopback connection is never added 3722dc39357SAndy Grover * to the conn hash, so we never trigger a reconnect on this 3732dc39357SAndy Grover * conn - the reconnect is always triggered by the active peer. */ 374d769ef81SSowmini Varadhan cancel_delayed_work_sync(&cp->cp_conn_w); 375bcf50ef2SChris Mason rcu_read_lock(); 376bcf50ef2SChris Mason if (!hlist_unhashed(&conn->c_hash_node)) { 377bcf50ef2SChris Mason rcu_read_unlock(); 378d769ef81SSowmini Varadhan rds_queue_reconnect(cp); 379bcf50ef2SChris Mason } else { 380bcf50ef2SChris Mason rcu_read_unlock(); 381bcf50ef2SChris Mason } 3822dc39357SAndy Grover } 3832dc39357SAndy Grover 3843ecc5693SSowmini Varadhan /* destroy a single rds_conn_path. rds_conn_destroy() iterates over 3853ecc5693SSowmini Varadhan * all paths using rds_conn_path_destroy() 3863ecc5693SSowmini Varadhan */ 3873ecc5693SSowmini Varadhan static void rds_conn_path_destroy(struct rds_conn_path *cp) 3883ecc5693SSowmini Varadhan { 3893ecc5693SSowmini Varadhan struct rds_message *rm, *rtmp; 3903ecc5693SSowmini Varadhan 39102105b2cSSowmini Varadhan if (!cp->cp_transport_data) 39202105b2cSSowmini Varadhan return; 39302105b2cSSowmini Varadhan 3943ecc5693SSowmini Varadhan /* make sure lingering queued work won't try to ref the conn */ 3953ecc5693SSowmini Varadhan cancel_delayed_work_sync(&cp->cp_send_w); 3963ecc5693SSowmini Varadhan cancel_delayed_work_sync(&cp->cp_recv_w); 3973ecc5693SSowmini Varadhan 398aed20a53SSowmini Varadhan rds_conn_path_drop(cp, true); 399aed20a53SSowmini Varadhan flush_work(&cp->cp_down_w); 400aed20a53SSowmini Varadhan 4013ecc5693SSowmini Varadhan /* tear down queued messages */ 4023ecc5693SSowmini Varadhan list_for_each_entry_safe(rm, rtmp, 4033ecc5693SSowmini Varadhan &cp->cp_send_queue, 4043ecc5693SSowmini Varadhan m_conn_item) { 4053ecc5693SSowmini Varadhan list_del_init(&rm->m_conn_item); 4063ecc5693SSowmini Varadhan BUG_ON(!list_empty(&rm->m_sock_item)); 4073ecc5693SSowmini Varadhan rds_message_put(rm); 4083ecc5693SSowmini Varadhan } 4093ecc5693SSowmini Varadhan if (cp->cp_xmit_rm) 4103ecc5693SSowmini Varadhan rds_message_put(cp->cp_xmit_rm); 4113ecc5693SSowmini Varadhan 4123db6e0d1SSowmini Varadhan WARN_ON(delayed_work_pending(&cp->cp_send_w)); 4133db6e0d1SSowmini Varadhan WARN_ON(delayed_work_pending(&cp->cp_recv_w)); 4143db6e0d1SSowmini Varadhan WARN_ON(delayed_work_pending(&cp->cp_conn_w)); 4153db6e0d1SSowmini Varadhan WARN_ON(work_pending(&cp->cp_down_w)); 4163db6e0d1SSowmini Varadhan 4173ecc5693SSowmini Varadhan cp->cp_conn->c_trans->conn_free(cp->cp_transport_data); 4183ecc5693SSowmini Varadhan } 4193ecc5693SSowmini Varadhan 4202dc39357SAndy Grover /* 4212dc39357SAndy Grover * Stop and free a connection. 422ffcec0e1SZach Brown * 423ffcec0e1SZach Brown * This can only be used in very limited circumstances. It assumes that once 424ffcec0e1SZach Brown * the conn has been shutdown that no one else is referencing the connection. 425ffcec0e1SZach Brown * We can only ensure this in the rmmod path in the current code. 4262dc39357SAndy Grover */ 42700e0f34cSAndy Grover void rds_conn_destroy(struct rds_connection *conn) 42800e0f34cSAndy Grover { 429fe8ff6b5SZach Brown unsigned long flags; 43002105b2cSSowmini Varadhan int i; 43102105b2cSSowmini Varadhan struct rds_conn_path *cp; 432840df162SSowmini Varadhan int npaths = (conn->c_trans->t_mp_capable ? RDS_MPATH_WORKERS : 1); 43300e0f34cSAndy Grover 43400e0f34cSAndy Grover rdsdebug("freeing conn %p for %pI4 -> " 43500e0f34cSAndy Grover "%pI4\n", conn, &conn->c_laddr, 43600e0f34cSAndy Grover &conn->c_faddr); 43700e0f34cSAndy Grover 438abf45439SChris Mason /* Ensure conn will not be scheduled for reconnect */ 439abf45439SChris Mason spin_lock_irq(&rds_conn_lock); 440bcf50ef2SChris Mason hlist_del_init_rcu(&conn->c_hash_node); 441abf45439SChris Mason spin_unlock_irq(&rds_conn_lock); 442bcf50ef2SChris Mason synchronize_rcu(); 443bcf50ef2SChris Mason 444ffcec0e1SZach Brown /* shut the connection down */ 445840df162SSowmini Varadhan for (i = 0; i < npaths; i++) { 4463ecc5693SSowmini Varadhan cp = &conn->c_path[i]; 4473ecc5693SSowmini Varadhan rds_conn_path_destroy(cp); 4483ecc5693SSowmini Varadhan BUG_ON(!list_empty(&cp->cp_retrans)); 44900e0f34cSAndy Grover } 45000e0f34cSAndy Grover 45100e0f34cSAndy Grover /* 45200e0f34cSAndy Grover * The congestion maps aren't freed up here. They're 45300e0f34cSAndy Grover * freed by rds_cong_exit() after all the connections 45400e0f34cSAndy Grover * have been freed. 45500e0f34cSAndy Grover */ 45600e0f34cSAndy Grover rds_cong_remove_conn(conn); 45700e0f34cSAndy Grover 458840df162SSowmini Varadhan kfree(conn->c_path); 45900e0f34cSAndy Grover kmem_cache_free(rds_conn_slab, conn); 46000e0f34cSAndy Grover 461fe8ff6b5SZach Brown spin_lock_irqsave(&rds_conn_lock, flags); 46200e0f34cSAndy Grover rds_conn_count--; 463fe8ff6b5SZach Brown spin_unlock_irqrestore(&rds_conn_lock, flags); 46400e0f34cSAndy Grover } 465616b757aSAndy Grover EXPORT_SYMBOL_GPL(rds_conn_destroy); 46600e0f34cSAndy Grover 46700e0f34cSAndy Grover static void rds_conn_message_info(struct socket *sock, unsigned int len, 46800e0f34cSAndy Grover struct rds_info_iterator *iter, 46900e0f34cSAndy Grover struct rds_info_lengths *lens, 47000e0f34cSAndy Grover int want_send) 47100e0f34cSAndy Grover { 47200e0f34cSAndy Grover struct hlist_head *head; 47300e0f34cSAndy Grover struct list_head *list; 47400e0f34cSAndy Grover struct rds_connection *conn; 47500e0f34cSAndy Grover struct rds_message *rm; 47600e0f34cSAndy Grover unsigned int total = 0; 477501dcccdSZach Brown unsigned long flags; 47800e0f34cSAndy Grover size_t i; 479992c9ec5SSowmini Varadhan int j; 48000e0f34cSAndy Grover 48100e0f34cSAndy Grover len /= sizeof(struct rds_info_message); 48200e0f34cSAndy Grover 483bcf50ef2SChris Mason rcu_read_lock(); 48400e0f34cSAndy Grover 48500e0f34cSAndy Grover for (i = 0, head = rds_conn_hash; i < ARRAY_SIZE(rds_conn_hash); 48600e0f34cSAndy Grover i++, head++) { 487b67bfe0dSSasha Levin hlist_for_each_entry_rcu(conn, head, c_hash_node) { 488992c9ec5SSowmini Varadhan struct rds_conn_path *cp; 489840df162SSowmini Varadhan int npaths; 49000e0f34cSAndy Grover 491840df162SSowmini Varadhan npaths = (conn->c_trans->t_mp_capable ? 492840df162SSowmini Varadhan RDS_MPATH_WORKERS : 1); 493840df162SSowmini Varadhan 494840df162SSowmini Varadhan for (j = 0; j < npaths; j++) { 495992c9ec5SSowmini Varadhan cp = &conn->c_path[j]; 496992c9ec5SSowmini Varadhan if (want_send) 497992c9ec5SSowmini Varadhan list = &cp->cp_send_queue; 498992c9ec5SSowmini Varadhan else 499992c9ec5SSowmini Varadhan list = &cp->cp_retrans; 500992c9ec5SSowmini Varadhan 501992c9ec5SSowmini Varadhan spin_lock_irqsave(&cp->cp_lock, flags); 50200e0f34cSAndy Grover 50300e0f34cSAndy Grover /* XXX too lazy to maintain counts.. */ 50400e0f34cSAndy Grover list_for_each_entry(rm, list, m_conn_item) { 50500e0f34cSAndy Grover total++; 50600e0f34cSAndy Grover if (total <= len) 507992c9ec5SSowmini Varadhan rds_inc_info_copy(&rm->m_inc, 508992c9ec5SSowmini Varadhan iter, 50900e0f34cSAndy Grover conn->c_laddr, 510992c9ec5SSowmini Varadhan conn->c_faddr, 511992c9ec5SSowmini Varadhan 0); 51200e0f34cSAndy Grover } 51300e0f34cSAndy Grover 514992c9ec5SSowmini Varadhan spin_unlock_irqrestore(&cp->cp_lock, flags); 515992c9ec5SSowmini Varadhan } 51600e0f34cSAndy Grover } 51700e0f34cSAndy Grover } 518bcf50ef2SChris Mason rcu_read_unlock(); 51900e0f34cSAndy Grover 52000e0f34cSAndy Grover lens->nr = total; 52100e0f34cSAndy Grover lens->each = sizeof(struct rds_info_message); 52200e0f34cSAndy Grover } 52300e0f34cSAndy Grover 52400e0f34cSAndy Grover static void rds_conn_message_info_send(struct socket *sock, unsigned int len, 52500e0f34cSAndy Grover struct rds_info_iterator *iter, 52600e0f34cSAndy Grover struct rds_info_lengths *lens) 52700e0f34cSAndy Grover { 52800e0f34cSAndy Grover rds_conn_message_info(sock, len, iter, lens, 1); 52900e0f34cSAndy Grover } 53000e0f34cSAndy Grover 53100e0f34cSAndy Grover static void rds_conn_message_info_retrans(struct socket *sock, 53200e0f34cSAndy Grover unsigned int len, 53300e0f34cSAndy Grover struct rds_info_iterator *iter, 53400e0f34cSAndy Grover struct rds_info_lengths *lens) 53500e0f34cSAndy Grover { 53600e0f34cSAndy Grover rds_conn_message_info(sock, len, iter, lens, 0); 53700e0f34cSAndy Grover } 53800e0f34cSAndy Grover 53900e0f34cSAndy Grover void rds_for_each_conn_info(struct socket *sock, unsigned int len, 54000e0f34cSAndy Grover struct rds_info_iterator *iter, 54100e0f34cSAndy Grover struct rds_info_lengths *lens, 54200e0f34cSAndy Grover int (*visitor)(struct rds_connection *, void *), 54300e0f34cSAndy Grover size_t item_len) 54400e0f34cSAndy Grover { 54500e0f34cSAndy Grover uint64_t buffer[(item_len + 7) / 8]; 54600e0f34cSAndy Grover struct hlist_head *head; 54700e0f34cSAndy Grover struct rds_connection *conn; 54800e0f34cSAndy Grover size_t i; 54900e0f34cSAndy Grover 550bcf50ef2SChris Mason rcu_read_lock(); 55100e0f34cSAndy Grover 55200e0f34cSAndy Grover lens->nr = 0; 55300e0f34cSAndy Grover lens->each = item_len; 55400e0f34cSAndy Grover 55500e0f34cSAndy Grover for (i = 0, head = rds_conn_hash; i < ARRAY_SIZE(rds_conn_hash); 55600e0f34cSAndy Grover i++, head++) { 557b67bfe0dSSasha Levin hlist_for_each_entry_rcu(conn, head, c_hash_node) { 55800e0f34cSAndy Grover 55900e0f34cSAndy Grover /* XXX no c_lock usage.. */ 56000e0f34cSAndy Grover if (!visitor(conn, buffer)) 56100e0f34cSAndy Grover continue; 56200e0f34cSAndy Grover 56300e0f34cSAndy Grover /* We copy as much as we can fit in the buffer, 56400e0f34cSAndy Grover * but we count all items so that the caller 56500e0f34cSAndy Grover * can resize the buffer. */ 56600e0f34cSAndy Grover if (len >= item_len) { 56700e0f34cSAndy Grover rds_info_copy(iter, buffer, item_len); 56800e0f34cSAndy Grover len -= item_len; 56900e0f34cSAndy Grover } 57000e0f34cSAndy Grover lens->nr++; 57100e0f34cSAndy Grover } 57200e0f34cSAndy Grover } 573bcf50ef2SChris Mason rcu_read_unlock(); 57400e0f34cSAndy Grover } 575616b757aSAndy Grover EXPORT_SYMBOL_GPL(rds_for_each_conn_info); 57600e0f34cSAndy Grover 577bb789763SSantosh Shilimkar static void rds_walk_conn_path_info(struct socket *sock, unsigned int len, 578992c9ec5SSowmini Varadhan struct rds_info_iterator *iter, 579992c9ec5SSowmini Varadhan struct rds_info_lengths *lens, 580992c9ec5SSowmini Varadhan int (*visitor)(struct rds_conn_path *, void *), 581992c9ec5SSowmini Varadhan size_t item_len) 582992c9ec5SSowmini Varadhan { 583992c9ec5SSowmini Varadhan u64 buffer[(item_len + 7) / 8]; 584992c9ec5SSowmini Varadhan struct hlist_head *head; 585992c9ec5SSowmini Varadhan struct rds_connection *conn; 586992c9ec5SSowmini Varadhan size_t i; 587992c9ec5SSowmini Varadhan int j; 588992c9ec5SSowmini Varadhan 589992c9ec5SSowmini Varadhan rcu_read_lock(); 590992c9ec5SSowmini Varadhan 591992c9ec5SSowmini Varadhan lens->nr = 0; 592992c9ec5SSowmini Varadhan lens->each = item_len; 593992c9ec5SSowmini Varadhan 594992c9ec5SSowmini Varadhan for (i = 0, head = rds_conn_hash; i < ARRAY_SIZE(rds_conn_hash); 595992c9ec5SSowmini Varadhan i++, head++) { 596992c9ec5SSowmini Varadhan hlist_for_each_entry_rcu(conn, head, c_hash_node) { 597992c9ec5SSowmini Varadhan struct rds_conn_path *cp; 598840df162SSowmini Varadhan int npaths; 599992c9ec5SSowmini Varadhan 600840df162SSowmini Varadhan npaths = (conn->c_trans->t_mp_capable ? 601840df162SSowmini Varadhan RDS_MPATH_WORKERS : 1); 602840df162SSowmini Varadhan for (j = 0; j < npaths; j++) { 603992c9ec5SSowmini Varadhan cp = &conn->c_path[j]; 604992c9ec5SSowmini Varadhan 605992c9ec5SSowmini Varadhan /* XXX no cp_lock usage.. */ 606992c9ec5SSowmini Varadhan if (!visitor(cp, buffer)) 607992c9ec5SSowmini Varadhan continue; 608992c9ec5SSowmini Varadhan } 609992c9ec5SSowmini Varadhan 610992c9ec5SSowmini Varadhan /* We copy as much as we can fit in the buffer, 611992c9ec5SSowmini Varadhan * but we count all items so that the caller 612992c9ec5SSowmini Varadhan * can resize the buffer. 613992c9ec5SSowmini Varadhan */ 614992c9ec5SSowmini Varadhan if (len >= item_len) { 615992c9ec5SSowmini Varadhan rds_info_copy(iter, buffer, item_len); 616992c9ec5SSowmini Varadhan len -= item_len; 617992c9ec5SSowmini Varadhan } 618992c9ec5SSowmini Varadhan lens->nr++; 619992c9ec5SSowmini Varadhan } 620992c9ec5SSowmini Varadhan } 621992c9ec5SSowmini Varadhan rcu_read_unlock(); 622992c9ec5SSowmini Varadhan } 623992c9ec5SSowmini Varadhan 624992c9ec5SSowmini Varadhan static int rds_conn_info_visitor(struct rds_conn_path *cp, void *buffer) 62500e0f34cSAndy Grover { 62600e0f34cSAndy Grover struct rds_info_connection *cinfo = buffer; 62700e0f34cSAndy Grover 628992c9ec5SSowmini Varadhan cinfo->next_tx_seq = cp->cp_next_tx_seq; 629992c9ec5SSowmini Varadhan cinfo->next_rx_seq = cp->cp_next_rx_seq; 630992c9ec5SSowmini Varadhan cinfo->laddr = cp->cp_conn->c_laddr; 631992c9ec5SSowmini Varadhan cinfo->faddr = cp->cp_conn->c_faddr; 632992c9ec5SSowmini Varadhan strncpy(cinfo->transport, cp->cp_conn->c_trans->t_name, 63300e0f34cSAndy Grover sizeof(cinfo->transport)); 63400e0f34cSAndy Grover cinfo->flags = 0; 63500e0f34cSAndy Grover 636992c9ec5SSowmini Varadhan rds_conn_info_set(cinfo->flags, test_bit(RDS_IN_XMIT, &cp->cp_flags), 6370f4b1c7eSZach Brown SENDING); 63800e0f34cSAndy Grover /* XXX Future: return the state rather than these funky bits */ 63900e0f34cSAndy Grover rds_conn_info_set(cinfo->flags, 640992c9ec5SSowmini Varadhan atomic_read(&cp->cp_state) == RDS_CONN_CONNECTING, 64100e0f34cSAndy Grover CONNECTING); 64200e0f34cSAndy Grover rds_conn_info_set(cinfo->flags, 643992c9ec5SSowmini Varadhan atomic_read(&cp->cp_state) == RDS_CONN_UP, 64400e0f34cSAndy Grover CONNECTED); 64500e0f34cSAndy Grover return 1; 64600e0f34cSAndy Grover } 64700e0f34cSAndy Grover 64800e0f34cSAndy Grover static void rds_conn_info(struct socket *sock, unsigned int len, 64900e0f34cSAndy Grover struct rds_info_iterator *iter, 65000e0f34cSAndy Grover struct rds_info_lengths *lens) 65100e0f34cSAndy Grover { 652992c9ec5SSowmini Varadhan rds_walk_conn_path_info(sock, len, iter, lens, 65300e0f34cSAndy Grover rds_conn_info_visitor, 65400e0f34cSAndy Grover sizeof(struct rds_info_connection)); 65500e0f34cSAndy Grover } 65600e0f34cSAndy Grover 657ef87b7eaSZach Brown int rds_conn_init(void) 65800e0f34cSAndy Grover { 65900e0f34cSAndy Grover rds_conn_slab = kmem_cache_create("rds_connection", 66000e0f34cSAndy Grover sizeof(struct rds_connection), 66100e0f34cSAndy Grover 0, 0, NULL); 6628690bfa1SAndy Grover if (!rds_conn_slab) 66300e0f34cSAndy Grover return -ENOMEM; 66400e0f34cSAndy Grover 66500e0f34cSAndy Grover rds_info_register_func(RDS_INFO_CONNECTIONS, rds_conn_info); 66600e0f34cSAndy Grover rds_info_register_func(RDS_INFO_SEND_MESSAGES, 66700e0f34cSAndy Grover rds_conn_message_info_send); 66800e0f34cSAndy Grover rds_info_register_func(RDS_INFO_RETRANS_MESSAGES, 66900e0f34cSAndy Grover rds_conn_message_info_retrans); 67000e0f34cSAndy Grover 67100e0f34cSAndy Grover return 0; 67200e0f34cSAndy Grover } 67300e0f34cSAndy Grover 67400e0f34cSAndy Grover void rds_conn_exit(void) 67500e0f34cSAndy Grover { 67600e0f34cSAndy Grover rds_loop_exit(); 67700e0f34cSAndy Grover 67800e0f34cSAndy Grover WARN_ON(!hlist_empty(rds_conn_hash)); 67900e0f34cSAndy Grover 68000e0f34cSAndy Grover kmem_cache_destroy(rds_conn_slab); 68100e0f34cSAndy Grover 68200e0f34cSAndy Grover rds_info_deregister_func(RDS_INFO_CONNECTIONS, rds_conn_info); 68300e0f34cSAndy Grover rds_info_deregister_func(RDS_INFO_SEND_MESSAGES, 68400e0f34cSAndy Grover rds_conn_message_info_send); 68500e0f34cSAndy Grover rds_info_deregister_func(RDS_INFO_RETRANS_MESSAGES, 68600e0f34cSAndy Grover rds_conn_message_info_retrans); 68700e0f34cSAndy Grover } 68800e0f34cSAndy Grover 68900e0f34cSAndy Grover /* 69000e0f34cSAndy Grover * Force a disconnect 69100e0f34cSAndy Grover */ 692aed20a53SSowmini Varadhan void rds_conn_path_drop(struct rds_conn_path *cp, bool destroy) 6930cb43965SSowmini Varadhan { 6940cb43965SSowmini Varadhan atomic_set(&cp->cp_state, RDS_CONN_ERROR); 695aed20a53SSowmini Varadhan 6963db6e0d1SSowmini Varadhan rcu_read_lock(); 697ebeeb1adSSowmini Varadhan if (!destroy && rds_destroy_pending(cp->cp_conn)) { 6983db6e0d1SSowmini Varadhan rcu_read_unlock(); 699aed20a53SSowmini Varadhan return; 7003db6e0d1SSowmini Varadhan } 7010cb43965SSowmini Varadhan queue_work(rds_wq, &cp->cp_down_w); 7023db6e0d1SSowmini Varadhan rcu_read_unlock(); 7030cb43965SSowmini Varadhan } 7040cb43965SSowmini Varadhan EXPORT_SYMBOL_GPL(rds_conn_path_drop); 7050cb43965SSowmini Varadhan 70600e0f34cSAndy Grover void rds_conn_drop(struct rds_connection *conn) 70700e0f34cSAndy Grover { 7085916e2c1SSowmini Varadhan WARN_ON(conn->c_trans->t_mp_capable); 709aed20a53SSowmini Varadhan rds_conn_path_drop(&conn->c_path[0], false); 71000e0f34cSAndy Grover } 711616b757aSAndy Grover EXPORT_SYMBOL_GPL(rds_conn_drop); 71200e0f34cSAndy Grover 71300e0f34cSAndy Grover /* 714f3c6808dSZach Brown * If the connection is down, trigger a connect. We may have scheduled a 715f3c6808dSZach Brown * delayed reconnect however - in this case we should not interfere. 716f3c6808dSZach Brown */ 7173c0a5900SSowmini Varadhan void rds_conn_path_connect_if_down(struct rds_conn_path *cp) 7183c0a5900SSowmini Varadhan { 7193db6e0d1SSowmini Varadhan rcu_read_lock(); 720ebeeb1adSSowmini Varadhan if (rds_destroy_pending(cp->cp_conn)) { 7213db6e0d1SSowmini Varadhan rcu_read_unlock(); 7223db6e0d1SSowmini Varadhan return; 7233db6e0d1SSowmini Varadhan } 7243c0a5900SSowmini Varadhan if (rds_conn_path_state(cp) == RDS_CONN_DOWN && 7253c0a5900SSowmini Varadhan !test_and_set_bit(RDS_RECONNECT_PENDING, &cp->cp_flags)) 7263c0a5900SSowmini Varadhan queue_delayed_work(rds_wq, &cp->cp_conn_w, 0); 7273db6e0d1SSowmini Varadhan rcu_read_unlock(); 7283c0a5900SSowmini Varadhan } 7291a0e100fSSowmini Varadhan EXPORT_SYMBOL_GPL(rds_conn_path_connect_if_down); 7303c0a5900SSowmini Varadhan 731f3c6808dSZach Brown void rds_conn_connect_if_down(struct rds_connection *conn) 732f3c6808dSZach Brown { 7333c0a5900SSowmini Varadhan WARN_ON(conn->c_trans->t_mp_capable); 7343c0a5900SSowmini Varadhan rds_conn_path_connect_if_down(&conn->c_path[0]); 735f3c6808dSZach Brown } 736f3c6808dSZach Brown EXPORT_SYMBOL_GPL(rds_conn_connect_if_down); 737f3c6808dSZach Brown 738fb1b3dc4SSowmini Varadhan void 739fb1b3dc4SSowmini Varadhan __rds_conn_path_error(struct rds_conn_path *cp, const char *fmt, ...) 740fb1b3dc4SSowmini Varadhan { 741fb1b3dc4SSowmini Varadhan va_list ap; 742fb1b3dc4SSowmini Varadhan 743fb1b3dc4SSowmini Varadhan va_start(ap, fmt); 744fb1b3dc4SSowmini Varadhan vprintk(fmt, ap); 745fb1b3dc4SSowmini Varadhan va_end(ap); 746fb1b3dc4SSowmini Varadhan 747aed20a53SSowmini Varadhan rds_conn_path_drop(cp, false); 748fb1b3dc4SSowmini Varadhan } 749