1 /* 2 * net/tipc/node.h: Include file for TIPC node management routines 3 * 4 * Copyright (c) 2000-2006, Ericsson AB 5 * Copyright (c) 2005, Wind River Systems 6 * All rights reserved. 7 * 8 * Redistribution and use in source and binary forms, with or without 9 * modification, are permitted provided that the following conditions are met: 10 * 11 * 1. Redistributions of source code must retain the above copyright 12 * notice, this list of conditions and the following disclaimer. 13 * 2. Redistributions in binary form must reproduce the above copyright 14 * notice, this list of conditions and the following disclaimer in the 15 * documentation and/or other materials provided with the distribution. 16 * 3. Neither the names of the copyright holders nor the names of its 17 * contributors may be used to endorse or promote products derived from 18 * this software without specific prior written permission. 19 * 20 * Alternatively, this software may be distributed under the terms of the 21 * GNU General Public License ("GPL") version 2 as published by the Free 22 * Software Foundation. 23 * 24 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" 25 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 26 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 27 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE 28 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 29 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 30 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 31 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 32 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 33 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 34 * POSSIBILITY OF SUCH DAMAGE. 35 */ 36 37 #ifndef _TIPC_NODE_H 38 #define _TIPC_NODE_H 39 40 #include "node_subscr.h" 41 #include "addr.h" 42 #include "cluster.h" 43 #include "bearer.h" 44 45 /** 46 * struct tipc_node - TIPC node structure 47 * @addr: network address of node 48 * @lock: spinlock governing access to structure 49 * @owner: pointer to cluster that node belongs to 50 * @next: pointer to next node in sorted list of cluster's nodes 51 * @nsub: list of "node down" subscriptions monitoring node 52 * @active_links: pointers to active links to node 53 * @links: pointers to all links to node 54 * @working_links: number of working links to node (both active and standby) 55 * @cleanup_required: non-zero if cleaning up after a prior loss of contact 56 * @link_cnt: number of links to node 57 * @permit_changeover: non-zero if node has redundant links to this system 58 * @routers: bitmap (used for multicluster communication) 59 * @last_router: (used for multicluster communication) 60 * @bclink: broadcast-related info 61 * @supported: non-zero if node supports TIPC b'cast capability 62 * @acked: sequence # of last outbound b'cast message acknowledged by node 63 * @last_in: sequence # of last in-sequence b'cast message received from node 64 * @gap_after: sequence # of last message not requiring a NAK request 65 * @gap_to: sequence # of last message requiring a NAK request 66 * @nack_sync: counter that determines when NAK requests should be sent 67 * @deferred_head: oldest OOS b'cast message received from node 68 * @deferred_tail: newest OOS b'cast message received from node 69 * @defragm: list of partially reassembled b'cast message fragments from node 70 */ 71 72 struct tipc_node { 73 u32 addr; 74 spinlock_t lock; 75 struct cluster *owner; 76 struct tipc_node *next; 77 struct list_head nsub; 78 struct link *active_links[2]; 79 struct link *links[MAX_BEARERS]; 80 int link_cnt; 81 int working_links; 82 int cleanup_required; 83 int permit_changeover; 84 u32 routers[512/32]; 85 int last_router; 86 struct { 87 int supported; 88 u32 acked; 89 u32 last_in; 90 u32 gap_after; 91 u32 gap_to; 92 u32 nack_sync; 93 struct sk_buff *deferred_head; 94 struct sk_buff *deferred_tail; 95 struct sk_buff *defragm; 96 } bclink; 97 }; 98 99 extern u32 tipc_own_tag; 100 101 struct tipc_node *tipc_node_create(u32 addr); 102 void tipc_node_delete(struct tipc_node *n_ptr); 103 struct tipc_node *tipc_node_attach_link(struct link *l_ptr); 104 void tipc_node_detach_link(struct tipc_node *n_ptr, struct link *l_ptr); 105 void tipc_node_link_down(struct tipc_node *n_ptr, struct link *l_ptr); 106 void tipc_node_link_up(struct tipc_node *n_ptr, struct link *l_ptr); 107 int tipc_node_has_active_links(struct tipc_node *n_ptr); 108 int tipc_node_has_redundant_links(struct tipc_node *n_ptr); 109 u32 tipc_node_select_router(struct tipc_node *n_ptr, u32 ref); 110 struct tipc_node *tipc_node_select_next_hop(u32 addr, u32 selector); 111 int tipc_node_is_up(struct tipc_node *n_ptr); 112 void tipc_node_add_router(struct tipc_node *n_ptr, u32 router); 113 void tipc_node_remove_router(struct tipc_node *n_ptr, u32 router); 114 struct sk_buff *tipc_node_get_links(const void *req_tlv_area, int req_tlv_space); 115 struct sk_buff *tipc_node_get_nodes(const void *req_tlv_area, int req_tlv_space); 116 117 static inline struct tipc_node *tipc_node_find(u32 addr) 118 { 119 if (likely(in_own_cluster(addr))) 120 return tipc_local_nodes[tipc_node(addr)]; 121 else if (tipc_addr_domain_valid(addr)) { 122 struct cluster *c_ptr = tipc_cltr_find(addr); 123 124 if (c_ptr) 125 return c_ptr->nodes[tipc_node(addr)]; 126 } 127 return NULL; 128 } 129 130 static inline struct tipc_node *tipc_node_select(u32 addr, u32 selector) 131 { 132 if (likely(in_own_cluster(addr))) 133 return tipc_local_nodes[tipc_node(addr)]; 134 return tipc_node_select_next_hop(addr, selector); 135 } 136 137 static inline void tipc_node_lock(struct tipc_node *n_ptr) 138 { 139 spin_lock_bh(&n_ptr->lock); 140 } 141 142 static inline void tipc_node_unlock(struct tipc_node *n_ptr) 143 { 144 spin_unlock_bh(&n_ptr->lock); 145 } 146 147 #endif 148