1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * Copyright (C)2003,2004 USAGI/WIDE Project 4 * 5 * Authors Mitsuru KANDA <mk@linux-ipv6.org> 6 * YOSHIFUJI Hideaki <yoshfuji@linux-ipv6.org> 7 * 8 * Based on net/ipv4/xfrm4_tunnel.c 9 */ 10 #include <linux/module.h> 11 #include <linux/xfrm.h> 12 #include <linux/slab.h> 13 #include <linux/rculist.h> 14 #include <net/ip.h> 15 #include <net/xfrm.h> 16 #include <net/ipv6.h> 17 #include <linux/ipv6.h> 18 #include <linux/icmpv6.h> 19 #include <linux/mutex.h> 20 #include <net/netns/generic.h> 21 22 #define XFRM6_TUNNEL_SPI_BYADDR_HSIZE 256 23 #define XFRM6_TUNNEL_SPI_BYSPI_HSIZE 256 24 25 #define XFRM6_TUNNEL_SPI_MIN 1 26 #define XFRM6_TUNNEL_SPI_MAX 0xffffffff 27 28 struct xfrm6_tunnel_net { 29 struct hlist_head spi_byaddr[XFRM6_TUNNEL_SPI_BYADDR_HSIZE]; 30 struct hlist_head spi_byspi[XFRM6_TUNNEL_SPI_BYSPI_HSIZE]; 31 u32 spi; 32 }; 33 34 static unsigned int xfrm6_tunnel_net_id __read_mostly; 35 static inline struct xfrm6_tunnel_net *xfrm6_tunnel_pernet(struct net *net) 36 { 37 return net_generic(net, xfrm6_tunnel_net_id); 38 } 39 40 /* 41 * xfrm_tunnel_spi things are for allocating unique id ("spi") 42 * per xfrm_address_t. 43 */ 44 struct xfrm6_tunnel_spi { 45 struct hlist_node list_byaddr; 46 struct hlist_node list_byspi; 47 xfrm_address_t addr; 48 u32 spi; 49 refcount_t refcnt; 50 struct rcu_head rcu_head; 51 }; 52 53 static DEFINE_SPINLOCK(xfrm6_tunnel_spi_lock); 54 55 static struct kmem_cache *xfrm6_tunnel_spi_kmem __read_mostly; 56 57 static inline unsigned int xfrm6_tunnel_spi_hash_byaddr(const xfrm_address_t *addr) 58 { 59 unsigned int h; 60 61 h = ipv6_addr_hash((const struct in6_addr *)addr); 62 h ^= h >> 16; 63 h ^= h >> 8; 64 h &= XFRM6_TUNNEL_SPI_BYADDR_HSIZE - 1; 65 66 return h; 67 } 68 69 static inline unsigned int xfrm6_tunnel_spi_hash_byspi(u32 spi) 70 { 71 return spi % XFRM6_TUNNEL_SPI_BYSPI_HSIZE; 72 } 73 74 static struct xfrm6_tunnel_spi *__xfrm6_tunnel_spi_lookup(struct net *net, const xfrm_address_t *saddr) 75 { 76 struct xfrm6_tunnel_net *xfrm6_tn = xfrm6_tunnel_pernet(net); 77 struct xfrm6_tunnel_spi *x6spi; 78 79 hlist_for_each_entry_rcu(x6spi, 80 &xfrm6_tn->spi_byaddr[xfrm6_tunnel_spi_hash_byaddr(saddr)], 81 list_byaddr, lockdep_is_held(&xfrm6_tunnel_spi_lock)) { 82 if (xfrm6_addr_equal(&x6spi->addr, saddr)) 83 return x6spi; 84 } 85 86 return NULL; 87 } 88 89 __be32 xfrm6_tunnel_spi_lookup(struct net *net, const xfrm_address_t *saddr) 90 { 91 struct xfrm6_tunnel_spi *x6spi; 92 u32 spi; 93 94 rcu_read_lock_bh(); 95 x6spi = __xfrm6_tunnel_spi_lookup(net, saddr); 96 spi = x6spi ? x6spi->spi : 0; 97 rcu_read_unlock_bh(); 98 return htonl(spi); 99 } 100 EXPORT_SYMBOL(xfrm6_tunnel_spi_lookup); 101 102 static int __xfrm6_tunnel_spi_check(struct net *net, u32 spi) 103 { 104 struct xfrm6_tunnel_net *xfrm6_tn = xfrm6_tunnel_pernet(net); 105 struct xfrm6_tunnel_spi *x6spi; 106 int index = xfrm6_tunnel_spi_hash_byspi(spi); 107 108 hlist_for_each_entry(x6spi, 109 &xfrm6_tn->spi_byspi[index], 110 list_byspi) { 111 if (x6spi->spi == spi) 112 return -1; 113 } 114 return index; 115 } 116 117 static u32 __xfrm6_tunnel_alloc_spi(struct net *net, xfrm_address_t *saddr) 118 { 119 struct xfrm6_tunnel_net *xfrm6_tn = xfrm6_tunnel_pernet(net); 120 u32 spi; 121 struct xfrm6_tunnel_spi *x6spi; 122 int index; 123 124 if (xfrm6_tn->spi < XFRM6_TUNNEL_SPI_MIN || 125 xfrm6_tn->spi >= XFRM6_TUNNEL_SPI_MAX) 126 xfrm6_tn->spi = XFRM6_TUNNEL_SPI_MIN; 127 else 128 xfrm6_tn->spi++; 129 130 for (spi = xfrm6_tn->spi; spi <= XFRM6_TUNNEL_SPI_MAX; spi++) { 131 index = __xfrm6_tunnel_spi_check(net, spi); 132 if (index >= 0) 133 goto alloc_spi; 134 135 if (spi == XFRM6_TUNNEL_SPI_MAX) 136 break; 137 } 138 for (spi = XFRM6_TUNNEL_SPI_MIN; spi < xfrm6_tn->spi; spi++) { 139 index = __xfrm6_tunnel_spi_check(net, spi); 140 if (index >= 0) 141 goto alloc_spi; 142 } 143 spi = 0; 144 goto out; 145 alloc_spi: 146 xfrm6_tn->spi = spi; 147 x6spi = kmem_cache_alloc(xfrm6_tunnel_spi_kmem, GFP_ATOMIC); 148 if (!x6spi) 149 goto out; 150 151 memcpy(&x6spi->addr, saddr, sizeof(x6spi->addr)); 152 x6spi->spi = spi; 153 refcount_set(&x6spi->refcnt, 1); 154 155 hlist_add_head_rcu(&x6spi->list_byspi, &xfrm6_tn->spi_byspi[index]); 156 157 index = xfrm6_tunnel_spi_hash_byaddr(saddr); 158 hlist_add_head_rcu(&x6spi->list_byaddr, &xfrm6_tn->spi_byaddr[index]); 159 out: 160 return spi; 161 } 162 163 __be32 xfrm6_tunnel_alloc_spi(struct net *net, xfrm_address_t *saddr) 164 { 165 struct xfrm6_tunnel_spi *x6spi; 166 u32 spi; 167 168 spin_lock_bh(&xfrm6_tunnel_spi_lock); 169 x6spi = __xfrm6_tunnel_spi_lookup(net, saddr); 170 if (x6spi) { 171 refcount_inc(&x6spi->refcnt); 172 spi = x6spi->spi; 173 } else 174 spi = __xfrm6_tunnel_alloc_spi(net, saddr); 175 spin_unlock_bh(&xfrm6_tunnel_spi_lock); 176 177 return htonl(spi); 178 } 179 EXPORT_SYMBOL(xfrm6_tunnel_alloc_spi); 180 181 static void x6spi_destroy_rcu(struct rcu_head *head) 182 { 183 kmem_cache_free(xfrm6_tunnel_spi_kmem, 184 container_of(head, struct xfrm6_tunnel_spi, rcu_head)); 185 } 186 187 static void xfrm6_tunnel_free_spi(struct net *net, xfrm_address_t *saddr) 188 { 189 struct xfrm6_tunnel_net *xfrm6_tn = xfrm6_tunnel_pernet(net); 190 struct xfrm6_tunnel_spi *x6spi; 191 struct hlist_node *n; 192 193 spin_lock_bh(&xfrm6_tunnel_spi_lock); 194 195 hlist_for_each_entry_safe(x6spi, n, 196 &xfrm6_tn->spi_byaddr[xfrm6_tunnel_spi_hash_byaddr(saddr)], 197 list_byaddr) 198 { 199 if (xfrm6_addr_equal(&x6spi->addr, saddr)) { 200 if (refcount_dec_and_test(&x6spi->refcnt)) { 201 hlist_del_rcu(&x6spi->list_byaddr); 202 hlist_del_rcu(&x6spi->list_byspi); 203 call_rcu(&x6spi->rcu_head, x6spi_destroy_rcu); 204 break; 205 } 206 } 207 } 208 spin_unlock_bh(&xfrm6_tunnel_spi_lock); 209 } 210 211 static int xfrm6_tunnel_output(struct xfrm_state *x, struct sk_buff *skb) 212 { 213 skb_push(skb, -skb_network_offset(skb)); 214 return 0; 215 } 216 217 static int xfrm6_tunnel_input(struct xfrm_state *x, struct sk_buff *skb) 218 { 219 return skb_network_header(skb)[IP6CB(skb)->nhoff]; 220 } 221 222 static int xfrm6_tunnel_rcv(struct sk_buff *skb) 223 { 224 struct net *net = dev_net(skb->dev); 225 const struct ipv6hdr *iph = ipv6_hdr(skb); 226 __be32 spi; 227 228 spi = xfrm6_tunnel_spi_lookup(net, (const xfrm_address_t *)&iph->saddr); 229 return xfrm6_rcv_spi(skb, IPPROTO_IPV6, spi, NULL); 230 } 231 232 static int xfrm6_tunnel_err(struct sk_buff *skb, struct inet6_skb_parm *opt, 233 u8 type, u8 code, int offset, __be32 info) 234 { 235 /* xfrm6_tunnel native err handling */ 236 switch (type) { 237 case ICMPV6_DEST_UNREACH: 238 switch (code) { 239 case ICMPV6_NOROUTE: 240 case ICMPV6_ADM_PROHIBITED: 241 case ICMPV6_NOT_NEIGHBOUR: 242 case ICMPV6_ADDR_UNREACH: 243 case ICMPV6_PORT_UNREACH: 244 default: 245 break; 246 } 247 break; 248 case ICMPV6_PKT_TOOBIG: 249 break; 250 case ICMPV6_TIME_EXCEED: 251 switch (code) { 252 case ICMPV6_EXC_HOPLIMIT: 253 break; 254 case ICMPV6_EXC_FRAGTIME: 255 default: 256 break; 257 } 258 break; 259 case ICMPV6_PARAMPROB: 260 switch (code) { 261 case ICMPV6_HDR_FIELD: break; 262 case ICMPV6_UNK_NEXTHDR: break; 263 case ICMPV6_UNK_OPTION: break; 264 } 265 break; 266 default: 267 break; 268 } 269 270 return 0; 271 } 272 273 static int xfrm6_tunnel_init_state(struct xfrm_state *x) 274 { 275 if (x->props.mode != XFRM_MODE_TUNNEL) 276 return -EINVAL; 277 278 if (x->encap) 279 return -EINVAL; 280 281 x->props.header_len = sizeof(struct ipv6hdr); 282 283 return 0; 284 } 285 286 static void xfrm6_tunnel_destroy(struct xfrm_state *x) 287 { 288 struct net *net = xs_net(x); 289 290 xfrm6_tunnel_free_spi(net, (xfrm_address_t *)&x->props.saddr); 291 } 292 293 static const struct xfrm_type xfrm6_tunnel_type = { 294 .owner = THIS_MODULE, 295 .proto = IPPROTO_IPV6, 296 .init_state = xfrm6_tunnel_init_state, 297 .destructor = xfrm6_tunnel_destroy, 298 .input = xfrm6_tunnel_input, 299 .output = xfrm6_tunnel_output, 300 }; 301 302 static struct xfrm6_tunnel xfrm6_tunnel_handler __read_mostly = { 303 .handler = xfrm6_tunnel_rcv, 304 .err_handler = xfrm6_tunnel_err, 305 .priority = 3, 306 }; 307 308 static struct xfrm6_tunnel xfrm46_tunnel_handler __read_mostly = { 309 .handler = xfrm6_tunnel_rcv, 310 .err_handler = xfrm6_tunnel_err, 311 .priority = 3, 312 }; 313 314 static int __net_init xfrm6_tunnel_net_init(struct net *net) 315 { 316 struct xfrm6_tunnel_net *xfrm6_tn = xfrm6_tunnel_pernet(net); 317 unsigned int i; 318 319 for (i = 0; i < XFRM6_TUNNEL_SPI_BYADDR_HSIZE; i++) 320 INIT_HLIST_HEAD(&xfrm6_tn->spi_byaddr[i]); 321 for (i = 0; i < XFRM6_TUNNEL_SPI_BYSPI_HSIZE; i++) 322 INIT_HLIST_HEAD(&xfrm6_tn->spi_byspi[i]); 323 xfrm6_tn->spi = 0; 324 325 return 0; 326 } 327 328 static void __net_exit xfrm6_tunnel_net_exit(struct net *net) 329 { 330 struct xfrm6_tunnel_net *xfrm6_tn = xfrm6_tunnel_pernet(net); 331 unsigned int i; 332 333 xfrm_flush_gc(); 334 xfrm_state_flush(net, 0, false, true); 335 336 for (i = 0; i < XFRM6_TUNNEL_SPI_BYADDR_HSIZE; i++) 337 WARN_ON_ONCE(!hlist_empty(&xfrm6_tn->spi_byaddr[i])); 338 339 for (i = 0; i < XFRM6_TUNNEL_SPI_BYSPI_HSIZE; i++) 340 WARN_ON_ONCE(!hlist_empty(&xfrm6_tn->spi_byspi[i])); 341 } 342 343 static struct pernet_operations xfrm6_tunnel_net_ops = { 344 .init = xfrm6_tunnel_net_init, 345 .exit = xfrm6_tunnel_net_exit, 346 .id = &xfrm6_tunnel_net_id, 347 .size = sizeof(struct xfrm6_tunnel_net), 348 }; 349 350 static int __init xfrm6_tunnel_init(void) 351 { 352 int rv; 353 354 xfrm6_tunnel_spi_kmem = kmem_cache_create("xfrm6_tunnel_spi", 355 sizeof(struct xfrm6_tunnel_spi), 356 0, SLAB_HWCACHE_ALIGN, 357 NULL); 358 if (!xfrm6_tunnel_spi_kmem) 359 return -ENOMEM; 360 rv = register_pernet_subsys(&xfrm6_tunnel_net_ops); 361 if (rv < 0) 362 goto out_pernet; 363 rv = xfrm_register_type(&xfrm6_tunnel_type, AF_INET6); 364 if (rv < 0) 365 goto out_type; 366 rv = xfrm6_tunnel_register(&xfrm6_tunnel_handler, AF_INET6); 367 if (rv < 0) 368 goto out_xfrm6; 369 rv = xfrm6_tunnel_register(&xfrm46_tunnel_handler, AF_INET); 370 if (rv < 0) 371 goto out_xfrm46; 372 return 0; 373 374 out_xfrm46: 375 xfrm6_tunnel_deregister(&xfrm6_tunnel_handler, AF_INET6); 376 out_xfrm6: 377 xfrm_unregister_type(&xfrm6_tunnel_type, AF_INET6); 378 out_type: 379 unregister_pernet_subsys(&xfrm6_tunnel_net_ops); 380 out_pernet: 381 kmem_cache_destroy(xfrm6_tunnel_spi_kmem); 382 return rv; 383 } 384 385 static void __exit xfrm6_tunnel_fini(void) 386 { 387 xfrm6_tunnel_deregister(&xfrm46_tunnel_handler, AF_INET); 388 xfrm6_tunnel_deregister(&xfrm6_tunnel_handler, AF_INET6); 389 xfrm_unregister_type(&xfrm6_tunnel_type, AF_INET6); 390 unregister_pernet_subsys(&xfrm6_tunnel_net_ops); 391 /* Someone maybe has gotten the xfrm6_tunnel_spi. 392 * So need to wait it. 393 */ 394 rcu_barrier(); 395 kmem_cache_destroy(xfrm6_tunnel_spi_kmem); 396 } 397 398 module_init(xfrm6_tunnel_init); 399 module_exit(xfrm6_tunnel_fini); 400 MODULE_LICENSE("GPL"); 401 MODULE_ALIAS_XFRM_TYPE(AF_INET6, XFRM_PROTO_IPV6); 402