1 /* netfilter.c: look after the filters for various protocols. 2 * Heavily influenced by the old firewall.c by David Bonn and Alan Cox. 3 * 4 * Thanks to Rob `CmdrTaco' Malda for not influencing this code in any 5 * way. 6 * 7 * Rusty Russell (C)2000 -- This code is GPL. 8 * Patrick McHardy (c) 2006-2012 9 */ 10 #include <linux/kernel.h> 11 #include <linux/netfilter.h> 12 #include <net/protocol.h> 13 #include <linux/init.h> 14 #include <linux/skbuff.h> 15 #include <linux/wait.h> 16 #include <linux/module.h> 17 #include <linux/interrupt.h> 18 #include <linux/if.h> 19 #include <linux/netdevice.h> 20 #include <linux/netfilter_ipv6.h> 21 #include <linux/inetdevice.h> 22 #include <linux/proc_fs.h> 23 #include <linux/mutex.h> 24 #include <linux/slab.h> 25 #include <linux/rcupdate.h> 26 #include <net/net_namespace.h> 27 #include <net/sock.h> 28 29 #include "nf_internals.h" 30 31 static DEFINE_MUTEX(afinfo_mutex); 32 33 const struct nf_afinfo __rcu *nf_afinfo[NFPROTO_NUMPROTO] __read_mostly; 34 EXPORT_SYMBOL(nf_afinfo); 35 const struct nf_ipv6_ops __rcu *nf_ipv6_ops __read_mostly; 36 EXPORT_SYMBOL_GPL(nf_ipv6_ops); 37 38 DEFINE_PER_CPU(bool, nf_skb_duplicated); 39 EXPORT_SYMBOL_GPL(nf_skb_duplicated); 40 41 int nf_register_afinfo(const struct nf_afinfo *afinfo) 42 { 43 mutex_lock(&afinfo_mutex); 44 RCU_INIT_POINTER(nf_afinfo[afinfo->family], afinfo); 45 mutex_unlock(&afinfo_mutex); 46 return 0; 47 } 48 EXPORT_SYMBOL_GPL(nf_register_afinfo); 49 50 void nf_unregister_afinfo(const struct nf_afinfo *afinfo) 51 { 52 mutex_lock(&afinfo_mutex); 53 RCU_INIT_POINTER(nf_afinfo[afinfo->family], NULL); 54 mutex_unlock(&afinfo_mutex); 55 synchronize_rcu(); 56 } 57 EXPORT_SYMBOL_GPL(nf_unregister_afinfo); 58 59 #ifdef HAVE_JUMP_LABEL 60 struct static_key nf_hooks_needed[NFPROTO_NUMPROTO][NF_MAX_HOOKS]; 61 EXPORT_SYMBOL(nf_hooks_needed); 62 #endif 63 64 static DEFINE_MUTEX(nf_hook_mutex); 65 #define nf_entry_dereference(e) \ 66 rcu_dereference_protected(e, lockdep_is_held(&nf_hook_mutex)) 67 68 static struct nf_hook_entry __rcu **nf_hook_entry_head(struct net *net, const struct nf_hook_ops *reg) 69 { 70 if (reg->pf != NFPROTO_NETDEV) 71 return net->nf.hooks[reg->pf]+reg->hooknum; 72 73 #ifdef CONFIG_NETFILTER_INGRESS 74 if (reg->hooknum == NF_NETDEV_INGRESS) { 75 if (reg->dev && dev_net(reg->dev) == net) 76 return ®->dev->nf_hooks_ingress; 77 } 78 #endif 79 return NULL; 80 } 81 82 int nf_register_net_hook(struct net *net, const struct nf_hook_ops *reg) 83 { 84 struct nf_hook_entry __rcu **pp; 85 struct nf_hook_entry *entry, *p; 86 87 if (reg->pf == NFPROTO_NETDEV) { 88 #ifndef CONFIG_NETFILTER_INGRESS 89 if (reg->hooknum == NF_NETDEV_INGRESS) 90 return -EOPNOTSUPP; 91 #endif 92 if (reg->hooknum != NF_NETDEV_INGRESS || 93 !reg->dev || dev_net(reg->dev) != net) 94 return -EINVAL; 95 } 96 97 pp = nf_hook_entry_head(net, reg); 98 if (!pp) 99 return -EINVAL; 100 101 entry = kmalloc(sizeof(*entry), GFP_KERNEL); 102 if (!entry) 103 return -ENOMEM; 104 105 entry->orig_ops = reg; 106 entry->ops = *reg; 107 entry->next = NULL; 108 109 mutex_lock(&nf_hook_mutex); 110 111 /* Find the spot in the list */ 112 while ((p = nf_entry_dereference(*pp)) != NULL) { 113 if (reg->priority < p->orig_ops->priority) 114 break; 115 pp = &p->next; 116 } 117 rcu_assign_pointer(entry->next, p); 118 rcu_assign_pointer(*pp, entry); 119 120 mutex_unlock(&nf_hook_mutex); 121 #ifdef CONFIG_NETFILTER_INGRESS 122 if (reg->pf == NFPROTO_NETDEV && reg->hooknum == NF_NETDEV_INGRESS) 123 net_inc_ingress_queue(); 124 #endif 125 #ifdef HAVE_JUMP_LABEL 126 static_key_slow_inc(&nf_hooks_needed[reg->pf][reg->hooknum]); 127 #endif 128 return 0; 129 } 130 EXPORT_SYMBOL(nf_register_net_hook); 131 132 void nf_unregister_net_hook(struct net *net, const struct nf_hook_ops *reg) 133 { 134 struct nf_hook_entry __rcu **pp; 135 struct nf_hook_entry *p; 136 137 pp = nf_hook_entry_head(net, reg); 138 if (WARN_ON_ONCE(!pp)) 139 return; 140 141 mutex_lock(&nf_hook_mutex); 142 while ((p = nf_entry_dereference(*pp)) != NULL) { 143 if (p->orig_ops == reg) { 144 rcu_assign_pointer(*pp, p->next); 145 break; 146 } 147 pp = &p->next; 148 } 149 mutex_unlock(&nf_hook_mutex); 150 if (!p) { 151 WARN(1, "nf_unregister_net_hook: hook not found!\n"); 152 return; 153 } 154 #ifdef CONFIG_NETFILTER_INGRESS 155 if (reg->pf == NFPROTO_NETDEV && reg->hooknum == NF_NETDEV_INGRESS) 156 net_dec_ingress_queue(); 157 #endif 158 #ifdef HAVE_JUMP_LABEL 159 static_key_slow_dec(&nf_hooks_needed[reg->pf][reg->hooknum]); 160 #endif 161 synchronize_net(); 162 nf_queue_nf_hook_drop(net, p); 163 /* other cpu might still process nfqueue verdict that used reg */ 164 synchronize_net(); 165 kfree(p); 166 } 167 EXPORT_SYMBOL(nf_unregister_net_hook); 168 169 int nf_register_net_hooks(struct net *net, const struct nf_hook_ops *reg, 170 unsigned int n) 171 { 172 unsigned int i; 173 int err = 0; 174 175 for (i = 0; i < n; i++) { 176 err = nf_register_net_hook(net, ®[i]); 177 if (err) 178 goto err; 179 } 180 return err; 181 182 err: 183 if (i > 0) 184 nf_unregister_net_hooks(net, reg, i); 185 return err; 186 } 187 EXPORT_SYMBOL(nf_register_net_hooks); 188 189 void nf_unregister_net_hooks(struct net *net, const struct nf_hook_ops *reg, 190 unsigned int n) 191 { 192 while (n-- > 0) 193 nf_unregister_net_hook(net, ®[n]); 194 } 195 EXPORT_SYMBOL(nf_unregister_net_hooks); 196 197 static LIST_HEAD(nf_hook_list); 198 199 static int _nf_register_hook(struct nf_hook_ops *reg) 200 { 201 struct net *net, *last; 202 int ret; 203 204 for_each_net(net) { 205 ret = nf_register_net_hook(net, reg); 206 if (ret && ret != -ENOENT) 207 goto rollback; 208 } 209 list_add_tail(®->list, &nf_hook_list); 210 211 return 0; 212 rollback: 213 last = net; 214 for_each_net(net) { 215 if (net == last) 216 break; 217 nf_unregister_net_hook(net, reg); 218 } 219 return ret; 220 } 221 222 int nf_register_hook(struct nf_hook_ops *reg) 223 { 224 int ret; 225 226 rtnl_lock(); 227 ret = _nf_register_hook(reg); 228 rtnl_unlock(); 229 230 return ret; 231 } 232 EXPORT_SYMBOL(nf_register_hook); 233 234 static void _nf_unregister_hook(struct nf_hook_ops *reg) 235 { 236 struct net *net; 237 238 list_del(®->list); 239 for_each_net(net) 240 nf_unregister_net_hook(net, reg); 241 } 242 243 void nf_unregister_hook(struct nf_hook_ops *reg) 244 { 245 rtnl_lock(); 246 _nf_unregister_hook(reg); 247 rtnl_unlock(); 248 } 249 EXPORT_SYMBOL(nf_unregister_hook); 250 251 int nf_register_hooks(struct nf_hook_ops *reg, unsigned int n) 252 { 253 unsigned int i; 254 int err = 0; 255 256 for (i = 0; i < n; i++) { 257 err = nf_register_hook(®[i]); 258 if (err) 259 goto err; 260 } 261 return err; 262 263 err: 264 if (i > 0) 265 nf_unregister_hooks(reg, i); 266 return err; 267 } 268 EXPORT_SYMBOL(nf_register_hooks); 269 270 /* Caller MUST take rtnl_lock() */ 271 int _nf_register_hooks(struct nf_hook_ops *reg, unsigned int n) 272 { 273 unsigned int i; 274 int err = 0; 275 276 for (i = 0; i < n; i++) { 277 err = _nf_register_hook(®[i]); 278 if (err) 279 goto err; 280 } 281 return err; 282 283 err: 284 if (i > 0) 285 _nf_unregister_hooks(reg, i); 286 return err; 287 } 288 EXPORT_SYMBOL(_nf_register_hooks); 289 290 void nf_unregister_hooks(struct nf_hook_ops *reg, unsigned int n) 291 { 292 while (n-- > 0) 293 nf_unregister_hook(®[n]); 294 } 295 EXPORT_SYMBOL(nf_unregister_hooks); 296 297 /* Caller MUST take rtnl_lock */ 298 void _nf_unregister_hooks(struct nf_hook_ops *reg, unsigned int n) 299 { 300 while (n-- > 0) 301 _nf_unregister_hook(®[n]); 302 } 303 EXPORT_SYMBOL(_nf_unregister_hooks); 304 305 unsigned int nf_iterate(struct sk_buff *skb, 306 struct nf_hook_state *state, 307 struct nf_hook_entry **entryp) 308 { 309 unsigned int verdict; 310 311 do { 312 repeat: 313 verdict = (*entryp)->ops.hook((*entryp)->ops.priv, skb, state); 314 if (verdict != NF_ACCEPT) { 315 if (verdict != NF_REPEAT) 316 return verdict; 317 goto repeat; 318 } 319 *entryp = rcu_dereference((*entryp)->next); 320 } while (*entryp); 321 return NF_ACCEPT; 322 } 323 324 325 /* Returns 1 if okfn() needs to be executed by the caller, 326 * -EPERM for NF_DROP, 0 otherwise. Caller must hold rcu_read_lock. */ 327 int nf_hook_slow(struct sk_buff *skb, struct nf_hook_state *state, 328 struct nf_hook_entry *entry) 329 { 330 unsigned int verdict; 331 int ret; 332 333 next_hook: 334 verdict = nf_iterate(skb, state, &entry); 335 switch (verdict & NF_VERDICT_MASK) { 336 case NF_ACCEPT: 337 ret = 1; 338 break; 339 case NF_DROP: 340 kfree_skb(skb); 341 ret = NF_DROP_GETERR(verdict); 342 if (ret == 0) 343 ret = -EPERM; 344 break; 345 case NF_QUEUE: 346 ret = nf_queue(skb, state, &entry, verdict); 347 if (ret == 1 && entry) 348 goto next_hook; 349 /* Fall through. */ 350 default: 351 /* Implicit handling for NF_STOLEN, as well as any other non 352 * conventional verdicts. 353 */ 354 ret = 0; 355 break; 356 } 357 return ret; 358 } 359 EXPORT_SYMBOL(nf_hook_slow); 360 361 362 int skb_make_writable(struct sk_buff *skb, unsigned int writable_len) 363 { 364 if (writable_len > skb->len) 365 return 0; 366 367 /* Not exclusive use of packet? Must copy. */ 368 if (!skb_cloned(skb)) { 369 if (writable_len <= skb_headlen(skb)) 370 return 1; 371 } else if (skb_clone_writable(skb, writable_len)) 372 return 1; 373 374 if (writable_len <= skb_headlen(skb)) 375 writable_len = 0; 376 else 377 writable_len -= skb_headlen(skb); 378 379 return !!__pskb_pull_tail(skb, writable_len); 380 } 381 EXPORT_SYMBOL(skb_make_writable); 382 383 /* This needs to be compiled in any case to avoid dependencies between the 384 * nfnetlink_queue code and nf_conntrack. 385 */ 386 struct nfnl_ct_hook __rcu *nfnl_ct_hook __read_mostly; 387 EXPORT_SYMBOL_GPL(nfnl_ct_hook); 388 389 #if IS_ENABLED(CONFIG_NF_CONNTRACK) 390 /* This does not belong here, but locally generated errors need it if connection 391 tracking in use: without this, connection may not be in hash table, and hence 392 manufactured ICMP or RST packets will not be associated with it. */ 393 void (*ip_ct_attach)(struct sk_buff *, const struct sk_buff *) 394 __rcu __read_mostly; 395 EXPORT_SYMBOL(ip_ct_attach); 396 397 void nf_ct_attach(struct sk_buff *new, const struct sk_buff *skb) 398 { 399 void (*attach)(struct sk_buff *, const struct sk_buff *); 400 401 if (skb->nfct) { 402 rcu_read_lock(); 403 attach = rcu_dereference(ip_ct_attach); 404 if (attach) 405 attach(new, skb); 406 rcu_read_unlock(); 407 } 408 } 409 EXPORT_SYMBOL(nf_ct_attach); 410 411 void (*nf_ct_destroy)(struct nf_conntrack *) __rcu __read_mostly; 412 EXPORT_SYMBOL(nf_ct_destroy); 413 414 void nf_conntrack_destroy(struct nf_conntrack *nfct) 415 { 416 void (*destroy)(struct nf_conntrack *); 417 418 rcu_read_lock(); 419 destroy = rcu_dereference(nf_ct_destroy); 420 BUG_ON(destroy == NULL); 421 destroy(nfct); 422 rcu_read_unlock(); 423 } 424 EXPORT_SYMBOL(nf_conntrack_destroy); 425 426 /* Built-in default zone used e.g. by modules. */ 427 const struct nf_conntrack_zone nf_ct_zone_dflt = { 428 .id = NF_CT_DEFAULT_ZONE_ID, 429 .dir = NF_CT_DEFAULT_ZONE_DIR, 430 }; 431 EXPORT_SYMBOL_GPL(nf_ct_zone_dflt); 432 #endif /* CONFIG_NF_CONNTRACK */ 433 434 #ifdef CONFIG_NF_NAT_NEEDED 435 void (*nf_nat_decode_session_hook)(struct sk_buff *, struct flowi *); 436 EXPORT_SYMBOL(nf_nat_decode_session_hook); 437 #endif 438 439 static int nf_register_hook_list(struct net *net) 440 { 441 struct nf_hook_ops *elem; 442 int ret; 443 444 rtnl_lock(); 445 list_for_each_entry(elem, &nf_hook_list, list) { 446 ret = nf_register_net_hook(net, elem); 447 if (ret && ret != -ENOENT) 448 goto out_undo; 449 } 450 rtnl_unlock(); 451 return 0; 452 453 out_undo: 454 list_for_each_entry_continue_reverse(elem, &nf_hook_list, list) 455 nf_unregister_net_hook(net, elem); 456 rtnl_unlock(); 457 return ret; 458 } 459 460 static void nf_unregister_hook_list(struct net *net) 461 { 462 struct nf_hook_ops *elem; 463 464 rtnl_lock(); 465 list_for_each_entry(elem, &nf_hook_list, list) 466 nf_unregister_net_hook(net, elem); 467 rtnl_unlock(); 468 } 469 470 static int __net_init netfilter_net_init(struct net *net) 471 { 472 int i, h, ret; 473 474 for (i = 0; i < ARRAY_SIZE(net->nf.hooks); i++) { 475 for (h = 0; h < NF_MAX_HOOKS; h++) 476 RCU_INIT_POINTER(net->nf.hooks[i][h], NULL); 477 } 478 479 #ifdef CONFIG_PROC_FS 480 net->nf.proc_netfilter = proc_net_mkdir(net, "netfilter", 481 net->proc_net); 482 if (!net->nf.proc_netfilter) { 483 if (!net_eq(net, &init_net)) 484 pr_err("cannot create netfilter proc entry"); 485 486 return -ENOMEM; 487 } 488 #endif 489 ret = nf_register_hook_list(net); 490 if (ret) 491 remove_proc_entry("netfilter", net->proc_net); 492 493 return ret; 494 } 495 496 static void __net_exit netfilter_net_exit(struct net *net) 497 { 498 nf_unregister_hook_list(net); 499 remove_proc_entry("netfilter", net->proc_net); 500 } 501 502 static struct pernet_operations netfilter_net_ops = { 503 .init = netfilter_net_init, 504 .exit = netfilter_net_exit, 505 }; 506 507 int __init netfilter_init(void) 508 { 509 int ret; 510 511 ret = register_pernet_subsys(&netfilter_net_ops); 512 if (ret < 0) 513 goto err; 514 515 ret = netfilter_log_init(); 516 if (ret < 0) 517 goto err_pernet; 518 519 return 0; 520 err_pernet: 521 unregister_pernet_subsys(&netfilter_net_ops); 522 err: 523 return ret; 524 } 525