1 /* IP Virtual Server 2 * data structure and functionality definitions 3 */ 4 5 #ifndef _NET_IP_VS_H 6 #define _NET_IP_VS_H 7 8 #include <linux/ip_vs.h> /* definitions shared with userland */ 9 10 #include <asm/types.h> /* for __uXX types */ 11 12 #include <linux/list.h> /* for struct list_head */ 13 #include <linux/spinlock.h> /* for struct rwlock_t */ 14 #include <linux/atomic.h> /* for struct atomic_t */ 15 #include <linux/compiler.h> 16 #include <linux/timer.h> 17 #include <linux/bug.h> 18 19 #include <net/checksum.h> 20 #include <linux/netfilter.h> /* for union nf_inet_addr */ 21 #include <linux/ip.h> 22 #include <linux/ipv6.h> /* for struct ipv6hdr */ 23 #include <net/ipv6.h> 24 #if IS_ENABLED(CONFIG_IP_VS_IPV6) 25 #include <linux/netfilter_ipv6/ip6_tables.h> 26 #endif 27 #if IS_ENABLED(CONFIG_NF_CONNTRACK) 28 #include <net/netfilter/nf_conntrack.h> 29 #endif 30 #include <net/net_namespace.h> /* Netw namespace */ 31 32 #define IP_VS_HDR_INVERSE 1 33 #define IP_VS_HDR_ICMP 2 34 35 /* Generic access of ipvs struct */ 36 static inline struct netns_ipvs *net_ipvs(struct net* net) 37 { 38 return net->ipvs; 39 } 40 41 /* Get net ptr from skb in traffic cases 42 * use skb_sknet when call is from userland (ioctl or netlink) 43 */ 44 static inline struct net *skb_net(const struct sk_buff *skb) 45 { 46 #ifdef CONFIG_NET_NS 47 #ifdef CONFIG_IP_VS_DEBUG 48 /* 49 * This is used for debug only. 50 * Start with the most likely hit 51 * End with BUG 52 */ 53 if (likely(skb->dev && dev_net(skb->dev))) 54 return dev_net(skb->dev); 55 if (skb_dst(skb) && skb_dst(skb)->dev) 56 return dev_net(skb_dst(skb)->dev); 57 WARN(skb->sk, "Maybe skb_sknet should be used in %s() at line:%d\n", 58 __func__, __LINE__); 59 if (likely(skb->sk && sock_net(skb->sk))) 60 return sock_net(skb->sk); 61 pr_err("There is no net ptr to find in the skb in %s() line:%d\n", 62 __func__, __LINE__); 63 BUG(); 64 #else 65 return dev_net(skb->dev ? : skb_dst(skb)->dev); 66 #endif 67 #else 68 return &init_net; 69 #endif 70 } 71 72 static inline struct net *skb_sknet(const struct sk_buff *skb) 73 { 74 #ifdef CONFIG_NET_NS 75 #ifdef CONFIG_IP_VS_DEBUG 76 /* Start with the most likely hit */ 77 if (likely(skb->sk && sock_net(skb->sk))) 78 return sock_net(skb->sk); 79 WARN(skb->dev, "Maybe skb_net should be used instead in %s() line:%d\n", 80 __func__, __LINE__); 81 if (likely(skb->dev && dev_net(skb->dev))) 82 return dev_net(skb->dev); 83 pr_err("There is no net ptr to find in the skb in %s() line:%d\n", 84 __func__, __LINE__); 85 BUG(); 86 #else 87 return sock_net(skb->sk); 88 #endif 89 #else 90 return &init_net; 91 #endif 92 } 93 94 /* This one needed for single_open_net since net is stored directly in 95 * private not as a struct i.e. seq_file_net can't be used. 96 */ 97 static inline struct net *seq_file_single_net(struct seq_file *seq) 98 { 99 #ifdef CONFIG_NET_NS 100 return (struct net *)seq->private; 101 #else 102 return &init_net; 103 #endif 104 } 105 106 /* Connections' size value needed by ip_vs_ctl.c */ 107 extern int ip_vs_conn_tab_size; 108 109 struct ip_vs_iphdr { 110 int hdr_flags; /* ipvs flags */ 111 __u32 off; /* Where IP or IPv4 header starts */ 112 __u32 len; /* IPv4 simply where L4 starts 113 * IPv6 where L4 Transport Header starts */ 114 __u16 fragoffs; /* IPv6 fragment offset, 0 if first frag (or not frag)*/ 115 __s16 protocol; 116 __s32 flags; 117 union nf_inet_addr saddr; 118 union nf_inet_addr daddr; 119 }; 120 121 static inline void *frag_safe_skb_hp(const struct sk_buff *skb, int offset, 122 int len, void *buffer, 123 const struct ip_vs_iphdr *ipvsh) 124 { 125 return skb_header_pointer(skb, offset, len, buffer); 126 } 127 128 /* This function handles filling *ip_vs_iphdr, both for IPv4 and IPv6. 129 * IPv6 requires some extra work, as finding proper header position, 130 * depend on the IPv6 extension headers. 131 */ 132 static inline int 133 ip_vs_fill_iph_skb_off(int af, const struct sk_buff *skb, int offset, 134 int hdr_flags, struct ip_vs_iphdr *iphdr) 135 { 136 iphdr->hdr_flags = hdr_flags; 137 iphdr->off = offset; 138 139 #ifdef CONFIG_IP_VS_IPV6 140 if (af == AF_INET6) { 141 struct ipv6hdr _iph; 142 const struct ipv6hdr *iph = skb_header_pointer( 143 skb, offset, sizeof(_iph), &_iph); 144 if (!iph) 145 return 0; 146 147 iphdr->saddr.in6 = iph->saddr; 148 iphdr->daddr.in6 = iph->daddr; 149 /* ipv6_find_hdr() updates len, flags */ 150 iphdr->len = offset; 151 iphdr->flags = 0; 152 iphdr->protocol = ipv6_find_hdr(skb, &iphdr->len, -1, 153 &iphdr->fragoffs, 154 &iphdr->flags); 155 if (iphdr->protocol < 0) 156 return 0; 157 } else 158 #endif 159 { 160 struct iphdr _iph; 161 const struct iphdr *iph = skb_header_pointer( 162 skb, offset, sizeof(_iph), &_iph); 163 if (!iph) 164 return 0; 165 166 iphdr->len = offset + iph->ihl * 4; 167 iphdr->fragoffs = 0; 168 iphdr->protocol = iph->protocol; 169 iphdr->saddr.ip = iph->saddr; 170 iphdr->daddr.ip = iph->daddr; 171 } 172 173 return 1; 174 } 175 176 static inline int 177 ip_vs_fill_iph_skb_icmp(int af, const struct sk_buff *skb, int offset, 178 bool inverse, struct ip_vs_iphdr *iphdr) 179 { 180 int hdr_flags = IP_VS_HDR_ICMP; 181 182 if (inverse) 183 hdr_flags |= IP_VS_HDR_INVERSE; 184 185 return ip_vs_fill_iph_skb_off(af, skb, offset, hdr_flags, iphdr); 186 } 187 188 static inline int 189 ip_vs_fill_iph_skb(int af, const struct sk_buff *skb, bool inverse, 190 struct ip_vs_iphdr *iphdr) 191 { 192 int hdr_flags = 0; 193 194 if (inverse) 195 hdr_flags |= IP_VS_HDR_INVERSE; 196 197 return ip_vs_fill_iph_skb_off(af, skb, skb_network_offset(skb), 198 hdr_flags, iphdr); 199 } 200 201 static inline bool 202 ip_vs_iph_inverse(const struct ip_vs_iphdr *iph) 203 { 204 return !!(iph->hdr_flags & IP_VS_HDR_INVERSE); 205 } 206 207 static inline bool 208 ip_vs_iph_icmp(const struct ip_vs_iphdr *iph) 209 { 210 return !!(iph->hdr_flags & IP_VS_HDR_ICMP); 211 } 212 213 static inline void ip_vs_addr_copy(int af, union nf_inet_addr *dst, 214 const union nf_inet_addr *src) 215 { 216 #ifdef CONFIG_IP_VS_IPV6 217 if (af == AF_INET6) 218 dst->in6 = src->in6; 219 else 220 #endif 221 dst->ip = src->ip; 222 } 223 224 static inline void ip_vs_addr_set(int af, union nf_inet_addr *dst, 225 const union nf_inet_addr *src) 226 { 227 #ifdef CONFIG_IP_VS_IPV6 228 if (af == AF_INET6) { 229 dst->in6 = src->in6; 230 return; 231 } 232 #endif 233 dst->ip = src->ip; 234 dst->all[1] = 0; 235 dst->all[2] = 0; 236 dst->all[3] = 0; 237 } 238 239 static inline int ip_vs_addr_equal(int af, const union nf_inet_addr *a, 240 const union nf_inet_addr *b) 241 { 242 #ifdef CONFIG_IP_VS_IPV6 243 if (af == AF_INET6) 244 return ipv6_addr_equal(&a->in6, &b->in6); 245 #endif 246 return a->ip == b->ip; 247 } 248 249 #ifdef CONFIG_IP_VS_DEBUG 250 #include <linux/net.h> 251 252 int ip_vs_get_debug_level(void); 253 254 static inline const char *ip_vs_dbg_addr(int af, char *buf, size_t buf_len, 255 const union nf_inet_addr *addr, 256 int *idx) 257 { 258 int len; 259 #ifdef CONFIG_IP_VS_IPV6 260 if (af == AF_INET6) 261 len = snprintf(&buf[*idx], buf_len - *idx, "[%pI6c]", 262 &addr->in6) + 1; 263 else 264 #endif 265 len = snprintf(&buf[*idx], buf_len - *idx, "%pI4", 266 &addr->ip) + 1; 267 268 *idx += len; 269 BUG_ON(*idx > buf_len + 1); 270 return &buf[*idx - len]; 271 } 272 273 #define IP_VS_DBG_BUF(level, msg, ...) \ 274 do { \ 275 char ip_vs_dbg_buf[160]; \ 276 int ip_vs_dbg_idx = 0; \ 277 if (level <= ip_vs_get_debug_level()) \ 278 printk(KERN_DEBUG pr_fmt(msg), ##__VA_ARGS__); \ 279 } while (0) 280 #define IP_VS_ERR_BUF(msg...) \ 281 do { \ 282 char ip_vs_dbg_buf[160]; \ 283 int ip_vs_dbg_idx = 0; \ 284 pr_err(msg); \ 285 } while (0) 286 287 /* Only use from within IP_VS_DBG_BUF() or IP_VS_ERR_BUF macros */ 288 #define IP_VS_DBG_ADDR(af, addr) \ 289 ip_vs_dbg_addr(af, ip_vs_dbg_buf, \ 290 sizeof(ip_vs_dbg_buf), addr, \ 291 &ip_vs_dbg_idx) 292 293 #define IP_VS_DBG(level, msg, ...) \ 294 do { \ 295 if (level <= ip_vs_get_debug_level()) \ 296 printk(KERN_DEBUG pr_fmt(msg), ##__VA_ARGS__); \ 297 } while (0) 298 #define IP_VS_DBG_RL(msg, ...) \ 299 do { \ 300 if (net_ratelimit()) \ 301 printk(KERN_DEBUG pr_fmt(msg), ##__VA_ARGS__); \ 302 } while (0) 303 #define IP_VS_DBG_PKT(level, af, pp, skb, ofs, msg) \ 304 do { \ 305 if (level <= ip_vs_get_debug_level()) \ 306 pp->debug_packet(af, pp, skb, ofs, msg); \ 307 } while (0) 308 #define IP_VS_DBG_RL_PKT(level, af, pp, skb, ofs, msg) \ 309 do { \ 310 if (level <= ip_vs_get_debug_level() && \ 311 net_ratelimit()) \ 312 pp->debug_packet(af, pp, skb, ofs, msg); \ 313 } while (0) 314 #else /* NO DEBUGGING at ALL */ 315 #define IP_VS_DBG_BUF(level, msg...) do {} while (0) 316 #define IP_VS_ERR_BUF(msg...) do {} while (0) 317 #define IP_VS_DBG(level, msg...) do {} while (0) 318 #define IP_VS_DBG_RL(msg...) do {} while (0) 319 #define IP_VS_DBG_PKT(level, af, pp, skb, ofs, msg) do {} while (0) 320 #define IP_VS_DBG_RL_PKT(level, af, pp, skb, ofs, msg) do {} while (0) 321 #endif 322 323 #define IP_VS_BUG() BUG() 324 #define IP_VS_ERR_RL(msg, ...) \ 325 do { \ 326 if (net_ratelimit()) \ 327 pr_err(msg, ##__VA_ARGS__); \ 328 } while (0) 329 330 #ifdef CONFIG_IP_VS_DEBUG 331 #define EnterFunction(level) \ 332 do { \ 333 if (level <= ip_vs_get_debug_level()) \ 334 printk(KERN_DEBUG \ 335 pr_fmt("Enter: %s, %s line %i\n"), \ 336 __func__, __FILE__, __LINE__); \ 337 } while (0) 338 #define LeaveFunction(level) \ 339 do { \ 340 if (level <= ip_vs_get_debug_level()) \ 341 printk(KERN_DEBUG \ 342 pr_fmt("Leave: %s, %s line %i\n"), \ 343 __func__, __FILE__, __LINE__); \ 344 } while (0) 345 #else 346 #define EnterFunction(level) do {} while (0) 347 #define LeaveFunction(level) do {} while (0) 348 #endif 349 350 /* The port number of FTP service (in network order). */ 351 #define FTPPORT cpu_to_be16(21) 352 #define FTPDATA cpu_to_be16(20) 353 354 /* TCP State Values */ 355 enum { 356 IP_VS_TCP_S_NONE = 0, 357 IP_VS_TCP_S_ESTABLISHED, 358 IP_VS_TCP_S_SYN_SENT, 359 IP_VS_TCP_S_SYN_RECV, 360 IP_VS_TCP_S_FIN_WAIT, 361 IP_VS_TCP_S_TIME_WAIT, 362 IP_VS_TCP_S_CLOSE, 363 IP_VS_TCP_S_CLOSE_WAIT, 364 IP_VS_TCP_S_LAST_ACK, 365 IP_VS_TCP_S_LISTEN, 366 IP_VS_TCP_S_SYNACK, 367 IP_VS_TCP_S_LAST 368 }; 369 370 /* UDP State Values */ 371 enum { 372 IP_VS_UDP_S_NORMAL, 373 IP_VS_UDP_S_LAST, 374 }; 375 376 /* ICMP State Values */ 377 enum { 378 IP_VS_ICMP_S_NORMAL, 379 IP_VS_ICMP_S_LAST, 380 }; 381 382 /* SCTP State Values */ 383 enum ip_vs_sctp_states { 384 IP_VS_SCTP_S_NONE, 385 IP_VS_SCTP_S_INIT1, 386 IP_VS_SCTP_S_INIT, 387 IP_VS_SCTP_S_COOKIE_SENT, 388 IP_VS_SCTP_S_COOKIE_REPLIED, 389 IP_VS_SCTP_S_COOKIE_WAIT, 390 IP_VS_SCTP_S_COOKIE, 391 IP_VS_SCTP_S_COOKIE_ECHOED, 392 IP_VS_SCTP_S_ESTABLISHED, 393 IP_VS_SCTP_S_SHUTDOWN_SENT, 394 IP_VS_SCTP_S_SHUTDOWN_RECEIVED, 395 IP_VS_SCTP_S_SHUTDOWN_ACK_SENT, 396 IP_VS_SCTP_S_REJECTED, 397 IP_VS_SCTP_S_CLOSED, 398 IP_VS_SCTP_S_LAST 399 }; 400 401 /* Delta sequence info structure 402 * Each ip_vs_conn has 2 (output AND input seq. changes). 403 * Only used in the VS/NAT. 404 */ 405 struct ip_vs_seq { 406 __u32 init_seq; /* Add delta from this seq */ 407 __u32 delta; /* Delta in sequence numbers */ 408 __u32 previous_delta; /* Delta in sequence numbers 409 * before last resized pkt */ 410 }; 411 412 /* counters per cpu */ 413 struct ip_vs_counters { 414 __u64 conns; /* connections scheduled */ 415 __u64 inpkts; /* incoming packets */ 416 __u64 outpkts; /* outgoing packets */ 417 __u64 inbytes; /* incoming bytes */ 418 __u64 outbytes; /* outgoing bytes */ 419 }; 420 /* Stats per cpu */ 421 struct ip_vs_cpu_stats { 422 struct ip_vs_counters cnt; 423 struct u64_stats_sync syncp; 424 }; 425 426 /* IPVS statistics objects */ 427 struct ip_vs_estimator { 428 struct list_head list; 429 430 u64 last_inbytes; 431 u64 last_outbytes; 432 u64 last_conns; 433 u64 last_inpkts; 434 u64 last_outpkts; 435 436 u64 cps; 437 u64 inpps; 438 u64 outpps; 439 u64 inbps; 440 u64 outbps; 441 }; 442 443 /* 444 * IPVS statistics object, 64-bit kernel version of struct ip_vs_stats_user 445 */ 446 struct ip_vs_kstats { 447 u64 conns; /* connections scheduled */ 448 u64 inpkts; /* incoming packets */ 449 u64 outpkts; /* outgoing packets */ 450 u64 inbytes; /* incoming bytes */ 451 u64 outbytes; /* outgoing bytes */ 452 453 u64 cps; /* current connection rate */ 454 u64 inpps; /* current in packet rate */ 455 u64 outpps; /* current out packet rate */ 456 u64 inbps; /* current in byte rate */ 457 u64 outbps; /* current out byte rate */ 458 }; 459 460 struct ip_vs_stats { 461 struct ip_vs_kstats kstats; /* kernel statistics */ 462 struct ip_vs_estimator est; /* estimator */ 463 struct ip_vs_cpu_stats __percpu *cpustats; /* per cpu counters */ 464 spinlock_t lock; /* spin lock */ 465 struct ip_vs_kstats kstats0; /* reset values */ 466 }; 467 468 struct dst_entry; 469 struct iphdr; 470 struct ip_vs_conn; 471 struct ip_vs_app; 472 struct sk_buff; 473 struct ip_vs_proto_data; 474 475 struct ip_vs_protocol { 476 struct ip_vs_protocol *next; 477 char *name; 478 u16 protocol; 479 u16 num_states; 480 int dont_defrag; 481 482 void (*init)(struct ip_vs_protocol *pp); 483 484 void (*exit)(struct ip_vs_protocol *pp); 485 486 int (*init_netns)(struct net *net, struct ip_vs_proto_data *pd); 487 488 void (*exit_netns)(struct net *net, struct ip_vs_proto_data *pd); 489 490 int (*conn_schedule)(int af, struct sk_buff *skb, 491 struct ip_vs_proto_data *pd, 492 int *verdict, struct ip_vs_conn **cpp, 493 struct ip_vs_iphdr *iph); 494 495 struct ip_vs_conn * 496 (*conn_in_get)(int af, 497 const struct sk_buff *skb, 498 const struct ip_vs_iphdr *iph); 499 500 struct ip_vs_conn * 501 (*conn_out_get)(int af, 502 const struct sk_buff *skb, 503 const struct ip_vs_iphdr *iph); 504 505 int (*snat_handler)(struct sk_buff *skb, struct ip_vs_protocol *pp, 506 struct ip_vs_conn *cp, struct ip_vs_iphdr *iph); 507 508 int (*dnat_handler)(struct sk_buff *skb, struct ip_vs_protocol *pp, 509 struct ip_vs_conn *cp, struct ip_vs_iphdr *iph); 510 511 int (*csum_check)(int af, struct sk_buff *skb, 512 struct ip_vs_protocol *pp); 513 514 const char *(*state_name)(int state); 515 516 void (*state_transition)(struct ip_vs_conn *cp, int direction, 517 const struct sk_buff *skb, 518 struct ip_vs_proto_data *pd); 519 520 int (*register_app)(struct net *net, struct ip_vs_app *inc); 521 522 void (*unregister_app)(struct net *net, struct ip_vs_app *inc); 523 524 int (*app_conn_bind)(struct ip_vs_conn *cp); 525 526 void (*debug_packet)(int af, struct ip_vs_protocol *pp, 527 const struct sk_buff *skb, 528 int offset, 529 const char *msg); 530 531 void (*timeout_change)(struct ip_vs_proto_data *pd, int flags); 532 }; 533 534 /* protocol data per netns */ 535 struct ip_vs_proto_data { 536 struct ip_vs_proto_data *next; 537 struct ip_vs_protocol *pp; 538 int *timeout_table; /* protocol timeout table */ 539 atomic_t appcnt; /* counter of proto app incs. */ 540 struct tcp_states_t *tcp_state_table; 541 }; 542 543 struct ip_vs_protocol *ip_vs_proto_get(unsigned short proto); 544 struct ip_vs_proto_data *ip_vs_proto_data_get(struct net *net, 545 unsigned short proto); 546 547 struct ip_vs_conn_param { 548 struct net *net; 549 const union nf_inet_addr *caddr; 550 const union nf_inet_addr *vaddr; 551 __be16 cport; 552 __be16 vport; 553 __u16 protocol; 554 u16 af; 555 556 const struct ip_vs_pe *pe; 557 char *pe_data; 558 __u8 pe_data_len; 559 }; 560 561 /* IP_VS structure allocated for each dynamically scheduled connection */ 562 struct ip_vs_conn { 563 struct hlist_node c_list; /* hashed list heads */ 564 /* Protocol, addresses and port numbers */ 565 __be16 cport; 566 __be16 dport; 567 __be16 vport; 568 u16 af; /* address family */ 569 union nf_inet_addr caddr; /* client address */ 570 union nf_inet_addr vaddr; /* virtual address */ 571 union nf_inet_addr daddr; /* destination address */ 572 volatile __u32 flags; /* status flags */ 573 __u16 protocol; /* Which protocol (TCP/UDP) */ 574 __u16 daf; /* Address family of the dest */ 575 #ifdef CONFIG_NET_NS 576 struct net *net; /* Name space */ 577 #endif 578 579 /* counter and timer */ 580 atomic_t refcnt; /* reference count */ 581 struct timer_list timer; /* Expiration timer */ 582 volatile unsigned long timeout; /* timeout */ 583 584 /* Flags and state transition */ 585 spinlock_t lock; /* lock for state transition */ 586 volatile __u16 state; /* state info */ 587 volatile __u16 old_state; /* old state, to be used for 588 * state transition triggerd 589 * synchronization 590 */ 591 __u32 fwmark; /* Fire wall mark from skb */ 592 unsigned long sync_endtime; /* jiffies + sent_retries */ 593 594 /* Control members */ 595 struct ip_vs_conn *control; /* Master control connection */ 596 atomic_t n_control; /* Number of controlled ones */ 597 struct ip_vs_dest *dest; /* real server */ 598 atomic_t in_pkts; /* incoming packet counter */ 599 600 /* Packet transmitter for different forwarding methods. If it 601 * mangles the packet, it must return NF_DROP or better NF_STOLEN, 602 * otherwise this must be changed to a sk_buff **. 603 * NF_ACCEPT can be returned when destination is local. 604 */ 605 int (*packet_xmit)(struct sk_buff *skb, struct ip_vs_conn *cp, 606 struct ip_vs_protocol *pp, struct ip_vs_iphdr *iph); 607 608 /* Note: we can group the following members into a structure, 609 * in order to save more space, and the following members are 610 * only used in VS/NAT anyway 611 */ 612 struct ip_vs_app *app; /* bound ip_vs_app object */ 613 void *app_data; /* Application private data */ 614 struct ip_vs_seq in_seq; /* incoming seq. struct */ 615 struct ip_vs_seq out_seq; /* outgoing seq. struct */ 616 617 const struct ip_vs_pe *pe; 618 char *pe_data; 619 __u8 pe_data_len; 620 621 struct rcu_head rcu_head; 622 }; 623 624 /* To save some memory in conn table when name space is disabled. */ 625 static inline struct net *ip_vs_conn_net(const struct ip_vs_conn *cp) 626 { 627 #ifdef CONFIG_NET_NS 628 return cp->net; 629 #else 630 return &init_net; 631 #endif 632 } 633 634 static inline void ip_vs_conn_net_set(struct ip_vs_conn *cp, struct net *net) 635 { 636 #ifdef CONFIG_NET_NS 637 cp->net = net; 638 #endif 639 } 640 641 static inline int ip_vs_conn_net_eq(const struct ip_vs_conn *cp, 642 struct net *net) 643 { 644 #ifdef CONFIG_NET_NS 645 return cp->net == net; 646 #else 647 return 1; 648 #endif 649 } 650 651 /* Extended internal versions of struct ip_vs_service_user and ip_vs_dest_user 652 * for IPv6 support. 653 * 654 * We need these to conveniently pass around service and destination 655 * options, but unfortunately, we also need to keep the old definitions to 656 * maintain userspace backwards compatibility for the setsockopt interface. 657 */ 658 struct ip_vs_service_user_kern { 659 /* virtual service addresses */ 660 u16 af; 661 u16 protocol; 662 union nf_inet_addr addr; /* virtual ip address */ 663 __be16 port; 664 u32 fwmark; /* firwall mark of service */ 665 666 /* virtual service options */ 667 char *sched_name; 668 char *pe_name; 669 unsigned int flags; /* virtual service flags */ 670 unsigned int timeout; /* persistent timeout in sec */ 671 __be32 netmask; /* persistent netmask or plen */ 672 }; 673 674 675 struct ip_vs_dest_user_kern { 676 /* destination server address */ 677 union nf_inet_addr addr; 678 __be16 port; 679 680 /* real server options */ 681 unsigned int conn_flags; /* connection flags */ 682 int weight; /* destination weight */ 683 684 /* thresholds for active connections */ 685 u32 u_threshold; /* upper threshold */ 686 u32 l_threshold; /* lower threshold */ 687 688 /* Address family of addr */ 689 u16 af; 690 }; 691 692 693 /* 694 * The information about the virtual service offered to the net and the 695 * forwarding entries. 696 */ 697 struct ip_vs_service { 698 struct hlist_node s_list; /* for normal service table */ 699 struct hlist_node f_list; /* for fwmark-based service table */ 700 atomic_t refcnt; /* reference counter */ 701 702 u16 af; /* address family */ 703 __u16 protocol; /* which protocol (TCP/UDP) */ 704 union nf_inet_addr addr; /* IP address for virtual service */ 705 __be16 port; /* port number for the service */ 706 __u32 fwmark; /* firewall mark of the service */ 707 unsigned int flags; /* service status flags */ 708 unsigned int timeout; /* persistent timeout in ticks */ 709 __be32 netmask; /* grouping granularity, mask/plen */ 710 struct net *net; 711 712 struct list_head destinations; /* real server d-linked list */ 713 __u32 num_dests; /* number of servers */ 714 struct ip_vs_stats stats; /* statistics for the service */ 715 716 /* for scheduling */ 717 struct ip_vs_scheduler __rcu *scheduler; /* bound scheduler object */ 718 spinlock_t sched_lock; /* lock sched_data */ 719 void *sched_data; /* scheduler application data */ 720 721 /* alternate persistence engine */ 722 struct ip_vs_pe __rcu *pe; 723 724 struct rcu_head rcu_head; 725 }; 726 727 /* Information for cached dst */ 728 struct ip_vs_dest_dst { 729 struct dst_entry *dst_cache; /* destination cache entry */ 730 u32 dst_cookie; 731 union nf_inet_addr dst_saddr; 732 struct rcu_head rcu_head; 733 }; 734 735 /* The real server destination forwarding entry with ip address, port number, 736 * and so on. 737 */ 738 struct ip_vs_dest { 739 struct list_head n_list; /* for the dests in the service */ 740 struct hlist_node d_list; /* for table with all the dests */ 741 742 u16 af; /* address family */ 743 __be16 port; /* port number of the server */ 744 union nf_inet_addr addr; /* IP address of the server */ 745 volatile unsigned int flags; /* dest status flags */ 746 atomic_t conn_flags; /* flags to copy to conn */ 747 atomic_t weight; /* server weight */ 748 749 atomic_t refcnt; /* reference counter */ 750 struct ip_vs_stats stats; /* statistics */ 751 unsigned long idle_start; /* start time, jiffies */ 752 753 /* connection counters and thresholds */ 754 atomic_t activeconns; /* active connections */ 755 atomic_t inactconns; /* inactive connections */ 756 atomic_t persistconns; /* persistent connections */ 757 __u32 u_threshold; /* upper threshold */ 758 __u32 l_threshold; /* lower threshold */ 759 760 /* for destination cache */ 761 spinlock_t dst_lock; /* lock of dst_cache */ 762 struct ip_vs_dest_dst __rcu *dest_dst; /* cached dst info */ 763 764 /* for virtual service */ 765 struct ip_vs_service __rcu *svc; /* service it belongs to */ 766 __u16 protocol; /* which protocol (TCP/UDP) */ 767 __be16 vport; /* virtual port number */ 768 union nf_inet_addr vaddr; /* virtual IP address */ 769 __u32 vfwmark; /* firewall mark of service */ 770 771 struct list_head t_list; /* in dest_trash */ 772 unsigned int in_rs_table:1; /* we are in rs_table */ 773 }; 774 775 /* The scheduler object */ 776 struct ip_vs_scheduler { 777 struct list_head n_list; /* d-linked list head */ 778 char *name; /* scheduler name */ 779 atomic_t refcnt; /* reference counter */ 780 struct module *module; /* THIS_MODULE/NULL */ 781 782 /* scheduler initializing service */ 783 int (*init_service)(struct ip_vs_service *svc); 784 /* scheduling service finish */ 785 void (*done_service)(struct ip_vs_service *svc); 786 /* dest is linked */ 787 int (*add_dest)(struct ip_vs_service *svc, struct ip_vs_dest *dest); 788 /* dest is unlinked */ 789 int (*del_dest)(struct ip_vs_service *svc, struct ip_vs_dest *dest); 790 /* dest is updated */ 791 int (*upd_dest)(struct ip_vs_service *svc, struct ip_vs_dest *dest); 792 793 /* selecting a server from the given service */ 794 struct ip_vs_dest* (*schedule)(struct ip_vs_service *svc, 795 const struct sk_buff *skb, 796 struct ip_vs_iphdr *iph); 797 }; 798 799 /* The persistence engine object */ 800 struct ip_vs_pe { 801 struct list_head n_list; /* d-linked list head */ 802 char *name; /* scheduler name */ 803 atomic_t refcnt; /* reference counter */ 804 struct module *module; /* THIS_MODULE/NULL */ 805 806 /* get the connection template, if any */ 807 int (*fill_param)(struct ip_vs_conn_param *p, struct sk_buff *skb); 808 bool (*ct_match)(const struct ip_vs_conn_param *p, 809 struct ip_vs_conn *ct); 810 u32 (*hashkey_raw)(const struct ip_vs_conn_param *p, u32 initval, 811 bool inverse); 812 int (*show_pe_data)(const struct ip_vs_conn *cp, char *buf); 813 }; 814 815 /* The application module object (a.k.a. app incarnation) */ 816 struct ip_vs_app { 817 struct list_head a_list; /* member in app list */ 818 int type; /* IP_VS_APP_TYPE_xxx */ 819 char *name; /* application module name */ 820 __u16 protocol; 821 struct module *module; /* THIS_MODULE/NULL */ 822 struct list_head incs_list; /* list of incarnations */ 823 824 /* members for application incarnations */ 825 struct list_head p_list; /* member in proto app list */ 826 struct ip_vs_app *app; /* its real application */ 827 __be16 port; /* port number in net order */ 828 atomic_t usecnt; /* usage counter */ 829 struct rcu_head rcu_head; 830 831 /* output hook: Process packet in inout direction, diff set for TCP. 832 * Return: 0=Error, 1=Payload Not Mangled/Mangled but checksum is ok, 833 * 2=Mangled but checksum was not updated 834 */ 835 int (*pkt_out)(struct ip_vs_app *, struct ip_vs_conn *, 836 struct sk_buff *, int *diff); 837 838 /* input hook: Process packet in outin direction, diff set for TCP. 839 * Return: 0=Error, 1=Payload Not Mangled/Mangled but checksum is ok, 840 * 2=Mangled but checksum was not updated 841 */ 842 int (*pkt_in)(struct ip_vs_app *, struct ip_vs_conn *, 843 struct sk_buff *, int *diff); 844 845 /* ip_vs_app initializer */ 846 int (*init_conn)(struct ip_vs_app *, struct ip_vs_conn *); 847 848 /* ip_vs_app finish */ 849 int (*done_conn)(struct ip_vs_app *, struct ip_vs_conn *); 850 851 852 /* not used now */ 853 int (*bind_conn)(struct ip_vs_app *, struct ip_vs_conn *, 854 struct ip_vs_protocol *); 855 856 void (*unbind_conn)(struct ip_vs_app *, struct ip_vs_conn *); 857 858 int * timeout_table; 859 int * timeouts; 860 int timeouts_size; 861 862 int (*conn_schedule)(struct sk_buff *skb, struct ip_vs_app *app, 863 int *verdict, struct ip_vs_conn **cpp); 864 865 struct ip_vs_conn * 866 (*conn_in_get)(const struct sk_buff *skb, struct ip_vs_app *app, 867 const struct iphdr *iph, int inverse); 868 869 struct ip_vs_conn * 870 (*conn_out_get)(const struct sk_buff *skb, struct ip_vs_app *app, 871 const struct iphdr *iph, int inverse); 872 873 int (*state_transition)(struct ip_vs_conn *cp, int direction, 874 const struct sk_buff *skb, 875 struct ip_vs_app *app); 876 877 void (*timeout_change)(struct ip_vs_app *app, int flags); 878 }; 879 880 struct ipvs_master_sync_state { 881 struct list_head sync_queue; 882 struct ip_vs_sync_buff *sync_buff; 883 unsigned long sync_queue_len; 884 unsigned int sync_queue_delay; 885 struct task_struct *master_thread; 886 struct delayed_work master_wakeup_work; 887 struct netns_ipvs *ipvs; 888 }; 889 890 /* How much time to keep dests in trash */ 891 #define IP_VS_DEST_TRASH_PERIOD (120 * HZ) 892 893 struct ipvs_sync_daemon_cfg { 894 union nf_inet_addr mcast_group; 895 int syncid; 896 u16 sync_maxlen; 897 u16 mcast_port; 898 u8 mcast_af; 899 u8 mcast_ttl; 900 /* multicast interface name */ 901 char mcast_ifn[IP_VS_IFNAME_MAXLEN]; 902 }; 903 904 /* IPVS in network namespace */ 905 struct netns_ipvs { 906 int gen; /* Generation */ 907 int enable; /* enable like nf_hooks do */ 908 /* Hash table: for real service lookups */ 909 #define IP_VS_RTAB_BITS 4 910 #define IP_VS_RTAB_SIZE (1 << IP_VS_RTAB_BITS) 911 #define IP_VS_RTAB_MASK (IP_VS_RTAB_SIZE - 1) 912 913 struct hlist_head rs_table[IP_VS_RTAB_SIZE]; 914 /* ip_vs_app */ 915 struct list_head app_list; 916 /* ip_vs_proto */ 917 #define IP_VS_PROTO_TAB_SIZE 32 /* must be power of 2 */ 918 struct ip_vs_proto_data *proto_data_table[IP_VS_PROTO_TAB_SIZE]; 919 /* ip_vs_proto_tcp */ 920 #ifdef CONFIG_IP_VS_PROTO_TCP 921 #define TCP_APP_TAB_BITS 4 922 #define TCP_APP_TAB_SIZE (1 << TCP_APP_TAB_BITS) 923 #define TCP_APP_TAB_MASK (TCP_APP_TAB_SIZE - 1) 924 struct list_head tcp_apps[TCP_APP_TAB_SIZE]; 925 #endif 926 /* ip_vs_proto_udp */ 927 #ifdef CONFIG_IP_VS_PROTO_UDP 928 #define UDP_APP_TAB_BITS 4 929 #define UDP_APP_TAB_SIZE (1 << UDP_APP_TAB_BITS) 930 #define UDP_APP_TAB_MASK (UDP_APP_TAB_SIZE - 1) 931 struct list_head udp_apps[UDP_APP_TAB_SIZE]; 932 #endif 933 /* ip_vs_proto_sctp */ 934 #ifdef CONFIG_IP_VS_PROTO_SCTP 935 #define SCTP_APP_TAB_BITS 4 936 #define SCTP_APP_TAB_SIZE (1 << SCTP_APP_TAB_BITS) 937 #define SCTP_APP_TAB_MASK (SCTP_APP_TAB_SIZE - 1) 938 /* Hash table for SCTP application incarnations */ 939 struct list_head sctp_apps[SCTP_APP_TAB_SIZE]; 940 #endif 941 /* ip_vs_conn */ 942 atomic_t conn_count; /* connection counter */ 943 944 /* ip_vs_ctl */ 945 struct ip_vs_stats tot_stats; /* Statistics & est. */ 946 947 int num_services; /* no of virtual services */ 948 949 /* Trash for destinations */ 950 struct list_head dest_trash; 951 spinlock_t dest_trash_lock; 952 struct timer_list dest_trash_timer; /* expiration timer */ 953 /* Service counters */ 954 atomic_t ftpsvc_counter; 955 atomic_t nullsvc_counter; 956 957 #ifdef CONFIG_SYSCTL 958 /* 1/rate drop and drop-entry variables */ 959 struct delayed_work defense_work; /* Work handler */ 960 int drop_rate; 961 int drop_counter; 962 atomic_t dropentry; 963 /* locks in ctl.c */ 964 spinlock_t dropentry_lock; /* drop entry handling */ 965 spinlock_t droppacket_lock; /* drop packet handling */ 966 spinlock_t securetcp_lock; /* state and timeout tables */ 967 968 /* sys-ctl struct */ 969 struct ctl_table_header *sysctl_hdr; 970 struct ctl_table *sysctl_tbl; 971 #endif 972 973 /* sysctl variables */ 974 int sysctl_amemthresh; 975 int sysctl_am_droprate; 976 int sysctl_drop_entry; 977 int sysctl_drop_packet; 978 int sysctl_secure_tcp; 979 #ifdef CONFIG_IP_VS_NFCT 980 int sysctl_conntrack; 981 #endif 982 int sysctl_snat_reroute; 983 int sysctl_sync_ver; 984 int sysctl_sync_ports; 985 int sysctl_sync_persist_mode; 986 unsigned long sysctl_sync_qlen_max; 987 int sysctl_sync_sock_size; 988 int sysctl_cache_bypass; 989 int sysctl_expire_nodest_conn; 990 int sysctl_sloppy_tcp; 991 int sysctl_sloppy_sctp; 992 int sysctl_expire_quiescent_template; 993 int sysctl_sync_threshold[2]; 994 unsigned int sysctl_sync_refresh_period; 995 int sysctl_sync_retries; 996 int sysctl_nat_icmp_send; 997 int sysctl_pmtu_disc; 998 int sysctl_backup_only; 999 int sysctl_conn_reuse_mode; 1000 int sysctl_schedule_icmp; 1001 int sysctl_ignore_tunneled; 1002 1003 /* ip_vs_lblc */ 1004 int sysctl_lblc_expiration; 1005 struct ctl_table_header *lblc_ctl_header; 1006 struct ctl_table *lblc_ctl_table; 1007 /* ip_vs_lblcr */ 1008 int sysctl_lblcr_expiration; 1009 struct ctl_table_header *lblcr_ctl_header; 1010 struct ctl_table *lblcr_ctl_table; 1011 /* ip_vs_est */ 1012 struct list_head est_list; /* estimator list */ 1013 spinlock_t est_lock; 1014 struct timer_list est_timer; /* Estimation timer */ 1015 /* ip_vs_sync */ 1016 spinlock_t sync_lock; 1017 struct ipvs_master_sync_state *ms; 1018 spinlock_t sync_buff_lock; 1019 struct task_struct **backup_threads; 1020 int threads_mask; 1021 volatile int sync_state; 1022 struct mutex sync_mutex; 1023 struct ipvs_sync_daemon_cfg mcfg; /* Master Configuration */ 1024 struct ipvs_sync_daemon_cfg bcfg; /* Backup Configuration */ 1025 /* net name space ptr */ 1026 struct net *net; /* Needed by timer routines */ 1027 /* Number of heterogeneous destinations, needed becaus heterogeneous 1028 * are not supported when synchronization is enabled. 1029 */ 1030 unsigned int mixed_address_family_dests; 1031 }; 1032 1033 #define DEFAULT_SYNC_THRESHOLD 3 1034 #define DEFAULT_SYNC_PERIOD 50 1035 #define DEFAULT_SYNC_VER 1 1036 #define DEFAULT_SLOPPY_TCP 0 1037 #define DEFAULT_SLOPPY_SCTP 0 1038 #define DEFAULT_SYNC_REFRESH_PERIOD (0U * HZ) 1039 #define DEFAULT_SYNC_RETRIES 0 1040 #define IPVS_SYNC_WAKEUP_RATE 8 1041 #define IPVS_SYNC_QLEN_MAX (IPVS_SYNC_WAKEUP_RATE * 4) 1042 #define IPVS_SYNC_SEND_DELAY (HZ / 50) 1043 #define IPVS_SYNC_CHECK_PERIOD HZ 1044 #define IPVS_SYNC_FLUSH_TIME (HZ * 2) 1045 #define IPVS_SYNC_PORTS_MAX (1 << 6) 1046 1047 #ifdef CONFIG_SYSCTL 1048 1049 static inline int sysctl_sync_threshold(struct netns_ipvs *ipvs) 1050 { 1051 return ipvs->sysctl_sync_threshold[0]; 1052 } 1053 1054 static inline int sysctl_sync_period(struct netns_ipvs *ipvs) 1055 { 1056 return ACCESS_ONCE(ipvs->sysctl_sync_threshold[1]); 1057 } 1058 1059 static inline unsigned int sysctl_sync_refresh_period(struct netns_ipvs *ipvs) 1060 { 1061 return ACCESS_ONCE(ipvs->sysctl_sync_refresh_period); 1062 } 1063 1064 static inline int sysctl_sync_retries(struct netns_ipvs *ipvs) 1065 { 1066 return ipvs->sysctl_sync_retries; 1067 } 1068 1069 static inline int sysctl_sync_ver(struct netns_ipvs *ipvs) 1070 { 1071 return ipvs->sysctl_sync_ver; 1072 } 1073 1074 static inline int sysctl_sloppy_tcp(struct netns_ipvs *ipvs) 1075 { 1076 return ipvs->sysctl_sloppy_tcp; 1077 } 1078 1079 static inline int sysctl_sloppy_sctp(struct netns_ipvs *ipvs) 1080 { 1081 return ipvs->sysctl_sloppy_sctp; 1082 } 1083 1084 static inline int sysctl_sync_ports(struct netns_ipvs *ipvs) 1085 { 1086 return ACCESS_ONCE(ipvs->sysctl_sync_ports); 1087 } 1088 1089 static inline int sysctl_sync_persist_mode(struct netns_ipvs *ipvs) 1090 { 1091 return ipvs->sysctl_sync_persist_mode; 1092 } 1093 1094 static inline unsigned long sysctl_sync_qlen_max(struct netns_ipvs *ipvs) 1095 { 1096 return ipvs->sysctl_sync_qlen_max; 1097 } 1098 1099 static inline int sysctl_sync_sock_size(struct netns_ipvs *ipvs) 1100 { 1101 return ipvs->sysctl_sync_sock_size; 1102 } 1103 1104 static inline int sysctl_pmtu_disc(struct netns_ipvs *ipvs) 1105 { 1106 return ipvs->sysctl_pmtu_disc; 1107 } 1108 1109 static inline int sysctl_backup_only(struct netns_ipvs *ipvs) 1110 { 1111 return ipvs->sync_state & IP_VS_STATE_BACKUP && 1112 ipvs->sysctl_backup_only; 1113 } 1114 1115 static inline int sysctl_conn_reuse_mode(struct netns_ipvs *ipvs) 1116 { 1117 return ipvs->sysctl_conn_reuse_mode; 1118 } 1119 1120 static inline int sysctl_schedule_icmp(struct netns_ipvs *ipvs) 1121 { 1122 return ipvs->sysctl_schedule_icmp; 1123 } 1124 1125 static inline int sysctl_ignore_tunneled(struct netns_ipvs *ipvs) 1126 { 1127 return ipvs->sysctl_ignore_tunneled; 1128 } 1129 1130 #else 1131 1132 static inline int sysctl_sync_threshold(struct netns_ipvs *ipvs) 1133 { 1134 return DEFAULT_SYNC_THRESHOLD; 1135 } 1136 1137 static inline int sysctl_sync_period(struct netns_ipvs *ipvs) 1138 { 1139 return DEFAULT_SYNC_PERIOD; 1140 } 1141 1142 static inline unsigned int sysctl_sync_refresh_period(struct netns_ipvs *ipvs) 1143 { 1144 return DEFAULT_SYNC_REFRESH_PERIOD; 1145 } 1146 1147 static inline int sysctl_sync_retries(struct netns_ipvs *ipvs) 1148 { 1149 return DEFAULT_SYNC_RETRIES & 3; 1150 } 1151 1152 static inline int sysctl_sync_ver(struct netns_ipvs *ipvs) 1153 { 1154 return DEFAULT_SYNC_VER; 1155 } 1156 1157 static inline int sysctl_sloppy_tcp(struct netns_ipvs *ipvs) 1158 { 1159 return DEFAULT_SLOPPY_TCP; 1160 } 1161 1162 static inline int sysctl_sloppy_sctp(struct netns_ipvs *ipvs) 1163 { 1164 return DEFAULT_SLOPPY_SCTP; 1165 } 1166 1167 static inline int sysctl_sync_ports(struct netns_ipvs *ipvs) 1168 { 1169 return 1; 1170 } 1171 1172 static inline int sysctl_sync_persist_mode(struct netns_ipvs *ipvs) 1173 { 1174 return 0; 1175 } 1176 1177 static inline unsigned long sysctl_sync_qlen_max(struct netns_ipvs *ipvs) 1178 { 1179 return IPVS_SYNC_QLEN_MAX; 1180 } 1181 1182 static inline int sysctl_sync_sock_size(struct netns_ipvs *ipvs) 1183 { 1184 return 0; 1185 } 1186 1187 static inline int sysctl_pmtu_disc(struct netns_ipvs *ipvs) 1188 { 1189 return 1; 1190 } 1191 1192 static inline int sysctl_backup_only(struct netns_ipvs *ipvs) 1193 { 1194 return 0; 1195 } 1196 1197 static inline int sysctl_conn_reuse_mode(struct netns_ipvs *ipvs) 1198 { 1199 return 1; 1200 } 1201 1202 static inline int sysctl_schedule_icmp(struct netns_ipvs *ipvs) 1203 { 1204 return 0; 1205 } 1206 1207 static inline int sysctl_ignore_tunneled(struct netns_ipvs *ipvs) 1208 { 1209 return 0; 1210 } 1211 1212 #endif 1213 1214 /* IPVS core functions 1215 * (from ip_vs_core.c) 1216 */ 1217 const char *ip_vs_proto_name(unsigned int proto); 1218 void ip_vs_init_hash_table(struct list_head *table, int rows); 1219 #define IP_VS_INIT_HASH_TABLE(t) ip_vs_init_hash_table((t), ARRAY_SIZE((t))) 1220 1221 #define IP_VS_APP_TYPE_FTP 1 1222 1223 /* ip_vs_conn handling functions 1224 * (from ip_vs_conn.c) 1225 */ 1226 enum { 1227 IP_VS_DIR_INPUT = 0, 1228 IP_VS_DIR_OUTPUT, 1229 IP_VS_DIR_INPUT_ONLY, 1230 IP_VS_DIR_LAST, 1231 }; 1232 1233 static inline void ip_vs_conn_fill_param(struct net *net, int af, int protocol, 1234 const union nf_inet_addr *caddr, 1235 __be16 cport, 1236 const union nf_inet_addr *vaddr, 1237 __be16 vport, 1238 struct ip_vs_conn_param *p) 1239 { 1240 p->net = net; 1241 p->af = af; 1242 p->protocol = protocol; 1243 p->caddr = caddr; 1244 p->cport = cport; 1245 p->vaddr = vaddr; 1246 p->vport = vport; 1247 p->pe = NULL; 1248 p->pe_data = NULL; 1249 } 1250 1251 struct ip_vs_conn *ip_vs_conn_in_get(const struct ip_vs_conn_param *p); 1252 struct ip_vs_conn *ip_vs_ct_in_get(const struct ip_vs_conn_param *p); 1253 1254 struct ip_vs_conn * ip_vs_conn_in_get_proto(int af, const struct sk_buff *skb, 1255 const struct ip_vs_iphdr *iph); 1256 1257 struct ip_vs_conn *ip_vs_conn_out_get(const struct ip_vs_conn_param *p); 1258 1259 struct ip_vs_conn * ip_vs_conn_out_get_proto(int af, const struct sk_buff *skb, 1260 const struct ip_vs_iphdr *iph); 1261 1262 /* Get reference to gain full access to conn. 1263 * By default, RCU read-side critical sections have access only to 1264 * conn fields and its PE data, see ip_vs_conn_rcu_free() for reference. 1265 */ 1266 static inline bool __ip_vs_conn_get(struct ip_vs_conn *cp) 1267 { 1268 return atomic_inc_not_zero(&cp->refcnt); 1269 } 1270 1271 /* put back the conn without restarting its timer */ 1272 static inline void __ip_vs_conn_put(struct ip_vs_conn *cp) 1273 { 1274 smp_mb__before_atomic(); 1275 atomic_dec(&cp->refcnt); 1276 } 1277 void ip_vs_conn_put(struct ip_vs_conn *cp); 1278 void ip_vs_conn_fill_cport(struct ip_vs_conn *cp, __be16 cport); 1279 1280 struct ip_vs_conn *ip_vs_conn_new(const struct ip_vs_conn_param *p, int dest_af, 1281 const union nf_inet_addr *daddr, 1282 __be16 dport, unsigned int flags, 1283 struct ip_vs_dest *dest, __u32 fwmark); 1284 void ip_vs_conn_expire_now(struct ip_vs_conn *cp); 1285 1286 const char *ip_vs_state_name(__u16 proto, int state); 1287 1288 void ip_vs_tcp_conn_listen(struct net *net, struct ip_vs_conn *cp); 1289 int ip_vs_check_template(struct ip_vs_conn *ct); 1290 void ip_vs_random_dropentry(struct net *net); 1291 int ip_vs_conn_init(void); 1292 void ip_vs_conn_cleanup(void); 1293 1294 static inline void ip_vs_control_del(struct ip_vs_conn *cp) 1295 { 1296 struct ip_vs_conn *ctl_cp = cp->control; 1297 if (!ctl_cp) { 1298 IP_VS_ERR_BUF("request control DEL for uncontrolled: " 1299 "%s:%d to %s:%d\n", 1300 IP_VS_DBG_ADDR(cp->af, &cp->caddr), 1301 ntohs(cp->cport), 1302 IP_VS_DBG_ADDR(cp->af, &cp->vaddr), 1303 ntohs(cp->vport)); 1304 1305 return; 1306 } 1307 1308 IP_VS_DBG_BUF(7, "DELeting control for: " 1309 "cp.dst=%s:%d ctl_cp.dst=%s:%d\n", 1310 IP_VS_DBG_ADDR(cp->af, &cp->caddr), 1311 ntohs(cp->cport), 1312 IP_VS_DBG_ADDR(cp->af, &ctl_cp->caddr), 1313 ntohs(ctl_cp->cport)); 1314 1315 cp->control = NULL; 1316 if (atomic_read(&ctl_cp->n_control) == 0) { 1317 IP_VS_ERR_BUF("BUG control DEL with n=0 : " 1318 "%s:%d to %s:%d\n", 1319 IP_VS_DBG_ADDR(cp->af, &cp->caddr), 1320 ntohs(cp->cport), 1321 IP_VS_DBG_ADDR(cp->af, &cp->vaddr), 1322 ntohs(cp->vport)); 1323 1324 return; 1325 } 1326 atomic_dec(&ctl_cp->n_control); 1327 } 1328 1329 static inline void 1330 ip_vs_control_add(struct ip_vs_conn *cp, struct ip_vs_conn *ctl_cp) 1331 { 1332 if (cp->control) { 1333 IP_VS_ERR_BUF("request control ADD for already controlled: " 1334 "%s:%d to %s:%d\n", 1335 IP_VS_DBG_ADDR(cp->af, &cp->caddr), 1336 ntohs(cp->cport), 1337 IP_VS_DBG_ADDR(cp->af, &cp->vaddr), 1338 ntohs(cp->vport)); 1339 1340 ip_vs_control_del(cp); 1341 } 1342 1343 IP_VS_DBG_BUF(7, "ADDing control for: " 1344 "cp.dst=%s:%d ctl_cp.dst=%s:%d\n", 1345 IP_VS_DBG_ADDR(cp->af, &cp->caddr), 1346 ntohs(cp->cport), 1347 IP_VS_DBG_ADDR(cp->af, &ctl_cp->caddr), 1348 ntohs(ctl_cp->cport)); 1349 1350 cp->control = ctl_cp; 1351 atomic_inc(&ctl_cp->n_control); 1352 } 1353 1354 /* IPVS netns init & cleanup functions */ 1355 int ip_vs_estimator_net_init(struct net *net); 1356 int ip_vs_control_net_init(struct net *net); 1357 int ip_vs_protocol_net_init(struct net *net); 1358 int ip_vs_app_net_init(struct net *net); 1359 int ip_vs_conn_net_init(struct net *net); 1360 int ip_vs_sync_net_init(struct net *net); 1361 void ip_vs_conn_net_cleanup(struct net *net); 1362 void ip_vs_app_net_cleanup(struct net *net); 1363 void ip_vs_protocol_net_cleanup(struct net *net); 1364 void ip_vs_control_net_cleanup(struct net *net); 1365 void ip_vs_estimator_net_cleanup(struct net *net); 1366 void ip_vs_sync_net_cleanup(struct net *net); 1367 void ip_vs_service_net_cleanup(struct net *net); 1368 1369 /* IPVS application functions 1370 * (from ip_vs_app.c) 1371 */ 1372 #define IP_VS_APP_MAX_PORTS 8 1373 struct ip_vs_app *register_ip_vs_app(struct net *net, struct ip_vs_app *app); 1374 void unregister_ip_vs_app(struct net *net, struct ip_vs_app *app); 1375 int ip_vs_bind_app(struct ip_vs_conn *cp, struct ip_vs_protocol *pp); 1376 void ip_vs_unbind_app(struct ip_vs_conn *cp); 1377 int register_ip_vs_app_inc(struct net *net, struct ip_vs_app *app, __u16 proto, 1378 __u16 port); 1379 int ip_vs_app_inc_get(struct ip_vs_app *inc); 1380 void ip_vs_app_inc_put(struct ip_vs_app *inc); 1381 1382 int ip_vs_app_pkt_out(struct ip_vs_conn *, struct sk_buff *skb); 1383 int ip_vs_app_pkt_in(struct ip_vs_conn *, struct sk_buff *skb); 1384 1385 int register_ip_vs_pe(struct ip_vs_pe *pe); 1386 int unregister_ip_vs_pe(struct ip_vs_pe *pe); 1387 struct ip_vs_pe *ip_vs_pe_getbyname(const char *name); 1388 struct ip_vs_pe *__ip_vs_pe_getbyname(const char *pe_name); 1389 1390 /* Use a #define to avoid all of module.h just for these trivial ops */ 1391 #define ip_vs_pe_get(pe) \ 1392 if (pe && pe->module) \ 1393 __module_get(pe->module); 1394 1395 #define ip_vs_pe_put(pe) \ 1396 if (pe && pe->module) \ 1397 module_put(pe->module); 1398 1399 /* IPVS protocol functions (from ip_vs_proto.c) */ 1400 int ip_vs_protocol_init(void); 1401 void ip_vs_protocol_cleanup(void); 1402 void ip_vs_protocol_timeout_change(struct netns_ipvs *ipvs, int flags); 1403 int *ip_vs_create_timeout_table(int *table, int size); 1404 int ip_vs_set_state_timeout(int *table, int num, const char *const *names, 1405 const char *name, int to); 1406 void ip_vs_tcpudp_debug_packet(int af, struct ip_vs_protocol *pp, 1407 const struct sk_buff *skb, int offset, 1408 const char *msg); 1409 1410 extern struct ip_vs_protocol ip_vs_protocol_tcp; 1411 extern struct ip_vs_protocol ip_vs_protocol_udp; 1412 extern struct ip_vs_protocol ip_vs_protocol_icmp; 1413 extern struct ip_vs_protocol ip_vs_protocol_esp; 1414 extern struct ip_vs_protocol ip_vs_protocol_ah; 1415 extern struct ip_vs_protocol ip_vs_protocol_sctp; 1416 1417 /* Registering/unregistering scheduler functions 1418 * (from ip_vs_sched.c) 1419 */ 1420 int register_ip_vs_scheduler(struct ip_vs_scheduler *scheduler); 1421 int unregister_ip_vs_scheduler(struct ip_vs_scheduler *scheduler); 1422 int ip_vs_bind_scheduler(struct ip_vs_service *svc, 1423 struct ip_vs_scheduler *scheduler); 1424 void ip_vs_unbind_scheduler(struct ip_vs_service *svc, 1425 struct ip_vs_scheduler *sched); 1426 struct ip_vs_scheduler *ip_vs_scheduler_get(const char *sched_name); 1427 void ip_vs_scheduler_put(struct ip_vs_scheduler *scheduler); 1428 struct ip_vs_conn * 1429 ip_vs_schedule(struct ip_vs_service *svc, struct sk_buff *skb, 1430 struct ip_vs_proto_data *pd, int *ignored, 1431 struct ip_vs_iphdr *iph); 1432 int ip_vs_leave(struct ip_vs_service *svc, struct sk_buff *skb, 1433 struct ip_vs_proto_data *pd, struct ip_vs_iphdr *iph); 1434 1435 void ip_vs_scheduler_err(struct ip_vs_service *svc, const char *msg); 1436 1437 /* IPVS control data and functions (from ip_vs_ctl.c) */ 1438 extern struct ip_vs_stats ip_vs_stats; 1439 extern int sysctl_ip_vs_sync_ver; 1440 1441 struct ip_vs_service * 1442 ip_vs_service_find(struct net *net, int af, __u32 fwmark, __u16 protocol, 1443 const union nf_inet_addr *vaddr, __be16 vport); 1444 1445 bool ip_vs_has_real_service(struct net *net, int af, __u16 protocol, 1446 const union nf_inet_addr *daddr, __be16 dport); 1447 1448 int ip_vs_use_count_inc(void); 1449 void ip_vs_use_count_dec(void); 1450 int ip_vs_register_nl_ioctl(void); 1451 void ip_vs_unregister_nl_ioctl(void); 1452 int ip_vs_control_init(void); 1453 void ip_vs_control_cleanup(void); 1454 struct ip_vs_dest * 1455 ip_vs_find_dest(struct net *net, int svc_af, int dest_af, 1456 const union nf_inet_addr *daddr, __be16 dport, 1457 const union nf_inet_addr *vaddr, __be16 vport, 1458 __u16 protocol, __u32 fwmark, __u32 flags); 1459 void ip_vs_try_bind_dest(struct ip_vs_conn *cp); 1460 1461 static inline void ip_vs_dest_hold(struct ip_vs_dest *dest) 1462 { 1463 atomic_inc(&dest->refcnt); 1464 } 1465 1466 static inline void ip_vs_dest_put(struct ip_vs_dest *dest) 1467 { 1468 smp_mb__before_atomic(); 1469 atomic_dec(&dest->refcnt); 1470 } 1471 1472 static inline void ip_vs_dest_put_and_free(struct ip_vs_dest *dest) 1473 { 1474 if (atomic_dec_return(&dest->refcnt) < 0) 1475 kfree(dest); 1476 } 1477 1478 /* IPVS sync daemon data and function prototypes 1479 * (from ip_vs_sync.c) 1480 */ 1481 int start_sync_thread(struct net *net, struct ipvs_sync_daemon_cfg *cfg, 1482 int state); 1483 int stop_sync_thread(struct net *net, int state); 1484 void ip_vs_sync_conn(struct net *net, struct ip_vs_conn *cp, int pkts); 1485 1486 /* IPVS rate estimator prototypes (from ip_vs_est.c) */ 1487 void ip_vs_start_estimator(struct net *net, struct ip_vs_stats *stats); 1488 void ip_vs_stop_estimator(struct net *net, struct ip_vs_stats *stats); 1489 void ip_vs_zero_estimator(struct ip_vs_stats *stats); 1490 void ip_vs_read_estimator(struct ip_vs_kstats *dst, struct ip_vs_stats *stats); 1491 1492 /* Various IPVS packet transmitters (from ip_vs_xmit.c) */ 1493 int ip_vs_null_xmit(struct sk_buff *skb, struct ip_vs_conn *cp, 1494 struct ip_vs_protocol *pp, struct ip_vs_iphdr *iph); 1495 int ip_vs_bypass_xmit(struct sk_buff *skb, struct ip_vs_conn *cp, 1496 struct ip_vs_protocol *pp, struct ip_vs_iphdr *iph); 1497 int ip_vs_nat_xmit(struct sk_buff *skb, struct ip_vs_conn *cp, 1498 struct ip_vs_protocol *pp, struct ip_vs_iphdr *iph); 1499 int ip_vs_tunnel_xmit(struct sk_buff *skb, struct ip_vs_conn *cp, 1500 struct ip_vs_protocol *pp, struct ip_vs_iphdr *iph); 1501 int ip_vs_dr_xmit(struct sk_buff *skb, struct ip_vs_conn *cp, 1502 struct ip_vs_protocol *pp, struct ip_vs_iphdr *iph); 1503 int ip_vs_icmp_xmit(struct sk_buff *skb, struct ip_vs_conn *cp, 1504 struct ip_vs_protocol *pp, int offset, 1505 unsigned int hooknum, struct ip_vs_iphdr *iph); 1506 void ip_vs_dest_dst_rcu_free(struct rcu_head *head); 1507 1508 #ifdef CONFIG_IP_VS_IPV6 1509 int ip_vs_bypass_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp, 1510 struct ip_vs_protocol *pp, struct ip_vs_iphdr *iph); 1511 int ip_vs_nat_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp, 1512 struct ip_vs_protocol *pp, struct ip_vs_iphdr *iph); 1513 int ip_vs_tunnel_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp, 1514 struct ip_vs_protocol *pp, struct ip_vs_iphdr *iph); 1515 int ip_vs_dr_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp, 1516 struct ip_vs_protocol *pp, struct ip_vs_iphdr *iph); 1517 int ip_vs_icmp_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp, 1518 struct ip_vs_protocol *pp, int offset, 1519 unsigned int hooknum, struct ip_vs_iphdr *iph); 1520 #endif 1521 1522 #ifdef CONFIG_SYSCTL 1523 /* This is a simple mechanism to ignore packets when 1524 * we are loaded. Just set ip_vs_drop_rate to 'n' and 1525 * we start to drop 1/rate of the packets 1526 */ 1527 static inline int ip_vs_todrop(struct netns_ipvs *ipvs) 1528 { 1529 if (!ipvs->drop_rate) 1530 return 0; 1531 if (--ipvs->drop_counter > 0) 1532 return 0; 1533 ipvs->drop_counter = ipvs->drop_rate; 1534 return 1; 1535 } 1536 #else 1537 static inline int ip_vs_todrop(struct netns_ipvs *ipvs) { return 0; } 1538 #endif 1539 1540 /* ip_vs_fwd_tag returns the forwarding tag of the connection */ 1541 #define IP_VS_FWD_METHOD(cp) (cp->flags & IP_VS_CONN_F_FWD_MASK) 1542 1543 static inline char ip_vs_fwd_tag(struct ip_vs_conn *cp) 1544 { 1545 char fwd; 1546 1547 switch (IP_VS_FWD_METHOD(cp)) { 1548 case IP_VS_CONN_F_MASQ: 1549 fwd = 'M'; break; 1550 case IP_VS_CONN_F_LOCALNODE: 1551 fwd = 'L'; break; 1552 case IP_VS_CONN_F_TUNNEL: 1553 fwd = 'T'; break; 1554 case IP_VS_CONN_F_DROUTE: 1555 fwd = 'R'; break; 1556 case IP_VS_CONN_F_BYPASS: 1557 fwd = 'B'; break; 1558 default: 1559 fwd = '?'; break; 1560 } 1561 return fwd; 1562 } 1563 1564 void ip_vs_nat_icmp(struct sk_buff *skb, struct ip_vs_protocol *pp, 1565 struct ip_vs_conn *cp, int dir); 1566 1567 #ifdef CONFIG_IP_VS_IPV6 1568 void ip_vs_nat_icmp_v6(struct sk_buff *skb, struct ip_vs_protocol *pp, 1569 struct ip_vs_conn *cp, int dir); 1570 #endif 1571 1572 __sum16 ip_vs_checksum_complete(struct sk_buff *skb, int offset); 1573 1574 static inline __wsum ip_vs_check_diff4(__be32 old, __be32 new, __wsum oldsum) 1575 { 1576 __be32 diff[2] = { ~old, new }; 1577 1578 return csum_partial(diff, sizeof(diff), oldsum); 1579 } 1580 1581 #ifdef CONFIG_IP_VS_IPV6 1582 static inline __wsum ip_vs_check_diff16(const __be32 *old, const __be32 *new, 1583 __wsum oldsum) 1584 { 1585 __be32 diff[8] = { ~old[3], ~old[2], ~old[1], ~old[0], 1586 new[3], new[2], new[1], new[0] }; 1587 1588 return csum_partial(diff, sizeof(diff), oldsum); 1589 } 1590 #endif 1591 1592 static inline __wsum ip_vs_check_diff2(__be16 old, __be16 new, __wsum oldsum) 1593 { 1594 __be16 diff[2] = { ~old, new }; 1595 1596 return csum_partial(diff, sizeof(diff), oldsum); 1597 } 1598 1599 /* Forget current conntrack (unconfirmed) and attach notrack entry */ 1600 static inline void ip_vs_notrack(struct sk_buff *skb) 1601 { 1602 #if defined(CONFIG_NF_CONNTRACK) || defined(CONFIG_NF_CONNTRACK_MODULE) 1603 enum ip_conntrack_info ctinfo; 1604 struct nf_conn *ct = nf_ct_get(skb, &ctinfo); 1605 1606 if (!ct || !nf_ct_is_untracked(ct)) { 1607 nf_conntrack_put(skb->nfct); 1608 skb->nfct = &nf_ct_untracked_get()->ct_general; 1609 skb->nfctinfo = IP_CT_NEW; 1610 nf_conntrack_get(skb->nfct); 1611 } 1612 #endif 1613 } 1614 1615 #ifdef CONFIG_IP_VS_NFCT 1616 /* Netfilter connection tracking 1617 * (from ip_vs_nfct.c) 1618 */ 1619 static inline int ip_vs_conntrack_enabled(struct netns_ipvs *ipvs) 1620 { 1621 #ifdef CONFIG_SYSCTL 1622 return ipvs->sysctl_conntrack; 1623 #else 1624 return 0; 1625 #endif 1626 } 1627 1628 void ip_vs_update_conntrack(struct sk_buff *skb, struct ip_vs_conn *cp, 1629 int outin); 1630 int ip_vs_confirm_conntrack(struct sk_buff *skb); 1631 void ip_vs_nfct_expect_related(struct sk_buff *skb, struct nf_conn *ct, 1632 struct ip_vs_conn *cp, u_int8_t proto, 1633 const __be16 port, int from_rs); 1634 void ip_vs_conn_drop_conntrack(struct ip_vs_conn *cp); 1635 1636 #else 1637 1638 static inline int ip_vs_conntrack_enabled(struct netns_ipvs *ipvs) 1639 { 1640 return 0; 1641 } 1642 1643 static inline void ip_vs_update_conntrack(struct sk_buff *skb, 1644 struct ip_vs_conn *cp, int outin) 1645 { 1646 } 1647 1648 static inline int ip_vs_confirm_conntrack(struct sk_buff *skb) 1649 { 1650 return NF_ACCEPT; 1651 } 1652 1653 static inline void ip_vs_conn_drop_conntrack(struct ip_vs_conn *cp) 1654 { 1655 } 1656 #endif /* CONFIG_IP_VS_NFCT */ 1657 1658 static inline int 1659 ip_vs_dest_conn_overhead(struct ip_vs_dest *dest) 1660 { 1661 /* We think the overhead of processing active connections is 256 1662 * times higher than that of inactive connections in average. (This 1663 * 256 times might not be accurate, we will change it later) We 1664 * use the following formula to estimate the overhead now: 1665 * dest->activeconns*256 + dest->inactconns 1666 */ 1667 return (atomic_read(&dest->activeconns) << 8) + 1668 atomic_read(&dest->inactconns); 1669 } 1670 1671 #endif /* _NET_IP_VS_H */ 1672