1 /* 2 * Copyright (C) 2017 Netronome Systems, Inc. 3 * 4 * This software is dual licensed under the GNU General License Version 2, 5 * June 1991 as shown in the file COPYING in the top-level directory of this 6 * source tree or the BSD 2-Clause License provided below. You have the 7 * option to license this software under the complete terms of either license. 8 * 9 * The BSD 2-Clause License: 10 * 11 * Redistribution and use in source and binary forms, with or 12 * without modification, are permitted provided that the following 13 * conditions are met: 14 * 15 * 1. Redistributions of source code must retain the above 16 * copyright notice, this list of conditions and the following 17 * disclaimer. 18 * 19 * 2. Redistributions in binary form must reproduce the above 20 * copyright notice, this list of conditions and the following 21 * disclaimer in the documentation and/or other materials 22 * provided with the distribution. 23 * 24 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 25 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 26 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 27 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 28 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 29 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 30 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 31 * SOFTWARE. 32 */ 33 34 #include <linux/bitfield.h> 35 #include <net/pkt_cls.h> 36 37 #include "cmsg.h" 38 #include "main.h" 39 40 static void 41 nfp_flower_compile_meta_tci(struct nfp_flower_meta_two *frame, 42 struct tc_cls_flower_offload *flow, u8 key_type, 43 bool mask_version) 44 { 45 struct fl_flow_key *target = mask_version ? flow->mask : flow->key; 46 struct flow_dissector_key_vlan *flow_vlan; 47 u16 tmp_tci; 48 49 memset(frame, 0, sizeof(struct nfp_flower_meta_two)); 50 /* Populate the metadata frame. */ 51 frame->nfp_flow_key_layer = key_type; 52 frame->mask_id = ~0; 53 54 if (dissector_uses_key(flow->dissector, FLOW_DISSECTOR_KEY_VLAN)) { 55 flow_vlan = skb_flow_dissector_target(flow->dissector, 56 FLOW_DISSECTOR_KEY_VLAN, 57 target); 58 /* Populate the tci field. */ 59 if (flow_vlan->vlan_id) { 60 tmp_tci = FIELD_PREP(NFP_FLOWER_MASK_VLAN_PRIO, 61 flow_vlan->vlan_priority) | 62 FIELD_PREP(NFP_FLOWER_MASK_VLAN_VID, 63 flow_vlan->vlan_id) | 64 NFP_FLOWER_MASK_VLAN_CFI; 65 frame->tci = cpu_to_be16(tmp_tci); 66 } 67 } 68 } 69 70 static void 71 nfp_flower_compile_meta(struct nfp_flower_meta_one *frame, u8 key_type) 72 { 73 frame->nfp_flow_key_layer = key_type; 74 frame->mask_id = 0; 75 frame->reserved = 0; 76 } 77 78 static int 79 nfp_flower_compile_port(struct nfp_flower_in_port *frame, u32 cmsg_port, 80 bool mask_version, enum nfp_flower_tun_type tun_type) 81 { 82 if (mask_version) { 83 frame->in_port = cpu_to_be32(~0); 84 return 0; 85 } 86 87 if (tun_type) 88 frame->in_port = cpu_to_be32(NFP_FL_PORT_TYPE_TUN | tun_type); 89 else 90 frame->in_port = cpu_to_be32(cmsg_port); 91 92 return 0; 93 } 94 95 static void 96 nfp_flower_compile_mac(struct nfp_flower_mac_mpls *frame, 97 struct tc_cls_flower_offload *flow, 98 bool mask_version) 99 { 100 struct fl_flow_key *target = mask_version ? flow->mask : flow->key; 101 struct flow_dissector_key_eth_addrs *addr; 102 103 memset(frame, 0, sizeof(struct nfp_flower_mac_mpls)); 104 105 if (dissector_uses_key(flow->dissector, FLOW_DISSECTOR_KEY_ETH_ADDRS)) { 106 addr = skb_flow_dissector_target(flow->dissector, 107 FLOW_DISSECTOR_KEY_ETH_ADDRS, 108 target); 109 /* Populate mac frame. */ 110 ether_addr_copy(frame->mac_dst, &addr->dst[0]); 111 ether_addr_copy(frame->mac_src, &addr->src[0]); 112 } 113 114 if (mask_version) 115 frame->mpls_lse = cpu_to_be32(~0); 116 } 117 118 static void 119 nfp_flower_compile_tport(struct nfp_flower_tp_ports *frame, 120 struct tc_cls_flower_offload *flow, 121 bool mask_version) 122 { 123 struct fl_flow_key *target = mask_version ? flow->mask : flow->key; 124 struct flow_dissector_key_ports *tp; 125 126 memset(frame, 0, sizeof(struct nfp_flower_tp_ports)); 127 128 if (dissector_uses_key(flow->dissector, FLOW_DISSECTOR_KEY_PORTS)) { 129 tp = skb_flow_dissector_target(flow->dissector, 130 FLOW_DISSECTOR_KEY_PORTS, 131 target); 132 frame->port_src = tp->src; 133 frame->port_dst = tp->dst; 134 } 135 } 136 137 static void 138 nfp_flower_compile_ipv4(struct nfp_flower_ipv4 *frame, 139 struct tc_cls_flower_offload *flow, 140 bool mask_version) 141 { 142 struct fl_flow_key *target = mask_version ? flow->mask : flow->key; 143 struct flow_dissector_key_ipv4_addrs *addr; 144 struct flow_dissector_key_basic *basic; 145 146 /* Wildcard TOS/TTL for now. */ 147 memset(frame, 0, sizeof(struct nfp_flower_ipv4)); 148 149 if (dissector_uses_key(flow->dissector, 150 FLOW_DISSECTOR_KEY_IPV4_ADDRS)) { 151 addr = skb_flow_dissector_target(flow->dissector, 152 FLOW_DISSECTOR_KEY_IPV4_ADDRS, 153 target); 154 frame->ipv4_src = addr->src; 155 frame->ipv4_dst = addr->dst; 156 } 157 158 if (dissector_uses_key(flow->dissector, FLOW_DISSECTOR_KEY_BASIC)) { 159 basic = skb_flow_dissector_target(flow->dissector, 160 FLOW_DISSECTOR_KEY_BASIC, 161 target); 162 frame->proto = basic->ip_proto; 163 } 164 } 165 166 static void 167 nfp_flower_compile_ipv6(struct nfp_flower_ipv6 *frame, 168 struct tc_cls_flower_offload *flow, 169 bool mask_version) 170 { 171 struct fl_flow_key *target = mask_version ? flow->mask : flow->key; 172 struct flow_dissector_key_ipv6_addrs *addr; 173 struct flow_dissector_key_basic *basic; 174 175 /* Wildcard LABEL/TOS/TTL for now. */ 176 memset(frame, 0, sizeof(struct nfp_flower_ipv6)); 177 178 if (dissector_uses_key(flow->dissector, 179 FLOW_DISSECTOR_KEY_IPV6_ADDRS)) { 180 addr = skb_flow_dissector_target(flow->dissector, 181 FLOW_DISSECTOR_KEY_IPV6_ADDRS, 182 target); 183 frame->ipv6_src = addr->src; 184 frame->ipv6_dst = addr->dst; 185 } 186 187 if (dissector_uses_key(flow->dissector, FLOW_DISSECTOR_KEY_BASIC)) { 188 basic = skb_flow_dissector_target(flow->dissector, 189 FLOW_DISSECTOR_KEY_BASIC, 190 target); 191 frame->proto = basic->ip_proto; 192 } 193 } 194 195 static void 196 nfp_flower_compile_vxlan(struct nfp_flower_vxlan *frame, 197 struct tc_cls_flower_offload *flow, 198 bool mask_version, __be32 *tun_dst) 199 { 200 struct fl_flow_key *target = mask_version ? flow->mask : flow->key; 201 struct flow_dissector_key_ipv4_addrs *vxlan_ips; 202 struct flow_dissector_key_keyid *vni; 203 204 /* Wildcard TOS/TTL/GPE_FLAGS/NXT_PROTO for now. */ 205 memset(frame, 0, sizeof(struct nfp_flower_vxlan)); 206 207 if (dissector_uses_key(flow->dissector, 208 FLOW_DISSECTOR_KEY_ENC_KEYID)) { 209 u32 temp_vni; 210 211 vni = skb_flow_dissector_target(flow->dissector, 212 FLOW_DISSECTOR_KEY_ENC_KEYID, 213 target); 214 temp_vni = be32_to_cpu(vni->keyid) << NFP_FL_TUN_VNI_OFFSET; 215 frame->tun_id = cpu_to_be32(temp_vni); 216 } 217 218 if (dissector_uses_key(flow->dissector, 219 FLOW_DISSECTOR_KEY_ENC_IPV4_ADDRS)) { 220 vxlan_ips = 221 skb_flow_dissector_target(flow->dissector, 222 FLOW_DISSECTOR_KEY_ENC_IPV4_ADDRS, 223 target); 224 frame->ip_src = vxlan_ips->src; 225 frame->ip_dst = vxlan_ips->dst; 226 *tun_dst = vxlan_ips->dst; 227 } 228 } 229 230 int nfp_flower_compile_flow_match(struct tc_cls_flower_offload *flow, 231 struct nfp_fl_key_ls *key_ls, 232 struct net_device *netdev, 233 struct nfp_fl_payload *nfp_flow) 234 { 235 enum nfp_flower_tun_type tun_type = NFP_FL_TUNNEL_NONE; 236 __be32 tun_dst, tun_dst_mask = 0; 237 struct nfp_repr *netdev_repr; 238 int err; 239 u8 *ext; 240 u8 *msk; 241 242 if (key_ls->key_layer & NFP_FLOWER_LAYER_VXLAN) 243 tun_type = NFP_FL_TUNNEL_VXLAN; 244 245 memset(nfp_flow->unmasked_data, 0, key_ls->key_size); 246 memset(nfp_flow->mask_data, 0, key_ls->key_size); 247 248 ext = nfp_flow->unmasked_data; 249 msk = nfp_flow->mask_data; 250 if (NFP_FLOWER_LAYER_PORT & key_ls->key_layer) { 251 /* Populate Exact Metadata. */ 252 nfp_flower_compile_meta_tci((struct nfp_flower_meta_two *)ext, 253 flow, key_ls->key_layer, false); 254 /* Populate Mask Metadata. */ 255 nfp_flower_compile_meta_tci((struct nfp_flower_meta_two *)msk, 256 flow, key_ls->key_layer, true); 257 ext += sizeof(struct nfp_flower_meta_two); 258 msk += sizeof(struct nfp_flower_meta_two); 259 260 /* Populate Exact Port data. */ 261 err = nfp_flower_compile_port((struct nfp_flower_in_port *)ext, 262 nfp_repr_get_port_id(netdev), 263 false, tun_type); 264 if (err) 265 return err; 266 267 /* Populate Mask Port Data. */ 268 err = nfp_flower_compile_port((struct nfp_flower_in_port *)msk, 269 nfp_repr_get_port_id(netdev), 270 true, tun_type); 271 if (err) 272 return err; 273 274 ext += sizeof(struct nfp_flower_in_port); 275 msk += sizeof(struct nfp_flower_in_port); 276 } else { 277 /* Populate Exact Metadata. */ 278 nfp_flower_compile_meta((struct nfp_flower_meta_one *)ext, 279 key_ls->key_layer); 280 /* Populate Mask Metadata. */ 281 nfp_flower_compile_meta((struct nfp_flower_meta_one *)msk, 282 key_ls->key_layer); 283 ext += sizeof(struct nfp_flower_meta_one); 284 msk += sizeof(struct nfp_flower_meta_one); 285 } 286 287 if (NFP_FLOWER_LAYER_META & key_ls->key_layer) { 288 /* Additional Metadata Fields. 289 * Currently unsupported. 290 */ 291 return -EOPNOTSUPP; 292 } 293 294 if (NFP_FLOWER_LAYER_MAC & key_ls->key_layer) { 295 /* Populate Exact MAC Data. */ 296 nfp_flower_compile_mac((struct nfp_flower_mac_mpls *)ext, 297 flow, false); 298 /* Populate Mask MAC Data. */ 299 nfp_flower_compile_mac((struct nfp_flower_mac_mpls *)msk, 300 flow, true); 301 ext += sizeof(struct nfp_flower_mac_mpls); 302 msk += sizeof(struct nfp_flower_mac_mpls); 303 } 304 305 if (NFP_FLOWER_LAYER_TP & key_ls->key_layer) { 306 /* Populate Exact TP Data. */ 307 nfp_flower_compile_tport((struct nfp_flower_tp_ports *)ext, 308 flow, false); 309 /* Populate Mask TP Data. */ 310 nfp_flower_compile_tport((struct nfp_flower_tp_ports *)msk, 311 flow, true); 312 ext += sizeof(struct nfp_flower_tp_ports); 313 msk += sizeof(struct nfp_flower_tp_ports); 314 } 315 316 if (NFP_FLOWER_LAYER_IPV4 & key_ls->key_layer) { 317 /* Populate Exact IPv4 Data. */ 318 nfp_flower_compile_ipv4((struct nfp_flower_ipv4 *)ext, 319 flow, false); 320 /* Populate Mask IPv4 Data. */ 321 nfp_flower_compile_ipv4((struct nfp_flower_ipv4 *)msk, 322 flow, true); 323 ext += sizeof(struct nfp_flower_ipv4); 324 msk += sizeof(struct nfp_flower_ipv4); 325 } 326 327 if (NFP_FLOWER_LAYER_IPV6 & key_ls->key_layer) { 328 /* Populate Exact IPv4 Data. */ 329 nfp_flower_compile_ipv6((struct nfp_flower_ipv6 *)ext, 330 flow, false); 331 /* Populate Mask IPv4 Data. */ 332 nfp_flower_compile_ipv6((struct nfp_flower_ipv6 *)msk, 333 flow, true); 334 ext += sizeof(struct nfp_flower_ipv6); 335 msk += sizeof(struct nfp_flower_ipv6); 336 } 337 338 if (key_ls->key_layer & NFP_FLOWER_LAYER_VXLAN) { 339 /* Populate Exact VXLAN Data. */ 340 nfp_flower_compile_vxlan((struct nfp_flower_vxlan *)ext, 341 flow, false, &tun_dst); 342 /* Populate Mask VXLAN Data. */ 343 nfp_flower_compile_vxlan((struct nfp_flower_vxlan *)msk, 344 flow, true, &tun_dst_mask); 345 ext += sizeof(struct nfp_flower_vxlan); 346 msk += sizeof(struct nfp_flower_vxlan); 347 348 /* Configure tunnel end point MAC. */ 349 if (nfp_netdev_is_nfp_repr(netdev)) { 350 netdev_repr = netdev_priv(netdev); 351 nfp_tunnel_write_macs(netdev_repr->app); 352 353 /* Store the tunnel destination in the rule data. 354 * This must be present and be an exact match. 355 */ 356 nfp_flow->nfp_tun_ipv4_addr = tun_dst; 357 nfp_tunnel_add_ipv4_off(netdev_repr->app, tun_dst); 358 } 359 } 360 361 return 0; 362 } 363