| 1 | #ifndef _NET_FLOW_OFFLOAD_H |
| 2 | #define _NET_FLOW_OFFLOAD_H |
| 3 | |
| 4 | #include <linux/kernel.h> |
| 5 | #include <linux/list.h> |
| 6 | #include <linux/netlink.h> |
| 7 | #include <net/flow_dissector.h> |
| 8 | |
| 9 | struct flow_match { |
| 10 | struct flow_dissector *dissector; |
| 11 | void *mask; |
| 12 | void *key; |
| 13 | }; |
| 14 | |
| 15 | struct flow_match_meta { |
| 16 | struct flow_dissector_key_meta *key, *mask; |
| 17 | }; |
| 18 | |
| 19 | struct flow_match_basic { |
| 20 | struct flow_dissector_key_basic *key, *mask; |
| 21 | }; |
| 22 | |
| 23 | struct flow_match_control { |
| 24 | struct flow_dissector_key_control *key, *mask; |
| 25 | }; |
| 26 | |
| 27 | struct flow_match_eth_addrs { |
| 28 | struct flow_dissector_key_eth_addrs *key, *mask; |
| 29 | }; |
| 30 | |
| 31 | struct flow_match_vlan { |
| 32 | struct flow_dissector_key_vlan *key, *mask; |
| 33 | }; |
| 34 | |
| 35 | struct flow_match_arp { |
| 36 | struct flow_dissector_key_arp *key, *mask; |
| 37 | }; |
| 38 | |
| 39 | struct flow_match_ipv4_addrs { |
| 40 | struct flow_dissector_key_ipv4_addrs *key, *mask; |
| 41 | }; |
| 42 | |
| 43 | struct flow_match_ipv6_addrs { |
| 44 | struct flow_dissector_key_ipv6_addrs *key, *mask; |
| 45 | }; |
| 46 | |
| 47 | struct flow_match_ip { |
| 48 | struct flow_dissector_key_ip *key, *mask; |
| 49 | }; |
| 50 | |
| 51 | struct flow_match_ports { |
| 52 | struct flow_dissector_key_ports *key, *mask; |
| 53 | }; |
| 54 | |
| 55 | struct flow_match_ports_range { |
| 56 | struct flow_dissector_key_ports_range *key, *mask; |
| 57 | }; |
| 58 | |
| 59 | struct flow_match_icmp { |
| 60 | struct flow_dissector_key_icmp *key, *mask; |
| 61 | }; |
| 62 | |
| 63 | struct flow_match_tcp { |
| 64 | struct flow_dissector_key_tcp *key, *mask; |
| 65 | }; |
| 66 | |
| 67 | struct flow_match_ipsec { |
| 68 | struct flow_dissector_key_ipsec *key, *mask; |
| 69 | }; |
| 70 | |
| 71 | struct flow_match_mpls { |
| 72 | struct flow_dissector_key_mpls *key, *mask; |
| 73 | }; |
| 74 | |
| 75 | struct flow_match_enc_keyid { |
| 76 | struct flow_dissector_key_keyid *key, *mask; |
| 77 | }; |
| 78 | |
| 79 | struct flow_match_enc_opts { |
| 80 | struct flow_dissector_key_enc_opts *key, *mask; |
| 81 | }; |
| 82 | |
| 83 | struct flow_match_ct { |
| 84 | struct flow_dissector_key_ct *key, *mask; |
| 85 | }; |
| 86 | |
| 87 | struct flow_match_pppoe { |
| 88 | struct flow_dissector_key_pppoe *key, *mask; |
| 89 | }; |
| 90 | |
| 91 | struct flow_match_l2tpv3 { |
| 92 | struct flow_dissector_key_l2tpv3 *key, *mask; |
| 93 | }; |
| 94 | |
| 95 | struct flow_rule; |
| 96 | |
| 97 | void flow_rule_match_meta(const struct flow_rule *rule, |
| 98 | struct flow_match_meta *out); |
| 99 | void flow_rule_match_basic(const struct flow_rule *rule, |
| 100 | struct flow_match_basic *out); |
| 101 | void flow_rule_match_control(const struct flow_rule *rule, |
| 102 | struct flow_match_control *out); |
| 103 | void flow_rule_match_eth_addrs(const struct flow_rule *rule, |
| 104 | struct flow_match_eth_addrs *out); |
| 105 | void flow_rule_match_vlan(const struct flow_rule *rule, |
| 106 | struct flow_match_vlan *out); |
| 107 | void flow_rule_match_cvlan(const struct flow_rule *rule, |
| 108 | struct flow_match_vlan *out); |
| 109 | void flow_rule_match_arp(const struct flow_rule *rule, |
| 110 | struct flow_match_arp *out); |
| 111 | void flow_rule_match_ipv4_addrs(const struct flow_rule *rule, |
| 112 | struct flow_match_ipv4_addrs *out); |
| 113 | void flow_rule_match_ipv6_addrs(const struct flow_rule *rule, |
| 114 | struct flow_match_ipv6_addrs *out); |
| 115 | void flow_rule_match_ip(const struct flow_rule *rule, |
| 116 | struct flow_match_ip *out); |
| 117 | void flow_rule_match_ports(const struct flow_rule *rule, |
| 118 | struct flow_match_ports *out); |
| 119 | void flow_rule_match_ports_range(const struct flow_rule *rule, |
| 120 | struct flow_match_ports_range *out); |
| 121 | void flow_rule_match_tcp(const struct flow_rule *rule, |
| 122 | struct flow_match_tcp *out); |
| 123 | void flow_rule_match_ipsec(const struct flow_rule *rule, |
| 124 | struct flow_match_ipsec *out); |
| 125 | void flow_rule_match_icmp(const struct flow_rule *rule, |
| 126 | struct flow_match_icmp *out); |
| 127 | void flow_rule_match_mpls(const struct flow_rule *rule, |
| 128 | struct flow_match_mpls *out); |
| 129 | void flow_rule_match_enc_control(const struct flow_rule *rule, |
| 130 | struct flow_match_control *out); |
| 131 | void flow_rule_match_enc_ipv4_addrs(const struct flow_rule *rule, |
| 132 | struct flow_match_ipv4_addrs *out); |
| 133 | void flow_rule_match_enc_ipv6_addrs(const struct flow_rule *rule, |
| 134 | struct flow_match_ipv6_addrs *out); |
| 135 | void flow_rule_match_enc_ip(const struct flow_rule *rule, |
| 136 | struct flow_match_ip *out); |
| 137 | void flow_rule_match_enc_ports(const struct flow_rule *rule, |
| 138 | struct flow_match_ports *out); |
| 139 | void flow_rule_match_enc_keyid(const struct flow_rule *rule, |
| 140 | struct flow_match_enc_keyid *out); |
| 141 | void flow_rule_match_enc_opts(const struct flow_rule *rule, |
| 142 | struct flow_match_enc_opts *out); |
| 143 | void flow_rule_match_ct(const struct flow_rule *rule, |
| 144 | struct flow_match_ct *out); |
| 145 | void flow_rule_match_pppoe(const struct flow_rule *rule, |
| 146 | struct flow_match_pppoe *out); |
| 147 | void flow_rule_match_l2tpv3(const struct flow_rule *rule, |
| 148 | struct flow_match_l2tpv3 *out); |
| 149 | |
| 150 | enum flow_action_id { |
| 151 | FLOW_ACTION_ACCEPT = 0, |
| 152 | FLOW_ACTION_DROP, |
| 153 | FLOW_ACTION_TRAP, |
| 154 | FLOW_ACTION_GOTO, |
| 155 | FLOW_ACTION_REDIRECT, |
| 156 | FLOW_ACTION_MIRRED, |
| 157 | FLOW_ACTION_REDIRECT_INGRESS, |
| 158 | FLOW_ACTION_MIRRED_INGRESS, |
| 159 | FLOW_ACTION_VLAN_PUSH, |
| 160 | FLOW_ACTION_VLAN_POP, |
| 161 | FLOW_ACTION_VLAN_MANGLE, |
| 162 | FLOW_ACTION_TUNNEL_ENCAP, |
| 163 | FLOW_ACTION_TUNNEL_DECAP, |
| 164 | FLOW_ACTION_MANGLE, |
| 165 | FLOW_ACTION_ADD, |
| 166 | FLOW_ACTION_CSUM, |
| 167 | FLOW_ACTION_MARK, |
| 168 | FLOW_ACTION_PTYPE, |
| 169 | FLOW_ACTION_PRIORITY, |
| 170 | FLOW_ACTION_RX_QUEUE_MAPPING, |
| 171 | FLOW_ACTION_WAKE, |
| 172 | FLOW_ACTION_QUEUE, |
| 173 | FLOW_ACTION_SAMPLE, |
| 174 | FLOW_ACTION_POLICE, |
| 175 | FLOW_ACTION_CT, |
| 176 | FLOW_ACTION_CT_METADATA, |
| 177 | FLOW_ACTION_MPLS_PUSH, |
| 178 | FLOW_ACTION_MPLS_POP, |
| 179 | FLOW_ACTION_MPLS_MANGLE, |
| 180 | FLOW_ACTION_GATE, |
| 181 | FLOW_ACTION_PPPOE_PUSH, |
| 182 | FLOW_ACTION_JUMP, |
| 183 | FLOW_ACTION_PIPE, |
| 184 | FLOW_ACTION_VLAN_PUSH_ETH, |
| 185 | FLOW_ACTION_VLAN_POP_ETH, |
| 186 | FLOW_ACTION_CONTINUE, |
| 187 | NUM_FLOW_ACTIONS, |
| 188 | }; |
| 189 | |
| 190 | /* This is mirroring enum pedit_header_type definition for easy mapping between |
| 191 | * tc pedit action. Legacy TCA_PEDIT_KEY_EX_HDR_TYPE_NETWORK is mapped to |
| 192 | * FLOW_ACT_MANGLE_UNSPEC, which is supported by no driver. |
| 193 | */ |
| 194 | enum flow_action_mangle_base { |
| 195 | FLOW_ACT_MANGLE_UNSPEC = 0, |
| 196 | FLOW_ACT_MANGLE_HDR_TYPE_ETH, |
| 197 | FLOW_ACT_MANGLE_HDR_TYPE_IP4, |
| 198 | FLOW_ACT_MANGLE_HDR_TYPE_IP6, |
| 199 | FLOW_ACT_MANGLE_HDR_TYPE_TCP, |
| 200 | FLOW_ACT_MANGLE_HDR_TYPE_UDP, |
| 201 | }; |
| 202 | |
| 203 | enum flow_action_hw_stats_bit { |
| 204 | FLOW_ACTION_HW_STATS_IMMEDIATE_BIT, |
| 205 | FLOW_ACTION_HW_STATS_DELAYED_BIT, |
| 206 | FLOW_ACTION_HW_STATS_DISABLED_BIT, |
| 207 | |
| 208 | FLOW_ACTION_HW_STATS_NUM_BITS |
| 209 | }; |
| 210 | |
| 211 | enum flow_action_hw_stats { |
| 212 | FLOW_ACTION_HW_STATS_IMMEDIATE = |
| 213 | BIT(FLOW_ACTION_HW_STATS_IMMEDIATE_BIT), |
| 214 | FLOW_ACTION_HW_STATS_DELAYED = BIT(FLOW_ACTION_HW_STATS_DELAYED_BIT), |
| 215 | FLOW_ACTION_HW_STATS_ANY = FLOW_ACTION_HW_STATS_IMMEDIATE | |
| 216 | FLOW_ACTION_HW_STATS_DELAYED, |
| 217 | FLOW_ACTION_HW_STATS_DISABLED = |
| 218 | BIT(FLOW_ACTION_HW_STATS_DISABLED_BIT), |
| 219 | FLOW_ACTION_HW_STATS_DONT_CARE = BIT(FLOW_ACTION_HW_STATS_NUM_BITS) - 1, |
| 220 | }; |
| 221 | |
| 222 | typedef void (*action_destr)(void *priv); |
| 223 | |
| 224 | struct flow_action_cookie { |
| 225 | u32 cookie_len; |
| 226 | u8 cookie[]; |
| 227 | }; |
| 228 | |
| 229 | struct flow_action_cookie *flow_action_cookie_create(void *data, |
| 230 | unsigned int len, |
| 231 | gfp_t gfp); |
| 232 | void flow_action_cookie_destroy(struct flow_action_cookie *cookie); |
| 233 | |
| 234 | struct flow_action_entry { |
| 235 | enum flow_action_id id; |
| 236 | u32 hw_index; |
| 237 | unsigned long cookie; |
| 238 | u64 miss_cookie; |
| 239 | enum flow_action_hw_stats hw_stats; |
| 240 | action_destr destructor; |
| 241 | void *destructor_priv; |
| 242 | union { |
| 243 | u32 chain_index; /* FLOW_ACTION_GOTO */ |
| 244 | struct net_device *dev; /* FLOW_ACTION_REDIRECT */ |
| 245 | struct { /* FLOW_ACTION_VLAN */ |
| 246 | u16 vid; |
| 247 | __be16 proto; |
| 248 | u8 prio; |
| 249 | } vlan; |
| 250 | struct { /* FLOW_ACTION_VLAN_PUSH_ETH */ |
| 251 | unsigned char dst[ETH_ALEN]; |
| 252 | unsigned char src[ETH_ALEN]; |
| 253 | } vlan_push_eth; |
| 254 | struct { /* FLOW_ACTION_MANGLE */ |
| 255 | /* FLOW_ACTION_ADD */ |
| 256 | enum flow_action_mangle_base htype; |
| 257 | u32 offset; |
| 258 | u32 mask; |
| 259 | u32 val; |
| 260 | } mangle; |
| 261 | struct ip_tunnel_info *tunnel; /* FLOW_ACTION_TUNNEL_ENCAP */ |
| 262 | u32 csum_flags; /* FLOW_ACTION_CSUM */ |
| 263 | u32 mark; /* FLOW_ACTION_MARK */ |
| 264 | u16 ptype; /* FLOW_ACTION_PTYPE */ |
| 265 | u16 rx_queue; /* FLOW_ACTION_RX_QUEUE_MAPPING */ |
| 266 | u32 priority; /* FLOW_ACTION_PRIORITY */ |
| 267 | struct { /* FLOW_ACTION_QUEUE */ |
| 268 | u32 ctx; |
| 269 | u32 index; |
| 270 | u8 vf; |
| 271 | } queue; |
| 272 | struct { /* FLOW_ACTION_SAMPLE */ |
| 273 | struct psample_group *psample_group; |
| 274 | u32 rate; |
| 275 | u32 trunc_size; |
| 276 | bool truncate; |
| 277 | } sample; |
| 278 | struct { /* FLOW_ACTION_POLICE */ |
| 279 | u32 burst; |
| 280 | u64 rate_bytes_ps; |
| 281 | u64 peakrate_bytes_ps; |
| 282 | u32 avrate; |
| 283 | u16 overhead; |
| 284 | u64 burst_pkt; |
| 285 | u64 rate_pkt_ps; |
| 286 | u32 mtu; |
| 287 | struct { |
| 288 | enum flow_action_id act_id; |
| 289 | u32 extval; |
| 290 | } exceed, notexceed; |
| 291 | } police; |
| 292 | struct { /* FLOW_ACTION_CT */ |
| 293 | int action; |
| 294 | u16 zone; |
| 295 | struct nf_flowtable *flow_table; |
| 296 | } ct; |
| 297 | struct { |
| 298 | unsigned long cookie; |
| 299 | u32 mark; |
| 300 | u32 labels[4]; |
| 301 | bool orig_dir; |
| 302 | } ct_metadata; |
| 303 | struct { /* FLOW_ACTION_MPLS_PUSH */ |
| 304 | u32 label; |
| 305 | __be16 proto; |
| 306 | u8 tc; |
| 307 | u8 bos; |
| 308 | u8 ttl; |
| 309 | } mpls_push; |
| 310 | struct { /* FLOW_ACTION_MPLS_POP */ |
| 311 | __be16 proto; |
| 312 | } mpls_pop; |
| 313 | struct { /* FLOW_ACTION_MPLS_MANGLE */ |
| 314 | u32 label; |
| 315 | u8 tc; |
| 316 | u8 bos; |
| 317 | u8 ttl; |
| 318 | } mpls_mangle; |
| 319 | struct { |
| 320 | s32 prio; |
| 321 | u64 basetime; |
| 322 | u64 cycletime; |
| 323 | u64 cycletimeext; |
| 324 | u32 num_entries; |
| 325 | struct action_gate_entry *entries; |
| 326 | } gate; |
| 327 | struct { /* FLOW_ACTION_PPPOE_PUSH */ |
| 328 | u16 sid; |
| 329 | } pppoe; |
| 330 | }; |
| 331 | struct flow_action_cookie *user_cookie; /* user defined action cookie */ |
| 332 | }; |
| 333 | |
| 334 | struct flow_action { |
| 335 | unsigned int num_entries; |
| 336 | struct flow_action_entry entries[] __counted_by(num_entries); |
| 337 | }; |
| 338 | |
| 339 | static inline bool flow_action_has_entries(const struct flow_action *action) |
| 340 | { |
| 341 | return action->num_entries; |
| 342 | } |
| 343 | |
| 344 | /** |
| 345 | * flow_offload_has_one_action() - check if exactly one action is present |
| 346 | * @action: tc filter flow offload action |
| 347 | * |
| 348 | * Return: true if exactly one action is present. |
| 349 | */ |
| 350 | static inline bool flow_offload_has_one_action(const struct flow_action *action) |
| 351 | { |
| 352 | return action->num_entries == 1; |
| 353 | } |
| 354 | |
| 355 | static inline bool flow_action_is_last_entry(const struct flow_action *action, |
| 356 | const struct flow_action_entry *entry) |
| 357 | { |
| 358 | return entry == &action->entries[action->num_entries - 1]; |
| 359 | } |
| 360 | |
| 361 | #define flow_action_for_each(__i, __act, __actions) \ |
| 362 | for (__i = 0, __act = &(__actions)->entries[0]; \ |
| 363 | __i < (__actions)->num_entries; \ |
| 364 | __act = &(__actions)->entries[++__i]) |
| 365 | |
| 366 | static inline bool |
| 367 | flow_action_mixed_hw_stats_check(const struct flow_action *action, |
| 368 | struct netlink_ext_ack *extack) |
| 369 | { |
| 370 | const struct flow_action_entry *action_entry; |
| 371 | u8 last_hw_stats; |
| 372 | int i; |
| 373 | |
| 374 | if (flow_offload_has_one_action(action)) |
| 375 | return true; |
| 376 | |
| 377 | flow_action_for_each(i, action_entry, action) { |
| 378 | if (i && action_entry->hw_stats != last_hw_stats) { |
| 379 | NL_SET_ERR_MSG_MOD(extack, "Mixing HW stats types for actions is not supported" ); |
| 380 | return false; |
| 381 | } |
| 382 | last_hw_stats = action_entry->hw_stats; |
| 383 | } |
| 384 | return true; |
| 385 | } |
| 386 | |
| 387 | static inline const struct flow_action_entry * |
| 388 | flow_action_first_entry_get(const struct flow_action *action) |
| 389 | { |
| 390 | WARN_ON(!flow_action_has_entries(action)); |
| 391 | return &action->entries[0]; |
| 392 | } |
| 393 | |
| 394 | static inline bool |
| 395 | __flow_action_hw_stats_check(const struct flow_action *action, |
| 396 | struct netlink_ext_ack *extack, |
| 397 | bool check_allow_bit, |
| 398 | enum flow_action_hw_stats_bit allow_bit) |
| 399 | { |
| 400 | const struct flow_action_entry *action_entry; |
| 401 | |
| 402 | if (!flow_action_has_entries(action)) |
| 403 | return true; |
| 404 | if (!flow_action_mixed_hw_stats_check(action, extack)) |
| 405 | return false; |
| 406 | |
| 407 | action_entry = flow_action_first_entry_get(action); |
| 408 | |
| 409 | /* Zero is not a legal value for hw_stats, catch anyone passing it */ |
| 410 | WARN_ON_ONCE(!action_entry->hw_stats); |
| 411 | |
| 412 | if (!check_allow_bit && |
| 413 | ~action_entry->hw_stats & FLOW_ACTION_HW_STATS_ANY) { |
| 414 | NL_SET_ERR_MSG_MOD(extack, "Driver supports only default HW stats type \"any\"" ); |
| 415 | return false; |
| 416 | } else if (check_allow_bit && |
| 417 | !(action_entry->hw_stats & BIT(allow_bit))) { |
| 418 | NL_SET_ERR_MSG_MOD(extack, "Driver does not support selected HW stats type" ); |
| 419 | return false; |
| 420 | } |
| 421 | return true; |
| 422 | } |
| 423 | |
| 424 | static inline bool |
| 425 | flow_action_hw_stats_check(const struct flow_action *action, |
| 426 | struct netlink_ext_ack *extack, |
| 427 | enum flow_action_hw_stats_bit allow_bit) |
| 428 | { |
| 429 | return __flow_action_hw_stats_check(action, extack, check_allow_bit: true, allow_bit); |
| 430 | } |
| 431 | |
| 432 | static inline bool |
| 433 | flow_action_basic_hw_stats_check(const struct flow_action *action, |
| 434 | struct netlink_ext_ack *extack) |
| 435 | { |
| 436 | return __flow_action_hw_stats_check(action, extack, check_allow_bit: false, allow_bit: 0); |
| 437 | } |
| 438 | |
| 439 | struct flow_rule { |
| 440 | struct flow_match match; |
| 441 | struct flow_action action; |
| 442 | }; |
| 443 | |
| 444 | struct flow_rule *flow_rule_alloc(unsigned int num_actions); |
| 445 | |
| 446 | static inline bool flow_rule_match_key(const struct flow_rule *rule, |
| 447 | enum flow_dissector_key_id key) |
| 448 | { |
| 449 | return dissector_uses_key(flow_dissector: rule->match.dissector, key_id: key); |
| 450 | } |
| 451 | |
| 452 | /** |
| 453 | * flow_rule_is_supp_control_flags() - check for supported control flags |
| 454 | * @supp_flags: control flags supported by driver |
| 455 | * @ctrl_flags: control flags present in rule |
| 456 | * @extack: The netlink extended ACK for reporting errors. |
| 457 | * |
| 458 | * Return: true if only supported control flags are set, false otherwise. |
| 459 | */ |
| 460 | static inline bool flow_rule_is_supp_control_flags(const u32 supp_flags, |
| 461 | const u32 ctrl_flags, |
| 462 | struct netlink_ext_ack *extack) |
| 463 | { |
| 464 | if (likely((ctrl_flags & ~supp_flags) == 0)) |
| 465 | return true; |
| 466 | |
| 467 | NL_SET_ERR_MSG_FMT_MOD(extack, |
| 468 | "Unsupported match on control.flags %#x" , |
| 469 | ctrl_flags); |
| 470 | |
| 471 | return false; |
| 472 | } |
| 473 | |
| 474 | /** |
| 475 | * flow_rule_is_supp_enc_control_flags() - check for supported control flags |
| 476 | * @supp_enc_flags: encapsulation control flags supported by driver |
| 477 | * @enc_ctrl_flags: encapsulation control flags present in rule |
| 478 | * @extack: The netlink extended ACK for reporting errors. |
| 479 | * |
| 480 | * Return: true if only supported control flags are set, false otherwise. |
| 481 | */ |
| 482 | static inline bool flow_rule_is_supp_enc_control_flags(const u32 supp_enc_flags, |
| 483 | const u32 enc_ctrl_flags, |
| 484 | struct netlink_ext_ack *extack) |
| 485 | { |
| 486 | if (likely((enc_ctrl_flags & ~supp_enc_flags) == 0)) |
| 487 | return true; |
| 488 | |
| 489 | NL_SET_ERR_MSG_FMT_MOD(extack, |
| 490 | "Unsupported match on enc_control.flags %#x" , |
| 491 | enc_ctrl_flags); |
| 492 | |
| 493 | return false; |
| 494 | } |
| 495 | |
| 496 | /** |
| 497 | * flow_rule_has_control_flags() - check for presence of any control flags |
| 498 | * @ctrl_flags: control flags present in rule |
| 499 | * @extack: The netlink extended ACK for reporting errors. |
| 500 | * |
| 501 | * Return: true if control flags are set, false otherwise. |
| 502 | */ |
| 503 | static inline bool flow_rule_has_control_flags(const u32 ctrl_flags, |
| 504 | struct netlink_ext_ack *extack) |
| 505 | { |
| 506 | return !flow_rule_is_supp_control_flags(supp_flags: 0, ctrl_flags, extack); |
| 507 | } |
| 508 | |
| 509 | /** |
| 510 | * flow_rule_has_enc_control_flags() - check for presence of any control flags |
| 511 | * @enc_ctrl_flags: encapsulation control flags present in rule |
| 512 | * @extack: The netlink extended ACK for reporting errors. |
| 513 | * |
| 514 | * Return: true if control flags are set, false otherwise. |
| 515 | */ |
| 516 | static inline bool flow_rule_has_enc_control_flags(const u32 enc_ctrl_flags, |
| 517 | struct netlink_ext_ack *extack) |
| 518 | { |
| 519 | return !flow_rule_is_supp_enc_control_flags(supp_enc_flags: 0, enc_ctrl_flags, extack); |
| 520 | } |
| 521 | |
| 522 | /** |
| 523 | * flow_rule_match_has_control_flags() - match and check for any control flags |
| 524 | * @rule: The flow_rule under evaluation. |
| 525 | * @extack: The netlink extended ACK for reporting errors. |
| 526 | * |
| 527 | * Return: true if control flags are set, false otherwise. |
| 528 | */ |
| 529 | static inline bool flow_rule_match_has_control_flags(struct flow_rule *rule, |
| 530 | struct netlink_ext_ack *extack) |
| 531 | { |
| 532 | struct flow_match_control match; |
| 533 | |
| 534 | if (!flow_rule_match_key(rule, key: FLOW_DISSECTOR_KEY_CONTROL)) |
| 535 | return false; |
| 536 | |
| 537 | flow_rule_match_control(rule, out: &match); |
| 538 | |
| 539 | return flow_rule_has_control_flags(ctrl_flags: match.mask->flags, extack); |
| 540 | } |
| 541 | |
| 542 | struct flow_stats { |
| 543 | u64 pkts; |
| 544 | u64 bytes; |
| 545 | u64 drops; |
| 546 | u64 lastused; |
| 547 | enum flow_action_hw_stats used_hw_stats; |
| 548 | bool used_hw_stats_valid; |
| 549 | }; |
| 550 | |
| 551 | static inline void flow_stats_update(struct flow_stats *flow_stats, |
| 552 | u64 bytes, u64 pkts, |
| 553 | u64 drops, u64 lastused, |
| 554 | enum flow_action_hw_stats used_hw_stats) |
| 555 | { |
| 556 | flow_stats->pkts += pkts; |
| 557 | flow_stats->bytes += bytes; |
| 558 | flow_stats->drops += drops; |
| 559 | flow_stats->lastused = max_t(u64, flow_stats->lastused, lastused); |
| 560 | |
| 561 | /* The driver should pass value with a maximum of one bit set. |
| 562 | * Passing FLOW_ACTION_HW_STATS_ANY is invalid. |
| 563 | */ |
| 564 | WARN_ON(used_hw_stats == FLOW_ACTION_HW_STATS_ANY); |
| 565 | flow_stats->used_hw_stats |= used_hw_stats; |
| 566 | flow_stats->used_hw_stats_valid = true; |
| 567 | } |
| 568 | |
| 569 | enum flow_block_command { |
| 570 | FLOW_BLOCK_BIND, |
| 571 | FLOW_BLOCK_UNBIND, |
| 572 | }; |
| 573 | |
| 574 | enum flow_block_binder_type { |
| 575 | FLOW_BLOCK_BINDER_TYPE_UNSPEC, |
| 576 | FLOW_BLOCK_BINDER_TYPE_CLSACT_INGRESS, |
| 577 | FLOW_BLOCK_BINDER_TYPE_CLSACT_EGRESS, |
| 578 | FLOW_BLOCK_BINDER_TYPE_RED_EARLY_DROP, |
| 579 | FLOW_BLOCK_BINDER_TYPE_RED_MARK, |
| 580 | }; |
| 581 | |
| 582 | struct flow_block { |
| 583 | struct list_head cb_list; |
| 584 | }; |
| 585 | |
| 586 | struct netlink_ext_ack; |
| 587 | |
| 588 | struct flow_block_offload { |
| 589 | enum flow_block_command command; |
| 590 | enum flow_block_binder_type binder_type; |
| 591 | bool block_shared; |
| 592 | bool unlocked_driver_cb; |
| 593 | struct net *net; |
| 594 | struct flow_block *block; |
| 595 | struct list_head cb_list; |
| 596 | struct list_head *driver_block_list; |
| 597 | struct netlink_ext_ack *extack; |
| 598 | struct Qdisc *sch; |
| 599 | struct list_head *cb_list_head; |
| 600 | }; |
| 601 | |
| 602 | enum tc_setup_type; |
| 603 | typedef int flow_setup_cb_t(enum tc_setup_type type, void *type_data, |
| 604 | void *cb_priv); |
| 605 | |
| 606 | struct flow_block_cb; |
| 607 | |
| 608 | struct flow_block_indr { |
| 609 | struct list_head list; |
| 610 | struct net_device *dev; |
| 611 | struct Qdisc *sch; |
| 612 | enum flow_block_binder_type binder_type; |
| 613 | void *data; |
| 614 | void *cb_priv; |
| 615 | void (*cleanup)(struct flow_block_cb *block_cb); |
| 616 | }; |
| 617 | |
| 618 | struct flow_block_cb { |
| 619 | struct list_head driver_list; |
| 620 | struct list_head list; |
| 621 | flow_setup_cb_t *cb; |
| 622 | void *cb_ident; |
| 623 | void *cb_priv; |
| 624 | void (*release)(void *cb_priv); |
| 625 | struct flow_block_indr indr; |
| 626 | unsigned int refcnt; |
| 627 | }; |
| 628 | |
| 629 | struct flow_block_cb *flow_block_cb_alloc(flow_setup_cb_t *cb, |
| 630 | void *cb_ident, void *cb_priv, |
| 631 | void (*release)(void *cb_priv)); |
| 632 | struct flow_block_cb *flow_indr_block_cb_alloc(flow_setup_cb_t *cb, |
| 633 | void *cb_ident, void *cb_priv, |
| 634 | void (*release)(void *cb_priv), |
| 635 | struct flow_block_offload *bo, |
| 636 | struct net_device *dev, |
| 637 | struct Qdisc *sch, void *data, |
| 638 | void *indr_cb_priv, |
| 639 | void (*cleanup)(struct flow_block_cb *block_cb)); |
| 640 | void flow_block_cb_free(struct flow_block_cb *block_cb); |
| 641 | |
| 642 | struct flow_block_cb *flow_block_cb_lookup(struct flow_block *block, |
| 643 | flow_setup_cb_t *cb, void *cb_ident); |
| 644 | |
| 645 | void *flow_block_cb_priv(struct flow_block_cb *block_cb); |
| 646 | void flow_block_cb_incref(struct flow_block_cb *block_cb); |
| 647 | unsigned int flow_block_cb_decref(struct flow_block_cb *block_cb); |
| 648 | |
| 649 | static inline void flow_block_cb_add(struct flow_block_cb *block_cb, |
| 650 | struct flow_block_offload *offload) |
| 651 | { |
| 652 | list_add_tail(new: &block_cb->list, head: &offload->cb_list); |
| 653 | } |
| 654 | |
| 655 | static inline void flow_block_cb_remove(struct flow_block_cb *block_cb, |
| 656 | struct flow_block_offload *offload) |
| 657 | { |
| 658 | list_move(list: &block_cb->list, head: &offload->cb_list); |
| 659 | } |
| 660 | |
| 661 | static inline void flow_indr_block_cb_remove(struct flow_block_cb *block_cb, |
| 662 | struct flow_block_offload *offload) |
| 663 | { |
| 664 | list_del(entry: &block_cb->indr.list); |
| 665 | list_move(list: &block_cb->list, head: &offload->cb_list); |
| 666 | } |
| 667 | |
| 668 | bool flow_block_cb_is_busy(flow_setup_cb_t *cb, void *cb_ident, |
| 669 | struct list_head *driver_block_list); |
| 670 | |
| 671 | int flow_block_cb_setup_simple(struct flow_block_offload *f, |
| 672 | struct list_head *driver_list, |
| 673 | flow_setup_cb_t *cb, |
| 674 | void *cb_ident, void *cb_priv, bool ingress_only); |
| 675 | |
| 676 | enum flow_cls_command { |
| 677 | FLOW_CLS_REPLACE, |
| 678 | FLOW_CLS_DESTROY, |
| 679 | FLOW_CLS_STATS, |
| 680 | FLOW_CLS_TMPLT_CREATE, |
| 681 | FLOW_CLS_TMPLT_DESTROY, |
| 682 | }; |
| 683 | |
| 684 | struct flow_cls_common_offload { |
| 685 | u32 chain_index; |
| 686 | __be16 protocol; |
| 687 | u32 prio; |
| 688 | bool skip_sw; |
| 689 | struct netlink_ext_ack *extack; |
| 690 | }; |
| 691 | |
| 692 | struct flow_cls_offload { |
| 693 | struct flow_cls_common_offload common; |
| 694 | enum flow_cls_command command; |
| 695 | bool use_act_stats; |
| 696 | unsigned long cookie; |
| 697 | struct flow_rule *rule; |
| 698 | struct flow_stats stats; |
| 699 | u32 classid; |
| 700 | }; |
| 701 | |
| 702 | enum offload_act_command { |
| 703 | FLOW_ACT_REPLACE, |
| 704 | FLOW_ACT_DESTROY, |
| 705 | FLOW_ACT_STATS, |
| 706 | }; |
| 707 | |
| 708 | struct flow_offload_action { |
| 709 | struct netlink_ext_ack *extack; /* NULL in FLOW_ACT_STATS process*/ |
| 710 | enum offload_act_command command; |
| 711 | enum flow_action_id id; |
| 712 | u32 index; |
| 713 | unsigned long cookie; |
| 714 | struct flow_stats stats; |
| 715 | struct flow_action action; |
| 716 | }; |
| 717 | |
| 718 | struct flow_offload_action *offload_action_alloc(unsigned int num_actions); |
| 719 | |
| 720 | static inline struct flow_rule * |
| 721 | flow_cls_offload_flow_rule(struct flow_cls_offload *flow_cmd) |
| 722 | { |
| 723 | return flow_cmd->rule; |
| 724 | } |
| 725 | |
| 726 | static inline void flow_block_init(struct flow_block *flow_block) |
| 727 | { |
| 728 | INIT_LIST_HEAD(list: &flow_block->cb_list); |
| 729 | } |
| 730 | |
| 731 | typedef int flow_indr_block_bind_cb_t(struct net_device *dev, struct Qdisc *sch, void *cb_priv, |
| 732 | enum tc_setup_type type, void *type_data, |
| 733 | void *data, |
| 734 | void (*cleanup)(struct flow_block_cb *block_cb)); |
| 735 | |
| 736 | int flow_indr_dev_register(flow_indr_block_bind_cb_t *cb, void *cb_priv); |
| 737 | void flow_indr_dev_unregister(flow_indr_block_bind_cb_t *cb, void *cb_priv, |
| 738 | void (*release)(void *cb_priv)); |
| 739 | int flow_indr_dev_setup_offload(struct net_device *dev, struct Qdisc *sch, |
| 740 | enum tc_setup_type type, void *data, |
| 741 | struct flow_block_offload *bo, |
| 742 | void (*cleanup)(struct flow_block_cb *block_cb)); |
| 743 | bool flow_indr_dev_exists(void); |
| 744 | |
| 745 | #endif /* _NET_FLOW_OFFLOAD_H */ |
| 746 | |