^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1) /* SPDX-License-Identifier: GPL-2.0 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2) #ifndef __NET_UDP_TUNNEL_H
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3) #define __NET_UDP_TUNNEL_H
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 5) #include <net/ip_tunnels.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 6) #include <net/udp.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 7)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 8) #if IS_ENABLED(CONFIG_IPV6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 9) #include <net/ipv6.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 10) #include <net/ipv6_stubs.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 11) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 12)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 13) struct udp_port_cfg {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 14) u8 family;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 15)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 16) /* Used only for kernel-created sockets */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 17) union {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 18) struct in_addr local_ip;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 19) #if IS_ENABLED(CONFIG_IPV6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 20) struct in6_addr local_ip6;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 21) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 22) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 23)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 24) union {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 25) struct in_addr peer_ip;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 26) #if IS_ENABLED(CONFIG_IPV6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 27) struct in6_addr peer_ip6;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 28) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 29) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 30)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 31) __be16 local_udp_port;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 32) __be16 peer_udp_port;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 33) int bind_ifindex;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 34) unsigned int use_udp_checksums:1,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 35) use_udp6_tx_checksums:1,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 36) use_udp6_rx_checksums:1,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 37) ipv6_v6only:1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 38) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 39)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 40) int udp_sock_create4(struct net *net, struct udp_port_cfg *cfg,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 41) struct socket **sockp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 42)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 43) #if IS_ENABLED(CONFIG_IPV6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 44) int udp_sock_create6(struct net *net, struct udp_port_cfg *cfg,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 45) struct socket **sockp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 46) #else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 47) static inline int udp_sock_create6(struct net *net, struct udp_port_cfg *cfg,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 48) struct socket **sockp)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 49) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 50) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 51) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 52) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 53)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 54) static inline int udp_sock_create(struct net *net,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 55) struct udp_port_cfg *cfg,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 56) struct socket **sockp)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 57) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 58) if (cfg->family == AF_INET)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 59) return udp_sock_create4(net, cfg, sockp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 60)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 61) if (cfg->family == AF_INET6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 62) return udp_sock_create6(net, cfg, sockp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 63)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 64) return -EPFNOSUPPORT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 65) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 66)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 67) typedef int (*udp_tunnel_encap_rcv_t)(struct sock *sk, struct sk_buff *skb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 68) typedef int (*udp_tunnel_encap_err_lookup_t)(struct sock *sk,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 69) struct sk_buff *skb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 70) typedef void (*udp_tunnel_encap_destroy_t)(struct sock *sk);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 71) typedef struct sk_buff *(*udp_tunnel_gro_receive_t)(struct sock *sk,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 72) struct list_head *head,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 73) struct sk_buff *skb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 74) typedef int (*udp_tunnel_gro_complete_t)(struct sock *sk, struct sk_buff *skb,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 75) int nhoff);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 76)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 77) struct udp_tunnel_sock_cfg {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 78) void *sk_user_data; /* user data used by encap_rcv call back */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 79) /* Used for setting up udp_sock fields, see udp.h for details */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 80) __u8 encap_type;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 81) udp_tunnel_encap_rcv_t encap_rcv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 82) udp_tunnel_encap_err_lookup_t encap_err_lookup;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 83) udp_tunnel_encap_destroy_t encap_destroy;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 84) udp_tunnel_gro_receive_t gro_receive;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 85) udp_tunnel_gro_complete_t gro_complete;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 86) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 87)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 88) /* Setup the given (UDP) sock to receive UDP encapsulated packets */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 89) void setup_udp_tunnel_sock(struct net *net, struct socket *sock,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 90) struct udp_tunnel_sock_cfg *sock_cfg);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 91)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 92) /* -- List of parsable UDP tunnel types --
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 93) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 94) * Adding to this list will result in serious debate. The main issue is
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 95) * that this list is essentially a list of workarounds for either poorly
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 96) * designed tunnels, or poorly designed device offloads.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 97) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 98) * The parsing supported via these types should really be used for Rx
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 99) * traffic only as the network stack will have already inserted offsets for
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 100) * the location of the headers in the skb. In addition any ports that are
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 101) * pushed should be kept within the namespace without leaking to other
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 102) * devices such as VFs or other ports on the same device.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 103) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 104) * It is strongly encouraged to use CHECKSUM_COMPLETE for Rx to avoid the
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 105) * need to use this for Rx checksum offload. It should not be necessary to
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 106) * call this function to perform Tx offloads on outgoing traffic.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 107) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 108) enum udp_parsable_tunnel_type {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 109) UDP_TUNNEL_TYPE_VXLAN = BIT(0), /* RFC 7348 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 110) UDP_TUNNEL_TYPE_GENEVE = BIT(1), /* draft-ietf-nvo3-geneve */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 111) UDP_TUNNEL_TYPE_VXLAN_GPE = BIT(2), /* draft-ietf-nvo3-vxlan-gpe */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 112) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 113)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 114) struct udp_tunnel_info {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 115) unsigned short type;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 116) sa_family_t sa_family;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 117) __be16 port;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 118) u8 hw_priv;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 119) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 120)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 121) /* Notify network devices of offloadable types */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 122) void udp_tunnel_push_rx_port(struct net_device *dev, struct socket *sock,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 123) unsigned short type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 124) void udp_tunnel_drop_rx_port(struct net_device *dev, struct socket *sock,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 125) unsigned short type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 126) void udp_tunnel_notify_add_rx_port(struct socket *sock, unsigned short type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 127) void udp_tunnel_notify_del_rx_port(struct socket *sock, unsigned short type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 128)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 129) static inline void udp_tunnel_get_rx_info(struct net_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 130) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 131) ASSERT_RTNL();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 132) call_netdevice_notifiers(NETDEV_UDP_TUNNEL_PUSH_INFO, dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 133) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 134)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 135) static inline void udp_tunnel_drop_rx_info(struct net_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 136) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 137) ASSERT_RTNL();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 138) call_netdevice_notifiers(NETDEV_UDP_TUNNEL_DROP_INFO, dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 139) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 140)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 141) /* Transmit the skb using UDP encapsulation. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 142) void udp_tunnel_xmit_skb(struct rtable *rt, struct sock *sk, struct sk_buff *skb,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 143) __be32 src, __be32 dst, __u8 tos, __u8 ttl,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 144) __be16 df, __be16 src_port, __be16 dst_port,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 145) bool xnet, bool nocheck);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 146)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 147) int udp_tunnel6_xmit_skb(struct dst_entry *dst, struct sock *sk,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 148) struct sk_buff *skb,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 149) struct net_device *dev, struct in6_addr *saddr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 150) struct in6_addr *daddr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 151) __u8 prio, __u8 ttl, __be32 label,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 152) __be16 src_port, __be16 dst_port, bool nocheck);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 153)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 154) void udp_tunnel_sock_release(struct socket *sock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 155)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 156) struct metadata_dst *udp_tun_rx_dst(struct sk_buff *skb, unsigned short family,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 157) __be16 flags, __be64 tunnel_id,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 158) int md_size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 159)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 160) #ifdef CONFIG_INET
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 161) static inline int udp_tunnel_handle_offloads(struct sk_buff *skb, bool udp_csum)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 162) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 163) int type = udp_csum ? SKB_GSO_UDP_TUNNEL_CSUM : SKB_GSO_UDP_TUNNEL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 164)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 165) return iptunnel_handle_offloads(skb, type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 166) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 167) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 168)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 169) static inline void udp_tunnel_encap_enable(struct socket *sock)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 170) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 171) struct udp_sock *up = udp_sk(sock->sk);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 172)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 173) if (up->encap_enabled)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 174) return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 175)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 176) up->encap_enabled = 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 177) #if IS_ENABLED(CONFIG_IPV6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 178) if (sock->sk->sk_family == PF_INET6)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 179) ipv6_stub->udpv6_encap_enable();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 180) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 181) udp_encap_enable();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 182) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 183)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 184) #define UDP_TUNNEL_NIC_MAX_TABLES 4
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 185)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 186) enum udp_tunnel_nic_info_flags {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 187) /* Device callbacks may sleep */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 188) UDP_TUNNEL_NIC_INFO_MAY_SLEEP = BIT(0),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 189) /* Device only supports offloads when it's open, all ports
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 190) * will be removed before close and re-added after open.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 191) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 192) UDP_TUNNEL_NIC_INFO_OPEN_ONLY = BIT(1),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 193) /* Device supports only IPv4 tunnels */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 194) UDP_TUNNEL_NIC_INFO_IPV4_ONLY = BIT(2),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 195) /* Device has hard-coded the IANA VXLAN port (4789) as VXLAN.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 196) * This port must not be counted towards n_entries of any table.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 197) * Driver will not receive any callback associated with port 4789.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 198) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 199) UDP_TUNNEL_NIC_INFO_STATIC_IANA_VXLAN = BIT(3),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 200) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 201)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 202) struct udp_tunnel_nic;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 203)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 204) #define UDP_TUNNEL_NIC_MAX_SHARING_DEVICES (U16_MAX / 2)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 205)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 206) struct udp_tunnel_nic_shared {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 207) struct udp_tunnel_nic *udp_tunnel_nic_info;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 208)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 209) struct list_head devices;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 210) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 211)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 212) struct udp_tunnel_nic_shared_node {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 213) struct net_device *dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 214) struct list_head list;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 215) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 216)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 217) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 218) * struct udp_tunnel_nic_info - driver UDP tunnel offload information
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 219) * @set_port: callback for adding a new port
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 220) * @unset_port: callback for removing a port
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 221) * @sync_table: callback for syncing the entire port table at once
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 222) * @shared: reference to device global state (optional)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 223) * @flags: device flags from enum udp_tunnel_nic_info_flags
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 224) * @tables: UDP port tables this device has
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 225) * @tables.n_entries: number of entries in this table
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 226) * @tables.tunnel_types: types of tunnels this table accepts
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 227) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 228) * Drivers are expected to provide either @set_port and @unset_port callbacks
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 229) * or the @sync_table callback. Callbacks are invoked with rtnl lock held.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 230) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 231) * Devices which (misguidedly) share the UDP tunnel port table across multiple
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 232) * netdevs should allocate an instance of struct udp_tunnel_nic_shared and
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 233) * point @shared at it.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 234) * There must never be more than %UDP_TUNNEL_NIC_MAX_SHARING_DEVICES devices
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 235) * sharing a table.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 236) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 237) * Known limitations:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 238) * - UDP tunnel port notifications are fundamentally best-effort -
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 239) * it is likely the driver will both see skbs which use a UDP tunnel port,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 240) * while not being a tunneled skb, and tunnel skbs from other ports -
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 241) * drivers should only use these ports for non-critical RX-side offloads,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 242) * e.g. the checksum offload;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 243) * - none of the devices care about the socket family at present, so we don't
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 244) * track it. Please extend this code if you care.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 245) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 246) struct udp_tunnel_nic_info {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 247) /* one-by-one */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 248) int (*set_port)(struct net_device *dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 249) unsigned int table, unsigned int entry,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 250) struct udp_tunnel_info *ti);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 251) int (*unset_port)(struct net_device *dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 252) unsigned int table, unsigned int entry,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 253) struct udp_tunnel_info *ti);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 254)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 255) /* all at once */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 256) int (*sync_table)(struct net_device *dev, unsigned int table);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 257)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 258) struct udp_tunnel_nic_shared *shared;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 259)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 260) unsigned int flags;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 261)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 262) struct udp_tunnel_nic_table_info {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 263) unsigned int n_entries;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 264) unsigned int tunnel_types;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 265) } tables[UDP_TUNNEL_NIC_MAX_TABLES];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 266) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 267)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 268) /* UDP tunnel module dependencies
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 269) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 270) * Tunnel drivers are expected to have a hard dependency on the udp_tunnel
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 271) * module. NIC drivers are not, they just attach their
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 272) * struct udp_tunnel_nic_info to the netdev and wait for callbacks to come.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 273) * Loading a tunnel driver will cause the udp_tunnel module to be loaded
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 274) * and only then will all the required state structures be allocated.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 275) * Since we want a weak dependency from the drivers and the core to udp_tunnel
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 276) * we call things through the following stubs.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 277) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 278) struct udp_tunnel_nic_ops {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 279) void (*get_port)(struct net_device *dev, unsigned int table,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 280) unsigned int idx, struct udp_tunnel_info *ti);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 281) void (*set_port_priv)(struct net_device *dev, unsigned int table,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 282) unsigned int idx, u8 priv);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 283) void (*add_port)(struct net_device *dev, struct udp_tunnel_info *ti);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 284) void (*del_port)(struct net_device *dev, struct udp_tunnel_info *ti);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 285) void (*reset_ntf)(struct net_device *dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 286)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 287) size_t (*dump_size)(struct net_device *dev, unsigned int table);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 288) int (*dump_write)(struct net_device *dev, unsigned int table,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 289) struct sk_buff *skb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 290) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 291)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 292) #ifdef CONFIG_INET
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 293) extern const struct udp_tunnel_nic_ops *udp_tunnel_nic_ops;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 294) #else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 295) #define udp_tunnel_nic_ops ((struct udp_tunnel_nic_ops *)NULL)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 296) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 297)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 298) static inline void
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 299) udp_tunnel_nic_get_port(struct net_device *dev, unsigned int table,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 300) unsigned int idx, struct udp_tunnel_info *ti)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 301) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 302) /* This helper is used from .sync_table, we indicate empty entries
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 303) * by zero'ed @ti. Drivers which need to know the details of a port
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 304) * when it gets deleted should use the .set_port / .unset_port
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 305) * callbacks.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 306) * Zero out here, otherwise !CONFIG_INET causes uninitilized warnings.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 307) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 308) memset(ti, 0, sizeof(*ti));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 309)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 310) if (udp_tunnel_nic_ops)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 311) udp_tunnel_nic_ops->get_port(dev, table, idx, ti);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 312) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 313)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 314) static inline void
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 315) udp_tunnel_nic_set_port_priv(struct net_device *dev, unsigned int table,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 316) unsigned int idx, u8 priv)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 317) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 318) if (udp_tunnel_nic_ops)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 319) udp_tunnel_nic_ops->set_port_priv(dev, table, idx, priv);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 320) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 321)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 322) static inline void
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 323) udp_tunnel_nic_add_port(struct net_device *dev, struct udp_tunnel_info *ti)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 324) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 325) if (udp_tunnel_nic_ops)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 326) udp_tunnel_nic_ops->add_port(dev, ti);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 327) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 328)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 329) static inline void
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 330) udp_tunnel_nic_del_port(struct net_device *dev, struct udp_tunnel_info *ti)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 331) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 332) if (udp_tunnel_nic_ops)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 333) udp_tunnel_nic_ops->del_port(dev, ti);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 334) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 335)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 336) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 337) * udp_tunnel_nic_reset_ntf() - device-originating reset notification
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 338) * @dev: network interface device structure
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 339) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 340) * Called by the driver to inform the core that the entire UDP tunnel port
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 341) * state has been lost, usually due to device reset. Core will assume device
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 342) * forgot all the ports and issue .set_port and .sync_table callbacks as
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 343) * necessary.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 344) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 345) * This function must be called with rtnl lock held, and will issue all
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 346) * the callbacks before returning.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 347) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 348) static inline void udp_tunnel_nic_reset_ntf(struct net_device *dev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 349) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 350) if (udp_tunnel_nic_ops)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 351) udp_tunnel_nic_ops->reset_ntf(dev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 352) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 353)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 354) static inline size_t
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 355) udp_tunnel_nic_dump_size(struct net_device *dev, unsigned int table)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 356) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 357) if (!udp_tunnel_nic_ops)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 358) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 359) return udp_tunnel_nic_ops->dump_size(dev, table);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 360) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 361)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 362) static inline int
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 363) udp_tunnel_nic_dump_write(struct net_device *dev, unsigned int table,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 364) struct sk_buff *skb)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 365) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 366) if (!udp_tunnel_nic_ops)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 367) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 368) return udp_tunnel_nic_ops->dump_write(dev, table, skb);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 369) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 370) #endif