| // SPDX-License-Identifier: GPL-2.0 |
| /* OpenVPN data channel offload |
| * |
| * Copyright (C) 2020-2025 OpenVPN, Inc. |
| * |
| * Author: James Yonan <james@openvpn.net> |
| * Antonio Quartulli <antonio@openvpn.net> |
| */ |
| |
| #include <linux/skbuff.h> |
| #include <linux/list.h> |
| #include <linux/hashtable.h> |
| #include <net/ip6_route.h> |
| |
| #include "ovpnpriv.h" |
| #include "bind.h" |
| #include "pktid.h" |
| #include "crypto.h" |
| #include "io.h" |
| #include "main.h" |
| #include "netlink.h" |
| #include "peer.h" |
| #include "socket.h" |
| |
| static void unlock_ovpn(struct ovpn_priv *ovpn, |
| struct llist_head *release_list) |
| __releases(&ovpn->lock) |
| { |
| struct ovpn_peer *peer; |
| |
| spin_unlock_bh(&ovpn->lock); |
| |
| llist_for_each_entry(peer, release_list->first, release_entry) { |
| ovpn_socket_release(peer); |
| ovpn_peer_put(peer); |
| } |
| } |
| |
| /** |
| * ovpn_peer_keepalive_set - configure keepalive values for peer |
| * @peer: the peer to configure |
| * @interval: outgoing keepalive interval |
| * @timeout: incoming keepalive timeout |
| */ |
| void ovpn_peer_keepalive_set(struct ovpn_peer *peer, u32 interval, u32 timeout) |
| { |
| time64_t now = ktime_get_real_seconds(); |
| |
| netdev_dbg(peer->ovpn->dev, |
| "scheduling keepalive for peer %u: interval=%u timeout=%u\n", |
| peer->id, interval, timeout); |
| |
| peer->keepalive_interval = interval; |
| WRITE_ONCE(peer->last_sent, now); |
| peer->keepalive_xmit_exp = now + interval; |
| |
| peer->keepalive_timeout = timeout; |
| WRITE_ONCE(peer->last_recv, now); |
| peer->keepalive_recv_exp = now + timeout; |
| |
| /* now that interval and timeout have been changed, kick |
| * off the worker so that the next delay can be recomputed |
| */ |
| mod_delayed_work(system_wq, &peer->ovpn->keepalive_work, 0); |
| } |
| |
| /** |
| * ovpn_peer_keepalive_send - periodic worker sending keepalive packets |
| * @work: pointer to the work member of the related peer object |
| * |
| * NOTE: the reference to peer is not dropped because it gets inherited |
| * by ovpn_xmit_special() |
| */ |
| static void ovpn_peer_keepalive_send(struct work_struct *work) |
| { |
| struct ovpn_peer *peer = container_of(work, struct ovpn_peer, |
| keepalive_work); |
| |
| local_bh_disable(); |
| ovpn_xmit_special(peer, ovpn_keepalive_message, |
| sizeof(ovpn_keepalive_message)); |
| local_bh_enable(); |
| } |
| |
| /** |
| * ovpn_peer_new - allocate and initialize a new peer object |
| * @ovpn: the openvpn instance inside which the peer should be created |
| * @id: the ID assigned to this peer |
| * |
| * Return: a pointer to the new peer on success or an error code otherwise |
| */ |
| struct ovpn_peer *ovpn_peer_new(struct ovpn_priv *ovpn, u32 id) |
| { |
| struct ovpn_peer *peer; |
| int ret; |
| |
| /* alloc and init peer object */ |
| peer = kzalloc(sizeof(*peer), GFP_KERNEL); |
| if (!peer) |
| return ERR_PTR(-ENOMEM); |
| |
| peer->id = id; |
| peer->ovpn = ovpn; |
| |
| peer->vpn_addrs.ipv4.s_addr = htonl(INADDR_ANY); |
| peer->vpn_addrs.ipv6 = in6addr_any; |
| |
| RCU_INIT_POINTER(peer->bind, NULL); |
| ovpn_crypto_state_init(&peer->crypto); |
| spin_lock_init(&peer->lock); |
| kref_init(&peer->refcount); |
| ovpn_peer_stats_init(&peer->vpn_stats); |
| ovpn_peer_stats_init(&peer->link_stats); |
| INIT_WORK(&peer->keepalive_work, ovpn_peer_keepalive_send); |
| |
| ret = dst_cache_init(&peer->dst_cache, GFP_KERNEL); |
| if (ret < 0) { |
| netdev_err(ovpn->dev, |
| "cannot initialize dst cache for peer %u\n", |
| peer->id); |
| kfree(peer); |
| return ERR_PTR(ret); |
| } |
| |
| netdev_hold(ovpn->dev, &peer->dev_tracker, GFP_KERNEL); |
| |
| return peer; |
| } |
| |
| /** |
| * ovpn_peer_reset_sockaddr - recreate binding for peer |
| * @peer: peer to recreate the binding for |
| * @ss: sockaddr to use as remote endpoint for the binding |
| * @local_ip: local IP for the binding |
| * |
| * Return: 0 on success or a negative error code otherwise |
| */ |
| int ovpn_peer_reset_sockaddr(struct ovpn_peer *peer, |
| const struct sockaddr_storage *ss, |
| const void *local_ip) |
| { |
| struct ovpn_bind *bind; |
| size_t ip_len; |
| |
| lockdep_assert_held(&peer->lock); |
| |
| /* create new ovpn_bind object */ |
| bind = ovpn_bind_from_sockaddr(ss); |
| if (IS_ERR(bind)) |
| return PTR_ERR(bind); |
| |
| if (local_ip) { |
| if (ss->ss_family == AF_INET) { |
| ip_len = sizeof(struct in_addr); |
| } else if (ss->ss_family == AF_INET6) { |
| ip_len = sizeof(struct in6_addr); |
| } else { |
| net_dbg_ratelimited("%s: invalid family %u for remote endpoint for peer %u\n", |
| netdev_name(peer->ovpn->dev), |
| ss->ss_family, peer->id); |
| kfree(bind); |
| return -EINVAL; |
| } |
| |
| memcpy(&bind->local, local_ip, ip_len); |
| } |
| |
| /* set binding */ |
| ovpn_bind_reset(peer, bind); |
| |
| return 0; |
| } |
| |
| /* variable name __tbl2 needs to be different from __tbl1 |
| * in the macro below to avoid confusing clang |
| */ |
| #define ovpn_get_hash_slot(_tbl, _key, _key_len) ({ \ |
| typeof(_tbl) *__tbl2 = &(_tbl); \ |
| jhash(_key, _key_len, 0) % HASH_SIZE(*__tbl2); \ |
| }) |
| |
| #define ovpn_get_hash_head(_tbl, _key, _key_len) ({ \ |
| typeof(_tbl) *__tbl1 = &(_tbl); \ |
| &(*__tbl1)[ovpn_get_hash_slot(*__tbl1, _key, _key_len)];\ |
| }) |
| |
| /** |
| * ovpn_peer_endpoints_update - update remote or local endpoint for peer |
| * @peer: peer to update the remote endpoint for |
| * @skb: incoming packet to retrieve the source/destination address from |
| */ |
| void ovpn_peer_endpoints_update(struct ovpn_peer *peer, struct sk_buff *skb) |
| { |
| struct hlist_nulls_head *nhead; |
| struct sockaddr_storage ss; |
| struct sockaddr_in6 *sa6; |
| bool reset_cache = false; |
| struct sockaddr_in *sa; |
| struct ovpn_bind *bind; |
| const void *local_ip; |
| size_t salen = 0; |
| |
| spin_lock_bh(&peer->lock); |
| bind = rcu_dereference_protected(peer->bind, |
| lockdep_is_held(&peer->lock)); |
| if (unlikely(!bind)) |
| goto unlock; |
| |
| switch (skb->protocol) { |
| case htons(ETH_P_IP): |
| /* float check */ |
| if (unlikely(!ovpn_bind_skb_src_match(bind, skb))) { |
| /* unconditionally save local endpoint in case |
| * of float, as it may have changed as well |
| */ |
| local_ip = &ip_hdr(skb)->daddr; |
| sa = (struct sockaddr_in *)&ss; |
| sa->sin_family = AF_INET; |
| sa->sin_addr.s_addr = ip_hdr(skb)->saddr; |
| sa->sin_port = udp_hdr(skb)->source; |
| salen = sizeof(*sa); |
| reset_cache = true; |
| break; |
| } |
| |
| /* if no float happened, let's double check if the local endpoint |
| * has changed |
| */ |
| if (unlikely(bind->local.ipv4.s_addr != ip_hdr(skb)->daddr)) { |
| net_dbg_ratelimited("%s: learning local IPv4 for peer %d (%pI4 -> %pI4)\n", |
| netdev_name(peer->ovpn->dev), |
| peer->id, &bind->local.ipv4.s_addr, |
| &ip_hdr(skb)->daddr); |
| bind->local.ipv4.s_addr = ip_hdr(skb)->daddr; |
| reset_cache = true; |
| } |
| break; |
| case htons(ETH_P_IPV6): |
| /* float check */ |
| if (unlikely(!ovpn_bind_skb_src_match(bind, skb))) { |
| /* unconditionally save local endpoint in case |
| * of float, as it may have changed as well |
| */ |
| local_ip = &ipv6_hdr(skb)->daddr; |
| sa6 = (struct sockaddr_in6 *)&ss; |
| sa6->sin6_family = AF_INET6; |
| sa6->sin6_addr = ipv6_hdr(skb)->saddr; |
| sa6->sin6_port = udp_hdr(skb)->source; |
| sa6->sin6_scope_id = ipv6_iface_scope_id(&ipv6_hdr(skb)->saddr, |
| skb->skb_iif); |
| salen = sizeof(*sa6); |
| reset_cache = true; |
| break; |
| } |
| |
| /* if no float happened, let's double check if the local endpoint |
| * has changed |
| */ |
| if (unlikely(!ipv6_addr_equal(&bind->local.ipv6, |
| &ipv6_hdr(skb)->daddr))) { |
| net_dbg_ratelimited("%s: learning local IPv6 for peer %d (%pI6c -> %pI6c)\n", |
| netdev_name(peer->ovpn->dev), |
| peer->id, &bind->local.ipv6, |
| &ipv6_hdr(skb)->daddr); |
| bind->local.ipv6 = ipv6_hdr(skb)->daddr; |
| reset_cache = true; |
| } |
| break; |
| default: |
| goto unlock; |
| } |
| |
| if (unlikely(reset_cache)) |
| dst_cache_reset(&peer->dst_cache); |
| |
| /* if the peer did not float, we can bail out now */ |
| if (likely(!salen)) |
| goto unlock; |
| |
| if (unlikely(ovpn_peer_reset_sockaddr(peer, |
| (struct sockaddr_storage *)&ss, |
| local_ip) < 0)) |
| goto unlock; |
| |
| net_dbg_ratelimited("%s: peer %d floated to %pIScp", |
| netdev_name(peer->ovpn->dev), peer->id, &ss); |
| |
| spin_unlock_bh(&peer->lock); |
| |
| /* rehashing is required only in MP mode as P2P has one peer |
| * only and thus there is no hashtable |
| */ |
| if (peer->ovpn->mode == OVPN_MODE_MP) { |
| spin_lock_bh(&peer->ovpn->lock); |
| spin_lock_bh(&peer->lock); |
| bind = rcu_dereference_protected(peer->bind, |
| lockdep_is_held(&peer->lock)); |
| if (unlikely(!bind)) { |
| spin_unlock_bh(&peer->lock); |
| spin_unlock_bh(&peer->ovpn->lock); |
| return; |
| } |
| |
| /* This function may be invoked concurrently, therefore another |
| * float may have happened in parallel: perform rehashing |
| * using the peer->bind->remote directly as key |
| */ |
| |
| switch (bind->remote.in4.sin_family) { |
| case AF_INET: |
| salen = sizeof(*sa); |
| break; |
| case AF_INET6: |
| salen = sizeof(*sa6); |
| break; |
| } |
| |
| /* remove old hashing */ |
| hlist_nulls_del_init_rcu(&peer->hash_entry_transp_addr); |
| /* re-add with new transport address */ |
| nhead = ovpn_get_hash_head(peer->ovpn->peers->by_transp_addr, |
| &bind->remote, salen); |
| hlist_nulls_add_head_rcu(&peer->hash_entry_transp_addr, nhead); |
| spin_unlock_bh(&peer->lock); |
| spin_unlock_bh(&peer->ovpn->lock); |
| } |
| return; |
| unlock: |
| spin_unlock_bh(&peer->lock); |
| } |
| |
| /** |
| * ovpn_peer_release_rcu - RCU callback performing last peer release steps |
| * @head: RCU member of the ovpn_peer |
| */ |
| static void ovpn_peer_release_rcu(struct rcu_head *head) |
| { |
| struct ovpn_peer *peer = container_of(head, struct ovpn_peer, rcu); |
| |
| /* this call will immediately free the dst_cache, therefore we |
| * perform it in the RCU callback, when all contexts are done |
| */ |
| dst_cache_destroy(&peer->dst_cache); |
| kfree(peer); |
| } |
| |
| /** |
| * ovpn_peer_release - release peer private members |
| * @peer: the peer to release |
| */ |
| void ovpn_peer_release(struct ovpn_peer *peer) |
| { |
| ovpn_crypto_state_release(&peer->crypto); |
| spin_lock_bh(&peer->lock); |
| ovpn_bind_reset(peer, NULL); |
| spin_unlock_bh(&peer->lock); |
| call_rcu(&peer->rcu, ovpn_peer_release_rcu); |
| netdev_put(peer->ovpn->dev, &peer->dev_tracker); |
| } |
| |
| /** |
| * ovpn_peer_release_kref - callback for kref_put |
| * @kref: the kref object belonging to the peer |
| */ |
| void ovpn_peer_release_kref(struct kref *kref) |
| { |
| struct ovpn_peer *peer = container_of(kref, struct ovpn_peer, refcount); |
| |
| ovpn_peer_release(peer); |
| } |
| |
| /** |
| * ovpn_peer_skb_to_sockaddr - fill sockaddr with skb source address |
| * @skb: the packet to extract data from |
| * @ss: the sockaddr to fill |
| * |
| * Return: sockaddr length on success or -1 otherwise |
| */ |
| static int ovpn_peer_skb_to_sockaddr(struct sk_buff *skb, |
| struct sockaddr_storage *ss) |
| { |
| struct sockaddr_in6 *sa6; |
| struct sockaddr_in *sa4; |
| |
| switch (skb->protocol) { |
| case htons(ETH_P_IP): |
| sa4 = (struct sockaddr_in *)ss; |
| sa4->sin_family = AF_INET; |
| sa4->sin_addr.s_addr = ip_hdr(skb)->saddr; |
| sa4->sin_port = udp_hdr(skb)->source; |
| return sizeof(*sa4); |
| case htons(ETH_P_IPV6): |
| sa6 = (struct sockaddr_in6 *)ss; |
| sa6->sin6_family = AF_INET6; |
| sa6->sin6_addr = ipv6_hdr(skb)->saddr; |
| sa6->sin6_port = udp_hdr(skb)->source; |
| return sizeof(*sa6); |
| } |
| |
| return -1; |
| } |
| |
| /** |
| * ovpn_nexthop_from_skb4 - retrieve IPv4 nexthop for outgoing skb |
| * @skb: the outgoing packet |
| * |
| * Return: the IPv4 of the nexthop |
| */ |
| static __be32 ovpn_nexthop_from_skb4(struct sk_buff *skb) |
| { |
| const struct rtable *rt = skb_rtable(skb); |
| |
| if (rt && rt->rt_uses_gateway) |
| return rt->rt_gw4; |
| |
| return ip_hdr(skb)->daddr; |
| } |
| |
| /** |
| * ovpn_nexthop_from_skb6 - retrieve IPv6 nexthop for outgoing skb |
| * @skb: the outgoing packet |
| * |
| * Return: the IPv6 of the nexthop |
| */ |
| static struct in6_addr ovpn_nexthop_from_skb6(struct sk_buff *skb) |
| { |
| const struct rt6_info *rt = skb_rt6_info(skb); |
| |
| if (!rt || !(rt->rt6i_flags & RTF_GATEWAY)) |
| return ipv6_hdr(skb)->daddr; |
| |
| return rt->rt6i_gateway; |
| } |
| |
| /** |
| * ovpn_peer_get_by_vpn_addr4 - retrieve peer by its VPN IPv4 address |
| * @ovpn: the openvpn instance to search |
| * @addr: VPN IPv4 to use as search key |
| * |
| * Refcounter is not increased for the returned peer. |
| * |
| * Return: the peer if found or NULL otherwise |
| */ |
| static struct ovpn_peer *ovpn_peer_get_by_vpn_addr4(struct ovpn_priv *ovpn, |
| __be32 addr) |
| { |
| struct hlist_nulls_head *nhead; |
| struct hlist_nulls_node *ntmp; |
| struct ovpn_peer *tmp; |
| unsigned int slot; |
| |
| begin: |
| slot = ovpn_get_hash_slot(ovpn->peers->by_vpn_addr4, &addr, |
| sizeof(addr)); |
| nhead = &ovpn->peers->by_vpn_addr4[slot]; |
| |
| hlist_nulls_for_each_entry_rcu(tmp, ntmp, nhead, hash_entry_addr4) |
| if (addr == tmp->vpn_addrs.ipv4.s_addr) |
| return tmp; |
| |
| /* item may have moved during lookup - check nulls and restart |
| * if that's the case |
| */ |
| if (get_nulls_value(ntmp) != slot) |
| goto begin; |
| |
| return NULL; |
| } |
| |
| /** |
| * ovpn_peer_get_by_vpn_addr6 - retrieve peer by its VPN IPv6 address |
| * @ovpn: the openvpn instance to search |
| * @addr: VPN IPv6 to use as search key |
| * |
| * Refcounter is not increased for the returned peer. |
| * |
| * Return: the peer if found or NULL otherwise |
| */ |
| static struct ovpn_peer *ovpn_peer_get_by_vpn_addr6(struct ovpn_priv *ovpn, |
| struct in6_addr *addr) |
| { |
| struct hlist_nulls_head *nhead; |
| struct hlist_nulls_node *ntmp; |
| struct ovpn_peer *tmp; |
| unsigned int slot; |
| |
| begin: |
| slot = ovpn_get_hash_slot(ovpn->peers->by_vpn_addr6, addr, |
| sizeof(*addr)); |
| nhead = &ovpn->peers->by_vpn_addr6[slot]; |
| |
| hlist_nulls_for_each_entry_rcu(tmp, ntmp, nhead, hash_entry_addr6) |
| if (ipv6_addr_equal(addr, &tmp->vpn_addrs.ipv6)) |
| return tmp; |
| |
| /* item may have moved during lookup - check nulls and restart |
| * if that's the case |
| */ |
| if (get_nulls_value(ntmp) != slot) |
| goto begin; |
| |
| return NULL; |
| } |
| |
| /** |
| * ovpn_peer_transp_match - check if sockaddr and peer binding match |
| * @peer: the peer to get the binding from |
| * @ss: the sockaddr to match |
| * |
| * Return: true if sockaddr and binding match or false otherwise |
| */ |
| static bool ovpn_peer_transp_match(const struct ovpn_peer *peer, |
| const struct sockaddr_storage *ss) |
| { |
| struct ovpn_bind *bind = rcu_dereference(peer->bind); |
| struct sockaddr_in6 *sa6; |
| struct sockaddr_in *sa4; |
| |
| if (unlikely(!bind)) |
| return false; |
| |
| if (ss->ss_family != bind->remote.in4.sin_family) |
| return false; |
| |
| switch (ss->ss_family) { |
| case AF_INET: |
| sa4 = (struct sockaddr_in *)ss; |
| if (sa4->sin_addr.s_addr != bind->remote.in4.sin_addr.s_addr) |
| return false; |
| if (sa4->sin_port != bind->remote.in4.sin_port) |
| return false; |
| break; |
| case AF_INET6: |
| sa6 = (struct sockaddr_in6 *)ss; |
| if (!ipv6_addr_equal(&sa6->sin6_addr, |
| &bind->remote.in6.sin6_addr)) |
| return false; |
| if (sa6->sin6_port != bind->remote.in6.sin6_port) |
| return false; |
| break; |
| default: |
| return false; |
| } |
| |
| return true; |
| } |
| |
| /** |
| * ovpn_peer_get_by_transp_addr_p2p - get peer by transport address in a P2P |
| * instance |
| * @ovpn: the openvpn instance to search |
| * @ss: the transport socket address |
| * |
| * Return: the peer if found or NULL otherwise |
| */ |
| static struct ovpn_peer * |
| ovpn_peer_get_by_transp_addr_p2p(struct ovpn_priv *ovpn, |
| struct sockaddr_storage *ss) |
| { |
| struct ovpn_peer *tmp, *peer = NULL; |
| |
| rcu_read_lock(); |
| tmp = rcu_dereference(ovpn->peer); |
| if (likely(tmp && ovpn_peer_transp_match(tmp, ss) && |
| ovpn_peer_hold(tmp))) |
| peer = tmp; |
| rcu_read_unlock(); |
| |
| return peer; |
| } |
| |
| /** |
| * ovpn_peer_get_by_transp_addr - retrieve peer by transport address |
| * @ovpn: the openvpn instance to search |
| * @skb: the skb to retrieve the source transport address from |
| * |
| * Return: a pointer to the peer if found or NULL otherwise |
| */ |
| struct ovpn_peer *ovpn_peer_get_by_transp_addr(struct ovpn_priv *ovpn, |
| struct sk_buff *skb) |
| { |
| struct ovpn_peer *tmp, *peer = NULL; |
| struct sockaddr_storage ss = { 0 }; |
| struct hlist_nulls_head *nhead; |
| struct hlist_nulls_node *ntmp; |
| unsigned int slot; |
| ssize_t sa_len; |
| |
| sa_len = ovpn_peer_skb_to_sockaddr(skb, &ss); |
| if (unlikely(sa_len < 0)) |
| return NULL; |
| |
| if (ovpn->mode == OVPN_MODE_P2P) |
| return ovpn_peer_get_by_transp_addr_p2p(ovpn, &ss); |
| |
| rcu_read_lock(); |
| begin: |
| slot = ovpn_get_hash_slot(ovpn->peers->by_transp_addr, &ss, sa_len); |
| nhead = &ovpn->peers->by_transp_addr[slot]; |
| |
| hlist_nulls_for_each_entry_rcu(tmp, ntmp, nhead, |
| hash_entry_transp_addr) { |
| if (!ovpn_peer_transp_match(tmp, &ss)) |
| continue; |
| |
| if (!ovpn_peer_hold(tmp)) |
| continue; |
| |
| peer = tmp; |
| break; |
| } |
| |
| /* item may have moved during lookup - check nulls and restart |
| * if that's the case |
| */ |
| if (!peer && get_nulls_value(ntmp) != slot) |
| goto begin; |
| rcu_read_unlock(); |
| |
| return peer; |
| } |
| |
| /** |
| * ovpn_peer_get_by_id_p2p - get peer by ID in a P2P instance |
| * @ovpn: the openvpn instance to search |
| * @peer_id: the ID of the peer to find |
| * |
| * Return: the peer if found or NULL otherwise |
| */ |
| static struct ovpn_peer *ovpn_peer_get_by_id_p2p(struct ovpn_priv *ovpn, |
| u32 peer_id) |
| { |
| struct ovpn_peer *tmp, *peer = NULL; |
| |
| rcu_read_lock(); |
| tmp = rcu_dereference(ovpn->peer); |
| if (likely(tmp && tmp->id == peer_id && ovpn_peer_hold(tmp))) |
| peer = tmp; |
| rcu_read_unlock(); |
| |
| return peer; |
| } |
| |
| /** |
| * ovpn_peer_get_by_id - retrieve peer by ID |
| * @ovpn: the openvpn instance to search |
| * @peer_id: the unique peer identifier to match |
| * |
| * Return: a pointer to the peer if found or NULL otherwise |
| */ |
| struct ovpn_peer *ovpn_peer_get_by_id(struct ovpn_priv *ovpn, u32 peer_id) |
| { |
| struct ovpn_peer *tmp, *peer = NULL; |
| struct hlist_head *head; |
| |
| if (ovpn->mode == OVPN_MODE_P2P) |
| return ovpn_peer_get_by_id_p2p(ovpn, peer_id); |
| |
| head = ovpn_get_hash_head(ovpn->peers->by_id, &peer_id, |
| sizeof(peer_id)); |
| |
| rcu_read_lock(); |
| hlist_for_each_entry_rcu(tmp, head, hash_entry_id) { |
| if (tmp->id != peer_id) |
| continue; |
| |
| if (!ovpn_peer_hold(tmp)) |
| continue; |
| |
| peer = tmp; |
| break; |
| } |
| rcu_read_unlock(); |
| |
| return peer; |
| } |
| |
| static void ovpn_peer_remove(struct ovpn_peer *peer, |
| enum ovpn_del_peer_reason reason, |
| struct llist_head *release_list) |
| { |
| lockdep_assert_held(&peer->ovpn->lock); |
| |
| switch (peer->ovpn->mode) { |
| case OVPN_MODE_MP: |
| /* prevent double remove */ |
| if (hlist_unhashed(&peer->hash_entry_id)) |
| return; |
| |
| hlist_del_init_rcu(&peer->hash_entry_id); |
| hlist_nulls_del_init_rcu(&peer->hash_entry_addr4); |
| hlist_nulls_del_init_rcu(&peer->hash_entry_addr6); |
| hlist_nulls_del_init_rcu(&peer->hash_entry_transp_addr); |
| break; |
| case OVPN_MODE_P2P: |
| /* prevent double remove */ |
| if (peer != rcu_access_pointer(peer->ovpn->peer)) |
| return; |
| |
| RCU_INIT_POINTER(peer->ovpn->peer, NULL); |
| /* in P2P mode the carrier is switched off when the peer is |
| * deleted so that third party protocols can react accordingly |
| */ |
| netif_carrier_off(peer->ovpn->dev); |
| break; |
| } |
| |
| peer->delete_reason = reason; |
| ovpn_nl_peer_del_notify(peer); |
| |
| /* append to provided list for later socket release and ref drop */ |
| llist_add(&peer->release_entry, release_list); |
| } |
| |
| /** |
| * ovpn_peer_get_by_dst - Lookup peer to send skb to |
| * @ovpn: the private data representing the current VPN session |
| * @skb: the skb to extract the destination address from |
| * |
| * This function takes a tunnel packet and looks up the peer to send it to |
| * after encapsulation. The skb is expected to be the in-tunnel packet, without |
| * any OpenVPN related header. |
| * |
| * Assume that the IP header is accessible in the skb data. |
| * |
| * Return: the peer if found or NULL otherwise. |
| */ |
| struct ovpn_peer *ovpn_peer_get_by_dst(struct ovpn_priv *ovpn, |
| struct sk_buff *skb) |
| { |
| struct ovpn_peer *peer = NULL; |
| struct in6_addr addr6; |
| __be32 addr4; |
| |
| /* in P2P mode, no matter the destination, packets are always sent to |
| * the single peer listening on the other side |
| */ |
| if (ovpn->mode == OVPN_MODE_P2P) { |
| rcu_read_lock(); |
| peer = rcu_dereference(ovpn->peer); |
| if (unlikely(peer && !ovpn_peer_hold(peer))) |
| peer = NULL; |
| rcu_read_unlock(); |
| return peer; |
| } |
| |
| rcu_read_lock(); |
| switch (skb->protocol) { |
| case htons(ETH_P_IP): |
| addr4 = ovpn_nexthop_from_skb4(skb); |
| peer = ovpn_peer_get_by_vpn_addr4(ovpn, addr4); |
| break; |
| case htons(ETH_P_IPV6): |
| addr6 = ovpn_nexthop_from_skb6(skb); |
| peer = ovpn_peer_get_by_vpn_addr6(ovpn, &addr6); |
| break; |
| } |
| |
| if (unlikely(peer && !ovpn_peer_hold(peer))) |
| peer = NULL; |
| rcu_read_unlock(); |
| |
| return peer; |
| } |
| |
| /** |
| * ovpn_nexthop_from_rt4 - look up the IPv4 nexthop for the given destination |
| * @ovpn: the private data representing the current VPN session |
| * @dest: the destination to be looked up |
| * |
| * Looks up in the IPv4 system routing table the IP of the nexthop to be used |
| * to reach the destination passed as argument. If no nexthop can be found, the |
| * destination itself is returned as it probably has to be used as nexthop. |
| * |
| * Return: the IP of the next hop if found or dest itself otherwise |
| */ |
| static __be32 ovpn_nexthop_from_rt4(struct ovpn_priv *ovpn, __be32 dest) |
| { |
| struct rtable *rt; |
| struct flowi4 fl = { |
| .daddr = dest |
| }; |
| |
| rt = ip_route_output_flow(dev_net(ovpn->dev), &fl, NULL); |
| if (IS_ERR(rt)) { |
| net_dbg_ratelimited("%s: no route to host %pI4\n", |
| netdev_name(ovpn->dev), &dest); |
| /* if we end up here this packet is probably going to be |
| * thrown away later |
| */ |
| return dest; |
| } |
| |
| if (!rt->rt_uses_gateway) |
| goto out; |
| |
| dest = rt->rt_gw4; |
| out: |
| ip_rt_put(rt); |
| return dest; |
| } |
| |
| /** |
| * ovpn_nexthop_from_rt6 - look up the IPv6 nexthop for the given destination |
| * @ovpn: the private data representing the current VPN session |
| * @dest: the destination to be looked up |
| * |
| * Looks up in the IPv6 system routing table the IP of the nexthop to be used |
| * to reach the destination passed as argument. If no nexthop can be found, the |
| * destination itself is returned as it probably has to be used as nexthop. |
| * |
| * Return: the IP of the next hop if found or dest itself otherwise |
| */ |
| static struct in6_addr ovpn_nexthop_from_rt6(struct ovpn_priv *ovpn, |
| struct in6_addr dest) |
| { |
| #if IS_ENABLED(CONFIG_IPV6) |
| struct dst_entry *entry; |
| struct rt6_info *rt; |
| struct flowi6 fl = { |
| .daddr = dest, |
| }; |
| |
| entry = ipv6_stub->ipv6_dst_lookup_flow(dev_net(ovpn->dev), NULL, &fl, |
| NULL); |
| if (IS_ERR(entry)) { |
| net_dbg_ratelimited("%s: no route to host %pI6c\n", |
| netdev_name(ovpn->dev), &dest); |
| /* if we end up here this packet is probably going to be |
| * thrown away later |
| */ |
| return dest; |
| } |
| |
| rt = dst_rt6_info(entry); |
| |
| if (!(rt->rt6i_flags & RTF_GATEWAY)) |
| goto out; |
| |
| dest = rt->rt6i_gateway; |
| out: |
| dst_release((struct dst_entry *)rt); |
| #endif |
| return dest; |
| } |
| |
| /** |
| * ovpn_peer_check_by_src - check that skb source is routed via peer |
| * @ovpn: the openvpn instance to search |
| * @skb: the packet to extract source address from |
| * @peer: the peer to check against the source address |
| * |
| * Return: true if the peer is matching or false otherwise |
| */ |
| bool ovpn_peer_check_by_src(struct ovpn_priv *ovpn, struct sk_buff *skb, |
| struct ovpn_peer *peer) |
| { |
| bool match = false; |
| struct in6_addr addr6; |
| __be32 addr4; |
| |
| if (ovpn->mode == OVPN_MODE_P2P) { |
| /* in P2P mode, no matter the destination, packets are always |
| * sent to the single peer listening on the other side |
| */ |
| return peer == rcu_access_pointer(ovpn->peer); |
| } |
| |
| /* This function performs a reverse path check, therefore we now |
| * lookup the nexthop we would use if we wanted to route a packet |
| * to the source IP. If the nexthop matches the sender we know the |
| * latter is valid and we allow the packet to come in |
| */ |
| |
| switch (skb->protocol) { |
| case htons(ETH_P_IP): |
| addr4 = ovpn_nexthop_from_rt4(ovpn, ip_hdr(skb)->saddr); |
| rcu_read_lock(); |
| match = (peer == ovpn_peer_get_by_vpn_addr4(ovpn, addr4)); |
| rcu_read_unlock(); |
| break; |
| case htons(ETH_P_IPV6): |
| addr6 = ovpn_nexthop_from_rt6(ovpn, ipv6_hdr(skb)->saddr); |
| rcu_read_lock(); |
| match = (peer == ovpn_peer_get_by_vpn_addr6(ovpn, &addr6)); |
| rcu_read_unlock(); |
| break; |
| } |
| |
| return match; |
| } |
| |
| void ovpn_peer_hash_vpn_ip(struct ovpn_peer *peer) |
| { |
| struct hlist_nulls_head *nhead; |
| |
| lockdep_assert_held(&peer->ovpn->lock); |
| |
| /* rehashing makes sense only in multipeer mode */ |
| if (peer->ovpn->mode != OVPN_MODE_MP) |
| return; |
| |
| if (peer->vpn_addrs.ipv4.s_addr != htonl(INADDR_ANY)) { |
| /* remove potential old hashing */ |
| hlist_nulls_del_init_rcu(&peer->hash_entry_addr4); |
| |
| nhead = ovpn_get_hash_head(peer->ovpn->peers->by_vpn_addr4, |
| &peer->vpn_addrs.ipv4, |
| sizeof(peer->vpn_addrs.ipv4)); |
| hlist_nulls_add_head_rcu(&peer->hash_entry_addr4, nhead); |
| } |
| |
| if (!ipv6_addr_any(&peer->vpn_addrs.ipv6)) { |
| /* remove potential old hashing */ |
| hlist_nulls_del_init_rcu(&peer->hash_entry_addr6); |
| |
| nhead = ovpn_get_hash_head(peer->ovpn->peers->by_vpn_addr6, |
| &peer->vpn_addrs.ipv6, |
| sizeof(peer->vpn_addrs.ipv6)); |
| hlist_nulls_add_head_rcu(&peer->hash_entry_addr6, nhead); |
| } |
| } |
| |
| /** |
| * ovpn_peer_add_mp - add peer to related tables in a MP instance |
| * @ovpn: the instance to add the peer to |
| * @peer: the peer to add |
| * |
| * Return: 0 on success or a negative error code otherwise |
| */ |
| static int ovpn_peer_add_mp(struct ovpn_priv *ovpn, struct ovpn_peer *peer) |
| { |
| struct sockaddr_storage sa = { 0 }; |
| struct hlist_nulls_head *nhead; |
| struct sockaddr_in6 *sa6; |
| struct sockaddr_in *sa4; |
| struct ovpn_bind *bind; |
| struct ovpn_peer *tmp; |
| size_t salen; |
| int ret = 0; |
| |
| spin_lock_bh(&ovpn->lock); |
| /* do not add duplicates */ |
| tmp = ovpn_peer_get_by_id(ovpn, peer->id); |
| if (tmp) { |
| ovpn_peer_put(tmp); |
| ret = -EEXIST; |
| goto out; |
| } |
| |
| bind = rcu_dereference_protected(peer->bind, true); |
| /* peers connected via TCP have bind == NULL */ |
| if (bind) { |
| switch (bind->remote.in4.sin_family) { |
| case AF_INET: |
| sa4 = (struct sockaddr_in *)&sa; |
| |
| sa4->sin_family = AF_INET; |
| sa4->sin_addr.s_addr = bind->remote.in4.sin_addr.s_addr; |
| sa4->sin_port = bind->remote.in4.sin_port; |
| salen = sizeof(*sa4); |
| break; |
| case AF_INET6: |
| sa6 = (struct sockaddr_in6 *)&sa; |
| |
| sa6->sin6_family = AF_INET6; |
| sa6->sin6_addr = bind->remote.in6.sin6_addr; |
| sa6->sin6_port = bind->remote.in6.sin6_port; |
| salen = sizeof(*sa6); |
| break; |
| default: |
| ret = -EPROTONOSUPPORT; |
| goto out; |
| } |
| |
| nhead = ovpn_get_hash_head(ovpn->peers->by_transp_addr, &sa, |
| salen); |
| hlist_nulls_add_head_rcu(&peer->hash_entry_transp_addr, nhead); |
| } |
| |
| hlist_add_head_rcu(&peer->hash_entry_id, |
| ovpn_get_hash_head(ovpn->peers->by_id, &peer->id, |
| sizeof(peer->id))); |
| |
| ovpn_peer_hash_vpn_ip(peer); |
| out: |
| spin_unlock_bh(&ovpn->lock); |
| return ret; |
| } |
| |
| /** |
| * ovpn_peer_add_p2p - add peer to related tables in a P2P instance |
| * @ovpn: the instance to add the peer to |
| * @peer: the peer to add |
| * |
| * Return: 0 on success or a negative error code otherwise |
| */ |
| static int ovpn_peer_add_p2p(struct ovpn_priv *ovpn, struct ovpn_peer *peer) |
| { |
| LLIST_HEAD(release_list); |
| struct ovpn_peer *tmp; |
| |
| spin_lock_bh(&ovpn->lock); |
| /* in p2p mode it is possible to have a single peer only, therefore the |
| * old one is released and substituted by the new one |
| */ |
| tmp = rcu_dereference_protected(ovpn->peer, |
| lockdep_is_held(&ovpn->lock)); |
| if (tmp) |
| ovpn_peer_remove(tmp, OVPN_DEL_PEER_REASON_TEARDOWN, |
| &release_list); |
| |
| rcu_assign_pointer(ovpn->peer, peer); |
| /* in P2P mode the carrier is switched on when the peer is added */ |
| netif_carrier_on(ovpn->dev); |
| unlock_ovpn(ovpn, &release_list); |
| |
| return 0; |
| } |
| |
| /** |
| * ovpn_peer_add - add peer to the related tables |
| * @ovpn: the openvpn instance the peer belongs to |
| * @peer: the peer object to add |
| * |
| * Assume refcounter was increased by caller |
| * |
| * Return: 0 on success or a negative error code otherwise |
| */ |
| int ovpn_peer_add(struct ovpn_priv *ovpn, struct ovpn_peer *peer) |
| { |
| switch (ovpn->mode) { |
| case OVPN_MODE_MP: |
| return ovpn_peer_add_mp(ovpn, peer); |
| case OVPN_MODE_P2P: |
| return ovpn_peer_add_p2p(ovpn, peer); |
| } |
| |
| return -EOPNOTSUPP; |
| } |
| |
| /** |
| * ovpn_peer_del_mp - delete peer from related tables in a MP instance |
| * @peer: the peer to delete |
| * @reason: reason why the peer was deleted (sent to userspace) |
| * @release_list: list where delete peer should be appended |
| * |
| * Return: 0 on success or a negative error code otherwise |
| */ |
| static int ovpn_peer_del_mp(struct ovpn_peer *peer, |
| enum ovpn_del_peer_reason reason, |
| struct llist_head *release_list) |
| { |
| struct ovpn_peer *tmp; |
| int ret = -ENOENT; |
| |
| lockdep_assert_held(&peer->ovpn->lock); |
| |
| tmp = ovpn_peer_get_by_id(peer->ovpn, peer->id); |
| if (tmp == peer) { |
| ovpn_peer_remove(peer, reason, release_list); |
| ret = 0; |
| } |
| |
| if (tmp) |
| ovpn_peer_put(tmp); |
| |
| return ret; |
| } |
| |
| /** |
| * ovpn_peer_del_p2p - delete peer from related tables in a P2P instance |
| * @peer: the peer to delete |
| * @reason: reason why the peer was deleted (sent to userspace) |
| * @release_list: list where delete peer should be appended |
| * |
| * Return: 0 on success or a negative error code otherwise |
| */ |
| static int ovpn_peer_del_p2p(struct ovpn_peer *peer, |
| enum ovpn_del_peer_reason reason, |
| struct llist_head *release_list) |
| { |
| struct ovpn_peer *tmp; |
| |
| lockdep_assert_held(&peer->ovpn->lock); |
| |
| tmp = rcu_dereference_protected(peer->ovpn->peer, |
| lockdep_is_held(&peer->ovpn->lock)); |
| if (tmp != peer) |
| return -ENOENT; |
| |
| ovpn_peer_remove(peer, reason, release_list); |
| |
| return 0; |
| } |
| |
| /** |
| * ovpn_peer_del - delete peer from related tables |
| * @peer: the peer object to delete |
| * @reason: reason for deleting peer (will be sent to userspace) |
| * |
| * Return: 0 on success or a negative error code otherwise |
| */ |
| int ovpn_peer_del(struct ovpn_peer *peer, enum ovpn_del_peer_reason reason) |
| { |
| LLIST_HEAD(release_list); |
| int ret = -EOPNOTSUPP; |
| |
| spin_lock_bh(&peer->ovpn->lock); |
| switch (peer->ovpn->mode) { |
| case OVPN_MODE_MP: |
| ret = ovpn_peer_del_mp(peer, reason, &release_list); |
| break; |
| case OVPN_MODE_P2P: |
| ret = ovpn_peer_del_p2p(peer, reason, &release_list); |
| break; |
| default: |
| break; |
| } |
| unlock_ovpn(peer->ovpn, &release_list); |
| |
| return ret; |
| } |
| |
| /** |
| * ovpn_peer_release_p2p - release peer upon P2P device teardown |
| * @ovpn: the instance being torn down |
| * @sk: if not NULL, release peer only if it's using this specific socket |
| * @reason: the reason for releasing the peer |
| */ |
| static void ovpn_peer_release_p2p(struct ovpn_priv *ovpn, struct sock *sk, |
| enum ovpn_del_peer_reason reason) |
| { |
| struct ovpn_socket *ovpn_sock; |
| LLIST_HEAD(release_list); |
| struct ovpn_peer *peer; |
| |
| spin_lock_bh(&ovpn->lock); |
| peer = rcu_dereference_protected(ovpn->peer, |
| lockdep_is_held(&ovpn->lock)); |
| if (!peer) { |
| spin_unlock_bh(&ovpn->lock); |
| return; |
| } |
| |
| if (sk) { |
| ovpn_sock = rcu_access_pointer(peer->sock); |
| if (!ovpn_sock || ovpn_sock->sk != sk) { |
| spin_unlock_bh(&ovpn->lock); |
| ovpn_peer_put(peer); |
| return; |
| } |
| } |
| |
| ovpn_peer_remove(peer, reason, &release_list); |
| unlock_ovpn(ovpn, &release_list); |
| } |
| |
| static void ovpn_peers_release_mp(struct ovpn_priv *ovpn, struct sock *sk, |
| enum ovpn_del_peer_reason reason) |
| { |
| struct ovpn_socket *ovpn_sock; |
| LLIST_HEAD(release_list); |
| struct ovpn_peer *peer; |
| struct hlist_node *tmp; |
| int bkt; |
| |
| spin_lock_bh(&ovpn->lock); |
| hash_for_each_safe(ovpn->peers->by_id, bkt, tmp, peer, hash_entry_id) { |
| bool remove = true; |
| |
| /* if a socket was passed as argument, skip all peers except |
| * those using it |
| */ |
| if (sk) { |
| rcu_read_lock(); |
| ovpn_sock = rcu_dereference(peer->sock); |
| remove = ovpn_sock && ovpn_sock->sk == sk; |
| rcu_read_unlock(); |
| } |
| |
| if (remove) |
| ovpn_peer_remove(peer, reason, &release_list); |
| } |
| unlock_ovpn(ovpn, &release_list); |
| } |
| |
| /** |
| * ovpn_peers_free - free all peers in the instance |
| * @ovpn: the instance whose peers should be released |
| * @sk: if not NULL, only peers using this socket are removed and the socket |
| * is released immediately |
| * @reason: the reason for releasing all peers |
| */ |
| void ovpn_peers_free(struct ovpn_priv *ovpn, struct sock *sk, |
| enum ovpn_del_peer_reason reason) |
| { |
| switch (ovpn->mode) { |
| case OVPN_MODE_P2P: |
| ovpn_peer_release_p2p(ovpn, sk, reason); |
| break; |
| case OVPN_MODE_MP: |
| ovpn_peers_release_mp(ovpn, sk, reason); |
| break; |
| } |
| } |
| |
| static time64_t ovpn_peer_keepalive_work_single(struct ovpn_peer *peer, |
| time64_t now, |
| struct llist_head *release_list) |
| { |
| time64_t last_recv, last_sent, next_run1, next_run2; |
| unsigned long timeout, interval; |
| bool expired; |
| |
| spin_lock_bh(&peer->lock); |
| /* we expect both timers to be configured at the same time, |
| * therefore bail out if either is not set |
| */ |
| if (!peer->keepalive_timeout || !peer->keepalive_interval) { |
| spin_unlock_bh(&peer->lock); |
| return 0; |
| } |
| |
| /* check for peer timeout */ |
| expired = false; |
| timeout = peer->keepalive_timeout; |
| last_recv = READ_ONCE(peer->last_recv); |
| if (now < last_recv + timeout) { |
| peer->keepalive_recv_exp = last_recv + timeout; |
| next_run1 = peer->keepalive_recv_exp; |
| } else if (peer->keepalive_recv_exp > now) { |
| next_run1 = peer->keepalive_recv_exp; |
| } else { |
| expired = true; |
| } |
| |
| if (expired) { |
| /* peer is dead -> kill it and move on */ |
| spin_unlock_bh(&peer->lock); |
| netdev_dbg(peer->ovpn->dev, "peer %u expired\n", |
| peer->id); |
| ovpn_peer_remove(peer, OVPN_DEL_PEER_REASON_EXPIRED, |
| release_list); |
| return 0; |
| } |
| |
| /* check for peer keepalive */ |
| expired = false; |
| interval = peer->keepalive_interval; |
| last_sent = READ_ONCE(peer->last_sent); |
| if (now < last_sent + interval) { |
| peer->keepalive_xmit_exp = last_sent + interval; |
| next_run2 = peer->keepalive_xmit_exp; |
| } else if (peer->keepalive_xmit_exp > now) { |
| next_run2 = peer->keepalive_xmit_exp; |
| } else { |
| expired = true; |
| next_run2 = now + interval; |
| } |
| spin_unlock_bh(&peer->lock); |
| |
| if (expired) { |
| /* a keepalive packet is required */ |
| netdev_dbg(peer->ovpn->dev, |
| "sending keepalive to peer %u\n", |
| peer->id); |
| if (schedule_work(&peer->keepalive_work)) |
| ovpn_peer_hold(peer); |
| } |
| |
| if (next_run1 < next_run2) |
| return next_run1; |
| |
| return next_run2; |
| } |
| |
| static time64_t ovpn_peer_keepalive_work_mp(struct ovpn_priv *ovpn, |
| time64_t now, |
| struct llist_head *release_list) |
| { |
| time64_t tmp_next_run, next_run = 0; |
| struct hlist_node *tmp; |
| struct ovpn_peer *peer; |
| int bkt; |
| |
| lockdep_assert_held(&ovpn->lock); |
| |
| hash_for_each_safe(ovpn->peers->by_id, bkt, tmp, peer, hash_entry_id) { |
| tmp_next_run = ovpn_peer_keepalive_work_single(peer, now, |
| release_list); |
| if (!tmp_next_run) |
| continue; |
| |
| /* the next worker run will be scheduled based on the shortest |
| * required interval across all peers |
| */ |
| if (!next_run || tmp_next_run < next_run) |
| next_run = tmp_next_run; |
| } |
| |
| return next_run; |
| } |
| |
| static time64_t ovpn_peer_keepalive_work_p2p(struct ovpn_priv *ovpn, |
| time64_t now, |
| struct llist_head *release_list) |
| { |
| struct ovpn_peer *peer; |
| time64_t next_run = 0; |
| |
| lockdep_assert_held(&ovpn->lock); |
| |
| peer = rcu_dereference_protected(ovpn->peer, |
| lockdep_is_held(&ovpn->lock)); |
| if (peer) |
| next_run = ovpn_peer_keepalive_work_single(peer, now, |
| release_list); |
| |
| return next_run; |
| } |
| |
| /** |
| * ovpn_peer_keepalive_work - run keepalive logic on each known peer |
| * @work: pointer to the work member of the related ovpn object |
| * |
| * Each peer has two timers (if configured): |
| * 1. peer timeout: when no data is received for a certain interval, |
| * the peer is considered dead and it gets killed. |
| * 2. peer keepalive: when no data is sent to a certain peer for a |
| * certain interval, a special 'keepalive' packet is explicitly sent. |
| * |
| * This function iterates across the whole peer collection while |
| * checking the timers described above. |
| */ |
| void ovpn_peer_keepalive_work(struct work_struct *work) |
| { |
| struct ovpn_priv *ovpn = container_of(work, struct ovpn_priv, |
| keepalive_work.work); |
| time64_t next_run = 0, now = ktime_get_real_seconds(); |
| LLIST_HEAD(release_list); |
| |
| spin_lock_bh(&ovpn->lock); |
| switch (ovpn->mode) { |
| case OVPN_MODE_MP: |
| next_run = ovpn_peer_keepalive_work_mp(ovpn, now, |
| &release_list); |
| break; |
| case OVPN_MODE_P2P: |
| next_run = ovpn_peer_keepalive_work_p2p(ovpn, now, |
| &release_list); |
| break; |
| } |
| |
| /* prevent rearming if the interface is being destroyed */ |
| if (next_run > 0) { |
| netdev_dbg(ovpn->dev, |
| "scheduling keepalive work: now=%llu next_run=%llu delta=%llu\n", |
| next_run, now, next_run - now); |
| schedule_delayed_work(&ovpn->keepalive_work, |
| (next_run - now) * HZ); |
| } |
| unlock_ovpn(ovpn, &release_list); |
| } |