OSDN Git Service

drivers: vxlan: vnifilter: add support for stats dumping
authorNikolay Aleksandrov <nikolay@nvidia.com>
Tue, 1 Mar 2022 05:04:39 +0000 (05:04 +0000)
committerDavid S. Miller <davem@davemloft.net>
Tue, 1 Mar 2022 08:38:02 +0000 (08:38 +0000)
Add support for VXLAN vni filter entries' stats dumping

Signed-off-by: Nikolay Aleksandrov <nikolay@nvidia.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
drivers/net/vxlan/vxlan_vnifilter.c
include/uapi/linux/if_link.h

index 2d23312..9f28d0b 100644 (file)
@@ -116,6 +116,34 @@ void vxlan_vs_del_vnigrp(struct vxlan_dev *vxlan)
        spin_unlock(&vn->sock_lock);
 }
 
+static void vxlan_vnifilter_stats_get(const struct vxlan_vni_node *vninode,
+                                     struct vxlan_vni_stats *dest)
+{
+       int i;
+
+       memset(dest, 0, sizeof(*dest));
+       for_each_possible_cpu(i) {
+               struct vxlan_vni_stats_pcpu *pstats;
+               struct vxlan_vni_stats temp;
+               unsigned int start;
+
+               pstats = per_cpu_ptr(vninode->stats, i);
+               do {
+                       start = u64_stats_fetch_begin_irq(&pstats->syncp);
+                       memcpy(&temp, &pstats->stats, sizeof(temp));
+               } while (u64_stats_fetch_retry_irq(&pstats->syncp, start));
+
+               dest->rx_packets += temp.rx_packets;
+               dest->rx_bytes += temp.rx_bytes;
+               dest->rx_drops += temp.rx_drops;
+               dest->rx_errors += temp.rx_errors;
+               dest->tx_packets += temp.tx_packets;
+               dest->tx_bytes += temp.tx_bytes;
+               dest->tx_drops += temp.tx_drops;
+               dest->tx_errors += temp.tx_errors;
+       }
+}
+
 static void vxlan_vnifilter_stats_add(struct vxlan_vni_node *vninode,
                                      int type, unsigned int len)
 {
@@ -182,9 +210,48 @@ static size_t vxlan_vnifilter_entry_nlmsg_size(void)
                + nla_total_size(sizeof(struct in6_addr));/* VXLAN_VNIFILTER_ENTRY_GROUP{6} */
 }
 
+static int __vnifilter_entry_fill_stats(struct sk_buff *skb,
+                                       const struct vxlan_vni_node *vbegin)
+{
+       struct vxlan_vni_stats vstats;
+       struct nlattr *vstats_attr;
+
+       vstats_attr = nla_nest_start(skb, VXLAN_VNIFILTER_ENTRY_STATS);
+       if (!vstats_attr)
+               goto out_stats_err;
+
+       vxlan_vnifilter_stats_get(vbegin, &vstats);
+       if (nla_put_u64_64bit(skb, VNIFILTER_ENTRY_STATS_RX_BYTES,
+                             vstats.rx_bytes, VNIFILTER_ENTRY_STATS_PAD) ||
+           nla_put_u64_64bit(skb, VNIFILTER_ENTRY_STATS_RX_PKTS,
+                             vstats.rx_packets, VNIFILTER_ENTRY_STATS_PAD) ||
+           nla_put_u64_64bit(skb, VNIFILTER_ENTRY_STATS_RX_DROPS,
+                             vstats.rx_drops, VNIFILTER_ENTRY_STATS_PAD) ||
+           nla_put_u64_64bit(skb, VNIFILTER_ENTRY_STATS_RX_ERRORS,
+                             vstats.rx_errors, VNIFILTER_ENTRY_STATS_PAD) ||
+           nla_put_u64_64bit(skb, VNIFILTER_ENTRY_STATS_TX_BYTES,
+                             vstats.tx_bytes, VNIFILTER_ENTRY_STATS_PAD) ||
+           nla_put_u64_64bit(skb, VNIFILTER_ENTRY_STATS_TX_PKTS,
+                             vstats.tx_packets, VNIFILTER_ENTRY_STATS_PAD) ||
+           nla_put_u64_64bit(skb, VNIFILTER_ENTRY_STATS_TX_DROPS,
+                             vstats.tx_drops, VNIFILTER_ENTRY_STATS_PAD) ||
+           nla_put_u64_64bit(skb, VNIFILTER_ENTRY_STATS_TX_ERRORS,
+                             vstats.tx_errors, VNIFILTER_ENTRY_STATS_PAD))
+               goto out_stats_err;
+
+       nla_nest_end(skb, vstats_attr);
+
+       return 0;
+
+out_stats_err:
+       nla_nest_cancel(skb, vstats_attr);
+       return -EMSGSIZE;
+}
+
 static bool vxlan_fill_vni_filter_entry(struct sk_buff *skb,
                                        struct vxlan_vni_node *vbegin,
-                                       struct vxlan_vni_node *vend)
+                                       struct vxlan_vni_node *vend,
+                                       bool fill_stats)
 {
        struct nlattr *ventry;
        u32 vs = be32_to_cpu(vbegin->vni);
@@ -217,6 +284,9 @@ static bool vxlan_fill_vni_filter_entry(struct sk_buff *skb,
                }
        }
 
+       if (fill_stats && __vnifilter_entry_fill_stats(skb, vbegin))
+               goto out_err;
+
        nla_nest_end(skb, ventry);
 
        return true;
@@ -249,7 +319,7 @@ static void vxlan_vnifilter_notify(const struct vxlan_dev *vxlan,
        tmsg->family = AF_BRIDGE;
        tmsg->ifindex = vxlan->dev->ifindex;
 
-       if (!vxlan_fill_vni_filter_entry(skb, vninode, vninode))
+       if (!vxlan_fill_vni_filter_entry(skb, vninode, vninode, false))
                goto out_err;
 
        nlmsg_end(skb, nlh);
@@ -269,10 +339,11 @@ static int vxlan_vnifilter_dump_dev(const struct net_device *dev,
 {
        struct vxlan_vni_node *tmp, *v, *vbegin = NULL, *vend = NULL;
        struct vxlan_dev *vxlan = netdev_priv(dev);
-       struct tunnel_msg *new_tmsg;
+       struct tunnel_msg *new_tmsg, *tmsg;
        int idx = 0, s_idx = cb->args[1];
        struct vxlan_vni_group *vg;
        struct nlmsghdr *nlh;
+       bool dump_stats;
        int err = 0;
 
        if (!(vxlan->cfg.flags & VXLAN_F_VNIFILTER))
@@ -283,6 +354,9 @@ static int vxlan_vnifilter_dump_dev(const struct net_device *dev,
        if (!vg || !vg->num_vnis)
                return 0;
 
+       tmsg = nlmsg_data(cb->nlh);
+       dump_stats = !!(tmsg->flags & TUNNEL_MSG_FLAG_STATS);
+
        nlh = nlmsg_put(skb, NETLINK_CB(cb->skb).portid, cb->nlh->nlmsg_seq,
                        RTM_NEWTUNNEL, sizeof(*new_tmsg), NLM_F_MULTI);
        if (!nlh)
@@ -302,11 +376,12 @@ static int vxlan_vnifilter_dump_dev(const struct net_device *dev,
                        vend = v;
                        continue;
                }
-               if (vnirange(vend, v) == 1 &&
+               if (!dump_stats && vnirange(vend, v) == 1 &&
                    vxlan_addr_equal(&v->remote_ip, &vend->remote_ip)) {
                        goto update_end;
                } else {
-                       if (!vxlan_fill_vni_filter_entry(skb, vbegin, vend)) {
+                       if (!vxlan_fill_vni_filter_entry(skb, vbegin, vend,
+                                                        dump_stats)) {
                                err = -EMSGSIZE;
                                break;
                        }
@@ -318,7 +393,7 @@ update_end:
        }
 
        if (!err && vbegin) {
-               if (!vxlan_fill_vni_filter_entry(skb, vbegin, vend))
+               if (!vxlan_fill_vni_filter_entry(skb, vbegin, vend, dump_stats))
                        err = -EMSGSIZE;
        }
 
@@ -338,6 +413,11 @@ static int vxlan_vnifilter_dump(struct sk_buff *skb, struct netlink_callback *cb
 
        tmsg = nlmsg_data(cb->nlh);
 
+       if (tmsg->flags & ~TUNNEL_MSG_VALID_USER_FLAGS) {
+               NL_SET_ERR_MSG(cb->extack, "Invalid tunnelmsg flags in ancillary header");
+               return -EINVAL;
+       }
+
        rcu_read_lock();
        if (tmsg->ifindex) {
                dev = dev_get_by_index_rcu(net, tmsg->ifindex);
index 3dfc9ff..e315e53 100644 (file)
@@ -716,18 +716,41 @@ enum ipvlan_mode {
 /* Tunnel RTM header */
 struct tunnel_msg {
        __u8 family;
-       __u8 reserved1;
+       __u8 flags;
        __u16 reserved2;
        __u32 ifindex;
 };
 
 /* VXLAN section */
+
+/* include statistics in the dump */
+#define TUNNEL_MSG_FLAG_STATS  0x01
+
+#define TUNNEL_MSG_VALID_USER_FLAGS TUNNEL_MSG_FLAG_STATS
+
+/* Embedded inside VXLAN_VNIFILTER_ENTRY_STATS */
+enum {
+       VNIFILTER_ENTRY_STATS_UNSPEC,
+       VNIFILTER_ENTRY_STATS_RX_BYTES,
+       VNIFILTER_ENTRY_STATS_RX_PKTS,
+       VNIFILTER_ENTRY_STATS_RX_DROPS,
+       VNIFILTER_ENTRY_STATS_RX_ERRORS,
+       VNIFILTER_ENTRY_STATS_TX_BYTES,
+       VNIFILTER_ENTRY_STATS_TX_PKTS,
+       VNIFILTER_ENTRY_STATS_TX_DROPS,
+       VNIFILTER_ENTRY_STATS_TX_ERRORS,
+       VNIFILTER_ENTRY_STATS_PAD,
+       __VNIFILTER_ENTRY_STATS_MAX
+};
+#define VNIFILTER_ENTRY_STATS_MAX (__VNIFILTER_ENTRY_STATS_MAX - 1)
+
 enum {
        VXLAN_VNIFILTER_ENTRY_UNSPEC,
        VXLAN_VNIFILTER_ENTRY_START,
        VXLAN_VNIFILTER_ENTRY_END,
        VXLAN_VNIFILTER_ENTRY_GROUP,
        VXLAN_VNIFILTER_ENTRY_GROUP6,
+       VXLAN_VNIFILTER_ENTRY_STATS,
        __VXLAN_VNIFILTER_ENTRY_MAX
 };
 #define VXLAN_VNIFILTER_ENTRY_MAX      (__VXLAN_VNIFILTER_ENTRY_MAX - 1)