OSDN Git Service

net: atlantic: QoS implementation: multi-TC support
authorDmitry Bezrukov <dbezrukov@marvell.com>
Fri, 22 May 2020 08:19:40 +0000 (11:19 +0300)
committerDavid S. Miller <davem@davemloft.net>
Fri, 22 May 2020 21:08:28 +0000 (14:08 -0700)
This patch adds multi-TC support.

PTP is automatically disabled when the user enables more than 2 TCs,
otherwise traffic on TC2 won't quite work, because it's reserved for PTP.

Signed-off-by: Dmitry Bezrukov <dbezrukov@marvell.com>
Co-developed-by: Dmitry Bogdanov <dbogdanov@marvell.com>
Signed-off-by: Dmitry Bogdanov <dbogdanov@marvell.com>
Co-developed-by: Mark Starovoytov <mstarovoitov@marvell.com>
Signed-off-by: Mark Starovoytov <mstarovoitov@marvell.com>
Signed-off-by: Igor Russkikh <irusskikh@marvell.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
15 files changed:
drivers/net/ethernet/aquantia/atlantic/aq_filters.c
drivers/net/ethernet/aquantia/atlantic/aq_hw.h
drivers/net/ethernet/aquantia/atlantic/aq_hw_utils.c
drivers/net/ethernet/aquantia/atlantic/aq_hw_utils.h
drivers/net/ethernet/aquantia/atlantic/aq_main.c
drivers/net/ethernet/aquantia/atlantic/aq_nic.c
drivers/net/ethernet/aquantia/atlantic/aq_nic.h
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2.c
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_internal.h
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh.c
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh.h
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh_internal.h

index 03ff92b..1bc4d33 100644 (file)
@@ -153,6 +153,8 @@ aq_check_approve_fvlan(struct aq_nic_s *aq_nic,
                       struct aq_hw_rx_fltrs_s *rx_fltrs,
                       struct ethtool_rx_flow_spec *fsp)
 {
+       struct aq_nic_cfg_s *cfg = &aq_nic->aq_nic_cfg;
+
        if (fsp->location < AQ_RX_FIRST_LOC_FVLANID ||
            fsp->location > AQ_RX_LAST_LOC_FVLANID) {
                netdev_err(aq_nic->ndev,
@@ -170,10 +172,10 @@ aq_check_approve_fvlan(struct aq_nic_s *aq_nic,
                return -EINVAL;
        }
 
-       if (fsp->ring_cookie > aq_nic->aq_nic_cfg.num_rss_queues) {
+       if (fsp->ring_cookie > cfg->num_rss_queues * cfg->tcs) {
                netdev_err(aq_nic->ndev,
                           "ethtool: queue number must be in range [0, %d]",
-                          aq_nic->aq_nic_cfg.num_rss_queues - 1);
+                          cfg->num_rss_queues * cfg->tcs - 1);
                return -EINVAL;
        }
        return 0;
@@ -262,6 +264,7 @@ static bool __must_check
 aq_rule_is_not_correct(struct aq_nic_s *aq_nic,
                       struct ethtool_rx_flow_spec *fsp)
 {
+       struct aq_nic_cfg_s *cfg = &aq_nic->aq_nic_cfg;
        bool rule_is_not_correct = false;
 
        if (!aq_nic) {
@@ -274,11 +277,11 @@ aq_rule_is_not_correct(struct aq_nic_s *aq_nic,
        } else if (aq_check_filter(aq_nic, fsp)) {
                rule_is_not_correct = true;
        } else if (fsp->ring_cookie != RX_CLS_FLOW_DISC) {
-               if (fsp->ring_cookie >= aq_nic->aq_nic_cfg.num_rss_queues) {
+               if (fsp->ring_cookie >= cfg->num_rss_queues * cfg->tcs) {
                        netdev_err(aq_nic->ndev,
                                   "ethtool: The specified action is invalid.\n"
                                   "Maximum allowable value action is %u.\n",
-                                  aq_nic->aq_nic_cfg.num_rss_queues - 1);
+                                  cfg->num_rss_queues * cfg->tcs - 1);
                        rule_is_not_correct = true;
                }
        }
index c3df9da..1dccaae 100644 (file)
@@ -124,6 +124,7 @@ struct aq_stats_s {
 #define AQ_HW_TXD_MULTIPLE 8U
 #define AQ_HW_RXD_MULTIPLE 8U
 
+#define AQ_HW_QUEUES_MAX                32U
 #define AQ_HW_MULTICAST_ADDRESS_MAX     32U
 
 #define AQ_HW_PTP_TC                    2U
index 7dbf49a..342c517 100644 (file)
@@ -79,3 +79,29 @@ int aq_hw_err_from_flags(struct aq_hw_s *hw)
 err_exit:
        return err;
 }
+
+int aq_hw_num_tcs(struct aq_hw_s *hw)
+{
+       switch (hw->aq_nic_cfg->tc_mode) {
+       case AQ_TC_MODE_8TCS:
+               return 8;
+       case AQ_TC_MODE_4TCS:
+               return 4;
+       default:
+               break;
+       }
+
+       return 1;
+}
+
+int aq_hw_q_per_tc(struct aq_hw_s *hw)
+{
+       switch (hw->aq_nic_cfg->tc_mode) {
+       case AQ_TC_MODE_8TCS:
+               return 4;
+       case AQ_TC_MODE_4TCS:
+               return 8;
+       default:
+               return 4;
+       }
+}
index 9ef82d4..32aa5f2 100644 (file)
@@ -34,5 +34,7 @@ u32 aq_hw_read_reg(struct aq_hw_s *hw, u32 reg);
 void aq_hw_write_reg(struct aq_hw_s *hw, u32 reg, u32 value);
 u64 aq_hw_read_reg64(struct aq_hw_s *hw, u32 reg);
 int aq_hw_err_from_flags(struct aq_hw_s *hw);
+int aq_hw_num_tcs(struct aq_hw_s *hw);
+int aq_hw_q_per_tc(struct aq_hw_s *hw);
 
 #endif /* AQ_HW_UTILS_H */
index 9fcab64..ef9e969 100644 (file)
 #include "aq_ethtool.h"
 #include "aq_ptp.h"
 #include "aq_filters.h"
+#include "aq_hw_utils.h"
 
 #include <linux/netdevice.h>
 #include <linux/module.h>
 #include <linux/ip.h>
 #include <linux/udp.h>
+#include <net/pkt_cls.h>
 
 MODULE_LICENSE("GPL v2");
 MODULE_AUTHOR(AQ_CFG_DRV_AUTHOR);
@@ -38,7 +40,7 @@ struct net_device *aq_ndev_alloc(void)
        struct net_device *ndev = NULL;
        struct aq_nic_s *aq_nic = NULL;
 
-       ndev = alloc_etherdev_mq(sizeof(struct aq_nic_s), AQ_CFG_VECS_MAX);
+       ndev = alloc_etherdev_mq(sizeof(struct aq_nic_s), AQ_HW_QUEUES_MAX);
        if (!ndev)
                return NULL;
 
@@ -330,6 +332,40 @@ static int aq_ndo_vlan_rx_kill_vid(struct net_device *ndev, __be16 proto,
        return 0;
 }
 
+static int aq_validate_mqprio_opt(struct aq_nic_s *self,
+                                 const unsigned int num_tc)
+{
+       if (num_tc > aq_hw_num_tcs(self->aq_hw)) {
+               netdev_err(self->ndev, "Too many TCs requested\n");
+               return -EOPNOTSUPP;
+       }
+
+       if (num_tc != 0 && !is_power_of_2(num_tc)) {
+               netdev_err(self->ndev, "TC count should be power of 2\n");
+               return -EOPNOTSUPP;
+       }
+
+       return 0;
+}
+
+static int aq_ndo_setup_tc(struct net_device *dev, enum tc_setup_type type,
+                          void *type_data)
+{
+       struct aq_nic_s *aq_nic = netdev_priv(dev);
+       struct tc_mqprio_qopt *mqprio = type_data;
+       int err;
+
+       if (type != TC_SETUP_QDISC_MQPRIO)
+               return -EOPNOTSUPP;
+
+       err = aq_validate_mqprio_opt(aq_nic, mqprio->num_tc);
+       if (err)
+               return err;
+
+       return aq_nic_setup_tc_mqprio(aq_nic, mqprio->num_tc,
+                                     mqprio->prio_tc_map);
+}
+
 static const struct net_device_ops aq_ndev_ops = {
        .ndo_open = aq_ndev_open,
        .ndo_stop = aq_ndev_close,
@@ -341,6 +377,7 @@ static const struct net_device_ops aq_ndev_ops = {
        .ndo_do_ioctl = aq_ndev_ioctl,
        .ndo_vlan_rx_add_vid = aq_ndo_vlan_rx_add_vid,
        .ndo_vlan_rx_kill_vid = aq_ndo_vlan_rx_kill_vid,
+       .ndo_setup_tc = aq_ndo_setup_tc,
 };
 
 static int __init aq_ndev_init_module(void)
index 3eeb652..da78108 100644 (file)
@@ -26,6 +26,7 @@
 #include <linux/ip.h>
 #include <linux/tcp.h>
 #include <net/ip.h>
+#include <net/pkt_cls.h>
 
 static unsigned int aq_itr = AQ_CFG_INTERRUPT_MODERATION_AUTO;
 module_param_named(aq_itr, aq_itr, uint, 0644);
@@ -68,6 +69,7 @@ static void aq_nic_rss_init(struct aq_nic_s *self, unsigned int num_rss_queues)
 void aq_nic_cfg_start(struct aq_nic_s *self)
 {
        struct aq_nic_cfg_s *cfg = &self->aq_nic_cfg;
+       int i;
 
        cfg->tcs = AQ_CFG_TCS_DEF;
 
@@ -142,6 +144,9 @@ void aq_nic_cfg_start(struct aq_nic_s *self)
        cfg->is_vlan_rx_strip = !!(cfg->features & NETIF_F_HW_VLAN_CTAG_RX);
        cfg->is_vlan_tx_insert = !!(cfg->features & NETIF_F_HW_VLAN_CTAG_TX);
        cfg->is_vlan_force_promisc = true;
+
+       for (i = 0; i < sizeof(cfg->prio_tc_map); i++)
+               cfg->prio_tc_map[i] = cfg->tcs * i / 8;
 }
 
 static int aq_nic_update_link_status(struct aq_nic_s *self)
@@ -517,14 +522,21 @@ int aq_nic_start(struct aq_nic_s *self)
                        goto err_exit;
        }
 
-       err = netif_set_real_num_tx_queues(self->ndev, self->aq_vecs);
+       err = netif_set_real_num_tx_queues(self->ndev,
+                                          self->aq_vecs * cfg->tcs);
        if (err < 0)
                goto err_exit;
 
-       err = netif_set_real_num_rx_queues(self->ndev, self->aq_vecs);
+       err = netif_set_real_num_rx_queues(self->ndev,
+                                          self->aq_vecs * cfg->tcs);
        if (err < 0)
                goto err_exit;
 
+       for (i = 0; i < cfg->tcs; i++) {
+               u16 offset = self->aq_vecs * i;
+
+               netdev_set_tc_queue(self->ndev, i, self->aq_vecs, offset);
+       }
        netif_tx_start_all_queues(self->ndev);
 
 err_exit:
@@ -690,10 +702,10 @@ exit:
 int aq_nic_xmit(struct aq_nic_s *self, struct sk_buff *skb)
 {
        unsigned int vec = skb->queue_mapping % self->aq_nic_cfg.vecs;
+       unsigned int tc = skb->queue_mapping / self->aq_nic_cfg.vecs;
        struct aq_ring_s *ring = NULL;
        unsigned int frags = 0U;
        int err = NETDEV_TX_OK;
-       unsigned int tc = 0U;
 
        frags = skb_shinfo(skb)->nr_frags + 1;
 
@@ -712,7 +724,8 @@ int aq_nic_xmit(struct aq_nic_s *self, struct sk_buff *skb)
        }
 
        /* Above status update may stop the queue. Check this. */
-       if (__netif_subqueue_stopped(self->ndev, ring->idx)) {
+       if (__netif_subqueue_stopped(self->ndev,
+                                    AQ_NIC_RING2QMAP(self, ring->idx))) {
                err = NETDEV_TX_BUSY;
                goto err_exit;
        }
@@ -1266,3 +1279,43 @@ void aq_nic_release_filter(struct aq_nic_s *self, enum aq_rx_filter_type type,
                break;
        }
 }
+
+int aq_nic_setup_tc_mqprio(struct aq_nic_s *self, u32 tcs, u8 *prio_tc_map)
+{
+       struct aq_nic_cfg_s *cfg = &self->aq_nic_cfg;
+       bool ndev_running;
+       int err = 0;
+       int i;
+
+       /* if already the same configuration or
+        * disable request (tcs is 0) and we already is disabled
+        */
+       if (tcs == cfg->tcs || (tcs == 0 && !cfg->is_qos))
+               return 0;
+
+       ndev_running = netif_running(self->ndev);
+       if (ndev_running)
+               dev_close(self->ndev);
+
+       cfg->tcs = tcs;
+       if (cfg->tcs == 0)
+               cfg->tcs = 1;
+       if (prio_tc_map)
+               memcpy(cfg->prio_tc_map, prio_tc_map, sizeof(cfg->prio_tc_map));
+       else
+               for (i = 0; i < sizeof(cfg->prio_tc_map); i++)
+                       cfg->prio_tc_map[i] = cfg->tcs * i / 8;
+
+       cfg->is_qos = (tcs != 0 ? true : false);
+       cfg->is_ptp = (cfg->tcs <= AQ_HW_PTP_TC);
+       if (!cfg->is_ptp)
+               netdev_warn(self->ndev, "%s\n",
+                           "PTP is auto disabled due to requested TC count.");
+
+       netdev_set_num_tc(self->ndev, cfg->tcs);
+
+       if (ndev_running)
+               err = dev_open(self->ndev, NULL);
+
+       return err;
+}
index 3434f82..29e1294 100644 (file)
@@ -59,10 +59,12 @@ struct aq_nic_cfg_s {
        bool is_polling;
        bool is_rss;
        bool is_lro;
+       bool is_qos;
        bool is_ptp;
        enum aq_tc_mode tc_mode;
        u32 priv_flags;
        u8  tcs;
+       u8 prio_tc_map[8];
        struct aq_rss_parameters aq_rss;
        u32 eee_speeds;
 };
@@ -79,8 +81,15 @@ struct aq_nic_cfg_s {
 #define AQ_NIC_WOL_MODES        (WAKE_MAGIC |\
                                 WAKE_PHY)
 
+#define AQ_NIC_RING_PER_TC(_NIC_) \
+       (((_NIC_)->aq_nic_cfg.tc_mode == AQ_TC_MODE_4TCS) ? 8 : 4)
+
 #define AQ_NIC_TCVEC2RING(_NIC_, _TC_, _VEC_) \
-       ((_TC_) * AQ_CFG_TCS_MAX + (_VEC_))
+       ((_TC_) * AQ_NIC_RING_PER_TC(_NIC_) + (_VEC_))
+
+#define AQ_NIC_RING2QMAP(_NIC_, _ID_) \
+       ((_ID_) / AQ_NIC_RING_PER_TC(_NIC_) * (_NIC_)->aq_vecs + \
+       ((_ID_) % AQ_NIC_RING_PER_TC(_NIC_)))
 
 struct aq_hw_rx_fl2 {
        struct aq_rx_filter_vlan aq_vlans[AQ_VLAN_MAX_FILTERS];
@@ -106,7 +115,7 @@ struct aq_nic_s {
        atomic_t flags;
        u32 msg_enable;
        struct aq_vec_s *aq_vec[AQ_CFG_VECS_MAX];
-       struct aq_ring_s *aq_ring_tx[AQ_CFG_VECS_MAX * AQ_CFG_TCS_MAX];
+       struct aq_ring_s *aq_ring_tx[AQ_HW_QUEUES_MAX];
        struct aq_hw_s *aq_hw;
        struct net_device *ndev;
        unsigned int aq_vecs;
@@ -183,4 +192,5 @@ void aq_nic_shutdown(struct aq_nic_s *self);
 u8 aq_nic_reserve_filter(struct aq_nic_s *self, enum aq_rx_filter_type type);
 void aq_nic_release_filter(struct aq_nic_s *self, enum aq_rx_filter_type type,
                           u32 location);
+int aq_nic_setup_tc_mqprio(struct aq_nic_s *self, u32 tcs, u8 *prio_tc_map);
 #endif /* AQ_NIC_H */
index bae95a6..68fdb39 100644 (file)
@@ -232,8 +232,11 @@ void aq_ring_queue_wake(struct aq_ring_s *ring)
 {
        struct net_device *ndev = aq_nic_get_ndev(ring->aq_nic);
 
-       if (__netif_subqueue_stopped(ndev, ring->idx)) {
-               netif_wake_subqueue(ndev, ring->idx);
+       if (__netif_subqueue_stopped(ndev,
+                                    AQ_NIC_RING2QMAP(ring->aq_nic,
+                                                     ring->idx))) {
+               netif_wake_subqueue(ndev,
+                                   AQ_NIC_RING2QMAP(ring->aq_nic, ring->idx));
                ring->stats.tx.queue_restarts++;
        }
 }
@@ -242,8 +245,11 @@ void aq_ring_queue_stop(struct aq_ring_s *ring)
 {
        struct net_device *ndev = aq_nic_get_ndev(ring->aq_nic);
 
-       if (!__netif_subqueue_stopped(ndev, ring->idx))
-               netif_stop_subqueue(ndev, ring->idx);
+       if (!__netif_subqueue_stopped(ndev,
+                                     AQ_NIC_RING2QMAP(ring->aq_nic,
+                                                      ring->idx)))
+               netif_stop_subqueue(ndev,
+                                   AQ_NIC_RING2QMAP(ring->aq_nic, ring->idx));
 }
 
 bool aq_ring_tx_clean(struct aq_ring_s *self)
@@ -466,7 +472,10 @@ int aq_ring_rx_clean(struct aq_ring_s *self,
                             buff->is_hash_l4 ? PKT_HASH_TYPE_L4 :
                             PKT_HASH_TYPE_NONE);
                /* Send all PTP traffic to 0 queue */
-               skb_record_rx_queue(skb, is_ptp_ring ? 0 : self->idx);
+               skb_record_rx_queue(skb,
+                                   is_ptp_ring ? 0
+                                               : AQ_NIC_RING2QMAP(self->aq_nic,
+                                                                  self->idx));
 
                ++self->stats.rx.packets;
                self->stats.rx.bytes += skb->len;
index 7caf586..7753824 100644 (file)
@@ -46,7 +46,8 @@
                        NETIF_F_HW_VLAN_CTAG_RX |     \
                        NETIF_F_HW_VLAN_CTAG_TX |     \
                        NETIF_F_GSO_UDP_L4      |     \
-                       NETIF_F_GSO_PARTIAL,          \
+                       NETIF_F_GSO_PARTIAL |         \
+                       NETIF_F_HW_TC,                \
        .hw_priv_flags = IFF_UNICAST_FLT, \
        .flow_control = true,             \
        .mtu = HW_ATL_B0_MTU_JUMBO,       \
@@ -134,7 +135,7 @@ static int hw_atl_b0_hw_qos_set(struct aq_hw_s *self)
        struct aq_nic_cfg_s *cfg = self->aq_nic_cfg;
        u32 tx_buff_size = HW_ATL_B0_TXBUF_MAX;
        u32 rx_buff_size = HW_ATL_B0_RXBUF_MAX;
-       unsigned int i_priority = 0U;
+       unsigned int prio = 0U;
        u32 tc = 0U;
 
        if (cfg->is_ptp) {
@@ -153,42 +154,45 @@ static int hw_atl_b0_hw_qos_set(struct aq_hw_s *self)
        hw_atl_tps_tx_pkt_shed_desc_tc_arb_mode_set(self, 0U);
        hw_atl_tps_tx_pkt_shed_data_arb_mode_set(self, 0U);
 
-       /* TX Packet Scheduler Data TC0 */
-       hw_atl_tps_tx_pkt_shed_tc_data_max_credit_set(self, 0xFFF, tc);
-       hw_atl_tps_tx_pkt_shed_tc_data_weight_set(self, 0x64, tc);
-       hw_atl_tps_tx_pkt_shed_desc_tc_max_credit_set(self, 0x50, tc);
-       hw_atl_tps_tx_pkt_shed_desc_tc_weight_set(self, 0x1E, tc);
-
-       /* Tx buf size TC0 */
-       hw_atl_tpb_tx_pkt_buff_size_per_tc_set(self, tx_buff_size, tc);
-       hw_atl_tpb_tx_buff_hi_threshold_per_tc_set(self,
-                                                  (tx_buff_size *
-                                                  (1024 / 32U) * 66U) /
-                                                  100U, tc);
-       hw_atl_tpb_tx_buff_lo_threshold_per_tc_set(self,
-                                                  (tx_buff_size *
-                                                  (1024 / 32U) * 50U) /
-                                                  100U, tc);
-
-       /* QoS Rx buf size per TC */
-       hw_atl_rpb_rx_pkt_buff_size_per_tc_set(self, rx_buff_size, tc);
-       hw_atl_rpb_rx_buff_hi_threshold_per_tc_set(self,
-                                                  (rx_buff_size *
-                                                  (1024U / 32U) * 66U) /
-                                                  100U, tc);
-       hw_atl_rpb_rx_buff_lo_threshold_per_tc_set(self,
-                                                  (rx_buff_size *
-                                                  (1024U / 32U) * 50U) /
-                                                  100U, tc);
-
-       hw_atl_b0_set_fc(self, self->aq_nic_cfg->fc.req, tc);
+       tx_buff_size /= cfg->tcs;
+       rx_buff_size /= cfg->tcs;
+       for (tc = 0; tc < cfg->tcs; tc++) {
+               u32 threshold = 0U;
+
+               /* TX Packet Scheduler Data TC0 */
+               hw_atl_tps_tx_pkt_shed_tc_data_max_credit_set(self, 0xFFF, tc);
+               hw_atl_tps_tx_pkt_shed_tc_data_weight_set(self, 0x64, tc);
+               hw_atl_tps_tx_pkt_shed_desc_tc_max_credit_set(self, 0x50, tc);
+               hw_atl_tps_tx_pkt_shed_desc_tc_weight_set(self, 0x1E, tc);
+
+               /* Tx buf size TC0 */
+               hw_atl_tpb_tx_pkt_buff_size_per_tc_set(self, tx_buff_size, tc);
+
+               threshold = (tx_buff_size * (1024 / 32U) * 66U) / 100U;
+               hw_atl_tpb_tx_buff_hi_threshold_per_tc_set(self, threshold, tc);
+
+               threshold = (tx_buff_size * (1024 / 32U) * 50U) / 100U;
+               hw_atl_tpb_tx_buff_lo_threshold_per_tc_set(self, threshold, tc);
+
+               /* QoS Rx buf size per TC */
+               hw_atl_rpb_rx_pkt_buff_size_per_tc_set(self, rx_buff_size, tc);
+
+               threshold = (rx_buff_size * (1024U / 32U) * 66U) / 100U;
+               hw_atl_rpb_rx_buff_hi_threshold_per_tc_set(self, threshold, tc);
+
+               threshold = (rx_buff_size * (1024U / 32U) * 50U) / 100U;
+               hw_atl_rpb_rx_buff_lo_threshold_per_tc_set(self, threshold, tc);
+
+               hw_atl_b0_set_fc(self, self->aq_nic_cfg->fc.req, tc);
+       }
 
        if (cfg->is_ptp)
                hw_atl_b0_tc_ptp_set(self);
 
        /* QoS 802.1p priority -> TC mapping */
-       for (i_priority = 8U; i_priority--;)
-               hw_atl_rpf_rpb_user_priority_tc_map_set(self, i_priority, 0U);
+       for (prio = 0; prio < 8; ++prio)
+               hw_atl_rpf_rpb_user_priority_tc_map_set(self, prio,
+                                                       cfg->prio_tc_map[prio]);
 
        return aq_hw_err_from_flags(self);
 }
@@ -319,7 +323,7 @@ int hw_atl_b0_hw_offload_set(struct aq_hw_s *self,
 static int hw_atl_b0_hw_init_tx_path(struct aq_hw_s *self)
 {
        /* Tx TC/Queue number config */
-       hw_atl_tpb_tps_tx_tc_mode_set(self, 1U);
+       hw_atl_tpb_tps_tx_tc_mode_set(self, self->aq_nic_cfg->tc_mode);
 
        hw_atl_thm_lso_tcp_flag_of_first_pkt_set(self, 0x0FF6U);
        hw_atl_thm_lso_tcp_flag_of_middle_pkt_set(self, 0x0FF6U);
@@ -345,7 +349,7 @@ static int hw_atl_b0_hw_init_rx_path(struct aq_hw_s *self)
        int i;
 
        /* Rx TC/RSS number config */
-       hw_atl_rpb_rpf_rx_traf_class_mode_set(self, 1U);
+       hw_atl_rpb_rpf_rx_traf_class_mode_set(self, cfg->tc_mode);
 
        /* Rx flow control */
        hw_atl_rpb_rx_flow_ctl_mode_set(self, 1U);
index 9e2d01a..8cb6765 100644 (file)
@@ -754,7 +754,7 @@ void hw_atl_rpfl2_accept_all_mc_packets_set(struct aq_hw_s *aq_hw,
 }
 
 void hw_atl_rpf_rpb_user_priority_tc_map_set(struct aq_hw_s *aq_hw,
-                                            u32 user_priority_tc_map, u32 tc)
+                                            u32 user_priority, u32 tc)
 {
 /* register address for bitfield rx_tc_up{t}[2:0] */
        static u32 rpf_rpb_rx_tc_upt_adr[8] = {
@@ -773,10 +773,9 @@ void hw_atl_rpf_rpb_user_priority_tc_map_set(struct aq_hw_s *aq_hw,
                        0U, 4U, 8U, 12U, 16U, 20U, 24U, 28U
                };
 
-       aq_hw_write_reg_bit(aq_hw, rpf_rpb_rx_tc_upt_adr[tc],
-                           rpf_rpb_rx_tc_upt_msk[tc],
-                           rpf_rpb_rx_tc_upt_shft[tc],
-                           user_priority_tc_map);
+       aq_hw_write_reg_bit(aq_hw, rpf_rpb_rx_tc_upt_adr[user_priority],
+                           rpf_rpb_rx_tc_upt_msk[user_priority],
+                           rpf_rpb_rx_tc_upt_shft[user_priority], tc);
 }
 
 void hw_atl_rpf_rss_key_addr_set(struct aq_hw_s *aq_hw, u32 rss_key_addr)
index a141185..05c0496 100644 (file)
@@ -47,7 +47,8 @@ static int hw_atl2_act_rslvr_table_set(struct aq_hw_s *self, u8 location,
                        NETIF_F_HW_VLAN_CTAG_RX |     \
                        NETIF_F_HW_VLAN_CTAG_TX |     \
                        NETIF_F_GSO_UDP_L4      |     \
-                       NETIF_F_GSO_PARTIAL,          \
+                       NETIF_F_GSO_PARTIAL     |     \
+                       NETIF_F_HW_TC,                \
        .hw_priv_flags = IFF_UNICAST_FLT, \
        .flow_control = true,             \
        .mtu = HW_ATL2_MTU_JUMBO,         \
@@ -132,7 +133,6 @@ static int hw_atl2_hw_qos_set(struct aq_hw_s *self)
        u32 tx_buff_size = HW_ATL2_TXBUF_MAX;
        u32 rx_buff_size = HW_ATL2_RXBUF_MAX;
        unsigned int prio = 0U;
-       u32 threshold = 0U;
        u32 tc = 0U;
 
        /* TPS Descriptor rate init */
@@ -146,34 +146,41 @@ static int hw_atl2_hw_qos_set(struct aq_hw_s *self)
        hw_atl_tps_tx_pkt_shed_desc_tc_arb_mode_set(self, 0U);
        hw_atl_tps_tx_pkt_shed_data_arb_mode_set(self, 0U);
 
-       /* TX Packet Scheduler Data TC0 */
-       hw_atl2_tps_tx_pkt_shed_tc_data_max_credit_set(self, 0xFFF0, tc);
-       hw_atl2_tps_tx_pkt_shed_tc_data_weight_set(self, 0x640, tc);
-       hw_atl_tps_tx_pkt_shed_desc_tc_max_credit_set(self, 0x50, tc);
-       hw_atl_tps_tx_pkt_shed_desc_tc_weight_set(self, 0x1E, tc);
+       tx_buff_size /= cfg->tcs;
+       rx_buff_size /= cfg->tcs;
+       for (tc = 0; tc < cfg->tcs; tc++) {
+               u32 threshold = 0U;
 
-       /* Tx buf size TC0 */
-       hw_atl_tpb_tx_pkt_buff_size_per_tc_set(self, tx_buff_size, tc);
+               /* TX Packet Scheduler Data TC0 */
+               hw_atl2_tps_tx_pkt_shed_tc_data_max_credit_set(self, 0xFFF0,
+                                                              tc);
+               hw_atl2_tps_tx_pkt_shed_tc_data_weight_set(self, 0x640, tc);
+               hw_atl_tps_tx_pkt_shed_desc_tc_max_credit_set(self, 0x50, tc);
+               hw_atl_tps_tx_pkt_shed_desc_tc_weight_set(self, 0x1E, tc);
 
-       threshold = (tx_buff_size * (1024 / 32U) * 66U) / 100U;
-       hw_atl_tpb_tx_buff_hi_threshold_per_tc_set(self, threshold, tc);
+               /* Tx buf size TC0 */
+               hw_atl_tpb_tx_pkt_buff_size_per_tc_set(self, tx_buff_size, tc);
 
-       threshold = (tx_buff_size * (1024 / 32U) * 50U) / 100U;
-       hw_atl_tpb_tx_buff_lo_threshold_per_tc_set(self, threshold, tc);
+               threshold = (tx_buff_size * (1024 / 32U) * 66U) / 100U;
+               hw_atl_tpb_tx_buff_hi_threshold_per_tc_set(self, threshold, tc);
 
-       /* QoS Rx buf size per TC */
-       hw_atl_rpb_rx_pkt_buff_size_per_tc_set(self, rx_buff_size, tc);
+               threshold = (tx_buff_size * (1024 / 32U) * 50U) / 100U;
+               hw_atl_tpb_tx_buff_lo_threshold_per_tc_set(self, threshold, tc);
 
-       threshold = (rx_buff_size * (1024U / 32U) * 66U) / 100U;
-       hw_atl_rpb_rx_buff_hi_threshold_per_tc_set(self, threshold, tc);
+               /* QoS Rx buf size per TC */
+               hw_atl_rpb_rx_pkt_buff_size_per_tc_set(self, rx_buff_size, tc);
 
-       threshold = (rx_buff_size * (1024U / 32U) * 50U) / 100U;
-       hw_atl_rpb_rx_buff_lo_threshold_per_tc_set(self, threshold, tc);
+               threshold = (rx_buff_size * (1024U / 32U) * 66U) / 100U;
+               hw_atl_rpb_rx_buff_hi_threshold_per_tc_set(self, threshold, tc);
+
+               threshold = (rx_buff_size * (1024U / 32U) * 50U) / 100U;
+               hw_atl_rpb_rx_buff_lo_threshold_per_tc_set(self, threshold, tc);
+       }
 
        /* QoS 802.1p priority -> TC mapping */
        for (prio = 0; prio < 8; ++prio)
                hw_atl_rpf_rpb_user_priority_tc_map_set(self, prio,
-                                                       cfg->tcs * prio / 8);
+                                                       cfg->prio_tc_map[prio]);
 
        /* ATL2 Apply legacy ring to TC mapping */
        hw_atl2_hw_queue_to_tc_map_set(self);
@@ -184,11 +191,24 @@ static int hw_atl2_hw_qos_set(struct aq_hw_s *self)
 static int hw_atl2_hw_rss_set(struct aq_hw_s *self,
                              struct aq_rss_parameters *rss_params)
 {
-       u8 *indirection_table = rss_params->indirection_table;
+       u8 *indirection_table = rss_params->indirection_table;
+       const u32 num_tcs = aq_hw_num_tcs(self);
+       u32 rpf_redir2_enable;
+       int tc;
        int i;
 
-       for (i = HW_ATL2_RSS_REDIRECTION_MAX; i--;)
-               hw_atl2_new_rpf_rss_redir_set(self, 0, i, indirection_table[i]);
+       rpf_redir2_enable = num_tcs > 4 ? 1 : 0;
+
+       hw_atl2_rpf_redirection_table2_select_set(self, rpf_redir2_enable);
+
+       for (i = HW_ATL2_RSS_REDIRECTION_MAX; i--;) {
+               for (tc = 0; tc != num_tcs; tc++) {
+                       hw_atl2_new_rpf_rss_redir_set(self, tc, i,
+                                                     tc *
+                                                     aq_hw_q_per_tc(self) +
+                                                     indirection_table[i]);
+               }
+       }
 
        return aq_hw_err_from_flags(self);
 }
@@ -196,7 +216,7 @@ static int hw_atl2_hw_rss_set(struct aq_hw_s *self,
 static int hw_atl2_hw_init_tx_path(struct aq_hw_s *self)
 {
        /* Tx TC/RSS number config */
-       hw_atl_tpb_tps_tx_tc_mode_set(self, 1U);
+       hw_atl_tpb_tps_tx_tc_mode_set(self, self->aq_nic_cfg->tc_mode);
 
        hw_atl_thm_lso_tcp_flag_of_first_pkt_set(self, 0x0FF6U);
        hw_atl_thm_lso_tcp_flag_of_middle_pkt_set(self, 0x0FF6U);
@@ -219,13 +239,29 @@ static int hw_atl2_hw_init_tx_path(struct aq_hw_s *self)
 static void hw_atl2_hw_init_new_rx_filters(struct aq_hw_s *self)
 {
        struct hw_atl2_priv *priv = (struct hw_atl2_priv *)self->priv;
+       u8 *prio_tc_map = self->aq_nic_cfg->prio_tc_map;
+       u16 action;
        u8 index;
+       int i;
 
+       /* Action Resolver Table (ART) is used by RPF to decide which action
+        * to take with a packet based upon input tag and tag mask, where:
+        *  - input tag is a combination of 3-bit VLan Prio (PTP) and
+        *    29-bit concatenation of all tags from filter block;
+        *  - tag mask is a mask used for matching against input tag.
+        * The input_tag is compared with the all the Requested_tags in the
+        * Record table to find a match. Action field of the selected matched
+        * REC entry is used for further processing. If multiple entries match,
+        * the lowest REC entry, Action field will be selected.
+        */
        hw_atl2_rpf_act_rslvr_section_en_set(self, 0xFFFF);
        hw_atl2_rpfl2_uc_flr_tag_set(self, HW_ATL2_RPF_TAG_BASE_UC,
                                     HW_ATL2_MAC_UC);
        hw_atl2_rpfl2_bc_flr_tag_set(self, HW_ATL2_RPF_TAG_BASE_UC);
 
+       /* FW reserves the beginning of ART, thus all driver entries must
+        * start from the offset specified in FW caps.
+        */
        index = priv->art_base_index + HW_ATL2_RPF_L2_PROMISC_OFF_INDEX;
        hw_atl2_act_rslvr_table_set(self, index, 0,
                                    HW_ATL2_RPF_TAG_UC_MASK |
@@ -238,33 +274,17 @@ static void hw_atl2_hw_init_new_rx_filters(struct aq_hw_s *self)
                                        HW_ATL2_RPF_TAG_UNTAG_MASK,
                                    HW_ATL2_ACTION_DROP);
 
-       index = priv->art_base_index + HW_ATL2_RPF_VLAN_INDEX;
-       hw_atl2_act_rslvr_table_set(self, index, HW_ATL2_RPF_TAG_BASE_VLAN,
-                                   HW_ATL2_RPF_TAG_VLAN_MASK,
-                                   HW_ATL2_ACTION_ASSIGN_TC(0));
+       /* Configure ART to map given VLan Prio (PCP) to the TC index for
+        * RSS redirection table.
+        */
+       for (i = 0; i < 8; i++) {
+               action = HW_ATL2_ACTION_ASSIGN_TC(prio_tc_map[i]);
 
-       index = priv->art_base_index + HW_ATL2_RPF_MAC_INDEX;
-       hw_atl2_act_rslvr_table_set(self, index, HW_ATL2_RPF_TAG_BASE_UC,
-                                   HW_ATL2_RPF_TAG_UC_MASK,
-                                   HW_ATL2_ACTION_ASSIGN_TC(0));
-
-       index = priv->art_base_index + HW_ATL2_RPF_ALLMC_INDEX;
-       hw_atl2_act_rslvr_table_set(self, index, HW_ATL2_RPF_TAG_BASE_ALLMC,
-                                   HW_ATL2_RPF_TAG_ALLMC_MASK,
-                                   HW_ATL2_ACTION_ASSIGN_TC(0));
-
-       index = priv->art_base_index + HW_ATL2_RPF_UNTAG_INDEX;
-       hw_atl2_act_rslvr_table_set(self, index, HW_ATL2_RPF_TAG_UNTAG_MASK,
-                                   HW_ATL2_RPF_TAG_UNTAG_MASK,
-                                   HW_ATL2_ACTION_ASSIGN_TC(0));
-
-       index = priv->art_base_index + HW_ATL2_RPF_VLAN_PROMISC_ON_INDEX;
-       hw_atl2_act_rslvr_table_set(self, index, 0, HW_ATL2_RPF_TAG_VLAN_MASK,
-                                   HW_ATL2_ACTION_DISABLE);
-
-       index = priv->art_base_index + HW_ATL2_RPF_L2_PROMISC_ON_INDEX;
-       hw_atl2_act_rslvr_table_set(self, index, 0, HW_ATL2_RPF_TAG_UC_MASK,
-                                   HW_ATL2_ACTION_DISABLE);
+               index = priv->art_base_index + HW_ATL2_RPF_PCP_TO_TC_INDEX + i;
+               hw_atl2_act_rslvr_table_set(self, index,
+                                           i << HW_ATL2_RPF_TAG_PCP_OFFSET,
+                                           HW_ATL2_RPF_TAG_PCP_MASK, action);
+       }
 }
 
 static void hw_atl2_hw_new_rx_filter_vlan_promisc(struct aq_hw_s *self,
@@ -327,7 +347,7 @@ static int hw_atl2_hw_init_rx_path(struct aq_hw_s *self)
        int i;
 
        /* Rx TC/RSS number config */
-       hw_atl_rpb_rpf_rx_traf_class_mode_set(self, 1U);
+       hw_atl_rpb_rpf_rx_traf_class_mode_set(self, cfg->tc_mode);
 
        /* Rx flow control */
        hw_atl_rpb_rx_flow_ctl_mode_set(self, 1U);
index be0c049..9ac1979 100644 (file)
@@ -82,13 +82,6 @@ enum HW_ATL2_RPF_ART_INDEX {
        HW_ATL2_RPF_VLAN_USER_INDEX     = HW_ATL2_RPF_ET_PCP_USER_INDEX + 16,
        HW_ATL2_RPF_PCP_TO_TC_INDEX     = HW_ATL2_RPF_VLAN_USER_INDEX +
                                          HW_ATL_VLAN_MAX_FILTERS,
-       HW_ATL2_RPF_VLAN_INDEX          = HW_ATL2_RPF_PCP_TO_TC_INDEX +
-                                         AQ_CFG_TCS_MAX,
-       HW_ATL2_RPF_MAC_INDEX,
-       HW_ATL2_RPF_ALLMC_INDEX,
-       HW_ATL2_RPF_UNTAG_INDEX,
-       HW_ATL2_RPF_VLAN_PROMISC_ON_INDEX,
-       HW_ATL2_RPF_L2_PROMISC_ON_INDEX,
 };
 
 #define HW_ATL2_ACTION(ACTION, RSS, INDEX, VALID) \
index e779d70..f096d0a 100644 (file)
@@ -7,6 +7,14 @@
 #include "hw_atl2_llh_internal.h"
 #include "aq_hw_utils.h"
 
+void hw_atl2_rpf_redirection_table2_select_set(struct aq_hw_s *aq_hw,
+                                              u32 select)
+{
+       aq_hw_write_reg_bit(aq_hw, HW_ATL2_RPF_PIF_RPF_REDIR2_ENI_ADR,
+                           HW_ATL2_RPF_PIF_RPF_REDIR2_ENI_MSK,
+                           HW_ATL2_RPF_PIF_RPF_REDIR2_ENI_SHIFT, select);
+}
+
 void hw_atl2_rpf_rss_hash_type_set(struct aq_hw_s *aq_hw, u32 rss_hash_type)
 {
        aq_hw_write_reg_bit(aq_hw, HW_ATL2_RPF_PIF_RPF_RSS_HASH_TYPEI_ADR,
index 8c6d78a..5c1ae75 100644 (file)
@@ -15,6 +15,10 @@ void hw_atl2_reg_tx_intr_moder_ctrl_set(struct aq_hw_s *aq_hw,
                                        u32 tx_intr_moderation_ctl,
                                        u32 queue);
 
+/* Set Redirection Table 2 Select */
+void hw_atl2_rpf_redirection_table2_select_set(struct aq_hw_s *aq_hw,
+                                              u32 select);
+
 /** Set RSS HASH type */
 void hw_atl2_rpf_rss_hash_type_set(struct aq_hw_s *aq_hw, u32 rss_hash_type);
 
index cde9e9d..b0ac8cd 100644 (file)
@@ -6,6 +6,16 @@
 #ifndef HW_ATL2_LLH_INTERNAL_H
 #define HW_ATL2_LLH_INTERNAL_H
 
+/* RX pif_rpf_redir_2_en_i Bitfield Definitions
+ * PORT="pif_rpf_redir_2_en_i"
+ */
+#define HW_ATL2_RPF_PIF_RPF_REDIR2_ENI_ADR 0x000054C8
+#define HW_ATL2_RPF_PIF_RPF_REDIR2_ENI_MSK 0x00001000
+#define HW_ATL2_RPF_PIF_RPF_REDIR2_ENI_MSKN 0xFFFFEFFF
+#define HW_ATL2_RPF_PIF_RPF_REDIR2_ENI_SHIFT 12
+#define HW_ATL2_RPF_PIF_RPF_REDIR2_ENI_WIDTH 1
+#define HW_ATL2_RPF_PIF_RPF_REDIR2_ENI_DEFAULT 0x0
+
 /* RX pif_rpf_rss_hash_type_i Bitfield Definitions
  */
 #define HW_ATL2_RPF_PIF_RPF_RSS_HASH_TYPEI_ADR 0x000054C8