Replace the byte based tail queue congestion support
with frame count based congestion groups.
It can easily map to number of RX descriptors for a queue.

Signed-off-by: Hemant Agrawal <hemant.agra...@nxp.com>
---
 drivers/bus/dpaa/rte_bus_dpaa_version.map |  5 ++
 drivers/net/dpaa/dpaa_ethdev.c            | 98 +++++++++++++++++++++++++++----
 drivers/net/dpaa/dpaa_ethdev.h            |  8 +--
 3 files changed, 97 insertions(+), 14 deletions(-)

diff --git a/drivers/bus/dpaa/rte_bus_dpaa_version.map 
b/drivers/bus/dpaa/rte_bus_dpaa_version.map
index eeeb458..f412362 100644
--- a/drivers/bus/dpaa/rte_bus_dpaa_version.map
+++ b/drivers/bus/dpaa/rte_bus_dpaa_version.map
@@ -69,6 +69,11 @@ DPDK_18.02 {
        global:
 
        dpaa_svr_family;
+       qman_alloc_cgrid_range;
+       qman_create_cgr;
+       qman_delete_cgr;
+       qman_modify_cgr;
+       qman_release_cgrid_range;
 
        local: *;
 } DPDK_17.11;
diff --git a/drivers/net/dpaa/dpaa_ethdev.c b/drivers/net/dpaa/dpaa_ethdev.c
index adcc219..6482998 100644
--- a/drivers/net/dpaa/dpaa_ethdev.c
+++ b/drivers/net/dpaa/dpaa_ethdev.c
@@ -73,6 +73,9 @@
 /* Keep track of whether QMAN and BMAN have been globally initialized */
 static int is_global_init;
 
+/* Per FQ Taildrop in frame count */
+static unsigned int td_threshold = CGR_RX_PERFQ_THRESH;
+
 struct rte_dpaa_xstats_name_off {
        char name[RTE_ETH_XSTATS_NAME_SIZE];
        uint32_t offset;
@@ -447,12 +450,13 @@ static void dpaa_eth_multicast_disable(struct rte_eth_dev 
*dev)
 
 static
 int dpaa_eth_rx_queue_setup(struct rte_eth_dev *dev, uint16_t queue_idx,
-                           uint16_t nb_desc __rte_unused,
+                           uint16_t nb_desc,
                            unsigned int socket_id __rte_unused,
                            const struct rte_eth_rxconf *rx_conf __rte_unused,
                            struct rte_mempool *mp)
 {
        struct dpaa_if *dpaa_intf = dev->data->dev_private;
+       struct qman_fq *rxq = &dpaa_intf->rx_queues[queue_idx];
 
        PMD_INIT_FUNC_TRACE();
 
@@ -488,7 +492,23 @@ int dpaa_eth_rx_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
                            dpaa_intf->name, fd_offset,
                        fman_if_get_fdoff(dpaa_intf->fif));
        }
-       dev->data->rx_queues[queue_idx] = &dpaa_intf->rx_queues[queue_idx];
+
+       dev->data->rx_queues[queue_idx] = rxq;
+
+       /* configure the CGR size as per the desc size */
+       if (dpaa_intf->cgr_rx) {
+               struct qm_mcc_initcgr cgr_opts = {0};
+               int ret;
+
+               /* Enable tail drop with cgr on this queue */
+               qm_cgr_cs_thres_set64(&cgr_opts.cgr.cs_thres, nb_desc, 0);
+               ret = qman_modify_cgr(dpaa_intf->cgr_rx, 0, &cgr_opts);
+               if (ret) {
+                       DPAA_PMD_WARN(
+                               "rx taildrop modify fail on fqid %d (ret=%d)",
+                               rxq->fqid, ret);
+               }
+       }
 
        return 0;
 }
@@ -724,11 +744,21 @@ static int dpaa_fc_set_default(struct dpaa_if *dpaa_intf)
 }
 
 /* Initialise an Rx FQ */
-static int dpaa_rx_queue_init(struct qman_fq *fq,
+static int dpaa_rx_queue_init(struct qman_fq *fq, struct qman_cgr *cgr_rx,
                              uint32_t fqid)
 {
        struct qm_mcc_initfq opts = {0};
        int ret;
+       u32 flags = 0;
+       struct qm_mcc_initcgr cgr_opts = {
+               .we_mask = QM_CGR_WE_CS_THRES |
+                               QM_CGR_WE_CSTD_EN |
+                               QM_CGR_WE_MODE,
+               .cgr = {
+                       .cstd_en = QM_CGR_EN,
+                       .mode = QMAN_CGR_MODE_FRAME
+               }
+       };
 
        PMD_INIT_FUNC_TRACE();
 
@@ -758,12 +788,24 @@ static int dpaa_rx_queue_init(struct qman_fq *fq,
        opts.fqd.context_a.stashing.data_cl = DPAA_IF_RX_DATA_STASH;
        opts.fqd.context_a.stashing.context_cl = DPAA_IF_RX_CONTEXT_STASH;
 
-       /*Enable tail drop */
-       opts.we_mask = opts.we_mask | QM_INITFQ_WE_TDTHRESH;
-       opts.fqd.fq_ctrl = opts.fqd.fq_ctrl | QM_FQCTRL_TDE;
-       qm_fqd_taildrop_set(&opts.fqd.td, CONG_THRESHOLD_RX_Q, 1);
-
-       ret = qman_init_fq(fq, 0, &opts);
+       if (cgr_rx) {
+               /* Enable tail drop with cgr on this queue */
+               qm_cgr_cs_thres_set64(&cgr_opts.cgr.cs_thres, td_threshold, 0);
+               cgr_rx->cb = NULL;
+               ret = qman_create_cgr(cgr_rx, QMAN_CGR_FLAG_USE_INIT,
+                                     &cgr_opts);
+               if (ret) {
+                       DPAA_PMD_WARN(
+                               "rx taildrop init fail on rx fqid %d (ret=%d)",
+                               fqid, ret);
+                       goto without_cgr;
+               }
+               opts.we_mask |= QM_INITFQ_WE_CGID;
+               opts.fqd.cgid = cgr_rx->cgrid;
+               opts.fqd.fq_ctrl |= QM_FQCTRL_CGE;
+       }
+without_cgr:
+       ret = qman_init_fq(fq, flags, &opts);
        if (ret)
                DPAA_PMD_ERR("init rx fqid %d failed with ret: %d", fqid, ret);
        return ret;
@@ -845,6 +887,7 @@ dpaa_dev_init(struct rte_eth_dev *eth_dev)
        struct fm_eth_port_cfg *cfg;
        struct fman_if *fman_intf;
        struct fman_if_bpool *bp, *tmp_bp;
+       uint32_t cgrid[DPAA_MAX_NUM_PCD_QUEUES];
 
        PMD_INIT_FUNC_TRACE();
 
@@ -881,10 +924,31 @@ dpaa_dev_init(struct rte_eth_dev *eth_dev)
 
        dpaa_intf->rx_queues = rte_zmalloc(NULL,
                sizeof(struct qman_fq) * num_rx_fqs, MAX_CACHELINE);
+
+       /* If congestion control is enabled globally*/
+       if (td_threshold) {
+               dpaa_intf->cgr_rx = rte_zmalloc(NULL,
+                       sizeof(struct qman_cgr) * num_rx_fqs, MAX_CACHELINE);
+
+               ret = qman_alloc_cgrid_range(&cgrid[0], num_rx_fqs, 1, 0);
+               if (ret != num_rx_fqs) {
+                       DPAA_PMD_WARN("insufficient CGRIDs available");
+                       return -EINVAL;
+               }
+       } else {
+               dpaa_intf->cgr_rx = NULL;
+       }
+
        for (loop = 0; loop < num_rx_fqs; loop++) {
                fqid = DPAA_PCD_FQID_START + dpaa_intf->ifid *
                        DPAA_PCD_FQID_MULTIPLIER + loop;
-               ret = dpaa_rx_queue_init(&dpaa_intf->rx_queues[loop], fqid);
+
+               if (dpaa_intf->cgr_rx)
+                       dpaa_intf->cgr_rx[loop].cgrid = cgrid[loop];
+
+               ret = dpaa_rx_queue_init(&dpaa_intf->rx_queues[loop],
+                       dpaa_intf->cgr_rx ? &dpaa_intf->cgr_rx[loop] : NULL,
+                       fqid);
                if (ret)
                        return ret;
                dpaa_intf->rx_queues[loop].dpaa_intf = dpaa_intf;
@@ -939,6 +1003,7 @@ dpaa_dev_init(struct rte_eth_dev *eth_dev)
                DPAA_PMD_ERR("Failed to allocate %d bytes needed to "
                                                "store MAC addresses",
                                ETHER_ADDR_LEN * DPAA_MAX_MAC_FILTER);
+               rte_free(dpaa_intf->cgr_rx);
                rte_free(dpaa_intf->rx_queues);
                rte_free(dpaa_intf->tx_queues);
                dpaa_intf->rx_queues = NULL;
@@ -977,6 +1042,7 @@ static int
 dpaa_dev_uninit(struct rte_eth_dev *dev)
 {
        struct dpaa_if *dpaa_intf = dev->data->dev_private;
+       int loop;
 
        PMD_INIT_FUNC_TRACE();
 
@@ -994,6 +1060,18 @@ dpaa_dev_uninit(struct rte_eth_dev *dev)
        if (dpaa_intf->fc_conf)
                rte_free(dpaa_intf->fc_conf);
 
+       /* Release RX congestion Groups */
+       if (dpaa_intf->cgr_rx) {
+               for (loop = 0; loop < dpaa_intf->nb_rx_queues; loop++)
+                       qman_delete_cgr(&dpaa_intf->cgr_rx[loop]);
+
+               qman_release_cgrid_range(dpaa_intf->cgr_rx[loop].cgrid,
+                                        dpaa_intf->nb_rx_queues);
+       }
+
+       rte_free(dpaa_intf->cgr_rx);
+       dpaa_intf->cgr_rx = NULL;
+
        rte_free(dpaa_intf->rx_queues);
        dpaa_intf->rx_queues = NULL;
 
diff --git a/drivers/net/dpaa/dpaa_ethdev.h b/drivers/net/dpaa/dpaa_ethdev.h
index ef726d3..b26e411 100644
--- a/drivers/net/dpaa/dpaa_ethdev.h
+++ b/drivers/net/dpaa/dpaa_ethdev.h
@@ -60,10 +60,8 @@
 #define DPAA_MIN_RX_BUF_SIZE 512
 #define DPAA_MAX_RX_PKT_LEN  10240
 
-/* RX queue tail drop threshold
- * currently considering 32 KB packets.
- */
-#define CONG_THRESHOLD_RX_Q  (32 * 1024)
+/* RX queue tail drop threshold (CGR Based) in frame count */
+#define CGR_RX_PERFQ_THRESH 256
 
 /*max mac filter for memac(8) including primary mac addr*/
 #define DPAA_MAX_MAC_FILTER (MEMAC_NUM_OF_PADDRS + 1)
@@ -79,6 +77,7 @@
 #define DPAA_PCD_FQID_START            0x400
 #define DPAA_PCD_FQID_MULTIPLIER       0x100
 #define DPAA_DEFAULT_NUM_PCD_QUEUES    1
+#define DPAA_MAX_NUM_PCD_QUEUES                32
 
 #define DPAA_IF_TX_PRIORITY            3
 #define DPAA_IF_RX_PRIORITY            4
@@ -128,6 +127,7 @@ struct dpaa_if {
        char *name;
        const struct fm_eth_port_cfg *cfg;
        struct qman_fq *rx_queues;
+       struct qman_cgr *cgr_rx;
        struct qman_fq *tx_queues;
        struct qman_fq debug_queues[2];
        uint16_t nb_rx_queues;
-- 
2.7.4

Reply via email to