net: hns3: change SSU's buffer allocation according to UM
authorYunsheng Lin <linyunsheng@huawei.com>
Fri, 28 Jun 2019 11:50:09 +0000 (19:50 +0800)
committerDavid S. Miller <davem@davemloft.net>
Fri, 28 Jun 2019 16:39:49 +0000 (09:39 -0700)
Currently when there is share buffer in the SSU(storage
switching unit), the low waterline for RX private buffer is
too low to keep the hardware running. Hardware may have
processed all the packet stored in the private buffer of the
low waterline before the new packet comes, because hardware
only tell the peer send packet again when the private buffer
is under the low waterline.

So this patch only allocate RX private buffer if there is
enough buffer according to hardware user manual.

This patch also reserve some buffer for reusing when TC num
is less than or equal to 2, and change PAUSE_TRANS_GAP &
HCLGE_NON_DCB_ADDITIONAL_BUF according to hardware user
manual.

Signed-off-by: Yunsheng Lin <linyunsheng@huawei.com>
Signed-off-by: Peng Li <lipeng321@huawei.com>
Signed-off-by: Huazhong Tan <tanhuazhong@huawei.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_cmd.h
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_tm.h

index cf52cdf13270a17f7593576b57d5d85498b47ba7..d23ab2b819d950d64eb69cbc82ca1adf82181747 100644 (file)
@@ -884,7 +884,7 @@ struct hclge_serdes_lb_cmd {
 #define HCLGE_TOTAL_PKT_BUF            0x108000 /* 1.03125M bytes */
 #define HCLGE_DEFAULT_DV               0xA000   /* 40k byte */
 #define HCLGE_DEFAULT_NON_DCB_DV       0x7800  /* 30K byte */
-#define HCLGE_NON_DCB_ADDITIONAL_BUF   0x200   /* 512 byte */
+#define HCLGE_NON_DCB_ADDITIONAL_BUF   0x1400  /* 5120 byte */
 
 #define HCLGE_TYPE_CRQ                 0
 #define HCLGE_TYPE_CSQ                 1
index b25365c343d10832d3a99c7d6603a31b1c9cdfbb..a2401c34a89d5a7e0d6b368bac7e8e06b3576d95 100644 (file)
@@ -30,6 +30,9 @@
 #define HCLGE_BUF_SIZE_UNIT    256U
 #define HCLGE_BUF_MUL_BY       2
 #define HCLGE_BUF_DIV_BY       2
+#define NEED_RESERVE_TC_NUM    2
+#define BUF_MAX_PERCENT                100
+#define BUF_RESERVE_PERCENT    90
 
 #define HCLGE_RESET_MAX_FAIL_CNT       5
 
@@ -1694,10 +1697,14 @@ static bool  hclge_is_rx_buf_ok(struct hclge_dev *hdev,
        }
 
        if (hnae3_dev_dcb_supported(hdev)) {
+               hi_thrd = shared_buf - hdev->dv_buf_size;
+
+               if (tc_num <= NEED_RESERVE_TC_NUM)
+                       hi_thrd = hi_thrd * BUF_RESERVE_PERCENT
+                                       / BUF_MAX_PERCENT;
+
                if (tc_num)
-                       hi_thrd = (shared_buf - hdev->dv_buf_size) / tc_num;
-               else
-                       hi_thrd = shared_buf - hdev->dv_buf_size;
+                       hi_thrd = hi_thrd / tc_num;
 
                hi_thrd = max_t(u32, hi_thrd, HCLGE_BUF_MUL_BY * aligned_mps);
                hi_thrd = rounddown(hi_thrd, HCLGE_BUF_SIZE_UNIT);
@@ -1837,6 +1844,55 @@ static bool hclge_drop_pfc_buf_till_fit(struct hclge_dev *hdev,
        return hclge_is_rx_buf_ok(hdev, buf_alloc, rx_all);
 }
 
+static int hclge_only_alloc_priv_buff(struct hclge_dev *hdev,
+                                     struct hclge_pkt_buf_alloc *buf_alloc)
+{
+#define COMPENSATE_BUFFER      0x3C00
+#define COMPENSATE_HALF_MPS_NUM        5
+#define PRIV_WL_GAP            0x1800
+
+       u32 rx_priv = hdev->pkt_buf_size - hclge_get_tx_buff_alloced(buf_alloc);
+       u32 tc_num = hclge_get_tc_num(hdev);
+       u32 half_mps = hdev->mps >> 1;
+       u32 min_rx_priv;
+       unsigned int i;
+
+       if (tc_num)
+               rx_priv = rx_priv / tc_num;
+
+       if (tc_num <= NEED_RESERVE_TC_NUM)
+               rx_priv = rx_priv * BUF_RESERVE_PERCENT / BUF_MAX_PERCENT;
+
+       min_rx_priv = hdev->dv_buf_size + COMPENSATE_BUFFER +
+                       COMPENSATE_HALF_MPS_NUM * half_mps;
+       min_rx_priv = round_up(min_rx_priv, HCLGE_BUF_SIZE_UNIT);
+       rx_priv = round_down(rx_priv, HCLGE_BUF_SIZE_UNIT);
+
+       if (rx_priv < min_rx_priv)
+               return false;
+
+       for (i = 0; i < HCLGE_MAX_TC_NUM; i++) {
+               struct hclge_priv_buf *priv = &buf_alloc->priv_buf[i];
+
+               priv->enable = 0;
+               priv->wl.low = 0;
+               priv->wl.high = 0;
+               priv->buf_size = 0;
+
+               if (!(hdev->hw_tc_map & BIT(i)))
+                       continue;
+
+               priv->enable = 1;
+               priv->buf_size = rx_priv;
+               priv->wl.high = rx_priv - hdev->dv_buf_size;
+               priv->wl.low = priv->wl.high - PRIV_WL_GAP;
+       }
+
+       buf_alloc->s_buf.buf_size = 0;
+
+       return true;
+}
+
 /* hclge_rx_buffer_calc: calculate the rx private buffer size for all TCs
  * @hdev: pointer to struct hclge_dev
  * @buf_alloc: pointer to buffer calculation data
@@ -1856,6 +1912,9 @@ static int hclge_rx_buffer_calc(struct hclge_dev *hdev,
                return 0;
        }
 
+       if (hclge_only_alloc_priv_buff(hdev, buf_alloc))
+               return 0;
+
        if (hclge_rx_buf_calc_all(hdev, true, buf_alloc))
                return 0;
 
index 5150daaac3454601d2cf7375d508e270b284b1ca..818610988d342fddaaf1474ccaf7f985536f828c 100644 (file)
@@ -12,7 +12,7 @@
 
 #define HCLGE_TM_PORT_BASE_MODE_MSK    BIT(0)
 
-#define HCLGE_DEFAULT_PAUSE_TRANS_GAP  0xFF
+#define HCLGE_DEFAULT_PAUSE_TRANS_GAP  0x7F
 #define HCLGE_DEFAULT_PAUSE_TRANS_TIME 0xFFFF
 
 /* SP or DWRR */