Skip to content

Commit

Permalink
net/hns3: fix vector Rx burst limitation
Browse files Browse the repository at this point in the history
[ upstream commit 2d408d0 ]

Currently, driver uses the macro HNS3_DEFAULT_RX_BURST whose value is
32 to limit the vector Rx burst size, as a result, the burst size
can't exceed 32.

This patch fixes this problem by support big burst size.
Also adjust HNS3_DEFAULT_RX_BURST to 64 as it performs better than 32.

Fixes: a3d4f4d ("net/hns3: support NEON Rx")
Fixes: 952ebac ("net/hns3: support SVE Rx")

Signed-off-by: Chengwen Feng <fengchengwen@huawei.com>
Signed-off-by: Min Hu (Connor) <humin29@huawei.com>
  • Loading branch information
fengchengwen authored and steevenlee committed Jun 8, 2021
1 parent 88be45d commit 29da2a8
Show file tree
Hide file tree
Showing 4 changed files with 58 additions and 15 deletions.
2 changes: 1 addition & 1 deletion drivers/net/hns3/hns3_rxtx.h
Expand Up @@ -20,7 +20,7 @@
#define HNS3_DEFAULT_TX_RS_THRESH 32
#define HNS3_TX_FAST_FREE_AHEAD 64

#define HNS3_DEFAULT_RX_BURST 32
#define HNS3_DEFAULT_RX_BURST 64
#if (HNS3_DEFAULT_RX_BURST > 64)
#error "PMD HNS3: HNS3_DEFAULT_RX_BURST must <= 64\n"
#endif
Expand Down
36 changes: 28 additions & 8 deletions drivers/net/hns3/hns3_rxtx_vec.c
Expand Up @@ -104,14 +104,13 @@ hns3_recv_pkts_vec(void *__restrict rx_queue,
{
struct hns3_rx_queue *rxq = rx_queue;
struct hns3_desc *rxdp = &rxq->rx_ring[rxq->next_to_use];
uint64_t bd_err_mask; /* bit mask indicate whick pkts is error */
uint64_t pkt_err_mask; /* bit mask indicate whick pkts is error */
uint16_t nb_rx;

nb_pkts = RTE_MIN(nb_pkts, HNS3_DEFAULT_RX_BURST);
nb_pkts = RTE_ALIGN_FLOOR(nb_pkts, HNS3_DEFAULT_DESCS_PER_LOOP);

rte_prefetch_non_temporal(rxdp);

nb_pkts = RTE_ALIGN_FLOOR(nb_pkts, HNS3_DEFAULT_DESCS_PER_LOOP);

if (rxq->rx_rearm_nb > HNS3_DEFAULT_RXQ_REARM_THRESH)
hns3_rxq_rearm_mbuf(rxq);

Expand All @@ -124,10 +123,31 @@ hns3_recv_pkts_vec(void *__restrict rx_queue,
rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 2].mbuf);
rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 3].mbuf);

bd_err_mask = 0;
nb_rx = hns3_recv_burst_vec(rxq, rx_pkts, nb_pkts, &bd_err_mask);
if (unlikely(bd_err_mask))
nb_rx = hns3_rx_reassemble_pkts(rx_pkts, nb_rx, bd_err_mask);
if (likely(nb_pkts <= HNS3_DEFAULT_RX_BURST)) {
pkt_err_mask = 0;
nb_rx = hns3_recv_burst_vec(rxq, rx_pkts, nb_pkts,
&pkt_err_mask);
nb_rx = hns3_rx_reassemble_pkts(rx_pkts, nb_rx, pkt_err_mask);
return nb_rx;
}

nb_rx = 0;
while (nb_pkts > 0) {
uint16_t ret, n;

n = RTE_MIN(nb_pkts, HNS3_DEFAULT_RX_BURST);
pkt_err_mask = 0;
ret = hns3_recv_burst_vec(rxq, &rx_pkts[nb_rx], n,
&pkt_err_mask);
nb_pkts -= ret;
nb_rx += hns3_rx_reassemble_pkts(&rx_pkts[nb_rx], ret,
pkt_err_mask);
if (ret < n)
break;

if (rxq->rx_rearm_nb > HNS3_DEFAULT_RXQ_REARM_THRESH)
hns3_rxq_rearm_mbuf(rxq);
}

return nb_rx;
}
Expand Down
3 changes: 3 additions & 0 deletions drivers/net/hns3/hns3_rxtx_vec.h
Expand Up @@ -71,6 +71,9 @@ hns3_rx_reassemble_pkts(struct rte_mbuf **rx_pkts,
uint16_t count, i;
uint64_t mask;

if (likely(pkt_err_mask == 0))
return nb_pkts;

count = 0;
for (i = 0; i < nb_pkts; i++) {
mask = ((uint64_t)1u) << i;
Expand Down
32 changes: 26 additions & 6 deletions drivers/net/hns3/hns3_rxtx_vec_sve.c
Expand Up @@ -287,12 +287,11 @@ hns3_recv_pkts_vec_sve(void *__restrict rx_queue,
{
struct hns3_rx_queue *rxq = rx_queue;
struct hns3_desc *rxdp = &rxq->rx_ring[rxq->next_to_use];
uint64_t bd_err_mask; /* bit mask indicate whick pkts is error */
uint64_t pkt_err_mask; /* bit mask indicate whick pkts is error */
uint16_t nb_rx;

rte_prefetch_non_temporal(rxdp);

nb_pkts = RTE_MIN(nb_pkts, HNS3_DEFAULT_RX_BURST);
nb_pkts = RTE_ALIGN_FLOOR(nb_pkts, HNS3_SVE_DEFAULT_DESCS_PER_LOOP);

if (rxq->rx_rearm_nb > HNS3_DEFAULT_RXQ_REARM_THRESH)
Expand All @@ -304,10 +303,31 @@ hns3_recv_pkts_vec_sve(void *__restrict rx_queue,

hns3_rx_prefetch_mbuf_sve(&rxq->sw_ring[rxq->next_to_use]);

bd_err_mask = 0;
nb_rx = hns3_recv_burst_vec_sve(rxq, rx_pkts, nb_pkts, &bd_err_mask);
if (unlikely(bd_err_mask))
nb_rx = hns3_rx_reassemble_pkts(rx_pkts, nb_rx, bd_err_mask);
if (likely(nb_pkts <= HNS3_DEFAULT_RX_BURST)) {
pkt_err_mask = 0;
nb_rx = hns3_recv_burst_vec_sve(rxq, rx_pkts, nb_pkts,
&pkt_err_mask);
nb_rx = hns3_rx_reassemble_pkts(rx_pkts, nb_rx, pkt_err_mask);
return nb_rx;
}

nb_rx = 0;
while (nb_pkts > 0) {
uint16_t ret, n;

n = RTE_MIN(nb_pkts, HNS3_DEFAULT_RX_BURST);
pkt_err_mask = 0;
ret = hns3_recv_burst_vec_sve(rxq, &rx_pkts[nb_rx], n,
&pkt_err_mask);
nb_pkts -= ret;
nb_rx += hns3_rx_reassemble_pkts(&rx_pkts[nb_rx], ret,
pkt_err_mask);
if (ret < n)
break;

if (rxq->rx_rearm_nb > HNS3_DEFAULT_RXQ_REARM_THRESH)
hns3_rxq_rearm_mbuf_sve(rxq);
}

return nb_rx;
}
Expand Down

0 comments on commit 29da2a8

Please sign in to comment.