if (tx_q->dirty_tx != tx_q->cur_tx)
return -EBUSY; /* still unfinished work */
-@@ -1294,7 +1294,7 @@ static void stmmac_display_rx_rings(stru
+@@ -1302,7 +1302,7 @@ static void stmmac_display_rx_rings(stru
/* Display RX rings */
for (queue = 0; queue < rx_cnt; queue++) {
pr_info("\tRX Queue %u rings\n", queue);
-@@ -1307,7 +1307,7 @@ static void stmmac_display_rx_rings(stru
+@@ -1315,7 +1315,7 @@ static void stmmac_display_rx_rings(stru
}
/* Display RX ring */
rx_q->dma_rx_phy, desc_size);
}
}
-@@ -1321,7 +1321,7 @@ static void stmmac_display_tx_rings(stru
+@@ -1329,7 +1329,7 @@ static void stmmac_display_tx_rings(stru
/* Display TX rings */
for (queue = 0; queue < tx_cnt; queue++) {
pr_info("\tTX Queue %d rings\n", queue);
-@@ -1336,7 +1336,7 @@ static void stmmac_display_tx_rings(stru
+@@ -1344,7 +1344,7 @@ static void stmmac_display_tx_rings(stru
desc_size = sizeof(struct dma_desc);
}
tx_q->dma_tx_phy, desc_size);
}
}
-@@ -1377,21 +1377,21 @@ static int stmmac_set_bfsize(int mtu, in
+@@ -1385,21 +1385,21 @@ static int stmmac_set_bfsize(int mtu, in
*/
static void stmmac_clear_rx_descriptors(struct stmmac_priv *priv, u32 queue)
{
}
/**
-@@ -1403,12 +1403,12 @@ static void stmmac_clear_rx_descriptors(
+@@ -1411,12 +1411,12 @@ static void stmmac_clear_rx_descriptors(
*/
static void stmmac_clear_tx_descriptors(struct stmmac_priv *priv, u32 queue)
{
struct dma_desc *p;
if (priv->extend_desc)
-@@ -1456,7 +1456,7 @@ static void stmmac_clear_descriptors(str
+@@ -1464,7 +1464,7 @@ static void stmmac_clear_descriptors(str
static int stmmac_init_rx_buffers(struct stmmac_priv *priv, struct dma_desc *p,
int i, gfp_t flags, u32 queue)
{
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
if (!buf->page) {
-@@ -1481,7 +1481,7 @@ static int stmmac_init_rx_buffers(struct
+@@ -1489,7 +1489,7 @@ static int stmmac_init_rx_buffers(struct
buf->addr = page_pool_get_dma_addr(buf->page) + buf->page_offset;
stmmac_set_desc_addr(priv, p, buf->addr);
stmmac_init_desc3(priv, p);
return 0;
-@@ -1495,7 +1495,7 @@ static int stmmac_init_rx_buffers(struct
+@@ -1503,7 +1503,7 @@ static int stmmac_init_rx_buffers(struct
*/
static void stmmac_free_rx_buffer(struct stmmac_priv *priv, u32 queue, int i)
{
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
if (buf->page)
-@@ -1515,7 +1515,7 @@ static void stmmac_free_rx_buffer(struct
+@@ -1523,7 +1523,7 @@ static void stmmac_free_rx_buffer(struct
*/
static void stmmac_free_tx_buffer(struct stmmac_priv *priv, u32 queue, int i)
{
if (tx_q->tx_skbuff_dma[i].buf &&
tx_q->tx_skbuff_dma[i].buf_type != STMMAC_TXBUF_T_XDP_TX) {
-@@ -1560,17 +1560,17 @@ static void dma_free_rx_skbufs(struct st
+@@ -1568,17 +1568,17 @@ static void dma_free_rx_skbufs(struct st
{
int i;
struct dma_desc *p;
int ret;
-@@ -1597,10 +1597,10 @@ static int stmmac_alloc_rx_buffers(struc
+@@ -1605,10 +1605,10 @@ static int stmmac_alloc_rx_buffers(struc
*/
static void dma_free_rx_xskbufs(struct stmmac_priv *priv, u32 queue)
{
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
if (!buf->xdp)
-@@ -1613,10 +1613,10 @@ static void dma_free_rx_xskbufs(struct s
+@@ -1621,10 +1621,10 @@ static void dma_free_rx_xskbufs(struct s
static int stmmac_alloc_rx_buffers_zc(struct stmmac_priv *priv, u32 queue)
{
struct stmmac_rx_buffer *buf;
dma_addr_t dma_addr;
struct dma_desc *p;
-@@ -1659,7 +1659,7 @@ static struct xsk_buff_pool *stmmac_get_
+@@ -1667,7 +1667,7 @@ static struct xsk_buff_pool *stmmac_get_
*/
static int __init_dma_rx_desc_rings(struct stmmac_priv *priv, u32 queue, gfp_t flags)
{
int ret;
netif_dbg(priv, probe, priv->dev,
-@@ -1705,11 +1705,11 @@ static int __init_dma_rx_desc_rings(stru
+@@ -1713,11 +1713,11 @@ static int __init_dma_rx_desc_rings(stru
if (priv->extend_desc)
stmmac_mode_init(priv, rx_q->dma_erx,
rx_q->dma_rx_phy,
}
return 0;
-@@ -1736,7 +1736,7 @@ static int init_dma_rx_desc_rings(struct
+@@ -1744,7 +1744,7 @@ static int init_dma_rx_desc_rings(struct
err_init_rx_buffers:
while (queue >= 0) {
if (rx_q->xsk_pool)
dma_free_rx_xskbufs(priv, queue);
-@@ -1765,7 +1765,7 @@ err_init_rx_buffers:
+@@ -1773,7 +1773,7 @@ err_init_rx_buffers:
*/
static int __init_dma_tx_desc_rings(struct stmmac_priv *priv, u32 queue)
{
int i;
netif_dbg(priv, probe, priv->dev,
-@@ -1777,16 +1777,16 @@ static int __init_dma_tx_desc_rings(stru
+@@ -1785,16 +1785,16 @@ static int __init_dma_tx_desc_rings(stru
if (priv->extend_desc)
stmmac_mode_init(priv, tx_q->dma_etx,
tx_q->dma_tx_phy,
struct dma_desc *p;
if (priv->extend_desc)
-@@ -1856,12 +1856,12 @@ static int init_dma_desc_rings(struct ne
+@@ -1864,12 +1864,12 @@ static int init_dma_desc_rings(struct ne
*/
static void dma_free_tx_skbufs(struct stmmac_priv *priv, u32 queue)
{
stmmac_free_tx_buffer(priv, queue, i);
if (tx_q->xsk_pool && tx_q->xsk_frames_done) {
-@@ -1891,7 +1891,7 @@ static void stmmac_free_tx_skbufs(struct
+@@ -1899,7 +1899,7 @@ static void stmmac_free_tx_skbufs(struct
*/
static void __free_dma_rx_desc_resources(struct stmmac_priv *priv, u32 queue)
{
/* Release the DMA RX socket buffers */
if (rx_q->xsk_pool)
-@@ -1904,11 +1904,11 @@ static void __free_dma_rx_desc_resources
+@@ -1912,11 +1912,11 @@ static void __free_dma_rx_desc_resources
/* Free DMA regions of consistent memory previously allocated */
if (!priv->extend_desc)
sizeof(struct dma_extended_desc),
rx_q->dma_erx, rx_q->dma_rx_phy);
-@@ -1937,7 +1937,7 @@ static void free_dma_rx_desc_resources(s
+@@ -1945,7 +1945,7 @@ static void free_dma_rx_desc_resources(s
*/
static void __free_dma_tx_desc_resources(struct stmmac_priv *priv, u32 queue)
{
size_t size;
void *addr;
-@@ -1955,7 +1955,7 @@ static void __free_dma_tx_desc_resources
+@@ -1963,7 +1963,7 @@ static void __free_dma_tx_desc_resources
addr = tx_q->dma_tx;
}
dma_free_coherent(priv->device, size, addr, tx_q->dma_tx_phy);
-@@ -1984,7 +1984,7 @@ static void free_dma_tx_desc_resources(s
+@@ -1992,7 +1992,7 @@ static void free_dma_tx_desc_resources(s
*/
static int __alloc_dma_rx_desc_resources(struct stmmac_priv *priv, u32 queue)
{
struct stmmac_channel *ch = &priv->channel[queue];
bool xdp_prog = stmmac_xdp_is_enabled(priv);
struct page_pool_params pp_params = { 0 };
-@@ -1996,8 +1996,8 @@ static int __alloc_dma_rx_desc_resources
+@@ -2004,8 +2004,8 @@ static int __alloc_dma_rx_desc_resources
rx_q->priv_data = priv;
pp_params.flags = PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV;
pp_params.order = ilog2(num_pages);
pp_params.nid = dev_to_node(priv->device);
pp_params.dev = priv->device;
-@@ -2012,7 +2012,7 @@ static int __alloc_dma_rx_desc_resources
+@@ -2020,7 +2020,7 @@ static int __alloc_dma_rx_desc_resources
return ret;
}
sizeof(*rx_q->buf_pool),
GFP_KERNEL);
if (!rx_q->buf_pool)
-@@ -2020,7 +2020,7 @@ static int __alloc_dma_rx_desc_resources
+@@ -2028,7 +2028,7 @@ static int __alloc_dma_rx_desc_resources
if (priv->extend_desc) {
rx_q->dma_erx = dma_alloc_coherent(priv->device,
sizeof(struct dma_extended_desc),
&rx_q->dma_rx_phy,
GFP_KERNEL);
-@@ -2029,7 +2029,7 @@ static int __alloc_dma_rx_desc_resources
+@@ -2037,7 +2037,7 @@ static int __alloc_dma_rx_desc_resources
} else {
rx_q->dma_rx = dma_alloc_coherent(priv->device,
sizeof(struct dma_desc),
&rx_q->dma_rx_phy,
GFP_KERNEL);
-@@ -2086,20 +2086,20 @@ err_dma:
+@@ -2094,20 +2094,20 @@ err_dma:
*/
static int __alloc_dma_tx_desc_resources(struct stmmac_priv *priv, u32 queue)
{
sizeof(struct sk_buff *),
GFP_KERNEL);
if (!tx_q->tx_skbuff)
-@@ -2112,7 +2112,7 @@ static int __alloc_dma_tx_desc_resources
+@@ -2120,7 +2120,7 @@ static int __alloc_dma_tx_desc_resources
else
size = sizeof(struct dma_desc);
addr = dma_alloc_coherent(priv->device, size,
&tx_q->dma_tx_phy, GFP_KERNEL);
-@@ -2356,7 +2356,7 @@ static void stmmac_dma_operation_mode(st
+@@ -2364,7 +2364,7 @@ static void stmmac_dma_operation_mode(st
/* configure all channels */
for (chan = 0; chan < rx_channels_count; chan++) {
u32 buf_size;
qmode = priv->plat->rx_queues_cfg[chan].mode_to_use;
-@@ -2371,7 +2371,7 @@ static void stmmac_dma_operation_mode(st
+@@ -2379,7 +2379,7 @@ static void stmmac_dma_operation_mode(st
chan);
} else {
stmmac_set_dma_bfsize(priv, priv->ioaddr,
chan);
}
}
-@@ -2387,7 +2387,7 @@ static void stmmac_dma_operation_mode(st
+@@ -2395,7 +2395,7 @@ static void stmmac_dma_operation_mode(st
static bool stmmac_xdp_xmit_zc(struct stmmac_priv *priv, u32 queue, u32 budget)
{
struct netdev_queue *nq = netdev_get_tx_queue(priv->dev, queue);
struct xsk_buff_pool *pool = tx_q->xsk_pool;
unsigned int entry = tx_q->cur_tx;
struct dma_desc *tx_desc = NULL;
-@@ -2462,7 +2462,7 @@ static bool stmmac_xdp_xmit_zc(struct st
+@@ -2470,7 +2470,7 @@ static bool stmmac_xdp_xmit_zc(struct st
stmmac_enable_dma_transmission(priv, priv->ioaddr);
entry = tx_q->cur_tx;
}
-@@ -2488,7 +2488,7 @@ static bool stmmac_xdp_xmit_zc(struct st
+@@ -2496,7 +2496,7 @@ static bool stmmac_xdp_xmit_zc(struct st
*/
static int stmmac_tx_clean(struct stmmac_priv *priv, int budget, u32 queue)
{
unsigned int bytes_compl = 0, pkts_compl = 0;
unsigned int entry, xmits = 0, count = 0;
-@@ -2501,7 +2501,7 @@ static int stmmac_tx_clean(struct stmmac
+@@ -2509,7 +2509,7 @@ static int stmmac_tx_clean(struct stmmac
entry = tx_q->dirty_tx;
/* Try to clean all TX complete frame in 1 shot */
struct xdp_frame *xdpf;
struct sk_buff *skb;
struct dma_desc *p;
-@@ -2601,7 +2601,7 @@ static int stmmac_tx_clean(struct stmmac
+@@ -2609,7 +2609,7 @@ static int stmmac_tx_clean(struct stmmac
stmmac_release_tx_desc(priv, p, priv->mode);
}
tx_q->dirty_tx = entry;
-@@ -2666,7 +2666,7 @@ static int stmmac_tx_clean(struct stmmac
+@@ -2674,7 +2674,7 @@ static int stmmac_tx_clean(struct stmmac
*/
static void stmmac_tx_err(struct stmmac_priv *priv, u32 chan)
{
netif_tx_stop_queue(netdev_get_tx_queue(priv->dev, chan));
-@@ -2733,8 +2733,8 @@ static int stmmac_napi_check(struct stmm
+@@ -2741,8 +2741,8 @@ static int stmmac_napi_check(struct stmm
{
int status = stmmac_dma_interrupt_status(priv, priv->ioaddr,
&priv->xstats, chan, dir);
struct stmmac_channel *ch = &priv->channel[chan];
struct napi_struct *rx_napi;
struct napi_struct *tx_napi;
-@@ -2910,7 +2910,7 @@ static int stmmac_init_dma_engine(struct
+@@ -2918,7 +2918,7 @@ static int stmmac_init_dma_engine(struct
/* DMA RX Channel Configuration */
for (chan = 0; chan < rx_channels_count; chan++) {
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
rx_q->dma_rx_phy, chan);
-@@ -2924,7 +2924,7 @@ static int stmmac_init_dma_engine(struct
+@@ -2932,7 +2932,7 @@ static int stmmac_init_dma_engine(struct
/* DMA TX Channel Configuration */
for (chan = 0; chan < tx_channels_count; chan++) {
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
tx_q->dma_tx_phy, chan);
-@@ -2939,7 +2939,7 @@ static int stmmac_init_dma_engine(struct
+@@ -2947,7 +2947,7 @@ static int stmmac_init_dma_engine(struct
static void stmmac_tx_timer_arm(struct stmmac_priv *priv, u32 queue)
{
hrtimer_start(&tx_q->txtimer,
STMMAC_COAL_TIMER(priv->tx_coal_timer[queue]),
-@@ -2989,7 +2989,7 @@ static void stmmac_init_coalesce(struct
+@@ -2997,7 +2997,7 @@ static void stmmac_init_coalesce(struct
u32 chan;
for (chan = 0; chan < tx_channel_count; chan++) {
priv->tx_coal_frames[chan] = STMMAC_TX_FRAMES;
priv->tx_coal_timer[chan] = STMMAC_COAL_TX_TIMER;
-@@ -3011,12 +3011,12 @@ static void stmmac_set_rings_length(stru
+@@ -3019,12 +3019,12 @@ static void stmmac_set_rings_length(stru
/* set TX ring length */
for (chan = 0; chan < tx_channels_count; chan++)
stmmac_set_tx_ring_len(priv, priv->ioaddr,
}
/**
-@@ -3351,7 +3351,7 @@ static int stmmac_hw_setup(struct net_de
+@@ -3359,7 +3359,7 @@ static int stmmac_hw_setup(struct net_de
/* Enable TSO */
if (priv->tso) {
for (chan = 0; chan < tx_cnt; chan++) {
/* TSO and TBS cannot co-exist */
if (tx_q->tbs & STMMAC_TBS_AVAIL)
-@@ -3373,7 +3373,7 @@ static int stmmac_hw_setup(struct net_de
+@@ -3381,7 +3381,7 @@ static int stmmac_hw_setup(struct net_de
/* TBS */
for (chan = 0; chan < tx_cnt; chan++) {
int enable = tx_q->tbs & STMMAC_TBS_AVAIL;
stmmac_enable_tbs(priv, priv->ioaddr, enable, chan);
-@@ -3417,7 +3417,7 @@ static void stmmac_free_irq(struct net_d
+@@ -3425,7 +3425,7 @@ static void stmmac_free_irq(struct net_d
for (j = irq_idx - 1; j >= 0; j--) {
if (priv->tx_irq[j] > 0) {
irq_set_affinity_hint(priv->tx_irq[j], NULL);
}
}
irq_idx = priv->plat->rx_queues_to_use;
-@@ -3426,7 +3426,7 @@ static void stmmac_free_irq(struct net_d
+@@ -3434,7 +3434,7 @@ static void stmmac_free_irq(struct net_d
for (j = irq_idx - 1; j >= 0; j--) {
if (priv->rx_irq[j] > 0) {
irq_set_affinity_hint(priv->rx_irq[j], NULL);
}
}
-@@ -3559,7 +3559,7 @@ static int stmmac_request_irq_multi_msi(
+@@ -3567,7 +3567,7 @@ static int stmmac_request_irq_multi_msi(
sprintf(int_name, "%s:%s-%d", dev->name, "rx", i);
ret = request_irq(priv->rx_irq[i],
stmmac_msi_intr_rx,
if (unlikely(ret < 0)) {
netdev_err(priv->dev,
"%s: alloc rx-%d MSI %d (error: %d)\n",
-@@ -3582,7 +3582,7 @@ static int stmmac_request_irq_multi_msi(
+@@ -3590,7 +3590,7 @@ static int stmmac_request_irq_multi_msi(
sprintf(int_name, "%s:%s-%d", dev->name, "tx", i);
ret = request_irq(priv->tx_irq[i],
stmmac_msi_intr_tx,
if (unlikely(ret < 0)) {
netdev_err(priv->dev,
"%s: alloc tx-%d MSI %d (error: %d)\n",
-@@ -3713,21 +3713,21 @@ static int stmmac_open(struct net_device
+@@ -3721,21 +3721,21 @@ static int stmmac_open(struct net_device
bfsize = 0;
if (bfsize < BUF_SIZE_16KiB)
int tbs_en = priv->plat->tx_queues_cfg[chan].tbs_en;
/* Setup per-TXQ tbs flag before TX descriptor alloc */
-@@ -3785,7 +3785,7 @@ irq_error:
+@@ -3793,7 +3793,7 @@ irq_error:
phylink_stop(priv->phylink);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
stmmac_hw_teardown(dev);
init_error:
-@@ -3827,7 +3827,7 @@ static int stmmac_release(struct net_dev
+@@ -3835,7 +3835,7 @@ static int stmmac_release(struct net_dev
stmmac_disable_all_queues(priv);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
netif_tx_disable(dev);
-@@ -3891,7 +3891,7 @@ static bool stmmac_vlan_insert(struct st
+@@ -3899,7 +3899,7 @@ static bool stmmac_vlan_insert(struct st
return false;
stmmac_set_tx_owner(priv, p);
return true;
}
-@@ -3909,7 +3909,7 @@ static bool stmmac_vlan_insert(struct st
+@@ -3917,7 +3917,7 @@ static bool stmmac_vlan_insert(struct st
static void stmmac_tso_allocator(struct stmmac_priv *priv, dma_addr_t des,
int total_len, bool last_segment, u32 queue)
{
struct dma_desc *desc;
u32 buff_size;
int tmp_len;
-@@ -3920,7 +3920,7 @@ static void stmmac_tso_allocator(struct
+@@ -3928,7 +3928,7 @@ static void stmmac_tso_allocator(struct
dma_addr_t curr_addr;
tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx,
WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]);
if (tx_q->tbs & STMMAC_TBS_AVAIL)
-@@ -3948,7 +3948,7 @@ static void stmmac_tso_allocator(struct
+@@ -3956,7 +3956,7 @@ static void stmmac_tso_allocator(struct
static void stmmac_flush_tx_descriptors(struct stmmac_priv *priv, int queue)
{
int desc_size;
if (likely(priv->extend_desc))
-@@ -4010,7 +4010,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
+@@ -4018,7 +4018,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
dma_addr_t des;
int i;
first_tx = tx_q->cur_tx;
/* Compute header lengths */
-@@ -4050,7 +4050,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
+@@ -4058,7 +4058,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
stmmac_set_mss(priv, mss_desc, mss);
tx_q->mss = mss;
tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx,
WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]);
}
-@@ -4162,7 +4162,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
+@@ -4170,7 +4170,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
* ndo_start_xmit will fill this descriptor the next time it's
* called and stmmac_tx_clean may clean up to this descriptor.
*/
if (unlikely(stmmac_tx_avail(priv, queue) <= (MAX_SKB_FRAGS + 1))) {
netif_dbg(priv, hw, priv->dev, "%s: stop transmitted packets\n",
-@@ -4250,7 +4250,7 @@ static netdev_tx_t stmmac_xmit(struct sk
+@@ -4258,7 +4258,7 @@ static netdev_tx_t stmmac_xmit(struct sk
int entry, first_tx;
dma_addr_t des;
first_tx = tx_q->cur_tx;
if (priv->tx_path_in_lpi_mode && priv->eee_sw_timer_en)
-@@ -4313,7 +4313,7 @@ static netdev_tx_t stmmac_xmit(struct sk
+@@ -4321,7 +4321,7 @@ static netdev_tx_t stmmac_xmit(struct sk
int len = skb_frag_size(frag);
bool last_segment = (i == (nfrags - 1));
WARN_ON(tx_q->tx_skbuff[entry]);
if (likely(priv->extend_desc))
-@@ -4384,7 +4384,7 @@ static netdev_tx_t stmmac_xmit(struct sk
+@@ -4392,7 +4392,7 @@ static netdev_tx_t stmmac_xmit(struct sk
* ndo_start_xmit will fill this descriptor the next time it's
* called and stmmac_tx_clean may clean up to this descriptor.
*/
tx_q->cur_tx = entry;
if (netif_msg_pktdata(priv)) {
-@@ -4499,7 +4499,7 @@ static void stmmac_rx_vlan(struct net_de
+@@ -4507,7 +4507,7 @@ static void stmmac_rx_vlan(struct net_de
*/
static inline void stmmac_rx_refill(struct stmmac_priv *priv, u32 queue)
{
int dirty = stmmac_rx_dirty(priv, queue);
unsigned int entry = rx_q->dirty_rx;
-@@ -4549,7 +4549,7 @@ static inline void stmmac_rx_refill(stru
+@@ -4557,7 +4557,7 @@ static inline void stmmac_rx_refill(stru
dma_wmb();
stmmac_set_rx_owner(priv, p, use_rx_wd);
}
rx_q->dirty_rx = entry;
rx_q->rx_tail_addr = rx_q->dma_rx_phy +
-@@ -4577,12 +4577,12 @@ static unsigned int stmmac_rx_buf1_len(s
+@@ -4585,12 +4585,12 @@ static unsigned int stmmac_rx_buf1_len(s
/* First descriptor, not last descriptor and not split header */
if (status & rx_not_ls)
}
static unsigned int stmmac_rx_buf2_len(struct stmmac_priv *priv,
-@@ -4598,7 +4598,7 @@ static unsigned int stmmac_rx_buf2_len(s
+@@ -4606,7 +4606,7 @@ static unsigned int stmmac_rx_buf2_len(s
/* Not last descriptor */
if (status & rx_not_ls)
plen = stmmac_get_rx_frame_len(priv, p, coe);
-@@ -4609,7 +4609,7 @@ static unsigned int stmmac_rx_buf2_len(s
+@@ -4617,7 +4617,7 @@ static unsigned int stmmac_rx_buf2_len(s
static int stmmac_xdp_xmit_xdpf(struct stmmac_priv *priv, int queue,
struct xdp_frame *xdpf, bool dma_map)
{
unsigned int entry = tx_q->cur_tx;
struct dma_desc *tx_desc;
dma_addr_t dma_addr;
-@@ -4672,7 +4672,7 @@ static int stmmac_xdp_xmit_xdpf(struct s
+@@ -4680,7 +4680,7 @@ static int stmmac_xdp_xmit_xdpf(struct s
stmmac_enable_dma_transmission(priv, priv->ioaddr);
tx_q->cur_tx = entry;
return STMMAC_XDP_TX;
-@@ -4846,7 +4846,7 @@ static void stmmac_dispatch_skb_zc(struc
+@@ -4854,7 +4854,7 @@ static void stmmac_dispatch_skb_zc(struc
static bool stmmac_rx_refill_zc(struct stmmac_priv *priv, u32 queue, u32 budget)
{
unsigned int entry = rx_q->dirty_rx;
struct dma_desc *rx_desc = NULL;
bool ret = true;
-@@ -4889,7 +4889,7 @@ static bool stmmac_rx_refill_zc(struct s
+@@ -4897,7 +4897,7 @@ static bool stmmac_rx_refill_zc(struct s
dma_wmb();
stmmac_set_rx_owner(priv, rx_desc, use_rx_wd);
}
if (rx_desc) {
-@@ -4904,7 +4904,7 @@ static bool stmmac_rx_refill_zc(struct s
+@@ -4912,7 +4912,7 @@ static bool stmmac_rx_refill_zc(struct s
static int stmmac_rx_zc(struct stmmac_priv *priv, int limit, u32 queue)
{
unsigned int count = 0, error = 0, len = 0;
int dirty = stmmac_rx_dirty(priv, queue);
unsigned int next_entry = rx_q->cur_rx;
-@@ -4926,7 +4926,7 @@ static int stmmac_rx_zc(struct stmmac_pr
+@@ -4934,7 +4934,7 @@ static int stmmac_rx_zc(struct stmmac_pr
desc_size = sizeof(struct dma_desc);
}
rx_q->dma_rx_phy, desc_size);
}
while (count < limit) {
-@@ -4973,7 +4973,7 @@ read_again:
+@@ -4981,7 +4981,7 @@ read_again:
/* Prefetch the next RX descriptor */
rx_q->cur_rx = STMMAC_GET_ENTRY(rx_q->cur_rx,
next_entry = rx_q->cur_rx;
if (priv->extend_desc)
-@@ -5094,7 +5094,7 @@ read_again:
+@@ -5102,7 +5102,7 @@ read_again:
*/
static int stmmac_rx(struct stmmac_priv *priv, int limit, u32 queue)
{
struct stmmac_channel *ch = &priv->channel[queue];
unsigned int count = 0, error = 0, len = 0;
int status = 0, coe = priv->hw->rx_csum;
-@@ -5107,7 +5107,7 @@ static int stmmac_rx(struct stmmac_priv
+@@ -5115,7 +5115,7 @@ static int stmmac_rx(struct stmmac_priv
int buf_sz;
dma_dir = page_pool_get_dma_dir(rx_q->page_pool);
if (netif_msg_rx_status(priv)) {
void *rx_head;
-@@ -5121,7 +5121,7 @@ static int stmmac_rx(struct stmmac_priv
+@@ -5129,7 +5129,7 @@ static int stmmac_rx(struct stmmac_priv
desc_size = sizeof(struct dma_desc);
}
rx_q->dma_rx_phy, desc_size);
}
while (count < limit) {
-@@ -5165,7 +5165,7 @@ read_again:
+@@ -5173,7 +5173,7 @@ read_again:
break;
rx_q->cur_rx = STMMAC_GET_ENTRY(rx_q->cur_rx,
next_entry = rx_q->cur_rx;
if (priv->extend_desc)
-@@ -5299,7 +5299,7 @@ read_again:
+@@ -5307,7 +5307,7 @@ read_again:
buf1_len, dma_dir);
skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags,
buf->page, buf->page_offset, buf1_len,
/* Data payload appended into SKB */
page_pool_release_page(rx_q->page_pool, buf->page);
-@@ -5311,7 +5311,7 @@ read_again:
+@@ -5319,7 +5319,7 @@ read_again:
buf2_len, dma_dir);
skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags,
buf->sec_page, 0, buf2_len,
/* Data payload appended into SKB */
page_pool_release_page(rx_q->page_pool, buf->sec_page);
-@@ -5753,11 +5753,13 @@ static irqreturn_t stmmac_safety_interru
+@@ -5761,11 +5761,13 @@ static irqreturn_t stmmac_safety_interru
static irqreturn_t stmmac_msi_intr_tx(int irq, void *data)
{
struct stmmac_tx_queue *tx_q = (struct stmmac_tx_queue *)data;
if (unlikely(!data)) {
netdev_err(priv->dev, "%s: invalid dev pointer\n", __func__);
-@@ -5797,10 +5799,12 @@ static irqreturn_t stmmac_msi_intr_tx(in
+@@ -5805,10 +5807,12 @@ static irqreturn_t stmmac_msi_intr_tx(in
static irqreturn_t stmmac_msi_intr_rx(int irq, void *data)
{
struct stmmac_rx_queue *rx_q = (struct stmmac_rx_queue *)data;
if (unlikely(!data)) {
netdev_err(priv->dev, "%s: invalid dev pointer\n", __func__);
-@@ -5831,10 +5835,10 @@ static void stmmac_poll_controller(struc
+@@ -5839,10 +5843,10 @@ static void stmmac_poll_controller(struc
if (priv->plat->multi_msi_en) {
for (i = 0; i < priv->plat->rx_queues_to_use; i++)
} else {
disable_irq(dev->irq);
stmmac_interrupt(dev->irq, dev);
-@@ -6015,34 +6019,34 @@ static int stmmac_rings_status_show(stru
+@@ -6023,34 +6027,34 @@ static int stmmac_rings_status_show(stru
return 0;
for (queue = 0; queue < rx_count; queue++) {
}
}
-@@ -6383,7 +6387,7 @@ void stmmac_disable_rx_queue(struct stmm
+@@ -6391,7 +6395,7 @@ void stmmac_disable_rx_queue(struct stmm
void stmmac_enable_rx_queue(struct stmmac_priv *priv, u32 queue)
{
struct stmmac_channel *ch = &priv->channel[queue];
unsigned long flags;
u32 buf_size;
-@@ -6420,7 +6424,7 @@ void stmmac_enable_rx_queue(struct stmma
+@@ -6428,7 +6432,7 @@ void stmmac_enable_rx_queue(struct stmma
rx_q->queue_index);
} else {
stmmac_set_dma_bfsize(priv, priv->ioaddr,
rx_q->queue_index);
}
-@@ -6446,7 +6450,7 @@ void stmmac_disable_tx_queue(struct stmm
+@@ -6454,7 +6458,7 @@ void stmmac_disable_tx_queue(struct stmm
void stmmac_enable_tx_queue(struct stmmac_priv *priv, u32 queue)
{
struct stmmac_channel *ch = &priv->channel[queue];
unsigned long flags;
int ret;
-@@ -6496,7 +6500,7 @@ void stmmac_xdp_release(struct net_devic
+@@ -6504,7 +6508,7 @@ void stmmac_xdp_release(struct net_devic
stmmac_disable_all_queues(priv);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
/* Free the IRQ lines */
stmmac_free_irq(dev, REQ_IRQ_ERR_ALL, 0);
-@@ -6555,7 +6559,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6563,7 +6567,7 @@ int stmmac_xdp_open(struct net_device *d
/* DMA RX Channel Configuration */
for (chan = 0; chan < rx_cnt; chan++) {
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
rx_q->dma_rx_phy, chan);
-@@ -6573,7 +6577,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6581,7 +6585,7 @@ int stmmac_xdp_open(struct net_device *d
rx_q->queue_index);
} else {
stmmac_set_dma_bfsize(priv, priv->ioaddr,
rx_q->queue_index);
}
-@@ -6582,7 +6586,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6590,7 +6594,7 @@ int stmmac_xdp_open(struct net_device *d
/* DMA TX Channel Configuration */
for (chan = 0; chan < tx_cnt; chan++) {
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
tx_q->dma_tx_phy, chan);
-@@ -6615,7 +6619,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6623,7 +6627,7 @@ int stmmac_xdp_open(struct net_device *d
irq_error:
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
stmmac_hw_teardown(dev);
init_error:
-@@ -6642,8 +6646,8 @@ int stmmac_xsk_wakeup(struct net_device
+@@ -6650,8 +6654,8 @@ int stmmac_xsk_wakeup(struct net_device
queue >= priv->plat->tx_queues_to_use)
return -EINVAL;
ch = &priv->channel[queue];
if (!rx_q->xsk_pool && !tx_q->xsk_pool)
-@@ -6899,8 +6903,8 @@ int stmmac_reinit_ringparam(struct net_d
+@@ -6907,8 +6911,8 @@ int stmmac_reinit_ringparam(struct net_d
if (netif_running(dev))
stmmac_release(dev);
if (netif_running(dev))
ret = stmmac_open(dev);
-@@ -7336,7 +7340,7 @@ int stmmac_suspend(struct device *dev)
+@@ -7344,7 +7348,7 @@ int stmmac_suspend(struct device *dev)
stmmac_disable_all_queues(priv);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
if (priv->eee_enabled) {
priv->tx_path_in_lpi_mode = false;
-@@ -7387,7 +7391,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend);
+@@ -7395,7 +7399,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend);
static void stmmac_reset_rx_queue(struct stmmac_priv *priv, u32 queue)
{
rx_q->cur_rx = 0;
rx_q->dirty_rx = 0;
-@@ -7395,7 +7399,7 @@ static void stmmac_reset_rx_queue(struct
+@@ -7403,7 +7407,7 @@ static void stmmac_reset_rx_queue(struct
static void stmmac_reset_tx_queue(struct stmmac_priv *priv, u32 queue)
{
--- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
+++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
-@@ -1285,7 +1285,8 @@ static int stmmac_phy_setup(struct stmma
+@@ -1293,7 +1293,8 @@ static int stmmac_phy_setup(struct stmma
return 0;
}
{
u32 rx_cnt = priv->plat->rx_queues_to_use;
unsigned int desc_size;
-@@ -1294,7 +1295,7 @@ static void stmmac_display_rx_rings(stru
+@@ -1302,7 +1303,7 @@ static void stmmac_display_rx_rings(stru
/* Display RX rings */
for (queue = 0; queue < rx_cnt; queue++) {
pr_info("\tRX Queue %u rings\n", queue);
-@@ -1307,12 +1308,13 @@ static void stmmac_display_rx_rings(stru
+@@ -1315,12 +1316,13 @@ static void stmmac_display_rx_rings(stru
}
/* Display RX ring */
{
u32 tx_cnt = priv->plat->tx_queues_to_use;
unsigned int desc_size;
-@@ -1321,7 +1323,7 @@ static void stmmac_display_tx_rings(stru
+@@ -1329,7 +1331,7 @@ static void stmmac_display_tx_rings(stru
/* Display TX rings */
for (queue = 0; queue < tx_cnt; queue++) {
pr_info("\tTX Queue %d rings\n", queue);
-@@ -1336,18 +1338,19 @@ static void stmmac_display_tx_rings(stru
+@@ -1344,18 +1346,19 @@ static void stmmac_display_tx_rings(stru
desc_size = sizeof(struct dma_desc);
}
}
static int stmmac_set_bfsize(int mtu, int bufsize)
-@@ -1371,44 +1374,50 @@ static int stmmac_set_bfsize(int mtu, in
+@@ -1379,44 +1382,50 @@ static int stmmac_set_bfsize(int mtu, in
/**
* stmmac_clear_rx_descriptors - clear RX descriptors
* @priv: driver private structure
struct dma_desc *p;
if (priv->extend_desc)
-@@ -1425,10 +1434,12 @@ static void stmmac_clear_tx_descriptors(
+@@ -1433,10 +1442,12 @@ static void stmmac_clear_tx_descriptors(
/**
* stmmac_clear_descriptors - clear descriptors
* @priv: driver private structure
{
u32 rx_queue_cnt = priv->plat->rx_queues_to_use;
u32 tx_queue_cnt = priv->plat->tx_queues_to_use;
-@@ -1436,16 +1447,17 @@ static void stmmac_clear_descriptors(str
+@@ -1444,16 +1455,17 @@ static void stmmac_clear_descriptors(str
/* Clear the RX descriptors */
for (queue = 0; queue < rx_queue_cnt; queue++)
* @p: descriptor pointer
* @i: descriptor index
* @flags: gfp flag
-@@ -1453,10 +1465,12 @@ static void stmmac_clear_descriptors(str
+@@ -1461,10 +1473,12 @@ static void stmmac_clear_descriptors(str
* Description: this function is called to allocate a receive buffer, perform
* the DMA mapping and init the descriptor.
*/
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
if (!buf->page) {
-@@ -1481,7 +1495,7 @@ static int stmmac_init_rx_buffers(struct
+@@ -1489,7 +1503,7 @@ static int stmmac_init_rx_buffers(struct
buf->addr = page_pool_get_dma_addr(buf->page) + buf->page_offset;
stmmac_set_desc_addr(priv, p, buf->addr);
stmmac_init_desc3(priv, p);
return 0;
-@@ -1490,12 +1504,13 @@ static int stmmac_init_rx_buffers(struct
+@@ -1498,12 +1512,13 @@ static int stmmac_init_rx_buffers(struct
/**
* stmmac_free_rx_buffer - free RX dma buffers
* @priv: private structure
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
if (buf->page)
-@@ -1510,12 +1525,15 @@ static void stmmac_free_rx_buffer(struct
+@@ -1518,12 +1533,15 @@ static void stmmac_free_rx_buffer(struct
/**
* stmmac_free_tx_buffer - free RX dma buffers
* @priv: private structure
if (tx_q->tx_skbuff_dma[i].buf &&
tx_q->tx_skbuff_dma[i].buf_type != STMMAC_TXBUF_T_XDP_TX) {
-@@ -1554,23 +1572,28 @@ static void stmmac_free_tx_buffer(struct
+@@ -1562,23 +1580,28 @@ static void stmmac_free_tx_buffer(struct
/**
* dma_free_rx_skbufs - free RX dma buffers
* @priv: private structure
struct dma_desc *p;
int ret;
-@@ -1579,7 +1602,7 @@ static int stmmac_alloc_rx_buffers(struc
+@@ -1587,7 +1610,7 @@ static int stmmac_alloc_rx_buffers(struc
else
p = rx_q->dma_rx + i;
queue);
if (ret)
return ret;
-@@ -1593,14 +1616,17 @@ static int stmmac_alloc_rx_buffers(struc
+@@ -1601,14 +1624,17 @@ static int stmmac_alloc_rx_buffers(struc
/**
* dma_free_rx_xskbufs - free RX dma buffers from XSK pool
* @priv: private structure
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
if (!buf->xdp)
-@@ -1611,12 +1637,14 @@ static void dma_free_rx_xskbufs(struct s
+@@ -1619,12 +1645,14 @@ static void dma_free_rx_xskbufs(struct s
}
}
struct stmmac_rx_buffer *buf;
dma_addr_t dma_addr;
struct dma_desc *p;
-@@ -1651,22 +1679,25 @@ static struct xsk_buff_pool *stmmac_get_
+@@ -1659,22 +1687,25 @@ static struct xsk_buff_pool *stmmac_get_
/**
* __init_dma_rx_desc_rings - init the RX descriptor ring (per queue)
* @priv: driver private structure
xdp_rxq_info_unreg_mem_model(&rx_q->xdp_rxq);
-@@ -1693,9 +1724,9 @@ static int __init_dma_rx_desc_rings(stru
+@@ -1701,9 +1732,9 @@ static int __init_dma_rx_desc_rings(stru
/* RX XDP ZC buffer pool may not be populated, e.g.
* xdpsock TX-only.
*/
if (ret < 0)
return -ENOMEM;
}
-@@ -1705,17 +1736,19 @@ static int __init_dma_rx_desc_rings(stru
+@@ -1713,17 +1744,19 @@ static int __init_dma_rx_desc_rings(stru
if (priv->extend_desc)
stmmac_mode_init(priv, rx_q->dma_erx,
rx_q->dma_rx_phy,
{
struct stmmac_priv *priv = netdev_priv(dev);
u32 rx_count = priv->plat->rx_queues_to_use;
-@@ -1727,7 +1760,7 @@ static int init_dma_rx_desc_rings(struct
+@@ -1735,7 +1768,7 @@ static int init_dma_rx_desc_rings(struct
"SKB addresses:\nskb\t\tskb data\tdma data\n");
for (queue = 0; queue < rx_count; queue++) {
if (ret)
goto err_init_rx_buffers;
}
-@@ -1736,12 +1769,12 @@ static int init_dma_rx_desc_rings(struct
+@@ -1744,12 +1777,12 @@ static int init_dma_rx_desc_rings(struct
err_init_rx_buffers:
while (queue >= 0) {
rx_q->buf_alloc_num = 0;
rx_q->xsk_pool = NULL;
-@@ -1758,14 +1791,17 @@ err_init_rx_buffers:
+@@ -1766,14 +1799,17 @@ err_init_rx_buffers:
/**
* __init_dma_tx_desc_rings - init the TX descriptor ring (per queue)
* @priv: driver private structure
int i;
netif_dbg(priv, probe, priv->dev,
-@@ -1777,16 +1813,16 @@ static int __init_dma_tx_desc_rings(stru
+@@ -1785,16 +1821,16 @@ static int __init_dma_tx_desc_rings(stru
if (priv->extend_desc)
stmmac_mode_init(priv, tx_q->dma_etx,
tx_q->dma_tx_phy,
struct dma_desc *p;
if (priv->extend_desc)
-@@ -1808,7 +1844,8 @@ static int __init_dma_tx_desc_rings(stru
+@@ -1816,7 +1852,8 @@ static int __init_dma_tx_desc_rings(stru
return 0;
}
{
struct stmmac_priv *priv = netdev_priv(dev);
u32 tx_queue_cnt;
-@@ -1817,7 +1854,7 @@ static int init_dma_tx_desc_rings(struct
+@@ -1825,7 +1862,7 @@ static int init_dma_tx_desc_rings(struct
tx_queue_cnt = priv->plat->tx_queues_to_use;
for (queue = 0; queue < tx_queue_cnt; queue++)
return 0;
}
-@@ -1825,26 +1862,29 @@ static int init_dma_tx_desc_rings(struct
+@@ -1833,26 +1870,29 @@ static int init_dma_tx_desc_rings(struct
/**
* init_dma_desc_rings - init the RX/TX descriptor rings
* @dev: net device structure
return ret;
}
-@@ -1852,17 +1892,20 @@ static int init_dma_desc_rings(struct ne
+@@ -1860,17 +1900,20 @@ static int init_dma_desc_rings(struct ne
/**
* dma_free_tx_skbufs - free TX dma buffers
* @priv: private structure
if (tx_q->xsk_pool && tx_q->xsk_frames_done) {
xsk_tx_completed(tx_q->xsk_pool, tx_q->xsk_frames_done);
-@@ -1881,34 +1924,37 @@ static void stmmac_free_tx_skbufs(struct
+@@ -1889,34 +1932,37 @@ static void stmmac_free_tx_skbufs(struct
u32 queue;
for (queue = 0; queue < tx_queue_cnt; queue++)
sizeof(struct dma_extended_desc),
rx_q->dma_erx, rx_q->dma_rx_phy);
-@@ -1920,29 +1966,33 @@ static void __free_dma_rx_desc_resources
+@@ -1928,29 +1974,33 @@ static void __free_dma_rx_desc_resources
page_pool_destroy(rx_q->page_pool);
}
if (priv->extend_desc) {
size = sizeof(struct dma_extended_desc);
-@@ -1955,7 +2005,7 @@ static void __free_dma_tx_desc_resources
+@@ -1963,7 +2013,7 @@ static void __free_dma_tx_desc_resources
addr = tx_q->dma_tx;
}
dma_free_coherent(priv->device, size, addr, tx_q->dma_tx_phy);
-@@ -1963,28 +2013,32 @@ static void __free_dma_tx_desc_resources
+@@ -1971,28 +2021,32 @@ static void __free_dma_tx_desc_resources
kfree(tx_q->tx_skbuff);
}
struct stmmac_channel *ch = &priv->channel[queue];
bool xdp_prog = stmmac_xdp_is_enabled(priv);
struct page_pool_params pp_params = { 0 };
-@@ -1996,8 +2050,8 @@ static int __alloc_dma_rx_desc_resources
+@@ -2004,8 +2058,8 @@ static int __alloc_dma_rx_desc_resources
rx_q->priv_data = priv;
pp_params.flags = PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV;
pp_params.order = ilog2(num_pages);
pp_params.nid = dev_to_node(priv->device);
pp_params.dev = priv->device;
-@@ -2012,7 +2066,7 @@ static int __alloc_dma_rx_desc_resources
+@@ -2020,7 +2074,7 @@ static int __alloc_dma_rx_desc_resources
return ret;
}
sizeof(*rx_q->buf_pool),
GFP_KERNEL);
if (!rx_q->buf_pool)
-@@ -2020,7 +2074,7 @@ static int __alloc_dma_rx_desc_resources
+@@ -2028,7 +2082,7 @@ static int __alloc_dma_rx_desc_resources
if (priv->extend_desc) {
rx_q->dma_erx = dma_alloc_coherent(priv->device,
sizeof(struct dma_extended_desc),
&rx_q->dma_rx_phy,
GFP_KERNEL);
-@@ -2029,7 +2083,7 @@ static int __alloc_dma_rx_desc_resources
+@@ -2037,7 +2091,7 @@ static int __alloc_dma_rx_desc_resources
} else {
rx_q->dma_rx = dma_alloc_coherent(priv->device,
sizeof(struct dma_desc),
&rx_q->dma_rx_phy,
GFP_KERNEL);
-@@ -2054,7 +2108,8 @@ static int __alloc_dma_rx_desc_resources
+@@ -2062,7 +2116,8 @@ static int __alloc_dma_rx_desc_resources
return 0;
}
{
u32 rx_count = priv->plat->rx_queues_to_use;
u32 queue;
-@@ -2062,7 +2117,7 @@ static int alloc_dma_rx_desc_resources(s
+@@ -2070,7 +2125,7 @@ static int alloc_dma_rx_desc_resources(s
/* RX queues buffers and DMA */
for (queue = 0; queue < rx_count; queue++) {
if (ret)
goto err_dma;
}
-@@ -2070,7 +2125,7 @@ static int alloc_dma_rx_desc_resources(s
+@@ -2078,7 +2133,7 @@ static int alloc_dma_rx_desc_resources(s
return 0;
err_dma:
return ret;
}
-@@ -2078,28 +2133,31 @@ err_dma:
+@@ -2086,28 +2141,31 @@ err_dma:
/**
* __alloc_dma_tx_desc_resources - alloc TX resources (per queue).
* @priv: private structure
sizeof(struct sk_buff *),
GFP_KERNEL);
if (!tx_q->tx_skbuff)
-@@ -2112,7 +2170,7 @@ static int __alloc_dma_tx_desc_resources
+@@ -2120,7 +2178,7 @@ static int __alloc_dma_tx_desc_resources
else
size = sizeof(struct dma_desc);
addr = dma_alloc_coherent(priv->device, size,
&tx_q->dma_tx_phy, GFP_KERNEL);
-@@ -2129,7 +2187,8 @@ static int __alloc_dma_tx_desc_resources
+@@ -2137,7 +2195,8 @@ static int __alloc_dma_tx_desc_resources
return 0;
}
{
u32 tx_count = priv->plat->tx_queues_to_use;
u32 queue;
-@@ -2137,7 +2196,7 @@ static int alloc_dma_tx_desc_resources(s
+@@ -2145,7 +2204,7 @@ static int alloc_dma_tx_desc_resources(s
/* TX queues buffers and DMA */
for (queue = 0; queue < tx_count; queue++) {
if (ret)
goto err_dma;
}
-@@ -2145,27 +2204,29 @@ static int alloc_dma_tx_desc_resources(s
+@@ -2153,27 +2212,29 @@ static int alloc_dma_tx_desc_resources(s
return 0;
err_dma:
return ret;
}
-@@ -2173,16 +2234,18 @@ static int alloc_dma_desc_resources(stru
+@@ -2181,16 +2242,18 @@ static int alloc_dma_desc_resources(stru
/**
* free_dma_desc_resources - free dma desc resources
* @priv: private structure
}
/**
-@@ -2671,8 +2734,8 @@ static void stmmac_tx_err(struct stmmac_
+@@ -2679,8 +2742,8 @@ static void stmmac_tx_err(struct stmmac_
netif_tx_stop_queue(netdev_get_tx_queue(priv->dev, chan));
stmmac_stop_tx_dma(priv, chan);
stmmac_reset_tx_queue(priv, chan);
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
tx_q->dma_tx_phy, chan);
-@@ -3669,19 +3732,93 @@ static int stmmac_request_irq(struct net
+@@ -3677,19 +3740,93 @@ static int stmmac_request_irq(struct net
}
/**
u32 chan;
int ret;
-@@ -3708,45 +3845,10 @@ static int stmmac_open(struct net_device
+@@ -3716,45 +3853,10 @@ static int stmmac_open(struct net_device
memset(&priv->xstats, 0, sizeof(struct stmmac_extra_stats));
priv->xstats.threshold = tc;
if (priv->plat->serdes_powerup) {
ret = priv->plat->serdes_powerup(dev, priv->plat->bsp_priv);
-@@ -3789,14 +3891,28 @@ irq_error:
+@@ -3797,14 +3899,28 @@ irq_error:
stmmac_hw_teardown(dev);
init_error:
static void stmmac_fpe_stop_wq(struct stmmac_priv *priv)
{
set_bit(__FPE_REMOVING, &priv->fpe_task_state);
-@@ -3843,7 +3959,7 @@ static int stmmac_release(struct net_dev
+@@ -3851,7 +3967,7 @@ static int stmmac_release(struct net_dev
stmmac_stop_all_dma(priv);
/* Release and free the Rx/Tx resources */
/* Disable the MAC Rx/Tx */
stmmac_mac_set(priv, priv->ioaddr, false);
-@@ -6382,7 +6498,7 @@ void stmmac_disable_rx_queue(struct stmm
+@@ -6390,7 +6506,7 @@ void stmmac_disable_rx_queue(struct stmm
spin_unlock_irqrestore(&ch->lock, flags);
stmmac_stop_rx_dma(priv, queue);
}
void stmmac_enable_rx_queue(struct stmmac_priv *priv, u32 queue)
-@@ -6393,21 +6509,21 @@ void stmmac_enable_rx_queue(struct stmma
+@@ -6401,21 +6517,21 @@ void stmmac_enable_rx_queue(struct stmma
u32 buf_size;
int ret;
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
rx_q->dma_rx_phy, rx_q->queue_index);
-@@ -6445,7 +6561,7 @@ void stmmac_disable_tx_queue(struct stmm
+@@ -6453,7 +6569,7 @@ void stmmac_disable_tx_queue(struct stmm
spin_unlock_irqrestore(&ch->lock, flags);
stmmac_stop_tx_dma(priv, queue);
}
void stmmac_enable_tx_queue(struct stmmac_priv *priv, u32 queue)
-@@ -6455,21 +6571,21 @@ void stmmac_enable_tx_queue(struct stmma
+@@ -6463,21 +6579,21 @@ void stmmac_enable_tx_queue(struct stmma
unsigned long flags;
int ret;
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
tx_q->dma_tx_phy, tx_q->queue_index);
-@@ -6509,7 +6625,7 @@ void stmmac_xdp_release(struct net_devic
+@@ -6517,7 +6633,7 @@ void stmmac_xdp_release(struct net_devic
stmmac_stop_all_dma(priv);
/* Release and free the Rx/Tx resources */
/* Disable the MAC Rx/Tx */
stmmac_mac_set(priv, priv->ioaddr, false);
-@@ -6534,14 +6650,14 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6542,14 +6658,14 @@ int stmmac_xdp_open(struct net_device *d
u32 chan;
int ret;
if (ret < 0) {
netdev_err(dev, "%s: DMA descriptors initialization failed\n",
__func__);
-@@ -6623,7 +6739,7 @@ irq_error:
+@@ -6631,7 +6747,7 @@ irq_error:
stmmac_hw_teardown(dev);
init_error:
dma_desc_error:
return ret;
}
-@@ -7482,7 +7598,7 @@ int stmmac_resume(struct device *dev)
+@@ -7490,7 +7606,7 @@ int stmmac_resume(struct device *dev)
stmmac_reset_queues_param(priv);
stmmac_free_tx_skbufs(priv);