if (unlikely(ret < 0)) {
netdev_err(priv->dev,
"%s: alloc tx-%d MSI %d (error: %d)\n",
-@@ -3732,21 +3732,21 @@ static int stmmac_open(struct net_device
+@@ -3733,21 +3733,21 @@ static int stmmac_open(struct net_device
bfsize = 0;
if (bfsize < BUF_SIZE_16KiB)
int tbs_en = priv->plat->tx_queues_cfg[chan].tbs_en;
/* Setup per-TXQ tbs flag before TX descriptor alloc */
-@@ -3804,7 +3804,7 @@ irq_error:
+@@ -3805,7 +3805,7 @@ irq_error:
phylink_stop(priv->phylink);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
stmmac_hw_teardown(dev);
init_error:
-@@ -3848,7 +3848,7 @@ static int stmmac_release(struct net_dev
+@@ -3849,7 +3849,7 @@ static int stmmac_release(struct net_dev
stmmac_disable_all_queues(priv);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
netif_tx_disable(dev);
-@@ -3912,7 +3912,7 @@ static bool stmmac_vlan_insert(struct st
+@@ -3913,7 +3913,7 @@ static bool stmmac_vlan_insert(struct st
return false;
stmmac_set_tx_owner(priv, p);
return true;
}
-@@ -3930,7 +3930,7 @@ static bool stmmac_vlan_insert(struct st
+@@ -3931,7 +3931,7 @@ static bool stmmac_vlan_insert(struct st
static void stmmac_tso_allocator(struct stmmac_priv *priv, dma_addr_t des,
int total_len, bool last_segment, u32 queue)
{
struct dma_desc *desc;
u32 buff_size;
int tmp_len;
-@@ -3941,7 +3941,7 @@ static void stmmac_tso_allocator(struct
+@@ -3942,7 +3942,7 @@ static void stmmac_tso_allocator(struct
dma_addr_t curr_addr;
tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx,
WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]);
if (tx_q->tbs & STMMAC_TBS_AVAIL)
-@@ -3969,7 +3969,7 @@ static void stmmac_tso_allocator(struct
+@@ -3970,7 +3970,7 @@ static void stmmac_tso_allocator(struct
static void stmmac_flush_tx_descriptors(struct stmmac_priv *priv, int queue)
{
int desc_size;
if (likely(priv->extend_desc))
-@@ -4031,7 +4031,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
+@@ -4032,7 +4032,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
dma_addr_t des;
int i;
first_tx = tx_q->cur_tx;
/* Compute header lengths */
-@@ -4071,7 +4071,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
+@@ -4072,7 +4072,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
stmmac_set_mss(priv, mss_desc, mss);
tx_q->mss = mss;
tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx,
WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]);
}
-@@ -4195,7 +4195,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
+@@ -4196,7 +4196,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
* ndo_start_xmit will fill this descriptor the next time it's
* called and stmmac_tx_clean may clean up to this descriptor.
*/
if (unlikely(stmmac_tx_avail(priv, queue) <= (MAX_SKB_FRAGS + 1))) {
netif_dbg(priv, hw, priv->dev, "%s: stop transmitted packets\n",
-@@ -4283,7 +4283,7 @@ static netdev_tx_t stmmac_xmit(struct sk
+@@ -4284,7 +4284,7 @@ static netdev_tx_t stmmac_xmit(struct sk
int entry, first_tx;
dma_addr_t des;
first_tx = tx_q->cur_tx;
if (priv->tx_path_in_lpi_mode && priv->eee_sw_timer_en)
-@@ -4346,7 +4346,7 @@ static netdev_tx_t stmmac_xmit(struct sk
+@@ -4347,7 +4347,7 @@ static netdev_tx_t stmmac_xmit(struct sk
int len = skb_frag_size(frag);
bool last_segment = (i == (nfrags - 1));
WARN_ON(tx_q->tx_skbuff[entry]);
if (likely(priv->extend_desc))
-@@ -4417,7 +4417,7 @@ static netdev_tx_t stmmac_xmit(struct sk
+@@ -4418,7 +4418,7 @@ static netdev_tx_t stmmac_xmit(struct sk
* ndo_start_xmit will fill this descriptor the next time it's
* called and stmmac_tx_clean may clean up to this descriptor.
*/
tx_q->cur_tx = entry;
if (netif_msg_pktdata(priv)) {
-@@ -4529,7 +4529,7 @@ static void stmmac_rx_vlan(struct net_de
+@@ -4530,7 +4530,7 @@ static void stmmac_rx_vlan(struct net_de
*/
static inline void stmmac_rx_refill(struct stmmac_priv *priv, u32 queue)
{
int dirty = stmmac_rx_dirty(priv, queue);
unsigned int entry = rx_q->dirty_rx;
-@@ -4579,7 +4579,7 @@ static inline void stmmac_rx_refill(stru
+@@ -4580,7 +4580,7 @@ static inline void stmmac_rx_refill(stru
dma_wmb();
stmmac_set_rx_owner(priv, p, use_rx_wd);
}
rx_q->dirty_rx = entry;
rx_q->rx_tail_addr = rx_q->dma_rx_phy +
-@@ -4607,12 +4607,12 @@ static unsigned int stmmac_rx_buf1_len(s
+@@ -4608,12 +4608,12 @@ static unsigned int stmmac_rx_buf1_len(s
/* First descriptor, not last descriptor and not split header */
if (status & rx_not_ls)
}
static unsigned int stmmac_rx_buf2_len(struct stmmac_priv *priv,
-@@ -4628,7 +4628,7 @@ static unsigned int stmmac_rx_buf2_len(s
+@@ -4629,7 +4629,7 @@ static unsigned int stmmac_rx_buf2_len(s
/* Not last descriptor */
if (status & rx_not_ls)
plen = stmmac_get_rx_frame_len(priv, p, coe);
-@@ -4639,7 +4639,7 @@ static unsigned int stmmac_rx_buf2_len(s
+@@ -4640,7 +4640,7 @@ static unsigned int stmmac_rx_buf2_len(s
static int stmmac_xdp_xmit_xdpf(struct stmmac_priv *priv, int queue,
struct xdp_frame *xdpf, bool dma_map)
{
unsigned int entry = tx_q->cur_tx;
struct dma_desc *tx_desc;
dma_addr_t dma_addr;
-@@ -4702,7 +4702,7 @@ static int stmmac_xdp_xmit_xdpf(struct s
+@@ -4703,7 +4703,7 @@ static int stmmac_xdp_xmit_xdpf(struct s
stmmac_enable_dma_transmission(priv, priv->ioaddr);
tx_q->cur_tx = entry;
return STMMAC_XDP_TX;
-@@ -4876,7 +4876,7 @@ static void stmmac_dispatch_skb_zc(struc
+@@ -4877,7 +4877,7 @@ static void stmmac_dispatch_skb_zc(struc
static bool stmmac_rx_refill_zc(struct stmmac_priv *priv, u32 queue, u32 budget)
{
unsigned int entry = rx_q->dirty_rx;
struct dma_desc *rx_desc = NULL;
bool ret = true;
-@@ -4919,7 +4919,7 @@ static bool stmmac_rx_refill_zc(struct s
+@@ -4920,7 +4920,7 @@ static bool stmmac_rx_refill_zc(struct s
dma_wmb();
stmmac_set_rx_owner(priv, rx_desc, use_rx_wd);
}
if (rx_desc) {
-@@ -4934,7 +4934,7 @@ static bool stmmac_rx_refill_zc(struct s
+@@ -4935,7 +4935,7 @@ static bool stmmac_rx_refill_zc(struct s
static int stmmac_rx_zc(struct stmmac_priv *priv, int limit, u32 queue)
{
unsigned int count = 0, error = 0, len = 0;
int dirty = stmmac_rx_dirty(priv, queue);
unsigned int next_entry = rx_q->cur_rx;
-@@ -4956,7 +4956,7 @@ static int stmmac_rx_zc(struct stmmac_pr
+@@ -4957,7 +4957,7 @@ static int stmmac_rx_zc(struct stmmac_pr
desc_size = sizeof(struct dma_desc);
}
rx_q->dma_rx_phy, desc_size);
}
while (count < limit) {
-@@ -5003,7 +5003,7 @@ read_again:
+@@ -5004,7 +5004,7 @@ read_again:
/* Prefetch the next RX descriptor */
rx_q->cur_rx = STMMAC_GET_ENTRY(rx_q->cur_rx,
next_entry = rx_q->cur_rx;
if (priv->extend_desc)
-@@ -5116,7 +5116,7 @@ read_again:
+@@ -5117,7 +5117,7 @@ read_again:
*/
static int stmmac_rx(struct stmmac_priv *priv, int limit, u32 queue)
{
struct stmmac_channel *ch = &priv->channel[queue];
unsigned int count = 0, error = 0, len = 0;
int status = 0, coe = priv->hw->rx_csum;
-@@ -5129,7 +5129,7 @@ static int stmmac_rx(struct stmmac_priv
+@@ -5130,7 +5130,7 @@ static int stmmac_rx(struct stmmac_priv
int buf_sz;
dma_dir = page_pool_get_dma_dir(rx_q->page_pool);
if (netif_msg_rx_status(priv)) {
void *rx_head;
-@@ -5143,7 +5143,7 @@ static int stmmac_rx(struct stmmac_priv
+@@ -5144,7 +5144,7 @@ static int stmmac_rx(struct stmmac_priv
desc_size = sizeof(struct dma_desc);
}
rx_q->dma_rx_phy, desc_size);
}
while (count < limit) {
-@@ -5187,7 +5187,7 @@ read_again:
+@@ -5188,7 +5188,7 @@ read_again:
break;
rx_q->cur_rx = STMMAC_GET_ENTRY(rx_q->cur_rx,
next_entry = rx_q->cur_rx;
if (priv->extend_desc)
-@@ -5314,7 +5314,7 @@ read_again:
+@@ -5315,7 +5315,7 @@ read_again:
buf1_len, dma_dir);
skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags,
buf->page, buf->page_offset, buf1_len,
/* Data payload appended into SKB */
page_pool_release_page(rx_q->page_pool, buf->page);
-@@ -5326,7 +5326,7 @@ read_again:
+@@ -5327,7 +5327,7 @@ read_again:
buf2_len, dma_dir);
skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags,
buf->sec_page, 0, buf2_len,
/* Data payload appended into SKB */
page_pool_release_page(rx_q->page_pool, buf->sec_page);
-@@ -5759,11 +5759,13 @@ static irqreturn_t stmmac_safety_interru
+@@ -5760,11 +5760,13 @@ static irqreturn_t stmmac_safety_interru
static irqreturn_t stmmac_msi_intr_tx(int irq, void *data)
{
struct stmmac_tx_queue *tx_q = (struct stmmac_tx_queue *)data;
/* Check if adapter is up */
if (test_bit(STMMAC_DOWN, &priv->state))
-@@ -5798,10 +5800,12 @@ static irqreturn_t stmmac_msi_intr_tx(in
+@@ -5799,10 +5801,12 @@ static irqreturn_t stmmac_msi_intr_tx(in
static irqreturn_t stmmac_msi_intr_rx(int irq, void *data)
{
struct stmmac_rx_queue *rx_q = (struct stmmac_rx_queue *)data;
/* Check if adapter is up */
if (test_bit(STMMAC_DOWN, &priv->state))
-@@ -5827,10 +5831,10 @@ static void stmmac_poll_controller(struc
+@@ -5828,10 +5832,10 @@ static void stmmac_poll_controller(struc
if (priv->plat->multi_msi_en) {
for (i = 0; i < priv->plat->rx_queues_to_use; i++)
} else {
disable_irq(dev->irq);
stmmac_interrupt(dev->irq, dev);
-@@ -6011,34 +6015,34 @@ static int stmmac_rings_status_show(stru
+@@ -6012,34 +6016,34 @@ static int stmmac_rings_status_show(stru
return 0;
for (queue = 0; queue < rx_count; queue++) {
}
}
-@@ -6385,7 +6389,7 @@ void stmmac_disable_rx_queue(struct stmm
+@@ -6386,7 +6390,7 @@ void stmmac_disable_rx_queue(struct stmm
void stmmac_enable_rx_queue(struct stmmac_priv *priv, u32 queue)
{
struct stmmac_channel *ch = &priv->channel[queue];
unsigned long flags;
u32 buf_size;
-@@ -6422,7 +6426,7 @@ void stmmac_enable_rx_queue(struct stmma
+@@ -6423,7 +6427,7 @@ void stmmac_enable_rx_queue(struct stmma
rx_q->queue_index);
} else {
stmmac_set_dma_bfsize(priv, priv->ioaddr,
rx_q->queue_index);
}
-@@ -6448,7 +6452,7 @@ void stmmac_disable_tx_queue(struct stmm
+@@ -6449,7 +6453,7 @@ void stmmac_disable_tx_queue(struct stmm
void stmmac_enable_tx_queue(struct stmmac_priv *priv, u32 queue)
{
struct stmmac_channel *ch = &priv->channel[queue];
unsigned long flags;
int ret;
-@@ -6498,7 +6502,7 @@ void stmmac_xdp_release(struct net_devic
+@@ -6499,7 +6503,7 @@ void stmmac_xdp_release(struct net_devic
stmmac_disable_all_queues(priv);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
/* Free the IRQ lines */
stmmac_free_irq(dev, REQ_IRQ_ERR_ALL, 0);
-@@ -6557,7 +6561,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6558,7 +6562,7 @@ int stmmac_xdp_open(struct net_device *d
/* DMA RX Channel Configuration */
for (chan = 0; chan < rx_cnt; chan++) {
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
rx_q->dma_rx_phy, chan);
-@@ -6575,7 +6579,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6576,7 +6580,7 @@ int stmmac_xdp_open(struct net_device *d
rx_q->queue_index);
} else {
stmmac_set_dma_bfsize(priv, priv->ioaddr,
rx_q->queue_index);
}
-@@ -6584,7 +6588,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6585,7 +6589,7 @@ int stmmac_xdp_open(struct net_device *d
/* DMA TX Channel Configuration */
for (chan = 0; chan < tx_cnt; chan++) {
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
tx_q->dma_tx_phy, chan);
-@@ -6617,7 +6621,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6618,7 +6622,7 @@ int stmmac_xdp_open(struct net_device *d
irq_error:
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
stmmac_hw_teardown(dev);
init_error:
-@@ -6644,8 +6648,8 @@ int stmmac_xsk_wakeup(struct net_device
+@@ -6645,8 +6649,8 @@ int stmmac_xsk_wakeup(struct net_device
queue >= priv->plat->tx_queues_to_use)
return -EINVAL;
ch = &priv->channel[queue];
if (!rx_q->xsk_pool && !tx_q->xsk_pool)
-@@ -6905,8 +6909,8 @@ int stmmac_reinit_ringparam(struct net_d
+@@ -6906,8 +6910,8 @@ int stmmac_reinit_ringparam(struct net_d
if (netif_running(dev))
stmmac_release(dev);
if (netif_running(dev))
ret = stmmac_open(dev);
-@@ -7344,7 +7348,7 @@ int stmmac_suspend(struct device *dev)
+@@ -7345,7 +7349,7 @@ int stmmac_suspend(struct device *dev)
stmmac_disable_all_queues(priv);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
if (priv->eee_enabled) {
priv->tx_path_in_lpi_mode = false;
-@@ -7396,7 +7400,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend);
+@@ -7397,7 +7401,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend);
static void stmmac_reset_rx_queue(struct stmmac_priv *priv, u32 queue)
{
rx_q->cur_rx = 0;
rx_q->dirty_rx = 0;
-@@ -7404,7 +7408,7 @@ static void stmmac_reset_rx_queue(struct
+@@ -7405,7 +7409,7 @@ static void stmmac_reset_rx_queue(struct
static void stmmac_reset_tx_queue(struct stmmac_priv *priv, u32 queue)
{
stmmac_reset_tx_queue(priv, chan);
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
tx_q->dma_tx_phy, chan);
-@@ -3688,19 +3751,93 @@ static int stmmac_request_irq(struct net
+@@ -3689,19 +3752,93 @@ static int stmmac_request_irq(struct net
}
/**
u32 chan;
int ret;
-@@ -3727,45 +3864,10 @@ static int stmmac_open(struct net_device
+@@ -3728,45 +3865,10 @@ static int stmmac_open(struct net_device
memset(&priv->xstats, 0, sizeof(struct stmmac_extra_stats));
priv->xstats.threshold = tc;
if (priv->plat->serdes_powerup) {
ret = priv->plat->serdes_powerup(dev, priv->plat->bsp_priv);
-@@ -3808,14 +3910,28 @@ irq_error:
+@@ -3809,14 +3911,28 @@ irq_error:
stmmac_hw_teardown(dev);
init_error:
static void stmmac_fpe_stop_wq(struct stmmac_priv *priv)
{
set_bit(__FPE_REMOVING, &priv->fpe_task_state);
-@@ -3864,7 +3980,7 @@ static int stmmac_release(struct net_dev
+@@ -3865,7 +3981,7 @@ static int stmmac_release(struct net_dev
stmmac_stop_all_dma(priv);
/* Release and free the Rx/Tx resources */
/* Disable the MAC Rx/Tx */
stmmac_mac_set(priv, priv->ioaddr, false);
-@@ -6384,7 +6500,7 @@ void stmmac_disable_rx_queue(struct stmm
+@@ -6385,7 +6501,7 @@ void stmmac_disable_rx_queue(struct stmm
spin_unlock_irqrestore(&ch->lock, flags);
stmmac_stop_rx_dma(priv, queue);
}
void stmmac_enable_rx_queue(struct stmmac_priv *priv, u32 queue)
-@@ -6395,21 +6511,21 @@ void stmmac_enable_rx_queue(struct stmma
+@@ -6396,21 +6512,21 @@ void stmmac_enable_rx_queue(struct stmma
u32 buf_size;
int ret;
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
rx_q->dma_rx_phy, rx_q->queue_index);
-@@ -6447,7 +6563,7 @@ void stmmac_disable_tx_queue(struct stmm
+@@ -6448,7 +6564,7 @@ void stmmac_disable_tx_queue(struct stmm
spin_unlock_irqrestore(&ch->lock, flags);
stmmac_stop_tx_dma(priv, queue);
}
void stmmac_enable_tx_queue(struct stmmac_priv *priv, u32 queue)
-@@ -6457,21 +6573,21 @@ void stmmac_enable_tx_queue(struct stmma
+@@ -6458,21 +6574,21 @@ void stmmac_enable_tx_queue(struct stmma
unsigned long flags;
int ret;
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
tx_q->dma_tx_phy, tx_q->queue_index);
-@@ -6511,7 +6627,7 @@ void stmmac_xdp_release(struct net_devic
+@@ -6512,7 +6628,7 @@ void stmmac_xdp_release(struct net_devic
stmmac_stop_all_dma(priv);
/* Release and free the Rx/Tx resources */
/* Disable the MAC Rx/Tx */
stmmac_mac_set(priv, priv->ioaddr, false);
-@@ -6536,14 +6652,14 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6537,14 +6653,14 @@ int stmmac_xdp_open(struct net_device *d
u32 chan;
int ret;
if (ret < 0) {
netdev_err(dev, "%s: DMA descriptors initialization failed\n",
__func__);
-@@ -6625,7 +6741,7 @@ irq_error:
+@@ -6626,7 +6742,7 @@ irq_error:
stmmac_hw_teardown(dev);
init_error:
dma_desc_error:
return ret;
}
-@@ -7491,7 +7607,7 @@ int stmmac_resume(struct device *dev)
+@@ -7492,7 +7608,7 @@ int stmmac_resume(struct device *dev)
stmmac_reset_queues_param(priv);
stmmac_free_tx_skbufs(priv);