if (tx_q->dirty_tx != tx_q->cur_tx)
return -EBUSY; /* still unfinished work */
-@@ -1307,7 +1307,7 @@ static void stmmac_display_rx_rings(stru
+@@ -1310,7 +1310,7 @@ static void stmmac_display_rx_rings(stru
/* Display RX rings */
for (queue = 0; queue < rx_cnt; queue++) {
pr_info("\tRX Queue %u rings\n", queue);
-@@ -1320,7 +1320,7 @@ static void stmmac_display_rx_rings(stru
+@@ -1323,7 +1323,7 @@ static void stmmac_display_rx_rings(stru
}
/* Display RX ring */
rx_q->dma_rx_phy, desc_size);
}
}
-@@ -1334,7 +1334,7 @@ static void stmmac_display_tx_rings(stru
+@@ -1337,7 +1337,7 @@ static void stmmac_display_tx_rings(stru
/* Display TX rings */
for (queue = 0; queue < tx_cnt; queue++) {
pr_info("\tTX Queue %d rings\n", queue);
-@@ -1349,7 +1349,7 @@ static void stmmac_display_tx_rings(stru
+@@ -1352,7 +1352,7 @@ static void stmmac_display_tx_rings(stru
desc_size = sizeof(struct dma_desc);
}
tx_q->dma_tx_phy, desc_size);
}
}
-@@ -1390,21 +1390,21 @@ static int stmmac_set_bfsize(int mtu, in
+@@ -1393,21 +1393,21 @@ static int stmmac_set_bfsize(int mtu, in
*/
static void stmmac_clear_rx_descriptors(struct stmmac_priv *priv, u32 queue)
{
}
/**
-@@ -1416,12 +1416,12 @@ static void stmmac_clear_rx_descriptors(
+@@ -1419,12 +1419,12 @@ static void stmmac_clear_rx_descriptors(
*/
static void stmmac_clear_tx_descriptors(struct stmmac_priv *priv, u32 queue)
{
struct dma_desc *p;
if (priv->extend_desc)
-@@ -1469,7 +1469,7 @@ static void stmmac_clear_descriptors(str
+@@ -1472,7 +1472,7 @@ static void stmmac_clear_descriptors(str
static int stmmac_init_rx_buffers(struct stmmac_priv *priv, struct dma_desc *p,
int i, gfp_t flags, u32 queue)
{
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
if (!buf->page) {
-@@ -1494,7 +1494,7 @@ static int stmmac_init_rx_buffers(struct
+@@ -1497,7 +1497,7 @@ static int stmmac_init_rx_buffers(struct
buf->addr = page_pool_get_dma_addr(buf->page) + buf->page_offset;
stmmac_set_desc_addr(priv, p, buf->addr);
stmmac_init_desc3(priv, p);
return 0;
-@@ -1508,7 +1508,7 @@ static int stmmac_init_rx_buffers(struct
+@@ -1511,7 +1511,7 @@ static int stmmac_init_rx_buffers(struct
*/
static void stmmac_free_rx_buffer(struct stmmac_priv *priv, u32 queue, int i)
{
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
if (buf->page)
-@@ -1528,7 +1528,7 @@ static void stmmac_free_rx_buffer(struct
+@@ -1531,7 +1531,7 @@ static void stmmac_free_rx_buffer(struct
*/
static void stmmac_free_tx_buffer(struct stmmac_priv *priv, u32 queue, int i)
{
if (tx_q->tx_skbuff_dma[i].buf &&
tx_q->tx_skbuff_dma[i].buf_type != STMMAC_TXBUF_T_XDP_TX) {
-@@ -1573,17 +1573,17 @@ static void dma_free_rx_skbufs(struct st
+@@ -1576,17 +1576,17 @@ static void dma_free_rx_skbufs(struct st
{
int i;
struct dma_desc *p;
int ret;
-@@ -1610,10 +1610,10 @@ static int stmmac_alloc_rx_buffers(struc
+@@ -1613,10 +1613,10 @@ static int stmmac_alloc_rx_buffers(struc
*/
static void dma_free_rx_xskbufs(struct stmmac_priv *priv, u32 queue)
{
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
if (!buf->xdp)
-@@ -1626,10 +1626,10 @@ static void dma_free_rx_xskbufs(struct s
+@@ -1629,10 +1629,10 @@ static void dma_free_rx_xskbufs(struct s
static int stmmac_alloc_rx_buffers_zc(struct stmmac_priv *priv, u32 queue)
{
struct stmmac_rx_buffer *buf;
dma_addr_t dma_addr;
struct dma_desc *p;
-@@ -1672,7 +1672,7 @@ static struct xsk_buff_pool *stmmac_get_
+@@ -1675,7 +1675,7 @@ static struct xsk_buff_pool *stmmac_get_
*/
static int __init_dma_rx_desc_rings(struct stmmac_priv *priv, u32 queue, gfp_t flags)
{
int ret;
netif_dbg(priv, probe, priv->dev,
-@@ -1718,11 +1718,11 @@ static int __init_dma_rx_desc_rings(stru
+@@ -1721,11 +1721,11 @@ static int __init_dma_rx_desc_rings(stru
if (priv->extend_desc)
stmmac_mode_init(priv, rx_q->dma_erx,
rx_q->dma_rx_phy,
}
return 0;
-@@ -1749,7 +1749,7 @@ static int init_dma_rx_desc_rings(struct
+@@ -1752,7 +1752,7 @@ static int init_dma_rx_desc_rings(struct
err_init_rx_buffers:
while (queue >= 0) {
if (rx_q->xsk_pool)
dma_free_rx_xskbufs(priv, queue);
-@@ -1778,7 +1778,7 @@ err_init_rx_buffers:
+@@ -1781,7 +1781,7 @@ err_init_rx_buffers:
*/
static int __init_dma_tx_desc_rings(struct stmmac_priv *priv, u32 queue)
{
int i;
netif_dbg(priv, probe, priv->dev,
-@@ -1790,16 +1790,16 @@ static int __init_dma_tx_desc_rings(stru
+@@ -1793,16 +1793,16 @@ static int __init_dma_tx_desc_rings(stru
if (priv->extend_desc)
stmmac_mode_init(priv, tx_q->dma_etx,
tx_q->dma_tx_phy,
struct dma_desc *p;
if (priv->extend_desc)
-@@ -1869,12 +1869,12 @@ static int init_dma_desc_rings(struct ne
+@@ -1872,12 +1872,12 @@ static int init_dma_desc_rings(struct ne
*/
static void dma_free_tx_skbufs(struct stmmac_priv *priv, u32 queue)
{
stmmac_free_tx_buffer(priv, queue, i);
if (tx_q->xsk_pool && tx_q->xsk_frames_done) {
-@@ -1904,7 +1904,7 @@ static void stmmac_free_tx_skbufs(struct
+@@ -1907,7 +1907,7 @@ static void stmmac_free_tx_skbufs(struct
*/
static void __free_dma_rx_desc_resources(struct stmmac_priv *priv, u32 queue)
{
/* Release the DMA RX socket buffers */
if (rx_q->xsk_pool)
-@@ -1917,11 +1917,11 @@ static void __free_dma_rx_desc_resources
+@@ -1920,11 +1920,11 @@ static void __free_dma_rx_desc_resources
/* Free DMA regions of consistent memory previously allocated */
if (!priv->extend_desc)
sizeof(struct dma_extended_desc),
rx_q->dma_erx, rx_q->dma_rx_phy);
-@@ -1950,7 +1950,7 @@ static void free_dma_rx_desc_resources(s
+@@ -1953,7 +1953,7 @@ static void free_dma_rx_desc_resources(s
*/
static void __free_dma_tx_desc_resources(struct stmmac_priv *priv, u32 queue)
{
size_t size;
void *addr;
-@@ -1968,7 +1968,7 @@ static void __free_dma_tx_desc_resources
+@@ -1971,7 +1971,7 @@ static void __free_dma_tx_desc_resources
addr = tx_q->dma_tx;
}
dma_free_coherent(priv->device, size, addr, tx_q->dma_tx_phy);
-@@ -1997,7 +1997,7 @@ static void free_dma_tx_desc_resources(s
+@@ -2000,7 +2000,7 @@ static void free_dma_tx_desc_resources(s
*/
static int __alloc_dma_rx_desc_resources(struct stmmac_priv *priv, u32 queue)
{
struct stmmac_channel *ch = &priv->channel[queue];
bool xdp_prog = stmmac_xdp_is_enabled(priv);
struct page_pool_params pp_params = { 0 };
-@@ -2009,8 +2009,8 @@ static int __alloc_dma_rx_desc_resources
+@@ -2012,8 +2012,8 @@ static int __alloc_dma_rx_desc_resources
rx_q->priv_data = priv;
pp_params.flags = PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV;
pp_params.order = ilog2(num_pages);
pp_params.nid = dev_to_node(priv->device);
pp_params.dev = priv->device;
-@@ -2025,7 +2025,7 @@ static int __alloc_dma_rx_desc_resources
+@@ -2028,7 +2028,7 @@ static int __alloc_dma_rx_desc_resources
return ret;
}
sizeof(*rx_q->buf_pool),
GFP_KERNEL);
if (!rx_q->buf_pool)
-@@ -2033,7 +2033,7 @@ static int __alloc_dma_rx_desc_resources
+@@ -2036,7 +2036,7 @@ static int __alloc_dma_rx_desc_resources
if (priv->extend_desc) {
rx_q->dma_erx = dma_alloc_coherent(priv->device,
sizeof(struct dma_extended_desc),
&rx_q->dma_rx_phy,
GFP_KERNEL);
-@@ -2042,7 +2042,7 @@ static int __alloc_dma_rx_desc_resources
+@@ -2045,7 +2045,7 @@ static int __alloc_dma_rx_desc_resources
} else {
rx_q->dma_rx = dma_alloc_coherent(priv->device,
sizeof(struct dma_desc),
&rx_q->dma_rx_phy,
GFP_KERNEL);
-@@ -2099,20 +2099,20 @@ err_dma:
+@@ -2102,20 +2102,20 @@ err_dma:
*/
static int __alloc_dma_tx_desc_resources(struct stmmac_priv *priv, u32 queue)
{
sizeof(struct sk_buff *),
GFP_KERNEL);
if (!tx_q->tx_skbuff)
-@@ -2125,7 +2125,7 @@ static int __alloc_dma_tx_desc_resources
+@@ -2128,7 +2128,7 @@ static int __alloc_dma_tx_desc_resources
else
size = sizeof(struct dma_desc);
addr = dma_alloc_coherent(priv->device, size,
&tx_q->dma_tx_phy, GFP_KERNEL);
-@@ -2369,7 +2369,7 @@ static void stmmac_dma_operation_mode(st
+@@ -2372,7 +2372,7 @@ static void stmmac_dma_operation_mode(st
/* configure all channels */
for (chan = 0; chan < rx_channels_count; chan++) {
u32 buf_size;
qmode = priv->plat->rx_queues_cfg[chan].mode_to_use;
-@@ -2384,7 +2384,7 @@ static void stmmac_dma_operation_mode(st
+@@ -2387,7 +2387,7 @@ static void stmmac_dma_operation_mode(st
chan);
} else {
stmmac_set_dma_bfsize(priv, priv->ioaddr,
chan);
}
}
-@@ -2400,7 +2400,7 @@ static void stmmac_dma_operation_mode(st
+@@ -2403,7 +2403,7 @@ static void stmmac_dma_operation_mode(st
static bool stmmac_xdp_xmit_zc(struct stmmac_priv *priv, u32 queue, u32 budget)
{
struct netdev_queue *nq = netdev_get_tx_queue(priv->dev, queue);
struct xsk_buff_pool *pool = tx_q->xsk_pool;
unsigned int entry = tx_q->cur_tx;
struct dma_desc *tx_desc = NULL;
-@@ -2475,7 +2475,7 @@ static bool stmmac_xdp_xmit_zc(struct st
+@@ -2478,7 +2478,7 @@ static bool stmmac_xdp_xmit_zc(struct st
stmmac_enable_dma_transmission(priv, priv->ioaddr);
entry = tx_q->cur_tx;
}
-@@ -2501,7 +2501,7 @@ static bool stmmac_xdp_xmit_zc(struct st
+@@ -2504,7 +2504,7 @@ static bool stmmac_xdp_xmit_zc(struct st
*/
static int stmmac_tx_clean(struct stmmac_priv *priv, int budget, u32 queue)
{
unsigned int bytes_compl = 0, pkts_compl = 0;
unsigned int entry, xmits = 0, count = 0;
-@@ -2514,7 +2514,7 @@ static int stmmac_tx_clean(struct stmmac
+@@ -2517,7 +2517,7 @@ static int stmmac_tx_clean(struct stmmac
entry = tx_q->dirty_tx;
/* Try to clean all TX complete frame in 1 shot */
struct xdp_frame *xdpf;
struct sk_buff *skb;
struct dma_desc *p;
-@@ -2614,7 +2614,7 @@ static int stmmac_tx_clean(struct stmmac
+@@ -2617,7 +2617,7 @@ static int stmmac_tx_clean(struct stmmac
stmmac_release_tx_desc(priv, p, priv->mode);
}
tx_q->dirty_tx = entry;
-@@ -2679,7 +2679,7 @@ static int stmmac_tx_clean(struct stmmac
+@@ -2682,7 +2682,7 @@ static int stmmac_tx_clean(struct stmmac
*/
static void stmmac_tx_err(struct stmmac_priv *priv, u32 chan)
{
netif_tx_stop_queue(netdev_get_tx_queue(priv->dev, chan));
-@@ -2746,8 +2746,8 @@ static int stmmac_napi_check(struct stmm
+@@ -2749,8 +2749,8 @@ static int stmmac_napi_check(struct stmm
{
int status = stmmac_dma_interrupt_status(priv, priv->ioaddr,
&priv->xstats, chan, dir);
struct stmmac_channel *ch = &priv->channel[chan];
struct napi_struct *rx_napi;
struct napi_struct *tx_napi;
-@@ -2923,7 +2923,7 @@ static int stmmac_init_dma_engine(struct
+@@ -2926,7 +2926,7 @@ static int stmmac_init_dma_engine(struct
/* DMA RX Channel Configuration */
for (chan = 0; chan < rx_channels_count; chan++) {
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
rx_q->dma_rx_phy, chan);
-@@ -2937,7 +2937,7 @@ static int stmmac_init_dma_engine(struct
+@@ -2940,7 +2940,7 @@ static int stmmac_init_dma_engine(struct
/* DMA TX Channel Configuration */
for (chan = 0; chan < tx_channels_count; chan++) {
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
tx_q->dma_tx_phy, chan);
-@@ -2952,7 +2952,7 @@ static int stmmac_init_dma_engine(struct
+@@ -2955,7 +2955,7 @@ static int stmmac_init_dma_engine(struct
static void stmmac_tx_timer_arm(struct stmmac_priv *priv, u32 queue)
{
hrtimer_start(&tx_q->txtimer,
STMMAC_COAL_TIMER(priv->tx_coal_timer[queue]),
-@@ -3002,7 +3002,7 @@ static void stmmac_init_coalesce(struct
+@@ -3005,7 +3005,7 @@ static void stmmac_init_coalesce(struct
u32 chan;
for (chan = 0; chan < tx_channel_count; chan++) {
priv->tx_coal_frames[chan] = STMMAC_TX_FRAMES;
priv->tx_coal_timer[chan] = STMMAC_COAL_TX_TIMER;
-@@ -3024,12 +3024,12 @@ static void stmmac_set_rings_length(stru
+@@ -3027,12 +3027,12 @@ static void stmmac_set_rings_length(stru
/* set TX ring length */
for (chan = 0; chan < tx_channels_count; chan++)
stmmac_set_tx_ring_len(priv, priv->ioaddr,
}
/**
-@@ -3364,7 +3364,7 @@ static int stmmac_hw_setup(struct net_de
+@@ -3367,7 +3367,7 @@ static int stmmac_hw_setup(struct net_de
/* Enable TSO */
if (priv->tso) {
for (chan = 0; chan < tx_cnt; chan++) {
/* TSO and TBS cannot co-exist */
if (tx_q->tbs & STMMAC_TBS_AVAIL)
-@@ -3386,7 +3386,7 @@ static int stmmac_hw_setup(struct net_de
+@@ -3389,7 +3389,7 @@ static int stmmac_hw_setup(struct net_de
/* TBS */
for (chan = 0; chan < tx_cnt; chan++) {
int enable = tx_q->tbs & STMMAC_TBS_AVAIL;
stmmac_enable_tbs(priv, priv->ioaddr, enable, chan);
-@@ -3430,7 +3430,7 @@ static void stmmac_free_irq(struct net_d
+@@ -3433,7 +3433,7 @@ static void stmmac_free_irq(struct net_d
for (j = irq_idx - 1; j >= 0; j--) {
if (priv->tx_irq[j] > 0) {
irq_set_affinity_hint(priv->tx_irq[j], NULL);
}
}
irq_idx = priv->plat->rx_queues_to_use;
-@@ -3439,7 +3439,7 @@ static void stmmac_free_irq(struct net_d
+@@ -3442,7 +3442,7 @@ static void stmmac_free_irq(struct net_d
for (j = irq_idx - 1; j >= 0; j--) {
if (priv->rx_irq[j] > 0) {
irq_set_affinity_hint(priv->rx_irq[j], NULL);
}
}
-@@ -3572,7 +3572,7 @@ static int stmmac_request_irq_multi_msi(
+@@ -3575,7 +3575,7 @@ static int stmmac_request_irq_multi_msi(
sprintf(int_name, "%s:%s-%d", dev->name, "rx", i);
ret = request_irq(priv->rx_irq[i],
stmmac_msi_intr_rx,
if (unlikely(ret < 0)) {
netdev_err(priv->dev,
"%s: alloc rx-%d MSI %d (error: %d)\n",
-@@ -3595,7 +3595,7 @@ static int stmmac_request_irq_multi_msi(
+@@ -3598,7 +3598,7 @@ static int stmmac_request_irq_multi_msi(
sprintf(int_name, "%s:%s-%d", dev->name, "tx", i);
ret = request_irq(priv->tx_irq[i],
stmmac_msi_intr_tx,
if (unlikely(ret < 0)) {
netdev_err(priv->dev,
"%s: alloc tx-%d MSI %d (error: %d)\n",
-@@ -3726,21 +3726,21 @@ static int stmmac_open(struct net_device
+@@ -3729,21 +3729,21 @@ static int stmmac_open(struct net_device
bfsize = 0;
if (bfsize < BUF_SIZE_16KiB)
int tbs_en = priv->plat->tx_queues_cfg[chan].tbs_en;
/* Setup per-TXQ tbs flag before TX descriptor alloc */
-@@ -3798,7 +3798,7 @@ irq_error:
+@@ -3801,7 +3801,7 @@ irq_error:
phylink_stop(priv->phylink);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
stmmac_hw_teardown(dev);
init_error:
-@@ -3840,7 +3840,7 @@ static int stmmac_release(struct net_dev
+@@ -3843,7 +3843,7 @@ static int stmmac_release(struct net_dev
stmmac_disable_all_queues(priv);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
netif_tx_disable(dev);
-@@ -3904,7 +3904,7 @@ static bool stmmac_vlan_insert(struct st
+@@ -3907,7 +3907,7 @@ static bool stmmac_vlan_insert(struct st
return false;
stmmac_set_tx_owner(priv, p);
return true;
}
-@@ -3922,7 +3922,7 @@ static bool stmmac_vlan_insert(struct st
+@@ -3925,7 +3925,7 @@ static bool stmmac_vlan_insert(struct st
static void stmmac_tso_allocator(struct stmmac_priv *priv, dma_addr_t des,
int total_len, bool last_segment, u32 queue)
{
struct dma_desc *desc;
u32 buff_size;
int tmp_len;
-@@ -3933,7 +3933,7 @@ static void stmmac_tso_allocator(struct
+@@ -3936,7 +3936,7 @@ static void stmmac_tso_allocator(struct
dma_addr_t curr_addr;
tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx,
WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]);
if (tx_q->tbs & STMMAC_TBS_AVAIL)
-@@ -3961,7 +3961,7 @@ static void stmmac_tso_allocator(struct
+@@ -3964,7 +3964,7 @@ static void stmmac_tso_allocator(struct
static void stmmac_flush_tx_descriptors(struct stmmac_priv *priv, int queue)
{
int desc_size;
if (likely(priv->extend_desc))
-@@ -4023,7 +4023,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
+@@ -4026,7 +4026,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
dma_addr_t des;
int i;
first_tx = tx_q->cur_tx;
/* Compute header lengths */
-@@ -4063,7 +4063,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
+@@ -4066,7 +4066,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
stmmac_set_mss(priv, mss_desc, mss);
tx_q->mss = mss;
tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx,
WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]);
}
-@@ -4175,7 +4175,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
+@@ -4178,7 +4178,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
* ndo_start_xmit will fill this descriptor the next time it's
* called and stmmac_tx_clean may clean up to this descriptor.
*/
if (unlikely(stmmac_tx_avail(priv, queue) <= (MAX_SKB_FRAGS + 1))) {
netif_dbg(priv, hw, priv->dev, "%s: stop transmitted packets\n",
-@@ -4263,7 +4263,7 @@ static netdev_tx_t stmmac_xmit(struct sk
+@@ -4266,7 +4266,7 @@ static netdev_tx_t stmmac_xmit(struct sk
int entry, first_tx;
dma_addr_t des;
first_tx = tx_q->cur_tx;
if (priv->tx_path_in_lpi_mode && priv->eee_sw_timer_en)
-@@ -4326,7 +4326,7 @@ static netdev_tx_t stmmac_xmit(struct sk
+@@ -4329,7 +4329,7 @@ static netdev_tx_t stmmac_xmit(struct sk
int len = skb_frag_size(frag);
bool last_segment = (i == (nfrags - 1));
WARN_ON(tx_q->tx_skbuff[entry]);
if (likely(priv->extend_desc))
-@@ -4397,7 +4397,7 @@ static netdev_tx_t stmmac_xmit(struct sk
+@@ -4400,7 +4400,7 @@ static netdev_tx_t stmmac_xmit(struct sk
* ndo_start_xmit will fill this descriptor the next time it's
* called and stmmac_tx_clean may clean up to this descriptor.
*/
/* Data payload appended into SKB */
page_pool_release_page(rx_q->page_pool, buf->sec_page);
-@@ -5766,11 +5766,13 @@ static irqreturn_t stmmac_safety_interru
+@@ -5767,11 +5767,13 @@ static irqreturn_t stmmac_safety_interru
static irqreturn_t stmmac_msi_intr_tx(int irq, void *data)
{
struct stmmac_tx_queue *tx_q = (struct stmmac_tx_queue *)data;
if (unlikely(!data)) {
netdev_err(priv->dev, "%s: invalid dev pointer\n", __func__);
-@@ -5810,10 +5812,12 @@ static irqreturn_t stmmac_msi_intr_tx(in
+@@ -5811,10 +5813,12 @@ static irqreturn_t stmmac_msi_intr_tx(in
static irqreturn_t stmmac_msi_intr_rx(int irq, void *data)
{
struct stmmac_rx_queue *rx_q = (struct stmmac_rx_queue *)data;
if (unlikely(!data)) {
netdev_err(priv->dev, "%s: invalid dev pointer\n", __func__);
-@@ -5844,10 +5848,10 @@ static void stmmac_poll_controller(struc
+@@ -5845,10 +5849,10 @@ static void stmmac_poll_controller(struc
if (priv->plat->multi_msi_en) {
for (i = 0; i < priv->plat->rx_queues_to_use; i++)
} else {
disable_irq(dev->irq);
stmmac_interrupt(dev->irq, dev);
-@@ -6028,34 +6032,34 @@ static int stmmac_rings_status_show(stru
+@@ -6029,34 +6033,34 @@ static int stmmac_rings_status_show(stru
return 0;
for (queue = 0; queue < rx_count; queue++) {
}
}
-@@ -6396,7 +6400,7 @@ void stmmac_disable_rx_queue(struct stmm
+@@ -6403,7 +6407,7 @@ void stmmac_disable_rx_queue(struct stmm
void stmmac_enable_rx_queue(struct stmmac_priv *priv, u32 queue)
{
struct stmmac_channel *ch = &priv->channel[queue];
unsigned long flags;
u32 buf_size;
-@@ -6433,7 +6437,7 @@ void stmmac_enable_rx_queue(struct stmma
+@@ -6440,7 +6444,7 @@ void stmmac_enable_rx_queue(struct stmma
rx_q->queue_index);
} else {
stmmac_set_dma_bfsize(priv, priv->ioaddr,
rx_q->queue_index);
}
-@@ -6459,7 +6463,7 @@ void stmmac_disable_tx_queue(struct stmm
+@@ -6466,7 +6470,7 @@ void stmmac_disable_tx_queue(struct stmm
void stmmac_enable_tx_queue(struct stmmac_priv *priv, u32 queue)
{
struct stmmac_channel *ch = &priv->channel[queue];
unsigned long flags;
int ret;
-@@ -6509,7 +6513,7 @@ void stmmac_xdp_release(struct net_devic
+@@ -6516,7 +6520,7 @@ void stmmac_xdp_release(struct net_devic
stmmac_disable_all_queues(priv);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
/* Free the IRQ lines */
stmmac_free_irq(dev, REQ_IRQ_ERR_ALL, 0);
-@@ -6568,7 +6572,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6575,7 +6579,7 @@ int stmmac_xdp_open(struct net_device *d
/* DMA RX Channel Configuration */
for (chan = 0; chan < rx_cnt; chan++) {
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
rx_q->dma_rx_phy, chan);
-@@ -6586,7 +6590,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6593,7 +6597,7 @@ int stmmac_xdp_open(struct net_device *d
rx_q->queue_index);
} else {
stmmac_set_dma_bfsize(priv, priv->ioaddr,
rx_q->queue_index);
}
-@@ -6595,7 +6599,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6602,7 +6606,7 @@ int stmmac_xdp_open(struct net_device *d
/* DMA TX Channel Configuration */
for (chan = 0; chan < tx_cnt; chan++) {
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
tx_q->dma_tx_phy, chan);
-@@ -6628,7 +6632,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6635,7 +6639,7 @@ int stmmac_xdp_open(struct net_device *d
irq_error:
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
stmmac_hw_teardown(dev);
init_error:
-@@ -6655,8 +6659,8 @@ int stmmac_xsk_wakeup(struct net_device
+@@ -6662,8 +6666,8 @@ int stmmac_xsk_wakeup(struct net_device
queue >= priv->plat->tx_queues_to_use)
return -EINVAL;
ch = &priv->channel[queue];
if (!rx_q->xsk_pool && !tx_q->xsk_pool)
-@@ -6912,8 +6916,8 @@ int stmmac_reinit_ringparam(struct net_d
+@@ -6923,8 +6927,8 @@ int stmmac_reinit_ringparam(struct net_d
if (netif_running(dev))
stmmac_release(dev);
if (netif_running(dev))
ret = stmmac_open(dev);
-@@ -7351,7 +7355,7 @@ int stmmac_suspend(struct device *dev)
+@@ -7359,7 +7363,7 @@ int stmmac_suspend(struct device *dev)
stmmac_disable_all_queues(priv);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
if (priv->eee_enabled) {
priv->tx_path_in_lpi_mode = false;
-@@ -7402,7 +7406,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend);
+@@ -7411,7 +7415,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend);
static void stmmac_reset_rx_queue(struct stmmac_priv *priv, u32 queue)
{
rx_q->cur_rx = 0;
rx_q->dirty_rx = 0;
-@@ -7410,7 +7414,7 @@ static void stmmac_reset_rx_queue(struct
+@@ -7419,7 +7423,7 @@ static void stmmac_reset_rx_queue(struct
static void stmmac_reset_tx_queue(struct stmmac_priv *priv, u32 queue)
{
if (i >= priv->plat->tx_queues_to_use)
--- a/drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
+++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
-@@ -970,13 +970,13 @@ static int tc_setup_etf(struct stmmac_pr
+@@ -971,13 +971,13 @@ static int tc_setup_etf(struct stmmac_pr
return -EOPNOTSUPP;
if (qopt->queue >= priv->plat->tx_queues_to_use)
return -EINVAL;