net: mvneta: reduce smp_processor_id() calling in mvneta_tx_done_gbe
authorJisheng Zhang <Jisheng.Zhang@synaptics.com>
Fri, 31 Aug 2018 08:11:09 +0000 (16:11 +0800)
committerDavid S. Miller <davem@davemloft.net>
Sun, 2 Sep 2018 21:13:31 +0000 (14:13 -0700)
In the loop of mvneta_tx_done_gbe(), we call the smp_processor_id()
each time, move the call out of the loop to optimize the code a bit.

Before the patch, the loop looks like(under arm64):

        ldr     x1, [x29,#120]
        ...
        ldr     w24, [x1,#36]
        ...
        bl      0 <_raw_spin_lock>
        str     w24, [x27,#132]
        ...

After the patch, the loop looks like(under arm64):

        ...
        bl      0 <_raw_spin_lock>
        str     w23, [x28,#132]
        ...
where w23 is loaded so be ready before the loop.

>From another side, mvneta_tx_done_gbe() is called from mvneta_poll()
which is in non-preemptible context, so it's safe to call the
smp_processor_id() function once.

Signed-off-by: Jisheng Zhang <Jisheng.Zhang@synaptics.com>
Reviewed-by: Gregory CLEMENT <gregory.clement@bootlin.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
drivers/net/ethernet/marvell/mvneta.c

index bcd20ebb2ebdfc20c1f87d7b428b01db010cf4c0..fe3edb3c2bf4c51987b970bdfa1a0e86e9c4066d 100644 (file)
@@ -2507,12 +2507,13 @@ static void mvneta_tx_done_gbe(struct mvneta_port *pp, u32 cause_tx_done)
 {
        struct mvneta_tx_queue *txq;
        struct netdev_queue *nq;
+       int cpu = smp_processor_id();
 
        while (cause_tx_done) {
                txq = mvneta_tx_done_policy(pp, cause_tx_done);
 
                nq = netdev_get_tx_queue(pp->dev, txq->id);
-               __netif_tx_lock(nq, smp_processor_id());
+               __netif_tx_lock(nq, cpu);
 
                if (txq->count)
                        mvneta_txq_done(pp, txq);