Skip to content

Commit

Permalink
rps: avoid one atomic in enqueue_to_backlog
Browse files Browse the repository at this point in the history
If CONFIG_SMP=y, then we own a queue spinlock, we can avoid the atomic
test_and_set_bit() from napi_schedule_prep().

We now have same number of atomic ops per netif_rx() calls than with
pre-RPS kernel.

Signed-off-by: Eric Dumazet <eric.dumazet@gmail.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
  • Loading branch information
Eric Dumazet authored and David S. Miller committed May 18, 2010
1 parent 3f78d1f commit ebda37c
Showing 1 changed file with 4 additions and 2 deletions.
6 changes: 4 additions & 2 deletions net/core/dev.c
Original file line number Diff line number Diff line change
Expand Up @@ -2432,8 +2432,10 @@ static int enqueue_to_backlog(struct sk_buff *skb, int cpu,
return NET_RX_SUCCESS;
}

/* Schedule NAPI for backlog device */
if (napi_schedule_prep(&sd->backlog)) {
/* Schedule NAPI for backlog device
* We can use non atomic operation since we own the queue lock
*/
if (!__test_and_set_bit(NAPI_STATE_SCHED, &sd->backlog.state)) {
if (!rps_ipi_queued(sd))
____napi_schedule(sd, &sd->backlog);
}
Expand Down

0 comments on commit ebda37c

Please sign in to comment.