diff options
author | Tom Herbert <therbert@google.com> | 2010-11-21 13:17:27 +0000 |
---|---|---|
committer | David S. Miller <davem@davemloft.net> | 2010-11-24 11:44:20 -0800 |
commit | 1d24eb4815d1e0e8b451ecc546645f8ef1176d4f (patch) | |
tree | 0172e72b9452dc46c4e1043817005979cec022a7 /drivers/w1 | |
parent | 3853b5841c01a3f492fe137afaad9c209e5162c6 (diff) | |
download | op-kernel-dev-1d24eb4815d1e0e8b451ecc546645f8ef1176d4f.zip op-kernel-dev-1d24eb4815d1e0e8b451ecc546645f8ef1176d4f.tar.gz |
xps: Transmit Packet Steering
This patch implements transmit packet steering (XPS) for multiqueue
devices. XPS selects a transmit queue during packet transmission based
on configuration. This is done by mapping the CPU transmitting the
packet to a queue. This is the transmit side analogue to RPS-- where
RPS is selecting a CPU based on receive queue, XPS selects a queue
based on the CPU (previously there was an XPS patch from Eric
Dumazet, but that might more appropriately be called transmit completion
steering).
Each transmit queue can be associated with a number of CPUs which will
use the queue to send packets. This is configured as a CPU mask on a
per queue basis in:
/sys/class/net/eth<n>/queues/tx-<n>/xps_cpus
The mappings are stored per device in an inverted data structure that
maps CPUs to queues. In the netdevice structure this is an array of
num_possible_cpu structures where each structure holds and array of
queue_indexes for queues which that CPU can use.
The benefits of XPS are improved locality in the per queue data
structures. Also, transmit completions are more likely to be done
nearer to the sending thread, so this should promote locality back
to the socket on free (e.g. UDP). The benefits of XPS are dependent on
cache hierarchy, application load, and other factors. XPS would
nominally be configured so that a queue would only be shared by CPUs
which are sharing a cache, the degenerative configuration woud be that
each CPU has it's own queue.
Below are some benchmark results which show the potential benfit of
this patch. The netperf test has 500 instances of netperf TCP_RR test
with 1 byte req. and resp.
bnx2x on 16 core AMD
XPS (16 queues, 1 TX queue per CPU) 1234K at 100% CPU
No XPS (16 queues) 996K at 100% CPU
Signed-off-by: Tom Herbert <therbert@google.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'drivers/w1')
0 files changed, 0 insertions, 0 deletions