[RFC 0/3] lro: Generic Large Receive Offload for TCP traffic
gallatin at gmail.com
Fri Jul 13 02:01:53 EST 2007
On 7/11/07, Jan-Bernd Themann <ossthema at de.ibm.com> wrote:
> Generic Large Receive Offload proposal
I'm very glad that somebody is stepping up to take responsibility
I'm the primary author of the Myricom Myri10GE driver, and its LRO mechanism
(which has been rejected several times when posted here). I don't subscribe
to the LKML or netdev lists (the b/w is way too much for me).
Thankfully, my colleague Brice (who maintains our driver in the kernel)
forwarded me this message.
I looked at your patch, and I believe that we can improve the
performance further by using a slightly different approach, or at
least making that approach optional.
When I first did our LRO implementation, it aggregated packets
essentially the same way you are doing it -- by appending packets to
the frag_list. I did quite extensive profiling, and the most expensive
operations seemed to be the allocation and freeing of memory. A
colleague of mine (Loic, CC'ed) came up with the brilliant idea of
receiving into pages. When I implemented his suggestion, it turned
out to be much, much more efficient to receive into pages, and to
accumulate the pages to the frags array.
The benchmarks I did on very low end machines in my lab (2GHz
amd64 x2 3800+) showed that the receiver was saturated at roughly
4.2Gb/s without LRO, 5.7Gb/s with frag_list based LRO, 8.6Gb/s with
frags array based LRO, and somewhat idle at line rate with frags array
based LRO and 32KB order=3 pages. (This is 1500b frames, BTW).
The savings comes from being able to do fewer allocations. For
example, 2 1500b packets fit in a single page. So, for a "full" frag
array, we have 16 1/2 4KB pages and a single skb holding them.
This works out to be 9 allocations for roughly 23KB of payload,
rather than 16. Using order 3 (32KB) pages, it gets even better,
and we have just 2 allocations per full skb frag list.
So... It would be wonderful if your patch could also deal with
data residing in pages, rather than in skbs. I can understand
how you might not want to modify your driver to do this, which
is why I'm asking about making it optional. However,
your driver would probably benefit from receiving into
pages, and I encourage you to investigate that.
I'm picturing an additional path into lro, such as:
int lro_maybe_receive_frags(struct net_lro_mgr *lro_mgr,
struct skb_frag_struct *frags,
int len, void *priv);
This would return 0 if the packet was "accepted", and an
error if it was not. It would then call a modified
__lro_proc_skb() (perhaps better named __lro_proc_segment())
which would have the length as an argument (so as to avoid
the need to pass the skb to lro_tcp_ip_check()) in addition
to the *frags.
The only real additional work would be in having an alternate path
inside lro_init_desc() which allocated an skb to hang the pages from
if the passed skb was null, and in having an alternate
lro_add_packet() path which added the frag(s), rather than chaining an
Also, your patch does not seem to maintain TCP checksums.
To be fair, we did not maintain checksums in our earlier LRO
work either, and it was much simpler that way! :) However,
not maintaining the TCP checksum can lead to user complaints
(invalid csum reported by traffic sniffers), as well as problems
when various filtering software is used which attempts to
re-write the TCP header.
I would very much appreciate it if you could look at my LRO
implementation in our Myri10GE driver which has been
posted (and rejected) several times in the past by Brice.
The source code is also available at:
It uses a page based approach, and it maintains correct TCP
This driver is Dual BSD/GPL licensed, and you are free to take
whatever you like.
Thank you again for stepping up with a generic implementation!
More information about the Linuxppc-dev