Can you say what causes the difference? I had a brief poke
around - generic_copy_from_user() dominates in both cases
of course, but nothing really stood out when comparing the
zerocopy kernel's profile with non-zc.
Varying the value of MAXPGS (all the way down to 1) and also
the amount of data which is sent with send() does change the
throughput, but not the ratio wrt non-zc.
> What do we get for this cost?
> Basically, the big win is not that the card checksums the packet.
> We could get that for free while copying the data from userspace
> into the kernel pages during the sendmsg(), using the combined
> "copy+checksum" hand-coded assembly routines we already have.
> It is in fact the better use of memory. Firstly, we use page
> allocations, only single ones. With linear buffers SLAB could
> use multiple pages which strain the memory subsystem quite a bit at
> times. Secondly, we fill pages with socket data precisely whereas
> SLAB can only get as tight packing as any general purpose memory
> allocator can.
> This, I feel, outweighs the slight performance decrease. And I would
> wager a bet that the better usage of memory will result in better
> all around performance.
ie: inappropriate test coverage. Not surprising. What
additional scenarios need to be tested? Zillions of
If anyone really needs that 10% they can use the `hw_checksums=0'
module parm, but SG+xsum is enabled by default - we need the testing.
> The problem with microscopic tests is that you do not see the world
> around the thing being focused on. I feel Andrew/Jamal's test are
> very valuable, but lets keep things in perspective when doing cost
> Finally, please do some tests on loopback. It is usually a great
> way to get "pure software overhead" measurements of our TCP stack.
BTW: can you suggest why I'm not observing any change in NFS client
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to email@example.com
Please read the FAQ at http://www.tux.org/lkml/