bad NFS/UDP performance

Matthew Dillon dillon at apollo.backplane.com
Fri Sep 26 18:30:32 UTC 2008


:> 	-vfs.nfs.realign_test: 22141777
:> 	+vfs.nfs.realign_test: 498351
:> 
:> 	-vfs.nfsrv.realign_test: 5005908
:> 	+vfs.nfsrv.realign_test: 0
:> 
:> 	+vfs.nfsrv.commit_miss: 0
:> 	+vfs.nfsrv.commit_blks: 0
:> 
:> changing them did nothing - or at least with respect to nfs throughput :-)
:
:I'm not sure what any of these do, as NFS is a bit out of my league.
::-)  I'll be following this thread though!
:
:-- 
:| Jeremy Chadwick                                jdc at parodius.com |

    A non-zero nfs_realign_count is bad, it means NFS had to copy the
    mbuf chain to fix the alignment.  nfs_realign_test is just the
    number of times it checked.  So nfs_realign_test is irrelevant.
    it's nfs_realign_count that matters.

    Several things can cause NFS payloads to be improperly aligned.
    Anything from older network drivers which can't start DMA on a 
    2-byte boundary, resulting in the 14-byte encapsulation header 
    causing improper alignment of the IP header & payload, to rpc
    embedded in NFS TCP streams winding up being misaligned.

    Modern network hardware either support 2-byte-aligned DMA, allowing
    the encapsulation to be 2-byte aligned so the payload winds up being
    4-byte aligned, or support DMA chaining allowing the payload to be
    placed in its own mbuf, or pad, etc.

    --

    One thing I would check is to be sure a couple of nfsiod's are running
    on the client when doing your tests.  If none are running the RPCs wind
    up being more synchronous and less pipelined.  Another thing I would
    check is IP fragment reassembly statistics (for UDP) - there should be
    none for TCP connections no matter what the NFS I/O size selected.

    (It does seem more likely to be scheduler-related, though).

						-Matt



More information about the freebsd-stable mailing list