cvs commit: src/sys/netinet in_pcb.c tcp_subr.c tcp_timer.c tcp_var.h

Gleb Smirnoff glebius at FreeBSD.org
Wed Sep 6 08:01:53 PDT 2006


  Mike,

On Wed, Sep 06, 2006 at 09:49:15AM -0500, Mike Silbersack wrote:
M> >Then we found the CPU hog in the in_pcblookup_local(). I've added
M> >counters and gathered stats via ktr(4). When a lag occured, the
M> >following data was gathered:
M> >
M> >112350 return 0x0, iterations 0, expired 0
M> >112349 return 0xc5154888, iterations 19998, expired 745
M> 
M> Ah, I think I see what's happening.  It's probably spinning because the 
M> heuristic isn't triggering on each entry, that doesn't surprise me.  What 
M> does surprise me is that it's expiring more than one entry - my original 
M> intent with that code was for it to free just one entry, which it would 
M> then use... meaning that I goofed up the implementation.
M> 
M> I had been thinking of rewriting that heuristic anyway, I'm sure that I 
M> can go back and find something far more efficient if you give me a few 
M> days.  (Or a week.)

I think we should free the oldmost tcptw entry in a case if we can't
find the local endpoint. We can tell definitely that we can't find one
only in in_pcbbind_setup() in the "do {} while (in_pcblookup_local)" cycle,
where EADDRNOTAVAIL is returned. We can't definitely tell this in
in_pcblookup_local() since we don't know whether tried port is the
last one.

The oldmost tcptw entry can be taken simply from the ordered list, like
tcp_timer_2msl_tw() does this.

However, I don't like the idea of "finding" the free port at all. This
makes connect()/bind() performance depending on number of busy endpoints.
Shouldn't we make an algorythm, where free endpoints are stored, and
we don't need to _find_ one, we could just _take_ one?

M> >1.78 hasn't yet been merged to RELENG_6, and we faced the problem on
M> >RELENG_6 boxes where the periodic merging cycle is present. So the
M> >problem is not in 1.78 of tcp_timer.c. We have a lot of tcptw entries
M> >because we have a very big connection rate, not because they are
M> >leaked or not purged.
M> 
M> Ok, just checking.
M> 
M> With this code removed, are you not seeing the web frontends delaying new 
M> connections when they can't find a free port to use?

No. We monitor the amount of entries in tcptw zone. It is the same
as before. So the periodic cycle purges tcptw states at the same
rate as in_pcblookup_local() was, except that it does this consuming
less CPU time.

-- 
Totus tuus, Glebius.
GLEBIUS-RIPN GLEB-RIPE


More information about the cvs-src mailing list