On Fri, Feb 13, 2026 at 11:41 AM Stuart Henderson <[email protected]> wrote:
>
> On 2026/02/13 10:18, K R wrote:
> >
> > It runs about 10 network daemons serving TCP clients. About 64-128
> > open sockets each, at any given time. Not much traffic, but around 4k
> > pf states.
>
> Yet it seems you ran into 100k states to be hitting PF state limits?
Right.
> I wonder if it's worth scripting a check on the number of states and
> dumping the state table (pfctl -ss -v at least) to get an idea what's in
> there when it's high.
Yes, good idea, thanks!
>
>
> > The resources:
> >
> > hw.model=Intel(R) Xeon(R) Gold 6338 CPU @ 2.00GHz
> > hw.vendor=VMware, Inc.
> > hw.physmem=4277600256
> > hw.ncpuonline=2
> >
> > > > ddb{1}> show all pools
> > > > Name Size Requests Fail Releases Pgreq Pgrel Npage Hiwat Minpg
> > > > Maxpg Idle
> > > > tcpcb 736 4309640 103 4091353 20236 382 19854 19854 0
> > > > 8 0
> > > > inpcb 328 5892193 0 5673850 18519 314 18205 18205 0
> > > > 8 0
> > > > sockpl 552 7621733 0 7403339 15972 364 15608 15608 0
> > > > 8 0
> > > > mbufpl 256 286232 0 0 13640 5 13635 13635 0
> > > > 8 0
> > >
> > > If I read this currectly the box has 20k+ TCP sockets open. Which results
> > > in high resrouce usage of tcpcb, inpcb, sockpl and for the TCP template
> > > mbufs.
> >
> > What I see now, using systat pool, sorted by Npage:
> >
> > NAME SIZE REQUESTS FAIL INUSE PGREQ PGREL
> > NPAGE HIWAT MINPG MAXPG
> > tcpcb 736 1670128 0 40124 4548 308
> > 4240 4297 0 8
> > inpcb 328 2438004 0 40182 4150 247
> > 3903 3944 0 8
> > sockpl 552 3299804 0 40236 3621 255
> > 3366 3385 0 8
> > mbufpl 256 49530665 0 39963 2949 5
> > 2944 2944 0 8
> >
> > > At least the tcpcb and sockpl use the kmem_map.
> > > Which is (19854 + 15608) * 4k or 141848K. Your kmem_map has a limit of
> > > 186616K
> > > so there is just not enough space. You may need to increase memory or you
> > > can also tune NKMEMPAGES via config(8).
> >
> > I see. It is odd, though, that we have similar machines (both VMs and
> > baremetal, similar resources) and the only one that panics is this
> > one, running under VMware.
> >
> > >
> > > > pfstate 384 16598777 5933960 16587284 239196 237883 1313 10001 0
> > > > 8 0
> > >
> > > There seems to be some strange bursts on the pfstate pool as well.
> > >
> > > --
> > > :wq Claudio
> >