Dave Chinner wrote: > So effectively the storage subsystem (NFS, filesystem, DM, MD, > device drivers) have about 4K of stack to work in now. That seems to > be a lot less than last time I looked at this, and we've been really > careful not to increase XFS's stack usage for quite some time now. OK. I should note that we have what appears to be a similar problem on a 2.6.28 distro kernel, so I'm not sure this is a very recent change. (We see the lockups on that kernel, we haven't tried larger stacks + stack instrumentation on the earlier kernel). Do you know if there are any obvious knobs to twiddle to make these codepaths less likely? The cluster is resilient against occasional server death, but frequent death is more annoying. We're currently running with sysctls: net.ipv4.ip_nonlocal_bind=1 kernel.panic=300 vm.dirty_background_ratio=3 vm.min_free_kbytes=16384 I'm not sure what circumstances force the memory reclaim (and why it doesn't come from discarding a cached page). Is the problem is the DMA/DMA32 zone and we should try playing with lowmem_reserve_ratio? Is there anything else we could do to keep dirty pages out of the low zones? Before trying THREAD_ORDER 2, we tried doubling the RAM in a couple of boxes from 2GB to 4GB without any significant reduction in the problem. Lastly - if we end up stuck with THREAD_ORDER 2, does anyone know what symptoms to look out for to know if unable to allocate thread stacks due to fragmentation? > I'll have to have a bit of a think on this one - if you could > provide further stack traces as they get deeper (esp. if they go > past 8k) that would be really handy. Two of the worst offenders below. We have plenty to send if you would like more. Please let us know if you'd like us to try anything else or would like other info. Thanks very much for your thoughts, suggestions and work so far, it's very much appreciated here. regards, jb