Hi Sven,

Which NSD server should I run top on, the fs manager? If so the CPU load is about 155%. I'm working on perf top but not off to a great start...

# perf top
PerfTop: 1095 irqs/sec kernel:61.9% exact: 0.0% [1000Hz cycles], (all, 28 CPUs)
---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
Segmentation fault

-Aaron

On 3/24/17 1:04 PM, Sven Oehme wrote:
while this is happening  run top and see if there is very high cpu
utilization at this time on the NSD Server.

if there is , run perf top (you might need to install perf command) and
see if the top cpu contender is a spinlock . if so send a screenshot of
perf top as i may know what that is and how to fix.

sven


On Fri, Mar 24, 2017 at 9:43 AM Aaron Knister <[email protected]
<mailto:[email protected]>> wrote:

    Since yesterday morning we've noticed some deadlocks on one of our
    filesystems that seem to be triggered by writing to it. The waiters on
    the clients look like this:

    0x19450B0 (   6730) waiting 2063.294589599 seconds, SyncHandlerThread:
    on ThCond 0x1802585CB10 (0xFFFFC9002585CB10) (InodeFlushCondVar), reason
    'waiting for the flush flag to commit metadata'
    0x7FFFDA65E200 (  22850) waiting 0.000246257 seconds,
    AllocReduceHelperThread: on ThCond 0x7FFFDAC7FE28 (0x7FFFDAC7FE28)
    (MsgRecordCondvar), reason 'RPC wait' for allocMsgTypeRelinquishRegion
    on node 10.1.52.33 <c0n3271>
    0x197EE70 (   6776) waiting 0.000198354 seconds,
    FileBlockWriteFetchHandlerThread: on ThCond 0x7FFFF00CD598
    (0x7FFFF00CD598) (MsgRecordCondvar), reason 'RPC wait' for
    allocMsgTypeRequestRegion on node 10.1.52.33 <c0n3271>

    (10.1.52.33/c0n3271 <http://10.1.52.33/c0n3271> is the fs manager
    for the filesystem in question)

    there's a single process running on this node writing to the filesystem
    in question (well, trying to write, it's been blocked doing nothing for
    half an hour now). There are ~10 other client nodes in this situation
    right now. We had many more last night before the problem seemed to
    disappear in the early hours of the morning and now its back.

    Waiters on the fs manager look like this. While the individual waiter is
    short it's a near constant stream:

    0x7FFF60003540 (   8269) waiting 0.001151588 seconds, Msg handler
    allocMsgTypeRequestRegion: on ThMutex 0x1802163A2E0 (0xFFFFC9002163A2E0)
    (AllocManagerMutex)
    0x7FFF601C8860 (  20606) waiting 0.001115712 seconds, Msg handler
    allocMsgTypeRelinquishRegion: on ThMutex 0x1802163A2E0
    (0xFFFFC9002163A2E0) (AllocManagerMutex)
    0x7FFF91C10080 (  14723) waiting 0.000959649 seconds, Msg handler
    allocMsgTypeRequestRegion: on ThMutex 0x1802163A2E0 (0xFFFFC9002163A2E0)
    (AllocManagerMutex)
    0x7FFFB03C2910 (  12636) waiting 0.000769611 seconds, Msg handler
    allocMsgTypeRequestRegion: on ThMutex 0x1802163A2E0 (0xFFFFC9002163A2E0)
    (AllocManagerMutex)
    0x7FFF8C092850 (  18215) waiting 0.000682275 seconds, Msg handler
    allocMsgTypeRelinquishRegion: on ThMutex 0x1802163A2E0
    (0xFFFFC9002163A2E0) (AllocManagerMutex)
    0x7FFF9423F730 (  12652) waiting 0.000641915 seconds, Msg handler
    allocMsgTypeRequestRegion: on ThMutex 0x1802163A2E0 (0xFFFFC9002163A2E0)
    (AllocManagerMutex)
    0x7FFF9422D770 (  12625) waiting 0.000494256 seconds, Msg handler
    allocMsgTypeRequestRegion: on ThMutex 0x1802163A2E0 (0xFFFFC9002163A2E0)
    (AllocManagerMutex)
    0x7FFF9423E310 (  12651) waiting 0.000437760 seconds, Msg handler
    allocMsgTypeRelinquishRegion: on ThMutex 0x1802163A2E0
    (0xFFFFC9002163A2E0) (AllocManagerMutex)

    I don't know if this data point is useful but both yesterday and today
    the metadata NSDs for this filesystem have had a constant aggregate
    stream of 25MB/s 4kop/s reads during each episode (very low latency
    though so I don't believe the storage is a bottleneck here). Writes are
    only a few hundred ops and didn't strike me as odd.

    I have a PMR open for this but I'm curious if folks have seen this in
    the wild and what it might mean.

    -Aaron

    --
    Aaron Knister
    NASA Center for Climate Simulation (Code 606.2)
    Goddard Space Flight Center
    (301) 286-2776 <tel:(301)%20286-2776>
    _______________________________________________
    gpfsug-discuss mailing list
    gpfsug-discuss at spectrumscale.org <http://spectrumscale.org>
    http://gpfsug.org/mailman/listinfo/gpfsug-discuss



_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss


--
Aaron Knister
NASA Center for Climate Simulation (Code 606.2)
Goddard Space Flight Center
(301) 286-2776
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss

Reply via email to