On Mon, Aug 11, 2014 at 10:44:23AM -0500, [email protected] wrote:
>
> Hi,
>
> I was running an fio job over the weekend and the main thread
> seems to have hung. This in on RHEL6u5 with kernel
> 2.6.32-431.20.3.el6.x86_64 but with a highly modified hpsa
> driver, just in case it matters.
>
> [root@msablackburn saperf]# uname -a
> Linux msablackburn.americas.hpqcorp.net 2.6.32-431.20.3.el6.x86_64 #1 SMP Fri
> Jun 6 18:30:54 EDT 2014 x86_64 x86_64 x86_64 GNU/Linux
> [root@msablackburn saperf]# /usr/local/bin/fio --version
> fio-2.1.11-3-g789d
>
> The jobfile looks like this (note the runtime=216000, that's 60 hours)
> It seems to have hung with about 25 hours remaining.)
>
> [root@msablackburn saperf]# cat 2drive_sdr.fio
> [global]
> direct=1
> ioengine=libaio
> norandommap
> randrepeat=0
> bs=4096
> iodepth=1011
> #iodepth=96
> #numjobs=1
> numjobs=12
> #numjobs=24
> # number_ios=1
> runtime=216000
> #runtime=54000
> time_based=1
> group_reporting
> thread
> gtod_reduce=1
> #iodepth_batch=4
> #iodepth_batch_complete=4
> cpus_allowed=0-5
> cpus_allowed_policy=split
> rw=randread
>
> [4_KiB_RR_drive_r]
> filename=/dev/sdr
>
> [4_KiB_RR_drive_s]
> filename=/dev/sds
>
>
> Here is all the output so far:
>
> ---------------8<------------8<-------------8<-----------------
> [root@msablackburn saperf]# fio 2drive_sdr.fio
> 4_KiB_RR_drive_r: (g=0): rw=randread, bs=4K-4K/4K-4K/4K-4K, ioengine=libaio,
> iodepth=1011
> ...
> 4_KiB_RR_drive_s: (g=0): rw=randread, bs=4K-4K/4K-4K/4K-4K, ioengine=libaio,
> iodepth=1011
> ...
> fio-2.1.11-3-g789d
> Starting 24 threads
> ^Cbs: 24 (f=24): [r(24)] [1.2% done] [1113MB/0KB/0KB /s] [285K/0/0 iops] [eta
> 14h:48m:47s]
> fio: terminating on signal 2
>
> 4_KiB_RR_drive_r: (groupid=0, jobs=24): err= 0: pid=7696: Fri Aug 8 16:49:19
> 2014
> read : io=881568MB, bw=1309.1MB/s, iops=335341, runt=672989msec
> cpu : usr=1.75%, sys=23.15%, ctx=1266117, majf=9, minf=835
> IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=100.0%
> submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%,
> >=64=0.0%
> complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%,
> >=64=0.1%
> issued : total=r=225681428/w=0/d=0, short=r=0/w=0/d=0
> latency : target=0, window=0, percentile=100.00%, depth=1011
>
> Run status group 0 (all jobs):
> READ: io=881568MB, aggrb=1309.1MB/s, minb=1309.1MB/s, maxb=1309.1MB/s,
> mint=672989msec, maxt=672989msec
>
> Disk stats (read/write):
> sdr: ios=112299535/0, merge=0/0, ticks=13349007/0, in_queue=13054755,
> util=98.57%
> sds: ios=113359581/0, merge=0/0, ticks=13710834/0, in_queue=13414931,
> util=98.71%
> [root@msablackburn saperf]# vi !$
> vi 2drive_sdr.fio
> [root@msablackburn saperf]# fio 2drive_sdr.fio
> 4_KiB_RR_drive_r: (g=0): rw=randread, bs=4K-4K/4K-4K/4K-4K, ioengine=libaio,
> iodepth=1011
> ...
> 4_KiB_RR_drive_s: (g=0): rw=randread, bs=4K-4K/4K-4K/4K-4K, ioengine=libaio,
> iodepth=1011
> ...
> fio-2.1.11-3-g789d
> Starting 24 threads
> Jobs: 24 (f=24): [r(24)] [57.1% done] [1408MB/0KB/0KB /s] [361K/0/0 iops]
> [eta 01d:01h:43m:07s]
> ---------------8<------------8<-------------8<-----------------
>
> And there it sits, the eta does not update.
>
>
> I tried connecting to the process with gdb, and doing "info threads" and "bt":
>
> PID of main thread is 8530:
>
> root 8530 598 1.0 2159116 170812 pts/0 Sl+ Aug08 23570:56 fio
> 2drive_sdr.fio
>
> ...
> (gdb) info threads
> 26 Thread 0x7f65da431700 (LWP 8531) 0x0000003aa4cacced in nanosleep ()
> from /lib64/libc.so.6
> 25 Thread 0x7f65d9a30700 (LWP 8532) 0x0000003c69800667 in io_submit ()
> from /lib64/libaio.so.1
> 24 Thread 0x7f65d902f700 (LWP 8533) 0x0000003c69800667 in io_submit ()
> from /lib64/libaio.so.1
> 23 Thread 0x7f65d862e700 (LWP 8534) 0x0000003c69800667 in io_submit ()
> from /lib64/libaio.so.1
> 22 Thread 0x7f65d7c2d700 (LWP 8535) 0x0000003c69800667 in io_submit ()
> from /lib64/libaio.so.1
> 21 Thread 0x7f65d722c700 (LWP 8536) 0x0000003c69800667 in io_submit ()
> from /lib64/libaio.so.1
> 20 Thread 0x7f65d682b700 (LWP 8537) 0x0000003c69800667 in io_submit ()
> from /lib64/libaio.so.1
> 19 Thread 0x7f65d5e2a700 (LWP 8538) 0x0000003c69800667 in io_submit ()
> from /lib64/libaio.so.1
> 18 Thread 0x7f65d5429700 (LWP 8539) 0x0000003c69800667 in io_submit ()
> from /lib64/libaio.so.1
> 17 Thread 0x7f65d4a28700 (LWP 8540) 0x0000003c69800667 in io_submit ()
> from /lib64/libaio.so.1
> 16 Thread 0x7f65d4027700 (LWP 8541) 0x0000003c69800667 in io_submit ()
> from /lib64/libaio.so.1
> 15 Thread 0x7f65d3626700 (LWP 8542) 0x0000003c69800667 in io_submit ()
> from /lib64/libaio.so.1
> 14 Thread 0x7f65d2c25700 (LWP 8543) 0x0000003c69800667 in io_submit ()
> from /lib64/libaio.so.1
> 13 Thread 0x7f65d2222700 (LWP 8544) 0x0000003c69800667 in io_submit ()
> from /lib64/libaio.so.1
> 12 Thread 0x7f65d1820700 (LWP 8545) 0x0000003c69800667 in io_submit ()
> from /lib64/libaio.so.1
> 11 Thread 0x7f65d0e1e700 (LWP 8546) td_io_queue (td=0x7f65da496410,
> io_u=0x7f656804d010) at ioengines.c:344
> 10 Thread 0x7f65d041c700 (LWP 8547) 0x0000003c69800667 in io_submit ()
> from /lib64/libaio.so.1
> 9 Thread 0x7f65cfa1a700 (LWP 8548) 0x0000003c69800667 in io_submit () from
> /lib64/libaio.so.1
> 8 Thread 0x7f65cf018700 (LWP 8549) 0x0000003c69800667 in io_submit () from
> /lib64/libaio.so.1
> 7 Thread 0x7f65ce616700 (LWP 8550) 0x0000003c69800667 in io_submit () from
> /lib64/libaio.so.1
> 6 Thread 0x7f65cdc14700 (LWP 8551) fio_libaio_event (td=0x7f65da4b06a8,
> event=<value optimized out>) at engines/libaio.c:65
> 5 Thread 0x7f65cd212700 (LWP 8552) __get_next_offset (td=0x7f65da4b5a60)
> at io_u.c:389
> 4 Thread 0x7f65cc810700 (LWP 8553) 0x0000003c69800667 in io_submit () from
> /lib64/libaio.so.1
> 3 Thread 0x7f65cbe0e700 (LWP 8554) td_io_u_unlock (td=0x7f65da4c01d0,
> min_evts=<value optimized out>, bytes=0x7f65cbe0dd10) at fio.h:619
> 2 Thread 0x7f65cb40c700 (LWP 8555) 0x0000003c69800667 in io_submit () from
> /lib64/libaio.so.1
> * 1 Thread 0x7f65e34c6720 (LWP 8530) 0x0000003aa4cacced in nanosleep () from
> /lib64/libc.so.6
> (gdb) bt
> #0 0x0000003aa4cacced in nanosleep () from /lib64/libc.so.6
> #1 0x0000003aa4ce1e44 in usleep () from /lib64/libc.so.6
> #2 0x000000000044afe4 in do_usleep () at backend.c:1727
> #3 run_threads () at backend.c:1965
> #4 0x000000000044b84d in fio_backend () at backend.c:2068
> #5 0x0000003aa4c1ed1d in __libc_start_main () from /lib64/libc.so.6
> #6 0x0000000000409c39 in _start ()
> (gdb) detach
> Detaching from program: /usr/local/bin/fio, process 8530
> (gdb) quit
> [root@msablackburn saperf]#
>
> I am guessing the 8530 thread is the one that prints out the eta updates?
> Well, maybe not, I see a call to create_disk_util_thread() in backend.c.
>
> i/o still seems to be running, and there are no messages in dmesg about
> hung i/o or the scsi mid layer getting angry and trying to abort things
> or anything like that. Seems to be happily cruising along, except that
> the eta isn't updating, and the job should have finished several hours
> ago now. I'm leaving it running for now just to try to figure out
> what happened.
>
> Any ideas? Anything else I should poke at?
A bit more info:
thread 26 looks to be the one that's supposed to print the eta info:
(gdb) thread 26
[Switching to thread 26 (Thread 0x7f65da431700 (LWP 8531))]#0
0x0000003aa4cacced in nanosleep () from /lib64/libc.so.6
(gdb) bt
#0 0x0000003aa4cacced in nanosleep () from /lib64/libc.so.6
#1 0x0000003aa4ce1e44 in usleep () from /lib64/libc.so.6
#2 0x00000000004474a0 in disk_thread_main (data=<value optimized out>) at
backend.c:1992
#3 0x0000003aa50079d1 in start_thread () from /lib64/libpthread.so.0
#4 0x0000003aa4ce8b5d in clone () from /lib64/libc.so.6
(gdb)
static void *disk_thread_main(void *data)
{
int ret = 0;
fio_mutex_up(startup_mutex);
while (threads && !ret) {
usleep(DISK_UTIL_MSEC * 1000);
if (!threads)
break;
ret = update_io_ticks();
if (!is_backend)
print_thread_status();
}
fio_mutex_up(disk_thread_mutex);
return NULL;
}
that usleep() is where it's in nanosleep, I guess.
from eta.c:
void print_thread_status(void)
{
struct jobs_eta *je;
size_t size;
je = get_jobs_eta(0, &size);
if (je)
display_thread_status(je);
free(je);
}
Maybe that je is coming back false? which is
probably the return value of calc_thread_status() which, well,
at a glance, I'm not sure what calc_thread_status() is doing.
-- steve
--
To unsubscribe from this list: send the line "unsubscribe fio" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html