If I stat on that box, I get nothing:
q782657@usadc-seaxd01:/mounts/ceph1/pubdata/tcga/raw$ cd BRCA
-bash: cd: BRCA: No such file or directory
perl -e 'print stat("BRCA")'
<no result>
If I access a mount on another machine, I can see the files:
q782657@usadc-nasea05:/mounts/ceph1/pubdata/tcga$ ls -l raw
total 0
drwxrwxr-x 1 q783775 pipeline 366462246414 May 8 12:00 BRCA
drwxrwxr-x 1 q783775 pipeline 161578200377 May 8 12:00 COAD
drwxrwxr-x 1 q783775 pipeline 367320207221 May 8 11:35 HNSC
drwxrwxr-x 1 q783775 pipeline 333587505256 May 8 13:27 LAML
drwxrwxr-x 1 q783775 pipeline 380346443564 May 8 13:27 LUSC
drwxrwxr-x 1 q783775 pipeline 357340261602 May 8 13:33 PAAD
drwxrwxr-x 1 q783775 pipeline 389882082560 May 8 13:33 PRAD
drwxrwxr-x 1 q783775 pipeline 634089122305 May 8 13:33 STAD
drwxrwxr-x 1 q783775 pipeline 430754940032 May 8 13:33 THCA
I will try updating the kernel, and rerunning some tests. Thanks.
-----Original Message-----
From: Lincoln Bryant [mailto:[email protected]]
Sent: Friday, May 09, 2014 10:39 AM
To: Aronesty, Erik
Cc: ceph-users
Subject: Re: [ceph-users] issues with ceph
Hi Erik,
What happens if you try to stat one of the "missing" files (assuming you know
the name of the file before you remount raw)?
I had a problem where files would disappear and reappear in CephFS, which I
believe was fixed in kernel 3.12.
Cheers,
Lincoln
On May 9, 2014, at 9:30 AM, Aronesty, Erik wrote:
> So we were attempting to stress test a cephfs installation, and last night,
> after copying 500GB of files, we got this:
>
> 570G in the "raw" directory
>
> q782657@usadc-seaxd01:/mounts/ceph1/pubdata/tcga$ ls -lh
> total 32M
> -rw-rw-r-- 1 q783775 pipeline 32M May 8 10:39
> 2014-02-25T12:00:01-0800_data_manifest.tsv
> -rw-rw-r-- 1 q783775 pipeline 144 May 8 10:42 cghub.key
> drwxrwxr-x 1 q783775 pipeline 234G May 8 11:31 fastqs
> drwxrwxr-x 1 q783775 pipeline 570G May 8 13:33 raw
> -rw-rw-r-- 1 q783775 pipeline 86 May 8 11:19 readme.txt
>
> But when I ls into the "raw" folder, I get zero files:
>
> q782657@usadc-seaxd01:/mounts/ceph1/pubdata/tcga$ ls -lh raw
> total 0
>
> If I mount that folder again... all the files "re-appear".
>
> Is this a bug that's been solved in a newer release?
>
> KERNEL:
> Linux usadc-nasea05 3.11.0-20-generic #34~precise1-Ubuntu SMP Thu Apr 3
> 17:25:07 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux
>
> CEPH:
> ii ceph 0.72.2-1precise
> distributed storage and file system
>
>
> ------ No errors that I could see on the client machine:
>
> q782657@usadc-seaxd01:/mounts/ceph1/pubdata/tcga$ dmesg | grep ceph
> [588560.047193] Key type ceph registered
> [588560.047334] libceph: loaded (mon/osd proto 15/24)
> [588560.102874] ceph: loaded (mds proto 32)
> [588560.117392] libceph: client6005 fsid f067539c-7426-47ee-afb0-7d2c6dfcbcd0
> [588560.126477] libceph: mon1 10.18.176.180:6789 session established
>
>
> ------ Ceph itself looks fine.
>
> root@usadc-nasea05:~# ceph health
> HEALTH_OK
>
> root@usadc-nasea05:~# ceph quorum_status
> {"election_epoch":668,"quorum":[0,1,2,3],"quorum_names":["usadc-nasea05","usadc-nasea06","usadc-nasea07","usadc-nasea08"],"quorum_leader_name":"usadc-nasea05","monmap":{"epoch":1,"fsid":"f067539c-7426-47ee-afb0-7d2c6dfcbcd0","modified":"0.000000","created":"0.000000","mons":[{"rank":0,"name":"usadc-nasea05","addr":"10.18.176.179:6789\/0"},{"rank":1,"name":"usadc-nasea06","addr":"10.18.176.180:6789\/0"},{"rank":2,"name":"usadc-nasea07","addr":"10.18.176.181:6789\/0"},{"rank":3,"name":"usadc-nasea08","addr":"10.18.176.182:6789\/0"}]}}
>
> root@usadc-nasea05:~# ceph mon dump
> dumped monmap epoch 1
> epoch 1
> fsid f067539c-7426-47ee-afb0-7d2c6dfcbcd0
> last_changed 0.000000
> created 0.000000
> 0: 10.18.176.179:6789/0 mon.usadc-nasea05
> 1: 10.18.176.180:6789/0 mon.usadc-nasea06
> 2: 10.18.176.181:6789/0 mon.usadc-nasea07
> 3: 10.18.176.182:6789/0 mon.usadc-nasea08
>
>
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com