The lack of verbosity is not Ganesha's fault; it only gets the single error code back from Ceph.
Try turning up all your client related logging in your ceph.conf, and check it's logging? Daniel On 01/14/2017 03:02 PM, Alessandro De Salvo wrote: > Hi Daniel, > > indeed, this is the root cause, but I do not understand what's wrong > here and the real cause of the failure. Since I was suspicious about the > ceph.conf setup I've already checked that it all works, and it does > indeed. I can issue ceph -s or rados df without any problem. > > Here I give you a couple of examples: > > > $ ceph -s > > cluster aac2c2c4-5953-44d7-b90c-9922a8ccd77a > health HEALTH_OK > monmap e4: 3 mons at > {mon1=<ipmon1>:6789/0,mon2=<ipmon1>:6789/0,mon3=<ipmon3>:6789/0} > election epoch 124, quorum 0,1,2 mon3,mon2,mon1 > fsmap e42: 1/1/1 up {0=mds1=up:active}, 1 up:standby > mgr no daemons active > osdmap e16477: 54 osds: 52 up, 52 in > flags sortbitwise,require_jewel_osds,require_kraken_osds > pgmap v656338: 800 pgs, 16 pools, 4136 MB data, 1326 objects > 48076 MB used, 71839 GB / 71886 GB avail > 800 active+clean > > > $ rados df > > POOL_NAME USED OBJECTS CLONES COPIES > MISSING_ON_PRIMARY UNFOUND DEGRAED RD_OPS RD WR_OPS WR > .rgw.root 1681 4 0 12 > 0 0 0 156 121k 4 5120 > cephfs_data 0 0 0 0 > 0 0 0 0 0 0 0 > cephfs_metadata 2148 20 0 60 > 0 0 0 23 24576 41 7168 > default.rgw.buckets.data 4136M 1092 0 3276 > 0 0 0 150 124k 9745 4136M > default.rgw.buckets.index 0 2 0 6 > 0 0 0 796 1264k 442 0 > default.rgw.buckets.non-ec 0 0 0 0 > 0 0 0 161 161k 128 0 > default.rgw.control 0 8 0 24 > 0 0 0 0 0 0 0 > default.rgw.data.root 1228 4 0 12 > 0 0 0 27 24576 61 15360 > default.rgw.gc 0 32 0 96 > 0 0 0 6620 6588k 4416 0 > default.rgw.lc 0 32 0 96 > 0 0 0 894 862k 448 0 > default.rgw.log 0 128 0 384 > 0 0 0 151070 147M 100778 0 > default.rgw.users.keys 11 1 0 3 > 0 0 0 21 14336 1 1024 > default.rgw.users.uid 358 2 0 6 > 0 0 0 4507 4503k 4385 1024 > kraken-test 0 1 0 3 > 0 0 0 652378 199G 371644 507G > rbd 0 0 0 0 > 0 0 0 1309 5196M 2518 5000M > scbench 0 0 0 0 > 0 0 0 1154 4608M 3458 4608M > > > Any other hint? Of course, having more verbosity from the rados init > would be very helpful, but even with FULL_DEBUG I always get the same > messages and nothing more. > Thanks, > > Alessandro > > Il 13/01/17 19:37, Daniel Gryniewicz ha scritto: >> Hi, Alessandro. >> >> This error (-5) is caused by the failure to initialize the RADOS client >> in librados. Can you perform ceph operations from that same host? >> (say, ceph -s) It's likely to be a problem in your ceph.conf, I think, >> such as wrong or unreachable MON addresses. >> >> Daniel >> >> On 01/13/2017 12:39 PM, Alessandro De Salvo wrote: >>> Hi, >>> I'm trying to use the RGW FSAL on CentOS 7 with ceph kraken v11.1.1 and >>> ganesha 2.4.1-2. I have rebuilt the RPMS from the rawhide fedora >>> version, who is now including the RGW FSAL. When trying to run the >>> ganesha daemon I get the following error: >>> >>> 13/01/2017 17:21:15 : epoch 58790c88 : node1 : ganesha.nfsd-1[main] >>> init :FSAL :DEBUG :RGW module registering. >>> 13/01/2017 17:21:15 : epoch 58790c88 : node1 : ganesha.nfsd-1[main] >>> init_config :FSAL :DEBUG :RGW module setup. >>> 13/01/2017 17:21:15 : epoch 58790c88 : node1 : ganesha.nfsd-1[main] >>> create_export :FSAL :CRIT :RGW module: librgw init failed (-5) >>> 13/01/2017 17:21:15 : epoch 58790c88 : node1 : ganesha.nfsd-1[main] >>> fsal_put :FSAL :INFO :FSAL RGW now unused >>> >>> >>> The daemon is run in a privileged docker container with >>> >>> /usr/bin/ganesha.nfsd -F -N NIV_DEBUG -L /var/log/ganesha.log >>> -f /etc/ganesha/ganesha.conf >>> >>> All the ceph.conf and keyrings are properly installed in the machine and >>> container, and in fact I can access the ceph cluster correctly and the >>> RGW instance. >>> The ganesha configuration is the following: >>> >>> EXPORT >>> { >>> Export_ID=1; >>> >>> Path = "/atlas"; >>> >>> Pseudo = "/atlas"; >>> >>> Access_Type = RW; >>> >>> SecType = "sys"; >>> >>> FSAL { >>> Name = RGW; >>> User_Id = "testuser"; >>> Access_Key_Id ="testkey"; >>> Secret_Access_Key = "testsecret"; >>> } >>> } >>> >>> RGW { >>> ceph_conf = "/etc/ceph/ceph.conf"; >>> name = "auth.node2.roma1.infn.it"; >>> cluster = "ceph"; >>> init_args = "-d --debug-rgw=16"; >>> } >>> >>> >>> Does anyone know how to increase the verbosity of ganesha in order to >>> have more info on this error? Only "init failed" is not really very >>> helpful. >>> Any hint on this error? >>> Thanks, >>> >>> Alessandro >>> >>> >>> ------------------------------------------------------------------------------ >>> Developer Access Program for Intel Xeon Phi Processors >>> Access to Intel Xeon Phi processor-based developer platforms. >>> With one year of Intel Parallel Studio XE. >>> Training and support from Colfax. >>> Order your platform today. http://sdm.link/xeonphi >>> _______________________________________________ >>> Nfs-ganesha-devel mailing list >>> Nfs-ganesha-devel@lists.sourceforge.net >>> https://lists.sourceforge.net/lists/listinfo/nfs-ganesha-devel >>> >> >> ------------------------------------------------------------------------------ >> Developer Access Program for Intel Xeon Phi Processors >> Access to Intel Xeon Phi processor-based developer platforms. >> With one year of Intel Parallel Studio XE. >> Training and support from Colfax. >> Order your platform today. http://sdm.link/xeonphi >> _______________________________________________ >> Nfs-ganesha-devel mailing list >> Nfs-ganesha-devel@lists.sourceforge.net >> https://lists.sourceforge.net/lists/listinfo/nfs-ganesha-devel > > > ------------------------------------------------------------------------------ > Developer Access Program for Intel Xeon Phi Processors > Access to Intel Xeon Phi processor-based developer platforms. > With one year of Intel Parallel Studio XE. > Training and support from Colfax. > Order your platform today. http://sdm.link/xeonphi > _______________________________________________ > Nfs-ganesha-devel mailing list > Nfs-ganesha-devel@lists.sourceforge.net > https://lists.sourceforge.net/lists/listinfo/nfs-ganesha-devel > ------------------------------------------------------------------------------ Developer Access Program for Intel Xeon Phi Processors Access to Intel Xeon Phi processor-based developer platforms. With one year of Intel Parallel Studio XE. Training and support from Colfax. Order your platform today. http://sdm.link/xeonphi _______________________________________________ Nfs-ganesha-devel mailing list Nfs-ganesha-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/nfs-ganesha-devel