Hi Daniel, indeed, this is the root cause, but I do not understand what's wrong here and the real cause of the failure. Since I was suspicious about the ceph.conf setup I've already checked that it all works, and it does indeed. I can issue ceph -s or rados df without any problem.
Here I give you a couple of examples: $ ceph -s cluster aac2c2c4-5953-44d7-b90c-9922a8ccd77a health HEALTH_OK monmap e4: 3 mons at {mon1=<ipmon1>:6789/0,mon2=<ipmon1>:6789/0,mon3=<ipmon3>:6789/0} election epoch 124, quorum 0,1,2 mon3,mon2,mon1 fsmap e42: 1/1/1 up {0=mds1=up:active}, 1 up:standby mgr no daemons active osdmap e16477: 54 osds: 52 up, 52 in flags sortbitwise,require_jewel_osds,require_kraken_osds pgmap v656338: 800 pgs, 16 pools, 4136 MB data, 1326 objects 48076 MB used, 71839 GB / 71886 GB avail 800 active+clean $ rados df POOL_NAME USED OBJECTS CLONES COPIES MISSING_ON_PRIMARY UNFOUND DEGRAED RD_OPS RD WR_OPS WR .rgw.root 1681 4 0 12 0 0 0 156 121k 4 5120 cephfs_data 0 0 0 0 0 0 0 0 0 0 0 cephfs_metadata 2148 20 0 60 0 0 0 23 24576 41 7168 default.rgw.buckets.data 4136M 1092 0 3276 0 0 0 150 124k 9745 4136M default.rgw.buckets.index 0 2 0 6 0 0 0 796 1264k 442 0 default.rgw.buckets.non-ec 0 0 0 0 0 0 0 161 161k 128 0 default.rgw.control 0 8 0 24 0 0 0 0 0 0 0 default.rgw.data.root 1228 4 0 12 0 0 0 27 24576 61 15360 default.rgw.gc 0 32 0 96 0 0 0 6620 6588k 4416 0 default.rgw.lc 0 32 0 96 0 0 0 894 862k 448 0 default.rgw.log 0 128 0 384 0 0 0 151070 147M 100778 0 default.rgw.users.keys 11 1 0 3 0 0 0 21 14336 1 1024 default.rgw.users.uid 358 2 0 6 0 0 0 4507 4503k 4385 1024 kraken-test 0 1 0 3 0 0 0 652378 199G 371644 507G rbd 0 0 0 0 0 0 0 1309 5196M 2518 5000M scbench 0 0 0 0 0 0 0 1154 4608M 3458 4608M Any other hint? Of course, having more verbosity from the rados init would be very helpful, but even with FULL_DEBUG I always get the same messages and nothing more. Thanks, Alessandro Il 13/01/17 19:37, Daniel Gryniewicz ha scritto: > Hi, Alessandro. > > This error (-5) is caused by the failure to initialize the RADOS client > in librados. Can you perform ceph operations from that same host? > (say, ceph -s) It's likely to be a problem in your ceph.conf, I think, > such as wrong or unreachable MON addresses. > > Daniel > > On 01/13/2017 12:39 PM, Alessandro De Salvo wrote: >> Hi, >> I'm trying to use the RGW FSAL on CentOS 7 with ceph kraken v11.1.1 and >> ganesha 2.4.1-2. I have rebuilt the RPMS from the rawhide fedora >> version, who is now including the RGW FSAL. When trying to run the >> ganesha daemon I get the following error: >> >> 13/01/2017 17:21:15 : epoch 58790c88 : node1 : ganesha.nfsd-1[main] >> init :FSAL :DEBUG :RGW module registering. >> 13/01/2017 17:21:15 : epoch 58790c88 : node1 : ganesha.nfsd-1[main] >> init_config :FSAL :DEBUG :RGW module setup. >> 13/01/2017 17:21:15 : epoch 58790c88 : node1 : ganesha.nfsd-1[main] >> create_export :FSAL :CRIT :RGW module: librgw init failed (-5) >> 13/01/2017 17:21:15 : epoch 58790c88 : node1 : ganesha.nfsd-1[main] >> fsal_put :FSAL :INFO :FSAL RGW now unused >> >> >> The daemon is run in a privileged docker container with >> >> /usr/bin/ganesha.nfsd -F -N NIV_DEBUG -L /var/log/ganesha.log >> -f /etc/ganesha/ganesha.conf >> >> All the ceph.conf and keyrings are properly installed in the machine and >> container, and in fact I can access the ceph cluster correctly and the >> RGW instance. >> The ganesha configuration is the following: >> >> EXPORT >> { >> Export_ID=1; >> >> Path = "/atlas"; >> >> Pseudo = "/atlas"; >> >> Access_Type = RW; >> >> SecType = "sys"; >> >> FSAL { >> Name = RGW; >> User_Id = "testuser"; >> Access_Key_Id ="testkey"; >> Secret_Access_Key = "testsecret"; >> } >> } >> >> RGW { >> ceph_conf = "/etc/ceph/ceph.conf"; >> name = "auth.node2.roma1.infn.it"; >> cluster = "ceph"; >> init_args = "-d --debug-rgw=16"; >> } >> >> >> Does anyone know how to increase the verbosity of ganesha in order to >> have more info on this error? Only "init failed" is not really very >> helpful. >> Any hint on this error? >> Thanks, >> >> Alessandro >> >> >> ------------------------------------------------------------------------------ >> Developer Access Program for Intel Xeon Phi Processors >> Access to Intel Xeon Phi processor-based developer platforms. >> With one year of Intel Parallel Studio XE. >> Training and support from Colfax. >> Order your platform today. http://sdm.link/xeonphi >> _______________________________________________ >> Nfs-ganesha-devel mailing list >> Nfs-ganesha-devel@lists.sourceforge.net >> https://lists.sourceforge.net/lists/listinfo/nfs-ganesha-devel >> > > ------------------------------------------------------------------------------ > Developer Access Program for Intel Xeon Phi Processors > Access to Intel Xeon Phi processor-based developer platforms. > With one year of Intel Parallel Studio XE. > Training and support from Colfax. > Order your platform today. http://sdm.link/xeonphi > _______________________________________________ > Nfs-ganesha-devel mailing list > Nfs-ganesha-devel@lists.sourceforge.net > https://lists.sourceforge.net/lists/listinfo/nfs-ganesha-devel ------------------------------------------------------------------------------ Developer Access Program for Intel Xeon Phi Processors Access to Intel Xeon Phi processor-based developer platforms. With one year of Intel Parallel Studio XE. Training and support from Colfax. Order your platform today. http://sdm.link/xeonphi _______________________________________________ Nfs-ganesha-devel mailing list Nfs-ganesha-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/nfs-ganesha-devel