Hi Daniel,
indeed, this is the root cause, but I do not understand what's wrong
here and the real cause of the failure. Since I was suspicious about the
ceph.conf setup I've already checked that it all works, and it does
indeed. I can issue ceph -s or rados df without any problem.
Here I give you a couple of examples:
$ ceph -s
cluster aac2c2c4-5953-44d7-b90c-9922a8ccd77a
health HEALTH_OK
monmap e4: 3 mons at
{mon1=<ipmon1>:6789/0,mon2=<ipmon1>:6789/0,mon3=<ipmon3>:6789/0}
election epoch 124, quorum 0,1,2 mon3,mon2,mon1
fsmap e42: 1/1/1 up {0=mds1=up:active}, 1 up:standby
mgr no daemons active
osdmap e16477: 54 osds: 52 up, 52 in
flags sortbitwise,require_jewel_osds,require_kraken_osds
pgmap v656338: 800 pgs, 16 pools, 4136 MB data, 1326 objects
48076 MB used, 71839 GB / 71886 GB avail
800 active+clean
$ rados df
POOL_NAME USED OBJECTS CLONES COPIES
MISSING_ON_PRIMARY UNFOUND DEGRAED RD_OPS RD WR_OPS WR
.rgw.root 1681 4 0 12
0 0 0 156 121k 4 5120
cephfs_data 0 0 0 0
0 0 0 0 0 0 0
cephfs_metadata 2148 20 0 60
0 0 0 23 24576 41 7168
default.rgw.buckets.data 4136M 1092 0 3276
0 0 0 150 124k 9745 4136M
default.rgw.buckets.index 0 2 0 6
0 0 0 796 1264k 442 0
default.rgw.buckets.non-ec 0 0 0 0
0 0 0 161 161k 128 0
default.rgw.control 0 8 0 24
0 0 0 0 0 0 0
default.rgw.data.root 1228 4 0 12
0 0 0 27 24576 61 15360
default.rgw.gc 0 32 0 96
0 0 0 6620 6588k 4416 0
default.rgw.lc 0 32 0 96
0 0 0 894 862k 448 0
default.rgw.log 0 128 0 384
0 0 0 151070 147M 100778 0
default.rgw.users.keys 11 1 0 3
0 0 0 21 14336 1 1024
default.rgw.users.uid 358 2 0 6
0 0 0 4507 4503k 4385 1024
kraken-test 0 1 0 3
0 0 0 652378 199G 371644 507G
rbd 0 0 0 0
0 0 0 1309 5196M 2518 5000M
scbench 0 0 0 0
0 0 0 1154 4608M 3458 4608M
Any other hint? Of course, having more verbosity from the rados init
would be very helpful, but even with FULL_DEBUG I always get the same
messages and nothing more.
Thanks,
Alessandro
Il 13/01/17 19:37, Daniel Gryniewicz ha scritto:
> Hi, Alessandro.
>
> This error (-5) is caused by the failure to initialize the RADOS client
> in librados. Can you perform ceph operations from that same host?
> (say, ceph -s) It's likely to be a problem in your ceph.conf, I think,
> such as wrong or unreachable MON addresses.
>
> Daniel
>
> On 01/13/2017 12:39 PM, Alessandro De Salvo wrote:
>> Hi,
>> I'm trying to use the RGW FSAL on CentOS 7 with ceph kraken v11.1.1 and
>> ganesha 2.4.1-2. I have rebuilt the RPMS from the rawhide fedora
>> version, who is now including the RGW FSAL. When trying to run the
>> ganesha daemon I get the following error:
>>
>> 13/01/2017 17:21:15 : epoch 58790c88 : node1 : ganesha.nfsd-1[main]
>> init :FSAL :DEBUG :RGW module registering.
>> 13/01/2017 17:21:15 : epoch 58790c88 : node1 : ganesha.nfsd-1[main]
>> init_config :FSAL :DEBUG :RGW module setup.
>> 13/01/2017 17:21:15 : epoch 58790c88 : node1 : ganesha.nfsd-1[main]
>> create_export :FSAL :CRIT :RGW module: librgw init failed (-5)
>> 13/01/2017 17:21:15 : epoch 58790c88 : node1 : ganesha.nfsd-1[main]
>> fsal_put :FSAL :INFO :FSAL RGW now unused
>>
>>
>> The daemon is run in a privileged docker container with
>>
>> /usr/bin/ganesha.nfsd -F -N NIV_DEBUG -L /var/log/ganesha.log
>> -f /etc/ganesha/ganesha.conf
>>
>> All the ceph.conf and keyrings are properly installed in the machine and
>> container, and in fact I can access the ceph cluster correctly and the
>> RGW instance.
>> The ganesha configuration is the following:
>>
>> EXPORT
>> {
>> Export_ID=1;
>>
>> Path = "/atlas";
>>
>> Pseudo = "/atlas";
>>
>> Access_Type = RW;
>>
>> SecType = "sys";
>>
>> FSAL {
>> Name = RGW;
>> User_Id = "testuser";
>> Access_Key_Id ="testkey";
>> Secret_Access_Key = "testsecret";
>> }
>> }
>>
>> RGW {
>> ceph_conf = "/etc/ceph/ceph.conf";
>> name = "auth.node2.roma1.infn.it";
>> cluster = "ceph";
>> init_args = "-d --debug-rgw=16";
>> }
>>
>>
>> Does anyone know how to increase the verbosity of ganesha in order to
>> have more info on this error? Only "init failed" is not really very
>> helpful.
>> Any hint on this error?
>> Thanks,
>>
>> Alessandro
>>
>>
>> ------------------------------------------------------------------------------
>> Developer Access Program for Intel Xeon Phi Processors
>> Access to Intel Xeon Phi processor-based developer platforms.
>> With one year of Intel Parallel Studio XE.
>> Training and support from Colfax.
>> Order your platform today. http://sdm.link/xeonphi
>> _______________________________________________
>> Nfs-ganesha-devel mailing list
>> [email protected]
>> https://lists.sourceforge.net/lists/listinfo/nfs-ganesha-devel
>>
>
> ------------------------------------------------------------------------------
> Developer Access Program for Intel Xeon Phi Processors
> Access to Intel Xeon Phi processor-based developer platforms.
> With one year of Intel Parallel Studio XE.
> Training and support from Colfax.
> Order your platform today. http://sdm.link/xeonphi
> _______________________________________________
> Nfs-ganesha-devel mailing list
> [email protected]
> https://lists.sourceforge.net/lists/listinfo/nfs-ganesha-devel
------------------------------------------------------------------------------
Developer Access Program for Intel Xeon Phi Processors
Access to Intel Xeon Phi processor-based developer platforms.
With one year of Intel Parallel Studio XE.
Training and support from Colfax.
Order your platform today. http://sdm.link/xeonphi
_______________________________________________
Nfs-ganesha-devel mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/nfs-ganesha-devel