Hi Karan your cinder.conf looks sensible to me, I have posted mine here:
--- cut --- [DEFAULT] rootwrap_config = /etc/cinder/rootwrap.conf api_paste_confg = /etc/cinder/api-paste.ini iscsi_helper = tgtadm volume_name_template = volume-%s volume_group = cinder-volumes verbose = True auth_strategy = keystone state_path = /var/lib/cinder lock_path = /var/lock/cinder volumes_dir = /var/lib/cinder/volumes volume_driver=cinder.volume.drivers.rbd.RBDDriver rbd_pool=volumes glance_api_version=2 rbd_user=volumes rbd_secret_uuid=e1915277-e3a5-4547-bc9e-xxxxxxx rpc_backend = cinder.openstack.common.rpc.impl_kombu rabbit_host = xxx.yyy.cc rabbit_port = 5672 quota_volumes=20 quota_snapshots=20 debug = False use_syslog = True syslog_log_facility = LOG_LOCAL0 [database] connection = mysql://cinder:x...@xxx.yyy.cc/cinder [keystone_authtoken] # keystone public API auth_protocol = https auth_host = xxx.yyy.cc auth_port = 5000 admin_tenant_name = service admin_user = cinder admin_password =xxx --- cut --- what are the different cinder*.log files telling you? Is /etc/ceph/ceph.conf readable for other processes? (chmod 644 /etc/ceph/ceph.conf) Are the key rings available and readable? good luck jc -- SWITCH Jens-Christian Fischer, Peta Solutions Werdstrasse 2, P.O. Box, 8021 Zurich, Switzerland phone +41 44 268 15 15, direct +41 44 268 15 71 jens-christian.fisc...@switch.ch http://www.switch.ch http://www.switch.ch/socialmedia On 27.11.2013, at 08:51, Karan Singh <ksi...@csc.fi> wrote: > Hello Sebastien / Community > > > I tried the commands mentioned in below email. > > > [root@rdo ~]# > [root@rdo ~]# cinder create 1 > +---------------------+--------------------------------------+ > | Property | Value | > +---------------------+--------------------------------------+ > | attachments | [] | > | availability_zone | nova | > | bootable | false | > | created_at | 2013-11-27T07:40:54.161478 | > | display_description | None | > | display_name | None | > | id | ae8cd686-5f1d-4c05-8c42-cb7622122a3e | > | metadata | {} | > | size | 1 | > | snapshot_id | None | > | source_volid | None | > | status | creating | > | volume_type | None | > +---------------------+--------------------------------------+ > [root@rdo ~]# > [root@rdo ~]# cinder list > +--------------------------------------+--------+--------------+------+-------------+----------+-------------+ > | ID | Status | Display Name | Size | > Volume Type | Bootable | Attached to | > +--------------------------------------+--------+--------------+------+-------------+----------+-------------+ > | ae8cd686-5f1d-4c05-8c42-cb7622122a3e | error | None | 1 | > None | false | | > +--------------------------------------+--------+--------------+------+-------------+----------+-------------+ > [root@rdo ~]# > [root@rdo ~]# > [root@rdo ~]# > [root@rdo ~]# rbd -p ceph-volumes ls > rbd: pool ceph-volumes doesn't contain rbd images > [root@rdo ~]# > [root@rdo ~]# > [root@rdo ~]# rados lspools > data > metadata > rbd > ceph-images > ceph-volumes > [root@rdo ~]# rbd -p rbd ls > [root@rdo ~]# rbd -p data ls > foo > foo1 > [root@rdo ~]# > > > > > I checked in cinder.log and got the below errors. > > > 2013-11-27 09:44:14.830 3273 INFO cinder.volume.manager [-] Updating volume > status > 2013-11-27 09:44:14.830 3273 WARNING cinder.volume.manager [-] Unable to > update stats, driver is uninitialized > 2013-11-27 09:44:42.407 12007 INFO cinder.volume.manager [-] Updating volume > status > 2013-11-27 09:44:42.408 12007 WARNING cinder.volume.manager [-] Unable to > update stats, driver is uninitialized > 2013-11-27 09:44:51.799 4943 INFO cinder.volume.manager [-] Updating volume > status > 2013-11-27 09:44:51.799 4943 WARNING cinder.volume.manager [-] Unable to > update stats, driver is uninitialized > 2013-11-27 09:45:14.834 3273 INFO cinder.volume.manager [-] Updating volume > status > 2013-11-27 09:45:14.834 3273 WARNING cinder.volume.manager [-] Unable to > update stats, driver is uninitialized > [root@rdo cinder]# > > > > > Output from my cinder.conf file > > > > # Options defined in cinder.volume.utils > # > > # The default block size used when copying/clearing volumes > # (string value) > #volume_dd_blocksize=1M > > > # Total option count: 382 > volume_driver=cinder.volume.drivers.rbd.RBDDriver > rbd_pool=ceph-volumes > glance_api_version=2 > rbd_user=volumes > rbd_secret_uuid=801a42ec-aec1-3ea8-d869-823c2de56b83 > > rootwrap_config=/etc/cinder/rootwrap.conf > sql_connection = mysql://cinder:root@localhost/cinder > api_paste_config = /etc/cinder/api-paste.ini > > iscsi_helper=tgtadm > volume_name_template = volume-%s > volume_group = cinder-volumes > verbose = True > auth_strategy = keystone > #osapi_volume_listen_port=5900 > [keystone_authtoken] > admin_tenant_name = admin > admin_user = admin > admin_password = root@123 > auth_host = 192.168.1.49 > auth_port = 35357 > auth_protocol = http > signing_dirname = /tmp/keystone-signing-cinder > [root@rdo cinder]# > > > > Can you suggest what is wrong with this configuration , i followed everything > mentioned on http://ceph.com/docs/master/rbd/rbd-openstack/ and there is > not much information available on internet WRT this error. > > Please help. > > > > Many Thanks > Karan Singh > > > ----- Original Message ----- > From: "Sebastien Han" <sebastien....@enovance.com> > To: "Karan Singh" <ksi...@csc.fi> > Cc: ceph-users@lists.ceph.com > Sent: Wednesday, 27 November, 2013 1:08:35 AM > Subject: Re: [ceph-users] how to Testing cinder and glance with CEPH > > Hi, > > Well after restarting the services run: > > $ cinder create 1 > > Then you can check both status in Cinder and Ceph: > > For Cinder run: > $ cinder list > > For Ceph run: > $ rbd -p <cinder-pool> ls > > If the image is there, you’re good. > > Cheers. > > –––– > Sébastien Han > Cloud Engineer > > "Always give 100%. Unless you're giving blood.” > > Phone: +33 (0)1 49 70 99 72 > Mail: sebastien....@enovance.com > Address : 10, rue de la Victoire - 75009 Paris > Web : www.enovance.com - Twitter : @enovance > > On 27 Nov 2013, at 00:04, Karan Singh <ksi...@csc.fi> wrote: > >> Hello Cephers >> >> I was following http://ceph.com/docs/master/rbd/rbd-openstack/ for ceph and >> openstack Integration , using this document ih ave done all the changes >> required for this integration. >> >> I am not sure how should i test my configuration , how should i make sure >> integration is successful. Can you suggest some test that i can perform to >> check my ceph and openstack integration . >> >> FYI , in the document http://ceph.com/docs/master/rbd/rbd-openstack/ , >> Nothing is mentioned after " Restart Openstack Services " heading , but >> there should be steps to test this inttegration , please suggest me here , i >> am new to openstack great if you can give me some commanes used for testing. >> >> >> >> Karan Singh >> CSC - IT Center for Science Ltd. >> P.O. Box 405, FI-02101 Espoo, FINLAND >> http://www.csc.fi/ | +358 (0) 503 812758 >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com