On Thu, Aug 21, 2014 at 8:29 AM, Jens-Christian Fischer
<jens-christian.fisc...@switch.ch> wrote:
> I am working with Cinder Multi Backends on an Icehouse installation and have 
> added another backend (Quobyte) to a previously running Cinder/Ceph 
> installation.
>
> I can now create QuoByte volumes, but no longer any ceph volumes. The 
> cinder-scheduler log get’s an incorrect number for the free size of the 
> volumes pool and disregards the RBD backend as a viable storage system:

I don't know much about Cinder, but given this output:

> 2014-08-21 16:42:49.847 1469 DEBUG 
> cinder.openstack.common.scheduler.filters.capabilities_filter [r...] 
> extra_spec requirement 'rbd' does not match 'quobyte' _satisfies_extra_specs 
> /usr/lib/python2.7/dist-packages/cinder/openstack/common/scheduler/filters/capabilities_filter.py:55
> 2014-08-21 16:42:49.848 1469 DEBUG 
> cinder.openstack.common.scheduler.filters.capabilities_filter [r...] host 
> 'controller@quobyte': free_capacity_gb: 156395.931061 fails resource_type 
> extra_specs requirements host_passes 
> /usr/lib/python2.7/dist-packages/cinder/openstack/common/scheduler/filters/capabilities_filter.py:68
> 2014-08-21 16:42:49.848 1469 WARNING cinder.scheduler.filters.capacity_filter 
> [r...-] Insufficient free space for volume creation (requested / avail): 
> 20/8.0
> 2014-08-21 16:42:49.849 1469 ERROR cinder.scheduler.flows.create_volume [r.] 
> Failed to schedule_create_volume: No valid host was found.

I suspect you'll have better luck on the Openstack mailing list. :)

Although for a random quick guess, I think maybe you need to match the
"rbd" and "rbd-volumes" (from your conf file) strings?
-Greg
Software Engineer #42 @ http://inktank.com | http://ceph.com


>
> here’s our /etc/cinder/cinder.conf
>
> — cut —
> [DEFAULT]
> rootwrap_config = /etc/cinder/rootwrap.conf
> api_paste_confg = /etc/cinder/api-paste.ini
> # iscsi_helper = tgtadm
> volume_name_template = volume-%s
> # volume_group = cinder-volumes
> verbose = True
> auth_strategy = keystone
> state_path = /var/lib/cinder
> lock_path = /var/lock/cinder
> volumes_dir = /var/lib/cinder/volumes
> rabbit_host=10.2.0.10
> use_syslog=False
> api_paste_config=/etc/cinder/api-paste.ini
> glance_num_retries=0
> debug=True
> storage_availability_zone=nova
> glance_api_ssl_compression=False
> glance_api_insecure=False
> rabbit_userid=openstack
> rabbit_use_ssl=False
> log_dir=/var/log/cinder
> osapi_volume_listen=0.0.0.0
> glance_api_servers=1.2.3.4:9292
> rabbit_virtual_host=/
> scheduler_driver=cinder.scheduler.filter_scheduler.FilterScheduler
> default_availability_zone=nova
> rabbit_hosts=10.2.0.10:5672
> control_exchange=openstack
> rabbit_ha_queues=False
> glance_api_version=2
> amqp_durable_queues=False
> rabbit_password=secret
> rabbit_port=5672
> rpc_backend=cinder.openstack.common.rpc.impl_kombu
> enabled_backends=quobyte,rbd
> default_volume_type=rbd
>
> [database]
> idle_timeout=3600
> connection=mysql://cinder:secret@10.2.0.10/cinder
>
> [quobyte]
> quobyte_volume_url=quobyte://hostname.cloud.example.com/openstack-volumes
> volume_driver=cinder.volume.drivers.quobyte.QuobyteDriver
>
> [rbd-volumes]
> volume_backend_name=rbd-volumes
> rbd_pool=volumes
> rbd_flatten_volume_from_snapshot=False
> rbd_user=cinder
> rbd_ceph_conf=/etc/ceph/ceph.conf
> rbd_secret_uuid=1234-5678-ABCD-…-DEF
> rbd_max_clone_depth=5
> volume_driver=cinder.volume.drivers.rbd.RBDDriver
>
> — cut ---
>
> any ideas?
>
> cheers
> Jens-Christian
>
> --
> SWITCH
> Jens-Christian Fischer, Peta Solutions
> Werdstrasse 2, P.O. Box, 8021 Zurich, Switzerland
> phone +41 44 268 15 15, direct +41 44 268 15 71
> jens-christian.fisc...@switch.ch
> http://www.switch.ch
>
> http://www.switch.ch/stories
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to