> It would be nice if glance was clever enough to convert where appropriate.

You're right, and it looks like that was added in the Kilo cycle:
https://review.openstack.org/#/c/159129/


-Chris

On Thu, May 28, 2015 at 3:34 PM, Warren Wang <war...@wangspeed.com> wrote:

> Even though we're using Ceph as a backend, we still use qcow2 images as
> our golden images, since we still have a significant (maybe majority)
> number of users using true ephemeral disks. It would be nice if glance was
> clever enough to convert where appropriate.
>
> Warren
>
> Warren
>
> On Thu, May 28, 2015 at 3:21 PM, Fox, Kevin M <kevin....@pnnl.gov> wrote:
>
>> I've experienced the opposite problem though. Downloading raw images and
>> uploading them to the cloud is very slow. Doing it through qcow2 allows
>> them to be compressed over the slow links. Ideally, the Ceph driver would
>> take a qcow2 and convert it to raw on glance ingest rather then at boot.
>>
>> Thanks,
>> Kevin
>> ________________________________________
>> From: Dmitry Borodaenko [dborodae...@mirantis.com]
>> Sent: Thursday, May 28, 2015 12:10 PM
>> To: David Medberry
>> Cc: openstack-operators@lists.openstack.org
>> Subject: Re: [Openstack-operators] what is the different in use Qcow2 or
>> Raw in Ceph
>>
>> David is right, Ceph implements volume snapshotting at the RBD level,
>> not even RADOS level: whole 2 levels of abstraction above file system.
>> It doesn't matter if it's XFS, BtrFS, Ext4, or VFAT (if Ceph supported
>> VFAT): Ceph RBD takes care of it before individual chunks of an RBD
>> volume are passed to RADOS as objects and get written into the file
>> system as files by an OSD process.
>>
>> The reason Fuel documentation recommends to disable QCOW2 format for
>> images is that RBD does not support QCOW2 disks directly, so Nova and
>> Cinder have to _convert_ a QCOW2 image into RAW format before passing
>> it to QEMU's RBD driver. This means that you end up downloading the
>> QCOW2 image from Ceph to a nova-compute node (first full copy),
>> converting it (second full copy), and uploading the resultant RAW
>> image back to Ceph (third full copy) just to launch a VM or create a
>> volume from an image.
>>
>>
>> On Thu, May 28, 2015 at 8:33 AM, David Medberry <openst...@medberry.net>
>> wrote:
>> > yep. It's at the CEPH level (not the XFS level.)
>> >
>> > On Thu, May 28, 2015 at 8:40 AM, Stephen Cousins <
>> steve.cous...@maine.edu>
>> > wrote:
>> >>
>> >> Hi David,
>> >>
>> >> So Ceph will use Copy-on-write even with XFS?
>> >>
>> >> Thanks,
>> >>
>> >> Steve
>> >>
>> >> On Thu, May 28, 2015 at 10:36 AM, David Medberry <
>> openst...@medberry.net>
>> >> wrote:
>> >>>
>> >>> This isn't remotely related to btrfs. It works fine with XFS. Not sure
>> >>> how that works in Fuel, never used it.
>> >>>
>> >>> On Thu, May 28, 2015 at 8:01 AM, Forrest Flagg <
>> fostro.fl...@gmail.com>
>> >>> wrote:
>> >>>>
>> >>>> I'm also curious about this.  Here are some other pieces of
>> information
>> >>>> relevant to the discussion.  Maybe someone here can clear this up
>> for me as
>> >>>> well.  The documentation for Fuel 6.0, not sure what they changed
>> for 6.1,
>> >>>> [1] states that when using Ceph one should disable qcow2 so that
>> images are
>> >>>> stored in raw format.  This is due to the fact that Ceph includes
>> its own
>> >>>> mechanisms for copy-on-write and snapshots.  According to the Ceph
>> >>>> documentation [2], this is true only when using a BTRFS file system,
>> but in
>> >>>> Fuel 6.0 Ceph uses XFS which doesn't provide this functionality.
>> Also, [2]
>> >>>> recommends not using BTRFS for production as it isn't considered
>> fully
>> >>>> mature.  In addition, Fuel 6.0 [3] states that OpenStack with raw
>> images
>> >>>> doesn't support snapshotting.
>> >>>>
>> >>>> Given this, why does Fuel suggest not using qcow2 with Ceph?  How can
>> >>>> Ceph be useful if snapshotting isn't an option with raw images and
>> qcow2
>> >>>> isn't recommended?  Are there other factors to take into
>> consideration that
>> >>>> I'm missing?
>> >>>>
>> >>>> [1]
>> >>>>
>> https://docs.mirantis.com/openstack/fuel/fuel-6.0/terminology.html#qcow2
>> >>>> [2]
>> >>>>
>> http://ceph.com/docs/master/rados/configuration/filesystem-recommendations/
>> >>>> [3]
>> >>>>
>> https://docs.mirantis.com/openstack/fuel/fuel-6.0/user-guide.html#qcow-format-ug
>> >>>>
>> >>>> Thanks,
>> >>>>
>> >>>> Forrest
>> >>>>
>> >>>> On Thu, May 28, 2015 at 8:02 AM, David Medberry <
>> openst...@medberry.net>
>> >>>> wrote:
>> >>>>>
>> >>>>> and better explained here:
>> >>>>> http://ceph.com/docs/master/rbd/qemu-rbd/
>> >>>>>
>> >>>>> On Thu, May 28, 2015 at 6:02 AM, David Medberry
>> >>>>> <openst...@medberry.net> wrote:
>> >>>>>>
>> >>>>>> The primary difference is the ability for CEPH to make zero byte
>> >>>>>> copies. When you use qcow2, ceph must actually create a complete
>> copy
>> >>>>>> instead of a zero byte copy as it cannot do its own copy-on-write
>> tricks
>> >>>>>> with a qcow2 image.
>> >>>>>>
>> >>>>>> So, yes, it will work fine with qcow2 images but it won't be as
>> >>>>>> performant as it is with RAW. Also, it will actually use more of
>> the native
>> >>>>>> underlying storage.
>> >>>>>>
>> >>>>>> This is also shown as an Important Note in the CEPH docs:
>> >>>>>> http://ceph.com/docs/master/rbd/rbd-openstack/
>> >>>>>>
>> >>>>>> On Thu, May 28, 2015 at 4:12 AM, Shake Chen <shake.c...@gmail.com>
>> >>>>>> wrote:
>> >>>>>>>
>> >>>>>>> Hi
>> >>>>>>>
>> >>>>>>> Now I try to use Fuel 6.1 deploy openstack Juno, use Ceph as
>> cinder,
>> >>>>>>> nova and glance backend.
>> >>>>>>>
>> >>>>>>> In Fuel document suggest if use ceph, suggest use RAW format
>> image.
>> >>>>>>>
>> >>>>>>> but if I upload qcow2 image, seem working well.
>> >>>>>>>
>> >>>>>>> what is the different use qcow2 and RAW in Ceph?
>> >>>>>>>
>> >>>>>>>
>> >>>>>>> --
>> >>>>>>> Shake Chen
>> >>>>>>>
>> >>>>>>>
>> >>>>>>> _______________________________________________
>> >>>>>>> OpenStack-operators mailing list
>> >>>>>>> OpenStack-operators@lists.openstack.org
>> >>>>>>>
>> >>>>>>>
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>> >>>>>>>
>> >>>>>>
>> >>>>>
>> >>>>>
>> >>>>> _______________________________________________
>> >>>>> OpenStack-operators mailing list
>> >>>>> OpenStack-operators@lists.openstack.org
>> >>>>>
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>> >>>>>
>> >>>>
>> >>>
>> >>>
>> >>> _______________________________________________
>> >>> OpenStack-operators mailing list
>> >>> OpenStack-operators@lists.openstack.org
>> >>>
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>> >>>
>> >>
>> >>
>> >>
>> >> --
>> >> ________________________________________________________________
>> >>  Steve Cousins             Supercomputer Engineer/Administrator
>> >>  Advanced Computing Group            University of Maine System
>> >>  244 Neville Hall (UMS Data Center)              (207) 561-3574
>> >>  Orono ME 04469                      steve.cousins at maine.edu
>> >>
>> >
>> >
>> > _______________________________________________
>> > OpenStack-operators mailing list
>> > OpenStack-operators@lists.openstack.org
>> > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>> >
>>
>>
>>
>> --
>> Dmitry Borodaenko
>>
>> _______________________________________________
>> OpenStack-operators mailing list
>> OpenStack-operators@lists.openstack.org
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>
>> _______________________________________________
>> OpenStack-operators mailing list
>> OpenStack-operators@lists.openstack.org
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>
>
>
> _______________________________________________
> OpenStack-operators mailing list
> OpenStack-operators@lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>
>
_______________________________________________
OpenStack-operators mailing list
OpenStack-operators@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators

Reply via email to