On Thu, May 28, 2015 at 3:21 PM, Fox, Kevin M <kevin....@pnnl.gov> wrote: > I've experienced the opposite problem though. Downloading raw images and > uploading them to the cloud is very slow. Doing it through qcow2 allows them > to be compressed over the slow links. Ideally, the Ceph driver would take a > qcow2 and convert it to raw on glance ingest rather then at boot.
There is a blueprint for this which if I read correctly has landed in kilo (I'm still on juno and haven't started kilo testing yet): https://blueprints.launchpad.net/glance/+spec/basic-import-conversion https://review.openstack.org/#/c/159129/ I hope that it's landed because I too am really looking forward to that feature -Jon > Thanks, > Kevin > ________________________________________ > From: Dmitry Borodaenko [dborodae...@mirantis.com] > Sent: Thursday, May 28, 2015 12:10 PM > To: David Medberry > Cc: openstack-operators@lists.openstack.org > Subject: Re: [Openstack-operators] what is the different in use Qcow2 or Raw > in Ceph > > David is right, Ceph implements volume snapshotting at the RBD level, > not even RADOS level: whole 2 levels of abstraction above file system. > It doesn't matter if it's XFS, BtrFS, Ext4, or VFAT (if Ceph supported > VFAT): Ceph RBD takes care of it before individual chunks of an RBD > volume are passed to RADOS as objects and get written into the file > system as files by an OSD process. > > The reason Fuel documentation recommends to disable QCOW2 format for > images is that RBD does not support QCOW2 disks directly, so Nova and > Cinder have to _convert_ a QCOW2 image into RAW format before passing > it to QEMU's RBD driver. This means that you end up downloading the > QCOW2 image from Ceph to a nova-compute node (first full copy), > converting it (second full copy), and uploading the resultant RAW > image back to Ceph (third full copy) just to launch a VM or create a > volume from an image. > > > On Thu, May 28, 2015 at 8:33 AM, David Medberry <openst...@medberry.net> > wrote: >> yep. It's at the CEPH level (not the XFS level.) >> >> On Thu, May 28, 2015 at 8:40 AM, Stephen Cousins <steve.cous...@maine.edu> >> wrote: >>> >>> Hi David, >>> >>> So Ceph will use Copy-on-write even with XFS? >>> >>> Thanks, >>> >>> Steve >>> >>> On Thu, May 28, 2015 at 10:36 AM, David Medberry <openst...@medberry.net> >>> wrote: >>>> >>>> This isn't remotely related to btrfs. It works fine with XFS. Not sure >>>> how that works in Fuel, never used it. >>>> >>>> On Thu, May 28, 2015 at 8:01 AM, Forrest Flagg <fostro.fl...@gmail.com> >>>> wrote: >>>>> >>>>> I'm also curious about this. Here are some other pieces of information >>>>> relevant to the discussion. Maybe someone here can clear this up for me >>>>> as >>>>> well. The documentation for Fuel 6.0, not sure what they changed for 6.1, >>>>> [1] states that when using Ceph one should disable qcow2 so that images >>>>> are >>>>> stored in raw format. This is due to the fact that Ceph includes its own >>>>> mechanisms for copy-on-write and snapshots. According to the Ceph >>>>> documentation [2], this is true only when using a BTRFS file system, but >>>>> in >>>>> Fuel 6.0 Ceph uses XFS which doesn't provide this functionality. Also, >>>>> [2] >>>>> recommends not using BTRFS for production as it isn't considered fully >>>>> mature. In addition, Fuel 6.0 [3] states that OpenStack with raw images >>>>> doesn't support snapshotting. >>>>> >>>>> Given this, why does Fuel suggest not using qcow2 with Ceph? How can >>>>> Ceph be useful if snapshotting isn't an option with raw images and qcow2 >>>>> isn't recommended? Are there other factors to take into consideration >>>>> that >>>>> I'm missing? >>>>> >>>>> [1] >>>>> https://docs.mirantis.com/openstack/fuel/fuel-6.0/terminology.html#qcow2 >>>>> [2] >>>>> http://ceph.com/docs/master/rados/configuration/filesystem-recommendations/ >>>>> [3] >>>>> https://docs.mirantis.com/openstack/fuel/fuel-6.0/user-guide.html#qcow-format-ug >>>>> >>>>> Thanks, >>>>> >>>>> Forrest >>>>> >>>>> On Thu, May 28, 2015 at 8:02 AM, David Medberry <openst...@medberry.net> >>>>> wrote: >>>>>> >>>>>> and better explained here: >>>>>> http://ceph.com/docs/master/rbd/qemu-rbd/ >>>>>> >>>>>> On Thu, May 28, 2015 at 6:02 AM, David Medberry >>>>>> <openst...@medberry.net> wrote: >>>>>>> >>>>>>> The primary difference is the ability for CEPH to make zero byte >>>>>>> copies. When you use qcow2, ceph must actually create a complete copy >>>>>>> instead of a zero byte copy as it cannot do its own copy-on-write tricks >>>>>>> with a qcow2 image. >>>>>>> >>>>>>> So, yes, it will work fine with qcow2 images but it won't be as >>>>>>> performant as it is with RAW. Also, it will actually use more of the >>>>>>> native >>>>>>> underlying storage. >>>>>>> >>>>>>> This is also shown as an Important Note in the CEPH docs: >>>>>>> http://ceph.com/docs/master/rbd/rbd-openstack/ >>>>>>> >>>>>>> On Thu, May 28, 2015 at 4:12 AM, Shake Chen <shake.c...@gmail.com> >>>>>>> wrote: >>>>>>>> >>>>>>>> Hi >>>>>>>> >>>>>>>> Now I try to use Fuel 6.1 deploy openstack Juno, use Ceph as cinder, >>>>>>>> nova and glance backend. >>>>>>>> >>>>>>>> In Fuel document suggest if use ceph, suggest use RAW format image. >>>>>>>> >>>>>>>> but if I upload qcow2 image, seem working well. >>>>>>>> >>>>>>>> what is the different use qcow2 and RAW in Ceph? >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Shake Chen >>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> OpenStack-operators mailing list >>>>>>>> OpenStack-operators@lists.openstack.org >>>>>>>> >>>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators >>>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> OpenStack-operators mailing list >>>>>> OpenStack-operators@lists.openstack.org >>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators >>>>>> >>>>> >>>> >>>> >>>> _______________________________________________ >>>> OpenStack-operators mailing list >>>> OpenStack-operators@lists.openstack.org >>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators >>>> >>> >>> >>> >>> -- >>> ________________________________________________________________ >>> Steve Cousins Supercomputer Engineer/Administrator >>> Advanced Computing Group University of Maine System >>> 244 Neville Hall (UMS Data Center) (207) 561-3574 >>> Orono ME 04469 steve.cousins at maine.edu >>> >> >> >> _______________________________________________ >> OpenStack-operators mailing list >> OpenStack-operators@lists.openstack.org >> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators >> > > > > -- > Dmitry Borodaenko > > _______________________________________________ > OpenStack-operators mailing list > OpenStack-operators@lists.openstack.org > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators > > _______________________________________________ > OpenStack-operators mailing list > OpenStack-operators@lists.openstack.org > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators _______________________________________________ OpenStack-operators mailing list OpenStack-operators@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators