Re: [ceph-users] Live Migrations with cephFS
On 06/14/2013 08:00 PM, Ilja Maslov wrote: Hi, Is live migration supported with RBD and KVM/OpenStack? Always wanted to know but was afraid to ask :) totally works in my productive setup. but we don't use openstack in this installation, just KVM/RBD. Pardon brevity and formatting, replying from the phone. Cheers, Ilja Robert Sander r.san...@heinlein-support.de wrote: On 14.06.2013 12:55, Alvaro Izquierdo Jimeno wrote: By default, openstack uses NFS but… other options are available….can we use cephFS instead of NFS? Wouldn't you use qemu-rbd for your virtual guests in OpenStack? AFAIK CephFS is not needed for KVM/qemu virtual machines. Regards -- Robert Sander Heinlein Support GmbH Schwedter Str. 8/9b, 10119 Berlin http://www.heinlein-support.de Tel: 030 / 405051-43 Fax: 030 / 405051-19 Zwangsangaben lt. §35a GmbHG: HRB 93818 B / Amtsgericht Berlin-Charlottenburg, Geschäftsführer: Peer Heinlein -- Sitz: Berlin ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com This email and any files transmitted with it are confidential and intended solely for the use of the individual or entity to whom they are addressed. If you are not the intended recipient, please note that any review, dissemination, disclosure, alteration, printing, circulation, retention or transmission of this e-mail and/or any file or attachment transmitted with it, is prohibited and may be unlawful. If you have received this e-mail or any file or attachment transmitted with it in error please notify postmas...@openet.com. Although Openet has taken reasonable precautions to ensure no viruses are present in this email, we cannot accept responsibility for any loss or damage arising from the use of this email or attachments. ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- DI (FH) Wolfgang Hennerbichler Software Development Unit Advanced Computing Technologies RISC Software GmbH A company of the Johannes Kepler University Linz IT-Center Softwarepark 35 4232 Hagenberg Austria Phone: +43 7236 3343 245 Fax: +43 7236 3343 250 wolfgang.hennerbich...@risc-software.at http://www.risc-software.at ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Live Migrations with cephFS
OpenStack grizzly VM can be started on rbd(0.61.3) with no problem. I didn't try live migration though. -- Original -- From: Wolfgang Hennerbichlerwolfgang.hennerbich...@risc-software.at; Date: Mon, Jun 17, 2013 02:00 PM To: ceph-usersceph-users@lists.ceph.com; Subject: Re: [ceph-users] Live Migrations with cephFS On 06/14/2013 08:00 PM, Ilja Maslov wrote: Hi, Is live migration supported with RBD and KVM/OpenStack? Always wanted to know but was afraid to ask :) totally works in my productive setup. but we don't use openstack in this installation, just KVM/RBD. Pardon brevity and formatting, replying from the phone. Cheers, Ilja Robert Sander r.san...@heinlein-support.de wrote: On 14.06.2013 12:55, Alvaro Izquierdo Jimeno wrote: By default, openstack uses NFS but… other options are available….can we use cephFS instead of NFS? Wouldn't you use qemu-rbd for your virtual guests in OpenStack? AFAIK CephFS is not needed for KVM/qemu virtual machines. Regards -- Robert Sander Heinlein Support GmbH Schwedter Str. 8/9b, 10119 Berlin http://www.heinlein-support.de Tel: 030 / 405051-43 Fax: 030 / 405051-19 Zwangsangaben lt. §35a GmbHG: HRB 93818 B / Amtsgericht Berlin-Charlottenburg, Geschäftsführer: Peer Heinlein -- Sitz: Berlin ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com This email and any files transmitted with it are confidential and intended solely for the use of the individual or entity to whom they are addressed. If you are not the intended recipient, please note that any review, dissemination, disclosure, alteration, printing, circulation, retention or transmission of this e-mail and/or any file or attachment transmitted with it, is prohibited and may be unlawful. If you have received this e-mail or any file or attachment transmitted with it in error please notify postmas...@openet.com. Although Openet has taken reasonable precautions to ensure no viruses are present in this email, we cannot accept responsibility for any loss or damage arising from the use of this email or attachments. ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- DI (FH) Wolfgang Hennerbichler Software Development Unit Advanced Computing Technologies RISC Software GmbH A company of the Johannes Kepler University Linz IT-Center Softwarepark 35 4232 Hagenberg Austria Phone: +43 7236 3343 245 Fax: +43 7236 3343 250 wolfgang.hennerbich...@risc-software.at http://www.risc-software.at ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Ceph and open source cloud software: Path of least resistance
On 06/16/2013 07:48 PM, Jens Kristian Søgaard wrote: ... Any advice you could give would be greatly appreciated! hi, I believe openstack is your best bet. I didn't try it with caching enabled, but you might as well write a patch for that, this shouldn't be too hard. anyways, my general opinion on the current cloud stacks is that they support way too much hypervisors, and are therefore slow in development, and they tend to be fragile and hard to configure because of this broad support (and not following the unix philosophy). they should really focus on kvm and backend storages, but this may just be my limited view of the world, and is way too off-topic for this mailinglist... Thanks, Wolfgang -- DI (FH) Wolfgang Hennerbichler Software Development Unit Advanced Computing Technologies RISC Software GmbH A company of the Johannes Kepler University Linz IT-Center Softwarepark 35 4232 Hagenberg Austria Phone: +43 7236 3343 245 Fax: +43 7236 3343 250 wolfgang.hennerbich...@risc-software.at http://www.risc-software.at ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Need help with Ceph error
Hi The issue it hangs when I type any ceph commands.. :( === root@xtream:~# ceph -s ^C root@xtream:~# service ceph start === mds.a === Starting Ceph mds.a on xtream...already running === osd.0 === Mounting xfs on xtream:/var/lib/ceph/osd/ceph-0 ^C root@xtream:~# == Thanks and regards Sreejith KJ -Original Message- From: Gregory Farnum [mailto:g...@inktank.com] Sent: Friday, June 14, 2013 9:07 PM To: Sreejith Keeriyattil Cc: ceph-us...@ceph.com Subject: Re: [ceph-users] Need help with Ceph error On Fri, Jun 14, 2013 at 12:20 AM, Sreejith Keeriyattil sreejith.keeriyat...@happiestminds.com wrote: Hi To keep it simple I disabled cephx authentication but after that am getting the below error. == == root@xtream:/etc/ceph# service ceph -v start /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n mon.a user /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n mds.a user === mds.a === /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n mds.a pid file --- xtream# mkdir -p /var/run/ceph /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n mds.a log dir /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n mds.a auto start --- xtream# [ -e /var/run/ceph/mds.a.pid ] || exit 1 # no pid, presumably not pid=`cat /var/run/ceph/mds.a.pid` [ -e /proc/$pid ] grep -q ceph-mds /proc/$pid/cmdline grep -qwe -i exit 1 # pid is something else Starting Ceph mds.a on xtream...already running /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 user === osd.0 === /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 pid file --- xtream# mkdir -p /var/run/ceph /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 log dir /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 auto start --- xtream# [ -e /var/run/ceph/osd.0.pid ] || exit 1 # no pid, presumably not pid=`cat /var/run/ceph/osd.0.pid` [ -e /proc/$pid ] grep -q ceph-osd /proc/$pid/cmdline grep -qwe -i exit 1 # pid is something else /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 copy executable to /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 osd data /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 fs path /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 devs /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 lock file /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 admin socket /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 max open files /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 restart on core dump /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 valgrind /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 pre mount command /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 osd mkfs type /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 osd mount options xfs --- xtream# true Mounting xfs on xtream:/var/lib/ceph/osd/ceph-0 --- xtream# modprobe xfs ; egrep -q '^[^ ]+ /var/lib/ceph/osd/ceph-0' /proc/moun /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 osd crush update on start /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 osd crush location /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 osd crush initial weight /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 keyring It hangs after this My ceph.conf file looks like this I think you're still having issues connecting to your monitors. Can you run ceph -s and provide the output? -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com Happiest Minds Disclaimer This message is for the sole use of the intended recipient(s) and may contain confidential, proprietary or legally privileged information. Any unauthorized review, use, disclosure or distribution is prohibited. If you are not the original intended recipient of the message, please contact the sender by reply email and destroy all copies of the original message. Happiest Minds Technologies http://www.happiestminds.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] RDB
Hi All, A bit of an update... I should have run the command from my-cluster directory. I am now receiving this error: alphaceph@cephadmin1:~/ceph-deploy/my-cluster$ rbd create fooimage --size 1024 --pool barpool -m cephserver1.zion.bt.co.uk -k /etc/ceph/ceph.client.admin.keyring 2013-06-17 08:55:14.751204 7f1b28950780 -1 monclient(hunting): ERROR: missing keyring, cannot use cephx for authentication 2013-06-17 08:55:14.751212 7f1b28950780 0 librados: client.admin initialization error (2) No such file or directory rbd: couldn't connect to the cluster! More FYI alphaceph@cephadmin1:~/ceph-deploy/my-cluster$ more ceph.client.admin.keyring [client.admin] key = AQCp5rNRkBLCHRAAOqfY/24mkYCQZ7sNy/8BDA== alphaceph@cephadmin1:~/ceph-deploy/my-cluster$ more ceph.conf [global] fsid = 5e29db66-a1f1-4220-aa19-ab82020adc78 mon initial members = cephserver1 mon host = 10.255.40.22 auth supported = cephx osd journal size = 1024 filestore xattr use omap = true osd crush chooseleaf type = 0 Thanks Gary On Mon, Jun 17, 2013 at 8:37 AM, Gary Bruce garyofscotl...@gmail.comwrote: Hi All, I finally got around to progressing with this but immediately got this message. Any thoughts? alphaceph@cephadmin1:~$ rbd create fooimage --size 1024 --pool barpool -m cephserver1.zion.bt.co.uk -k /etc/ceph/ceph.client.admin.keyring 2013-06-17 08:38:43.955683 7f76a6b72780 -1 did not load config file, using default settings. 2013-06-17 08:38:43.962518 7f76a6b72780 -1 monclient(hunting): authenticate NOTE: no keyring found; disabled cephx authentication 2013-06-17 08:38:43.962541 7f76a6b72780 0 librados: client.admin authentication error (95) Operation not supported rbd: couldn't connect to the cluster! FYI... alphaceph@cephserver1:~$ sudo more /etc/ceph/ceph.client.admin.keyring [client.admin] key = AQCp5rNRkBLCHRAAOqfY/24mkYCQZ7sNy/8BDA== alphaceph@cephserver1:~$ sudo more /etc/ceph/ceph.conf [global] fsid = 5e29db66-a1f1-4220-aa19-ab82020adc78 mon_initial_members = cephserver1 mon_host = 10.255.40.22 auth_supported = cephx osd_journal_size = 1024 filestore_xattr_use_omap = true osd_crush_chooseleaf_type = 0 Thanks in advance. Gary On Tue, Jun 11, 2013 at 8:14 PM, John Wilkins john.wilk...@inktank.comwrote: Gary, I've added that instruction to the docs. It should be up shortly. Let me know if you have other feedback for the docs. Regards, John On Mon, Jun 10, 2013 at 9:13 AM, Gary Bruce garyofscotl...@gmail.com wrote: Hi again, I don't see anything in http://ceph.com/docs/master/start/ that mentions installing ceph-common or a package that would have it as a dependency on the admin server. If there's a gap in the documentation, I'd like to help address it. If I need to install ceph-common on my admin node, how should I go about doing it as this is not clear from the documentation. Some possible approaches are to run one of these commands from my admin node, cephadmin1: *** sudo apt-get install ceph-common *** sudo apt-get install ceph *** ceph-deploy install --stable cuttlefish cephadmin1// I used ceph-deploy install --stable cuttlefish cephserver1 to install ceph on my server node from my admin node. Any thoughts on the most appropriate way to install ceph-common (and other required packages) on cephadmin? Thanks Gary On Sun, Jun 9, 2013 at 10:03 AM, Smart Weblications GmbH i...@smart-weblications.de wrote: Hi, Am 09.06.2013 10:42, schrieb Gary Bruce: Hi, I'm trying to run this from my admin node, have I missed a step? alphaceph@cephadmin1:~/ceph-deploy/my-cluster$ rbd create fooimage --size 1024 --pool barpool -m cephserver1.zion.bt.co.uk http://cephserver1.zion.bt.co.uk -k /etc/ceph/ceph.client.admin.keyring Look: The program 'rbd' is currently not installed. To run 'rbd' please ask your administrator to install the package 'ceph-common' Maybe you missed installing ceph-common on your host cephadmin1 -- Mit freundlichen Grüßen, Smart Weblications GmbH Martinsberger Str. 1 D-95119 Naila fon.: +49 9282 9638 200 fax.: +49 9282 9638 205 24/7: +49 900 144 000 00 - 0,99 EUR/Min* http://www.smart-weblications.de -- Sitz der Gesellschaft: Naila Geschäftsführer: Florian Wiessner HRB-Nr.: HRB 3840 Amtsgericht Hof *aus dem dt. Festnetz, ggf. abweichende Preise aus dem Mobilfunknetz ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- John Wilkins Senior Technical Writer Intank john.wilk...@inktank.com (415) 425-9599 http://inktank.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Disaster recovery of monitor
On 2013-06-14 19:59, Joao Eduardo Luis wrote: On 06/14/2013 02:39 PM, pe...@2force.nl wrote: On 2013-06-13 20:10, pe...@2force.nl wrote: On 2013-06-13 18:57, Joao Eduardo Luis wrote: On 06/13/2013 05:25 PM, pe...@2force.nl wrote: On 2013-06-13 18:06, Gregory Farnum wrote: On Thursday, June 13, 2013, wrote: Hello, We ran into a problem with our test cluster after adding monitors. It now seems that our main monitor doesn't want to start anymore. The logs are flooded with: 2013-06-13 11:41:05.316982 7f7689ca4780 7 mon.a@0(leader).osd e2809 update_from_paxos applying incremental 2810 2013-06-13 11:41:05.317043 7f7689ca4780 1 mon.a@0(leader).osd e2809 e2809: 9 osds: 9 up, 9 in 2013-06-13 11:41:05.317064 7f7689ca4780 7 mon.a@0(leader).osd e2809 update_from_paxos applying incremental 2810 Is this accurate? It's applying the *same* incremental I've and over again? Yes, this is the current state: Peter, Can you point me to the full log of the monitor caught in this apparent loop? -Joao Hi Joao, Here it is: http://www.2force.nl/ceph/ceph-mon.a.log.gz Thanks, Peter Hi Joao, Did you happen to figure out what is going on? If you need more log files let me know. Peter, You can find all the updates on #5343 [1]. It is my understanding that you are running a test cluster; is this correct? If so, our suggestion is to start your monitor fresh. We've been able to figure out all the caused for this issue (thanks for your help!): - Injecting a monmap with a wrong fsid was the main culprit. Given you are on a version suffering from a bug that won't kill the monitor if some sanity checks fail when the monitor is started, the monitor was started even though said fsid mismatch was present. A fix for that will be hitting 0.61.4 soon, and has already hit master a few days back. - There was a bug in OSDMonitor::update_from_paxos() that would ignore the return from OSDMap::apply_incremental(), thus leading to the infinite loop in case the incremental failed to be applied. That should go into master soon. However, with regard to getting the monitor running back again, there's little we can do at the moment. We don't believe the fix to correct the incremental's fsid is necessary, as it should never happen again once the patches are in and shouldn't even have happened in the first place were the fsid in the injected monmap to be correct. So, if this is indeed a test cluster, it would be better to just start off fresh; otherwise, let me know and we may be able to put a quick and dirty fix to get your cluster back again. Thanks! -Joao [1] - http://tracker.ceph.com/issues/5343 Hi Joao, You're welcome! Happy that we could help. I was at first hesitant to post to the mailinglist because I thought it was just user error. In this case it seems that due to our user error we uncovered a bug or at least something that should have never happened :) So if anyone out there is having the same feeling, just post. You never know what comes out. Are there any other tips you might have for us and other users? Is it possible to have a backup of your monitor directory? Or is ensuring you have enough monitors enough? Is it possible for errors like this to be propagated to other monitors? It would be really nice of there will be tools that can help with disaster recovery and some more documentation on this. I'm sure nobody would play around like we did with their live cluster but strange things do tend to happen (and bugs) and it is always nice to know if there is a way out. You don't want to end up with those petabytes sitting there :) Thanks! Peter ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Live Migrations with cephFS
Thank you, Sebastien Han. I am sure many are thankful you've published your thoughts and experiences with Ceph and even OpemStack.Thanks Bo! :)If I may, I would like to reword my question/statement with greater clarity: To force all instances to always boot from RBD volumes, would a person would have to make changes to something more than Horizon (demonstration GUI)? If the changes need only be inHorizon, the provider would then likely need to restrict or deny their customers access to their unmodified APIs. If they do not, then the unchanged APIs would allow for behavior the provider does not want.Thoughts? Corrections? Feel free to teach.This is correct. Forcing the boot from volume requires a modified version of the API which kinda tricky and GUI modifications. There are 2 cases:1. you're an ISP (public provider), you should forget about the idea unless you want to provide a _really_ close service.2.you're the only one managing your platform (private cloud) this might be doable but even so you'll encounter a lot of problems while upgrading. At the end it's up to you, if you're 100% sure that you have the complete control of your infra and that you know when, who and how new instances are booted (and occasionally don't care about update and compatibility).You can always hack the dashboard but it's more than that you have to automate the action that each time someone is booting a VM you have to create a volume from an image for this. This will prolong the process. At this point, I'll recommend you to push this blueprint, it'll run all the VM through ceph even the one not using the boot-from-volume option.https://blueprints.launchpad.net/nova/+spec/bring-rbd-support-libvirt-images-typeAn article is coming next week and will cover the entire subject.Cheers!Sébastien HanCloud Engineer"Always give 100%. Unless you're giving blood."Phone :+33 (0)1 49 70 99 72–Mobile :+33 (0)6 52 84 44 70Email :sebastien@enovance.com–Skype :han.sbastienAddress :10, rue de la Victoire – 75009 ParisWeb :www.enovance.com–Twitter :@enovance On Jun 17, 2013, at 8:00 AM, Wolfgang Hennerbichler wolfgang.hennerbich...@risc-software.at wrote:On 06/14/2013 08:00 PM, Ilja Maslov wrote:Hi,Is live migration supported with RBD and KVM/OpenStack?Always wanted to know but was afraid to ask :)totally works in my productive setup. but we don't use openstack in thisinstallation, just KVM/RBD.Pardon brevity and formatting, replying from the phone.Cheers,IljaRobert Sander r.san...@heinlein-support.de wrote:On 14.06.2013 12:55, Alvaro Izquierdo Jimeno wrote:By default, openstack uses NFS but… other options are available….can weuse cephFS instead of NFS?Wouldn't you use qemu-rbd for your virtual guests in OpenStack?AFAIK CephFS is not needed for KVM/qemu virtual machines.Regards--Robert SanderHeinlein Support GmbHSchwedter Str. 8/9b, 10119 Berlinhttp://www.heinlein-support.deTel: 030 / 405051-43Fax: 030 / 405051-19Zwangsangaben lt. §35a GmbHG:HRB 93818 B / Amtsgericht Berlin-Charlottenburg,Geschäftsführer: Peer Heinlein -- Sitz: Berlin___ceph-users mailing listceph-users@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.comThis email and any files transmitted with it are confidential and intended solely for the use of the individual or entity to whom they are addressed. If you are not the intended recipient, please note that any review, dissemination, disclosure, alteration, printing,circulation, retention or transmission of this e-mail and/or any file or attachment transmitted with it, is prohibited and may be unlawful. If you have received this e-mail or any file or attachment transmitted with it in error please notify postmas...@openet.com.Although Openet has taken reasonable precautions to ensure no viruses are present in this email, we cannot accept responsibility for any loss or damage arising from the use of this email or attachments.___ceph-users mailing listceph-users@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com--DI (FH) Wolfgang HennerbichlerSoftware DevelopmentUnit Advanced Computing TechnologiesRISC Software GmbHA company of the Johannes Kepler University LinzIT-CenterSoftwarepark 354232 HagenbergAustriaPhone: +43 7236 3343 245Fax: +43 7236 3343 250wolfgang.hennerbich...@risc-software.athttp://www.risc-software.at___ceph-users mailing listceph-users@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Ceph and open source cloud software: Path of least resistance
Hi Jens, with regard to OpenNebula I would like to point out a couple of things. OpenNebula has official support not just for CentOS but for three other distros, among which there's Ubuntu, which as far as I know has Ceph and rbd supported libvirt and qemu-kvm versions. Also, as far as I know there's a current effort at CentOS to provide an rpm repo with a full Ceph stack with updated libvirt and qemu-kvm versions which should be ready by the end of the month. So setting up a working cluster with Ceph (that works as of now) and deploying newer libvirt and qemu-kvm version compatible with rbd will be very easy in just a matter of weeks. However, from the point-of-view of OpenNebula all of this is pretty much transparent, the Ceph support is working and as long as you have the requirements stated above you should be good to go. regards, Jaime On Sun, Jun 16, 2013 at 7:48 PM, Jens Kristian Søgaard j...@mermaidconsulting.dk wrote: Hi guys, I'm looking to setup an open source cloud IaaS system that will work well together with Ceph. I'm looking for a system that will handle running KVM virtual servers with persistent storage on a number of physical servers with a multi-tenant dashboard. I have now tried a number of systems, but having difficulties finding something that will work with Ceph in an optimal way. Or at least, having difficulties finding hints on how to achieve that. By optimal I mean: a) To have Ceph as the only storage, so that I don't have a NFS SPoF nor have to wait for images to be copied from server to server. b) To run KVM with the async flush feature in 1.4.2 (or backported) and with the librbd cache. Any of you guys are doing this? - have hints to offer? I have tried CloudStack, but found that it was not possible to rely fully on Ceph storage. I learnt that it would potentially be possible with the upcoming 4.2 release, so I tried installed CloudStack from the development source code tree. I wasn't able to get this working because of various bugs (to be expected when running a development version ofcourse). I also tried OpenNebula, but found that it was very hard to get working on the recommended CentOS 6.4 distribution. By upgrading all sorts of systems and manually patching parts of the system I was able to get it almost working. However in the end, I ended up in a dilemma where OpenNebula needed a newer qemu version to support RBDs and that newer qemu didn't work well with the older libvirt. On the other hand if I upgraded libvirt, I couldn't get it to work with the older qemu versions with backported RBD support, as the newer libvirt where setting an auth_supported=none option that stopped it from working. It didn't seem possible to convince OpenNebula to store a secret for Ceph with libvirt. I have been looking at OpenStack, but by reading the documentation and googling it seems that it is not possible to configure OpenStack to use the librbd cache with Ceph. Could this be right? Or is it merely the case, that you cannot configure it on a per-VM basis, so that you have to rely on the default settings in ceph.conf? (which wouldn't be a problem for me) Any advice you could give would be greatly appreciated! Thanks, -- Jens Kristian Søgaard, Mermaid Consulting ApS, j...@mermaidconsulting.dk, http://www.mermaidconsulting.**com/ http://www.mermaidconsulting.com/ __**_ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/**listinfo.cgi/ceph-users-ceph.**comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Join us at OpenNebulaConf2013 http://opennebulaconf.com/ in Berlin, 24-26 September, 2013 -- Jaime Melis Project Engineer OpenNebula - The Open Source Toolkit for Cloud Computing www.OpenNebula.org | jme...@opennebula.org ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Ceph and open source cloud software: Path of least resistance
Hi Jaime, We spoke on IRC when I was trying to setup OpenNebula. Thanks for all the help and hints there! It is right that I found that my primary problem was that I choose CentOS 6.4 from the list of supported distributions, as that is the one I'm most comfortable with. If I had chosen Ubuntu from the get go, I would have have run into far fewer problems. However, I don't think OpenNebula currently fulfills the goals I set up. If it indeed does, it would be really nice - and I would start over setting up OpenNebula on Ubuntu instead. My problems with OpenNebula as far as goals go are: Reg. goal a) From my initial experience, it seems I cannot rely solely on Ceph storage. Images have to be copied back and forth between the servers. Reg. goal b) The qemu-kvm binary in the supported Ubuntu 12.10 distribution does not include async flush. I don't know if this is available as a backport from somewhere else, as my attempts to simply upgrade qemu didn't go well. -- Jens Kristian Søgaard, Mermaid Consulting ApS, j...@mermaidconsulting.dk, http://www.mermaidconsulting.com/ ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Ceph and open source cloud software: Path of least resistance
On 06/17/2013 12:51 PM, Jens Kristian Søgaard wrote: Reg. goal b) The qemu-kvm binary in the supported Ubuntu 12.10 distribution does not include async flush. I don't know if this is available as a backport from somewhere else, as my attempts to simply upgrade qemu didn't go well. I've packaged those for ubuntu 12.04 amd64, and you can download them here: http://www.wogri.at/Qemu-Ceph-Packages.343.0.html Wolfgang ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Ceph and open source cloud software: Path of least resistance
Hi Wolfgang, I've packaged those for ubuntu 12.04 amd64, and you can download them here: Thanks for the link! I'm not that familiar with Ubuntu, so sorry for the stupid question. Will this .dev be compatible with 12.10? OpenNebula doesn't list 12.04 as a supported distribution, so I'm more inclined to 12.10. -- Jens Kristian Søgaard, Mermaid Consulting ApS, j...@mermaidconsulting.dk, http://www.mermaidconsulting.com/ ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Ceph and open source cloud software: Path of least resistance
On 06/17/2013 01:03 PM, Jens Kristian Søgaard wrote: Hi Wolfgang, I've packaged those for ubuntu 12.04 amd64, and you can download them here: Thanks for the link! no problem. I'm not that familiar with Ubuntu, so sorry for the stupid question. Will this .dev be compatible with 12.10? hm. I don't think so. But I haven't tried. The debian packages define dependencies (just like RPM), and if those dependencies aren't met it won't install. So the worst that can happen to you is that you have to build qemu by hand, which wasn't really too hard (and I'm not a big fan of do-it-yourself-compiling or makefiles, too) OpenNebula doesn't list 12.04 as a supported distribution, so I'm more inclined to 12.10. it seems you're doomed :) -- DI (FH) Wolfgang Hennerbichler Software Development Unit Advanced Computing Technologies RISC Software GmbH A company of the Johannes Kepler University Linz IT-Center Softwarepark 35 4232 Hagenberg Austria Phone: +43 7236 3343 245 Fax: +43 7236 3343 250 wolfgang.hennerbich...@risc-software.at http://www.risc-software.at ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Ceph and open source cloud software: Path of least resistance
On 06/16/2013 08:48 PM, Jens Kristian Søgaard wrote: Hi guys, I'm looking to setup an open source cloud IaaS system that will work well together with Ceph. I'm looking for a system that will handle running KVM virtual servers with persistent storage on a number of physical servers with a multi-tenant dashboard. I have now tried a number of systems, but having difficulties finding something that will work with Ceph in an optimal way. Or at least, having difficulties finding hints on how to achieve that. By optimal I mean: a) To have Ceph as the only storage, so that I don't have a NFS SPoF nor have to wait for images to be copied from server to server. b) To run KVM with the async flush feature in 1.4.2 (or backported) and with the librbd cache. Any of you guys are doing this? - have hints to offer? I have tried CloudStack, but found that it was not possible to rely fully on Ceph storage. I learnt that it would potentially be possible with the upcoming 4.2 release, so I tried installed CloudStack from the development source code tree. I wasn't able to get this working because of various bugs (to be expected when running a development version ofcourse). I also tried OpenNebula, but found that it was very hard to get working on the recommended CentOS 6.4 distribution. By upgrading all sorts of systems and manually patching parts of the system I was able to get it almost working. However in the end, I ended up in a dilemma where OpenNebula needed a newer qemu version to support RBDs and that newer qemu didn't work well with the older libvirt. On the other hand if I upgraded libvirt, I couldn't get it to work with the older qemu versions with backported RBD support, as the newer libvirt where setting an auth_supported=none option that stopped it from working. It didn't seem possible to convince OpenNebula to store a secret for Ceph with libvirt. I have been looking at OpenStack, but by reading the documentation and googling it seems that it is not possible to configure OpenStack to use the librbd cache with Ceph. Could this be right? Or is it merely the case, that you cannot configure it on a per-VM basis, so that you have to rely on the default settings in ceph.conf? (which wouldn't be a problem for me) Any advice you could give would be greatly appreciated! Thanks, Hi, you might want to take a look at Synnefo. [1] Synnefo is a complete open source cloud IaaS platform, which uses Google Ganeti [2] for the VM cluster management at the backend and implements / exposes OpenStack APIs at the frontend. Synnefo supports Ceph / RBD on the API layer, as a 'disk template' when creating VMs, and passes that information to Ganeti, which actually does the RBD device handling. At the moment Ganeti only supports the in-kernel RBD driver, although support for the qemu-rbd driver should be implemented soon. Using the in-kernel RBD driver means that you should probably run a relatively modern kernel, but it also means that caching and flushing is handled by the kernel mechanisms (page cache, block layer etc), without the need to rely on specific qemu / libvirt versions to support them. Ganeti does *not* use libvirt in the backend and supports out-of-the-box both KVM and Xen. You can also read this blog post [3] for more information, to see how we use Synnefo + Ganeti + Ceph to power a large scale public cloud service. [1] http://www.synnefo.org [2] https://code.google.com/p/ganeti/ [3] http://synnefo-software.blogspot.gr/2013/02/we-are-happy-to-announce-that-synnefo_11.html Thanks, Stratos -- Stratos Psomadakis s.psomada...@gmail.com signature.asc Description: OpenPGP digital signature ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Ceph and open source cloud software: Path of least resistance
Hi Stratos, you might want to take a look at Synnefo. [1] I did take a look at it earlier, but decided not to test it. Mainly I was deterred because I found the documentation a bit lacking. I opened up the section on File Storage and found that there were only chapter titles, but no actual content. Perhaps I was too quick to dismiss it. A bit more practical problem for me was that my test equipment consists of a single server (besides the Ceph cluster). As far as I understood the docs, there was a bug that makes it impossible to run Synnefo on a single server (to be fixed in the next version)? Regarding my goals, I read through the installation guide and it recommends setting up an NFS server on one of the servers to serve images to the rest. This is what I wanted to avoid. Is that optional and/or could be replaced with Ceph? At the moment Ganeti only supports the in-kernel RBD driver, although support for the qemu-rbd driver should be implemented soon. Using the Hmm, I wanted to avoid using the in-kernel RBD driver, as I figured it lead to various problems. Is it not a problem in practice? I was thinking it would be wisest to stay with the distribution kernel, but I guess you swap it out for a later version? The rbds for all my existing VMs would probably have to be converted back from format 2 to format 1, right? -- Jens Kristian Søgaard, Mermaid Consulting ApS, j...@mermaidconsulting.dk, http://www.mermaidconsulting.com/ ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
[ceph-users] Upgrade from bobtail
Hi, i'm planning to Upgrade my bobtail (latest) cluster to cuttlefish. Are there any outstanding issues that I should be aware of? Anything that could brake my productive setup? Wolfgang -- Sent from my mobile device ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Need help with Ceph error
Yep, you can't connect to your monitors so nothing else is going to work either. There's a wealth of conversations about debugging monitor connection issues in the mailing list and irc archives (and I think some in the docs), but as a quick start list: 1) make sure the monitor processes are actually running in top 2) connect to them using the admin socket and see what state they think they're in 3) see if you can connect to them from their host instead of a different one (different keys might be present). -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com On Sun, Jun 16, 2013 at 11:48 PM, Sreejith Keeriyattil sreejith.keeriyat...@happiestminds.com wrote: Hi The issue it hangs when I type any ceph commands.. :( === root@xtream:~# ceph -s ^C root@xtream:~# service ceph start === mds.a === Starting Ceph mds.a on xtream...already running === osd.0 === Mounting xfs on xtream:/var/lib/ceph/osd/ceph-0 ^C root@xtream:~# == Thanks and regards Sreejith KJ -Original Message- From: Gregory Farnum [mailto:g...@inktank.com] Sent: Friday, June 14, 2013 9:07 PM To: Sreejith Keeriyattil Cc: ceph-us...@ceph.com Subject: Re: [ceph-users] Need help with Ceph error On Fri, Jun 14, 2013 at 12:20 AM, Sreejith Keeriyattil sreejith.keeriyat...@happiestminds.com wrote: Hi To keep it simple I disabled cephx authentication but after that am getting the below error. == == root@xtream:/etc/ceph# service ceph -v start /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n mon.a user /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n mds.a user === mds.a === /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n mds.a pid file --- xtream# mkdir -p /var/run/ceph /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n mds.a log dir /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n mds.a auto start --- xtream# [ -e /var/run/ceph/mds.a.pid ] || exit 1 # no pid, presumably not pid=`cat /var/run/ceph/mds.a.pid` [ -e /proc/$pid ] grep -q ceph-mds /proc/$pid/cmdline grep -qwe -i exit 1 # pid is something else Starting Ceph mds.a on xtream...already running /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 user === osd.0 === /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 pid file --- xtream# mkdir -p /var/run/ceph /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 log dir /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 auto start --- xtream# [ -e /var/run/ceph/osd.0.pid ] || exit 1 # no pid, presumably not pid=`cat /var/run/ceph/osd.0.pid` [ -e /proc/$pid ] grep -q ceph-osd /proc/$pid/cmdline grep -qwe -i exit 1 # pid is something else /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 copy executable to /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 osd data /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 fs path /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 devs /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 lock file /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 admin socket /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 max open files /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 restart on core dump /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 valgrind /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 pre mount command /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 osd mkfs type /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 osd mount options xfs --- xtream# true Mounting xfs on xtream:/var/lib/ceph/osd/ceph-0 --- xtream# modprobe xfs ; egrep -q '^[^ ]+ /var/lib/ceph/osd/ceph-0' /proc/moun /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 osd crush update on start /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 osd crush location /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 osd crush initial weight /usr/bin/ceph-conf -c /etc/ceph/ceph.conf -n osd.0 keyring It hangs after this My ceph.conf file looks like this I think you're still having issues connecting to your monitors. Can you run ceph -s and provide the output? -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com Happiest Minds Disclaimer This message is for the sole use of the intended recipient(s) and may contain confidential, proprietary or legally privileged information. Any unauthorized review, use, disclosure or distribution is prohibited. If you are not the original intended recipient of the message, please contact the sender by reply email and destroy all copies of the original message. Happiest Minds Technologies http://www.happiestminds.com ___ ceph-users mailing list ceph-users@lists.ceph.com
Re: [ceph-users] Upgrade from bobtail
I'm actually planning this same upgrade on Saturday. Is the memory leak from Bobtail during deep-scrub known to be squashed? I've been seeing that a lot lately. I know Bobtail-Cuttlefish is only one way, due to the mon re-architecting. But in general, whenever we do upgrades we usually have a fall-back/reversion plan in case things go wrong. Is that ever going to be possible with Ceph? - Travis On Mon, Jun 17, 2013 at 12:27 PM, Sage Weil s...@inktank.com wrote: On Mon, 17 Jun 2013, Wolfgang Hennerbichler wrote: Hi, i'm planning to Upgrade my bobtail (latest) cluster to cuttlefish. Are there any outstanding issues that I should be aware of? Anything that could brake my productive setup? There will be another point release out in the next day or two that resolves a rare sequence of errors during the upgrade that can be problematic (see the 0.61.3 release notes). There are also several fixes for udev/ceph-disk/ceph-deploy on rpm-based distros that will be included. If you can wait a couple days I would suggest that. sage ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Influencing reads/writes
On Sun, Jun 16, 2013 at 11:10 PM, Wolfgang Hennerbichler wolfgang.hennerbich...@risc-software.at wrote: On 06/16/2013 01:27 AM, Matthew Walster wrote: In the same way that we have CRUSH maps for determining placement groups, I was wondering if anyone had stumbled across a way to influence a *client* (be it CephFS or RBD) as to where they should read/write data from/to. I think the concept of CRUSH doesn't really involve your wish, to write to specific locations (i was wishing for the same some time in the past, and then I RTFM'ed more, and in the end found out that this wish is not very trivial to implement). Although reading locally is possible, as michael lowe stated in his other e-mail. This is correct. What you can do is set up pools with different rules, such that you have a West and East pool, you split up your OSDs into west and east groups, and then the pool selects a primary from the matching set of OSDs and a secondary/tertiary from the other. Then put the RBD images in the appropriate pool. On Sat, Jun 15, 2013 at 6:25 PM, Michael Lowe j.michael.l...@gmail.com wrote: My read of http://ceph.com/releases/v0-63-released/ has this for rbd reads in the dev branch. FYI, right now this really is just *local*, ie you're on the same host. We've had low-intensity discussions around enhancing this core functionality to be a read-from-closest model for a while now, but have yet to implement anything because it's much more complex than the simple hack we're currently using and requires stuff like interpreting a closeness model (probably based on CRUSH, but we don't have any similar functionality to borrow from). -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
[ceph-users] subscribe
___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
[ceph-users] Ceph auth get-or-create
Hello! How do you add new pool access to existing Ceph Client? e.g. At first create a new user -- openstack-volumes: ceph auth get-or-create client.openstack-volumes mon 'allow r' osd 'allow class-read object_prefix rbd_children, allow rwx *pool=openstack-volumes*, allow rx pool=openstack-images' Add another pool for this user to access -- openstack-volumes: ceph auth get-or-create client.openstack-volumes mon 'allow r' osd 'allow class-read object_prefix rbd_children, allow rwx *pool=openstack-volumes*, allow rwx *pool=openstack-volumes-2*, allow rx pool=openstack-images' Thanks! Amit Vijairania | 978.319.3684 --*-- ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] How to remove /var/lib/ceph/osd/ceph-2?
If you followed the standard setup, each OSD is it's own disk + filesystem. /var/lib/ceph/osd/ceph-2 is in use, as the mount point for the OSD.2 filesystem. Double check by examining the output of the `mount` command. I get the same error when I try to rename a directory that's used as a mount point. Try `umount /var/lib/ceph/osd/ceph-2` instead of the mv and rm. The fuser command is telling you that the kernel has a filesystem mounted in that directory. Nothing else appears to be using it, so the umount should complete successfully. Also, you should fix that time skew on mon.ceph-node5. The mailing list archives have several posts with good answers. On 6/15/2013 2:14 AM, Da Chun wrote: Hi all, On Ubuntu 13.04 with ceph 0.61.3. I want to remove osd.2 from my cluster. The following steps were performed: root@ceph-node6:~# ceph osd out osd.2 marked out osd.2. root@ceph-node6:~# ceph -w health HEALTH_WARN clock skew detected on mon.ceph-node5 monmap e1: 3 mons at {ceph-node4=172.18.46.34:6789/0,ceph-node5=172.18.46.35:6789/0,ceph-node6=172.18.46.36:6789/0}, election epoch 124, quorum 0,1,2 ceph-node4,ceph-node5,ceph-node6 osdmap e414: 6 osds: 5 up, 5 in pgmap v10540: 456 pgs: 456 active+clean; 12171 MB data, 24325 MB used, 50360 MB / 74685 MB avail mdsmap e102: 1/1/1 up {0=ceph-node4=up:active} 2013-06-15 16:55:22.096059 mon.0 [INF] pgmap v10540: 456 pgs: 456 active+clean; 12171 MB data, 24325 MB used, 50360 MB / 74685 MB avail ^C root@ceph-node6:~# stop ceph-osd id=2 ceph-osd stop/waiting root@ceph-node6:~# ceph osd crush remove osd.2 removed item id 2 name 'osd.2' from crush map root@ceph-node6:~# ceph auth del osd.2 updated root@ceph-node6:~# ceph osd rm 2 removed osd.2 root@ceph-node6:~# mv /var/lib/ceph/osd/ceph-2 /var/lib/ceph/osd/ceph-2.bak mv: cannot move '/var/lib/ceph/osd/ceph-2' to '/var/lib/ceph/osd/ceph-2.bak': Device or resource busy Everything was working OK until the last step to remove the osd.2 directory /var/lib/ceph/osd/ceph-2. root@ceph-node6:~# fuser -v /var/lib/ceph/osd/ceph-2 USERPID ACCESS COMMAND /var/lib/ceph/osd/ceph-2: root kernel mount /var/lib/ceph/osd/ceph-2 // What does this mean? root@ceph-node6:~# lsof +D /var/lib/ceph/osd/ceph-2 root@ceph-node6:~# I restarted the system, and found that the osd.2 daemon was still running: root@ceph-node6:~# ps aux | grep osd root 1264 1.4 12.3 550940 125732 ? Ssl 16:41 0:20 /usr/bin/ceph-osd --cluster=ceph -i 2 -f root 2876 0.0 0.0 4440 628 ?Ss 16:44 0:00 /bin/sh -e -c /usr/bin/ceph-osd --cluster=${cluster:-ceph} -i $id -f /bin/sh root 2877 4.9 18.2 613780 185676 ? Sl 16:44 1:04 /usr/bin/ceph-osd --cluster=ceph -i 5 -f I have to take this workaround: root@ceph-node6:~# rm -rf /var/lib/ceph/osd/ceph-2 rm: cannot remove '/var/lib/ceph/osd/ceph-2': Device or resource busy root@ceph-node6:~# ls /var/lib/ceph/osd/ceph-2 root@ceph-node6:~# shutdown -r now root@ceph-node6:~# ps aux | grep osd root 1416 0.0 0.0 4440 628 ?Ss 17:10 0:00 /bin/sh -e -c /usr/bin/ceph-osd --cluster=${cluster:-ceph} -i $id -f /bin/sh root 1417 8.9 5.8 468052 59868 ?Sl 17:10 0:02 /usr/bin/ceph-osd --cluster=ceph -i 5 -f root@ceph-node6:~# rm -r /var/lib/ceph/osd/ceph-2 root@ceph-node6:~# Any idea? HELP! ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Re: [ceph-users] Simulating Disk Failure
Thanks. I'll have to get more creative. :-) On 6/14/13 18:19 , Gregory Farnum wrote: Yeah. You've picked up on some warty bits of Ceph's error handling here for sure, but it's exacerbated by the fact that you're not simulating what you think. In a real disk error situation the filesystem would be returning EIO or something, but here it's returning ENOENT. Since the OSD is authoritative for that key space and the filesystem says there is no such object, presto! It doesn't exist. If you restart the OSD it does a scan of the PGs on-disk as well as what it should have, and can pick up on the data not being there and recover. But correctly handling data that has been (from the local FS' perspective) properly deleted under a running process would require huge and expensive contortions on the part of the daemon (in any distributed system that I can think of). -Greg On Friday, June 14, 2013, Craig Lewis wrote: So I'm trying to break my test cluster, and figure out how to put it back together again. I'm able to fix this, but the behavior seems strange to me, so I wanted to run it past more experienced people. I'm doing these tests using RadosGW. I currently have 2 nodes, with replication=2. (I haven't gotten to the cluster expansion testing yet). I'm going to upload a file, then simulate a disk failure by deleting some PGs on one of the OSDs. I have seen this mentioned as the way to fix OSDs that filled up during recovery/backfill. I expected the cluster to detect the error, change the cluster health to warn, then return the data from another copy. Instead, I got a 404 error. me@client ~ $ s3cmd ls 2013-06-12 00:02 s3://bucket1 me@client ~ $ s3cmd ls s3://bucket1 2013-06-12 00:0213 8ddd8be4b179a529afa5f2ffae4b9858 s3://bucket1/hello.txt me@client ~ $ s3cmd put Object1 s3://bucket1 Object1 - s3://bucket1/Object1 [1 of 1] 4 of 4 100% in 62s 6.13 MB/s done me@client ~ $ s3cmd ls s3://bucket1 2013-06-13 01:10 381M 15bdad3e014ca5f5c9e5c706e17d65f3 s3://bucket1/Object1 2013-06-12 00:0213 8ddd8be4b179a529afa5f2ffae4b9858 s3://bucket1/hello.txt So at this point, the cluster is healthy, and we can download objects from RGW. me@dev-ceph0:/var/lib/ceph/osd/ceph-0/current$ ceph status health HEALTH_OK monmap e2: 2 mons at {dev-ceph0=192.168.18.24:6789/0,dev-ceph1=192.168.18.25:6789/0 http://192.168.18.24:6789/0,dev-ceph1=192.168.18.25:6789/0}, election epoch 12, quorum 0,1 dev-ceph0,dev-ceph1 osdmap e44: 2 osds: 2 up, 2 in pgmap v4055: 248 pgs: 248 active+clean; 2852 MB data, 7941 MB used, 94406 MB / 102347 MB avail; 17B/s rd, 0op/s mdsmap e1: 0/0/1 up me@client ~ $ s3cmd get s3://bucket1/Object1 ./Object.Download1 s3://bucket1/Object1 - ./Object.Download1 [1 of 1] 4 of 4 100% in 13s27.63 MB/s done Time to simulate a failure. Let's delete all the PGs used by .rgw.buckets on OSD.0. me@dev-ceph0:~$ ceph osd tree # idweighttype nameup/downreweight -10.09998root default -20.04999host dev-ceph0 00.04999osd.0up1 -30.04999host dev-ceph1 10.04999osd.1up1 me@dev-ceph0:~$ ceph osd dump | grep .rgw.buckets pool 9 '.rgw.buckets' rep size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 21 owner 18446744073709551615 me@dev-ceph0:~$ cd /var/lib/ceph/osd/ceph-0/current me@dev-ceph0:/var/lib/ceph/osd/ceph-0/current$ du -sh 9.* 321M9.0_head 289M9.1_head 425M9.2_head 357M9.3_head 358M9.4_head 309M9.5_head 401M9.6_head 397M9.7_head me@dev-ceph0:/var/lib/ceph/osd/ceph-0/current$ sudo rm -rf 9.* The cluster is still healthy me@dev-ceph0:/var/lib/ceph/osd/ceph-0/current$ ceph status health HEALTH_OK monmap e2: 2 mons at {dev-ceph0=192.168.18.24:6789/0,dev-ceph1=192.168.18.25:6789/0 http://192.168.18.24:6789/0,dev-ceph1=192.168.18.25:6789/0}, election epoch 12, quorum 0,1 dev-ceph0,dev-ceph1 osdmap e44: 2 osds: 2 up, 2 in pgmap v4059: 248 pgs: 248 active+clean; 2852 MB data, 7941 MB used, 94406 MB / 102347 MB avail; 16071KB/s rd, 3op/s mdsmap e1: 0/0/1 up It probably hasn't noticed the damage yet, there's no I/O on this test cluster unless I generate it. Lets retrieve some data, that'll make the cluster notice. me@client ~ $ s3cmd get s3://bucket1/Object1 ./Object.Download2 s3://bucket1/Object1 - ./Object.Download2 [1 of 1] ERROR: S3 error: 404 (Not Found): me@client ~ $ s3cmd ls s3://bucket1 ERROR: S3 error: 404 (NoSuchKey): I wasn't expecting that. I expected my object to
Re: [ceph-users] Upgrade from bobtail
On Mon, Jun 17, 2013 at 02:10:27PM -0400, Travis Rhoden wrote: I'm actually planning this same upgrade on Saturday. Is the memory leak from Bobtail during deep-scrub known to be squashed? I've been seeing that a lot lately. this is actually the reason why we're planning to upgrade, too. one of the OSD's went nuts yesterday, and ate up all the memory. Ceph exploded, but - and this is the good news - it recovered smoothly. I know Bobtail-Cuttlefish is only one way, due to the mon re-architecting. But in general, whenever we do upgrades we usually have a fall-back/reversion plan in case things go wrong. Is that ever going to be possible with Ceph? just from my guts i guess this will stabilize when the mon architecture stabilizes. but ceph is young, and young means going forward only. - Travis On Mon, Jun 17, 2013 at 12:27 PM, Sage Weil s...@inktank.com wrote: On Mon, 17 Jun 2013, Wolfgang Hennerbichler wrote: Hi, i'm planning to Upgrade my bobtail (latest) cluster to cuttlefish. Are there any outstanding issues that I should be aware of? Anything that could brake my productive setup? There will be another point release out in the next day or two that resolves a rare sequence of errors during the upgrade that can be problematic (see the 0.61.3 release notes). There are also several fixes for udev/ceph-disk/ceph-deploy on rpm-based distros that will be included. If you can wait a couple days I would suggest that. sage ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- http://www.wogri.com ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
[ceph-users] ceph iscsi questions
Hi List, I want to deploy a ceph cluster with latest cuttlefish, and export it with iscsi interface to my applications. Some questions here: 1. Which Linux distro and release would you recommend? I used Ubuntu 13.04 for testing purpose before. 2. Which iscsi target is better? LIO, SCST, or others? 3. The system for the iscsi target will be a single point of failure. How to eliminate it and make good use of ceph's nature of distribution? Thanks!___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com