Re: [ceph-users] Why RBD is not enough [was: Inconsistent view on mounted CephFS]

2013-09-15 Thread Simon Leinen
Maciej Gałkiewicz writes:
 On 13 September 2013 17:12, Simon Leinen simon.lei...@switch.ch wrote:
 
 [We're not using is *instead* of rbd, we're using it *in addition to*
 rbd.  For example, our OpenStack (users') cinder volumes are stored in
 rbd.]

 So you probably have cinder volumes in rbd but you boot instances from
 images. This is why you need cephfs for /var/lib/nova/instances. I
 suggest creating volumes from images and booting instances from them.
 Cephfs is not required then

Thanks, I know that we could boot from volume.  Two problems:

1.) Our OpenStack installation is not a private cloud; we allow
external users to set up VMs.  These users need to be able to use
the standard workflow (Horizon) to start VMs from an image.

2.) We didn't manage to make boot from volume work with RBD in Folsom.
Yes, presumably it works fine in Grizzly and above, so we should
just upgrade.

 What we want to achieve is to have a shared instance store
 (i.e. /var/lib/nova/instances) across all our nova-compute nodes, so
 that we can e.g. live-migrate instances between different hosts.  And we
 want to use Ceph for that.
 
 In Folsom (but also in Grizzly, I think), this isn't straightforward to
 do with RBD.  A feature[1] to make it more straightforward was merged in
 Havana(-3) just two weeks ago.

 I dont get it. I am successfully using live-migration (in Grizzly,
 havent try Folsom) of instances booted from cinder volumes stored as
 rbd volumes. What is not straightforward to do? Are you using KVM?

As I said, boot from volume is not really an option for us.

 Yes, people want shared storage that they can access in a POSIXly way
 from multiple VMs.  CephFS is a relatively easy way to give them that,
 though I don't consider it production-ready - mostly because secure
 isolation between different tenants is hard to achieve.

 For now GlusterFS may fits better here.

Possibly, but it's another system we'd have to learn, configure and
support.  And CephFS is already in standard kernels (though obviously
it's not reliable, and there may be horrible regressions such as this
bug in 3.10).
-- 
Simon.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Inconsistent view on mounted CephFS

2013-09-15 Thread Simon Leinen
Yan, Zheng writes:
 On Fri, Sep 13, 2013 at 3:09 PM, Jens-Christian Fischer
 The problem we see, is that the different hosts have different
 views on the filesystem (i.e. they see different amount of files).
[...]

 The bug was introduced in 3.10 kernel, will be fixed in 3.12 kernel by
 commit 590fb51f1c (vfs: call d_op-d_prune() before unhashing dentry).
 Sage may backport the fix to 3.11 and 3.10 kernel soon.

This would be great! 3.12 won't be out until November.

(Downgrading to e.g.3.9.11 should also fix the issue, right?)

 please use ceph-fuse at present.

That's what we're doing now, but it seems slower.

Best regards,
-- 
Simon.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] mds stuck in rejoin

2013-09-15 Thread Serge Slipchenko
Hi,

I'm testing ceph 0.67.3 (408cd61584c72c0d97b774b3d8f95c6b1b06341a) under
load.
My configuration has 2 mds, 3 mon and 16 osd - mon and mds are on separate
servers, osd distributed on 8 servers

3 servers with several processes read and write via libcephfs.

Restart of active mds leads to infinite rejoin and complete inaccessibility
of the cephfs.

It seems related to the bug http://tracker.ceph.com/issues/4637

-- 
Kind regards, Serge Slipchenko
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] problem with ceph-deploy hanging

2013-09-15 Thread Gruher, Joseph R

From: Gruher, Joseph R
From: Alfredo Deza [mailto:alfredo.d...@inktank.com]
On Fri, Sep 13, 2013 at 5:06 PM, Gruher, Joseph R
joseph.r.gru...@intel.com wrote:

 root@cephtest01:~# ssh cephtest02 wget -q -O-
 'https://ceph.com/git/?p=ceph.git;a=blob_plain;f=keys/release.asc' |
 apt-key add -

 gpg: no valid OpenPGP data found.


This is clearly part of the problem. Can you try getting to this with
something other than wget (e.g. curl) ?

OK, I am seeing the problem here after turning off quiet mode on wget.  You
can see in the wget output that part of the URL is lost when executing the
command over SSH.  However, I'm still unsure how to fix this, I've tried a
number of ways of enclosing the command and this keeps happening.

SSH command leads to incomplete URL and returns web page (note URL
truncated at ceph.git):

root@cephtest01:~# ssh cephtest02 sudo wget -O-
'https://ceph.com/git/?p=ceph.git;a=blob_plain;f=keys/release.asc'
--2013-09-13 16:37:06--  https://ceph.com/git/?p=ceph.git

When run locally complete URL returns PGP key:

root@cephtest02:/# wget -O-
'https://ceph.com/git/?p=ceph.git;a=blob_plain;f=keys/release.asc'
--2013-09-13 16:37:30--
https://ceph.com/git/?p=ceph.git;a=blob_plain;f=keys/release.asc

I was able to show the wget command does succeed if properly formatted (have to 
double-enclose in quotes as SSH strips the outer set) as does the apt-key add 
if prefaced with a sudo.

So, I'm still stuck on the problem of ceph deploy hanging at the point shown 
below.  Any tips on how to debug further?  Has anyone else experienced a 
similar problem?  Is it possible to enable any additional output from 
ceph-deploy?  Is there any documentation on how to deploy without using 
ceph-deploy install?  Thanks!

Here's where it hangs:

root@cephtest01:~# ceph-deploy install cephtest02 cephtest03 cephtest04 
[ceph_deploy.install][DEBUG ] Installing stable version dumpling on cluster 
ceph hosts cephtest02 cephtest03 cephtest04
[ceph_deploy.install][DEBUG ] Detecting platform for host cephtest02 ...
[ceph_deploy.install][INFO  ] Distro info: Ubuntu 12.04 precise
[cephtest02][INFO  ] installing ceph on cephtest02
[cephtest02][INFO  ] Running command: env DEBIAN_FRONTEND=noninteractive 
apt-get -q install --assume-yes ca-certificates
[cephtest02][INFO  ] Reading package lists...
[cephtest02][INFO  ] Building dependency tree...
[cephtest02][INFO  ] Reading state information...
[cephtest02][INFO  ] ca-certificates is already the newest version.
[cephtest02][INFO  ] 0 upgraded, 0 newly installed, 0 to remove and 4 not 
upgraded.
[cephtest02][INFO  ] Running command: wget -q -O- 
'https://ceph.com/git/?p=ceph.git;a=blob_plain;f=keys/release.asc' | apt-key 
add -

Here's the command it seems to be hanging on succeeding when manually run on 
the command line:

root@cephtest01:~# ssh cephtest02 wget -q -O- 
'https://ceph.com/git/?p=ceph.git;a=blob_plain;f=keys/release.asc' | sudo 
apt-key add -
OK
root@cephtest01:~#

Thanks,
Joe
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Module rbd not found on Ubuntu 13.04

2013-09-15 Thread Alek Paunov

On 11.09.2013 20:05, Prasanna Gholap wrote:


By the link about aws, rbd.ko isn't included yet in linux aws . I'll try to
build the kernel manually and proceed for rbd.
Thanks for your help.


If your requirement is modern Linux (not Ubuntu exclusive) you can use 
Fedora (AMIs are built with unmodified Fedora kernel which of course 
includes recent rbd)


http://fedoraproject.org/en/get-fedora-options#clouds

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] mds stuck in rejoin

2013-09-15 Thread Gregory Farnum
What's the output of ceph -s, and have you tried running the MDS
with any logging enabled that we can check out?
-Greg
Software Engineer #42 @ http://inktank.com | http://ceph.com


On Sun, Sep 15, 2013 at 8:24 AM, Serge Slipchenko
serge.slipche...@gmail.com wrote:
 Hi,

 I'm testing ceph 0.67.3 (408cd61584c72c0d97b774b3d8f95c6b1b06341a) under
 load.
 My configuration has 2 mds, 3 mon and 16 osd - mon and mds are on separate
 servers, osd distributed on 8 servers

 3 servers with several processes read and write via libcephfs.

 Restart of active mds leads to infinite rejoin and complete inaccessibility
 of the cephfs.

 It seems related to the bug http://tracker.ceph.com/issues/4637

 --
 Kind regards, Serge Slipchenko

 ___
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Rbd cp empty block

2013-09-15 Thread 王根意
Hi all:

I have a 30G rbd block device as virtual machine disk, Aleady installed
ubuntu 12.04. About 1G space used.

When I want to deploy vm, I made a rbd cp. Then problem came, it copy 30G
data instead of 1G. And this action take lots of time.

Any ideal? I just want make it faster to deploy vm.

-- 
OPS 王根意
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] ceph-deploy and ceph-disk cannot prepare disks

2013-09-15 Thread Andy Schuette
First-time list poster here, and I'm pretty stumped on this one. My
problem hasn't really been discussed on the list before, so I'm hoping
that I can get this figured out since it's stopping me from learning
more about ceph. I've tried this with the journal on the same disk and
on a separate SSD, both with the same error stopping me.

I'm using ceph-deploy 1.2.3, and ceph is version 0.67.2 on the osd
node. OS is Ubuntu 13.04, kernel is 3.8.0-29, architecture is x86_64.

Here is my log from ceph-disk prepare:

ceph-disk prepare /dev/sdd
INFO:ceph-disk:Will colocate journal with data on /dev/sdd
Information: Moved requested sector from 34 to 2048 in
order to align on 2048-sector boundaries.
The operation has completed successfully.
Information: Moved requested sector from 2097153 to 2099200 in
order to align on 2048-sector boundaries.
The operation has completed successfully.
meta-data=/dev/sdd1  isize=2048   agcount=4, agsize=122029061 blks
 =   sectsz=512   attr=2, projid32bit=0
data =   bsize=4096   blocks=488116241, imaxpct=5
 =   sunit=0  swidth=0 blks
naming   =version 2  bsize=4096   ascii-ci=0
log  =internal log   bsize=4096   blocks=238338, version=2
 =   sectsz=512   sunit=0 blks, lazy-count=1
realtime =none   extsz=4096   blocks=0, rtextents=0
umount: /var/lib/ceph/tmp/mnt.X21v8V: device is busy.
(In some cases useful info about processes that use
 the device is found by lsof(8) or fuser(1))
ceph-disk: Unmounting filesystem failed: Command '['/bin/umount',
'--', '/var/lib/ceph/tmp/mnt.X21v8V']' returned non-zero exit status 1

And the log from ceph-deploy is the same (I truncated since it's the
same for all 3 in the following):

2013-09-02 11:42:47,658 [ceph_deploy.osd][DEBUG ] Preparing cluster
ceph disks ACU1:/dev/sdd:/dev/sdc1 ACU1:/dev/sde:/dev/sdc2
ACU1:/dev/sdf:/dev/sdc3
2013-09-02 11:42:49,855 [ceph_deploy.osd][DEBUG ] Deploying osd to ACU1
2013-09-02 11:42:49,966 [ceph_deploy.osd][DEBUG ] Host ACU1 is now
ready for osd use.
2013-09-02 11:42:49,967 [ceph_deploy.osd][DEBUG ] Preparing host ACU1
disk /dev/sdd journal /dev/sdc1 activate False
2013-09-02 11:43:03,489 [ceph_deploy.osd][ERROR ] ceph-disk-prepare
--cluster ceph -- /dev/sdd /dev/sdc1 returned 1
Information: Moved requested sector from 34 to 2048 in
order to align on 2048-sector boundaries.
The operation has completed successfully.
meta-data=/dev/sdd1  isize=2048   agcount=4, agsize=122094597 blks
 =   sectsz=512   attr=2, projid32bit=0
data =   bsize=4096   blocks=488378385, imaxpct=5
 =   sunit=0  swidth=0 blks
naming   =version 2  bsize=4096   ascii-ci=0
log  =internal log   bsize=4096   blocks=238466, version=2
 =   sectsz=512   sunit=0 blks, lazy-count=1
realtime =none   extsz=4096   blocks=0, rtextents=0

WARNING:ceph-disk:OSD will not be hot-swappable if journal is not the
same device as the osd data
umount: /var/lib/ceph/tmp/mnt.68dFXq: device is busy.
(In some cases useful info about processes that use
 the device is found by lsof(8) or fuser(1))
ceph-disk: Unmounting filesystem failed: Command '['/bin/umount',
'--', '/var/lib/ceph/tmp/mnt.68dFXq']' returned non-zero exit status 1

When I go to the host machine I can umount all day with no indication
of anything holding up the process, and lsof isn't yielding anything
useful for me. Any pointers to what is going wrong would be
appreciated.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com