CephFS is I believe not very production ready. Use production quality
clustered filesystems or consider using NFS or Samba shares.
The exact setup depends on what you need.

Cheers, Vasily.


On Thu, May 21, 2015 at 6:47 PM, gjprabu <gjpr...@zohocorp.com> wrote:

> Hi Angapov,
>
>   I have seen below message in ceph official sites. How it's consider to
> use in production.
>
> "ImportantCephFS currently lacks a robust ‘fsck’ check and repair
> function. Please use caution when storing important data as the disaster
> recovery tools are still under development. For more information about
> using CephFS today, see CephFS for early adopters"
>
> Regards
> Prabu
>
> Regards
> G.J
>
>
> ---- On Thu, 21 May 2015 19:57:09 +0530 * anga...@gmail.com
> <anga...@gmail.com> * wrote ----
>
> Hi, Prabu!
>
> This behavior is expected because you are using non-clustered filesystem
> (ext4 or xfs or whatever), which is not expected to be mounted to multiple
> hosts at the same time.
> What's more - you can lose data when doing like this. That's the nature of
> local filesystems.
> So if you need to access filesystem simultaneously at two or more hosts -
> consider using clustered filesystem like CephFS, OCFS etc or using network
> mounts like NFS or Samba.
>
> Regards, Vasily.
>
> On Thu, May 21, 2015 at 5:10 PM, gjprabu <gjpr...@zohocorp.com> wrote:
>
> Hi All,
>
>         We are using rbd and map the same rbd image to the rbd device on
> two different client but i can't see the data until i umount and mount -a
> partition. Kindly share the solution for this issue.
>
> *Example*
> create rbd image named foo
> map foo to /dev/rbd0 on server A,   mount /dev/rbd0 to /mnt
> map foo to /dev/rbd0 on server B,   mount /dev/rbd0 to /mnt
>
> Regards
> Prabu
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to