Re: [ceph-users] Ceph for "home lab" / hobbyist use?

2019-09-06 Thread William Ferrell
On Fri, Sep 6, 2019 at 6:37 PM Peter Woodman  wrote:
>
> 2GB ram is gonna be really tight, probably.

Bummer. So it won't be enough for any Ceph component to run reliably?

Drat. The nice thing about the HC2 is the fact that it can power the
attached SATA disk and itself through one barrel connector and
includes a stackable case to mount the disk and act as a heat sink.
Would have been so easy to stack those up and use fewer cables/power
adapters to do it. I wish I could convince the Hardkernel guys to drop
a new model with this form factor that uses their 64-bit 4GB chipset.

> However, I do something similar at home with a bunch of rock64 4gb boards, 
> and it works well.

Nice! Would you mind describing your setup in more detail? It's
encouraging to hear this works on arm (I wasn't sure if that was still
supported). I'm also curious how you manage all the power supplies,
cables and disks. Assuming you're using external 3.5" USB 3 disks that
require their own power, you're already at 8 power adapters for just 4
nodes.

> There are sometimes issues with the released ARM packages (frequently crc32 
> doesn;'t work, which isn't great), so you may have to build your own on the 
> board you're targeting or on something like scaleway, YMMV.

Thanks, that's good to know.
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Ceph for "home lab" / hobbyist use?

2019-09-06 Thread Peter Woodman
2GB ram is gonna be really tight, probably. However, I do something similar
at home with a bunch of rock64 4gb boards, and it works well. There are
sometimes issues with the released ARM packages (frequently crc32 doesn;'t
work, which isn't great), so you may have to build your own on the board
you're targeting or on something like scaleway, YMMV.

On Fri, Sep 6, 2019 at 6:16 PM Cranage, Steve 
wrote:

> I use those HC2 nodes for my home Ceph cluster, but my setup only has to
> support the librados API, my software does HSM between regular XFS file
> systems and the RADOS api so I don’t need the other MDS and the rest so I
> can’t tell you if you’ll be happy in your configuration.
>
>
>
> Steve Cranage
>
> Principal Architect, Co-Founder
>
> DeepSpace Storage
>
> 719-930-6960
>
>
> --
> *From:* ceph-users  on behalf of
> William Ferrell 
> *Sent:* Friday, September 6, 2019 3:16:30 PM
> *To:* ceph-users@lists.ceph.com 
> *Subject:* [ceph-users] Ceph for "home lab" / hobbyist use?
>
> Hello everyone!
>
> After years of running several ZFS pools on a home server and several
> disk failures along the way, I've decided that my current home storage
> setup stinks. So far there hasn't been any data loss, but
> recovering/"resilvering" a ZFS pool after a disk failure is a
> nail-biting experience. I also think the way things are set up now
> isn't making the best use of all the disks attached to the server;
> they were acquired over time instead of all at once, so I've got 4
> 4-disk raidz1 pools, each in their own enclosures. If any enclosure
> dies, all that pool's data is lost. Despite having a total of 16 disks
> in use for storage, the entire system can only "safely" lose one disk
> before there's a risk of a second failure taking a bunch of data with
> it.
>
> I'd like to ask the list's opinions on running a Ceph cluster in a
> home environment as a filer using cheap, low-power systems. I don't
> have any expectations for high performance (this will be built on a
> gigabit network, and just used for backups and streaming videos,
> music, etc. for two people); the main concern is resiliency if one or
> two disks fail, and the secondary concern is having a decent usable
> storage capacity. Being able to slowly add capacity to the cluster one
> disk at a time is a very appealing bonus.
>
> I'm interested in using these things as OSDs (and hopefully monitors
> and metadata servers):
> https://www.hardkernel.com/shop/odroid-hc2-home-cloud-two/
>
> They're about $50 each, can boot from MicroSD or eMMC flash (basically
> an SSD with a custom connector), and have one SATA port. They have
> 8-core 32-bit CPUs, 2GB of RAM and a gigabit ethernet port. Four of
> them (including disks) can run off a single 12V/8A power adapter
> (basically 100 watts per set of 4). The obvious appeal is price, plus
> they're stackable so they'd be easy to hide away in a closet.
>
> Is it feasible for these to work as OSDs at all? The Ceph hardware
> recommendations page suggests OSDs need 1GB per TB of space, so does
> this mean these wouldn't be suitable with, say, a 4TB or 8TB disk? Or
> would they work, but just more slowly?
>
> Pushing my luck further (assuming the HC2 can handle OSD duties at
> all), is that enough muscle to run the monitor and/or metadata
> servers? Should monitors and MDS's be run separately, or can/should
> they piggyback on hosts running OSDs?
>
> I'd be perfectly happy with a setup like this even if it could only
> achieve speeds in the 20-30MB/sec range.
>
> Is this a dumb idea, or could it actually work? Are there any other
> recommendations among Ceph users for low-end hardware to cobble
> together a working cluster?
>
> Any feedback is sincerely appreciated.
>
> Thanks!
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Ceph for "home lab" / hobbyist use?

2019-09-06 Thread Cranage, Steve
I use those HC2 nodes for my home Ceph cluster, but my setup only has to 
support the librados API, my software does HSM between regular XFS file systems 
and the RADOS api so I don’t need the other MDS and the rest so I can’t tell 
you if you’ll be happy in your configuration.



Steve Cranage

Principal Architect, Co-Founder

DeepSpace Storage

719-930-6960

[cid:image001.png@01D3FCBC.58FDB6F0]




From: ceph-users  on behalf of William 
Ferrell 
Sent: Friday, September 6, 2019 3:16:30 PM
To: ceph-users@lists.ceph.com 
Subject: [ceph-users] Ceph for "home lab" / hobbyist use?

Hello everyone!

After years of running several ZFS pools on a home server and several
disk failures along the way, I've decided that my current home storage
setup stinks. So far there hasn't been any data loss, but
recovering/"resilvering" a ZFS pool after a disk failure is a
nail-biting experience. I also think the way things are set up now
isn't making the best use of all the disks attached to the server;
they were acquired over time instead of all at once, so I've got 4
4-disk raidz1 pools, each in their own enclosures. If any enclosure
dies, all that pool's data is lost. Despite having a total of 16 disks
in use for storage, the entire system can only "safely" lose one disk
before there's a risk of a second failure taking a bunch of data with
it.

I'd like to ask the list's opinions on running a Ceph cluster in a
home environment as a filer using cheap, low-power systems. I don't
have any expectations for high performance (this will be built on a
gigabit network, and just used for backups and streaming videos,
music, etc. for two people); the main concern is resiliency if one or
two disks fail, and the secondary concern is having a decent usable
storage capacity. Being able to slowly add capacity to the cluster one
disk at a time is a very appealing bonus.

I'm interested in using these things as OSDs (and hopefully monitors
and metadata servers):
https://www.hardkernel.com/shop/odroid-hc2-home-cloud-two/

They're about $50 each, can boot from MicroSD or eMMC flash (basically
an SSD with a custom connector), and have one SATA port. They have
8-core 32-bit CPUs, 2GB of RAM and a gigabit ethernet port. Four of
them (including disks) can run off a single 12V/8A power adapter
(basically 100 watts per set of 4). The obvious appeal is price, plus
they're stackable so they'd be easy to hide away in a closet.

Is it feasible for these to work as OSDs at all? The Ceph hardware
recommendations page suggests OSDs need 1GB per TB of space, so does
this mean these wouldn't be suitable with, say, a 4TB or 8TB disk? Or
would they work, but just more slowly?

Pushing my luck further (assuming the HC2 can handle OSD duties at
all), is that enough muscle to run the monitor and/or metadata
servers? Should monitors and MDS's be run separately, or can/should
they piggyback on hosts running OSDs?

I'd be perfectly happy with a setup like this even if it could only
achieve speeds in the 20-30MB/sec range.

Is this a dumb idea, or could it actually work? Are there any other
recommendations among Ceph users for low-end hardware to cobble
together a working cluster?

Any feedback is sincerely appreciated.

Thanks!
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] Ceph for "home lab" / hobbyist use?

2019-09-06 Thread William Ferrell
Hello everyone!

After years of running several ZFS pools on a home server and several
disk failures along the way, I've decided that my current home storage
setup stinks. So far there hasn't been any data loss, but
recovering/"resilvering" a ZFS pool after a disk failure is a
nail-biting experience. I also think the way things are set up now
isn't making the best use of all the disks attached to the server;
they were acquired over time instead of all at once, so I've got 4
4-disk raidz1 pools, each in their own enclosures. If any enclosure
dies, all that pool's data is lost. Despite having a total of 16 disks
in use for storage, the entire system can only "safely" lose one disk
before there's a risk of a second failure taking a bunch of data with
it.

I'd like to ask the list's opinions on running a Ceph cluster in a
home environment as a filer using cheap, low-power systems. I don't
have any expectations for high performance (this will be built on a
gigabit network, and just used for backups and streaming videos,
music, etc. for two people); the main concern is resiliency if one or
two disks fail, and the secondary concern is having a decent usable
storage capacity. Being able to slowly add capacity to the cluster one
disk at a time is a very appealing bonus.

I'm interested in using these things as OSDs (and hopefully monitors
and metadata servers):
https://www.hardkernel.com/shop/odroid-hc2-home-cloud-two/

They're about $50 each, can boot from MicroSD or eMMC flash (basically
an SSD with a custom connector), and have one SATA port. They have
8-core 32-bit CPUs, 2GB of RAM and a gigabit ethernet port. Four of
them (including disks) can run off a single 12V/8A power adapter
(basically 100 watts per set of 4). The obvious appeal is price, plus
they're stackable so they'd be easy to hide away in a closet.

Is it feasible for these to work as OSDs at all? The Ceph hardware
recommendations page suggests OSDs need 1GB per TB of space, so does
this mean these wouldn't be suitable with, say, a 4TB or 8TB disk? Or
would they work, but just more slowly?

Pushing my luck further (assuming the HC2 can handle OSD duties at
all), is that enough muscle to run the monitor and/or metadata
servers? Should monitors and MDS's be run separately, or can/should
they piggyback on hosts running OSDs?

I'd be perfectly happy with a setup like this even if it could only
achieve speeds in the 20-30MB/sec range.

Is this a dumb idea, or could it actually work? Are there any other
recommendations among Ceph users for low-end hardware to cobble
together a working cluster?

Any feedback is sincerely appreciated.

Thanks!
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] regurlary 'no space left on device' when deleting on cephfs

2019-09-06 Thread Paul Emmerich
Yeah, no ENOSPC error code on deletion is a little bit unintuitive,
but what it means is: the purge queue is full.
You've already told the MDS to purge faster.

Not sure how to tell it to increase the maximum backlog for
deletes/purges, though, but you should be able to find something with
the search term "purge queue". :)


Paul

-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90

On Fri, Sep 6, 2019 at 4:00 PM Stefan Kooman  wrote:
>
> Quoting Kenneth Waegeman (kenneth.waege...@ugent.be):
> > The cluster is healthy at this moment, and we have certainly enough space
> > (see also osd df below)
>
> It's not well balanced though ... do you use ceph balancer (with
> balancer in upmap mode)?
>
> Gr. Stefan
>
>
> --
> | BIT BV  https://www.bit.nl/Kamer van Koophandel 09090351
> | GPG: 0xD14839C6   +31 318 648 688 / i...@bit.nl
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] regurlary 'no space left on device' when deleting on cephfs

2019-09-06 Thread Stefan Kooman
Quoting Kenneth Waegeman (kenneth.waege...@ugent.be):
> The cluster is healthy at this moment, and we have certainly enough space
> (see also osd df below)

It's not well balanced though ... do you use ceph balancer (with
balancer in upmap mode)?

Gr. Stefan


-- 
| BIT BV  https://www.bit.nl/Kamer van Koophandel 09090351
| GPG: 0xD14839C6   +31 318 648 688 / i...@bit.nl
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] regurlary 'no space left on device' when deleting on cephfs

2019-09-06 Thread Kenneth Waegeman

Hi all,

We are using cephfs to make a copy of another fs via rsync, and also use 
snapshots.


I'm seeing this issue now and then when I try to delete files on cephFS:

|[root@osd001 ~]# rm -f /mnt/ceph/backups/osd00*||
||rm: cannot remove 
‘/mnt/ceph/backups/osd001.gigalith.os-3eea7740.1542483’: No space left 
on device||
||rm: cannot remove 
‘/mnt/ceph/backups/osd001.gigalith.os-b2f21740.1557247’: No space left 
on device||
||rm: cannot remove 
‘/mnt/ceph/backups/osd001.gigalith.os-ca3be740.1549780’: No space left 
on device||
||rm: cannot remove 
‘/mnt/ceph/backups/osd002.gigalith.os-1b437740.1173950’: No space left 
on device||
||rm: cannot remove 
‘/mnt/ceph/backups/osd002.gigalith.os-92186740.1169503’: No space left 
on device||
||rm: cannot remove 
‘/mnt/ceph/backups/osd002.gigalith.os-e9260740.1178280’: No space left 
on device||
||rm: cannot remove 
‘/mnt/ceph/backups/osd003.gigalith.os-2dec5740.2025571’: No space left 
on device||
||rm: cannot remove 
‘/mnt/ceph/backups/osd003.gigalith.os-e5f94740.2029993’: No space left 
on device||
||rm: cannot remove 
‘/mnt/ceph/backups/osd004.gigalith.os-f4a9740.364609’: No space left on 
device||

|


The cluster is healthy at this moment, and we have certainly enough 
space (see also osd df below)



[root@osd001 ~]# ceph -s
  cluster:
    id: 92bfcf0a-1d39-43b3-b60f-44f01b630e47
    health: HEALTH_OK

  services:
    mon: 3 daemons, quorum mds01,mds02,mds03
    mgr: mds02(active), standbys: mds03, mds01
    mds: ceph_fs-2/2/2 up {0=mds01=up:active,1=mds02=up:active}, 1 
up:standby

    osd: 536 osds: 536 up, 536 in

  data:
    pools:   3 pools, 3328 pgs
    objects: 451.5 M objects, 762 TiB
    usage:   1.1 PiB used, 2.0 PiB / 3.2 PiB avail
    pgs: 1280 active+clean
 1138 active+clean+snaptrim_wait
 899  active+clean+snaptrim
 9    active+clean+scrubbing+deep
 2    active+clean+scrubbing

  io:
    client:   0 B/s wr, 0 op/s rd, 0 op/s wr


There is also nothing in the mds log.

We tuned the mds already before:

|[mds]||
||mds_cache_memory_limit=21474836480||
||mds_log_max_expiring=200||
||mds_log_max_segments=200||
||mds_max_purge_files=2560||
||mds_max_purge_ops=327600||
||mds_max_purge_ops_per_pg=20|


I tried restarting the mds, flushing the mds journals, and do a remount 
on the client, but that does not help - after some time it just works 
again..



What can I do to debug / tune  this further ?


Thanks!!

Kenneth



ceph osd df :


ID  CLASS WEIGHT  REWEIGHT SIZE    USE  AVAIL   %USE  VAR PGS
546   ssd 0.14699  1.0 151 GiB   67 GiB  84 GiB 44.30 1.23 512
547   ssd 0.14699  1.0 151 GiB   68 GiB  83 GiB 45.27 1.26 512
  0   ssd 0.14699  1.0 151 GiB   66 GiB  85 GiB 43.88 1.22 515
  1   ssd 0.14699  1.0 151 GiB   66 GiB  85 GiB 43.65 1.21 509
  2   ssd 0.14699  1.0 151 GiB   67 GiB  84 GiB 44.53 1.24 511
  3   ssd 0.14699  1.0 151 GiB   67 GiB  84 GiB 44.39 1.23 513
540  fast 0.14799  1.0 151 GiB  3.1 GiB 148 GiB  2.05 0.06   0
541  fast 0.14799  1.0 151 GiB  3.1 GiB 148 GiB  2.05 0.06   0
528   hdd 3.63899  1.0 3.6 TiB  1.4 TiB 2.2 TiB 38.50 1.07  31
529   hdd 3.63899  1.0 3.6 TiB  2.1 TiB 1.6 TiB 56.96 1.58  42
530   hdd 3.63899  1.0 3.6 TiB  1.5 TiB 2.2 TiB 39.85 1.11  31
531   hdd 3.63899  1.0 3.6 TiB  1.5 TiB 2.2 TiB 39.91 1.11  30
532   hdd 3.63899  1.0 3.6 TiB  1.6 TiB 2.1 TiB 42.75 1.19  34
533   hdd 3.63899  1.0 3.6 TiB  1.3 TiB 2.3 TiB 35.60 0.99  31
534   hdd 3.63899  1.0 3.6 TiB  2.2 TiB 1.4 TiB 61.22 1.70  46
535   hdd 3.63899  1.0 3.6 TiB  1.3 TiB 2.3 TiB 37.02 1.03  33
536   hdd 3.63899  1.0 3.6 TiB  1.5 TiB 2.2 TiB 39.89 1.11  30
537   hdd 3.63899  1.0 3.6 TiB  1.1 TiB 2.5 TiB 31.35 0.87  24
538   hdd 3.63899  1.0 3.6 TiB  2.3 TiB 1.4 TiB 62.68 1.74  53
539   hdd 3.63899  1.0 3.6 TiB  1.9 TiB 1.8 TiB 51.27 1.43  40
542   hdd 3.63899  1.0 3.6 TiB  2.2 TiB 1.5 TiB 59.81 1.66  46
543   hdd 3.63899  1.0 3.6 TiB  1.5 TiB 2.1 TiB 41.27 1.15  35
544   hdd 3.63899  1.0 3.6 TiB  1.3 TiB 2.3 TiB 35.58 0.99  28
545   hdd 3.63899  1.0 3.6 TiB  1.2 TiB 2.4 TiB 32.73 0.91  28
520  fast 0.14799  1.0 151 GiB  3.8 GiB 147 GiB  2.53 0.07   0
522  fast 0.14799  1.0 151 GiB  3.8 GiB 147 GiB  2.53 0.07   0
496   hdd 3.63899  1.0 3.6 TiB  1.8 TiB 1.9 TiB 48.41 1.35  35
498   hdd 3.63899  1.0 3.6 TiB  1.1 TiB 2.6 TiB 29.88 0.83  27
500   hdd 3.63899  1.0 3.6 TiB  2.0 TiB 1.6 TiB 55.49 1.54  43
502   hdd 3.63899  1.0 3.6 TiB  1.4 TiB 2.2 TiB 38.48 1.07  31
504   hdd 3.63899  1.0 3.6 TiB  1.1 TiB 2.6 TiB 29.90 0.83  24
510   hdd 3.63899  1.0 3.6 TiB  1.2 TiB 2.4 TiB 34.16 0.95  28
512   hdd 3.63899  1.0 3.6 TiB  955 GiB 2.7 TiB 25.64 0.71  22
514   hdd 3.63899  1.0 3.6 TiB  1.3 TiB 2.3 TiB 37.03 1.03  31
516   hdd 3.63899  1.0 3.6 TiB  1.6 TiB 2.1 TiB 42.73 1.19  37
518   hdd 3.63899  1.0 3.6 TiB  1.8 TiB 1.9 TiB 48.39 1.35  38
524   hdd 3.63899  1.0 

[ceph-users] Automatic balancing vs supervised optimization

2019-09-06 Thread Massimo Sgaravatto
Hi

I have question regarding supervised/automatic balancing using upmap.

I created a  plan in supervised mode, but its score was not expected to
improve the data distribution. But the automatic balancer triggered a
considerable rebalance.
Is this normal ? I thought that automatic balancing basically means running
from time to time the balancer in supervised mode,


More details on what I tried:

I changed the balancer mode of my luminous cluster from crush-compact to
upmap:


ceph balancer off
ceph osd set-require-min-compat-client luminous
ceph osd crush weight-set rm-compat

This triggered some rebalance. When done (i.e. when the status was
HEALTH_OK) I issued:

ceph balancer mode upmap

I then tried to run the balancer in supervised mode:

# ceph balancer optimize myplan
# ceph balancer eval myplan

But the returned value was greater than the value returned by "ceph
balancer eval"

I then tried with the automatic balancer:

ceph balancer on

which started moving objects improving A LOT the data distribution.


Thanks, Massimo
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[ceph-users] 14.2.2 -> 14.2.3 upgrade [WRN] failed to encode map e905 with expected crc

2019-09-06 Thread Stefan Kooman
Hi,

While upgrading the monitors on a Nautilus test cluster warning messages
apear:

[WRN] failed to encode map e905 with expected crc

Is this expected?

I have only seen this in the past when mixing different releases (major
versions), not when upgrading within a release.

What is the impact of this?

Gr. Stefan

-- 
| BIT BV  https://www.bit.nl/Kamer van Koophandel 09090351
| GPG: 0xD14839C6   +31 318 648 688 / i...@bit.nl
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com