[ceph-users] CfP Software Defined Storage devroom

2020-12-08 Thread Jan Fajerski
FOSDEM is a free software event that offers open source communities a place to 
meet, share ideas and collaborate.  It is well known for being highly 
developer-oriented and in the past brought together 8000+ participants from all 
over the world.  It's home is in the city of Brussels (Belgium).


FOSDEM 2021 will take place as an online event during the weekend of February 
6./7. 2021. More details about the event can be found at http://fosdem.org/


** Call For Participation

The Software Defined Storage devroom will go into it's fifth round for talks 
around Open Source Software Defined Storage projects, management tools

and real world deployments.

Presentation topics could include but are not limited too:

- Your work on a SDS project like Ceph, Gluster, OpenEBS, CORTX or Longhorn

- Your work on or with SDS related projects like OpenStack SWIFT or Container 
  Storage Interface


- Management tools for SDS deployments

- Monitoring tools for SDS clusters

** Important dates:

- Dec 27th 2020:  submission deadline for talk proposals
- Dec 31st 2020:  announcement of the final schedule
- Feb  6th 2021:  Software Defined Storage dev room

Talk proposals will be reviewed by a steering committee:
- Niels de Vos (OpenShift Container Storage Developer - Red Hat)
- Jan Fajerski (Ceph Developer - SUSE)
- TBD

Use the FOSDEM 'pentabarf' tool to submit your proposal:
https://penta.fosdem.org/submission/FOSDEM21

- If necessary, create a Pentabarf account and activate it.
Please reuse your account from previous years if you have
already created it.
https://penta.fosdem.org/user/new_account/FOSDEM21

- In the "Person" section, provide First name, Last name
(in the "General" tab), Email (in the "Contact" tab)
and Bio ("Abstract" field in the "Description" tab).

- Submit a proposal by clicking on "Create event".

- If you plan to register your proposal in several tracks to increase your chances, 
don't! Register your talk once, in the most accurate track.


- Presentations have to be pre-recorded before the event and will be streamed on 
  the event weekend.


- Important! Select the "Software Defined Storage devroom" track
(on the "General" tab).

- Provide the title of your talk ("Event title" in the "General" tab).

- Provide a description of the subject of the talk and the
intended audience (in the "Abstract" field of the "Description" tab)

- Provide a rough outline of the talk or goals of the session (a short
list of bullet points covering topics that will be discussed) in the
"Full description" field in the "Description" tab

- Provide an expected length of your talk in the "Duration" field.
  We suggest a length between 15 and 45 minutes.

** For accepted talks

Once your proposal is accepted we will assign you a volunteer deputy who will 
help you to produce the talk recording.  The volunteer will also try to ensure 
the recording is of good quality, help with uploading it to the system, 
broadcasting it during the event and moderate the Q session after the 
broadcast.  Please note that as a presenter you're expected to be available 
online during and especially after the broadcast of you talk.  The schedule will 
be available under 
https://fosdem.org/2021/schedule/track/software_defined_storage/


Hope to hear from you soon! And please forward this announcement.

If you have any further questions, please write to the mailing list at
storage-devr...@lists.fosdem.org and we will try to answer as soon as
possible.

Thanks!
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: How to use ceph-volume to create multiple OSDs per NVMe disk, and with fixed WAL/DB partition on another device?

2020-11-11 Thread Jan Fajerski

On Fri, Nov 06, 2020 at 10:15:52AM -, victorh...@yahoo.com wrote:

I'm building a new 4-node Proxmox/Ceph cluster, to hold disk images for our 
VMs. (Ceph version is 15.2.5).

Each node has 6 x NVMe SSDs (4TB), and 1 x Optane drive (960GB).

CPU is AMD Rome 7442, so there should be plenty of CPU capacity to spare.

My aim is to create 4 x OSDs per NVMe SSD (to make more effective use of the 
NVMe performance) and use the Optane drive to store the WAL/DB partition for 
each OSD. (I.e. total of 24 x 35GB WAL/DB partitions).

However, I am struggling to get the right ceph-volume command to achieve this.

Thanks to a very kind Redditor, I was able to get close:

/dev/nvme0n1 is an Optane device (900GB).

/dev/nvme2n1 is an Intel NVMe SSD (4TB).

```
# ceph-volume lvm batch --osds-per-device 4 /dev/nvme2n1 --db-devices 
/dev/nvme0n1

Total OSDs: 4

Solid State VG:
 Targets:   block.db  Total size: 893.00 GB
 Total LVs: 16Size per LV: 223.25 GB
 Devices:   /dev/nvme0n1

 TypePathLV 
Size % of device

 [data]  /dev/nvme2n1931.25 
GB   25.0%
 [block.db]  vg: vg/lv   223.25 
GB   25%

 [data]  /dev/nvme2n1931.25 
GB   25.0%
 [block.db]  vg: vg/lv   223.25 
GB   25%

 [data]  /dev/nvme2n1931.25 
GB   25.0%
 [block.db]  vg: vg/lv   223.25 
GB   25%

 [data]  /dev/nvme2n1931.25 
GB   25.0%
 [block.db]  vg: vg/lv   223.25 
GB   25%
--> The above OSDs would be created if the operation continues
--> do you want to proceed? (yes/no)
```

This does split up the NVMe disk into 4 OSDs, and creates WAL/DB partition on 
the Optane drive - however, it creates 4 x 223 GB partitions on the Optane 
(whereas I want 35GB partitions).

Is there any way to specify the WAL/DB partition size in the above?

And can it be done, such that you can run successive ceph-volume commands, to 
add the OSDs and WAL/DB partitions for each NVMe disk?
Is there is particular reason you want to run ceph-volume multiple times? The 
batch subcommand can handle that in one go without the need to explicitly 
specify any sizes as another reply proposed (though that will work nicely).


Something like this should get you there:
ceph-volume lvm batch --osds-per-device 4 /dev/nvme1n1 /dev/nvme2n1 
/dev/nvme3n1 /dev/nvme4n1 --db-devices /dev/nvme0n1

This of course makes assumption regarding device names, adjust accordingly.

Another option to size the volumes on the Optane drive would be to rely on the 
*slots arguments of the batch subcommand. See either ceph-volume lvm batch --help 
or https://docs.ceph.com/en/latest/ceph-volume/lvm/batch/#implicit-sizing




(Or if there's an easier way to achieve the above layout, please let me know).

That being said - I also just saw this ceph-users thread:

https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/thread/3Y6DEJCF7ZMXJL2NRLXUUEX76W7PPYXK/

It talks there about "osd op num shards" and "osd op num threads per shard" - 
is there some way to set those, to achieve similar performance to say, 4 x OSDs per NVMe drive, but 
with only 1 x NVMe? Has anybody done any testing/benchmarking on this they can share?
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph Octopus and Snapshot Schedules

2020-10-27 Thread Jan Fajerski

It sure is: https://docs.ceph.com/en/latest/releases/octopus/#rbd-block-storage

On Tue, Oct 27, 2020 at 10:29:18AM -0400, Adam Boyhan wrote:

  That is exactly what I am thinking.  My mistake, I should have
  specified RBD.
  Is snapshots scheduling/retention for RBD already in Octopus as well?

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph Octopus and Snapshot Schedules

2020-10-27 Thread Jan Fajerski

Maybe you are thinking of the rbd variety of this?
https://docs.ceph.com/en/latest/rbd/rbd-mirroring/

On Tue, Oct 27, 2020 at 10:19:25AM -0400, Adam Boyhan wrote:

  I thought Octopus brought the new snapshot replication feature to the
  table?  Was there issues with it?

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph Octopus and Snapshot Schedules

2020-10-27 Thread Jan Fajerski
snapshot scheduling for CephFS landed in master already, The octopus backport is 
pending still but on the agenda (https://github.com/ceph/ceph/pull/37142)
Replication of those snapshots is being worked on and is planned for pacific, 
but rsync can help you out until then surely.


On Thu, Oct 22, 2020 at 09:37:50AM -0400, Adam Boyhan wrote:

Hey all.

I was wondering if Ceph Octopus is capable of automating/managing snapshot 
creation/retention and then replication? Ive seen some notes about it, but 
can't seem to find anything solid.

Open to suggestions as well. Appreciate any input!
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: [doc] drivegroups advanced case

2020-08-24 Thread Jan Fajerski

On Fri, Aug 21, 2020 at 06:52:19PM +, Tony Liu wrote:

Hi,

Regarding to the YAML in this section,
https://ceph.readthedocs.io/en/latest/cephadm/drivegroups/#the-advanced-case

Is the "rotational" supposed to be "1" meaning spinning HDD?

Correct. This corresponds to /sys/block//queue/rotational.



Thanks!
Tony
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io



--
Jan Fajerski
Senior Software Engineer Enterprise Storage
SUSE Software Solutions Germany GmbH
Maxfeldstr. 5, 90409 Nürnberg, Germany
(HRB 36809, AG Nürnberg)
Geschäftsführer: Felix Imendörffer
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: can't get healthy cluster to trim osdmaps (13.2.8)

2020-03-23 Thread Jan Fajerski

https://tracker.ceph.com/issues/44184
Looks similar, maybe you're also seeing other symptoms listed there? In any case 
would be good to track this in one place.


On Mon, Mar 23, 2020 at 11:29:53AM +0100, Nikola Ciprich wrote:

OK, so after some debugging, I've pinned the problem down to
OSDMonitor::get_trim_to:

   std::lock_guard l(creating_pgs_lock);
   if (!creating_pgs.pgs.empty()) {
 return 0;
   }

apparently creating_pgs.pgs.empty() is not true, do I understand it
correctly that cluster thinks the list of creating pgs is not empty?

all pgs are in clean+active state, so maybe there's something malformed
in the db? How can I check?

I tried dumping list of creating_pgs according to
http://lists.ceph.com/pipermail/ceph-users-ceph.com/2018-October/030297.html
but to no avail

On Tue, Mar 17, 2020 at 12:25:29PM +0100, Nikola Ciprich wrote:

Hello dear cephers,

lately, there's been some discussion about slow requests hanging
in "wait for new map" status. At least in my case, it's being caused
by osdmaps not being properly trimmed. I tried all possible steps
to force osdmap pruning (restarting mons, restarting everyging,
poking crushmap), to no avail. Still all OSDs keep min osdmap version
1, while newest is 4734. Otherwise cluster is healthy, with no down
OSDs, network communication works flawlessly, all seems to be fine.
Just can't get old osdmaps to go away.. I's very small cluster and I've
moved all production traffic elsewhere, so I'm free to investigate
and debug, however I'm out of ideas on what to try or where to look.

Any ideas somebody please?

The cluster is running 13.2.8

I'd be very grateful for any tips

with best regards

nikola ciprich

--
-
Ing. Nikola CIPRICH
LinuxBox.cz, s.r.o.
28.rijna 168, 709 00 Ostrava

tel.:   +420 591 166 214
fax:+420 596 621 273
mobil:  +420 777 093 799
www.linuxbox.cz

mobil servis: +420 737 238 656
email servis: ser...@linuxbox.cz
-



--
-
Ing. Nikola CIPRICH
LinuxBox.cz, s.r.o.
28.rijna 168, 709 00 Ostrava

tel.:   +420 591 166 214
fax:+420 596 621 273
mobil:  +420 777 093 799
www.linuxbox.cz

mobil servis: +420 737 238 656
email servis: ser...@linuxbox.cz
-
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


--
Jan Fajerski
Senior Software Engineer Enterprise Storage
SUSE Software Solutions Germany GmbH
Maxfeldstr. 5, 90409 Nürnberg, Germany
(HRB 36809, AG Nürnberg)
Geschäftsführer: Felix Imendörffer
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: ceph prometheus module no export content

2020-02-26 Thread Jan Fajerski

On Thu, Feb 27, 2020 at 10:20:07AM +0800, 黄明友 wrote:

  hi,all
  I had enable the prometheus module on my ceph cluster , the
  ceph version
  is ceph version 14.2.5 (ad5bd132e1492173c85fda2cc863152730b16a92) nauti
  lus (stable).
  when I eanble this modle, can get exported content from the prometheus
  module.
  but , when I restart all the ceph cluster's node , the module can't
  expor any content , curl http://host/metrics ; get blan;
  the tcp/9283 port is listen;   I had disable and enable this module,but
  still no content export.
You're probably querying a standby mgr daemon. Check ceph -s for the current 
active and query that one. For a prometheus server you always want to scrape all 
mgr daemons. Also see https://docs.ceph.com/docs/master/mgr/prometheus/

  黄明友
  IT基础架构部经理
  V.Photos 云摄影
  移动电话:  +86 13540630430
  客服电话:400 - 806 - 5775
  电子邮件:  hmy@v.photos
  官方网址:  www.v.photos
  上海 黄浦区中山东二路88号外滩SOHO3Q F栋 2层
  北京 朝阳区光华路9号光华路SOHO二期南二门SOHO3Q 1层
  广州 天河区林和中路136号天誉花园二期3Wcoffice 天誉青创社区
  深圳 南山区蛇口网谷科技大厦二期A座102网谷双创街 1层
  成都 成华区建设路世贸广场 7层
  [cid:__aliyun158277000719672821]





___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io



--
Jan Fajerski
Senior Software Engineer Enterprise Storage
SUSE Software Solutions Germany GmbH
Maxfeldstr. 5, 90409 Nürnberg, Germany
(HRB 36809, AG Nürnberg)
Geschäftsführer: Felix Imendörffer
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Nautilus 14.2.6 ceph-volume bluestore _read_fsid unparsable uuid

2020-01-29 Thread Jan Fajerski
On Tue, Jan 28, 2020 at 08:03:35PM +0100, bauen1 wrote:
>Hi,
>
>I've run into the same issue while testing:
>
>ceph version 14.2.6 (f0aa067ac7a02ee46ea48aa26c6e298b5ea272e9) 
>nautilus (stable)
>
>debian bullseye
>
>Ceph was installed using ceph-ansible on a vm from the repo 
>http://download.ceph.com/debian-nautilus
>
>The output of `sudo sh -c 'CEPH_VOLUME_DEBUG=true ceph-volume 
>--cluster test lvm batch --bluestore /dev/vdb'` has been attached.
Thx, I opened https://tracker.ceph.com/issues/43868.
This looks like a bluestore/osd issue to me, though it might end up being 
ceph-volumes fault.
>
>Also worth noting might be that '/var/lib/ceph/osd/test-0/fsid' is 
>empty (but I don't know too much about the internals)
>
>- bauen1
>
>On 1/28/20 4:54 PM, Dave Hall wrote:
>>Jan,
>>
>>Unfortunately I'm under immense pressure right now to get some form 
>>of Ceph into production, so it's going to be Luminous for now, or 
>>maybe a live upgrade to Nautilus without recreating the OSDs (if 
>>that's possible).
>>
>>The good news is that in the next couple months I expect to add more 
>>hardware that should be nearly identical.  I will gladly give it a 
>>go at that time and see if I can recreate.  (Or, if I manage to 
>>thoroughly crash my current fledgling cluster, I'll give it another 
>>go on one node while I'm up all night recovering.)
>>
>>If you could tell me where to look I'd gladly read some code and see 
>>if I can find anything that way.  Or if there's any sort of design 
>>document describing the deep internals I'd be glad to scan it to see 
>>if I've hit a corner case of some sort.  Actually, I'd be interested 
>>in reading those documents anyway if I could.
>>
>>Thanks.
>>
>>-Dave
>>
>>Dave Hall
>>
>>On 1/28/2020 3:05 AM, Jan Fajerski wrote:
>>>On Mon, Jan 27, 2020 at 03:23:55PM -0500, Dave Hall wrote:
>>>>All,
>>>>
>>>>I've just spent a significant amount of time unsuccessfully chasing
>>>>the  _read_fsid unparsable uuid error on Debian 10 / Natilus 14.2.6.
>>>>Since this is a brand new cluster, last night I gave up and moved back
>>>>to Debian 9 / Luminous 12.2.11.  In both cases I'm using the packages
>>>>from Debian Backports with ceph-ansible as my deployment tool.
>>>>Note that above I said 'the _read_fsid unparsable uuid' error. I've
>>>>searched around a bit and found some previously reported issues, but I
>>>>did not see any conclusive resolutions.
>>>>
>>>>I would like to get to Nautilus as quickly as possible, so I'd gladly
>>>>provide additional information to help track down the cause of this
>>>>symptom.  I can confirm that, looking at the ceph-volume.log on the
>>>>OSD host I see no difference between the ceph-volume lvm batch command
>>>>generated by the ceph-ansible versions associated with these two Ceph
>>>>releases:
>>>>
>>>>   ceph-volume --cluster ceph lvm batch --bluestore --yes
>>>>   --block-db-size 133358734540 /dev/sdc /dev/sdd /dev/sde /dev/sdf
>>>>   /dev/sdg /dev/sdh /dev/sdi /dev/sdj /dev/nvme0n1
>>>>
>>>>Note that I'm using --block-db-size to divide my NVMe into 12 segments
>>>>as I have 4 empty drive bays on my OSD servers that I may eventually
>>>>be able to fill.
>>>>
>>>>My OSD hardware is:
>>>>
>>>>   Disk /dev/nvme0n1: 1.5 TiB, 1600321314816 bytes, 3125627568 sectors
>>>>   Disk /dev/sdc: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>>>>   Disk /dev/sdd: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>>>>   Disk /dev/sde: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>>>>   Disk /dev/sdf: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>>>>   Disk /dev/sdg: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>>>>   Disk /dev/sdh: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>>>>   Disk /dev/sdi: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>>>>   Disk /dev/sdj: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>>>>
>>>>I'd send the output of ceph-volume inventory on Luminous, but I'm
>>>>getting  -->: KeyError: 'human_readable_size'.
>>>>
>>>>Please let me know if I can provide any further information.
>>>Mind re-running you ceph-volume command with  debug output
>>>enabled:
>>>CEPH_VOLUME_DEBUG=true ceph-volume --cluster ceph lvm ba

[ceph-users] Re: Nautilus 14.2.6 ceph-volume bluestore _read_fsid unparsable uuid

2020-01-28 Thread Jan Fajerski
On Mon, Jan 27, 2020 at 03:23:55PM -0500, Dave Hall wrote:
>All,
>
>I've just spent a significant amount of time unsuccessfully chasing 
>the  _read_fsid unparsable uuid error on Debian 10 / Natilus 14.2.6.  
>Since this is a brand new cluster, last night I gave up and moved back 
>to Debian 9 / Luminous 12.2.11.  In both cases I'm using the packages 
>from Debian Backports with ceph-ansible as my deployment tool.
>
>Note that above I said 'the _read_fsid unparsable uuid' error. I've 
>searched around a bit and found some previously reported issues, but I 
>did not see any conclusive resolutions.
>
>I would like to get to Nautilus as quickly as possible, so I'd gladly 
>provide additional information to help track down the cause of this 
>symptom.  I can confirm that, looking at the ceph-volume.log on the 
>OSD host I see no difference between the ceph-volume lvm batch command 
>generated by the ceph-ansible versions associated with these two Ceph 
>releases:
>
>   ceph-volume --cluster ceph lvm batch --bluestore --yes
>   --block-db-size 133358734540 /dev/sdc /dev/sdd /dev/sde /dev/sdf
>   /dev/sdg /dev/sdh /dev/sdi /dev/sdj /dev/nvme0n1
>
>Note that I'm using --block-db-size to divide my NVMe into 12 segments 
>as I have 4 empty drive bays on my OSD servers that I may eventually 
>be able to fill.
>
>My OSD hardware is:
>
>   Disk /dev/nvme0n1: 1.5 TiB, 1600321314816 bytes, 3125627568 sectors
>   Disk /dev/sdc: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>   Disk /dev/sdd: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>   Disk /dev/sde: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>   Disk /dev/sdf: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>   Disk /dev/sdg: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>   Disk /dev/sdh: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>   Disk /dev/sdi: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>   Disk /dev/sdj: 10.9 TiB, 12000138625024 bytes, 23437770752 sectors
>
>I'd send the output of ceph-volume inventory on Luminous, but I'm 
>getting  -->: KeyError: 'human_readable_size'.
>
>Please let me know if I can provide any further information.
Mind re-running you ceph-volume command with  debug output 
enabled:
CEPH_VOLUME_DEBUG=true ceph-volume --cluster ceph lvm batch --bluestore ...

Ideally you could also openen a bug report here 
https://tracker.ceph.com/projects/ceph-volume/issues/new

Thanks!
>
>Thanks.
>
>-Dave
>
>-- 
>Dave Hall
>Binghamton University
>
>___
>ceph-users mailing list -- ceph-users@ceph.io
>To unsubscribe send an email to ceph-users-le...@ceph.io

-- 
Jan Fajerski
Senior Software Engineer Enterprise Storage
SUSE Software Solutions Germany GmbH
Maxfeldstr. 5, 90409 Nürnberg, Germany
(HRB 36809, AG Nürnberg)
Geschäftsführer: Felix Imendörffer
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Ceph-volume lvm batch: strategy changed after filtering

2020-01-24 Thread Jan Fajerski
On Fri, Jan 24, 2020 at 11:48:10AM +, Stolte, Felix wrote:
>Hey guys,
>
>I’m struggling with the ceph-volume command in nautilus 14.2.6. I have 12 
>disks on each server,  3 of them ssds (sda,sdb,sdc) and 9 spinning disks (sdd 
>.. sdl). Initial deploy witch ceph-volume batch works fine, one ssd is used 
>for wal and db for 3 spinning disks. But running the ‘ceph-volume lvm batch 
>--report /dev/sda /dev/sdb /dev/sdc /dev/sdd /dev/sde /dev/sdf /dev/sdg 
>/dev/sdh /dev/sdi /dev/sdj /dev/sdk /dev/sdl’ after the initial deploy I get 
>'Aborting because strategy changed from bluestore.MixedType to 
>bluestore.SingleType after filtering'
>
>Looking at /var/log/ceph/ceph-volume.log I see:
>[ceph_volume.devices.lvm.batch][INFO  ] Ignoring devices already used by ceph: 
>/dev/sdd, /dev/sde, /dev/sdf, /dev/sdg, /dev/sdh, /dev/sdi, /dev/sdj, 
>/dev/sdk, /dev/sdl
>
>I do understand that before filtering I have MixedType (ssd + sata) and after 
>filtering (only ssds left) it is single type. So the error message is correct.
>
>But why does ceph-volume filter out the spinning disks with data partitions 
>but not the ssds with wal/db? Is this intended?
Intended as in coded that way yes. See 
https://github.com/ceph/ceph/blob/master/src/ceph-volume/ceph_volume/util/device.py#L381
I'm not exactly sure why journal/wal/db was explicitly excluded here. I'm in 
the 
process of re-writing how batch creates OSDs to address inconsistencies like 
this.

Is there maybe space on your SSD drives? In this case its certainly valid to 
not 
filter those, as one might want to co-locate additional volumes.
What exactly are you trying to do?

Nest,
Jan
>
>Regards Felix
>
>
>IT-Services
>Telefon 02461 61-9243
>E-Mail: f.sto...@fz-juelich.de
>-
>-
>Forschungszentrum Juelich GmbH
>52425 Juelich
>Sitz der Gesellschaft: Juelich
>Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
>Vorsitzender des Aufsichtsrats: MinDir Dr. Karl Eugen Huthmacher
>Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
>Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
>Prof. Dr. Sebastian M. Schmidt
>-
>-
>
>
>_______
>ceph-users mailing list -- ceph-users@ceph.io
>To unsubscribe send an email to ceph-users-le...@ceph.io

-- 
Jan Fajerski
Senior Software Engineer Enterprise Storage
SUSE Software Solutions Germany GmbH
Maxfeldstr. 5, 90409 Nürnberg, Germany
(HRB 36809, AG Nürnberg)
Geschäftsführer: Felix Imendörffer
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: ceph-volume lvm filestore OSDs fail to start on reboot. Permission denied on journal partition

2020-01-23 Thread Jan Fajerski
On Wed, Jan 22, 2020 at 12:00:28PM -0500, Wesley Dillingham wrote:
>   After upgrading to Nautilus 14.2.6 from Luminous 12.2.12 we are seeing
>   the following behavior on OSDs which were created with "ceph-volume lvm
>   create --filestore --osd-id  --data  --journal "
>   Upon restart of the server containing these OSDs they fail to start
>   with the following error in the logs:
>2020-01-21 13:36:11.635 7fee633e8a80 -1 filestore(/var/lib/ceph/osd/ceph-199) 
>mo
>unt(1928): failed to open journal /var/lib/ceph/osd/ceph-199/journal: (13) 
>Permi
>ssion denied
>
>   /var/lib/ceph/osd/ceph-199/journal symlinks to /dev/sdc5 in our case
>   and inspecting the ownership on /dev/sdc5 it is root:root, chowning
>   that to ceph:ceph causes the osd to start and come back up and in near
>   instantly.
>   As a note these OSDs we experience this with are OSDs which have
>   previously failed and been replaced using the above ceph-volume, longer
>   running OSDs in the same server created with ceph-disk or ceph-volume
>   simple (that have a corresponding .json in /etc/ceph/osd) start up fine
>   and get ceph:ceph on their journal partition. Bluestore OSDs also do
>   not have any issue.
>   My hope is that I can preemptively fix these OSDs before shutting them
>   down so that reboots happen seamlessly. Thanks for any insight.
ceph-volume is supposed to take care of this via the ceph-volume@ systemd unit. 
 
This is a one shot unit, that should set things up and then start the osd.
The unit name is a bit convoluted: ceph-volume@-, there 
should 
be symbolic link in /etc/systemd/system/multi-user.target.wants/

You can also check cat /var/log/ceph/ceph-volume-systemd.log for any errors.
Feel free to open a tracker ticket on 
https://tracker.ceph.com/projects/ceph-volume

>
>   Respectfully,
>   Wes Dillingham
>   [1]w...@wesdillingham.com
>   [2]LinkedIn
>
>References
>
>   1. mailto:w...@wesdillingham.com
>   2. http://www.linkedin.com/in/wesleydillingham

>___
>ceph-users mailing list -- ceph-users@ceph.io
>To unsubscribe send an email to ceph-users-le...@ceph.io


-- 
Jan Fajerski
Senior Software Engineer Enterprise Storage
SUSE Software Solutions Germany GmbH
Maxfeldstr. 5, 90409 Nürnberg, Germany
(HRB 36809, AG Nürnberg)
Geschäftsführer: Felix Imendörffer
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: ceph-volume sizing osds

2019-12-14 Thread Jan Fajerski
On Fri, Dec 13, 2019 at 11:49:50PM +0100, Oscar Segarra wrote:
>   Hi,
>   I have recently started working with Ceph Nautilus release and I have
>   realized that you have to start working with LVM to create OSD instead
>   of the "old fashioned" ceph-disk.
>   In terms of performance and best practices, as far as I must use LVM I
>   can create volume groups that joins or extends two or more physical
>   disks. In this scenario (many disks for server) where ceph-volume is
>   manatory, It still remains the rule of one OSD for each physical
>   device? or I can reduce the number of OSDs?

That is hard to answer generically. ceph-volume will not create multi-disk 
volumes. One disk, one OSD is still the recommendation afaik.
However there are certainly scenarios where multi-disk OSDs are sensible. For 
example if you have a machine that has _a lot_ of disks attached but lacks CPU 
and RAM to run as many OSDs as disks, it can make sense to deploy one OSD on 
multiple disks. 

>   Thanks in advance,
>   Óscar

>___
>ceph-users mailing list
>ceph-us...@lists.ceph.com
>http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Prometheus endpoint hanging with 13.2.7 release?

2019-12-10 Thread Jan Fajerski
On Mon, Dec 09, 2019 at 05:01:04PM -0800, Paul Choi wrote:
>   Hello,
>   Anybody seeing the Prometheus endpoint hanging with the new 13.2.7
>   release?
>   With 13.2.6 the endpoint would respond with a payload of 15MB in less
>   than 10 seconds.

I'd guess its not the prometheus module itself:
$ git  diff v13.2.6 v13.2.7 src/pybind/mgr/prometheus
diff --git a/src/pybind/mgr/prometheus/module.py 
b/src/pybind/mgr/prometheus/module.py
index 2d4472434a..3a398e0b0c 100644
--- a/src/pybind/mgr/prometheus/module.py
+++ b/src/pybind/mgr/prometheus/module.py
@@ -142,7 +142,8 @@ class Metric(object):

  def promethize(path):
  ''' replace illegal metric name characters '''
-result = path.replace('.', '_').replace('+', 
'_plus').replace('::', '_')
+result = path.replace('.', '_').replace(
+'+', '_plus').replace('::', '_').replace(' ', '_')

  # Hyphens usually turn into underscores, unless they are
  # trailing
@@ -720,7 +721,8 @@ class Module(MgrModule):
  raise cherrypy.HTTPError(503, 'No MON connection')

  # Make the cache timeout for collecting configurable
-self.collect_timeout = self.get_localized_config('scrape_interval', 
5.0)
+self.collect_timeout = float(self.get_localized_config(
+'scrape_interval', 5.0))

  server_addr = self.get_localized_config('server_addr', DEFAULT_ADDR)
  server_port = self.get_localized_config('server_port', DEFAULT_PORT)

So the mgr would be a likely suspect. If you could open a tracker ticket, 
ideally with mgr debug logs attached, this can be looked at.

>   Now, if you restart ceph-mgr, the Prometheus endpoint responds quickly
>   for the first run, then successive runs get slower and slower, until it
>   takes several minutes.
>   I have no customization for the mgr module. Except for the Prometheus
>   module, the "status" module and "Zabbix" module are working fine.
>   This is on Ubuntu 16 LTS:
>   ii  ceph-mgr 13.2.7-1xenial
>   amd64manager for the ceph distributed storage
>   system
>   I'd love to know if there's a way to diagnose this issue - I tried
>   upping the debug ms level but that doesn't seem to yield useful
>   information.
>   I don't know if this useful, but "prometheus self-test" is fine too.
>   $ sudo ceph tell mgr.0 prometheus self-test
>   Self-test OK
>   pchoi@pchoi-desktop:~$ ceph mgr module ls
>   {
>   "enabled_modules": [
>   "prometheus",
>   "status",
>   "zabbix"
>   ],
>   "disabled_modules": [
>   {
>   "name": "balancer",
>   "can_run": true,
>   "error_string": ""
>   },
>   {
>   "name": "crash",
>   "can_run": true,
>   "error_string": ""
>   },
>   {
>   "name": "dashboard",
>   "can_run": true,
>   "error_string": ""
>   },
>   {
>   "name": "hello",
>   "can_run": true,
>   "error_string": ""
>   },
>   {
>   "name": "influx",
>   "can_run": false,
>   "error_string": "influxdb python module not found"
>   },
>   {
>   "name": "iostat",
>   "can_run": true,
>   "error_string": ""
>   },
>   {
>   "name": "localpool",
>   "can_run": true,
>   "error_string": ""
>   },
>   {
>   "name": "restful",
>   "can_run": true,
>   "error_string": ""
>   },
>   {
>   "name": "selftest",
>   "can_run": true,
>   "error_string": ""
>   },
>   {
>   "name": "smart",
>   "can_run": true,
>   "error_string": ""
>   },
>   {
>   "name": "telegraf",
>   "can_run": true,
>   "error_string": ""
>   },
>   {
>   "name": "telemetry",
>   "can_run": true,
>   "error_string": ""
>   }
>   ]
>   }
>   pchoi@pchoi-desktop:~$ ceph mgr services
>   {
>   "prometheus": "[1]http://woodenbox2:9283/;
>   }
>
>References
>
>   1. http://woodenbox2:9283/

>___
>ceph-users mailing list -- ceph-users@ceph.io
>To unsubscribe send an email to ceph-users-le...@ceph.io


-- 
Jan Fajerski
Senior Software Engineer Enterprise Storage
SUSE Software Solutions Germany GmbH
Maxfeldstr. 5, 90409 Nürnberg, Germany
(HRB 36809, AG Nürnberg)
Geschäftsführer: Felix Imendörffer
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: ceph-volume lvm create leaves half-built OSDs lying around

2019-09-11 Thread Jan Fajerski
On Wed, Sep 11, 2019 at 11:17:47AM +0100, Matthew Vernon wrote:
>Hi,
>
>We keep finding part-made OSDs (they appear not attached to any host,
>and down and out; but still counting towards the number of OSDs); we
>never saw this with ceph-disk. On investigation, this is because
>ceph-volume lvm create makes the OSD (ID and auth at least) too early in
>the process and is then unable to roll-back cleanly (because the
>bootstrap-osd credential isn't allowed to remove OSDs).
>
>As an example (very truncated):
>
>Running command: /usr/bin/ceph --cluster ceph --name
>client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring
>-i - osd new 20cea174-4c1b-4330-ad33-505a03156c33
>Running command: vgcreate --force --yes
>ceph-9d66ec60-c71b-49e0-8c1a-e74e98eafb0e /dev/sdbh
> stderr: Device /dev/sdbh not found (or ignored by filtering).
>  Unable to add physical volume '/dev/sdbh' to volume group
>'ceph-9d66ec60-c71b-49e0-8c1a-e74e98eafb0e'.
>--> Was unable to complete a new OSD, will rollback changes
>--> OSD will be fully purged from the cluster, because the ID was generated
>Running command: ceph osd purge osd.828 --yes-i-really-mean-it
> stderr: 2019-09-10 15:07:53.396528 7fbca2caf700 -1 auth: unable to find
>a keyring on
>/etc/ceph/ceph.client.admin.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin,:
>(2) No such file or directory
> stderr: 2019-09-10 15:07:53.397318 7fbca2caf700 -1 monclient:
>authenticate NOTE: no keyring found; disabled cephx authentication
>2019-09-10 15:07:53.397334 7fbca2caf700  0 librados: client.admin
>authentication error (95) Operation not supported
>
>This is annoying to have to clear up, and it seems to me could be
>avoided by either:
>
>i) ceph-volume should (attempt to) set up the LVM volumes  before
>making the new OSD id
>or
>ii) allow the bootstrap-osd credential to purge OSDs
>
>i) seems like clearly the better answer...?
Agreed. Would you mind opening a bug report on
https://tracker.ceph.com/projects/ceph-volume.

I have found other situation where a roll-back is working as it should, though
not with as much impact as this.
>
>Regards,
>
>Matthew
>




>_______
>ceph-users mailing list
>ceph-us...@lists.ceph.com
>http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


-- 
Jan Fajerski
Senior Software Engineer Enterprise Storage
SUSE Software Solutions Germany GmbH
(HRB 247165, AG München)
Geschäftsführer: Felix Imendörffer
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io