[ceph-users] Re: Running Mons on msgrv2/3300 only.

2020-12-09 Thread 胡 玮文
Out of curiosity, how DNS lookup [1] works with v2 only mons? Since we can’t 
specify whether the port is v1 or v2 in SRV records. Or it just works if I 
specify 3300 port in SRV records?

[1]: https://docs.ceph.com/en/latest/rados/configuration/mon-lookup-dns/

在 2020年12月9日,18:10,Wido den Hollander  写道:

On 08/12/2020 20:17, Wesley Dillingham wrote:
We rebuilt all of our mons in one cluster such that they bind only to port 3300 
with msgrv2. Previous to this we were binding to both 6789 and 3300. All of our 
server and client components are sufficiently new (14.2.x) and we haven’t 
observed any disruption but I am inquiring if this may be problematic for any 
unforeseen reason. We don’t intend to have any older clients connecting. 
https://docs.ceph.com/en/latest/rados/configuration/msgr2/ doesn’t mention much 
about running with only v2 so I just want to make sure we aren’t setting 
ourselves up for trouble. Thanks.

No, not really. I have a couple of those servers as well where v1 was disabled 
completely.

Wido

--
Respectfully,
Wes Dillingham
Site Reliability Engineer IV
Storage Engineering / Ceph
wdillingham(at)godaddy.com
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: CentOS

2020-12-09 Thread Mark Nelson

Hi Folks,


We have CentOS Stream 8 installed on some of our community performance 
test gear.  A couple of months ago there was a change introduced to the 
repository names that broke install-deps.sh, but that has now been 
fixed.  Otherwise, Ceph master builds fine on these nodes and I don't 
see why packages would have any issue.  I believe Octopus and Nautilus 
would also need that change to install-deps.sh to build properly, but 
otherwise I don't believe there are any major issues preventing them 
from working.



Mark


On 12/8/20 3:01 PM, Marc Roos wrote:
  
I did not. Thanks for the info. But if I understand this[1] explanation

correctly. CentOS stream is some sort of trial environment for rhel. So
who is ever going to put SDS on such an OS?

Last post on this blog "But if you read the FAQ, you also learn that
once they start work on RHEL 9, CentOS Stream 8 ceases to exist..."

[1]
https://www.youtube.com/watch?v=IEEdOogPMY8






-Original Message-
To: ceph-users@ceph.io
Subject: [ceph-users] CentOS

All;

As you may or may not know; this morning RedHat announced the end of
CentOS as a rebuild distribution[1].  "CentOS" will be retired in favor
of the recently announced "CentOS Stream."

Can Ceph be installed on CentOS Stream?

Since CentOS Stream is currently at 8, the question really is: Can Ceph
Octopus be installed on CentOS Stream 8?  How about Nautilus?

Thank you,

Dominic L. Hilsbos, MBA
Director - Information Technology
Perform Air International Inc.
dhils...@performair.com
www.PerformAir.com

[1: https://blog.centos.org/2020/12/future-is-centos-stream/]
___
ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an
email to ceph-users-le...@ceph.io

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Larger number of OSDs, cheroot, cherrypy, limits + containers == broken

2020-12-09 Thread David Orman
Ken,

We have rebuilt the container images of 15.2.7 with this RPM applied, and
will be deploying it to a larger (504 OSD) cluster to test - this cluster
had the issue previously until we disabled polling via Prometheus. We will
update as soon as it's run for a day or two and we've been able to verify
the mgr issues we saw no longer occur after extended polling via external
and internal prometheus instances.

Thank you again for the quick update, we'll let you know as soon as we have
more feedback,
David

On Tue, Dec 8, 2020 at 10:37 AM David Orman  wrote:

> Hi Ken,
>
> Thank you for the update! As per:
> https://github.com/ceph/ceph-container/issues/1748
>
> We implemented the (dropping ulimit to 1024:4096 for mgr) suggested change
> last night, and on our test cluster of 504 OSDs, being polled by the
> internal prometheus and our external instance, the mgrs stopped responding
> and dropped out of the cluster entirely. This is impacting not just
> metrics, but the mgr itself. I think this is a high priority issue, as
> metrics are critical for prod, but mgr itself seems to be impacted on a
> moderately sized cluster.
>
> Respectfully,
> David Orman
>
> On Mon, Dec 7, 2020 at 1:50 PM Ken Dreyer  wrote:
>
>> Thanks for bringing this up.
>>
>> We need to update Cheroot in Fedora and EPEL 8. I've opened
>> https://src.fedoraproject.org/rpms/python-cheroot/pull-request/3 to
>> get this into Fedora first.
>>
>> I've published an el8 RPM at
>> https://fedorapeople.org/~ktdreyer/bz1868629/ for early testing. I can
>> bring up a "hello world" cherrypy app with this, but I've not tested
>> it with Ceph.
>>
>> - Ken
>>
>> On Mon, Dec 7, 2020 at 9:57 AM David Orman  wrote:
>> >
>> > Hi,
>> >
>> > We have a ceph 15.2.7 deployment using cephadm under podman w/ systemd.
>> > We've run into what we believe is:
>> >
>> > https://github.com/ceph/ceph-container/issues/1748
>> > https://tracker.ceph.com/issues/47875
>> >
>> > In our case, eventually the mgr container stops emitting
>> output/logging. We
>> > are polling with external prometheus clusters, which is likely what
>> > triggers the issue, as it appears some amount of time after the
>> container
>> > is spawned.
>> >
>> > Unfortunately, setting limits in the systemd service file for the mgr
>> > service on the host OS doesn't work, nor does modifying the unit.run
>> file
>> > which is used to start the container under podman to include the
>> --ulimit
>> > settings as suggested. Looking inside the container:
>> >
>> > lib/systemd/system/ceph-mgr@.service:LimitNOFILE=1048576
>> >
>> > This prevents us from deploying medium to large ceph clusters, so I
>> would
>> > argue it's a high priority bug that should not be closed, unless there
>> is a
>> > workaround that works until EPEL 8 contains the fixed version of cheroot
>> > and the ceph containers include it.
>> >
>> > My understanding is this was fixed in cheroot 8.4.0:
>> >
>> > https://github.com/cherrypy/cheroot/issues/249
>> > https://github.com/cherrypy/cheroot/pull/301
>> >
>> > Thank you in advance for any suggestions,
>> > David
>> > ___
>> > ceph-users mailing list -- ceph-users@ceph.io
>> > To unsubscribe send an email to ceph-users-le...@ceph.io
>> >
>>
>>
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] DocuBetter Meeting cancelled this week.

2020-12-09 Thread John Zachary Dover
The DocuBetter meeting scheduled for one hour from the time of this email
is cancelled.

Send documentation concerns to me at zac.do...@gmail.com.

Zac Dover
Upstream Docs
Ceph
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


[ceph-users] Re: Running Mons on msgrv2/3300 only.

2020-12-09 Thread Wido den Hollander




On 08/12/2020 20:17, Wesley Dillingham wrote:

We rebuilt all of our mons in one cluster such that they bind only to port 3300 
with msgrv2. Previous to this we were binding to both 6789 and 3300. All of our 
server and client components are sufficiently new (14.2.x) and we haven’t 
observed any disruption but I am inquiring if this may be problematic for any 
unforeseen reason. We don’t intend to have any older clients connecting. 
https://docs.ceph.com/en/latest/rados/configuration/msgr2/ doesn’t mention much 
about running with only v2 so I just want to make sure we aren’t setting 
ourselves up for trouble. Thanks.




No, not really. I have a couple of those servers as well where v1 was 
disabled completely.


Wido


--
Respectfully,

Wes Dillingham
Site Reliability Engineer IV
Storage Engineering / Ceph
wdillingham(at)godaddy.com

___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io