Re: [ceph-users] PG calculator improvement

2017-04-13 Thread Michael Kidd
Hello Frédéric,
  Thank you very much for the input.  I would like to ask for some feedback
from you, as well as the ceph-users list at large.

The PGCalc tool was created to help steer new Ceph users in the right
direction, but it's certainly difficult to account for every possible
scenario.  I'm struggling to find a way to implement something that would
work better for the scenario that you (Frédéric) describe, while still
being a useful starting point for the novice / more mainstream use cases.
I've also gotten complaints at the other end of the spectrum, that the tool
expects the user to know too much already, so accounting for the number of
objects is bound to add to this sentiment.

As the Ceph user base expands and the use cases diverge, we are definitely
finding more edge cases that are causing pain.  I'd love to make something
to help prevent these types of issues, but again, I worry about the
complexity introduced.

With this, I see a few possible ways forward:
* Simply re-wroding the %data to be % object count -- but this seems more
abstract, again leading to more confusion of new users.
* Increase complexity of the PG Calc tool, at the risk of further
alienating novice/mainstream users
* Add a disclaimer about the tool being a base for decision making, but
that certain edge cases require adjustments to the recommended PG count
and/or ceph.conf & sysctl values.
* Add a disclaimer urging the end user to secure storage consulting if
their use case falls into certain categories or they are new to Ceph to
ensure the cluster will meet their needs.

Having been on the storage consulting team and knowing the expertise they
have, I strongly believe that newcomers to Ceph (or new use cases inside of
established customers) should secure consulting before final decisions are
made on hardware... let alone the cluster is deployed.  I know it seems a
bit self-serving to make this suggestion as I work at Red Hat, but there is
a lot on the line when any establishment is storing potentially business
critical data.

I suspect the answer lies in a combination of the above or in something
I've not thought of.  Please do weigh in as any and all suggestions are
more than welcome.

Thanks,
Michael J. Kidd
Principal Software Maintenance Engineer
Red Hat Ceph Storage
+1 919-442-8878


On Wed, Apr 12, 2017 at 6:35 AM, Frédéric Nass <
frederic.n...@univ-lorraine.fr> wrote:

>
> Hi,
>
> I wanted to share a bad experience we had due to how the PG calculator
> works.
>
> When we set our production cluster months ago, we had to decide on the
> number of PGs to give to each pool in the cluster.
> As you know, the PG calc would recommended to give a lot of PGs to heavy
> pools in size, regardless the number of objects in the pools. How bad...
>
> We essentially had 3 pools to set on 144 OSDs :
>
> 1. a EC5+4 pool for the radosGW (.rgw.buckets) that would hold 80% of all
> datas in the cluster. PG calc recommended 2048 PGs.
> 2. a EC5+4 pool for zimbra's data (emails) that would hold 20% of all
> datas. PG calc recommended 512 PGs.
> 3. a replicated pool for zimbra's metadata (null size objects holding
> xattrs - used for deduplication) that would hold 0% of all datas. PG calc
> recommended 128 PGs, but we decided on 256.
>
> With 120M of objects in pool #3, as soon as we upgraded to Jewel, we hit
> the Jewel scrubbing bug (OSDs flapping).
> Before we could upgrade to patched Jewel, scrub all the cluster again
> prior to increasing the number of PGs on this pool, we had to take more
> than a hundred of snapshots (for backup/restoration purposes), with the
> number of objects still increasing in the pool. Then when a snapshot was
> removed, we hit the current Jewel snap trimming bug affecting pools with
> too many objects for the number of PGs. The only way we could stop the
> trimming was to stop OSDs resulting in PGs being degraded and not trimming
> anymore (snap trimming only happens on active+clean PGs).
>
> We're now just getting out of this hole, thanks to Nick's post regarding
> osd_snap_trim_sleep and RHCS support expertise.
>
> If the PG calc had considered not only the pools weight but also the
> number of expected objects in the pool (which we knew by that time), we
> wouldn't have it these 2 bugs.
> We hope this will help improving the ceph.com and RHCS PG calculators.
>
> Regards,
>
> Frédéric.
>
> --
>
> Frédéric Nass
>
> Sous-direction Infrastructures
> Direction du Numérique
> Université de Lorraine
>
> Tél : +33 3 72 74 11 35
>
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Javascript error at http://ceph.com/pgcalc/

2017-01-11 Thread Michael Kidd
Hello John,
  Thanks for the bug report.  Unfortunately, I'm not able to reproduce the
error.  I tested from both Firefox and Chrome on linux.  Can you let me
know what os/browser you're using?  Also, I've not tested any non 'en-US'
characters, so I can't attest to how it will behave with other alphabets.

Thanks,

Michael J. Kidd
Sr. Software Maintenance Engineer
Red Hat Ceph Storage
+1 919-442-8878 <(919)%20442-8878>

On Wed, Jan 11, 2017 at 6:01 PM, 林自均  wrote:

> Hi Michael,
>
> Thanks for your link!
>
> However, when I am using your clone of pgcalc, the newly created pool
> didn't follow my values in the "Add Pool" dialog. For example, no matter
> what I fill in "Pool Name", I always get "newPool" as the name.
>
> By the way, where can I find the git repository of pgcalc? I can't find it
> at https://github.com/ceph/. Thanks!
>
> Best,
> John Lin
>
> Michael Kidd  於 2017年1月12日 週四 上午7:02寫道:
>
>> Hello John,
>>   Apologies for the error.  We will be working to correct it, but in the
>> interim, you can use http://linuxkidd.com/ceph/pgcalc.html
>>
>> Thanks,
>>
>> Michael J. Kidd
>> Sr. Software Maintenance Engineer
>> Red Hat Ceph Storage
>> +1 919-442-8878 <(919)%20442-8878>
>>
>> On Wed, Jan 11, 2017 at 12:03 AM, 林自均  wrote:
>>
>> Hi all,
>>
>> I am not sure if this is the correct mailing list. Correct me if I am
>> wrong.
>>
>> I failed to add a pool at http://ceph.com/pgcalc/ because of a
>> Javascript error:
>>
>> (index):345 Uncaught TypeError: $(...).dialog is not a function
>> at addPool (http://ceph.com/pgcalc/:345:31)
>> at HTMLButtonElement.onclick (http://ceph.com/pgcalc/:534:191)
>>
>> Where can I report this error? Thanks.
>>
>> Best,
>> John Lin
>>
>> ___
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Javascript error at http://ceph.com/pgcalc/

2017-01-11 Thread Michael Kidd
Hello John,
  Apologies for the error.  We will be working to correct it, but in the
interim, you can use http://linuxkidd.com/ceph/pgcalc.html

Thanks,

Michael J. Kidd
Sr. Software Maintenance Engineer
Red Hat Ceph Storage
+1 919-442-8878

On Wed, Jan 11, 2017 at 12:03 AM, 林自均  wrote:

> Hi all,
>
> I am not sure if this is the correct mailing list. Correct me if I am
> wrong.
>
> I failed to add a pool at http://ceph.com/pgcalc/ because of a Javascript
> error:
>
> (index):345 Uncaught TypeError: $(...).dialog is not a function
> at addPool (http://ceph.com/pgcalc/:345:31)
> at HTMLButtonElement.onclick (http://ceph.com/pgcalc/:534:191)
>
> Where can I report this error? Thanks.
>
> Best,
> John Lin
>
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Calculating PG in an mixed environment

2016-03-15 Thread Michael Kidd
Hello Martin,
  The proper way is to perform the following process:

For all Pools utilizing the same bucket of OSDs:
(Pool1_pg_num * Pool1_size) + (Pool2_pg_num * Pool2_size) + ...
(Pool(n)_pg_num * Pool(n)_size)
--
OSD count

This value should be between 100 and 200 PGs and is the actual ratio of PGs
per OSD in that bucket of OSDs.

For the actual recommendation from Ceph Devs (and written by myself),
please see:
http://ceph.com/pgcalc/

NOTE: The tool is partially broken, but the explanation at the top/bottom
is sound.  I'll work to get the tool fully functional again.

Thanks,

Michael J. Kidd
Sr. Software Maintenance Engineer
Red Hat Ceph Storage
+1 919-442-8878

On Tue, Mar 15, 2016 at 11:41 AM, Martin Palma  wrote:

> Hi all,
>
> The documentation [0] gives us the following formula for calculating
> the number of PG if the cluster is bigger than 50 OSDs:
>
>  (OSDs * 100)
> Total PGs =  
>  pool size
>
> When we have mixed storage server (HDD disks and SSD disks) and we
> have defined different roots in our crush map to map some pools only
> to HDD disk and some to SSD disks like described by Sebastien Han [1].
>
> In the above formula what number of OSDs should be use to calculate
> the  PGs for a pool only on the HDD disks? The total number of OSDs in
> a cluster or only the number of OSDs which have an HDD disk as
> backend?
>
> Best,
> Martin
>
>
> [0]
> http://docs.ceph.com/docs/master/rados/operations/placement-groups/#choosing-the-number-of-placement-groups
> [1]
> http://www.sebastien-han.fr/blog/2014/08/25/ceph-mix-sata-and-ssd-within-the-same-box/
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Why is this pg incomplete?

2016-01-04 Thread Michael Kidd
Bryan,
  If you can read the disk that was osd.102, you may wish to attempt this
process to recover your data:
https://ceph.com/community/incomplete-pgs-oh-my/

Good luck!

Michael J. Kidd
Sr. Software Maintenance Engineer
Red Hat Ceph Storage

On Mon, Jan 4, 2016 at 8:32 AM, Bryan Wright  wrote:

> Gregory Farnum  writes:
>
> > I can't parse all of that output, but the most important and
> > easiest-to-understand bit is:
> > "blocked_by": [
> > 102
> > ],
> >
> > And indeed in the past_intervals section there are a bunch where it's
> > just 102. You really want min_size >=2 for exactly this reason. :/ But
> > if you get 102 up stuff should recover; if you can't you can mark it
> > as "lost" and RADOS ought to resume processing, with potential
> > data/metadata loss...
> > -Greg
> >
>
>
> Ack!  I thought min_size was 2, but I see:
>
> ceph osd pool get data min_size
> min_size: 1
>
> Well that's a fine kettle of fish.
>
> The osd in question (102) has actually already been marked as lost, via
> "ceph osd lost 102 --yes-i-really-mean-it", and it shows up in "ceph osd
> tree" as "DNE".  If I can manage to read the disk, how should I try to add
> it back in?
>
>
>
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] ceph-mon always election when change crushmap in firefly

2015-09-23 Thread Michael Kidd
Hello Alexander,
  One other point on your email.. You indicate you desire each OSD to have
~100 PGs, but depending on your pool size, it seems you may have forgetting
about the additional PGs associated with replication itself.

Assuming 3x replication in your environment:
70,000 * 3

800 OSDs

=~ 262.5 PGs per OSD on average

While this PG to OSD ratio shouldn't cause significant pain, I would not go
to any higher PG count without adding more spindles.

For more specific PG count guidance and modeling, please see:
http://ceph.com/pgcalc

Hope this helps,

Michael J. Kidd
Sr. Storage Consultant
Red Hat Global Storage Consulting
+1 919-442-8878

On Wed, Sep 23, 2015 at 8:34 AM, Sage Weil  wrote:

> On Wed, 23 Sep 2015, Alexander Yang wrote:
> > hello,
> > We use Ceph+Openstack in our private cloud. In our cluster, we
> have
> > 5 mons and 800 osds, the Capacity is about 1Pb. And run about 700 vms and
> > 1100 volumes,
> > recently, we increase our pg_num , now the cluster have about
> 7
> > pgs. In my real intention? I want every osd have 100pgs. but after
> increase
> > pg_num, I find I'm wrong. Because the different crush weight for
> different
> > osd, the osd's pg_num is different, some osd have exceed  500pgs.
> > Now, the problem is  appear?cause some reason when i want to
> change
> > some osd  weight, that means change the crushmap.  This change cause
> about
> > 0.03% data to migrate. the mon is always begin to election. It's will
> hung
> > the cluster, and when they end, the  original  leader still is the new
> > leader. And during the mon eclection?On the upper layer, vm have too many
> > slow request will appear. so now i dare to do any operation about change
> > crushmap. But i worry about an important thing, If  when our cluster
> down
> >  one host even down one rack.   By the time, the cluster curshmap will
> > change large, and the migrate data also large. I worry the cluster will
> > hung  long time. and result on upper layer, all vm became to  shutdown.
> > In my opinion, I guess when I change the crushmap,* the leader
> mon
> > maybe calculate the too many information*, or* too many client want to
> get
> > the new crushmap from leader mon*.  It must be hung the mon thread, so
> the
> > leader mon can't heatbeat to other mons, the other mons think the leader
> is
> > down then begin the new election.  I am sorry if i guess is wrong.
> > The crushmap in accessory. So who can give me some advice or
> guide,
> > Thanks very much!
>
> There were huge improvements made in hammer in terms of mon efficiency in
> these cases where it is under load.  I recommend upgrading as that will
> help.
>
> You can also mitigate the problem somewhat by adjusting the mon_lease and
> associated settings up.  Scale all of mon_lease, mon_lease_renew_interval,
> mon_lease_ack_timeout, mon_accept_timeout by 2x or 3x.
>
> It also sounds like you may be using some older tunables/settings
> for your pools or crush rules.  Can you attach the output of 'ceph osd
> dump' and 'ceph osd crush dump | tail -n 20' ?
>
> sage
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


Re: [ceph-users] Firefly - Giant : CentOS 7 : install failed ceph-deploy

2015-05-04 Thread Michael Kidd
For Firefly / Giant installs, I've had success with the following:

yum install ceph ceph-common --disablerepo=base --disablerepo=epel

Let us know if this works for you as well.

Thanks,

Michael J. Kidd
Sr. Storage Consultant
Inktank Professional Services
 - by Red Hat

On Wed, Apr 8, 2015 at 8:55 PM, Travis Rhoden  wrote:

> I did also confirm that, as Ken mentioned, this is not a problem on Hammer
> since Hammer includes the package split (python-ceph became python-rados
> and python-rbd).
>
>  - Travis
>
> On Wed, Apr 8, 2015 at 5:00 PM, Travis Rhoden  wrote:
>
>> Hi Vickey,
>>
>> The easiest way I know of to get around this right now is to add the
>> following line in section for epel in /etc/yum.repos.d/epel.repo
>>
>> exclude=python-rados python-rbd
>>
>> So this is what my epel.repo file looks like: http://fpaste.org/208681/
>>
>> It is those two packages in EPEL that are causing problems.  I also tried
>> enabling epel-testing, but that didn't work either.
>>
>> Unfortunately you would need to add this line on each node where Ceph
>> Giant is being installed.
>>
>>  - Travis
>>
>> On Wed, Apr 8, 2015 at 4:11 PM, Vickey Singh > > wrote:
>>
>>> Community , need help.
>>>
>>>
>>> -VS-
>>>
>>> On Wed, Apr 8, 2015 at 4:36 PM, Vickey Singh <
>>> vickey.singh22...@gmail.com> wrote:
>>>
 Any suggestion  geeks


 VS

 On Wed, Apr 8, 2015 at 2:15 PM, Vickey Singh <
 vickey.singh22...@gmail.com> wrote:

>
> Hi
>
>
> The below suggestion also didn’t worked
>
>
> Full logs here : http://paste.ubuntu.com/10771939/
>
>
>
>
> [root@rgw-node1 yum.repos.d]# yum --showduplicates list ceph
>
> Loaded plugins: fastestmirror, priorities
>
> Loading mirror speeds from cached hostfile
>
>  * base: mirror.zetup.net
>
>  * epel: ftp.fi.muni.cz
>
>  * extras: mirror.zetup.net
>
>  * updates: mirror.zetup.net
>
> 25 packages excluded due to repository priority protections
>
> Available Packages
>
> ceph.x86_64
> 0.80.6-0.el7.centos
> Ceph
>
> ceph.x86_64
> 0.80.7-0.el7.centos
> Ceph
>
> ceph.x86_64
> 0.80.8-0.el7.centos
> Ceph
>
> ceph.x86_64
> 0.80.9-0.el7.centos
> Ceph
>
> [root@rgw-node1 yum.repos.d]#
>
>
>
>
>
> Its not able to install latest available package , yum is getting
> confused with other DOT releases.
>
>
> Any other suggestion to fix this ???
>
>
>
> --> Processing Dependency: libboost_system-mt.so.1.53.0()(64bit) for
> package: librbd1-0.80.9-0.el7.centos.x86_64
>
> --> Processing Dependency: libboost_thread-mt.so.1.53.0()(64bit) for
> package: librbd1-0.80.9-0.el7.centos.x86_64
>
> --> Finished Dependency Resolution
>
> Error: Package: librbd1-0.80.7-0.el7.centos.x86_64 (Ceph)
>
>Requires: libboost_system-mt.so.1.53.0()(64bit)
>
> Error: Package: ceph-0.80.7-0.el7.centos.x86_64 (Ceph)
>
>Requires: libboost_system-mt.so.1.53.0()(64bit)
>
> Error: Package: ceph-0.80.7-0.el7.centos.x86_64 (Ceph)
>
>Requires: libaio.so.1(LIBAIO_0.4)(64bit)
>
> Error: Package: ceph-common-0.80.7-0.el7.centos.x86_64 (Ceph)
>
>Requires: libboost_thread-mt.so.1.53.0()(64bit)
>
> Error: Package: ceph-common-0.80.7-0.el7.centos.x86_64 (Ceph)
>
>Requires: librados2 = 0.80.7-0.el7.centos
>
>Available: librados2-0.80.6-0.el7.centos.x86_64 (Ceph)
>
>librados2 = 0.80.6-0.el7.centos
>
>Available: librados2-0.80.7-0.el7.centos.x86_64 (Ceph)
>
>librados2 = 0.80.7-0.el7.centos
>
>Available: librados2-0.80.8-0.el7.centos.x86_64 (Ceph)
>
>librados2 = 0.80.8-0.el7.centos
>
>Installing: librados2-0.80.9-0.el7.centos.x86_64 (Ceph)
>
>librados2 = 0.80.9-0.el7.centos
>
> Error: Package: libcephfs1-0.80.7-0.el7.centos.x86_64 (Ceph)
>
>Requires: libboost_thread-mt.so.1.53.0()(64bit)
>
> Error: Package: ceph-common-0.80.7-0.el7.centos.x86_64 (Ceph)
>
>Requires: python-requests
>
> Error: Package: ceph-common-0.80.7-0.el7.centos.x86_64 (Ceph)
>
>Requires: librbd1 = 0.80.7-0.el7.centos
>
>Available: librbd1-0.80.6-0.el7.centos.x86_64 (Ceph)
>
>librbd1 = 0.80.6-0.el7.centos
>
>Available: librbd1-0.80.7-0.el7.centos.x86_64 (Ceph)
>
>librbd1 = 0.80.7-0.el7.centos
>
>Available: librbd1-0.80.8-0.el7.centos.x86_64 (Ceph)
>
>librbd1 = 0.80.8-0.el7.centos
>
>Installing: librbd1-0.80.9-0.el7.centos.x86_64 (Ceph)
>
>   

Re: [ceph-users] Firefly - Giant : CentOS 7 : install failed ceph-deploy

2015-04-08 Thread Michael Kidd
I don't think this came through the first time.. resending.. If it's a
dupe, my apologies..

For Firefly / Giant installs, I've had success with the following:

yum install ceph ceph-common --disablerepo=base --disablerepo=epel

Let us know if this works for you as well.

Thanks,

Michael J. Kidd
Sr. Storage Consultant
Inktank Professional Services
 - by Red Hat

On Wed, Apr 8, 2015 at 9:07 PM, Michael Kidd  wrote:

> For Firefly / Giant installs, I've had success with the following:
>
> yum install ceph ceph-common --disablerepo=base --disablerepo=epel
>
> Let us know if this works for you as well.
>
> Thanks,
>
> Michael J. Kidd
> Sr. Storage Consultant
> Inktank Professional Services
>  - by Red Hat
>
> On Wed, Apr 8, 2015 at 8:55 PM, Travis Rhoden  wrote:
>
>> I did also confirm that, as Ken mentioned, this is not a problem on
>> Hammer since Hammer includes the package split (python-ceph became
>> python-rados and python-rbd).
>>
>>  - Travis
>>
>> On Wed, Apr 8, 2015 at 5:00 PM, Travis Rhoden  wrote:
>>
>>> Hi Vickey,
>>>
>>> The easiest way I know of to get around this right now is to add the
>>> following line in section for epel in /etc/yum.repos.d/epel.repo
>>>
>>> exclude=python-rados python-rbd
>>>
>>> So this is what my epel.repo file looks like: http://fpaste.org/208681/
>>>
>>> It is those two packages in EPEL that are causing problems.  I also
>>> tried enabling epel-testing, but that didn't work either.
>>>
>>> Unfortunately you would need to add this line on each node where Ceph
>>> Giant is being installed.
>>>
>>>  - Travis
>>>
>>> On Wed, Apr 8, 2015 at 4:11 PM, Vickey Singh <
>>> vickey.singh22...@gmail.com> wrote:
>>>
>>>> Community , need help.
>>>>
>>>>
>>>> -VS-
>>>>
>>>> On Wed, Apr 8, 2015 at 4:36 PM, Vickey Singh <
>>>> vickey.singh22...@gmail.com> wrote:
>>>>
>>>>> Any suggestion  geeks
>>>>>
>>>>>
>>>>> VS
>>>>>
>>>>> On Wed, Apr 8, 2015 at 2:15 PM, Vickey Singh <
>>>>> vickey.singh22...@gmail.com> wrote:
>>>>>
>>>>>>
>>>>>> Hi
>>>>>>
>>>>>>
>>>>>> The below suggestion also didn’t worked
>>>>>>
>>>>>>
>>>>>> Full logs here : http://paste.ubuntu.com/10771939/
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> [root@rgw-node1 yum.repos.d]# yum --showduplicates list ceph
>>>>>>
>>>>>> Loaded plugins: fastestmirror, priorities
>>>>>>
>>>>>> Loading mirror speeds from cached hostfile
>>>>>>
>>>>>>  * base: mirror.zetup.net
>>>>>>
>>>>>>  * epel: ftp.fi.muni.cz
>>>>>>
>>>>>>  * extras: mirror.zetup.net
>>>>>>
>>>>>>  * updates: mirror.zetup.net
>>>>>>
>>>>>> 25 packages excluded due to repository priority protections
>>>>>>
>>>>>> Available Packages
>>>>>>
>>>>>> ceph.x86_64
>>>>>> 0.80.6-0.el7.centos
>>>>>> Ceph
>>>>>>
>>>>>> ceph.x86_64
>>>>>> 0.80.7-0.el7.centos
>>>>>> Ceph
>>>>>>
>>>>>> ceph.x86_64
>>>>>> 0.80.8-0.el7.centos
>>>>>> Ceph
>>>>>>
>>>>>> ceph.x86_64
>>>>>> 0.80.9-0.el7.centos
>>>>>> Ceph
>>>>>>
>>>>>> [root@rgw-node1 yum.repos.d]#
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> Its not able to install latest available package , yum is getting
>>>>>> confused with other DOT releases.
>>>>>>
>>>>>>
>>>>>> Any other suggestion to fix this ???
>>>>>>
>>>>>>
>>>>>>
>>>>>> --> Processing Dependency: libboost_system-mt.so.1.53.0()(64bit) for
>>>>>> package: librbd1-0.80.9-0.el7.centos.x86_64
>>>>>>
>>>>>> --> Processing Dependen

Re: [ceph-users] mon add problem

2013-12-16 Thread Michael Kidd
This indicates you have multiple networks on the new mon host, but no
definition in your ceph.conf as to which network is public.

In your ceph.conf, add:

public network = 192.168.1.0/24
cluster network = 192.168.2.0/24

(Fix the subnet definitions for your environment)

Then, re-try your new mon deploy.

Thanks,
Michael J. Kidd

Michael J. Kidd
Sr. Storage Consultant
Inktank Professional Services


On Mon, Dec 16, 2013 at 4:27 PM, Umar Draz  wrote:

> Hi,
>
> I am try to add mon host using ceph-deploy mon create kvm2, but its not
> working and giving me an error.
>
> [kvm2][DEBUG ] determining if provided host has same hostname in remote
> [kvm2][DEBUG ] get remote short hostname
> [kvm2][DEBUG ] deploying mon to kvm2
> [kvm2][DEBUG ] get remote short hostname
> [kvm2][DEBUG ] remote hostname: kvm2
> [kvm2][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
> [kvm2][DEBUG ] create the mon path if it does not exist
> [kvm2][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-kvm2/done
> [kvm2][DEBUG ] create a done file to avoid re-doing the mon deployment
> [kvm2][DEBUG ] create the init path if it does not exist
> [kvm2][DEBUG ] locating the `service` executable...
> [kvm2][INFO  ] Running command: initctl emit ceph-mon cluster=ceph id=kvm2
> [kvm2][INFO  ] Running command: ceph --cluster=ceph --admin-daemon
> /var/run/ceph/ceph-mon.kvm2.asok mon_status
> [kvm2][ERROR ] admin_socket: exception getting command descriptions:
> [Errno 2] No such file or directory
> [kvm2][WARNIN] monitor: mon.kvm2, might not be running yet
> [kvm2][INFO  ] Running command: ceph --cluster=ceph --admin-daemon
> /var/run/ceph/ceph-mon.kvm2.asok mon_status
> [kvm2][ERROR ] admin_socket: exception getting command descriptions:
> [Errno 2] No such file or directory
> [kvm2][WARNIN] kvm2 is not defined in `mon initial members`
> [kvm2][WARNIN] monitor kvm2 does not exist in monmap
> [kvm2][WARNIN] neither `public_addr` nor `public_network` keys are defined
> for monitors
> [kvm2][WARNIN] monitors may not be able to form quorum
> root@kvm1:/home/umar/ceph-cluster# ceph-deploy mon create kvm2
>
>
> would you please help me how to solve this problem?
>
> Br.
>
> Umar
>
> ___
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
___
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com