On Tue, Jun 25, 2019 at 2:40 PM Tarek Zegar <tze...@us.ibm.com> wrote:

> Sasha,
>
> Sorry I don't get it, the documentation for the command states that in
> order to see the config DB for all do: *"ceph config dump"*
> To see what's in the config DB for a particular daemon do: *"ceph config
> get <daemon>"*
> To see what's set for a particular daemon (be it from the config db,
> override, conf file, etc): *"ceph config show <daemon>"*
>
> I don't see anywhere where the command you mentioned is valid: "ceph
> config get client.admin"
>
> Here is output from a monitor node on bare metal
> root@hostmonitor1:~# ceph config dump
> WHO MASK LEVEL OPTION VALUE RO
> mon.hostmonitor1 advanced mon_osd_down_out_interval 30
> mon.hostmonitor1 advanced mon_osd_min_in_ratio 0.100000
> mgr unknown mgr/balancer/active 1 *
> mgr unknown mgr/balancer/mode upmap *
> osd.* advanced debug_ms 20/20
> osd.* advanced osd_max_backfills 2
>
> root@hostmonitor1:~# ceph config get mon.hostmonitor1
> WHO MASK LEVEL OPTION VALUE RO
> mon.hostmonitor1 advanced mon_osd_down_out_interval 30
> mon.hostmonitor1 advanced mon_osd_min_in_ratio 0.100000
>
> root@hostmonitor1:~# ceph config get client.admin
> WHO MASK LEVEL OPTION VALUE RO <-----blank
>

>
> What am I missing from what you're suggesting?
>

This just means that you don't have any MON config store overrides that
apply to the "client.admin" Ceph user:

$ ceph config get client.admin
WHO MASK LEVEL OPTION VALUE RO
$ ceph config set client debug_rbd 20
$ ceph config set client.admin debug_rbd_mirror 20
$ ceph config get client.admin
WHO          MASK LEVEL    OPTION                    VALUE RO
client            advanced debug_rbd                 20/20
client.admin      advanced debug_rbd_mirror          20/20

The first parameter to "ceph config get" is a "who" (i.e. a Ceph entity
like a daemon or client).

>
>
> Thank you for clarifying,
> Tarek Zegar
> Senior SDS Engineer
> Email *tze...@us.ibm.com* <email%20address>
> Mobile *630.974.7172*
>
>
>
>
> [image: Inactive hide details for Sasha Litvak ---06/25/2019 10:38:09
> AM---Tarek, Of course you are correct about the client nodes. I]Sasha
> Litvak ---06/25/2019 10:38:09 AM---Tarek, Of course you are correct about
> the client nodes. I executed this command
>
> From: Sasha Litvak <alexander.v.lit...@gmail.com>
> To: Tarek Zegar <tze...@us.ibm.com>, ceph-users@lists.ceph.com
> Date: 06/25/2019 10:38 AM
> Subject: [EXTERNAL] Re: Re: Re: [ceph-users] Client admin socket for RBD
> ------------------------------
>
>
>
> Tarek,
>
> Of course you are correct about the client nodes.  I executed this command
> inside of container that runs mon.  Or it can be done on the bare metal
> node that runs mon.  You essentially quering mon configuration database.
>
>
> On Tue, Jun 25, 2019 at 8:53 AM Tarek Zegar <*tze...@us.ibm.com*
> <tze...@us.ibm.com>> wrote:
>
>    "config get" on a client.admin? There is no daemon for client.admin, I
>    get nothing. Can you please explain?
>
>
>    Tarek Zegar
>    Senior SDS Engineer
> *Email **tze...@us.ibm.com* <email%20address>
>    Mobile *630.974.7172*
>
>
>
>
>    [image: Inactive hide details for Sasha Litvak ---06/24/2019 07:48:46
>    PM---ceph config get client.admin On Mon, Jun 24, 2019, 1:10 PM T]Sasha
>    Litvak ---06/24/2019 07:48:46 PM---ceph config get client.admin On Mon, Jun
>    24, 2019, 1:10 PM Tarek Zegar <*tze...@us.ibm.com* <tze...@us.ibm.com>>
>    wrote:
>
>    From: Sasha Litvak <*alexander.v.lit...@gmail.com*
>    <alexander.v.lit...@gmail.com>>
>    To: Tarek Zegar <*tze...@us.ibm.com* <tze...@us.ibm.com>>
>    Date: 06/24/2019 07:48 PM
>    Subject: [EXTERNAL] Re: Re: [ceph-users] Client admin socket for RBD
>    ------------------------------
>
>
>
>    ceph config get client.admin
>
>    On Mon, Jun 24, 2019, 1:10 PM Tarek Zegar <*tze...@us.ibm.com*
>    <tze...@us.ibm.com>> wrote:
>       Alex,
>
>          Sorry real quick, what did you type to get that last bit of info?
>
>          Tarek Zegar
>          Senior SDS Engineer
> *Email **tze...@us.ibm.com* <email%20address>
>          Mobile *630.974.7172*
>
>
>
>
>          Alex Litvak ---06/24/2019 01:07:28 PM---Jason, Here you go:
>
>          From: Alex Litvak <*alexander.v.lit...@gmail.com*
>          <alexander.v.lit...@gmail.com>>
>          To: *ceph-users@lists.ceph.com* <ceph-users@lists.ceph.com>
>          Cc: ceph-users <
>          *public-ceph-users-idqoxfivofjgjs9i8mt...@plane.gmane.org*
>          <public-ceph-users-idqoxfivofjgjs9i8mt...@plane.gmane.org>>
>          Date: 06/24/2019 01:07 PM
>          Subject: [EXTERNAL] Re: [ceph-users] Client admin socket for RBD
>          Sent by: "ceph-users" <*ceph-users-boun...@lists.ceph.com*
>          <ceph-users-boun...@lists.ceph.com>>
>          ------------------------------
>
>
>
>          Jason,
>
>          Here you go:
>
>          WHO    MASK LEVEL    OPTION                      VALUE
>                        RO
>          client      advanced admin_socket
>           /var/run/ceph/$name.$pid.asok *
>          global      advanced cluster_network             *10.0.42.0/23*
>          <http://10.0.42.0/23>                  *
>          global      advanced debug_asok                  0/0
>          global      advanced debug_auth                  0/0
>          global      advanced debug_bdev                  0/0
>          global      advanced debug_bluefs                0/0
>          global      advanced debug_bluestore             0/0
>          global      advanced debug_buffer                0/0
>          global      advanced debug_civetweb              0/0
>          global      advanced debug_client                0/0
>          global      advanced debug_compressor            0/0
>          global      advanced debug_context               0/0
>          global      advanced debug_crush                 0/0
>          global      advanced debug_crypto                0/0
>          global      advanced debug_dpdk                  0/0
>          global      advanced debug_eventtrace            0/0
>          global      advanced debug_filer                 0/0
>          global      advanced debug_filestore             0/0
>          global      advanced debug_finisher              0/0
>          global      advanced debug_fuse                  0/0
>          global      advanced debug_heartbeatmap          0/0
>          global      advanced debug_javaclient            0/0
>          global      advanced debug_journal               0/0
>          global      advanced debug_journaler             0/0
>          global      advanced debug_kinetic               0/0
>          global      advanced debug_kstore                0/0
>          global      advanced debug_leveldb               0/0
>          global      advanced debug_lockdep               0/0
>          global      advanced debug_mds                   0/0
>          global      advanced debug_mds_balancer          0/0
>          global      advanced debug_mds_locker            0/0
>          global      advanced debug_mds_log               0/0
>          global      advanced debug_mds_log_expire        0/0
>          global      advanced debug_mds_migrator          0/0
>          global      advanced debug_memdb                 0/0
>          global      advanced debug_mgr                   0/0
>          global      advanced debug_mgrc                  0/0
>          global      advanced debug_mon                   0/0
>          global      advanced debug_monc                  0/00
>          global      advanced debug_ms                    0/0
>          global      advanced debug_none                  0/0
>          global      advanced debug_objclass              0/0
>          global      advanced debug_objectcacher          0/0
>          global      advanced debug_objecter              0/0
>          global      advanced debug_optracker             0/0
>          global      advanced debug_osd                   0/0
>          global      advanced debug_paxos                 0/0
>          global      advanced debug_perfcounter           0/0
>          global      advanced debug_rados                 0/0
>          global      advanced debug_rbd                   0/0
>          global      advanced debug_rbd_mirror            0/0
>          global      advanced debug_rbd_replay            0/0
>          global      advanced debug_refs                  0/0
>          global      basic    log_file                    /dev/null
>                        *
>          global      advanced mon_cluster_log_file        /dev/null
>                        *
>          global      advanced osd_pool_default_crush_rule -1
>          global      advanced osd_scrub_begin_hour        19
>          global      advanced osd_scrub_end_hour          4
>          global      advanced osd_scrub_load_threshold    0.010000
>          global      advanced osd_scrub_sleep             0.100000
>          global      advanced perf                        true
>          global      advanced public_network              *10.0.40.0/23*
>          <http://10.0.40.0/23>                  *
>          global      advanced rocksdb_perf                true
>
>          On 6/24/2019 11:50 AM, Jason Dillaman wrote:
>          > On Sun, Jun 23, 2019 at 4:27 PM Alex Litvak
>          > <*alexander.v.lit...@gmail.com* <alexander.v.lit...@gmail.com>>
>          wrote:
>          >>
>          >> Hello everyone,
>          >>
>          >> I encounter this in nautilus client and not with mimic.
>          Removing admin socket entry from config on client makes no difference
>          >>
>          >> Error:
>          >>
>          >> rbd ls -p one
>          >> 2019-06-23 12:58:29.344 7ff2710b0700 -1 set_mon_vals failed
>          to set admin_socket = /var/run/ceph/$name.$pid.asok: Configuration 
> option
>          'admin_socket' may not be modified at runtime
>          >> 2019-06-23 12:58:29.348 7ff2708af700 -1 set_mon_vals failed
>          to set admin_socket = /var/run/ceph/$name.$pid.asok: Configuration 
> option
>          'admin_socket' may not be modified at runtime
>          >>
>          >> I have no issues running other ceph clients (no messages on
>          the screen with ceph -s or ceph iostat from the same box.)
>          >> I connected to a few other client nodes and as root I can do
>          the same string
>          >> rbd ls -p one
>          >>
>          >>
>          >> On all the nodes with user libvirt I have seen the
>          admin_socket messages
>          >>
>          >> oneadmin@virt3n1-la:~$  rbd ls -p one --id libvirt
>          >> 2019-06-23 13:16:41.626 7f9ea0ff9700 -1 set_mon_vals failed
>          to set admin_socket = /var/run/ceph/$name.$pid.asok: Configuration 
> option
>          'admin_socket' may not be modified at runtime
>          >> 2019-06-23 13:16:41.626 7f9e8bfff700 -1 set_mon_vals failed
>          to set admin_socket = /var/run/ceph/$name.$pid.asok: Configuration 
> option
>          'admin_socket' may not be modified at runtime
>          >>
>          >> I can execute all rbd operations on the cluster from client
>          otherwise.  Commenting client in config file makes no difference
>          >>
>          >> This is an optimiised config distributed across the clients
>          it is almost the same as on servers (no libvirt on servers)
>          >>
>          >> [client]
>          >> admin_socket = /var/run/ceph/$name.$pid.asok
>          >>
>          >> [client.libvirt]
>          >> admin socket =
>          /var/run/ceph/$cluster-$type.$id.$pid.$cctid.asok # must be writable 
> by
>          QEMU and allowed by SELinux or AppArmor
>          >> log file = /var/log/ceph/qemu-guest-$pid.log # must be
>          writable by QEMU and allowed by SELinux or AppArmor
>          >>
>          >> # Please do not change this file directly since it is managed
>          by Ansible and will be overwritten
>          >> [global]
>          >> cluster network = *10.0.42.0/23* <http://10.0.42.0/23>
>          >> fsid = 3947ba2d-1b01-4909-8e3a-f9714f427483
>          >> log file = /dev/null
>          >> mon cluster log file = /dev/null
>          >> mon host = [v2:*10.0.40.121:3300* <http://10.0.40.121:3300>
>          ,v1:*10.0.40.121:6789* <http://10.0.40.121:6789>],[v2:
>          *10.0.40.122:3300* <http://10.0.40.122:3300>,v1:
>          *10.0.40.122:6789* <http://10.0.40.122:6789>],[v2:
>          *10.0.40.123:3300* <http://10.0.40.123:3300>,v1:
>          *10.0.40.123:6789* <http://10.0.40.123:6789>]
>          >> perf = True
>          >> public network = *10.0.40.0/23* <http://10.0.40.0/23>
>          >> rocksdb_perf = True
>          >>
>          >>
>          >> Here is config from mon
>          >>
>          >> NAME                        VALUE
>
>                          SOURCE   OVERRIDES                      IGNORES
>          >> cluster_network             *10.0.42.0/23*
>          <http://10.0.42.0/23>
>                                                                               
>   file
>              (mon[*10.0.42.0/23* <http://10.0.42.0/23>])
>          >> daemonize                   false
>
>                          override
>          >> debug_asok                  0/0
>
>                          mon
>          >> debug_auth                  0/0
>
>                          mon
>          >> debug_bdev                  0/0
>
>                          mon
>          >> debug_bluefs                0/0
>
>                          mon
>          >> debug_bluestore             0/0
>
>                          mon
>          >> debug_buffer                0/0
>
>                          mon
>          >> debug_civetweb              0/0
>
>                          mon
>          >> debug_client                0/0
>
>                          mon
>          >> debug_compressor            0/0
>
>                          mon
>          >> debug_context               0/0
>
>                          mon
>          >> debug_crush                 0/0
>
>                          mon
>          >> debug_crypto                0/0
>
>                          mon
>          >> debug_dpdk                  0/0
>
>                          mon
>          >> debug_eventtrace            0/0
>
>                          mon
>          >> debug_filer                 0/0
>
>                          mon
>          >> debug_filestore             0/0
>
>                          mon
>          >> debug_finisher            0/0
>
>                        mon
>          >> debug_fuse                  0/0
>
>                          mon
>          >> debug_heartbeatmap          0/0
>
>                          mon
>          >> debug_javaclient            0/0
>
>                          mon
>          >> debug_journal               0/0
>
>                          mon
>          >> debug_journaler             0/0
>
>                          mon
>          >> debug_kinetic               0/0
>
>                          mon
>          >> debug_kstore                0/0
>
>                          mon
>          >> debug_leveldb               0/0
>
>                          mon
>          >> debug_lockdep               0/0
>
>                          mon
>          >> debug_mds                   0/0
>
>                          mon
>          >> debug_mds_balancer          0/0
>
>                          mon
>          >> debug_mds_locker            0/0
>
>                        mon
>          >> debug_mds_log               0/0
>
>                          mon
>          >> debug_mds_log_expire        0/0
>
>                          mon
>          >> debug_mds_migrator          0/0
>
>                          mon
>          >> debug_memdb                 0/0
>
>                          mon
>          >> debug_mgr                   0/0
>
>                          mon
>          >> debug_mgrc                  0/0
>
>                          mon
>          >> debug_mon                   0/0
>
>                          mon
>          >> debug_monc                  0/00
>
>                           mon
>          >> debug_ms                    0/0
>
>                          mon
>          >> debug_none                  0/0
>
>                          mon
>          >> debug_objclass              0/0
>
>                          mon
>          >> debug_objectcacher          0/0
>
>                          mon
>          >> debug_objecter              0/0
>
>                          mon
>          >> debug_optracker             0/0
>
>                          mon
>          >> debug_osd                   0/0
>
>                          mon
>          >> debug_paxos                 0/0
>
>                          mon
>          >> debug_perfcounter           0/0
>
>                          mon
>          >> debug_rados                 0/0
>
>                          mon
>          >> debug_rbd                   0/0
>
>                          mon
>          >> debug_rbd_mirror            0/0
>
>                          mon
>          >> debug_rbd_replay            0/0
>
>                          mon
>          >> debug_refs                  0/0
>
>                          mon
>          >> err_to_stderr               true
>
>                           override
>          >> keyring                     $mon_data/keyring
>
>                          default
>          >> leveldb_block_size          65536
>
>                          default
>          >> leveldb_cache_size          536870912
>
>                          default
>          >> leveldb_compression         false
>
>                          default
>          >> leveldb_log
>
>                          default
>          >> leveldb_write_buffer_size   33554432
>
>                           default
>          >> log_file
>
>                           override file[/dev/null],mon[/dev/null]
>          >> log_stderr_prefix           debug
>
>                          cmdline
>          >> log_to_stderr               true
>
>                           override
>          >> log_to_syslog               false
>
>                          override
>          >> mon_allow_pool_delete       true
>
>                           mon
>          >> mon_cluster_log_file        /dev/null
>
>                          file     (mon[/dev/null])
>          >> mon_cluster_log_to_stderr   true
>
>                           cmdline
>          >> mon_data
>           /var/lib/ceph/mon/ceph-storage2n2-la
>                                                             cmdline
>          >> mon_host                    [v2:*10.0.40.121:3300*
>          <http://10.0.40.121:3300>,v1:*10.0.40.121:6789*
>          <http://10.0.40.121:6789>],[v2:*10.0.40.122:3300*
>          <http://10.0.40.122:3300>,v1:*10.0.40.122:6789*
>          <http://10.0.40.122:6789>],[v2:*10.0.40.123:3300*
>          <http://10.0.40.123:3300>,v1:*10.0.40.123:6789*
>          <http://10.0.40.123:6789>] file
>          >> mon_initial_members
>          storage2n1-la,storage2n2-la,storage2n3-la
>                                                            file
>          >> mon_osd_down_out_interval   300
>
>                          mon
>          >> osd_pool_default_crush_rule -1
>
>                           file     (mon[-1])
>          >> osd_scrub_begin_hour        19
>
>                           mon
>          >> osd_scrub_end_hour          4
>
>                          mon
>          >> osd_scrub_load_threshold    0.010000
>
>                           mon
>          >> osd_scrub_sleep             0.100000
>
>                           mon
>          >> perf                        true
>
>                           file     (mon[true])
>          >> public_addr                 v2:*10.0.40.122:0/0*
>          <http://10.0.40.122:0/0>
>                                                                             
> cmdline
>          >> public_network              *10.0.40.0/23*
>          <http://10.0.40.0/23>
>                                                                               
>   file
>              (mon[*10.0.40.0/23* <http://10.0.40.0/23>])
>          >> rbd_default_features        61
>
>                           default
>          >> rocksdb_perf                true
>
>                           file     (mon[true])
>          >> setgroup                    ceph
>
>                           cmdline
>          >> setuser                     ceph
>
>                           cmdline
>          >
>          > What's the mon config for the "client.admin" user? "ceph
>          config get
>          > client.admin"
>          >
>          >>
>          >> I am not sure why I am  getting this messages and why are
>          they inconsistent across the nodes.  For example I am not getting 
> those
>          when I execute rbd in containers running ceph daemons on server
>          >> cluster nodes.   Any clue would be appreciated.
>          >>
>          >>
>          >>
>          >>
>          >>
>          >>
>          >>
>          >>
>          >>
>          >> _______________________________________________
>          >> ceph-users mailing list
>          >> *ceph-users@lists.ceph.com* <ceph-users@lists.ceph.com>
>          >> *http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com*
>          <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>          >
>          >
>          >
>
>
>          _______________________________________________
>          ceph-users mailing list
> *ceph-users@lists.ceph.com* <ceph-users@lists.ceph.com>
> *http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com*
>          <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>
>
>
> * (See attached file: 1C790723.gif)*
>
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>


-- 
Jason
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to