Tarek,

Of course you are correct about the client nodes.  I executed this command
inside of container that runs mon.  Or it can be done on the bare metal
node that runs mon.  You essentially quering mon configuration database.


On Tue, Jun 25, 2019 at 8:53 AM Tarek Zegar <tze...@us.ibm.com> wrote:

> "config get" on a client.admin? There is no daemon for client.admin, I get
> nothing. Can you please explain?
>
>
> Tarek Zegar
> Senior SDS Engineer
> Email *tze...@us.ibm.com* <email%20address>
> Mobile *630.974.7172*
>
>
>
>
> [image: Inactive hide details for Sasha Litvak ---06/24/2019 07:48:46
> PM---ceph config get client.admin On Mon, Jun 24, 2019, 1:10 PM T]Sasha
> Litvak ---06/24/2019 07:48:46 PM---ceph config get client.admin On Mon, Jun
> 24, 2019, 1:10 PM Tarek Zegar <tze...@us.ibm.com> wrote:
>
> From: Sasha Litvak <alexander.v.lit...@gmail.com>
> To: Tarek Zegar <tze...@us.ibm.com>
> Date: 06/24/2019 07:48 PM
> Subject: [EXTERNAL] Re: Re: [ceph-users] Client admin socket for RBD
> ------------------------------
>
>
>
> ceph config get client.admin
>
> On Mon, Jun 24, 2019, 1:10 PM Tarek Zegar <*tze...@us.ibm.com*
> <tze...@us.ibm.com>> wrote:
>
>    Alex,
>
>    Sorry real quick, what did you type to get that last bit of info?
>
>    Tarek Zegar
>    Senior SDS Engineer
> *Email **tze...@us.ibm.com* <email%20address>
>    Mobile *630.974.7172*
>
>
>
>
>    Alex Litvak ---06/24/2019 01:07:28 PM---Jason, Here you go:
>
>    From: Alex Litvak <*alexander.v.lit...@gmail.com*
>    <alexander.v.lit...@gmail.com>>
>    To: *ceph-users@lists.ceph.com* <ceph-users@lists.ceph.com>
>    Cc: ceph-users <
>    *public-ceph-users-idqoxfivofjgjs9i8mt...@plane.gmane.org*
>    <public-ceph-users-idqoxfivofjgjs9i8mt...@plane.gmane.org>>
>    Date: 06/24/2019 01:07 PM
>    Subject: [EXTERNAL] Re: [ceph-users] Client admin socket for RBD
>    Sent by: "ceph-users" <*ceph-users-boun...@lists.ceph.com*
>    <ceph-users-boun...@lists.ceph.com>>
>    ------------------------------
>
>
>
>    Jason,
>
>    Here you go:
>
>    WHO    MASK LEVEL    OPTION                      VALUE
>            RO
>    client      advanced admin_socket
>     /var/run/ceph/$name.$pid.asok *
>    global      advanced cluster_network             *10.0.42.0/23*
>    <http://10.0.42.0/23>                  *
>    global      advanced debug_asok                  0/0
>    global      advanced debug_auth                  0/0
>    global      advanced debug_bdev                  0/0
>    global      advanced debug_bluefs                0/0
>    global      advanced debug_bluestore             0/0
>    global      advanced debug_buffer                0/0
>    global      advanced debug_civetweb              0/0
>    global      advanced debug_client                0/0
>    global      advanced debug_compressor            0/0
>    global      advanced debug_context               0/0
>    global      advanced debug_crush                 0/0
>    global      advanced debug_crypto                0/0
>    global      advanced debug_dpdk                  0/0
>    global      advanced debug_eventtrace            0/0
>    global      advanced debug_filer                 0/0
>    global      advanced debug_filestore             0/0
>    global      advanced debug_finisher              0/0
>    global      advanced debug_fuse                  0/0
>    global      advanced debug_heartbeatmap          0/0
>    global      advanced debug_javaclient            0/0
>    global      advanced debug_journal               0/0
>    global      advanced debug_journaler             0/0
>    global      advanced debug_kinetic               0/0
>    global      advanced debug_kstore                0/0
>    global      advanced debug_leveldb               0/0
>    global      advanced debug_lockdep               0/0
>    global      advanced debug_mds                   0/0
>    global      advanced debug_mds_balancer          0/0
>    global      advanced debug_mds_locker            0/0
>    global      advanced debug_mds_log               0/0
>    global      advanced debug_mds_log_expire        0/0
>    global      advanced debug_mds_migrator          0/0
>    global      advanced debug_memdb                 0/0
>    global      advanced debug_mgr                   0/0
>    global      advanced debug_mgrc                  0/0
>    global      advanced debug_mon                   0/0
>    global      advanced debug_monc                  0/00
>    global      advanced debug_ms                    0/0
>    global      advanced debug_none                  0/0
>    global      advanced debug_objclass              0/0
>    global      advanced debug_objectcacher          0/0
>    global      advanced debug_objecter              0/0
>    global      advanced debug_optracker             0/0
>    global      advanced debug_osd                   0/0
>    global      advanced debug_paxos                 0/0
>    global      advanced debug_perfcounter           0/0
>    global      advanced debug_rados                 0/0
>    global      advanced debug_rbd                   0/0
>    global      advanced debug_rbd_mirror            0/0
>    global      advanced debug_rbd_replay            0/0
>    global      advanced debug_refs                  0/0
>    global      basic    log_file                    /dev/null
>            *
>    global      advanced mon_cluster_log_file        /dev/null
>            *
>    global      advanced osd_pool_default_crush_rule -1
>    global      advanced osd_scrub_begin_hour        19
>    global      advanced osd_scrub_end_hour          4
>    global      advanced osd_scrub_load_threshold    0.010000
>    global      advanced osd_scrub_sleep             0.100000
>    global      advanced perf                        true
>    global      advanced public_network              *10.0.40.0/23*
>    <http://10.0.40.0/23>                  *
>    global      advanced rocksdb_perf                true
>
>    On 6/24/2019 11:50 AM, Jason Dillaman wrote:
>    > On Sun, Jun 23, 2019 at 4:27 PM Alex Litvak
>    > <*alexander.v.lit...@gmail.com* <alexander.v.lit...@gmail.com>>
>    wrote:
>    >>
>    >> Hello everyone,
>    >>
>    >> I encounter this in nautilus client and not with mimic.  Removing
>    admin socket entry from config on client makes no difference
>    >>
>    >> Error:
>    >>
>    >> rbd ls -p one
>    >> 2019-06-23 12:58:29.344 7ff2710b0700 -1 set_mon_vals failed to set
>    admin_socket = /var/run/ceph/$name.$pid.asok: Configuration option
>    'admin_socket' may not be modified at runtime
>    >> 2019-06-23 12:58:29.348 7ff2708af700 -1 set_mon_vals failed to set
>    admin_socket = /var/run/ceph/$name.$pid.asok: Configuration option
>    'admin_socket' may not be modified at runtime
>    >>
>    >> I have no issues running other ceph clients (no messages on the
>    screen with ceph -s or ceph iostat from the same box.)
>    >> I connected to a few other client nodes and as root I can do the
>    same string
>    >> rbd ls -p one
>    >>
>    >>
>    >> On all the nodes with user libvirt I have seen the admin_socket
>    messages
>    >>
>    >> oneadmin@virt3n1-la:~$  rbd ls -p one --id libvirt
>    >> 2019-06-23 13:16:41.626 7f9ea0ff9700 -1 set_mon_vals failed to set
>    admin_socket = /var/run/ceph/$name.$pid.asok: Configuration option
>    'admin_socket' may not be modified at runtime
>    >> 2019-06-23 13:16:41.626 7f9e8bfff700 -1 set_mon_vals failed to set
>    admin_socket = /var/run/ceph/$name.$pid.asok: Configuration option
>    'admin_socket' may not be modified at runtime
>    >>
>    >> I can execute all rbd operations on the cluster from client
>    otherwise.  Commenting client in config file makes no difference
>    >>
>    >> This is an optimiised config distributed across the clients it is
>    almost the same as on servers (no libvirt on servers)
>    >>
>    >> [client]
>    >> admin_socket = /var/run/ceph/$name.$pid.asok
>    >>
>    >> [client.libvirt]
>    >> admin socket = /var/run/ceph/$cluster-$type.$id.$pid.$cctid.asok #
>    must be writable by QEMU and allowed by SELinux or AppArmor
>    >> log file = /var/log/ceph/qemu-guest-$pid.log # must be writable by
>    QEMU and allowed by SELinux or AppArmor
>    >>
>    >> # Please do not change this file directly since it is managed by
>    Ansible and will be overwritten
>    >> [global]
>    >> cluster network = *10.0.42.0/23* <http://10.0.42.0/23>
>    >> fsid = 3947ba2d-1b01-4909-8e3a-f9714f427483
>    >> log file = /dev/null
>    >> mon cluster log file = /dev/null
>    >> mon host = [v2:*10.0.40.121:3300* <http://10.0.40.121:3300>,v1:
>    *10.0.40.121:6789* <http://10.0.40.121:6789>],[v2:*10.0.40.122:3300*
>    <http://10.0.40.122:3300>,v1:*10.0.40.122:6789*
>    <http://10.0.40.122:6789>],[v2:*10.0.40.123:3300*
>    <http://10.0.40.123:3300>,v1:*10.0.40.123:6789*
>    <http://10.0.40.123:6789>]
>    >> perf = True
>    >> public network = *10.0.40.0/23* <http://10.0.40.0/23>
>    >> rocksdb_perf = True
>    >>
>    >>
>    >> Here is config from mon
>    >>
>    >> NAME                        VALUE
>
>              SOURCE   OVERRIDES                      IGNORES
>    >> cluster_network             *10.0.42.0/23* <http://10.0.42.0/23>
>
>                                       file     (mon[*10.0.42.0/23*
>    <http://10.0.42.0/23>])
>    >> daemonize                   false
>
>              override
>    >> debug_asok                  0/0
>
>              mon
>    >> debug_auth                  0/0
>
>              mon
>    >> debug_bdev                  0/0
>
>              mon
>    >> debug_bluefs                0/0
>
>              mon
>    >> debug_bluestore             0/0
>
>              mon
>    >> debug_buffer                0/0
>
>              mon
>    >> debug_civetweb              0/0
>
>              mon
>    >> debug_client                0/0
>
>              mon
>    >> debug_compressor            0/0
>
>              mon
>    >> debug_context               0/0
>
>              mon
>    >> debug_crush                 0/0
>
>              mon
>    >> debug_crypto                0/0
>
>              mon
>    >> debug_dpdk                  0/0
>
>              mon
>    >> debug_eventtrace            0/0
>
>              mon
>    >> debug_filer                 0/0
>
>              mon
>    >> debug_filestore             0/0
>
>              mon
>    >> debug_finisher              0/0
>
>              mon
>    >> debug_fuse                  0/0
>
>              mon
>    >> debug_heartbeatmap          0/0
>
>              mon
>    >> debug_javaclient            0/0
>
>              mon
>    >> debug_journal               0/0
>
>              mon
>    >> debug_journaler             0/0
>
>              mon
>    >> debug_kinetic               0/0
>
>              mon
>    >> debug_kstore                0/0
>
>              mon
>    >> debug_leveldb               0/0
>
>              mon
>    >> debug_lockdep               0/0
>
>              mon
>    >> debug_mds                   0/0
>
>              mon
>    >> debug_mds_balancer          0/0
>
>              mon
>    >> debug_mds_locker            0/0
>
>            mon
>    >> debug_mds_log               0/0
>
>              mon
>    >> debug_mds_log_expire        0/0
>
>              mon
>    >> debug_mds_migrator          0/0
>
>              mon
>    >> debug_memdb                 0/0
>
>              mon
>    >> debug_mgr                   0/0
>
>              mon
>    >> debug_mgrc                  0/0
>
>              mon
>    >> debug_mon                   0/0
>
>              mon
>    >> debug_monc                  0/00
>
>               mon
>    >> debug_ms                    0/0
>
>              mon
>    >> debug_none                  0/0
>
>              mon
>    >> debug_objclass              0/0
>
>              mon
>    >> debug_objectcacher          0/0
>
>              mon
>    >> debug_objecter              0/0
>
>              mon
>    >> debug_optracker             0/0
>
>              mon
>    >> debug_osd                   0/0
>
>              mon
>    >> debug_paxos                 0/0
>
>              mon
>    >> debug_perfcounter           0/0
>
>              mon
>    >> debug_rados                 0/0
>
>              mon
>    >> debug_rbd                   0/0
>
>              mon
>    >> debug_rbd_mirror            0/0
>
>              mon
>    >> debug_rbd_replay            0/0
>
>              mon
>    >> debug_refs                  0/0
>
>              mon
>    >> err_to_stderr               true
>
>               override
>    >> keyring                     $mon_data/keyring
>
>              default
>    >> leveldb_block_size          65536
>
>              default
>    >> leveldb_cache_size          536870912
>
>              default
>    >> leveldb_compression         false
>
>              default
>    >> leveldb_log
>
>              default
>    >> leveldb_write_buffer_size   33554432
>
>               default
>    >> log_file
>
>               override file[/dev/null],mon[/dev/null]
>    >> log_stderr_prefix           debug
>
>              cmdline
>    >> log_to_stderr               true
>
>               override
>    >> log_to_syslog               false
>
>              override
>    >> mon_allow_pool_delete       true
>
>               mon
>    >> mon_cluster_log_file        /dev/null
>
>              file     (mon[/dev/null])
>    >> mon_cluster_log_to_stderr   true
>
>               cmdline
>    >> mon_data                    /var/lib/ceph/mon/ceph-storage2n2-la
>
>               cmdline
>    >> mon_host                    [v2:*10.0.40.121:3300*
>    <http://10.0.40.121:3300>,v1:*10.0.40.121:6789*
>    <http://10.0.40.121:6789>],[v2:*10.0.40.122:3300*
>    <http://10.0.40.122:3300>,v1:*10.0.40.122:6789*
>    <http://10.0.40.122:6789>],[v2:*10.0.40.123:3300*
>    <http://10.0.40.123:3300>,v1:*10.0.40.123:6789*
>    <http://10.0.40.123:6789>] file
>    >> mon_initial_members
>    storage2n1-la,storage2n2-la,storage2n3-la
>                                                      file
>    >> mon_osd_down_out_interval   300
>
>              mon
>    >> osd_pool_default_crush_rule -1
>
>               file     (mon[-1])
>    >> osd_scrub_begin_hour        19
>
>               mon
>    >> osd_scrub_end_hour          4
>
>              mon
>    >> osd_scrub_load_threshold    0.010000
>
>               mon
>    >> osd_scrub_sleep             0.100000
>
>               mon
>    >> perf                        true
>
>               file     (mon[true])
>    >> public_addr                 v2:*10.0.40.122:0/0*
>    <http://10.0.40.122:0/0>
>                                                                 cmdline
>    >> public_network              *10.0.40.0/23* <http://10.0.40.0/23>
>
>                                       file     (mon[*10.0.40.0/23*
>    <http://10.0.40.0/23>])
>    >> rbd_default_features        61
>
>               default
>    >> rocksdb_perf                true
>
>               file     (mon[true])
>    >> setgroup                    ceph
>
>               cmdline
>    >> setuser                     ceph
>
>               cmdline
>    >
>    > What's the mon config for the "client.admin" user? "ceph config get
>    > client.admin"
>    >
>    >>
>    >> I am not sure why I am  getting this messages and why are they
>    inconsistent across the nodes.  For example I am not getting those when I
>    execute rbd in containers running ceph daemons on server
>    >> cluster nodes.   Any clue would be appreciated.
>    >>
>    >>
>    >>
>    >>
>    >>
>    >>
>    >>
>    >>
>    >>
>    >> _______________________________________________
>    >> ceph-users mailing list
>    >> *ceph-users@lists.ceph.com* <ceph-users@lists.ceph.com>
>    >> *http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com*
>    <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>    >
>    >
>    >
>
>
>    _______________________________________________
>    ceph-users mailing list
> *ceph-users@lists.ceph.com* <ceph-users@lists.ceph.com>
> *http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com*
>    <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
>
>
>
>    *(See attached file: 1C790723.gif)*
>
>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to