Tarek, Of course you are correct about the client nodes. I executed this command inside of container that runs mon. Or it can be done on the bare metal node that runs mon. You essentially quering mon configuration database.
On Tue, Jun 25, 2019 at 8:53 AM Tarek Zegar <tze...@us.ibm.com> wrote: > "config get" on a client.admin? There is no daemon for client.admin, I get > nothing. Can you please explain? > > > Tarek Zegar > Senior SDS Engineer > Email *tze...@us.ibm.com* <email%20address> > Mobile *630.974.7172* > > > > > [image: Inactive hide details for Sasha Litvak ---06/24/2019 07:48:46 > PM---ceph config get client.admin On Mon, Jun 24, 2019, 1:10 PM T]Sasha > Litvak ---06/24/2019 07:48:46 PM---ceph config get client.admin On Mon, Jun > 24, 2019, 1:10 PM Tarek Zegar <tze...@us.ibm.com> wrote: > > From: Sasha Litvak <alexander.v.lit...@gmail.com> > To: Tarek Zegar <tze...@us.ibm.com> > Date: 06/24/2019 07:48 PM > Subject: [EXTERNAL] Re: Re: [ceph-users] Client admin socket for RBD > ------------------------------ > > > > ceph config get client.admin > > On Mon, Jun 24, 2019, 1:10 PM Tarek Zegar <*tze...@us.ibm.com* > <tze...@us.ibm.com>> wrote: > > Alex, > > Sorry real quick, what did you type to get that last bit of info? > > Tarek Zegar > Senior SDS Engineer > *Email **tze...@us.ibm.com* <email%20address> > Mobile *630.974.7172* > > > > > Alex Litvak ---06/24/2019 01:07:28 PM---Jason, Here you go: > > From: Alex Litvak <*alexander.v.lit...@gmail.com* > <alexander.v.lit...@gmail.com>> > To: *ceph-users@lists.ceph.com* <ceph-users@lists.ceph.com> > Cc: ceph-users < > *public-ceph-users-idqoxfivofjgjs9i8mt...@plane.gmane.org* > <public-ceph-users-idqoxfivofjgjs9i8mt...@plane.gmane.org>> > Date: 06/24/2019 01:07 PM > Subject: [EXTERNAL] Re: [ceph-users] Client admin socket for RBD > Sent by: "ceph-users" <*ceph-users-boun...@lists.ceph.com* > <ceph-users-boun...@lists.ceph.com>> > ------------------------------ > > > > Jason, > > Here you go: > > WHO MASK LEVEL OPTION VALUE > RO > client advanced admin_socket > /var/run/ceph/$name.$pid.asok * > global advanced cluster_network *10.0.42.0/23* > <http://10.0.42.0/23> * > global advanced debug_asok 0/0 > global advanced debug_auth 0/0 > global advanced debug_bdev 0/0 > global advanced debug_bluefs 0/0 > global advanced debug_bluestore 0/0 > global advanced debug_buffer 0/0 > global advanced debug_civetweb 0/0 > global advanced debug_client 0/0 > global advanced debug_compressor 0/0 > global advanced debug_context 0/0 > global advanced debug_crush 0/0 > global advanced debug_crypto 0/0 > global advanced debug_dpdk 0/0 > global advanced debug_eventtrace 0/0 > global advanced debug_filer 0/0 > global advanced debug_filestore 0/0 > global advanced debug_finisher 0/0 > global advanced debug_fuse 0/0 > global advanced debug_heartbeatmap 0/0 > global advanced debug_javaclient 0/0 > global advanced debug_journal 0/0 > global advanced debug_journaler 0/0 > global advanced debug_kinetic 0/0 > global advanced debug_kstore 0/0 > global advanced debug_leveldb 0/0 > global advanced debug_lockdep 0/0 > global advanced debug_mds 0/0 > global advanced debug_mds_balancer 0/0 > global advanced debug_mds_locker 0/0 > global advanced debug_mds_log 0/0 > global advanced debug_mds_log_expire 0/0 > global advanced debug_mds_migrator 0/0 > global advanced debug_memdb 0/0 > global advanced debug_mgr 0/0 > global advanced debug_mgrc 0/0 > global advanced debug_mon 0/0 > global advanced debug_monc 0/00 > global advanced debug_ms 0/0 > global advanced debug_none 0/0 > global advanced debug_objclass 0/0 > global advanced debug_objectcacher 0/0 > global advanced debug_objecter 0/0 > global advanced debug_optracker 0/0 > global advanced debug_osd 0/0 > global advanced debug_paxos 0/0 > global advanced debug_perfcounter 0/0 > global advanced debug_rados 0/0 > global advanced debug_rbd 0/0 > global advanced debug_rbd_mirror 0/0 > global advanced debug_rbd_replay 0/0 > global advanced debug_refs 0/0 > global basic log_file /dev/null > * > global advanced mon_cluster_log_file /dev/null > * > global advanced osd_pool_default_crush_rule -1 > global advanced osd_scrub_begin_hour 19 > global advanced osd_scrub_end_hour 4 > global advanced osd_scrub_load_threshold 0.010000 > global advanced osd_scrub_sleep 0.100000 > global advanced perf true > global advanced public_network *10.0.40.0/23* > <http://10.0.40.0/23> * > global advanced rocksdb_perf true > > On 6/24/2019 11:50 AM, Jason Dillaman wrote: > > On Sun, Jun 23, 2019 at 4:27 PM Alex Litvak > > <*alexander.v.lit...@gmail.com* <alexander.v.lit...@gmail.com>> > wrote: > >> > >> Hello everyone, > >> > >> I encounter this in nautilus client and not with mimic. Removing > admin socket entry from config on client makes no difference > >> > >> Error: > >> > >> rbd ls -p one > >> 2019-06-23 12:58:29.344 7ff2710b0700 -1 set_mon_vals failed to set > admin_socket = /var/run/ceph/$name.$pid.asok: Configuration option > 'admin_socket' may not be modified at runtime > >> 2019-06-23 12:58:29.348 7ff2708af700 -1 set_mon_vals failed to set > admin_socket = /var/run/ceph/$name.$pid.asok: Configuration option > 'admin_socket' may not be modified at runtime > >> > >> I have no issues running other ceph clients (no messages on the > screen with ceph -s or ceph iostat from the same box.) > >> I connected to a few other client nodes and as root I can do the > same string > >> rbd ls -p one > >> > >> > >> On all the nodes with user libvirt I have seen the admin_socket > messages > >> > >> oneadmin@virt3n1-la:~$ rbd ls -p one --id libvirt > >> 2019-06-23 13:16:41.626 7f9ea0ff9700 -1 set_mon_vals failed to set > admin_socket = /var/run/ceph/$name.$pid.asok: Configuration option > 'admin_socket' may not be modified at runtime > >> 2019-06-23 13:16:41.626 7f9e8bfff700 -1 set_mon_vals failed to set > admin_socket = /var/run/ceph/$name.$pid.asok: Configuration option > 'admin_socket' may not be modified at runtime > >> > >> I can execute all rbd operations on the cluster from client > otherwise. Commenting client in config file makes no difference > >> > >> This is an optimiised config distributed across the clients it is > almost the same as on servers (no libvirt on servers) > >> > >> [client] > >> admin_socket = /var/run/ceph/$name.$pid.asok > >> > >> [client.libvirt] > >> admin socket = /var/run/ceph/$cluster-$type.$id.$pid.$cctid.asok # > must be writable by QEMU and allowed by SELinux or AppArmor > >> log file = /var/log/ceph/qemu-guest-$pid.log # must be writable by > QEMU and allowed by SELinux or AppArmor > >> > >> # Please do not change this file directly since it is managed by > Ansible and will be overwritten > >> [global] > >> cluster network = *10.0.42.0/23* <http://10.0.42.0/23> > >> fsid = 3947ba2d-1b01-4909-8e3a-f9714f427483 > >> log file = /dev/null > >> mon cluster log file = /dev/null > >> mon host = [v2:*10.0.40.121:3300* <http://10.0.40.121:3300>,v1: > *10.0.40.121:6789* <http://10.0.40.121:6789>],[v2:*10.0.40.122:3300* > <http://10.0.40.122:3300>,v1:*10.0.40.122:6789* > <http://10.0.40.122:6789>],[v2:*10.0.40.123:3300* > <http://10.0.40.123:3300>,v1:*10.0.40.123:6789* > <http://10.0.40.123:6789>] > >> perf = True > >> public network = *10.0.40.0/23* <http://10.0.40.0/23> > >> rocksdb_perf = True > >> > >> > >> Here is config from mon > >> > >> NAME VALUE > > SOURCE OVERRIDES IGNORES > >> cluster_network *10.0.42.0/23* <http://10.0.42.0/23> > > file (mon[*10.0.42.0/23* > <http://10.0.42.0/23>]) > >> daemonize false > > override > >> debug_asok 0/0 > > mon > >> debug_auth 0/0 > > mon > >> debug_bdev 0/0 > > mon > >> debug_bluefs 0/0 > > mon > >> debug_bluestore 0/0 > > mon > >> debug_buffer 0/0 > > mon > >> debug_civetweb 0/0 > > mon > >> debug_client 0/0 > > mon > >> debug_compressor 0/0 > > mon > >> debug_context 0/0 > > mon > >> debug_crush 0/0 > > mon > >> debug_crypto 0/0 > > mon > >> debug_dpdk 0/0 > > mon > >> debug_eventtrace 0/0 > > mon > >> debug_filer 0/0 > > mon > >> debug_filestore 0/0 > > mon > >> debug_finisher 0/0 > > mon > >> debug_fuse 0/0 > > mon > >> debug_heartbeatmap 0/0 > > mon > >> debug_javaclient 0/0 > > mon > >> debug_journal 0/0 > > mon > >> debug_journaler 0/0 > > mon > >> debug_kinetic 0/0 > > mon > >> debug_kstore 0/0 > > mon > >> debug_leveldb 0/0 > > mon > >> debug_lockdep 0/0 > > mon > >> debug_mds 0/0 > > mon > >> debug_mds_balancer 0/0 > > mon > >> debug_mds_locker 0/0 > > mon > >> debug_mds_log 0/0 > > mon > >> debug_mds_log_expire 0/0 > > mon > >> debug_mds_migrator 0/0 > > mon > >> debug_memdb 0/0 > > mon > >> debug_mgr 0/0 > > mon > >> debug_mgrc 0/0 > > mon > >> debug_mon 0/0 > > mon > >> debug_monc 0/00 > > mon > >> debug_ms 0/0 > > mon > >> debug_none 0/0 > > mon > >> debug_objclass 0/0 > > mon > >> debug_objectcacher 0/0 > > mon > >> debug_objecter 0/0 > > mon > >> debug_optracker 0/0 > > mon > >> debug_osd 0/0 > > mon > >> debug_paxos 0/0 > > mon > >> debug_perfcounter 0/0 > > mon > >> debug_rados 0/0 > > mon > >> debug_rbd 0/0 > > mon > >> debug_rbd_mirror 0/0 > > mon > >> debug_rbd_replay 0/0 > > mon > >> debug_refs 0/0 > > mon > >> err_to_stderr true > > override > >> keyring $mon_data/keyring > > default > >> leveldb_block_size 65536 > > default > >> leveldb_cache_size 536870912 > > default > >> leveldb_compression false > > default > >> leveldb_log > > default > >> leveldb_write_buffer_size 33554432 > > default > >> log_file > > override file[/dev/null],mon[/dev/null] > >> log_stderr_prefix debug > > cmdline > >> log_to_stderr true > > override > >> log_to_syslog false > > override > >> mon_allow_pool_delete true > > mon > >> mon_cluster_log_file /dev/null > > file (mon[/dev/null]) > >> mon_cluster_log_to_stderr true > > cmdline > >> mon_data /var/lib/ceph/mon/ceph-storage2n2-la > > cmdline > >> mon_host [v2:*10.0.40.121:3300* > <http://10.0.40.121:3300>,v1:*10.0.40.121:6789* > <http://10.0.40.121:6789>],[v2:*10.0.40.122:3300* > <http://10.0.40.122:3300>,v1:*10.0.40.122:6789* > <http://10.0.40.122:6789>],[v2:*10.0.40.123:3300* > <http://10.0.40.123:3300>,v1:*10.0.40.123:6789* > <http://10.0.40.123:6789>] file > >> mon_initial_members > storage2n1-la,storage2n2-la,storage2n3-la > file > >> mon_osd_down_out_interval 300 > > mon > >> osd_pool_default_crush_rule -1 > > file (mon[-1]) > >> osd_scrub_begin_hour 19 > > mon > >> osd_scrub_end_hour 4 > > mon > >> osd_scrub_load_threshold 0.010000 > > mon > >> osd_scrub_sleep 0.100000 > > mon > >> perf true > > file (mon[true]) > >> public_addr v2:*10.0.40.122:0/0* > <http://10.0.40.122:0/0> > cmdline > >> public_network *10.0.40.0/23* <http://10.0.40.0/23> > > file (mon[*10.0.40.0/23* > <http://10.0.40.0/23>]) > >> rbd_default_features 61 > > default > >> rocksdb_perf true > > file (mon[true]) > >> setgroup ceph > > cmdline > >> setuser ceph > > cmdline > > > > What's the mon config for the "client.admin" user? "ceph config get > > client.admin" > > > >> > >> I am not sure why I am getting this messages and why are they > inconsistent across the nodes. For example I am not getting those when I > execute rbd in containers running ceph daemons on server > >> cluster nodes. Any clue would be appreciated. > >> > >> > >> > >> > >> > >> > >> > >> > >> > >> _______________________________________________ > >> ceph-users mailing list > >> *ceph-users@lists.ceph.com* <ceph-users@lists.ceph.com> > >> *http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com* > <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com> > > > > > > > > > _______________________________________________ > ceph-users mailing list > *ceph-users@lists.ceph.com* <ceph-users@lists.ceph.com> > *http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com* > <http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com> > > > > *(See attached file: 1C790723.gif)* > > > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com