Jason,
Here you go:
WHO MASK LEVEL OPTION VALUE
RO
client advanced admin_socket /var/run/ceph/$name.$pid.asok *
global advanced cluster_network 10.0.42.0/23 *
global advanced debug_asok 0/0
global advanced debug_auth 0/0
global advanced debug_bdev 0/0
global advanced debug_bluefs 0/0
global advanced debug_bluestore 0/0
global advanced debug_buffer 0/0
global advanced debug_civetweb 0/0
global advanced debug_client 0/0
global advanced debug_compressor 0/0
global advanced debug_context 0/0
global advanced debug_crush 0/0
global advanced debug_crypto 0/0
global advanced debug_dpdk 0/0
global advanced debug_eventtrace 0/0
global advanced debug_filer 0/0
global advanced debug_filestore 0/0
global advanced debug_finisher 0/0
global advanced debug_fuse 0/0
global advanced debug_heartbeatmap 0/0
global advanced debug_javaclient 0/0
global advanced debug_journal 0/0
global advanced debug_journaler 0/0
global advanced debug_kinetic 0/0
global advanced debug_kstore 0/0
global advanced debug_leveldb 0/0
global advanced debug_lockdep 0/0
global advanced debug_mds 0/0
global advanced debug_mds_balancer 0/0
global advanced debug_mds_locker 0/0
global advanced debug_mds_log 0/0
global advanced debug_mds_log_expire 0/0
global advanced debug_mds_migrator 0/0
global advanced debug_memdb 0/0
global advanced debug_mgr 0/0
global advanced debug_mgrc 0/0
global advanced debug_mon 0/0
global advanced debug_monc 0/00
global advanced debug_ms 0/0
global advanced debug_none 0/0
global advanced debug_objclass 0/0
global advanced debug_objectcacher 0/0
global advanced debug_objecter 0/0
global advanced debug_optracker 0/0
global advanced debug_osd 0/0
global advanced debug_paxos 0/0
global advanced debug_perfcounter 0/0
global advanced debug_rados 0/0
global advanced debug_rbd 0/0
global advanced debug_rbd_mirror 0/0
global advanced debug_rbd_replay 0/0
global advanced debug_refs 0/0
global basic log_file /dev/null *
global advanced mon_cluster_log_file /dev/null *
global advanced osd_pool_default_crush_rule -1
global advanced osd_scrub_begin_hour 19
global advanced osd_scrub_end_hour 4
global advanced osd_scrub_load_threshold 0.010000
global advanced osd_scrub_sleep 0.100000
global advanced perf true
global advanced public_network 10.0.40.0/23 *
global advanced rocksdb_perf true
On 6/24/2019 11:50 AM, Jason Dillaman wrote:
On Sun, Jun 23, 2019 at 4:27 PM Alex Litvak
<alexander.v.lit...@gmail.com> wrote:
Hello everyone,
I encounter this in nautilus client and not with mimic. Removing admin socket
entry from config on client makes no difference
Error:
rbd ls -p one
2019-06-23 12:58:29.344 7ff2710b0700 -1 set_mon_vals failed to set admin_socket
= /var/run/ceph/$name.$pid.asok: Configuration option 'admin_socket' may not be
modified at runtime
2019-06-23 12:58:29.348 7ff2708af700 -1 set_mon_vals failed to set admin_socket
= /var/run/ceph/$name.$pid.asok: Configuration option 'admin_socket' may not be
modified at runtime
I have no issues running other ceph clients (no messages on the screen with
ceph -s or ceph iostat from the same box.)
I connected to a few other client nodes and as root I can do the same string
rbd ls -p one
On all the nodes with user libvirt I have seen the admin_socket messages
oneadmin@virt3n1-la:~$ rbd ls -p one --id libvirt
2019-06-23 13:16:41.626 7f9ea0ff9700 -1 set_mon_vals failed to set admin_socket
= /var/run/ceph/$name.$pid.asok: Configuration option 'admin_socket' may not be
modified at runtime
2019-06-23 13:16:41.626 7f9e8bfff700 -1 set_mon_vals failed to set admin_socket
= /var/run/ceph/$name.$pid.asok: Configuration option 'admin_socket' may not be
modified at runtime
I can execute all rbd operations on the cluster from client otherwise.
Commenting client in config file makes no difference
This is an optimiised config distributed across the clients it is almost the
same as on servers (no libvirt on servers)
[client]
admin_socket = /var/run/ceph/$name.$pid.asok
[client.libvirt]
admin socket = /var/run/ceph/$cluster-$type.$id.$pid.$cctid.asok # must be
writable by QEMU and allowed by SELinux or AppArmor
log file = /var/log/ceph/qemu-guest-$pid.log # must be writable by QEMU and
allowed by SELinux or AppArmor
# Please do not change this file directly since it is managed by Ansible and
will be overwritten
[global]
cluster network = 10.0.42.0/23
fsid = 3947ba2d-1b01-4909-8e3a-f9714f427483
log file = /dev/null
mon cluster log file = /dev/null
mon host =
[v2:10.0.40.121:3300,v1:10.0.40.121:6789],[v2:10.0.40.122:3300,v1:10.0.40.122:6789],[v2:10.0.40.123:3300,v1:10.0.40.123:6789]
perf = True
public network = 10.0.40.0/23
rocksdb_perf = True
Here is config from mon
NAME VALUE
SOURCE OVERRIDES IGNORES
cluster_network 10.0.42.0/23
file
(mon[10.0.42.0/23])
daemonize false
override
debug_asok 0/0
mon
debug_auth 0/0
mon
debug_bdev 0/0
mon
debug_bluefs 0/0
mon
debug_bluestore 0/0
mon
debug_buffer 0/0
mon
debug_civetweb 0/0
mon
debug_client 0/0
mon
debug_compressor 0/0
mon
debug_context 0/0
mon
debug_crush 0/0
mon
debug_crypto 0/0
mon
debug_dpdk 0/0
mon
debug_eventtrace 0/0
mon
debug_filer 0/0
mon
debug_filestore 0/0
mon
debug_finisher 0/0
mon
debug_fuse 0/0
mon
debug_heartbeatmap 0/0
mon
debug_javaclient 0/0
mon
debug_journal 0/0
mon
debug_journaler 0/0
mon
debug_kinetic 0/0
mon
debug_kstore 0/0
mon
debug_leveldb 0/0
mon
debug_lockdep 0/0
mon
debug_mds 0/0
mon
debug_mds_balancer 0/0
mon
debug_mds_locker 0/0
mon
debug_mds_log 0/0
mon
debug_mds_log_expire 0/0
mon
debug_mds_migrator 0/0
mon
debug_memdb 0/0
mon
debug_mgr 0/0
mon
debug_mgrc 0/0
mon
debug_mon 0/0
mon
debug_monc 0/00
mon
debug_ms 0/0
mon
debug_none 0/0
mon
debug_objclass 0/0
mon
debug_objectcacher 0/0
mon
debug_objecter 0/0
mon
debug_optracker 0/0
mon
debug_osd 0/0
mon
debug_paxos 0/0
mon
debug_perfcounter 0/0
mon
debug_rados 0/0
mon
debug_rbd 0/0
mon
debug_rbd_mirror 0/0
mon
debug_rbd_replay 0/0
mon
debug_refs 0/0
mon
err_to_stderr true
override
keyring $mon_data/keyring
default
leveldb_block_size 65536
default
leveldb_cache_size 536870912
default
leveldb_compression false
default
leveldb_log
default
leveldb_write_buffer_size 33554432
default
log_file
override file[/dev/null],mon[/dev/null]
log_stderr_prefix debug
cmdline
log_to_stderr true
override
log_to_syslog false
override
mon_allow_pool_delete true
mon
mon_cluster_log_file /dev/null
file
(mon[/dev/null])
mon_cluster_log_to_stderr true
cmdline
mon_data /var/lib/ceph/mon/ceph-storage2n2-la
cmdline
mon_host
[v2:10.0.40.121:3300,v1:10.0.40.121:6789],[v2:10.0.40.122:3300,v1:10.0.40.122:6789],[v2:10.0.40.123:3300,v1:10.0.40.123:6789]
file
mon_initial_members storage2n1-la,storage2n2-la,storage2n3-la
file
mon_osd_down_out_interval 300
mon
osd_pool_default_crush_rule -1
file
(mon[-1])
osd_scrub_begin_hour 19
mon
osd_scrub_end_hour 4
mon
osd_scrub_load_threshold 0.010000
mon
osd_scrub_sleep 0.100000
mon
perf true
file
(mon[true])
public_addr v2:10.0.40.122:0/0
cmdline
public_network 10.0.40.0/23
file
(mon[10.0.40.0/23])
rbd_default_features 61
default
rocksdb_perf true
file
(mon[true])
setgroup ceph
cmdline
setuser ceph
cmdline
What's the mon config for the "client.admin" user? "ceph config get
client.admin"
I am not sure why I am getting this messages and why are they inconsistent
across the nodes. For example I am not getting those when I execute rbd in
containers running ceph daemons on server
cluster nodes. Any clue would be appreciated.
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com