Hello everyone,
I encounter this in nautilus client and not with mimic. Removing admin socket
entry from config on client makes no difference
Error:
rbd ls -p one
2019-06-23 12:58:29.344 7ff2710b0700 -1 set_mon_vals failed to set admin_socket
= /var/run/ceph/$name.$pid.asok: Configuration option 'admin_socket' may not be
modified at runtime
2019-06-23 12:58:29.348 7ff2708af700 -1 set_mon_vals failed to set admin_socket
= /var/run/ceph/$name.$pid.asok: Configuration option 'admin_socket' may not be
modified at runtime
I have no issues running other ceph clients (no messages on the screen with
ceph -s or ceph iostat from the same box.)
I connected to a few other client nodes and as root I can do the same string
rbd ls -p one
On all the nodes with user libvirt I have seen the admin_socket messages
oneadmin@virt3n1-la:~$ rbd ls -p one --id libvirt
2019-06-23 13:16:41.626 7f9ea0ff9700 -1 set_mon_vals failed to set admin_socket
= /var/run/ceph/$name.$pid.asok: Configuration option 'admin_socket' may not be
modified at runtime
2019-06-23 13:16:41.626 7f9e8bfff700 -1 set_mon_vals failed to set admin_socket
= /var/run/ceph/$name.$pid.asok: Configuration option 'admin_socket' may not be
modified at runtime
I can execute all rbd operations on the cluster from client otherwise.
Commenting client in config file makes no difference
This is an optimiised config distributed across the clients it is almost the
same as on servers (no libvirt on servers)
[client]
admin_socket = /var/run/ceph/$name.$pid.asok
[client.libvirt]
admin socket = /var/run/ceph/$cluster-$type.$id.$pid.$cctid.asok # must be
writable by QEMU and allowed by SELinux or AppArmor
log file = /var/log/ceph/qemu-guest-$pid.log # must be writable by QEMU and
allowed by SELinux or AppArmor
# Please do not change this file directly since it is managed by Ansible and
will be overwritten
[global]
cluster network = 10.0.42.0/23
fsid = 3947ba2d-1b01-4909-8e3a-f9714f427483
log file = /dev/null
mon cluster log file = /dev/null
mon host =
[v2:10.0.40.121:3300,v1:10.0.40.121:6789],[v2:10.0.40.122:3300,v1:10.0.40.122:6789],[v2:10.0.40.123:3300,v1:10.0.40.123:6789]
perf = True
public network = 10.0.40.0/23
rocksdb_perf = True
Here is config from mon
NAME VALUE
SOURCE OVERRIDES IGNORES
cluster_network 10.0.42.0/23
file
(mon[10.0.42.0/23])
daemonize false
override
debug_asok 0/0
mon
debug_auth 0/0
mon
debug_bdev 0/0
mon
debug_bluefs 0/0
mon
debug_bluestore 0/0
mon
debug_buffer 0/0
mon
debug_civetweb 0/0
mon
debug_client 0/0
mon
debug_compressor 0/0
mon
debug_context 0/0
mon
debug_crush 0/0
mon
debug_crypto 0/0
mon
debug_dpdk 0/0
mon
debug_eventtrace 0/0
mon
debug_filer 0/0
mon
debug_filestore 0/0
mon
debug_finisher 0/0
mon
debug_fuse 0/0
mon
debug_heartbeatmap 0/0
mon
debug_javaclient 0/0
mon
debug_journal 0/0
mon
debug_journaler 0/0
mon
debug_kinetic 0/0
mon
debug_kstore 0/0
mon
debug_leveldb 0/0
mon
debug_lockdep 0/0
mon
debug_mds 0/0
mon
debug_mds_balancer 0/0
mon
debug_mds_locker 0/0
mon
debug_mds_log 0/0
mon
debug_mds_log_expire 0/0
mon
debug_mds_migrator 0/0
mon
debug_memdb 0/0
mon
debug_mgr 0/0
mon
debug_mgrc 0/0
mon
debug_mon 0/0
mon
debug_monc 0/00
mon
debug_ms 0/0
mon
debug_none 0/0
mon
debug_objclass 0/0
mon
debug_objectcacher 0/0
mon
debug_objecter 0/0
mon
debug_optracker 0/0
mon
debug_osd 0/0
mon
debug_paxos 0/0
mon
debug_perfcounter 0/0
mon
debug_rados 0/0
mon
debug_rbd 0/0
mon
debug_rbd_mirror 0/0
mon
debug_rbd_replay 0/0
mon
debug_refs 0/0
mon
err_to_stderr true
override
keyring $mon_data/keyring
default
leveldb_block_size 65536
default
leveldb_cache_size 536870912
default
leveldb_compression false
default
leveldb_log
default
leveldb_write_buffer_size 33554432
default
log_file
override file[/dev/null],mon[/dev/null]
log_stderr_prefix debug
cmdline
log_to_stderr true
override
log_to_syslog false
override
mon_allow_pool_delete true
mon
mon_cluster_log_file /dev/null
file
(mon[/dev/null])
mon_cluster_log_to_stderr true
cmdline
mon_data /var/lib/ceph/mon/ceph-storage2n2-la
cmdline
mon_host
[v2:10.0.40.121:3300,v1:10.0.40.121:6789],[v2:10.0.40.122:3300,v1:10.0.40.122:6789],[v2:10.0.40.123:3300,v1:10.0.40.123:6789]
file
mon_initial_members storage2n1-la,storage2n2-la,storage2n3-la
file
mon_osd_down_out_interval 300
mon
osd_pool_default_crush_rule -1
file
(mon[-1])
osd_scrub_begin_hour 19
mon
osd_scrub_end_hour 4
mon
osd_scrub_load_threshold 0.010000
mon
osd_scrub_sleep 0.100000
mon
perf true
file
(mon[true])
public_addr v2:10.0.40.122:0/0
cmdline
public_network 10.0.40.0/23
file
(mon[10.0.40.0/23])
rbd_default_features 61
default
rocksdb_perf true
file
(mon[true])
setgroup ceph
cmdline
setuser ceph
cmdline
I am not sure why I am getting this messages and why are they inconsistent across the nodes. For example I am not getting those when I execute rbd in containers running ceph daemons on server
cluster nodes. Any clue would be appreciated.
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com