Just checked cluster logs and they are full of:cephadm exited with an error 
code: 1, stderr:Reconfig daemon osd.16 ... Traceback (most recent call last): 
File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 7931, in <module> main() File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 7919, in main r = ctx.func(ctx) File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 1717, in defaultimage return func(ctx) File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 4162, in command_deploy c = get_container(ctx, ctx.fsid, daemon_type, 
daemon_id, File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b
 697d119482", line 2451, in get_container 
volume_mounts=get_container_mounts(ctx, fsid, daemon_type, daemon_id), File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 2292, in get_container_mounts if HostFacts(ctx).selinux_enabled: File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 6451, in selinux_enabled return (self.kernel_security['type'] == 
'SELinux') and \ File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 6434, in kernel_security ret = _fetch_apparmor() File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 6415, in _fetch_apparmor item, mode = line.split(' ') ValueError: not 
enough values to unpack (expected 2, got 1) Traceback (most recent call
  last): File "/usr/share/ceph/mgr/cephadm/serve.py", line 1172, in 
_remote_connection yield (conn, connr) File 
"/usr/share/ceph/mgr/cephadm/serve.py", line 1087, in _run_cephadm code, 
'\n'.join(err))) orchestrator._interface.OrchestratorError: cephadm exited with 
an error code: 1, stderr:Reconfig daemon osd.16 ... Traceback (most recent call 
last): File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 7931, in <module> main() File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 7919, in main r = ctx.func(ctx) File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 1717, in _default_image return func(ctx) File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 
 4162, in command_deploy c = get_container(ctx, ctx.fsid, daemon_type, 
daemon_id, File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 2451, in get_container volume_mounts=get_container_mounts(ctx, fsid, 
daemon_type, daemon_id), File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 2292, in get_container_mounts if HostFacts(ctx).selinux_enabled: File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 6451, in selinux_enabled return (self.kernel_security['type'] == 
'SELinux') and \ File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bdc911a9c50d6408adfca696c2faaa65c018d660a3b697d119482",
 line 6434, in kernel_security ret = _fetch_apparmor() File 
"/var/lib/ceph/30449cba-44e4-11eb-ba64-dda10beff041/cephadm.17068a0b484bd
 c911a9c50d6408adfca696c2faaa65c018d660a3b697d119482", line 6415, in 
_fetch_apparmor item, mode = line.split(' ') ValueError: not enough values to 
unpack (expected 2, got 1)being repeated over and over again for each OSD.Again 
listing "ValueError: not enough values to unpack (expected 2, got 1)"
> On Mon May 03 2021 17:20:59 GMT+0800 (Singapore Standard Time), Ashley 
> Merrick <ash...@amerrick.co.uk> wrote:
> Hello,Wondering if anyone had any feedback on some commands I could try to 
> manually update the current OSD that is down to 16.2.1 so I can at least get 
> around this upgrade bug and back to 100%?If there is any log's or if it seems 
> a new bug and I should create a bugzilla report do let me know.Thanks
>> On Fri Apr 30 2021 21:54:30 GMT+0800 (Singapore Standard Time), Ashley 
>> Merrick <ash...@amerrick.co.uk> wrote:
>> Hello All,I was running 15.2.8 via cephadm on docker Ubuntu 20.04I just 
>> attempted to upgrade to 16.2.1 via the automated method, it successfully 
>> upgraded the mon/mgr/mds and some OSD's, however it then failed on an OSD 
>> and hasn't been able to pass even after stopping and restarting the 
>> upgrade.It reported the following ""message": "Error: UPGRADEREDEPLOYDAEMON: 
>> Upgrading daemon osd.35 on host sn-s01 failed.""If I run 'ceph health 
>> detail' I get lot's of the following error : "ValueError: not enough values 
>> to unpack (expected 2, got 1)" throughout the detail reportUpon googling, it 
>> looks like I am hitting something along the lines of 
>> https://158.69.68.89/issues/48924 & 
>> https://tracker.ceph.com/issues/49522What do I need to do to either get 
>> around this bug, or a way I can manually upgrade the remaining ceph OSD's to 
>> 16.2.1, currently my cluster is working but the last OSD it failed to 
>> upgrade is currently offline (I guess as no image attached to it now as it 
>> failed to pull it), and I
  have a cluster with OSD's from not 15.2.8 and 16.2.1Thanks
>>  
>> Sent via MXlogin
> 
>  
> Sent via MXlogin

 
Sent via MXlogin
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to