Hi Erich, im not sure about this specific error message, but "ceph fs status" sometimes did fail me end of last year/in the beginning of the year.
Restarting ALL mon, mgr AND mds fixed it at the time. Best regards, Gunnar ======================================================= Gunnar Bandelow (dipl. phys.) Universitätsrechenzentrum (URZ) Universität Greifswald Felix-Hausdorff-Straße 18 17489 GreifswaldGermany --- Original Nachricht --- Betreff: [ceph-users] Re: 'ceph fs status' no longer works? Von: "Erich Weiler" An: "Eugen Block" , ceph-users@ceph.io Datum: 02-05-2024 21:05 Hi Eugen, Thanks for the tip! I just ran: ceph orch daemon restart mgr.pr-md-01.jemmdf (my specific mgr instance) And it restarted my primary mgr daemon, and in the process failed over to my standby mgr daemon on another server. That went smoothly. Unfortunately, I still cannot get 'ceph fs status' to work (on any node)... # ceph fs status Error EINVAL: Traceback (most recent call last): File "/usr/share/ceph/mgr/mgr_module.py", line 1811, in _handle_command return CLICommand.COMMANDS[cmd['prefix']].call(self, cmd, inbuf) File "/usr/share/ceph/mgr/mgr_module.py", line 474, in call return self.func(mgr, **kwargs) File "/usr/share/ceph/mgr/status/module.py", line 109, in handle_fs_status assert metadata AssertionError -erich On 5/2/24 11:07 AM, Eugen Block wrote: > Yep, seen this a couple of times during upgrades. I’ll have to check my > notes if I wrote anything down for that. But try a mgr failover first, > that could help. > > Zitat von Erich Weiler : > >> Hi All, >> >> For a while now I've been using 'ceph fs status' to show current MDS >> active servers, filesystem status, etc. I recently took down my MDS >> servers and added RAM to them (one by one, so the filesystem stayed >> online). After doing that with my four MDS servers (I had two active >> and two standby), all looks OK, 'ceph -s' reports HEALTH_OK. But when >> I do 'ceph fs status' now, I get this: >> >> # ceph fs status >> Error EINVAL: Traceback (most recent call last): >> File "/usr/share/ceph/mgr/mgr_module.py", line 1811, in _handle_command >> return CLICommand.COMMANDS[cmd['prefix']].call(self, cmd, inbuf) >> File "/usr/share/ceph/mgr/mgr_module.py", line 474, in call >> return self.func(mgr, **kwargs) >> File "/usr/share/ceph/mgr/status/module.py", line 109, in >> handle_fs_status >> assert metadata >> AssertionError >> >> This is on ceph 18.2.1 reef. This is very odd - can anyone think of a >> reason why 'ceph fs status' would stop working after taking each of >> the servers down for maintenance? >> >> The filesystem is online and working just fine however. This ceph >> instance is deployed via the cephadm method on RHEL 9.3, so the >> everything is containerized in podman. >> >> Thanks again, >> erich >> _______________________________________________ >> ceph-users mailing list -- ceph-users@ceph.io >> To unsubscribe send an email to ceph-users-le...@ceph.io > > > _______________________________________________ > ceph-users mailing list -- ceph-users@ceph.io > To unsubscribe send an email to ceph-users-le...@ceph.io _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io
smime.p7s
Description: S/MIME cryptographic signature
_______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io