On February 7, 2020 11:28:17 AM GMT+02:00, "Goorkate, B.J." <b.j.goork...@umcutrecht.nl> wrote: >Hi, > >There is just one unsynced entry which comes back every time: the >dom_md/ids file. > >When I add/delete a VM it's gone for a short while, but then it >reappears. > >I did a brick replace to a node which is already upgraded. After that >I'll do the same with the brick on the remaining not-yet-upgraded node, >just to be sure. Luckily we have the hardware for it. > >We're migrating from 4.2.8 (I think) to 4.3.7.2. > >Regards, > >Bertjan > > >On Thu, Feb 06, 2020 at 12:43:48PM +0200, Strahil Nikolov wrote: >> On February 6, 2020 9:42:13 AM GMT+02:00, "Goorkate, B.J." ><b.j.goork...@umcutrecht.nl> wrote: >> >Hi Strahil, >> > >> >The 0-xlator-message still occurs, but not frequent. Yesterdag a >couple >> >of times, but the 4th there were no entries at all. >> > >> >What I did find out, was that the unsynced entries belonged to >> >VM-images which were on specific hosts. >> > >> >Yesterday I migrated them all to other hosts and the unsynced >entries >> >were gone except for 3. After a 'stat' of those files/directories, >they >> >were gone too. >> > >> >I think I can migrate the remaining hosts now. An option would be to >> >move the bricks of the not-yet-upgraded hosts to upgraded hosts. I >have >> >spare disks. What do you think? >> > >> >Regards, >> > >> >Bertjan >> > >> > >> > >> > >> >On Tue, Feb 04, 2020 at 06:40:21PM +0000, Strahil Nikolov wrote: >> >> Did you manage to fix the issue with the 0-xlator ? If yes, >most >> >probably >> >> the issue will be OK. Yet 'probably' doesn't meant that they >> >'will' be OK. >> >> If it was my lab - I would go ahead only if the 0-xlator issue >is >> >> over.Yet, a lab is a different thing than prod - so it is your >> >sole >> >> decision. >> >> Did you test the upgrade prior moving to Prod ? >> >> About the hooks - I had such issue before and I had to >reinstall >> >the >> >> gluster rpms to solve it. >> >> Best Regards, >> >> Strahil Nikolov >> >> В вторник, 4 февруари 2020 г., 16:35:27 ч. Гринуич+2, Goorkate, >> >B.J. >> >> <b.j.goork...@umcutrecht.nl> написа: >> >> Hi Strahil, >> >> >> >> Thanks for your time so far! >> >> >> >> The packages seem fine on all of the 3 nodes. >> >> Only /var/lib/glusterd/glusterd.info is modified and on the not >> >yet >> >> upgraded nodes these files are missing: >> >> >> >> missing /var/lib/glusterd/hooks/1/set/post/S30samba-set.sh >> >> missing /var/lib/glusterd/hooks/1/start/post/S29CTDBsetup.sh >> >> missing >/var/lib/glusterd/hooks/1/start/post/S30samba-start.sh >> >> missing >/var/lib/glusterd/hooks/1/stop/pre/S29CTDB-teardown.sh >> >> missing /var/lib/glusterd/hooks/1/stop/pre/S30samba-stop.sh >> >> >> >> But that doesn't seem relevant... >> >> >> >> I stopped a couple of virtual machines with image files with >> >unsynched >> >> entries. When they were turned off, I couldn't find them >anymore >> >in the >> >> unsyced entries list. When I turned them on again, some >> >re-appeared, some >> >> didn't. >> >> >> >> I really don't know where to look next. >> >> >> >> The big question is: will the problems be resolved when I >upgrade >> >the two >> >> remaining nodes, or will it get worse? >> >> >> >> Regards, >> >> >> >> Bertjan >> >> >> >> On Sun, Feb 02, 2020 at 08:06:58PM +0000, Strahil Nikolov >wrote: >> >> > It seems that something went "off". >> >> > That '0-xlator: dlsym(xlator_api) missing: >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so: >undefined >> >symbol: >> >> > xlator_api' is really worrisome. >> >> > I think that you might have a bad package and based on the >> >info it >> >> could >> >> > be glusterfs-libs which should provide >> >/usr/lib64/libgfrpc.so.0 . I'm >> >> > currently on gluster v7.0 and I can't check it on my >> >installation. >> >> > Run a for loop to check the rpms: >> >> > for i in $(rpm -qa | grep gluster) ; do echo "$i :" ; rpm >-V >> >$i'; >> >> > echo;echo; done >> >> > Most probably you can safely redo (on the upgraded node) >the >> >last yum >> >> > transaction: >> >> > yum history >> >> > yum history info <ID> -> verify the gluster packages were >> >installed >> >> here >> >> > yum history redo <ID> >> >> > If kernel, glibc, systemd were not in this transaction , >you >> >can stop >> >> > gluster and start it again: >> >> > Node in maintenance in oVirt >> >> > systemctl stop ovirt-ha-agent ovirt-ha-broker vdsmd >> >supervdsmd >> >> > systemctl stop sanlock >> >> > systemctl stop glusterd >> >> > /usr/share/glusterfs/scripts/stop-all-gluster-processes.sh >> >> > And then power up again: >> >> > systemctl start glusterd >> >> > gluster volume status -> check all connected >> >> > systemctl start sanlock supervdsmd vdsmd >> >> > systemctl start ovirt-ha-broker ovirt-ha-agent >> >> > Check situation. >> >> > Yet, you need to make gluster stop complaining , before >you >> >can take >> >> care >> >> > of the heal. Usually 'rsync' is my best friend - but this >is >> >when >> >> gluster >> >> > is working normally - and your case is far from normal. >> >> > If redo doesn't work for you -> try the "yum history >> >rollback" to >> >> recover >> >> > to last good state. >> >> > I think that 'BOOM Boot Manager' is best in such cases. >> >> > Note: Never take any of my words for granted. I'm not >running >> >oVirt >> >> in >> >> > production and some of my methods might not be OK for your >> >> environment. >> >> > Best Regards, >> >> > Strahil Nikolov >> >> > В неделя, 2 февруари 2020 г., 08:56:01 ч. Гринуич+2, Mark >> >Lamers >> >> > <mark.r.lam...@gmail.com> написа: >> >> > >> >> > Hi Strahil, >> >> > >> >> > Bertjan is not in the office today, so I will reply if >okay >> >with you. >> >> > >> >> > First I like to describe the status of our network >> >> > >> >> > There are three bricks: >> >> > >> >> > and 5 nodes: >> >> > >> >> > Every host has a management and migrate vlan iface on a >> >different >> >> bond >> >> > iface. >> >> > >> >> > The last octet of the ipaddress is similar >> >> > >> >> > The output from 'gluster volume heal <volname> info' gives >a >> >long >> >> list of >> >> > shards from all three nodes, the file is attached as >> >> > 'gluster_volume_heal_info.txt'. The node without shards in >> >the list >> >> is the >> >> > updated node to gluster6, oVirt 4.3. That is curious i >think. >> >> > >> >> > Further more I find the following errors ' E ' in the >> >glusterd.log >> >> of the >> >> > upgrated host: >> >> > >> >> > [2020-01-26 17:40:46.147730] E [MSGID: 106118] >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-26 22:47:16.655651] E [MSGID: 106118] >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-27 07:07:51.815490] E [MSGID: 106118] >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-27 18:28:14.953974] E [MSGID: 106118] >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-27 18:58:22.629457] E [MSGID: 106118] >> >> > [glusterd-op-sm.c:4133:glusterd_op_ac_lock] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-27 18:58:22.629595] E [MSGID: 106376] >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management: >handler >> >> returned: -1 >> >> > [2020-01-27 18:58:22.756430] E [MSGID: 106117] >> >> > [glusterd-op-sm.c:4189:glusterd_op_ac_unlock] >0-management: >> >Unable to >> >> > release lock for vmstore1 >> >> > [2020-01-27 18:58:22.756581] E [MSGID: 106376] >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management: >handler >> >> returned: -1 >> >> > [2020-01-28 05:31:52.427196] E [MSGID: 106118] >> >> > [glusterd-op-sm.c:4133:glusterd_op_ac_lock] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-28 05:31:52.427315] E [MSGID: 106376] >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management: >handler >> >> returned: -1 >> >> > [2020-01-28 05:31:52.537799] E [MSGID: 106115] >> >> > [glusterd-mgmt.c:117:gd_mgmt_v3_collate_errors] >0-management: >> >Locking >> >> > failed on 10.13.250.14. Please check log file for details. >> >> > [2020-01-28 05:31:52.537973] E [MSGID: 106150] >> >> > [glusterd-syncop.c:1931:gd_sync_task_begin] 0-management: >> >Locking >> >> Peers >> >> > Failed. >> >> > [2020-01-28 05:31:52.539620] E [MSGID: 106117] >> >> > [glusterd-op-sm.c:4189:glusterd_op_ac_unlock] >0-management: >> >Unable to >> >> > release lock for vmstore1 >> >> > [2020-01-28 05:31:52.539759] E [MSGID: 106376] >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management: >handler >> >> returned: -1 >> >> > [2020-01-28 05:31:52.539937] E [MSGID: 106115] >> >> > [glusterd-mgmt.c:117:gd_mgmt_v3_collate_errors] >0-management: >> >> Unlocking >> >> > failed on 10.13.250.14. Please check log file for details. >> >> > [2020-01-28 05:31:52.540446] E [MSGID: 106151] >> >> > [glusterd-syncop.c:1616:gd_unlock_op_phase] 0-management: >> >Failed to >> >> unlock >> >> > on some peer(s) >> >> > [2020-01-28 08:51:45.638694] E [MSGID: 106118] >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-28 08:52:45.709950] E [MSGID: 106118] >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-28 08:54:12.455555] E [MSGID: 101097] >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator: >> >dlsym(xlator_api) >> >> missing: >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so: >undefined >> >symbol: >> >> > xlator_api >> >> > [2020-01-28 08:54:16.214779] E [MSGID: 106118] >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-28 08:57:59.496842] E [MSGID: 106118] >> >> > [glusterd-op-sm.c:4133:glusterd_op_ac_lock] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-28 08:57:59.496905] E [MSGID: 106376] >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management: >handler >> >> returned: -1 >> >> > [2020-01-28 08:57:59.505119] E >> >[rpcsvc.c:1577:rpcsvc_submit_generic] >> >> > 0-rpc-service: failed to submit message (XID: 0x2, >Program: >> >GlusterD >> >> svc >> >> > cli read-only, ProgVers: 2, Proc: 5) to rpc-transport >> >> (socket.management) >> >> > [2020-01-28 08:57:59.505135] E [MSGID: 106430] >> >> > [glusterd-utils.c:558:glusterd_submit_reply] 0-glusterd: >> >Reply >> >> submission >> >> > failed >> >> > [2020-01-28 08:57:59.647456] E [MSGID: 106117] >> >> > [glusterd-op-sm.c:4189:glusterd_op_ac_unlock] >0-management: >> >Unable to >> >> > release lock for vmstore1 >> >> > [2020-01-28 08:57:59.647508] E [MSGID: 106376] >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management: >handler >> >> returned: 1 >> >> > [2020-01-28 08:57:59.654929] E >> >[rpcsvc.c:1577:rpcsvc_submit_generic] >> >> > 0-rpc-service: failed to submit message (XID: 0x2, >Program: >> >GlusterD >> >> svc >> >> > cli, ProgVers: 2, Proc: 27) to rpc-transport >> >(socket.management) >> >> > [2020-01-28 08:57:59.654943] E [MSGID: 106430] >> >> > [glusterd-utils.c:558:glusterd_submit_reply] 0-glusterd: >> >Reply >> >> submission >> >> > failed >> >> > [2020-01-28 09:07:34.941350] E [MSGID: 106118] >> >> > [glusterd-op-sm.c:4133:glusterd_op_ac_lock] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-28 09:07:34.941391] E [MSGID: 106376] >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management: >handler >> >> returned: -1 >> >> > [2020-01-28 09:07:35.042466] E [MSGID: 106117] >> >> > [glusterd-op-sm.c:4189:glusterd_op_ac_unlock] >0-management: >> >Unable to >> >> > release lock for vmstore1 >> >> > [2020-01-28 09:07:35.042510] E [MSGID: 106376] >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management: >handler >> >> returned: -1 >> >> > [2020-01-28 09:13:32.329172] E [MSGID: 106244] >> >[glusterd.c:1785:init] >> >> > 0-management: creation of 1 listeners failed, continuing >with >> >> succeeded >> >> > transport >> >> > [2020-01-28 09:13:44.024431] E [MSGID: 101097] >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator: >> >dlsym(xlator_api) >> >> missing: >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so: >undefined >> >symbol: >> >> > xlator_api >> >> > [2020-01-28 09:45:46.347499] E [MSGID: 106118] >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-28 09:46:45.837466] E [MSGID: 106118] >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-28 09:47:45.976186] E [MSGID: 106118] >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-28 09:48:24.976568] E >> >[rpc-clnt.c:346:saved_frames_unwind] >> >> (--> >> >> > >> >/lib64/libglusterfs.so.0(_gf_log_callingfn+0x13b)[0x7f61c7c07adb] >> >> (--> >> >> > /lib64/libgfrpc.so.0(+0xd7e4)[0x7f61c79ae7e4] (--> >> >> > /lib64/libgfrpc.so.0(+0xd8fe)[0x7f61c79ae8fe] (--> >> >> > >> >> >> >>/lib64/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x97)[0x7f61c79af987] >> >> > (--> /lib64/libgfrpc.so.0(+0xf518)[0x7f61c79b0518] ))))) >> >> 0-management: >> >> > forced unwinding frame type(glusterd mgmt v3) op(--(1)) >> >called at >> >> > 2020-01-28 09:44:45.743365 (xid=0x52) >> >> > [2020-01-28 09:48:24.976640] E [MSGID: 106115] >> >> > [glusterd-mgmt.c:117:gd_mgmt_v3_collate_errors] >0-management: >> >Locking >> >> > failed on 10.13.250.11. Please check log file for details. >> >> > [2020-01-28 09:48:24.976743] E [MSGID: 106150] >> >> > [glusterd-syncop.c:1931:gd_sync_task_begin] 0-management: >> >Locking >> >> Peers >> >> > Failed. >> >> > [2020-01-28 09:48:24.977361] E >> >[rpc-clnt.c:346:saved_frames_unwind] >> >> (--> >> >> > >> >/lib64/libglusterfs.so.0(_gf_log_callingfn+0x13b)[0x7f61c7c07adb] >> >> (--> >> >> > /lib64/libgfrpc.so.0(+0xd7e4)[0x7f61c79ae7e4] (--> >> >> > /lib64/libgfrpc.so.0(+0xd8fe)[0x7f61c79ae8fe] (--> >> >> > >> >> >> >>/lib64/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x97)[0x7f61c79af987] >> >> > (--> /lib64/libgfrpc.so.0(+0xf518)[0x7f61c79b0518] ))))) >> >> 0-management: >> >> > forced unwinding frame type(glusterd mgmt) op(--(3)) >called >> >at >> >> 2020-01-28 >> >> > 09:46:03.586874 (xid=0x53) >> >> > [2020-01-28 09:48:24.977417] E [MSGID: 106152] >> >> > [glusterd-syncop.c:104:gd_collate_errors] 0-glusterd: >Staging >> >failed >> >> on >> >> > 10.13.250.11. Please check log file for details. >> >> > [2020-01-28 09:48:24.977631] E >> >[rpcsvc.c:1577:rpcsvc_submit_generic] >> >> > 0-rpc-service: failed to submit message (XID: 0x2, >Program: >> >GlusterD >> >> svc >> >> > cli, ProgVers: 2, Proc: 22) to rpc-transport >> >(socket.management) >> >> > [2020-01-28 09:48:24.977664] E [MSGID: 106430] >> >> > [glusterd-utils.c:558:glusterd_submit_reply] 0-glusterd: >> >Reply >> >> submission >> >> > failed >> >> > [2020-01-28 09:48:24.977744] E [MSGID: 106115] >> >> > [glusterd-mgmt.c:117:gd_mgmt_v3_collate_errors] >0-management: >> >> Unlocking >> >> > failed on 10.13.250.13. Please check log file for details. >> >> > [2020-01-28 09:48:24.977861] E [MSGID: 106115] >> >> > [glusterd-mgmt.c:117:gd_mgmt_v3_collate_errors] >0-management: >> >> Unlocking >> >> > failed on 10.13.250.15. Please check log file for details. >> >> > [2020-01-28 09:48:25.012037] E [MSGID: 106115] >> >> > [glusterd-mgmt.c:117:gd_mgmt_v3_collate_errors] >0-management: >> >> Unlocking >> >> > failed on 10.13.250.14. Please check log file for details. >> >> > [2020-01-28 09:48:25.012148] E [MSGID: 106151] >> >> > [glusterd-syncop.c:1616:gd_unlock_op_phase] 0-management: >> >Failed to >> >> unlock >> >> > on some peer(s) >> >> > [2020-01-28 09:48:25.012315] E [MSGID: 106117] >> >> > [glusterd-syncop.c:1640:gd_unlock_op_phase] 0-management: >> >Unable to >> >> > release lock for vmstore1 >> >> > [2020-01-28 09:48:25.012425] E >> >[rpcsvc.c:1577:rpcsvc_submit_generic] >> >> > 0-rpc-service: failed to submit message (XID: 0x2, >Program: >> >GlusterD >> >> svc >> >> > cli, ProgVers: 2, Proc: 27) to rpc-transport >> >(socket.management) >> >> > [2020-01-28 09:48:25.012452] E [MSGID: 106430] >> >> > [glusterd-utils.c:558:glusterd_submit_reply] 0-glusterd: >> >Reply >> >> submission >> >> > failed >> >> > [2020-01-28 09:50:46.105492] E [MSGID: 101097] >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator: >> >dlsym(xlator_api) >> >> missing: >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so: >undefined >> >symbol: >> >> > xlator_api >> >> > [2020-01-28 10:13:16.199982] E [MSGID: 106118] >> >> > [glusterd-op-sm.c:4133:glusterd_op_ac_lock] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-28 10:13:16.200078] E [MSGID: 106376] >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management: >handler >> >> returned: -1 >> >> > [2020-01-28 10:13:16.345918] E [MSGID: 106117] >> >> > [glusterd-op-sm.c:4189:glusterd_op_ac_unlock] >0-management: >> >Unable to >> >> > release lock for vmstore1 >> >> > [2020-01-28 10:13:16.346012] E [MSGID: 106376] >> >> > [glusterd-op-sm.c:8150:glusterd_op_sm] 0-management: >handler >> >> returned: -1 >> >> > [2020-01-28 12:42:46.212783] E [MSGID: 106118] >> >> > [glusterd-syncop.c:1896:gd_sync_task_begin] 0-management: >> >Unable to >> >> > acquire lock for vmstore1 >> >> > [2020-01-28 13:25:49.857368] E [MSGID: 101097] >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator: >> >dlsym(xlator_api) >> >> missing: >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so: >undefined >> >symbol: >> >> > xlator_api >> >> > [2020-01-29 13:26:03.031179] E [MSGID: 101097] >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator: >> >dlsym(xlator_api) >> >> missing: >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so: >undefined >> >symbol: >> >> > xlator_api >> >> > [2020-01-30 12:33:18.590063] E [MSGID: 101097] >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator: >> >dlsym(xlator_api) >> >> missing: >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so: >undefined >> >symbol: >> >> > xlator_api >> >> > [2020-01-30 13:34:57.027468] E [MSGID: 101097] >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator: >> >dlsym(xlator_api) >> >> missing: >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so: >undefined >> >symbol: >> >> > xlator_api >> >> > [2020-01-30 15:08:10.814931] E [MSGID: 101097] >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator: >> >dlsym(xlator_api) >> >> missing: >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so: >undefined >> >symbol: >> >> > xlator_api >> >> > [2020-01-31 09:40:22.725825] E [MSGID: 101097] >> >> > [xlator.c:218:xlator_volopt_dynload] 0-xlator: >> >dlsym(xlator_api) >> >> missing: >> >> > /usr/lib64/glusterfs/6.6/rpc-transport/socket.so: >undefined >> >symbol: >> >> > xlator_api >> >> > >> >> > I 'm curious for your opinion of the data sofar. >> >> > >> >> > Love to here from you (or ather members of the list) >> >> > >> >> > Regards Mark Lamers >> >> > >> >> > _______________________________________________ >> >> > Users mailing list -- users@ovirt.org >> >> > To unsubscribe send an email to users-le...@ovirt.org >> >> > Privacy Statement: >https://www.ovirt.org/site/privacy-policy/ >> >> > oVirt Code of Conduct: >> >> > >https://www.ovirt.org/community/about/community-guidelines/ >> >> > List Archives: >> >> > >> >> >> >>https://lists.ovirt.org/archives/list/users@ovirt.org/message/VUBWMDUN3GMKGOT63J2Y2A5X7VXXW3H3/ >> >> >> >> > _______________________________________________ >> >> > Users mailing list -- users@ovirt.org >> >> > To unsubscribe send an email to users-le...@ovirt.org >> >> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >> >> > oVirt Code of Conduct: >> >> https://www.ovirt.org/community/about/community-guidelines/ >> >> > List Archives: >> >> >> >>https://lists.ovirt.org/archives/list/users@ovirt.org/message/PF2ONI7EHIXWXRBCJIZCVCFHYYKCYZSI/ >> >> >> >> >> >>------------------------------------------------------------------------------ >> >> >> >> De informatie opgenomen in dit bericht kan vertrouwelijk zijn >en >> >is >> >> uitsluitend bestemd voor de geadresseerde. Indien u dit bericht >> >onterecht >> >> ontvangt, wordt u verzocht de inhoud niet te gebruiken en de >> >afzender >> >> direct >> >> te informeren door het bericht te retourneren. Het Universitair >> >Medisch >> >> Centrum Utrecht is een publiekrechtelijke rechtspersoon in de >zin >> >van de >> >> W.H.W. >> >> (Wet Hoger Onderwijs en Wetenschappelijk Onderzoek) en staat >> >geregistreerd >> >> bij >> >> de Kamer van Koophandel voor Midden-Nederland onder nr. >30244197. >> >> >> >> Denk s.v.p aan het milieu voor u deze e-mail afdrukt. >> >> >> >> >> >>------------------------------------------------------------------------------ >> >> >> >> This message may contain confidential information and is >intended >> >> exclusively >> >> for the addressee. If you receive this message unintentionally, >> >please do >> >> not >> >> use the contents but notify the sender immediately by return >> >e-mail. >> >> University >> >> Medical Center Utrecht is a legal person by public law and is >> >registered >> >> at >> >> the Chamber of Commerce for Midden-Nederland under no. >30244197. >> >> >> >> Please consider the environment before printing this e-mail. >> >> Hi Bertjan, >> >> If you healed everything - I think there is no need to migrate >bricks. >> Just a precaution - keep a test VM for testing power off and power >on after each node is upgraded. >> >> What version are you upgrading to ? >> >> Best Regards, >> Strahil Nikolov >_______________________________________________ >Users mailing list -- users@ovirt.org >To unsubscribe send an email to users-le...@ovirt.org >Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >oVirt Code o
You can check on the bricks which has the newest (usually the diff is in the timestamp inside) and copy that on the rest of the bricks. Then a heal can clear the pending heal entry. Sadly this happens for some time and I never had such issues with oVirt 4.2 on gluster v3. Best Regards, Strahil Nikolov _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/ETSAW3DN7TKIQQRMDTE4ONSQ3FTXOFEJ/