On Thu, Nov 21, 2019 at 6:03 AM Strahil Nikolov <hunter86...@yahoo.com>
wrote:

> Hi All,
>
> another clue in the logs :
> [2019-11-21 00:29:50.536631] W [MSGID: 114031]
> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-1:
> remote operation failed. Path:
> /.shard/b0af2b81-22cf-482e-9b2f-c431b6449dae.79
> (00000000-0000-0000-0000-000000000000) [Permission denied]
> [2019-11-21 00:29:50.536798] W [MSGID: 114031]
> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-0:
> remote operation failed. Path:
> /.shard/b0af2b81-22cf-482e-9b2f-c431b6449dae.79
> (00000000-0000-0000-0000-000000000000) [Permission denied]
> [2019-11-21 00:29:50.536959] W [MSGID: 114031]
> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-2:
> remote operation failed. Path:
> /.shard/b0af2b81-22cf-482e-9b2f-c431b6449dae.79
> (00000000-0000-0000-0000-000000000000) [Permission denied]
> [2019-11-21 00:29:50.537007] E [MSGID: 133010]
> [shard.c:2327:shard_common_lookup_shards_cbk] 0-data_fast-shard: Lookup on
> shard 79 failed. Base file gfid = b0af2b81-22cf-482e-9b2f-c431b6449dae
> [Permission denied]
> [2019-11-21 00:29:50.537066] W [fuse-bridge.c:2830:fuse_readv_cbk]
> 0-glusterfs-fuse: 12458: READ => -1
> gfid=b0af2b81-22cf-482e-9b2f-c431b6449dae fd=0x7fc63c00fe18 (Permission
> denied)
> [2019-11-21 00:30:01.177665] I [MSGID: 133022]
> [shard.c:3674:shard_delete_shards] 0-data_fast-shard: Deleted shards of
> gfid=eb103fbf-80dc-425d-882f-1e4efe510db5 from backend
> [2019-11-21 00:30:13.132756] W [MSGID: 114031]
> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-0:
> remote operation failed. Path:
> /.shard/17c663c2-f582-455b-b806-3b9d01fb2c6c.79
> (00000000-0000-0000-0000-000000000000) [Permission denied]
> [2019-11-21 00:30:13.132824] W [MSGID: 114031]
> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-1:
> remote operation failed. Path:
> /.shard/17c663c2-f582-455b-b806-3b9d01fb2c6c.79
> (00000000-0000-0000-0000-000000000000) [Permission denied]
> [2019-11-21 00:30:13.133217] W [MSGID: 114031]
> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-2:
> remote operation failed. Path:
> /.shard/17c663c2-f582-455b-b806-3b9d01fb2c6c.79
> (00000000-0000-0000-0000-000000000000) [Permission denied]
> [2019-11-21 00:30:13.133238] E [MSGID: 133010]
> [shard.c:2327:shard_common_lookup_shards_cbk] 0-data_fast-shard: Lookup on
> shard 79 failed. Base file gfid = 17c663c2-f582-455b-b806-3b9d01fb2c6c
> [Permission denied]
> [2019-11-21 00:30:13.133264] W [fuse-bridge.c:2830:fuse_readv_cbk]
> 0-glusterfs-fuse: 12660: READ => -1
> gfid=17c663c2-f582-455b-b806-3b9d01fb2c6c fd=0x7fc63c007038 (Permission
> denied)
> [2019-11-21 00:30:38.489449] W [MSGID: 114031]
> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-0:
> remote operation failed. Path:
> /.shard/a10a5ae8-108b-4d78-9e65-cca188c27fc4.6
> (00000000-0000-0000-0000-000000000000) [Permission denied]
> [2019-11-21 00:30:38.489520] W [MSGID: 114031]
> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-1:
> remote operation failed. Path:
> /.shard/a10a5ae8-108b-4d78-9e65-cca188c27fc4.6
> (00000000-0000-0000-0000-000000000000) [Permission denied]
> [2019-11-21 00:30:38.489669] W [MSGID: 114031]
> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-2:
> remote operation failed. Path:
> /.shard/a10a5ae8-108b-4d78-9e65-cca188c27fc4.6
> (00000000-0000-0000-0000-000000000000) [Permission denied]
> [2019-11-21 00:30:38.489717] E [MSGID: 133010]
> [shard.c:2327:shard_common_lookup_shards_cbk] 0-data_fast-shard: Lookup on
> shard 6 failed. Base file gfid = a10a5ae8-108b-4d78-9e65-cca188c27fc4
> [Permission denied]
> [2019-11-21 00:30:38.489777] W [fuse-bridge.c:2830:fuse_readv_cbk]
> 0-glusterfs-fuse: 12928: READ => -1
> gfid=a10a5ae8-108b-4d78-9e65-cca188c27fc4 fd=0x7fc63c01a058 (Permission
> denied)
>
>
> Anyone got an idea why is it happening?
> I checked user/group and selinux permissions - all OK
> I even replaced one of the disks and healed , but the result is the same
> for all my VMs.
>

Have you checked the permission for user/group are set correctly across all
the bricks in the cluster?
What does ls -la on the images directory from mount of the volume show you.

Adding Krutika and Rafi as they ran into a similar issue in the past.


> Best Regards,
> Strahil Nikolov
>
>
> В сряда, 20 ноември 2019 г., 18:17:18 ч. Гринуич+2, Strahil Nikolov <
> hunter86...@yahoo.com> написа:
>
>
> Hello All,
>
> my engine is back online , but I'm still having difficulties to make vdsm
> powerup the systems.
> I think that the events generated today can lead me to the right
> direction(just an example , many more are there):
>
> VDSM ovirt3.localdomain command SpmStatusVDS failed: Cannot inquire
> Lease(name='SDM',
> path=u'/rhev/data-center/mnt/glusterSD/gluster1:_data__fast3/ecc3bf0e-8214-45c1-98a6-0afa642e591f/dom_md/leases',
> offset=1048576): (2, 'Sanlock get hosts failure', 'No such file or
> directory')
>
> I will try to collect a fresh log and see what is it complaining about
> this time.
>
> Best Regards,
> Strahil Nikolov
>
> >Hi Sahina,
>
> >I have a strange situation:
> >1. When I try to access the file via 'sudo -u vdsm dd if=disk of=test
> bs=4M' the command fails on aprox 60MB.
> >2. If I run same command as root , remove the file and then run again via
> vdsm user -> this time no i/o error reported.
>
> >My guess is that I need to check what's going on the bricks themselve ...
>
> >Best Regards,
> >Strahil Nikolov
>
>
> В вторник, 19 ноември 2019 г., 0:02:16 ч. Гринуич-5, Sahina Bose <
> sab...@redhat.com> написа:
>
>
>
>
> On Tue, Nov 19, 2019 at 10:10 AM Strahil Nikolov <hunter86...@yahoo.com>
> wrote:
>
> Hi Sahina,
>
> Sadly engine logs have no errors.
> I've got only an I/O error, but in the debug of the vdsm I can clearly see
> that "qemu-img" is giving an "OK".
> During the upgrade I got some metadata files pending heal, but I have
> recovered the conflict manually and should be OK.
> Today I have defined one of the VMs manually (virsh define) and then
> started it , but the issue is the same.
> It seems to be storage-related issue,as VMs that are on specific domain
> can be started , but most of my VMs are on the fast storage domains and
> none of them can be started.
>
> After the gluster snapshot restore , the engine is having issues and I
> have to separately investigate that (as I poweroff my HostedEngine before
> creating the snapshot).
>
> The logs can be find at :
> https://drive.google.com/open?id=1VAZFZWWrpimDeVuZT0sWFVXy76scr4NM
>
>
> Any ideas where to look at , as I can definitely read (using "dd if=disk"
> or qemu-img info) the disks of the rhel7 VM ?
>
>
> The vdsm logs have this:
> 2019-11-17 10:21:23,892+0200 INFO  (libvirt/events) [virt.vm]
> (vmId='b3c4d84a-9784-470c-b70e-7ad7cc45e913') abnormal vm stop device
> ua-94f763e9-fd96-4bee-a6b2-31af841a918b error eother (vm:5075)
> 2019-11-17 10:21:23,892+0200 INFO  (libvirt/events) [virt.vm]
> (vmId='b3c4d84a-9784-470c-b70e-7ad7cc45e913') CPU stopped: onIOError
> (vm:6062)
> 2019-11-17 10:21:23,893+0200 DEBUG (libvirt/events) [jsonrpc.Notification]
> Sending event {"params": {"notify_time": 4356025830,
> "b3c4d84a-9784-470c-b70e-7ad7cc45e913": {"status": "WaitForLaunch",
> "ioerror": {"alias": "ua-94f763e9-fd96-4bee-a6b2-31af841a918b", "name":
> "sda", "path":
> "/rhev/data-center/mnt/glusterSD/gluster1:_data__fast/396604d9-2a9e-49cd-9563-fdc79981f67b/images/94f763e9-fd96-4bee-a6b2-31af841a918b/5b1d3113-5cca-4582-9029-634b16338a2f"},
> "pauseCode": "EOTHER"}}, "jsonrpc": "2.0", "method":
> "|virt|VM_status|b3c4d84a-9784-470c-b70e-7ad7cc45e913"} (__init__:181)
>
> Can you check the permissions of the file
> /rhev/data-center/mnt/glusterSD/gluster1:_data__fast/396604d9-2a9e-49cd-9563-fdc79981f67b/images/94f763e9-fd96-4bee-a6b2-31af841a918b/5b1d3113-5cca-4582-9029-634b16338a2f.
> Was it reset after upgrade?
>
> Are you able to copy this file to a different location and try running a
> VM with this image?
>
> Any errors in the mount log of gluster1:_data__fast volume?
>
>
> Best Regards,
> Strahil Nikolov
>
>
>
>
> В понеделник, 18 ноември 2019 г., 11:38:13 ч. Гринуич+2, Sahina Bose <
> sab...@redhat.com> написа:
>
>
>
>
> On Mon, Nov 18, 2019 at 2:58 PM Sandro Bonazzola <sbona...@redhat.com>
> wrote:
>
> +Sahina Bose <sab...@redhat.com> +Gobinda Das <go...@redhat.com> +Nir
> Soffer <nsof...@redhat.com> +Tal Nisan <tni...@redhat.com> can you please
> help here?
>
>
> Il giorno dom 17 nov 2019 alle ore 16:00 Strahil Nikolov <
> hunter86...@yahoo.com> ha scritto:
>
> So far,
>
> I have rolled back the engine and the 3 hosts - still cannot manipulate
> the storage.
> It seems that gluster itself is working, but vdsm and the oVirt stack
> cannot access the storage - cannot create new VM disks, cannot start a VM
> and I'm on the verge of redeploy.
>
>
>
> Any errors in vdsm logs? engine logs?
>
>
>
> Best Regards,
> Strahil Nikolov
>
> В събота, 16 ноември 2019 г., 15:40:25 ч. Гринуич+2, Strahil <
> hunter86...@yahoo.com> написа:
>
>
> I got upgraded  to  RC3 and now  cannot power any VM .
> Constantly getting I/O error, but checking at gluster level - I can dd
> from each disk or even create a new one.
>
> Removing the HighAvailability doesn't help.
>
> I guess I should restore  the engine from the gluster snapshot and
> rollback via 'yum history undo last'.
>
> Does anyone else have my issues  ?
>
> Best Regards,
> Strahil Nikolov
> On Nov 13, 2019 15:31, Sandro Bonazzola <sbona...@redhat.com> wrote:
>
>
>
> Il giorno mer 13 nov 2019 alle ore 14:25 Sandro Bonazzola <
> sbona...@redhat.com> ha scritto:
>
>
>
> Il giorno mer 13 nov 2019 alle ore 13:56 Florian Schmid <
> fsch...@ubimet.com> ha scritto:
>
> Hello,
>
> I have a question about bugs, which are flagged as [downstream clone -
> 4.3.7], but are not yet released.
>
> I'm talking about this bug:
> https://bugzilla.redhat.com/show_bug.cgi?id=1749202
>
> I can't see it in 4.3.7 release notes. Will it be included in a further
> release candidate? This fix is very important I think and I can't upgrade
> yet because of this bug.
>
>
>
> Looking at the bug, the fix was done with $ git tag --contains
> 12bd5cb1fe7c95e29b4065fca968913722fe9eaa
> ovirt-engine-4.3.6.6
> ovirt-engine-4.3.6.7
> ovirt-engine-4.3.7.0
> ovirt-engine-4.3.7.1
>
> So the fix is already included in release oVirt 4.3.6.
>
>
> Sent a fix to 4.3.6 release notes:
> https://github.com/oVirt/ovirt-site/pull/2143. @Ryan Barry
> <rba...@redhat.com> can you please review?
>
>
>
>
>
>
>
>
>
>
> BR Florian Schmid
>
> ------------------------------
> *Von: *"Sandro Bonazzola" <sbona...@redhat.com>
> *An: *"users" <users@ovirt.org>
> *Gesendet: *Mittwoch, 13. November 2019 13:34:59
> *Betreff: *[ovirt-users] [ANN] oVirt 4.3.7 Third Release Candidate is now
> available for testing
>
> The oVirt Project is pleased to announce the availability of the oVirt
> 4.3.7 Third Release Candidate for testing, as of November 13th, 2019.
>
> This update is a release candidate of the seventh in a series of
> stabilization updates to the 4.3 series.
> This is pre-release software. This pre-release should not to be used in
> production.
>
> This release is available now on x86_64 architecture for:
> * Red Hat Enterprise Linux 7.7 or later (but <8)
> * CentOS Linux (or similar) 7.7 or later (but <8)
>
> This release supports Hypervisor Hosts on x86_64 and ppc64le architectures
> for:
> * Red Hat Enterprise Linux 7.7 or later (but <8)
> * CentOS Linux (or similar) 7.7 or later (but <8)
> * oVirt Node 4.3 (available for x86_64 only) has been built consuming
> CentOS 7.7 Release
>
> See the release notes [1] for known issues, new features and bugs fixed.
>
> While testing this release candidate please note that oVirt node now
> includes:
> - ansible 2.9.0
> - GlusterFS 6.6
>
> Notes:
> - oVirt Appliance is already available
> - oVirt Node is already available
>
> Additional Resources:
> * Read more about the oVirt 4.3.7 release highlights:
> http://www.ovirt.org/release/4.3.7/
> * Get more oVirt Project updates on Twitter: https://twitter.com/ovirt
> * Check out the latest project news on the oVirt blog:
> http://www.ovirt.org/blog/
>
> [1] http://www.ovirt.org/release/4.3.7/
> [2] http://resources.ovirt.org/pub/ovirt-4.3-pre/iso/
>
> --
>
> Sandro Bonazzola
>
> MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
>
> Red Hat EMEA <https://www.redhat.com/>
>
> sbona...@redhat.com
> <https://www.redhat.com/>*Red Hat respects your work life balance.
> Therefore there is no need to answer this email out of your office hours.*
>
> _______________________________________________
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/24QUREJPZHTSMHLDYBUDVZML2DEF7PKQ/
>
>
>
> --
>
> Sandro Bonazzola
>
> MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
>
> Red Hat EMEA <https://www.redhat.com/>
>
> sbona...@redhat.com
> <https://www.redhat.com/>*Red Hat respects your work life balance.
> Therefore there is no need to answer this email out of your office hours.
> <https://mojo.redhat.com/docs/DOC-1199578>*
>
>
>
> --
>
> Sandro Bonazzola
>
> MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
>
> Red Hat EMEA <https://www.redhat.com/>
>
> sbona...@redhat.com
> <https://www.redhat.com/>*Red Hat respects your work life balance.
> Therefore there is no need to answer this email out of your office hours.
> <https://mojo.redhat.com/docs/DOC-1199578>*
>
>
>
> --
>
> Sandro Bonazzola
>
> MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
>
> Red Hat EMEA <https://www.redhat.com/>
>
> sbona...@redhat.com
> <https://www.redhat.com/>*Red Hat respects your work life balance.
> Therefore there is no need to answer this email out of your office hours.
> <https://mojo.redhat.com/docs/DOC-1199578>*
>
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3ZPGKJ4JJWVFD3AFLTMLVSOZL34USCOF/

Reply via email to