On Thu, Nov 21, 2019 at 8:20 AM Sahina Bose <sab...@redhat.com> wrote:

>
>
> On Thu, Nov 21, 2019 at 6:03 AM Strahil Nikolov <hunter86...@yahoo.com>
> wrote:
>
>> Hi All,
>>
>> another clue in the logs :
>> [2019-11-21 00:29:50.536631] W [MSGID: 114031]
>> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-1:
>> remote operation failed. Path:
>> /.shard/b0af2b81-22cf-482e-9b2f-c431b6449dae.79
>> (00000000-0000-0000-0000-000000000000) [Permission denied]
>> [2019-11-21 00:29:50.536798] W [MSGID: 114031]
>> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-0:
>> remote operation failed. Path:
>> /.shard/b0af2b81-22cf-482e-9b2f-c431b6449dae.79
>> (00000000-0000-0000-0000-000000000000) [Permission denied]
>> [2019-11-21 00:29:50.536959] W [MSGID: 114031]
>> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-2:
>> remote operation failed. Path:
>> /.shard/b0af2b81-22cf-482e-9b2f-c431b6449dae.79
>> (00000000-0000-0000-0000-000000000000) [Permission denied]
>> [2019-11-21 00:29:50.537007] E [MSGID: 133010]
>> [shard.c:2327:shard_common_lookup_shards_cbk] 0-data_fast-shard: Lookup on
>> shard 79 failed. Base file gfid = b0af2b81-22cf-482e-9b2f-c431b6449dae
>> [Permission denied]
>> [2019-11-21 00:29:50.537066] W [fuse-bridge.c:2830:fuse_readv_cbk]
>> 0-glusterfs-fuse: 12458: READ => -1
>> gfid=b0af2b81-22cf-482e-9b2f-c431b6449dae fd=0x7fc63c00fe18 (Permission
>> denied)
>> [2019-11-21 00:30:01.177665] I [MSGID: 133022]
>> [shard.c:3674:shard_delete_shards] 0-data_fast-shard: Deleted shards of
>> gfid=eb103fbf-80dc-425d-882f-1e4efe510db5 from backend
>> [2019-11-21 00:30:13.132756] W [MSGID: 114031]
>> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-0:
>> remote operation failed. Path:
>> /.shard/17c663c2-f582-455b-b806-3b9d01fb2c6c.79
>> (00000000-0000-0000-0000-000000000000) [Permission denied]
>> [2019-11-21 00:30:13.132824] W [MSGID: 114031]
>> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-1:
>> remote operation failed. Path:
>> /.shard/17c663c2-f582-455b-b806-3b9d01fb2c6c.79
>> (00000000-0000-0000-0000-000000000000) [Permission denied]
>> [2019-11-21 00:30:13.133217] W [MSGID: 114031]
>> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-2:
>> remote operation failed. Path:
>> /.shard/17c663c2-f582-455b-b806-3b9d01fb2c6c.79
>> (00000000-0000-0000-0000-000000000000) [Permission denied]
>> [2019-11-21 00:30:13.133238] E [MSGID: 133010]
>> [shard.c:2327:shard_common_lookup_shards_cbk] 0-data_fast-shard: Lookup on
>> shard 79 failed. Base file gfid = 17c663c2-f582-455b-b806-3b9d01fb2c6c
>> [Permission denied]
>> [2019-11-21 00:30:13.133264] W [fuse-bridge.c:2830:fuse_readv_cbk]
>> 0-glusterfs-fuse: 12660: READ => -1
>> gfid=17c663c2-f582-455b-b806-3b9d01fb2c6c fd=0x7fc63c007038 (Permission
>> denied)
>> [2019-11-21 00:30:38.489449] W [MSGID: 114031]
>> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-0:
>> remote operation failed. Path:
>> /.shard/a10a5ae8-108b-4d78-9e65-cca188c27fc4.6
>> (00000000-0000-0000-0000-000000000000) [Permission denied]
>> [2019-11-21 00:30:38.489520] W [MSGID: 114031]
>> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-1:
>> remote operation failed. Path:
>> /.shard/a10a5ae8-108b-4d78-9e65-cca188c27fc4.6
>> (00000000-0000-0000-0000-000000000000) [Permission denied]
>> [2019-11-21 00:30:38.489669] W [MSGID: 114031]
>> [client-rpc-fops_v2.c:2634:client4_0_lookup_cbk] 0-data_fast-client-2:
>> remote operation failed. Path:
>> /.shard/a10a5ae8-108b-4d78-9e65-cca188c27fc4.6
>> (00000000-0000-0000-0000-000000000000) [Permission denied]
>> [2019-11-21 00:30:38.489717] E [MSGID: 133010]
>> [shard.c:2327:shard_common_lookup_shards_cbk] 0-data_fast-shard: Lookup on
>> shard 6 failed. Base file gfid = a10a5ae8-108b-4d78-9e65-cca188c27fc4
>> [Permission denied]
>> [2019-11-21 00:30:38.489777] W [fuse-bridge.c:2830:fuse_readv_cbk]
>> 0-glusterfs-fuse: 12928: READ => -1
>> gfid=a10a5ae8-108b-4d78-9e65-cca188c27fc4 fd=0x7fc63c01a058 (Permission
>> denied)
>>
>>
>> Anyone got an idea why is it happening?
>> I checked user/group and selinux permissions - all OK
>>
>
Can you share the commands (and output) used to check this?

I would try this on the hypervisor to check what vdsm/qemu see:

$ ls -lahRZ /rhv/data-center/mnt/glusterSD/gluster-server:_path

Also, to make sure we don't have selinux issue on the hypervisor, you can
change
selinux to permissive mode:

    setenforce 0

And then try again. If this was selinux issue the permission denied issue
will disappear.
If this is the case please provide the output of:

    ausearh -m AVC -ts today

If the issue still exists, we eliminated selinux, and you can enable it
again:

    setenforce 1

Nir


> I even replaced one of the disks and healed , but the result is the same
>> for all my VMs.
>>
>
> Have you checked the permission for user/group are set correctly across
> all the bricks in the cluster?
> What does ls -la on the images directory from mount of the volume show you.
>
> Adding Krutika and Rafi as they ran into a similar issue in the past.
>
>
>> Best Regards,
>> Strahil Nikolov
>>
>>
>> В сряда, 20 ноември 2019 г., 18:17:18 ч. Гринуич+2, Strahil Nikolov <
>> hunter86...@yahoo.com> написа:
>>
>>
>> Hello All,
>>
>> my engine is back online , but I'm still having difficulties to make vdsm
>> powerup the systems.
>> I think that the events generated today can lead me to the right
>> direction(just an example , many more are there):
>>
>> VDSM ovirt3.localdomain command SpmStatusVDS failed: Cannot inquire
>> Lease(name='SDM',
>> path=u'/rhev/data-center/mnt/glusterSD/gluster1:_data__fast3/ecc3bf0e-8214-45c1-98a6-0afa642e591f/dom_md/leases',
>> offset=1048576): (2, 'Sanlock get hosts failure', 'No such file or
>> directory')
>>
>> I will try to collect a fresh log and see what is it complaining about
>> this time.
>>
>> Best Regards,
>> Strahil Nikolov
>>
>> >Hi Sahina,
>>
>> >I have a strange situation:
>> >1. When I try to access the file via 'sudo -u vdsm dd if=disk of=test
>> bs=4M' the command fails on aprox 60MB.
>> >2. If I run same command as root , remove the file and then run again
>> via vdsm user -> this time no i/o error reported.
>>
>> >My guess is that I need to check what's going on the bricks themselve ...
>>
>> >Best Regards,
>> >Strahil Nikolov
>>
>>
>> В вторник, 19 ноември 2019 г., 0:02:16 ч. Гринуич-5, Sahina Bose <
>> sab...@redhat.com> написа:
>>
>>
>>
>>
>> On Tue, Nov 19, 2019 at 10:10 AM Strahil Nikolov <hunter86...@yahoo.com>
>> wrote:
>>
>> Hi Sahina,
>>
>> Sadly engine logs have no errors.
>> I've got only an I/O error, but in the debug of the vdsm I can clearly
>> see that "qemu-img" is giving an "OK".
>> During the upgrade I got some metadata files pending heal, but I have
>> recovered the conflict manually and should be OK.
>> Today I have defined one of the VMs manually (virsh define) and then
>> started it , but the issue is the same.
>> It seems to be storage-related issue,as VMs that are on specific domain
>> can be started , but most of my VMs are on the fast storage domains and
>> none of them can be started.
>>
>> After the gluster snapshot restore , the engine is having issues and I
>> have to separately investigate that (as I poweroff my HostedEngine before
>> creating the snapshot).
>>
>> The logs can be find at :
>> https://drive.google.com/open?id=1VAZFZWWrpimDeVuZT0sWFVXy76scr4NM
>>
>>
>> Any ideas where to look at , as I can definitely read (using "dd if=disk"
>> or qemu-img info) the disks of the rhel7 VM ?
>>
>>
>> The vdsm logs have this:
>> 2019-11-17 10:21:23,892+0200 INFO  (libvirt/events) [virt.vm]
>> (vmId='b3c4d84a-9784-470c-b70e-7ad7cc45e913') abnormal vm stop device
>> ua-94f763e9-fd96-4bee-a6b2-31af841a918b error eother (vm:5075)
>> 2019-11-17 10:21:23,892+0200 INFO  (libvirt/events) [virt.vm]
>> (vmId='b3c4d84a-9784-470c-b70e-7ad7cc45e913') CPU stopped: onIOError
>> (vm:6062)
>> 2019-11-17 10:21:23,893+0200 DEBUG (libvirt/events)
>> [jsonrpc.Notification] Sending event {"params": {"notify_time": 4356025830,
>> "b3c4d84a-9784-470c-b70e-7ad7cc45e913": {"status": "WaitForLaunch",
>> "ioerror": {"alias": "ua-94f763e9-fd96-4bee-a6b2-31af841a918b", "name":
>> "sda", "path":
>> "/rhev/data-center/mnt/glusterSD/gluster1:_data__fast/396604d9-2a9e-49cd-9563-fdc79981f67b/images/94f763e9-fd96-4bee-a6b2-31af841a918b/5b1d3113-5cca-4582-9029-634b16338a2f"},
>> "pauseCode": "EOTHER"}}, "jsonrpc": "2.0", "method":
>> "|virt|VM_status|b3c4d84a-9784-470c-b70e-7ad7cc45e913"} (__init__:181)
>>
>> Can you check the permissions of the file
>> /rhev/data-center/mnt/glusterSD/gluster1:_data__fast/396604d9-2a9e-49cd-9563-fdc79981f67b/images/94f763e9-fd96-4bee-a6b2-31af841a918b/5b1d3113-5cca-4582-9029-634b16338a2f.
>> Was it reset after upgrade?
>>
>> Are you able to copy this file to a different location and try running a
>> VM with this image?
>>
>> Any errors in the mount log of gluster1:_data__fast volume?
>>
>>
>> Best Regards,
>> Strahil Nikolov
>>
>>
>>
>>
>> В понеделник, 18 ноември 2019 г., 11:38:13 ч. Гринуич+2, Sahina Bose <
>> sab...@redhat.com> написа:
>>
>>
>>
>>
>> On Mon, Nov 18, 2019 at 2:58 PM Sandro Bonazzola <sbona...@redhat.com>
>> wrote:
>>
>> +Sahina Bose <sab...@redhat.com> +Gobinda Das <go...@redhat.com> +Nir
>> Soffer <nsof...@redhat.com> +Tal Nisan <tni...@redhat.com> can you
>> please help here?
>>
>>
>> Il giorno dom 17 nov 2019 alle ore 16:00 Strahil Nikolov <
>> hunter86...@yahoo.com> ha scritto:
>>
>> So far,
>>
>> I have rolled back the engine and the 3 hosts - still cannot manipulate
>> the storage.
>> It seems that gluster itself is working, but vdsm and the oVirt stack
>> cannot access the storage - cannot create new VM disks, cannot start a VM
>> and I'm on the verge of redeploy.
>>
>>
>>
>> Any errors in vdsm logs? engine logs?
>>
>>
>>
>> Best Regards,
>> Strahil Nikolov
>>
>> В събота, 16 ноември 2019 г., 15:40:25 ч. Гринуич+2, Strahil <
>> hunter86...@yahoo.com> написа:
>>
>>
>> I got upgraded  to  RC3 and now  cannot power any VM .
>> Constantly getting I/O error, but checking at gluster level - I can dd
>> from each disk or even create a new one.
>>
>> Removing the HighAvailability doesn't help.
>>
>> I guess I should restore  the engine from the gluster snapshot and
>> rollback via 'yum history undo last'.
>>
>> Does anyone else have my issues  ?
>>
>> Best Regards,
>> Strahil Nikolov
>> On Nov 13, 2019 15:31, Sandro Bonazzola <sbona...@redhat.com> wrote:
>>
>>
>>
>> Il giorno mer 13 nov 2019 alle ore 14:25 Sandro Bonazzola <
>> sbona...@redhat.com> ha scritto:
>>
>>
>>
>> Il giorno mer 13 nov 2019 alle ore 13:56 Florian Schmid <
>> fsch...@ubimet.com> ha scritto:
>>
>> Hello,
>>
>> I have a question about bugs, which are flagged as [downstream clone -
>> 4.3.7], but are not yet released.
>>
>> I'm talking about this bug:
>> https://bugzilla.redhat.com/show_bug.cgi?id=1749202
>>
>> I can't see it in 4.3.7 release notes. Will it be included in a further
>> release candidate? This fix is very important I think and I can't upgrade
>> yet because of this bug.
>>
>>
>>
>> Looking at the bug, the fix was done with $ git tag --contains
>> 12bd5cb1fe7c95e29b4065fca968913722fe9eaa
>> ovirt-engine-4.3.6.6
>> ovirt-engine-4.3.6.7
>> ovirt-engine-4.3.7.0
>> ovirt-engine-4.3.7.1
>>
>> So the fix is already included in release oVirt 4.3.6.
>>
>>
>> Sent a fix to 4.3.6 release notes:
>> https://github.com/oVirt/ovirt-site/pull/2143. @Ryan Barry
>> <rba...@redhat.com> can you please review?
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> BR Florian Schmid
>>
>> ------------------------------
>> *Von: *"Sandro Bonazzola" <sbona...@redhat.com>
>> *An: *"users" <users@ovirt.org>
>> *Gesendet: *Mittwoch, 13. November 2019 13:34:59
>> *Betreff: *[ovirt-users] [ANN] oVirt 4.3.7 Third Release Candidate is
>> now available for testing
>>
>> The oVirt Project is pleased to announce the availability of the oVirt
>> 4.3.7 Third Release Candidate for testing, as of November 13th, 2019.
>>
>> This update is a release candidate of the seventh in a series of
>> stabilization updates to the 4.3 series.
>> This is pre-release software. This pre-release should not to be used in
>> production.
>>
>> This release is available now on x86_64 architecture for:
>> * Red Hat Enterprise Linux 7.7 or later (but <8)
>> * CentOS Linux (or similar) 7.7 or later (but <8)
>>
>> This release supports Hypervisor Hosts on x86_64 and ppc64le
>> architectures for:
>> * Red Hat Enterprise Linux 7.7 or later (but <8)
>> * CentOS Linux (or similar) 7.7 or later (but <8)
>> * oVirt Node 4.3 (available for x86_64 only) has been built consuming
>> CentOS 7.7 Release
>>
>> See the release notes [1] for known issues, new features and bugs fixed.
>>
>> While testing this release candidate please note that oVirt node now
>> includes:
>> - ansible 2.9.0
>> - GlusterFS 6.6
>>
>> Notes:
>> - oVirt Appliance is already available
>> - oVirt Node is already available
>>
>> Additional Resources:
>> * Read more about the oVirt 4.3.7 release highlights:
>> http://www.ovirt.org/release/4.3.7/
>> * Get more oVirt Project updates on Twitter: https://twitter.com/ovirt
>> * Check out the latest project news on the oVirt blog:
>> http://www.ovirt.org/blog/
>>
>> [1] http://www.ovirt.org/release/4.3.7/
>> [2] http://resources.ovirt.org/pub/ovirt-4.3-pre/iso/
>>
>> --
>>
>> Sandro Bonazzola
>>
>> MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
>>
>> Red Hat EMEA <https://www.redhat.com/>
>>
>> sbona...@redhat.com
>> <https://www.redhat.com/>*Red Hat respects your work life balance.
>> Therefore there is no need to answer this email out of your office hours.*
>>
>> _______________________________________________
>> Users mailing list -- users@ovirt.org
>> To unsubscribe send an email to users-le...@ovirt.org
>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>> oVirt Code of Conduct:
>> https://www.ovirt.org/community/about/community-guidelines/
>> List Archives:
>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/24QUREJPZHTSMHLDYBUDVZML2DEF7PKQ/
>>
>>
>>
>> --
>>
>> Sandro Bonazzola
>>
>> MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
>>
>> Red Hat EMEA <https://www.redhat.com/>
>>
>> sbona...@redhat.com
>> <https://www.redhat.com/>*Red Hat respects your work life balance.
>> Therefore there is no need to answer this email out of your office hours.
>> <https://mojo.redhat.com/docs/DOC-1199578>*
>>
>>
>>
>> --
>>
>> Sandro Bonazzola
>>
>> MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
>>
>> Red Hat EMEA <https://www.redhat.com/>
>>
>> sbona...@redhat.com
>> <https://www.redhat.com/>*Red Hat respects your work life balance.
>> Therefore there is no need to answer this email out of your office hours.
>> <https://mojo.redhat.com/docs/DOC-1199578>*
>>
>>
>>
>> --
>>
>> Sandro Bonazzola
>>
>> MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
>>
>> Red Hat EMEA <https://www.redhat.com/>
>>
>> sbona...@redhat.com
>> <https://www.redhat.com/>*Red Hat respects your work life balance.
>> Therefore there is no need to answer this email out of your office hours.
>> <https://mojo.redhat.com/docs/DOC-1199578>*
>>
>>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/4YLTOEMD2REQHCYEGZINTJRFF6B3VROV/

Reply via email to