[ovirt-users] Re: NVMe disk cause kernel panic in installer

2019-07-19 Thread bossman90
Unfortunately, it didn't help.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/2YSRWPT536PAEWIJE6K5DPLZ7BDV24YJ/


[ovirt-users] Re: NVMe disk cause kernel panic in installer

2019-07-19 Thread Strahil Nikolov
 I have seen similar behaviour in Anakonda even with regular disks.Please try 
to boot from the instalaltion media and go for troubleshooting.Then , just get 
a console (Ctrl + Alt + Fx - where x is between 1 and 12) and wipe the 
beginning of your NVMe.Something like:dd if=/dev/zero of=/dev/nvme0n1 bs=4M 
count=100 status=progress
Then reboot and try again.
I assume that you do not have anything on it.
Best Regards,Strahil Nikolov
В четвъртък, 18 юли 2019 г., 19:53:57 ч. Гринуич+3, bossma...@gmail.com 
 написа:  
 
 While installer is loading.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/NBNQE62UUKZ34ANXYIDB2SUOW7EY2KA3/
  ___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QI2CMCXLDO2QIBKW3STJWJUFNJLGJL47/


[ovirt-users] Re: Storage domain 'Inactive' but still functional

2019-07-19 Thread Strahil Nikolov
 Hi Martin,
First check what went wrong with the VG -as it could be something 
simple.vgcfgbackup -f VGname will create a file which you can use to compare 
current metadata with a previous version.
If you have Linux boxes - you can add disks from another storage and then 
pvmove the data inside the VM. Of course , you will need to reinstall grub on 
the new OS disk , or you won't be able to boot afterwards.If possible, try with 
a test VM before proceeding with important ones.

Backing up the VMs is very important , because working on LVM metadata is quite 
risky.Last time I had such an issue , I was working on clustered LVs which got 
their PVs "Missing". For me , restore from VG backup fixed the issue - but that 
might not be always the case.
Just get the vgcfgbackup's output and compare with diff or vimdiff and check 
what is different.
Sadly, I think that this is more a Linux problem , than an oVirt problem.
Best Regards,Strahil Nikolov  
В четвъртък, 18 юли 2019 г., 18:51:32 ч. Гринуич+3, Martijn Grendelman 
 написа:  
 
  Hi!

Thanks. Like I wrote, I have metadata backups from /etc/lvm/backup and 
-/archive, and I also have the current metadata as it exists on disk. What I'm 
most concerned about, is the proposed procedure.

I would create a backup of the VG, but I'm not sure what would be the most 
sensible way to do it. I could make a new iSCSI target and simply 'dd' the 
whole disk over, but that would take quite some time (it's 2,5 TB) and there 
are VMs that can't really be down for that long. And I'm not even sure that 
dd'ing the disk like that is a sensible strategy.

Moving disks out of the domain is currently not possible. oVirt says 'Source 
Storage Domain is not active'.

Thanks,
Martijn.


Op 18-7-2019 om 17:44 schreef Strahil Nikolov:

Can you check the /etc/lvm/backup and /etc/lvm/archive on your SPM host (check 
the other hosts, just in case you find anything useful) ?Usually LVM makes 
backup of everything.
I would recommend you to:1. Create a backup of the problematic VG2. Compare the 
backup file and a file from backup/archive folders for the same VGCheck what is 
different with diff/vimdiff . It might give you a clue.
I had some issues (non-related to oVirt) and restoring the VG from older backup 
did help me .Still ,any operation on block devices should be considered risky 
and a proper backup is needed.You could  try to move a less important VM's 
disks out of this storage domain to another one.
If it succeeds - then you can evacuate all VMs away before you can start 
"breaking" the storage domain.
Best Regards,Strahil Nikolov
  
В четвъртък, 18 юли 2019 г., 16:59:46 ч. Гринуич+3, Martijn Grendelman 
 написа: 

Hi,

It appears that O365 has trouble delivering mails to this list, so two
earlier mails of mine are still somewhere in a queue and may yet be delivered.

This mail has all of the content of 3 successive mails. I apologize for this
format.

Op 18-7-2019 om 11:20 schreef Martijn Grendelman:

Op 18-7-2019 om 10:16 schreef Martijn Grendelman:

Hi,

For the first time in many months I have run into some trouble with oVirt 
(4.3.4.3) and I need some help.

Yesterday, I noticed one of my iSCSI storage domains was almost full, and tried 
to move a disk image off of it, to another domain. This failed, and somewhere 
in the process, the whole storage domain went to status 'Inactive'.

>From engine.log:

2019-07-17 16:30:35,319+02 INFO  
[org.ovirt.engine.core.vdsbroker.irsbroker.IrsProxy] 
(EE-ManagedThreadFactory-engine-Thread-1836383) [] starting 
processDomainRecovery for domain 
'875847b6-29a4-4419-be92-9315f4435429:HQST0_ISCSI02'.
2019-07-17 16:30:35,337+02 ERROR 
[org.ovirt.engine.core.vdsbroker.irsbroker.IrsProxy] 
(EE-ManagedThreadFactory-engine-Thread-1836383) [] Domain 
'875847b6-29a4-4419-be92-9315f4435429:HQST0_ISCSI02' was reported by all hosts 
in status UP as problematic. Moving the domain to NonOperational.
2019-07-17 16:30:35,410+02 WARN  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(EE-ManagedThreadFactory-engine-Thread-1836383) [5f6fd35e] EVENT_ID: 
SYSTEM_DEACTIVATED_STORAGE_DOMAIN(970), Storage Domain HQST0_ISCSI02 (Data 
Center ISAAC01) was deactivated by system because it's not visible by any of 
the hosts.

The thing is, the domain is still functional on all my hosts. It carries over 
50 disks, and all involved VMs are up and running, and don't seem to have any 
problems. Also, 'iscsiadm' on all hosts seems to indiciate that everything is 
fine with this specific target and reading from the device with dd, or getting 
its size with 'blockdev' all works without issue.

When I try to reactivate the domain, these errors are logged:


2019-07-18 09:34:53,631+02 ERROR 
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(EE-ManagedThreadFactory-engine-Thread-43475) [79e386e] EVENT_ID: 
IRS_BROKER_COMMAND_FAILURE(10,803), VDSM command ActivateStorageDomainVDS 
failed: Storage domain does not exist: 

[ovirt-users] Re: Error exporting into ova

2019-07-19 Thread Gianluca Cecchi
On Fri, Jul 19, 2019 at 4:14 PM Gianluca Cecchi 
wrote:

> On Fri, Jul 19, 2019 at 3:15 PM Gianluca Cecchi 
> wrote:
>
>>
>>
>> In engine.log the first error I see is 30 minutes after start
>>
>> 2019-07-19 12:25:31,563+02 ERROR
>> [org.ovirt.engine.core.common.utils.ansible.AnsibleExecutor]
>> (EE-ManagedThreadFactory-engineScheduled-Thread-64) [2001ddf4] Ansible
>> playbook execution failed: Timeout occurred while executing Ansible
>> playbook.
>>
>
> In the mean time, as the playbook seems this one ( I run the job from
> engine) : /usr/share/ovirt-engine/playbooks/ovirt-ova-export.yml
>
>
Based on what described in bugzilla
https://bugzilla.redhat.com/show_bug.cgi?id=1697301
I created at the moment the file
/etc/ovirt-engine/engine.conf.d/99-ansible-playbook-timeout.conf
with
ANSIBLE_PLAYBOOK_EXEC_DEFAULT_TIMEOUT=80
and restarted the engine and the python script to verify

Just to see if it completes, even if in my case with a 30Gb preallocated
disk, the source problem is qemu-img convert command very slow in I/O.
It reads from iscsi multipath (2 paths) with 2x3MB/s and it writes on nfs
If I run a dd command from iscsi device mapper device to an nfs file I have
140MB/s rate that is what expected based on my storage array performances
and my network.

Not understood why the qemu-img command is so slow
The question still applies in case I have to do an appliance from a VM with
a very big disk, where the copy could potentially have an elapsed of more
that 30 minutes...
Gianluca
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/DMUARHYE34JLJRAX3BLEPV54EK2CS5SD/


[ovirt-users] Re: Error exporting into ova

2019-07-19 Thread Gianluca Cecchi
On Fri, Jul 19, 2019 at 3:15 PM Gianluca Cecchi 
wrote:

>
>
> In engine.log the first error I see is 30 minutes after start
>
> 2019-07-19 12:25:31,563+02 ERROR
> [org.ovirt.engine.core.common.utils.ansible.AnsibleExecutor]
> (EE-ManagedThreadFactory-engineScheduled-Thread-64) [2001ddf4] Ansible
> playbook execution failed: Timeout occurred while executing Ansible
> playbook.
>

In the mean time, as the playbook seems this one ( I run the job from
engine) : /usr/share/ovirt-engine/playbooks/ovirt-ova-export.yml

- hosts: all
  remote_user: root
  gather_facts: no

  roles:
- ovirt-ova-export-pre-pack
- ovirt-ova-pack
- ovirt-ova-export-post-pack

and it seems async is not supported, I temporary changed the main.yml
inside ovirt-ova-pack role from

--
- name: Run packing script
  script: >
pack_ova.py
...

to

---
- name: Copy packing script
  copy:
src: pack_ova.py
dest: /tmp/pack_ova.py
mode: '0755'

- name: Run packing script
  command: timeout 4800 /tmp/pack_ova.py
...

just to check... see how it goes now
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/3QSRA47E7STVEFX3WLPFPUC3YNNSCPWR/


[ovirt-users] Error exporting into ova

2019-07-19 Thread Gianluca Cecchi
Hello,
I'm playing with export_vm_as_ova.py downloaded from the examples github:
https://github.com/oVirt/ovirt-engine-sdk/blob/master/sdk/examples/export_vm_as_ova.py

My environment is oVirt 4.3.3.7 with iSCSI storage domain.
It fails leaving an ova.tmp file

In webadmin gui:
Starting to export Vm enginecopy1 as a Virtual Appliance
7/19/1911:55:12 AM

VDSM ov301 command TeardownImageVDS failed: Cannot deactivate Logical
Volume: ('General Storage Exception: ("5 [] [\' Logical volume
fa33df49-b09d-4f86-9719-ede649542c21/0420ef47-0ad0-4cf9-babd-d89383f7536b
in
use.\']\\nfa33df49-b09d-4f86-9719-ede649542c21/[\'a7480dc5-b5ca-4cb3-986d-77bc12165be4\',
\'0420ef47-0ad0-4cf9-babd-d89383f7536b\']",)',)
7/19/1912:25:36 PM

Failed to export Vm enginecopy1 as a Virtual Appliance to path
/save_ova/base/dump/myvm2.ova on Host ov301
7/19/1912:25:37 PM

During export I have this qemu-img process creating the disk over the loop
device:

root 30878 30871  0 11:55 pts/200:00:00 su -p -c qemu-img convert
-T none -O qcow2
'/rhev/data-center/mnt/blockSD/fa33df49-b09d-4f86-9719-ede649542c21/images/59a4a324-4c99-4ff5-abb1-e9bbac83292a/0420ef47-0ad0-4cf9-babd-d89383f7536b'
'/dev/loop1' vdsm
vdsm 30882 30878 10 11:55 ?00:00:00 qemu-img convert -T none -O
qcow2
/rhev/data-center/mnt/blockSD/fa33df49-b09d-4f86-9719-ede649542c21/images/59a4a324-4c99-4ff5-abb1-e9bbac83292a/0420ef47-0ad0-4cf9-babd-d89383f7536b
/dev/loop1

The ova.tmp file is getting filled while command runs
eg:
[root@ov301 ]# du -sh /save_ova/base/dump/myvm2.ova.tmp
416M /save_ova/base/dump/myvm2.ova.tmp
[root@ov301 sysctl.d]#

[root@ov301 sysctl.d]# du -sh /save_ova/base/dump/myvm2.ova.tmp
911M /save_ova/base/dump/myvm2.ova.tmp
[root@ov301 ]#

and the final generated / not completed file is in this state:
[root@ov301 ]# qemu-img info /save_ova/base/dump/myvm2.ova.tmp
image: /save_ova/base/dump/myvm2.ova.tmp
file format: raw
virtual size: 30G (32217446400 bytes)
disk size: 30G
[root@ov301 sysctl.d]#

But I notice that the timestamp of the file is about 67 minutes after start
of job and well after the notice of its failure

[root@ov301 sysctl.d]# ll /save_ova/base/dump/
total 30963632
-rw---. 1 root root 32217446400 Jul 19 13:02 myvm2.ova.tmp
[root@ov301 sysctl.d]#

[root@ov301 sysctl.d]# du -sh /save_ova/base/dump/myvm2.ova.tmp
30G /save_ova/base/dump/myvm2.ova.tmp
[root@ov301 sysctl.d]#

In engine.log the first error I see is 30 minutes after start

2019-07-19 12:25:31,563+02 ERROR
[org.ovirt.engine.core.common.utils.ansible.AnsibleExecutor]
(EE-ManagedThreadFactory-engineScheduled-Thread-64) [2001ddf4] Ansible
playbook execution failed: Timeout occurred while executing Ansible
playbook.
2019-07-19 12:25:31,563+02 INFO
 [org.ovirt.engine.core.common.utils.ansible.AnsibleExecutor]
(EE-ManagedThreadFactory-engineScheduled-Thread-64) [2001ddf4] Ansible
playbook command has exited with value: 1
2019-07-19 12:25:31,564+02 ERROR
[org.ovirt.engine.core.bll.CreateOvaCommand]
(EE-ManagedThreadFactory-engineScheduled-Thread-64) [2001ddf4] Failed to
create OVA. Please check logs for more details:
/var/log/ovirt-engine/ova/ovirt-export-ova-ansible-20190719115531-ov301-2001ddf4.log
2019-07-19 12:25:31,565+02 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.TeardownImageVDSCommand]
(EE-ManagedThreadFactory-engineScheduled-Thread-64) [2001ddf4] START,
TeardownImageVDSCommand(HostName = ov301,
ImageActionsVDSCommandParameters:{hostId='8ef1ce6f-4e38-486c-b3a4-58235f1f1d06'}),
log id: 3d2246f7
2019-07-19 12:25:36,569+02 ERROR
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(EE-ManagedThreadFactory-engineScheduled-Thread-64) [2001ddf4] EVENT_ID:
VDS_BROKER_COMMAND_FAILURE(10,802), VDSM ov301 command TeardownImageVDS
failed: Cannot deactivate Logical Volume: ('General Storage Exception: ("5
[] [\'  Logical volume
fa33df49-b09d-4f86-9719-ede649542c21/0420ef47-0ad0-4cf9-babd-d89383f7536b
in
use.\']\\nfa33df49-b09d-4f86-9719-ede649542c21/[\'a7480dc5-b5ca-4cb3-986d-77bc12165be4\',
\'0420ef47-0ad0-4cf9-babd-d89383f7536b\']",)',)
In ansible playbook suggested log file I don't see anything useful.
It ends with timestamps when the script has been launched.
Last lines are:

2019-07-19 11:55:33,877 p=5699 u=ovirt |  TASK [ovirt-ova-export-pre-pack :
Retrieving the temporary path for the OVA file] ***
2019-07-19 11:55:34,198 p=5699 u=ovirt |  changed: [ov301] => {
"changed": true,
"dest": "/save_ova/base/dump/myvm2.ova.tmp",
"gid": 0,
"group": "root",
"mode": "0600",
"owner": "root",
"secontext": "system_u:object_r:nfs_t:s0",
"size": 32217446912,
"state": "file",
"uid": 0
}
2019-07-19 11:55:34,204 p=5699 u=ovirt |  TASK [ovirt-ova-pack : Run
packing script] *

It seems 30 minutes... for timeout? About what, ansible job?
Or possibly implicit user session created when running the python script?
The snapshot has been correctly deleted (as I see also in engine.log), I
don't 

[ovirt-users] Re: Stuck in "Finalizing" disk upload phase

2019-07-19 Thread Vrgotic, Marko
Hi Daniel,

Is it possible that upgrade from 4.3.3 to 4.3.4  affected/reset/changed the 
certificate for imageio?

Reason why I am asking is that, while running 4.3.3 version, I was uploading 
Ubuntu 16-18, CentOS 7 cloud images, and CentOS6 no issues.
Now issue is with CentOS6. 
I have tried to upload centos6.5 image via Ansible ovirt_disk and than one 
ended stuck in “Finalizing” state.
Than I tried with to Upload same image from UI and that ended in “Paused” state.

I will try Test Connection and let you know.

Sent from my iPhone

> On 18 Jul 2019, at 16:13, Daniel Erez  wrote:
> 
> Hi Marko,
> 
> According to the log, it seems there's an issue with a secured
> connection to the proxy.
> Can you please try the 'Test Connection' button on upload disk dialog,
> to ensure the certificate is properly installed?
> Also, is the disk currently in 'Paused by System' status?
> 
> Thanks,
> Daniel
> 
> On Thu, Jul 18, 2019 at 2:35 PM Vrgotic, Marko
>  wrote:
>> 
>> Hi Nir,
>> 
>> 
>> 
>> Since this is already compressed file, I have added output of grep for 
>> moment this command was issue and following 300lines.
>> 
>> 
>> 
>> [root@ovirt-engine ovirt-engine]# zgrep "2019-07-17 13:36:11,121Z" 
>> engine.log-20190* -A300
>> 
>> engine.log-20190718.gz:2019-07-17 13:36:11,121Z INFO  
>> [org.ovirt.engine.core.bll.storage.disk.image.TransferDiskImageCommand] 
>> (default task-376) [a43180ec-afc7-429e-9f30-9e851eaf7ce7] Running command: 
>> TransferDiskImageCommand internal: false. Entities affected :  ID: 
>> 3452459d-aec6-430e-9509-1d9ca815b2d8 Type: DiskAction group 
>> EDIT_DISK_PROPERTIES with role type USER
>> 
>> engine.log-20190718.gz:2019-07-17 13:36:11,121Z INFO  
>> [org.ovirt.engine.core.bll.storage.disk.image.TransferDiskImageCommand] 
>> (default task-376) [a43180ec-afc7-429e-9f30-9e851eaf7ce7] Creating 
>> ImageTransfer entity for command '28eda0c2-e36b-4e70-91ea-2ecf4a030d19'
>> 
>> gzip: -A300.gz: No such file or directory
>> 
>> [root@ovirt-engine ovirt-engine]# zgrep -A 300 "2019-07-17 13:36:11,121Z" 
>> engine.log-20190*
>> 
>> engine.log-20190718.gz:2019-07-17 13:36:11,121Z INFO  
>> [org.ovirt.engine.core.bll.storage.disk.image.TransferDiskImageCommand] 
>> (default task-376) [a43180ec-afc7-429e-9f30-9e851eaf7ce7] Running command: 
>> TransferDiskImageCommand internal: false. Entities affected :  ID: 
>> 3452459d-aec6-430e-9509-1d9ca815b2d8 Type: DiskAction group 
>> EDIT_DISK_PROPERTIES with role type USER
>> 
>> engine.log-20190718.gz:2019-07-17 13:36:11,121Z INFO  
>> [org.ovirt.engine.core.bll.storage.disk.image.TransferDiskImageCommand] 
>> (default task-376) [a43180ec-afc7-429e-9f30-9e851eaf7ce7] Creating 
>> ImageTransfer entity for command '28eda0c2-e36b-4e70-91ea-2ecf4a030d19'
>> 
>> engine.log-20190718.gz:2019-07-17 13:36:11,142Z INFO  
>> [org.ovirt.engine.core.bll.storage.disk.image.TransferDiskImageCommand] 
>> (default task-376) [a43180ec-afc7-429e-9f30-9e851eaf7ce7] Successfully added 
>> Upload disk 'av-07-centos-65-base' (disk id: 
>> '3452459d-aec6-430e-9509-1d9ca815b2d8', image id: 
>> 'b44659a9-607a-4eeb-a255-99532fd4fce4') for image transfer command 
>> '28eda0c2-e36b-4e70-91ea-2ecf4a030d19'
>> 
>> engine.log-20190718.gz:2019-07-17 13:36:11,188Z INFO  
>> [org.ovirt.engine.core.vdsbroker.vdsbroker.PrepareImageVDSCommand] (default 
>> task-376) [a43180ec-afc7-429e-9f30-9e851eaf7ce7] START, 
>> PrepareImageVDSCommand(HostName = ovirt-hv-01.avinity.tv, 
>> PrepareImageVDSCommandParameters:{hostId='e7e3f1dc-8037-4e74-a44c-442bdb02197d'}),
>>  log id: 3bd84b38
>> 
>> engine.log-20190718.gz:2019-07-17 13:36:11,351Z INFO  
>> [org.ovirt.engine.core.vdsbroker.vdsbroker.PrepareImageVDSCommand] (default 
>> task-376) [a43180ec-afc7-429e-9f30-9e851eaf7ce7] FINISH, 
>> PrepareImageVDSCommand, return: PrepareImageReturn:{status='Status [code=0, 
>> message=Done]'}, log id: 3bd84b38
>> 
>> engine.log-20190718.gz:2019-07-17 13:36:11,356Z INFO  
>> [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeLegalityVDSCommand] 
>> (default task-376) [a43180ec-afc7-429e-9f30-9e851eaf7ce7] START, 
>> SetVolumeLegalityVDSCommand( 
>> SetVolumeLegalityVDSCommandParameters:{storagePoolId='10525aa0-839d-11e9-a016-00163e4f2a6d',
>>  ignoreFailoverLimit='false', 
>> storageDomainId='644aacaa-12e1-4fcd-b3aa-941678cf95bd', 
>> imageGroupId='3452459d-aec6-430e-9509-1d9ca815b2d8', 
>> imageId='b44659a9-607a-4eeb-a255-99532fd4fce4'}), log id: 5e0f7654
>> 
>> engine.log-20190718.gz:2019-07-17 13:36:11,390Z INFO  
>> [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeLegalityVDSCommand] 
>> (default task-376) [a43180ec-afc7-429e-9f30-9e851eaf7ce7] FINISH, 
>> SetVolumeLegalityVDSCommand, return: , log id: 5e0f7654
>> 
>> engine.log-20190718.gz:2019-07-17 13:36:11,394Z INFO  
>> [org.ovirt.engine.core.vdsbroker.vdsbroker.AddImageTicketVDSCommand] 
>> (default task-376) [a43180ec-afc7-429e-9f30-9e851eaf7ce7] START, 
>> AddImageTicketVDSCommand(HostName = ovirt-hv-01.avinity.tv, 
>>