Hm...
It's strage it doesn't detect the VG , but could be related to the issue.

Accordi g to this:

lvthinpool_tmeta {
id = "WBut10-rAOP-FzA7-bJvr-ZdxL-lB70-jzz1Tv"
status = ["READ", "WRITE"]
flags = []
creation_time = 1545495487 # 2018-12-22 10:18:07 -0600
creation_host = "vmh.cyber-range.lan"
segment_count = 1

segment1 {
start_extent = 0
extent_count = 16192 # 15.8125 Gigabytes

You got 15GiB of metadata, so create your new metadata LV at least 30 GiB.

Best Regards,
Strahil NikolovOn Oct 2, 2019 04:49, jeremy_tourvi...@hotmail.com wrote:
>
> I don't know why I didn't think to get some more info regarding my storage 
> environment and post it here earlier.  My gluster_vg1 volume is on /dev/sda1. 
>  I can access the engine storage directory but I think that is because it is 
> not thin provisioned.  I guess I was too bogged down in solving the problem 
> when I'm stuck in emergency mode.  I had to  sneaker net my USB drive to my 
> system so I could capture some info.  Anyhow here it is: 
>
> # lsblk 
> NAME                                                       MAJ:MIN RM   SIZE 
> RO TYPE  MOUNTPOINT 
> sda                                                          8:0    0   5.5T  
> 0 disk  
> └─sda1                                                       8:1    0   5.5T  
> 0 part  
> sdb                                                          8:16   0 223.6G  
> 0 disk  
> ├─sdb1                                                       8:17   0     1G  
> 0 part  /boot 
> └─sdb2                                                       8:18   0 222.6G  
> 0 part  
>   └─md127                                                    9:127  0 222.5G  
> 0 raid1 
>     ├─onn_vmh-swap                                         253:0    0     4G  
> 0 lvm   [SWAP] 
>     ├─onn_vmh-pool00_tmeta                                 253:1    0     1G  
> 0 lvm   
>     │ └─onn_vmh-pool00-tpool                               253:3    0 173.6G  
> 0 lvm   
>     │   ├─onn_vmh-ovirt--node--ng--4.2.7.1--0.20181216.0+1 253:4    0 146.6G  
> 0 lvm   / 
>     │   ├─onn_vmh-pool00                                   253:5    0 173.6G  
> 0 lvm   
>     │   ├─onn_vmh-root                                     253:6    0 146.6G  
> 0 lvm   
>     │   ├─onn_vmh-home                                     253:7    0     1G  
> 0 lvm   /home 
>     │   ├─onn_vmh-tmp                                      253:8    0     1G  
> 0 lvm   /tmp 
>     │   ├─onn_vmh-var                                      253:9    0    15G  
> 0 lvm   /var 
>     │   ├─onn_vmh-var_log                                  253:10   0     8G  
> 0 lvm   /var/log 
>     │   ├─onn_vmh-var_log_audit                            253:11   0     2G  
> 0 lvm   /var/log/audit 
>     │   └─onn_vmh-var_crash                                253:12   0    10G  
> 0 lvm   /var/crash 
>     └─onn_vmh-pool00_tdata                                 253:2    0 173.6G  
> 0 lvm   
>       └─onn_vmh-pool00-tpool                               253:3    0 173.6G  
> 0 lvm   
>         ├─onn_vmh-ovirt--node--ng--4.2.7.1--0.20181216.0+1 253:4    0 146.6G  
> 0 lvm   / 
>         ├─onn_vmh-pool00                                   253:5    0 173.6G  
> 0 lvm   
>         ├─onn_vmh-root                                     253:6    0 146.6G  
> 0 lvm   
>         ├─onn_vmh-home                                     253:7    0     1G  
> 0 lvm   /home 
>         ├─onn_vmh-tmp                                      253:8    0     1G  
> 0 lvm   /tmp 
>         ├─onn_vmh-var                                      253:9    0    15G  
> 0 lvm   /var 
>         ├─onn_vmh-var_log                                  253:10   0     8G  
> 0 lvm   /var/log 
>         ├─onn_vmh-var_log_audit                            253:11   0     2G  
> 0 lvm   /var/log/audit 
>         └─onn_vmh-var_crash                                253:12   0    10G  
> 0 lvm   /var/crash 
> sdc                                                          8:32   0 223.6G  
> 0 disk  
> └─sdc1                                                       8:33   0 222.6G  
> 0 part  
>   └─md127                                                    9:127  0 222.5G  
> 0 raid1 
>     ├─onn_vmh-swap                                         253:0    0     4G  
> 0 lvm   [SWAP] 
>     ├─onn_vmh-pool00_tmeta                                 253:1    0     1G  
> 0 lvm   
>     │ └─onn_vmh-pool00-tpool                               253:3    0 173.6G  
> 0 lvm   
>     │   ├─onn_vmh-ovirt--node--ng--4.2.7.1--0.20181216.0+1 253:4    0 146.6G  
> 0 lvm   / 
>     │   ├─onn_vmh-pool00                                   253:5    0 173.6G  
> 0 lvm   
>     │   ├─onn_vmh-root                                     253:6    0 146.6G  
> 0 lvm   
>     │   ├─onn_vmh-home                                     253:7    0     1G  
> 0 lvm   /home 
>     │   ├─onn_vmh-tmp                                      253:8    0     1G  
> 0 lvm   /tmp 
>     │   ├─onn_vmh-var                                      253:9    0    15G  
> 0 lvm   /var 
>     │   ├─onn_vmh-var_log                                  253:10   0     8G  
> 0 lvm   /var/log 
>     │   ├─onn_vmh-var_log_audit                            253:11   0     2G  
> 0 lvm   /var/log/audit 
>     │   └─onn_vmh-var_crash                                253:12   0    10G  
> 0 lvm   /var/crash 
>     └─onn_vmh-pool00_tdata                                 253:2    0 173.6G  
> 0 lvm   
>       └─onn_vmh-pool00-tpool                               253:3    0 173.6G  
> 0 lvm   
>         ├─onn_vmh-ovirt--node--ng--4.2.7.1--0.20181216.0+1 253:4    0 146.6G  
> 0 lvm   / 
>         ├─onn_vmh-pool00                                   253:5    0 173.6G  
> 0 lvm   
>         ├─onn_vmh-root                                     253:6    0 146.6G  
> 0 lvm   
>         ├─onn_vmh-home                                     253:7    0     1G  
> 0 lvm   /home 
>         ├─onn_vmh-tmp                                      253:8    0     1G  
> 0 lvm   /tmp 
>         ├─onn_vmh-var                                      253:9    0    15G  
> 0 lvm   /var 
>         ├─onn_vmh-var_log                                  253:10   0     8G  
> 0 lvm   /var/log 
>         ├─onn_vmh-var_log_audit                            253:11   0     2G  
> 0 lvm   /var/log/audit 
>         └─onn_vmh-var_crash                                253:12   0    10G  
> 0 lvm   /var/crash 
> sdd                                                          8:48   0 596.2G  
> 0 disk  
> └─sdd1                                                       8:49   0     4G  
> 0 part  
>   └─gluster_vg3-tmpLV                                      253:13   0     2G  
> 0 lvm   
> sde                                                          8:64   1   7.5G  
> 0 disk  
> └─sde1                                                       8:65   1   7.5G  
> 0 part  /mnt 
>
> # blkid 
> /dev/sda1: UUID="f026a2dc-201a-4b43-974e-2419a8783bce" TYPE="xfs" 
> PARTLABEL="Linux filesystem" PARTUUID="4bca8a3a-42f0-4877-aa60-f544bf1fdce7" 
> /dev/sdc1: UUID="e5f4acf5-a4bc-6470-7b6f-415e3f4077ff" 
> UUID_SUB="a895900e-5585-8f31-7515-1ff7534e39d7" 
> LABEL="vmh.cyber-range.lan:pv00" TYPE="linux_raid_member" 
> /dev/sdb1: UUID="9b9546f9-25d2-42a6-835b-303f32aee4b1" TYPE="ext4" 
> /dev/sdb2: UUID="e5f4acf5-a4bc-6470-7b6f-415e3f4077ff" 
> UUID_SUB="6e20b5dd-0152-7f42-22a7-c17133fbce45" 
> LABEL="vmh.cyber-range.lan:pv00" TYPE="linux_raid_member" 
> /dev/sdd1: UUID="2nLjVF-sh3N-0qkm-aUQ1-jnls-3e8W-tUkBw5" TYPE="LVM2_member" 
> /dev/md127: UUID="Mq1chn-6XhF-WCwF-LYhl-tZEz-Y8lq-8R2Ifq" TYPE="LVM2_member" 
> /dev/mapper/onn_vmh-swap: UUID="1b0b9c91-22ed-41d1-aebf-e22fd9aa05d9" 
> TYPE="swap" 
> /dev/mapper/onn_vmh-ovirt--node--ng--4.2.7.1--0.20181216.0+1: 
> UUID="b0e1c479-9696-4e19-b799-7f81236026b7" TYPE="ext4" 
> /dev/mapper/onn_vmh-root: UUID="60905f5d-ed91-4ca9-9729-9a72a4678ddd" 
> TYPE="ext4" 
> /dev/mapper/onn_vmh-home: UUID="82a1d567-f8af-4b96-bfbf-5f79dff7384f" 
> TYPE="ext4" 
> /dev/mapper/onn_vmh-tmp: UUID="7dd9d3ae-3af7-4763-9683-19f583d8d15b" 
> TYPE="ext4" 
> /dev/mapper/onn_vmh-var: UUID="f206e030-876b-45a9-8a90-a0e54005b85c" 
> TYPE="ext4" 
> /dev/mapper/onn_vmh-var_log: UUID="b8a12f56-0818-416c-9fb7-33b48ef29eed" 
> TYPE="ext4" 
> /dev/mapper/onn_vmh-var_log_audit: 
> UUID="bc78ad0c-9ab6-4f57-a69f-5b1ddf898552" TYPE="ext4" 
> /dev/mapper/onn_vmh-var_crash: UUID="a941d416-4d7d-41ae-bcd4-8c1ec9d0f744" 
> TYPE="ext4" 
> /dev/sde1: UUID="44aa40d0-6c82-4e8e-8218-177e5c8474f4" TYPE="ext4" 
>
> # pvscan 
>   PV /dev/md127   VG onn_vmh         lvm2 [222.44 GiB / 43.66 GiB free] 
>   PV /dev/sdd1    VG gluster_vg3     lvm2 [<4.00 GiB / <2.00 GiB free] 
>   Total: 2 [<226.44 GiB] / in use: 2 [<226.44 GiB] / in no VG: 0 [0   ] 
>   Reading all physical volumes.  This may take a while... 
>   Found volume group "onn_vmh" using metadata type lvm2 
>   Found volume group "gluster_vg3" using metadata type lvm2 
>
> # lvscan 
>   ACTIVE            '/dev/onn_vmh/pool00' [173.60 GiB] inherit 
>   ACTIVE            '/dev/onn_vmh/root' [146.60 GiB] inherit 
>   ACTIVE            '/dev/onn_vmh/home' [1.00 GiB] inherit 
>   ACTIVE            '/dev/onn_vmh/tmp' [1.00 GiB] inherit 
>   ACTIVE            '/dev/onn_vmh/var' [15.00 GiB] inherit 
>   ACTIVE            '/dev/onn_vmh/var_log' [8.00 GiB] inherit 
>   ACTIVE            '/dev/onn_vmh/var_log_audit' [2.00 GiB] inherit 
>   ACTIVE            '/dev/onn_vmh/swap' [4.00 GiB] inherit 
>   inactive          '/dev/onn_vmh/ovirt-node-ng-4.2.7.1-0.20181216.0' [146.60 
> GiB] inherit 
>   ACTIVE            '/dev/onn_vmh/ovirt-node-ng-4.2.7.1-0.20181216.0+1' 
> [146.60 GiB] inherit 
>   ACTIVE            '/dev/onn_vmh/var_crash' [10.00 GiB] inherit 
>   ACTIVE            '/dev/gluster_vg3/tmpLV' [2.00 GiB] inherit 
>
> [/etc/lvm/backup/gluster_vg1] 
> # Generated by LVM2 version 2.02.180(2)-RHEL7 (2018-07-20): Sat Dec 22 
> 10:18:46 2018 
>
> contents = "Text Format Volume Group" 
> version = 1 
>
> description = "Created *after* executing '/usr/sbin/lvcreate --virtualsize 
> 500GB --name lv_datadisks -T gluster_vg1/lvthinpool'" 
>
> creation_host = "vmh.cyber-range.lan" # Linux vmh.cyber-range.lan 
> 3.10.0-957.1.3.el7.x86_64 #1 SMP Thu Nov 29 14:49:43 UTC 2018 x86_64 
> creation_time = 1545495526 # Sat Dec 22 10:18:46 2018 
>
> gluster_vg1 { 
> id = "TfNqtn-2eX6-i5gC-w4ye-h29n-5Zfy-UFHSvU" 
> seqno = 9 
> format = "lvm2" # informational 
> status = ["RESIZEABLE", "READ", "WRITE"] 
> flags = [] 
> extent_size = 2048 # 1024 Kilobytes 
> max_lv = 0 
> max_pv = 0 
> metadata_copies = 0 
>
> physical_volumes { 
>
> pv0 { 
> id = "DRfoKl-TUhb-cirx-Oz9P-mZEY-XoiB-mdLy6v" 
> device = "/dev/sda" # Hint only 
>
> status = ["ALLOCATABLE"] 
> flags = [] 
> dev_size = 11718885376 # 5.45703 Terabytes 
> pe_start = 2048 
> pe_count = 5722111 # 5.45703 Terabytes 
> } 
> } 
>
> logical_volumes { 
>
> engine_lv { 
> id = "P2ScEB-ws3V-iqVv-XTWh-Y2Jh-pg0c-6KnJXC" 
> status = ["READ", "WRITE", "VISIBLE"] 
> flags = [] 
> creation_time = 1545495460 # 2018-12-22 10:17:40 -0600 
> creation_host = "vmh.cyber-range.lan" 
> segment_count = 1 
>
> segment1 { 
> start_extent = 0 
> extent_count = 102400 # 100 Gigabytes 
>
> type = "striped" 
> stripe_count = 1 # linear 
>
> stripes = [ 
> "pv0", 0 
> ] 
> } 
> } 
>
> lvthinpool { 
> id = "c0yaNn-DcaB-cYjj-9ZRv-M2Em-rzLW-qY9WpI" 
> status = ["READ", "WRITE", "VISIBLE"] 
> flags = [] 
> creation_time = 1545495487 # 2018-12-22 10:18:07 -0600 
> creation_host = "vmh.cyber-range.lan" 
> segment_count = 1 
>
> segment1 { 
> start_extent = 0 
> extent_count = 512000 # 500 Gigabytes 
>
> type = "thin-pool" 
> metadata = "lvthinpool_tmeta" 
> pool = "lvthinpool_tdata" 
> transaction_id = 2 
> chunk_size = 2048 # 1024 Kilobytes 
> discards = "passdown" 
> zero_new_blocks = 1 
> } 
> } 
>
> lv_vmdisks { 
> id = "erpXRi-nPUq-mCf2-ga2J-3a0l-OWiC-i0xr8M" 
> status = ["READ", "WRITE", "VISIBLE"] 
> flags = [] 
> creation_time = 1545495493 # 2018-12-22 10:18:13 -0600 
> creation_host = "vmh.cyber-range.lan" 
> segment_count = 1 
>
> segment1 { 
> start_extent = 0 
> extent_count = 4613735 # 4.4 Terabytes 
>
> type = "thin" 
> thin_pool = "lvthinpool" 
> transaction_id = 0 
> device_id = 1 
> } 
> } 
>
> lv_datadisks { 
> id = "hKim3z-1QCh-dwhU-st2O-t4tG-wIss-UpLMZw" 
> status = ["READ", "WRITE", "VISIBLE"] 
> flags = [] 
> creation_time = 1545495526 # 2018-12-22 10:18:46 -0600 
> creation_host = "vmh.cyber-range.lan" 
> segment_count = 1 
>
> segment1 { 
> start_extent = 0 
> extent_count = 512000 # 500 Gigabytes 
>
> type = "thin" 
> thin_pool = "lvthinpool" 
> transaction_id = 1 
> device_id = 2 
> } 
> } 
>
> lvol0_pmspare { 
> id = "bHc0eC-Z4Ed-mV47-QTU4-SCvo-FWbE-L8NV7Q" 
> status = ["READ", "WRITE"] 
> flags = [] 
> creation_time = 1545495487 # 2018-12-22 10:18:07 -0600 
> creation_host = "vmh.cyber-range.lan" 
> segment_count = 1 
>
> segment1 { 
> start_extent = 0 
> extent_count = 16192 # 15.8125 Gigabytes 
>
> type = "striped" 
> stripe_count = 1 # linear 
>
> stripes = [ 
> "pv0", 102400 
> ] 
> } 
> } 
>
> lvthinpool_tmeta { 
> id = "WBut10-rAOP-FzA7-bJvr-ZdxL-lB70-jzz1Tv" 
> status = ["READ", "WRITE"] 
> flags = [] 
> creation_time = 1545495487 # 2018-12-22 10:18:07 -0600 
> creation_host = "vmh.cyber-range.lan" 
> segment_count = 1 
>
> segment1 { 
> start_extent = 0 
> extent_count = 16192 # 15.8125 Gigabytes 
>
> type = "striped" 
> stripe_count = 1 # linear 
>
> stripes = [ 
> "pv0", 630592 
> ] 
> } 
> } 
>
> lvthinpool_tdata { 
> id = "rwNZux-1fz1-dv8J-yN2j-LcES-f6ml-231td5" 
> status = ["READ", "WRITE"] 
> flags = [] 
> creation_time = 1545495487 # 2018-12-22 10:18:07 -0600 
> creation_host = "vmh.cyber-range.lan" 
> segment_count = 1 
>
> segment1 { 
> start_extent = 0 
> extent_count = 512000 # 500 Gigabytes 
>
> type = "striped" 
> stripe_count = 1 # linear 
>
> stripes = [ 
> "pv0", 118592 
> ] 
> } 
> } 
> } 
>
> } 
>
>
> # cd /var/log 
> # grep -ri gluster_vg1-lvthinpool-tpool 
> messages-20190922:Sep 15 03:58:15 vmh lvm[14072]: Failed command for 
> gluster_vg1-lvthinpool-tpool. 
> messages:Sep 22 03:44:05 vmh lvm[14072]: Failed command for 
> gluster_vg1-lvthinpool-tpool. 
> messages-20190908:Sep  1 21:27:14 vmh lvm[14062]: Monitoring thin pool 
> gluster_vg1-lvthinpool-tpool. 
> messages-20190908:Sep  1 21:27:24 vmh lvm[14062]: WARNING: Thin pool 
> gluster_vg1-lvthinpool-tpool data is now 100.00% full. 
> messages-20190908:Sep  2 00:19:05 vmh lvm[14072]: Monitoring thin pool 
> gluster_vg1-lvthinpool-tpool. 
> messages-20190908:Sep  2 00:19:15 vmh lvm[14072]: WARNING: Thin pool 
> gluster_vg1-lvthinpool-tpool data is now 100.00% full. 
> messages-20190908:Sep  2 20:16:34 vmh lvm[14072]: Failed command for 
> gluster_vg1-lvthinpool-tpool. 
>
> _______________________________________________
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct: 
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives: 
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/PPTNPPD3SQL3IRYKGUVPZTT3ECPGPNGO/
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/2XJHGLYWK33PNUHXNA2KWYPLB6Z2JHN4/

Reply via email to