Hi guys,
thanks for trying to help.
Here's the whole story: The two 1 TB disks originally were in a Synology
NAS unit, not configured (to my knowledge) as any kind of RAID. The second
disk was only physically installed in the NAS after the first one had been
filling up. The NAS GUI showed the disks as two separate volume groups, vg1
and vg1000. I wanted to upgrade the disks to bigger size, so I took them
out of the NAS and connected them directly to my Linux PC (using a USB
docking station), wanting to copy the data to a couple of new disks
installed in the NAS. I then went through a very steep (and incomplete)
learning curve about mdadm and lvm and finally managed to mount the
volumes. I could see and read all my files and stuff.
Because I wasn't yet ready to initiate the big data transfer, I unmounted
the LVM disks, stopped the md stuff, and turned off the computer. When I
turned it back on, I couldn't get the LVM working again. All that happens
is that a tiny device "/dev/vg1/syno_vg_reserved_area" is created.
>From my first trial I remember that when I read data from one of the
volumes, only the corresponding LED from one drive of the USB dock flashed,
from which I conclude that I definetely don't have any kind of "real" RAID
on those disks. Note however the output of vgchange (below) which mentions
something about a degraded RAID.
I tried re-installing the disks into the NAS. One of the disks is now
reported as "Crashed", and nothing is mounted. As far as I can see all LVM
metadata is intact on the disks. I also think that the data must be largely
intact because so far I only did reading operations on the disks (but
didn't mount them read-only the first time round).
Fortunately the data on the disks is not really essential, but a huge
amount of work archiving a large CD and DVD collection... is there any way
to rescure this stuff?
Here's what I tried on the Linux box so far:
root@dotcom:/home/dh# mdadm --assemble --scan
mdadm: /dev/md/3 has been started with 1 drive.
mdadm: /dev/md/2 has been started with 1 drive.
root@dotcom:/home/dh# lvmdiskscan
/dev/loop0 [ 213.51 MiB]
/dev/loop1 [ 208.05 MiB]
/dev/sda1 [ 30.00 GiB]
/dev/loop2 [ 220.89 MiB]
/dev/md2 [ 926.91 GiB] LVM physical volume
/dev/loop3 [ 206.55 MiB]
/dev/md3 [ 926.91 GiB] LVM physical volume
/dev/sda5 [ 2.00 GiB]
/dev/sda6 [ 30.00 GiB]
/dev/sda7 [ 200.00 GiB]
/dev/sda8 [ 203.76 GiB]
/dev/sdb1 [ 465.76 GiB]
/dev/sdc1 [ 149.05 GiB]
/dev/sdd1 [ 2.37 GiB]
/dev/sdd2 [ 2.00 GiB]
/dev/sde1 [ 2.37 GiB]
/dev/sde2 [ 2.00 GiB]
0 disks
15 partitions
0 LVM physical volume whole disks
2 LVM physical volumes
root@dotcom:/home/dh# vgdisplay
--- Volume group ---
VG Name vg1000
System ID
Format lvm2
Metadata Areas 1
Metadata Sequence No 3
VG Access read/write
VG Status resizable
MAX LV 0
Cur LV 0
Open LV 0
Max PV 0
Cur PV 1
Act PV 1
VG Size 926.90 GiB
PE Size 4.00 MiB
Total PE 237287
Alloc PE / Size 0 / 0
Free PE / Size 237287 / 926.90 GiB
VG UUID 9PQXmK-0dqN-3I11-1CbR-3ND3-okUD-WcySR3
--- Volume group ---
VG Name vg1
System ID
Format lvm2
Metadata Areas 1
Metadata Sequence No 6
VG Access read/write
VG Status resizable
MAX LV 0
Cur LV 1
Open LV 0
Max PV 0
Cur PV 1
Act PV 1
VG Size 926.91 GiB
PE Size 4.00 MiB
Total PE 237290
Alloc PE / Size 3 / 12.00 MiB
Free PE / Size 237287 / 926.90 GiB
VG UUID srS6Ku-i7PP-9xoZ-6kFf-IWiG-Uu9J-DcicaM
root@dotcom:~# vgchange -v -ay vg1000
DEGRADED MODE. Incomplete RAID LVs will be processed.
Using volume group(s) on command line
Finding volume group "vg1000"
0 logical volume(s) in volume group "vg1000" now active
root@dotcom:~# vgchange -v -ay vg1
DEGRADED MODE. Incomplete RAID LVs will be processed.
Using volume group(s) on command line
Finding volume group "vg1"
Activating logical volume "syno_vg_reserved_area".
activation/volume_list configuration setting not defined: Checking only
host tags for vg1/syno_vg_reserved_area
Creating vg1-syno_vg_reserved_area
Loading vg1-syno_vg_reserved_area table (254:0)
Resuming vg1-syno_vg_reserved_area (254:0)
Activated 1 logical volumes in volume group vg1
1 logical volume(s) in volume group "vg1" now active
root@dotcom:~# lvdisplay
--- Logical volume ---
LV Path /dev/vg1/syno_vg_reserved_area
LV Name syno_vg_reserved_area
VG Name vg1
LV UUID ZvVm6T-FDWz-zaF4-7M4P-20i5-gXyB-4M56Hj
LV Write Access read/write
LV Creation host, time ,
LV Status available
# open 0
LV Size 12.00 MiB
Current LE 3
Segments 1
Allocation inherit
Read ahead sectors auto
- currently set to 256
Block device 254:0
root@dotcom:~# vgs
VG #PV #LV #SN Attr VSize VFree
vg1 1 1 0 wz--n- 926.91g 926.90g
vg1000 1 0 0 wz--n- 926.90g 926.90g
On Thu, Nov 24, 2016 at 10:57 PM, Roland Müller <
[email protected]> wrote:
> Hello,
>
>
> On 11/24/2016 08:35 PM, Robert Latest wrote:
>
>> Hey all,
>>
>> I got it to work ONCE, but for the life of me I can't figure out how to
>> do it again.
>>
>> This is what I think I did the first time, but for the second time it
>> just doesn't work.
>>
>> root@dotcom:~# mdadm --assemble --scan
>> mdadm: /dev/md/2 has been started with 1 drive.
>> mdadm: /dev/md/3 has been started with 1 drive.
>>
>> ---OK, good so far. Now let's find the LVs
>>
>> root@dotcom:~# lvmdiskscan
>> /dev/loop0 [ 213.51 MiB]
>> /dev/loop1 [ 206.55 MiB]
>> /dev/sda1 [ 30.00 GiB]
>> /dev/loop2 [ 220.89 MiB]
>> /dev/md2 [ 926.91 GiB] LVM physical volume
>> /dev/loop3 [ 208.05 MiB]
>> /dev/md3 [ 926.91 GiB] LVM physical volume
>> /dev/sda5 [ 2.00 GiB]
>> /dev/sda6 [ 30.00 GiB]
>> /dev/sda7 [ 200.00 GiB]
>> /dev/sda8 [ 203.76 GiB]
>> /dev/sdb1 [ 2.37 GiB]
>> /dev/sdb2 [ 2.00 GiB]
>> /dev/sdc1 [ 2.37 GiB]
>> /dev/sdc2 [ 2.00 GiB]
>> /dev/sdd1 [ 465.76 GiB]
>> /dev/sde1 [ 149.05 GiB]
>> 0 disks
>> 15 partitions
>> 0 LVM physical volume whole disks
>> 2 LVM physical volumes
>>
>> ---Still looking good. Now I'm supposed to find the logical volumes,
>> ---but lvdisplay simply doesn't show anything.
>>
>> root@dotcom:~# lvdisplay
>> root@dotcom:~#
>>
>> ---Now I'm stuck. All LVM instructions I find on the Internet say that I
>> find the path of the LVM device by using lvdisplay. Also I know that one
>> hour ago I had my volumes mounted and was copying data from them. After
>> properly syncing and unmounting them, and stopping the LVM and md thingys,
>> I'm stuck now.
>>
>> Any suggestions?
>>
>> robert
>>
>
> I have a problem to understand what is actually your issue with LVM. Was
> your system working before and now logical volumes that were existing
> before disappeared?
>
> What is the situation with volume group or groups? What does vgscan or vgs
> commands say?
>
> BR,
>
> Roland
>
>