https://bugs.freebsd.org/bugzilla/show_bug.cgi?id=230910

--- Comment #5 from Volodymyr Kostyrko <arc...@b1t.name> ---
Looks like I partially know whats going on here. First when I tried to
partition drives they were just freebsd-boot, freebsd-swap and freebsd-zfs. But
when that failed to work I sliced out 1G out of last one into other partition:

# gpart show da0
=>       40  781422688  da0  GPT  (373G)
         40        256    1  freebsd-boot  (128K)
        296   67108864    2  freebsd-swap  (32G)
   67109160    2097152    3  freebsd-ufs  (1.0G)
   69206312  712216408    4  freebsd-zfs  (340G)
  781422720          8       - free -  (4.0K)

arcade@cis\/home/arcade# zpool status mycon
  pool: mycon
 state: ONLINE
  scan: scrub repaired 0 in 0h33m with 0 errors on Mon Aug 13 04:45:19 2018
config:

        NAME        STATE     READ WRITE CKSUM
        mycon       ONLINE       0     0     0
          raidz1-0  ONLINE       0     0     0
            da0p4   ONLINE       0     0     0
            da1p4   ONLINE       0     0     0
            da2p4   ONLINE       0     0     0

errors: No known data errors

Here what it shows during "good" boot:

Reconstructed root pool config:
version: 5000
name: mycon
state: 0
txg: 19961181
pool_guid: 14318237666280658641
hostid: 2404943215
hostname: cis.mix101.info
vdev_children: 1
features_for_read:
        com.delphix:hole_birth: false
        com.delphix:embedded_data: false

vdev_tree:
        type: root
        id: 0
        guid: 14318237666280658641
        children:
        children[0]:
                type: raidz
                id: 0
                guid: 17901688662798610243
                nparity: 1
                metaslab_array: 39
                metaslab_shift: 33
                ashift: 12
                asize: 1093949718528
                is_log: 0
                create_txg: 4
                children:
                children[0]:
                        type: disk
                        id: 0
                        guid: 3986367950682223964
                        path: /dev/da0p4
                        whole_disk: 1
                        DTL: 177
                        create_txg: 4
                children[1]:
                        type: disk
                        id: 1
                        guid: 9245294472961478945
                        path: /dev/da1p4
                        whole_disk: 1
                        DTL: 176
                        create_txg: 4
                children[2]:
                        type: disk
                        id: 2
                        guid: 5056913690226738568
                        path: /dev/da2p4
                        whole_disk: 1
                        DTL: 175
                        create_txg: 4


This means that ZFS finds correct pool marks on partition already grabbed by
geom_mirror and fails when trying to construct pool because kernel would not
allow rw access there.

I'll go zero out my ufs partitions...

-- 
You are receiving this mail because:
You are the assignee for the bug.
_______________________________________________
freebsd-geom@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-geom
To unsubscribe, send any mail to "freebsd-geom-unsubscr...@freebsd.org"

Reply via email to