On 09/20/2016 03:02 AM, Anthony DeRobertis wrote:
On 09/20/2016 01:38 AM, Guoqing Jiang wrote:

Thanks for report, could you try the latest tree git://git.kernel.org/pub/scm/utils/mdadm/mdadm.git? I guess 45a87c2f31335a759190dff663a881bc78ca5443 should resolve it , and I can add a spare disk to native raid (internal bitmap) with different metadatas (0.9, 1.0 to 1.2).

(please keep me cc'd, I'm not subscribed)

$ git rev-parse --short HEAD
676e87a
$ make -j4
...

# ./mdadm -a /dev/md/pv0 /dev/sdc3
mdadm: add new device failed for /dev/sdc3 as 8: Invalid argument

[375036.613907] md: sdc3 does not have a valid v1.0 superblock, not importing!
[375036.613926] md: md_import_device returned -22

The md-cluster code should only work raid1 with 1.2 metadata, and your array (/dev/md/pv0) is raid10 with 1.0 metadata (if I read bug correctly), so it is weird that your array can invoke the code for md-cluster.

I assume it only happens with existed array, a new created one doesn't have the problem, right? And I can't
reproduce it from my side.

Which kernel version are you used to created the array in case the kernel was updated? Also pls show the output of "mdadm -X $DISK", and your bitmap is a little weird (but I don't try with 10 level before, so maybe
it is correct).

Internal Bitmap : -234 sectors from superblock


Thanks,
Guoqing

Reply via email to