On 01/02/17 19:30, lejeczek wrote:
On 01/02/17 14:44, Atin Mukherjee wrote:
I think you have hit
https://bugzilla.redhat.com/show_bug.cgi?id=1406411 which
has been fixed in mainline and will be available in
release-3.10 which is slated for next month.
To prove you have hit the same problem can you please
confirm the following:
1. Which Gluster version are you running?
2. Was any of the existing brick down?
2. Did you mounted the volume? If not you have two ways
(1) bring up the brick and restart glusterd followed by
add-brick or (2) if the existing brick(s) is bad for some
reason, restarting glusterd and mounting the volume
followed by a look up and then attempting add-brick
should succeed.
a chance to properly investigate it has been lost I think.
I all started with one peer I missed was not migrated from
3.7 to 3.8 and unfortunately it was a system I could not
tamper with until late evening, which is now.
This problem though occurred after I already upgraded that
gluster to 3.8. I even removed that failing node's bricks
and detached it, re-attached it and still, those errors I
described earlier... until now when I restarted that one
last one peer... now all seems ok, well, at least I don't
see those errors any more.
Should I now be looking at something particular more closely?
b.w.
L.
I realized I might have not been clear there, that box
where I missed that upgrade 3.7->3.8 was not the one where
problem existed, it was a different box, if that
changes/helps anything.
On Wed, Feb 1, 2017 at 7:49 PM, lejeczek
<pelj...@yahoo.co.uk <mailto:pelj...@yahoo.co.uk>> wrote:
hi,
I have a four peers gluster and one is failing, well,
kind of..
If on a working peer I do:
$ gluster volume add-brick QEMU-VMs replica 3
10.5.6.49:/__.aLocalStorages/0/0-GLUSTERs/0GLUSTER-QEMU-VMs
force
volume add-brick: failed: Commit failed on whale.priv
Please check log file for details.
but:
$ gluster vol info QEMU-VMs
Volume Name: QEMU-VMs
Type: Replicate
Volume ID: 8709782a-daa5-4434-a816-c4e0aef8fef2
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1:
10.5.6.100:/__.aLocalStorages/1/0-GLUSTERs/1GLUSTER-QEMU-VMs
Brick2:
10.5.6.17:/__.aLocalStorages/1/0-GLUSTERs/QEMU-VMs
Brick3:
10.5.6.49:/__.aLocalStorages/0/0-GLUSTERs/0GLUSTER-QEMU-VMs
# <= so it is here, also this command on that failing
peers reports correctly.
Interestingly,
$ gluster volume remove-brick
removes no errors, but this change is not propagated
to the failing peer. Vol info still reports its brick
is part of the volume.
And the failing completely part: every command on
failing peer reports:
$ gluster volume remove-brick QEMU-VMs replica 2
10.5.6.49:/__.aLocalStorages/0/0-GLUSTERs/0GLUSTER-QEMU-VMs
force
Removing brick(s) can result in data loss. Do you
want to Continue? (y/n) y
volume remove-brick commit force: failed: Commit
failed on 10.5.6.32. Please check log file for details.
Commit failed on rider.priv Please check log file for
details.
Commit failed on 10.5.6.17. Please check log file for
details.
I've been watching logs but honestly, don't know
which one(s) I should paste in here.
b.w.
L.
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
<mailto:Gluster-users@gluster.org>
http://lists.gluster.org/mailman/listinfo/gluster-users
<http://lists.gluster.org/mailman/listinfo/gluster-users>
--
~ Atin (atinm)
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users