Hi,

I'm trying to build a new cluster of servers using debian plus DRBD9 and drbdmanage.

After a number of attempts, I thought I had everything right, and it's all been "ok" for a couple of weeks.

Today, I rebooted both machines (new power being installed for the UPS), and then I tried to create a new volume of 700GB.

Here is what I did:
san2:~# drbdmanage add-volume oldNFS 700GB --deploy 2
Operation completed successfully
Operation completed successfully
san2:~# drbdmanage list-nodes
+------------------------------------------------------------------------------------------------------------+
| Name | Pool Size | Pool Free | | State |
|------------------------------------------------------------------------------------------------------------|
| san2.websitemanagers.com.au | 3777040 | 3109316 | | ok | | san3 | 1830932 | 1830924 | | ok |
+------------------------------------------------------------------------------------------------------------+
san2:~# drbdmanage list-volumes --show Port
+------------------------------------------------------------------------------------------------------------+
| Name | Vol ID | Size | Minor | Port | | State |
|------------------------------------------------------------------------------------------------------------|
| oldNFS | 0 | 667572 | 101 | 7001 | | ok | | test1 | 0 | 9536 | 100 | 7000 | | ok |
+------------------------------------------------------------------------------------------------------------+
san2:~# lvs
LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert
  .drbdctrl_0 drbdpool -wi-ao---- 4.00m
  .drbdctrl_1 drbdpool -wi-ao---- 4.00m
  oldNFS_00   drbdpool -wi-ao---- 652.07g
san2:~# dpkg -l | grep drbd
ii drbd-utils 8.9.6-1 amd64 RAID 1 over TCP/IP for Linux (user utilities) ii python-drbdmanage 0.94-1 all DRBD distributed resource management utility
san2:~# cat /proc/drbd
version: 9.0.1-1 (api:2/proto:86-111)
GIT-hash: f57acfc22d29a95697e683fb6bbacd9a1ad4348e build by r...@san2.websitemanagers.com.au, 2016-03-01 00:38:53
Transports (api:14): tcp (1.0.0)

*So far, everything looks good, so I thought to check out the other node, and see what is happening there....*

root@san3:~# lvs
LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert
  .drbdctrl_0                   drbdpool  -wi-ao---- 4.00m
  .drbdctrl_1                   drbdpool  -wi-ao---- 4.00m
  backup_system_20141006_193935 san1small -wi-a----- 8.00g
  swap                          san1small -wi-ao---- 3.72g
  system                        san1small -wi-ao---- 13.97g

*Hmmm, that's strange, we don't have any new LV **here?*

root@san3:~# drbdmanage list-nodes
+------------------------------------------------------------------------------------------------------------+
| Name | Pool Size | Pool Free | | State |
|------------------------------------------------------------------------------------------------------------|
| san2.websitemanagers.com.au | 3777040 | 3109316 | | ok | | san3 | 1830932 | 1830924 | | ok |
+------------------------------------------------------------------------------------------------------------+
root@san3:~# drbdmanage list-volumes --show Port
+------------------------------------------------------------------------------------------------------------+
| Name | Vol ID | Size | Minor | Port | | State |
|------------------------------------------------------------------------------------------------------------|
| oldNFS | 0 | 667572 | 101 | 7001 | | ok | | test1 | 0 | 9536 | 100 | 7000 | | ok |
+------------------------------------------------------------------------------------------------------------+
root@san3:~# dpkg -l | grep drbd
ii drbd-utils 8.9.6-1 amd64 RAID 1 over TCP/IP for Linux (user utilities) ii python-drbdmanage 0.94-1 all DRBD distributed resource management utility
root@san3:~# cat /proc/drbd
version: 9.0.1-1 (api:2/proto:86-111)
GIT-hash: f57acfc22d29a95697e683fb6bbacd9a1ad4348e build by root@san1, 2016-03-01 00:38:33
Transports (api:14): tcp (1.0.0)

Reading more docs, I then find this section:
http://www.drbd.org/doc/users-guide-90/s-check-status
san2:~# drbd-overview
  0:.drbdctrl/0  Connected(2*) Secondary(2*) UpToDa/UpToDa
  1:.drbdctrl/1  Connected(2*) Secondary(2*) UpToDa/UpToDa
101:oldNFS/0     Connec/C'ting Second/Unknow UpToDa/DUnkno

root@san3:/etc/drbd.d# drbd-overview
 0:.drbdctrl/0  Connected(2*) Secondary(2*) UpToDa/UpToDa
 1:.drbdctrl/1  Connected(2*) Secondary(2*) UpToDa/UpToDa

So, it would seem that the problem is the config hasn't been sent to the other node, and it just doesn't know anything about it.....

san2:~# drbdadm status oldNFS --verbose
drbdsetup status oldNFS
oldNFS role:Secondary
  disk:UpToDate
  san3 connection:Connecting

Can anyone help advise where I should look, or what I might need to do to get this working?

Thanks,
Adam

_______________________________________________
drbd-user mailing list
drbd-user@lists.linbit.com
http://lists.linbit.com/mailman/listinfo/drbd-user

Reply via email to