Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Hello all, Thought I would provide an update as someone may find this useful. So I was able to resolve my situation. Here is what I did: - Changed drbd config to use the raw [/dev/sdc] device on the secondary node - Wiped out /dev/sdc on the secondary node [dd if=/dev/zero of=/dev/sdc bs=1M count=10] - Recreated the internal metadata [LVM_SYSTEM_DIR= drbdadm create-md vmkfs_vol1] - Brought up the resource and did a full sync - Failed-over the cluster and repeated the above steps With kind regards, -Peter Brunnengräber ----- Original Message ----- From: "Peter Brunnengraeber" <pbrunnen at bccglobal.com> To: drbd-user at lists.linbit.com Sent: Sunday, August 21, 2016 3:40:47 PM Subject: DRBD resource expansion trouble Hello, I have been having an issue attempting to expand a DRBD storage resource. Our system is hardware raid (lsi megaraid) -> drbd (internal metadata) -> lvm storage The storage software we use requires control of LVM and we backed it with DRBD, thus the strange setup with LVM ontop of DRBD. - We added disks and extended the raid storage backend - Issued 'echo 1 >/sys/class/scsi_device/1\:2\:2\:0/device/rescan' - Kernel and 'parted /dev/sdc print free' sees the new size and free space We attempted to do an online expansion based on the notes in a previous posting: Based on> http://lists.linbit.com/pipermail/drbd-user/2014-February/020663.html - Attempted 'drbdmeta /dev/drbd1 v08 /dev/sdc1 internal check-resize' but that did not do anything that we could tell - 'pvdisplay /dev/sdc' still shows the old size; 'parted /dev/sdc print free' also still shows the old size and free space - Attempted 'drbdadm -v resize vmkfs_vol1', also does nothing. No error, but no additional space added to drbd So we attempted to fall back to the drbd documented offline way: Based on> https://www.drbd.org/en/doc/users-guide-83/s-resizing - Took down the drbd resource and dumped the metadata: 'drbdadm dump-md vmkfs_vol1 >/tmp/metadata' - Edited /tmp/metadata to change "la-size-sect = DevSizeInSect - ( (DevSizeInSect / 32768 * 8) + 72 )" - Used 'LVM_SYSTEM_DIR= drbdadm create-md vmkfs_vol1' to create the new metadata and circumventing the LVM detection problem for the LVM contained within the DRBD resource. - Reimported the metadata to the drbd resource - Attempted to bring up the drbd resouce, but errors "Low.dev. smaller than requested DRBD-dev. size." Looking at parted's output, we see the free space still. In parted I tried to: - Used sector mode - Resize, but fails with "unknown filesystem type" - Remove and recreate fails with "closest location we can manage" message which only provided the option to recreate it up to a few thousand sectors after the original sector size. At this point, I assume that parted won't let me create the larger partition because it see something there like the drbd metadata... I guess this is a parted bug/limitation more than anything. The disk is GPT table because of the size, so fdisk was off the table. As a last resort, I assume that if I was to low level wipe the data from the RAID on one node that I could recreate the partition and metadata from scratch. I believe that I can do a full resync from a smaller to larger resource, failover to the system with the expanded storage, and then wipe/resync/repeat back to the other node. Does this last resort make sense and/or has anyone else run across this or have a way to move forward? Also, can I just back DRBD with the device directly and no partition (/dev/sdc instead of /dev/sdc1)? I'm on drbd v8.3.11 Very much appreciated! -With kind regards, Peter Brunnengräber