[DRBD-user] “The peer's disk size is too small!” messages on attempts to add rebuilt pee

Anthony G. agenerette at hotmail.com
Fri Dec 21 18:39:48 CET 2012

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.





> well, you could try the one I put in my previous answer ... and it does
> not need to be of the exact size on nfs1 ... equal or more
> 
I will try that.  It's probably apparent, but I'm new to LVM and DRBD.  Is the"drbdadm adjust nfs" on nfs2 something that I can do while that system isup-and-running and servicing Production requests?
Thanks, again,
-Anthony
> Date: Fri, 21 Dec 2012 18:12:23 +0100
> From: andreas at hastexo.com
> To: drbd-user at lists.linbit.com
> CC: agenerette at hotmail.com
> Subject: Re: [DRBD-user] “The peer's disk size is too small!” messages on attempts to add rebuilt pee
> 
> 
> Please don't bypass the mailing-list ...
> 
> On 12/21/2012 06:04 PM, Anthony G. wrote:
> > Thank you for your input.  That was my first thought, but I caught hell
> > trying
> > to get the partition sizes to match.  I'm not sure which size reading I
> > need to 
> > take on -nfs2 and then which specific lvcreate command I need to execute on 
> > -nfs1 to get the size on the latter set properly.
> 
> well, you could try the one I put in my previous answer ... and it does
> not need to be of the exact size on nfs1 ... equal or more
> 
> > 
> > I've recreated the lv, though (just to try and make some progress), and
> > am now 
> > getting the following, when I try to 'service drbd start' on -nfs1:
> > 
> > DRBD's startup script waits for the peer node(s) to appear.
> >  - In case this node was already a degraded cluster before the
> >    reboot the timeout is 0 seconds. [degr-wfc-timeout]
> >  - If the peer was available before the reboot the timeout will
> >    expire after 0 seconds. [wfc-timeout]
> >    (These values are for resource 'nfs'; 0 sec -> wait forever)
> >  To abort waiting enter 'yes' [ 123]:yes
> > 
> > 'netstat -a' doesn't show -nfs2 listening on port 7789, but I do see
> > drbd-related
> > processes running on that box.  
> 
> so the resource on nfs2 is in disconnected state .... do a "drbdadm
> adjust nfs" on nfs2
> 
> Regards,
> Andreas
> 
> > 
> > -Anthony
> > 
> > Date: Fri, 21 Dec 2012 17:25:01 +0100
> > From: andreas at hastexo.com
> > To: drbd-user at lists.linbit.com
> > Subject: Re: [DRBD-user] “The peer's disk size is too small!” messages
> > on attempts to add rebuilt pee
> > 
> > On 12/21/2012 12:13 AM, Anthony G. wrote:
> >> Hi,
> >> 
> >> There's so much information relating to my current configuration, that
> >> I'm not sure what I should post here.  Let me start by saying that I had
> >> two Ubuntu 10.04 hosts configured in a DRBD relationship:  sf02-nfs1
> >> (primary) and sf0-nfs2 (secondary).  -nfs1 suffered a major filesystem
> >> fault.  I had to make -nfs2 primary and rebuild -nfs1.  I want to
> >> eventually have all of my machines on 12.04, so I took this as an
> >> opportunity to set -nfs1 on that OS.
> >> 
> >> Here is a copy of my main configuration file (/etc/drbd.d/nfs.res):
> >> 
> >> resource nfs {
> >>   on sf02-nfs2 {
> >>     device    /dev/drbd0;
> >>     disk      /dev/ubuntu/drbd-nfs;
> >>     address   10.0.6.2:7789;
> >>     meta-disk internal;
> >>   }
> >>   on sf02-nfs1 {
> >>     device    /dev/drbd0;
> >>     disk      /dev/ubuntuvg/drbd-nfs;
> >>     address   10.0.6.1:7789;
> >>     meta-disk internal;
> >>   }
> >> }
> >> 
> >> 
> >> I'm trying to re-introduce -nfs1 into the DRBD relationship and am
> >> having trouble.  I have:
> >> 
> >> 
> >> 1.) created the resource "nfs" on -nfs1 ('drbdadm create-md nfs')
> >> 
> >> 2.) run 'drbdadm primary nfs' on -nfs2 and 'drbdadm secondary nfs' on -nfs1.
> >> 
> >> 3.) run drbdadm -- --overwrite-data-of-peer primary all' from -nfs2.
> >> 
> >> 
> >> But /var/log/kern.log shows:
> >> 
> >> =====
> >> 
> >> Dec 19 19:55:47 sf02-nfs2 kernel: [9284165.843938] block drbd0:
> >> Handshake successful: Agreed network protocol version 91
> >> 
> >> Dec 19 19:55:47 sf02-nfs2 kernel: [9284165.843949] block drbd0: conn(
> >> WFConnection -> WFReportParams )
> >> 
> >> Dec 19 19:55:47 sf02-nfs2 kernel: [9284165.844171] block drbd0: Starting
> >> asender thread (from drbd0_receiver [2452])
> >> 
> >> Dec 19 19:55:47 sf02-nfs2 kernel: [9284165.844539] block drbd0:
> >> data-integrity-alg: <not-used>
> >> 
> >> Dec 19 19:55:47 sf02-nfs2 kernel: [9284165.844610] block drbd0: *The
> >> peer's disk size is too small!*
> >> 
> >> Dec 19 19:55:47 sf02-nfs2 kernel: [9284165.844617] block drbd0: conn(
> >> WFReportParams -> Disconnecting )
> >> 
> >> Dec 19 19:55:47 sf02-nfs2 kernel: [9284165.844626] block drbd0: error
> >> receiving ReportSizes, l: 32!
> >> 
> >> Dec 19 19:55:47 sf02-nfs2 kernel: [9284165.844680] block drbd0: asender
> >> terminated
> >> 
> >> Dec 19 19:55:47 sf02-nfs2 kernel: [9284165.844691] block drbd0:
> >> Terminating asender thread
> >> 
> >> Dec 19 19:55:47 sf02-nfs2 kernel: [9284165.844746] block drbd0:
> >> Connection closed
> >> 
> >> Dec 19 19:55:47 sf02-nfs2 kernel: [9284165.844755] block drbd0: conn(
> >> Disconnecting -> StandAlone )
> >> 
> >> Dec 19 19:55:47 sf02-nfs2 kernel: [9284165.844791] block drbd0: receiver
> >> terminated
> >> 
> >> Dec 19 19:55:47 sf02-nfs2 kernel: [9284165.844794] block drbd0:
> >> Terminating receiver thread
> >> 
> >> =====
> >> 
> >> 
> >> So, it seems that a difference in the size of drbd0 on the respective
> >> machines is the source of my trouble.  'cat /proc/partitions' (output
> >> pasted at the end of this message) on each machine tells me that -nfs2's
> >> partition is around 348148 blocks larger than -nfs1's.  -nfs2 contains
> >> my company's Production data, so I do not, of course, want to do
> >> anything destructive there.  I can, however, certainly recreate the
> >> resource on -nfs1.  
> >> 
> >> 
> >> Does anyone out there know what steps I need to take to make the
> >> partition sizes match?  Of course, I'm working under the belief that the
> >> "peer's disk size is too small" message points up the source of my
> >> trouble.  Let me know, of course, if I need to post more information on
> >> my setup.
> >  
> > You are using LVM, so simply resize the lv below DRBD on nfs1 to be at
> > least of the same size or bigger ala:
> >  
> > lvresize -L+200M ubuntuvg/drbd-nfs
> >  
> > ... then recreate meta-data on that resized lv on nfs1 and on nfs1 do a:
> >  
> > drbdadm up nfs
> >  
> >  
> > Regards,
> > Andreas
> >  
> > -- 
> > Need help with DRBD?
> > http://www.hastexo.com/now
> >  
> >> 
> >> 
> >> Thanks,
> >> 
> >> 
> >> -Anthony
> >> 
> >> 
> >> 
> >> 
> >> 
> >> 
> >> ==========
> >> 
> >> root at sf02-nfs1:/dev/ubuntuvg# cat /proc/partitions 
> >> 
> >> major minor  #blocks  name
> >> 
> >> 
> >>    8       0 1952448512 sda
> >> 
> >>    8       1     512000 sda1
> >> 
> >>    8       2          1 sda2
> >> 
> >>    8        5 1886388224sda5
> >> 
> >>  252        0   20971520 dm-0
> >> 
> >>  252        1    5242880 dm-1
> >> 
> >>  252       2 1706033152 dm-2
> >> 
> >>  147        0 1705981052 drbd0
> >> 
> >> 
> >> 
> >> root at sf02-nfs2:/etc/drbd.d# cat /proc/partitions 
> >> 
> >> major minor  #blocks  name
> >> 
> >> 
> >>    8        0 1952448512 sda
> >> 
> >>    8        1     248832 sda1
> >> 
> >>    8        2          1 sda2
> >> 
> >>    8        5 1952196608 sda5
> >> 
> >>  252        0  209715200 dm-0ubuntuvg-root
> >> 
> >>  252        1  36098048 dm-1ubuntuvg-swap
> >> 
> >>  252        2 1706381312 dm-2ubuntuvg-drbd--nfs
> >> 
> >>  147        0 1706329200 drbd0
> >> 
> >> 
> >> 
> >> 
> >> _______________________________________________
> >> drbd-user mailing list
> >> drbd-user at lists.linbit.com
> >> http://lists.linbit.com/mailman/listinfo/drbd-user
> >> 
> >  
> >  
> >  
> > 
> > 
> > _______________________________________________ drbd-user mailing list
> > drbd-user at lists.linbit.com
> > http://lists.linbit.com/mailman/listinfo/drbd-user
> 
> 

 		 	   		  
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20121221/538878ab/attachment.htm>


More information about the drbd-user mailing list