[DRBD-user] drbd sycing again from the start
Digimer
lists at alteeve.ca
Thu Oct 17 17:21:00 CEST 2019
8.4.4 is old. Can you upgrade to the latest 8.4.11? I believe 8.4.4 was
older than the other reporters with a similar issue, so this may be
fixed. Upgrading to .11 should not cause any issues.
PS - Please keep replies on the list. These discussions help others by
being in the archives.
digimer
On 2019-10-17 9:54 a.m., Paras pradhan wrote:
> drbd version is drbd-8.4.4-0.27.4.2 and yes are upgrading it to version
> 9 in the near future.
>
> No it is not a live snapshot. Both drbd nodes were shutdown and used
> clonezilla bootable image to take backup and also to restore.
>
> Thanks
> Paras.
>
> On Wed, Oct 16, 2019 at 7:26 PM Digimer <lists at alteeve.ca
> <mailto:lists at alteeve.ca>> wrote:
>
> I don't see the version, but looking in the mailing list archives, the
> common recommendation is to upgrade. What version of DRBD 8 are you
> using, exactly?
>
> Does the resync happen only after recovery? Is the backups of the nodes
> done via live-snapshot? If so, then if there is _any_ time between the
> two nodes being snapped, the UUID will differ and could be causing this.
>
> digimer
>
> On 2019-10-16 10:04 a.m., Paras pradhan wrote:
> > Hi
> >
> > Here is the log for one of the drbd resource (which is 300GB).
> >
> > --
> > [ 194.780377] block drbd1: disk( Diskless -> Attaching )
> > [ 194.780536] block drbd1: max BIO size = 1048576
> > [ 194.780548] block drbd1: drbd_bm_resize called with capacity ==
> 629126328
> > [ 194.783069] block drbd1: resync bitmap: bits=78640791 words=1228763
> > pages=2400
> > [ 194.783077] block drbd1: size = 300 GB (314563164 KB)
> > [ 194.793958] block drbd1: bitmap READ of 2400 pages took 3 jiffies
> > [ 194.796342] block drbd1: recounting of set bits took additional
> 1 jiffies
> > [ 194.796348] block drbd1: 0 KB (0 bits) marked out-of-sync by on
> disk
> > bit-map.
> > [ 194.796359] block drbd1: disk( Attaching -> Outdated )
> > [ 194.796366] block drbd1: attached to UUIDs
> > 56E4CF14A115440C:0000000000000000:02DCAB23D758DA48:02DBAB23D758DA49
> > [ 475.740272] block drbd1: drbd_sync_handshake:
> > [ 475.740280] block drbd1: self
> > 56E4CF14A115440C:0000000000000000:02DCAB23D758DA48:02DBAB23D758DA49
> > bits:0 flags:0
> > [ 475.740288] block drbd1: peer
> > F5A226CE3F2DA2F2:0000000000000000:56E5CF14A115440D:56E4CF14A115440D
> > bits:0 flags:0
> > [ 475.740295] block drbd1: uuid_compare()=-2 by rule 60
> > [ 475.740299] block drbd1: Writing the whole bitmap, full sync
> required
> > after drbd_sync_handshake.
> > [ 475.757877] block drbd1: bitmap WRITE of 2400 pages took 4 jiffies
> > [ 475.757888] block drbd1: 300 GB (78640791 bits) marked
> out-of-sync by
> > on disk bit-map.
> > [ 475.758018] block drbd1: peer( Unknown -> Secondary ) conn(
> > WFReportParams -> WFBitMapT ) pdsk( DUnknown -> UpToDate )
> > [ 475.800134] block drbd1: receive bitmap stats [Bytes(packets)]:
> plain
> > 0(0), RLE 23(1), total 23; compression: 100.0%
> > [ 475.802697] block drbd1: send bitmap stats [Bytes(packets)]: plain
> > 0(0), RLE 23(1), total 23; compression: 100.0%
> > [ 475.802717] block drbd1: conn( WFBitMapT -> WFSyncUUID )
> > [ 475.815155] block drbd1: updated sync uuid
> > CEF5B26573C154CC:0000000000000000:02DCAB23D758DA48:02DBAB23D758DA49
> > [ 475.815377] block drbd1: helper command: /sbin/drbdadm
> > before-resync-target minor-1
> > [ 475.820270] block drbd1: helper command: /sbin/drbdadm
> > before-resync-target minor-1 exit code 0 (0x0)
> > [ 475.820293] block drbd1: conn( WFSyncUUID -> SyncTarget ) disk(
> > Outdated -> Inconsistent )
> > [ 475.820306] block drbd1: Began resync as SyncTarget (will sync
> > 314563164 KB [78640791 bits set]).
> > [ 538.518371] block drbd1: peer( Secondary -> Primary )
> > [ 538.548954] block drbd1: role( Secondary -> Primary )
> > [ 2201.521232] block drbd1: conn( SyncTarget -> PausedSyncT )
> user_isp(
> > 0 -> 1 )
> > [ 2201.521237] block drbd1: Resync suspended
> > [ 2301.930484] block drbd1: conn( PausedSyncT -> SyncTarget )
> user_isp(
> > 1 -> 0 )
> > [ 2301.930490] block drbd1: Syncer continues.
> > [ 5216.750314] block drbd1: Resync done (total 4740 sec; paused
> 100 sec;
> > 67792 K/sec)
> > [ 5216.750323] block drbd1: 98 % had equal checksums, eliminated:
> > 311395164K; transferred 3168000K total 314563164K
> > [ 5216.750333] block drbd1: updated UUIDs
> > F5A226CE3F2DA2F3:0000000000000000:CEF5B26573C154CD:56E5CF14A115440D
> > [ 5216.750343] block drbd1: conn( SyncTarget -> Connected ) disk(
> > Inconsistent -> UpToDate )
> > [ 5216.750518] block drbd1: helper command: /sbin/drbdadm
> > after-resync-target minor-1
> > [ 5216.845211] block drbd1: helper command: /sbin/drbdadm
> > after-resync-target minor-1 exit code 0 (0x0)
> > ---
> >
> >
> > Thanks!
> >
> > On Wed, Oct 16, 2019 at 12:46 AM Digimer <lists at alteeve.ca
> <mailto:lists at alteeve.ca>
> > <mailto:lists at alteeve.ca <mailto:lists at alteeve.ca>>> wrote:
> >
> > On 2019-10-15 4:58 p.m., Paras pradhan wrote:
> > > Hi
> > >
> > > I have a two node drbd 8 cluster. We are doing some test and
> while
> > drbd
> > > resources are consistent/synced on both nodes, I powered off
> both
> > nodes
> > > and took a bare metal backup using clonezilla.
> > >
> > > Then I restored the both nodes using the backup and started drbd
> > on both
> > > nodes. On one of the nodes it starts to sync all over again.
> > >
> > > My question is: when I took the backup drbd resources are synced
> > and why
> > > it is starting all over again? I hope I explained clearly.
> > >
> > > Thanks in advance !
> > > Paras.
> >
> > Do you have the system logs from when you started DRBD on the
> nodes
> > post-recovery? There should be DRBD log entries on both nodes
> as DRBD
> > started. The reason/trigger of the resync will likely be
> explained in
> > there. If not, please share the logs.
> >
> > --
> > Digimer
> > Papers and Projects: https://alteeve.com/w/
> > "I am, somehow, less interested in the weight and convolutions of
> > Einstein’s brain than in the near certainty that people of
> equal talent
> > have lived and died in cotton fields and sweatshops." -
> Stephen Jay
> > Gould
> >
>
>
> --
> Digimer
> Papers and Projects: https://alteeve.com/w/
> "I am, somehow, less interested in the weight and convolutions of
> Einstein’s brain than in the near certainty that people of equal talent
> have lived and died in cotton fields and sweatshops." - Stephen Jay
> Gould
>
--
Digimer
Papers and Projects: https://alteeve.com/w/
"I am, somehow, less interested in the weight and convolutions of
Einstein’s brain than in the near certainty that people of equal talent
have lived and died in cotton fields and sweatshops." - Stephen Jay Gould
More information about the drbd-user
mailing list