Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Ok I didn't realise this was that kind of an issue, I just assumed DRBD wasn't happy with my setup due to the second node not being online. The partition /dev/sdb1 is only set to about 40GB while I'm testing the setup, once I've nutted everything the machine will be rebuilt to have it's full capacity. (around 7TB) it's an LVM partition I created a single Logical Volume around 20GB have it exported via iscsi-scst. Single VMWare ESXi server (again this is all testing this phase), setup iscsi to connect to the SAN (testing MPIO) as soon as I "scan" from the storage adapter I start getting the [28609.797320] block drbd1: al_complete_io() called on inactive extent 623 ete.. see http://www.genis-x.com/files/syslog Where the log ends is where I shutdown the ESXi server If there are any additional debugging steps/options I can enable that would help, let me know. I haven't done anything with /dev/drdb0 yet this is going to hold config files etc. Background: Ubuntu 10.4 with recompiled linux-source (2.6.32.15) kernel due to iscsi-scst patches. iscsi-scst as target engine heartbeat (not yet though just getting the base working, drbd + scst) then I'll integrate heartbeat lvm2 on top of drbd root at storage00:/etc# cat drbd.conf global { usage-count no; } common { syncer { rate 110M; } } resource meta { protocol C; handlers { pri-on-incon-degr "echo O > /proc/sysrq-trigger ; halt -f"; pri-lost-after-sb "echo O > /proc/sysrq-trigger ; halt -f"; local-io-error "echo O > /proc/sysrq-trigger ; halt -f"; } startup { # wfc-timeout 0; degr-wfc-timeout 120; # 2 minutes. } disk { on-io-error detach; } net { after-sb-0pri disconnect; after-sb-1pri disconnect; after-sb-2pri disconnect; rr-conflict disconnect; } syncer { # rate 110M; # after "r2"; al-extents 257; } on storage00 { device /dev/drbd0; disk /dev/sda3; address 192.168.1.10:7788; meta-disk internal; } on storage01 { device /dev/drbd0; disk /dev/sda3; address 192.168.1.11:7788; meta-disk internal; } } resource data { protocol C; startup { wfc-timeout 0; ## Infinite! degr-wfc-timeout 120; ## 2 minutes. } disk { on-io-error detach; } net { # timeout 60; # connect-int 10; # ping-int 10; # max-buffers 2048; # max-epoch-size 2048; } syncer { after "meta"; } on storage00 { device /dev/drbd1; disk /dev/sdb1; address 192.168.1.10:7789; meta-disk internal; } on storage01 { device /dev/drbd1; disk /dev/sdb1; address 192.168.1.11:7789; meta-disk internal; } } root at storage00:/etc# fdisk -l Disk /dev/sda: 21.5 GB, 21474836480 bytes 255 heads, 63 sectors/track, 2610 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disk identifier: 0x00004968 Device Boot Start End Blocks Id System /dev/sda1 * 1 2432 19530752 83 Linux /dev/sda2 2432 2494 499712 82 Linux swap / Solaris /dev/sda3 2494 2611 939008 83 Linux Disk /dev/sdb: 42.9 GB, 42949672960 bytes 255 heads, 63 sectors/track, 5221 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disk identifier: 0x00064540 Device Boot Start End Blocks Id System /dev/sdb1 1 5222 41940992 8e Linux LVM Regards, Adam Niedzwiedzki c: eye media australia pty ltd a: suite 109, 685 burke rd, camberwell, victoria, 3124, au, earth p: +61 (0)3 8080 5850 d: +61 (0)3 8080 5674 f: +61 (0)3 8080 5840 m: +61 (0)439 646 941 w: http://www.eyemedia.com.au pgp: http://www.eyemedia.com.au/adamn.asc From: Lars Ellenberg <lars.ellenberg at linbit.com> To: drbd-user at lists.linbit.com Date: 14/07/2010 23:17 Subject: Re: [DRBD-user] 1 Primary Node Sent by: drbd-user-bounces at lists.linbit.com On Wed, Jul 14, 2010 at 03:54:48PM +1000, adamn at eyemedia.com.au wrote: > Hi all, > > I have a physical server that I'm setting up as a SAN, the second node > won't be available to put online for at least a week. (migrating FROM it > to the NEW SAN setup) > > How do I tell DRBD to I guess stop trying to synchronise? I get pages and > pages of > > [28609.791929] block drbd1: al_complete_io() called on inactive extent 623 > [28609.797320] block drbd1: al_complete_io() called on inactive extent 623 > [28609.801638] block drbd1: al_complete_io() called on inactive extent 623 > [28609.807021] block drbd1: al_complete_io() called on inactive extent 623 > [28609.811691] block drbd1: al_complete_io() called on inactive extent 623 > [28609.816680] block drbd1: al_complete_io() called on inactive extent 623 > [28609.821531] block drbd1: al_complete_io() called on inactive extent 623 > [28609.826190] block drbd1: al_complete_io() called on inactive extent 623 > [28609.829968] block drbd1: al_complete_io() called on inactive extent 623 > [28609.833730] block drbd1: al_complete_io() called on inactive extent 623 > [28609.837759] block drbd1: al_complete_io() called on inactive extent 623 Is that the only message that gets logged there? Nothing before? "That should not happen", and actually if it does happen, either something is broken in your setup, or DRBD has a bug. > I'm running version: 8.3.8.1 (api:88/proto:86-94) > > So in short I'm wanting to put my primary node online, migrate a bunch of > data from a "current" SAN (it's NOT running DRBD, around 7TB of data) to > the primary, format rebuild the current THEN sync. That plan sounds ok, and should just work. -- : Lars Ellenberg : LINBIT | Your Way to High Availability : DRBD/HA support and consulting http://www.linbit.com DRBD® and LINBIT® are registered trademarks of LINBIT, Austria. __ please don't Cc me, but send to list -- I'm subscribed _______________________________________________ drbd-user mailing list drbd-user at lists.linbit.com http://lists.linbit.com/mailman/listinfo/drbd-user -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20100715/da04ab32/attachment.htm>