[DRBD-user] 1 Primary Node

adamn at eyemedia.com.au adamn at eyemedia.com.au
Thu Jul 15 01:16:43 CEST 2010

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


Ok I didn't realise this was that kind of an issue, I just assumed DRBD 
wasn't happy with my setup due to the second node not being online.
The partition /dev/sdb1 is only set to about 40GB while I'm testing the 
setup, once I've nutted everything the machine will be rebuilt to have 
it's full capacity. (around 7TB) it's an LVM partition

I created a single Logical Volume around 20GB have it exported via 
iscsi-scst. 

Single VMWare ESXi server (again this is all testing this phase), setup 
iscsi to connect to the SAN (testing MPIO) as soon as I "scan" from the 
storage adapter I start getting the 
[28609.797320] block drbd1: al_complete_io() called on inactive extent 623 

ete..

see http://www.genis-x.com/files/syslog

Where the log ends is where I shutdown the ESXi server

If there are any additional debugging steps/options I can enable that 
would help, let me know.
I haven't done anything with /dev/drdb0 yet this is going to hold config 
files etc.

Background:
Ubuntu 10.4 with recompiled linux-source (2.6.32.15) kernel due to 
iscsi-scst patches.
iscsi-scst as target engine
heartbeat (not yet though just getting the base working, drbd + scst) then 
I'll integrate heartbeat
lvm2 on top of drbd

root at storage00:/etc# cat drbd.conf
global {
        usage-count no;
}

common {
        syncer { rate 110M; }
}
resource meta {
        protocol C;
        handlers {
         pri-on-incon-degr "echo O > /proc/sysrq-trigger ; halt -f";
         pri-lost-after-sb "echo O > /proc/sysrq-trigger ; halt -f";
         local-io-error "echo O > /proc/sysrq-trigger ; halt -f";
        }

        startup {
         # wfc-timeout 0;
         degr-wfc-timeout 120; # 2 minutes.
        }
        disk {
         on-io-error detach;
        }

        net {
         after-sb-0pri disconnect;
         after-sb-1pri disconnect;
         after-sb-2pri disconnect;
         rr-conflict disconnect;
        }

        syncer {
         # rate 110M;
         # after "r2";
         al-extents 257;
        }

        on storage00 {
         device /dev/drbd0;
         disk /dev/sda3;
         address 192.168.1.10:7788;
         meta-disk internal;
        }

        on storage01 {
         device /dev/drbd0;
         disk /dev/sda3;
         address 192.168.1.11:7788;
         meta-disk internal;
        }
}

resource data {
        protocol C;
        startup {
         wfc-timeout 0; ## Infinite!
         degr-wfc-timeout 120; ## 2 minutes.
        }

        disk {
         on-io-error detach;
        }

        net {
         # timeout 60;
         # connect-int 10;
         # ping-int 10;
         # max-buffers 2048;
         # max-epoch-size 2048;
        }

        syncer {
         after "meta";
        }

        on storage00 {
         device /dev/drbd1;
         disk /dev/sdb1;
         address 192.168.1.10:7789;
         meta-disk internal;
        }
        on storage01 {
         device /dev/drbd1;
         disk /dev/sdb1;
         address 192.168.1.11:7789;
         meta-disk internal;
        }
}

root at storage00:/etc# fdisk -l

Disk /dev/sda: 21.5 GB, 21474836480 bytes
255 heads, 63 sectors/track, 2610 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00004968

   Device Boot      Start         End      Blocks   Id  System
/dev/sda1   *           1        2432    19530752   83  Linux
/dev/sda2            2432        2494      499712   82  Linux swap / 
Solaris
/dev/sda3            2494        2611      939008   83  Linux

Disk /dev/sdb: 42.9 GB, 42949672960 bytes
255 heads, 63 sectors/track, 5221 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00064540

   Device Boot      Start         End      Blocks   Id  System
/dev/sdb1               1        5222    41940992   8e  Linux LVM

Regards,
Adam Niedzwiedzki

c: eye media australia pty ltd
a: suite 109, 685 burke rd, camberwell, victoria, 3124, au, earth
p: +61 (0)3 8080 5850
d: +61 (0)3 8080 5674
f: +61 (0)3 8080 5840
m: +61 (0)439 646 941
w: http://www.eyemedia.com.au
pgp: http://www.eyemedia.com.au/adamn.asc



From:
Lars Ellenberg <lars.ellenberg at linbit.com>
To:
drbd-user at lists.linbit.com
Date:
14/07/2010 23:17
Subject:
Re: [DRBD-user] 1 Primary Node
Sent by:
drbd-user-bounces at lists.linbit.com



On Wed, Jul 14, 2010 at 03:54:48PM +1000, adamn at eyemedia.com.au wrote:
> Hi all,
> 
> I have a physical server that I'm setting up as a SAN, the second node 
> won't be available to put online for at least a week. (migrating FROM it 

> to the NEW SAN setup)
> 
> How do I tell DRBD to I guess stop trying to synchronise? I get pages 
and 
> pages of
> 
> [28609.791929] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.797320] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.801638] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.807021] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.811691] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.816680] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.821531] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.826190] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.829968] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.833730] block drbd1: al_complete_io() called on inactive extent 
623
> [28609.837759] block drbd1: al_complete_io() called on inactive extent 
623

Is that the only message that gets logged there?
Nothing before?

"That should not happen", and actually if it does happen, either
something is broken in your setup, or DRBD has a bug.

> I'm running version: 8.3.8.1 (api:88/proto:86-94)
> 
> So in short I'm wanting to put my primary node online, migrate a bunch 
of 
> data from a "current" SAN (it's NOT running DRBD, around 7TB of data) to 

> the primary, format rebuild the current THEN sync.

That plan sounds ok, and should just work.

-- 
: Lars Ellenberg
: LINBIT | Your Way to High Availability
: DRBD/HA support and consulting http://www.linbit.com

DRBD® and LINBIT® are registered trademarks of LINBIT, Austria.
__
please don't Cc me, but send to list   --   I'm subscribed
_______________________________________________
drbd-user mailing list
drbd-user at lists.linbit.com
http://lists.linbit.com/mailman/listinfo/drbd-user

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20100715/da04ab32/attachment.htm>


More information about the drbd-user mailing list