Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Ok I didn't realise this was that kind of an issue, I just assumed DRBD
wasn't happy with my setup due to the second node not being online.
The partition /dev/sdb1 is only set to about 40GB while I'm testing the
setup, once I've nutted everything the machine will be rebuilt to have
it's full capacity. (around 7TB) it's an LVM partition
I created a single Logical Volume around 20GB have it exported via
iscsi-scst.
Single VMWare ESXi server (again this is all testing this phase), setup
iscsi to connect to the SAN (testing MPIO) as soon as I "scan" from the
storage adapter I start getting the
[28609.797320] block drbd1: al_complete_io() called on inactive extent 623
ete..
see http://www.genis-x.com/files/syslog
Where the log ends is where I shutdown the ESXi server
If there are any additional debugging steps/options I can enable that
would help, let me know.
I haven't done anything with /dev/drdb0 yet this is going to hold config
files etc.
Background:
Ubuntu 10.4 with recompiled linux-source (2.6.32.15) kernel due to
iscsi-scst patches.
iscsi-scst as target engine
heartbeat (not yet though just getting the base working, drbd + scst) then
I'll integrate heartbeat
lvm2 on top of drbd
root at storage00:/etc# cat drbd.conf
global {
usage-count no;
}
common {
syncer { rate 110M; }
}
resource meta {
protocol C;
handlers {
pri-on-incon-degr "echo O > /proc/sysrq-trigger ; halt -f";
pri-lost-after-sb "echo O > /proc/sysrq-trigger ; halt -f";
local-io-error "echo O > /proc/sysrq-trigger ; halt -f";
}
startup {
# wfc-timeout 0;
degr-wfc-timeout 120; # 2 minutes.
}
disk {
on-io-error detach;
}
net {
after-sb-0pri disconnect;
after-sb-1pri disconnect;
after-sb-2pri disconnect;
rr-conflict disconnect;
}
syncer {
# rate 110M;
# after "r2";
al-extents 257;
}
on storage00 {
device /dev/drbd0;
disk /dev/sda3;
address 192.168.1.10:7788;
meta-disk internal;
}
on storage01 {
device /dev/drbd0;
disk /dev/sda3;
address 192.168.1.11:7788;
meta-disk internal;
}
}
resource data {
protocol C;
startup {
wfc-timeout 0; ## Infinite!
degr-wfc-timeout 120; ## 2 minutes.
}
disk {
on-io-error detach;
}
net {
# timeout 60;
# connect-int 10;
# ping-int 10;
# max-buffers 2048;
# max-epoch-size 2048;
}
syncer {
after "meta";
}
on storage00 {
device /dev/drbd1;
disk /dev/sdb1;
address 192.168.1.10:7789;
meta-disk internal;
}
on storage01 {
device /dev/drbd1;
disk /dev/sdb1;
address 192.168.1.11:7789;
meta-disk internal;
}
}
root at storage00:/etc# fdisk -l
Disk /dev/sda: 21.5 GB, 21474836480 bytes
255 heads, 63 sectors/track, 2610 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00004968
Device Boot Start End Blocks Id System
/dev/sda1 * 1 2432 19530752 83 Linux
/dev/sda2 2432 2494 499712 82 Linux swap /
Solaris
/dev/sda3 2494 2611 939008 83 Linux
Disk /dev/sdb: 42.9 GB, 42949672960 bytes
255 heads, 63 sectors/track, 5221 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00064540
Device Boot Start End Blocks Id System
/dev/sdb1 1 5222 41940992 8e Linux LVM
Regards,
Adam Niedzwiedzki
c: eye media australia pty ltd
a: suite 109, 685 burke rd, camberwell, victoria, 3124, au, earth
p: +61 (0)3 8080 5850
d: +61 (0)3 8080 5674
f: +61 (0)3 8080 5840
m: +61 (0)439 646 941
w: http://www.eyemedia.com.au
pgp: http://www.eyemedia.com.au/adamn.asc
From:
Lars Ellenberg <lars.ellenberg at linbit.com>
To:
drbd-user at lists.linbit.com
Date:
14/07/2010 23:17
Subject:
Re: [DRBD-user] 1 Primary Node
Sent by:
drbd-user-bounces at lists.linbit.com
On Wed, Jul 14, 2010 at 03:54:48PM +1000, adamn at eyemedia.com.au wrote:
> Hi all,
>
> I have a physical server that I'm setting up as a SAN, the second node
> won't be available to put online for at least a week. (migrating FROM it
> to the NEW SAN setup)
>
> How do I tell DRBD to I guess stop trying to synchronise? I get pages
and
> pages of
>
> [28609.791929] block drbd1: al_complete_io() called on inactive extent
623
> [28609.797320] block drbd1: al_complete_io() called on inactive extent
623
> [28609.801638] block drbd1: al_complete_io() called on inactive extent
623
> [28609.807021] block drbd1: al_complete_io() called on inactive extent
623
> [28609.811691] block drbd1: al_complete_io() called on inactive extent
623
> [28609.816680] block drbd1: al_complete_io() called on inactive extent
623
> [28609.821531] block drbd1: al_complete_io() called on inactive extent
623
> [28609.826190] block drbd1: al_complete_io() called on inactive extent
623
> [28609.829968] block drbd1: al_complete_io() called on inactive extent
623
> [28609.833730] block drbd1: al_complete_io() called on inactive extent
623
> [28609.837759] block drbd1: al_complete_io() called on inactive extent
623
Is that the only message that gets logged there?
Nothing before?
"That should not happen", and actually if it does happen, either
something is broken in your setup, or DRBD has a bug.
> I'm running version: 8.3.8.1 (api:88/proto:86-94)
>
> So in short I'm wanting to put my primary node online, migrate a bunch
of
> data from a "current" SAN (it's NOT running DRBD, around 7TB of data) to
> the primary, format rebuild the current THEN sync.
That plan sounds ok, and should just work.
--
: Lars Ellenberg
: LINBIT | Your Way to High Availability
: DRBD/HA support and consulting http://www.linbit.com
DRBD® and LINBIT® are registered trademarks of LINBIT, Austria.
__
please don't Cc me, but send to list -- I'm subscribed
_______________________________________________
drbd-user mailing list
drbd-user at lists.linbit.com
http://lists.linbit.com/mailman/listinfo/drbd-user
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20100715/da04ab32/attachment.htm>