[DRBD-user] H.A Xen

gilmarlinux at agrovale.com.br gilmarlinux at agrovale.com.br
Sat Feb 19 21:43:15 CET 2011

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.



I'm using google translator because I do not speak English and I appreciate very
much your help.I got it, and the next, I have a script
and placed in / etc / ha / haresources that starts the domU, getting so low./etc/ha.d/haresourcesservidor1
xendomains::buriti.cfg xendomains::jatoba.cfg xendomains::mangaba.cfg
xendomains::cagaita.cfg xendomains::tigui.cfg xendomains::deladeiro.cfg
xendomains::babacu.cfg/etc/ha.d/resource.d/xendomains# / Bin
/ bashXM = "/ usr / sbin /
xm"CONFPATH = "/
home / vm /"RES = "$ 1"CMD = "$ 2"NOME_BASE = `basename $ res. Cfg`IsRunning = falsecase "$ CMD" in    start)        XM create-f $ $ $ CONFPATH RES        ;;    stop)         XM exec $ shutdown - halt - wait $ NOME_BASE        ;;    status)        $ XM list | awk '{print $
1}' | grep $ res> / dev / null        if [$?
-Eq 0]        then                echo running        else                echo stopped        fi        ;;    *)        echo "Usage: xendomain [filename]
{start | stop | status}"        exit 1        ;;esacexit 0I'm guessing you're using Google because the translation
is a bit Inglêsflaky.Anyway, I gather you Have the VG on top of DRBD. What should Happen isIndeed That your DomU config
points to the LV in the VG, and That HeartbeatACTIVATES it after making the primary DRBD resource. Mind you, thissetup does not allow for live
migration.> I'm guessing you're using Google
translate because the English is a bit> flaky.> > Anyway, I
gather you have a VG on top of DRBD. What should happen is> that your DomU cfg
indeed points to the LV in the VG, and that Heartbeat> activates it after
making the DRBD resource primary. Mind you, this> setup does not allow for live
migration.> > > On 02/19/11 13:12,
gilmarlinux at agrovale.com.br wrote:>> I understood, but in this my case, I
applied on lvm / dev/drbd0 not use>> the script for the heartbeat drbdisk
Monster to partition, because the>> servers domu.cfg and declared the disc
to be used lvm. example:>> disk = ['phy: / dev/VGxen/debian_buriti.disk1,
xvda1, w', 'phy: />> dev/VGxen/debian_buriti.disk2, xvda2, w']>>
Did not facilitate this admistração of drbd, because if every time foce>> created a VG I would have to reconfigure the drbd.conf.>>>> Result pvs:>> / # pvs etc/rc2.d>>    PV VG Fmt Attr
psize pfree>>    / dev/drbd0 VGxen lvm2 a-924.55 g 205.55 g>>>>>> Part of drbd.conf>>    servidor1 on {>>
     device / dev/drbd0;>>      disk / dev/sda7;>>     
10.1.1.50:7788 address;>>      meta-disk internal;>>>>   servidor2 on {>>      device / dev/drbd0;>>     
disk / dev/sda7;>>      10.1.1.51:7788 address;>>      meta-disk
internal;>>    }>> }>>>>> this is
plain wrong. heartbeat makes the DRBD resource primary and>>> activates
the LVM VG, so it should run before xen(d).>>>>>>>>> On 02/19/11 12:40, gilmarlinux at agrovale.com.br wrote:>>>> Yes they are lvm, drbd when he starts running vgchange-a y. for him
when>>>> he runs the vgchange-an, I informed these commands within the
script>>>> drbd startup.>>>> The order of startup
services are first initializes the drbd -> xen ->>>>>
heartbeat.>>>>>>>>> What's the deal with the
storage for the DomU's? LVM block devices? Are>>>>> the VG
activated first?>>>>>>>>>> On 02/19/11 10:52,
gilmarlinux at agrovale.com.br wrote:>>>>>> Hello!>>>>>> I am using the solution with drbd + heartbeat xen xen to have
an>>>>>> environment with high availability.>>>>>> 2 servers with two primaries where five runs on server1 and 2
domus>>>>>> domus in server2, uses the same block device, but
there is no recording>>>>>> simultaneously in the same VG.>>>>>> The problem with that when server2 2 domU, has a fall, after
a few>>>>>> seconds a message appears on server1 and the same
screen and restarted.>>>>>> Below is the message>>>>>> Thanks.>>>>>>>>>>>> [ 302.797651] BUG: soft lockup - CPU#6 stuck for 61s!
[vgchange:3054]>>>>>> [ 302.797713] Modules linked in: bridge
stp xen_evtchn xenfs drbd>>>>>> lru_cache cn loop snd_pcm
snd_timer snd soundcore snd_page_alloc>> psmouse>>>>>> joydev evdev serio_raw power_meter dcdbas processor pcspkr
button>>>>>> acpi_processor ext4 mbcache jbd2 crc16 dm_mod
usbhid hid sg sr_mod>> cdrom>>>>>> ata_generic
sd_mod ses usb_storage enclosure crc_t10dif uhci_hcd>>>>>>
ata_piix libata megaraid_sas ehci_hcd usbcore nls_base scsi_mod thermal>>>>>> bnx2 thermal_sys [last unloaded: scsi_wait_scan]>>>>>> [ 302.799472] CPU 6:>>>>>> [
302.799551] Modules linked in: bridge stp xen_evtchn xenfs drbd>>>>>> lru_cache cn loop snd_pcm snd_timer snd soundcore
snd_page_alloc>> psmouse>>>>>> joydev evdev
serio_raw power_meter dcdbas processor pcspkr button>>>>>>
acpi_processor ext4 mbcache jbd2 crc16 dm_mod usbhid hid sg sr_mod>>
cdrom>>>>>> ata_generic sd_mod ses usb_storage enclosure
crc_t10dif uhci_hcd>>>>>> ata_piix libata megaraid_sas ehci_hcd
usbcore nls_base scsi_mod thermal>>>>>> bnx2 thermal_sys [last
unloaded: scsi_wait_scan]>>>>>> [ 302.801304] Pid: 3054, comm:
vgchange Tainted: G D>>>>>> 2.6.32-5-xen-amd64 #1 PowerEdge
T610>>>>>> [ 302.801379] RIP: e030:[<ffffffff8130cb18>]
[<ffffffff8130cb18>]>>>>>> _spin_lock+0x15/0x1b>>>>>> [ 302.801491] RSP: e02b:ffff88013de65b10 EFLAGS: 00000297>>>>>> [ 302.801546] RAX: 0000000000000049 RBX: ffff88013de65b28
RCX:>>>>>> ffff88013de65b68>>>>>> [
302.801604] RDX: 0000000000000048 RSI: 0000000000000200 RDI:>>>>>> ffff88013afa3880>>>>>> [ 302.801662]
RBP: 0000000000000200 R08: 0000000000000008 R09:>>>>>>
ffffffff814eb870>>>>>> [ 302.801720] R10: 000000000000000b R11:
ffffffff811ba945 R12:>>>>>> ffff88013afa3800>>>>>> [ 302.801778] R13: 000000000000c580 R14: ffff88013de65b28
R15:>>>>>> ffffffff814eb830>>>>>> [
302.801839] FS: 00007fb85aeee7a0(0000) GS:ffff88000678b000(0000)>>>>>> knlGS:0000000000000000>>>>>> [
302.806903] CS: e033 DS: 0000 ES: 0000 CR0: 000000008005003b>>>>>> [ 302.806959] CR2: 00000000028f03f8 CR3: 0000000001001000
CR4:>>>>>> 0000000000002660>>>>>> [
302.807017] DR0: 0000000000000000 DR1: 0000000000000000 DR2:>>>>>> 0000000000000000>>>>>> [ 302.807075]
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7:>>>>>>
0000000000000400>>>>>> [ 302.807133] Call Trace:>>>>>> [ 302.807189] [<ffffffff8100dd87>] ?
xen_exit_mmap+0xf8/0x136>>>>>> [ 302.807245]
[<ffffffff8130cb3a>] ? _spin_unlock_irqrestore+0xd/0xe>>>>>> [ 302.807303] [<ffffffff8100922a>] ?
hypercall_page+0x22a/0x1001>>>>>> [ 302.807361]
[<ffffffff810d1208>] ? exit_mmap+0x5a/0x148>>>>>> [
302.807417] [<ffffffff8100ecf2>] ? check_events+0x12/0x20>>>>>> [ 302.807475] [<ffffffff811ba945>] ?
vgacon_cursor+0x0/0x140>>>>>> [ 302.807533]
[<ffffffff8104cb09>] ? mmput+0x3c/0xdf>>>>>> [ 302.807588]
[<ffffffff81050702>] ? exit_mm+0x102/0x10d>>>>>> [
302.807644] [<ffffffff8100ec99>] ? xen_irq_enable_direct_end+0x0/0x7>>>>>> [ 302.807702] [<ffffffff81052127>] ?
do_exit+0x1f8/0x6c6>>>>>> [ 302.807758]
[<ffffffff8100ecdf>] ? xen_restore_fl_direct_end+0x0/0x1>>>>>> [ 302.807816] [<ffffffff8130cb3a>] ?
_spin_unlock_irqrestore+0xd/0xe>>>>>> [ 302.807873]
[<ffffffff8104f3af>] ? release_console_sem+0x17e/0x1af>>>>>> [ 302.807931] [<ffffffff8130d9dd>] ?
oops_end+0xaf/0xb4>>>>>> [ 302.807988]
[<ffffffff810135f0>] ? do_invalid_op+0x8b/0x95>>>>>> [
302.808045] [<ffffffff8100c694>] ? pin_pagetable_pfn+0x2d/0x36>>>>>> [ 302.808105] [<ffffffffa01079ea>] ?
copy_params+0x71/0xb1 [dm_mod]>>>>>> [ 302.808164]
[<ffffffff810baf07>] ? __alloc_pages_nodemask+0x11c/0x5f5>>>>>> [ 302.808222] [<ffffffff8101293b>] ?
invalid_op+0x1b/0x20>>>>>> [ 302.808278]
[<ffffffff8100c694>] ? pin_pagetable_pfn+0x2d/0x36>>>>>> [
302.808335] [<ffffffff8100c690>] ? pin_pagetable_pfn+0x29/0x36>>>>>> [ 302.808392] [<ffffffff810cd4e2>] ?
__pte_alloc+0x6b/0xc6>>>>>> [ 302.808449]
[<ffffffff810cb394>] ? pmd_alloc+0x28/0x5b>>>>>> [
302.808505] [<ffffffff810cd60b>] ? handle_mm_fault+0xce/0x80f>>>>>> [ 302.808561] [<ffffffff810d2bcc>] ?
do_brk+0x227/0x301>>>>>> [ 302.808619]
[<ffffffff8130f016>] ? do_page_fault+0x2e0/0x2fc>>>>>> [
302.808675] [<ffffffff8130ceb5>] ? page_fault+0x25/0x30>>>>>>>>>>>>>>>>>>>>>>>>
_______________________________________________>>>>>> drbd-user
mailing list>>>>>> drbd-user at lists.linbit.com>>>>>> http://lists.linbit.com/mailman/listinfo/drbd-user>>>>>>>>> 
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20110219/df68ea2d/attachment.htm>


More information about the drbd-user mailing list