[DRBD-user] H.A Xen

Bart Coninckx bart.coninckx at telenet.be
Sat Feb 19 21:59:44 CET 2011

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


Hi,

I see. Well, I appreciate your effort for getting your message across,
Google translator or not.

Personally, I would not do this with heartbeat (old and depricated) but
with Pacemaker (http://www.clusterlabs.org). Pacemaker has resource
agants for LVM and Xen that will nicely take care of this.

Looking at your haresources file, not sure if the LVM stuff is tackled
in there properly. Neither is setting the DRBD primary.

Bart




On 02/19/11 21:43, gilmarlinux at agrovale.com.br wrote:
> I'm using google translator because I do not speak English and I
> appreciate very much your help.
> I got it, and the next, I have a script and placed in / etc / ha /
> haresources that starts the domU, getting so low.
> 
> /etc/ha.d/haresources
> servidor1 xendomains::buriti.cfg xendomains::jatoba.cfg
> xendomains::mangaba.cfg xendomains::cagaita.cfg xendomains::tigui.cfg
> xendomains::deladeiro.cfg xendomains::babacu.cfg
> 
> /etc/ha.d/resource.d/xendomains
> # / Bin / bash
> XM = "/ usr / sbin / xm"
> CONFPATH = "/ home / vm /"
> RES = "$ 1"
> CMD = "$ 2"
> NOME_BASE = `basename $ res. Cfg`
> IsRunning = false
> case "$ CMD" in
>     start)
>         XM create-f $ $ $ CONFPATH RES
>         ;;
>     stop)
>          XM exec $ shutdown - halt - wait $ NOME_BASE
>         ;;
>     status)
>         $ XM list | awk '{print $ 1}' | grep $ res> / dev / null
>         if [$? -Eq 0]
>         then
>                 echo running
>         else
>                 echo stopped
>         fi
>         ;;
>     *)
>         echo "Usage: xendomain [filename] {start | stop | status}"
>         exit 1
>         ;;
> esac
> 
> exit 0
> 
> 
> 
> 
> I'm guessing you're using Google because the translation is a bit Inglês
> flaky.
> 
> Anyway, I gather you Have the VG on top of DRBD. What should Happen is
> Indeed That your DomU config points to the LV in the VG, and That Heartbeat
> ACTIVATES it after making the primary DRBD resource. Mind you, this
> setup does not allow for live migration.
> 
> 
>> I'm guessing you're using Google translate because the English is a bit
>> flaky.
>>
>> Anyway, I gather you have a VG on top of DRBD. What should happen is
>> that your DomU cfg indeed points to the LV in the VG, and that Heartbeat
>> activates it after making the DRBD resource primary. Mind you, this
>> setup does not allow for live migration.
>>
>>
>> On 02/19/11 13:12, gilmarlinux at agrovale.com.br wrote:
>>> I understood, but in this my case, I applied on lvm / dev/drbd0 not use
>>> the script for the heartbeat drbdisk Monster to partition, because the
>>> servers domu.cfg and declared the disc to be used lvm. example:
>>> disk = ['phy: / dev/VGxen/debian_buriti.disk1, xvda1, w', 'phy: /
>>> dev/VGxen/debian_buriti.disk2, xvda2, w']
>>> Did not facilitate this admistração of drbd, because if every time foce
>>> created a VG I would have to reconfigure the drbd.conf.
>>>
>>> Result pvs:
>>> / # pvs etc/rc2.d
>>> PV VG Fmt Attr psize pfree
>>> / dev/drbd0 VGxen lvm2 a-924.55 g 205.55 g
>>>
>>>
>>> Part of drbd.conf
>>> servidor1 on {
>>> device / dev/drbd0;
>>> disk / dev/sda7;
>>> 10.1.1.50:7788 address;
>>> meta-disk internal;
>>>
>>> servidor2 on {
>>> device / dev/drbd0;
>>> disk / dev/sda7;
>>> 10.1.1.51:7788 address;
>>> meta-disk internal;
>>> }
>>> }
>>>
>>>> this is plain wrong. heartbeat makes the DRBD resource primary and
>>>> activates the LVM VG, so it should run before xen(d).
>>>>
>>>>
>>>> On 02/19/11 12:40, gilmarlinux at agrovale.com.br wrote:
>>>>> Yes they are lvm, drbd when he starts running vgchange-a y. for him
> when
>>>>> he runs the vgchange-an, I informed these commands within the script
>>>>> drbd startup.
>>>>> The order of startup services are first initializes the drbd -> xen ->
>>>>> heartbeat.
>>>>>
>>>>>> What's the deal with the storage for the DomU's? LVM block
> devices? Are
>>>>>> the VG activated first?
>>>>>>
>>>>>> On 02/19/11 10:52, gilmarlinux at agrovale.com.br wrote:
>>>>>>> Hello!
>>>>>>> I am using the solution with drbd + heartbeat xen xen to have an
>>>>>>> environment with high availability.
>>>>>>> 2 servers with two primaries where five runs on server1 and 2 domus
>>>>>>> domus in server2, uses the same block device, but there is no
> recording
>>>>>>> simultaneously in the same VG.
>>>>>>> The problem with that when server2 2 domU, has a fall, after a few
>>>>>>> seconds a message appears on server1 and the same screen and
> restarted.
>>>>>>> Below is the message
>>>>>>> Thanks.
>>>>>>>
>>>>>>> [ 302.797651] BUG: soft lockup - CPU#6 stuck for 61s! [vgchange:3054]
>>>>>>> [ 302.797713] Modules linked in: bridge stp xen_evtchn xenfs drbd
>>>>>>> lru_cache cn loop snd_pcm snd_timer snd soundcore snd_page_alloc
>>> psmouse
>>>>>>> joydev evdev serio_raw power_meter dcdbas processor pcspkr button
>>>>>>> acpi_processor ext4 mbcache jbd2 crc16 dm_mod usbhid hid sg sr_mod
>>> cdrom
>>>>>>> ata_generic sd_mod ses usb_storage enclosure crc_t10dif uhci_hcd
>>>>>>> ata_piix libata megaraid_sas ehci_hcd usbcore nls_base scsi_mod
> thermal
>>>>>>> bnx2 thermal_sys [last unloaded: scsi_wait_scan]
>>>>>>> [ 302.799472] CPU 6:
>>>>>>> [ 302.799551] Modules linked in: bridge stp xen_evtchn xenfs drbd
>>>>>>> lru_cache cn loop snd_pcm snd_timer snd soundcore snd_page_alloc
>>> psmouse
>>>>>>> joydev evdev serio_raw power_meter dcdbas processor pcspkr button
>>>>>>> acpi_processor ext4 mbcache jbd2 crc16 dm_mod usbhid hid sg sr_mod
>>> cdrom
>>>>>>> ata_generic sd_mod ses usb_storage enclosure crc_t10dif uhci_hcd
>>>>>>> ata_piix libata megaraid_sas ehci_hcd usbcore nls_base scsi_mod
> thermal
>>>>>>> bnx2 thermal_sys [last unloaded: scsi_wait_scan]
>>>>>>> [ 302.801304] Pid: 3054, comm: vgchange Tainted: G D
>>>>>>> 2.6.32-5-xen-amd64 #1 PowerEdge T610
>>>>>>> [ 302.801379] RIP: e030:[<ffffffff8130cb18>] [<ffffffff8130cb18>]
>>>>>>> _spin_lock+0x15/0x1b
>>>>>>> [ 302.801491] RSP: e02b:ffff88013de65b10 EFLAGS: 00000297
>>>>>>> [ 302.801546] RAX: 0000000000000049 RBX: ffff88013de65b28 RCX:
>>>>>>> ffff88013de65b68
>>>>>>> [ 302.801604] RDX: 0000000000000048 RSI: 0000000000000200 RDI:
>>>>>>> ffff88013afa3880
>>>>>>> [ 302.801662] RBP: 0000000000000200 R08: 0000000000000008 R09:
>>>>>>> ffffffff814eb870
>>>>>>> [ 302.801720] R10: 000000000000000b R11: ffffffff811ba945 R12:
>>>>>>> ffff88013afa3800
>>>>>>> [ 302.801778] R13: 000000000000c580 R14: ffff88013de65b28 R15:
>>>>>>> ffffffff814eb830
>>>>>>> [ 302.801839] FS: 00007fb85aeee7a0(0000) GS:ffff88000678b000(0000)
>>>>>>> knlGS:0000000000000000
>>>>>>> [ 302.806903] CS: e033 DS: 0000 ES: 0000 CR0: 000000008005003b
>>>>>>> [ 302.806959] CR2: 00000000028f03f8 CR3: 0000000001001000 CR4:
>>>>>>> 0000000000002660
>>>>>>> [ 302.807017] DR0: 0000000000000000 DR1: 0000000000000000 DR2:
>>>>>>> 0000000000000000
>>>>>>> [ 302.807075] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7:
>>>>>>> 0000000000000400
>>>>>>> [ 302.807133] Call Trace:
>>>>>>> [ 302.807189] [<ffffffff8100dd87>] ? xen_exit_mmap+0xf8/0x136
>>>>>>> [ 302.807245] [<ffffffff8130cb3a>] ? _spin_unlock_irqrestore+0xd/0xe
>>>>>>> [ 302.807303] [<ffffffff8100922a>] ? hypercall_page+0x22a/0x1001
>>>>>>> [ 302.807361] [<ffffffff810d1208>] ? exit_mmap+0x5a/0x148
>>>>>>> [ 302.807417] [<ffffffff8100ecf2>] ? check_events+0x12/0x20
>>>>>>> [ 302.807475] [<ffffffff811ba945>] ? vgacon_cursor+0x0/0x140
>>>>>>> [ 302.807533] [<ffffffff8104cb09>] ? mmput+0x3c/0xdf
>>>>>>> [ 302.807588] [<ffffffff81050702>] ? exit_mm+0x102/0x10d
>>>>>>> [ 302.807644] [<ffffffff8100ec99>] ?
> xen_irq_enable_direct_end+0x0/0x7
>>>>>>> [ 302.807702] [<ffffffff81052127>] ? do_exit+0x1f8/0x6c6
>>>>>>> [ 302.807758] [<ffffffff8100ecdf>] ?
> xen_restore_fl_direct_end+0x0/0x1
>>>>>>> [ 302.807816] [<ffffffff8130cb3a>] ? _spin_unlock_irqrestore+0xd/0xe
>>>>>>> [ 302.807873] [<ffffffff8104f3af>] ? release_console_sem+0x17e/0x1af
>>>>>>> [ 302.807931] [<ffffffff8130d9dd>] ? oops_end+0xaf/0xb4
>>>>>>> [ 302.807988] [<ffffffff810135f0>] ? do_invalid_op+0x8b/0x95
>>>>>>> [ 302.808045] [<ffffffff8100c694>] ? pin_pagetable_pfn+0x2d/0x36
>>>>>>> [ 302.808105] [<ffffffffa01079ea>] ? copy_params+0x71/0xb1 [dm_mod]
>>>>>>> [ 302.808164] [<ffffffff810baf07>] ?
> __alloc_pages_nodemask+0x11c/0x5f5
>>>>>>> [ 302.808222] [<ffffffff8101293b>] ? invalid_op+0x1b/0x20
>>>>>>> [ 302.808278] [<ffffffff8100c694>] ? pin_pagetable_pfn+0x2d/0x36
>>>>>>> [ 302.808335] [<ffffffff8100c690>] ? pin_pagetable_pfn+0x29/0x36
>>>>>>> [ 302.808392] [<ffffffff810cd4e2>] ? __pte_alloc+0x6b/0xc6
>>>>>>> [ 302.808449] [<ffffffff810cb394>] ? pmd_alloc+0x28/0x5b
>>>>>>> [ 302.808505] [<ffffffff810cd60b>] ? handle_mm_fault+0xce/0x80f
>>>>>>> [ 302.808561] [<ffffffff810d2bcc>] ? do_brk+0x227/0x301
>>>>>>> [ 302.808619] [<ffffffff8130f016>] ? do_page_fault+0x2e0/0x2fc
>>>>>>> [ 302.808675] [<ffffffff8130ceb5>] ? page_fault+0x25/0x30
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> drbd-user mailing list
>>>>>>> drbd-user at lists.linbit.com
>>>>>>> http://lists.linbit.com/mailman/listinfo/drbd-user
>>>>>>
>>>>
>>
> 



More information about the drbd-user mailing list