[DRBD-user] H.A Xen

gilmarlinux at agrovale.com.br gilmarlinux at agrovale.com.br
Sat Feb 19 13:12:52 CET 2011

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.



I understood, but in this my case, I applied on lvm / dev/drbd0 not use the script for the heartbeat drbdisk Monster to partition, because the servers domu.cfg and declared the disc to be used lvm. example: disk = ['phy: / dev/VGxen/debian_buriti.disk1, xvda1, w', 'phy: / dev/VGxen/debian_buriti.disk2, xvda2, w'] Did not facilitate this admistração of drbd, because if every time foce created a VG I would have to reconfigure the drbd.conf. Result pvs: / # pvs etc/rc2.d   PV VG Fmt Attr psize pfree   / dev/drbd0 VGxen lvm2 a-924.55 g 205.55 g Part of drbd.conf   servidor1 on {     device / dev/drbd0;    
disk / dev/sda7;     10.1.1.50:7788 address;     meta-disk internal;   servidor2 on {     device / dev/drbd0;     disk / dev/sda7;     10.1.1.51:7788 address;     meta-disk internal;  
}
}> this is plain wrong. heartbeat makes the DRBD resource primary and> activates the LVM VG, so it should run before xen(d).> > > On 02/19/11 12:40, gilmarlinux at agrovale.com.br wrote:>> Yes they are
lvm, drbd when he starts running vgchange-a y. for him when>> he runs the
vgchange-an, I informed these commands within the script>> drbd startup.>> The order of startup services are first initializes the drbd -> xen
->>> heartbeat.>>>>> What's the deal with the
storage for the DomU's? LVM block devices? Are>>> the VG activated
first?>>>>>> On 02/19/11 10:52,
gilmarlinux at agrovale.com.br wrote:>>>> Hello!>>>> I
am using the solution with drbd + heartbeat xen xen to have an>>>>
environment with high availability.>>>> 2 servers with two primaries
where five runs on server1 and 2 domus>>>> domus in server2, uses the
same block device, but there is no recording>>>> simultaneously in the
same VG.>>>> The problem with that when server2 2 domU, has a fall,
after a few>>>> seconds a message appears on server1 and the same
screen and restarted.>>>> Below is the message>>>>
Thanks.>>>>>>>> [ 302.797651] BUG: soft lockup -
CPU#6 stuck for 61s! [vgchange:3054]>>>> [ 302.797713] Modules linked
in: bridge stp xen_evtchn xenfs drbd>>>> lru_cache cn loop snd_pcm
snd_timer snd soundcore snd_page_alloc psmouse>>>> joydev evdev
serio_raw power_meter dcdbas processor pcspkr button>>>>
acpi_processor ext4 mbcache jbd2 crc16 dm_mod usbhid hid sg sr_mod cdrom>>>> ata_generic sd_mod ses usb_storage enclosure crc_t10dif uhci_hcd>>>> ata_piix libata megaraid_sas ehci_hcd usbcore nls_base scsi_mod
thermal>>>> bnx2 thermal_sys [last unloaded: scsi_wait_scan]>>>> [ 302.799472] CPU 6:>>>> [ 302.799551] Modules
linked in: bridge stp xen_evtchn xenfs drbd>>>> lru_cache cn loop
snd_pcm snd_timer snd soundcore snd_page_alloc psmouse>>>> joydev
evdev serio_raw power_meter dcdbas processor pcspkr button>>>>
acpi_processor ext4 mbcache jbd2 crc16 dm_mod usbhid hid sg sr_mod cdrom>>>> ata_generic sd_mod ses usb_storage enclosure crc_t10dif uhci_hcd>>>> ata_piix libata megaraid_sas ehci_hcd usbcore nls_base scsi_mod
thermal>>>> bnx2 thermal_sys [last unloaded: scsi_wait_scan]>>>> [ 302.801304] Pid: 3054, comm: vgchange Tainted: G D>>>> 2.6.32-5-xen-amd64 #1 PowerEdge T610>>>> [
302.801379] RIP: e030:[<ffffffff8130cb18>] [<ffffffff8130cb18>]>>>> _spin_lock+0x15/0x1b>>>> [ 302.801491] RSP:
e02b:ffff88013de65b10 EFLAGS: 00000297>>>> [ 302.801546] RAX:
0000000000000049 RBX: ffff88013de65b28 RCX:>>>> ffff88013de65b68>>>> [ 302.801604] RDX: 0000000000000048 RSI: 0000000000000200 RDI:>>>> ffff88013afa3880>>>> [ 302.801662] RBP:
0000000000000200 R08: 0000000000000008 R09:>>>> ffffffff814eb870>>>> [ 302.801720] R10: 000000000000000b R11: ffffffff811ba945 R12:>>>> ffff88013afa3800>>>> [ 302.801778] R13:
000000000000c580 R14: ffff88013de65b28 R15:>>>> ffffffff814eb830>>>> [ 302.801839] FS: 00007fb85aeee7a0(0000) GS:ffff88000678b000(0000)>>>> knlGS:0000000000000000>>>> [ 302.806903] CS: e033
DS: 0000 ES: 0000 CR0: 000000008005003b>>>> [ 302.806959] CR2:
00000000028f03f8 CR3: 0000000001001000 CR4:>>>> 0000000000002660>>>> [ 302.807017] DR0: 0000000000000000 DR1: 0000000000000000 DR2:>>>> 0000000000000000>>>> [ 302.807075] DR3:
0000000000000000 DR6: 00000000ffff0ff0 DR7:>>>> 0000000000000400>>>> [ 302.807133] Call Trace:>>>> [ 302.807189]
[<ffffffff8100dd87>] ? xen_exit_mmap+0xf8/0x136>>>> [
302.807245] [<ffffffff8130cb3a>] ? _spin_unlock_irqrestore+0xd/0xe>>>> [ 302.807303] [<ffffffff8100922a>] ?
hypercall_page+0x22a/0x1001>>>> [ 302.807361]
[<ffffffff810d1208>] ? exit_mmap+0x5a/0x148>>>> [ 302.807417]
[<ffffffff8100ecf2>] ? check_events+0x12/0x20>>>> [ 302.807475]
[<ffffffff811ba945>] ? vgacon_cursor+0x0/0x140>>>> [ 302.807533]
[<ffffffff8104cb09>] ? mmput+0x3c/0xdf>>>> [ 302.807588]
[<ffffffff81050702>] ? exit_mm+0x102/0x10d>>>> [ 302.807644]
[<ffffffff8100ec99>] ? xen_irq_enable_direct_end+0x0/0x7>>>> [
302.807702] [<ffffffff81052127>] ? do_exit+0x1f8/0x6c6>>>> [
302.807758] [<ffffffff8100ecdf>] ? xen_restore_fl_direct_end+0x0/0x1>>>> [ 302.807816] [<ffffffff8130cb3a>] ?
_spin_unlock_irqrestore+0xd/0xe>>>> [ 302.807873]
[<ffffffff8104f3af>] ? release_console_sem+0x17e/0x1af>>>> [
302.807931] [<ffffffff8130d9dd>] ? oops_end+0xaf/0xb4>>>> [
302.807988] [<ffffffff810135f0>] ? do_invalid_op+0x8b/0x95>>>> [
302.808045] [<ffffffff8100c694>] ? pin_pagetable_pfn+0x2d/0x36>>>> [ 302.808105] [<ffffffffa01079ea>] ? copy_params+0x71/0xb1
[dm_mod]>>>> [ 302.808164] [<ffffffff810baf07>] ?
__alloc_pages_nodemask+0x11c/0x5f5>>>> [ 302.808222]
[<ffffffff8101293b>] ? invalid_op+0x1b/0x20>>>> [ 302.808278]
[<ffffffff8100c694>] ? pin_pagetable_pfn+0x2d/0x36>>>> [
302.808335] [<ffffffff8100c690>] ? pin_pagetable_pfn+0x29/0x36>>>> [ 302.808392] [<ffffffff810cd4e2>] ? __pte_alloc+0x6b/0xc6>>>> [ 302.808449] [<ffffffff810cb394>] ? pmd_alloc+0x28/0x5b>>>> [ 302.808505] [<ffffffff810cd60b>] ?
handle_mm_fault+0xce/0x80f>>>> [ 302.808561]
[<ffffffff810d2bcc>] ? do_brk+0x227/0x301>>>> [ 302.808619]
[<ffffffff8130f016>] ? do_page_fault+0x2e0/0x2fc>>>> [
302.808675] [<ffffffff8130ceb5>] ? page_fault+0x25/0x30>>>>>>>>>>>>>>>>
_______________________________________________>>>> drbd-user mailing
list>>>> drbd-user at lists.linbit.com>>>>
http://lists.linbit.com/mailman/listinfo/drbd-user>>>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20110219/f06ba02f/attachment.htm>


More information about the drbd-user mailing list