Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
I understood, but in this my case, I applied on lvm / dev/drbd0 not use the script for the heartbeat drbdisk Monster to partition, because the servers domu.cfg and declared the disc to be used lvm. example: disk = ['phy: / dev/VGxen/debian_buriti.disk1, xvda1, w', 'phy: / dev/VGxen/debian_buriti.disk2, xvda2, w'] Did not facilitate this admistração of drbd, because if every time foce created a VG I would have to reconfigure the drbd.conf. Result pvs: / # pvs etc/rc2.d PV VG Fmt Attr psize pfree / dev/drbd0 VGxen lvm2 a-924.55 g 205.55 g Part of drbd.conf servidor1 on { device / dev/drbd0; disk / dev/sda7; 10.1.1.50:7788 address; meta-disk internal; servidor2 on { device / dev/drbd0; disk / dev/sda7; 10.1.1.51:7788 address; meta-disk internal; } }> this is plain wrong. heartbeat makes the DRBD resource primary and> activates the LVM VG, so it should run before xen(d).> > > On 02/19/11 12:40, gilmarlinux at agrovale.com.br wrote:>> Yes they are lvm, drbd when he starts running vgchange-a y. for him when>> he runs the vgchange-an, I informed these commands within the script>> drbd startup.>> The order of startup services are first initializes the drbd -> xen ->>> heartbeat.>>>>> What's the deal with the storage for the DomU's? LVM block devices? Are>>> the VG activated first?>>>>>> On 02/19/11 10:52, gilmarlinux at agrovale.com.br wrote:>>>> Hello!>>>> I am using the solution with drbd + heartbeat xen xen to have an>>>> environment with high availability.>>>> 2 servers with two primaries where five runs on server1 and 2 domus>>>> domus in server2, uses the same block device, but there is no recording>>>> simultaneously in the same VG.>>>> The problem with that when server2 2 domU, has a fall, after a few>>>> seconds a message appears on server1 and the same screen and restarted.>>>> Below is the message>>>> Thanks.>>>>>>>> [ 302.797651] BUG: soft lockup - CPU#6 stuck for 61s! [vgchange:3054]>>>> [ 302.797713] Modules linked in: bridge stp xen_evtchn xenfs drbd>>>> lru_cache cn loop snd_pcm snd_timer snd soundcore snd_page_alloc psmouse>>>> joydev evdev serio_raw power_meter dcdbas processor pcspkr button>>>> acpi_processor ext4 mbcache jbd2 crc16 dm_mod usbhid hid sg sr_mod cdrom>>>> ata_generic sd_mod ses usb_storage enclosure crc_t10dif uhci_hcd>>>> ata_piix libata megaraid_sas ehci_hcd usbcore nls_base scsi_mod thermal>>>> bnx2 thermal_sys [last unloaded: scsi_wait_scan]>>>> [ 302.799472] CPU 6:>>>> [ 302.799551] Modules linked in: bridge stp xen_evtchn xenfs drbd>>>> lru_cache cn loop snd_pcm snd_timer snd soundcore snd_page_alloc psmouse>>>> joydev evdev serio_raw power_meter dcdbas processor pcspkr button>>>> acpi_processor ext4 mbcache jbd2 crc16 dm_mod usbhid hid sg sr_mod cdrom>>>> ata_generic sd_mod ses usb_storage enclosure crc_t10dif uhci_hcd>>>> ata_piix libata megaraid_sas ehci_hcd usbcore nls_base scsi_mod thermal>>>> bnx2 thermal_sys [last unloaded: scsi_wait_scan]>>>> [ 302.801304] Pid: 3054, comm: vgchange Tainted: G D>>>> 2.6.32-5-xen-amd64 #1 PowerEdge T610>>>> [ 302.801379] RIP: e030:[<ffffffff8130cb18>] [<ffffffff8130cb18>]>>>> _spin_lock+0x15/0x1b>>>> [ 302.801491] RSP: e02b:ffff88013de65b10 EFLAGS: 00000297>>>> [ 302.801546] RAX: 0000000000000049 RBX: ffff88013de65b28 RCX:>>>> ffff88013de65b68>>>> [ 302.801604] RDX: 0000000000000048 RSI: 0000000000000200 RDI:>>>> ffff88013afa3880>>>> [ 302.801662] RBP: 0000000000000200 R08: 0000000000000008 R09:>>>> ffffffff814eb870>>>> [ 302.801720] R10: 000000000000000b R11: ffffffff811ba945 R12:>>>> ffff88013afa3800>>>> [ 302.801778] R13: 000000000000c580 R14: ffff88013de65b28 R15:>>>> ffffffff814eb830>>>> [ 302.801839] FS: 00007fb85aeee7a0(0000) GS:ffff88000678b000(0000)>>>> knlGS:0000000000000000>>>> [ 302.806903] CS: e033 DS: 0000 ES: 0000 CR0: 000000008005003b>>>> [ 302.806959] CR2: 00000000028f03f8 CR3: 0000000001001000 CR4:>>>> 0000000000002660>>>> [ 302.807017] DR0: 0000000000000000 DR1: 0000000000000000 DR2:>>>> 0000000000000000>>>> [ 302.807075] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7:>>>> 0000000000000400>>>> [ 302.807133] Call Trace:>>>> [ 302.807189] [<ffffffff8100dd87>] ? xen_exit_mmap+0xf8/0x136>>>> [ 302.807245] [<ffffffff8130cb3a>] ? _spin_unlock_irqrestore+0xd/0xe>>>> [ 302.807303] [<ffffffff8100922a>] ? hypercall_page+0x22a/0x1001>>>> [ 302.807361] [<ffffffff810d1208>] ? exit_mmap+0x5a/0x148>>>> [ 302.807417] [<ffffffff8100ecf2>] ? check_events+0x12/0x20>>>> [ 302.807475] [<ffffffff811ba945>] ? vgacon_cursor+0x0/0x140>>>> [ 302.807533] [<ffffffff8104cb09>] ? mmput+0x3c/0xdf>>>> [ 302.807588] [<ffffffff81050702>] ? exit_mm+0x102/0x10d>>>> [ 302.807644] [<ffffffff8100ec99>] ? xen_irq_enable_direct_end+0x0/0x7>>>> [ 302.807702] [<ffffffff81052127>] ? do_exit+0x1f8/0x6c6>>>> [ 302.807758] [<ffffffff8100ecdf>] ? xen_restore_fl_direct_end+0x0/0x1>>>> [ 302.807816] [<ffffffff8130cb3a>] ? _spin_unlock_irqrestore+0xd/0xe>>>> [ 302.807873] [<ffffffff8104f3af>] ? release_console_sem+0x17e/0x1af>>>> [ 302.807931] [<ffffffff8130d9dd>] ? oops_end+0xaf/0xb4>>>> [ 302.807988] [<ffffffff810135f0>] ? do_invalid_op+0x8b/0x95>>>> [ 302.808045] [<ffffffff8100c694>] ? pin_pagetable_pfn+0x2d/0x36>>>> [ 302.808105] [<ffffffffa01079ea>] ? copy_params+0x71/0xb1 [dm_mod]>>>> [ 302.808164] [<ffffffff810baf07>] ? __alloc_pages_nodemask+0x11c/0x5f5>>>> [ 302.808222] [<ffffffff8101293b>] ? invalid_op+0x1b/0x20>>>> [ 302.808278] [<ffffffff8100c694>] ? pin_pagetable_pfn+0x2d/0x36>>>> [ 302.808335] [<ffffffff8100c690>] ? pin_pagetable_pfn+0x29/0x36>>>> [ 302.808392] [<ffffffff810cd4e2>] ? __pte_alloc+0x6b/0xc6>>>> [ 302.808449] [<ffffffff810cb394>] ? pmd_alloc+0x28/0x5b>>>> [ 302.808505] [<ffffffff810cd60b>] ? handle_mm_fault+0xce/0x80f>>>> [ 302.808561] [<ffffffff810d2bcc>] ? do_brk+0x227/0x301>>>> [ 302.808619] [<ffffffff8130f016>] ? do_page_fault+0x2e0/0x2fc>>>> [ 302.808675] [<ffffffff8130ceb5>] ? page_fault+0x25/0x30>>>>>>>>>>>>>>>> _______________________________________________>>>> drbd-user mailing list>>>> drbd-user at lists.linbit.com>>>> http://lists.linbit.com/mailman/listinfo/drbd-user>>>> -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20110219/f06ba02f/attachment.htm>