[DRBD-user] problem setting up a logical volume with drbd

listslut at outofoptions.net listslut at outofoptions.net
Tue Jul 19 18:47:52 CEST 2011

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


Digimer.  Thanks for the help.  I went back and partitioned the drive  
for lvm and then it didn't show up at all.  I removed the partitions  
and suddenly everything went according to plan.  I'm guessing this  
just got everything back to a known state.  Not sure.  The changes on  
the first machine would always show up on the second machine. So, who  
knows...

Thanks Again.
Ken

Quoting listslut at outofoptions.net:

> Quoting Digimer <linux at alteeve.com>:
>
>> On 07/19/2011 11:46 AM, listslut at outofoptions.net wrote:
>>> Quoting Digimer <linux at alteeve.com>:
>>>
>>>> On 07/19/2011 11:13 AM, listslut at outofoptions.net wrote:
>>>>> Quoting Digimer <linux at alteeve.com>:
>>>>>
>>>>>> On 07/19/2011 10:55 AM, listslut at outofoptions.net wrote:
>>>>>>> I can create a Volume Group and the data gets replicated from one
>>>>>>> machine to the other.  This command fails though:
>>>>>>>
>>>>>>> [root at thing1 lvm]# /usr/sbin/lvcreate -n vmdata -l 69972 thing0
>>>>>>>    Logging initialised at Tue Jul 19 10:53:21 2011
>>>>>>>    Set umask to 0077
>>>>>>>    Setting logging type to disk
>>>>>>>    Finding volume group "thing0"
>>>>>>>    Archiving volume group "thing0" metadata (seqno 15).
>>>>>>>    Creating logical volume vmdata
>>>>>>>    Creating volume group backup "/etc/lvm/backup/thing0" (seqno 16).
>>>>>>>  Error locking on node thing1.eyemg.com: device-mapper: create ioctl
>>>>>>> failed: Device or resource busy
>>>>>>>  Error locking on node thing2.eyemg.com: device-mapper: create ioctl
>>>>>>> failed: Device or resource busy
>>>>>>>  Failed to activate new LV.
>>>>>>>    Creating volume group backup "/etc/lvm/backup/thing0" (seqno 17).
>>>>>>>    Wiping internal VG cache
>>>>>>>
>>>>>>> I don't know if this is a drbd issue or an lvm.  I can't seem to find
>>>>>>> anything on it.
>>>>>>>
>>>>>>> Thanks
>>>>>>> Ken Lowther
>>>>>>
>>>>>> Obvius question first; Is the DRBD in Primary?
>>>>>
>>>>> It was primary/primary but this gave me an idea.  I changed the 'other'
>>>>> node to secondary and it did change the error message some in that it
>>>>> gives a 'uuid not found' error instead on the second resource instead of
>>>>> the lock error.
>>>>>
>>>>> [root at thing1 lvm]# /usr/sbin/lvcreate -n vmdata -l 69972 thing0
>>>>>    Logging initialised at Tue Jul 19 11:09:02 2011
>>>>>    Set umask to 0077
>>>>>    Setting logging type to disk
>>>>>    Finding volume group "thing0"
>>>>>    Archiving volume group "thing0" metadata (seqno 19).
>>>>>    Creating logical volume vmdata
>>>>>    Creating volume group backup "/etc/lvm/backup/thing0" (seqno 20).
>>>>>  Error locking on node thing1.eyemg.com: device-mapper: create ioctl
>>>>> failed: Device or resource busy
>>>>>  Error locking on node thing2.eyemg.com: Volume group for uuid not
>>>>> found: hqcys8c9fDoBtX4UGLV0lmAbTZ7FMW8516YBHLfh64TzKNxRBqDH1wYg7IQHMRul
>>>>>  Failed to activate new LV.
>>>>>    Creating volume group backup "/etc/lvm/backup/thing0" (seqno 21).
>>>>>    Wiping internal VG cache
>>>>>
>>>>>
>>>>>>
>>>>>> Please share more details about your config so that folks can better
>>>>>> help you, rather than making wild stabs in the dark. :)
>>>>>>
>>>>> Using locking 3 in lvm with clvmd running.
>>>>> resource drbd0 {
>>>>>        on thing1.eyemg.com {
>>>>>                disk /dev/cciss/c0d1;
>>>>>                device /dev/drbd0;
>>>>>                meta-disk internal;
>>>>>                address 192.168.244.1:7788;
>>>>>        }
>>>>>        on thing2.eyemg.com {
>>>>>                disk /dev/cciss/c0d1;
>>>>>                device /dev/drbd0;
>>>>>                meta-disk internal;
>>>>>                address 192.168.244.2:7788;
>>>>>        }
>>>>> Ken
>>>>
>>>> Can I assume then that the cluster itself is running and quorate?
>>>
>>> [root at thing1 ~]# cman_tool nodes
>>> Node  Sts   Inc   Joined               Name
>>>   1   M    392   2011-07-18 15:35:22  thing1.eyemg.com
>>>   2   M    592   2011-07-18 15:37:02  thing2.eyemg.com
>>> [root at thing1 ~]#
>>>
>>>> Also,
>>>> is anything else (trying to) use the backing devices?
>>>
>>> Not that I know of.  I'm setting up a new cluster.  What types of things
>>> might I look for that I don't know about?
>>>
>>>> Does this error
>>>> occur on both nodes?
>>>
>>> Yes.
>>>
>>> Thanks
>>> Ken
>>
>> Can you paste the following:
>>
>> /etc/cluster/cluster.conf (only sanitize the PWs please)
>
> <?xml version="1.0"?>
> <cluster config_version="8" name="thing0">
> 	<fence_daemon post_fail_delay="0" post_join_delay="3"/>
> 	<clusternodes>
> 		<clusternode name="thing1.eyemg.com" nodeid="1" votes="1">
> 			<fence/>
> 		</clusternode>
> 		<clusternode name="thing2.eyemg.com" nodeid="2" votes="1">
> 			<fence/>
> 		</clusternode>
> 	</clusternodes>
> 	<cman expected_votes="1" two_node="1"/>
> 	<fencedevices>
> 		<fencedevice agent="fence_manual" name="thing1-ilo.eyemg.com"/>
> 		<fencedevice agent="fence_manual" name="thing2-ilo.eyemg.com"/>
> 	</fencedevices>
> 	<rm>
> 		<failoverdomains/>
> 		<resources>
> 			<script file="/etc/init.d/httpd" name="Apache"/>
> 			<ip address="10.0.8.177" monitor_link="1"/>
> 			<script file="/etc/init.d/mysqld" name="mysqld"/>
> 			<ip address="10.0.8.178" monitor_link="1"/>
> 		</resources>
> 		<service autostart="0" name="httpd" recovery="disable">
> 			<ip ref="10.0.8.177"/>
> 			<script ref="Apache"/>
> 		</service>
> 		<service autostart="0" name="mysqld" recovery="disable">
> 			<ip ref="10.0.8.178"/>
> 			<script ref="mysqld"/>
> 		</service>
> 	</rm>
> </cluster>
>
>
>> cman_tool status
>
> [root at thing1 software]#  cman_tool status
> Version: 6.2.0
> Config Version: 8
> Cluster Name: thing0
> Cluster Id: 6910
> Cluster Member: Yes
> Cluster Generation: 720
> Membership state: Cluster-Member
> Nodes: 3
> Expected votes: 1
> Total votes: 2
> Quorum: 1
> Active subsystems: 9
> Flags: 2node Dirty
> Ports Bound: 0 11 177
> Node name: thing1.eyemg.com
> Node ID: 1
> Multicast addresses: 239.192.26.25
> Node addresses: 10.0.8.201
>
>
>> cat /proc/drbd
> root at thing1 software]# cat /proc/drbd
> version: 8.3.11 (api:88/proto:86-96)
> GIT-hash: 0de839cee13a4160eed6037c4bddd066645e23c5 build by  
> root at thing1.eyemg.com, 2011-07-15 09:53:15
>  0: cs:Connected ro:Primary/Primary ds:UpToDate/UpToDate C r-----
>     ns:0 nr:0 dw:0 dr:344 al:0 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:b oos:0
>
>
>>
>> Also, is there anything of interest in the log files of note starting
>> with the cluster coming online?
>
> I do notice that I can't seem to use lvm tools to do much.  I  
> rebooted to a clean state and started over.  Since then:
>
> [root at thing2 backup]# pvdisplay
>   Skipping clustered volume group thing0
>   Skipping volume group thing0
> [root at thing2 backup]# lvm pvs
>   Skipping clustered volume group thing0
>   Skipping volume group thing0
> [root at thing2 backup]# vgdisplay
>   Skipping clustered volume group thing0
> [root at thing2 backup]# lvm vgs
>   Skipping clustered volume group thing0
> [root at thing2 backup]# lvdisplay
>   Skipping clustered volume group thing0
>   Skipping volume group thing0
> [root at thing2 backup]# lvm lvs
>   Skipping clustered volume group thing0
>   Skipping volume group thing0
>
> I appreciate your help.  I can't even wipe and start over using the  
> tools because it skips the clustered volume......
>
> Ken
>
> _______________________________________________
> drbd-user mailing list
> drbd-user at lists.linbit.com
> http://lists.linbit.com/mailman/listinfo/drbd-user
>






More information about the drbd-user mailing list