[DRBD-user] problem setting up a logical volume with drbd

listslut at outofoptions.net listslut at outofoptions.net
Tue Jul 19 18:24:28 CEST 2011

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


Quoting Digimer <linux at alteeve.com>:

> On 07/19/2011 11:46 AM, listslut at outofoptions.net wrote:
>> Quoting Digimer <linux at alteeve.com>:
>>
>>> On 07/19/2011 11:13 AM, listslut at outofoptions.net wrote:
>>>> Quoting Digimer <linux at alteeve.com>:
>>>>
>>>>> On 07/19/2011 10:55 AM, listslut at outofoptions.net wrote:
>>>>>> I can create a Volume Group and the data gets replicated from one
>>>>>> machine to the other.  This command fails though:
>>>>>>
>>>>>> [root at thing1 lvm]# /usr/sbin/lvcreate -n vmdata -l 69972 thing0
>>>>>>     Logging initialised at Tue Jul 19 10:53:21 2011
>>>>>>     Set umask to 0077
>>>>>>     Setting logging type to disk
>>>>>>     Finding volume group "thing0"
>>>>>>     Archiving volume group "thing0" metadata (seqno 15).
>>>>>>     Creating logical volume vmdata
>>>>>>     Creating volume group backup "/etc/lvm/backup/thing0" (seqno 16).
>>>>>>   Error locking on node thing1.eyemg.com: device-mapper: create ioctl
>>>>>> failed: Device or resource busy
>>>>>>   Error locking on node thing2.eyemg.com: device-mapper: create ioctl
>>>>>> failed: Device or resource busy
>>>>>>   Failed to activate new LV.
>>>>>>     Creating volume group backup "/etc/lvm/backup/thing0" (seqno 17).
>>>>>>     Wiping internal VG cache
>>>>>>
>>>>>> I don't know if this is a drbd issue or an lvm.  I can't seem to find
>>>>>> anything on it.
>>>>>>
>>>>>> Thanks
>>>>>> Ken Lowther
>>>>>
>>>>> Obvius question first; Is the DRBD in Primary?
>>>>
>>>> It was primary/primary but this gave me an idea.  I changed the 'other'
>>>> node to secondary and it did change the error message some in that it
>>>> gives a 'uuid not found' error instead on the second resource instead of
>>>> the lock error.
>>>>
>>>> [root at thing1 lvm]# /usr/sbin/lvcreate -n vmdata -l 69972 thing0
>>>>     Logging initialised at Tue Jul 19 11:09:02 2011
>>>>     Set umask to 0077
>>>>     Setting logging type to disk
>>>>     Finding volume group "thing0"
>>>>     Archiving volume group "thing0" metadata (seqno 19).
>>>>     Creating logical volume vmdata
>>>>     Creating volume group backup "/etc/lvm/backup/thing0" (seqno 20).
>>>>   Error locking on node thing1.eyemg.com: device-mapper: create ioctl
>>>> failed: Device or resource busy
>>>>   Error locking on node thing2.eyemg.com: Volume group for uuid not
>>>> found: hqcys8c9fDoBtX4UGLV0lmAbTZ7FMW8516YBHLfh64TzKNxRBqDH1wYg7IQHMRul
>>>>   Failed to activate new LV.
>>>>     Creating volume group backup "/etc/lvm/backup/thing0" (seqno 21).
>>>>     Wiping internal VG cache
>>>>
>>>>
>>>>>
>>>>> Please share more details about your config so that folks can better
>>>>> help you, rather than making wild stabs in the dark. :)
>>>>>
>>>> Using locking 3 in lvm with clvmd running.
>>>> resource drbd0 {
>>>>         on thing1.eyemg.com {
>>>>                 disk /dev/cciss/c0d1;
>>>>                 device /dev/drbd0;
>>>>                 meta-disk internal;
>>>>                 address 192.168.244.1:7788;
>>>>         }
>>>>         on thing2.eyemg.com {
>>>>                 disk /dev/cciss/c0d1;
>>>>                 device /dev/drbd0;
>>>>                 meta-disk internal;
>>>>                 address 192.168.244.2:7788;
>>>>         }
>>>> Ken
>>>
>>> Can I assume then that the cluster itself is running and quorate?
>>
>> [root at thing1 ~]# cman_tool nodes
>> Node  Sts   Inc   Joined               Name
>>    1   M    392   2011-07-18 15:35:22  thing1.eyemg.com
>>    2   M    592   2011-07-18 15:37:02  thing2.eyemg.com
>> [root at thing1 ~]#
>>
>>> Also,
>>> is anything else (trying to) use the backing devices?
>>
>> Not that I know of.  I'm setting up a new cluster.  What types of things
>> might I look for that I don't know about?
>>
>>> Does this error
>>> occur on both nodes?
>>
>> Yes.
>>
>> Thanks
>> Ken
>
> Can you paste the following:
>
> /etc/cluster/cluster.conf (only sanitize the PWs please)

<?xml version="1.0"?>
<cluster config_version="8" name="thing0">
	<fence_daemon post_fail_delay="0" post_join_delay="3"/>
	<clusternodes>
		<clusternode name="thing1.eyemg.com" nodeid="1" votes="1">
			<fence/>
		</clusternode>
		<clusternode name="thing2.eyemg.com" nodeid="2" votes="1">
			<fence/>
		</clusternode>
	</clusternodes>
	<cman expected_votes="1" two_node="1"/>
	<fencedevices>
		<fencedevice agent="fence_manual" name="thing1-ilo.eyemg.com"/>
		<fencedevice agent="fence_manual" name="thing2-ilo.eyemg.com"/>
	</fencedevices>
	<rm>
		<failoverdomains/>
		<resources>
			<script file="/etc/init.d/httpd" name="Apache"/>
			<ip address="10.0.8.177" monitor_link="1"/>
			<script file="/etc/init.d/mysqld" name="mysqld"/>
			<ip address="10.0.8.178" monitor_link="1"/>
		</resources>
		<service autostart="0" name="httpd" recovery="disable">
			<ip ref="10.0.8.177"/>
			<script ref="Apache"/>
		</service>
		<service autostart="0" name="mysqld" recovery="disable">
			<ip ref="10.0.8.178"/>
			<script ref="mysqld"/>
		</service>
	</rm>
</cluster>


> cman_tool status

[root at thing1 software]#  cman_tool status
Version: 6.2.0
Config Version: 8
Cluster Name: thing0
Cluster Id: 6910
Cluster Member: Yes
Cluster Generation: 720
Membership state: Cluster-Member
Nodes: 3
Expected votes: 1
Total votes: 2
Quorum: 1
Active subsystems: 9
Flags: 2node Dirty
Ports Bound: 0 11 177
Node name: thing1.eyemg.com
Node ID: 1
Multicast addresses: 239.192.26.25
Node addresses: 10.0.8.201


> cat /proc/drbd
root at thing1 software]# cat /proc/drbd
version: 8.3.11 (api:88/proto:86-96)
GIT-hash: 0de839cee13a4160eed6037c4bddd066645e23c5 build by  
root at thing1.eyemg.com, 2011-07-15 09:53:15
  0: cs:Connected ro:Primary/Primary ds:UpToDate/UpToDate C r-----
     ns:0 nr:0 dw:0 dr:344 al:0 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:b oos:0


>
> Also, is there anything of interest in the log files of note starting
> with the cluster coming online?

I do notice that I can't seem to use lvm tools to do much.  I rebooted  
to a clean state and started over.  Since then:

[root at thing2 backup]# pvdisplay
   Skipping clustered volume group thing0
   Skipping volume group thing0
[root at thing2 backup]# lvm pvs
   Skipping clustered volume group thing0
   Skipping volume group thing0
[root at thing2 backup]# vgdisplay
   Skipping clustered volume group thing0
[root at thing2 backup]# lvm vgs
   Skipping clustered volume group thing0
[root at thing2 backup]# lvdisplay
   Skipping clustered volume group thing0
   Skipping volume group thing0
[root at thing2 backup]# lvm lvs
   Skipping clustered volume group thing0
   Skipping volume group thing0

I appreciate your help.  I can't even wipe and start over using the  
tools because it skips the clustered volume......

Ken




More information about the drbd-user mailing list