Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Hello guys, we have a problem getting pacemaker and drbd-8.3.11 running und RHEL 6.1 We also tried drbd-8.3.9. Pacemaker itself works perfectly when you take over ips or start daemons, but we cant get it to manage drbd resources and mount a drbd filesystem. The problem is: When pacemaker tries to mount the drbd filesystem we get the error: fs0_start_0 (node=lb02, call=31, rc=1, status=complete): unknown error We compiled drbd using: ./configure --enable-spec --with-km --prefix=/ make tgz cp drbd*.tar.gz `rpm -E _sourcedir` rpmbuild -bb drbd.spec rpmbuild -bb drbd-km.spec and then installed the resulting rpms. We know drbd very well and the drbd resources are ok and running, we can manage them using drbdadm and see everything is ok. Afterwards we hand over the drbd to pacemaker. When we cold boot the two drbd servers the first server comes up with a valid mounted filesystem but whenever we try to migrate or failover we will see the unknown error. The Logfiles give no more hints. cluster was configured using: crm configure <<END property no-quorum-policy=ignore property stonith-enabled=false primitive drbd0 ocf:linbit:drbd \ params drbd_resource=app \ op monitor role=Master interval=59s timeout=30s \ op monitor role=Slave interval=60s timeout=30s ms ms-drbd0 drbd0 \ meta clone-max=2 notify=true globally-unique=false primitive fs0 ocf:heartbeat:Filesystem params fstype=ext3 directory=/app \ device=/dev/drbd/by-res/app commit exit END crm status shows: Stack: openais Current DC: lb01 - partition with quorum Version: 1.1.5-5.el6-01e86afaaa6d4a8c4836f68df80ababd6ca3902f 2 Nodes configured, 2 expected votes 2 Resources configured. ============ Online: [ lb02 lb01 ] Master/Slave Set: ms-drbd0 [drbd0] Masters: [ lb01 ] Slaves: [ lb02 ] Failed actions: fs0_start_0 (node=lb02, call=31, rc=1, status=complete): unknown error fs0_start_0 (node=lb01, call=8, rc=1, status=complete): unknown error We can run the /usr/lib/ocf/resource.d/heartbeat/Filesystem System script with the above parameters on the shell succesfully without getting any errors. The drbd setup is quite easy, uses standard global_common.conf and the following res : resource app { protocol C; startup { wfc-timeout 0; degr-wfc-timeout 120; } disk { on-io-error detach; } syncer { rate 60M; } on lb01.domain.net { device /dev/drbd0; disk /dev/mapper/vg0-app; address 192.168.0.1:7792; meta-disk internal; } on lb02.domain.net { device /dev/drbd0; disk /dev/mapper/vg0-app; address 192.168.0.2:7792; meta-disk internal; } } Any ideas how we can debug that problem ? Christoph -- Christoph Adomeit GATWORKS GmbH Reststrauch 191 41199 Moenchengladbach Sitz: Moenchengladbach Amtsgericht Moenchengladbach, HRB 6303 Geschaeftsfuehrer: Christoph Adomeit, Hans Wilhelm Terstappen Christoph.Adomeit at gatworks.de Internetloesungen vom Feinsten Fon. +49 2166 9149-32 Fax. +49 2166 9149-10