[DRBD-user] OCFS2 over DRBDv8

Rene Mayrhofer rene.mayrhofer at gibraltar.at
Tue Feb 27 14:49:09 CET 2007

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


Hi all,

[Please CC me in replies, I am not subscribed to this list.]

I am trying to to use ocfs2 over drbd8 and am failing right now with seemingly 
the same error as described in 
http://lists.linbit.com/pipermail/drbd-user/2006-August/005483.html

Packages are from Debian etch (drbd8-utils 2:8.0.0-1 with the modules compiled 
from drbd8-module-source 8.0.0-1, ocfs2-tools 1.2.1-1.3 with kernel modules 
from the Debian kernel package linux-image-2.6.18-4-xen-686 
2.6.18.dfsg.1-11). Kernel command line from /proc/cmdline is more or less 
boring:

[root at jupiter2 ~]# cat /proc/cmdline
root=/dev/md1 ro selinux=1 enforcing=0 irqpoll console=tty0

With this setup, I get

[root at jupiter ~]# mount -t ocfs2 /dev/drbd0 /home/
ocfs2_hb_ctl: I/O error on channel while starting heartbeat
mount.ocfs2: Error when attempting to run /sbin/ocfs2_hb_ctl: "Operation not 
permitted"

and dmesg then tells me

(21368,0):o2hb_setup_one_bio:290 ERROR: Error adding page to bio i = 7, 
vec_len = 4096, len = 0
, start = 0
(21368,0):o2hb_read_slots:385 ERROR: status = -5
(21368,0):o2hb_populate_slot_data:1299 ERROR: status = -5
(21368,0):o2hb_region_dev_write:1399 ERROR: status = -5

on both nodes of the cluster. drbd8 seems to be up correctly:

[root at jupiter2 ~]# cat /proc/drbd
version: 8.0.0 (api:86/proto:86)
SVN Revision: 2713 build by rene at moon, 2007-02-26 14:34:42
 0: cs:Connected st:Primary/Primary ds:UpToDate/UpToDate C r---
    ns:492376 nr:0 dw:492376 dr:3107 al:150 bm:0 lo:0 pe:0 ua:0 ap:0
        resync: used:0/31 hits:0 misses:0 starving:0 dirty:0 changed:0
        act_log: used:0/257 hits:15922 misses:150 starving:0 dirty:0 
changed:150

and the ocfs2 configuration is pretty simple:

[root at jupiter2 ~]# cat /etc/ocfs2/cluster.conf
cluster:
        node_count = 2
        name = ocfs2

node:
        ip_port = 7777
        ip_address = 192.168.255.30
        number = 0
        name = jupiter
        cluster = ocfs2

node:
        ip_port = 7777
        ip_address = 192.168.255.29
        number = 1
        name = jupiter2
        cluster = ocfs2

(and copied to both hosts). o2cb seems to be fine as well (equal on both 
nodes)

[root at jupiter2 ~]# /etc/init.d/o2cb status
Module "configfs": Loaded
Filesystem "configfs": Mounted
Module "ocfs2_nodemanager": Loaded
Module "ocfs2_dlm": Loaded
Module "ocfs2_dlmfs": Loaded
Filesystem "ocfs2_dlmfs": Mounted
Checking cluster ocfs2: Online
Checking heartbeat: Not active

mkfs.ocfs2 has been called without specific options for cluster block sizes, 
so it used the default of 4k. Underlying /dev/drbd0 are software RAID1 
devices (which work flawlessly).

Has this issue seen any solution since last year? And help is greatly 
appreciated!

with best regards,
Rene

-- 
-------------------------------------------------
Gibraltar firewall       http://www.gibraltar.at/
-------------- next part --------------
A non-text attachment was scrubbed...
Name: not available
Type: application/pgp-signature
Size: 189 bytes
Desc: not available
URL: <http://lists.linbit.com/pipermail/drbd-user/attachments/20070227/ef607538/attachment.pgp>


More information about the drbd-user mailing list