[DRBD-user] NFS not starting with heartbeat

Marcel Kraan marcel at kraan.net
Wed May 16 22:48:32 CEST 2012

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


Hello Matt, Lars,

While installing PaceMaker the "nfslock nfs" solution frm Matt worked when i restart heartbeat
when i reboot the server the ifs is still down? (very strange)

root at kvmstorage1 ~]# ifconfig
eth0      Link encap:Ethernet  HWaddr 52:54:00:18:BC:3D  
          inet addr:192.168.123.211  Bcast:192.168.123.255  Mask:255.255.255.0
          inet6 addr: fe80::5054:ff:fe18:bc3d/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:92869 errors:0 dropped:15 overruns:0 frame:0
          TX packets:97696 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:1000 
          RX bytes:113420015 (108.1 MiB)  TX bytes:114931967 (109.6 MiB)
          Interrupt:10 Base address:0x2000 

eth0:0    Link encap:Ethernet  HWaddr 52:54:00:18:BC:3D  
          inet addr:192.168.123.209  Bcast:192.168.123.255  Mask:255.255.255.0
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          Interrupt:10 Base address:0x2000 

lo        Link encap:Local Loopback  
          inet addr:127.0.0.1  Mask:255.0.0.0
          inet6 addr: ::1/128 Scope:Host
          UP LOOPBACK RUNNING  MTU:16436  Metric:1
          RX packets:12 errors:0 dropped:0 overruns:0 frame:0
          TX packets:12 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0 
          RX bytes:840 (840.0 b)  TX bytes:840 (840.0 b)

[root at kvmstorage1 ~]# service nfs status
rpc.svcgssd is stopped
rpc.mountd is stopped
nfsd is stopped
rpc.rquotad (pid 1342) is running...


# haresources
kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfslock nfs rpcidmapd mysqld

# logfile

May 16 22:44:39 kvmstorage1.localdomain heartbeat: [8177]: info: All HA resources relinquished.
May 16 22:44:40 kvmstorage1.localdomain heartbeat: [6569]: WARN: 1 lost packet(s) for [kvmstorage2.localdomain] [51:53]
May 16 22:44:40 kvmstorage1.localdomain heartbeat: [6569]: info: No pkts missing from kvmstorage2.localdomain!
May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBFIFO process 6572 with signal 15
May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBWRITE process 6573 with signal 15
May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBREAD process 6574 with signal 15
May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6574 exited. 3 remaining
May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6572 exited. 2 remaining
May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6573 exited. 1 remaining
May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: kvmstorage1.localdomain Heartbeat shutdown complete.
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: Pacemaker support: false
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: WARN: Logging daemon is disabled --enabling logging daemon is recommended
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: **************************
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: Configuration validated. Starting heartbeat 3.0.4
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: heartbeat: version 3.0.4
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Heartbeat generation: 1335851675
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: glib: UDP Broadcast heartbeat started on port 694 (694) interface eth0
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: glib: UDP Broadcast heartbeat closed on port 694 interface eth0 - Status: 1
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_TriggerHandler: Added signal manual handler
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_TriggerHandler: Added signal manual handler
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_SignalHandler: Added signal handler for signal 17
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Local status now set to: 'up'
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Link kvmstorage1.localdomain:eth0 up.
May 16 22:45:24 kvmstorage1.localdomain heartbeat: [1528]: info: Link kvmstorage2.localdomain:eth0 up.
May 16 22:45:24 kvmstorage1.localdomain heartbeat: [1528]: info: Status update for node kvmstorage2.localdomain: status active
harc[1630]:	2012/05/16_22:45:24 info: Running /etc/ha.d//rc.d/status status
May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Comm_now_up(): updating status to active
May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Local status now set to: 'active'
May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed.
May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed.
May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Local Resource acquisition completed. (none)
May 16 22:45:26 kvmstorage1.localdomain heartbeat: [1528]: info: kvmstorage2.localdomain wants to go standby [foreign]
May 16 22:45:31 kvmstorage1.localdomain heartbeat: [1528]: info: standby: acquire [foreign] resources from kvmstorage2.localdomain
May 16 22:45:31 kvmstorage1.localdomain heartbeat: [2253]: info: acquire local HA resources (standby).
ResourceManager[2275]:	2012/05/16_22:45:31 info: Acquiring resource group: kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfslock nfs rpcidmapd mysqld
IPaddr[2321]:	2012/05/16_22:45:32 INFO:  Resource is stopped
ResourceManager[2275]:	2012/05/16_22:45:32 info: Running /etc/ha.d/resource.d/IPaddr 192.168.123.209/24/eth0 start
IPaddr[2439]:	2012/05/16_22:45:32 INFO: Using calculated netmask for 192.168.123.209: 255.255.255.0
IPaddr[2439]:	2012/05/16_22:45:32 INFO: eval ifconfig eth0:0 192.168.123.209 netmask 255.255.255.0 broadcast 192.168.123.255
IPaddr[2402]:	2012/05/16_22:45:32 INFO:  Success
ResourceManager[2275]:	2012/05/16_22:45:32 info: Running /etc/ha.d/resource.d/drbddisk main start
Filesystem[2627]:	2012/05/16_22:45:33 INFO:  Resource is stopped
ResourceManager[2275]:	2012/05/16_22:45:33 info: Running /etc/ha.d/resource.d/Filesystem /dev/drbd0 /datastore ext4 start
Filesystem[2741]:	2012/05/16_22:45:33 INFO: Running start for /dev/drbd0 on /datastore
Filesystem[2729]:	2012/05/16_22:45:33 INFO:  Success
ResourceManager[2275]:	2012/05/16_22:45:33 info: Running /etc/init.d/nfslock  start
ResourceManager[2275]:	2012/05/16_22:45:34 info: Running /etc/init.d/rpcidmapd  start
ResourceManager[2275]:	2012/05/16_22:45:35 info: Running /etc/init.d/mysqld  start
May 16 22:45:36 kvmstorage1.localdomain heartbeat: [2253]: info: local HA resource acquisition completed (standby).
May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: Standby resource acquisition done [foreign].
May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: Initial resource acquisition complete (auto_failback)
May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed.














On 16 mei 2012, at 22:36, Matt Graham wrote:

> From: Lars Ellenberg <lars.ellenberg at linbit.com>
> [snippage]
>> May I ask why you chose to use heartbeat haresource mode instead of
>> RHCS or Pacemaker, or any other potential candidate for the job?
>> 
>> Just curious here. I'm trying to figure out how common it is nowadays
>> to do new haresources deployments with new projects.
> 
> The last time I tried pacemaker on a test cluster at work, it didn't work at
> all.  This was apparently because multicast was required, and the people who
> run networking at work said they wouldn't turn multicast on for any of the
> switches.  heartbeat, though, works reasonably well because it doesn't require
> multicast.
> 
> FWIW, Marcel's haresources had "nfs nfslock", while a heartbeat-using NFS
> cluster here has "nfslock nfs" in the same place.  The /etc/init.d/ scripts
> have chkconfig comments showing nfslock starting before nfs.  So that's
> another thing for Marcel to look at.
> 
> -- 
> Matt G / Dances With Crows
> The Crow202 Blog:  http://crow202.org/wordpress/
> There is no Darkness in Eternity/But only Light too dim for us to see
> 
> _______________________________________________
> drbd-user mailing list
> drbd-user at lists.linbit.com
> http://lists.linbit.com/mailman/listinfo/drbd-user




More information about the drbd-user mailing list