[DRBD-user] NFS not starting with heartbeat

Marcel Kraan marcel at kraan.net
Wed May 16 23:37:29 CEST 2012

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


When i turn "off" the services ifs and nfslock at start
chkconfig nfs off
chkconfig nfslock off

then everything is working. also after reboot.
so my problem is solved and i will go into pacemaker next week..



On 16 mei 2012, at 22:48, Marcel Kraan wrote:

> Hello Matt, Lars,
> 
> While installing PaceMaker the "nfslock nfs" solution frm Matt worked when i restart heartbeat
> when i reboot the server the ifs is still down? (very strange)
> 
> root at kvmstorage1 ~]# ifconfig
> eth0      Link encap:Ethernet  HWaddr 52:54:00:18:BC:3D  
>          inet addr:192.168.123.211  Bcast:192.168.123.255  Mask:255.255.255.0
>          inet6 addr: fe80::5054:ff:fe18:bc3d/64 Scope:Link
>          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>          RX packets:92869 errors:0 dropped:15 overruns:0 frame:0
>          TX packets:97696 errors:0 dropped:0 overruns:0 carrier:0
>          collisions:0 txqueuelen:1000 
>          RX bytes:113420015 (108.1 MiB)  TX bytes:114931967 (109.6 MiB)
>          Interrupt:10 Base address:0x2000 
> 
> eth0:0    Link encap:Ethernet  HWaddr 52:54:00:18:BC:3D  
>          inet addr:192.168.123.209  Bcast:192.168.123.255  Mask:255.255.255.0
>          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>          Interrupt:10 Base address:0x2000 
> 
> lo        Link encap:Local Loopback  
>          inet addr:127.0.0.1  Mask:255.0.0.0
>          inet6 addr: ::1/128 Scope:Host
>          UP LOOPBACK RUNNING  MTU:16436  Metric:1
>          RX packets:12 errors:0 dropped:0 overruns:0 frame:0
>          TX packets:12 errors:0 dropped:0 overruns:0 carrier:0
>          collisions:0 txqueuelen:0 
>          RX bytes:840 (840.0 b)  TX bytes:840 (840.0 b)
> 
> [root at kvmstorage1 ~]# service nfs status
> rpc.svcgssd is stopped
> rpc.mountd is stopped
> nfsd is stopped
> rpc.rquotad (pid 1342) is running...
> 
> 
> # haresources
> kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfslock nfs rpcidmapd mysqld
> 
> # logfile
> 
> May 16 22:44:39 kvmstorage1.localdomain heartbeat: [8177]: info: All HA resources relinquished.
> May 16 22:44:40 kvmstorage1.localdomain heartbeat: [6569]: WARN: 1 lost packet(s) for [kvmstorage2.localdomain] [51:53]
> May 16 22:44:40 kvmstorage1.localdomain heartbeat: [6569]: info: No pkts missing from kvmstorage2.localdomain!
> May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBFIFO process 6572 with signal 15
> May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBWRITE process 6573 with signal 15
> May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBREAD process 6574 with signal 15
> May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6574 exited. 3 remaining
> May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6572 exited. 2 remaining
> May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6573 exited. 1 remaining
> May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: kvmstorage1.localdomain Heartbeat shutdown complete.
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: Pacemaker support: false
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: WARN: Logging daemon is disabled --enabling logging daemon is recommended
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: **************************
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: Configuration validated. Starting heartbeat 3.0.4
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: heartbeat: version 3.0.4
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Heartbeat generation: 1335851675
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: glib: UDP Broadcast heartbeat started on port 694 (694) interface eth0
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: glib: UDP Broadcast heartbeat closed on port 694 interface eth0 - Status: 1
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_TriggerHandler: Added signal manual handler
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_TriggerHandler: Added signal manual handler
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_SignalHandler: Added signal handler for signal 17
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Local status now set to: 'up'
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Link kvmstorage1.localdomain:eth0 up.
> May 16 22:45:24 kvmstorage1.localdomain heartbeat: [1528]: info: Link kvmstorage2.localdomain:eth0 up.
> May 16 22:45:24 kvmstorage1.localdomain heartbeat: [1528]: info: Status update for node kvmstorage2.localdomain: status active
> harc[1630]:	2012/05/16_22:45:24 info: Running /etc/ha.d//rc.d/status status
> May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Comm_now_up(): updating status to active
> May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Local status now set to: 'active'
> May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed.
> May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed.
> May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Local Resource acquisition completed. (none)
> May 16 22:45:26 kvmstorage1.localdomain heartbeat: [1528]: info: kvmstorage2.localdomain wants to go standby [foreign]
> May 16 22:45:31 kvmstorage1.localdomain heartbeat: [1528]: info: standby: acquire [foreign] resources from kvmstorage2.localdomain
> May 16 22:45:31 kvmstorage1.localdomain heartbeat: [2253]: info: acquire local HA resources (standby).
> ResourceManager[2275]:	2012/05/16_22:45:31 info: Acquiring resource group: kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfslock nfs rpcidmapd mysqld
> IPaddr[2321]:	2012/05/16_22:45:32 INFO:  Resource is stopped
> ResourceManager[2275]:	2012/05/16_22:45:32 info: Running /etc/ha.d/resource.d/IPaddr 192.168.123.209/24/eth0 start
> IPaddr[2439]:	2012/05/16_22:45:32 INFO: Using calculated netmask for 192.168.123.209: 255.255.255.0
> IPaddr[2439]:	2012/05/16_22:45:32 INFO: eval ifconfig eth0:0 192.168.123.209 netmask 255.255.255.0 broadcast 192.168.123.255
> IPaddr[2402]:	2012/05/16_22:45:32 INFO:  Success
> ResourceManager[2275]:	2012/05/16_22:45:32 info: Running /etc/ha.d/resource.d/drbddisk main start
> Filesystem[2627]:	2012/05/16_22:45:33 INFO:  Resource is stopped
> ResourceManager[2275]:	2012/05/16_22:45:33 info: Running /etc/ha.d/resource.d/Filesystem /dev/drbd0 /datastore ext4 start
> Filesystem[2741]:	2012/05/16_22:45:33 INFO: Running start for /dev/drbd0 on /datastore
> Filesystem[2729]:	2012/05/16_22:45:33 INFO:  Success
> ResourceManager[2275]:	2012/05/16_22:45:33 info: Running /etc/init.d/nfslock  start
> ResourceManager[2275]:	2012/05/16_22:45:34 info: Running /etc/init.d/rpcidmapd  start
> ResourceManager[2275]:	2012/05/16_22:45:35 info: Running /etc/init.d/mysqld  start
> May 16 22:45:36 kvmstorage1.localdomain heartbeat: [2253]: info: local HA resource acquisition completed (standby).
> May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: Standby resource acquisition done [foreign].
> May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: Initial resource acquisition complete (auto_failback)
> May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed.
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> 
> On 16 mei 2012, at 22:36, Matt Graham wrote:
> 
>> From: Lars Ellenberg <lars.ellenberg at linbit.com>
>> [snippage]
>>> May I ask why you chose to use heartbeat haresource mode instead of
>>> RHCS or Pacemaker, or any other potential candidate for the job?
>>> 
>>> Just curious here. I'm trying to figure out how common it is nowadays
>>> to do new haresources deployments with new projects.
>> 
>> The last time I tried pacemaker on a test cluster at work, it didn't work at
>> all.  This was apparently because multicast was required, and the people who
>> run networking at work said they wouldn't turn multicast on for any of the
>> switches.  heartbeat, though, works reasonably well because it doesn't require
>> multicast.
>> 
>> FWIW, Marcel's haresources had "nfs nfslock", while a heartbeat-using NFS
>> cluster here has "nfslock nfs" in the same place.  The /etc/init.d/ scripts
>> have chkconfig comments showing nfslock starting before nfs.  So that's
>> another thing for Marcel to look at.
>> 
>> -- 
>> Matt G / Dances With Crows
>> The Crow202 Blog:  http://crow202.org/wordpress/
>> There is no Darkness in Eternity/But only Light too dim for us to see
>> 
>> _______________________________________________
>> drbd-user mailing list
>> drbd-user at lists.linbit.com
>> http://lists.linbit.com/mailman/listinfo/drbd-user
> 
> _______________________________________________
> drbd-user mailing list
> drbd-user at lists.linbit.com
> http://lists.linbit.com/mailman/listinfo/drbd-user




More information about the drbd-user mailing list