Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
Hello Matt, Lars, While installing PaceMaker the "nfslock nfs" solution frm Matt worked when i restart heartbeat when i reboot the server the ifs is still down? (very strange) root at kvmstorage1 ~]# ifconfig eth0 Link encap:Ethernet HWaddr 52:54:00:18:BC:3D inet addr:192.168.123.211 Bcast:192.168.123.255 Mask:255.255.255.0 inet6 addr: fe80::5054:ff:fe18:bc3d/64 Scope:Link UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 RX packets:92869 errors:0 dropped:15 overruns:0 frame:0 TX packets:97696 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:1000 RX bytes:113420015 (108.1 MiB) TX bytes:114931967 (109.6 MiB) Interrupt:10 Base address:0x2000 eth0:0 Link encap:Ethernet HWaddr 52:54:00:18:BC:3D inet addr:192.168.123.209 Bcast:192.168.123.255 Mask:255.255.255.0 UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 Interrupt:10 Base address:0x2000 lo Link encap:Local Loopback inet addr:127.0.0.1 Mask:255.0.0.0 inet6 addr: ::1/128 Scope:Host UP LOOPBACK RUNNING MTU:16436 Metric:1 RX packets:12 errors:0 dropped:0 overruns:0 frame:0 TX packets:12 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:0 RX bytes:840 (840.0 b) TX bytes:840 (840.0 b) [root at kvmstorage1 ~]# service nfs status rpc.svcgssd is stopped rpc.mountd is stopped nfsd is stopped rpc.rquotad (pid 1342) is running... # haresources kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfslock nfs rpcidmapd mysqld # logfile May 16 22:44:39 kvmstorage1.localdomain heartbeat: [8177]: info: All HA resources relinquished. May 16 22:44:40 kvmstorage1.localdomain heartbeat: [6569]: WARN: 1 lost packet(s) for [kvmstorage2.localdomain] [51:53] May 16 22:44:40 kvmstorage1.localdomain heartbeat: [6569]: info: No pkts missing from kvmstorage2.localdomain! May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBFIFO process 6572 with signal 15 May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBWRITE process 6573 with signal 15 May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBREAD process 6574 with signal 15 May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6574 exited. 3 remaining May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6572 exited. 2 remaining May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6573 exited. 1 remaining May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: kvmstorage1.localdomain Heartbeat shutdown complete. May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: Pacemaker support: false May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: WARN: Logging daemon is disabled --enabling logging daemon is recommended May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: ************************** May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: Configuration validated. Starting heartbeat 3.0.4 May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: heartbeat: version 3.0.4 May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Heartbeat generation: 1335851675 May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: glib: UDP Broadcast heartbeat started on port 694 (694) interface eth0 May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: glib: UDP Broadcast heartbeat closed on port 694 interface eth0 - Status: 1 May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_TriggerHandler: Added signal manual handler May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_TriggerHandler: Added signal manual handler May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_SignalHandler: Added signal handler for signal 17 May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Local status now set to: 'up' May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Link kvmstorage1.localdomain:eth0 up. May 16 22:45:24 kvmstorage1.localdomain heartbeat: [1528]: info: Link kvmstorage2.localdomain:eth0 up. May 16 22:45:24 kvmstorage1.localdomain heartbeat: [1528]: info: Status update for node kvmstorage2.localdomain: status active harc[1630]: 2012/05/16_22:45:24 info: Running /etc/ha.d//rc.d/status status May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Comm_now_up(): updating status to active May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Local status now set to: 'active' May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed. May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed. May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Local Resource acquisition completed. (none) May 16 22:45:26 kvmstorage1.localdomain heartbeat: [1528]: info: kvmstorage2.localdomain wants to go standby [foreign] May 16 22:45:31 kvmstorage1.localdomain heartbeat: [1528]: info: standby: acquire [foreign] resources from kvmstorage2.localdomain May 16 22:45:31 kvmstorage1.localdomain heartbeat: [2253]: info: acquire local HA resources (standby). ResourceManager[2275]: 2012/05/16_22:45:31 info: Acquiring resource group: kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfslock nfs rpcidmapd mysqld IPaddr[2321]: 2012/05/16_22:45:32 INFO: Resource is stopped ResourceManager[2275]: 2012/05/16_22:45:32 info: Running /etc/ha.d/resource.d/IPaddr 192.168.123.209/24/eth0 start IPaddr[2439]: 2012/05/16_22:45:32 INFO: Using calculated netmask for 192.168.123.209: 255.255.255.0 IPaddr[2439]: 2012/05/16_22:45:32 INFO: eval ifconfig eth0:0 192.168.123.209 netmask 255.255.255.0 broadcast 192.168.123.255 IPaddr[2402]: 2012/05/16_22:45:32 INFO: Success ResourceManager[2275]: 2012/05/16_22:45:32 info: Running /etc/ha.d/resource.d/drbddisk main start Filesystem[2627]: 2012/05/16_22:45:33 INFO: Resource is stopped ResourceManager[2275]: 2012/05/16_22:45:33 info: Running /etc/ha.d/resource.d/Filesystem /dev/drbd0 /datastore ext4 start Filesystem[2741]: 2012/05/16_22:45:33 INFO: Running start for /dev/drbd0 on /datastore Filesystem[2729]: 2012/05/16_22:45:33 INFO: Success ResourceManager[2275]: 2012/05/16_22:45:33 info: Running /etc/init.d/nfslock start ResourceManager[2275]: 2012/05/16_22:45:34 info: Running /etc/init.d/rpcidmapd start ResourceManager[2275]: 2012/05/16_22:45:35 info: Running /etc/init.d/mysqld start May 16 22:45:36 kvmstorage1.localdomain heartbeat: [2253]: info: local HA resource acquisition completed (standby). May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: Standby resource acquisition done [foreign]. May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: Initial resource acquisition complete (auto_failback) May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed. On 16 mei 2012, at 22:36, Matt Graham wrote: > From: Lars Ellenberg <lars.ellenberg at linbit.com> > [snippage] >> May I ask why you chose to use heartbeat haresource mode instead of >> RHCS or Pacemaker, or any other potential candidate for the job? >> >> Just curious here. I'm trying to figure out how common it is nowadays >> to do new haresources deployments with new projects. > > The last time I tried pacemaker on a test cluster at work, it didn't work at > all. This was apparently because multicast was required, and the people who > run networking at work said they wouldn't turn multicast on for any of the > switches. heartbeat, though, works reasonably well because it doesn't require > multicast. > > FWIW, Marcel's haresources had "nfs nfslock", while a heartbeat-using NFS > cluster here has "nfslock nfs" in the same place. The /etc/init.d/ scripts > have chkconfig comments showing nfslock starting before nfs. So that's > another thing for Marcel to look at. > > -- > Matt G / Dances With Crows > The Crow202 Blog: http://crow202.org/wordpress/ > There is no Darkness in Eternity/But only Light too dim for us to see > > _______________________________________________ > drbd-user mailing list > drbd-user at lists.linbit.com > http://lists.linbit.com/mailman/listinfo/drbd-user