Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.
When i turn "off" the services ifs and nfslock at start chkconfig nfs off chkconfig nfslock off then everything is working. also after reboot. so my problem is solved and i will go into pacemaker next week.. On 16 mei 2012, at 22:48, Marcel Kraan wrote: > Hello Matt, Lars, > > While installing PaceMaker the "nfslock nfs" solution frm Matt worked when i restart heartbeat > when i reboot the server the ifs is still down? (very strange) > > root at kvmstorage1 ~]# ifconfig > eth0 Link encap:Ethernet HWaddr 52:54:00:18:BC:3D > inet addr:192.168.123.211 Bcast:192.168.123.255 Mask:255.255.255.0 > inet6 addr: fe80::5054:ff:fe18:bc3d/64 Scope:Link > UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 > RX packets:92869 errors:0 dropped:15 overruns:0 frame:0 > TX packets:97696 errors:0 dropped:0 overruns:0 carrier:0 > collisions:0 txqueuelen:1000 > RX bytes:113420015 (108.1 MiB) TX bytes:114931967 (109.6 MiB) > Interrupt:10 Base address:0x2000 > > eth0:0 Link encap:Ethernet HWaddr 52:54:00:18:BC:3D > inet addr:192.168.123.209 Bcast:192.168.123.255 Mask:255.255.255.0 > UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 > Interrupt:10 Base address:0x2000 > > lo Link encap:Local Loopback > inet addr:127.0.0.1 Mask:255.0.0.0 > inet6 addr: ::1/128 Scope:Host > UP LOOPBACK RUNNING MTU:16436 Metric:1 > RX packets:12 errors:0 dropped:0 overruns:0 frame:0 > TX packets:12 errors:0 dropped:0 overruns:0 carrier:0 > collisions:0 txqueuelen:0 > RX bytes:840 (840.0 b) TX bytes:840 (840.0 b) > > [root at kvmstorage1 ~]# service nfs status > rpc.svcgssd is stopped > rpc.mountd is stopped > nfsd is stopped > rpc.rquotad (pid 1342) is running... > > > # haresources > kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfslock nfs rpcidmapd mysqld > > # logfile > > May 16 22:44:39 kvmstorage1.localdomain heartbeat: [8177]: info: All HA resources relinquished. > May 16 22:44:40 kvmstorage1.localdomain heartbeat: [6569]: WARN: 1 lost packet(s) for [kvmstorage2.localdomain] [51:53] > May 16 22:44:40 kvmstorage1.localdomain heartbeat: [6569]: info: No pkts missing from kvmstorage2.localdomain! > May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBFIFO process 6572 with signal 15 > May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBWRITE process 6573 with signal 15 > May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBREAD process 6574 with signal 15 > May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6574 exited. 3 remaining > May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6572 exited. 2 remaining > May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6573 exited. 1 remaining > May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: kvmstorage1.localdomain Heartbeat shutdown complete. > May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: Pacemaker support: false > May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: WARN: Logging daemon is disabled --enabling logging daemon is recommended > May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: ************************** > May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: Configuration validated. Starting heartbeat 3.0.4 > May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: heartbeat: version 3.0.4 > May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Heartbeat generation: 1335851675 > May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: glib: UDP Broadcast heartbeat started on port 694 (694) interface eth0 > May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: glib: UDP Broadcast heartbeat closed on port 694 interface eth0 - Status: 1 > May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_TriggerHandler: Added signal manual handler > May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_TriggerHandler: Added signal manual handler > May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_SignalHandler: Added signal handler for signal 17 > May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Local status now set to: 'up' > May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Link kvmstorage1.localdomain:eth0 up. > May 16 22:45:24 kvmstorage1.localdomain heartbeat: [1528]: info: Link kvmstorage2.localdomain:eth0 up. > May 16 22:45:24 kvmstorage1.localdomain heartbeat: [1528]: info: Status update for node kvmstorage2.localdomain: status active > harc[1630]: 2012/05/16_22:45:24 info: Running /etc/ha.d//rc.d/status status > May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Comm_now_up(): updating status to active > May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Local status now set to: 'active' > May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed. > May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed. > May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Local Resource acquisition completed. (none) > May 16 22:45:26 kvmstorage1.localdomain heartbeat: [1528]: info: kvmstorage2.localdomain wants to go standby [foreign] > May 16 22:45:31 kvmstorage1.localdomain heartbeat: [1528]: info: standby: acquire [foreign] resources from kvmstorage2.localdomain > May 16 22:45:31 kvmstorage1.localdomain heartbeat: [2253]: info: acquire local HA resources (standby). > ResourceManager[2275]: 2012/05/16_22:45:31 info: Acquiring resource group: kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfslock nfs rpcidmapd mysqld > IPaddr[2321]: 2012/05/16_22:45:32 INFO: Resource is stopped > ResourceManager[2275]: 2012/05/16_22:45:32 info: Running /etc/ha.d/resource.d/IPaddr 192.168.123.209/24/eth0 start > IPaddr[2439]: 2012/05/16_22:45:32 INFO: Using calculated netmask for 192.168.123.209: 255.255.255.0 > IPaddr[2439]: 2012/05/16_22:45:32 INFO: eval ifconfig eth0:0 192.168.123.209 netmask 255.255.255.0 broadcast 192.168.123.255 > IPaddr[2402]: 2012/05/16_22:45:32 INFO: Success > ResourceManager[2275]: 2012/05/16_22:45:32 info: Running /etc/ha.d/resource.d/drbddisk main start > Filesystem[2627]: 2012/05/16_22:45:33 INFO: Resource is stopped > ResourceManager[2275]: 2012/05/16_22:45:33 info: Running /etc/ha.d/resource.d/Filesystem /dev/drbd0 /datastore ext4 start > Filesystem[2741]: 2012/05/16_22:45:33 INFO: Running start for /dev/drbd0 on /datastore > Filesystem[2729]: 2012/05/16_22:45:33 INFO: Success > ResourceManager[2275]: 2012/05/16_22:45:33 info: Running /etc/init.d/nfslock start > ResourceManager[2275]: 2012/05/16_22:45:34 info: Running /etc/init.d/rpcidmapd start > ResourceManager[2275]: 2012/05/16_22:45:35 info: Running /etc/init.d/mysqld start > May 16 22:45:36 kvmstorage1.localdomain heartbeat: [2253]: info: local HA resource acquisition completed (standby). > May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: Standby resource acquisition done [foreign]. > May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: Initial resource acquisition complete (auto_failback) > May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed. > > > > > > > > > > > > > > > On 16 mei 2012, at 22:36, Matt Graham wrote: > >> From: Lars Ellenberg <lars.ellenberg at linbit.com> >> [snippage] >>> May I ask why you chose to use heartbeat haresource mode instead of >>> RHCS or Pacemaker, or any other potential candidate for the job? >>> >>> Just curious here. I'm trying to figure out how common it is nowadays >>> to do new haresources deployments with new projects. >> >> The last time I tried pacemaker on a test cluster at work, it didn't work at >> all. This was apparently because multicast was required, and the people who >> run networking at work said they wouldn't turn multicast on for any of the >> switches. heartbeat, though, works reasonably well because it doesn't require >> multicast. >> >> FWIW, Marcel's haresources had "nfs nfslock", while a heartbeat-using NFS >> cluster here has "nfslock nfs" in the same place. The /etc/init.d/ scripts >> have chkconfig comments showing nfslock starting before nfs. So that's >> another thing for Marcel to look at. >> >> -- >> Matt G / Dances With Crows >> The Crow202 Blog: http://crow202.org/wordpress/ >> There is no Darkness in Eternity/But only Light too dim for us to see >> >> _______________________________________________ >> drbd-user mailing list >> drbd-user at lists.linbit.com >> http://lists.linbit.com/mailman/listinfo/drbd-user > > _______________________________________________ > drbd-user mailing list > drbd-user at lists.linbit.com > http://lists.linbit.com/mailman/listinfo/drbd-user