[DRBD-user] Starting up in degraded mode

N.J. van der Horn (Nico) nico at vanderhorn.nl
Sun Feb 10 16:07:25 CET 2008

Note: "permalinks" may not be as permanent as we would like,
direct links of old sources may well be a few messages off.


Hi Gordan,

Just use "drbdadm primary <res>".
When your 2nd node is connected and configured, the sync will start 
instantly.
It is only once needed to include the special syntax "-- 
--overwrite-data-of-peer"
to tell DRBD that you are sure what you are doing.
This status is "remembered", even over reboots.

Greetz, Nico

अनुज Anuj Singh schreef:
> On Feb 8, 2008 9:40 PM,  <drbd at bobich.net> wrote:
>   
>> Hi,
>>
>> I'm building a 2-node cluster, and I only have 1 node built at the moment.
>> I'm planning to use DRBD as a primary-primary shared storage device. When
>> I reboot the machine, the DRBD service starts, but it doesn't activate the
>> resource and create the /dev/drbd1 device node.
>>
>> My config file is below:
>>
>> global { usage-count no; }
>> common { syncer { rate 10M; } }
>>
>> resource r0
>> {
>>          protocol C;
>>          net
>>          {
>>                  cram-hmac-alg sha1;
>>                  shared-secret "password";
>>          }
>>
>>          on server1
>>          {
>>                  device          /dev/drbd1;
>>                  disk            /dev/sda6;
>>                  address         10.0.0.1:7789;
>>                  meta-disk       internal;
>>          }
>>
>>          on server2
>>          {
>>                  device          /dev/drbd1;
>>                  disk            /dev/sda6;
>>                  address         10.0.0.1:7789;
>>                  meta-disk       internal;
>>          }
>>
>> I am using GFS.
>>
>> 1) Would it come up correctly if both peers were available and accessible?
>>     
>
> allow-two-primaries;
>
>   
>> 2) Is there a standard way of ensuring that if the 2nd node is not
>> accessible after some timeout (e.g. 10-30 seconds), the current node sets
>> the local instance as primary and create the device node?
>>
>>     
>
> yes. use heartbeat
>
> http://www.linux-ha.org/GettingStarted
>
>   
>> 3) Is there a sane way to handle the condition where both nodes come up
>> individually and only then the connection is restored? Obviously, the
>> disks would not be consistent, but they would both be working by that
>> point. Resyncing the BD underneath GFS would probably trash whichever
>> node's data is being overwritten. Is there a method available to prevent
>> this split-brain condition? One option I can see is to not sync. GFS would
>> ry to mount, notice the other node up but not using it's journal, and
>> cluster would end up fencing one node. It'd be a race on which one gets
>> fenced, but that isn't a huge problem.
>>
>>     
>
>
>
>
>             after-sb-0pri discard-younger-primary;
>
>
>
>             after-sb-1pri discard-secondary;
>
>
>
>             after-sb-2pri call-pri-lost-after-sb;
>
> go threw man pages.
>
>
>   
>> But first I need to ensure that the local DRBD powers up even if the peer
>> isn't around. Is there a config option for that?
>>
>> Thanks.
>>
>> Gordan
>> _______________________________________________
>> drbd-user mailing list
>> drbd-user at lists.linbit.com
>> http://lists.linbit.com/mailman/listinfo/drbd-user
>>
>>     
> _______________________________________________
> drbd-user mailing list
> drbd-user at lists.linbit.com
> http://lists.linbit.com/mailman/listinfo/drbd-user
>   

-- 
Behandeld door / Handled by: N.J. van der Horn (Nico)
---
ICT Support Vanderhorn IT-works, www.vanderhorn.nl,
Voorstraat 55, 3135 HW Vlaardingen, The Netherlands,
Tel +31 10 2486060, Fax +31 10 2486061





More information about the drbd-user mailing list