[prev in list] [next in list] [prev in thread] [next in thread]
List: redhat-linux-cluster
Subject: Re: [Linux-cluster] Linux-cluster Digest, Vol 73, Issue 15
From: parshuram prasad <parshu001 () gmail ! com>
Date: 2010-05-18 5:40:56
Message-ID: AANLkTilOp8-ZHTefEZn4CCqZIAEqtLyQRYixBP-4vfDm () mail ! gmail ! com
[Download RAW message or body]
[Attachment #2 (multipart/alternative)]
please send me cluster script . i want to create two node clustering on
linux 5.3
thx
parshuram
On Sat, May 15, 2010 at 6:57 PM, <linux-cluster-request@redhat.com> wrote:
> Send Linux-cluster mailing list submissions to
> linux-cluster@redhat.com
>
> To subscribe or unsubscribe via the World Wide Web, visit
> https://www.redhat.com/mailman/listinfo/linux-cluster
> or, via email, send a message with subject or body 'help' to
> linux-cluster-request@redhat.com
>
> You can reach the person managing the list at
> linux-cluster-owner@redhat.com
>
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of Linux-cluster digest..."
>
>
> Today's Topics:
>
> 1. GFS on Debian Lenny (Brent Clark)
> 2. pull plug on node, service never relocates (Dusty)
> 3. Re: GFS on Debian Lenny (Joao Ferreira gmail)
> 4. Re: pull plug on node, service never relocates (Corey Kovacs)
> 5. Re: pull plug on node, service never relocates (Kit Gerrits)
>
>
> ----------------------------------------------------------------------
>
> Message: 1
> Date: Fri, 14 May 2010 20:26:46 +0200
> From: Brent Clark <brentgclarklist@gmail.com>
> To: linux clustering <linux-cluster@redhat.com>
> Subject: [Linux-cluster] GFS on Debian Lenny
> Message-ID: <4BED95E6.4040006@gmail.com>
> Content-Type: text/plain; charset=ISO-8859-1; format=flowed
>
> Hiya
>
> Im trying to get GFS working on Debian Lenny. Unfortuantely
> documentation seems to be non existent. And the one site that google
> recommends, gcharriere.com, is down.
>
> I used googles caching mechanism to try and make head and tails of whats
> needed to be done, but unfortunately Im unsuccessful.
>
> Would anyone have any documentation or any sites or if you have a heart,
> provide a howto to get GFS working.
>
> From myside, all ive done is:
>
> aptitude install gfs2-tools
> modprobe gfs2
> gfs_mkfs -p lock_dlm -t lolcats:drbdtest /dev/drbd0 -j 2
>
> thats all, ive done. No editting of configs etc.
>
> When I try,
>
> mount -t gfs2 /dev/drbd0 /drbd/
>
> I get the following message:
>
> /sbin/mount.gfs2: can't connect to gfs_controld: Connection refused
> /sbin/mount.gfs2: gfs_controld not running
> /sbin/mount.gfs2: error mounting lockproto lock_dlm
>
> If anyone can help, it would be appreciated.
>
> Kind Regards
> Brent Clark
>
>
>
> ------------------------------
>
> Message: 2
> Date: Fri, 14 May 2010 14:45:11 -0500
> From: Dusty <dhoffutt@gmail.com>
> To: Linux-cluster@redhat.com
> Subject: [Linux-cluster] pull plug on node, service never relocates
> Message-ID:
> <AANLkTil1ssNgEYRs71I_xmsLV3enagF76kEQYAt-Tdse@mail.gmail.com>
> Content-Type: text/plain; charset="iso-8859-1"
>
> Greetings,
>
> Using stock "clustering" and "cluster-storage" from RHEL5 update 4 X86_64
> ISO.
>
> As an example using my below config:
>
> Node1 is running service1, node2 is running service2, etc, etc, node5 is
> spare and available for the relocation of any failover domain / cluster
> service.
>
> If I go into the APC PDU and turn off the electrical port to node1, node2
> will fence node1 (going into the APC PDU and doing and off, on on node1's
> port), this is fine. Works well. When node1 comes back up, then it shuts
> down service1 and service1 relocates to node5.
>
> Now if I go in the lab and literally pull the plug on node5 running
> service1, another node fences node5 via the APC - can check the APC PDU log
> and see that it has done an off/on on node5's electrical port just fine.
>
> But I pulled the plug on node5 - resetting the power doesn't matter. I want
> to simulate a completely dead node, and have the service relocate in this
> case of complete node failure.
>
> In this RHEL5.4 cluster, the service never relocates. I can similate this
> on
> any node for any service. What if a node's motherboard fries?
>
> What can I set to have the remaining nodes stop waiting for the reboot of a
> failed node and just go ahead and relocate the cluster service that had
> been
> running on the now failed node?
>
> Thank you!
>
> versions:
>
> cman-2.0.115-1.el5
> openais-0.80.6-8.el5
> modcluster-0.12.1-2.el5
> lvm2-cluster-2.02.46-8.el5
> rgmanager-2.0.52-1.el5
> ricci-0.12.2-6.el5
>
> cluster.conf (sanitized, real scripts removed, all gfs2 mounts gone for
> clarity):
> <?xml version="1.0"?>
> <cluster config_version="1"
> name="alderaanDefenseShieldRebelAllianceCluster">
> <fence_daemon clean_start="0" post_fail_delay="3" post_join_delay="60"/>
> <clusternodes>
> <clusternode name="192.168.1.1" nodeid="1" votes="1">
> <fence>
> <method name="1">
> <device name="apc_pdu" port="1" switch="1"/>
> </method>
> </fence>
> </clusternode>
> <clusternode name="192.168.1.2" nodeid="2" votes="1">
> <fence>
> <method name="1">
> <device name="apc_pdu" port="2" switch="1"/>
> </method>
> </fence>
> </clusternode>
> <clusternode name="192.168.1.3" nodeid="3" votes="1">
> <fence>
> <method name="1">
> <device name="apc_pdu" port="3" switch="1"/>
> </method>
> </fence>
> </clusternode>
> <clusternode name="192.168.1.4" nodeid="4" votes="1">
> <fence>
> <method name="1">
> <device name="apc_pdu" port="4" switch="1"/>
> </method>
> </fence>
> </clusternode>
> <clusternode name="192.168.1.5" nodeid="5" votes="1">
> <fence>
> <method name="1">
> <device name="apc_pdu" port="5" switch="1"/>
> </method>
> </fence>
> </clusternode>
> </clusternodes>
> <cman expected_votes="6"/>
> <fencedevices>
> <fencedevice agent="fence_apc" ipaddr="192.168.1.20" login="device"
> name="apc_pdu" passwd="wonderwomanWasAPrettyCoolSuperhero"/>
> </fencedevices>
> <rm>
> <failoverdomains>
> <failoverdomain name="fd1" nofailback="0" ordered="1"
> restricted="1">
> <failoverdomainnode name="192.168.1.1" priority="1"/>
> <failoverdomainnode name="192.168.1.2" priority="2"/>
> <failoverdomainnode name="192.168.1.3" priority="3"/>
> <failoverdomainnode name="192.168.1.4" priority="4"/>
> <failoverdomainnode name="192.168.1.5" priority="5"/>
> </failoverdomain>
> <failoverdomain name="fd2" nofailback="0" ordered="1"
> restricted="1">
> <failoverdomainnode name="192.168.1.1" priority="5"/>
> <failoverdomainnode name="192.168.1.2" priority="1"/>
> <failoverdomainnode name="192.168.1.3" priority="2"/>
> <failoverdomainnode name="192.168.1.4" priority="3"/>
> <failoverdomainnode name="192.168.1.5" priority="4"/>
> </failoverdomain>
> <failoverdomain name="fd3" nofailback="0" ordered="1"
> restricted="1">
> <failoverdomainnode name="192.168.1.1" priority="4"/>
> <failoverdomainnode name="192.168.1.2" priority="5"/>
> <failoverdomainnode name="192.168.1.3" priority="1"/>
> <failoverdomainnode name="192.168.1.4" priority="2"/>
> <failoverdomainnode name="192.168.1.5" priority="3"/>
> </failoverdomain>
> <failoverdomain name="fd4" nofailback="0" ordered="1"
> restricted="1">
> <failoverdomainnode name="192.168.1.1" priority="3"/>
> <failoverdomainnode name="192.168.1.2" priority="4"/>
> <failoverdomainnode name="192.168.1.3" priority="5"/>
> <failoverdomainnode name="192.168.1.4" priority="1"/>
> <failoverdomainnode name="192.168.1.5" priority="2"/>
> </failoverdomain>
> </failoverdomains>
> <resources>
> <ip address="10.1.1.1" monitor_link="1"/>
> <ip address="10.1.1.2" monitor_link="1"/>
> <ip address="10.1.1.3" monitor_link="1"/>
> <ip address="10.1.1.4" monitor_link="1"/>
> <ip address="10.1.1.5" monitor_link="1"/>
> <script file="/usr/local/bin/service1" name="service1"/>
> <script file="/usr/local/bin/service2" name="service2"/>
> <script file="/usr/local/bin/service3" name="service3"/>
> <script file="/usr/local/bin/service4" name="service4"/>
> </resources>
> <service autostart="1" domain="fd1" exclusive="1" name="service1"
> recovery="relocate">
> <ip ref="10.1.1.1"/>
> <script ref="service1"/>
> </service>
> <service autostart="1" domain="fd2" exclusive="1" name="service2"
> recovery="relocate">
> <ip ref="10.1.1.2"/>
> <script ref="service2"/>
> </service>
> <service autostart="1" domain="fd3" exclusive="1" name="service3"
> recovery="relocate">
> <ip ref="10.1.1.3"/>
> <script ref="service3"/>
> </service>
> <service autostart="1" domain="fd4" exclusive="1" name="service4"
> recovery="relocate">
> <ip ref="10.1.1.4"/>
> <script ref="service4"/>
> </service>
> </rm>
> </cluster>
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: <
> https://www.redhat.com/archives/linux-cluster/attachments/20100514/c892bf86/attachment.html
> >
>
> ------------------------------
>
> Message: 3
> Date: Fri, 14 May 2010 23:31:41 +0100
> From: Joao Ferreira gmail <joao.miguel.c.ferreira@gmail.com>
> To: linux clustering <linux-cluster@redhat.com>
> Subject: Re: [Linux-cluster] GFS on Debian Lenny
> Message-ID: <1273876301.5298.1.camel@debj5n.critical.pt>
> Content-Type: text/plain
>
> Have you checked the docs at the drbd site ?
>
> it contains some short info regarding usage of gsf over drbd..
>
> http://www.drbd.org/docs/applications/
>
> cheers
> Joao
>
> On Fri, 2010-05-14 at 20:26 +0200, Brent Clark wrote:
> > Hiya
> >
> > Im trying to get GFS working on Debian Lenny. Unfortuantely
> > documentation seems to be non existent. And the one site that google
> > recommends, gcharriere.com, is down.
> >
> > I used googles caching mechanism to try and make head and tails of whats
> > needed to be done, but unfortunately Im unsuccessful.
> >
> > Would anyone have any documentation or any sites or if you have a heart,
> > provide a howto to get GFS working.
> >
> > From myside, all ive done is:
> >
> > aptitude install gfs2-tools
> > modprobe gfs2
> > gfs_mkfs -p lock_dlm -t lolcats:drbdtest /dev/drbd0 -j 2
> >
> > thats all, ive done. No editting of configs etc.
> >
> > When I try,
> >
> > mount -t gfs2 /dev/drbd0 /drbd/
> >
> > I get the following message:
> >
> > /sbin/mount.gfs2: can't connect to gfs_controld: Connection refused
> > /sbin/mount.gfs2: gfs_controld not running
> > /sbin/mount.gfs2: error mounting lockproto lock_dlm
> >
> > If anyone can help, it would be appreciated.
> >
> > Kind Regards
> > Brent Clark
> >
> > --
> > Linux-cluster mailing list
> > Linux-cluster@redhat.com
> > https://www.redhat.com/mailman/listinfo/linux-cluster
>
>
>
> ------------------------------
>
> Message: 4
> Date: Sat, 15 May 2010 04:59:23 +0100
> From: Corey Kovacs <corey.kovacs@gmail.com>
> To: linux clustering <linux-cluster@redhat.com>
> Subject: Re: [Linux-cluster] pull plug on node, service never
> relocates
> Message-ID:
> <AANLkTinYVvrit1oPb76TfLa9vmp1AMHcGI3eoZALHxrJ@mail.gmail.com>
> Content-Type: text/plain; charset=ISO-8859-1
>
> What happens when you do ...
>
> fence_node 192.168.1.4
>
> from any of the other nodes?
>
> if that doesn't work, then fencing is not configured correctly and you
> should try to invoke the fence agent directly.
> Also, it would help if you included the APC model and firmware rev.
> The fence_apc agent can be finicky about such things.
>
>
> Hope this helps.
>
> -Core
>
> On Fri, May 14, 2010 at 8:45 PM, Dusty <dhoffutt@gmail.com> wrote:
> > Greetings,
> >
> > Using stock "clustering" and "cluster-storage" from RHEL5 update 4 X86_64
> > ISO.
> >
> > As an example using my below config:
> >
> > Node1 is running service1, node2 is running service2, etc, etc, node5 is
> > spare and available for the relocation of any failover domain / cluster
> > service.
> >
> > If I go into the APC PDU and turn off the electrical port to node1, node2
> > will fence node1 (going into the APC PDU and doing and off, on on node1's
> > port), this is fine. Works well. When node1 comes back up, then it shuts
> > down service1 and service1 relocates to node5.
> >
> > Now if I go in the lab and literally pull the plug on node5 running
> > service1, another node fences node5 via the APC - can check the APC PDU
> log
> > and see that it has done an off/on on node5's electrical port just fine.
> >
> > But I pulled the plug on node5 - resetting the power doesn't matter. I
> want
> > to simulate a completely dead node, and have the service relocate in this
> > case of complete node failure.
> >
> > In this RHEL5.4 cluster, the service never relocates. I can similate this
> on
> > any node for any service. What if a node's motherboard fries?
> >
> > What can I set to have the remaining nodes stop waiting for the reboot of
> a
> > failed node and just go ahead and relocate the cluster service that had
> been
> > running on the now failed node?
> >
> > Thank you!
> >
> > versions:
> >
> > cman-2.0.115-1.el5
> > openais-0.80.6-8.el5
> > modcluster-0.12.1-2.el5
> > lvm2-cluster-2.02.46-8.el5
> > rgmanager-2.0.52-1.el5
> > ricci-0.12.2-6.el5
> >
> > cluster.conf (sanitized, real scripts removed, all gfs2 mounts gone for
> > clarity):
> > <?xml version="1.0"?>
> > <cluster config_version="1"
> > name="alderaanDefenseShieldRebelAllianceCluster">
> > ??? <fence_daemon clean_start="0" post_fail_delay="3"
> post_join_delay="60"/>
> > ??? <clusternodes>
> > ??????? <clusternode name="192.168.1.1" nodeid="1" votes="1">
> > ??????????? <fence>
> > ??????????????? <method name="1">
> > ??????????????????? <device name="apc_pdu" port="1" switch="1"/>
> > ??????????????? </method>
> > ??????????? </fence>
> > ??????? </clusternode>
> > ??????? <clusternode name="192.168.1.2" nodeid="2" votes="1">
> > ??????????? <fence>
> > ??????????????? <method name="1">
> > ??????????????????? <device name="apc_pdu" port="2" switch="1"/>
> > ??????????????? </method>
> > ??????????? </fence>
> > ??????? </clusternode>
> > ??????? <clusternode name="192.168.1.3" nodeid="3" votes="1">
> > ??????????? <fence>
> > ??????????????? <method name="1">
> > ??????????????????? <device name="apc_pdu" port="3" switch="1"/>
> > ??????????????? </method>
> > ??????????? </fence>
> > ??????? </clusternode>
> > ??????? <clusternode name="192.168.1.4" nodeid="4" votes="1">
> > ??????????? <fence>
> > ??????????????? <method name="1">
> > ??????????????????? <device name="apc_pdu" port="4" switch="1"/>
> > ??????????????? </method>
> > ??????????? </fence>
> > ??????? </clusternode>
> > ??????? <clusternode name="192.168.1.5" nodeid="5" votes="1">
> > ??????????? <fence>
> > ??????????????? <method name="1">
> > ??????????????????? <device name="apc_pdu" port="5" switch="1"/>
> > ??????????????? </method>
> > ??????????? </fence>
> > ??????? </clusternode>
> > ??? </clusternodes>
> > ??? <cman expected_votes="6"/>
> > ??? <fencedevices>
> > ??????? <fencedevice agent="fence_apc" ipaddr="192.168.1.20"
> login="device"
> > name="apc_pdu" passwd="wonderwomanWasAPrettyCoolSuperhero"/>
> > ??? </fencedevices>
> > ??? <rm>
> > ??????? <failoverdomains>
> > ??????????? <failoverdomain name="fd1" nofailback="0" ordered="1"
> > restricted="1">
> > ??????????????? <failoverdomainnode name="192.168.1.1" priority="1"/>
> > ??????????????? <failoverdomainnode name="192.168.1.2" priority="2"/>
> > ??????????????? <failoverdomainnode name="192.168.1.3" priority="3"/>
> > ??????????????? <failoverdomainnode name="192.168.1.4" priority="4"/>
> > ??????????????? <failoverdomainnode name="192.168.1.5" priority="5"/>
> > ??????????? </failoverdomain>
> > ??????????? <failoverdomain name="fd2" nofailback="0" ordered="1"
> > restricted="1">
> > ??????????????? <failoverdomainnode name="192.168.1.1" priority="5"/>
> > ??????????????? <failoverdomainnode name="192.168.1.2" priority="1"/>
> > ??????????????? <failoverdomainnode name="192.168.1.3" priority="2"/>
> > ??????????????? <failoverdomainnode name="192.168.1.4" priority="3"/>
> > ??????????????? <failoverdomainnode name="192.168.1.5" priority="4"/>
> > ??????????? </failoverdomain>
> > ??????????? <failoverdomain name="fd3" nofailback="0" ordered="1"
> > restricted="1">
> > ??????????????? <failoverdomainnode name="192.168.1.1" priority="4"/>
> > ??????????????? <failoverdomainnode name="192.168.1.2" priority="5"/>
> > ??????????????? <failoverdomainnode name="192.168.1.3" priority="1"/>
> > ??????????????? <failoverdomainnode name="192.168.1.4" priority="2"/>
> > ??????????????? <failoverdomainnode name="192.168.1.5" priority="3"/>
> > ??????????? </failoverdomain>
> > ??????????? <failoverdomain name="fd4" nofailback="0" ordered="1"
> > restricted="1">
> > ??????????????? <failoverdomainnode name="192.168.1.1" priority="3"/>
> > ??????????????? <failoverdomainnode name="192.168.1.2" priority="4"/>
> > ??????????????? <failoverdomainnode name="192.168.1.3" priority="5"/>
> > ??????????????? <failoverdomainnode name="192.168.1.4" priority="1"/>
> > ??????????????? <failoverdomainnode name="192.168.1.5" priority="2"/>
> > ??????????? </failoverdomain>
> > ??????? </failoverdomains>
> > ??????? <resources>
> > ??????????? <ip address="10.1.1.1" monitor_link="1"/>
> > ??????????? <ip address="10.1.1.2" monitor_link="1"/>
> > ??????????? <ip address="10.1.1.3" monitor_link="1"/>
> > ??????????? <ip address="10.1.1.4" monitor_link="1"/>
> > ??????????? <ip address="10.1.1.5" monitor_link="1"/>
> > ??????????? <script file="/usr/local/bin/service1" name="service1"/>
> > ??????????? <script file="/usr/local/bin/service2" name="service2"/>
> > ??????????? <script file="/usr/local/bin/service3" name="service3"/>
> > ??????????? <script file="/usr/local/bin/service4" name="service4"/>
> > ?????? </resources>
> > ??????? <service autostart="1" domain="fd1" exclusive="1" name="service1"
> > recovery="relocate">
> > ??????????? <ip ref="10.1.1.1"/>
> > ??????????? <script ref="service1"/>
> > ??????? </service>
> > ??????? <service autostart="1" domain="fd2" exclusive="1" name="service2"
> > recovery="relocate">
> > ??????????? <ip ref="10.1.1.2"/>
> > ??????????? <script ref="service2"/>
> > ??????? </service>
> > ??????? <service autostart="1" domain="fd3" exclusive="1" name="service3"
> > recovery="relocate">
> > ??????????? <ip ref="10.1.1.3"/>
> > ??????????? <script ref="service3"/>
> > ??????? </service>
> > ??????? <service autostart="1" domain="fd4" exclusive="1" name="service4"
> > recovery="relocate">
> > ??????????? <ip ref="10.1.1.4"/>
> > ??????????? <script ref="service4"/>
> > ??????? </service>
> > ??? </rm>
> > </cluster>
> >
> >
> > --
> > Linux-cluster mailing list
> > Linux-cluster@redhat.com
> > https://www.redhat.com/mailman/listinfo/linux-cluster
> >
>
>
>
> ------------------------------
>
> Message: 5
> Date: Sat, 15 May 2010 15:26:49 +0200
> From: "Kit Gerrits" <kitgerrits@gmail.com>
> To: "'linux clustering'" <linux-cluster@redhat.com>
> Subject: Re: [Linux-cluster] pull plug on node, service never
> relocates
> Message-ID: <4beea118.1067f10a.4a1f.ffff8975@mx.google.com>
> Content-Type: text/plain; charset="us-ascii"
>
>
> Hello,
>
> You might want to check the syslog to see if the cluster has noticed the
> outage and what is has tried to do about it.
> You can also check the node status via 'cman nodes' (explanaation of states
> in the cman manpage).
> Does the server have another power source, by any chance?
> (if not make sure you DO have dual power supplies. These things die Often)
>
>
> Regards,
>
> Kit
>
> _____
>
> From: linux-cluster-bounces@redhat.com
> [mailto:linux-cluster-bounces@redhat.com] On Behalf Of Dusty
> Sent: vrijdag 14 mei 2010 21:45
> To: Linux-cluster@redhat.com
> Subject: [Linux-cluster] pull plug on node, service never relocates
>
>
> Greetings,
>
> Using stock "clustering" and "cluster-storage" from RHEL5 update 4 X86_64
> ISO.
>
> As an example using my below config:
>
> Node1 is running service1, node2 is running service2, etc, etc, node5 is
> spare and available for the relocation of any failover domain / cluster
> service.
>
> If I go into the APC PDU and turn off the electrical port to node1, node2
> will fence node1 (going into the APC PDU and doing and off, on on node1's
> port), this is fine. Works well. When node1 comes back up, then it shuts
> down service1 and service1 relocates to node5.
>
> Now if I go in the lab and literally pull the plug on node5 running
> service1, another node fences node5 via the APC - can check the APC PDU log
> and see that it has done an off/on on node5's electrical port just fine.
>
> But I pulled the plug on node5 - resetting the power doesn't matter. I want
> to simulate a completely dead node, and have the service relocate in this
> case of complete node failure.
>
> In this RHEL5.4 cluster, the service never relocates. I can similate this
> on
> any node for any service. What if a node's motherboard fries?
>
> What can I set to have the remaining nodes stop waiting for the reboot of a
> failed node and just go ahead and relocate the cluster service that had
> been
> running on the now failed node?
>
> Thank you!
>
> versions:
>
> cman-2.0.115-1.el5
> openais-0.80.6-8.el5
> modcluster-0.12.1-2.el5
> lvm2-cluster-2.02.46-8.el5
> rgmanager-2.0.52-1.el5
> ricci-0.12.2-6.el5
>
> cluster.conf (sanitized, real scripts removed, all gfs2 mounts gone for
> clarity):
> <?xml version="1.0"?>
> <cluster config_version="1"
> name="alderaanDefenseShieldRebelAllianceCluster">
> <fence_daemon clean_start="0" post_fail_delay="3" post_join_delay="60"/>
> <clusternodes>
> <clusternode name="192.168.1.1" nodeid="1" votes="1">
> <fence>
> <method name="1">
> <device name="apc_pdu" port="1" switch="1"/>
> </method>
> </fence>
> </clusternode>
> <clusternode name="192.168.1.2" nodeid="2" votes="1">
> <fence>
> <method name="1">
> <device name="apc_pdu" port="2" switch="1"/>
> </method>
> </fence>
> </clusternode>
> <clusternode name="192.168.1.3" nodeid="3" votes="1">
> <fence>
> <method name="1">
> <device name="apc_pdu" port="3" switch="1"/>
> </method>
> </fence>
> </clusternode>
> <clusternode name="192.168.1.4" nodeid="4" votes="1">
> <fence>
> <method name="1">
> <device name="apc_pdu" port="4" switch="1"/>
> </method>
> </fence>
> </clusternode>
> <clusternode name="192.168.1.5" nodeid="5" votes="1">
> <fence>
> <method name="1">
> <device name="apc_pdu" port="5" switch="1"/>
> </method>
> </fence>
> </clusternode>
> </clusternodes>
> <cman expected_votes="6"/>
> <fencedevices>
> <fencedevice agent="fence_apc" ipaddr="192.168.1.20" login="device"
> name="apc_pdu" passwd="wonderwomanWasAPrettyCoolSuperhero"/>
> </fencedevices>
> <rm>
> <failoverdomains>
> <failoverdomain name="fd1" nofailback="0" ordered="1"
> restricted="1">
> <failoverdomainnode name="192.168.1.1" priority="1"/>
> <failoverdomainnode name="192.168.1.2" priority="2"/>
> <failoverdomainnode name="192.168.1.3" priority="3"/>
> <failoverdomainnode name="192.168.1.4" priority="4"/>
> <failoverdomainnode name="192.168.1.5" priority="5"/>
> </failoverdomain>
> <failoverdomain name="fd2" nofailback="0" ordered="1"
> restricted="1">
> <failoverdomainnode name="192.168.1.1" priority="5"/>
> <failoverdomainnode name="192.168.1.2" priority="1"/>
> <failoverdomainnode name="192.168.1.3" priority="2"/>
> <failoverdomainnode name="192.168.1.4" priority="3"/>
> <failoverdomainnode name="192.168.1.5" priority="4"/>
> </failoverdomain>
> <failoverdomain name="fd3" nofailback="0" ordered="1"
> restricted="1">
> <failoverdomainnode name="192.168.1.1" priority="4"/>
> <failoverdomainnode name="192.168.1.2" priority="5"/>
> <failoverdomainnode name="192.168.1.3" priority="1"/>
> <failoverdomainnode name="192.168.1.4" priority="2"/>
> <failoverdomainnode name="192.168.1.5" priority="3"/>
> </failoverdomain>
> <failoverdomain name="fd4" nofailback="0" ordered="1"
> restricted="1">
> <failoverdomainnode name="192.168.1.1" priority="3"/>
> <failoverdomainnode name="192.168.1.2" priority="4"/>
> <failoverdomainnode name="192.168.1.3" priority="5"/>
> <failoverdomainnode name="192.168.1.4" priority="1"/>
> <failoverdomainnode name="192.168.1.5" priority="2"/>
> </failoverdomain>
> </failoverdomains>
> <resources>
> <ip address="10.1.1.1" monitor_link="1"/>
> <ip address="10.1.1.2" monitor_link="1"/>
> <ip address="10.1.1.3" monitor_link="1"/>
> <ip address="10.1.1.4" monitor_link="1"/>
> <ip address="10.1.1.5" monitor_link="1"/>
> <script file="/usr/local/bin/service1" name="service1"/>
> <script file="/usr/local/bin/service2" name="service2"/>
> <script file="/usr/local/bin/service3" name="service3"/>
> <script file="/usr/local/bin/service4" name="service4"/>
> </resources>
> <service autostart="1" domain="fd1" exclusive="1" name="service1"
> recovery="relocate">
> <ip ref="10.1.1.1"/>
> <script ref="service1"/>
> </service>
> <service autostart="1" domain="fd2" exclusive="1" name="service2"
> recovery="relocate">
> <ip ref="10.1.1.2"/>
> <script ref="service2"/>
> </service>
> <service autostart="1" domain="fd3" exclusive="1" name="service3"
> recovery="relocate">
> <ip ref="10.1.1.3"/>
> <script ref="service3"/>
> </service>
> <service autostart="1" domain="fd4" exclusive="1" name="service4"
> recovery="relocate">
> <ip ref="10.1.1.4"/>
> <script ref="service4"/>
> </service>
> </rm>
> </cluster>
>
>
>
> No virus found in this incoming message.
> Checked by AVG - www.avg.com
> Version: 9.0.819 / Virus Database: 271.1.1/2874 - Release Date: 05/14/10
> 20:26:00
>
>
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: <
> https://www.redhat.com/archives/linux-cluster/attachments/20100515/4bc55bbe/attachment.html
> >
>
> ------------------------------
>
> --
> Linux-cluster mailing list
> Linux-cluster@redhat.com
> https://www.redhat.com/mailman/listinfo/linux-cluster
>
> End of Linux-cluster Digest, Vol 73, Issue 15
> *********************************************
>
--
Warm Regards
Parshuram Prasad
+91-9560170372
Sr. System Administrator & Database Administrator
Stratoshear Technology Pvt. Ltd.
BPS House Green Park -16
www.stratoshear.com
[Attachment #5 (text/html)]
please send me cluster script . i want to create two node clustering on linux \
5.3<br><br>thx <br>parshuram<br><br><br><div class="gmail_quote">On Sat, May 15, 2010 \
at 6:57 PM, <span dir="ltr"><<a \
href="mailto:linux-cluster-request@redhat.com">linux-cluster-request@redhat.com</a>></span> \
wrote:<br> <blockquote class="gmail_quote" style="border-left: 1px solid rgb(204, \
204, 204); margin: 0pt 0pt 0pt 0.8ex; padding-left: 1ex;">Send Linux-cluster mailing \
list submissions to<br>
<a href="mailto:linux-cluster@redhat.com">linux-cluster@redhat.com</a><br>
<br>
To subscribe or unsubscribe via the World Wide Web, visit<br>
<a href="https://www.redhat.com/mailman/listinfo/linux-cluster" \
target="_blank">https://www.redhat.com/mailman/listinfo/linux-cluster</a><br> or, via \
email, send a message with subject or body 'help' to<br>
<a href="mailto:linux-cluster-request@redhat.com">linux-cluster-request@redhat.com</a><br>
<br>
You can reach the person managing the list at<br>
<a href="mailto:linux-cluster-owner@redhat.com">linux-cluster-owner@redhat.com</a><br>
<br>
When replying, please edit your Subject line so it is more specific<br>
than "Re: Contents of Linux-cluster digest..."<br>
<br>
<br>
Today's Topics:<br>
<br>
1. GFS on Debian Lenny (Brent Clark)<br>
2. pull plug on node, service never relocates (Dusty)<br>
3. Re: GFS on Debian Lenny (Joao Ferreira gmail)<br>
4. Re: pull plug on node, service never relocates (Corey Kovacs)<br>
5. Re: pull plug on node, service never relocates (Kit Gerrits)<br>
<br>
<br>
----------------------------------------------------------------------<br>
<br>
Message: 1<br>
Date: Fri, 14 May 2010 20:26:46 +0200<br>
From: Brent Clark <<a \
href="mailto:brentgclarklist@gmail.com">brentgclarklist@gmail.com</a>><br>
To: linux clustering <<a \
href="mailto:linux-cluster@redhat.com">linux-cluster@redhat.com</a>><br>
Subject: [Linux-cluster] GFS on Debian Lenny<br>
Message-ID: <<a href="mailto:4BED95E6.4040006@gmail.com">4BED95E6.4040006@gmail.com</a>><br>
Content-Type: text/plain; charset=ISO-8859-1; format=flowed<br>
<br>
Hiya<br>
<br>
Im trying to get GFS working on Debian Lenny. Unfortuantely<br>
documentation seems to be non existent. And the one site that google<br>
recommends, <a href="http://gcharriere.com" target="_blank">gcharriere.com</a>, is \
down.<br> <br>
I used googles caching mechanism to try and make head and tails of whats<br>
needed to be done, but unfortunately Im unsuccessful.<br>
<br>
Would anyone have any documentation or any sites or if you have a heart,<br>
provide a howto to get GFS working.<br>
<br>
From myside, all ive done is:<br>
<br>
aptitude install gfs2-tools<br>
modprobe gfs2<br>
gfs_mkfs -p lock_dlm -t lolcats:drbdtest /dev/drbd0 -j 2<br>
<br>
thats all, ive done. No editting of configs etc.<br>
<br>
When I try,<br>
<br>
mount -t gfs2 /dev/drbd0 /drbd/<br>
<br>
I get the following message:<br>
<br>
/sbin/mount.gfs2: can't connect to gfs_controld: Connection refused<br>
/sbin/mount.gfs2: gfs_controld not running<br>
/sbin/mount.gfs2: error mounting lockproto lock_dlm<br>
<br>
If anyone can help, it would be appreciated.<br>
<br>
Kind Regards<br>
Brent Clark<br>
<br>
<br>
<br>
------------------------------<br>
<br>
Message: 2<br>
Date: Fri, 14 May 2010 14:45:11 -0500<br>
From: Dusty <<a href="mailto:dhoffutt@gmail.com">dhoffutt@gmail.com</a>><br>
To: <a href="mailto:Linux-cluster@redhat.com">Linux-cluster@redhat.com</a><br>
Subject: [Linux-cluster] pull plug on node, service never relocates<br>
Message-ID:<br>
<<a href="mailto:AANLkTil1ssNgEYRs71I_xmsLV3enagF76kEQYAt-Tdse@mail.gmail.com">AANLkTil1ssNgEYRs71I_xmsLV3enagF76kEQYAt-Tdse@mail.gmail.com</a>><br>
Content-Type: text/plain; charset="iso-8859-1"<br>
<br>
Greetings,<br>
<br>
Using stock "clustering" and "cluster-storage" from RHEL5 update \
4 X86_64<br> ISO.<br>
<br>
As an example using my below config:<br>
<br>
Node1 is running service1, node2 is running service2, etc, etc, node5 is<br>
spare and available for the relocation of any failover domain / cluster<br>
service.<br>
<br>
If I go into the APC PDU and turn off the electrical port to node1, node2<br>
will fence node1 (going into the APC PDU and doing and off, on on node1's<br>
port), this is fine. Works well. When node1 comes back up, then it shuts<br>
down service1 and service1 relocates to node5.<br>
<br>
Now if I go in the lab and literally pull the plug on node5 running<br>
service1, another node fences node5 via the APC - can check the APC PDU log<br>
and see that it has done an off/on on node5's electrical port just fine.<br>
<br>
But I pulled the plug on node5 - resetting the power doesn't matter. I want<br>
to simulate a completely dead node, and have the service relocate in this<br>
case of complete node failure.<br>
<br>
In this RHEL5.4 cluster, the service never relocates. I can similate this on<br>
any node for any service. What if a node's motherboard fries?<br>
<br>
What can I set to have the remaining nodes stop waiting for the reboot of a<br>
failed node and just go ahead and relocate the cluster service that had been<br>
running on the now failed node?<br>
<br>
Thank you!<br>
<br>
versions:<br>
<br>
cman-2.0.115-1.el5<br>
openais-0.80.6-8.el5<br>
modcluster-0.12.1-2.el5<br>
lvm2-cluster-2.02.46-8.el5<br>
rgmanager-2.0.52-1.el5<br>
ricci-0.12.2-6.el5<br>
<br>
cluster.conf (sanitized, real scripts removed, all gfs2 mounts gone for<br>
clarity):<br>
<?xml version="1.0"?><br>
<cluster config_version="1"<br>
name="alderaanDefenseShieldRebelAllianceCluster"><br>
<fence_daemon clean_start="0" post_fail_delay="3" \
post_join_delay="60"/><br> <clusternodes><br>
<clusternode name="192.168.1.1" nodeid="1" \
votes="1"><br> <fence><br>
<method name="1"><br>
<device name="apc_pdu" port="1" \
switch="1"/><br> </method><br>
</fence><br>
</clusternode><br>
<clusternode name="192.168.1.2" nodeid="2" \
votes="1"><br> <fence><br>
<method name="1"><br>
<device name="apc_pdu" port="2" \
switch="1"/><br> </method><br>
</fence><br>
</clusternode><br>
<clusternode name="192.168.1.3" nodeid="3" \
votes="1"><br> <fence><br>
<method name="1"><br>
<device name="apc_pdu" port="3" \
switch="1"/><br> </method><br>
</fence><br>
</clusternode><br>
<clusternode name="192.168.1.4" nodeid="4" \
votes="1"><br> <fence><br>
<method name="1"><br>
<device name="apc_pdu" port="4" \
switch="1"/><br> </method><br>
</fence><br>
</clusternode><br>
<clusternode name="192.168.1.5" nodeid="5" \
votes="1"><br> <fence><br>
<method name="1"><br>
<device name="apc_pdu" port="5" \
switch="1"/><br> </method><br>
</fence><br>
</clusternode><br>
</clusternodes><br>
<cman expected_votes="6"/><br>
<fencedevices><br>
<fencedevice agent="fence_apc" ipaddr="192.168.1.20" \
login="device"<br> name="apc_pdu" \
passwd="wonderwomanWasAPrettyCoolSuperhero"/><br> \
</fencedevices><br> <rm><br>
<failoverdomains><br>
<failoverdomain name="fd1" nofailback="0" \
ordered="1"<br> restricted="1"><br>
<failoverdomainnode name="192.168.1.1" \
priority="1"/><br>
<failoverdomainnode name="192.168.1.2" \
priority="2"/><br>
<failoverdomainnode name="192.168.1.3" \
priority="3"/><br>
<failoverdomainnode name="192.168.1.4" \
priority="4"/><br>
<failoverdomainnode name="192.168.1.5" \
priority="5"/><br> </failoverdomain><br>
<failoverdomain name="fd2" nofailback="0" \
ordered="1"<br> restricted="1"><br>
<failoverdomainnode name="192.168.1.1" \
priority="5"/><br>
<failoverdomainnode name="192.168.1.2" \
priority="1"/><br>
<failoverdomainnode name="192.168.1.3" \
priority="2"/><br>
<failoverdomainnode name="192.168.1.4" \
priority="3"/><br>
<failoverdomainnode name="192.168.1.5" \
priority="4"/><br> </failoverdomain><br>
<failoverdomain name="fd3" nofailback="0" \
ordered="1"<br> restricted="1"><br>
<failoverdomainnode name="192.168.1.1" \
priority="4"/><br>
<failoverdomainnode name="192.168.1.2" \
priority="5"/><br>
<failoverdomainnode name="192.168.1.3" \
priority="1"/><br>
<failoverdomainnode name="192.168.1.4" \
priority="2"/><br>
<failoverdomainnode name="192.168.1.5" \
priority="3"/><br> </failoverdomain><br>
<failoverdomain name="fd4" nofailback="0" \
ordered="1"<br> restricted="1"><br>
<failoverdomainnode name="192.168.1.1" \
priority="3"/><br>
<failoverdomainnode name="192.168.1.2" \
priority="4"/><br>
<failoverdomainnode name="192.168.1.3" \
priority="5"/><br>
<failoverdomainnode name="192.168.1.4" \
priority="1"/><br>
<failoverdomainnode name="192.168.1.5" \
priority="2"/><br> </failoverdomain><br>
</failoverdomains><br>
<resources><br>
<ip address="10.1.1.1" monitor_link="1"/><br>
<ip address="10.1.1.2" monitor_link="1"/><br>
<ip address="10.1.1.3" monitor_link="1"/><br>
<ip address="10.1.1.4" monitor_link="1"/><br>
<ip address="10.1.1.5" monitor_link="1"/><br>
<script file="/usr/local/bin/service1" \
name="service1"/><br>
<script file="/usr/local/bin/service2" \
name="service2"/><br>
<script file="/usr/local/bin/service3" \
name="service3"/><br>
<script file="/usr/local/bin/service4" \
name="service4"/><br> </resources><br>
<service autostart="1" domain="fd1" \
exclusive="1" name="service1"<br> \
recovery="relocate"><br> <ip ref="10.1.1.1"/><br>
<script ref="service1"/><br>
</service><br>
<service autostart="1" domain="fd2" \
exclusive="1" name="service2"<br> \
recovery="relocate"><br> <ip ref="10.1.1.2"/><br>
<script ref="service2"/><br>
</service><br>
<service autostart="1" domain="fd3" \
exclusive="1" name="service3"<br> \
recovery="relocate"><br> <ip ref="10.1.1.3"/><br>
<script ref="service3"/><br>
</service><br>
<service autostart="1" domain="fd4" \
exclusive="1" name="service4"<br> \
recovery="relocate"><br> <ip ref="10.1.1.4"/><br>
<script ref="service4"/><br>
</service><br>
</rm><br>
</cluster><br>
-------------- next part --------------<br>
An HTML attachment was scrubbed...<br>
URL: <<a href="https://www.redhat.com/archives/linux-cluster/attachments/20100514/c892bf86/attachment.html" \
target="_blank">https://www.redhat.com/archives/linux-cluster/attachments/20100514/c892bf86/attachment.html</a>><br>
<br>
------------------------------<br>
<br>
Message: 3<br>
Date: Fri, 14 May 2010 23:31:41 +0100<br>
From: Joao Ferreira gmail <<a \
href="mailto:joao.miguel.c.ferreira@gmail.com">joao.miguel.c.ferreira@gmail.com</a>><br>
To: linux clustering <<a \
href="mailto:linux-cluster@redhat.com">linux-cluster@redhat.com</a>><br>
Subject: Re: [Linux-cluster] GFS on Debian Lenny<br>
Message-ID: <<a href="mailto:1273876301.5298.1.camel@debj5n.critical.pt">1273876301.5298.1.camel@debj5n.critical.pt</a>><br>
Content-Type: text/plain<br>
<br>
Have you checked the docs at the drbd site ?<br>
<br>
it contains some short info regarding usage of gsf over drbd..<br>
<br>
<a href="http://www.drbd.org/docs/applications/" \
target="_blank">http://www.drbd.org/docs/applications/</a><br> <br>
cheers<br>
Joao<br>
<br>
On Fri, 2010-05-14 at 20:26 +0200, Brent Clark wrote:<br>
> Hiya<br>
><br>
> Im trying to get GFS working on Debian Lenny. Unfortuantely<br>
> documentation seems to be non existent. And the one site that google<br>
> recommends, <a href="http://gcharriere.com" target="_blank">gcharriere.com</a>, \
is down.<br> ><br>
> I used googles caching mechanism to try and make head and tails of whats<br>
> needed to be done, but unfortunately Im unsuccessful.<br>
><br>
> Would anyone have any documentation or any sites or if you have a heart,<br>
> provide a howto to get GFS working.<br>
><br>
> From myside, all ive done is:<br>
><br>
> aptitude install gfs2-tools<br>
> modprobe gfs2<br>
> gfs_mkfs -p lock_dlm -t lolcats:drbdtest /dev/drbd0 -j 2<br>
><br>
> thats all, ive done. No editting of configs etc.<br>
><br>
> When I try,<br>
><br>
> mount -t gfs2 /dev/drbd0 /drbd/<br>
><br>
> I get the following message:<br>
><br>
> /sbin/mount.gfs2: can't connect to gfs_controld: Connection refused<br>
> /sbin/mount.gfs2: gfs_controld not running<br>
> /sbin/mount.gfs2: error mounting lockproto lock_dlm<br>
><br>
> If anyone can help, it would be appreciated.<br>
><br>
> Kind Regards<br>
> Brent Clark<br>
><br>
> --<br>
> Linux-cluster mailing list<br>
> <a href="mailto:Linux-cluster@redhat.com">Linux-cluster@redhat.com</a><br>
> <a href="https://www.redhat.com/mailman/listinfo/linux-cluster" \
target="_blank">https://www.redhat.com/mailman/listinfo/linux-cluster</a><br> <br>
<br>
<br>
------------------------------<br>
<br>
Message: 4<br>
Date: Sat, 15 May 2010 04:59:23 +0100<br>
From: Corey Kovacs <<a \
href="mailto:corey.kovacs@gmail.com">corey.kovacs@gmail.com</a>><br>
To: linux clustering <<a \
href="mailto:linux-cluster@redhat.com">linux-cluster@redhat.com</a>><br>
Subject: Re: [Linux-cluster] pull plug on node, service never<br>
relocates<br>
Message-ID:<br>
<<a href="mailto:AANLkTinYVvrit1oPb76TfLa9vmp1AMHcGI3eoZALHxrJ@mail.gmail.com">AANLkTinYVvrit1oPb76TfLa9vmp1AMHcGI3eoZALHxrJ@mail.gmail.com</a>><br>
Content-Type: text/plain; charset=ISO-8859-1<br>
<br>
What happens when you do ...<br>
<br>
fence_node 192.168.1.4<br>
<br>
from any of the other nodes?<br>
<br>
if that doesn't work, then fencing is not configured correctly and you<br>
should try to invoke the fence agent directly.<br>
Also, it would help if you included the APC model and firmware rev.<br>
The fence_apc agent can be finicky about such things.<br>
<br>
<br>
Hope this helps.<br>
<br>
-Core<br>
<br>
On Fri, May 14, 2010 at 8:45 PM, Dusty <<a \
href="mailto:dhoffutt@gmail.com">dhoffutt@gmail.com</a>> wrote:<br> > \
Greetings,<br> ><br>
> Using stock "clustering" and "cluster-storage" from RHEL5 \
update 4 X86_64<br> > ISO.<br>
><br>
> As an example using my below config:<br>
><br>
> Node1 is running service1, node2 is running service2, etc, etc, node5 is<br>
> spare and available for the relocation of any failover domain / cluster<br>
> service.<br>
><br>
> If I go into the APC PDU and turn off the electrical port to node1, node2<br>
> will fence node1 (going into the APC PDU and doing and off, on on \
node1's<br> > port), this is fine. Works well. When node1 comes back up, then \
it shuts<br> > down service1 and service1 relocates to node5.<br>
><br>
> Now if I go in the lab and literally pull the plug on node5 running<br>
> service1, another node fences node5 via the APC - can check the APC PDU log<br>
> and see that it has done an off/on on node5's electrical port just fine.<br>
><br>
> But I pulled the plug on node5 - resetting the power doesn't matter. I \
want<br> > to simulate a completely dead node, and have the service relocate in \
this<br> > case of complete node failure.<br>
><br>
> In this RHEL5.4 cluster, the service never relocates. I can similate this on<br>
> any node for any service. What if a node's motherboard fries?<br>
><br>
> What can I set to have the remaining nodes stop waiting for the reboot of a<br>
> failed node and just go ahead and relocate the cluster service that had been<br>
> running on the now failed node?<br>
><br>
> Thank you!<br>
><br>
> versions:<br>
><br>
> cman-2.0.115-1.el5<br>
> openais-0.80.6-8.el5<br>
> modcluster-0.12.1-2.el5<br>
> lvm2-cluster-2.02.46-8.el5<br>
> rgmanager-2.0.52-1.el5<br>
> ricci-0.12.2-6.el5<br>
><br>
> cluster.conf (sanitized, real scripts removed, all gfs2 mounts gone for<br>
> clarity):<br>
> <?xml version="1.0"?><br>
> <cluster config_version="1"<br>
> name="alderaanDefenseShieldRebelAllianceCluster"><br>
> ??? <fence_daemon clean_start="0" post_fail_delay="3" \
post_join_delay="60"/><br> > ??? <clusternodes><br>
> ??????? <clusternode name="192.168.1.1" nodeid="1" \
votes="1"><br> > ??????????? <fence><br>
> ??????????????? <method name="1"><br>
> ??????????????????? <device name="apc_pdu" port="1" \
switch="1"/><br> > ??????????????? </method><br>
> ??????????? </fence><br>
> ??????? </clusternode><br>
> ??????? <clusternode name="192.168.1.2" nodeid="2" \
votes="1"><br> > ??????????? <fence><br>
> ??????????????? <method name="1"><br>
> ??????????????????? <device name="apc_pdu" port="2" \
switch="1"/><br> > ??????????????? </method><br>
> ??????????? </fence><br>
> ??????? </clusternode><br>
> ??????? <clusternode name="192.168.1.3" nodeid="3" \
votes="1"><br> > ??????????? <fence><br>
> ??????????????? <method name="1"><br>
> ??????????????????? <device name="apc_pdu" port="3" \
switch="1"/><br> > ??????????????? </method><br>
> ??????????? </fence><br>
> ??????? </clusternode><br>
> ??????? <clusternode name="192.168.1.4" nodeid="4" \
votes="1"><br> > ??????????? <fence><br>
> ??????????????? <method name="1"><br>
> ??????????????????? <device name="apc_pdu" port="4" \
switch="1"/><br> > ??????????????? </method><br>
> ??????????? </fence><br>
> ??????? </clusternode><br>
> ??????? <clusternode name="192.168.1.5" nodeid="5" \
votes="1"><br> > ??????????? <fence><br>
> ??????????????? <method name="1"><br>
> ??????????????????? <device name="apc_pdu" port="5" \
switch="1"/><br> > ??????????????? </method><br>
> ??????????? </fence><br>
> ??????? </clusternode><br>
> ??? </clusternodes><br>
> ??? <cman expected_votes="6"/><br>
> ??? <fencedevices><br>
> ??????? <fencedevice agent="fence_apc" \
ipaddr="192.168.1.20" login="device"<br> > \
name="apc_pdu" \
passwd="wonderwomanWasAPrettyCoolSuperhero"/><br> > ??? \
</fencedevices><br> > ??? <rm><br>
> ??????? <failoverdomains><br>
> ??????????? <failoverdomain name="fd1" nofailback="0" \
ordered="1"<br> > restricted="1"><br>
> ??????????????? <failoverdomainnode name="192.168.1.1" \
priority="1"/><br> > ??????????????? <failoverdomainnode \
name="192.168.1.2" priority="2"/><br> > ??????????????? \
<failoverdomainnode name="192.168.1.3" priority="3"/><br> \
> ??????????????? <failoverdomainnode name="192.168.1.4" \
priority="4"/><br> > ??????????????? <failoverdomainnode \
name="192.168.1.5" priority="5"/><br> > ??????????? \
</failoverdomain><br> > ??????????? <failoverdomain name="fd2" \
nofailback="0" ordered="1"<br> > \
restricted="1"><br> > ??????????????? <failoverdomainnode \
name="192.168.1.1" priority="5"/><br> > ??????????????? \
<failoverdomainnode name="192.168.1.2" priority="1"/><br> \
> ??????????????? <failoverdomainnode name="192.168.1.3" \
priority="2"/><br> > ??????????????? <failoverdomainnode \
name="192.168.1.4" priority="3"/><br> > ??????????????? \
<failoverdomainnode name="192.168.1.5" priority="4"/><br> \
> ??????????? </failoverdomain><br> > ??????????? <failoverdomain \
name="fd3" nofailback="0" ordered="1"<br> > \
restricted="1"><br> > ??????????????? <failoverdomainnode \
name="192.168.1.1" priority="4"/><br> > ??????????????? \
<failoverdomainnode name="192.168.1.2" priority="5"/><br> \
> ??????????????? <failoverdomainnode name="192.168.1.3" \
priority="1"/><br> > ??????????????? <failoverdomainnode \
name="192.168.1.4" priority="2"/><br> > ??????????????? \
<failoverdomainnode name="192.168.1.5" priority="3"/><br> \
> ??????????? </failoverdomain><br> > ??????????? <failoverdomain \
name="fd4" nofailback="0" ordered="1"<br> > \
restricted="1"><br> > ??????????????? <failoverdomainnode \
name="192.168.1.1" priority="3"/><br> > ??????????????? \
<failoverdomainnode name="192.168.1.2" priority="4"/><br> \
> ??????????????? <failoverdomainnode name="192.168.1.3" \
priority="5"/><br> > ??????????????? <failoverdomainnode \
name="192.168.1.4" priority="1"/><br> > ??????????????? \
<failoverdomainnode name="192.168.1.5" priority="2"/><br> \
> ??????????? </failoverdomain><br> > ??????? \
</failoverdomains><br> > ??????? <resources><br>
> ??????????? <ip address="10.1.1.1" \
monitor_link="1"/><br> > ??????????? <ip \
address="10.1.1.2" monitor_link="1"/><br> > ??????????? \
<ip address="10.1.1.3" monitor_link="1"/><br> > \
??????????? <ip address="10.1.1.4" monitor_link="1"/><br> \
> ??????????? <ip address="10.1.1.5" \
monitor_link="1"/><br> > ??????????? <script \
file="/usr/local/bin/service1" name="service1"/><br> > \
??????????? <script file="/usr/local/bin/service2" \
name="service2"/><br> > ??????????? <script \
file="/usr/local/bin/service3" name="service3"/><br> > \
??????????? <script file="/usr/local/bin/service4" \
name="service4"/><br> > ?????? </resources><br>
> ??????? <service autostart="1" domain="fd1" \
exclusive="1" name="service1"<br> > \
recovery="relocate"><br> > ??????????? <ip \
ref="10.1.1.1"/><br> > ??????????? <script \
ref="service1"/><br> > ??????? </service><br>
> ??????? <service autostart="1" domain="fd2" \
exclusive="1" name="service2"<br> > \
recovery="relocate"><br> > ??????????? <ip \
ref="10.1.1.2"/><br> > ??????????? <script \
ref="service2"/><br> > ??????? </service><br>
> ??????? <service autostart="1" domain="fd3" \
exclusive="1" name="service3"<br> > \
recovery="relocate"><br> > ??????????? <ip \
ref="10.1.1.3"/><br> > ??????????? <script \
ref="service3"/><br> > ??????? </service><br>
> ??????? <service autostart="1" domain="fd4" \
exclusive="1" name="service4"<br> > \
recovery="relocate"><br> > ??????????? <ip \
ref="10.1.1.4"/><br> > ??????????? <script \
ref="service4"/><br> > ??????? </service><br>
> ??? </rm><br>
> </cluster><br>
><br>
><br>
> --<br>
> Linux-cluster mailing list<br>
> <a href="mailto:Linux-cluster@redhat.com">Linux-cluster@redhat.com</a><br>
> <a href="https://www.redhat.com/mailman/listinfo/linux-cluster" \
target="_blank">https://www.redhat.com/mailman/listinfo/linux-cluster</a><br> \
><br> <br>
<br>
<br>
------------------------------<br>
<br>
Message: 5<br>
Date: Sat, 15 May 2010 15:26:49 +0200<br>
From: "Kit Gerrits" <<a \
href="mailto:kitgerrits@gmail.com">kitgerrits@gmail.com</a>><br>
To: "'linux clustering'" <<a \
href="mailto:linux-cluster@redhat.com">linux-cluster@redhat.com</a>><br>
Subject: Re: [Linux-cluster] pull plug on node, service never<br>
relocates<br>
Message-ID: <<a href="mailto:4beea118.1067f10a.4a1f.ffff8975@mx.google.com">4beea118.1067f10a.4a1f.ffff8975@mx.google.com</a>><br>
Content-Type: text/plain; charset="us-ascii"<br>
<br>
<br>
Hello,<br>
<br>
You might want to check the syslog to see if the cluster has noticed the<br>
outage and what is has tried to do about it.<br>
You can also check the node status via 'cman nodes' (explanaation of \
states<br> in the cman manpage).<br>
Does the server have another power source, by any chance?<br>
(if not make sure you DO have dual power supplies. These things die Often)<br>
<br>
<br>
Regards,<br>
<br>
Kit<br>
<br>
_____<br>
<br>
From: <a href="mailto:linux-cluster-bounces@redhat.com">linux-cluster-bounces@redhat.com</a><br>
[mailto:<a href="mailto:linux-cluster-bounces@redhat.com">linux-cluster-bounces@redhat.com</a>] \
On Behalf Of Dusty<br>
Sent: vrijdag 14 mei 2010 21:45<br>
To: <a href="mailto:Linux-cluster@redhat.com">Linux-cluster@redhat.com</a><br>
Subject: [Linux-cluster] pull plug on node, service never relocates<br>
<br>
<br>
Greetings,<br>
<br>
Using stock "clustering" and "cluster-storage" from RHEL5 update \
4 X86_64<br> ISO.<br>
<br>
As an example using my below config:<br>
<br>
Node1 is running service1, node2 is running service2, etc, etc, node5 is<br>
spare and available for the relocation of any failover domain / cluster<br>
service.<br>
<br>
If I go into the APC PDU and turn off the electrical port to node1, node2<br>
will fence node1 (going into the APC PDU and doing and off, on on node1's<br>
port), this is fine. Works well. When node1 comes back up, then it shuts<br>
down service1 and service1 relocates to node5.<br>
<br>
Now if I go in the lab and literally pull the plug on node5 running<br>
service1, another node fences node5 via the APC - can check the APC PDU log<br>
and see that it has done an off/on on node5's electrical port just fine.<br>
<br>
But I pulled the plug on node5 - resetting the power doesn't matter. I want<br>
to simulate a completely dead node, and have the service relocate in this<br>
case of complete node failure.<br>
<br>
In this RHEL5.4 cluster, the service never relocates. I can similate this on<br>
any node for any service. What if a node's motherboard fries?<br>
<br>
What can I set to have the remaining nodes stop waiting for the reboot of a<br>
failed node and just go ahead and relocate the cluster service that had been<br>
running on the now failed node?<br>
<br>
Thank you!<br>
<br>
versions:<br>
<br>
cman-2.0.115-1.el5<br>
openais-0.80.6-8.el5<br>
modcluster-0.12.1-2.el5<br>
lvm2-cluster-2.02.46-8.el5<br>
rgmanager-2.0.52-1.el5<br>
ricci-0.12.2-6.el5<br>
<br>
cluster.conf (sanitized, real scripts removed, all gfs2 mounts gone for<br>
clarity):<br>
<?xml version="1.0"?><br>
<cluster config_version="1"<br>
name="alderaanDefenseShieldRebelAllianceCluster"><br>
<fence_daemon clean_start="0" post_fail_delay="3" \
post_join_delay="60"/><br> <clusternodes><br>
<clusternode name="192.168.1.1" nodeid="1" \
votes="1"><br> <fence><br>
<method name="1"><br>
<device name="apc_pdu" port="1" \
switch="1"/><br> </method><br>
</fence><br>
</clusternode><br>
<clusternode name="192.168.1.2" nodeid="2" \
votes="1"><br> <fence><br>
<method name="1"><br>
<device name="apc_pdu" port="2" \
switch="1"/><br> </method><br>
</fence><br>
</clusternode><br>
<clusternode name="192.168.1.3" nodeid="3" \
votes="1"><br> <fence><br>
<method name="1"><br>
<device name="apc_pdu" port="3" \
switch="1"/><br> </method><br>
</fence><br>
</clusternode><br>
<clusternode name="192.168.1.4" nodeid="4" \
votes="1"><br> <fence><br>
<method name="1"><br>
<device name="apc_pdu" port="4" \
switch="1"/><br> </method><br>
</fence><br>
</clusternode><br>
<clusternode name="192.168.1.5" nodeid="5" \
votes="1"><br> <fence><br>
<method name="1"><br>
<device name="apc_pdu" port="5" \
switch="1"/><br> </method><br>
</fence><br>
</clusternode><br>
</clusternodes><br>
<cman expected_votes="6"/><br>
<fencedevices><br>
<fencedevice agent="fence_apc" ipaddr="192.168.1.20" \
login="device"<br> name="apc_pdu" \
passwd="wonderwomanWasAPrettyCoolSuperhero"/><br> \
</fencedevices><br> <rm><br>
<failoverdomains><br>
<failoverdomain name="fd1" nofailback="0" \
ordered="1"<br> restricted="1"><br>
<failoverdomainnode name="192.168.1.1" \
priority="1"/><br>
<failoverdomainnode name="192.168.1.2" \
priority="2"/><br>
<failoverdomainnode name="192.168.1.3" \
priority="3"/><br>
<failoverdomainnode name="192.168.1.4" \
priority="4"/><br>
<failoverdomainnode name="192.168.1.5" \
priority="5"/><br> </failoverdomain><br>
<failoverdomain name="fd2" nofailback="0" \
ordered="1"<br> restricted="1"><br>
<failoverdomainnode name="192.168.1.1" \
priority="5"/><br>
<failoverdomainnode name="192.168.1.2" \
priority="1"/><br>
<failoverdomainnode name="192.168.1.3" \
priority="2"/><br>
<failoverdomainnode name="192.168.1.4" \
priority="3"/><br>
<failoverdomainnode name="192.168.1.5" \
priority="4"/><br> </failoverdomain><br>
<failoverdomain name="fd3" nofailback="0" \
ordered="1"<br> restricted="1"><br>
<failoverdomainnode name="192.168.1.1" \
priority="4"/><br>
<failoverdomainnode name="192.168.1.2" \
priority="5"/><br>
<failoverdomainnode name="192.168.1.3" \
priority="1"/><br>
<failoverdomainnode name="192.168.1.4" \
priority="2"/><br>
<failoverdomainnode name="192.168.1.5" \
priority="3"/><br> </failoverdomain><br>
<failoverdomain name="fd4" nofailback="0" \
ordered="1"<br> restricted="1"><br>
<failoverdomainnode name="192.168.1.1" \
priority="3"/><br>
<failoverdomainnode name="192.168.1.2" \
priority="4"/><br>
<failoverdomainnode name="192.168.1.3" \
priority="5"/><br>
<failoverdomainnode name="192.168.1.4" \
priority="1"/><br>
<failoverdomainnode name="192.168.1.5" \
priority="2"/><br> </failoverdomain><br>
</failoverdomains><br>
<resources><br>
<ip address="10.1.1.1" monitor_link="1"/><br>
<ip address="10.1.1.2" monitor_link="1"/><br>
<ip address="10.1.1.3" monitor_link="1"/><br>
<ip address="10.1.1.4" monitor_link="1"/><br>
<ip address="10.1.1.5" monitor_link="1"/><br>
<script file="/usr/local/bin/service1" \
name="service1"/><br>
<script file="/usr/local/bin/service2" \
name="service2"/><br>
<script file="/usr/local/bin/service3" \
name="service3"/><br>
<script file="/usr/local/bin/service4" \
name="service4"/><br> </resources><br>
<service autostart="1" domain="fd1" \
exclusive="1" name="service1"<br> \
recovery="relocate"><br> <ip ref="10.1.1.1"/><br>
<script ref="service1"/><br>
</service><br>
<service autostart="1" domain="fd2" \
exclusive="1" name="service2"<br> \
recovery="relocate"><br> <ip ref="10.1.1.2"/><br>
<script ref="service2"/><br>
</service><br>
<service autostart="1" domain="fd3" \
exclusive="1" name="service3"<br> \
recovery="relocate"><br> <ip ref="10.1.1.3"/><br>
<script ref="service3"/><br>
</service><br>
<service autostart="1" domain="fd4" \
exclusive="1" name="service4"<br> \
recovery="relocate"><br> <ip ref="10.1.1.4"/><br>
<script ref="service4"/><br>
</service><br>
</rm><br>
</cluster><br>
<br>
<br>
<br>
No virus found in this incoming message.<br>
Checked by AVG - <a href="http://www.avg.com" target="_blank">www.avg.com</a><br>
Version: 9.0.819 / Virus Database: 271.1.1/2874 - Release Date: 05/14/10<br>
20:26:00<br>
<br>
<br>
-------------- next part --------------<br>
An HTML attachment was scrubbed...<br>
URL: <<a href="https://www.redhat.com/archives/linux-cluster/attachments/20100515/4bc55bbe/attachment.html" \
target="_blank">https://www.redhat.com/archives/linux-cluster/attachments/20100515/4bc55bbe/attachment.html</a>><br>
<br>
------------------------------<br>
<font color="#888888"><br>
--<br>
Linux-cluster mailing list<br>
<a href="mailto:Linux-cluster@redhat.com">Linux-cluster@redhat.com</a><br>
<a href="https://www.redhat.com/mailman/listinfo/linux-cluster" \
target="_blank">https://www.redhat.com/mailman/listinfo/linux-cluster</a><br> <br>
End of Linux-cluster Digest, Vol 73, Issue 15<br>
*********************************************<br>
</font></blockquote></div><br><br clear="all"><br>-- <br>Warm Regards<br>Parshuram \
Prasad<br>+91-9560170372<br>Sr. System Administrator & Database Administrator \
<br><br>Stratoshear Technology Pvt. Ltd.<br><br>BPS House Green Park -16<br> <a \
href="http://www.stratoshear.com">www.stratoshear.com</a><br> <br><br>
--
Linux-cluster mailing list
Linux-cluster@redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster
[prev in list] [next in list] [prev in thread] [next in thread]
Configure |
About |
News |
Add a list |
Sponsored by KoreLogic