[prev in list] [next in list] [prev in thread] [next in thread] 

List:       drbd-user
Subject:    [DRBD-user] Pacemaker drbd fail to promote to primary
From:       Neil Schneider <nschneider () awarepoint ! com>
Date:       2016-09-02 4:56:01
Message-ID: BY1PR0401MB1612741A7E59794F48DDE47EBEE50 () BY1PR0401MB1612 ! namprd04 ! prod ! outlook ! com
[Download RAW message or body]

Below are my configs. 
The issue I'm experiencing is that when stonith reboots the primary, the secondary \
doesn't get promoted.  I thought the handlers in drbd.conf were supposed to "handle" \
that.  Anybody know what I'm missing? 
I've been looking at logs, but nothing stands out to me. The logging is pretty \
verbose. Perhaps I could make the logs a little less verbose. 
I don't know where those options are. 
I had this working with this identical configuration with the same nodes but simple \
hostnames. I switched it up to simulate a real world change, where I am using fqdn \
style hostnames, and also in the configuration files below. This is a pair of \
"appliances" that run out proprietary software.  FQDNs are part of our requirements \
for our software.   So I or someone I train, is going to have to configure this for \
each pair we deliver.  The old HA was much simpler. 
Have to always move forward. 

/etc/drbd.conf
global {
        usage-count no;
}
common {
        protocol C;
        handlers {
                fence-peer "/usr/lib/drbd/crm-fence-peer.sh";
                after-resync-target "/usr/lib/drbd/crm_unfence-peer.sh";
                split-brain "/usr/lib/drbd/notify-split-brain.sh root";
        }

        startup {
        }
        disk {
                fencing resource-and-stonith;
                #on-io-error detach;
        }
        net {
                allow-two-primaries yes;
                after-sb-0pri discard-zero-changes;
                after-sb-1pri discard-secondary;
                after-sb-2pri disconnect;
                rr-conflict disconnect;
        }
}
resource mysql {
        protocol C;
        meta-disk internal;
        device /dev/drbd1;
        syncer {
                verify-alg sha1;
                rate 33M;
                csums-alg sha1;
        }
on node1 {
        disk /dev/sda1;
        address 10.6.7.24:7789;
            }
on awpnode2 {
        disk /dev/sda1;
        address 10.6.7.27:7789;
            }
}

/etc/corosync/corosync.conf
# Please read the corosync.conf.5 manual page
compatibility: whitetank

totem {
        version: 2
        secauth: off
        threads: 0
        interface {
                member {
                        memberaddr: 10.6.7.24
                        }
                member  {
                        memberaddr: 10.6.7.27
                        }
                ringnumber: 0
                bindnetaddr: 10.6.7.0
                mcastport: 5405
                ttl: 1
        }
        transport: udpu
}
logging {
        fileline: off
        to_stderr: no
        to_logfile: yes
        logfile: /var/log/cluster/corosync.log
        to_syslog: yes
        debug: off
        timestamp: on
        logger_subsys {
                subsys: AMF
                debug: off
        }
}

pcs config show

 Master/Slave Set: mysql_data_clone [mysql_data]
     Masters: [ node2 ]
     Slaves: [ node1i ]

pcs config show
Cluster Name: awpcluster
Corosync Nodes:
 node1 node2
Pacemaker Nodes:
 node1 node2

Resources:
 Master: mysql_data_clone
  Meta Attrs: master-max=1 master-node-max=1 clone-max=2 clone-node-max=1 notify      \
=true  Resource: mysql_data (class=ocf provider=linbit type=drbd)
   Attributes: drbd_resource=mysql
   Operations: start interval=0s timeout=240 (mysql_data-start-interval-0s)
               promote interval=0s timeout=90 (mysql_data-promote-interval-0s)
               demote interval=0s timeout=90 (mysql_data-demote-interval-0s)
               stop interval=0s timeout=100 (mysql_data-stop-interval-0s)
               monitor interval=30s (mysql_data-monitor-interval-30s)

Stonith Devices:
 Resource: fence_node1_kvm (class=stonith type=fence_virsh)
  Attributes: pcmk_host_list=node1 ipaddr=10.6.7.10 action=reboot login=root \
passwd=password port=node1  Operations: monitor interval=30s \
(fence_node1_kvm-monitor-interval-30s)  Resource: fence_node2_kvm (class=stonith \
type=fence_virsh)  Attributes: pcmk_host_list=node2 ipaddr=10.6.7.12 action=reboot \
login=root passwd=password port=awpnode2 delay=15  Operations: monitor interval=30s \
(fence_awpnode2_kvm-monitor-interval-30s) Fencing Levels:

Location Constraints:
  Resource: mysql_data_clone
    Constraint: drbd-fence-by-handler-mysql-mysql_data_clone
      Rule: score=-INFINITY role=Master  \
                (id:drbd-fence-by-handler-mysql-rule-mysql_data_clone)
        Expression: #uname ne cleardata-awpnode2.awarepoint.com  \
(id:drbd-fence-by-handler-mysql-expr-mysql_data_clone) Ordering Constraints:
Colocation Constraints:

Resources Defaults:
 resource-stickiness: 200
Operations Defaults:
 No defaults set

Cluster Properties:
 cluster-infrastructure: cman
 dc-version: 1.1.14-8.el6_8.1-70404b0
 have-watchdog: false
 last-lrm-refresh: 1472768147
 no-quorum-policy: ignore
 stonith-enabled: true

Thank you.

Neil Schneider
DevOps Engineer

_______________________________________________
drbd-user mailing list
drbd-user@lists.linbit.com
http://lists.linbit.com/mailman/listinfo/drbd-user


[prev in list] [next in list] [prev in thread] [next in thread] 

Configure | About | News | Add a list | Sponsored by KoreLogic