[Date Prev][Date Next] [Thread Prev][Thread Next] [Date Index] [Thread Index]

Re: new to pacemaker and heartbeat on debian...getting error..



Okay..I was able to get my server back to its original problem state:
here is what i have installed:

root@deb1:/etc/ha.d# dpkg -l |grep pacemaker
ii  pacemaker                            1.0.9.1+hg15626-1
    HA cluster resource manager
root@deb1:/etc/ha.d# dpkg -l |grep heartbeat
ii  heartbeat                            1:3.0.3-2
    Subsystem for High-Availability Linux
ii  libheartbeat2                        1:3.0.3-2
    Subsystem for High-Availability Linux (libraries)
ii  libheartbeat2-dev                    1:3.0.3-2
    Subsystem for High-Availability Linux (development files)
root@deb1:/etc/ha.d# dpkg -l |grep corosync
ii  corosync                             1.2.1-4
    Standards-based cluster framework (daemon and modules)
ii  libcorosync-dev                      1.2.1-4
    Standards-based cluster framework (developer files)
ii  libcorosync4                         1.2.1-4
    Standards-based cluster framework (libraries)

Can not really tell if i am using heartbeat 1 or 2...it looks like it
is 2 by the lib files but debian package says ver. 1.

When i run the corosync command, i get this :
root@deb1:/etc/ha.d# crm_mon --one-shot
============
Last updated: Mon Oct 24 07:57:32 2011
Stack: openais
Current DC: deb1 - partition with quorum
Version: 1.0.9-74392a28b7f31d7ddc86689598bd23114f58978b
2 Nodes configured, 2 expected votes
1 Resources configured.
============

Online: [ deb1 deb2 ]


Failed actions:
    failover-ip_start_0 (node=deb1, call=3, rc=1, status=complete):
unknown error
    failover-ip_start_0 (node=deb2, call=41, rc=1, status=complete):
unknown error


when i run the command you said earlier, i get this:
root@deb1:/etc/ha.d# cibadmin -Q
<cib epoch="19" num_updates="50" admin_epoch="0"
validate-with="pacemaker-1.0" crm_feature_set="3.0.1" have-quorum="1"
cib-last-written="Sun Oct 23 18:58:30 2011" dc-uuid="deb1">
  <configuration>
    <crm_config>
      <cluster_property_set id="cib-bootstrap-options">
        <nvpair id="cib-bootstrap-options-dc-version"
name="dc-version"
value="1.0.9-74392a28b7f31d7ddc86689598bd23114f58978b"/>
        <nvpair id="cib-bootstrap-options-cluster-infrastructure"
name="cluster-infrastructure" value="openais"/>
        <nvpair id="cib-bootstrap-options-expected-quorum-votes"
name="expected-quorum-votes" value="2"/>
        <nvpair id="cib-bootstrap-options-stonith-enabled"
name="stonith-enabled" value="false"/>
      </cluster_property_set>
    </crm_config>
    <nodes>
      <node id="deb1" uname="deb1" type="normal"/>
      <node id="deb2" uname="deb2" type="normal"/>
    </nodes>
    <resources>
      <primitive class="ocf" id="failover-ip" provider="heartbeat"
type="IPaddr">
        <instance_attributes id="failover-ip-instance_attributes">
          <nvpair id="failover-ip-instance_attributes-ip" name="ip"
value="192.168.2.113"/>
        </instance_attributes>
        <operations>
          <op id="failover-ip-monitor-10s" interval="10s" name="monitor"/>
        </operations>
      </primitive>
    </resources>
    <constraints/>
    <rsc_defaults/>
    <op_defaults/>
  </configuration>
  <status>
    <node_state id="deb1" uname="deb1" ha="active" in_ccm="true"
crmd="online" join="member" expected="member"
crm-debug-origin="do_state_transition" shutdown="0">
      <transient_attributes id="deb1">
        <instance_attributes id="status-deb1">
          <nvpair id="status-deb1-probe_complete"
name="probe_complete" value="true"/>
          <nvpair id="status-deb1-fail-count-failover-ip"
name="fail-count-failover-ip" value="INFINITY"/>
          <nvpair id="status-deb1-last-failure-failover-ip"
name="last-failure-failover-ip" value="1319410734"/>
        </instance_attributes>
      </transient_attributes>
      <lrm id="deb1">
        <lrm_resources>
          <lrm_resource id="failover-ip" type="IPaddr" class="ocf"
provider="heartbeat">
            <lrm_rsc_op id="failover-ip_monitor_0" operation="monitor"
crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1"
transition-key="4:10:7:2413e953-0f5c-415b-9705-55d1ab7df826"
transition-magic="0:7;4:10:7:2413e953-0f5c-415b-9705-55d1ab7df826"
call-id="2" rc-code="7" op-status="0" interval="0"
last-run="1319410733" last-rc-change="1319410733" exec-time="60"
queue-time="0" op-digest="fe94f52163446180249c8ee0fc7615d9"/>
            <lrm_rsc_op id="failover-ip_start_0" operation="start"
crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1"
transition-key="7:10:0:2413e953-0f5c-415b-9705-55d1ab7df826"
transition-magic="0:1;7:10:0:2413e953-0f5c-415b-9705-55d1ab7df826"
call-id="3" rc-code="1" op-status="0" interval="0"
last-run="1319410733" last-rc-change="1319410733" exec-time="90"
queue-time="0" op-digest="fe94f52163446180249c8ee0fc7615d9"/>
            <lrm_rsc_op id="failover-ip_stop_0" operation="stop"
crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1"
transition-key="1:12:0:2413e953-0f5c-415b-9705-55d1ab7df826"
transition-magic="0:0;1:12:0:2413e953-0f5c-415b-9705-55d1ab7df826"
call-id="4" rc-code="0" op-status="0" interval="0"
last-run="1319410733" last-rc-change="1319410733" exec-time="40"
queue-time="0" op-digest="fe94f52163446180249c8ee0fc7615d9"/>
          </lrm_resource>
        </lrm_resources>
      </lrm>
    </node_state>
    <node_state id="deb2" uname="deb2" ha="active" in_ccm="true"
crmd="online" crm-debug-origin="do_update_resource" join="member"
expected="member" shutdown="0">
      <lrm id="deb2">
        <lrm_resources>
          <lrm_resource id="failover-ip" type="IPaddr" class="ocf"
provider="heartbeat">
            <lrm_rsc_op id="failover-ip_monitor_0" operation="monitor"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="5:22:7:2413e953-0f5c-415b-9705-55d1ab7df826"
transition-magic="0:7;5:22:7:2413e953-0f5c-415b-9705-55d1ab7df826"
call-id="2" rc-code="7" op-status="0" interval="0"
last-run="1319415286" last-rc-change="1319415286" exec-time="30"
queue-time="0" op-digest="fe94f52163446180249c8ee0fc7615d9"/>
            <lrm_rsc_op id="failover-ip_start_0" operation="start"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="6:41:0:2413e953-0f5c-415b-9705-55d1ab7df826"
transition-magic="0:1;6:41:0:2413e953-0f5c-415b-9705-55d1ab7df826"
call-id="41" rc-code="1" op-status="0" interval="0"
last-run="1319415289" last-rc-change="1319415289" exec-time="80"
queue-time="0" op-digest="fe94f52163446180249c8ee0fc7615d9"/>
            <lrm_rsc_op id="failover-ip_stop_0" operation="stop"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="1:43:0:2413e953-0f5c-415b-9705-55d1ab7df826"
transition-magic="0:0;1:43:0:2413e953-0f5c-415b-9705-55d1ab7df826"
call-id="42" rc-code="0" op-status="0" interval="0"
last-run="1319415289" last-rc-change="1319415289" exec-time="20"
queue-time="0" op-digest="fe94f52163446180249c8ee0fc7615d9"/>
          </lrm_resource>
        </lrm_resources>
      </lrm>
      <transient_attributes id="deb2">
        <instance_attributes id="status-deb2">
          <nvpair id="status-deb2-fail-count-failover-ip"
name="fail-count-failover-ip" value="INFINITY"/>
          <nvpair id="status-deb2-last-failure-failover-ip"
name="last-failure-failover-ip" value="1319415290"/>
          <nvpair id="status-deb2-probe_complete"
name="probe_complete" value="true"/>
        </instance_attributes>
      </transient_attributes>
    </node_state>
  </status>
</cib>

hope this helps..will try to get a clean syslog for you.
thanks much for your efforts.


Reply to: