Re: new to pacemaker and heartbeat on debian...getting error..
Okay..I was able to get my server back to its original problem state:
here is what i have installed:
root@deb1:/etc/ha.d# dpkg -l |grep pacemaker
ii pacemaker 1.0.9.1+hg15626-1
HA cluster resource manager
root@deb1:/etc/ha.d# dpkg -l |grep heartbeat
ii heartbeat 1:3.0.3-2
Subsystem for High-Availability Linux
ii libheartbeat2 1:3.0.3-2
Subsystem for High-Availability Linux (libraries)
ii libheartbeat2-dev 1:3.0.3-2
Subsystem for High-Availability Linux (development files)
root@deb1:/etc/ha.d# dpkg -l |grep corosync
ii corosync 1.2.1-4
Standards-based cluster framework (daemon and modules)
ii libcorosync-dev 1.2.1-4
Standards-based cluster framework (developer files)
ii libcorosync4 1.2.1-4
Standards-based cluster framework (libraries)
Can not really tell if i am using heartbeat 1 or 2...it looks like it
is 2 by the lib files but debian package says ver. 1.
When i run the corosync command, i get this :
root@deb1:/etc/ha.d# crm_mon --one-shot
============
Last updated: Mon Oct 24 07:57:32 2011
Stack: openais
Current DC: deb1 - partition with quorum
Version: 1.0.9-74392a28b7f31d7ddc86689598bd23114f58978b
2 Nodes configured, 2 expected votes
1 Resources configured.
============
Online: [ deb1 deb2 ]
Failed actions:
failover-ip_start_0 (node=deb1, call=3, rc=1, status=complete):
unknown error
failover-ip_start_0 (node=deb2, call=41, rc=1, status=complete):
unknown error
when i run the command you said earlier, i get this:
root@deb1:/etc/ha.d# cibadmin -Q
<cib epoch="19" num_updates="50" admin_epoch="0"
validate-with="pacemaker-1.0" crm_feature_set="3.0.1" have-quorum="1"
cib-last-written="Sun Oct 23 18:58:30 2011" dc-uuid="deb1">
<configuration>
<crm_config>
<cluster_property_set id="cib-bootstrap-options">
<nvpair id="cib-bootstrap-options-dc-version"
name="dc-version"
value="1.0.9-74392a28b7f31d7ddc86689598bd23114f58978b"/>
<nvpair id="cib-bootstrap-options-cluster-infrastructure"
name="cluster-infrastructure" value="openais"/>
<nvpair id="cib-bootstrap-options-expected-quorum-votes"
name="expected-quorum-votes" value="2"/>
<nvpair id="cib-bootstrap-options-stonith-enabled"
name="stonith-enabled" value="false"/>
</cluster_property_set>
</crm_config>
<nodes>
<node id="deb1" uname="deb1" type="normal"/>
<node id="deb2" uname="deb2" type="normal"/>
</nodes>
<resources>
<primitive class="ocf" id="failover-ip" provider="heartbeat"
type="IPaddr">
<instance_attributes id="failover-ip-instance_attributes">
<nvpair id="failover-ip-instance_attributes-ip" name="ip"
value="192.168.2.113"/>
</instance_attributes>
<operations>
<op id="failover-ip-monitor-10s" interval="10s" name="monitor"/>
</operations>
</primitive>
</resources>
<constraints/>
<rsc_defaults/>
<op_defaults/>
</configuration>
<status>
<node_state id="deb1" uname="deb1" ha="active" in_ccm="true"
crmd="online" join="member" expected="member"
crm-debug-origin="do_state_transition" shutdown="0">
<transient_attributes id="deb1">
<instance_attributes id="status-deb1">
<nvpair id="status-deb1-probe_complete"
name="probe_complete" value="true"/>
<nvpair id="status-deb1-fail-count-failover-ip"
name="fail-count-failover-ip" value="INFINITY"/>
<nvpair id="status-deb1-last-failure-failover-ip"
name="last-failure-failover-ip" value="1319410734"/>
</instance_attributes>
</transient_attributes>
<lrm id="deb1">
<lrm_resources>
<lrm_resource id="failover-ip" type="IPaddr" class="ocf"
provider="heartbeat">
<lrm_rsc_op id="failover-ip_monitor_0" operation="monitor"
crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1"
transition-key="4:10:7:2413e953-0f5c-415b-9705-55d1ab7df826"
transition-magic="0:7;4:10:7:2413e953-0f5c-415b-9705-55d1ab7df826"
call-id="2" rc-code="7" op-status="0" interval="0"
last-run="1319410733" last-rc-change="1319410733" exec-time="60"
queue-time="0" op-digest="fe94f52163446180249c8ee0fc7615d9"/>
<lrm_rsc_op id="failover-ip_start_0" operation="start"
crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1"
transition-key="7:10:0:2413e953-0f5c-415b-9705-55d1ab7df826"
transition-magic="0:1;7:10:0:2413e953-0f5c-415b-9705-55d1ab7df826"
call-id="3" rc-code="1" op-status="0" interval="0"
last-run="1319410733" last-rc-change="1319410733" exec-time="90"
queue-time="0" op-digest="fe94f52163446180249c8ee0fc7615d9"/>
<lrm_rsc_op id="failover-ip_stop_0" operation="stop"
crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1"
transition-key="1:12:0:2413e953-0f5c-415b-9705-55d1ab7df826"
transition-magic="0:0;1:12:0:2413e953-0f5c-415b-9705-55d1ab7df826"
call-id="4" rc-code="0" op-status="0" interval="0"
last-run="1319410733" last-rc-change="1319410733" exec-time="40"
queue-time="0" op-digest="fe94f52163446180249c8ee0fc7615d9"/>
</lrm_resource>
</lrm_resources>
</lrm>
</node_state>
<node_state id="deb2" uname="deb2" ha="active" in_ccm="true"
crmd="online" crm-debug-origin="do_update_resource" join="member"
expected="member" shutdown="0">
<lrm id="deb2">
<lrm_resources>
<lrm_resource id="failover-ip" type="IPaddr" class="ocf"
provider="heartbeat">
<lrm_rsc_op id="failover-ip_monitor_0" operation="monitor"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="5:22:7:2413e953-0f5c-415b-9705-55d1ab7df826"
transition-magic="0:7;5:22:7:2413e953-0f5c-415b-9705-55d1ab7df826"
call-id="2" rc-code="7" op-status="0" interval="0"
last-run="1319415286" last-rc-change="1319415286" exec-time="30"
queue-time="0" op-digest="fe94f52163446180249c8ee0fc7615d9"/>
<lrm_rsc_op id="failover-ip_start_0" operation="start"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="6:41:0:2413e953-0f5c-415b-9705-55d1ab7df826"
transition-magic="0:1;6:41:0:2413e953-0f5c-415b-9705-55d1ab7df826"
call-id="41" rc-code="1" op-status="0" interval="0"
last-run="1319415289" last-rc-change="1319415289" exec-time="80"
queue-time="0" op-digest="fe94f52163446180249c8ee0fc7615d9"/>
<lrm_rsc_op id="failover-ip_stop_0" operation="stop"
crm-debug-origin="do_update_resource" crm_feature_set="3.0.1"
transition-key="1:43:0:2413e953-0f5c-415b-9705-55d1ab7df826"
transition-magic="0:0;1:43:0:2413e953-0f5c-415b-9705-55d1ab7df826"
call-id="42" rc-code="0" op-status="0" interval="0"
last-run="1319415289" last-rc-change="1319415289" exec-time="20"
queue-time="0" op-digest="fe94f52163446180249c8ee0fc7615d9"/>
</lrm_resource>
</lrm_resources>
</lrm>
<transient_attributes id="deb2">
<instance_attributes id="status-deb2">
<nvpair id="status-deb2-fail-count-failover-ip"
name="fail-count-failover-ip" value="INFINITY"/>
<nvpair id="status-deb2-last-failure-failover-ip"
name="last-failure-failover-ip" value="1319415290"/>
<nvpair id="status-deb2-probe_complete"
name="probe_complete" value="true"/>
</instance_attributes>
</transient_attributes>
</node_state>
</status>
</cib>
hope this helps..will try to get a clean syslog for you.
thanks much for your efforts.
Reply to: