我已经配置了 corosync,然后我得到了这些问题 0 个节点和 0 个资源配置。
[root@sitampan2 verdana]# crm status
Stack: unknown
Current DC: NONE
Last updated: Thu Jun 16 15:15:40 2016 Last change: Wed Jun 15 14:25:43 2016 by root via cibadmin on sitampan2
0 nodes and 0 resources configured: 160547800 resources DISABLED and 4352 BLOCKED from being started due to failures
Full list of resources:
这是我的 corosync 配置:
compatibility: whitetank
aisexec {
# Run as root - this is necessary to be able to manage resources with Pacemaker
user: root
group: root
}
service {
# Load the Pacemaker Cluster Resource Manager
ver: 1
name: pacemaker
use_mgmtd: no
use_logd: no
}
totem {
version: 2
#How long before declaring a token lost (ms)
token: 5000
# How many token retransmits before forming a new configuration
token_retransmits_before_loss_const: 10
# How long to wait for join messages in the membership protocol (ms)
join: 1000
# How long to wait for consensus to be achieved before starting a new
# round of membership configuration (ms)
consensus: 7500
# Turn off the virtual synchrony filter
vsftype: none
# Number of messages that may be sent by one processor on receipt of the token
max_messages: 20
# Stagger sending the node join messages by 1..send_join ms
send_join: 45
# Limit generated nodeids to 31-bits (positive signed integers)
clear_node_high_bit: yes
# Disable encryption
secauth: off
# How many threads to use for encryption/decryption
threads: 0
# Optionally assign a fixed node id (integer)
# nodeid: 1234interface {
interface {
ringnumber: 0
member {
memberaddr: 192.168.88.202
}
member {
memberaddr: 192.168.88.201
}
bindnetaddr: 192.168.88.0
mcastaddr: 226.94.1.1
mcastport: 5405
ttl: 1
}
}
logging {
fileline: off
to_stderr: no
to_logfile: yes
to_syslog: yes
logfile: /var/log/cluster/corosync.log
debug: off
timestamp: on
logger_subsys {
subsys: AMF
debug: off
}
}
amf {
mode: disabled
}
我正在使用 centos 6.7
这是 corosync 日志文件:
Jun 15 14:25:22 [1461] sitampan2 corosync notice [MAIN ] Node was shut down by a signal
Jun 15 14:25:22 [1461] sitampan2 corosync notice [SERV ] Unloading all Corosync service engines.
Jun 15 14:25:22 [1461] sitampan2 corosync info [QB ] withdrawing server sockets
Jun 15 14:25:22 [1461] sitampan2 corosync notice [SERV ] Service engine unloaded: corosync configuration map access
Jun 15 14:25:23 [13024] sitampan2 pacemakerd: error: cfg_connection_destroy: Connection destroyed
Jun 15 14:25:23 [13024] sitampan2 pacemakerd: notice: pcmk_shutdown_worker: Shutting down Pacemaker
Jun 15 14:25:23 [13024] sitampan2 pacemakerd: notice: stop_child: Stopping pengine: Sent -15 to process 13030
Jun 15 14:25:23 [13030] sitampan2 pengine: notice: crm_signal_dispatch: Invoking handler for signal 15: Terminated
Jun 15 14:25:23 [1461] sitampan2 corosync info [QB ] withdrawing server sockets
Jun 15 14:25:23 [13030] sitampan2 pengine: info: qb_ipcs_us_withdraw: withdrawing server sockets
Jun 15 14:25:23 [1461] sitampan2 corosync notice [SERV ] Service engine unloaded: corosync configuration service
Jun 15 14:25:23 [13030] sitampan2 pengine: info: crm_xml_cleanup: Cleaning up memory from libxml2
Jun 15 14:25:23 [13024] sitampan2 pacemakerd: info: pcmk_child_exit: The pengine process (13030) exited: OK (0)
Jun 15 14:25:23 [13024] sitampan2 pacemakerd: notice: stop_child: Stopping attrd: Sent -15 to process 13029
Jun 15 14:25:23 [13029] sitampan2 attrd: notice: crm_signal_dispatch: Invoking handler for signal 15: Terminated
Jun 15 14:25:23 [13029] sitampan2 attrd: info: attrd_shutdown: Shutting down
Jun 15 14:25:23 [13029] sitampan2 attrd: notice: main: Cleaning up before exit
Jun 15 14:25:23 [13029] sitampan2 attrd: info: qb_ipcs_us_withdraw: withdrawing server sockets
Jun 15 14:25:23 [13029] sitampan2 attrd: info: attrd_cib_destroy_cb: Connection disconnection complete
Jun 15 14:25:23 [13029] sitampan2 attrd: info: crm_xml_cleanup: Cleaning up memory from libxml2
Jun 15 14:25:23 [13026] sitampan2 cib: error: pcmk_cpg_dispatch: Connection to the CPG API failed: Library error (2)
Jun 15 14:25:23 [13026] sitampan2 cib: error: cib_cs_destroy: Corosync connection lost! Exiting.
Jun 15 14:25:23 [13026] sitampan2 cib: info: terminate_cib: cib_cs_destroy: Exiting fast...
Jun 15 14:25:23 [13024] sitampan2 pacemakerd: info: pcmk_child_exit: The attrd process (13029) exited: OK (0)
Jun 15 14:25:23 [13026] sitampan2 cib: info: qb_ipcs_us_withdraw: withdrawing server sockets
Jun 15 14:25:23 [13024] sitampan2 pacemakerd: notice: stop_child: Stopping lrmd: Sent -15 to process 13028
Jun 15 14:25:23 [13027] sitampan2 stonith-ng: error: crm_ipc_read: Connection to cib_rw failed
Jun 15 14:25:23 [13026] sitampan2 cib: info: qb_ipcs_us_withdraw: withdrawing server sockets
Jun 15 14:25:23 [13027] sitampan2 stonith-ng: error: mainloop_gio_callback: Connection to cib_rw[0x96c2c50] closed (I/O condition=17)
Jun 15 14:25:23 [13028] sitampan2 lrmd: notice: crm_signal_dispatch: Invoking handler for signal 15: Terminated
Jun 15 14:25:23 [13026] sitampan2 cib: info: qb_ipcs_us_withdraw: withdrawing server sockets
Jun 15 14:25:23 [13028] sitampan2 lrmd: info: lrmd_exit: Terminating with 0 clients
Jun 15 14:25:23 [13028] sitampan2 lrmd: info: qb_ipcs_us_withdraw: withdrawing server sockets
Jun 15 14:25:23 [13026] sitampan2 cib: info: crm_xml_cleanup: Cleaning up memory from libxml2
Jun 15 14:25:23 [13028] sitampan2 lrmd: info: crm_xml_cleanup: Cleaning up memory from libxml2
Jun 15 14:25:23 [13027] sitampan2 stonith-ng: notice: cib_connection_destroy: Connection to the CIB terminated. Shutting down.
Jun 15 14:25:23 [13027] sitampan2 stonith-ng: info: stonith_shutdown: Terminating with 0 clients
Jun 15 14:25:23 [13027] sitampan2 stonith-ng: info: qb_ipcs_us_withdraw: withdrawing server sockets
Jun 15 14:25:23 [13027] sitampan2 stonith-ng: info: main: Done
Jun 15 14:25:23 [13027] sitampan2 stonith-ng: info: qb_ipcs_us_withdraw: withdrawing server sockets
Jun 15 14:25:23 [13027] sitampan2 stonith-ng: info: main: Done
Jun 15 14:25:23 [13027] sitampan2 stonith-ng: info: crm_xml_cleanup: Cleaning up memory from libxml2
Jun 15 14:25:23 [13024] sitampan2 pacemakerd: info: pcmk_child_exit: The lrmd process (13028) exited: OK (0)
Jun 15 14:25:23 [13024] sitampan2 pacemakerd: notice: stop_child: Stopping stonith-ng: Sent -15 to process 13027
Jun 15 14:25:23 [13024] sitampan2 pacemakerd: error: pcmk_cpg_dispatch: Connection to the CPG API failed: Library error (2)
Jun 15 14:25:23 [13024] sitampan2 pacemakerd: error: mcp_cpg_destroy: Connection destroyed
Jun 15 14:25:23 [13024] sitampan2 pacemakerd: info: crm_xml_cleanup: Cleaning up memory from libxml2
Jun 15 14:25:23 [1461] sitampan2 corosync info [QB ] withdrawing server sockets
Jun 15 14:25:23 [1461] sitampan2 corosync notice [SERV ] Service engine unloaded: corosync cluster closed process group service v1.01
Jun 15 14:25:23 [1461] sitampan2 corosync info [QB ] withdrawing server sockets
Jun 15 14:25:23 [1461] sitampan2 corosync notice [SERV ] Service engine unloaded: corosync cluster quorum service v0.1
Jun 15 14:25:23 [1461] sitampan2 corosync notice [SERV ] Service engine unloaded: corosync profile loading service
Jun 15 14:25:23 [1461] sitampan2 corosync notice [MAIN ] Corosync Cluster Engine exiting normally
这是起搏器日志:
Jun 20 11:23:32 [2006] sitampan2 lrmd: notice: crm_signal_dispatch: Invoking handler for signal 15: Terminated
Jun 20 11:23:32 [2006] sitampan2 lrmd: info: lrmd_exit: Terminating with 0 clients
Jun 20 11:23:32 [2006] sitampan2 lrmd: info: qb_ipcs_us_withdraw: withdrawing server sockets
Jun 20 11:23:32 [2006] sitampan2 lrmd: info: crm_xml_cleanup: Cleaning up memory from libxml2
Jun 20 11:23:42 [8620] sitampan2 pacemakerd: info: get_cluster_type: Detected an active 'corosync' cluster
Jun 20 11:23:42 [8620] sitampan2 pacemakerd: info: mcp_read_config: Reading configure for stack: corosync
Jun 20 11:23:42 [8620] sitampan2 pacemakerd: notice: crm_add_logfile: Switching to /var/log/cluster/corosync.log
Jun 20 11:29:10 [1985] sitampan2 pacemakerd: info: crm_log_init: Changed active directory to /var/lib/pacemaker/cores/root
Jun 20 11:29:10 [1985] sitampan2 pacemakerd: info: get_cluster_type: Detected an active 'corosync' cluster
Jun 20 11:29:10 [1985] sitampan2 pacemakerd: info: mcp_read_config: Reading configure for stack: corosync
Jun 20 11:29:10 [1985] sitampan2 pacemakerd: notice: crm_add_logfile: Switching to /var/log/cluster/corosync.log
Jun 20 11:29:24 [2093] sitampan2 pacemakerd: info: get_cluster_type: Detected an active 'corosync' cluster
Jun 20 11:29:24 [2093] sitampan2 pacemakerd: info: mcp_read_config: Reading configure for stack: corosync
Jun 20 11:29:24 [2093] sitampan2 pacemakerd: notice: crm_add_logfile: Switching to /var/log/cluster/corosync.log
这是 server2 中 cibadmin --query > config.txt 失败的结果:
<cib crm_feature_set="3.0.10" validate-with="pacemaker-2.4" epoch="1" num_updates="0" admin_epoch="0" cib-last-written="Fri Jun 17 15:10:08 2016" update-origin="sitamp$
<configuration>
<crm_config>
<cluster_property_set id="cib-bootstrap-options">
<nvpair name="no-quorum-policy" value="ignore" id="cib-bootstrap-options-no-quorum-policy"/>
<nvpair name="stonith-enabled" value="false" id="cib-bootstrap-options-stonith-enabled"/>
<nvpair name="default-resource-stickiness" value="1000" id="cib-bootstrap-options-default-resource-stickiness"/>
</cluster_property_set>
</crm_config>
<nodes/>
<resources/>
<constraints/>
</configuration>
<status/>
</cib>
这是在 server1 中执行 cibadmin --query > config.txt 的结果:
<cib crm_feature_set="3.0.10" validate-with="pacemaker-2.4" epoch="4" num_updates="5" admin_epoch="0" cib-last-written="Fri Jun 17 13:16:36 2016" update-origin="sitamp$
<configuration>
<crm_config>
<cluster_property_set id="cib-bootstrap-options">
<nvpair id="cib-bootstrap-options-have-watchdog" name="have-watchdog" value="true"/>
<nvpair id="cib-bootstrap-options-dc-version" name="dc-version" value="1.1.14-03fa134"/>
<nvpair id="cib-bootstrap-options-cluster-infrastructure" name="cluster-infrastructure" value="corosync"/>
</cluster_property_set>
</crm_config>
<nodes>
<node id="1084774601" uname="sitampan1"/>
</nodes>
<resources/>
<constraints/>
</configuration>
<status>
<node_state id="1084774601" uname="sitampan1" in_ccm="true" crmd="online" crm-debug-origin="do_state_transition" join="member" expected="member">
<lrm id="1084774601">
<lrm_resources/>
</lrm>
<transient_attributes id="1084774601">
<instance_attributes id="status-1084774601">
<nvpair id="status-1084774601-shutdown" name="shutdown" value="0"/>
</instance_attributes>
</transient_attributes>
</node_state>
</status>
</cib>
我已经重新安装,server1 和 server2 已启动,但尚未同步。以下是屏幕截图。
有人可以帮帮我吗?
答案1
看起来您只有两个节点,您可能想要添加这个:
quorum {
provider: corosync_votequorum
expected_votes: 2
}
答案2
在图腾部分,您可以尝试添加此行:
rrp_mode: none
如果仍然有错误,请尝试添加更多此部分:
quorum {
provider: corosync_votequorum
expected_votes: 3
}