High-Availability
crm 狀態:已配置 0 個節點和 0 個資源
我已經配置了 corosync,然後配置了這些問題 0 個節點和 0 個資源。
[root@sitampan2 verdana]# crm status Stack: unknown Current DC: NONE Last updated: Thu Jun 16 15:15:40 2016 Last change: Wed Jun 15 14:25:43 2016 by root via cibadmin on sitampan2 0 nodes and 0 resources configured: 160547800 resources DISABLED and 4352 BLOCKED from being started due to failures Full list of resources:
這是我的 corosync 配置:
compatibility: whitetank aisexec { # Run as root - this is necessary to be able to manage resources with Pacemaker user: root group: root } service { # Load the Pacemaker Cluster Resource Manager ver: 1 name: pacemaker use_mgmtd: no use_logd: no } totem { version: 2 #How long before declaring a token lost (ms) token: 5000 # How many token retransmits before forming a new configuration token_retransmits_before_loss_const: 10 # How long to wait for join messages in the membership protocol (ms) join: 1000 # How long to wait for consensus to be achieved before starting a new # round of membership configuration (ms) consensus: 7500 # Turn off the virtual synchrony filter vsftype: none # Number of messages that may be sent by one processor on receipt of the token max_messages: 20 # Stagger sending the node join messages by 1..send_join ms send_join: 45 # Limit generated nodeids to 31-bits (positive signed integers) clear_node_high_bit: yes # Disable encryption secauth: off # How many threads to use for encryption/decryption threads: 0 # Optionally assign a fixed node id (integer) # nodeid: 1234interface { interface { ringnumber: 0 member { memberaddr: 192.168.88.202 } member { memberaddr: 192.168.88.201 } bindnetaddr: 192.168.88.0 mcastaddr: 226.94.1.1 mcastport: 5405 ttl: 1 } } logging { fileline: off to_stderr: no to_logfile: yes to_syslog: yes logfile: /var/log/cluster/corosync.log debug: off timestamp: on logger_subsys { subsys: AMF debug: off } } amf { mode: disabled }
我正在使用centos 6.7
這是 corosync 日誌文件:
Jun 15 14:25:22 [1461] sitampan2 corosync notice [MAIN ] Node was shut down by a signal Jun 15 14:25:22 [1461] sitampan2 corosync notice [SERV ] Unloading all Corosync service engines. Jun 15 14:25:22 [1461] sitampan2 corosync info [QB ] withdrawing server sockets Jun 15 14:25:22 [1461] sitampan2 corosync notice [SERV ] Service engine unloaded: corosync configuration map access Jun 15 14:25:23 [13024] sitampan2 pacemakerd: error: cfg_connection_destroy: Connection destroyed Jun 15 14:25:23 [13024] sitampan2 pacemakerd: notice: pcmk_shutdown_worker: Shutting down Pacemaker Jun 15 14:25:23 [13024] sitampan2 pacemakerd: notice: stop_child: Stopping pengine: Sent -15 to process 13030 Jun 15 14:25:23 [13030] sitampan2 pengine: notice: crm_signal_dispatch: Invoking handler for signal 15: Terminated Jun 15 14:25:23 [1461] sitampan2 corosync info [QB ] withdrawing server sockets Jun 15 14:25:23 [13030] sitampan2 pengine: info: qb_ipcs_us_withdraw: withdrawing server sockets Jun 15 14:25:23 [1461] sitampan2 corosync notice [SERV ] Service engine unloaded: corosync configuration service Jun 15 14:25:23 [13030] sitampan2 pengine: info: crm_xml_cleanup: Cleaning up memory from libxml2 Jun 15 14:25:23 [13024] sitampan2 pacemakerd: info: pcmk_child_exit: The pengine process (13030) exited: OK (0) Jun 15 14:25:23 [13024] sitampan2 pacemakerd: notice: stop_child: Stopping attrd: Sent -15 to process 13029 Jun 15 14:25:23 [13029] sitampan2 attrd: notice: crm_signal_dispatch: Invoking handler for signal 15: Terminated Jun 15 14:25:23 [13029] sitampan2 attrd: info: attrd_shutdown: Shutting down Jun 15 14:25:23 [13029] sitampan2 attrd: notice: main: Cleaning up before exit Jun 15 14:25:23 [13029] sitampan2 attrd: info: qb_ipcs_us_withdraw: withdrawing server sockets Jun 15 14:25:23 [13029] sitampan2 attrd: info: attrd_cib_destroy_cb: Connection disconnection complete Jun 15 14:25:23 [13029] sitampan2 attrd: info: crm_xml_cleanup: Cleaning up memory from libxml2 Jun 15 14:25:23 [13026] sitampan2 cib: error: pcmk_cpg_dispatch: Connection to the CPG API failed: Library error (2) Jun 15 14:25:23 [13026] sitampan2 cib: error: cib_cs_destroy: Corosync connection lost! Exiting. Jun 15 14:25:23 [13026] sitampan2 cib: info: terminate_cib: cib_cs_destroy: Exiting fast... Jun 15 14:25:23 [13024] sitampan2 pacemakerd: info: pcmk_child_exit: The attrd process (13029) exited: OK (0) Jun 15 14:25:23 [13026] sitampan2 cib: info: qb_ipcs_us_withdraw: withdrawing server sockets Jun 15 14:25:23 [13024] sitampan2 pacemakerd: notice: stop_child: Stopping lrmd: Sent -15 to process 13028 Jun 15 14:25:23 [13027] sitampan2 stonith-ng: error: crm_ipc_read: Connection to cib_rw failed Jun 15 14:25:23 [13026] sitampan2 cib: info: qb_ipcs_us_withdraw: withdrawing server sockets Jun 15 14:25:23 [13027] sitampan2 stonith-ng: error: mainloop_gio_callback: Connection to cib_rw[0x96c2c50] closed (I/O condition=17) Jun 15 14:25:23 [13028] sitampan2 lrmd: notice: crm_signal_dispatch: Invoking handler for signal 15: Terminated Jun 15 14:25:23 [13026] sitampan2 cib: info: qb_ipcs_us_withdraw: withdrawing server sockets Jun 15 14:25:23 [13028] sitampan2 lrmd: info: lrmd_exit: Terminating with 0 clients Jun 15 14:25:23 [13028] sitampan2 lrmd: info: qb_ipcs_us_withdraw: withdrawing server sockets Jun 15 14:25:23 [13026] sitampan2 cib: info: crm_xml_cleanup: Cleaning up memory from libxml2 Jun 15 14:25:23 [13028] sitampan2 lrmd: info: crm_xml_cleanup: Cleaning up memory from libxml2 Jun 15 14:25:23 [13027] sitampan2 stonith-ng: notice: cib_connection_destroy: Connection to the CIB terminated. Shutting down. Jun 15 14:25:23 [13027] sitampan2 stonith-ng: info: stonith_shutdown: Terminating with 0 clients Jun 15 14:25:23 [13027] sitampan2 stonith-ng: info: qb_ipcs_us_withdraw: withdrawing server sockets Jun 15 14:25:23 [13027] sitampan2 stonith-ng: info: main: Done Jun 15 14:25:23 [13027] sitampan2 stonith-ng: info: qb_ipcs_us_withdraw: withdrawing server sockets Jun 15 14:25:23 [13027] sitampan2 stonith-ng: info: main: Done Jun 15 14:25:23 [13027] sitampan2 stonith-ng: info: crm_xml_cleanup: Cleaning up memory from libxml2 Jun 15 14:25:23 [13024] sitampan2 pacemakerd: info: pcmk_child_exit: The lrmd process (13028) exited: OK (0) Jun 15 14:25:23 [13024] sitampan2 pacemakerd: notice: stop_child: Stopping stonith-ng: Sent -15 to process 13027 Jun 15 14:25:23 [13024] sitampan2 pacemakerd: error: pcmk_cpg_dispatch: Connection to the CPG API failed: Library error (2) Jun 15 14:25:23 [13024] sitampan2 pacemakerd: error: mcp_cpg_destroy: Connection destroyed Jun 15 14:25:23 [13024] sitampan2 pacemakerd: info: crm_xml_cleanup: Cleaning up memory from libxml2 Jun 15 14:25:23 [1461] sitampan2 corosync info [QB ] withdrawing server sockets Jun 15 14:25:23 [1461] sitampan2 corosync notice [SERV ] Service engine unloaded: corosync cluster closed process group service v1.01 Jun 15 14:25:23 [1461] sitampan2 corosync info [QB ] withdrawing server sockets Jun 15 14:25:23 [1461] sitampan2 corosync notice [SERV ] Service engine unloaded: corosync cluster quorum service v0.1 Jun 15 14:25:23 [1461] sitampan2 corosync notice [SERV ] Service engine unloaded: corosync profile loading service Jun 15 14:25:23 [1461] sitampan2 corosync notice [MAIN ] Corosync Cluster Engine exiting normally
這是起搏器日誌:
Jun 20 11:23:32 [2006] sitampan2 lrmd: notice: crm_signal_dispatch: Invoking handler for signal 15: Terminated Jun 20 11:23:32 [2006] sitampan2 lrmd: info: lrmd_exit: Terminating with 0 clients Jun 20 11:23:32 [2006] sitampan2 lrmd: info: qb_ipcs_us_withdraw: withdrawing server sockets Jun 20 11:23:32 [2006] sitampan2 lrmd: info: crm_xml_cleanup: Cleaning up memory from libxml2 Jun 20 11:23:42 [8620] sitampan2 pacemakerd: info: get_cluster_type: Detected an active 'corosync' cluster Jun 20 11:23:42 [8620] sitampan2 pacemakerd: info: mcp_read_config: Reading configure for stack: corosync Jun 20 11:23:42 [8620] sitampan2 pacemakerd: notice: crm_add_logfile: Switching to /var/log/cluster/corosync.log Jun 20 11:29:10 [1985] sitampan2 pacemakerd: info: crm_log_init: Changed active directory to /var/lib/pacemaker/cores/root Jun 20 11:29:10 [1985] sitampan2 pacemakerd: info: get_cluster_type: Detected an active 'corosync' cluster Jun 20 11:29:10 [1985] sitampan2 pacemakerd: info: mcp_read_config: Reading configure for stack: corosync Jun 20 11:29:10 [1985] sitampan2 pacemakerd: notice: crm_add_logfile: Switching to /var/log/cluster/corosync.log Jun 20 11:29:24 [2093] sitampan2 pacemakerd: info: get_cluster_type: Detected an active 'corosync' cluster Jun 20 11:29:24 [2093] sitampan2 pacemakerd: info: mcp_read_config: Reading configure for stack: corosync Jun 20 11:29:24 [2093] sitampan2 pacemakerd: notice: crm_add_logfile: Switching to /var/log/cluster/corosync.log
這是 server2 中的 cibadmin –query > config.txt 失敗的結果:
<cib crm_feature_set="3.0.10" validate-with="pacemaker-2.4" epoch="1" num_updates="0" admin_epoch="0" cib-last-written="Fri Jun 17 15:10:08 2016" update-origin="sitamp$ <configuration> <crm_config> <cluster_property_set id="cib-bootstrap-options"> <nvpair name="no-quorum-policy" value="ignore" id="cib-bootstrap-options-no-quorum-policy"/> <nvpair name="stonith-enabled" value="false" id="cib-bootstrap-options-stonith-enabled"/> <nvpair name="default-resource-stickiness" value="1000" id="cib-bootstrap-options-default-resource-stickiness"/> </cluster_property_set> </crm_config> <nodes/> <resources/> <constraints/> </configuration> <status/> </cib>
這是 server1 中 cibadmin –query > config.txt 的結果:
<cib crm_feature_set="3.0.10" validate-with="pacemaker-2.4" epoch="4" num_updates="5" admin_epoch="0" cib-last-written="Fri Jun 17 13:16:36 2016" update-origin="sitamp$ <configuration> <crm_config> <cluster_property_set id="cib-bootstrap-options"> <nvpair id="cib-bootstrap-options-have-watchdog" name="have-watchdog" value="true"/> <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" value="1.1.14-03fa134"/> <nvpair id="cib-bootstrap-options-cluster-infrastructure" name="cluster-infrastructure" value="corosync"/> </cluster_property_set> </crm_config> <nodes> <node id="1084774601" uname="sitampan1"/> </nodes> <resources/> <constraints/> </configuration> <status> <node_state id="1084774601" uname="sitampan1" in_ccm="true" crmd="online" crm-debug-origin="do_state_transition" join="member" expected="member"> <lrm id="1084774601"> <lrm_resources/> </lrm> <transient_attributes id="1084774601"> <instance_attributes id="status-1084774601"> <nvpair id="status-1084774601-shutdown" name="shutdown" value="0"/> </instance_attributes> </transient_attributes> </node_state> </status> </cib>
我已經重新安裝並且 server1 和 server2 已啟動但未同步。這裡是列印螢幕。
有人可以幫助我嗎?
看來你只有兩個節點,你想添加這個:
quorum { provider: corosync_votequorum expected_votes: 2 }
在圖騰部分,你可以嘗試添加這一行:
rrp_mode: none
如果仍然有錯誤,請嘗試添加更多此部分:
quorum { provider: corosync_votequorum expected_votes: 3 }