High-Availability

crm 狀態:已配置 0 個節點和 0 個資源

  • June 21, 2016

我已經配置了 corosync,然後配置了這些問題 0 個節點和 0 個資源。

[root@sitampan2 verdana]# crm status
Stack: unknown
Current DC: NONE
Last updated: Thu Jun 16 15:15:40 2016          Last change: Wed Jun 15 14:25:43 2016 by root via cibadmin on sitampan2

0 nodes and 0 resources configured: 160547800 resources DISABLED and 4352 BLOCKED from being started due to failures

Full list of resources:

這是我的 corosync 配置:

compatibility: whitetank
aisexec {
   # Run as root - this is necessary to be able to manage resources with Pacemaker
   user: root
   group: root
}
service {
   # Load the Pacemaker Cluster Resource Manager
   ver: 1
   name: pacemaker
   use_mgmtd: no
   use_logd: no
}
totem {
   version: 2
   #How long before declaring a token lost (ms)
       token: 5000
   # How many token retransmits before forming a new configuration
       token_retransmits_before_loss_const: 10
   # How long to wait for join messages in the membership protocol (ms)
       join: 1000
   # How long to wait for consensus to be achieved before starting a new
   # round of membership configuration (ms)
       consensus: 7500
   # Turn off the virtual synchrony filter
       vsftype: none
   # Number of messages that may be sent by one processor on receipt of the token
       max_messages: 20
   # Stagger sending the node join messages by 1..send_join ms
       send_join: 45
   # Limit generated nodeids to 31-bits (positive signed integers)
       clear_node_high_bit: yes
   # Disable encryption
       secauth: off
   # How many threads to use for encryption/decryption
       threads: 0
   # Optionally assign a fixed node id (integer)
   # nodeid: 1234interface {
   interface {
       ringnumber: 0  
       member {
           memberaddr: 192.168.88.202
       }
       member {
           memberaddr: 192.168.88.201
       }
       bindnetaddr: 192.168.88.0
               mcastaddr: 226.94.1.1
               mcastport: 5405
               ttl: 1
       }
}
logging {
   fileline: off
   to_stderr: no
   to_logfile: yes
   to_syslog: yes
   logfile: /var/log/cluster/corosync.log
   debug: off
   timestamp: on
   logger_subsys {
       subsys: AMF
       debug: off
   }
}
amf {
   mode: disabled
}

我正在使用centos 6.7

這是 corosync 日誌文件:

Jun 15 14:25:22 [1461] sitampan2 corosync notice  [MAIN  ] Node was shut down by a signal
Jun 15 14:25:22 [1461] sitampan2 corosync notice  [SERV  ] Unloading all Corosync service engines.
Jun 15 14:25:22 [1461] sitampan2 corosync info    [QB    ] withdrawing server sockets
Jun 15 14:25:22 [1461] sitampan2 corosync notice  [SERV  ] Service engine unloaded: corosync configuration map access
Jun 15 14:25:23 [13024] sitampan2 pacemakerd:    error: cfg_connection_destroy: Connection destroyed
Jun 15 14:25:23 [13024] sitampan2 pacemakerd:   notice: pcmk_shutdown_worker:   Shutting down Pacemaker
Jun 15 14:25:23 [13024] sitampan2 pacemakerd:   notice: stop_child:     Stopping pengine: Sent -15 to process 13030
Jun 15 14:25:23 [13030] sitampan2    pengine:   notice: crm_signal_dispatch:    Invoking handler for signal 15: Terminated
Jun 15 14:25:23 [1461] sitampan2 corosync info    [QB    ] withdrawing server sockets
Jun 15 14:25:23 [13030] sitampan2    pengine:     info: qb_ipcs_us_withdraw:    withdrawing server sockets
Jun 15 14:25:23 [1461] sitampan2 corosync notice  [SERV  ] Service engine unloaded: corosync configuration service
Jun 15 14:25:23 [13030] sitampan2    pengine:     info: crm_xml_cleanup:        Cleaning up memory from libxml2
Jun 15 14:25:23 [13024] sitampan2 pacemakerd:     info: pcmk_child_exit:        The pengine process (13030) exited: OK (0)
Jun 15 14:25:23 [13024] sitampan2 pacemakerd:   notice: stop_child:     Stopping attrd: Sent -15 to process 13029
Jun 15 14:25:23 [13029] sitampan2      attrd:   notice: crm_signal_dispatch:    Invoking handler for signal 15: Terminated
Jun 15 14:25:23 [13029] sitampan2      attrd:     info: attrd_shutdown: Shutting down
Jun 15 14:25:23 [13029] sitampan2      attrd:   notice: main:   Cleaning up before exit
Jun 15 14:25:23 [13029] sitampan2      attrd:     info: qb_ipcs_us_withdraw:    withdrawing server sockets
Jun 15 14:25:23 [13029] sitampan2      attrd:     info: attrd_cib_destroy_cb:   Connection disconnection complete
Jun 15 14:25:23 [13029] sitampan2      attrd:     info: crm_xml_cleanup:        Cleaning up memory from libxml2
Jun 15 14:25:23 [13026] sitampan2        cib:    error: pcmk_cpg_dispatch:      Connection to the CPG API failed: Library error (2)
Jun 15 14:25:23 [13026] sitampan2        cib:    error: cib_cs_destroy: Corosync connection lost!  Exiting.
Jun 15 14:25:23 [13026] sitampan2        cib:     info: terminate_cib:  cib_cs_destroy: Exiting fast...
Jun 15 14:25:23 [13024] sitampan2 pacemakerd:     info: pcmk_child_exit:        The attrd process (13029) exited: OK (0)
Jun 15 14:25:23 [13026] sitampan2        cib:     info: qb_ipcs_us_withdraw:    withdrawing server sockets
Jun 15 14:25:23 [13024] sitampan2 pacemakerd:   notice: stop_child:     Stopping lrmd: Sent -15 to process 13028
Jun 15 14:25:23 [13027] sitampan2 stonith-ng:    error: crm_ipc_read:   Connection to cib_rw failed
Jun 15 14:25:23 [13026] sitampan2        cib:     info: qb_ipcs_us_withdraw:    withdrawing server sockets
Jun 15 14:25:23 [13027] sitampan2 stonith-ng:    error: mainloop_gio_callback:  Connection to cib_rw[0x96c2c50] closed (I/O condition=17)
Jun 15 14:25:23 [13028] sitampan2       lrmd:   notice: crm_signal_dispatch:    Invoking handler for signal 15: Terminated
Jun 15 14:25:23 [13026] sitampan2        cib:     info: qb_ipcs_us_withdraw:    withdrawing server sockets
Jun 15 14:25:23 [13028] sitampan2       lrmd:     info: lrmd_exit:      Terminating with  0 clients
Jun 15 14:25:23 [13028] sitampan2       lrmd:     info: qb_ipcs_us_withdraw:    withdrawing server sockets
Jun 15 14:25:23 [13026] sitampan2        cib:     info: crm_xml_cleanup:        Cleaning up memory from libxml2
Jun 15 14:25:23 [13028] sitampan2       lrmd:     info: crm_xml_cleanup:        Cleaning up memory from libxml2
Jun 15 14:25:23 [13027] sitampan2 stonith-ng:   notice: cib_connection_destroy: Connection to the CIB terminated. Shutting down.
Jun 15 14:25:23 [13027] sitampan2 stonith-ng:     info: stonith_shutdown:       Terminating with  0 clients
Jun 15 14:25:23 [13027] sitampan2 stonith-ng:     info: qb_ipcs_us_withdraw:    withdrawing server sockets
Jun 15 14:25:23 [13027] sitampan2 stonith-ng:     info: main:   Done
Jun 15 14:25:23 [13027] sitampan2 stonith-ng:     info: qb_ipcs_us_withdraw:    withdrawing server sockets
Jun 15 14:25:23 [13027] sitampan2 stonith-ng:     info: main:   Done
Jun 15 14:25:23 [13027] sitampan2 stonith-ng:     info: crm_xml_cleanup:        Cleaning up memory from libxml2
Jun 15 14:25:23 [13024] sitampan2 pacemakerd:     info: pcmk_child_exit:        The lrmd process (13028) exited: OK (0)
Jun 15 14:25:23 [13024] sitampan2 pacemakerd:   notice: stop_child:     Stopping stonith-ng: Sent -15 to process 13027
Jun 15 14:25:23 [13024] sitampan2 pacemakerd:    error: pcmk_cpg_dispatch:      Connection to the CPG API failed: Library error (2)
Jun 15 14:25:23 [13024] sitampan2 pacemakerd:    error: mcp_cpg_destroy:        Connection destroyed
Jun 15 14:25:23 [13024] sitampan2 pacemakerd:     info: crm_xml_cleanup:        Cleaning up memory from libxml2
Jun 15 14:25:23 [1461] sitampan2 corosync info    [QB    ] withdrawing server sockets
Jun 15 14:25:23 [1461] sitampan2 corosync notice  [SERV  ] Service engine unloaded: corosync cluster closed process group service v1.01
Jun 15 14:25:23 [1461] sitampan2 corosync info    [QB    ] withdrawing server sockets
Jun 15 14:25:23 [1461] sitampan2 corosync notice  [SERV  ] Service engine unloaded: corosync cluster quorum service v0.1
Jun 15 14:25:23 [1461] sitampan2 corosync notice  [SERV  ] Service engine unloaded: corosync profile loading service
Jun 15 14:25:23 [1461] sitampan2 corosync notice  [MAIN  ] Corosync Cluster Engine exiting normally

這是起搏器日誌:

   Jun 20 11:23:32 [2006] sitampan2       lrmd:   notice: crm_signal_dispatch:     Invoking handler for signal 15: Terminated
Jun 20 11:23:32 [2006] sitampan2       lrmd:     info: lrmd_exit:       Terminating with  0 clients
Jun 20 11:23:32 [2006] sitampan2       lrmd:     info: qb_ipcs_us_withdraw:     withdrawing server sockets
Jun 20 11:23:32 [2006] sitampan2       lrmd:     info: crm_xml_cleanup: Cleaning up memory from libxml2
Jun 20 11:23:42 [8620] sitampan2 pacemakerd:     info: get_cluster_type:        Detected an active 'corosync' cluster
Jun 20 11:23:42 [8620] sitampan2 pacemakerd:     info: mcp_read_config: Reading configure for stack: corosync
Jun 20 11:23:42 [8620] sitampan2 pacemakerd:   notice: crm_add_logfile: Switching to /var/log/cluster/corosync.log
Jun 20 11:29:10 [1985] sitampan2 pacemakerd:     info: crm_log_init:    Changed active directory to /var/lib/pacemaker/cores/root
Jun 20 11:29:10 [1985] sitampan2 pacemakerd:     info: get_cluster_type:        Detected an active 'corosync' cluster
Jun 20 11:29:10 [1985] sitampan2 pacemakerd:     info: mcp_read_config: Reading configure for stack: corosync
Jun 20 11:29:10 [1985] sitampan2 pacemakerd:   notice: crm_add_logfile: Switching to /var/log/cluster/corosync.log
Jun 20 11:29:24 [2093] sitampan2 pacemakerd:     info: get_cluster_type:        Detected an active 'corosync' cluster
Jun 20 11:29:24 [2093] sitampan2 pacemakerd:     info: mcp_read_config: Reading configure for stack: corosync
Jun 20 11:29:24 [2093] sitampan2 pacemakerd:   notice: crm_add_logfile: Switching to /var/log/cluster/corosync.log

這是 server2 中的 cibadmin –query > config.txt 失敗的結果:

<cib crm_feature_set="3.0.10" validate-with="pacemaker-2.4" epoch="1" num_updates="0" admin_epoch="0" cib-last-written="Fri Jun 17 15:10:08 2016" update-origin="sitamp$
 <configuration>
   <crm_config>
     <cluster_property_set id="cib-bootstrap-options">
       <nvpair name="no-quorum-policy" value="ignore" id="cib-bootstrap-options-no-quorum-policy"/>
       <nvpair name="stonith-enabled" value="false" id="cib-bootstrap-options-stonith-enabled"/>
       <nvpair name="default-resource-stickiness" value="1000" id="cib-bootstrap-options-default-resource-stickiness"/>
     </cluster_property_set>
   </crm_config>
   <nodes/>
   <resources/>
   <constraints/>
 </configuration>
 <status/>
</cib>

這是 server1 中 cibadmin –query > config.txt 的結果:

<cib crm_feature_set="3.0.10" validate-with="pacemaker-2.4" epoch="4" num_updates="5" admin_epoch="0" cib-last-written="Fri Jun 17 13:16:36 2016" update-origin="sitamp$
 <configuration>
   <crm_config>
     <cluster_property_set id="cib-bootstrap-options">
       <nvpair id="cib-bootstrap-options-have-watchdog" name="have-watchdog" value="true"/>
       <nvpair id="cib-bootstrap-options-dc-version" name="dc-version" value="1.1.14-03fa134"/>
       <nvpair id="cib-bootstrap-options-cluster-infrastructure" name="cluster-infrastructure" value="corosync"/>
     </cluster_property_set>
   </crm_config>
   <nodes>
     <node id="1084774601" uname="sitampan1"/>
   </nodes>
   <resources/>
   <constraints/>
 </configuration>
 <status>
   <node_state id="1084774601" uname="sitampan1" in_ccm="true" crmd="online" crm-debug-origin="do_state_transition" join="member" expected="member">
     <lrm id="1084774601">
       <lrm_resources/>
     </lrm>
     <transient_attributes id="1084774601">
       <instance_attributes id="status-1084774601">
         <nvpair id="status-1084774601-shutdown" name="shutdown" value="0"/>
       </instance_attributes>
     </transient_attributes>
   </node_state>
 </status>
</cib>

我已經重新安裝並且 server1 和 server2 已啟動但未同步。這裡是列印螢幕。

伺服器 1 伺服器2

有人可以幫助我嗎?

看來你只有兩個節點,你想添加這個:

quorum {
 provider: corosync_votequorum
 expected_votes: 2
}

在圖騰部分,你可以嘗試添加這一行:

rrp_mode: none

如果仍然有錯誤,請嘗試添加更多此部分:

quorum {
       provider: corosync_votequorum
       expected_votes: 3
}

引用自:https://serverfault.com/questions/784312