Cluster

每當 Solaris 集群重新啟動或重新啟動電源時,zpool 就會消失

  • May 27, 2015

我有以下設置

Oracle Solaris 10 -> 5.10 Generic_147147-26 sun4v sparc

Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64 位生產

適用於 Solaris 10 sparc 的 Oracle Solaris Cluster 3.3u2

適用於 Solaris 10 sparc 的 Oracle Solaris Cluster 地理版 3.3u2

我安裝了帶有 ZFS 的 Oracle Solaris 10 我有一個用於 /oradata 的池 每當我重新啟動/重新啟動集群時,我的集群 ZFS 池就會消失,因為該集群無法啟動 oracle 數據庫資源/組 每次重新啟動/重新啟動集群後,我都必須這樣做手動

zpool import db
clrg online ora-rg 
...

可能是什麼原因?

我唯一知道的 db zpool,池是使用我創建的 ora-has 資源導入的,如下所示(使用 Zpools 選項)

# /usr/cluster/bin/clresourcegroup create ora-rg
# /usr/cluster/bin/clresourcetype register SUNW.HAStoragePlus 
# /usr/cluster/bin/clresource create -g ora-rg -t SUNW.HAStoragePlus -p Zpools=db ora-has

# zpool status db
 pool: db
 state: ONLINE
scan: none requested
config:

       NAME        STATE     READ WRITE CKSUM
       db          ONLINE       0     0     0
         mirror-0  ONLINE       0     0     0
           c0t2d0  ONLINE       0     0     0
           c0t3d0  ONLINE       0     0     0

errors: No known data errors


Booting in cluster mode

impdneilab1 console login: Apr 21 17:12:24 impdneilab1 cl_runtime:     NOTICE: CMM: Node impdneilab1 (nodeid = 1) with votecount = 1 added.
Apr 21 17:12:24 impdneilab1 sendmail[642]: My unqualified host name (impdneilab1) unknown; sleeping for retry
Apr 21 17:12:24 impdneilab1 cl_runtime: NOTICE: CMM: Node impdneilab1: attempting to join cluster.
Apr 21 17:12:24 impdneilab1 cl_runtime: NOTICE: CMM: Cluster has reached quorum.
Apr 21 17:12:24 impdneilab1 cl_runtime: NOTICE: CMM: Node impdneilab1 (nodeid = 1) is up; new incarnation number = 1429629142.
Apr 21 17:12:24 impdneilab1 cl_runtime: NOTICE: CMM: Cluster members: impdneilab1.
Apr 21 17:12:24 impdneilab1 cl_runtime: NOTICE: CMM: node reconfiguration #1 completed.
Apr 21 17:12:24 impdneilab1 cl_runtime: NOTICE: CMM: Node impdneilab1: joined cluster.
Apr 21 17:12:24 impdneilab1 in.mpathd[262]: Successfully failed over from NIC nxge1 to NIC e1000g1
Apr 21 17:12:24 impdneilab1 in.mpathd[262]: Successfully failed over from NIC nxge0 to NIC e1000g0
obtaining access to all attached disks

親愛的,我找到了答案

https://community.oracle.com/thread/3714952?sr=inbox

地理集群配置中的單節點集群會出現這種行為:


如果整個集群出現故障然後又重新啟動,則預期行為是地理版本在啟動時會停止本地集群上的保護組。這樣做的原因是可能已發出接管或儲存/數據可能不完整或不可用(如果主站點完全失敗,儘管集群節點已恢復,但這並不意味著儲存/數據完好無損並準備好承擔該站點在故障之前的角色)。這與我們在添加到保護組的應用程序 rgs 上要求 auto_start_on_new_cluster=false 的原因相同。集群重新啟動後,使用者需要根據需要進行干預並啟動或執行故障恢復過程。

引用自:https://serverfault.com/questions/686738