Cluster
每當 Solaris 集群重新啟動或重新啟動電源時,zpool 就會消失
我有以下設置
Oracle Solaris 10 -> 5.10 Generic_147147-26 sun4v sparc
Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64 位生產
適用於 Solaris 10 sparc 的 Oracle Solaris Cluster 3.3u2
適用於 Solaris 10 sparc 的 Oracle Solaris Cluster 地理版 3.3u2
我安裝了帶有 ZFS 的 Oracle Solaris 10 我有一個用於 /oradata 的池 每當我重新啟動/重新啟動集群時,我的集群 ZFS 池就會消失,因為該集群無法啟動 oracle 數據庫資源/組 每次重新啟動/重新啟動集群後,我都必須這樣做手動
zpool import db clrg online ora-rg ...
可能是什麼原因?
我唯一知道的 db zpool,池是使用我創建的 ora-has 資源導入的,如下所示(使用 Zpools 選項)
# /usr/cluster/bin/clresourcegroup create ora-rg # /usr/cluster/bin/clresourcetype register SUNW.HAStoragePlus # /usr/cluster/bin/clresource create -g ora-rg -t SUNW.HAStoragePlus -p Zpools=db ora-has # zpool status db pool: db state: ONLINE scan: none requested config: NAME STATE READ WRITE CKSUM db ONLINE 0 0 0 mirror-0 ONLINE 0 0 0 c0t2d0 ONLINE 0 0 0 c0t3d0 ONLINE 0 0 0 errors: No known data errors Booting in cluster mode impdneilab1 console login: Apr 21 17:12:24 impdneilab1 cl_runtime: NOTICE: CMM: Node impdneilab1 (nodeid = 1) with votecount = 1 added. Apr 21 17:12:24 impdneilab1 sendmail[642]: My unqualified host name (impdneilab1) unknown; sleeping for retry Apr 21 17:12:24 impdneilab1 cl_runtime: NOTICE: CMM: Node impdneilab1: attempting to join cluster. Apr 21 17:12:24 impdneilab1 cl_runtime: NOTICE: CMM: Cluster has reached quorum. Apr 21 17:12:24 impdneilab1 cl_runtime: NOTICE: CMM: Node impdneilab1 (nodeid = 1) is up; new incarnation number = 1429629142. Apr 21 17:12:24 impdneilab1 cl_runtime: NOTICE: CMM: Cluster members: impdneilab1. Apr 21 17:12:24 impdneilab1 cl_runtime: NOTICE: CMM: node reconfiguration #1 completed. Apr 21 17:12:24 impdneilab1 cl_runtime: NOTICE: CMM: Node impdneilab1: joined cluster. Apr 21 17:12:24 impdneilab1 in.mpathd[262]: Successfully failed over from NIC nxge1 to NIC e1000g1 Apr 21 17:12:24 impdneilab1 in.mpathd[262]: Successfully failed over from NIC nxge0 to NIC e1000g0 obtaining access to all attached disks
親愛的,我找到了答案
https://community.oracle.com/thread/3714952?sr=inbox
地理集群配置中的單節點集群會出現這種行為:
如果整個集群出現故障然後又重新啟動,則預期行為是地理版本在啟動時會停止本地集群上的保護組。這樣做的原因是可能已發出接管或儲存/數據可能不完整或不可用(如果主站點完全失敗,儘管集群節點已恢復,但這並不意味著儲存/數據完好無損並準備好承擔該站點在故障之前的角色)。這與我們在添加到保護組的應用程序 rgs 上要求 auto_start_on_new_cluster=false 的原因相同。集群重新啟動後,使用者需要根據需要進行干預並啟動或執行故障恢復過程。