Root.sh Failed on Second Node: Configuration of ASM Failed: Disk Group ... Already Exists [ID 138472
2012-08-10 13:39
696 查看
In this Document
Symptoms
Cause
Solution
Oracle Server - Enterprise Edition - Version: 11.2.0.2 and later [Release: 11.2 and later ]
Information in this document applies to any platform.
Root.sh failed on second node
ASMCA log shows:
For the first node, root.sh did not failed and Oracle Clusterware is up/running as well as +ASM1 which has the ASM diskgroup mounted.
Checked the ASM disks from second node, and these are visible/accessible, so the problem is not caused by the storage.
A 'ps -ef|grep -i asm' from the second node shows several background processes running for ASM but these are for ASM1? This should be +ASM2 instead. A 'cat /etc/oratab' from the second node, confirms that +ASM1 (not +ASM2) is configured on
the second node.
Having +ASM1 on more than one node occurs when root.sh is executed on the second node before root.sh completes on the first node. Since the Oracle Clusterware has not completed registering +ASM1 on the first node (where root.sh is executed first but yet still
running), root.sh on the second node is not aware that +ASM1 configuration is in progress so it proceed to configure it on the second node which eventually fails.
To confirm this, check the timestamps for rootcrs_node<x>.log from all nodes.
To correct this please do the following manual steps since ASMCA does not have yet an option to remove an ASM instance:
To recreate ASM on the node where it initially failed:
1) Deconfigure Grid Infrastructure (GI, gi_home) on the second node:
As root
# cd <gi_home>/crs/install
# ./rootcrs.pl -verbose -deconfig -force
Remove the ASM entry from the second node's /etc/oratab. Example: +ASM1:/u01/app/11.2.0/grid:N # line added by Agent
Backup and remove any ASM named files from the second node's <gi_home>/dbs directory: Example: ab_+ASM1.dat hc_+ASM2.dat
2) Reconfigure GI on the node where it initially failed (second node):
As root
# cd <gi_home>
# ./root.sh
3) Checks:
root.sh on node2 must succeed: Example (last line): Configure Oracle Grid Infrastructure for a Cluster ... succeeded
As the gi_home owner/user:
$ cd <gi_home>/bin
$ ./crsctl stat res -t <-- must show Oracle Clusterware, ASM, etc running on both nodes (+ASM1 on first node, +ASM2 on second node)
Symptoms
Cause
Solution
This document is being delivered to you via Oracle Support's Rapid Visibility (RaV) process and therefore has not been subject to an independent technical review. |
Applies to:
Oracle Server - Enterprise Edition - Version: 11.2.0.2 and later [Release: 11.2 and later ]Information in this document applies to any platform.
Symptoms
Root.sh failed on second noderootcrs_node2.log: ... 2011-10-14 12:19:00: Configuring ASM via ASMCA 2011-10-14 12:19:00: Executing as oracle: /oracle/app/product/11.2.0/grid/bin/asmca -silent -diskGroupName DATA -diskList '/asmdisks/10GB/c0t60000970000192604380533031384543d0s4,/asmdisks/10GB/c0t60000970000192604380533031384544d0s4,/asmdisks/10GB/c0t60000970000192604380533031384545d0s4' -redundancy NORMAL -diskString '/asmdisks/10GB/*' -configureLocalASM 2011-10-14 12:19:00: Running as user oracle: /oracle/app/product/11.2.0/grid/bin/asmca -silent -diskGroupName DATA -diskList '/asmdisks/10GB/c0t60000970000192604380533031384543d0s4,/asmdisks/10GB/c0t60000970000192604380533031384544d0s4,/asmdisks/10GB/c0t60000970000192604380533031384545d0s4' -redundancy NORMAL -diskString '/asmdisks/10GB/*' -configureLocalASM 2011-10-14 12:19:00: Invoking "/oracle/app/product/11.2.0/grid/bin/asmca -silent -diskGroupName DATA -diskList '/asmdisks/10GB/c0t60000970000192604380533031384543d0s4,/asmdisks/10GB/c0t60000970000192604380533031384544d0s4,/asmdisks/10GB/c0t60000970000192604380533031384545d0s4' -redundancy NORMAL -diskString '/asmdisks/10GB/*' -configureLocalASM" as user "oracle" 2011-10-14 12:19:00: Executing /bin/su oracle -c "/oracle/app/product/11.2.0/grid/bin/asmca -silent -diskGroupName DATA -diskList '/asmdisks/10GB/c0t60000970000192604380533031384543d0s4,/asmdisks/10GB/c0t60000970000192604380533031384544d0s4,/asmdisks/10GB/c0t60000970000192604380533031384545d0s4' -redundancy NORMAL -diskString '/asmdisks/10GB/*' -configureLocalASM" 2011-10-14 12:19:00: Executing cmd: /bin/su oracle -c "/oracle/app/product/11.2.0/grid/bin/asmca -silent -diskGroupName DATA -diskList '/asmdisks/10GB/c0t60000970000192604380533031384543d0s4,/asmdisks/10GB/c0t60000970000192604380533031384544d0s4,/asmdisks/10GB/c0t60000970000192604380533031384545d0s4' -redundancy NORMAL -diskString '/asmdisks/10GB/*' -configureLocalASM" 2011-10-14 12:19:29: Command output: > > Disk Group DATA creation failed with the following message: > Disk Group DATA already exists. Cannot be created again > >End Command output 2011-10-14 12:19:29: Configuration of ASM ... failed
ASMCA log shows:
[main] [ 2011-10-14 12:19:27.813 EDT ] [UsmcaLogger.logInfo:142] Instance running true [main] [ 2011-10-14 12:19:27.823 EDT ] [UsmcaLogger.logInfo:142] Diskgroup does not exist, creating.. ... [main] [ 2011-10-14 12:19:28.380 EDT ] [UsmcaLogger.logInfo:142] Disk Group DATA creation failed with the following message: [main] [ 2011-10-14 12:19:28.380 EDT ] [UsmcaLogger.logInfo:142] Disk Group DATA already exists. Cannot be created again ... [main] [ 2011-10-14 12:19:28.381 EDT ] [UsmcaLogger.logInfo:142] Diskgroup creation is not successful.
For the first node, root.sh did not failed and Oracle Clusterware is up/running as well as +ASM1 which has the ASM diskgroup mounted.
Cause
Checked the ASM disks from second node, and these are visible/accessible, so the problem is not caused by the storage.A 'ps -ef|grep -i asm' from the second node shows several background processes running for ASM but these are for ASM1? This should be +ASM2 instead. A 'cat /etc/oratab' from the second node, confirms that +ASM1 (not +ASM2) is configured on
the second node.
Having +ASM1 on more than one node occurs when root.sh is executed on the second node before root.sh completes on the first node. Since the Oracle Clusterware has not completed registering +ASM1 on the first node (where root.sh is executed first but yet still
running), root.sh on the second node is not aware that +ASM1 configuration is in progress so it proceed to configure it on the second node which eventually fails.
To confirm this, check the timestamps for rootcrs_node<x>.log from all nodes.
Solution
To correct this please do the following manual steps since ASMCA does not have yet an option to remove an ASM instance:To recreate ASM on the node where it initially failed:
1) Deconfigure Grid Infrastructure (GI, gi_home) on the second node:
As root
# cd <gi_home>/crs/install
# ./rootcrs.pl -verbose -deconfig -force
Remove the ASM entry from the second node's /etc/oratab. Example: +ASM1:/u01/app/11.2.0/grid:N # line added by Agent
Backup and remove any ASM named files from the second node's <gi_home>/dbs directory: Example: ab_+ASM1.dat hc_+ASM2.dat
2) Reconfigure GI on the node where it initially failed (second node):
As root
# cd <gi_home>
# ./root.sh
3) Checks:
root.sh on node2 must succeed: Example (last line): Configure Oracle Grid Infrastructure for a Cluster ... succeeded
As the gi_home owner/user:
$ cd <gi_home>/bin
$ ./crsctl stat res -t <-- must show Oracle Clusterware, ASM, etc running on both nodes (+ASM1 on first node, +ASM2 on second node)
相关文章推荐
- How to restore ASM based OCR after complete loss of the CRS diskgroup on Linux systems[ID 1062983.1]
- RAC root.sh Configuration of ASM failed Command return code of 1 (256) from command 解决方法
- RAC root.sh Configuration of ASM failed Command return code of 1 (256) from command 解决方法
- How to restore ASM based OCR after complete loss of the CRS diskgroup on Linux/Unix systems [ID 1062
- Install 11.2.0.3(RAC) Grid Infrastructure on linux x86-64, run root.sh on second node failed 失败一列
- Executing root.sh errors with "Failed To Upgrade Oracle Cluster Registry Configuration" [ID 466673.1]
- Executing root.sh errors with "Failed To Upgrade Oracle Cluster Registry Configuration" [ID 466673.1]
- HowTo Restore RMAN Disk backups of RAC Database to Single Instance On Another Node (文档 ID 415579.1)
- Executing root.sh errors with "Failed To Upgrade Oracle Cluster Registry Configuration" [ID 466673.1]
- Executing root.sh errors with "Failed To Upgrade Oracle Cluster Registry Configuration" [ID 466673.1]
- Executing root.sh errors with "Failed To Upgrade Oracle Cluster Registry Configuration" [ID 466673.1]
- Executing root.sh errors with "Failed To Upgrade Oracle Cluster Registry Configuration" [ID 466673.1]
- If the root.sh script fails on a node during the install of the Grid Infrastructure with Oracle Data
- Executing root.sh errors with "Failed To Upgrade Oracle Cluster Registry Configuration" [ID 466673.1
- HowTo Restore RMAN Disk backups of RAC Database to Single Instance On Another Node [ID 415579.1]
- ASM Diskgroup Creation Failed with "IBM AIX RISC system/6000: 6:no such device or address" [ID 12634
- Running Root.Sh On Second Node Fails PRKN-1008 Unable to load the shared library srvmhas10 解决
- HowTo Restore RMAN Disk backups of RAC Database to Single Instance On Another Node [ID 415579.1]
- Running Root.Sh On Second Node Fails PRKN-1008 Unable to load the shared library srvmhas10
- Executing root.sh errors with "Failed To Upgrade Oracle Cluster R 4000 egistry Configuration" [ID 466673.