su - grid
. oraenv
+ASM1 or +ASM2
Check the VD Location
crsctl query css votedisk
Check the OCR Location
cat /etc/oracle/ocr.loc
ocrcheck -config
asmcmd spget
asmcmd lsdg
asmcmd lsdsk -kp -G CRSDG78
or
asmcmd -p
lsdsk -kp -G CRSDG78
Login as a Root#
#. oraenv
+ASM1 or +ASM2
To view the backup files
ocrconfig -showbackup
To take OCR/VD Backup(Backup contains both OCR and Voting Disk)
ocrconfig -manualbackup
To view the backup files
ocrconfig -showbackup manual
or
ocrconfig -showbackup (this will display both auto and manual backup)
Check the Below Logfiles when cluster issue occures
---------------------------------------------------
1. Cluster Alertlog File Path/Filename: (locate alertnodename.log)
/u01/app/11.2.0/grid/log/racnode7/alertracnode7.log
2. CSSD Logfile Path/Filename: (locate ocssd.log)
/u01/app/11.2.0/grid/log/racnode7/cssd/ocssd.log
3. CRSD Logfile Path/Filename: (locate crsd.log)
/u01/app/11.2.0/grid/log/racnode7/crsd/crsd.log
4. ASM Instance Alert Logfile Path/Filename: (locate alert_+ASM)
/u01/app/grid/diag/asm/+asm/+ASM1/trace/alert_+ASM1.log
5. OS Logfile.
ls -lrt /var/log/messages*
Do Not Use in Production(Real Time)
dd if=/dev/zero of=/dev/sdc2 count=100 bs=1024
/etc/init.d/oracleasm scandisks
crsctl stat res -t
crsctl stop crs -f (
*********************************************************************************************
1. Locate the latest automatic OCR backup
#. oraenv
+ASM1/+ASM2
#ocrconfig -showbackup
2. Stop CRS on All Nodes (As a root User)
#. oraenv
+ASM1/+ASM2
#crsctl stop crs -f
3. Start the CRS stack in exclusive mode (Only on ONE Node) (As a root User)
On the node that has the most recent OCR backup, log on as root and start CRS in exclusive mode, this mode will allow ASM to start & stay up without the presence of a Voting disk and without the CRS daemon process (crsd.bin) running.
# crsctl start crs -excl -nocrs (Only on ONE Node)
4. Label the CRS disk for ASMLIB use (Only on ONE Node)
/etc/init.d/oracleasm createdisk DISKNEW2 /dev/sdc2 (/dev/sdc2 would be different in your case)
5. Create the CRS diskgroup via sqlplus (Only on ONE Node)(As a grid User)
su - grid
. oraenv
+ASM1 or +ASM2
$ sqlplus / as sysasm
SQL> show parameter asm_diskstring
SQL> alter system set asm_diskstring='/dev/oracleasm/disks/*';
SQL> create diskgroup CRS external redundancy disk '/dev/oracleasm/disks/KIRAN_1' attribute 'COMPATIBLE.ASM' = '11.2';
6. Restore the latest OCR backup (As a root User)
#ocrconfig -restore <backuppath/filename>
7. Recreate the Voting file (As a grid/root User)
crsctl replace votedisk +CRSDG78
8. Recreate the SPFILE for ASM (optional)(Only on ONE Node) (As a grid User)
su - grid
. oraenv
+ASM1 or +ASM2
vi /home/grid/asm_pfile.ora
*.asm_power_limit=1
*.diagnostic_dest='/u01/app/grid'
*.instance_type='asm'
*.large_pool_size=12M
*.remote_login_passwordfile='EXCLUSIVE'
sqlplus / as sysasm
SQL> create spfile='+CRSDG78' from pfile='/home/grid/asm_pfile.ora';
9. Shutdown CRS (Only on ONE Node) (As a root User)
crsctl stop crs -f
10. Rescan ASM disks (If using ASMLIB rescan all ASM disks on each node As the root user)
/etc/init.d/oracleasm scandisks
11. Start CRS (On All Nodes) (As a root User)
crsctl start crs
12. Verify CRS (Only on ONE Node) (As a root/grid User)
crsctl stat res -t
. oraenv
+ASM1 or +ASM2
Check the VD Location
crsctl query css votedisk
Check the OCR Location
cat /etc/oracle/ocr.loc
ocrcheck -config
asmcmd spget
asmcmd lsdg
asmcmd lsdsk -kp -G CRSDG78
or
asmcmd -p
lsdsk -kp -G CRSDG78
Login as a Root#
#. oraenv
+ASM1 or +ASM2
To view the backup files
ocrconfig -showbackup
To take OCR/VD Backup(Backup contains both OCR and Voting Disk)
ocrconfig -manualbackup
To view the backup files
ocrconfig -showbackup manual
or
ocrconfig -showbackup (this will display both auto and manual backup)
Check the Below Logfiles when cluster issue occures
---------------------------------------------------
1. Cluster Alertlog File Path/Filename: (locate alertnodename.log)
/u01/app/11.2.0/grid/log/racnode7/alertracnode7.log
2. CSSD Logfile Path/Filename: (locate ocssd.log)
/u01/app/11.2.0/grid/log/racnode7/cssd/ocssd.log
3. CRSD Logfile Path/Filename: (locate crsd.log)
/u01/app/11.2.0/grid/log/racnode7/crsd/crsd.log
4. ASM Instance Alert Logfile Path/Filename: (locate alert_+ASM)
/u01/app/grid/diag/asm/+asm/+ASM1/trace/alert_+ASM1.log
5. OS Logfile.
ls -lrt /var/log/messages*
Do Not Use in Production(Real Time)
dd if=/dev/zero of=/dev/sdc2 count=100 bs=1024
/etc/init.d/oracleasm scandisks
crsctl stat res -t
crsctl stop crs -f (
*********************************************************************************************
1. Locate the latest automatic OCR backup
#. oraenv
+ASM1/+ASM2
#ocrconfig -showbackup
2. Stop CRS on All Nodes (As a root User)
#. oraenv
+ASM1/+ASM2
#crsctl stop crs -f
3. Start the CRS stack in exclusive mode (Only on ONE Node) (As a root User)
On the node that has the most recent OCR backup, log on as root and start CRS in exclusive mode, this mode will allow ASM to start & stay up without the presence of a Voting disk and without the CRS daemon process (crsd.bin) running.
# crsctl start crs -excl -nocrs (Only on ONE Node)
4. Label the CRS disk for ASMLIB use (Only on ONE Node)
/etc/init.d/oracleasm createdisk DISKNEW2 /dev/sdc2 (/dev/sdc2 would be different in your case)
5. Create the CRS diskgroup via sqlplus (Only on ONE Node)(As a grid User)
su - grid
. oraenv
+ASM1 or +ASM2
$ sqlplus / as sysasm
SQL> show parameter asm_diskstring
SQL> alter system set asm_diskstring='/dev/oracleasm/disks/*';
SQL> create diskgroup CRS external redundancy disk '/dev/oracleasm/disks/KIRAN_1' attribute 'COMPATIBLE.ASM' = '11.2';
6. Restore the latest OCR backup (As a root User)
#ocrconfig -restore <backuppath/filename>
7. Recreate the Voting file (As a grid/root User)
crsctl replace votedisk +CRSDG78
8. Recreate the SPFILE for ASM (optional)(Only on ONE Node) (As a grid User)
su - grid
. oraenv
+ASM1 or +ASM2
vi /home/grid/asm_pfile.ora
*.asm_power_limit=1
*.diagnostic_dest='/u01/app/grid'
*.instance_type='asm'
*.large_pool_size=12M
*.remote_login_passwordfile='EXCLUSIVE'
sqlplus / as sysasm
SQL> create spfile='+CRSDG78' from pfile='/home/grid/asm_pfile.ora';
9. Shutdown CRS (Only on ONE Node) (As a root User)
crsctl stop crs -f
10. Rescan ASM disks (If using ASMLIB rescan all ASM disks on each node As the root user)
/etc/init.d/oracleasm scandisks
11. Start CRS (On All Nodes) (As a root User)
crsctl start crs
12. Verify CRS (Only on ONE Node) (As a root/grid User)
crsctl stat res -t