Documente Academic
Documente Profesional
Documente Cultură
Search
Search LOGIN | REGISTER
Home Platforms Oracle Oracle Blog Troubleshooting Oracle Clusterware common startup
failures
Troubleshooting Oracle
Clusterware common startup
failures
Oracle Community Join
Blog
Options
Print
Comment
RSS Feed
Share Tweet
Like 0
0
Related
Posts
Oracle Clusterware
12c: What-if
command evaluation
- Part-I
latest revision 4
months ago
by Anju Garg
Oracle Clusterware
12c: What-if
command evaluation
- Part-II
1 of 8 18-02-2017 18:32
Troubleshooting Oracle Clusterware common startup failures - Oracle ... https://www.toadworld.com/platforms/oracle/b/weblog/archive/2014/0...
latest revision 3
months ago
by Anju Garg
Oracle Clusterware
latest revision 4
months ago
by Anju Garg
View More
Node eviction
Cluster becoming unhealthy
Unable to start cluster and some of the Clusterware components
The diagram below depicts Oracle Cluster stack (components) startup sequence at
various levels:
2 of 8 18-02-2017 18:32
Troubleshooting Oracle Clusterware common startup failures - Oracle ... https://www.toadworld.com/platforms/oracle/b/weblog/archive/2014/0...
The entire Oracle Cluster stack and the services registered on the cluster
automatically comes up when a node reboots or if the cluster stack manually
starts. The startup process is segregated in ve (05) levels, at each level, dierent
processes are got started in a sequence.
On node reboot/crash, the init process on the OS spawns init.ohasd (as mentioned
in the /etc/inittab le), which in turn commence Oracle High Availability Service
Daemon (ohasd). The ohasd daemon is then responsible of starting o the other
critical cluster daemon processes.
Imagine your crsctl check cluster/crs command and its gives the following errors:
3 of 8 18-02-2017 18:32
Troubleshooting Oracle Clusterware common startup failures - Oracle ... https://www.toadworld.com/platforms/oracle/b/weblog/archive/2014/0...
OR
We will start taking the issues of the components startup failures in the same
sequence they usually start. Lets talk about ohasd startup failures which would
result in CRS-4639/4124/4000 errors. The following are the main causes:
Verify if the /etc/inittab le contains the entry to start the ohasd process
automatically.
Ensure Cluster auto startup is congured using the crsctl cong crs
command. For any reasons it is not auto start congured, enable the auto
start and start the cluster manually.
Refer alert.log and ohasd.log les under $ORA_GRID_HOME/log/hosname
location.
Ensure the node has no issues accessing the OLR and OCR/Voting Disks.
This can be veried in the alert or ohasd log les.
--------------------------------------------------------------------------------
NAME TARGET STATE SERVER
STATE_DETAILS
--------------------------------------------------------------------------------
Cluster Resources
--------------------------------------------------------------------------------
ora.asm
1 ONLINE ONLINE rac2
Started
ora.cluster_interconnect.haip
1 ONLINE ONLINE
rac2
ora.crsd
1 ONLINE OFFLINE
rac2
ora.cssd
1 ONLINE ONLINE
rac2
ora.cssdmonitor
1 ONLINE ONLINE
rac2
ora.ctssd
1 ONLINE ONLINE rac2
OBSERVER
ora.diskmon
1 ONLINE ONLINE
rac2
ora.evmd
4 of 8 18-02-2017 18:32
Troubleshooting Oracle Clusterware common startup failures - Oracle ... https://www.toadworld.com/platforms/oracle/b/weblog/archive/2014/0...
If you couldnt resolve the problem reviewing the above guide lines, you might
need to contact Oracle support and provide all required information to investigate
further on the problem.
The following are some of the most referred logs as part of cluster maintenance
and troubleshooting various Cluster related problems, such as, Node eviction,
Cluster stack heath, OCR/Voting disk related problems etc:
alert_hostname.log
crsd.log
ocssd.log
ohasd.log
alert_hostname.log Each cluster node has its own alert log le and write
important and useful information about cluster startup, node eviction, any cluster
5 of 8 18-02-2017 18:32
Troubleshooting Oracle Clusterware common startup failures - Oracle ... https://www.toadworld.com/platforms/oracle/b/weblog/archive/2014/0...
ocssd.log Yet another very critical and important log le which needs your
attention. Whenever cluster encounters any serious snags with regards to Cluster
Synchronization Service Daemon (CSSD) process, this is the le that needs to refer
to understand the nature of the problem and to resolve the problem too. This is
one of the busiest log les that is writing continuously and maintained by the
cluster automatically. Manual maintenance on the le is not recommended, and
once the le size reaches to 50MB, Oracle automatically archives the les and
create a new log le. In the context, total 10 archive log les are maintenance at
any given point in time.
crsd.log Cluster Ready Service Daemon (CRSD) process writes all important events
to the le, such as, cluster resources startup, stop, failure and CRSD health status.
If you have issues starting/stopping any cluster and non-cluster resources on the
node, refer this log le to diagnose the issue. This les also maintained by the
Oracle and remove the le is not recommended. Once the size of the le reaches
to 10MB, it will be archived automatically and a new log le will be created. There
will be 10 archived copies of the le will be kept under any point of time.
ohasd.log The log le is accessed and managed by the new Oracle High
Availability Service Daemon (ohasd) process which was rst introduced in Oracle
11gR2. If you encounter any issues running root.sh or rootupgrade.sh scripts, refer
to this log le to understand troubleshooting the problem. If you face issues
starting up the process, and if Oracle Local Registry (OLR) has any corruption or
inaccessibility issues, also refer to this le. Like crsd and ocssd log les, this le
also is maintained automatically by Oracle and archives the log le upon reaching
10MB size. Total of 10 archived log les are maintained at any given point in time.
Fortunately, Oracle let you adjust the default trace/debug levels of any cluster
component or resource dynamically. To list the default trace/debug settings of a
component or sub-component, login as root user and execute the following
command from the GRID_HOME:
You will have to seek Oracle support advice before you set or adjust any default
settings. To disable the trace/debug level, set the level to value 0. You can set the
levels from 1 to 5. The more the value, the more information will be generated and
you must closely watch the log le growth and space on the lesystem to avoid any
space related issues.
You can also turn on trace through OS settings; use the following OS specic
command:
$ export SRVM_TRACE=true
Once the above is set on any UNIX OS system, trace/log les are generated under
$GRID_HOME/cv/log destination. Once you exit from the terminal, tracing will end.
6 of 8 18-02-2017 18:32
Troubleshooting Oracle Clusterware common startup failures - Oracle ... https://www.toadworld.com/platforms/oracle/b/weblog/archive/2014/0...
Let me zip through and explain you the benets of some of the very
important and mostly used tools and utilities here.
Cluster Verication U tility (CVU) is used to collect pre and post cluster
installation conguration details at various levels and various components.
With 11gR2, it also provides the ability to verify the cluster health. Look at
some of the useful commands below:
Example:
7 of 8 18-02-2017 18:32
Troubleshooting Oracle Clusterware common startup failures - Oracle ... https://www.toadworld.com/platforms/oracle/b/weblog/archive/2014/0...
Above all, there is many other important and useful tools: Cluster Health
Monitoring (CHM) to diagnose node eviction issues, DB Hanganalysis, OSWatcher,
The Light on Monitor, ProWatcher etc are available for your use under dierent
circumstances.
In a nutshell, this paper make you understand the Cluster startup sequence, how
things get started on node reboot and provides guide lines to solve most
commonly faced Cluster startup issues by analyzing the cluster logs, using
appropriate tools/utilities discussed.
59113 0 /
About Toad World Quest Communities Privacy Policy Terms of Use Contact Us
Send Feedback About Quest
Toad World is sponsored Copyright 2017 Quest Software Inc. ALL RIGHTS
by RESERVED.
8 of 8 18-02-2017 18:32