11g RAC | Using Duplicate target database 11g Active Database option
I have a 2 Node RAC Standby database (STDBY) . I need to replicate it as a Load TEST database (LDTEST) in a read/write mode.
I have a 2 Node RAC Standby database (STDBY) . I need to replicate it as a Load TEST database (LDTEST) in a read/write mode.
Here I am duplicating 11g RAC database from one RHEL Server to Another by old 10g method.
I could have used 11g “DUPLICATE TARGET DATABASE TO TARGET_DB FROM ACTIVE DATABASE” which doesn’t need previous rman backup at source. But it may not be a good option for large databases or at places with narrow network bandwidth.
If you have already uninstalled Oracle Clusterware using the Oracle Universal Installer - Please download attached file cleanup.zip that contains a copy of the logpartformat.exe and the guioracleobjmanage.exe (plus necessary dll).
Note: You must be logged in as the root user, because root owns the OCR files. Make sure there is a recent copy of the OCR file before making any changes: ocrconfig showbackup
SQL> desc v$mystat
Name Null? Type
—————————————– ——– —————————-
SID NUMBER
STATISTIC# NUMBER
VALUE NUMBER
1. Grep occsd Log file
[oracle @ tadrac1]: /u1/app/../cssd >grep -i “master node” ocssd.log | tail -1
[CSSD]CLSS-3001: local node number 1, master node number 1
Step A> Sequence of events to pull cluster database down..
1. Bring down load balanced/TAF service
srvctl stop service -d orcl -s RAC
CRS DAEMON FUNCTIONALITY
CRSD: Performs high availability recovery and management operations such as maintaining the OCR and managing application resources.
- Engine for HA operation
- Manages ‘application resources’
- Starts, stops, and fails ‘application resources’ over
- Spawns separate ‘actions’ to start/stop/check application resources
- Maintains configuration profiles in the OCR (Oracle Configuration Repository)
- Stores current known state in the OCR.
- Runs as root
- Is restarted automatically on failure
column tsname format a30 heading ‘Tablespace Name’
column tbs_size_mb format 99999,999 heading ‘Size|(MB)’
column used format 99999,999 heading ‘Used|(MB)’
column avail format 99999,999 heading ‘Free|(MB)’
column used_visual format a11 heading ‘Used’
column pct_used format 999 heading ‘% Used’
Login to one of the nodes (i.e. linux1) and disable the cluster instance parameter by setting cluster_database to FALSE from the current instance:
$ sqlplus “/ as sysdba”
SQL> alter system set cluster_database=false scope=spfile sid=’orcl1′;
Sequence of events to bring cluster database back..
1. Start all services using “start -nodeapps”
2. Start ASM instnace using “srvctl start asm -n (node)”
3. Start RAC instances using “srvctl start instance -d (database) -I (instance)”
4. Finish up by bringing our load balanced/TAF service online “srvctl start service -d orcl -s RAC”
Check out current configuration information