Asynchronous CDC on a Logical Standby database (10gR2)
Hi,
Question:
Is it possible to setup Asynchronous CDC on a Logical Standby database (10gR2)?
Background:
In our Oracle Data Warehouse environment where we need to provide up-to-the-second "operational" reporting against a Logical Standby database (10gR2 Data Guard) as the production application database owners do not permit direct reporting against the Primary database (10gR2); we also have a Data Warehouse database (10gR2) that requires more sophisticated/summary reporting which is updated at less regular intervals from the Logical Standby database.
The complication is that although the Logical Standby database will serve the purpose of a 'syncronized' version of the Primary database by giving us a current-state view of the data, we require more detailed change data capture on certain tables to audit change history between the points we extract data for the Data Warehouse.
The DBAs do not want to set-up Asynchronous CDC on the Primary database, so is it possible to setup Asynchronous CDC on a Logical Standby database?
Anyone done it before?
Any issues?
Any alternatives?
Any help would be much appreciated.
Thanks
As far as I know it is but it would be easier, and more robust, to just create a second logical standby and feed them both from the primary.
Logical Data Guard and Streams are essentially just two implementations of the same technology written by the same team of developers.
Similar Messages
-
Asynchronous CDC on a Logical Standby Database
Hi,
Question:
Is it possible to setup Asynchronous CDC on a Logical Standby database (10gR2)?
Background:
In our Oracle Data Warehouse environment where we need to provide up-to-the-second "operational" reporting against a Logical Standby database (10gR2 Data Guard) as the production application database owners do not permit direct reporting against the Primary database (10gR2); we also have a Data Warehouse database (10gR2) that requires more sophisticated/summary reporting which is updated at less regular intervals from the Logical Standby database.
The complication is that although the Logical Standby database will serve the purpose of a 'syncronized' version of the Primary database by giving us a current-state view of the data, we require more detailed change data capture on certain tables to audit change history between the points we extract data for the Data Warehouse.
The DBAs do not want to set-up Asynchronous CDC on the Primary database, so is it possible to setup Asynchronous CDC on a Logical Standby database?
Anyone done it before?
Any issues?
Any alternatives?
Any help would be much appreciated.
Thanks
TimAutolog CDC could be one possible solution. It uses a feature of the downstream capture.
See details from the following link
http://search.oracle.com/search/search?default=true&keyword=change+data+capture&start=1&nodeid=&fid=&showSimilarDoc=true&group=Oracle+Technology+Network -
Is CDC feature working on the logical standby database
I would like to implement CDC on the standby logical database instead of primary production database. Is CDC working on the logical standby database??
Thanks.Thanks. My database version is: 10.2.0.4.0 and no plan to upgrade to 11g.
Either asyn or syn CDC on logical standby database is ok for me. I searched but couldn't find any docs for how to set CDC for logical standby, they are all for primary database.
Please help. -
Log mining is taking too much time in logical standby database
dear DBAs,
today i found a gap between the production database and the logical standby database and i found that the log mining is taking more than 1 hour to complete an archivelog (size: 500M)
note that the MAX_SGA is 1500M and the MAX_SERVERS=45
the databases is 10gR2 (10.2.0.5.0) running on a linux machine RHEL 4
please your help.
thx in advance
Eliehi,
can you check metalink id [ID 241512.1]
thanks -
Problem building logical standby database
Hi all,
i am trying to build a logical standby database on platform Sun OS 10/Oracle 10gR2. I am following the Oracle document http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/create_ls.htm#BEIGHEIA
I have created a physical standby database and converting it to a logical standby database. I ensured that my physical Standby is in Sync with primary.
Procedure DBMS_LOGSTDBY.BUILD executes successfully on primary.
The problem is that the command *'alter database recover to logical standby test;'* DOESN'T END. No error in archive log. I have identified the archived redo log that contains the LogMiner dictionary and the starting SCN and applied that archive log on standby. Still the the above command doesn't end.
Any Help is appreciated.SQL> alter database recover to logical standby m2test;
This command doesn't return an sql> prompt. Alert log says it is waiting for log sequence 25. The command is running is for more than 5 hours, but still not competed.
Alertlog:
Thu Feb 5 22:14:25 2009
alter database recover to logical standby m2test
Thu Feb 5 22:14:25 2009
Media Recovery Start: Managed Standby Recovery (mtest)
Thu Feb 5 22:14:25 2009
Managed Standby Recovery not using Real Time Apply
parallel recovery started with 2 processes
Media Recovery Waiting for thread 1 sequence 25
Document says :-
If a dictionary build is not successfully performed on the primary database, this command will never complete.
But the dictionary build on primary is successful.
SQL> execute dbms_logstdby.build;
PL/SQL procedure successfully completed.
I used the following queries and to find which archive log contains dictionary build and made sure that the log archive sequence 22 is applied on standby.
SQL> SELECT NAME FROM V$ARCHIVED_LOG
WHERE (SEQUENCE#=(SELECT MAX(SEQUENCE#)
FROM V$ARCHIVED_LOG
WHERE DICTIONARY_BEGIN = 'YES' AND STANDBY_DEST='NO')); 2 3 4
NAME
/oradata/mtest/archive/mtest_1_22_677975686.arc
SQL> SELECT MAX(FIRST_CHANGE#) FROM V$ARCHIVED_LOG
WHERE DICTIONARY_BEGIN='YES'; 2
MAX(FIRST_CHANGE#)
177407
SQL>
Edited by: user592715 on Feb 6, 2009 3:22 PM -
Create table on Logical Standby Database
Dear DBAs,
here is my situation;
1. My primary database (where the tables' owner is "ICC" having a DBA) is used for insert/update transactions.
2. The logical stdby DB (dataguard structure with MAX availability) is used to generate reports, after updating some tables owned by the same user "ICC". note that these tables are excluded from the replication.
3. The developers might need to create this kind of tables to generate another kind of reports.
The issue is that when the standby apply is enabled or disabled, the user ICC is not able to create a new table, it gives ORA-01031: insufficient privileges. note that it's not practical to ask the DBA to disable the "APPLY" at any table creation.
Do you have an idea about how to resolve.
My database is 10gR2 path.10.2.0.4.0 on Windows 2003
Thank you in advanceIf you stop applying log on the logical standby you can easily create a table over there. See oracle docs. The following list shows >how to re-create a table and restart SQL Apply on that table:
>
Stop SQL Apply:
SQL> ALTER DATABASE STOP LOGICAL STANDBY APPLY;
Ensure no operations are being skipped for the table in question by querying the DBA_LOGSTDBY_SKIP view:
SQL> SELECT * FROM DBA_LOGSTDBY_SKIP;
ERROR STATEMENT_OPT OWNER NAME PROC
N SCHEMA_DDL HR EMPLOYEES
N DML HR EMPLOYEES
N SCHEMA_DDL OE TEST_ORDER
N DML OE TEST_ORDER
Because you already have skip rules associated with the table that you want to re-create on the logical standby >database, you must first delete those rules. You can accomplish that by calling the DBMS_LOGSTDBY.UNSKIP >procedure. For example:
SQL> EXECUTE DBMS_LOGSTDBY.UNSKIP(stmt => 'DML', -
schema_name => 'HR', - object_name => 'EMPLOYEES');SQL> EXECUTE DBMS_LOGSTDBY.UNSKIP(stmt >=> 'SCHEMA_DDL', -
schema_name => 'HR', - object_name => 'EMPLOYEES');
Re-create the table HR.EMPLOYEES with all its data in the logical standby database by using the >DBMS_LOGSTDBY.INSTANTIATE_TABLE procedure. For example:
SQL> EXECUTE DBMS_LOGSTDBY.INSTANTIATE_TABLE(shema_name => 'HR', -
object-_name => 'EMPLOYEES', -
dblink => 'BOSTON');
Start SQL Apply:
SQL> ALTER DATABASE START LOGICAL STANDBY APPLY IMMEDIATE;Regards. -
Error when logical standby database
Hii all,
First time i am creating logical standby database on oracle 10gR2 by oracle doc.
but When i try to create logical standby database it give error
DG 2011-07-11-09:44:24 0 2 755997031 Operation CTL_GET_STATUS cancelled during phase 2, error = ORA-16797
DG 2011-07-11-09:44:24 0 2 755997031 Operation CTL_GET_STATUS cancelled during phase 2, error = ORA-16797
DG 2011-07-11-09:45:24 0 2 0 RSM0: HEALTH CHECK ERROR: ORA-16797: database is not using a server parameter file.
when i try to create spfile from pfile it give error.
ORA-01078 failure in processing initialization parameters.
LRM-00123 invalid character number found in the input file.
Please help
Vaibhav Dixit.Vaibhav Dixit wrote:
stan.__db_cache_size=432013312
stan.__java_pool_size=4194304
stan.__large_pool_size=8388608
stan.__shared_pool_size=163577856
stan.__streams_pool_size=0
*.audit_file_dest='D:\oracle\product\10.2.0/admin/stan/adump'
*.background_dump_dest='D:\oracle\product\10.2.0/admin/stan/bdump'
*.compatible='10.2.0.1.0'
*.control_files='D:\oracle\product\10.2.0/oradata/stan/stan05.ctl'
*.core_dump_dest='D:\oracle\product\10.2.0/admin/stan/cdump'
*.db_block_size=8192
*.db_domain=''
*.db_file_multiblock_read_count=16
*.db_name='stan'
*.DB_UNIQUE_NAME='stan'
*.LOG_ARCHIVE_CONFIG='DG_CONFIG=(prim,stan)'
*.DB_FILE_NAME_CONVERT='D:\oracle\product\10.2.0\oradata\prim\','D:\oracle\product\10.2.0\oradata\stan\'
*.LOG_FILE_NAME_CONVERT= ’D:\oracle\product\10.2.0\oradata\prim\’,’D:\oracle\product\10.2.0\oradata\stan\’,'D:\oracle\product\10.2.0\flash_recovery_area\prim\ONLINELOG','D:\oracle\product\10.2.0\flash_recovery_area\stan\ONLINELOG'
*.LOG_ARCHIVE_FORMAT='%t_%s_%r.arc'
*.LOG_ARCHIVE_DEST_1= 'LOCATION=D:\oracle\product\10.2.0\flash_recovery_area\stan\ VALID_FOR=(ONLINE_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=stan'
*.LOG_ARCHIVE_DEST_2= 'SERVICE=prim LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=prim'
*.LOG_ARCHIVE_DEST_3= 'LOCATION=D:\oracle\product\10.2.0\flash_recovery_area\STAN2\ VALID_FOR=(STANDBY_LOGFILES,STANDBY_ROLE) DB_UNIQUE_NAME=stan'
*.LOG_ARCHIVE_DEST_STATE_1='ENABLE'
*.LOG_ARCHIVE_DEST_STATE_2='ENABLE'
*.LOG_ARCHIVE_DEST_STATE_3='ENABLE'
*.REMOTE_LOGIN_PASSWORDFILE='EXCLUSIVE'
*.LOG_ARCHIVE_MAX_PROCESSES=30
*.STANDBY_FILE_MANAGEMENT='AUTO'
*.FAL_SERVER='prim'
*.FAL_CLIENT='stan'
*.db_recovery_file_dest_size=9390363648
*.dispatchers='(protocol=TCP)'
*.job_queue_processes=10
*.nls_territory='INDIA'
*.open_cursors=300
*.pga_aggregate_target=202375168
*.processes=150
*.sga_target=609222656
*.shared_servers=1
*.undo_management='AUTO'
*.undo_tablespace='UNDOTBS1'
*.user_dump_dest='D:\oracle\product\10.2.0/admin/stan/udump'
*.undo_retention=3600
*.dg_broker_start='TRUE'from
'D:\oracle\product\10.2.0/admin/stan/udump'
please change or replace the forward slash with backward slash in all file locations
to
eg:- 'D:\oracle\product\10.2.0\admin\stan\udump'
and create a spfile from pfile. -
Creating a new schema in a Logical Standby Database
Hi All,
I am experimenting with logical standby databases for the purpose of reporting, and have not been able to create a new schema in the logical standby database - one of the key features of logical standbys.
I have setup primary and logical standby databases, and they seem to be running just fine - changes are moved from the primary to the standby and queries on the standby seem to run ok.
However, If I try to create a new schema on the logical standby, that does not exist on the primary, I get "ORA-01031: insufficient privileges" errors when I try to create new objects.
Show below are the steps I have taken to create the new schema on the logical standby. Any help would be greatly appreciated.
SYS@UATDR> connect / as sysdba
Connected.
SYS@UATDR>
SYS@UATDR> select name, log_mode, database_role, guard_status, force_logging, flashback_on, db_unique_name
2 from v$database
3 /
NAME LOG_MODE DATABASE_ROLE GUARD_S FOR FLASHBACK_ON DB_UNIQUE_NAME
UATDR ARCHIVELOG LOGICAL STANDBY ALL YES YES UATDR
SYS@UATDR>
SYS@UATDR> create tablespace ts_new
2 /
Tablespace created.
SYS@UATDR>
SYS@UATDR> create user new
2 identified by new
3 default tablespace ts_new
4 temporary tablespace temp
5 quota unlimited on ts_new
6 /
User created.
SYS@UATDR>
SYS@UATDR> grant connect, resource to new
2 /
Grant succeeded.
SYS@UATDR> grant unlimited tablespace, create table, create any table to new
2 /
Grant succeeded.
SYS@UATDR>
SYS@UATDR> -- show privs given to new
SYS@UATDR> select * from dba_sys_privs where grantee='NEW'
2 /
GRANTEE PRIVILEGE ADM
NEW CREATE ANY TABLE NO
NEW CREATE TABLE NO
NEW UNLIMITED TABLESPACE NO
SYS@UATDR>
SYS@UATDR> -- create objects in schema
SYS@UATDR> connect new/new
Connected.
NEW@UATDR>
NEW@UATDR> -- prove ability to create tables
NEW@UATDR> create table new
2 (col1 number not null)
3 tablespace ts_new
4 /
create table new
ERROR at line 1:
ORA-01031: insufficient privileges
NEW@UATDR>
NEW@UATDR>HI Daniel,
I appreciate your quick response.
My choice of name may not have been ideal, however changing new to another name - like gav - does not solve the problem.
SYS@UATDR> connect / as sysdba
Connected.
SYS@UATDR>
SYS@UATDR> select name, log_mode, database_role, guard_status, force_logging, flashback_on, db_unique_name
2 from v$database
3 /
NAME LOG_MODE DATABASE_ROLE GUARD_S FOR FLASHBACK_ON DB_UNIQUE_NAME
UATDR ARCHIVELOG LOGICAL STANDBY ALL YES YES UATDR
SYS@UATDR>
SYS@UATDR> create tablespace ts_gav
2 /
Tablespace created.
SYS@UATDR>
SYS@UATDR> create user gav
2 identified by gav
3 default tablespace ts_gav
4 temporary tablespace temp
5 quota unlimited on ts_gav
6 /
User created.
SYS@UATDR>
SYS@UATDR> grant connect, resource to gav
2 /
Grant succeeded.
SYS@UATDR> grant unlimited tablespace, create table, create any table to gav
2 /
Grant succeeded.
SYS@UATDR>
SYS@UATDR> -- show privs given to gav
SYS@UATDR> select * from dba_sys_privs where grantee='GAV'
2 /
GRANTEE PRIVILEGE ADM
GAV CREATE TABLE NO
GAV CREATE ANY TABLE NO
GAV UNLIMITED TABLESPACE NO
SYS@UATDR>
SYS@UATDR> -- create objects in schema
SYS@UATDR> connect gav/gav
Connected.
GAV@UATDR>
GAV@UATDR> -- prove ability to create tables
GAV@UATDR> create table gav
2 (col1 number not null)
3 tablespace ts_gav
4 /
create table gav
ERROR at line 1:
ORA-01031: insufficient privileges
GAV@UATDR> -
Logical Standby Database in NOARCHIVE Mode
Hi,
I have configured a Logical Standby Database for Reporting purposes. A Physical Standby Database is running for MAA. i.e. in case of Role Transition (switch/Failover) the Physical Stdby Db will get the role of the Primary.
The logical standby database is creating a lot of Archive Redologs files, nearly every minute. Redolog files are 50MB and there is no work done in db during the time. I'm NOT using Standby Redolog files.
Is there a need for logical standby database to be in NOARCHIVELOG mode? The Primary is definatley in ARCHIVELOG mode.
Thanks for any responses.
regards
Sahbahi,
well there are two things to the above:-
1. there was an archive file nearly every minute:
this is due to a db recovery. for some reason, the db was in inconsistent state, after a sudden shutdownof the OS. I was on a test environment, on windows vista, unfortunately. unimportant ... a reboot solved it.
2. Logical standby db in NOARCHIVE MODE when setup for the purpose of Reporting.
As long as the MAA configured for the primary db, such as physical standby db, and a second, the logical standby db setup purely for the purpose of reporting, which then can run with NOARCHIVELOG mode, after converting the physical standby db to logical.
logical standby db uses Streams architecture, so this method brings cost, time and performance advantages to the customer.
regards
Sahba -
How to delete the foreign archivelogs in a Logical Standby database
How do I remove the foreign archive logs that are being sent to my logical standby database. I have files in the FRA of ASM going back weeks ago. I thought RMAN would delete them.
I am doing hot backups of the databases to FRA for both databases. Using ASM, FRA, in a Data Guard environment.
I am not backing up anything to tape yet.
The ASM FRA foreign_archivelog directory on the logical standby FRA keeps growing and nothing is get deleted when
I run the following command every day.
delete expired backup;
delete noprompt force obsolete;
Primary database RMAN settings (Not all of them)
CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 9 DAYS;
CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
CONFIGURE DB_UNIQUE_NAME 'WMRTPRD' CONNECT IDENTIFIER 'WMRTPRD_CWY';
CONFIGURE DB_UNIQUE_NAME 'WMRTPRD2' CONNECT IDENTIFIER 'WMRTPRD2_CWY';
CONFIGURE DB_UNIQUE_NAME 'WMRTPRD3' CONNECT IDENTIFIER 'WMRTPRD3_DG';
CONFIGURE ARCHIVELOG DELETION POLICY TO APPLIED ON ALL STANDBY;
Logical standby database RMAN setting (not all of them)
CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 9 DAYS;
CONFIGURE ARCHIVELOG BACKUP COPIES FOR DEVICE TYPE DISK TO 1; # default
CONFIGURE ARCHIVELOG DELETION POLICY TO NONE; # default
How do I cleanup/delete the old ASM foreign_archivelog files?OK, the default is TRUE which is what it is now
from DBA_LOGSTDBY_PARAMETERS
LOG_AUTO_DELETE TRUE SYSTEM YES
I am not talking about deleting the Archive logs files for the Logical database that it is creating, but the Standby archive log files being sent to the Logical Database after they have been applied.
They are in the alert log as follows under RFS LogMiner: Registered logfile
RFS[1]: Selected log 4 for thread 1 sequence 159 dbid -86802306 branch 763744382
Thu Jan 12 15:44:57 2012
*RFS LogMiner: Registered logfile [+FRA/wmrtprd2/foreign_archivelog/wmrtprd/2012_01_12/thread_1_seq_158.322.772386297] to LogM*
iner session id [1]
Thu Jan 12 15:44:58 2012
LOGMINER: Alternate logfile found. Transition to mining archived logfile for session 1 thread 1 sequence 158, +FRA/wmrtprd2/
foreign_archivelog/wmrtprd/2012_01_12/thread_1_seq_158.322.772386297
LOGMINER: End mining logfile for session 1 thread 1 sequence 158, +FRA/wmrtprd2/foreign_archivelog/wmrtprd/2012_01_12/threa
d_1_seq_158.322.772386297
LOGMINER: Begin mining logfile for session 1 thread 1 sequence 159, +DG1/wmrtprd2/onlinelog/group_4.284.771760923 -
Hi Friends,
I am getting the following exception in logical standby database at the time of Sql Apply.
After run the command alter database start logical standby apply sql apply services start but after few second automatically stop and getting following exception.
alter database start logical standby apply
Tue May 17 06:42:00 2011
No optional part
Attempt to start background Logical Standby process
LOGSTDBY Parameter: MAX_SERVERS = 20
LOGSTDBY Parameter: MAX_SGA = 100
LOGSTDBY Parameter: APPLY_SERVERS = 10
LSP0 started with pid=30, OS id=4988
Tue May 17 06:42:00 2011
Completed: alter database start logical standby apply
Tue May 17 06:42:00 2011
LOGSTDBY status: ORA-16111: log mining and apply setting up
Tue May 17 06:42:00 2011
LOGMINER: Parameters summary for session# = 1
LOGMINER: Number of processes = 4, Transaction Chunk Size = 201
LOGMINER: Memory Size = 100M, Checkpoint interval = 500M
Tue May 17 06:42:00 2011
LOGMINER: krvxpsr summary for session# = 1
LOGMINER: StartScn: 0 (0x0000.00000000)
LOGMINER: EndScn: 0 (0x0000.00000000)
LOGMINER: HighConsumedScn: 2660033 (0x0000.002896c1)
LOGMINER: session_flag 0x1
LOGMINER: session# = 1, preparer process P002 started with pid=35 OS id=4244
LOGSTDBY Apply process P014 started with pid=47 OS id=5456
LOGSTDBY Apply process P010 started with pid=43 OS id=6484
LOGMINER: session# = 1, reader process P000 started with pid=33 OS id=4732
Tue May 17 06:42:01 2011
LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1417, X:\TANVI\ARCHIVE2\ARC01417_0748170313.001
Tue May 17 06:42:01 2011
LOGMINER: Turning ON Log Auto Delete
Tue May 17 06:42:01 2011
LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01417_0748170313.001
Tue May 17 06:42:01 2011
LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1418, X:\TANVI\ARCHIVE2\ARC01418_0748170313.001
LOGSTDBY Apply process P008 started with pid=41 OS id=4740
LOGSTDBY Apply process P013 started with pid=46 OS id=7864
LOGSTDBY Apply process P006 started with pid=39 OS id=5500
LOGMINER: session# = 1, builder process P001 started with pid=34 OS id=4796
Tue May 17 06:42:02 2011
LOGMINER: skipped redo. Thread 1, RBA 0x00058a.00000950.0010, nCV 6
LOGMINER: op 4.1 (Control File)
Tue May 17 06:42:02 2011
LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01418_0748170313.001
Tue May 17 06:42:03 2011
LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1419, X:\TANVI\ARCHIVE2\ARC01419_0748170313.001
Tue May 17 06:42:03 2011
LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01419_0748170313.001
Tue May 17 06:42:03 2011
LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1420, X:\TANVI\ARCHIVE2\ARC01420_0748170313.001
Tue May 17 06:42:03 2011
LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01420_0748170313.001
Tue May 17 06:42:03 2011
LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1421, X:\TANVI\ARCHIVE2\ARC01421_0748170313.001
LOGSTDBY Analyzer process P004 started with pid=37 OS id=5096
Tue May 17 06:42:03 2011
LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01421_0748170313.001
LOGSTDBY Apply process P007 started with pid=40 OS id=2760
Tue May 17 06:42:03 2011
Errors in file x:\oracle\product\10.2.0\admin\tanvi\bdump\tanvi_p001_4796.trc:
ORA-00600: internal error code, arguments: [krvxbpx20], [1], [1418], [2380], [16], [], [], []
LOGSTDBY Apply process P012 started with pid=45 OS id=7152
Tue May 17 06:42:03 2011
LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1422, X:\TANVI\ARCHIVE2\ARC01422_0748170313.001
Tue May 17 06:42:03 2011
LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01422_0748170313.001
Tue May 17 06:42:03 2011
LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1423, X:\TANVI\ARCHIVE2\ARC01423_0748170313.001
Tue May 17 06:42:03 2011
LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01423_0748170313.001
Tue May 17 06:42:03 2011
LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1424, X:\TANVI\ARCHIVE2\ARC01424_0748170313.001
LOGMINER: session# = 1, preparer process P003 started with pid=36 OS id=5468
Tue May 17 06:42:03 2011
LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01424_0748170313.001
Tue May 17 06:42:04 2011
LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1425, X:\TANVI\ARCHIVE2\ARC01425_0748170313.001
LOGSTDBY Apply process P011 started with pid=44 OS id=6816
LOGSTDBY Apply process P005 started with pid=38 OS id=5792
LOGSTDBY Apply process P009 started with pid=42 OS id=752
Tue May 17 06:42:05 2011
krvxerpt: Errors detected in process 34, role builder.
Tue May 17 06:42:05 2011
krvxmrs: Leaving by exception: 600
Tue May 17 06:42:05 2011
Errors in file x:\oracle\product\10.2.0\admin\tanvi\bdump\tanvi_p001_4796.trc:
ORA-00600: internal error code, arguments: [krvxbpx20], [1], [1418], [2380], [16], [], [], []
LOGSTDBY status: ORA-00600: internal error code, arguments: [krvxbpx20], [1], [1418], [2380], [16], [], [], []
Tue May 17 06:42:06 2011
Errors in file x:\oracle\product\10.2.0\admin\tanvi\bdump\tanvi_lsp0_4988.trc:
ORA-12801: error signaled in parallel query server P001
ORA-00600: internal error code, arguments: [krvxbpx20], [1], [1418], [2380], [16], [], [], []
Tue May 17 06:42:06 2011
LogMiner process death detected
Tue May 17 06:42:06 2011
logminer process death detected, exiting logical standby
LOGSTDBY Analyzer process P004 pid=37 OS id=5096 stopped
LOGSTDBY Apply process P010 pid=43 OS id=6484 stopped
LOGSTDBY Apply process P008 pid=41 OS id=4740 stopped
LOGSTDBY Apply process P012 pid=45 OS id=7152 stopped
LOGSTDBY Apply process P014 pid=47 OS id=5456 stopped
LOGSTDBY Apply process P005 pid=38 OS id=5792 stopped
LOGSTDBY Apply process P006 pid=39 OS id=5500 stopped
LOGSTDBY Apply process P007 pid=40 OS id=2760 stopped
LOGSTDBY Apply process P011 pid=44 OS id=6816 stopped
Tue May 17 06:42:10 2011Errors in file x:\oracle\product\10.2.0\admin\tanvi\bdump\tanvi_p001_4796.trc:
ORA-00600: internal error code, arguments: [krvxbpx20], [1], [1418], [2380], [16], [], [], []submit an SR to ORACLE SUPPORT.
refer these too
*ORA-600/ORA-7445 Error Look-up Tool [ID 153788.1]*
*Bug 6022014: ORA-600 [KRVXBPX20] ON LOGICAL STANDBY* -
Creating logical standby database
Hi all,
10.2.0.1
Following this link
http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/create_ls.htm
Where do i need to issue these statements:
SQL> EXECUTE DBMS_LOGSTDBY.BUILD;
SQL> ALTER DATABASE RECOVER TO LOGICAL STANDBY db_name;
on physical standby or primary database.
I issued the first stament on primary and second on physical standby .
IN the alert log of standby,i have the following entries.
Wed Jan 20 15:34:28 2010
Converting standby mount to primary mount.
Wed Jan 20 15:34:28 2010
ACTIVATE STANDBY: Complete - Database mounted as primary (treasury)
*** DBNEWID utility started ***
DBID will be changed from 306589979 to new DBID of 330710340 for database
.........................I am trying to create a logical standby database after creating a physical standby database.
It seems standby changed to primary which was not desired.
ThanksNot tried myself, but you might want to have a look at this URL.
It appears to suggest that you need to change names of datafiles as well as value db_name explcitly on standby. -
Creation of Logical Standby Database Using RMAN ACTIVE DATABASE COMMAND
Hi All,
I am in confusion how to create logical standby database from primary database using rman active database command.
What i did:-
Create primary database on machine 1 on RHEL 5 with Oracle 11gR2
Create standby database on machine 2 on RHEL 5 With Oracle 11gR2 from primary using RMAN active database command
Trying to create logical standby database on machine 3 on RHEL 5 with Oracle 11gR2 using RMAN active database command from primary.
The point which confuse me is to start the logical standby in nomount mode on machine 3 with which pfile like i create the pfile for standby database do i need to create the pfile for logical standby db.
I done the creation of logical standby database by converting physical standby to logical standby database
I am following the below mentioned doc for the same:
Creating a physical and a logical standby database in a DR environment | Chen Guang's Blog
Kindly guide me how to work over the same or please provide me the steps of the same.
Thanks in advance.Thanks for your reply
I already started the logical standby database with pfile in nomount mode. And successfully completed the duplication of database. by mentioning the DB_FILE_NAME_CONVERT and LOG_FILE_NAME_CONVERT parameter.
But i am not able to receive the logs on the above mentioned blog i run the sql command to check the logs but getting "no rows selected"
My primary database pfile is:
pc01prmy.__db_cache_size=83886080
pc01prmy.__java_pool_size=12582912
pc01prmy.__large_pool_size=4194304
pc01prmy.__oracle_base='/u01/app/oracle'#ORACLE_BASE set from environment
pc01prmy.__pga_aggregate_target=79691776
pc01prmy.__sga_target=239075328
pc01prmy.__shared_io_pool_size=0
pc01prmy.__shared_pool_size=134217728
pc01prmy.__streams_pool_size=0
*.audit_file_dest='/u01/app/oracle/admin/pc01prmy/adump'
*.audit_trail='db'
*.compatible='11.1.0.0.0'
*.control_files='/u01/app/oracle/oradata/PC01PRMY/controlfile/o1_mf_91g3mdtr_.ctl','/u01/app/oracle/flash_recovery_area/PC01PRMY/controlfile/o1_mf_91g3mf6v_.ctl'
*.db_block_size=8192
*.db_create_file_dest='/u01/app/oracle/oradata'
*.db_domain=''
*.db_file_name_convert='/u01/app/oracle/oradata/PC01SBY/datafile','/u01/app/oracle/oradata/PC01PRMY/datafile'
*.db_name='pc01prmy'
*.db_recovery_file_dest='/u01/app/oracle/flash_recovery_area'
*.db_recovery_file_dest_size=2147483648
*.diagnostic_dest='/u01/app/oracle'
*.dispatchers='(PROTOCOL=TCP) (SERVICE=pc01prmyXDB)'
*.fal_client='PC01PRMY'
*.fal_server='PC01SBY'
*.log_archive_config='DG_CONFIG=(pc01prmy,pc01sby,pc01ls)'
*.log_archive_dest_1='LOCATION=/u01/app/oracle/flash_recovery_area/PC01PRMY/ VALID_FOR=(ONLINE_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=pc01prmy'
*.log_archive_dest_2='SERVICE=pc01sby LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=pc01sby'
*.log_archive_dest_3='SERVICE=pc01ls LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES, PRIMARY_ROLE) DB_UNIQUE_NAME=pc01ls'
*.log_archive_dest_state_1='ENABLE'
*.log_archive_dest_state_2='DEFER'
*.log_archive_dest_state_3='DEFER'
*.log_archive_max_processes=30
*.log_file_name_convert='/u01/app/oracle/oradata/PC01SBY/onlinelog','/u01/app/oracle/oradata/PC01PRMY/onlinelog'
*.open_cursors=300
*.pga_aggregate_target=78643200
*.processes=150
*.remote_login_passwordfile='EXCLUSIVE'
*.sga_target=236978176
*.undo_tablespace='UNDOTBS1'
My logical standby pfile is:-
pc01ls.__db_cache_size=92274688
pc01ls.__java_pool_size=12582912
pc01ls.__large_pool_size=4194304
pc01ls.__oracle_base='/u01/app/oracle'#ORACLE_BASE set from environment
pc01ls.__pga_aggregate_target=79691776
pc01ls.__sga_target=239075328
pc01ls.__shared_io_pool_size=0
pc01ls.__shared_pool_size=125829120
pc01ls.__streams_pool_size=0
*.audit_file_dest='/u01/app/oracle/admin/pc01ls/adump'
*.audit_trail='db'
*.compatible='11.1.0.0.0'
*.control_files='/u01/app/oracle/oradata/PC01LS/controlfile/o1_mf_91g3mdtr_.ctl','/u01/app/oracle/flash_recovery_area/PC01LS/controlfile/o1_mf_91g3mf6v_.ctl'
*.db_block_size=8192
*.db_create_file_dest='/u01/app/oracle/oradata'
*.db_domain=''
*.db_file_name_convert='/u01/app/oracle/oradata/PC01SBY/datafile','/u01/app/oracle/oradata/PC01PRMY/datafile'
*.db_name='pc01prmy'
*.db_unique_name='pc01ls'
*.db_recovery_file_dest='/u01/app/oracle/flash_recovery_area'
*.db_recovery_file_dest_size=2147483648
*.diagnostic_dest='/u01/app/oracle'
*.dispatchers='(PROTOCOL=TCP) (SERVICE=pc01prmyXDB)'
*.log_archive_config='DG_CONFIG=(pc01prmy,pc01sby,pc01ls)'
*.log_archive_dest_1='LOCATION=/u01/app/oracle/flash_recovery_area/PC01PRMY/ VALID_FOR=(ONLINE_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=pc01prmy'
*.log_archive_dest_2='LOCATION=/u01/app/oracle/flash_recovery_area/PC01LS/ VALID_FOR=(STANDBY_LOGFILES,STANDBY_ROLE) DB_UNIQUE_NAME=pc01ls'
*.log_archive_dest_3='SERVICE=pc01ls LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES, PRIMARY_ROLE) DB_UNIQUE_NAME=pc01ls'
*.log_archive_dest_state_1='ENABLE'
*.log_archive_dest_state_2='ENABLE'
*.log_archive_max_processes=30
*.log_file_name_convert='/u01/app/oracle/oradata/PC01SBY/onlinelog','/u01/app/oracle/oradata/PC01PRMY/onlinelog'
*.open_cursors=300
*.pga_aggregate_target=78643200
*.processes=150
*.remote_login_passwordfile='EXCLUSIVE'
*.sga_target=236978176
*.undo_tablespace='UNDOTBS1'
Kindly advice over the same -
Logical Standby Database and XMLDB
I couldn't find a proper group to post this message and thought I would try here.
I want to set up a Logical Standby Database for our production database server 9.2.0.4.0 with XMLDB. I am having problem with some system tables that were created when I registered some XML schemas.
These tables are in the standby database but Oracle complains the object does not exist.
Any idea?>
The "may be" is because I have tested flashback of a physical standby to before resetlogs, but not a logical standby.
>
A physical standby keeps the DBID of the primary - a logical standby does not. That is exactly the problem that restricts the reconversion into physical from logical, and you did not encounter that problem.
>
I haven't used "keep identity" but from what I read it relates to "convert to physical" not "flashback database".
>
Exactly. And that is what the OP wants to do: convert to physical (from logical).
You mentioned that this might be possible with flashback.
Problem: During the conversion from physical to logical, the DBID gets changed unless you specify (in 11g) KEEP IDENTITY. This would make it possible to reconvert into phyiscal from logical.
In short: If there is no solution for the changed DBID of the logical standby in order to flashback it into physical as you suggested, then it is not possible .
When I saw your first answer, I thought that you might have a solution in mind in order to solve that obvious problem. Sorry for having bothered you.
Kind regards
Uwe
http://uhesse.wordpress.com -
Hi,
When we say "Logical Standby Databases are logically identical to primary databases although the physical organization and structure of the data can be different." what does it exactly means?
Does it mean that in logical standby tablespace name, schema name, table name, column names etc can be different and still has the same data as primary?
Does it mean that we can exclude indexes and constraints as present in primary?
Only the data should match with primary word by word, value by value?
I am asking this as i have never worked in a logical standby database but i seriously want to know.
Please answer.
Regards,
SIDPhysical standby differs from logical standby:
Physical standby schema matches exactly the source database.
Archived redo logs and FTP'ed directly to the standby database which is always running in "recover" mode. Upon arrival, the archived redo logs are applied directly to the standby database.
Logical standby is different from physical standby:
Logical standby database does not have to match the schema structure of the source database.
Logical standby uses LogMiner techniques to transform the archived redo logs into native DML statements (insert, update, delete). This DML is transported and applied to the standby database.
Logical standby tables can be open for SQL queries (read only), and all other standby tables can be open for updates.
Logical standby database can have additional materialized views and indexes added for faster performance.
Installing Physical standbys offers these benefits:
An identical physical copy of the primary database
Disaster recovery and high availability
High Data protection
Reduction in primary database workload
Performance Faster
Installing Logical standbys offer:
Simultaneous use for reporting, summations and queries
Efficient use of standby hardware resources
Reduction in primary database workload
Some limitations on the use of certain datatypes
Maybe you are looking for
-
I have a new PC and two iTune accounts, how can I add my ipod to my new pc
I have an ipod nano that i would like to add to the same itune account as that of my iphone. I need to put them all together. How do I do this?
-
Cannot Start Microsoft Outlook. Cannot open the Outlook Window. The set fo folders cannot be opened. You must connect to Microsoft Exchange with the current profile before you can synchronize your folders with your Outlook data file (.ost) OK. This i
-
I'm trying to convert an XML message to an ABAP internal table. However, the Transformation returns 2 empty rows in the internal table (no data). When I run XSLT to debug the transformation it seems to work. Here is the ABAP program, XSLT program
-
Why are we locked to use Walkman, when it **bleep** as it is now?
Hi, Is it really that hard to add good functionality to the Walkman app? We are forced to use it but we don't get the functionality we need! It would be so simple to set which folder to scan! I really would like to use it, if i could set it not to sc
-
Adding a markup field to a map?
Hello, I am looking to add an image of a map to a form, where people could put a mark where they want to. I have tried making 100's of check boxes, but that is.... cumbersome. Is there a way to have a field that will place a mark wherever they click?