Logical Standby Data Consistency issues

Hi all,
We have been running a logical standby instance for about three weeks now. Both our primary and logical are 11g (11.1.0.7) databases running on Sun Solaris.
We have off-loaded our Discoverer reporting to the logical standby.
About three days ago, we started getting the following error message (initially for three tables, but from this morning on a whole lot more)
ORA-26787: The row with key (<coulmn>) = (<value>) does not exist in table <schema>.<table>
This error implies that we have data consistency issues between our primary and logical standby databases, but we find that hard to believe
because the "data guard" status is set to "standby", implying that schemas' being replicated by data guard are not available for user modification.
any assistance in this regard would be greatly appreciated.
thanks
Mel

It is a bug : Bug 10302680 . Apply the corresponding Patch 10302680 to your standby db.

Similar Messages

  • Logical standby rolling upgrade issue

    Hello together,
    im experiencing some strange error during the setup of a logical standby environment.
    primary system running: 10.1.0.3.0 - 64bit Production
    log. standby system running: 10.1.0.5.0 - 64bit Production
    compatibility flags set to 10.1.0.3.0 on the standby side.
    now everything works fine till it comes to the point where I have to register the logical logfile after building the logminer dictionary on the primary site.
    Execution flow on the primary site:
    SQL> EXECUTE DBMS_LOGSTDBY.BUILD;
    SQL> SELECT NAME FROM V$ARCHIVED_LOG WHERE (SEQUENCE#=(SELECT MAX(SEQUENCE#) FROM V$ARCHIVED_LOG WHERE DICTIONARY_BEGIN = 'YES' AND STANDBY_DEST= 'NO'));
    gives me:
    /oradata/primary/arch/585226884_1_10.arc
    Copied to the standby side and tried to register the logfile:
    SQL> alter database register logical logfile '/oradata/standby/arch/585226884_1_10.arc';
    alter database register logical logfile '/oradata/standby/arch/585226884_1_10.arc'
    ERROR at line 1:
    ORA-16225: Missing LogMiner session name for Streams
    anyone knows what could cause ?
    thanks in advance
    Chris

    It isn't documented anywhere that I know of but yes you can.
    Stop SQL Apply, set your Guaranteed Restore Point (GRP) in the Logical standby, run your replay, get the results, flashback the standby to the GRP, open reset logs and restart the apply.
    You would need to figure out at what point you would want to run the capture at the primary and stop the SQL Apply at that point or else you will need to flash the logical standby back to the point you started the capture at the Primary and set your Guaranteed Restore point at that point in the database so that Replay can operate correctly (i.e. on the same set of data).
    Bear in mind that you during this process you will have two GRP's. One on the standby and one on the Primary. So you need room for the flashback logs at the standby to hold the flashback data and archive logs generated by the replay as well as room for the archive logs coming in from the Primary. And you will need room for the flashback logs at the Primary because it will be still running production during your replay period and generating flashback data there as well.
    Larry

  • Logical Standby....

    Hi, I have Created Logical Standby data base on my windows 2003 Advance Server Machine and Oracle 10.2.0.3.
    now I am having some problem that my LOGSTDBY is not in sync with the Primary DB from few days and in alertlog it is showing this error....
    LOGMINER: Parameters summary for session# = 1
    LOGMINER: Number of processes = 3, Transaction Chunk Size = 201
    LOGMINER: Memory Size = 30M, Checkpoint interval = 150M
    LOGMINER: session# = 1, builder process P001 started with pid=18 OS id=2820
    LOGMINER: session# = 1, reader process P000 started with pid=17 OS id=2816
    LOGSTDBY Analyzer process P003 started with pid=20 OS id=2828
    LOGSTDBY Apply process P005 started with pid=39 OS id=2984
    LOGSTDBY Apply process P006 started with pid=40 OS id=2988
    Fri Aug 21 11:56:17 2009
    LOGMINER: Begin mining logfile: G:\ORACLE\ORADATA\SOS\ARCHIVE\SOST001S63660.ARC
    Fri Aug 21 11:56:17 2009
    LOGMINER: Turning ON Log Auto Delete
    LOGSTDBY Apply process P004 started with pid=22 OS id=2980
    LOGSTDBY Apply process P008 started with pid=42 OS id=2996
    LOGMINER: session# = 1, preparer process P002 started with pid=19 OS id=2824
    LOGSTDBY Apply process P007 started with pid=41 OS id=2992
    Fri Aug 21 11:56:26 2009
    Errors in file d:\oracle\admin\sos\bdump\logstdby_p004_2980.trc:
    ORA-00604: error occurred at recursive SQL level 2
    ORA-00955: name is already used by an existing object
    Fri Aug 21 11:56:26 2009
    LOGSTDBY status: ORA-00955: name is already used by an existing object
    LOGSTDBY Apply process P004 pid=22 OS id=2980 stopped
    Fri Aug 21 11:56:29 2009
    Errors in file d:\oracle\admin\sos\bdump\logstdby_lsp0_2944.trc:
    ORA-12801: error signaled in parallel query server P004
    ORA-00955: name is already used by an existing object
    LOGSTDBY Apply process P006 pid=40 OS id=2988 stopped
    LOGSTDBY Analyzer process P003 pid=20 OS id=2828 stopped
    LOGSTDBY Apply process P005 pid=39 OS id=2984 stopped
    LOGSTDBY Apply process P007 pid=41 OS id=2992 stopped
    LOGSTDBY Apply process P008 pid=42 OS id=2996 stopped
    Fri Aug 21 11:58:50 2009
    Redo Shipping Client Connected as PUBLIC
    Plz rply ASAP

    Here are the steps to identify the object in question:
    1) Start the replication on logical standby
    2) Once the replication starts it will show you the same error in the alert log file as you mentioned above.
    3) Query DBA_LOGSTDBY_EVENTS order by EVENT_TIMESTAMP. The last event will capture the name of the object that is creating the problem.
    4) Based on your knowledge of database /application objects, you can skip that table using DBMS_LOGSTADBY.SKIP and start the replication again without any issue.
    5) If you dont need that object in the logical DB to be replicated then let it be skipped, otherwise reinstate the object again in logical standby.
    Amar.

  • Transportable tablespaces with Logical Standby

    Does anyone know whether or not transportable tablespaces can be used in a Logical Standby environment? I know they can be used with a Physical Standby, and the documentation covers how to do this, but there's nothing about Logical Standby. This is Oracle 10.2.0.4/SLES 10.
    Thanks.

    Using transportable tablespaces to a logical standby environment. Actually, there is a paragraph in the 10.2 Data Guard Concepts and Administration manual. I will summarize:
    After creating the transportable tablespace export file on the source system, copy the datafiles to both the primary and logical standby systems.
    On the dataguard logical standby, within SQLplus, issue SQL> alter session disable guard;
    Import the tablespaces into the logical standby database
    SQL> alter session enable guard;
    Import the tablespaces into the primary database.
    In 10.1, we imported the tablespaces into an environment with a physical standby and then converted the physical standby to a logical standby.
    I tested the procedure for 10.2 and it seemed to work. Has anyone else done this and if so, are you aware of any issues? Like I said, the documentation is one very brief paragraph.
    Thanks.

  • Oracle 10g Logical Standby

    The question is on Oracle 10g Logical Standby Data Guard. Below is the short description of what we wanted to do.
    We currently run our Remedy application on Oracle10g. We wanted to setup Data Guard (logical standby) so we can perform our application upgrade on the logical standby. While we upgrade our application, our customers perform the dml on the primary instance.
    During the upgrade, we will change the ddl on our application in the logical standby instance. After the application upgrade, we do a role switch between the two instances (primary became a new logical standby and logical standby became new primary). Now, we wanted the data or the dml that our customers perform on the old primary replicate over to the new primary.
    It is possible for the logical standby to perform the ddl at the same time user perform the ddl and dml on the primary instance?
    After role switch between the two instances, is the data changed on the primary replicate over to the new primary?
    Thank you for your helps.
    Loc Nguyen
    [email protected]

    Hi Shital,
    Guard_status protects the data from being changed.
    ALL- By default it is not possible for a non-privileged user to modify data on a data guard SQL apply database. This is because the database guard is automatically set to ALL.
    With this level of security, only SYS user can modify the data.
    STANDBY- When you set this level of security, users are able to modify data that is not maintained by the Logical apply engine.
    NONE permits any users to access the standby database as long as they have correct privileges. This is the normal security for all data in the database.
    You can change the guard status value from ALL to NONE in order to allow non-privileged users to modify data and Yes you can create user and extra tablespace in logical standby database..this is what the one of advantage of using Logical standby database.
    SQL> ALTER DATABASE GUARD NONE;
    Thanks

  • Logical Standby working issues Oracle 9i, Windows

    Hi,
    Set up Oracle 9i Logical Standby on Windows. (instructions as per Oracle Documentation)
    Did not have any issues setting up.
    While setting up the Logical Standby, Recovered the Primary Database until Oct 10/09 8:16 pm
    Registered the archive log in the logical standby generated hence and the FAL took care of copying/registering the rest of the archivelogs.
    Created and inserted some records in Primary database and could see them in Standby.
    So far so good.
    On Oct11 data was entered into Primary database. Archivelogs were shipped to Standby, I could see them registered in DBA_LOGSTDBY_LOG.
    The APPLIED_SCN,NEWEST_SCN were in sync as per DBA_LOGSTDBY_PROGRESS.
    Today, we had some issues with data and when we queried the user tables: (no skip settings)
    Couldn't see any data in standby past the recovery...
    No errors reported in DBA_LOGSTDBY_EVENTS. No errors in Alert log also.
    What could be happening?
    Thanks,
    Madhuri

    I figured it out...
    Today, we had some issues with data and when we queried the user tables: (no skip settings)
    Couldn't see any data in standby past the recovery...I was using two tables as random spot check and both did not get updated. So, I was under the impression SQL APPLY did not do anything.
    But, it did apply the redo on the rest of the tables.
    These 2 tables in question were skipped because both of them had Function Based indexes.
    They are very huge individual tables .
    So, exporting them from Primary database and Importing them into Standby Database. Skipping DML in DataGuard.
    That solved the problem.
    --Madhuri                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

  • Logical Standby Issue

    Hi Friends,
    I have 4 doubts. Please help me to clear my confusion.
    Doubt 1:-
    I have one primary and one logical standby database. When i restart the primary database or logical standby database or whenever data is not transffering to standby database, I execute the package "EXECUTE DBMS_LOGSTDBY.BUILD;". Then it starts to transfer the data to standby. My question is executing this package many times will affect the primary database? or my configuration is wrong?
    Doubt 2:-
    I configured DGMGRL for both database and i can switch over and switch back to logical standby database. It automatically start to apply the data to new logical standby database. But when i switch back to old primary database, It didnt transfer the data automatically. So i execute the package "EXECUTE DBMS_LOGSTDBY.BUILD;" in primary. Then it starts to transfer the data. My question is my switchback is wrong? or it happens like that only?
    Note:- I check the verbose of both database before switchback. It show success.
    Doubt 3:-
    When i check the switchover status in logical standby database, It show "NOT ALLOWED". Even though i can switch over by using dgmgrl. Why logical standby database showing "NOT ALLOWED"?
    Doubt 4:-
    Is it any possible to view that what commands executing by dgmgrl process in background while it switchovering to logical standby database.?
    Because when i try to switchover manually it didnt work. Many times i tried. But in dgmgrl, I simply give "switchover to logdb;", It works.
    Please help to me to understand clearly.
    Thank you in advance.

    Wrong forum mate. This forum deals with SQL and PL/SQL languages and related issues - not database administration and configuration issues.
    The shared memory error is not unusual though ito SQL and PL/SQL. It is often a result of not using sharable SQL - in other words, SQL statements that are not using bind variables.
    This causes additional consumption and fragmentation of shared pool memory. The very worse thing to do in this case would be to increase the size of the shared pool. That is akin to moving the wall a bit further away so that one can run even faster and harder into it.
    But I'm merely speculating as you may have another cause for this problem. I suggest that you:
    - post this in the [url http://forums.oracle.com/forums/forum.jspa?forumID=61]Database Forum
    - research this error on [url http://metalink.oracle.com]Metalink

  • DATA Guard Logical Standby v.s. Streams Apply (10gR2, rac -- rac )

    Greetings -
    We currently have a 3 node cluster, doing a 'Schema' capture, with a 2 node cluster serving as the apply side. Both are on 10gR2 (solaris 10)
    We have been experiencing apply latency due to large transactions. The way logminer/streams evaluates the arch logs, it converts each updated row into a transaction as part of a transaction set, using 'decode (' statements.
    I am under the impression a physical standby will do the same thing. But what about a logical Standby ?
    (this from the Oracle Documentation)
    [10g Concepts|http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/concepts.htm#SBYDB00010]
    ...Contains the same logical information as the production database, although the physical organization and structure of the data can be different. The logical standby database is kept synchronized with the primary database though SQL Apply, which transforms the data in the redo received from the primary database into SQL statements and then executing the SQL statements on the standby database.
    Does anyone know firsthand if SQL APPLY re-issues the original sql, or if it relies on the 'decode' command as well ?
    Thanks
    The Nets Edge

    A Physical standby does not do anything with SQL. It is running media recovery under the covers. A Logical standby uses the Logminer to read the redo, converts it to SQL and Data and applies those transactions to the Logical standby.
    The both build logical change records, figure out order, dependencies etc etc and applies the transaction. Both are susceptible to long running transactions on the Primary (source) database.
    Streams uses Logminer, Logical Standby use part of the Streams apply so as Mr. Morgan said, they are very similar :^)
    If you are having apply performance issues you may want to look at Active data Guard.
    Larry

  • Data Guard Logical Standby DB Questions

    Hi,
    I am creating Oracle 9i Logical Standby database on AIX 5 on same server ( for
    testing only ).
    In Data Guard Manual ( PDF file page no : 86 ) it is said that
    Step 4 Create a backup copy of the control file for the standby database. On the
    primary database, create a backup copy of the control file for the standby database:
    SQL> ALTER DATABASE BACKUP CONTROLFILE TO
    2> '/disk1/oracle/oradata/payroll/standby/payroll3.ctl';
    My Questions regarding this are
    1. Does it missing "backup standby controlfile" word ? if not what is the use of
    this control file ? Because as per this manual section 4.2.2 step no 1 and 2 I
    have stopped the primary database and copied datafiles,logfiles and controlfiles
    as specified which are having consistent stage.
    2. On primary database I am mirroring 2 controlfiles at two diffrent locations.
    On Standby Database I am going to keep same policy. Then on standby database do I have to copy this controlfile to two locations and rename datafiles or I have
    to user controlfiles which I have backed up along with datafiles ?
    3. Suppose my primary database is "infod" and standby database is "standby".
    Then what should be value of log_file_format parameter of standby database ? On
    primary db I have defined "infod LOG_ARCHIVE_FORMAT=infod_log%s_%t.arc. Do I have to keep same or change it ? If on standby db I change the value do I
    require other parameters to be set ?
    regards & thanks
    pjp

    Q/A1) Its correct you dont need the standby keyword. Not sure abut the reason why but we have created logical standby running in prod using the oracle doc.
    Q/A2) Yo can specify any location and name for your controlfiles as it is instance specific and not depending on primary database.
    Q/A3) You can set any format for your archlogs on standby side. It doesnt bother the primary db or DG configuration.
    Regards,
    http://askyogesh.com

  • ORA-01403: no data found on LOGICAL STANDBY database

    Hi ,
    Logical Standby issue :
    Oracle 10.2.0.2 enterprise edition .
    M Working on LOGICAL Standby since 1 yrs but still i havent got this ......................................
    I m getting countinuously no data foud errror on logical standby database .
    I found the table causing the proble(db_logstdby_events) and skipped that table and instanciated table using bwlow package:
    exec dbms_logstdby.instantiate_table (.......................................
    but when i start apply process on logical standby it again give no data found for new table :
    Even i tried to instantiate the table using EXPORT/IMPORT during down time but the same facing same problem .
    As much as i known abt the error that is :
    table1
    id
    10
    20
    30
    Now if sql apply process on logical standby tries to performe the update transaction(for example) as belows
    update table1 set id=100 where id=50;
    above query will not be completed cos it will never find the values 50 which is not in table .Thts why this error comming ..
    Now my worry is ... no users dare to change/make such changes on Logical standby .So if there is no changes in tables then sqll apply should get all the values to be needded for an update ......
    watingggg guyssss/......

    Troubleshooting ORA-1403 errors with Flashback Transaction
    In the event that the SQL Apply engine errors out with an ORA-1403, it may be possible to utilize flashback transaction on the standby database to reconstruct the missing data. This is reliant upon the undo_retention parameter specified on the standby database instance.
    ORA-1403: No Data Found
    Under normal circumstances the ORA-1403 error should not be seen in a Logical Standby environment. The error occurs when data in a SQL Apply managed table is modified directly on the standby database, and then the same data is modified on the primary database.
    When the modified data is updated on the primary database and received by the SQL Apply engine, the SQL Apply engine verifies the original version of the data is present on the standby database before updating the record. When this verification fails, an ORA-1403: No Data Found error is thrown by Oracle Data Guard: SQL Apply.
    The initial error
    When the SQL Apply engine verification fails, the error thrown by the SQL Apply engine is reported in the alert log of the logical standby database as well as a record being inserted into the DBA_LOGSTDBY_EVENTS view. The information in the alert log is truncated, while the error is reported in it's entirety in the database view.
    LOGSTDBY stmt: update "SCOTT"."MASTER"
    set
    "NAME" = 'john'
    where
    "PK" = 1 and
    "NAME" = 'andrew' and
    ROWID = 'AAAAAAAAEAAAAAPAAA'
    LOGSTDBY status: ORA-01403: no data found
    LOGSTDBY PID 1006, oracle@staco03 (P004)
    LOGSTDBY XID 0x0006.00e.00000417, Thread 1, RBA 0x02dd.00002221.10
    The Investigation
    The first step is to analyze the historical data of the table that threw the error. This can be achieved using the VERSIONS clause of the SELECT statement.
    SQL> select versions_xid
    , versions_startscn
    , versions_endscn
    , versions_operation
    , pk
    , name
    from scott.master
    versions between scn minvalue and maxvalue
    where pk = 1
    order by nvl(versions_startscn,0);
    VERSIONS_XID VERSIONS_STARTSCN VERSIONS_ENDSCN V PK NAME
    03001900EE070000 3492279 3492290 I 1 andrew
    02000D00E4070000 3492290 D 1 andrew
    Depending upon the amount of undo retention that the database is configured to retain (undo_retention) and the activity on the table, the information returned might be extensive and the versions between syntax might need to be changed to restrict the amount of information returned.
    From the information returned, it can be seen that the record was first inserted at scn 3492279 and then was deleted at scn 3492290 as part of transaction ID 02000D00E4070000. Using the transaction ID, the database should be queried to find the scope of the transaction. This is achieved by querying the flashback_transaction_query view.
    SQL> select operation
    , undo_sql
    from flashback_transaction_query
    where xid = hextoraw('02000D00E4070000');
    OPERATION UNDO_SQL
    DELETE insert into "SCOTT"."MASTER"("PK","NAME") values
    ('1','andrew');
    BEGIN
    Note that there is always one row returned representing the start of the transaction. In this transaction, only one row was deleted in the master table. The undo_sql column when executed will restore the original data into the table.
    SQL> insert into "SCOTT"."MASTER"("PK","NAME") values ('1','andrew');
    SQL> commit;
    The SQL Apply engine may now be restarted and the transaction will be applied to the standby database.
    SQL> alter database start logical standby apply;

  • Logical standby issue after adding a new user at production db server

    Hello Experts,
    I am facing a issue related to logical standby db server where after adding a new user at production db server logical standby has stopped sql apply. Can anybody give some idea to find out the root cause and give some tips to resolve the issue.
    tried below .....
    SELECT EVENT_TIME, STATUS, EVENT FROM DBA_LOGSTDBY_EVENTS
    ORDER BY EVENT_TIMESTAMP, COMMIT_SCN, CURRENT_SCN;
    got error
    ORA-16227: DDL skipped due to missing object
    ORA-01403: no data found
    ORA-00942: table or view does not exist
    OS : RHEL 4 (IBM P5)
    DB : Oracle 10g (10.2.0.1)
    Achyot

    Please let me know if can skip the DDL at which it is failing.
    ALTER DATABASE STOP LOGICAL STANDBY APPLY;
    Use EXECUTE DBMS_LOGSTDBY.SKIP
    ALTER DATABASE START LOGICAL STANDBY APPLY;
    Thanks,
    Renu

  • Goldengate from physical standby with Active data guard or from logical standby - which is better?

    Hi,
    We have Oracle 11.2.0.4 and are currently using Active Data Guard on a physical standby. My boss wants me to use Goldengate
    to replicate several schema from Oracle to SqlSserver. She doesn't want to touch the Primary, which is in production. Would it be better
    to setup a Logical standby for Goldendate to use, in addition to the Physical standby? Or will it be sufficient to use the Physical standby?
    The application tables don't all have primary keys. I'm trying to decide if the Logical standby adds another layer of complexity or would
    introduce delays in getting the data from Oracle to Sqlserver.
    Thanks in advance,

    Hi,
    In this case I recommend you to use physical standby + Golden gate with ALO mode. You might have sligthly more trouble in logical standby. it means you might have more trouble at Goldengate's extract side.
    You said
    The application tables don't all have primary keys.
    so you have to add keycols on replicat side.
    I recommend you to add an index which should consist of keycols columns. it will be necessary for replicat performance.
    Best Regards

  • Interesting issue with Logical Standby and database triggers

    We have a Logical Standby that each month we export (expdp) a schema (CSPAN) that is being maintained by SQL Apply and import (impdp)it to a 'frozen copy' (eg CSPAN201104) using REMAP_SCHEMA.
    This works fine although we've noticed that because triggers on the original schema being exported have the original schema (CSPAN) hard-referenced in the definition are imported into and owned by the new 'frozen' schema but are still 'attached' to the original schema's tables.
    This is currently causing the issue where the frozen schema trigger is INVALID and causing the SQL Apply to fail. This is the error:
    'CSPAN201104.AUD_R_TRG_PEOPLE' is
    invalid and failed re-validation
    Failed SQL update "CSPAN"."PEOPLE" set "ORG_ID" = 2, "ACTIVE_IND" = 'Y', "CREATE_DT" = TO_DATE('22-JUL-08','DD-MON-RR'),"CREATOR_NM" = 'LC', "FIRST_NM" = 'Test', "LAST_PERSON" ='log'...
    Note: this trigger references the CSPAN schema (...AFTER INSERT ON CSPAN.PEOPLE...)
    I suspect that triggers on a SQL Apply Maintained schema in a Logical Standby do not need to be valid (since they do not fire) but what if they reference a SQL Apply schema but are 'owned' by a non-SQL Apply schema? This trigger references a SQL Apply table so it should not fire
    This is 10gR2 (10.2.0.4) on 64 bit Windows.
    Regards
    Graeme King

    OK, I've finally got around to actually test this and it looks like you are not quite correct Larry in this statement...
    'Since this trigger belongs to a new schema that is not controlled by SQL Apply (CSPAN201105) it will fire. But the trigger references a schema that is controlled by SQL Apply (CSPAN) so it will fail because it has to be validated.'
    My testing concludes that even though the trigger belongs to a schema CSPAN201105 (not controlled by SQL Apply) and references a schema controlled by SQL Apply - it does not fire. However it DOES need to be valid or it breaks SQL Apply.
    My testing was as follows:
    Primary DB
    Create new EMP table in CSPAN schema on Primary
    Create new table TRIGGER_LOG in CSPAN schema on Primary
    Create AFTER INSERT/UPDATE trigger on CSPAN.EMP table (that inserts into TRIGGER_LOG table)
    **All of the above replicates to Standby**
    Standby DB
    Create new table TRIGGER_LOG_STNDBY in CSPAN201105 schema on Primary
    Create new trigger in CSPAN201105 schema that fires on INSERT/UPDATE on CSPAN.EMP but that inserts into CSPAN201105.TRIGGER_LOG_STNDBY table)
    Primary DB
    Insert 4 rows into CSPAN.EMP
    Update 2 rows in CSPAN.EMP
    TRIGGER_LOG table has 6 rows as expected
    Standby DB
    TRIGGER_LOG table has 6 rows as expected
    TRIGGER_LOG_STNDBY table has **0 rows**
    Re-create trigger in CSPAN201105 schema that fires on INSERT/UPDATE on CSPAN.EMP but that inserts into CSPAN201105.TRIGGER_LOG_STNDBY table) **but with syntax error**
    Primary DB
    Update 1 row in CSPAN.EMP
    TRIGGER_LOG table has 7 rows as expected
    Standby DB
    SQL Apply is broken - ORA-04098: trigger 'CSPAN201105.TEST_TRIGGER_TRG' is invalid and failed re-validation

  • Can a physical standby propagate data to a logical standby?

    Here is my situation. I have a RAC database that currently has a physical standby (also RAC cluster) for disaster recovery (DR) and I am about to create an additional logical standby for real-time reporting. My customer is asking if I can also create a logical standby for reporting that resides in the DR environment that would be available in a disaster. My question is can a physical standby propagate data to a logical standby? Is there any special configuration that would be required and what might it be if this is possible?
    Thanks.

    Hi,
    Welcome. Yes it's possible, example in 11.2 :
    A cascading standby database can cascade primary database redo to up to 30 cascaded destinations, each of which can be a physical, logical, or snapshot standby database.
    http://download.oracle.com/docs/cd/E11882_01/server.112/e17022/log_transport.htm#SBYDB5122
    Best regards
    Phil

  • Data file time stamp on Logical Standby

    Dear All,
    My Logical Standby Database that is Oracle 10gR2 on Wondows 2003 Server is fully synced with the production database.
    If i query number of records on Logical Standby they are the same as on production.
    And when i query applied_time from dba_logstby_progress view, there is hardly a difference of second since I have configured real time apply.
    BUT the date and time stamp of my data files and index files on Logical Standby is more than 12 days old.
    I am confused, why this its like this. Is there any workout to fix that? Is it normal? or what?
    Kindly help.
    Regards, Imran

    misterimran wrote:
    Dear All,
    My Logical Standby Database that is Oracle 10gR2 on Wondows 2003 Server is fully synced with the production database.
    If i query number of records on Logical Standby they are the same as on production.
    And when i query applied_time from dba_logstby_progress view, there is hardly a difference of second since I have configured real time apply.
    BUT the date and time stamp of my data files and index files on Logical Standby is more than 12 days old.
    I am confused, why this its like this. Is there any workout to fix that? Is it normal? or what?
    Kindly help.
    Regards, Imran
    Is there any workout to fix that?
    Who said it is something to "fix"?

Maybe you are looking for

  • Changes in holiday and factory calender ?

    Hi Gurus, i have to delete a holiday from my holiday calendar ? i  have some doubts about transporting the changes to quality and prd can anybody explain how we have to trans Port CALENDER

  • Downgrading BI7.0 Queries to BW3.5

    Hi Experts,             I have created some queries in BI7.0 query designer and now my client wants me recreate those in Bw3.5 version due to some technical compatibility issue.          So, can anybody let me know is there any program/FM/method to d

  • Employee tax types and tax rates

    Hi gurus, I have a requirement in that i need to get the employee tax  rates. Is tax rates are maintained in SAP  or BSI. If tax rates are maintained in sap what are the tax rate tables. If they are maintained in BSI, is their any FM'S to get the tho

  • [SOLVED] Installing the latest development version of GNOME

    Is there a way to install GNOME (2.26 development version) from SVN or from an Arch repo? I can't find any information- Last edited by gjoellee (2009-03-01 13:56:34)

  • Collecting values from multiple check boxes with same name

    Hi, I have a jsp-page with one form with several check boxes sharing the same name (but with different values). Can I collect the values in an array/list or something? /jsp-beginner and google ain't my friend in this case