Logical standby stuck at initializing SQL apply only coordinator process up

Hi
OS: solaris 5.10
Hardware: sun sparc
Oracle database: 11.2.0.1.0
Primary database name: asadmin
Standby database name: test
I had been trying to convert a physical standby to logical standby database. Both the primary and standby reside on the same machine.
The physical standby was created with a hot backup of primary.
I had been following document id 278371.1 to convert the physical to logical standby and used the following steps:
Relevant init parameters on primary:
*.db_name='asadmin'
*.db_unique_name='asadmin'
*.log_archive_config='dg_config=(asadmin,test)'
*.log_archive_dest_1='location=/u01/asadmin/archive valid_for=(all_logfiles,all_roles) db_unique_name=asadmin'
*.log_archive_dest_2='SERVICE=test async valid_for=(online_logfiles,primary_role) db_unique_name=test'
*.log_archive_dest_state_1='enable'
*.log_archive_dest_state_2='enable'
*.fal_client='asadmin'
*.fal_server='test'
*.remote_login_passwordfile='EXCLUSIVE'
Relevant init parameters on standby database:
*.db_name='test' -- Was asadmin before I renamed the DB during conversion to logical standby
*.db_unique_name='test'
*.log_archive_dest_1='location=/u01/test/archive valid_for=(all_logfiles,all_roles) db_unique_name=test'
*.log_archive_dest_2='service=asadmin async valid_for=(online_logfiles,primary_role) db_unique_name=asadmin'
*.log_archive_dest_state_1=enable
*.log_archive_dest_state_2=defer
*.remote_login_passwordfile='EXCLUSIVE'*.fal_server=test
*.fal_client=asadmin
Steps on primary:
1) alter system set log_archive_dest_state_2=defer;
2) shutdown immediate;
3) Made sure that the physical standby has applied all of the redo sent to it following the shutdown.
4) startup mount;
5) ALTER DATABASE BACKUP CONTROLFILE to '/home/oracle/control01.ctl';
6) ALTER SYSTEM ENABLE RESTRICTED SESSION;
7) ALTER DATABASE OPEN;
8) Verified that the supplemental logging is on.
9) ALTER SYSTEM ARCHIVE LOG CURRENT;
10) Checked for the checkpoint change no. at this point which is 72403818 and is present in archive log file 1_62_775102253.dbf
11) EXECUTE DBMS_LOGSTDBY.BUILD;
12) ALTER SYSTEM ARCHIVE LOG CURRENT;
13) Checked for the archive log containing dictionary build which is 1_64_775102253.dbf
14) ALTER SYSTEM DISABLE RESTRICTED SESSION;
Details of archive logs and related checkpoint change nos:
NAME FIRST_CHANGE# NEXT_CHANGE#
/u01/asadmin/archive/1_61_775102253.dbf 72402901 72403817
/u01/asadmin/archive/1_62_775102253.dbf 72403817 72404069
/u01/asadmin/archive/1_63_775102253.dbf 72404069 72404211
/u01/asadmin/archive/1_64_775102253.dbf 72404211 72405700
Steps on standby:
1) shutdown immediate;
2) Copy the archivelog file 61(was created at primary after apply stopped at standby), 62(contains checkpoint no. 72403818), 63 and 64(contains dictionary build). Copy the backup controlfile from step 5 above to the controlfile location in standby init.
3) startup mount;
4) Rename all datafiles and redo log files (including standby redo log files) to the correct path on standby.
5) alter database recover automatic from '/u01/test/archive' until change 72405700 using backup controlfile; -- This completed error-free
6) alter database guard all; -- this completed error free
7) alter database open resetlogs; -- this completed error free.
8) nid target=sys/oracle12 dbname=test
9) Changed the db_name in init file to new name test.
10) Added a tempfile to temp tablespaces.
11) ALTER DATABASE REGISTER LOGICAL LOGFILE '/u01/test/archive/1_61_775102253.dbf'; -- ORA-16225: Missing LogMiner session name for Streams
12) ALTER DATABASE START LOGICAL STANDBY APPLY INITIAL 72405700; -- This completed error free.
Also enabled the log_archive_dest_state_2 on primary.
After this output from some views:
SQL> SELECT SESSION_ID, STATE FROM V$LOGSTDBY_STATE;
SESSION_ID STATE
1 INITIALIZING
SQL> SELECT SID, SERIAL#, SPID, TYPE FROM V$LOGSTDBY_PROCESS;
SID SERIAL# SPID TYPE
587 22 15476 COORDINATOR
SELECT PERCENT_DONE, COMMAND
FROM V$LOGMNR_DICTIONARY_LOAD
WHERE SESSION_ID = (SELECT SESSION_ID FROM V$LOGSTDBY_STATE);
PERCENT_DONE
COMMAND
0
SQL> SELECT TYPE, HIGH_SCN, STATUS FROM V$LOGSTDBY;
TYPE HIGH_SCN STATUS
COORDINATOR ORA-16111: log mining and apply setting up
SQL> SELECT APPLIED_SCN, NEWEST_SCN FROM DBA_LOGSTDBY_PROGRESS;
APPLIED_SCN NEWEST_SCN
72405700 72411501
SELECT THREAD#, SEQUENCE#, FILE_NAME FROM DBA_LOGSTDBY_LOG L
WHERE NEXT_CHANGE# NOT IN
(SELECT FIRST_CHANGE# FROM DBA_LOGSTDBY_LOG WHERE L.THREAD# = THREAD#)
ORDER BY THREAD#,SEQUENCE#;
no rows selected
SQL> SELECT EVENT_TIME, STATUS, EVENT
FROM DBA_LOGSTDBY_EVENTS
ORDER BY EVENT_TIMESTAMP, COMMIT_SCN; 2 3
EVENT_TIME STATUS EVENT
14-FEB-12 02:00:50 ORA-16111: log mining and apply setting up
14-FEB-12 02:00:50 Apply LWM 72405699, HWM 72405699, SCN 72405699
14-FEB-12 02:20:11 ORA-16128: User initiated stop apply successfully
completed
14-FEB-12 02:20:39 ORA-16111: log mining and apply setting up
14-FEB-12 02:20:39 Apply LWM 72405699, HWM 72405699, SCN 72405699
14-FEB-12 02:54:15 ORA-16128: User initiated stop apply successfully
completed
14-FEB-12 02:57:38 ORA-16111: log mining and apply setting up
EVENT_TIME STATUS EVENT
14-FEB-12 02:57:38 Apply LWM 72405699, HWM 72405699, SCN 72405699
14-FEB-12 03:01:36 ORA-16128: User initiated stop apply successfully
completed
14-FEB-12 03:13:44 ORA-16111: log mining and apply setting up
14-FEB-12 03:13:44 Apply LWM 72405699, HWM 72405699, SCN 72405699
14-FEB-12 04:32:23 ORA-16128: User initiated stop apply successfully
completed
14-FEB-12 04:34:17 ORA-16111: log mining and apply setting up
14-FEB-12 04:34:17 Apply LWM 72405699, HWM 72405699, SCN 72405699
EVENT_TIME STATUS EVENT
14-FEB-12 04:36:16 ORA-16128: User initiated stop apply successfully
completed
14-FEB-12 04:36:21 ORA-16111: log mining and apply setting up
14-FEB-12 04:36:21 Apply LWM 72405699, HWM 72405699, SCN 72405699
14-FEB-12 05:15:22 ORA-16128: User initiated stop apply successfully
completed
14-FEB-12 05:15:29 ORA-16111: log mining and apply setting up
14-FEB-12 05:15:29 Apply LWM 72405699, HWM 72405699, SCN 72405699
I also greped for lsp and lcr processes and found that lsp is up but do not see any lcr.
The logs are getting transported to the archive destination on standby whenever they are archived on primary but are not getting applied to standby.
Also in case the standby is down while a log is generated on primary it is not automatically transported to standby once the standby is up, means gap resolution is also not working.
I see the following in alert log every time I try to restart the log apply, everything seems to be stuck at initialization.
ALTER DATABASE START LOGICAL STANDBY APPLY (test)
with optional part
IMMEDIATE
Attempt to start background Logical Standby process
Tue Feb 14 05:15:28 2012
LSP0 started with pid=28, OS id=23391
Completed: alter database start logical standby apply immediate
LOGMINER: Parameters summary for session# = 1
LOGMINER: Number of processes = 3, Transaction Chunk Size = 201
LOGMINER: Memory Size = 30M, Checkpoint interval = 150M
LOGMINER: SpillScn 0, ResetLogScn 0
-- NOTHING AFTER THIS

Hello;
I noticed some of your parameters seem to be wrong.
fal_client - This is Obsolete in 11.2
You have db_name='test' on the Standby, it should be 'asadmin'
fal_server=test is set like this on the standby, it should be 'asadmin'
I might consider changing VALID_FOR to this :
VALID_FOR=(ONLINE_LOGFILES,ALL_ROLES)Would review 4.2 Step-by-Step Instructions for Creating a Logical Standby Database of Oracle Document E10700-02
Document 278371.1 is showing its age in my humble opinion.
-----Wait on this until you fix your parameters----------------------
Try restarting the SQL Apply
ALTER DATABASE START LOGICAL STANDBY APPLY IMMEDIATEI don't see the parameter MAX_SERVERS, try setting it to 8 times the number of cores.
Use these statements to trouble shoot :
SELECT NAME, VALUE, UNIT FROM V$DATAGUARD_STATS;
SELECT NAME, VALUE FROM V$LOGSTDBY_STATS WHERE NAME LIKE ;TRANSACTIONS%';
SELECT COUNT(1) AS IDLE_PREPARERS FROM V$LOGSTDBY_PROCESS WHERE
TYPE = 'PREPERER' AND STATUS_CODE = 16166;Best Regards
mseberg
Edited by: mseberg on Feb 14, 2012 7:37 AM

Similar Messages

  • DATA Guard Logical Standby v.s. Streams Apply (10gR2, rac -- rac )

    Greetings -
    We currently have a 3 node cluster, doing a 'Schema' capture, with a 2 node cluster serving as the apply side. Both are on 10gR2 (solaris 10)
    We have been experiencing apply latency due to large transactions. The way logminer/streams evaluates the arch logs, it converts each updated row into a transaction as part of a transaction set, using 'decode (' statements.
    I am under the impression a physical standby will do the same thing. But what about a logical Standby ?
    (this from the Oracle Documentation)
    [10g Concepts|http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/concepts.htm#SBYDB00010]
    ...Contains the same logical information as the production database, although the physical organization and structure of the data can be different. The logical standby database is kept synchronized with the primary database though SQL Apply, which transforms the data in the redo received from the primary database into SQL statements and then executing the SQL statements on the standby database.
    Does anyone know firsthand if SQL APPLY re-issues the original sql, or if it relies on the 'decode' command as well ?
    Thanks
    The Nets Edge

    A Physical standby does not do anything with SQL. It is running media recovery under the covers. A Logical standby uses the Logminer to read the redo, converts it to SQL and Data and applies those transactions to the Logical standby.
    The both build logical change records, figure out order, dependencies etc etc and applies the transaction. Both are susceptible to long running transactions on the Primary (source) database.
    Streams uses Logminer, Logical Standby use part of the Streams apply so as Mr. Morgan said, they are very similar :^)
    If you are having apply performance issues you may want to look at Active data Guard.
    Larry

  • Logical standby stuck waiting for dictionary logs

    Hello,
    I created a logical standby (v. 10.2.0.3). It is in state 'WAITING FOR DICTIONARY LOGS', for a few days now.
    Does anyone know how to get past this? It looks like the log that contained the Logminer Dictionary was applied
    already.
    Thanks in advance for any insight you can provide.

    Make sure the archivelogs are reaching the standby. Check your alert logs on both databases to confirm there is no archival problem.

  • Determining the last archive applied on Logical Standby

    Hi,
    I am trying to determine the last log received and applied on my Logical Standby
    SQL> select thread#, max(sequence#) "Last Standby Seq Received"
    from v$archived_log val, v$database vdb
    where val.resetlogs_change# = vdb.resetlogs_change#
    group by thread# order by 1;
       THREAD# Last Standby Seq Received
             1                       14
    SQL> select thread#, max(sequence#) "Last Standby Seq Applied"
    from v$archived_log val, v$database vdb
    where val.resetlogs_change# = vdb.resetlogs_change#
    and applied='YES'
    group by thread# order by 1;
    Does not return anything
    These statements work ok on a physical standby.
    I know Sql Apply is enabled on my logical standby - my broker configuration is enabled and for the logical standby database, it is showing the Intended State as APPLY-ON with no Transport or Apply Lag on any of the databases
    Q. How do I determine the last seq applied on my logical standby ?
    thanks,
    Jim

    Hello;
    I have this from my notes:
    SELECT
      L.SEQUENCE#,
      L.FIRST_TIME,
      (CASE WHEN L.NEXT_CHANGE# < P.READ_SCN THEN 'YES'WHEN L.FIRST_CHANGE# < P.APPLIED_SCN THEN 'CURRENT' ELSE 'NO' END) APPLIED
    FROM
      DBA_LOGSTDBY_LOG L, DBA_LOGSTDBY_PROGRESS P
    ORDER BY SEQUENCE#;
    Best Regards
    mseberg

  • Logical Standby table not supported 11.2.0.1

    I have a table with CLOB in primary and runs the SQL to determine which tables are unsupported.
    They, nor I, can see why this table is unsupported. Please let me know what I am missing.
    Are there additional steps to see why?
    Please also see the DDL for the table at the end.
    SQL> SELECT COLUMN_NAME,DATA_TYPE FROM DBA_LOGSTDBY_UNSUPPORTED WHERE OWNER='P_RPMX_AUDIT_DB' AND TABLE_NAME = 'AUDIT_TAB' ;
    COLUMN_NAME DATA_TYPE
    TABLE_SCRIPT CLOB
    SQL> SELECT OWNER FROM DBA_LOGSTDBY_SKIP WHERE STATEMENT_OPT = 'INTERNAL SCHEMA';
    OWNER
    DBSNMP
    SYS
    SYSTEM
    WMSYS
    ORDDATA
    OUTLN
    DIP
    EXFSYS
    XDB
    ORDPLUGINS
    ANONYMOUS
    APPQOSSYS
    ORDSYS
    SI_INFORMTN_SCHEMA
    XS$NULL
    15 rows selected.
    SQL> SELECT DISTINCT OWNER,TABLE_NAME FROM DBA_LOGSTDBY_UNSUPPORTED ORDER BY OWNER,TABLE_NAME;
    OWNER TABLE_NAME
    APG APG_REJECT_TAB
    APG ERR$_ACCOUNT
    APG ERR$_ACCOUNT_ADDRESS
    APG ERR$_ACCOUNT_ATTRIBUTES
    APG ERR$_ACCOUNT_CYCLE_HISTORY
    APG ERR$_ACCOUNT_USERS
    APG ERR$_ACCOUNT_XREFERENCE
    APG ERR$_ACTIONABLE_ITEMS
    APG ERR$_IDENTIFICATION
    APG ERR$_LOAD_RECON_METRICS
    APG ERR$_PHONE
    APG ERR$_POOL
    APG ERR$_POOL_ACCOUNTS
    APG ERR$_POOL_ACCOUNT_ENROLL_HISTO
    APG ERR$_USERS
    P_IM_EXTRACT_MASK ERR$_REWARDS_ACTIVITY
    P_IM_EXTRACT_MASK ERR$_REWARDS_TRANSACTION
    P_IM_EXTRACT_MASK ERR$_REWARDS_TRANSACTION_ITEM
    P_IM_EXTRACT_MASK ERR$_TRANSACTION
    P_RPMX_AUDIT_DB AUDIT_TAB
    P_RPMX_JOBS ERR$_ACCOUNT
    P_RPMX_JOBS ERR$_ACCOUNT_XREFERENCE
    22 rows selected.
    CREATE TABLE P_RPMX_AUDIT_DB.AUDIT_TAB
    TABLE_SCRIPT CLOB
    LOB (TABLE_SCRIPT) STORE AS (
    TABLESPACE P_RPMX_AUD_DB_DATA
    ENABLE STORAGE IN ROW
    CHUNK 8192
    RETENTION
    NOCACHE
    LOGGING
    STORAGE (
    INITIAL 1M
    NEXT 1M
    MINEXTENTS 1
    MAXEXTENTS UNLIMITED
    PCTINCREASE 0
    BUFFER_POOL DEFAULT
    FLASH_CACHE DEFAULT
    CELL_FLASH_CACHE DEFAULT
    TABLESPACE P_RPMX_AUD_DB_DATA
    RESULT_CACHE (MODE DEFAULT)
    PCTUSED 0
    PCTFREE 10
    INITRANS 1
    MAXTRANS 255
    STORAGE (
    INITIAL 1M
    NEXT 1M
    MINEXTENTS 1
    MAXEXTENTS UNLIMITED
    PCTINCREASE 0
    BUFFER_POOL DEFAULT
    FLASH_CACHE DEFAULT
    CELL_FLASH_CACHE DEFAULT
    LOGGING
    NOCOMPRESS
    NOCACHE
    NOPARALLEL
    MONITORING;
    Thanks, Chris

    Hi I was able to find the following answer. Thanks
    "Logical standby has never supported tables that only contain LOBs. We require some scalar column that can be used for row identification during update and delete processing.
    There is some discussion of row identification issues in section 4.1.2 of Oracle Data Guard Concepts and Administration. "
    http://docs.oracle.com/cd/E11882_01/server.112/e25608/create_ls.htm#i77026
    "If there is no primary key and no nonnull unique constraint/index, then all columns of bounded size are logged as part of the UPDATE statement to identify the modified row. All columns are logged except the following: LONG, LOB, LONG RAW, object type, and collections."
    So you see with only a LOB column it cannot identify the row. You need a primary key or at least some columns that can uniquely identify the row. The one LOB column is no logged with the supplemental logging.
    So it is not supported.

  • How to skip delete on logical standby

    Dear guys and gurus,
    On next month i will have plan to delete history data on my Primary database, but i still want to keep them in logical standby as the data warehouse. So, how can i skip delete statement from primary database apply to logical standby.
    Oracle SQL apply only support for Skip DML but not for keep delete statemen (on 10g and 11g)
    SQL>exec dbms_logstdby.skip(statement => 'DELETE',schema_name => 'SCOTT', object_name => 'DEPT');
    BEGIN dbms_logstdby.skip(statement => 'DELETE',schema_name => 'SCOTT', object_name => 'DEPT'); END;
    ERROR at line 1:
    ORA-06550: line 1, column 7:
    PLS-00306: wrong number or types of arguments in call to 'SKIP'
    ORA-06550: line 1, column 7:
    PL/SQL: Statement ignored
    I do search and see one post about 3 years age about this question but still not answered.So, do you have any idea for this ???
    So thanks and best regards.
    Ch

    You might want to play with this but here is an idea.
    CREATE OR REPLACE PROCEDURE schema.handle_delete (
    old_stmt IN VARCHAR2,
    stmt_type IN VARCHAR2,
    schema IN VARCHAR2,
    name IN VARCHAR2,
    xidusn IN NUMBER,
    xidslt IN NUMBER,
    xidsqn IN NUMBER,
    error IN VARCHAR2,
    new_stmt OUT VARCHAR2
    ) AS
    BEGIN
    -- Default to what we already have
    new_stmt := old_stmt;
    -- Ignore any GRANT errors on SYS or HR schemas
    IF INSTR(UPPER(old_stmt),'DELETE') > 0
    THEN
    IF schema IS NULL
    OR (schema IS NOT NULL AND
    (UPPER(schema) = '<SCHEMA>' OR UPPER(SCHEMA) = '<SCHEMA>' )
    THEN
    new_stmt := NULL;
    -- record the fact that we just skipped
    END IF;
    END IF;
    END handle_delete;
    Register the error handler with SQL Apply:
    SQL> EXECUTE DBMS_LOGSTDBY.SKIP_ERROR ( -
    statement => 'DML', -
    schema_name => NULL, -
    object_name => NULL, -
    proc_name => 'SYS.HANDLE_DELETE');

  • Doubt regarding switchover from logical standby to primary

    I am using Oracle 10g databases on Linux environment
    I was trying creating a data guard configuration and once I was successfully done I created a logical standby database from the physical standby . Then I was trying to do a switchover .
    By mistake I gave the prepare switchover command in the logical standby before executing it in the primary . And then when I gave it in the primary both the status turned to be "PREPARING SWITCHOVER"
    In logical standby
    SQL> ALTER DATABASE PREPARE TO SWITCHOVER TO PRIMARY;
    Database altered.
    SQL>  SELECT SWITCHOVER_STATUS FROM V$DATABASE;
    SWITCHOVER_STATUS
    PREPARING SWITCHOVER
    In primary
    SQL> ALTER DATABASE PREPARE TO SWITCHOVER TO LOGICAL STANDBY;
    Database altered.
    SQL> select switchover_status from v$database;
    SWITCHOVER_STATUS
    PREPARING SWITCHOVERNow I am not able to commit any of them to primary or logical standby
    In primary
    SQL> ALTER DATABASE COMMIT TO SWITCHOVER TO LOGICAL STANDBY;
    ALTER DATABASE COMMIT TO SWITCHOVER TO LOGICAL STANDBY
    ERROR at line 1:
    ORA-16217: prepare to switchover has not completed
    ORA-06512: at line 1
    In logical standby
    SQL> ALTER DATABASE COMMIT TO SWITCHOVER TO PRIMARY;
    ALTER DATABASE commit  TO SWITCHOVER TO PRIMARY
    ERROR at line 1:
    ORA-16109: failed to apply log data from previous primaryLet me know what should be done now ? Is there anyway to solve this issue ?

    Can you try using the following first on Primary and then on Standby?
    ALTER DATABASE PREPARE TO SWITCHOVER CANCEL;

  • Logical standby versus Streams?

    Anybody can explain me difference between logical standby and streams in terms of architecture and functionality?

    Both Logical standby and streams share the same underlying architecture. The difference is the requirements u have
    1) the Logical standby database is the complete replica of the whole database. while streams enable you to replicate table, schema or the database.
    2) The logical standby database can opened in read only mode. ( by the way u can open it in read/write but you have to stop the recovery first and u have flash back database enabled). Streams allow you to open the replicated database read/write mode.
    3) The logical standby database allows one way replication. with streams you can implement bi-directional replication as well.
    4) Streams are far more customizable than logical data guard.
    5) in my opinion streams are much harder to implement, administer and monitor than logical standby database

  • Creating Logical Standby, DBMS_LOGSTDBY.BUILD; Hangs on Primary..

    My two node RAC database Primary hangs upon executing SQL> exec DBMS_LOGSTDBY.BUILD; when creating a Logical Standby. I have stopped all the recovery processes on the Physical Standby. I have also taken down the instances on the RAC apart from the one I'm working on. Additionally not much info in the alert log...
    Logminer Bld: Build started
    Wed Jan 23 06:01:57 2013
    ALTER SYSTEM SWITCH ALL LOGFILE start (UDSRED1)
    Wed Jan 23 06:01:57 2013
    ALTER SYSTEM SWITCH ALL LOGFILE complete (UDSRED1)
    Wed Jan 23 06:01:57 2013
    Thread 1 advanced to log sequence 2323 (LGWR switch)
    Current log# 2 seq# 2323 mem# 0: /data03/oracle/oradata/UDSRED/redo02.rdo
    Wed Jan 23 06:01:57 2013
    Wed Jan 23 06:01:57 2013
    Logminer Bld: Lockdown Complete. DB_TXN_SCN is 2548 1551493736
    Wed Jan 23 06:01:57 2013
    LNS: Standby redo logfile selected for thread 1 sequence 2323 for destination LOG_ARCHIVE_DEST_2
    Wed Jan 23 06:01:57 2013
    LNS: Standby redo logfile selected for thread 1 sequence 2323 for destination LOG_ARCHIVE_DEST_3
    Wed Jan 23 06:01:59 2013
    ALTER SYSTEM ARCHIVE LOG
    Wed Jan 23 06:01:59 2013
    Thread 1 advanced to log sequence 2324 (LGWR switch)
    Current log# 3 seq# 2324 mem# 0: /data03/oracle/oradata/UDSRED/redo03.rdo
    Wed Jan 23 06:01:59 2013
    LNS: Standby redo logfile selected for thread 1 sequence 2324 for destination LOG_ARCHIVE_DEST_3
    Wed Jan 23 06:01:59 2013
    LNS: Standby redo logfile selected for thread 1 sequence 2324 for destination LOG_ARCHIVE_DEST_2
    Wed Jan 23 06:02:00 2013
    Wed Jan 23 06:02:00 2013
    Logminer Bld: Done
    Edited by: Imran on Jan 23, 2013 2:10 PM

    Imran,
    The information from alert log file is not sufficient.
    Can you log switch on primary and monitor whether it is archiving in standby or not.
    and you can monitor the changes in "v$managed_standby"
    Also check for the errors using "v$dataguard_status"
    SQL> select error_cdoe, message, timestamp from v$dataguard_status where dest_id=<n>;

  • Transportable tablespaces with Logical Standby

    Does anyone know whether or not transportable tablespaces can be used in a Logical Standby environment? I know they can be used with a Physical Standby, and the documentation covers how to do this, but there's nothing about Logical Standby. This is Oracle 10.2.0.4/SLES 10.
    Thanks.

    Using transportable tablespaces to a logical standby environment. Actually, there is a paragraph in the 10.2 Data Guard Concepts and Administration manual. I will summarize:
    After creating the transportable tablespace export file on the source system, copy the datafiles to both the primary and logical standby systems.
    On the dataguard logical standby, within SQLplus, issue SQL> alter session disable guard;
    Import the tablespaces into the logical standby database
    SQL> alter session enable guard;
    Import the tablespaces into the primary database.
    In 10.1, we imported the tablespaces into an environment with a physical standby and then converted the physical standby to a logical standby.
    I tested the procedure for 10.2 and it seemed to work. Has anyone else done this and if so, are you aware of any issues? Like I said, the documentation is one very brief paragraph.
    Thanks.

  • Webintelligenceprocessingserver stucks in "initializing" status

    Hello everybody!
    I have serious problem whith Web Intelligence Processing Server. Today after rebooting SAP BO server it stucked in "initializing" status.
    I've tried recreating webi server, tried different setting for it (Cannot initialize report engine server). Nothing, every new server also can't starts.
    Please help.
    By the way, where can i find Web Intelligence Processing Server logs, if they exist?
    Thanks!

    I've solved this issue.
    The problem was in extremly large amount of audit files in (Drive)\Install Directory\SAP BusinessObjects\SAP BusinessObjects Enterprise XI 12.0\Auditing\.
    After deleting all of these files and desabling audit for WebIntelligenceProcessingServer this service starts fine.
    Some helpfull notes:
    1919999 - Web Intelligence Processing Server stuck in Initializing status
    1494758 - Web Intelligence Processing Server does not start , remains in Failed mode

  • Logical Standby SQL Apply Using Incorrect Decode Statement

    We are seeing statements erroring out on our logical standby that have been rewritten (presumably by sql apply) with decode statements that don't appear to be correct. For example, here is one of the rewritten statements.
    update /*+ streams restrict_all_ref_cons */ "CADPROD"."OMS_SQL_STATEMENT" p
    set *"APPLICATION"=decode(:1,'N',"APPLICATION",:2)*,
    "STATEMENT"=dbms_reputil2.get_final_lob(:3,"STATEMENT",:4)
    where (:5='N' or(1=1 and (:6='N' or(dbms_lob.compare(:7,"STATEMENT")=0)or(:7 is null and "STATEMENT" is null)))) and(:8="APPLICATION")
    The problem comes in, we believe, with the attempt to write the value "APPLICATION" to the application column which is only a 10 character field. the value for the :1 bind variable is "N" and the value for :2 is null.
    We see the following error on the logical standby:
    ORA-00600: internal error code, arguments: [kgh_heap_sizes:ds], [0x01FCDBE60], [], [], [], [], [], []
    ORA-07445: exception encountered: core dump [ACCESS_VIOLATION] [kxtoedu+54] [PC:0x2542308] [ADDR:0xFFFFFFFFFFFFFFFF] [UNABLE_TO_READ] []
    ORA-12899: value too large for column "CADPROD"."OMS_SQL_STATEMENT"."APPLICATION" (actual: 19576, maximum: 10)
    Is this a configuration issue or is it normal for SQL Apply to convert statements from logminer into decode statements?
    We have an Oracle 10.2.0.4 database running on windows 2003 R2 64-bit os. We have 3 physical and 2 logical standby's, no problems on the physical standbys.

    Hello;
    I noticed some of your parameters seem to be wrong.
    fal_client - This is Obsolete in 11.2
    You have db_name='test' on the Standby, it should be 'asadmin'
    fal_server=test is set like this on the standby, it should be 'asadmin'
    I might consider changing VALID_FOR to this :
    VALID_FOR=(ONLINE_LOGFILES,ALL_ROLES)Would review 4.2 Step-by-Step Instructions for Creating a Logical Standby Database of Oracle Document E10700-02
    Document 278371.1 is showing its age in my humble opinion.
    -----Wait on this until you fix your parameters----------------------
    Try restarting the SQL Apply
    ALTER DATABASE START LOGICAL STANDBY APPLY IMMEDIATEI don't see the parameter MAX_SERVERS, try setting it to 8 times the number of cores.
    Use these statements to trouble shoot :
    SELECT NAME, VALUE, UNIT FROM V$DATAGUARD_STATS;
    SELECT NAME, VALUE FROM V$LOGSTDBY_STATS WHERE NAME LIKE ;TRANSACTIONS%';
    SELECT COUNT(1) AS IDLE_PREPARERS FROM V$LOGSTDBY_PROCESS WHERE
    TYPE = 'PREPERER' AND STATUS_CODE = 16166;Best Regards
    mseberg
    Edited by: mseberg on Feb 14, 2012 7:37 AM

  • SQL apply is very slow on Logical Standby..!!

    Hello all,
    We are having Data Guard setup in our environment where we are having Primary, Physical Standby as well as logical standby databases..
    DB Version : 10.2.0.1 in all databases (Pri, Phy and Logical)
    OS : RHEL4
    Only Oracle is running on this Box..
    Since last month we are facing problems in Logical Standby database where it seems SQL apply has become very slow..
    Archive log files are successfully transferring from Primary but since SQL apply has become slow logical standby is lagging behind primary by two days..
    How do i speed up this SQL apply..?? Any ideas and suggestions are most welcome..
    I checked TOP command to find out which oracle processes are consuming maximum CPU and i have found ora_p000_oracle, ora_p001_oracle, ora_p002_oracle, ora_p003_oracle, ora_p004_oracle, ora_p005_oracle, etc processes are consuming highest CPU and Load Average has always been above 1..
    Any help would be greatly appreciated..
    Thanks - HP

    Hello;
    These Oracle notes might help :
    Slow Performance In Logical Standby Database Due To Lots Of Activity On Sys.Aud$ [ID 862173.1]
    Oracle10g Data Guard SQL Apply Troubleshooting [ID 312434.1]
    Developer and DBA Tips to Optimize SQL Apply [ID 603361.1]
    Best Regards
    mseberg

  • How to monitor SQL Apply for 10.2.0.3 logical standby database

    We have a logical standby database setup for reporting purposes. Users want to monitor whether sql apply is working or failed closely as it has reporting repercations.
    In case of 9i databases , there was "Data Not Applied (logs)" metric which we used for alerting and paging, in case a backlog of more than 5 log files developed.
    With 10.2.0.3 onwards, the metric no more exists.
    I would like to learn from other, how to monitor the setup, so that if the backlog in logs shipping or applying develops, we get page.
    Regards.

    regather the statistics on the table with method_opt=>for all columns or for all indexed columns or whatever size 1
    The 'size 1' directive will remove the histogram statistics.
    Sorry, didn't read ur post in a hurry. Below article (http://www.freelists.org/post/oracle-l/Any-quick-way-to-remove-histograms,13) removes histogram without re-analyzing the table. Hope that helps!!!
    On 3/16/07, Wolfgang Breitling <breitliw@xxxxxxxxxxxxx> wrote:
    I also did a quick check and just using
    exec
    dbms_stats.set_column_stats(user,'table_name',colname=>'column_name',d
    istcnt=>
    <num_distinct>);
    will remove the histogram without removing the low_value andhigh_value.
    At 01:40 PM 3/16/2007, Alberto Dell'Era wrote:
    On 3/16/07, Allen, Brandon <Brandon.Allen@xxxxxxxxxxx> wrote:
    Is there any faster way to remove histograms other than
    re-analyzing
    >
    the table? I want to keep the existing table, index & columnstats,
    >
    but with only 1 bucket (i.e. no histograms).You might try the attached script, that reads the stats using
    dbms_stats.get_column_stats and re-sets them, minus the histogram,
    using dbms_stats.set_column_stats.
    I haven't fully tested it - it's only 10 minutes old, even if Ihave
    slightly modified for you another script I've used for quite some
    time - and the spool on 10.2.0.3 seems to confirmthat the histogram
    is, indeed, removed, while all the other statistics are preserved.I
    have also reset density to 1/num_distinct, that is the value youget
    if no histogram is collected.regards,
    naren
    Edited by: fuzzydba on Oct 25, 2010 10:52 AM

  • Sql Apply issue in logical standby database--(10.2.0.5.0) x86 platform

    Hi Friends,
    I am getting the following exception in logical standby database at the time of Sql Apply.
    After run the command alter database start logical standby apply sql apply services start but after few second automatically stop and getting following exception.
    alter database start logical standby apply
    Tue May 17 06:42:00 2011
    No optional part
    Attempt to start background Logical Standby process
    LOGSTDBY Parameter: MAX_SERVERS = 20
    LOGSTDBY Parameter: MAX_SGA = 100
    LOGSTDBY Parameter: APPLY_SERVERS = 10
    LSP0 started with pid=30, OS id=4988
    Tue May 17 06:42:00 2011
    Completed: alter database start logical standby apply
    Tue May 17 06:42:00 2011
    LOGSTDBY status: ORA-16111: log mining and apply setting up
    Tue May 17 06:42:00 2011
    LOGMINER: Parameters summary for session# = 1
    LOGMINER: Number of processes = 4, Transaction Chunk Size = 201
    LOGMINER: Memory Size = 100M, Checkpoint interval = 500M
    Tue May 17 06:42:00 2011
    LOGMINER: krvxpsr summary for session# = 1
    LOGMINER: StartScn: 0 (0x0000.00000000)
    LOGMINER: EndScn: 0 (0x0000.00000000)
    LOGMINER: HighConsumedScn: 2660033 (0x0000.002896c1)
    LOGMINER: session_flag 0x1
    LOGMINER: session# = 1, preparer process P002 started with pid=35 OS id=4244
    LOGSTDBY Apply process P014 started with pid=47 OS id=5456
    LOGSTDBY Apply process P010 started with pid=43 OS id=6484
    LOGMINER: session# = 1, reader process P000 started with pid=33 OS id=4732
    Tue May 17 06:42:01 2011
    LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1417, X:\TANVI\ARCHIVE2\ARC01417_0748170313.001
    Tue May 17 06:42:01 2011
    LOGMINER: Turning ON Log Auto Delete
    Tue May 17 06:42:01 2011
    LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01417_0748170313.001
    Tue May 17 06:42:01 2011
    LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1418, X:\TANVI\ARCHIVE2\ARC01418_0748170313.001
    LOGSTDBY Apply process P008 started with pid=41 OS id=4740
    LOGSTDBY Apply process P013 started with pid=46 OS id=7864
    LOGSTDBY Apply process P006 started with pid=39 OS id=5500
    LOGMINER: session# = 1, builder process P001 started with pid=34 OS id=4796
    Tue May 17 06:42:02 2011
    LOGMINER: skipped redo. Thread 1, RBA 0x00058a.00000950.0010, nCV 6
    LOGMINER: op 4.1 (Control File)
    Tue May 17 06:42:02 2011
    LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01418_0748170313.001
    Tue May 17 06:42:03 2011
    LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1419, X:\TANVI\ARCHIVE2\ARC01419_0748170313.001
    Tue May 17 06:42:03 2011
    LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01419_0748170313.001
    Tue May 17 06:42:03 2011
    LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1420, X:\TANVI\ARCHIVE2\ARC01420_0748170313.001
    Tue May 17 06:42:03 2011
    LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01420_0748170313.001
    Tue May 17 06:42:03 2011
    LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1421, X:\TANVI\ARCHIVE2\ARC01421_0748170313.001
    LOGSTDBY Analyzer process P004 started with pid=37 OS id=5096
    Tue May 17 06:42:03 2011
    LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01421_0748170313.001
    LOGSTDBY Apply process P007 started with pid=40 OS id=2760
    Tue May 17 06:42:03 2011
    Errors in file x:\oracle\product\10.2.0\admin\tanvi\bdump\tanvi_p001_4796.trc:
    ORA-00600: internal error code, arguments: [krvxbpx20], [1], [1418], [2380], [16], [], [], []
    LOGSTDBY Apply process P012 started with pid=45 OS id=7152
    Tue May 17 06:42:03 2011
    LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1422, X:\TANVI\ARCHIVE2\ARC01422_0748170313.001
    Tue May 17 06:42:03 2011
    LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01422_0748170313.001
    Tue May 17 06:42:03 2011
    LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1423, X:\TANVI\ARCHIVE2\ARC01423_0748170313.001
    Tue May 17 06:42:03 2011
    LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01423_0748170313.001
    Tue May 17 06:42:03 2011
    LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1424, X:\TANVI\ARCHIVE2\ARC01424_0748170313.001
    LOGMINER: session# = 1, preparer process P003 started with pid=36 OS id=5468
    Tue May 17 06:42:03 2011
    LOGMINER: End mining logfile: X:\TANVI\ARCHIVE2\ARC01424_0748170313.001
    Tue May 17 06:42:04 2011
    LOGMINER: Begin mining logfile for session 1 thread 1 sequence 1425, X:\TANVI\ARCHIVE2\ARC01425_0748170313.001
    LOGSTDBY Apply process P011 started with pid=44 OS id=6816
    LOGSTDBY Apply process P005 started with pid=38 OS id=5792
    LOGSTDBY Apply process P009 started with pid=42 OS id=752
    Tue May 17 06:42:05 2011
    krvxerpt: Errors detected in process 34, role builder.
    Tue May 17 06:42:05 2011
    krvxmrs: Leaving by exception: 600
    Tue May 17 06:42:05 2011
    Errors in file x:\oracle\product\10.2.0\admin\tanvi\bdump\tanvi_p001_4796.trc:
    ORA-00600: internal error code, arguments: [krvxbpx20], [1], [1418], [2380], [16], [], [], []
    LOGSTDBY status: ORA-00600: internal error code, arguments: [krvxbpx20], [1], [1418], [2380], [16], [], [], []
    Tue May 17 06:42:06 2011
    Errors in file x:\oracle\product\10.2.0\admin\tanvi\bdump\tanvi_lsp0_4988.trc:
    ORA-12801: error signaled in parallel query server P001
    ORA-00600: internal error code, arguments: [krvxbpx20], [1], [1418], [2380], [16], [], [], []
    Tue May 17 06:42:06 2011
    LogMiner process death detected
    Tue May 17 06:42:06 2011
    logminer process death detected, exiting logical standby
    LOGSTDBY Analyzer process P004 pid=37 OS id=5096 stopped
    LOGSTDBY Apply process P010 pid=43 OS id=6484 stopped
    LOGSTDBY Apply process P008 pid=41 OS id=4740 stopped
    LOGSTDBY Apply process P012 pid=45 OS id=7152 stopped
    LOGSTDBY Apply process P014 pid=47 OS id=5456 stopped
    LOGSTDBY Apply process P005 pid=38 OS id=5792 stopped
    LOGSTDBY Apply process P006 pid=39 OS id=5500 stopped
    LOGSTDBY Apply process P007 pid=40 OS id=2760 stopped
    LOGSTDBY Apply process P011 pid=44 OS id=6816 stopped
    Tue May 17 06:42:10 2011

    Errors in file x:\oracle\product\10.2.0\admin\tanvi\bdump\tanvi_p001_4796.trc:
    ORA-00600: internal error code, arguments: [krvxbpx20], [1], [1418], [2380], [16], [], [], []submit an SR to ORACLE SUPPORT.
    refer these too
    *ORA-600/ORA-7445 Error Look-up Tool [ID 153788.1]*
    *Bug 6022014: ORA-600 [KRVXBPX20] ON LOGICAL STANDBY*

Maybe you are looking for

  • How do I transfer a gift card balance to family members in family share

    Ok well I just realized that the account associated with my Family share will only access my credit card and not my gift card balance. I had two $100 gift cards which I redeemed into my account thinking we could all share the gift card.  Little did I

  • Meta Data JCo Error

    Hi All, I am Created Application JCo's it is successfully tested then i created Metadata JCo then i tested and i got the following Error.  com.sap.mw.jco.JCO$Exception: (102) RFC_ERROR_COMMUNICATION: Connect to message server host failed Connect_PM 

  • Changing primary user Name

    My ex husband originally registered us for Broadband so the primary email is his. However we are now divorced and I have been paying the bill and the contract has been in my name for over a year now. However he gets alerts when I approach my usage li

  • Sales order condition type

    Hi, Does anyone know what needs to be maintained where to allow a user to change the CnTy (condition type column) in a sales order at the line item? Thank you

  • FRM-13008 after conversion from 6i

    First, let me indicate that I am dumb as dirt about Forms, IAS, Java. I am a DBA but have next to zero experience in this area. In a test environment, I have converted all application 6i Forms & Reports to 10G. I have only one problem form to convert