EM: Creation process for a physical standby DB failed with SQL Exception

Version: EM Oracle 10.2.5 (agents running, repository running, primary db running, all targets visible and reachable with EM)
I try to create a physical standby database with the enterprise manager and each time the process is aborted with a SQL exception during the preparation of the job by the EM. I have added a part of the OMs log containing the error at the end of the excerpt.
=============
2010-04-29 16:00:39,856 [EMUI_16_00_39_/console/targets] WARN pref.SubtabPref getFolders.710 - Unknown folder id: VirtualServers retrieved from repository
2010-04-29 16:01:04,765 [EMUI_16_01_04_/console/database/dataguard/create] ERROR em.dataguard validate.1329 - CreateBean: ClassNotFoundException: null
2010-04-29 16:02:05,476 [EMUI_16_02_05_/console/database/dataguard/create] ERROR jobs.dbclone checkSetFileError.79 - DatabaseFileAttributes.checkSetFileError(): Null database file!
2010-04-29 16:02:05,476 [EMUI_16_02_05_/console/database/dataguard/create] ERROR jobs.dbclone setControlfiles.160 - DatabaseFileAttributes.setDatafiles(): Invalid control file!
2010-04-29 16:02:05,492 [EMUI_16_02_05_/console/database/dataguard/create] ERROR jobs.dbclone getControlFileNames.616 - DatabaseFileAttributes.getDatafileNames(): null datafile names!
2010-04-29 16:02:32,823 [Thread-28] ERROR em.jobs remoteOp.2389 - DBVerify.remoteOp(): Error: max_stamp# 6071384
2010-04-29 16:02:32,823 [Thread-28] ERROR jobs.dbclone submitJobPreparation.3297 - DBCloneObject.submitJobPreparation(): getMaxLogSequenceNum: Während der Vorbereitung des Jobs ist eine SQL Exception aufgetreten. Um das Problem zu diagnostizieren, legen Sie das Agent Perl-Skript-Tracing auf DEBUG fest und wiederholen den Vorgang
2010-04-29 16:02:32,823 [Thread-28] ERROR jobs.dbclone submitJobPreparation.3501 - DBCloneObject.submitJobPreparation(): Exception: java.lang.Exception: Während der Vorbereitung des Jobs ist eine SQL Exception aufgetreten. Um das Problem zu diagnostizieren, legen Sie das Agent Perl-Skript-Tracing auf DEBUG fest und wiederholen den Vorgang
2010-04-29 16:02:32,823 [Thread-28] ERROR jobs.dbclone submitDBCloneJob.3716 - DBCloneObject.submitDBCloneJob(): Exception: Während der Vorbereitung des Jobs ist eine SQL Exception aufgetreten. Um das Problem zu diagnostizieren, legen Sie das Agent Perl-Skript-Tracing auf DEBUG fest und wiederholen den Vorgang
2010-04-29 16:02:37,496 [EMUI_16_02_37_/console/database/dataguard/create] ERROR em.dataguard onEvent.1243 - CreateConfigController: Exception: oracle.sysman.db.dg.util.VxxStandbyException: Während der Vorbereitung des Jobs ist eine SQL Exception aufgetreten. Um das Problem zu diagnostizieren, legen Sie das Agent Perl-Skript-Tracing auf DEBUG fest und wiederholen den Vorgang
=========
I have set the agent perl script tracing to DEBUG, but can't find any reason, why the job preparation failed.
Has anyone an idea why the job cannot be prepared? Thanks in advance for investigation :-)

gromit wrote:
Is there a difference if:
I install EM Version 10.2.0.1/10.2.0.2 -> upgrade it to 10.2.0.3 -> upgrade it to 10.2.0.5
OR
I install EM Version 10.2.0.1/10.2.0.2 and directly upgrade it to 10.2.0.5 ???Hi,
Both are same. Result will be same. Finally you are reaching 10.2.0.5 version..
But I would suggest to choose 2nd option.. i.e. install EM Version 10.2.0.1/10.2.0.2 and directly upgrade it to 10.2.0.5.
Because it will save your time..
Regards
Rajesh

Similar Messages

  • CSSCAN for database character set conversion failing with ORA-01578

    Hi ,
    CSSCAN for database character set conversion failing with ORA-01578: ORACLE data block corrupted (file # 84, block # 23930). please help me out in this regard.
    Thanks,
    Sravan.

    Hi Anand,
    Thanks for your update. The segment is a table not an index in my case. And i got this error while running CSSCAN on Apps database for character set conversion to UTF8 from WE8ISO8859P1. Please find the snapshot below for your reference.
    SQL> select segment_name, segment_type, owner from dba_extents where file_id = 84 and 23930 between block_id and block_id + blocks - 1;
    SEGMENT_NAME
    SEGMENT_TYPE OWNER
    EDW_LOOKUP_M
    TABLE POA
    SQL> ANALYZE TABLE POA.EDW_LOOKUP_M VALIDATE STRUCTURE CASCADE;
    ANALYZE TABLE POA.EDW_LOOKUP_M VALIDATE STRUCTURE CASCADE
    ERROR at line 1:
    ORA-01578: ORACLE data block corrupted (file # 84, block # 23930)
    ORA-01110: data file 84: '/d911/oracle/dbcondata/poad01.dbf'
    Thanks,
    Sravan.

  • Tns entry on client machine for a physical standby!

    Hi all.
    i am working on 10gR2 on RHEL4.I have created a physical standby on the same host. The primary and the standby database has different sid as prod and prodsb
    can i have one tns entry on the client machine so that they can connect to the database even after teh switchover or failover??? or will they have to change the connection string while connecting to the database??

    Hi, you can use Oracle Transparent Failover, please review the next link for get more information.
    http://download.oracle.com/docs/cd/B19306_01/network.102/b14212/advcfg.htm#sthref1278
    Please let me know if you have some doubt or comment about this.
    Regards.

  • Physical standby database fail-over

    Hi,
    I am working on Oracle 10.2.0.3 on Solaris SPARC 64-bit.
    I have a Dataguard configuration with a single Physical standby database that uses real time application. We had a major application upgrade yesterday and before the start of upgrade, we cancelled the media recovery and disabled the log_archive_dest_n so that it doesn't ship the archive logs to standby site. We left the dataguard configuration in this mode incase of a rollback.
    Primary:
    alter system set log_archive_dest_state_2='DEFER';
    alter system switch logfile;
    Standby:
    alter database recover managed standby database cancel;Due to application upgrade induced problems we had to failover to the physical standby, which was not in sync with primary from yesterday. I used the following method to fail-over since i do not want to apply any redo from yesterday.
    Standby:
    alter database activate physical standby database;
    alter database open;
    shutdown immediate;
    startupSo, after this step, the database was a stand alone database, which doesn't have any standby databases yet (but it still has log_archive_config parameter set and log_archive_dest_n parameters set but i have 'DEFER' the log_archive_dest_n pointing to the old primary). I have even changed the "archive log deletion policy to NONE"
    RMAN> configure archivelog deletion policy to none;After the fail-over was completed, the log sequence started from Sequence 1. We cleared the FRA to make space for the new archive logs and started off a FULL database backup (backup incremental level 0 database plus archivelog delete input). The backup succeded but we got these alerts in the backup log that RMAN cannot delete the archivelogs.
    RMAN-08137: WARNING: archive log not deleted as it is still neededMy question here is
    1) Even though i have disabled the log_archive_dest_n parameters, why is RMAN not able to delete the archivelogs after backup when there is no standby database for this failed-over database?
    2) Are all the old backups marked unusable after a fail-over is performed?
    FYI... flashback database was not used in this case as it did not server our purpose.
    Any information or documentation links would be greatly appreciated.
    Thanks,
    Harris.

    Thanks for the reply.
    The FINISH FORCE works in some cases but if there is an archive gap (though it didn't report in our case), it might not work some times (DOCID: 846087.1). So, we followed the Switch-over & Fail-Over best practices where it mentioned about this "ACTIVE PHYSICAL STANDBY" for a fail-over if you intend not to apply any archivelogs. The process we followed is the Right one.
    Anyhow, we got the issue resolved. Below is the resolution path.
    1) Even though if you DEFER the LOG_ARCHIVE_DEST_STATE_N parameter's on the primary, there are some situations where the Primary database in a dataguard configuration where it will not delete the archive logs due to some SCN issues. This issue may or may not arise in all fail-over scenarios. If it does, then do the following checks
    Follow DOCID: 803635.1, which talks about a PLSQL procedure to check for problematic SCN's in a dataguard configuration even though the physical standby databases are no available (i.e., if the dataguard parameters are set, log_archive_config, log_archive_dest_n='SERVICE=..." still set and even though corresponding LOG_ARCHIVE_DEST_STATE_N parameters are DEFERRED).
    If this procedure returns any rows, then the primary database is not able to delete the archivelogs because it is still thinking there is a standby database and trying to save the archive logs because of the SCN conflict.
    So, the best thing to do is, remove the DG related parameters from the spfile (log_archive_config, log_archive_dest_n parameters).
    After i made these changes, i ran a test backup using "backup archivelog all delete input", the archive logs got deleted after backup without any issues.
    Thanks,
    Harris.
    Edited by: user11971589 on Nov 18, 2010 2:55 PM

  • Attribute Change Run in Process Chain failed with SQL 942 error

    Hi,
    Attribute change run in th eprocess chain failed with the message " SQL Error 942" and ran succesfully after repeating the step.Searched in Forums for similar error but the solutions were not apt.
    Any pointers would be of great help!!
    Thanks in advance,
    Bhagyarekha.

    Hi Bhagyarekha,
    The reason behind that error could be:
    An attempt is made to activate a characteristic whose compounding was replaced and whose navigation attributes were simultaneously deleted. The replacement of the compounding temporarily deletes the M view of the characteristic during internal activation. The process of deleting navigation attributes involves converting the attribute SID tables. However, the M-view of the characteristic, which has just been deleted, is needed for this conversion.
    Check this SAP Note:
    Note 684405 - Replacing the compounding and deleting navigation attr.
    Hope it helps.
    Veerendra.

  • Help for migrating DB2 to Oracle11 g, with Sql Developer

    Hello, Good Afternoon,
    Today I started a series of exercises to migrate DB2 9 to Oracle 11 g, with SQL Developer in order to evaluate this tool for a project we have in door.
    But it is setting a number of issues, because the DB2 tables that we are trying to migrate do not have primary key and load the data sets the error ORA-02270.
    It is generating an error when trying to convert the following triggers.
    CREATE TRIGGER min_salary NO CASCADE
    BEFORE INSERT ON staff REFERENCING NEW AS newstaff
    FOR EACH ROW MODE DB2SQL BEGIN ATOMIC
    SET newstaff.salary =
    CASE
    WHEN newstaff.job = 'Mgr' AND newstaff.salary < 17000.00 THEN 17000.00
    WHEN newstaff.job = 'Sales' AND newstaff.salary < 14000.00 THEN 14000.00
    WHEN newstaff.job = 'Clerk' AND newstaff.salary < 10000.00 THEN 10000.00
    ELSE newstaff.salary END;
    END
    CREATE TRIGGER do_not_del_sales NO CASCADE
    BEFORE DELETE ON staff REFERENCING OLD AS oldstaff
    FOR EACH ROW MODE DB2SQL
    WHEN(oldstaff.job = 'Sales') BEGIN ATOMIC SIGNAL SQLSTATE '75000'
    ('Sales staff cannot be deleted... see the DO_NOT_DEL_SALES trigger.'); END
    SEVERE 197 625 oracle.dbtools.migration.workbench.core.MigrationLogResourceBundle Failed to Convert Trigger MIN_SALARY>: class org.antlr.runtime.NoViableAltException -> null: Line 1 Column 0
    SEVERE 196 1328 oracle.dbtools.migration.workbench.core.MigrationLogResourceBundle Failed to Convert Trigger DO_NOT_DEL_SALES>: class org.antlr.runtime.NoViableAltException -> null: Line 1 Column 0
    I can do for Work Bench (SQL Developer) not validate that DB2 must carry primary keys?
    I can do for Work Bench (SQL Developer), I can turn triggers the 2 above?
    Greetings

    to follow up the root cause of ORA-02270 it would be good to get the DB2 table definition as the table is defined in the Oracle database.
    For the other issue with the trigger, please have a look at:
    http://www.oracle.com/technetwork/database/migration/db2-084087.html
    => it lists which objects the migration supports as:
    Supported Migration Objects
         Tables      
         Views
         Indexes
         Users
         Constraints
    => so no triggers.

  • Automating the Creation Process for 1 video for 1000s?

    Here's what I want to do, but I don't know if it's possible:
    For my company, I have designed a fairly simple motion graphics type video for various products that we sell online. It involves the image of the project, some descriptions of it and a title, with some simple movement and transitions with music underneath. The problem though is that we have 1000s of products. Is there a way to create, buy, or have written a script that will be able to take the template project that i've created and use it to do a similar thing with the 1000s of other products (over a long time, presumably)? Just replace the image and text with text and image in a predetermined location.
    Does this make sense? Is it possible? If so, how?
    If not, what other approaches would you guys suggest to approaching this problem?
    Thanks a lot.
    William

    hi william!
    in short: yes.
    what you're asking for is very much possible, and actually not too complicated to do.
    i would recommend using AE's javaScript interface and not the C++ API.
    it's one hell of a lot easier to develop and also to adjust, and it has all the facilities you need.
    in essence:
    1. you supply a file with a list of the data you want to input to each product. (i.e. file names, text, ect...)
    2. the script parses that file and iterates though the entries, each time importing the base project and replacing the footage and text.
    3. either render each project and move to the next, or create a big project containing any number of the entries and render it all together.
    4. tada!

  • Update for Camera Raw 6.6 fails with PSE 10 on Mac OSX

    I have been trying to update my just purchased from App Store Photoshop Elements 10 to the latest Camera Raw 6.6
    #1 I cannot see an update option in the Help Menu
    #2 Try to do a Manual Update after mounting the dmg and runnning AdobePatchInstaller.app says Update is Not Applicable
    Please Help
    Tarique Sani

    Okay. The file paths for the app store version are totally different from that of other versions because apple requires that everything live inside the app package itself for all store apps.
    1. Go here and download ACR 6.6:
    http://www.adobe.com/support/downloads/detail.jsp?ftpID=5315
    Once it's downloaded, double click the dmg file to mount it. Double click Payloads. You are going to want two things from the payloads folder: the plugin and the profiles.
    2. Open AdobeCameraRawProfile6.0All-(followed by a long string of numbers--not sure if they vary by localization). This will be the top folder if you're in list view. Open the .dmg file in there:
    and drag the Profiles folder to the desktop to copy it.
    3. Now go back to the contents of the Payloads folder and do the same thing for the CameraRawforElements6.4All-followed by a long string of numbers). Again, double click the .dmg file in there, then, in the mounted .dmg go to Assets>Universal>and drag cameraraw.plugin to the desktop. You can now eject all the dmg files you've opened.
    So now you have the plug-in file and the profiles folder. Those are the two items you will need.
    4. Now right-click the actual PSE editor in your Applications folder and choose Show Package Contents>Contents>Plug-ins>File Formats, and replace the camera raw.plugin there with the new version (take the old one out and put it somewhere safe in case this doesn't work). In Resources, replace the camera profiles folder and the Lens profiles folder with the ones from the Profiles folder you removed from the .dmg file.
    Hopefully that should work, but I'm not sure if something else may need to be replaced. I don't have a copy of the app store version of PSE 10 to try this, so let me know if that's all that's necessary.

  • Create physical standby database via grid control Errror

    Hi
    I wonder if anyone has seen the following problem
    I am creating a standby database using EM Grid Control wizard. It all configures fine and starts the job goes through all steps including copying datafiles etc. but when it hits the step CREATE STANDBY DATABASE it fails with no explanation. In the oms log all it says is an error about cant find the sid name
    EM Grid COntrol is 10.2.0.5 version
    Windows Primary is Windows 2003 server 10.2.0.4 db
    Trying to create on windows 2003 server with same oracle version
    Thanks
    D

    Did you search the same on Metalink ?
    please also cehck
    EM: Creation process for a physical standby DB failed with SQL Exception
    Regards
    Rajesh

  • Cannot Clear Critical Alert for Physical Standby Databases

    10.2.0.4.0 Grid Control monitoring 10.2.0.4.0 databases and standby databases with 10.2.0.4.0 agents.
    The standby databases are running on Xen guests. The O/S is Red Hat 4 Advanced Server.
    I had a failure of Xen guest on Saturday that caused me to have to rebuild the standby servers. EM Grid Control successfully verifies the configurations and the status is normal for both physical standby databases. (A rebuilt logical standby database shows no alerts.)
    Grid Control database targets page shows a single critical alert for each physical standby database. The alert is for "number of missing media files is 4". The metric graph shows the count as 0 since before the rebuilds. I cannot clear the alerts from the Critical Alerts page. Grid Control reports "The selected alert(s) cannot be manually cleared. They will clear automatically once the metric is no longer in a critical or warning state.".
    Any suggestions?
    Thanks,
    Ray Westphal

    Thanks for the reply Anthony.
    The result of the query on both standby databases is '0'. The metric graph also shows the value at '0' since before rebuilds.
    And the OMS db and agents have been reset several times since I posted this.
    Ray Westphal.

  • Compressing redo data for Physical Standby

    Hi All,
    I have been checking option of compressing redo data for a Physical Standby database. I did not find many details in the docs. Any one has implemented this kind of compression.
    Cheers/Aditya

    per Oracle support, redo compression in DG makes sense only if the bandwidth between the primary & standby server is less than 100mbps and is expected to be heavily taxing on CPU.
    you may check metalink article ID: 729551.1 where they have given some test stats. Not a whole lot of data, but it might help you make a good judgement.
    Cheers.

  • Logical standby stuck at initializing SQL apply only coordinator process up

    Hi
    OS: solaris 5.10
    Hardware: sun sparc
    Oracle database: 11.2.0.1.0
    Primary database name: asadmin
    Standby database name: test
    I had been trying to convert a physical standby to logical standby database. Both the primary and standby reside on the same machine.
    The physical standby was created with a hot backup of primary.
    I had been following document id 278371.1 to convert the physical to logical standby and used the following steps:
    Relevant init parameters on primary:
    *.db_name='asadmin'
    *.db_unique_name='asadmin'
    *.log_archive_config='dg_config=(asadmin,test)'
    *.log_archive_dest_1='location=/u01/asadmin/archive valid_for=(all_logfiles,all_roles) db_unique_name=asadmin'
    *.log_archive_dest_2='SERVICE=test async valid_for=(online_logfiles,primary_role) db_unique_name=test'
    *.log_archive_dest_state_1='enable'
    *.log_archive_dest_state_2='enable'
    *.fal_client='asadmin'
    *.fal_server='test'
    *.remote_login_passwordfile='EXCLUSIVE'
    Relevant init parameters on standby database:
    *.db_name='test' -- Was asadmin before I renamed the DB during conversion to logical standby
    *.db_unique_name='test'
    *.log_archive_dest_1='location=/u01/test/archive valid_for=(all_logfiles,all_roles) db_unique_name=test'
    *.log_archive_dest_2='service=asadmin async valid_for=(online_logfiles,primary_role) db_unique_name=asadmin'
    *.log_archive_dest_state_1=enable
    *.log_archive_dest_state_2=defer
    *.remote_login_passwordfile='EXCLUSIVE'*.fal_server=test
    *.fal_client=asadmin
    Steps on primary:
    1) alter system set log_archive_dest_state_2=defer;
    2) shutdown immediate;
    3) Made sure that the physical standby has applied all of the redo sent to it following the shutdown.
    4) startup mount;
    5) ALTER DATABASE BACKUP CONTROLFILE to '/home/oracle/control01.ctl';
    6) ALTER SYSTEM ENABLE RESTRICTED SESSION;
    7) ALTER DATABASE OPEN;
    8) Verified that the supplemental logging is on.
    9) ALTER SYSTEM ARCHIVE LOG CURRENT;
    10) Checked for the checkpoint change no. at this point which is 72403818 and is present in archive log file 1_62_775102253.dbf
    11) EXECUTE DBMS_LOGSTDBY.BUILD;
    12) ALTER SYSTEM ARCHIVE LOG CURRENT;
    13) Checked for the archive log containing dictionary build which is 1_64_775102253.dbf
    14) ALTER SYSTEM DISABLE RESTRICTED SESSION;
    Details of archive logs and related checkpoint change nos:
    NAME FIRST_CHANGE# NEXT_CHANGE#
    /u01/asadmin/archive/1_61_775102253.dbf 72402901 72403817
    /u01/asadmin/archive/1_62_775102253.dbf 72403817 72404069
    /u01/asadmin/archive/1_63_775102253.dbf 72404069 72404211
    /u01/asadmin/archive/1_64_775102253.dbf 72404211 72405700
    Steps on standby:
    1) shutdown immediate;
    2) Copy the archivelog file 61(was created at primary after apply stopped at standby), 62(contains checkpoint no. 72403818), 63 and 64(contains dictionary build). Copy the backup controlfile from step 5 above to the controlfile location in standby init.
    3) startup mount;
    4) Rename all datafiles and redo log files (including standby redo log files) to the correct path on standby.
    5) alter database recover automatic from '/u01/test/archive' until change 72405700 using backup controlfile; -- This completed error-free
    6) alter database guard all; -- this completed error free
    7) alter database open resetlogs; -- this completed error free.
    8) nid target=sys/oracle12 dbname=test
    9) Changed the db_name in init file to new name test.
    10) Added a tempfile to temp tablespaces.
    11) ALTER DATABASE REGISTER LOGICAL LOGFILE '/u01/test/archive/1_61_775102253.dbf'; -- ORA-16225: Missing LogMiner session name for Streams
    12) ALTER DATABASE START LOGICAL STANDBY APPLY INITIAL 72405700; -- This completed error free.
    Also enabled the log_archive_dest_state_2 on primary.
    After this output from some views:
    SQL> SELECT SESSION_ID, STATE FROM V$LOGSTDBY_STATE;
    SESSION_ID STATE
    1 INITIALIZING
    SQL> SELECT SID, SERIAL#, SPID, TYPE FROM V$LOGSTDBY_PROCESS;
    SID SERIAL# SPID TYPE
    587 22 15476 COORDINATOR
    SELECT PERCENT_DONE, COMMAND
    FROM V$LOGMNR_DICTIONARY_LOAD
    WHERE SESSION_ID = (SELECT SESSION_ID FROM V$LOGSTDBY_STATE);
    PERCENT_DONE
    COMMAND
    0
    SQL> SELECT TYPE, HIGH_SCN, STATUS FROM V$LOGSTDBY;
    TYPE HIGH_SCN STATUS
    COORDINATOR ORA-16111: log mining and apply setting up
    SQL> SELECT APPLIED_SCN, NEWEST_SCN FROM DBA_LOGSTDBY_PROGRESS;
    APPLIED_SCN NEWEST_SCN
    72405700 72411501
    SELECT THREAD#, SEQUENCE#, FILE_NAME FROM DBA_LOGSTDBY_LOG L
    WHERE NEXT_CHANGE# NOT IN
    (SELECT FIRST_CHANGE# FROM DBA_LOGSTDBY_LOG WHERE L.THREAD# = THREAD#)
    ORDER BY THREAD#,SEQUENCE#;
    no rows selected
    SQL> SELECT EVENT_TIME, STATUS, EVENT
    FROM DBA_LOGSTDBY_EVENTS
    ORDER BY EVENT_TIMESTAMP, COMMIT_SCN; 2 3
    EVENT_TIME STATUS EVENT
    14-FEB-12 02:00:50 ORA-16111: log mining and apply setting up
    14-FEB-12 02:00:50 Apply LWM 72405699, HWM 72405699, SCN 72405699
    14-FEB-12 02:20:11 ORA-16128: User initiated stop apply successfully
    completed
    14-FEB-12 02:20:39 ORA-16111: log mining and apply setting up
    14-FEB-12 02:20:39 Apply LWM 72405699, HWM 72405699, SCN 72405699
    14-FEB-12 02:54:15 ORA-16128: User initiated stop apply successfully
    completed
    14-FEB-12 02:57:38 ORA-16111: log mining and apply setting up
    EVENT_TIME STATUS EVENT
    14-FEB-12 02:57:38 Apply LWM 72405699, HWM 72405699, SCN 72405699
    14-FEB-12 03:01:36 ORA-16128: User initiated stop apply successfully
    completed
    14-FEB-12 03:13:44 ORA-16111: log mining and apply setting up
    14-FEB-12 03:13:44 Apply LWM 72405699, HWM 72405699, SCN 72405699
    14-FEB-12 04:32:23 ORA-16128: User initiated stop apply successfully
    completed
    14-FEB-12 04:34:17 ORA-16111: log mining and apply setting up
    14-FEB-12 04:34:17 Apply LWM 72405699, HWM 72405699, SCN 72405699
    EVENT_TIME STATUS EVENT
    14-FEB-12 04:36:16 ORA-16128: User initiated stop apply successfully
    completed
    14-FEB-12 04:36:21 ORA-16111: log mining and apply setting up
    14-FEB-12 04:36:21 Apply LWM 72405699, HWM 72405699, SCN 72405699
    14-FEB-12 05:15:22 ORA-16128: User initiated stop apply successfully
    completed
    14-FEB-12 05:15:29 ORA-16111: log mining and apply setting up
    14-FEB-12 05:15:29 Apply LWM 72405699, HWM 72405699, SCN 72405699
    I also greped for lsp and lcr processes and found that lsp is up but do not see any lcr.
    The logs are getting transported to the archive destination on standby whenever they are archived on primary but are not getting applied to standby.
    Also in case the standby is down while a log is generated on primary it is not automatically transported to standby once the standby is up, means gap resolution is also not working.
    I see the following in alert log every time I try to restart the log apply, everything seems to be stuck at initialization.
    ALTER DATABASE START LOGICAL STANDBY APPLY (test)
    with optional part
    IMMEDIATE
    Attempt to start background Logical Standby process
    Tue Feb 14 05:15:28 2012
    LSP0 started with pid=28, OS id=23391
    Completed: alter database start logical standby apply immediate
    LOGMINER: Parameters summary for session# = 1
    LOGMINER: Number of processes = 3, Transaction Chunk Size = 201
    LOGMINER: Memory Size = 30M, Checkpoint interval = 150M
    LOGMINER: SpillScn 0, ResetLogScn 0
    -- NOTHING AFTER THIS

    Hello;
    I noticed some of your parameters seem to be wrong.
    fal_client - This is Obsolete in 11.2
    You have db_name='test' on the Standby, it should be 'asadmin'
    fal_server=test is set like this on the standby, it should be 'asadmin'
    I might consider changing VALID_FOR to this :
    VALID_FOR=(ONLINE_LOGFILES,ALL_ROLES)Would review 4.2 Step-by-Step Instructions for Creating a Logical Standby Database of Oracle Document E10700-02
    Document 278371.1 is showing its age in my humble opinion.
    -----Wait on this until you fix your parameters----------------------
    Try restarting the SQL Apply
    ALTER DATABASE START LOGICAL STANDBY APPLY IMMEDIATEI don't see the parameter MAX_SERVERS, try setting it to 8 times the number of cores.
    Use these statements to trouble shoot :
    SELECT NAME, VALUE, UNIT FROM V$DATAGUARD_STATS;
    SELECT NAME, VALUE FROM V$LOGSTDBY_STATS WHERE NAME LIKE ;TRANSACTIONS%';
    SELECT COUNT(1) AS IDLE_PREPARERS FROM V$LOGSTDBY_PROCESS WHERE
    TYPE = 'PREPERER' AND STATUS_CODE = 16166;Best Regards
    mseberg
    Edited by: mseberg on Feb 14, 2012 7:37 AM

  • URGENT:Diagnose The Steps for Creating a Standby Database

    Why Following Steps Can Not Succeed? What are the problems?
    The Steps of Creating a Physical Standby Database
    SQL>alter system set db_recovery_file_dest_size=2000m scope=both;
    SQL>alter system set db_recovery_file_dest='C:\oracle\product\10.2.0/flash_recovery_area' scope=both;
    SQL> shutdown immediate
    SQL> startup mount
    SQL> alter database archivelog;
    SQL>alter system set db_flashback_retention_target=120 scope=both;
    SQL> alter database flashback on;
    SQL> alter database force logging;
    #The Primary Database Pfile after being edited
    # Copyright (c) 1991, 2001, 2002 by Oracle Corporation
    # SGA Memory
    sga_target=370147328
    # Job Queues
    job_queue_processes=10
    # Shared Server
    dispatchers="(PROTOCOL=TCP) (SERVICE=orcl10XDB)"
    # Miscellaneous
    compatible=10.2.0.1.0
    # Security and Auditing
    audit_file_dest=C:\oracle\product\10.2.0/admin/orcl10/adump
    remote_login_passwordfile=EXCLUSIVE
    # Sort, Hash Joins, Bitmap Indexes
    pga_aggregate_target=122683392
    # Database Identification
    db_domain=""
    db_name=orcl10
    # File Configuration
    control_files=("C:\oracle\product\10.2.0\oradata\orcl10\control01.ctl", "C:\oracle\product\10.2.0\oradata\orcl10\control02.ctl", "C:\oracle\product\10.2.0\oradata\orcl10\control03.ctl")
    db_recovery_file_dest=C:\oracle\product\10.2.0/flash_recovery_area
    db_recovery_file_dest_size=2147483648
    # Cursors and Library Cache
    open_cursors=300
    # System Managed Undo and Rollback Segments
    undo_management=AUTO
    undo_tablespace=UNDOTBS1
    # Diagnostics and Statistics
    background_dump_dest=C:\oracle\product\10.2.0/admin/orcl10/bdump
    core_dump_dest=C:\oracle\product\10.2.0/admin/orcl10/cdump
    user_dump_dest=C:\oracle\product\10.2.0/admin/orcl10/udump
    # Processes and Sessions
    processes=150
    # Cache and I/O
    db_block_size=8192
    db_file_multiblock_read_count=16
    DB_NAME=orcl10
    DB_UNIQUE_NAME=orcl10
    LOG_ARCHIVE_CONFIG='DG_CONFIG=(orcl10,stdby4)'
    LOG_ARCHIVE_DEST_1=
    'LOCATION=/arch1/orcl10/
    VALID_FOR=(ALL_LOGFILES,ALL_ROLES)
    DB_UNIQUE_NAME=orcl10'
    LOG_ARCHIVE_DEST_2=
    'SERVICE=stdby4 LGWR ASYNC
    VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE)
    DB_UNIQUE_NAME=stdby4'
    LOG_ARCHIVE_DEST_STATE_1=ENABLE
    LOG_ARCHIVE_DEST_STATE_2=ENABLE
    REMOTE_LOGIN_PASSWORDFILE=EXCLUSIVE
    LOG_ARCHIVE_FORMAT=%t_%s_%r.arc
    LOG_ARCHIVE_MAX_PROCESSES=30
    FAL_SERVER=stdby4
    FAL_CLIENT=orcl10
    DB_FILE_NAME_CONVERT='stdby4','orcl10'
    LOG_FILE_NAME_CONVERT=
    '/arch1/stdby4/','/arch1/orcl10/','/arch2/stdby4/','/arch2/orcl10/'
    STANDBY_FILE_MANAGEMENT=AUTO
    SQL> create spfile from pfile;
    RMAN> connect target
    RMAN> backup database;
    ###TheBackupHandle==C:\ORACLE\PRODUCT\10.2.0\FLASH_RECOVERY_AREA\ORCL10\BACKUPSET\2007_11_06\O1_MF_NCSNF_TAG20071106T064915_3LZ7LRN8_.BKP
    RMAN> backup datafile1,2,3,4,5;
    C:\DocumentsandSettings\Administrator>mkdir C:\ORACLE\PRODUCT\10.2.0\ORADATA\temp\stdby4
    SQL> alter database create standby controlfile as '/temp/stdby4.ctl';
    SQL> create pfile='/temp/initstdby4.ora' from spfile;
    #Pfile Edited for The Physical Standby Database
    # Copyright (c) 1991, 2001, 2002 by Oracle Corporation
    # SGA Memory
    sga_target=370147328
    # Job Queues
    job_queue_processes=10
    # Shared Server
    dispatchers="(PROTOCOL=TCP) (SERVICE=stdby4XDB)"
    # Miscellaneous
    compatible=10.2.0.1.0
    # Security and Auditing
    audit_file_dest=C:\oracle\product\10.2.0/admin/stdby4/adump
    remote_login_passwordfile=EXCLUSIVE
    # Sort, Hash Joins, Bitmap Indexes
    pga_aggregate_target=122683392
    # Database Identification
    db_domain=""
    db_name=orcl9
    # File Configuration
    control_files=("C:\oracle\product\10.2.0\oradata\stdby4\control01.ctl", "C:\oracle\product\10.2.0\oradata\stdby4\control02.ctl", "C:\oracle\product\10.2.0\oradata\stdby4\control03.ctl")
    db_recovery_file_dest=C:\oracle\product\10.2.0/stdby4/flash_recovery_area
    db_recovery_file_dest_size=2147483648
    # Cursors and Library Cache
    open_cursors=300
    # System Managed Undo and Rollback Segments
    undo_management=AUTO
    undo_tablespace=UNDOTBS1
    # Diagnostics and Statistics
    background_dump_dest=C:\oracle\product\10.2.0/admin/stdby4/bdump
    core_dump_dest=C:\oracle\product\10.2.0/admin/stdby4/cdump
    user_dump_dest=C:\oracle\product\10.2.0/admin/stdby4/udump
    # Processes and Sessions
    processes=150
    # Cache and I/O
    db_block_size=8192
    db_file_multiblock_read_count=16
    DB_UNIQUE_NAME=stdby4
    LOG_ARCHIVE_CONFIG='DG_CONFIG=(orcl10,stdby4)'
    DB_FILE_NAME_CONVERT='orcl10','stdby4'
    LOG_FILE_NAME_CONVERT=
    '/arch1/orcl10/','/arch1/stdby4/','/arch2/orcl10/','/arch2/stdby4/'
    LOG_ARCHIVE_FORMAT=log%t_%s_%r.arc
    LOG_ARCHIVE_DEST_1=
    'LOCATION=/arch1/stdby4/
    VALID_FOR=(ALL_LOGFILES,ALL_ROLES)
    DB_UNIQUE_NAME=stdby4'
    LOG_ARCHIVE_DEST_2=
    'SERVICE=orcl10 LGWR ASYNC
    VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE)
    DB_UNIQUE_NAME=orcl10'
    LOG_ARCHIVE_DEST_STATE_1=ENABLE
    LOG_ARCHIVE_DEST_STATE_2=ENABLE
    REMOTE_LOGIN_PASSWORDFILE=EXCLUSIVE
    STANDBY_FILE_MANAGEMENT=AUTO
    FAL_SERVER=orcl10
    FAL_CLIENT=stdby4
    C:\Documents and Settings\Administrator>oradim -new -sid stdby4 -startmode manual
    Create Listener and Network Service Name for Standby Database.Stop and start listener(not need)
    SQL>create spfile from pfile='/temp/initstdby4.ora';
    C:\Documents and Settings\Administrator>set oracle_sid=stdby4
    SQL> startup mount;
    ORA-01078: failure in processing system parameters
    LRM-00109: ???????????????? 'C:\ORACLE\PRODUCT\10.2.0\DB_10\DATABASE\INITSTDBY4.ORA'
    Message was edited by:
    frank.qian
    Message was edited by:
    frank.qian

    you have now two different pfiles: one in c:\temp\initstdby4.ora. and %ORACLE_HOME%\database\initstdby4.ora.
    You are using the latter, as you didn't specify pfile= in your startup mount.
    Also, if you are using a spfile, the only line in a pfile should be: spfile=....
    In your pfile you are mixing up Unix like file locations with \ and Windows file locations with /
    Could you please read the docs
    more closely
    more closely
    more closely
    and work
    more carefully
    more carefully
    more carefully
    It is very simple, but you must work carefully.
    Sybrand Bakker
    Senior Oracle DBA

  • Trace in Physical Standby Database

    I'm getting the following messages in the trace file for a Physical Standby database:
    Background Managed Standby Recovery process started
    Start recovery at thread 1 ckpt scn 8253476346531 logseq 4391 block 2
    *** 2008-02-13 09:09:44.366
    Managed Recovery: Active posted.
    *** 2008-02-13 09:09:44.367
    Acquiring global enqueue on thread 1 sequence 4391
    *** 2008-02-13 09:09:44.367
    Failed to acquire global enqueue on thread 1 sequence 4391
    Has anyone encountered this before?

    found this...
    http://ora-00569.ora-code.com/
    ORA-00569:     Failed to acquire global enqueue.
    Cause:     A prior error occurred on one of the instances in the cluster. Typically errors are caused by shared pool resource contention.
    Action:     Check for and resolve prior errors on all instances in the cluster. If there is shared pool resource contention, increase the SHARED_POOL_SIZE, DML_LOCKS, PROCESSES, TRANSACTIONS, CLUSTER_DATABASE_INSTANCES and PARALLEL_MAX_SERVERS initialization parameters.

  • Physical Standby: RFS: 'Attach' request version mismatch

    I have a 10.1.0.4 database that I want to move to a new machine and also upgrade to 10.2.0.1. I'm trying to follow the instructions for using SQL Apply to perforn a rolling upgrade with minimal downtime. My understanding is that I must first create a physical standby, and then transition the physical standby to a logical standby. Well, my physical standby is failing. I see the following in the standby's alert log:
    RFS: 'Attach' request version mismatch
    Both databases have their COMPATIBLE parameter set to 10.1.0.2.
    As an aside, if I must first create a working physical standby, then why do I need to transition that standby to a logical standby? I mean, can't I just get the physical standby working, perform a switchover, and be done with it?

    I was confused. I thought I could just create a 10.2.0.1 standby for a 10.1.0.4 primary. The process, though, is to first get a working 10.1.0.4 (logical) standby for the 10.1.0.4 primary. At that point dbua can be used to upgrade the standby, and I can go from there.
    I'm sure everything will work, now that I understand the steps required.

Maybe you are looking for