Anyone Here want a Data Guard Certification?

There is a Oracle RAC Certfied Expert Certification . . . Why not one for Data Guard?
It would be nice for those of us who have spent time with Data Guard
Post a comment and let's see if they will make one.

OCP is meaningless. You wouldn't go to a dentist who got their degree with a multiple choice test with so low a bar to passing. Or an attorney, or a surgeon. Why is OCP acceptable as a criterion for anything other than private schools to make money?
To be meaningful a certification must be hands-on and must include proving competence. If you can not build a 3 node RAC cluster, unaided, you should not get the piece of paper.

Similar Messages

  • Will Oracle ever create a Data Guard certification?

    Will Oracle ever create a Data Guard Certification?
    If so would it be more of a "HA" - high availability,
    or would it be only about Active Data Guard.
    Thanks

    We have a Data Guard Admin certification coming soon. Unfortunately, I don't have a closer date than that.
    https://blogs.oracle.com/certification/entry/0856_31
    Regards,
    Brandye Barrington
    Forum Moderator

  • Should e-book Data Guard Concepts and Administration be useless for us?

    I think The Data Guard Broker is an easy and efficient way to manage a Data Guard configuration. The Data Guard Broker provides us with two efficient tools: EM Console and DGMGRL. We can do everything we want for Data Guard system.
    Why we spend so much time to read Oracle e-book Data Guard Concepts and Administration? I think it is useless. Only Broker is enough?

    Frank,
    The thing is GUI tool itself is also program that need some one to manage. The developer of such tools cannot take care unlimited possibilities in the real world.
    Just check how many threads in this forum about why emca, dbca and EM console not working.
    Being a DBA, we should understand the underlying technology that driven the system. It's like driving a car. A regular customer could say I don't care about mechanics of a car. As long as it's running, that's enough. But we are not regular customer, we suppose to be the auto repair mechanician in Oracle world. We could enjoy driving a car, but at same time we should know why and how it's working.

  • Data Guard Gap Monitoring script

    Hello,
    Can anyone please provide me data guard gap monitoring script for databases(primary,standby) on RAC.
    Oracle RDBMS 11.2.0.2(4-node RAC) on RHEL 5.6.
    Thanks
    Edited by: 951368 on Dec 26, 2012 9:21 AM

    951368 wrote:
    Hello,
    Can anyone please provide me data guard gap monitoring script for databases(primary,standby) on RAC.
    Oracle RDBMS 11.2.0.2(4-node RAC) on RHEL 5.6.
    Thanks
    Edited by: 951368 on Dec 26, 2012 9:21 AMUse the script of MSeberg, Modify v$instance as gv$instance for RAC

  • SAP & Data Guard

    hi,
    i want to Data Guard on SAP. have you got any idea?
    thanks.

    Hi,
    First read Dataguard documentation see below link
    http://download.oracle.com/docs/cd/B10501_01/server.920/a96653/create_ps.htm
    https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/d2827bef-0c01-0010-75a3-f96aa47fe920?overridelayout=true
    SurendraJain

  • I Want to Buy MacBook Pro 15" In Installments Can Anyone Here Tell me The Plans Of Installments

    I Want to Buy MacBook Pro 15" In Installments Can Anyone Here Tell me The Plans Of Installments... Please Give Me Proper Information With Interest..
    Thanks In Advance

    Hello,
    Apple offers a Barclay's VISA card that provides 6-12 months at 0% interest on your first purchase from Apple.  You can even sign up for the card & get approved at the same time that you place your order. 
    If your purchase is under $999 USD you get 6 months at 0% interest.  If more than $999 you get 12 months at 0% interest.  As long as you make the minimum payment on time every month you continue at 0% until the promotional period runs out.
    Note: if any payment is late or you don't pay off the balance within the promotional period, you get charged all the interest that accrued from day one; as of today that rate is over 22%, so it is important to make all your payments on time and pay off of the full balance before the end of the promotional period.
    Last year I applied for the card at the same time I ordered my iMac; got approved immediately; and then set up an automatic payment every month with Barclays so I wouldn't risk being late; and then I paid off the remaining balance 30 days before the end of the promotional period.  Everything went smoothly, Barclays' online system worked perfectly; they did not play any billing tricks nor did they ever spam me with junk promotions.  Recently they offered another 12 months at 0% financing, so it's not just a one time offer!
    Your profile doesn't say where you are located, so if you are outside the USA you will need to check the Apple website for your country to see if there is a similar financing offer in your location.

  • Here is a question does anyone ever want to just kill final cut sometimes

    Here is a question does anyone ever want to just kill final cut sometimes for doing things that seriously just make no sense

    I would, but then would be out 1200 dollars.

  • HT4221 I really want my Apple TV, iPad and iPhone to sort them in date taken like in iPhoto on my iMac. How do I do that? Sorting on date modified seems so stupid to me, why would anyone need this? Date taken gives a timeline in your event.

    I really want my Apple TV, iPad and iPhone to sort them in date taken like in iPhoto on my iMac. How do I do that? Sorting on date modified seems so stupid to me, why would anyone need this? Date taken gives a timeline in your event.

    The unix commands you need are:
    GetFileInfo
    SetFileInfo
    and maybe find
    for cryptic details use the man command
    Macintosh-HD -> Applications -> Utilities -> Terminal
    man SetFileInfo
    You may use the SetFileInfo command to set the file type & the program which will open the file.
    # This little gem will do a get info on all files in a directory.
    mac $ ls  | xargs -I {} GetFileInfo "{}"
    file: "/Users/mac/playdoc/oddadocodd"
    type: ""
    creator: ""
    attributes: avbstclinmedz
    created: 05/01/2011 14:53:22
    modified: 05/01/2011 14:53:22
    file: "/Users/mac/playdoc/one.docx"
    type: ""
    creator: ""
    attributes: avbstclinmedz
    created: 05/01/2011 13:57:48
    modified: 05/01/2011 13:57:48
    file: "/Users/mac/playdoc/oneLineFile"
    type: "TEXT"
    creator: "!Rch"
    attributes: avbstclinmedz
    created: 05/07/2011 14:27:17
    modified: 05/07/2011 14:27:17
    file: "/Users/mac/playdoc/oneLineFile.txt"
    type: "TEXT"
    creator: "!Rch"
    attributes: avbstclinmedz
    created: 05/07/2011 14:27:49
    modified: 05/07/2011 14:27:49
    file: "/Users/mac/playdoc/three.docx"
    type: ""
    creator: ""
    attributes: avbstclinmedz
    created: 05/01/2011 13:58:03
    modified: 05/01/2011 13:58:03
    file: "/Users/mac/playdoc/two.docx"
    type: ""
    creator: ""
    attributes: avbstclinmedz
    created: 05/01/2011 13:57:56
    modified: 05/01/2011 13:57:56
    file: "/Users/mac/playdoc/weirder.doc.docx"
    type: ""
    creator: ""
    attributes: avbstclinmedz
    created: 05/01/2011 14:50:03
    modified: 05/01/2011 14:50:03
    # well, ! is a funnie character so we escape it.
    mac $ SetFile -t TEXT -c \!Rch two.docx
    mac $ GetFileInfo two.docx
    file: "/Users/mac/playdoc/two.docx"
    type: "TEXT"
    creator: "!Rch"
    attributes: avbstclinmedz
    created: 05/01/2011 13:57:56
    modified: 05/01/2011 13:57:56
    mac $
    mac $ date
    Sat May  7 14:40:56 EDT 2011
    mac $

  • Data Guard two Production sites

    I am exploring the best (and least expensive) way to configure Data Guard in our environment.
    Has anyone had two production environments, each in a different location and set up Standby Data Guard databases across each.
    For example -
    Site 1 - Prod RAC 1 with database 1 and a standby for database 2
    Site 2 - Prod RAC 2 with database 2 and a standby for database 1
    Is there any load to the resources in either cluster when also maintaining a standby database along with production?
    I searched Metalink and was not able to find information on this type of configuration.
    Thanks!

    Yes, your understanding is correct. I simplified our environment for the question, actually we have two sites in different geographic locations.
    Site 1 is production and Site 2 is Development, but we want to protect each site for Disaster Recovery, etc.
    Site 1 has a three node rac cluster with a 'PRD_A' database and Site 2 has a three node rac cluster with a 'PRD_B'.
    So, I would like to create, just as you said, a standby for PRD_A in the Site 2 cluster & create standby for PRD_B in the Site 1 cluster.
    And, actually, each site has 2-3 databases running that I would like to create standby's for. This is why I am wondering if anyone has done this and if the load would be an issue. Yo Yo got it ;-)
    See PRD_A has 3 instances let suppose PRD_A1, PRD_A2, PRD_A3.
    PRD_B has 3 instances let suppose PRD_B1, PRD_B2, PRD_B3
    So for PRD_A you can create either one standby database as PRD_A_STBY.
    or you can create three standby databases as PRD_A1_STBY, PRD_A2_STBY, PRD_A3_STBY on CLUSTER B
    For PRD_B you can create either one standby database as PRD_B_STBY.
    or you can create three standby databases as PRD_B1_STBY, PRD_B2_STBY, PRD_B3_STBY on CLUSTER A
    But Recovery process will be performed any one of the standby of one database, Remaining standby databases will sit idle.
    (or)
    You can have only one standby database, So the database itself clustered so you can add instances if in case of Switchover/Failover.
    Here you will have two database on each instance.
    Here the things
    1) how much RAM
    2) is it server capable to hold two instances?

  • Data Guard Agent, Authentication Failure

    I'm working with two Windows 2003 servers, attempting to use one as a standby and one as a primary database use Data Guard. However, I'm having a bit of trouble when trying to get one server to communicate through the Management Agent and Management Service. I've done Management Agent installs on about 20 XP workstations and they've also worked wonderfully with the Oracle Grid Control.
    When the agent on my would-be standby database instance starts up I'm receiving the following errors in emagent.trc:
    2005-11-01 15:16:54 Thread-3836 WARN main: clear collection state due to OMS_version difference
    2005-11-01 15:16:54 Thread-3836 WARN command: Job Subsystem Timeout set at 600 seconds
    2005-11-01 15:16:54 Thread-3836 WARN upload: Upload manager has no Failure script: disabled
    2005-11-01 15:16:54 Thread-3836 WARN upload: Recovering left over xml files in upload directory
    2005-11-01 15:16:54 Thread-3836 WARN upload: Recovered 0 left over xml files in upload directory
    2005-11-01 15:16:54 Thread-3836 WARN metadata: Metric RuntimeLog does not have any data columns
    2005-11-01 15:16:54 Thread-3836 WARN metadata: Metric collectSnapshot does not have any data columns
    2005-11-01 15:16:54 Thread-3836 ERROR engine: [oracle_bc4j] CategoryProp NAME [VersionCategory] is not one of the valid choices
    2005-11-01 15:16:54 Thread-3836 ERROR engine: ParseError: File=D:\oracle\product\10.1.0\dg\sysman\admin\metadata\oracle_bc4j.xml, Line=486, Msg=attribute NAME in <CategoryProp> cannot be NULL
    2005-11-01 15:16:54 Thread-3836 WARN metadata: Column name EFFICIENCY__BYTES_SAVED_WITH_COMPRESSION__AVG_PER_SEC_SINCE_START too long, truncating to EFFICIENCY__BYTES_SAVED_WITH_COMPRESSION__AVG_PER_SEC_SINCE_STAR
    2005-11-01 15:16:54 Thread-3836 WARN metadata: Column name ESI__ERRORS__ESI_DEFAULT_FRAGMENT_SERVED__AVG_PER_SEC_SINCE_START too long, truncating to ESI__ERRORS__ESI_DEFAULT_FRAGMENT_SERVED__AVG_PER_SEC_SINCE_STAR
    2005-11-01 15:16:54 Thread-3836 WARN metadata: Column name SERVERS__APP_SRVR_STATS__SERVER__REQUESTS__AVG_PER_SEC_SINCE_START too long, truncating to SERVERS__APP_SRVR_STATS__SERVER__REQUESTS__AVG_PER_SEC_SINCE_STA
    2005-11-01 15:16:54 Thread-3836 WARN metadata: Column name SERVERS__APP_SRVR_STATS__SERVER__LATENCY__MAX_PER_SEC_SINCE_START too long, truncating to SERVERS__APP_SRVR_STATS__SERVER__LATENCY__MAX_PER_SEC_SINCE_STAR
    2005-11-01 15:16:54 Thread-3836 WARN metadata: Column name SERVERS__APP_SRVR_STATS__SERVER__LATENCY__AVG_PER_SEC_SINCE_START too long, truncating to SERVERS__APP_SRVR_STATS__SERVER__LATENCY__AVG_PER_SEC_SINCE_STAR
    2005-11-01 15:16:54 Thread-3836 WARN metadata: Column name SERVERS__APP_SRVR_STATS__SERVER__OPEN_CONNECTIONS__MAX_SINCE_START too long, truncating to SERVERS__APP_SRVR_STATS__SERVER__OPEN_CONNECTIONS__MAX_SINCE_STA
    2005-11-01 15:16:54 Thread-3836 WARN metadata: Column name SERVER__APP_SRVR_STATS__SERVER__REQUESTS__MAX_PER_SEC_SINCE_START too long, truncating to SERVER__APP_SRVR_STATS__SERVER__REQUESTS__MAX_PER_SEC_SINCE_STAR
    2005-11-01 15:16:54 Thread-3836 WARN metadata: Column name SERVER__APP_SRVR_STATS__SERVER__REQUESTS__AVG_PER_SEC_SINCE_START too long, truncating to SERVER__APP_SRVR_STATS__SERVER__REQUESTS__AVG_PER_SEC_SINCE_STAR
    2005-11-01 15:16:54 Thread-3836 WARN metadata: Column name SERVER__APP_SRVR_STATS__SERVER__OPEN_CONNECTIONS__MAX_SINCE_START too long, truncating to SERVER__APP_SRVR_STATS__SERVER__OPEN_CONNECTIONS__MAX_SINCE_STAR
    2005-11-01 15:16:54 Thread-3836 WARN metadata: Metric Wireless_PID does not have any data columns
    2005-11-01 15:16:54 Thread-3836 WARN metadata: Metric numberOfAppDownloadsOverInterval_instance does not have any data columns
    2005-11-01 15:17:00 Thread-4172 WARN vpxoci: OCI Error -- ErrorCode(1017): ORA-01017: invalid username/password; logon denied
    SQL = " OCISessionBegin"...
    LOGIN = dbsnmp/<PW>@(DESCRIPTION=(ADDRESS=(PROTOCOL=TCP)(HOST=MY_DATABASE)(PORT=1521))(CONNECT_DATA=(SID=CPD2DB)))
    2005-11-01 15:17:00 Thread-4172 ERROR vpxoci: ORA-01017: invalid username/password; logon denied
    2005-11-01 15:17:00 Thread-4172 WARN vpxoci: Login 0xe8c220 failed, error=ORA-01017: invalid username/password; logon denied
    2005-11-01 15:17:00 Thread-4172 WARN TargetManager: Exception in computing dynamic properties of {MY_DATABASE, oracle_database },MonitorConfigStatus::ORA-01017: invalid username/password; logon denied
    2005-11-01 15:17:01 Thread-4172 WARN vpxoci: OCI Error -- ErrorCode(1017): ORA-01017: invalid username/password; logon denied
    I've already toggled the Local Security Policy (Log On As Batch Job) setting in Windows, unlocked the Monitoring Profile account, etc. I've also tried to set the Preferred Host Credentials for the database, but it doesn't seem to want to authenticate the Windows 2003 Administrator user.
    Anyone have any other suggestions?

    Check the following:
    Does the user have administrative privilege on the system?
    Is the user running this part of ORA_DBA group?
    Does the user have the local security policy "Logon as Batch Job"?
    Have you set the OS Preferred Credential? If you are a domain user, this will be looking for domain\user name instead of just the user name.
    On another note:
    Have you doen any upgrades to the OMS repository?
    If yes, is the new Repository compatible with the EM Console?

  • Clarification on Data Guard(Physical Standyb db)

    Hi guys,
    I have been trying to setup up Data Guard with a physical standby database for the past few weeks and I think I have managed to setup it up and also perform a switchover. I have been reading a lot of websites and even Oracle Docs for this.
    However I need clarification on the setup and whether or not it is working as expected.
    My environment is Windows 32bit (Windows 2003)
    Oracle 10.2.0.2 (Client/Server)
    2 Physical machines
    Here is what I have done.
    Machine 1
    1. Create a primary database using standard DBCA, hence the Oracle service(oradgp) and password file are also created along with the listener service.
    2. Modify the pfile to include the following:-
    oradgp.__db_cache_size=436207616
    oradgp.__java_pool_size=4194304
    oradgp.__large_pool_size=4194304
    oradgp.__shared_pool_size=159383552
    oradgp.__streams_pool_size=0
    *.audit_file_dest='M:\oracle\product\10.2.0\admin\oradgp\adump'
    *.background_dump_dest='M:\oracle\product\10.2.0\admin\oradgp\bdump'
    *.compatible='10.2.0.3.0'
    *.control_files='M:\oracle\product\10.2.0\oradata\oradgp\control01.ctl','M:\oracle\product\10.2.0\oradata\oradgp\control02.ctl','M:\oracle\product\10.2.0\oradata\oradgp\control03.ctl'
    *.core_dump_dest='M:\oracle\product\10.2.0\admin\oradgp\cdump'
    *.db_block_size=8192
    *.db_domain=''
    *.db_file_multiblock_read_count=16
    *.db_name='oradgp'
    *.db_recovery_file_dest='M:\oracle\product\10.2.0\flash_recovery_area'
    *.db_recovery_file_dest_size=21474836480
    *.fal_client='oradgp'
    *.fal_server='oradgs'
    *.job_queue_processes=10
    *.log_archive_dest_1='LOCATION=E:\ArchLogs VALID_FOR=(ALL_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=oradgp'
    *.log_archive_dest_2='SERVICE=oradgs LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=oradgs'
    *.log_archive_format='ARC%S_%R.%T'
    *.log_archive_max_processes=30
    *.nls_territory='IRELAND'
    *.open_cursors=300
    *.pga_aggregate_target=203423744
    *.processes=150
    *.remote_login_passwordfile='EXCLUSIVE'
    *.sga_target=612368384
    *.standby_file_management='auto'
    *.undo_management='AUTO'
    *.undo_tablespace='UNDOTBS1'
    *.user_dump_dest='M:\oracle\product\10.2.0\admin\oradgp\udump'
    *.service_names=oradgp
    The locations on the harddisk are all available and archived redo are created (e:\archlogs)
    3. I then add the necessary (4) standby logs on primary.
    4. To replicate the db on the machine 2(standby db), I did an RMAN backup as:-
    RMAN> run
    {allocate channel d1 type disk format='M:\DGBackup\stby_%U.bak';
    backup database plus archivelog delete input;
    5. I then copied over the standby~.bak files created from machine1 to machine2 to the same directory (M:\DBBackup) since I maintained the directory structure exactly the same between the 2 machines.
    6. Then created a standby controlfile. (At this time the db was in open/write mode).
    7. I then copied this standby ctl file to machine2 under the same directory structure (M:\oracle\product\10.2.0\oradata\oradgp) and replicated the same ctl file into 3 different files such as: CONTROL01.CTL, CONTROL02.CTL & CONTROL03.CTL
    Machine2
    8. I created an Oracle service called the same as primary (oradgp).
    9. Created a listener also.
    9. Set the Oracle Home & SID to the same name as primary (oradgp) <<<-- I am not sure about the sid one.
    10. I then copied over the pfile from the primary to standby and created an spfile with this one.
    It looks like this:-
    oradgp.__db_cache_size=436207616
    oradgp.__java_pool_size=4194304
    oradgp.__large_pool_size=4194304
    oradgp.__shared_pool_size=159383552
    oradgp.__streams_pool_size=0
    *.audit_file_dest='M:\oracle\product\10.2.0\admin\oradgp\adump'
    *.background_dump_dest='M:\oracle\product\10.2.0\admin\oradgp\bdump'
    *.compatible='10.2.0.3.0'
    *.control_files='M:\oracle\product\10.2.0\oradata\oradgp\control01.ctl','M:\oracle\product\10.2.0\oradata\oradgp\control02.ctl','M:\oracle\product\10.2.0\oradata\oradgp\control03.ctl'
    *.core_dump_dest='M:\oracle\product\10.2.0\admin\oradgp\cdump'
    *.db_block_size=8192
    *.db_domain=''
    *.db_file_multiblock_read_count=16
    *.db_name='oradgp'
    *.db_recovery_file_dest='M:\oracle\product\10.2.0\flash_recovery_area'
    *.db_recovery_file_dest_size=21474836480
    *.fal_client='oradgs'
    *.fal_server='oradgp'
    *.job_queue_processes=10
    *.log_archive_dest_1='LOCATION=E:\ArchLogs VALID_FOR=(ALL_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=oradgs'
    *.log_archive_dest_2='SERVICE=oradgp LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=oradgp'
    *.log_archive_format='ARC%S_%R.%T'
    *.log_archive_max_processes=30
    *.nls_territory='IRELAND'
    *.open_cursors=300
    *.pga_aggregate_target=203423744
    *.processes=150
    *.remote_login_passwordfile='EXCLUSIVE'
    *.sga_target=612368384
    *.standby_file_management='auto'
    *.undo_management='AUTO'
    *.undo_tablespace='UNDOTBS1'
    *.user_dump_dest='M:\oracle\product\10.2.0\admin\oradgp\udump'
    *.service_names=oradgs
    log_file_name_convert='junk','junk'
    11. User RMAN to restore the db as:-
    RMAN> startup mount;
    RMAN> restore database;
    Then RMAN created the datafiles.
    12. I then added the same number (4) of standby redo logs to machine2.
    13. Also added a tempfile though the temp tablespace was created per the restore via RMAN, I think the actual file (temp01.dbf) didn't get created, so I manually created the tempfile.
    14. Ensuring the listener and Oracle service were running and that the database on machine2 was in MOUNT mode, I then started the redo apply using:-
    SQL> ALTER DATABASE RECOVER MANAGED STANDBY DATABASE DISCONNECT FROM SESSION;
    It seems to have started the redo apply as I've checked the alert log and noticed that the sequence# was all "YES" for applied.
    ****However I noticed that in the alert log the standby was complaining about the online REDO log not being present****
    So copied over the REDO logs from the primary machine and placed them in the same directory structure of the standby.
    ########Q1. I understand that the standby database does not need online REDO Logs but why is it reporting in the alert log then??########
    I wanted to enable realtime apply so, I cancelled the recover by :-
    SQL> ALTER DATABASE RECOVER MANAGED STANDBY DATABASE CANCEL;
    and issued:-
    SQL> ALTER DATABASE RECOVER MANAGED STANDBY DATABASE USING CURRENT LOGFILE DISCONNECT;
    This too was successful and I noticed that the recovery mode is set to MANAGED REAL TIME APPLY.
    Checked this via the primary database also and it too reported that the DEST_2 is in MANAGED REAL TIME APPLY.
    Also performed a log swith on primary and it got transported to the standby and was applied (YES).
    Also ensured that there are no gaps via some queries where no rows were returned.
    15. I now wanted to perform a switchover, hence issued:-
    Primary_SQL> ALTER DATABASE COMMIT TO SWITCHOVER TO PHYSICAL STANDBY WITH SESSION SHUTDOWN;
    All the archivers stopped as expected.
    16. Now on machine2:
    Stdby_SQL> ALTER DATABASE COMMIT TO SWITCHOVER TO PRIMARY;
    17. On machine1:
    Primary_Now_Standby_SQL>SHUTDOWN IMMEDIATE;
    Primary_Now_Standby_SQL>STARTUP MOUNT;
    Primary_Now_Standby_SQL> ALTER DATABASE RECOVER MANAGED STANDBY DATABASE USING CURRENT LOGFILE DISCONNECT;
    17. On machine2:
    Stdby_Now_Primary_SQL>ALTER DATABASE OPEN;
    Checked by switching the logfile on the new primary and ensured that the standby received this logfile and was applied (YES).
    However, here are my questions for clarifications:-
    Q1. There is a question about ONLINE REDO LOGS within "#" characters.
    Q2. Do you see me doing anything wrong in regards to naming the directory structures? Should I have renamed the dbname directory in the Oracle Home to oradgs rather than oradgp?
    Q3. When I enabled real time apply does that mean, that I am not in 'MANAGED' mode anymore? Is there an un-managed mode also?
    Q4. After the switchover, I have noticed that the MRP0 process is "APPLYING LOG" status to a sequence# which is not even the latest sequence# as per v$archived_log. By this I mean:-
    SQL> SELECT PROCESS, STATUS, THREAD#, SEQUENCE#, BLOCK#, BLOCKS,FROM V$MANAGED_STANDBY;
    MRP0 APPLYING_LOG 1 47 452 1024000
    but :
    SQL> select max(sequence#) from v$archived_log;
    46
    Why is that? Also I have noticed that one of the sequence#s is NOT applied but the later ones are:-
    SQL> SELECT SEQUENCE#,APPLIED FROM V$ARCHIVED_LOG ORDER BY SEQUENCE#;
    42 NO
    43 YES
    44 YES
    45 YES
    46 YES
    What could be the possible reasons why sequence# 42 didn't get applied but the others did?
    After reading several documents I am confused at this stage because I have read that you can setup standby databases using 'standby' logs but is there another method without using standby logs?
    Q5. The log switch isn't happening automatically on the primary database where I could see the whole process happening on it own, such as generation of a new logfile, that being transported to the standby and then being applied on the standby.
    Could this be due to inactivity on the primary database as I am not doing anything on it?
    Sorry if I have missed out something guys but I tried to put in as much detail as I remember...
    Thank you very much in advance.
    Regards,
    Bharath
    Edited by: Bharath3 on Jan 22, 2010 2:13 AM

    Parameters:
    Missing on the Primary:
    DB_UNIQUE_NAME=oradgp
    LOG_ARCHIVE_CONFIG=DG_CONFIG=(oradgp, oradgs)
    Missing on the Standby:
    DB_UNIQUE_NAME=oradgs
    LOG_ARCHIVE_CONFIG=DG_CONFIG=(oradgp, oradgs)
    You said: Also added a tempfile though the temp tablespace was created per the restore via RMAN, I think the actual file (temp01.dbf) didn't get created, so I manually created the tempfile.
    RMAN should have also added the temp file. Note that as of 11g RMAN duplicate for standby will also add the standby redo log files at the standby if they already existed on the Primary when you took the backup.
    You said: ****However I noticed that in the alert log the standby was complaining about the online REDO log not being present****
    That is just the weird error that the RDBMS returns when the database tries to find the online redo log files. You see that at the start of the MRP because it tries to open them and if it gets the error it will manually create them based on their file definition in the controlfile combined with LOG_FILE_NAME_CONVERT if they are in a different place from the Primary.
    Your questions (Q1 answered above):
    You said: Q2. Do you see me doing anything wrong in regards to naming the directory structures? Should I have renamed the dbname directory in the Oracle Home to oradgs rather than oradgp?
    Up to you. Not a requirement.
    You said: Q3. When I enabled real time apply does that mean, that I am not in 'MANAGED' mode anymore? Is there an un-managed mode also?
    You are always in MANAGED mode when you use the RECOVER MANAGED STANDBY DATABASE command. If you use manual recovery "RECOVER STANDBY DATABASE" (NOT RECOMMENDED EVER ON A STANDBY DATABASE) then you are effectively in 'non-managed' mode although we do not call it that.
    You said: Q4. After the switchover, I have noticed that the MRP0 process is "APPLYING LOG" status to a sequence# which is not even the latest sequence# as per v$archived_log. By this I mean:-
    Log 46 (in your example) is the last FULL and ARCHIVED log hence that is the latest one to show up in V$ARCHIVED_LOG as that is a list of fully archived log files. Sequence 47 is the one that is current in the Primary online redo log and also current in the standby's standby redo log and as you are using real time apply that is the one it is applying.
    You said: What could be the possible reasons why sequence# 42 didn't get applied but the others did?
    42 was probably a gap. Select the FAL columns as well and it will proably say 'YES' for FAL. We do not update the Primary's controlfile everytime we resolve a gap. Try the same command on the standby and you will see that 42 was indeed applied. Redo can never be applied out of order so the max(sequence#) from v$archived_log where applied = 'YES' will tell you that every sequence before that number has to have been applied.
    You said: After reading several documents I am confused at this stage because I have read that you can setup standby databases using 'standby' logs but is there another method without using standby logs?
    Yes, If you do not have standby redo log files on the standby then we write directly to an archive log. Which means potential large data loss at failover and no real time apply. That was the old 9i method for ARCH. Don't do that. Always have standby redo logs (SRL)
    You said: Q5. The log switch isn't happening automatically on the primary database where I could see the whole process happening on it own, such as generation of a new logfile, that being transported to the standby and then being applied on the standby.
    Could this be due to inactivity on the primary database as I am not doing anything on it?
    Log switches on the Primary happen when the current log gets full, a log switch has not happened for the number of seconds you specified in the ARCHIVE_LAG_TARGET parameter or you say ALTER SYSTEM SWITCH LOGFILE (or the other methods for switching log files. The heartbeat redo will eventually fill up an online log file but it is about 13 bytes so you do the math on how long that would take :^)
    You are shipping redo with ASYNC so we send the redo as it is commited, there is no wait for the log switch. And we are in real time apply so there is no wait for the log switch to apply that redo. In theroy you could create an online log file large enough to hold an entire day's worth of redo and never switch for the whole day and the standby would still be caught up with the primary.

  • Anyone here make a living with Photoshop?

    Does anyone here make a living with PS (or close to it)? I don't really mean in combination with anything else (like a graphics arts degree, or "I use it along with these six other programs I'm an expert in, too").
    From what I'm gathering, relying solely on PS for income would leave one starving, and I understand that an employer - if one isn't a freelancer - would want an employee to have multiple areas of expertise.

    I'm 53 now so that made me 38 when I started using Ps. Not that far off from where you are today. All I can say is that it happened very gradually for me, first just playing around then trying to emulate what prepress houses were doing for my clients at the time on images I had shot, to finally doing small jobs myself, usually learning just enough at one time to figure out how to get that one project done.
    Then after reading the early Real World books and later the Professional Photoshop series, I began to get a grasp of what it took to start to take on the responsibility of taking on projects where my contribution was a small part of a much larger budget - annual reports that cost a quarter million to print, large campaigns for public utilities that had billboards, busboards and ads all across L.A. I had to step up my game and not just be "as good" as those prepress and retouching houses that my clients used to use, I had to be significantly better. This involved not only being able to do anything they through at me but also saying yes when when they told me I could say no. It also involved learning as much as possible about digital color, digital color management and how to make digital images reproduce their best wherever and however the clients wanted to print. And that led to investment in hardware and software - tens or thousands of dollars in spectrophotometers and color management software that gave me a competitive advantage over my competition and tools that even though many print shops owned, very few knew how to use. It's that constant pushing forward and drive to improve your skills that keeps you interested in a very fascinating place in history.
    There's room for anyone good, and even if you're just starting out, and somehow consider yourself mid-career, that's no reason to give up. Before you know it, you'll be able to offer services that others aren't willing to do for themselves and if you don't try to take on too much at once, you'll look back in a couple of years and ask yourself why you were so worried at the time.
    There is so much more good information and better, less expensive tools readily available that it's easier today that ever before. Hell, I can remember participating in a Monday night live AOL chat where Chris Cox, the Adobe engineer was the guest and asking him when they were going to start supporting ColorSync and ColorSync profiles, to which he replied something on the order that that was not a very good direction to go in. My how things have changed.
    I just finished what it probably one of the most difficult Ps jobs I've done to date - the 2010 Manhattan Wine Auction Poster - for Manhattan Beach, Ca. not Manhattan, NY. It really pushed the limits of my own capabilities, both in the actual shooting and certainly in the digital post production. Here's a link to their website which has the poster on the first page.
    http://manhattanwineauction.com/
    This was every bit as hard as it looks but most importantly, involved another helicopter ride, taking the doors off a Robinson 44 and hanging out over the ocean. There were three bottles of wine involved in what you see here - a bottle of two buck chuck from Trader Joe's that I drilled a 1/4 inch vent hole in to facilitate a smoother pour, an opened $400 bottle of wine for the open neck and a full bottle so it wouldn't be too transparent - all blended together to form one perfect union. The splash was from a rubber ball dropped on a fishing line into a washtub full of wine colored Kool-Aid and then the concentric ripples from a can of Dust-Off compressed air shot into the same washtube, all photographed with the Canon 1DsMKIII and a 90mm Tilt-Shift lens with a Broncolor strobe that had a 1/7500th of a second flash duration to stop the action. I populated the pier with people from alternate aerial shots and finally used surfing shots from Body Glove for the three surfers at the bottom. The final file was prepped for printing at Lithocraft in Anaheim using a custom profile I made for their press and the printed piece looked as close to what I saw I my screen as anything else I have taken to press.

  • Data Guard Broker and Cold Failover clusters

    Hi,
    I wanted to use Data Guard Broker to control primary/standby systems on AIX clusters but found that without Oracle clusterware it is not supported on failover of a database to another node (unless you drop and recreate the configuration which in my opinion is not support at all!).
    The 11g documentation states that DG Broker offers support for single instance databases configured for HA using Oracle Clusterware and "cold failover clusters".
    Does anyone know whether this support for cold failover clusters in 11g means "proper support" i.e the configuration detects that the hostname has changed and automatically renames it so it continues to work?, or is the support in 11g the same as that in 10g?.
    Thanks,
    Andy

    Hi Jan,
    We already use virtual hostnames.
    When I set up the configuration the hostname for the databases default to the server name. The only way I know of changing this is to disable the configuration, then use the "edit database...set property" command but when I enable the configuration again it reverts back to its previous value.
    regards,
    Andy

  • Using Data Guard and hot backups - 9.2.0.6.0

    Hi all,
    I have an existing 9.2.0.6.0 database that is setup in a DataGuard environment - one primary database with a physical standby in a separate datacenter. It is all setup and it works beautifully. On our primary database, we currently have 2 different types of backups we are doing - we do an export of the main schema (all of the application data is all in this one schema) 4 times a day, and we do a full database hot backup once a night.
    My question is in regards to the hot backup - I don't know that it is even worth doing a hot backup of this database? I am trying to think of a situation where we would actually want to restore a hot backup of the primary database... If we ran into some kind of a data issue, it would probably be quickest and easiest to restore data from one of the exports, and when we did that restore (import), I assume that data change would be replicated through DataGuard to the standby site. But if there was some kind of situation where we wanted to restore a recent hot backup of the primary database, that would essentially break the Data Guard configuration, and I assume that after the hot backup was restored, we would have to somehow re-instantiate Data Guard on the standby site.
    Does anyone have any input on this? If you are running with DataGuard, is it even worth it to be doing hot backups? What kind of situation would call for restoring a hot backup, instead of just failing over to the standby?
    Thanks!
    --Brad                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

    If we ran into some kind of a data issue, it would probably be quickest and easiest to restore data from one of the exportsIt would be quickest to failover to stand by database than restroing from dump file. After all you maintaing stand by db for that reason.
    How can you restore database up to the latest changes using export/import ? You may have to restore using rman and apply the logs.
    You can backup from stand by database. You do not need to back up primary.

  • DATA Guard Logical Standby v.s. Streams Apply (10gR2, rac -- rac )

    Greetings -
    We currently have a 3 node cluster, doing a 'Schema' capture, with a 2 node cluster serving as the apply side. Both are on 10gR2 (solaris 10)
    We have been experiencing apply latency due to large transactions. The way logminer/streams evaluates the arch logs, it converts each updated row into a transaction as part of a transaction set, using 'decode (' statements.
    I am under the impression a physical standby will do the same thing. But what about a logical Standby ?
    (this from the Oracle Documentation)
    [10g Concepts|http://download.oracle.com/docs/cd/B19306_01/server.102/b14239/concepts.htm#SBYDB00010]
    ...Contains the same logical information as the production database, although the physical organization and structure of the data can be different. The logical standby database is kept synchronized with the primary database though SQL Apply, which transforms the data in the redo received from the primary database into SQL statements and then executing the SQL statements on the standby database.
    Does anyone know firsthand if SQL APPLY re-issues the original sql, or if it relies on the 'decode' command as well ?
    Thanks
    The Nets Edge

    A Physical standby does not do anything with SQL. It is running media recovery under the covers. A Logical standby uses the Logminer to read the redo, converts it to SQL and Data and applies those transactions to the Logical standby.
    The both build logical change records, figure out order, dependencies etc etc and applies the transaction. Both are susceptible to long running transactions on the Primary (source) database.
    Streams uses Logminer, Logical Standby use part of the Streams apply so as Mr. Morgan said, they are very similar :^)
    If you are having apply performance issues you may want to look at Active data Guard.
    Larry

Maybe you are looking for

  • Asset retirement - Error - AA416 & AA324

    Dear All, I am getting following error while doing asset retirement through t-code F-92. My legacy transfer date is 31.03.2009 I am retiring last year legacy asset. For few of the asset class I am getting this error. For few of the asset I am not get

  • Adding Custom fields to VL02N...very urgent

    Hi Friends, I have come up with a requirement, where i need to add 2 custom fields namely "EOD INDICATOR and FREIGHT AMOUNT" in a customized additional tab at the header level in VL02N transaction.I need to update the freight amount in the deliveries

  • Supports Design View in Linux ?

    Hi All, am new to Flex and am using ubuntu os(Linux) . I have downloaded and installed Flex Builder alpha version for Linux and i read that Linux doesn't support Design View and some other views for Flex Builder from the following URL: http://labs.ad

  • HR IDOCS - Restrict IDOC Authorization

    Hi Gurus, I need some guidance and recomedations on how to secure IDOC authorization so users won't be able to view HR data. I have the parameters for  the LOGICAL MESSAGE  and PARTNER TYPE.  All need is some input on how to go about doing this. Than

  • I purchased Illustrator 8.0 on ebay and apparently it is not compatable with windows 7, is there anything I can do to use this program?

    I purchased illustrator 8.0 on ebay and apparently it is not compatible with windows 7.  Is there anything I can do to use this program?