Data gaurd Standby logs

Hi,
I have been configuring standby on a test machine, I am following document.
It has, standby redo log files creation in Primary database using following statement.
SQL> ALTER DATABASE ADD STANDBY LOGFILE GROUP 4 SIZE 50M;
Database altered.
I know the importance standby redo log files in Standby database but why we need to create them in Primary database.
Also,
Why we need to create number of oneline redo log files + 1 for standby rego log files.
Thanks!

Standby Redo Logs are required if you are using Real-Time Apply on Standby Database, or your standby standby database is in maximum protection or maximum availability mode. In addition you need them if using Cascading Destinations.
But why we need to create them in Primary database?
Although the standby redo log is only used when the database is running in the standby role, Oracle recommends that you create a standby redo log on the primary database so that the primary database can switch over quickly to the standby role without the need for additional DBA intervention.
Why we need to create number of oneline redo log files + 1 for standby rego log files?
This question is a little hard to understand. I would create at least 3 standby redo log files. Standby Redo Logs are similar to Online Redo Logs and only difference between two is that Standby Redo Log is used to store redo data received from another database (primary database).
Best Regards
mseberg

Similar Messages

  • Unable to ship Redo logs to standby DB (Oracle Data Gaurd)

    Hi all,
    We have configured Oracle Data Gaurd between our Production (NPP) & Standby (NPP_DR).
    The configuration is complete however, the production is unable to ship redo logs to standby DB.
    We keep getting the error "PING[ARC0]: Heartbeat failed to connect to standby 'NPP_DR'. Error is 12154." in Primary DB
    Primary & DR are on different boxes.
    Please see the logs below in the production alert log file & npp_arc0_18944.trc trace files:
    npp_arc0_18944.trc :*
    2011-01-19 09:17:38.007 62692 kcrr.c
    Error 12154 received logging on to the standby
    Error 12154 connecting to destination LOG_ARCHIVE_DEST_2 standby host 'NPP_DR'
    Error 12154 attaching to destination LOG_ARCHIVE_DEST_2 standby host 'NPP_DR'
    2011-01-19 09:17:38.007 62692 kcrr.c
    PING[ARC0]: Heartbeat failed to connect to standby 'NPP_DR'. Error is 12154.
    2011-01-19 09:17:38.007 60970 kcrr.c
    kcrrfail: dest:2 err:12154 force:0 blast:1
    2011-01-19 09:22:38.863
    Redo shipping client performing standby login
    OCIServerAttach failed -1
    .. Detailed OCI error val is 12154 and errmsg is 'ORA-12154: TNS:could not resolve the connect identifier specified
    alert log file on Primary*
    Error 12154 received logging on to the standby
    Wed Jan 19 09:02:35 2011
    Error 12154 received logging on to the standby
    Wed Jan 19 09:07:36 2011
    Error 12154 received logging on to the standby
    Wed Jan 19 09:12:37 2011
    Error 12154 received logging on to the standby
    Wed Jan 19 09:13:10 2011
    Incremental checkpoint up to RBA [0x2cc.2fe0.0], current log tail at RBA [0x2cc.2fe9.0]
    Wed Jan 19 09:17:38 2011
    Error 12154 received logging on to the standby
    Wed Jan 19 09:22:38 2011
    Error 12154 received logging on to the standby
    Wed Jan 19 09:27:39 2011
    Error 12154 received logging on to the standby
    However, we are able to tnsping from primary to DR
    Tnsping Results
    From Primary:
    juemdbp1:oranpp 19> tnsping NPP_DR
    TNS Ping Utility for HPUX: Version 10.2.0.4.0 - Production on 19-JAN-2011 09:32:50
    Copyright (c) 1997,  2007, Oracle.  All rights reserved.
    Used parameter files:
    /oracle/NPP/102_64/network/admin/sqlnet.ora
    Used TNSNAMES adapter to resolve the alias
    Attempting to contact (DESCRIPTION = (ADDRESS_LIST = (ADDRESS = (COMMUNITY = SAP.WORLD) (PROTOCOL = TCP) (HOST = 10.80.51.101) (PORT = 49160))) (CONNECT_DATA = (SID = NPP) (SERVER = DEDICATED)))
    OK (60 msec)
    Tnsnames.ora in Primary:
    Filename......: tnsnames.ora
    Created.......: created by SAP AG, R/3 Rel. >= 6.10
    Name..........:
    Date..........:
    @(#) $Id: //bc/700-1_REL/src/ins/SAPINST/impl/tpls/ora/ind/TNSNAMES.ORA#4 $
    NPP.WORLD=
      (DESCRIPTION =
        (ADDRESS_LIST =
            (ADDRESS =
              (COMMUNITY = SAP.WORLD)
              (PROTOCOL = TCP)
              (HOST = nppjorp)
              (PORT = 49160)
        (CONNECT_DATA =
           (SID = NPP)
           (GLOBAL_NAME = NPP.WORLD)
    NPP_HQ.WORLD=
    (DESCRIPTION =
       (ADDRESS_LIST =
            (ADDRESS =
              (COMMUNITY = SAP.WORLD)
              (PROTOCOL = TCP)
              (HOST = nppjorp)
              (PORT = 49160)
       (CONNECT_DATA =
            (SID = NPP)
            (SERVER = DEDICATED)
    NPP_DR.WORLD=
    (DESCRIPTION =
      (ADDRESS_LIST =
       (ADDRESS =
            (COMMUNITY = SAP.WORLD)
            (PROTOCOL = TCP)
            (HOST = 10.80.51.101)
            (PORT = 49160)
    (CONNECT_DATA =
            (SID = NPP)
            (SERVER = DEDICATED)
    NPPLISTENER.WORLD=
            (DESCRIPTION =
              (ADDRESS =
                    (PROTOCOL = TCP)
                    (HOST = nppjorp)
                    (PORT = 49160)
    Listener.ora in Primary
    Filename......: listener.ora
    Created.......: created by SAP AG, R/3 Rel. >= 6.10
    Name..........:
    Date..........:
    @(#) $Id: //bc/700-1_REL/src/ins/SAPINST/impl/tpls/ora/ind/LISTENER.ORA#4 $
    ADMIN_RESTRICTIONS_LISTENER = on
    LISTENER =
      (ADDRESS_LIST =
            (ADDRESS =
              (PROTOCOL = IPC)
              (KEY = NPP.WORLD)
            (ADDRESS=
              (PROTOCOL = IPC)
              (KEY = NPP)
            (ADDRESS =
              (COMMUNITY = SAP.WORLD)
              (PROTOCOL = TCP)
              (HOST = nppjorp)
              (PORT = 49160)
    STARTUP_WAIT_TIME_LISTENER = 0
    CONNECT_TIMEOUT_LISTENER = 10
    TRACE_LEVEL_LISTENER = OFF
    SID_LIST_LISTENER =
      (SID_LIST =
        (SID_DESC =
          (SID_NAME = NPP)
          (ORACLE_HOME = /oracle/NPP/102_64)
    Thank You,
    Salman Qayyum
    Edited by: Salman M.A. Qayyum on Jan 19, 2011 8:12 AM

    Hi,
    Please find the remaining post ...
    Tnsnames.ora in DR:
    Filename......: tnsnames.ora
    Created.......: created by SAP AG, R/3 Rel. >= 6.10
    Name..........:
    Date..........:
    @(#) $Id: //bc/700-1_REL/src/ins/SAPINST/impl/tpls/ora/ind/TNSNAMES.ORA#4 $
    NPP.WORLD=
      (DESCRIPTION =
        (ADDRESS_LIST =
            (ADDRESS =
              (COMMUNITY = SAP.WORLD)
              (PROTOCOL = TCP)
              (HOST = nppjor)
              (PORT = 49160)
        (CONNECT_DATA =
           (SID = NPP)
           (GLOBAL_NAME = NPP.WORLD)
    NPP_HQ.WORLD=
      (DESCRIPTION =
        (ADDRESS_LIST =
          (ADDRESS =
            (COMMUNITY = SAP.WORLD)
            (PROTOCOL = TCP)
            (HOST = hq_nppjor)
            (PORT = 49160)
        (CONNECT_DATA =
            (SID = NPP)
            (SERVER = DEDICATED)
    NPP_DR.WORLD=
      (DESCRIPTION =
        (ADDRESS_LIST =
          (ADDRESS =
            (COMMUNITY = SAP.WORLD)
            (PROTOCOL = TCP)
            (HOST = nppjor)
            (PORT = 49160)
        (CONNECT_DATA =
            (SID = NPP)
            (SERVER = DEDICATED)
            (SERVICE_NAME = NPP_DR)
    NPPLISTENER.WORLD=
            (DESCRIPTION =
              (ADDRESS =
                    (PROTOCOL = TCP)
                    (HOST = nppjor)
                    (PORT = 49160)
    Listener.ora in DR
    Filename......: listener.ora
    Created.......: created by SAP AG, R/3 Rel. >= 6.10
    Name..........:
    Date..........:
    @(#) $Id: //bc/700-1_REL/src/ins/SAPINST/impl/tpls/ora/ind/LISTENER.ORA#4 $
    ADMIN_RESTRICTIONS_LISTENER = on
    LISTENER =
      (ADDRESS_LIST =
            (ADDRESS =
              (PROTOCOL = IPC)
              (KEY = NPP.WORLD)
            (ADDRESS=
              (PROTOCOL = IPC)
              (KEY = NPP)
            (ADDRESS =
              (COMMUNITY = SAP.WORLD)
              (PROTOCOL = TCP)
              (HOST = nppjor)
              (PORT = 49160)
           (ADDRESS =
             (COMMUNITY = SAP.WORLD)
             (PROTOCOL = TCP)
             (HOST = 10.80.50.101)
             (PORT = 49160)
    STARTUP_WAIT_TIME_LISTENER = 0
    CONNECT_TIMEOUT_LISTENER = 10
    TRACE_LEVEL_LISTENER = OFF
    SID_LIST_LISTENER =
      (SID_LIST =
        (SID_DESC =
          (SID_NAME = NPP)
          (ORACLE_HOME = /oracle/NPP/102_64)
    /etc/hosts settings in Primary
    host:oranpp 25> grep nppjor /etc/hosts
    10.32.243.54    nppjor.sabic.com        nppjor
    10.32.50.115    nppjorp.sabic.com        nppjorp
    /etc/hosts settings in DR
    host:oranpp 11> grep nppjor /etc/hosts
    10.32.243.54    hq_nppjor.sabic.com     hq_nppjor
    10.80.243.54    nppjor.sabic.com        nppjor
    10.80.50.115    nppjorp.sabic.com        nppjorp
    Thank You,
    Salman Qayyum

  • Convert Standyby to Primary and Primary to Standby using Data Gaurd

    What are the steps involved to do the following.
    Convert Standyby to Primary and Primary to Standby databse server using Data Gaurd.
    Pls help.

    That will be a role Switchovers.
    Check the Oracle document and follow the steps.
    http://download-west.oracle.com/docs/cd/B19306_01/server.102/b14239/role_management.htm#i1030646

  • EM console HA page says "Data Gaurd not configured for this database"

    I am attempting to use EM console's "High Availability Console" to create a standby database, but, the option is not available. Under "Data Gaurd Summary" the page shows "Oracle DataGaurd is not configured on this database."
    How do I configure this option so that I can use the console to create Standy databases? I have looked everywhere I can think of and cannot find any instructions!

    Click on your database name, and you'll be lead to a page which shows you all the statistics of your Db
    Click on the "Availability" Sub Tab.
    You'll see the option to "Add Standby Database". You can take it from there. (you'll be prompted to log in to your database -- log in with sys)
    Now click on "Add Standby Database"
    The process I normally follow is this
    1. RMAN your primary
    2. Create standby control file from primary.
    3. Get your init file from primary
    4. Copy everything over.
    5. Make the init file changes (fal client, fal server on primary, any kind of directory mapping, point the init file to the standby control file on the standby side)
    6. Bring up the standby db (doing all the other changes that are required)
    Then you can import that Db into OEM,
    or you could just use the Wizard from OEM :)

  • Standby log files in Oracle Dataguard

    Hi,
    What is the difference between standby log files and online redo log files in a Dataguard environment?
    What is the use of standby log files?
    Thanks,
    Charith.

    You're probably familiar with the Online Redo Logs (ORLs). Transaction changes are written from the Log Buffer to the ORLs by the LGWR process.
    If you are setting up a physical standby, then you will want to create Standby Redo Logs (SRLs) in the standby database. When SRL's are in place, a process called LNS will transport redo from the Log Buffer of the primary to the RFS process on the standby which will write the redo to the SRLs. If the SRL does not exist, RFS can't do this. The biggest benefit of using SRLs is that you will experience much less data loss, even in MAX PERFORMANCE mode. Redo will constantly be shipped. You won't have to wait for ARCH to transport a full archived redo log.
    Cheers,
    Brian

  • Data Guard - Redo log files

    Hi....
    I have done data guard .......every thing is fine.......archives are bring transferred to standby..........
    Also, during configuration, I had created standby redolog groups 4,5,6 and copied to standby.....
    But in real time apply.......the standby is not using standby redolog groups 4,5,6........when i am query v$log it is showing group 1,2,3.
    Actually, its should use standby redo log for maximum availability.
    Please help....
    Thanks in advance.

    There was a similar question here just a few days ago:
    Data Guard - redo log files

  • Is this possible in Oracle Data Gaurd 11g?

    Dear All,
    I was looking for replication scenerio while this thing came in my mind.
    We have setup 2 database servers on 2 different location, thousands of miles apart. Both servers have oracle 11gR1 and are on linux operating system.
    We have been facing connectivity problems as users are accessing our application connected to one of these servers from all over the world and when there is connectivity problems applications are no more accessible.
    I want to setup a stand by database server, like if there is connectivity problem on one server, or its down due to any reason applications are automatically shift to the second available server. The primary server is obvioulsy the first server and in case of no connectivity or any other problem that should bein use.
    In this case we have to keep an exact copy of the primary server on second server. I want to know if Oracle data Gaurd will provide us such a functionality when the servers are on remote locations? How the data will be copied to secondry server?
    Kindly let me know if there is any document available to setup such a scenerio. I will be thankful to you.
    Regards,
    Imran Baig

    DataGuard Physical standby will keep a copy of your database on a standby server. Utilizing fast start failover with an observer (3rd party) that sees both the primary and standby server will give you the capability to automatically failover to the standby server should connectivity to the primary be interrupted. Fast start failover is configured with a threshold that will determine the amount of time it waits for lost connectivity prior to actually doing the failover allowing you to configure the threshold prior to failover. Your application accessing the database will need to reconnect to the new primary after failover.

  • Standby log files

    I want to convert our old script based primary /standby database into a dataguard config using the LGWR as log transport.
    I already have old log files on the standby database, but the data in them is from 2004. Not entirely interesting since the database gets recovered from the arch log files every night.
    Point is, can I use these as standby log files, or do I have to (somehow) drop these and re-create new standby logfiles. I cant drop them anyway since when I try, I get "ORA-01624, log 1 needed for crash recovery". (Like h*ll, since the data is older than Noah).
    Will these just get re-written?
    null
    null

    Note:219344.1 This note from metalink gives "Usage, Benefits and Limitations of Standby Redo Logs (SRL)".
    Standby Redo Logs are only supported for the Physical Standby Database in
    Oracle 9i and as well for Logical Standby Databases in 10g. Standby Redo Logs
    are only used if you have the LGWR activated for archival to the Remote Standby
    Database.
    The great Advantage of Standby Redo Logs is that every Entry written into
    the Online RedoLogs of the Primary Database is transfered to the Standby
    Site and written into the Standby Redo Logs at the same time; threfore, you
    reduce the probability of Data Loss on the Standby Database.

  • I'm a bit confused about standby log files

    Hi all,
    I'm a bit confused about something and wondering if someone can explain.
    I have a Primary database that ships logs to a Logical Standby database.
    Everything appears to be working properly. If I check the v$archived_log table in the Primary and compare it to the dba_logstdby_log view in the Logical Standby, I'm seeing that logs are being applied.
    On the logical standby, I have the following configured for log_archive_dest_n parameters:
    *.log_archive_dest_1='LOCATION=/u01/oracle/archivedlogs/ORADB1
    VALID_FOR=(ALL_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=PNX8A_GMD'
    *.log_archive_dest_2='LOCATION=/u02/oracle/archivedlogs/ORADB1
    VALID_FOR=(ALL_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=PNX8A_GMD'
    *.log_archive_dest_3='LOCATION=/u03/oracle/archivedlogs/ORADB1
    VALID_FOR=(ALL_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=PNX8A_GMD'
    *.log_archive_dest_4='SERVICE=PNX8A_WDC ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=PNX8A_WDC'
    *.log_archive_dest_5='LOCATION=/u01/oracle/standbylogs/ORADB1
    VALID_FOR=(STANDBY_LOGFILES,STANDBY_ROLE) DB_UNIQUE_NAME=PNX8A_GMD'
    *.log_archive_dest_6='LOCATION=/u02/oracle/standbylogs/ORADB1
    VALID_FOR=(STANDBY_LOGFILES,STANDBY_ROLE) DB_UNIQUE_NAME=PNX8A_GMD'
    *.log_archive_dest_7='LOCATION=/u03/oracle/standbylogs/ORADB1
    VALID_FOR=(STANDBY_LOGFILES,STANDBY_ROLE) DB_UNIQUE_NAME=PNX8A_GMD'
    Here is my confusion now. Before converting from a Physical standby database to a Logical Standby database, I was under the impression that I needed the standby logs (i.e. log_archive_dest_5, 6 and 7 above) because a Physical Standby database would receive the redo from the primary and write it into the standby logs before applying the redo in the standby logs to the Physical standby database.
    I've now converted to a Logical Standby database. What's happening is that the standby logs are accumulating in the directory pointed to by log_archive_dest_6 above (/u02/oracle/standbylogs/ORADB1). They do not appear to be getting cleaned up by the database.
    In the Logical Standby database I do have STANDBY_FILE_MANAGEMENT parameter set to AUTO. Can anyone explain to me why standby log files would continue to accumulate and how I can get the Logical Standby database to remove them after they are no longer needed on the LSB db?
    Thanks in advance.
    John S

    JSebastian wrote:
    I assume you mean in your question, why on the standby database I am using three standby log locations (i.e. log_archive_dest_5, 6, and 7)?
    If that is your question, my answer is that I just figured more than one location would be safer but I could be wrong about this. Can you tell me if only one location should be sufficient for the standby logs? The more I think of this, that is probably correct because I assume that Log Transport services will re-request the log from the Primary database if there is some kind of error at the standby location with the standby log. Is this correct?As simple configure as below. Why more multiple destinations for standby?
    check notes Step by Step Guide on How to Create Logical Standby [ID 738643.1]
    >
    LOG_ARCHIVE_DEST_1='LOCATION=/arch1/boston VALID_FOR=(ONLINE_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=boston'
    LOG_ARCHIVE_DEST_2='SERVICE=chicago LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=chicago'
    LOG_ARCHIVE_DEST_3='LOCATION=/arch2/boston/ VALID_FOR=(STANDBY_LOGFILES,STANDBY_ROLE) DB_UNIQUE_NAME=boston'
    The following table describes the archival processing defined by the initialization parameters shown in Example 4-2.
         When the Boston Database Is Running in the Primary Role      When the Boston Database Is Running in the Logical Standby Role
    LOG_ARCHIVE_DEST_1      Directs archival of redo data generated by the primary database from the local online redo log files to the local archived redo log files in /arch1/boston/.      Directs archival of redo data generated by the logical standby database from the local online redo log files to the local archived redo log files in /arch1/boston/.
    LOG_ARCHIVE_DEST_2      Directs transmission of redo data to the remote logical standby database chicago.      Is ignored; LOG_ARCHIVE_DEST_2 is valid only when boston is running in the primary role.
    LOG_ARCHIVE_DEST_3      Is ignored; LOG_ARCHIVE_DEST_3 is valid only when boston is running in the standby role.      Directs archival of redo data received from the primary database to the local archived redo log files in /arch2/boston/.
    >
    Source:-
    http://docs.oracle.com/cd/B19306_01/server.102/b14239/create_ls.htm

  • Data gaurd replication

    hi all
    i have a doubt in replication method in Data gaurd
    I have configured data gaurd for my 11g database.Evrything is working fine. I want to empty tables in standby or may be drop the schema and re create them with empty tables.
    And after that is it possible to duplicate the primary again and sync it up with primary.
    if i ponder on this i can use data pump to load the data, but how will i sync it with primary.
    thanks in advance...
    feel free to ask questions...

    >
    And after that is it possible to duplicate the primary again and sync it up with primary.
    That means you simply want to recreate your Standby from Primary which is not impacted by whatever changes you made earlier to the Standby. If my understanding to your statement is correct, you are duplicating Primary again to create standby. So the changes you made in Standby is no longer available because it is recreated.

  • Data Gaurd Modes

    1.What is maximum availability mode?
    2.What is the maximum protection mode?
    which one is highest mode

    Hi
    Maximum availability=this mode is mixture of maximum protection and maximum performance mode.Primary database dont tell operation completed to end user until write redo data both primary online log and standby online log file.if it can not write redo data to standby redo log during the time specified by net_timeout parameter it operates as in maximum performance
    Maximum protection=this mode ensures that no data loss while  primary database is fail.Primary database dont tell operation completed to end user until write redo data both primary online log and standby online log file.primary database has failed if primary database can not write redo information to at least one standby database logfile.

  • Need suggestion to get data from change log table of ODS.

    Hello,
    There is a case where i am loading opportunity header data from header ODS and opportunity item data from item ODS in the opportunity cube.
    Status (1= OPEN, 2= WON ETC) of the opportunity are available only in header ODS and not in item ODS.
    While loading data from header ODS to cube, I am loading it directly but while loading data from item ODS to cube i am using active data table of header ODS as a lookup in the update rule from item ODS to cube. I am selecting status from the active data table of header ODS while loading data from item ODS to cube.
    Since active data table will have only after image records, there is some data mismatch in the report as i am selecting data from active data table of header ODS while loading data from item ODS to cube.
    I need to select data from Change log in order to get before image also instead of active data table in order to overcome this issue. Is there any way by which i can do selection from Change log instead of active data table as change logs are generated at run time.
    Please let me know if you have any suggestions.
    Regards,
    Sanjay Chaurasia.

    Hi,
    You can use the changelog table of the DSO.
    Right click manage the Header DSO, go to the contents tab and click Change Log table. There you can see the technical name of the Change Log table.
    In the update rule Routine, give the tech name of Change log table instead of Active table name.
    Hope it helps.
    Krishna

  • Read data from a log file to flex textarea while deploying

    Hi,
    I am new to flex. I am able to read data from a log file that placed in C drive and able to write that data into flex textarea, but I am unable to read the same file while it is deployed into weblogic sever.
    Could anyone please tell me how to solve this problem.
    Thanks,
    Sri.

    Do you have it trying to read that same file on your C drive or is the file actually deployed somewhere on the server?

  • How to transfer data in change log table of dso to z-table using abap code

    Hi  can you please explain me how to transfer data in change log table of dso to z-table using abap code ,with out using Function module concept

    PROGRAM NAME:   ZBW_DELTA_TO_GSTAR                                 **
    report ZBW_DELTA_TO_GSTAR no standard page heading
                                     line-size 120
                                     line-count 75
                                     message-id ZBW_MSG_CLS.
    tables:   ZGIV_DLTA_EBV_BB,
              ZGIV_DLTA_EM2_BL,
              ZGIV_DLTA_EM2_BK.
    Selection Screen Definitions
    SELECTION-SCREEN: BEGIN OF BLOCK INNER WITH FRAME TITLE TEXT-001.
    SELECTION-SCREEN: SKIP 1.
    PARAMETERS:       EBVBB RADIOBUTTON GROUP ROLL,
                      EM2BL RADIOBUTTON GROUP ROLL,
                      EM2BK RADIOBUTTON GROUP ROLL.
    SELECTION-SCREEN: END OF BLOCK INNER.
    Data:  WS_UPDATE_FLAG  Type C,
           UCounter(9)      Type N,
           ICounter(9)      Type N.
    DATA:  T_ZGIV_DLTA_EBV_BB Type Standard Table of ZGIV_DLTA_EBV_BB,
           s_ZGIV_DLTA_EBV_BB LIKE line of T_ZGIV_DLTA_EBV_BB.
    DATA:  T_ZGIV_DLTA_EM2_BK Type Standard Table of ZGIV_DLTA_EM2_BK,
           s_ZGIV_DLTA_EM2_BK LIKE line of T_ZGIV_DLTA_EM2_BK.
    DATA:  T_ZGIV_DLTA_EM2_BL Type Standard Table of ZGIV_DLTA_EM2_BL,
           s_ZGIV_DLTA_EM2_BL LIKE line of T_ZGIV_DLTA_EM2_BL.
    Standard Internal Tables - Describe usage.
    data: begin of i_AEPSD_O0140 occurs 0.
            include structure /BIC/AEPSD_O0140.
    data: end of i_AEPSD_O0140.
    data: begin of i_AEPSD_O0240 occurs 0.
            include structure /BIC/AEPSD_O0240.
    data: end of i_AEPSD_O0240.
    data: begin of i_AEPSD_O0340 occurs 0.
            include structure /BIC/AEPSD_O0340.
    data: end of i_AEPSD_O0340.
    data: begin of i_GIV_DLTA_EBV_BB occurs 0.
            include structure ZGIV_DLTA_EBV_BB.
    data: end of i_GIV_DLTA_EBV_BB.
    data: begin of i_GIV_DLTA_EM2_BK occurs 0.
            include structure ZGIV_DLTA_EM2_BK.
    data: end of i_GIV_DLTA_EM2_BK.
    data: begin of i_GIV_DLTA_EM2_BL occurs 0.
            include structure ZGIV_DLTA_EM2_BL.
    data: end of i_GIV_DLTA_EM2_BL.
    Miscellaneous Program Variables and Constants.
    TOP-OF-PAGE
    top-of-page.
    START-OF-SELECTION
    start-of-selection.
      Clear: i_GIV_DLTA_EBV_BB,
             i_GIV_DLTA_EM2_BK,
             i_GIV_DLTA_EM2_BL,
             UCounter, ICounter.
      IF EBVBB = 'X'.
        PERFORM 100_EXTRACT_EBV_BB_DELTA_RECS.
      ELSEIF EM2BK = 'X'.
        PERFORM 100_EXTRACT_EM2_BK_DELTA_RECS.
      ELSE.
        PERFORM 100_EXTRACT_EM2_BL_DELTA_RECS.
      ENDIF.
    FORM 100_EXTRACT_EBV_BB_DELTA_RECS
    FORM 100_EXTRACT_EBV_BB_DELTA_RECS.
      Refresh:   i_AEPSD_O0140,
                 i_GIV_DLTA_EBV_BB.
      Clear:      UCounter, ICounter, s_ZGIV_DLTA_EBV_BB .
      Select * From /BIC/AEPSD_O0140
        Into TABLE i_AEPSD_O0140.
      IF SY-Subrc = 0.
        LOOP AT i_AEPSD_O0140.
          MOVE-CORRESPONDING i_AEPSD_O0140 TO s_ZGIV_DLTA_EBV_BB.
          MOVE SY-DATUM to s_ZGIV_DLTA_EBV_BB-create_dt.
          INSERT ZGIV_DLTA_EBV_BB FROM s_ZGIV_DLTA_EBV_BB.
          IF SY-Subrc = 0.
            ICounter = ICounter + 1.
          ELSE.
            UPDATE ZGIV_DLTA_EBV_BB FROM  s_ZGIV_DLTA_EBV_BB.
            IF SY-Subrc = 0.
              UCounter = UCounter + 1.
            ELSE.
              Message E067 with SY-DATUM ' ' SY-UZEIT ' '.
            ENDIF.
          ENDIF.
        ENDLOOP.
      ENDIF.
    ENDFORM.                    "100_EXTRACT_EBV_BB_DELTA_RECS
    FORM 100_EXTRACT_EM2_BK_DELTA_RECS
    FORM 100_EXTRACT_EM2_BK_DELTA_RECS.
    Refresh:   i_AEPSD_O0240,
               i_GIV_DLTA_EM2_BK.
      Clear:      UCounter, ICounter, s_ZGIV_DLTA_EM2_BK .
      Select * From /BIC/AEPSD_O0240
        Into TABLE i_AEPSD_O0240.
      IF SY-Subrc = 0.
        LOOP AT i_AEPSD_O0240.
          MOVE-CORRESPONDING i_AEPSD_O0240 TO s_ZGIV_DLTA_EM2_BK.
          MOVE SY-DATUM to s_ZGIV_DLTA_EM2_BK-create_dt.
            INSERT ZGIV_DLTA_EM2_BK FROM s_ZGIV_DLTA_EM2_BK.
          IF SY-Subrc = 0.
            ICounter = ICounter + 1.
          ELSE.
            UPDATE ZGIV_DLTA_EM2_BK FROM  s_ZGIV_DLTA_EM2_BK.
            IF SY-Subrc = 0.
              UCounter = UCounter + 1.
            ELSE.
              Message E067 with SY-DATUM ' ' SY-UZEIT ' '.
            ENDIF.
          ENDIF.
        ENDLOOP.
      ENDIF.
    ENDFORM.                    "100_EXTRACT_EM2_BK_DELTA_RECS
    FORM 100_EXTRACT_EM2_BL_DELTA_RECS
    FORM 100_EXTRACT_EM2_BL_DELTA_RECS.
    Refresh:   i_AEPSD_O0340,
               i_GIV_DLTA_EM2_BL.
      Clear:      UCounter, ICounter, s_ZGIV_DLTA_EM2_BL .
      Select * From /BIC/AEPSD_O0340
        Into TABLE i_AEPSD_O0340.
      IF SY-Subrc = 0.
        LOOP AT i_AEPSD_O0340.
          MOVE-CORRESPONDING i_AEPSD_O0340 TO s_ZGIV_DLTA_EM2_BL.
          MOVE SY-DATUM to s_ZGIV_DLTA_EM2_BL-create_dt.
            INSERT ZGIV_DLTA_EM2_BL FROM s_ZGIV_DLTA_EM2_BL.
          IF SY-Subrc = 0.
            ICounter = ICounter + 1.
          ELSE.
            UPDATE ZGIV_DLTA_EM2_BL FROM  s_ZGIV_DLTA_EM2_BL.
            IF SY-Subrc = 0.
              UCounter = UCounter + 1.
            ELSE.
              Message E067 with SY-DATUM ' ' SY-UZEIT ' '.
            ENDIF.
          ENDIF.
        ENDLOOP.
      ENDIF.
    ENDFORM.                    "100_EXTRACT_EM2_BL_DELTA_RECS
    END-OF-SELECTION
    end-of-selection.
      perform D1000_REPORT_DATA.
    D1000_REPORT_DATA
    form D1000_REPORT_DATA.
    *Display the title of the program
      write: /25 SY-TITLE.
      skip.
    Diaplay the details of the user and time
      write: /1 'Executed by', 15 SY-UNAME, 30 'Date',
      38 SY-DATUM, 53 'Time', 60 SY-UZEIT.
      skip 2.
      write: /  'Delta Records have been extracted  ',
             /   'Updates : ', UCounter,
             /   'Inserts : ', ICounter.
      skip.
      skip 3.
      write: /20 'End of the report'.
    endform.                                           "D1000_REPORT_DATA
    chgeck it out this also may hep you

  • I need to move my "auto-complete" data to my log-in on the companies computer.

    I need to move my "auto-complete" data to my log-in on the companies computer.
    How do i do this? Replicating the date

    See:
    * http://kb.mozillazine.org/Transferring_data_to_a_new_profile_-_Firefox
    * [[Recovering important data from an old profile]]

Maybe you are looking for

  • How to get the value retruned by java script function into my jsp page

    Hai all, I had a particular java script function which returns a date. function getDate() {      var sDate; // This code executes when the user clicks on a day in the calendar. if ("TD" == event.srcElement.tagName) // Test whether day is valid. if ("

  • Can't grant privilege on column to user via role?

    Hi: From what I read in the docs I should be able to create a role that has UPDATE privs on a column of a table, and then grant that role to a user, who should be able to update the column of the table. I get "insufficient privileges" when I try that

  • MS ODBC error SQL_HANDLE_ENV

    I've just installed Oracle 9i personal edition and developer suite. Also installed Oracle ODBC drivers. Trying to set up a system DSN I get an error: SQL_HANDLE_ENV. Looked up on Google, says Net8 wasn't installed. Also says, however, that Net8 is in

  • Complex (Indian) scripts not rendering in latest build of Jdk8

    Complex (Indian) scripts not rendering in JavaFx Text components of latest build of JDK8. I am using early access build b89 . I remember it was working in earlier JDK8 release when the button components were not bigger.

  • CS4 & CS5 Indy with same serial number

    Can CS4 & CS5 Indy with same serial number both be open on the same machine? Thanks!