Generating lots of archive logs

Hi Friends,
We have an EBS 11i on AIX 5L ...which has just been setup and ready for UAT...but the AppsDBA/Functional Consultant who didi it are not around anymore to ask for questiones. I just noticed the there are archive logs generated everyday...like 30 logs...when in fact the application is not being used. Is there concurrent programs
that has been setup to update data on a background process, just like recursive updating which is not really necessary. How do I check if there are updates being done.
Thanks a lot

Do not stop this concurrent program as it is used to synchronize the Workflow local tables with the user and role information stored in the product application tables until each affected product performs the synchronization automatically.
More details can be found in the following note:
Note: 171703.1 - 11.5.x: Implementing Oracle Workflow Directory Service Synchronization
https://metalink2.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=171703.1
Did you check the total size of the log files? I believe you should not be worried now until the system is delivered to the users, you can monitor the number of log files generated daily then and based on that start your investigation.

Similar Messages

  • OID (iasdb) 9.0.1.4.0 generating tons of archive logs

    Hi,
    When the tspurge jobs run, from dba_jobs it looks like there is a lot of archive logs being generated. Is this normal ? If so, are records being deleted from certain log and audit tables (PORTAL, ORASSO etc). I have a situation when the tspurgeDirObject ('cn=secrefresh events purgeconfig' ); END; runs, 125MB archive logs are generated every 2 minutes and is filling up disk space.
    Thanks for your help.
    Ramesh.

    Hi,
    When the tspurge jobs run, from dba_jobs it looks like there is a lot of archive logs being generated. Is this normal ? If so, are records being deleted from certain log and audit tables (PORTAL, ORASSO etc). I have a situation when the tspurgeDirObject ('cn=secrefresh events purgeconfig' ); END; runs, 125MB archive logs are generated every 2 minutes and is filling up disk space.
    Thanks for your help.
    Ramesh.

  • Are there any possible reason RMAN generates some corrupt archive log files

    Dear all,
    Are there any possible reason RMAN generates some corrupt archive log files?
    Best Regards,
    Amy

    Because I try to perform daily backup at lunch time and found out it takes more than 1 hour had no any progress. Normally we take around 40 minus. The following is the log file:
    RMAN> Run
    2> {
    3> CONFIGURE CONTROLFILE AUTOBACKUP ON;
    4> CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO '/u03/db/backup/RMAN/%F.bck';
    5> CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 1 DAYS;
    6> allocate channel ch1 type disk format '/u03/db/backup/RMAN/backup_%d_%t_%s_%p_%U.bck';
    7> backup incremental level 1 cumulative database plus archivelog delete all input;
    8> backup current controlfile;
    9> backup spfile;
    10> release channel ch1;
    11> }
    12> allocate channel for maintenance type disk;
    13> delete noprompt obsolete;
    14> delete noprompt archivelog all backed up 2 times to disk;
    15>
    16>
    using target database controlfile instead of recovery catalog
    old RMAN configuration parameters:
    CONFIGURE CONTROLFILE AUTOBACKUP ON;
    new RMAN configuration parameters:
    CONFIGURE CONTROLFILE AUTOBACKUP ON;
    new RMAN configuration parameters are successfully stored
    old RMAN configuration parameters:
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO '/u03/db/backup/RMAN/%F.bck';
    new RMAN configuration parameters:
    CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK TO '/u03/db/backup/RMAN/%F.bck';
    new RMAN configuration parameters are successfully stored
    old RMAN configuration parameters:
    CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 1 DAYS;
    new RMAN configuration parameters:
    CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 1 DAYS;
    new RMAN configuration parameters are successfully stored
    allocated channel: ch1
    channel ch1: sid=99 devtype=DISK
    Starting backup at 31-MAR-09
    current log archived
    After that I go to archive log directory "/u02/oracle/uat/uatdb/9.2.0/dbs" and use ls -lt command to see how many archive logs and my screen just hang. After we found out that we cannot use ls -lt command to read arch1_171.dbf
    archive log, the rest of archive logs able to use ls -lt command.
    We cannot delete this file as well. We shutdown database abort and perform check disk...... and fix the disk error and then open database again. Everything seems back to normal and we can use ls -lt command to read arch1_171.dbf.
    The strange problem is we have the same problem in Development and Production..... one ore more archive logs seems to be corrupted under the same directories /u02/oracle/uat/uatdb/9.2.0/dbs.
    Does anyone encounter the same problem?
    Amy

  • OVM 3.0 Database Creating Lots of Archive Logs

    Greetings - ever since we initially installed OVM 3.0 earlier this fall (~October), the OVM database has generated archive logs at a very rapid rate. It continually threatens to fill up our 16 GB filesystem dedicated to archive logs, even after daily backup and purging.
    Our OVM database itself is about 4-6 GB large, and we would need to increase the archive log filesystem to about 20-25 GB in size, which we see as unreasonable for such a small database.
    What is causing OVM to generate so many redo logs? Our best guess is that OVM is continuously gathering guest VM CPU usage on each physical server.
    Is there a way to configure the OVM application in order to reduce the amount of redo/archive logs being created?
    We are currently running 3.0.3, having upgraded each time a 3.0.* patch was released. OVMM running on OEL 6.1, database running on latest HP-UX.

    majedian21 wrote:
    Greetings - ever since we initially installed OVM 3.0 earlier this fall (~October), the OVM database has generated archive logs at a very rapid rate. It continually threatens to fill up our 16 GB filesystem dedicated to archive logs, even after daily backup and purging.I would log an SR with Oracle Support for this, so that Development can look at it. Sounds like your environment has lots of VMs running and yes, collecting usage stats for all of those environments. However, there may be some old data from the previous versions that's causing more stats to be collected than necessary.

  • Create procedure is generating too many archive logs

    Hi
    The following procedure was run on one of our databases and it hung since there were too many archive logs being generated.
    What would be the answer? The db must remain in archivelog mode.
    I understand the nologging concept, but as I know this applies to creating tables, views, indexes and tablespaces. This script is creating procedure.
    CREATE OR REPLACE PROCEDURE APPS.Dfc_Payroll_Dw_Prc(Errbuf OUT VARCHAR2, Retcode OUT NUMBER
    ,P_GRE NUMBER
    ,P_SDATE VARCHAR2
    ,P_EDATE VARCHAR2
    ,P_ssn VARCHAR2
    ) IS
    CURSOR MainCsr IS
    SELECT DISTINCT
    PPF.NATIONAL_IDENTIFIER SSN
    ,ppf.full_name FULL_NAME
    ,ppa.effective_date Pay_date
    ,ppa.DATE_EARNED period_end
    ,pet.ELEMENT_NAME
    ,SUM(TO_NUMBER(prv.result_value)) VALOR
    ,PET.ELEMENT_INFORMATION_CATEGORY
    ,PET.CLASSIFICATION_ID
    ,PET.ELEMENT_INFORMATION1
    ,pet.ELEMENT_TYPE_ID
    ,paa.tax_unit_id
    ,PAf.ASSIGNMENT_ID ASSG_ID
    ,paf.ORGANIZATION_ID
    FROM
    pay_element_classifications pec
    , pay_element_types_f pet
    , pay_input_values_f piv
    , pay_run_result_values prv
    , pay_run_results prr
    , pay_assignment_actions paa
    , pay_payroll_actions ppa
    , APPS.pay_all_payrolls_f pap
    ,Per_Assignments_f paf
    ,per_people_f ppf
    WHERE
    ppa.effective_date BETWEEN TO_DATE(p_sdate) AND TO_DATE(p_edate)
    AND ppa.payroll_id = pap.payroll_id
    AND paa.tax_unit_id = NVL(p_GRE, paa.tax_unit_id)
    AND ppa.payroll_action_id = paa.payroll_action_id
    AND paa.action_status = 'C'
    AND ppa.action_type IN ('Q', 'R', 'V', 'B', 'I')
    AND ppa.action_status = 'C'
    --AND PEC.CLASSIFICATION_NAME IN ('Earnings','Alien/Expat Earnings','Supplemental Earnings','Imputed Earnings','Non-payroll Payments')
    AND paa.assignment_action_id = prr.assignment_action_id
    AND prr.run_result_id = prv.run_result_id
    AND prv.input_value_id = piv.input_value_id
    AND piv.name = 'Pay Value'
    AND piv.element_type_id = pet.element_type_id
    AND pet.element_type_id = prr.element_type_id
    AND pet.classification_id = pec.classification_id
    AND pec.non_payments_flag = 'N'
    AND prv.result_value <> '0'
    --AND( PET.ELEMENT_INFORMATION_CATEGORY LIKE '%EARNINGS'
    -- OR PET.element_type_id IN (1425, 1428, 1438, 1441, 1444, 1443) )
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN PET.EFFECTIVE_START_DATE AND PET.EFFECTIVE_END_DATE
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN PIV.EFFECTIVE_START_DATE AND PIV.EFFECTIVE_END_DATE --dcc
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN Pap.EFFECTIVE_START_DATE AND Pap.EFFECTIVE_END_DATE --dcc
    AND paf.ASSIGNMENT_ID = paa.ASSIGNMENT_ID
    AND ppf.NATIONAL_IDENTIFIER = NVL(p_ssn, ppf.NATIONAL_IDENTIFIER)
    ------------------------------------------------------------------TO get emp.
    AND ppf.person_id = paf.person_id
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN ppf.EFFECTIVE_START_DATE AND ppf.EFFECTIVE_END_DATE
    ------------------------------------------------------------------TO get emp. ASSIGNMENT
    --AND paf.assignment_status_type_id NOT IN (7,3)
    AND NVL(PPA.DATE_EARNED, PPA.EFFECTIVE_DATE) BETWEEN paf.effective_start_date AND paf.effective_end_date
    GROUP BY PPF.NATIONAL_IDENTIFIER
    ,ppf.full_name
    ,ppa.effective_date
    ,ppa.DATE_EARNED
    ,pet.ELEMENT_NAME
    ,PET.ELEMENT_INFORMATION_CATEGORY
    ,PET.CLASSIFICATION_ID
    ,PET.ELEMENT_INFORMATION1
    ,pet.ELEMENT_TYPE_ID
    ,paa.tax_unit_id
    ,PAF.ASSIGNMENT_ID
    ,paf.ORGANIZATION_ID
    BEGIN
    DELETE cust.DFC_PAYROLL_DW
    WHERE PAY_DATE BETWEEN TO_DATE(p_sdate) AND TO_DATE(p_edate)
    AND tax_unit_id = NVL(p_GRE, tax_unit_id)
    AND ssn = NVL(p_ssn, ssn)
    COMMIT;
    FOR V_REC IN MainCsr LOOP
    INSERT INTO cust.DFC_PAYROLL_DW(SSN, FULL_NAME, PAY_DATE, PERIOD_END, ELEMENT_NAME, ELEMENT_INFORMATION_CATEGORY, CLASSIFICATION_ID, ELEMENT_INFORMATION1, VALOR, TAX_UNIT_ID, ASSG_ID,ELEMENT_TYPE_ID,ORGANIZATION_ID)
    VALUES(V_REC.SSN,V_REC.FULL_NAME,v_rec.PAY_DATE,V_REC.PERIOD_END,V_REC.ELEMENT_NAME,V_REC.ELEMENT_INFORMATION_CATEGORY, V_REC.CLASSIFICATION_ID, V_REC.ELEMENT_INFORMATION1, V_REC.VALOR,V_REC.TAX_UNIT_ID,V_REC.ASSG_ID, v_rec.ELEMENT_TYPE_ID, v_rec.ORGANIZATION_ID);
    COMMIT;
    END LOOP;
    END ;
    So, how could I assist our developer with this, so that she can run it again without it generating a ton of logs ? ?
    Thanks
    Oracle 9.2.0.5
    AIX 5.2

    The amount of redo generated is a direct function of how much data is changing. If you insert 'x' number of rows, you are going to generate 'y' mbytes of redo. If your procedure is destined to insert 1000 rows, then it is destined to create a certain amount of redo. Period.
    I would question the <i>performance</i> of the procedure shown ... using a cursor loop with a commit after every row is going to be a slug on performance but that doesn't change the fact 'x' inserts will always generate 'y' redo.

  • Archive log generation in every 7 minute interval

    One of the HP Unix 11.11 hosts two databases uiivc and uiivc1. It is found that there is heavy archive log generation in every 7 minute in both databases. Redo log size is 100mb and configured with 2 members each on three groups for these databases.Version of the database is 9.2.0.8. Can anyone help me to find out how to monitor the redo log file contents which is filling up more frequently making more archived redo to generate (filling up the mount point)?
    Current settings are
    fast_start_mttr_target integer 300
    log_buffer integer 5242880
    Regards
    Manoj

    You can try to find the sessions which are generating lots of redo logs, check metalink doc id: 167492.1
    1) Query V$SESS_IO. This view contains the column BLOCK_CHANGES which indicates
    how much blocks have been changed by the session. High values indicate a
    session generating lots of redo.
    The query you can use is:
    SQL> SELECT s.sid, s.serial#, s.username, s.program,
    2 i.block_changes
    3 FROM v$session s, v$sess_io i
    4 WHERE s.sid = i.sid
    5 ORDER BY 5 desc, 1, 2, 3, 4;
    Run the query multiple times and examine the delta between each occurrence
    of BLOCK_CHANGES. Large deltas indicate high redo generation by the session.
    2) Query V$TRANSACTION. This view contains information about the amount of
    undo blocks and undo records accessed by the transaction (as found in the
    USED_UBLK and USED_UREC columns).
    The query you can use is:
    SQL> SELECT s.sid, s.serial#, s.username, s.program,
    2 t.used_ublk, t.used_urec
    3 FROM v$session s, v$transaction t
    4 WHERE s.taddr = t.addr
    5 ORDER BY 5 desc, 6 desc, 1, 2, 3, 4;
    Run the query multiple times and examine the delta between each occurrence
    of USED_UBLK and USED_UREC. Large deltas indicate high redo generation by
    the session.

  • One process generating too much archives

    Hi all,
    DB-Oracle 9.2
    OS- Aix
    One of my friend ask me how to find out which process generating too mcuh archive logs
    below are some of the parameters
    log_checkpoint_interval integer 30000
    log_checkpoint_timeout integer 1800
    log_checkpoints_to_alert boolean FALSE
    log_buffer  1048576
    log_archive_max_processes  2
    fast_start_mttr_target 0please suggest me

    hi,
    you have killed the seesion.. but insert operations are not needed or what? if you don't want any insert from that user then give priveleges acccordingly.
    coz, the import may be needed for some application. or else increase the redo log files.
    to find which process
    check the holding session from dba_waiter and check whether the session using any insert or update statement..(find the query).. even u can check from ps-ef |grep arch..
    instead of killing you can do any other alteration or inform user to import at a less consuption DB time..
    regards,
    Deepak

  • "specification does not match any archived log in the recovery catalog"

    I would like to know about this message.
    I Validated my backup and archivelogs yesterday.
    Today when I tried to do - CROSSCHECK ARCHIVELOG ALL;
    I gave me the following error -
    RMAN> crosscheck archivelog all;
    released channel: ORA_SBT_TAPE_1
    released channel: ORA_DISK_1
    allocated channel: ORA_DISK_1
    channel ORA_DISK_1: SID=291 device type=DISK
    specification does not match any archived log in the recovery catalog
    when I tried to list all the archivelogs then also I am getting this message -
    RMAN> list archivelog all;
    specification does not match any archived log in the recovery catalog
    I have retention policy of 15 Days.
    Waiting for your inputs.

    Thanks Werner..This means that it is just a message.
    One more question related to Archive Log backups -
    traditionally DBAs used to take backup of an Archivelog by spupplying "alter system archive log current" before taking the backup. somewhat like -
    run
    allocate channel t1 device type 'sbt_tape' PARMS="ENV=(TDPO_OPTFILE=/oracle/TESTDB/tdpo.opt)";
    sql 'alter system archive log current';
    backup archivelog all delete input;
    release channel t1;
    But, since RMAN automatically archives CURRENT redo log at the start of the archivelog backup, do we REALLY need to pass sql 'alter system archive log current'; before starting the archivelog backup in the script?
    RMAN> backup archivelog all;
    Starting backup at 11-MAR-10
    current log archived ===========> RMAN automatically archived the CURRENT redo log
    released channel: ORA_DISK_1
    allocated channel: ORA_SBT_TAPE_1
    related question to this is - Since it generates one archived log everytime it takes the archived log backup, this backup may get hung if the archived destination is 100% full. Is there any way we can take the archive log backup without generating an extra archived log before the start of the backup?
    Thanks,
    Roopesh

  • My Backups and Archive Logs

    Hi All,
    I have an Oracle 10G installation on a solaris host and at the moment, there are a lot of archive log files on the server, full level 0 backups are executed every week, and level 1 differentials are done on a daily basis, now the problem is that I have a lot of archive log which has not been deleted, I am not sure if I can delete it now.
    At the moment, the last backkup which was done today also backed up the archive log as well and the latest sequence ID is 2367, the backup done this morning covered the archive logs up till sequence ID 2351 - 2360. The oldest backup goes as far back as sequence ID 2070. I was wondering whether I can delete all archive logs prior to this sequence ID.
    Also, is there a functionality in Enterprise Manager that automatically deletes.
    I look forward to your replies.
    thanks

    If it's a simple "no change to existing scripts" sort of thing you are after, then just add another job to run after the existing ones, as follows (you'll want to follow whatever approach your existing scripts use - this is just a demo.):
    $ORACLE_HOME/bin/rman <<EOF
    connect target
    backup archivelog all delete input;
    exit;
    EOFAnd add whatever error checking the existing script has (if any).
    Or, if changing the existing script/s is OK, add PLUS ARCHIVELOG DELETE INPUT to the script/s.
    And TEST it first, before putting it into production.

  • Generate archive logs are not in sequence number?

    On last friday... the latest archive log number was ARC00024.ARC. Tomorrow when I come backup, the archive logs ARC00001.ARC and ARC00002.ARC were being generated by oracle itself. I wondering the archive log sequence should be in sequence. What is happening?
    SQL> archive log list;
    Database log mode Archive Mode
    Automatic archival Enabled
    Archive destination C:\oracle\ora92\RDBMS
    Oldest online log sequence 1
    Next log sequence to archive 3
    Current log sequence 3
    SQL>
    FAN
    Edited by: user623471 on Jun 7, 2009 7:35 PM

    khurram,
    Its our production instance and havent issued resetlogs option but when listing the arvchives it shows in different sequence number...
    and also while copying the archives by RMAN it doesnt copy in sequence
    -rw-r----- 1 xxx dba 69363859 May 28 19:16 2_10373.arc.gz
    -rw-r----- 1 xxx dba 43446622 May 28 19:16 1_10553.arc.gz
    -rw-r----- 1 xxx dba 52587365 May 28 19:16 1_10578.arc.gz
    -rw-r----- 1 xxx dba 45251820 May 28 19:16 1_10543.arc.gz
    -rw-r----- 1 xxx dba 60890256 May 28 19:17 1_10579.arc.gz
    -rw-r----- 1 xxx dba 46659008 May 28 19:17 1_10548.arc.gz
    -rw-r----- 1 xxx dba 116899466 May 28 19:17 2_10353.arc.gz
    -rw-r----- 1 xxx dba 77769517 May 28 19:17 1_10531.arc.gz
    -rw-r----- 1 xxx dba 66401923 May 28 19:18 1_10530.arc.gz
    -rw-r----- 1 xxx dba 45972697 May 28 19:18 1_10605.arc.gz
    -rw-r----- 1 xxx dba 55082543 May 28 19:18 1_10600.arc.gz
    -rw-r----- 1 xxxq dba 42682207 May 28 19:19 1_10547.arc.gz
    thanks,
    baskar.l

  • To generate email alert when archive logs fill

    Hi. I was away yesterday and the archive logs were filling up and were at 92% when I got in this morning.
    I need to have an email generated once the /archlogs directory fills past 90%.
    The total size of that directory is 8064M. So once it hits around 7257M I would like an email fired to me and my boss.
    Oracle 9.2.0.5.0
    UNIX AIX 5.2

    In addition to Sybrand's reply, you might want to consider some other factors.
    First, when setting up email notifications, you need to adjust your expectations to account for the possible latency in email delivery. I have seen 'critical' emails take several hours to get through the system. This is not a function of Oracle, but of the email servers.
    Second, if archivelog destination filling up is an ongoing problem, I'd be looking at how my archivelogs are backed up and deleted - how I do my housekeeping on that destination. You should have more important things to do than constant monitoring and responding to 'destination full' conditions.

  • NOLOGGING but still generating lot of logs!

    I'm doing a bulk loading using 'sqlldr' with the DIRECT OPTION set to true. The table in question is set no NOLOGGING, yet, I'm getting a lot of redo (resulting in gigabytes of archived logs). The Force Logging option is set to FALSE.
    I really have no clue where is my problem, but I hope someone can shed some light and help me out.

    If you have any indexes on the table, this can result in redo generation. If you have not done so, you may want to use the 'skip_index_maintenance' parameter. This will mark any indexes as unusable. After the load has completed, you can rebuild the indexes in nolooging mode (alter index index_name rebuild nologging)

  • How can I turn off archive logs are being generated by system? (ugrent)

    Dear all,
    How can I turn off archive logs are being generated by system?
    Best Regards,
    Amy

    Sorry not to you @kamran its to OP.accidently it reply button pressed for you
    SQL> shutdown immediate
    Database closed.
    Database dismounted.
    ORACLE instance shut down.
    SQL> startup mount
    ORACLE instance started.
    Total System Global Area  171966464 bytes
    Fixed Size                   787988 bytes
    Variable Size             145750508 bytes
    Database Buffers           25165824 bytes
    Redo Buffers                 262144 bytes
    Database mounted.
    SQL> alter database noarchivelog
      2  /
    Database altered.
    SQL> Khurram

  • Reduce amount of archived log generated.

    RDBMS version : 9.2.0.8
    SQL> SELECT tablespace_name, force_logging FROM dba_tablespaces;
    TABLESPACE_NAME FORCE_LOGGING
    SYSTEM NO
    Above is what status of database, but when I do maintenance work of rebuilding index tablespace I get day or two worth of archived log files.
    and I dont' think ALTER DATABASE no force logging will reduce the amount of log generated.
    Is there any other method available?
    thanks

    Hi,
    if you force logging for a tablespace or for the database, then this means only that any nologging clause that comes with statements related to segments in that tablespace/database is ignored. No force logging is the default.
    In order to reduce the amount of redo protocol, you may consider to use NOLOGGING for the rebuild of your indexes:
    create index <indexname> on <table(column)> nologging;Or you put the tablespace in NOLOGGING in which the indexes are created in:
    alter tablespace <indextablespace> nologging;Or (perhaps even better) simply leave the indexes as they are. Most indexes do not need a rebuild anyway.
    Kind regards
    Uwe
    http://uhesse.wordpress.com

  • Refresh Materilaized Views without generating Archive Logs

    Hello Gurus,
    I'm facing an emebarassing situation,
    I've a job every night to execute a dbms_refresh on several materialiled Views, but there are 2 Mviews that take almost 30 minutes each but another issue thay generates almost 10G of data and my File system becomes full in a minute.
    I've done an alter Mview Nologging but it doesn't change anything.
    Does soemon ahve an idea how to do a refresh without generating Archive logs - or
    without logging?
    thxs in advance
    C.

    Since this is a TRUNCATE and INSERT, any user / application querying the MV while the Refresh is running will see ZERO rows. Furthermore if the INSERT fails (e.g. insufficent space to add an extent), the MV remains with ZERO rows.
    A DELETE and INSERT avoids such situations (even if the refresh fails, both INSERT and DELETE are rolled-back and the MV is reverted to the state as it was before the Refresh began -- thus showing data from the previous refresh)
    Hemant K Chitale
    http://hemantoracledba.blogspot.com
    Just to clarify : My paragraphs above are not about data "missing" but about precautions to be taken. A COMPLETE Refresh is done during defined outages -- users/applications are made aware that data is not available while the Refresh is running.
    Edited by: Hemant K Chitale on Sep 1, 2009 10:53 AM

Maybe you are looking for

  • Macbook Pro 13" Screen Model 9CAO Calibration help?

    This is the screen that I have on my new Macbook Pro. Would anyone out there have a screen profile calibration for this screen model that I could add to my screen (display) profiles. I have tried to do simple calibrations on my own but unfortunately

  • Hose Bank & FBZP Config

    Hi Our client is  not using APP (Automatic Payment Program) Payment to the majority Vendors will be manage through hand written checks.Only for some Banks they are going to use check printing functionality with T Code F-58 To faciliate  this  is it r

  • Managed library - Drobo or Standalone FW800 Drive?

    Greetings to all from Sydney, Australia! I currently have my Aperture 3 managed library on a Drobo connected via FW800 to my 24" iMac 2.8 GHz Core 2 Duo running under Leopard 10.5.8. It seems slow when I am editing images - there is always a delay be

  • System.in newbie problem

    I'm new in java and i can't find anywhere: Like i have PrintWriter object (and OutputStreamWriter(System.out,"codepage"); I want the same thing, to read form windows console, for example a String... (is there an object, that is similar to OutputStrea

  • Period Year conversion

    Hi, I need to convert Period & year to a calender date. How to acheive this ?. I sthere any Function module? Pls help Thanks Ram