Steps for Data Archiving

Dear Friends,
Can i know the steps for creating the archive file in data archiving.
or
If you have the links,please give it to me.
I promise to Reward.
Thanx in advance.
Geetu

Hi,
Check the links below for data archiving
http://help.sap.com/saphelp_46c/helpdata/en/6d/56a06a463411d189000000e8323d3a/frameset.htm
http://help.sap.com/saphelp_46c/helpdata/en/2e/9396345788c131e10000009b38f83b/frameset.htm
http://help.sap.com/saphelp_nw04s/helpdata/en/60/16d93f2e93f106e10000000a1550b0/content.htm
http://help.sap.com/saphelp_nw2004s/helpdata/en/ff/ff6c00409f11d1896b0000e8322d00/frameset.htm
http://help.sap.com/saphelp_nw2004s/helpdata/en/b6/ff0d794b9e11d192f20000e829fbc6/frameset.htm
http://help.sap.com/saphelp_nw2004s/helpdata/en/b0/3a618cd1d411d4b645006094192fe3/frameset.htm
http://help.sap.com/saphelp_nw2004s/helpdata/en/85/8d3584c47211d5993d00508b6b8b11/frameset.htm
http://help.sap.com/saphelp_nw2004s/helpdata/en/4f/99335d446d11d189700000e8322d00/frameset.htm
http://help.sap.com/saphelp_nw2004s/helpdata/en/b0/3a6183d1d411d4b645006094192fe3/frameset.htm
http://help.sap.com/saphelp_nw2004s/helpdata/en/4f/993685446d11d189700000e8322d00 /frameset.htm
http://help.sap.com/saphelp_nw2004s/helpdata/en/60/16d93f2e93f106e10000000a1550b0/frameset.htm
http://help.sap.com/saphelp_nw2004s/helpdata/en/89/43a239dac4b56be10000000a114084/frameset.htm
http://help.sap.com/saphelp_nw2004s/helpdata/en/4f/99366b446d11d189700000e8322d00/frameset.htm

Similar Messages

  • Best Practise for Data Archiving

    Hi All,
    I have query in SAP that when business will do the data archving in their R3 systems.
    On which basis they will do data archiving?
    Is there any minimum period that data need's to be archived?
    before doing the data archiving what they will do ?
    Regards
    Srini

    Hi Srini,
    1. SAP suggestes to implement data archiving strategy  as early as possible to manage database growth .
    However pople think of archiving when they realise the  problems like large data volumes,slow system resonse time,performance issues etc...
    2. There is a proper way to implement Data Archiving . Database has to be anaylzed first for getting the top DB tables and Archiving objects.
    3. Based on the DB analysis ,Data archiving plan has to be implenemted according to data management guide.
    4. There is a minimum period known as residence time has to be completed before any data to be archived. Once the document is business completed and serverd its minimum required period in the Database ,it can be archived.
    5, Before going for data archiving there are many steps to be followed like analysis,configuration etc that you can see in details at the link below :
    http://help.sap.com/saphelp_47x200/helpdata/en/2e/9396345788c131e10000009b38f83b/frameset.htm
    let me know if this helps you .
    -Supriya
    Edited by: Supriya  Shrivastava on May 4, 2009 10:38 AM

  • Steps for Data extraction from SAP r/3

    Dear all,
    I am New to SAP Bw.
    I have done data extraction from Excel into SAP BW system.
    that is like
    Create info objects > info area> Catalog
                                                --> Character catalog
                                                --> Key catalog
    Create info source
    Upload data.
    create info cube
    I need similar steps for data extraction for SAP R/3
    1. when data is in Ztables ( using Views/Infosets/Function etc)
    2. When data is with Standard SAP using Business Content.
    Thanks and Regards,
    Gaurav Sood

    hi,
    chk the links
    Generic Extraction
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/84bf4d68-0601-0010-13b5-b062adbb3e33
    CO-PA
    http://help.sap.com/saphelp_46c/helpdata/en/7a/4c37ef4a0111d1894c0000e829fbbd/content.htm
    CO-PC
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/fb07ab90-0201-0010-c489-d527d39cc0c6
    iNVENTORY
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/f83be790-0201-0010-4fb0-98bd7c01e328
    Extractions in BI
    https://www.sdn.sap.com/irj/sdn/wiki
    LO Extraction:
    /people/sap.user72/blog/2004/12/16/logistic-cockpit-delta-mechanism--episode-one-v3-update-the-145serializer146
    /people/sap.user72/blog/2005/01/19/logistic-cockpit-delta-mechanism--episode-three-the-new-update-methods
    /people/sap.user72/blog/2005/02/14/logistic-cockpit--when-you-need-more--first-option-enhance-it
    /people/sap.user72/blog/2004/12/23/logistic-cockpit-delta-mechanism--episode-two-v3-update-when-some-problems-can-occur
    /people/sap.user72/blog/2005/04/19/logistic-cockpit-a-new-deal-overshadowed-by-the-old-fashioned-lis
    Remya

  • A question on different options for data archival and compression

    Hi Experts,
    I have production database of about 5 terabyte size in production and about 50 GB in development/QA. I am on Oracle 11.2.0.3 on Linux. We have RMAN backups configured. I have a question about data archival stretegy.  To keep the OLTP database size optimum, what options can be suggested for data archival?
    1) Should each table have a archival stretegy?
    2) What is the best way to archive data - should it be sent to a seperate archival database?
    In our environment, only for about 15 tables we have the archival stretegy defined. For these tables, we copy their data every night to a seperate schema meant to store this archived data and then transfer it eventualy to a diffent archival database. For all other tables, there is no data archival stretegy.
    What are the different options and best practices about it that can be reviewed to put a practical data archival strategy in place? I will be most thankful for your inputs.
    Also what are the different data compression stregegies? For example we have about 25 tables that are read-only. Should they be compressed using the default oracle 9i basic compression? (alter table compress).
    Thanks,
    OrauserN

    You are using 11g and you can compress read only as well as read-write tables in 11g. Both read only and read write tables are the candidate for compression always. This will save space and could increase the performance but always test it before. This was not an option in 10g. Read the following docs.
    http://www.oracle.com/technetwork/database/storage/advanced-compression-whitepaper-130502.pdf
    http://www.oracle.com/technetwork/database/options/compression/faq-092157.html

  • Any pre-requiste for data archiving

    hi,
      is there any pre-requiste for data archiving..
         i.e.,   for a particular data we have to archive..
              or any condition
    regards
    sree

    Hi
    Archive Selection
    Use
    For some archiving actions it is necessary to select individual archive files or whole archive sessions and make them available for the relevant action. You do this in the Archive Selection screen, which is accessible from the initial screen of the relevant action.
    Archive selection is relevant for the following actions:
    Delete
    Read
    Build index
    Reload
    Postprocessing
    Prerequisites
    Depending on the action, there are four different criteria for whether archiving sessions and archive files are displayed in the archive selection:
    Criterion 1
    Archive file status
    Action
    Criterion
    Delete
    Archive file status
    Archived or
    Stored (if sequence "Store before deletion" has been selected in Customizing)
    Read
    Archive file status Archived
    or
    Deleted
    or
    Saved
    Build index
    Archive file has the status Deleted and the file is in the file system
    Reload
    Archive file is in file system and the session has the corresponding status (see criterion 5)
    Postprocessing
    All archive files have status Deleted
    Criterion 2
    The archive file must contain at least one data object. If you have performed an archiving session and maintained the variants so that no data objects were archived, the archive file is automatically deleted and no longer offered for selection.
    Criterion 3
    If a file is not accessible in the file system, the file is not offered for selection. A file is only offered for read if all the files in a session are accessible.
    While checking the accessibility of a file, the system performs a read access for each file. This check process can take a long time if there are a lot of files. To optimize this process the access check can be de-/activated according to archive files in file system and archive files in an optical archive in archiving object-specific Customizing.
    Criterion 4
    For client-dependent archiving objects only the sessions for the current client are offered.
    Criterion 5
    An archiving session can only be reloaded if it has the status Complete or Archiving session is reloaded.
    Features
    In the Archive Selection screen, which you can access by choosing Archive Selection from the initial screen of the archiving action, you can either select complete archive sessions or individual archive files.

  • Need Step By Step Guide For Data Archiving

    Can anybody provide me step by step Data Archiving guide for ECC 6.0 system.
    I searched OSS but did not get any guide.
    Thanks

    Hi
    You can go through the below links related to archiving which transactions we use for archiving the objects and how we do that.
    http://help.sap.com/saphelp_srm30/helpdata/en/15/c9df3b6ac34b44e10000000a114084/content.htm
    http://help.sap.com/SAPHELP_NW04S/helpdata/EN/43/f0ed9f81917063e10000000a1553f6/content.htm
    Basically Data archiving process comprises three major phases. They are:
    1. Creating an archive file: The archived files of data are created in the SAP database by the Archiving Management system. The Management system reads the data from the database and writes it to the archive files in the background. In instances of archived files exceeding the maximum specified limit, or if the number of data objects exceed the stipulated limit in the system, then the system automatically creates new archive files.
    At the end of the process of saving data into archive files, ADK triggers the system event SAP_ARCHIVING_WRITE_FINISHED, which is an indicator to the system to start next phase of archiving process.
    2. Removing the archived data from the database:
    While archiving management system writes data on the archive files, another program deletes it from the database permanently. The program checks whether the data has been transferred to the archive. It is quite important too as it is the last check performed by the system before deleting data permanently from the database. Several deletion programs run simultaneously, because the archiving program is much more faster than the deletion programs. This is important as it increases the efficiency of archiving process.
    3. Transferring the archived files to a location outside the SAP database
    Once the Archive management system has finished archiving the data, the next step is to save the archived files at a different location other than the SAP database. This can be accomplished by an automated process in the system or by a manual process. This step is optional since many enterprises may wish to keep the archived files within the current database. However, large enterprises transfer their data periodically as a part of their data archiving processes.
    Hope this helps!!
    Thanks and Kind Regards
    Esha

  • The best practice for data archiving

    Hi
    My client has been using OnDemand for almost 2 years, there are around 2M records in the system(Activities), so just want to know what is the best practice of data archiving, we dont care much about the data in the 6 month ago.

    Hi Erik,
    Archival is nothing but deletion.
    Create a backup cube in BW. Copy the data from your planning cube to the backup cube, and then delete that data region from your planning cube.
    Archival will definitely improve the performance of your templates, scripts, etc; since the system will now search from a smaller dataset.
    Hope this helps.

  • Regarding sending certificate to the third party for data archiving.

    Hi,
    I am stuck at one point in SAP. The steps I have followed are:
    1. go to transaction code "oac0" (Display content repositories).
    2. double click to any archive.
    3. Here we can see a "Send Certificate", button.
         By pressing this button, an http request is fired over the network, addressing the application server in its configuration configured for that archive. Together with it an certificate is also sent to the server outside SAP.
    Generally we do validation of the certificate outside SAP in our application, e.g. in our JAVA code, using iaik jars (security jars).
    Now I need to validate this certificate inside SAP itself prior to be sent over the network from SAP.
    Is it possible any way.

    Mridul,
    Depending on your ERP version you can use STRUST to import the certificate into the ABAP system.
    Regards,
    Salvatore Castro
    Sr. Solution Architect

  • Post Installation steps for Messages archiving from Adapter engine

    Hi,
    Can any let me know the post installation steps for Archiving messages from AE.
    We are archiving the messages from Integration Engine (BC_XMB).
    This note refers to the Post installation steps... But i could not find them anywhere.
    Note 790226 - Messages in AdapterEngine/PCK database do not get archived
    Thanks,
    Tanuj
    Message was edited by:
            Tanuj Kumar Bolisetty

    Hi Tanuj,
    have a look at note 872388
    "2) For messages in the Adapter Framework: note 816022, question 8
    (deletion) and
    http://help.sap.com/saphelp_nw04/helpdata/en/cf/b2fc3f48ecc742e10000000a
    1550b0/content.htm (archiving)"
    Additionally consult the troubleshooting guide,page 110
    Set the adapter service properties (note 791379)
    Hope it helps
    Jaime

  • SAP Support for Data Archiving

    Hello,
    We just wanted to confirm if SAP will support issues if we encounter issues while restoring data archived via SARA. For example, if in the process of restoring archived data a defect is found, is this still supported by SAP?

    Hi
    Supported and the SAP Data Archiving Support Tools  are SARA , DART and SARI
    SARA --- SAP Archive Administration
    All data archiving activities commence with Archive administration (transaction SARA). SARA provides the overall administration of archiving schedules and manages the archiving sessions. The process includes customization of objects, their conversion to archived sequential archived files and most importantly, their overall management. In addition to these, the archive administration process also retrieves the archived files and converts the data through an automated process, if there is a change in software, hardware or the data structure. The data archiving process is streamlined and simplified through the Central command of archiving administration
    DART --- Data Retention Tool
    A data retention strategy is required for retaining enterprise information for long periods of time. The Data Retention Tool (DART) provides this functionality. It is capable of extracting data, which is period-specific as well as any supporting information. DART transforms database objects into flat files that can be read by any third-party software designed for flat files. It is available as an add-on for older versions of SAP R/3, but is an integrated feature of the recent versions of SAP R/3.
    SARI --- SAP Archive Information System
    SARI provides much-needed retrieval capabilities against previously archived data. It requires the archive files be loaded into new tables in the database. In this approach, AS requires detailed, technical knowledge of the database table structure and field layouts. Database structure of an enterprise needs to be flexible and dynamic to be able to quickly adapt to organisational changes. It might happen that the current database structure is not compatible with the structure of database when archive files were initially generated. Archive Information System is a standard tool delivered by SAP that facilitates customized access to the archived data
    Regards
    Venkat

  • Steps for Data Guard with one primary and 2 standby

    Hi,
    Database :10.2.0.4, 11.2.0.1
    Os: Windows , Unix
    A ----------------> Primary database
    B ----------------> Standby Database 1
    C ----------------> Standby Database 2
    I want to configure *2 standby* databases for single primary database.
    Lets take, A ,B and C are my machines.My data guard configuration will be like,*archive logs will be moving* from A to B and A to C.
    If i do any switchover in between A and B , now B is primary and remaining A and C are standby databases.At this stage also , archive logs should move from B to A and B to C. Also, same should happen from C to A and C to B,If i do switchover in between B and C.If everything is fine , then i will do switchback to main Primary database(A).
    How do i have to mention PFILE in all machines ,the parameters like
    LOG_ARCHIVE_DEST_1=LOCATION=<PATH> -- LOCAL ARCHIVE PATH
    LOG_ARCHIVE_DEST_2=SERVICE=
    LOG_ARCHIVE_DEST_3=SERVICE=
    FAL_SERVER=
    FAL_CLIENT=
    STANDBY_FILE_MANAGEMENT=
    In my tnsnames.ora , primary,standby1 and standby2 are my service entries and these are same in all of my machines.
    Please suggest me , how do i can configure my pfiles in all machines ?.
    Thanks,
    Sunand

    Not yet, but now you have me interested.
    Please consider Flashback.
    I still have to test but here's my take:
    PRIMARY SETTINGS
    *.FAL_SERVER=STANDBY
    *.FAL_CLIENT=PRIMARY
    *.STANDBY_FILE_MANAGEMENT=AUTO
    *.DB_UNIQUE_NAME=PRIMARY
    *.LOG_FILE_NAME_CONVERT='STANDBY','PRIMARY'
    *.log_archive_dest_1='LOCATION=USE_DB_RECOVERY_FILE_DEST VALID_FOR=(ALL_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=PRIMARY'
    *.log_archive_dest_2='SERVICE=STANDBY LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=STANDBY'
    *.log_archive_dest_3='SERVICE=STANDBY2 LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=STANDBY2'
    *.LOG_ARCHIVE_DEST_STATE_1=ENABLE
    *.LOG_ARCHIVE_DEST_STATE_2=ENABLE
    *.LOG_ARCHIVE_DEST_STATE_3=ENABLE
    *.LOG_ARCHIVE_MAX_PROCESSES=30
    STANDBY 1 SETTINGS
    *.FAL_SERVER=PRIMARY
    *.FAL_CLIENT=STANDBY
    *.STANDBY_FILE_MANAGEMENT=AUTO
    *.DB_UNIQUE_NAME=STANDBY
    *.LOG_FILE_NAME_CONVERT='PRIMARY','STANDBY'
    *.log_archive_dest_1=LOCATION=USE_DB_RECOVERY_FILE_DEST VALID_FOR=(ALL_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=STANDBY'
    *.log_archive_dest_2='SERVICE=PRIMARY LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=PRIMARY'
    *.log_archive_dest_3='SERVICE=STANDBY2 LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=STANDBY2'
    *.LOG_ARCHIVE_DEST_STATE_1=ENABLE
    *.LOG_ARCHIVE_DEST_STATE_2=DEFER
    *.LOG_ARCHIVE_DEST_STATE_3=DEFER
    *.LOG_ARCHIVE_MAX_PROCESSES=30
    STANDBY2 SETTINGS
    *.FAL_SERVER=PRIMARY
    *.FAL_CLIENT=STANDBY2
    *.STANDBY_FILE_MANAGEMENT=AUTO
    *.DB_UNIQUE_NAME=STANDBY2
    *.LOG_FILE_NAME_CONVERT='PRIMARY','STANDBY2'
    *.log_archive_dest_1=LOCATION=USE_DB_RECOVERY_FILE_DEST VALID_FOR=(ALL_LOGFILES,ALL_ROLES) DB_UNIQUE_NAME=STANDBY2'
    *.log_archive_dest_2='SERVICE=STANDBY LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=STANDBY'
    *.log_archive_dest_3='SERVICE=PRIMARY LGWR ASYNC VALID_FOR=(ONLINE_LOGFILES,PRIMARY_ROLE) DB_UNIQUE_NAME=PRIMARY'
    *.LOG_ARCHIVE_DEST_STATE_1=ENABLE
    *.LOG_ARCHIVE_DEST_STATE_2=DEFER
    *.LOG_ARCHIVE_DEST_STATE_3=DEFER
    *.LOG_ARCHIVE_MAX_PROCESSES=30
    Edited by: mseberg on Nov 29, 2010 9:39 AM
    The first test slapped me. Looking at 409013.1 Cascaded Standby Databases
    Edited by: mseberg on Nov 29, 2010 12:49 PM

  • Steps for data guard RAC primary RAC standby

    Hi,
    I have some doubt in configuring data guard for rac DR to rac primary.
    1) In RAC DR, physical standby will be in RAC before starting data guard for RAC Primary , right?.
    2) While configuring RAC DR , should one node be down in standby?.
    3) After creating the auxiliary database in standby, i mean restoring the rman backup from staging directory in DR,
    Shall we have to register the services(database,instance,asm,etc..) to CRS(OCR)?. If yes ,Why?.
    4) After DR configure over , Shall i shut/down one node?. Should all the nodes in DR be in up , while applying logs to
    standby ?.
    If anybody have made the setup and have the prepared document , requesting you to share with me.
    Thank you very much,
    Sunand

    Hi Sunand,
    Please follow the following My Oracle Support (MOS) Document ID:
    MAA - Creating a RAC Physical Standby for a RAC Primary [ID 380449.1]
    For further details, see the referred documents in previous post. Try practicing on VMWare because it will give you the flexibility for creating snapshots to save your work. Please see the answers for your question below:
    Q1. In RAC DR, physical standby will be in RAC before starting data guard for RAC Primary , right?.
    Answer: Yes, this is recommend but not mandatory. You can add a second node later on.
    Q2. While configuring RAC DR , should one node be down in standby?
    Answer: Yes, there are some database configuration commands that should be run in exclusive mode. You can start the remaining instance(s) after DB creation. Do not forget to set CLUSTER_DATABASE parameter to TRUE. Also, all instances except one should be close while performing switchover/failover.
    Q3. After creating the auxiliary database in standby, i mean restoring the rman backup from staging directory in DR, Shall we have to register the services(database,instance,asm,etc..) to CRS(OCR)?. If yes ,Why?
    Answer: Yes, this is recommended for RAC configuration to take advantage of high availability services.
    Q4. After DR configure over , Shall i shut/down one node?. Should all the nodes in DR be in up , while applying logs to
    standby ?
    Answer: You can start up all the instances after DR creation. In this case if your one node goes down log apply services will continue to apply changes in DR.
    Hope, this helps to clear your questions.
    Regards,
    Shahid

  • Write custom logic for data archiving using Open text(SAP archiving)

    Hi,
    We are currently doing archiving for sales documents and billing documents using Open text. As part of this we need to have a search functionality based on Sales document,customer number, customer address etc...
    since customer address doesn't exist in sales order tables we need to write logic so that the document gets archived and retrieved using the address.
    can you please help me out in how and where to write logic for this...can this be achieved?
    i observed an area on Attribute Object for a Node which gives an option for Event and User exit. can this be used..if so how?
    Thanks,
    Anil

    hi,
    please write more detailed information what do you really need...

  • How to create a repository for data archiving?

    We have a content server installed 2 months ago.
    We want to store the archived data either into the content server DB (even we know it's NOT
    recommended) or better into the file system on the content server.
    Could you help tell how to create the repository to hold the archived data in above 2 scenarios?
    Thanks a lot!

    content server DB (even we know it's NOT recommended)
    I have use Content Server with MaxDB in more than 1 ocation without any issues.
    You need to set up the repositories in OAC0.
    Read,
    http://help.sap.com/saphelp_nw04s/helpdata/en/ed/cefa3882cd8229e10000000a11402f/frameset.htm
    Regards
    Juan

  • About the physical path for data archive

    Hi guys,
    Now we're doing the archive work.As our PRD system is a CI+DI system which has two servers just let me
    call them 63 and 64 here.We mount a media on 63's root directory.So 64
    cannot recognize the media indeed.And the background job of delete
    programs,in fact,are dilivered on the two servers depend something like
    load.
    So parts of the jobs finished,but others are unsuccessful.
    Now I don't know why the 64 server cannnot recognize the media's
    directory.
    How can I solve this problem?I've configured the directory parameter in t-code AL11,but it still only can be recognized by 63 which the media mounted on.
    Now I have two ideas about this.
    One is that using FTP to read or write files,so it has nothing to do with mount on which server.But I don't know whether it works and how can I do this.
    The other is that using SAMBA to mount the media which is an area of a Windows Server onto the 63 and 64 servers simutaneously.
    Do you have some much better advices?Please give me some advices.
    Thank you very much.
    Brooksq
    Moderator message: not related to ABAP development, please search for a better SCN forum for your issue.
    Edited by: Thomas Zloch on Mar 2, 2012

    The only product I'm aware of is the Griffin Cable Lock.
    My personal preference is for a bike cable lock through the hole in the stand to an eyebolt in the wall behind.

Maybe you are looking for