Data Archiving and ABAP

Hi All,
I am an ABAPer and recently came accross the process of data archiving..
Can anyone tell me what significance data archiving has w.r.t ABAP.Are they interrelated at any point??
Waiting for Reply..
Shilpa

Hi Shilpa,
As a developer you must be aware of the fact that what ever application are provided by SAP they all are using codes written in ABAP in background .
Similarly we have a tool Archive Developent Kit (ADK) provided by SAP that uses various programs written in ABAP grouped in the form of Archive object to perform Archiving Successfully in any system.
Yes Archiving and ABAP both are interrelated and an ABPER can very well understand how actually there programs functions at runtime while we Archive data from the database.
Another imporeant thing is apart from the standard Archive Objects provided by SAP sometimes  requiremnets comes for  custom Objects to be created which requires good indepth knowledge of ABAP.
I hope my answer will help you understand the assosiation of ABAP and Data Archiving.
For more you can go through the below link ::
http://help.sap.com/saphelp_47x200/helpdata/en/8d/3e4d22462a11d189000000e8323d3a/frameset.htm
-Supriya

Similar Messages

  • What is data archiving and DMS(Data Management System) in SAP

    what is data archiving and DMS(Data Management System) in SAP
    Welcome to SCN. Before posting questions please search for available information here and in the web. Please also read the Rules of Engagement before further posting.
    Edited by: kishan P on Aug 31, 2010 1:06 PM

    Hi,
    Filtering at the IDoc Level
    Identify the filter object (BD59)
    Modify the distribution model
    Segment Filtering
    specify the segments to be filtered (BD56)
    The Reduced IDoc Type
    Analyze the data.
    Reduce the IDoc type (BD53)
    Thanks and regards.

  • Data archiving and data cleansing

    hi experts,
    Can any one tell me step by step guide for Data archiving and Data cleansing of SAP-ISU Object.
    what is the difference between  Data archiving and Data cleansing .
    Thanx & Rgds

    Data Archiving: So many objects are there you can look some of them..........
    ISU_BBP IS-U Archiving: Budget Billing Plan
    ISU_BCONT Business Partner Contacts (Contract A/R + A/P)
    ISU_BILL IS-U Archiving: Billing Document Header
    ISU_BILLZ IS-U Archiving: Billing Line Item
    ISU_EABL IS-U Archiving: Meter Reading Results
    ISU_EORDER IS-U Archiving: Waste Disposal Order
    ISU_EUFASS Archiving of Usage Factors
    ISU_FACTS Installation Facts
    ISU_INSPEC IS-U Archiving: Campaigns for Inspection List
    ISU_PPM Prepayment Documents
    ISU_PRDOCH IS-U Archiving: Print Document Header
    ISU_PRDOCL IS-U Archiving: Print Document Line Item
    ISU_PROFV IS-U Archiving: EDM Profile Values
    ISU_ROUTE IS-U Archiving: Waste Disposal Route
    ISU_SETTLB Settlement Document
    ISU_SWTDOC Archive Switch Document
    Go to SARA t-code and enter the object CA_BUPA for business partner then Press F6 you will get  step by step documentation. please follow the procedure for all the objects.
    Regards,
    Siva

  • Data archival and purging for OLTP database

    Hi All,
    Need your suggestion regarding data archival and purging solution for OLTP db.
    currently, we are planning to generate flat files from table..before purging the inactive data and move them to tapes/disks for archiving then purge the data from system. we have many retention requirements and conditions before archival of data. so partition alone is not sufficient.
    Is there any better approach for archival and purging other than this flat file approach..
    thank you.
    regards,
    vara

    user11261773 wrote:
    Hi All,
    Need your suggestion regarding data archival and purging solution for OLTP db.
    currently, we are planning to generate flat files from table..before purging the inactive data and move them to tapes/disks for archiving then purge the data from system. we have many retention requirements and conditions before archival of data. so partition alone is not sufficient.
    Is there any better approach for archival and purging other than this flat file approach..
    thank you.
    regards,
    varaFBDA is the better option option .Check the below link :
    http://www.oracle.com/pls/db111/search?remark=quick_search&word=flashback+data+archive
    Good luck
    --neeraj                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • What should i do to implement Data Archiving and Data Reporting

    we want to implement Data Archiving and Data Reporting in our product. Can someone tell me what techniques or approaches people take in general to implement Data Archiving and Data reporting.
    i am currently looking into DataWarehousing. is this the right apporach ? i have no idea as where i should start on this. can someone give me a good direction as a starting point.
    thank you,
    Puja

    Did you setup Find My Mac before it was stolen?

  • Goldengate Extracts reads slow during Table Data Archiving and Index Rebuilding Operations.

    We have configured OGG on a  near-DR server. The extracts are configured to work in ALO Mode.
    During the day, extracts work as expected and are in sync. But during any dialy maintenance task, the extracts starts lagging, and read the same archives very slow.
    This usually happens during Table Data Archiving (DELETE from prod tables, INSERT into history tables) and during Index Rebuilding on those tables.
    Points to be noted:
    1) The Tables on which Archiving is done and whose Indexes are rebuilt are not captured by GoldenGate Extract.
    2) The extracts are configured to capture DML opeartions. Only INSERT and UPDATE operations are captured, DELETES are ignored by the extracts. Also DDL extraction is not configured.
    3) There is no connection to PROD or DR Database
    4) System functions normally all the time, but just during table data archiving and index rebuild it starts lagging.
    Q 1. As mentioned above, even though the tables are not a part of capture, the extracts lags ? What are the possible reasons for the lag ?
    Q 2. I understand that Index Rebuild is a DDL operation, then too it induces a lag into the system. how ?
    Q 3. We have been trying to find a way to overcome the lag, which ideally shouldn't have arised. Is there any extract parameter or some work around for this situation ?

    Hi Nick.W,
    The amount of redo logs generated is huge. Approximately 200-250 GB in 45-60 minutes.
    I agree that the extract has to parse the extra object-id's. During the day, there is a redo switch every 2-3 minutes. The source is a 3-Node RAC. So approximately, 80-90 archives generated in an hour.
    The reason to mention this was, that while reading these archives also, the extract would be parsing extra Object ID's, as we are capturing data only for 3 tables. The effect of parsing extract object id's should have been seen during the day also. The reason being archive size is same, amount of data is same, the number of records to be scanned is same.
    The extract slows down and read at half the speed. If normally it would take 45-50 secs to read an archive log of normal day functioning, then it would take approx 90-100 secs to read the archives of the mentioned activities.
    Regarding the 3rd point,
    a. The extract is a classic extract, the archived logs are on local file system. No ASM, NO SAN/NAS.
    b. We have added  "TRANLOGOPTIONS BUFSIZE" parameter in our extract. We'll update as soon as we see any kind of improvements.

  • Difference between Data Dictionay and ABAP/4 dictionary

    Hi all,
      can any body please say me the difference between Data Dictionay and ABAP/4 dictionary
                                                                                    Ranjith

    Check this PDF:
    http://www.sappoint.com/faq/faqabdic.pdf#search=%22Data%20Dictionary%20and%20ABAP%2F4%20dictionary%22
    Regards,
    Prakash.

  • A question on different options for data archival and compression

    Hi Experts,
    I have production database of about 5 terabyte size in production and about 50 GB in development/QA. I am on Oracle 11.2.0.3 on Linux. We have RMAN backups configured. I have a question about data archival stretegy.  To keep the OLTP database size optimum, what options can be suggested for data archival?
    1) Should each table have a archival stretegy?
    2) What is the best way to archive data - should it be sent to a seperate archival database?
    In our environment, only for about 15 tables we have the archival stretegy defined. For these tables, we copy their data every night to a seperate schema meant to store this archived data and then transfer it eventualy to a diffent archival database. For all other tables, there is no data archival stretegy.
    What are the different options and best practices about it that can be reviewed to put a practical data archival strategy in place? I will be most thankful for your inputs.
    Also what are the different data compression stregegies? For example we have about 25 tables that are read-only. Should they be compressed using the default oracle 9i basic compression? (alter table compress).
    Thanks,
    OrauserN

    You are using 11g and you can compress read only as well as read-write tables in 11g. Both read only and read write tables are the candidate for compression always. This will save space and could increase the performance but always test it before. This was not an option in 10g. Read the following docs.
    http://www.oracle.com/technetwork/database/storage/advanced-compression-whitepaper-130502.pdf
    http://www.oracle.com/technetwork/database/options/compression/faq-092157.html

  • Business Partner Data Archiving - Help Required

    Hi,
    Am New to Data Archiving and need help to Archive Business Partner in CRM. I tried to Archive some BP's but it was not archiving. Kindly throw some light on it.
    Problem we face are:
    1) When we try to write a Business Partner to an Archive File the Job Log Shows finished but no files are created in the System.
    2) When we try to delete the BP from Data Base it doesn't show the Archived BP's that could be deleted (I guess this is because there are no archived file).
    Archive Object Used is CA_BUPA. We have created a Variant and set the Time as Immediate and Spool to LOCL.
    Kindly let us know if there is any step we are missing here and if we are on the wrong track.
    All answers are appreciated.
    Thanks,
    Prashanth
    Message was edited by:
            Prashanth KR

    Hi,
    To archive a BP following steps are to be followed.
    A. Mark the BP to be archived in transaction BP >status tab
    B. Run dacheck
    FYI : Steps on how to perform dacheck :
    1.  In transaction DACONTROL change the following parameter to the
    values : for CA_BUPA .
    No. calls/packages   1
    Number of Objects   50
    Package SizeRFC     20
    Number of Days       1 (if you use mobile client it should be more)
    2.If a business partner should be archived directly after the
      archiving note was set, this can be done by reseting the check
       date with report CRM_BUPA_ARCHIVING_FLAGS.
       here only check (set) the options :
       - Archiving Flag
       - Reset Check Date
      Reset all options should be unchecked.
    3. go to dacheck and run the process
    4. This will change the status of the BPs to 'Archivable'
       *Only those bp's which are not involved in any business transactions,
        install base, Product Partner Range (PPR) will be set to archivable
    The BP's with status 'Archivable' will be used by the archiving
         run.
    Kindly refer note 646425 before goint to step C ***
    C.Then run transaction CRM_BUPA_ARC.
    - Make sure that the "selection program" in transaction "DACONTROL"
       is maintained as " CRM_BUPA_DA_SELECTIONS"
      Also create a variant, which can done by excecuting
      CRM_BUPA_DA_SELECTION and enter the variants name for the field
      Variant. This variant is the buisness partner range.
    - Also please check note 554460.
    Regards, Gerhard

  • Sys tables from flashback data archive?

    Hi,
    I am prototyping flashback data archive and I have been following some tutorials.
    When I query
    user_flashback_archive_tables
    I can see the table for which I have configured FBDA. I can see also the name of the History table, like SYS_FBA_HIST_xxxxx (xxxxxx being the object_id of the table).
    But I just cannot find the table in USER_TABLES or USER_TAB_PARTITIONS.
    The history seems to be getting stored... The sql-developer display the history in the "flashback" tab. But why aren't the SYS_FBA_HIST_xxxx tables getting created?
    (I am using 11gR2 running on EL5 Linux)

    The SYS_FBA_xxx belongs to SYS user. To be able to see them you must query DBA_TABLES and have SELECT privilege on DBA_TABLES or role SELECT_CATALOG_ROLE.You can also just try to connect as SYS:
    SQL> select table_name from dba_tables where table_name like 'SYS_FBA%';
    TABLE_NAME
    SYS_FBA_HIST_73563
    SYS_FBA_TCRV_73563
    SYS_FBA_DDL_COLMAP_73563
    SYS_FBA_DL
    SYS_FBA_USERS
    SYS_FBA_PARTITIONS
    SYS_FBA_TRACKEDTABLES
    SYS_FBA_BARRIERSCN
    SYS_FBA_TSFA
    SYS_FBA_FA
    10 rows selected.Edited by: P. Forstmann on 7 oct. 2010 20:52

  • CATS data archiving

    Hi...
    Can any one help me out in giving the information of CATS data archiving and i require OSS notes for this.
    Vijay

    Hi Srinivas,
    Thanks for quick response.
    Can you please provide me the OSS notes of these CATS data archiving.
    Vijay

  • Archive and Purge provisions in GRC suite

    Hi,
    I'm trying to understand the Archive and Purge options provided / offered by GRC suite but have very skeletal information. Can somebody help me on that front?
    Balaji K

    In a nutshell,purging is removal of data from a system or storage device with the intent that the data can not be reconstructed by any known technique. In GRC Suite purging has similar concept to purge incident stored after a long period of tie and storage (AACG). You can also purge snapshop or change tracking definitions in CCG. General rule of thumb, ensure you follow your company policies with regards to data archive and purging and access to these functionalities within the GRC Control Suite should be restricted.

  • Unable to access Archived data with an ABAP Query

    I have an ABAP Query that uses Logical Database KDF (Vendor) for reporting.
    KDF is Archive Enabled and I can access archived document via SAP standard programs that used the KDF Logical Database.
    the query appears accessing the data but nothing shows on the report.
    Has anyone successfully written and ABAP query and included access to the Archived data?
    Any help is appreciated.

    Hi Gena,
    I'm facing exactly the same problem as you...
    Since this post is an old one, I imagine that you may not remember, but I have to try...
    Have you solved it? If yes, could you please tell me how?
    I've tried to use CL_ABAP_GZIP and CL_ABAP_UNGZIP_BINARY_STREAM and I'm getting the same error 30 at the  IctDecompressStream function.
    Tks in advance,
    Flavio.

  • SAP Data Archiving in R/3 4.6C and ECC 6.0.

    Hi Guys,
        Need some suggestions,
    We currently working on SAP R/3 4.6 and have plans to upgrade it to ECC 6.0,
    In the mean time there is an requirement for SAP Data Archiving for reducing the database size and increase system performance.So wanted to know if its better to do Data Archiving before upgrade or after. Technically which will be comfortable, Also wanted to know if any advanced method available in ECC 6.0 compared to SAP R/3 4.6.
    Please provide your valuable suggestion.
    Thanks and Regards
    Deepu

    Hi Deepu,
    With respect to archiving process there will be no major difference in 4.6 and ECC 6.0 system. However you may get more advantage in ECC6.0 system because of Archive routing and also upgraded Write, Delete programs (upgraded program will depend on your current program in 4.6 system). For example In 4.6 system for archive MM_EKKO write program RM06EW30, archiving of purchase document is based on company code in the selection criteria and there is no preprocessing functionality. In ECC 6.0 you can archive by purchase organization selection criteria and preprocessing functionality will additional help in archiving of PO.
    In case if you archive documents in 4.6 and later upgrade it to ECC 6.0, SAP system will assure you to retrieve the archived data.
    With this i can say that going with archiving after upgrade to ECC 6.0 system will be better with respect to archiving process.
    -Thanks,
    Ajay

  • Differrences between structure and table in data dictionary in ABAP?

    What is the differrences between structure and table in data dictionary in ABAP?
    are they same?

    Tables :
    1. The place where the data is stored so that you can retrieve at any time.
    2. There can be more than one record stored
    Structures :
    1. The data / info stays only during the runtime of the application and will not get stored at all during the run time ....
    2. Only one record can be stored at the runtime .....
    A structure forms the skeleton of the table.
    A structure comprises components i.e., fields. Types are defined for the components A component can refer to an elementary type (via a data element or by directly specifying the data type and length in the structure definition), another structure or a table type. A structure can therefore be nested to any depth
    Tables can be defined independently of the database in the ABAP Dictionary. The fields of the table are defined with their (database-independent) data types and lengths.
    When the table is activated, a physical table definition is created in the database for the table definition stored in the ABAP Dictionary. The table definition is translated from the ABAP Dictionary to a definition of the particular database.

Maybe you are looking for

  • Using Time Machine to backup FROM a network drive

    Can time machine be configured to backup files that are stored on a network drive? Drive is mounted via SMB. Thanks Simon

  • How do I download the newer version of iTunes on a computer ?

    I'm trying to get on iTunes from my computer which has Windows . But it won't let me open the app . A little window pops up saying "The file 'iTunes Library .itl' cannot be read because it was created by a newer version of iTunes . I want to open it

  • Color Laserjet 2600n Windows 7 -64 Bit Looking for HP Toolbox

    ‎01-09-2014 08:30 AM I'm having trouble finding the HP Toolbox.    I have a Color Laserjet 2600n.   OS is Windows 7 - 64 Bit.  When I print checks, I notice a slight shadowing of the numbers.  Thinking it may need cleaned, but I cant find my toolbox

  • RFC Connection

    Hi We have upgraded BW and applied the JAVA stack etc but we are getting a problem when trying to use the report designer. When we try and open something or create a new report the following message pops up 'Error while loading query metadata, check

  • Array of culster

    Hi all : please see attach file ,, I bulid array (0-15) of cluster each cluster had 6 number SGL type .. ,, i want to separate this array  and culster  but according to me idea if i build  array (0-15) of cluster each cluster had 6 number SGL type  ,