What is data archiving and DMS(Data Management System) in SAP

what is data archiving and DMS(Data Management System) in SAP
Welcome to SCN. Before posting questions please search for available information here and in the web. Please also read the Rules of Engagement before further posting.
Edited by: kishan P on Aug 31, 2010 1:06 PM

Hi,
Filtering at the IDoc Level
Identify the filter object (BD59)
Modify the distribution model
Segment Filtering
specify the segments to be filtered (BD56)
The Reduced IDoc Type
Analyze the data.
Reduce the IDoc type (BD53)
Thanks and regards.

Similar Messages

  • Link between table EKKO-EBELN and DMS - Document Management System

    Hi to all,
    I wanted to know the form to connect or link the Document Info Record (inside "Define Object Link") to object DRAW (DMS) with table EKKO-EBELN (order head). In standard form it is possible to do it with EKPO (Purchase order item) This is in transaction DC10 Define Object Link.
    Thanks!
    Greetings
    Martin

    Hi Martin,
    Within table DRAD you can see all linked objects to a specific document info record (DIR).
    To be able to link objects to DIRs you have to maintain the correct objects in customizing.
    First of it is necessary to maintain the correct screen numbers for all needed objects within the customizing. Therefore go to transaction SPRO and select
    Cross-Application Components
    > Document Management
    > Control Data
    > Maintain Screen for Object Link
    Here you can enter all needed SAP objects with the name and
    the related screen number. If you do not know the object or the screen number you can display a list of all SAP standard objects in transaction SE80 under function group CV130 ("Screens").
    So for material master you maintain the following entry:
    SAP object Screen Description Auth.
    MARA 201 material master 1
    Please note that the system automatically adds the leading "1" which is displayed in function group CV130 under
    "Screens". So for object MARA you just have to enter 201 instead of 1201.
    The value "1" in the column "Auth." is necessary to grant that always the latest dynpros and authorizations were used when linking this object to an document info record. This will avoid different problems by displaying linked documents from the object side.
    The next step is to decide which objects should be used in the different document types. Therefore please go to transaction DC10 and choose the relevant document type. Afterwards please select "Define object links". Here you can
    add all needed objects only by entering the object. The correct screen number is taken automatically. Also note the blue information buttons in the upper right corner which contain further necessary information about the possible settings of the current screen.
    After doing these settings in customizing you should be able to link the maintained objects to document info records and oppositely by using the DMS transaction CV02N (tab "object links") or the object related transactions.
    Important notes regarding the object link functionalities are:
    Note 880077: Performance problem using object links
    Note 885977: Extension to note 867241 on object links
    Note 926060: Extension of note 885977 for object link Purchase Order
    Note 960202: Object Links Deleted While Adding New Objects
    Note 1002648: Wrong behavior while using RFC_CHANGE_DOCUMENT_MASTER

  • What should i do to implement Data Archiving and Data Reporting

    we want to implement Data Archiving and Data Reporting in our product. Can someone tell me what techniques or approaches people take in general to implement Data Archiving and Data reporting.
    i am currently looking into DataWarehousing. is this the right apporach ? i have no idea as where i should start on this. can someone give me a good direction as a starting point.
    thank you,
    Puja

    Did you setup Find My Mac before it was stolen?

  • Data archiving and data cleansing

    hi experts,
    Can any one tell me step by step guide for Data archiving and Data cleansing of SAP-ISU Object.
    what is the difference between  Data archiving and Data cleansing .
    Thanx & Rgds

    Data Archiving: So many objects are there you can look some of them..........
    ISU_BBP IS-U Archiving: Budget Billing Plan
    ISU_BCONT Business Partner Contacts (Contract A/R + A/P)
    ISU_BILL IS-U Archiving: Billing Document Header
    ISU_BILLZ IS-U Archiving: Billing Line Item
    ISU_EABL IS-U Archiving: Meter Reading Results
    ISU_EORDER IS-U Archiving: Waste Disposal Order
    ISU_EUFASS Archiving of Usage Factors
    ISU_FACTS Installation Facts
    ISU_INSPEC IS-U Archiving: Campaigns for Inspection List
    ISU_PPM Prepayment Documents
    ISU_PRDOCH IS-U Archiving: Print Document Header
    ISU_PRDOCL IS-U Archiving: Print Document Line Item
    ISU_PROFV IS-U Archiving: EDM Profile Values
    ISU_ROUTE IS-U Archiving: Waste Disposal Route
    ISU_SETTLB Settlement Document
    ISU_SWTDOC Archive Switch Document
    Go to SARA t-code and enter the object CA_BUPA for business partner then Press F6 you will get  step by step documentation. please follow the procedure for all the objects.
    Regards,
    Siva

  • Goldengate Extracts reads slow during Table Data Archiving and Index Rebuilding Operations.

    We have configured OGG on a  near-DR server. The extracts are configured to work in ALO Mode.
    During the day, extracts work as expected and are in sync. But during any dialy maintenance task, the extracts starts lagging, and read the same archives very slow.
    This usually happens during Table Data Archiving (DELETE from prod tables, INSERT into history tables) and during Index Rebuilding on those tables.
    Points to be noted:
    1) The Tables on which Archiving is done and whose Indexes are rebuilt are not captured by GoldenGate Extract.
    2) The extracts are configured to capture DML opeartions. Only INSERT and UPDATE operations are captured, DELETES are ignored by the extracts. Also DDL extraction is not configured.
    3) There is no connection to PROD or DR Database
    4) System functions normally all the time, but just during table data archiving and index rebuild it starts lagging.
    Q 1. As mentioned above, even though the tables are not a part of capture, the extracts lags ? What are the possible reasons for the lag ?
    Q 2. I understand that Index Rebuild is a DDL operation, then too it induces a lag into the system. how ?
    Q 3. We have been trying to find a way to overcome the lag, which ideally shouldn't have arised. Is there any extract parameter or some work around for this situation ?

    Hi Nick.W,
    The amount of redo logs generated is huge. Approximately 200-250 GB in 45-60 minutes.
    I agree that the extract has to parse the extra object-id's. During the day, there is a redo switch every 2-3 minutes. The source is a 3-Node RAC. So approximately, 80-90 archives generated in an hour.
    The reason to mention this was, that while reading these archives also, the extract would be parsing extra Object ID's, as we are capturing data only for 3 tables. The effect of parsing extract object id's should have been seen during the day also. The reason being archive size is same, amount of data is same, the number of records to be scanned is same.
    The extract slows down and read at half the speed. If normally it would take 45-50 secs to read an archive log of normal day functioning, then it would take approx 90-100 secs to read the archives of the mentioned activities.
    Regarding the 3rd point,
    a. The extract is a classic extract, the archived logs are on local file system. No ASM, NO SAN/NAS.
    b. We have added  "TRANLOGOPTIONS BUFSIZE" parameter in our extract. We'll update as soon as we see any kind of improvements.

  • What is open data set and close data set

    what is open data set and close data set,
    how to use the files in sap directories ?

    hi,
    Open Dataset is used to read or write on to application server ... other than that i am not sure that there exists any way to do the same ... here is a short description for that
    FILE HANDLING IN SAP
    Introduction
    • Files on application server are sequential files.
    • Files on presentation server / workstation are local files.
    • A sequential file is also called a dataset.
    Handling of Sequential file
    Three steps are involved in sequential file handling
    • OPEN
    • PROCESS
    • CLOSE
    Here processing of file can be READING a file or WRITING on to a file.
    OPEN FILE
    Before data can be processed, a file needs to be opened.
    After processing file is closed.
    Syntax:
    OPEN DATASET <file name> FOR {OUTPUT/INPUT/APPENDING}
    IN {TEXT/BINARY} MODE
    This statement returns SY_SUBRC as 0 for successful opening of file or 8, if unsuccessful.
    OUTPUT: Opens the file for writing. If the dataset already exists, this will place the cursor at the start of the dataset, the old contents get deleted at the end of the program or when the CLOSE DATASET is encountered.
    INPUT: Opens a file for READ and places the cursor at the beginning of the file.
    FOR APPENDING: Opens the file for writing and places the cursor at the end of file. If the file does not exist, it is generated.
    BINARY MODE: The READ or TRANSFER will be character wise. Each time ‘n’’ characters are READ or transferred. The next READ or TRANSFER will start from the next character position and not on the next line.
    IN TEXT MODE: The READ or TRANSFER will start at the beginning of a new line each time. If for READ, the destination is shorter than the source, it gets truncated. If destination is longer, then it is padded with spaces.
    Defaults: If nothing is mentioned, then defaults are FOR INPUT and in BINARY MODE.
    PROCESS FILE:
    Processing a file involves READing the file or Writing on to file TRANSFER.
    TRANSFER Statement
    Syntax:
    TRANSFER <field> TO <file name>.
    <Field> can also be a field string / work area / DDIC structure.
    Each transfer statement writes a statement to the dataset. In binary mode, it writes the length of the field to the dataset. In text mode, it writes one line to the dataset.
    If the file is not already open, TRANSFER tries to OPEN file FOR OUTPUT (IN BINARY MODE) or using the last OPEN DATASET statement for this file.
    IF FILE HANDLING, TRANSFER IS THE ONLY STATEMENT WHICH DOES NOT RETURN SY-SUBRC
    READ Statement
    Syntax:
    READ DATASET <file name> INTO <field>.
    <Field> can also be a field string / work area / DDIC structure.
    Each READ will get one record from the dataset. In binary mode it reads the length of the field and in text mode it reads each line.
    CLOSE FILE:
    The program will close all sequential files, which are open at the end of the program. However, it is a good programming practice to explicitly close all the datasets that were opened.
    Syntax:
    CLOSE DATASET <file name>.
    SY-SUBRC will be set to 0 or 8 depending on whether the CLOSE is successful or not.
    DELETE FILE:
    A dataset can be deleted.
    Syntax:
    DELETE DATASET <file name>.
    SY-SUBRC will be set to 0 or 8 depending on whether the DELETE is successful or not.
    Pseudo logic for processing the sequential files:
    For reading:
    Open dataset for input in a particular mode.
    Start DO loop.
    Read dataset into a field.
    If READ is not successful.
    Exit the loop.
    Endif.
    Do relevant processing for that record.
    End the do loop.
    Close the dataset.
    For writing:
    Open dataset for output / Appending in a particular mode.
    Populate the field that is to be transferred.
    TRANSFER the filed to a dataset.
    Close the dataset.
    Regards
    Anver
    if hlped pls mark points

  • How to upload the data to SAP-DMS(Document Management System)

    Dear sir,
    How to upload the data to SAP-DMS (Document Management System) please give me the code for this .ex for (excel format or tab  deliminated) to sap system through BAPI i have to pass.
    regards
    R M Patil

    Hi,
    May be following link may help you.
    /thread/315373 [original link is broken]
    /message/205534#205534 [original link is broken]
    Thanks,
    Sree.

  • Data archival and purging for OLTP database

    Hi All,
    Need your suggestion regarding data archival and purging solution for OLTP db.
    currently, we are planning to generate flat files from table..before purging the inactive data and move them to tapes/disks for archiving then purge the data from system. we have many retention requirements and conditions before archival of data. so partition alone is not sufficient.
    Is there any better approach for archival and purging other than this flat file approach..
    thank you.
    regards,
    vara

    user11261773 wrote:
    Hi All,
    Need your suggestion regarding data archival and purging solution for OLTP db.
    currently, we are planning to generate flat files from table..before purging the inactive data and move them to tapes/disks for archiving then purge the data from system. we have many retention requirements and conditions before archival of data. so partition alone is not sufficient.
    Is there any better approach for archival and purging other than this flat file approach..
    thank you.
    regards,
    varaFBDA is the better option option .Check the below link :
    http://www.oracle.com/pls/db111/search?remark=quick_search&word=flashback+data+archive
    Good luck
    --neeraj                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • HT4847 How i can download my backup data? And how to manage the data on i Cloud?

    How i can download my backup data? And how to manage the data on i Cloud?

    You can't download an iCloud backup, except to restore it to your device should you ever need to.
    iCloud data can be managed within the apps on your iPad.  Any changes to the data within the apps corresponding to the data you are syncing with iCloud will take place in iCloud.  You can also manage some of this data on icloud.com from your computer.
    This article explains ways to manage your iCloud storage space, should you need to reduce you iCloud storage: http://support.apple.com/kb/ht4847.

  • How to make a Infocube as data source and upload data? urgent

    Hi all,
    I have 2 same InfoCubes, and I want to make the test InfoCube(ZCO_1T) as data source and upload data to another Cube(ZCO_1), Now I has Generated Export DataSource on ZCO_1T and create a update rule between ZCO_1T and ZCO_1. what will I do the next step? Thanks.

    hi delve,
    once u generate datasource from test cube it acts as an data source for further upload.
    now u choose update target3.x in additional functions in context menu of the test cube.
    u will get an infopackage pop up with data target to the cube ZCO_1,processing to datatarget only.
    selection criterion can bve selected from the first tab.
    under schedule tab you can start extraction.
    reward points if helpful.

  • UTC Date Time and Normal Date Time

    Hi All,
    1. How UTC date time and Normal date time differs in siebel.
    2. If legacy data needed to be loaded into siebel, in siebel few fields are date time and UTC date time fields. what would happen if we load both normal date time and UTC date time without considering them techinically?
    3. UTC date time holds any specific format in physical database? If we want to load legacy data to UTC date time format what is the query to manipulate it?
    Thankyou
    Sean

    Sean,
    Please check document below, I believe it has most of the answers to the questions you have:
    http://download.oracle.com/docs/cd/E14004_01/books/GlobDep/GlobDepUTC.html
    Hope it helps,
    Wilson

  • Difference between Realtime Data Acqisition and Direct Data Access ?

    Hi experts,
    What is RDA and DDA and whrere these two are used ?
    What is the Difference between Realtime Data Acqisition and Direct Data Access?
    Please explain me in a detail manner, i will assign points for ur valuable answers..............

    Hi,
    Eventhough the aim of the two methods are to report on the latest data, there is a difference in both.
    In Realtime Data Acqisition you are loading the data into BW and then reporting on that data from BW Infoproviders.
    http://help.sap.com/saphelp_nw04s/helpdata/en/52/777e403566c65de10000000a155106/content.htm
    I think you are refering to Virtual providers by the term Direct Data Access. In  Direct Data Access you are not loading data into BW. Data is residing in R/3 itself. Your reports will fetch data from R/3 directly. The data is called during execution of a query in the source system.
    http://help.sap.com/saphelp_nw04s/helpdata/en/23/c0234239f75733e10000000a155106/frameset.htm
    http://help.sap.com/saphelp_nw70/helpdata/EN/13/3e34429692b76be10000000a155106/frameset.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/da/5909392a430303e10000000a114084/frameset.htm
    Hope this helps.
    Thanks,
    JituK

  • Need help regarding Simple Data Client and Simple Data Server VIs

    Hi everyone.
    I have a simple objective. I just want to test the 2 example VIs, "Simple Data Client" and "Simple Data Server" between 2 computers. I just want to check whether is this working between the 2 computers.
    What I have done for now is that I changed the "address", from "localhost" in the "Simple Data Client.vi" to the IP address of the computer running the "Simple Data Server". I runned the "Simple Data Server" VI in one of the computers first followed by the "Simple Data Client" in the other computer. Nothing is received and the client just timed out.
    Can anyone please help me troubleshoot and tell me what are the possible problems for this? Are there any wires connections between 2 computers that I am missing or any other configurations I have to make before I can successfully do this?
    Thanks.
    Regards,
    Jonathan

    Hi Lee.P.
    I understand that. I was just feeling frustrated about the project not working. Sincere apologies from me.
    I was wrong about the error number. It is not Error 60. It is Error 59 - The network is down, unreachable, or has been reset.. Yes, I have tried changing the port numbers at the 2 computers when trying to send/receive.
    Could anything else be the problem?
    Regards,
    Jonathan  

  • Download using open data set and close data set

    can any body please send some sample pgm using open data set and close data set .the data should get downloaded in application server
    very simple pgm needed

    Hi Arun,
    See the Sample code for BDC using OPEN DATASET.
    report ZSDBDCP_PRICING no standard page heading
    line-size 255.
    include zbdcrecx1.
    *--Internal Table To hold condition records data from flat file.
    Data: begin of it_pricing occurs 0,
    key(4),
    f1(4),
    f2(4),
    f3(2),
    f4(18),
    f5(16),
    end of it_pricing.
    *--Internal Table To hold condition records header .
    data : begin of it_header occurs 0,
    key(4),
    f1(4),
    f2(4),
    f3(2),
    end of it_header.
    *--Internal Table To hold condition records details .
    data : begin of it_details occurs 0,
    key(4),
    f4(18),
    f5(16),
    end of it_details.
    data : v_sno(2),
    v_rows type i,
    v_fname(40).
    start-of-selection.
    refresh : it_pricing,it_header,it_details.
    clear : it_pricing,it_header,it_details.
    CALL FUNCTION 'UPLOAD'
    EXPORTING
    FILENAME = 'C:\WINDOWS\Desktop\pricing.txt'
    FILETYPE = 'DAT'
    TABLES
    DATA_TAB = it_pricing
    EXCEPTIONS
    CONVERSION_ERROR = 1
    INVALID_TABLE_WIDTH = 2
    INVALID_TYPE = 3
    NO_BATCH = 4
    UNKNOWN_ERROR = 5
    GUI_REFUSE_FILETRANSFER = 6
    OTHERS = 7.
    WRITE : / 'Condition Records ', P_FNAME, ' on ', SY-DATUM.
    OPEN DATASET P_FNAME FOR INPUT IN TEXT MODE.
    if sy-subrc ne 0.
    write : / 'File could not be uploaded.. Check file name.'.
    stop.
    endif.
    CLEAR : it_pricing[], it_pricing.
    DO.
    READ DATASET P_FNAME INTO V_STR.
    IF SY-SUBRC NE 0.
    EXIT.
    ENDIF.
    write v_str.
    translate v_str using '#/'.
    SPLIT V_STR AT ',' INTO it_pricing-key
    it_pricing-F1 it_pricing-F2 it_pricing-F3
    it_pricing-F4 it_pricing-F5 .
    APPEND it_pricing.
    CLEAR it_pricing.
    ENDDO.
    IF it_pricing[] IS INITIAL.
    WRITE : / 'No data found to upload'.
    STOP.
    ENDIF.
    loop at it_pricing.
    At new key.
    read table it_pricing index sy-tabix.
    move-corresponding it_pricing to it_header.
    append it_header.
    clear it_header.
    endat.
    move-corresponding it_pricing to it_details.
    append it_details.
    clear it_details.
    endloop.
    perform open_group.
    v_rows = sy-srows - 8.
    loop at it_header.
    perform bdc_dynpro using 'SAPMV13A' '0100'.
    perform bdc_field using 'BDC_CURSOR'
    'RV13A-KSCHL'.
    perform bdc_field using 'BDC_OKCODE'
    '/00'.
    perform bdc_field using 'RV13A-KSCHL'
    it_header-f1.
    perform bdc_dynpro using 'SAPMV13A' '1004'.
    perform bdc_field using 'BDC_CURSOR'
    'KONP-KBETR(01)'.
    perform bdc_field using 'BDC_OKCODE'
    '/00'.
    perform bdc_field using 'KOMG-VKORG'
    it_header-f2.
    perform bdc_field using 'KOMG-VTWEG'
    it_header-f3.
    **Table Control
    v_sno = 0.
    loop at it_details where key eq it_header-key.
    v_sno = v_sno + 1.
    clear v_fname.
    CONCATENATE 'KOMG-MATNR(' V_SNO ')' INTO V_FNAME.
    perform bdc_field using v_fname
    it_details-f4.
    clear v_fname.
    CONCATENATE 'KONP-KBETR(' V_SNO ')' INTO V_FNAME.
    perform bdc_field using v_fname
    it_details-f5.
    if v_sno eq v_rows.
    v_sno = 0.
    perform bdc_dynpro using 'SAPMV13A' '1004'.
    perform bdc_field using 'BDC_OKCODE'
    '=P+'.
    perform bdc_dynpro using 'SAPMV13A' '1004'.
    perform bdc_field using 'BDC_OKCODE'
    '/00'.
    endif.
    endloop.
    *--Save
    perform bdc_dynpro using 'SAPMV13A' '1004'.
    perform bdc_field using 'BDC_OKCODE'
    '=SICH'.
    perform bdc_transaction using 'VK11'.
    endloop.
    perform close_group.
    Hope this resolves your query.
    Reward all the helpful answers.
    Regards

  • My Extensis suitcase 11.0.4 is no longer working with my system OSX10.6.8, what is the best and cheapest font management software to use that's compatible with this system and CS software

    my Extensis suitcaseX 11.0.4 is no longer working with my system OSX10.6.8, what is the best and cheapest font management software to use that's compatible with this system and my existing CS design software?

    If you upgraded to Snow Leopard, then it would seem that Suitcase 11 is not compatible with Snow Leopard and you'd need to buy Suitcase Fusion 3 for $99.95.

Maybe you are looking for

  • Problems deploying a simple EJB on Weblogic 8.1 using JDeveloper

    Hey guys, Title says it all. First I had one problem in that I could not test the connection to the Weblogic server. I moved the weblogic.jar file into the lib/ext folder and that test worked then. After that I wrote the EJB...a very simple one...and

  • Having trouble with image processor...help please!

    HI there...First off I just bought a new iMac and haven't a clue what I'm doing so my problem could be related to that.  While I'm waiting for Adobe to transfer over my platform I downloaded the trial of CS5 Extended to get me started.  I'm having tr

  • Error in saproot.sh

    hi just today i update kernel patch to 137(not latest) but when i run saproot.sh it show below: ./saproot.sh DEV The database environment must be set for your DB6 database. Login as <sid>adm and do 'su root' before starting this script or specify the

  • NAC and Multiple AD

    Hi all, I have a customer who is deploying NAC appliance in a multi-tenancy building. The solution is likely have NAC act as an authentication proxy for Active Directory. What isn't clear is whether there will be a single instance of Active Directory

  • Debugging J2EE Engine Portal Related Service

    Hi, I have AP5 Engine running in my machine. While starting the engine all the services related to Portal are starting, but once I stop and restart the engine again, the Portal Related service is not starting. I dont have files related to this. Can a