Data Extraction template

Hi,
    In my current project there is a requirement for Data migration from legacy, So can anyone please help me on providing the data extraction template for Vendor and customer open line items, G/L balances and for Bank directory,
   Your help in this regard is highly appreciated
Thanks
Rajesh. R

Hi Rajesh,
When you extract the data from the legacy system the point that you should keep in mind are,
1. How the organisation structure in legacy system is mapped in SAP system. Becuase the data upload into sap should also happen in the way the reports are expected from SAP.
There is no standard layout which is used while extracting, you need to make sure that you extract all the necessary information from legacy system which is need to be uploaded in SAP.
I can give you an example of the field that you may include in the layout of extraction
Vendor account, Document data, Posting date, Document Type, Company code, Amount in Document currency, and local currency, etc
2. Please keep in mind that there are some GL account which will be maintained in SAP as open item basis. Therefore your extraction should also possibly happen each transaction wise.
3. There are certain GL which will be maintained in foreign currency, line bank GL which are in foreign currency. In such cases you need to extract the balance in foreign currency.
My suggestion to you will be thinking through the precess first and then go ahead with the extraction.
Hope this helps
Regards
Paul

Similar Messages

  • SAP HR Data Extract Templates - Download

    Hi Gurus,
    I would appreciate if any one can help me in downloading the Infotype templates used for data conversion.
    I wanted to know if SAP has provided any templates for the standard infotypes with fields. ( If its available in Service Market Place ) to be downloaded
    Regards,
    SAP HCM

    There was a LSMW training manual with templates for all the common infotypes for HR floating about one of the 'other' forums few years back.
    Or also see
    OSS notes (they have attachments)
    1348460 SAP Best Practices for Data Migration v1.603
    1253725 New Data Migration package for SAP Business All-in-One

  • How to add new fields to a data extract

    The following data extract program generates an output file which is displayed using a publisher template as a check format report.
    Oracle Payments Funds Disbursement Payment Instruction Extract 1.0
    How would I add new fields to the generated file? In other words what is the procedure to add new fields to the data extract?
    Thanks

    Do anyone pls advise how to customize the payment extraction program? We also have the similar requirement to extract extra fields to format a payment file.

  • Infocube data extraction using ABAP

    Hi,
    We need to extract Charactristics values data from Infocube but LISTCUBE cannot be used because of reasons of flexibilty.
    can any body suggest a way of the data extraction from the Infocube.I tried using the 'RSD_CUBE_DATA_GET' but not to much success also SAP will not support for later releases.
    If anybody has a template program can you plaese send it.
    at email id [email protected]
    Thanks and Regards,
    Arunava

    Hi
    In BW v3.1 have been able to use function module RSDRI_INFOPROV_READ,  (- see demo program RSDRI_INFOPROV_READ_DEMO).
    Hope this is of some help.
    Regards,

  • External definition and used them as Data Type Templates

    hi all
    how to Develop external definition and used them as Data Type Templates .
    Thank u
    Swari

    HI
    External Definitions: If a definition of the message structure exists in wsdl, xsd, dtd formats, we can use it in the integration repository by importing it as an external definition.
    External Definations:
    An external definition enables you to import a local WSDL, XSD, or DTD file to the Integration Repository and specify which parts of the schema to extract as the description for a message. You can use these extracted message schema as:
    Output or input messages in message interfaces
    Source or target structures for message mappings
    There is a series of standard schema for describing the message structure at runtime. WSDL (Web Service Description Language), XSD (XML Schema Definition Language), or DTDs (Document Type Definitions) are often used to describe message schema. If a description of the message structure already exists in one of the above formats, you can use it in the Integration Repository by importing it as an external definition rather than re-entering it manually using the data type editor.
    Have a look at
    http://help.sap.com/saphelp_nw04/helpdata/en/43/ce49ef1019087ae10000000a1553f6/content.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/ba/d4c23b95c8466ce10000000a114084/content.htm
    http://help.sap.com/saphelp_nw04/helpdata/en/26/9e97b0f525d743882936c2d6f375c7/content.htm
    cheers

  • MDMGX - After Data Extraction

    Hi,
    I am new for MDM Generic Extraction concept and understand the process up to generation of XSD and XML file generation. Below are the doubts in my mind. Please help me to get clear.
    1. What is the use of XSD file generation. Is it used for source template for MDM import manager?
    2. What is the use of Time out option in 'Define Repositories and FTP Server Details'.
    3. How will import generated multiple XML files into MDM server via Import Manager. For Ex
    1. Repository Fields :
    Product ID -- UNIQUE Key FIELD
    Product Desc
    Country
    Counry Description
    ISO Code
    Field1
    Field2
    2. Extracted XML files from MDMGX will be
    File1 --> Product ID & Desc
    File2 ---> Country & Country Desc
    Please take the above ex or any valid examples and explain.
    Thanks in advance.

    Hi Rakesh,
    SAP has delivered standard extraction for Reference Data and for Master Data
    The T-Code MDMGX is for Reference Data and it is used to load the Sub Table of MDM
    The MDM business content contains the standard ports and maps which is required for reference data
    The below thread explains the procedure to configure the MDMGX
    Extract Data usnig MDMGX
    There are sequences in extarcting the data.You can load only the revelant sub tables which is used in your repository using the selection creteria.
    There is no XI/PI required as you can configure FTP or you can download it your desktop and manually load
    Master Data Extraction
    The T-Code MDM_CLNT_EXTR is used for the Master Data Extraction.
    The distribution model is required for this and the configuration at PI is required
    Follow the below link for more details
    MDM_CLNT_EXTR
    Regards,
    Antony

  • View/Modify Funds Disbursement Data Extract

    In Oracle R12 payment format tied to the bank account dictates which data extract would be used and which XML template would be applied to the data. My question is how would I see the content of a data extract and I would I add more fields to the extract? I tried using XML Publisher responsibility but could not get answer to my 2 questions. I would really appreciate if someone could guide in the right direction. Thanks.

    Well, when you create a template in XML Publisher, you have to assign the Data Definition you want to use for this template. The Data Definition also sits in XML Publisher. Click on the Data Definitions tab in XML Publisher if you want to see the Data Definition.
    So the process, if you were creating something new, is this - 1) Create the Data Definition in XML Publisher. Have to create this first. 2. Then create the template and assign the proper data definition in that template.
    Does this make sense? Maybe you have an XML Publisher guru there you can talk to. Or you can download the XML Publisher Administration and Developer's Guide for R12, and study up on this that way. Pretty hard in a forum like this to give you detailed information on something like this.
    Hopefully this helps a bit.
    John Dickey
    Edited by: John Dickey, McCarthy on Mar 26, 2010 2:50 PM

  • IPayment Data extract customization

    For a payment process we are using Data extract  'Oracle Payments Funds Disbursement Payment Instruction Extract 1.0' and a custom template.
    We need some data which is not available in this extract. So I think either we have customize this extract.
    I would like to known how this Data extract calls the package and and how to make changes to that PKG.

    Check MEtalink, there's plenty of information on how to customize the extract programs to add fields not already included.

  • I am using the big date calendar template and when I submit it to apple for printing I lose the name of two months. These names are not text boxes. I see the names when I send it in but something happens during the transmission to apple. It was suggested

    I am using the big date calendar template in iPhoto. I am on Lion 10.7.2, macbook air. The names of the months are on each calendar page but something happens when I send the data to Apple. The names are part of the template. They are not text boxes. I lose two names on the calendar after it is sent to Apple. Apple suggested I make a pdf file of my calendar before sending it in and check to make sure every name shows. I did this with a calendar I just sent in. The calendar was correct. All names of the months were showing. After sending the data two month names disappeard because when it arrived by mail, it was incorrect. Apple looked at my calendar via a pdf file and it was incorrect.  This is second time this has happened. I called Apple and they had me delete several folders in the Library folder, some preferences and do a complete reinstall of iPhoto.  I have not yet remade the defective calendar. I am wondering if anyone else has had this problem?
    kathy

    Control-click on the background of the view all pages window and select "Preview Calendar" from the contextual menu.
    You can also save the pdf as a file to compare to the printed calendar.  If the two names are visible in the pdf file then the printed copy should show them.  Contact Apple for a refund.  Apple Print Products - Apple Store (U.S.)

  • BODS 3.1 : SAP R/3 data extraction -What is the difference in 2 dataflows?

    Hi.
    Can anyone advise as to what is the difference  in using the data extraction flow for extracting Data from SAP R/3 ?
    1)DF1 >> SAPR/3 (R3/Table -query transformation-dat file) >>query transformation >> target
    This ABAP flow generates a ABAP program and a dat file.
    We can also upload this program and run jobs as execute preloaded option on datastore.
    This works fine.
    2) We also can pull the SAP R/3 table directly.
    DF2>>SAPR/3 table (this has a red arrow like in OHD) >> Query transformation >> target
    THIS ALSO Works fine. And we are able to see the data directly into oracle.
    Which can also be scheduled on a job.
    BUT am unable to understand the purpose of using the different types of data extraction flows.
    When to use which type of flow for data extraction.
    Advantage / disadvantage - over the 2 data flows.
    What we are not understanding is that :
    if we can directly pull data from R/3 table directly thro a query transformation into the target table,
    why use the Flow of creating a R/3 data flow,
    and then do a query transformation again
    and then populate the target database?
    There might be some practical reasons for using these 2 different types of flows in doing the data extraction. Which I would like to understand.  Can anyone advise please.
    Many thanks
    indu
    Edited by: Indumathy Narayanan on Aug 22, 2011 3:25 PM

    Hi Jeff.
    Greetings. And many thanks for your response.
    Generally we pull the entire SAP R/3 table thro query transformation into oracle.
    For which we use R/3 data flow and the ABAP program, which we upload on the R/3 system
    so as to be able to use the option of Execute preloaded - and run the jobs.
    Since we do not have any control on our R/3 servers nor we have anyone on ABAP programming,
    we do not do anything at the SAP R/3 level
    I was doing this trial and error testing on our Worflows for our new requirement
    WF 1 : which has some 15 R/3 TABLES.
    For each table we have created a separate Dataflow.
    And finally in between in some dataflows, wherein, the SAP tables which had lot of rows, i decided to pull it directly,
    by-passing the ABAP flow.
    And still the entire work flow and data extraction happens ok.
    In fact i tried creating a new sample data flow and tested.
    Using direct download and - and also execute preloaded.
    I did not see any major difference in time taken for data extraction;
    Because anyhow we pull the entire Table, then choose whatever we want to bring into oracle thro a view for our BO reporting or aggregate and then bring data as a table for Universe consumption.
    Actually, I was looking at other options to avoid this ABAP generation - and the R/3 data flow because we are having problems on our dev and qa environments - giving delimiter errors.  Whereas in production it works fine. Production environment is a old set up of BODS 3.1. QA and Dev are relatively new enviornments of BODS. Which is having this delimiter error.
    I did not understand how to resolve it as per this post : https://cw.sdn.sap.com/cw/ideas/2596
    And trying to resolve this problem, I ended up with the option of trying to pull directly the R/3 table. Without using ABAP workflow.  Just by trial and error of each and every drag and drop option. Because we had to urgently do a POC and deliver the data for the entire e recruiting module of SAP. 
    I dont know whether i could do this direct pulling of data - for the new job which i have created,
    which has 2 workflows with 15 Dataflows in each worflow.
    And and push this job into production.
    And also whether i could by-pass this ABAP flow and do a direct pulling of R/3 data, in all the Dataflows in the future for ANY of our SAP R/3 data extraction requirement.  And this technical understanding is not clear to us as regards the difference between the 2 flows.  And being new to this whole of ETL - I just wanted to know the pros and cons of this particular data extraction. 
    As advised I shall check the schedules for a week, and then we shall move it probably into production.
    Thanks again.
    Kind Regards
    Indu
    Edited by: Indumathy Narayanan on Aug 22, 2011 7:02 PM

  • Open data extraction orders -  Applying Support Packs

    Dear All,
    I have done the IDES 4.6C SR2 installation.
    While updating the support packs, i get the message saying
    CHECK_REQUIREMENTS phase.
    Open data extraction orders
    There are still open data extraction orders in the system
    process these before the start of the object import because changes to the ABAP Dictionary structures could lead to data extraction orders not being able to be read after the import and their processing terminating
    For more details about this problem, see Note 328181.
    Go to the Customizing cockpit for data extraction and start the processing of all open extraction orders.
    I have checked the Note.
    But this is something m facing for the first time.
    Any suggestion!!!
    Rgds,
    NK

    The exact message is :
    Phase CHECK_REQUIREMENTS: Explanation of the Errors
    Open Data Extraction Requests
    The system has found a number of open data extraction requests. These
    should be processed before starting the object import process, as
    changes to DDIC structures could prevent data extraction requests from
    being read after the import, thus causing them to terminate. You can
    find more information about this problem in SAP Note 328181.
    Call the Customizing Cockpit data extraction transaction and process all
    open extraction requests.

  • Bulk API V2.0 Data extract support for additional objects (Campaign,Email,Form,FormData,LandingPage)?

    allison.moore
    Any plans for adding following objects under Bulk API V2.0 for data extraction from Eloqua. Extracting the data using the REST API for these objects makes it complicated.

    Thanks for quick response. Extracting these objects using REST API in depth=Complete poses lots of complication from the code perspective since these object(s) contains multiple nested or embedded objects within it. is there any guideline on how to extract these objects using REST so that we can get all the data which is required for analysis/reporting.

  • Data Extraction and ODS/Cube loading: New date key field added

    Good morning.
    Your expert advise is required with the following:
    1. A data extract was done previously from a source with a full upload to the ODS and cube. An event is triggered from the source when data is available and then the process chain will first clear all the data in the ODS and cube and then reload, activate etc.
    2. In the ODS, the 'forecast period' field was now moved from data fields to 'Key field' as the user would like to report per period in future. The source will in future only provide the data for a specific period and not all the data as before.
    3) Data must be appended in future.
    4) the current InfoPackage in the ODS is a full upload.
    5) The 'old' data in the ODS and cube must not be deleted as the source cannot provide it again. They will report on the data per forecast period key in future.
    I am not sure what to do in BW as far as the InfoPackages are concerned, loading the data and updating the cube.
    My questions are:
    Q1) How will I ensure that BW will append the data for each forecast period to the ODS and cube in future? What do I check in the InfoPackages?
    Q2) I have now removed the process chain event that used to delete the data in the ODS and cube before reloading it again. Was that the right thing to do?
    Your assistance will be highly appreciated. Thanks
    Cornelius Faurie

    Hi Cornelius,
    Q1) How will I ensure that BW will append the data for each forecast period to the ODS and cube in future? What do I check in the InfoPackages?
    -->> Try to load data into ODS in Overwrite mode full update asbefore(adds new records and changes previous records with latest). Pust delta from this ODS to CUBE.
    If existing ODS loading in addition, introduce one more ODS with same granularity of source and load in Overwrite mode if possible delta or Full and push delta only subsequently.
    Q2) I have now removed the process chain event that used to delete the data in the ODS and cube before reloading it again. Was that the right thing to do?
    --> Yes, It is correct. Otherwise you will loose historic data.
    Hope it Helps
    Srini

  • FI data extraction help

    HI All,
    I have gone thorugh the sdn link for FI extarction and founf out to be very  useful.
    Still i have some doubts....
    For line item data extraction...........Do we need to etract data from 0FI_GL_4, 0FI_AP_4, 0FI_AR_4, into ODS0FIGL_O02(General Ledger: Line Items) ? If so do we need to maintain transformation between ods and all three DS?
    ALso Please educate me on 0FI_AP_3 and 0FI_AP_4 data sources.......

    >
    Jacob Jansen wrote:
    > Hi Raj.
    >
    > Yes, you should run GL_4 first. If not, AP_4 and AR_4 will be lagging behind. You can see in R/3 in table BWOM_TIMEST how the deltas are "behaving" with respect to the date selection.
    >
    > br
    > jacob
    Not necessarily for systems above plug in 2002.
    As of Plug-In 2002.2, it is no longer necessary to have DataSources linked. This means that you can now load 0FI_GL_4, 0FI_AR_4, 0FI_AP_4, and 0FI_TX_4 in any order. You also have the option of using DataSources 0FI_AR_4, 0FI_AP_4 and 0FI_TX_4 separately without 0FI_GL_4. The DataSources can then be used independently of one another (see note 551044).
    Source:http://help.sap.com/saphelp_nw04s/helpdata/EN/af/16533bbb15b762e10000000a114084/frameset.htm

  • Generic Data Extraction From SAP R/3 to BI server using Function Module

    Hi,
    I want step by step procedure for Generic Extraction from SAP R/3 application to BI application
    using Functional module.
    If any body have any Document or any PPT then please reply and post it in forum, i will give point for them.
    Thanks & Regards
    Subhasis Pan

    please go thr this link
    [SAP BI Generic Extraction Using a Function Module|https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/com.sap.km.cm.docs/library/business-intelligence/s-u/sap%20bi%20generic%20extraction%20using%20a%20function%20module.pdf]
    [Generic Data Extraction Using Function Module |Re: Generic Data Extraction Using Function Module;

Maybe you are looking for

  • Workbook Migration-System error in program CL_RSR_XLS_VIEW and from ASSIGN_

    Hi Experts, We are migrating workbooks from 3.x to BI 7. While opening the workboon from BI 7 Analyzer we are getting the following System Error  and Disconnecting from BW server. 'Program Error in class SAPMSSY1 method UNCAUGHT_EXCEPTION. System err

  • New Aluminum Mac Display

    I have a MacBook Pro and a new iMac. The Text on the iMac is a little bit blurry compared to the text in the MacBook Pro (Core Dual 1). Any suggestions would be appreciated. I tried to switch off the zoom feature but it did not improve the text rende

  • How to install Cisco Device Manger on Window 7

    Hello Is Cisco device manager supported on windows 7 ? If yes how to install Cisco Device Manger on Window 7? Please advise Thanks Chetan R

  • Download Firefox Installer for Windows xp for disconnected networks

    Hello. I want to obtain a complete install package (exe, etc) of Firefox 18 or 19 as my preferred browser on non-internet facing nodes. I have read that installers are calling out for updates upon completion and the nodes in questions cannot connect

  • OTL Employee Rotation Plan Query

    Hello, I need help on below Select Statement. We have created OTL Work plan as below : Work week starts on Sunday. Example: Su M Tu W Th F Sa (1) REGDY - Tu W Th (2) OFFDY - Su M F (3) RESTDY - Sa i need to fetch above information from backend i have