Data Conversion mapping u0096 BDC

I am coming across a scenario, as we are working on conversions, we are getting data from some 3rd party system into SAP and we need to update Infotypes using uploads.
Right now we are facing problem, which they have the separate data I mean their naming standard like they have different codes 01, 02, 03 and in SAP we have different ones.
What are the different ways that we map and validate our data before uploading it into SAP? They give us a master file extract from their 3rd party system and they want us to upload into SAP. In Sap we might have lot of difference of fields and value, which we might need to make sure and validate before we upload it in to SAP.
Could you please write your suggestions/inputs, how I can achieve this task.
Thanks

Hi,
You can also pre-populate an internal table with all the mapping values.  For example: if the input value is 01, the corresponding SAP value = 'One'.
So you create an internal table with two fields, first with the legacy values and second with SAP values.
Then when you are looping through the internal table of the input file read the above internal table with key = input file value and then use the corresponding value.
And if it is an interface (not conversion) and the values keep changing, they I would suggest you t maintian a Ztable.
Good luck.
Venu

Similar Messages

  • Data extraction / Conversion / Mapping / Miration

    Hi All,
       Can any one please explain the meaning of the below terms and how data is moved from Non SAP system to SAP system?.
    "Data extraction / Conversion / Mapping / Miration"
    Kindly give me detailed step by step instructions on how to accomplish this. Also if there is any documentation please forward.
    I appreciate the help in advance.
    Raj

    Hi Raj,
    We can use LSMW, BDC, eCATT to upload the data from Legacy to SAP system.
    Data Extraction - Data has to extract from legacy in to some flat file like  Xls.
    Conversion - Convert as per the need of SAP.
    Mapping - Mapping the extracted field with SAP field for upload.
    Migration - Data migration from legacy to SAP using any tool like LSMW, BDC, CATT.
    Good Luck
    Om

  • Conversion mapping is losing time zone data during daylight saving time

    We have a problem with conversion of Calendars to timestamp with timezone for the last hour of Daylight Saving Time (e.g. 01:00 EDT - 01:59 EDT) where it is being interpreted as Standard Time which is in reality 60 minutes later.
    We've written a JUnit test case that runs directly against TopLink to avoid any issues with WAS and its connection pooling.
    The Calendar theDateTime comes from an object called TimeEntry which is mapped to a TIMESTAMP WITH TIMEZONE field using conversion mapping with Data Type TIMESTAMPTZ (oracle.sql) and Attribute Type Calendar (java.util).
    We are using:
    Oracle TopLink - 10g Release 3 (10.1.3.0.0) (Build Patch for Bugs 5145690 and 5156075)
    Oracle9i Enterprise Edition Release 9.2.0.7.0 - 64bit Production
    Oracle JDBC driver Version: 10.2.0.1.0
    platform=>Oracle9Platform
    Execute this Java:
    SimpleDateFormat format = new SimpleDateFormat("MM/dd/yyyy HH:mm z");
    TimeZone tzEasternRegion = TimeZone.getTimeZone("US/Eastern");
    Calendar theDateTime = Calendar.getInstance(tzEasternRegion);
    theDateTime.setTime(format.parse("10/29/2006 01:00 EDT"));
    Persist to the database and execute this SQL:
    SELECT the_date_time, EXTRACT(TIMEZONE_REGION FROM the_date_time), EXTRACT(TIMEZONE_ABBR FROM the_date_time), EXTRACT(TIMEZONE_HOUR FROM the_date_time)
    FROM time_table WHERE time_table_id=1
    This provides the following results:
    THE_DATE_TIME EXTRACT(TIMEZONE_REGION FROM the_date_time) EXTRACT(TIMEZONE_ABBR FROM the_date_time) EXTRACT(TIMEZONE_HOUR FROM the_date_time)
    29-OCT-06 01.00.00.000000 AM US/EASTERN US/Eastern EST -5
    The wrong time zone is in the database. It should be EDT -4. Let's test the SQL that should be generated by TopLink. It should look like the following.
    Execute this SQL:
    UPDATE time_table SET the_date_time = TO_TIMESTAMP_TZ('10/29/2006 01:00 US/Eastern','mm/dd/yyyy HH24:MI TZR') WHERE (time_table_id=1)
    SELECT the_date_time, EXTRACT(TIMEZONE_REGION FROM the_date_time), EXTRACT(TIMEZONE_ABBR FROM the_date_time), EXTRACT(TIMEZONE_HOUR FROM the_date_time)
    FROM time_table WHERE time_table_id=1
    This provides the same results:
    THE_DATE_TIME EXTRACT(TIMEZONE_REGION FROM the_date_time) EXTRACT(TIMEZONE_ABBR FROM the_date_time) EXTRACT(TIMEZONE_HOUR FROM the_date_time)
    29-OCT-06 01.00.00.000000 AM US/EASTERN US/Eastern EST -5
    Now, execute this SQL:
    UPDATE time_table SET the_date_time = TO_TIMESTAMP_TZ('10/29/2006 01:00 EDT US/Eastern','mm/dd/yyyy HH24:MI TZD TZR') WHERE (time_table_id=1)
    SELECT the_date_time, EXTRACT(TIMEZONE_REGION FROM the_date_time), EXTRACT(TIMEZONE_ABBR FROM the_date_time), EXTRACT(TIMEZONE_HOUR FROM the_date_time)
    FROM time_table WHERE time_table_id=1
    This provides better results:
    THE_DATE_TIME EXTRACT(TIMEZONE_REGION FROM the_date_time) EXTRACT(TIMEZONE_ABBR FROM the_date_time) EXTRACT(TIMEZONE_HOUR FROM the_date_time)
    29-OCT-06 01.00.00.000000 AM US/EASTERN US/Eastern EDT -4
    The correct time zone is now in the database. Let's test reading this with the following Java:
    System.out.println("cal= " + theDateTime);
    System.out.println("date= " + theDateTime.getTime());
    System.out.println("millis= " + theDateTime.getTimeInMillis());
    System.out.println("zone= " + theDateTime.getTimeZone());
    This provides the following results:
    cal= java.util.GregorianCalendar[...]
    date= Sun Oct 29 01:00:00 EST 2006
    millis= 1162101600000
    zone= sun.util.calendar.ZoneInfo[id="US/Eastern",...]
    The TimeZone object is correct since we are using the US/Eastern regional time zone, but the millis are wrong which makes the time EST instead of EDT. The millis should be 1162098000000.
    The conversion from java.util.Calendar to TIMESTAMPTZ loses the actual offset when setting to a regional time zone. It can maintain this info by specifying it explicitly.
    The conversion from TIMSTAMPTZ to java.util.Calendar also loses the actual offset even if the correct offset is in the database.
    Has anyone else encountered this conversion problem? It appears to be a conversion problem in both directions. I know that the Calendar is lenient by default and will assume Standard Time if time is entered during the repeated 1 o'clock hour at the end of Daylight Saving Time, but the Calendars we are using are explicit in their time, so this would be classified as data corruption.
    Nik

    Opened an SR. Looks like there is a problem with conversion either in TopLink or in JDBC.

  • GL Data Conversion from Legacy to Oracle EBS

    Hi All,
    I am tasked with converting legacy data into GL. I have worked on other conversions like supplier, AR and AP conversions but never worked on GL conversion, so I don't know how and where to start, what are the pre-requisites for entering data into gl_interface table how to ensure that data is correct in the .csv file etc.
    Kindly request you all to guide me in right direction. Hope to hear from you all very soon!!! :)
    Regards

    Hi,
    When you talk about GL-Data it is the trial balance to be loaded from legacy to oracle applications.
    You can use Web-ADI to upload it ,the check list are
    The balance for each account comibnation in the legacy system to be mapped to GL-Oracle code combination balance
    With respect to open AR, AP Invoices if the invoices are converted using a control account for migration in GL then
    the the balance transfered from AP,AR need not be reversed in GL from the source receivables and payables.
    In case if the same account combination are for migrating the balances from AP,AR to GL ..Then the balances transferred from AP,AR should be reversed ..so that it does not affect the TB-GL-Balance
    Finally ensure that the balance for the TB-tallys with your legacy system and upload it using Web-ADI ,import and
    review and post it.
    Hope this points helps your GL-Data conversion.
    Regards,
    Ramaa

  • Data Conversion in SAP E-Recruiting.

    Hi SAP E-Recruiters,
    Our client is using a Legacy application for recruiting candidates where it has 5000 candidates data base (internal and external candidates' applications).
    Now they are asking for movement of this whole data to the newly implementing SAP E-recruiting system. We are exploring various data conversion methods and found that BDC, LSMW will not work for E-recruiting data upload as we have every thing in BSP/Web Dynpro pages here.
    We need your help in finding a solution for moving this data from legacy to SAP E-Recruiting. Pls do the needful.
    Thanks,
    Sudheer

    Sudheer,
    So far there are no SAP delivered migration methods or BAPI's available conversion of candidates/applications, as you have already specified BDC's wouldnot work. Internal candidates always can be brought over in eRecruitment with integration techniques between PA and eRecruit depending upon the landscape.
    However, for external candidates custom conversion programs have to be written by making use of methods/function modules that are used in BSP/WebDynpro.
    regards
    Sridhar

  • Data conversion for New GL - Going live during fiscal year

    Hi Experts,
    My client is going live on Oct 1 (fiscal year in Jan -Dec). So for data conversion we need to load P&L and BalanceSheet as of Sept 30, 2009.  I need your help to resolve few open issues. (we have activated Doc Splitting with Zero Balance and Segment )
    1) When we load P&L balances - how can I assign Profit Centers to all these P&L iitems .  
    2) What is the procedure (best practices) to assign Profit Centers to all the Balance Sheet items since we are going live during Fiscal Year
    3) When we process open items  (AR, AP) in month of October how the Profit Center assignment will work?
    (My client will be using Profit /Cost center terminology first time, so we need to map all the B&S and PnL items to Profit centers)
    4) What is the best way to map all the Balance Sheet GL accounts to Profit Centers.
    Sorry for asking too many questions. Any help would me very much appreciated.
    Thanks,
    Sam

    1) When we load P&L balances - how can I assign Profit Centers to all these P&L iitems .
    Each P&L account balance has to be upload with cost centre, then automatically PC will be triggered.
    2) What is the procedure (best practices) to assign Profit Centers to all the Balance Sheet items since we are going live during Fiscal Year
    Let client decide the profit centre for balance sheet items, since they knows which balance relates to which pc
    3) When we process open items (AR, AP) in month of October how the Profit Center assignment will work?
    While uploading AR,AP balances, each line item to be uploaded with profit centre, automatically in october PC assignment will work.
    (My client will be using Profit /Cost center terminology first time, so we need to map all the B&S and PnL items to Profit centers)
    4) What is the best way to map all the Balance Sheet GL accounts to Profit Centers
    Try to identify each balance sheet gl balance to each profit centre, this will help in future reports. otherwise put in common pc

  • Data Conversion Logic

    I am now trying to implement a data conversion logic(such as UOM field), I want to load all the data conversion rules in the cache from file and keep them in the cache. So when mapping starts it just picks up the data conversion rules from the cache.
    Using GlobalContainer it is ok for one mapping even several UOM fields(just load from file once). But it will load from file again when the mapping is triggered again.
    Is there any logic I can use to implememt the data conversion just load from file once?
    Regards,
    Nick

    Hi Arnold,
    normally every input should be made in LC, why you enter the data in group currency?
    However if you have to start from group currency the only solution to avoid this error is to put at the end of sprunconversion the ON_ERROR_CONTINUE clause.
    But remember he will stop to notify every type of error so you have to be very careful and have this clear in mind when some issue occurs with the currency conversion.
    Regards
         Roberto

  • Issue with Date Conversion when loading XML File into Oracle 10g Database

    Hello all,
    I have the interface shown in the screenshot below. In it, amongst other actions, I'm mapping an XML file element representing a date to an Oracle table column defined as DATE. The source and target columns are highlighted in the screenshot.
    !http://img223.imageshack.us/img223/1565/odiscr275.jpg!
    When I execute the interface, I get the following error message:
    java.lang.IllegalArgumentException at java.sql.Date.valueOf(Date.java:103)
    I'm assuming this refers to the date conversion!
    I've already tried replacing SRC_TRADES.DEAL_DATE with TO_DATE( SRC_TRADES.DEAL_DATE, 'DD/MM/YYYY' ) in the Implementation tab. This function was not recognised when I executed the interface, so it didn't work! The date value in the XML file is in DD/MM/YYYY format.
    I'm guessing that Oracle SQL Date functions don't work in the Implementation tab. Please could somebody let me know:
    1. Which Date Conversion function I could use instead?
    2. Where I can find a reference for the methods/functions I can use in the Implementation tab (if such a reference exists)?
    Cheers.
    James

    Hi.
    Try to change the execution area to staging area. After You change it, write in the mapping box just SRC_TRADERS.DEAL_DATE. When You use TO_DATE, the source field typu should be varchar2, not date (as it is in your source datastore)

  • GL Legacy Data Conversion

    I have a question for the data conversion strategy.
    We are planning to store 2 yr detailed transactions and 4yr balances in Oracle system. In terms of the data conversion process, we were also going to take the same method. It would, however, cause schedule and workload constraints. Actually, we have not yet discussed this from realistic view points. In other words, we do not insist on storing all of data which are 2yr transaction and 4yr balances.
    Although I believe that we will be able to eliminate a volume of the data to be converted to new chart of accounts from old ones, I am a little bit concerned about a couple of things as follows:
    1. the workload for reporting processes
    Assuming that we will not use Oracle standard reports so much, it would not be a big issue, even if we do not convert all of data I mentioned above. We will store historical report data somewhere, and be able to generate certain reports using both Oracle data and historical data in the repository. If not, it could cause us extra efforts.
    2. Audit trails / Examination trails
    If we give up converting the data to fit new system, and that fiscal year has not been examined yet, how should we handle non-converted data for the examination? I am just wondering whether or not, we have only to prepare the conversion table, that ties Oracle balances back to ABC detailed transactions for their reference.
    If anyone is are aware of anything you can advise me, could you please provide some information or guidance? Thanks to ALL.

    Hi,
    When you talk about GL-Data it is the trial balance to be loaded from legacy to oracle applications.
    You can use Web-ADI to upload it ,the check list are
    The balance for each account comibnation in the legacy system to be mapped to GL-Oracle code combination balance
    With respect to open AR, AP Invoices if the invoices are converted using a control account for migration in GL then
    the the balance transfered from AP,AR need not be reversed in GL from the source receivables and payables.
    In case if the same account combination are for migrating the balances from AP,AR to GL ..Then the balances transferred from AP,AR should be reversed ..so that it does not affect the TB-GL-Balance
    Finally ensure that the balance for the TB-tallys with your legacy system and upload it using Web-ADI ,import and
    review and post it.
    Hope this points helps your GL-Data conversion.
    Regards,
    Ramaa

  • XI for managing data conversions

    My customer is considering using XI for managing their data conversions.  This is a very large implementation with many source data systems.  We will also have a very large volume of data to migrate.  For example, we'll have around 8 million materials.  For these materials, we're thinking we can extract this data and map it to an idoc structure and pass into SAP via XI.  First, has anyone used XI for data conversions?  Second, have you dealt with this volume of transactions?  Can XI handle this?  Any thoughts or comments would be greatly appreciated.
    Regards,
    Chris

    Hi Chris,
    SAP XI is used for those exact purposes.  Of course as volume increases, your customer will have to increase memory and hardware.  They have also consider the whether the messages are synchronous.  In case of migration, I think they will be asynchronous.  SAP has a customer who before going live, performed some massive load tests that will simulate an average "real" working day, of 1,200,000 (million, two hundreds thousands) transactions, and it worked smoothly yet fast.  Your customer will most likely have to break up 8 million material into daily batches for now.
    Enjoy!
    John Ta

  • ASAP Data Migration / Data Conversion / Data Transfer Templates

    Hi,
    I'm looking for the ASAP Data Migration / Data Conversion / Data Transfer Templates that used to be available in ASAP.
    Some time I ago, and I believe it was on this site and can no longer find it, I found a zip file which contained the above .xls template to manage all the data conversions on a project. Also included was the data mapping template .xls
    Many thanks.

    Hi,
    I'm looking for the ASAP Data Migration / Data Conversion / Data Transfer Templates that used to be available in ASAP.
    Some time I ago, and I believe it was on this site and can no longer find it, I found a zip file which contained the above .xls template to manage all the data conversions on a project. Also included was the data mapping template .xls
    Many thanks.

  • BPC Data Conversion File

    Dear Experts,
    I am trying to upload a Data file. Below is my Transformation file, Here i am also using a Conversional file. Data Conversion is working fine. After running "Apptend to Fact Tabel" package.
    Converted data is not getting inserted.
    *OPTIONS
    FORMAT = DELIMITED
    HEADER = YES
    DELIMITER =
    SKIP = 0
    SKIPIF =
    VALIDATERECORDS=YES
    CREDITPOSITIVE=YES
    MAXREJECTCOUNT=10
    ROUNDAMOUNT=
    *MAPPING
    Account=*Col(1)
    Categories=*Col(2)
    RPTCurrency=*Col(3)
    Employees=*Col(4)
    Levels=*Col(5)
    Time=*Col(6)
    Project=*Col(7)
    *CONVERSION
    Project=[company]Test.xls!Conversion
    Please need help.
    Thanks,
    Satish

    Satish,
    Why are you using the "append to fact table" package? I have to say I have never used that package before, but have you tried running the "Import" package to see if that runs fine with your datafile?
    Regards,
    Joost

  • VA41 data Conversion

    Hi All,
    Just quick question What’s best option for VA41 data Conversion (single/multiple line items). Custom BDC / Custom Program with BAPI / Standard Idoc / LSMW /...
    Thanks in advance.
    Regards,
    Tim

    hi Tim,
    you will need to either write a program to upload the data to an internal table, or use LSMW to handle all that. Either way, you should use BAPI_CONTRACT_CREATEFROMDATA.
    Regards,
    Naveen

  • Data conversion rule manager & repository

    Hi All,
    We are running a fairly large legacy data conversion project from IMS and DB2 to Oracle which will involve quite a variety range of conversion/mapping rules (complex data manipulation to simple mapping). These requirement and rules involve both tech and business collaboration. Are there any data conversion rule managers tool out there that provide these type of capabilities:
    1. Data conversion rule repository and version control
    2. collaboration features (e.g. collaboration editing, commenting, decision capture, approval...etc)
    We have check someed of the tools out there, and they are either generic requirement gathering tools, which is basically a few tabs with drop downs and a big text box, or some end-to-end data conversion automation solutions, which focus the execution of the data conversion instead of managing & tracking conversion rules.
    Can anyone you help? please let me know if there is anything comes close that is practical and useful. Thank you.

    Tubby, thank you for the info. Yes, we actually use ODI as a DEV tool for data conversion rule implementation and execution between our staging DB and target DB. However, ODI does not really have the capability to allow our tech and business to collaborate together and focus on "getting all these rules defined correctly". Having say that, ODI does have many very useful side features for our DEV to leverage and share information such as knowledge module and data mapping...etc
    We would hope to have a practical data rule management tool & repository that allows tech and business to work together and gather all the rules and build up all the knowledge base in 1 single place. DEV/QA can then take away and focus on the implementation and validation.
    Please let me know. Thank you very much.

  • XI for new data conversion?

    Hello experts,
    You recommendations please on data conversion with XI versus using the LSMW tool?
    We are working on a new implementation, planning to use XI for the interfaces anyway. Now the idea is to use XI for all dataconversion from the legacy systems.
    Good idea, not suitable, better to use LSMW?
    Your thoughts please!
    Many thanks,
    Albert

    Hi again.
    The response to this question cannot be a general answer, because it depends on several aspects. I'll try to order by importance.
    1 - What kind of files do you have for material and BOM creation? XML files?
    XI is a powerfull tool for transfer that kind of files. With an ABAP proxy, you can easily decompose a XML file into a strucuture in ABAP, so it's easy to import files. To send, is also easy. You fill a structure and XI will convert it to a XML file and sent to the respective receiver, using it's communication channel. So, if it's XML, I would say YES, use XI. If not, XI only can copy those flat files and that way it's not easy to get it's data in a proxy. You can copy those file to a directory and import files with a program, but, if so, why using XI for your purpose? Or you can do some mappings to convert flat file into XML file ... the mapping can be easy or very difficult.
    2 - Experience in XI or LSMW? And Urgence?
    This question is important. If you have experience in one of those tools, you should use it. As far as I can catch from your post, XI is a new tool for you, right? To master XI is a process that will take you some time ... not too much ... but consider this part too in your decision. However, since you will start using XI for interfaces, this could be a good start!
    3 - Many other reasons can be found in the link that Murthy provided, but this two should be the most important (in my opinion).
    Hope this helps.
    Regards,
    Valter Oliveira.

Maybe you are looking for

  • ITunes 8.2.1 update bug (-50) and changes to iTunes Photo tab

    Many thanks in advance for your attention. 1. Since I updated to iTunes 8.2.1, have had the following bug whenever I try to sync my iPod Touch (2G). "The iPod " " cannot be synced - An unknown error occurred (-50)." pops up. 2. I have a file in iPhot

  • How to get file name from file adapter (Read Operation)

    Hi All, I am reading files from local location, in jca file i gave filename property =*.* ( which reads all files), and now I want to read the incoming file name inside bpel and I want to copy the file name to some variable. Is this possible?

  • Autocreate-tables="false" orion-application.xml

    using oc4j standalone(9.0.3). CMP2.0. jeveloper(9.0.3). The jdeveloper Domain UML entity diagram and generated entity beans are deployed with default tables generated. The EJB2.0 Dev Guide(9.0.3), says to copy the generated orion-ejb-jar.xml from the

  • Measuring with both channels of virtual bench simultaneously in labview, error 375903

    Hi, I am trying to perform two measurements simultanously using both channels of the Virtual Bench's analog input. I have selected MSO channel 1 and 2 for the measurements yet I am getting error 375903 returned evewry time saying the resource request

  • Cairo-compmgr + Openbox + xf86-video-intel + mouse clicks issue

    When I start cairo-compmgr with Software Rendering, I face issue of clikable items not being highlighted when I hover mouse over them. Also, I have the issue of first item being selected automatically when I right click on certain apps. Is there anyt