How to load infotype 2010 data in to R/3?

Hi,
I posted the same question earlier but with a wrong subject line.
My question is how can I load infotype 2010 data in to ECC 6.0? I am using PI to convert a flat file, but depending on the feature available on R/3 side(either IDOC, BAPI, or a flat file) I am going to convert to the required format.
Can anyone tell me what is the best way to push this data in to R/3.
Appreciate your answers.
Thanks
Raju

Sorry My Bad.
Use this BAPI
BAPI_PTMGREXTREMSPEC_INSERT and below id the description of the BAPI.
This method stores the data records in the interface tables PTEX2010 and PTEXDIR. Then, the report RPTEXTPT (External Transfer -> Time Management reads the data from PTEX2010 and stores it in the Employee Remuneration Info infotype 2010. Neither cost assignment nor activity allocation is supported with this method.
You can try these as well.
BAPI_PTMGREXTREMSPEC_INSWACT Insert EE remuneration info in table (with activity allocation)
BAPI_PTMGREXTREMSPEC_INSWCOST Insert external EE remuneration info in table (w/ cost assignment)

Similar Messages

  • How to load the exist Data Mart Request in to cube.

    Hi Friends,
    I have a scenario... i got DSO and one Cube , iam loading the data from DSO to Cube .
    The number of records are more in the DSO and due to some clculation in the routine i need to load the data year by year(i got 2007 and 2008 year daya). I have loaded the 2007 data to the Infocube and in the DSO i can see the Data Mart Symbal against the Request Id when th load is finished successfully.
    When iam trying to load the 2008 data from DSO to Cube iam getting the "0" records. I realised that the Data Mart symbal is exist in the DSO.
    How to load the 2008 data in this scenario. If i delete the data mart sysbel means iam deleting the Cube request.
    Can any have an idea on this.
    Thaks in advance.

    HI,
    Things are not clear.
    How is loading happening if its delta or full load through DTP or you using 3.5 flow??
    In any cases if you do a full load or full repair based on the year selection it should pick the records from the source...there is nothing to do with the data mart status in the case of full loads.
    Data mart status will come into picture only when you schedule the delta loads.
    Do a full load based on selections on year from the DSO to the cube...no need to delete the data mart or it will bring that request again delta is scheduled...does that request in DSO contains the data for 2008 only ...if yes then you can just delete the data mart status for that and do a delta...if not then do full loads as said.
    Thanks
    Ajeet

  • How to load Matrix report data into basic table data using ODI

    Hi,
    How to load Matrix report data into basic table data using oracle Data Integrator ?
    Requirement Description:
    Following is the matrix report data:
    JOB                       DEPT10                DEPT20 
    ANALYST                                           6000
    CLERK                   1300                     1900 Need to convert it into below format:
    JOB                             Dept                        Salary
    ANALYST                  DEPT10     
    ANALYST                  DEPT20                     6000
    CLERK                       DEPT10                    1300
    CLERK                       DEPT20                    1900
        Thanks for your help in advance. Let me know if any further explanation is required.

    Your list seems to be a little restrictive, you can do a lot more with ODI procedures.
    If you create new procedure, and add a step. In the 'command on source' tab set you technology and schema as per your source database. Use the unpivot functionality as described in the link, please, rather than using 'SELECT *' use the appropriate column names and alias them for eg:
    SELECT job as job,
    deptsal as deptsal,
    saldesc as saledesc
    FROM pivoted_data
    UNPIVOT (
    deptsal --<-- unpivot_clause
    FOR saldesc --<-- unpivot_for_clause
    IN (d10_sal, d20_sal, d30_sal, d40_sal) --<-- unpivot_in_clause
    Then in your 'command on target' tab set the technology and schema to your target db, then put your INSERT statement for eg:
    INSERT INTO job_sales
    (job,
    deptsal,
    saledesc
    VALUES
    :job,
    :deptsal,
    :saledesc
    Therefore you are using bind variables from source to load data into target.
    Obviously if the source and target table are in the same database, then you can have it all in one statement in the 'command on target' as
    INSERT INTO job_sales
    (job,
    deptsal,
    saledesc
    SELECT job as job,
    deptsal as deptsal,
    saldesc as saledesc
    FROM pivoted_data
    UNPIVOT (
    deptsal --<-- unpivot_clause
    FOR saldesc --<-- unpivot_for_clause
    IN (d10_sal, d20_sal, d30_sal, d40_sal) --<-- unpivot_in_clause
    also set the log counter as 'Insert' on the tab where your INSERT statement is, so you know how many rows you insert into the table.
    Hope this helps.
    BUT remember that this feature only came out in Oracle 11g.

  • How to load text file data to Oracle Database table?

    By using Oracle Forms, how to load text file data to Oracle Database table?

    Metalink note 33247.1 explains how to use text_io as suggested by Robin to read the file into a Multi-Row block. However, that article was written for forms 4.5 and uses CREATE_RECORD in a loop. There was another article, 91513.1 describing the more elegant method of 'querying' the file into the block by transactional triggers. Unfortunately this more recent article has disappeared without trace and Oracle deny its existence. I know it existed as I have a printed copy in front of me, and very useful it is too.

  • HOW TO LOAD R/3 DATA INTO SAP BI USING PROCESS CHAINS?

    Hi,
    Can we load R/3 data into BI using process chains?... I loaded data from R/3 into Infocube using generic extraction using view... took 2 tables EBKN and EBAN and ceated view.
    In PSA I can find all the 2388 records but when I load into datatarget in transferred tab there are 2388 records but in added colum i could find only 2096.....
    I deleted the request and want to load through process chains....... but  how to do ?????? without flat file ...can we laod using process chains?
    I appreciate any inputs.......
    Regards,
    Prasanthi.

    did you even bother looking at the links in my previous posts???
    read the docs...try yourself...if you encounter specific issues, you can post them on the forum...
    if you're really expecting somebody to post a step by step for process chain, i think you can wait a long, long time...

  • How to Load the Master Data for custom defined Z infoobjects?

    Hi Gurus,
    This is the requirement.
    Some screen enhancements has to be made by adding  new Custom fields( like two text boxes,eg Name and city ) in such a way if we enter data into that text box,that data value for both of the Enhancements has to be stored in some table.If this data is a master data,then how would I pull this master data into BW.
    Please someone throw some light on this problem.
    Lets say a Screen is enhanced with NAME and CITY with text text boxes.Let the user enters the value for NAME and CITY.Let them be stored under two different tables ZTAB1 and ZTAB2.How would I load this master data into custom Defined  Z infoobjects ZNAME AND ZCITY in BW.
    Please help me out.
    Thanks is Advance,
    Regards

    you can by creating generic datasources on thoase tables with Full load or delta.
    if the number entries is less.. you can try full load...
    if it is more... along with required fields...put date also as one fields. which needs to get date of execution of program or date of uploading to Tables. this will be useful build Delta Mechanism.
    all the best.
    Nagesh Ganisetti.

  • How to load xml file data

    hi ,
    I know how to load data from csv file...
    Anyone can guide me for how to load data from xml file
    i need step by step guidence
    from
    naman

    hi naman shah,
    you can used XML DOM in SAP in order to read data from xml file
    first step to do that, you need to read the xml file using OPEN DATASET
    after that you convert the content of the file to XML DOM class in SAP using function modul SDIXML_XML_TO_DOM
    after that you can read / manipulate the data from xml.
    for the XML DOM class documentation in SAP, please refer to http://help.sap.com/saphelp_crm50/helpdata/en/bb/576637dca511d4990b00508b6b8b11/frameset.htm
    hopes it helps
    regards
    eddhie

  • How to load the XML data of SAP to  a target database

    i use an IDocs to read from SAP and converted into XML (ie .XSD ) and now i want to move that XML data to target data base.
    i am trying to move XML data to template table, but i am getting this mapping error , even though i used NRDM and mapping while moving to template table.
    ERROR MESSAGE :
    Invalid mapping expression for column <Query_1.SNDPRN>. Additional information: <Cannot parse expression.
    Table <Query.EDI_DC40> for column <SNDPRN> does not occur in the FROM clause. For a top-level query, all columns must belong to some table in the FROM clause.. (BODI-1112351)>. (BODI-1111081)

    Hi Srikanth,
    Please find my understanding below.Correct me if my understanding is wrong.
    1)You have moved the data from IDOC to XML file (i.e Creating an XSD and using it as target).
    2)Now this XML file, will you your source and you are going to load it into the target table.
    The error looks like BODS is unable to parse the Hierarchy structure to the flat structure.
    How are you  un-nesting  your data (Are you using XML pipeline) .
    Thanks
    Arun.

  • SQL*LOADER how to load blanks when data is null

    I need to load data which contains null values into a table which cannot contains nulls. How can I tell SQL*LOADER to load blanks in a column when the input data is nulls?

    I think something like this in your control file should work:
    col1 char(10) nvl(col1,' ')

  • How to load a huge data file (export from mysql) to an oracle table?

    I have a data file which contains 1.3G bytes table data from mysql. It's format is csv.
    How can i load these data to an oracle table in java way?
    I thought about using sqlldr. But it requires a oracle client which may not be permitted in my environment.
    If i generate "insert " statement and use batch update the batch update will stop if it comes across an exception when executing the insert one by one. I find no "insert ingore" statement in oracle like mysql.
    Is there anyone can give me a solution with high performance? It's urgent and Thanks!

    hsy541 wrote:
    I have a data file which contains 1.3G bytes table data from mysql. It's format is csv.
    How can i load these data to an oracle table in java way?
    Why must it be a Java way? The point is just to move the data.
    I thought about using sqlldr. But it requires a oracle client which may not be permitted in my environment.
    Why not just import the .csv file into Oracle? I'm sure there's a script or utility that will do it.
    If i generate "insert " statement and use batch update the batch update will stop if it comes across an exception when executing the insert one by one. I find no "insert ingore" statement in oracle like mysql.
    You want to turn off transactions? I'm not an Oracle expert, but I'd be shocked if there was something you could do in MySQL that didn't have an analog in Oracle. Find a better resource, like somebody that actually knows Oracle. Who set up that database for you? Consult with them.
    Is there anyone can give me a solution with high performance? It's urgent and Thanks!Urgent to you....
    %

  • How I load all the data in a database table into a cache?

    Hi
    Lets assume I have a lookup table I want to load it upfront into the cache so that the application can use this data from cache. One way is to write code to create a map from the database and call putALL. Is there any better approach to this?
    Another thing is if the data is more than 1 GB and if I construct a map of one GB before putting it into the cache the JVM might crash. The question is how do I distribute upront loading of the cache to multiple JVMs. Do I have to use invocation service or is there any thing out of box.

    I only use one JVMs and I use a distributed cache, the configuration XML file like this:
    <p> <distributed-scheme>
    <scheme-name>default-distributed</scheme-name>
    <service-name>DistributedCachestore</service-name>
    <backing-map-scheme>
    <read-write-backing-map-scheme>
    <scheme-ref>example-read-write</scheme-ref>
    </read-write-backing-map-scheme>
    </backing-map-scheme>
         <listener-scheme> </listener-scheme> 
         <autostart>true</autostart>
    </distributed-scheme>
    <read-write-backing-map-scheme>
    <scheme-name>example-read-write</scheme-name>
    <internal-cache-scheme>
    <local-scheme>
    <scheme-ref>default-eviction</scheme-ref>
    <high-units>50050</high-units>
    <low-units>40000</low-units>
    <expiry-delay>3600</expiry-delay> 
    </local-scheme>
    </internal-cache-scheme>
    <cachestore-scheme>
    <class-scheme>
    <scheme-ref>example-cachestore</scheme-ref>
    </class-scheme>
    </cachestore-scheme>
    <read-only>false</read-only>
    <write-delay-seconds>0</write-delay-seconds>
    </read-write-backing-map-scheme>
    <!-- 
    CacheStore test definition 
    -->
    <class-scheme>
    <scheme-name>example-cachestore</scheme-name>
    <class-name>com.tangosol.examples.coherence.DBCrudStore</class-name>
    <init-params>
    <init-param>
    <param-type>java.lang.String</param-type>
    <param-value>cachec</param-value>
    </init-param>
    </init-params>
    </class-scheme></p>
    The data in my table will reach about 1000000+ and I want to load about 50000 into cache.
    Thanks.

  • How to load thousands of data files?

    My database is installed on windows, and I have thousands of data files that I need to load. Trying to figure out how to do it without manually editing my CTL file every time I load one data file.
    Can you specifiy your control file to load directories instead of just a file?
    Or can some looping script be written that makes the changes?
    Thanks!
    Message was edited by:
    user447781

    Don't list the data file in the CTL file ... use the "data=" option on the command line to pass in the data file.
    A wrapper BAT file should be easy to create/generate.
    D:\>sqlldr
    SQL*Loader: Release 9.2.0.6.0 - Production on Mon Aug 29 13:52:54 2
    Copyright (c) 1982, 2002, Oracle Corporation.  All rights reserved.
    Usage: SQLLDR keyword=value [,keyword=value,...]
    Valid Keywords:
        userid -- ORACLE username/password
       control -- Control file name
           log -- Log file name
           bad -- Bad file name
          data -- Data file name
       discard -- Discard file name
    discardmax -- Number of discards to allow          (Default all)
    ...

  • How to load oracle views data in essbase via sql interface

    <p>Could someone explain to me how to use Essbasesql-interface.</p><p>We have some oracle views for data and a certain number of DLRsfor otl building and data loading.</p><p>I know that there is an INCBUILDDIM syntax for sql source inwich you just specify the dlr and the oracle user-password toconnect.</p><p>How does it find the appropriate view to fill the dlr? How doesit work?</p><p> </p><p>Thanks in advance</p><p>eakp</p>

    Hi,
    How are you loading, if you are loading from a SQL table you would just load a null into the database
    e.g. loading data into sample basic
    Measures,Period,Scenario,Market,Product,Data
    Sales,Jan,Actual,Florida,100-10,10
    Cogs,Jan,Actual,Florida,100-10,+NULL+
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • How to load from a data model located in SharePointOnline/PowerBI?

    PowerPivot has the option to load 'From Analysis Services Or PowerPivot'.
    The Table Import Wizard usually expects the full path of the workbook (including the data model) in SharePoint.
    However, 'http://myname.sharepoint.com/shared documents/myfile.xlsx' doesn't work.
    How can I access  my online data model from PowerPivot ?

    SharePoint needs to have Power Pivot Enabled. Speak to your SharePoint Admin if you have any questions about that. Once enabled, you would be able to create a Power Pivot Gallery - where you can upload your Power Pivot Models. Once uploaded the easiest way
    is to go to the Gallery and click on the "Excel" button next to the model. This would provide you a connection to the file on SharePoint. 
    See screenshot below OR watch this video for a demo: Power Pivot to SharePoint: Create thin reports connected to SharePoint Model
    Related Blog Article: Power Pivot to SSAS Tabular in less than 30 minutes (speaks about separating Model from Reports using SSAS or SharePoint)
    Regards, Avi
    www.powerpivotpro.com
    Wiki:How to ask a Power Pivot Question to get a prompt, accurate and helpful response

  • How to load the existing data from the databse in the CMP entity bean?

    hello
    my problem is as follows--
    i am creating a CMP entity bean.in these beans client create the data instances using create() function.
    now a entity bean will be created for the newly created data and it's EJBObject will also be formed.
    now we will be able to perform finder methods over them.
    remember these data has been newly created in the database.
    but now if i want to perform the finder methods on the
    existing data in the database( i don't need to use create() fn as i am not "creating" the data.). how will i perform the query over the existing data.
    basically i have to make a CMP bean to extract the data corresponding to a particular condition , from the database table. i don't want to create the data in the database but just find out what data satisfies my condition.
    thanking you
    Prashant

    you have to use findByXXX methods. depending on your appserver you may have to declare them (and eventually the query you want to be used) in the vendor specific deploiement descriptor.
    (e.g. using jboss i don't remember having explicitly declare the findBySomeField method - where someField is a persistent field - in the dd, whereas with bas all finder methods have to be explicitly declared)
    [note that obviously these methods have to be present in your home interface]

Maybe you are looking for

  • White Border Around Flash in Dreamweaver

    When I insert my Flash file (.swf) into my Dreamweaver page, Dreamweaver inserts a thin white border around the flash video. I have tried several potential solutions I have seen but none of them seem to solve the problem. Looking for suggestions!

  • Safari 2.0.3 : certain web sites don't recognize correct login/password

    Just upgraded to 10.4.4 and with this to safari 2.0.3. From this day on I can't gain access to some web sites that require a log/pass because it says it is incorrect. ( Note the password I type is the correct one and works with explorer...). One of t

  • Memory leak? amd 64 bit using 1.4.2_03 program runs fine on other computers

    I have a program that runs fine on a red hat linux desktop and a pentium 1.6ghz xp pro laptop but on my amd 64 bit 3400+ xp pro desktop after ~20 seconds it slows to a crawl. i was originally using j2sdk1.4.2_04 but noticed this was not available for

  • Sending iMovie to Windows

    I have tried to send an iMovie to a Windows machine via Yahoo Messenger, person on the windows has received the movie but cannot open it . Help please thanks

  • Itunes driver's registry settings for burning CDs/DVDs are missing,

    Links provided by the Apple support staff thus far 1.Windows: http://docs.info.apple.com/article.html?artnum=93286 2, http://www.info.apple.com/kbnum/n93071 are generic and do not address the problem I've described in numerous posts - I have read and