BW data Upload issue

Hello Gurus,
We have upgraded our APO system to SCM5.0.
When I am trying to upload upload a flatfile into BI7.0 I am getting an error.
Here are the error details:
Error 1 when loading external data
Message no. RSAR234
Diagnosis
Error number 1 occurred when loading external data:
1. Error when reading the file (access rights, file name, ...)
2. File size or number of records does not correspond to the data in the control file
3. Error when generating the IDoc
4. File contains invalid data (errors with an arithmetic operation or data conversion)
Procedure
Check whether you have the required access rights and whether the data in the control file is correct (file names, record length, number of records, ...). Correct the data in the control file if necessary and check the data file for invalid data (values of the wrong type, values in the wrong format for conversion exit,...). Check whether the file has headers that have not been specified.
I checked all the things and everything looks good. I even tried to upload it into PSA but still the problem persists.
Could anyone please help me out with this
Thanks in advance
Sai
Message was edited by:
        sai balakavi
Message was edited by:
        sai balakavi

I have a question for you.
When you sent me your error i can see that your source system is PW5CLNT300 so, i think that you have a connection with an R3 system.
Why are you trying to load a file with an R3 datasource?
I think that you are sending me an incorrect error, because your load monitor must show PC FILE source system instead of PW5CLNT300.
Regards.
Marín.
Message was edited by:
        Ulises Marí

Similar Messages

  • Cost center data uploading issue in OKENN transaction

    Hi Gurus,
    I have uploaded cost center data using LSMW IDOC method. the data uploaded in the CSKS table successfully. after that i have checked for the cost centers in OKENN transaction at standard hierarchy level. but there cost center are not exist under hierarchy level. then i have changed the hierarchy data(KHINR)(for example From SUS2021100 to SUS2021200, ihave changed in KS02.... then the cost displyed under SUS2021200.... again i have changed the hierarchy from SUS2021200 to SUS2021100.... then the cost center displyed under SUS2021100..... But actualy when i have loaded the data using LSMW all the loaded.... but the cost center not displyed under hierarchy.... after manualy chnged the hierarchy data at KS02 transaction.... it is displying.... idon't understand what is the problem.... can anybody solve my issue.... thanks in advance

    Hi Gurus,
    I am facing the same issue while uploading cost centers through COSMAS01 basic type.
    Cost centers are created properly, we can display them with KS02. Every data are loaded correctly, even the standard hierarchy. Into table CSKS, field KHINR for standard hierarchy is populated, nevertheless, we do not find the cost center into the OKEON Tcode.
    How could I solve the issue I am facing ?
    Thx for your help.
    Regards,
    Aurelien

  • FIM data upload issues

    Hi
    When i am using FIM 10.0 to upload data from flat file i get an error as follows:
    Failed to generate ATL file: The job launch for job 'test_FF' failed with error 'The repository bods for the batch job test_FF cannot be found.'
    Experts please advise
    regards
    Divneet

    Hej Marc,
    thank you for your trouble.
    The problem was solved and the job was executed successfully.
    The error was a wrong description of the DataServices URL. Instead of the name of the server “http://localhost:8080/” was used.
    Although the system message was “Data Services Information. The information you provided is correct”, this was the cause of the failed job execution.
    Sometimes one has to review it again and again to find such a snag.
    Thank you!
    Manuel

  • SPL MK DATA Upload issue

    Hi Expert,
    We are having a batch job running everyday for uploading the latest SPL data from MK Data.
    From 2009/01/23, we found there is some error of the SPL upload logs.
    It display in detail like below:
    No data records were transferred;
    No List type is assigned to the SPL entity
    Checking consistency of SPL entities
    SPL record 25335 is consisiten
    There is 2 questoin here,
    1. I found there is a lot of List type configured for the legal Reg. SPLUS, and how can I knowwhic list type is no included in the configuration?
    2. Why all error log have the message 'SPL record 25335 is consisiten', if there is something wrong for this record?
    Please advice,

    Please also check the following thread  http://scn.sap.com/thread/3298134
    There was a change in XML update file format at header level that is being sent by our data provider, that is causing the message.   I'm currently in process in implementing note 1817221 that was suggested in the thread.
    Regards,
    Ann

  • Asset Data Upload Issue in as91

    Dear Sapgurus,
    Company code Z100, Asset Takeover date is 31.01.2010, While Creating asset master in as91 Asset class 2100, Com Code, Z100, Short text : Sofas, Long Text Sofas Capitalisation Date 20.12.2008 Business Area z100, Cost Center Z1000000, In depreciation area data tab Dep Area tab : 01, Deo Key D001, Use ful life 20, Period 2 Dep Start Date 01.02.2010, Expired Life 44 Years While saving the asset document system popup error message
    E698 SYST: Abnormal termination (ANLB-LGJAN 1 not equal to ANLC-GJAHR
    But in T Code oaaq systems it shows 2009 10  Perid.
    Regards
    My

    You have a takeover date of 31st Jan
    A Depn Start date of 1st Feb
    You therefore cant have any depn for this asset.
    Can you also confirm what date your posting period is for period 10 - FY 2009 - is that November 2009?

  • Issues with 4.1 Data Upload

    I've got some issues with the new feature 'data upload' in 4.1. I had already made the pages and the data loading and it worked perfect. Now a new column was added to my data loading table and I have to add a new table lookup for this column. Everything is set up like it has to be. I've got 4 table lookups. When I try to upload some data and indicate the table columns to the data, I've always get the error: Failed to retrieve the lookup value. There is no problem when I do a data load where he only have to retrieve one column from the lookup table, when he has to retrieve data from more tables for more columns, I always get the FAILED message. Does anyone know the problem of this situation? I already tried to make a totally new data loading, but this also failed to do the job.

    Hi Ellenvanhees,
    I dont think the number of lookups that you defined is an issue here. If possible try to share more details about your data and tables. The few things that come to my mind are probably your data.
    But if you are able to do one by one lookup without problem then I think your upload is failing due to null values. The current status of data upload feature returns a failed lookup even if a null value was to be uploaded. This is a bug #13582661 and has been fixed for the future release.
    Patrick

  • Issues in mass data Upload

    Hi All,
            Hope you all are doing fine.
            I have to do master data upload for my next project. I have gone through LSMW, DX-WB, Recording etc. and now I am quite comfortable with all of these.
            As i have tested these tools for maximum 8-10 records, I am interested in hearing from you all, your experiences regarding actual data upload where volume is high and data may be difficult to verify manualy.  Particularly I am interested in how to make upload faster,Error free,consistent(No record being posted twice etc.)
            Your inputs would be higlhly appreciated.
            Thanks a lot for patient reading.
    Bye and Regards.

    Hi navdeep,
    the mass data upload depending on tht data u want to upload, there ara several function module to upload data like create reservation or upload long text inspection method tell me what the data u want to upload
    Best Regard
    Waleed Sadat

  • BW upgrade EHP1, data uploads should stop?

    Dear experts,
    we have plan for system upgrade. current system is BW 7.0 sp 17, Now planned for EHP1 and SP9.
    I know the there some post upgrade activities, which include consistancy check for objects (infoobject, transfer rules, cubes,dso, etc...)
    Could some one pls confirm, do we need to stop the data uploads/stop process chains during system upgrade?
    Thanks in advance!
    Best Regards,
    Mannu

    Hi Ingo,
    RSRT was giving proper results. We have now implemented few SAP notes and the issues got resolved.
    The following are the notes:
    1499233 - MDX:bXML flattening, unbalanced hierarchy, empty columns
    1485648 - MDX: bXML flattening and hierarchies and displaced columns
    1446245 - MDX: Error when RSR_MDX_BXML_GET_GZIP_DATA is called
    1441767 - MDX: No data for bXML if only ALL member is requested
    1438091 - MDX: basXML: Object MEASURE 0 not found
    1435844 - MDX:Wrong no. decimal places for basXML flattening interface
    1432162 - MDX: Flattening problems when using hierarchies
    1420169 - MDX: bXML flattening: Subsequent note
    1411491 - MDX: bXML flattening in transac. MDXTEST: Selecting packages
    1404328 - MDX: bXML flattening and PROPERTIES: Columns overwritten
    Thanks for your inputs.
    Regards,
    shesha.

  • Photo Gallery display by date uploaded

    I would really like to have the photo gallery display images by date uploaded not alphabetical. I know this isn't possible but I remember somewhere on the old forum someone posting a workaround for this issue. Does any one know a workaround? I'm sure it had something to do with renaming the files.
    Thanks
    Karl

    Hi Sidney, thanks loads for your reply. If i try your method the most recent image still wont be placed at the beginning eg.
    If today's date is 01022012 then tomorrows date will be 02022012 which will still be placed after 01022012
    Eg
    01022012
    02022012
    Thanks again but is there any other method?
    Karl

  • Optimization for bulk data upload

    Hi everyone!
    I've got the following issue:
    I have to do a bulk data upload using JMS deploy in a glassfish 2.1, to process and validate data in a Oracle 10g DB System before it is insert.
    I have my web interface that loads a file and then delegates the process to a Stateless Session Bean which read a N number of lines and after that send a message to a JMS Queue. The JMS has to parse the each line, validate them with the data already in the DB, and finally persist the new data.
    This process is high process consuming, and I need to improve the performance time. I tried to change glassfish default JMS and JDBC pool size, but I have no a huge difference.
    Do you have any advice that could help me?
    Thanks in advance!

    Hi! thank you for you answer!
    High process consuming is in the MDB
    I'm grouping each N number of read lines in the EJB and then send the message to the JMS. The MDB process the persists each line as info in different related tables.
    Thanks again!

  • Data Upload Using BDC

    Hi Friends,
    While creating the task list user defined field is in input mode, but when i do data uploading thru BDC program it's always coming in display mode, so system is not taking the data input for that field.
    Field Name : User Defined Field ( PLPOD-USR00 )
    Kindly suggest the solution.
    Regards
    Pankaj

    Hi ~
    1. Are you  trying bothn online and BDC with the same user id? Please check that if there is any authorization issue.
    2. Some times in BDC the screen flow is different. Check whether there is any difference in the sequence of the screens in between the BDC and online for creating task list.

  • Data Uploads done with one User Id not visible to other users in SPM

    Hi,
    Data uploads were successfully carried out by one of the SPM users. However, other users (with different user id) are not able to see anything in the Data Upload Summary screen.
    Is there a restriction on the visibility of Data Upload Summary for data uploads carried out with one User id to other users in SPM? A similar behaviour is observed for other screens within the Data upload workbench.
    Incase this is not the expected behaviour, it would be great if you could please provide pointers to possible reasons for this.
    Just for your information, all users have been granted same privileges in the SPM application.
    Thanks in advance.
    Regards,
    Ashish Sharma

    Hi Ashish,
    No this is not the expected behavior. We have seen this issue for other customers in the past but the reason has always turned out to be role related.
    Can you ensure that the required SPM roles are assigned to the user who do not see the DM data both in ABAP as well as portal.
    Thanks,
    Divyesh

  • Data upload from R3 to BPC using filters

    We are facing a performance issue on the uploading data flow from R3 to BPC. Two steps have been set up.
    Step1: R/3 --> ODS --> BW CUBE (No filters and full mode)
    Step2: BW  CUBE --> BPC CUBE (Standard BW upload package)
    The first one (from R3 to a BW Cube) has no filters, so it loads every data and it takes too much time. However, the second one (from the Cube to BPC)  has filtering options, so we can only load the data we need to (we usually use entity and time filters).
    Both are executed from the BPC DataManager.
    Any advice to improve the performance of the first step? Could we join the step 1 and step 2 and get the filters from the standard BW upload process using a custom process chain?
    Does anybody have any experience using filters to reduce the amount of records on the data uploading from R/3 to BPC?
    BPC Version 7.50.15

    Hi,
    for example in BI data is stored as below
    GroupAcc - Date - Balance
    100000 - 12.12.10 - 400
    **I'm not quite sure how you are getting the above data from infocube.Are you directly loading data from PSA to infocube with out loading data into DSO -0FIGL_O10(General ledger(New):Transaction figures) ?
    To get required format of following data, no need to do any thing while laoding the data into BPC. Your BI infocube should maitain the data in below format.
    GroupAcc - Date - Balance
    100000 - 12.12.10 - 100 - (LC)
    100000 - 12.12.10 - 100 - (TC)
    100000 - 12.12.10 - 200 - (GC)
    To acheive above format of data in the infocube please follow standard way of data flow given by SAP in BI Contect(http://help.sap.com/saphelp_nw70ehp1/helpdata/en/e6/f16940c3c7bf49e10000000a1550b0/frameset.htm).
    Flow should be
    R/3 --> PSA --> DSO -->Infocube.
    In this flow its important to remember about infoobect 0CURKEY_TC in DSO.
    This is always the currency key of the transaction currency; it is also filled for records with a different currency type. Without this key field, postings with different transaction currencies would be overwritten after summarization and thereby be lost.
    (http://help.sap.com/saphelp_nw70ehp1/helpdata/en/a8/e26840b151181ce10000000a1550b0/content.htm)
    Once you follow the standard flow, your infocube contains the requried format (LC,TC and GC )of data to load it into BPC application.
    hope it helps...
    regards,
    Raju

  • BOM Data upload

    Hi,
      We plan to create new fields in the BOM header through INCLUDE structure CI_STKO and do a data upload into these fields through LSMW. However, the issue is that CI_STKO structure is not included in the BI structure BICSK used in LSMW, hence the fields added to the BOM header through CI_STKO cannot be loaded through LSMW.
    This works fine at the item level - since CI_STPO is included in BICSP.
    Can anyone let me know if they have faced similar problem and are aware of a solution.
    Thanks,
    Vikas

    Hi,
    Insted of LSMW goo for BDC.. if not goo for BAPI.. you have an enhancemnet concept by using this concept you can enhance the standard BAPI.. and you can process for upload ...
    BAPI_MATERIAL_BOM_GROUP_CREATE

  • Regarding PA data upload

    Hi Experts,
    I just wanted to have a discussion regarding PA data upload.
    What we feel is that, it should be ok if we go with BDC for each of the infotrype.
    hr_infotype_operation can be used but  again it  also use bdcdata only so it should not  be a big performmance issue.
    Both approach can be used as per the data upload and its' going to be one time only.
    Can anyone of you suggest something regarding the same.
    I have searched the forum and i got responses like bdc can be used ( some might sue FM also).
    Can i have some comments from you guys regardimng the same ?

    Hi Rinki;
    Try LSMW recording method for data upload in PA Infotype.
    Regards

Maybe you are looking for

  • Report Server Configuration Issue

    Hi all  , We are trying to install Report Server 11gR2 ( 11.1.2.1.0) , Weblogic Server is 10.3.6.0 on Sun Solaris on Sparc (64-bit) version 11. Java used is JDK1.7_21 . We first installed Weblogic and did a software only install of Forms and Reports

  • Material Account Assignment Group, in Sales Order Mass Update

    Sd Experts, We just went live, our consultant updated Account assignment group at the material level but did not updated existing materials. When we caught the problem and updated exisisting materials there were some existing orders already created.

  • Losing counts using asynchrono​us reads

    Hi guys, I have been working on an application to allow me to do imaging on a microscope. Using an external clock source I need to "bin" photons coming from my sample for a set period of time. To do these measurements I devised an APD object in C#. T

  • Nokia 5300

    Hi! My nokia 5300's manu button and navigation buttons are not working. How to solve this problem? Thanks,

  • MSI Vox: major humming noise in background

    Hey folks, I'm new to the forum and have a major problem. While the sound is playing there is a major humming distortion in the background. I've connnected the Vox on 2 individual systems and it occurs on both. I'm curious, I wonder if the "optional