Error while uploading master data attributes thru flat file

while uploading master data attributes thru flat file i am facing the below errow. can nyone let me know plz.
Error 8 when compiling the upload program: row 227, message: Data type /BIC/CCDEIO_MAT5_ATTR was found in a new

This just means that you changed the transfer rules/transfer structure and have not refreshed the workbench yet.  The infopackage data hasn't been updated ... this is similar to a datasource update, where you need to replicate to make sure BW knows the latest version. 
The best way to correct this error is to reload RSA1.  Every time I make a change to the tr/ts, I always run /nrsa1 to refresh the workbench. 
Brian

Similar Messages

  • Error while loading master data attribute

    Hi Guru's
    I am getting below mention error while loading master  data attribute.
    0MANU_MATNR : Data record 1035 ('000000000000229058 '): Version 'BDM755s ' is not valid
    Please sugggest how to resolve ?
    Regards
    Karan

    Hi,
    I think there is problem with "0MANU_MATNR" Data, check the records 1033 to 1036 and find the differnce between 1035 record and other records, some values is giving problem so by looking at other reocrds i.e 1033,1034,1036. Check in PSA in Monitor. 
    Note: If any problem with Upper/Lowe case then use TOUPPER in Transfer Rules
    If need use Conversion exits in transfer rules.
    ALL FUNCTION 'CONVERSION_EXIT_ALPHA_INPUT'
      EXPORTING
        INPUT         = 0MANU_MATNR
    IMPORTING
       OUTPUT        =   Result .
    or
    ALL FUNCTION 'CONVERSION_EXIT_ALPHA_OUTPUT'
      EXPORTING
        INPUT         = 0MANU_MATNR
    IMPORTING
       OUTPUT        =   Result .
    Thanks
    Reddy
    Edited by: Surendra Reddy on Mar 3, 2009 11:46 AM

  • Error while deleting master data (Message no. RSDMD118)

    Dear all,
    I am performing a test to create and delete new Company Codes in Consolidation Workbench.
    Creation of new Company Codes went on successfully but deleting them give me the following error message upon saving:
    Error while deleting master data
    Message no. RSDMD118
    Since they are new Company Codes, they have not been used and stored in any BW cubes. Any idea on why BCS prevents me from deleting them?
    Appreciate assistance to troubleshoot and resolve this error.
    Thank you.

    Thank you both for your replies.
    I am sure those are not the case here as I deleted the Company immediately after creating it.
    Just to make sure, I performed "Where-Used List" again and BCS confirmed that it is not being used anywhere:
    Use of: Company UAT1 UAT Company 1
    Number of Uses Found: 0 
    However, the same error message was displayed again after I tried to delete the Company and save the configuration. Still can't find any solid reason behind this.

  • Error while uploading the data using FM"upload"

    Hi,
    I encountering an error while uploading the data using text file with FM " UPLOAD"
    The error is "File does not exist or cannot be opened "
    But there is a file with name and extenstion right.
    Regards
    Vishnu

    You have to create RC29P-IDNRK(var) using concatenate statement. Try this.
    DATA: new_mark TYPE bdcdata-fnam.
    CONCATENATE 'RC29P-IDNRK(' var ')' INTO new_mark.
    PERFORM bdc_field USING new_mark W_BOM-QTY

  • Error while loading master data.

    I am facing error while loading master data in a infoobject.
    The error comes when I run DTP fo rloading data from PSA to infoobject.
    The error wordings are : "MCK_PRDT1 : Data record 1 ('Bleach '): Version 'Bleach ' is not valid"
    Please help me in understanding this Version errror and suggest a solution for this.

    SAP metodology integration MDM & BW you can look at:
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/b0f1f135-0015-2b10-5e83-acca118e17aa
    You can turn matching strategy in MDM Data manager for delete dublicate of records
    and then use MDM Syndicator for export data to file and then load that file to BW
    If you use OHD you cannot load from BW to MDM repository 10 records and unload only 5 because
    in BW every record has unique key.
    also you can use MDM ABAP or Java API for delete dublicate of records
    Edited by: Kanstantsin Chernichenka on May 8, 2009 1:02 PM

  • BI 7.0 - Duplicate Record Error while loading master data

    I am working on BI 7.0 and I am trying to load master data to an info object.
    I created an Infopackage and loaded into PSA.
    I created transformation and DTP and I get an error after I execute the DTP about duplicate records.
    I have read all previous threads about duplicate record error while loading master data and most of the them suggested to check 'Ignore duplicate records' option in the infopackage. But in 7.0, I can only load to PSA with infopackage and it doesn't have any option for me to ignore duplicate records.
    My data is getting loaded to PSA fine and I get this error while loading to info object using DTP.
    I would appreciate your help to resolve this issue.
    Regards,
    Ram.

    Hi,
    Refer:
    http://help.sap.com/saphelp_nw2004s/helpdata/en/45/2a8430131e03c3e10000000a1553f6/frameset.htm
    http://help.sap.com/saphelp_nw2004s/helpdata/en/d0/538f3b294a7f2de10000000a11402f/frameset.htm
    With rgds,
    Anil Kumar Sharma .P

  • 36 duplicate record  found. -- error while loading master data

    Hello BW Experts,
    Error while loading master data
    ( green light )Update PSA ( 50000 Records posted ) No errors
    ( green light )Transfer rules ( 50000  „³ 50000 Records ): No errors
    ( green light )Update rules ( 50000 „³ 50000 Records ): No errors
    ( green light ) Update ( 0 new / 50000 changed ): No errors
    ( red light )Processing end: errors occurred
         Processing 2 finished
    ƒÞ     36 duplicate record found. 15718 recordings used in table /BIC/PZMATNUM
    ƒÞ     36 duplicate record found. 15718 recordings used in table /BIC/XZMATNUM
    This error repeats with all the data-packets.
    what could be the reason of the error. how to correct the error.
    Any suggestions appreciated.
    Thanks,
    BWer

    BWer,
    We have exactly the same issue when loading the infoobject 0PM_ORDER, the datasource is 0PM_ORDER_ATTR.
    The workaround we have been using is 'Manual push' from the details tab of the monitor. Weirdly we don't have this issue in our test and dev systems, even in Production, this doesn't occur somedays. We have all the necessary settings enabled in the infopackage.
    Did you figure out a solution for your issue, if so please let me know and likewise I will.
    -Venkat

  • Strange error while loading master data

    Hi,
      Has anybody seen the following error while loading master data?
    6 duplicate record found. 886 recordings used in table /BI0/XCOSTCENTER
    6 duplicate record found. 886 recordings used in table /BI0/PCOSTCENTER     
    and repeat again once (so total 4 lines of error message)
    The update mode is "PSA and then in the infoobject (Package by Package)"
    Surprisingly the step by step error analysis shows all green except in step "Processing error in Warehouse reported?" which is red.  The over all status i.e. Total and  Technical are both red.
    It all means something went wrong but with the step by step error analysis all green even in the parts like "Processing successfully finished" and "All reported data packet received" I am confused.
    To make matter more confusing there are no error records in corresponding PSA data package.
    any comments would be greatly appreciated.

    Hi Novice,
    In your InfoPackage > Processing tab, select the option Only PSA and Update subsequently into Data Target, and also add the check mark for Ignore double data records. Then try the load again...it should be okay.
    Hope this helps...

  • Caller-70 Error while loading master data into infoobject

    hi ,
    I am getting following error while loading master data into infoobject (0tb-account). I am loading this master data in production environment for the first time. there are about 300000 records. All have got loaded upto PSA. Infopackage settings were PSA and then into data target.
    Short dump in the Warehouse
    Diagnosis
    The data update was not finished. A short dump has probably been logged in BI. This provides information about the error.
    System Response
    "Caller 70" is missing.
    ST22 dump analysis is as below:
    Termination occurred in the ABAP program "GP476CZYBEF2WX53UZ8TXFG6XOS" - in                  
    "VALUE_TO_SID_CONVERT_DB".                                                                  
    The main program was "RSMO1_RSM2 ".
    Please help as soon as you can..Production problem....
    Regards
    Rakesh

    Hi rakesh,
    May be IDOCs not processed completely,
    Idoc Problem, Either wait till time out & process Idoc from detail monitor screen, or go to BD87 & process Idoc with status = YELLOW ( be careful while processing IDOCS from BD87, choose only relevant Idocs
    Cheers
    Raj

  • Error while performing Master Data transfer from SAPR/3- ECC5.0 to SCM4.1

    Hi all,
    Am getting an error while transfering Master data from SAP R/3 system ECC5.0 to SCM4.1 through Core Interface (CIF).
    After creating an Integration model and while activating the same am getting the following error,
    " System: SCMCLNT800    User:  SMART03 06/17/2006 13:32:14
    Function: /SAPAPO/CIF_LOC_INBOUND
    Text:Error in address data for location DC2000/ECC800 "
    Could anyone provide me how to fix this?.
    Regards
    Bharanidharan

    Hi Veera,
    Am sorry for the delayed response. Well am using XI as integration engine for transfer of transactional data from R/3 to SCM4.1. I resolved this issue which was because of master data problem. Thanks for your input.
    Regards
    Bharanidharan

  • Uploading the data from a flat file into ztable

    Hi,
    I have a requirement where I have to upload the data from 2 flat files into 2 z tables(ZRB_HDR,ZRB_ITM).From the 1st flat file only data for few fields have to be uploaded into ztable(ZRB_HRD) .Fromthe 2nd flat file data for all the fields have to me uploaded into ztable(ZRB_ITM). How can I do this?
    Regards,
    Hema

    hi,
    declare two internal table with structur of your tables.
    your flat files should be .txt files.
    now make use of GUI_UPLOAD function module to upload your flatfile into internal tables.
    CALL FUNCTION 'GUI_UPLOAD'
        EXPORTING
          filename            = 'c:\file1.txt'
          has_field_separator = 'X'
        TABLES
          data_tab            = itab1
        EXCEPTIONS
          OTHERS              = 1.
    use this function twice for two tables.
    then loop them individually and make use of insert command.

  • Problem in the BDC program to upload the data from a flat file.

    Hi,
    I am required to write a BDC program to upload the data from a flat file. The conditions are as mentioned below:-
    1) Selection Screen will be prompted to user and user needs to provide:- File Path on presentation server (with F4 help for this obligatory parameter) and File Separator e.g. @,#,$,%,... etc(fields in the file will be separated by using this special character) or fields may be separated by tab(tab delimited).
    2) Finally after the data is uploaded, following messages need to be displayed:-
    a) Total Number of records successfully uploaded.
    b) Session Name
    c) Number of Sessions created.
    Problem is when each record is fetched from flat file, the record needs to be split into individual fields separated by delimiter or in case tab separated, then proceeding in usual manner.
    It would be great if you provide me either the logic, pseudocode, or sample code for this BDC program.
    Thanks,

    Here is an example program,  if you require the delimitor to be a TAB, then enter TAB on the selection screen, if you require the delimitor to be a comma, slash, pipe, whatever, then simply enter that value.  This example is simply the uploading of the file, not the BDC, I assume that you know what to do once you have the data into the internal table.
    REPORT zrich_0001.
    TYPES: BEGIN OF ttab,
            rec TYPE string,
           END OF ttab.
    TYPES: BEGIN OF tdat,
           fld1(10) TYPE c,
           fld2(10) TYPE c,
           fld3(10) TYPE c,
           fld4(10) TYPE c,
           END OF tdat.
    DATA: itab TYPE TABLE OF ttab.
    data: xtab like line of itab.
    DATA: idat TYPE TABLE OF tdat.
    data: xdat like line of idat.
    DATA: file_str TYPE string.
    DATA: delimitor TYPE string.
    PARAMETERS: p_file TYPE localfile.
    PARAMETERS: p_del(5) TYPE c.
    AT SELECTION-SCREEN ON VALUE-REQUEST FOR p_file.
      DATA: ifiletab TYPE filetable.
      DATA: xfiletab LIKE LINE OF ifiletab.
      DATA: rc TYPE i.
      CALL METHOD cl_gui_frontend_services=>file_open_dialog
        CHANGING
          file_table = ifiletab
          rc         = rc.
      READ TABLE ifiletab INTO xfiletab INDEX 1.
      IF sy-subrc = 0.
        p_file = xfiletab-filename.
      ENDIF.
    START-OF-SELECTION.
      TRANSLATE p_del TO UPPER CASE.
      CASE p_del.
        WHEN 'TAB'.
          delimitor = cl_abap_char_utilities=>horizontal_tab.
        WHEN others.
          delimitor = p_del.
      ENDCASE.
      file_str = p_file.
      CALL METHOD cl_gui_frontend_services=>gui_upload
        EXPORTING
          filename = file_str
        CHANGING
          data_tab = itab.
      LOOP AT itab into xtab.
        CLEAR xdat.
        SPLIT xtab-rec AT delimitor INTO xdat-fld1
                                         xdat-fld2
                                         xdat-fld3
                                         xdat-fld4.
        APPEND xdat to idat.
      ENDLOOP.
      LOOP AT idat into xdat.
        WRITE:/ xdat-fld1, xdat-fld2, xdat-fld3, xdat-fld4.
      ENDLOOP.
    Regards,
    Rich Heilman

  • Error while loading master data to attribute ZMAJORGU(Major organisational unit)

    Hi Experts,
    While loading master data to  ZMAJORGU(MOU)(info object texts) from  datasource (ZOS_HRP1000)datasource iam getting error in DTP.
    below is the error...
    Diagnosis
        Data record 349 & with the key 'Ship &' is invalid in value 'Ship &' of
        the attribute/characteristic ZMAJORGU &.
    System Response
        The system has recognized that the value mentioned above is invalid, and
        has processed this general error message. A subsequent message may give
        you more information on the error. This message refers to the same
        value, even though it does not state this explicitly.
    Procedure
        If this message appears during a data load, maintain the attribute in
        the PSA maintenance screens. If this message appears in the master data
        maintenance screens, leave the transaction and call it again. This
        allows you to maintain your master data.
    Please find the below screen shots
    I have done the following regarding the issue ....Looking at the error i have gone trough the transformation and i didnt find any routines or formuales
    I have checked the PSA and down loaded the data to spread sheet and put a filter on SHORT and STEXT assuming i will find the value
    'Ship &' as mentioned in the Diagnosis description. However i was not able to find the relavent data(i went to se16 gave teh request ID from PSA and displayed the entire data)tarting with Ship .The closest  value i could find in both the fileds was was AVP Shipment .
    I have the following questions
    1)In the screen short 2 it says record number 349 is it right to go to PSA table and check against the record number 349(i tried that but i was not able to find the error against that record)
    2)why am i not finding the value 'Ship' in PSA ...but this is the value wich is creating a poblem in updating to the target ....does it mean the record value is Ship or it starts with Ship or is ir part of some word?
    Please find the attached PSA data file (u have to convert it to Excel)
    I request people to help me analyse the issue and fix the issue.
    Regards
    Kiran

    Ram
    I checked other records as well they were a lot of smallcharecters upper case charectres ,numeric special charecters .but nothing stopped ...when the system is even accpeting all other charectres  like smallof caps or numeric... why on this record is creating an issue...
    please have a look at error stack
    and i have one more question when i look at PSA there are close to 300000 records
    but when i do an SE16 with request id i see only close to one lakh records (i came to conclusion b dragging down and the last record number was (below screen shot)
    Can you tell me what my mistake is numebr or records are not tallying in SE16 and PSA..

  • Error while pulling master data

    Hello Friends,
    While i was pulling my Master data, i got following errors.
    Lock NOT set for: Loading master data attributes
    Error 1 in the update     
    The attributes for characteristic 0INSP_LOT are locked by a change run     
    while Error 1 in the update had long text as
    Diagnosis
        The update delivered the error code 1 .
    Procedure
        You can find further information on this error in the error message of
        the update.
    now where can i view the error message of the update??
    does this indicate that, the data has been transferred but not been updated?
    Thanks in advance.
    Regards,
    anuradha

    Hi,
    Seems Some attribute change run is happening in the system. So wait till that background job(SM50/SM37) is over.And then schedule this upload.
    With rgds,
    Anil Kumar Sharma .P
    Message was edited by:
            Anil Kumar Sharma

  • Where to set lock while loading master data attributes

    Hello Experts,
    I have two different datasources from ECC which loads into 0CUST_SALES attributes. I'm using two different DTP & transformation which are loading data in the this infoobject attributes.
    When i try to load data, the DTP fails with an error message at "Updating attributes for InfoObject 0CUST_SALES Processing Terminated" & the job log says: "Lock NOT set for: Loading master data attributes".
    Here what exactly  LOCK means?
    Thanks in advance.
    Zakir Khan.

    Hi,
    The lock means that some program is updating the master data (a table) and at the same time other program can not update that same master data. When the first program completes the updation it will release the lock so that the other program can update this.
    In your case you and schedule the DTP one after another.
    Thanks

Maybe you are looking for