Master Data Load Failure- duplicate records

Hi Gurus,
I am a new member in SDN.
Now,  work in BW 3.5 . I got a data load failure today. The error message saying that there is 5 duplicate records . The processing is in to PSA and then to infoobject. I checked the PSA and data available in PSA. How can I avoid these duplicate records.
Please help me, I want to fix this issue immediately.
regards
Milu

Hi Milu,
If it is a direct update, you willn't have any request for that.
The Data directly goes to Masterdata tables so don't have Manage Tab for that infoobject to see the request.
Where as in the case of flexible update you will have update rules from your infosource to the infoobject so you can delete the request in this case.
Check this link for flexible update of master data
https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/37dda990-0201-0010-198f-9fdfefc02412

Similar Messages

  • Master data Load Failure

    Hi All,
    Could any body clearly explain the following options in the Info Package of master data Info Object with full update
    Update data
    1. PSA and then into Info Obejcts
    2  . Only PSA   ( Update Subsequently in data targets )
          Data source trasfers no duplicate records
         Igonore double data records
    While I am doing master data load with the 1 and 3 options I am getting same no of records updated in PSA but loadng is getting failured with the 1st option (Caller 01 02 .. 20 Messages ) .and request with 3rd option is success..
    Thanks in Advance...
    Regards,
    Nagamani

    Hi
    U r master data might have the Dublicate records
    select the PSA
    and select the option
    Ignore dublicate data records
    IF  you have the Deblicates
    EX
    CNO
    1010
    1020
    1010
    ITs not alowed the 1010 record
    Regards

  • SAP BPC 7.5 SP 7 - Master Data Load Detected duplicate member ID

    Hi Gurus I have  a requirement.
    I am loading the Master data for Cost Center - initially while loading master data i didnot load hierarchy. Now I started to load the Master data with Hierarchy but when ever I tried vailidate the transformation file then it is throwing the "Detected Duplicate member "
    Let me show you the transformation - what I wrote there
    *OPTIONS
    FORMAT = DELIMITED
    HEADER = YES
    DELIMITER = TAB
    *MAPPING
    ID=ID
    *CONVERSION
    ID=Master_Data_Conversion.xls
    Conversion Master_Data_Conversion.xls has the following code this is to remove the spaces EXTERNALINTERNAL
    *js:%external%.toString().replace(/\s+/g,"")
    Let me show you how I have selected the Data type - Master Data /text from NW BW Infoobject
    Selection of infoobject - 0costcenter
    Format - External Format
    Set Selection - first Tab - Attribute I only wante the Control Area which is 1000 so - Control Area = 1000
    2nd Tab - Hierarchy - Import Text node - yes Hierarchy node - xxxxxxx  Version - Empty Member ID - first member    Level - Blank
    3rd Tab Language - English
    4th Tab Attribute list - Controling Area only is selected
    Note: Let me tell you that when I am loading the Master data with out Hierarchy inthe Set Selection the loading is succesful but when I am doing it with the Hierarchy as mentioned in the 2nd Tab I am getting the error as following.
    Master data (dealt by table level) has errors
    Detected duplicate member ID '201100'
    Also, the Master data for Cost Center in the BW is a time dependent so it is having Valid to and Valid from fields which is not there to be delt in the BPC.
    Please Help

    @Vinay let me tell you when the BW is having the master data which is time dependent then it will have this Duplicate member as we will have duplicate cost center names which are compounded with the Time, so BW will not show an error as there is compounding but BPC do not have that feature.
    The was raised with SAP and they resolved this issue.
    SAP Note 1641529 - Loading Master Data gets duplicate error in special case
    When running Data Manager Package 'Loading Master Data from BW
    InfoObjects', 'Duplicate Members are found' error may be reported in below
    case:
    o In the source BW InfoObjects, there are master data that have
    different lengths, and their IDs are all numeric characters. And
    o If the members are sorted by their lengths first and then IDs gets
    a different order compared with the one if they are sorted directly
    by their IDs. Take member '122' and '1101' for instance. In BW,
    they are sorted as [122, 1102]; if they are sorted directly by the
    IDs, the order is [1102, 122]. And
    o when running the package, the members are in both 'Attribute' and
    'Hierarchy' selection, and the option Filter members by Attributes
    or Hierarchies is used. And
    o Select 'External Format' when running the package.
    Other terms
    DM, Loading Master Data from BW, Duplicate Members
    Reason and Prerequisites
    It's a program error.
    Solution
    Please apply this note or upgrade to SP11.
    I hope this will help you else let me know the entire requirements so that I can provide some kind of assistance.
    Do check how the master data in ur BW is then how the hierarchy nodes and costcenter nodes are there?
    Good Luck
    Vijay Sumith

  • Master data load failure. RSRV check resulted in inconsistencies.

    Hi...
    In our production system, master data load to 0EMPLOYEE is failing every alternate day. When I check the same in RSRV, following checks are red:
    1. Time intervals in Q table for a characteristic with time-dep. master data
    2. Compare sizes of P or Q and X or Y tables for characteristic 0EMPLOYEE
    3. SID values in X and Y table: Characteristic '0EMPLOYEE'
    If I repair it, it becomes green and the load is fine. Next day the load fails again. When I check in RSRV, I get the same 3 errors. So, again I need to repair it. Let me know the permanent solution for this.
    I ran the programs: RSDG_IOBJ_REORG and RSDMD_CHECKPRG_ALL but these fixes are also temporary. Moving a new version of the object from Dev to QA and then to Production is not preferable right now as this involves a high amount of visibility.
    I know the SID tables and all are corrupted from the logs I see. But is there any permanent solution for this without transports?
    Thanks,
    Srinivas

    Hi
    Chk this link will help you: Master data deletion
    Regards
    Ashwin.

  • Master data failing with Duplicate records

    Dear All,
    Daily the ODS XD0SDO08 is getting loaded with four requests.
    While loading Master data Delta from ODS XD0SDO08 to 0Doc_Number and Cube XD_SDC08, regularly it's getting failed with the error "54 duplicate record found. 1130 recordings used in table /BI0/XDOC_NUMBER" But after deleting the Request from the 0Doc_number and the cube XD_SDC08, and reconstructing from PSA, it is succcessfull.
    If I check the PSA,there are few records in PSA, in which each Sales document number has two after image records and two before image records.If I count the number of these records they are almost equal to the number of Duplicate records we are getting in the error message.
    As we are loading to cube XD_SDC08 and 0Doc_Number, we don't have the option of Ignore Duplicate records in the Infopackge.
    Please suggest me a solution as I have to delete the request manually and reconstruct it daily.
    Regards,
    Ugendhar

    Hi ughi,
    As ur telling that data is sucessesful in cube but not in infoobject check the records in  infoobject wheather their is M version and A version if both are there ..some times load fail with duplicate data record due to this problem ...Run Attribute change run for that infoobject and  check ..
    if this will also be not sucessesful ,and as per ur update after reconstruction it is sucessful means their will be problem in update rule heck once
    another thing u will not get ignore duplicate data records because here infoobject as marked as data target ..
    if u think my suggestion is helpful assigin point to this thread
    regards,
    Gurudatt Bellary

  • Master data info object - duplicate records

    Hi All,
    I have a flat file which has department data that looks like this
    0     OTHER
    0      00 Other
    1      01 Language Arts
    2      02 Mathematics
    3      03 Reading
    4      04 Social Studies
    5      05 Science
    6      06 Student Success
    7     07 Write Traits
    1     READING
    2     LANGUAGE ARTS
    3     MATHEMATICS
    4     SCIENCE
    5     SOCIAL STUDIES
    50     50 Profesional Development
    6     BUSINESS EDUCATION
    65     65 FRABOOMERANG
    7     FOREIGN LANGUAGE
    75     75 KA Software Product
    8     Legacy = 8: No Description
    80     80 CCI
    89     89 Money Only EDC Wrkshps
    9     GENERIC
    99     99 Money Only Miscellaneou
    I was asked to create master data info object for this. I created it but when I am loading it , I am getting an error that there are duplicate records.
    What should be done? Will such data be still created as master data or an ODS ? Because there are duplicate records.
    Thanks

    Hi Eugene and Bhanu
    I have the same issue. I have this master data flat file, with 2 fields . I am trying to load it . And the data looks like this
    0----
    1----
    001 Work'g Words Spelling     
    2----
    002 Vocab for Achievmt           
    3----
    003 ETS95 Packets Program     
    4----
    004 Every Day Counts            
    5----
    005 Phonics We Use            
    6----
    006 US94 History of US         
    7----
    007 Masterng SAT 1/PSAT        
    I tried Bhanu's suggestion of ignoring duplicate records . I did not get any error but that I lost some data. So I tried Eugene's suggestion of shifting the fields .. I moved the second field to the frist place and saved and loaded it. I got this error.
    "There are duplicates of the data record 2. with the key '3.' for characteristic 1.."
    Any suggestions.
    Thanks

  • Data Loader inserting duplicate records

    Hi,
    There is an import that we need to run everyday in order to load data from another system into CRM On Demand . I have set up a data loader script which is scheduled to run every morning. The script should perform insert operation.
    Every morning a file with new insert data is available in the same location(generated by someone else) & same name. The data loader script must insert all records in it.
    One morning , there was a problem in the other job and a new file was not produced. When the data loader script ran , it found the old file and re-inserted the records (there were 3 in file). I had specified the -duplicatecheckoption parameter as the external id, since the records come from another system, but I came to know that the option works in the case of update operations only.
    How can a situation like this handled in future? The external id should be checked for duplicates before the insert operation is performed. If we cant check on the data loader side, is it possible to somehow specify the field as 'unique' in the UI so that there is an error if a duplicate record is inserted? Please suggest.
    Regards,

    Hi
    You can use something like this:
    cursor crs is select distinct deptno,dname,loc from dept.
    Now you can insert all the records present in this cursor.
    Assumption: You do not have duplicate entry in the dept table initially.
    Cheers
    Sudhir

  • Duplicate records during master data loading.

    hello guys,
    I am reading one blog where the blogger wrote about 'Various issues in a BW Production project'.....I came across one issue which I couldnot understand...
    Data loading failed due to Duplicate records during master data loading.......
    Why do this error occur?How can we rectify this in a production environment?
    Thanks and Regards,
    S

    Hi SChandx200 ,
          May I ask where you get "Various issues in a BW production project"?
    Many Thanks,

  • Master Data Load - Number records not match

    I am loading the 0PLANT table.  In R3, it shows 966 records.   After running the InfoPackage, there are 966 records in the PSA, none in error.  But I get a message - Number of records  requested does not match number transferred (packet 1).
    This is a first use of a new source system to this BW.  All master data loads give me the same error.
    What is setup wrong?

    Hello RoberT,
    & How r u ?
    For the same 0PLANT we load data on daily basis. Not only for this but also for more Master Data Objects. For some of them we used to get this message but not as an Error. On pressing the Update Tree (Refresh) it will change the status.
    Anyway u got the records in the PSA so there is no problem with the data load. What signal u find in the Monitor ? Yellow or Red ? If yellow wait for some time, press Update Tree.
    Best Regards....
    Sankar Kumar
    +91 98403 47141

  • Master data loading issue

    Hi gurus,
        Presently i am working on BI 7.0.I have small issue regarding master data loading.
        I have generic data soruce for master data loading.i have to fetch this data to BW side.Always i have to do full data load to the master data object.first time i have scheduled info package and run DTP to load data to master data object, no issues, data got loaded successfully.whenever i run infopacage for second time and run DTP i am getting error saying that duplicated records.
       How can i handle this.
    Best Regards
    Prasad

    Hi Prasad,
    Following is happening in your case:
    <b>Loading 1st Time:</b>
    1. Data loaded to PSA through ipack.It is a full load.
    2. data loaded to infoobject through DTP.
    <b>Loading 2nd Time:</b>
    1. Data is again loaded to PSA. It is a full load.
    2. At this point, data in PSA itself is duplicate. So when you are running the DTP, it picks up the data of both the requests that were loaded to PSA. And hence, you are getting the Duplicate record error.
    Please clear the PSA after the data is loaded to infoobject.
    Assign points if helpful.
    Regards,
    Tej Trivedi

  • 0EQUIPMENT MASTER DATA LOAD ERROR

    Hi,
    I am trying to load master data for 0EQUIPMENT in PM,an error occured saying
    "The time interval ['20061203'/'20061203'] & (from/to) for the data
    records 1 & and 2 & overlaps in characteristic 0EQUIPMENT &. "    
    Please give your suggestions to fix this error.
    Regards,

    Check with the folks in the source system and find out why there are duplicate records. For a given period the keyfields can only appear once or else the load would fail. If it is ok (upon confirmation with functional folks) you can use the option of loading to PSA (3rd option in the infopackage) and then load subsequently to data targets & ignore duplicate records. But like I said, if I were you I would first try to find out why there are duplicate records before I start ignoring them.
           Hope this helps.

  • Sap bw 7.3 master data load error

    sap bw 7.3 master data load error
    erroe: Exception in Substep Rules

    Hi Jayram,
    I am assuming that you are getting this error when you are loading data from PSA to IO. If so, the error might be because of
    1. Duplicate records
    PSA might have duplicate records. But you will be able to load only one record to Master data IO. In DTP Update tab, there is an option for "Handle Duplicate Record Keys" . Just enable this & try to load again.
    2. Erroneous records.
    Some special reocrds are allowed in PSA where as thye might not be allowed in Master Data/Data Targets like Lower case letters/ some speacial characters etc. If error is because of this, then you might need to correct the data at PSA level & reload the data to Master data. Else get it corrected in source system itself & fetch the data to BW. Or write some code to take care of these special characters.
    Hope it helps!
    Regards,
    Pavan

  • Master data load for 0COSTCENTER (Cost Center) failing

    Hi Experts
    I have a master data load for 0COSTCENTER (Cost Center). The Load has started failing from a couple of days at DTP.
    (R) Filter Out New Records with the Same Key
          (G) Starting Processing...
          (R) Dump: ABAP/4 processor: DBIF_RSQL_SQL_ERROR
    I am unable to understand the reason for this failure. I tried loading the data 1 costcenter at a time it is still failing so i doubt if its a internal table storage issue as suggested by the dump.
    Could you help me on this one please.
    Regards
    Akshay Chonkar

    Thanks All
    I have got the issue resolved
    The Error Stack for the DTP had accumulated so many entries that it was unable to process further.
    So i deleted the entries in Error Stack usin SE14. Then again executed my DTP.
    Everything is fine now thanks for your help.
    Regards
    Akshay Chonkar

  • Master Data loading got failed: error "Update mode R is not supported by th

    Hello Experts,
    I use to load master data for 0Customer_Attr though daily process chain, it was running successfully.
    For last 2 days master data loading for 0Customer_Attr got failed and it gives following error message:
    "Update mode R is not supported by the extraction API"
    Can anyone tell me what is that error for? how to resolve this issue?
    Regards,
    Nirav

    Hi
    Update mode R error will come in the below case
    You are running a delta (for master data) which afils due to some error. to resolve that error, you make the load red and try to repeat the load.
    This time the load will fail with update mode R.
    As repeat delta is not supported.
    So, now, the only thing you can do is to reinit the delta(as told in above posts) and then you can proceed. The earlier problem has nothing to do with update mode R.
    example your fiorst delta failed with replication issue.
    only replicating and repeaing will not solve the update mode R.
    you will have to do both replication of the data source and re-int for the update mode R.
    One more thing I would like to add is.
    If the the delat which failed with error the first time(not update mode R), then
    you have to do init with data transfer
    if it failed without picking any records,
    then do init without data transfer.
    Hope this helps
    Regards
    Shilpa
    Edited by: Shilpa Vinayak on Oct 14, 2008 12:48 PM

  • Problem with Master Data Load

    Dear Experts,
    If somebody can help me by the following case, please give me some solution. Iu2019m working in a project BI 7.0 were needed to delete master data for an InfoObject material. The way that I took for this was through tcode u201CS14u201D. After that, I have tried to load again the master data, but the process was broken and the load done to half data.
    This it is the error:
    Second attempt to write record 'YY99993' to /BIC/PYYYY00006 failed
    Message no. RSDMD218
    Diagnosis
    During the master data update, the master data tables are read to determine which records of the data package that was passed have to be inserted, updated, or modified. Some records are inserted in the master data table by a concurrently running request between reading the tables at the start of package processing and the actual record insertion at the end of package processing.
    The master data update tries to overwrite the records inserted by the concurrently running process, but the database record modification returns an unexpected error.
    Procedure
    u2022     Check if the values of the master data record with the key specified in this message are updated correctly.
    u2022     Run the RSRV master data test "Time Overlaps of Load Requests" and enter the current request to analyze which requests are running concurrently and may have affected the master data update process.
    u2022     Re-schedule the master data load process to avoid such situations in future.
    u2022     Read SAP note 668466 to get more information about master data update scheduling.
    Other hand, the SID table in the master data product is empty.
    Thanks for you well!
    Luis

    Dear Daya,
    Thank for your help, but I was applied your suggesting. I sent to OSS with the following details:
    We are on BI 7.0 (system ID DXX)
    While loading Master Data for infoobject XXXX00001 (main characteristic in our system u2013 like material) we are facing the following error:
    Yellow warning u201CSecond attempt to write record u20182.347.263u2019 to BIC/ XXXX00001 was successfulu201D
    We are loading the Master data from data source ZD_BW_XXXXXXX (from APO system) through the DTP ZD_BW_XXXXX / XXX130 -> XXXX00001
    The Master Data tables (S, P, X) are not updated properly.
    The following reparing actions have been taken so far:
    1.     Delete all related transactional and master data, by checking all relation (tcode SLG1 à RSDMD, MD_DEL)
    2.     Follow instructions from OSS 632931 (tcode RSRV)
    3.     Run report RSDMD_CHECKPRG_ALL from tcode SE38 (using both check and repair options).
    After deleting all data, the previous tests were ok, but once we load new master data, the same problem appears again, and the report RSDMD_CHECKPRG_ALL gives the following error.
    u201CCharacteristic XXXX00001: error fund during this test.u201D
    The RSRV check for u201CCompare sizes of P and X and/or Q and Y tables for characteristic XXXX00001u201D is shown below:
    Characteristic XXXX00001: Table /BIC/ PXXXX00001, /BIC/ XXXXX00001 are not consistent 351.196 derivation.
    It seems that our problem is described in OSS 1143433 (SP13), even if we already are in SP16.
    Could somebody please help us, and let us know how to solve the problem?
    Thank for all,
    Luis

Maybe you are looking for

  • Typewriter problems in Acrobat 9 Pro

    I've been using the typewriter tool in Acrobat 9 Pro, and up until yesterday was able to change the font and size I was typing in. As of now, the whole typewriter tool bar at the top of my PDF is greyed out and seems to be locked. Please help! I have

  • Changing the document page size after the fact.

    Does anyone know of a way to change the page size globally after the fact? I have sent a 200 page book created with Pages '08 to my printer. The work has a full bleed watermark in the background on a custom 7" X 11" document. I've just been informed

  • Connecting AE to Dlink Router DI-624

    Greetings all ! just purchased the AE and already having difficulty :P I want to transfer music from my PC to my living room stereo. Tried connecting the AE to my network with no sucess. Here's my config and if someone can help me make this thing wor

  • Regarding link button by generating query

    Hi all .. Can any one tel me solution for the folloeing problem . i wrote a query for getting doc num based on card code. when i run that query i tshows the doc nums with a link when we click that link it will show that particular transaction .......

  • Right path for sap hana

    Is this a right track for a fresher to get into SAP HANA? I am sap abap consultant and have interest to learn sap hana... According to me,I wanted to first learn sap bi through a small or local institute and the n go for sap hana training from Sap ba