Getting duplicate records in cube from each data packet.

Hi Guys,
I am using 3.x BI version. I am getting duplicate records in cube. for deleting these duplicate records i have written code. still it is giving same result. Actually i have written a start routine for deleting duplicate records.
These duplication is occurring depending on the number of packets.
Eg: If the number of packets are 2 then it is giving me 2 duplicate records.
If the number of packets are 7 then it is giving me 7 duplicate records.
How can i modify my code so that it can fetch only one record by eliminating duplicate records? Or any other solution is welcomed.
Thanks in advance.

Hi  Andreas, Mayank.
  Thanks for your reply.
  I created my own DSO, but its giving error. And I tried with the stanadard DSO too. Still its giving the same error as could not activate.
In error its giving a name of function module RSB1_OLTPSOURCE_GENERATE.
I searched in R3 but could not get that one.
Even I tried creating DSO for trial basis, they are also giving the same problem.
I think its the problem from BASIS side.
Please help if you have any idea.
Thanks.

Similar Messages

  • Deleting duplicate records from different data packets in BI data source.

    Hi,
    I am getting same (duplicate) records from different data packets in BI data source, after completion of extraction.
    I tried to store key fields of the first data packet in an internal table. But this internal table is not carrying the previous data at the time of extraction of second data packet.
    Is there any other way to remove duplicate records after completion of extraction.
    Thanks in advance.

    I did not extensively worked in BI routenes. But I recon there will be routene which will het executed before data mapping part there will be a start routene in which you can validate the existense of data before beeing passed from data source to cube.
    Hope this helps,
    Regards,
    Murthy.

  • Duplicate records in cube.

    Hi,Xperts,
    i have checked in my PSA which has no duplicate records.but when i am loading the data to cube,in cube i am getting duplicate records.
    can any one help me on this?

    Hi Satish,
    please check in R/3,
    U told that it is delta load,  go to RSO2, select the required one and enter
    there click on the generic delta tab and check the settings: what u have given:
    Safety level lower limit: give the particular value to eliminate the data records so that the system knows from where exactly the data has to be loaded.
    Check the two options below:
    New status for changed records
    Additive Delta
    Select the appropriate one.
    if helpful, please try this.
    Regards
    Swathi

  • BI 7.0 - Duplicate Record Error while loading master data

    I am working on BI 7.0 and I am trying to load master data to an info object.
    I created an Infopackage and loaded into PSA.
    I created transformation and DTP and I get an error after I execute the DTP about duplicate records.
    I have read all previous threads about duplicate record error while loading master data and most of the them suggested to check 'Ignore duplicate records' option in the infopackage. But in 7.0, I can only load to PSA with infopackage and it doesn't have any option for me to ignore duplicate records.
    My data is getting loaded to PSA fine and I get this error while loading to info object using DTP.
    I would appreciate your help to resolve this issue.
    Regards,
    Ram.

    Hi,
    Refer:
    http://help.sap.com/saphelp_nw2004s/helpdata/en/45/2a8430131e03c3e10000000a1553f6/frameset.htm
    http://help.sap.com/saphelp_nw2004s/helpdata/en/d0/538f3b294a7f2de10000000a11402f/frameset.htm
    With rgds,
    Anil Kumar Sharma .P

  • OLAP cubes from heterogeneous data sources using XML DB

    hi.
    Q:1
    How we can create an OLAP cube (XML cubes) from XML data sources if using XML DB.
    Q:2
    How we can create an OLAP cubes from warehouses and flat files and covert these cubes into XML cubes.
    Q:3
    Is there any other tool (Except Analytical workspace manager AWM) which supports the construction of OLAP cubes in XML format from heterogeneous data sources?
    Edited by: user11236392 on Aug 21, 2009 3:50 AM

    Hi Stuart!
    Your undersatnding is partially correct. i am working for providing an architecture for XOLAP, XML is one of my data source.
    The idea is to generate uniform cubes from heterogeneous sources that can be integrated into a global cube. instead of building Oracle OLAP cubes from all the sources (this work is already done) i want to generate an XCube from XML data sources using XQuery.
    On the other hand if we have generated the Oracle OLAP cube from other sources like warehouses or flat files. i have to convert these Oracle OLAP cube into XML cube for uniformity. in an research paper i find that there is an operator Xcube embedded in XQuery which converts the multidimensional data (cube) into XML cube. im looking for the implementation of this operator in Query that how this operator works.
    hope u understand my architecture. but if u still have some confusion, kindly give me ur mail id. i will mail the diagram of my architecture.
    thanks.
    saqib

  • ODI CDC - Getting Duplicate Records in Journal Data

    Dear Gurus
    I am getting the following issues in CDC using Oracle Logminer
    01) All operations on Source Data (Insert, Update) are shown as only Insert in Model -> CDC -> Journal Data
    02) The records in Model -> CDC -> Journal Data are double
    03) These are not travelling to Desitination table, I want to load the destiation table
    04) Is it possible to have the last value and the new value both available on same screen as output in ODI, I want to see what data changed before actually populating the tables

    Hi  Andreas, Mayank.
      Thanks for your reply.
      I created my own DSO, but its giving error. And I tried with the stanadard DSO too. Still its giving the same error as could not activate.
    In error its giving a name of function module RSB1_OLTPSOURCE_GENERATE.
    I searched in R3 but could not get that one.
    Even I tried creating DSO for trial basis, they are also giving the same problem.
    I think its the problem from BASIS side.
    Please help if you have any idea.
    Thanks.

  • Duplicate records in Cube Level ( BI 7.0 )

    Dear All
    I am working on BI 7.0 , I have an issue , i am  loading the data from  Flat File to ODS  and from ODS to Cube . In ODS  we are selected Overwrite option, in cube level we have an Summation option. the problem is while loading the data from Flatfile to ODS  records are fine . from ODS  to cube data  loading also fine but here i am getting the Duplicate records .
    what are the  best options to go ahead in such a situations??
    Regards
    KK

    I am sharing a case occured for me. Please see if its applicable for you.
    Sometimes, in the step loading to Cube, when any type of problem occurs, we restart the load. If the cube load prompts saying 'the lost load was unsuccessful.... reload?', this problem may occur. It will load the records in the previous load also.
    Verify what has been duplicated from the ODS changelog table and the cube load record count. If you see the number of records updated being the total of the records in the different ODSR request (in the change log table). Delete the previous load in the cube. (provided no other side effect is produced e.g from start routine etc)
    Cheers.

  • How to delete duplicate records in cube

    Hi,
    can u help me how to delete the duplicate records in my cube
    and tell me some predifined cubes and data sourcess for MM and SD modules

    Hi Anne,
    about "duplicate records" could you be more precise?.
    The must be at least one different Characteristic to distinguish one record from the other (at least Request ID). In order to delete Data from InfoCubes (selectively) use ABAP Report RSDRD_DELETE_FACTS (be carefull it does not request any confirmation as in RSA1 ...).
    About MM and SD Cubes see RSA1 -> Business Content -> InfoProvider by InfoAreas. See also for MetadataRepository about the same InfoProviders.
    About DataSources just execute TCode LBWE in you source sys: there you see all LO-Cockipt Extrators.
    Hope it helps (and if so remember reward points)
    GFV

  • Duplicate Records in Details for ECC data source. Help.

    Hello. First post on SDN. I have been searching prior posts, but have come up empty. I am in the middle of creating a report linking directly into 4 tables in ECC 6.0. I am having trouble in getting either the table links set up correctly, or filtering out duplicate record sets that are being reporting in the details section of my report. It appears that I have 119 records being displayed, when the parameters values should only yeild 7. The details section is repeating the 7 records 17 times (there are 17 matching records for the parameter choices in one of the other tables which I think is the cause).
    I think this is due to the other table links for my parameter values. But, I need to keep the links the way they are for other aspects of the report (header information). The tables in question are using an Inner Join, Enforced Both, =. I tried the other link options, with no luck.
    I am unable to use the "Select Disctinct Records" option in the Database menu since this is not supported when connecting to ECC.
    Any ideas would be greatly appreciated.
    Thanks,
    Barret
    PS. I come from more of a Functional background, so development is sort of new to me. Take it easy on the newbie.

    If you can't establish links to bring back unique data then use a group to diplay data.
    Group report by a filed which is the lowest commom denominator.
    Move all fields into group footer and suppress Group header and details
    You will not be able to use normal summaries as they will count/sum all the duplicated data, use Running Totals instead and select evaluate on change of the introduced group
    Ian

  • Combining Data into one Cube from two Data-sources..

    Dear Experts,
    I am pulling data from two data sources and trying to combine in one Info-Cube. The data are like
    Data-Source 01
    1. GUID  --Common
    2.Document No ( User Entry)
    3.Dist. Channel
    4.Transaction Type
    5.Date and Quantity
    Data-Source 02
    1.GUID -- Common
    2.Billing Document ( If User drill down according to Document No , Billing Document should come in the report )
    3.Billing date
    4.Net Value
    Out of the datas , The GUID is common between the 2 data-sources.  I was thinking that, tha data will take according to its place and If i select the Document No in Report, it will atomatically fetch all the data like Tran type, dist ch, Billing Document No , Billing date.. .
    The problem is , in the report Tha data is not coming as I was thinking.
    And Another problem is , In future I need to create a Multiprovider between the above mentioned Info-cube and One ODS. And  DOCUMENT NO is common in Cube and ODS.
    Please Suggest,
    How can I proceed for the following requirement.
    Thanks,
    Sanjana

    Hi Sanjana,
    In your case cube will create a problem because it will have multiple records . For example :
    Data-Source 01 :
    1. GUID -- 101
    2.Document No - 999
    3.Dist. Channel - DL
    4.Transaction Type - GPRO
    5.Date and Quantity - 20.02.2011 & 20
    Data-Source 02
    1.GUID -- 101
    2.Billing Document - 6000
    3.Billing date - 03.03.2011
    4.Net Value - 500
    Your cube will have 2 records . And your requirement is to show above two records in 1 record in the report .
    Why dont you make an ODS in between , where you can put GUID as the Key field and rest all the fields as data fields. Create 2 transformations to this DSO from the 2 datasources . And let it get updated one by one . Your DSO will have 1 record only . Now either you do reporting on this DSO or take the data to the cube .
    Hope the above reply was helpful.
    Kind Regards,
    Ashutosh Singh
    Edited by: Ashutosh Singh on May 19, 2011 1:34 PM

  • Master Data: 2 records with different FROM-TO dates

    Hi,
    I am extracting the master data from R/3. In R/3 the master data record is created with valid FROM: 01/01/1900 and valid TO: 12/31/9999. But when I extract the data into BW it is coming as 2 records one with 01/01/1000-12/31/1899 range and the other with 01/01/1900-12/31/9999.
    What settings/steps I need to change to get only one record for now and of course in future any change to the records will comes as different valid date range.
    When I schedule the info package, the update tab consist of Fixed time interval start date: 10/20/2005; end date: 12/31/9999.
    Time being to avoid this in the transfer rule of FROM date info object selected constant option with date 01/01/1900 as in R/3????
    Appreciate your help on this.
    Thanks.
    Sudha.

    Hi Nagesh,
    Thanks for reply.
    I have assigned one-to-one from R/3 to BW.
    In R/3 I have FROM and TO date and assigned these two fields to Transfer structure FROM and TO info objects.
    I have maintained Transfer rule only after getting two records into BW for every one entry in R/3. In the rule I just assigned constant date to avoid 2 records for every entry from r/3. So nothing to do with the routine.
    I want to see one record in BW for every entry coming from R/3. At present in the BW master data sitting as 2 records.
    Example:
    one record in R/3 with:
    FROM date: 01/01/1900; TO date: 12/31/9999.
    Inserted 2 records into BW with:
    FROM date: 01/01/1000; TO date: 12/31/1899.
    FROM date: 01/01/1900; TO date: 12/31/9999.
    Thanks.
    Sudha.

  • I keep getting duplicate photos when downloading from my card reader.

    In my PSE11 Camera and Card Reader preferences I have checked "Copy new files only".  Yet whenever I download a camera card that still contains photos that were previously downloaded, I end up with duplicate photos in PSE11. 
    I prefer to leave my photos on the camera card as a "backup" in case of a hard drive crash or other loss of data on the computer.  Once the card is full I archive it.
    Why do I get duplicates even though this setting makes it clear that I should not?  It is very annoying having to look through my photo collection and manually delete dozens or hundreds of duplicate photos.
    Thank you.
    Jon

    First, try contacting the developer of the app here >  http://support.aspyr.com/home
    If you do not get a reply within three to five business days, report the issue to Apple here > Apple - Support - Mac App Store - Contact Support
    Apple's policy states, "all sales are final" > TUNES STORE - MAC APP STORE -TERMS AND CONDITIONS but if you explain the error and you haven't received areply from the developer, it can't hurt to request a refund.
    You're sure there's enough free space on the startup disk?
    edited by:  cs

  • Duplicate records found while loading master data(very urgent)

    Hi all,
    One infopackage in the process chain failed while laoding the master data(full update).Its showing the following error-->duplicate record found ..1 record used in /BI0/PTCTQUERY and the same record occured in /BI0/PTCTQUERY tables.
    can anyone give me the solution...its very urgent...
    Thanks & Regards,
    Manjula

    Hi
    You can see the check box in the Processing tab page. Make a tick mark for the check box Ignore Duplicate Data Records indicator . When multiple data records that have the same key are transferred, the last data record in the request is updated to BI. Any other data records in the request with the same key are ignored.
    Help says that:
    To maintain consistency, ignoring duplicate data records is only possible if the data is updated serially. A serial update is when data is first updated into the PSA then, after it has been successfully written to the PSA, it is updated into the master data or text tables of the InfoObject.
    If a DataSource transfers potentially duplicate data records or if you manually set the Ignore Duplicate Data Records indicator, the PSA Only update type is automatically selected in the scheduler.
    hope it clears ur doubt, otherwise let me know.
    Regards
    Kiran

  • Number of records in cube from Query Designer

    I don't have access to the cube in BW(listschema). I only have access to pull reports from cube from Query Designer. How can I tell the total number of records in a cube?
    Thanks.

    Hi
    you can use the tech content in the query designer to display the count for no of records or you can do the same via creating a new CKF
    or see the below link to display the count
    https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/009819ab-c96e-2910-bbb2-c85f7bdec04a

  • How to get name of the month from current date.

    Hi,
       How to get the name of the month from current date.
    Thanks,
    Senthil

    Sethil,
    Use your date(let us say Date1) instead of sy-datum.
    CALL FUNCTION 'MONTH_NAMES_GET'
    EXPORTING
    LANGUAGE = SY-LANGU
    IMPORTING
    RETURN_CODE =
    TABLES
    MONTH_NAMES = itab_month
    EXCEPTIONS
    MONTH_NAMES_NOT_FOUND = 1
    OTHERS = 2
    IF SY-SUBRC <> 0.
    MESSAGE ID SY-MSGID TYPE SY-MSGTY NUMBER SY-MSGNO
    WITH SY-MSGV1 SY-MSGV2 SY-MSGV3 SY-MSGV4.
    ENDIF.
    READ TABLE itab_month
    WITH KEY MNR = date1+4(2).
    itab_month-LTX will contain the value you are looking for
    Good luck
    Raghava

Maybe you are looking for