Duplicate Data Records indicator / the handle duplicate records

Hi All,
I am getting double data in two request. How can I delete extra data using "Duplicate Data Records indicator ".
I am not able to see this option in PSA as well as DTP for "the handle duplicate records".
can u help me to find the option in PSA/DTP.
Regards
Amit Srivastava

What Arvind said is correct.
But if you can try this out in an End Routine, this may work, Not sure though.
Because then you will be dealing with the entire result_package.
Also, say if the target that you are talking about is a DSO, then you can Delete Adjacant Duplicates in Start Routine while updating it into your next target. That can be a cube, for ex.

Similar Messages

  • Master data infoobject can't handle duplicate records after SP10

    Hi
    I am trying to load master data which happened to contain duplicate records from the source system.  In the DTP of the master data infoobject, I have ticked the 'Handle Duplicate Record Keys' checkbox.  After executing this DTP, the duplicate master data records were trapped in the Error Stack.  I am expecting overwriting of the duplicate master data to take place instead.  I understand that this error was fixed in Note 954661 - Updating master data texts when error in data package which is from SP9.  After applying Support Pack 10, the master data infoobject just can't handle records with duplicate keys.
    Please let me know if you manage to fix this problem.
    Many thanks,
    Anthony

    Found a fix for this problem.  Just applied this OSS note  Note 986196 - Error during duplicate record handling of master data texts.

  • I want to diaplay my records from the databasewith 10 records per page. i l

    i want to diaplay my records from the databasewith 10 records per page. i learned that it can be done through pagination . Please can anyone there help me and give me the code of how to page my records in my jsp pages
    I have small code ,but its not working....
    below is the code....
    public ResultSet codeResult(int i){
    if(dataSource!= null){
         try {
                             con=dataSource.getConnection();
         stmt=con.createStatement (ResultSet.TYPE_SCROLL_SENSITIVE,ResultSet.CONCUR_READ_ONLY);
         int maxRows=stmt.getMaxRows();
         System.out.println("the maxRows is:"+maxRows);
         System.out.println("the I value is:"+i);
         if((i < 0) || (i > maxRows)){
              i=0;
         stmt.setFetchSize(i);
         long t = System.currentTimeMillis();
         rs=stmt.executeQuery("select asc_code,countrycode,asc_name,asc_address1,asc_address2,asc_city,asc_country,asc_region,asc_contact_person,asc_contact_no,asc_email_id,asc_fax_no,nso_incharge,nso_email,ar_ap_code from asc_master order by asc_code asc");
    but,its giving all records...
    any one help me on this........
    Thank you,
    D.Nagireddy.                    
                        } catch (SQLException e) {
                             e.printStackTrace();
              return rs;
              }

    hey, here my code for paging. hope it helps...
         public void DueRecordInfo(int page,int MAX_PER_PAGE)
              con = new DBConnection().getConnection();
             try
                     Statement stmt = con.createStatement();
                     String sqlcount = "SELECT COUNT(*) FROM INFO WHERE TO_DATE(DATEUPDATED,'Month DD, YYYY') < ADD_MONTHS(SYSDATE,-6) AND OBSOLETE LIKE 'No'";
                     ResultSet rscount = stmt.executeQuery(sqlcount);     
                    rscount.next();
                    total_row_count = rscount.getInt(1);
                  lowerbound = (page * MAX_PER_PAGE - (MAX_PER_PAGE))+1;
                  upperbound = lowerbound - 1 + MAX_PER_PAGE;
                  String sql = "Select * from "+
                                    "(SELECT row_number() over(ORDER BY TO_DATE(DATEUPDATED,'Month DD, YYYY') DESC) AS RN,ID,SPECNO,REVISEDNO,DATEREVISED,TITLE,ORIGINATOR,DCBNUMBER,DCBLOCATION,OBSOLETE,DATEOBSOLETE,WORDFILE,PDFFILE,DATEUPDATED FROM INFO WHERE TO_DATE(DATEUPDATED,'Month DD, YYYY') < ADD_MONTHS(SYSDATE,-6) AND OBSOLETE LIKE 'No') " +
                                    "WHERE RN BETWEEN "+ lowerbound + " AND " +upperbound;
                   ResultSet rs = stmt.executeQuery(sql);
                  while (rs.next())
                         id.add(num, rs.getString("ID"));
                         specno.add(num, rs.getString("SPECNO"));
                         revisedno.add(num, rs.getString("REVISEDNO"));
                       daterevised.add(num, rs.getString("DATEREVISED"));
                       title.add(num, rs.getString("TITLE"));
                       originator.add(num, rs.getString("ORIGINATOR"));
                       dcbnumber.add(num, rs.getString("DCBNUMBER"));
                       dcblocation.add(num, rs.getString("DCBLOCATION"));
                       obsolete.add(num, rs.getString("OBSOLETE"));
                       dateobsolete.add(num, rs.getString("DATEOBSOLETE"));
                       wordfile.add(num, rs.getString("WORDFILE"));
                       pdffile.add(num, rs.getString("PDFFILE"));
                       dateupdated.add(num, rs.getString("DATEUPDATED"));
                         num++;
                   con.close();
                   rscount.close();
                   rs.close();
                   stmt.close();
             }

  • Page level validation to prevent duplicate data entry into the database

    Hello,
    Can anyone please help me out with this issue.
    I have a form with two items based on a table. I already have an item level validation to check for null. Now I would like to create a page level validation to check that duplicate data are not entered into the database. I would like to check the database when the user clicks on ‘Create’ button to ensure they are not inserting duplicate record. If data already exist, then show the error message and redirect them to another page. I am using apex 3.2
    Thanks

    Hi,
    Have you tried writing a PLSQL function to check this?
    I haven't tested this specifically, but something like this should work:
    1) Create a Page Level Validation
    2) Choose PLSQL for the method
    3) Choose Function Returning Boolean for the Type
    For the validation code, you could do something like this:
    DECLARE
        v_cnt number;
    BEGIN
        select count(*)
        into v_cnt
        from
        your_table
        where
        col1 = :P1_field1 AND
        col2 = :P2_field2;
        if v_cnt > 0 then return false;
        else return true;
        end if;
    END;If the query returns false, then your error message will be displayed.
    Not sure how you would go about redirecting after this page though. Maybe just allow the user to try again with another value (in case they made a mistake) or just press a 'cancel' button to finish trying to create a new record.
    Amanda.

  • Multiple info records for the same material record in the catalog

    Hi Guys,
    I have got sort of a puzzle I can not solve at the moment, maybe some one can help me to solve it. The situation in our system landscape  is as follows:
    Systems:
    SRM 5.0 (Classic implementation scenario)
    ECC 6.0
    SRM-MDM 2.0
    PI 2005
    Material master is maintained in R/3 and the materials are replicated to SRM. From SRM we replicate all materials to the catalog repository in  SRM-MDM.
    As the material master in R/3 is vendor independent, the materials replicated to SRM also have no vendor attached. For the first initial material replication from SRM to the catalog SRM-MDM this also means that the products in the main table of the repository, u2018catalog positionsu2019, will not have a vendor or vendor number attached (i.e. no reference to a source of supply).
    To assign a source of supply in the repository to a product (record) we will use info records which are replicated from R/3 to SRM-MDM.  The mapping for this in the Import Manager can only be based on the product number. As this is the only unique value available which can map a info record to a already existing (but vendor less) ptoduct item in the catalog.
    As long as every item in the repository has only one unique info record, meaning one source of supply with a specific price, there is no issue in the above scenario. Based on the product-id, the info record data will be matched with the material item in the catalog.
    The issue for me arises when multiple info records exists for the same material. The info records can heave different vendors and different prices. All possible vendors should be available in the catalog for the same material (at least this is the requirement ). 
    At this point the product-id will no longer be an unique value. Importing the info records will cause problems, as only one material record is available in the repository for the specific product-id (remember that when doing a initial replication of the material master to the catalog, no vendor data is replicated).
    Does anyone had this issue before, and knows a solution? Is it for example possible during the import of the info records in the Import manager, to duplicate material records in the destination data, based on the number of info records available in the source data for the same product-id. Or is there an other solution that I am missing?
    Your help would be appreciated!
    Regards,
    Skander

    Dear Poster
    Your thread has had no response since it's creation over
    2 weeks ago, therefore, I recommend that you either:
    - Rephrase the question.
    - Provide additional Information to prompt a response.
    - Close the thread if the answer is already known.
    Thank you for your compliance in this regard.
    Jason Boggans
    SAP SRM SDN Moderator

  • Retrive a record before the last inserted record

    Hi,
    I use simple select statement to retrive records where cola = 'Yes':
    select * from tbl1
    where cola = 'Yes';
    The result gives me about 10 records with different date.
    From the above result, I want to get the record right before the last inserted record according to the date field; which means I can not use MAX function (all I want just one record).
    If I write a function to return a date and passes it back to the SELECT statement, I'll get a date for each record, which I don't want.
    I know there's a way to do that in PL/SQL.
    Plese help me.
    pem

    select * from (
    select * from
    (select * from (select * from (table> order by <date> desc) where rownum < 3)
    order by <date>) where rownum =1

  • Merging records - how to handle deleted records?

    hi,
    can anyone tell me what is the proper way to handle
    deleted records during records update ?
    let's say I have 2 customers and I merge them into one
    so I need to send to transactional systems that the deleted one
    should not longer be active - but how to achive that
    (sending deletion info) in MDM ?
    thank you for your help,
    Regards,
    michal

    Hi Michal,
    MDM does not delete any records when it merges them. whenever you merge any two records the local id is never lost, it is always present in the edit key mapping.
    You can see this info by right clicking on record and select edit key mapping. You will find both the record entries there.
    With proper mapping in syndicator, you can get this information for both the records. while mapping map the 'VALUE' field in the source items to the 'LOCAL ID' field in the destination item.
    Please let me know if this not help completely and if you have more doubts.
    - An

  • No 'Handle Duplicate records' in update tab

    Hi,
    I've a DTP from ODS/DSO to ODS/DSO and I got a Duplicate record error, which I find rather strange for standard ODS/DSO. I read in the help and within these forums that it can be fixed with the 'Handle Duplicate records' checkbox in the update tab. Trouble is that there isn't such a checkbox in our BI7 SP15 installation.
    Any suggestion on the reason for that and how to fix this and more important on how to get rid of the duplicate record error (and the reason why it's occurring)?
    Many thanks in advance
    Eddy

    Hi Eddy,
    I am confused -:)
    Have u tried by checking or by unchecking..
    My suggestion is to try by selecting the setting unique data records...
    Cheers
    Siva

  • Handling Duplicated Records in DTP

    Dear Experts,
    I am trying to load to the master data of 0PLANT using datasource 0BBP_PLANT_LOCMAP_ATTR (Location/Plant Mapping) using DTP.
    This standard datasource is neither language- nor time-dependent. Also, in the source system, it is not marked to handle duplicate records.
    I have also referred to OSS Note 1147563 - Consulting note: Indicator "Duplicate records". One of the key highlight is "If you use a DTP to transfer master data or texts, you can set the indicator 'Duplicate records' in the DTP tab page 'Data transfer'." I would suppose that this means the "Handle Duplicated Record Keys" option under tab "Update" of the respective DTP.
    In this OSS Note, it was also mentioned that
    "You must not set the indicator if the following prerequisites apply:
    The indicator 'DataSource delivers duplicate records' is not set in the DataSource."
    >> which is currently the case of datasource 0BBP_PLANT_LOCMAP_ATTR
    Checked in SAP Help [link|http://help.sap.com/saphelp_nw04s/helpdata/en/42/fbd598481e1a61e10000000a422035/frameset.htm]:
    You can specify how duplicate data records within a request are handled, independently of whether the setting that allows DataSources to deliver potentially duplicate data records has been made. This is useful if the setting was not made to the DataSource, but the system knows from other sources that duplicate data records are transferred (for example, when flat files are loaded).
    My question is, I can't load the master data mainly because of these duplicated record key errors and when I checked on the indicator to handle duplicated record keys, I was given the error message "Enter a valid value". Therafter, I can't do anything at all - activate DTP, click on other tabs - and it just got stucked at this point, and I could only choose to exit the transaction.
    Can anyone advise if I have basically missed anything?
    Thank you in advance.
    Regards,
    Adelynn

    Hi,
    Handling Duplicate Data Records 
    Use
    DataSources for texts or attributes can transfer data records with the same key into BI in one request. Whether the DataSource transfers multiple data records with the same key in one request is a property of the DataSource. There may be cases in which you want to transfer multiple data records with the same key (referred to as duplicate data records below) to BI more than once within a request; this is not always an error. BI provides functions to handle duplicate data records so that you can accommodate this.
    Features
    In a dataflow that is modeled using a transformation, you can work with duplicate data records for time-dependent and time-independent attributes and texts.
    If you are updating attributes or texts from a DataSource to an InfoObject using a data transfer process (DTP), you can specify the number of data records with the same record key within a request that the system can process. In DTP maintenance on the Update tab page, you set the Handle Duplicate Record Keys indicator to specify the number of data records.
    This indicator is not set by default.
    If you set the indicator, duplicate data records (multiple records with identical key values) are handled as follows:
    ●      Time-independent data:
    If data records have the same key, the last data record in the data package is interpreted as being valid and is updated to the target.
    ●      Time-Dependent Data
    If data records have the same key, the system calculates new time intervals for the data record values. The system calculates new time intervals on the basis of the intersecting time intervals and the sequence of the data records.
    Data record 1 is valid from 01.01.2006 to 31.12.2006
    Data record 2 has the same key but is valid from 01.07.2006 to 31.12.2007
    The system corrects the time interval for data record 1 to 01.01.2006 to 30.06.2006. As of 01.07.2006, the next data record in the data package (data record 2) is valid.
    If you set the indicator for time-dependent data, note the following:
    You cannot include the data source field that contains the DATETO information in the semantic key of the DTP. This may cause duplicate data records to be sorted incorrectly and time intervals to be incorrectly calculated.
    The semantic key specifies the structure of the data packages that are read from the source.
    Example
    You have two data records with the same key within one data package.
    In the following graphic, DATETO is not an element of the key:
    In the data package, the data records are in sequence DS2, DS1. In this case, the time interval for data record 1 is corrected:
    Data record 1 is valid from 1.1.2002 to 31.12.2006.
    Data record 2 is valid from 1.1.2000 to 31.12.2001.
    In the following graphic, DATETO is an element of the key:
    If DATETO is an element of the key, the records are sorted by DATETO. In this case, the data record with the earliest date is put before the data record with the most recent date. In the data package, the data records are in sequence DS2, DS1. In this case, the time interval for data record 2 is corrected:
    Data record 2 is valid from 1.1.2000 to 31.12.2000.
    Data record 1 is valid from 1.1.2001 to 31.12.2006.
    If you do not set this indicator, data records that have the same key are written to the error stack of the DTP.
    Note
    You can specify how duplicate data records within a request are handled, independently of whether the setting that allows DataSources to deliver potentially duplicate data records has been made. This is useful if the setting was not made to the DataSource, but the system knows from other sources that duplicate data records are transferred (for example, when flat files are loaded).

  • Record filtered in advance as error records with the same key exist

    Dear Masters
       I am getting below error for master data (0MATERIAL_ATTR), while loading data from PSA to Info object through DTP (I have set the Handle duliocate record also). When I checked that error data in Error Stack, I am getting below error for all records(100)
    "Record filtered in advance as error records with the same key exist"
    Could u please help me to resolve this issue
    Thanks in Advance
    Raja

    Hi
    Thanks for reply
    I have loaded the Delta data in PSA. In PSA data was loaded successfully. I got the failure only when I loaded the from PSA to Maste data
    Thanks in advance
    Raja

  • How to limit the number of records in WebI Report

    Hi,
    I have a requirement where I need to display the WebI Report data only if the number if record is less than 65k.
    Incase if the records are more than 65k then the report should be blank and a text message needs to be displayed.
    The client dont want to set the number of records either in universe or at the edit query at the report. They want to limit once the report is executed.
    Please let me know if its possible to do it at the report level.
    Thanks in advance!!
    Best Regards
    SAP BO

    Optional prompts are not a problem.
    Assuming you first query uses the following SQL statement
    select a,b,c from t where x=y and z=@prompt(...)
    The query for checking the number of rows should use the following SQL code
    select count(*) from t where x=y and z=@prompt(...)
    The original query should be modified as follows
    select a,b,c from t where x=y and z=@prompt(...) AND (select count(*) from t where x=y and z=@prompt(...)) < 65000
    Regards,
    Stratos

  • Do Data Found Error on Inserting a Record

    I am working in Apex 4.0.
    I'm getting this error when I click the "Create" button to insert a record.
    ORA-01403: no data found
         Error      Unable to fetch row.
    Since I'm inserting rather than fetching, why would I get this error? What do I need to do to suppress it?

    Arie, thanks for the response. I suspected the automatic Row Processing might be the problem. But I did not (and still don't) know how to fix it.
    Here's what's happening. The page involved is part of an insurance tracking system. At the time of annual renewal of a policy, we create a new record for the next year. This is done by including in the new record all usuable data fields from the old expiring record. I use a button to navigate to the current page (page 22 to page 22) and set values (from the old page 22) of items in the new page 22. The user then makes a couple of manual entries and then clicks the "Create" button which appears on the new page 22 since the primary key is null. The new record is then correctly inserted into the table, and the system branches back to a calling page. It is at this time that the error occurs.
    ORA-01403: no data found
    Error Unable to fetch row.
    A database trigger gets the new primary key from a sequence generator and puts it into the record on insert. Here's the SQL on the trigger:
    CREATE OR REPLACE TRIGGER "UCC_AUTO_SEQ"
    BEFORE
    insert on "UCC"
    for each row
    begin
    select ucc_seq.nextval
    into :new.uccseq
    from dual;
    end;
    May be that I can fix this if you will explain to me what you mean by this: "My guess is that your page includes an ARF process and you didn’t set the return key item, or you didn’t use the RETURNING clause, so the ARF process can’t fetch the newly inserted record." Forgive my ignorance, but I do not know what a "RETURNING clause" is. I checked out the Source: Automatic Row Processing (DML) and found that the item labeled "Return Key Into Item" was blank. I put "P22_INSURESEQ" (the primary key item) into this field, but the error is still being thrown. The "Action Processed" message always appears in its own window just above the error message.

  • Prevent duplicate data

    I would like to make sure there are no duplicate data entries in my JSP that populates an Oracle database with a table called MainTable. MainTable has an Id field that is the primary key, ValData with a varchar data type, Fid and Fid2 are number data types. Fid and Fid2 are foreign key values that are taken from another table.
    Id   ValData   Fid   Fid2
    1    abc         34    2
    2    efg          23    34
    3    zeo         25    43Sometimes someone can enter duplicate data and the ValData, Fid and Fid2 will end up like this:
    Id   ValData   Fid   Fid2
    1    abc          34    2
    2    efg           23    34
    3    zeo          25    43
    4    zeo          25    43Is there anything in Java I can implement to prevent duplicate data entry in the above example?

    Thanks,
    It now works after I added unique constraints to each of the fields.
    Google results showed me that ORA-00001 is Oracle's Duplicate message and I added a condition in the SQLException Catch area and it catches all my Duplicate entry attempts and shows message to the user.
    Here is what I have and would like to know if this is an efficient way I am doing this??
    try
    // db stuff
    catch (SQLException sqle)
        String message = sqle.getMessage();  
        if (message.indexOf("ORA-00001") != -1)
            out.println("Duplicate Data entry.");  
        else
          out.println("some other ORA error");  
    }

  • Detecting Duplicate Data

    Application server is: Sun Java System Application Server Platform Edition 8.1.
    I am using CMP and when I call home.create() with duplicate data I get the following exception.
    javax.ejb.EJBException: Transaction aborted; nested exception is: javax.transaction.RollbackException: Transaction marked for rollback.
    javax.transaction.RollbackException: Transaction marked for rollback.
    Shouldn't I get the CreateException or DuplicateKeyException.
    I can check for duplicates by calling home.find() before home.create(). This will create an extra database hit. If I have more that one unique key on the table I will have to call multiple home.find() methods.
    Is there any other way?

    With 8.1 version you can set flush on on the corresponding create method, unless it is executed
    in a separate transaction.
    In all other cases, you do need to use find. You can construct a single query that checks
    all unique constraints.
    regards,
    -marina

  • Recording from the phone to computer

    Hey all. I am trying to use Captivate to capture a phone
    call. I have used a plug in microphone in the past, however have
    not done so trying to capture a webinar. I need to have the phone
    line run into the computer to have captiviate to record.
    Please advise.
    Thanks.

    Hi Richard
    Indeed, Camtasia may well be the better tool for this. It's
    because of the inherent way Camtasia records VS the way Captivate
    records.
    I think this is what Larry was saying. I think Larry was
    thinking that in order to record a meeting this way, one would want
    to kick off full-motion recording and record the whole movie. This
    would be a bad idea for Captivate. You would end up with a single
    slide that you couldn't edit if you did this. But I could see where
    you might record a meeting using Captivate and taking screen
    captures here and there while possibly also recording audio.
    Camtasia records a large AVI file that you may later edit
    like you can the Captivate slides.
    Cheers all... Rick

Maybe you are looking for

  • How to identify the file type in a safe way

    I need to identify the filetype of a file in my disc. I watched over internet and i've found this good site: http://filext.com/ that supplys the "magic bytes" for many kind of files. I have also looked over internet for some ready class that identify

  • TEMPLATE CHANGE OF BUSINESS BLUEPRINT-SOLAR01

    Dear Friends, Our requiremnet is to replace the Standard business blue print template with our project specific template. I did the settings as below. SAP Solution Manager Implementation Guide->SAP Solution Manager->Scenario-Specific Settings->Implem

  • Installing tomcat JSP engine

    I am installing TOMCAT so i can run JSP files on my apache webserver. Everything works fine. Installed a new version of Solaris9 installed GCC and after that Apache..... No problem I get the basic apache page.. so it is working fine. :-) Now we have

  • Lightroom serial number doesn't work on new MAC.  How can I fix this?

    I loaded the student edition of Lightroom 5 on my PC, and am trying to convert everything to a MAC.  Do I need to remove the Lightroom software from my PC to load Lightroom 5 to my MAC?  I have the student teacher edition software.

  • SetNull(2, java.sql.Types.DATE);

    If I use setNull(1, java.sql.Types.DATE); function, it will insert null on my table's column. I want to select the date from another table and insert into my table's column, what should I do? I am sure I cannot call setNull(1,java.sql.Types.DATE), it