Duplciuate Data record error while reloading a DTP

I have a DTP created to load etxt datasource and put in the process chain. I have set "delta mode" so that it picks up if additional records are available.
But, I noiticc that when there are no additional records (for eg: earlier load had 269 records and current load has same 269 records), I am gettign the error - "Duplciate data record" and status is set as "Red". I do not want to see error in this case as it did not find any new records and want the status to be green if there are no new records.
Could you please suggest what settings will do that.
Regards
Raj

Delta DTP will fetch only unloaded requests ( requests which donot exist in target ) and not additional records.
Is the text datasource delta enabled ?
Do you have Infopackage of Update type Delta setup?
Did you run a Full DTP before Delta DTP?
Assuming Full Infopackage has loaded 269 records to PSA and the same will be loaded ( if no additional records in source system ) again.
Req 1 - 269 - Yesterday
Req 2 - 269  - Today
Full DTP ran yesterday will load 269 records to target.
Delta DTP ran today will load Req1 & Req 2 to target ( master data will be overwritten ) - Reason Delta DTP is acting like Init with data transfer.
You can start off with a Delta DTP instead Full, if Full DTP is ran before a Delta DTP maje sure you delete requests loaded by Full DTP.
This can be ignored as this is Master Data which wil be overwritten.
To get rid of error jst Check " Handle Duplicate Records " in  Update Tab of DTP.

Similar Messages

  • Getting Duplicate data Records error while loading the Master data.

    Hi All,
    We are getting Duplicate data Records error while loading the Profit centre Master data. Master data contains time dependent attributes.
    the load is direct update. So i made it red and tried to reloaded from PSA even though it is throwing same error.
    I checked in PSA. Showing red which records have same Profit centre.
    Could any one give us any suggestions to resolve the issues please.
    Thanks & Regards,
    Raju

    Hi Raju,
            I assume there are no routines written in the update rules and also you ae directly loading the data from R/3 ( not from any ODS). If that is the case then it could be that, the data maintained in R/3 has overlapping time intervals (since  time dependency of attributes is involved). Check your PSA if the same profit center has time intervals which over lap. In that case, you need to get this fixed in R/3. If there are no overlapping time intervals, you can simply increase the error tolerance limit in your info-package and repeat the load.
    Hope this helps you.
    Thanks & Regards,
    Nithin Reddy.

  • BI 7.0 - Duplicate Record Error while loading master data

    I am working on BI 7.0 and I am trying to load master data to an info object.
    I created an Infopackage and loaded into PSA.
    I created transformation and DTP and I get an error after I execute the DTP about duplicate records.
    I have read all previous threads about duplicate record error while loading master data and most of the them suggested to check 'Ignore duplicate records' option in the infopackage. But in 7.0, I can only load to PSA with infopackage and it doesn't have any option for me to ignore duplicate records.
    My data is getting loaded to PSA fine and I get this error while loading to info object using DTP.
    I would appreciate your help to resolve this issue.
    Regards,
    Ram.

    Hi,
    Refer:
    http://help.sap.com/saphelp_nw2004s/helpdata/en/45/2a8430131e03c3e10000000a1553f6/frameset.htm
    http://help.sap.com/saphelp_nw2004s/helpdata/en/d0/538f3b294a7f2de10000000a11402f/frameset.htm
    With rgds,
    Anil Kumar Sharma .P

  • Error while executing a DTP

    Hi All ,
    I got  an error while executing a DTP. Short Dump analysis points this error. :   DBIF_REPO_SQL_ERROR.
    Please provide in your valuable inputs to resolve this error.
    Thank & Regards,
    Rohit Garg

    This error comes majorly when you have a deadlock in  your code. Check ST22 if you have any short dump corresponding to it and see the source code where it is getting stuck.
    If you find this dump explaining table space issue, then contact Basis, they will be able to help you with table space problem.
    Edited by: Deepmala Agarwal on Aug 18, 2009 7:40 AM

  • OSB: Cannot acquire data source error while using JCA DBAdapter in OSB

    Hi All,
    I've entered 'Cannot acquire data source' error while using JCA DBAdapter in OSB.
    Error infor are as follows:
    The invocation resulted in an error: Invoke JCA outbound service failed with application error, exception: com.bea.wli.sb.transports.jca.JCATransportException: oracle.tip.adapter.sa.api.JCABindingException: oracle.tip.adapter.sa.impl.fw.ext.org.collaxa.thirdparty.apache.wsif.WSIFException: servicebus:/WSDL/DBAdapter1/RetrievePersonService [ RetrievePersonService_ptt::RetrievePersonServiceSelect(RetrievePersonServiceSelect_inputParameters,PersonTCollection) ] - WSIF JCA Execute of operation 'RetrievePersonServiceSelect' failed due to: Could not create/access the TopLink Session.
    This session is used to connect to the datastore.
    Caused by Exception [EclipseLink-7060] (Eclipse Persistence Services - 2.0.2.v20100323-r6872): org.eclipse.persistence.exceptions.ValidationException
    Exception Description: Cannot acquire data source [jdbc/soademoDatabase].
    Internal Exception: javax.naming.NameNotFoundException: Unable to resolve 'jdbc.soademoDatabase'. Resolved 'jdbc'; remaining name 'soademoDatabase'.
    ; nested exception is:
    BINDING.JCA-11622
    Could not create/access the TopLink Session.
    This session is used to connect to the datastore.
    JNDI Name for the Database pool: eis/DB/soademoDatabase
    JNDI Name for the Data source: jdbc/soademoDatabase
    I created a basic DBAdapter in JDeveloper, got the xsd file, wsdl file, .jca file and the topLink mapping file imported them into OSB project.
    Then I used the .jca file to generate a business service, and tested, then the error occurs as described above.
    Login info in RetrievePersonService-or-mappings.xml
    <login xsi:type="database-login">
    <platform-class>org.eclipse.persistence.platform.database.oracle.Oracle9Platform</platform-class>
    <user-name></user-name>
    <connection-url></connection-url>
    </login>
    jca file content are as follows:
    <adapter-config name="RetrievePersonService" adapter="Database Adapter" wsdlLocation="RetrievePersonService.wsdl" xmlns="http://platform.integration.oracle/blocks/adapter/fw/metadata">
    <connection-factory location="eis/DB/soademoDatabase" UIConnectionName="Connection1" adapterRef=""/>
    <endpoint-interaction portType="RetrievePersonService_ptt" operation="RetrievePersonServiceSelect">
    <interaction-spec className="oracle.tip.adapter.db.DBReadInteractionSpec">
    <property name="DescriptorName" value="RetrievePersonService.PersonT"/>
    <property name="QueryName" value="RetrievePersonServiceSelect"/>
    <property name="MappingsMetaDataURL" value="RetrievePersonService-or-mappings.xml"/>
    <property name="ReturnSingleResultSet" value="false"/>
    <property name="GetActiveUnitOfWork" value="false"/>
    </interaction-spec>
    </endpoint-interaction>
    </adapter-config>
    RetrievePersonService_db.wsdl are as follows:
    <?xml version="1.0" encoding="UTF-8"?>
    <WL5G3N0:definitions name="RetrievePersonService-concrete" targetNamespace="http://xmlns.oracle.com/pcbpel/adapter/db/KnowledeMgmtSOAApplication/AdapterJDevProject/RetrievePersonService" xmlns:WL5G3N0="http://schemas.xmlsoap.org/wsdl/" xmlns:WL5G3N1="http://xmlns.oracle.com/pcbpel/adapter/db/KnowledeMgmtSOAApplication/AdapterJDevProject/RetrievePersonService" xmlns:WL5G3N2="http://schemas.xmlsoap.org/wsdl/soap/">
    <WL5G3N0:import location="RetrievePersonService.wsdl" namespace="http://xmlns.oracle.com/pcbpel/adapter/db/KnowledeMgmtSOAApplication/AdapterJDevProject/RetrievePersonService"/>
    <WL5G3N0:binding name="RetrievePersonService_ptt-binding" type="WL5G3N1:RetrievePersonService_ptt">
    <WL5G3N2:binding style="document" transport="http://www.bea.com/transport/2007/05/jca"/>
    <WL5G3N0:operation name="RetrievePersonServiceSelect">
    <WL5G3N2:operation soapAction="RetrievePersonServiceSelect"/>
    <WL5G3N0:input>
    <WL5G3N2:body use="literal"/>
    </WL5G3N0:input>
    <WL5G3N0:output>
    <WL5G3N2:body use="literal"/>
    </WL5G3N0:output>
    </WL5G3N0:operation>
    </WL5G3N0:binding>
    <WL5G3N0:service name="RetrievePersonService_ptt-bindingQSService">
    <WL5G3N0:port binding="WL5G3N1:RetrievePersonService_ptt-binding" name="RetrievePersonService_ptt-bindingQSPort">
    <WL5G3N2:address location="jca://eis/DB/soademoDatabase"/>
    </WL5G3N0:port>
    </WL5G3N0:service>
    </WL5G3N0:definitions>
    Any suggestion is appricated .
    Thanks in advance!
    Edited by: user11262117 on Jan 26, 2011 5:28 PM

    Hi Anuj,
    Thanks for your reply!
    I found that the data source is registered on server soa_server1 as follows:
    Binding Name: jdbc.soademoDatabase
    Class: weblogic.jdbc.common.internal.RmiDataSource_1033_WLStub
    Hash Code: 80328036
    toString Results: ClusterableRemoteRef(8348400613458600489S:10.2.1.143:[8001,8001,-1,-1,-1,-1,-1]:base_domain:soa_server1 [8348400613458600489S:10.2.1.143:[8001,8001,-1,-1,-1,-1,-1]:base_domain:soa_server1/291])/291
    Binding Name: jdbc.SOADataSource
    Class: weblogic.jdbc.common.internal.RmiDataSource_1033_WLStub
    Hash Code: 92966755
    toString Results: ClusterableRemoteRef(8348400613458600489S:10.2.1.143:[8001,8001,-1,-1,-1,-1,-1]:base_domain:soa_server1 [8348400613458600489S:10.2.1.143:[8001,8001,-1,-1,-1,-1,-1]:base_domain:soa_server1/285])/285
    I don't know how to determine which server the DBAdapter is targetted to.
    But I found the following information:
    Under Deoloyment->DBAdapter->Monitoring->Outbound Connection Pools
    Outbound Connection Pool Server State Current Connections Created Connections
    eis/DB/SOADemo AdminServer Running 1 1
    eis/DB/SOADemo soa_server1 Running 1 1
    eis/DB/soademoDatabase AdminServer Running 1 1
    eis/DB/soademoDatabase soa_server1 Running 1 1
    The DbAdapter is related to the following files:
    C:\ Oracle\ Middleware\ home_11gR1\ Oracle_SOA1\ soa\ connectors\ DbAdapter. rar
    C:\ Oracle\ Middleware\ home_11gR1\ Oracle_SOA1\ soa\ DBPlan\ Plan. xml
    I unzipped DbAdapter.rar, opened weblogic-ra.xml and found that there's only one data source is registered:
    <?xml version="1.0"?>
    <weblogic-connector xmlns="http://www.bea.com/ns/weblogic/90">
    <enable-global-access-to-classes>true</enable-global-access-to-classes>
    <outbound-resource-adapter>
    <default-connection-properties>
    <pool-params>
    <initial-capacity>1</initial-capacity>
    <max-capacity>1000</max-capacity>
    </pool-params>
    <properties>
    <property>
    <name>usesNativeSequencing</name>
    <value>true</value>
    </property>
    <property>
    <name>sequencePreallocationSize</name>
    <value>50</value>
    </property>
    <property>
    <name>defaultNChar</name>
    <value>false</value>
    </property>
    <property>
    <name>usesBatchWriting</name>
    <value>true</value>
    </property>
    <property>
    <name>usesSkipLocking</name>
    <value>true</value>
    </property>
    </properties>
              </default-connection-properties>
    <connection-definition-group>
    <connection-factory-interface>javax.resource.cci.ConnectionFactory</connection-factory-interface>
    <connection-instance>
    <jndi-name>eis/DB/SOADemo</jndi-name>
              <connection-properties>
                   <properties>
                   <property>
                   <name>xADataSourceName</name>
                   <value>jdbc/SOADataSource</value>
                   </property>
                   <property>
                   <name>dataSourceName</name>
                   <value></value>
                   </property>
                   <property>
                   <name>platformClassName</name>
                   <value>org.eclipse.persistence.platform.database.Oracle10Platform</value>
                   </property>
                   </properties>
              </connection-properties>
    </connection-instance>
    </connection-definition-group>
    </outbound-resource-adapter>
    </weblogic-connector>
    Then I decided to use eis/DB/SOADemo for testing.
    For JDeveloper project, after I deployed to weblogic server, it works fine.
    But for OSB project referencing wsdl, jca and mapping file from JDeveloper project, still got the same error as follows:
    BEA-380001: Invoke JCA outbound service failed with application error, exception:
    com.bea.wli.sb.transports.jca.JCATransportException: oracle.tip.adapter.sa.api.JCABindingException: oracle.tip.adapter.sa.impl.fw.ext.org.collaxa.thirdparty.apache.wsif.WSIFException: servicebus:/WSDL/DBAdapterTest/DBReader [ DBReader_ptt::DBReaderSelect(DBReaderSelect_inputParameters,PersonTCollection) ] - WSIF JCA Execute of operation 'DBReaderSelect' failed due to: Could not create/access the TopLink Session.
    This session is used to connect to the datastore.
    Caused by Exception [EclipseLink-7060] (Eclipse Persistence Services - 2.0.2.v20100323-r6872): org.eclipse.persistence.exceptions.ValidationException
    Exception Description: Cannot acquire data source [jdbc/SOADataSource].
    Internal Exception: javax.naming.NameNotFoundException: Unable to resolve 'jdbc.SOADataSource'. Resolved 'jdbc'; remaining name 'SOADataSource'.
    ; nested exception is:
    BINDING.JCA-11622
    Could not create/access the TopLink Session.
    This session is used to connect to the datastore.
    Caused by Exception [EclipseLink-7060] (Eclipse Persistence Services - 2.0.2.v20100323-r6872): org.eclipse.persistence.exceptions.ValidationException
    Exception Description: Cannot acquire data source [jdbc/SOADataSource].
    Internal Exception: javax.naming.NameNotFoundException: Unable to resolve 'jdbc.SOADataSource'. Resolved 'jdbc'; remaining name 'SOADataSource'.
    You may need to configure the connection settings in the deployment descriptor (i.e. DbAdapter.rar#META-INF/weblogic-ra.xml) and restart the server. This exception is considered not retriable, likely due to a modelling mistake.
    It almost drive me crazy!!:-(
    What's the purpose of 'weblogic-ra.xml' under the folder of 'C:\Oracle\Middleware\home_11gR1\Oracle_OSB1\lib\external\adapters\META-INF'?
    ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------

  • No data Exists error while retrieving the data from a table

    Hi Everyone!
    I am getting "No Data Exists" error while retrieving the data from a table....where i need to check...if possible please give me example link.....please help me regarding this

    Hi !
    thanks for ur response...
    I have written vo.executeQuery for the table...My page is running but i m not getting the data....I have to select two LOV's and when I click on Go button the data has to display in the table...After the selection of LOV's when I click on Go in the table "No data exists" message is appearing...the table has the data and the query is also executing...please give info where i did the mistake....

  • How to avoid 'duplicate data record' error message when loading master data

    Dear Experts
    We have a custom extractor on table CSKS called ZCOSTCENTER_ATTR. The settings of this datasource are the same as the settings of 0COSTCENTER_ATTR. The problem is that when loading to BW it seems that validity (DATEFROM and DATETO) is not taken into account. If there is a cost center with several entries having different validity, I get this duplicate data record error. There is no error when loading 0COSTCENTER_ATTR.
    Enhancing 0COSTCENTER_ATTR to have one datasource instead of two is not an option.
    I know that you can set ignore duplicates in the infopackage, but that is not a nice solution. 0COSTCENTER_ATTR can run without this!
    Is there a trick you know to tell the system that the date fields are also part of the key??
    Thank you for your help
    Peter

    Alessandro - ZCOSTCENTER_ATTR is loading 0COSTCENTER, just like 0COSTCENTER_ATTR.
    Siggi - I don't have the error message described in the note.
    "There are duplicates of the data record 2 & with the key 'NO010000122077 &' for characteristic 0COSTCENTER &."
    In PSA the records are marked red with the same message (MSG no 191).
    As you see the key does not contain the date when the record is valid. How do I add it? How is it working for 0COSTCENTER_ATTR with the same records? Is it done on the R/3 or on the BW side?
    Thanks
    Peter

  • Error while loading from DTP

    Hi,
    We are facing some issue while loading the data from R/3 to BW using DTP.
    We are getting the error (both for full and delta)as follows:
    Data package 1:Errors during Processing
    *     Extraction datasource*
    *     Filter out new records with same key.*
    *     RSDS material*
    *     Update to datastore object*
    *          Unpack data package*
    *          Exception in substep:write data package*
    *          Processing terminated.*
    *     Set technical status to red*
    *     Set overall status to red*
    *     Set overall status to green*
    *     Further processing started*
    *     Set status to u2018Processed furtheru2019*
    The expansion of the error Exception in substep: write data package:
    Record 0, segment 0001 is not in teh cross-record table
    Error while writing error stack
    Record 0, segment 0001 is not in the cross-record table
    Error in substep update to DataStore Object
    We are using both start routine and end routine in the transformation.
    Any help on this is highly appreciable.
    End routine:
        IF NOT RESULT_PACKAGE IS INITIAL.
          SELECT salesorg comp_code FROM /bi0/psalesorg
          INTO TABLE it_salesorg
          FOR ALL ENTRIES IN RESULT_PACKAGE
          WHERE salesorg = RESULT_PACKAGE-SALESORG
          AND objvers = 'A'.
          SELECT /bic/ZMATRTUGP FROM /bic/pZMATRTUGP
          INTO TABLE it_mat
          FOR ALL ENTRIES IN RESULT_PACKAGE
          WHERE /bic/ZMATRTUGP = RESULT_PACKAGE-/bic/ZMATRTUGP.
          it_data[] = RESULT_PACKAGE[].
        ENDIF.
    get company code from salesorg for all material
        LOOP AT it_data INTO st_data.
          w_tabix = sy-tabix.
          READ TABLE it_salesorg INTO st_salesorg
          WITH TABLE KEY salesorg = st_data-salesorg.
          IF sy-subrc = 0.
            st_data-comp_code = st_salesorg-compcode.
            MODIFY it_data FROM st_data INDEX w_tabix.
          ENDIF.
        ENDLOOP.
        REFRESH RESULT_PACKAGE.
        RESULT_PACKAGE[] = it_data[].
    *Finally check for each material if something has changed since the last
    *loading : if yes compare and update, otherwise no upload
        IF NOT it_mat[] IS INITIAL.
          SELECT * FROM /bic/azalomrtu00
          INTO TABLE it_check
          FOR ALL ENTRIES IN it_mat
          WHERE /bic/ZMATRTUGP = it_mat-zmaterial.
        ENDIF.
        DELETE it_check WHERE salesorg IS INITIAL.
        CLEAR st_result_package.
        LOOP AT RESULT_PACKAGE INTO st_RESULT_PACKAGE.
          READ TABLE it_check INTO st_check
          WITH TABLE KEY
          /bic/ZMATRTUGP = st_result_package-/bic/ZMATRTUGP
          CUST_GRP4 = st_result_package-CUST_GRP4
          SALESORG = st_result_package-SALESORG
          VALIDTO = st_result_package-VALIDTO.
          IF sy-subrc = 0.
    *If one of the characteristic is different, let that entry in the
    *result_package otherwise no use updating it
            IF st_check-/BIC/ZCSBRTUGP = st_result_package-/BIC/ZCSBRTUGP
            AND st_check-/BIC/ZCONDTYPE = st_result_package-/BIC/ZCONDTYPE
            AND st_check-comp_code = st_result_package-comp_code
            AND st_check-/BIC/ZCNT_RTU = st_result_package-/bic/ZCNT_RTU
            AND st_check-VALIDFROM = st_result_package-VALIDFROM.
              DELETE RESULT_PACKAGE.
            ELSE.
    *entry is new : let it updated.
            ENDIF.
            DELETE it_check WHERE
                  /bic/ZMATRTUGP = st_result_package-/bic/ZMATRTUGP
          AND CUST_GRP4 = st_result_package-CUST_GRP4
          AND SALESORG = st_result_package-SALESORG
          AND VALIDTO = st_result_package-VALIDTO.
          ENDIF.
        ENDLOOP.
    *if some entries are in the ODS but not in the datapackage, they have to
    *be deleted in the ODS since they don't exist anymore
        LOOP AT it_check INTO st_check.
          CLEAR st_result_package.
          st_result_package-/bic/ZMATRTUGP = st_check-/bic/ZMATRTUGP.
          st_result_package-CUST_GRP4 = st_check-CUST_GRP4.
          st_result_package-SALESORG = st_check-SALESORG.
          st_result_package-VALIDTO = st_check-VALIDTO.
          st_result_package-recordmode = 'D'.
          APPEND st_result_package TO RESULT_PACKAGE.
        ENDLOOP.
    Thanks in advance.
    Thanks,
    Meghana

    Hi Meghana
    HAve you got an short dump in ST22 transaction ?
    If yes, what is it ?
    In your routine, maybe you should as well add a counter for index in your loop :
    DATA : w_idx like sy-tabix.
    When you loop at result_package, do :      w_idx = sy-tabix.
    And make :     DELETE result_package INDEX w_idex.
    Hope it helps you
    Mickael

  • Runtime error while running a DTP

    I have changed a keyfigure from integer to fltp.  now i try loading the masterdata for some object which consists this keyfigure as an attribute.
    the data is loaded successfully in PSA and while runnng the dtp i get a dump with the message that "Runtime Errors         OBJECTS_TABLES_NOT_COMPATIBLE"
    "Short text:
       Two internal tables are neither compatible nor convertible."
    Please suggest what actions need to be taken to rectify this error.

    Hi Ajeet,
    Thanks for the reply.
    When i see my data source...
    the Type in dataSource shown is FLTP 8 and type from Source system is NUMBER 22 for this field
    in transformation its showing FLTP8 [source] ---> FLTP8 [target] for this particular rule.
    please let me know if i am missing on some information.

  • Getting Incorrect data type error while trying to do a CAST in table

    Getting an error while trying to compile the following piece of code
    CREATE OR REPLACE PACKAGE BODY A_pkg AS
    FUNCTION A(O_error_message IN OUT varchar2)
    RETURN BOOLEAN IS
    --Declaring the local variables and CURSORs used in the program unit
    L_attrib_tbl CFA_SQL.TYP_attrib_tbl;
    cursor c1 is
    select list_first1.number_11
    from TABLE (CAST (L_attrib_tbl AS "CFA_SQL.TYP_attrib_tbl")) list_first1;
    BEGIN
    L_group_id = '22'
    IF L_merch_type_value = 'G' OR L_merch_type_value = 'I' THEN
    CFA_SQL.QUERY_ATTRIB(L_attrib_tbl,
    L_group_id) ;
    END IF;
    open C1;
    Fetch C1 into L_number;
    close C1;
    return true;
    END A;
    END;
    Also pasting the Spec for CFA_SQL which contains TYP_attrib_tbl
    TYPE TYP_attrib_rec IS RECORD
    group_id CFA_ATTRIB_GROUP.GROUP_ID%TYPE,
    varchar2_1 VARCHAR2,
    number_11 number(10,0));
    TYPE TYP_attrib_tbl is TABLE of TYP_attrib_rec INDEX BY BINARY_INTEGER;
    The error is coming in the line
    cursor c1 is
    select list_first1.number_11
    from TABLE (CAST (L_attrib_tbl AS "CFA_SQL.TYP_attrib_tbl")) list_first1;
    with the error as Invalid data type pointing to CFA_SQL.TYP_attrib_tbl as invalid, but I have initialized L_attrib_tbl as of that datatype only. Ahy help would be greatly appreciated.
    Regards,
    Joydeep

    Hi Kelly,
    There is no data that is entered in that period.The data will be entered for that period next month.
    POV is also not set to that period,but still the problem persists.
    Thanks

  • Duplicate Data records occured while loading

    Hai Experts
    While loading Duplicate Data records occured earlier there was less no of records i used to delete the duplicate records in background in PSA. Now  i am having more records which i am confusing which to delete and which not to delete.  Since it is a flatfile loading with delta update mode. and when i went to infopackage it is showing   update subsequently in data target in hide position and i went through process chainipdisplay variant---- and made status red--- selected updated subsequently in data target
    ignore duplicate data records. and i want to trigger the subsequent process. if we go through process monitor i can rspc_process_finish. I i go display variant then what is the process.......................... is there any function module.

    Select the check box " Handle duplicate records key" in DTP on update tab.
    Thanks....
    Shambhu

  • Duplicate data records error

    Hello
    IO is updated from ODS. System generates an error : duplicate data records.
    This is because keys of ods dont match keys of infoobject.
    I can build an other ods to aggregate data before loading to infobjects.
    Are there any ways to do it in start/end routine?
    Thanks

    there is Handle duplicate recors option in DTP
    Indicator: Handling Duplicate Data Records
    Use
    If this indicator is set, duplicate data records (that is, records with the same key) are handled during an update in the order in which they occur in a data package.
    For time-independent attributes of a characteristic, the last data record with the corresponding key within a data package defines the valid attribute value for the update for a given data record key.
    Issue solved
    Thanks a lot !

  • I was getting data load error while loading profit center in BCS

    Hi,
    I was getting error message while i was loading profit center master data.The error message is the total 6 records were readed from the file 6 records contain parameter setting exceeded the user limitation of execution the data collection task"
    Just i want to load Profit center into BCS.I have create method master data upload,where i have specified fields layout as below
    Header
    Controling area
    Fiscal Year
    Posting Period
    Data Row
    Profit Cneter
    When i execute Flexible upload and method where i was getting error message saying the 6 recors are ignored
    My flat files is in this format
    Controling area  fiscal year   posting period
    2000             2008          8
    50030
    50035
    50040
    50045
    50050
    50055          
    Please some one advise me how to load it into BCS.
    Its urgent requirment and i appreciate your solutions
    Edited by: Prasad B on Oct 22, 2008 7:42 PM

    Thanks alot Dan and Eugene.
    I was following Eugene Blog for flexible update master data.I have create the way you have specififed in the document.
    I have created Method to upload Profit center and try to load the data where i am getting error.
    I am loading attributes,but not hierachy.I am not understanding why i am not able to load flat file profit center master data into BCS.
    The error is "                                                                               
    7 of 7 data rows were ignored                                                                               
    Message no. UCF7017                                                                               
    Diagnosis                                                                               
    A total of 7 data records were read from file C:\Documents and            
         Settings\Laxmikant.dube\Desktop\P. Of those, 7 records contain parameter  
         settings that exceed the user limitations for executing the data          
         collection task.                                                                               
    System response                                                                               
    The ignored records were not checked any further and will not be          
         written.                                                                               
    Procedure                                                                               
    Make sure that the ignored records are irrelevant for the current task.   
         If some of the ignored records were supposed to be written, then the      
      file probably contains an error.                                                                               
    In this case, closely examine the settings for the fields Version,       
      Fiscal Year, Period, Group Currency, and Consolidation Units in the      
      file.                                                                               
    If you have more than one characteristic with the role consolidation     
      unit (a socalled "matrix organization"), also check whether the settings 
      in the ignored records refer to valid combinations of consolidation      
      units.                                                                               
    Character string "2000,2008,8" is cut off after 4 characters                                                                               
    Message no. UCF7003                                                                               
    Diagnosis                                                                               
    The current data row contains for characteristic Controlling area the     
         value 2000,2008,8. But the maximum length for values of this              
         characteristic is only 4 characters.                                                                               
    System response                                                                               
    The character string "2000,2008,8" is cut off after 4 characters.                                                                               
    Procedure                                                                               
    Check if the string "2000,2008,8" is indeed supposed to be interpreted    
         as the value of characteristic Controlling area. If this is the case,     
         avoid exceeding the maximum length for this characteristic.                                                                               
    However, if there has been a mix-up, compare the structure of the file 
    with the definition in the upload method. If variable column widths are
    used, particularly pay attention to the number of field separators in  
    the data row involved.                                                 
    Please advise me on this.It would be great help

  • Date parse error while importing users from OIM to OIA (SRM 5.0.3)

    Hi All,
    Env Details:
    OIA (SRM 5.0.3), Weblogic and Oracle 10g DB
    We have integrated OIM to OIA with extended attributes mapping by modifying iam-context.xml file to load users. Its done successfully. But when we map "Date" related attribute, its giving "Date Parsing error" and its not loading the users.
    We have tried loading users using flatFile mechanism, its also giving same result.
    Please suggest me. Thanks in Advance !!!
    Regards,
    Ravi G.

    Hi,
    Its a problem with OOB's OIMIAMSolution.class file, which is called while importing users from OIM. It used DateParse () conversion method only for all attributes which OIA attributes' name is ends with "Date". It defined, the conversion of date from (yyyy-MM-dd). So its expecting the input value should be in defined format(yyyy-MM-dd), if not, it gives a parse error.
    We found work around for this as follows,
    We have used other related OIA attribute which name ends other than "Date" string.
    Thanks,
    Ravi G.

  • Data access error while running  the script in fdm workbench script editor

    Hello Experts,
    I have been getting the error when i run the script in script editor of fdqm workbench client.this is the error
                              -2147467259 Data access error and it is navigating the error at line 30 about this script "Set rsAppend = DW.DataAccess.farsTable(strWorkTableName)"
    Please revert me asap where i should load the data from peoplesoft to one of my planning database through fdm.
    Thanks in advance...

    Hi,
    Thanks for your suggestion's here i am going to run the datapump script so while i am running in workbench it is throwing data access error.So as you guy's suggested i created one import format and added this script to that import format which is added to one of my location even it is throwing the same error.I am getting the error at this sentence   "strWorkTableName" I want to know what is this where it reflect's?So my question is Is there any difference while running data pump or integration script.Please let me know a detailed navigation steps to run this script's with out data access error to run these kind of scripts.
    Note:Again my target is to load the data to planning database through fdqm which i am doing through sql statement's.
    Suggest your valuable suggestion's
    Thanks in advance...
    I am getting the error at this sentence   "strWorkTableName" I want to know what is this where it reflect's?

Maybe you are looking for

  • Toggle on/off color correction for a/b comparison

    How is this done?

  • Reporting - Filters

    Hello Experts When we copy the report from the existing reports. In the filters if the company code for the past report exist, then do we have to change the filter - company code in  the present report. PLease help. Thank you Swetha

  • Add customized task to user screen for IdM 7.0

    The user screen in IdM 7.0 is a lot more different from the previous versions. A navigation bar is used for launching different tasks(workflows). I want to add a new user request task to MenuBar Requests or add a new task to the Launch Requests page.

  • Installation Succeeded but Patch Failed. Download Error Support (49)

    Installation Succeeded but Patch Failed. Download Error contact Support (49)

  • Multiple updates on a table

    Hi, My DB is 10.0.2 I have a batch program, which fires multiple updates on a single table (without commit) on different conditions. For Ex: update tablea set end_date=sysdate-1 where trans_id=2010; update tablea set end_date=sysdate where trans_id=2