Problem loading data from jena

Hi, two issues when loading data into Oracle from a jena model:
1. The incremental and batch load both works well except when we add a triple with a literal types as double:
triple = new Triple(dirNode.asNode(), Node.createURI("http://www.w3.org/2003/01/geo/wgs84_pos#long"), Node.createLiteral(geopos.getLongitude().toString(), null,      (RDFDatatype) XSDDatatype.XSDdouble));
graph.add(triple);
We get the error:
GRAVE: Could not add triple
java.sql.BatchUpdateException: ORA-55303: Fallo en el constructor SDO_RDF_TRIPLE_S: Simple case: SQLERRM=ORA-55328: fallo al intentar insertar el valor literal "-5.9278863"^^<http://www.w3.org/2001/XMLSchema#double>
ORA-06512: en "MDSYS.MD", línea 1723
ORA-06512: en "MDSYS.MDERR", línea 17
ORA-06512: en "MDSYS.SDO_RDF_TRIPLE_S", línea 211
     at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1335)
     at oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3449)
     at oracle.jdbc.driver.OraclePreparedStatement.executeUpdate(OraclePreparedStatement.java:3530)
     at oracle.jdbc.driver.OraclePreparedStatementWrapper.executeUpdate(OraclePreparedStatementWrapper.java:1062)
2. The bulk load simply does not work:
((OracleBulkUpdateHandler) graph.getBulkUpdateHandler()).addInBulk(GraphUtil.findAll(model.getGraph()), "sem_ts");
We get:
01-oct-2009 13:11:39 oracle.spatial.rdf.client.jena.SimpleLog warn
ADVERTENCIA: addInBulk: [92 ] sqle
java.sql.SQLException: ORA-44004: nombre de SQL cualificado no válido
ORA-06512: en "SYS.DBMS_ASSERT", línea 188
ORA-06512: en "MDSYS.SDO_RDF", línea 242
ORA-06512: en "MDSYS.RDF_APIS", línea 693
ORA-06512: en línea 1
     at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:439)
     at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:395)
     at oracle.jdbc.driver.T4C8Oall.processError(T4C8Oall.java:802)
In both case our conexion is something like:
public static String conexion = "jdbc:oracle:thin:user/pass@ourserver:1521:ourdb";
Any idea? Thanks

Hi Wu, we have included your code in a java test ang got the same problem.
Our installation is Oracle Database 11.2.0.1.0. Then we added the 'Semantic patch' 11_2_sem and followed the instructions to create a tablespace and the RDF/SEM network. Finally we create a model as [explained here|http://download.oracle.com/docs/cd/E11882_01/appdev.112/e11828/sdo_rdf_concepts.htm#CHDEDFFA] .
Some text in the exception is spanish; basically it seems to say 'fails inserting the literal value'... The rest of data in the app has been correctly inserted.
This is the java test:
public class PruebaOracleTest extends TestCase {
     String jdbcUrl = "jdbc:oracle:thin:user/pass@server:1521:bd";
     public void testInsertData() throws Exception {
          Oracle oracle = new Oracle(jdbcUrl, null, null);
          GraphOracleSem graph = new GraphOracleSem(oracle, "ARTICLES");
          ModelOracleSem model = new ModelOracleSem(graph);
          Model inMemoryJenaModel = ModelFactory.createDefaultModel();
          long lStartTime = System.currentTimeMillis();
          System.out.println("testCustomerMisc: start");
          Triple t = new Triple(Node.createURI("http://sub"), Node
                    .createURI("http://www.w3.org/2003/01/geo/wgs84_pos#long"),
                    Node.createLiteral("-5.9278863", null,
                              (RDFDatatype) XSDDatatype.XSDdouble));
          graph.add(t);
          graph.flushAdd();
          String queryString = "SELECT * " + "WHERE { "
                    + " ?subject ?predicate ?object . " + "} ";
          Query query = QueryFactory.create(queryString);
          QueryExecution qexec = QueryExecutionFactory.create(query, model);
          ResultSet results;
               results = qexec.execSelect();
               ResultSetFormatter.out(System.out, results, query);
     public void testListTriples() throws Exception {
          Oracle oracle = new Oracle(jdbcUrl, null, null);
          GraphOracleSem graph = new GraphOracleSem(oracle, "ARTICLES");
          int cont = 0;
          ExtendedIterator it = graph.find(Triple.ANY);
          while (it.hasNext() && cont<100) {
               Triple t = (Triple) it.next();
               System.out.println(t.toString());
               cont++;
          graph.close();
          oracle.dispose();
     public void testCleanModel() throws Exception {
          Oracle oracle = new Oracle(jdbcUrl, null, null);
          GraphOracleSem graph = new GraphOracleSem(oracle, "ARTICLES");
          ModelOracleSem model = new ModelOracleSem(graph);
          model.removeAll();
          graph.close();
          oracle.dispose();
And this the exception we get:
java.sql.SQLException: ORA-55303: Fallo en el constructor SDO_RDF_TRIPLE_S: Simple case: SQLERRM=ORA-55328: fallo al intentar insertar el valor literal "-5.9278863"^^<http://www.w3.org/2001/XMLSchema#double>
ORA-06512: en "MDSYS.MD", línea 1723
ORA-06512: en "MDSYS.MDERR", línea 17
ORA-06512: en "MDSYS.SDO_RDF_TRIPLE_S", línea 211
     at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:439)
     at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:395)
     at oracle.jdbc.driver.T4C8Oall.processError(T4C8Oall.java:802)
     at oracle.jdbc.driver.T4CTTIfun.receive(T4CTTIfun.java:436)
     at oracle.jdbc.driver.T4CTTIfun.doRPC(T4CTTIfun.java:186)
     at oracle.jdbc.driver.T4C8Oall.doOALL(T4C8Oall.java:521)
     at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:205)
     at oracle.jdbc.driver.T4CPreparedStatement.executeForRows(T4CPreparedStatement.java:1008)
     at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1307)
     at oracle.jdbc.driver.OraclePreparedStatement.sendBatch(OraclePreparedStatement.java:3753)
     at oracle.jdbc.driver.OraclePreparedStatementWrapper.sendBatch(OraclePreparedStatementWrapper.java:1140)
     at oracle.spatial.rdf.client.jena.GraphOracleSem.flushAdd(GraphOracleSem.java:1219)
     at org.fundacionctic.ogd.data.support.PruebaOracleTest.testInsertData(PruebaOracleTest.java:42)
     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
     at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
     at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
     at java.lang.reflect.Method.invoke(Method.java:597)
     at junit.framework.TestCase.runTest(TestCase.java:154)
     at junit.framework.TestCase.runBare(TestCase.java:127)
     at junit.framework.TestResult$1.protect(TestResult.java:106)
     at junit.framework.TestResult.runProtected(TestResult.java:124)
     at junit.framework.TestResult.run(TestResult.java:109)
     at junit.framework.TestCase.run(TestCase.java:118)
     at org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
     at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
     at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:460)
     at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:673)
     at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:386)
     at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:196)

Similar Messages

  • Problem loading data from the PSA to the InfoCube

    Hello experts.
    I'm having a problem loading data from the PSA to the InfoCube.
    I'm using a DTP for this process but is happening the following error:
    "Diagnosis
          An error occurred while executing the transformation rule:
          The exact error message is:
          Overflow converting from''
          The error was triggered at the point in the Following Program:
          GP4KMDU7EAUOSBIZVE233WNLPIG 718
      System Response
          Processing the record date has Been terminated.
    Procedure
          The Following is additional information included in the higher-level
         node of the monitor:
         Transformation ID
         Data record number of the source record
         Number and the name of the rule Which produced the error
    Procedure for System Administration
    Have already created new DTP's deactivate and reactivate the InfoCube, the transformation, but solves nothing.
    Does anyone have any idea what to do?
    Thank you.

    HI,
    Is it a flat file load or loading frm any data source?
    try to execute the program GP4KMDU7EAUOSBIZVE233WNLPIG 718 in Se38 and check if its active and no syntax errors are there.
    Check the mapping of the fileds in transformations weather
    some data fileds are mapped to decimal or char 32 filed is mapped to Raw 16
    or calweek, calmonth mapped to calday etc.
    Check in St22 if there any short dumps ..
    Regards
    KP

  • PROBLEM LOADING DATA FROM A TEXT FILE.

    Hi,
    Im having a problem in loading my csv file to the database. Im using Oracle Database 10g for Linux. Im in p. 228 in the book. This is my csv file look.
    db_name     db_version     host_id
    db10     9.2.0.7     1
    db11     10.2.0.1     1
    db12     10.2.0.1     1
    db13     9.2.0.7     1
    db14     10.2.0.1     1
    db15     9.2.0.7     1
    I loaded this data to an existing table called DATABASES loaded from tab delimited. FILE CHARACTER SET is UNICODE UTF-8. Then I browsed the name of the csv file to be uploaded. It looked like this.
    File Name      F23757437/db2.csv     Reupload File      
    Separator      
    Optionally Enclosed By      
         First row contains column names.
    File Character Set      
    I CLICKED NEXT, THIS IS WHAT IT LOOKED LIKE.
    Schema:      HANDSONXE06
    Table Name:      DATABASES
    Define Column Mapping      
    Column Names     %
    Format     
    Upload     yes
    Row 1     "db10" "9.2.0.7" 1
    Row 2     "db11" "10.2.0.1" 1
    Row 3     "db12" "10.2.0.1" 1
    Row 4     "db13" "9.2.0.7" 1
    Row 5     "db14" "10.2.0.1" 1
    Row 6     "db15" "9.2.0.7" 1
    I CLICKED LOAD AND THIS WAS THE RESULT.
    * There are NOT NULL columns in HANDSONXE06.DATABASES. Select to upload the data without an error.
    Schema
    Down
    Table Name
    Down
    File Details
    Down
    Column Mapping
    Load Data      
    Schema:      HANDSONXE06
    Table Name:      DATABASES
    Define Column Mapping      
    Column Names     COLUMN_NAMES
    Format     FORMAT
    Upload     UPLOAD
    Row 1     "db10" "9.2.0.7" 1
    Row 2     "db11" "10.2.0.1" 1
    Row 3     "db12" "10.2.0.1" 1
    Row 4     "db13" "9.2.0.7" 1
    Row 5     "db14" "10.2.0.1" 1
    Row 6     "db15" "9.2.0.7" 1
    I WAS REALLY WONDERING WHAT WAS REALLY WRONG. AN ERROR MESSAGE SAID, THERE ARE NOT NULL COLUMNS IN THE HANDSONXE06.DATABASES. I DIDN'T KNOW HOW TO FIX IT. WHAT DO I NEED TO CHANGE TO LOAD THE DATA WITHOUT AN ERROR? IT REALLY CONFUSED ME A LOT AND HOW COME I HAVE AN ERROR? PLEASE HELP ME. I NEED AND ANSWER TO MY PROBLEM PLEASE. I CANNOT GO FORWARD BECAUSE OF THIS.
    THANKS,
    JOCELYN

    I'm not certain of the utility you are using to load the data, however, I completed the following test using SQL Loader to insert the data into my table. Your process should work similar if the trigger and sequence are created for the table you are loading.
    SQL> create table load_tbl
      2  (db_id number(3) not null,
      3   db_name varchar2(100) not null,
      4   db_version varchar2(25),
      5   host_id number(3) not null)
      6  /
    Table created.
    SQL> desc load_tbl
    Name                                      Null?    Type
    DB_ID                                     NOT NULL NUMBER(3)
    DB_NAME                                   NOT NULL VARCHAR2(100)
    DB_VERSION                                         VARCHAR2(25)
    HOST_ID                                   NOT NULL NUMBER(3)
    SQL> create sequence db_id_seq;
    Sequence created.
    SQL> create or replace trigger db_id_trig
      2  before insert on load_tbl
      3  for each row
      4  when (new.db_id is null)
      5  begin
      6    select db_id_seq.nextval into :new.db_id from dual;
      7  end;
      8  /
    Trigger created.
    The contents of the data file, control file and log file are below for the load into load_tbl.
    C:\>sqlldr userid=username/password@db control=db_id_load.ctl log=db_id_load.log
    SQL*Loader: Release 9.2.0.6.0 - Production on Thu Jan 18 17:21:47 2007
    Copyright (c) 1982, 2002, Oracle Corporation.  All rights reserved.
    Commit point reached - logical record count 6
    C:\>
    SQL> select * from load_tbl
      2  /
         DB_ID DB_NAME              DB_VERSION                   HOST_ID
             1 db10                 9.2.0.7                            1
             2 db11                 10.2.0.1                           1
             3 db12                 10.2.0.1                           1
             4 db13                 9.2.0.7                            1
             5 db14                 10.2.0.1                           1
             6 db15                 9.2.0.7                            1
    6 rows selected.
    SQL>
    Data File"db10" "9.2.0.7" 1
    "db11" "10.2.0.1" 1
    "db12" "10.2.0.1" 1
    "db13" "9.2.0.7" 1
    "db14" "10.2.0.1" 1
    "db15" "9.2.0.7" 1
    Control FileLOAD DATA
    INFILE "C:\db_id_load.dat"
    APPEND INTO TABLE load_tbl
    FIELDS TERMINATED BY WHITESPACE OPTIONALLY ENCLOSED BY '"'
    TRAILING NULLCOLS
    (db_name CHAR,
    db_version CHAR,
    host_id "TO_NUMBER(:host_id,'99999999999')"
    Log FileSQL*Loader: Release 9.2.0.6.0 - Production on Thu Jan 18 17:21:47 2007
    Copyright (c) 1982, 2002, Oracle Corporation.  All rights reserved.
    Control File:   db_id_load.ctl
    Data File:      C:\db_id_load.dat
      Bad File:     db_id_load.bad
      Discard File:  none specified
    (Allow all discards)
    Number to load: ALL
    Number to skip: 0
    Errors allowed: 50
    Bind array:     64 rows, maximum of 256000 bytes
    Continuation:    none specified
    Path used:      Conventional
    Table LOAD_TBL, loaded from every logical record.
    Insert option in effect for this table: APPEND
    TRAILING NULLCOLS option in effect
       Column Name                  Position   Len  Term Encl Datatype
    DB_NAME                             FIRST     *  WHT O(") CHARACTER           
    DB_VERSION                           NEXT     *  WHT O(") CHARACTER           
    HOST_ID                              NEXT     *  WHT O(") CHARACTER           
        SQL string for column : "TO_NUMBER(:host_id,'99999999999')"
    Table LOAD_TBL:
      6 Rows successfully loaded.
      0 Rows not loaded due to data errors.
      0 Rows not loaded because all WHEN clauses were failed.
      0 Rows not loaded because all fields were null.
    Space allocated for bind array:                  49536 bytes(64 rows)
    Read   buffer bytes: 1048576
    Total logical records skipped:          0
    Total logical records read:             6
    Total logical records rejected:         0
    Total logical records discarded:        0
    Run began on Thu Jan 18 17:21:47 2007
    Run ended on Thu Jan 18 17:21:47 2007
    Elapsed time was:     00:00:00.39
    CPU time was:         00:00:00.13

  • Problem in Loading Data from SQL Server 2000 to Oracle 10g

    Hi All,
    I am a university student and using ODI for my final project on real-time data warehousing. I am trying to load data from MS SQL Server 2000 into Oracle 10g target table. Everything goes fine until I execute the interface for the initial load. When I choose the CKM Oracle(Create unique index on the I$ table) km, the following step fails:
    21 - Integration - Prj_Dim_RegionInterface - Create Unique Index on flow table
    Where Prj_Dim_Region is the name of my target table in Oracle.
    The error message is:
    955 : 42000 : java.sql.SQLException: ORA-00955: name is already used by an existing object
    java.sql.SQLException: ORA-00955: name is already used by an existing object
         at oracle.jdbc.driver.DatabaseError.throwSqlException(DatabaseError.java:125)
         at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:316)
         at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:282)
         at oracle.jdbc.driver.T4C8Oall.receive(T4C8Oall.java:639)
         at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:185)
         at oracle.jdbc.driver.T4CPreparedStatement.execute_for_rows(T4CPreparedStatement.java:633)
         at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1086)
         at oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:2984)
         at oracle.jdbc.driver.OraclePreparedStatement.executeUpdate(OraclePreparedStatement.java:3057)
         at com.sunopsis.sql.SnpsQuery.executeUpdate(SnpsQuery.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execStdOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.i(e.java)
         at com.sunopsis.dwg.cmd.g.y(g.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)
    I am using a surrogate key column in my target table alongwith the natural key. The natural key is populated by the primary key of my source table, but for the surrogate key, I have created a sequence in my oracle schema where the target table exists and have used the following code for mapping:
    <%=snpRef.getObjectName( "L" , "SQ_PRJ_DIM_REGION" , "D" )%>.nextval
    I have chosen to execute this code on target.
    Among my attempts to solve this problem was to set Create Index option of the CKM Oracle(Create Index for the I$ Table) to No so that it wont create any index on the flow table. I also tried to use the simple CKM Oracle km . Both solutions allowed the interface to execute successfully without any errors, but the data was not loaded into the target table.
    When I right-click on the Prj_Dim_Region data store and choose Data, it shows empty. Pressing the SQL button in this data store shows a dialog box " New Query" where I see this query:
    select * from NOVELTYFURNITUREDW.PRJ_DIM_REGION
    But when i press OK to run it, I get this error message:
    java.lang.IllegalArgumentException: Row index out of range
         at javax.swing.JTable.boundRow(Unknown Source)
         at javax.swing.JTable.setRowSelectionInterval(Unknown Source)
         at com.borland.dbswing.JdbTable.accessChange(JdbTable.java:2959)
         at com.borland.dx.dataset.AccessEvent.dispatch(Unknown Source)
         at com.borland.jb.util.EventMulticaster.dispatch(Unknown Source)
         at com.borland.dx.dataset.DataSet.a(Unknown Source)
         at com.borland.dx.dataset.DataSet.a(Unknown Source)
         at com.borland.dx.dataset.DataSet.a(Unknown Source)
         at com.borland.dx.dataset.DataSet.open(Unknown Source)
         at com.borland.dx.dataset.StorageDataSet.refresh(Unknown Source)
         at com.borland.dx.sql.dataset.QueryDataSet.refresh(Unknown Source)
         at com.borland.dx.sql.dataset.QueryDataSet.executeQuery(Unknown Source)
         at com.sunopsis.graphical.frame.a.cg.actionPerformed(cg.java)
         at javax.swing.AbstractButton.fireActionPerformed(Unknown Source)
         at javax.swing.AbstractButton$ForwardActionEvents.actionPerformed(Unknown Source)
         at javax.swing.DefaultButtonModel.fireActionPerformed(Unknown Source)
         at javax.swing.DefaultButtonModel.setPressed(Unknown Source)
         at javax.swing.plaf.basic.BasicButtonListener.mouseReleased(Unknown Source)
         at java.awt.AWTEventMulticaster.mouseReleased(Unknown Source)
         at java.awt.Component.processMouseEvent(Unknown Source)
         at java.awt.Component.processEvent(Unknown Source)
         at java.awt.Container.processEvent(Unknown Source)
         at java.awt.Component.dispatchEventImpl(Unknown Source)
         at java.awt.Container.dispatchEventImpl(Unknown Source)
         at java.awt.Component.dispatchEvent(Unknown Source)
         at java.awt.LightweightDispatcher.retargetMouseEvent(Unknown Source)
         at java.awt.LightweightDispatcher.processMouseEvent(Unknown Source)
         at java.awt.LightweightDispatcher.dispatchEvent(Unknown Source)
         at java.awt.Container.dispatchEventImpl(Unknown Source)
         at java.awt.Window.dispatchEventImpl(Unknown Source)
         at java.awt.Component.dispatchEvent(Unknown Source)
         at java.awt.EventQueue.dispatchEvent(Unknown Source)
         at java.awt.EventDispatchThread.pumpOneEventForHierarchy(Unknown Source)
         at java.awt.EventDispatchThread.pumpEventsForHierarchy(Unknown Source)
         at java.awt.EventDispatchThread.pumpEvents(Unknown Source)
         at java.awt.EventDispatchThread.pumpEvents(Unknown Source)
         at java.awt.EventDispatchThread.run(Unknown Source)
    I do not understand what the problem is and wasting days to figure it out. Any help will be highly appreciated as my deadline is too close for this project.
    Thank you so much in advance.
    Neel

    Hi Cezar,
    Can u plz help me with this scenario?
    I have one Oracle data model with 19 source tables and one SQL Server data model with 10 target tables. I have created 10 interfaces which use JournalizedDataOnly on one of the tables in the interface; e.g in interface for DimCustomer target table, I have 2 tables, namely Customer and Address, but the journalizing filter appear only on Customer table and this option is disabled for Address automatically.
    Now I want to create a package using OdiWaitForLog event detection. Is it possible to put all these 10 interfaces in just one package to populate the target tables? It works fine when I have only one interface and I use the name of one table in the interface for Table Name parameter of OdiWaitForLogData event, but when I try a comma seperated list of table names[Customer, Address] this error happens
    java.sql.SQLException: ORA-00942: table or view does not exist
    and if I use this method <%=odiRef.getObjectName("L","model_code","logical_schema","D")%>, I get this error
    "-CDC_SET_NAME=Exception getObjectName("L", "model_code", "logical_schema", "D") : SnpLSchema.getLSchemaByName : SnpLschema does not exist" "
    Please let me know how to make it work?
    Do I need to create separate data models each including only those tables which appear in their corresponding interface and package? Or do I need to create multiple packages each with only one journalized interface to populate only one target table?
    Thank you for your time in advance.
    Regards,
    Neel

  • Problem while loading data from ODS to infoobject

    Hi guys,
    I am facing problem while loading data from <b>ODS to infoobject</b>.
    If I load data via PSA it works fine but
    if I load data without PSA its giving error as Duplicate records.
    Do u have any idea why it is so.
    Thanks in advance
    savio

    Hi,
    when you load the data via the PSA, what did you select? Serial or Paralel?
    If you select serial most likely you don't have duplicates within the same datapackage and your load can go through.
    Loading directly in the IObj will happen thefore if you have the same key in two different packages, the "duplicate records" will be raised; you can perhaps flag your IPack with the option "ignore duplicate records" as suggested...
    hope this helps...
    Olivier.

  • Problem in Loading data from R/3 to ODS

    Whn i load data from r3 - tables ,I could not get the data to the data target ODS.
    What is the problem ? can any one help this out .

    Hi Lakshmanan,
    Important point : It is always good to detail your issue to get a prompt response.
    There could be lot of issues.
    1. Find the datasource which is feeding this ODS and
          Goto R/3 system -> RSA3 and check whether any records are populated, If no records are populated then there is issue with extracting the records..
    2. if records are populating in the first step..on the BW Side, check the PSA table..and if you find any records in PSA then this is issue on BW Side..and there could be many issues..
    May be if you can provide more information..we will of help to you...
    Regards, Vj

  • Problem in loading data from one DSO to another

    Hi All,
       I am using Multicube 0PM_MP03, this mutli cube use DSO 0PM_DS02 & 0PM_DS01, data will flow from PSA to 0PM_DS01 & then to 0PM_DS02 using a Update rule, i have triggered my infopackage & i can see data in 0PM_DS01 & i have activated the data in 0PM_DS01, now on right click on 0PM_DS01 i click on the option  Additional Functions -> Update 3.x Data to Target, now the request status is green in 0PM_DS02 but i cannot see any data in this DSO, neither in new or active, can anyone help me out in this where i am doing wrong.
    Thanks,
    Pritam.

    Hi Pritam,
    This is strange, Check the below points.
    1. If the number of records added 0, in manage of DSO2?
    2. Confirm if there is a transfprmation between ODS1 & ODS2. If there is, you need to use a DTP (not Additional Functions -> Update 3.x Data to Target).
    3. Activate the request and then check.
    4. Maintain PSA for the datasource 80PM_DS01 and when loading data from DSO1 to DSO2 choose option PSA and then into targets.
    Regards,
    Panksj.

  • Problem in loading data from one DSO to onother DSO in BI7

    Hi
    Iam loading data from one DSO to onother, After executing DTP Iam getting error message as "Aggregation NOP not allowed  on DB (key figure 0DSCT_DAYS1)     ", Here 0DSCT_DAYS1 is the keyfigure aggragation type is "NOP" , exception aggregation is " No aggregation(X if more than one value unique to 0 occurs)"
    Please kindly give me solution .
    Radha Krishna

    We had this issue in our project. It mainly arises when you have written start routine in transformation. At start routine level it can not aggregate the fields for which NOP is set & that's why it throws an error.
    Please check the load by removing the start routine. If it works then introduce a data target in between which has exactly same structure as previous data target except the fields with NOP aggregation.

  • Problem in loading data from flat file to info object

    Hi experts,
      I am new on bw. I am trying to load data from .csv file to info object. CSV file contains two colums Mat No, Mat Name and are separated by comma and similarly Mat Name is attribute of Mat No in info object. I am doing all steps by following the book but whenever I load data into info object it is loaded in one column. I mean both Mat no and Mat name apears in Mat No column whereas Mat Name column appears empty in info object. I have changed exel file into .csv. Any one can help me sort it out.
    regards

    Hi,
    your flatfile should be in the form of csv ie it does not mean that in a word u seperate it with coma
    create an excel file and say save as in teh file name give the name of your flatfile and in the save as type select CSV(comma delimited) from the drop down list
    and then in this file enter the data as u fill in excel sheet.
    After closing the file right click it and say open with wordpad then the data should appear seperated with a comma.
    Try this and then revert back in case of any doubts.........
    Are u loading data into BI 7.0 or 3.x
    Regards
    Madhavi

  • ERROR :  "Exceptions in Substep: Rules" while loading data from DSO to CUBE

    Hi All,
    I have a scenario like this.
    I was loading data from ODS to cube daily. everything was working fine till 2  days before. i added some characteristic fields in my cube while i didn't do any change in my key figures. now i started reloading of data from ODS to cubes. i got the following error  "Exceptions in Substep: Rules" after some package of data loaded successfully.
    I deleted the added fields from the infocube and then i tried toreload data again but i am facing the same error again  "Exceptions in Substep: Rules".
    Can any one tell me how to come out from this situation ?
    Regards,
    Komik Shah

    Dear,
    Check some related SDN posts:
    zDSo Loading Problem
    Re: Regarding Data load in cube
    Re: Regarding Data load in cube
    Re: Error : Exceptions in Substep: Rules
    Regards,
    Syed Hussain.

  • DTP load data from DSO to CUBE can't debug

    Hi,all expert.
    I met a problem yestoday.
    When I load data from datasource to DSO,and debug the start routine,it was ok.But when I load data from DSO to CUBE,and debug the routine,it got an error like this:
    inconsistent input parameter(parameter:<unknow>,value <unknown>)
    message no: RS_EXCEPTION 101.
    I don't know why,PLZ help me,thank you very much.

    Hi,
    An alternative way to do this is to have a small code for infinite loop in your start routine. And when you start the load goto SM50 and debug the process which is executing the load, from there you can close the infinite loop and carry on with the debugging. Please let me know if you require further information.
    Thanks,
    Arminder

  • Error when loading data from DSO to Cube

    Hi,
    After upgrading to 2004s we get the following error message when trying to load data from a DSO to a Cube:
    Unexpected error: RSDRC_CONVERT_RANGES_TO_SELDR
    Has anyone experienced a similar problem or have some guidelines on how to resolve it?
    Kind regards,
    Fredrik

    Hej Martin,
    Tack!
    Problem solved
    //Fredrik

  • Loading Data from R/3 to 3.5 AND 7.0

    Hi,
    We have a BW 3.5 that has been extracting Sales data from R/3 for the past few years. We now need to load the same data into a new BI 7.0. I saw that we can load data from one R/3 system to multiple BW systems through note 775568. But the question is, how will we INIT for a new BI system when the data extraction has already been happening for a long time. The other option would be move historical data from 3.5 TO 7.0 But is there any way for loading data from R/3 into the new BI.?
    Thanks in advance.
    Cheers
    AN

    Hi,
    It is possible to have multiple BW/BI systems for one R3/ECC systems. In order for you to load data into both 3.5 and 7.0 systems, you need to difine BI 7.0 logical system in R3/ECC just like your BW 3.5 logical system in R3/ECC. In R3/ECC, the outbound queue (SMQ1) differentiates the deltas based on the logical system name that you have defined (3.5 and 7.0) in R3/ECC system. So, in RSA7 you will see same delta queue name twice but differentiated by logical system name. So you will not have problems loading deltas. Hope this will shed some light.
    Regards,
    Rao.

  • How to load data from a ODS to CUBE Request ID - by - Request ID?

    <i>How to load data from a ODS to CUBE Request ID - by - Request ID?</i>
    The problem is that... some requests had been eliminated of the cube and the delta control between the ODS and CUBE was lost. The flag "<b>data mart status of request</b>" of all the requests of the ODS had been blank.
    Now it is necessary to load some requests from the ODS for the cube.
    Notes:
    - it is not possible to make a complete load selecting the data to be loaded;
    - the PSA is not being used;
    - considering the data volume it is impracticable to reload the cube completely.
    Thanks in advance,
    Wesley.

    Dear R B,
    Considering the following:
    -> the delta control was lost;
    -> the data already are active in the ODS;
    -> part of the data of the ODS already is in the cube.
    The indicated procedure it only guarantees the load of the data that are in the ODS and that are not in the cube.
    Tks,
    Wesley.

  • Load data from an infocube into other infocube

    Hi guys....i try to load data from an infocube to other infocube.....i generated the infosource for export without problems i can visulize the datasource in my BW System assigned well...but how i can run the extraction from one infocube into the other?
    I view that in ODSs there's an option to run load of data into datatargets but in the infocube i can´t find this option......
    is there anything else to do?
    Regards

    The Infosurce and Datasource was linked automatically i can see this in Source System Myself but i can´t see it  in Infosources Tab and i can´t create a Infopackage....that´s my problem

Maybe you are looking for

  • SQL *Loader and External Table

    Hi, Can anyone tell me the difference between SQL* Loader and External table? What are the conditions under we can use SQL * Loader and External Table. Thanx

  • De activating / un registering cs2 from my mac

    Hi all Hope someone can help here. . . My friend installed cs2 on my mac and im the second user for his cd. He now wants me to delete it off my mac as he wants to use it himself and has asked me to unregister it. I have dragged everything from applic

  • Upgrade to Yosemite 10.10.3 bluetooth keyboard not working

    after the latest upgrade to Yosemite 10.10.3 some keys on my bluetooth keyboard are not working, any idea how to fix this? Example my "a" key does nothing, I can't get a ? and when I hold down shift to get a capital letter (example S) I get a lower c

  • Key combination to choose startup disk

    My Windows partition is hosed. How do I get back to a screen that lets me choose my start-up disk? I can't remember the key or key combination. Thank you, Nancy

  • Closing GPlayer

    Have been trying to download flash player update and I continue to get a prompt telling me to close "gplayer." What is it and how do I close it so I can install the update?