Issue in loading data from data mart

Actual load is from APO to BW is happening, but load gets failed when it is loaded from base cube to target cube (data mart), this flow is in 3.5.
Attached document contains the system msg.

Your load is still in process.
Where load failure mesasge?
Try to activate source 8DSO/Cube and later actiavte update rules thru prorgams.
Can you check your job details from SM37.
From info pack monitor, copy request id.
Go SM37,Job name - <enter copied request id>.
Use job status - Active/cancel/finished.
Use proper data ranges.
User - *

Similar Messages

  • Special character issue while loading data from SAP HR through VDS

    Hello,
    We have a special character issue, while loading data from SAP HR to IdM, using a VDS and following the standard documentation: http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/e09fa547-f7c9-2b10-3d9e-da93fd15dca1?quicklink=index&overridelayout=true
    French accent like (é,à,è,ù), are correctly loaded but Turkish special ones (like : Ş, İ, ł ) are transformed into u201C#u201D in Idm.
    The question is : does someone know any special setting to do in the VDS or in IdM for special characters to solve this issue??
    Our SAP HR version is ECC6.0 (ABA/BASIS7.0 SP21, SAP_HR6.0 SP54) and we are using a VDS 7.1 SP5 and SAP NW IdM 7.1 SP5 Patch1 on oracle 10.2.
    Thanks

    We are importing directly to the HR staging area, using the transactions/programs "HRLDAP_MAP", "LDAP" and "/RPLDAP_EXTRACT", then we have a job which extract data from the staging area to a CSV file.
    So before the import, the character appears correctly in SAP HR, but by the time it comes through the VDS to the IDM's temporary table, it becomes "#".
    Yes, our data is coming from a Unicode system.
    So, could it be a java parameter to change / add in the VDS??
    Regards.

  • Issue to load data from database to XML file. error: ODI-40768

    Hi,
    While I am trying to load data from database to XML following error is appearing.
    ODI-1228: Task TEST_XML_DATA (Integration) fails on the target XML connection MyLOCALXSD.
    Caused By: java.sql.SQLException: ODI-40768: Could not save the file <default>:C:\DATA_FILE\www.xml because a class java.io.IOException occurred and said: The filename, directory name, or volume label syntax is incorrect
         at com.sunopsis.jdbc.driver.xml.SnpsXmlFile.writeToFile(SnpsXmlFile.java:751)
         at com.sunopsis.jdbc.driver.xml.SnpsXmlConnection.internalExecute(SnpsXmlConnection.java:769)
         at com.sunopsis.jdbc.driver.xml.SnpsXmlPreparedStatement.execute(SnpsXmlPreparedStatement.java:46)
         at oracle.odi.runtime.agent.execution.sql.SQLCommand.execute(SQLCommand.java:166)
         at oracle.odi.runtime.agent.execution.sql.SQLExecutor.execute(SQLExecutor.java:102)
         at oracle.odi.runtime.agent.execution.sql.SQLExecutor.execute(SQLExecutor.java:1)
         at oracle.odi.runtime.agent.execution.TaskExecutionHandler.handleTask(TaskExecutionHandler.java:50)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.processTask(SnpSessTaskSql.java:2906)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2609)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:540)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:453)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:1740)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$2.doAction(StartSessRequestProcessor.java:338)
         at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:214)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:272)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$0(StartSessRequestProcessor.java:263)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$StartSessTask.doExecute(StartSessRequestProcessor.java:822)
         at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:123)
         at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$2.run(DefaultAgentTaskExecutor.java:83)
         at java.lang.Thread.run(Thread.java:662)
    jdbc url: jdbc:snps:xml?d=C:\test_xsd.xsd&s=TESTSQL&re=employeesX&ro=true
    physical schema: TESTSQL
    Knowledge modules are:
    LKM: LKM SQL to SQL
    IKM: IKM XML Control Append
    CKM: CKM SQL
    Parameters set are:
    CREATE_XML_FILE: true
    XML_PATH:<default>:C:\DATA_FILE\www.xml
    I have tried to change the directory path but error is still there.
    Could some one please help me to resolve the issue?
    Regards,
    user1672911

    Hi,
    The trouble in "<default>:"  - if you set XML_PATH as C:\DATA_FILE\www.xml instead  <default>:C:\DATA_FILE\www.xml- it will work correctly.
    Greetings,
    Eugene

  • TT Issue "9994 Loading data store from disk into RAM in progress"

    Getting this strange issue. Sometimes (2-3 times a day) TT just becomes unavailable and when trying to connect to it getting "9994 Loading data store from disk into RAM in progress".
    Any advise on what the root cause for this behavior might be?

    In all likelihood this is exactly what it sounds like - your datastore is being reloaded from disk into memory. Connections can't be completed until the datastore has gone through recovery and is fully in-memory. You need to look back through your ttmesg.log files and find out why the datastore was taken out of memory in the first place. Possibly there's been a crash and it has been invalidated, meaning it needs to be recovered from the checkpoint files and transaction logs. Or maybe there were no user connections and it was taken out of memory and checkpointed to disk in a normal, controlled fashion. The ttmesg.log files will tell you what happened.

  • Issue while loading data from DSO to Cube

    Gurus,
    I'm facing a typical problem while loading Cube from a DSO. The load is based upon certain conditions.
    Though the data in DSO satisfies those condition, but the Cube is still not being loaded.
    All the loads are managed through Process Chains.
    Would there be any reason in specific/particular for this type of problem?
    Any pointers would be of greatest help !
    Regards,
    Yaseen & Soujanya

    Yaseen & Soujanya,
    It is very hard to guess the problem with the amount of information you have provided.
    - What do you mean by cube is not being loaded? No records are extracted from DSO? Is the load completing with 0 records?
    - How is data loaded from DSO to InfoCube? Full? Are you first loading to PSA or not?
    - Is there data already in the InfoCube?
    - Is there change log data for DSO or did someone delete all the PSA data?
    Sincere there are so many reasons for the behavior you are witnessing, your best option is to approach your resident expert.
    Good luck.
    Sudhi Karkada
    <a href="http://main.nationalmssociety.org/site/TR/Bike/TXHBikeEvents?px=5888378&pg=personal&fr_id=10222">Biking for MS Relief</a>

  • Issue while loading data to sample essbase app using odi

    while executing data load the error is
    Now, while loading data to an essbase app i am getting the following error:
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 23, in ?
    com.hyperion.odi.essbase.ODIEssbaseException: Error records reached the maximum error threshold : 1
    at com.hyperion.odi.essbase.ODIEssbaseDataWriter.loadData(Unknown Source)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
    at java.lang.reflect.Method.invoke(Unknown Source)
    at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java)
    at org.python.core.PyMethod.__call__(PyMethod.java)
    at org.python.core.PyObject.__call__(PyObject.java)
    at org.python.core.PyInstance.invoke(PyInstance.java)
    at org.python.pycode._pyx4.f$0(<string>:23)
    at org.python.pycode._pyx4.call_function(<string>)
    at org.python.core.PyTableCode.call(PyTableCode.java)
    at org.python.core.PyCode.call(PyCode.java)
    at org.python.core.Py.runCode(Py.java)
    at org.python.core.Py.exec(Py.java)
    at org.python.util.PythonInterpreter.exec(PythonInterpreter.java)
    at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:144)
    at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
    at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
    at com.sunopsis.dwg.cmd.e.i(e.java)
    at com.sunopsis.dwg.cmd.g.y(g.java)
    at com.sunopsis.dwg.cmd.e.run(e.java)
    at java.lang.Thread.run(Unknown Source)
    Caused by: com.hyperion.odi.essbase.ODIEssbaseException: Error records reached the maximum error threshold : 1
    at com.hyperion.odi.essbase.ODIEssbaseDataWriter.sendRecordArrayToEsbase(Unknown Source)
    ... 32 more
    com.hyperion.odi.essbase.ODIEssbaseException: com.hyperion.odi.essbase.ODIEssbaseException: Error records reached the maximum error threshold : 1
    at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
    at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
    at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
    at com.sunopsis.dwg.cmd.e.i(e.java)
    at com.sunopsis.dwg.cmd.g.y(g.java)
    at com.sunopsis.dwg.cmd.e.run(e.java)
    at java.lang.Thread.run(Unknown Source)

    Hi,
    It means in your options of the KM you have it set to quit when it hits one error, if you set it to 0 (infinite) then it will not stop no matter how many data load errors it hits.
    If you set it to 0 and run the interface, depending on how you set up the options in the KM it can write to two log files, which you should check.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Issues while loading data to the Essbase.

    Hi All,,
    I'm trying to load data to essbase and i'm getting the following error.It says unknown source.But when i do the reverse it is working fine without any errors.
    I'm doing a simple data load from the file..Please help me.
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 23, in ?
    com.hyperion.odi.essbase.ODIEssbaseException: Error records reached the maximum error threshold : 1
         at com.hyperion.odi.essbase.ODIEssbaseDataWriter.loadData(Unknown Source)
         at sun.reflect.GeneratedMethodAccessor100.invoke(Unknown Source)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
         at java.lang.reflect.Method.invoke(Unknown Source)
         at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java)
         at org.python.core.PyMethod.__call__(PyMethod.java)
         at org.python.core.PyObject.__call__(PyObject.java)
         at org.python.core.PyInstance.invoke(PyInstance.java)
         at org.python.pycode._pyx125.f$0(<string>:23)
         at org.python.pycode._pyx125.call_function(<string>)
         at org.python.core.PyTableCode.call(PyTableCode.java)
         at org.python.core.PyCode.call(PyCode.java)
         at org.python.core.Py.runCode(Py.java)
         at org.python.core.Py.exec(Py.java)
         at org.python.util.PythonInterpreter.exec(PythonInterpreter.java)
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:144)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.j(e.java)
         at com.sunopsis.dwg.cmd.h.z(h.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)
    Caused by: com.hyperion.odi.essbase.ODIEssbaseException: Error records reached the maximum error threshold : 1
         at com.hyperion.odi.essbase.ODIEssbaseDataWriter.checkMaxError(Unknown Source)
         at com.hyperion.odi.essbase.ODIEssbaseDataWriter.sendRecordArrayToEsbase(Unknown Source)
         ... 31 more
    com.hyperion.odi.essbase.ODIEssbaseException: com.hyperion.odi.essbase.ODIEssbaseException: Error records reached the maximum error threshold : 1
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.j(e.java)
         at com.sunopsis.dwg.cmd.h.z(h.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)
    Thanks
    K

    Sorry John,,
    I setup the log file and i see the following in my log file..it says that cannot connect to the essbase server...
    How ever the user id which is setup in the physical schema essbase server has full rights to essbase server and the application.
    2009-09-30 12:03:32,561 INFO [DwgCmdExecutionThread]: ODI Hyperion Essbase Adapter Version 9.3.1.1
    2009-09-30 12:03:32,561 INFO [DwgCmdExecutionThread]: Connecting to Essbase application [Test] on [Server]:[1423] using username [hypadmin].
    2009-09-30 12:03:33,311 INFO [DwgCmdExecutionThread]: Successfully connected to the Essbase application.
    2009-09-30 12:03:33,311 INFO [DwgCmdExecutionThread]: Essbase Load IKM option RULES_FILE = AMORT3
    2009-09-30 12:03:33,311 INFO [DwgCmdExecutionThread]: Essbase Load IKM option RULE_SEPARATOR = |
    2009-09-30 12:03:33,311 INFO [DwgCmdExecutionThread]: Essbase Load IKM option PRE_LOAD_MAXL_SCRIPT =
    2009-09-30 12:03:33,311 INFO [DwgCmdExecutionThread]: Essbase Load IKM option POST_LOAD_MAXL_SCRIPT =
    2009-09-30 12:03:33,327 INFO [DwgCmdExecutionThread]: Essbase Load IKM option ABORT_ON_PRE_MAXL_ERROR = true
    2009-09-30 12:03:33,327 INFO [DwgCmdExecutionThread]: Essbase Load IKM option CLEAR_DATABASE = None
    2009-09-30 12:03:33,327 INFO [DwgCmdExecutionThread]: Essbase Load IKM option COMMIT_INTERVAL = 1000
    2009-09-30 12:03:33,327 INFO [DwgCmdExecutionThread]: Essbase Load IKM option CALCULATION_SCRIPT = null
    2009-09-30 12:03:33,327 INFO [DwgCmdExecutionThread]: Essbase Load IKM option RUN_CALC_SCRIPT_ONLY = false
    2009-09-30 12:03:33,358 DEBUG [DwgCmdExecutionThread]: LoadData Begins
    _2009-09-30 12:03:33,358 DEBUG [DwgCmdExecutionThread]: Error occured in sending record chunk...Cannot begin data load. Analytic Server Error(1042015): Network error: Cannot Locate Connect Information For [server]_
    2009-09-30 12:03:33,374 DEBUG [DwgCmdExecutionThread]: Sending data record by record to essbase
    2009-09-30 12:03:33,374 INFO [DwgCmdExecutionThread]: Logging out and disconnecting from the essbase application.
    K
    Edited by: kanna143 on Sep 30, 2009 10:35 AM

  • Issue with loading data into cube with duplicate members in different dimensions

    Have a cube with allowed dublicate members only in two dims (Period1 and Period2).
    Outlines saves with no errors, but when i try to load data, i get some errors:
    "Member 2012-12 is a duplicate member in the outline
    A6179398-68BD-7843-E0C2-B5EE81808D0B    01011    cd00    st01    2905110000    EK    fo0000    NNNN-NN-NN    2012-12    cust00000$$$    1" 
    These dims represent two similar periods of time. Am I to change member names and aliases a little in one of them(f.e. yyyy1-mm --> yyyy1 - mm1)?
    Users wouldnt like it...
    Period1
      yyyy1
        yyyy1/q1
          yyyy1-mm1
          yyyy1-mm2
          yyyy1-mm3   
        yyyy1/q2
    Period2
      yyyy1
        yyyy1/q1
          yyyy1-mm1
            yyyy1-mm1-dd1 
            yyyy1-mm1-dd2
          yyyy1-mm2
          yyyy1-mm3   
        yyyy1/q2
    Tnanx

    You may have to use fully qualified name something like
    [Period1].[2012-12]
    [Period2].[2012-12]
    You can refer to ESSBASE admin guide on "Creating and Working With Duplicate Member Outlines"
    Regards,
    Sunil

  • Issue in loading images from Apache webserver in JSF image tags on WLS 10

    I have configured the apache webserver2.0 for passing on the request to weblogic 10.0. I have images in apache webserver. In my application on weblogic i am using JSF f. In JSF for <h:graphics> tag i am not able to load images from apache webserver. issue we found out that it prepend a context url on it.
              pls provide if u have any solution for it.
              Edited by neerajr at 11/16/2007 3:33 PM

    Hi, Neeraj,
              Are you actually telling <h:graphicImage>? As per html_basic.tld, the "uri" attribute is context-relative.
              Using "value" attribute instead of "url" attribute may be helpful. Something like
              <h:graphicImage id="fooImage" value="http://www.foosite.com/image/foo.jpg"/>
              This should work on WLS10.0 if you are using the JSF library bundled in it.
              Thanks,
              -Fred.

  • Issues with Loading Images from a Jar File

    This code snippet basically loops through a jar of gifs and loads them into a hashmap to be used later. The images all load into the HashMap just fine, I tested and made sure their widths and heights were changing as well as the buffer size from gif to gif. The problem comes in when some of the images are loaded to be painted they are incomplete it looks as though part of the image came through but not all of it, while other images look just fine. The old way in which we loaded the graphics didn't involve getting them from a jar file. My question is, is this a common problem with loading images from a jar from an applet? For a while I had tried to approach the problem by getting the URL of the image in a jar and passing that into the toolkit and creating the image that way, I was unsuccessful in getting that to work.
    //app is the Japplet
    MediaTracker tracker = new MediaTracker(app);
    //jf represents the jar file obj, enum for looping through jar entries
    Enumeration e = jf.entries();
    Toolkit toolkit = Toolkit.getDefaultToolkit();
    //buffer for reading image stream
    byte buffer [];
    while(e.hasMoreElements())
    fileName = e.nextElement().toString();
    InputStream inputstream = jf.getInputStream(jf.getEntry(fileName));
    buffer = new byte[inputstream.available()];
    inputstream.read(buffer);
    currentIm = toolkit.createImage(buffer);
    tracker.addImage(currentIm, 0);
    tracker.waitForAll();
    images.put(fileName.substring(0, fileName.indexOf(".")), currentIm);
    } //while
    }//try
    catch(Exception e)
    e.printStackTrace();
    }

    compressed files are not the problem. It is just the problem of the read not returning all the bytes. Here is a working implementation:
    InputStream is = jar.getInputStream(entry);
    ByteArrayOutputStream os = new ByteArrayOutputStream(1024);
    try{
    byte[] buf = new byte[1024];
    int read;
    while((read = is.read(buf)) > 0) {
    os.write(buf, 0, read);
    catch(Exception e){
         e.printStackTrace();
         return null;
    image = Toolkit.getDefaultToolkit().createImage(os.toByteArray());
    This works but I think you end up opening the jar a second time and downloading it from the server again. Another way of getting the images is using the class loader:
    InputStream is = MyApplet.class.getResourceAsStream(strImageName);
    In this case, the image file needs to be at the same level than MyApplet.class but you don't get the benefit of enumerating of the images available in the jar.

  • Semantic Partitioning Delta issue while load data from DSO to Cube -BW 7.3

    Hi All,
    We have created the semantic partion with the help of BADI to perform Partitions. Everthing looks good,
    first time i have loaded the data it is full update.
    Second time i init the data load. I pulled the delta  from DSO to Cube . DSO is standard where as Cube is partition with the help of semantic partition. What i can see is the records which are updated in Latest delta are shown into the report rest all are ignored by the system.
    I tried compression but still it did not worked.
    Can some has face this kind
    Thanks

    Yaseen & Soujanya,
    It is very hard to guess the problem with the amount of information you have provided.
    - What do you mean by cube is not being loaded? No records are extracted from DSO? Is the load completing with 0 records?
    - How is data loaded from DSO to InfoCube? Full? Are you first loading to PSA or not?
    - Is there data already in the InfoCube?
    - Is there change log data for DSO or did someone delete all the PSA data?
    Sincere there are so many reasons for the behavior you are witnessing, your best option is to approach your resident expert.
    Good luck.
    Sudhi Karkada
    <a href="http://main.nationalmssociety.org/site/TR/Bike/TXHBikeEvents?px=5888378&pg=personal&fr_id=10222">Biking for MS Relief</a>

  • Issue in loading data to BI

    Hi,
    I will list the steps that I used to transfer data from ECC to BI.
    I created a characteristic in my infoobject catalog. With master data/attributes.
    I then created a generic datasource for master data attributes in ECC and replicated it in BI. The datasource is a view in ECC.
    Then I created an infopackage for the datasource that was replicated in BI.
    and when I clicked on start.
    I get the following.
    Request still running
    Diagnosis
    No errors found. The current process has probably not finished yet.
    System Response
    The ALE inbox of BI is identical to the ALE outbox of the source system
    or
    the maximum wait time for this request has not yet been exceeded
    or
    the background job has not yet finished in the source system.
    Current status
    No Idocs arrived from the source system.
    I am in the dev system and the data is only a few hundred records.
    Alternatively,
    I also created a transformation for my attribute under the characteristic and a data transfer process. I get no error. But no data found is returned.
    Can anyone please suggest what am I doing wrong here.
    Thanks,
    Abdullah Ismail.

    Hi,
    sounds to me, as if the User which connects BI and Source system is disabled in the source system, or SAP Basis in the source system thought that BI only uses a background (batch) User. Take a look on the parameters of the "Source System" .
    In the connection parameters check the tab "Logon & Security". The user mentioned there has to be a dialog user in the source system (!). It has to be set to a password which never expires (otherwise after expiration time BI-Source connection will be lost).
    If the Basis management of the source system has locked the user, or the user is changed to batch, or the passwort expiration time is set, you have to convince them, that these are mandatory requirements to keep BI and Source system connected.
    Kind regards,
    Jürgen
    Edited by: Jürgen Kirsch on Oct 20, 2009 4:05 PM

  • Short dump issue while loading data

    Hello All,
    We are trying to load the data to  the cube , but every time it is throwing us a short dump with the below error.
    We have found a SAP Note as well  for the below error , Please suggest, how to implement the same.
    Error:
    Runtime Errors         CONVT_NO_NUMBER
    METHOD end_routine.
    === Segments ===
       FIELD-SYMBOLS:
         <RESULT_FIELDS>    TYPE _ty_s_TG_1.
       DATA:
         MONITOR_REC     TYPE rstmonitor.
       ... "insert your code here
    --  fill table "MONITOR" with values of structure "MONITOR_REC"
    -   to make monitor entries
       ... "to cancel the update process
        raise exception type CX_RSROUT_ABORT.
    *Fiscal year quarter addition
       loop at RESULT_PACKAGE assigning <RESULT_FIELDS>.
         if <RESULT_FIELDS>-/bic/zsp_Mon BETWEEN 1 and 3.
           CONCATENATE 'FY' <RESULT_FIELDS>-fiscper+2(2) 'Q3' INTO
           <RESULT_FIELDS>-/bic/zsp_FYQA.
         elseif <RESULT_FIELDS>-/bic/zsp_Mon BETWEEN 4 and 6.
           CONCATENATE 'FY' <RESULT_FIELDS>-fiscper+2(2) 'Q4' INTO
           <RESULT_FIELDS>-/bic/zsp_FYQA.
         elseif <RESULT_FIELDS>-/bic/zsp_Mon BETWEEN 7 and 9.
           CONCATENATE 'FY' <RESULT_FIELDS>-fiscper+2(2) 'Q1' INTO
           <RESULT_FIELDS>-/bic/zsp_FYQA.
         elseif <RESULT_FIELDS>-/bic/zsp_Mon BETWEEN 10 and 12.
           CONCATENATE 'FY' <RESULT_FIELDS>-fiscper+2(2) 'Q2' INTO
           <RESULT_FIELDS>-/bic/zsp_FYQA.
      endif.
    endloop.
    ENDDMETHOD.                    "end_routine
    THOD inverse_end_routine.
    IMPORTING

    Hi,
    As suggested by note, you need to put this code in end routine.
    Check the syntax , if error go through it and fix syntax (take help of ABAPer if required) and
    activate Transformation and DTP.
    Thank-You.
    Regards,
    VB

  • Issue in loading data into OLAP dim/cube.

    All,
    Facing this issue while executing the dimension in OLAP.
    Warning:-_
    O/*RA-20101: 17:39:12 ***Error Occured in BUILD_DRIVER: In __XML_SEQUENTIAL_LOADER: In __XML_LOAD_DIMVALS: Error loading dimension values for DIM_SST.DIMENSION, level DIM_SST.SOL_LEVEL.LEVEL, hierarchy DIM_SST.AW$NONE.HIERARCHY, mapping group DIM_SST.SOL_LEVEL.MAPGROUP1.DIMENSIONMAPGROUP. In __XML_LOAD_DIMVALS_ITEM: In ___XML_LOAD_TEMPPRG: All dimension values must be single line values.
    TRUNCATE_LOAD=false
    AW Execution status: Success
    17:39:12 Started Build(Refresh) of MIS Analytic Workspace.
    17:39:12 Attached AW MIS in RW Mode.
    17:39:12 Started Loading Dimensions.
    17:39:12 Started Loading Dimension Members.
    17:39:12 Started Loading Dimension Members for DIM_SST.DIMENSION (1 out of 1 Dimensions).
    17:39:12 Failed to Build(Refresh) MIS Analytic Workspace.
    17:39:12 ***Error Occured in BUILD_DRIVER: In __XML_SEQUENTIAL_LOADER: In __XML_LOAD_DIMVALS: Error loading dimension values for DIM_SST.DIMENSION, level DIM_SST.SOL_LEVEL.LEVEL, hierarchy DIM_SST.AW$NONE.HIERARCHY, mapping group DIM_SST.SOL_LEVEL.MAPGROUP1.DIMENSIONMAPGROUP. In __XML_LOAD_DIMVALS_ITEM: In ___XML_LOAD_TEMPPRG: All dimension values must be single line values.

    Hi there,
    What version of 10g OLAP are you using?
    It can be very important to apply the latest patches where possible - more details can be found here:- http://www.oracle.com/technology/products/bi/olap/collateral/olap_certification.html
    Thanks,
    Stuart Bunby
    OLAP Blog: http://oracleOLAP.blogspot.com
    OLAP Wiki: http://wiki.oracle.com/page/Oracle+OLAP+Option
    OLAP on OTN: http://www.oracle.com/technology/products/bi/olap/index.html
    DW on OTN : http://www.oracle.com/technology/products/bi/db/11g/index.html

  • After upgrade to BI NW4.0S issue in loading cube from ODS

    Hi,
      I have an SAP delivered ODS(0CRM_OPPT_H) which I was loading for some time. The cube 0CRM_C04 was loaded from the ODS a couple of times. I have now deleted all the data from the cube and want to load the data from the cube to ODS, In between our BW was upgrade frpm 3.5 to BI NW0S..
      I have done all prerequisite , replicated the 80 datsource , deleted and re activated the update rule and transfer rule again.. and then tried to load the Cube from ODS.. however everytime I am getting the following error
    DataSource 80CRM_OPPI does not have the same status as the source system in the Business Information Warehouse.
    The time stamp in the source system is 12/23/2005 10:09:36.
    The time stamp in the BW system is 11/09/2005 13:02:29.
    Is it due to the upgrade?? I have done everything so that this dosen't happen, still could not resolve..
    Thanks

    Are you sure you've replicated the DataSource 80CRM_OPPI? Try replicating it individually again. If that doesn't help, have you tried reactivating the ODS Object itself?
    Regards, Klaus

Maybe you are looking for