Issue in loading data to BI

Hi,
I will list the steps that I used to transfer data from ECC to BI.
I created a characteristic in my infoobject catalog. With master data/attributes.
I then created a generic datasource for master data attributes in ECC and replicated it in BI. The datasource is a view in ECC.
Then I created an infopackage for the datasource that was replicated in BI.
and when I clicked on start.
I get the following.
Request still running
Diagnosis
No errors found. The current process has probably not finished yet.
System Response
The ALE inbox of BI is identical to the ALE outbox of the source system
or
the maximum wait time for this request has not yet been exceeded
or
the background job has not yet finished in the source system.
Current status
No Idocs arrived from the source system.
I am in the dev system and the data is only a few hundred records.
Alternatively,
I also created a transformation for my attribute under the characteristic and a data transfer process. I get no error. But no data found is returned.
Can anyone please suggest what am I doing wrong here.
Thanks,
Abdullah Ismail.

Hi,
sounds to me, as if the User which connects BI and Source system is disabled in the source system, or SAP Basis in the source system thought that BI only uses a background (batch) User. Take a look on the parameters of the "Source System" .
In the connection parameters check the tab "Logon & Security". The user mentioned there has to be a dialog user in the source system (!). It has to be set to a password which never expires (otherwise after expiration time BI-Source connection will be lost).
If the Basis management of the source system has locked the user, or the user is changed to batch, or the passwort expiration time is set, you have to convince them, that these are mandatory requirements to keep BI and Source system connected.
Kind regards,
Jürgen
Edited by: Jürgen Kirsch on Oct 20, 2009 4:05 PM

Similar Messages

  • Special character issue while loading data from SAP HR through VDS

    Hello,
    We have a special character issue, while loading data from SAP HR to IdM, using a VDS and following the standard documentation: http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/e09fa547-f7c9-2b10-3d9e-da93fd15dca1?quicklink=index&overridelayout=true
    French accent like (é,à,è,ù), are correctly loaded but Turkish special ones (like : Ş, İ, ł ) are transformed into u201C#u201D in Idm.
    The question is : does someone know any special setting to do in the VDS or in IdM for special characters to solve this issue??
    Our SAP HR version is ECC6.0 (ABA/BASIS7.0 SP21, SAP_HR6.0 SP54) and we are using a VDS 7.1 SP5 and SAP NW IdM 7.1 SP5 Patch1 on oracle 10.2.
    Thanks

    We are importing directly to the HR staging area, using the transactions/programs "HRLDAP_MAP", "LDAP" and "/RPLDAP_EXTRACT", then we have a job which extract data from the staging area to a CSV file.
    So before the import, the character appears correctly in SAP HR, but by the time it comes through the VDS to the IDM's temporary table, it becomes "#".
    Yes, our data is coming from a Unicode system.
    So, could it be a java parameter to change / add in the VDS??
    Regards.

  • Issue to load data from database to XML file. error: ODI-40768

    Hi,
    While I am trying to load data from database to XML following error is appearing.
    ODI-1228: Task TEST_XML_DATA (Integration) fails on the target XML connection MyLOCALXSD.
    Caused By: java.sql.SQLException: ODI-40768: Could not save the file <default>:C:\DATA_FILE\www.xml because a class java.io.IOException occurred and said: The filename, directory name, or volume label syntax is incorrect
         at com.sunopsis.jdbc.driver.xml.SnpsXmlFile.writeToFile(SnpsXmlFile.java:751)
         at com.sunopsis.jdbc.driver.xml.SnpsXmlConnection.internalExecute(SnpsXmlConnection.java:769)
         at com.sunopsis.jdbc.driver.xml.SnpsXmlPreparedStatement.execute(SnpsXmlPreparedStatement.java:46)
         at oracle.odi.runtime.agent.execution.sql.SQLCommand.execute(SQLCommand.java:166)
         at oracle.odi.runtime.agent.execution.sql.SQLExecutor.execute(SQLExecutor.java:102)
         at oracle.odi.runtime.agent.execution.sql.SQLExecutor.execute(SQLExecutor.java:1)
         at oracle.odi.runtime.agent.execution.TaskExecutionHandler.handleTask(TaskExecutionHandler.java:50)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.processTask(SnpSessTaskSql.java:2906)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2609)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:540)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:453)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:1740)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$2.doAction(StartSessRequestProcessor.java:338)
         at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:214)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:272)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$0(StartSessRequestProcessor.java:263)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$StartSessTask.doExecute(StartSessRequestProcessor.java:822)
         at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:123)
         at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$2.run(DefaultAgentTaskExecutor.java:83)
         at java.lang.Thread.run(Thread.java:662)
    jdbc url: jdbc:snps:xml?d=C:\test_xsd.xsd&s=TESTSQL&re=employeesX&ro=true
    physical schema: TESTSQL
    Knowledge modules are:
    LKM: LKM SQL to SQL
    IKM: IKM XML Control Append
    CKM: CKM SQL
    Parameters set are:
    CREATE_XML_FILE: true
    XML_PATH:<default>:C:\DATA_FILE\www.xml
    I have tried to change the directory path but error is still there.
    Could some one please help me to resolve the issue?
    Regards,
    user1672911

    Hi,
    The trouble in "<default>:"  - if you set XML_PATH as C:\DATA_FILE\www.xml instead  <default>:C:\DATA_FILE\www.xml- it will work correctly.
    Greetings,
    Eugene

  • TT Issue "9994 Loading data store from disk into RAM in progress"

    Getting this strange issue. Sometimes (2-3 times a day) TT just becomes unavailable and when trying to connect to it getting "9994 Loading data store from disk into RAM in progress".
    Any advise on what the root cause for this behavior might be?

    In all likelihood this is exactly what it sounds like - your datastore is being reloaded from disk into memory. Connections can't be completed until the datastore has gone through recovery and is fully in-memory. You need to look back through your ttmesg.log files and find out why the datastore was taken out of memory in the first place. Possibly there's been a crash and it has been invalidated, meaning it needs to be recovered from the checkpoint files and transaction logs. Or maybe there were no user connections and it was taken out of memory and checkpointed to disk in a normal, controlled fashion. The ttmesg.log files will tell you what happened.

  • Issue while loading data to sample essbase app using odi

    while executing data load the error is
    Now, while loading data to an essbase app i am getting the following error:
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 23, in ?
    com.hyperion.odi.essbase.ODIEssbaseException: Error records reached the maximum error threshold : 1
    at com.hyperion.odi.essbase.ODIEssbaseDataWriter.loadData(Unknown Source)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
    at java.lang.reflect.Method.invoke(Unknown Source)
    at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java)
    at org.python.core.PyMethod.__call__(PyMethod.java)
    at org.python.core.PyObject.__call__(PyObject.java)
    at org.python.core.PyInstance.invoke(PyInstance.java)
    at org.python.pycode._pyx4.f$0(<string>:23)
    at org.python.pycode._pyx4.call_function(<string>)
    at org.python.core.PyTableCode.call(PyTableCode.java)
    at org.python.core.PyCode.call(PyCode.java)
    at org.python.core.Py.runCode(Py.java)
    at org.python.core.Py.exec(Py.java)
    at org.python.util.PythonInterpreter.exec(PythonInterpreter.java)
    at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:144)
    at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
    at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
    at com.sunopsis.dwg.cmd.e.i(e.java)
    at com.sunopsis.dwg.cmd.g.y(g.java)
    at com.sunopsis.dwg.cmd.e.run(e.java)
    at java.lang.Thread.run(Unknown Source)
    Caused by: com.hyperion.odi.essbase.ODIEssbaseException: Error records reached the maximum error threshold : 1
    at com.hyperion.odi.essbase.ODIEssbaseDataWriter.sendRecordArrayToEsbase(Unknown Source)
    ... 32 more
    com.hyperion.odi.essbase.ODIEssbaseException: com.hyperion.odi.essbase.ODIEssbaseException: Error records reached the maximum error threshold : 1
    at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
    at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
    at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
    at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
    at com.sunopsis.dwg.cmd.e.i(e.java)
    at com.sunopsis.dwg.cmd.g.y(g.java)
    at com.sunopsis.dwg.cmd.e.run(e.java)
    at java.lang.Thread.run(Unknown Source)

    Hi,
    It means in your options of the KM you have it set to quit when it hits one error, if you set it to 0 (infinite) then it will not stop no matter how many data load errors it hits.
    If you set it to 0 and run the interface, depending on how you set up the options in the KM it can write to two log files, which you should check.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Issues while loading data to the Essbase.

    Hi All,,
    I'm trying to load data to essbase and i'm getting the following error.It says unknown source.But when i do the reverse it is working fine without any errors.
    I'm doing a simple data load from the file..Please help me.
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (innermost last):
    File "<string>", line 23, in ?
    com.hyperion.odi.essbase.ODIEssbaseException: Error records reached the maximum error threshold : 1
         at com.hyperion.odi.essbase.ODIEssbaseDataWriter.loadData(Unknown Source)
         at sun.reflect.GeneratedMethodAccessor100.invoke(Unknown Source)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
         at java.lang.reflect.Method.invoke(Unknown Source)
         at org.python.core.PyReflectedFunction.__call__(PyReflectedFunction.java)
         at org.python.core.PyMethod.__call__(PyMethod.java)
         at org.python.core.PyObject.__call__(PyObject.java)
         at org.python.core.PyInstance.invoke(PyInstance.java)
         at org.python.pycode._pyx125.f$0(<string>:23)
         at org.python.pycode._pyx125.call_function(<string>)
         at org.python.core.PyTableCode.call(PyTableCode.java)
         at org.python.core.PyCode.call(PyCode.java)
         at org.python.core.Py.runCode(Py.java)
         at org.python.core.Py.exec(Py.java)
         at org.python.util.PythonInterpreter.exec(PythonInterpreter.java)
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:144)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.j(e.java)
         at com.sunopsis.dwg.cmd.h.z(h.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)
    Caused by: com.hyperion.odi.essbase.ODIEssbaseException: Error records reached the maximum error threshold : 1
         at com.hyperion.odi.essbase.ODIEssbaseDataWriter.checkMaxError(Unknown Source)
         at com.hyperion.odi.essbase.ODIEssbaseDataWriter.sendRecordArrayToEsbase(Unknown Source)
         ... 31 more
    com.hyperion.odi.essbase.ODIEssbaseException: com.hyperion.odi.essbase.ODIEssbaseException: Error records reached the maximum error threshold : 1
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
         at com.sunopsis.dwg.codeinterpretor.k.a(k.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.execScriptingOrders(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTaskTrt(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSqlI.treatTaskTrt(SnpSessTaskSqlI.java)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandSession.treatCommand(DwgCommandSession.java)
         at com.sunopsis.dwg.cmd.DwgCommandBase.execute(DwgCommandBase.java)
         at com.sunopsis.dwg.cmd.e.j(e.java)
         at com.sunopsis.dwg.cmd.h.z(h.java)
         at com.sunopsis.dwg.cmd.e.run(e.java)
         at java.lang.Thread.run(Unknown Source)
    Thanks
    K

    Sorry John,,
    I setup the log file and i see the following in my log file..it says that cannot connect to the essbase server...
    How ever the user id which is setup in the physical schema essbase server has full rights to essbase server and the application.
    2009-09-30 12:03:32,561 INFO [DwgCmdExecutionThread]: ODI Hyperion Essbase Adapter Version 9.3.1.1
    2009-09-30 12:03:32,561 INFO [DwgCmdExecutionThread]: Connecting to Essbase application [Test] on [Server]:[1423] using username [hypadmin].
    2009-09-30 12:03:33,311 INFO [DwgCmdExecutionThread]: Successfully connected to the Essbase application.
    2009-09-30 12:03:33,311 INFO [DwgCmdExecutionThread]: Essbase Load IKM option RULES_FILE = AMORT3
    2009-09-30 12:03:33,311 INFO [DwgCmdExecutionThread]: Essbase Load IKM option RULE_SEPARATOR = |
    2009-09-30 12:03:33,311 INFO [DwgCmdExecutionThread]: Essbase Load IKM option PRE_LOAD_MAXL_SCRIPT =
    2009-09-30 12:03:33,311 INFO [DwgCmdExecutionThread]: Essbase Load IKM option POST_LOAD_MAXL_SCRIPT =
    2009-09-30 12:03:33,327 INFO [DwgCmdExecutionThread]: Essbase Load IKM option ABORT_ON_PRE_MAXL_ERROR = true
    2009-09-30 12:03:33,327 INFO [DwgCmdExecutionThread]: Essbase Load IKM option CLEAR_DATABASE = None
    2009-09-30 12:03:33,327 INFO [DwgCmdExecutionThread]: Essbase Load IKM option COMMIT_INTERVAL = 1000
    2009-09-30 12:03:33,327 INFO [DwgCmdExecutionThread]: Essbase Load IKM option CALCULATION_SCRIPT = null
    2009-09-30 12:03:33,327 INFO [DwgCmdExecutionThread]: Essbase Load IKM option RUN_CALC_SCRIPT_ONLY = false
    2009-09-30 12:03:33,358 DEBUG [DwgCmdExecutionThread]: LoadData Begins
    _2009-09-30 12:03:33,358 DEBUG [DwgCmdExecutionThread]: Error occured in sending record chunk...Cannot begin data load. Analytic Server Error(1042015): Network error: Cannot Locate Connect Information For [server]_
    2009-09-30 12:03:33,374 DEBUG [DwgCmdExecutionThread]: Sending data record by record to essbase
    2009-09-30 12:03:33,374 INFO [DwgCmdExecutionThread]: Logging out and disconnecting from the essbase application.
    K
    Edited by: kanna143 on Sep 30, 2009 10:35 AM

  • Issue with loading data into cube with duplicate members in different dimensions

    Have a cube with allowed dublicate members only in two dims (Period1 and Period2).
    Outlines saves with no errors, but when i try to load data, i get some errors:
    "Member 2012-12 is a duplicate member in the outline
    A6179398-68BD-7843-E0C2-B5EE81808D0B    01011    cd00    st01    2905110000    EK    fo0000    NNNN-NN-NN    2012-12    cust00000$$$    1" 
    These dims represent two similar periods of time. Am I to change member names and aliases a little in one of them(f.e. yyyy1-mm --> yyyy1 - mm1)?
    Users wouldnt like it...
    Period1
      yyyy1
        yyyy1/q1
          yyyy1-mm1
          yyyy1-mm2
          yyyy1-mm3   
        yyyy1/q2
    Period2
      yyyy1
        yyyy1/q1
          yyyy1-mm1
            yyyy1-mm1-dd1 
            yyyy1-mm1-dd2
          yyyy1-mm2
          yyyy1-mm3   
        yyyy1/q2
    Tnanx

    You may have to use fully qualified name something like
    [Period1].[2012-12]
    [Period2].[2012-12]
    You can refer to ESSBASE admin guide on "Creating and Working With Duplicate Member Outlines"
    Regards,
    Sunil

  • Semantic Partitioning Delta issue while load data from DSO to Cube -BW 7.3

    Hi All,
    We have created the semantic partion with the help of BADI to perform Partitions. Everthing looks good,
    first time i have loaded the data it is full update.
    Second time i init the data load. I pulled the delta  from DSO to Cube . DSO is standard where as Cube is partition with the help of semantic partition. What i can see is the records which are updated in Latest delta are shown into the report rest all are ignored by the system.
    I tried compression but still it did not worked.
    Can some has face this kind
    Thanks

    Yaseen & Soujanya,
    It is very hard to guess the problem with the amount of information you have provided.
    - What do you mean by cube is not being loaded? No records are extracted from DSO? Is the load completing with 0 records?
    - How is data loaded from DSO to InfoCube? Full? Are you first loading to PSA or not?
    - Is there data already in the InfoCube?
    - Is there change log data for DSO or did someone delete all the PSA data?
    Sincere there are so many reasons for the behavior you are witnessing, your best option is to approach your resident expert.
    Good luck.
    Sudhi Karkada
    <a href="http://main.nationalmssociety.org/site/TR/Bike/TXHBikeEvents?px=5888378&pg=personal&fr_id=10222">Biking for MS Relief</a>

  • Issue while loading data from DSO to Cube

    Gurus,
    I'm facing a typical problem while loading Cube from a DSO. The load is based upon certain conditions.
    Though the data in DSO satisfies those condition, but the Cube is still not being loaded.
    All the loads are managed through Process Chains.
    Would there be any reason in specific/particular for this type of problem?
    Any pointers would be of greatest help !
    Regards,
    Yaseen & Soujanya

    Yaseen & Soujanya,
    It is very hard to guess the problem with the amount of information you have provided.
    - What do you mean by cube is not being loaded? No records are extracted from DSO? Is the load completing with 0 records?
    - How is data loaded from DSO to InfoCube? Full? Are you first loading to PSA or not?
    - Is there data already in the InfoCube?
    - Is there change log data for DSO or did someone delete all the PSA data?
    Sincere there are so many reasons for the behavior you are witnessing, your best option is to approach your resident expert.
    Good luck.
    Sudhi Karkada
    <a href="http://main.nationalmssociety.org/site/TR/Bike/TXHBikeEvents?px=5888378&pg=personal&fr_id=10222">Biking for MS Relief</a>

  • Short dump issue while loading data

    Hello All,
    We are trying to load the data to  the cube , but every time it is throwing us a short dump with the below error.
    We have found a SAP Note as well  for the below error , Please suggest, how to implement the same.
    Error:
    Runtime Errors         CONVT_NO_NUMBER
    METHOD end_routine.
    === Segments ===
       FIELD-SYMBOLS:
         <RESULT_FIELDS>    TYPE _ty_s_TG_1.
       DATA:
         MONITOR_REC     TYPE rstmonitor.
       ... "insert your code here
    --  fill table "MONITOR" with values of structure "MONITOR_REC"
    -   to make monitor entries
       ... "to cancel the update process
        raise exception type CX_RSROUT_ABORT.
    *Fiscal year quarter addition
       loop at RESULT_PACKAGE assigning <RESULT_FIELDS>.
         if <RESULT_FIELDS>-/bic/zsp_Mon BETWEEN 1 and 3.
           CONCATENATE 'FY' <RESULT_FIELDS>-fiscper+2(2) 'Q3' INTO
           <RESULT_FIELDS>-/bic/zsp_FYQA.
         elseif <RESULT_FIELDS>-/bic/zsp_Mon BETWEEN 4 and 6.
           CONCATENATE 'FY' <RESULT_FIELDS>-fiscper+2(2) 'Q4' INTO
           <RESULT_FIELDS>-/bic/zsp_FYQA.
         elseif <RESULT_FIELDS>-/bic/zsp_Mon BETWEEN 7 and 9.
           CONCATENATE 'FY' <RESULT_FIELDS>-fiscper+2(2) 'Q1' INTO
           <RESULT_FIELDS>-/bic/zsp_FYQA.
         elseif <RESULT_FIELDS>-/bic/zsp_Mon BETWEEN 10 and 12.
           CONCATENATE 'FY' <RESULT_FIELDS>-fiscper+2(2) 'Q2' INTO
           <RESULT_FIELDS>-/bic/zsp_FYQA.
      endif.
    endloop.
    ENDDMETHOD.                    "end_routine
    THOD inverse_end_routine.
    IMPORTING

    Hi,
    As suggested by note, you need to put this code in end routine.
    Check the syntax , if error go through it and fix syntax (take help of ABAPer if required) and
    activate Transformation and DTP.
    Thank-You.
    Regards,
    VB

  • Issue in loading data from data mart

    Actual load is from APO to BW is happening, but load gets failed when it is loaded from base cube to target cube (data mart), this flow is in 3.5.
    Attached document contains the system msg.

    Your load is still in process.
    Where load failure mesasge?
    Try to activate source 8DSO/Cube and later actiavte update rules thru prorgams.
    Can you check your job details from SM37.
    From info pack monitor, copy request id.
    Go SM37,Job name - <enter copied request id>.
    Use job status - Active/cancel/finished.
    Use proper data ranges.
    User - *

  • Issue in loading data into OLAP dim/cube.

    All,
    Facing this issue while executing the dimension in OLAP.
    Warning:-_
    O/*RA-20101: 17:39:12 ***Error Occured in BUILD_DRIVER: In __XML_SEQUENTIAL_LOADER: In __XML_LOAD_DIMVALS: Error loading dimension values for DIM_SST.DIMENSION, level DIM_SST.SOL_LEVEL.LEVEL, hierarchy DIM_SST.AW$NONE.HIERARCHY, mapping group DIM_SST.SOL_LEVEL.MAPGROUP1.DIMENSIONMAPGROUP. In __XML_LOAD_DIMVALS_ITEM: In ___XML_LOAD_TEMPPRG: All dimension values must be single line values.
    TRUNCATE_LOAD=false
    AW Execution status: Success
    17:39:12 Started Build(Refresh) of MIS Analytic Workspace.
    17:39:12 Attached AW MIS in RW Mode.
    17:39:12 Started Loading Dimensions.
    17:39:12 Started Loading Dimension Members.
    17:39:12 Started Loading Dimension Members for DIM_SST.DIMENSION (1 out of 1 Dimensions).
    17:39:12 Failed to Build(Refresh) MIS Analytic Workspace.
    17:39:12 ***Error Occured in BUILD_DRIVER: In __XML_SEQUENTIAL_LOADER: In __XML_LOAD_DIMVALS: Error loading dimension values for DIM_SST.DIMENSION, level DIM_SST.SOL_LEVEL.LEVEL, hierarchy DIM_SST.AW$NONE.HIERARCHY, mapping group DIM_SST.SOL_LEVEL.MAPGROUP1.DIMENSIONMAPGROUP. In __XML_LOAD_DIMVALS_ITEM: In ___XML_LOAD_TEMPPRG: All dimension values must be single line values.

    Hi there,
    What version of 10g OLAP are you using?
    It can be very important to apply the latest patches where possible - more details can be found here:- http://www.oracle.com/technology/products/bi/olap/collateral/olap_certification.html
    Thanks,
    Stuart Bunby
    OLAP Blog: http://oracleOLAP.blogspot.com
    OLAP Wiki: http://wiki.oracle.com/page/Oracle+OLAP+Option
    OLAP on OTN: http://www.oracle.com/technology/products/bi/olap/index.html
    DW on OTN : http://www.oracle.com/technology/products/bi/db/11g/index.html

  • Data load issue with export data source - BW 3.5

    Hi,
    We are facing issues in loading data with the help of export data source.
    We have created export data source of 0PCA_C01 cube. With the help of this export datasource,  we are loading data to other custom cube. Scenario is working fine in development server.
    But when we transported objects to quality server data is not getting loaded to custom target cube.
    It is extracting zero records.  All transports are ok and we have generated export datasource in quality before transports .Also regenerated export datasource after transport and activated infosource, update rule via RS* programs.  Every object is active but data is not getting extracted.
    RSA3 for 80PCA_C01 datasource isn't extracting any record in Quality. Records getting extracted in development.   We are in BW 3.5 with patch level 19.
    Please guide us to resolve the issue.
    Thanks,
    Aditya

    Hi
    Make sure that you have relevant Role & Authorization at Quality/PRS.
    You have to Transport the Source Cube first and then Create a Generate Export Data Source in QAS. Then, replicate data sources for BW QAS Soruce System. Make sure this replicated Data Source in QAS. Only then can transport new update rules for second cube.
    Hope it helps and clear

  • Error while loading data into ODS

    Hi we have BI for NW04S and CRM4.0. While loading data for a CRM ODS , I have received a error message which I never seen before. Yesterdaya data load was smooth and I only see issue while loading data today
    The error message in the monitor is as follows
    "Port 'A000000003' does not exists in the table of Port Description' ( message ID E0, NO 31 ).
    have anybody encountered this.
    Thanks
    Arun

    Thanks Marc,
      we found the table EDIPOA and found that the port that is missing. It seems that the enytry( row)  for our CRM system is missing which has this port as a field in this table.
      Not sure whether to manually populate it or it's part of the source system repair etc.
      please let me know
      Thanks
    Arunava

  • Memory dump while loading data

    Hello,
    We are facing memory issues while loading data in BW system.
    This particular loading process is taking more than 4GB (3GB of EM and 1GB of heap) and still failing.
    We cannot increase the memory beyond this. Can you please suggest a way to load the data ?
    Runtime Errors         TSV_TNEW_PAGE_ALLOC_FAILED
    Date and Time          09.03.2010 22:15:49
    Error analysis
         The internal table "{A:2*\TYPE=%_T00003S00000114O0000015636}" could not be
          further extended. To enable
         error handling, the table had to be delete before this log was written.
         As a result, the table is displayed further down or, if you branch to
         the ABAP Debugger, with 0 rows.
         At the time of the termination, the following data was determined for
         the relevant internal table:
         Memory location: "Session memory"
         Row width: 1140
         Number of rows: 30824
         Allocated rows: 30824
         Newly requested rows: 8 (in 1 blocks)
    How to correct the error
         The amount of storage space (in bytes) filled at termination time was:
         Roll area...................... 14627872
         Extended memory (EM)........... 3004121016
         Assigned memory (HEAP)......... 1000024976
         Short area..................... " "
         Paging area.................... 24576
         Maximum address space.......... 4294967295
         If the error occures in a non-modified SAP program, you may be able to
         find an interim solution in an SAP Note.
         If you have access to SAP Notes, carry out a search with the following
         keywords:
         "TSV_TNEW_PAGE_ALLOC_FAILED" " "
         "GP44XYX8R5ZMOGVGGMHG0U1U04O" or "GP44XYX8R5ZMOGVGGMHG0U1U04O"
         "CONVERT_FROM_MEMORY"

    Try to load by reducing data package size.
    I think you have some routine in transformation/update rule. Check that whether you are using "SELECT *" in that. It may be one reason for this error.
    You can check OSS Note 396642, 379266 and 997930.
    Edited by: Pravender on Mar 10, 2010 1:09 PM

Maybe you are looking for