Cannot load data into infocube in BW NW2004s sneak preview

For BW with NW2004s preview.
I can load data into PSA but there is an error happens when loading data from PSA to infocube using DTP. The error is as 'Join table TESTDATRNRPART0 does not exist - read SAP note 871084'. In note 871084, it says the SP05 fix the problem. But I check the status of the BW server, it is 'release 700 / level 0005 / highest support package SAPKW0005'.
Does it mean the BW server has SP05?

Hi Parvendar and Vinod,
Thanks for the information. I checked in SM58. For today it shows 4 TRFC. 1 with user bwremote and 3 with user BASIS(the user i am logged in as). All have error saying "An exception occurred that was uncaught".
I also tried to reschedule the Data for the infopackage alone. We are having 3 infopackages in the process chain and I tried iwth one of them.
When i scheduled it, the Infocube -> Manage shows the status of the request as Yellow. When i try to see the Load monitor, it displays the same message as I saw yesterday.
In Extraction Phase it shows
Extraction (messages): Errors Occured
  Data Request recieved - green
  Data selection scheduled - green
  41 Records Sent (41 Records received_) - yellow
         Data Package1(41 records): Missing Messages - yellow
Date Selection ended - green
Transfer (Idocs and TRFC) Missing messages or warnings
Data package1: arrived in BW: processing: data packet not yet processed
Info IDOC1,2,3,4 are all green
Update PSA, Transfer Rules and Update rules are all green
Update (0 new/0 changed) : warning recieved
  message missing: update finished  for 0TCT_C21
Please let me know what could be the possible issue
Regards,
Raghavan
Edited by: Raghavan Guruswaminathan on May 11, 2010 8:10 AM

Similar Messages

  • ABAP short dump when loading data into Infocube

    Hello All,
    I am getting an ABAP short dump error when loading data into Infocube. I tried to load data from PSA and ODS.
    I ran the code to compress the Infocube, no improvement still getting the same error.
    Error Analysis:
    A raise statement in the program "SAPLRSDU_PART" raised in the exception.
    Internal Notes:
    The termination occured in the function "ab_i fune" of the SAP Basis system. Specifically in the line 2316 of the module "//bas/620/src/krn/runt/abfunc. C#18".
    The internal operation just processed is "FUNE"
    Advance thanks !!!
    Siv

    Hello Siv,
    try to run program SAP_DROP_EMPTY_FPARTITIONS to display existing partitions. Then compare it to the number of uncompressed requests. Maybe there's a mismatch.
    SAP notes that might help: 385163, 590370
    Regards,
    Marc
    SAP NetWeaver RIG, US BI

  • Load data into infocube

    hello can anybody please say me how to load data into infocube in bw7.0
    thanks a lot
    komal.

    Hi,
    These are the steps for creating an Info cube and data loading
    1. It is same as creating an Info cube in BW 3.5
    2. Here once u created an info cube,assign dimesions to it.
    3. Create application component, info source similar to BW 3.5.
    4. After creating info source, right click on it
    5. Here all the dimensions and key figures are mapped automatically in transformation rules.
    6. Right click on info cube find the option : Create Data transfer Process
    7. create an info package, as usual and schedule the package to load the data into an Info provider.

  • Loading data into infocube in bi 7.0 from flat file.

    Hello  All,
    I need the complete procedure for loading data into the infocube from flat file in bi 7.0 i.e.using transformation,DTP etc.
    Please help me with some gud documents.

    Hi Pratighya,
    Step by step procedure for loading data from flat file.
    1. Create the infoobjects you might need in BI
    2. Create your target infoprovider
    3. . Create a source system
    4. Create a datasource
    5. Create and configure an Infopackage that will bring your records to the PSA
    6. Create a transformation from the datasource to the Infoprovider
    7. Create a Data Transfer Process (DTP) from the datasource to the Infoprovider
    8. Schedule the infopackage
    9. Once succesful, run the DTP
    10. This will fill your target.
    Hope this helps
    Regards
    Karthik

  • Runtime error while loading data into infocube

    Hi all,
    I am getting a runtime error while loading the transaction data into the InfoCube.
    Exception name: CX_SY_GENERATE_SUBPOOL_FULL
    Runtime error: GENERATE_SUBPOOL_DIR_FULL
    Anyone has any idea about this error? Any help will be really helpful. Thanks in advance.
    Regards,
    Sathesh

    Hi Sathesh,
    Please provide more details for the problem.
    You can go and check in ST22. Go through the Dump Analysis and Let us know.
    Regards
    Hemant Khemani

  • Steps for loading data into the infocube in BI7, with dso in between

    Dear All,
    I am loading data into the infocube in BI7, with dso in between. By data flow looks like....
    Top to bottom:
    InfoCube (Customized)
    Transformation
    DSO (Customized)
    Transformation
    DataSource (Customized).
    The mapping and everything else looks fine and data is also seen in the cube on the FULL Load.
    But due to some minor error (i guess), i am unable to see the DELTA data in DSO, although it is loaded in DataSource through process chains.
    Kindly advise me, where did i miss.
    OR .. Step by step instructions for loading data into the infocube in BI7, with dso in between would be really helpful.
    Regards,

    Hi,
    my first impulse would be to check if the DSO is set to "direct update". In this case there is no Delta possible, because the Change log is not maintained.
    My second thought would be to check the DTP moving data between the DSO and the target cube. If this is set to full, you will not get a delta. It is only possible to create one DTP. So if you created one in FULL mode you can't switch to Delta. Just create the DTP in Delta mode.
    Hope this helps.
    Kind regards,
    Jürgen

  • Unable to load data into Planning cube

    Hi,
    I am trying to load data into a planning cube using a DTP.
    But the system throws a message saying that, real time data loading cannot be performed to the planning cube.
    What would be the possible reason for the same?
    Thanks & Regards,
    Surjit P

    Hi Surjit,
    To load data into cube using DTP, it should be put in loading mode only.
    The real time cube behaviour is set to planning mode only when the data is input to the cube via the planning layouts / or through the Fileupload program.
    You can change the behaviour of the cube in RSA1 -> Right click on the cube -> Change Real-time load behaviour and select the first option(real time cube can be loaded with data; planning not allowed)
    Best Rgds
    Shyam
    Edited by: Syam K on Mar 14, 2008 7:57 AM
    Edited by: Syam K on Mar 14, 2008 7:59 AM

  • Error in loading data into essbase while using Rule file through ODI

    Hi Experts,
    Refering my previous post Error while using Rule file in loading data into Essbase through ODI
    I am facing problem while loading data into Essbase. I am able to load data into Essbase successfully. But when i used Rule file to add values to existing values I am getting error.
    test is my Rule file.
    com.hyperion.odi.essbase.ODIEssbaseException: com.hyperion.odi.essbase.ODIEssbaseException: Cannot put olap file object. Essbase Error(1053025): Object [test] already exists and is not locked by user [admin@Native Directory]
    at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
    at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.execInBSFEngine(SnpScriptingInterpretor.java:346)
    at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.exec(SnpScriptingInterpretor.java:170)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java:2458)
    at oracle.odi.runtime.agent.execution.cmd.ScriptingExecutor.execute(ScriptingExecutor.java:48)
    at oracle.odi.runtime.agent.execution.cmd.ScriptingExecutor.execute(ScriptingExecutor.java:1)
    at oracle.odi.runtime.agent.execution.TaskExecutionHandler.handleTask(TaskExecutionHandler.java:50)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.processTask(SnpSessTaskSql.java:2906)
    at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2609)
    at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:540)
    at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:453)
    at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:1740)
    at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$2.doAction(StartSessRequestProcessor.java:338)
    at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:214)
    at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:272)
    at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$0(StartSessRequestProcessor.java:263)
    at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$StartSessTask.doExecute(StartSessRequestProcessor.java:822)
    at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:123)
    at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$2.run(DefaultAgentTaskExecutor.java:83)
    at java.lang.Thread.run(Thread.java:662)
    from com.hyperion.odi.common import ODIConstants
    from com.hyperion.odi.connection import HypAppConnectionFactory
    from java.lang import Class
    from java.lang import Boolean
    from java.sql import *
    from java.util import HashMap
    # Get the select statement on the staging area:
    sql= """select C1_HSP_RATES "HSP_Rates",C2_ACCOUNT "Account",C3_PERIOD "Period",C4_YEAR "Year",C5_SCENARIO "Scenario",C6_VERSION "Version",C7_CURRENCY "Currency",C8_ENTITY "Entity",C9_VERTICAL "Vertical",C10_HORIZONTAL "Horizontal",C11_SALES_HIERARICHY "Sales Hierarchy",C12_DATA "Data" from PLANAPP."C$_0HexaApp_PLData" where (1=1) """
    srcCx = odiRef.getJDBCConnection("SRC")
    stmt = srcCx.createStatement()
    srcFetchSize=30
    #stmt.setFetchSize(srcFetchSize)
    stmt.setFetchSize(1)
    print "executing query"
    rs = stmt.executeQuery(sql)
    print "done executing query"
    #load the data
    print "loading data"
    stats = pWriter.loadData(rs)
    print "done loading data"
    #close the database result set, connection
    rs.close()
    stmt.close()
    Please help me on this...
    Thanks & Regards,
    Chinnu

    Hi Priya,
    Thanks for giving reply. I already checked that no lock are available for rule file. I don't know what's the problem. It is working fine without the Rule file, but throwing error only when using rule file.
    Please help on this.
    Thanks,
    Chinnu

  • Error while loading data into 0pur_c03 using 2lis_02_scl

    Hi,
    While trying to load data into 0pur_c03 by using Data Source 2lis_02_scl from R/3 the error message "The material 100 does not exist or not active" is being displayed. I have loaded master data for 0Material and performed change Run also.Still the problem is not resolved.Please suggest in resolving this issue.
    Regards,
    Hari Prasad B.

    I have the same problem in my system and i found that the material XXXXX had movements.....sales documents and more but in certain moment it material was deleted in R3 and it is not loaded into 0MATERIAL infoobject because is marked for deletion....
    If you have this error must activate the load request manually and your infocube can hold the data without any problem.....
    Regards

  • Error while using Rule file in loading data into Essbase through ODI

    Hi Experts,
    I am facing problem while loading data into Essbase. I am able to load data into Essbase successfully. But when i used Rule fule to add values to existing values am getting error.
    test is my Rule file.
    com.hyperion.odi.essbase.ODIEssbaseException: com.hyperion.odi.essbase.ODIEssbaseException: Cannot put olap file object. Essbase Error(1053025): Object [test] already exists and is not locked by user [admin@Native Directory]
         at org.apache.bsf.engines.jython.JythonEngine.exec(JythonEngine.java:146)
         at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.execInBSFEngine(SnpScriptingInterpretor.java:346)
         at com.sunopsis.dwg.codeinterpretor.SnpScriptingInterpretor.exec(SnpScriptingInterpretor.java:170)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.scripting(SnpSessTaskSql.java:2458)
         at oracle.odi.runtime.agent.execution.cmd.ScriptingExecutor.execute(ScriptingExecutor.java:48)
         at oracle.odi.runtime.agent.execution.cmd.ScriptingExecutor.execute(ScriptingExecutor.java:1)
         at oracle.odi.runtime.agent.execution.TaskExecutionHandler.handleTask(TaskExecutionHandler.java:50)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.processTask(SnpSessTaskSql.java:2906)
         at com.sunopsis.dwg.dbobj.SnpSessTaskSql.treatTask(SnpSessTaskSql.java:2609)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatAttachedTasks(SnpSessStep.java:540)
         at com.sunopsis.dwg.dbobj.SnpSessStep.treatSessStep(SnpSessStep.java:453)
         at com.sunopsis.dwg.dbobj.SnpSession.treatSession(SnpSession.java:1740)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$2.doAction(StartSessRequestProcessor.java:338)
         at oracle.odi.core.persistence.dwgobject.DwgObjectTemplate.execute(DwgObjectTemplate.java:214)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.doProcessStartSessTask(StartSessRequestProcessor.java:272)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor.access$0(StartSessRequestProcessor.java:263)
         at oracle.odi.runtime.agent.processor.impl.StartSessRequestProcessor$StartSessTask.doExecute(StartSessRequestProcessor.java:822)
         at oracle.odi.runtime.agent.processor.task.AgentTask.execute(AgentTask.java:123)
         at oracle.odi.runtime.agent.support.DefaultAgentTaskExecutor$2.run(DefaultAgentTaskExecutor.java:83)
         at java.lang.Thread.run(Thread.java:662)
    from com.hyperion.odi.common import ODIConstants
    from com.hyperion.odi.connection import HypAppConnectionFactory
    from java.lang import Class
    from java.lang import Boolean
    from java.sql import *
    from java.util import HashMap
    # Get the select statement on the staging area:
    sql= """select C1_HSP_RATES "HSP_Rates",C2_ACCOUNT "Account",C3_PERIOD "Period",C4_YEAR "Year",C5_SCENARIO "Scenario",C6_VERSION "Version",C7_CURRENCY "Currency",C8_ENTITY "Entity",C9_VERTICAL "Vertical",C10_HORIZONTAL "Horizontal",C11_SALES_HIERARICHY "Sales Hierarchy",C12_DATA "Data" from PLANAPP."C$_0HexaApp_PLData" where      (1=1) """
    srcCx = odiRef.getJDBCConnection("SRC")
    stmt = srcCx.createStatement()
    srcFetchSize=30
    #stmt.setFetchSize(srcFetchSize)
    stmt.setFetchSize(1)
    print "executing query"
    rs = stmt.executeQuery(sql)
    print "done executing query"
    #load the data
    print "loading data"
    stats = pWriter.loadData(rs)
    print "done loading data"
    #close the database result set, connection
    rs.close()
    stmt.close()
    Please help me on this...
    Thanks & Regards,
    Chinnu

    Hi Priya,
    Thanks for giving reply. I already checked that no lock are available for rule file. I don't know what's the problem. It is working fine without the Rule file, but throwing error only when using rule file.
    Please help on this.
    Thanks,
    Chinnu

  • Music on my iPod that I cannot load/sync into iTunes on my new computer

    The original computer that I loaded my iTunes app and db on crashed hard and I could not retrieve the data or deauthorize it. I have been trying to setup iTunes on my new computer. Problem - I have music on my iPod that I cannot load/sync into iTunes on my new computer.

    Click here and follow the instructions.
    (34700)

  • Error while loading data into cube

    hi BW gurus,
    when ever i am trying to load data into the cube from flat file after scheduling iam getting short dump in BW system. I checked it in st22 it is giving me a error as exception add_partition_failed. please help me to sort out this problem. If you know the error recovery please give me the answer in detail.
    I will assign points for good answers.

    This is what the note says:
    Symptom
    The process of loading transaction data fails because a new partition cannot be added to the F fact table.The loading process terminates with a short dump.
    Other terms
    RSDU_TABLE_ADD_PARTITION_ORA, RSDU_TABLE_ADD_PARTITION_FAILED, TABART_INCONSITENCY, TSORA, TAORA , CATALOG
    Reason and Prerequisites
    The possible causes are:
    SQL errors when creating the partition
    Inconsistencies in the Data Dictionary control tables TAORA and TSORA
    Solution
    BW 3.0A & BW 3.0B
    In the case of SQL errors:Analyze the SQL code in the system log or short dump and if possible, eliminate the cause. The cause is often a disk space problem or lock situations on the database catalog or, less frequently: the partitioning option in the ORACLE database is not installed.
    The most common cause of the problem is inconsistencies in the TAORA and TSORA tables. As of Support Package 14 for BW 3.0B/Support Package 8 for BW 3.1C, the TABART_INCONSITENCY exception is issued in this case. The reason is almost always missing entries in TSORA for the tablespaces of the DDIM, DFACT and DODS data classes.
    The TAORA table contains the assignment of data classes to data tablespaces and their attributes, for example:
    Tablespace Data class
    DDIM PSAPDIMD ........
    DFACT PSAPFACTD ........
    DODS PSAPODSD .......
    Foreach data tablespace, the TSORA table must contain an entry for the corresponding index tablespace, for example:
    TABSPACE INDSPACE
    PSAPDIMD PSAPDIMD
    PSAPFACTD PSAPFACTD
    PSAPODSD PSAPODSD
    In most cases, these entries are missing and have to be added. See also notes 502989 and 46272.

  • No dataload.err: Using ODI to load data into Essbase

    Hi,
    I am using ODI to load data into Essbase. In the load data KM options, I select the rule file and mention that errors be logged.
    After executing the data load through ODI, I get a log file and an error file. But the error file does not have the bad record like the "dataload.err" has when I load data through EAS and use the same data and rules file.
    Am I missing some option? Is there some option to log the bad data to the error file or some option to generate a dataload.err file while using ODI?
    Thanks.

    Hi John,
    I have set LOG_ERRORS to Yes and given a file name. A file is generated everytime the interface is executed, but the file is blank.
    I have also set the Log_enabled to Yes. The log file is generated and has the error in it. The error line in the Log file ( not the error file) looks like this -
    2009-04-16 10:45:38,457 INFO [DwgCmdExecutionThread]: ODI Hyperion Essbase Adapter Version 9.3.1.1
    2009-04-16 10:45:38,457 INFO [DwgCmdExecutionThread]: Connecting to Essbase application [XXXXXX] on [XXXXXX]:[1423] using username [admin].
    2009-04-16 10:45:39,315 INFO [DwgCmdExecutionThread]: Successfully connected to the Essbase application.
    2009-04-16 10:45:39,315 INFO [DwgCmdExecutionThread]: Essbase Load IKM option RULES_FILE = d_PL
    2009-04-16 10:45:39,315 INFO [DwgCmdExecutionThread]: Essbase Load IKM option RULE_SEPARATOR = ,
    2009-04-16 10:45:39,315 INFO [DwgCmdExecutionThread]: Essbase Load IKM option PRE_LOAD_MAXL_SCRIPT =
    2009-04-16 10:45:39,315 INFO [DwgCmdExecutionThread]: Essbase Load IKM option POST_LOAD_MAXL_SCRIPT =
    2009-04-16 10:45:39,315 INFO [DwgCmdExecutionThread]: Essbase Load IKM option ABORT_ON_PRE_MAXL_ERROR = true
    2009-04-16 10:45:39,315 INFO [DwgCmdExecutionThread]: Essbase Load IKM option CLEAR_DATABASE = None
    2009-04-16 10:45:39,315 INFO [DwgCmdExecutionThread]: Essbase Load IKM option COMMIT_INTERVAL = 1000
    2009-04-16 10:45:39,315 INFO [DwgCmdExecutionThread]: Essbase Load IKM option CALCULATION_SCRIPT = null
    2009-04-16 10:45:39,315 INFO [DwgCmdExecutionThread]: Essbase Load IKM option RUN_CALC_SCRIPT_ONLY = false
    2009-04-16 10:45:39,346 DEBUG [DwgCmdExecutionThread]: LoadData Begins
    2009-04-16 10:45:39,549 DEBUG [DwgCmdExecutionThread]: Error occured in sending record chunk...Cannot end dataload. Analytic Server Error(1003014): Unknown Member [DPT_DFDFDF] in Data Load, [2] Records Completed
    2009-04-16 10:45:39,549 DEBUG [DwgCmdExecutionThread]: Sending data record by record to essbase
    2009-04-16 10:45:40,968 INFO [DwgCmdExecutionThread]: Logging out and disconnecting from the essbase application.
    There is nothing in the error file. There should have been a record for the unknown member, DPT_DFDFDF.
    I am using ODI version 10.1.3.5.
    and John, thanks for the ODI blog. It has helped me a lot in the past couple of months.
    Thank you.

  • Error when loading data into Planning

    Unable to load data into planning 11.1.2.3.200 using ODI 11.1.1.7
    Please find the errors from below logs:
    INFO  [SimpleAsyncTaskExecutor-2]: Oracle Data Integrator Adapter for Hyperion Planning
    INFO  [SimpleAsyncTaskExecutor-2]: Connecting to planning application [xxxxxxx] on [xxxxxxxxxxx]:[xxxx] using username [admin].
    INFO  [SimpleAsyncTaskExecutor-2]: Successfully connected to the planning application.
    INFO  [SimpleAsyncTaskExecutor-2]: The load options for the planning load are
      Dimension Name: Account Sort Parent Child : false
      Load Order By Input : false
      Refresh Database : false
    INFO  [SimpleAsyncTaskExecutor-2]: Begining the load process.
    DEBUG [SimpleAsyncTaskExecutor-2]: Number of columns in the source result set does not  match the number of planning target columns.
    INFO  [SimpleAsyncTaskExecutor-2]: Load type is [Load dimension member].
    ERROR [SimpleAsyncTaskExecutor-2]: Record [[A603010, null, null, null, null, null, null, null, null, null, null, null, xxxxx, -100, F3E0,C011,E7172_93275,FY17,Stage 1,Current Service Level,Jul, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null]] was rejected by the Planning Server.
    ERROR [SimpleAsyncTaskExecutor-2]: Record [[A601060, null, null, null, null, null, null, null, null, null, null, null, xxxxx, -250, F3E0,C011,E7172_93275,FY17,Stage 1,Current Service Level,Jul, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null]] was rejected by the Planning Server.
    log.err
    Account,Data Load Cube Name,Budget,Point-of-View,Error_Reason
    A603010,xxxxx,-100,F3E0,C011,E7172_93275,FY17,Stage 1,Current Service Level,Jul,Cannot load dimension member, error message is: RemoteException occurred in server thread; nested exception is:
      java.rmi.UnmarshalException: unrecognized method hash: method not supported by remote object
    A601060,xxxxx,-250,F3E0,C011,E7172_93275,FY17,Stage 1,Current Service Level,Jul,Cannot load dimension member, error message is: RemoteException occurred in server thread; nested exception is:
      java.rmi.UnmarshalException: unrecognized method hash: method not supported by remote object
    FDMEE log:
    ERROR [AIF]: Error: No records exist for Period 'Pd 2 - 2014-08-01'
    ERROR [AIF]: Error: No records exist for Period 'Pd 3 - 2014-09-01'
    FDMEE Logging Level is set to 5

    Are you sure that you haven't got Planning 11.1.2.3.500 in your environment? This sounds spookily similar to the issue described in note 1678759.1 which was seen after applying Planning 11.1.2.3.500. If it's definitely Planning 11.1.2.3.500 is there any chance that someone has applied ODI patch 18687916 to the ODI home (C:\Oracle\Middleware\odi)?
    If you /*are*/ running Planning 11.1.2.3.500 then you might need to apply patch 18687916 to C:\Oracle\Middleware\odi - you might need note 1683307.1 to take some of the pain out of applying this patch though.
    Given the error that you're seeing though there's a mismatch in versions of JAR files between ODI and the Planning server somewhere.
    Regards
    Craig

  • Can not load data into the ods.

    When I was trying to load data into BW3.5 from R3 system, I am getting this message
    "For InfoSource ZSD_A507_011, there is no Hierarchies for master data in source system R3PCLNT300
    "  any idea.  There is no hierarchies for this.

    Hello Mat,
    What are you trying to load??
    If the data source is a hierarchy datasource??
    just check this thread
    There are no hierarchies for this InfoSource in source system
    Thanks
    Ajeet

Maybe you are looking for

  • Query

    I have the following query in the report. select b.sup_name,vendor_number,        decode(div,null,'ANY',div) div,        decode(dept,null,'ANY',dept) dept,         decode(per,100,'ANY',(100-per)) per,         decode(per,0,'ANY',1) rate,         decod

  • AIR won't uninstall or install

    here are the log files 9/27/13 7:50:38.042 AM Install Adobe Muse: Begin runtime download from http://airdownload.adobe.com/air/3/nai/macos10.7/x86/installer 9/27/13 7:54:35.121 AM Install Adobe Muse: Begin signature download from http://airdownload.a

  • How to enable firefox addon if it is disbaled.(need help in pref.js values)

    Addon are coming but they are in disbaled mode.how to enable those addons using pref.js for all users.

  • Changing exif: time

    Hello to everybody, Being new in this forum, I would like to introduce myself as an photo-fan working regular with Aperture. That does not mean, I know that program in all details. So here is my first question: in mobile me, I published my fotos in a

  • Gx630 gray screen on start up

    i had my G630 8 months recently in the last few weeks it has been doing something strange when i shut down and turn it back on the screen starts up gray and sometimes times the screen freezes and color lines come across when ever it does this i just