Unable to load data into Planning cube

Hi,
I am trying to load data into a planning cube using a DTP.
But the system throws a message saying that, real time data loading cannot be performed to the planning cube.
What would be the possible reason for the same?
Thanks & Regards,
Surjit P

Hi Surjit,
To load data into cube using DTP, it should be put in loading mode only.
The real time cube behaviour is set to planning mode only when the data is input to the cube via the planning layouts / or through the Fileupload program.
You can change the behaviour of the cube in RSA1 -> Right click on the cube -> Change Real-time load behaviour and select the first option(real time cube can be loaded with data; planning not allowed)
Best Rgds
Shyam
Edited by: Syam K on Mar 14, 2008 7:57 AM
Edited by: Syam K on Mar 14, 2008 7:59 AM

Similar Messages

  • Unable to load data into Essbase cube using Essbase Studio

    Hi
    We are creating an essbase cube using Essbase Studio using flat files as data sources.
    We have taken different hierarchies into different flat files and created one fact file having dimension intersection along with data.
    We are able to create the cube and the hierarchy but not able to load any data.
    We are getting the following error
    Failed to deploy Essbase cube.
    Caused by: Unable to perform dataload from more than one flat file.
    Could anyone please help on this?

    Oh this was killing me, so I did this test in 11.1.1.3:
    1) Excel 2007 format -- no go, Essbase didn't see it
    2) Excel 2003 format, three sheets -- only the first sheet was read into an empty rule
    3) Excel 2003 format, one sheet -- the first sheet was read into an empty rule
    4) Excel 95 format, one sheet -- the first sheet was read into an empty rule
    The lesson?
    1) Excel 2007/2010 sheets don't work (no surprise there as the .xlsx format isn't supported).
    2) Excel 2003 and lower (hey, if you have Excel 4, I'll bet that works as well) work, but only the first sheet is recognized.
    Regards,
    Cameron Lackpour

  • Error when loading data into Planning

    Unable to load data into planning 11.1.2.3.200 using ODI 11.1.1.7
    Please find the errors from below logs:
    INFO  [SimpleAsyncTaskExecutor-2]: Oracle Data Integrator Adapter for Hyperion Planning
    INFO  [SimpleAsyncTaskExecutor-2]: Connecting to planning application [xxxxxxx] on [xxxxxxxxxxx]:[xxxx] using username [admin].
    INFO  [SimpleAsyncTaskExecutor-2]: Successfully connected to the planning application.
    INFO  [SimpleAsyncTaskExecutor-2]: The load options for the planning load are
      Dimension Name: Account Sort Parent Child : false
      Load Order By Input : false
      Refresh Database : false
    INFO  [SimpleAsyncTaskExecutor-2]: Begining the load process.
    DEBUG [SimpleAsyncTaskExecutor-2]: Number of columns in the source result set does not  match the number of planning target columns.
    INFO  [SimpleAsyncTaskExecutor-2]: Load type is [Load dimension member].
    ERROR [SimpleAsyncTaskExecutor-2]: Record [[A603010, null, null, null, null, null, null, null, null, null, null, null, xxxxx, -100, F3E0,C011,E7172_93275,FY17,Stage 1,Current Service Level,Jul, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null]] was rejected by the Planning Server.
    ERROR [SimpleAsyncTaskExecutor-2]: Record [[A601060, null, null, null, null, null, null, null, null, null, null, null, xxxxx, -250, F3E0,C011,E7172_93275,FY17,Stage 1,Current Service Level,Jul, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null, null]] was rejected by the Planning Server.
    log.err
    Account,Data Load Cube Name,Budget,Point-of-View,Error_Reason
    A603010,xxxxx,-100,F3E0,C011,E7172_93275,FY17,Stage 1,Current Service Level,Jul,Cannot load dimension member, error message is: RemoteException occurred in server thread; nested exception is:
      java.rmi.UnmarshalException: unrecognized method hash: method not supported by remote object
    A601060,xxxxx,-250,F3E0,C011,E7172_93275,FY17,Stage 1,Current Service Level,Jul,Cannot load dimension member, error message is: RemoteException occurred in server thread; nested exception is:
      java.rmi.UnmarshalException: unrecognized method hash: method not supported by remote object
    FDMEE log:
    ERROR [AIF]: Error: No records exist for Period 'Pd 2 - 2014-08-01'
    ERROR [AIF]: Error: No records exist for Period 'Pd 3 - 2014-09-01'
    FDMEE Logging Level is set to 5

    Are you sure that you haven't got Planning 11.1.2.3.500 in your environment? This sounds spookily similar to the issue described in note 1678759.1 which was seen after applying Planning 11.1.2.3.500. If it's definitely Planning 11.1.2.3.500 is there any chance that someone has applied ODI patch 18687916 to the ODI home (C:\Oracle\Middleware\odi)?
    If you /*are*/ running Planning 11.1.2.3.500 then you might need to apply patch 18687916 to C:\Oracle\Middleware\odi - you might need note 1683307.1 to take some of the pain out of applying this patch though.
    Given the error that you're seeing though there's a mismatch in versions of JAR files between ODI and the Planning server somewhere.
    Regards
    Craig

  • Automatically trigger the event to load data from Planning cube to Standard Cube

    Hello,
    We have a below set up in our system..
    1. A Planning BEx query using which user makes certain entries and writes back data to the planning cube.
    2. An actual reporting cube which gets data from the planning cube above.
    Now, what we want to do is to automate the data load from Planning cube to Reporting cube.
    This involves 2 things..
    1. Change the setting " Change real time load behaviour " of the planing cube to Planning.
    2. Trigger the DTP which loads data from Planning cube to reporting cube.
    We want to automate the above two steps...
    I have tried few things to achieve the same..
    1. Created an event in SM64,
    2. In the Planning cube "Manage" Screen,  clicked on "Subsequent Processing" and provided the event details. ( Not sure if its the correct place to provide the event detail ).
    3. Wrote a ABAP program which changes the setting of the planning cube (  " Change real time load behaviour " to Loading )
    4. Created a process chain, where we have used the event as a start variant, used the ABAP program as a next step, DTP run as the last step..
    This i hoped, would trigger the event as soon as a new request comes and sits in planning cube, which in turn would trigger the process chain which would load the data from Planning cube to reporting cube..
    This is not working. I dont think the event is triggering, and even if does i am not sure whether it will start the process chain automatically. Any ideas please ?

    hi,
    try to do the transformation directly in the input cube by using CR of type exit, more details :
    http://help.sap.com/saphelp_nw70ehp2/helpdata/en/43/1c3d0f31b70701e10000000a422035/content.htm
    hope it helps.

  • Weblogic  message ---error while loading data into planning database

    Hi,
    Does anyone have idea on this error message .We are loading data into planning 11.1.2.1 using ERPI standalone 11.1.2.1.501 but everytime it load gives out this message and data load is unsuccessful.platform is Microsoft windows 64-bit 2008R2 ----Database Microsoft SQL Server 2008.
    Please help.
    <Mar 2, 2012 12:20:51 PM PST> <Notice> <Log Management> <BEA-170027> <The Server has established connection with the Domain level Diagnostic Service successfully.>
    <Mar 2, 2012 12:36:24 PM PST> <Warning> <JMX> <BEA-149515> <An error was encountered getting the attribute DatabaseProductName on the MBean com.bea:ServerRuntime=ErpIntegrator0,Name=aif_datasource,Type=JDBCDataSourceRuntime during a call to getAttributes>
    <Mar 2, 2012 12:36:24 PM PST> <Warning> <JMX> <BEA-149515> <An error was encountered getting the attribute DriverVersion on the MBean com.bea:ServerRuntime=ErpIntegrator0,Name=aif_datasource,Type=JDBCDataSourceRuntime during a call to getAttributes>
    <Mar 2, 2012 12:36:24 PM PST> <Warning> <JMX> <BEA-149515> <An error was encountered getting the attribute DriverName on the MBean com.bea:ServerRuntime=ErpIntegrator0,Name=aif_datasource,Type=JDBCDataSourceRuntime during a call to getAttributes>
    <Mar 2, 2012 12:36:24 PM PST> <Warning> <JMX> <BEA-149515> <An error was encountered getting the attribute DatabaseProductVersion on the MBean com.bea:ServerRuntime=ErpIntegrator0,Name=aif_datasource,Type=JDBCDataSourceRuntime during a call to getAttributes>
    <Mar 2, 2012 12:36:24 PM PST> <Warning> <JMX> <BEA-149515> <An error was encountered getting the attribute DatabaseProductName on the MBean com.bea:ServerRuntime=ErpIntegrator0,Name=EPMSystemRegistry,Type=JDBCDataSourceRuntime during a call to getAttributes>
    <Mar 2, 2012 12:36:24 PM PST> <Warning> <JMX> <BEA-149515> <An error was encountered getting the attribute DriverVersion on the MBean com.bea:ServerRuntime=ErpIntegrator0,Name=EPMSystemRegistry,Type=JDBCDataSourceRuntime during a call to getAttributes>
    <Mar 2, 2012 12:36:24 PM PST> <Warning> <JMX> <BEA-149515> <An error was encountered getting the attribute DriverName on the MBean com.bea:ServerRuntime=ErpIntegrator0,Name=EPMSystemRegistry,Type=JDBCDataSourceRuntime during a call to getAttributes>
    <Mar 2, 2012 12:36:24 PM PST> <Warning> <JMX> <BEA-149515> <An error was encountered getting the attribute DatabaseProductVersion on the MBean com.bea:ServerRuntime=ErpIntegrator0,Name=EPMSystemRegistry,Type=JDBCDataSourceRuntime during a call to getAttributes>

    Hi Found solution.login to weblogic console and change the timeout of ERPI from 600 sec to 20000 sec.\
    It worked no more error.
    Cheers

  • Getting error while loading  Data into ASO cube by flat file.

    Hi All,
    i am getting this error Essbase error 1270040: Data load buffer[1] does not exist while loading data into ASO cube.
    does anyone have solution.
    Regards,
    VM

    Are you using ODI to load the data or maxl? If you are using an ODI interface, are you using a load rule also which version of essbase and ODI are you using
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Loading data into two cubes

    We want to set up a delta refresh from R/3 data that will pull data into two cubes.  One cube I want all the records to be loaded, in the second cube I want to filter the records that are loaded into that cube.  I can figure out how to load the data into two cubes, but can I place a filter on one of the loads.  (i.e. only load records of where type = 'X')
    Thanks,
    Chris

    You can do that in the Update Rules to the second cube... In the Start Routine...
    DELETE FROM DATA_PACKAGE
    WHERE type EQ 'X'.
    (Please, verify the right syntax)
    Then with only one Infopackage, you can load both cube with different conditions.
    Hope it helps.
    Regards,
    Luis

  • Unable to load data into any application.database on server.

    I have a rather odd problem that's been vexing my for a few days.
    I am unable to do a data import into any cubes within Essbase, its as if the cube is in read only mode, though everything seems OK. I'm not running any sort of archiving, and to manually check, within essmsh, I did an "alter database end archive". I've tried with our existing and unchanged data load script as well as from within the EAS ( Right-click Data Load ). If I check the processes running on the server ( LINUX server ) the ESSSVR process is the top process eating 100% CPU. I can do outline builds OK.
    From within the application log file, the last few entries are:
    [Wed Apr  3 09:44:58 2013]Local/OP_ACC/Accounts/svc_biserver/Info(1021044)
    Starting to execute query
    [Wed Apr  3 09:44:58 2013]Local/OP_ACC/Accounts/svc_biserver/Info(1021045)
    Finished executing query, and started to fetch records
    [Wed Apr  3 09:44:58 2013]Local/OP_ACC/Accounts/svc_biserver/Info(1021000)
    Connection With SQL Database Server is Established
    [Wed Apr  3 09:44:58 2013]Local/OP_ACC/Accounts/svc_biserver/Info(1003040)
    Parallel dataload enabled: [1] block prepare threads, [1] block write threads.
    [Wed Apr  3 09:45:19 2013]Local/OP_ACC/Accounts/svc_biserver/Info(1021047)
    Finished fetching data
    Nothing has changed on this server for a few weeks so I'm somewhat flummoxed and no other errors show in any other logs ( nohup.out ).
    Essbase 11.1.1.2
    Redhat LINUX_x64 2.6

    Yes, I tried rebooting the whole box, server has 9Gb free space, same story when I load just 1 row via a text file in the EAS, it seems to hang and the ESSSVR process for that cube goes to 100%.
    I've tried a few different cubes, and they all have the same problem, so I suspect its something to do with Essbase itself rather than the specific cube I'm having problems with.
    This is our test server I'm experiencing the problem with and I tried migrating 1 app from UAT back to test which still didnt work.
    Stumped!

  • Unable to load entities into Planning using OutlineLoad utitlity

    Good day all,
    One of our Planning admins needs us to load some new entities for her. We've done this off and on with her Planning app for the last year without a problem. However, over the last couple of days we have not been able to load her latest metadata. She's massaged the load file almost half a dozen times and we get the same error message every time. If anyone has any ideas what it does not like about the header values, I would very much appreciate any feedback.
    We are currently running EPM Planning 11.1.1.3.04
    Successfully logged into "HPEGGD" application, Release 11.113, Adapter Interface Version 5, Workforce supported and not enabled, CapEx not supported and not enabled, CSS Version 3
    [Wed Nov 23 09:47:30 CST 2011]Successfully located and opened input file "D:\FILES\Entity_ISG_11_23_11.txt".
    [Wed Nov 23 09:47:30 CST 2011]Header record fields: Parent, Entity, Alias: Default, Alias: Member-Alias, Aggregation (HPEGGD), Aggregation (HPEGGD2), Aggregation (HPEGGD3), Two Pass Calc, Expense, Time Balance, Time Balance Skip, Data Storage, Level Number, Attributed, WF_Group, Shared_Projects, Contract_Type, Agency, Bus_CC_Mgr, Dom-Intl, ID_Number, Segment, Source_System
    [Wed Nov 23 09:47:30 CST 2011]Located and using "Entity" dimension for loading data in "HPEGGD" application.
    [Wed Nov 23 09:47:30 CST 2011]Unable to obtain dimension information and/or perform a data load: Unrecognized column header value(s), refer to previous messages. (Note: column header values are case sensitive.)
    [Wed Nov 23 09:47:30 CST 2011]Examine the Essbase log files for status if Essbase data was loaded.
    [Wed Nov 23 09:47:30 CST 2011]Planning Outline data store load process finished with exceptions: not all input records were read due to errors (or an empty input file). 0 data records were read, 0 data records were processed, 0 were successfully loaded, 0 were rejected.
    Thanks in advance for the assistance.
    Regards,
    -Bob

    One more massage did the trick.

  • Unable to Load data in Planning

    Hi,
    I am receiving this error when ever i am trying to load data.
    Launch error message:
    "An error occurred while running the specified calc script. Check the log for details."
    and some times
    "You will be unable to update data when essbase in archive mode"
    But i have checked the essbase, no "Back up process" is running.
    Can Anyone advice me what's the issue..i am restarting the server everytime to fix the issue but not sure how i need to rectify it with out restarting the server.
    Regards
    Saravanakumar.Vijay

    I take it you are running a script that loads data and then executes a rule/calc
    "An error occurred while running the specified calc script. Check the log for details."
    This refers to a calc/rule that has been run
    "You will be unable to update data when essbase in archive mode"
    This means that the database is still in archive mode, you can use Maxl to take it out of archive mode.
    e.g. Alter database app.db end archive;
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Error while loading data into the cube

    Hi,
    I loaded data on to PSA and when I am loading the data to the cube through DataTransferProcess, I get an error (red color).
    Through "Manage", I could see teh request in red. How do I get to knoe the exact error? Also what could be the possibel reason for this?
    Also can some one explain the Datatransfer process(not in process chain)?
    Regards,
    Sam

    Hi Sam
        after you load the data  through DTP(after click on execute button..) > just go to monitor screen.. in that press  the refresh button..> in that it self.. you can find the  logs..
       otherwise.. in the request  screen also..  beside of the request number... you can see the logs icon.. you can click on this..
    DTP  means..
    DTP-used for data transfer process from psa to data target..
    check thi link..for DTP:
    http://help.sap.com/saphelp_nw04s/helpdata/en/42/f98e07cc483255e10000000a1553f7/frameset.htm
    to load data in to the datatargets or infoproviders like DSO, cube....
    in your  case.. the problem may be.. check the date formats.. special cherectrs... and
    REGARDS
    @JAY

  • Loading data into Fact/Cube with surrogate keys from SCD2

    We have created 2 dimensions, CUSTOMER & PRODUCT with surrogate keys to reflect SCD Type 2.
    We now have the transactional data that we need to load.
    The data has a customer id that relates to the natural key of the customer dimension and a product id that relates to the natural key of the product dimension.
    Can anyone point us in the direction of some documentation that explains the steps necessary to populate our fact table with the appropriate surrgoate key?
    We assume that we need to have an lookup table between the current version of the customer and the incoming transaction data - but not sure how to go about this.
    Thanks in advance for your help.
    Laura

    Hi Laura
    There is another way to handling SCD and changing Facts. This is to use a different table for the history. Let me explain.
    The standard approach has these three steps:
    1. Determine if a change has occurred
    2. End Date the existing record
    3. Insert a new record into the same table with a new Start Date and dummy End Date, using a new surrogate key
    The modified approach also has three steps:
    1. Determine if a change has occurred
    2. Copy the existing record to a history table, setting the appropriate End Date en route
    3. Update the existing record with the changed information giving the record a new Start Date, but retaining the original surrogate key
    What else do you need to do?
    The current table can use the surrogate key as the primary key with the natural key being being a unique key. The history table has the surrogate key and the end date in the primary key, with a unique key on the natural key and the end date. For end user queries which in more than 90% of the time go against current data this method is much faster because only current records are in the main table and no filters are needed on dates. If a user wants to query history and current combined then a view which uses a union of the main and historical data can be used. One more thing to note is that if you adopt this approach for your dimension tables then they always keep the same surrogate key for the index. This means that if you follow a strict Kimball approach to make the primary key of the fact table be a composite key made up of the foreign keys from each dimension, you NEVER have to rework this primary key. It always points to the correct dimension, thereby eliminating the need for a surrogate key on the fact table!
    I am using this technique to great effect in my current contract and performance is excellent. The splitter at the end of the map splits the data into three sets. Set one is for an insert into the main table when there is no match on the natural key. Set two is when there is a match on the natural key and the delta comparison has determined that a change occurred. In this case the current row needs to be copied into history, setting the End Date to the system date en route. Set three is also when there is a match on the natural key and the delta comparison has determined that a change occurred. In this case the main record is simply updated with the Start Date being reset to the system date.
    By the way, I intend to put a white paper together on this approach if anyone is interested.
    Hope this helps
    Regards
    Michael

  • How to load data into Planning/Essbase from multiple data column

    Dear All,
    I have a interface file which contains multiple data column as follows.
    Year,Department,Account,Jan,Feb,Mar,Apr,May,Jun,Jul,Aug,Sep,Oct,Nov,Dec
    FY10,Department1,Account1,1,2,3,4,5,6,7,8,9,10,11,12
    FY10,Department2,Account1,1,2,3,4,5,6,7,8,9,10,11,12
    FY10,Department3,Account1,1,2,3,4,5,6,7,8,9,10,11,12
    I created a data rule to load these interface file.
    I want to use ODI to upload this interface. I try to specify the rule name in ODI and run the interface.
    But, it came out following errors.
    2010-02-22 11:40:25,609 DEBUG [DwgCmdExecutionThread]: Error occured in sending record chunk...Cannot end dataload. Analytic Server Error(1003014): Unknown Member [FY09,032003,910201,99,0,0,0,0,0,0,0,0,0,0,0,0] in Data Load, [1] Records Completed
    Any idea to fix the column, I sure the member name is correct as I can load data from data load rule correctly.
    Thanks

    Dear John,
    I updated the data load rule delimter to "," and found different error message as follows.
    'A910201.99','HSP_InputValue','HKDepart','No_WT','032003','NO_Lease','FY09','Actual','Final','Local','0','0','0','0','0','0','0','0','0','0','0','0','Cannot end dataload. Analytic Server Error(1003014): Unknown Member [0] in Data Load, [1] Records Completed'
    It seems that the data load rule can recognize the some member except the figures of Jan to Dec..
    thanks for your help

  • Unable to load data to Hyperion planning application using odi

    Hi All,
    When I try to load data into planning using odi, the odi process completes successfully with the following status in the operator ReportStatistics as shown below but the data doesn't seem to appear in the planning data form or essbase
    can anyone please help
    org.apache.bsf.BSFException: exception from Jython:
    Traceback (most recent call last):
    File "<string>", line 2, in <module>
    Planning Writer Load Summary:
         Number of rows successfully processed: 20
         Number of rows rejected: 0
    source is oracle database
    target account dimension
    LKM SQL TO SQL
    IKM SQL TO HYPERION PLANNING is used
    In Target the following columns were mapped
    Account(load dimension)
    Data load cube name
    driverdimensionmetadata
    Point of view
    LOG FILE
    2012-08-27 09:46:43,214 INFO [SimpleAsyncTaskExecutor-3]: Oracle Data Integrator Adapter for Hyperion Planning
    2012-08-27 09:46:43,214 INFO [SimpleAsyncTaskExecutor-3]: Connecting to planning application [OPAPP] on [mcg-b055]:[11333] using username [admin].
    2012-08-27 09:46:43,277 INFO [SimpleAsyncTaskExecutor-3]: Successfully connected to the planning application.
    2012-08-27 09:46:43,277 INFO [SimpleAsyncTaskExecutor-3]: The load options for the planning load are
         Dimension Name: Account Sort Parent Child : false
         Load Order By Input : false
         Refresh Database : false
    2012-08-27 09:46:43,339 INFO [SimpleAsyncTaskExecutor-3]: Begining the load process.
    2012-08-27 09:46:43,355 DEBUG [SimpleAsyncTaskExecutor-3]: Number of columns in the source result set does not match the number of planning target columns.
    2012-08-27 09:46:43,371 INFO [SimpleAsyncTaskExecutor-3]: Load type is [Load dimension member].
    2012-08-27 09:46:43,996 INFO [SimpleAsyncTaskExecutor-3]: Load process completed.

    Do any members exist in the account dimension before the load? if not can you try adding one member manually then trying the load again.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • How to Get the Current data into Planning Layout from the Planning cube

    Hi,
    I have a problem in BPS.   I am selecting the data from cube based on Month org details but I want to see the latest data.   How can i get this data into planning cube.
    Like
    data
    Tran           cal month            org               amt
    1                 jan                        a                  100
    1                 feb                        a                   200
    if i want to read based on Tran org as input values I shoud get the below data but I am getting the previous data.   
    Tran           cal month            org               amt
    1                 feb                        a                   200
    Kindly help me in this regard
    Thanks
    Naveen

    Naveen,
        Are yo having issue when you save something in the layout, the data doesn't appear in the listcube ? Or do you have issue that the latest data you see in the cube doesn't appear in the layout ?
    For the former issue, please look at your selections of listcube and for the second issue, please check your planning level definition, make sure all the restrictions you have applied are valid for this latest data to be presented in the layout.
    Hope this helps.
    Cheers
    Srini

Maybe you are looking for

  • How to read/write local file in lync silverlight application?

    I tried doing that by various method but an exception pop-ups. My lync silverlight application works fine in browser or out of browser (with elevated permissions) but when I try to do so in lync's window extension (CWE) is throws an exception saying

  • Event handlers question

    Which event handler is called first when navigating from a page? Say I am passing parameters from page 1 to page 2. In page 2, will the onCreate get called first or onRequest get called first? In onRequest, I have calmoyr = request->get_form_field( '

  • The Data Access service is either not running or not yet initialized. Check the event log for more information

    Hi, I have SCSM with remote SQL and the SCSM Management server give below error Message: Failed to connect to server 'Name of Server' Microsoft.EnterpriseManagement.Common.ServiceNotRunningException: The Data Access service is either not running or n

  • Quicktime Player fails to launch.

    I can't get QuickTime to work whatsoever. Nothing happens when I click on it. What am I doing wrong?

  • Why AP Management Address is Untagged

    Anyone want to venture a guess as to why Cisco doesn't allow a management address on an AP on a tagged interface?  I believe that it has to do with 'bridge 1 route ip' and that if you were to place a management address on another VLAN you'd be restri