Defragmenting Planning Cube with Refresh

Has any body ever worked across a script which will defragment the planning app cubes(3 cubes in an app say).
Essbase cubes its easy I know. clearing lev0 data/reset cube/load data ang aggregate.
For planning cubes the process need to be 100% automatic. Also it requires automatic refresh of Planning app before the aggregation step, to bring currency rates to essbase.
If essbase and planning servers are different how to accomplish this a fully automatic script without any manual intervention.
Any suggestion will be appreciated.
Thanks,
Tarini

Try the following utility to refresh your cube:
[http://www.network54.com/Forum/58296/message/1201822528/Re-+CubeRefresh-exe]
Use a batch file or shell script to launch the utility above, then use MaxL (launched from the same batch file) to export the data from the Essbase cubes, clear and reload to defrag. I would install the Essbase client on your Planning Server (so you can run MaxL from your Planning server). Schedule with whatever scheduling tool you prefer.
- Jake

Similar Messages

  • Does exist a way to read data from IP planning cube with ABAP?

    Hello All.
    My scenario is as follows:
    I have an ODS where we store costcenters to be planned. This ODS is loaded via a manual falt file load (regular dta transfer process).  In order to avoid inconsisten data in the planning cube, I want to check in my ODS load process, if any costcenter which is already planned (it has planned data in my planning cube) is being deleted in the loaded file.
    To do this, I need to access my planning data cube ( i guess also planning buffer), so my question is if there is any function module which retrieves data from a planning level as reference data.
    Thanks a lot and best regards,
    Alfonso.

    Hi Alfonso,
    note 1101726 shows how the plan buffer can be read.
         l_r_plan_buffer = cl_rsplfa_plan_buffer=>if_rsplfa_plan_buffer~get_instance( i_infoprov ).
         l_r_plan_buffer->get_data( EXPORTING i_t_charsel = l_t_charsel       
                                              i_include_zero_records = rc_false
                                              i_r_msg = l_r_msg
                                    IMPORTING e_r_th_data  =  l_r_th_data
                                    EXCEPTIONS OTHERS  = 2 ).  
    Please take a look at the note. You do not need to implement the after_burn_selection exit, but you can find sample code how to read the planning buffer. Please give it a try.
    Another solution would be to use the function module RSDRI_INFOPROV_READ. But you need to make shure that you first close the yellow request. This can be done using function module RSAPO_SWITCH_BATCH_TO_TRANS.
    Hope this helps
    Matthias Nutt
    SAP Consulting Switzerland

  • A/P & Consignment data pull to Planning Cube.....

    I put this in "BC & Extractors" section, but did not get any response.....
    Need couple of clarifications (brief background is provided.....):
    1. We do DELTA extractions using 0FI_AP_3 adn 0FI_AR_4 to a base cube which includes both 'Open' and 'Cleared'
    items. From this cube we do a DELTA load to a Planning Cube on a DAILY basis. We filter the data on Company Code
    and 'Item Status' (OnlyOPEN Items) for Planning purposes. The key figure pulled to Planning Cube is '0DEB_CRE_DC'.
    The bsae cube has data at Line Item Document level, where as the planning cube is only at Customer & Vendor level.
    The question is - Will this Daily DELTA to Planning cube with filter on Open Item bring correct key figure? For example,
    if the status of one of the items pulled as OPEN in earlier Delta changes to "CLEARED" the subsequent day, will the Key
    Figure reflect the changes and show correct value? Is there any impact on the key figure due to the filter on Open Item?
    My thinking is the key figure 0DEB_DRE_DC will reflect he correct number (adjsuted) since it is a summated key figure.
    I could be totally wrong and want to understand this with comments from experts.
    2. Has anyone worked with pulling Consignment (SMI) data into Planning cube? We pull this data using a generic extractor
    from a SAP table "RKWA" with some logic. Since I pull A/P data that contains SMI & Non-SMI transactions. To avoid
    Double Counting, how can I avoid bringing in the SMI transactions from A/P? There is no flag to diffferentiate the SMI /
    Non-SMI transactions....
    I am hoping someone has worked in these area to offer expert comments/ recommendations.
    Thanks.. Shaun
    Please post one question only once
    Edited by: Vikram Srivastava on Aug 13, 2010 1:02 PM

    Hello Lee,
    For your first question as you are using the Filter on the status of the record,
    lets say today the status is OPen and as a part of the filter the Delta will be able to pick this record as the status is OPen but if the status is changed to Closed tomorrow then this record will be completely ignored and hence in the next level target you will not have the right status at all.
    So you need to plan your filters accordingly.
    Hope this helps.
    Thanks
    Murali

  • Planning problems with russian language during refresh application

    Hello all ))
    I have problems with Planning during refresh my Application with Essbase.
    I intsalled Essbase on RHEL4 and set ESSLANG=Russian_Russia.ISO-8859-5@Default also I set locale on ru_RU.iso88595
    Ater that I installed Planning on Solaris 9 which uses Oracle RDBMS with NLS_CHARACTERSET=AL32UTF8 and set locale on server to ru_RU.iso88595
    Now I can create dementions, members on russion language in Essbase without problems also I can create dementions,members,etc on russian language in Planning. But if I create demention or member in Planning on Russian languge and ) try refresh my Application with Essbase I give follow error
    "=====(HspCubeCreation.java)sQueryString:?Application=Test1
    com.hyperion.planning.olap.EssbaseException: (1060010)
    at com.hyperion.planning.olap.HspEssbaseOutlineAPI.EssAddMemberEx(Native Method)
    at com.hyperion.planning.olap.HspCubeRefreshTask.addChildren(Unknown Source)
    at com.hyperion.planning.olap.HspCubeRefreshTask.addMembers(Unknown Source)
    at com.hyperion.planning.olap.HspCubeRefreshTask.addDimensionsAndMembers(Unknown Source)
    at com.hyperion.planning.olap.HspCubeRefreshTask.buildOutlines(Unknown Source)
    at com.hyperion.planning.olap.HspCubeRefreshTask.run(Unknown Source)
    at java.lang.Thread.run(Unknown Source)"
    Why it's happened?
    Please anybody help me

    since march 5 we are on planning 9.3.1.1 and also having problems with refreshing:
    com.hyperion.planning.HspCubeRefreshInProgressException: Cannot complete your request because the database is being refreshed.
         at com.hyperion.planning.HspJSImpl.lockApp(I)V(Unknown Source)
         at com.hyperion.planning.HyperionPlanningBean.lockApp()V(Unknown Source)
         at HspSecurityFilter.Handle(Ljavax.servlet.http.HttpServletRequest;Ljavax.servlet.http.HttpServletResponse;)V(Unknown Source)
         at HspSecurityFilter.doPost(Ljavax.servlet.http.HttpServletRequest;Ljavax.servlet.http.HttpServletResponse;)V(Unknown Source)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:763)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:856)
         at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:225)
         at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:127)
         at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:272)
         at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:165)
         at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3150)
         at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
         at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:121)
         at weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:1973)
         at weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:1880)
         at weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1310)
         at weblogic.work.ExecuteThread.execute(ExecuteThread.java:207)
         at weblogic.work.ExecuteThread.run(ExecuteThread.java:179)
    com.hyperion.planning.HspCubeRefreshInProgressException: Cannot complete your request because the database is being refreshed.
         at com.hyperion.planning.olap.HspCubeRefreshTask.run()V(Unknown Source)
         at java.lang.Thread.run()V(Unknown Source)
    anyone a idea?

  • Issues with Writing back data to Planning Cube After adding a new characterstic

    Hello,
    We have two cubes and a MP on top of it. ( Reporting Cube and Planning Cube ). An aggregation layer has also been built and so does a planning BEx query on top of it.
    Before Scenario: things were working fine as we were able to write back data to planning cube using the BEx.
    After Scenario: We created a new Dimension in the planning cube and added a new characteristic ( lets call it ZX_KEY ) . This Characteristic holds the concatenated values of some of the other characteristics in its master data tables . ( For ex: Account_Customer ID_Country-code_Industry-code ).
    This new characteristic is also present in the reporting cube and has been added and transferred to MP.
    This new characterstic was added to the planning BEx query as one of the rows and we executed it the analyzer. After entering the required values for all characterstics ( including the newly added ZX_KEY ) and the key figure values, we tried saving it. Now this is where the Analyzer is throwing up this error..
    ~~Characterstic combination cannot be assigned to part provider ~~
    ~~Characterstic 'ZX_KEY'; Characterstic value 'R123000000_12_US_TX'~~
    ~~ Entered values are incorrect:Correct before navigation ~~
    The value that we are entering ( R123000000_12_US_TX ) is picked up from the master data table of ZX_KEY by double clicking and selecting it from the fetched values. So i am not sure as to why its throwing up the above error. Request your help on this please.
    regards,
    Karthik

    Hi,
    Try to check that all characteristics are correctly assigned at multiprovider level.
    hope it helps

  • Create and Populate a Hyperion Planning Cube using Hyperion Data Integratio

    Friends,
    I am new to Essbase and have worked extensively in Informatica. Hyperion DIM (OEM version of Informatica) is chosen to create and populate a Hyperion Planning System (with Essbase cube in the backend).
    I am using Hyperion DIM 9.3.
    Can someone let me know (or share a document) how I can do the following
    1) Create a Planning application with a Essbase Cube in the backend using Hyperion Data Integration Management
    2) How to populate the Essbase outline and the actuals Essbase cube with data using DIM.
    Thanks a lot for all help.

    Hi,
    You cannot create planning applications using DIM.
    To load metadata have a look at :- http://www.oracle.com/technology/obe/hyp_fp/DIM_Planning/OBE_Dim_Planning.html
    You can refresh planning database in DIM by
    To enable the Refresh Database property for a session:
    In Workflow Manager, right-click the session and select Edit.
    Click the Mapping tab.
    Select a Planning target.
    Check the Refresh Database box.
    Ok?
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Planning Security Filter refresh and Essbase is Crashing

    Hi,
    I am on System 9.3.1 and I am doing a Planning security filter refresh and Essbase is Crashing. Please advise in this case.
    Security is fine on this cube. It used to work fine earlier. Now from few days we are facing this issue.

    We see same issue in 9.2.0.3, but only with 1 of our 5 applications. Was never able to get support to identify a cause, our resolution is to push users in small groups and save the essbase.sec file. Real pain. You could try stopping essbase service, using essbase.bak file and restarting then try to push and see if it works. I did identify issues with our openLDAP which I felt was the cause, I was able to clean it up in TEST and it works, but did same in Prod and worked once, then reverted back to crashing again, unless we push users in really small groups. Interested to see if you get a 'fix' from someone...

  • Key figure display in planning book with respect to Time bucket profile

    Hi,
    I am loading a key figure to planning area from the info cube for the current month. When I review the key figure in planning book with monthly time bucket profile it shows 85 for the current month. In the same planning book with weekly bucket profile, it shows 55 from the current week and future weeks and the remaining 30 goes into the past weeks of the current month.
    How to make the total quantity 85 to show in the current and future weeks only.
    thanks and regards
    Murugesan

    Hi Murugesan,
    Within the Planning Area, the data is stored at the lowest level granularity that you maintain in storage bucket profile. Then during display, system will decide what data to show depending on what kind of time bucket profile you use in the planning view, and the time based disaggregation that you maintain for Key Figure.
    In this below case, what time characteristic do you have in cube? Is it date, week or month?
    If it's date, check how much KF data is maintained on the dates which belong to week which has days both in this month/last month e.g. if I talk about Dec 2011, how much data is stored 1,2,3 & 4 th of Dec, 2011.
    This data would appear in Dec in monthly view, but in week, it would appear in the week starting 28th November.
    If data is maintained in cube in weeks, then you need to calculate how time based disaggregation would show it to you in months.
    If it's months, then you would need to find out how much data would go to the days in the past week of the month.
    The time based disaggregation may be causing you some issues, but in that case, I would not expect 30 out of 85 to go in the past week, unless you have data in cube in days.
    Data shown in weekly view for week starting 28th Nov should ideally be a small proportion of 85, unless you are using a time stream/fiscal year variant, due to which most of December is in holidays.  The only other exception I can think of is that you have data in teh days mentioned above.
    It would be best to help the business understand this disaggregation logic, rather than thinking of manipulating the data to shift to a later week.
    If this logic doesn't explain your situation, then please provide the date/week/month at which you have data in cube, and what quantity.
    Thanks - Pawan

  • Currency Translation in Planning sequence with variable for target currency

    Hello ,
    i have a requirement, the cube contains multiple data in the Group Currency for multiple company codes. & I want to convert these amounts to another KF as local currency using planning sequence only (not using transformation, as the users want to do this when they require.)
    For this, i defined a Aggr level , filter & Planning function with Currency translation selecting the source field as AMOUNT_GC & target field as AMOUNT_LC using the Currency Translation type ZCTGCLC.
    The target currency is not know beforehand, as it depends on the company code of each records that comes in.
    Hence, Currency translation type uses  a Customer exit variable 'Z_R_0026' for target currency.
    This is customer exit variable which searches the Master data table of the Company code & replies back with the local currency of that particular company code (derived value from the company code variable also in teh filter).
    However, when i execute this planning sequence, it gives me an error saying, "no value found for variable 'Z_R_0026'
    1. when edit the CMOD code, to return only say "EUR" the planning sequence runs fine.However when i use this select statement to fetch the corresponding value of the currency using the company code reference from the company code master data table, it gives the above error. it seems to be a problem with the select statement.
    below is the code for that exit
    when 'Z_R_0026'.
         data: l_curr type /BI0/OICURRENCY.
         IF i_step = 2.
           LOOP AT i_t_var_range INTO loc_var_range
                              WHERE vnam = 'Z_S_0003'. " Company Code Single
             CLEAR l_s_range.
        Filling Currency Key from Company Code
             select single currency
             from /bi0/pcomp_code
             into l_curr
             where comp_code = loc_var_range-low
             and objvers = 'A'.
             if sy-subrc = 0.
               l_s_range-low = l_curr.
             else.
               l_s_range-low      = 'ABC'.
             endif.
             l_s_range-opt      = 'EQ'.
             l_s_range-sign     = 'I'.
             APPEND l_s_range TO e_t_range.
             clear l_s_range.
             EXIT.
           ENDLOOP.
         ENDIF.
    2. there is a option of using infoobject for target currency in the RSCUR itself, however, when i try to input the 0COMP_CODE for finding the attribute currency from this infoobject, it doesnt accept.
    has anyone tried this earlier ?
    Any help/suggestions , guys, pls?
    Thanks
    John

    Hi Praveen ,
    Thanks for that.
    I have a Question.
    If RSCUR were to only accept a Date InfoObject , it would be placed under a differnt heading.
    The Tab is "Currrency"  & the sub-Tab is "Target Currency"
    so i guess, it should accept other infoObjects as well.
    Moreover, the help on that InfoObject buttons says
    "Specification of an InfoObject (Characteristic) in which a currency
    attribute is stored in InfoObject maintenance. At runtime, the
    associated currency attribute is determined using the characteristic
    value and this is interpreted as source or target currency."
    & i have already defined 0currency as a attribute of 0comp_code. so this should work ideally.
    almost similiar CMOD code i have used for other too, but i dont know why only this one is getting struck.
    any help, pls.
    Thanks
    John
    Edited by: John Lewis on Apr 17, 2009 3:24 PM
    Edited by: John Lewis on Apr 17, 2009 3:25 PM

  • Unable to load data into Planning cube

    Hi,
    I am trying to load data into a planning cube using a DTP.
    But the system throws a message saying that, real time data loading cannot be performed to the planning cube.
    What would be the possible reason for the same?
    Thanks & Regards,
    Surjit P

    Hi Surjit,
    To load data into cube using DTP, it should be put in loading mode only.
    The real time cube behaviour is set to planning mode only when the data is input to the cube via the planning layouts / or through the Fileupload program.
    You can change the behaviour of the cube in RSA1 -> Right click on the cube -> Change Real-time load behaviour and select the first option(real time cube can be loaded with data; planning not allowed)
    Best Rgds
    Shyam
    Edited by: Syam K on Mar 14, 2008 7:57 AM
    Edited by: Syam K on Mar 14, 2008 7:59 AM

  • Error in delta loading from a planning cube

    hi all,
    I am using a delta load from one planning cube to another.When the changed records are of a magnitude of 100 then it is successful but when the records are of magnitude of 100000 then it stays yellow with no data records sent and eventually fails.the packet size is 20000.
    any help is appreciated.

    Hello Ajay,
    Have you checked that your Request is closed in the planning cube? FM RSAPO_CLOSE_TRANS_REQUEST will do it
    It is a customizing issue to tell the infopackage to end "green" when it has no data.
    Transaction Spro ==> SAP Netweaver ==> SAP Business Information Warehouse ==> Automated Processes ==> Monitor Settings ==> Set Traffic Light Color. There you have to set the Traffic light on green if no data is available in the system.
    Hope that helps
    best regards
    Yannick

  • Planning cube Key figure over flow error

    Hi Experts,
    I am getting the following  Key figure errors in Planning cube while running the Planning sequences.
    1.Over flow occurred when calculating KF
    KF details : Data type: DEC-Counter or amount filed  with comma and sign.
    2.ORA-01455: converting column overflows integer data type
    Data type : INT4- 4byte integer,Integer number with sign.
    The number s are growing in cube and it is giving overflow error .We are compressing the cubes once in a week.we are using these KF's in two cubes.Please help me to resolve the issue.
    Regards
    Prasad
    Edited by: PRASAD on Feb 1, 2009 9:39 AM

    Hi Prasad,
             Check here.........
    https://websmp130.sap-ag.de/sap(bD1lbiZjPTAwMQ==)/bc/bsp/sno/ui_entry/entry.htm?param=69765F6D6F64653D3030312669765F7361706E6F7465735F6E756D6265723D3131393536373226
    https://websmp130.sap-ag.de/sap(bD1lbiZjPTAwMQ==)/bc/bsp/spn/sapnotes/index2.htm?numm=460652
    Thanks,
    Vijay.

  • Data from planning cube to TEXT FILE IN directories of BW

    hi all
    i have 2 transactional planning cubes in BW. the data in these cubes in bw needs to be uploaded monthly in to a single text file so that APO system can access the data for some purpose.
    my question is how to upload the data from transactional cubes to the text file, which is more detailed. is that use ful to construct a ODS or CUBE(basic) on top of the transactional cubes.
    if i keep the file in BW directory whethere APO can access that.?..
    please help me out
    praveen.

    Hi,
    You could create a multicube over your 2 transactional cubes and extract your data with a query on the multicube.
    Make sure you restrict 0REQUID 'request id' using 0S_RQMRC 'most current data' to be sure you extract the most recently planned data.
    No need to convert the transactional cubes to basic cubes.
    To download the query I think you have to use the reporting agent of the broadcaster.
    Thanks for you points if you found this helpful or if this solved your problem.
    Best Regards,
    Filip

  • Newbie questions regarding pulling data from hyperion planning cube w/OBIEE

    Hello there,
    we've recently implemented Essbase and are currently pumping it full of revenue/expense data from out source systems to calculate NOI. This data is stored in a staging table at the detail level where it is sourced into Hyperion and aggregated. We also have OBIEE 10g (plan to upgrade to 11g later this year) and we would like to connect to and report out of Essase. Our ultimate goal is to be able to report on the NOI numbers in the planning cube but have the ability to drill down to the detail level which is not stored in Essbase. We've heard it is possible, albeit not native for OBIEE 10g to do this. We've also heard that it is not best practice to use our transactional cube for this type of reporting but to actually create a second "reporting cube".
    What are best practices for getting this NOI data out of Hyperion and merging it with our relational detail reporting? Can we somehow export the data from the cube and store it in a relational database? Should we clone the cube (if even possible) and configure both it and our relational source in the BI repository and setup all drill-throughs there?
    Any info is GREATLY appreciated. Thank you.
    Edited by: cisGuy on Sep 20, 2010 5:31 PM

    i have found information on how to use ODI to extract data for the cube. What I'm really trying to find out though is best practices for reporting off summary level data in Essbase with the ability to drill-through to the detail.
    We've heard reporting off the same cube that users are writing back and transacting on is bad. Do we need to make a "reporting cube" and then bring that in OBIEE and merge it with a relational source or is it better to extract the data from Essbase into flat files and join it to detail tables in our relational source?

  • How to give planners read only access to Planning cube for ad-hoc analysis?

    Hi Everyone,
    I am trying to issue smartview reports to the planners. Most of the reports are coming off Essbase (reporting) application. However, we have to generate one report from planning server because of the text fields. Currently, on planning server, planners have access only to the forms. They dont have access to the planning cube. But to make them able to refresh this report, we need to give them access to the planning cube.
    But we dont want them to do ad-hoc data write on planning cube. So my question goes like this - Is there a way for planners to have write access on the web forms but only read access to the cube through samrtview ??

    The access permissions will match between planning and smart view if a planning connection is used.
    Cheers
    John
    http://john-goodwin.blogspot.com/

Maybe you are looking for