EIS 7.1 - Load Data in ASO - Performance tuning

Hi,
Right now I've EIS integration to load member and data in ASO cube using 2 separate meta outlines. It takes long time to load the data in ASO cube. To enhance the data loading performance, I'm planning to take the following actions. Can you pls let me know if you've any better idea to do it in EIS/Essbase 7.1 version.
1. Load the data using freeform loading/own rule file while members will be built using EIS meta outline. So that drill through reports won't be affected.
2. Use custom sqls to fetch the data from relational database and while building the data file, do the required transformations in the sql scripts itself.
3. Can split the source data file into multiple files and data load buffer can be used to load the data in ASO cube.
Thanks in advance!

Hi,
You should try the essbase forum you will probably get more responses :- Essbase
Cheers
John
http://john-goodwin.blogspot.com/

Similar Messages

  • Getting error while loading  Data into ASO cube by flat file.

    Hi All,
    i am getting this error Essbase error 1270040: Data load buffer[1] does not exist while loading data into ASO cube.
    does anyone have solution.
    Regards,
    VM

    Are you using ODI to load the data or maxl? If you are using an ODI interface, are you using a load rule also which version of essbase and ODI are you using
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Loading data in ASO using sql

    hi...
    is it possible to load data in an aso cube using sql....similar to what we can do fro a bso cube..
    in total how many ways are there for loading data in aso cube??

    Hi Srivastava,
    Load data between aso and bso are also the same. so, you can use sql to load your data into aso cube.
    Thanks
    Regards,
    VieN

  • Error while loading data to ASO database using rules files and buffer id

    Hello all,
    While loading the data to our ASO cube from a data file and using a server rules file, I am getting a warning: "Incorrect Syntax near ['$'] ". But my script no where mentions the usage of $. Also the console throws no error while loading the data. It displays the progress bar 0-100% while the data is being loaded, but actually no data is getting loaded (when retrieval is checked). Also the script is unable to form any log file or error file. Can anyone throw some light what might be the reason behind this kind of data load behaviour.
    Thanks in advance.
    Amit

    is that the real script? I don't see you using any parameter. (That's where $ comes to play). Try a double quote with echo (I never used echo in MaxL)
    Also don't post passwords and server names on a forum.
    Regards
    Celvin
    http://www.orahyplabs.com

  • Report Slow Due to Mass Data , Soln for Performance Tuning

    Dear All,
    I am making report with mass data so for this i have to put For All enteries in & Ranges at lodz of places.
    1. For all enteries in is making my report works very slow.
    2. If i change for all enteries by Ranges then if the no. of records are large the system will thn
        throw dump.
    If the no. of records are large or the logic applied in complex then report is taking very long
    execution time , Can anyone suggest me the method by which i can optimize my report and make it
    run smoother in performance.
    Thanks
    Ankesh Jindal

    Hi,
    >
    Ankesh Jindal wrote:
    > The problem is with FAE and ranges acc to uptill i have discovered,
    > As I have mentioned if I take FAE and number of records are large the execution will take
    > very large amount of time for that i have changed FAE to ranges but still if no. of records are
    > large in ranges system will throw dump
    >
    so far so good. SQL statements must not get too large.
    >
    Ankesh Jindal wrote:
    > so the best soln for this which i have considered for
    > my reports is to use ranges but with some logic applied that is ;
    >  suppose i have 20,000 records then send data to ranges in 3k or 4k lots .
    > so in this case i am using ranges with 3k or 4k lots so the sytem will not throw dump and i will get
    > faster execution of query with ranges...
    >
    General question: How big is the runtime difference for your SELECT with Range (4k) and the default FAE (5 in case of ORACLE)... more than 20%? could you post your actual run time figures?
    Your range approach is faster because you do less database calls. 5 db calls if you have 4 k entries in the range.
    You can influence the number of database calls as well for the FAE.
    Assuming you are running on ORACLE with a default configuration you have 5 entries per call.
    (Parameters rsdb/max_in_blocking_factor, rsdb/max_blocking_factor). So you will end up
    with 4000 db calls with 5 records each. 
    You compare that with 5 db calls with 4000 records for your range... this is not fair
    Hint your FAE with this, this would lead to 5 db calls for the FAE as well.:
        %_hints oracle '&max_in_blocking_factor 4000&'.
    now compare again... .
    Note1: Be care full with big ranges and blocking factors... cost based optimizers may react sensitive to big inlists or or concatenations and may change plans suddenly... .
    Note2: If you are not on ORACLE your blocking factors may be considerably higher (30, 60, ...).
    Kind regards,
    Hermann

  • DATA LOAD WORNINGS IN ASO CUBES

    Hi Every one,
    While loading data into ASO cubes in Essbae we are getting wornings like *"Data load strem contains 1.25797e 08 and [0] #misssing cells ".*My data file have #missing values and o's and sepecial carecters like E .I want To load the complete data with out warnings.Kindly let me know if any one's know the solution .Whether i need to change any settings in rule file or how to ingnore those cells .
    Thanks,
    Vikram

    The warnings are really informational messages to let you know it loaded and did not load those values. Which is fine as they tend to bloat a cube (The zeros). #missing is not going to load anyway and the E is exponential format of numbers whinch should not be a problem. Excel will display it this way, but you can format it without the E. You don't mention if you are doing this from EAS or MaxL and what version you are on. In Veraion 11, in EAS there are options in the load dialog to turn on or off the loading of zzeros and missing across the top. In MaxL, I don't see the syntax in the Tech reference, but I thought it was there in 9.

  • EIS error while automating data loading

    Hi,
    We daily load ASO cube using EIS through planned jobs (.cbs).
    This morning the load data process was still running after 3 hours... it last 20 mins normally. The only error I was having was the following, and nothing explicit in the olapisvr.log!
    Error :
    Environment variable ARBORPATH is not set.
    Unable to open input file
    Essbase Integration Services Shell - 11.1.1.3.0 (Build EIS111110B021)
    Copyright(c) 1991, 2009, Oracle Corporation
    All rights reserved
    [LOCAL] ->
    Script :
    %EIS_BIN%\olapicmd.exe -f%EIS_BATCH%\dwh_cos.cbs

    Hi,
    While creating batch file did you set the ARBORPATH?
    If not include SET ARBORPATH=Location of the essbase server in batch file and give a try.
    Thanks,
    Prathap

  • Possible to load data to an account with formula in ASO ?

    For example, if we calculate an account for the forecast periods and load data from HFM to the same account for the actual periods, it seems that it’s not feasible?
    If you put formula on one account, even if the periods are specified not to calculate for the actual months, it will still not let me load anything to it.
    Is there a common way to solve this? Or do I just have to create a different account?

    As Glenn said, can't be loaded.
    Your other options are:
    1. Calculate the value outside of Essbase / derive it in ETL / fudge it via load rule (yuk) etc...
    2. Create an extra 'Input' account, tell the formula to pick from the 'Input' account for actuals and derive for forecast
    3. Calculate in Essbase with an ASO procedural calc - trickier to implement if you haven't done it before, but would look cleaner than an extra account - might give better retrieval performance too, depending on what your derivation of forecast does
    4. Similar to the 'extra' account but in the Scenario dimension; add a 'ForecastInput' member and place a formula on the 'Forecast' member. You'd have to make Scenario dynamic but beyond the overhead of the Forecast derivation itself I can't see that this would cause problems with a typical, non-aggregating Scenario dimension
    1, 3 and 4 may look cleaner if you end up having to apply the same calculation to lots of different accounts in future, rather than just one.

  • FDM to load data in Essbase ASO cube

    Anybody have used FDM to load data in Essbase ASO cube? How do you clear and run calc on ASO cube?
    Thanks

    Does the Essbase Adapter for FDM Support ASO Cubes? [ID 1168153.1]
    Modified 17-AUG-2010 Type HOWTO Status PUBLISHED
    Applies to:
    Hyperion Financial Data Quality Management - Version: 11.1.1.3.00 and later [Release: 11.1 and later ]
    Information in this document applies to any platform.
    Goal:
    Does the Essbase adapter for FDQM support ASO cubes?
    Solution:
    ASO cubes are not currently supported in FDQM.
    Unpublished Enhancement 6568323 has been created and it is currently under consideration for a future release.
    References
    BUG:6568323 - 8-529236080 - CUSTOMER WANTS TO TAKE ADVANTAGE OF THE ASO FUNCTIONS IN ESSBASE.
    Related
    Products
    Middleware > Enterprise Performance Management > Financial Data Quality Management > Hyperion Financial Data Quality Management

  • Data Load MAXLs in ASO

    Hi All,
    Greetings of the day !!!!
    Want to understand the difference between "Add values create slice" and "override values create slice" used in data loading MAXLs
    Suppose we initialized buffer and loaded data in buffer then we can use following two MAXLs
    1)
    import database AsoSamp.Sample data
    from load_buffer with buffer_id 1
    add values create slice;
    2)
    import database AsoSamp.Sample data
    from load_buffer with buffer_id 1
    override values create slice;
    Q1
    What i am thinking logically is if i am again loading the data in the same intersections from which slice is created ADD VALUE will add it and override value will overwrite it .... e.g if 100 was present earlier and we are again loading 200 then ADD will make 300 and overwrite will result 200. Let me know if my understanding is correct
    Q2
    Why do we use "create slice" ? What is the use? Is it for better performance for data loading? Is it compulsary to merge the slices after dataloading??
    Cant we just use add value or override values if we dont want to create slice...
    Q3
    I saw two MAXLs for merging also ... one was Merge ALL DATA and other was MERGE incremental data ... Whats the diff ? In which case we use what?
    Pls help me in resolving my doubts... Thanks a lot !!!!

    Q1 - Your understanding is correct. The buffer commit specification determines how what is in the buffer is applied to what is already in the cube. Note that there are also buffer initialization specifications for 'sum' and 'use last' that apply only to data loaded to the buffer.
    Q2 - Load performance. Loading data to an ASO cube without 'create slice' takes time (per the DBAG) proportional to the amount of data already in the cube. So loading one value to a 100GB cube may take a very long time. Loading data to an ASO cube with 'create slice' takes time proportional to the amount of data being loaded - much faster in my example. There is no requirement to immediately merge slices, but it will have to be done to design / process aggregations or restructure the cube (in the case of restructure, it happens automatically IIRC). The extra slices are like extra cubes, so when you query Essbase now has to look at both the main cube and the slice. There is a statistic that tells you how much time Essbase spends querying slices vs querying the main cube, but no real guidance on what a 'good' or 'bad' number is! See http://docs.oracle.com/cd/E17236_01/epm.1112/esb_tech_ref/aggstor_runtime_stats.html.
    The other reason you might want to create a slice is that it's possible to overwrite (or even remove, by committing an empty buffer with the 'override incremental data' clause in the buffer commit specification) only the slice data without having to do physical or logical clears. So if you are continually updating current period data, for example, it might make sense to load that data to an incremental slice.
    Q3 - You can merge the incremental slices into the rest of the cube, or you can merge multiple incremental slices into one single incremental slice, but not into the rest of the cube. Honestly, I've only ever wanted to use the first option. I'm not really sure when or why you would want to do the second, although I'm sure it's in there for a reason.

  • Can we load data for all levels in ASO?

    Hi All,
    Im creating cube in ASO
    can i load data for all levels in ASO
    we can load data for all Levels In BSO but in ASO i need confirmation????
    and one more
    wat is the consider all levels option in ASO is used for ? wat is the purpose?
    Can any one help ,it would be appriciate.
    Thanks

    In an ASO cube you can only load to level zero
    The consider all levels if used for aggregation hints. It allows you to tell the aggregation optimizer to look at all levels when deciding if aggregation needs to be done on the dimension

  • Can we load data for all levels in ASO cube

    Hi All,
    Can we load data for all levels of members in ASO cube in 9.3.1.
    Regards

    Yes you can load data for all levels in an ASO cube in any version HOWEVER, none of the upper level data in any cube will be there when you look for it. You will get a warning message in the load because ASO cubes don't store data at upper levels. It is the same as loading data into dynamic calc members in BSO cube. It will do the load without compalints, but there will be no data there (At least you get the warning in ASO)

  • Getting error while loading data from sql to ASO

    Hi There,
    we are working on ASO cube(Hyperion 9.3.0.1) and my data source is Oracle 10 g R2 while we are trying to load data or build dimension from sql, we are getting an error. Although, previously it was working properly. we are getting different error different time. I have updated essbase.cfg file for NETDELAY 800 & NETRETRYCOUNT 1000.I have changed panding cache size limit in my application from 48 to 96 also but still getting same error. Following are the errors
    ERROR
    Database IBasic loaded
    Application IdeaBas2 loaded - connection established
    Application [IdeaBas2] started with process id [10920]
    Object [IBasic] is locked by user [dhanjit]
    Cannot read SQL driver name for [Hyperion Client Sample] from [ODBC.INI]
    Cannot read SQL driver name for [Hyperion BIplus Client Sample1] from [ODBC.INI]
    Cannot read SQL driver name for [Hyperion BIplus Client Sample2] from [ODBC.INI]
    Cannot read SQL driver name for [tmw2k_1] from [ODBC.INI]
    Connection String is generated
    Connection With SQL Database Server is Established
    SQL Connection is Freed
    Building Dimensions Elapsed Time : [10.641] seconds
    Reading Parameters For Database [Drxxxxxx]
    Declared Dimension Sizes = [109 165 15 80 1785 1938 1242 93 25 14 10 6 20 1502 21 5 211 2 ]
    Actual Dimension Sizes = [108 158 15 78 1785 1938 1241 93 25 14 10 6 20 1502 21 5 211 1 ]
    Network error [10054]: Cannot Send Data
    Network error [10054]: Cannot Send Data
    Unexpected Essbase error 1042012
    Object [IBasic] unlocked by user [dhanjit]
    regards,
    Dhanjit

    Hi There,
    we are working on ASO cube(Hyperion 9.3.0.1) and my data source is Oracle 10 g R2 while we are trying to load data or build dimension from sql, we are getting an error. Although, previously it was working properly. we are getting different error different time. I have updated essbase.cfg file for NETDELAY 800 & NETRETRYCOUNT 1000.I have changed panding cache size limit in my application from 48 to 96 also but still getting same error. Following are the errors
    ERROR
    Database IBasic loaded
    Application IdeaBas2 loaded - connection established
    Application [IdeaBas2] started with process id [10920]
    Object [IBasic] is locked by user [dhanjit]
    Cannot read SQL driver name for [Hyperion Client Sample] from [ODBC.INI]
    Cannot read SQL driver name for [Hyperion BIplus Client Sample1] from [ODBC.INI]
    Cannot read SQL driver name for [Hyperion BIplus Client Sample2] from [ODBC.INI]
    Cannot read SQL driver name for [tmw2k_1] from [ODBC.INI]
    Connection String is generated
    Connection With SQL Database Server is Established
    SQL Connection is Freed
    Building Dimensions Elapsed Time : [10.641] seconds
    Reading Parameters For Database [Drxxxxxx]
    Declared Dimension Sizes = [109 165 15 80 1785 1938 1242 93 25 14 10 6 20 1502 21 5 211 2 ]
    Actual Dimension Sizes = [108 158 15 78 1785 1938 1241 93 25 14 10 6 20 1502 21 5 211 1 ]
    Network error [10054]: Cannot Send Data
    Network error [10054]: Cannot Send Data
    Unexpected Essbase error 1042012
    Object [IBasic] unlocked by user [dhanjit]
    regards,
    Dhanjit

  • How to load data only at the upper level in ASO

    We are loading the Days member with values 31/30/29/28 for the top level combinations in Block storage cube using a rule file.<BR>Now we need to carry out the same thing in ASO but ASO has a constraint that we can load data only at zero level.<BR>If we do that, then data at higher levels, after aggregation is incorrect. What can be done ?

    Also possible but a little more complicated is to create a separate cube with a DayCount account and only accounts and time dimensions. The time dimension goes down to day. You load it with 1 record per day. Then, you use this cube and your original cube as source, transparent partitions to a BSO cube. <BR><BR>This way, you have no bogus "input" members anywhere. Formulas that combine DayCount with other things would be written as outline formulas in BSO, since it doesn't support MDX. If you need those to be in MDX, then this wouldn't work for you.

  • Data load in Essbase ASO cube

    Hi,
    I have not been using ASO cube before and had worked only on BSO cubes. Now I have a requirement to create a rule file to load data in to an ASO Essbase cube. I have created a data load rule file as I was creating for a BSO cube which is correctly validating. However when I am doing the data load I am getting following warning:
    "Aggregate storage applications ignore update to derived cells. [480] cells skipped"
    I have investigated further and found that ASO cube does not allow data loading at upper levels & on members calculated through formulas. After this I have ensured that I am loading the data in to zero level members and members which are not calculated through formula. But still I am not able to do the data load & getting the same warning.
    Could you please help me and let me know if there is anything else which I am missing here?
    Thanks in advance...
    AKW

    Hi AKW,
    "Aggregate storage applications ignore update to derived cells. [480] cells skipped"This is only a warning message that means only those many cells were skipped might be for some reasons like any member pointing to those cells will be missing.
    If you want to copy the Data of your BSO cube to an ASO Application why dont you use an PARTIONING it will copy your whole data from BSO to ASO (If Outline is common in both then copy any member of Sparse dimension like "Scenario 1" from Source i.e. BSO, to same member like "Scenario 1" in Target i.e ASO ),
    This is only an alternate wayThanks
    Avneet Singh Bhatia

Maybe you are looking for

  • Forms Migration Assistant Usage

    Hi all, I am not able to understand the clear usage of Forms Migration Assistant yet. I have seen in Forums that Forms 6i applications can be migrated to 11g just by recompiling the forms and running it on 11g. If this is the case , whats the clear u

  • How to set  color of a  row depending on the value of column in JTable?

    Hi All, I have a JTable that add rows when the user clicks on the button. In this way there can be any no. of rows in my table. My table contains five columns. When a new row is added , it is added with new data each time. Also the data of the rows k

  • Getting "The Application Disk Utility quit unexpectedly"

    Hi all. New to the forum. Went to do my weekly disk Utility this afternoon and it won't open from my applications. Keep getting the error "The Application Disk Utility quit unexpectedly". Now This morning i did allow firefox to do an update, not sure

  • Need help with getStyleDeclaration

    Hi, I'm trying to move my Flex 3 project to new Flash Builder environ. I've a class named "CanvasButtonAccordionHeader" which is extended from an mx:Button instance. There was a code as: var selector:CSSStyleDeclaration = StyleManager.getStyleDeclara

  • Approval tasks showing incorrect approval by xelsysadm

    Hi! We discovered a strange behaviour with approval tasks and assignment adapters. We configured an Approval Process with three approval tasks. The first one is approved by the target user's manager. The second one is approved by a group, determined