Size of a Essbase cube

Hi All,
can anyone plz tell what are the parameters(like stored dimesions,rule file) we need to consider to estimate the size of the Essbase cube??
Is there any formula to calculate the same??
Regards,
Lolita

Have you had a look at the section in dbag that covers estimating - http://download.oracle.com/docs/cd/E17236_01/epm.1112/esb_dbag/dstsiz.html
Cheers
John
http://john-goodwin.blogspot.com/

Similar Messages

  • Essbase cube/pag file size reduction-bitmap compression

    We are seeing some huge pag files in our essbase cube, and Oracle had suggested changing to bitmap compression to see if that will help.(currently, we have 'no compression')
    That said, what is the difference between using RLE versus bitmap encoding for BSO cubes?
    Would like to hear other's experiences.

    (currently, we have 'no compression')^^^You are going to be very happy. Very, very happy.
    You can read the Database Administrator's Guide -- just search for "comrpession".
    Here it is: http://download.oracle.com/docs/cd/E17236_01/epm.1112/esb_dbag/dstalloc.html
    There are a bunch of caveats and rules re the "better" compression method -- having said that, I almost always use RLE as the method it seems the most efficient wrt disk space. This makes sense as this method examines each block and applies the most efficient method: RLE, bitmap, or Index Value pair. I can't see a calculation difference between bitmap and RLE.
    How on earth did you get into the position of no compression in the first place?
    Regards,
    Cameron Lackpour
    Edited by: CL on Jul 18, 2011 10:48 AM
    Whoops, a couple of things:
    1) After you apply the change in compression in EAS or via MaxL, you must stop and then restart the database for the setting to take affect.
    2) You must export the database, clear it out, and reload it using the new compression method -- the change in #1 only affects new blocks so you could have a mix of compressed and uncompressed bliocks in your db which won't do anything for storage space.
    3) The only way to really and truly know which method is more efficient is to benchmark the calculation (think writing to disk which could be a bit slower than it is now) and compression methods is to do it. Try it as bitmap and write down sample calc times and IND and PAG file size, then try it as RLE and do the same.

  • Essbase cube optimization

    Hi All,
    I know a lot has been discussed on optimizing Essbase cubes and others but need to get some more guidance on the same.
    Scenario:
    Essbase app 11.1.2.2 with 2 dense and 6 sparse as below:
    Account ~500 stored (dense)
    Period ~13 stored (dense)
    Fin Element ~2000 stored (sparse)
    Entity ~10000 stored (sparse)
    Prod ~10000 stored (sparse)
    * Year, Version and Scenario are non-aggregating dimensions
    Stats: Block size ~50kb and lev0 size ~400MB
    The issue is that there are few monthly end calc scripts that takes more than ~3-4hrs /script to execute on the cube.
    I had done the following things so far:
    1- Provide custom hour glass outline: Period-Accounts-Fin Element- Prod - Entity - Yr - Version -Scenario
    2- Changed few member properties
    3- Optimized script
    However the run time for the scripts is equal to what was earlier.
    A quick question:
    If I make dimension 'Fin element' as dense, block size is ~80MB. I guess 100KB is max recommended, but need to confirm on this.
    Kindly guide ahead as in what else should I try ahead and do..
    Regards..

    Hi
    This is all a little subjective....but....
    Are your upper level members of dense dims dynamic or are you having to run any CALC DIM commands against them within the lengthy BR's? If you are then these will probably be responsible for a fair chunk of the time - could always pull these out and run separately to see exactly how much time they take.
    JB

  • Size Limits to Essbase

    Hi,
    I am working on a Planning application and as most of you would now, it is based on a BSO Essbase cube.
    I wanted to know, if there are any limitations to the size of the dimensions? And if yes, then what is the limit on the number of members?
    For your reference, I have two dimensions that have roughly 10,000 to 15,000 members.
    At the moment, it takes time to deploy the application but otherwise, everything seems to work. I want to know if there is any limit to expansion so I can better prepare for that scenario. Thanks.
    Regards,
    Shehzad

    Please,
    check the limits here DBAG .
    Before you hit these limits, I assume that your hardware will no longer be able to run calculations (as DISK IO latency on per block operations increases linear).
    I always prefer designs that would only have the amount of detail, that is usable for analyzing major effects. Business people are happy when it takes the database to calculate all dependencies after changing a value in a few seconds.
    In 11.1.1.3 dense restructures are still single thread operations that issue IOs for each block (IOs > 10K). This is a llot of stress for our IO subsystems - thus restructures run at 2 MB/s. Therefore large databases consume too much time during maintenance operations.
    I managed to bring a 200GB database down to just 12 GB by:
    - removing things that were unimportant to the business
    - use dynamic calculation where possible (even for currencies)
    - split the database into three
    Don't get me wrong - if it's the business demand to have 100.000 members, thats fine - I just never found a REAL reason why this adds to a business process.
    Sascha.

  • Increased calc time in Essbase cube

    Hello-
    I have an Essbase cube that just recently started experiencing very long calc times. The cube has been in a 9.3.1 environment for approximately 4 months and there haven't been any major additions to the data. Just adding Actuals every month. I had a CalcAll running in approximately 2 minutes and now it's running in 2+ hours. This essentially happened overnight. The size of the page file is 267Mb and index file is 25Mb. The data cache and index cache are currently set at 400,000Kb and 120,000Kb respectively.
    This is a 7 dimensional cube with 2 dense dimensions... Accounts and Time. My block size is high due to the large amount of level 0 Accounts... 215,384Kb.
    The number of index page writes and data block reads & writes is pretty high... 128,214 - 3,809,875. And the hit ratio on the data cache is .07.
    I've tried adjusting the data cache and index cache both up and down... but with no success.
    Any help would be appreciated.
    Thanks!

    Here are a couple of things to think about
    How big is the databasename.ind file? How often do you restructure the database? if it is large, it means the database is fragmented (you can also look at the database stats to help figure it out) Every time you runn the calc all, you cause fragmentation.
    If you have agg missing off it also increases the calc time. Also, you will find that calculation a loaded cube takes longer than if you have a cube with just level zero data.
    If your database is not too big, you might try exporting level zero data, clearing the database, reloading the data and doing your calc all. (of course I would try this on a copy of the production database not the actual database itself). You might fined it quicker than your straight calc.
    Since you do a straight calc all, you might also consider converting this to an ASO cube, you really get rid of the call all in total.

  • Linking existing Essbase Cube to Planning

    My business team created an Essbase cube, and now they want us to link that cube to Planning.
    I opened an SR and I was told the following:
    1. Ensure you have a test environment to create this new Planning app before moving to production.
    2. Create a new Planning app in the test environment with the same name as the production Essbase cube.
    3. Export all Essbase data.
    4. Copy over the outline to the new Essbase cube in the test env. You can copy over the rest of the Essbase objects as needed (calc scripts, report scripts, etc).
    5. Import the Essbase data into the new Essbase cube.
    6. refresh Planning
    Everything went well except for Step 6. When I do the refresh from Planning, the Essbase Outline gets updated with the one from Planning which doesn't have anything since it was just created.
    I need to refresh the Planning Outline with the the one from Essbase that does have the data the users have been working on.
    Thanks.

    Hi,
    You will have to define all the dimension membes etc to Planning. Because Essbase is the database, it will be updated with planning information... There are no shortcuts.. But you can export your dimension members from existing Essbase and load into planning using HAL/ODI/DIM..
    Use outline extractor tool to get the existing dimensionality and create 1 metadata file per dimensions.
    I believe Oracle support thought that you want to create a new Essbase cube from existing Planning cube.
    Cheers
    RS

  • Unable to load data into Essbase cube using Essbase Studio

    Hi
    We are creating an essbase cube using Essbase Studio using flat files as data sources.
    We have taken different hierarchies into different flat files and created one fact file having dimension intersection along with data.
    We are able to create the cube and the hierarchy but not able to load any data.
    We are getting the following error
    Failed to deploy Essbase cube.
    Caused by: Unable to perform dataload from more than one flat file.
    Could anyone please help on this?

    Oh this was killing me, so I did this test in 11.1.1.3:
    1) Excel 2007 format -- no go, Essbase didn't see it
    2) Excel 2003 format, three sheets -- only the first sheet was read into an empty rule
    3) Excel 2003 format, one sheet -- the first sheet was read into an empty rule
    4) Excel 95 format, one sheet -- the first sheet was read into an empty rule
    The lesson?
    1) Excel 2007/2010 sheets don't work (no surprise there as the .xlsx format isn't supported).
    2) Excel 2003 and lower (hey, if you have Excel 4, I'll bet that works as well) work, but only the first sheet is recognized.
    Regards,
    Cameron Lackpour

  • Workspace error when drilling down on Essbase Cube

    An Interactive Reporting Service error has occurred.-Failed to acquire requested service.
    (2001)
    We're trying to create OLAP queries in IR to deploy through workspace. When I drill down on any dimension, after deploying to workspace, I get the generic error above. Has anyone seen this error?

    I have created one report using Essbase cubes in OBIEE11G and when I am archiving the same report at one local server and unarchiving it at some other server then at the other server I am facing this error.
    Any replies will indeed be helpful.

  • E-mail Alerts from Essbase Cubes

    Hi,
    I have Essbase 7.1.3 windows server and We do lots of Cube processing through Batch Files.
    I would like to have an option to send email alerts to the business users once we are done with the processing.
    Say for example , I have a Deposits Essbase cube and once i am done with the calculation and Enable connects to the users i would like to send a mail notification " The Deposit Cube now available....."
    Essbase have some thing called " Triggers..." But i am not sure how far will it help me.
    Looking forward for your inputs.
    Thnx,
    MS

    Okay, yes, I still use esscmd, so this example is for an esscmd script (some day I will convert them, I promise...).
    The shell (batch) script that I use calls esscmd to do the job, and then scans the script (using the for command) for tokens. These tokens determine who gets an email depending on the outcome of the script. By encoding the messaging into the script, I can control the business logic at an appropriate level. The TITLE token is also used during the notification, but I left the logic out that builds the SUBJECT variable (for brevity)...
    :TITLE Load Budget data to Daily Trends
    :Job Notifications:
    :ONERR [email protected]
    :ONERR [email protected]
    :ONSUCCESS [email protected]
    :ONSUCCESS [email protected]
    :ONSUCCESS [email protected]
    :ONSUCCESS [email protected]
    LOGIN "LocalHost" "<superuser>" "<superpassword>";
    ... etc.Here is the relevant part of the shell job that scans for tokens (the RESULT environment variable is set to either "ERR" or "SUCCESS" based on the output, with other variables set as needed):
    TYPE %SCRFILE% | FIND ":ON%RESULT%" >Notify.tmp
    FOR /f "tokens=2" %%a in (Notify.tmp) DO SendEmail -f %FROM% -t %%a -u "%SUBJECT%" -m "%MESSAGE%" -s %RELAY% -a %ATTACH%
    ERASE /Q Notify.tmp

  • Using ODI to load 2 similar Essbase cubes

    Hello all,
    I am trying to create an ODI routine that incorporates the following details:
    -Loads specific Essbase member properties from a single flat file to 2 similar Essbase cubes
    -The Essbase cubes are virtually identical, except that one is ASO and one is BSO, therefore the properties may differ for specific members (i.e. DataStorage, Consolidation, Formula, etc.). There are also other differences in dimensionality, but those details are inconsequential to this routine...
    -For each type of member property, the flat file has 2 sets of columns - one for ASO and one for BSO. Therefore, one column might be called "ASODataStorage" while another one is called "BSODataStorage". They have different values in them respective to the Essbase architecture.
    -There is also a set of columns called "ASOFlag" and "BSOFlag" that flag which members should even be loaded to that respective cube. So, one sub hierarchy may be needed for the ASO cube, but not the BSO cube.
    Earlier, I created 2 contexts for these cubes so I could use the same ODI routine and source metadata to update both cubes with a simple change of context. This part is working correctly. The challenge I'm facing now, however, is how to design a routine that will select the appropriate columns of data based on my context. Therefore, if one context is called "ASO", it should select the "ASO" columns and apply only those properties/members. And a similar situation should happen when the "BSO" context is created.
    I have tinkered with the idea of using a variable that stores the current context information, and becomes evaluated at runtime and then changes the source columns based on the variable value. However, I'm having trouble getting this to work correctly (I am a newbie to variables). I am also having trouble figuring out how to map all of the source columns to the target columns, as I am going from double property columns in the source to a single set of property columns in the Essbase metdata target.
    Is there a better way to approach this in ODI? Please note that I am aware that EIS can handle this as well...but I am looking for an ODI solution.
    Thanks in advance!
    -OS

    Hi,
    In simple terms you should be able to create a package, drag the variable on to the package and set the value as ASOflag
    Then drag your interface on to the diagram to run the ASO load (the interface should have a filter that uses the variable to filter the required records)
    Then drag the variable on to the diagram and set the value to BSOflag.
    Then drag your interface on to the diagram to run the BSO load (the interface should have a filter that uses the variable to filter the required records)
    Though you say you are changing contexts so you might have to go down the route of generating a scenario from a package for the ASO load, then a scenario for the BSO load, then create another package using the ODIstartscen tool and execute the scenarios with the specified scenarios.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • How to check the size of an info cube?

    Dear all
    Is there any t-code which can be used to see the total disk space occupied by a specific cube?
    I know ST14 can be used for the TOP 30 cubes and stuff.
    but my cube is not in top 30.....so how should i see the total size of my cube in bytes?
    please help me...i need it urgently
    Please dont recommend the SAP note on Sizing...i have it...the caliculation gave me wrong answer!
    Total Points assured!!!!
    Edited by: reddy reddy on Jan 27, 2008 9:34 AM

    Hi,
    we can calculate the size of the info cube.using this information
    Each keyfiure occupies 10 Bytes of memory
    Each Char occupies 6 Bytes of memory
    So in an Infocube the maximum number of fields are 256 out of which 233 keyfigure, 16 Dimesions and 6 special char.
    So The maximum capacity of a cube
    = 233(Key figure)10 + 16(Characteristics)6 + 6(Sp.Char)*6
    In general InfoCube size should not exceed 100 GB of data.
    While considering sizing, you must consider factors like years of data rewuired for reporting, acceptable repsonse time, database limitations, archiving options, etc.
    pl read this document:
    [https://websmp209.sap-ag.de/sizing]
    Hope this helps,
    regards
    CSM Reddy

  • Essbase Studio: Failed to deploy Essbase cube

    Hi
    I have started working with Essbase studio sometime back and I am able to deploy BSO cube with success using the TBCSample Database which comes along with Essbase. Now I wanted to deploy ASO cube, as no sample database is available I thought to create one, I extracted ASOSamp using ODI to CSV files. Then I bulk inserted the csv extracts into MSSQL 2003 server which created 11 tables (Age, Geography, IncomeLevel, Measures, PaymentType, Product, Stores, Time TransactionType, Year). The above mentioned table does not have any keys(Primary, Foreign) as it is an Essbase export.
    I then successful created ASO Cube Schema using the newly created sample database in MSSQL, validated cube schema without any errors.
    Essbase Property Setting:
    Measures Hierarchy is tagged as Dynamic Compression at dimension level
    Time, Product and Year Hierarchy is tagged as Multiple Hierarchies Enabled, Year does not have multiple hierarchies but it has formulas for Variance and Variance % member. Is there a way to tag Year as Dynamic hierarchy?
    But when I try to deploy the cube to Essbase I receive following errors:
    Failed to deploy Essbase cube
    Caused By: Cannot end incremental build. Essbase Error(1060053): Outline has errors
    \\Record #1 - Member name (Time) already used
    + S Time + S
    \\Record #6 - Member name (1st Half) already used
    MTD + S 1st Half + S
    \\Record #7 - Member name (2nd Half) already used
    MTD + S 2nd Half + S
    \\Record #21 - Member name (Qtr1) already used
    Qtr1 + S Feb + S
    \\Record #22 - Member name (Qtr1) already used
    Qtr1 + S Jan + S
    \\Record #23 - Member name (Qtr1) already used
    Qtr1 + S Mar + S
    \\Record #24 - Member name (Qtr2) already used
    Qtr2 + S Apr + S
    \\Record #25 - Member name (Qtr2) already used
    Qtr2 + S Jun + S
    \\Record #26 - Member name (Qtr2) already used
    Qtr2 + S May + S
    \\Record #27 - Member name (Qtr3) already used
    Qtr3 + S Aug + S
    \\Record #28 - Member name (Qtr3) already used
    Qtr3 + S Jul + S
    \\Record #29 - Member name (Qtr3) already used
    Qtr3 + S Sep + S
    \\Record #30 - Member name (Qtr4) already used
    Qtr4 + S Dec + S
    \\Record #31 - Member name (Qtr4) already used
    Qtr4 + S Nov + S
    \\Record #32 - Member name (Qtr4) already used
    Qtr4 + S Oct + S
    \\Record #33 - Member name (Time) already used
    Time + S MTD + S
    \\Record #34 - Member name (Time) already used
    Time ~ S QTD ~ S
    \\Record #35 - Member name (Time) already used
    Time ~ S YTD ~ S
    \\Record #9 - Error adding Attribute to member QTD(Jan) (3320)
    \\Record #9 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    QTD + S QTD(Jan) + S [Jan]
    \\Record #10 - Error adding Attribute to member QTD(Apr) (3320)
    \\Record #10 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    QTD ~ S QTD(Apr) ~ S [Apr]
    \\Record #11 - Error adding Attribute to member QTD(Aug) (3320)
    \\Record #11 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    QTD ~ S QTD(Aug) ~ S [Jul]+[Aug]
    \\Record #12 - Error adding Attribute to member QTD(Dec) (3320)
    \\Record #12 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    QTD ~ S QTD(Dec) ~ S [Oct]+[Nov]+[Dec]
    \\Record #13 - Error adding Attribute to member QTD(Feb) (3320)
    \\Record #13 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    QTD ~ S QTD(Feb) ~ S [Jan]+[Feb]
    \\Record #14 - Error adding Attribute to member QTD(Jul) (3320)
    \\Record #14 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    QTD ~ S QTD(Jul) ~ S [Jul]
    \\Record #15 - Error adding Attribute to member QTD(Jun) (3320)
    \\Record #15 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    QTD ~ S QTD(Jun) ~ S [Apr]+[May]+[Jun]
    \\Record #16 - Error adding Attribute to member QTD(Mar) (3320)
    \\Record #16 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    QTD ~ S QTD(Mar) ~ S [Jan]+[Feb]+[Mar]
    \\Record #17 - Error adding Attribute to member QTD(May) (3320)
    \\Record #17 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    QTD ~ S QTD(May) ~ S [Apr]+[May]
    \\Record #18 - Error adding Attribute to member QTD(Nov) (3320)
    \\Record #18 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    QTD ~ S QTD(Nov) ~ S [Oct]+[Nov]
    \\Record #19 - Error adding Attribute to member QTD(Oct) (3320)
    \\Record #19 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    QTD ~ S QTD(Oct) ~ S [Oct]
    \\Record #20 - Error adding Attribute to member QTD(Sep) (3320)
    \\Record #20 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    QTD ~ S QTD(Sep) ~ S [Jul]+[Aug]+[Sep]
    \\Record #36 - Error adding Attribute to member YTD(Jan) (3320)
    \\Record #36 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    YTD + S YTD(Jan) + S [Jan]
    \\Record #37 - Error adding Attribute to member YTD(Apr) (3320)
    \\Record #37 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    YTD ~ S YTD(Apr) ~ S [Qtr1]+[Apr]
    \\Record #38 - Error adding Attribute to member YTD(Aug) (3320)
    \\Record #38 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    YTD ~ S YTD(Aug) ~ S [1st Half]+[Jul]+[Aug]
    \\Record #39 - Error adding Attribute to member YTD(Dec) (3320)
    \\Record #39 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    YTD ~ S YTD(Dec) ~ S [1st Half]+[Qtr3]+[Qtr4]
    \\Record #40 - Error adding Attribute to member YTD(Feb) (3320)
    \\Record #40 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    YTD ~ S YTD(Feb) ~ S [Jan]+[Feb]
    \\Record #41 - Error adding Attribute to member YTD(Jul) (3320)
    \\Record #41 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    YTD ~ S YTD(Jul) ~ S [1st Half]+[Jul]
    \\Record #42 - Error adding Attribute to member YTD(Jun) (3320)
    \\Record #42 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    YTD ~ S YTD(Jun) ~ S [1st Half]
    \\Record #43 - Error adding Attribute to member YTD(Mar) (3320)
    \\Record #43 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    YTD ~ S YTD(Mar) ~ S [Qtr1]
    \\Record #44 - Error adding Attribute to member YTD(May) (3320)
    \\Record #44 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    YTD ~ S YTD(May) ~ S [Qtr1]+[Apr]+[May]
    \\Record #45 - Error adding Attribute to member YTD(Nov) (3320)
    \\Record #45 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    YTD ~ S YTD(Nov) ~ S [1st Half]+[Qtr3]+[Oct]+[Nov]
    \\Record #46 - Error adding Attribute to member YTD(Oct) (3320)
    \\Record #46 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    YTD ~ S YTD(Oct) ~ S [1st Half]+[Qtr3]+[Oct]
    \\Record #47 - Error adding Attribute to member YTD(Sep) (3320)
    \\Record #47 - Aggregate storage outlines only allow formulas in compression dimension or dynamic hierarchies.
    YTD ~ S YTD(Sep) ~ S [1st Half]+[Qtr3]
    \\Record #2 - Incorrect Dimension [Year] For Member [ParentName] (3308)
    ParentName Consolidation DataStorage MemberName Consolidation DataStorage Formula
    \\Record #1 - Member name (Promotions) already used
    S Promotions S
    \\Record #2 - Incorrect Dimension [Promotions] For Member [ParentName] (3308)
    ParentName DataStorage MemberName DataStorage
    \\Record #3 - Member name (Promotions) already used
    Promotions S Coupon S
    \\Record #4 - Member name (Promotions) already used
    Promotions S Newspaper Ad S
    \\Record #5 - Member name (Promotions) already used
    Promotions S No Promotion S
    \\Record #6 - Member name (Promotions) already used
    Promotions S Temporary Price Reduction S
    \\Record #7 - Member name (Promotions) already used
    Promotions S Year End Sale S
    \\Record #2 - Incorrect Dimension [Payment Type] For Member [ParentName] (3308)
    ParentName DataStorage MemberName DataStorage
    \\Record #2 - Incorrect Dimension [Transation Type] For Member [ParentName] (3308)
    ParentName DataStorage MemberName DataStorage
    \\Record #22 - Member name (Home Entertainment) already used
    Home Entertainment + S Home Audio/Video + S
    \\Record #23 - Member name (Home Entertainment) already used
    Home Entertainment + S Televisions + S
    \\Record #24 - Member name (Other) already used
    Other + S Computers and Peripherals + S
    \\Record #25 - Incorrect Dimension [Product] For Member [ParentName] (3308)
    ParentName Consolidation DataStorage MemberName Consolidation DataStorage
    \\Record #26 - Member name (Personal Electronics) already used
    Personal Electronics + S Digital Cameras/Camcorders + S
    \\Record #27 - Member name (Personal Electronics) already used
    Personal Electronics + S Handhelds/PDAs + S
    \\Record #28 - Member name (Personal Electronics) already used
    Personal Electronics + S Portable Audio + S
    \\Record #31 - Member name (All Merchandise) already used
    Products + S All Merchandise + S
    \\Record #32 - Member name (High End Merchandise) already used
    Products ~ S High End Merchandise ~ S
    \\Record #33 - Member name (Systems) already used
    Systems + S Desktops + S
    \\Record #34 - Member name (Systems) already used
    Systems + S Notebooks + S
    \\Record #18 - Error adding Attribute to member Digital Recorders (3320)
    Home Audio/Video + S Digital Recorders + S
    \\Record #36 - Error adding Attribute to member Flat Panel (3320)
    Televisions + S Flat Panel + S
    \\Record #37 - Error adding Attribute to member HDTV (3320)
    Televisions + S HDTV + S
    \\Record #8 - Incorrect Dimension [Income Level] For Member [ParentName] (3308)
    ParentName DataStorage MemberName DataStorage
    \\Record #1 - Member name (Geography) already used
    S Geography S
    \\Record #2 - Error adding member 27425 (3317)
    \\Record #2 - Aggregate storage outlines only allow any shared member once in a stored hierarchy, including prototype.
    A M F GREENSBORO - NC S 27425 S 336
    \\Record #3 - Error adding member 36310 (3317)
    \\Record #3 - Aggregate storage outlines only allow any shared member once in a stored hierarchy, including prototype.
    ABBEVILLE - AL S 36310 S 334
    \\Record #4 - Error adding member 29620 (3317)
    \\Record #4 - Aggregate storage outlines only allow any shared member once in a stored hierarchy, including prototype.
    ABBEVILLE - SC S 29620 S 864
    \\Record #5 - Error adding member 67510 (3317)
    \\Record #5 - Aggregate storage outlines only allow any shared member once in a stored hierarchy, including prototype.
    ABBYVILLE - KS S 67510 S 316
    \\Record #6 - Error adding member 58001 (3317)
    \\Record #6 - Aggregate storage outlines only allow any shared member once in a stored hierarchy, including prototype.
    ABERCROMBIE - ND S 58001 S 701
    \\Record #7 - Error adding member 42201 (3317)
    \\Record #7 - Aggregate storage outlines only allow any shared member once in a stored hierarchy, including prototype.
    ABERDEEN - KY S 42201 S 502
    \\Record #8 - Error adding member 21001 (3317)
    \\Record #8 - Aggregate storage outlines only allow any shared member once in a stored hierarchy, including prototype.
    ABERDEEN - MD S 21001 S 410
    \\Record #9 - Error adding member 39730 (3317)
    \\Record #9 - Aggregate storage outlines only allow any shared member once in a stored hierarchy, including prototype.
    ABERDEEN - MS S 39730 S 601
    \\Record #10 - Error adding member 28315 (3317)
    \\Record #10 - Aggregate storage outlines only allow any shared member once in a stored hierarchy, including prototype.
    ABERDEEN - NC S 28315 S 910
    \\Record #11 - Error adding member 79311 (3317)
    \\Record #11 - Aggregate storage outlines only allow any shared member once in a stored hierarchy, including prototype.
    ABERNATHY - TX S 79311 S 806
    \\Record #12 - Error adding member 79601 (3317)
    \\Record #12 - Aggregate storage outlines only allow any shared member once in a stored hierarchy, including prototype.
    ABILENE - TX S 79601 S 915
    \\Record #13 - Error adding member 79608 (3317)
    \\Record #13 - Aggregate storage outlines only allow any shared member once in a stored hierarchy, including prototype.
    ABILENE - TX S 79608 S 915
    \\Record #14 - Error adding member 79698 (3317)
    \\Record #14 - Aggregate storage outlines only allow any shared member once in a stored hierarchy, including prototype.
    Are these errors due to data source, if yes what could be possible work around?
    Is there any problem with Essbase properties which I have set if so then when I validate cube schema why I dn't get any errors?
    Please help me, I am stuck here not able to deploy ASO Cube.
    Thanks in advance

    Hii
    I have the same problem , you have.
    did you manage to solve it ??
    Thanks in advance

  • View Display Error in OBIEE with Essbase Cubes

    Hi All,
    Currently we are generating the Reports from Essbase Cubes.
    We have an hierarchy in OBIEE and when we are trying to drill down one hierarchy(Tech Executive) we are getting below Error.
    " Error
    View Display Error
    Odbc driver returned an error (SQLExecDirectW).
    Error Details
    Error Codes: OPR4ONWY:U9IM8TAC:OI2DL65P
    State: HY000. Code: 10058. [NQODBC] [SQL_STATE: HY000] [nQSError: 10058] A general error has occurred. [nQSError: 43113] Message returned from OBIS. [nQSError: 43119] Query Failed: [nQSError: 96002] Essbase Error: Unknown Member [H1_Rel_TecExec_Ini_appl].[Unknown] used in query (HY000)
    SQL Issued: SELECT s_0, s_1, s_2, s_3, s_4, s_5 FROM ( SELECT 0 s_0, "Defect7M"."H1_Rel_TecExec_Ini_appl"."H1 Initiative Name" s_1, "Defect7M"."H1_Rel_TecExec_Ini_appl"."H1 Tech Executive" s_2, SORTKEY("Defect7M"."H1_Rel_TecExec_Ini_appl"."H1 Initiative Name") s_3, SORTKEY("Defect7M"."H1_Rel_TecExec_Ini_appl"."H1 Tech Executive") s_4, "Defect7M"."Defect7M#1"."Defect7M - measure" s_5 FROM "Defect7M" WHERE ("Defect7M"."H1_Rel_TecExec_Ini_appl"."H1 Tech Executive" = 'Unknown') ) djm "
    Can someone assist me how to resolve this error
    Thanks,
    SatyaB

    Satya,
    Have you done anything to modify the essbase drill logic within your BMM?
    Remember when modeling essbase you should just try to use the defaults first to ensure that all works correctly the first time through. Then you can adjust any hiearchies, federate, etc.

  • Problems in accessing Essbase cube using Interactive Reporting studio

    Hi All,
    I have developed a report in Interactive studio for which Essbase is the source. When i try to process my query i am getting the below error
    "Request [Report] from user[username] was terminated since it exceeded the time limit"
    Any idea as to how the time limit can be increased at essbase side.

    I still can not resolve the problem. I am pulling up an attribute dimension in my report and that seems to be the root cause of the problem. When i remove this attribute dimension and generate the report with rest of the details the report renders without any issues.
    can this be becuase the attribute dimension is "Dynamic Calc and Store" in Essbase cube and IR does not support certain essbase features such as Dynamic attribute dimensions ? This is quiet urgent and hence would need a quick response from ya'll

  • Combining relation facts with dimensions from an Essbase cube

    Hi!
    I am having trouble combining relational measures (from EBS) with dimensions from an Essbase cube. The dimensions that we want to use for reporting (drilling etc) are in an Essbase cube and the facts are in EBS.
    I have managed to import both the EBS tables and the cube into OBIEE (11.1.15) and I have created a business model on the cube. For the cube I converted the accounts dimension to a value based dimension, other than that it was basically just drag and drop.
    In this business model I created a new logical table with an LTS consisting of three tables from the relational database.
    The relational data has an account key that conforms to the member key of the accounts dimension in the Essbase cube. So in the accounts dimension (in the BMM layer) I mapped the relational column to correct column (that is already mapped to the cube) - this column now has two sources; the relational table and the cube. This account key is also available in the LTS of my fact table.
    The content levels for the LTS in the fact table have all been set to detail level for the accounts dimension.
    So far I am able to report on the data from the fact table (only relational data) and I can combine this report with account key from the account dimension (because this column is mapped to the relational source as well as the cube). But if expand the report with a column (from the accounts dimension) that is mapped only to the cube (the alias column that contains the description of the accounts key), I get an error (NQSError 14025 - see below).
    Seeing as how I have modeled that the facts are connected to the dimension through the common accounts key, I cannot understand why OBIEE doesn't seem to understand which other columns - from the same dimension - to fetch.
    If this had been in a relational database I could have done this very easily with SQL; something along the lines of select * from relational_fact, dim_accounts where relational_fact.account_key=dim_accounts.account_key.
    Error message:
    [nQSError: 14025] No fact table exists at the requested level of detail
    Edit:
    Regards
    Mogens
    Edited by: user13050224 on Jun 19, 2012 6:40 AM

    Avneet gave you the beginnings of one way, but left out that a couple of things. First, you would want to do the export of level zero only. Second, the export needs to be in column format and third, you need to make sure the load rule you use is set to be additive otherwise the last row will overwrite the previouse values.
    A couple of other wats I can think of doing this
    Create a replicated partition that maps the 3 non used dimensiosn to null (Pick the member at the top of the dimension in your mapping area)
    Create a report script to extract the data putting the three dimensions in the page so they don't show up.
    Use the custom defined function jexport in a calc script to get what you want

Maybe you are looking for