Conditions Need to decided for ASO or BSO cube?

Hi
Could you please tell me on what conditions or business needs we need to decided whether we need to build an ASO or BSO cube?,
which dimension needs to be Label only or Stored or Accounts or Multiple Hierarchy Enabled property?
Which needs to be Dense & which needs to be Sparse if it is BSO cube?
Could you please let me with an example with above question that would be helpfull.
Thanks,
Raj

For dense data sets and Analytic databases ( databases with complex calculations ) and read/write databases we will go for BSO.
In BSO we can load high and apply the pushdown calculations.
For large dimensionality and sparse data sets and for read only database we will go for ASO.
In ASO we can load at leaf level only.

Similar Messages

  • Export - import between ASO and BSO cubes of different structure

    Hi everybody!
    I have a 'large' ASO cube with information for a period of several years.
    I need to be able to load a smaller slice (e.g. one quarter) from the 'large' cube into a BSO cube with a different (simpler) structure , perform calculations and then load the updated data back to the 'large' ASO cube.
    What's the best way to approach this (import - export)?
    Thanks in advance,
    Alex

    Hi GlennS,
    Hello and thank you again for the help.
    I'm running version 9.3.1
    The ASO cube has 12 dimensions plus 8 attribute dimensions.
    The BSO cube - 8 dimensions plus 2 attribute dimensions.
    Data size:
    The ASO cube is about 1 GB, about 16 M input level cells. Only about 10 % of it will be extracted and loaded into the BSO cube.
    The BSO cube is ~ 300 MB fully calculated. 300,000 existing blocks
    I prefer not to load the BSO cube from the same source as because ASO cube contains not only the source data, but also modifications - for example - allocations etc. I would like to keep these modifications and to be able to load (modified) data to BSO, make additional modifications and then 'save' the modified data back to the ASO.
    How should I approach loading data back to the ASO? I will need to 'replace' a slice of ASO cube - delete the 'old version' cells and load new version from the BSO cube.
    Thanks,
    Alex

  • Sharing complex substitution variable values between ASO and BSO databases

    We have ASO and BSO Essbase database member names with spaces in, and need to store some of these member names in substitution variables. However, this has to be done differently for ASO and BSO, due to calc script syntax requiring double quotes and MDX requiring square brackets. For example:
    ASO:
    &CurWeek value = Week 1
    MDX: [&CurWeek]
    BSO:
    &CurWeek value = "Week 1"
    Calc Script: &CurWeek
    As a result, the substitution variables cannot be shared between the ASO and BSO cubes, since the BSO variable value requires double quotes due to the space in the member name.
    Is there a way to get the above to work with both ASO and BSO? Can the double quotes be escaped in calc script syntax? Or can the double quotes be removed in the MDX formula?

    Hi TimG,
    Apologies for such a late reponse to this, genuinely haven't had a spare second to reply until now!
    Yes, I suspect a complex alias name may be the best solution here, and to remove the spaces from the actual member names.
    I was not aware of the latter part at all. My colleague has confirmed as much on this too - DBAG 11.1.2.1 pp117 & 118:
    "Note: If a substitution variable value is numeric or a member name starting with a
    numeral or containing the special characters referred to above is to be used both
    in MDX and non-MDX situations, create two substitution variables, one without
    the value enclosed in quotation marks and one with the value in quotation marks."
    "To ensure that a new substitution variable value is available in formulas, partition definitions,
    and security filters, stop and restart the application. All other uses of substitution variables are
    dynamically resolved when used."
    This last paragraph is the most concerning since we were planning to be able to update substitution variables values and then access the new values from calc scripts and formulae instantaneously. This quirk is unexpected and a little inconvenient. We may have to look at scheduling a change of substitution variable value overnight, followed by a stop/start of the app ready for the next day, and to work around needing to access changed values instantly.

  • ASO and BSO Transparent partition

    Hi
    I have couple of questions regarding transparent partition, please help me out,
    1) I have two ASO cubes and one empty BSO cube, do I need to run security (apply filters) on both ASO and BSO cubes?
    2) How should I give access to users? For example: what should I write as "..."identified by" ....
    TO 'VZB_ISu'.'VB_IS' AT "localhost"
    AS "....." IDENTIFIED BY "....."
    AREA '"2006","2005",@IDESC("Month"),@IDESC("Scenario"),@IDESC("Measure_Type"),@IDESC("Measures"),@IDESC("Accounts"),@IDESC("RegionCompany"),@IDESC("Company"),@IDESC("RegionOrganization"),@IDESC("Organization"),@IDESC("Affiliate"),@IDESC("Product"),@IDESC("Ledger_Code"),@IDESC("Region"),@IDESC("Country")';
    3) Are their specific ways to tune an empty BSO cube transparently linked to ASO cube? BSO cube has no data.
    Thank you
    Namita

    Hi,
    1) You only need to set security on the BSO db (I'm assuming the users will not be connecting directly to the ASO db). When you setup your partition you define what user will be used to login to the ASO database. When ever an end-user retrieves data over the partition it's the partition user that logs into the ASO database to retrieve the data. You only need to make sure that the user defined in the partition has sufficient rights to retrieve that data.
    2) You can either use EAS to create the filter or MaxL...this is the example for create/replace filter: create filter sample.basic.filt1 read on 'Jan, sales', no_access on '@CHILDREN(Qtr2)';
    3) Try an avoid too many dynamic calcs in the target. These will obviously affect the retrieval times.
    Gee

  • SSaudit for ASO cubes in 11.1.1.3?

    Hello Gurus -
    I was wondering if anyone has been able to get the SSaudit feature working on ASO cube on version 11.1.1.3 to get the .atx and .alg file generated?
    We tried but it doesn't seem to work for us. Also, the "send" activity itself doesn't seem to get captured in either apps log or essbase logs.
    I was wondering if there is a workaround or especial setting we need to do for ASO.
    Pls, throw some light if anyone know anything on it.
    Regards

    SSAudit is not supported for ASO cubes and unfortunatly neither is transaction logging.
    If you need to do something like this I suggest using Dodeca which can save sends in a relational table

  • I am a studio graphic designer needing to purchase a color jet printer to handle small business takes.  I am trying to decide on the Epson WorkForce Pro WP-4530 All-In-One Printer  OR the HP Officejet Pro Premium e-All-in-One (need second tray for envelop

    I am a studio graphic designer needing to purchase a color jet printer, for my iMac, to handle small business tasks. I am trying to decide between the Epson WorkForce Pro WP-4530 All-In-One Printer  OR the HP Officejet Pro Premium e-All-in-One (need second tray for envelopes).   
    Do you have any recommendations?

    I am a studio graphic designer needing to purchase a color jet printer, for my iMac, to handle small business tasks. I am trying to decide between the Epson WorkForce Pro WP-4530 All-In-One Printer  OR the HP Officejet Pro Premium e-All-in-One (need second tray for envelopes).   
    Do you have any recommendations?

  • Drill through Funtionality Across different cubes (ASO to BSO to ASO)

    Hi All,
       I have one complicated problem. I have one essbase ASO cube where the We load the account data on most granular level (Generation 7) and then it rolls up the data in generation 5 and that generation 5 data is pushed to another cube ( BSO ) for which generation 5 is in BSO cube which in turn push the data in ASO1 at generation 5.
    Now user want to log into ASO1 to see the data and want that if user need to see more granular level data (Generation 7), user can be able to see without getting logged out or loggin into another cube.
    Any solution or suggestion how it can achived.
    Cheers
    AB

    Hi Celvin,
    Yes you are right. The data is ASO rolls ups to become Level0 in ASO1. but there is mapping which needs to be done between ASO and ASO1 which is kinda i am worry about.
    Eg Profit centre in ASO is 11000 and 12000 mapped to  ASO1 at PC2078 (both level zero).
    Account in ASO(Level zero) rolls to Gen5 which mappes to Level zero in ASO1.
    i was thinking of making alternate hierarchy in ASO instead of using mapping and integration tool. (Since we are messing up ASO cube).
    in account and
    in this way the Drill will be good. and no data mapping required between ASO to BSO and ASO to ASO1.
    Please let me know your thoughts/input/suggestions on this solution.
    Cheers
    AB

  • Unix setting for ASO?

    In a Unix environment there are certain settings that by default aren't set to the max when essbase is configured.
    For instance, the STACK setting defaults to 8k and you can increase it to the max (64k) by placing
    ulimit -s 65536
    in the essbase owners .profile file.
    For BSO this helps run time on large calcs as the system doesn't need to page its stack entries.
    Is there something similar for ASO cubes?
    I tried increasing the stack to 64K, but it didn't make any difference on the retrieval time of a rather large (14 dimensions) ASO cube.
    TIA,
    Robert

    Dear Gajanan,
    Create asset class in OAOA. Assign account determination & screen layout, number range.
    You have a section down as Status of AUC. Select the second radio button of "Line item Settlement"
    Regards,
    Ajay

  • ASO Vs BSO

    Hi
    I have two cubes (metaoutlines) derrived from the same OLAP model. The only difference between the metaoutlines is one BSO (Block Storage Option) and other ASO (Aggregate Storage Option). They both loaded with the same number of records of members (dimension data) & data (fact data). When I queried the cubes, ASO cube returned data but BSO cube didn't. Anybody could tell me what's going on please?
    Regards
    Chandra

    Sean is right.
    ASO aggregates on the fly, so if you loaded Level 0 data and queried the top of the database, you would get data.
    If you want to see summarized data in the BSO cube, you need to create a calculation script, or run CALC ALL.
    If you look for any of your the records in your data file in both cubes, you should be able to find the record level data.

  • Performance Tuning Data Load for ASO cube

    Hi,
    Anyone can help how to fine tune data load on ASO cube.
    We have ASO cube which load around 110 million records from a total of 20 data files.
    18 of the data files has 4 million records each and the last two has around 18 million records.
    On average, to load 4 million records it took 130 seconds.
    The data file has 157 data column representing period dimension.
    With BSO cube, sorting the data file normally help. But with ASO, it does not seem to have
    any impact. Any suggestion how to improve the data load performance for ASO cube?
    Thanks,
    Lian

    Yes TimG it sure looks identical - except for the last BSO reference.
    Well nevermind as long as those that count remember where the words come from.
    To the Original Poster and to 960127 (come on create a profile already will you?):
    The sort order WILL matter IF you are using a compression dimension. In this case the compression dimension acts just like a BSO Dense dimension. If you load part of it in one record then when the next record comes along it has to be added to the already existing part. The ASO "load buffer" is really a file named <dbname.dat> that is built in your temp tablespace.
    The most recent x records that can fit in the ASO cache are still retained on the disk drive in the cache. So if the record is still there it will not have to be reread from the disk drive. So you could (instead of sorting) create an ASO cache as large as your final dat file. Then the record would already still be on the disk.
    BUT WAIT BEFORE YOU GO RAISING YOUR ASO CACHE. All operating systems use memory mapped IO therefore even if it is not in the cache it will likely still be in on the disk in "Standby" memory (the dark blue memory as seen in Resource Monitor) this will continue until the system runs out of "Free" memory (light blue in resource monitor).
    So in conclusion if your system still has Free memory there is no need (in a data load) to increase your ASO cache. And if you are out of Free memory then all you will do is slow down the other applications running on your system by increasing ASO Cache during a data load - so don't do it.
    Finally, if you have enough memory so that the entire data file fits in StandBY + Free memory then don't bother to sort it first. But if you do not have enough then sort it.
    Of course you have 20 data files so I hope that you do not have compression members spread out amongst these files!!!
    Finally, you did not say if you were using parallel load threads. If you need to have 20 files read up on having parrallel load buffers and parallel load scripts. that will make it faster.
    But if you do not really need 20 files and just broke them up to load parallel then create one single file and raise your DLTHREADSPREPARE and DLTHREADSWRITE settings. Heck these will help even if you do go parallel and really help if you don't but still keep 20 separate files.

  • Performance of Financial Reports against a ASO Cube to a BSO Cube

    Hi All,
    I am working on Financial Reporting and Essbase. I wanted to understand and find some relevant documentation, which tells about the performance issues or difference when a financial report hits a BSO cube or an ASO cube.
    1. If there is a difference in the performance for an ASO vs BSO for Financial Reports, where can I find the document or details for it?
    2. If there is a difference in the performance for an ASO vs BSO for Financial Reports, what is the reason for the same?
    3. How can we improve the ASO performance for the reports?
    Any insights for the same, would be highly appreciated.
    Thanks
    Ankur Jain.

    Thanks Sean V,
    Its quite amazing for me as well, and that is why I am trying drill into any of the FR documentation which might contain something like this. As of now,since I don't have access to the cube, I have nothing to add, nor have the insights to the cube, based upon which I could explain, more on the Cube and Outline design.
    But as soon as I get the access, I will bring this back and have a discussion on the forum.
    Thanks for confirming the same, of what I had been explaining to my Client side as well. Though, I still need some basis to explain and prove them exactly and a prototype as well, which might be needed as well.
    Thanks,
    Ankur Jain.

  • ASO to BSO Transparent Partition problem

    Hi,
    I am doing a partition from ASO(Source) to BSO(Target). Outlines are same except one dimension:
    Aso cube
    dimension has 10 Level 0 members in Level 1.
    BSO cube
    Same dimension has 2 level 0 members(these 2 members are same as in ASo) in Level 1.
    Do rest 8 members be available while retrieving data in BSo cube? [Right now when i connect can only see 2 members]
    Was wondering with Transparent partition is it possible to see the additional outline members in BS ?.
    Any help to clarify would help thanks.

    I think i am understanding your question.
    A user will only "see" what members/dimensions are in the Target cube. You would need to add the other 8 members to the target outline to be able to pull the data for them from the source cube.

  • Outline optimization for ASO 7.1

    Hi,
    I've the time hierarchy which contains years,quarters,months and days. Now will it be a good idea to break the time hierarchy in 3 separate dimensions like year (which will contain the years only), time period (contain quater and months) and Days (1-31). It'll defntly reduce the member numbers in the outline since there is no repeated member. But will that help me to enhance the data load performance in the cube? And if yes, How?
    Thanks!

    Given that it is an ASO cube, there are no considerations for dense/sparse like there are with BSO cubes. My first thought is that you won't get much of a performance increase. You're right, you could reduce the number of members in the outline but usually the reason I need to use ASO is partly because I have tons of members in the Measures dimension, meaning that the number of members in Time is fairly trivial.
    The other thing you'd have to consider is that the way your users retrieve data from the cube would be entirely different if you change the dimensions. If you already have lots of reports out there or users that are accustomed to drilling into the cube in a certain way, changing the dimensions could be a very bad idea for that reason alone.
    Of course, the easiest way to find out if you get any data load performance gains is to try it. Are you using any of the other ASO loading tricks, like using more threads specified in your essbase.cfg file?
    Jason
    [jasons hyperion blog|http://www.jasonwjones.com]

  • Advice needed on monitor for Photoshop and Lightroom use

    Hi. I am aserious amature photographer wishing to move to the next level and sell some of my work.
    I just had a custom pc built to work with the new copies of Photoshop CS5 and Lightroom 3 I bought (lots or RAM and HD space, ssd, etc....)
    The last piece of my system is to purchase a monitor. I want to be somewhere in the better then Best Buy but less than NEC/Eizo range in price, or between $500 - $900. I have worked with cameras since the early 80's and moved to digital several years ago but the only post processing I have done is with Photoshop Elements.  I would be doing mostly prints to sell but also need to have a web site to do so. Will also use the pc for daily net surfing... but do not game or watch a lot of video on the pc.
    Being really new to this whole process I have a few questions.
    The first thing I need to decide is whether I need to look for a wide gamut display or not.
    I realize the whole chain must be 10 bit (Adobe -OS - graphics driver - graphics card - display port.
    I have Adobe Photoshop CS5 and Lightroom 3, Win7 64 bit, Zotac ZT 50701 10M video card (which uses GeForce GTX 560  fermi and an nvidia chipset. It does have displayport). I am having a hard time determining whether my video card actually supports wide gamut (10bit).
    Standard vs Wide Gamut? Is wide gamut important enough to deal with the issues it brings (calibration, viewing things other than PS and LR or color managed, which appear to be rare?) Is sRGB good enough for most prints (don't do fine art, mostly nature and portraits but starting to do some HRD things). If wide gamut is the way to go I have no problem with that and have the time to learn about calibration, color management, etc... But I also want to make sure the juice is worth the squeeze.
    24" vs 27"? Is there any advantage to one or the other when editing photos?
    IPS vs PLS? I realize they are similar but are there differences worth noting?
    Glossy vs Matte Anti-Glare? seems to be a lot of comments regarding the anti-glare coating, mostly poor. Yet I can see issues using a glossy screen in my study with a window to my back.
    One manufacturer vs another? I realize Eizo, NEC and LaCie are at the top of the heap. But with my budget, after upgrading my pc and camera equipment, I can't make that work now. So I need to choose from the next group down (Dell, HP, Samsung, Asus...)
    One or two monitors? It looks like many (mid-grade) wide gamut monitors do a lousy job of displaying anything but color managed sites. Is that necessarily true of all the mid-grades? Or can some be used for graphics but as well for routine net surfing, MS Office, etc...? Or am I better off getting two monitors, one for graphics and one for the rest? That would pretty much limit me to 24" or less given my budget (used to using a Dell 21" TN monitor that oddly crapped out just as my new pc was done).
    The more I read reviews the more confusing it gets. There seems to be a difference of opinion even among pros on whether to go wide gamut or stick with an easier sRGB. Realizing a standard gamut monitor would be cheaper, I do want to make the right decision up front, given my budget.
    The one thing I have found astounding is that there is nowhere to actually see many of the monitors I am considering. We live in Nashville TN but my wife is from Atlanta Ga so we drove there a few weeks ago to visit family and for me to visit monitor shops. Even the largest ones there (Fry's and Microcenter) had minimal IPS monitors, a few Dells and HP's. The knowledge of their sales folks was so poor I finally gave up. Felt bad about this until I posted this on another board and got a reply from a guy in LA (second largest city in the US) that he wanted to see a particular monitor and there was no place even there to do so.
    Anyhow, here is what I have considered:
    24" Wide Gamut: Dell U2410 and Asus PA246Q. Dells appear to be good IF you get a good one. The Asus appears to be a clone of the Dell that gets a lot of good press.
    27" Wide Gamut: Dell U2711 that also gets a ton of good reviews
    24: Standard Gamut: Dell U2412 and HP ZR2440.
    27" Standard Gamut: Samsung S27A850D and Apple Cinema- The Samsung uses PLS technology versus IPS while the Apple is a glossy screen that will work with a pc.
    Sorry for the long post. Any comments are greatly appreciated.

    dkg62 wrote:
    I realize the whole chain must be 10 bit
    Not trying to talk you out of setting up a 10 bit pipeline, but it's still not very mature, and it really isn't a necessity to get a good editing experience.
    Personally I find advantage in using two 4:3 ratio monitors for Photoshop work.  All my panels are on the right monitor, while pretty much the entire left one shows the Photoshop main window and the working canvas space.  My desktop is 3200 x 1200 pixels overall, and I find having the panels remain visible all the time is important.
    Regarding whether a wide gamut is important...  Will you be printing to devices that deliver a wide gamut?  What other things will you be doing with your system?
    It's not a no-brainer whether a wider gamut monitor is always "better" for everything, since it can accentuate the differences between the output from color-managed and non-color-managed applications, and it's definitely true that not everything is color-managed.  With a monitor that's close to sRGB, for example, you might find Internet Explorer output acceptable, while using a wide gamut monitor will result in garishly oversaturated IE displays.  On the other hand, FireFox (with a settings tweak) seems to get color management right, so there is an alternative.
    I think, as John has implied above, you should work to get your head completely around how color-management works, soup to nuts.  If you don't, there will always be things that are a mystery or which surprise you at the wrong times.  Being able to order a print and have it come back with the expected color can be very important, as you might imagine.
    -Noel

  • Need a code for Customer exit for extractor 0WBS_ELEMT_ATTR

    Hi Guys,
    I need a code for following requirement.
    I have appended some fields to standard extractor 0WBS_ELEMT_ATTR which normally takes data from PRPS table. But the new fields will be getting data from PRTE and PROJ table. with common keys. Following is my code which is not working. Please help me out, I am not good in abap. Please note some lines are commented.
    Thanks,
    H
    CODE:::::
    form 0wbs_elemt_attr  tables i_t_data structure biw_prps.
      data: it_wbs_prps like biw_prps occurs 0.
      data wa_prps like it_wbs_prps.
    data: begin of it_wbs_prps occurs 0,
             PSPNR like prps-pspnr,
             POSID like prps-posid,
             PBUKR like prps-PBUKR,
             IZWEK like prps-IZWEK,
             USR10 like prps-USR10,
             USR00 like prps-USR00,
             USR02 like prps-USR02,
             OBJNR like prps-OBJNR,
           end of it_wbs_prps.
    data wa_prps like it_wbs_prps.
      data: begin of it_wbs_prps1 occurs 0,
              PSPNR like prps-pspnr,
              POSID like prps-posid,
            end of it_wbs_prps1.
      data wa_prps1 like it_wbs_prps1.
      data: begin of it_wbs_prte occurs 0,
              posnr like prte-posnr,
              PSTRT like prte-PSTRT,
              PENDE like prte-PENDE,
              ESTRT like prte-ESTRT,
              EENDE like prte-EENDE,
              ISTRT like prte-ISTRT,
              IENDE like prte-IENDE,
            end of it_wbs_prte.
      data wa_prte like it_wbs_prte.
      data: begin of it_wbs_proj occurs 0,
              PSPNR like proj-pspnr,
              PLFAZ like proj-plfaz,
              PLSEZ like proj-plsez,
            end of it_wbs_proj.
      data wa_proj like it_wbs_proj.
    refresh: it_wbs_prte, it_wbs_proj.
    *it_wbs_jest.
    if i_t_data[] is not initial.
    Get data from PRPS WBS Master Data
         select pspnr posid PBUKR IZWEK USR10 USR00 USR02 objnr
           into corresponding fields of table it_wbs_prps
           from prps
            for all entries in i_t_data
          where POSID = i_t_data-POSID.
         sort it_wbs_prps by posid.
    Get data from PRPS1 WBS Master Data
          select pspnr posid
           into corresponding fields of table it_wbs_prps1
            from prps
            for all entries in i_t_data
         where POSID = i_t_data-POSID.
        sort it_wbs_prps1 by posid.
    Get data from PRTE Scheduling Data for Project Item
         if it_wbs_prps1[] is not initial.
         select posnr PSTRT PENDE ESTRT EENDE ISTRT IENDE
           into corresponding fields of table it_wbs_prte
           from prte
            for all entries in it_wbs_prps1
          where posnr = it_wbs_prps1-pspnr.
         sort it_wbs_prte by posnr.
         endif.
    Get data from PROJ Project Master Data
         if it_wbs_prps1[] is not initial.
         select pspnr PLFAZ PLSEZ
           into corresponding fields of table it_wbs_proj
           from proj
           for all entries in it_wbs_prps
          where pspnr = it_wbs_proj-pspnr.
         sort it_wbs_proj by pspnr.
         endif.
    loop at i_t_data.
       clear: wa_prps1, wa_prte, wa_proj.
       read table it_wbs_prps into wa_prps with key
                  posid = i_t_data-posid binary search.
       read table it_wbs_prte into wa_prte with key
                  posnr = wa_prps1-pspnr.
       read table it_wbs_proj into wa_proj with key
                  pspnr = it_wbs_proj-pspnr.
             i_t_data-zzPBUKR = wa_prps-pbukr.
             i_t_data-zzIZWEK = wa_prps-IZWEK.
             i_t_data-zzUSR10 = wa_prps-USR10.
             i_t_data-zzUSR00 = wa_prps-USR00.
             i_t_data-zzUSR02 = wa_prps-USR02.
              i_t_data-zzPSTRT = wa_prte-PSTRT.
              i_t_data-zzPENDE = wa_prte-PENDE.
              i_t_data-zzESTRT = wa_prte-ESTRT.
              i_t_data-zzEENDE = wa_prte-EENDE.
              i_t_data-zzISTRT = wa_prte-ISTRT.
              i_t_data-zzIENDE = wa_prte-IENDE.
              i_t_data-zzPLFAZ = wa_proj-PLFAZ.
              i_t_data-zzPLSEZ = wa_proj-PLSEZ.
        modify i_t_data.
    endloop.
    endif.
    endform.                    " 0wbs_elemt_attr

    Hi,
    it_wbs_prps1 .
    If you are going to use this internal table to check for a null condition, shouldn't you first assign the data package to it?
    For example,   it_wbs_prps1 = i_t_data[].
    Because you are now using  it_wbs_prps1 to select data from a table and it doesn't have any data.
    Please check.
    -RMP
    Edit : Please ignore this. I didn't see that you are actually filling the internal table. Sorry!
    Edited by: RMP on Oct 18, 2010 12:16 PM

Maybe you are looking for