Help me in Conerting BSO to ASO cube

IF(@ISLEV("Time Service",0)) @SUMRANGE("# of Requests",@CURRMBRRANGE("Time Service",LEV,0,,0))/"# of Requests"->"Time Service"; ELSEIF(@ISLEV("Time Service",1)) @SUMRANGE("# of Requests",@CURRMBRRANGE("Time Service",LEV,1,,0))/"# of Requests"->"Time Service"; ELSEIF(@ISLEV("Time Service",2)) 1; ELSE #Missing; ENDIF;
ASO ONE IS
CASE
WHEN
IsLevel ([Time Service], 0) Then
Sum([ # of Requests], CurrentMemberRange([TimeSerive],[Lev],0,,0)]/[# of Requests],[Time Serive]
When
IsLevel [TimeService], 1) Then
Sum ([# of Requests], CurrentMemberrange([TimeService],[Lev]1,,0)/[# of Request],[Time Service]
When
IsLevel([Time Service],2) 1 Then
[# of Missing]
End
cAN ANY ONE HELP IN FIXING ERRORS PLEASE.

Try this instead. This should be correct for syntax. See if it returns the correct result.
CASE WHEN ISLEVEL([Time Service].CurrentMember, 0) THEN
SUM(CROSSJOIN({[# of Requests]},
{OPENINGPERIOD([Time Service].levels(0), [Time Service]) : [Time Service].CurrentMember})) /
([# of Requests], [Time Service])
WHEN ISLEVEL([Time Service].CurrentMember, 1) THEN
SUM(CROSSJOIN({[# of Requests]},
{OPENINGPERIOD([Time Service].levels(1), [Time Service]) : [Time Service].CurrentMember})) /
([# of Requests], [Time Service])
WHEN ISLEVEL([Time Service].CurrentMember, 2) THEN 1
ELSE MISSING
END

Similar Messages

  • BSO to ASO Partitioning Help

    Can anyone help me with a couple questions in regards to our partitioning strategy from our Block Storage planning cube to our Aggregate storage planning cube in 11.1.2? Our original approach was to use a transparent partition for planning scenario (Budget, Forecast) data from BSO source to ASO target for level-zero only. ASO would handle aggregations. Also loading Actuals data directly into ASO cube from source system. We would like to have additional dimensions not required for Planning to be stored just in the ASO database in order to minimize database size and calc times on the BSO side. Two questions:
    1) We've found that the Actuals scenario data can not be queried in the ASO target (return #missing). Is this standard behavior for a transparent partitioned ASO target in 11.1.2?
    2) Regardless of partition type (Replicated or Transparent) can we have a different number of dimensions in our BSO and ASO cubes? Again my understanding is no, but wanted to confirm.
    Thanks in advance for any help that can be provided.

    data that is loaded directly to an ASO cube that is the target of a transparent partition is not visible^^^Isn't this the key? I read that the natively loaded data is overwritten (not really, more like overpartitioned) by the partition definition and the partition takes precedence.
    So the answer then would be to not have that conflict.
    Unless Tom was talking about something else/I'm misunderstanding.
    Regards,
    Cameron Lackpour
    Edited by: CL on Mar 10, 2011 8:13 AM
    Whoops, missed this part:
    This is the case even when the intersection of the data is outside the area defined by the partition.I still wonder if there's an overwrite of intersections going on.

  • ASO Cube with BSO Partition as a Target

    Hi,
    Can someone please explain me the following because after reading different blogs as well as some other Essbase documents I am not able to understand,
    What is the actual use of having a BSO Transparent partition (target) of an ASO cube (source)?
    How exactly does the Write-Back functionality works if I have a BSO Transparent partition (target) of an ASO cube (source)?
    And lastly, in what business scenario one would implement a BSO Transparent partition (target) of an ASO cube (source)?
    Your help is much appreciated.
    thanks,
    fikes

    I have a situation where I have a BSO (target) on top of an ASO (source).  This was to try and resolve the dynamic time series (where time and periods are in different dimensions).  The DTS works almost great in the BSO partition.
    The other problem (or why we wanted to go to ASO) was that BSO cubes do not handle attributes across partitions (we have quarterly data in transparent partitions) to the target cube. -- but we do have dynamic time series.
    What I'm facing now it that the attributes work in the BSO (target) except when using a dynamic time series.
    The attributes are defined in both the BSO and ASO cubes.
    Any suggestions on how to handle attributes AND dynamic time series ?

  • Automatic outline sync between a BSO & ASO cube

    We have a BSO and ASO cube connected through replicated partition.The partition & data update seems to work fine.Our next objective is to sync any changes made to BSO outline to the ASO outline automatically.
    Can anyone provide any pointers as to how to achieve this?

    Agreed, hence the suggestion of staging the data. This way you create a comprehensive set of properties covering (in an EPMA-esque way) all properties required for BSO, ASO, Planning, HFM etc... as necessary; and then using views limited to the required fields for each application to generate files in the correct format for the build of its particular app. E.g. Simple parent child tables with Essbase properties for use with Maxl and Essbase load rules, (different for BSO and ASO), slightly differently formatted parent child files for use with command line utility for Planning..etc..
    Edited by: user0124 on Sep 28, 2012 11:05 AM

  • Partitioning BSO to ASO to get around FDM limitations

    FDM only currently loads to BSO Essbase cubes. However, one of our FDM-loaded BSO cube is far better suited to being ASO, not least because it has:
    - Twelve dimensions.
    - A huge "customer" hierarchy (over 40,000 members), which is purely aggregating.
    - Most of the other dimensions are also purely aggregating (although there are a few complex calcs on specific members).
    The sheer size of the BSO cube outline and the data being loaded results in prohibitive calculation times, so we are looking for a workaround.
    Thoughts at present are to create an ASO cube (effectively copying and converting the current outline to ASO using the Wizard), create a replicated partition to transfer data from the BSO to ASO cube (mostly from level-0 members) and to use the BSO cube purely as a data-entry point. Users would then report from the ASO cube, which could be refreshed from the BSO cube when required.
    This would then mean that upper-level members and any calculations in the BSO cube are pretty much redundant.
    Does anyone agree with, or have any reservations on the above idea? Is there potentially a better method here? Also, could anyone please hazard a guess on how long the replicated partition refresh might take?

    Is the BSO cube only loaded from FDM / other loads? Or are users entering data? Is it an Essbase or Planning app?
    While it is true that there isn't a FDM adapter to load into ASO, you do not have to go from a BSO cube to the ASO cube for the data loads. Just have FDM export to a flat file and load that into ASO using maxl / batch script.
    Deanna Sunde
    Sr. Director
    The Hackett Group

  • BSO to ASO Formula Conversion - Help Needed

    ASO is not my forte...I am trying to get a simple formula to work in ASO cube as part of BSO->ASO conversion.
    If it is level 0 entity, the formula is (Last Shipped (PU) * Part Volume)...works fine.
    If its upper level, it should just sum the values of its children. The BSO equivalent that we typically use is @SUM(@CHILDREN(@CURRMBR("Entity"))).
    I attempted the same in the ASO MDX script as shown below, it is not yielding me the correct result....any thoughts why this is not working?
    Thanks in advance.
    Nima
    Member Formula*
    CASE WHEN ISLEVEL(Entity.CurrentMember, 0) THEN+
    Last Shipped (PU) x Part Volume+
    ELSE+
    SUM(Children(Entity.CurrentMember))+
    END+
    Edited by: Nima.V on Nov 12, 2008 7:12 PM

    Have you tried specifying which measure you want it to sum?
    e.g.
    CASE WHEN ISLEVEL(Entity.CurrentMember, 0) THEN
    Last Shipped (PU) x Part Volume
    ELSE
    SUM(Children(Entity.CurrentMember), [Measure].[Part Volume])
    END
    also I'm not sure whether the SUM is correct. This might work as it definately specifies the tuples:
    SUM({Products.CurrentMember.children},[Measures].[Part Volume])
    hope it helps,
    Gee

  • Need a process to update BSO Planning Web form into an ASO Cube

    Need a process to update information from the BSO Planning Web form into an ASO Report Cube every (5) five minutes.
    Please advise

    Thanks here is Part 2 of my questions
    Details:
    User Input data into a web form for Intersection are the following:
    Dim1,Dim2,Dim3,Dim4 , Account AAAA, 1000
    Dim1,Dim2,Dim3,Dim4 , Account BBBB, 2000
    Dim1,Dim2,Dim3,Dim4 , Account TTTT, 4000
    Dim1,Dim2,Dim3,Dim4 , Account ZZZZ, 7000
    Design Question:
    In the ASO Reporting Cube
    AAAA and BBBB are both at level 0 in both cube
    TTTT and ZZZZ are being loaded to a parent which is Level 1
    MY Issue with the Design Question  ???
    ASO Cubes can only be load at a level 0.
    So how can I complete this task, what are my option:
    My thinking process is the XRef Calc Script could push data the cube, than by built a maxl scripts that  would run the Calc script and put this scripts into some kind of system scheduler running the Maxl Scripts Every  (5) Mintues
    or am i way off base here ????
    Please advise
    Thanks in Advance

  • Drill through Funtionality Across different cubes (ASO to BSO to ASO)

    Hi All,
       I have one complicated problem. I have one essbase ASO cube where the We load the account data on most granular level (Generation 7) and then it rolls up the data in generation 5 and that generation 5 data is pushed to another cube ( BSO ) for which generation 5 is in BSO cube which in turn push the data in ASO1 at generation 5.
    Now user want to log into ASO1 to see the data and want that if user need to see more granular level data (Generation 7), user can be able to see without getting logged out or loggin into another cube.
    Any solution or suggestion how it can achived.
    Cheers
    AB

    Hi Celvin,
    Yes you are right. The data is ASO rolls ups to become Level0 in ASO1. but there is mapping which needs to be done between ASO and ASO1 which is kinda i am worry about.
    Eg Profit centre in ASO is 11000 and 12000 mapped to  ASO1 at PC2078 (both level zero).
    Account in ASO(Level zero) rolls to Gen5 which mappes to Level zero in ASO1.
    i was thinking of making alternate hierarchy in ASO instead of using mapping and integration tool. (Since we are messing up ASO cube).
    in account and
    in this way the Drill will be good. and no data mapping required between ASO to BSO and ASO to ASO1.
    Please let me know your thoughts/input/suggestions on this solution.
    Cheers
    AB

  • Performance of Financial Reports against a ASO Cube to a BSO Cube

    Hi All,
    I am working on Financial Reporting and Essbase. I wanted to understand and find some relevant documentation, which tells about the performance issues or difference when a financial report hits a BSO cube or an ASO cube.
    1. If there is a difference in the performance for an ASO vs BSO for Financial Reports, where can I find the document or details for it?
    2. If there is a difference in the performance for an ASO vs BSO for Financial Reports, what is the reason for the same?
    3. How can we improve the ASO performance for the reports?
    Any insights for the same, would be highly appreciated.
    Thanks
    Ankur Jain.

    Thanks Sean V,
    Its quite amazing for me as well, and that is why I am trying drill into any of the FR documentation which might contain something like this. As of now,since I don't have access to the cube, I have nothing to add, nor have the insights to the cube, based upon which I could explain, more on the Cube and Outline design.
    But as soon as I get the access, I will bring this back and have a discussion on the forum.
    Thanks for confirming the same, of what I had been explaining to my Client side as well. Though, I still need some basis to explain and prove them exactly and a prototype as well, which might be needed as well.
    Thanks,
    Ankur Jain.

  • Impacts on converting ASO cubes to BSO cubes

    <p>Hi,</p><p> </p><p>We want to convert 7 ASO cubes to BSO because we areexperimenting some bugs on ASO,</p><p> </p><p>what are the impacts of doing this ?</p><p> </p><p>Thanks in advance</p><p> </p><p>kbbft</p>

    Usually people are trying to convert BSO to ASO. It's interesting to see this coming full circle with new users of Essbase starting off with ASO and not having the BSO experience and now having to work backwards from there. I think it shows how powerful ASO is becoming.
    For your situation your MDX needs to be converted to Calc Script language. Tuples are represented using cross dimensional operators (->) and you need to end your statements with a semi colon (;).
    + Current Month vs Prior Month (+) Formulae : &CurMon->&CurYr - &PrMon->&PrMonYear;
    + Current Month vs Prior Month % (+) Formulae : ((&CurMon->&CurYr - &PrMon->&PrMonYear) /&PrMon->&PrMonYear) * 100;

  • Building BSO using ASO Rules file

    Hi All,
    We are building a new application where in we want to have both ASO and BSO. Eventually we wanted to do partition between these two cubes.
    We already have a fully functional ASO with data loaded into it. This ASO also has bunch of rules files using which the ASO cube was built.
    We need to build a new BSO and it should have same outline as ASO and in order to achieve this as used the same rules file that are in ASO. I already built the BSO usinng ASO rules files
    My issue is with the dimension which are deifned as Multiplw Heirarchy and I would appreciate if some one can help me how do i define them in BSO. Also if some one canhelp me understand what are additional changes that needs to be taken care when I am building BSO using ASO rules files
    I am trying to apply a caluclation which is in ASO and I am getting following error -- Error(1200337) - Error compiling formula for [Feb YTD] (line 1): [)] without [(]. As part of my research i found that error code is very generic. I checked the caluclation for missing ) or ( but everything looks good
    Any help is greatly appreciated
    Thanks
    Satish

    Your ASO formulas will not work in BSO cube. ASO uses MDX formulas and BSO uses Essbase calc script language. There is a guide in the technical reference that helps convert from one to the other.
    As far as multiple hierarchies, BSO does not have this concept. You should be able to load the hierarchies as is with perhaps a few changes to the load rule to address how to handle shared members.
    You will not have a direct one to one from BSO to ASO. They are a different architecture and as such have variances that are not 1 to 1 compatible.

  • Essbase ASO Cube query performance from OBI EE

    Hi all
    I have serious problems of performance when I query an ASO cube from OBI EE. The problem born when I implement a filter in some dimension of model in the Business Model and Mapping layer. This filter is to level-0 of the dimension, the values are obtained from a session variable in OBI EE. The objetive of this is apply filters depending of users. Then, for session variable I've a table in relational dabase base with relation between user and "access", then my dimensions (not all) have as level-0 the "access" of users (as duplicated members).
    The session variable in OBI EE is filled with row-wise option, so it has all values of "access" that correspond to user (:USER system variabe).
    When I query only by one of this filtered dimensions the respond is very fast, When I query for one of this filtered dimensions and a metric the respond is fast (10 seconds). But when I query for two of this filtered dimensions and metric the respond take 25 minutes. I checked Essbase app log and found this:
    +[Mon Nov 15 19:56:01 2010]Local/TestSec5/TestSec5/admin/Info(1013091)+
    +Received Command [MdxReport] from user [admin]+
    +[Mon Nov 15 20:28:28 2010]Local/TestSec5/TestSec5/admin/Info(1260039)+
    MaxL DML Execution Elapsed Time : [1947.18] seconds
    When I look the MDX query generated by OBI I see that the aggregation process is doing in the fly in the members filtered of the crossjoin of two dimensions:
    With
    set [CATALOGO_INSTITUCIONAL2] as '[CATALOGO_INSTITUCIONAL].Generations(2).members'
    set [CATALOGO_PRESUPUESTARIO2] as '[CATALOGO_PRESUPUESTARIO].Generations(2).members'
    *member [METRICAS_PRESUPUESTARIAS].[MS1] as 'AGGREGATE(filter(crossjoin (Descendants([CATALOGO_INSTITUCIONAL].currentmember,[CATALOGO_INSTITUCIONAL].Generations(7)),Descendants([CATALOGO_PRESUPUESTARIO].currentmember,[CATALOGO_PRESUPUESTARIO].Generations(7))),(([CATALOGO_INSTITUCIONAL].CurrentMember.MEMBER_ALIAS = "01.01" OR [CATALOGO_INSTITUCIONAL].CurrentMember.MEMBER_Name = "01.01")) AND (([CATALOGO_PRESUPUESTARIO].CurrentMember.MEMBER_ALIAS = "G" OR [CATALOGO_PRESUPUESTARIO].CurrentMember.MEMBER_Name = "G") OR ([CATALOGO_PRESUPUESTARIO].CurrentMember.MEMBER_ALIAS = "I0101" OR [CATALOGO_PRESUPUESTARIO].CurrentMember.MEMBER_Name = "I0101") OR ([CATALOGO_PRESUPUESTARIO].CurrentMember.MEMBER_ALIAS = "S01" OR [CATALOGO_PRESUPUESTARIO].CurrentMember.MEMBER_Name = "S01"))),METRICAS_PRESUPUESTARIAS.[Compromiso])', SOLVE_ORDER = 100*
    select
    { [METRICAS_PRESUPUESTARIAS].[MS1]
    } on columns,
    NON EMPTY {crossjoin ({[CATALOGO_INSTITUCIONAL2]},{[CATALOGO_PRESUPUESTARIO2]})} properties ANCESTOR_NAMES, GEN_NUMBER on rows
    from [TestSec5.TestSec5]
    Can somebody tell me if is possible to change the way in that OBI built the query or if is possible to use aggregations previously materialized of essbase?

    hi Amol,
    1. On what basis , did you estimate your cube to around 400GB to 600GB.
    2. If ASO is an option, its huge advantage lies in space, its does not take more space , unlike BSO.
    3. I have seen cubes ,who size was around 300-400GB in BSO,when made the same cube into ASO , its consumed space of 40GB-45GB.
    HOpe this helps
    Sandeep Reddy Enti
    HCC
    http://hyperionconsutlancy.com/

  • Problem working on member formulas in ASO cube

    Hi
    I was asked to convert a planning application BSO cubes into a single ASO cube.I have managed to convert one BSO Cube into ASO through EAS console and added the members in the other BSO cube throgh rulesfile.
    Now i have to write the member formulas for the level0 members of Account dimension.These are very simple formulas like
    if(@ISMBR("New_Seats"))
    "Assets Value"="Total Asset Cost";
    else
    "Assets Value"="Asset_Value";
    endif;
    and
    "Empty_Seats"=(("New Seat Additions"+"Available_Seats")-"Required_Seats");
    This is the first time i am working on ASO.I am getting this error while writing these formulas
    "Error(1260052) syntax error in input mdx query on line 1 at token '=' Empty_Seats..."
    Help me in writing these formulas and also in selecting proper member properties.

    The hierarchy is like
    Account(gen1)
    -> Statistical_Account(gen2)
    ->New Seat Additions(gen3)
    I have tried tthis and the formula is verified successfully
    *([Statistical_Account].[New Seat Additions]+[Statistical_Account].[Available_Seats])-([Statistical_Account].[Required_Seats])*
    if i use *([New Seat Additions]+[Available_Seats])-[Required_Seats]* it is not veryfying throwing some syntax error
    Thanks

  • Performance Tuning Data Load for ASO cube

    Hi,
    Anyone can help how to fine tune data load on ASO cube.
    We have ASO cube which load around 110 million records from a total of 20 data files.
    18 of the data files has 4 million records each and the last two has around 18 million records.
    On average, to load 4 million records it took 130 seconds.
    The data file has 157 data column representing period dimension.
    With BSO cube, sorting the data file normally help. But with ASO, it does not seem to have
    any impact. Any suggestion how to improve the data load performance for ASO cube?
    Thanks,
    Lian

    Yes TimG it sure looks identical - except for the last BSO reference.
    Well nevermind as long as those that count remember where the words come from.
    To the Original Poster and to 960127 (come on create a profile already will you?):
    The sort order WILL matter IF you are using a compression dimension. In this case the compression dimension acts just like a BSO Dense dimension. If you load part of it in one record then when the next record comes along it has to be added to the already existing part. The ASO "load buffer" is really a file named <dbname.dat> that is built in your temp tablespace.
    The most recent x records that can fit in the ASO cache are still retained on the disk drive in the cache. So if the record is still there it will not have to be reread from the disk drive. So you could (instead of sorting) create an ASO cache as large as your final dat file. Then the record would already still be on the disk.
    BUT WAIT BEFORE YOU GO RAISING YOUR ASO CACHE. All operating systems use memory mapped IO therefore even if it is not in the cache it will likely still be in on the disk in "Standby" memory (the dark blue memory as seen in Resource Monitor) this will continue until the system runs out of "Free" memory (light blue in resource monitor).
    So in conclusion if your system still has Free memory there is no need (in a data load) to increase your ASO cache. And if you are out of Free memory then all you will do is slow down the other applications running on your system by increasing ASO Cache during a data load - so don't do it.
    Finally, if you have enough memory so that the entire data file fits in StandBY + Free memory then don't bother to sort it first. But if you do not have enough then sort it.
    Of course you have 20 data files so I hope that you do not have compression members spread out amongst these files!!!
    Finally, you did not say if you were using parallel load threads. If you need to have 20 files read up on having parrallel load buffers and parallel load scripts. that will make it faster.
    But if you do not really need 20 files and just broke them up to load parallel then create one single file and raise your DLTHREADSPREPARE and DLTHREADSWRITE settings. Heck these will help even if you do go parallel and really help if you don't but still keep 20 separate files.

  • Data load in Essbase ASO cube

    Hi,
    I have not been using ASO cube before and had worked only on BSO cubes. Now I have a requirement to create a rule file to load data in to an ASO Essbase cube. I have created a data load rule file as I was creating for a BSO cube which is correctly validating. However when I am doing the data load I am getting following warning:
    "Aggregate storage applications ignore update to derived cells. [480] cells skipped"
    I have investigated further and found that ASO cube does not allow data loading at upper levels & on members calculated through formulas. After this I have ensured that I am loading the data in to zero level members and members which are not calculated through formula. But still I am not able to do the data load & getting the same warning.
    Could you please help me and let me know if there is anything else which I am missing here?
    Thanks in advance...
    AKW

    Hi AKW,
    "Aggregate storage applications ignore update to derived cells. [480] cells skipped"This is only a warning message that means only those many cells were skipped might be for some reasons like any member pointing to those cells will be missing.
    If you want to copy the Data of your BSO cube to an ASO Application why dont you use an PARTIONING it will copy your whole data from BSO to ASO (If Outline is common in both then copy any member of Sparse dimension like "Scenario 1" from Source i.e. BSO, to same member like "Scenario 1" in Target i.e ASO ),
    This is only an alternate wayThanks
    Avneet Singh Bhatia

Maybe you are looking for