Wrong date value in Essbase ASO cube

Hi All,
I'm trying to load a date value in mm-dd-yy format into an Essbase ASO cube. I'm using is a txt tab delimited file. The load rule is working fine. The outline properties is set with the proper format "mm-dd-yy". I loaded the data and when I retrieve the data using Smart View I see all the dates decreased by one day in my Smartview report.
Would you have any ideas why that is happening?
Thanks

this is a bug and fixed in 11.1.2

Similar Messages

  • Data load in Essbase ASO cube

    Hi,
    I have not been using ASO cube before and had worked only on BSO cubes. Now I have a requirement to create a rule file to load data in to an ASO Essbase cube. I have created a data load rule file as I was creating for a BSO cube which is correctly validating. However when I am doing the data load I am getting following warning:
    "Aggregate storage applications ignore update to derived cells. [480] cells skipped"
    I have investigated further and found that ASO cube does not allow data loading at upper levels & on members calculated through formulas. After this I have ensured that I am loading the data in to zero level members and members which are not calculated through formula. But still I am not able to do the data load & getting the same warning.
    Could you please help me and let me know if there is anything else which I am missing here?
    Thanks in advance...
    AKW

    Hi AKW,
    "Aggregate storage applications ignore update to derived cells. [480] cells skipped"This is only a warning message that means only those many cells were skipped might be for some reasons like any member pointing to those cells will be missing.
    If you want to copy the Data of your BSO cube to an ASO Application why dont you use an PARTIONING it will copy your whole data from BSO to ASO (If Outline is common in both then copy any member of Sparse dimension like "Scenario 1" from Source i.e. BSO, to same member like "Scenario 1" in Target i.e ASO ),
    This is only an alternate wayThanks
    Avneet Singh Bhatia

  • FDM to load data in Essbase ASO cube

    Anybody have used FDM to load data in Essbase ASO cube? How do you clear and run calc on ASO cube?
    Thanks

    Does the Essbase Adapter for FDM Support ASO Cubes? [ID 1168153.1]
    Modified 17-AUG-2010 Type HOWTO Status PUBLISHED
    Applies to:
    Hyperion Financial Data Quality Management - Version: 11.1.1.3.00 and later [Release: 11.1 and later ]
    Information in this document applies to any platform.
    Goal:
    Does the Essbase adapter for FDQM support ASO cubes?
    Solution:
    ASO cubes are not currently supported in FDQM.
    Unpublished Enhancement 6568323 has been created and it is currently under consideration for a future release.
    References
    BUG:6568323 - 8-529236080 - CUSTOMER WANTS TO TAKE ADVANTAGE OF THE ASO FUNCTIONS IN ESSBASE.
    Related
    Products
    Middleware > Enterprise Performance Management > Financial Data Quality Management > Hyperion Financial Data Quality Management

  • Accessing Essbase ASO Cube from Oracle Relational database

    Hi All,
    I am a Oracle database developer. We have a requirement where we need to access Hyperion Essbase ASO cube data directly from Relational Database. We have identified below options.
    1. Use Hyperion web service and UTL_HTTP oracle utility
    2. Use JAVA API to access ASO cube. The code of the Java, will be written in Informatica Java Transformation.
    Unfortunetly, i am not getting good resources in Google on how to do?
    Appreciate, if someone share the knowledge if they have implemented this.?

    I am not competent to recommend any particular approach but Essbase.ru has some blog entries on using XMLA / 11.1.2.2 services and a Google Code project...
    http://essbase.ru/archives/category/performance/essbase-api/xmla
    Google will translate if you don't read Russian!

  • Essbase ASO Cube query performance from OBI EE

    Hi all
    I have serious problems of performance when I query an ASO cube from OBI EE. The problem born when I implement a filter in some dimension of model in the Business Model and Mapping layer. This filter is to level-0 of the dimension, the values are obtained from a session variable in OBI EE. The objetive of this is apply filters depending of users. Then, for session variable I've a table in relational dabase base with relation between user and "access", then my dimensions (not all) have as level-0 the "access" of users (as duplicated members).
    The session variable in OBI EE is filled with row-wise option, so it has all values of "access" that correspond to user (:USER system variabe).
    When I query only by one of this filtered dimensions the respond is very fast, When I query for one of this filtered dimensions and a metric the respond is fast (10 seconds). But when I query for two of this filtered dimensions and metric the respond take 25 minutes. I checked Essbase app log and found this:
    +[Mon Nov 15 19:56:01 2010]Local/TestSec5/TestSec5/admin/Info(1013091)+
    +Received Command [MdxReport] from user [admin]+
    +[Mon Nov 15 20:28:28 2010]Local/TestSec5/TestSec5/admin/Info(1260039)+
    MaxL DML Execution Elapsed Time : [1947.18] seconds
    When I look the MDX query generated by OBI I see that the aggregation process is doing in the fly in the members filtered of the crossjoin of two dimensions:
    With
    set [CATALOGO_INSTITUCIONAL2] as '[CATALOGO_INSTITUCIONAL].Generations(2).members'
    set [CATALOGO_PRESUPUESTARIO2] as '[CATALOGO_PRESUPUESTARIO].Generations(2).members'
    *member [METRICAS_PRESUPUESTARIAS].[MS1] as 'AGGREGATE(filter(crossjoin (Descendants([CATALOGO_INSTITUCIONAL].currentmember,[CATALOGO_INSTITUCIONAL].Generations(7)),Descendants([CATALOGO_PRESUPUESTARIO].currentmember,[CATALOGO_PRESUPUESTARIO].Generations(7))),(([CATALOGO_INSTITUCIONAL].CurrentMember.MEMBER_ALIAS = "01.01" OR [CATALOGO_INSTITUCIONAL].CurrentMember.MEMBER_Name = "01.01")) AND (([CATALOGO_PRESUPUESTARIO].CurrentMember.MEMBER_ALIAS = "G" OR [CATALOGO_PRESUPUESTARIO].CurrentMember.MEMBER_Name = "G") OR ([CATALOGO_PRESUPUESTARIO].CurrentMember.MEMBER_ALIAS = "I0101" OR [CATALOGO_PRESUPUESTARIO].CurrentMember.MEMBER_Name = "I0101") OR ([CATALOGO_PRESUPUESTARIO].CurrentMember.MEMBER_ALIAS = "S01" OR [CATALOGO_PRESUPUESTARIO].CurrentMember.MEMBER_Name = "S01"))),METRICAS_PRESUPUESTARIAS.[Compromiso])', SOLVE_ORDER = 100*
    select
    { [METRICAS_PRESUPUESTARIAS].[MS1]
    } on columns,
    NON EMPTY {crossjoin ({[CATALOGO_INSTITUCIONAL2]},{[CATALOGO_PRESUPUESTARIO2]})} properties ANCESTOR_NAMES, GEN_NUMBER on rows
    from [TestSec5.TestSec5]
    Can somebody tell me if is possible to change the way in that OBI built the query or if is possible to use aggregations previously materialized of essbase?

    hi Amol,
    1. On what basis , did you estimate your cube to around 400GB to 600GB.
    2. If ASO is an option, its huge advantage lies in space, its does not take more space , unlike BSO.
    3. I have seen cubes ,who size was around 300-400GB in BSO,when made the same cube into ASO , its consumed space of 40GB-45GB.
    HOpe this helps
    Sandeep Reddy Enti
    HCC
    http://hyperionconsutlancy.com/

  • DATA LOAD WORNINGS IN ASO CUBES

    Hi Every one,
    While loading data into ASO cubes in Essbae we are getting wornings like *"Data load strem contains 1.25797e 08 and [0] #misssing cells ".*My data file have #missing values and o's and sepecial carecters like E .I want To load the complete data with out warnings.Kindly let me know if any one's know the solution .Whether i need to change any settings in rule file or how to ingnore those cells .
    Thanks,
    Vikram

    The warnings are really informational messages to let you know it loaded and did not load those values. Which is fine as they tend to bloat a cube (The zeros). #missing is not going to load anyway and the E is exponential format of numbers whinch should not be a problem. Excel will display it this way, but you can format it without the E. You don't mention if you are doing this from EAS or MaxL and what version you are on. In Veraion 11, in EAS there are options in the load dialog to turn on or off the loading of zzeros and missing across the top. In MaxL, I don't see the syntax in the Tech reference, but I thought it was there in 9.

  • Jobs and Tasks for Automation of load data in Essbase ASO Cube

    Hi, all people,
    my question is about creating jobs or tasks for automation loading data in essbase cubes.
    I use .bat file witch run MaxL script for loading data form oracle database into an essbase cube. For running .bat file is made a Windows Task.
    Is there any other opportunities to run .bat file not using windows tasks? Maybe there is special utilities in Oracle EPM Systems or any other.....
    I am using essbase version 11.1.1.2.0.
    Thanks for reply))

    There is no internal scheduler.You either have to use Windows Scheduler, Unix Cron or a third party tool. Take a look at Star Analytics Command center, it is designed specifically for Hyperion applications http://www.staranalytics.com/products/command_center.htm

  • Wrong data values displayed for YTD in reports

    Hi,
    In my reports when i try to select measures as YTD, eg. March 2010 YTD, i get wrong values for entities. But, apart from YTD, i get absolutely correct values if i select measures as anually, quarterly or monthly.
    I have checked ISBEGINNING property in time dimension and assigned values as 0 and 1 for the respective values. I have also checked if i have any wrong statements in dbo.measureformula in SQL server Management studio which might give wrong calculations but could not find any.
    Please guide me through.
    Many thanks in advance.
    Regards,
    Poonam

    HI Nilanjan,
    Thanks for your suggestions but i do not have any data in Apshell to test if it is being displayed correct or not.
    Please advice me further.
    Regards,

  • Wrong date value is inserting

    Hi All!
    When I call the procedure insertDate that inserting the date it inserts wrong value, sometimes the date is more than a year off, otherwise the value of new Timestamp(date.getTime()) is giving the right date at the moment of setting the parameter.
    Why is it happening and how to fix that?
    Here is the source:
    private static Date formatDate(int year, int mon, int mday, int hour, int min, int sec) {
    Date date = null;
    SimpleDateFormat format = new SimpleDateFormat("yyyy-MM-dd HH:mm:ss");
    String textDate = (1900 + year) + "-" + mon + "-" + mday + " " + hour + ":" + min + ":" + sec;
    System.out.println(textDate);
    try {
    date = format.parse(textDate);
    System.out.println(date.toString());
    } catch (ParseException ex) {
    Logger.getLogger(Main.class.getName()).log(Level.SEVERE, null, ex);
    return date;
    Date date = formatDate(111, 6, 29, 12, 59, 45);
    CallableStatement cs = connection.prepareCall("{ call insertDate(?)}");
    cs.the pasetTimestamp(1, new Timestamp(date.getTime()));
    Thank you!
    Regards,
    Astrid

    I'm using MySQL.
    If I store Timestamp mostly I got '0000-00-00 00:00:00' and sometimes something like '2013-11-08 15:09:46'.
    I've made changes in order to insert not Timestamp directly, but string. And it works properly by inserting exact date and time - '2011-07-01 08:29:33'.
    My procedure now looks like
    CREATE PROCEDURE `saveQueryResult`(date1 VARCHAR(50))
    BEGIN
         DECLARE d DATETIME;
         SET d = TIMESTAMP(date1);
         INSERT INTO values_ (submit_date)
         VALUES (d);
    END$$
    DELIMITER ;

  • Issue with federation between OBIEE 11.1.1.5 and Essbase ASO Cube 11.1.2.1

    Hi All,
    I am trying to retrieve relational attributes from a confirmed dimension to the Essbase Cube.
    However I am facing lot of issues/errors during rpd development.
    Both the Essbase Cube hierarchy and Table that contians attributes are from same relational source.
    Essbase 11.1.2.1 and OBIEE 11.1.1.5 are both cmpatabile versions.
    I read this blog and didn't got the solution
    http://www.rittmanmead.com/2009/11/oracle-bi-ee-10-1-3-4-1-essbase-connectivity-enriching-essbase-reports-with-relational-attributes/
    Can someone assist me with some better option to proceed further for horizontal federation
    Thanks,
    SatyaB

    You shouldn't have to restart Essbase as it should automatically sync, there may be additional information in SharedServices_Security_Client.log though it is probably worth logging with Oracle.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Loading data to Parent in ASO cube

    I created an Excel data file to test different load scenarios. One scenario was using a parent level in one of the dimensions. I expected the record to show up in the dropped records file but it doesn't. Does anyone know why it doesn't? Is there any way to tell that the record wasn't loaded? We are on v11.

    You get a warning message 1003055: "Aggregate storage applications ignore update to derived cells. [X] cells skipped".

  • Performance Tuning Data Load for ASO cube

    Hi,
    Anyone can help how to fine tune data load on ASO cube.
    We have ASO cube which load around 110 million records from a total of 20 data files.
    18 of the data files has 4 million records each and the last two has around 18 million records.
    On average, to load 4 million records it took 130 seconds.
    The data file has 157 data column representing period dimension.
    With BSO cube, sorting the data file normally help. But with ASO, it does not seem to have
    any impact. Any suggestion how to improve the data load performance for ASO cube?
    Thanks,
    Lian

    Yes TimG it sure looks identical - except for the last BSO reference.
    Well nevermind as long as those that count remember where the words come from.
    To the Original Poster and to 960127 (come on create a profile already will you?):
    The sort order WILL matter IF you are using a compression dimension. In this case the compression dimension acts just like a BSO Dense dimension. If you load part of it in one record then when the next record comes along it has to be added to the already existing part. The ASO "load buffer" is really a file named <dbname.dat> that is built in your temp tablespace.
    The most recent x records that can fit in the ASO cache are still retained on the disk drive in the cache. So if the record is still there it will not have to be reread from the disk drive. So you could (instead of sorting) create an ASO cache as large as your final dat file. Then the record would already still be on the disk.
    BUT WAIT BEFORE YOU GO RAISING YOUR ASO CACHE. All operating systems use memory mapped IO therefore even if it is not in the cache it will likely still be in on the disk in "Standby" memory (the dark blue memory as seen in Resource Monitor) this will continue until the system runs out of "Free" memory (light blue in resource monitor).
    So in conclusion if your system still has Free memory there is no need (in a data load) to increase your ASO cache. And if you are out of Free memory then all you will do is slow down the other applications running on your system by increasing ASO Cache during a data load - so don't do it.
    Finally, if you have enough memory so that the entire data file fits in StandBY + Free memory then don't bother to sort it first. But if you do not have enough then sort it.
    Of course you have 20 data files so I hope that you do not have compression members spread out amongst these files!!!
    Finally, you did not say if you were using parallel load threads. If you need to have 20 files read up on having parrallel load buffers and parallel load scripts. that will make it faster.
    But if you do not really need 20 files and just broke them up to load parallel then create one single file and raise your DLTHREADSPREPARE and DLTHREADSWRITE settings. Heck these will help even if you do go parallel and really help if you don't but still keep 20 separate files.

  • Ncome Statement & Balance Sheet Reporting via a single ASO cube

    Hi All,
    I wanted to get some perspective on industry best practices as it relates to performing Income Statement & Balance Sheet reporting via a single Essbase ASO cube. As both these areas share a lot of common dimensions, do most of the companies implement a single cube for integrated reporting or split cubes and have tools like HFR, Web Analysis etc. combine the information from those cubes for integrated cube.
    Appreciate any thoughts on this.
    Thanks!

    In 16 years of Essbase/Hyperion experience (14 of them consulting and training), I have seen just as many clients combining income statements and balance sheets in the same cube (BSO or ASO) as I have seen separate the two into individual cubes. Just as the BSO/ASO decision should factor in data volumes and hierarchy size, the combined/separate cube decision must come from good analysis of the situation at hand. Sometimes it is just a design preference for the company at hand.
    Your question seems weighted toward ASO, so I would encourage you to make sure you are using ASO for the right reasons. I reserver ASO for cubes where the current or planned amount of history carried will be massive and/or when a very large hierarchy is required. Otherwise, I prefer to have the full flexibility of BSO.
    One favorite project was for a re-insurance company in Bermuda (you can probably guess some of the other reasons I considered it a favorite project!?!). The consolidated P&L and B/S were in the same cube. This allowed us to properly "connect" the two statements such that retained earnings at the end of a quarter could flow over to hit the balance sheet. Of course, a two-cube design wouldn't necessarily prevent this thanks to @XREF (which didn't exist back in those days) or partitioning to name a couple of alternative. This makes my point that, to some extent, it all comes back to a matter of design preference for you and your project team.
    Darrell Barr

  • Building Drill through reports from aso cube using odi in webanalysis11.1.3

    I need some urgent help as i have a important requirement for an essbase aso cube .I was trying to establish a database connection to build a drill through report in webanalysis from oracle datawarehouse.the integration tool which we are using to get data from oracle datawarehouse to hyperion is odi.My question are
    1) Can drill through reports be built using odi like eis from an aso cube?if yes whats the procedure?
    2) is there any alternate ways to bring transaction level or relational data for reporting in webanalysis?
    regards,
    praveen.
    Edited by: user13070887 on Oct 11, 2010 3:48 PM

    Hi Glenn,
    We tried optimizing the drill through SQL query but actually when we run the directly in TOAD it takes *23 secs* but when we do drill through on the same intersection
    it took more than 25 mins. Following is our query structure :
    (SELECT *
    FROM "Table A" cp_594
    INNER JOIN "Table B" cp_595 ON (cp_594.key = cp_595.key)
    WHERE (Upper(cp_595.*"Dim1"*) in (select Upper(CHILD) from (SELECT * FROM DIM_TABLE_1 where CUBE = 'ALL') WHERE CONNECT_BY_ISLEAF = 1 START WITH PARENT = $$Dim1$$ CONNECT BY PRIOR CHILD = PARENT UNION ALL select Upper(CHILD) from DIM_TABLE_1 where CUBE = 'ALL' AND REPLACE('GL_'||CHILD, 'GL_IC_', 'IC_') = $$Dim1$$))
    And ----same for 5 more dimensions
    Can you suggest some improvement ? Please advice.
    Thanks

  • Format of Date Values in a Condition...Please Help!

    Hi,
    I just downloaded Oracle 11g XE and built an Oracle 11g XE database and is using the Application Express included to build simple reports. I used the Query Builder inside APEX and entered a condition on a Date column as BETWEEN '01-SEP-2011' AND '01-SEP-2012' and it gave me the error of the wrong data value, something like DATE expected. The actual values of that column are such as, 01-DEC-02, and there are no issues with the data column.
    This method has been working with the Query Builder in Oracle 10g. The generated SQL in this APEX is the same as the generated SQL in the query builder of 10g.
    I also tried different things, such as using 11 instead of 2011, removing the single quotes, etc. None has worked.
    Can someone tell me what is the correct way to key in a condition when the user wants to specify a date range?
    Thanks!

    Can you try like BETWEEN TO_DATE('01-SEP-2011', 'dd-mm-yyyy')?

Maybe you are looking for