Logical reporting cube - slow data retrieval

OLAP 11.2
I have seen discussion on this board recommending using a logical reporting cube to expose data if there are multiple physical cubes in the workspace.  The problem I am having is that the data retrieval seems very slow from the logical cube while it is fine if I select from the actual physical cubes. Here is my set up.
Physical cube A - Dimensions x,y,z
Physical Cube B - Dimension w,x
Physical Cube C, Dimension y,z
Logical Cube  D. Dimension w,x,y,z with calculated measures mapped to the measures from the cubes A,B and C as well as some other calculated measures based on these.
Data retrieval from A, B and C is quick but the query hangs if I select certain measures from D. On some measures, it seems to work without issues.
Is there some thing I am missing?
Thanks,
Usman

Usman,
Its probably because of LOOP_VAR  and/or LOOP_DENSE settings of the measures in your Logical reporting cube.
I have posted on this topic in the past.   With this new Oracle-Forum interface its difficult to search old postings.
(1). Execute the following:
exec dbms_cube_log.enable(dbms_cube_log.TYPE_OPERATIONS, dbms_cube_log.TARGET_TABLE, dbms_cube_log.LEVEL_HIGH);
(2). Run the olap query.
(3). Query CUBE_OPERATIONS_LOG table
select *
from cube_operations_log
where   upper(suboperation) like '%LOOP%' or upper(name) like '%LOOP%'
order by time;
Tell us what you see in the 'NAME  and  'VALUE'  columns, after your query execution.
Also take a look at this post:   Delay when querying from CUBE_TABLE object, what is it?
Nasar

Similar Messages

  • Slow data retrieval on 8i

    I am using oracle 8i database on Windows 2000 server on compaq
    proliant 350 server machine. The problem is that sometimes
    connecting from a client is very slow and after connection data
    retrieval is also slow. I am usring TCP/IP and net8. I came to
    know from internet that there is a patch which is actually a
    work around to solve this problem. Can anybody help me to locate
    this patch?
    Thanks in advance.
    G. Rajan.

    You can probably prove that this is the issue by creating a retrieve using the excel addin or smart view to replicate the form and see how long it takes to retrieve.
    You will also see in the essbase app logs how long it is taking to perform the retrieve.
    Cheers
    John
    http://john-goodwin.blgspot.com/

  • Developing Z report of Workflow : Data retrieve for work id

    HI all,
    Iam  developing a workflow report for account payables and receivable , I have got the linkage between the work id and the document . Now i want
    1. Reason for which the document was parked
    2. Reason for which the document was Rejected
    3 Additional Note parking/rejecting the the dialog box
    I have checked this FM  SWW_WI_CONTAINER_READ but iam not clear how to retrieve the data from this FM
    Regards
    Madhukar Shetty
    Edited by: Madhukar Shetty on Sep 29, 2010 12:49 PM

    Hi Madhukar,
    The Input for the FM SWW_WI_CONTAINER_READ is Work item ID.
    Pass your Workitem ID to this FM. It will give you all your Workflow container data's in Element-Value format.
    Just loop it and get the required data's.
    Thanks,
    Viji.

  • Slow data retrieval from essbase using Planning web

    Hi
    Well, When we try to access the data form on ultimate parent level, it takes long time to retrieve data from the server. Pls advice
    Regards
    SNK

    You can probably prove that this is the issue by creating a retrieve using the excel addin or smart view to replicate the form and see how long it takes to retrieve.
    You will also see in the essbase app logs how long it is taking to perform the retrieve.
    Cheers
    John
    http://john-goodwin.blgspot.com/

  • BPC 10 - EPM data retrieval very slow!

    Hi BPCers,
    We are using an Excel EPM Input Schedules as a Resource Management tool - using VBA to provide the functionality we need.
    Performance is generally good, but quickly deteriorates when handling larger data sets - even 500-600 rows of transactional data is enough to slow data retrieval from our BPC Cube to EPM to an unusable speed. This is in relative terms pretty small so  there should be some option for optimisation.
    Does anybody have any experience with this? All suggestions welcome. We are operating on EPM Service Pack 7 Patch 1, but I'm not sure that EPM is necessarily the problem here.
    Thanks,
    Tom

    Thanks Gersh,
    Had a look through fiddler and have identified the job that is causing the delay - some rooting around in the ABAP debugger produced the answer as to why adding more data slows processing speed so dramatically.
    When we take data from the back end, we select a couple of parameters which limit the range of data that we are pulling through - a certain set of people, and a certain range of days. Once this is pulled through, allocations are made to any combination of person and day within this range, which generates and extra two properties - a project ID and a work status.
    This makes 4 properties, and when BPC pulls data it attempts to find every combination of every one of the properties that exist within this range - so the more allocations made the more this slows down as it dramatically increases the number of combinations.
    The result is that BPC runs through a couple of hundred thousand generated tables, most of which are nonsense.
    Not sure what to do from here. This is how BPC reads data so approaching a fix could be difficult.
    Tom

  • Passing values to subreport in SSRS throwing an error - Data Retrieval failed for the report, please check the log for more details.

    Hi,
    I have the subreport calling from the main report. The subreport is based on MDX query agianst the SSAS cube. some dimensions in cube has values 0 and 1.
    when I try to pass '0' to the sub report as the parameter value, it gives an error "Data Retrieval failed for the report, please check the log for more details".
    Actually I am using table for storing these parameter values. In the main report I am calling this table (dataset) and passing these values to subreport.
    so I have given like [0],[1] and this works fine. when I give only either [0] or [1] then it is throwing an error.
    Could you please advise on this.
    Appreciate all and any help.
    Thanks,
    Divya

    Hi Divya,
    Based on the current description, I understand that there is no issue if you pass two values from main report to subreport, while the issue occurs when passing one value to subreport.
    To narrow down the issue, I want to confirm whether the subreport can run if there is only [0] or [1] in the subreport. If so, it indicates the query statements exist error in the subreport. If it’s not the case, this shows the issue occurs during passing
    values from main report to subreport. To make further analysis, please post the details of query statements of the subreport to the forum.
    Regards,
    Heidi Duan
    Heidi Duan
    TechNet Community Support

  • Sharepoint 2013 Reporting Services & OLAP Cubes for Data Modeling.

    I've been using PowerPivot & PowerView in Excel 2013 Pro for some time now so am now eager to get set up with Sharepoint 2013 Reporting Services.
    Before set up Reporting Services  I have just one question to resolve.
    What are the benefits/differences of using a normal flat table set up, compared to an OLAP cube?
    Should I base my Data Model on an OLAP Cube or just Connect to tables in my SQL 2012 database?
    I realize that OLAP Cubes aggregate data making it faster to return results, but am unclear if this is needed with Data Modeling for Sharepoint 2013.
    Many thanks,
    Mike

    So yes, PV is an in-memory cube. When data is loaded from the data source, it's cached in memory, and stored (compressed) in the Excel file. (also, same concept for SSAS Tabular mode... loads from source, cached in mem, but also stored (compressed) in data
    files, in the event that the server reboots, or something similar).
    As far as performance, tabular uses memory, but has a shorter load process (no ETL, no cube processing)... OLAP/MDX uses less memory, by requiring ETL and cube processing... technically tabular uses column compression, so the memory consumption will be based
    on the type of data (numeric data is GREAT, text not as much)... but the decision to use OLAP (MDX)/TAB (DAX) is just dependent on the type of load and your needs... both platforms CAN do realtime queries (ROLAP in multidimensional, or DirectQuery for tabular),
    or can use their processed/in-memory cache (MOLAP in multidimensional, xVelocity for tabular) to process queries.
    if you have a cube, there's no need to reinvent the wheel (especially since there's no way to convert/import the BIDS/SSDT project from MDX to DAX). If you have SSAS 2012 SP1 CU4 or later, you can connect PV (from Excel OR from within SP) directly to the
    MDX cube.
    Generally, the benefit of PP is for the power users who can build models quickly and easily (without needing to talk to the BI dept)... SharePoint lets those people share the reports with a team... if it's worthy of including in an enterprise warehouse,
    it gets handed off to the BI folks who vet the process and calculations... but by that time, the business has received value from the self-service (Excel) and team (SharePoint) analytics... and the BI team has less effort since the PP model includes data sources
    and calculations - aside from verifying the sources and calculations, BI can just port the effort into the existing enterprise ETL / warehouse / cubes / reports... shorter dev cycle.
    I'll be speaking on this very topic (done so several times already) this weekend in Chicago at SharePoint Saturday!
    http://www.spschicagosuburbs.com/Pages/Sessions.aspx
    Scott Brickey
    MCTS, MCPD, MCITP
    www.sbrickey.com
    Strategic Data Systems - for all your SharePoint needs

  • Report Developed in Webi Rich Client Consuming more time in Data Retrieval

    Dear All,
    I am a BO Consultant, recently in my project I have developed one report in Webi Rich Client., at the time of development and subsequent days the report was working fine (taking Data Retrieval time less than 1 minute), but after some days its taking much time (increasing day by day and now its taking more than 11 minutes).
    Can anybody point out what could be the reason?????
    We are using,
    1. SAP BI 7.0
    2. SAP BO XI 3.1 Edge
    3. Webi Rich Client Version :12.3.0 and Build 601
    This report is made on a Multiprovider (Sales).
    What are the important points that should be considered so that we can improve the performance of Webi Reports????
    Waiting for a suitable solution.....................
    Regards,
    Arun Krishnan.G
    SAP BO Consultant
    Edited by: ArunKG on Oct 11, 2011 3:50 PM

    Hi,
    Please come back here with a copy/paste of the 2 MDX statements from the MDA.log to compare the good/bad runtimes.
    & the 2 equivalent DPCOMMANDS clauses (good and bad) from the WebI trace logs.
    Can u explain what u really mean in the bold text above..................Actually I didn't get you..........
    Pardon, I have only 3 months experience in BO.
    Regards,
    Arun
    Edited by: ArunKG on Oct 11, 2011 4:28 PM

  • Can we generate a report on master data alone without Info cube?

    Could some one let me know whether
    Can we generate a report on master data alone in BEx without Info cube?
    Thanks in advance.

    yes you can generate a report .
    use the master data infoobject and insert that infoobject as a infoprovider and design a query using this infoprovider and view your query on this master data in Bex Query...
    cheers
    Sunil

  • Updating metadata/data in ASO reporting cube from BSO

    Hi Experts,
    I am creating an ASO reporting cube and would like to know the best way to automate the update of metadata and data from a BSO cube without the use of a partition? This is my first time working with ASO. Any insight or examples would be much appreciated. Please advise.
    Thanks in advance!

    Thank you for the responses!
    The dimensions will be the same. I will be exporting level0 data then loading it into the ASO cube. I believe I should be able to use the same load rule that I am using for the BSO cube, correct? Also, I am planning on only loading the 'Final' versions of 'Budget' and 'Forecast'. In my data export I believe I can put a condition to only export the 'Final' version, please correct me if I'm wrong?
    One last thing, I would like to keep 2 years of 'Actuals' and 1 year of 'Budget' and 'Forecast'. At year end I would to clear previous 'Budget' and 'Forecast' and 2nd prior year after the current 'Budget' and 'Forecast' are closed. Can this be handled in a separate script? Please advise.
    Thanks again for the responses.

  • User complaining report very slow?

    Hi Gurus,
    There is an interview question in EBS 11I:
    “We have lot of issues in here, We facing some problem, the Discoverer users always complaining getting report very slow, sometimes it gives time-out error. What you think, how to resolve the problem?”
    What are the answers they are looking from us?
    Usually they are not satisfied with couple of answers, if we answered/verified each answer, the interviewer reply “No, …No…No……….. ”.
    So tell me the answer what they are looking for. Please provide full possible scenarios.
    Thanks,

    Here is my 2 cent comments:
    1. network issue
    you can compare the discoverer reports with Oracle reports or Application module to identify this, usually this is not the case
    2. tune the backend query
    you can run the discovery report backend sql query in TOAD, if the query is very slow, then the query has performance issue, you can modify the join condition to improve the performance
    3. add Parameter to Discoverer report
    sometimes it is very difficult to modify the backend sql query, you can add parameter to discoverer report, when end user run the report based on the parameter, data can be quickly displayed. for example, currently report will retrieve all invoices for the company, if you add a year parameter, 2010, the report only retrieve the invoice of 2010.
    4. change timeout time
    timeout issue is a discoverer report setup issue, you can change the timeout time for example from 30 minutes to 60 minutes.
    Thanks

  • Multi server data retrieval performance

    Hi experts,
    I have a question regarding the data retrieval performance (EVDRE) on a multi server installation environment on Microsoft SQL Server 2008.
    We have succesfully migrated Outlooksoft 4.2 SP03 to SAP BPC 7.0 SP07 for a customer. During this project we have also set up a complete new server environment consisting of:
    Development server: dedicated single server, Windows 2003 Standard SP2 32 bit, SQL Server 2008 SP1 with cumulative update package 6, SAP BPC 7.0 SP07, 2 quad core processors, 4 GB RAM
    QA server: dedicated multi servers - 1 database server (SQL/OLAP), Windows 2003 Standard SP2 64 bit, SQL Server 2008 SP1 with cumulative update package 6, 2 quad core processors, 32 GB RAM - 1 dedicated application/web server, Windows 2003 Standard SP2 32 bit, SQL Server 2008 SP1 with cumulative update package 6 (shared components / reporting services), SAP BPC 7.0 SP07, 2 quad core processors, 4 GB RAM
    Production server: dedicated multi servers - 1 database server (SQL/OLAP/Reporting services), Windows 2003 Standard SP2 64 bit, SQL Server 2008 SP1 with cumulative update package 6, 2 quad core processors, 32 GB RAM - 2 dedicated application/web server, Windows 2003 Standard SP2 32 bit, SQL Server 2008 SP1 with cumulative update package 6 (shared components), SAP BPC 7.0 SP07, 2 quad core processors, 4 GB RAM
    Furthermore, two terminal servers with the SAP BPC client.
    All servers have good performancve and we have great times on cube processing and SQL processing. However, to our great surprise we find that the single development server is much faster with a single user to retrieve data using EVDRE than the multi-server environment. About 2x as fast. A reporting book with more then 10 sheets and about 25 EVDRE's takes about 42 seconds on the development server and 93 seconds on the multi server.
    It seems that EVDRE is taking up a lot of time to communicate between the application server and the database server in a multi server environment while being much faster on a single server. This is not what we want :-). The network speed in the domain consist of all 1 GB lines so that should not be the issue.
    Do you have any experience with this? How can we upgrade the speed of the multi server, are there specific settings?
    Hope the get some useful answers. Thanks in advance.
    Damien
    Edited by: DWiegman on Feb 20, 2010 4:15 PM

    Hi,
       You have to activate also the EVDRE logs on the client and server level, just to understand from where is coming the problem (appserver-db comuncication or client-appserver communication). You have to check also if there is any proxy firewall between client and application server.
        In case you are using NLB, please verify if afinity is setup to true.
        The performance problems can be come from db level. Did you verify how many records do you have into WB table for the specific application? Are you keeping the DB in full mode? How big is the log of the databse?
        The are a lot of things that can have impact of this, but it looks to be a setup problem.
    Hope this can help you,
    Mihaela

  • Data Retrieval from SSRS 2008 R2 to SQL Server 2012 standard

    Hello,
    We have a virtual machine running SSRS 2008 R2 and a physical server running SQL server 2012 standard.  We are having an issue with a report that is taking almost 10 minutes for the data to be retrieved from the sql 2012 database.  We used the
    execution logging query to debug and analyze.  We see that it renders and processes very fast, but the data retrieval time is very slow.  When the same query is run using sql studio mgr. from the SSRS server it runs just fine...it
    takes seconds for the results to come up.
    I don't believe there is a compatibilty issue because other reports run fine.  We're just having problems with one report.

    A long shot, but I've seen where the data type of a report parameter is mismatched against the data type of the table it is being compared against.  SQL profiler would show this clearly.  
    This would explain why the query returns quickly in ssms but not with the report.  

  • Data Retrieval in task audit

    Hello,
    What is the exact meaning of Data Retrieval in task audit? Is it just data retrieved by reports?
    Thanks
    Lilach

    Hi Lilach,
    It would mean data retrieval through adhoc analysis or grids,HsGetValue Functions etc.
    You could also check Task Audit table you can see activity codes which has details below.
    ActivityID
    ActivityName
    0
    Rules Load
    1
    Rules Scan
    2
    Rules Extract
    3
    Consolidation 
    4
    Chart Logic
    5
    Translation 
    6
    Custom Logic
    7
    Allocate 
    8
    Data Load
    9
    Data Extract
    10
    Data Extract via HAL
    11
    Data Entry
    12
    Data Retrieval
    13
    Data Clear
    14
    Data Copy
    15
    Journal Entry
    16
    Journal Retrieval
    17
    Journal Posting
    18
    Journal Unposting
    You can query AppName_Task_Audit & HSV_Activity_Users tables by joining them and applying condition on Activity id from App_Task_Audit.
    Thanks
    Amith

  • Performance: How to manage large reports with high data volume

    Hi everybody,
    we actually make some tests on our BO server system, to define limitations and oppertunities. Among other things we constructed a large report with a high data volume (about 250.000 data records).
    When executing the query in SAP Query Designer it takes about 10 minutes to display it. In Crystal Reports we rebult the row and column structure of the query. The data retrieval in Crystal Reports Designer last about 9 minutes - even faster as in the query.
    Unfortunately in BO InfoView the report is not displayed. After 30 minutes of loading time we get a timeout error RCIRAS0244.
    com.crystaldecisions.sdk.occa.managedreports.ras.internal.ManagedRASException:
    Cannot open report document. ---
    The request timed out because there has been no reply from the server for 600.000 milliseconds.
    Also a refresh of an report with saved data is not possible.
    Now we are asking us some questions:
    1. Where can we set the timeout for InfoView to a value larger than 30 minutes?
    2. Why is InfoView so slow compared to Crystal Designer? Where is the bottleneck?
    3. Whats the impact of SAP single sign-on compared to Enterprise logon on the performance?
    Thanks for any helps and comments!
    Sebastian

    Hi Ingo,
    thank you for your reply.
    I will check the servers and maybe change the time limits.
    Unfortunately we have a quite slow server system that probably cause this timeout. In CR Designer we have no problems, its really quick. Is it to expect that CR Designer and InfoView have almost the same performance?
    Another nice point: When we execute the query in SAP BEx Query Designer it takes about 10 minutes to open it, in Crystal Designer it needs just about 5-6 minutes. We integrated exactly the same fields in the report, which exist in die SAP BEx query.
    What may cause the difference?
    - Exceptions and conditions in the query?
    - Free characteristics in the query?
    - anything else?
    Best regards,
    Sebastian

Maybe you are looking for