Query performance on remote/virtual cube

All:
We are on BW 3.5.
Can anyone suggest anything about improving query performance on remote/virtual cubes? Analysis shows query performance is sufferring at the database level.
I am looking for some advise beyond hardware and database parameters. It seems current hardware and database parametrs work fine with basis cubes.
Another solution is datamart.But can anything be done before/without going to datamart?
Any help will be appreciated.
Thanks

Hi,
In this case try to find where more time is utilized by using the ST03.
if there is more time consuming in front end, rearrange the query by using less in the row and using more free chars and filter areas. using the variable also will work.
use reporting agent and schedule the query in background to fill the cache and rerun the query.
Reg,
Vishwa

Similar Messages

  • Query Performance Issues on a cube sized 64GB.

    Hi,
    We have a non-time based cube whose size is 64 GB . Effectively, I can't use time dimension for partitioning. The transaction table has ~ 850 million records. We have 20+ dimensions among which 2 of the dimensions have 50 million records.
    I have equally distributed the fact table records among 60 partitions. Each partition size is around 900 MB.
    The processing of the cube is not an issue as it completes in 3.5 hours. The issue is with the query performance of the cube.
    When an MDX query is submitted, unfortunately, in majority of the cases the storage engine has to scan all the partitions (as our cube  is not time dependent and we can't find a suitable dimension that will fit the bill to partition measure group based
    on it.)
    I'm aware of the cache warming and  usage based aggregation(UBO) techniques.
    However, the cube is available for users to perform adhoc queries and hence the benefits of cache warming and UBO may cease to contribute to the performance gain as there is a high probability that each user may look at the data from different perspectives
    (especially when we have 20 + dimensions) as day(s) progress.
    Also, we have 15 + average calculations (calculated measures) in the cube. So, the storage engine sends all the granular data that the formula engine might have requested (could be millions of rows) and then perform the average calculation.
    A look at the profiler suggested that considerable amount of time has been spent by storage engine to gather the records (from 60 partitions).
    FYI - Our server has RAM 32 GB and 8 cores  and it is exclusive to Analysis Services.
    I would appreciate comments from anyone who has worked on a large cube that is not time dependent and the steps they took to improve the adhoc query performance for the users.
    Thanks
    CoolP

    Hello CoolP,
    Here is a good articles regarding how to tuning query performance in SSAS, please see:
    Analysis Services Query Performance Top 10 Best Practices:
    http://technet.microsoft.com/en-us/library/cc966527.aspx
    Hope you can find some helpful clues to tuning your SSAS Server query performance. Moreover, there are two ways to improve the query response time for an increasing number of end-users:
    Adding more power to the existing server (scale up)
    Distributing the load among several small servers (scale out)
    For detail information, please see:
    http://technet.microsoft.com/en-us/library/cc966449.aspx
    Regards,
    Elvis Long
    TechNet Community Support

  • Query Data cached? (Virtual Cube)

    Hi folks,
    I have some problems with a Query-Objekt which gets data out of an Virtual Cube. The Virtual Cube is based on a 3.X Infosource an gets data out of a table in the ERP.
    When i call the Query from VC everything works fine and current data is shown. But if i manipulate data of the table which the Virtual Cube points to and send a refresh event to the Query-Objekt, the manipulated data is not shown. It always returns the data which it fetched at the first call. If i refresh the whole application in the browser (via F5), the manipulated data is shown. I disabled cachmode in rsrt for this query but it doesn't work.
    Any chance to get the current data by just sending a refresh action and call the Query-Objekt again without reloading the whole application? Any Idea?
    Points will be awarded for usefull information.

    Hello,
    The reason why the data manipulated is not showing up in the query even after refresh is sent is because the cache for the virtual provider does not get reset as it would for a normal InfoCube.
    So, it does not know when to reset the cache for itself even when data is manipulated.
    The way we have worked around this is have a temporary process chain which runs on a frequent basis and executes the function module RSDMD_SET_DTA_TIMESTAMP for the virtual cube in consideration.
    Thanks
    Dharma.

  • Performance issue on a virtual cube

    Hi BW gurus,
    I am working on Consolidation virtual cube and the query performance thru that cube is very bad. I know that we cannot build aggregates or partition on a virtual cube ..what should be my approach then.....
    Your suggestions will be appreciated with lots of points

    Hi Nick,
    If you can not move out of the virtualcube option, then I think you should try to improve the performance of the virtual cube. This mainly ABAP work. You can use SE30 to analyze what parts of the code is taking too much time. Follow this steps:
    1) Create a breakpoint in the function module of your virtualcube
    2) Go to listcube and initiate an extraction from your virtualcube.
    3) On a separate session, run SE30 to start an analysis of the extraction process for your virtual cube.
    You can use the report from SE30 as a starting point on your performance optimization work.
    Note: Transaction ST05 can help you determine what database calls are taking a long time.
    Hope this helps.

  • BEx query based on virtual cube donu00B4t display a valid List of Value (LOV)

    Hello
    I have a problem with an invalid LOV. The scenario is the following; There´s a BEx query based on a virtual cube. The query has an exit variable on caracteristic that is based on 0CALMONTH.
    At Universe Designer I simply create a connection, a universe based on this query and export.
    At Web Intelligence (also at Live Office), when I try to execute de query, the prompt to fill my exit variable display a list of value that doesn´t match with the values of the caracteristic at the cube.
    Actually, the list at the prompt starts with 01.0000 and finishes with 05.0968.
    In Universe Designer, the option to edit the list of values is not available. But I think that editing the LOV is not the correct way.
    I´ve tried creating a new query based on the DSO that is the source of the virtual cube. In this case, I had a valid list. Unfortunately, I can´t use this DSO.
    Did anyone already have this problem?

    Hi James,
    can you explain what you mean with "input length for that filed" ?
    The field in the table is varchar2(120). I coudn't found options for the List of value.
    Thanks for your response
    Carsten
    null

  • Query performance on Multiprovider(Remote Cube)

    Hi All,
    I have to increase the query performance for a report which built on Multi provider.
    This multiprovide designed from several remote cubes,but for this report data will bring through one remote cube from R/3.
    In filter i had one remote cube, which bring data from R/3.
    Now in ST03 the stats are like
    %init Time - 0, %DB time - 0, %OLAP time - 16.67, %Front end - 83.33.
    Now i have to improve the %Front end lapsed time.
    Could you please guide me.
    Thanks
    Srinivas

    Hi Srinivas,
    Please see this document
    https://websmp105.sap-ag.de/~sapidb/011000358700001394912002
    And this Discussion Thread
    Re: Deactivate Hierarchy symbols in excel
    See whether this is helpful in case of Remote Cubes.
    Thanks
    CK

  • ACCESS REMOTE AND VIRTUAL CUBES

    Hi all,
    I installed standard cube:
    0FIGL_C10       General Ledger (New): Transaction Figures
    and it came with the following ones:
    0FIGL_R10       General Ledger (New): Transaction Figures, Remote
    0FIGL_V10     General Ledger (New): Financial Statement
    0FIGL_V11     General Ledger (New): Financial Statement, Remote
    How do i link these remote and virtual cubes to 0FIGL_C10 and how do i populate these cubes to execute query's from them?
    Thanks

    Hi SAPPER BI,
    Check this links:
    Sap Help: http://help.sap.com/saphelp_nw70/helpdata/en/e2/f16940c3c7bf49e10000000a1550b0/frameset.htm
    Best Practices: http://help.sap.com/bp_biv270/BBLibrary/HTML/B70_EN_DE.htm
    Only 0FIGL_V10 is connected to that cube, and it uses a function module. You won't see a transformation between them. You should activate them from BCT, then load and information will get populated to your reports on the virtual and remote cubes.
    Hope this helps.
    Regards,
    Diego

  • Querying on aggregates created on Virtual Cube

    Hello,
    I have implemented a virtual InfoProvider with Services.When I create queries directly on the Virtual Infoprovider the query runs fine and I see the report.
    As per my requirement I create an aggregate on the Virtual Infoprovider .Then I define a query on the aggregate .But when I execute this query I get the following errors :
    Error reading the data of InfoProvider AG4
    An exception with the type CX_SY_REF_IS_INITIAL occurred, but was neither handled locally, nor declared in a RAISING clause
    Dereferencing of the NULL reference.
    Would appreciate any assistance on this topic.
    Thanks
    Priyadarshi

    Yes it is possible to create aggregates on Virtual cubes.
    I will be grateful if hope anybody who is aware of the method of aggreagate creation and who has faced similar issues comes forward and throws some light on what could be the error.
    Thanks

  • Query on virtual cube showing incorrect values

    Experts,
    I have create a virtual cube on a base cube using services (info-source). both virtual as well base cube are z-development and are in BI.
    Base cube has 6 records, now when I check data in virtual cube using display data it shows correct values based on routine written in virtual cube transformation.
    Problem is when I created query on this virtual cube it doesnt show correct data or sometime no applicable data. After debugging I found, in end routine itself result package contains records with blank values in it on which I have written routine. But it does this only while executing query not when I display data in cube.
    Regards,
    Akshay

    Hi!!!
    And how do you solved it?

  • Query ID in Virtual Cube with services-Function module

    Hi,
    I am using virtual cube with services linked to a function module.
    The function module has fixed parameters(such as infoprovider name). None of these parameters consists of query information such as query  ID OR queryname .
    Do any one know how to determine query which was executed this function module?
    Best Regards,
    Anil

    Hi Claudio,
    I never implemented Virtual InfoCube with services with a FM, but I know there is a couple of How To Documents about named:
    - How to Reporting from External Data via Virtual InfoProvider
    -How to Implement a Virtual InfoCube with Services
    both with some code samples: did you read it?
    Hope it helps
    GFV

  • Query on BCS virtual cube is not using the aggregates on BCS basic cube

    Hi all,
    I have BCS Virtual cube which is linked to BCS Basic cube. I built aggregates on BCS Basic cube.
    I created simple query on BCS basic cube and ran in debug mode of rsrt, it showed the aggregates on bcs basic cube. But when I created the same query on BCS vitual cube and ran it rsrt debug mode the query did not show any aggregates, that was strange.
    So My questions is whether query built on virtual bcs can utilize the aggregates built BCS basic cube, if possible please let me knows the tweaks.
    Thanks,
    Raj.

    1. Goto se37. Enter RSDRI_INFOPROV_READ and choose Display.
    2. In line 82 (in a BW 3.5) there is a line that says:
      CLEAR: e_t_data, e_end_of_data, e_aggregate, e_split_occurred.
    Put the cursor in there and press the 'stop shield' or use CtrlShiftF12.
    3. In the same mode open transaction RSRT and choose your query. Execute it. If you stop at the breakpoint, enter I_TH_SFC into one of the fields in the lower left area and press Enter. You should see a table with the characteristics you need in the system.
    As I said I'm not quite sure if it works. I have access to a BCS system on monday. I'll try then to find out more.
    Best regards
    Dirk

  • Impact of real time cube on query performance and OLAP cache

    Hi:
    We have actual and plan cubes both setup as real time cubes (only plan cube is being planned against, not actual cube) and both cubes are compressed once a day.
    We are planning on implementing BIA accelerator and have questions related to query performance optimization:
    1/ Is there are any query performance benefits in changing the actual cube to be a basic cube (using program SAP_CONVERT_NORMAL_TRANS) if the F table is fully compressed
    2/ Can OLAP cache be leveraged for the queries run against the real time cubes e.g. actual cubes
    3/ What is the impact on BIA of having the actual cube as real time (whetehr or not there is data being loaded/planned during the day in that cube)
    Thank you in advance,
    Catherine

    1) Is there are any query performance benefits in changing the actual cube to be a basic cube (using program SAP_CONVERT_NORMAL_TRANS) if the F table is fully compressed
    From the performance point of view, tha actual cubes i.e std cubes are relatively better.
    2) Yes OLAP cache can be leveraged for bringing up the plan query but all the calculations are done in the planning buffer.
    3) Not sure.

  • Query on Virtual cube 0FIGL_V10 show old "Last Data Updated:" on Portal

    Hello all,
    We are using the SAP delivered virtual cube for G/L financial statements reporting. We are running these queries on Portal (BI 7.0) we are using the 0ANALYSIS_PATTERN template. Now since we are not actually loading data in this cube the template show "Last Data updated" as long time back, I am guessing when the cube was last activated or transported.
    The template and time stamp is working perfect for queries on all other cubes or Multiproviders.
    The business users are asking can we have the actual time stamp there (when actually the data was last loaded),
    Does anyone have any idea how we can do it without actually rewriting the queries on standard cube.
    Thanks in advance,

    Hi Kiran,
    Always last update data status only will see in our reports on last data updated date. It may be activation of data in cube. please check is there any option of activate data in data flow level.
    Try to create one record in remote data source  and load it again. It may be find the latest status.
    Thanks,
    Chandra

  • BCS Virtual Cubes Performance Tuning

    Hi All,
    We are working on improving the performanc tuning of queries on BCS Virtual Cubes (with services).
    Any specific changes (from RAM to specific properties on queries, Virtual Cubes with services) that you have seen working in improving the performance in yuor environemnt is greately appreciated.
    Thanks,
    - Shashi

    Thanks a lot Marc,
    We are on NW2004, with the following support pack levels
    SAP_BW     350     0016
    FINBASIS     300     0012
    BI_CONT     353     0008
    SEM-BW     400     0012
    I have checked the service market place and the current SP's available are:
    SAP_BW     350     0019
    FINBASIS     300     0015
    BI_CONT     353     0013
    SEM-BW     400     0015
    Which Service Packs that you suggest us to go with for performance issues on BCS Virtual Infoprovider queries?
    Thanks,
    - Shashi

  • Building a new Cube Vs Restricted Key figure in Query - Performance issue

    Hi,
    I have a requirement to create  a OPEX restricted key figure in Query. The problem is that the key figure should be restricted to about 30 GL Accounts and almost 300 Cost centers.
    I do not know if this might cause performance issue in the query. At the moment, I am thinking of creating a new OPEX cube and load only those 30 GL Accounts, 300 cost  centers and Amount. and  include OPEX  in multiprovider in order to get OPEX
    amount in the report.
    whats the best solution - creating OPEX restricted key figure or OPEX cube ?
    thanks,
    Bhat

    I think you should go for cube as all the restrcited key figure are calculated at OLAP runtime so it will definitely affect the query performance.There are a lot of costcenter for which you have to restrict it so definitely during the runtime of query it will take a lot of time to fetch tha data from infoprovider.Its better that you create a cube with the restrictions and include it in MP.It will definitely save a lot of time during query execution
    Edited by: shyam agarwal on Feb 29, 2012 10:26 AM

Maybe you are looking for

  • How can I fit 75 pages into Illustrator?

    I'm working on a rulebook for a boardgame.  Each page has vector art and text blocks intermixed to explain the rule mechanics.  It now exists as 75 individual Illustrator files that need to be combined into a single PDF document.  Obstacle 1:  Illust

  • Maps not showing on iPad

    Road maps suddenly no longer work on my iPad, the problem affects both the Map App and The Weather Channel App (radar portion).  If I try to view the road map I get nothing but grid lines, the only maps that seem to work (so far) are the Satellite ma

  • Transport Requests

    Hello Experts, I am facing couple of issues in transporting the requests from Dev to QA environment. Firstly when trying to replicate DS in QA, DS is replicated but it is in MOdified status (and Cannot activate because QA system is non modifiable sta

  • Can't install a downloaded MX

    I had to erase and do a clean install on my Mac OS10.4.11. Can't find my original CD of Freehand MX to install it again. Tried to order it from Customer service but it will take too long to get it here. I've got to finish a design job. So I downloade

  • How do you  copy mac to mac

    I Want to COPY Iphoto   FROM old mac to new mac .