Steps involved in Query execution

Can anyone please suggest the exact steps that take place in a query execution. I know on the broad level. But I want to know to basic level. For example we would like to know all the steps that an OLAP processor does to get the data, process it and then send it for display.
Also, say the total time taken for the query execution is xx msec, we would like to know the complete break up of this time for all the activities that are completed from the time of query execution till the display of result.
Thanks

Hi Suchitra,
Please go through these links:
/people/vikash.agrawal/blog/2006/04/26/in-what-147sequence148-olap-processor-look-for-query-relevant-data
/people/prakash.darji/blog/2006/01/26/query-optimization
Assign points if helpful
Thanks,
Amith

Similar Messages

  • An error has occurred during report processing. (rsProcessingAborted) Query execution failed for dataset 'dsPriority'. (rsErrorExecutingCommand)

    click report error:
    log file:
    processing!ReportServer_0-2!104c!04/27/2015-19:15:21:: e ERROR: Throwing Microsoft.ReportingServices.ReportProcessing.ProcessingAbortedException: An error has occurred during report processing., ;
     Info: Microsoft.ReportingServices.ReportProcessing.ProcessingAbortedException: An error has occurred during report processing.
     ---> Microsoft.ReportingServices.ReportProcessing.ReportProcessingException: Query execution failed for dataset 'dsPriority'.
     ---> Microsoft.AnalysisServices.AdomdClient.AdomdErrorResponseException: Query (26, 25) The ALLMEMBERS function expects a hierarchy expression for the  argument. A member expression was used.
       at Microsoft.AnalysisServices.AdomdClient.AdomdDataReader..ctor(XmlReader xmlReader, CommandBehavior commandBehavior, AdomdConnection connection)
       at Microsoft.AnalysisServices.AdomdClient.AdomdCommand.ExecuteReader(CommandBehavior behavior)
       at Microsoft.AnalysisServices.AdomdClient.AdomdCommand.System.Data.IDbCommand.ExecuteReader(CommandBehavior behavior)
       at Microsoft.ReportingServices.DataExtensions.AdoMdCommand.ExecuteReader(CommandBehavior behavior)
       at Microsoft.ReportingServices.OnDemandProcessing.RuntimeDataSet.RunDataSetQuery()
       --- End of inner exception stack trace ---
       at Microsoft.ReportingServices.OnDemandProcessing.RuntimeDataSet.RunDataSetQuery()
       at Microsoft.ReportingServices.OnDemandProcessing.RuntimeDataSet.Process()
       at Microsoft.ReportingServices.OnDemandProcessing.RuntimeParameterDataSet.Process()
       at Microsoft.ReportingServices.OnDemandProcessing.RuntimeDataSet.ProcessConcurrent(Object threadSet)
       --- End of inner exception stack trace ---
    open http://localhost:8080/tfs/TeamFoundation/Administration/v3.0/WarehouseControlService.asmx?op=ProcessWarehouse click Invoke:
    System.Web.Services.Protocols.SoapException: TF221029: Reporting for Team Foundation Server does not have any warehouse jobs defined. Use the Team Foundation Administration Console to rebuild the reporting. : 2015-04-27T19:30:29:782 ---> System.InvalidOperationException:
    TF221029: Reporting for Team Foundation Server does not have any warehouse jobs defined. Use the Team Foundation Administration Console to rebuild the reporting. at Microsoft.TeamFoundation.Warehouse.WarehouseAdmin.QueueJobs(String collectionName, String jobName)
    at Microsoft.TeamFoundation.Warehouse.WarehouseControlWebService.ProcessWarehouse(String collectionName, String jobName) --- End of inner exception stack trace --- at Microsoft.TeamFoundation.Warehouse.WarehouseControlWebService.ProcessWarehouse(String collectionName,
    String jobName)

    Hi shelman,
    I'd like to know whether you configured TFS reporting service properly, and if you can get the report normally before. From the error message, you might has wrong data source or the parameters of the dataset 'dsPriority'. You can check whether the data source
    for your report is available, try to use windows authentication. Or check the parameters of dataset 'dsPriority' make sure the query works when you execute it manually.
    To process the TFS data warehouse and analysis services cube, I'd like to know which operation you selected before clicking Invoke button. Please follow the instructions on this
    page to process TFS data warehouse and analysis service cube manually. You can also refer the James's last reply in this
    thread or this
    blog to check if the solutions work for you.
    Another option is run TFS best practice analyzer to check if there any configure issues on your TFS server machine. And check event logs to see if there any useful information, elaborate more details about your scenario including reproduce steps if the problem
    persists.
    Best regards,

  • Finding number of working days during query execution

    Hi ,
    We have a requirement to calculate number of working days during query execution. In the cube we got "received date", this will be used as a variable in the query. Based on the received date value entered, query needs to display number of products received from past 1, 2,3,4 days (working days should include weekends or public holidays).
    1day     2 days     3 days     4 days   5 days 
    10         12            20            15          20
    to me it appears as I need to create virtual key figures.
    Regards,
    Ramz

    Hi ramesh,
    First try to create formula variables for start data and end date to calculate the no of working days.
    Create a Variable ZSTDT --of type replacement path.
    Type of Variable = Formula
    Variable Name = ZSTDT
    Characteristic = <info object>
    For Intervals Use = Select From Value
    Offset Start = 0000 Offset Length = 0000
    Replacement var with : key
    use the same steps to create one more formula variable to get end date.
    Now create a formula to get the No of days.
    once after getting no of days you can create buckets as per your need.
    for more info check the below links
    http://www.sd-solutions.com/SAP-HCM-BW-Replacement-Path-Variables.html
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/30f15839-0cf1-2b10-c6a7-ebe68cc87cdc?quicklink=index&overridelayout=true

  • Query Execution plan

    Hi,
    I have a question about reading query execution plans. I don't understand what is meant by "Hash join" in a plan step. I understand "Nested_Loops" and "Merge_Join", but after looking at different documentations, I am still not clear about "Hash_Join". Please don't give links to Oracle documentation, because I have seen a lot of it, but it is still not clear. I also don't fully understand what is meant by a "Hash Table", what are all these "Hashesssssss"???
    Please explain in your own words, if possible, and with the help of some example.
    Thanks in Advance.,

    user566817 wrote:
    Can any one else answer my question? I would be highly thankful.Hash joins are one of the 3 techniques used by Oracle to join to sets of data.
    As a brief, and very very crude overview
    Generally after applying the filter to one data source, the join key is hashed into a memory table (as described by link sb provided). The second data set areas filtered and it's join key is also hashed, but the hash is used to find matches in the memory table created by the first data source. Assuming no hash chaining, found matches are joins; if hassh chaining exists, the chain needs to be traversed to verify joins.
    This, of course, has a lot of holes which would be filled in by the documentation and by external links that you do not seem to want to use.
    (Remember: School papers need to cite references. Teachers these days DO check.)

  • Slow query execution time

    Hi,
    I have a query which fetches around 100 records from a table which has approximately 30 million records. Unfortunately, I have to use the same table and can't go ahead with a new table.
    The query executes within a second from RapidSQL. The problem I'm facing is it takes more than 10 minutes when I run it through the Java application. It doesn't throw any exceptions, it executes properly.
    The query:
    SELECT aaa, bbb, SUM(ccc), SUM(ddd), etc
    FROM MyTable
    WHERE SomeDate= date_entered_by_user  AND SomeString IN ("aaa","bbb")
    GROUP BY aaa,bbbI have an existing clustered index on SomeDate and SomeString fields.
    To check I replaced the where clause with
    WHERE SomeDate= date_entered_by_user  AND SomeString = "aaa"No improvements.
    What could be the problem?
    Thank you,
    Lobo

    It's hard for me to see how a stored proc will address this problem. I don't think it changes anything. Can you explain? The problem is slow query execution time. One way to speed up the execution time inside the RDBMS is to streamline the internal operations inside the interpreter.
    When the engine receives a command to execute a SQL statement, it does a few things before actually executing the statement. These things take time. First, it checks to make sure there are no syntax errors in the SQL statement. Second, it checks to make sure all of the tables, columns and relationships "are in order." Third, it formulates an execution plan. This last step takes the most time out of the three. But, they all take time. The speed of these processes may vary from product to product.
    When you create a stored procedure in a RDBMS, the processes above occur when you create the procedure. Most importantly, once an execution plan is created it is stored and reused whenever the stored procedure is ran. So, whenever an application calls the stored procedure, the execution plan has already been created. The engine does not have to anaylze the SELECT|INSERT|UPDATE|DELETE statements and create the plan (over and over again).
    The stored execution plan will enable the engine to execute the query faster.
    />

  • Analyzer Query Execution vs. APD Query Execution

    hi gurus,
    i have a problem with an execution of a query in BI 7.0. the query is based on a DSO.
    if i execute the query in bex analyzer, i get the result within 30 seconds.
    but if i execute the query in an APD, it will take several hours to complete the job.
    i checked the RSRT to view the query execution plan - here it is:
    SELECT "O1"."COMP_CODE" AS "K____240" ,"O1"."RECONTRACT" AS "K____301" ,"O1"."FISCVARNT" AS "K____181"
    ,"O1"."/BIC/ZIMIMNPER" AS "K____535" ,"O1"."BUSENTITY" AS "K____237" ,"O1"."/BIC/ZIMIMGEBL" AS "K____1098"
    , COUNT( * )  AS "Z____131" , SUM (  "O1"."/BIC/ZIMIK_113"  )  AS "Z____1886" ,
    SUM (  "O1"."/BIC/ZIMIK_183"  )  AS "Z____1908" , SUM (  "O1"."/BIC/ZIMIK_201"  )  AS "Z____1998"
    FROM "/BIC/AZIMODSSM00" "O1"
    WHERE
    ( ( ( ( "O1"."/BIC/ZIMIMNPER" = '2010009'
    ) ) AND  ( ( "O1"."/BIC/ZIMIMZBR" IN ( 'BEQ' ,'REFX3' ,'REFX4' ,'REFX6') ) ) ) )
    GROUP BY "O1"."COMP_CODE" ,"O1"."RECONTRACT" ,"O1"."FISCVARNT" ,
    "O1"."/BIC/ZIMIMNPER" ,"O1"."BUSENTITY" ,"O1"."/BIC/ZIMIMGEBL"
    but when i check the SQL statement for the APD in SM51, i found this statement:
    SELECT   "S0000"."SID" ,"P0000"."COMP_CODE" ,"P0000"."BUSENTITY"
    FROM   "/BI0/PBUSENTITY" "P0000" JOIN "/BI0/SBUSENTITY" "S0000" ON  "P0000" . "BUSENTITY" = "S0000" . "BUSENTITY" AND "P0000" . "COMP_CODE" = "S0000" . "COMP_CODE"
    WHERE   "P0000"."OBJVERS" = 'A' AND "S0000"."SID" IN ( SELECT "S"."SID" FROM "/BIC/AZIMODSSM00" "O" JOIN
      "/BI0/SBUSENTITY" "S" ON  "O" . "BUSENTITY" = "S" . "BUSENTITY" AND "O" . "COMP_CODE" = "S" .
      "COMP_CODE" )
    ORDER BY
      "COMP_CODE" ASC , "BUSENTITY" ASC#
    could you help me understanding the difference and the cause for it?
    i think, there are some validitycheckings executed for masterdata!? is it possible to disable these steps?
    here some more information:
    in the query, i'm using two compound characteristics and an constant selection for a restricted keyfigure. the characteristics all have access type for result values: posted values.
    regards,
    philipp

    Hello,
    Check this note:
    Note 751577 - APD-FAQ: Data source query
    This note helps with the following problems:
    Question 1: Query execution terminates or takes too long. What can I do?
    Question 2: The data in the analysis process does not correspond with the data in the Business Explorer. What is the cause of this?
    Question 3: When you execute the analysis process, the system issues error message B1 888: "For the ISO code &1 there is no unique SAP currency code for ALE". What does this mean?
    Question 4: When you execute the analysis process, the system issues error message RSAN_WB 010 (CX_RSAN_SVC_CHECK_ANALYSIS_N=>NODE_FIELDNAME_NOT_UNQIUE). What does this mean?
    Regards,
    Jorge Diogo

  • Remote Query execution option

    Hello Experts,
    I have installed and configured Intercompany 2.0 for testing, during which I could not find remote query execution option which was there in earlier version.
    Is there any setting to be done to enable this feature in new version.
    Thanks
    Deepak

    Hi Deepak,
    To access the Remote query execution feature of Intercompany, you are required to enable the SAP B1 user for the same. Follow the steps below to enable Intercompany solution Remote Query Execution feature:
    Log in SAP B1, navigate to Administration--> Setup--> Users and search the user.
    On Users-Setup screen, select Can Use Remote Query Execution check box and choose Update button.
      3. Now, using this user, login SAP B1 and open Query Manager window.
    4. Now on Query Manager window; Select a query, choose either of following options and execute          the query:
         Local: To execute and pull data from the logged in company.
         ALL   : To execute and pull data from all intercompany configured companies.
         Selection: To execute and pull data from a selected set of intercompany configured companies.
    Refer to Intercompany Integration Solution User Guide for further information related to Remote query execution.
    Regards,
    Agneesh Jain
    SAP Intercompany Team

  • BW statistics: Query Execution Time per User

    I am struggling to get something meaningful out of the BW statistics. The only information I really need is:
    The time from the first call of a query until the first selection screen and
    The time from the execution request until the data is displayed (i.e. after filling selection screen)
    With the possibility to see this information by user.
    So with other words:  KF Duration of query call + KF Duration of query execution, with Characteristics: user, period, query
    This could sound like a very simple universal query, but indeed absolutely not straightforward when looking at statistics contents. Any hints on the KF calculations are most welcome.
    The BW version in 7.0, with Bex as well as WAD reporting.

    I was looking for the same KFs, but did not find anything meaningful in ST03N as well.
    But I would recommend to execute the Web Report (from WAD or from Portal, if you are using Enterprise Portal) and then look at table view RSDDSTAT_OLAP. In this table view, BW Statistic data for each step is saved on event level. So if you execute the Web Report, you will find entries for the step corresponding to displaying the selection screen and for the step of generating the Web Report after the selection screen. The column RUNTIME displays the runtime of each step, which is a good indicator for the average runtime KFs you want to have.
    Maybe there are more meaningful KFs in technical BI Content, but I did not find them yet.

  • Explain the different steps involved in the process.

    very simple question.
    When you execute a query in oracle what happens. Explain the different steps involved in the process.

    Check out the documentation, explains it all:
    Overview of SQL Processing

  • 0BWTC_C02 - Viewing number of query executions and refreshes

    Hi,
    I having been testing some queries I created for infocube 0BWTC_C02.  Is there any way to limit the key figure 0TCTNAVCTR to only show navigations that were query refreshes (e.g. ignoring navigations such as removing/adding a drilldown, results row, ect.)?  Do any of the available characteristics show what type of navigation was being performed, so that I could restrict the query to only look at certain types of navigations?
    Thanks,
    Andy

    Hi Pcrao,
    I am using the BW statistics infocube 0BWTC_C02.
    But as far as I see, the key figure 0TCTNAVCTR( number of navigation steps)now only includes query execution, but also includes all other navigation steps like drill down etc.
    Can we do some restriction on navigation type etc to find just the number times query execution takes place?
    Thanks,
    James

  • Methods to reduce query execution time

    hi experts
                  can anybody sugest the steps /methods to reduce the time taken for query execution
    Thanks and regards
    Pradeep

    Hi Pradeep.........
    I think u hav already posted a similar thread.........
    query and load performance steps
    Anyways also check these notes......
    SAP Note : 557870............ 'FAQ BW Query Performance'
    567746 'Composite note BW 3.x performance Query and Web'.
    How to design a good query........
    /people/prakash.darji/blog/2006/01/27/query-creation-checklist
    /people/prakash.darji/blog/2006/01/26/query-optimization
    Also Check this.......
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/cccad390-0201-0010-5093-fd9ec8157802
    Business Intelligence Performance Tuning [original link is broken]
    Query Performance Improvement Tools
    Query Performance Improvement Tools
    Regards,
    Debjani........

  • Query Execution during Data Loads (extarction)

    I think BI 7.0 permitts that but I would still like to confirm from Gurus.
    Can we/users continue to access data or execute queries while extraction is going on?? Can we load data during query execution?
    What are the pros and cons of doing that??
    Always appreciative of your help.
    Suresh

    Hi,
    Query execution will not really hamper data loading or vice versa. But freshly loaded data would not be available for reporting before it gets activated in the infoprovider. Also in case of a cube, if the 'delete overlapping requests' step is to be performed, there could be erroneous looking data in the report till the time this step runs - that is, between the time when the new load has come in and the old request is deleted. That is why loads are best scheduled when users are not working on the system.

  • Data Driven Subscriptions Error - the query processor could not start the necessary thread resources for parallel query execution

    Hi,
    We are getting the following error when certain data driven subscriptions are fired off: "the query processor could not start the necessary thread resources for parallel query execution".  I've read other posts that have the same error, and
    the solution usually involves adjusting MaxDOP to limit the number of queries that are fired off in parallel.  
    Unfortunately, we cannot change this setting on our server for performance reasons (outside of data driven subscriptions, it negatively impacts our ETL processing times).  We tried putting query hints like "OPTION (MAXDOP 2);" in the reports
    that are causing the error, but it did not resolve the problem.
    Are there any settings within Reporting Services that can be adjusted to limit the number of subscriptions that get fired off in parallel?
    Any help is appreciated - thanks!

    Yes, that is correct.  It's a painful problem, because you don't know which specific subscription failed. For example, we have a data driven subscription that sends out about 800 emails. Lately, we've been having a handful of them fail. You don't know
    which ones out of the 800 failed though, even from the RS log files - all it tells you is that "the
    query processor could not start the necessary thread resources for parallel query execution".
    Thanks, I'll try changing <MaxQueueThreads> and will let you know if it works.
    On a side note: I've noticed that it is only reports with cascading parameters (ex. where parameter 2 is dependent on the selection from parameter 1) that get this error message...

  • Issue while query execution on web analyser.

    Hi,
    I am getting an error message while query execution on web ie Record set too large , data retrieval restricted by configuration .I am able to run the same query in bex analyser without any issue .Any idea what could be the reason and solution for this issue .
    Regards,
    Neetika.

    Hi Neetika,
    The Query is exceeding the set limits,i suggest you to Reduce the time LIne for the Query, as it may be having more number of Cells in terms of Rows and Columns.
    Execute the Query for Less number of Days,if u r executing it for 1 Month then execute it for 10 Days.
    Rgds
    SVU123

  • SSRS Error : Query execution failed for dataset 'dataSet'. (rsErrorExecutingCommand) Semantic query execution failed. Invalid object name 'RPT.*********'. (rsSemanticQueryEngineError)

    I am new to SSRS and I am trying to migrate reports from 2008 to 2012. As I have so many reports to migrate, I simply got the back up of ReportServer,
    ReportServerTempDB, and Encryption Key and restored them to test environment. I made necessary configuration from RS configuration tool. I am able to see the reports now when I browse //hostname/reports. But when I open any particular report I am getting some
    error.
    · An error has occurred during report processing.
    (rsProcessingAborted)
    Query execution       failed for dataset 'dataSet'.
          (rsErrorExecutingCommand
    Semantic query        execution failed. Invalid object name
           'RPT. ******'. (rsSemanticQueryEngineError)
    ****** - I am assuming this is a custom data class.
    Does anyone have insight on this? or any better way that I can migrate the reports to new server with less efforts.
    I don’t have the reports solution file to deploy the reports, so I have followed backup and restore process.

    Hi Kishore237,
    According to your description, you migrated some reports from Reporting Services (SSRS) 2008 to 2012. Now you get error when accessing the reports on SSRS 2012. Right?
    In this scenario, did you modify the report data source in database after migration? You can try to open the report in Report Builder or Report designer and check the report dataset. If you can preview the report in Report builder or Report designer,
    please try to redeploy the report to Report Server. If it is still not working, please try to restore the database from backup. And for migrating reports, please follow the "Content-Only Migration" in the link below:
    http://msdn.microsoft.com/en-us/library/ms143724(v=sql.110).aspx
    If you have any question, please feel free to ask.
    Best Regards,
    Simon Hou

Maybe you are looking for

  • Preview not working in 10.6

    Preview not working in 10.6. Not able to open multiple JPG files. No problem in 10.5.7. Any solution?

  • CANNOT CHANGE PLANNING DATE

    HI GURUS, We posted a SD document 180000XXXX Base line date 16.12.2008 and Payt terms C003 (30 days) thE planning date must be 15.01.2009 but the Planning date is 12.01.2008. We intented change the base line date, the Payment terms and the Planning d

  • Transaction code ME1L

    Hi I Need to make changes in program RM06IL00. Tcode for this program is ME1L. I Want to apply ALV to the output format. Pls. help to select data/fields which can be placed in an internal table for further  use in ALV. Thanks & Regards Harshada

  • When I try to print a 4X6 photo using Elements 12 it prints out 2X3.  Does anyone know how you can get hold a live person at Adobe?

    When attempting to print a 4X6 photo using Elements 12 it prints out 2X3.  Does anyone know how you can get hold of a live person at Adobe?

  • Best Composite Out Solution?

    I have a Mac Pro with FCP. I would like to have a card that does composite out (while not replacing the current card i have, 512mb ATI RAEDEON). What is the cheapest solution for this? Basically I just want to plug into my SD TV to look at the image