Restricting a Query execution

Hi,
I'am using Report Builder 6.0.8.11.3. I'm getting parameters from the user and want to execute specific query based on it.
Is it possible to restrict the execution of a query at runtime....
Thank you....

I don't know whether your problem is solved or not? I did the similar work combining two different tasks in a report.
But here is the solution, yes you can enter a parameter at runtime to select a specific query and ignore the rest.
Create a parameter as OPTION which takes list of values.
for example as 'SUMMARY' and 'DETAIL'( You can see List of values in properties of parameter).
For example if you have summary and detail reports as different queries. you can combine them as one.
in summary report query add :p_option = 'SUMMARY'
in detail report query add :p_option = 'DETAIL'
And on the main frame of the layout u can include a pl/sql code to select that specific report.
Example:
========== for summary report ==========
function M_1FormatTrigger return boolean is
begin
IF :P_OPTION = 'SUMMARY'
THEN
RETURN(TRUE);
ELSE
RETURN (FALSE);
END IF;
return (TRUE);
end;
=========for detail report =============
function M_2FormatTrigger return boolean is
begin
IF :P_OPTION = 'DETAIL'
THEN
RETURN(TRUE);
ELSE
RETURN (FALSE);
END IF;
return (TRUE);
end;
Hope this gives an idea!! If you have any other method to achieve that do let me know.
Message was edited by:
Prathima
Message was edited by:
Prathima

Similar Messages

  • Unwanted query execution before adding Find page restrictions

    Hello,
    When the JHS (10.1.2.1.27) application I made is started, the whole table content is queried without restrictions BEFORE the Find page has loaded.
    Because it's a fair amount of data, I get out of memory messages.
    I have used 2 specific settings that are related to this:
    1) Property "Use Table Range?" has been turned off
    2) Query Bind Parameters are used
    Looking in the log files I noticed the following:
    13:24:00 DEBUG (JhsActionServlet) -Request URI: /PppHseSisQuery/FindPppVLastBhy.do
    13:24:00 DEBUG (JhsActionServlet) -Request Character Encoding: ISO-8859-1
    13:24:04 DEBUG (JhsDataAction) -Executing action /FindPppVLastBhy
    13:24:04 DEBUG (JhsDataAction) -Created searchBean map and stored on session
    13:24:04 DEBUG (JhsDataAction) -Created new searchBean for PppVLastBhyUIModel and added to quick search bean map
    13:24:04 DEBUG (JhsDataAction) -Stored searchBean for PppVLastBhyUIModel on request
    13:24:04 DEBUG (JhsDataAction) -ViewObject PppVLastBhyView: value of bind param 0 set to null
    13:24:04 DEBUG (JhsDataAction) -ViewObject PppVLastBhyView: value of bind param 1 set to null
    13:24:04 DEBUG (JhsDataAction) -ViewObject PppVLastBhyView: value of bind param 2 set to null
    13:24:04 DEBUG (JhsDataAction) -ViewObject PppVLastBhyView: value of bind param 3 set to null
    13:24:04 DEBUG (JhsDataAction) -ViewObject PppVLastBhyView: value of bind param 4 set to null
    13:24:04 DEBUG (JhsDataAction) -ViewObject PppVLastBhyView: value of bind param 5 set to null
    13:24:04 DEBUG (JhsDataAction) -ViewObject PppVLastBhyView: value of bind param 6 set to null
    13:24:04 DEBUG (JhsDataAction) -ViewObject PppVLastBhyView: executing query, bind parameter values have changed
    It seems that the bind parameters get initialized in some kind of way at startup.
    This is detected as a change in value apparently, after which the query execution is triggered.
    So far so good. The problem is that my main query restriction is not related to any of the query bind parameters.
    Because the Find page has not been loaded yet, this mandatory restriction has not been added to the searchBean yet and will therefore not be added to the view object's where clause.
    The result is that all table records will be queried.
    My question: Is there any way to prevent the query to be executed before the Find page has loaded?

    Have you tried setting the group checkbox "Auto Query?" to unchecked? This should prevent the query from executing before the find page is loaded.
    Steven Davelaar,
    JHeadstart Team.

  • Issue while query execution on web analyser.

    Hi,
    I am getting an error message while query execution on web ie Record set too large , data retrieval restricted by configuration .I am able to run the same query in bex analyser without any issue .Any idea what could be the reason and solution for this issue .
    Regards,
    Neetika.

    Hi Neetika,
    The Query is exceeding the set limits,i suggest you to Reduce the time LIne for the Query, as it may be having more number of Cells in terms of Rows and Columns.
    Execute the Query for Less number of Days,if u r executing it for 1 Month then execute it for 10 Days.
    Rgds
    SVU123

  • Query execution time

    Dear SCN,
    I am new to BOBJ Environment. I have created a webi report on top of bex query by using BISC connection. Bex query is build for Vendor Ageing Analysis. My bex query will take very less time to execute the report (max 1 min). But in case of webi is takeing around 5 min when i click on refresh. I have not used any conditions,filters,restrictions are done at webi level all are done at bex level only.
    Please let me know techniques to optimize the query execution time in webi. Currently we are in BO 4.0.
    Regards,
    PRK

    Hi Praveen
    Go through this document for performance optimization using BICS connection
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/d0e3c552-e419-3010-1298-b32e6210b58d?QuickLink=index&…

  • BEX problem: F4 help to filter values of F4 value list at query execution

    Hi all,
    description of the problem we face:
    at query execution the selection pop-up appears and "ask" for values. To select those values I try the F4 button and receive a long list of e.g. vendors. (Our vendors are compounded to a special characteristic due to several sources.) To select now several individual values from the list I choose the filter button and try to shorten the list by searching with values of the atributes, here by 0NAME of the vendor. From the result shown I choose one value compounded to a unique value of the special characteristic, there are values for vendors which are the same, but  with different compound values and of course different meanings.
    When I mark one vendor value shown with his compounded value in front of the row, I expect the compounded value will be transferred to the selection pop-up too, but it won't!!
    Only the vendor is transferred, so the result of the query shows more information then wanted.
    Any solutions for this?
    Thank you in advance

    Hi Guido,
    I dont know any direct solution to your problem, a workaround is
    May be add a variable on the Special characterstics, and ask users to enter the Special character also along with your Vendor Code.
    As while selecting the vendor user already know special charater value, then can just enter it in Selection screen and restrict the output.
    Thanks
    CK

  • Analyzer Query Execution vs. APD Query Execution

    hi gurus,
    i have a problem with an execution of a query in BI 7.0. the query is based on a DSO.
    if i execute the query in bex analyzer, i get the result within 30 seconds.
    but if i execute the query in an APD, it will take several hours to complete the job.
    i checked the RSRT to view the query execution plan - here it is:
    SELECT "O1"."COMP_CODE" AS "K____240" ,"O1"."RECONTRACT" AS "K____301" ,"O1"."FISCVARNT" AS "K____181"
    ,"O1"."/BIC/ZIMIMNPER" AS "K____535" ,"O1"."BUSENTITY" AS "K____237" ,"O1"."/BIC/ZIMIMGEBL" AS "K____1098"
    , COUNT( * )  AS "Z____131" , SUM (  "O1"."/BIC/ZIMIK_113"  )  AS "Z____1886" ,
    SUM (  "O1"."/BIC/ZIMIK_183"  )  AS "Z____1908" , SUM (  "O1"."/BIC/ZIMIK_201"  )  AS "Z____1998"
    FROM "/BIC/AZIMODSSM00" "O1"
    WHERE
    ( ( ( ( "O1"."/BIC/ZIMIMNPER" = '2010009'
    ) ) AND  ( ( "O1"."/BIC/ZIMIMZBR" IN ( 'BEQ' ,'REFX3' ,'REFX4' ,'REFX6') ) ) ) )
    GROUP BY "O1"."COMP_CODE" ,"O1"."RECONTRACT" ,"O1"."FISCVARNT" ,
    "O1"."/BIC/ZIMIMNPER" ,"O1"."BUSENTITY" ,"O1"."/BIC/ZIMIMGEBL"
    but when i check the SQL statement for the APD in SM51, i found this statement:
    SELECT   "S0000"."SID" ,"P0000"."COMP_CODE" ,"P0000"."BUSENTITY"
    FROM   "/BI0/PBUSENTITY" "P0000" JOIN "/BI0/SBUSENTITY" "S0000" ON  "P0000" . "BUSENTITY" = "S0000" . "BUSENTITY" AND "P0000" . "COMP_CODE" = "S0000" . "COMP_CODE"
    WHERE   "P0000"."OBJVERS" = 'A' AND "S0000"."SID" IN ( SELECT "S"."SID" FROM "/BIC/AZIMODSSM00" "O" JOIN
      "/BI0/SBUSENTITY" "S" ON  "O" . "BUSENTITY" = "S" . "BUSENTITY" AND "O" . "COMP_CODE" = "S" .
      "COMP_CODE" )
    ORDER BY
      "COMP_CODE" ASC , "BUSENTITY" ASC#
    could you help me understanding the difference and the cause for it?
    i think, there are some validitycheckings executed for masterdata!? is it possible to disable these steps?
    here some more information:
    in the query, i'm using two compound characteristics and an constant selection for a restricted keyfigure. the characteristics all have access type for result values: posted values.
    regards,
    philipp

    Hello,
    Check this note:
    Note 751577 - APD-FAQ: Data source query
    This note helps with the following problems:
    Question 1: Query execution terminates or takes too long. What can I do?
    Question 2: The data in the analysis process does not correspond with the data in the Business Explorer. What is the cause of this?
    Question 3: When you execute the analysis process, the system issues error message B1 888: "For the ISO code &1 there is no unique SAP currency code for ALE". What does this mean?
    Question 4: When you execute the analysis process, the system issues error message RSAN_WB 010 (CX_RSAN_SVC_CHECK_ANALYSIS_N=>NODE_FIELDNAME_NOT_UNQIUE). What does this mean?
    Regards,
    Jorge Diogo

  • 0BWTC_C02 - Viewing number of query executions and refreshes

    Hi,
    I having been testing some queries I created for infocube 0BWTC_C02.  Is there any way to limit the key figure 0TCTNAVCTR to only show navigations that were query refreshes (e.g. ignoring navigations such as removing/adding a drilldown, results row, ect.)?  Do any of the available characteristics show what type of navigation was being performed, so that I could restrict the query to only look at certain types of navigations?
    Thanks,
    Andy

    Hi Pcrao,
    I am using the BW statistics infocube 0BWTC_C02.
    But as far as I see, the key figure 0TCTNAVCTR( number of navigation steps)now only includes query execution, but also includes all other navigation steps like drill down etc.
    Can we do some restriction on navigation type etc to find just the number times query execution takes place?
    Thanks,
    James

  • SSRS Error : Query execution failed for dataset 'dataSet'. (rsErrorExecutingCommand) Semantic query execution failed. Invalid object name 'RPT.*********'. (rsSemanticQueryEngineError)

    I am new to SSRS and I am trying to migrate reports from 2008 to 2012. As I have so many reports to migrate, I simply got the back up of ReportServer,
    ReportServerTempDB, and Encryption Key and restored them to test environment. I made necessary configuration from RS configuration tool. I am able to see the reports now when I browse //hostname/reports. But when I open any particular report I am getting some
    error.
    · An error has occurred during report processing.
    (rsProcessingAborted)
    Query execution       failed for dataset 'dataSet'.
          (rsErrorExecutingCommand
    Semantic query        execution failed. Invalid object name
           'RPT. ******'. (rsSemanticQueryEngineError)
    ****** - I am assuming this is a custom data class.
    Does anyone have insight on this? or any better way that I can migrate the reports to new server with less efforts.
    I don’t have the reports solution file to deploy the reports, so I have followed backup and restore process.

    Hi Kishore237,
    According to your description, you migrated some reports from Reporting Services (SSRS) 2008 to 2012. Now you get error when accessing the reports on SSRS 2012. Right?
    In this scenario, did you modify the report data source in database after migration? You can try to open the report in Report Builder or Report designer and check the report dataset. If you can preview the report in Report builder or Report designer,
    please try to redeploy the report to Report Server. If it is still not working, please try to restore the database from backup. And for migrating reports, please follow the "Content-Only Migration" in the link below:
    http://msdn.microsoft.com/en-us/library/ms143724(v=sql.110).aspx
    If you have any question, please feel free to ask.
    Best Regards,
    Simon Hou

  • Unable to select the filter value after query execution

    hi,
    I am unable to drill down my keyfigures.
    1.i have a lC and GC values out of which GC is hidden so after the query execution i would like to filter my value between GC and LC but i do get only LC value in the filter and i do not get GC value to select .
    Could any one tell me how this can be done.this is very urgent.
    Thx
    Subha

    resolved on my own,so i am closing this.
    Thx
    Subha

  • Asset query execution performance after upgrade from 4.6C to ECC 6.0+EHP4

    Hi,guys
    I am encounted a weird problems about asset query execution performance after upgrade to ECC 6.0.
    Our client had migrated sap system from 4.6c to ECC 6.0. We test all transaction code and related stand report and query.
    Everything is working normally except this asset depreciation query report. It is created based on ANLP, ANLZ, ANLA, ANLB, ANLC table; there is also some ABAP code for additional field.
    This report execution costed about 6 minutes in 4.6C system; however it will take 25 minutes in ECC 6.0 with same selection parameter.
    At first, I am trying to find some difference in table index ,structure between 4.6c and ECC 6.0,but there is no difference about it.
    i am wondering why the other query reports is running normally but only this report running with too long time execution dump messages even though we do not make any changes for it.
    your reply is very appreciated
    Regards
    Brian

    Thanks for your replies.
    I check these notes, unfortunately it is different our situation.
    Our situation is all standard asset report and query (sq01) is running normally except this query report.
    I executed se30 for this query (SQ01) at both 4.6C and ECC 6.0.
    I find there is some difference in select sequence logic even though same query without any changes.
    I list there for your reference.
    4.6C
    AQA0FI==========S2============
    Open Cursor ANLP                                    38,702  39,329,356  = 39,329,356      34.6     AQA0FI==========S2============   DB     Opens
    Fetch ANLP                                         292,177  30,378,351  = 30,378,351      26.7    26.7  AQA0FI==========S2============   DB     OpenS
    Select Single ANLC                                  15,012  19,965,172  = 19,965,172      17.5    17.5  AQA0FI==========S2============   DB     OpenS
    Select Single ANLA                                  13,721  11,754,305  = 11,754,305      10.3    10.3  AQA0FI==========S2============   DB     OpenS
    Select Single ANLZ                                   3,753   3,259,308  =  3,259,308       2.9     2.9  AQA0FI==========S2============   DB     OpenS
    Select Single ANLB                                   3,753   3,069,119  =  3,069,119       2.7     2.7  AQA0FI==========S2============   DB     OpenS
    ECC 6.0
    Perform FUNKTION_AUSFUEHREN     2     358,620,931          355
    Perform COMMAND_QSUB     1     358,620,062          68
    Call Func. RSAQ_SUBMIT_QUERY_REPORT     1     358,569,656          88
    Program AQIWFI==========S2============     2     358,558,488          1,350
    Select Single ANLA     160,306     75,576,052     =     75,576,052
    Open Cursor ANLP     71,136     42,096,314     =     42,096,314
    Select Single ANLC     71,134     38,799,393     =     38,799,393
    Select Single ANLB     61,888     26,007,721     =     26,007,721
    Select Single ANLZ     61,888     24,072,111     =     24,072,111
    Fetch ANLP     234,524     13,510,646     =     13,510,646
    Close Cursor ANLP     71,136     2,017,654     =     2,017,654
    We can see first open cursor ANLP ,fetch ANLP then select ANLC,ANLA,ANLZ,ANLB at 4.C.
    But it changed to first select ANLA,and open cursor ANLP,then select  ANLC,ANLB,ANLZ,at last fetch ANLP.
    Probably,it is the real reason why it is running long time in ECC 6.0.
    Is there any changes for query selcection logic(table join function) in ECC 6.0.

  • An error has occurred during report processing. (rsProcessingAborted). Query execution failed for dataset 'DimUserWorkCentre'. (rsErrorExecutingCommand). The Cube either does not exists or has not been processed

    Hi,
    I'm having issues with the report created using SSAS cube.
    An error has occurred during report processing. (rsProcessingAborted)
    Query execution failed for dataset 'DimUserWorkCentre'. (rsErrorExecutingCommand)
    The Operator_Performance cube either does not exist or has not been processed.
    I have searched through internet and tried all the solutions, but didn't worked for me.
    SSRS services running as NETEWORK SERVICE user.
    SSRS Execution running as a Different user, which is the login is used to logon ot that server. I have also verified this user has access to database. I'm using Shared DataSource(SSAS Source) for this report.
    Can any one please help me.
    Thank You,
    Praveen.
    Praveen

    Hello,
    Have you tried it to execute on report manager , Is your data source properly configured in Report Manager  and your report is mapped with Datset correctly?
    Have you executed the Dataset query MDX editor  now?
    What is the volume of data you are fetching in the report. Try it to execute in other than IE , I don't know the exact reason but some of our report with large volume of data  are failing on IE , on the other hand these reports are running fine Google
    Chrome
    blog:My Blog/
    Hope this will help you !!!
    Sanjeewan

  • Steps involved in Query execution

    Can anyone please suggest the exact steps that take place in a query execution. I know on the broad level. But I want to know to basic level. For example we would like to know all the steps that an OLAP processor does to get the data, process it and then send it for display.
    Also, say the total time taken for the query execution is xx msec, we would like to know the complete break up of this time for all the activities that are completed from the time of query execution till the display of result.
    Thanks

    Hi Suchitra,
    Please go through these links:
    /people/vikash.agrawal/blog/2006/04/26/in-what-147sequence148-olap-processor-look-for-query-relevant-data
    /people/prakash.darji/blog/2006/01/26/query-optimization
    Assign points if helpful
    Thanks,
    Amith

  • An error has occurred during report processing. (rsProcessingAborted).. Query execution failed for data set

    Hi All,
    I'm facing a strange problem..
    I've developed few reports. they are working fine in develop environment. after successfull testing they were published on web.
    in web version, all reports are executing for first time.. if I change any of parameters values or without chaning also..
    if I press "View Report"  following error occurs..
    An error has occurred during report processing. (rsProcessingAborted)
    Query execution failed for data set 'dsMLGDB2Odbc'. (rsErrorExecutingCommand)
    For more information about this error navigate to the report server on the local server machine, or enable remote errors
    please suggest any alternative ways to overcome this issue
    thanks in adv.

    in my case the problem is
    one virtual machine is for developers
    other for testers
    in developers i created a report, then save like *.rdl and copy to testers machine, does not work there
    the error what testers get is
    Error during the local report processing.
    Could not find a web-based application at http://developersMachine/AnalyticsReports/DataBaseConnector.rsds
    and the solution is to use alternative url or in some cases http://localhost/

  • An error has occurred during report processing. (rsProcessingAborted) Query execution failed for dataset 'dsPriority'. (rsErrorExecutingCommand)

    click report error:
    log file:
    processing!ReportServer_0-2!104c!04/27/2015-19:15:21:: e ERROR: Throwing Microsoft.ReportingServices.ReportProcessing.ProcessingAbortedException: An error has occurred during report processing., ;
     Info: Microsoft.ReportingServices.ReportProcessing.ProcessingAbortedException: An error has occurred during report processing.
     ---> Microsoft.ReportingServices.ReportProcessing.ReportProcessingException: Query execution failed for dataset 'dsPriority'.
     ---> Microsoft.AnalysisServices.AdomdClient.AdomdErrorResponseException: Query (26, 25) The ALLMEMBERS function expects a hierarchy expression for the  argument. A member expression was used.
       at Microsoft.AnalysisServices.AdomdClient.AdomdDataReader..ctor(XmlReader xmlReader, CommandBehavior commandBehavior, AdomdConnection connection)
       at Microsoft.AnalysisServices.AdomdClient.AdomdCommand.ExecuteReader(CommandBehavior behavior)
       at Microsoft.AnalysisServices.AdomdClient.AdomdCommand.System.Data.IDbCommand.ExecuteReader(CommandBehavior behavior)
       at Microsoft.ReportingServices.DataExtensions.AdoMdCommand.ExecuteReader(CommandBehavior behavior)
       at Microsoft.ReportingServices.OnDemandProcessing.RuntimeDataSet.RunDataSetQuery()
       --- End of inner exception stack trace ---
       at Microsoft.ReportingServices.OnDemandProcessing.RuntimeDataSet.RunDataSetQuery()
       at Microsoft.ReportingServices.OnDemandProcessing.RuntimeDataSet.Process()
       at Microsoft.ReportingServices.OnDemandProcessing.RuntimeParameterDataSet.Process()
       at Microsoft.ReportingServices.OnDemandProcessing.RuntimeDataSet.ProcessConcurrent(Object threadSet)
       --- End of inner exception stack trace ---
    open http://localhost:8080/tfs/TeamFoundation/Administration/v3.0/WarehouseControlService.asmx?op=ProcessWarehouse click Invoke:
    System.Web.Services.Protocols.SoapException: TF221029: Reporting for Team Foundation Server does not have any warehouse jobs defined. Use the Team Foundation Administration Console to rebuild the reporting. : 2015-04-27T19:30:29:782 ---> System.InvalidOperationException:
    TF221029: Reporting for Team Foundation Server does not have any warehouse jobs defined. Use the Team Foundation Administration Console to rebuild the reporting. at Microsoft.TeamFoundation.Warehouse.WarehouseAdmin.QueueJobs(String collectionName, String jobName)
    at Microsoft.TeamFoundation.Warehouse.WarehouseControlWebService.ProcessWarehouse(String collectionName, String jobName) --- End of inner exception stack trace --- at Microsoft.TeamFoundation.Warehouse.WarehouseControlWebService.ProcessWarehouse(String collectionName,
    String jobName)

    Hi shelman,
    I'd like to know whether you configured TFS reporting service properly, and if you can get the report normally before. From the error message, you might has wrong data source or the parameters of the dataset 'dsPriority'. You can check whether the data source
    for your report is available, try to use windows authentication. Or check the parameters of dataset 'dsPriority' make sure the query works when you execute it manually.
    To process the TFS data warehouse and analysis services cube, I'd like to know which operation you selected before clicking Invoke button. Please follow the instructions on this
    page to process TFS data warehouse and analysis service cube manually. You can also refer the James's last reply in this
    thread or this
    blog to check if the solutions work for you.
    Another option is run TFS best practice analyzer to check if there any configure issues on your TFS server machine. And check event logs to see if there any useful information, elaborate more details about your scenario including reproduce steps if the problem
    persists.
    Best regards,

  • Query Execution/Elapsed Time and Oracle Data Blocks

    Hi,
    I have created 3 tables with one column only. As an example Table 1 below:
    SQL> create table T1 ( x char(2000));
    So 3 tables are created in this way i.e. T1,T2 and T3.
    T1 = in the default database tablespace of 8k (11g v11.1.0.6.0 - Production) (O.S=Windows).
    T2 = I created in a Tablespace with Blocksize 16k.
    T3 = I created in a Tablespace with Blocksize 4k. In the same Instance.
    Each table has approx. 500 rows (So, table sizes are same in all the cases to test Query execution time ). As these 3 tables are created under different data block sizes so the ALLOCATED no. of data blocks are different in all cases.
    T1  =   8k  = 256 Blocks =  00:00:04.76 (query execution time/elapsed time)
    T2  = 16k=121 Blocks =  00:00:04.64
    T3 =   4k =  490 Blocks =  00:00:04.91
    Table Access is FULL i.e. I have used select * from table_name; in all 3 cases. No Index nothing.
    My Question is why query execution time is nearly the same in all 3 cases because Oracle has to read all the data blocks in each case to fetch the records and there is a much difference in the allocated no. of blocks ???
    In 4k block size example, Oracle has to read just 121 blocks and it's taking nearly the same time as it's taking to read 490 blocks???
    This is just 1 example of different data blocks. I have around 40 tables in each block size tablespace and the result are nearly the same. It's very strange for me because there is a much difference in the no. of allocated blocks but execution time is almost the same, only difference in milliseconds.
    I'll highly appreciate the expert opinions.
    Bundle of thanks in advance.
    Best Regards,

    Hi Chris,
    No I'm not using separate databases, it's 8k database with non-standard blocksizes of 16k and 4k.
    Actually I wanted to test the Elapsed time of these 3 tables, so for that I tried to create the same size
    tables.
    And how I equalize these is like I have created one column table with char(2000).
    555 MB is the figure I wanted to use for these 3 tables ( no special figure, just to make it bigger than the
    RAM used for my db at the db startup to be sure of not retrieving the records from cache).
    so row size with overhead is 2006 * 290,000 rows = 581740000(bytes) / 1024 = 568105KB / 1024 = 555MB.
    Through this math calculation I thought It will be the total table size. So I Created the same no. of rows in 3 blocksizes.
    If it's wrong then what a mes because I was calculating tables sizes in the same way from the last few months.
    Can you please explain a little how you found out the tables sizes in different block sizes.Though I understood how you
    calculated size in MB from these 3 block sizes
    T8K =97177 BLOCKS=759MB *( 97177*8 = 777416KB / 1024 = 759MB )*
    T16K=41639 BLOCKS=650MB
    BT4K=293656 BLOCKS=1147MB
    For me it's new to calculate the size of a table. Can you please tell me then how many rows I can create in each of
    these 3 tables to make them equal in MB to test for elapsed time.
    Then I'll again run my test and put the results here. Because If I've wrongly calculated table sizes then there is no need to talk about elapsed time. First I must equalize the table sizes properly.
    SQL> select sum(bytes)/1024/1024 "Size in MB" from dba_segments> 2 where segment_name = 'T16K';
    Size in MB
    655
    Is above SQL is correct to calculate the size or is it the correct alternative way to your method of calculating the size??
    I created the same table again with everything same and the result is :
    SQL> select num_rows,blocks from user_tables where table_name = 'T16K';NUM_ROWS BLOCKS
    290000 41703
    64 more blocks are allocated this time so may be that's y it's showing total size of 655 instead of 650.
    Thanks alot for your help.
    Best Regards,
    KAm
    Edited by: kam555 on Nov 20, 2009 5:57 PM

Maybe you are looking for

  • How to view the old data in a column if that column is dropped or rename

    Using workspace manager I am not able to track the old data of a column which is dropped. I have tried by using import and export features that Workspace manager have.. But import is not working properly in this case. And savepoint option is work wit

  • How can I tell if a message received is via iMessage?

    My daughter has an iPhone  4 with iOS6 on, and I have an iPhone 5 with iOS6 on. We both have iMessage enabled and use it. Now I can see when I have sent a message and its an iMessage as it is on blue. But how can tell if a message from my daughter is

  • Apple TV and old audio system

    Here's what I want to do. I want to put all the CDs on an external drive connected to the network (do I need a Time Machine or other Ethernet connected hard drive?) and share with everybody in the house. Then I want to be able to get the music to pla

  • Add an avergae row in a spreadsheet

    I would like to add an additional 5th row to a spreadsheet file that contains a row of headers and 3 rows of data. The 5th row will consist on the average values of the 3 corresponding cells above(rows 2, 3 and 4). See the attached files as example o

  • Iphone camera issue

    i have a green "casting"/hue when taking pictures... are there any setting to do white balancing? I have checked someone elses and it appeared fine....