Query time lengthened after additional indexes added

I had been frequently getting results from my query in about .5 - 1.5 seconds. However, I added additional indexes to my container on fields that I am not even using in my query at the moment. When I did this, my query went to about 35 seconds. Can anyone tell me how to fix this problem
Here is my query:
for $record in collection("Catalog.dbxml")
/*[T245[contains(.,"augustine") or contains(.,"Augustine") or contains(.,"AUGUSTINE")]] return $record
Here are my indexes:
Index: node-element-substring-string for node {}:T050
Index: node-element-substring-string for node {}:T090
Index: node-element-substring-string for node {}:T100
Index: node-element-substring-string for node {}:T110
Index: node-element-substring-string for node {}:T111
Index: node-element-substring-string for node {}:T245
Index: node-element-substring-string for node {}:T246
Index: node-element-substring-string for node {}:T700
Index: node-element-substring-string for node {}:T710
Index: node-element-substring-string for node {}:T711
Here is my queryplan:
<XQuery>
<Navigation>
<QueryPlanFunction result="collection" container="Catalog5.dbxml">
<OQPlan>n(V(node-element-substring-string,T245,=,'ine'),V(node-element-substring-string,T245,=,'tin'),V(node-element-substring-string,T245,=,'sti'),V(node-element-substring-string,T245,=,'ust'),V(node-element-substring-string,T245,=,'gus'),V(node-element-substring-string,T245,=,'aug'),V(node-element-substring-string,T245,=,'ugu'))</OQPlan>
</QueryPlanFunction>
<Step axis="child" uri="*" name="*" nodeType="element"/>
<DbXmlFilter>
<Navigation>
<Step axis="child" name="T245" nodeType="element">
<Predicates>
<Operator name="or">
<Navigation>
<DbXmlContains>
<OQPlan>n(V(node-element-substring-string,T245,=,'ugu'),V(node-element-substring-string,T245,=,'gus'),V(node-element-substring-string,T245,=,'ine'),V(node-element-substring-string,T245,=,'sti'),V(node-element-substring-string,T245,=,'tin'),V(node-element-substring-string,T245,=,'ust'),V(node-element-substring-string,T245,=,'aug'))</OQPlan>
<Sequence>
<AnyAtomicTypeConstructor value="augustine" typeuri="http://www.w3.org/2001/XMLSchema" typename="string"/>
</Sequence>
</DbXmlContains>
</Navigation>
<Navigation>
<DbXmlContains>
<OQPlan>n(V(node-element-substring-string,T245,=,'ine'),V(node-element-substring-string,T245,=,'tin'),V(node-element-substring-string,T245,=,'ust'),V(node-element-substring-string,T245,=,'sti'),V(node-element-substring-string,T245,=,'gus'),V(node-element-substring-string,T245,=,'ugu'),V(node-element-substring-string,T245,=,'aug'))</OQPlan>
<Sequence>
<AnyAtomicTypeConstructor value="Augustine" typeuri="http://www.w3.org/2001/XMLSchema" typename="string"/>
</Sequence>
</DbXmlContains>
</Navigation>
<Navigation>
<DbXmlContains>
<OQPlan>n(V(node-element-substring-string,T245,=,'ust'),V(node-element-substring-string,T245,=,'aug'),V(node-element-substring-string,T245,=,'ugu'),V(node-element-substring-string,T245,=,'gus'),V(node-element-substring-string,T245,=,'tin'),V(node-element-substring-string,T245,=,'sti'),V(node-element-substring-string,T245,=,'ine'))</OQPlan>
<Sequence>
<AnyAtomicTypeConstructor value="AUGUSTINE" typeuri="http://www.w3.org/2001/XMLSchema" typename="string"/>
</Sequence>
</DbXmlContains>
</Navigation>
</Operator>
</Predicates>
</Step>
</Navigation>
</DbXmlFilter>
</Navigation>
</XQuery>

Hi Andrew,
You experience surprises me a little - I'd like to look into reproducing what you are seeing.
Can you tell me which indexes you had to start with and which indexes you then added to cause the slow down? Is it possible for you to send me some sample data that exhibits this behavior - either the XML, or your container? If so, can you upload it to ftp.sleepycat.com/incoming and let me know what you've named the file.
Thanks,
John

Similar Messages

  • Restricting a characteristic & query time out problem

    Hi. We have the follwoing problem:
    Before our BI upgrade, we have had a number of users querying different info areas, that are now having difficulty restricting on a characteristic in a query.  This was performed with no problems before the upgrade. 
    Many times, the list of records for the particular characteristic is quite small.  Either it takes 20-30 minutes for the list to appear, or the query times out (after 6000 seconds).
    Any ideas?

    Any time you have an existing query that has been running quickly in production, and it suddenly starts to run much longer, you should be suspicious of some change to DB statisitics and or indexes.
    So the first thing to do is work with your DBA to make sure the DB statistics are current for the tables involved in the query, then get an Explain Plan for the query which will show you how the DB is trying to execute the query, what indexes it uses, etc.  Perhaps stats are not being refreshed after the upgrade, or something happened to an index

  • How to speed up the query time after adding order by

    Hi,
    In Form 6.0 front-end and 8i server, I put an "order by date
    column desc" in Block Property and when I hit query button the
    screen will bring up the most current date data instead of
    oldest date data. The problem is slowing down the query time,
    because the "order by" does the resorting work in the user temp
    table space. The base table to be queried also contains two
    varchar2(1500) columns and that also slow down the query time.
    I am wondering if there is any other way to bring the current
    records first without slow down the query speed ?
    Any thought will be helpful and appreciated.
    Thnaks!
    Tony
    null

    anthony hsu (guest) wrote:
    : Hi,
    : In Form 6.0 front-end and 8i server, I put an "order by date
    : column desc" in Block Property and when I hit query button the
    : screen will bring up the most current date data instead of
    : oldest date data. The problem is slowing down the query time,
    : because the "order by" does the resorting work in the user
    temp
    : table space. The base table to be queried also contains two
    : varchar2(1500) columns and that also slow down the query time.
    : I am wondering if there is any other way to bring the current
    : records first without slow down the query speed ?
    : Any thought will be helpful and appreciated.
    : Thnaks!
    : Tony
    What I have done is use hints to force the use of an index in
    descending order. The part of the following example that
    pertains to your question is the index_desc hint. Syntax and
    suggestions for using hints are in the Oracle documentation.
    The column you want to order desc by would have to be in an
    index. The following example creates a view which you might not
    need to do.
    create or replace view v_sample_result as
    select
    /*+
    first_rows
    index ( sa )
    index_desc ( t c_test_altk )
    index ( r c_result_pk )
    index ( lt )
    null

  • Query slows down after second run for Index Organised Tables

    We are trying to optimise our application which supports MSSQL to run with Oracle 9i for one of our customers.
    We have created one database with normal tables and PK constraints/indexes and turned caching on for the tables, this seems to work well but no way as fast as MSSQL on similar hardware. The first run of query was slower but as the caching became more effective the query times came down.
    So we investigated turning the tables into Index Organised Tables. We ran analyze on the new indexed tables and the response time of one of our more complex queries became akin to MSSQL. We ran the same query 5 seconds later and it took about 3 times longer to return the same data. Subsequent runs produced the same result.
    We have run the same query on both styles of tables and also run showplans on the two queries, the regular table returns a cost of 190 and the IOT 340. This would point to the fact that we should use the regular tables for our queries but why did the IOT set return much faster for the first run after the analyze then slow down as if the stats were missing, but the execution plan remain the same.
    Any help would be appreiciated.
    Darren Fitzgibbon

    could be a lot of reasons:
    1. Is Oracle the only process that runs on this server? Could it be any other process (i.e. MSSQL) that took the the resources during the secodn run?
    2. Is this the only query that was running during your tests? Could there be another query that put the load on the database when you were running the second test?
    3. The autotrace statistics and explain plan would be useful for first and second run
    (how to use autotrace:
    http://download-west.oracle.com/docs/cd/B10501_01/server.920/a96533/autotrac.htm#1018
    4. If you run the same process again several times - how does the response time change?
    Mike

  • Weird result of oracle query before and after function base index creation

    Hi All,
    Here is the unique situation we are facing after creating just an index.
    The query result before the index and the query result after the index do not match.
    This is very illogical situation. Shidhar and me have done lot of R&D, also tried to get lots info from Google but we couldn't decipher the reason for that.
    I am giving you all the details about the query, index and tables with following steps.
    Please let us know if anything is going wrong from our side or is it a bug at oracle level which is a rarest possibility but a possibility.
    Step 1 :- Create table
    create table TEMP_COMP
    ID VARCHAR2(10),
    GROUP_ID VARCHAR2(10),
    TRAN_DATE DATE,
    AMT_1 NUMBER,
    AMT_2 NUMBER,
    AMT_3 NUMBER
    Step 2 :- Insert Sample data
    set feedback off
    set define off
    prompt Deleting TEMP_COMP...
    delete from TEMP_COMP;
    commit;
    prompt Loading TEMP_COMP...
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('01', 'G01', to_date('01-03-2007', 'dd-mm-yyyy'), 1, 11, 111);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('02', 'G01', to_date('02-03-2007', 'dd-mm-yyyy'), 2, 22, 222);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('03', 'G01', to_date('03-03-2007', 'dd-mm-yyyy'), 3, 33, 333);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('04', 'G01', to_date('04-03-2007', 'dd-mm-yyyy'), 4, 44, 444);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('05', 'G01', to_date('05-03-2007', 'dd-mm-yyyy'), 5, 55, 555);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('06', 'G01', to_date('01-03-2008', 'dd-mm-yyyy'), 6, 66, 666);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('07', 'G01', to_date('02-03-2008', 'dd-mm-yyyy'), 7, 77, 777);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('08', 'G01', to_date('03-03-2008', 'dd-mm-yyyy'), 8, 88, 888);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('09', 'G01', to_date('04-03-2008', 'dd-mm-yyyy'), 9, 99, 999);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('10', 'G01', to_date('05-03-2008', 'dd-mm-yyyy'), 10, 100, 1000);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('01', 'G01', to_date('01-03-2007', 'dd-mm-yyyy'), 1, 11, 111);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('02', 'G01', to_date('02-03-2007', 'dd-mm-yyyy'), 2, 22, 222);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('03', 'G01', to_date('03-03-2007', 'dd-mm-yyyy'), 3, 33, 333);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('04', 'G01', to_date('04-03-2007', 'dd-mm-yyyy'), 4, 44, 444);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('05', 'G01', to_date('05-03-2007', 'dd-mm-yyyy'), 5, 55, 555);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('06', 'G01', to_date('01-03-2008', 'dd-mm-yyyy'), 6, 66, 666);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('07', 'G01', to_date('02-03-2008', 'dd-mm-yyyy'), 7, 77, 777);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('08', 'G01', to_date('03-03-2008', 'dd-mm-yyyy'), 8, 88, 888);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('09', 'G01', to_date('04-03-2008', 'dd-mm-yyyy'), 9, 99, 999);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('10', 'G01', to_date('05-03-2008', 'dd-mm-yyyy'), 10, 100, 1000);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('01', 'G02', to_date('01-03-2007', 'dd-mm-yyyy'), 1, 11, 111);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('02', 'G02', to_date('02-03-2007', 'dd-mm-yyyy'), 2, 22, 222);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('03', 'G02', to_date('03-03-2007', 'dd-mm-yyyy'), 3, 33, 333);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('04', 'G02', to_date('04-03-2007', 'dd-mm-yyyy'), 4, 44, 444);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('05', 'G02', to_date('05-03-2007', 'dd-mm-yyyy'), 5, 55, 555);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('06', 'G02', to_date('01-03-2008', 'dd-mm-yyyy'), 6, 66, 666);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('07', 'G02', to_date('02-03-2008', 'dd-mm-yyyy'), 7, 77, 777);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('08', 'G02', to_date('03-03-2008', 'dd-mm-yyyy'), 8, 88, 888);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('09', 'G02', to_date('04-03-2008', 'dd-mm-yyyy'), 9, 99, 999);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('10', 'G02', to_date('05-03-2008', 'dd-mm-yyyy'), 10, 100, 1000);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('01', 'G03', to_date('01-03-2007', 'dd-mm-yyyy'), 1, 11, 111);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('02', 'G03', to_date('02-03-2007', 'dd-mm-yyyy'), 2, 22, 222);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('03', 'G03', to_date('03-03-2007', 'dd-mm-yyyy'), 3, 33, 333);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('04', 'G03', to_date('04-03-2007', 'dd-mm-yyyy'), 4, 44, 444);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('05', 'G03', to_date('05-03-2007', 'dd-mm-yyyy'), 5, 55, 555);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('06', 'G03', to_date('01-03-2008', 'dd-mm-yyyy'), 6, 66, 666);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('07', 'G03', to_date('02-03-2008', 'dd-mm-yyyy'), 7, 77, 777);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('08', 'G03', to_date('03-03-2008', 'dd-mm-yyyy'), 8, 88, 888);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('09', 'G03', to_date('04-03-2008', 'dd-mm-yyyy'), 9, 99, 999);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('10', 'G03', to_date('05-03-2008', 'dd-mm-yyyy'), 10, 100, 1000);
    insert into TEMP_COMP (ID, GROUP_ID, TRAN_DATE, AMT_1, AMT_2, AMT_3)
    values ('11', 'G01', to_date('03-03-2008', 'dd-mm-yyyy'), 100, 200, 300);
    commit;
    prompt 41 records loaded
    set feedback on
    set define on
    prompt Done.
    Step 3 :- Execute the query.
    SELECT GROUP_ID
    , SUM(LAST_YR_REV) as "Year_2007_Amt"
    , SUM(CURR_YR_REV) as "Year_2008_Amt"
    FROM (
    SELECT GROUP_ID,
    CASE WHEN TO_CHAR(TRAN_DATE,'YYYYMM') BETWEEN'200701'AND'200712'THEN SUM(AMT_1) ELSE 0 END AS LAST_YR_REV ,
    CASE WHEN TO_CHAR(TRAN_DATE,'YYYYMM') BETWEEN'200801'AND'200812'THEN SUM(AMT_1) ELSE 0 END AS CURR_YR_REV
    FROM TEMP_COMP t
    WHERE GROUP_ID ='G01'
    AND TO_CHAR(TRAN_DATE,'YYYYMM') BETWEEN'200601'AND'200912'
    GROUP BY GROUP_ID, TRAN_DATE
    GROUP BY GROUP_ID
    The result of above query
    GROUP_ID Year_2007_Amt Year_2008_Amt
    G01 30 180
    Step 4 : Create composite index
    create index GROUP_ID_TRAN_DATE_IDX on TEMP_COMP (GROUP_ID, TO_CHAR(TRAN_DATE,'YYYYMM'))
    Step 5 : Execute once again query from step 3.
    SELECT GROUP_ID
    , SUM(LAST_YR_REV) as "Year_2007_Amt"
    , SUM(CURR_YR_REV) as "Year_2008_Amt"
    FROM (
    SELECT GROUP_ID,
    CASE WHEN TO_CHAR(TRAN_DATE,'YYYYMM') BETWEEN'200701'AND'200712'THEN SUM(AMT_1) ELSE 0 END AS LAST_YR_REV ,
    CASE WHEN TO_CHAR(TRAN_DATE,'YYYYMM') BETWEEN'200801'AND'200812'THEN SUM(AMT_1) ELSE 0 END AS CURR_YR_REV
    FROM TEMP_COMP t
    WHERE GROUP_ID ='G01'
    AND TO_CHAR(TRAN_DATE,'YYYYMM') BETWEEN'200601'AND'200912'
    GROUP BY GROUP_ID, TRAN_DATE
    GROUP BY GROUP_ID
    The result of above query
    GROUP_ID Year_2007_Amt Year_2008_Amt
    G01 0 210
    Thanks
    Sunil

    I just wanted to make a comment. The predicates in both your queries are flawed I believe. You convert a date column to a character and then you say only pick the converted result between two sets of characters.
    TO_CHAR(TRAN_DATE,'YYYYMM') BETWEEN'200601'AND'200912'It should be coded like this for a proper date range comparison:
    TRAN_DATE BETWEEN TO_DATE('200601','YYYYMM') AND TO_DATE('200912','YYYYMM')That will eliminate the need for you to create FBI that you created. You should now be able to create a regular (B*Tree) index on the TRAN_DATE column.
    Also, now that we have the structure of your tables and sample data, what business question are you trying to answer? I think there is a better query that can be written if we know the requirements.
    Hope this helps!

  • MacBook Pro and Time Capsule after adding new wireless router

    I have a MacBook Pro and Time Capsule (which functions wirelessly to router and computer). I added a new wireless router (netgear, not my idea) and had to change name of wireless network. Now Time Capsule flashes amber and does not show up in Airport Utility.
    Do I have to set up Time Capsule from scratch? If so, if you can aim me to directions I won't have to waste tech support's time.
    My additional question on a similar vein: I also have a Time Capsule (second time capsule) which is plugged into the wireless router using ethernet cable for use with a windows notebook running windows 7 which backs up wirelessly. Do I have to change anything on this Time Capsule (second Time Capsule) as a result of change in name of network (I've already adjusted computer to connect to network)?

    Hi DD,
    If you don't get answers here, try posting in the Time Capsule forum (http://discussions.apple.com/forum.jspa?forumID=1253) and/or the Time Machine forum (http://discussions.apple.com/forum.jspa?forumID=1342).

  • TS4009 before upgrading to Lion, I switched from MobileMe to iCloud.  At that time I used my MobileMe user name and password.  I purchased additional storage space to 20GB.  I was unabl to send any emails during this time.  After researching, I upgraded t

    before upgrading to Lion, I switched from MobileMe to iCloud.  At that time I used my MobileMe user name and password.  I purchased additional storage space to 20GB.  I was unable to send any emails during this time.  After researching, I upgraded to Lion and changed my icloud mail usuer name and password from the original I used from mobileme.  How do I get my upgraded storage to my new account.  I intend to delete the first one.
    Please help!

    Welcome to the Apple Community.
    Unfortunately you can't, the additional storage belongs to your first account, you can't transfer it.

  • Asset query execution performance after upgrade from 4.6C to ECC 6.0+EHP4

    Hi,guys
    I am encounted a weird problems about asset query execution performance after upgrade to ECC 6.0.
    Our client had migrated sap system from 4.6c to ECC 6.0. We test all transaction code and related stand report and query.
    Everything is working normally except this asset depreciation query report. It is created based on ANLP, ANLZ, ANLA, ANLB, ANLC table; there is also some ABAP code for additional field.
    This report execution costed about 6 minutes in 4.6C system; however it will take 25 minutes in ECC 6.0 with same selection parameter.
    At first, I am trying to find some difference in table index ,structure between 4.6c and ECC 6.0,but there is no difference about it.
    i am wondering why the other query reports is running normally but only this report running with too long time execution dump messages even though we do not make any changes for it.
    your reply is very appreciated
    Regards
    Brian

    Thanks for your replies.
    I check these notes, unfortunately it is different our situation.
    Our situation is all standard asset report and query (sq01) is running normally except this query report.
    I executed se30 for this query (SQ01) at both 4.6C and ECC 6.0.
    I find there is some difference in select sequence logic even though same query without any changes.
    I list there for your reference.
    4.6C
    AQA0FI==========S2============
    Open Cursor ANLP                                    38,702  39,329,356  = 39,329,356      34.6     AQA0FI==========S2============   DB     Opens
    Fetch ANLP                                         292,177  30,378,351  = 30,378,351      26.7    26.7  AQA0FI==========S2============   DB     OpenS
    Select Single ANLC                                  15,012  19,965,172  = 19,965,172      17.5    17.5  AQA0FI==========S2============   DB     OpenS
    Select Single ANLA                                  13,721  11,754,305  = 11,754,305      10.3    10.3  AQA0FI==========S2============   DB     OpenS
    Select Single ANLZ                                   3,753   3,259,308  =  3,259,308       2.9     2.9  AQA0FI==========S2============   DB     OpenS
    Select Single ANLB                                   3,753   3,069,119  =  3,069,119       2.7     2.7  AQA0FI==========S2============   DB     OpenS
    ECC 6.0
    Perform FUNKTION_AUSFUEHREN     2     358,620,931          355
    Perform COMMAND_QSUB     1     358,620,062          68
    Call Func. RSAQ_SUBMIT_QUERY_REPORT     1     358,569,656          88
    Program AQIWFI==========S2============     2     358,558,488          1,350
    Select Single ANLA     160,306     75,576,052     =     75,576,052
    Open Cursor ANLP     71,136     42,096,314     =     42,096,314
    Select Single ANLC     71,134     38,799,393     =     38,799,393
    Select Single ANLB     61,888     26,007,721     =     26,007,721
    Select Single ANLZ     61,888     24,072,111     =     24,072,111
    Fetch ANLP     234,524     13,510,646     =     13,510,646
    Close Cursor ANLP     71,136     2,017,654     =     2,017,654
    We can see first open cursor ANLP ,fetch ANLP then select ANLC,ANLA,ANLZ,ANLB at 4.C.
    But it changed to first select ANLA,and open cursor ANLP,then select  ANLC,ANLB,ANLZ,at last fetch ANLP.
    Probably,it is the real reason why it is running long time in ECC 6.0.
    Is there any changes for query selcection logic(table join function) in ECC 6.0.

  • How to track querying time in ASO

    Hi
    I am working on an ASO cube and have lot of member formulas in the outline.
    The formulas are really big and could be modified with some tweeks.
    Now i want to check, whether on changing formulas, my querying time is increasing or decreasing. Is there any way or log file to track, how much time did it take when i query for set of members from my spreadsheet.
    I want to compare how much seconds/min did a querying event took.

    Dude,
    I believe the best practice here is going to be to Enable Query Tracking in the Essbase Database (Right-Click database > Query Tracking > Enable).
    At this point we typically write several report scripts that simulate queries against HFR reports, etc. that the business would be pulling on a daily basis (or other frequency). You can even schedule the report scripts so that they execute at some intense frequency to check how the server handles concurrent requests, etc.
    The query tracking output should then provide you with most of the infromation you seek.
    Be sure to turn off query tracking for that database after you've completed your testing/exercise as it add additional processing to the database which will slow down a production server.
    Some References:
    http://download.oracle.com/docs/cd/E17236_01/epm.1112/eas_help/frameset.htm?qrytrack.html
    http://download.oracle.com/docs/cd/E17236_01/epm.1112/eas_help/frameset.htm?dbwzagg_3.html
    If this was helpful or the correct answer please award points.
    Cheer,
    Christian
    http://www.artofbi.com

  • Time Machine breaks at "Indexing Backup"

    Hello!
    After I upgrade to 10.6. Snow Leopard I can not backup my TM.I have try many times,but every time TM breaks by "Indexing Backup" and don't go forward.What shoud I do??(Sorry about my English)

    Pondini wrote:
    When you double-click the TC/sparse bundle, does it appear on your Desktop? Drag it from there. (I have an Airport Extreme, not a TC, and it works a bit differently: when my iMac is connected to the AEBS, the sparsebundle appears in the DU sidebar automaticaly. The instructions came from a TC user).
    I guess I'm a bit confused on what the sparsebundle is as opposed to a partition as opposed to the disk image as opposed to the...
    Feel free to (and please do) correct any terminology that I am using incorrectly.
    The disk image (assuming that is the icon that shows up on the desktop when accessing the TC) of the sparsebundle (assuming that is what is labeled "Backup of David Nash's MacBook Pro") does automatically appear in my disk utility sidebar when the TC disk image is mounted on the desktop. I am guessing from your last post that the "dragging the sparsebundle" step is unnecessary in light of this fact.
    If the format matching is as big of a deal as the "User Tips" author makes it sound...
    That would be me.
    That's funny.
    So format matching doesn't seem to matter for TC? To be honest, It wouldn't be the end of the world if I had to completely erase my TC and start over. I am tempted to just do that and start over fresh if that would solve the problem. I'll hold off on that for now...
    I assume you've tried removing it from the Spotlight > Privacy box, then re-adding it? If you can't, it's a bit of a conundrum: when you dismount the TC/sparsebundle, it will disappear from the Privacy display, but come back when it's remounted.
    I originally just added the whole time capsule disk image to the privacy list. I later noticed that the "Backup of David Nash's MacBook Pro" disk image was automatically added to the privacy list when the spotlight is trying to index, and the "Backup..." disk image is removed from the privacy list when the spotlight is not trying to index. I have since tried to remove and re-add the sparsebundle to the list. It was in the process of trying to index, and re-adding it to the list didn't stop the indexing. I don't know if I was supposed to stop the indexing by restarting the computer or what, but as far as I can tell, being on the privacy list doesn't seem to have any effect on spotlight indexing.
    If you can't do the +*Repair Disk,+* and nothing else works, you may have to delete the sparsebundle and let TM start fresh.
    Like I mentioned above, this would not be the end of the world to delete the sparsebundle (or the whole TC, for that matter). I haven't yet tried to do the Repair Disk, because I don't know if it is something that I have to do just right or I'll mess something up. Does it matter if spotlight is trying to index? Is there any other considerations I should take? As you can probably tell, I don't really know what I'm doing, so I appreciate your help.
    One last question about terminology - you said that it is not possible to "partition" the TC's drive; I was under the impression that the separate sparsebundles for the different computers (mine and my wife's) were "partitions". So this is incorrect terminology?

  • After building index is crawler deletes the data from crawl db?

    HI
    how the search service crawl component works,
    each time after building index is crawler deletes the data from crawl db
    adil

    Hi Adil,
    Index Database is usually called Crawl Database.
    I found a series of articles which goes through what these components are, how they work together, and how to provision them:
    http://blogs.msdn.com/b/russmax/archive/2010/04/16/search-2010-architecture-and-scale-part-1-crawl.aspx
    http://blogs.msdn.com/b/russmax/archive/2010/04/23/search-2010-architecture-and-scale-part-2-query.aspx
    SharePoint Server 2010 Enterprise Search Evaluation Guide:
    https://www.microsoft.com/en-us/download/details.aspx?id=24189
    Regards,
    Rebecca Tu
    TechNet Community Support

  • Bi Dashboards issue works properly some times and after not display pointers data and give un expected error and throw login window

    HI,
    I have toff situation where  we deployed bi dashboards in our site and it works some time and some times not,
    and some times it keep loading and throw a login window.
    works when
    we restart performance point service  and if not work
    we re create secure store service and generated new key and create new pps service application
    before that we stopped secure store and pps service  using c.a in application server
    we are using ssas for datasource to connect  from dashboard designer
    we have seperate server for ssas dbs
    seperate application server running: pps service and secure store service
    two web front ends:  one of running secure store
    1 Domain controller + central admin service running
    some times after 4 -5 hours bi dashboards works , we dashboards not display data and throw un expected errors,
    also when we try to connect using a dashboard designer from a seperate client machine in same domain ,we have same issue , we unable to connect to ssas datasource server, and if connect some times unable to deploy dashboards.
    we tried every thin icreased server time out values  in ssas server
    :\Program Files\Microsoft SQL Server\MSAS10_50.MSSQLSERVER\OLAP\Config\msmdsrv.ini. in this location
    and in 
    c:\program files\common files\microsoft shared\web server extensions\14\webclients\ppsmonitoringServer\client.config
    maxStringContentLength="2147483647"
    maxNameTableCharCount="2147483647"
    maxBytesPerRead="2147483647"
    maxArrayLength="2147483647"
    maxDepth="2147483647" />
    even we have same issue
    is some thing happening when dashboards working some time and  after some time not 
    is a secure store crashing or its still request pending in IIS and not authenticating to data soruce why 
    below are the logs
    Here are the results of the log analysis :
    server wfe1 
    09/01/2014 14:43:44.66 w3wp.exe (0x2128) 0x158C PerformancePoint Service PerformancePoint Services 39 Critical A PerformancePoint service application call was aborted by the caller.  This may indicate the HttpRuntime executionTimeout for the Web Application
    is configured to a value smaller than the DataSourceQueryTimeout for the PerformancePoint Service Application. 8bc44f14-acef-49bf-b1c1-2755ea7e6e24
    09/01/2014 14:43:44.66 w3wp.exe (0x2128) 0x158C PerformancePoint Service PerformancePoint Services ef8z Critical ‏‏حدث
    استثناء
    أثناء عرض
    عنصر ويب.
    قد تساعد
    معلومات التشخيص
    التالية في
    تحديد سبب
    هذه المشكلة:  Microsoft.PerformancePoint.Scorecards.BpmException: ‏‏لقد
    انتهت
    مهلة الطلب
    أو تم
    إجهاضها. تم
    تسجيل تفاصيل
    إضافية من
    أجل المسؤول.  ‏‏رمز
    خطأ "خدمات PerformancePoint"
    هو 20700. 8bc44f14-acef-49bf-b1c1-2755ea7e6e24
    09/01/2014 14:43:44.66 w3wp.exe (0x2128) 0x158C SharePoint Foundation Runtime ba3q Medium Redirect to error.aspx?ErrorText=Request%20timed%20out%2E failed. Exception: System.Web.HttpException: The remote host closed the connection. The error code is 0x800704CD.    
    at System.Web.Hosting.IIS7WorkerRequest.RaiseCommunicationError(Int32 result, Boolean throwOnDisconnect)     at System.Web.Hosting.IIS7WorkerRequest.ExplicitFlush()     at System.Web.HttpResponse.Flush(Boolean finalFlush)    
    at System.Web.HttpResponse.End()     at Microsoft.SharePoint.Utilities.SPUtility.Redirect(String url, SPRedirectFlags flags, HttpContext context, String queryString) 8bc44f14-acef-49bf-b1c1-2755ea7e6e24
    09/01/2014 14:47:32.69 w3wp.exe (0x2128) 0x2154 Web Content Management Publishing Cache 7363 Critical Object Cache: The super reader account utilized by the cache does not have sufficient permissions to SharePoint databases. To configure the account use
    the following command 'stsadm -o setproperty -propertyname portalsuperreaderaccount -propertyvalue account -url webappurl'. It should be configured to be an account that has Read access to the SharePoint databases.  Additional Data:  Current default
    super reader account: NT AUTHORITY\LOCAL SERVICE 78c107cd-0d4b-46f5-8e1a-0d9b4ebc7b29
    09/01/2014 15:07:40.74 w3wp.exe (0x2128) 0x209C PerformancePoint Service PerformancePoint Services 20 Critical The user "i:Anonymous" attempted to access an item in the following location:
    http://www.xxx.xxx.sa/ar-sa/Notary/Rspointer/Lists/PerformancePoint Content/28_.000  Verify that the location exists and that the user has the "Read Items" permission. 
    Exception details: Microsoft.PerformancePoint.Scorecards.BpmException: ‏‏لقد
    انتهت
    مهلة الطلب
    أو تم
    إجهاضها. تم
    تسجيل تفاصيل
    إضافية من
    أجل المسؤول.     at Microsoft.PerformancePoint.Scorecards.Store.Dao.SPListItemDao.Get(ISPFcoCache cache, Guid webSiteID, RepositoryLocation location, SPListItem item)    
    at Microsoft.PerformancePoint.Scorecards.Store.Dao.SPListItemDao.Get(ISPFcoCache cache, Guid webSiteID, RepositoryLocation location)     at Microsoft.PerformancePoint.Scorecards.Store.SPDataStore.SPItemGet(RepositoryLocation location, String
    spObjectTypeId) abf0263d-7333-4e4f-9cd4-a3a4dcb700c0
    09/01/2014 15:13:25.92 w3wp.exe (0x2128) 0x18DC PerformancePoint Service PerformancePoint Services 20 Critical The user "i:Anonymous" attempted to access an item in the following location:
    http://www.xxx.xxx.sa/ar-sa/Notary/Rspointer/Lists/PerformancePoint Content/218_.000  Verify that the location exists and that the user has the "Read Items" permission. 
    Exception details: Microsoft.PerformancePoint.Scorecards.BpmException: ‏‏لقد
    انتهت
    مهلة الطلب
    أو تم
    إجهاضها. تم
    تسجيل تفاصيل
    إضافية من
    أجل المسؤول.     at Microsoft.PerformancePoint.Scorecards.Store.Dao.SPListItemDao.Get(ISPFcoCache cache, Guid webSiteID, RepositoryLocation location, SPListItem item)    
    at Microsoft.PerformancePoint.Scorecards.Store.Dao.SPListItemDao.Get(ISPFcoCache cache, Guid webSiteID, RepositoryLocation location)     at Microsoft.PerformancePoint.Scorecards.Store.SPDataStore.SPItemGet(RepositoryLocation location, String
    spObjectTypeId) b35bf864-54f4-43fa-88a8-44cdf938bfa2
    09/01/2014 15:11:10.87 w3wp.exe (0x2128) 0x1F88 PerformancePoint Service PerformancePoint Services 20 Critical The user "i:Anonymous" attempted to access an item in the following location:
    http://www.xxx.xxx.sa/ar-sa/Courts/Bic/Lists/PerformancePoint Content/4_.000  Verify that the location exists and that the user has the "Read Items" permission. 
    Exception details: Microsoft.PerformancePoint.Scorecards.BpmException: ‏‏لقد
    انتهت
    مهلة الطلب
    أو تم
    إجهاضها. تم
    تسجيل تفاصيل
    إضافية من
    أجل المسؤول.     at Microsoft.PerformancePoint.Scorecards.Store.Dao.SPListItemDao.Get(ISPFcoCache cache, Guid webSiteID, RepositoryLocation location, SPListItem item)    
    at Microsoft.PerformancePoint.Scorecards.Store.Dao.SPListItemDao.Get(ISPFcoCache cache, Guid webSiteID, RepositoryLocation location)     at Microsoft.PerformancePoint.Scorecards.Store.SPDataStore.SPItemGet(RepositoryLocation location, String
    spObjectTypeId) 7cea0eb6-214f-4d7b-a6e7-97a0fe11c956
    09/01/2014 15:11:40.87 w3wp.exe (0x2128) 0x2150 PerformancePoint Service PerformancePoint Services 20 Critical The user "i:Anonymous" attempted to access an item in the following location:
    http://www.xxx.xxx.sa/ar-sa/Notary/Rspointer/Lists/PerformancePoint Content/162_.000  Verify that the location exists and that the user has the "Read Items" permission. 
    Exception details: Microsoft.PerformancePoint.Scorecards.BpmException: ‏‏لقد
    انتهت
    مهلة الطلب
    أو تم
    إجهاضها. تم
    تسجيل تفاصيل
    إضافية من
    أجل المسؤول.     at Microsoft.PerformancePoint.Scorecards.Store.Dao.SPListItemDao.Get(ISPFcoCache cache, Guid webSiteID, RepositoryLocation location, SPListItem item)    
    at Microsoft.PerformancePoint.Scorecards.Store.Dao.SPListItemDao.Get(ISPFcoCache cache, Guid webSiteID, RepositoryLocation location)     at Microsoft.PerformancePoint.Scorecards.Store.SPDataStore.SPItemGet(RepositoryLocation location, String
    spObjectTypeId) 4154222b-6ff0-4b64-81b8-d08501a19278
    server wfe 2
    09/01/2014 14:50:59.05 w3wp.exe (0x1294) 0x1D88 Web Content Management Publishing Cache 7363 Critical Object Cache: The super reader account utilized by the cache does not have sufficient permissions to SharePoint databases. To configure the account use
    the following command 'stsadm -o setproperty -propertyname portalsuperreaderaccount -propertyvalue account -url webappurl'. It should be configured to be an account that has Read access to the SharePoint databases.  Additional Data:  Current default
    super reader account: NT AUTHORITY\LOCAL SERVICE 382f711d-cdad-4c5d-be88-2e7d6f36dde2
    09/01/2014 14:50:59.05 w3wp.exe (0x1294) 0x1D88 Web Content Management Publishing Cache 7363 Critical Object Cache: The super reader account utilized by the cache does not have sufficient permissions to SharePoint databases. To configure the account use
    the following command 'stsadm -o setproperty -propertyname portalsuperreaderaccount -propertyvalue account -url webappurl'. It should be configured to be an account that has Read access to the SharePoint databases.  Additional Data:  Current default
    super reader account: NT AUTHORITY\LOCAL SERVICE 382f711d-cdad-4c5d-be88-2e7d6f36dde2
    09/01/2014 14:50:59.05 w3wp.exe (0x1294) 0x1D88 Web Content Management Publishing Cache 7363 Critical Object Cache: The super reader account utilized by the cache does not have sufficient permissions to SharePoint databases. To configure the account use
    the following command 'stsadm -o setproperty -propertyname portalsuperreaderaccount -propertyvalue account -url webappurl'. It should be configured to be an account that has Read access to the SharePoint databases.  Additional Data:  Current default
    super reader account: NT AUTHORITY\LOCAL SERVICE 382f711d-cdad-4c5d-be88-2e7d6f36dde2
    09/01/2014 14:50:59.05 w3wp.exe (0x1294) 0x1D88 Web Content Management Publishing Cache 7363 Critical Object Cache: The super reader account utilized by the cache does not have sufficient permissions to SharePoint databases. To configure the account use
    the following command 'stsadm -o setproperty -propertyname portalsuperreaderaccount -propertyvalue account -url webappurl'. It should be configured to be an account that has Read access to the SharePoint databases.  Additional Data:  Current default
    super reader account: NT AUTHORITY\LOCAL SERVICE 382f711d-cdad-4c5d-be88-2e7d6f36dde2
    In the Roiscan logs we get an error in regards to the English Language pack:
    Review Items
    ============
    Error:                       Product {90140000-1015-0409-1000-0000000FF1CE} - Microsoft SharePoint Foundation 2010 1033 Lang Pack:  has unexpected
    file state(s).
    adil

    chinnijagadeesh,
    You suck!
    Signed: The rest of the universe.
    ... and quit crossposting FFS... it really is quite annoying!

  • Query not considering function based index in oracle 11g

    I have a query which used Function Based Index when run in oracle 9i but when I run the same query
    without any changes, it does not consider index. Below is the query:
    SELECT distinct patient_role.domain_key, patient_role.patient_role_key,
    patient_role.emergency_contact_name,
    patient_role.emergency_contact_phone, patient_role.emergency_contact_note,
    patient_role.emergency_contact_relation_id,
    patient_role.financial_class_desc_id, no_known_allergies, patient_role.CREATED_BY,
    patient_role.CREATED_TIMESTAMP,
    patient_role.CREATED_TIMESTAMP_TZ, patient_role.UPDATED_BY, patient_role.UPDATED_TIMESTAMP,
    patient_role.UPDATED_TIMESTAMP_TZ,
    patient_role.discontinued_date
    FROM encounter, patient_role
    WHERE patient_role.patient_role_key = encounter.patient_role_key
    AND UPPER(TRIM(leading :SYS_B_0 from encounter.account_number)) = UPPER(TRIM(leading :SYS_B_1 from
    :SYS_B_2))
    AND patient_role.discontinued_date IS null
    AND encounter.discontinued_date IS null ;
    Index definition:
    CREATE INDEX "user1"."IX_TRIM_ACCOUNT_NUMBER" ON "user1."ENCOUNTER" (UPPER(TRIM(LEADING
    '0' FROM "ACCOUNT_NUMBER")), "PATIENT_ROLE_KEY", "DOMAIN_KEY", "DISCONTINUED_DATE")
    PCTFREE 10 INITRANS 2 MAXTRANS 255 COMPUTE STATISTICS
    STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
    PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1
    BUFFER_POOL DEFAULT)
    TABLESPACE "user1"
    Database : Oracle 11g (11.2.0.3)
    O/S : Linux 64 bit (the query does not consider index even on windows os)
    Any suggestions?
    -Onkar
    Edited by: onkar.nath on Jul 2, 2012 3:32 PM

    Onkar,
    I don't appreciate you posting this question in several forums at the same time.
    If I would know you also posted this on Asktom, I wouldn't even have bothered.
    As to your 'issue':
    First of all: somehow cursor_sharing MUST have been set to FORCE. Oracle is a predictable system, not a fruitmachine.
    Your statement the '0' is replaced by a bind variable anyway is simply false. If you really believe it is not false, SUBMIT a SR.
    But your real issue is not Oracle: it is your 'application', which is a mess anyway. Allowing for alphanumeric numbers is a really bad idea.
    Right now you are already putting workaround on workaround on workaround on workaround.
    Issue is the application: it is terminal., and you either need to kill it, or to replace it.
    Sybrand Bakker
    Senior Oracle DBA

  • How do I change the length of multiple photo's after I've added them?

    How do I change the length of multiple photo's after I've added them?

    Hey,
    The best way to do this is to select all the clips by holding down the shift key or command key if the clips you want to change are not all right next to each other and then selecting each clip that you want to change. Then double-click on the clips which will open up the inspector tab and then you can change the duration. If you want to just change all clips to the same duration just double-click on one of the photos and in the inspector tab next to the durationt tab you will see a box that says "add duration to all clips". If you click that, when you change the duration all the photos will change to that duration. That will then change the duration of all the clips at the same time.

  • Query Transformation in Function Based Index's Expression

    Hi All,
    I am having a wired issue with Function Based Indexes on Orcale 10.2.0.3 running on Solaris 5.9.
    I have created two FBI on two tables using following syntax.
    CREATE INDEX EQHUB.IDX_BBO_WRT_DES_FBI ON EQHUB.BBO_WARRANT_PRICING (CASE WHEN latest_version_flag = 'Y' THEN 1 ELSE NULL END);
    CREATE INDEX EQHUB.IDX_BBO_DES_FBI ON EQHUB.BBO_DESCRIPTIVE_DATA (CASE WHEN latest_version_flag = 'Y' THEN 1 ELSE NULL END);
    For the second command (IDX_BBO_DES_FBI), when i query DBA_IND_EXPRESSIONS view, i found that Oracle has done some kind of QUERY TRANSFORMATION (?) and converted
    FBI expression to CASE "LATEST_VERSION_FLAG" WHEN 'Y' THEN 1 ELSE NULL END.At the same time,EXPRESSION on first index is not changed.
    Now,my question is what has made transformation to occure only for second index.
    I also found that inspite of highly SELECTIVE nature of both the indexes, only SECOND index is being used by CBO (for which trasnformation occured)
    and IDX_BBO_WRT_DES_FBI is not being used(FTS is happening instead).
    Query is using same expression for both the tables as
    (CASE WHEN latest_version_flag = 'Y' THEN 1 ELSE NULL END)=1
    INDEX_NAME                          TABLE_NAME                          COLUMN_EXPRESSION
    IDX_BBO_WRT_DES_FBI                 BBO_WARRANT_PRICING                 CASE  WHEN "LATEST_VERSION_FLAG"='Y' THEN 1 ELSE NULL END
    IDX_BBO_DES_FBI                     BBO_DESCRIPTIVE_DATA                CASE "LATEST_VERSION_FLAG" WHEN 'Y' THEN 1 ELSE NULL ENDI read that expression should be evaluated including CASE of characters and spaces in query.Is that true?
    Appreciating responses in advance.

    Randolf.
    It's a shame that I forgot to look into the full execution plan information to check how Oracle really handles my queries.
    Look here(edited for clarity):
    explain plan for
    select /*+ case1 ordered use_nl(x, y) */ count(case c1
        when '1' then 1
        when '2' then 2
        when '3' then 3
        else 4
      end) from
    (select level from dual connect by level <= 300000) x,
    (select
    from t1
    ) y;
    | Id  | Operation                       | Name | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT                |      |     1 |     2 |     5   (0)| 00:00:01 |
    |   1 |  SORT AGGREGATE                 |      |     1 |     2 |            |          |
    |   2 |   NESTED LOOPS                  |      |     3 |     6 |     5   (0)| 00:00:01 |
    |   3 |    VIEW                         |      |     1 |       |     2   (0)| 00:00:01 |
    |   4 |     CONNECT BY WITHOUT FILTERING|      |       |       |            |          |
    |   5 |      FAST DUAL                  |      |     1 |       |     2   (0)| 00:00:01 |
    |   6 |    TABLE ACCESS FULL            | T1   |     3 |     6 |     3   (0)| 00:00:01 |
    Column Projection Information (identified by operation id):
       1 - (#keys=0) COUNT(CASE  WHEN "T1"."C1"='1' THEN 1 WHEN "T1"."C1"='2' THEN
           2 WHEN "T1"."C1"='3' THEN 3 ELSE 4 END )[22]
       2 - (#keys=0) "T1"."C1"[CHARACTER,1]
       4 - "DUAL".ROWID[ROWID,10], LEVEL[4]
       5 - "DUAL".ROWID[ROWID,10]
       6 - "T1"."C1"[CHARACTER,1]
    32 rows selected.
    explain plan for select /*+ case2 ordered use_nl(x, y) */ count(case
        when c1 = '1' then 1
        when c1 = '2' then 2
        when c1 = '3' then 3
        else 4
      end) from
    (select level from dual connect by level <= 300000) x,
    (select
    from t1
    ) y;
    | Id  | Operation                       | Name | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT                |      |     1 |     2 |     5   (0)| 00:00:01 |
    |   1 |  SORT AGGREGATE                 |      |     1 |     2 |            |          |
    |   2 |   NESTED LOOPS                  |      |     3 |     6 |     5   (0)| 00:00:01 |
    |   3 |    VIEW                         |      |     1 |       |     2   (0)| 00:00:01 |
    |   4 |     CONNECT BY WITHOUT FILTERING|      |       |       |            |          |
    |   5 |      FAST DUAL                  |      |     1 |       |     2   (0)| 00:00:01 |
    |   6 |    TABLE ACCESS FULL            | T1   |     3 |     6 |     3   (0)| 00:00:01 |
    Column Projection Information (identified by operation id):
       1 - (#keys=0) COUNT(CASE  WHEN "T1"."C1"='1' THEN 1 WHEN "T1"."C1"='2' THEN
           2 WHEN "T1"."C1"='3' THEN 3 ELSE 4 END )[22]
       2 - (#keys=0) "T1"."C1"[CHARACTER,1]
       4 - "DUAL".ROWID[ROWID,10], LEVEL[4]
       5 - "DUAL".ROWID[ROWID,10]
       6 - "T1"."C1"[CHARACTER,1]As you exactly mentioned, I was executing different SQL but actually same queries!
    Thanks for pointing my flaws.
    PS) OP, forgive me for bothering you with off-topic things. :)
    ================================
    Dion Cho - Oracle Performance Storyteller
    http://dioncho.wordpress.com (english)
    http://ukja.tistory.com (korean)
    ================================
    Edited by: Dion_Cho on Feb 10, 2009 5:45 AM
    Typo

Maybe you are looking for

  • COGS workflow-Using Different Cost of Goods Sold Account

    hi all, Could some one interpret this and tell me exactly what it is. This is the customization of cogs workflow. Ive to use different COGS account when the fares are sold on promotion.and COGS account should be different between regular fares and pr

  • Selection screen versions in LdB

    The following questions are related to SAP delivered LdBs that are not for HR. a) Do you know where the selection screen versions for a delivered LdB are stored? b) Can additional selection versions be created /delivered selection screen versions mod

  • Itunes has experienced an error ....

    hello, so i've had this problem before of trying to upgrade itunes and then the dumb thing doesnt work. last time it took at least 2 weeks to figure it out i hope this one is shorter. i downloaded 6.0 and the quicktime works perfectly but itunes does

  • No Update of iMovie 10.0 with Mavericks on iMac from 2007

    After upgrading my iMac (24") from 2007 with Maverick it is not possible to update the new iMovie because the Graphic-Board seems to be too slow or too less of RAM. Question: Is it possible to upgrade the Graphic-Board?

  • Error: unexpected symbol; expected ' ', ' /', entity reference

    I am getting below error in Sxmb_moni. Any idea whwn does it comes. Looks like system error. My senario is file to file . I am using  java mapping and even doing RFC look up in the  senario: <SAP:Error xmlns:SAP="http://sap.com/xi/XI/Message/30" xmln