Which of these takes less time?

Which of these takes less time?
INSERT INTO TABLE_1 AS SELECT * FROM TABLE_2;
CREATE TABLE TABLE_1 SELECT * FROM TABLE_2;
Thanks,
Bhagat

here's what the documentation says about a table created with the NOLOGGING clause:
"Specify NOLOGGING if you do not want these operations to be logged.
For a nonpartitioned object, the value specified for this clause is the actual physical attribute of the segment associated with the object."
I read that to mean that segment information is not logged. I don't see where it says that inserts are not logged.
And my testing on versions 9, 10.1 and 10.2 have shown that normal inserts to a table created with nologging mode will generate the same amount of redo as inserts to a table created with logging mode.
Using APPEND will drastically reduce the amount of redo generated. Here's some of my findings showing the amount of redo (in bytes) generated for different methods:
v10.2.0.3.0
table structure:
t( a number, b number)
insert 50,000 rows per test
created regular table, did NOT enable parallel dml
insert append parallel :  141808 redo
insert parallel        :  142032 redo
insert append          :    1156 redo
insert                 :  754120 redo
regular table, alter session enable parallel dml:
insert append parallel :  141180
insert parallel        :  141876
insert append          :   16156
insert                 :  748552
create table with nologging option, did NOT enable parallel dml:
insert append          :   16000
insert                 :  748616let the debate begin (again)

Similar Messages

  • I bought a usb power adapter for my ipod nano so I can plug inti a power outlet. Will this take less time to charge my ipod or same amout of time if using the computer?

             I bought a USB power adapter for my ipod nano so i can take it camping and use an outlet. Will this take less time for charging or the same amount as the computer?

    Hi, Atcandela.
    I've a Macally external fire/usb enclosure that powers up and mounts directly on to my PB Aluminum 1.67 laptop but the USB portion doesn't draw power.
    Do you mean it mounts and works OK when connected via FireWire, but not when connected via USB? If so, that's unsurprising. I would be surprised to hear that it doesn't mount when you use the FireWire connnection, because most drives are able to get enough power from that port.
    Does MacAlly not offer a cable for your drive that allows it to draw power from both your USB ports? Some manufacturers of external drives or enclosures do offer such a cable, and it's well worth buying because it's much smaller and more portable than an AC adapter. Have a look at this one from Western Digital — it might fill the bill for you.

  • I have the newest imovie as i got a macbook for christmas, i have made a video and want to upload it to youtube and it is a quicktime file, but yet it is still taking ages to upload and youtube says if i change the format it will take less time,help?

    I need help for how to upload a video on youtube when it is a quicktime file but how i change it to a streamable format in order for it to not take as long to upload, how do i do this?

    I have the same issue.  If you are exporting from iMovie version 10.0.1, I run into the same issues if I export directly to YouTube or if I export to file then upload to YouTube.
    Google's help file doesn't properly address this problem and thier help is also out of date.  In talking to a support rep at Apple, they said the export to YouTube should just work or the exporting to a file then uploading otherwise it's an issue for Google support.
    If someone can post a solution that works (is a streamable format - which is different than the format that YouTube accepts as it will accept many formats but take a long time to process them), I'd be interested to hear also.
    Thanks,
    Rob.

  • How do I make a VI that contols a 6534 card take less time to load 1Mbyte and output data ?

    Hello-
    I'm developing a pattern generation burst mode type VI that reads byte characters from a text file, manipulates the bytes that it reads and then outputs the data from a 6534 card. The REQ1 pin is pulled high which makes the card continually output data until the internal buffer is empty. An internal clock is used. The problem is that the size of the file that is read by the VI can be relatively large, in the 1MByte range. After the VI manipulation, the output array is 16 times larger than the input string. So, when the VI reads a 1MByte text file, the output array contains 16 million entries (each array entry is 3 bits by the way). It takes about 20 seco
    nds to run and output data from the 6534 even when reading files that are 200kbyte or so. I'm outputting the data at 5Mbps. I'm not sure where all the delay is coming from. Using the 'Profile VI" function under the 'Tools' menu shows a snapshot of time that really doesn't help me since it only shows that the time is spent in my VI. Is this just the result of loading large files ? Or, is there a way to preload the 1MByte file into the 32MB of RAM on the 6534 card before the VI is actually run. What I mean is that I do have the option in my system to have a time when all text files will somehow be loaded into the 6534 memory, during some initialization period, and then the VI would be run with the text file already loaded. This of course assumes that the delay is loading the file from the harddrive, but since the harddrive transfer speed is fast, I have no idea where to speed things up. Any ideas ? Thanks !

    Your question actually belongs to the Measurement Devices >> Digital IO section since it concerns operating the 653x family of digital boards. You will see other related questions and answers there. In the future, please direct your digital questions to that forum.
    If you are trying to profile your VI, you can break it into different pieces. For instance, you can make a subVI out of the file data manipulation.
    Now, are you doing pattern generation or burst mode handshaking output? From your description of the signal connections, it looks like burst mode. You can double-check in the 653x User Manual at the http://www.ni.com/manuals pages. Still, it looks like you could accomplish your task with pattern generation. There are example VIs for both operations that
    I suggest you start with. They are located in the LabVIEW >> Examples >> Daq >> Digital >> 653x.llb directory.
    It also sounds like you are doing a single buffered operation, where you have a set amount of data that you want to output, and once done, the VI will stop. You are preloading the data to the buffer at DIO Write.vi before the 6534 actually outputs the data with DIO Start.vi when you program like the examples.
    Thus, start with the appropriate example and then add the File I/O to it. You can even test with a smaller file to see the effects. You will also find useful information concerning the 6534 in the KnowledgeBase, particularly KnowledgeBase 298CGF9M "How Many Samples of Data Can I Store in the Onboard Memory of the NI 6534?". You can find this online by going to the http://www.ni.com/support and choosing KnowledgeBase from the Technical Resources pull-down menu. Then, type in keywords to search for, such as "6534".
    Regards,
    Geneva L.
    Applications Engineering
    Nati
    onal Instruments
    http://www.ni.com/ask

  • Bind variable code takes more time to complete?

    Hello, My database is oracle11g.
    I have same plsql code and first one is without bind variable and second one is with bind variable. Usually, bind variable should take less time. But here
    the bind variable takes more time than the regular code... Can any one please explain why?
    SQL> alter system flush shared_pool;
    System altered.
    SQL> declare
    2 cursor c1 is select * from emp where rownum < 50;
    3 l_start NUMBER DEFAULT DBMS_UTILITY.GET_TIME;
    4 v_cnt number;
    5 begin
    6 for i in c1 loop
    7 SELECT count(*) into v_cnt
    8 FROM rate
    9 WHERE rate_id IN (SELECT rate_id
    10 FROM ratedetail
    11 WHERE benefit_id = i.benefit_id)
    12 AND effective_date =
    13 TO_DATE ('2011-01-23 00:00:00', 'yyyy-MM-dd HH24:MI:SS')
    14 AND rate_type_id = 1;
    15 end loop;
    16 DBMS_OUTPUT.PUT_LINE('total minutes....'||ROUND(ROUND((DBMS_UTILITY.GET_TIME - l_start)/100, 2)
    /60,3));
    17 end;
    18 /
    total minutes.....06
    PL/SQL procedure successfully completed.
    SQL> alter system flush shared_pool;
    System altered.
    SQL>
    SQL> declare
    2 cursor c1 is select benefit_id from emp where rownum < 50;
    3 l_start NUMBER DEFAULT DBMS_UTILITY.GET_TIME;
    4 v_cnt number;
    5 begin
    6 for i in c1 loop
    7 execute immediate 'SELECT count(*)
    8 FROM rate
    9 WHERE rate_id IN (SELECT rate_id
    10 FROM ratedetail
    11 WHERE benefit_id = :x)
    12 AND effective_date = trunc(sysdate)-202
    13 AND rate_type_id = 1'
    14 into v_cnt using i.benefit_id;
    15 end loop;
    16 DBMS_OUTPUT.PUT_LINE('total minutes....'||ROUND(ROUND((DBMS_UTILITY.GET_TIME - l_start)/100, 2)
    /60,3));
    17 end;
    18 /
    total minutes.....061
    PL/SQL procedure successfully completed.
    SQL>

    Shrinika wrote:
    Thanks for the clarification.. Now i understand...
    One final question on this thread before i close this thread....
    My database is set to CURSOR_SHARING=FORCE for some reason. It seems somebody applied a "quick and dirty fix" to "database is slow" problem. BAD PRACTICE
    My question is, when we use bind variable, does it parse the sql code every time? or does it reuse the execution plan?
    In my database, it reuse the execution plan... Just checking... When we set CURSOR_SHARING=FORCE, it should generate the execution plan
    for every unqiue sql code... Is that correct? Am i confusing?If by "parse" you mean a "hard parse" (which generates execution plan), then the answer is NO. As you observed, it reuses execution plan.
    For e.g. with CURSOR_SHARING=FORCE setting, following SQLs
    select employee_no, first_name, last_name from employees where dept_no = 10 ;and
    select employee_no, first_name, last_name from employees where dept_no = 20 ;would tend to reuse the same execution plan since both of these will be rewritten by oracle (before execution) as
    select employee_no, first_name, last_name from employees where dept_no = :SYS01 ;Hope this helps.
    Edited by: user503699 on Aug 14, 2010 3:55 AM

  • Publishing EJBs takes more time in Oracle Weblogic 10.3 workshop

    Hi,
    We recently migrated from Weblogic 9.2 to Oracle Weblogic 10.3 workshop.
    I have a ejb project which has more than 30 ejb's(both entity and session ejbs).
    The problem is: whenever I modify the code inside a method and save, it builds the project automatically. After that when I try to publish the module, it takes around 5 - 10 minutes for publishing.
    Points to note are : The ejb I am modifying has reference to other (10 - 20) ejbs. Hence it takes more time for even for single ejb change.
    however, when i modify an ejb which has no reference to ejb, takes less time to publish.
    My question is : is there a way to reduce the publishing time in this scenario.
    Thanks in advance!

    There is a special forum for Workshop issues:
    Workshop
    Try there.

  • MRP job takes more time (Duration / Sec)

    Hello PP Guru’s,
    I have below situation in production environment,
    MRP job for plant – A (Takes more time – 14.650)
    MRP job for plant – B (Takes less time – 4.512)
    When I compare the variant / attributes with plant A and B, I observe only difference in attribute is scheduling (2 - Lead Time Scheduling and Capacity Planning) for plant A.
    For plant B scheduling (1- Determination of Basic Dates for Planned) it was updated.
    So in my observation this scheduling is playing a major role for MRB job which takes more time in plant – A.
    I am in process of changing the variant attribute for plant – A from scheduling 2 to scheduling 1.
    I wanted to know from experts whether if I change any impact / problem will happen in future for plant A or not.
    Please let me know what all the hidden impacts are there if I change the scheduling in variant attribute.
    I look forward to your valuable input to reduce time for my MRP related job.
    Regards,
    Kumar S

    Hi Kumar,
    You have no need to change the inhouse production time you just need to update the Lot size dependent  inhouse production time in work scheduling view of material master. That you can do by scheduling the routing/recipe.
    Transactions CA97 or CA97N can be used to update the in-house production time with the information from the routing.
    If business don't want to have the capacity planning for planned orders then you change the scheduling from 2 to 1 basic date scheduling.
    Expert Caetano already answer you query
    The reports listed below can be used to compare the MRP past executions regarding the runtime:
    RMMDMONI: This report compares the runtime of the MRP execution and also provides the total of planning elements (planned orders, purchase requisitions, etc) changed, created or deleted. It also shows which planning parameters where used and how much time MRP spent on each step (database read, BOM explosion, MRP calculation, scheduling, BAdIs, etc). With this information is possible to observe the relation of runtime and number of elements changed/created/deleted and also to see on which step MRP is taking more time.
    RMMDPERF: This report shows the "material hit list", that means, which materials had the highest runtime during the MRP execution and also on which step MRP is taking more time. Knowing which materials have the highest runtime, allow you to isolate the problem and reproduce it on MD03, where it is possible to run an ABAP or SQL trace for a more detailed analysis.
    Regards,
    R.Brahmankar

  • Query Take more time on timesten

    Hi
    One query takes lot of time on timesten , while the same query takes less time on oracle
    Query :-
    select *
                    from (SELECT RELD_EM_ENTITY_ID,
                                 RELD_EXM_EXCH_ID,
                                 RELD_SEGMENT_TYPE,
                                NVL(RELD_ACC_TYPE, 0) ACC_TYPE, --END ASHA
                                 ROUND(NVL(sum(RELD_RTO_EXP), -1), 2) RTO_EXP,
                                 ROUND(NVL(sum(RELD_NE_EXP), -1), 2) NET_EXP,
                                 ROUND(NVL(sum(RELD_MAR_UTILIZATION), -1),
                                       2) MAR_EXP,
                                 ROUND(decode(sign(sum(C.reld_m2m_exp)),
                                              -1,
                                              abs(sum(C.reld_m2m_exp)),
                                              0),
                                       2) M2M_EXP,
                                 NVL(OLM_SUSPNSN_FLG, 'A') SUSPNSN_FLG,
                                 EM_RP_PROFILE_ID
                            FROM ENTITY_MASTER         A,
                                 ORDER_LMT_MASTER      B,
                                 RMS_ENTITY_LIMIT_DTLS C
                           WHERE A.EM_CONTROLLER_ID = 100100010000
                             AND A.EM_ENTITY_ID = C.RELD_EM_ENTITY_ID
                             AND C.RELD_EM_ENTITY_ID = B.OLM_EPM_EM_ENTITY_ID(+)
                             AND C.RELD_SEGMENT_TYPE = 'E'
                             AND C.RELD_EXM_EXCH_ID = B.OLM_EXCH_ID(+)
                             AND C.RELD_EXM_EXCH_ID <> 'ALL'
                             AND B.OLM_SEM_SMST_SECURITY_ID(+) =
                                 'ALL'
                             AND ((A.EM_ENTITY_TYPE IN ('CL') AND
                                 A.EM_CLIENT_TYPE <> 'CC') OR
                                 A.EM_ENTITY_TYPE <> 'CL')
                             AND B.OLM_PRODUCT_ID(+) = 'M' --Added by Harshit Shah on 4th June 09
                           GROUP BY RELD_EM_ENTITY_ID,
                                    RELD_EXM_EXCH_ID,
                                    RELD_SEGMENT_TYPE,
                                    RELD_ACC_TYPE,
                                    OLM_SUSPNSN_FLG,
                                    EM_RP_PROFILE_ID,
                                    OLM_PRODUCT_ID
                          UNION --union all removed by pramod on 08-jan-2012 as it was giving multiple rows
                          SELECT RELD_EM_ENTITY_ID,
                                 RELD_EXM_EXCH_ID,
                                 RELD_SEGMENT_TYPE SEGMENTID,
                               NVL(RELD_ACC_TYPE, 0) ACC_TYPE, --END ASHA
                                 ROUND(NVL(SUM(RELD_RTO_EXP), -1), 2) RTO_EXP,
                                 ROUND(NVL(SUM(RELD_NE_EXP), -1), 2) NET_EXP,
                                 ROUND(NVL(SUM(RELD_MAR_UTILIZATION), -1),
                                       2) MAR_EXP,
                                 ROUND(decode(sign(SUM(C.reld_m2m_exp)),
                                              -1,
                                              abs(SUM(C.reld_m2m_exp)),
                                              0),
                                       2) M2M_EXP,
                                 NVL(OLM_SUSPNSN_FLG, 'A') SUSPNSN_FLG,
                                 EM_RP_PROFILE_ID
                            FROM ENTITY_MASTER         A,
                                 ORDER_LMT_MASTER      B,
                                 RMS_ENTITY_LIMIT_DTLS C
                           WHERE A.EM_CONTROLLER_ID = 100100010000
                             AND A.EM_ENTITY_ID = B.OLM_EPM_EM_ENTITY_ID
                             AND B.OLM_EPM_EM_ENTITY_ID = C.RELD_EM_ENTITY_ID(+)
                             AND C.RELD_SEGMENT_TYPE = 'E'
                             AND B.OLM_EXCH_ID = 'ALL'
                             AND B.OLM_SEM_SMST_SECURITY_ID(+) =
                                 'ALL'
                             AND ((A.EM_ENTITY_TYPE IN ('CL') AND
                                 A.EM_CLIENT_TYPE <> 'CC') OR
                                 A.EM_ENTITY_TYPE <> 'CL')
                             AND B.OLM_PRODUCT_ID(+) = 'M' --Added by Harshit Shah on 4th June 09
                           GROUP BY RELD_EM_ENTITY_ID,
                                    RELD_EXM_EXCH_ID,
                                    RELD_SEGMENT_TYPE,
                                    RELD_ACC_TYPE,
                                    OLM_SUSPNSN_FLG,
                                    EM_RP_PROFILE_ID,
                                    OLM_PRODUCT_ID
                          UNION --union all removed by pramod on 08-jan-2012 as it was giving multiple rows
                          SELECT RELD_EM_ENTITY_ID,
                                 RELD_EXM_EXCH_ID,
                                 RELD_SEGMENT_TYPE,
                                 NVL(RELD_ACC_TYPE, 0) ACC_TYPE, --END ASHA
                                 ROUND(NVL(sum(RELD_RTO_EXP), -1), 2) RTO_EXP,
                                 ROUND(NVL(sum(RELD_NE_EXP), -1), 2) NET_EXP,
                                 ROUND(NVL(sum(RELD_MAR_UTILIZATION), -1),
                                       2) MAR_EXP,
                                 ROUND(decode(sign(sum(C.reld_m2m_exp)),
                                              -1,
                                              abs(sum(C.reld_m2m_exp)),
                                              0),
                                       2) M2M_EXP,
                                 NVL(OLIM_SUSPNSN_FLG, 'A') SUSPNSN_FLG,
                                 EM_RP_PROFILE_ID
                            FROM ENTITY_MASTER             A,
                                 DRV_ORDER_INST_LMT_MASTER B,
                                 RMS_ENTITY_LIMIT_DTLS     C
                           WHERE A.EM_CONTROLLER_ID = 100100010000
                             AND A.EM_ENTITY_ID = C.RELD_EM_ENTITY_ID
                             AND C.RELD_EM_ENTITY_ID =
                                 B.OLIM_EPM_EM_ENTITY_ID(+)
                             AND C.RELD_SEGMENT_TYPE = 'D'
                             AND C.RELD_EXM_EXCH_ID = B.OLIM_EXCH_ID(+)
                             AND C.RELD_EXM_EXCH_ID <> 'ALL'
                             AND B.OLIM_INSTRUMENT_ID(+) = 'ALL'
                             AND ((A.EM_ENTITY_TYPE IN ('CL') AND
                                 A.EM_CLIENT_TYPE <> 'CC') OR
                                 A.EM_ENTITY_TYPE <> 'CL')
                           GROUP BY RELD_EM_ENTITY_ID,
                                    RELD_EXM_EXCH_ID,
                                    RELD_SEGMENT_TYPE,
                                    RELD_ACC_TYPE,
                                    OLIM_SUSPNSN_FLG,
                                    EM_RP_PROFILE_ID
                          UNION --union all removed by pramod on 08-jan-2012 as it was giving multiple rows
                          SELECT RELD_EM_ENTITY_ID,
                                 RELD_EXM_EXCH_ID,
                                 RELD_SEGMENT_TYPE SEGMENTID,
                                 NVL(RELD_ACC_TYPE, 0) ACC_TYPE, --END ASHA
                                 ROUND(NVL(SUM(RELD_RTO_EXP), -1), 2) RTO_EXP,
                                 ROUND(NVL(SUM(RELD_NE_EXP), -1), 2) NET_EXP,
                                 ROUND(NVL(SUM(RELD_MAR_UTILIZATION), -1),
                                       2) MAR_EXP,
                                 ROUND(decode(sign(SUM(C.reld_m2m_exp)),
                                              -1,
                                              abs(SUM(C.reld_m2m_exp)),
                                              0),
                                       2) M2M_EXP,
                                 NVL(OLIM_SUSPNSN_FLG, 'A') SUSPNSN_FLG,
                                 EM_RP_PROFILE_ID
                            FROM ENTITY_MASTER             A,
                                 DRV_ORDER_INST_LMT_MASTER B,
                                 RMS_ENTITY_LIMIT_DTLS     C
                           WHERE A.EM_CONTROLLER_ID = 100100010000
                             AND A.EM_ENTITY_ID = B.OLIM_EPM_EM_ENTITY_ID
                             AND B.OLIM_EPM_EM_ENTITY_ID =
                                 C.RELD_EM_ENTITY_ID(+)
                             AND C.RELD_SEGMENT_TYPE = 'D'
                             AND B.OLIM_EXCH_ID = 'ALL'
                             AND B.OLIM_INSTRUMENT_ID(+) = 'ALL'
                             AND ((A.EM_ENTITY_TYPE IN ('CL') AND
                                 A.EM_CLIENT_TYPE <> 'CC') OR
                                 A.EM_ENTITY_TYPE <> 'CL')
                           GROUP BY RELD_EM_ENTITY_ID,
                                    RELD_EXM_EXCH_ID,
                                    RELD_SEGMENT_TYPE,
                                    RELD_ACC_TYPE,
                                    OLIM_SUSPNSN_FLG,
                                    EM_RP_PROFILE_ID)
                   ORDER BY RELD_EM_ENTITY_ID,
                            RELD_SEGMENT_TYPE,
                            RELD_EXM_EXCH_ID;
    Please suggest  what should i check for this.

    As always when examining SQL performance, start by checking the query execution plan. If you use ttIsql you can just prepend EXPLAIN to the query and the plan will be displayed. e.g.
    EXPLAIN  select ...........;
    Check the plan is optimal and all necessary indexes are in place. You may need to add indexes depending o what the plan shows.
    Please note that Oracle database can, and usually does, execute many types of query in parallel using multiple CPU cores. TimesTen does not currently support parallelisation of individual queries. Hence in some cases Oracle database may indeed be faster than TimesTen due to the parallel execution that occurs in Oracle.
    Chris

  • Payables Account Analysis report takes long time to produce xml output

    Hi,
    I am trying to get xml data for the Payables Account Analysis report. I have changed the output format of the concurrent program to XML.
    The report takes long time to produce the xml data irrespective of the number of rows fetched. But the same report with Text output runs very fast.
    Any reason why the xml output takes long time?
    thanks in advance
    Malathi.

    Hi,
    Thanks for the reply.
    As mentioned above, i deleted the Q_FLEXDATA and ran the report. it takes less time.
    But Will the report data not affected when we delete Q_FLEXDATA Group? and Why this flexdata group affects the running time?
    Thanks,
    Malathi.

  • Oracle9i reports take longer time while running in web

    Hi,
    I have developed few reports in Oracle9i and I am trying to run the reports in web. Running a report through report builder takes lesser time compare to running the same report in web using web.show_document. This also depends on the file size. If my report file size(.jsp file) is less than 100KB then it takes 1 minute to show the parameter form and another 1 minute to show the report output. If my file size is around 190KB then the system takes atleast 15 minutes to show the parameter form. Another 10 to 15 minutes to show the report output. I don't understand why the system takes long time to show the parameter form.
    I have a similar problem while opening the file in reports builder also. If my file size is more than 150KB then it takes more than 15 minutes to open the file.
    Could anyone please help me on this.
    Thanks, Radha

    This problem exists only with .jsp reports. I saved the reports in .rdf format and they run faster on web now. Opening a .jsp report takes longer time(file with 600KB takes atleast 2 hours) but the same report in .rdf format takes few seconds to get opened in reports builder.

  • SSRS report is consuming much more time to fetch data from DB even direct run SP takes less than a second

    Hi,
    we are using SQL SERVER 2008R2 X64 RTM version. 
    One of the SSRS report designed by developer  is  consuming  much more time( 5 to 6 minutes ) to fetch data from DB. Even direct run of  Stored Procedure (Called in report )takes less than a second to display the result set.
    Please help.
    Regards Naveen MSSQL DBA

    Hi Naveen,
    Based on my understanding, you spend a little time retrieving data with a stored procedure from database in dataset designer. However, it takes long time to run the report to display the data, right?
    In Reporting Services, the total time to generate a report include TimeDataRetreval, TimeProcessing and TimeRendering. In your scenario, since you mentioned retrieving data costs a little time, you should check the table
    Executionlog3 in the ReportServer database to find which section costs most of time, TimeProcessing or TimeRendering. Then you can refer to this article to optimize your report:
    Troubleshooting Reports: Report Performance.
    Besides, if parameters exist in the report, you should declare variables inside of the stored procedure and assign the incoming parameters to the variables. For more information, please refer to the similar thread:
    Fast query runs slow in SSRS.
    If you have any question, please feel free to ask.
    Best regards,
    Qiuyun Yu
    Qiuyun Yu
    TechNet Community Support

  • Optimizing the query - which takes more time

    Hi,
    Am having a query which was returning the results pretty fast one week back but now the same query takes more time to respond, nothing much changed in the table data, what could be the problem. Am using IN in the where clause, whether that could be an issue? if so what is the best method of rewriting the query.
    SELECT  RI.RESOURCE_NAME,TR.MSISDN,MAX(TR.ADDRESS1_GOOGLE) KEEP(DENSE_RANK LAST ORDER BY TR.MSG_DATE_INFO) ADDRESS1_GOOGLE,
                    MAX(TR.TIME_STAMP) MSG_DATE_INFO FROM  TRACKING_REPORT TR, RESOURCE_INFO RI
                    WHERE TR.MSISDN IN ( SELECT  MSISDN FROM  RESOURCE_INFO WHERE GROUP_ID ='4'
                   AND COM_ID='12') AND RI.MSISDN = TR.MSISDN
                   GROUP BY  RI.RESOURCE_NAME,TR.MSISDN ORDER BY MSG_DATE_INFO DESC

    Hi
    i have followed this link http://www.lorentzcenter.nl/awcourse/oracle/server.920/a96533/sqltrace.htm in enabling the trace and found out the following trace output, can you explain the problem here and its remedial action pls.
    SELECT  RI.RESOURCE_NAME,TR.MSISDN,MAX(TR.ADDRESS1_GOOGLE) KEEP(DENSE_RANK
      LAST ORDER BY TR.MSG_DATE_INFO) ADDRESS1_GOOGLE,                      MAX(TR.TIME_STAMP)
      MSG_DATE_INFO
    FROM
      TRACKING_REPORT TR, RESOURCE_INFO RI                          WHERE RI.GROUP_ID ='426'                         AND
      RI.COM_ID='122' AND RI.MSISDN = TR.MSISDN                      GROUP BY  RI.RESOURCE_NAME,
      TR.MSISDN
    call     count       cpu    elapsed       disk      query    current        rows
    Parse        1      0.01       0.02          0          0          0           0
    Execute      1      0.00       0.00          0          0          0           0
    Fetch        6     13.69     389.03      81747     280722          0          72
    total        8     13.70     389.05      81747     280722          0          72
    Misses in library cache during parse: 1
    Optimizer goal: CHOOSE
    Parsing user id: 281 
    Rows     Row Source Operation
         72  SORT GROUP BY
    276558   NESTED LOOPS 
         79    TABLE ACCESS FULL RESOURCE_INFO
    276558    TABLE ACCESS BY INDEX ROWID TRACKING_REPORT
    276558     INDEX RANGE SCAN TR_INDX_ON_MSISDN_TIME (object id 60507)
    ********************************************************************************and the plan_table output is
    STATEMENT_ID     TIMESTAMP     REMARKS     OPERATION     OPTIONS     OBJECT_NODE     OBJECT_OWNER     OBJECT_NAME     OBJECT_INSTANCE     OBJECT_TYPE     OPTIMIZER     SEARCH_COLUMNS     ID     PARENT_ID     POSITION     COST     CARDINALITY     BYTES     OTHER_TAG     PARTITION_START     PARTITION_STOP     PARTITION_ID     OTHER     DISTRIBUTION     CPU_COST     IO_COST     TEMP_SPACE     ACCESS_PREDICATES     FILTER_PREDICATES
         23-Mar-11 23:36:45          SELECT STATEMENT                                   CHOOSE          0          115     115     1058     111090                                        115               
         23-Mar-11 23:36:45          SORT     GROUP BY                                        1     0     1     115     1058     111090                                        115               
         23-Mar-11 23:36:45          NESTED LOOPS                                             2     1     1     9     4603     483315                                        9               
         23-Mar-11 23:36:45          TABLE ACCESS     FULL          BSNL_RTMS     RESOURCE_INFO     2          ANALYZED          3     2     1     8     1     30                                        8               "RI"."GROUP_ID"=426 AND "RI"."COM_ID"='122'
         23-Mar-11 23:36:45          TABLE ACCESS     BY INDEX ROWID          BSNL_RTMS     TRACKING_REPORT     1          ANALYZED          4     2     2     1     3293     246975                                        1               
         23-Mar-11 23:36:45          INDEX     RANGE SCAN          BSNL_RTMS     TR_INDX_ON_MSISDN_TIME          NON-UNIQUE          1     5     4     1     1     3293                                             1          "RI"."MSISDN"="TR"."MSISDN"     

  • Analyze a Query which takes longer time in Production server with ST03 only

    Hi,
    I want to Analyze a Query which takes longer time in Production server with ST03 t-code only.
    Please provide me with detail steps as to perform the same with ST03
    ST03 - Expert mode- then I need to know the steps after this. I have checked many threads. So please don't send me the links.
    Write steps in detail please.
    <REMOVED BY MODERATOR>
    Regards,
    Sameer
    Edited by: Alvaro Tejada Galindo on Jun 12, 2008 12:14 PM

    Then please close the thread.
    Greetings,
    Blag.

  • A block of code takes more time in JRE 6_20 but less in the previous versio

    while (entries.hasMoreElements()) {
    ZipEntry zipEntry = (ZipEntry) entries.nextElement();
    is = zipFile.getInputStream(zipEntry);
    File file = new File(unzipDir, zipEntry.getName());
    if (is.available() == 0) {
    file.mkdir();
    is.close();
    } else {
    file.createNewFile();
    fos = new FileOutputStream(file);
    CommonUtils.connectIO(is, fos, -1, true);
    }

    Sorry a type,. The above code takes more time when I run withJDK 6_20. But its less time if the previous version is used. Cannot figure whats wrong. Here is the full method.
    * Unzips the specified file into the specified directory. The optional file names list allows the caller
    * to specify the actual files that get unzipped.
    * @param srcFile the file to unzip
    * @param unzipDir the directory where the unzipped files will be put
    * @param fileNames the optional list of name strings for the zip entries to unzip or <code>null</code> to unzip
    * all entries
    * @throws NullPointerException if either the source file or the unzip directory is <code>null</code>
    * @throws javax.faces.FacesException if the
    public static void unzipFile(File srcFile, File unzipDir, List fileNames) {
    if (srcFile == null) {
    throw new NullPointerException("The zip file argument is null");
    if (unzipDir == null) {
    throw new NullPointerException("The unzip directory argument is null");
    ZipFile zipFile = null;
    InputStream is = null;
    FileOutputStream fos = null;
    try {
    zipFile = new ZipFile(srcFile);
    Enumeration entries = null;
    if (fileNames != null) {
    // Use a vector only so we can abstract away the zip entries enumeration...
    Vector v = new Vector();
    Iterator it = fileNames.iterator();
    while (it.hasNext()) {
    String name = (String) it.next();
    v.add(new ZipEntry(name));
    entries = v.elements();
    } else {
    entries = zipFile.entries();
    while (entries.hasMoreElements()) {
    ZipEntry zipEntry = (ZipEntry) entries.nextElement();
    is = zipFile.getInputStream(zipEntry);
    File file = new File(unzipDir, zipEntry.getName());
    if (is.available() == 0) {
    file.mkdir();
    is.close();
    } else {
    file.createNewFile();
    fos = new FileOutputStream(file);
    CommonUtils.connectIO(is, fos, -1, true);
    catch (IOException e) {
    throw new FacesException("Problem unzipping file " + srcFile.getAbsolutePath(), e);
    finally {
    try {
    if (is != null) {
    is.close();
    if (fos != null) {
    fos.close();
    if (zipFile != null) {
    zipFile.close();
    catch (IOException e) {
    throw new FacesException("Problem closing resources when unzipping file " + srcFile.getAbsolutePath(), e);
    }

  • Macbook Pro acting like my good ol'PC  (which cost a lot less).   Won't load Yahoo emails....this since I got Lion.  I don't have time for this $%^

    Mac book pro acting like my good ol'PC which cost a lot less.   Won't load yahoo emails, get the beachball, slow, etc etc.   Not what I paid for....and Apple care on the phone is even less help they $^&% around for some time without a solution, then tell you they will send you to a "senior tech".....huh?   Like my money back for this so called service.  I end up taking my time to go to the genius bar where they have helped.   But this is getting tedious.
    ----lost photos which had to be recovered, even the geniuses don't know whu...
    ---PC, er, excuse me, macbook froze up and essentially crashed.....AC did not know why.  Said they'd call back but never did.
    ---now the latest issue with my personal email account.   I always love it when if it's a non Apple item, there is little they can do.  Well for the money guys, you should and you owe your users that.
    Getting VERY disilliouned with apple products.....

    Your post is in the wrong branch of the community.
    If you have complaints on Apple products feel free to give them a feedback.
    For that follow that link -> http://www.apple.com/contact/feedback.html

Maybe you are looking for

  • Wireless printer and wifi wont connect at same time

    Hello I have a Mac Mini with software version 10.6.8  I also have a HP wireless printer and for some reason cannot get my wirless internet and wireless printer to connect at the same time,  to print anything I have to disconnect from the internet and

  • Video out to Samsung A650 not working correctly

    I have a MBP 2.2Ghz, and I recently purchased a 46" Samsung LN46A650. I used the DVI to VGA adapter and a VGA cable for the first week and it worked. then i wanted to move further from the tv, so i bought a VGA extension cable. tried it, it didn't wo

  • Adobe Acrobat Distiller 9 Problems

    Hi, New to the forum - hopefully someone can help me out. I have also tried the Adobe Forums, Quark Forums etc. etc. with no result - all are blaming Apple - so here I am. (Mod's please move this thread if it isn't in the right place). Specs first -

  • [ANDROID] Authenticate user with Google Account?

    Hello I'm currently developing in Flash Builder 4.5 (Flex Mobile Project) and I'm about to launch a closed beta, and I'd like to do that by authenticating the testers with their Google Account that they use on their Android phones. So now I wonder if

  • How to add trigger on materialized view?

    Hi, I'm trying to add trigger on materialized view. I don't know how to do it. I'm new to oracle and PL/SQL. Can you guys put some working example please? Thank you for every advise.