Delete DML statment tales more time than Update or Insert.
i want to know whether a delete statement takes more time than an update or insert DML command. Please help in solving the doubt.
Regards.
I agree: the amount of ROLLBACK (called UNDO) and ROLLFORWARD (called REDO) information written by the various statement has a crucial impact on the speed.
I did some simple benchmarks for INSERT, UPDATE and DELETE using a 1 million row simple table. As an alternative to the long UPDATEs and DELETEs, I tested also the usual workarounds (which have only partial applicability).
Here are the conclusions (quite important in my opinion, but not to be taken as universal truth):
1. Duration of DML statements for 1 million rows operations (with the size of redo generated):
--- INSERT: 3.5 sec (redo: 3.8 MB)
--- UPDATE: 24.8 sec (redo: 240 MB)
--- DELETE: 26.1 sec (redo: 228 MB)
2. Replacement of DELETE with TRUNCATE
--- DELETE: 26.1 sec (rollback: 228 MB)
--- TRUNCATE: 0.1 sec (rollback: 0.1 MB)
3. Replacement of UPDATE with CREATE new TABLE AS SELECT (followed by DROP ols and RENAME new AS old)
--- UPDATE: 24.8 sec (redo_size: 240 MB)
--- replacement: 3.5 sec (rollback: 0.3 MB)
-- * Preparation *
CREATE TABLE ao AS
SELECT rownum AS id,
'N' || rownum AS name
FROM all_objects, all_objects
WHERE rownum <= 1000000;
CREATE OR REPLACE PROCEDURE print_my_stat(p_name IN v$statname.NAME%TYPE) IS
v_value v$mystat.VALUE%TYPE;
BEGIN
SELECT b.VALUE
INTO v_value
FROM v$statname a,
v$mystat b
WHERE a.statistic# = b.statistic# AND lower(a.NAME) LIKE lower(p_name);
dbms_output.put_line('*' || p_name || ': ' || v_value);
END print_my_stat;
-- * Test 1: Comparison of INSERT, UPDATE and DELETE *
CREATE TABLE ao1 AS
SELECT * FROM ao WHERE 1 = 2;
exec print_my_stat('redo_size')
*redo_size= 277,220,544
INSERT INTO ao1 SELECT * FROM ao;
1000000 rows inserted
executed in 3.465 seconds
exec print_my_stat('redo_size')
*redo_size= 301,058,852
commit;
UPDATE ao1 SET name = 'M' || SUBSTR(name, 2);
1000000 rows updated
executed in 24.786 seconds
exec print_my_stat('redo_size')
*redo_size= 545,996,280
commit;
DELETE FROM ao1;
1000000 rows deleted
executed in 26.128 seconds
exec print_my_stat('redo_size')
*redo_size= 783,655,196
commit;
-- * Test 2: Replace DELETE with TRUNCATE *
DROP TABLE ao1;
CREATE TABLE ao1 AS
SELECT * FROM ao;
exec print_my_stat('redo_size')
*redo_size= 807,554,512
TRUNCATE TABLE ao1;
executed in 0.08 seconds
exec print_my_stat('redo_size')
*redo_size= 807,616,528
-- * Test 3: Replace UPDATE with CREATE TABLE AS SELECT *
INSERT INTO ao1 SELECT * FROM ao;
commit;
exec print_my_stat('redo_size')
*redo_size= 831,525,556
CREATE TABLE ao2 AS
SELECT id, 'M' || SUBSTR(name, 2) name FROM ao1;
executed in 3.125 seconds
DROP TABLE ao1;
executed in 0.32 seconds
RENAME ao2 TO ao1;
executed in 0.01 seconds
exec print_my_stat('redo_size')
*redo_size= 831,797,608
Similar Messages
-
Delete DML statment takes more time than Update or Insert.
i want to know whether a delete statement takes more time than an update or insert DML command. Please help in solving the doubt.
Regards.i do not get good answers sometimes, so, i ask again.I think Alex answer to your post was quite complete. If you missed some information, continue the same post, instead of opening a new thread with the same subject and content.
You should be satistied with the answers you get, I also answered your question about global indexes, and I do think my answer was very complete. You may ask more if you want, but stop multiposting please. It is quite annoying.
Ok, have a nice day -
I use itunes on a Dell XPS502 with W7/64. In some cases have have problems to import CD's. The sound is very disturbed and the import need a lot more time than in normal cases. Is there a problem between itunes and W7/64 or a known hardware issue?
Example-CD : "Tracy Chapman , Telling stories" is not able to import . I have more such negative cases. But in other cases it works fine and the sound is great.
The firmware at the inbuild CD/DVD DS-6E2SH is the latest version.
What can I do??hi b noir,
I don't no about virtuel drives like you mententioned. In the mean time I have rebooted the XPS and run again the iTunes diagnostics. I think the back - chance in the registry was not ready to use. Now there are another results. They are the same in case of a running CD or a not running CD. The difference in total is like before. It takes more time that iTunes reads the (bad) CD and at the there is no good sound. In both cases ( running or not running CD) iTunes diagnostics gives as a result :
(the copie from ITunes shows the result of the not running CD from Tracy Chapman)
Microsoft Windows 7 x64 Ultimate Edition Service Pack 1 (Build 7601)
Dell Inc. Dell System XPS L502X
iTunes 10.3.1.55
QuickTime 7.6.9
FairPlay 1.11.17
Apple Application Support 1.5.2
iPod Updater-Bibliothek 10.0d2
CD Driver 2.2.0.1
CD Driver DLL 2.1.1.1
Apple Mobile Device 3.4.0.25
Apple Mobile Device Treiber 1.55.0.0
Bonjour 2.0.5.0 (214.3)
Gracenote SDK 1.8.2.457
Gracenote MusicID 1.8.2.89
Gracenote Submit 1.8.2.123
Gracenote DSP 1.8.2.34
iTunes-Seriennummer 00D7B2B00CD25750
Aktueller Benutzer ist kein Administrator.
Aktuelles Datum und Uhrzeit sind 2011-06-11 19:33:22.
iTunes befindet sich nicht im abgesicherten Modus.
WebKit Accelerated Compositing ist aktiviert.
HDCP wird unterstützt.
Core Media wird unterstützt.
Info zu Video-Anzeige
NVIDIA, NVIDIA GeForce GT 540M
Intel Corporation, Intel(R) HD Graphics Family
**** Info für externe Plug-Ins ****
Keine externen Plug-Ins installiert.
iPodService 10.3.1.55 (x64) arbeitet zurzeit.
iTunesHelper 10.3.1.55 arbeitet zurzeit.
Apple Mobile Device service 3.3.0.0 arbeitet zurzeit.
**** CD/DVD-Laufwerkstests****
LowerFilters: PxHlpa64 (2.0.0.0),
UpperFilters: GEARAspiWDM (2.2.0.1),
D: PLDS DVDRWBD DS-6E2SH, Rev CD11
Audio-CD im Laufwerk
11 Titel auf der CD gefunden, Spieldauer: 42:07 auf Audio-CD
Titel 1, Startzeit: 00:02:00
Titel 2, Startzeit: 03:59:47
Titel 3, Startzeit: 07:19:27
Titel 4, Startzeit: 11:31:30
Titel 5, Startzeit: 15:31:50
Titel 6, Startzeit: 20:07:50
Titel 7, Startzeit: 24:27:15
Titel 8, Startzeit: 27:49:10
Titel 9, Startzeit: 32:41:25
Titel 10, Startzeit: 35:29:65
Titel 11, Startzeit: 38:38:00
Audio-CD erfolgreich gelesen (Suche nach alter Firmware).
Laufwerksgeschwindigkeit erfolgreich erkannt
Die CDR-Geschwindigkeiten des Laufwerks sind: 4 10 16 24
Die CDRW-Geschwindigkeiten des Laufwerks sind: 4
Die DVDR-Geschwindigkeiten des Laufwerks sind: 4
Die DVDRW-Geschwindigkeiten des Laufwerks sind: 4
After starting the import it is going slower and slower. If it is helpful I can send you a soundfile with these distortions.
best regards
tcgerd -
I'm getting "exceeding my limit" messages more times than not on my iPad. I feel I'm not sending anything extraordinary. No mass sending but only replying to emails sent to me. I have 7.5 G remaining on my iCloud account. Is it because I have a lot of emails files? Please help because I'm getting frustrated with this.
Hello, PriestessJeanann.
Thank you for visiting Apple Support Communities.
Here is an article I would recommend going through when experiencing issues with mail. The usual fix would be to delete the email account in question and add this account via the preset AOL option. I would also recommend checking with your email provider for security procedures such as two-step verification as this could cause this issue.
iOS: Troubleshooting Mail
http://support.apple.com/kb/ts3899
Cheers,
Jason H. -
Query in timesten taking more time than query in oracle database
Hi,
Can anyone please explain me why query in timesten taking more time
than query in oracle database.
I am mentioning in detail what are my settings and what have I done
step by step.........
1.This is the table I created in Oracle datababase
(Oracle Database 10g Enterprise Edition Release 10.2.0.1.0)...
CREATE TABLE student (
id NUMBER(9) primary keY ,
first_name VARCHAR2(10),
last_name VARCHAR2(10)
2.THIS IS THE ANONYMOUS BLOCK I USE TO
POPULATE THE STUDENT TABLE(TOTAL 2599999 ROWS)...
declare
firstname varchar2(12);
lastname varchar2(12);
catt number(9);
begin
for cntr in 1..2599999 loop
firstname:=(cntr+8)||'f';
lastname:=(cntr+2)||'l';
if cntr like '%9999' then
dbms_output.put_line(cntr);
end if;
insert into student values(cntr,firstname, lastname);
end loop;
end;
3. MY DSN IS SET THE FOLLWING WAY..
DATA STORE PATH- G:\dipesh3repo\db
LOG DIRECTORY- G:\dipesh3repo\log
PERM DATA SIZE-1000
TEMP DATA SIZE-1000
MY TIMESTEN VERSION-
C:\Documents and Settings\dipesh>ttversion
TimesTen Release 7.0.3.0.0 (32 bit NT) (tt70_32:17000) 2007-09-19T16:04:16Z
Instance admin: dipesh
Instance home directory: G:\TimestTen\TT70_32
Daemon home directory: G:\TimestTen\TT70_32\srv\info
THEN I CONNECT TO THE TIMESTEN DATABASE
C:\Documents and Settings\dipesh> ttisql
command>connect "dsn=dipesh3;oraclepwd=tiger";
4. THEN I START THE AGENT
call ttCacheUidPwdSet('SCOTT','TIGER');
Command> CALL ttCacheStart();
5.THEN I CREATE THE READ ONLY CACHE GROUP AND LOAD IT
create readonly cache group rc_student autorefresh
interval 5 seconds from student
(id int not null primary key, first_name varchar2(10), last_name varchar2(10));
load cache group rc_student commit every 100 rows;
6.NOW I CAN ACCESS THE TABLES FROM TIMESTEN AND PERFORM THE QUERY
I SET THE TIMING..
command>TIMING 1;
consider this query now..
Command> select * from student where first_name='2155666f';
< 2155658, 2155666f, 2155660l >
1 row found.
Execution time (SQLExecute + Fetch Loop) = 0.668822 seconds.
another query-
Command> SELECT * FROM STUDENTS WHERE FIRST_NAME='2340009f';
2206: Table SCOTT.STUDENTS not found
Execution time (SQLPrepare) = 0.074964 seconds.
The command failed.
Command> SELECT * FROM STUDENT where first_name='2093434f';
< 2093426, 2093434f, 2093428l >
1 row found.
Execution time (SQLExecute + Fetch Loop) = 0.585897 seconds.
Command>
7.NOW I PERFORM THE SIMILAR QUERIES FROM SQLPLUS...
SQL> SELECT * FROM STUDENT WHERE FIRST_NAME='1498671f';
ID FIRST_NAME LAST_NAME
1498663 1498671f 1498665l
Elapsed: 00:00:00.15
Can anyone please explain me why query in timesten taking more time
that query in oracle database.
Message was edited by: Dipesh Majumdar
user542575
Message was edited by:
user542575TimesTen
Hardware: Windows Server 2003 R2 Enterprise x64; 8 x Dual-core AMD 8216 2.41GHz processors; 32 GB RAM
Version: 7.0.4.0.0 64 bit
Schema:
create usermanaged cache group factCache from
MV_US_DATAMART
ORDER_DATE DATE,
IF_SYSTEM VARCHAR2(32) NOT NULL,
GROUPING_ID TT_BIGINT,
TIME_DIM_ID TT_INTEGER NOT NULL,
BUSINESS_DIM_ID TT_INTEGER NOT NULL,
ACCOUNT_DIM_ID TT_INTEGER NOT NULL,
ORDERTYPE_DIM_ID TT_INTEGER NOT NULL,
INSTR_DIM_ID TT_INTEGER NOT NULL,
EXECUTION_DIM_ID TT_INTEGER NOT NULL,
EXEC_EXCHANGE_DIM_ID TT_INTEGER NOT NULL,
NO_ORDERS TT_BIGINT,
FILLED_QUANTITY TT_BIGINT,
CNT_FILLED_QUANTITY TT_BIGINT,
QUANTITY TT_BIGINT,
CNT_QUANTITY TT_BIGINT,
COMMISSION BINARY_FLOAT,
CNT_COMMISSION TT_BIGINT,
FILLS_NUMBER TT_BIGINT,
CNT_FILLS_NUMBER TT_BIGINT,
AGGRESSIVE_FILLS TT_BIGINT,
CNT_AGGRESSIVE_FILLS TT_BIGINT,
NOTIONAL BINARY_FLOAT,
CNT_NOTIONAL TT_BIGINT,
TOTAL_PRICE BINARY_FLOAT,
CNT_TOTAL_PRICE TT_BIGINT,
CANCELLED_ORDERS_COUNT TT_BIGINT,
CNT_CANCELLED_ORDERS_COUNT TT_BIGINT,
ROUTED_ORDERS_NO TT_BIGINT,
CNT_ROUTED_ORDERS_NO TT_BIGINT,
ROUTED_LIQUIDITY_QTY TT_BIGINT,
CNT_ROUTED_LIQUIDITY_QTY TT_BIGINT,
REMOVED_LIQUIDITY_QTY TT_BIGINT,
CNT_REMOVED_LIQUIDITY_QTY TT_BIGINT,
ADDED_LIQUIDITY_QTY TT_BIGINT,
CNT_ADDED_LIQUIDITY_QTY TT_BIGINT,
AGENT_CHARGES BINARY_FLOAT,
CNT_AGENT_CHARGES TT_BIGINT,
CLEARING_CHARGES BINARY_FLOAT,
CNT_CLEARING_CHARGES TT_BIGINT,
EXECUTION_CHARGES BINARY_FLOAT,
CNT_EXECUTION_CHARGES TT_BIGINT,
TRANSACTION_CHARGES BINARY_FLOAT,
CNT_TRANSACTION_CHARGES TT_BIGINT,
ORDER_MANAGEMENT BINARY_FLOAT,
CNT_ORDER_MANAGEMENT TT_BIGINT,
SETTLEMENT_CHARGES BINARY_FLOAT,
CNT_SETTLEMENT_CHARGES TT_BIGINT,
RECOVERED_AGENT BINARY_FLOAT,
CNT_RECOVERED_AGENT TT_BIGINT,
RECOVERED_CLEARING BINARY_FLOAT,
CNT_RECOVERED_CLEARING TT_BIGINT,
RECOVERED_EXECUTION BINARY_FLOAT,
CNT_RECOVERED_EXECUTION TT_BIGINT,
RECOVERED_TRANSACTION BINARY_FLOAT,
CNT_RECOVERED_TRANSACTION TT_BIGINT,
RECOVERED_ORD_MGT BINARY_FLOAT,
CNT_RECOVERED_ORD_MGT TT_BIGINT,
RECOVERED_SETTLEMENT BINARY_FLOAT,
CNT_RECOVERED_SETTLEMENT TT_BIGINT,
CLIENT_AGENT BINARY_FLOAT,
CNT_CLIENT_AGENT TT_BIGINT,
CLIENT_ORDER_MGT BINARY_FLOAT,
CNT_CLIENT_ORDER_MGT TT_BIGINT,
CLIENT_EXEC BINARY_FLOAT,
CNT_CLIENT_EXEC TT_BIGINT,
CLIENT_TRANS BINARY_FLOAT,
CNT_CLIENT_TRANS TT_BIGINT,
CLIENT_CLEARING BINARY_FLOAT,
CNT_CLIENT_CLEARING TT_BIGINT,
CLIENT_SETTLE BINARY_FLOAT,
CNT_CLIENT_SETTLE TT_BIGINT,
CHARGEABLE_TAXES BINARY_FLOAT,
CNT_CHARGEABLE_TAXES TT_BIGINT,
VENDOR_CHARGE BINARY_FLOAT,
CNT_VENDOR_CHARGE TT_BIGINT,
ROUTING_CHARGES BINARY_FLOAT,
CNT_ROUTING_CHARGES TT_BIGINT,
RECOVERED_ROUTING BINARY_FLOAT,
CNT_RECOVERED_ROUTING TT_BIGINT,
CLIENT_ROUTING BINARY_FLOAT,
CNT_CLIENT_ROUTING TT_BIGINT,
TICKET_CHARGES BINARY_FLOAT,
CNT_TICKET_CHARGES TT_BIGINT,
RECOVERED_TICKET_CHARGES BINARY_FLOAT,
CNT_RECOVERED_TICKET_CHARGES TT_BIGINT,
PRIMARY KEY(ORDER_DATE, TIME_DIM_ID, BUSINESS_DIM_ID, ACCOUNT_DIM_ID, ORDERTYPE_DIM_ID, INSTR_DIM_ID, EXECUTION_DIM_ID,EXEC_EXCHANGE_DIM_ID),
READONLY);
No of rows: 2228558
Config:
< CkptFrequency, 600 >
< CkptLogVolume, 0 >
< CkptRate, 0 >
< ConnectionCharacterSet, US7ASCII >
< ConnectionName, tt_us_dma >
< Connections, 64 >
< DataBaseCharacterSet, AL32UTF8 >
< DataStore, e:\andrew\datacache\usDMA >
< DurableCommits, 0 >
< GroupRestrict, <NULL> >
< LockLevel, 0 >
< LockWait, 10 >
< LogBuffSize, 65536 >
< LogDir, e:\andrew\datacache\ >
< LogFileSize, 64 >
< LogFlushMethod, 1 >
< LogPurge, 0 >
< Logging, 1 >
< MemoryLock, 0 >
< NLS_LENGTH_SEMANTICS, BYTE >
< NLS_NCHAR_CONV_EXCP, 0 >
< NLS_SORT, BINARY >
< OracleID, NYCATP1 >
< PassThrough, 0 >
< PermSize, 4000 >
< PermWarnThreshold, 90 >
< PrivateCommands, 0 >
< Preallocate, 0 >
< QueryThreshold, 0 >
< RACCallback, 0 >
< SQLQueryTimeout, 0 >
< TempSize, 514 >
< TempWarnThreshold, 90 >
< Temporary, 1 >
< TransparentLoad, 0 >
< TypeMode, 0 >
< UID, OS_OWNER >
ORACLE:
Hardware: Sunos 5.10; 24x1.8Ghz (unsure of type); 82 GB RAM
Version 10g Enterprise Edition Release 10.2.0.3.0 - 64bit Production
Schema:
CREATE MATERIALIZED VIEW OS_OWNER.MV_US_DATAMART
TABLESPACE TS_OS
PARTITION BY RANGE (ORDER_DATE)
PARTITION MV_US_DATAMART_MINVAL VALUES LESS THAN (TO_DATE(' 2007-11-01 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_07_NOV_D1 VALUES LESS THAN (TO_DATE(' 2007-11-11 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_07_NOV_D2 VALUES LESS THAN (TO_DATE(' 2007-11-21 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_07_NOV_D3 VALUES LESS THAN (TO_DATE(' 2007-12-01 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_07_DEC_D1 VALUES LESS THAN (TO_DATE(' 2007-12-11 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_07_DEC_D2 VALUES LESS THAN (TO_DATE(' 2007-12-21 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_07_DEC_D3 VALUES LESS THAN (TO_DATE(' 2008-01-01 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_08_JAN_D1 VALUES LESS THAN (TO_DATE(' 2008-01-11 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_08_JAN_D2 VALUES LESS THAN (TO_DATE(' 2008-01-21 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_08_JAN_D3 VALUES LESS THAN (TO_DATE(' 2008-02-01 00:00:00', 'SYYYY-MM-DD HH24:MI:SS', 'NLS_CALENDAR=GREGORIAN'))
LOGGING
NOCOMPRESS
TABLESPACE TS_OS,
PARTITION MV_US_DATAMART_MAXVAL VALUES LESS THAN (MAXVALUE)
LOGGING
NOCOMPRESS
TABLESPACE TS_OS
NOCACHE
NOCOMPRESS
NOPARALLEL
BUILD DEFERRED
USING INDEX
TABLESPACE TS_OS_INDEX
REFRESH FAST ON DEMAND
WITH PRIMARY KEY
ENABLE QUERY REWRITE
AS
SELECT order_date, if_system,
GROUPING_ID (order_date,
if_system,
business_dim_id,
time_dim_id,
account_dim_id,
ordertype_dim_id,
instr_dim_id,
execution_dim_id,
exec_exchange_dim_id
) GROUPING_ID,
/* ============ DIMENSIONS ============ */
time_dim_id, business_dim_id, account_dim_id, ordertype_dim_id,
instr_dim_id, execution_dim_id, exec_exchange_dim_id,
/* ============ MEASURES ============ */
-- o.FX_RATE /* FX_RATE */,
COUNT (*) no_orders,
-- SUM(NO_ORDERS) NO_ORDERS,
-- COUNT(NO_ORDERS) CNT_NO_ORDERS,
SUM (filled_quantity) filled_quantity,
COUNT (filled_quantity) cnt_filled_quantity, SUM (quantity) quantity,
COUNT (quantity) cnt_quantity, SUM (commission) commission,
COUNT (commission) cnt_commission, SUM (fills_number) fills_number,
COUNT (fills_number) cnt_fills_number,
SUM (aggressive_fills) aggressive_fills,
COUNT (aggressive_fills) cnt_aggressive_fills,
SUM (fx_rate * filled_quantity * average_price) notional,
COUNT (fx_rate * filled_quantity * average_price) cnt_notional,
SUM (fx_rate * fills_number * average_price) total_price,
COUNT (fx_rate * fills_number * average_price) cnt_total_price,
SUM (CASE
WHEN order_status = 'C'
THEN 1
ELSE 0
END) cancelled_orders_count,
COUNT (CASE
WHEN order_status = 'C'
THEN 1
ELSE 0
END
) cnt_cancelled_orders_count,
-- SUM(t.FX_RATE*t.NO_FILLS*t.AVG_PRICE) AVERAGE_PRICE,
-- SUM(FILLS_NUMBER*AVERAGE_PRICE) STAGING_AVERAGE_PRICE,
-- COUNT(FILLS_NUMBER*AVERAGE_PRICE) CNT_STAGING_AVERAGE_PRICE,
SUM (routed_orders_no) routed_orders_no,
COUNT (routed_orders_no) cnt_routed_orders_no,
SUM (routed_liquidity_qty) routed_liquidity_qty,
COUNT (routed_liquidity_qty) cnt_routed_liquidity_qty,
SUM (removed_liquidity_qty) removed_liquidity_qty,
COUNT (removed_liquidity_qty) cnt_removed_liquidity_qty,
SUM (added_liquidity_qty) added_liquidity_qty,
COUNT (added_liquidity_qty) cnt_added_liquidity_qty,
SUM (agent_charges) agent_charges,
COUNT (agent_charges) cnt_agent_charges,
SUM (clearing_charges) clearing_charges,
COUNT (clearing_charges) cnt_clearing_charges,
SUM (execution_charges) execution_charges,
COUNT (execution_charges) cnt_execution_charges,
SUM (transaction_charges) transaction_charges,
COUNT (transaction_charges) cnt_transaction_charges,
SUM (order_management) order_management,
COUNT (order_management) cnt_order_management,
SUM (settlement_charges) settlement_charges,
COUNT (settlement_charges) cnt_settlement_charges,
SUM (recovered_agent) recovered_agent,
COUNT (recovered_agent) cnt_recovered_agent,
SUM (recovered_clearing) recovered_clearing,
COUNT (recovered_clearing) cnt_recovered_clearing,
SUM (recovered_execution) recovered_execution,
COUNT (recovered_execution) cnt_recovered_execution,
SUM (recovered_transaction) recovered_transaction,
COUNT (recovered_transaction) cnt_recovered_transaction,
SUM (recovered_ord_mgt) recovered_ord_mgt,
COUNT (recovered_ord_mgt) cnt_recovered_ord_mgt,
SUM (recovered_settlement) recovered_settlement,
COUNT (recovered_settlement) cnt_recovered_settlement,
SUM (client_agent) client_agent,
COUNT (client_agent) cnt_client_agent,
SUM (client_order_mgt) client_order_mgt,
COUNT (client_order_mgt) cnt_client_order_mgt,
SUM (client_exec) client_exec, COUNT (client_exec) cnt_client_exec,
SUM (client_trans) client_trans,
COUNT (client_trans) cnt_client_trans,
SUM (client_clearing) client_clearing,
COUNT (client_clearing) cnt_client_clearing,
SUM (client_settle) client_settle,
COUNT (client_settle) cnt_client_settle,
SUM (chargeable_taxes) chargeable_taxes,
COUNT (chargeable_taxes) cnt_chargeable_taxes,
SUM (vendor_charge) vendor_charge,
COUNT (vendor_charge) cnt_vendor_charge,
SUM (routing_charges) routing_charges,
COUNT (routing_charges) cnt_routing_charges,
SUM (recovered_routing) recovered_routing,
COUNT (recovered_routing) cnt_recovered_routing,
SUM (client_routing) client_routing,
COUNT (client_routing) cnt_client_routing,
SUM (ticket_charges) ticket_charges,
COUNT (ticket_charges) cnt_ticket_charges,
SUM (recovered_ticket_charges) recovered_ticket_charges,
COUNT (recovered_ticket_charges) cnt_recovered_ticket_charges
FROM us_datamart_raw
GROUP BY order_date,
if_system,
business_dim_id,
time_dim_id,
account_dim_id,
ordertype_dim_id,
instr_dim_id,
execution_dim_id,
exec_exchange_dim_id;
-- Note: Index I_SNAP$_MV_US_DATAMART will be created automatically
-- by Oracle with the associated materialized view.
CREATE UNIQUE INDEX OS_OWNER.MV_US_DATAMART_UDX ON OS_OWNER.MV_US_DATAMART
(ORDER_DATE, TIME_DIM_ID, BUSINESS_DIM_ID, ACCOUNT_DIM_ID, ORDERTYPE_DIM_ID,
INSTR_DIM_ID, EXECUTION_DIM_ID, EXEC_EXCHANGE_DIM_ID)
NOLOGGING
NOPARALLEL
COMPRESS 7;
No of rows: 2228558
The query (taken Mondrian) I run against each of them is:
select sum("MV_US_DATAMART"."NOTIONAL") as "m0"
--, sum("MV_US_DATAMART"."FILLED_QUANTITY") as "m1"
--, sum("MV_US_DATAMART"."AGENT_CHARGES") as "m2"
--, sum("MV_US_DATAMART"."CLEARING_CHARGES") as "m3"
--, sum("MV_US_DATAMART"."EXECUTION_CHARGES") as "m4"
--, sum("MV_US_DATAMART"."TRANSACTION_CHARGES") as "m5"
--, sum("MV_US_DATAMART"."ROUTING_CHARGES") as "m6"
--, sum("MV_US_DATAMART"."ORDER_MANAGEMENT") as "m7"
--, sum("MV_US_DATAMART"."SETTLEMENT_CHARGES") as "m8"
--, sum("MV_US_DATAMART"."COMMISSION") as "m9"
--, sum("MV_US_DATAMART"."RECOVERED_AGENT") as "m10"
--, sum("MV_US_DATAMART"."RECOVERED_CLEARING") as "m11"
--,sum("MV_US_DATAMART"."RECOVERED_EXECUTION") as "m12"
--,sum("MV_US_DATAMART"."RECOVERED_TRANSACTION") as "m13"
--, sum("MV_US_DATAMART"."RECOVERED_ROUTING") as "m14"
--, sum("MV_US_DATAMART"."RECOVERED_ORD_MGT") as "m15"
--, sum("MV_US_DATAMART"."RECOVERED_SETTLEMENT") as "m16"
--, sum("MV_US_DATAMART"."RECOVERED_TICKET_CHARGES") as "m17"
--,sum("MV_US_DATAMART"."TICKET_CHARGES") as "m18"
--, sum("MV_US_DATAMART"."VENDOR_CHARGE") as "m19"
from "OS_OWNER"."MV_US_DATAMART" "MV_US_DATAMART"
where I uncomment a column at a time and rerun. I improved the TimesTen results since my first post, by retyping the NUMBER columns to BINARY_FLOAT. The results I got were:
No Columns ORACLE TimesTen
1 1.05 0.94
2 1.07 1.47
3 2.04 1.8
4 2.06 2.08
5 2.09 2.4
6 3.01 2.67
7 4.02 3.06
8 4.03 3.37
9 4.04 3.62
10 4.06 4.02
11 4.08 4.31
12 4.09 4.61
13 5.01 4.76
14 5.02 5.06
15 5.04 5.25
16 5.05 5.48
17 5.08 5.84
18 6 6.21
19 6.02 6.34
20 6.04 6.75 -
Error in sql query as "loop has run more times than expected (Loop Counter went negative)"
Hello,
When I run the query as below
DECLARE @LoopCount int
SET @LoopCount = (SELECT Count(*) FROM KC_PaymentTransactionIDConversion with (nolock) Where KC_Transaction_ID is NULL and TransactionYear is NOT NULL)
WHILE (
SELECT Count(*)
FROM KC_PaymentTransactionIDConversion with (nolock)
Where KC_Transaction_ID is NULL
and TransactionYear is NOT NULL
) > 0
BEGIN
IF @LoopCount < 0
RAISERROR ('Issue with data in KC_PaymentTransactionIDConversion, loop has run more times than expected (Loop Counter went negative).', -- Message text.
16, -- Severity.
1 -- State.
SET @LoopCount = @LoopCount - 1
end
I am getting error as "loop has run more times than expected (Loop Counter went negative)"
Could any one help on this issue ASAP.
Thanks ,
VinayHi Vinay,
According to your code above, the error message make sense. Because once the value returned by “SELECT Count(*) FROM KC_PaymentTransactionIDConversion with (nolock) Where KC_Transaction_ID is NULL and TransactionYear is NOT NULL” is bigger than 0,
then decrease @LoopCount. Without changing the table data, the returned value always bigger than 0, always decrease @LoopCount until it's negative and raise the error.
To fix this issue with the current information, we should make the following modification:
Change the code
WHILE (
SELECT Count(*)
FROM KC_PaymentTransactionIDConversion with (nolock)
Where KC_Transaction_ID is NULL
and TransactionYear is NOT NULL
) > 0
To
WHILE @LoopCount > 0
Besides, since the current query is senseless, please modify the query based on your requirement.
If there are any other questions, please feel free to ask.
Thanks,
Katherine Xiong
Katherine Xiong
TechNet Community Support -
Count (*) for select stmt take more time than execute a that sql stmt
HI
count (*) for select stmt take more time than execute a that sql stmt
executing particular select stmt take 2.47 mins but select stmt is using the /*+parallel*/ (sql optimer) in that sql command for faster execute .
but if i tried to find out total number of rows in that query it takes more time ..
almost 2.30 hrs still running to find count(col)
please help me to get count of row faster.
thanks in advance...797525 wrote:
HI
count (*) for select stmt take more time than execute a that sql stmt
executing particular select stmt take 2.47 mins but select stmt is using the /*+parallel*/ (sql optimer) in that sql command for faster execute .
but if i tried to find out total number of rows in that query it takes more time ..
almost 2.30 hrs still running to find count(col)
please help me to get count of row faster.
thanks in advance...That may be because your client is displaying only the first few records when you are running the "SELECT *". But when you run "COUNT(*)", the whole records has to be counted.
As already mentined please read teh FAQ to post tuning questions. -
Why import of change request in production takes more time than quality?
Hello All,
why import of change request in production takes more time than import into quality?Hi jahangeer,
I believe it takes same time to import a request in both quality and production as they will be in sync.
Even then if it takes more time in production that may depend on the change request.
Thanks
Pavan -
Level1 backup is taking more time than Level0
The Level1 backup is taking more time than Level0, I really am frustated how could it happen. I have 6.5GB of database. Level0 took 8 hrs but level1 is taking more than 8hrs . please help me in this regard.
Ogan Ozdogan wrote:
Charles,
By enabling the block change tracking will be indeed faster than before he have got. I think this does not address the question of the OP unless you are saying the incremental backup without the block change tracking is slower than a level 0 (full) backup?
Thank you in anticipating.
OganOgan,
I can't explain why a 6.5GB level 0 RMAN backup would require 8 hours to complete (maybe a very slow destination device connected by 10Mb/s Ethernet) - I would expect that it should complete in a couple of minutes.
An incremental level 1 backup without a block change tracking file could take longer than a level 0 backup. I encountered a good written description of why that could happen, but I can't seem to locate the source at the moment. The longer run time might have been related to the additional code paths required to constantly compare the SCN of each block, and the variable write rate which may affect some devices, such as a tape device.
A paraphrase from the book "Oracle Database 10g RMAN Backup & Recovery"
"Incremental backups must check the header of each block to discover if it has changed since the last incremental backup - that means an incremental backup may not complete much faster than a full backup."
Charles Hooper
Co-author of "Expert Oracle Practices: Oracle Database Administration from the Oak Table"
http://hoopercharles.wordpress.com/
IT Manager/Oracle DBA
K&M Machine-Fabricating, Inc. -
When i put my Mac for sleep it takes more time than normal (>20 secs). Sometimes, coming back from sleep the system is not responding (freeze).
Perform SMC and NVRAM resets:
http://support.apple.com/en-us/HT201295
http://support.apple.com/en-us/HT204063
The try a safe boot:
http://support.apple.com/en-us/HT201262
Any change?
Ciao. -
When will there be fix for AirPrint? Since upgrading to Yosemite I cannot print from my MacBook Air which I use mostly. (I can print from the my old 2005 G5 desktop and iPad2 just fine on the same network.) I could print just fine when I had Mavericks. I have tried EVERYTHING under the sun: resetting print system, deleting and adding the printer, checking for updates, uninstalling and reinstalling printer software and folders, turning off modem, router etc. NOTHING works.
The laptop shows that it is printing but nothing happens. Then it says it cannot locate the printer or there is a broken pipe. But it does tell me that I have a low level in one of the ink cartridges (which is true.)
Help! I have warned everyone I know not to upgrade. Is there a way to go back to Mavericks?
MacBook Air (mid 2012), OS 10.10, HP Officejet 7500aThis is what appeared on Console:
11/12/14 12:42:32.558 PM com.apple.preference.printfax.remoteservice[3099]: <NSViewServiceMarshal: 0x7fa398514bd0> failed to complete rights grant 6B212CA9-7872-4726-BA05-68BD5C994479 due to 3
timestamp: 12:42:32.557 Wednesday 12 November 2014
process/thread/queue: com.apple.preference.printfax.remoteservice (3099) / 0x7fff78cb4300 / com.apple.main-thread
code: line 834 of /SourceCache/ViewBridge/ViewBridge-99/NSViewServiceMarshal.m in -[NSViewServiceMarshal invalidateWindowRights]
#exceptions
11/12/14 12:42:32.598 PM com.apple.xpc.launchd[1]: (com.apple.preference.printfax.remoteservice[3099]) Service exited due to signal: Killed: 9
11/12/14 12:42:37.845 PM com.apple.appkit.xpc.openAndSavePanelService[3298]: assertion failed: 14A389: libxpc.dylib + 97940 [9437C02E-A07B-38C8-91CB-299FAA63083D]: 0x89
11/12/14 12:42:37.879 PM sandboxd[246]: ([3295]) Pages(3295) deny file-read-data /Users/Mom/Downloads/CONM 136 - Study Guide Test 3 (2014).docx
11/12/14 12:42:37.971 PM sandboxd[246]: ([3295]) Pages(3295) deny file-read-data /Users/Mom/Downloads/12%20angry%20men.docx
11/12/14 12:42:38.033 PM sandboxd[246]: ([3295]) Pages(3295) deny file-read-data /Users/Mom/Downloads/Quiz%20#2.docx
11/12/14 12:42:39.712 PM sandboxd[246]: ([3295]) Pages(3295) deny file-read-data /Users/Mom/Downloads/12%20angry%20men.docx
11/12/14 12:42:39.000 PM kernel[0]: Sandbox: Pages(3295) deny file-read-data /Users/Mom/Downloads/Quiz%20#2.docx
11/12/14 12:42:39.748 PM sandboxd[246]: ([3295]) Pages(3295) deny file-read-data /Users/Mom/Downloads/12%20angry%20men.docx
11/12/14 12:42:39.850 PM sandboxd[246]: ([3295]) Pages(3295) deny file-read-data /Users/Mom/Downloads/12%20angry%20men.docx
11/12/14 12:42:39.874 PM sandboxd[246]: ([3295]) Pages(3295) deny file-read-data /Users/Mom/Downloads/Quiz%20#2.docx
11/12/14 12:42:39.945 PM sandboxd[246]: ([3295]) Pages(3295) deny file-read-data /Users/Mom/Downloads/Quiz%20#2.docx
11/12/14 12:42:41.000 PM kernel[0]: Sandbox: storeuid(3051) deny mach-lookup com.apple.dock.server
11/12/14 12:42:48.968 PM WindowServer[132]: WSGetSurfaceInWindow : Invalid surface 702894100 for window 1378
11/12/14 12:42:48.968 PM WindowServer[132]: WSGetSurfaceInWindow : Invalid surface 702894100 for window 1378
11/12/14 12:43:17.000 PM kernel[0]: Sandbox: storeuid(3051) deny mach-lookup com.apple.dock.server
11/12/14 12:43:23.225 PM com.apple.xpc.launchd[1]: (com.apple.quicklook[3316]) Endpoint has been activated through legacy launch(3) APIs. Please switch to XPC or bootstrap_check_in(): com.apple.quicklook
11/12/14 12:43:24.355 PM WindowServer[132]: disable_update_timeout: UI updates were forcibly disabled by application "PrinterProxy" for over 1.00 seconds. Server has re-enabled them.
11/12/14 12:43:24.946 PM WindowServer[132]: common_reenable_update: UI updates were finally reenabled by application "PrinterProxy" after 1.59 seconds (server forcibly re-enabled them after 1.00 seconds)
11/12/14 12:43:34.503 PM mdworker[3319]: code validation failed in the process of getting signing information: Error Domain=NSOSStatusErrorDomain Code=-67062 "The operation couldn’t be completed. (OSStatus error -67062.)"
11/12/14 12:43:34.514 PM mdworker[3320]: code validation failed in the process of getting signing information: Error Domain=NSOSStatusErrorDomain Code=-67062 "The operation couldn’t be completed. (OSStatus error -67062.)"
11/12/14 12:43:34.604 PM mdworker[3318]: code validation failed in the process of getting signing information: Error Domain=NSOSStatusErrorDomain Code=-67062 "The operation couldn’t be completed. (OSStatus error -67062.)"
11/12/14 12:44:38.628 PM sandboxd[246]: ([3051]) storeuid(3051) deny mach-lookup com.apple.dock.server
11/12/14 12:44:41.384 PM locationd[63]: Location icon should now be in state 'Active'
11/12/14 12:44:44.653 PM sandboxd[246]: ([375]) com.apple.metada(375) deny mach-lookup com.apple.cfnetwork.cfnetworkagent
11/12/14 12:44:44.793 PM sandboxd[246]: ([375]) com.apple.metada(375) deny mach-lookup com.apple.cfnetwork.cfnetworkagent
11/12/14 12:44:45.149 PM sandboxd[246]: ([375]) com.apple.metada(375) deny mach-lookup com.apple.cfnetwork.cfnetworkagent
11/12/14 12:44:56.515 PM Console[3327]: Failed to connect (_consoleX) outlet from (NSApplication) to (ConsoleX): missing setter or instance variable
11/12/14 12:44:56.641 PM locationd[63]: Location icon should now be in state 'Inactive'
11/12/14 12:44:56.834 PM sandboxd[246]: ([3051]) storeuid(3051) deny mach-lookup com.apple.dock.server
Not sure what any of this means. -
Cube content deletion is taking more time than usual.
Hi Experts,
We have a Process chain which ideally should run in every two hours. This chain has a delete data cube content step before the new data is loaded in the cube. This chain is running fine for one instance & the other instance is taking more time so it is quite intermittent.
In the process chain we are also deleting contents from the Dimension tables (in the delete content step). Need your inputs to improve the performance of this step.
Thanks & Regards
Mayank Tyagi.Hi Mayank ,
You can delete the indexes of the cube before deleting the contents of the cube . The concept is same as of data loading that data loads happens faster when indexes are deleted .
If you are having aggregates over this cube , then that aggregate will be also adjusted .
Kind Regards,
Ashutosh Singh -
Zfs destroy command takes more time than usual
Hi,
When I run the destroy command it takes more than usual.
I have exported the lun form this zfs volume ealier.
Later I have removed the lun view and deleted the lun.After that when I run the below command it takes more time (more than 5mins and still running)
#zfs destroy storage/luIs there a way to quickly destroy the filesystem.
It looks it removing the allocated files.
capacity operations bandwidth
pool alloc free read write read write
storage0 107G 116T 3.32K 2.52K 3.48M 37.7M
storage0 107G 116T 840 551 1.80M 6.01M
storage0 106G 116T 273 0 586K 0
storage0 106G 116T 1.19K 0 2.61M 0
storage0 106G 116T 1.47K 0 3.20M -
A sql suddenly spent more time than before
9/28号,业务部门发现一条sql运行时间突然增长,原来只需1秒,现在要10秒左右。
database 10.2.0.4 rac 2node
server aix 5.3
语句如下:
update J_ORGANIZATION c
set c.MODIFY_TIME = sysdate, c.MODIFY_EMPL_ID = 1111
where 1 = 1
and c.ORG_ID ='BJ0000270551'
and c.DEAL_STATUS = '1' ;
检查执行计划,发现没有显著影响性能
Rows Execution Plan
0 UPDATE STATEMENT MODE: ALL_ROWS
1 UPDATE OF 'J_ORGANIZATION'
1 TABLE ACCESS MODE: ANALYZED (BY GLOBAL INDEX ROWID) OF
'J_ORGANIZATION' (TABLE) PARTITION:ROW LOCATION
1 INDEX MODE: ANALYZED (UNIQUE SCAN) OF 'UNI_ORG_ID' (INDEX
(UNIQUE))
做了10046在trace中发现多了条sql
UPDATE BDP_ZQINFO.J_OUT_ORG_FOR_TRS B SET OPER_TYPE = 0, CREATE_TIME = SYSDATE
WHERE
EXISTS (SELECT 1 FROM BDP_ZQINFO.J_OUT_ORG_FOR_TRS A WHERE A.ORG_SERIAL_ID =
B.ORG_SERIAL_ID AND B.ORG_SERIAL_ID = :B1 );
这个是触发器生成,触发器语句如下:
CREATE OR REPLACE TRIGGER T_UPD_J_ORGANIZATION
before update of modify_time -- synchro_status
on J_ORGANIZATION
for each row
declare
org_serialid number:=:old.org_serial_id;
org_prov varchar2(8):=:old.prov_region_code;
begin
-- dbms_output.put_line('bbbbbbbbbbbbbbbbbbb');
/* IF
:old.synchro_status=1 and :new.synchro_status=0 OR
:old.synchro_status is null and :new.synchro_status=0
THEN*/
sp_j_out_org_for_trs(org_serialid,org_prov);
EXCEPTION
WHEN OTHERS THEN
-- Consider logging the error and then re-raise
RAISE;
-- END IF;
end T_UPD_J_ORGANIZATION;
这个update应是最先运行,J_OUT_ORG_FOR_TRS表为list分区(字段为PROV_REGION_CODE) 约有4百多万行,
其中字段ORG_SERIAL_ID索引J_OUT_ORG_FOR_TRS_ID为 global normal index,索引信息如下:
OWNER INDEX_NAME INDEX_TYPE TABLE_OWNER TABLE_NAME BLEVEL LEAF_BLOCKS DISTINCT_KEYS AVG_LEAF_BLOCKS_PER_KEY AVG_DATA_BLOCKS_PER_KEY CLUSTERING_FACTOR STATUS NUM_ROWS SAMPLE_SIZE LAST_ANALYZED
BDP_ZQINFO IDX_ORGANIZATION_1 NORMAL BDP_ZQINFO J_ORGANIZATION 3 154,241 18,181,606 1 1 17,794,548 VALID 18,307,431 131,394 10-05-2012 23:34:12
BDP_ZQINFO J_OUT_ORG_FOR_TRS_ID NORMAL BDP_ZQINFO J_OUT_ORG_FOR_TRS 2 15,188 4,429,794 1 1 3,576,204 VALID 4,491,419 330,908 10-01-2012 22:24:58
此条sql执行计划如下:
WORKLOAD REPOSITORY SQL Report
Snapshot Period Summary
DB Name DB Id Instance Inst num Release RAC Host
BSTTEST 1834441837 bsttest1 1 10.2.0.4.0 YES olap1
Snap Id Snap Time Sessions Cursors/Session
Begin Snap: 35139 08-10月-12 13:00:32 231 1.7
End Snap: 35140 08-10月-12 14:00:49 249 1.7
Elapsed: 60.28 (mins)
DB Time: 212.30 (mins)
SQL Summary
SQL Id Elapsed Time (ms) Module Action SQL Text
av6s7vnuqkhh6 100,066 UPDATE J_OUT_ORG_FOR_TRS B SET OPER_TYPE = 0, CREATE_TIME = SYSDATE W...
Back to Top
SQL ID: av6s7vnuqkhh6
1st Capture and Last Capture Snap IDs refer to Snapshot IDs witin the snapshot range
UPDATE J_OUT_ORG_FOR_TRS B SET OPER_TYPE = 0, CREATE_TIME = SYSDATE WH...
# Plan Hash Value Total Elapsed Time(ms) Executions 1st Capture Snap ID Last Capture Snap ID
1 1602621420 100,066 14 35140 35140
Back to Top
Plan 1(PHV: 1602621420)
Plan Statistics
Execution Plan
Back to Top
Plan Statistics
% Total DB Time is the Elapsed Time of the SQL statement divided into the Total Database Time multiplied by 100
Stat Name Statement Total Per Execution % Snap Total
Elapsed Time (ms) 100,066 7,147.57 0.79
CPU Time (ms) 97,695 6,978.23 1.18
Executions 14
Buffer Gets 335,669 23,976.36 0.09
Disk Reads 39 2.79 0.00
Parse Calls 5 0.36 0.00
Rows 40 2.86
User I/O Wait Time (ms) 173
Cluster Wait Time (ms) 2,632
Application Wait Time (ms) 0
Concurrency Wait Time (ms) 22
Invalidations 0
Version Count 5
Sharable Mem(KB) 27
Back to Plan 1(PHV: 1602621420)
Back to Top
Execution Plan
Id Operation Name Rows Bytes Cost (%CPU) Time Pstart Pstop
0 UPDATE STATEMENT 11M(100)
1 UPDATE J_OUT_ORG_FOR_TRS
2 FILTER
3 PARTITION LIST ALL 3843K 65M 4486 (3) 00:00:54 1 32
4 TABLE ACCESS FULL J_OUT_ORG_FOR_TRS 3843K 65M 4486 (3) 00:00:54 1 32
5 FILTER
6 INDEX RANGE SCAN J_OUT_ORG_FOR_TRS_ID 1 7 3 (0) 00:00:01
Back to Plan 1(PHV: 1602621420)
Back to Top
其中走了全表,花费大量时间,问如何优化?
下面是详细的trace文件:
/oracle/orabase/admin/bsttest/udump/bsttest1_ora_1421390.trc
Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
With the Partitioning, Real Application Clusters, OLAP, Data Mining
and Real Application Testing options
ORACLE_HOME = /oracle/orabase/product/10g/olap
System name: AIX
Node name: olap1
Release: 3
Version: 5
Machine: 00C6A8824C00
Instance name: bsttest1
Redo thread mounted by this instance: 1
Oracle process number: 156
Unix process pid: 1421390, image: oracle@olap1 (TNS V1-V3)
*** ACTION NAME:() 2012-10-08 13:25:14.174
*** MODULE NAME:(SQL*Plus) 2012-10-08 13:25:14.174
*** SERVICE NAME:(SYS$USERS) 2012-10-08 13:25:14.174
*** SESSION ID:(819.44232) 2012-10-08 13:25:14.174
WAIT #2: nam='SQL*Net message to client' ela= 2 driver id=1650815232 #bytes=1 p3=0 obj#=-1 tim=50576422823402
*** 2012-10-08 13:25:24.664
WAIT #2: nam='SQL*Net message from client' ela= 10235104 driver id=1650815232 #bytes=1 p3=0 obj#=-1 tim=50576433067587
WAIT #1: nam='library cache lock' ela= 277 handle address=504403169219121680 lock address=504403168553346632 100*mode+namespace=301 obj#=-1 tim=50576433068658
=====================
PARSING IN CURSOR #2 len=227 dep=1 uid=0 oct=3 lid=0 tim=50576433069481 hv=2190775527 ad='8ad0f328'
select u.name,o.name, t.update$, t.insert$, t.delete$, t.enabled from obj$ o,user$ u,trigger$ t where t.baseobject=:1 and t.obj#=o.obj# and o.owner#=u.user# and bitand(property,16)=0 and bitand(property,8)=0 order by o.obj#
END OF STMT
PARSE #2:c=0,e=37,p=0,cr=0,cu=0,mis=0,r=0,dep=1,og=4,tim=50576433069477
BINDS #2:
kkscoacd
Bind#0
oacdty=02 mxl=22(22) mxlc=00 mal=00 scl=00 pre=00
oacflg=00 fl2=0001 frm=00 csi=00 siz=24 off=0
kxsbbbfp=1104edb68 bln=22 avl=04 flg=05
value=182495
EXEC #2:c=0,e=146,p=0,cr=0,cu=0,mis=0,r=0,dep=1,og=4,tim=50576433069691
FETCH #2:c=0,e=328,p=0,cr=15,cu=0,mis=0,r=1,dep=1,og=4,tim=50576433070035
FETCH #2:c=0,e=4,p=0,cr=0,cu=0,mis=0,r=1,dep=1,og=4,tim=50576433070074
FETCH #2:c=0,e=12,p=0,cr=0,cu=0,mis=0,r=0,dep=1,og=4,tim=50576433070107
STAT #2 id=1 cnt=2 pid=0 pos=1 obj=0 op='SORT ORDER BY (cr=15 pr=0 pw=0 time=352 us)'
STAT #2 id=2 cnt=2 pid=1 pos=1 obj=0 op='NESTED LOOPS (cr=15 pr=0 pw=0 time=294 us)'
STAT #2 id=3 cnt=2 pid=2 pos=1 obj=0 op='NESTED LOOPS (cr=11 pr=0 pw=0 time=248 us)'
STAT #2 id=4 cnt=2 pid=3 pos=1 obj=81 op='TABLE ACCESS BY INDEX ROWID TRIGGER$ (cr=3 pr=0 pw=0 time=169 us)'
STAT #2 id=5 cnt=2 pid=4 pos=1 obj=125 op='INDEX RANGE SCAN I_TRIGGER1 (cr=1 pr=0 pw=0 time=116 us)'
STAT #2 id=6 cnt=2 pid=3 pos=2 obj=18 op='TABLE ACCESS BY INDEX ROWID OBJ$ (cr=8 pr=0 pw=0 time=79 us)'
STAT #2 id=7 cnt=2 pid=6 pos=1 obj=36 op='INDEX UNIQUE SCAN I_OBJ1 (cr=6 pr=0 pw=0 time=29 us)'
STAT #2 id=8 cnt=2 pid=2 pos=2 obj=22 op='TABLE ACCESS CLUSTER USER$ (cr=4 pr=0 pw=0 time=39 us)'
STAT #2 id=9 cnt=2 pid=8 pos=1 obj=11 op='INDEX UNIQUE SCAN I_USER# (cr=2 pr=0 pw=0 time=13 us)'
=====================
PARSING IN CURSOR #1 len=155 dep=0 uid=5 oct=6 lid=5 tim=50576433073227 hv=500046959 ad='8de25730'
update BDP_ZQINFO.J_ORGANIZATION c set c.MODIFY_TIME = sysdate, c.MODIFY_EMPL_ID = 1111 where 1 = 1 and c.ORG_ID ='BJ0000270551' and c.DEAL_STATUS = '1'
END OF STMT
PARSE #1:c=0,e=5500,p=0,cr=15,cu=0,mis=1,r=0,dep=0,og=1,tim=50576433073224
BINDS #1:
WAIT #1: nam='db file sequential read' ela= 189 file#=179 block#=277463 blocks=1 obj#=183635 tim=50576433073777
WAIT #2: nam='library cache lock' ela= 301 handle address=504403169246502016 lock address=504403168550871736 100*mode+namespace=301 obj#=183635 tim=50576433074697
=====================
PARSING IN CURSOR #2 len=183 dep=1 uid=138 oct=6 lid=138 tim=50576433074784 hv=896090630 ad='1e6739b8'
UPDATE J_OUT_ORG_FOR_TRS B SET OPER_TYPE = 0, CREATE_TIME = SYSDATE WHERE EXISTS (SELECT 1 FROM J_OUT_ORG_FOR_TRS A WHERE A.ORG_SERIAL_ID = B.ORG_SERIAL_ID AND B.ORG_SERIAL_ID = :B1 )
END OF STMT
PARSE #2:c=0,e=511,p=0,cr=0,cu=0,mis=0,r=0,dep=1,og=1,tim=50576433074781
BINDS #2:
kkscoacd
Bind#0
oacdty=02 mxl=22(21) mxlc=00 mal=00 scl=00 pre=00
oacflg=13 fl2=206001 frm=00 csi=00 siz=24 off=0
kxsbbbfp=110562a00 bln=22 avl=06 flg=09
value=-23145297
WAIT #2: nam='gc current grant busy' ela= 559 p1=442 p2=296026 p3=33619969 obj#=184185 tim=50576433356596
WAIT #2: nam='gc current block 2-way' ela= 321 p1=30 p2=90709 p3=16777217 obj#=184219 tim=50576433357396
WAIT #2: nam='gc current block 2-way' ela= 514 p1=122 p2=50469 p3=33554433 obj#=184219 tim=50576433358061
WAIT #2: nam='gc current block 2-way' ela= 521 p1=122 p2=50461 p3=33554433 obj#=184219 tim=50576433358753
WAIT #2: nam='gc cr block 2-way' ela= 466 p1=189 p2=73366 p3=1 obj#=184187 tim=50576434468561
WAIT #2: nam='gc current block 2-way' ela= 613 p1=317 p2=239612 p3=1 obj#=184190 tim=50576437055206
WAIT #2: nam='gc cr block 2-way' ela= 654 p1=319 p2=228768 p3=1 obj#=184190 tim=50576437392221
WAIT #2: nam='gc cr block 2-way' ela= 333 p1=319 p2=228994 p3=1 obj#=184190 tim=50576437697191
WAIT #2: nam='gc current block 2-way' ela= 361 p1=361 p2=270819 p3=1 obj#=184190 tim=50576437997483
WAIT #2: nam='gc cr block busy' ela= 912 p1=363 p2=459877 p3=1 obj#=184191 tim=50576438627331
WAIT #2: nam='gc cr block 2-way' ela= 358 p1=87 p2=230733 p3=1 obj#=184192 tim=50576439432074
WAIT #2: nam='gc cr block 2-way' ela= 385 p1=366 p2=121340 p3=1 obj#=184199 tim=50576441543979
WAIT #2: nam='gc cr block 2-way' ela= 428 p1=366 p2=121393 p3=1 obj#=184199 tim=50576441854612
WAIT #2: nam='gc cr block 2-way' ela= 377 p1=366 p2=121412 p3=1 obj#=184199 tim=50576441864739
WAIT #2: nam='gc current block 2-way' ela= 408 p1=366 p2=121509 p3=1 obj#=184199 tim=50576442172533
WAIT #2: nam='gc cr block 2-way' ela= 392 p1=366 p2=121653 p3=1 obj#=184199 tim=50576442467808
*** 2012-10-08 13:25:37.010
EXEC #2:c=11740000,e=12049484,p=0,cr=23912,cu=7,mis=0,r=1,dep=1,og=1,tim=50576445124325
EXEC #1:c=11750000,e=12051323,p=1,cr=23916,cu=10,mis=0,r=1,dep=0,og=1,tim=50576445124609
WAIT #1: nam='SQL*Net message to client' ela= 3 driver id=1650815232 #bytes=1 p3=0 obj#=184199 tim=50576445124710
*** 2012-10-08 13:26:08.750
WAIT #1: nam='SQL*Net message from client' ela= 30995944 driver id=1650815232 #bytes=1 p3=0 obj#=184199 tim=50576476120696
STAT #1 id=1 cnt=1 pid=0 pos=1 obj=0 op='UPDATE J_ORGANIZATION (cr=23916 pr=1 pw=0 time=12051201 us)'
STAT #1 id=2 cnt=1 pid=1 pos=1 obj=182495 op='TABLE ACCESS BY GLOBAL INDEX ROWID J_ORGANIZATION PARTITION: ROW LOCATION ROW LOCATION (cr=4 pr=1 pw=0 time=428 us)'
STAT #1 id=3 cnt=1 pid=2 pos=1 obj=183635 op='INDEX UNIQUE SCAN UNI_ORG_ID (cr=3 pr=1 pw=0 time=409 us)'
=====================
PARSING IN CURSOR #1 len=56 dep=0 uid=5 oct=42 lid=5 tim=50576476121198 hv=1729844458 ad='0'
alter session set events '10046 trace name context off'
END OF STMT
PARSE #1:c=0,e=173,p=0,cr=0,cu=0,mis=0,r=0,dep=0,og=0,tim=50576476121194
EXEC #1:c=0,e=120,p=0,cr=0,cu=0,mis=0,r=0,dep=0,og=0,tim=50576476121370Edited by: 小电工II on 2012-10-7 下午11:43此为
UPDATE BDP_ZQINFO.J_OUT_ORG_FOR_TRS B SET OPER_TYPE = 0, CREATE_TIME = SYSDATE
WHERE
EXISTS (SELECT 1 FROM BDP_ZQINFO.J_OUT_ORG_FOR_TRS A WHERE A.ORG_SERIAL_ID =
B.ORG_SERIAL_ID AND B.ORG_SERIAL_ID = :B1 );
生成的
Enter value for sqlid: av6s7vnuqkhh6
old 21: ('&SQLID') order by s.snap_id
new 21: ('av6s7vnuqkhh6') order by s.snap_id
SnapId PLAN_HASH_VALUE Date time No. of exec LIO/exec CPUTIM/exec ETIME/exec PIO/exec ROWs/exec
34967 1602621420 10/01/12_0800_0900 26 23159.35 5.66 5.83 12.27 260.88
34968 1602621420 10/01/12_0900_1000 70 23162.23 5.63 5.70 1.20 97.33
34969 1602621420 10/01/12_1000_1100 58 23162.43 5.72 5.84 .91 117.95
34970 1602621420 10/01/12_1100_1200 62 23161.10 5.84 5.97 .92 110.71
34971 1602621420 10/01/12_1200_1300 5 23164.40 5.52 5.59 2.40 .60
34974 1602621420 10/01/12_1500_1600 6 23157.17 5.60 5.60 .00 .00
34991 1602621420 10/02/12_0800_0900 11 23166.36 5.50 5.54 1.00 625.00
34992 1602621420 10/02/12_0900_1000 19 23158.05 5.39 5.39 .47 362.00
34994 1602621420 10/02/12_1100_1200 1 23161.00 5.77 5.77 .00 .00
34998 1602621420 10/02/12_1500_1600 7 23253.00 5.70 5.70 1.57 7.71
34999 1602621420 10/02/12_1600_1700 9 23159.22 5.61 5.71 .33 6.22
35002 1602621420 10/02/12_1900_2000 1 23169.00 5.93 6.16 1.00 56.00
35016 1602621420 10/03/12_0900_1000 6 23188.83 5.79 5.96 2.17 3.50
35017 1602621420 10/03/12_1000_1100 17 23175.59 5.58 5.58 1.41 3.06
35018 1602621420 10/03/12_1100_1200 9 23165.89 5.60 5.60 .89 6.56
35019 1602621420 10/03/12_1200_1300 17 23170.24 5.66 5.67 1.59 413.29
35020 1602621420 10/03/12_1300_1400 116 23160.33 5.61 5.61 .47 60.44
35021 1602621420 10/03/12_1400_1500 107 23161.92 5.61 5.61 .83 65.96
35022 1602621420 10/03/12_1500_1600 51 23163.78 5.65 5.65 1.69 138.96
35023 1602621420 10/03/12_1600_1700 59 23161.88 5.62 5.62 1.12 120.61
35024 1602621420 10/03/12_1700_1800 45 23165.49 5.63 5.63 1.80 158.91
35025 1602621420 10/03/12_1800_1900 21 23161.24 5.60 5.60 1.00 340.95
35040 1602621420 10/04/12_0900_1000 165 23168.90 5.67 5.68 .79 .96
35041 1602621420 10/04/12_1000_1100 387 23160.60 5.62 5.63 .59 .73
35042 1602621420 10/04/12_1100_1200 79 23166.33 5.72 5.74 1.32 4.43
35045 1602621420 10/04/12_1400_1500 25 23170.24 5.63 5.63 1.48 1.24
35046 1602621420 10/04/12_1500_1600 79 23164.59 5.69 5.70 .70 1.20
35047 1602621420 10/04/12_1600_1700 125 23162.08 5.74 5.74 .42 1.31
35048 1602621420 10/04/12_1700_1800 77 23166.55 5.62 5.62 .62 3.10
35050 1602621420 10/04/12_1900_2000 1 23467.00 5.65 8.39 1213.00 7405.00
35054 1602621420 10/04/12_2300_0000 366 23166.97 5.50 5.51 11.54 20.58
35055 1602621420 10/05/12_0000_0100 634 23198.98 5.53 5.53 .31 12.88
35056 1602621420 10/05/12_0100_0200 619 23133.01 5.62 5.65 .26 14.19
35057 1602621420 10/05/12_0200_0300 616 23166.56 5.57 5.58 .42 15.28
35058 1602621420 10/05/12_0300_0400 634 23166.36 5.54 5.55 .38 15.85
35059 1602621420 10/05/12_0400_0500 629 23167.02 5.49 5.49 .62 17.04
35060 1602621420 10/05/12_0500_0600 637 23166.48 5.48 5.48 .52 17.82
35061 1602621420 10/05/12_0600_0700 626 23166.50 5.52 5.52 .52 19.14
35062 1602621420 10/05/12_0700_0800 643 23166.47 5.44 5.45 .47 19.63
35063 1602621420 10/05/12_0800_0900 721 23103.33 5.42 5.42 .64 18.40
35064 1602621420 10/05/12_0900_1000 1101 23145.79 5.54 5.59 .96 12.91
35065 1602621420 10/05/12_1000_1100 396 23163.76 5.55 5.59 1.00 36.46
35066 1602621420 10/05/12_1100_1200 153 23255.38 5.55 5.59 1.51 104.88
35067 1602621420 10/05/12_1200_1300 417 23160.47 5.43 5.45 .64 .31
35068 1602621420 10/05/12_1300_1400 395 23162.84 5.44 5.45 .63 .88
35069 1602621420 10/05/12_1400_1500 2 23161.50 5.42 5.64 .50 .50
35070 1602621420 10/05/12_1500_1600 10 23160.30 5.93 6.15 .60 .50
35071 1602621420 10/05/12_1600_1700 18 23181.94 5.72 5.74 2.11 892.39
35073 1602621420 10/05/12_1800_1900 3 23312.00 5.76 5.98 1.00 5370.33
35091 1602621420 10/06/12_1200_1300 68 23165.46 5.49 5.51 1.01 .82
35092 1602621420 10/06/12_1300_1400 164 23161.13 5.49 5.49 .43 .77
35093 1602621420 10/06/12_1400_1500 225 23159.11 5.52 5.54 .32 .82
35094 1602621420 10/06/12_1500_1600 109 23162.06 5.52 5.56 .93 2.23
35095 1602621420 10/06/12_1600_1700 24 23160.13 5.50 5.52 .58 10.46
35096 1602621420 10/06/12_1700_1800 56 23167.84 5.49 5.50 .89 5.68
35097 1602621420 10/06/12_1800_1900 71 23165.11 5.49 5.50 .83 5.15
35098 1602621420 10/06/12_1900_2000 437 23166.72 5.55 5.56 1.04 1.83
35099 1602621420 10/06/12_2000_2100 624 23167.51 5.52 5.52 .89 2.36
35100 1602621420 10/06/12_2100_2200 642 23191.14 5.44 5.44 .84 5.75
35101 1602621420 10/06/12_2200_2300 640 23167.84 5.45 5.45 .58 6.89
35102 1602621420 10/06/12_2300_0000 637 23166.58 5.43 5.43 .48 7.92
35103 1602621420 10/07/12_0000_0100 643 23169.93 5.43 5.44 .42 9.17
35104 1602621420 10/07/12_0100_0200 627 23169.19 5.49 5.51 .53 10.66
35105 1602621420 10/07/12_0200_0300 640 23166.57 5.44 5.44 .32 11.44
35106 1602621420 10/07/12_0300_0400 641 23166.40 5.45 5.46 .26 12.42
35107 1602621420 10/07/12_0400_0500 582 23181.24 5.43 5.43 .34 16.28
35113 1602621420 10/07/12_1000_1100 14 23235.79 5.79 5.86 1.21 1157.21
35117 1602621420 10/07/12_1400_1500 56 23158.70 5.41 5.43 .30 288.66
35118 1602621420 10/07/12_1500_1600 55 23161.78 5.43 5.45 .87 294.40
35119 1602621420 10/07/12_1600_1700 55 23157.67 5.45 5.47 .33 294.53
35120 1602621420 10/07/12_1700_1800 51 23178.39 5.40 5.40 1.00 319.88
35121 1602621420 10/07/12_1800_1900 7 23161.57 5.40 5.43 1.71 2331.00
35122 1602621420 10/07/12_1900_2000 12 23303.58 5.38 5.38 1.58 1375.08
35136 1602621420 10/08/12_0900_1000 197 23371.71 5.64 5.89 .79 2.98
35137 1602621420 10/08/12_1000_1100 96 23930.24 6.01 6.14 .93 6.52
35138 1602621420 10/08/12_1100_1200 81 23923.36 5.90 5.93 .94 8.17
35139 1602621420 10/08/12_1200_1300 41 23929.37 5.84 5.85 1.07 1.15
35140 1602621420 10/08/12_1300_1400 14 23976.36 6.98 7.15 2.79 1182.71
35141 1602621420 10/08/12_1400_1500 109 38385.74 6.24 6.34 .67 .56
79 rows selected. -
Why SQL2 took much more time than SQL1?
I run these 2 SQLs sequencely.
--- SQL1: It took 245 seconds.
create table PORTAL_DAYLOG_100118_bak
as
select * from PORTAL_DAYLOG_100118;
--- SQL2: It took 3105 seconds.
create table PORTAL_DAYLOG_100121_bak
as
select * from PORTAL_DAYLOG_100121;
It is really strange that SQL2 took almost 13 times than SQL1, with nearly same data amount and same data structure in the same tablespace.
Could anyone tell me the reason? or How could I find out why?
Here is more detail info. for my case,
--- Server:
[@wapbi.no.sohu.com ~]$ uname -a
Linux test 2.6.18-128.el5 #1 SMP Wed Dec 17 11:41:38 EST 2008 x86_64 x86_64 x86_64 GNU/Linux
--- DB
Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
--- Tablespace:
CREATE TABLESPACE PORTAL DATAFILE
'/data/oradata/wapbi/portal01.dbf' SIZE 19456M AUTOEXTEND ON NEXT 1024M MAXSIZE UNLIMITED,
'/data/oradata/wapbi/portal02.dbf' SIZE 17408M AUTOEXTEND ON NEXT 1024M MAXSIZE UNLIMITED
LOGGING
ONLINE
PERMANENT
EXTENT MANAGEMENT LOCAL AUTOALLOCATE
BLOCKSIZE 8K
SEGMENT SPACE MANAGEMENT AUTO
FLASHBACK ON;
--- Tables:
SQL> select table_name,num_rows,blocks,avg_row_len from dba_tables
2 where table_name in ('PORTAL_DAYLOG_100118','PORTAL_DAYLOG_100121');
TABLE_NAME NUM_ROWS BLOCKS AVG_ROW_LEN
PORTAL_DAYLOG_100118 20808536 269760 85
PORTAL_DAYLOG_100121 33747911 440512 86
CREATE TABLE PORTAL_DAYLOG_100118
IP VARCHAR2(20 BYTE),
NODEPATH VARCHAR2(50 BYTE),
PG VARCHAR2(20 BYTE),
PAGETYPE INTEGER,
CLK VARCHAR2(20 BYTE),
FR VARCHAR2(20 BYTE),
PHID INTEGER,
ANONYMOUSID VARCHAR2(50 BYTE),
USID VARCHAR2(50 BYTE),
PASSPORT VARCHAR2(200 BYTE),
M_TIME CHAR(4 BYTE) NOT NULL,
M_DATE CHAR(6 BYTE) NOT NULL,
LOGDATE DATE
LOGGING
NOCOMPRESS
NOCACHE
NOPARALLEL
MONITORING;
CREATE TABLE PORTAL_DAYLOG_100121
IP VARCHAR2(20 BYTE),
NODEPATH VARCHAR2(50 BYTE),
PG VARCHAR2(20 BYTE),
PAGETYPE INTEGER,
CLK VARCHAR2(20 BYTE),
FR VARCHAR2(20 BYTE),
PHID INTEGER,
ANONYMOUSID VARCHAR2(50 BYTE),
USID VARCHAR2(50 BYTE),
PASSPORT VARCHAR2(200 BYTE),
M_TIME CHAR(4 BYTE) NOT NULL,
M_DATE CHAR(6 BYTE) NOT NULL,
LOGDATE DATE
LOGGING
NOCOMPRESS
NOCACHE
NOPARALLEL
MONITORING;Any comment will be really appeciated!!!
SatineHey Anurag,
Thank you for your help!
Here it is.
SQL1:
create table portal.PORTAL_DAYLOG_100118_TEST
as
select * from portal.PORTAL_DAYLOG_100118
call count cpu elapsed disk query current rows
Parse 1 0.00 0.00 0 0 0 0
Execute 1 374.69 519.05 264982 265815 274858 20808536
Fetch 0 0.00 0.00 0 0 0 0
total 2 374.69 519.05 264982 265815 274858 20808536
Misses in library cache during parse: 1
Optimizer mode: ALL_ROWS
Parsing user id: SYS
Rows Row Source Operation
0 LOAD AS SELECT (cr=268138 pr=264982 pw=264413 time=0 us)
20808536 TABLE ACCESS FULL PORTAL_DAYLOG_100118 (cr=265175 pr=264981 pw=0 time=45792172 us cost=73478 size=1768725560 card=20808536)SQL2:
create table portal.PORTAL_DAYLOG_100121_TEST
as
select * from portal.PORTAL_DAYLOG_100121
call count cpu elapsed disk query current rows
Parse 1 0.00 0.00 0 0 0 0
Execute 1 1465.72 1753.35 290959 291904 300738 22753695
Fetch 0 0.00 0.00 0 0 0 0
total 2 1465.72 1753.35 290959 291904 300738 22753695
Misses in library cache during parse: 1
Optimizer mode: ALL_ROWS
Parsing user id: SYS
Rows Row Source Operation
0 LOAD AS SELECT (cr=295377 pr=290960 pw=289966 time=0 us)
22753695 TABLE ACCESS FULL PORTAL_DAYLOG_100121 (cr=291255 pr=290958 pw=0 time=56167952 us cost=80752 size=1956817770 card=22753695)Best wishes,
Satine
Maybe you are looking for
-
Design problem with dynamically added row in datatable
Hello all, I have a dataTable and have developped a method that dynamically adds rows to the dataTable. Each row is made up of two form controls. The problem I have now is that I don't know how to retrieve the values from the mutiple rows. Ideally I
-
Updating Iphone to IOS5.1 error "the network connection could not be establish"
Hi recently when I was in Boston I purchased the new Iphone 4S at apple store. and everything worked fine. Now I wanted to update my phone to the new IOS5.1 but I get this error message "The network connectiong could not be established". I get the sa
-
PO Print when mark a Delete option for a Line Item
Hi Friends, I am in a Fix, I have changed a PO line item by giving a header text and then Marked the line item as deleted, when I tried to take a PO Print It does not show me the Header text, Please help me out with problem.
-
Show Swing Container JPanel in ADF Component
Hi ADF-Folk, I come out of the forms developer world. in forms i used pluggable java components to show graphical components in forms. Is it possible to use a show a Swing JPanel Class in a ADF Component like a PanelBox or something else? Application
-
1.0.10 Still caching in AppData
So "thank" you for reimplementing a different cache location (why was this removed to begin with?)However for me - version 1.0.10 still caches down into AppData eating up my system-drive.So right now it just looks to use twice as much cache - or spli