Need to tune my index
I have a table with is created as below
CREATE UNIQUE INDEX schema.table1 ON table1
(Name,Roll No, Class, age)
With all 4 colums as primary key in listed order.
Now Roll No is the most queried column and query becomes very slow as there is a full table scan always .
Can u let me know . how can I tune this index to reduce time on query execution
Impatience is an unpleasant trait. Expecting people to jump to your requirements when you don't seem to have made any effort yourself is unlikely to encourage people to help you.
You haven't even bothered to research the correct forum to post to. As the title indicates this forum is for the SQL Developer tool. General SQL questions should go in the SQL and PL/SQL forum.
When you repost your question there, provide more information including your database version and what options you have tried.
Similar Messages
-
Hi,
I'm using below query inside the procedure.I need to tune this query as it is taking more time for execution.Could you some help on this
*version.*
10.2.0
Query
INSERT INTO TEMP_TEST_STAT(
SELECT S.STUDENT_ID_PK, A.RESPONSE_KEY GENDER, B.RESPONSE_KEY ETHNICITY, T.SCORE SCALE_SCORE ,D.TEST_NAME
FROM TEST_SESSION_DETAIL T, TEST_SESSION S, STUDENT U, TEST_DETAIL D,
SELECT I.ITEM_ID, I.UIN, BG.BG_SESSION_DETAIL_ID, BG.RESPONSE_KEY , B.TEST_SESSION_ID, C.STUDENT_ID_PK
FROM BG_QUESTION_USER_RESPONSE BG, ITEM I, BG_SESSION_DETAIL B, TEST_SESSION C
WHERE
BG.ITEM_ID = I.ITEM_ID AND I.UIN = 'SBG002' AND
BG.BG_SESSION_DETAIL_ID = B.BG_SESSION_DETAIL_ID
AND B.TEST_SESSION_ID = C.TEST_SESSION_ID ) A,
SELECT I.ITEM_ID, I.UIN, BG.BG_SESSION_DETAIL_ID, BG.RESPONSE_KEY , B.TEST_SESSION_ID, C.STUDENT_ID_PK
FROM BG_QUESTION_USER_RESPONSE BG, ITEM I, BG_SESSION_DETAIL B, TEST_SESSION C
WHERE
BG.ITEM_ID = I.ITEM_ID AND I.UIN = 'SBG001' AND
BG.BG_SESSION_DETAIL_ID = B.BG_SESSION_DETAIL_ID
AND B.TEST_SESSION_ID = C.TEST_SESSION_ID ) B
,INSTITUTION E
WHERE
D.TEST_NAME = 'Reading Comprehension' AND
T.TEST_SESSION_STATUS_ID = 3
AND TRUNC(T.TEST_END_DATE) BETWEEN TO_DATE('01-09-09','DD-MON-YY') AND TO_DATE('30-09-09','DD-MON-YY')
AND T.TEST_DETAIL_ID = D.TEST_DETAIL_ID
AND S.STUDENT_ID_PK = U.STUDENT_ID_PK
AND S.TEST_SESSION_ID = T.TEST_SESSION_ID
AND A.TEST_SESSION_ID = T.TEST_SESSION_ID
AND A.STUDENT_ID_PK = S.STUDENT_ID_PK
AND B.TEST_SESSION_ID = T.TEST_SESSION_ID
AND B.STUDENT_ID_PK = S.STUDENT_ID_PK
AND S.INSTITUTION_ID=E.INSTITUTION_ID
AND SUBSTR(E.INSTITUTION_ID_DISPLAY,8,3) <> '000');
Explain plan
PLAN_TABLE_OUTPUT
Plan hash value: 790379856
| Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time | Pstart| Pstop |
| 0 | INSERT STATEMENT | | 1 | 204 | 1567 (1)| 00:00:19 | | |
|* 1 | FILTER | | | | | | | |
| 2 | NESTED LOOPS | | 1 | 204 | 1567 (1)| 00:00:19 | | |
| 3 | NESTED LOOPS | | 1 | 189 | 1566 (1)| 00:00:19 | | |
| 4 | NESTED LOOPS | | 1 | 167 | 1561 (1)| 00:00:19 | | |
| 5 | NESTED LOOPS | | 1 | 155 | 1560 (1)| 00:00:19 | | |
| 6 | NESTED LOOPS | | 1 | 143 | 1556 (1)| 00:00:19 | | |
| 7 | NESTED LOOPS | | 1 | 128 | 1555 (1)| 00:00:19 | | |
| 8 | NESTED LOOPS | | 1 | 106 | 1550 (1)| 00:00:19 | | |
| 9 | NESTED LOOPS | | 186 | 17484 | 1364 (1)| 00:00:17 | | |
|* 10 | HASH JOIN | | 86 | 7052 | 1020 (1)| 00:00:13 | | |
| 11 | NESTED LOOPS | | 86 | 5848 | 1002 (1)| 00:00:13 | | |
|* 12 | HASH JOIN | | 86 | 4472 | 916 (1)| 00:00:11 | | |
|* 13 | INDEX RANGE SCAN | TEST_DETAIL_AK_1 | 1 | 21 | 2 (0)| 00:00:01 | | |
|* 14 | TABLE ACCESS BY GLOBAL INDEX ROWID| TEST_SESSION_DETAIL | 8549 | 258K| 914 (1)| 00:00:11 | ROWID | ROWID |
|* 15 | INDEX RANGE SCAN | TEST_END_DATE_IDX | 15672 | | 53 (0)| 00:00:01 | | |
| 16 | PARTITION HASH ITERATOR | | 1 | 16 | 1 (0)| 00:00:01 | KEY | KEY |
| 17 | TABLE ACCESS BY LOCAL INDEX ROWID | TEST_SESSION | 1 | 16 | 1 (0)| 00:00:01 | KEY | KEY |
|* 18 | INDEX UNIQUE SCAN | PK_TEST_SESSION | 1 | | 0 (0)| 00:00:01 | KEY | KEY |
|* 19 | INDEX FAST FULL SCAN | INSTI_ID_NAME_COUN_DISP_IDX | 5930 | 83020 | 17 (0)| 00:00:01 | | |
| 20 | TABLE ACCESS BY GLOBAL INDEX ROWID | BG_SESSION_DETAIL | 2 | 24 | 4 (0)| 00:00:01 | ROWID | ROWID |
|* 21 | INDEX RANGE SCAN | BG_SESS_DETAIL_TEST_SE_FK1_I | 2 | | 2 (0)| 00:00:01 | | |
| 22 | PARTITION HASH ITERATOR | | 1 | 12 | 1 (0)| 00:00:01 | KEY | KEY |
|* 23 | TABLE ACCESS BY LOCAL INDEX ROWID | TEST_SESSION | 1 | 12 | 1 (0)| 00:00:01 | KEY | KEY |
|* 24 | INDEX UNIQUE SCAN | PK_TEST_SESSION | 1 | | 0 (0)| 00:00:01 | KEY | KEY |
| 25 | TABLE ACCESS BY GLOBAL INDEX ROWID | BG_QUESTION_USER_RESPONSE | 4 | 88 | 5 (0)| 00:00:01 | ROWID | ROWID |
|* 26 | INDEX RANGE SCAN | BG_QUESTION_USER_RESP_IDX1 | 4 | | 2 (0)| 00:00:01 | | |
| 27 | PARTITION HASH ITERATOR | | 1 | 15 | 1 (0)| 00:00:01 | KEY | KEY |
|* 28 | TABLE ACCESS BY LOCAL INDEX ROWID | ITEM | 1 | 15 | 1 (0)| 00:00:01 | KEY | KEY |
|* 29 | INDEX UNIQUE SCAN | PK_ITEM | 1 | | 0 (0)| 00:00:01 | KEY | KEY |
| 30 | TABLE ACCESS BY GLOBAL INDEX ROWID | BG_SESSION_DETAIL | 2 | 24 | 4 (0)| 00:00:01 | ROWID | ROWID |
|* 31 | INDEX RANGE SCAN | BG_SESS_DETAIL_TEST_SE_FK1_I | 2 | | 2 (0)| 00:00:01 | | |
| 32 | PARTITION HASH ITERATOR | | 1 | 12 | 1 (0)| 00:00:01 | KEY | KEY |
|* 33 | TABLE ACCESS BY LOCAL INDEX ROWID | TEST_SESSION | 1 | 12 | 1 (0)| 00:00:01 | KEY | KEY |
|* 34 | INDEX UNIQUE SCAN | PK_TEST_SESSION | 1 | | 0 (0)| 00:00:01 | KEY | KEY |
| 35 | TABLE ACCESS BY GLOBAL INDEX ROWID | BG_QUESTION_USER_RESPONSE | 4 | 88 | 5 (0)| 00:00:01 | ROWID | ROWID |
|* 36 | INDEX RANGE SCAN | BG_QUESTION_USER_RESP_IDX1 | 4 | | 2 (0)| 00:00:01 | | |
| 37 | PARTITION HASH ITERATOR | | 1 | 15 | 1 (0)| 00:00:01 | KEY | KEY |
|* 38 | TABLE ACCESS BY LOCAL INDEX ROWID | ITEM | 1 | 15 | 1 (0)| 00:00:01 | KEY | KEY |
|* 39 | INDEX UNIQUE SCAN | PK_ITEM | 1 | | 0 (0)| 00:00:01 | KEY | KEY |
Predicate Information (identified by operation id):
1 - filter(TO_DATE('01-09-09','DD-MON-YY')<=TO_DATE('30-09-09','DD-MON-YY'))
10 - access("S"."INSTITUTION_ID"="E"."INSTITUTION_ID")
12 - access("T"."TEST_DETAIL_ID"="D"."TEST_DETAIL_ID")
13 - access("D"."TEST_NAME"='Reading Comprehension')
14 - filter("T"."TEST_SESSION_STATUS_ID"=3)
15 - access(TRUNC(INTERNAL_FUNCTION("TEST_END_DATE"))>=TO_DATE('01-09-09','DD-MON-YY') AND
TRUNC(INTERNAL_FUNCTION("TEST_END_DATE"))<=TO_DATE('30-09-09','DD-MON-YY'))
18 - access("S"."TEST_SESSION_ID"="T"."TEST_SESSION_ID")
19 - filter(SUBSTR("E"."INSTITUTION_ID_DISPLAY",8,3)<>'000')
21 - access("B"."TEST_SESSION_ID"="T"."TEST_SESSION_ID")
23 - filter("C"."STUDENT_ID_PK"="S"."STUDENT_ID_PK")
24 - access("B"."TEST_SESSION_ID"="C"."TEST_SESSION_ID")
26 - access("BG"."BG_SESSION_DETAIL_ID"="B"."BG_SESSION_DETAIL_ID")
28 - filter("I"."UIN"='SBG001')
29 - access("BG"."ITEM_ID"="I"."ITEM_ID")
31 - access("B"."TEST_SESSION_ID"="T"."TEST_SESSION_ID")
33 - filter("C"."STUDENT_ID_PK"="S"."STUDENT_ID_PK")
34 - access("B"."TEST_SESSION_ID"="C"."TEST_SESSION_ID")
36 - access("BG"."BG_SESSION_DETAIL_ID"="B"."BG_SESSION_DETAIL_ID")
38 - filter("I"."UIN"='SBG002')
39 - access("BG"."ITEM_ID"="I"."ITEM_ID")Rob,
How i can my insert statement here.
I tried like..but error.
select /*+ gather_plan_statistics */ INSERT INTO TEMP_TEST_STAT(
SELECT S.STUDENT_ID_PK, A.RESPONSE_KEY GENDER, B.RESPONSE_KEY ETHNICITY, T.SCORE SCALE_SCORE ,D.TEST_NAME
FROM TEST_SESSION_DETAIL T, TEST_SESSION S, STUDENT U, TEST_DETAIL D,
SELECT I.ITEM_ID, I.UIN, BG.BG_SESSION_DETAIL_ID, BG.RESPONSE_KEY , B.TEST_SESSION_ID, C.STUDENT_ID_PK
FROM BG_QUESTION_USER_RESPONSE BG, ITEM I, BG_SESSION_DETAIL B, TEST_SESSION C
WHERE
BG.ITEM_ID = I.ITEM_ID AND I.UIN = 'SBG002' AND
BG.BG_SESSION_DETAIL_ID = B.BG_SESSION_DETAIL_ID
AND B.TEST_SESSION_ID = C.TEST_SESSION_ID ) A,
SELECT I.ITEM_ID, I.UIN, BG.BG_SESSION_DETAIL_ID, BG.RESPONSE_KEY , B.TEST_SESSION_ID, C.STUDENT_ID_PK
FROM BG_QUESTION_USER_RESPONSE BG, ITEM I, BG_SESSION_DETAIL B, TEST_SESSION C
WHERE
BG.ITEM_ID = I.ITEM_ID AND I.UIN = 'SBG001' AND
BG.BG_SESSION_DETAIL_ID = B.BG_SESSION_DETAIL_ID
AND B.TEST_SESSION_ID = C.TEST_SESSION_ID ) B
,INSTITUTION E
WHERE
D.TEST_NAME = 'Reading Comprehension' AND
T.TEST_SESSION_STATUS_ID = 3
AND TRUNC(T.TEST_END_DATE) BETWEEN TO_DATE('01-09-09','DD-MON-YY') AND TO_DATE('30-09-09','DD-MON-YY')
AND T.TEST_DETAIL_ID = D.TEST_DETAIL_ID
AND S.STUDENT_ID_PK = U.STUDENT_ID_PK
AND S.TEST_SESSION_ID = T.TEST_SESSION_ID
AND A.TEST_SESSION_ID = T.TEST_SESSION_ID
AND A.STUDENT_ID_PK = S.STUDENT_ID_PK
AND B.TEST_SESSION_ID = T.TEST_SESSION_ID
AND B.STUDENT_ID_PK = S.STUDENT_ID_PK
AND S.INSTITUTION_ID=E.INSTITUTION_ID
AND SUBSTR(E.INSTITUTION_ID_DISPLAY,8,3) '000') from dual;please advice -
Need to generate a Index xml file for corresponding Report PDF file.
Need to generate a Index xml file for corresponding Report PDF file.
Currently in fusion we are generating a pdf file using given Rtf template and dataModal source through Ess BIPJobType.xml .
This is generating pdf successfully.
As per requirement from Oracle GSI team, they need index xml file of corresponding generated pdf file for their own business scenario.
Please see the following attached sample file .
PDf file : https://kix.oraclecorp.com/KIX/uploads1/Jan-2013/354962/docs/BPA_Print_Trx-_output.pdf
Index file : https://kix.oraclecorp.com/KIX/uploads1/Jan-2013/354962/docs/o39861053.out.idx.txt
In R12 ,
We are doing this through java API call to FOProcessor and build the pdf. Here is sample snapshot :
xmlStream = PrintInvoiceThread.generateXML(pCpContext, logFile, outFile, dbCon, list, aLog, debugFlag);
OADocumentProcessor docProc = new OADocumentProcessor(xmlStream, tmpDir);
docProc.process();
PrintInvoiceThread :
out.println("<?xml version=\"1.0\" encoding=\"UTF-8\" ?>");
out.print("<xapi:requestset ");
out.println("<xapi:filesystem output=\"" + outFile.getFileName() + "\"/>");
out.println("<xapi:indexfile output=\"" + outFile.getFileName() + ".idx\">");
out.println(" <totalpages>${VAR_TOTAL_PAGES}</totalpages>");
out.println(" <totaldocuments>${VAR_TOTAL_DOCS}</totaldocuments>");
out.println("</xapi:indexfile>");
out.println("<xapi:document output-type=\"pdf\">");
out.println("<xapi:customcontents>");
XMLDocument idxDoc = new XMLDocument();
idxDoc.setEncoding("UTF-8");
((XMLElement)(generator.buildIndexItems(idxDoc, am, row)).getDocumentElement()).print(out);
idxDoc = null;
out.println("</xapi:customcontents>");
In r12 we have a privilege to use page number variable through oracle.apps.xdo.batch.ControlFile
public static final String VAR_BEGIN_PAGE = "${VAR_BEGIN_PAGE}";
public static final String VAR_END_PAGE = "${VAR_END_PAGE}";
public static final String VAR_TOTAL_DOCS = "${VAR_TOTAL_DOCS}";
public static final String VAR_TOTAL_PAGES = "${VAR_TOTAL_PAGES}";
Is there any similar java library which do the same thing in fusion .
Note: I checked in the BIP doc http://docs.oracle.com/cd/E21764_01/bi.1111/e18863/javaapis.htm#CIHHDDEH
Section 7.11.3.2 Invoking Processors with InputStream .
But this is not helping much to me. Is there any other document/view-let which covers these thing .
Appreciate any help/suggestions.
-anjani prasad
I have attached these java file in kixs : https://kix.oraclecorp.com/KIX/display.php?labelId=3755&articleId=354962
PrintInvoiceThread
InvoiceXmlBuilder
Control.javaYou can find the steps here.
http://weblogic-wonders.com/weblogic/2009/11/29/plan-xml-usage-for-message-driven-bean/
http://weblogic-wonders.com/weblogic/2009/12/16/invalidation-interval-secs/ -
I cannot get itunes to work/install on myPC. I am running windows 8 in a 64 bit machine. I also want to purchase books via itunes for my Nook. Are you deliberately excluding PC owners so we cannot use ituens on our computers???? I have selected a community but have no idea what community I should be working with. Can somone explai these communities to me? Also. you don't allow much time to go thru the sign up process. I had to start over SEVERAL times as you kept timing me out. Jesus...when people are signing up for the first time, we have no idea of the questions that will be asked. Give us a break and give us time to sign up. Not everone is disability free and able to go thru your process quickly...you are NOT user friendly. You should be for people who are wanting to join your community. I used to want an Apple product but now that I have had first hand experience with your web page, I think I will probably stay away from Apple products. VERY DISAPPOINTING
Jan 11, 2014 10:26 AM Re: I NEED I TUNES FOR A 64 BIT PC. OPERATING SYSTEM IS WINDOWS 8, NON TOUCH SCREEN. WHICH IS THE ABSOLUTELY BEST WAY TO GET ITUNES OPERATING CORRECTLY ON MY PC????//I HAVE TRIED SEVERAL THINGS, NONE OF WHICH ARE WORKING. IS ITUNES DELIBERATELY EXCLUDING
Re: I NEED I TUNES FOR A 64 BIT PC. OPERATING SYSTEM IS WINDOWS 8, NON TOUCH SCREEN. WHICH IS THE ABSOLUTELY BEST WAY TO GET ITUNES OPERATING CORRECTLY ON MY PC????//I HAVE TRIED SEVERAL THINGS, NONE OF WHICH ARE WORKING. IS ITUNES DELIBERATELY EXCLUDING PC'S in response to Birdlover1
Post over in the iTunes for Windows forum, here:
https://discussions.apple.com/community/itunes/itunes_for_windows
sebastian -
I'm trying to download the newest version of I tunes because I just got the I phone 5c..it wont download because I have to have I tunes 10.6.8 but it wont let me download that either because I need I tunes 10.6 to get it and I can't find that download anywhere..I need to get the book for school on my I pad or phone by tomorrow...what do I do?
Welcome to Apple Support Communities
The iPhone 5c needs iTunes 11.1, which requires Mac OS X 10.6.8. See if your Mac can be upgraded to Mac OS X Snow Leopard > http://support.apple.com/kb/sp575 If so, buy Mac OS X Snow Leopard > http://store.apple.com/us/product/MC573Z/A/mac-os-x-106-snow-leopard
Then, make a backup of your data with Time Machine or Disk Utility's Restore, insert the disc and follow its steps to upgrade to Mac OS X Snow Leopard. You shouldn't lose compatibility with any application. After upgrading, open > Software Update, and install the latest version. It will install iTunes 11.1, so you will be able to sync the iPhone 5c with your Mac -
Do I need to tune the base query of my materialized view ?
Hi gurus,
We implement materialized view (FAST REFRESH ON COMMIT) on base table used in OLTP system. In fact the MV itself is also queried by the OLTP system itself, interactively.
Currently user complain that after they input transaction, they see that the MV is not refresh as soon as they commit the transaction.
(it takes 5 to 10 minutes before they see the update on the MV)
The query of the MV is attached below, it consists of union all of two queries but actually the real query has 13 queries on a UNION ALL.
The question is :
Do I need to tune the query that forms the MV ?
Does it impact the refresh time of the MV ?
Thank you for your help,
xtanto
SELECT GRNH.DOC_CODE ||' '|| GRNH.GRN_NO, GRNH.GRN_DATE, GRNH.REGS_CODE,
F_Convert_Qty_K(GRND.PRODUCT, GRND.QTY, GRND.UOM) QTY_IN, 0 QTY_OUT, 0 QTY_ADJ, 'N/A'.....
FROM GRN_LOCAL_D GRND
JOIN GRN_LOCAL_H GRNH ON GRNH.GRNH_ID = GRND.GRNH_ID
WHERE GRND.PRODUCT IS NOT NULL AND GRNH.WH_CODE IS NOT NULL
UNION ALL
SELECT GRNH.DOC_CODE ||' '|| GRNH.GRN_NO, GRNH.GRN_DATE, GRNH.REGS_CODE,
F_Convert_Qty_K(GRND.PRODUCT_B, GRND.QTY_B, GRND.UOM_B) QTY_IN, 0 QTY_OUT, 0 QTY_ADJ, 'N/A'...
FROM GRN_LOCAL_D GRND
JOIN GRN_LOCAL_H GRNH ON GRNH.GRNH_ID = GRND.GRNH_ID
WHERE GRND.PRODUCT_B IS NOT NULL AND GRNH.WH_CODE IS NOT NULL
UNION ALL
......Hi Justin & APC,
Are they seeing a 5-10 minute delay in being able to commit their transaction?NO, the commit itself is completed normally, but after that when they go to the page (this is webbased app) that display data from the MV, the newly committed transactionis not there yet. Only after 5-10 minutes we can see the new data.
What does the F_Convert_Qty_K() function do?It is a function that convert the quantity from whatever Unit Of Measurement (UOM) into the smallest UOM. Below is the code.
So the question remains:
why intermittently theere is delay before the new data visible in the MV ?
Thanks you,
xtanto
CREATE OR REPLACE FUNCTION F_Convert_Qty (pProduct VARCHAR2, pQty NUMBER, pUOM VARCHAR2)
RETURN NUMBER
IS
vQtyBase NUMBER(13,5);
BEGIN
SELECT CASE WHEN pUOM = ( SELECT UOM_BASE FROM PRODUCTS WHERE PRODUCT = pProduct)
THEN pQty
ELSE
SELECT ROUND(pQty * Coefficient / Divisor, 5)
FROM (
SELECT Coefficient, Divisor FROM CONVERT_UOM
WHERE PRODUCT = pProduct AND UOM_CONVERT = pUOM
) tmp
END INTO vQtyBase FROM DUAL;
--RETURN (vQtyBase) ;
RETURN ( NVL(vQtyBase,0) ) ;
END F_Convert_Qty;
/ -
Need to create multiple indexes - will this work?
For a catalog of about 350 pages, I need both a manufacturer index and a product index. I have read some of the creative solutions already posted here and in other web discussions, and I have an idea.
Why not create two Book files? one could be "Catalog-Product.indb" and the other could be "Catalog-Manufacturer.indb". Basically two separate contexts for the series of layouts. When in the Product book, do all the indexing for that context, update the Product index and close the set of files (saving the Catalog-Product.indb as well and then closing it). Then open the Catalog-Manufacturer book and do the indexing for the Manufacturer index, generate it, save and close.
The question is: will each of the two book files be able to ignore the invisible index metadata that belongs to the other book when generating the index? will there be trash index entries in the Index palette - representing those of the other book file? Or, can these two sets of metadata be properly kept separate?
One consequence that I predict is that the Hyperlinks option during PDF Export from InDesign will be compromised - I would expect whichever book was inactive during the PDF export from the Book file would be the one whose index links would be non-functional.
Can anyone tell me though whether this strategy could generate two proper indexes at least for a print document?Bill Planey wrote:
For a catalog of about 350 pages, I need both a manufacturer index and a product index. I have read some of the creative solutions already posted here and in other web discussions, and I have an idea.
Why not create two Book files? one could be "Catalog-Product.indb" and the other could be "Catalog-Manufacturer.indb". Basically two separate contexts for the series of layouts. When in the Product book, do all the indexing for that context, update the Product index and close the set of files (saving the Catalog-Product.indb as well and then closing it). Then open the Catalog-Manufacturer book and do the indexing for the Manufacturer index, generate it, save and close.
The question is: will each of the two book files be able to ignore the invisible index metadata that belongs to the other book when generating the index? will there be trash index entries in the Index palette - representing those of the other book file? Or, can these two sets of metadata be properly kept separate?
One consequence that I predict is that the Hyperlinks option during PDF Export from InDesign will be compromised - I would expect whichever book was inactive during the PDF export from the Book file would be the one whose index links would be non-functional.
Can anyone tell me though whether this strategy could generate two proper indexes at least for a print document?
Hi, Bill:
Search Google for terms like "InDesign multiple index" without quotes. There are some good links for this; admittedly they're workarounds. If you take the time to file a formal request for this feature at https://www.adobe.com/cfusion/mmform/index.cfm?name=wishform, there's a chance the developers will eventually run out of excuses or higher-priority tasks, and include it in a future release.
HTH
Regards,
Peter
Peter Gold
KnowHow ProServices -
Privilege need to create functional index
Greeting gurus?
What privilege I needed for create functional index
I am writing a definer righted procedure to create index however when creating functional index, it gives me ora-0131 insufficent but unique, bitmap and reverse index works
and I granted create any index and create any indextype already. what else is missing ?CREATE INDEX
To create a function-based index, in addition to the prerequisites for creating a conventional index, if the index is based on user-defined functions, then those functions must be marked DETERMINISTIC. Also, you must have the EXECUTE object privilege on any user-defined function(s) used in the function-based index if those functions are owned by another user. -
i need i tunes 10.5 for my i phone 4s and cant download onto 64 bit windows 7 laptop tried everything anyone help me please????
true. but he is right itunes automatically sees your running 64 bit and as soon as you go to the site on left hand side of sceen you see download 10.5 just click it and you should see this it automatically points to 64 bit install then you just hit download again. pretty simple
Tunes 10.5 for Windows (64-bit)
Email me New On iTunes and special iTunes offers.
Keep me up to date with Apple news, software updates, and the latest information on products and services.
then right below this it says download
or just click this link http://www.apple.com/itunes/download/ -
hi there, I am not sure if I need to perform process Index? Do I have to do that after I process full on measure groups?
thanks
--Currently using Reporting Service 2000; Visual Studio .NET 2003; Visual Source Safe SSIS 2008 SSAS 2008, SVN --I did process update on all dimensions... after that, I did process full on each measure groups...
In this case, do I need to process index on the whole cube?
Hi Cat_ca,
ProcessFull - Simple to digest, applies all objects which means it discards the storage contents of the object and rebuilds them.
ProcessIndexes - Applies to dimension, cube, measure group, and partition.
Since you did process full on each measure groups, I don't think you need do a processindex, here are some links about process options on Analysis Services for your reference.
Different Kinds of SSAS Processing in simple words
Difference between SSAS ProcessUpdate and ProcessFull
Regards,
Charlie Liao
If you have any feedback on our support, please click
here.
Charlie Liao
TechNet Community Support -
Do i need to tune the redolog?
10g R2 on AIX 5.3L
this is a typical alert file about the redo log file switching of the day.
Mon May 28 01:26:33 2007
Thread 1 advanced to log sequence 24017
Current log# 4 seq# 24017 mem# 0: /u020/apps/oracle/oradata/BAAN/redo04.log
Mon May 28 08:04:13 2007
Thread 1 advanced to log sequence 24018
Current log# 5 seq# 24018 mem# 0: /u020/apps/oracle/oradata/BAAN/redo05.log
Mon May 28 08:05:56 2007
Thread 1 advanced to log sequence 24019
Current log# 3 seq# 24019 mem# 0: /u020/apps/oracle/oradata/BAAN/redo03.log
Mon May 28 08:07:29 2007
Thread 1 advanced to log sequence 24020
Current log# 2 seq# 24020 mem# 0: /u020/apps/oracle/oradata/BAAN/redo02.log
Mon May 28 08:08:55 2007
Thread 1 advanced to log sequence 24021
Current log# 1 seq# 24021 mem# 0: /u020/apps/oracle/oradata/BAAN/redo01.log
Mon May 28[b] 08:10:26 2007
Thread 1 advanced to log sequence 24022
Current log# 4 seq# 24022 mem# 0: /u020/apps/oracle/oradata/BAAN/redo04.log
Mon May 28 08:12:05 2007
Thread 1 advanced to log sequence 24023
Current log# 5 seq# 24023 mem# 0: /u020/apps/oracle/oradata/BAAN/redo05.log
Mon May 28 08:13:55 2007
Thread 1 advanced to log sequence 24024
Current log# 3 seq# 24024 mem# 0: /u020/apps/oracle/oradata/BAAN/redo03.log
Mon May 28 11:31:05 2007
Thread 1 advanced to log sequence 24025
Current log# 2 seq# 24025 mem# 0: /u020/apps/oracle/oradata/BAAN/redo02.log
Mon May 28 14:36:13 2007
Thread 1 advanced to log sequence 24026
Current log# 1 seq# 24026 mem# 0: /u020/apps/oracle/oradata/BAAN/redo01.log
Do I need to tune the redo log? Each of redo log file is 50MB. It switched more frequent when there are not many users than real busy hours.
nullSelect may or may not generate redo. Delayed block cleanout is one of the reasons for a query to produce redo: https://metalink.oracle.com/metalink/plsql/f?p=130:14:2631437878498686237::::p14_database_id,p14_docid,p14_show_header,p14_show_help,p14_black_frame,p14_font:NOT,40689.1,1,1,1,helvetica#aref1
Another reason is locking the rows with "for update":
SQL> create table t1 as select * from all_objects;
Table created.
SQL> set autot trace stat
SQL> select * from t1;
39310 rows selected.
Statistics
0 recursive calls
0 db block gets
3117 consistent gets
538 physical reads
0 redo size
2524195 bytes sent via SQL*Net to client
18612 bytes received via SQL*Net from client
2622 SQL*Net roundtrips to/from client
0 sorts (memory)
0 sorts (disk)
39310 rows processed
SQL> select * from t1 for update;
39310 rows selected.
Statistics
24 recursive calls
40020 db block gets
43446 consistent gets
625 physical reads
7754052 redo size
3501635 bytes sent via SQL*Net to client
18612 bytes received via SQL*Net from client
2622 SQL*Net roundtrips to/from client
0 sorts (memory)
0 sorts (disk)
39310 rows processed -
Hi ,
I had one ZEE program which has 25 files as input file (input data's . Here the logic of the program is as mention below.
1.initializition .
2.get_file_names.(All 25 files name is get in the same perform )
3.write_job_log.
4.open_files.(Open 25 files at a time )
5.read_surr_ctl_file.
6.load_internal_tabs.( Load 25 files data to ITAB )
7.create_edi214_tables.
8.process_ekko_ekpo.(Proces the ITAB data at the time.)
9.refresh_internal_tables.
Here they are openning 25 files , then loading all 25 files to ITAB and processing all the loaded datas.
When this progarm is schedule in background it takes 1.25 hrs to complete the process .
I need to tune this program which ever possible way.
My thought and analysis in this program is -> This performance problem is due to like openning 25 files , then loading all 25 files to ITAB and processing all the loaded datas which may hit the performance . Kindly let me know you IDEAS and experence on this issue . This program is having 10000 lines of code ..
Regards,
VeeraHi,
This program has a lot of steps and the best way to solve its performance problems is with SE30.
I think you should execute SE30 in development environment e check the steps where there is bad performance.
If you improve the performance in the development environment you will improve also in production environment.
Regards,
Fernando -
Space needed to create an index
I have a table with about 30 million records. I needed to create an index for one of the fields in this table. How can I determine how much space I will need in the tablespace where I need to create the index for the index to be created ?
ThanksThere's a nice discussion regarding this on asktom:
http://asktom.oracle.com/pls/ask/f?p=4950:8:::::F4950_P8_DISPLAYID:1295520444744
Hope this helps.
Kailash. -
Help needed to tune the Query:Statistics added
Can someone DBA please help me to tune this query:
SELECT DISTINCT K.ATTRIBUTE_VALUE AGENCY_ID,B.PROFILE_NM ,NVL(G.OFFICE_DESC,'--') OFFICE_DESC,f.OFFICE_ID,B.PROFILE_ID,'%' ROLE,'%' LAYOUT,
CASE
WHEN 'flagB' = '%' THEN
NVL(J.ISS_GRP_DESC,'BILLING')
WHEN 'flagO' = '%' THEN
NVL(J.ISS_GRP_DESC,'ORDERING')
WHEN 'flag' = '%' THEN
NVL(J.ISS_GRP_DESC,'BILLING/ORDERING')
ELSE
NVL(J.ISS_GRP_DESC,' ')
END ISS_GRP_DESC,
DECODE(NVL(H.USERID,' ') ,' ','--','<a sbcuid_in=' || H.USERID || ' target=NEW >'||H.FIRSTNAME || ' ' || H.LASTNAME || '( ' || H.USERID || ' )</a>' ) USER_NAME
FROM
PROFILE_PORTAL B ,
TBL_BDA_AGENCY_RESP_REP C ,
TBL_BDA_AGENCY_OFFICE F,
TBL_BDA_OFFICE G,
USERS_PORTAL H,
TBL_BDA_USR_ISS_GRP I ,
TBL_BDA_ISS_GROUP J,
ATTRIBUTE_VALUES_PORTAL K,
PROFILE_TYPE_PORTAL L
WHERE
B.PROFILE_ID = F.AGENCY_ID (+)
AND B.PROFILE_ID = C.AGENCY_ID (+)
AND G.OFFICE_ID (+)= F.OFFICE_ID
AND H.USERID (+)= C.RESP_USR_ID
AND C.ISS_GRP_ID = I.ISS_GRP_ID (+)
AND I.ISS_GRP_ID = J.ISS_GRP_ID(+)
AND 'PROFILE.'||B.PROFILE_ID = K.ENTITY_ID(+)
AND K.ATTRIBUTE_VALUE IS NOT NULL
AND L.PROFILE_TYPE_ID = B.PROFILE_TYPE_ID
AND L.APPLICATION_CD='BDA'
AND NOT EXISTS (SELECT agency_id
FROM TBL_BDA_AGENCY_RESP_REP t
WHERE t.ISS_GRP_ID IN ('%')
AND t.AGENCY_ID = C.AGENCY_ID)
AND K.ATTRIBUTE_VALUE LIKE '%'
AND UPPER(B.PROFILE_NM) LIKE UPPER('%')
AND (to_char(NVL(B.PROFILE_ID,0)) LIKE '%' OR NVL(B.PROFILE_ID,0) IN ('a'))
AND NVL(G.OFFICE_ID,0) IN ('%')
AND (to_char(NVL(C.RESP_USR_ID,'0')) LIKE '%' OR NVL(C.RESP_USR_ID,'0') IN ('k'))
ORDER BY PROFILE_NM
The number of rows in these tables are as follows:
PROFILE_PORTAL -- 2392
TBL_BDA_AGENCY_RESP_REP 3508
TBL_BDA_AGENCY_OFFICE 2151
TBL_BDA_OFFICE 3
USERS_PORTAL 270500
TBL_BDA_USR_ISS_GRP 234
TBL_BDA_ISS_GROUP 2
ATTRIBUTE_VALUES_PORTAL 2790
PROFILE_TYPE_PORTAL 3
The Explain pal nhas given this o/p to me:
SQL> select * from table(dbms_xplan.display) dual;
PLAN_TABLE_OUTPUT
| Id | Operation | Name | Rows | Bytes |TempSpc| Cost |
| 0 | SELECT STATEMENT | | 807 | 102K| | 2533 |
| 1 | SORT UNIQUE | | 807 | 102K| 232K| 82 |
|* 2 | FILTER | | | | | |
|* 3 | HASH JOIN OUTER | | 807 | 102K| | 52 |
|* 4 | HASH JOIN OUTER | | 807 | 95226 | | 40 |
|* 5 | TABLE ACCESS BY INDEX ROWID | ATTRIBUTE_VALUES | 1 | 23 | | 2 |
| 6 | NESTED LOOPS | | 7 | 805 | | 37 |
| 7 | NESTED LOOPS OUTER | | 6 | 552 | | 25 |
|* 8 | FILTER | | | | | |
| 9 | NESTED LOOPS OUTER | | | | | |
|* 10 | FILTER | | | | | |
| 11 | NESTED LOOPS OUTER | | | | | |
| 12 | NESTED LOOPS OUTER | | 3 | 141 | | 10 |
|* 13 | HASH JOIN | | 3 | 120 | | 7 |
|* 14 | TABLE ACCESS FULL | PROFILE | 6 | 198 | | 4 |
|* 15 | TABLE ACCESS FULL | PROFILE_TYPE | 1 | 7 | | 2 |
|* 16 | INDEX RANGE SCAN | SYS_C0019777 | 1 | 7 | | 1 |
| 17 | TABLE ACCESS BY INDEX ROWID| TBL_BDA_OFFICE | 1 | 10 | | 1 |
|* 18 | INDEX UNIQUE SCAN | SYS_C0019800 | 1 | | | |
| 19 | TABLE ACCESS BY INDEX ROWID | TBL_BDA_AGENCY_RESP_REP | 2 | 26 | | 2 |
|* 20 | INDEX RANGE SCAN | IDX_AGECYRESP_AGNCYID | 2 | | | 1 |
| 21 | TABLE ACCESS BY INDEX ROWID | USER_ | 1 | 22 | | 1 |
|* 22 | INDEX UNIQUE SCAN | USER_PK | 1 | | | |
|* 23 | INDEX RANGE SCAN | IDX_ATTVAL_ENTATTID | 1 | | | 1 |
| 24 | TABLE ACCESS FULL | TBL_BDA_USR_ISS_GRP | 234 | 702 | | 2 |
| 25 | TABLE ACCESS FULL | TBL_BDA_ISS_GROUP | 2 | 24 | | 2 |
|* 26 | TABLE ACCESS BY INDEX ROWID | TBL_BDA_AGENCY_RESP_REP | 1 | 7 | | 3 |
|* 27 | INDEX RANGE SCAN | IDX_AGECYRESP_AGNCYID | 2 | | | 1 |
Predicate Information (identified by operation id):
2 - filter( NOT EXISTS (SELECT /*+ */ 0 FROM "TBL_BDA_AGENCY_RESP_REP" "T" WHERE "T"."AGENCY_ID"=:B1
AND "T"."ISS_GRP_ID"=TO_NUMBER('%')))
3 - access("I"."ISS_GRP_ID"="J"."ISS_GRP_ID"(+))
4 - access("SYS_ALIAS_1"."ISS_GRP_ID"="I"."ISS_GRP_ID"(+))
5 - filter("K"."ATTRIBUTE_VALUE" IS NOT NULL AND "K"."ATTRIBUTE_VALUE" LIKE '%')
8 - filter(NVL("SYS_ALIAS_1"."RESP_USR_ID",'0') LIKE '%' OR NVL("SYS_ALIAS_1"."RESP_USR_ID",'0')='k')
10 - filter(NVL("G"."OFFICE_ID",0)=TO_NUMBER('%'))
13 - access("L"."PROFILE_TYPE_ID"="B"."PROFILE_TYPE_ID")
14 - filter(UPPER("B"."PROFILE_NM") LIKE '%' AND (TO_CHAR(NVL("B"."PROFILE_ID",0)) LIKE '%' OR
NVL("B"."PROFILE_ID",0)=TO_NUMBER('a')))
15 - filter("L"."APPLICATION_CD"='BDA')
16 - access("B"."PROFILE_ID"="F"."AGENCY_ID"(+))
18 - access("G"."OFFICE_ID"(+)="F"."OFFICE_ID")
20 - access("B"."PROFILE_ID"="SYS_ALIAS_1"."AGENCY_ID"(+))
22 - access("H"."USERID"(+)="SYS_ALIAS_1"."RESP_USR_ID")
23 - access("K"."ENTITY_ID"='PROFILE.'||TO_CHAR("B"."PROFILE_ID"))
26 - filter("T"."ISS_GRP_ID"=TO_NUMBER('%'))
27 - access("T"."AGENCY_ID"=:B1)
Note: cpu costing is off
57 rows selected.
Elapsed: 00:00:01.08
Please help me.
Aashish S.Hello Eric,
Here is the code:
SELECT DISTINCT
K.ATTRIBUTE_VALUE AGENCY_ID,
B.PROFILE_NM ,
NVL(G.OFFICE_DESC,'--') OFFICE_DESC,
f.OFFICE_ID,
B.PROFILE_ID,
'%' ROLE,
'%' LAYOUT,
case
WHEN 'flagB' = '%' THEN
NVL(J.ISS_GRP_DESC,'BILLING')
WHEN 'flagO' = '%' THEN
NVL(J.ISS_GRP_DESC,'ORDERING')
WHEN 'flag' = '%' THEN
NVL(J.ISS_GRP_DESC,'BILLING/ORDERING')
else
NVL(J.ISS_GRP_DESC,' ')
END ISS_GRP_DESC,
DECODE(NVL(H.USERID,' ') ,' ','--','<a sbcuid_in=' || H.USERID || ' target=NEW >'||H.FIRSTNAME || ' ' || H.LASTNAME ||
'( ' || H.USERID || ' )</a>' ) USER_NAME
from
PROFILE_PORTAL B ,
TBL_BDA_AGENCY_RESP_REP C ,
TBL_BDA_AGENCY_OFFICE F,
TBL_BDA_OFFICE G,
USERS_PORTAL H,
TBL_BDA_USR_ISS_GRP I ,
TBL_BDA_ISS_GROUP J,
ATTRIBUTE_VALUES_PORTAL K,
PROFILE_TYPE_PORTAL L
WHERE
B.PROFILE_ID = F.AGENCY_ID (+)
AND B.PROFILE_ID = C.AGENCY_ID (+)
AND G.OFFICE_ID (+)= F.OFFICE_ID
AND H.USERID (+)= C.RESP_USR_ID
AND C.ISS_GRP_ID = I.ISS_GRP_ID (+)
AND I.ISS_GRP_ID = J.ISS_GRP_ID(+)
AND 'PROFILE.'||B.PROFILE_ID = K.ENTITY_ID(+)
AND K.ATTRIBUTE_VALUE IS NOT NULL
AND L.PROFILE_TYPE_ID = B.PROFILE_TYPE_ID
AND L.APPLICATION_CD='BDA'
AND NOT EXISTS
(SELECT agency_id
FROM TBL_BDA_AGENCY_RESP_REP t
WHERE t.ISS_GRP_ID IN (1)
AND t.AGENCY_ID = C.AGENCY_ID)
AND K.ATTRIBUTE_VALUE LIKE '%'
AND UPPER(B.PROFILE_NM) LIKE UPPER('%')
AND (to_char(NVL(B.PROFILE_ID,0))
LIKE '%'
OR NVL(B.PROFILE_ID,0) IN (1))
AND NVL(G.OFFICE_ID,0) IN (1)
AND (to_char(NVL(C.RESP_USR_ID,'0'))
LIKE '%'
OR NVL(C.RESP_USR_ID,'0') IN ('%'))
ORDER BY PROFILE_NM
This is the Query and the query takes some mins. to run in prod environment.
From the Query plan ,I am not able to get any idea for optimization.
Now,Can you tell me which steps I need to follow to run it faster and which all modifications should be made?
Thanks.
Aashish S. -
This below query's performance is ver slow,
SELECT res.*
FROM (SELECT ROWNUM rx,
e.*
FROM (SELECT Nvl(company_nm, individual_last_nm
|| ','
|| Substr(individual_first_nm, 1, 1))
customer_name,
h.company_nm
company_nm
a.individual_last_nm
indiv_last_nm,
a.individual_first_nm
indiv_first_nm,
b.proposal_nbr
proposal_nbr,
i.retailer_id
retailerid
c.proposal_id
pid,
c.proposal_status_gid
statusid,
p.config_value
status,
e.vehicle_make_nm
make,
e.vehicle_model_nm
model,
k.product_type_desc
product_type,
c.revision_id
revision_id,
c.credit_score
credit_score,
c.total_of_guarantor
noofguar,
c.total_of_applicant
noofapp,
c.update_dttm
last_modified_date
j.user_nm
business_manager_nm,
l.underwriting_appln_num
underwriting_appln_num,
Nvl2(intrnl_held_offering_comnts, 'Y',
Nvl2(dealer_held_offering_comnts, 'Y',
'N')) AS held_offering,
last_updated_user_flag AS
hld_offr_usr_flag
FROM individual a,
customer b,
proposal c,
quote_vehicle e,
finance_quote f,
quote_finance_plan g,
system_config p,
company h,
application_retailer i,
gss_user j,
finance_product_type k,
underwriting_result l
WHERE b.customer_nbr = a.customer_nbr (+)
AND b.customer_nbr = h.customer_nbr (+)
AND b.proposal_nbr = c.proposal_nbr
AND c.proposal_nbr = e.proposal_nbr
AND c.proposal_nbr = f.proposal_nbr
AND c.proposal_nbr = l.proposal_nbr (+)
AND f.finance_quote_nbr = g.finance_quote_nbr
AND f.proposal_nbr = g.proposal_nbr
AND Substr(c.dealer_branch_nbr, 4, 4) = i.dealer
AND Substr(c.dealer_branch_nbr, 1, 3) = i.branch
AND b.applicant_flg = 'A'
AND b.customer_sequence = 1
AND i.application_type_id = 1
AND f.selected_flg = 'Y'
AND c.country_cd = i.country_cd
AND g.selected_product_nm = k.product_type_nm
AND c.country_cd = k.country_cd
AND i.active_flg = 'Y'
AND c.update_dttm BETWEEN To_date(' 2010-07-20 ',
'yyyy-mm-dd')
AND
To_date(
' 2010-07-26'
||
' 23:59:59',
'yyyy-mm-dd hh24:mi:ss')
AND c.country_cd = :a
AND c.country_cd = p.country_cd
AND p.config_key LIKE 'PROPOSAL_STATUS_%'
AND Substr(config_key, Instr(config_key, '_', -1) + 1) =
c.proposal_status_gid
AND p.language_cd = :a
AND ( c.retailer_id IN ( 'QGNLBEMY', 'YQIWOVRE' ) )
AND c.brand_id IN ( '2', '1', '3', '52', '11' )
AND c.fi_manager_id = j.user_id (+)
ORDER BY last_modified_date DESC) e
WHERE ROWNUM <= 15) res
WHERE rx >= 1Contd...
Explain Plan for above query
==============================
SQL> /
PLAN_TABLE_OUTPUT
| Id | Operation | Name | Rows | Bytes | Cost | Pstart| Pstop |
| 0 | SELECT STATEMENT | | 1 | 1842 | 16 | | |
|* 1 | VIEW | | 1 | 1842 | 16 | | |
|* 2 | COUNT STOPKEY | | | | | | |
| 3 | VIEW | | 1 | 1829 | 16 | | |
| 4 | NESTED LOOPS | | 1 | 380 | 16 | | |
| 5 | NESTED LOOPS OUTER | | 1 | 354 | 15 | | |
| 6 | NESTED LOOPS OUTER | | 1 | 322 | 14 | | |
| 7 | NESTED LOOPS | | 1 | 289 | 13 | | |
| 8 | NESTED LOOPS | | 1 | 271 | 11 | | |
| 9 | NESTED LOOPS | | 1 | 221 | 10 | | |
| 10 | NESTED LOOPS | | 1 | 207 | 9 | | |
| 11 | NESTED LOOPS OUTER | | 1 | 194 | 8 | | |
| 12 | NESTED LOOPS OUTER | | 1 | 162 | 7 | | |
| 13 | NESTED LOOPS | | 1 | 148 | 6 | | |
| 14 | NESTED LOOPS | | 1 | 124 | 5 | | |
|* 15 | TABLE ACCESS BY GLOBAL INDEX ROWID| PROPOSAL | 1 | 74 | 4 | ROWID | ROW L |
|* 16 | INDEX RANGE SCAN DESCENDING | ALT_IDX_PROP_DTTM | 1 | | 2 | | |
| 17 | PARTITION LIST ITERATOR | | | | | KEY | KEY |
|* 18 | TABLE ACCESS FULL | SYSTEM_CONFIG | 1 | 50 | 1 | KEY | KEY |
| 19 | TABLE ACCESS BY GLOBAL INDEX ROWID | QUOTE_VEHICLE | 1 | 24 | 1 | ROWID | ROW L |
|* 20 | INDEX UNIQUE SCAN | SYS_C007438 | 1 | | | | |
| 21 | TABLE ACCESS BY GLOBAL INDEX ROWID | UNDERWRITING_RESULT | 1 | 14 | 1 | ROWID | ROW L |
|* 22 | INDEX UNIQUE SCAN | SYS_C007550 | 1 | | | | |
| 23 | TABLE ACCESS BY INDEX ROWID | GSS_USER | 1 | 32 | 1 | | |
|* 24 | INDEX UNIQUE SCAN | XPKGSS_USER | 1 | | | | |
|* 25 | INDEX RANGE SCAN | XAK1_PERF_FINANCE_QUOTE | 1 | 13 | 1 | | |
| 26 | TABLE ACCESS BY GLOBAL INDEX ROWID | QUOTE_FINANCE_PLAN | 1 | 14 | 1 | ROWID | ROW L |
|* 27 | INDEX UNIQUE SCAN | SYS_C007421 | 1 | | | | |
| 28 | TABLE ACCESS BY INDEX ROWID | FINANCE_PRODUCT_TYPE | 1 | 50 | 1 | | |
|* 29 | INDEX UNIQUE SCAN | XPKFINANCE_PRODUCT_TYPE | 1 | | | | |
|* 30 | TABLE ACCESS BY GLOBAL INDEX ROWID | CUSTOMER | 1 | 18 | 2 | ROWID | ROW L |
|* 31 | INDEX RANGE SCAN | XAK1_PERF_CUSTOMER | 1 | | 1 | | |
| 32 | TABLE ACCESS BY GLOBAL INDEX ROWID | COMPANY | 1 | 33 | 1 | ROWID | ROW L |
|* 33 | INDEX UNIQUE SCAN | SYS_C007184 | 1 | | | | |
| 34 | TABLE ACCESS BY GLOBAL INDEX ROWID | INDIVIDUAL | 1 | 32 | 1 | ROWID | ROW L |
|* 35 | INDEX UNIQUE SCAN | SYS_C007329 | 1 | | | | |
|* 36 | TABLE ACCESS BY GLOBAL INDEX ROWID | APPLICATION_RETAILER | 1 | 26 | 1 | ROWID | ROW L |
|* 37 | INDEX UNIQUE SCAN | SYS_C004725 | 2 | | | | |
Predicate Information (identified by operation id):
1 - filter("RES"."RX">=1)
2 - filter(ROWNUM<=15)
15 - filter("C"."COUNTRY_CD"=:Z AND ("C"."RETAILER_ID"='QGNLBEMY' OR "C"."RETAILER_ID"='YQIWOVRE') AND ("C"."BRAND_ID"=1
OR "C"."BRAND_ID"=2 OR "C"."BRAND_ID"=3 OR "C"."BRAND_ID"=11 OR "C"."BRAND_ID"=52))
16 - access("C"."UPDATE_DTTM">=TIMESTAMP'2010-07-20 00:00:00' AND "C"."UPDATE_DTTM"<=TIMESTAMP'2010-07-26 23:59:59')
filter("C"."UPDATE_DTTM"<=TIMESTAMP'2010-07-26 23:59:59' AND "C"."UPDATE_DTTM">=TIMESTAMP'2010-07-20 00:00:00')
18 - filter("C"."COUNTRY_CD"="P"."COUNTRY_CD" AND "P"."CONFIG_KEY" LIKE 'PROPOSAL_STATUS_%' AND "C"."PROPOSAL_STATUS_GID
"=TO_NUMBER(SUBSTR("P"."CONFIG_KEY",INSTR("P"."CONFIG_KEY",'_',-1)+1)) AND "P"."LANGUAGE_CD"=:Z)
20 - access("C"."PROPOSAL_NBR"="E"."PROPOSAL_NBR")
22 - access("C"."PROPOSAL_NBR"="L"."PROPOSAL_NBR"(+))
24 - access("C"."FI_MANAGER_ID"="J"."USER_ID"(+))
25 - access("C"."PROPOSAL_NBR"="F"."PROPOSAL_NBR" AND "F"."SELECTED_FLG"='Y')
filter("F"."SELECTED_FLG"='Y')
27 - access("F"."PROPOSAL_NBR"="G"."PROPOSAL_NBR" AND "F"."FINANCE_QUOTE_NBR"="G"."FINANCE_QUOTE_NBR")
29 - access("G"."SELECTED_PRODUCT_NM"="K"."PRODUCT_TYPE_NM" AND "C"."COUNTRY_CD"="K"."COUNTRY_CD")
30 - filter("B"."APPLICANT_FLG"='A' AND "B"."CUSTOMER_SEQUENCE"=1)
31 - access("B"."PROPOSAL_NBR"="C"."PROPOSAL_NBR")
33 - access("B"."CUSTOMER_NBR"="H"."CUSTOMER_NBR"(+))
35 - access("B"."CUSTOMER_NBR"="A"."CUSTOMER_NBR"(+))
36 - filter("I"."ACTIVE_FLG"='Y')
37 - access("I"."DEALER"=SUBSTR("C"."DEALER_BRANCH_NBR",4,4) AND "C"."COUNTRY_CD"="I"."COUNTRY_CD" AND "I"."BRANCH"=SUBS
TR("C"."DEALER_BRANCH_NBR",1,3) AND "I"."APPLICATION_TYPE_ID"=1
Note: cpu costing is off
Please suggest for tunning process(what need to do)
Thanks..
Maybe you are looking for
-
Reloaded Itunes on an XP machine. Reloaded because Itunes would not open. Now when I play a movie that is in my library, the movie plays for aprx. 60 seconds and then reverts back to my Itunes library. Does it with all movies in the library, Most whe
-
Changing the Color of an Object in Elements 4.0
Neither of the two methods documented in Elements 4.0 to change the color of an object seem to work on my object. It is a fleur-de-lis that I extracted from a PDF an artist made for me some time ago. It's now black and I want to make it gold. What am
-
Buffer_pool and db_keep_cache_size
hi all, i want to change the buffer_pool of one table to KEEP and my parameter settings are sga_target big integer 7516192768 buffer_pool_recycle string db_recycle_cache_size big integer 0 db_keep_cache_size big integer 0 if i change the buffer_pool
-
Outbound program s for idoc in ale
hi to all abap gurus I have doubt regarding idocs we know that outbound programs are message specific ie for each message type there is one outbound program . we know that orders 05 idoc transfers both purchse and sales orders data by using two mes
-
# Question I am informed by FF that there are updates. When I click on them the window just says it's trying to connect to the server and never does. I am never able to update and I believe FF is now up to 3.6.17. I have a Mac Powerbook with 10.4.11