Query on data dictionary results in large number of physical reads

I don't understand why I am getting 80,000 physicals for this query. I am not looking for help re-writing this. I just don't understand why I would hit the disk at all.
My understanding had been that v$views where sql structures that pointed to x$tables. These x$tables are sql structures.
underneath, the x$tables were linked lists stored in memory. This is why when you bounce the database, all the data, gets reset. Since it is not saved to disk.
I am doing a simple insert/select off of v$open_cursor that is resulting in 80,000+ physical reads. I am posting the tkprof. It is all from v$open_cursor.
mysid_table has 6 records. It is 1 mb in size
if I index mysid_table.sid the query reduces to 20,000 physical reads. (but all the physical reads are on v$session_event)
the sequence number I am passing returns 2 SIDs
insert into my_save_table
select *
from v$session_event
where sid in (select sid
  from my_sid_table
where id = vseq);
vrowcount := sql%rowcount;
call     count       cpu    elapsed       disk      query    current        rows
Parse        1      0.01       0.01          0          0          0           0
Execute      1     31.70      47.57      88570         22          0           1
Fetch        0      0.00       0.00          0          0          0           0
total        2     31.71      47.58      88570         22          0           1
Misses in library cache during parse: 1
Optimizer mode: CHOOSE
Parsing user id: 22 
Elapsed times include waiting on following events:
  Event waited on                             Times   Max. Wait  Total Waited
  ----------------------------------------   Waited  ----------  ------------
  latch: row cache objects                        1        0.00          0.00
  log file sync                                   1        0.00          0.00
  SQL*Net message to client                       1        0.00          0.00
  SQL*Net message from client                     1        0.00          0.00
********************************************************************************

It seems like there is some missing information.
You have a wait for a log file sync, but no commit.
Your table my_sid_table is 1 MB for only 6 records?
Does the target table you are inserting into (my_save_table) have indexes on it?

Similar Messages

  • SQL remote query and data-dictionary

    Hi,
    We are considering the implementation of SAP Business One 2005. I've installed a demo version but in order for it to continue to support some business functions I need to perform a number of SQL queries against the data. Unfortunately I'm having difficulties in getting the answer from our reseller.
    Does anyone know the answers to the following question. I'd very much appreciate some help.
    1) Is there a data-dictionary document showing which tables perform which roles and the associated joins/associations.
    2) Is it 'advised' to open the SAP datasource directly to perform a SELECT query to obtain the results or should a third-party tool be used?
    3) Do I need the SDK if I want to perform direct SELECT queries or if I want to automate posting invoices to SAP? We have competent programmers in the company - I believe they use VB6 and VB.net.
    4) Is there an example if I simply wanted to query all suppliers within SAP? (e.g. SELECT NAME FROM Supplier_tbl...etc..etc)
    Any help would be very much appreciated - it's a huge application and I'm struggling a little.
    Many thanks,
    Gavin Russell.

    Gavin,
    The forum that you have posted your questions in is or questions related to the SAP Business One Integration for SAP NetWeaver.  Your questions seem to be along the line of general data access to SAP Business One.  You should typically post those types of questions in the SAP Business One Discussion Forum or the SAP Business One SDK Discussion Forum.
    As to your questions ...
    1.  The definition of the SAP Business One table schema is part of the SAP Business One SDK Help Center documentation.  You simply need to install the SAP Business One SDK Help and you will have it.  There is not a "data dictionary" other than this documentation.
    2.  Typically any direct access to SAP Business One data using SQL is discouraged as if data corruption occurs as a result of it, the environment would not then be supported by SAP.
    3.  If you want to push data into SAP Business One, as stated, direct access to the Business One database is forbidden.  You would use the SAP Business One SDK for this.
    4.  You can use the SAP Business One Query Manager/Wizard in the SAP Business One product which in essence is performing SQL type SELECTS for data on Business One tables.
    HTH,
    Eddy

  • Large number of physical IOs for a simple query

    oracle 10.2.0.3
    redhat 4.5
    table size 160 GB
    partition by day
    I do:
    select count(*)
    from mytable
    where rownum < 2;
    does 122,000 physical reads
    I run a 10046 trace and look at tkprof all of my wait time is on the physical reads. It takes 30 seconds to return.
    I look at the raw trace file. None of the individual reads seem to take that long. Its just that there are so many of them.
    I then take the same query and restrict it by
    day field (partitioned field) < a value I dont have in my table
    by my assumption this should only scan the first partition. This query has run for 15 minutes and is not returning.
    Nothing else is running. The individual physical IO calls are not slow
    No problems with any other tables.
    This is on a SAN using ASM.

    I wonder if the problem is similar to the one in this thread:
    http://groups.google.com/group/comp.databases.oracle.server/browse_thread/thread/79d0a3d85381dc64/ed3ab1dbeeaa1cb2
    Posts 26-29 arrived at the conclusion that in that particular case the index either needed to be rebuilt or coalesced. The plan showed "INDEX FULL SCAN (MIN/MAX)", which should have only required a very small number of IOs (likely less than 5), but actually required 118,525 logical IOs with 118,248 single block reads from disk.
    Charles Hooper
    IT Manager/Oracle DBA
    K&M Machine-Fabricating, Inc.

  • Large amount of physical reads

    Any query using a CONTAINS accessing our main IMT index does a huge amount of physical reads. The cost according to an explain plan is great, only 3, but usually over 2000 physical reads each time. Even if you do the same one over and over, it never decreases. What might I do about this?
    Steve Karam
    Lead DBA www.traderonline.com www.harmonhomes.com

    Our application runs on ColdFusion and points to our database from 24 seperate webservers all armed with Oracle Client. The 'databii' are both 8.1.6.3 on Sun SPARC Solaris 2.7. The index is created with 160M memory, our STORAGE preference puts the tables and index in a seperate tablespace about 4 times their size. We use a USER type Datastore to convert the indexed field to an XML appearance, then use the Section Group to read through it XML style. It indexes about 430,000 documents. We rebuild 'optimize full' every morning and keep frag levels under 2%. Our queries use CONTAINS and makes use of the SYN and PT function to query against the thesaurus, and in a few spots we use the WITHIN clause for our 2 zones, MAKE_NAME and MODEL_NAME (with MAK and MOD as tags respectively).

  • Filter item limits - search not returning any results with large number of elements otherwise ok

    Hi,
    We are working through a problem we've encountered with Azure Search. We are building a filter string based on an "id eq 'xxx' or id eq 'ccc' or id eq 'vvv' etc. The id's are provided in a collection and we loop through, building the string until it's
    ready to apply.
    We are using 2015-02-28 preview at the moment.
    We are encountering a situation where, after approximately 20 id's Azure Search doesn't return any results, nor does it appear to return any error code. I'm pretty sure that the url length is less than 8K.
    Is there any limit on the number of filter elements in a query?

    We followed up offline.
    The symptom in this case was a 200 response with no body. The underlying cause is a URL parsing bug that tries to interpret colons in the query string as the delimiter of a URL scheme (like https:), but with a hard length limit of 1KB. We will work
    on a fix for both the underlying URL parsing issue and the issue that caused it to surface as a body-less 200.
    In the meantime, the workaround is to put colons as close to the beginning of the URL query string as possible. Specifically, putting $filter and facets first, and putting expressions with colons within those first, will mitigate this in most cases.
    Note that the .NET SDK puts $filter and facets near the beginning of the query string by default, so if you're consuming Azure Search you might want to give it a try:
    http://www.nuget.org/packages/Microsoft.Azure.Search/

  • Using cursor to query from data dictionary

    Hi
    I am having a weird problem.
    In a procedure, I like to compress some tables
    so I wrote a cursor like this:
    CURSOR crs IS SELECT OWNER,TABLE_NAME FROM ALL_TABLES WHERE TABLE_NAME LIKE '%ABC';then I'm opening the cursor in for loop and running
    xSql := 'ALTER TABLE ' || X.OWNER || '.' || X.TABLE_NAME || ' MOVE COMPRESS';
    EXECUTE IMMEDIATE (xSql);Now strangely enough, my cursor is not fetching any record at all inside procedure!
    But when I run the cursor select statement of its own in SqlPlus, it does fetch lots of record!
    So, what's wrong??
    Edited by: movilogo on Jun 30, 2010 9:25 AM

    Full code
    CREATE OR REPLACE PROCEDURE COMPRESS_ABC_TABLES IS
        -- to compress tables
        CURSOR crs IS SELECT OWNER,TABLE_NAME FROM ALL_TABLES WHERE TABLE_NAME LIKE '%ABC';
        xSql VARCHAR2(1000);
    BEGIN
        FOR X IN crs
        LOOP
                BEGIN
                    xSql := 'ALTER TABLE ' || X.OWNER || '.' || X.TABLE_NAME || ' MOVE COMPRESS';
                    DBMS_OUTPUT.PUT_LINE(xSql);
                    EXECUTE IMMEDIATE (xSql);
                EXCEPTION
                    WHEN OTHERS THEN
                        DBMS_OUTPUT.PUT_LINE(SQLERRM);
                END;
        END LOOP;
    END;Edited by: movilogo on Jun 30, 2010 9:53 AM

  • Transfer data from Result Set (Execute SQL Task) into Flat File

    Hi to all,
    My problem is this:
     -) I have a stored procedure which use a temporary table in join with another "real" table in select statement. In particular the stored procedure look like this:
    create table #tmp
    col1 varchar(20),
    col2 varchar(50)
    --PUT SOME VALUE IN #TMP
    insert into #tmp
    values ('abc','xyz')
    --SELECT SOME VALUE
    select rt.*
    from realTable rt, #tmp
    where rt.col1 = #tmp.col1
    -) I cannot modify the stored procedure because I'm not admin of database.
    -) I HAVE THE REQUIREMENT OF TRANSFER DATA OF SELECT STATEMENT OF SOTRED PROCEDURE INTO FLAT FILE
    -) THE PROBLEM is that if I use an OLEDB source Task within a Data Flow Task I'm not be able of mapping column from OLEDB source to flat file destination. The reason for this, is that in the "Column page" of OLEDB source task, SSIS do not retrieves
    any column when we using a temporary table. The reason for this, is that SSIS is not be able to retrieve metadata related to temporary table. 
    -) One possible solution to this problem is to use an Execute SQL Task to invoke the stored procedure, store the result returned from stored procedure in a Result Set through a Object type user variable.
    -) The problem now is: How to transfer data from result set to flat file?
    -) Reading here on this forum the solution look be like to use a Script task to transfer data from result set to flat file.
    QUESTIONS: How to transfer data from result set to flat file using a script task?? (I'm not an expert of vb .net) Is it really possible?? P.S.: The number of row returned of stored procedure is very high!
    thanks in advance.

    Hi  Visakh16<abbr
    class="affil"></abbr>
    thanks for the response.
    Your is a good Idea, but I do not have permission to use DDL statement on the database in production environment.

  • Need to configure auto-complete off on large number of workstations

    We currently have a large number of Acrobat Reader installs at our facility on various version levels.  Is there a simple way to push the latest version of Reader to all clients and default to auto-complete off?
    Also, is there a way to lock the auto-complete to the off state?  We are a hospital and must not cache patient information in the forms due to HIPAA regulations.
    Thank you!
    Dave Pyatt
    [signature deleted by host]

    We currently have a large number of Acrobat Reader installs at our facility on various version levels.  Is there a simple way to push the latest version of Reader to all clients and default to auto-complete off?
    Also, is there a way to lock the auto-complete to the off state?  We are a hospital and must not cache patient information in the forms due to HIPAA regulations.
    Thank you!
    Dave Pyatt
    [signature deleted by host]

  • Query tuning-how to reduce physical reads-help me

    1* select * from masterbillingInvoiceview Where SiteIID =300964 and InvoiceID like '%' order by Invoice asc
    SQL> /
    33 rows selected.
    Elapsed: 00:00:00.34
    Execution Plan
    Plan hash value: 352896138
    | Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time |
    | 0 | SELECT STATEMENT | | 16 | 6304 | 341 (3)| 00:00:05 |
    | 1 | SORT ORDER BY | | 16 | 6304 | 341 (3)| 00:00:05 |
    | 2 | TABLE ACCESS BY INDEX ROWID | ADDRESS | 1 | 48 | 2 (0)| 00:00:01
    | 3 | NESTED LOOPS | | 16 | 6304 | 340 (3)| 00:00:05 |
    | 4 | NESTED LOOPS | | 16 | 5536 | 316 (3)| 00:00:04 |
    |* 5 | HASH JOIN OUTER | | 16 | 5168 | 308 (3)| 00:00:04 |
    |* 6 | HASH JOIN | | 16 | 4992 | 302 (3)| 00:00:04 |
    |* 7 | HASH JOIN | | 16 | 4752 | 298 (3)| 00:00:04 |
    | 8 | NESTED LOOPS | | 17 | 4828 | 44 (0)| 00:00:01 |
    | 9 | NESTED LOOPS | | 17 | 3451 | 35 (0)| 00:00:01 |
    | 10 | NESTED LOOPS OUTER | | 1 | 91 | 2 (0)| 00:00:01 |
    | 11 | TABLE ACCESS BY INDEX ROWID| SITEMASTER | 1 | 74 | 1 (0)| 00:00:01 |
    |* 12 | INDEX UNIQUE SCAN | PK_SITEMASTER | 1 | | 1 (0)| 00:00:01 |
    | 13 | TABLE ACCESS BY INDEX ROWID| INVGROUPS | 1735 | 29495 | 1 (0)| 00:00:01 |
    |* 14 | INDEX UNIQUE SCAN | PK_INVGROUPS | 1 | | 1 (0)| 00:00:01 |
    |* 15 | TABLE ACCESS BY INDEX ROWID | INVOICEHEAD | 17 | 1904 | 33 (0)| 00:00:01 |
    |* 16 | INDEX RANGE SCAN | IDX_INVOICEHEADSITEIID | 271 | | 1 (0)| 00:00:01 |
    | 17 | TABLE ACCESS BY INDEX ROWID | CUSTOMER | 1 | 81 | 1 (0)| 00:00:01 |
    |* 18 | INDEX UNIQUE SCAN | PK_CUSTOMER | 1 | | 1 (0)| 00:00:01 |
    | 19 | VIEW | | 35844 | 455K| 254 (3)| 00:00:04 |
    | 20 | HASH GROUP BY | | 35844 | 455K| 254 (3)| 00:00:04 |
    | 21 | TABLE ACCESS FULL | CONTACT | 56100 | 712K| 250 (1)| 00:00:03 |
    | 22 | VIEW | index$_join$_009 | 7 | 105 | 3 (0)| 00:00:01 |
    |* 23 | HASH JOIN | | | | | |
    | 24 | INDEX FAST FULL SCAN | IDX_PAYMENTTERMSID | 7 | 105 | 1 (0)| 00:00:01
    | 25 | INDEX FAST FULL SCAN | PK_PAYMENTTERMS | 7 | 105 | 1 (0)| 00:00:01
    | 26 | VIEW | index$_join$_011 | 1428 | 15708 | 6 (0)| 00:00:01 |
    |* 27 | HASH JOIN | | | | | |
    | 28 | INDEX FAST FULL SCAN | PK_EMPLOYEE | 1428 | 15708 | 3 (0)| 00:00:01 |
    | 29 | INDEX FAST FULL SCAN | IDX_EMPLOYEEEMPLOYEEID | 1428 | 15708 | 3 (0)| 00:00:01
    | 30 | TABLE ACCESS BY INDEX ROWID | CONTACT | 1 | 23 | 1 (0)| 00:00:
    |* 31 | INDEX UNIQUE SCAN | PK_CONTACT | 1 | | 1 (0)| 00:00:01 |
    |* 32 | INDEX RANGE SCAN | PK_ADDRESS | 1 | | 1 (0)| 00:00:01 |
    Predicate Information (identified by operation id):
    5 - access("A"."CREATEDBY"="J"."EMPIID"(+))
    6 - access("A"."TERMSIID"="H"."PAYMENTTERMSIID")
    7 - access("D"."CUSTOMERIID"="F"."CUSTOMERIID")
    12 - access("B"."SITEIID"=300964)
    14 - access("B"."PARENTIID"="I"."GROUPIID"(+))
    15 - filter("A"."TYPE"=4 AND "A"."INVOICEID" LIKE '%')
    16 - access("A"."SITEIID"=300964)
    18 - access("A"."BILLINGCUSTOMERIID"="D"."CUSTOMERIID")
    23 - access(ROWID=ROWID)
    27 - access(ROWID=ROWID)
    31 - access("F"."CONTACTIID"="G"."CONTACTIID")
    32 - access("E"."ADDRESSIID"=NVL("D"."BILLINGADDRESSIID","D"."ADDRESSIID"))
    Statistics
    107 recursive calls
    0 db block gets
    2819 consistent gets
    0 physical reads
    0 redo size
    6586 bytes sent via SQL*Net to client
    356 bytes received via SQL*Net from client
    4 SQL*Net roundtrips to/from client
    1 sorts (memory)
    0 sorts (disk)
    33 rows processed
    SQL> ed
    Wrote file afiedt.buf
    1* select * from masterbillingInvoiceview Where SiteIID =300964 and InvoiceID like '%%%' order by Invoice asc
    SQL> /
    33 rows selected.
    Elapsed: 00:06:15.23
    Execution Plan
    Plan hash value: 1828716447
    | Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time |
    | 0 | SELECT STATEMENT | | 2 | 964 | 295 (3)| 00:00:04 |
    |* 1 | FILTER | | | | | |
    | 2 | SORT GROUP BY | | 2 | 964 | 295 (3)| 00:00:04 |
    | 3 | MERGE JOIN CARTESIAN | | 72315 | 33M| 291 (1)| 00:00:04 |
    | 4 | TABLE ACCESS BY INDEX ROWID | ADDRESS | 1 | 52 | 2 (0)| 00:00:01 |
    | 5 | NESTED LOOPS | | 1 | 447 | 41 (0)| 00:00:01 |
    | 6 | NESTED LOOPS | | 1 | 395 | 40 (0)| 00:00:01 |
    | 7 | NESTED LOOPS | | 1 | 382 | 38 (0)| 00:00:01 |
    | 8 | NESTED LOOPS OUTER | | 1 | 289 | 37 (0)| 00:00:01 |
    | 9 | NESTED LOOPS | | 1 | 266 | 36 (0)| 00:00:01 |
    | 10 | NESTED LOOPS | | 1 | 239 | 35 (0)| 00:00:01 |
    | 11 | NESTED LOOPS OUTER | | 1 | 115 | 2 (0)| 00:00:01 |
    | 12 | TABLE ACCESS BY INDEX ROWID| SITEMASTER | 1 | 86 | 1 (0)| 00:00:01 |
    |* 13 | INDEX UNIQUE SCAN | PK_SITEMASTER | 1 | | 1 (0)| 00:00:01 |
    | 14 | TABLE ACCESS BY INDEX ROWID| INVGROUPS | 1735 | 50315 | 1 (0)| 00:00:01 |
    |* 15 | INDEX UNIQUE SCAN | PK_INVGROUPS | 1 | | 1 (0)| 00:00:01 |
    |* 16 | TABLE ACCESS BY INDEX ROWID | INVOICEHEAD | 1 | 124 | 33 (0)| 00:00:01 |
    |* 17 | INDEX RANGE SCAN | IDX_INVOICEHEADSITEIID | 271 | | 1 (0)| 00:00:01 |
    | 18 | TABLE ACCESS BY INDEX ROWID | PAYMENTTERMS | 1 | 27 | 1 (0)| 00:00:01 |
    |* 19 | INDEX UNIQUE SCAN | PK_PAYMENTTERMS | 1 | | 1 (0)| 00:00:01 |
    | 20 | TABLE ACCESS BY INDEX ROWID | EMPLOYEE | 1 | 23 | 1 (0)| 00:00:01 |
    |* 21 | INDEX UNIQUE SCAN | PK_EMPLOYEE | 1 | | 1 (0)| 00:00:01 |
    | 22 | TABLE ACCESS BY INDEX ROWID | CUSTOMER | 1 | 93 | 1 (0)| 00:00:01 |
    |* 23 | INDEX UNIQUE SCAN | PK_CUSTOMER | 1 | | 1 (0)| 00:00:01 |
    | 24 | TABLE ACCESS BY INDEX ROWID | CONTACT | 1 | 13 | 2 (0)| 00:00:01 |
    |* 25 | INDEX RANGE SCAN | IDX_CONTACTCUSTOMERIID | 2 | | 1 (0)| 00:00:01 |
    |* 26 | INDEX RANGE SCAN | PK_ADDRESS | 1 | | 1 (0)| 00:00:01 |
    | 27 | BUFFER SORT | | 56100 | 1917K| 294 (3)| 00:00:04 |
    | 28 | TABLE ACCESS FULL | CONTACT | 56100 | 1917K| 250 (1)| 00:00:03 |
    Predicate Information (identified by operation id):
    1 - filter("G"."CONTACTIID"=MAX("CONTACTIID"))
    13 - access("B"."SITEIID"=300964)
    15 - access("B"."PARENTIID"="I"."GROUPIID"(+))
    16 - filter("A"."TYPE"=4 AND "A"."INVOICEID" LIKE '%%%')
    17 - access("A"."SITEIID"=300964)
    19 - access("A"."TERMSIID"="H"."PAYMENTTERMSIID")
    21 - access("A"."CREATEDBY"="J"."EMPIID"(+))
    23 - access("A"."BILLINGCUSTOMERIID"="D"."CUSTOMERIID")
    25 - access("D"."CUSTOMERIID"="CUSTOMERIID")
    filter("CUSTOMERIID" IS NOT NULL)
    26 - access("E"."ADDRESSIID"=NVL("D"."BILLINGADDRESSIID","D"."ADDRESSIID"))
    Statistics
    1952 recursive calls
    78 db block gets
    4649 consistent gets
    236151 physical reads
    0 redo size
    6586 bytes sent via SQL*Net to client
    356 bytes received via SQL*Net from client
    4 SQL*Net roundtrips to/from client
    1 sorts (memory)
    1 sorts (disk)
    33 rows processed
    this is the execution plan of my one query with a small difference, bt there is large diffference in physical reads..
    can anyone help me out of this
    thanks
    aju

    Hi,
    Can you please format your explain plan using
    { code } --without space
    Explain plan
    { code } -- without any space
    What is your DB version?
    There are differences in access and filter criteria...
    -- FRIST QUERY
    5 - access("A"."CREATEDBY"="J"."EMPIID"(+))
    6 - access("A"."TERMSIID"="H"."PAYMENTTERMSIID")
    7 - access("D"."CUSTOMERIID"="F"."CUSTOMERIID")
    12 - access("B"."SITEIID"=300964)
    14 - access("B"."PARENTIID"="I"."GROUPIID"(+))
    15 - filter("A"."TYPE"=4 AND "A"."INVOICEID" LIKE '%')
    16 - access("A"."SITEIID"=300964)
    18 - access("A"."BILLINGCUSTOMERIID"="D"."CUSTOMERIID")
    23 - access(ROWID=ROWID)
    27 - access(ROWID=ROWID)
    31 - access("F"."CONTACTIID"="G"."CONTACTIID")
    32 - access("E"."ADDRESSIID"=NVL("D"."BILLINGADDRESSIID","D"."ADDRESSIID"))
    ------SECOND QUERY
    1 - filter("G"."CONTACTIID"=MAX("CONTACTIID"))
    13 - access("B"."SITEIID"=300964)
    15 - access("B"."PARENTIID"="I"."GROUPIID"(+))
    16 - filter("A"."TYPE"=4 AND "A"."INVOICEID" LIKE '%%%')
    17 - access("A"."SITEIID"=300964)
    19 - access("A"."TERMSIID"="H"."PAYMENTTERMSIID")
    21 - access("A"."CREATEDBY"="J"."EMPIID"(+))
    23 - access("A"."BILLINGCUSTOMERIID"="D"."CUSTOMERIID")
    25 - access("D"."CUSTOMERIID"="CUSTOMERIID")
    filter("CUSTOMERIID" IS NOT NULL)
    26 - access("E"."ADDRESSIID"=NVL("D"."BILLINGADDRESSIID","D"."ADDRESSIID"))-Avinash

  • Large number of attributes reduces data display in Query (PLS HELP).

    Dear Guru's
    We have newly upgraded our system to 7.0, Currently i have a Query based on master data info object(0 equipment) which contain several attributes, due to huge number of attributes, i cannot see data for few  attributes( Ex:-0service descrip), but data exists for these atrrib, when i check in RSA1. my client want to use heavy number of 0Euipment numbers(nearly 10000) in selection screen. thats the reason for data not to be seen.
    WE are under SP 16 also(Note 562162 - Additional selection of attribute reduces query result set is allready exists), and did not found any proper note.
    So guru's can any one please tell me how i can get the data in Query for these attribute by keeping the same selection option for Equipment numbers (10000).
    Dude's give me some good suggestion please...client is waiting from 1 week
    Thanks in adv
    Dev
    Edited by: Srinivas dev on Jun 10, 2008 11:52 AM

    Dear Yogesh,
    Thanks a lot dude...
    I assigned points for your kindly help....
    Got any idea about by initial post   Large number of attributes reduces data display in Query
    Thanks and Regards
    Dev

  • How to handle a large number of query parameters for a Browse screen

    I need to implement an advanced search functionality in a browse screen for a large table.  The table has 80+ columns and therefore will have a large number of possible query parameters.  The screen will be built on a modeled query with all
    of the parameters marked as optional.  Given the large number of parameters, I am thinking that it would be better to use a separate screen to receive the parameter input from the user, rather than a Popup.  Is it possible for example to have a search
    button on the browse screen (screen a) open a new screen (screen b) that contains all of the search parameters, have the user enter the parameters they want, then click a button to send all of the parameters back to screen a where the query is executed and
    the search results are returned to the table control?  This would effectively make screen b an advanced modal window for screen a.  In addition, if the user were to execute the query, then want to change a parameter, they would need to be able to
    re-open screen b and have all of their original parameters still set.  How would you implement this, or otherwise deal with a large number of optional query parameters in the html client?  My initial thinking is to store all of the parameters in
    an object and use beforeShown/afterClosed to pass them between the screens, but I'm not quite sure how to make that work.  TIA

    Wow Josh, thanks.  I have a lot of reading to do.  What I ultimately plan to do with this (my other posts relate to this too), is have a separate screen for advanced filtering that also allows the user to save their queries if desired. 
    There is an excellent way to get at all of the query information in the Query_Executed() method.  I just put an extra Boolean parameter in the query called "SaveQuery" and when true, the Query_Executed event triggers an entry into a table with
    the query name, user name, and parameter value pairs that the user entered.  Upon revisiting the screen, I want the user to be able to select from their saved queries and load all the screen parameters (screen properties) from their selected query. 
    I almost have it working.  It may be as easy as marking all of the screen properties that are query parameters as screen parameters (not required), then passing them in from the saved query data (filtered by username, queryname, and selected
    item).  I'll post an update once I get it.  Probably will have some more questions as I go through it.  Thanks again! 

  • I have a large number of photos imported into iPhoto with the dates wrong.  How can I adjust multiple photos (with varying dates) to the same, correct, date?

    I have a large number of photos imported into iPhoto with the dates wrong.  How can I adjust multiple photos (with varying dates) to the same, correct, date?

    If I understand you correctly, when you enter a date in the Adjust Date and Time window, the picture does not update with the date you enter.  If that is the case then something is wrong with iPhoto or your perhaps your library.
    How large a date change are you putting in?  iPhoto currently has an issue with date changes beyond about 60 years at a time.  If the difference between the current date on the image and the date you are entering is beyond that range that may explain why this is not working.
    If that is not the case:
    Remove the following to the trash and restart the computer and try again:
    Home > Library > Caches > com.apple.iphoto
    Home > Library > Preferences > com.apple.iPhoto (There may be more than one. Remove them all.)
    ---NOTE: to get to the "home > library" hold down option on the keyboard and click on "Go" > "Library" while in the Finder.
    Let me know the results.

  • How to calculate the area of a large number of polygons in a single query

    Hi forum
    Is it possible to calculate the area of a large number of polygons in a single query using a combination of SDO_AGGR_UNION and SDO_AREA? So far, I have tried doing something similar to this:
    select sdo_geom.sdo_area((
    select sdo_aggr_union (   sdoaggrtype(mg.geoloc, 0.005))
    from mapv_gravsted_00182 mg 
    where mg.dblink = 521 or mg.dblink = 94 or mg.dblink = 38 <many many more....>),
    0.0005) calc_area from dualThe table MAPV_GRAVSTED_00182 contains 2 fields - geoloc (SDO_GEOMETRY) and dblink (Id field) needed for querying specific polygons.
    As far as I can see, I need to first somehow get a single SDO_GEOMETRY object and use this as input for the SDO_AREA function. But I'm not 100% sure, that I'm doing this the right way. This query is very inefficient, and sometimes fails with strange errors like "No more data to read from socket" when executed from SQL Developer. I even tried with the latest JDBC driver from Oracle without much difference.
    Would a better approach be to write some kind of stored procedure, that adds up all the single geometries by adding each call to SDO_AREA on each single geometry object - or what is the best approach?
    Any advice would be appreciated.
    Thanks in advance,
    Jacob

    Hi
    I am now trying to update all my spatial table with SRID's. To do this, I try to drop the spatial index first to recreate it after the update. But for a lot of tables I can't drop the spatial index. Whenever I try to DROP INDEX <spatial index name>, I get this error - anyone know what this means?
    Thanks,
    Jacob
    Error starting at line 2 in command:
    drop index BSSYS.STIER_00182_SX
    Error report:
    SQL Error: ORA-29856: error occurred in the execution of ODCIINDEXDROP routine
    ORA-13249: Error in Spatial index: cannot drop sequence BSSYS.MDRS_1424B$
    ORA-13249: Stmt-Execute Failure: DROP SEQUENCE BSSYS.MDRS_1424B$
    ORA-29400: data cartridge error
    ORA-02289: sequence does not exist
    ORA-06512: at "MDSYS.SDO_INDEX_METHOD_10I", line 27
    29856. 00000 - "error occurred in the execution of ODCIINDEXDROP routine"
    *Cause:    Failed to successfully execute the ODCIIndexDrop routine.
    *Action:   Check to see if the routine has been coded correctly.
    Edit - just found the answer for this in MetaLink note 241003.1. Apparently there is some internal problem when dropping spatial indexes, some objects gets dropped that shouldn't be. Solution is to manually create the sequence it complains it can't drop, then it works... Weird error.

  • Selection and data binding in a large BW query...

    Hi,
    I've created a dashbaord using Present 2008. It uses a BW query that returns a fairly large amount of data (around 9K-10K rows in Excel). In the initial version of the dashboard I was using a map with individual pushbuttons on it representing locations for my corps "offices" (reduced down to large "dots'). The dashboard displays quite a large number of metrics so for each location the query returns a block of data around 50 rows by 12 columns from which I pull each metric and display it, typically, in either a chart or the "value" component.
    The problem - binding data to each button this way is easy. When the user pushes the button, I select the 50 X 12 area as source and a similar area for the destination. If I wanted to move to using a different selector (like a combobox), e.g. the map isn't greatest metaphor for some metrics, how do I bind that much data to one of the label-based selectors? I've tried several different ways to do this but XCelsius seems to balk at using that amount of data. I tried using filtered rows by labeling each location and it seemed it limited the amount of data to around 500 rows or so.
    So, I am just going about this wrong way or misunderstanding how to bind the label-based selectors to that amount of data or is this not possible?
    Please bear in mind I'm relatively new to XCelsius and have only been using it off and on for a few weeks and I welcome any attempts to alleviate my ignorance.

    "click="{requestNode(rp.currentItem.id);}" will not work
    because the "currentItem" only exists when the repeater is
    rendering and not when the click event happens. Try this:
    click="{requestNode(rp.getRepeaterItem().id);}"
    If that does not work like it is, look up getRepeaterItem in
    the docs, I may not have the syntax exactly right.
    Tracy

  • Fastest Way to Delete a large number of data

    I m trying to delete a large number of Data
    this Proc will be running every day without disturbing the other Process
    SET sql = 'DELETE TOP (100) FROM ' + @TableName + ' WHERE Id IN (Select Distinct Id FROM [state] WHERE ToDel = 1)'SET @Deleted_Rows = 1;SET @Rows = 0; WHILE (@Deleted_Rows > 0)BEGIN BEGIN TRANEXEC sp_executesql @sql SET @Deleted_Rows = (SELECT @@ROWCOUNT);SET @Rows = (@Rows + @Deleted_Rows);COMMIT TRANWaitFor DELAY '00:00:00:01' END
    Have you any Idea how can I optimize my Query ?
    Thank you very much

    Hi, following are few thoughts.
    1. If you can manage without dynamic sql then it may yield better performance as except table name there is no need for it as per your query.
    2. Ensure indexes are in place.
    3. If the table in question is not participant in replication then you may slightly go for higher batch size and again it depends on volume.
    Please take a look at the tweaked code.
    declare @Deleted_Rows int, @Rows int;
    SET @Deleted_Rows = 1;
    SET @Rows = 0;
    WHILE (@Deleted_Rows > 0)
    BEGIN
    BEGIN TRAN;
    DELETE TOP (100) t1
    FROM
    tabName as t1
    WHERE exists (Select * FROM [state] as t2 WHERE t2.ToDel = 1 and t2.Id = t1.id);
    SET @Deleted_Rows = @@ROWCOUNT ;
    SET @Rows = @Rows + @Deleted_Rows;
    COMMIT TRAN;
    WaitFor DELAY '00:00:00:01';
    END
    Print 'Total rows deleted : '+convert(varchar(50), @Rows);

Maybe you are looking for

  • How to find rubber feet for mbp 13"?

    I've lost one of the rubber feet on the bottom of my macbook pro 13" 2.6 GHz core duo. What can I do please? Can anyone inform me, so that I can find a solution?

  • 8610e - can Ipad switch mono/colour before printing?

    Hi, HP 8610e printer and Ipad 2 When I print the only options available are to change the number of copies and single/double sided printing. It prints in colours and there is no option shown on the ipad for switching to mono. Is there a way to do it

  • Reducing WSDL Size

    Hi, Is there a way in teo to control which processes end up in the wsdl for the Northbound Web Service?  It seems like this should be an option on a process by process basis but I can't seem to find it. As it is the wsdl is pretty large with a lot of

  • How do I change the volume in the middle of a clip up and down?

    I need to make my questions I ask during the interview quieter in my Final Cut x. The man I was interviewing was pretty near to deaf so for him to hear I almost yell the questions. I remember that somehow I need to put start and stop makers on the vo

  • Trouble importing Axis packages

    Hi, I am trying to deploy a test web services implementation on JBoss 3.2.1 and Axis 1.1RC2. The problem is, I cannot even compile my client code (the one that invoke the web services)...it seems that the java compiler cannot locate the following pac