Query on a table runs more than 45mins(after stats) and same query runs 19secs(before stats - rebuild)

Query on a table runs more than 45mins(after stats) and same query runs 19secs(before stats - rebuild) - Not sure what the cause is.
- Analysed the explain the plan
- different explain plan used afterr stats gather.
Any idea what could be the cause with this kind of difference.
Thank you!

What is the difference you see in the explain plan ? Where it spends most time. All these needs to be analysed.

Similar Messages

  • Running more than one batch on the same resource at the same time.

    Dear all,
    as i'm working in paper industry , i have to release more than one batch at the same time using batchable resource , but i have problem in declaring the resource usage , if i run 3 batches on the same resource for i hour , what is the resource usage for each batch and how the system will help me to know that.
    Regards.

    Hi
    My question is If I run 3 batches on the same resoures......what do you mean
    Assume you have 3 batches names b001/boo2/boo3
    for B001 you have planned resoues as 1, but Actual resouce count is 2
    for B002 you have planned resourse as 3 but Actual resource count is 3
    for B003 you have planned resourse as 2 but actual resouree count is 2
    Here the the Acutal resource usage will calculate as per the actual count and that too within Batch only. And it will not consolidte Batches altogehter.
    Rgds
    Raj

  • When will this thing run more than one app at the same time?

    When will we be able to run more than one app at a time?

    As has been suggested, we won't know about future features until they announce them, but by understanding the current situation we can guess.
    The current limitations on third-party applications running in the background make sense. It's not a technical limitation. App store downloads, and the phone functionality run in the background while other applications are up front.
    It's a resource limitation. Processor power and probably more importantly battery life are the key limiting factors. It's not that Apple doesn't want you to have more functionality. It's that it currently struggling to deliver a quality experience that satisfies as many customers as possible. As it is the battery life is only just acceptable enough for most current customers, many are unhappy with it.
    So if more efficient chips or batteries become available, or some genius programmer manages to squeeze out a whole lot more out of less, then we'll see more background processes allowed.

  • How can I run more than one VST on a PC slave running Logic 5.5?

    I have moved from PC to Intel Mac (good!) and wish to run several non Intel VST apps from a slave PC running Logic 5.5 (ie Trilogy, Atmosphere).
    My question is (and forgive my stupidity if the answer is obvious) how can I address each virtual instrument from the Mac? Currently I operate MidioverLAN CP which enables my PC computer to receive midi but I seem only to be able to access one instrument at a time, not permitting me to run several apps at once.
    many thanks in advance for any kind help

    Logic 5.5 is not an ideal VST host, as by default, only one virtual instrument is in Live mode, and you can't rig up environment solutions without a large latency (which makes it unusable).
    There is a workaround involving putting a bypassed IO plugin on your instrument tracks, and then you can use environment cabling without the large latency (it tricks Logic into putting all instruments in Live mode). Then you'd cable your physical input through a channel splitter and onto the audio instruments to assign each a MIDI channel as has been discussed to death on this forum.
    If there isn't a reason you must use Logic 5.5 PC to do this, then there are much better PC VST hosting apps and I recommend looking into them...

  • Actionscript 3 Class Not Running More Than Once After Using The GoToAndStop function

    I'm having a strange issue, and I am sure it's due to my lack of knowledge about how classes fully work in Adobe AIR (Flash CC). I've created a login button that is connected to a function, and based on a set of conditions it will take you to a different frame with the GoToAndStop function. I've created a logout button that is suppose to take you "back" to the login frame (that has the login button) if you click the logout button using the GoToAndStop function. Up to this point, everything works until I try to click the Login button again, it does not do anything. It's like it does not even try to execute my class file anymore.
    Does the class file only run at startup and never runs again, and if so, how can I create a section of code that is constantly monitoring what the user is doing no matter what frame you are on and no matter if I use the GoToAndStop function or not?
    Thanks

    Based on your qyestion, I think I was wrong with my second guess.
    When I want to create a button, what I do is draw a bitmap that looks like a button and then use a mouse event listener that looks for any part of the stage that is clicked.  You would use the same code for touch as well.  As3 will return the x.y. coordinate of the area that was clicked or touched.  When the xy coordinate is known, if it corrseponds to the area where the button is, I code the appropriate response.  Hence, I only use one event listener for all my buttons, as long as thet involve simple clicks or taps of the screen.
    Good luck, Justin

  • Row chaining in table with more than 255 columns

    Hi,
    I have a table with 1000 columns.
    I saw the following citation: "Any table with more then 255 columns will have chained
    rows (we break really wide tables up)."
    If I insert a row populated with only the first 3 columns (the others are null), is a row chaining occurred?
    I tried to insert a row described above and no row chaining occurred.
    As I understand, a row chaining occurs in a table with 1000 columns only when the populated data increases
    the block size OR when more than 255 columns are populated. Am I right?
    Thanks
    dyahav

    user10952094 wrote:
    Hi,
    I have a table with 1000 columns.
    I saw the following citation: "Any table with more then 255 columns will have chained
    rows (we break really wide tables up)."
    If I insert a row populated with only the first 3 columns (the others are null), is a row chaining occurred?
    I tried to insert a row described above and no row chaining occurred.
    As I understand, a row chaining occurs in a table with 1000 columns only when the populated data increases
    the block size OR when more than 255 columns are populated. Am I right?
    Thanks
    dyahavYesterday, I stated this on the forum "Tables with more than 255 columns will always have chained rows." My statement needs clarification. It was based on the following:
    http://download.oracle.com/docs/cd/B28359_01/server.111/b28318/schema.htm#i4383
    "Oracle Database can only store 255 columns in a row piece. Thus, if you insert a row into a table that has 1000 columns, then the database creates 4 row pieces, typically chained over multiple blocks."
    And this paraphrase from "Practical Oracle 8i":
    V$SYSSTAT will show increasing values for CONTINUED ROW FETCH as table rows are read for tables containing more than 255 columns.
    Related information may also be found here:
    http://download.oracle.com/docs/cd/B10501_01/server.920/a96524/c11schem.htm
    "When a table has more than 255 columns, rows that have data after the 255th column are likely to be chained within the same block. This is called intra-block chaining. A chained row's pieces are chained together using the rowids of the pieces. With intra-block chaining, users receive all the data in the same block. If the row fits in the block, users do not see an effect in I/O performance, because no extra I/O operation is required to retrieve the rest of the row."
    http://download.oracle.com/docs/html/B14340_01/data.htm
    "For a table with several columns, the key question to consider is the (average) row length, not the number of columns. Having more than 255 columns in a table built with a smaller block size typically results in intrablock chaining.
    Oracle stores multiple row pieces in the same block, but the overhead to maintain the column information is minimal as long as all row pieces fit in a single data block. If the rows don't fit in a single data block, you may consider using a larger database block size (or use multiple block sizes in the same database). "
    Why not a test case?
    Create a test table named T4 with 1000 columns.
    With the table created, insert 1,000 rows into the table, populating the first 257 columns each with a random 3 byte string which should result in an average row length of about 771 bytes.
    SPOOL C:\TESTME.TXT
    SELECT
      SN.NAME,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    INSERT INTO T4 (
    COL1,
    COL2,
    COL3,
    COL255,
    COL256,
    COL257)
    SELECT
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3)
    FROM
      DUAL
    CONNECT BY
      LEVEL<=1000;
    SELECT
      SN.NAME,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    SET AUTOTRACE TRACEONLY STATISTICS
    SELECT
    FROM
      T4;
    SET AUTOTRACE OFF
    SELECT
      SN.NAME,
      SN.STATISTIC#,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    SPOOL OFFWhat are the results of the above?
    Before the insert:
    NAME                      VALUE                                                
    table fetch continue        166
    After the insert:
    NAME                      VALUE                                                
    table fetch continue        166                                                
    After the select:
    NAME                 STATISTIC#      VALUE                                     
    table fetch continue        252        332  Another test, this time with an average row length of about 12 bytes:
    DELETE FROM T4;
    COMMIT;
    SPOOL C:\TESTME2.TXT
    SELECT
      SN.NAME,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    INSERT INTO T4 (
      COL1,
      COL256,
      COL257,
      COL999)
    SELECT
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3)
    FROM
      DUAL
    CONNECT BY
      LEVEL<=100000;
    SELECT
      SN.NAME,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    SET AUTOTRACE TRACEONLY STATISTICS
    SELECT
    FROM
      T4;
    SET AUTOTRACE OFF
    SELECT
      SN.NAME,
      SN.STATISTIC#,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    SPOOL OFFWith 100,000 rows each containing about 12 bytes, what should the 'table fetch continued row' statistic show?
    Before the insert:
    NAME                      VALUE                                                
    table fetch continue        332 
    After the insert:
    NAME                      VALUE                                                
    table fetch continue        332
    After the select:
    NAME                 STATISTIC#      VALUE                                     
    table fetch continue        252      33695The final test only inserts data into the first 4 columns:
    DELETE FROM T4;
    COMMIT;
    SPOOL C:\TESTME3.TXT
    SELECT
      SN.NAME,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    INSERT INTO T4 (
      COL1,
      COL2,
      COL3,
      COL4)
    SELECT
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3),
    DBMS_RANDOM.STRING('A',3)
    FROM
      DUAL
    CONNECT BY
      LEVEL<=100000;
    SELECT
      SN.NAME,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    SET AUTOTRACE TRACEONLY STATISTICS
    SELECT
    FROM
      T4;
    SET AUTOTRACE OFF
    SELECT
      SN.NAME,
      SN.STATISTIC#,
      MS.VALUE
    FROM
      V$MYSTAT MS,
      V$STATNAME SN
    WHERE
      SN.NAME = 'table fetch continued row'
      AND SN.STATISTIC#=MS.STATISTIC#;
    SPOOL OFFWhat should the 'table fetch continued row' show?
    Before the insert:
    NAME                      VALUE                                                
    table fetch continue      33695
    After the insert:
    NAME                      VALUE                                                
    table fetch continue      33695
    After the select:
    NAME                 STATISTIC#      VALUE                                     
    table fetch continue        252      33695 My statement "Tables with more than 255 columns will always have chained rows." needs to be clarified:
    "Tables with more than 255 columns will always have chained rows +(row pieces)+ if a column beyond column 255 is used, but the 'table fetch continued row' statistic +may+ only increase in value if the remaining row pieces are found in a different block."
    Charles Hooper
    IT Manager/Oracle DBA
    K&M Machine-Fabricating, Inc.
    Edited by: Charles Hooper on Aug 5, 2009 9:52 AM
    Paraphrase misspelled the view name "V$SYSSTAT", corrected a couple minor typos, and changed "will" to "may" in the closing paragraph as this appears to be the behavior based on the test case.

  • More than one reports at the same time?

    Hello
    Is there any way that i can call more than one reports at the same time through the .FMX file. At the moment i can call only one report, if i need to call another report then i have to close the 1st report, then it will show me the 2nd report. Report Background Engine can run more than one reports at the same time?
    Nametullah kalair

    Yes i know, we can submit multiple - asynchronous- jobs, but the problem is, is not possible to open more than one previewer to see multiple reports at the same time for analysis.
    Regards,
    Nametullah kalair

  • Error while running spatial queries on a table with more than one geometry.

    Hello,
    I'm using GeoServer with Oracle Spatial database, and this is a second time I run into some problems because we use tables with more than one geometry.
    When GeoServer renders objects with more than one geometry on the map, it creates a query where it asks for objects which one of the two geometries interacts with the query window. This type of query always fails with "End of TNS data channel" error.
    We are running Oracle Standard 11.1.0.7.0.
    Here is a small script to demonstrate the error. Could anyone confirm that they also have this type of error? Or suggest a fix?
    What this script does:
    1. Create table object1 with two geometry columns, geom1, geom2.
    2. Create metadata (projected coordinate system).
    3. Insert a row.
    4. Create spacial indices on both columns.
    5. Run a SDO_RELATE query on one column. Everything is fine.
    6. Run a SDO_RELATE query on both columns. ERROR: "End of TNS data channel"
    7. Clean.
    CREATE TABLE object1
    id NUMBER PRIMARY KEY,
    geom1 SDO_GEOMETRY,
    geom2 SDO_GEOMETRY
    INSERT INTO user_sdo_geom_metadata (table_name, column_name, srid, diminfo)
    VALUES
    'OBJECT1',
    'GEOM1',
    2180,
    SDO_DIM_ARRAY
    SDO_DIM_ELEMENT('X', 400000, 700000, 0.05),
    SDO_DIM_ELEMENT('Y', 300000, 600000, 0.05)
    INSERT INTO user_sdo_geom_metadata (table_name, column_name, srid, diminfo)
    VALUES
    'OBJECT1',
    'GEOM2',
    2180,
    SDO_DIM_ARRAY
    SDO_DIM_ELEMENT('X', 400000, 700000, 0.05),
    SDO_DIM_ELEMENT('Y', 300000, 600000, 0.05)
    INSERT INTO object1 VALUES(1, SDO_GEOMETRY(2001, 2180, SDO_POINT_TYPE(500000, 400000, NULL), NULL, NULL), SDO_GEOMETRY(2001, 2180, SDO_POINT_TYPE(550000, 450000, NULL), NULL, NULL));
    CREATE INDEX object1_geom1_sidx ON object1(geom1) INDEXTYPE IS MDSYS.SPATIAL_INDEX;
    CREATE INDEX object1_geom2_sidx ON object1(geom2) INDEXTYPE IS MDSYS.SPATIAL_INDEX;
    SELECT *
    FROM object1
    WHERE
    SDO_RELATE("GEOM1", SDO_GEOMETRY(2001, 2180, SDO_POINT_TYPE(500000, 400000, NULL), NULL, NULL), 'MASK=ANYINTERACT') = 'TRUE';
    SELECT *
    FROM object1
    WHERE
    SDO_RELATE("GEOM1", SDO_GEOMETRY(2001, 2180, SDO_POINT_TYPE(500000, 400000, NULL), NULL, NULL), 'MASK=ANYINTERACT') = 'TRUE' OR
    SDO_RELATE("GEOM2", SDO_GEOMETRY(2001, 2180, SDO_POINT_TYPE(500000, 400000, NULL), NULL, NULL), 'MASK=ANYINTERACT') = 'TRUE';
    DELETE FROM user_sdo_geom_metadata WHERE table_name = 'OBJECT1';
    DROP INDEX object1_geom1_sidx;
    DROP INDEX object1_geom2_sidx;
    DROP TABLE object1;
    Thanks for help.

    This error appears in GeoServer and SQLPLUS.
    I have set up a completly new database installation to test this error and everything works fine. I tried it again on the previous database but I still get the same error. I also tried to restart the database, but with no luck, the error is still there. I geuss something is wrong with the database installation.
    Anyone knows what could cause an error like this "End of TNS data channel"?

  • Query with trim function running more than 15 hours..

    Hi,
    I have one select query which is running more than 15 hours.I have many trim() functions in that.and i have index for some columns for the tables that i used in that query.but it is not considering those indexes..
    select
    DISTINCT
    trim(A.x) x,
    trim(B.y) y,....... from A,B
    Can anyone tell why this query is taking more time?and is there any other way for using the trim function?
    Thanks in advance..

    Note the name of this forum is "SQL Developer *(Not for general SQL/PLSQL questions)*", so only for issues with the SQL Developer tool. Please post these questions under the dedicated [SQL And PL/SQL|https://forums.oracle.com/forums/forum.jspa?forumID=75] forum.
    Regards,
    K.

  • Query which the Reports which not run more than 15 days

    Hello,
    I need I query to show all the Concurrent Programs which are attached in Application but not run more than 15 days. Also filter will segregate concurrent Reports and concurrent processes.
    regards,

    Please clarify are you interested in conc programs/reports which has not been run in last 15 days or the ones which were run only in the first 15 days from creation date.
    I am pasting the base query which gives you the details like concurrent program name, executable name, executable type, and execuation file name ( you may add condition like fcp.enabled_flag = 'Y' for all active programs etc)
    SELECT  fe.executable_name,
            fe.execution_file_name,
            fl.meaning,
            fcp.concurrent_program_name,
            fcpt.USER_CONCURRENT_PROGRAM_NAME
    FROM    apps.fnd_executables fe,
            apps.fnd_lookups     fl,
            apps.fnd_concurrent_programs fcp,
            apps.fnd_concurrent_programs_tl fcpt
    WHERE  fcp.executable_id = fe.executable_id
    AND    fe.execution_method_code = fl.lookup_code
    AND    fl.lookup_type =  'CP_EXECUTION_METHOD_CODE'
    AND    fcp.concurrent_program_id = fcpt.concurrent_program_idEdited by: 936671 on Sep 4, 2012 5:58 AM

  • Analyse a partitioned table with more than 50 million rows

    Hi,
    I have a partitioned table with more than 50 million rows. The last analyse is on 1/25/2007. Do I need to analyse him? (query runs on this table is very slow).
    If I need to analyse him, what is the best way? Use DBMS_STATS and schedule a job?
    Thanks

    A partitioned table has global statistics as well as partition (and subpartition if the table is subpartitioned) statistics. My guess is that you mean to say that the last time that global statistics were gathered was in 2007. Is that guess accurate? Are the partition-level statistics more recent?
    Do any of your queries actually use global statistics? Or would you expect that every query involving this table would specify one or more values for the partitioning key and thus force partition pruning to take place? If all your queries are doing partition pruning, global statistics are irrelevant, so it doesn't matter how old and out of date they are.
    Are you seeing any performance problems that are potentially attributable to stale statistics on this table? If you're not seeing any performance problems, leaving the statistics well enough alone may be the most prudent course of action. Gathering statistics would only have the potential to change query plans. And since the cost of a query plan regressing is orders of magnitude greater than the benefit of a different query performing faster (at least for most queries in most systems), the balance of risks would argue for leaving the stats alone if there is no problem you're trying to solve.
    If your system does actually use global statistics and there are performance problems that you believe are potentially attributable to stale global statistics and your partition level statistics are accurate, you can gather just global statistics on the table probably with a reasonably small sample size. Make sure, though, that you back up your existing statistics just in case a query plan goes south. Ideally, you'd also have a test environment with identical (or nearly identical) data volumes that you could use to verify that gathering statistics doesn't cause any problems.
    Justin

  • Regarding the thread running more than 24 hours

    The below is the query executed from oracle side and is running more than 24 hours and not yet over. And multiple threads are running.
    SELECT COUNT(*)
    FROM snurk_cmms_csht009_rfs,
    snurk_cmms_csht008_rfs,
    snurk_cmms_csht001_rfs
    WHERE snurk_cmms_csht001_rfs.cd_plant = snurk_cmms_csht008_rfs.cd_plant (+)
    AND snurk_cmms_csht001_rfs.no_rfs = snurk_cmms_csht008_rfs.no_rfs (+)
    AND snurk_cmms_csht008_rfs.cd_plant = snurk_cmms_csht009_rfs.cd_plant (+)
    AND snurk_cmms_csht008_rfs.no_rfs = snurk_cmms_csht009_rfs.no_rfs (+)
    AND snurk_cmms_csht008_rfs.ct_misc_item = snurk_cmms_csht009_rfs.ct_misc_item (+)
    AND (snurk_cmms_csht009_rfs.ct_misc_item_page = 1
    OR snurk_cmms_csht009_rfs.ct_misc_item_page IS NULL);
    snurk_cmms_csht009_rfs,snurk_cmms_csht008_rfs,snurk_cmms_csht001_rfs are synonyms for views that are pointing to DB2 tables with the help of dblink.
    The dba saying that the where clause in the above select is not sent to the db2 side so creating a cartesian product and asking us to add something in the where clause.
    Thread detail
    ENCLAVE DETAIL INFORMATION
    PLAN
    Thread: Plan=DISTSERV Connid=SERVER Corrid=G422001E Authid=CPPRXPS
    Dist : Type=DATABASE ACCESS, Luwid=H36A45B9.A712.C8ADB744CDC6=132705
    Location : 19.106.69.185
    enc
    ENCLAVE TOKEN: F0001B77B8 Enclave Type: Original Indep
    Owning System: SYB1 Owning Job: DB27DIST
    WLM Mode: Goal Enclave CPU Time 01:26:00.935
    THREAD DETAIL
    PLAN
    Thread: Plan=DISTSERV Connid=SERVER Corrid=G422001E Authid=CPPRXPS
    Dist : Type=DATABASE ACCESS, Luwid=H36A45B9.A712.C8ADB744CDC6=132705
    Location : 19.106.69.185
    act
    Thread Activity User Defined Functions
    DB2 Status = WAIT-REMREQ TCB Time (SQL) = 00:00:00.000
    MVS Status = Wait for TCB Time = 00:00:00.000
    Total Elapsed Time = 01-17:00 Elapsed Time = 00:00:00.000
    CP CPU Utilization = 00.7% Elapsed Time (SQL) = 00:00:00.000
    Total CP CPU Time = N/A SQL Events = 0
    IIP CPU Time = 00:00:00.000
    Total Parallel Tasks = 0
    Current Parallel Tasks= 0
    SQL CALL BEING EXECUTED
    PLAN
    Thread: Plan=DISTSERV Connid=SERVER Corrid=G422001E Authid=CPPRXPS
    Dist : Type=DATABASE ACCESS, Luwid=H36A45B9.A712.C8ADB744CDC6=132705
    Location : 19.106.69.185
    call
    SQL call is active, call information is as follows :
    Thread Status = WAIT-REMREQ SQL Request Type = DYNAMIC
    Total SQL Reqs = 3589969 SQL Call Type = FETCH
    SQL DBRM Name = GDRDA11 SQL Statement Number = 00000
    Collection ID = SDCSDRDA
    SELECT A1."CD_PLANT", A1."NO_RFS", A2."CD_PLANT", A2."NO_RFS", A2."CD_PL
    ANT", A2."NO_RFS", A2."CT_MISC_ITEM" FROM "MMDBDP7"."CSHT001_RFS" A1, "M
    MDBDP7"."CSHT008_RFS_MISC" A2
    Please suggest on what to be added in the where clause.
    thanks,
    Vinodh

    We cannot create a view in db2 side because of security reasons. I tried to run the query as below.
    SELECT count(a),count(b),count(c) FROM
    (Select cd_plant,no_rfs,ct_misc_item, ct_misc_item_page from snurk_cmms_csht009_rfs) a,
    (Select cd_plant,no_rfs,ct_misc_item from snurk_cmms_csht008_rfs) b,
    (Select cd_plant,no_rfs from snurk_cmms_csht001_rfs ) c
    Where c.cd_plant = b.cd_plant (+)
    AND c.no_rfs = b.no_rfs (+)
    AND b.cd_plant = a.cd_plant (+)
    AND b.no_rfs = a.no_rfs (+)
    AND b.ct_misc_item = a.ct_misc_item (+)
    AND (a.ct_misc_item_page = 1
    OR a.ct_misc_item_page IS NULL)
    I am getting the error as below:
    ERROR at line 4:
    ORA-28545: error diagnosed by Net8 when connecting to an agent
    Unable to retrieve text of NETWORK/NCR message 65535
    ORA-02063: preceding 2 lines from FAST_SNURK_CMMS
    How to troubleshoot this.
    thanks,
    Vinodh

  • Tables with more than one cell with a high number of rowspan (ej. 619). This cell can not be print in a page and it must be cut. But I don't know how  indesign can do this action.

    Tables with more than one cell with a high number of rowspan (ej. 619). This cell can not be print in a page and it must be cut. But I don’t know how  indesign can do this action.

    set the wake-on lan on the main computer
    The laptop's too far away from the router to be connected by ethernet. It's all wifi.
    No separate server app on the laptop, it's all samba
    The files are on a windows laptop and a hard drive hooked up to the windows laptop. The windows share server is pants, so I'd need some sort of third party server running. Maybe you weren't suggesting to use Samba to connect to the windows share though?
    I'm glad that you've all understood my ramblings and taken and interest, thanks The way I see it, I can't be the only netbook user these days looking for this kind of convenience, and I certainly won't be once chrome and moblin hit the market.
    Last edited by saft (2010-03-18 20:38:08)

  • Trouble with the SQL smt to :list tables having more than 1000 rows

    Please I trying to list only tables having more than 1000 rows, but the sql stmt below doesn't work, can someone gives me a tips
    select table_name from user_tables where table_name in ( select table_name from user_tables where rownum > 1000 ) : The result is no rows!
    But I know that I have at lest 50 tables having more than 1000 rows
    Thanks a lot for the help

    If your tables are reasonably analyzed, then you can simply query:
    SELECT table_name,
           num_rows
      FROM user_tables
    WHERE num_rows >= 1000This will give you quite a reasonable estimate.
    Otherwise you have to go for dynamic sql or use the data dictionary to help you generate suitable scripts ....

  • Can't SSAS engines make use of more than one aggregation to answer a query?!

    I have a very simple cube (just for testing and training) . This cube contains two dimensions: [Dim Soccer Player] contains one attribute hierarchy [player name], the other dimension
    is [Dim Match Acts] contains also one attribute [Acts] which has values like fouls, goals, saves, tackles… etc. And of course a Fact that contains one measure of Just Count ... that simple ... so this cube can
    answers a question like how many goals scored by "Messi", for example ... a very simple trivial cube.
    I'm testing aggregations and their effect. so first I've designed one aggregation (Aggregation 0) on the granularity level of [Player name], then
    I run a query to get the count of ALL the[Acts] done by each [Player name] ... I've checked the SQL Profiler and I found that the aggregation was used.
    Then I cleared the cache, and I run another query, but this time to get just the number of Fouls committed by each [Player name], I checked the Profiler but the Aggregation 0 was NOT used.
    I went back to the aggregations design tab in BIDS, and I added another new aggregation (Aggregation 1) on the level of [Acts], so now I have two aggregation one on the granularity level of
    [Player name] and the second on the level of the [Acts].... I cleared the cache again and rerun the last query. NONE of the aggregation was used!
    In the third test I deleted Aggregation 1 and added [Acts] to Aggregation 0. so Aggregation 0 now on both [Player name] AND [Acts]... cleared the cache and rerun the last query. Aggregation
    0 appeared again.
    I just want to make sure (and if possible know why) the SSAS engine can't make use of and combine more than one aggregation to serve a query (point number 2), and that to design an aggregation
    that will serve a query which contains attributes from different dimensions, I have to add ALL the attributes in that query in that one aggregation, like point 3 ... is this true?!

    I think you are on the right track. You need to include all the attributes used in one query in the same aggregation (like #3) for it to be used. Example #2 works as I would expect. Queries above the grain of the agg (query by player name and an agg by
    player/act can be used) can be used. Queries below grain of the agg (example #2) can't use the agg.
    http://artisconsulting.com/Blogs/GregGalloway

Maybe you are looking for

  • How to print the Actual Check on 1st Page if line items are more for F110_p

    Hi, How to print the Actual Check on 1st page if the line items are more in F110_PRENUM_CHCK script. The standard script is printing at the last page of line items. I tried using IF &PAGE& EQ '1 ' /E 545 --> text element ENDIF but it is of no use. Pl

  • [solved] skype not in repos?

    Hi, I wanted to test skype in my 64 bit installation (with KDEmod, of course). Searched and found http://wiki.archlinux.org/index.php/Skype. As per the wiki entry, though skype has only 32 bit binary, it can be run in 64 bit environment. In my system

  • Pipeline Steps and JCo Calls

    Hi All, I wanted a few links which can provide me with indepth knowledge of these two points 1. Pipeline Steps in XI 2. JCo Calls during processing of messages I am looking specifically to understand the message processing in steps where the Mapping

  • Why having ORA-24344 error on my query?

    There are two tables one is sectors_main and emp_rem_view table. I don't know why i am having ORA-24344 error? Can anyone tell me why? SELECT ms_id, COUNT(*) cnt, NVL(ROUND(percentile_cont(0.25) within group (order by base_salary),0),0) bs_q1 FROM (S

  • Android Market on your pc

    Here is an app which will let you browse the market on your pc then you can sync with your phone and install what ever you want. WWW.APPBRAIN.COM