Which SQL statement is he running?

I have some oracle processes which are consuming a big fraction of the CPU resources and the load is so big that no users can connect (except with connect / as sysdba). Knowing their PID (from the UNIX top command), and connecting internally on Oracle, is there a way to know which statements are currently being performed by these users?
NB: Database in question is Oracle 10g.

Just because users cannot connect does not mean the problem is a "load" problem.
What does v$session_wait show that connected sessions are waiting on? Are any sessions lock waited?
If there are lockwaited sessions are any of the Oracle rdbms background sessions involved in the lock problem?
In situations where people cannot connect or take five minutes to do so the problem usually comes down to improper configuration of the database, poor design of the most heavily hit section of an application, or your are especially lucky and have managed to hit a bug.
In the absence of a locking problem or every session waiting on the same resouce then a statspack for a short duration is worth a shot (providing it will run).
HTH -- Mark D Powell --

Similar Messages

  • Content Server - Which SQL statements are running?

    Hi sap lovers! Does anyone aware of which sql statements being run on the maxDB part when, e.g saving a document (or retrieving a document) (on the ABAP part) within our DMS landscape?
    And at the whole, Is It possible to monitor sqls running on MAXDB, like we can see in ST04
    -> Oracle Session
    Regards,
    Vladimir

    You can find that out doing the following:
    - on any ABAP system start DB59
    - enter the data to connec to the content server database
    - doubleclick on the made entry so you get in "DB50 for remote databases"
    - switch on the command monitor
    - then use the task manager and press on "running commands"
    Markus

  • BPC 7 MS: Which SQL statements are created for writing with input schedule?

    Hi,
    I wanted to know which SQL statements are created and executed if a user submits values using an Excel input schedule to an application.
    When I check the correspoinding MS SQL server log files, I see that data is read from the three partitions belonging to the application and put into a temporary table, but I can't find anything about writing back to the application (presumably the WB partiton...) in the log.
    There are some cryptic entries in the log file as well, but they are not human-readale... are there any BPC logfiles that could tell me which SQL statements are created and executed to write back the new values to the application? Thanks!

    Hi,
    As far as i know, when a user send data entry from an excel schedule, it will be written in the WB table of the application (for each application, you have 3 Data tables : WB, Fact2 and Fact).
    I presume that the SQL statement may be an INSERT or UPDATE statement.
    Technically, the update is done by the send governor service (hosted on your BPC application server).
    There is no log that will show you the SQL Statement besides a SQL trace that you have to setup in SQL Server 2005 Manager Studio.
    btw, the data are written in the Relationnal database but are read from the OLAP cube. The olap Cube is split in 3 partitions (ROLAP on WB table / and MOLAP on fact and fat2). Wich mean that every new entry in WB will be automatically "updated" in the cube.
    Some DM packages can directly write data in fact2 table. In this case you need to reprocess the cube to get it loaded.

  • How to find what sql statement is currently running by scheduler job?

    Hi
    I scheduled a stored procedure to run every 5 minutes using dbms_scheduler.
    The stored procedure internally call serveral other stored procedures.
    The scheduled job is in running state and right now I want to find out what is the sql statement the job executing.
    Is there any sql query to find out what is sql query the schedular job is running?? Please help me.
    Thanks in Advance.

    The previous sql id is null in this case.
    I already quoted as I am running a pl/sql block in my scheduler job.
    begin
    usp_test('praram');
    end;
    I am using the pl/sql block as above in my scheduler job. The USP_TEST inturn calls some other stored procedures and each stored procedure has several inserts and update statements.
    The scheduler job is still running. I know the sid of the running scheduler job.
    I want to know the job is at which stored procedure and at which insert/update statement.
    Please let me know is there any query to fulfil my requirement?
    I greatly appreciate your help
    Thanks

  • How do i see the sql statements that are run ?

    i want to see (in a log file) what are the sql statements that oracle had ran.
    how do i do it ? what configurations should i do ? how ?
    please help ... thanks

    Current SQL statements are viewed in dynamic tables:
    SELECT SCHEMANAME, SQL_ADDRESS, SQL_TEXT, last_call_et
    FROM V$SESSION, V$SQLAREA
    WHERE V$SESSION.SQL_ADDRESS=V$SQLAREA.ADDRESS
    A historical log of all SQL statements is recorded in Oracle Archived Logs, and can be viewed using the Oracle Logminer approach. This process requires your database to be in ARCHIVELOG mode, and that Logminer be configured to read the archived logs.

  • ORA-00900: invalid SQL statement error while running jobs in APEX 4.2

    Hi,
    the following procedure compiled successfully in Apex 4.2:
    create or replace procedure "DEVLOE" AS
    L_TEMPLATE VARCHAR2(32000);
    L_base_TEMPLATE VARCHAR2(32000);
    begin
    l_base_template := '<html>
    <head>
    <title>HTML Editor Sample Page</title>
    </head>
    <body>
    <p>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;"><span class="plsql-literal" style="line-height: 14.65625px; white-space: nowrap;">Hi #Contact#,</span></span></span></p>
    <p>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;"><span class="plsql-literal" style="line-height: 14.65625px; white-space: nowrap;">This is a notification message to inform you that <strong>Development </strong>effort </span><span class="plsql-literal" style="line-height: 14.65625px; white-space: nowrap;">and milestones are yet to be updated for ECR: </span><span class="plsql-operator" style="line-height: 14.65625px; white-space: nowrap;"></span><span class="plsql-word" style="line-height: 14.65625px; white-space: nowrap;">#ECR_NAME#. This ECR has priority #p#. </span></span></span></p>
    <table align="center" border="1" cellpadding="1" cellspacing="1" style="width: 500px;">
    <tbody>
    <tr>
    <td>
    <span style="font-size:14px;"><span style="color: rgb(255, 0, 0);"><span style="font-family: verdana, geneva, sans-serif;"><u><strong>Department</strong></u></span></span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;"><span style="color: rgb(255, 0, 0);"><u><strong><span style="background-color: rgb(255, 255, 255);">Contact Name</span></strong></u></span></span></span></td>
    </tr>
    <tr>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">Product Management Contact</span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">#pdm#</span></span></td>
    </tr>
    <tr>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">PMO Contact</span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">#PMO#</span></span></td>
    </tr>
    <tr>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">Dev Contact</span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">#Dev#</span></span></td>
    </tr>
    <tr>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">QA Contact</span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">#QA#</span></span></td>
    </tr>
    <tr>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">Infodev Contact</span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">#Info#</span></span></td>
    </tr>
    </tbody>
    </table>
    <p>
     </p>
    <span fckbookmark="1" style="display: none;"> </span><span fckbookmark="1" style="display: none;"> </span>
    <table align="center" border="1" cellpadding="1" cellspacing="1" style="width: 500px;">
    <tbody>
    <tr>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;"><span style="color: rgb(0, 0, 255);">Team</span></span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;"><span style="color: rgb(0, 0, 255);">Start Date</span></span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;"><span style="color: rgb(0, 0, 255);">End Date</span></span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;"><span style="color: rgb(0, 0, 255);">Effort(in person days)</span></span></span></td>
    </tr>
    <tr>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">Development</span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">#d1#</span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">#e1#</span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">#L1#</span></span></td>
    </tr>
    <tr>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">Quality Assurance</span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">#d2#</span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">#e2#</span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">#L2#</span></span></td>
    </tr>
    <tr>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">Information Development</span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">#d3#</span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">#e3#</span></span></td>
    <td>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">#L3#</span></span></td>
    </tr>
    </tbody>
    </table>
    <p>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">The ECR application is located</span> at </span>: https://apex.oraclecorp.com/pls/apex/f?p=1472:1<span fckbookmark="1" style="display: none;"> </span></p>
    <span fckbookmark="1" style="display: none;"> </span>
    <p>
    <span style="font-family: verdana, geneva, sans-serif; font-size: 14px;">Regards,</span></p>
    <p>
     </p>
    <p>
    <span style="font-size:14px;"><span style="font-family: verdana, geneva, sans-serif;">ECR Team</span></span></p>
    </body>
    </html>
    FOR c1 IN (select ECR_NAME,PMO_CONTACT, QE_LOE,INFODEV_LOE, DEV_START_DATE, DEV_END_DATE, QE_CONTACT,QE_START_DATE,INFODEV_START_DATE,INFODEV_END_DATE, QE_END_DATE, INFODEV_CONTACT, PDM_CONTACT, DEV_LOE,ECR_PRIORITY, DEV_CONTACT from ECR_NAME WHERE DEV_LOE IS NULL AND ECR_STATUS = 'Effort Estimation is in Progress' AND ECR_PRIORITY <=5)
    LOOP
    l_template := l_base_template;
    l_template:= replace(l_template, '#Contact#', c1.DEV_CONTACT);
    l_template:= replace(l_template, '#ECR_NAME#', c1.ECR_NAME);
    l_template:= replace(l_template, '#p#', c1.ECR_PRIORITY);
    l_template:= replace(l_template, '#PMO#', c1.PMO_CONTACT);
    l_template:= replace(l_template, '#pdm#', c1.PDM_CONTACT);
    l_template:= replace(l_template, '#Dev#', c1.DEV_CONTACT);
    l_template:= replace(l_template, '#QA#', c1.QE_CONTACT);
    l_template:= replace(l_template, '#Info#', c1.INFODEV_CONTACT);
    l_template:= replace(l_template, '#d1#', c1.DEV_START_DATE);
    l_template:= replace(l_template, '#L1#', c1.DEV_LOE);
    l_template:= replace(l_template, '#e1#', c1.DEV_END_DATE);
    l_template:= replace(l_template, '#d2#', c1.QE_START_DATE);
    l_template:= replace(l_template, '#L2#', c1.QE_LOE);
    l_template:= replace(l_template, '#e2#', c1.QE_END_DATE);
    l_template:= replace(l_template, '#d3#', c1.INFODEV_START_DATE);
    l_template:= replace(l_template, '#L3#', c1.INFODEV_LOE);
    l_template:= replace(l_template, '#e3#', c1.INFODEV_END_DATE);
    htmldb_mail.Send(p_to => '[email protected]',
    p_cc => '[email protected]',
    p_from => '[email protected]',
    p_body => l_template,
    p_body_html => l_template,
    p_subj => 'cxxxx:' ||c1.ECR_NAME );
    END LOOP;
    end;
    the when I ran:
    begin
    devloe();
    end;
    I received the mail as well.
    However, when I try to schedule this in the job:
    declare
    JobNo user_jobs.job%TYPE;
    begin
    dbms_job.submit(JobNo, 'BEGIN
    DEVLOE();
    END;', sysdate, 'sysdate + 1/24/60'); commit; end;
    The job status shows failures every minute.
    dbms_job.run(job_no);
    It gave : ORA-00900: invalid SQL statement

    First of all I don't see a procedure header.
    Secondly I see you commit inside your procedure. This is a bug.
    Thirdly I see you also commit inside a loop. This is also a bug, and needs to be removed asap.
    The error indicates a statement doesn't parse. As you don't post the error stack, nor a table definition no one can reproduce the problem.
    You need to isolate the statements, one by one, and run them through sql*plus to see what happens.
    Sybrand Bakker
    Senior Oracle DBA

  • SQL Statement error when running ETL from DAC

    Dear all,
    I installed and configured biapps 7.9.6.3, then I run the full load of Subject CRM - Loyalty in DAC. And I got the error of task fail, I check the Session log files in Informatica server.
    $ view SEBL_VERT_811.DATAWAREHOUSE.SDE_SBL_Vert_811_Adaptor.SDE_GeographyDimension_Business.log
    DIRECTOR> VAR_27028 Use override value [0] for user-defined workflow/worklet variable:[$$passInStatus].
    DIRECTOR> VAR_27028 Use override value [DataWarehouse] for session parameter:[$DBConnection_OLAP].
    DIRECTOR> VAR_27028 Use override value [SEBL_VERT_811] for session parameter:[$DBConnection_OLTP].
    DIRECTOR> VAR_27028 Use override value [SEBL_VERT_811.DATAWAREHOUSE.SDE_SBL_Vert_811_Adaptor.SDE_GeographyDimension_Business.log] for session parameter:[$PMSessionLogFile].
    DIRECTOR> VAR_27028 Use override value [1] for mapping parameter:[MPLT_LOAD_W_GEO_DS.$$DATASOURCE_NUM_ID].
    DIRECTOR> VAR_27028 Use override value [] for mapping parameter:[$$Hint1].
    DIRECTOR> VAR_27028 Use override value [] for mapping parameter:[$$Hint2].
    DIRECTOR> TM_6014 Initializing session [SDE_GeographyDimension_Business] at [Mon Jul 25 17:29:47 2011].
    DIRECTOR> TM_6683 Repository Name: [Oracle_BI_DW_Base]
    DIRECTOR> TM_6684 Server Name: [Oracle_BI_DW_Server]
    DIRECTOR> TM_6686 Folder: [SDE_SBL_Vert_811_Adaptor]
    DIRECTOR> TM_6685 Workflow: [SDE_GeographyDimension_Business] Run Instance Name: [] Run Id: [260]
    DIRECTOR> TM_6101 Mapping name: SDE_GeographyDimension_Business [version 1].
    DIRECTOR> TM_6963 Pre 85 Timestamp Compatibility is Enabled
    DIRECTOR> TM_6964 Date format for the Session is [MM/DD/YYYY HH24:MI:SS]
    DIRECTOR> TM_6827 [u01/app/oracle/biapps/dev/Informatica/9.0.1/server/infa_shared/Storage] will be used as storage directory for session [SDE_GeographyDimension_Business].
    DIRECTOR> CMN_1802 Session recovery cache initialization is complete.
    DIRECTOR> TM_6708 Using configuration property [DisableDB2BulkMode,Yes]
    DIRECTOR> TM_6708 Using configuration property [ServerPort,6325]
    DIRECTOR> TM_6708 Using configuration property [overrideMpltVarWithMapVar,Yes]
    DIRECTOR> TM_6708 Using configuration property [SiebelUnicodeDB,SIEBEL@ANSDEV dwhadmin@ANBDEV]
    DIRECTOR> TM_6703 Session [SDE_GeographyDimension_Business] is run by 64-bit Integration Service [node01_hkhgc01dvapp01], version [9.0.1 HotFix2], build [1111].
    MANAGER> PETL_24058 Running Partition Group [1].
    MANAGER> PETL_24000 Parallel Pipeline Engine initializing.
    MANAGER> PETL_24001 Parallel Pipeline Engine running.
    MANAGER> PETL_24003 Initializing session run.
    MAPPING> CMN_1569 Server Mode: [UNICODE]
    MAPPING> CMN_1570 Server Code page: [UTF-8 encoding of Unicode]
    MAPPING> TM_6151 The session sort order is [Binary].
    MAPPING> TM_6185 Warning. Code page validation is disabled in this session.
    MAPPING> TM_6156 Using low precision processing.
    MAPPING> TM_6180 Deadlock retry logic will not be implemented.
    MAPPING> TM_6187 Session target-based commit interval is [10000].
    MAPPING> TM_6307 DTM error log disabled.
    MAPPING> TE_7022 TShmWriter: Initialized
    MAPPING> DBG_21075 Connecting to database [ANBDEV], user [dwhadmin]
    MAPPING> CMN_1716 Lookup [MPLT_LOAD_W_GEO_DS.LKP_W_LST_OF_VAL_G] uses database connection [Relational:DataWarehouse] in code page [UTF-8 encoding of Unicode]
    MAPPING> CMN_1716 Lookup [MPLT_LOAD_W_GEO_DS.LKP_W_GEO_DS] uses database connection [Relational:DataWarehouse] in code page [UTF-8 encoding of Unicode]
    MAPPING> DBG_21694 AGG_COUNTRY_CITY_ZIPCODE Partition [0]: Index cache size = [1048576], Data cache size = [2097152]
    MAPPING> TE_7212 Increasing [Index Cache] size for transformation [AGG_COUNTRY_CITY_ZIPCODE] from [1048576] to [2402304].
    MAPPING> TE_7212 Increasing [Data Cache] size for transformation [AGG_COUNTRY_CITY_ZIPCODE] from [2097152] to [2097528].
    MAPPING> TE_7029 Aggregate Information: Creating New Index and Data Files
    MAPPING> TE_7034 Aggregate Information: Index file is [u01/app/oracle/biapps/dev/Informatica/9.0.1/server/infa_shared/Cache/PMAGG14527_3_0_260.idx]
    MAPPING> TE_7035 Aggregate Information: Data file is [u01/app/oracle/biapps/dev/Informatica/9.0.1/server/infa_shared/Cache/PMAGG14527_3_0_260.dat]
    MAPPING> TM_6007 DTM initialized successfully for session [SDE_GeographyDimension_Business]
    DIRECTOR> PETL_24033 All DTM Connection Info: [<NONE>].
    MANAGER> PETL_24004 Starting pre-session tasks. : (Mon Jul 25 17:29:47 2011)
    MANAGER> PETL_24027 Pre-session task completed successfully. : (Mon Jul 25 17:29:47 2011)
    DIRECTOR> PETL_24006 Starting data movement.
    MAPPING> TM_6660 Total Buffer Pool size is 36000000 bytes and Block size is 128000 bytes.
    LKPDP_2> DBG_21097 Lookup Transformation [MPLT_LOAD_W_GEO_DS.LKP_W_GEO_DS]: Default sql to create lookup cache: SELECT CITY,COUNTRY,ZIPCODE,STATE_PROV FROM W_GEO_DS ORDER BY CITY,COUNTRY,ZIPCODE,STATE_PROV
    LKPDP_1> DBG_21312 Lookup Transformation [MPLT_LOAD_W_GEO_DS.LKP_W_LST_OF_VAL_G]: Lookup override sql to create cache: SELECT W_LST_OF_VAL_G.VAL AS VAL, W_LST_OF_VAL_G.R_TYPE AS R_TYPE FROM W_LST_OF_VAL_G
    WHERE
         W_LST_OF_VAL_G.R_TYPE LIKE 'ETL%' ORDER BY R_TYPE,VAL
    LKPDP_1> TE_7212 Increasing [Index Cache] size for transformation [MPLT_LOAD_W_GEO_DS.LKP_W_LST_OF_VAL_G] from [1048576] to [1050000].
    LKPDP_2> TE_7212 Increasing [Index Cache] size for transformation [MPLT_LOAD_W_GEO_DS.LKP_W_GEO_DS] from [20000000] to [20006400].
    LKPDP_2> TE_7212 Increasing [Data Cache] size for transformation [MPLT_LOAD_W_GEO_DS.LKP_W_GEO_DS] from [20000000] to [20004864].
    READER_1_1_1> DBG_21438 Reader: Source is [ANSDEV], user [SIEBEL]
    READER_1_1_1> BLKR_16051 Source database connection [SEBL_VERT_811] code page: [UTF-8 encoding of Unicode]
    READER_1_1_1> BLKR_16003 Initialization completed successfully.
    WRITER_1_*_1> WRT_8146 Writer: Target is database [ANBDEV], user [dwhadmin], bulk mode [ON]
    WRITER_1_*_1> WRT_8106 Warning! Bulk Mode session - recovery is not guaranteed.
    WRITER_1_*_1> WRT_8221 Target database connection [DataWarehouse] code page: [UTF-8 encoding of Unicode]
    WRITER_1_*_1> WRT_8124 Target Table W_GEO_DS :SQL INSERT statement:
    INSERT INTO W_GEO_DS(CITY,CONTINENT,COUNTRY,COUNTY,STATE_PROV,ZIPCODE,DATASOURCE_NUM_ID,X_CUSTOM) VALUES ( ?, ?, ?, ?, ?, ?, ?, ?)
    WRITER_1_*_1> WRT_8020 No column marked as primary key for table [W_GEO_DS]. UPDATEs Not Supported.
    WRITER_1_*_1> WRT_8270 Target connection group #1 consists of target(s) [W_GEO_DS]
    WRITER_1_*_1> WRT_8003 Writer initialization complete.
    READER_1_1_1> BLKR_16007 Reader run started.
    WRITER_1_*_1> WRT_8005 Writer run started.
    WRITER_1_*_1> WRT_8158
    *****START LOAD SESSION*****
    Load Start Time: Mon Jul 25 17:29:47 2011
    Target tables:
    W_GEO_DS
    READER_1_1_1> RR_4029 SQ Instance [SQ_S_ADDR_ORG] User specified SQL Query [SELECT  DISTINCT
    S_ADDR_ORG.CITY,
    S_ADDR_ORG.COUNTRY,
    S_ADDR_ORG.COUNTY,
    S_ADDR_ORG.PROVINCE,
    S_ADDR_ORG.STATE,
    S_ADDR_ORG.ZIPCODE,
          '0' AS X_CUSTOM
    FROM
    V_ADDR_ORG S_ADDR_ORG
    READER_1_1_1> RR_4049 SQL Query issued to database : (Mon Jul 25 17:29:47 2011)
    READER_1_1_1> CMN_1761 Timestamp Event: [Mon Jul 25 17:29:47 2011]
    READER_1_1_1> RR_4035 SQL Error [
    ORA-00942: table or view does not exist
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    S_ADDR_ORG.CITY,
    S_ADDR_ORG.COUNTRY,
    S_ADDR_ORG.COUNTY,
    S_ADDR_ORG.PROVINCE,
    S_ADDR_ORG.STATE,
    S_ADDR_ORG.ZIPCODE,
    '0' AS X_CUSTOM
    FROM
    V_ADDR_ORG S_ADDR_ORG
    Oracle Fatal Error
    Database driver error...
    Function Name : Execute
    SQL Stmt : SELECT DISTINCT
    S_ADDR_ORG.CITY,
    S_ADDR_ORG.COUNTRY,
    S_ADDR_ORG.COUNTY,
    S_ADDR_ORG.PROVINCE,
    S_ADDR_ORG.STATE,
    S_ADDR_ORG.ZIPCODE,
    '0' AS X_CUSTOM
    FROM
    V_ADDR_ORG S_ADDR_ORG
    Oracle Fatal Error].
    READER_1_1_1> CMN_1761 Timestamp Event: [Mon Jul 25 17:29:47 2011]
    READER_1_1_1> BLKR_16004 ERROR: Prepare failed.
    WRITER_1_*_1> WRT_8333 Rolling back all the targets due to fatal session error.
    WRITER_1_*_1> WRT_8325 Final rollback executed for the target [W_GEO_DS] at end of load
    WRITER_1_*_1> WRT_8035 Load complete time: Mon Jul 25 17:29:47 2011
    LOAD SUMMARY
    ============
    WRT_8036 Target: W_GEO_DS (Instance Name: [W_GEO_DS])
    WRT_8044 No data loaded for this target
    WRITER_1__1> WRT_8043 ****END LOAD SESSION*****
    MANAGER> PETL_24031
    ***** RUN INFO FOR TGT LOAD ORDER GROUP [1], CONCURRENT SET [1] *****
    Thread [READER_1_1_1] created for [the read stage] of partition point [SQ_S_ADDR_ORG] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [TRANSF_1_1_1] created for [the transformation stage] of partition point [SQ_S_ADDR_ORG] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [TRANSF_1_2_1] created for [the transformation stage] of partition point [AGG_COUNTRY_CITY_ZIPCODE] has completed. The total run time was insufficient for any meaningful statistics.
    Thread [WRITER_1_*_1] created for [the write stage] of partition point [W_GEO_DS] has completed. The total run time was insufficient for any meaningful statistics.
    MAPPING> CMN_1791 The index cache size that would hold [0] aggregate groups of input rows for [AGG_COUNTRY_CITY_ZIPCODE], in memory, is [0] bytes
    MAPPING> CMN_1790 The data cache size that would hold [0] aggregate groups of input rows for [AGG_COUNTRY_CITY_ZIPCODE], in memory, is [0] bytes
    MAPPING> CMN_1793 The index cache size that would hold [0] rows in the lookup table for [MPLT_LOAD_W_GEO_DS.LKP_W_LST_OF_VAL_G], in memory, is [0] bytes
    MAPPING> CMN_1792 The data cache size that would hold [0] rows in the lookup table for [MPLT_LOAD_W_GEO_DS.LKP_W_LST_OF_VAL_G], in memory, is [0] bytes
    MAPPING> CMN_1793 The index cache size that would hold [0] rows in the lookup table for [MPLT_LOAD_W_GEO_DS.LKP_W_GEO_DS], in memory, is [0] bytes
    MAPPING> CMN_1792 The data cache size that would hold [0] rows in the lookup table for [MPLT_LOAD_W_GEO_DS.LKP_W_GEO_DS], in memory, is [0] bytes
    MANAGER> PETL_24005 Starting post-session tasks. : (Mon Jul 25 17:29:47 2011)
    MANAGER> PETL_24029 Post-session task completed successfully. : (Mon Jul 25 17:29:47 2011)
    MAPPING> TE_7216 Deleting cache files [PMLKUP14527_524289_0_260L64] for transformation [MPLT_LOAD_W_GEO_DS.LKP_W_LST_OF_VAL_G].
    MAPPING> TE_7216 Deleting cache files [PMLKUP14527_524293_0_260L64] for transformation [MPLT_LOAD_W_GEO_DS.LKP_W_GEO_DS].
    MAPPING> TM_6018 The session completed with [0] row transformation errors.
    MANAGER> TE_7216 Deleting cache files [u01/app/oracle/biapps/dev/Informatica/9.0.1/server/infa_shared/Cache/PMAGG14527_3_0_260.idx] for transformation [AGG_COUNTRY_CITY_ZIPCODE].
    MANAGER> TE_7216 Deleting cache files [u01/app/oracle/biapps/dev/Informatica/9.0.1/server/infa_shared/Cache/PMAGG14527_3_0_260.dat] for transformation [AGG_COUNTRY_CITY_ZIPCODE].
    MANAGER> PETL_24002 Parallel Pipeline Engine finished.
    DIRECTOR> PETL_24013 Session run completed with failure.
    DIRECTOR> TM_6022
    SESSION LOAD SUMMARY
    ================================================
    DIRECTOR> TM_6252 Source Load Summary.
    DIRECTOR> CMN_1740 Table: [SQ_S_ADDR_ORG] (Instance Name: [SQ_S_ADDR_ORG])
         Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6253 Target Load Summary.
    DIRECTOR> CMN_1740 Table: [W_GEO_DS] (Instance Name: [W_GEO_DS])
         Output Rows [0], Affected Rows [0], Applied Rows [0], Rejected Rows [0]
    DIRECTOR> TM_6023
    ===================================================
    DIRECTOR> TM_6020 Session [SDE_GeographyDimension_Business] completed at [Mon Jul 25 17:29:48 2011].
    After review the log, I found the select statement was fail, the SQL below was wrong:
    SELECT DISTINCT
    S_ADDR_ORG.CITY,
    S_ADDR_ORG.COUNTRY,
    S_ADDR_ORG.COUNTY,
    S_ADDR_ORG.PROVINCE,
    S_ADDR_ORG.STATE,
    S_ADDR_ORG.ZIPCODE,
    '0' AS X_CUSTOM
    FROM
    V_ADDR_ORG S_ADDR_ORG
    There is no table V_ADDR_ORG but S_ADDR_ORG in the Siebel transaction database. So I don't know why it generate this sql when transfering data to OBAW. This is the out of box bi application.
    Experts! How could I fix this problem?? Please help, thank you very much!!
    Best regards,
    Ryan

    Ryan,
    Yes you missed to create views for source tables. so in dac design>> tables tab, select any table and right clieck on it then click on change capture scripts>> generate view scripts. it will ask you whether it can generate script for all the tables. so it will generate view script for you. now run the entire esript in source database. then your problem will be solved.
    if this answered your question . make my answer correct.
    Thanks
    Jay.

  • Current running sql statement in procedure?

    How to identify the current running sql statement in procedure or package? From the v$sqlarea, i can see the only procedure name with parameters, but not current execution statement? is there any dba view to know the current execution statement in 9i & 10g ?

    hi
    u can also use dba_source view to know which sql statement issued by which user

  • For Finding sql statements which consumes High CPU

    Hai,
    Can any help me for finding, which sql statements are comsumes high CPU

    generaly:
    1. from unix site :
    us top command or ps (but this depend on platform) -> use pid in select * from v$process where spid <>
    2. from db site
    select * from v$sql_area order by buffer_gets desc ;
    this can You join with v$session where v$session.sql_address=v$sqlarea.address
    and when You wont to see % of os select *spid from v$process where addr=<v$session.paddr>
    and then in unix ps -ef | grep <spid>
    But point 2 is maybe discutable because also parse is one style how to consume CPU ... etc
    In 10g
    from AWR dba_hist_osstat -> DBA_HIST_SQLSTAT You can see high usage CPU interval , but this show You not momentaly action , but historical data
    momentali from ASH
    now I study viraq_sh script ... I thing this answer your question tottally
    Message was edited by:
    branislav.dobrotka

  • CPU consumption 100%: How do I see which SQL is consuming?

    Hi,
    I need help badly with this because I cannot get my head around the problem. I have a Windows server 2003 Enterprise 64-bit running Oracle 10.2.0.3 and last night it started consuming 100% CPU. The Windows performance monitor shows 100% CPU in any case, and it was difficult to even login via remote desktop.
    It is Oracle.exe that is consuming, and there is only one database on this server. It is an Opteron 875 2.2Ghz with 8GB ram.
    I want to see which SQL statement is responsible for the CPU consumption. I know exactly which user/programapplication is using it... it is useless to kick the session because the application connects immediately again from the app. server and runs the same thing. Restarting the database also has the same effect: session reconnects and CPU goes through the roof. Customer says that they are running the same stuff as always, and app. has not been updated recently. 100 % CPU to them, means something is wrong.
    Incidently, I tried to CPU patch this database a couple of days beforehand, but there was not enough space on disk (needed 1.6GB) so opatch complained and would not continue. I simply restarted the database without having run opatch. I cannot imagine that this is the cause of this problem.
    Anyhoo...
    I have a script from Kyle Haileys web site to get the CPU for each sql_id using ash tables:
    -- (c) Kyle Hailey 2007
    col type for a10
    select * from (
    select
    ash.SQL_ID , ash.SQL_PLAN_HASH_VALUE Plan_hash, aud.name type,
    sum(decode(ash.session_state,'ON CPU',1,0)) "CPU",
    sum(decode(ash.session_state,'WAITING',1,0)) -
    sum(decode(ash.session_state,'WAITING', decode(wait_class, 'User I/O',1,0),0)) "WAIT" ,
    sum(decode(ash.session_state,'WAITING', decode(wait_class, 'User I/O',1,0),0)) "IO" ,
    sum(decode(ash.session_state,'ON CPU',1,1)) "TOTAL"
    from v$active_session_history ash,
    audit_actions aud
    where SQL_ID is not NULL
    and ash.sql_opcode=aud.action
    and ash.sample_time > sysdate - &minutes /( 60*24)
    group by sql_id, SQL_PLAN_HASH_VALUE , aud.name
    order by sum(decode(session_state,'ON CPU',1,1)) desc
    ) where rownum < 10
    This give following output:
    SQL_ID PLAN_HASH TYPE CPU WAIT IO TOTAL
    cxk376gut06wn 3831248992 SELECT 12 145 5855 6012
    4ffvdt72n46yt 4094592372 SELECT 0 3 1131 1134
    dnxgm7bg8wqg1 4025412530 SELECT 3 13 337 353
    3y0dyvkc9sx69 1764840833 SELECT 18 169 56 243
    abz05715h08ct 1967470119 SELECT 0 7 175 182
    4gd6b1r53yt88 0 UNKNOWN 0 92 4 96
    9phhuzxpqwsgc 2662599208 SELECT 0 67 0 67
    275bbj5czbrwp 629013258 SELECT 0 2 64 66
    1dvrrg3aah2mu 75914095 SELECT 0 6 57 63
    My assumption is that this gives the average CPU usage over the last x minutes and the total cpu usage for this sql_id. There are no sql running that show a very high average CPU, but the total is up in 6000 for the longest running sql, so I get the statement for the top sql_id that has highest total, and it is this:
    SQL_ID cxk376gut06wn, child number 2
    SELECT NVL(MIN(ID),0) FROM TABLE_OBFUSCATED WHERE TASK = :B1 AND
    STATUS='JobQueue' AND MESSAGE_ID IS NULL
    Plan hash value: 3831248992
    | Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time |
    | 0 | SELECT STATEMENT | | | | 3231 (100)| |
    | 1 | SORT AGGREGATE | | 1 | 42 | | |
    |* 2 | TABLE ACCESS FULL| TABLE_OBFUSCATED | 3 | 126 | 3231 (1)| 00:00:39 |
    Predicate Information (identified by operation id):
    2 - filter(("TASK"=:B1 AND "STATUS"='JobQueue' AND "MESSAGE_ID" IS NULL))
    It doesn't seem to fit! It's such a simple statement, but according to ash this is the sql using the most total cpu over the last 30 minutes?? So I create a temp index on the filter and check the plan for the statement again, and it uses the index, but I need to kill the session or bounce the database for the new plan to be used because the currently running session is still using the old plan. Customer was desperate to get it down so I even tried flushing the shared pool but new plan does not get used. I did not want to restart the database. Do I have to kill all sessions using that plan for the new one to take effect?
    Now im stuck. I don't really know how else to get an accurate picture of which sql is currently consuming CPU. I need to know this to try to fix the SQL because stopping the application is not an option, even if it is throwing too much at the database. The only other option I can think of is setting up profile limits for cpu consumption in oracle, but then the job will likely take an unacceptable amount of time for the customer.
    So my questions are:
    1. How else can I check which SQL is using CPU accurately?
    2. Does the SQL I show above accuratlely display current average CPU over x minutes with ASH? Why is it so low for the top consumer but my CPU is still at 100% on the OS?
    Thankyou for your help.

    Yeah, OK, I can see an arbitary number for CPU used since session startup (what does that refer to by the way? It can't be percent), and I can see that one user has consumed massive amounts of CPU compared to all other sessions, but he is currently not running anything, so how can I see which SQL he ran in the past day that cuased all that CPU consumption? Can I do this without AWR?
    By the way, do you need a Diagnostic Pack license for simply viewing the Performance tab of dbconsole? I Can go into dbconsole and see history for 2 days back and see the massive espike that caused CPU consumption, but when dbconsole displays the list of top SQL and top sessions under the graph, is it the top listed SQL that consumed the most cpu or do you need to correlate the colours, i.e. find the SQL listed in top SQL that has the same colour as CPU usage on the graph (green) even if it is not listed as the top "consumer" (which I assume means top AAS consumer: overall consumption of all resources)?

  • Query cannot be parsed within the Builder ORA-00900: invalid SQL statement

    Why am I getting the error message "Query cannot be parsed within the Builder. If you believe your query is syntactically correct, check the ''generic columns'' checkbox below the region source to proceed without parsing. ORA-00900: invalid SQL statement"
    If I run the query in sqlplus it comes back with "no rows selected" which is what I want.
    The function is valid and the user has privileges to run it.
    Is there something I can do about it?
    If I only have any field instead of the function the error goes away.
    Can someone explain what is happening? I'm using 2.0.
    select
    F_TOTAL_AMOUNT_PAID('fees',a.acct_fee_id,0) "Fee Paid"
    from fee_component a, fee b, license c
    where a.acctfee_id = b.acctfee_id
    and b.license_id = c.license_id
    and b.fee_status = 'U'
    and c.client_id = :P301_CLIENT_ID
    Thanks I really appreciate any help I can get.

    Does your parsing schema have direct granted privileges to execute the function or privileges via a role? If it is through a role, you need to directly grant execute on the function to your parsing schema.
    Mike

  • SQL Statement using Undo Tablespace

    Hi
    Is there a way to find out which SQL Statement is using the highest amount of Undo Tablespace ?

    Query v$undostat.
    Use column "MAXQUERYID" for identgifying long running queries.
    Refer Metalink Note 262066.1 for undo information.
    Regards
    Sethu

  • PHP - Multiple sql statements

    I am using several different sql statements to be used
    depending on which variables the user has chosen on the previous
    page. Sometimes more than one variable will be selected
    (checkboxes) so more than one of the sql statements will be run,
    but I want all the results to go into the same $num variable.
    I have tried this with the following scipt but it doesn't
    work. How can I modify this to make it work???
    Many Thanks
    if ($buying_and_merchandising){
    $query = "SELECT * FROM bwjobs WHERE location LIKE
    '$location' AND description LIKE '$keyword' AND title LIKE
    '$keyword' AND buying_and_merchandising =
    '$buying_and_merchandising' AND visible = 'yes' ORDER BY
    posted_date DESC";
    if ($design){
    $query2 = "SELECT * FROM bwjobs WHERE location LIKE
    '$location' AND description LIKE '$keyword' AND title LIKE
    '$keyword' AND design = '$design' AND visible = 'yes' ORDER BY
    posted_date DESC";
    if ($executive){
    $query3 = "SELECT * FROM bwjobs WHERE location LIKE
    '$location' AND description LIKE '$keyword' AND title LIKE
    '$keyword' AND executive = '$executive' AND visible = 'yes' ORDER
    BY posted_date DESC";
    if ($human_resources){
    $query4 = "SELECT * FROM bwjobs WHERE location LIKE
    '$location' AND description LIKE '$keyword' AND title LIKE
    '$keyword' AND human_resources = '$human_resources' AND visible =
    'yes' ORDER BY posted_date DESC";
    if ($manufacturing){
    $query5 = "SELECT * FROM bwjobs WHERE location LIKE
    '$location' AND description LIKE '$keyword' AND title LIKE
    '$keyword' AND manufacturing = '$manufacturing' AND visible = 'yes'
    ORDER BY posted_date DESC";
    if ($retail_operations){
    $query6 = "SELECT * FROM bwjobs WHERE location LIKE
    '$location' AND description LIKE '$keyword' AND title LIKE
    '$keyword' AND retail_operations = '$retail_operations' AND visible
    = 'yes' ORDER BY posted_date DESC";
    if ($sales_and_marketing){
    $query7 = "SELECT * FROM bwjobs WHERE location LIKE
    '$location' AND description LIKE '$keyword' AND title LIKE
    '$keyword' AND sales_and_marketing = '$sales_and_marketing' AND
    visible = 'yes' ORDER BY posted_date DESC";
    if ($technical){
    $query8 = "SELECT * FROM bwjobs WHERE location LIKE
    '$location' AND description LIKE '$keyword' AND title LIKE
    '$keyword' AND technical = '$technical' AND visible = 'yes' ORDER
    BY posted_date DESC";
    if ($visual_merchandising){
    $query9 = "SELECT * FROM bwjobs WHERE location LIKE
    '$location' AND description LIKE '$keyword' AND title LIKE
    '$keyword' AND visual_merchandising = '$visual_merchandising' AND
    visible = 'yes' ORDER BY posted_date DESC";
    $result = @mysql_query ($query);
    $result2 = @mysql_query ($query2);
    $result3 = @mysql_query ($query3);
    $result4 = @mysql_query ($query4);
    $result5 = @mysql_query ($query5);
    $result6 = @mysql_query ($query6);
    $result7 = @mysql_query ($query7);
    $result8 = @mysql_query ($query8);
    $result9 = @mysql_query ($query9);
    $num = mysql_num_rows ($result, $result2, $result3, $result4,
    $result5, $result6, $result7, $result8, $result9);

    You could just combine all into one query statement as
    follows:
    $query = "SELECT * FROM bwjobs WHERE location LIKE
    '$location' AND description LIKE '$keyword' AND title LIKE
    '$keyword' AND (buying_and_merchandising =
    '$buying_and_merchandising' OR design = '$design' OR executive =
    '$executive' OR human_resources = '$human_resources' OR
    manufacturing = '$manufacturing' OR retail_operations =
    '$retail_operations' OR sales_and_marketing =
    '$sales_and_marketing' OR technical = '$technical' OR
    visual_merchandising = '$visual_merchandising' ) AND visible =
    'yes' ORDER BY posted_date DESC";

  • SQL Trace(ST05) and Run Time Analysys(SE30)

    Hi ABAP Experts,
    Can any one explain what is the use of
    1) ST05 and
    2) SE30
    ponits will be given for clear cut explanation
    Thanks and Regards
    Vijaya

    Hi,
    The SQL Trace part of the Performance Trace tool allows you to see how the OPEN SQL statements that you use in ABAP programs are converted to standard SQL statements (see Embedded SQL) and the parameters with which the embedded SQL statements are passed to the database system.
    Overview
    From the time you turn on the trace function to the time you turn it off again, all database activity occurring either for a specific user or for an entire system is recorded. The SAP System takes OPEN SQL statements and converts them in to embedded SQL statements that it passes to the database and makes the results available. The embedded SQL statement and its parameters are recorded in the SQL Trace file. The results of the SQL statement, like return code, number of entries retrieved, inserted, or deleted by the database are recorded in the SQL Trace file as well. The log file also contains the runtime of the statement and the place in the application program, respectively transaction, from which it was called which enables additional analyses (supported by the SQL Trace function).
    From the recorded SQL trace you can deduce:
    ·        which SQL statements your application carries out
    ·        which values the system uses for specific database accesses and changes
    ·        how the system translates ABAP OPEN SQL commands (such as SELECT) into standard SQL commands
    ·        where your application positions COMMIT statements
    ·        where your application makes repeated database accesses
    ·        what database accesses or changes occur in the update section of your application
    Create a Trace (SQL Server Profiler)
    To create a trace
    On the File menu, click New Trace, and connect to an instance of SQL Server.
    The Trace Properties dialog box appears.
    Note: 
    The Trace Properties dialog box fails to appear, and the trace begins instead, if Start tracing immediately after making connection is selected. To turn off this setting, on the Tools menu, click Options, and clear the Start tracing immediately after making connection check box.
    In the Trace name box, type a name for the trace.
    In the Use the template list, select a trace template on which to base the trace, or select Blank if you do not want to use a template.
    To save the trace results, do one of the following:
    Click Save to file to capture the trace to a file. Specify a value for Set maximum file size. The default value is 5 megabytes (MB).
    Optionally, select Enable file rollover to automatically create new files when the maximum file size is reached. You can also optionally select Server processes trace data, which causes the service that is running the trace to process trace data instead of the client application. When the server processes trace data, no events are skipped even under stress conditions, but server performance may be affected.
    Click Save to table to capture the trace to a database table.
    Optionally, click Set maximum rows, and specify a value.
    Caution: 
    When you do not save the trace results to a file or table, you can view the trace while SQL Server Profiler is open. However, you lose the trace results after you stop the trace and close SQL Server Profiler. To avoid losing the trace results in this way, click Save on the File menu to save the results before you close SQL Server Profiler.
    Optionally, select the Enable trace stop time check box, and specify a stop date and time.
    To add or remove events, data columns or filters, click the Events Selection tab. For more information, see: How to: Specify Events and Data Columns for a Trace File (SQL Server Profiler)
    Click Run to start the trace.
    Runtime analysis:
    se30 is the tcode for run time analysis, is a tool it
    display report based on lps for abap pragram execution
    time , data base load, system load. these are all will
    display in micro seconds.this tool will use for Transaction
    code, function modules , abap progrmes
    In Transaction SE30, fill in the transaction name or the program name which needs to be analyzed for performance tuning.
    For our case, let this be “ZABAP_PERF_TUNING”
    After giving the required inputs to the program, execute it. After the final output list has been displayed, PRESS the “BACK” button.
    On the original SE30 screen, now click on “ANALYZE” button.
    The percentage across each of the areas ABAP/ Database/System shows the percentage of total time used for those areas and load on these areas while running the program . The lesser the database load faster the program runs.
    SQL trace:
    /people/siegfried.boes/blog/2007/09/05/the-sql-trace-st05-150-quick-and-easy
    SE30
    /people/siegfried.boes/blog/2007/11/13/the-abap-runtime-trace-se30--quick-and-easy
    regards,
    vasavi.
    kindly reward if helpful.

  • SQL statement question

    I'm trying compare two table in Oracle and with a firstname and
    lastname
    matches fill-in a pager pin number.
    This is the sql statement I'm running
    update addressbook2 set pin =
    (select pin from phonebook where
    phonebook.firstname=addressbook2.firstname
    and
    phonebook.lastname=addressbook2.lastname)
    where addressbook2.firstname IN (select firstname from phonebook)
    and addressbook2.lastname IN (select lastname from phonebook)
    but I get an error message saying:
    ORA-01427: single-row subquery returns more than one row
    My question is can I update the table even when there are
    duplicates in the tables. The query runs perfect when both
    tables are unique.
    Thank you for any help.
    MN
    null

    I presume you are trying to update the pin of only those people
    in the addressbook table that also exist in the phonebook table.
    The following will work:
    update addressbook a
    set a.pin = (
    select f.pin from phonebook f
    where f.fname = a.fname
    and f.lname = a.lname
    ie remove the last two lines from your DML statement.. be aware
    that you are using a denormalised design (pin is not normalised
    on the primary key) and the use of first and last names as a
    method of identifying people is not a good idea (does 'smith' =
    'smyth'?)
    MN (guest) wrote:
    : I'm trying compare two table in Oracle and with a firstname and
    : lastname
    : matches fill-in a pager pin number.
    : This is the sql statement I'm running
    : update addressbook2 set pin =
    : (select pin from phonebook where
    : phonebook.firstname=addressbook2.firstname
    : and
    : phonebook.lastname=addressbook2.lastname)
    : where addressbook2.firstname IN (select firstname from
    phonebook)
    : and addressbook2.lastname IN (select lastname from phonebook)
    : but I get an error message saying:
    : ORA-01427: single-row subquery returns more than one row
    : My question is can I update the table even when there are
    : duplicates in the tables. The query runs perfect when both
    : tables are unique.
    : Thank you for any help.
    : MN
    null

Maybe you are looking for

  • Retrieving latest record of an employee for a same scholar ID from pa9048

    Hi Experties...! i want to retrieve the latest record from the pa9048 table where the scholar ID (sname) is same for an employee.........can anyone tell me how i do that..........! latest record in means of date.......! Thanx in advance..............

  • Remove scroll bar from html snippet?

    Hello there, I have a little html sniipet on the fron page of my website which includes text from a small Goole sites page. The page it references has some google related text at the bottom ad is thus higher than the size of the snippet 'box'. O my i

  • Import Image in Shockwave

    Good morning, my Shockwave movie downloads a "jpg" into dswmedia folder and then imports this downloaded image into a castmember for Sprite(5). It displays perfectly in Windows (both XP and Vista) but the image does not display in sprite(5) on the ma

  • LR 1.1  Goodwill lost is hard to regain - What motivates Adobe

    Thanks to Adobe for trying to address some of the many criticisms of its LR software. However, the experience of using LR has been an eye opener in regards to what apparently motivates Adobe. Prior to using LR, I viewed Adobe at the high end of corpo

  • Please confirm, before I submit a bug report.

    Open any file in audtion CS6. In a spectral display chose a marquee selection tool and select a portion of the track in the lower frequency range. Adjust gain or delete. Observe the spicke appearing just outside the vertical edges of the selection. A