WF_EVENT.raise  performance issue

Hi,
We have built with JDeveloper an Extension for Applications.
When we execute the function WF_EVENT.raise ... an event is raised that handles the backoffice execution of all data.
The problem is that the process is waiting on the raise object as long as the backoffice program is running. This means that the screen keeps locked as long as the backoffice process is bussy. We just wanted to use events to disconnect the front application and the backoffice process. The event has a subscription that specifies a workflow to be launched.
Does somebody has an valid explanation for this ? Why does the raise event stays locked so long ? Even when I separate the code into an pl/sql block it takes 60 seconds to launch the event.
PL/SQL block used :
declare
l_event_name VARCHAR2(50) := 'oracle.apps.xxint.cp.startup';
l_document_id VARCHAR2(250):= '20CD413C17573E54E04400005A998170';
l_user VARCHAR2(100);
l_retcode NUMBER;
l_retmsg VARCHAR2(2000);
l_paramlist_t apps.wf_parameter_list_t;
BEGIN
wf_event.addparametertolist('INTERFACENAME', 'XXTN_PO_WORK_ORDER', l_paramlist_t);
wf_event.addparametertolist('ECX_DOCUMENT_ID', l_document_id, l_paramlist_t);
wf_event.addparametertolist('ECX_DOCUMENT_NUMBER', 118356, l_paramlist_t);
wf_event.addparametertolist('ECX_DEBUG_LEVEL', '3', l_paramlist_t);
wf_event.raise(
l_event_Name -- p_event_name in varchar2
,118356 eventKey p_event_key in varchar2
, NULL -- p_event_data in clob default NULL
, l_paramlist_t -- p_parameters in wf_parameter_list_t default NULL
, NULL -- p_send_date in date default NULL
l_paramlist_t.DELETE;
COMMIT;
END;

Hi,
When the event is raised, the event manager will process the synchronous subscriptions in the same database session, then return control to the calling code. Any asynchronous subscriptions will be executed when an agent listener is scheduled to run for the deferred agent.
If the subscription starts or continues a workflow, then this will continue to flow in the same database session (as any Workflow does). The flow continues until it can go no further, and then returns control.
So you have two possible solutions to your problem:
1 - customize the workflow to defer to the background once it has been initiated;
2 - change the subscription to be asynchronous (set the phase to > 100).
Of these two, the second one is the best way forward.
HTH,
Matt
WorkflowFAQ.com - the ONLY independent resource for Oracle Workflow development
Alpha review chapters from my book "Developing With Oracle Workflow" are available via my website http://www.workflowfaq.com
Have you read the blog at http://thoughts.workflowfaq.com ?
WorkflowFAQ support forum: http://forum.workflowfaq.com

Similar Messages

  • Inventory Cube performance Issue

    Hi All,
    This is not something new, but an old issue traditionally with this cube. I have customized 0IC_C03 for my requirement and having serious performance issues. It has 0CALDAY, 0MATERIAL, 0PLANT as non-cumulative value parameters. i hav eadded movement types (temproarily for validation purpose). But my query always timed out, unless I specify the material. There are close to 40K materials are being maintained. The values are all fine between ECC and BI afterdata loads. So we are thinking may be snap shot approach would hlp us resolve the performanc eissues.
    Anybody has implementeted snap-shot approach for inventory? I know it is a loading issue, but we think we could deal with that rather than performanc eissue when the users execute the query.
    if anybody has done it, could ou provide the steps?
    Thanks,
    Alex.

    Hi Jameson - Thanks for your response.
    We thought that would be the case. Have raised a SR with oracle and they are investigating on it. We have also sent an EIFF file to Oracle for investigation.
    Both the DBs are in the same environment (AIX 6.1) and DBAs have confirmed both the DBs have the same system parameters.
    Even if we keep aside comparing to 11.2.0.1, for some reason 11.2.0.3 seems to be very slow. Even a simple cube (2 Dim and 2 Measures) with 9K records takes around 15 min to get refreshed and it takes ages to view the data.
    Havent generated the AWR report, will see if we can do the same.
    rgds,
    Prakash S

  • Oracle BPEL 11G performance issue

    Hi
    We are facing performance issues in executing our composite process in Oracle SOA 11g .
    We have installed an admin server and 2 managed servers in cluster in same box. The machine utilization reached almost 95% when i started the admin server and 2 managed server (min n max size of heap given as 1GB each in start up). So i shut down one managed server and increased the JVM size of other to 2 GB and found that the heap size reaches 1.5 GB on start up (observed the heap size using Jconsole)
    The machine capacity is windows server with 4 GB RAM.
    Our process requries multiple records to be processed which are retrieved using Database query.
    We have created 2 composites
    the first composite has 2 BPEL process. the First BPEL 1 executes the DB query and retrieves the result and based on result retrieved we invoke the second BPEL2
    which does around 4 DB calls and passed the result to the next composite. The final BPEL process 3 has multiple select and update query involving DB intensive process.
    When we retrieve 500 records from the BPEL 1 and process , half way through we face out of memory exception. So we are using throttling but even then while executing the process of BPEL3 we are facing out of memory excetion.
    Can you let me know how to find the memory space taken from heap by each BPEL process during it execution. Where in console can i get the memory used details so that i can find which BPEL sis consuming more memory and we can work on optimising.
    Actually we are expecting around 1Lakh and above messages per day and need to check on how this process can handle and also how to increase or determine the capacity of the windows box.
    any immediate help is highly appreciated
    thanks

    Always raise a case with Oracle Support for such issues.
    Regards,
    Anuj

  • Oracle Apps Database severe Performance Issue

    Hi Gurus,
    This is regarding a severe performance issue running in our Production E-Business Suite Instance.
    its an R12.1.3 setup installed with 11.2.0.1 Database. All the servers are Solaris Sparc 64 (Solaris 10)
    Let me brief you about the instance first:
    2 Node Application
    - Main Application Server hosting web/forms/concurrent/admin servers
    - iSupplier server hosting web services (placed in DMZ, used by external suppliers via Internet)
    1 Node Database Server
    Database Server Specs
    Memory: 144G phys mem 20G total swap
    - CPUs (8Px4cores, 2Px2cores)
    - I/O - fiber channel hard disk (hitachi SAN Storage) - 7 DATA_TOPs (7 drives with RAID 5) - current DB size 1.6 TB
    - at peak load, around 1000 concurrent forms session and 2000 web sessions.
    We have been facing some serious performance issues and we raised an SR with Oracle Support.
    The Support analyzed a bunch of AWR Reports we provided them and they asked us to increase the DB_CACHE from its current usage of 27G to 40G
    So, we changed SGA_TARGET from 35G to 50G and PGA was increased from 35G to 40G as v$pgastat was also suggesting some lack of memory.
    We made these changes last night.
    Today morning we observed the following:
    1. after start of office hours, we checked in the home page of EM DB Console that ADDM was showing reduced impact due to lack of SGA memory which seemed to be a good sign. Earlier it was around 25% which was now at 12%.
    However, negative aspects were:
    1. lot of swapping was reported by the System Administrators on the DB Server
    2. High CPU Usage
    3. EM DB Console showed a lot of "Concurrency Wait Class" events ...throughout the day lot of blocking sessions were reported which were making other sessions to wait.
    in the AWR Report, following foreground reports were listed:
    Top 5 Timed Foreground Events
    Event
    Waits
    Time(s)
    Avg wait (ms)
    % DB time
    Wait Class
    DB CPU
    132,577
    61.46
    library cache lock
    3,539
    40,683
    11496
    18.86
    Concurrency
    library cache: mutex X
    4,014,083
    21,011
    5
    9.74
    Concurrency
    db file sequential read
    4,138,014
    20,767
    5
    9.63
    User I/O
    latch free
    381,916
    5,897
    15
    2.73
    Other
    This is showing "library cache lock" events as the main culprit apart from the usual suspect, the CPU.
    I am attaching the AWR Report. Please let me know if  i should revert back the memory changes or is there anything else i could do.
    Please help us resolving it because the performance is going worst.
    Regards,
    Muneer.

    Pl do not post duplicates - Oracle Apps Database severe Performance Issue
    For all critical production issues, pl work with Support thru SRs - using the forums to troubleshoot production issues is not wise

  • Oracle SOA 11g Performance Issue

    Hi,
    We have set up Oracle SOA Suite on AIX environment. Java which we are using is IBM Jdk 1.6. Recently we are hit with performance issue. Frequently we are getting out of memory exception and we need to restart the server and sometimes physically reboot the machine, because out of 16 GB of RAM 4GB we have given as heap space to Admin Server, 7 GB to SOA Server but it is taking more than 7 GB as heap space. On stopping or killing both the services memory is not getting released
    SOA Suite Version : 11.1.1.3
    Instance Node: Single Node
    I collected the logs and tried to analyze in Thread Dump Analyzer and i could see objects(Reserved) is taking 100% of the CPU Utilization.
    We are getting the following error highlighed in the analyzer. There are about 200+ threads got stuck.
    "HTTPThreadGroup-42" prio=10 tid=0x6382ba28 nid=0x20bf4 waiting on condition [0x6904f000..0x6904fb94]
    at sun.misc.Unsafe.park(Native Method)
    at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:146)
    at java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireNanos(AbstractQueuedSynchronizer.java:772)
    at java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireNanos(AbstractQueuedSynchronizer.java:1087)
    at java.util.concurrent.SynchronousQueue$Node.waitForPut(SynchronousQueue.java:291)
    at java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:443)
    at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:475)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:674)
    at java.lang.Thread.run(Thread.java:595)
    "HTTPThreadGroup-41" prio=10 tid=0x6ae3cce0 nid=0x20bf0 waiting on condition [0x68d8f000..0x68d8fc14]
    at sun.misc.Unsafe.park(Native Method)
    at java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:146)
    at java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireNanos(AbstractQueuedSynchronizer.java:772)
    at java.util.concurrent.locks.AbstractQueuedSynchronizer.tryAcquireNanos(AbstractQueuedSynchronizer.java:1087)
    at java.util.concurrent.SynchronousQueue$Node.waitForPut(SynchronousQueue.java:291)
    at java.util.concurrent.SynchronousQueue.poll(SynchronousQueue.java:443)
    at java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:475)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:674)
    at java.lang.Thread.run(Thread.java:595)
    Has anyone faced same issue? We are badly hit with this performance issue in UAT
    Please concider this on high priority and someone please help us
    Regards,
    Sundar

    Always raise a case with Oracle Support for such issues.
    Regards,
    Anuj

  • As a number of Oracle Connections created - Performance Issue

    Hello,
    We are using Oracle 9i as a backend and VB 6.0 as Frontend. For one client one connection will be opened in the Database Server, if 200 clients open the application then 200 connections will be opened in the Database Server, due to that performance issue is raising.
    We are doing: If the frontend application is not in use for 10-minutes then the connection is closed and at the same time front-end application is also terminated or if the Client terminates the application the connection is closed in the Server.
    Eventhought if 200-Clients connect the Server only 4 to 5 Clients send a request to the Server for retriving or for saving the data.
    I want to Open a fixed number of connections in the DataBase Server eg., around 10-connections, and i want to use this 10 connections for 200 clients.
    At First 10 Connections are opened and 10 Clients are using this 10 connections, among 10 Clients 1 released the database connection after finishing the Job and if 11'th Client opens then this released connection as to be used without opening a new connection, in the same way if some of the connections from 9 get free and if any other clients open application then this connections which are free as to be used for opening the application or else for doing any transaction. So, that the number of connection to be create can be redused.
    Please give me the suggestion how to make use the released connection for doing the transaction or else for opening the application.
    Thanking U with Regards,
    Sravan,
    Hyderabad.

    As Satish mentioned, Shared Server can be a good solution, but it would also be worth telling us , how did you quantify this thing that due to the number of connections, you have a performance issue? 200 is not that big number I guess. What's the o/s , exact database version and the system details with database details? Also if you have statspack report, post that too over here.
    HTH
    Aman....

  • Wf_event.raise : database trigger : ORA-04092 : differs from documented behavior

    Product Management :
    Calling wf_event.raise from an after insert row trigger produces ORA-04092. This is exactly opposite from the supposed documented behavior.
    Thanks in advance for your attention to this matter. Needless to say, raising business events from fundamental database activities like inserting & updating rows is critically important.
    Version : 2.6.2 as distributed with 9iAS 9.0.2.0.1 on Win2K
    Error message :
    ORA-04092: cannot ROLLBACK in a trigger
    ORA-06512: at "OWF_MGR.WF_EVENT", line 830
    ORA-04092: cannot SET SAVEPOINT in a trigger
    ORA-06512: at "OWF_MGR.WF_EVENT", line 451
    ORA-06512: at "CASEINT.CG$AIR_FOLDERS", line 7
    ORA-04088: error during execution of trigger 'CASEINT.CG$AIR_FOLDERS'
    From the Oracle Workflow Guide :
    "For environments such as database triggers or distributed
    transactions that do not allow savepoints, the Workflow Engine
    automatically traps Savepoint not allowed errors and defers
    the execution of the activity to the background engine."

    The Business Event System and Workflow Engine are two different components, hence the documentation is actually
    correct as it refers to the 'Workflow Engine' specifically.
    A fix is currently being developed for the problem that you are encountering and will be included in Workflow Standalone
    Version 2.6.3 (iAS 9.0.4).
    There is also a Bug raised for this issue in Workflow 2.6.2 - Bug 2427354 - if you wish to escalate this then please contact
    Oracle Support to raise a TAR against this Bug.
    As a workaround in the meantime you could launch a deferred workflow process that raises the required event.

  • Performance issue with Oracle Text index

    Hi Experts,
    We are on Oracle 11.2..0.3 on Solaris 10. I have implemented Oracle Text in our environment and I am facing a strange performance issue that is happening in our environment.
    One sql having CONTAINS clause is taking forever - more than 20 minutes and still does not complete. This sql has a contains clause and an exists clause and a not exists clause.
    Now if I remove the exists clause and a not exists clause , it completes fast. but with those two clauses it is just taking forever. It is late night so i am not able to post the table and sql query details and will do so tomorrow but based on this general description, are there any pointers for me to review?
    sql query doing fine:
    SELECT
        U.CLNT_OID, U.USR_OID, S.MAILADDR
    FROM
        access_usr U
        INNER JOIN access_sia S
            ON S.USR_OID = U.USR_OID AND S.CLNT_OID = U.CLNT_OID
        WHERE U.CLNT_OID = 'ABCX32S'
        AND CONTAINS(LAST_NAME , 'TO%' ) >0
    --sql query that hangs forever:
    SELECT
        U.CLNT_OID, U.USR_OID, S.MAILADDR
    FROM
        access_usr U
        INNER JOIN access_sia S
            ON S.USR_OID = U.USR_OID AND S.CLNT_OID = U.CLNT_OID
        WHERE U.CLNT_OID = 'ABCX32S'
        AND CONTAINS(LAST_NAME , 'TO%' ) >0
    and exists (--one clause here wiht a few table joins)
    and not exists (--one clause here wiht a few table joins);
    --Now another strange thing I found is if instead of 'TO%' in this sql, if I were to use 'ZZ%' or 'L1%' it works fast but for 'TO%' it goes slow with those two exists not exists clauses!
    I will be most thankful for the inputs.
    OrauserN

    Hi Barbara,
    First of all, thanks a lot for reviewing the issue.
    Unluckily making the change to empty_stoplist did not work out. I am today copying the entire sql here that has this issue and will be most thankful for more insights/pointers on what can be done.
    Here is the entire sql:
    SELECT U.CLNT_OID,
           U.USR_OID,
           S.EMAILADDRESS,
           U.FIRST_NAME,
           U.LAST_NAME,
           S.JOBCODE,
           S.LOCATION,
           S.DEPARTMENT,
           S.ASSOCIATEID,
           S.ENTERPRISECOMPANYCODE,
           S.EMPLOYEEID,
           S.PAYGROUP,
           S.PRODUCTLOCALE
      FROM    ACCESS_USR U
           INNER JOIN
              ACCESS_SIA S
           ON S.USR_OID = U.USR_OID AND S.CLNT_OID = U.CLNT_OID
    WHERE     U.CLNT_OID = 'G39NY3D25942TXDA'
           AND EXISTS
                  (SELECT 1
                     FROM ACCESS_USR_GROUP_XREF UGX
                          INNER JOIN ACCESS_GROUP RELG
                             ON     RELG.CLNT_OID = UGX.CLNT_OID
                                AND RELG.GROUP_OID = UGX.GROUP_OID
                          INNER JOIN ACCESS_GROUP G
                             ON     G.CLNT_OID = RELG.CLNT_OID
                                AND G.GROUP_TYPE_OID = RELG.GROUP_TYPE_OID
                    WHERE     UGX.CLNT_OID = U.CLNT_OID
                          AND UGX.USR_OID = U.USR_OID
                          AND G.GROUP_OID = 920512943
                          AND UGX.INCLUDED = 1)
           AND NOT EXISTS
                      (SELECT 1
                         FROM    ACCESS_USR_GROUP_XREF UGX
                              INNER JOIN
                                 ACCESS_GROUP G
                              ON     G.CLNT_OID = UGX.CLNT_OID
                                 AND G.GROUP_OID = UGX.GROUP_OID
                        WHERE     UGX.CLNT_OID = U.CLNT_OID
                              AND UGX.USR_OID = U.USR_OID
                              AND G.GROUP_OID = 920512943
                              AND UGX.INCLUDED = 1)
           AND CONTAINS (U.LAST_NAME, 'Bon%') > 0;
    Like I said before if the EXISTS and NOT EXISTS clause are removed it works in sub-second. But with those EXISTS and NOT EXISTS CLAUSE IT TAKES ANY WHERE FROM 25 minutes to more than one hour.
    NOte also that it was not TO% but Bon% in the CONTAINS clause that is giving the issue - sorry that was wrong on my part.
    Also please see below the ORACLE TEXT index defined on the table ACCESS_USER:
    --definition of preferences used in the index:
    SET SERVEROUTPUT ON size unlimited
    WHENEVER SQLERROR EXIT SQL.SQLCODE
    DECLARE
       v_err       VARCHAR2 (1000);
       v_sqlcode   NUMBER;
       v_count     NUMBER;
    BEGIN
       ctxsys.ctx_ddl.create_preference ('cust_lexer', 'BASIC_LEXER');
       ctxsys.ctx_ddl.set_attribute ('cust_lexer', 'base_letter', 'YES'); -- removes diacritics
    EXCEPTION
       WHEN OTHERS
       THEN
          v_err := SQLERRM;
          v_sqlcode := SQLCODE;
          v_count := INSTR (v_err, 'DRG-10701');
          IF v_count > 0
          THEN
             DBMS_OUTPUT.put_line (
                'The required preference named CUST_LEXER with BASIC LEXER is already set up');
          ELSE
             RAISE;
          END IF;
    END;
    DECLARE
       v_err       VARCHAR2 (1000);
       v_sqlcode   NUMBER;
       v_count     NUMBER;
    BEGIN
       ctxsys.ctx_ddl.create_preference ('cust_wl', 'BASIC_WORDLIST');
       ctxsys.ctx_ddl.set_attribute ('cust_wl', 'SUBSTRING_INDEX', 'true'); -- to improve performance
    EXCEPTION
       WHEN OTHERS
       THEN
          v_err := SQLERRM;
          v_sqlcode := SQLCODE;
          v_count := INSTR (v_err, 'DRG-10701');
          IF v_count > 0
          THEN
             DBMS_OUTPUT.put_line (
                'The required preference named CUST_WL with BASIC WORDLIST is already set up');
          ELSE
             RAISE;
          END IF;
    END;
    --now below is the code of the index:
    CREATE INDEX ACCESS_USR_IDX3 ON ACCESS_USR
    (FIRST_NAME)
    INDEXTYPE IS CTXSYS.CONTEXT
    PARAMETERS('LEXER cust_lexer WORDLIST cust_wl SYNC (ON COMMIT)');
    CREATE INDEX ACCESS_USR_IDX4 ON ACCESS_USR
    (LAST_NAME)
    INDEXTYPE IS CTXSYS.CONTEXT
    PARAMETERS('LEXER cust_lexer WORDLIST cust_wl SYNC (ON COMMIT)');
    The strange thing is that, like I said, If I remove the exists clause the query returns very fast. Also if I modify the query to use only one NOT EXISTS clause and remove the other EXISTS clause it returns in less than one second.  Also if I remove the EXISTS clause and use only the NOT EXISTS  clause it returns in less than 4 seconds. But with both clauses it runs forever!
    When I tried to get dbms_xplan.display_cursor to get the query plan (for the case of both exists and not exists clause in the query), it said that previous statement's sql id was 0 or something like that so that I was not able to see the query plan. I will keep trying to get this plan (it takes 25 minutes to one hour each time but will get this info soon). Again any pointers are most helpful.
    Regards
    OrauserN

  • Performance issue is most probably caused by the fact that IE can not handle such amount of requests ? is it ?

    Hi,
    I'm running web applications and there is performance issue is most probably caused by the fact that IE can not handle such amount of requests that are forwarded to the server at the same time.
    Data:
    On IE(10) – search variety 11- it takes a long time (about 25 seconds for response) .
    in general, the responding time of "search variety" in IE is much longer than Chrome.
    IE - see:http://screencast.com/t/kSeT3hC2mRfV
    Chrome see:http://screencast.com/t/NYf6fskU
    Why? any solution ?
    Thanks,

    Hi,
    without a link to your site (so we can use the same tools mentioned above)... we can only guess.
    these are peer to peer support forums... your favorite web search engine will help you to learn about web development.
    display the Developer tool console in each browser to view the suppressed error messages.
    All modern browsers support the performance api or your can use the Networking tab to view the latency metrics for the site.... commonly you may be using meta directives for caching instead of server headers or you have different settings between browsers
    which determine how long resources are cached. Commonly, IE has a different security model to other browsers... ensure that you are using the Default IE security zone settings.... Actions like innerHTML go through a sanitization process to remove script content...
    Commonly AJAX calls are made during the onload event...
    Please post questions about web site development to the IE Web Developer forum. Include with your questions a link to your website or a publicly accessible mashup that shows your issue.
    Questions regarding Internet Explorer 8, 9 and 10 and Internet Explorer 11 for the IT Pro Audience. Topics covered are: Installation, Deployment, Configuration, Security, Group Policy, Management questions. If you are a consumer looking for answers or to
    raise a question, it's highly recommended you head on over to http://answers.microsoft.com
    Rob^_^

  • Performance Issues during Upgrade of EBS from 11.5.10.2 to 12.1.1

    Hi,
    We're upgrading our EBS , from Rel 11.5.10.2 to 12.1.1.
    We're stuck , while running script ar120bnk.sql (ran more than 20 Hours) :
    Regarding the tables involved in this Process :
    select owner , table_name,num_rows,last_analyzed,sample_size
    from dba_tables
    where table_name in
    'RA_CUSTOMER_TRX_ALL',
    'RA_CUST_TRX_LINE_GL_DIST_ALL',
    'XLA_UPGRADE_DATES',
    'AR_SYSTEM_PARAMETERS_ALL',
    'RA_CUST_TRX_TYPES_ALL',
    'RA_CUST_TRX_LINE_GL_DIST_ALL',
    'XLA_TRANSACTION_ENTITIES_UPG')
    AR RA_CUSTOMER_TRX_ALL 55,540,740 04/02/2012 12:41:56 5554074
    AR RA_CUST_TRX_LINE_GL_DIST_ALL 380,513,830 04/02/2012 13:54:12 38051383
    AR RA_CUST_TRX_TYPES_ALL 90 04/02/2012 14:04:54 90
    AR AR_SYSTEM_PARAMETERS_ALL 6 04/02/2012 12:19:49 6
    XLA XLA_UPGRADE_DATES 4 05/02/2012 17:12:57 4
    As you can see: RA_CUST_TRX_LINE_GL_DIST_ALL is more tan 380 million rows !
    and RA_CUSTOMER_TRX_ALL is more than 55 million rows.
    We have more huge tables in the AR schema , and we would like to know if we are unique customer
    with huge AR schema objects , and if NOT how come that we are getting stuck on threed statment in
    AR schema.
    Bellow an output of all the objects that have more than 10 million rows in AR schema :
    select owner , table_name,to_char(num_rows,'999,999,999') ,last_analyzed
    from dba_tables
    where owner = 'AR'
    and num_rows > 10000000
    order by num_rows desc nulls last
    AR AR_DISTRIBUTIONS_ALL 408,567,520 04/02/2012 11:49:57
    AR RA_CUST_TRX_LINE_GL_DIST_ALL 380,513,830 04/02/2012 13:54:12
    AR MLOG$_AR_CASH_RECEIPTS_ALL 310,777,690 04/02/2012 12:30:33
    AR RA_CUSTOMER_TRX_LINES_ALL 260,211,090 04/02/2012 13:30:26
    AR AR_RECEIVABLE_APPLICATIONS_ALL 166,834,930 04/02/2012 12:16:54
    AR MLOG$_RA_CUSTOMER_TRX_ALL 150,962,980 04/02/2012 12:33:23
    AR AR_CASH_RECEIPT_HISTORY_ALL 145,737,410 04/02/2012 11:40:31
    AR RA_CUST_TRX_LINE_SALESREPS_ALL 130,287,580 04/02/2012 14:03:54
    AR AR_PAYMENT_SCHEDULES_ALL 108,652,480 04/02/2012 12:05:32
    AR RA_CUSTOMER_TRX_ALL 55,540,740 04/02/2012 12:41:56
    AR AR_CASH_RECEIPTS_ALL 53,182,340 04/02/2012 11:29:53
    AR AR_DOC_SEQUENCE_AUDIT 52,865,150 04/02/2012 11:52:46
    AR RA_MC_TRX_LINE_GL_DIST 17,317,730 04/02/2012 14:05:18
    AR AR_MC_DISTRIBUTIONS_ALL 13,037,030 04/02/2012 11:53:35
    AR AR_MC_RECEIVABLE_APPS 12,672,050 04/02/2012 11:53:57
    AR AR_TRX_SUMMARY 12,457,560 04/02/2012 12:20:16
    AR RA_CUST_RECEIPT_METHODS 11,105,750 04/02/2012 13:35:38
    AR HZ_ORGANIZATION_PROFILES 10,271,640 04/02/2012 12:24:44
    How to Upgrade AR tables whith Huge amount of Datas ( > 50 Millions Rows ) ?

    Hi,
    Dont worry, you are not the only one even we have one customer whose AR_DISTRIBUTIONS_ALL table is 80 GB now and i can even do select count(*) for this table.
    We had to keep this much data for business requirements, but i wonder if this is a bug or users mistake.
    Due to this we are facing seruios performance issues for AR reports and raised SRs but no resolution yet. And this guy who is assigned to us is reall ynot been helpful to fix the issue.
    Although we did not upgrade for this customer, but we migrated from 11.5.9 to R12.1.1 by re-implementation. But all these increasing size of these tables happened after migration.
    And i believe most of the time in your upgrade is going to building the indexes. You can ask Oracle if they can edit the driver file to skip building the indxes and rebuild them after upgrade. But again it will also take time.
    Another option for you is to "Archive and Purge the data" as per chapter 10 of
    11i Receivables user guide.
    http://docs.oracle.com/cd/B25284_01/current/acrobat/115arug.zip
    Thanks
    Edited by: EBSDBA on Feb 8, 2012 10:04 PM

  • ATPO Check and performance Issue

    Hi All
    We had performance issues for ATP check from R/3 to APO and while we asked SAP for help SAP responded saying the data volume for table TTSTR is high and we need to delete unnecessary data by running report -/SAPAPO/SCHED_DELETE_TSTRS and ZZSTRDELE. After running these two reports in test environment we found the data volume has reduced to 96 records from 105000 records. We asked SAP to know whether this is safe to run this and got response that we need to be carefull while running these as data is permanently deleted by this
    Can anybody tell whether its completely safe to run these two reports or we may delete genuine Time Streams by running these?
    Regards
    Mridul

    Mridul,
    It seems very odd that you would open a dialogue with SAP, and then when you needed further information about their proposal, you didn't go back and ask them the question you have now posted here.
    Although I have only run this report a couple of times, I have never had any problem  running report  /SAPAPO/SCHED_DELETE_TSTRS  For more info, refer to their supporting note.
    https://service.sap.com/sap/support/notes/1254606
    I can't comment on report ZZSTRDELE, since I have never heard of it, nor does it exist in any APO system I am currently working on.
    If you have doubts about SAP's recommendations,  I suggest you go back to your message and respond to their last reply, or if the old message is closed, raise a new message.  That's what you pay your support money for.
    Best Regards,
    DB49

  • Voyager Performance Issues

    We are having slow response times with BO Voyager, I'm not sure if this is due to the distance between the client session (US) and the physical location of the Server (Europe). Has anyone else experienced this before?
    Are there any general performance improvement measures that can be adopted/employed for Voyager as there doesn't seem to be any obvious measures that can be adopted?
    Is the performance entirely dependent on the speed of the OLAP connections supporting the Voyager workspaces
    Thanks

    Hello Shariff
    You don't mention what OLAP server you are using or any specifics about performance. Are your comments based on a comparison against another tool. The network distance won't help. However are all components located in Europe (OLAP server, BOE and MDAS) or just the OLAP server?
    The following are general guidelines that we follow when tracking down performance issues. I am assuming you are using Microsoft Analysis Services.
    Many of the performance issues raised against Voyager turn out to be problems with cube design. Microsoft has fairly extensive resources talking about best practices and performance tuning so it is always good to make sure customers are aware and are using these best practices. This post contains links to some of the resources:
    MSAS 2005
    [http://www.microsoft.com/technet/prodtechnol/sql/bestpractice/ssasqptb.mspx |http://www.microsoft.com/technet/prodtechnol/sql/bestpractice/ssasqptb.mspx] and
    [http://download.microsoft.com/download/8/5/e/85eea4fa-b3bb-4426-97d0-7f7151b2011c/SSAS2005PerfGuide.doc|http://download.microsoft.com/download/8/5/e/85eea4fa-b3bb-4426-97d0-7f7151b2011c/SSAS2005PerfGuide.doc]
    MSAS 2008
    [http://www.microsoft.com/downloads/details.aspx?FamilyID=3be0488d-e7aa-4078-a050-ae39912d2e43&DisplayLang=en|http://www.microsoft.com/downloads/details.aspx?FamilyID=3be0488d-e7aa-4078-a050-ae39912d2e43&DisplayLang=en]
    As for the cube optimization, here's the advise from Microsoft about how to speed up the cube response time:
    These are the two whitepapers that I recommend for background on performance tuning SSAS 2005:
    u2022         [Identifying and Resolving MDX Bottlenecks|http://tinyurl.com/33uxob]: This was produced by the SQL Customer Advisory Team, with input from a variety of sources. This is very focused on query tuning. It will explain how to determine whether a bottleneck is in the storage engine (responsible for retrieving data from partitions and aggregations) or in the formula engine (responsible for pretty much everything else). It explains relevant information about these components and how they work, and provides guidance on tuning issues in either component. Frankly, tuning problems in the storage engine is a much simpler task than tuning problems in the formula engine.  <URL:  http://sqlcat.com/whitepapers/archive/2007/12/16/identifying-and-resolving-mdx-query-performance-bottlenecks-in-sql-server-2005-analysis-services.aspx >
    u2022         [SQL Server 2005 Analysis Services Performance Guide|http://tinyurl.com/yr5hrv]: This predates the previous whitepaper and discusses performance more generally, including query and processing performance. It provides some very important guidance, especially with relation to cube design, that will help to achieve better performance with SSAS.
    Once you identified the query that's running very slow, then download Microsoft Business Intelligence Developer Studio Helper ([http://www.codeplex.com/bidshelper |http://www.codeplex.com/bidshelper]) and MANUALLY modify the aggregation at the level that takes the longest time.  Here's the writeup from Microsoft about step-by-step guide to BIDS: <see attachment>.
    Here is the general strategy we want to pursue to see if we can get Voyager to perform faster:
    First it would be good to get a better quantification of the problem. For example are they experiencing Voyager being in general u201Cxu201D times slower than Excel or another BI tool? And does this mean:
    u2022             Most operations in Excel take less than a second and most Voyager operations take 5 or 6 seconds
    u2022             Operations that take 10 seconds in Excel take around about a minute in Voyager
    u2022             Operations that take a minute in Excel take five or six minutes in Voyager
    u2022             All of the above
    The first answer is most likely to be Voyageru2019s scalable N tier environment versus Excelu2019s 2 tier environment. The others more likely suggest problems with the way that Voyager is retrieving the data from the OLAP server. However, these are not hard and fast rules as the reason for a performance difference is often workflow specific.
    The second step would be to do a sanity check on their environment. For example, make sure they havenu2019t deployed the web app server, mdas server on a VM machine with only 500 megs of ram. The things to check would be  CPU and memory usage:
    u2022             On the browser
    u2022             On the web application server machine
    u2022             On the MDAS machine
    u2022             On the Analysis Services machine (though this it could be assumed is OK if the performance of other tools is acceptable)
    The objective here would be to see if they have an under resourced environment. Unfortunately we donu2019t have a Voyager sizing guide so I cannot give pro-active recommendations about what things should look like other than to say obvious things like u2018if your MDAS machine is running out of memory with one user you need more memoryu2019. Our very general comment is that one MDAS can support 15 queries.
    The final step would be to identify any specific workflows that demonstrate the problem, in particular any workflows that demonstrate an extreme difference between Voyager and the other products, especially if itu2019s a high value workflow to the customer. The objective here would be to do some detailed profiling to see if there is an opportunity for a bug fix which could be released as a patch.
    It would also be good to know which version Voyager they are using (which fix packs etc.), which browser they are using, and their deployment environment (OS, how many machines, CPU speed, memory). There are various fixes that give better performance for specific  workflows that have gone into Voyager over time, so (without specific knowledge of what the problems are) it would be great to see if an XI 3.1 version of Voyager meets their expectations.
    General performance problems are unlikely to be fixed with bug fixes , and having a better system configuration will only take you so far before you hit the limits of what you can get out of an N tier system opposed to a 2 tier thick client one. Historically we have tended to find very specific problems which we were able to identify the root cause of and fix by issuing patches.
    I hope this helps somewhat.
    Regards

  • "Test Error" - "raise the issue with your service ...

    Help!
    I have had performance problems on my broadband for quite a while but not been able to progress. The usual number based test from BT says we should get between 4 and 7 mbps. In reality we struggle to get 2 mbps, frequently lanquishing at 1.5 mbps.
    I have been onto http://speedtester.bt.com/ to get analysis on the line (previously having had problems with our profile) but whenever I try to run a test (over several weeks) I now always get the reply:
    The Performance Tester is currently unable to run a speed test for your broadband connection. Please try again shortly, however if this problem persists, raise the issue with your service provider.
    Following up on this with the Indian call centre was an exercise in futility, they just couldn't understand what I was talking about and just wanted to follow the usual script for speed problems. The last time I tried they said they'd found an error at the exchange which they would fix shortly. The result of the "fix" was that our phone and broadband disappeared for 4 days and resolving this came with the usual threat of charging us £99 is the problem wasn't BTs fault. The fault was fixed without any explanation but subsequently I've been reluctant to follow up the problem through this route.
    I don't know how to go about getting someone to fix the problem without being able to run the speed tests that BT will accept.
    I would greatly appreciate some assistance from this usually very helpful group.
    Thanks,
    Andrew
    Solved!
    Go to Solution.

    John,
    Thank you for answering.
    The details from my ASDL statistics are below. I cannot provide a BT Speedtester result - this is one of the problems I was raising. I have done a quiet line test and there is no noise. I don't have an internal master socket; I have an external grey box from which a number of wires run to various sockets.
    Thanks,
    Andrew
    ADSL Line Status
    Connection Information
    Line state:
    Connected
    Connection time:
    1 days, 16:42:25
    Downstream:
    4.063 Mbps
    Upstream:
    448 Kbps
    ADSL Settings
    VPI/VCI:
    0/38
    Type:
    PPPoA
    Modulation:
    G.992.1 Annex A
    Latency type:
    Interleaved
    Noise margin (Down/Up):
    9.6 dB / 20.0 dB
    Line attenuation (Down/Up):
    44.0 dB / 25.0 dB
    Output power (Down/Up):
    20.1 dBm / 12.4 dBm
    FEC Events (Down/Up):
    4667988 / 202
    CRC Events (Down/Up):
    4713 / 192
    Loss of Framing (Local/Remote):
    0 / 0
    Loss of Signal (Local/Remote):
    0 / 0
    Loss of Power (Local/Remote):
    0 / 0
    HEC Events (Down/Up):
    86618 / 162
    Error Seconds (Local/Remote):
    33661 / 5662

  • Performance issue SCSM 2012 R2

    Hellol
    We have SCSM 2012 R2 (no RU), we have a performance issue, when we set implemener for Manual Activity and press OK, CPU utilization raises to 25-30% for 1,5 minutes and during this period console is not working. What can be a reason of it? On the SQL Servers
    (CMDB) there are enough resources, servers are virtual (cluster mode). How can we investigate this issue? SCSM is not used in the production yet.
    Disabled 4 subscriptions for manual activity, but did not help.

    In the UR2/UR3 of SCSM 2012 R2 is a console performance fix. You should consider to install the latest UR and give it a try again. 
    Andreas Baumgarten | H&D International Group
    We installed UR3 today and problem is not solved, also we defined problem more clear, it happens when we specify user AD login without pressing check button even.... also same problem when we press "..."button to find AD users. So problem not when
    we save MA, it is related to searching AD users. SQL servers have 8 cores each, cluster mode, enough free space. "Users" tab shows ~57 000 user accounts.

  • ALI Scripting - Performance Issues

    We are using ALI scripting to raise events so that other portlets in the page can listen to the events and also passing some data to other portlets using "PTPortlet.setSessionPref" scripting API. In my local testing, the scripting is taking just 0.3 seconds, however when we deployed this code to common portal which has few other portlets, the scripting is taking 5 seconds. Does any one know if there are any known best practices around ALI scripting to avoid performance issues especially to transfer events between portlets?
    Thanks
    Sampath

    Hi, I would like to provide additional information on the performance issue we are facing with ALI scripting. In the code highlighted below, we are using “setSessionPref” methods to set values in session variables using ALI scripting APIs and raising events for IPC.
    var prefName = 'selectionString';
    var prefValue = xmlFile;
    var gpConfigValue = "<%=gpPromptsConfigID%>";
    PTPortlet.setSessionPref(prefName,prefValue);
    PTPortlet.setSessionPref('gpFormObj',frmObj.name);
    PTPortlet.setSessionPref('gpPromptsConfigID', gpConfigValue);
    myportlet$$PORTLET_ID$$.raiseEvent('onSelectionSubmitFormObj',false);
    myportlet$$PORTLET_ID$$.raiseEvent('onSelectionSubmit', false);
    This code does not take lot of time(less than a second) to execute in my local machine however, it takes lof of time (around 5 seconds) in integrated development environment.
    After lot of debugging, I observed that if I change “PTPortalContext.GET_SESSION_PREFS_URL” to point to http://localhost:7001 instead of integrated development environemnt, the processing time is considerably reduced. This code below is inserted automatically by ALUI in every portal page. Does any one know the signficance of the below code automaitcally inserted by ALUI and how it can impact IPC.
    // Define PTPortalContext for CSAPI
    PTPortalContext = new Object();
    PTPortalContext.GET_SESSION_PREFS_URL = 'http://dcdev.pg.com/portal/server.pt?space=SessionPrefs&control=SessionPrefs&action=getprefs';
    PTPortalContext.SET_SESSION_PREFS_URL = 'http://dcdev.pg.com/portal/server.pt?space=SessionPrefs&control=SessionPrefs&action=setprefs';
    Thanks
    Sampath

Maybe you are looking for