Retry recovery action is taking long time to complete

Hi,
I am in the process of testing error handling for my BPEL process.
As part of it, I have given the default recover action in the fault policy of the process as retry with a retry count of 3 times with retry interval as 2.
To test this, i have changed the target end point uri of the target service to some invalid url and ran the process.
First I was not able to find the BPEL process instance at all in the console.
Later I was able to find the instance after half an hour. When I saw the instance, if retry had happened or not.
I found that the retry happened for 3 times, but it took 30 mins for the first retry to complete and start the second retry.
Can anyone please let me know why the retry is taking long time although the retry interval I gave is only 2 seconds.
THanks in advance!!!!

Hi,
Please find below the fault policy file:
<?xml version="1.0" encoding="UTF-8"?>
<faultPolicy version="2.0.1" id="TestFaultPolicy" xmlns:env="http://schemas.xmlsoap.org/soap/envelope/" xmlns:xs="http://www.w3.org/2001/XMLSchema" xmlns="http://schemas.oracle.com/bpel/faultpolicy" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
     <Conditions>
     <faultName>
     <condition>
     <action ref="aia-ora-retry"/>
     </condition>
     </faultName>
     </Conditions>
     <Actions>
          <Action id="aia-ora-retry">
               <retry>
                    <retryCount>3</retryCount>
                    <retryInterval>2</retryInterval>
                    <exponentialBackoff/>
                    <retryFailureAction ref="aia-ora-java"/>
                    <retrySuccessAction ref="aia-ora-java"/>
               </retry>
          </Action>
          <!-- This is an action will cause a replay scope fault-->
          <Action id="ora-replay-scope">
               <replayScope/>
          </Action>
          <!-- This is an action will bubble up the fault-->
          <Action id="ora-rethrow-fault">
               <rethrowFault/>
          </Action>
          <!-- This is an action will mark the work item to be "pending recovery from console"-->
          <Action id="ora-human-intervention">
               <humanIntervention/>
          </Action>
          <!-- This action will cause the instance to terminate-->
          <Action id="ora-terminate">
               <abort/>
          </Action>
          <Action id="aia-ora-java">
          <javaAction className="oracle.apps.aia.core.eh.BPELJavaAction" defaultAction="ora-rethrow-fault">
          <returnValue value="REPLAY" ref="ora-terminate"/>
          <returnValue value="RETRHOW" ref="ora-rethrow-fault"/>
          <returnValue value="ABORT" ref="ora-terminate"/>
          <returnValue value="RETRY" ref="aia-ora-retry"/>
          <returnValue value="MANUAL" ref="ora-human-intervention"/>
          </javaAction>
</Action>     
     </Actions>
</faultPolicy>

Similar Messages

  • Process chain taking long time to complete

    Hi,
    I am having following issues in the process chain daily loads
    1) The PSA deletion is taking very long time on fridays only (nearly 3 hrs). Other days it gets deleted in 1 hr max.
    2) Just to load 381 records via dtp from DSO to cube takes nearly 2 hrs(Delta load). No major routines are written in the transformation.
    How to analyse a process in process chain which does not fail, but takes very long time to complete. Is there any tool or transaction in BW which can help us to analyse why the process chain takes so long time to complete in different days. One day it completes in 8 hrs, another day it takes 12 hrs.
    None of the transactions, SM37, SM50, SLG1, LOGS etc are giving me any help to analyse this issue.
    If it fails we have error logs to check and analyse, but without failing how we can analyse and fix the delay and reduce the time of loading data. Please guide me.
    Thanks in advance.
    Vishwanath

    Hi.......
    1) This might be due to the poor performance of the system ..there wont be enoughwork processes avilabe in teh system .- What needs to be done if it does not have enough work processes available. Though I can find that lots of dialog processes were free during these jobs as well some Background processes were also free.?
    Look to increase Work Process.......................u can do two things..............U can cancel some Job...........which r not progressing...........or which r not very important for the time being..........
    Or if possible........u can increase number of servers..........
    Look When a Job will start...............first it will run in background.............u can start in in Dialog also..........but for Dialogue after some time ..load will go to time out.................now many Child jobs will be created for a single background jobs..............they may run in Dialogue..........u can monitor them through SM66............
    2) Check ST04 for lock waits and dead locks ..if the same lock persist for long then check with basis team.
    What needs to be conveyed to the basis team when these locks happen?
    generally these will be temporary Locks.............after some time locks will be realeased..............if they persist for a long time...........then u can contact Basis people..........Or if u can understand that which job is creating the lock.............like if u double click on the job >> click on Job details...............from there u will find the PID................copy that.............and check in ST04..........whether they matches or not..............now suppose it match.............and u know that job is not very important............then u can cancel the job.............
    Check also the table space availba in St04 . the usuage should nt be more than 90%
    What needs to be conveyed to the basis team if the usage is more than 90%.?
    U can ask them to increase the Table space........
    Check the sm37 >> delay column it should nt be high
    What needs to be done if delay column is high. Actually it is high for these jobs (PSA deletion and load from DSO to Cube visa DTP (Delta load)?
    If Delay is more.............it means ...........that there is no free work process...jobs r going to Release state............solution I had already given u.......
    Check the PID refeclting in lock waits in sm66 wether they are progressing or nt ..
    If they are not progressing what action needs to be taken?
    Actually...........suppose one Dialogue job is running............it may change..........ie  that Dialogue job will go to Stop status.........some other job will start............anyways if it does'nt progress............u can cacel the JOb.........
    Check OS07 for DB if the idle time is less then 20% then its a problem ..
    If the Idle time is less than 20% what action needs to be taken?
    U can contact the Basis people.......
    Check sm21 and RFC connection with other source systems in sm59 ..
    What needs to be checked in SM21 and SM59 specifically? What parameter i need to check?
    In SM21.............if red status is there............check the Log beside that..............may be it will be some Terminal disconnected.......
    In SM59...........clck on test connection..........
    if the problem is occuring due to only one source system then check the performance of that system ..
    How to check the performance of these systems? Any tools available in R/3 system to check its performance?
    No.....in this way u can check performance for all the source system...........go to SM59............double click on the desired source system>> click on test connection......
    If All these are persisting then its a performance problem and check with basis team ..
    Is there any special settings which needs to be maintained to achieve better performance of Process chain loads?
    U can Improve the performance of process chain by Parallel Processing..........ie split the loads by giving selections...........and execute thenm in parallel..........
    Regards,
    Debjani........

  • VB Macro in Bex Analyser is taking long time to complete execution

    Hi Experts,
    In a FI query , we have a VB macro which update the excel sheets by taking values from the previous excel sheets .
    The issue is its taking long time for query execution and if we are keep on pressing 'ENTER' button . The query is running very fastly and is giving the results ,but its not a correct way to do.
    Its a critical issue in Production and can anyone help me to resolve the issue.
    Regards,
    Anju

    Hi Anju,
    I need to create a VB macro in one of my query but i not familiar how i can plug the VB code into the query.
    Please can you give me some basic procedure on how to do that?
    What i need to do exactly is to bring a KF from a query into another query. Not sure if this is possible using VB.
    Let me know.
    Thanks.
    Cesar

  • Master Dataload is taking long time to complete

    Hi
    In last couple of weeks, we are experience the delay for master dataload job. These dataload should take 5-8 minute to complete but it is taking more then 5 hour to complete. here we are using process chain for master dataload.
    Any idea? We don't understand where to start to check the root cause of this problem.
    Regards
    Amar

    Hi Freinds
    Some update of our master dataload. Still not resolved.
    Seem to me, our problem is related to Hierarchy in BW server.It is taking long time to update the hierarchy in BW server. As i told you, problem occur only when loading master dataload. For transaction dataload, no IDOC get stuck in R/3 (NPR) server. There are enough resources in both R/3 and BW server for trfc.
    There are total 9 hierarchy in the master dataload that is delaying in the process for update. If you have seen this type of problem for the following hierarchy, pls update me.
    A     0ACCOUNT (InfoSource)
         Hierarchy
    1     Fin statement vers FSV5
    2     Cost Element by fun area
    B     0COSTCENTER (InfoSource)
         Hierarchy
    3     Mitel Network
    4     NA MNS Total
    5     CEO
    6     Global Support Cost center
    C     0COORDER (infosource)
         Hierarchy
    7     Total R&D (Ron Wellard)
    8     RD YRD1
    9     RD YRD2
    Regards
    Amarjit

  • Requisition Import is taking long time to complete

    Hi,
    I'm working with EBS 11.5.10 CU2, I have Requisition Import program scheduled, usually it's completing normal in around 5minutes; but sometimes it's taking around 30 minutes to complete. I compared the output files of program completing normal and taking long time, latter one has processed less requisitions than former one.
    Could somebody help what could be possible reasons to check for program to take long time to complete.
    Please help.
    Thanks

    945756 wrote:
    Hi,
    I'm working with EBS 11.5.10 CU2, I have Requisition Import program scheduled, usually it's completing normal in around 5minutes; but sometimes it's taking around 30 minutes to complete. I compared the output files of program completing normal and taking long time, latter one has processed less requisitions than former one.
    Could somebody help what could be possible reasons to check for program to take long time to complete.
    Please help.
    ThanksDo you have the statistics collected up to date?
    Please enable trace and generate the TKPROF file to find out why it takes that long -- https://forums.oracle.com/forums/search.jspa?threadID=&q=TKPROF+AND+Concurrent&objID=c3&dateRange=all&userID=&numResults=15&rankBy=10001
    Thanks,
    Hussein

  • BAPI_MOSRVAPS_SAVEMULTI3 Taking longer time to complete

    I am calling this bapi passing 100k records in the background job and it takes more than 5hrs to complete. Is there any way this run time can be reduced.
    I cannot slice the file as I need to create the master schedule either all or none.
    also I tried 'In Background' while calling the BAPI using the qRFC but this transaction async will not return the result back us.
    So my question to here is how do we get the return data from the method of calling the bapi.
    Any help/suggestion  is appreciated.

    Dear Balamuru,
    there is no generally applicable procedure for reducing runtime in this BAPI.
    The best would be to make a trace in ST12 to see where the performance issue really happens.
    Also please check if note 1276375 can help further.
    I hope I could help.
    Regards,
    Tibor

  • 11.5.10.2 -12.1.1 Upgrade- American English patch driver taking long time

    Hi,
    I have performed test upgrade from 11.5.10.2 to 12.1.1.It was successful.
    The American English Upgrade patch driver had taken 20 hours time to complete.
    Now i am doing a second test run by following the same steps, but now the same patch is taking long time. It is running from past 24 hours.
    And still 54000 jobs to complete.
    The main difference between my two test runs is
    1.For first instance both database and apps are in one server. (single node)
    2.For this new test instance, we have database running one server and 12.1.1 file system on another server.
    Will it make any difference? How can i diagnose the issue?
    The patch log is not showing any errors, but taking long time to complete.
    Thanks,

    Hi,
    From m note 12.1.1 UPGRADE FROM 11.5.10.2 VERY SLOW XDOLOADER TASK HANGS [ID 1065393.1],
    after changing mv /dev/random /dev/random_bak
    ln -s /dev/urandom /dev/random
    do i need to reboot the server?
    After this can i continue the previous session of the patch? or start new session?
    Thanks,

  • SSRS Reports taking long time to load

    Hello,
    Problem : SSRS Reports taking long time to load
    My System environment : Visual Studio 2008 SP1  and SQL Server 2008 R2
    Production Environment : Visual Studio 2008 SP1  and SQL Server 2008 R2
    I have created a Parameterized report (6 parameters), it will fetch data from 1 table. table has 1 year and 6 months data,      I am selecting parameters for only 1 month (about 2500 records). It is taking almost 2 minutes and 30 seconds
    to load the report.
    This report running efficiently in my system (report load takes only 5 to 6 seconds) but in
    production it is taking 2 minutes 30 seconds.
    I have checked the Execution log from production so I found the timing for
    Data retrieval (approx~)       Processing (approx~)               Rendering (approx~)
    10 second                                      15 sec                        
                2 mins and 5 sec.
    But Confusing point is that , if I run the same report at different time overall output time is same (approx) 2 min 30 sec but
    Data retrieval (approx~)       Processing (approx~)                Rendering (approx~)
    more than 1 min                            15 sec                                     
    more than 1 min
    so 1 question why timings are different ?
    My doubts are
    1) If query(procedure to retrieve the data) is the problem then it should take more time always,
    2) If Report structure is problem then rendering will also take same time (long time)
    for this (2nd point) I checked on blog that Rendering depends on environment structure e.g. Network bandwidth, RAM, CPU Usage , Number of users accessing same report at a time.
    So I did testing of report when no other user working on any report But failed (same result  output is 2 min 30 sec)
    From network team I got the result is that there is no issue or overload in CPU usage or RAM also No issue in Network bandwidth.
    Production Database Server and Report server are different (but in same network).
    I checked that database server the SQL Server is using almost Full RAM (23 GB out of 24 GB)
    I tried to allocate the memory to less amount up to 2GB (Trial solution I got from Blogs) but this on also failed.
    one hint I got from colleague that , change the allocated memory setting from static memory to dynamic to SQL Server
    (I guess above point is the same) I could not find that option Static and Dynamic memory setting.
    I did below steps
    Connected to SQL Server Instance
    Right click on Instance go to properties, Go to Memory Tab
    I found three options 1) Server Memory   2) Other memory   3) Section for "Configured values and Running values"
    Then I tried to reduce Maximum  Server memory up to 2 GB (As mentioned above)
    All trials failed, this issue I could not find the roots for this issue.
    Can anyone please help (it's bit urgent).

    Hi UdayKGR,
    According to your description, your report takes too long to load on your production environment. Right?
    In this scenario, since the report runs quickly in developing environment, we initially think it supposed to be the issue on data retrieval. However, based on the information in execution log, it takes longest time on rendering part. So we suggest you optimize
    the report itself to reduce the time for rendering. Please refer to the link below:
    My report takes too long to render
    Here is another article about overall performance optimization for Reporting Services:
    Reporting Services Performance and Optimization
    If you have any question, please feel free to ask.
    Best Regards,
    Simon Hou

  • Connecting to the database taking long time to connect database server

    Hi
    When I execute procedure i am getting the below message at bottom of the Oracle SQL Developer
    "Connecting to the database"
    it is taking more than 10 min plz guide

    Hi
    have you installed a normal Oracle Client also on your Host? normal Oracle Client
    Did you connect with host:port:sid or with a Oracle Naming Service? through TNS Service
    Can you test tnsping <alias> yes, It is working fine
    Did other user have the same problem? yes
    Did you connect through WAN or LAN connection? LAN (Intranet)
    Can you tell more about you client/database setup?
    Database setup:
    OS: Window 2008 server
    version: 11.1.0
    Client: 11.1.0
    OS: Window 2008 server
    Now I am not able to execute single select query which table contains 6 records and 15 columns it is taking long time I have waited 30 min still no resutls
    only one table is behaving like this remaining is working fine
    Edited by: user9235224 on Oct 6, 2012 7:06 PM

  • The ODS activation is taking long time

    Hi,
    We are on SAP NetWeaver BI 701 (Support Package 5).
    We create a Z ODS, it will contain a lot of data (180.000.000 month-end) and we want to generate specific reports about it.
    The activation is taking long time, I assume is because we checked the flag "SIDs Generation upon Activation". I am confused about this check. I really need it? is this check the only problem.
    Thanks for you help.
    Victoria

    Hi Victoria:
       If your Z DSO is used only for staging purposes (you don't have queries based on this DSO and you send the data to another DSO or to an InfoCube) then you don't need to check the "SIDs Generation Upon Activation" box.
    Even more, to achieve better performance during data loads in this scenario, you might consider using a Write Optimized DSO instead of a Standard DSO, but if you decide to take this alternative don't forget to select the "Do Not check Uniqueness of Data" box if you need to write several records with the same Semantic Key.
    Regards,
    Francisco Milán.

  • F4 Help is taking long time

    Hi All,
    We are working on BI 7.0. version
    In the varaible pop-up screen we have two info objects.
    1. Fiscal year Period
    2. JOA(Joint operating aggriment)
    If u press F4 for JOA, it is taking long time to execute and finally the application is getting closed.same situation is there in RSRT also.
    If i enter with out JOA the query is giving the output. Here i have to restrict the query by JOA.
    i have changed the JOA peroperties in query designer.
    Query execution for filter value selection = Values in master data table.......
    but still the situation is the same.......
    Could you please suggest any solution for this.....
    Thanks & Regards,
    PK

    Hi Kamal,
    You can set that at the query level in the query designer for each query.
    1. Select the corresponding characteristic in the query designer.
    2. Goto to the "Extended tab" in the properties
    3. Select the "Values in the Master data table" in the "Query execution in the filter value selection.
    Also see some recomendations:
    Note 748623 - Input help (F4) has a very long runtime - recommendations
    Hope this helps.
    CK

  • Update ztable is taking long time

    Hi All,
    i have run the 5 jobs with the same program at a time but when we check the db trace
    zs01 is taking long time as shown below.here zs01 is having small amount of data.
    in the below dbtrace for updating zs01 is taking 2,315,485 seconds .how to reduce this?
    HH:MM:SS.MS Duration     Program   ObjectName  Op.   Curs   Array   Rec     RC     Conn     
    2:36:15 AM     2,315,485     SAPLZS01  ZS01       FETCH  294     1     1     0     R/3     
    The code as shown below
    you can check the code in the program SAPLZS01 include LZS01F01.
    FORM UPDATE_ZS01.
    IF ZS02-STATUS = '3'.
        IF Z_ZS02_STATUS = '3'.            "previous status is ERROR
          EXIT.
        ELSE.
          SELECT SINGLE FOR UPDATE * FROM  ZS01
                 WHERE  PROC_NUM    = ZS02-PROC_NUM.
          CHECK SY-SUBRC = 0.
          ADD ZS02-MF_AMT TO ZS01-ERR_AMT.
          ADD 1           TO ZS01-ERR_INVOI.
          UPDATE ZS01.
        ENDIF.
      ENDIF.
    my question is when updating the ztable why it is taking such long time,
    how to reduce the time or how to make faster to update the ztable .
    Thanks in advance,
    regards
    Suni

    Try the code like this..
    data: wa_zs01 type zs01.
    FORM UPDATE_ZS01.
    IF ZS02-STATUS = '3'.
        IF Z_ZS02_STATUS = '3'.            "previous status is ERROR
          EXIT.
        ELSE.
          SELECT SINGLE FOR UPDATE * FROM  ZS01
                 WHERE  PROC_NUM    = ZS02-PROC_NUM.
    -- change
      CHECK SY-SUBRC = 0.
          ADD ZS02-MF_AMT TO wa_ZS01-ERR_AMT.
          ADD 1           TO wa_ZS01-ERR_INVOI.
          UPDATE ZS01 from wa_zs01.
        ENDIF.
      ENDIF.
    And i think this Select query for ZS01 is inside the ZS02 SELECT statement,
    This might also make slow process.
    If you want to make database access always use Workarea/Internal table to fetch the data
    and work with that.
    Accessing database like this or with Select.... endselect is an inefficient programming.

  • Simple query is taking long time

    Hi Experts,
    The below query is taking long time.
    [code]SELECT   FS.*
      FROM   ORL.FAX_STAGE FS
             INNER JOIN
                   ORL.FAX_SOURCE FSRC
                INNER JOIN
                   GLOBAL_BU_MAPPING GBM
                ON GBM.BU_ID = FSRC.BUID
             ON UPPER (FSRC.FAX_NUMBER) = UPPER (FS.DESTINATION)
    WHERE       FSRC.IS_DELETED = 'N'
             AND GBM.BU_ID IS NOT NULL
             AND UPPER (FS.FAX_STATUS) ='COMPLETED';[/code]
    this query is returning 1645457 records.
    [code]PLAN_TABLE_OUTPUT
    | Id  | Operation           | Name                   | Rows  | Bytes | Cost (%CPU)|
    |   0 | SELECT STATEMENT    |                        |   625K|   341M| 45113   (1)|
    |   1 |  HASH JOIN          |                        |   625K|   341M| 45113   (1)|
    |   2 |   NESTED LOOPS      |                        |   611 | 14664 |    22   (0)|
    |   3 |    TABLE ACCESS FULL| FAX_SOURCE             |  2290 | 48090 |    22   (0)|
    |   4 |    INDEX RANGE SCAN | GLOBAL_BU_MAPPING_BUID |     1 |     3 |     0   (0)|
    |   5 |   TABLE ACCESS FULL | FAX_STAGE              |  2324K|  1214M| 45076   (1)|
    PLAN_TABLE_OUTPUT
    Note
       - 'PLAN_TABLE' is old version
    15 rows selected.[/code]
    The distinct number of records in each table.
    [code]SELECT FAX_STATUS,count(*)
    FROM fax_STAGE
    GROUP BY FAX_STATUS;
    FAX_STATUS    COUNT(*)
    BROKEN          10
    Broken - New    9
    Completed    2324493
    New             20
    SELECT is_deleted,COUNT(*)
    FROM  FAX_SOURCE
    GROUP BY IS_DELETED;
    IS_DELETED COUNT(*)
    N         2290
    Y         78[/code]
    Total number of records in each table.
    [code]SELECT COUNT(*) FROM ORL.FAX_SOURCE FSRC-- 2368
    SELECT COUNT(*) FROM ORL.FAX_STAGE--2324532
    SELECT COUNT(*) FROM APPS_GLOBAL.GLOBAL_BU_MAPPING--9
    [/code]
    To improve the performance of this query I have created the following indexes.
    [code]Functional based index on UPPER (FSRC.FAX_NUMBER) ,UPPER (FS.DESTINATION) and UPPER (FS.FAX_STATUS).
    Bitmap index on FSRC.IS_DELETED.
    Normal Index on GBM.BU_ID and FSRC.BUID.
    [/code]
    But still the performance is bad for this query.
    What can I do apart from this to improve the performance of this query.
    Please help me .
    Thanks in advance.

    <I have created the following indexes.
    CREATE INDEX ORL.IDX_DESTINATION_RAM ON ORL.FAX_STAGE(UPPER("DESTINATION"))
    CREATE INDEX ORL.IDX_FAX_STATUS_RAM ON ORL.FAX_STAGE(LOWER("FAX_STATUS"))
    CREATE INDEX ORL.IDX_UPPER_FAX_STATUS_RAM ON ORL.FAX_STAGE(UPPER("FAX_STATUS"))
    CREATE INDEX ORL.IDX_BUID_RAM ON ORL.FAX_SOURCE(BUID)
    CREATE INDEX ORL.IDX_FAX_NUMBER_RAM ON ORL.FAX_SOURCE(UPPER("FAX_NUMBER"))
    CREATE BITMAP INDEX ORL.IDX_IS_DELETED_RAM ON ORL.FAX_SOURCE(IS_DELETED)
    After creating the following indexes performance got improved.
    But our DBA said that new BITMAP index at FAX_SOURCE table (ORL.IDX_IS_DELETED_RAM) can cause locks
    on multiple rows if IS_DELETED column is in use. Please proceed with detailed tests.
    I am sending the explain plan before creating indexes and after indexes has been created.
    SELECT  FS.*
    FROM  ORL.FAX_STAGE FS
                    INNER JOIN
                    ORL.FAX_SOURCE FSRC
                  INNER JOIN
                      GLOBAL_BU_MAPPING GBM
                    ON GBM.BU_ID = FSRC.BUID
                ON UPPER (FSRC.FAX_NUMBER) = UPPER (FS.DESTINATION)
    WHERE      FSRC.IS_DELETED = 'N'
              AND GBM.BU_ID IS NOT NULL
              AND UPPER (FS.FAX_STATUS) =:B1;
    --OLD without indexes
    PLAN_TABLE_OUTPUT
    Plan hash value: 3076973749
    | Id  | Operation          | Name                  | Rows  | Bytes | Cost (%CPU)| Time    |
    |  0 | SELECT STATEMENT    |                        |  141K|    85M| 45130  (1)| 00:09:02 |
    |*  1 |  HASH JOIN          |                        |  141K|    85M| 45130  (1)| 00:09:02 |
    |  2 |  NESTED LOOPS      |                        |  611 | 18330 |    22  (0)| 00:00:01 |
    |*  3 |    TABLE ACCESS FULL| FAX_SOURCE            |  2290 | 59540 |    22  (0)| 00:00:01 |
    |*  4 |    INDEX RANGE SCAN | GLOBAL_BU_MAPPING_BUID |    1 |    4 |    0  (0)| 00:00:01 |
    |*  5 |  TABLE ACCESS FULL | FAX_STAGE              | 23245 |    13M| 45106  (1)| 00:09:02 |
    PLAN_TABLE_OUTPUT
    Predicate Information (identified by operation id):
      1 - access(UPPER("FSRC"."FAX_NUMBER")=UPPER("FS"."DESTINATION"))
      3 - filter("FSRC"."IS_DELETED"='N')
      4 - access("GBM"."BU_ID"="FSRC"."BUID")
          filter("GBM"."BU_ID" IS NOT NULL)
      5 - filter(UPPER("FS"."FAX_STATUS")=SYS_OP_C2C(:B1))
    21 rows selected.
    --NEW with indexes.
    PLAN_TABLE_OUTPUT
    Plan hash value: 665032407
    | Id  | Operation                        | Name                    | Rows  | Bytes | Cost (%CPU)| Time    |
    |  0 | SELECT STATEMENT                |                          |  5995 |  3986K|  3117  (1)| 00:00:38 |
    |*  1 |  HASH JOIN                      |                          |  5995 |  3986K|  3117  (1)| 00:00:38 |
    |  2 |  NESTED LOOPS                  |                          |  611 | 47658 |    20  (5)| 00:00:01 |
    |*  3 |    VIEW                          | index$_join$_002        |  2290 |  165K|    20  (5)| 00:00:01 |
    |*  4 |    HASH JOIN                    |                          |      |      |            |      |
    |*  5 |      HASH JOIN                  |                          |      |      |            |      |
    PLAN_TABLE_OUTPUT
    |  6 |      BITMAP CONVERSION TO ROWIDS|                          |  2290 |  165K|    1  (0)| 00:00:01 |
    |*  7 |        BITMAP INDEX SINGLE VALUE | IDX_IS_DELETED_RAM      |      |      |            |      |
    |  8 |      INDEX FAST FULL SCAN      | IDX_BUID_RAM            |  2290 |  165K|    8  (0)| 00:00:01 |
    |  9 |      INDEX FAST FULL SCAN        | IDX_FAX_NUMBER_RAM      |  2290 |  165K|    14  (0)| 00:00:01 |
    |* 10 |    INDEX RANGE SCAN              | GLOBAL_BU_MAPPING_BUID  |    1 |    4 |    0  (0)| 00:00:01 |
    |  11 |  TABLE ACCESS BY INDEX ROWID    | FAX_STAGE                | 23245 |    13M|  3096  (1)| 00:00:38 |
    |* 12 |    INDEX RANGE SCAN              | IDX_UPPER_FAX_STATUS_RAM |  9298 |      |  2434  (1)| 00:00:30 |
    Predicate Information (identified by operation id):
    PLAN_TABLE_OUTPUT
      1 - access(UPPER("DESTINATION")="FSRC"."SYS_NC00035$")
      3 - filter("FSRC"."IS_DELETED"='N')
      4 - access(ROWID=ROWID)
      5 - access(ROWID=ROWID)
      7 - access("FSRC"."IS_DELETED"='N')
      10 - access("GBM"."BU_ID"="FSRC"."BUID")
          filter("GBM"."BU_ID" IS NOT NULL)
      12 - access(UPPER("FAX_STATUS")=SYS_OP_C2C(:B1))
    31 rows selected
    Please confirm on the DBA comment.Is this bitmap index locks rows in my case.
    Thanks.>

  • Fwrite() and fread() of a shared FAT32 formatted file is taking long time in MAC osx Lion C program

    Hi
    Is there any provision or api in MAC to open a file in shared mode same as windows
       hUSBdrive = CreateFile(pDriveName, GENERIC_READ | GENERIC_WRITE, FILE_SHARE_READ | FILE_SHARE_WRITE, NULL, OPEN_EXISTING, FILE_FLAG_NO_BUFFERING, NULL);
    we have the follwing scenario where a file is shared among two processes for read/write.one is running on Linux and the other one is running on MAC.where both the processses are reading/writing into the same memory location in the file say "X"
    FAT 32 formatted raw data file which is located on the device, is shared among two processes.
    One process is running on Linux device which is connected to MAC book through usb.In this linux process, the file is opened using fopen() and we have used fcntl() with O_DIRECT flag.This process continuously reads/writes data on memory location "X" in the shared file .
    The other process is running on Mac which has simple c program that opens the file on the connected device i.e from usb drive and reads/writes data using fread()/fwrite().fopen() is used to open the file and FILE_NOCACHE flag is used to avoid caching.
    The value at memory location "X" is updated by mac by using fwrite() and the linux process reads the memory location "X" by using fread(). Linux process is taking around 30 sec to get the updated value.
    If the value is updated by Linux process at memory location "X" by using fwrite() the MAC process is also taking long time more than a minute to read the updated value by usng fread().
    fwrite()/fread() on mac is taking long time where as the windows application which uses the same apis is taking msec time.
    Do we need to use other api s or flags to open file?
    thanks in advance.......

    does any one face this kind of problem?
    fwrite() and fread() takes long time?
    Is there any problem in read/write to a fat32 file on MAC?

  • RSPCM is taking long time

    Hi,
    RSPCM is taking long time to open at the same time RSPC is working fine
    when I try to change the process chain status through RSPC_PROCESS_FINISH . it is executing for long time no response. I tried executing it backgrond also from past 2 days it is running, no progress.
    Our basis team created index to all backend table of RSPCM, still issue persist. Please suggest me some to get rid of this.
    Br,
    Harish

    hi,
    Please check the below thread
    RSPCM T-Code was executing very slow
    Please check note 1372931
    hope it helps!
    Edited by: Lavanya J on Nov 4, 2011 1:32 PM

Maybe you are looking for

  • Cannot place order ERROR: "There was a problem adding your product to cart"

    Hey guys, Well, once again Best Buy's online system is preventing me from making an order. It seems like nearly every time I log in I experience some new problem. This time, I'm unable to add anything to my cart. It doesn't matter what item I try to

  • Issue in PO Status

    Hi There is an issue with a Service PO.When I created a service p.o the p.o value was 1,543,695.60 USD . I created service entry sheet and GR for 1,029,130.40 USD. The still to be delivered amount should be 1,543,695.60-1,029,130.40 =514565.20 USD Bu

  • Vista and BEFW11S4 issues

    Hi Everyone. I have a Wireless B Router model BEFW11S4 v 4.0. We recently bought a new computer that has Windows Vista. Recently we started having some problems with the internet. Our Windows Firewall is OFF as we have Norton 360 v 2.0. Our problem i

  • No connection to server via LAN

    When I say _nc:NetConnection = new NetConnection(); _nc.connect( "rtmp://127.0.0.1/ImmuneDefense" ); It works Now I want to let another computer in the local network connect to the server. _nc.connect( "rtmp://192.168.0.9/ImmuneDefense" ); It doesnt

  • Webpart causing error while accessing the subsite.

    Hi I am getting an error message saying that webpart causes error in accessing the site in MOSS 2007. PFA of the error: Please someone help with the error webpart: how to find which custom webpart is affecting the site access error. Thanks, Badri