Video files taking longer time when uploading using java sdk

Hi,
I am using java APIs for uploading image/video files to azure blob storage.And i see that video files take
much time to get uploaded.On average a 25 MB file takes more than 2 minutes...Is it expected with Java API?Is there any way with azure java APIs to reduce the upload time?
Thanks!

Hi,
>>On average a 25 MB file takes more than 2 minutes
I think the internet connection is a factor, please test the upload speed at:
http://www.azurespeed.com/Azure/Upload
If you want to reduce the upload time, you could try to use the chunk size and concurrent threads.
Best Regards,
Jambor 
We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
Click
HERE to participate the survey.

Similar Messages

  • GC taking long time when using Concurrent Mark Sweep GC with Sun JDK 150_12

    We are having problem of Garbage collection taking too long. We are using Weblogic 9.2 and Sun JDK 150_12
    Below are the memory arguments we are using. We are using Concurrent Mark Sweep GC. What we are observing is Young Generation is getting filled up and consequent tenured generation also hangs with long pauses.
    Below are the JVM arguments we are using
    -Xms2560M -Xmx2560M -Xloggc:${LOGDIR}/gc.log -XX:+PrintGCTimeStamps -XX:+PrintGCDetails -XX:+UseParNewGC -XX:+UseConcMarkSweepGC -XX:+CMSParallelRemarkEnabled -XX:SurvivorRatio=128 -XX:MaxTenuringThreshold=0 -XX:CMSInitiatingOccupancyFraction=60 -XX:NewSize=512m -XX:MaxNewSize=512m -XX:MaxPermSize=256m
    I have seen many forums where there are many reported issues with Concurrent Mark Sweep garbage collection with Sun JDK, but with different recommendations. But did not find any defnite recommendation. Please advice.
    - - Tarun

    We are having problem of Garbage collection taking too long. We are using Weblogic 9.2 and Sun JDK 150_12
    Below are the memory arguments we are using. We are using Concurrent Mark Sweep GC. What we are observing is Young Generation is getting filled up and consequent tenured generation also hangs with long pauses.
    Below are the JVM arguments we are using
    -Xms2560M -Xmx2560M -Xloggc:${LOGDIR}/gc.log -XX:+PrintGCTimeStamps -XX:+PrintGCDetails -XX:+UseParNewGC -XX:+UseConcMarkSweepGC -XX:+CMSParallelRemarkEnabled -XX:SurvivorRatio=128 -XX:MaxTenuringThreshold=0 -XX:CMSInitiatingOccupancyFraction=60 -XX:NewSize=512m -XX:MaxNewSize=512m -XX:MaxPermSize=256m
    I have seen many forums where there are many reported issues with Concurrent Mark Sweep garbage collection with Sun JDK, but with different recommendations. But did not find any defnite recommendation. Please advice.
    - - Tarun

  • Rag: Query performence - Report taking long time when drill down a Char.

    Hi All,
    After executing the report when we try to Drilldown PO Char, then the report used to take long time nearly 60 to 90 minutes and thorough an errors that  "An error is occurred during loading" or "Connection Timed Out".
    The report is created on Multiprovider which is build on 7 infocubes and the data for the PO char is coming from all the 7 cubes.
    So we have created Aggregates on the 5 Infocubes based on system proposal.
    Then,
    We checked in RSRT, the data is coming from all 5 Aggregates and also when we drill down the PO also the data is coming from 5 Aggregates itself.
    But,
    The report is running fine and also when we drilldown the PO to show all the PO values, then the report showing all the POu2019s details by taking nearly 15 mins only.
    Here our concern is why the report is taking 15 mins when we drill a PO char.
    Could any one give suggestions please.
    Thanks in Adv.
    Raju. E

    Hi,
    please check this list for query performance:
    http://sapbwneelam.blogspot.com/2007/10/checklist-for-query-performance.html
    Regards
    Erwin

  • Full GC taking longer time

    hi Team,
    We are running SLD on separate hardware with Windows 2003. But since couple of months we been observing that 'full GC taking longer time' then it used to be. So wondering what are the possible parameters, configuration items that i need to look back to find tune this issue.
    any good reference docs related to find tuning the GC is much appriciated.
    Thanks
    Sekhar

    Hi ,
    Hope this might help.
    http://www.ibm.com/developerworks/library/i-gctroub/
    http://my.safaribooksonline.com/0596003773/javapt2-CHP-3-SECT-4#X2ludGVybmFsX1NlY3Rpb25Db250ZW50P3htbGlkPTA1OTYwMDM3NzMvamF2YXB0Mi1DSFAtMy1TRUNULTE=
    http://java.sun.com/docs/hotspot/gc5.0/gc_tuning_5.html
    Also check out the pdfs
    https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/7fdca26e-0601-0010-369d-b3fc87d3a2d9
    https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/22baa590-0201-0010-26a3-f1cfa2469973
    Rgds
    joel

  • Fwrite() and fread() of a shared FAT32 formatted file is taking long time in MAC osx Lion C program

    Hi
    Is there any provision or api in MAC to open a file in shared mode same as windows
       hUSBdrive = CreateFile(pDriveName, GENERIC_READ | GENERIC_WRITE, FILE_SHARE_READ | FILE_SHARE_WRITE, NULL, OPEN_EXISTING, FILE_FLAG_NO_BUFFERING, NULL);
    we have the follwing scenario where a file is shared among two processes for read/write.one is running on Linux and the other one is running on MAC.where both the processses are reading/writing into the same memory location in the file say "X"
    FAT 32 formatted raw data file which is located on the device, is shared among two processes.
    One process is running on Linux device which is connected to MAC book through usb.In this linux process, the file is opened using fopen() and we have used fcntl() with O_DIRECT flag.This process continuously reads/writes data on memory location "X" in the shared file .
    The other process is running on Mac which has simple c program that opens the file on the connected device i.e from usb drive and reads/writes data using fread()/fwrite().fopen() is used to open the file and FILE_NOCACHE flag is used to avoid caching.
    The value at memory location "X" is updated by mac by using fwrite() and the linux process reads the memory location "X" by using fread(). Linux process is taking around 30 sec to get the updated value.
    If the value is updated by Linux process at memory location "X" by using fwrite() the MAC process is also taking long time more than a minute to read the updated value by usng fread().
    fwrite()/fread() on mac is taking long time where as the windows application which uses the same apis is taking msec time.
    Do we need to use other api s or flags to open file?
    thanks in advance.......

    does any one face this kind of problem?
    fwrite() and fread() takes long time?
    Is there any problem in read/write to a fat32 file on MAC?

  • Takes a long time to upload a publish to folder site

    i have charter.net internet service...3mb high speed internet service...
    i recently built a site that is approximately 19mb...when i use either fetch or captain ftp, it takes a long time to upload (hours sometimes)...most of the time is gets hung up on the .png files...it is as it stops...i have contacted charter about the slowness...there response is we don't support mac...you don't want to know what i told them...
    anyway...is there anything i may be missing or don't understand about uploading...the first few times several weeks ago, i could upload a site about half the size in about 10 minutes...charter will not admit anything on their own end...
    any ideas would be nice to know...
    michael

    Well, the 3mb speed that you quote are the theoretical maximum DOWNLOAD speeds. Reality might be a little bit different depending on a host of factors. Anyways, UPLOAD speeds are always capped at some fraction of that just like with asynchronous DSL. But your subjective experiences are still different. As long as it is not reporting out errors and your site is actually being uploaded you should be fine.
    As to why it is taking so long....I can't say. At first I was concerned that maybe Charter had a thing about filenames with spaces or a certain length or something. But you tell me, is your site being uploaded?

  • New Photos App take a very long time to upload 20 Pictures, while the Activity Monitor is showing much more data sent (MB) than these two pictues

    The New photos app is taking a very long time to upload images. It stays a very long time at 20,104 for example. When I check the Activity Monitor (Network) it shows that the computer is actually uploading stuff. After about 2 hours, only 20 Pics&Videos are uploaded. The uploaded pics and videos are about 500MB while the data sent is around 1GB. I am positive that Photos is the only thing uploading data at the moment.
    Anyone else facing this problem?

    It can take days, even weeks.
    Tell me 2 things and I will tell you the estimated upload time.
    1 How big is your Library (in GB)
    2 How fast is your upload speed.
    Or you can just use the upload time calculator and do it yourself

  • Interactive Report general Search is taking long time

    Hi All,
    I have two IR reports in my Application, Both are using same tables , same where clause, displaying column is different.
    Both query returning 60000 records, In my first report , after my report gets loaded , i am using the general search button for getting some data,
    when i try to find 100 in my reports , my first report displaying the search result in 5 seconds, But my second report i am doing the same steps , but its taking long time to respond.
    Could some one tell me what is the cause for this problem also in what basis this search will work means how this will work?
    Cheers,
    San

    Hi Mike,
    You are right test is the View
    After i enabled the debug i found the below query
    Report 1 Query
    SELECT "Checkbox",
      "RS_TRX_NUMBER",
      "EXT_TRX_NUMBER",
      "ARRANGEMENT_NUMBER",
      "TRX_DATE",
      "TRX_STATUS",
      "SOLD_TO_CUSTOMER_NAME",
      "TRX_TYPE",
      "CURRENCY",
      "COMMITTED_AMOUNT",
      "BUDGET_AMOUNT",
      "TRANSACTION_BUDGET",
      "TRX_DESCRIPTION",
      "FILES",
      "Lines",
      "Events",
      "TRX_Revenue",
      "ARGMT_Revenue",
      "Holds",
      "RS_TRX_ID",
      COUNT(DISTINCT "ARRANGEMENT_NUMBER") over (),
      COUNT(*) over () AS apxws_row_cnt
    FROM
      (SELECT *
      FROM
        (SELECT a.* ,
          'Events' "Events" ,
          'Billing' "Billing" ,
          'Transaction Revenue' "TRX_Revenue" ,
          'Arrangement Revenue' "ARGMT_Revenue" ,
          'Lines' "Lines" ,
          'Invoice' "Invoice" ,
          'Holds' "Holds" ,
          apex_item.hidden ( 2, a.rs_trx_number)
          || apex_item.checkbox ( 1, a.rs_trx_id, 'onchange="spCheckChange(this, ''x01'', ''P13_TRX_AJAX_NAME'', ''P13_TRX_COLLECTION_NAME'',''SELECTED_TRX'');"', c.n001, ':' ) "Checkbox"
        FROM test a,
          apex_collections c
        WHERE a.rs_trx_id        = c.n001(+)
        AND c.collection_name(+) = :P13_TRX_COLLECTION_NAME
        AND a.org_id            IN
          (SELECT DISTINCT org_id
          FROM test1
          WHERE (role_name = :APPUSER_CURRENT_ROLE)
        ) r
      ) r
    where rownum <= TO_NUMBER(:APXWS_MAX_ROW_CNT)
    Second Report
    SELECT "Checkbox",
      "EXT_TRX_NUMBER",
      "RS_TRX_NUMBER",
      "ARRANGEMENT_NUMBER",
      "TRX_RM_STATUS",
      "SOLD_TO_CUSTOMER_NAME",
      "TRX_TYPE",
      "TRX_DESCRIPTION",
      "CURRENCY",
      "COMMITTED_AMOUNT",
      "UNEARNED",
      "EARNED",
      "REV_TRANSFERRED",
      "REV_LINES",
      "SCHEDULES",
      "EVENTS",
      "EXT_TRX_ID",
      "PRIMARY_SALESREP_NUMBER",
      "RS_TRX_ID",
      COUNT(*) over () AS apxws_row_cnt
    FROM
      (SELECT *
      FROM
        (SELECT a.* ,
          'Rev Lines' REV_LINES ,
          'Schedules' Schedules ,
          'Events' Events ,
          apex_item.hidden ( 2, a.rs_trx_number)
          || apex_item.checkbox ( 1, a.rs_trx_id, 'onchange="spCheckChange(this, ''x01'', ''P5_REV_TRX_AJAX_NAME'', ''P5_RTCN'',''SELECTED_REV_TRX'');"', c.n001, ':' ) "Checkbox"
        FROM Test a,
          apex_collections c
        WHERE a.rs_trx_id        = c.n001(+)
        AND C.COLLECTION_NAME(+) = :P5_RTCN
        AND a.org_id            IN
          (SELECT DISTINCT org_id
          FROM Test1
          WHERE (role_name = :APPUSER_CURRENT_ROLE)
        ) r
      ) r
    WHERE rownum <= to_number(:APXWS_MAX_ROW_CNT)
    ORDER BY "ARRANGEMENT_NUMBER" DESC,
      "EXT_TRX_NUMBER"
    In Source i am using the same query in both report , But how come order by clause has been adding in my Second query.
    Also in Report 1 :
      COUNT(DISTINCT "ARRANGEMENT_NUMBER") over (),
      COUNT(*) over () AS apxws_row_cnt
    The same is not there in Second report query .
    How can i remove that order by clause from my id also how can i resolve the above change (
      COUNT(DISTINCT "ARRANGEMENT_NUMBER") over (),
      COUNT(*) over () AS apxws_row_cnt)
    Help me out to resolve.
    Cheers,
    San

  • Java Applet taking long time to load

    Hi,
    I have a swing based Menu generator Applet. On clicking a Menubar item, I have called a new class [ MyEntry1 my1 = new MyEntry() ], which called one or more utility classes. My main applet (which I have used in <Aplpet> tag in .html file) is very small. But from that, I have called severel classes which are not very small.
    My Problem is, when i am trying to run my applet through IE 5, it is taking long time (180-250 secs.)
    Can anyone give me some tips, how to minimise the loading time? I am using Applet/Servlet/JDBC to develop a web application, which will be used remotely. Using Tomcat server. Any code snippets will be very helpfull for me.
    Thanking all gurus in advance.

    Hope someone can tell me why my Java takes so long to
    load when I go into a online game..I don't think anyone in here can. You could call your local computer repair-shop to have them take a look at your PC.

  • Export (exp) taking long time and reading UNDO

    Hi Guys,
    Oracle 9.2.0.7 on AIX 5.3
    A schema level export job is scheduled at night. Since day before yesterday it has been taking really long time. It used to finish in 8 hours or so but yesterday it took around 20 hours and was still running. The schema size to be exported is around 1 TB. (I know it is bit stupid to take such daily exports but customer requirement, you know ;) ) Today again it is still running although i scheduled it to run even earlier by 1 and 1/2 hour.
    The command used is:
    exp userid=abc/abc file=expabc.pipe buffer=100000 rows=y direct=y
    recordlength=65535 indexes=n triggers=n grants=y
    constraints=y statistics=none log=expabc.log owner=abcI have monitored the session and all the time the wait event is db file sequential read. From p1 i figured out that all the datafiles it reads belong to UNDO tablespace. What surprises me is that when consistent=Y is not specified should it go to read UNDO so frequently ?
    There is total of around 1800 tables in the schema; what i can see from the export log is that it exported around 60 tables and has been stuck since then. The logfile, dumpfile both has not been updated since long time.
    Any hints, clues in which direction to diagnose please.
    Any other information required, please let me know.
    Regards,
    Amardeep Sidhu

    Thanks Hemant.
    As i wrote above, it runs from a cron job.
    Here is the output from a simple SQL querying v$session_wait & v$datafile:
    13:50:00 SQL> l
      1* select a.sid,a.p1,a.p2,a.p3,b.file#,b.name
      from v$session_wait a,v$datafile b where a.p1=b.file# and a.sid=154
    13:50:01 SQL> /
           SID         P1         P2         P3      FILE# NAME
           154        509     158244          1        509 /<some_path_here>/undotbs_45.dbf
    13:50:03 SQL> /
           SID         P1         P2         P3      FILE# NAME
           154        509     157566          1        509 /<some_path_here>/undotbs_45.dbf
    13:50:07 SQL> /
           SID         P1         P2         P3      FILE# NAME
           154        509     157016          1        509 /<some_path_here>/undotbs_45.dbf
    13:50:11 SQL> /
           SID         P1         P2         P3      FILE# NAME
           154        509     156269          1        509 /<some_path_here>/undotbs_45.dbf
    13:50:16 SQL> /
           SID         P1         P2         P3      FILE# NAME
           154        508     167362          1        508 /<some_path_here>/undotbs_44.dbf
    13:50:58 SQL> /
           SID         P1         P2         P3      FILE# NAME
           154        508     166816          1        508 /<some_path_here>/undotbs_44.dbf
    13:51:02 SQL> /
           SID         P1         P2         P3      FILE# NAME
           154        508     165024          1        508 /<some_path_here>/undotbs_44.dbf
    13:51:14 SQL> /
           SID         P1         P2         P3      FILE# NAME
           154        507     159019          1        507 /<some_path_here>/undotbs_43.dbf
    13:52:09 SQL> /
           SID         P1         P2         P3      FILE# NAME
           154        506     193598          1        506 /<some_path_here>/undotbs_42.dbf
    13:52:12 SQL> /
           SID         P1         P2         P3      FILE# NAME
           154        506     193178          1        506 /<some_path_here>/undotbs_42.dbf
    13:52:14 SQL>Regards,
    Amardeep Sidhu
    Edited by: Amardeep Sidhu on Jun 9, 2010 2:26 PM
    Replaced few paths with <some_path_here> ;)

  • Impdp taking long time for only few MBs data...

    Hi All,
    I have one query related to impdp. I have one expdp file and size is 47M. When I restore this dmp using impdp it will take more time. Also initially table_data loaded finsih very fast but then later on alter function/procedure/view taking a lot time almost 4 to 5 hrs.
    I have no idea why its taking long time... Earlier I can see one DB link is failed and it given error for "TNS name could not resolved" so I create DB link also before run impdp but the same result. Can any one suggest what could be the cause to taking long time for only 47MB data?
    Note - Both expdp and impdp database version is 11.2.0.3.0. If I import the same dmp file into 11.2.0.1.0 then its done in few mins.
    Thanks...

    Also Read
    Checklist For Slow Performance Of DataPump Export (expdp) And Import (impdp) [ID 453895.1]
    DataPump Import (IMPDP) is Very Slow at Object/System/Role Grants, Default Roles [ID 1267951.1]

  • SSRS Reports taking long time to load

    Hello,
    Problem : SSRS Reports taking long time to load
    My System environment : Visual Studio 2008 SP1  and SQL Server 2008 R2
    Production Environment : Visual Studio 2008 SP1  and SQL Server 2008 R2
    I have created a Parameterized report (6 parameters), it will fetch data from 1 table. table has 1 year and 6 months data,      I am selecting parameters for only 1 month (about 2500 records). It is taking almost 2 minutes and 30 seconds
    to load the report.
    This report running efficiently in my system (report load takes only 5 to 6 seconds) but in
    production it is taking 2 minutes 30 seconds.
    I have checked the Execution log from production so I found the timing for
    Data retrieval (approx~)       Processing (approx~)               Rendering (approx~)
    10 second                                      15 sec                        
                2 mins and 5 sec.
    But Confusing point is that , if I run the same report at different time overall output time is same (approx) 2 min 30 sec but
    Data retrieval (approx~)       Processing (approx~)                Rendering (approx~)
    more than 1 min                            15 sec                                     
    more than 1 min
    so 1 question why timings are different ?
    My doubts are
    1) If query(procedure to retrieve the data) is the problem then it should take more time always,
    2) If Report structure is problem then rendering will also take same time (long time)
    for this (2nd point) I checked on blog that Rendering depends on environment structure e.g. Network bandwidth, RAM, CPU Usage , Number of users accessing same report at a time.
    So I did testing of report when no other user working on any report But failed (same result  output is 2 min 30 sec)
    From network team I got the result is that there is no issue or overload in CPU usage or RAM also No issue in Network bandwidth.
    Production Database Server and Report server are different (but in same network).
    I checked that database server the SQL Server is using almost Full RAM (23 GB out of 24 GB)
    I tried to allocate the memory to less amount up to 2GB (Trial solution I got from Blogs) but this on also failed.
    one hint I got from colleague that , change the allocated memory setting from static memory to dynamic to SQL Server
    (I guess above point is the same) I could not find that option Static and Dynamic memory setting.
    I did below steps
    Connected to SQL Server Instance
    Right click on Instance go to properties, Go to Memory Tab
    I found three options 1) Server Memory   2) Other memory   3) Section for "Configured values and Running values"
    Then I tried to reduce Maximum  Server memory up to 2 GB (As mentioned above)
    All trials failed, this issue I could not find the roots for this issue.
    Can anyone please help (it's bit urgent).

    Hi UdayKGR,
    According to your description, your report takes too long to load on your production environment. Right?
    In this scenario, since the report runs quickly in developing environment, we initially think it supposed to be the issue on data retrieval. However, based on the information in execution log, it takes longest time on rendering part. So we suggest you optimize
    the report itself to reduce the time for rendering. Please refer to the link below:
    My report takes too long to render
    Here is another article about overall performance optimization for Reporting Services:
    Reporting Services Performance and Optimization
    If you have any question, please feel free to ask.
    Best Regards,
    Simon Hou

  • Update ztable is taking long time

    Hi All,
    i have run the 5 jobs with the same program at a time but when we check the db trace
    zs01 is taking long time as shown below.here zs01 is having small amount of data.
    in the below dbtrace for updating zs01 is taking 2,315,485 seconds .how to reduce this?
    HH:MM:SS.MS Duration     Program   ObjectName  Op.   Curs   Array   Rec     RC     Conn     
    2:36:15 AM     2,315,485     SAPLZS01  ZS01       FETCH  294     1     1     0     R/3     
    The code as shown below
    you can check the code in the program SAPLZS01 include LZS01F01.
    FORM UPDATE_ZS01.
    IF ZS02-STATUS = '3'.
        IF Z_ZS02_STATUS = '3'.            "previous status is ERROR
          EXIT.
        ELSE.
          SELECT SINGLE FOR UPDATE * FROM  ZS01
                 WHERE  PROC_NUM    = ZS02-PROC_NUM.
          CHECK SY-SUBRC = 0.
          ADD ZS02-MF_AMT TO ZS01-ERR_AMT.
          ADD 1           TO ZS01-ERR_INVOI.
          UPDATE ZS01.
        ENDIF.
      ENDIF.
    my question is when updating the ztable why it is taking such long time,
    how to reduce the time or how to make faster to update the ztable .
    Thanks in advance,
    regards
    Suni

    Try the code like this..
    data: wa_zs01 type zs01.
    FORM UPDATE_ZS01.
    IF ZS02-STATUS = '3'.
        IF Z_ZS02_STATUS = '3'.            "previous status is ERROR
          EXIT.
        ELSE.
          SELECT SINGLE FOR UPDATE * FROM  ZS01
                 WHERE  PROC_NUM    = ZS02-PROC_NUM.
    -- change
      CHECK SY-SUBRC = 0.
          ADD ZS02-MF_AMT TO wa_ZS01-ERR_AMT.
          ADD 1           TO wa_ZS01-ERR_INVOI.
          UPDATE ZS01 from wa_zs01.
        ENDIF.
      ENDIF.
    And i think this Select query for ZS01 is inside the ZS02 SELECT statement,
    This might also make slow process.
    If you want to make database access always use Workarea/Internal table to fetch the data
    and work with that.
    Accessing database like this or with Select.... endselect is an inefficient programming.

  • Retry recovery action is taking long time to complete

    Hi,
    I am in the process of testing error handling for my BPEL process.
    As part of it, I have given the default recover action in the fault policy of the process as retry with a retry count of 3 times with retry interval as 2.
    To test this, i have changed the target end point uri of the target service to some invalid url and ran the process.
    First I was not able to find the BPEL process instance at all in the console.
    Later I was able to find the instance after half an hour. When I saw the instance, if retry had happened or not.
    I found that the retry happened for 3 times, but it took 30 mins for the first retry to complete and start the second retry.
    Can anyone please let me know why the retry is taking long time although the retry interval I gave is only 2 seconds.
    THanks in advance!!!!

    Hi,
    Please find below the fault policy file:
    <?xml version="1.0" encoding="UTF-8"?>
    <faultPolicy version="2.0.1" id="TestFaultPolicy" xmlns:env="http://schemas.xmlsoap.org/soap/envelope/" xmlns:xs="http://www.w3.org/2001/XMLSchema" xmlns="http://schemas.oracle.com/bpel/faultpolicy" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
         <Conditions>
         <faultName>
         <condition>
         <action ref="aia-ora-retry"/>
         </condition>
         </faultName>
         </Conditions>
         <Actions>
              <Action id="aia-ora-retry">
                   <retry>
                        <retryCount>3</retryCount>
                        <retryInterval>2</retryInterval>
                        <exponentialBackoff/>
                        <retryFailureAction ref="aia-ora-java"/>
                        <retrySuccessAction ref="aia-ora-java"/>
                   </retry>
              </Action>
              <!-- This is an action will cause a replay scope fault-->
              <Action id="ora-replay-scope">
                   <replayScope/>
              </Action>
              <!-- This is an action will bubble up the fault-->
              <Action id="ora-rethrow-fault">
                   <rethrowFault/>
              </Action>
              <!-- This is an action will mark the work item to be "pending recovery from console"-->
              <Action id="ora-human-intervention">
                   <humanIntervention/>
              </Action>
              <!-- This action will cause the instance to terminate-->
              <Action id="ora-terminate">
                   <abort/>
              </Action>
              <Action id="aia-ora-java">
              <javaAction className="oracle.apps.aia.core.eh.BPELJavaAction" defaultAction="ora-rethrow-fault">
              <returnValue value="REPLAY" ref="ora-terminate"/>
              <returnValue value="RETRHOW" ref="ora-rethrow-fault"/>
              <returnValue value="ABORT" ref="ora-terminate"/>
              <returnValue value="RETRY" ref="aia-ora-retry"/>
              <returnValue value="MANUAL" ref="ora-human-intervention"/>
              </javaAction>
    </Action>     
         </Actions>
    </faultPolicy>

  • Master Dataload is taking long time to complete

    Hi
    In last couple of weeks, we are experience the delay for master dataload job. These dataload should take 5-8 minute to complete but it is taking more then 5 hour to complete. here we are using process chain for master dataload.
    Any idea? We don't understand where to start to check the root cause of this problem.
    Regards
    Amar

    Hi Freinds
    Some update of our master dataload. Still not resolved.
    Seem to me, our problem is related to Hierarchy in BW server.It is taking long time to update the hierarchy in BW server. As i told you, problem occur only when loading master dataload. For transaction dataload, no IDOC get stuck in R/3 (NPR) server. There are enough resources in both R/3 and BW server for trfc.
    There are total 9 hierarchy in the master dataload that is delaying in the process for update. If you have seen this type of problem for the following hierarchy, pls update me.
    A     0ACCOUNT (InfoSource)
         Hierarchy
    1     Fin statement vers FSV5
    2     Cost Element by fun area
    B     0COSTCENTER (InfoSource)
         Hierarchy
    3     Mitel Network
    4     NA MNS Total
    5     CEO
    6     Global Support Cost center
    C     0COORDER (infosource)
         Hierarchy
    7     Total R&D (Ron Wellard)
    8     RD YRD1
    9     RD YRD2
    Regards
    Amarjit

Maybe you are looking for

  • How can I open an existing document

    I would like to create a button that would open a ".doc" file. I'm using JSP. I know this is a JAVA forum but i'm sure that somebody could help me on that thx

  • How do I change the font color in Live Type?

    How can I change the font color in Live Type?

  • Sort Array of References by Label-Name

    I am developing on an DSC application with shared variables and alarming and events with a hugh amount of variables. Each variable is displayed on the front panel within tab controls. To reduce time and effort when expanding the variables I would lik

  • Number of TB displays you can add to MBP with TB

    Simple enough I guess. Is it possible to daisy chain 3 monitors with TB to a MBP with thunderbolt display port. 

  • MBP Clamshell Mode via DVI LG Monitor

    Hi I know this one has been done to death but....... I have a MBP 2.2SR and yesterday I got a LG 22" monitor. Originally I tried to connect via DVI, usual process Ensure MBP is connected to the mains Connect Display Ensure Keyboard is working Close L