Upload large file on doSubmit in javascript

Hi all,
I'm currently having some problems uploading large files using a custom Upload button (button calls doSubmit from javascript).
When using an Upload button with unconditional branch to the same page (no js) it works perfectly. Now I have to use my own Upload button that calls some javascript and AJAX and then Submit the page (with doSubmit); in this case I can only upload very small files (< 100k) ...
I understand the mechanism of the file browse and its uploading to the table WWV_FLOW_FILE_OBJECTS, but I cannot figure out why my configuration does not work when it does stricly the same but from js ...
Someone experienced the same issue ? How to submit the upload in javascript side with doSubmit ??
Best regards,
Othman

Hi Othman,
did you find a solution for this. I'm facing nearly the same problem, I'm calling a popup, where the user can upload up to 5 files. Then on clicking a button, javascript is called, that's calling the doSubmit and some others things and at the end the popup is closed.
Sometimes the upload of 5 files is working, sometimes only 1 file is possible (obvisiously it's depending on the filesize).
function addFile()
doSubmit('FILE');
opener.actualizeProject('TAB4');
close();
}chrissy

Similar Messages

  • OOM happens inside Weblogic 10.3.6 when application uploads large files

    Oracle Fusion BI Apps application is uploading large files (100+ MB) onto Oracle Cloud Storage. This application works properly when ran outside weblogic server. When deployed on Fusion Middleware Weblogic 10.3.6, during upload of large files we get this OOM error
    java.lang.OutOfMemoryError: allocLargeObjectOrArray: [B, size 268435472
        at jrockit/vm/Allocator.allocLargeObjectOrArray(JIZ)Ljava/lang/Object;(Native Method)
        at jrockit/vm/Allocator.allocObjectOrArray(Allocator.java:349)[optimized]
        at weblogic/utils/io/UnsyncByteArrayOutputStream.resizeBuffer(UnsyncByteArrayOutputStream.java:59)[inlined]
        at weblogic/utils/io/UnsyncByteArrayOutputStream.write(UnsyncByteArrayOutputStream.java:89)[optimized]
        at com/sun/jersey/api/client/CommittingOutputStream.write(CommittingOutputStream.java:90)
        at com/sun/jersey/core/util/ReaderWriter.writeTo(ReaderWriter.java:115)
        at com/sun/jersey/core/provider/AbstractMessageReaderWriterProvider.writeTo(AbstractMessageReaderWriterProvider.java:76)
        at com/sun/jersey/core/impl/provider/entity/InputStreamProvider.writeTo(InputStreamProvider.java:98)
        at com/sun/jersey/core/impl/provider/entity/InputStreamProvider.writeTo(InputStreamProvider.java:59)
        at com/sun/jersey/api/client/RequestWriter.writeRequestEntity(RequestWriter.java:300)
        at com/sun/jersey/client/urlconnection/URLConnectionClientHandler._invoke(URLConnectionClientHandler.java:213)
        at com/sun/jersey/client/urlconnection/URLConnectionClientHandler.handle(URLConnectionClientHandler.java:149)
    Looks like weblogic is using its default Weblogic HTTP handler, switching to Sun HTTP handler via start up JVM/Java Option "-DUseSunHttpHandler=true" solves the OOM issue.
    Seems instead of streaming the file content with some fixed size byte array its being on the whole file into memory during upload.
    Is it possible to solve this OOM by changing any setting of Weblogic HTPP handler without switching to Sun HTTP handler as there are many other application deployed on this weblogic instance?
    We are concerned whether there will be any impact on performance or any other issue.
    Please advice, highly appreciate your response.
    Thanks!

    Hi,
    If you have a back then restore the below file back and then try to start weblogic:
    \Oracle\Middleware\user_projects\domains\<domain_name>\config\config.lok
    Thanks,
    Sharmela

  • Upload large file in sharepoint

    hi
    is it possible upload large file in sharepoint in our scenario we need upload large file >3gb in document libraries
    where sharepoint store files? is there any way we can have a custom aspx page for uploading large files in chunks in document library?
    kindly suggest.
    Thankyou

    is not possbile upload more than 2Gb
    If you try to upload big files you will need to make some change on your Sharepoint to support the 2Gb.
    Error message when you try to upload a large file to a document library on a Windows SharePoint Services 3.0 site: "Request timed out"
    http://support.microsoft.com/?id=925083
    With files bigger than 2GB can recomend creeate linked files to that files where you can download.
    André Lage Microsoft SharePoint and CRM Consultant
    Blog:http://aaclage.blogspot.com
    Codeplex:http://spupload.codeplex.com/http://simplecamlsearch.codeplex.com/

  • How to upload large file with http via post

    Hi guys,
    Does anybody know how to upload large file (>100 MB) use applet to servlet with http via post method? Thanks in advance.
    Regards,
    Mark.

    Hi SuckRatE
    Thanks for your reply. Could you give me some client side code to upload a large file. I use URL to connect to server. It throws out of memory exception. The part of client code is below:
    // connect to the servlet
    URL theServlet = new URL(servletLocation);
    URLConnection servletConnection = theServlet.openConnection();
    // inform the connection that we will send output and accept input
    servletConnection.setDoInput(true);
    servletConnection.setDoOutput(true);
    // Don't used a cached version of URL connection.
    servletConnection.setUseCaches (false);
    servletConnection.setDefaultUseCaches(false);
    // Specify the content type that we will send text data
    servletConnection.setRequestProperty("Content-Type",
    +"application/octet-stream");
    // send the user string to the servlet.
    OutputStream outStream = servletConnection.getOutputStream();
    FileInputStream filein = new FileInputStream(largeFile);
    //BufferedReader in = new BufferedReader(new InputStreamReader
    +(servletConnection.getInputStream()));
    //System.out.println("tempCurrent = "+in.readLine());
    byte abyte[] = new byte[2048];
    int cnt = 0;
    while((cnt = filein.read(abyte)) > 0)
    outStream.write(abyte, 0, cnt);
    filein.close();
    outStream.flush();
    outStream.close();
    Regards,
    Mark.

  • Uploading larger files in iPlanet 4.1.11

    I am using iPlanet 4.1.11 as my web server and WebSphere 4.0.4 as my application server.
    I am trying to upload files of size greater than 10MB using multipart in my application. But I can upload files of size less than 10MB. If it is above 10 MB, HTTP connection is closed without any response.
    In WebLogic we have maximum post size parameter which can be set to enable uploading larger file. Is there any similar configuration settings in iPlanet or WebSphere? Or is there any relation between file cache and upload size? Is there any hardware limitations? Please help.

    I'm not aware of any config settings on iPlanet Webserver side that restrict upload file size. Probably, its something in the application or app server that is causing this..
    Note: 4.1 version has reached end of support life, you should consider upgrading to a currently supported web server version.
    Thanks
    Manish

  • UploadedFileProcessor Uploading Large File Optimization

    According to the documentation, one should create their own implementation of UploadedFileProcessor if they need to support uploading large files or special uploading processing.
    I am wondering if anyone has done this or knows of an article. I have a requirement for large uploads up to 4 GB though most of the time they'll be 100 - 400 MB.
    Thanks,
    Kris

    No

  • "Not connected to internet" error only when uploading large files.

    Hi,
    I have a strange problem that occurs when im uploading large files to any of my servers....
    The files upload just fine but while im uploading im unable to use the internet for anything else. For example if I try to open a web page I get an error saying im not connected to the internet. This error will persist until the file finishes uploading.
    Its a strange and annoying problem and was wondering if anybody could throw any light on the problem.
    I cant be sure but it may have coincided with me connecting a new Netgear wireless modem router.
    Many thanks
    Matt

    Hi Sorry, only just noticed your reply. Didn't get an email alert like I usually do.
    Thanks for that. To be honest I have no idea what QOS is but I will take a look at the routers settings to see if it makes sense to me.
    Thanks
    Matt

  • Is it possible to upload large files through FTP to server with iWeb?

    Is it possible to upload large files through FTP to server with iWeb like for example with Cyberduck?
    I don't need to publish website with files, I just need to upload files on server. Can it be done somehow with iWeb?

    Here's some info about FTP...
    http://www.iwebformusicians.com/Search-Engine-Optimization/Upload.html
    Make sure you are allowed to store files on your server. Most hosting services don't allow anything to be stored that can't be accessed via the internet. Check the fine print!

  • Web Service to upload large files

    Hi Guys
    I'm want to create a Web Service that uploads large files (excess of 100MB). Can some help me with the best way I can do that, either C# or VB.NET.
    Thanks 

    I've done this before with a web service. It wasn't that hard. As jdweng alludes to, you need a web service and a client that knows how to repeatedly transfer the file sending it in chunks.
    The web service needs to take the file name and a byte array. It needs to return the number of bytes it received. The client calls the service endpoint passing in the name of the file and an array populated with x number of bytes. This should/could be tuned
    based on the size of the file and available bandwidth. 10 MB is a good upper limit but that may not be practical over a really slow connection... you could potentially time-out the call. The service endpoint receives the byte array and the file name. It creates/opens
    the file and appends the byte array onto the end in a synchronous manner. (You need to do it synchronous to avoid any potential race conditions with multiple calls overlapping.) When the server side is done, it returns the number of bytes written to the caller.
    The web service should close the file handle after each call since there is no guarantee that there will be another call.
    The caller then uses the return value to increment the offset pointer into the source file. When the call returns, the caller increments into the source file by the number of bytes returned then refills the buffer with the next chunk. It then re-calls the
    web service sending the same file name and the new byte array. The process continues until all of the bytes in the source file have been transferred to the server.
    I uploaded 4GB disk images to a Sharepoint server doing this so it works great and totally avoids the HTTP max payload problem.
    Good luck!

  • Uploading large files to share fails - advice appreciated

    I wondered if anyone could help me with uploading large files. 
    The ability to distribute files to clients with the included 20Gb of server space was part of the reason I signed up to Creative Cloud; unfortunatly I'm having very mixed results and I'm constantly having to fall back on the free DropBox service which seems much more reliable.
    I'm working on a MacPro - 10.6.8 with a optical fibre based ISP 20Mb/s service; the service is, dare I say, usually very quick and reliable.  Before uploading a large file I set energy saver to "never sleep" in case I wonder off for a cup of tea, but enevitably come back to what looks like a "stuck" upload.  I've tried three times today to send a 285Mb file and they all failed so I reverted to sending the file by DropBox... with no problem - I've also sent files in excess of 1Gb by DropBox in the past without an issue.
    Todays file had reached about 50% according to the little progress bar.. it then stopped.   (it's a shame the progrees bar can't relay the actual upload speed and a percentage like DropBox so you can see what's going on)
    I have to state, I'm not a DropBox "fanboy" but it just seems to work...
    Are their any tricks or settings that could make Adobe's 20Gb of space a functioning part of my workflow?  It would be great to feel confident in it.
    Thanks in advance for any advice.

    Either Firefox or Chrome would be a good alternative.
    I would see if one of these browsers works for upload but we can also debug the issue on Safari.
    Here are the steps to debug on Safari:
    Add the Develop menu to Safari. From the menu choose Safari > Preferences. Select the Advanced tab and check the box for "Show Develop menu in menu bar".
    Once logged onto the Creative Cloud Files page (https://creative.adobe.com/files) Ctrl + Click anywhere on the page and from the browser context menu choose Inspect Element.
    In the window that opens in the bottom half of the browser click on the Clock icon (mouse over info tip says Instrument) and on the left select Network Requests.
    Try the upload and you will see various network requests. Look for those whose Name is "notification". When you select one in the list there will be an arrow next to "notification" and clicking on it will take you to the details. See screen shot. If successful you will have a Status of 200. Some of these "notification" items are for your storage quota being updated, but the one we want is for the successful upload.

  • An error message uploading "large" files

    Hello!
    I'm getting an error message uploading "large" files (40-70 mb) - "unable to sync "file_name" due to server error". This message occurs only with a "large" files. The files up to 10 mb syncs perfectly and all of my files are correct (no issues with a "Windows-reserved / trailing" characters or "files in use").
    I pay 90 usd montly to Adobe and still need to use free services like Wetransfer or Dropbox. Please help.
    Thank you!
    Mindaugas

    I'm having this same problem and have written other posts but no replies. The file size for a single file is supposed to be something like 1-2 gig. How can a 40-70 meg file not work? It seems like the cloud storage feature is a selling point only and not fully realized yet. I love the interface and the idea but it's unrelaible to sync up, leave work, arrive home and no "larger" files are available. Yep, back to Dropbox and Google drive unfortunately.

  • Beach Ball of Death on OSX while trying to upload large files(updated)

    Hi,
    I did a search on this and while I got lots of hits none on this issue.
    We have a web portal that allows customers to upload files to a DAM server, we have control server.
    The issue is not the back end, the back end is php and all set up correctly, we have 2gb ram assigned to php, the upload_max_filesize is set to 1.8gb max post size is set accordingly.
    The flex app loads, user selects large file say 1.6gb ( we do file size check to make sure its below the limit), click upload button, beach ball of death shows, a little while later the file uploads.
    Great so far, now the user tries another files same size, beach ball of death, script times out (we capture this and end gracefully). If the user restarts safari then the file will upload just fine.
    Seems you can upload a large file first time and small files subsequently below 1.2gb, but you can not upload multiple large files.
    Seems some sort of memory leak, I was looking at converting this upload app to send files via ftp, but if the flash player needs to process/load the files first then I don't think this will work looking to increase the file size ceiling to 4gb via ftp.
    Code is a bit involved but in the end just calls file reference upload, then beach ball appears
    Any ideas? player version 10.0.32.18 debug
    UPDATED 09_17_09
    It appears to be a memory leak in the player when in safari, the file is read but the memory not freed up after completion. Firefox frees the used memory after upload and you can continue. Not sure if Apple or Adobe issue.
    However why bother reading the file first? can we have an HTP stream file upload in the web flash player like AIR as. This way the player would not care on the file size and the limitation would reside on the server, which we can deal with.
    Message was edited by: flashharry!

    Hi,
    I did a search on this and while I got lots of hits none on this issue.
    We have a web portal that allows customers to upload files to a DAM server, we have control server.
    The issue is not the back end, the back end is php and all set up correctly, we have 2gb ram assigned to php, the upload_max_filesize is set to 1.8gb max post size is set accordingly.
    The flex app loads, user selects large file say 1.6gb ( we do file size check to make sure its below the limit), click upload button, beach ball of death shows, a little while later the file uploads.
    Great so far, now the user tries another files same size, beach ball of death, script times out (we capture this and end gracefully). If the user restarts safari then the file will upload just fine.
    Seems you can upload a large file first time and small files subsequently below 1.2gb, but you can not upload multiple large files.
    Seems some sort of memory leak, I was looking at converting this upload app to send files via ftp, but if the flash player needs to process/load the files first then I don't think this will work looking to increase the file size ceiling to 4gb via ftp.
    Code is a bit involved but in the end just calls file reference upload, then beach ball appears
    Any ideas? player version 10.0.32.18 debug
    UPDATED 09_17_09
    It appears to be a memory leak in the player when in safari, the file is read but the memory not freed up after completion. Firefox frees the used memory after upload and you can continue. Not sure if Apple or Adobe issue.
    However why bother reading the file first? can we have an HTP stream file upload in the web flash player like AIR as. This way the player would not care on the file size and the limitation would reside on the server, which we can deal with.
    Message was edited by: flashharry!

  • Flat-File Upload - Large File Size

    Hello,
    i have to upload a very large flat-file with a size of 200 to 300 MB. The upload from the presentation server (local win xp client) fails after exceeding the time limit, because the upload performs in online-mode. I have to approaches to fix the problem.
    1. Upload the file to the application server and start upload in batch-mode.
    2. Upload the file to the application server and access/read the with "open dataset & read read dataset" line-by-line and write with a function-module into an transactional ODS-Object.
    Please provide me with informations/hints about your experience in dealing with large flat files.
    Thanks & best regards

    Hello Markus
    I would suggest you to place the file in the app server and upload in background mode.
    This is from SAP Help,
    If you want to upload a large amount of transaction data from a flat file, and you are able to specify the file type of the flat file, you should create the flat file as an ASCII file. From a performance point of view, uploading the data from an ASCII file is the most cost-effective method.
    In certain circumstances, generating an ASCII file might involve a larger workload.
    Hope it helps
    Thanks
    Chandran

  • "Unable to sync" when uploading large files to Creative Cloud

    Hi,
    I recently tried to upload some large files (1.1 GB, 2 GB) to my Creative Cloud. After completing the upload after nine hours I got the error message "unable to sync".
    Any suggestions?
    Regards,
    pdm208

    I'm having the same issue. I removed the offending big files. I turned sync on and off, but it's still unable to sync. Any thoughts?

  • Can't Upload Large Files (Upload Fails using Internet Explorer but works with Google Chrome)

    I've been experience an issue uploading large (75MB & greater) PDF files to a SharePoint 2010 document library. Using normal upload procedures using Internet Explorer 7 (our company standard for the time being) the upload fails. No error message is thrown,
    the upload screen goes away and the page refreshes and the document isn't there. I tried upload multiple and it says throws a failed error after a while.
    Using google chrome I made an attempt just to see what it did and the file using the "Add a document" uploaded in seconds. Can't figure out why one browser worked and the other doesn't. We are getting sporadic inquiries with the same issue.
    We have previously setup large file support in the appropriate areas and large files are uploaded to the sites successfully. Any thoughts?

    File size upload has to be configured on the server farm level. Your administrator most likely set
    up the limit to size of files that can be uploaded. This size can be increased and you would then be able to upload your documents.

Maybe you are looking for

  • HT4885 how do i use a macbookpro with  a thunderbolt display and have an external projector

    I bought the latest macbook pro with a 500 gig solid state drive 8 gig ram etc and a thunderbolt display, is use it with lappy lid down and external key  board and mouse, so like an imac or tower. i want to add a matrox 3 display splitter in line to

  • Employee Photo Upload Fail - "System error when archiving"

    Hi all,    When i am trying upload a photo of an employee in our ESS development portal its working fine. But the issue here is when i am trying to upload the same in testing ESS portal its displaying an error message as "System error when archiving

  • Photo uploading error in Portal

    HI Experts, I have configured the ARCHIVELINK to store the photo in Content Server. Manage to upload the photo and display the photo in PA20/30 from the backend ECC system. but my requirement is when we click and go back to same screen where we uploa

  • Array structure

    I'm stillnew to array. So i'm trying to create a dynamic structure like this: this_array = new Array(); some_values = new Array(); some_values = param.split ( " , " ) for ( i = 0 ; i < param.length ; i ++ ){ this[ "values_" + i ] = param[ i ] ; retur

  • How do i re intall photoshop on a new laptop?

    On my old laptop I purchased Photoshop but a couple months later my laptop broke down and i had to get a new one. now i cant work out how to re intall the Photoshop i have already brought to my new laptop. My laptop is windows