WebLogic Apache bridge problems on uploading large files via HTTP post

I have a problem uploading files larger than quarter a mega, the jsp
page does a POST
to a servlet which reads the input stream and writes to a file.
Configuration: Apache webserver 1.3.12 connected to the Weblogic 5.1
application server
via the bridge(mod_wl_ssl.so) from WebLogic Service pack 4.
The upload goes on for about 30 secs and throws the following error.
"Failure of WebLogic APACHE bridge:
IO error writing POST data to 100.12.1.2:7001; sys err#: [32] sys err
msg [Broken pipe]
Build date/time: Jul 10 2000 12:29:18 "
The same upload(in fact I uploaded a 8 MEG file) using the
Netscape(NSAPI) WebLogic
connector.
Any answers would be deeply appreciated.

I have a problem uploading files larger than quarter a mega, the jsp
page does a POST
to a servlet which reads the input stream and writes to a file.
Configuration: Apache webserver 1.3.12 connected to the Weblogic 5.1
application server
via the bridge(mod_wl_ssl.so) from WebLogic Service pack 4.
The upload goes on for about 30 secs and throws the following error.
"Failure of WebLogic APACHE bridge:
IO error writing POST data to 100.12.1.2:7001; sys err#: [32] sys err
msg [Broken pipe]
Build date/time: Jul 10 2000 12:29:18 "
The same upload(in fact I uploaded a 8 MEG file) using the
Netscape(NSAPI) WebLogic
connector.
Any answers would be deeply appreciated.

Similar Messages

  • Uploading big files via http post to XDB table

    Hello,
    I ve created a web form for to upload files as blobs to the database using XML DB, DBPS_EPG package and preconfigured DAD.
    The issue is that small files (~10kb) are being uploaded ok, but file which is 100K during http post returns connection reset from the server side.
    To my opinion there might be some max file size parameter for oracle server to accept during oracle listener tcp connection (as apach has maxrequestlimit).
    Is there any workaround for to load large files to the XDB table via webform?
    Here is a piece of code:
    CREATE USER web IDENTIFIED BY web_upload;
    ALTER USER web DEFAULT TABLESPACE XML_DATA;
    ALTER USER web QUOTA UNLIMITED ON XML_DATA;
    ALTER USER web TEMPORARY TABLESPACE TEMP;
    GRANT CONNECT, ALTER SESSION TO web;
    GRANT CREATE TABLE, CREATE PROCEDURE TO web;
    ALTER SESSION SET CURRENT_SCHEMA = XDB;
    BEGIN DBMS_EPG.CREATE_DAD('WEB', '/upload/*'); END;
    BEGIN
        DBMS_EPG.SET_DAD_ATTRIBUTE('WEB', 'database-username', 'WEB');
        DBMS_EPG.SET_DAD_ATTRIBUTE('WEB', 'document-table-name', 'uploads');
        DBMS_EPG.SET_DAD_ATTRIBUTE('WEB', 'nls-language', '.al32utf8');
        DBMS_EPG.SET_DAD_ATTRIBUTE('WEB', 'default-page', 'upload');
        COMMIT;
    END; 
    ALTER SESSION SET CURRENT_SCHEMA = SYS;
    CREATE TABLE web.uploads (
        name VARCHAR2(256) NOT NULL
            CONSTRAINT pk_uploads PRIMARY KEY,
        mime_type VARCHAR2(128),
        doc_size NUMBER,
        dad_charset VARCHAR2(128),
        last_updated DATE,
        content_type VARCHAR2(128) DEFAULT 'BLOB',
        blob_content BLOB
    CREATE OR REPLACE PROCEDURE web.upload
    AS
      url VARCHAR2(256) := 'http://demo.test.com:9090/upload/uploaded';
    BEGIN
        HTP.P('<html>');
        HTP.P('<head>');
        HTP.P('  <title>Upload</title>');
        HTP.P('</head>');
        HTP.P('<body>');
        HTP.P('  <h1>Upload</h1>');
        HTP.P('  <form method="post"');
        HTP.P('      action="' || url || '"');
        HTP.P('      enctype="multipart/form-data">');
        HTP.P('    <p><input type="file" name="binaryfile" /></p>');
        HTP.P('    <p><input type="file" name="binaryfile" /></p>');
        HTP.P('    <p><button type="submit">Upload</button></p>');
        HTP.P('  </form>');
        HTP.P('</body>');
        HTP.P('</html>');
    END;
    CREATE OR REPLACE PROCEDURE web.uploaded (
        binaryfile OWA.VC_ARR
    AS
    BEGIN
        HTP.P('<html>');
        HTP.P('<head>');
        HTP.P('  <title>Uploaded</title>');
        HTP.P('</head>');
        HTP.P('<body>');
        HTP.P('  <h1>Uploaded</h1>');
        FOR i IN 1 .. binaryfile.COUNT LOOP
            IF binaryfile(i) IS NOT NULL THEN
                HTP.P('  <p>File: ' || binaryfile(i) || '</p>');
            END IF;
        END LOOP;
        HTP.P('</body>');
        HTP.P('</html>');
    END;
    /帖子经 anatoly编辑过

    Welcome to Apple Discussions!
    Much of what is available on Bittorrent is not legal, beta, or improperly labelled versions. If you want public domain software, see my FAQ*:
    http://www.macmaps.com/macosxnative.html#NATIVE
    for search engines of legitimate public domain sites.
    http://www.rbrowser.com/ has a light mode that supports binary without SSH security.
    http://rsug.itd.umich.edu/software/fugu/ has ssh secure FTP.
    Both I find are quick and a lot more reliable than Fetch. I know Fetch used to be the staple FTP program, but it grew too big for my use.
    - * Links to my pages may give me compensation.

  • Uploading Very Large Files via HTTP

    I am developing some classes that must upload files to a web server via HTTP and multipart/form-data. I am using Apache's Tomcat FileUpload library contained within the commons-fileupload-1.0.jar file on the server side. My code fails on large files or large quantities of small files because of the memory restriction of the VM. For example when uploading a 429 MB file I get this exception:
    java.lang.OutOfMemoryError
    Exception in thread "main"I have never been successful in uploading, regardless of the server-side component, more than ~30 MB.
    In a production environment I cannot alter the clients VM memory setting, so I must code my client classes to handle such cases.
    How can this be done in Java? This is the method that reads in a selected file and immediately writes it upon the output stream to the web resource as referenced by bufferedOutputStream:
    private void write(File file) throws IOException {
      byte[] buffer = new byte[bufferSize];
      BufferedInputStream fileInputStream = new BufferedInputStream(new FileInputStream(file));
      // read in the file
      if (file.isFile()) {
        System.out.print("----- " + file.getName() + " -----");
        while (fileInputStream.available() > 0) {
          if (fileInputStream.available() >= 0 &&
              fileInputStream.available() < bufferSize) {
            buffer = new byte[fileInputStream.available()];
          fileInputStream.read(buffer, 0, buffer.length);
          bufferedOutputStream.write(buffer);
          bufferedOutputStream.flush();
        // close the files input stream
        try {
          fileInputStream.close();
        } catch (IOException ignored) {
          fileInputStream = null;
      else {
        // do nothing for now
    }The problem is, the entire file, and any subsequent files being read in, are all being packed onto the output stream and don't begin actually moving until close() is called. Eventually the VM gives way.
    I require my client code to behave no different than the typcial web browser when uploading or downloading a file via HTTP. I know of several commercial applets that can do this, why can't I? Can someone please educate me or at least point me to a useful resource?
    Thank you,
    Henryiv

    Are you guys suggesting that the failures I'm
    experiencing in my client code is a direct result of
    the web resource's (servlet) caching of my request
    (files)? Because the exception that I am catching is
    on the client machine and is not generated by the web
    server.
    trumpetinc, your last statement intrigues me. It
    sounds as if you are suggesting having the client code
    and the servlet code open sockets and talk directly
    with one another. I don't think out customers would
    like that too much.Answering your first question:
    Your original post made it sound like the server is running out of memory. Is the out of memory error happening in your client code???
    If so, then the code you provided is a bit confusing - you don't tell us where you are getting the bufferedOutputStream - I guess I'll just assume that it is a properly configured member variable.
    OK - so now, on to what is actually causing your problem:
    You are sending the stream in a very odd way. I highly suspect that your call to
    buffer = new byte[fileInputStream.available()];is resulting in a massive buffer (fileInputStream.available() probably just returns the size of the file).
    This is what is causing your out of memory problem.
    The proper way to send a stream is as follows:
         static public void sendStream(InputStream is, OutputStream os, int bufsize)
                     throws IOException {
              byte[] buf = new byte[bufsize];
              int n;
              while ((n = is.read(buf)) > 0) {
                   os.write(buf, 0, n);
         static public void sendStream(InputStream is, OutputStream os)
                     throws IOException {
              sendStream(is, os, 2048);
         }the simple implementation with the hard coded 2048 buffer size is fine for almost any situation.
    Note that in your code, you are allocating a new buffer every time through your loop. The purpose of a buffer is to have a block of memory allocated that you then move data into and out of.
    Answering your second question:
    No - actually, I'm suggesting that you use an HTTPUrlConnection to connect to your servlet directly - no need for special sockets or ports, or even custom protocols.
    Just emulate what your browser does, but do it in the applet instead.
    There's nothing that says that you can't send a large payload to an http servlet without multi-part mime encoding it. It's just that is what browsers do when uploading a file using a standard HTML form tag.
    I can't see that a customer would have anything to say on the matter at all - you are using standard ports and standard communication protocols... Unless you are not in control of the server side implementation, and they've already dictated that you will mime-encode the upload. If that is the case, and they are really supporting uploads of huge files like this, then their architect should be encouraged to think of a more efficient upload mechanism (like the one I describe) that does NOT mime encode the file contents.
    - K

  • How to upload large file with http via post

    Hi guys,
    Does anybody know how to upload large file (>100 MB) use applet to servlet with http via post method? Thanks in advance.
    Regards,
    Mark.

    Hi SuckRatE
    Thanks for your reply. Could you give me some client side code to upload a large file. I use URL to connect to server. It throws out of memory exception. The part of client code is below:
    // connect to the servlet
    URL theServlet = new URL(servletLocation);
    URLConnection servletConnection = theServlet.openConnection();
    // inform the connection that we will send output and accept input
    servletConnection.setDoInput(true);
    servletConnection.setDoOutput(true);
    // Don't used a cached version of URL connection.
    servletConnection.setUseCaches (false);
    servletConnection.setDefaultUseCaches(false);
    // Specify the content type that we will send text data
    servletConnection.setRequestProperty("Content-Type",
    +"application/octet-stream");
    // send the user string to the servlet.
    OutputStream outStream = servletConnection.getOutputStream();
    FileInputStream filein = new FileInputStream(largeFile);
    //BufferedReader in = new BufferedReader(new InputStreamReader
    +(servletConnection.getInputStream()));
    //System.out.println("tempCurrent = "+in.readLine());
    byte abyte[] = new byte[2048];
    int cnt = 0;
    while((cnt = filein.read(abyte)) > 0)
    outStream.write(abyte, 0, cnt);
    filein.close();
    outStream.flush();
    outStream.close();
    Regards,
    Mark.

  • OOM happens inside Weblogic 10.3.6 when application uploads large files

    Oracle Fusion BI Apps application is uploading large files (100+ MB) onto Oracle Cloud Storage. This application works properly when ran outside weblogic server. When deployed on Fusion Middleware Weblogic 10.3.6, during upload of large files we get this OOM error
    java.lang.OutOfMemoryError: allocLargeObjectOrArray: [B, size 268435472
        at jrockit/vm/Allocator.allocLargeObjectOrArray(JIZ)Ljava/lang/Object;(Native Method)
        at jrockit/vm/Allocator.allocObjectOrArray(Allocator.java:349)[optimized]
        at weblogic/utils/io/UnsyncByteArrayOutputStream.resizeBuffer(UnsyncByteArrayOutputStream.java:59)[inlined]
        at weblogic/utils/io/UnsyncByteArrayOutputStream.write(UnsyncByteArrayOutputStream.java:89)[optimized]
        at com/sun/jersey/api/client/CommittingOutputStream.write(CommittingOutputStream.java:90)
        at com/sun/jersey/core/util/ReaderWriter.writeTo(ReaderWriter.java:115)
        at com/sun/jersey/core/provider/AbstractMessageReaderWriterProvider.writeTo(AbstractMessageReaderWriterProvider.java:76)
        at com/sun/jersey/core/impl/provider/entity/InputStreamProvider.writeTo(InputStreamProvider.java:98)
        at com/sun/jersey/core/impl/provider/entity/InputStreamProvider.writeTo(InputStreamProvider.java:59)
        at com/sun/jersey/api/client/RequestWriter.writeRequestEntity(RequestWriter.java:300)
        at com/sun/jersey/client/urlconnection/URLConnectionClientHandler._invoke(URLConnectionClientHandler.java:213)
        at com/sun/jersey/client/urlconnection/URLConnectionClientHandler.handle(URLConnectionClientHandler.java:149)
    Looks like weblogic is using its default Weblogic HTTP handler, switching to Sun HTTP handler via start up JVM/Java Option "-DUseSunHttpHandler=true" solves the OOM issue.
    Seems instead of streaming the file content with some fixed size byte array its being on the whole file into memory during upload.
    Is it possible to solve this OOM by changing any setting of Weblogic HTPP handler without switching to Sun HTTP handler as there are many other application deployed on this weblogic instance?
    We are concerned whether there will be any impact on performance or any other issue.
    Please advice, highly appreciate your response.
    Thanks!

    Hi,
    If you have a back then restore the below file back and then try to start weblogic:
    \Oracle\Middleware\user_projects\domains\<domain_name>\config\config.lok
    Thanks,
    Sharmela

  • Upload large file on doSubmit in javascript

    Hi all,
    I'm currently having some problems uploading large files using a custom Upload button (button calls doSubmit from javascript).
    When using an Upload button with unconditional branch to the same page (no js) it works perfectly. Now I have to use my own Upload button that calls some javascript and AJAX and then Submit the page (with doSubmit); in this case I can only upload very small files (< 100k) ...
    I understand the mechanism of the file browse and its uploading to the table WWV_FLOW_FILE_OBJECTS, but I cannot figure out why my configuration does not work when it does stricly the same but from js ...
    Someone experienced the same issue ? How to submit the upload in javascript side with doSubmit ??
    Best regards,
    Othman

    Hi Othman,
    did you find a solution for this. I'm facing nearly the same problem, I'm calling a popup, where the user can upload up to 5 files. Then on clicking a button, javascript is called, that's calling the doSubmit and some others things and at the end the popup is closed.
    Sometimes the upload of 5 files is working, sometimes only 1 file is possible (obvisiously it's depending on the filesize).
    function addFile()
    doSubmit('FILE');
    opener.actualizeProject('TAB4');
    close();
    }chrissy

  • Filesize a problem when uploading a file to Oracle via JDBC?

    Hello,
    I am trying to upload a file via JDBC to an Oracle Database (into a BLOB field). The file is converted to a byte array for uploading.
    While processing the request the following error is thrown :
    [31.08.05 09:58:33:151 CEST] 40cc40cc SystemErr R Caused by: java.sql.SQLException: Datengr��e gr��er als max. Gr��e f�r diesen Typ: 485323
    [31.08.05 09:58:33:151 CEST] 40cc40cc SystemErr R      at java.lang.Throwable.<init>(Throwable.java)
    [31.08.05 09:58:33:151 CEST] 40cc40cc SystemErr R      at java.lang.Throwable.<init>(Throwable.java)
    [31.08.05 09:58:33:151 CEST] 40cc40cc SystemErr R      at java.sql.SQLException.<init>(SQLException.java:52)
    [31.08.05 09:58:33:151 CEST] 40cc40cc SystemErr R      at oracle.jdbc.dbaccess.DBError.throwSqlException(DBError.java:134)
    [31.08.05 09:58:33:151 CEST] 40cc40cc SystemErr R      at oracle.jdbc.dbaccess.DBError.throwSqlException(DBError.java:179)
    [31.08.05 09:58:33:151 CEST] 40cc40cc SystemErr R      at oracle.jdbc.ttc7.TTCItem.setArrayData(TTCItem.java:95)
    [31.08.05 09:58:33:151 CEST] 40cc40cc SystemErr R      at oracle.jdbc.dbaccess.DBDataSetImpl.setBytesBindItem(DBDataSetImpl.java:2414)
    [31.08.05 09:58:33:151 CEST] 40cc40cc SystemErr R      at oracle.jdbc.driver.OraclePreparedStatement.setItem(OraclePreparedStatement.java:1134)
    [31.08.05 09:58:33:151 CEST] 40cc40cc SystemErr R      at oracle.jdbc.driver.OraclePreparedStatement.setBytes(OraclePreparedStatement.java:2170)
    ("Datengr��e gr��er als max. Gr��e f�r diesen Typ" means: "Filesize bigger than maximum allowed size for this type")
    A blob is able to hold files with gigabyte size, a simple 460k file should not cause this JDBC error. I had a test with a 127k file, too, same error (different file types tested: html, doc, pdf, ...). Is this a problem of the driver or is there anythig I am doing wrong?
    Many thanks in advance for your help,
    Rommie.

    My problem is not to retrieve the file but to upload
    it. I checked the examples and only saw examples for
    retrieving a BLOB or CLOB, but not for inserting it
    into the database
    Maybe I am blind ... if so, please publish your link
    to the example.The page at:
    http://www.oracle.com/technology/sample_code/tech/java/sqlj_jdbc/files/advanced/advanced.html
    Has a subsection that reads:
    LOB Datatype [19-Feb-2003]
    The term large object (LOB) refers to a data item that is too large to be stored directly in a database table. This sample application demonstrates how to read and write LOB datatype.The Oracle JDBC drivers provide support for two types of LOBs: BLOBs (unstructured binary data) and CLOBs (character data). BLOB and CLOB data is accessed and referenced using a locator stored in the database table.
    Download Now (ZIP, 76KB)
    Readme
    Source
    The zip is a standalone Java application that will create any needed tables, populate them with example data (both LOBs and CLOBs) from the data files in the zip and then put up a Swing gui that will allow you to add additional LOB and CLOB data to the table. Everything is there, the readme explains the 2 properties files you need to edit. The only other thing besides reading comprehension that you need to supply is DB connection information with the necessary privileges.
    The Source link is a link to the main code, but does not include all the code in the package. However, it contains the code to write CLOBs and BLOBs into the database.
    The Readme explains everything. It also has code excerpts, including the code used to read the sample data from file and write it into the LOBs and CLOBs.
    http://www.oracle.com/technology/sample_code/tech/java/sqlj_jdbc/files/advanced/LOBSample/Readme.html
    Look for "loadSamples"

  • Uploading larger files in iPlanet 4.1.11

    I am using iPlanet 4.1.11 as my web server and WebSphere 4.0.4 as my application server.
    I am trying to upload files of size greater than 10MB using multipart in my application. But I can upload files of size less than 10MB. If it is above 10 MB, HTTP connection is closed without any response.
    In WebLogic we have maximum post size parameter which can be set to enable uploading larger file. Is there any similar configuration settings in iPlanet or WebSphere? Or is there any relation between file cache and upload size? Is there any hardware limitations? Please help.

    I'm not aware of any config settings on iPlanet Webserver side that restrict upload file size. Probably, its something in the application or app server that is causing this..
    Note: 4.1 version has reached end of support life, you should consider upgrading to a currently supported web server version.
    Thanks
    Manish

  • "Not connected to internet" error only when uploading large files.

    Hi,
    I have a strange problem that occurs when im uploading large files to any of my servers....
    The files upload just fine but while im uploading im unable to use the internet for anything else. For example if I try to open a web page I get an error saying im not connected to the internet. This error will persist until the file finishes uploading.
    Its a strange and annoying problem and was wondering if anybody could throw any light on the problem.
    I cant be sure but it may have coincided with me connecting a new Netgear wireless modem router.
    Many thanks
    Matt

    Hi Sorry, only just noticed your reply. Didn't get an email alert like I usually do.
    Thanks for that. To be honest I have no idea what QOS is but I will take a look at the routers settings to see if it makes sense to me.
    Thanks
    Matt

  • ITunes (Win) crashing when trying to match/upload large files

    Hi there,
    I've got a little problem with iTunes Match again. Imho it is still veeery buggy but it certainly improved a bit!
    So the problem is:
    I am listening to mixes a lot - meaning: large mp3 files with runtimes up to 3hrs.
    When I load files like these into my iTunes lib on Windows 7 and then try to upload them into the cloud 9 out of 10 times iTunes simply crashes while analyzing these large files ("iTunes stopped working"). Nothing I can do. Even trying to upload these files one by one didn't help.
    Does anyone have any ideas on what to do about this?
    thanks, cheers
    P

    Thanks for the reply!
    I know that 200mb is the limit. But as you already noted it should just say "not compatible" or whatever but not just simply crash. Problem is I can't have these tracks in my iTunes library then because every time I try to sync my lib with Match it tries to sync these files as well and iTunes just simply crashes.
    I also have a Macbook so I will try to upload the files via iTunes on the Mac. If that doesn't help last idea may be to reduce file size ... although I hate doing this as I prefer 256kbit files.
    Split into two would be another idea.
    Anyways I had these problems with large files couple of times already but could always fix it by syncing one file at a time. not this time though.

  • Web Service to upload large files

    Hi Guys
    I'm want to create a Web Service that uploads large files (excess of 100MB). Can some help me with the best way I can do that, either C# or VB.NET.
    Thanks 

    I've done this before with a web service. It wasn't that hard. As jdweng alludes to, you need a web service and a client that knows how to repeatedly transfer the file sending it in chunks.
    The web service needs to take the file name and a byte array. It needs to return the number of bytes it received. The client calls the service endpoint passing in the name of the file and an array populated with x number of bytes. This should/could be tuned
    based on the size of the file and available bandwidth. 10 MB is a good upper limit but that may not be practical over a really slow connection... you could potentially time-out the call. The service endpoint receives the byte array and the file name. It creates/opens
    the file and appends the byte array onto the end in a synchronous manner. (You need to do it synchronous to avoid any potential race conditions with multiple calls overlapping.) When the server side is done, it returns the number of bytes written to the caller.
    The web service should close the file handle after each call since there is no guarantee that there will be another call.
    The caller then uses the return value to increment the offset pointer into the source file. When the call returns, the caller increments into the source file by the number of bytes returned then refills the buffer with the next chunk. It then re-calls the
    web service sending the same file name and the new byte array. The process continues until all of the bytes in the source file have been transferred to the server.
    I uploaded 4GB disk images to a Sharepoint server doing this so it works great and totally avoids the HTTP max payload problem.
    Good luck!

  • Uploading large files to share fails - advice appreciated

    I wondered if anyone could help me with uploading large files. 
    The ability to distribute files to clients with the included 20Gb of server space was part of the reason I signed up to Creative Cloud; unfortunatly I'm having very mixed results and I'm constantly having to fall back on the free DropBox service which seems much more reliable.
    I'm working on a MacPro - 10.6.8 with a optical fibre based ISP 20Mb/s service; the service is, dare I say, usually very quick and reliable.  Before uploading a large file I set energy saver to "never sleep" in case I wonder off for a cup of tea, but enevitably come back to what looks like a "stuck" upload.  I've tried three times today to send a 285Mb file and they all failed so I reverted to sending the file by DropBox... with no problem - I've also sent files in excess of 1Gb by DropBox in the past without an issue.
    Todays file had reached about 50% according to the little progress bar.. it then stopped.   (it's a shame the progrees bar can't relay the actual upload speed and a percentage like DropBox so you can see what's going on)
    I have to state, I'm not a DropBox "fanboy" but it just seems to work...
    Are their any tricks or settings that could make Adobe's 20Gb of space a functioning part of my workflow?  It would be great to feel confident in it.
    Thanks in advance for any advice.

    Either Firefox or Chrome would be a good alternative.
    I would see if one of these browsers works for upload but we can also debug the issue on Safari.
    Here are the steps to debug on Safari:
    Add the Develop menu to Safari. From the menu choose Safari > Preferences. Select the Advanced tab and check the box for "Show Develop menu in menu bar".
    Once logged onto the Creative Cloud Files page (https://creative.adobe.com/files) Ctrl + Click anywhere on the page and from the browser context menu choose Inspect Element.
    In the window that opens in the bottom half of the browser click on the Clock icon (mouse over info tip says Instrument) and on the left select Network Requests.
    Try the upload and you will see various network requests. Look for those whose Name is "notification". When you select one in the list there will be an arrow next to "notification" and clicking on it will take you to the details. See screen shot. If successful you will have a Status of 200. Some of these "notification" items are for your storage quota being updated, but the one we want is for the successful upload.

  • An error message uploading "large" files

    Hello!
    I'm getting an error message uploading "large" files (40-70 mb) - "unable to sync "file_name" due to server error". This message occurs only with a "large" files. The files up to 10 mb syncs perfectly and all of my files are correct (no issues with a "Windows-reserved / trailing" characters or "files in use").
    I pay 90 usd montly to Adobe and still need to use free services like Wetransfer or Dropbox. Please help.
    Thank you!
    Mindaugas

    I'm having this same problem and have written other posts but no replies. The file size for a single file is supposed to be something like 1-2 gig. How can a 40-70 meg file not work? It seems like the cloud storage feature is a selling point only and not fully realized yet. I love the interface and the idea but it's unrelaible to sync up, leave work, arrive home and no "larger" files are available. Yep, back to Dropbox and Google drive unfortunately.

  • Problem in Uploading a File by Applet

    Hi Members,
    * I have faced problem while uploading a file from client to server by ftp protocol using APPLET(No JSP) only
    * I am getting exception while running....
    * My source code is as follows,
    import java.awt.*;
    import java.applet.*;
    import java.awt.event.*;
    import java.io.BufferedInputStream;
    import java.io.BufferedOutputStream;
    import java.io.File;
    import java.io.FileInputStream;
    import java.io.IOException;
    import java.net.MalformedURLException;
    import java.net.URL;
    import java.net.URLConnection;
    public class UploadAndDownload extends Applet implements ActionListener {
         Button upload;
         Button browse;
         TextField filename;
         File source = null;
         Label name;
         StringBuffer sb;
         BufferedInputStream bis = null;
         BufferedOutputStream bos = null;
         public void init() {
              setLayout(new FlowLayout());
              upload = new Button("Upload");
              browse = new Button("Browse");
              name = new Label("Filename");
              filename = new TextField("", 45);
              add(name);
              add(filename);
              add(upload);
              add(browse);
              upload.addActionListener(this);
              browse.addActionListener(this);
         public void actionPerformed(ActionEvent evt) {
              // Code for browsing a file
              String input_file_name = "";
              if (evt.getSource() == browse)
                   Frame parent = new Frame();
                   FileDialog fd = new FileDialog(parent, "Select a file", FileDialog.LOAD);
                   fd.setVisible(true);
                   input_file_name = fd.getFile();
                   filename.setText(input_file_name);
                   // Gets the file from the file dialog and assign it to the source
                   source = new File(input_file_name);
                   repaint();
              // Code for Uploading a file to the server
              if (evt.getSource() == upload) {
                   // Appending the server pathname in string buffer
                   sb = new StringBuffer("ftp://");
                   sb.append("2847");
                   sb.append(':');
                   sb.append("Websphere25");
                   sb.append("@");
                   sb.append("172.16.1.111");
                   sb.append('/');
                   sb.append(input_file_name);
                   sb.append(";type=i");
                   try {
                        URL url = new URL(sb.toString());
                        URLConnection urlc = url.openConnection();
                        bos = new BufferedOutputStream(urlc.getOutputStream());
                        bis = new BufferedInputStream(new FileInputStream(source));
                        int i;
                        // Read from the inputstream and write it to the outputstream
                        while ((i = bis.read()) != -1) {
                             bos.write(i);
                   } catch (MalformedURLException e) {
                        e.printStackTrace();
                   } catch (IOException e) {
                        e.printStackTrace();
                   } finally {
                        if (bis != null)
                             try {
                                  bis.close();
                             } catch (IOException ioe) {
                                  ioe.printStackTrace();
                        if (bos != null)
                             try {
                                  bos.close();
                             } catch (IOException ioe) {
                                  ioe.printStackTrace();
    MY EXCEPTION IS,
    Exception in thread "AWT-EventQueue-2" java.security.AccessControlException: access denied (java.net.SocketPermission 172.16.1.111:80 connect,resolve)
         at java.security.AccessControlContext.checkPermission(Unknown Source)
         at java.security.AccessController.checkPermission(Unknown Source)
         at java.lang.SecurityManager.checkPermission(Unknown Source)
         at java.lang.SecurityManager.checkConnect(Unknown Source)
         at sun.net.www.http.HttpClient.openServer(Unknown Source)
         at sun.net.www.http.HttpClient.<init>(Unknown Source)
         at sun.net.www.http.HttpClient.New(Unknown Source)
         at sun.net.www.http.HttpClient.New(Unknown Source)
         at sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(Unknown Source)
         at sun.net.www.protocol.http.HttpURLConnection.plainConnect(Unknown Source)
         at sun.net.www.protocol.http.HttpURLConnection.connect(Unknown Source)
         at sun.net.www.protocol.ftp.FtpURLConnection.connect(Unknown Source)
         at sun.net.www.protocol.ftp.FtpURLConnection.getOutputStream(Unknown Source)
         at UploadAndDownload.actionPerformed(UploadAndDownload.java:68)
         at java.awt.Button.processActionEvent(Unknown Source)
         at java.awt.Button.processEvent(Unknown Source)
         at java.awt.Component.dispatchEventImpl(Unknown Source)
         at java.awt.Component.dispatchEvent(Unknown Source)
         at java.awt.EventQueue.dispatchEvent(Unknown Source)
         at java.awt.EventDispatchThread.pumpOneEventForHierarchy(Unknown Source)
         at java.awt.EventDispatchThread.pumpEventsForHierarchy(Unknown Source)
         at java.awt.EventDispatchThread.pumpEvents(Unknown Source)
         at java.awt.EventDispatchThread.pumpEvents(Unknown Source)
         at java.awt.EventDispatchThread.run(Unknown Source)* Please let me know what problem in my code....
    * Thanks in advance....

    * Thanks for your reply....
    * I have signed my policy file by giving AllPermission and mentioned in java.security file in bin folder....
    * My question is , by giving AllPermission , can we access and do all permissions like ( SecurityPermission, AWTPermission, SocketPermission, NetPermission, FilePermission, SecurityPermission etc )...
    * My policy file is looks like follow,
    /* AUTOMATICALLY GENERATED ON Tue Apr 16 17:20:59 EDT 2002*/
    /* DO NOT EDIT */
    grant {
      permission java.security.AllPermission;
    };* If i signed the policy like above, and when i run the applet file in InternetExplorer now , it thorws the following exception on my console,
    java.net.ProtocolException: cannot write to a URLConnection if doOutput=false - call setDoOutput(true)
         at sun.net.www.protocol.http.HttpURLConnection.getOutputStream(Unknown Source)
         at sun.net.www.protocol.ftp.FtpURLConnection.getOutputStream(Unknown Source)
         at UploadAndDownload.actionPerformed(UploadAndDownload.java:68)
         at java.awt.Button.processActionEvent(Unknown Source)
         at java.awt.Button.processEvent(Unknown Source)
         at java.awt.Component.dispatchEventImpl(Unknown Source)
         at java.awt.Component.dispatchEvent(Unknown Source)
         at java.awt.EventQueue.dispatchEvent(Unknown Source)
         at java.awt.EventDispatchThread.pumpOneEventForHierarchy(Unknown Source)
         at java.awt.EventDispatchThread.pumpEventsForHierarchy(Unknown Source)
         at java.awt.EventDispatchThread.pumpEvents(Unknown Source)
         at java.awt.EventDispatchThread.pumpEvents(Unknown Source)
         at java.awt.EventDispatchThread.run(Unknown Source)* Please let me know , how to solve this and give me your suggestion on the above process...
    * Thanks in advance...
    Regards,
    JavaImran

  • Problem with uploading a file in Clustered Environment

    Hi,
    I have a problem with uploading a file in a clustered environment. I have an iview component which facilitates an upload action of an xml config file. The problem is that the upload of the modified XML file is reflected only in the central instance of the cluster and not in the dialog instances. The dialog instances hold the old config file.
    Is there any solution to upload the file to all the nodes in the cluster.
    Thanks
    Kiran

    Hi,
    This is a known problem with clustered environment. Remember that your portal component runs on just on dialog instance and it doesn't automatically have access to the other nodes.  However, there are some ways to get around this
    1. Use KM to store files. KM is a common repository for all application servers and therefore you needn't worry more
    2. Use an external batch oriented product (suresync/robocopy) to synch folders on the different DIs. You basically use your existing portal component, but there is a batch job which makes sure the upload folder is identical on all DIs (however, there is a slight delay depending on how often you run the batch job)
    3. Store the files on a shared disk directly from the portal component.
    Cheers
    Dagfinn

Maybe you are looking for

  • Missing Java??

    I recently downloaded the new Safari 5.0.6 on my MDD 1.25 G4. I have a special program i use in my work called Retailgis that uses my Java to link with photos in iphoto.I've used this for 2 years without any problems. I tried to use it the other day

  • In dealer Invoice , Excise duty values not captured from MIGO

    In dealer invoice , I have changed excise duty values in MIGO . This is not captured in MIRO. But the system took excise value from default values ( from FV11 condition record ) . This is happening for New plants.  Old plants works good. Regards., La

  • Grouping set

    What is the difference between grouping sets and grouping. Can any one explain to me in easy way. I will be appreciate. Thanks.

  • SAP EHP Update for Large Database

    Dear Experts, We are planning for the SAP EHP7 update for our system. Please find the system details below Source system: SAP ERP6.0 OS: AIX DB: Oracle 11.2.0.3 Target System: SAP ERP6.0 EHP7 OS: AIX DB: 11.2.0.3 RAM: 32 GB The main concern over here

  • If you want to learn the Displace Filter

    If you want to learn the Displace Filter Welles Goodrich - 09:45am May 9, 2004 Pacific Recently I decided to undertake a study of the Displace Filter and found three tutorials which were excellent. Steve's series of tutorials presents a wide variety