"Unable to sync" when uploading large files to Creative Cloud

Hi,
I recently tried to upload some large files (1.1 GB, 2 GB) to my Creative Cloud. After completing the upload after nine hours I got the error message "unable to sync".
Any suggestions?
Regards,
pdm208

I'm having the same issue. I removed the offending big files. I turned sync on and off, but it's still unable to sync. Any thoughts?

Similar Messages

  • "Not connected to internet" error only when uploading large files.

    Hi,
    I have a strange problem that occurs when im uploading large files to any of my servers....
    The files upload just fine but while im uploading im unable to use the internet for anything else. For example if I try to open a web page I get an error saying im not connected to the internet. This error will persist until the file finishes uploading.
    Its a strange and annoying problem and was wondering if anybody could throw any light on the problem.
    I cant be sure but it may have coincided with me connecting a new Netgear wireless modem router.
    Many thanks
    Matt

    Hi Sorry, only just noticed your reply. Didn't get an email alert like I usually do.
    Thanks for that. To be honest I have no idea what QOS is but I will take a look at the routers settings to see if it makes sense to me.
    Thanks
    Matt

  • When sharing InDesign files via Creative Cloud, is there a way to display spreads instead of individual pages?

    We're designing a booklet, and the client needs to see the spreads, not just one page at a time. I can export a .pdf... but I'd like to just share the .indd file.
    [Moved from the generic Cloud forum to the specific Program forum... Mod]

    I will be sending him a link to the file in Creative Cloud, so he'll be using his internet browser.

  • When uploading large file from network, all other requests are not being sent to server

    In our application we are facing a weird scenario in which when we try to upload a huge file(600+ MB) from network, all other requests(AJAX) are getting blocked. But when the same file is being uploaded from local location(E drive) then everything works fine.

    How are you uploading the file via network?
    If you are troubleshooting the network bandwidth? or the threads where one process takes precedent over the other you will have to troubleshoot the QoS of the network you see this on.
    To give a better visual please use Firebug or the Web Developer tool called Network to analyze the requests. The "XMLHttpRequest" will give better clues for Ajax. Reference [http://ajaxian.com/archives/ajax-debugging-with-firebug]
    There may be a about:config option that allows mutithreading or max persistent connections per server you can check.

  • Upload Dialog Box Won't Go Away When Uploading Larger Files

    I'm uploading a 200 MB Excel file to Power BI and the upload dialog box never goes away, even after the upload is completed. I verified this by checking network activity from Internet Explorer and additionally by opening a new Power BI window and verifying
    that the file load took place. I think this is a bug.

    Guy, what are the next steps? Should Joey file a bug on Connect?
    Thanks!
    Ed Price, Power BI & SQL Server Customer Program Manager (Blog,
    Small Basic,
    Wiki Ninjas,
    Wiki)
    Answer an interesting question?
    Create a wiki article about it!

  • Can you selectively sync files for Creative Cloud?

    On my laptop I only want to sync specific folders or files with Creative Cloud. On my Desktop I want to sync all the files between both devices. Dropbox has a feature that allows you to select which cloud files to sync per device, can I accomplish this with CC?

    I second this request as I only have a 250gb SSD on my laptop so syncing all my creative cloud files (70gb+) just takes up too much space although on my iMac (1TB) it's no problem. Here is how it's handled in Dropbox:

  • OOM happens inside Weblogic 10.3.6 when application uploads large files

    Oracle Fusion BI Apps application is uploading large files (100+ MB) onto Oracle Cloud Storage. This application works properly when ran outside weblogic server. When deployed on Fusion Middleware Weblogic 10.3.6, during upload of large files we get this OOM error
    java.lang.OutOfMemoryError: allocLargeObjectOrArray: [B, size 268435472
        at jrockit/vm/Allocator.allocLargeObjectOrArray(JIZ)Ljava/lang/Object;(Native Method)
        at jrockit/vm/Allocator.allocObjectOrArray(Allocator.java:349)[optimized]
        at weblogic/utils/io/UnsyncByteArrayOutputStream.resizeBuffer(UnsyncByteArrayOutputStream.java:59)[inlined]
        at weblogic/utils/io/UnsyncByteArrayOutputStream.write(UnsyncByteArrayOutputStream.java:89)[optimized]
        at com/sun/jersey/api/client/CommittingOutputStream.write(CommittingOutputStream.java:90)
        at com/sun/jersey/core/util/ReaderWriter.writeTo(ReaderWriter.java:115)
        at com/sun/jersey/core/provider/AbstractMessageReaderWriterProvider.writeTo(AbstractMessageReaderWriterProvider.java:76)
        at com/sun/jersey/core/impl/provider/entity/InputStreamProvider.writeTo(InputStreamProvider.java:98)
        at com/sun/jersey/core/impl/provider/entity/InputStreamProvider.writeTo(InputStreamProvider.java:59)
        at com/sun/jersey/api/client/RequestWriter.writeRequestEntity(RequestWriter.java:300)
        at com/sun/jersey/client/urlconnection/URLConnectionClientHandler._invoke(URLConnectionClientHandler.java:213)
        at com/sun/jersey/client/urlconnection/URLConnectionClientHandler.handle(URLConnectionClientHandler.java:149)
    Looks like weblogic is using its default Weblogic HTTP handler, switching to Sun HTTP handler via start up JVM/Java Option "-DUseSunHttpHandler=true" solves the OOM issue.
    Seems instead of streaming the file content with some fixed size byte array its being on the whole file into memory during upload.
    Is it possible to solve this OOM by changing any setting of Weblogic HTPP handler without switching to Sun HTTP handler as there are many other application deployed on this weblogic instance?
    We are concerned whether there will be any impact on performance or any other issue.
    Please advice, highly appreciate your response.
    Thanks!

    Hi,
    If you have a back then restore the below file back and then try to start weblogic:
    \Oracle\Middleware\user_projects\domains\<domain_name>\config\config.lok
    Thanks,
    Sharmela

  • An error message uploading "large" files

    Hello!
    I'm getting an error message uploading "large" files (40-70 mb) - "unable to sync "file_name" due to server error". This message occurs only with a "large" files. The files up to 10 mb syncs perfectly and all of my files are correct (no issues with a "Windows-reserved / trailing" characters or "files in use").
    I pay 90 usd montly to Adobe and still need to use free services like Wetransfer or Dropbox. Please help.
    Thank you!
    Mindaugas

    I'm having this same problem and have written other posts but no replies. The file size for a single file is supposed to be something like 1-2 gig. How can a 40-70 meg file not work? It seems like the cloud storage feature is a selling point only and not fully realized yet. I love the interface and the idea but it's unrelaible to sync up, leave work, arrive home and no "larger" files are available. Yep, back to Dropbox and Google drive unfortunately.

  • Error: java.lang.OutOfMemoryError when uploading CSV files to web server

    Hi experts,
    I have made a JSP page from which clients load csv files to web server. I am using Tomca 4.1 as my web server and JDK 1.3.1_09.
    The system works fine when uploadiing small csv files, but it crashes when uploading large CSV files.
    It gives me the following error:
    java.lang.OutOfMemoryError
         <<no stack trace available>>
    This is the code that I used to load files....
    <%
    String saveFile = "";
    String contentType = request.getContentType();
    if ((contentType != null) && (contentType.indexOf("multipart/form-data") >= 0))
         DataInputStream in = new DataInputStream(request.getInputStream());
         int formDataLength = request.getContentLength();
         byte dataBytes[] = new byte[formDataLength];
         int byteRead = 0;
         int totalBytesRead = 0;
         while (totalBytesRead < formDataLength)
              byteRead = in.read(dataBytes, totalBytesRead, formDataLength);
              totalBytesRead += byteRead;
         String file = new String(dataBytes);
         saveFile = file.substring(file.indexOf("filename=\"") + 10);
         saveFile = saveFile.substring(0, saveFile.indexOf("\n"));
         saveFile = saveFile.substring(saveFile.lastIndexOf("\\") + 1,saveFile.indexOf("\""));
         int lastIndex = contentType.lastIndexOf("=");
         String boundary = contentType.substring(lastIndex + 1,contentType.length());
         int pos;
         pos = file.indexOf("filename=\"");
         pos = file.indexOf("\n", pos) + 1;
         pos = file.indexOf("\n", pos) + 1;
         pos = file.indexOf("\n", pos) + 1;
         int boundaryLocation = file.indexOf(boundary, pos) - 4;
         int startPos = ((file.substring(0, pos)).getBytes()).length;
         int endPos = ((file.substring(0, boundaryLocation)).getBytes()).length;
         String folder = "f:/Program Files/Apache Group/Tomcat 4.1/webapps/broadcast/file/";
         //String folder = "10.28.12.58/bulksms/";
         FileOutputStream fileOut = new FileOutputStream(folder + saveFile);
         //out.print("Saved here: " + saveFile);
         //fileOut.write(dataBytes);
         fileOut.write(dataBytes, startPos, (endPos - startPos));
         fileOut.flush();
         fileOut.close();
         out.println("File loaded successfully");
    //f:/Program Files/Apache Group/Tomcat 4.1/webapps/sms/file/
    %>
    Please can anyone help me solve this problem for me...
    Thanx...
    Deepak

    I know it may be hard to throw away all this code, but consider using the jakarta fileupload component.
    I think it would simplify your code down to
    // Create a factory for disk-based file items
    FileItemFactory factory = new DiskFileItemFactory();
    // Create a new file upload handler
    ServletFileUpload upload = new ServletFileUpload(factory);
    // Parse the request
    List /* FileItem */ items = upload.parseRequest(request);
    // Process the uploaded items
    Iterator iter = items.iterator();
    while (iter.hasNext()) {
        FileItem item = (FileItem) iter.next();
        if (item.isFormField()) {
            processFormField(item);
        } else {
            // item is a file.  write it
            File saveFolder = application.getRealPath("/file");          
            File uploadedFile = new File(saveFolder, item.getName());
            item.write(uploadedFile);
    }Most of this code was hijacked from http://jakarta.apache.org/commons/fileupload/using.html
    Check it out. It will solve your memory problem by writing the file to disk temporarily if necessary.
    Cheers,
    evnafets

  • Upload large file in sharepoint

    hi
    is it possible upload large file in sharepoint in our scenario we need upload large file >3gb in document libraries
    where sharepoint store files? is there any way we can have a custom aspx page for uploading large files in chunks in document library?
    kindly suggest.
    Thankyou

    is not possbile upload more than 2Gb
    If you try to upload big files you will need to make some change on your Sharepoint to support the 2Gb.
    Error message when you try to upload a large file to a document library on a Windows SharePoint Services 3.0 site: "Request timed out"
    http://support.microsoft.com/?id=925083
    With files bigger than 2GB can recomend creeate linked files to that files where you can download.
    André Lage Microsoft SharePoint and CRM Consultant
    Blog:http://aaclage.blogspot.com
    Codeplex:http://spupload.codeplex.com/http://simplecamlsearch.codeplex.com/

  • Upload large file on doSubmit in javascript

    Hi all,
    I'm currently having some problems uploading large files using a custom Upload button (button calls doSubmit from javascript).
    When using an Upload button with unconditional branch to the same page (no js) it works perfectly. Now I have to use my own Upload button that calls some javascript and AJAX and then Submit the page (with doSubmit); in this case I can only upload very small files (< 100k) ...
    I understand the mechanism of the file browse and its uploading to the table WWV_FLOW_FILE_OBJECTS, but I cannot figure out why my configuration does not work when it does stricly the same but from js ...
    Someone experienced the same issue ? How to submit the upload in javascript side with doSubmit ??
    Best regards,
    Othman

    Hi Othman,
    did you find a solution for this. I'm facing nearly the same problem, I'm calling a popup, where the user can upload up to 5 files. Then on clicking a button, javascript is called, that's calling the doSubmit and some others things and at the end the popup is closed.
    Sometimes the upload of 5 files is working, sometimes only 1 file is possible (obvisiously it's depending on the filesize).
    function addFile()
    doSubmit('FILE');
    opener.actualizeProject('TAB4');
    close();
    }chrissy

  • FTP Delivery error when uploading big file

    Hi,
    I've setup a Schedule with Delivery through SFTP. Usually (when the output size is not very big) it works fine.
    However, when my report (pdf) is too big, the job's status remains "Running" forever. When I go to my FTP directory the file's size is always 91.906.794 bytes. I checked BI Publisher's cache and the generated output is there but its size is always bigger than 91.906.794 bytes. That file works fine when I open it in Adobe Reader.
    Somehow, it seems that BI Publisher can't upload files bigger than 91.906.794 bytes but I can't find any maximum size in the documentation. And besides, I guess that if there was any maximum size, it should throw an error instead of getting stuck in "Running" status.
    I'm running Oracle BI Publisher version 11.1.1.6.0 in Red Hat Enterprise Linux Server release 5.5 (Tikanga). I've tested this feature in 2 different servers and the result is exactly the same.
    Anyone had this problem?
    Thank you,
    Tiago

    Hello,
    Here is the all possible fixes to upload large file, please check and verify with your settings:
    http://blogs.technet.com/b/praveenh/archive/2012/11/16/issues-with-uploading-large-documents-on-document-library-wss-3-0-amp-moss-2007.aspx
    http://blogs.technet.com/b/sammykailini/archive/2013/11/06/how-to-increase-the-maximum-upload-size-in-sharepoint-2013.aspx
    Hemendra:Yesterday is just a memory,Tomorrow we may never see
    Please remember to mark the replies as answers if they help and unmark them if they provide no help

  • Web Service to upload large files

    Hi Guys
    I'm want to create a Web Service that uploads large files (excess of 100MB). Can some help me with the best way I can do that, either C# or VB.NET.
    Thanks 

    I've done this before with a web service. It wasn't that hard. As jdweng alludes to, you need a web service and a client that knows how to repeatedly transfer the file sending it in chunks.
    The web service needs to take the file name and a byte array. It needs to return the number of bytes it received. The client calls the service endpoint passing in the name of the file and an array populated with x number of bytes. This should/could be tuned
    based on the size of the file and available bandwidth. 10 MB is a good upper limit but that may not be practical over a really slow connection... you could potentially time-out the call. The service endpoint receives the byte array and the file name. It creates/opens
    the file and appends the byte array onto the end in a synchronous manner. (You need to do it synchronous to avoid any potential race conditions with multiple calls overlapping.) When the server side is done, it returns the number of bytes written to the caller.
    The web service should close the file handle after each call since there is no guarantee that there will be another call.
    The caller then uses the return value to increment the offset pointer into the source file. When the call returns, the caller increments into the source file by the number of bytes returned then refills the buffer with the next chunk. It then re-calls the
    web service sending the same file name and the new byte array. The process continues until all of the bytes in the source file have been transferred to the server.
    I uploaded 4GB disk images to a Sharepoint server doing this so it works great and totally avoids the HTTP max payload problem.
    Good luck!

  • Uploading large files to share fails - advice appreciated

    I wondered if anyone could help me with uploading large files. 
    The ability to distribute files to clients with the included 20Gb of server space was part of the reason I signed up to Creative Cloud; unfortunatly I'm having very mixed results and I'm constantly having to fall back on the free DropBox service which seems much more reliable.
    I'm working on a MacPro - 10.6.8 with a optical fibre based ISP 20Mb/s service; the service is, dare I say, usually very quick and reliable.  Before uploading a large file I set energy saver to "never sleep" in case I wonder off for a cup of tea, but enevitably come back to what looks like a "stuck" upload.  I've tried three times today to send a 285Mb file and they all failed so I reverted to sending the file by DropBox... with no problem - I've also sent files in excess of 1Gb by DropBox in the past without an issue.
    Todays file had reached about 50% according to the little progress bar.. it then stopped.   (it's a shame the progrees bar can't relay the actual upload speed and a percentage like DropBox so you can see what's going on)
    I have to state, I'm not a DropBox "fanboy" but it just seems to work...
    Are their any tricks or settings that could make Adobe's 20Gb of space a functioning part of my workflow?  It would be great to feel confident in it.
    Thanks in advance for any advice.

    Either Firefox or Chrome would be a good alternative.
    I would see if one of these browsers works for upload but we can also debug the issue on Safari.
    Here are the steps to debug on Safari:
    Add the Develop menu to Safari. From the menu choose Safari > Preferences. Select the Advanced tab and check the box for "Show Develop menu in menu bar".
    Once logged onto the Creative Cloud Files page (https://creative.adobe.com/files) Ctrl + Click anywhere on the page and from the browser context menu choose Inspect Element.
    In the window that opens in the bottom half of the browser click on the Clock icon (mouse over info tip says Instrument) and on the left select Network Requests.
    Try the upload and you will see various network requests. Look for those whose Name is "notification". When you select one in the list there will be an arrow next to "notification" and clicking on it will take you to the details. See screen shot. If successful you will have a Status of 200. Some of these "notification" items are for your storage quota being updated, but the one we want is for the successful upload.

  • Code doesn't update when uploading new files to website.

    code doesn't update when uploading new files to website.   I was using an old DW version for a long time then forced to upgrade when I bought my MacBook Pro. Now when I make changes to my pages and link to new files etc on my harddrive, when i upload the pages it won't image because the code still points to harddrive (this did not used to be a problem)... I manually changed the code on the web host site to point to the site and it loaded fine. But now my files on my harddrive also point to the web host files and now don't image on my computer. I never had to worry about this so far and wonder if the new version has a setting I'm supposed to be aware of so that I can design on my hardrive and upload and it knows that the files are being uploaded with it?  Ugh hpe this makes sense - I'm NOT a webs designer and the new DW is way more than I use in it's technology.

    Hi
    This sounds like you have not defined your site in the site definition.
    Follow the steps in these 2 tutorials to set-up your site locally and for the remote server -
    http://www.adobe.com/devnet/dreamweaver/articles/first_website_pt1.html
    http://www.adobe.com/devnet/dreamweaver/articles/first_website_pt6.html
    PZ

Maybe you are looking for

  • SSRS Expression with OR

    Hi,  Can anyone tell me how to use OR in an SSRS expression? I would like to set the BorderStyle Left (& Right) property to "None" for two rows so I need to use an OR (6 or 9)   This is what I have so far: =IIf(RowNumber(nothing)=6,"None","Default") 

  • How to run Oracle workflow in windows 7 os

    Hi, Am new to oracle workflow. I have installed oracle workflow successfully in my laptop. After installation I dont see any shortcut's in my desktop. Can any one sugess me how to run oracle workflow. Steps to begin from basic level. Thanks Balaji

  • How a payment type can be customised using - pay_magtape_generic.new_formula

    Hi All, We are working on R12.1.3 for South Africa legislation. The electronic payment method used over here is ACB. In the payment type window (responsibility Global HRMS Manager) we notice a procedure "pay_magtape_generic.new_formula" attached with

  • Material transfer

    Gurus, I need to transfer from material from one plant to another.I am using mvmt type 303 & 305.What is the financial implication.Can i transfer thro same process for interco.cd/plant transfer also .

  • Renaming a existing org unit

    Hello All, My client wants to rename a existing org unit, so that the history should be available of the org unit. Scenario:- From dt. 01.01.2005 to 31.12.9999 the name of the org unit is ABC org unit. Now I want to rename the org unit as XYZ org uni