Size limit of 16TiB to sparsebundle files?

I manage several servers with higher disk volumes. Some use OsX.
At one particular customer, the OsX 10.7.5 fileserver has 21TB available and the OsX 10.8.4 backupserver has 25TB. Using TimeMachine we stumbled across an apparent size limit of 16TiB (17.59TB) for sparsebundle files. Whenever we try to expand the sparsebundle beyond that size, we receive error messages about exceeding `the´ size limit.
Which one? I found no information regarding such limitations. AFAIK, a sparsebundle is a virtual "Mac OS Extended" filesystem [partition in a virtual disk], which should be able to grow up to 8EiB since OsX 10.4 and the sparsebundle itself resides on a filesystem of that type also. So it's size limit should be 8EiB too. I see no reason, why a sparsebundle should have limits far less.
How can I create (resize) sparsebundles to, say, 20TB?

If there's a fixed limit to the size of an OS X disk image, it's not publicly documented anywhere that I know of. That doesn't mean there is no limit. Your customer may be the first person in history to have tried to create one that large. He should be mentioned in the Guinness Book of World Records.

Similar Messages

  • Is there a practical size limit on JPEG format image files?

    I have noticed when working with Photoshop v7.0 on Windows XP that, when creating a JPEG format image file from a very large scanned image (more than 800 Mb, maybe), the .jpg file will usually be saved on disk at any of the possible compression ratios available, without any reported error, yet when the saved file is re-opened, an error is reported - something along the lines of "missing or corrupted end of file marker" - and the image cannot be loaded.
    If before saving the original image is substantially reduced in size, e.g. by reducing the resolution of the image, a viable JPEG format file can be created, which will subsequently load successfully, though technically with data loss.
    This appears to happen only with JPEG format files - PDF format files can be saved from the original image at similar JPEG compression ratios, without error on reloading.
    I do not know if this also occurs with any other Photoshop versions.  I am working with Adobe Photoshop v7.0 on Windows XP, although we are preparing to upgrade to CS (which version I don't know) in the near future, for different reasons to those which are detailed above.
    Does anyone know if there is a technical explanation for this behaviour?  Is there a file size limit for writing viable JPEG files (which Photoshop does not report as an error when exceeded)?
    Thanks

    Chris Cox wrote:
    Where did you get the idea that we weren't doing something about it?
    that's what I gathered from "JPEG cannot [...] in Photoshop (code difficult to replace)".
    But if you say this is in the works right now, that's good news indeed.
    Chris Cox wrote:
    And where did you get the idea that posts get deleted? (other than SPAM)
    Are you sure you're reading the same forum as the rest of us?
    That's what happened:
    The content you posted below was reported for abuse and removed by our
    moderators:
    Subject: Re: Is there a practical size limit on JPEG format image files?
    Posted: 4/29/11 9:11 AM
    Anyway, you gave an honest answer on this technical issue, Chris, and that's what counts.

  • What is the size limit for a servlet .java file?

    I have been working on a jsp page with over 500 controls so I have getParameter() for those at least in my servlet. It is too big I think because the java file size is 40K but the class file is 25KB. Any sugguestion? Thanks,J

    Yes, there is a maximum length for the URL but if you use POST with a form the parameters do not appear in the URL... They would if you'd use GET however.
    Still, I guess it is just too much. And if that's a form a user would be supposed to fill out (usually they are supposed to do just that) they'd probably get tired and quit after the first 100 fields... 500 fields on one page is definately too much. I'd say try splitting it up to have about 10 fields on a page but then you'de have 50 pages to go through, which is less that good...
    May I ask what the 500 fields are for? What data is the form supposed to process? There might be a more practicable way to solve this.

  • LabView RT FTP file size limit

    I have created a few very large AVI video clips on my PXIe-8135RT (LabView RT 2014).  When i try to download these from the controller's drive to a host laptop (Windows 7) with FileZilla, the transfer stops at 1GB (The file size is actually 10GB).
    What's going on?  The file appears to be created correctly and I can even use AVI2 Open and AVI2 Get Info to see that the video file contains the frames I stored.  Reading up about LVRT, there is nothing but older information which claim the file size limit is 4GB, yet the file was created at 10GB using the AVI2 VIs.
    Thanks,
    Robert

    As usual, the answer was staring me right in the face.  FileZilla was reporting the size in an odd manner and the file was actually 1GB.  The vi I used was failing.  After fixing it, it failed at 2GB with error -1074395965 (AVI max file size reached).

  • Lite 10g DB File Size Limit

    Hello, everyone !
    I know, that Oracle Lite 5.x.x had a database file size limit = 4M per a db file. There is a statement in the Oracle® Database Lite 10g Release Notes, that db file size limit is 4Gb, but it is "... affected by the operating system. Maximum file size allowed by the operating system". Our company uses Oracle Lite on Windows XP operating system. XP allows file size more than 4Gb. So the question is - can the 10g Lite db file size exceed 4Gb limit ?
    Regards,
    Sergey Malykhin

    I don't know how Oracle Lite behave on PocketPC, 'cause we use it on Win32 platform. But ubder Windows, actually when .odb file reaches the max available size, the Lite database driver reports about I/O error ... after the next write operation (sorry, I just don't remember exact error message number).
    Sorry, I'm not sure what do you mean by "configure the situation" in this case ...

  • WCI 10.3 migration package size limit

    Hi,
    I'm trying to migrate content from old wci-server to new server (win2003 -->win2008) and some packages cannot be imported with portal interface.
    Is there a way to increase the size limit to allow bigger pte files than 20 MB?
    I'm aware of the command line tool, but there seems to be no way to exclude unwanted content.
    Regards,
    Johanna

    Hi,
    I'm trying to import certain user group with dependencies and the package is about 30 MB.
    This is One group containing about 4 other groups and dynamic members.
    When trying to import to a new environment nothing happens and after a while the error page is displayed.
    Logging spy gives 2 errors:
    Unable to load request data.
    com.plumtree.openfoundation.io.XPIOException: Posted content length of 29289005 exceeds limit of 20971520
        at com.plumtree.openfoundation.util.XPException.GetInstance(XPException.java:385)
        at com.plumtree.openfoundation.util.XPException.GetInstance(XPException.java:350)
        at com.plumtree.openfoundation.web.XPRequest.ParseRequest(XPRequest.java:935)
        at com.plumtree.uiinfrastructure.interpreter.Interpreter.HandleRequest(Interpreter.java:264)
        at com.plumtree.uiinfrastructure.interpreter.Interpreter.DoService(Interpreter.java:190)
        at com.plumtree.uiinfrastructure.web.XPPage.service(XPPage.java:300)
        at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)
        at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:227)
        at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:125)
        at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:292)
        at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:175)
        at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(Unknown Source)
        at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
        at weblogic.security.service.SecurityManager.runAs(Unknown Source)
        at weblogic.servlet.internal.WebAppServletContext.securedExecute(Unknown Source)
        at weblogic.servlet.internal.WebAppServletContext.execute(Unknown Source)
        at weblogic.servlet.internal.ServletRequestImpl.run(Unknown Source)
        at weblogic.work.ExecuteThread.execute(ExecuteThread.java:201)
        at weblogic.work.ExecuteThread.run(ExecuteThread.java:173)
    Caused by: java.io.IOException: Posted content length of 29289005 exceeds limit of 20971520
        at com.oreilly.servlet.multipart.MultipartParser.<init>(MultipartParser.java:172)
        at com.plumtree.openfoundation.web.XPRequest.ParseRequest(XPRequest.java:782)
        at com.plumtree.uiinfrastructure.interpreter.Interpreter.HandleRequest(Interpreter.java:264)
        at com.plumtree.uiinfrastructure.interpreter.Interpreter.DoService(Interpreter.java:191)
        at com.plumtree.uiinfrastructure.web.XPPage.service(XPPage.java:300)
        at javax.servlet.http.HttpServlet.service(HttpServlet.java:821)
        at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:227)
        at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:125)
        at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:292)
        at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:176)
        ... 8 more
    Server error.
    com.plumtree.openfoundation.io.XPIOException: Posted content length of 29289005 exceeds limit of 20971520
        at com.plumtree.openfoundation.util.XPException.GetInstance(XPException.java:385)
        at com.plumtree.openfoundation.util.XPException.GetInstance(XPException.java:350)
        at com.plumtree.openfoundation.web.XPRequest.ParseRequest(XPRequest.java:935)
        at com.plumtree.uiinfrastructure.interpreter.Interpreter.HandleRequest(Interpreter.java:264)
        at com.plumtree.uiinfrastructure.interpreter.Interpreter.DoService(Interpreter.java:190)
        at com.plumtree.uiinfrastructure.web.XPPage.service(XPPage.java:300)
        at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)
        at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:227)
        at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:125)
        at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:292)
        at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:175)
        at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(Unknown Source)
        at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
        at weblogic.security.service.SecurityManager.runAs(Unknown Source)
        at weblogic.servlet.internal.WebAppServletContext.securedExecute(Unknown Source)
        at weblogic.servlet.internal.WebAppServletContext.execute(Unknown Source)
        at weblogic.servlet.internal.ServletRequestImpl.run(Unknown Source)
        at weblogic.work.ExecuteThread.execute(ExecuteThread.java:201)
        at weblogic.work.ExecuteThread.run(ExecuteThread.java:173)
    Caused by: java.io.IOException: Posted content length of 29289005 exceeds limit of 20971520
        at com.oreilly.servlet.multipart.MultipartParser.<init>(MultipartParser.java:172)
        at com.plumtree.openfoundation.web.XPRequest.ParseRequest(XPRequest.java:782)
        at com.plumtree.uiinfrastructure.interpreter.Interpreter.HandleRequest(Interpreter.java:264)
        at com.plumtree.uiinfrastructure.interpreter.Interpreter.DoService(Interpreter.java:191)
        at com.plumtree.uiinfrastructure.web.XPPage.service(XPPage.java:300)
        at javax.servlet.http.HttpServlet.service(HttpServlet.java:821)
        at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:227)
        at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:125)
        at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:292)
        at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:176)
        ... 8 more
    Google gives a lot of results but nothing for this product for it's old stuff
    Regards,
    Johanna

  • HT4863 How can I increase the file size limit for outgoing mail. I need to send a file that is 50MB?

    How can I increase the file size limit for outgoing mail. I need to send a file that is 50MB?

    You can't change it, and I suspect few email providers would allow a file that big.  Consider uploading it to a service like Dropbox, then email the link allowing the recipient to download it.

  • FILE and FTP Adapter file size limit

    Hi,
    Oracle SOA Suite ESB related:
    I see that there is a file size limit of 7MB for transferring using File and FTP adapter and that debatching can be used to overcome this issue. Also see that debatching can be done only for strucutred files.
    1) What can be done to transfer unstructured files larger than 7MB from one server to the other using FTP adapter?
    2) For structured files, could someone help me in debatching a file with the following structure.
    000|SEC-US-MF|1234|POPOC|679
    100|PO_226312|1234|7130667
    200|PO_226312|1234|Line_id_1
    300|Line_id_1|1234|Location_ID_1
    400|Location_ID_1|1234|Dist_ID_1
    100|PO_226355|1234|7136890
    200|PO_226355|1234|Line_id_2
    300|Line_id_2|1234|Location_ID_2
    400|Location_ID_2|1234|Dist_ID_2
    100|PO_226355|1234|7136890
    200|PO_226355|1234|Line_id_N
    300|Line_id_N|1234|Location_ID_N
    400|Location_ID_N|1234|Dist_ID_N
    999|SSS|1234|88|158
    I would need a the complete data in a single file at the destination for each file in the source. If there are as many number of files as the number of batches at the destination, I would need the file output file structure be as follows:
    000|SEC-US-MF|1234|POPOC|679
    100|PO_226312|1234|7130667
    200|PO_226312|1234|Line_id_1
    300|Line_id_1|1234|Location_ID_1
    400|Location_ID_1|1234|Dist_ID_1
    999|SSS|1234|88|158
    Thanks in advance,
    RV
    Edited by: user10236075 on May 25, 2009 4:12 PM
    Edited by: user10236075 on May 25, 2009 4:14 PM

    Ok Here are the steps
    1. Create an inbound file adapter as you normally would. The schema is opaque, set the polling as required.
    2. Create an outbound file adapter as you normally would, it doesn't really matter what xsd you use as you will modify the wsdl manually.
    3. Create a xsd that will read your file. This would typically be the xsd you would use for the inbound adapter. I call this address-csv.xsd.
    4. Create a xsd that is the desired output. This would typically be the xsd you would use for the outbound adapter. I have called this address-fixed-length.xsd. So I want to map csv to fixed length format.
    5. Create the xslt that will map between the 2 xsd. Do this in JDev, select the BPEL project, right-click -> New -> General -> XSL Map
    6. Edit the outbound file partner link wsdl, the the jca operations as the doc specifies, this is my example.
    <jca:binding  />
            <operation name="MoveWithXlate">
          <jca:operation
              InteractionSpec="oracle.tip.adapter.file.outbound.FileIoInteractionSpec"
              SourcePhysicalDirectory="foo1"
              SourceFileName="bar1"
              TargetPhysicalDirectory="C:\JDevOOW\jdev\FileIoOperationApps\MoveHugeFileWithXlate\out"
              TargetFileName="purchase_fixed.txt"
              SourceSchema="address-csv.xsd" 
              SourceSchemaRoot ="Root-Element"
              SourceType="native"
              TargetSchema="address-fixedLength.xsd" 
              TargetSchemaRoot ="Root-Element"
              TargetType="native"
              Xsl="addr1Toaddr2.xsl"
              Type="MOVE">
          </jca:operation> 7. Edit the outbound header to look as follows
        <types>
            <schema attributeFormDefault="qualified" elementFormDefault="qualified"
                    targetNamespace="http://xmlns.oracle.com/pcbpel/adapter/file/"
                    xmlns="http://www.w3.org/2001/XMLSchema"
                    xmlns:FILEAPP="http://xmlns.oracle.com/pcbpel/adapter/file/">
                <element name="OutboundFileHeaderType">
                    <complexType>
                        <sequence>
                            <element name="fileName" type="string"/>
                            <element name="sourceDirectory" type="string"/>
                            <element name="sourceFileName" type="string"/>
                            <element name="targetDirectory" type="string"/>
                            <element name="targetFileName" type="string"/>                       
                        </sequence>
                    </complexType>
                </element> 
            </schema>
        </types>   8. the last trick is to have an assign between the inbound header to the outbound header partner link that copies the headers. You only need to copy the sourceDirectory and SourceGileName
        <assign name="Assign_Headers">
          <copy>
            <from variable="inboundHeader" part="inboundHeader"
                  query="/ns2:InboundFileHeaderType/ns2:fileName"/>
            <to variable="outboundHeader" part="outboundHeader"
                query="/ns2:OutboundFileHeaderType/ns2:sourceFileName"/>
          </copy>
          <copy>
            <from variable="inboundHeader" part="inboundHeader"
                  query="/ns2:InboundFileHeaderType/ns2:directory"/>
            <to variable="outboundHeader" part="outboundHeader"
                query="/ns2:OutboundFileHeaderType/ns2:sourceDirectory"/>
          </copy>
        </assign>you should be good to go. If you just want pass through then you don't need the native format set to opaque, with no XSLT
    cheers
    James

  • S1000 Data file size limit is reached in statement

    I am new to Java and was given the task to trouble shoot a java application that was written a few years ago and no longer supported. The java application creates database files the user's directory: diwdb.properties, diwdb.data, diwdb.lproperties, diwdb.script. The purpose of the application is to open a zip file and insert the files into a table in the database.
    The values that are populated in the diwdb.properties file are as follows:
    #HSQL Database Engine
    #Wed Jan 30 08:55:05 GMT 2013
    hsqldb.script_format=0
    runtime.gc_interval=0
    sql.enforce_strict_size=false
    hsqldb.cache_size_scale=8
    readonly=false
    hsqldb.nio_data_file=true
    hsqldb.cache_scale=14
    version=1.8.0
    hsqldb.default_table_type=memory
    hsqldb.cache_file_scale=1
    hsqldb.log_size=200
    modified=yes
    hsqldb.cache_version=1.7.0
    hsqldb.original_version=1.8.0
    hsqldb.compatible_version=1.8.0
    Once the databsae file gets to 2GB it brings up the error meessage 'S1000 Data file size limit is reached in statement (Insert into <tablename>......
    From searching on the itnernet it appeared that the parameter hsqldb.cache_file_scale needed to be increased & 8 was a suggested value.
    I have the distribution files (.jar & .jnlp) that are used to run the application. And I have a source directory that was found that contains java files. But I do not see any properties files to set any parameters. I was able to load both directories into NetBeans but really don't know if the files can be rebuilt for distribution as I'm not clear on what I'm doing and NetBeans shows errors in some of the directories.
    I have also tried to add parameters to the startup url: http://uknt117.uk.infores.com/DIW/DIW.jnlp?hsqldb.large_data=true?hsqldb.cache_file_scale=8 but that does not affect the application.
    I have been struggling with this for quite some time. Would greatly appreciate any assistance to help resolve this.
    Thanks!

    Thanks! But where would I run the sql statement. When anyone launches the application it creates the database files in their user directory. How would I connect to the database after that to execute the statement?
    I see the create table statements in the files I have pulled into NetBeans in both the source folder and the distribution folder. Could I add the statement there before the table is created in the jar file in the distribution folder and then re-compile it for distribution? OR would I need to add it to the file in source directory and recompile those to create a new distribution?
    Thanks!

  • Size limit of access database with image files stored outside database

    I have read 2GB limit on size of Access database.  We are using Access 2013 database with fields that are data type attachment and storing attached images outside the database.  Does the size of the folder that is holding these images count
    in any way toward 2GB size limit?

    Hi,
    As far as I know, Access database limit size 2GB based on itself, not related to the folder size.  Attachment data
    type is similar to attaching files to e-mail messages. We need to control the Access database size is smaller then 2GB.
    When a field is defined with the attachment data type, you can store one or more files for each record in it.
    Attachments can dramatically increase the size of the database, but since the attached file is stored as a part of the dabatbase, you are not dependent on network drives being available as you would be if you inculded a hyperlink to the file. As a matter of
    fact, feel free to backup the origainal file to other disk after you attach it to the database.
    Regards,
    George Zhao
    TechNet Community Support

  • Size limit for flar file with Sol 10 WANBOOT

    There seems to be some kind of a size limit for flash archive files used by wanboot. When I try to wanboot jumpstart, I get the following:
    Processing profile
    - Opening Flash archive
    ERROR: HTTP server returned an invalid archive file size: <-1256985017> bytes
    ERROR: Invalid HTTP headers were returned from the server
    ERROR: Flash installation failed
    Solaris installation program exited.
    The flash archive is about 3GB. I'm doing the WANBOOT using the "boot cdrom" work-around, because the V100 I'm jumpstarting doesn't support the newer OBP parameters.
    If create a new smaller flash archive like so:
    mv sol10.flar sol10.flar.old
    dd if=sol10.flar.old of=sol10.flar bs=1024K count=1024
    cutting the flar down to 1GB. Then I do the wanboot. I don't get the error, and the jumpstart proceeds. Of course it would eventually have a problem when the cpio file reached premature end.
    So is there a prescribed limit to flars with WANBOOT. I'm assuming it's 2GB, but if so, why?
    No I haven't tried creating a compressed flar: that's next.
    Thanks,
    Chip Bennett
    Laurus Technologies, Inc.
    Itasca, IL

    nope, 64-bit apache2.2 and still the same error:
    Processing profile
    - Opening Flash archive
    HTTP server returned an invalid archive file size:
    <-2116410290> bytes
    ERROR: Invalid HTTP headers were returned from the
    server
    ERROR: Flash installation failed
    Solaris installation program exited.
    apache log:
    192.168.224.112 - - [06/Jul/2006:20:52:45 +0200]
    "HEAD /flash/archives/pt_sol_10_oem.flar HTTP/1.1"
    200 -
    192.168.224.112 - - [06/Jul/2006:20:52:45 +0200] "GET
    /cgi-bin/bootlog-cgi?%3Ctime%3E+wanboot_client+ident:+
    %5BID+380565+user.panic%5D+HTTP+server+returned+an+inv
    alid+archive+file+size:+%3C-2116410290%3E+bytes
    HTTP/1.1" 200 32
    192.168.224.112 - - [06/Jul/2006:20:52:45 +0200] "GET
    /cgi-bin/bootlog-cgi?%3Ctime%3E+wanboot_client+ident:+
    %5BID+592631+user.panic%5D+Invalid+HTTP+headers+were+r
    eturned+from+the+server HTTP/1.1" 200 32
    192.168.224.112 - - [06/Jul/2006:20:52:45 +0200] "GET
    /cgi-bin/bootlog-cgi?%3Ctime%3E+wanboot_client+ident:+
    %5BID+915281+user.panic%5D+Flash+installation+failed
    HTTP/1.1" 200 32
    Any help would be much appreciated.hi. i spent a long time on this issue...and got it working reliably for more than 10 gb flars.
    there are two issues- 1) apache must be 64 bit capable, and 2) you need the latest wanboot miniroot, for solaris 9 and one for solaris 10. i worked for a different division of a major car company based in stuttgart (take a guess) than i do now, and we eventually received the support we needed from sun, through the early release program. i do not know any details about if, how, or when those miniroots are otherwise available. at some point, i can only assume they will be generally available to all, if not already.
    if someone knows more about this availability issue, i would also be interested in this issue.
    thanks, dls

  • 4GB File Size Limit in Finder for Windows/Samba Shares?

    I am unable to copy a 4.75GB video file from my Mac Pro to a network drive with XFS file system using the Finder. Files under 4GB can be dragged and dropped without problems. The drag and drop method produces an "unexepcted error" (code 0) message.
    I went into Terminal and used the cp command to successfully copy the 4.75GB file to the NAS drive, so obviously there's a 4GB file size limit that Finder is opposing?
    I was also able to use Quicktime and save a copy of the file to the network drive, so applications have no problem, either.
    XFS file system supports terabyte size files, so this shouldn't be a problem on the receiving end, and it's not, as the terminal copy worked.
    Why would they do that? Is there a setting I can use to override this? Google searching found some flags to use with the mount command in Linux terminal to work around this, but I'd rather just be able to use the GUI in OS X (10.5.1) - I mean, that's why we like Macs, right?

    I have frequently worked with 8 to 10 gigabyte capture files in both OS9 and OS X so any limit does not seem to be in QT or in the Player. 2 GIg limits would perhaps be something left over from pre-OS 9 versions of your software, as there was a general 2 gig limit in those earlier versions of the operating system. I have also seen people refer to 2 gig limits in QT for Windows but never in OS 9 or later MacOS.

  • Is there a Maximum file size limit when combining pdf files

    I am trying to combine files 8 PDF files created with i Explore using Acrobat XI Standard, the process is completed but when opening the file not all 8 individual files have been combined only the first two have been combined, the process shows all eight being read and combined.
    I can open and view each of the original 8 individual files and read the content without any issue, the largest of the eight files is 15,559kb and the smallest is 9,435kb
    if anyone can assist that would be most appreciated

    Hi Tayls450,
    This should not have happened as there is no maximum file size limit when combining PDFs.
    Please let me know if there are some password protected PDFs also?
    Also, I would suggest you to choose 'Repair Acrobat Installation' option from the Help menu.
    Regards,
    Anubha

  • Size limit for how much video can be included in an AIR file?

    I'm making an AIR app to play a video locally.  It is on a PC with no internet connection, so the video (fv4) is  included with the air install file.  Is there a size limit for how much  video can be included in an AIR file? I seem to get the "...file is  damaged..." error when I include a video over 200mb. I use Flash CS4 to  create and AIR is v 1.53.

    I'm authoring in CS4 on a MAC with the latest OS and 4GB ram.  It plays fine on the Mac.  On a WIN7 PC with all updates and 4GB ram I get the errors if the air file that contains a video file over 200mb.  I heard that AIR 2 should solve the issue and that it is just a PC issue.. Maybe...   I have not tried using ADT to package the file.

  • Folio file size limit

    In the past their were file size limits for folios and folio resources. I believe folios and their resources could not go over 100MB.
    Is their a file size limit to a folio and its resources? What is the limit?
    What happens if the limit is exceeded?
    Dave

    Article size is currently capped at 2GB. We've seen individual articles of up to 750MB in size but as mentioned previously you do start bumping into timeout issues uploading content of that size. We're currently looking at implementing a maximum article size based on the practical limitations of upload and download times and timeout values being hit.
    This check will be applied at article creation time and you will not be allowed to create articles larger than our maximum size. Please note that we are working on architecture changes that will allow the creation of larger content but that will take some time to implement.
    Currently to get the best performance you should have all your content on your local machine (Not on a network drive), shut down any extra applications, get the fastest machine you can, with the most memory. The size of article is direclty proportional to the speed at which you can bundle an article and get it uploaded to the server.

Maybe you are looking for

  • Is it possible to re-acticvate a 3G iPhone through itunes?

    I recently purchased a 3G, and shortly there after, I ported a number and changed the original number, however, both my phone and iTunes show the old number, I am guessing that this is the number that's on my sim card. Is there a way to re-activate t

  • Lots of ipod problems

    Hi this is my first post...i recently got a new 30GB Black iPod. A few days after i got it i started having all kinds of problems. I was trying to watch a movie and then none of the buttons would work. I reset my iPod and then got the sad faced icon.

  • Slow iPod sync with network drive

    I run a small network for my home/office (I am self employed) with about 4 PCs for me and my family. We have 3 iPods in our family and we just use one PC to host iTunes, but because we have so much music, I have put the music on a network drive serve

  • How to create alias and folders

    I have a MM as headless server. It has to users both with admin rights. I am working in the folder Library/application support/X It refuses to let me create more folders in folder X or aliases. Anybody knows how to get this changed so I can do that?

  • Method Call when click a link

    Hi All I have jspx page which have several links to go to different pages. How do i call a java method when I click a link Thank You Padma Kumara