Auto save and very large files

Hi  All,
I produce very high dim artworks, 6m x 2.7m(+) image at 200ppi, sometimes when I load the files Photoshop cc2014 would randomly will crash. This never happened with my CS5 version. I have auto save checked, could this be the issue?
I am running a high end computer with 32gb ram, 8 ssd's with two dedicated to PS scratch disk....the images, so far, load eventually but the heart beat stop moment when you think a file is corrupt really scares me, especially when $$$$$ are involved.
Any ideas?

Thanks JJMack
I will try it. It is possible it's not the auto-save feature, cs5 was rock solid and gave me confidence as it never crashed even with files over 5gb. The crash report states the faulting model is unknown for the last crash but I have seen ntdll.dll as the faulting module in previous crashes.

Similar Messages

  • Very odd and very large file found in system library: maremoted

    Hello,
    Using a single 1.25 G4 with 10.3.9. It's been acting up and running out of space.
    We located a file called maremoted, (which we're guessing is some kind of Remote Daemon), fil, which is upwards of 30 GB. It's run by the root user and trying to trash it causes a Kernel Panic. We took out the drive and sey it as a slave to mine and I deleted it.
    The machine started up fine, but it's come back and growing. It's currently 1 GB.
    Anyone have any idea what this thing is and how to get rid of it?
    Thank you for any suggestions.
    -Vincent

    This is a daemon that is running as part of the Track-It! auditing software by Numara. I have it crashing on several Macs and respawning with a new PID which makes it impossible to kill the process. The only file that I have found growing in size is the log file found in /Library/Logs/Track-It!/Mac Audit. Both the maremoted.log and the maauditord.log are growing in size from the crashdumps. Possibly due to the server and the clients being out of sync version wise?
    The symptom is as you describe; sluggishness and unresponsiveness while performing various tasks such as Application switching, network file transfer, and typing.
    Placed a call into Numara and the two Mac experts are not in to assist. Will be calling first thing tomorrow.

  • Tips or tools for handling very large file uploads and downloads?

    I am working on a site that has a document repository feature. The documents are stored as BLOBs in an Oracle database and for reasonably sized files its not problem to stream the files out directly from the database. For file uploads, I am using the Struts module to get them on disk and am then putting the blob in the database.
    We are now being asked to support very large files of 250MB+. I am concerned about problems I've heard of with HTTP not being reliable for files over 256MB. I'd also like a solution that would give the user a status bar and allow for restarts of broken uploads or downloads.
    Does anyone know of an off-the-shelf module that might help in this regard? I suspect an ActiveX control or Applet on the client side would be necessary. Freeware or Commercial software would be ok.
    Thanks in advance for any help/ideas.

    Hi. There is nothing wrong with HTTP handling 250MB+ files (per se).
    However, connections can get reset.
    Consider offering the files via FTP. Most FTP clients are good about resuming transfers.
    Or if you want to keep using HTTP, try supporting chunked encoding. Then a user can use something like 'GetRight' to auto resume HTTP downloads.
    Hope that helps,
    Peter
    http://rimuhosting.com - JBoss EJB/JSP hosting specialists

  • Would a 347 MB file be slow? Auto save and auto type turn off?

    I am doing an art inventory and add jpegs of the art. I don't resize them they range from 750k to 3MB. I guess if it is slowing the program down i should? It currently is 267 rows by about 18 columns.
    Thanks in advance.
    Oh can I turn off auto save and auto type? That may help. How do I do that?

    Done.
    Enter Scrip Editor
    paste the posted script
    File > Save > as Script  on the Desktop
    Move the script to the folder :
    Macintosh HD:Library:Scripts:Folder Action Scripts:
    CAUTION, you will be asked to enter your pasword.
    Create a folder to do the job. I named mine Normalized. I created mine on the Desktop where it's easy to reach.
    Go to :
    Macintosh HD:Library:Scripts:Folder Actions:
    Double click the alias :  Configure Folder Actions
    Below the left column, click
    navigate to select your new folder
    Below the right column, click
    select the script image - normalize400.scpt
    After that, drag and drop a picture file onto your folder.
    The original will be move in the folder Originals (isn’t it original ?)
    and a reduced copy  400 x height will be stored in the folder Normalized images.
    And now, here is the script :
    --{code}
    Image - Normalize
    This Folder Action handler is triggered whenever items are added to the attached folder.
    The script rotates the image counter-clockwise (left).
    Copyright © 2002–2007 Apple Inc.
    You may incorporate this Apple sample code into your program(s) without
    restriction.  This Apple sample code has been provided "AS IS" and the
    responsibility for its operation is yours.  You are not permitted to
    redistribute this Apple sample code as "Apple sample code" after having
    made changes.  If you're going to redistribute the code, we require
    that you make it clear that the code was descended from Apple sample
    code, but that you've made changes.
    modified by Yvan KOENIG (VALLAURIS, France)
    2011/12/08
    This version normalize pictures so that
    (1) the greater dimension become the width one (rotate left if needed)
    (2) this greater dimension is ruled by the property maxWidth defined below.
    property maxWidth : 400
    -- set it to fit your needs
    property done_foldername : "Normalized Images"
    property originals_foldername : "Original Images"
    property newimage_extension : "jpg"
    -- the list of file types which will be processed
    -- eg: {"PICT", "JPEG", "TIFF", "GIFf"}
    property type_list : {"TIFF", "GIFf", "PNGf", "PICT", "JPEG"}
    -- since file types are optional in Mac OS X,
    -- check the name extension if there is no file type
    -- NOTE: do not use periods (.) with the items in the name extensions list
    -- eg: {"txt", "text", "jpg", "jpeg"}, NOT: {".txt", ".text", ".jpg", ".jpeg"}
    property extension_list : {"tif", "tiff", "gif", "png", "pict", "pct", "jpeg", "jpg"}
    on adding folder items to this_folder after receiving these_items
              tell application "Finder"
                        if not (exists folder done_foldername of this_folder) then
      make new folder at this_folder with properties {name:done_foldername}
                        end if
                        set the results_folder to (folder done_foldername of this_folder) as alias
                        if not (exists folder originals_foldername of this_folder) then
      make new folder at this_folder with properties {name:originals_foldername}
                                  set current view of container window of this_folder to list view
                        end if
                        set the originals_folder to folder originals_foldername of this_folder
              end tell
              try
                        repeat with i from 1 to number of items in these_items
                                  set this_item to item i of these_items
                                  set the item_info to the info for this_item
                                  if (alias of the item_info is false and the file type of the item_info is in the type_list) or (the name extension of the item_info is in the extension_list) then
                                            tell application "Finder"
      --set name of this_item to "YK#" & (text -4 thru -1 of ("0000" & i))
                                                      my resolve_conflicts(this_item, originals_folder, "")
                                                      set the new_name to my resolve_conflicts(this_item, results_folder, newimage_extension)
                                                      set the source_file to (move this_item to the originals_folder with replacing) as alias
                                            end tell
      process_item(source_file, new_name, results_folder)
                                  end if
                        end repeat
              on error error_message number error_number
                        if the error_number is not -128 then
                                  tell application "Finder"
      activate
      display dialog error_message buttons {"Cancel"} default button 1 giving up after 120
                                  end tell
                        end if
              end try
    end adding folder items to
    on resolve_conflicts(this_item, target_folder, new_extension)
              tell application "Finder"
                        set the file_name to the name of this_item
                        set file_extension to the name extension of this_item
                        if the file_extension is "" then
                                  set the trimmed_name to the file_name
                        else
                                  set the trimmed_name to text 1 thru -((length of file_extension) + 2) of the file_name
                        end if
                        if the new_extension is "" then
                                  set target_name to file_name
                                  set target_extension to file_extension
                        else
                                  set target_extension to new_extension
                                  set target_name to (the trimmed_name & "." & target_extension) as string
                        end if
                        if (exists document file target_name of target_folder) then
                                  set the name_increment to 1
                                  repeat
                                            set the new_name to (the trimmed_name & "." & (name_increment as string) & "." & target_extension) as string
                                            if not (exists document file new_name of the target_folder) then
      -- rename to conflicting file
                                                      set the name of document file target_name of the target_folder to the new_name
                                                      exit repeat
                                            else
                                                      set the name_increment to the name_increment + 1
                                            end if
                                  end repeat
                        end if
              end tell
              return the target_name
    end resolve_conflicts
    -- this sub-routine processes files
    on process_item(source_file, new_name, results_folder)
      -- NOTE that the variable this_item is a file reference in alias format
      -- FILE PROCESSING STATEMENTS GOES HERE
              try
      -- the target path is the destination folder and the new file name
                        set the target_path to ((results_folder as string) & new_name) as string
                        with timeout of 900 seconds
                                  tell application "Image Events"
      launch -- always use with Folder Actions
                                            set this_image to open file (source_file as string)
                                            set {oldW, oldH} to dimensions of this_image
                                            if oldH > oldW then
                                                      set {oldW, oldH} to {oldH, oldW}
      rotate this_image to angle 270.0
                                            end if
                                            set |échelle| to maxWidth / oldW
                                            if |échelle| < 1 then scale this_image by factor |échelle|
      save this_image as JPEG in file target_path with icon
      close this_image
                                  end tell
                        end timeout
              on error error_message
                        tell application "Finder"
      activate
      display dialog error_message buttons {"Cancel"} default button 1 giving up after 120
                        end tell
              end try
    end process_item
    --{code}
    Yvan KOENIG (VALLAURIS, France)  jeudi 8 décembre 2011 21:25:33
    iMac 21”5, i7, 2.8 GHz, 12 Gbytes, 1 Tbytes, mac OS X 10.6.8 and 10.7.2
    My iDisk is : <http://public.me.com/koenigyvan>
    Please : Search for questions similar to your own before submitting them to the community

  • Hello, I am having issues open very large files I created, one being 1,241,776 KB. I have PS 12.1 with 64 bit version. I am not sure if they issues I am having is because of the PS version I have, and whether or not I have to upgrade?

    Hello, I am having issues open very large files I created, one being 1,241,776 KB. I have PS 12.1 with 64 bit version. I am not sure if they issues I am having is because of the PS version I have, and whether or not I have to upgrade?

    I think more likely, it's a memory / scratch disk issue.  1.25 gigabytes is a very big image file!!
    Nancy O.

  • Upload very large file to SharePoint Online

    Hi,
    I tried uploading very large files to SharePoint online via the REST API using JAVA. Uploading files with a size of ~160 mb works well, but if the filesize is between 400mb and 1,6 gb I either receive a Socket Exception (Connection reset by peer) or an error
    of SharePoint telling me "the security validation for the page is invalid". 
    So first of all I want to ask you, how to work around the security validation error. As far as I understand the token, which I have added to my X-RequestDigest header of the http post request seems to be expired (by default it expires after 1800 seconds).
    Uploading such large files is time consuming, so I can't change the token in my header while uploading the file. How can I extend the expiration time of the token (if it is possible)? Could it help to send post requests with the same token
    continuously while uploading the file and therefore prevent the token from expiration? Is there any other strategy to upload such large files which need much more than 1800 seconds for upload?
    Additionally any thoughts on the socket exception? It happens quite sporadically. So sometimes it happens after uploading 100 mb, and the other time I have already uploaded 1 gb.
    Thanks in advance!

    Hi,
    Thanks for the reply, the reason I'm looking into this is so users can migrate there files to SharePoint online/Office 365. The max file limit is 2gb, so thought I would try and cover this.
    I've looked into that link before, and when I try and use those end points ie StartUpload I get the following response
    {"error":{"code":"-1, System.NotImplementedException","message":{"lang":"en-US","value":"The
    method or operation is not implemented."}}}
    I can only presume is that these endpoints have not been implemented yet, even though the documentation says they only available for Office 365 which is what  I am using. 
    Also, the other strange thing is that I can actually upload a file of 512mb to the server as I can see it in the web UI. The problem I am experiencing is get a response, it hangs on this line until the TimeOut is reached.
    WebResponse wresp = wreq.GetResponse();
    I've tried flushing and closing the stream, sendchunked but still no success

  • Unable to copy very large file to eSATA external HDD

    I am trying to copy a VMWare Fusion virtual machine, 57 GB, from my Macbook Pro's laptop hard drive to an external, eSATA hard drive, which is attached through an ExpressPort adapter. VMWare Fusion is not running and the external drive has lots of room. The disk utility finds no problems with either drive. I have excluded both the external disk and the folder on my laptop hard drive that contains my virtual machine from my Time Machihne backups. At about the 42 GB mark, an error message appears:
    The Finder cannot complete the operation because some data in "Windows1-Snapshot6.vmem" could not be read or written. (Error code -36)
    After I press OK to remove the dialog, the copy does not continue, and I cannot cancel the copy. I have to force-quit the Finder to make the copy dialog go away before I can attempt the copy again. I've tried rebooting between attempts, still no luck. I have tried a total of 4 times now, exact same result at the exact same place, 42 GB / 57 GB.
    Any ideas?

    Still no breakthrough from Apple. They're telling me to terminate the VMWare processes before attempting the copy, but had they actually read my description of the problem first, they would have known that I already tried this. Hopefully they'll continue to investigate.
    From a correspondence with Tim, a support representative at Apple:
    Hi Tim,
    Thank you for getting back to me, I got your message. Although it is true that at the time I ran the Capture Data program there were some VMWare-related processes running (PID's 105, 106, 107 and 108), this was not the case when the issue occurred earlier. After initially experiencing the problem, this possibility had occurred to me so I took the time to terminate all VMWare processes using the activity monitor before again attempting to copy the files, including the processes mentioned by your engineering department. I documented this in my posting to apple's forum as follows: (quote is from my post of Feb 19, 2008, 1:28pm, to the thread "Unable to copy very large file to eSATA external HDD", relevant section in >bold print<)
    Thanks for the suggestions. I have since tried this operation with 3 different drives through two different interface types. Two of the drives are identical - 3.5" 7200 RPM 1TB Western Digital WD10EACS (WD Caviar SE16) in external hard drive enclosures, and the other is a smaller USB2 100GB Western Digital WD1200U0170-001 external drive. I tried the two 1TB drives through eSATA - ExpressPort and also over USB2. I have tried the 100GB drive only over USB2 since that is the only interface on the drive. In all cases the result is the same. All 3 drives are formatted Mac OS Extended (Journaled).
    I know the files work on my laptop's hard drive. They are a VMWare virtual machine that works just fine when I use it every day. >Before attempting the copy, I shut down VMWare and terminated all VMWare processes using the Activity Monitor for good measure.< I have tried the copy operation both through the finder and through the Unix command prompt using the drive's mount point of /Volumes/jfinney-ext-3.
    Any more ideas?
    Furthermore, to prove that there were no file locks present on the affected files, I moved them to a different location on my laptop's HDD and renamed them, which would not have been possible if there had been interference from vmware-related processes. So, that's not it.
    Your suggested workaround, to compress the files before copying them to the external drive, may serve as a temporary workaround but it is not a solution. This VM will grow over time to the point where even the compressed version is larger than the 42GB maximum, and compressing and uncompressing the files will take me a lot of time for files of this size. Could you please continue to pursue this issue and identify the underlying cause?
    Thank you,
    - Jeremy

  • Web Service to transmit a file (a very large file)

    Greetings,
    Is is possible to use a web service to transmit a file, sometimes a very large file? If so, what are the security implications of doing such a thing?
    Thanks,
    Charles

    Quick answer: It depends... Did you try using Google to look-up examples of doing this? Here is a url to try: http://www.google.com/#sclient=psy&hl=en&source=hp&q=web+service+transfer+file&pbx=1&oq=web+service+transfer+file&aq=f&aqi=g1g-j3g-b1&aql=&gs_sm=e&gs_upl=1984731l1997518l0l1997874l25l19l0l1l1l0l304l3879l0.7.10.1l18l0&bav=on.2,or.r_gc.r_pw.&fp=7caaed5eb0414d97&biw=1280&bih=871
    Thank you,
    Tony Miller
    Webster, TX
    Never Surrender Dreams!
    JMS
    If this question is answered, please mark the thread as closed and assign points where earned..

  • Exporting to iTunes results in a very large file

    I have been making music on Garageband and when I export it to itunes, it becomes a very large file. This makes disables me from being able to post my music on my website or email it. What can I do to make the file smaller?
    imac G5   Mac OS X (10.4.5)   1.9 GHz PowerPC G5

    You should convert your file to an MP3 in iTunes:
    - Make sure that in the import options of iTunes, MP3 is selected.
    - Mark your song in iTunes and select "convert to MP3" in the menu.
    - This creates a new file in iTunes with the same title. If you don't know which is which, apple-I will tell you if it's an AIFF or an MP3, and the file size.

  • How do share a very large file?

    How do share a very large file?

    Do you want to send a GarageBand project or the bounced audio file?  To send an audio file is not critical, but if you want to send the project use "File > Compress" to create a .zip file of the project before you send it.
    If you have a Dropbox account,  I'd simply copy the file into the Dropbox "public" folder and mail the link. Right-click the file in the Dropbox, then choose Dropbox > Copy Public Link. This copies an Internet link to your file that you can paste anywhere: emails, instant messages, blogs, etc.
    2 GB on Dropbox are free.     https://www.dropbox.com/help/category/Sharing

  • I inadvertently created a very large file. Not needing it I sent it to the Trash Can. However, the deletion process never completes. Any suggestions as to how to delete it ?

    I inadvertently created a very large file on my hard drive. Not needing it I sent it to the Trash Can. However, the deletion process never completes. Any suggestions as to how to delete it ? I re-installed the OS but the file was still there taking up needed space.

    Command-click all of the files you want to delete and don't actually move them to the Trash until you've gone through the entire folder.
    (63341)

  • Can I save and read text files on a server that I host?

    Hello everyone,
    I am a java hobbiest. I was wondering if I set up my own server, running out of my house, could I have my applets save to my computer and read from my computer without having to learn JDBC and a DATABASE language. In other words could i just have my applet save and read text files from and to my server?
    I'm trying to set up a sight for my 5th grade class where parents can log into. Thanks for your time.
    Oh yeah, which is easier, learning how to set up a server or learning JDBC and a DATABASE language?
    If you have any other good idease please tell me them
    Thank you, Bryan

    Short answer: This isn't gonna work
    Long answer: For this to work, the first thing you're going to need is a static IP address and a DNS name registered -actually you don't necessarily need #2 but you're probably gonna want it and it's by far the easier of the steps.
    As far as I know to get a fixed IP address you've either gotta be directly attached to a larger network (ie university network) or get a leased line from an ISP.
    Once you've got that done come back to us.

  • Handling a very larger file from Sender system

    Hi Sdn,
    Can you please tell me, what should be done from an SAP+XI side to handle a very large file coming from the sender system to be posted to the receiver? What should be done in XI end so that there is no problem in handling the larger file.
    Thanks and regards,
    Aniruddha Bhattacharya

    Hi Aniruddha,
    Go to below link and check point number 3.1.1 maintain parameters as mentioned in that point.
    https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/70ada5ef-0201-0010-1f8b-c935e444b0ad

  • Sorting very large file

    Hi,
    I've a very large file (1.3G, more than 10 million records) that I need to sort, what is the best possible way to do that, without using a database?
    Thanks,
    Gary

    Just a suggestion:
    If you create a sorted red/black tree of record numbers (i.e. offsets in the data file), you could get an equivalent sort as follows (probably without a huge amount of memory...):
    Here's the strategy:
    Create a comparitor that compares two Long objects. It performs a seek in the data file and retrieves the records pointed to by those two Longs and compares them using whatever record comparitor you were going to use.
    Create a sorted list based on that comparitor (a tree based list will probably be best)
    Run a for loop going from 0 to the total records in the data file, create Long objects for each value, and insert that value into the list.
    When you are done, the list will contain the record numbers in sorted order.
    Create an iterator for the list
    For each element in the list, grab the corresponding data file record from the source file and write it to the destination file.
    voila
    Some optimizations may be possible (and highly desirable, actually):
    When you read records from the file, place them into an intermediate cache (linked list?) of limited length (say 1000 records). When a record is requested, scan the cache for that record ID - if found, move it to the top of the list, then return the associated data. If it is not found, read from the file, then add to the top of the list. If the list has more than X records, remove the last item from the list.
    The upper pieces of the sort tree are going to get hammered, so having a cache like this is a good idea.
    Another possibility: Instead of inserting the records in order (i.e. 0 through X), it may be desirable to insert them in pseudo-random order - I seem to remember that there was a really nice algorithm for doing this (without repeating any records) in an issue of Dr. Dobbs from awhile back. You need to get a string of random numbers between 0 and X without repeats (like dealing from a card deck).
    If your data is already in relatively sorted order, the cache idea will not provide significant bennefit without randomizing the inputs.
    One final note: Try the algorithm without the cache first and see how bad it is. The OSes disk caching may provide sufficient performance. If you encapsulate the data file (i.e. an object that has a getRecordBytes(int recNumber) method), then it will be easy to implement the cache later if it is needed.
    I hope this provides some food for thought!
    - K

  • Very large files checkin to DMS/content server

    Dear experts,
    We have DMS with content server up and running.
    However, there is a problem with very large files (up to 2GB!) - these files cannot be checked in due to extremely long upload times and maybe GUI limitations.
    Does anyone have suggestions how we could get very large files into the content server ? I only want to check them in (either via DMS or directly to the content server) and then get the URL back.
    best regards,
    Johannes

    Hi Johannes,
    unfortunatly there is a limit for files regarding the Content Server which is about 2GB. Please note that such large files will cause very long upload times. If possible I would recommend you to split the files to smaller parts and try to check it in.
    From my point of view it is not recommended to put files directly on the Content Server, because this could lead to inconsistencies on the Content Server.
    Best regards,
    Christoph

Maybe you are looking for

  • Open VI Reference blocks event loop

    Run the attached VI.  If you have the WORKS/DOESN'T WORK toggle in the WORKS position and hold down the HOLD ME button with the mouse, the ticking clock keeps on ticking.  If you let go of the mouse and toggle to the DOESN'T WORK positions then hold

  • OS X 10.9 System Preferences - Engery Saver UPS Configuration

    There used to be an option for a UPS configuration under System Preferences - Engery Saver for previous versions of OS X.  Can this feature still be accessed for OS X 10.9, and if so, how and where?

  • Want to modify the partner in the fast entry tab of the trnx : crmd_order

    there r 4 partners in the fast entry tab. such as sold to party, ship to parth, service business partner and execution partner. in this i want to modify the execution partner with new value replacing the old one. i tried using crmd_order_maintain, bu

  • 'number of retries allowed & retry interval in seconds' does not working

    Hi Guys, Is there a way to schedule a report to run,and if the report job fail for whatever reason, retry the job again immediately, and automatically? I've tried use 'number of retries allowed' and 'retry interval in seconds', but it doesn't work. I

  • Iam getting no signals in my 3gs 32gb model

    hi there is no network on my phone,it keeps on saying 'searching' and 'no service', i have updated the phone to 6.1.3, still showing the same again  i have tried factory resetting, the outcome is again same but the phone is unlocked and doing well..