I am running out of memory on my hard drive and need to delete files. How can I see all the files/applications on my hard drive so I can see what is taking up a lot of room?

I am running out of memory on my hard drive and need to delete files. How can I see all the files/applications on my hard drive so I can see what is taking up a lot of room?
Thanks!
David

Either of these should help.
http://grandperspectiv.sourceforge.net/
http://www.whatsizemac.com/
Or search 'disk size' in the App Store.
Be carefull with what you delete & have a backup BEFORE you start, you may also want to reboot to try to free any memory that may have been written to disk.

Similar Messages

  • HT201250 when i m running out of space in my internal drive and need to delete some of my pictures / songs, will the time machine delete those update as well (which i have previously backed up)?

    i m afraid of losing the items when time machine sync with my internal drive.
    so is time machine works like a "mirror" or solely on "backup"

    Yes, Time Machine will eventually delete those things. TM is an incremental backup utility. It is not an archiving utility.
    If you want to keep them, archive (ie copy) them onto some other media (CD, DVD, hard drive, etc.).
    iPhoto used to create iPhoto archives that would mount up in iPhoto just like another Library. I haven't tried on any of the recent versions of iPhoto, though.

  • I'm about to hulk out guys! Brand new iPhone 5 and need to update iTunes for recognition. Keep getting "the installer encountered an error that caused the installation to fail." Can anyone help? I may smash something very soon.

    I'm about to hulk out guys! Brand new iPhone 5 and need to update iTunes for recognition. Keep getting "the installer encountered an error that caused the installation to fail." Can anyone help? I may smash something very soon.

    iTunes:
    AppName: itunes.exe          
    AppVer: 10.5.0.142          
    ModName: unknown
    ModVer: 0.0.0.0   
    Offset: 001be6f6
    Quicktime:
    EventType : BEX   
    P1 : QuickTimePlayer.exe   
    P2 : 7.71.80.42   
    P3 : 4ea5c6ef   
    P4 : QuickTimePlayer.dll   
    P5 : 7.71.80.42   
    P6 : 4ea5c6d5
    P7 : 00005b6d   
    P8 : c0000409   
    P9 : 00000000
    That's consistent with another application (other than QuickTime) having dropped old QuickTime componentry in non-standard locations on the PC.
    So we'll go looking for the older QuickTime componentry in the most common locations for it to be stashed.
    First we'll need to change some view options on the PC.
    1. Open My Computer from the desktop or Start menu.
    2. In the Tools menu, click Folder Options.
    3. Click the View tab.
    4. In the "Advanced settings" pane under "Hidden files and folders" make sure that the "Show hidden files and folders" option is selected, and the "Hide extensions for known file types" option is unchecked.
    5. Click OK.
    Now in My Computer, open your C drive.
    Open the "Windows" folder.
    Open the "system32" folder.
    What files and/or folders can you see in there with QuickTime in the file/folder name? (In a standard installation of QuickTime, you should see precisely two files ... QuickTime.qts and QuickTimeVR.qtx ... and no QuickTime folders whatsoever.)

  • Running out of memory watching HD

    So, I'm watching an HD 1080p video with VLC in full screen. I have 8GB ram and when I start the film I close everything else, leaving about 6.5gb free. Before the film is finishes, it starts to stutter and jerk and everything incurs a spinning beach ball. When I check the available memory I find it's completely empty and the inactive memory is full. I have to restart the mac in order to finish watching it (tried using ifreemem which helps but doesn't give me all the ram back). Does anyone else experience this? Is it a memory leak issue with VLC? Any suggestions as to how I watch something all the way through?

    VLC for Mac has a default setting that "eats" all your RAM. In the preferences of VLC, you have to turn off Input Codecs -> Access Modules -> MMap -> Use file memory mapping.
    Your VLC will not eat more RAM...

  • How can I avoid running out of memory when creating components dynamically

    Hello everyone,
    Recently, I am planning to design a web application. It will be used by all middle school teachers in a region to make examination papers and it must contain the following main functions.
    1)Generate test questions dynamically. For instance, a teacher who logs on the web application will only see a select one menu and a Next Quiz button. The former is used for determining the number of options for the current multiple/single choice question. The later is dedicated to creating appropriate input text elements according to the selected option number. That is to say, if the teacher selects 4 in the menu and presses the Next Quiz button, 5 input text form elements will appear. The first one is for the question to be asked such as "1.What is the biggest planet in the solar system?", the others are optional answers like a)Uranus. b) Saturn. c)Jupiter. d)Earch. Each answer stands for an input text elements. When the teacher fills in the fourth answer, another select one menu and Next Quiz button will emerge on the fly just under this answer, allowing the teacher to make the second question. The same thing repeats for the following questions.
    2)Undo and Redo. Whenever a teacher wants to roll back or redo what he has done, just press the Undo or[i] Redo button. In the previous example, if the teacher selects the third answer and presses the Delete button to drop this answer, it will delete both the literal string content[i] and the input text element, changing the answer d to c automatically. After that, he decides to get back the original answer c, Jupiter, he can just click the Undo button as if he hadn�ft made the deleting operation.
    3)Save the unfinished working in the client side. If a teacher has done half of his work, he can choose to press the Save button to store what he has done in his computer. The reason for doing so is simply to alleviate the burden of the server. Although all finished test papers must be saved in a database on the server, sometimes the unfinished papers could be dropped forever or could form the ultimate testing papers after several months. So if these papers keep in the server, it will waste the server computer�fs room. Next time the teacher can press the Restore button on the page to get the previously stored part of the test paper from his own computer and continue to finish the whole paper.
    4)Allow at least 1,000 teachers to make test papers at the same time. The maximum question number per examination paper is 60.
    Here are my two rough solutions,
    A.Using JSF.
    B.Using JavaScript and plain JSP[b] without JSF.
    The comparison of the two solutions:
    1)Both schemas can implement the first and the second requirements. In JSF page I could add a standard panelGird tag and use its binding attribute. In the backing bean, the method specified by the binding attribute is responsible for generating HtmlInput objects and adding them to the HtmlPanelGird object on the fly. Every HtmlInput object is corresponding to a question subject or an optional answer. The method is called by an actionListener, which is registered in the Next Quiz commandButton, triggering by the clicking on this button in the client side. Using JSF can also be prone to managing the HtmlInput objects, e.g. panelGird.getChildren().add(HtmlInput) and panelGird.getChildren().remove(HtmlInput) respond to the undoing operation of deleting an optional answer and the redoing operation of the deleting action respectively. I know JavaScript can also achieve these goals. It could be more complex since I don�ft know well about JavaScript.
    2)I can not find a way to meet the third demand right now. I am eager to know your suggestion.
    3)Using JSF, I think, can�ft allow 1,000 teachers to do their own papers at the same time. Because in this scenario, suppose each questionnaire having 60 questions and 4 answers per question, there will be approximately 300,000 HtmlInput objects (1,000X60X(4+1)) creating on the server side. The server must run out of memory undoubtedly. To make things better, we can use a custom component which can be rendered as a whole question including its all optional answers. That is to say, a new custom component on the server side stands for a whole question on the client side. Even so, about 60,000(1,000X60) this type of custom components will be created progressively and dynamically, plus other UISelectOne and UICommand objects, it also can�ft afford for most servers. Do I have to use JavaScript to avoid occupying the server's memory in this way? If so, I have to go back and use JavaScript and plain JSP without JSF.
    Thank you in advance!
    Best Regards, Ailsa
    2007/5/4

    Thank you for your quick response, BalusC. I really appreciate your answer.
    Yes, you are right. If I manually code the same amount of those components in the JSF pages instead of generating them dynamically, the server will still run out of memory. That is to say, JSF pages might not accommodate a great deal of concurrent visiting. If I upgrade the server to just allow 1,000 teachers making their own test papers at the same time, but when over 2,000 students take the same questionnaire simultaneously, the server will need another upgrading. So I have to do what you have told me, using JS+DOM instead of upgrading the server endlessly.
    Best Regards, Ailsa

  • I want to use Meteor app but run out of memory.

    Is there a way I can store data on the cloud, so freeing up memory to work on any given app? I run out of space all the time. I use lots of apps in different media so
    can I say, do some work on one app then save the files to the Cloud, then reload later when I need it. Basically, I feel I should only Keep the apps on my iPad, then on each separate app save the work as I finish for the day to Keep my memory uncluttered? Any help would be appreciated. I spent lots on apps but run out of memory. Thanks.

    Only if the individual Apps support saving to the cloud. Otherwise no. There is no user access to the iCloud storage area.
    Its only there for backups, and data syncronization between certain Apps that support it.

  • Oracle 9i running out of memory

    Folks !
    I have a simple 3 table schema with a few thousand entries each. After dedicating gigabytes of hard disk space and 50% of my 1+ GB memory, I do a few simple Oracle Text "contains" searches (see below) on these tables and oracle seems to grow some 25 MB after each query (which typically return less than a dozen rows each) till it eventually runs out of memory and I have to reboot the system (Sun Solaris).
    This is on Solaris 9/Sparc with Oracle 9.2 . My query is simple right outer join. I think the memory growth is related to Oracle Text index/caching since memory utilization seems pretty stable with simple like '%xx%' queries.
    "top" shows a dozen or so processes each with about 400MB RSS/SIZE. It has been a while since I did Oracle DBA work but I am nothing special here. Databse has all the default settings that you get when you create an Oracle database.
    I have played with SGA sizes and no matter how large or small the size of SGA/PGA, Oracle runs out of memory and crashes the system. Pretty stupid to an Enterprise databas to die like that.
    Any clue on how to arrest the fatal growth of memory for Oracle 9i r2?
    thanks a lot.
    -Sanjay
    PS: The query is:
    SELECT substr(sdn_name,1,32) as name, substr(alt_name,1,32) as alt_name, sdn.ent_num, alt_num, score(1), score(2)
    FROM sdn, alt
    where sdn.ent_num = alt.ent_num(+)
    and (contains(sdn_name,'$BIN, $LADEN',1) > 0 or
    contains(alt_name,'$BIN, $LADEN',2) > 0)
    order by ent_num, score(1), score(2) desc;
    There are following two indexes on the two tables:
    create index sdn_name on sdn(sdn_name) indextype is ctxsys.context;
    create index alt_name on alt(alt_name) indextype is ctxsys.context;

    I am already using MTS.
    Atached is the init.ora file below.
    may be I should repost this article with subject "memory leak in Oracle" to catch developer attention. I posted this a few weeks back in Oracle Text groiup and no response there either.
    Thanks for you help.
    -Sanjay
    # Copyright (c) 1991, 2001, 2002 by Oracle Corporation
    # Cache and I/O
    db_block_size=8192
    db_cache_size=33554432
    db_file_multiblock_read_count=16
    # Cursors and Library Cache
    open_cursors=300
    # Database Identification
    db_domain=""
    db_name=ofac
    # Diagnostics and Statistics
    background_dump_dest=/space/oracle/admin/ofac/bdump
    core_dump_dest=/space/oracle/admin/ofac/cdump
    timed_statistics=TRUE
    user_dump_dest=/space/oracle/admin/ofac/udump
    # File Configuration
    control_files=("/space/oracle/oradata/ofac/control01.ctl", "/space/oracle/oradata/ofac/control02.ctl", "/space/oracle/oradata/ofac/control03.ctl")
    # Instance Identification
    instance_name=ofac
    # Job Queues
    job_queue_processes=10
    # MTS
    dispatchers="(PROTOCOL=TCP) (SERVICE=ofacXDB)"
    # Miscellaneous
    aq_tm_processes=1
    compatible=9.2.0.0.0
    # Optimizer
    hash_join_enabled=TRUE
    query_rewrite_enabled=FALSE
    star_transformation_enabled=FALSE
    # Pools
    java_pool_size=117440512
    large_pool_size=16777216
    shared_pool_size=117440512
    # Processes and Sessions
    processes=150
    # Redo Log and Recovery
    fast_start_mttr_target=300
    # Security and Auditing
    remote_login_passwordfile=EXCLUSIVE
    # Sort, Hash Joins, Bitmap Indexes
    pga_aggregate_target=25165824
    sort_area_size=524288
    # System Managed Undo and Rollback Segments
    undo_management=AUTO
    undo_retention=10800
    undo_tablespace=UNDOTBS1

  • Running out of memory on iPhone

    hi all,
    when my Flash app is just getting started on the iPhone 3G, it runs out of memory.  I see this in the crash report:  "Count resident pages" is "11727 (jettisoned) (active)." 
    How do I prevent this?
    If it's relevant, my app has lots of graphics and audio, and I'm guessing it's all being loaded upfront.  But it's presented sequentially, so maybe I can just load one set of assets at a time.  Is there some way to break it up?
    thanks

    Hi,
    I was having 3 iphone apps which is in .swf format and i tried to convert them into .ipa, in order to deploy on iPad. and I used Adobe iPhone packager and issued the command.
    How ever I was successfully converted one app and deployed on iPad and it is working fine. But for 2nd app I am getting the fallowing error.
    Exception in thread “main” java.lang.OutOfMemoryError: Java heap space
    at adobe.abc.GlobalOptimizer$FrameState.(GlobalOptimizer.java:9154)
    at adobe.abc.GlobalOptimizer.createBlock(GlobalOptimizer.java:9170)
    at adobe.abc.GlobalOptimizer.merge(GlobalOptimizer.java:9217)
    at adobe.abc.GlobalOptimizer$InputAbc.readCode(GlobalOptimizer.java:1051)
    at adobe.abc.GlobalOptimizer$InputAbc.readBody(GlobalOptimizer.java:531)
    at adobe.abc.GlobalOptimizer$InputAbc.readAbc(GlobalOptimizer.java:404)
    at adobe.abc.GlobalOptimizer$InputAbc.readAbc(GlobalOptimizer.java:280)
    at adobe.abc.LLVMEmitter.generateBitcode(LLVMEmitter.java:160)
    at com.adobe.air.ipa.AOTCompiler.convertAbcToLlvmBitcode(AOTCompiler.java:329)
    at com.adobe.air.ipa.AOTCompiler.GenerateMacBinary(AOTCompiler.java:600)
    at com.adobe.air.ipa.IPAOutputStream.compileRootSwf(IPAOutputStream.java:196)
    at com.adobe.air.ipa.IPAOutputStream.finalizeSig(IPAOutputStream.java:355)
    at com.adobe.air.ADTPackager.createPackage(ADTPackager.java:65)
    at com.adobe.air.ipa.IPAPackager.createPackage(IPAPackager.java:165)
    at com.adobe.air.ADTEntrypoint.parseArgsAndGo(ADTEntrypoint.java:132)
    at com.adobe.air.ipa.PFI.parseArgsAndGo(PFI.java:152)
    at com.adobe.air.ADTEntrypoint.run(ADTEntrypoint.java:68)
    at com.adobe.air.ipa.PFI.main(PFI.java:112)
    And 3rd app is converted into .ipa file but when i deployed on ipad and tried to run it is crashing.
    Can any one, please let me know the reason for the last two scenarios .
    Thanks & Regards
    Ravi Bukka
    [email protected]

  • Workfow Iterate to subprocess runs out of memory

    I have a workflow that returns all suspended tasks and then calls a subprocess for each task. The subprocess decides whether the task needs to be deleted and if so it processes the task in various ways before deleting the taskinstance.
    I have no issues when there are not too many tasks returned by the query but when the workflow returns 2000+ items, I run out of memory.
    What is the best way to workflow to call the subprocess without running out of memory?
    Do I need to cleanup something at the end of subprocess?
    Do I need to add something in the workflow to beakup the list of tasks into smaller chunks?
    <Activity id='3' name='ProcessTasks'>
    <Action id='0' name='processTasks' process='processTheTask'>
    <Iterate for='taskInstanceName' in='mytasks'/>
    <Argument name='taskInstanceName' value='$(taskInstanceName)'/>
    </Action>
    Edited by: user1937458 on Mar 14, 2012 3:12 PM

    I didnt think that this would put that much stress on the system.
    1) Use IDM best practice to generate low memory tasks, use exposedVariables and extendedVariables in manual actions to generate low memory tasks, that will save lots of memory.
    No manual action. This is a scheduled task.
    2) Run this workflow on dedicated server which is responsible to run this task only.
    I have run this when no one else was using the system but that did not help either.
    3) You can put some more conditions to get the limited return data which your server can handle in one go.
    we normally have 8000 tasks in the system. About 5000 are completed so I can ignore those in the workflow. The rest need to be looked at to determine if we need to update the request. Let's say that I can use a rule to determine that in the workflow before the subprocess is called and I end up with a list of 500 taskinstance names, I think that the process will still run out of memory unless there is some other solution.
    2000 task names in a list should not take up that much space. I am pretty sure that the subprocess which determines if the task needs to be deleted is chewing up resources. This is going to be a scheduled task with no manual actions.
    My thinking was that workflow calls the subprocess and the subprocess does a lot of work as far as canceling a request, disabling accounts in some cases, auditing and notifying users that their request was cancelled. Upon return to the workflow to get the next taskinstance name, there is probably some variable that keeps getting larger with each iteration.
    I have run smaller lists and the flow diagram that returns at the end shows the flowchart for every item that was deleted so that is probably 1 place where the variable keeps getting larger.
    Is there a way to clean everything so that each subprocess acts as if it was the 1st and only time it was getting called?
    I tried the following at the end of the subprocess but that did not help:
    <Action id='0' name='CleanUp'>
    <expression>
    <set name='WF_CASE_RESULT'/>
    </expression>
    </Action>
    I will try to debug and see what variables are getting larger and larger but any other suggestions are appreciated.

  • System hanging when it runs out of memory

    Hello,
    my system has a finite amount of RAM and swap (it doesn't matter, to my purpose, if it's 16GB or 128MB, I'm not interested in increasing it anyway).
    Sometimes my apps completely use all the available memory. I would expect that in these cases the kernel kills some apps to keep working correctly. The OOM Killer exists just for this, doesn't it?
    What happens, instead, is that the system hangs. Even the mouse stops working. Sometimes it manages to get back to life in a few seconds/minutes, other times hours pass by and nothing changes.
    I don't want to add more memory, I just want that the kernel kills some application when it's running out of memory.
    Why isn't this happening?
    Right now I'm writing a bash script that will kill the most memory-hungry process when available memory gets below 10%, because I'm sick of freezing my machine.
    But why do I have to do this? Why do I need an user space tool polling memory usage and sentencing applications according to a cheap policy? What the hell is wrong with my OOM Killer, why isn't it doing is job?!

    Alright, you won, now quit pointing out my ignorance
    Your awkish oom killer is a lot cooler than mine, switching to it, thanks!
    I did some testing (initially just to test the oom killing script) and found out that if a program tries to allocate all the memory it can, it gets eventually killed by linux's oom killer. If instead it stops allocating new memory when there are less than 4MB of free memory (or similar values) the OOM won't do anything, and the system will get stuck, like if a forkbomb was running.
    That's it, this program with MINSIZE=1 will be killed, while with MINSIZE=4MB will force me to hard reboot:
    #include <string.h>
    #include <stdlib.h>
    #define MINSIZE (1024*1024*4) // 4MB
    int main( )
    int block = 1024*1024*1024; // 1GB
    void *p;
    while( 1 ) {
    p = malloc( block );
    if( p ) {
    memset( p, 85, block );
    else if( block > MINSIZE ) {
    block /= 2;
    return 0;
    Guess I'd need to go deeper to understand why Linux' oom killer works like that, but I won't (assuming the oom killing script will behave).

  • Premiere Pro CS5 runs out of memory

    I have the same problem with Premiere running out of memory loading a project with lots of images in it.  Running task manager, you can see it eat memory at 250MB per second until it pegs the meter after a short time and then things start failing.  The CS4 version does not have this problem and opens the project without error.

    What are the pixel x pixel dimensions of those still images? If they are overly-large, there is a very heavy load, when working with that Project. For Scaling them to fit the Sequence Frame Size, with automation in PS, see this ARTICLE.
    If one's computer is already on the borderline for resources, overly-large stills can shove it over the brink. There are few reasons to use overly-large images, as one can only ever see the pixels in the Sequence's Frame Size. The only reason for using images larger than that Frame Size are to Pan on Zoomed out images, with Effect>Motion>Position.
    Good luck,
    Hunt

  • Problems with PNGs... Overall compression + Running out of memory!

    We're having a number of issues with PNGs while working on our first iPhone project, and any assistance would be greatly appreciated!
    Our game is using a large number of PNG assets, some of which are full-frame (though the full-frame files tend to mostly be transparent/use alphas, though apparently that doesn't help memory issues much).
    We're running into two huge problems -
    1) We're running out of memory on device when calling to these full frame sequences, which tend to be anywhere from 10-40 frames each at anywhere from 50 to 250kb in size.
    2) Our overall package size is huge, sitting at around 60mb. I've already compressed the pngs through photoshop to the best of my ability, and I'm not having much luck with downloadable compressors like pngcrush. IS there a way to compress all the pngs through xcode/c++/objective C? The programmers are informing me right now that the only compression possible is whatever I apply directly to my pngs on my end - nothing through code.
    I'm stumped as to how I'm seeing seemingly-complex apps with plenty of content at 1-5mbs, and running smoothly with full-frame animations. I'm imagining the problem is that we're not using a proprietary engine to properly manage things, but I'm wondering if there is a simple solution.
    Thanks in advance, guys!
    P.S. - Already done a bunch of research on my own, and not having much luck. Just wondering if there is something obvious that both the programmers and I are missing!

    * Make image frame smaller? A fully transparent border around an image is pure overhead -- there is still data in the actual pixels (even though you cannot see them), plus the transparent border itself. All it needs is an x/y coordinate and a tiny code adjustment.
    * Use less transparency bits? Perhaps you could do with 1-bit transparency on some images. It'll use 1/8 times the memory (well, globally).
    * Use less colors? 24 bit color might compress down to 16 bits without visual artifacts (esp. when you don't have lots of gradients). Perhaps even 8 bit palettized.
    * Make images smaller? You might be able to get away with enlarging some images.
    Just a few things you could check right away, without major rewrites.

  • A lot of my pictures are in I Photo and also in Finder.  Can I delete them from finder?  They are taking up a lot of room.

    A lot of my pictures are in I Photo and also in Finder.  Can I delete them from finder?  They are taking up a lot of room.

    Just make certain that the "Originals" or "Masters" are in your iPhoto Library before you trash the images.
    You can find out by selecting the iPhoto Library file with the control key held down, then open the Library (which is, actually, a folder). You'll see a folder named "Originals" or "Masters" and you can check there to make certain that you have all of the high-res images in your iPhoto Library. If you do, you can safely trash the original files, although I would, if it were me, make an archive of the images before trashing (either on an external HD or on archival DVDs).
    Make sense?
    Clinton

  • Logic Pro 7 running out of memory!?!

    Hi!
    I am working on a project were I basically want to cut up every beat of a drum recording on 12 tracks and correct their timing individually (I have my reasons for doing this).
    This creates a decent amount of regions and for some reason Logic starts complaining about running short on memory and not having enough memory for the undo history.
    It bugs me because I am working on a Dual 2.5GHz G5 with 2 GB of RAM. Activity monitor says I have 480 MB of inactive memory and 1.25GB of free memory. Logic uses 300 MB of real memory and 762 MB of virtual memory. No other apps running.
    How can I make Logic use ALL the available memory?
    Is there a limit on the amount of memory that Logic can address? What is this limit, and how much memory is reasonably motivated for a PM running Logic?
    How about plug ins? Do they have separate memory spaces not limited by Logics potential limits? "Drumkit from ****" can by itself use a couple of Gigs of RAM if available...
    Thanks in advance!
    /Jont Olof

    Jont Olof Lyttkens, "Logic Pro 7 running out of memory!?!" #5, 03:19pm Sep 7, 2005 CDT
    However in the earlier days, running under Mac OS pre 9.x the ability to allocate memory to a specific application could increase the amount for the song event memory
    No, it didn't. It increased application memory, but the song event memory remained the same, just as it does now.
    I think, though, that with all Logic's new features, there is more actual data in an equivalent song, so the limit may be coming earlier. (I can't be sure of that though)
    I know you don't want to, but did you at least try reducing the undo history size (in prefs) ?
    Does deleting undo history and then saving help ?
    Another thing to consider, although I don't think it generates warnings like you are seeing, is Virtual Memory. If the computer has been up for a long time, running memory-intensive apps, like Logic, then it can get overloaded (too many pageouts) which can slow things down a lot. If you do notice things getting slower, then restarting the computer will flush the Swap File and get things going again.

  • Lightroom 5 permanently runs out of memory

    Lightroom 5 on Windows 7 32 Bit and 8 Gigabytes of memory (more than the 32 Bit system can use) permanently runs out of memory when doing some more complex edits on a RAW file, especially when exporting to 16 Bit TIFF. The RAW files were created by cameras with 10 up to 16 megapixel sensors with bit depths between 12 and 14.
    After exporting one or two images to 16 Bit uncompressed TIFF an error message "Not enough memory" will be displayed and only a Lightroom restart solves that - for the next one to two exports. If an image has much brush stroke edits, every additional stroke takes more and more time to see the result until the image disappears followed by the same "Not enough memory" error message.
    A tab character in the XMP sidecar file is *not* the reason (ensured that), as mentioned in a post. It seems that Lightroom in general does not allocate enough memory and frees too less/late allocated.
    Please fix that bug, it's not productive permanently quit and restart Lightroom when editing/exporting a few RAW files. Versions prior to Lightroom 4 did not have that bug.
    P.S. Posting here, because it was not possible to post it at http://feedback.photoshop.com/photoshop_family/topics/new It's very bad design, to let a user take much time to write and then say: "Log in", but a log in with the Adobe ID and password does not work (creating accounts on Facebook etc. is not an acceptable option, Adobe ID should be enough). Also a bugtracker such as Bugzilla would be a much better tool for improving a software and finding relevant issues to avoid duplicate postings.

    First of all: I personally agree with your comments regarding the feedback webpage. But that is out of our hands since this is a user-to-user forum, and there is nothing we users can do about it.
    Regarding your RAM: You are running Win7 32-bit, so 4 GB of your 8 GB of RAM sit idle since the system cannot use it. And, frankly, 4 GB is very scant for running Lr, considering that the system uses 1 GB of that. So there's only 3 GB for Lr - and that only if you are not running any other programs at the same time.
    Since you have a 8 GB system already, why don't you go for Win7 64-bit. Then you can also install Lr 64-bit and that - together with 8 GB of RAM - will bring a great boost in Lr performance.
    Adobe recommends to run Lr in the 64-bit version. For more on their suggestion on improving Lr performance see here:
    http://helpx.adobe.com/lightroom/kb/performance-hints.html?sdid=KBQWU
    for more: http://forums.adobe.com/thread/1110408?tstart=0

Maybe you are looking for

  • ITouch 4th Gen, crashes almost daily, how do I fix it?

    Ok, a month ago the iTouch started to act up, when iOS 5 was released.  Clearly it is a software issue.  At that time, I came to this site because it would not start, it kept the apple logo up.  I eventually used the white apple PC connection cord an

  • How can i request just the navigation or just the product list when i query the dgraph?

    Is there a way to make a request for navigation only or record list only?

  • BOM-MRP

    Hi, I have created a BOM. In our sale process, I want to invoice header item while I implement MRP for sub items. I double checked schedule line, item category and every thing ... I face a problem in this process. When I create a sales order for a su

  • Acrobat 9/9.1  - Coruption when export to Word

    Hi, I have been using Adobe Acrobat 9 for a few months now, both in version 9 and 9.1 I have recently been asked to utilise the export to Word feature, which from experience in version 8, has and should be a relatively painless process. However, to m

  • IPhone: Additional Examples available....

    I just discovered that the latest iPhone docs come with some additional example apps not referenced in the Dev Center. For example there is "iPhone Recipes", a nice custom list app, and "SimpleAnimation". Go to the doc browser and look for "Recipes,"