U0093Critical program error occurred .u0093Client out of memory error u0093 - Query

I have a problem with Query in BI 7.0
Query works perfectly in BW3.5 environment.  BI 7.0 Vista and Excel 2007 environment – I have date range in the query – If I provide date range 4 months interval it is working fine.
If I provide 5, 6,7, months interval – I am getting the error ,
“Critical program error occurred .The program has to close. Please refer to the trace for further information.”
Communication error, CPIC return code 020, SAP returns code 223
“Client out of memory error “
When I execute each interval this query and calculated the no of records – it is only less than 19,000 records.
The same query working fine in Bw3.5  for more than year interval.
Advance Thanks .

It depends on the ssytem settings. The problem with your query is that it is not able to fetch all the data due to lack of memory. However, when you are giving smaller range, you are getting the output. You can either ask the basis guy to increase the memory space or try running the query by giving smaller date range. You can have some filters on your query also.
Thanks...
Shambhu

Similar Messages

  • ERROR [B3108]: Unrecoverable out of memory error during a cluster operation

    We are using Sun Java(tm) System Message Queue Version: 3.5 SP1 (Build 48-G). We are using two JMS servers as a cluster.
    But we frequently getting the out of memory issue during the cluster operation.
    Messages also got queued up in the Topics. Eventhough listeners have the capability to reconnect with the Server after the broker restarting, usually we are restarting consumer instances to get work this.
    Here is detailed log :
    Jan 5 13:45:40 polar1-18.eastern.com imqbrokerd_cns-jms-18[8980]: [ID 478930 daemon.error] ERROR [B3108]: Unrecoverable out of memory error during a cluster operation. Shutting down the broker.
    Jan 5 13:45:57 polar1-18.eastern18.chntva1-dc1.cscehub.com imqbrokerd: [ID 702911 daemon.notice] Message Queue broker terminated abnormally -- restarting.
    Expecting your attention on this.
    Thanks

    Hi,
    If you do not use any special cmdline options, how do you configure your servers/
    brokers to 1 Gb or 2 Gb JVM heap ?
    Regarding your question on why the consumers appear to be connecting to just
    one of the brokers -
    How are the connection factories that the consumers use configured ?
    Is the connection factory configured using the imqAddressList and
    imqAddressListBehavior attributes ? Documentation for this is at:
    http://docs.sun.com/source/819-2571/ref_adminobj_props.html#wp62463
    imqAddressList should contain a list of brokers (i.e. 2 for you) in the cluster
    e.g.
    mq://server1:7676/jms,mq://server2:7676/jms
    imqAddressListBehavior defines how the 2 brokers in the above list are picked.
    The default is in the order of the list - so mq://server1:7676/jms will always be
    picked by default. If you want random behavior (which will hopefully even out the
    load), set imqAddressListBehavior to RANDOM.
    regards,
    -i
    http://www.sun.com/software/products/message_queue/index.xml

  • Out of memory error occurred while processing message at RFC receiver

    Hi Experts
    Hi Experts
    i have a scenario file to RFC BPM bridge Sync/Async,
    i´m getting this error when this BPM is trying to read  the RFC
    com.sap.engine.interfaces.messaging.api.exception.MessagingException: com.sap.aii.adapter.rfc.afcommunication.RfcAFWException: error while processing message to remote system:com.sap.aii.adapter.rfc.core.client.RfcClientException: Out of memory error occurred while processing message at RFC receiver
    please somebody can you help me

    Hi,
    No need to restart the server. In PI there is a parameter 'java Heap' indicating the size of data processed adapters. Check if it is necessary to increase the size of this parameter.
    Check the note: 1173398
    I hope to help you
    Regards,

  • Error due to out of memory condition

    Hi,
    system: Windows XP, 2 GM RAM, Indesign CS3
    I placed two files in Indesign: the first is .ai (1.1 MB, probably exported form a CAD program), the second is .psd (53 MB). When I try to print to .ps file, Indesign displays the message: "Export error: error due to out of memory condition". If I: 1. export to PDF, everything is OK, 2. rasterize .ai in Photoshop, then import the resulting .psd to Indesign, print to .ps file is OK.
    Does somebody has an answer to my problem?

    My gues is "mistakes" was referred to simply not outputting as intended by the designer.  I had output issues when going directly to PDF and had to also continue the old way of doing thinkgs by outputting to .ps then distill to PDF.  The "mistakes" that occured for me were nearly unnoticed.  The publication was a 128+4 product catalog that had a header at the top of each page.  I never had any problems outputting in the past.  At the time, I just had updated to CS3 (XP Pro SP3 Dual Core 4 gigs ram).  I was advised that outputting directly to PDF had been durastically improved.  I output my files without "mistakes" or errors, or so I thought.  After proofing, I didn't realize that all of the drop shadows for the header headline were not applied.  The problem though, wasn't that they were applied to all headers, just beginning at page 60 or so, halfway through.  After getting the finished publication back, I didn't notice it for a few months.  After noticing it, I investigated.  Several troubleshooting hours later (after forum posts and other expert help), I/we concluded that it was simply a program deficiency.  During my trouble shooting, I looked at several areas mainly focusing on the transparency settings.  I output the document several times.  I finally exported each page separately and found that that was the only way that I could output every page to include the dropshadow in each header page (past page 60 or so).
    I have been past that issue for a while now.  Now I am on Win 7, Quad core, 6 gigs of ram, and don't have the outputting problem that I used to.  Now I occasionally get this "Error due to out of memory condition" which is BS because my hardware specs are beyond reasonable for what I typically create in my workflow.  Most of the time, my memory is only at about half of its capacity when I get this error.  Not when I have a PS open with a large multipage file that I am working on.  My out of memory error happens during file output or just when performing ordinary layout functions in InDesign.
    When on my personal computer (Aluminum iMac 2.4 Core2 duo, 4 gigs of ram, 512 video, CS4 & CS5) I have neither of the above issues.

  • Out of memory error - from parsing a "fixed width file"

    This may be fairly simple for someone out there but I am trying to write a simple program that can go through a "fixed width" flat txt file and parse it to be comma dilmeted.
    I use a xml file with data dictionary specifications to do the work. I do this because there are over 430 fields that need to be parsed from a fixed width with close to 250,000 lines I can read the xml file fine to get the width dimensions but when I try to apply the parsing instructions, I get an out of memory error.
    I am hoping it is an error with code and not the large files. If it is the latter, does anyone out there know some techniques for getting at this data?
    Here is the code
       import java.io.*;
       import org.w3c.dom.Document;
       import org.w3c.dom.*;
       import javax.xml.parsers.DocumentBuilderFactory;
       import javax.xml.parsers.DocumentBuilder;
       import org.xml.sax.SAXException;
       import org.xml.sax.SAXParseException;
        public class FixedWidthConverter{
          String[] fieldNameArray;
          String[] fieldTypeArray;
          String[] fieldSizeArray;      
           public static void main(String args []){
             FixedWidthConverter fwc = new FixedWidthConverter();
             fwc.go();
             fwc.loadFixedWidthFile();
            //System.exit (0);
          }//end of main
           public void go(){
             try {
                DocumentBuilderFactory docBuilderFactory = DocumentBuilderFactory.newInstance();
                DocumentBuilder docBuilder = docBuilderFactory.newDocumentBuilder();
                Document doc = docBuilder.parse (new File("files/dic.xml"));
                // normalize text representation            doc.getDocumentElement ().normalize ();
                System.out.println ("Root element of the doc is " +
                     doc.getDocumentElement().getNodeName());
                NodeList listOfFields = doc.getElementsByTagName("FIELD");
                int totalFields = listOfFields.getLength();
                System.out.println("Total no of fields : " + totalFields);
                String[] fldNameArray = new String[totalFields];
                String[] fldTypeArray = new String[totalFields];
                String[] fldSizeArray = new String[totalFields];
                for(int s=0; s<listOfFields.getLength() ; s++){
                   Node firstFieldNode = listOfFields.item(s);
                   if(firstFieldNode.getNodeType() == Node.ELEMENT_NODE){
                      Element firstFieldElement = (Element)firstFieldNode;
                      NodeList firstFieldNMList = firstFieldElement.getElementsByTagName("FIELD_NM");
                      Element firstFieldNMElement = (Element)firstFieldNMList.item(0);
                      NodeList textFNList = firstFieldNMElement.getChildNodes();
                      //System.out.println("Field Name : " +
                               //((Node)textFNList.item(0)).getNodeValue().trim());
                      //loads values into an array
                      //fldNameArray[s] = ((Node)textFNList.item(0)).getNodeValue().trim();
                      NodeList typeList = firstFieldElement.getElementsByTagName("TYPE");
                      Element typeElement = (Element)typeList.item(0);
                      NodeList textTypList = typeElement.getChildNodes();
                      //System.out.println("Field Type : " +
                               //((Node)textTypList.item(0)).getNodeValue().trim());
                      //loads values into an array
                      //fldTypeArray[s] = ((Node)textTypList.item(0)).getNodeValue().trim(); 
                      NodeList sizeList = firstFieldElement.getElementsByTagName("SIZE");
                      Element sizeElement = (Element)sizeList.item(0);
                      NodeList textSizeList = sizeElement.getChildNodes();
                      //System.out.println("Field Size : " +
                               //((Node)textSizeList.item(0)).getNodeValue().trim());
                      //loads values into an array
                      fldSizeArray[s] = ((Node)textSizeList.item(0)).getNodeValue().trim();   
                   }//end of if clause
                }//end of for loop with s var
                //setFldNameArray(fldNameArray);
                //setFldTypeArray(fldTypeArray);
                setFldSizeArray(fldSizeArray);
                 catch (SAXParseException err) {
                   System.out.println ("** Parsing error" + ", line "
                      + err.getLineNumber () + ", uri " + err.getSystemId ());
                   System.out.println(" " + err.getMessage ());
                 catch (SAXException e) {
                   Exception x = e.getException ();
                   ((x == null) ? e : x).printStackTrace ();
                 catch (Throwable t) {
                   t.printStackTrace ();
          }//end go();
           public void setFldNameArray(String[] s){
             fieldNameArray = s;
          }//end setFldNameArray
           public void setFldTypeArray(String[] s){
             fieldTypeArray = s;
          }//end setFldTypeArray
           public void setFldSizeArray(String[] s){
             fieldSizeArray = s;
          }//end setFldSizeArray
           public String[] getFldNameArray(){
             return fieldNameArray;
          }//end setFldNameArray
           public String[] getFldTypeArray(){
             return fieldTypeArray;
          }//end setFldTypeArray
           public String[] getFldSizeArray(){
             return fieldSizeArray;
          }//end setFldSizeArray 
           public int getNumLines(){
             int countLines = 0;
             try {
                    //File must be in same director and be the name of the string below
                BufferedReader in = new BufferedReader(new FileReader("files/FLAT.txt"));
                String str;
                while ((str = in.readLine()) != null) {
                   countLines++;
                in.close();
                 catch (IOException e) {}    
             return countLines;
          }//end of getNumLines
           public void loadFixedWidthFile(){
             int c = getNumLines();
             int i = 0;
             String[] lineProcessed = new String[c];
             String chars;
             try {
                    //File must be in same director and be the name of the string below
                BufferedReader in = new BufferedReader(new FileReader("files/FLAT.txt"));
                String str;
                while ((str = in.readLine()) != null) {
                   //System.out.println(str.length());
                   lineProcessed[i] = parseThatLine(str);
                   i++;
                in.close();
                 catch (IOException e) {}     
                //write out the lineProcess[] array to another file
             writeThatFile(lineProcessed);
          }//end loadFixedWidthFile()
           public void writeThatFile(String[] s){
             try {
                BufferedWriter out = new BufferedWriter(new FileWriter("files/outfilename.txt"));
                for(int i = 0; i < s.length -1; i++){
                   out.write(s);
    }//end for loop
    out.close();
    catch (IOException e) {}
    }//end writeThatFile
    public String parseThatLine(String s){
    int start = 0;
    int end = 0;
    String parsedLine = "";
    int numChars = getFldSizeArray().length;
    //Print number of lines for testing
    //System.out.println(numChars);
    String[] oArray = getFldSizeArray();
    //String chars = oArray[0];
    //System.out.println(chars.length());
    //oArray
    for(int i = 0; i < numChars -1; i++ ){
    if(i == 0){
    start = 0;
    end = end + Integer.parseInt(oArray[i])-1;
    else
    start = end;
    end = end + Integer.parseInt(oArray[i]);
    parsedLine = parsedLine + s.substring(start, end) + "~";
    }//end for loop
    return parsedLine;
    }//End of parseThatLine
    I have tried to illeminate as many arrays as I can thinking that was chewing up the memory but to no avail.
    Any thoughts or ideas?
    Message was edited by:
    SaipanMan2005

    You should not keep a String array of all the lines of the file read.
    Instead for each line read, parse it, then write the parsed line in the other file:      public void loadFixedWidthFile() {
             BufferedReader in = null;
             BufferedWriter out = null;
             try {
                //File must be in same director and be the name of the string below
                in = new BufferedReader(new FileReader("files/FLAT.txt"));
                out = new BufferedWriter(new FileWriter("files/outfilename.txt"));
                String str;
                while ((str = in.readLine()) != null) {
                   //System.out.println(str.length());
                   str = parseThatLine(str);
                   //write out the parsed str to another file
                   out.write(str);
             catch (IOException e) {
                e.printStackTrace(); // At least print the exception - never swallow an exception
             finally { // Use a finally block to be sure of closing the files even when exception occurs
                try { in.close(); }
                catch (Exception e) {}
                try { out.close(); }
                catch (Exception e) {}
          }//end loadFixedWidthFile()Regards

  • How can I solve out of memory error on excell file in PL/SQL

    Hi,
    I'm new on PL/SQL. One of the PL/SQL code which is created excell report got out of Memory error. The first reason of this error, excell file not supported more than 65536 data. So I change the excell file separeted sheets. So that the single sheet size cannot exceed 65536 data.
    All the data are held on system cach and if many user want to take the report the they would get an out of memory error.
    So I want to change the code like that; when out of memory exception raises,
    the old excell file save to disk and new excell file is created,
    and go on to write the new file without exiting the program.
    At the end of the data all the excell file append and show only one file to the user.
    I do know how to save the file and create a new file. But I don't know how can PL/SQL program to turn back to loop again when the exception occurs.
    Is anyone help me on this issue?
    Here is my code
    Thank you
    dworkbook:=hssfworkbook.new;
    dCurrentItem := Get_Block_Property(pCurrentBlock, FIRST_ITEM);     
    while not (name_in('system.last_record')='TRUE') loop
    /* The data would be written to the excell file column order. */
    if (dRow=0) then
              /* Create a new sheet */
    elsif (dRow <= dMaxWorksheetNum) then
         /* Data of the report are written here. The data are written in column order */
    if (dRow > dMaxWorksheetNum) then
         /* give dRow and dColumn intial value */
    /* increase worksheet number */
    end if; /* End of if (dRow=1) */
    if (isWritten) and not name_in('system.last_record')='TRUE'then
         /* if not at the end of the record and the previously read record is written to the file
         , then go to next record */
         next_record;
    end if;
    /* save excell report */
    workbookwriter.save(dworkbook,global.gethome||dFileName);
    web.show_document('/users/'||dFileName,'_BLANK');
    /* when exceptions occurs */
    EXCEPTION
    WHEN ORA_JAVA.EXCEPTION_THROWN THEN
    begin
         javaException := ORA_JAVA.LAST_EXCEPTION;
         -- Print out the Exception by using the toString()
         -- Method of the exception Object
         javaException2 := Exception_.new(javaException);
         mess(27002,Exception_.getMessage(javaException2));
         -- and clean up
         ORA_JAVA.CLEAR_EXCEPTION;
    exception
         WHEN ORA_JAVA.JAVA_ERROR THEN
    -- In this case, the error would be
    -- Argument 1 can not be null
    mess(27002,ORA_JAVA.LAST_ERROR);
    --Clean up
    ORA_JAVA.CLEAR_ERROR;
    end;
    WHEN ORA_JAVA.JAVA_ERROR THEN
    -- In this case, the error would be
    -- Argument 1 can not be null
    message(ORA_JAVA.LAST_ERROR);
    --Clean up
    ORA_JAVA.CLEAR_ERROR;

    No need to double-post... most questions are answered pretty quickly...

  • Acrobat XI Pro "Out of Memory" Error.

    We just received a new Dell T7600 workstation (Win7, 64-bit, 64GB RAM, 8TB disk storage, and more processors than you can shake a stick at). We installed the Adobe CS6 Master Collection which had Acrobat Pro X. Each time we open a PDF of size greater than roughly 4MB, the program returns an "out of memory" error. After running updates, uninstalling and reinstalling (several times), I bought a copy of Acrobat XI Pro hoping this would solve the problem. Same problem still exists upon opening the larger PDFs. Our business depends on opening very large PDF files and we've paid for the Master Collection, so I'd rather not use an freeware PDF reader. Any help, thoughts, and/or suggestions are greatly appreciated.
    Regards,
    Chris

    As mentioned, the TEMP folder is typically the problem. MS limits the size of this folder and you have 2 choices: 1. empty it or 2. increase the size limit. I am not positive this is the issue, but it does crop up at times. It does not matter how big your harddrive is, it is a matter of the amount of space that MS has allocated for virtual memory. I am surprised that there is an issue with 64GB of RAM, but MS is real good at letting you know you can't have it all for use because you might want to open up something else. That is why a lot of big packages turn off some of the limits of Windows or use Linux.

  • RoboHelp 9 gives an out of memory error and crashes when I try to import or link a Frame 10 file or

    I have Tech Suite 3. If I start a new RoboHelp project and try to import or link Frame files, RoboHelp tries for a while, then I get an Out of Memory error and the program crashes.
    I opened one of the sample projects and was able to link to one of my frame files without any problem, so it seems to be an issue with creating something new.
    Any suggestions?

    It happens when I create a new project and then try to import or link frame docs to make up the content. It starts scanning, then crashes. I did get it to the conversion setting page once, but no further.
    It does not happen if I open one of the supplied example projects and link a file. But then it doesn't let me choose, during import, any style mapping. And I can't delete the sample project fold
    Twice now it has told me when I tried to import (not link, but import) that my .fm file could not be opened, and told me to verify that Frame is installed (it is) and that the file is a valid frame file (it is).
    The docs and project are in separate folders on my C: drive.

  • Adobe X Pro 10.1.10 Out of Memory error message

    Hello One, Hello All
    Since updating our Adobe X Pro machines to version update 10.1.10, we occasionally receive an "Out of Memory" error message, which depending on what we are doing, may force us to shutdown all Adobe windows and re-open Adobe, or simply click OK and continue working with the PDF document like nothing is wrong. It seems to have no rhyme or reason to when or why it occurs. I do not see any warnings or errors in the Windows Event Log. It happens with files we have created, and also happens with files received from internal and external sources through email. It is affecting our high end machines: Lenovo X Series and W Series laptops with Windows 8.1 3rd/4th gen i5 CPUs, 8+GB RAM, 128+GB SSD's. We have all system and Windows updates applied. We have Trend Micro and Malwarebytes real-time protection and system scans do not find any malware.
    I have found a few other recent threads on Adobe forum related to this error message but the responses are weak at best with no definitive fix. The system and user temp folders size are less than 100MB each so this cannot be the issue. When the error occurs I check task manager and system utilization, including memory, is well below 100%.
    We did not have this issue prior to version update 10.1.10.
    Really hoping Adobe can step in to help here and hoping boilerplate responses are not used.

    Hi wayne103,
    We released a new security update yesterday that is v10.1.11
    Please update to this version and check the performance.
    I have seen this error message occur while opening a 3rd party created pdf that has corrupted file structure.
    Please let me know if the issue occurs for some specific files or random
    Also let me know the PDF producer for these pdfs.
    Regards,
    Rave

  • Printing error due to out of memory condition...

    hi i got this problem of "printing error due to out of memory condition" and i cannot save it as a pdf file too. How to solve this prob? i using cs3 and acro 9.. pls help me... got a proj due and cannot afford to waste too much time...

    I've also seen this error message for the first time today "printing error due to out of memory condition" - I have 2GB of RAM, 300Gb+ free on my hard drive and have closed all other programs that were using some memory, however have had no problems printing to pdf ever before in Indesign... is there a solution to this problem?

  • Error compiling movie: out of memory !!! Premiere CS3

    Hey every1 --- desperate need of some guidance here. I've been all over the internet and these boards --- have seen this come up with a lot of people but no real solid answers that I'd seen or that have helped.
    First off, my comp is an imac 3.06 Ghz Core 2 Duo, 4GB memory. (just bought the computer last year)
    I'm shooting with a canon 7d in HD (1920 x 1080p) and I'm editing using Premiere CS3. I've noticed that editing has become very difficult since I started using HD; playback is very choppy and the 'quality settings' in the preview monitor are a joke and do absolutely nothing whether it's on draft or automatic.
    Anyway, despite the choppiness and slow editing, I was able to get a 3 min clip completed. These are my export settings:
    Export > Movie
    H.264
    960 x 540
    23.976 fps
    Square pixels,
    100% quality,
    recompress: unchecked.
    Once exported (I was able to export smaller timelines previously--- say, 50 secs of footage or so) the files look AMAZING and from there I use them for online video via vimeo -- stuff looks really awesome.
    Anyway, recently I'm getting: 'Error compiling movie. Out of Memory. Change optimization preferences to memory', etc. etc. Well, I did as it asked and that did absolutely nothing.
    So, I decided to open up my activity monitor and watch what's going on with my memory while premiere is exporting. I also closed ALL unused applications. System runs at about 2.16gb free memory before I start exporting. Memory seems stable during export UNTIL around the 50% exporting mark, and then its starts climbing like crazy and my memory just gets eaten up --- finally resulting in a crash at about 70% render.
    I really don't know what else to do. Does my computer not have the power to export 3 min of HD footage??? This seems surprising to me ... I don't know what else to try.
    Thanks for any help !!!
    -Mark

    Hi Mark,
    Hope you are having a wonderful day.
    Sorry to hear that you are having such a hard time. Although I did not specifically get the out of memory error, I can only be convinced that that is the problem.
    Although I am not a computer whiz, I have learned my share of things. I have been troubleshooting a video all this week (really, a couple of months), but I had to put it aside and come back to it.
    I tend to be talkative; so I will get back on topic or at least try to stay on topic, so bare with me. Smile....
    Here is what I have learned and maybe it will help you with yours if you have not found a solution already.
    Based on your hard drive specs and ram, I was thinking the same thing e.g. hard to believe that you are not able to render 3 minutes of video in HD. I primarily work in HD/1080p.
    The place where we are similar is that I eventually could get the video to render by reducing the workspace that is rendered piece by piece until I was able to render the entire file. When I tried to encode the video, I kept getting error compiling messages. I tried different codecs, deleting xmpses, not rendering xmpses, etc. I could not figure out the keyframe trick so I just put in a black video from after effects. By the way, all of my videos intro the same and no problems with the other videos fading from black to a still photo.
    Here is the problem from what I could gather.
    Simply, out of memory. I was rendering and trying to encode a 45 minute video with after effects compositions and the dynamic linking kept timing out once the file reached a certain size. There were no other true fancy effects. I do alot on the green screen and just use after effects to make a white background. Smaller videos, rendering has been slow, but doable. This massive video would never compile. I won't bore you with my system specs, but know that similar to your situation, the system in theory should have been adequate.
    Work around...
    I ended up exporting the video as 3 separate 15 minute video clips. I could not get them to encode as .wmv files so I encoded them as .mp4 files. By the way, I still encoded in HD.
    Here is the kool stuff.
    After what seems like forever, I found a converter that could convert .mp4 to .wmv and rejoin the files to play as one large file with the same quality as my source file and no audio synching issues. By the way, I do not work for this company. I just found this encoder at 4 o'clock this morning after trying all of these other converters that were straight garbage or were free to try and not really free if you wanted quality. By the way, the conversion tool is not only user friendly, but free which is awesome.
    The name of the software is Media Cope. Just visit www.mediacope.com.... Be sure to send the owner an e-mail to thank him and let him know that Mark sent you. I just sent him an e-mail earlier to let him know that I was going to post on Adobe what a wonderful software he created. I just needed something free, quick, easy, and user-friendly.
    Whoops... got off topic....
    I'll do a summary at the end for my recommendation to you.
    Here are some other tidbits that I learned as well that may save you some time.
    When editing this is what I do now...
    1) I use CCleaner from www.piriform.com to quickly clean my system especially temporary files before I start editing and if I am switching from project to project.
    2) In all of my troubleshooting this week, I found a kool little software call Smart Close. (Just google to find this one.) It closes all of your non-critical background files in order to maximize system resources. (When I am done editing, I just reboot and all of my background programs start up again. You do not have to do it that way, that's just how I do it. The nice thing is that it is a wizard interface; so you do not have to try to figure out if a process in the background is critical or non-critical.)
    3) Depending on my mood and if I did not forget, smile.... I change the priority to high for the adobe products in the task manager.
    More things that I learned...
    1) We are not able to change the temp file for how Adobe encodes. My logic was if I could change the encoding to a hard drive that was simply empty so Adobe can render there, then I would be good. Wrong.... Adobe does not allow you to change the temp location. (Note: This is what I discovered in the different forums; so encoding to an external hard drive is not the best move. Read more to se why.)
    2) It is best to have your scratch disks as your internal disk. If you have multiple internal disks, then you shoud check some other forums for the best way to set up your scratch disks. When I was just starting out, I bought an external hard drive enclosure with the thought that I would work primarily off of those external drives with the option of doing some RAID work. Bad idea... Even the simplest of tasks took forever. It was best to work off of the internal drive. By the way, I still use the enclosure hard drives for backups in order to allow me to free up some space on the primary hard drive. When rendering, it is truely a use of the RAM and the hard drive space in the editing process.
    I'll leave you with this...
    I take back what I said earlier, you do need a new computer with more RAM. If you can get 16 GB, great. The primary reason is that if you are editing in Premiere and if you do anything with dynamic linking into your after effects, it will cause your ram to climb as the file gets larger. Of course, be sure to get the largest hard drive possible as an internal drive, the maximum on processor speed, 64 bit operating system, etc.
    I know you are like by now, Mark, what is the point of all of your gibberish... Really, it was to share with you and others on some really kool work arounds.
    My recommendation for you are as follows.
    1) You have already proven that you can render and encode 50 second clips and you are very happy with those clips, so create separate smaller clips for this particular project.
    2) Once you have the final clips outputted, then use Media Cope to rejoin the files as one large file and you are done.
    (Note: The one thing that I did notice is that there is a small break right between where the files are merged. The nice thing is that as a video editor/film maker, you will be the only person likely to notice it because you know where the breaks are. The other nice thing is that the untrained eye will not notice it and if you had to submit your work to a television station especially on DVD, the video system should recognize the file as one large file and play all the way through. Note: For the last sentence, I said 'should play all the way through' as I have not had a chance to check this one yet as I just finished this project today and still have to turn in all of my videos. I still have some more videos to merge.)
    By the way, the only specs that I will share are that I work on a PC and it has 8 GB of ram.
    The summary of the summary is that sometimes an error compiling message could simply be that your system does not have resources to render and the encode that large of a file. At least that is what I learned this week or at least my specific situation. The confirmation is that I just finished shooting a documentary which was one hour and some change and was able to render and encode to .wmv with no problems. Of course, I was not using after effects for this particular project.
    In either case, hope all the information helps you with your project.
    Have a wonderful day and happy holidays everyone.
    Mark

  • Indesign cs5.5 out of memory error

    Hi,
    I am having constant problems with out of memory errors in indesign cs5.5 (version 7.5.3).  I am on a mac. Everything is up to date, I have preflight off, my display performance is on typical view and have tried restarted my computer multiple times.  I do not have any other programs open and the document I have open is 3 pages (8.5x11) with little content and a couple links to small images.  I do not know what to do!  This is affecting my job and my performance!! Please help!!!  Thank you!

    close the application
    Rename the preferences from the following location
    /Users/[User Name]/Library/Preferences/Adobe InDesign/Version #
    /Users/[User Name]/Library/Caches/Adobe InDesign/Version [#]
    then open a new document and then open the document you are facing an issue with it should clear the memory issue temporally
    Alternate :
    copy and paste the content in a new document and check if the 1st option does not work
    PS : changing the folder name would reset the preferences
    refer to KB : InDesign preferences and support file locations

  • Out of Memory Error when transcoding

    I am using winXP and Encore CS3. I don't have anything else open. My project is complete, but I cannot burn to DVD because I get this Out of Memory error, save and close to avoid loss.  I am trying to transcode the audio track for the DVD. I've made 3 different projects. On one of the projects, I just downloaded, it did transcode. All of my menu work is in another program. Is there a way to get this one transcoded file over to the other project. Then I could be finished and burn. I don't understand why this one transcode will not work. the other music I brought in as assets for the menu and submenu transcoded fine. Sometime it will say 'transcode preset error'. I went all the way back to Premiere and changed setting there in hopes. can anyone help please

    Thanks, disheartening? I'm distraught.
    Everything else works, but not Encore. I tried to reinstall, but it keeps saying it cannot install until I close Bridge.
    Bridge is not open.
    I restarted. I uninstalled an earlier verion of Bridge.
    Nothing doing.
    still wants me to close Bridge.
    I hate to do the whole thing over because of encore.
    Help me think of a workaround!

  • Out of Memory error on some PDF's, not all PDF's.

    Hi there,
    I have read most of the posts in the forums that relate to 'Out of Memory' issues with PDF's and I have to say that there is still no solution that I have found.
    I have tried reinstalling Adobe Reader, Flash Player and tried clearing my Temp Files. None of these fixed the issue.
    The PDF's that receive this memory error are downloaded off a CMS website that has many offices and logins. Only one office is experiencing this out of memory issue so we know that it is not an issue with generating the PDF's on the CMS website, otherwise all the offices and logins would have this issue, since they use the same system.
    I am an admin of that CMS website and even I receive the same out of memory issue when I try and view the PDF's from that office, as the customer does.
    When we open these PDF's with another PDF reading program, the PDF's open fine. So that tells me that the issue lies with Adobe Reader and not the PDF file itself.
    These PDF files that are receiving the error about 1MB-4MB, so they are not large.
    Could you please tell me the solution to this out of memory error as we may lose a customer because your product is producing an error that does not seem to have been fixed after years of being reported by your customers.
    Thanks.

    As an Adobe Reader XI user, if I may put my two cents in, that "Out of Memory" problem is not with the Adobe Reader XI application, which was installed on our computers! I think, the problem is with how the PDF documents, which we were trying to open, were generated (or regenerated).
    I remember I was able to open my old credit card statement online without any problem and now I am not able to open the same old statement (out of memory) because it could have been regenerated! In fact, I could not open any of my credit card statements (old or new) on this specific credit card web site, which makes me believe those statments were regenerated or somthing with some new Adobe software.
    As others mentioned, even I'm able to view other PDF documents without any problem.

  • Thread: Out of memory error

    I'm running a program that reads a file of more than 20k lines. Each line of this file is being processed by a thread. Sometimes the program is being halted with the out of memory error.
    What can I do to solve this problem ? Is it better to change the logic of this program and instead of calling a new thread for each line, call few threads to be responsible of processing multiple lines ?
    The current java version is 1.4.2.13. Upgrading it to the most recent version can solve the problem ?
    I'm running it with the following configuration:
    "-Xms1024m -Xmx2048m -XX urvivorRatio=10"
    Thanks a lot,
    Marcos.

    Indeed this design is flawed. It is an old system that came to my hands and now I have to fix some bugs on it.
    What is the difference between calling thousand of threads and having this backport of concurrent to take care of it ? I mean, isn't the jvm responsible for managing the both cases ?
    Using the backport of concurrent and defining a limited size thread pool, what will happen if the number of lines in the file is greater than the size of this pool ?
    Do you have an example or a link to where I can get a good example on how to use it ?
    Thanks for helping!

Maybe you are looking for