Importing CR content throws out of memory

Hi all.
While we are installing JDI server, there is step: import the SAP master Component Information.
the step requires us to import the CR content. But it throws out out of memory error.
How we can adjust the parameters to avoid this error ?

Hello Jianhong gu,
We used to face this error. The reason was that the heap size was not set correctly. Please refer to note 889038 for more info.
Also please do adhere to the hardware requirements for WAS and NWDI as specified in note 737368.
Regards
Sidharth

Similar Messages

  • XSOMParser throwing out of memory error

    Hello,
    Currently we are using XSOM parser with DomAnnotationParserFactory to parse XSD file. For small files it is working fine. However is was throwing out of memory error while parsing 9MB file. We could understood reason behind this. Is there any way to resolve this issue?
    Code :
         XSOMParser parser = new XSOMParser();
    parser.setAnnotationParser(new DomAnnotationParserFactory());
    XSSchemaSet schemaSet = null;
    XSSchema xsSchema = null;
    parser.parse(configFilePath);
    Here we are getting error on parser.parse() method. (using 128 MB heap memory using -Xrs -Xmx128m).
    Stack Trace :
    Exception in thread "main" java.lang.OutOfMemoryError: Java heap space
         at oracle.xml.parser.v2.XMLDocument.xdkIncCurrentId(XMLDocument.java:3020)
         at oracle.xml.parser.v2.XMLNode.xdkInit(XMLNode.java:2758)
         at oracle.xml.parser.v2.XMLNode.<init>(XMLNode.java:423)
         at oracle.xml.parser.v2.XMLNSNode.<init>(XMLNSNode.java:144)
         at oracle.xml.parser.v2.XMLElement.<init>(XMLElement.java:373)
         at oracle.xml.parser.v2.XMLDocument.createNodeFromType(XMLDocument.java:2865)
         at oracle.xml.parser.v2.XMLDocument.createElement(XMLDocument.java:1896)
         at oracle.xml.parser.v2.DocumentBuilder.startElement(DocumentBuilder.java:224)
         at oracle.xml.parser.v2.XMLElement.reportStartElement(XMLElement.java:3188)
         at oracle.xml.parser.v2.XMLElement.reportSAXEvents(XMLElement.java:2164)
         at oracle.xml.jaxp.JXTransformer.transform(JXTransformer.java:337)
         at oracle.xml.jaxp.JXTransformerHandler.endDocument(JXTransformerHandler.java:141)
         at com.sun.xml.xsom.impl.parser.state.NGCCRuntime.endElement(NGCCRuntime.java:267)
         at org.xml.sax.helpers.XMLFilterImpl.endElement(Unknown Source)
         at oracle.xml.parser.v2.NonValidatingParser.parseElement(NonValidatingParser.java:1257)
         at oracle.xml.parser.v2.NonValidatingParser.parseRootElement(NonValidatingParser.java:314)
         at oracle.xml.parser.v2.NonValidatingParser.parseDocument(NonValidatingParser.java:281)
         at oracle.xml.parser.v2.XMLParser.parse(XMLParser.java:196)
         at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source)
         at com.sun.xml.xsom.parser.JAXPParser.parse(JAXPParser.java:79)
         at com.sun.xml.xsom.impl.parser.NGCCRuntimeEx.parseEntity(NGCCRuntimeEx.java:298)
         at com.sun.xml.xsom.impl.parser.ParserContext.parse(ParserContext.java:87)
         at com.sun.xml.xsom.parser.XSOMParser.parse(XSOMParser.java:147)
         at com.sun.xml.xsom.parser.XSOMParser.parse(XSOMParser.java:136)
         at com.sun.xml.xsom.parser.XSOMParser.parse(XSOMParser.java:129)
         at com.sun.xml.xsom.parser.XSOMParser.parse(XSOMParser.java:122)
    Please let me know if anyone has comment on this.
    Also let me know if there any other parser which handles large input files efficiently.

    Hello,
    Currently we are using XSOM parser with DomAnnotationParserFactory to parse XSD file. For small files it is working fine. However is was throwing out of memory error while parsing 9MB file. We could understood reason behind this. Is there any way to resolve this issue?
    Code :
         XSOMParser parser = new XSOMParser();
    parser.setAnnotationParser(new DomAnnotationParserFactory());
    XSSchemaSet schemaSet = null;
    XSSchema xsSchema = null;
    parser.parse(configFilePath);
    Here we are getting error on parser.parse() method. (using 128 MB heap memory using -Xrs -Xmx128m).
    Stack Trace :
    Exception in thread "main" java.lang.OutOfMemoryError: Java heap space
         at oracle.xml.parser.v2.XMLDocument.xdkIncCurrentId(XMLDocument.java:3020)
         at oracle.xml.parser.v2.XMLNode.xdkInit(XMLNode.java:2758)
         at oracle.xml.parser.v2.XMLNode.<init>(XMLNode.java:423)
         at oracle.xml.parser.v2.XMLNSNode.<init>(XMLNSNode.java:144)
         at oracle.xml.parser.v2.XMLElement.<init>(XMLElement.java:373)
         at oracle.xml.parser.v2.XMLDocument.createNodeFromType(XMLDocument.java:2865)
         at oracle.xml.parser.v2.XMLDocument.createElement(XMLDocument.java:1896)
         at oracle.xml.parser.v2.DocumentBuilder.startElement(DocumentBuilder.java:224)
         at oracle.xml.parser.v2.XMLElement.reportStartElement(XMLElement.java:3188)
         at oracle.xml.parser.v2.XMLElement.reportSAXEvents(XMLElement.java:2164)
         at oracle.xml.jaxp.JXTransformer.transform(JXTransformer.java:337)
         at oracle.xml.jaxp.JXTransformerHandler.endDocument(JXTransformerHandler.java:141)
         at com.sun.xml.xsom.impl.parser.state.NGCCRuntime.endElement(NGCCRuntime.java:267)
         at org.xml.sax.helpers.XMLFilterImpl.endElement(Unknown Source)
         at oracle.xml.parser.v2.NonValidatingParser.parseElement(NonValidatingParser.java:1257)
         at oracle.xml.parser.v2.NonValidatingParser.parseRootElement(NonValidatingParser.java:314)
         at oracle.xml.parser.v2.NonValidatingParser.parseDocument(NonValidatingParser.java:281)
         at oracle.xml.parser.v2.XMLParser.parse(XMLParser.java:196)
         at org.xml.sax.helpers.XMLFilterImpl.parse(Unknown Source)
         at com.sun.xml.xsom.parser.JAXPParser.parse(JAXPParser.java:79)
         at com.sun.xml.xsom.impl.parser.NGCCRuntimeEx.parseEntity(NGCCRuntimeEx.java:298)
         at com.sun.xml.xsom.impl.parser.ParserContext.parse(ParserContext.java:87)
         at com.sun.xml.xsom.parser.XSOMParser.parse(XSOMParser.java:147)
         at com.sun.xml.xsom.parser.XSOMParser.parse(XSOMParser.java:136)
         at com.sun.xml.xsom.parser.XSOMParser.parse(XSOMParser.java:129)
         at com.sun.xml.xsom.parser.XSOMParser.parse(XSOMParser.java:122)
    Please let me know if anyone has comment on this.
    Also let me know if there any other parser which handles large input files efficiently.

  • Uploading large files from applet to servlet throws out of memory error

    I have a java applet that needs to upload files from a client machine
    to a web server using a servlet. the problem i am having is that in
    the current scheme, files larger than 17-20MB throw an out of memory
    error. is there any way we can get around this problem? i will post
    the client and server side code for reference.
    Client Side Code:
    import java.io.*;
    import java.net.*;
    // this class is a client that enables transfer of files from client
    // to server. This client connects to a servlet running on the server
    // and transmits the file.
    public class fileTransferClient
    private static final String FILENAME_HEADER = "fileName";
    private static final String FILELASTMOD_HEADER = "fileLastMod";
    // this method transfers the prescribed file to the server.
    // if the destination directory is "", it transfers the file to
    "d:\\".
    //11-21-02 Changes : This method now has a new parameter that
    references the item
    //that is being transferred in the import list.
    public static String transferFile(String srcFileName, String
    destFileName,
    String destDir, int itemID)
    if (destDir.equals(""))
    destDir = "E:\\FTP\\incoming\\";
    // get the fully qualified filename and the mere filename.
    String fqfn = srcFileName;
    String fname =
    fqfn.substring(fqfn.lastIndexOf(File.separator)+1);
    try
    //importTable importer = jbInit.getImportTable();
    // create the file to be uploaded and a connection to
    servlet.
    File fileToUpload = new File(fqfn);
    long fileSize = fileToUpload.length();
    // get last mod of this file.
    // The last mod is sent to the servlet as a header.
    long lastMod = fileToUpload.lastModified();
    String strLastMod = String.valueOf(lastMod);
    URL serverURL = new URL(webadminApplet.strServletURL);
    URLConnection serverCon = serverURL.openConnection();
    // a bunch of connection setup related things.
    serverCon.setDoInput(true);
    serverCon.setDoOutput(true);
    // Don't use a cached version of URL connection.
    serverCon.setUseCaches (false);
    serverCon.setDefaultUseCaches (false);
    // set headers and their values.
    serverCon.setRequestProperty("Content-Type",
    "application/octet-stream");
    serverCon.setRequestProperty("Content-Length",
    Long.toString(fileToUpload.length()));
    serverCon.setRequestProperty(FILENAME_HEADER, destDir +
    destFileName);
    serverCon.setRequestProperty(FILELASTMOD_HEADER, strLastMod);
    if (webadminApplet.DEBUG) System.out.println("Connection with
    FTP server established");
    // create file stream and write stream to write file data.
    FileInputStream fis = new FileInputStream(fileToUpload);
    OutputStream os = serverCon.getOutputStream();
    try
    // transfer the file in 4K chunks.
    byte[] buffer = new byte[4096];
    long byteCnt = 0;
    //long percent = 0;
    int newPercent = 0;
    int oldPercent = 0;
    while (true)
    int bytes = fis.read(buffer);
    byteCnt += bytes;
    //11-21-02 :
    //If itemID is greater than -1 this is an import file
    transfer
    //otherwise this is a header graphic file transfer.
    if (itemID > -1)
    newPercent = (int) ((double) byteCnt/ (double)
    fileSize * 100.0);
    int diff = newPercent - oldPercent;
    if (newPercent == 0 || diff >= 20)
    oldPercent = newPercent;
    jbInit.getImportTable().displayFileTransferStatus
    (itemID,
    newPercent);
    if (bytes < 0) break;
    os.write(buffer, 0, bytes);
    os.flush();
    if (webadminApplet.DEBUG) System.out.println("No of bytes
    sent: " + byteCnt);
    finally
    // close related streams.
    os.close();
    fis.close();
    if (webadminApplet.DEBUG) System.out.println("File
    Transmission complete");
    // find out what the servlet has got to say in response.
    BufferedReader reader = new BufferedReader(
    new
    InputStreamReader(serverCon.getInputStream()));
    try
    String line;
    while ((line = reader.readLine()) != null)
    if (webadminApplet.DEBUG) System.out.println(line);
    finally
    // close the reader stream from servlet.
    reader.close();
    } // end of the big try block.
    catch (Exception e)
    System.out.println("Exception during file transfer:\n" + e);
    e.printStackTrace();
    return("FTP failed. See Java Console for Errors.");
    } // end of catch block.
    return("File: " + fname + " successfully transferred.");
    } // end of method transferFile().
    } // end of class fileTransferClient
    Server side code:
    import java.io.*;
    import javax.servlet.*;
    import javax.servlet.http.*;
    import java.util.*;
    import java.net.*;
    // This servlet class acts as an FTP server to enable transfer of
    files
    // from client side.
    public class FtpServerServlet extends HttpServlet
    String ftpDir = "D:\\pub\\FTP\\";
    private static final String FILENAME_HEADER = "fileName";
    private static final String FILELASTMOD_HEADER = "fileLastMod";
    public void doGet(HttpServletRequest req, HttpServletResponse resp)
    throws ServletException,
    IOException
    doPost(req, resp);
    public void doPost(HttpServletRequest req, HttpServletResponse
    resp)
    throws ServletException,
    IOException
    // ### for now enable overwrite by default.
    boolean overwrite = true;
    // get the fileName for this transmission.
    String fileName = req.getHeader(FILENAME_HEADER);
    // also get the last mod of this file.
    String strLastMod = req.getHeader(FILELASTMOD_HEADER);
    String message = "Filename: " + fileName + " saved
    successfully.";
    int status = HttpServletResponse.SC_OK;
    System.out.println("fileName from client: " + fileName);
    // if filename is not specified, complain.
    if (fileName == null)
    message = "Filename not specified";
    status = HttpServletResponse.SC_INTERNAL_SERVER_ERROR;
    else
    // open the file stream for the file about to be transferred.
    File uploadedFile = new File(fileName);
    // check if file already exists - and overwrite if necessary.
    if (uploadedFile.exists())
    if (overwrite)
    // delete the file.
    uploadedFile.delete();
    // ensure the directory is writable - and a new file may be
    created.
    if (!uploadedFile.createNewFile())
    message = "Unable to create file on server. FTP failed.";
    status = HttpServletResponse.SC_INTERNAL_SERVER_ERROR;
    else
    // get the necessary streams for file creation.
    FileOutputStream fos = new FileOutputStream(uploadedFile);
    InputStream is = req.getInputStream();
    try
    // create a buffer. 4K!
    byte[] buffer = new byte[4096];
    // read from input stream and write to file stream.
    int byteCnt = 0;
    while (true)
    int bytes = is.read(buffer);
    if (bytes < 0) break;
    byteCnt += bytes;
    // System.out.println(buffer);
    fos.write(buffer, 0, bytes);
    // flush the stream.
    fos.flush();
    } // end of try block.
    finally
    is.close();
    fos.close();
    // set last mod date for this file.
    uploadedFile.setLastModified((new
    Long(strLastMod)).longValue());
    } // end of finally block.
    } // end - the new file may be created on server.
    } // end - we have a valid filename.
    // set response headers.
    resp.setContentType("text/plain");
    resp.setStatus(status);
    if (status != HttpServletResponse.SC_OK)
    getServletContext().log("ERROR: " + message);
    // get output stream.
    PrintWriter out = resp.getWriter();
    out.println(message);
    } // end of doPost().
    } // end of class FtpServerServlet

    OK - the problem you describe is definitely what's giving you grief.
    The workaround is to use a socket connection and send your own request headers, with the content length filled in. You may have to multi-part mime encode the stream on its way out as well (I'm not about that...).
    You can use the following:
    http://porsche.cis.udel.edu:8080/cis479/lectures/slides-04/slide-02.html
    on your server to get a feel for the format that the request headers need to take.
    - Kevin
    I get the out of Memory Error on the client side. I
    was told that this might be a bug in the URLConnection
    class implementation that basically it wont know the
    content length until all the data has been written to
    the output stream, so it uses an in memory buffer to
    store the data which basically causes memory issues..
    do you think there might be a workaround of any kind..
    or maybe a way that the buffer might be flushed after
    a certain size of file has been uploaded.. ?? do you
    have any ideas?

  • URLStream.readBytes always throw out of memory exception (errorID=1000)

    When I try to load a file as 180MBytes by using URLStream.readBytes(). 
    In some PCs, it's OK.
    But in some PCs, there are always [out of memory] exception (errorID=1000) even such PC still had enough memory.
    For example:
       Total memory is 2G, current used is 1.43G, but URLStream.readBytes() still throw such exception.
    For file with little size such as 30M, there isn't such problem.
    Could any body give some suggestion?
    Best regards,
    Sourcecode is very simple like:
      var myStream:URLStream;
      var inputBytes: ByteArray = new ByteArray();
      ... do load ...
      // load by progressive event. I only deal progress load once, and ignore all later progressive load events.
      // for example first 500KBytes was read.
      myStream(inputBytes,inputBytes.length);
      // load again when all datas is loaded completely. For example, 180MBytes should be read.
      try {
        myStream(inputBytes,inputBytes.length);
      }  catch(e:*) {
        // warning log

    When I try to load a file as 180MBytes by using URLStream.readBytes(). 
    In some PCs, it's OK.
    But in some PCs, there are always [out of memory] exception (errorID=1000) even such PC still had enough memory.
    For example:
       Total memory is 2G, current used is 1.43G, but URLStream.readBytes() still throw such exception.
    For file with little size such as 30M, there isn't such problem.
    Could any body give some suggestion?
    Best regards,
    Sourcecode is very simple like:
      var myStream:URLStream;
      var inputBytes: ByteArray = new ByteArray();
      ... do load ...
      // load by progressive event. I only deal progress load once, and ignore all later progressive load events.
      // for example first 500KBytes was read.
      myStream(inputBytes,inputBytes.length);
      // load again when all datas is loaded completely. For example, 180MBytes should be read.
      try {
        myStream(inputBytes,inputBytes.length);
      }  catch(e:*) {
        // warning log

  • Openscript throws out of memory exception

    Hi,
    I see the following exceptions very frequently while using open script 12.3.0.1:
    java.lang.OutOfMemoryError: Java heap space
    java.lang.OutOfMemoryError: PermGen space
    I have increases the vmargs values in the eclipse.ini of OracleATS/openscript folder where OATS is installed on my machine.
    -startup
    plugins/org.eclipse.equinox.launcher_1.0.101.R34x_v20080819.jar
    -showsplash
    org.eclipse.platform
    --launcher.XXMaxPermSize
    2048m
    -vmargs
    -Xms512m
    -Xmx4096m
    -XX:MaxPermSize=2048m
    I have also tried two other ways like setting JAVA_OPTS in environment variables and heap space in the control panel for java, but nothing worked for me.
    Note: I use windows7.
    Thanks,
    suseela.

    Hi Guys,
    Will PreparedStatement.addBatch() throw Out of
    f Memory EXception??It depends on how many objects are created during that method's execution and how much heap is left at the time you call it.
    Current i m populate 3000 update statement and
    nd using addBatch() method and execute at the end and
    it works fine for me...
    I m just wondering what if i have like 100.000
    00 record or something??
    Any Idea??It depends.
    Try it and see.
    And if it works one time, don't assume it will next time, unless your code and its data are very regular.

  • Download Servlet throwing Out Of Memory Exception

    I am trying to download file of more than 500 mb through servlet but getting out of memory exception .
    Before downloading i am zipping that huge file .
    try {
         String zipFileName = doZip(file);
          file =null;
          System.gc();
          File inputFile = new File(zipFileName);
          InputStream fileToDownload = new FileInputStream(
                                            inputFile);
           response.setContentType("application/zip");
         response.setHeader("Content-Disposition","attachment; filename=\""
                                                      + fileName.replaceAll("tmx", "zip")
                                                                .concat("\""));
         response.setContentLength(fileToDownload.available());
         byte buf[] = new byte[BUF_SIZE];
         int read;
         while ((read = fileToDownload.read(buf)) != -1) {
                   outs.write(buf, 0, read);
              fileToDownload.close();
                   outs.flush();
                                  outs.close();
    }catch(Exception e ) {
      //Getting out of memory.
    }Please suggest solution for this .

    cotton.m wrote:
    My zip suggestion was as follows.
    Take the file. Do not set the Content length header. Do set the Content encoding header to gzip. Create a GZIP output stream using the servlet output stream. Read the unzipped file in and output it through the gzip output stream.
    This cuts out one full cycle of file reading and writing from what you are doing currently.Thanks for u r reply
    InputStream fileToDownload = new FileInputStream(
                                            file);
    response.setContentType("application/gzip");
    response.setHeader("Transfer-Encoding", "chunked");
    response.setContentLength((int) file.length());
    GZIPOutputStream gzipoutputstream = new GZIPOutputStream(outs);
    byte buf[] = new byte[BUF_SIZE];
    int read;
    while ((read = fileToDownload.read(buf)) != -1) {
         gzipoutputstream.write(buf, 0, read);
    fileToDownload.close();
    outs.flush();
    outs.close();I made changes accordingly . Please provide u r view on this .

  • SAP AS JAVA installation in Solaris Zone envirn throws OUT of Memory error.

    Hi,
    We are installing SAP NW2004s 7.0 SR3 AS JAVA on Solaris 10 in zone environment. Thsi was prod server getting build where on top of this we will install GRC 5.3.
    We have faced no issues in development build.
    But during the Prod build at databse create step, we are getting the below error
    ORA-27102: out of memory
    SVR4 Error: 22: Invalid argument
    Disconnected
    SAPinst log entries where error messages started:
    04:43:58.128
    Execute step createDatabase of component |NW_Onehost|ind|ind|ind|ind|0|0|NW_Onehost_System|ind|ind|ind|ind|2|0|NW_CreateDBandLoad|ind|ind|ind|ind|10|0|NW_CreateDB|ind|ind|ind|ind|0|0|NW_OraDBCheck|ind|ind|ind|ind|0|0|NW_OraDBMain|ind|ind|ind|ind|0|0|NW_OraDBStd|ind|ind|ind|ind|3|0|NW_OraDbBuild|ind|ind|ind|ind|5|0
    INFO 2011-04-01 04:45:14.590
    Working directory changed to /tmp/sapinst_exe.16718.1301647358.
    INFO 2011-04-01 04:45:14.595
    Working directory changed to /tmp/sapinst_instdir/NW04S/SYSTEM/ORA/CENTRAL/AS.
    INFO 2011-04-01 04:45:14.609
    Working directory changed to /tmp/sapinst_exe.16718.1301647358.
    INFO 2011-04-01 04:45:14.621
    Working directory changed to /tmp/sapinst_instdir/NW04S/SYSTEM/ORA/CENTRAL/AS.
    INFO 2011-04-01 04:45:14.850
    Account oraac5 already exists.
    INFO 2011-04-01 04:45:14.852
    Account dba already exists.
    INFO 2011-04-01 04:45:14.852
    Account oraac5 already exists.
    INFO 2011-04-01 04:45:14.853
    Account dba already exists.
    INFO 2011-04-01 04:45:14.867
    Working directory changed to /tmp/sapinst_exe.16718.1301647358.
    INFO 2011-04-01 04:45:14.899
    Working directory changed to /tmp/sapinst_instdir/NW04S/SYSTEM/ORA/CENTRAL/AS.
    ERROR 2011-04-01 04:45:32.280
    CJS-00084  SQL statement or script failed. DIAGNOSIS: Error message: ORA-27102: out of memory
    SVR4 Error: 22: Invalid argument
    Disconnected
    . SOLUTION: See ora_sql_results.log and the Oracle documentation for details.
    ERROR 2011-04-01 04:45:32.286
    MUT-03025  Caught ESAPinstException in Modulecall: ORA-27102: out of memory
    SVR4 Error: 22: Invalid argument
    Disconnected
    ERROR 2011-04-01 04:45:32.453
    FCO-00011  The step createDatabase with step key |NW_Onehost|ind|ind|ind|ind|0|0|NW_Onehost_System|ind|ind|ind|ind|2|0|NW_CreateDBandLoad|ind|ind|ind|ind|10|0|NW_CreateDB|ind|ind|ind|ind|0|0|NW_OraDBCheck|ind|ind|ind|ind|0|0|NW_OraDBMain|ind|ind|ind|ind|0|0|NW_OraDBStd|ind|ind|ind|ind|3|0|NW_OraDbBuild|ind|ind|ind|ind|5|0|createDatabase was executed with status ERROR ( Last error reported by the step :Caught ESAPinstException in Modulecall: ORA-27102: out of memory
    SVR4 Error: 22: Invalid argument
    Disconnected
    ora_sql_results.log
    04:45:15 SAPINST ORACLE start logging for
    SHUTDOWN IMMEDIATE;
    exit;
    Output of SQL executing program:
    SQL*Plus: Release 10.2.0.4.0 - Production on Fri Apr 1 04:45:15 2011
    Copyright (c) 1982, 2007, Oracle.  All Rights Reserved.
    Connected to an idle instance.
    ORA-01034: ORACLE not available
    ORA-27101: shared memory realm does not exist
    SVR4 Error: 2: No such file or directory
    Disconnected
    SAPINST: End of output of SQL executing program /oracle/AC5/102_64/bin/sqlplus.
    SAPINST found errors.
    SAPINST The current process environment may be found in sapinst_ora_environment.log.
    2011-04-01, 04:45:15 SAPINST ORACLE stop logging
    ================================================================================
    2011-04-01, 04:45:15 SAPINST ORACLE start logging for
    STARTUP NOMOUNT;
    exit;
    Output of SQL executing program:
    SQL*Plus: Release 10.2.0.4.0 - Production on Fri Apr 1 04:45:15 2011
    Copyright (c) 1982, 2007, Oracle.  All Rights Reserved.
    Connected to an idle instance.
    ORA-27102: out of memory
    SVR4 Error: 22: Invalid argument
    Disconnected
    SAPINST: End of output of SQL executing program /oracle/AC5/102_64/bin/sqlplus.
    SAPINST found errors.
    SAPINST The current process environment may be found in sapinst_ora_environment.log.
    2011-04-01, 04:45:32 SAPINST ORACLE stop logging
    Already viewed S-note
    724713 - parameter settings for Solaris 10 - (Parameters are set as per this note)
    743328 - Composite SAP note: ORA-27102 - (no much information in memory problem on zones)
    Please provide your suggestions/resolution.
    Thankyou.

    Hi,
    @ Sunny: Thanks for response, the referred note was already checked and parameters are in sync as per note.
    @Mohit: SAP wouldn't proceed till create database if Oracle software was not installed. thanks for the response.
    @Markus: Thanks I agree with you, but I have doubt in this area. Isn't it like project.max-shm-memory was new parameter we need to set in local zone rather using shmsys:shminfo_shmmax in /etc/system. Do we need to still maintain this parameter in /etc/system in global zone.
    As per SUN doc below parameter was obsolete from Solaris 10.
    The following parameters are obsolete.
    ■ shmsys:shminfo_shmmni
    ■ shmsys:shminfo_shmmax
    As per your suggestion, do we need to set below parameters in that case, please clarify.
    Parameter                           Replaced by Resource Control      Recommended Value
    semsys:seminfo_semmni   project.max-sem-ids                      100
    semsys:seminfo_semmsl   process.max-sem-nsems               256
    shmsys:shminfo_shmmax  project.max-shm-memory             4294967295
    shmsys:shminfo_shmmni   project.max-shm-ids                       100
    Also   findings of /etc/release
    more /etc/release
    Solaris 10 10/08 s10s_u6wos_07b SPARC
    Regards,
    Sitarama.

  • Import Manager Out of Memory error

    I am running MDM 5.5.24.06 server on Windows XP Professional. The server has 4.0 GB RAM with 1.5 GB Virtual memory.
    I am trying to load 129 material master records from R/3 4.6 ( XML file size 8 MB), into MDM using Import Manager.
    When I click on import icon (Import Status tab selected), MDM Import Manager rightway returns 'Out of Memory' error.
    Tried to import the file again after rebooting the machine, restarting MDM Server, but still it returned the same error.
    Has anybody tried loading the R/3 material data into MDM? What's the experience? What was the load (# of records)?
    What was MDM IM performance? What's your hardware/software config?
    Appreciate your help.
    Thanks,
    Abhay

    Hi Abhay,
    My workstation has Windows 2000, Pentium 4 CPU 3.2 Ghz, 515,555 KB RAM.  The recommended specifications can be found on service market place of SAP, and we used those to procure our workstations.
    MDM server must be more pwerful than a workstation.  Try running the import manager from the server machine itself.  If it works there but not on your machine, it is memory of your workstation.
    Sometimes, I think this error comes when import map some issues- fields not mapped accurately or remote key mapped wrong etc.
    It is probably judicious to rpove the system memory as first step.
    Good luck!

  • Large Pdf using XML XSL - Out of Memory Error

    Hi Friends.
    I am trying to generate a PDF from XML, XSL and FO in java. It works fine if the PDF to be generated is small.
    But if the PDF to be generated is big, then it throws "Out of Memory" error. Can some one please give me some pointers about the possible reasons for this errors. Thanks for your help.
    RM
    Code:
    import java.io.*;
    import javax.servlet.*;
    import javax.servlet.http.*;
    import org.xml.sax.InputSource;
    import org.xml.sax.XMLReader;
    import org.apache.fop.apps.Driver;
    import org.apache.fop.apps.Version;
    import org.apache.fop.apps.XSLTInputHandler;
    import org.apache.fop.messaging.MessageHandler;
    import org.apache.avalon.framework.logger.ConsoleLogger;
    import org.apache.avalon.framework.logger.Logger;
    public class PdfServlet extends HttpServlet {
    public static final String FO_REQUEST_PARAM = "fo";
    public static final String XML_REQUEST_PARAM = "xml";
    public static final String XSL_REQUEST_PARAM = "xsl";
    Logger log = null;
         Com_BUtil myBu = new Com_BUtil();
    public void doGet(HttpServletRequest request,
    HttpServletResponse response) throws ServletException {
    if(log == null) {
         log = new ConsoleLogger(ConsoleLogger.LEVEL_WARN);
         MessageHandler.setScreenLogger(log);
    try {
    String foParam = request.getParameter(FO_REQUEST_PARAM);
    String xmlParam = myBu.getConfigVal("filePath") +"/"+request.getParameter(XML_REQUEST_PARAM);
    String xslParam = myBu.SERVERROOT + "/jsp/servlet/"+request.getParameter(XSL_REQUEST_PARAM)+".xsl";
         if((xmlParam != null) && (xslParam != null)) {
    XSLTInputHandler input = new XSLTInputHandler(new File(xmlParam), new File(xslParam));
    renderXML(input, response);
    } else {
    PrintWriter out = response.getWriter();
    out.println("<html><head><title>Error</title></head>\n"+
    "<body><h1>PdfServlet Error</h1><h3>No 'fo' "+
    "request param given.</body></html>");
    } catch (ServletException ex) {
    throw ex;
    catch (Exception ex) {
    throw new ServletException(ex);
    public void renderXML(XSLTInputHandler input,
    HttpServletResponse response) throws ServletException {
    try {
    ByteArrayOutputStream out = new ByteArrayOutputStream();
    response.setContentType("application/pdf");
    Driver driver = new Driver();
    driver.setLogger(log);
    driver.setRenderer(Driver.RENDER_PDF);
    driver.setOutputStream(out);
    driver.render(input.getParser(), input.getInputSource());
    byte[] content = out.toByteArray();
    response.setContentLength(content.length);
    response.getOutputStream().write(content);
    response.getOutputStream().flush();
    } catch (Exception ex) {
    throw new ServletException(ex);
    * creates a SAX parser, using the value of org.xml.sax.parser
    * defaulting to org.apache.xerces.parsers.SAXParser
    * @return the created SAX parser
    static XMLReader createParser() throws ServletException {
    String parserClassName = System.getProperty("org.xml.sax.parser");
    if (parserClassName == null) {
    parserClassName = "org.apache.xerces.parsers.SAXParser";
    try {
    return (XMLReader) Class.forName(
    parserClassName).newInstance();
    } catch (Exception e) {
    throw new ServletException(e);

    Hi,
    I did try that initially. After executing the command I get this message.
    C:\>java -Xms128M -Xmx256M
    Usage: java [-options] class [args...]
    (to execute a class)
    or java -jar [-options] jarfile [args...]
    (to execute a jar file)
    where options include:
    -cp -classpath <directories and zip/jar files separated by ;>
    set search path for application classes and resources
    -D<name>=<value>
    set a system property
    -verbose[:class|gc|jni]
    enable verbose output
    -version print product version and exit
    -showversion print product version and continue
    -? -help print this help message
    -X print help on non-standard options
    Thanks for your help.
    RM

  • Acrobat XI Pro "Out of Memory" error after Office 2010 install

    Good Afternoon,
    We recently pushed Office 2010 to our users and are now getting reports of previous installs of Adobe Acrobat XI Pro no longer working but throwing "Out of Memory" errors.
    We are in a Windows XP environment. All machines are HP 8440p/6930p/6910 with the same Service pack level (3) and all up to date on security patches.
    All machines are running Office 2010 SP1.
    All machines have 2GB or 4GB of RAM (Only 3.25GB recognized as we are a 32bit OS environment).
    All machines have adequate free space (ranging from 50gb to 200gb of free space).
    All machines are set to 4096mb initial page file size with 8192mb maximum page file size.
    All machines with Acrobat XI Pro *DO NOT* have Reader XI installed alongside. If Reader is installed, it is Reader 10.1 or higher.
    The following troubleshooting steps have been taken:
    Verify page file size (4096mb - 8192mb).
    Deleted local user and Windows temp files (%temp% and c:\WINDOWS\Temp both emptied).
    Repair on Adobe Acrobat XI Pro install. No change.
    Uninstall Acrobat Pro XI, reboot, re-install. No change.
    Uninstall Acrobat Pro XI Pro along with *ALL* other Adobe applications presently installed (Flash Player, Air), delete all Adobe folders and files found in a full search of the C drive, delete all orphaned Registry entries for all Adobe products, re-empty all temp folders, reboot.
    Re-install Adobe Acrobat XI Pro. No change.
    Disable enhanced security in Acrobat XI Pro. No change.
    Renamed Acrobat XI's plug_ins folder to plug_ins.old.
    You *can* get Acrobat to open once this is done but when you attempt to edit a file or enter data into a form, you get the message, "The "Updater" plug-in has been removed. Please re-install Acrobat to continue viewing the current file."
    A repair on the Office 2010 install and re-installing Office 2010 also had no effect.
    At this point, short of re-imaging the machines (which is *not* an option), we are stumped.
    We have not yet tried rolling back a user to Office 2007 as the upgrade initiative is enterprise-wide and rolling back would not be considered a solution.
    Anyone have any ideas beyond what has been tried so far?

    As mentioned, the TEMP folder is typically the problem. MS limits the size of this folder and you have 2 choices: 1. empty it or 2. increase the size limit. I am not positive this is the issue, but it does crop up at times. It does not matter how big your harddrive is, it is a matter of the amount of space that MS has allocated for virtual memory. I am surprised that there is an issue with 64GB of RAM, but MS is real good at letting you know you can't have it all for use because you might want to open up something else. That is why a lot of big packages turn off some of the limits of Windows or use Linux.

  • Can't place tiff or eps, can't export pdf -  "out of memory" masage

    Hi, I have have CS2 prem, have problems with cmap files in Indesign and resolve them but now after installing again and starting I can import jpg but can't import tiff or eps "out of memory".
    When opening an old document display performance are typical but some of the images look typical as they must be but some look like fast preview even in tipical mode.
    Have enought memory - check already. Also Indesign works correct ... 2 years on this computer.
    So, i can't use the program for finel projects )))
    Can anyone help me?

    Answered in the other thread.

  • Out of Memory Error in Oracle 8i

    When i ty to fetch data from view it throws "Out OF Memory Error",Normally it sholud return 300000 records as output uisng joing condition.
    Any body please help me as this is Bit Urgent ....
    Thanks.

    user7725408 wrote:
    When i ty to fetch data from view it throws "Out OF Memory Error",Normally it sholud return 300000 records as output uisng joing condition.
    Any body please help me as this is Bit Urgent ....
    Thanks.There are several specific "out of memory" errors. How about an acutal error code (ORA-nnnnn). Is this on a 32-bit Windows system?

  • Out of Memory error while builng HTML String from a Large HashMap.

    Hi,
    I am building an HTML string from a large map oject that consits of about 32000 objects using the Transformer class in java. As this HTML string needs to be displayed in the JSP page, the reponse time was too high and also some times it is throwing out of memory error.
    Please let me know how i can implement the concept of building the library tree(folder structure) HTML string for the first set of say 1000 entries and then display in the web page and then detect an onScroll event and handle it in java Script functions and come back and build the tree for the next set of entries in the map and append this string to the previous one and accordingly display it.
    please let me know whether
    1. the suggested solution was the advisable one.
    2. how to build tree(HTML String) for a set of entries in the map while iterating over the map.
    3. How to detect a onScroll event and handle it.
    Note : Handling the events in the JavaScript functions and displaying the tree is now being done using AJAX.
    Thanks for help in Advance,
    Kartheek

    Hi
    Sorry,
    I haven't seen any error in the browser as this may be Out of memory error which was not handled. I got the the following error from the web logic console
    org.apache.struts.actions.DispatchAction">Dispatch[serviceCenterHome] to method 'getUserLibraryTree' returned an exceptionjava.lang.reflect.InvocationTargetException
         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
         at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
         at java.lang.reflect.Method.invoke(Method.java:324)
         at org.apache.struts.actions.DispatchAction.dispatchMethod(DispatchAction.java:276)
         at org.apache.struts.actions.DispatchAction.execute(DispatchAction.java:196)
         at org.apache.struts.action.RequestProcessor.processActionPerform(RequestProcessor.java:421)
         at org.apache.struts.action.RequestProcessor.process(RequestProcessor.java:226)
         at org.apache.struts.action.ActionServlet.process(ActionServlet.java:1164)
         at org.apache.struts.action.ActionServlet.doPost(ActionServlet.java:415)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:760)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
         at weblogic.servlet.internal.ServletStubImpl$ServletInvocationAction.run(ServletStubImpl.java:996)
         at weblogic.servlet.internal.ServletStubImpl.invokeServlet(ServletStubImpl.java:419)
         at weblogic.servlet.internal.ServletStubImpl.invokeServlet(ServletStubImpl.java:315)
         at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:6452)
         at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
         at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:118)
         at weblogic.servlet.internal.WebAppServletContext.invokeServlet(WebAppServletContext.java:3661)
         at weblogic.servlet.internal.ServletRequestImpl.execute(ServletRequestImpl.java:2630)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:219)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:178)
    Caused by: java.lang.OutOfMemoryError
    </L_MSG>
    <L_MSG MN="ILHD-1109" PID="adminserver" TID="ExecuteThread: '14' for queue: 'weblogic.kernel.Default'" DT="2012/04/18 7:56:17:146" PT="WARN" AP="" DN="" SN="" SR="org.apache.struts.action.RequestProcessor">Unhandled Exception thrown: class javax.servlet.ServletException</L_MSG>
    <Apr 18, 2012 7:56:17 AM CDT> <Error> <HTTP> <BEA-101017> <[ServletContext(id=26367546,name=fcsi,context-path=/fcsi)] Root cause of ServletException.
    *java.lang.OutOfMemoryError*
    Please Advise.
    Thanks for your help in advance,
    Kartheek

  • Out of memory issue when executing wlappc ant task

    I am using weblogic 8.1 sp1 and doing a compilation for a little bit large EAR file using ant task. But it always throws out of memory error when the wlappc invoked the compiler to compile the jsp files.
    According to BEA's documentation<CR104610>, I put the runtimeflags with "-J-ms256m -J-mx512m" option into the wlappc tag and Ant seems recognizing this option but it didn't work. I tried every possible memory size to get rid of this problem, but the process still failed with the error message:
    The system is out of resources.
    Consult the following stack trace for details.
    java.lang.OutOfMemoryError
    - with nested exception:
    [Compilation errors : ]
    My computer has 1G memory, so it shouldn't be the hardware problem.
    Does anyone have an idea on this?
    Thanks in advance,
    Jacky

    Hi,
    Ant seems to use its own JVM and thus in the javac options during runtime,try specifying ANT java options for setting the memory parameters.
    ANT options would be memoryInitialSize and memoryMaximumSize.
    http://ant.apache.org/manual/index.html
    Hope this helps.

  • RoboHelp 9 gives an out of memory error and crashes when I try to import or link a Frame 10 file or

    I have Tech Suite 3. If I start a new RoboHelp project and try to import or link Frame files, RoboHelp tries for a while, then I get an Out of Memory error and the program crashes.
    I opened one of the sample projects and was able to link to one of my frame files without any problem, so it seems to be an issue with creating something new.
    Any suggestions?

    It happens when I create a new project and then try to import or link frame docs to make up the content. It starts scanning, then crashes. I did get it to the conversion setting page once, but no further.
    It does not happen if I open one of the supplied example projects and link a file. But then it doesn't let me choose, during import, any style mapping. And I can't delete the sample project fold
    Twice now it has told me when I tried to import (not link, but import) that my .fm file could not be opened, and told me to verify that Frame is installed (it is) and that the file is a valid frame file (it is).
    The docs and project are in separate folders on my C: drive.

Maybe you are looking for