Out of memory error  - admin server

Hi,
I have installed the Oracle SOA suite 11.1.1.7.1.
I have installed both SOA and OSB on admin server.
My laptop which has following configuration
4GB RAM, Intel Core I3 Processor, Windows 7 Ultimate 32 bit.
But I am getting the out of memory error ..I have the changed the settings in setSOADomainEnv.cmd as follows
set JAVA_OPTIONS=%JAVA_OPTIONS%
set DEFAULT_MEM_ARGS=-Xms768m -Xmx768m
set PORT_MEM_ARGS=-Xms768m -Xmx768m
if "%JAVA_VENDOR%" == "Oracle" goto OracleJVM
set DEFAULT_MEM_ARGS=%DEFAULT_MEM_ARGS% -XX:PermSize=128m -XX:MaxPermSize=512m
set PORT_MEM_ARGS=%PORT_MEM_ARGS% -XX:PermSize=256m -XX:MaxPermSize=512m
are these the right changes I am done if not what needs to be done to get it working?
thanks
Sanjeeva TJ

the correct file that you need modify is setDomainEnv.sh
File: $DOMAIN_HOME/bin/setDomainEnv.sh
Add the next lines:
if [ "${SERVER_NAME}" = "soa_server1" ] ; then
        EXTRA_JAVA_PROPERTIES="-Xms256m -Xmx5000m -XX:PermSize=256m -XX:MaxPermSize=5000m  ${EXTRA_JAVA_PROPERTIES}"
        export EXTRA_JAVA_PROPERTIES
fi      
Note: Replace "soa_server1" for your AdminServer Name.
After that you need to restart your weblogic.

Similar Messages

  • Out Of Memory Error - Weblogic server

    Hi,
    We are facing following errors in our production environment frequently...
    Please guide us how can we identify root cause.. because every time this error come in different functionalities so we can not blame one functionality (Code)...
    Error # 1
    ======
    <Jun 18, 2009 4:24:41 PM IST> <Error> <Socket> <BEA-000405> <Uncaught Throwable in processSockets
    java.lang.OutOfMemoryError: .
    java.lang.OutOfMemoryError:
    >
    <Jun 18, 2009 4:24:45 PM IST> <Error> <Server> <BEA-002608> <The ListenThread because of an error: java.lang.OutOfMemoryError: memsystemAllocObject - Object size: 24
    java.lang.OutOfMemoryError: memsystemAllocObject - Object size: 24
         at java.net.PlainSocketImpl.socketAccept(Native Method)
         at java.net.PlainSocketImpl.accept(PlainSocketImpl.java:384)
         at java.net.ServerSocket.implAccept(ServerSocket.java:450)
         at java.net.ServerSocket.accept(ServerSocket.java:421)
         at weblogic.socket.WeblogicServerSocket.accept(WeblogicServerSocket.java:34)
         Truncated. see log file for complete stacktrace
    >
    Error # 2
    ======
    <Jun 18, 2009 7:32:27 PM IST> <Error> <Server> <BEA-002608> <The ListenThread because of an error: java.lang.OutOfMemoryError:
    java.lang.OutOfMemoryError:
    >
    <Jun 18, 2009 7:32:27 PM IST> <Error> <Server> <BEA-002608> <The ListenThread because of an error: java.lang.OutOfMemoryError: memsystemAllocObject - Object size: 24
    java.lang.OutOfMemoryError: memsystemAllocObject - Object size: 24
         at java.net.PlainSocketImpl.socketAccept(Native Method)
         at java.net.PlainSocketImpl.accept(PlainSocketImpl.java:384)
         at java.net.ServerSocket.implAccept(ServerSocket.java:450)
         at java.net.ServerSocket.accept(ServerSocket.java:421)
         at weblogic.socket.WeblogicServerSocket.accept(WeblogicServerSocket.java:34)
         Truncated. see log file for complete stacktrace
    >
    Thanks in Advance.
    Edited by: Rushang.Kansara on Jul 1, 2009 6:46 AM

    Once you get to the state where you're out of memory, you'll get any number of possible variations on the error message. These variations will give you absolutely zero help for why you've run out of memory. Without more information, it's impossible to tell whether you have a memory leak, or whether you simply have too small a value for "maxmem" in the JVM.
    Typically, I would assume that if the server has been running for a "while" (whatever that means) without error, then it probably isn't a simple issue with setting the "maxmem" value high enough. If it does fail pretty quickly after startup (and processing of initial messages), then it very well could be a maxmem problem. This is done by adding or changing the "-Xmxnnn" command-line parameter to the JVM. For instance, if you wanted to give it 2g of max heap, you would put "-Xmx2048m" on the command line. That value won't work on Windows (not sure about Vista or 64-bit), but it will work perfectly well on Unix/Linux.
    If you have a memory leak, it's good to first verify this by sitting on the "Server"->"Performance" page in the WebLogic console. If you watch the "memory free" (or "memory inuse") percentage, you'll see that it varies by a wide margin over time. What you want to do is to be able to force a GC (there should still be a button there to do that) and then note the percent free immediately after that. You should do this every few minutes while the server is actively working. If the percent free value (after the GC) is consistently going down, even slowly, then you likely have a memory leak. This is where it gets a little more difficult. You could look at the free VisualVM tool which provides some memory debugging features, but they are somewhat limited (I only know this from reading some literature, I haven't tried it), or you could look at the commercial tools JProfiler, JProbe, or YourKit (in decreasing order of my personal preference).

  • Report with more than 600 kb image - BO Server getting Out Of memory Error

    Hi,
       We have a report which displays images and size is above 600 KB.We are getting an "Out Of memory" Error while previewing this report in Business Object Server.
    In another situation we tried by giving a dynamic path to the OLE object, but the report is not showing error. We are not able to view the image whose size is more than 600KB
    Notes: 1. Image is stored as a BLOB  in Oracle Database
               2. Connection used inside the report is OLEDB
               3. UNIX Environment
    Regards,
    Sathish

    Please re-post if this is still an issue to the Business Objects Forum or if you have a valid support contract create a case on line.

  • Out of Memory Error While deploying as EAR file

    Hai,
    I was trying to deploy an EAR file of size 63 MB which inturn containing about 60 EJB.jars. No WARs. application.xml has all the entries for the JARs. While I am deploying it is giving Out of Memory Error. Is there any way to tweak this problem. I am using my own hand written java application which uses the SunONE deployment APIs for deployment. Can u please tell how to tackle this problem. I am running my application through a batch file which uses jdk1.4.
    Please help me regarding this issue.

    You can set the initial heap size and maximum heap size for the JVM, either in the app-server admin console, or maybe in one of your scripts. You look-up the syntax!...
    I had this error yesterday. I too had run out of memory (150Mb). You simply need to allocate more to the app-server.

  • Out of memory errors

    I'm having problems starting 8i. When started from /etc/rc.d/
    init.d/dbora during bootup or manually by root, I get an out of
    memory error.
    If I start it from the dbadmin user (oracle), I either get the
    same out of memory error, or I end up with several hundred shell
    logins and the database still doesn't respond.
    This is RedHat Linux 6.0, kernel 2.2.5-22smp. Here is a sample of
    what happens when I get the out of memory error:
    Mem: 160448K av, 53488K used, 106960K free, 26264K shrd,
    6532K buff
    Swap: 656360K av, 0K used, 656360K free
    36328K cached
    Oracle Server Manager Release 3.1.5.0.0 - Production
    (c) Copyright 1997, Oracle Corporation. All Rights Reserved.
    Oracle8i Release 8.1.5.0.0 - Production
    With the Java option
    PL/SQL Release 8.1.5.0.0 - Production
    SVRMGR> Connected.
    SVRMGR> ORA-27102: out of memory
    Linux Error: 22: Invalid argument
    SVRMGR>
    Server Manager complete.
    Database "ORCL" warm started.
    null

    It turns out that the problem I was having with a large number of
    shell processes being created was due to the use of oraenv in my
    .bashrc file (so much for following the instructions!) It was
    calling itself recursively until the process limit was reached.
    However, even with this fixed, the out of memory error still
    exists.
    max (guest) wrote:
    : dan....
    : check your init.ora......
    Aside from comments, it has these lines, which were created by
    dbassist:
    db_name = test
    instance_name = ORCL
    service_names = test
    control_files = ("/u02/oradata/test/control01.ctl", "/u02/oradata/
    test/control02.ctl")
    db_block_buffers = 8192
    shared_pool_size = 4194304
    log_checkpoint_interval = 10000
    log_checkpoint_timeout = 1800
    # I reduced processes to see if it would help
    processes = 10
    log_buffer = 163840
    background_dump_dest = /u01/admin/test/bdump
    core_dump_dest = /u01/admin/test/cdump
    user_dump_dest = /u01/admin/test/udump
    db_block_size = 2048
    remote_login_passwordfile = exclusive
    os_authent_prefix = ""
    compatible = "8.1.0"
    : also check ulimit
    Here's ulimit -a:
    core file size (blocks) 1000000
    data seg size (kbytes) unlimited
    file size (blocks) unlimited
    max memory size (kbytes) unlimited
    stack size (kbytes) 8192
    cpu time (seconds) unlimited
    max user processes 256
    pipe size (512 bytes) 8
    open files 1024
    virtual memory (kbytes) 2105343
    Everything looks pretty large to me.
    : Dan Wilga (guest) wrote:
    : : I'm having problems starting 8i. When started from /etc/rc.d/
    : : init.d/dbora during bootup or manually by root, I get an out
    : of
    : : memory error.
    : : If I start it from the dbadmin user (oracle), I either get
    the
    : : same out of memory error, or I end up with several hundred
    : shell
    : : logins and the database still doesn't respond.
    : : This is RedHat Linux 6.0, kernel 2.2.5-22smp. Here is a
    sample
    : of
    : : what happens when I get the out of memory error:
    : : Mem: 160448K av, 53488K used, 106960K free, 26264K shrd,
    : : 6532K buff
    : : Swap: 656360K av, 0K used, 656360K free
    : : 36328K cached
    : : Oracle Server Manager Release 3.1.5.0.0 - Production
    : : (c) Copyright 1997, Oracle Corporation. All Rights Reserved.
    : : Oracle8i Release 8.1.5.0.0 - Production
    : : With the Java option
    : : PL/SQL Release 8.1.5.0.0 - Production
    : : SVRMGR> Connected.
    : : SVRMGR> ORA-27102: out of memory
    : : Linux Error: 22: Invalid argument
    : : SVRMGR>
    : : Server Manager complete.
    : : Database "ORCL" warm started.
    null

  • Photoshop out of memory error

    Hi, I am running in a mixed enviroment with Open Directory master for authentication on a Snow Leopard OS 10.68 server, with both Lion and Snow Leopard client machines. My Users, due to Adobe products crashing with none mobile accounts, are using external hard drives formatted OS Extended (Journaled) with their mobile account setup on that Hard Drive. This year it has been touch and go on login issues but working throught that. My problem is when a user uses a Lion machine 10.7.4 and Photoshop extended and then moves to our study lab which has Snow Leopard, photoshop workspaces are corrupted and they cannot open any photo's or settings, they get an "Out of memory" error. However when they go back to the Lion machine and after resetting their workspace they can use the software without issues. Anyone else hiving these issues? Ive tried chflags nohidden library in Snow Leopard to view settings, even deleted all photoshop settings in App data and preferences and still cannot access photoshop on SL machine.
    Thanks

    Thanks Kurt for the reply. I'll give more info. All machines have latest updates for both CS6 and current version of OS either 10.6.8 or 10.7.4.
    The only thing on the external hard drive is the user's home folder and their Data, I have to have permissions enabled so their home preferences and WGM settings work correctly. BTW all accounts have admin rights to machine, I have WGM settings preventing access to machine settings and can re-image if i get corrupted machines.
    PS is installed on each machine not the  External Hard Drive.
    All machines have the same Volume name for the internal boot drive, which is set as the PS scratch drive.
    I thought this issue was to do with the memory and may still be so.
    However when a clean profile is connected to our towers with Lion which has 12 gb ram and 1024 MB Video Memory, the settings are at 70% which is around 8 gb.
    When i take same clean profile to our other lab of iMacs which has 8 gb ram. and 512 mb video memory, PS adjusts the performance ram accordingly which is around 5.5 gb ram at 70%.
    I then take that same external drive to the problem iMac's (early 2009 and 2008) which has 4gb ram 256 video memory and is running Lion, PS adjust to 2.4 gb ram.
    Now i put that same drive on the same model, 2009 or 2008 iMac that has Snow Leopard running on it from same model Lion iMac,PS opens fine.
    However after moving from one of the other larger Lion machines and then back to this Snow Leopard machine, the profile gets corrupted, the workspace is corrupted and cannot reset it. Also I am unable to access any settings I get the "Could not complete the operation because there is not enough memory (RAM)" error.
    Now when I go back to a same model Lion machine with same minimum memory i get the same error, however when I go to a larger Lion machine all I get is Color profile cannot sync and the workspace is still corrupted and not showing but I can reset it.
    I then move the performance size to match that of the lower model’s 70%, I still get the error when I go back to the lower end Lion or Snow Leopard machine.
    I tried clearing PS preferences by opening with command+option+shift and delete PS preferences the issue is still there.
    I then remove all PS ~/Library/ settings for PS and it is still present.
    I had to re-create the profile all together to get this to work. As long as I don’t connect to a low end Snow Leopard machine things seem to be going well and PS readjust according to the machine, note, when I set the performance level to a low setting let's say 2.5 gb as on the early iMacs and plug into another machine, PS adjusts to the current machine' mamory availability and does not keep that lower setting setting.
    I have a copy of console error message below.
    9/18/12 2:31:51 PM Adobe Photoshop CS6[254] *** __NSAutoreleaseNoPool(): Object 0x10aa114d0 of class NSCFString autoreleased with no pool in place - just leaking
    9/18/12 2:31:51 PM Adobe Photoshop CS6[254] *** __NSAutoreleaseNoPool(): Object 0x106d65010 of class NSBundle autoreleased with no pool in place - just leaking
    9/18/12 2:31:51 PM Adobe Photoshop CS6[254] *** __NSAutoreleaseNoPool(): Object 0x11db6aa60 of class NSCFString autoreleased with no pool in place - just leaking
    9/18/12 2:31:51 PM Adobe Photoshop CS6[254] *** __NSAutoreleaseNoPool(): Object 0x10a98bc40 of class NSCFString autoreleased with no pool in place - just leaking
    9/18/12 2:31:51 PM Adobe Photoshop CS6[254] *** __NSAutoreleaseNoPool(): Object 0x1371a73e0 of class NSCFData autoreleased with no pool in place - just leaking
    9/18/12 2:31:51 PM Adobe Photoshop CS6[254] *** __NSAutoreleaseNoPool(): Object 0x1371f7320 of class NSCFData autoreleased with no pool in place - just leaking
    I have 12 Snow Leopard machines that are early 2008 and 2009 imac and no way to up upgrading to Lion and I am not ready for Mountain Lion to go into production untill I can upgrade my OD masters to Mountain Lion.
    I suspect it's not the ram settings that are affected but the video ram is not adjusting from machine to machine, is it possible to upgrade the 2009 20" iMac video cards and get proper firmware support?
    any help is appreciated.

  • Can anyone help with an out of memory error on CS2?

    Hello,
    I have been battling with an out of memory error on a dell laptop running 2gb of ram. I have had no prior issues with this laptop running CS2 until last week when I started getting the out of memory errors. I have seen several articles about this occuring on later versions of illustrator. Just prior to this issue appearing I had to delete and rebuild the main user profile on the computer due to it consistently logging into a temporary account.
    Once the new profile was established and all user documents and profile settings were transferred back into it, this problem started showing up. The error occurs when clicking on files listed as Illustrator files, and when opening the illustrator suite. Once I click on the out of memory error it still continues to load illustrator. The files are located on a server and not on the remote computer. I have tried to copy the files to the local hard drive with the same result when I try to open them. I also tried to change the extension to .pdf and got no further. I tried to open a file and recreate the work and all I got was the black outline of the fonts. none of the other artwork appeared. once rebuilt, when trying to save, It gives the out of memory error and does not save the file.
    Illustrator was installed on another dell laptop, same model, etc and works fine. no problem accessing network files. I just attempted to use Illustrator from the admin profile on the machine in question and it works fine and lightening fast. Just not from the new rebuilt profile. I have tried to uninstall and reinstall the software. it has not helped. could there be stray configuration settings that we transfered over to the new profile that should not have been?
    Any insights will be helpful.
    Thank You,
    Brad Yeutter

    Here are some steps that could help you:
    http://kb2.adobe.com/cps/320/320782.html
    http://kb2.adobe.com/cps/401/kb401053.html
    I hope this helps!

  • Jrockit out of memory error

    We are getting out of memory error with jrockit1.4.2_08. We used memory debugger tools to find the leak. Apparently there do not seem to be a leak from application perspective. Here is the core dump. Please help me resolve this.
    ===== BEGIN DUMP =============================================================
    JRockit dump produced after 0 days, 05:37:55 on Thu Jun 5 20:39:11 2008
    Additional information is available in:
    /opt/obs/obs_app/obs31.0/user_projects/admin/jrockit.17430.dump
    No core file will be created because core dumps have been
    disabled. To enable core dumping, try "ulimit -c unlimited"
    before starting JRockit again.
    If you see this dump, please open a support case with BEA and
    supply as much information as you can on your system setup and
    the program you were running. You can also search for solutions
    to your problem at http://forums.bea.com in
    the forum jrockit.developer.interest.general.
    Error code: 52
    Error Message: Null pointer exception in native code
    Signal info : si_signo=11, si_code=2
    Version : BEA WebLogic JRockit(TM) 1.4.2_08 JVM R24.5.0-61 ari-49095-20050826-1856-linux-ia32
    Threads / GC : Native Threads, GC strategy: parallel
    : mmHeap->data = 0x40416000, mmHeap->top = 0x7ec16000
    : mmStartCompaction = 0x45236000, mmEndCompaction = 0x4a056000
    CPU : Intel Pentium 4 (HT)
    Number CPUs : 4
    Tot Phys Mem : 4021997568
    OS version : Red Hat Enterprise Linux ES release 2.1 (Panama)
    Linux version 2.4.9-e.72smp ([email protected]) (gcc version 2.96 20000731 (Red Hat Linux 7.2 2.96-129.7.2)) #1 SMP Tue Jul 3 22:04:51 EDT 2007
    State : JVM is running
    Command Line : -Djava.class.path=<my own jars list>
    java.library.path=/opt/obs/wls/jrockit-j2sdk1.4.2_08/jre/lib/i386/jrockit:/opt/obs/wls/jrockit-j2sdk1.4.2_08/jre/lib/i386:/opt/obs/wls/jrockit-j2sdk1.4.2_08/jre/../lib/i386:/opt/mqm/java/lib:/opt/tib/rv7.1.2/lib:/opt/obs/obs_app/obs31.0/thirdparty:/opt/obs/wls/wls8.1sp3/weblogic81/server/lib/linux/i686:/opt/obs/wls/wls8.1sp3/weblogic81/server/lib/linux/i686/oci920_8
    C Heap : Good; no memory allocations have failed
    Registers (from context struct at 0x8d5b724/0x8d5b80c):
    EAX = 000005f4 EBX = 0000007f
    ECX = 7ff8bd90 EDX = 24a18f94
    ESI = 7ff46af0 EDI = 00000100
    ESP = 7f262bc0 EIP = 4026d58f
    EBP = 7f262bd8 EFL = 00010203
    CS = 0023 DS = 002b ES = 002b
    SS = 002b FS = 00df GS = 00df
    Stack:
    7f262bc0 :00000000 7ff46af0 00000f94 00000000 00028000 00000000
    7f262bd8 :7f262c08 401ffcae 7ff8bd90 24a18f94 00028000 a0000000
    7f262bf0 :0a5f1800 7f262c74 08d137a0 00000050 00000000 7f262c34
    7f262c08 :7f262c48 4028a7fb 24a18f94 00000000 08d137a0 08d139e4
    7f262c20 :581a664c 00000003 08d137a0 00000001 7f262c74 00000000
    7f262c38 :7f262ca8 402c28bd 7f262c74 20040144 7f262ca8 402c292f
    7f262c50 :7f262c74 00000000 00000000 7f262c84 581a6640 08d137a0
    7f262c68 :08d1382c 414d2db8 210d41e8 0a5f1800 08d137a0 7f262d9c
    7f262c80 :24a18f95 00000002 00000000 7f262cf8 08d1382c 581a6640
    7f262c98 :00000000 00000000 00000000 00000000 7f262ce8 402e0861
    7f262cb0 :08d137a0 581a6640 00000020 7f262cf8 210cc499 00000001
    7f262cc8 :00000000 7f262ce4 08d139e8 7f262cf0 581a66c0 53b8d780
    7f262ce0 :00000000 00000020 08d137a0 210cc43e 08d1382c 08d139e4
    7f262cf8 :08116840 08d139e4 00000000 581a6400 210b41ff 00000001
    7f262d10 :581a6618 581a6560 210cc499 581a6565 581a6618 581a6560
    7f262d28 :210d2744 581a62f0 581a6560 210d2725 2112a995 2112a985
    7f262d40 :2112a975 23a67584 581a6578 00000001 00000001 220e35a9
    7f262d58 :23a53381 581a62f0 00000000 00000000 581a6548 00000046
    7f262d70 :00000000 00000002 00000000 581a62f0 00000001 00000000
    7f262d88 :00000000 00000000 581a6530 581a62f0 23a56e15 24a18f95
    7f262da0 :24a18ba3 00000000 581a6130 210bead9 581a6038 581a6038
    7f262db8 :581a6038 53b88f50 40fa9ca8 581a6308 55b54b00 40fa9c38
    7f262dd0 :80000000 55a06030 581a6020 581a6038 581a6020 24a189bb
    7f262de8 :40fa9c38 581a6020 40fa9ca8 581a60c0 55a06030 53942ff8
    7f262e00 :55b54b00 40fa9ca8 40fa9c38 08d137a0 7f262e24 210b34d8
    7f262e18 :40fa9c38 40fa9c68 40fa9ca8 24a18800 402e3a4d 7f262fb3
    7f262e30 :0000002c 00000003 7f262ec4 08d1382c 24b196e0 249c4f84
    7f262e48 :581a5390 00000000 00000058 7f262e84 402b861a 249c4f84
    7f262e60 :7f262e80 08d1382c 00000001 00000001 00000008 00000013
    7f262e78 :00000005 581a5fd8 08d139dc 7f262e18 00000003 00000005
    7f262e90 :00000005 00000000 00000000 55a06030 55b54b00 7f262e24
    7f262ea8 :7f262e24 08d137a0 7f262ee0 24a18800 210b34a0 0000002c
    7f262ec0 :7f262fb3 7f262f04 402c8931 08d1382c 1aea85c0 24b196e0
    7f262ed8 :00000000 7f262f5c 402c8b74 7f262f54 401aaab0 00000000
    7f262ef0 :7f262f5c 7f262f54 401aaab0 401aa6a0 1aea85c0 7f262f64
    7f262f08 :402c8c41 08d1382c 249c4f84 00000000 7f262f5c 402c8b74
    7f262f20 :00000000 7f262f54 08d13dd0 08d1382c 7f262f9c 00000000
    7f262f38 :00000000 7f262f7c 402f2b2d 08d137a0 00000001 7f262f7c
    7f262f50 :402f2b24 08d13dd0 402f2b24 7f262fd4 08d1382c 7f262f94
    7f262f68 :402dbd4c 08d1382c 08d139dc 249c4f84 7f262fd4 08d139e4
    7f262f80 :581a5ee0 00000000 476bd4c0 210b67f6 00000000 08d137a0
    7f262f98 :210e7b1b 08d1382c 08d139e0 249c4f84 08d139dc 7f262fd4
    7f262fb0 :083e6898 08d139dc 00000000 210eb30d 40fa9ca8 210eb184
    7f262fc8 :405bff10 00000005 210e7c48 581a5fb8 581a5fb8 581a5fe0
    7f262fe0 :00000000 00000000 581a5690 210e7253 581a5fb8 00000009
    7f262ff8 :73a44aa8 73a44aa8 00000005 581a5fb8 00000009 581a5690
    7f263010 :246e7cd0 581a5fb8 581a53e8 581a5f48 40fa9b90 40fa9be8
    7f263028 :55a06030 55b54b00 40fa9be8 40fa9bd0 246e7b13 55b54b00
    7f263040 :55a06030 40fa9b60 55b54b00 55a06030 581a5378 246e79b6
    7f263058 :55a06030 55b54b00 57712968 405c6858 40fa9b40 00000001
    7f263070 :40fa9b40 2426bab4 55a06030 57712968 57712928 57712968
    7f263088 :55b54b00 55b54b00 4b345448 2426ba1b 2426b717 53b88f50
    7f2630a0 :405c6858 5393c038 53b88f50 00000002 53b88f40 24266f1f
    7f2630b8 :53b88f50 53b88f50 53b88f40 5393c038 53b849c0 00000000
    7f2630d0 :53912b90 5393acd0 5393c038 53912b90 242613c8 5393c038
    7f2630e8 :00000000 5393acd0 00000000 5393aa60 5393c038 24658a31
    7f263100 :5393acd0 4a0dbf40 00000000 4a0dbf40 539125c0 23a74a88
    7f263118 :539126e8 00000001 6553be00 539125c0 246587dc 53ebb610
    7f263130 :4a0dbf40 00000000 4a124210 00000000 00000000 5393aa60
    7f263148 :218aff64 09ccec00 221aa582 5393a440 4a1507e8 4a1507e8
    7f263160 :536f07d8 533cdee0 5393a440 533cdee0 000378b6 536f07d8
    7f263178 :5393a430 00000004 210b48f0 00000050 0000000f 211080dd
    7f263190 :40576ec8 21108145 5393a480 5393aa60 5393aa60 21108074
    7f2631a8 :00000003 536f07d8 4a124210 7d1c9048 5393a430 654d5040
    7f2631c0 :4a090428 232a5328 7d1c9048 4a124210 229f2aeb 7d1c9048
    7f2631d8 :4a124210 4a127090 538f80a8 654d5040 536f07d8 4a127090
    7f2631f0 :536f07d8 242438f0 536f07d8 7d1c9048 7d1c9048 538f80a8
    7f263208 :5393a418 7d1c9048 229f2a12 536f07d8 7d1c9048 22147208
    7f263220 :40528fa8 4056b138 00000000 00000000 405d4858 7d1c9048
    7f263238 :2214718a 5393a418 7d1c9048 229f239d 5393a418 536f07d8
    7f263250 :7d1c9048 4056b138 41356f20 00000000 536f07d8 40528fa8
    7f263268 :4a127090 7d1c9048 5393a418 538f80a8 42562b68 538f81e0
    7f263280 :229f2230 219161d1 41356f20 218e2089 42562b6d 41356f20
    7f263298 :7f2632b4 210b8c73 402e37ef 7f2633e8 08d137a0 7f2632b4
    7f2632b0 :210b3478 210b8c60 402e3a4d ffffffff 00000010 00000000
    7f2632c8 :7f263354 08d1382c 200920e0 200282cc 00000000 00000000
    7f2632e0 :00000000 7f263314 402b861a 200282cc 7f263310 08d1382c
    7f2632f8 :00000001 00000001 00000006 00000006 00000001 08d137a0
    7f263310 :00000000 7f2632b4 00000000 00000001 00000001 00000000
    7f263328 :00000000 00000000 42562b68 7f2632b4 7f2632b4 08d137a0
    7f263340 :00000000 210b8c60 210b3440 00000010 ffffffff 7f263394
    7f263358 :402c8931 08d1382c 080d97e0 200920e0 00000000 7f2633e8
    7f263370 :402e3550 7f2633cc 00000000 08d1382c 08d1382c 7f263aac
    7f263388 :00000000 00000000 080d97e0 7f2633d4 402c9e9f 08d1382c
    7f2633a0 :200282cc 00000000 7f2633e8 402e3550 00000001 7f2633cc
    7f2633b8 :00000000 00000000 00000000 00000000 00000000 00000000
    Code:
    4026d48f :83403d9a 008bf4c4 166be850 c4830009 f4c48310 a817e853
    4026d4a7 :b70f0009 c0010643 06438966 0ff8c483 048dc0b7 02e0c140
    4026d4bf :08438b50 57b7e850 43890003 20c48308 53f4c483 09a87ce8
    4026d4d7 :9a38a100 8b64403d 30050300 83403d9a c48310c4 50008bf4
    4026d4ef :0915f4e8 084b8b00 4574c985 0f0c558b 8d0443b7 14894004
    4026d507 :10558b81 0443b70f c140048d 430302e0 04508908 0f14558b
    4026d51f :8d0443b7 e0c14004 08430302 66085089 b80443ff 00000001
    4026d537 :b48d09eb 00000026 8bc03100 ec89e85d 768dc35d e5895500
    4026d54f :570cec83 458b5356 08558b08 00fc45c7 83000000 b70ff4c4
    4026d567 :478d0478 8df8d101 e852ff58 0009a7ed 8310c483 4974fffb
    4026d57f :084d8b90 8b5b148d 048d0841 0c558b90 77045039 0850391c
    4026d597 :188b0f72 51f4c483 09a7fce8 ebd88900 fc5d892e 768d05eb
    4026d5af :8bdf8900 de89fc4d 8939148d 1fe8c1d0 d1101c8d 75f339fb
    4026d5c7 :08458bb8 50f4c483 09a7cce8 8dc03100 5e5be865 5dec895f
    4026d5df :e58955c3 571cec83 458b5356 fc45c70c 00000000 04788366
    4026d5f7 :c7627400 0000f845 558b0000 f8458b0c 8b084203 8b118b08
    4026d60f :388b0442 8b08428b 8b028b30 f8c48300 8b04598b 08418b10
    4026d627 :5040d829 145d2b52 52575653 10458b53 2c406850 558b4036
    4026d63f :bae85208 8bc7ddb3 c4830c55 f8458330 fc45ff0c 0442b70f
    4026d657 :7cfc4539 d8658da5 895f5e5b 89c35dec e58955f6 572cec83
    4026d66f :38a15356 64403d9a 0503008b 403d9a30 8bf4c483 5ee85000
    4026d687 :83000914 c48310c4 e45d8df4 19bbe853 b2e80009 e8000918
    Loaded modules:
    (* denotes the module causing the exception)
    0x08048000-0x0804ce46 /opt/obs/wls/jrockit-j2sdk1.4.2_08/bin/java
    0x4001d000-0x40029fab /lib/i686/libpthread.so.0
    0x4004e000-0x4006fb42 /lib/i686/libm.so.6
    0x40071000-0x4007300c /lib/libdl.so.2
    0x40075000-0x401a76e5 /lib/i686/libc.so.6
    0x401b2000-0x40388eef* /opt/obs/wls/jrockit-j2sdk1.4.2_08/jre/lib/i386/jrockit/libjvm.so
    0x7ec57000-0x7ec66fa5 /opt/obs/wls/jrockit-j2sdk1.4.2_08/jre/lib/i386/libverify.so
    0x7ec75000-0x7ec94a0f /opt/obs/wls/jrockit-j2sdk1.4.2_08/jre/lib/i386/libjava.so
    0x7ec9c000-0x7ecae3ba /lib/libnsl.so.1
    0x7ec97000-0x7ec98705 /opt/obs/wls/wls8.1sp3/weblogic81/server/lib/linux/i686/libweblogicunix1.so
    0x7f5f2000-0x7f5f3eff /opt/obs/wls/wls8.1sp3/weblogic81/server/lib/linux/i686/libmuxer.so
    0x7f5f6000-0x7f5f95c1 /opt/obs/wls/jrockit-j2sdk1.4.2_08/jre/lib/i386/libioser12.so
    0x80318000-0x8031f8bf /opt/obs/obs_app/obs31.0/thirdparty/libXMLC.so
    0x80344000-0x8034c66f /opt/tib/rv7.1.2/lib/libtibrvj.so
    0x8034e000-0x803563e0 /opt/tib/rv7.1.2/lib/libtibrvcmq.so
    0x803d2000-0x803e66bb /opt/tib/rv7.1.2/lib/libtibrvcm.so
    0x7fef9000-0x7fefdcaf /opt/tib/rv7.1.2/lib/libtibrvft.so
    0x803e8000-0x804341a4 /opt/tib/rv7.1.2/lib/libtibrv.so
    Thread Stack Trace:
    WARNING: Memory exhausted
    at java/lang/Throwable.fillInStackTrace0(Native Method)@0x210cc3f0
    at java/lang/Throwable.fillInStackTrace(Unknown Source)@0x210cc499
    at java/lang/Throwable.<init>(Unknown Source)@0x210d2744
    at java/lang/Exception.<init>(Exception.java:41)@0x210d2725
    at java/lang/RuntimeException.<init>(RuntimeException.java:43)@0x2112a995
    at java/lang/IllegalArgumentException.<init>(IllegalArgumentException.java:36)@0x2112a985
    at java/lang/NumberFormatException.<init>(NumberFormatException.java:38)@0x2112a975
    at java/lang/NumberFormatException.forInputString(NumberFormatException.java:48)@0x23a67584
    at java/lang/FloatingDecimal.readJavaFormatString(FloatingDecimal.java:1207)@0x220e35a9
    at java/lang/Double.parseDouble(Double.java:220)@0x23a56e15

    I can't tell from the crash dump what the exact problem is - maybe you are running out of native memory?
    You are running a very old JRockit version. Since you are a Oracle/BEA customer, you can get a newer one from BEA Support, or through edelivery.oracle.com.
    -- Henrik

  • Out of memory using Windows Server 2003 and Oracle 9.2.0.3

    Hi,
    I just installed both Windows Server 2003 and Oracle 9.2.0.3. The server I am using has 4 gigabytes of memory installed, and windows shows that to be the actual physical memory on the server. The paging file is 8 gigabytes for a total of 12 gigabytes of virtual memory.
    Using the Oracle Database Configuration Assistant, I create a new database using default values. After verifying the installation worked OK (database started, queries ran), I attempted to change some of the memory parameters and was able to start the database using about 1.4 gigabytes of total SGA, with the O/S using about .3 gigabytes for a total of 1.7 gigabytes. I then attempted to change the buffer cache + 1 gigabyte for a total of 2 gigabytes. Now the database will not start and I get an out of memory error ora-27102 with an additional O/S memory error of "O/S Error: (OS 8) Not enough storage is available to process this command".
    So, I have two problems:
    1) With 4 gigabytes installed, and the O/S using only about 300 megabytes, why isn't Oracle able to take advantage of the memory on this machine past the 1.7 gigabytes (as reported by Windows task manager).
    2) How do I change the spfile memory parameters so the database will start using spfile? I can start the database pointing to a pfile with lesser memory parameters, but I don't know how to effect change on the spfile, since I cannot start with the spfile.
    Both these technologies (9.2.0.3 and Server 2003) are new to me, so any insight would be appreciated!
    HERE ARE THE SPFILE PARAMETERS THAT WORKED:
    SGA (M)
    Shared Pool=104
    Buffer Cache=960 ***
    Large Pool=0
    Java Pool=64
    Total SGA=1128
    SGA Max Size=1321
    Aggregate PGA Target=1500
    Current PGA Allocated=22868KB
    WINDOWS MEMORY INFO FROM TASK MANAGER SHOW:
    PF Usage=1.66GB
    Physical Memory (K)
    Total=4062704
    Available=3203460
    System Cache=2499096
    Commit Charge (K)
    Total=1748184
    Limit=12288628
    Peak=1925276
    *** When I change this to 2048 the database did not start.
    Thank you.

    CREATE SPFILE
    Purpose
    Use the CREATE SPFILE statement to create a server parameter file from a client-side initialization parameter file. Server parameter files are binary files that exist only on the server and are called from client locations to start up the database.
    Server parameter files let you make persistent changes to individual parameters. When you use a server parameter file, you can specify in an ALTER SYSTEM SET parameter statement that the new parameter value should be persistent. This means that the new value applies not only in the current instance, but also to any instances that are started up subsequently. Traditional client-side parameter files do not let you make persistent changes to parameter values. Because they are located on the server, these files allow for automatic database tuning by Oracle and for backup by Recovery Manager (RMAN).
    To use a server parameter file when starting up the database, you must create it from a traditional text initialization parameter file using the CREATE SPFILE statement.
    All instances in an Real Application Clusters environment must use the same server parameter file. However, when otherwise permitted, individual instances can have different settings of the same parameter within this one file. Instance-specific parameter definitions are specified as SID.parameter = value, where SID is the instance identifier.
    The method of starting up the database with a server parameter file depends on whether you create a default or nondefault server parameter file. Please refer to "Creating a Server Parameter File: Examples" for examples of how to use server parameter files.
    See Also:
    CREATE PFILE for information on creating a regular text parameter file from a binary server parameter file
    Oracle9i Database Administrator's Guide for information on pre-Oracle9i initialization parameter files and Oracle9i server parameter files
    Oracle9i Real Application Clusters Administration for information on using server parameter files in a Real Application Clusters environment
    Prerequisites
    You must have the SYSDBA or the SYSOPER system privilege to execute this statement. You can execute this statement before or after instance startup. However, if you have already started an instance using spfile_name, you cannot specify the same spfile_name in this statement.
    Syntax
    create_spfile::=
    Text description of create_spfile
    Semantics
    spfile_name
    This clause lets you specify a name for the server parameter file you are creating.
    If you do not specify spfile_name, Oracle uses the platform-specific default server parameter filename. If spfile_name already exists on the server, this statement will overwrite it. When using a default server parameter file, you start up the database without referring to the file by name.
    If you do specify spfile_name, you are creating a nondefault server parameter file. In this case, to start up the database, you must first create a single-line traditional parameter file that points to the server parameter file, and then name the single-line file in your STARTUP command.
    See Also:
    "Creating a Server Parameter File: Examples" for information on starting up the database with default and nondefault server parameter files
    Oracle9i Database Administrator's Guide for Windows (or other appropriate operating system specific documentation) for default parameter file names
    pfile_name
    Specify the traditional initialization parameter file from which you want to create a server parameter file.
    If you specify pfile_name, the parameter file must reside on the server. If it does not reside in the default directory for parameter files on your operating system, you must specify the full path.
    If you do not specify pfile_name, Oracle looks in the default directory for parameter files on your operating system for the default parameter filename, and uses that file. If that file does not exist in the expected directory, Oracle returns an error.
    Note:
    In a Real Application Clusters environment, you must first combine all instance parameter files into one file before specifying it in this statement to create a server parameter file. For information on accomplishing this step, see Oracle9i Real Application Clusters Setup and Configuration.
    Examples
    Creating a Server Parameter File: Examples
    The following example creates a default server parameter file from a client initialization parameter file named t_init1.ora:
    CREATE SPFILE
    FROM PFILE = '$ORACLE_HOME/work/t_init1.ora';
    Note:
    Typically you will need to specify the full path and filename for parameter files on your operating system. Please refer to your Oracle operating system documentation for path information.
    When you create a default server parameter file, you subsequently start up the database using that server parameter file by using the SQL*Plus command STARTUP without the PFILE parameter, as follows:
    STARTUP
    The following example creates a nondefault server parameter file s_params.ora from a client initialization file named t_init1.ora:
    CREATE SPFILE = 's_params.ora'
    FROM PFILE = '$ORACLE_HOME/work/t_init1.ora';
    When you create a nondefault server parameter file, you subsequently start up the database by first creating a traditional parameter file containing the following single line:
    spfile = 's_params.ora'
    The name of this parameter file must comply with the naming conventions of your operating system. You then use the single-line parameter file in the STARTUP command. The following example shows how to start up the database, assuming that the single-line parameter file is named new_param.ora:
    STARTUP PFILE=new_param.ora
    http://download-west.oracle.com/docs/cd/B10501_01/server.920/a96540/statements_616a.htm#SQLRF01315
    Joel P�rez

  • Uploading large files from applet to servlet throws out of memory error

    I have a java applet that needs to upload files from a client machine
    to a web server using a servlet. the problem i am having is that in
    the current scheme, files larger than 17-20MB throw an out of memory
    error. is there any way we can get around this problem? i will post
    the client and server side code for reference.
    Client Side Code:
    import java.io.*;
    import java.net.*;
    // this class is a client that enables transfer of files from client
    // to server. This client connects to a servlet running on the server
    // and transmits the file.
    public class fileTransferClient
    private static final String FILENAME_HEADER = "fileName";
    private static final String FILELASTMOD_HEADER = "fileLastMod";
    // this method transfers the prescribed file to the server.
    // if the destination directory is "", it transfers the file to
    "d:\\".
    //11-21-02 Changes : This method now has a new parameter that
    references the item
    //that is being transferred in the import list.
    public static String transferFile(String srcFileName, String
    destFileName,
    String destDir, int itemID)
    if (destDir.equals(""))
    destDir = "E:\\FTP\\incoming\\";
    // get the fully qualified filename and the mere filename.
    String fqfn = srcFileName;
    String fname =
    fqfn.substring(fqfn.lastIndexOf(File.separator)+1);
    try
    //importTable importer = jbInit.getImportTable();
    // create the file to be uploaded and a connection to
    servlet.
    File fileToUpload = new File(fqfn);
    long fileSize = fileToUpload.length();
    // get last mod of this file.
    // The last mod is sent to the servlet as a header.
    long lastMod = fileToUpload.lastModified();
    String strLastMod = String.valueOf(lastMod);
    URL serverURL = new URL(webadminApplet.strServletURL);
    URLConnection serverCon = serverURL.openConnection();
    // a bunch of connection setup related things.
    serverCon.setDoInput(true);
    serverCon.setDoOutput(true);
    // Don't use a cached version of URL connection.
    serverCon.setUseCaches (false);
    serverCon.setDefaultUseCaches (false);
    // set headers and their values.
    serverCon.setRequestProperty("Content-Type",
    "application/octet-stream");
    serverCon.setRequestProperty("Content-Length",
    Long.toString(fileToUpload.length()));
    serverCon.setRequestProperty(FILENAME_HEADER, destDir +
    destFileName);
    serverCon.setRequestProperty(FILELASTMOD_HEADER, strLastMod);
    if (webadminApplet.DEBUG) System.out.println("Connection with
    FTP server established");
    // create file stream and write stream to write file data.
    FileInputStream fis = new FileInputStream(fileToUpload);
    OutputStream os = serverCon.getOutputStream();
    try
    // transfer the file in 4K chunks.
    byte[] buffer = new byte[4096];
    long byteCnt = 0;
    //long percent = 0;
    int newPercent = 0;
    int oldPercent = 0;
    while (true)
    int bytes = fis.read(buffer);
    byteCnt += bytes;
    //11-21-02 :
    //If itemID is greater than -1 this is an import file
    transfer
    //otherwise this is a header graphic file transfer.
    if (itemID > -1)
    newPercent = (int) ((double) byteCnt/ (double)
    fileSize * 100.0);
    int diff = newPercent - oldPercent;
    if (newPercent == 0 || diff >= 20)
    oldPercent = newPercent;
    jbInit.getImportTable().displayFileTransferStatus
    (itemID,
    newPercent);
    if (bytes < 0) break;
    os.write(buffer, 0, bytes);
    os.flush();
    if (webadminApplet.DEBUG) System.out.println("No of bytes
    sent: " + byteCnt);
    finally
    // close related streams.
    os.close();
    fis.close();
    if (webadminApplet.DEBUG) System.out.println("File
    Transmission complete");
    // find out what the servlet has got to say in response.
    BufferedReader reader = new BufferedReader(
    new
    InputStreamReader(serverCon.getInputStream()));
    try
    String line;
    while ((line = reader.readLine()) != null)
    if (webadminApplet.DEBUG) System.out.println(line);
    finally
    // close the reader stream from servlet.
    reader.close();
    } // end of the big try block.
    catch (Exception e)
    System.out.println("Exception during file transfer:\n" + e);
    e.printStackTrace();
    return("FTP failed. See Java Console for Errors.");
    } // end of catch block.
    return("File: " + fname + " successfully transferred.");
    } // end of method transferFile().
    } // end of class fileTransferClient
    Server side code:
    import java.io.*;
    import javax.servlet.*;
    import javax.servlet.http.*;
    import java.util.*;
    import java.net.*;
    // This servlet class acts as an FTP server to enable transfer of
    files
    // from client side.
    public class FtpServerServlet extends HttpServlet
    String ftpDir = "D:\\pub\\FTP\\";
    private static final String FILENAME_HEADER = "fileName";
    private static final String FILELASTMOD_HEADER = "fileLastMod";
    public void doGet(HttpServletRequest req, HttpServletResponse resp)
    throws ServletException,
    IOException
    doPost(req, resp);
    public void doPost(HttpServletRequest req, HttpServletResponse
    resp)
    throws ServletException,
    IOException
    // ### for now enable overwrite by default.
    boolean overwrite = true;
    // get the fileName for this transmission.
    String fileName = req.getHeader(FILENAME_HEADER);
    // also get the last mod of this file.
    String strLastMod = req.getHeader(FILELASTMOD_HEADER);
    String message = "Filename: " + fileName + " saved
    successfully.";
    int status = HttpServletResponse.SC_OK;
    System.out.println("fileName from client: " + fileName);
    // if filename is not specified, complain.
    if (fileName == null)
    message = "Filename not specified";
    status = HttpServletResponse.SC_INTERNAL_SERVER_ERROR;
    else
    // open the file stream for the file about to be transferred.
    File uploadedFile = new File(fileName);
    // check if file already exists - and overwrite if necessary.
    if (uploadedFile.exists())
    if (overwrite)
    // delete the file.
    uploadedFile.delete();
    // ensure the directory is writable - and a new file may be
    created.
    if (!uploadedFile.createNewFile())
    message = "Unable to create file on server. FTP failed.";
    status = HttpServletResponse.SC_INTERNAL_SERVER_ERROR;
    else
    // get the necessary streams for file creation.
    FileOutputStream fos = new FileOutputStream(uploadedFile);
    InputStream is = req.getInputStream();
    try
    // create a buffer. 4K!
    byte[] buffer = new byte[4096];
    // read from input stream and write to file stream.
    int byteCnt = 0;
    while (true)
    int bytes = is.read(buffer);
    if (bytes < 0) break;
    byteCnt += bytes;
    // System.out.println(buffer);
    fos.write(buffer, 0, bytes);
    // flush the stream.
    fos.flush();
    } // end of try block.
    finally
    is.close();
    fos.close();
    // set last mod date for this file.
    uploadedFile.setLastModified((new
    Long(strLastMod)).longValue());
    } // end of finally block.
    } // end - the new file may be created on server.
    } // end - we have a valid filename.
    // set response headers.
    resp.setContentType("text/plain");
    resp.setStatus(status);
    if (status != HttpServletResponse.SC_OK)
    getServletContext().log("ERROR: " + message);
    // get output stream.
    PrintWriter out = resp.getWriter();
    out.println(message);
    } // end of doPost().
    } // end of class FtpServerServlet

    OK - the problem you describe is definitely what's giving you grief.
    The workaround is to use a socket connection and send your own request headers, with the content length filled in. You may have to multi-part mime encode the stream on its way out as well (I'm not about that...).
    You can use the following:
    http://porsche.cis.udel.edu:8080/cis479/lectures/slides-04/slide-02.html
    on your server to get a feel for the format that the request headers need to take.
    - Kevin
    I get the out of Memory Error on the client side. I
    was told that this might be a bug in the URLConnection
    class implementation that basically it wont know the
    content length until all the data has been written to
    the output stream, so it uses an in memory buffer to
    store the data which basically causes memory issues..
    do you think there might be a workaround of any kind..
    or maybe a way that the buffer might be flushed after
    a certain size of file has been uploaded.. ?? do you
    have any ideas?

  • JavaScript Out of Memory Error on Portal timeout.

    Hello All,
    I am using jsf and Inline navigation in all our portlets and when user leave the browser idle for portal timeout we have 2 problems. 1: Login portlet shows in that specific portlet. 2: we get a javascript alert saying out of memory at line 40. and the porltet shows error message as "Gateway was not able to access requested content. If the error persists, contact your portal Administrator."
    We are using Plumtree 5.0.4 Java version.
    any help is highly appreciated.
    Thanks
    A.J.

    Both are valid behaviors unfortunately.
    1) login portlet is showing up in specific portlet b/c inline navigation allows for you to create and load pages without affecting the overall portal.
    This happens when you use iframes (which behave in a similar fashion).
    - your only workaround is really to write some javascript function to "listen" to the portal login page getting loaded and then throwing the session into the parent browser (which is Portal). At least this is the only solution that I ever came up with when using Iframes.
    2) Don't know about out of memory error actually, but getting the "gatewy was not able to access requested content" is valid b/c the session died.
    - javascript errors require javascript solutions. Sorry I couldn't be more helpful than that.
    Maybe someone else will have better suggestions.
    The other suggestion is to use your app server to listen to the logout event and redirect appropriately to somewhere else, or have it do what you want it to do in situations as this.

  • ERROR [B3108]: Unrecoverable out of memory error during a cluster operation

    We are using Sun Java(tm) System Message Queue Version: 3.5 SP1 (Build 48-G). We are using two JMS servers as a cluster.
    But we frequently getting the out of memory issue during the cluster operation.
    Messages also got queued up in the Topics. Eventhough listeners have the capability to reconnect with the Server after the broker restarting, usually we are restarting consumer instances to get work this.
    Here is detailed log :
    Jan 5 13:45:40 polar1-18.eastern.com imqbrokerd_cns-jms-18[8980]: [ID 478930 daemon.error] ERROR [B3108]: Unrecoverable out of memory error during a cluster operation. Shutting down the broker.
    Jan 5 13:45:57 polar1-18.eastern18.chntva1-dc1.cscehub.com imqbrokerd: [ID 702911 daemon.notice] Message Queue broker terminated abnormally -- restarting.
    Expecting your attention on this.
    Thanks

    Hi,
    If you do not use any special cmdline options, how do you configure your servers/
    brokers to 1 Gb or 2 Gb JVM heap ?
    Regarding your question on why the consumers appear to be connecting to just
    one of the brokers -
    How are the connection factories that the consumers use configured ?
    Is the connection factory configured using the imqAddressList and
    imqAddressListBehavior attributes ? Documentation for this is at:
    http://docs.sun.com/source/819-2571/ref_adminobj_props.html#wp62463
    imqAddressList should contain a list of brokers (i.e. 2 for you) in the cluster
    e.g.
    mq://server1:7676/jms,mq://server2:7676/jms
    imqAddressListBehavior defines how the 2 brokers in the above list are picked.
    The default is in the order of the list - so mq://server1:7676/jms will always be
    picked by default. If you want random behavior (which will hopefully even out the
    load), set imqAddressListBehavior to RANDOM.
    regards,
    -i
    http://www.sun.com/software/products/message_queue/index.xml

  • Out of Memory error on some PDF's, not all PDF's.

    Hi there,
    I have read most of the posts in the forums that relate to 'Out of Memory' issues with PDF's and I have to say that there is still no solution that I have found.
    I have tried reinstalling Adobe Reader, Flash Player and tried clearing my Temp Files. None of these fixed the issue.
    The PDF's that receive this memory error are downloaded off a CMS website that has many offices and logins. Only one office is experiencing this out of memory issue so we know that it is not an issue with generating the PDF's on the CMS website, otherwise all the offices and logins would have this issue, since they use the same system.
    I am an admin of that CMS website and even I receive the same out of memory issue when I try and view the PDF's from that office, as the customer does.
    When we open these PDF's with another PDF reading program, the PDF's open fine. So that tells me that the issue lies with Adobe Reader and not the PDF file itself.
    These PDF files that are receiving the error about 1MB-4MB, so they are not large.
    Could you please tell me the solution to this out of memory error as we may lose a customer because your product is producing an error that does not seem to have been fixed after years of being reported by your customers.
    Thanks.

    As an Adobe Reader XI user, if I may put my two cents in, that "Out of Memory" problem is not with the Adobe Reader XI application, which was installed on our computers! I think, the problem is with how the PDF documents, which we were trying to open, were generated (or regenerated).
    I remember I was able to open my old credit card statement online without any problem and now I am not able to open the same old statement (out of memory) because it could have been regenerated! In fact, I could not open any of my credit card statements (old or new) on this specific credit card web site, which makes me believe those statments were regenerated or somthing with some new Adobe software.
    As others mentioned, even I'm able to view other PDF documents without any problem.

  • Out of memory error on Hyperion Analyzer

    When I tried to access a report through Hyperion analyzer HTML client, Out Of memory error comes up.I tried changing the virtual memory on windows 2000.Even then I am experiencing the same problem.Environment:Windows2000/Tomcat 4.04/IIS/Analyzer 6.2.1.Is there any parameter or setting to be set on properties file on App server side?Can anyone help me out of this problem?Thanks in advance.Jaya.

    How much RAM do you have on the analyzer server. You may need more... In the mean time try adjusting these settingsThe analyzer_startup.bat or file can be edited to change the amount of memory allocated to the Java Runtime Environment (JRE). Instead of: set CATALINA_OPTS=-Xms128m -Xmx256m -Xss2mChanged to: set CATALINA_OPTS=-Xms256m -Xmx512m -Xss4mFor Tomcat, 1024m is the highest setting for –Xmx.

  • Import Manager Out of Memory error

    I am running MDM 5.5.24.06 server on Windows XP Professional. The server has 4.0 GB RAM with 1.5 GB Virtual memory.
    I am trying to load 129 material master records from R/3 4.6 ( XML file size 8 MB), into MDM using Import Manager.
    When I click on import icon (Import Status tab selected), MDM Import Manager rightway returns 'Out of Memory' error.
    Tried to import the file again after rebooting the machine, restarting MDM Server, but still it returned the same error.
    Has anybody tried loading the R/3 material data into MDM? What's the experience? What was the load (# of records)?
    What was MDM IM performance? What's your hardware/software config?
    Appreciate your help.
    Thanks,
    Abhay

    Hi Abhay,
    My workstation has Windows 2000, Pentium 4 CPU 3.2 Ghz, 515,555 KB RAM.  The recommended specifications can be found on service market place of SAP, and we used those to procure our workstations.
    MDM server must be more pwerful than a workstation.  Try running the import manager from the server machine itself.  If it works there but not on your machine, it is memory of your workstation.
    Sometimes, I think this error comes when import map some issues- fields not mapped accurately or remote key mapped wrong etc.
    It is probably judicious to rpove the system memory as first step.
    Good luck!

Maybe you are looking for