Solaris 10 Zones: Out of Memory

I have struggled with installing Oracle 11.2 in a guest zone on Solaris 10 (SPARC). The machine has 16GB RAM and the zone is capped to 4GB, which ought to be enough; and project.max-shm-memory is 6GB for the default project. Still, when I try to run dbca, I get and "Out of memory" error very early on. The error occurs 3 times in a row, just after a log entry saying "starting with pfile=..."
There is obviously a parameter or several that I need to set; but the docs aren't telling. Does anybody know?

Hi;
Please check below notes and see its helpful:
Init.ora Parameter "SGA_MAX_SIZE" Reference Note [ID 153377.1]
Database Release 11.2 Released on Sparc Solaris [ID 966152.1]
Oracle Database on Unix AIX,HP-UX,Linux,Mac OS X,Solaris,Tru64 Unix Operating Systems Installation and Configuration Requirements Quick Reference (8.0.5 to 11.2) [ID 169706.1]
Direct I/O Not Used on SUN SPARC Solaris with filesystemio_options = DIRECTIO or SETALL [ID 555607.1]
Regard
Helios

Similar Messages

  • SAP AS JAVA installation in Solaris Zone envirn throws OUT of Memory error.

    Hi,
    We are installing SAP NW2004s 7.0 SR3 AS JAVA on Solaris 10 in zone environment. Thsi was prod server getting build where on top of this we will install GRC 5.3.
    We have faced no issues in development build.
    But during the Prod build at databse create step, we are getting the below error
    ORA-27102: out of memory
    SVR4 Error: 22: Invalid argument
    Disconnected
    SAPinst log entries where error messages started:
    04:43:58.128
    Execute step createDatabase of component |NW_Onehost|ind|ind|ind|ind|0|0|NW_Onehost_System|ind|ind|ind|ind|2|0|NW_CreateDBandLoad|ind|ind|ind|ind|10|0|NW_CreateDB|ind|ind|ind|ind|0|0|NW_OraDBCheck|ind|ind|ind|ind|0|0|NW_OraDBMain|ind|ind|ind|ind|0|0|NW_OraDBStd|ind|ind|ind|ind|3|0|NW_OraDbBuild|ind|ind|ind|ind|5|0
    INFO 2011-04-01 04:45:14.590
    Working directory changed to /tmp/sapinst_exe.16718.1301647358.
    INFO 2011-04-01 04:45:14.595
    Working directory changed to /tmp/sapinst_instdir/NW04S/SYSTEM/ORA/CENTRAL/AS.
    INFO 2011-04-01 04:45:14.609
    Working directory changed to /tmp/sapinst_exe.16718.1301647358.
    INFO 2011-04-01 04:45:14.621
    Working directory changed to /tmp/sapinst_instdir/NW04S/SYSTEM/ORA/CENTRAL/AS.
    INFO 2011-04-01 04:45:14.850
    Account oraac5 already exists.
    INFO 2011-04-01 04:45:14.852
    Account dba already exists.
    INFO 2011-04-01 04:45:14.852
    Account oraac5 already exists.
    INFO 2011-04-01 04:45:14.853
    Account dba already exists.
    INFO 2011-04-01 04:45:14.867
    Working directory changed to /tmp/sapinst_exe.16718.1301647358.
    INFO 2011-04-01 04:45:14.899
    Working directory changed to /tmp/sapinst_instdir/NW04S/SYSTEM/ORA/CENTRAL/AS.
    ERROR 2011-04-01 04:45:32.280
    CJS-00084  SQL statement or script failed. DIAGNOSIS: Error message: ORA-27102: out of memory
    SVR4 Error: 22: Invalid argument
    Disconnected
    . SOLUTION: See ora_sql_results.log and the Oracle documentation for details.
    ERROR 2011-04-01 04:45:32.286
    MUT-03025  Caught ESAPinstException in Modulecall: ORA-27102: out of memory
    SVR4 Error: 22: Invalid argument
    Disconnected
    ERROR 2011-04-01 04:45:32.453
    FCO-00011  The step createDatabase with step key |NW_Onehost|ind|ind|ind|ind|0|0|NW_Onehost_System|ind|ind|ind|ind|2|0|NW_CreateDBandLoad|ind|ind|ind|ind|10|0|NW_CreateDB|ind|ind|ind|ind|0|0|NW_OraDBCheck|ind|ind|ind|ind|0|0|NW_OraDBMain|ind|ind|ind|ind|0|0|NW_OraDBStd|ind|ind|ind|ind|3|0|NW_OraDbBuild|ind|ind|ind|ind|5|0|createDatabase was executed with status ERROR ( Last error reported by the step :Caught ESAPinstException in Modulecall: ORA-27102: out of memory
    SVR4 Error: 22: Invalid argument
    Disconnected
    ora_sql_results.log
    04:45:15 SAPINST ORACLE start logging for
    SHUTDOWN IMMEDIATE;
    exit;
    Output of SQL executing program:
    SQL*Plus: Release 10.2.0.4.0 - Production on Fri Apr 1 04:45:15 2011
    Copyright (c) 1982, 2007, Oracle.  All Rights Reserved.
    Connected to an idle instance.
    ORA-01034: ORACLE not available
    ORA-27101: shared memory realm does not exist
    SVR4 Error: 2: No such file or directory
    Disconnected
    SAPINST: End of output of SQL executing program /oracle/AC5/102_64/bin/sqlplus.
    SAPINST found errors.
    SAPINST The current process environment may be found in sapinst_ora_environment.log.
    2011-04-01, 04:45:15 SAPINST ORACLE stop logging
    ================================================================================
    2011-04-01, 04:45:15 SAPINST ORACLE start logging for
    STARTUP NOMOUNT;
    exit;
    Output of SQL executing program:
    SQL*Plus: Release 10.2.0.4.0 - Production on Fri Apr 1 04:45:15 2011
    Copyright (c) 1982, 2007, Oracle.  All Rights Reserved.
    Connected to an idle instance.
    ORA-27102: out of memory
    SVR4 Error: 22: Invalid argument
    Disconnected
    SAPINST: End of output of SQL executing program /oracle/AC5/102_64/bin/sqlplus.
    SAPINST found errors.
    SAPINST The current process environment may be found in sapinst_ora_environment.log.
    2011-04-01, 04:45:32 SAPINST ORACLE stop logging
    Already viewed S-note
    724713 - parameter settings for Solaris 10 - (Parameters are set as per this note)
    743328 - Composite SAP note: ORA-27102 - (no much information in memory problem on zones)
    Please provide your suggestions/resolution.
    Thankyou.

    Hi,
    @ Sunny: Thanks for response, the referred note was already checked and parameters are in sync as per note.
    @Mohit: SAP wouldn't proceed till create database if Oracle software was not installed. thanks for the response.
    @Markus: Thanks I agree with you, but I have doubt in this area. Isn't it like project.max-shm-memory was new parameter we need to set in local zone rather using shmsys:shminfo_shmmax in /etc/system. Do we need to still maintain this parameter in /etc/system in global zone.
    As per SUN doc below parameter was obsolete from Solaris 10.
    The following parameters are obsolete.
    ■ shmsys:shminfo_shmmni
    ■ shmsys:shminfo_shmmax
    As per your suggestion, do we need to set below parameters in that case, please clarify.
    Parameter                           Replaced by Resource Control      Recommended Value
    semsys:seminfo_semmni   project.max-sem-ids                      100
    semsys:seminfo_semmsl   process.max-sem-nsems               256
    shmsys:shminfo_shmmax  project.max-shm-memory             4294967295
    shmsys:shminfo_shmmni   project.max-shm-ids                       100
    Also   findings of /etc/release
    more /etc/release
    Solaris 10 10/08 s10s_u6wos_07b SPARC
    Regards,
    Sitarama.

  • Ora-27102: out of memory error on Solaris 10 SPARC zone install or Oracle10

    All
    I'm stuck!! Tried for days now ,and can't get this working.
    I"m getting the classic: "ora-27102: out of memory" error. However, my memory settings seem fine.
    I'm running Solaris 10 on a Zone, while installing Oracle 10.2.0.1.
    I've changed the max-memory to 6Gig's, but still, I get this error message.
    Could it be possible that this error message means something else entirely?
    Thank you for your assistance!!
    Anne

    Hi V
    Thanks for the response.
    I fixed the problem. Turns out it was because my physical memory for box is 16 gig, but my max-shm-memory was only at 6 GIG. I upped to 8 gig, and everything worked.
    I'm sure you were going to tell me to do just that! I found another post that explained it.
    Thanks!
    Anne

  • ORA-27102: out of memory Solaris 10 Oracle 11g Container

    Hi,
    I work with SPARCT3 32Gb Ram, Solaris 10. Trabajo con contenedores/zonas.
    I try to install some DB in the same container.
    The max-shm-memory variable set to 8G, 10G, 12G, 14G and 16G but its the same.
    The first installation its work, and i set the SGA to 2G, in the second installation with dbca, it showme the out of memory, i had to turn off the first installation to complete the second.
    When the second has install, i try to startup the first installation but it showme the out of memory.
    # prctl -n project.max-shm-memory -i project 100
    project: 100: pro.oracle
    NAME PRIVILEGE VALUE FLAG ACTION RECIPIENT
    project.max-shm-memory
    privileged 10.0GB - deny -
    system 16.0EB max deny -
    In the alert log only show me this:
    Wed Apr 18 13:03:42 2012
    Adjusting the default value of parameter parallel_max_servers
    from 1280 to 135 due to the value of parameter processes (150)
    Starting ORACLE instance (normal)
    Whe i install and i set the SGA to 2Gb, Oracle get this 2Gb or it wants more?
    Thanks.

    PVOWUSU wrote:
    Result of the command is below:
    [root@AGB-DC-12 ~]# find / -name alert_ghadb.log -print 2>/dev/nul
    /ORACLE/product/10.2.0/db_1/rdbms/log/alert_ghadb.log
    /ORACLE/admin/ghadb/bdump/alert_ghadb.log
    [root@AGB-DC-12 ~]#
    display from /ORACLE/product/10.2.0/db_1/rdbms/log/alert_ghadb.log
    [root@AGB-DC-12 ~]# vi /ORACLE/product/10.2.0/db_1/rdbms/log/alert_ghadb.log
    Tue Mar 20 00:02:21 2012
    Adjusting the default value of parameter parallel_max_servers
    from 1280 to 135 due to the value of parameter processes (150)
    Tue Mar 20 00:03:25 2012
    Shutting down instance (abort)
    Thanks.how many other Oracle DB are currently running on this system?
    who/what made spfile or pfile used to attempt starting this new DB?
    post results from following OS command
    ls -ltr $ORACLE_HOME/dbs/

  • Solaris 10, Oracle 10g and ORA-27102 out of memory

    After much STFW'ing, RTFM'ing and experimenting I finally got Oracle 10g to start a database on a SunBlade 100 with 512MB of memory and Solaris 10. As most of the replies on the web about error 27102 "out of memory" address solutions in Solaris 8 and 9, I thought I should post a clear description of the solution here.
    Unlike earlier releases of Solaris, most of the system parameters needed to run Oracle are already set properly, so the only one you need is the maximum shared memory parameter. In earlier versions this was called SHMMAX and was set by editing the /etc/system file and rebooting. With Solaris 10 you set this by modifying a "Resource Control Value". You can do this temporarily by using prctl, but that is lost at reboot so you will need to add the command to the oracle user's .profile. The other option is to create a default project for the oracle user:
    projadd -U oracle -K "project.max-shm-memory=(priv,2048MB,deny)" user.oraclewhat this does:
    - makes a project named "user.oracle" with the user oracle as it's only member.
    - because the name was of the form "user.username" it becomes the oracle user's default project.
    - the value of the maximum shared memory is set to 2gb, you might want to use a larger value here if you have more memory and swap.
    - no reboot is needed, the user will get the new value at their next login.

    thanks buddy. I was in this same tricky situation. and ur post solved my prblem.
    Is this a bug with solaris 10.
    Later i discovered.. on Oracle documentation that.. there r sum parameters that needed to be taken care while pre-installation of Oracle binaries.
    Here is the excerpts.
    The Oracle installer recognizes kernel parameters set using this method, but it is now deprecated in favour of resource control projects, explained below.
    As the root user, issue the following command.
    projadd oracle
    Append the following line to the "/etc/user_attr" file.
    oracle::::project=oracle
    If you've performed a default installation, it is likely that the only kernel parameter you need to alter is "max-shm-memory". To check the current value issue the following command.
    # prctl -n project.max-shm-memory -i project oracle
    project: 100: oracle
    NAME    PRIVILEGE       VALUE    FLAG   ACTION                       RECIPIENT
    project.max-shm-memory
            privileged       254MB      -   deny
            system          16.0EB    max   deny     
    To reset this value, make sure at least one session is logged in as the oracle user, then from the root user issue the following commands.
    # prctl -n project.max-shm-memory -v 4gb -r -i project oracle
    # projmod -s -K "project.max-shm-memory=(priv,4gb,deny)" oracle
    The first dynamically resets the value, while the second makes changes to the "/etc/project" file so the value is persistent between reboots.
    # cat /etc/project
    system:0::::
    user.root:1::::
    noproject:2::::
    default:3::::
    group.staff:10::::
    oracle:100::::project.max-shm-memory=(priv,4294967296,deny)
    The Oracle installer seems incapable of recognising kernel parameter set using resource control projects, but if you ignore the warnings the installation completes successfully.Even this method works.
    Regards
    Monu Koshy

  • ORA-27102: out of memory. Faild to install oracle 10gR2 on Solaris 10

    Hi, I want to install oracle on my solaris machine. I have 2.5G RAM and more than 5G swap file. But the ORA-27102: out of memory error occurred at installing the Oracle Database Configuration Assistant step(Copying database files Creating and starting Oracle instance).
    The only warnings bellow:
    Checking kernel parameters
    Checking for BIT_SIZE=64; found BIT_SIZE=64. Passed
    Checking for shmsys:shminfo_shmmax=4294967295; found no entry. Failed <<<<
    Checking for shmsys:shminfo_shmmni=100; found no entry. Failed <<<<
    Checking for semsys:seminfo_semmni=100; found no entry. Failed <<<<
    Checking for semsys:seminfo_semmsl=256; found no entry. Failed <<<<
    Check complete. The overall result of this check is: Failed <<<<
    Problem: The kernel parameters do not meet the minimum requirements (see above).
    What's the reason and how may I resolve it?
    Thanks!

    I set some kernel parameters like: set shmsys:shminfo_shmmax=4294967295 and so on. Restart the computer and run dbca+ and configure in advance mode. There's a stage require the size of flash_recover_segment of which the default size number is 2048M. WOW, out of memory on my pc of course.
    So I think it's the parameter value during the installation progress bring on the ORA-27102 problem. However, during the first installation, there's no prompt for me to enter the parameter; or I didn't see.

  • How can I avoid running out of memory when creating components dynamically

    Hello everyone,
    Recently, I am planning to design a web application. It will be used by all middle school teachers in a region to make examination papers and it must contain the following main functions.
    1)Generate test questions dynamically. For instance, a teacher who logs on the web application will only see a select one menu and a Next Quiz button. The former is used for determining the number of options for the current multiple/single choice question. The later is dedicated to creating appropriate input text elements according to the selected option number. That is to say, if the teacher selects 4 in the menu and presses the Next Quiz button, 5 input text form elements will appear. The first one is for the question to be asked such as "1.What is the biggest planet in the solar system?", the others are optional answers like a)Uranus. b) Saturn. c)Jupiter. d)Earch. Each answer stands for an input text elements. When the teacher fills in the fourth answer, another select one menu and Next Quiz button will emerge on the fly just under this answer, allowing the teacher to make the second question. The same thing repeats for the following questions.
    2)Undo and Redo. Whenever a teacher wants to roll back or redo what he has done, just press the Undo or[i] Redo button. In the previous example, if the teacher selects the third answer and presses the Delete button to drop this answer, it will delete both the literal string content[i] and the input text element, changing the answer d to c automatically. After that, he decides to get back the original answer c, Jupiter, he can just click the Undo button as if he hadn�ft made the deleting operation.
    3)Save the unfinished working in the client side. If a teacher has done half of his work, he can choose to press the Save button to store what he has done in his computer. The reason for doing so is simply to alleviate the burden of the server. Although all finished test papers must be saved in a database on the server, sometimes the unfinished papers could be dropped forever or could form the ultimate testing papers after several months. So if these papers keep in the server, it will waste the server computer�fs room. Next time the teacher can press the Restore button on the page to get the previously stored part of the test paper from his own computer and continue to finish the whole paper.
    4)Allow at least 1,000 teachers to make test papers at the same time. The maximum question number per examination paper is 60.
    Here are my two rough solutions,
    A.Using JSF.
    B.Using JavaScript and plain JSP[b] without JSF.
    The comparison of the two solutions:
    1)Both schemas can implement the first and the second requirements. In JSF page I could add a standard panelGird tag and use its binding attribute. In the backing bean, the method specified by the binding attribute is responsible for generating HtmlInput objects and adding them to the HtmlPanelGird object on the fly. Every HtmlInput object is corresponding to a question subject or an optional answer. The method is called by an actionListener, which is registered in the Next Quiz commandButton, triggering by the clicking on this button in the client side. Using JSF can also be prone to managing the HtmlInput objects, e.g. panelGird.getChildren().add(HtmlInput) and panelGird.getChildren().remove(HtmlInput) respond to the undoing operation of deleting an optional answer and the redoing operation of the deleting action respectively. I know JavaScript can also achieve these goals. It could be more complex since I don�ft know well about JavaScript.
    2)I can not find a way to meet the third demand right now. I am eager to know your suggestion.
    3)Using JSF, I think, can�ft allow 1,000 teachers to do their own papers at the same time. Because in this scenario, suppose each questionnaire having 60 questions and 4 answers per question, there will be approximately 300,000 HtmlInput objects (1,000X60X(4+1)) creating on the server side. The server must run out of memory undoubtedly. To make things better, we can use a custom component which can be rendered as a whole question including its all optional answers. That is to say, a new custom component on the server side stands for a whole question on the client side. Even so, about 60,000(1,000X60) this type of custom components will be created progressively and dynamically, plus other UISelectOne and UICommand objects, it also can�ft afford for most servers. Do I have to use JavaScript to avoid occupying the server's memory in this way? If so, I have to go back and use JavaScript and plain JSP without JSF.
    Thank you in advance!
    Best Regards, Ailsa
    2007/5/4

    Thank you for your quick response, BalusC. I really appreciate your answer.
    Yes, you are right. If I manually code the same amount of those components in the JSF pages instead of generating them dynamically, the server will still run out of memory. That is to say, JSF pages might not accommodate a great deal of concurrent visiting. If I upgrade the server to just allow 1,000 teachers making their own test papers at the same time, but when over 2,000 students take the same questionnaire simultaneously, the server will need another upgrading. So I have to do what you have told me, using JS+DOM instead of upgrading the server endlessly.
    Best Regards, Ailsa

  • Oracle 9i running out of memory

    Folks !
    I have a simple 3 table schema with a few thousand entries each. After dedicating gigabytes of hard disk space and 50% of my 1+ GB memory, I do a few simple Oracle Text "contains" searches (see below) on these tables and oracle seems to grow some 25 MB after each query (which typically return less than a dozen rows each) till it eventually runs out of memory and I have to reboot the system (Sun Solaris).
    This is on Solaris 9/Sparc with Oracle 9.2 . My query is simple right outer join. I think the memory growth is related to Oracle Text index/caching since memory utilization seems pretty stable with simple like '%xx%' queries.
    "top" shows a dozen or so processes each with about 400MB RSS/SIZE. It has been a while since I did Oracle DBA work but I am nothing special here. Databse has all the default settings that you get when you create an Oracle database.
    I have played with SGA sizes and no matter how large or small the size of SGA/PGA, Oracle runs out of memory and crashes the system. Pretty stupid to an Enterprise databas to die like that.
    Any clue on how to arrest the fatal growth of memory for Oracle 9i r2?
    thanks a lot.
    -Sanjay
    PS: The query is:
    SELECT substr(sdn_name,1,32) as name, substr(alt_name,1,32) as alt_name, sdn.ent_num, alt_num, score(1), score(2)
    FROM sdn, alt
    where sdn.ent_num = alt.ent_num(+)
    and (contains(sdn_name,'$BIN, $LADEN',1) > 0 or
    contains(alt_name,'$BIN, $LADEN',2) > 0)
    order by ent_num, score(1), score(2) desc;
    There are following two indexes on the two tables:
    create index sdn_name on sdn(sdn_name) indextype is ctxsys.context;
    create index alt_name on alt(alt_name) indextype is ctxsys.context;

    I am already using MTS.
    Atached is the init.ora file below.
    may be I should repost this article with subject "memory leak in Oracle" to catch developer attention. I posted this a few weeks back in Oracle Text groiup and no response there either.
    Thanks for you help.
    -Sanjay
    # Copyright (c) 1991, 2001, 2002 by Oracle Corporation
    # Cache and I/O
    db_block_size=8192
    db_cache_size=33554432
    db_file_multiblock_read_count=16
    # Cursors and Library Cache
    open_cursors=300
    # Database Identification
    db_domain=""
    db_name=ofac
    # Diagnostics and Statistics
    background_dump_dest=/space/oracle/admin/ofac/bdump
    core_dump_dest=/space/oracle/admin/ofac/cdump
    timed_statistics=TRUE
    user_dump_dest=/space/oracle/admin/ofac/udump
    # File Configuration
    control_files=("/space/oracle/oradata/ofac/control01.ctl", "/space/oracle/oradata/ofac/control02.ctl", "/space/oracle/oradata/ofac/control03.ctl")
    # Instance Identification
    instance_name=ofac
    # Job Queues
    job_queue_processes=10
    # MTS
    dispatchers="(PROTOCOL=TCP) (SERVICE=ofacXDB)"
    # Miscellaneous
    aq_tm_processes=1
    compatible=9.2.0.0.0
    # Optimizer
    hash_join_enabled=TRUE
    query_rewrite_enabled=FALSE
    star_transformation_enabled=FALSE
    # Pools
    java_pool_size=117440512
    large_pool_size=16777216
    shared_pool_size=117440512
    # Processes and Sessions
    processes=150
    # Redo Log and Recovery
    fast_start_mttr_target=300
    # Security and Auditing
    remote_login_passwordfile=EXCLUSIVE
    # Sort, Hash Joins, Bitmap Indexes
    pga_aggregate_target=25165824
    sort_area_size=524288
    # System Managed Undo and Rollback Segments
    undo_management=AUTO
    undo_retention=10800
    undo_tablespace=UNDOTBS1

  • Problem with Java in solaris 10 zone

    HI ,
    i get this error message in my solaris 10 zone :
    #./java version
    Exception java.lang.OutOfMemoryError: requested -4 bytes for size_t in /BUILD_AREA/jdk1.5.0_11/hotspot/src/os/solaris/vm/os_solaris.cpp. Out of swap space?
    in global zone , java works fine !
    uname -a
    SunOS Server 5.10 Generic_141414-07 sun4u sparc SUNW,SPARC-Enterprise
    best regards

    I see mention of a similar problem, but only with early development releases of solaris10 branded zones running on Solaris 11 development bits. It seems as though this is not what you are running.
    Since there are no swap caps nor brand emulation in place, the next thing I would suspect would be that you somehow have inconsistencies between key libraries (e.g. libc) and the kernel. This could happen if you have a full-root zone that has been force attached (zoneadm -z <zone> attach -F). In the past, I've seen products that do this under the covers (e.g. Veritas Cluster) so it may have happened without your knowledge. If it is a sparse root zone, you may want to run "zoneadm -z <zone> detach; zoneadm -z <zone> attach -u". Prior to doing so, take whatever precautions you would normally take when patching a system.
    If you open a support call about this I bet the first advice you are going to get is to patch, as you seem to be 2+ years behind. Also, newer releases of Java 5 are available at http://www.oracle.com/technetwork/java/javasebusiness/downloads/java-archive-downloads-javase5-419410.html.

  • Solaris 10 zone config for JES

    Setup a zone to install JES - Web servers, but when I launch the installer I get the following:
    WARNING: Unsupported non-global zone
    EntSysResources:zoneCheckPanel-Gui-LocalZoneSupported
    Something in the zone setup is incorrect, but not sure what.
    Anyone else out running JES in Solaris 10 zones?
    Thanks!
    Steve
    [email protected]

    From docs.sun.com -
    The following Java ES and zones scenarios are supported in this release of Java ES:
    * Installing Java ES in a non-global zone that uses the whole root file system (Using the default sparse root file system is not supported.)
    * Installing Java ES in the global zone with no non-global zones
    Zone can be 2 type - sparse root and whole root . sparse root model has following folders inherited - /lib, /platform, /sbin and /usr.
    If a non-global zone doesn't have any inherit-pkg-dir resources, such a zone is said to be a whole root one
    If you want to create a whole root zone but default shared file systems resources have been added by using inherit-pkg-dir, you must remove these default inherit-pkg-dir resources using zonecfg before you install the zone:
    zonecfg:my-zone> remove inherit-pkg-dir dir=/lib
    zonecfg:my-zone> remove inherit-pkg-dir dir=/platform
    zonecfg:my-zone> remove inherit-pkg-dir dir=/sbin
    zonecfg:my-zone> remove inherit-pkg-dir dir=/usr

  • Inetd service/program crashes with core dump in Solaris 8 zone/container

    I have developed a service in C that is launched from inetd when something comes on a specific port.
    When a connection is opened to the port a core dump is created in the same directory where the executable file is located.
    If you run the same service program from the command line everything is working perfect.
    This is running in a Solaris 8 zone/container on a Solaris 10 machine.
    Everything is set correctly in /etc/inetd.conf and in /etc/services.
    I have even stripped down the program to a hello world program that is just printing a string to the screen and it is still crashing with a core dump.
    # ldd test_srv
    /usr/lib/secure/s8_preload.so.1
    libc.so.1 => /usr/lib/libc.so.1
    libdl.so.1 => /usr/lib/libdl.so.1
    /usr/platform/SUNW,Sun-Fire-480R/lib/libc_psr.so.1
    The same service is running on a Linux machine and on a Solaris 10 machine without zones/containers without any problems.
    Can you please help me figure out what am I missing. Is there something specific with zones/containers that should be set / configured?
    Do I have to set some specific env. variables to work in a Solaris 8 zone/container environment or is it something very simple that I'm missing?

    Could you please examine the truss log and advice what the problem is and how to fix it?
    (some lines deleted)
    bash-2.03# truss -f -p 18361 #### /usr/sbin/inetd -s -t &
    18361:  poll(0xFFBFF528, 53, -1)        (sleeping...)
    18361:  poll(0xFFBFF528, 53, -1)                        = 1
    18361:  accept(63, 0xFFBFF870, 0xFFBFF914, 1)           = 3
    18361:  sigprocmask(SIG_BLOCK, 0xFFBFF5F0, 0xFFBFF600)  = 0
    18361:  lwp_sigtimedwait(0xFFBFF600, 0xFFBFF578, 0x00000010) = 0
    18361:  lwp_sigtimedwait(0xFFBFF568, 0xFFBFF728, 0x00000010) = 0
    18361:  fork()                                          = 1921
    1921:   fork()          (returning as child ...)        = 18361
    1921:   sigprocmask(0, 0x00000000, 0xFFBFF600)          = 0
    18361:  sigprocmask(0, 0x00000000, 0xFFBFF600)          = 0
    1921:   lwp_sigtimedwait(0xFFBFF600, 0xFFBFF578, 0x00000010) = 0
    18361:  sigprocmask(SIG_SETMASK, 0xFFBFF5F0, 0xFFBFF600) = 0
    18361:  close(3)                                        = 0
    18361:  sigprocmask(0, 0x00000000, 0xFFBFF600)          = 0
    1921:   lwp_sigtimedwait(0xFFBFF668, 0xFFBFF528, 0x00000020) = 0
    1921:   sigaction(SIGHUP, 0xFFBFF528, 0xFFBFF500)       = 0
    18361:  lwp_sigtimedwait(0xFFBFF568, 0xFFBFF5F0, 0x00000010) = 0
    1921:   lwp_sigtimedwait(0xFFBFF508, 0xFFBFF458, 0x00000010) = 0
    18361:  sigprocmask(SIG_SETMASK, 0xFFBFF5F0, 0xFFBFF600) = 0
    1921:   sigprocmask(SIG_SETMASK, 0xFFBFF5F0, 0xFFBFF600) = 0
    1921:   lwp_sigtimedwait(0xFFBFF600, 0xFFBFF578, 0x00000010) = 0
    1921:   lwp_sigtimedwait(0xFFBFF568, 0xFFBFF728, 0x00000010) = 0
    1921:   fcntl(3, F_DUP2FD, 0x00000000)                  = 0
    1921:   close(3)                                        = 0
    1921:   fcntl(0, F_DUP2FD, 0x00000001)                  = 1
    1921:   fcntl(0, F_DUP2FD, 0x00000002)                  = 2
    1921:   open64("/etc/.name_service_door", O_RDONLY)     = 3
    1921:   fcntl(3, F_SETFD, 0x00000001)                   = 0
    1921:   door_info(3, 0xFF0C2748)                        = 0
    1921:   door_call(3, 0xFFBFF278)                        = 0
    1921:   close(67)                                       Err#9 EBADF
    1921:   close(66)                                       Err#9 EBADF
    1921:   close(65)                                       Err#9 EBADF
    1921:   close(64)                                       Err#9 EBADF
    1921:   close(63)                                       = 0
    1921:   close(62)                                       = 0
    1921:   close(12)                                       = 0
    1921:   close(11)                                       = 0
    1921:   close(10)                                       Err#9 EBADF
    1921:   close(9)                                        Err#9 EBADF
    1921:   close(8)                                        Err#9 EBADF
    1921:   close(7)                                        Err#9 EBADF
    1921:   close(6)                                        Err#9 EBADF
    1921:   close(5)                                        Err#9 EBADF
    1921:   close(4)                                        Err#9 EBADF
    1921:   setrlimit(RLIMIT_NOFILE, 0xFFBFFD20)            = 0
    1921:   xenix(398872, 0xFFBFF5E4, 0x00000040)           = 38
    1921:   execve("/tmp/srv/t_srv", 0x0008B5FC, 0xFFBFFDA0)  argc = 0
    1921:   getuid()                                        = 0 [0]
    1921:   resolvepath("/usr/lib/ld.so.1", "/usr/lib/ld.so.1", 1023) = 16
    1921:   open("/var/ld/ld.config", O_RDONLY)             = 3
    1921:   fstat(3, 0xFFBFF5E8)                            = 0
    1921:   mmap(0x00000000, 148, PROT_READ, MAP_SHARED, 3, 0) = 0xFF3E0000
    1921:   close(3)                                        = 0
    1921:   stat("/usr/lib/libc.so.1", 0xFFBFF648)          = 0
    1921:   resolvepath("/usr/lib/libc.so.1", "/usr/lib/libc.so.1", 1023) = 18
    1921:   open("/usr/lib/libc.so.1", O_RDONLY)            = 3
    1921:   mmap(0x00000000, 8192, PROT_READ|PROT_EXEC, MAP_PRIVATE, 3, 0) = 0xFF340000
    1921:   mmap(0x00000000, 802816, PROT_NONE, MAP_PRIVATE|MAP_NORESERVE|MAP_ANON, -1, 0) = 0xFF200000
    1921:   mmap(0xFF200000, 703520, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED, 3, 0) = 0xFF200000
    1921:   mmap(0xFF2BC000, 24772, PROT_READ|PROT_WRITE|PROT_EXEC, MAP_PRIVATE|MAP_FIXED, 3, 704512) = 0xFF2BC000
    1921:   munmap(0xFF2AC000, 65536)                       = 0
    1921:   memcntl(0xFF200000, 113528, MC_ADVISE, MADV_WILLNEED, 0, 0) = 0
    1921:   close(3)                                        = 0
    1921:   stat("/usr/lib/libdl.so.1", 0xFFBFF648)         = 0
    1921:   resolvepath("/usr/lib/libdl.so.1", "/usr/lib/libdl.so.1", 1023) = 19
    1921:   open("/usr/lib/libdl.so.1", O_RDONLY)           = 3
    1921:   mmap(0xFF340000, 8192, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED, 3, 0) = 0xFF340000
    1921:   mmap(0x00000000, 8192, PROT_NONE, MAP_PRIVATE|MAP_NORESERVE|MAP_ANON, -1, 0) = 0xFF330000
    1921:   mmap(0xFF330000, 2638, PROT_READ|PROT_WRITE|PROT_EXEC, MAP_PRIVATE|MAP_FIXED, 3, 0) = 0xFF330000
    1921:   close(3)                                        = 0
    1921:   stat("/usr/platform/SUNW,Sun-Fire-480R/lib/libc_psr.so.1", 0xFFBFF368) = 0
    1921:   resolvepath("/usr/platform/SUNW,Sun-Fire-480R/lib/libc_psr.so.1", "/usr/platform/sun4u-us3/lib/libc_psr.so.1", 1023) = 41
    1921:   open("/usr/platform/SUNW,Sun-Fire-480R/lib/libc_psr.so.1", O_RDONLY) = 3
    1921:   mmap(0xFF340000, 8192, PROT_READ|PROT_EXEC, MAP_PRIVATE|MAP_FIXED, 3, 0) = 0xFF340000
    1921:   close(3)                                        = 0
    1921:   mmap(0x00000000, 8192, PROT_READ|PROT_WRITE|PROT_EXEC, MAP_PRIVATE|MAP_ANON, -1, 0) = 0xFF320000
    1921:   dup(0)                                          = 3
    1921:   llseek(0, 0, SEEK_CUR)                          Err#29 ESPIPE
    1921:   close(0)                                        = 0
    1921:   fcntl(3, F_DUP2FD, 0x00000000)                  = 0
    1921:   close(3)                                        = 0
    1921:   dup(1)                                          = 3
    1921:   close(1)                                        = 0
    1921:   fcntl(3, F_DUP2FD, 0x00000001)                  = 1
    1921:   close(3)                                        = 0
    1921:   dup(2)                                          = 3
    1921:   close(2)                                        = 0
    1921:   fcntl(3, F_DUP2FD, 0x00000002)                  = 2
    1921:   close(3)                                        = 0
    1921:   sys#177(0x00000080, 0xFFBFFB7C, 0xFF3F0518, 0x00000000, 0xFF3C2EF8, 0xFF2C0284) = 0x00000000 [0xFFBFFB7C]
    1921:   sys#227(0x00000006, 0x00000000, 0x0001ADF0, 0xFF3F0518, 0xFF3C3C18, 0xFF3C2670) = 0x0000000C [0x00000000]
    1921:   sys#227(0x00000002, 0x0000000C, 0x0000000E, 0xFFBFFCAE, 0x00000002, 0xFF3C2670) = 0x00000002 [0x00000000]
    1921:       Incurred fault #6, FLTBOUNDS  %pc = 0xFF232E2C
    1921:         siginfo: SIGSEGV SEGV_MAPERR addr=0x00000000
    1921:       Received signal #11, SIGSEGV [default]
    1921:         siginfo: SIGSEGV SEGV_MAPERR addr=0x00000000
    1921:           *** process killed ***
    18361:      Received signal #18, SIGCLD, in poll() [caught]
    18361:        siginfo: SIGCLD CLD_DUMPED pid=1921 status=0x000B
    18361:  poll(0xFFBFF528, 53, -1)                        Err#4 EINTR
    18361:  lwp_sigtimedwait(0xFFBFF218, 0xFFBFF140, 0x00000010) = 0
    18361:  lwp_sigtimedwait(0xFFBFF130, 0xFFBFF218, 0x00000010) = 0
    18361:  sigprocmask(0, 0x00000000, 0xFFBFEF28)          = 0
    18361:  poll(0xFFBFF528, 53, -1)        (sleeping...)Thank you in advance

  • Oracle Error 'ORA-27102: out of memory' - Shared memory parameters correct.

    Advice please!
    We’ve recently shut down our Oracle test server in order to increase file system capacity.  When we rebooted some of the databases wouldn’t start up.  It started the first 4 instances and then errored out saying “ORA-27102: out of memory“.
    I’m pretty sure it’s nothing to do with the file system because we actually reverted back to the old file system and the databases still wouldn’t start.  I think it’s more likely that something’s gone awry whilst the databases were actually running, and the problem has only manifested itself once we stopped and restarted them.
    I have researched the error and forud this artical and similar:  http://var-adm.blogspot.co.uk/2013/04/adjust-solaris-10-shared-memory-to.html 
    Everything suggests that Oracle is trying to create a larger shared memory segment than is allowed.  The thing is, we’ve never changed our shared memory settings, and one minute it was working, the next it isn’t.  To confirm this I checked the shared memory, which is as follows:
    sswift4# prctl -n project.max-shm-memory $$
    process: 926: bash
    NAME    PRIVILEGE       VALUE    FLAG   ACTION                       RECIPIENT
    project.max-shm-memory
            privileged      7.64GB      -   deny                                 -
            system          16.0EB    max   deny                                 -
    As suggested in the above article, I checked the alert log and found the ‘WARNING: EINVAL’ message which is as follows:
    WARNING: EINVAL creating segment of size 0x000000005e002000
    Converting this to decimal, it’s trying to create something of 1.5 GB, well within the shared memory settings, which suggests that this isn’t the problem.
    We are running Oracle 10g and 11g on Solaris 10 Sparc. The error does not seem to be instance specific, we have 8 instances on this box all with SGA max of 2000m. The server has 32GB of memory available.
    Any advice would be helpful.
    Thanks in advance.
    Debs

    Thanks for your quick responses - we have now resolved the issue.
    The shared memory value was set on the command line but not saved.
    Therefore once we rebooted it lost the configuration. this has been altered by our UNIX admin and all DBs have started without issue.
    Thanks
    Debs

  • Out Of Memory Error : unable to create new native thread

    Hi Experts,
    The details are given below -
    1. Sun Java System Application Server 7
    2. Java 1.4.2_05
    3. Solaris 9 (OS)
    4. RAM 8 GB
    There are 3-4 applications deployed on this Sun Server 7. Some times we got "Out Of Memory" Error while displaying any jsp. Then we restart the Sun Server, the problem is resolved for some time.
    This all is happening on production environment. We can not start the Sun Server again & again on production.
    We have also set the java parameters as -Xms 3584M & -Xmx 3584M i.e 3.5 GB around.
    If we change this parameter means less or more from 3584M, then our site becomes down.
    Please help us out as soon as possible.

    How do you expect anyone to give you a sensible answer to this? What server are you using to execute the JSP's? Tomcat? Is the code in the JSP causing the out of mem or is it a server related issue (unlikely...)?
    Drill down to the core of the problem before posting...

  • Error messages during booting of Solaris 8 zone

    I have installed a Solaris 8 zone from a flash archive of a Solaris 8 server that is currently in production. I get no errors during the installation of the zone and when I boot the zone I get the following messages on the zone console:
    SunOS Release 5.8 Version Generic_Virtual 64-bit
    Copyright 1983-2000 Sun Microsystems, Inc. All rights reserved
    Hostname: madison1xl
    The system is coming up. Please wait.
    NIS domainname is eam.rjf.com
    operation failed, Not owner
    device name given is not a stream device (No such device or address)
    device name given is not a stream device (No such device or address)
    device name given is not a stream device (No such device or address)
    device name given is not a stream device (No such device or address)
    device name given is not a stream device (No such device or address)
    device name given is not a stream device (No such device or address)
    device name given is not a stream device (No such device or address)
    device name given is not a stream device (No such device or address)
    device name given is not a stream device (No such device or address)
    device name given is not a stream device (No such device or address)
    device name given is not a stream device (No such device or address)
    device name given is not a stream device (No such device or address)
    device name given is not a stream device (No such device or address)
    device name given is not a stream device (No such device or address)
    device name given is not a stream device (No such device or address)
    device name given is not a stream device (No such device or address)
    operation failed, Not owner
    operation failed, Not owner
    operation failed, Not owner
    operation failed, Not owner
    operation failed, Not owner
    operation failed, Not owner
    operation failed, Not owner
    operation failed, Not owner
    operation failed, Not owner
    operation failed, Not owner
    Sun BluePrints network security settings applied.
    syslog service starting.
    I am not sure what is causing these errors. The zone appears to be running properly, but I am concerned that something is not starting during the boot process. /var/adm/messages file does not show any errors. Thanks for any help you can give me.

    Thanks. That helped alot. I found out that all the messages came from a start up script named nddconfig. Since the ngz cannot change anything about devices this was causing a lot of errors. I stopped the script from starting and now do not receive any errors.

  • How do I get from Branded Zones to "real" Solaris 10 Zones?

    My head is beginning to spin! I am beginning to think I have been led down a blind alley.
    We have vast numbers of old Solaris 8 systems that we want to consolidate on new hardware in Solaris 10 Zones. We have been bombarded with advice that the way to go is to move it first to Branded Zones, as a stepping stone, and then migrate/upgrade it from there to Solaris 10. So we have started on our journey.
    We now have our first bunch of Solaris 8 containers happily installed and running. And yes, it was relatively easy and painless, and the tool support was good. But now what? What are the steps and tools that will make the second step, from Solaris 8 to Solaris 10?
    Everything out there hints that it is supposed to be easy. But every time I try and spec out the steps for our teams to follow I come unstuck. Does Live Upgrade feature in this anywhere? (And if so, at what point does my zone lose its Brand identity?) Is there some way to use archives? (And if so, how do I only include the bits I need to move?) Is there something magic under the covers of the Zone administration commands? (And if so, what pulls in my Solaris 8 stuff?) Is there at least some tool somewhere that will capture the configuration I need and build me a new one?
    If the answer is that I just define the new Zone by hand and re-install all my applications from scratch, then can somebody explain just what I have achieved so far, and in what way it can be construed as a "stepping stone"?

    Correct. But we have now invested in the time taken to learn about Branded Zones, do the migrations, teach the support teams how to manage them... Plus we are paying the subscription for the Branded Zone software. And we still have pure Solaris 8 applications staring end-of-support in the face. We could have made the same investment to re-install the applications in a real Solaris 10 Zone and make them work "properly". All the publicity tells me I've made my migration to Solaris 10 much easier, but I still can't see how I'm any nearer my end point.Here's where your losing it. There is no real distinction between what you call a normal Solaris 10 zone and a Solaris Branded 8 zone. They could have called it something else and you would still insist that somehow one is normal and one is not. There are a variety of Branded Zones. There's a Solaris 8, Solaris 9, and Linux Branded Zones. They differ from what you call "normal" zones in that you basically are adding an OS to the Base OS and are not just copying files from the Base OS. That's the only difference. In all other respects they are the same. The only differentiators is for a Branded Zone you are adding the files from some other OS, even if it's just an earlier version of Solaris.
    How do you migrate from a Solaris 8 zone to a Solaris 10 zone. The same way that you moved from the Solaris 8 physical box to the Solaris 8 branded zone. The steps should be the same. Implicite_Order pretty much said the same thing.
    You seem to have a hang up on the words "normal zone." Yea I know, you've spent time and money and educated everyone and now your looking for the payoff. As I already stated and has been echoed by I_O, contact the vendor to see if they have Solaris 10 support, if they do, create a zone on a dev box, do some testing, and then create the zone on the real box, do the install, migrate the data. Just as you already did from the Solaris 8 box. If they don't have a Solaris 10 specific version you can still setup a test box, create a "normal" zone, install the app and some test data, make sure it works, and then create the "normal" zone, install the app, migrate the data, and relax.
    alan

Maybe you are looking for