SAPINST 'space' error on Solaris 10 Zone

Hi All,
I am attempting to install a 4.7E X 110 system on Oracle 9 and Solaris 10. I am in the CI phase of the installation and am getting an error in phase "Check/Adapt Filesystem". I'm not sure where - it does not say exactly. But the 3 main areas - /sapmnt , /usr/sap and /oracle all share the same space - "1927165003 blocks" (about 919GB).
This makes no sense for the space check function to be failing. It is possible that the SAPINST is checking the /usr directory at the global level even though it cannot install it there. (We are running a local zone for the server). We have a link set up /usrsap for the /usr/sap folder.
Has anyone ever seen this? Any ideas on how to push this forward? SAPINST does not offer an explicit option for the /usr/sap/ folder.
Points for all helpful answers!!!
Thanks!!

Due to Zones are virtualized Solaris 10 environments during the installation the standard Solaris installation problems may occur. For instance, the /home directory is read-only due to the auto-mount feature is enabled by default. This usually causes problems with the installer when creation of the users is required.
The concept of zones includes that the kernel of the host system is used. Therefore other limitations may occur when adapting the OS kernel is required. Also the read-only directories mounted from the global zone may cause problems.
Please read;724713,828268And important 1 zone note also.sorry to say i realy forgot that note number.
--Sreejesh

Similar Messages

  • SAP AS JAVA installation in Solaris Zone envirn throws OUT of Memory error.

    Hi,
    We are installing SAP NW2004s 7.0 SR3 AS JAVA on Solaris 10 in zone environment. Thsi was prod server getting build where on top of this we will install GRC 5.3.
    We have faced no issues in development build.
    But during the Prod build at databse create step, we are getting the below error
    ORA-27102: out of memory
    SVR4 Error: 22: Invalid argument
    Disconnected
    SAPinst log entries where error messages started:
    04:43:58.128
    Execute step createDatabase of component |NW_Onehost|ind|ind|ind|ind|0|0|NW_Onehost_System|ind|ind|ind|ind|2|0|NW_CreateDBandLoad|ind|ind|ind|ind|10|0|NW_CreateDB|ind|ind|ind|ind|0|0|NW_OraDBCheck|ind|ind|ind|ind|0|0|NW_OraDBMain|ind|ind|ind|ind|0|0|NW_OraDBStd|ind|ind|ind|ind|3|0|NW_OraDbBuild|ind|ind|ind|ind|5|0
    INFO 2011-04-01 04:45:14.590
    Working directory changed to /tmp/sapinst_exe.16718.1301647358.
    INFO 2011-04-01 04:45:14.595
    Working directory changed to /tmp/sapinst_instdir/NW04S/SYSTEM/ORA/CENTRAL/AS.
    INFO 2011-04-01 04:45:14.609
    Working directory changed to /tmp/sapinst_exe.16718.1301647358.
    INFO 2011-04-01 04:45:14.621
    Working directory changed to /tmp/sapinst_instdir/NW04S/SYSTEM/ORA/CENTRAL/AS.
    INFO 2011-04-01 04:45:14.850
    Account oraac5 already exists.
    INFO 2011-04-01 04:45:14.852
    Account dba already exists.
    INFO 2011-04-01 04:45:14.852
    Account oraac5 already exists.
    INFO 2011-04-01 04:45:14.853
    Account dba already exists.
    INFO 2011-04-01 04:45:14.867
    Working directory changed to /tmp/sapinst_exe.16718.1301647358.
    INFO 2011-04-01 04:45:14.899
    Working directory changed to /tmp/sapinst_instdir/NW04S/SYSTEM/ORA/CENTRAL/AS.
    ERROR 2011-04-01 04:45:32.280
    CJS-00084  SQL statement or script failed. DIAGNOSIS: Error message: ORA-27102: out of memory
    SVR4 Error: 22: Invalid argument
    Disconnected
    . SOLUTION: See ora_sql_results.log and the Oracle documentation for details.
    ERROR 2011-04-01 04:45:32.286
    MUT-03025  Caught ESAPinstException in Modulecall: ORA-27102: out of memory
    SVR4 Error: 22: Invalid argument
    Disconnected
    ERROR 2011-04-01 04:45:32.453
    FCO-00011  The step createDatabase with step key |NW_Onehost|ind|ind|ind|ind|0|0|NW_Onehost_System|ind|ind|ind|ind|2|0|NW_CreateDBandLoad|ind|ind|ind|ind|10|0|NW_CreateDB|ind|ind|ind|ind|0|0|NW_OraDBCheck|ind|ind|ind|ind|0|0|NW_OraDBMain|ind|ind|ind|ind|0|0|NW_OraDBStd|ind|ind|ind|ind|3|0|NW_OraDbBuild|ind|ind|ind|ind|5|0|createDatabase was executed with status ERROR ( Last error reported by the step :Caught ESAPinstException in Modulecall: ORA-27102: out of memory
    SVR4 Error: 22: Invalid argument
    Disconnected
    ora_sql_results.log
    04:45:15 SAPINST ORACLE start logging for
    SHUTDOWN IMMEDIATE;
    exit;
    Output of SQL executing program:
    SQL*Plus: Release 10.2.0.4.0 - Production on Fri Apr 1 04:45:15 2011
    Copyright (c) 1982, 2007, Oracle.  All Rights Reserved.
    Connected to an idle instance.
    ORA-01034: ORACLE not available
    ORA-27101: shared memory realm does not exist
    SVR4 Error: 2: No such file or directory
    Disconnected
    SAPINST: End of output of SQL executing program /oracle/AC5/102_64/bin/sqlplus.
    SAPINST found errors.
    SAPINST The current process environment may be found in sapinst_ora_environment.log.
    2011-04-01, 04:45:15 SAPINST ORACLE stop logging
    ================================================================================
    2011-04-01, 04:45:15 SAPINST ORACLE start logging for
    STARTUP NOMOUNT;
    exit;
    Output of SQL executing program:
    SQL*Plus: Release 10.2.0.4.0 - Production on Fri Apr 1 04:45:15 2011
    Copyright (c) 1982, 2007, Oracle.  All Rights Reserved.
    Connected to an idle instance.
    ORA-27102: out of memory
    SVR4 Error: 22: Invalid argument
    Disconnected
    SAPINST: End of output of SQL executing program /oracle/AC5/102_64/bin/sqlplus.
    SAPINST found errors.
    SAPINST The current process environment may be found in sapinst_ora_environment.log.
    2011-04-01, 04:45:32 SAPINST ORACLE stop logging
    Already viewed S-note
    724713 - parameter settings for Solaris 10 - (Parameters are set as per this note)
    743328 - Composite SAP note: ORA-27102 - (no much information in memory problem on zones)
    Please provide your suggestions/resolution.
    Thankyou.

    Hi,
    @ Sunny: Thanks for response, the referred note was already checked and parameters are in sync as per note.
    @Mohit: SAP wouldn't proceed till create database if Oracle software was not installed. thanks for the response.
    @Markus: Thanks I agree with you, but I have doubt in this area. Isn't it like project.max-shm-memory was new parameter we need to set in local zone rather using shmsys:shminfo_shmmax in /etc/system. Do we need to still maintain this parameter in /etc/system in global zone.
    As per SUN doc below parameter was obsolete from Solaris 10.
    The following parameters are obsolete.
    ■ shmsys:shminfo_shmmni
    ■ shmsys:shminfo_shmmax
    As per your suggestion, do we need to set below parameters in that case, please clarify.
    Parameter                           Replaced by Resource Control      Recommended Value
    semsys:seminfo_semmni   project.max-sem-ids                      100
    semsys:seminfo_semmsl   process.max-sem-nsems               256
    shmsys:shminfo_shmmax  project.max-shm-memory             4294967295
    shmsys:shminfo_shmmni   project.max-shm-ids                       100
    Also   findings of /etc/release
    more /etc/release
    Solaris 10 10/08 s10s_u6wos_07b SPARC
    Regards,
    Sitarama.

  • Solaris10 branded zone creation error on Solaris 11.1

    Hello,
    We are having the problem in creating the Solaris10 branded zone on the solaris 11.1 version. We are trying to install this branded zone using the solaris10 VM template. Please can some one through some pointers ?
    The installation steps stops at "Waiting for boot to complete" but never successfully boots even waiting for long times. We can see that Zone status in configured only.
    root@localhost:~# /home/solaris-10u10-sparc.bin -p /solaris10 -i net0 -a xxx.xxx.xxx.xxx -z zone10
    This is an Oracle VM Template for Oracle Solaris Zones.
    Copyright © 2011, Oracle and/or its affiliates. All rights reserved.
    This software and related documentation are provided under a license
    agreement containing restrictions on use and disclosure and are
    protected by intellectual property laws. Except as expressly
    permitted in your license agreement or allowed by law, you may not
    use, copy, reproduce, translate, broadcast, modify, license,
    transmit, distribute, exhibit, perform, publish, or display any
    part, in any form, or by any means. Reverse engineering,
    disassembly, or decompilation of this software, unless required by
    law for interoperability, is prohibited.
    Checking disk-space for extraction
    Ok
    Extracting in /root/bootimage.dnaqaD ...
    100% [===============================>]
    Checking data integrity
    Ok
    Checking platform compatibility
    The host and the image do not have the same Solaris release:
    host Solaris release: 5.11
    image Solaris release: 5.10
    Will create a Solaris 10 branded zone.
    IMAGE: /home/solaris-10u10-sparc.bin
    ZONE: zone10
    ZONEPATH: /solaris10/zone10
    INTERFACE: net0
    VNIC: vnicZBI6783
    MAC ADDR: 2:8:20:5c:92:f5
    IP ADDR: xxx.xxx.xxx.xxx
    NETMASK: 255.0.0.0
    DEFROUTER: # # This file is deprecated. Default routes will be created for any router # addresses specified here, but they will not change when the underlying # network configuration profile (NCP) changes. For NCP-specific static # routes, the '-p' option of the route(1M) command should be used. # # See netcfg(1M) for information about network configuration profiles. #
    TIMEZONE: Asia/Kolkata
    Checking disk-space for installation
    Ok
    Installing in /solaris10/zone10 ...
    100% [===============================>]
    Using a static exclusive-IP
    ./setupSysidcfgStaticIP[50]: file/sysidcfg.zone10: cannot create [No such file or directory]
    cp: cannot access file/sysidcfg.zone10
    chmod: WARNING: can't access file/sysidcfg.zone10
    Attaching zone10
    Booting zone10
    zoneadm: zone 'zone10': must be installed before boot.
    Waiting for boot to complete
    booting...
    booting...
    booting...
    root@localhost:# zoneadm list -cv
    ID NAME STATUS PATH BRAND IP
    0 global running / solaris shared
    4 zone1 running /zone1 solaris excl
    5 zone2 running /zone2 solaris excl
    - zone10 configured /solaris10/zone10 solaris10 excl
    root@localhost:#
    I have installed the package pkg:/system/zones/brand/brand-solaris10 on the system.
    # cat /etc/release
    Oracle Solaris 11.1 SPARC
    Copyright (c) 1983, 2012, Oracle and/or its affiliates. All rights reserved.
    Assembled 19 September 2012

    Sorry, please ignore my previous message about the wrong ip address (I assume you entered a valid ip address).
    The problem you are facing is related to the defaultrouter, and I assume you are running Solaris 11 update 1.
    If so, to workaround the problem for the time being you may edit the /etc/defaultrouter file in the global-zone and enter the ip address of the default router (you can remove this change when the template deployment is completed).
    If you have no idea about the default router ip address, just rename /etc/defaultrouter (e.g. to /etc/defaultrouter.orig) so that it cannot be found by the template. Again, you can undo the change when deployment is complete.
    Thierry

  • Error when creating zone cluster

    Hello,
    I have the following setup: Solaris 11.2 x86, cluster 4.2. I have already configured the cluster and it's up and running. I am trying to create a zone cluster, but getting the following error:
    >>> Result of the Creation for the Zone cluster(ztestcluster) <<<
        The zone cluster is being configured with the following configuration
            /usr/cluster/bin/clzonecluster configure ztestcluster
            create
            set zonepath=/zclusterpool/znode
            set brand=cluster
            set ip-type=shared
            set enable_priv_net=true
            add sysid
            set  root_password=********
            end
            add node
            set physical-host=node2
            set hostname=zclnode2
            add net
            set address=192.168.10.52
            set physical=net1
            end
            end
            add node
            set physical-host=node1
            set hostname=zclnode1
            add net
            set address=192.168.10.51
            set physical=net1
            end
            end
            add net
            set address=192.168.10.55
            end
    java.lang.NullPointerException
            at java.util.regex.Matcher.getTextLength(Matcher.java:1234)
            at java.util.regex.Matcher.reset(Matcher.java:308)
            at java.util.regex.Matcher.<init>(Matcher.java:228)
            at java.util.regex.Pattern.matcher(Pattern.java:1088)
            at com.sun.cluster.zcwizards.zonecluster.ZCWizardResultPanel.consoleInteraction(ZCWizardResultPanel.java:181)
            at com.sun.cluster.dswizards.clisdk.core.IteratorLayout.cliConsoleInteraction(IteratorLayout.java:563)
            at com.sun.cluster.dswizards.clisdk.core.IteratorLayout.displayPanel(IteratorLayout.java:623)
            at com.sun.cluster.dswizards.clisdk.core.IteratorLayout.run(IteratorLayout.java:607)
            at java.lang.Thread.run(Thread.java:745)
                 ERROR: System configuration error
                 As a result of a change to the system configuration, a resource that this
                 wizard will create is now invalid. Review any changes that were made to the
                 system after you started this wizard to determine which changes might have
                 caused this error. Then quit and restart this wizard.
        Press RETURN to close the wizard
    No errors in /var/adm/messages.
    Any ideas?
    Thank you!

    I must be doing some obvious, stupid mistake, cause I still get that "not enough space" error
    root@node1:~# clzonecluster show ztestcluster
    === Zone Clusters ===
    Zone Cluster Name:                              ztestcluster
      zonename:                                        ztestcluster
      zonepath:                                        /zcluster/znode
      autoboot:                                        TRUE
      brand:                                           solaris
      bootargs:                                        <NULL>
      pool:                                            <NULL>
      limitpriv:                                       <NULL>
      scheduling-class:                                <NULL>
      ip-type:                                         shared
      enable_priv_net:                                 TRUE
      resource_security:                               SECURE
      --- Solaris Resources for ztestcluster ---
      Resource Name:                                net
        address:                                       192.168.10.55
        physical:                                      auto
      --- Zone Cluster Nodes for ztestcluster ---
      Node Name:                                    node2
        physical-host:                                 node2
        hostname:                                      zclnode2
        --- Solaris Resources for node2 ---
      Node Name:                                    node1
        physical-host:                                 node1
        hostname:                                      zclnode1
        --- Solaris Resources for node1 ---
    root@node1:~# clzonecluster install ztestcluster
    Waiting for zone install commands to complete on all the nodes of the zone cluster "ztestcluster"...
    clzonecluster:  (C801046) Command execution failed on node node2. Please refer to the console for more information
    clzonecluster:  (C801046) Command execution failed on node node1. Please refer to the console for more information
    But I have enough FS space. I increased the virtual HDD to 25GB on each node. After global cluster installation, I still have 16GB free on each node. During the install I constantly check the free space and it should be enough (only about 500MB is consumed by downloaded packages, which leaves about 15.5GB free).  And every time the installation fails at "apply-sysconfig checkpoint"...

  • Zfs core dumps in Solaris Zones after patchcluster installation

    Hi
    I've installed the recommended patch cluster for Solaris 10 (SPARC) on my T2000 Server.
    Now after the reboot all the Solaris Zones (got about 10 of them) instantly do zfs core dumps when their booted.
    Some Output:
    Dec 15 15:34:43 XXXXX genunix: NOTICE: core_log: zfs[1711] core dumped: /var/core/core_XXXXXX_zfs_0_0_1292423682_1711
    bash-3.00# svcs
    STATE STIME FMRI
    online 13:58:50 svc:/system/svc/restarter:default
    online 13:58:51 svc:/system/filesystem/root:default
    online 13:58:51 svc:/network/loopback:default
    online 13:58:53 svc:/network/pfil:default
    online 13:58:54 svc:/system/installupdates:default
    online 13:58:55 svc:/system/boot-archive:default
    online 13:58:56 svc:/network/physical:default
    online 13:58:57 svc:/system/filesystem/usr:default
    online 13:58:57 svc:/system/identity:node
    online 13:58:58 svc:/system/device/local:default
    online 13:58:58 svc:/system/keymap:default
    online 13:58:58 svc:/milestone/devices:default
    online 13:58:58 svc:/system/filesystem/minimal:default
    online 13:58:58 svc:/system/rmtmpfiles:default
    online 13:58:58 svc:/system/cryptosvc:default
    online 13:58:59 svc:/system/identity:domain
    online 13:58:59 svc:/system/name-service-cache:default
    online 13:58:59 svc:/system/coreadm:default
    online 13:58:59 svc:/network/ipsec/ipsecalgs:default
    online 13:58:59 svc:/application/print/ppd-cache-update:default
    online 13:58:59 svc:/network/ipsec/policy:default
    online 13:59:00 svc:/milestone/network:default
    online 13:59:00 svc:/network/initial:default
    online 13:59:00 svc:/system/manifest-import:default
    online 13:59:00 svc:/network/service:default
    online 13:59:00 svc:/milestone/single-user:default
    online 13:59:00 svc:/network/dns/client:default
    online 13:59:00 svc:/network/routing-setup:default
    online 13:59:00 svc:/milestone/name-services:default
    online 13:59:03 svc:/milestone/sysconfig:default
    online 13:59:03 svc:/system/utmp:default
    online 13:59:03 svc:/system/console-login:default
    offline 13:58:51 svc:/system/sysidtool:net
    offline 13:58:51 svc:/network/rpc/bind:default
    offline 13:58:51 svc:/system/sysidtool:system
    offline 13:58:52 svc:/network/nfs/status:default
    offline 13:58:52 svc:/network/nfs/nlockmgr:default
    offline 13:58:52 svc:/network/nfs/cbd:default
    offline 13:58:52 svc:/network/nfs/mapid:default
    offline 13:58:52 svc:/network/inetd:default
    offline 13:58:52 svc:/network/nfs/client:default
    offline 13:58:52 svc:/system/filesystem/autofs:default
    offline 13:58:53 svc:/system/system-log:default
    offline 13:58:53 svc:/network/smtp:sendmail
    offline 13:58:53 svc:/system/cron:default
    offline 13:58:53 svc:/milestone/multi-user:default
    offline 13:58:54 svc:/application/management/snmpdx:default
    offline 13:58:54 svc:/application/management/dmi:default
    offline 13:58:54 svc:/application/management/seaport:default
    offline 13:58:54 svc:/network/ssh:default
    offline 13:58:54 svc:/milestone/multi-user-server:default
    offline 13:58:54 svc:/application/font/fc-cache:default
    offline 13:58:55 svc:/application/management/sma:default
    offline 13:58:58 svc:/system/sac:default
    offline 13:59:00 svc:/network/shares/group:default
    offline 13:59:00 svc:/system/boot-archive-update:default
    maintenance 15:34:43 svc:/system/filesystem/local:default
    uninitialized 13:58:52 svc:/network/rpc/gss:default
    uninitialized 13:58:54 svc:/application/font/stfsloader:default
    uninitialized 13:58:55 svc:/network/rpc/rstat:default
    uninitialized 13:58:55 svc:/application/print/rfc1179:default
    uninitialized 13:58:55 svc:/application/x11/xfs:default
    uninitialized 13:58:55 svc:/network/finger:default
    uninitialized 13:58:55 svc:/network/ftp:default
    uninitialized 13:58:56 svc:/network/login:rlogin
    uninitialized 13:58:56 svc:/network/nfs/rquota:default
    uninitialized 13:58:56 svc:/network/rpc/rusers:default
    uninitialized 13:58:56 svc:/network/rpc/smserver:default
    uninitialized 13:58:57 svc:/network/security/ktkt_warn:default
    uninitialized 13:58:57 svc:/network/shell:default
    uninitialized 13:58:57 svc:/network/telnet:default
    uninitialized 13:58:58 svc:/network/rpc-100235_1/rpc_ticotsord:default
    uninitialized 13:58:58 svc:/network/rpc/cde-calendar-manager:default
    uninitialized 13:58:59 svc:/network/rpc/cde-ttdbserver:tcp
    bash-3.00# svcs -x
    svc:/system/filesystem/local:default (local file system mounts)
    State: maintenance since Wed Dec 15 15:34:43 2010
    Reason: Start method exited with $SMF_EXIT_ERR_FATAL.
    See: http://sun.com/msg/SMF-8000-KS
    See: /var/svc/log/system-filesystem-local:default.log
    Impact: 24 dependent services are not running. (Use -v for list.)
    svc:/network/rpc/gss:default (Generic Security Service)
    State: uninitialized since Wed Dec 15 13:58:52 2010
    Reason: Restarter svc:/network/inetd:default is not running.
    See: http://sun.com/msg/SMF-8000-5H
    See: gssd(1M)
    Impact: 11 dependent services are not running. (Use -v for list.)
    svc:/network/rpc/rstat:default (kernel statistics server)
    State: uninitialized since Wed Dec 15 13:58:55 2010
    Reason: Restarter svc:/network/inetd:default is not running.
    See: http://sun.com/msg/SMF-8000-5H
    See: rpc.rstatd(1M)
    See: rstatd(1M)
    Impact: 1 dependent service is not running. (Use -v for list.)
    bash-3.00# cat /var/svc/log/system-filesystem-local:default.log
    [ Jul 10 03:41:35 Enabled. ]
    [ Jul 10 03:43:45 Rereading configuration. ]
    [ Jul 10 03:44:17 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 10 03:44:18 Method "start" exited with status 0 ]
    [ Jul 10 12:44:36 Enabled. ]
    [ Jul 10 12:44:45 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 10 12:44:45 Method "start" exited with status 0 ]
    [ Jul 10 12:45:03 Enabled. ]
    [ Jul 10 12:45:09 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 10 12:45:09 Method "start" exited with status 0 ]
    [ Jul 10 13:54:39 Enabled. ]
    [ Jul 10 13:54:46 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 10 13:54:46 Method "start" exited with status 0 ]
    [ Jul 18 12:34:19 Enabled. ]
    [ Jul 18 12:34:27 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 18 12:34:27 Method "start" exited with status 0 ]
    [ Jul 18 17:29:08 Enabled. ]
    [ Jul 18 17:29:15 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 18 17:29:15 Method "start" exited with status 0 ]
    [ Jul 19 11:13:34 Enabled. ]
    [ Jul 19 11:13:42 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 19 11:13:42 Method "start" exited with status 0 ]
    [ Jul 19 11:32:23 Enabled. ]
    [ Jul 19 11:32:30 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 19 11:32:31 Method "start" exited with status 0 ]
    [ Jul 19 11:57:06 Enabled. ]
    [ Jul 19 11:57:13 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 19 11:57:13 Method "start" exited with status 0 ]
    [ Jul 19 12:40:05 Enabled. ]
    [ Jul 19 12:40:12 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 19 12:40:12 Method "start" exited with status 0 ]
    [ Jul 20 15:51:46 Enabled. ]
    [ Jul 20 15:51:54 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 20 15:51:54 Method "start" exited with status 0 ]
    [ Jul 24 11:42:17 Enabled. ]
    [ Jul 24 11:42:25 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 24 11:42:25 Method "start" exited with status 0 ]
    [ Jul 27 12:05:06 Stopping because service disabled. ]
    [ Jul 27 12:05:07 Executing stop method (null) ]
    [ Jul 27 12:13:10 Enabled. ]
    [ Jul 27 12:13:17 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 27 12:13:17 Method "start" exited with status 0 ]
    [ Jul 27 13:10:26 Stopping because service disabled. ]
    [ Jul 27 13:10:26 Executing stop method (null) ]
    [ Jul 27 13:18:19 Enabled. ]
    [ Jul 27 13:18:26 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 27 13:18:26 Method "start" exited with status 0 ]
    [ Jul 31 10:12:31 Enabled. ]
    [ Jul 31 10:12:38 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jul 31 10:12:39 Method "start" exited with status 0 ]
    [ Aug  7 10:08:52 Stopping because service disabled. ]
    [ Aug  7 10:08:52 Executing stop method (null) ]
    [ Aug  7 10:34:50 Enabled. ]
    [ Aug  7 10:34:57 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug  7 10:34:58 Method "start" exited with status 0 ]
    [ Aug  9 09:14:50 Stopping because service disabled. ]
    [ Aug  9 09:14:50 Executing stop method (null) ]
    [ Aug  9 09:55:03 Enabled. ]
    [ Aug  9 09:55:10 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug  9 09:55:11 Method "start" exited with status 0 ]
    [ Aug 16 10:46:43 Enabled. ]
    [ Aug 16 10:46:50 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 16 10:46:50 Method "start" exited with status 0 ]
    [ Aug 16 12:06:56 Enabled. ]
    [ Aug 16 12:07:02 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 16 12:07:02 Method "start" exited with status 0 ]
    [ Aug 16 13:45:10 Enabled. ]
    [ Aug 16 13:45:16 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 16 13:45:16 Method "start" exited with status 0 ]
    [ Aug 16 14:36:59 Enabled. ]
    [ Aug 16 14:37:07 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 16 14:37:07 Method "start" exited with status 0 ]
    [ Aug 16 15:00:40 Enabled. ]
    [ Aug 16 15:00:47 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 16 15:00:47 Method "start" exited with status 0 ]
    [ Aug 22 16:00:13 Enabled. ]
    [ Aug 22 16:00:21 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 22 16:00:21 Method "start" exited with status 0 ]
    [ Aug 23 09:49:04 Enabled. ]
    [ Aug 23 09:49:10 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 23 09:49:11 Method "start" exited with status 0 ]
    [ Aug 23 09:49:23 Stopping because service disabled. ]
    [ Aug 23 09:49:23 Executing stop method (null) ]
    [ Aug 23 10:04:13 Enabled. ]
    [ Aug 23 10:04:20 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 23 10:04:20 Method "start" exited with status 0 ]
    [ Aug 23 10:06:57 Enabled. ]
    [ Aug 23 10:07:04 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 23 10:07:05 Method "start" exited with status 0 ]
    [ Oct 23 08:43:05 Enabled. ]
    [ Oct 23 08:43:14 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Oct 23 08:43:14 Method "start" exited with status 0 ]
    [ Nov 21 14:51:50 Enabled. ]
    [ Nov 21 14:51:58 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Nov 21 14:51:59 Method "start" exited with status 0 ]
    [ Nov 29 15:33:19 Enabled. ]
    [ Nov 29 15:33:26 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Nov 29 15:33:27 Method "start" exited with status 0 ]
    [ Dec 18 13:37:35 Enabled. ]
    [ Dec 18 13:37:42 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Dec 18 13:37:42 Method "start" exited with status 0 ]
    [ Dec 19 08:48:07 Enabled. ]
    [ Dec 19 08:48:14 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Dec 19 08:48:14 Method "start" exited with status 0 ]
    [ Dec 19 08:56:07 Enabled. ]
    [ Dec 19 08:56:14 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Dec 19 08:56:14 Method "start" exited with status 0 ]
    [ Dec 19 08:57:41 Enabled. ]
    [ Dec 19 08:57:48 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Dec 19 08:57:48 Method "start" exited with status 0 ]
    [ Feb  1 10:39:48 Enabled. ]
    [ Feb  1 10:39:55 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Feb  1 10:39:56 Method "start" exited with status 0 ]
    [ Feb  8 11:12:50 Enabled. ]
    [ Feb  8 11:12:57 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Feb  8 11:12:57 Method "start" exited with status 0 ]
    [ Jun  6 10:08:34 Enabled. ]
    [ Jun  6 10:08:44 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Jun  6 10:08:46 Method "start" exited with status 0 ]
    [ Aug  8 09:45:48 Enabled. ]
    [ Aug  8 09:46:35 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug  8 09:46:36 Method "start" exited with status 0 ]
    [ Aug 14 17:53:59 Enabled. ]
    [ Aug 14 17:54:06 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 14 17:54:06 Method "start" exited with status 0 ]
    [ Aug 14 17:55:20 Enabled. ]
    [ Aug 14 17:55:27 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 14 17:55:27 Method "start" exited with status 0 ]
    [ Aug 18 15:19:17 Enabled. ]
    [ Aug 18 15:19:24 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 18 15:19:24 Method "start" exited with status 0 ]
    [ Aug 29 12:56:48 Enabled. ]
    [ Aug 29 12:56:55 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 29 12:56:55 Method "start" exited with status 0 ]
    [ Oct 13 11:59:43 Enabled. ]
    [ Oct 13 11:59:50 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Oct 13 11:59:51 Method "start" exited with status 0 ]
    [ Feb 26 16:47:00 Enabled. ]
    [ Feb 26 16:47:07 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Feb 26 16:47:07 Method "start" exited with status 0 ]
    [ Apr 14 15:54:17 Enabled. ]
    [ Apr 14 15:54:24 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Apr 14 15:54:25 Method "start" exited with status 0 ]
    [ Aug 28 08:49:58 Stopping because service disabled. ]
    [ Aug 28 08:49:58 Executing stop method (null) ]
    [ Aug 28 08:52:54 Enabled. ]
    [ Aug 28 08:53:01 Executing start method ("/lib/svc/method/fs-local") ]
    bootadm: this operation is not supported on sparc
    [ Aug 28 08:53:01 Method "start" exited with status 0 ]
    [ Aug 28 09:08:20 Stopping because service disabled. ]
    [ Aug 28 09:08:20 Executing stop method (null) ]
    [ Sep  1 16:05:36 Enabled. ]
    [ Sep  1 16:06:05 Executing start method ("/lib/svc/method/fs-local") ]
    [ Sep  1 16:06:05 Method "start" exited with status 0 ]
    [ Sep  1 19:29:19 Enabled. ]
    [ Sep  1 19:29:27 Executing start method ("/lib/svc/method/fs-local") ]
    [ Sep  1 19:29:27 Method "start" exited with status 0 ]
    [ Sep  1 23:04:00 Enabled. ]
    [ Sep  1 23:04:08 Executing start method ("/lib/svc/method/fs-local") ]
    [ Sep  1 23:04:08 Method "start" exited with status 0 ]
    [ Sep  2 00:23:41 Stopping because service disabled. ]
    [ Sep  2 00:23:42 Executing stop method (null) ]
    [ Sep  2 00:46:46 Enabled. ]
    [ Sep  2 00:46:53 Executing start method ("/lib/svc/method/fs-local") ]
    [ Sep  2 00:46:54 Method "start" exited with status 0 ]
    [ Sep 26 12:04:34 Enabled. ]
    [ Sep 26 12:04:41 Executing start method ("/lib/svc/method/fs-local") ]
    [ Sep 26 12:04:42 Method "start" exited with status 0 ]
    [ Oct 20 13:05:09 Stopping because service disabled. ]
    [ Oct 20 13:05:10 Executing stop method (null) ]
    [ Oct 20 13:24:01 Enabled. ]
    [ Oct 20 13:24:08 Executing start method ("/lib/svc/method/fs-local") ]
    [ Oct 20 13:24:09 Method "start" exited with status 0 ]
    [ Oct 21 16:30:59 Enabled. ]
    [ Oct 21 16:31:09 Executing start method ("/lib/svc/method/fs-local") ]
    [ Oct 21 16:31:09 Method "start" exited with status 0 ]
    [ Oct 21 16:32:42 Stopping because service disabled. ]
    [ Oct 21 16:32:42 Executing stop method (null) ]
    [ Oct 21 16:42:05 Enabled. ]
    [ Oct 21 16:42:15 Executing start method ("/lib/svc/method/fs-local") ]
    [ Oct 21 16:42:15 Method "start" exited with status 0 ]
    [ Oct 21 17:27:37 Enabled. ]
    [ Oct 21 17:27:47 Executing start method ("/lib/svc/method/fs-local") ]
    [ Oct 21 17:27:47 Method "start" exited with status 0 ]
    [ Oct 22 11:08:34 Enabled. ]
    [ Oct 22 11:08:43 Executing start method ("/lib/svc/method/fs-local") ]
    [ Oct 22 11:08:43 Method "start" exited with status 0 ]
    [ Dec  9 09:17:33 Stopping because service disabled. ]
    [ Dec  9 09:17:33 Executing stop method (null) ]
    [ Dec 10 11:47:14 Enabled. ]
    [ Dec 10 11:47:24 Executing start method ("/lib/svc/method/fs-local") ]
    [ Dec 10 11:47:25 Method "start" exited with status 0 ]
    [ Dec 15 10:52:25 Stopping because service disabled. ]
    [ Dec 15 10:52:25 Executing stop method (null) ]
    [ Dec 15 12:42:16 Enabled. ]
    [ Dec 15 12:42:26 Executing start method ("/lib/svc/method/fs-local") ]
    Abort - core dumped
    WARNING: /usr/sbin/zfs mount -a failed: exit status 134
    [ Dec 15 12:42:28 Method "start" exited with status 95 ]
    [ Dec 15 13:13:19 Enabled. ]
    [ Dec 15 13:13:28 Executing start method ("/lib/svc/method/fs-local") ]
    Abort - core dumped
    WARNING: /usr/sbin/zfs mount -a failed: exit status 134
    [ Dec 15 13:13:30 Method "start" exited with status 95 ]
    [ Dec 15 13:58:51 Enabled. ]
    [ Dec 15 13:59:00 Executing start method ("/lib/svc/method/fs-local") ]
    Abort - core dumped
    WARNING: /usr/sbin/zfs mount -a failed: exit status 134
    [ Dec 15 13:59:03 Method "start" exited with status 95 ]
    [ Dec 15 15:34:41 Leaving maintenance because clear requested. ]
    [ Dec 15 15:34:41 Enabled. ]
    [ Dec 15 15:34:41 Executing start method ("/lib/svc/method/fs-local") ]
    Abort - core dumped
    WARNING: /usr/sbin/zfs mount -a failed: exit status 134
    [ Dec 15 15:34:43 Method "start" exited with status 95 ]
    bash-3.00# /usr/sbin/zfs mount -a
    internal error: Unknown error
    Abort (core dumped)
    Before that I've installed the patch cluster with halted solaris zones which produced following output:
    Application of patches finished : 2010.12.15 12:08:43
    Following patches were applied :
    119254-77 119812-10 125719-31 140159-03 142911-01
    141588-04 119900-11 125731-05 142292-01 142933-02
    142251-02 119906-16 126206-05 141444-09 142909-17
    118666-28 120460-17 126363-08 141500-07 143140-04
    118667-28 121012-03 126365-16 141502-02 143502-01
    118777-16 121308-20 126868-04 141506-09 143506-01
    140860-02 122212-40 127724-02 141514-02 143615-02
    119059-56 122261-03 136998-09 141518-12 143731-01
    119213-23 122675-05 137000-07 141552-03 143733-01
    124628-10 123003-04 137080-05 141558-01 143977-01
    119252-29 123590-12 137147-06 141586-01 144053-04
    119280-23 123893-22 138195-04 141590-02 144106-01
    124188-03 124393-11 138822-07 141874-09 144254-01
    120199-15 124457-02 138826-07 141876-07 144488-04
    119534-19 124630-42 138880-02 142084-04 144492-01
    120272-28 125215-03 139099-04 142397-01 145124-01
    119757-18 125388-03 139620-01 142529-01 145796-01
    119783-15 125555-07
    Following patches were skipped :
    Patches already applied
    120900-04 119063-01 119810-05 123005-07 137093-01
    121133-02 119081-25 119986-03 124444-01 138866-03
    119317-01 119130-33 120061-02 124939-03 137137-09
    121296-01 123611-04 120201-05 124943-01 137871-02
    138215-01 140899-01 120292-02 124997-01 138181-01
    127884-01 122640-05 120329-02 125279-05 138361-01
    118712-23 126897-02 121975-01 125539-06 138373-02
    118918-24 127755-01 120719-02 125891-01 138647-01
    138217-01 125503-02 120830-06 126440-01 141016-01
    119578-30 125547-02 121095-02 126540-02 139555-08
    121453-02 140796-01 121606-04 127127-11 139967-01
    121453-02 120011-14 124171-07 136882-02 140455-01
    121118-16 139520-02 123630-03 137032-01 140563-01
    118833-36 119764-06
    Patches obsoleted by one or more patches already applied
    118731-01 124204-04 122660-10
    Patches not applicable to packages on the system
    121181-03 120410-33 121211-02 125533-15 143317-03
    119115-35 120412-11 122259-03 125541-06 143510-01
    119117-52 120414-27 122470-03 125952-20 143725-01
    119315-19 120543-21 122911-24 137004-08 143727-01
    119548-14 120739-06 122958-06 138387-01 143739-01
    119903-02 120811-09 123938-02 138824-07 144325-01
    120094-30 120849-04 125136-24 138876-01 145006-02
    120185-21 121104-11 125137-24 139986-01 145080-01
    119368-04 121136-02 125332-14 142244-02 145200-01
    120286-03
    Installation of patch set complete. PLEASE REBOOT THE SYSTEM.
    when i tried to reboot the system by init 6 the service iscsi hanged and produced syslog errors about a unkown iocall -> killed the process. System rebooted w/o errors.
    Does anyone have got an idea how to find out which patch or whatever do produce these errors?

    Hi,
    I'm having the same issue of diskmon core dumps after upgrading grid and database from 11.2.0.1 to 11.2.0.2 on Solaris 10 x86-64. Following this thread, I'm trying to deinstall the Oracle 11.2.0.1 grid home.
    I answered the following questions asked by the grid deinstall tool:
    1. ASM configuration was not detected in this Oracle home. Was ASM configured in this Oracle home (y|n): y
    2. Specify the ASM Diagnostic Destination: /db01/app/oracle/diag/asm/+ASM
    3. Specify the diskgroups that are managed by this ASM instance: DATA FRA
    4. De-configuring ASM will drop all the diskgroups at cleanup time. Do you want deconfig tool to drop the diskgroups:
    If I answer no to question 1, the tool exits with error.
    How do I answer question #4? (confused about this question of dropping diskgroups, it's already been upgraded)
    Deinstall of Oracle db Home 11.2.0.1 was successful.
    The +ASM instance has been upgraded and is running fine under 11.2.0.2 except for the excessive diskmon core dumps.
    Thanks,
    Lisa
    Edited by: user12018917 on Apr 19, 2011 10:21 AM

  • SAP Java and Solaris Zones SolMan 4.0

    May require Solaris Zones experience to continue.
    I have three SAP database instances/central instances running in three sparse Solaris zones with no problem.
    I have created a new sparse zone for a new SAP installation (Solution Manager 4.0) and started the installation. SAP requires a 1.4.2 SDK even though Java 1.5 comes with Solaris 10. The 1.4.2 SDK is in /usr/j2se. The installation in the sparse zone errors out because it can't get "write" rights to /usr/j2se/jre/lib/security/local_policy.jar as it is trying to install some security encryption JCE component.
    I have thought about creating a /usr/j2se_zonename file system, copying the contents of /usr/j2se into it and then mounting /usr/j2se_zonename in the zone as a lofs with the name /usr/j2se. However when I do the copy of /usr/j2se I get some recursion errors.
    Any thoughts about how to add a writable /usr/j2se into the sparse zone with the least amount of effort ? Otherwise plan B would be to create a "large" zone with a writable /usr directory.
    Received a great answer, that while it may not be architecturally "pure" it may get the job done.
    You might just download the relevant JDK tarball and unpack that
    somewhere in your zone (anywhere you like), and point SAP at it...
    http://java.sun.com/j2se/1.4.2/download.html
    Get the one called "self extracting file"-- you can unpack that anywhere
    you want.
    Message was edited by: Atis Purins

    Hi Russ,
    no you only have to generate two RFCs to your R/3 and assign them in SMSY to for system monitoring
    Then you need a Solution, assign your R/3 to the Solution, setup the system monitoring.
    Regards,
    uDo

  • BIND 9.3.2 error in in zone

    I have set up the necessary named.conf and forward/reverse/root files and get the following error after I kickoff the named daemon (in /var/adm/messages):
    named[24451]: [ID 873579 daemon.notice] command channel listening on 127.0.0.1#953
    named[24451]: [ID 873579 daemon.notice] couldn't add command channel ::1#953: address not available
    I know this is probably not unique to solaris zones, but thought I would ask in this forum in case there are any issues with zones and bind of this nature.

    hello,
    just configure rndc
    Run rndc-confgen, and put lines in /etc/rndc.conf and /etc/named.conf
    Restart named and this message never appear.
    I don't think it's zones-related
    I ran named in zone and in global zone, and the message was the same
    hth
    gerard

  • How to make space in the heap zone

    i made a 3 minute and 40 second video and when i went to uploud it you youtube it says there is no space in the heap zone .. and i just got this computer .. please help

    Hi
    Error -108 memFullErr  Ran out of memory [not enough room in heap zone]
    Turn off - TimeMachine usually works - re-try.
    Can mean many thing's - My first thought is
    • Free Space on Start-Up hard disk. How much ? (other disks do not count)
    Yours Bengt W

  • Heap space error while creating XML document from Resultset

    I am getting Heap space error while creating XML document from Resultset.
    It was working fine from small result set object but when the size of resultset was more than 25,000, heap space error
    I am already using -Xms32m -Xmx1024m
    Is there a way to directly write to xml file from resultset instead of creating the whole document first and then writing it to file? Code examples please?
    here is my code:
    stmt = conn.prepareStatement(sql);
    result = stmt.executeQuery();
    result.setFetchSize(999);
    Document doc = JDBCUtil.toDocument(result, Application.BANK_ID, interfaceType, Application.VERSION);
    JDBCUtil.write(doc, fileName);
    public static Document toDocument(ResultSet rs, String bankId, String interfaceFileType, String version)
        throws ParserConfigurationException, SQLException {
            DocumentBuilderFactory factory = DocumentBuilderFactory.newInstance();
            DocumentBuilder builder = factory.newDocumentBuilder();
            Document doc = builder.newDocument();
            Element results = doc.createElement("sims");
            results.setAttribute("bank", bankId);
            results.setAttribute("record_type", "HEADER");
            results.setAttribute("file_type", interfaceFileType);
            results.setAttribute("version", version);
            doc.appendChild(results);
            ResultSetMetaData rsmd = rs.getMetaData();
            int colCount = rsmd.getColumnCount();
            String columnName="";
            Object value;
            while (rs.next()) {
                Element row = doc.createElement("rec");
                results.appendChild(row);
                for (int i = 1; i <= colCount; i++) {
                    columnName = rsmd.getColumnLabel(i);
                    value = rs.getObject(i);
                    Element node = doc.createElement(columnName);
                    if(value != null)
                        node.appendChild(doc.createTextNode(value.toString()));
                    else
                        node.appendChild(doc.createTextNode(""));
                    row.appendChild(node);
            return doc;
    public static void write(Document document, String filename) {
            //long start = System.currentTimeMillis();
            // lets write to a file
            OutputFormat format = new OutputFormat(document); // Serialize DOM
            format.setIndent(2);
            format.setLineSeparator(System.getProperty("line.separator"));
            format.setLineWidth(80);
            try {
                FileWriter writer = new FileWriter(filename);
                BufferedWriter buf = new BufferedWriter(writer);
                XMLSerializer FileSerial = new XMLSerializer(writer, format);
                FileSerial.asDOMSerializer(); // As a DOM Serializer
                FileSerial.serialize(document);
                writer.close();
            } catch (IOException ioe) {
                ioe.printStackTrace();
            //long end = System.currentTimeMillis();
            //System.err.println("W3C File write time :" + (end - start) + "  " + filename);
        }

    you can increase your heap size..... try setting this as your environment variable.....
    variable: JAVA_OPTS
    value: -Xms512m -Xmx1024m -XX:PermSize=256m -XX:MaxPermSize=512m

  • Java Heap Space error while reading from Excel

    Hello,
    I have a data flow which reads from excel and updates records in an oracle table.
    When the data flow tries to read the excel I get a java Heap space Error.
    Error preparing to read <SUPP_STG3_FILE1>: <Java heap space>.
    Data flow <DF_SUPP_STG3_UPD> is terminated due to error <58105>.
    I have configured Excel Adapter since the server is on UNIX. BOExcelAdapter shows status as red once the execution fails. Before that BOExcel adapter status is green.
    BOExcelAdapter Trace:
    1/13/12 10:57:38 PM Thread[Thread-6,5,main]      Received an error message from the message broker. [Communication failure. Read timed out]. Will exit adapter immediately.
    1/13/12 10:57:47 PM Thread[main,5,main]      Create broker client for life-cycle management. Host:port = hikalxboes01:4001
    1/13/12 10:57:47 PM Thread[main,5,main]      Create broker client for data exchange with ASP. Host:port = hikalxboes01:4001. Client ID = com.acta.adapter.msexceladapter.MSExcelAdapter.data_18
    1/13/12 10:57:47 PM Thread[Thread-11,5,main]      Create broker client for real-time services with ASP. Host:port = 'hikalxboes01:4000''.
    1/13/12 10:57:47 PM Thread[Thread-11,5,main]      Start instance [BOExcelAdapter] of adapter [com.acta.adapter.msexceladapter.MSExcelAdapter].
    1/13/12 10:57:47 PM Thread[Thread-11,5,main]      Register the client ID [com.acta.adapter.msexceladapter.MSExcelAdapter.data_18] and adapter destination [hikalxboes01:4001] with the adapter manager.
    1/13/12 10:57:47 PM Thread[Thread-11,5,main]      Registration complete.
    1/13/12 10:58:10 PM Thread[Thread-10,5,main]      java.lang.OutOfMemoryError: Java heap space
         at org.apache.poi.hssf.usermodel.HSSFRow.addCell(HSSFRow.java:273)
         at org.apache.poi.hssf.usermodel.HSSFRow.createCellFromRecord(HSSFRow.java:196)
         at org.apache.poi.hssf.usermodel.HSSFSheet.setPropertiesFromSheet(HSSFSheet.java:183)
         at org.apache.poi.hssf.usermodel.HSSFSheet.<init>(HSSFSheet.java:128)
         at org.apache.poi.hssf.usermodel.HSSFWorkbook.<init>(HSSFWorkbook.java:283)
         at org.apache.poi.hssf.usermodel.HSSFWorkbook.<init>(HSSFWorkbook.java:196)
         at org.apache.poi.hssf.usermodel.HSSFWorkbook.<init>(HSSFWorkbook.java:178)
         at com.acta.adapter.msexceladapter.MSExcelAdapterReadTable.begin(MSExcelAdapterReadTable.java:537)
         at com.acta.adapter.sdk.StreamOperationWrapper.beginStream(StreamOperationWrapper.java:88)
         at com.acta.adapter.sdk.AdapterRunnerImpl.processBeginOperation(AdapterRunnerImpl.java:1762)
         at com.acta.adapter.sdk.AdapterServiceDataListener.handleBrokerMessage(AdapterServiceDataListener.java:90)
         at com.acta.brokerclient.BrokerClient.handleMessage(BrokerClient.java:439)
         at com.acta.brokerclient.BrokerClient.access$100(BrokerClient.java:53)
         at com.acta.brokerclient.BrokerClient$MessageHandler.run(BrokerClient.java:1621)
         at com.acta.brokerclient.ThreadPool$PoolThread.run(ThreadPool.java:100)
    BOExcelAdapter Error Log:
    1/13/12 10:58:10 PM      java.lang.OutOfMemoryError: Java heap space
         at org.apache.poi.hssf.usermodel.HSSFRow.addCell(HSSFRow.java:273)
         at org.apache.poi.hssf.usermodel.HSSFRow.createCellFromRecord(HSSFRow.java:196)
         at org.apache.poi.hssf.usermodel.HSSFSheet.setPropertiesFromSheet(HSSFSheet.java:183)
         at org.apache.poi.hssf.usermodel.HSSFSheet.<init>(HSSFSheet.java:128)
         at org.apache.poi.hssf.usermodel.HSSFWorkbook.<init>(HSSFWorkbook.java:283)
         at org.apache.poi.hssf.usermodel.HSSFWorkbook.<init>(HSSFWorkbook.java:196)
         at org.apache.poi.hssf.usermodel.HSSFWorkbook.<init>(HSSFWorkbook.java:178)
         at com.acta.adapter.msexceladapter.MSExcelAdapterReadTable.begin(MSExcelAdapterReadTable.java:537)
         at com.acta.adapter.sdk.StreamOperationWrapper.beginStream(StreamOperationWrapper.java:88)
         at com.acta.adapter.sdk.AdapterRunnerImpl.processBeginOperation(AdapterRunnerImpl.java:1762)
         at com.acta.adapter.sdk.AdapterServiceDataListener.handleBrokerMessage(AdapterServiceDataListener.java:90)
         at com.acta.brokerclient.BrokerClient.handleMessage(BrokerClient.java:439)
         at com.acta.brokerclient.BrokerClient.access$100(BrokerClient.java:53)
         at com.acta.brokerclient.BrokerClient$MessageHandler.run(BrokerClient.java:1621)
         at com.acta.brokerclient.ThreadPool$PoolThread.run(ThreadPool.java:100)

    As Jay mentioned above it might be due to the size of the Excel spreadsheet as you might be hitting the limitation in Apache POI library used for adapter.
    What is the size of the Excel file?
    What is the value of -Xmx parameter - see DS Management Console > Adapter configuration page.
    Thank you,
    Viacheslav.

  • Java Heap Space Error in BI Publisher

    Hi,
    I have an issue with BIP reports.
    My report is working fine in development instance as it has very less data.
    But the same report is throwing "Java Heap Space" error in another instance which is having huge amount of data.
    I tried to modify the oc4j.cmd file by increasing the values -XX:MaxPermSize=128m -Xmx1024m.
    After restarting the oc4j serivices,my OBIEE and BIP URL's becomes dead slow.
    Is there any solution to improve the performance (or) is there any other way to resolve this issue.
    I am running on
    OBIEE 10.1.3.4.2.
    oc4j Server
    Windows box.
    Thanks in advance.
    Pramod.

    Changing os's may help, or it may not. Java requires contiguous memory addresses for the heap, and that varies with the os, os parameters, what has previously been loaded into memory, and how much physical memory is available. The heap can't be assigned to the os's swap space; the os handles paging in and out of memory. All modern os's provide paging to disk.

  • While generating the huge report getting Java Heap Space Error

    Hi All,
    We are using XML Publisher 5.6.2. we have to generate the huge report which contains more than 300 sql statements. Since our schema structure we can not reduce the queries. Now we are facing issue while generating the report. It takes lot of time and gives 'java heap space error'. We are using data template and RTF template for generating excel report.
    Can anybody help me to resolve the issue? or is there any other way to generate the huge reports?
    Thanks,

    Hi,
    I had seen the forum, but we had already adjusted the parameters in the oc4j "-XX:MaxPermSize=128m -Xmx512m" values to 4096. but still we are getting the java heap space error. since we increased the size of xmx value, XML Publisher occupies 61% of memory of the server. B'coz of this our other applications are getting affected.
    and also, since we are using XML Publisher 5.6.2, can you please assist us how to set the scalable property to 'true'.
    Thanks,
    Ramya.
    Edited by: user753355 on Aug 12, 2009 3:35 AM

  • Clean Install G4 Dual 1 GHz-Not enough Disk Space Error? -SLOW USB copies

    HELP puhleeeeze!
    This seems to be an issue on various sites with no one being able to figure out what is the cause and a solution. One person stated his problem went away with a fullwipe and clean-reinstall of everything. But this is what I did already. 
    What is the reason for the "Not Enough Disk Space" error message when both my hard drives have gigs of space available. If I had hair I'd be pulling it out by now!   When trying to copy a file to my 2 internal hard drives that have a lot of space available, but only by the USB port?!!
    I have a Power Mac G4 Mirrored Door (MDD) Dual 1 GHz with a clean install of 10.5.8 from the Original Mac OS X Leopard DVD. Both internal hard drives are 60GB and 100GB. The hard drives were wiped clean with Disc Utilities and reformatted to Mac OS Extended (Journaled). The Power Mac only has the OS and a few music apps installed. Easily over 30 GB of space left on the the main G4 interal hard drive and about 75 GB on the other internal hard drive.
    The error seems to only happen when I connect a USB flash drive to the USB port. I tried to copy a 50MB update to the desktop-"Not Enough Disk Space" error message happened.
    I know that the G4 MDD Power Macs only have USB 1.0. Is this a USB 2.0 issue? Another issue is that it has very slow USB speed (dramtically slower than normal USB 1.0 copying of files. And this is only when it is successful. Is it a USB format issue? I know I couldn't get an external USB 2.0 drive to mount on the Power Mac at all. Disc Utilities could see it but it still would not mount. I have a friend who has an older G4 single processor Mac that doesn't experience these issues with its USB 1.0 ports.
    Now here's the kicker:
    If I put in a CD disc in the DVD/RW drive, it copies with no problems to either hard drive. If I connect to the Power Mac  via the ethernet network, files can copy with no problems. But any USB connection -  I get a the "Not Enough Disk Space" error message. I know this is an old mac... but it's never been problematic until this latest wipe and update to Leopard 10.5.8.
    Is there a solution to this? Any and all help, feedback, comments, thoughts are greatly appreciated in advance!!!

    Hello everyone,
    Ok, I did the ultimate task. I erased the hard drives, AGAIN. Just to start fresh from scratch. Hard drives were erased. Did a new install of Mac OS 10.5. And all was well. I could copy with no problems. I gave the computer to my friend and all was good for a couple of days and then the issue came back again. He couldn't copy anything to his internal hard drives and we received the "Not enough disk space" error message! I was baffled. A complete hard drive erase 2 times, 2 fresh Mac OS 10.5 installs and it works for a few days and then coping to HD we got the same **** problem again! THEN - I remembered that the PRAM battery was dead on this G4 Mirrored-Door Dual 1GHz PowerMac. I had recommended that he get a new battery, because this computer would always have the incorrect date and time after every restart. So with recalling this, I had him reset the PRAM by holding down the CMD+OPT+P+R keys after a restart (start-up Chime). I had him hold the keys for at least 3 start-up-chimes. Soon as the G4 Power Mac started up everything was fine again. He was able to copy files to his hard drives again with the "Not enough disk space" error message.
    So everyone! Try zapping your PRAM (resetting the PRAM) and see if this solves your problem. I told my friend to get a new PRAM battery to replace the old one that is completely dead. I have not heard back from him.  I will ask him again if the problem has come back.
    If that fails: My only other solution would be to consider that it is possible that Mac OS 10.5 has a conflict or issue with the G4 Processors and maybe it is safer to install Mac OS 10.3 or 10.4. Because this problem only has occured after installing Mac OS 10.5, the highest version of the Mac OS the Dual 1GHz Mirrored-Door G4 can be installed.
    Cheers,
    B

  • After clean install of Snow Leopard, Time Machine did not restore all files.  When I try to restore these files I get "Not enough disk space" error.  What do I need to do to get these important files back onto my Mac from my external hard drive??

    After clean install of Snow Leopard, Time Machine did not restore all files.  When I try to restore these files I get "Not enough disk space" error.  What do I need to do to get these important files back onto my Mac from my external hard drive?? 

    Janet.b wrote:
    About 3 years old...pretty dated now I guess.  Am thinking I may just need to have it upgraded by a Mac tech.
    For what the Apple Store's charge you can buy almost buy a new computer.
    Just for giggles I brought my old laptop in for a drive upgrade and asked, they wanted $600 for a drive that only cost $120 at the time and it was a slow 5,400 RPM drive. The new Mac's of a similar make were going for $1000.
    I did the drive replacement myself, got a faster drive of better quality for $300 instead.
    Then another time I asked for a RAM upgrade from 4GB to 8GB, they wanted to charge $400 + $35 for the labor.
    The same RAM is on Crucial.com for $90 and all I need is a 00 micro phillips screwdriver.
    So you see what's going on here.
    I think what you should do, if your out of warranty/AppleCare, is to call up the local PC tech guy who also does Mac's and have them fix you right up.
    They can offer more personalized care which you need. Clone your old drive to the new and everything.
    With Apple they are overworked at the Genius Bar and just place roadblocks so you buy a new machine instead.
    For instance, all new iMac's now have proprietary drive software installed on the hard drives for heat monitoring. So now the only choice a user has is to bring their iMac into a Apple Store for a very expensive drive upgrade, which a person then decides the money is better spent buying a new machine.

  • "Not Enough Storage Space" error - Acrobat 9 Pro

    Hey gang,
    I get a "not enough storage space" error message when trying to install Acrobat 9 Pro.  I've got administrative rights and I'm trying to install from my local drive.  Any ideas?  Thanks!  I've attached a screen shot of the error message.
    My system particulars:
    Windows 7 Pro (64-bit)
    Acrobat 9 Professional
    8 GB RAM
    132 GB free on hard drive
    Thanks again,
    -terry-

    I cleaned the Temp directory and rebooted, making sure to kill the processes in the Startup folder.  It went smooth like a gravy sandwich.  Thanks for the help!
    -terry-

Maybe you are looking for

  • How to accessing current row report column value in Lov Query?

    Hi, which access methods (eg. bind variables, substitutions, ...) for getting the current row report column value can be used in the "Lov Query" property of a report column? As what I know of and what I have read on the forum there are no bind variab

  • CPr-PS integration - cProjects costs not showing up in PS

    Hi, We are on PPM5.0. Accounting integration results in successful replicaition of cPro structure as WBS in PS. However, the costs do not flow. We have used PS06 (ECP) as the costing variant. All configurations seem to be done, and yet, we do not see

  • Returning no output (intentionally)

    Hi, I have a bi publisher report that generates pdf and is printed out automatically. The problem is that I want to be able to stop this from happening for some data. If I return empty xml a blank pdf file is generated which is no good so the only me

  • Batch file naming in Photoshop 6.0

    I have tried every which way I can to rename photos I have scanned in using the Automate Batch process. I have used the batch to perform other tasks, but I am having no luck with renaming using the year, a 3 digit number (to keep the photos in order)

  • When is Pixel Bender hardware accelerated

    Pixel Bender is NOT hardware accelerated ? But when will support GPU? Thanks !