Jboss configuration on Sun Cluster 3.1

Hi.
I am using generic Data Services to manage JBoss instance under Sun Cluster. the command is as follows.
scrgadm -a -j jboss_resource -g cluster_failover_rg -t SUNW.gds \
-y Scalable=false -y Start_timeout=900 \
-y Stop_timeout=420 -x Probe_timeout=300 \
-y Port_list="8080/tcp" \
-y Resource_dependencies=oracle_server_resource \
-x Start_command='/bin/su mform -c "/usr/msm40/scripts/startup/jboss.sh start"' \
-x Stop_command='/bin/su mform -c "/usr/msm40/scripts/startup/jboss.sh stop"' \
-x Child_mon_level=0 -x Failover_enabled=true -x Stop_signal=9
My jboss script will take about 8 to 10 minutes to start completely as it is designed to start about 10 child processes. Hence I set the time out as 15 minutes.
But while starting the resource I found following messages on the console.
Oct 6 12:45:29 MFIN-SOL01 SC[SUNW.gds:5,cluster_failover_rg,jboss_resource,gds_
svc_start]: Failed to connect to host msm and port 8080: Connection refused.
Oct 6 12:45:29 MFIN-SOL01 SC[SUNW.gds:5,cluster_failover_rg,jboss_resource,gds_
svc_start]: Failed to connect to the host <msm> and port <8080>.
Oct 6 12:45:31 MFIN-SOL01 SC[SUNW.gds:5,cluster_failover_rg,jboss_resource,gds_
svc_start]: Failed to connect to host msm and port 8080: Connection refused.
Oct 6 12:45:31 MFIN-SOL01 SC[SUNW.gds:5,cluster_failover_rg,jboss_resource,gds_
svc_start]: Failed to connect to the host <msm> and port <8080>.
Oct 6 12:45:33 MFIN-SOL01 SC[SUNW.gds:5,cluster_failover_rg,jboss_resource,gds_
svc_start]: Failed to connect to host msm and port 8080: Connection refused.
Oct 6 12:45:33 MFIN-SOL01 SC[SUNW.gds:5,cluster_failover_rg,jboss_resource,gds_
svc_start]: Failed to connect to the host <msm> and port <8080>.
Oct 6 12:45:35 MFIN-SOL01 SC[SUNW.gds:5,cluster_failover_rg,jboss_resource,gds_
svc_start]: Failed to connect to host msm and port 8080: Connection refused.
Oct 6 12:45:35 MFIN-SOL01 SC[SUNW.gds:5,cluster_failover_rg,jboss_resource,gds_
svc_start]: Failed to connect to the host <msm> and port <8080>.
here msm is the logical hostname i have selected and port 8080 is used by jboss instance.
after throwing these error messages the cluster software failes over to the other node and changes the status to offline after several attempts.
I tried starting the instance manually and it worked fine.
Please let me know if I am missing something.
Thanks in advance for the help.

Found the solution. Added delay at the end of start script. This may be because jboss takes some time to bind the ports and the hostname.

Similar Messages

  • LDOM SUN Cluster Interconnect failure

    I am making a test SUN-Cluster on Solaris 10 in LDOM 1.3.
    in my environment, i have T5120, i have setup two guest OS with some configurations, setup sun cluster software, when executed, scinstall, it failed.
    node 2 come up, but node 1 throws following messgaes:
    Boot device: /virtual-devices@100/channel-devices@200/disk@0:a File and args:
    SunOS Release 5.10 Version Generic_139555-08 64-bit
    Copyright 1983-2009 Sun Microsystems, Inc. All rights reserved.
    Use is subject to license terms.
    Hostname: test1
    Configuring devices.
    Loading smf(5) service descriptions: 37/37
    /usr/cluster/bin/scdidadm: Could not load DID instance list.
    /usr/cluster/bin/scdidadm: Cannot open /etc/cluster/ccr/did_instances.
    Booting as part of a cluster
    NOTICE: CMM: Node test2 (nodeid = 1) with votecount = 1 added.
    NOTICE: CMM: Node test1 (nodeid = 2) with votecount = 0 added.
    NOTICE: clcomm: Adapter vnet2 constructed
    NOTICE: clcomm: Adapter vnet1 constructed
    NOTICE: CMM: Node test1: attempting to join cluster.
    NOTICE: CMM: Cluster doesn't have operational quorum yet; waiting for quorum.
    NOTICE: clcomm: Path test1:vnet1 - test2:vnet1 errors during initiation
    NOTICE: clcomm: Path test1:vnet2 - test2:vnet2 errors during initiation
    WARNING: Path test1:vnet1 - test2:vnet1 initiation encountered errors, errno = 62. Remote node may be down or unreachable through this path.
    WARNING: Path test1:vnet2 - test2:vnet2 initiation encountered errors, errno = 62. Remote node may be down or unreachable through this path.
    clcomm: Path test1:vnet2 - test2:vnet2 errors during initiation
    CREATED VIRTUAL SWITCH AND VNETS ON PRIMARY DOMAIN LIKE:<>
    532 ldm add-vsw mode=sc cluster-vsw0 primary
    533 ldm add-vsw mode=sc cluster-vsw1 primary
    535 ldm add-vnet vnet2 cluster-vsw0 test1
    536 ldm add-vnet vnet3 cluster-vsw1 test1
    540 ldm add-vnet vnet2 cluster-vsw0 test2
    541 ldm add-vnet vnet3 cluster-vsw1 test2
    Primary DOmain<>
    bash-3.00# dladm show-dev
    vsw0 link: up speed: 1000 Mbps duplex: full
    vsw1 link: up speed: 0 Mbps duplex: unknown
    vsw2 link: up speed: 0 Mbps duplex: unknown
    e1000g0 link: up speed: 1000 Mbps duplex: full
    e1000g1 link: down speed: 0 Mbps duplex: half
    e1000g2 link: down speed: 0 Mbps duplex: half
    e1000g3 link: up speed: 1000 Mbps duplex: full
    bash-3.00# dladm show-link
    vsw0 type: non-vlan mtu: 1500 device: vsw0
    vsw1 type: non-vlan mtu: 1500 device: vsw1
    vsw2 type: non-vlan mtu: 1500 device: vsw2
    e1000g0 type: non-vlan mtu: 1500 device: e1000g0
    e1000g1 type: non-vlan mtu: 1500 device: e1000g1
    e1000g2 type: non-vlan mtu: 1500 device: e1000g2
    e1000g3 type: non-vlan mtu: 1500 device: e1000g3
    bash-3.00#
    NOde1<>
    -bash-3.00# dladm show-link
    vnet0 type: non-vlan mtu: 1500 device: vnet0
    vnet1 type: non-vlan mtu: 1500 device: vnet1
    vnet2 type: non-vlan mtu: 1500 device: vnet2
    -bash-3.00# dladm show-dev
    vnet0 link: unknown speed: 0 Mbps duplex: unknown
    vnet1 link: unknown speed: 0 Mbps duplex: unknown
    vnet2 link: unknown speed: 0 Mbps duplex: unknown
    -bash-3.00#
    NODE2<>
    -bash-3.00# dladm show-link
    vnet0 type: non-vlan mtu: 1500 device: vnet0
    vnet1 type: non-vlan mtu: 1500 device: vnet1
    vnet2 type: non-vlan mtu: 1500 device: vnet2
    -bash-3.00#
    -bash-3.00#
    -bash-3.00# dladm show-dev
    vnet0 link: unknown speed: 0 Mbps duplex: unknown
    vnet1 link: unknown speed: 0 Mbps duplex: unknown
    vnet2 link: unknown speed: 0 Mbps duplex: unknown
    -bash-3.00#
    and this configuration i give while setting up scinstall
    Cluster Transport Adapters and Cables <<<You must identify the two cluster transport adapters which attach
    this node to the private cluster interconnect.
    For node "test1",
    What is the name of the first cluster transport adapter [vnet1]?
    Will this be a dedicated cluster transport adapter (yes/no) [yes]?
    All transport adapters support the "dlpi" transport type. Ethernet
    and Infiniband adapters are supported only with the "dlpi" transport;
    however, other adapter types may support other types of transport.
    For node "test1",
    Is "vnet1" an Ethernet adapter (yes/no) [yes]?
    Is "vnet1" an Infiniband adapter (yes/no) [yes]? no
    For node "test1",
    What is the name of the second cluster transport adapter [vnet3]? vnet2
    Will this be a dedicated cluster transport adapter (yes/no) [yes]?
    For node "test1",
    Name of the switch to which "vnet2" is connected [switch2]?
    For node "test1",
    Use the default port name for the "vnet2" connection (yes/no) [yes]?
    For node "test2",
    What is the name of the first cluster transport adapter [vnet1]?
    Will this be a dedicated cluster transport adapter (yes/no) [yes]?
    For node "test2",
    Name of the switch to which "vnet1" is connected [switch1]?
    For node "test2",
    Use the default port name for the "vnet1" connection (yes/no) [yes]?
    For node "test2",
    What is the name of the second cluster transport adapter [vnet2]?
    Will this be a dedicated cluster transport adapter (yes/no) [yes]?
    For node "test2",
    Name of the switch to which "vnet2" is connected [switch2]?
    For node "test2",
    Use the default port name for the "vnet2" connection (yes/no) [yes]?
    i have setup the configurations like.
    ldm list -l nodename
    NODE1<>
    NETWORK
    NAME SERVICE ID DEVICE MAC MODE PVID VID MTU LINKPROP
    vnet1 primary-vsw0@primary 0 network@0 00:14:4f:f9:61:63 1 1500
    vnet2 cluster-vsw0@primary 1 network@1 00:14:4f:f8:87:27 1 1500
    vnet3 cluster-vsw1@primary 2 network@2 00:14:4f:f8:f0:db 1 1500
    ldm list -l nodename
    NODE2<>
    NETWORK
    NAME SERVICE ID DEVICE MAC MODE PVID VID MTU LINKPROP
    vnet1 primary-vsw0@primary 0 network@0 00:14:4f:f9:a1:68 1 1500
    vnet2 cluster-vsw0@primary 1 network@1 00:14:4f:f9:3e:3d 1 1500
    vnet3 cluster-vsw1@primary 2 network@2 00:14:4f:fb:03:83 1 1500
    ldm list-services
    VSW
    NAME LDOM MAC NET-DEV ID DEVICE LINKPROP DEFAULT-VLAN-ID PVID VID MTU MODE INTER-VNET-LINK
    primary-vsw0 primary 00:14:4f:f9:25:5e e1000g0 0 switch@0 1 1 1500 on
    cluster-vsw0 primary 00:14:4f:fb:db:cb 1 switch@1 1 1 1500 sc on
    cluster-vsw1 primary 00:14:4f:fa:c1:58 2 switch@2 1 1 1500 sc on
    ldm list-bindings primary
    VSW
    NAME MAC NET-DEV ID DEVICE LINKPROP DEFAULT-VLAN-ID PVID VID MTU MODE INTER-VNET-LINK
    primary-vsw0 00:14:4f:f9:25:5e e1000g0 0 switch@0 1 1 1500 on
    PEER MAC PVID VID MTU LINKPROP INTERVNETLINK
    vnet1@gitserver 00:14:4f:f8:c0:5f 1 1500
    vnet1@racc2 00:14:4f:f8:2e:37 1 1500
    vnet1@test1 00:14:4f:f9:61:63 1 1500
    vnet1@test2 00:14:4f:f9:a1:68 1 1500
    NAME MAC NET-DEV ID DEVICE LINKPROP DEFAULT-VLAN-ID PVID VID MTU MODE INTER-VNET-LINK
    cluster-vsw0 00:14:4f:fb:db:cb 1 switch@1 1 1 1500 sc on
    PEER MAC PVID VID MTU LINKPROP INTERVNETLINK
    vnet2@test1 00:14:4f:f8:87:27 1 1500
    vnet2@test2 00:14:4f:f9:3e:3d 1 1500
    NAME MAC NET-DEV ID DEVICE LINKPROP DEFAULT-VLAN-ID PVID VID MTU MODE INTER-VNET-LINK
    cluster-vsw1 00:14:4f:fa:c1:58 2 switch@2 1 1 1500 sc on
    PEER MAC PVID VID MTU LINKPROP INTERVNETLINK
    vnet3@test1 00:14:4f:f8:f0:db 1 1500
    vnet3@test2 00:14:4f:fb:03:83 1 1500
    Any Idea Team, i beleive the cluster interconnect adapters were not successfull.
    I need any guidance/any clue, how to correct the private interconnect for clustering in two guest LDOMS.

    You dont have to stick to default IP's or subnet . You can change to whatever IP's you need. Whatever subnet mask you need. Even change the private names.
    You can do all this during install or even after install.
    Read the cluster install doc at docs.sun.com

  • Sun Cluster

    Hi Experts,
    I am new to Sun Cluster and trying to build a Sun Cluster on Solaris 10 x86 boxes. Is there a document that will walk me through step by step on configuring the Sun Cluster.
    Thanks in Advance.
    Sunil.

    Sunil,
    You can download the software and the documentation for free from the Sun web-site. Try reading the manuals. They are very detailed and contains many examples how to install the software.
    Cheers
    Andreas

  • Sun Cluster and DS6 within zones

    It is possible to configure the sun cluster agent for cluster 3.1/ 3.2 in a DS6.0 zone?
    It's supported?
    thank you.

    Sun Cluster 3.1 u4 does support Zones for failover but as a blackbox ie failover of the complete zone.
    I am not sure this will interact well with DS 6 Cluster Agent.
    DS 6.0 has it's own Agent (SUNW.ds6ldap delivered in the SUNWldap-directory-ha package).
    This package is available with JavaES 5 as part of the Cluster Agents bundle.
    Now with regards to the documentation of the DS cluster agent and its use, unfortunately the only documentation is what I posted on my blog early March : <http://blogs.sun.com/Ludo/entry/directory_server_6_ha_with>
    Regards,
    Ludovic.

  • Configuration of LUN's to Sun Cluster

    Hi,
    I have a 2 node Sun Cluster (V3.2) running on 2xE2900, Solaris 10...
    Basically, there are 3 installed Databases running on the development environment and I need to cluster all 3 in the Global Zone do some failovers and then engage Sun PS to come on site and configure the production cluster environment...
    Usually I have already configured metasets or ZFS and then the DBA installs the DB while everything is nice and neat, my question however is what is the best way to cluster the LUN's when they already have data which I cannot (or would prefer not) to loose.
    I believe the creation of LUN's in a metaset will destroy the data and obviously zfs pools will also destroy any data... hopefully this is a simple question from an SC novice :)
    Thanks...

    Thanks Tim, that answer the question... one more though :)
    I was advised to install a single node cluster then add the 2nd node to the config later. Ive done this but when I try to do the add it seems I have a problem with the cluster interconnects and receive the messages:-
    Adding cable to the cluster configuration ... failed
    scrconf: Failed to add cluster transport cable - does not exist
    scinstall: Failed to update cluster configuration ("-m endpoint=<server>:ce3,endpoint=switch1")
    The heartbeats are ce3 and ce7 which I know are working ok, ive tried everything from the 1st node but when I enter:-
    # scstat -W
    Nothing is shown, although when I do a scconf -p I can see the node transport adapters ok... so how do I let the 2nd node access to the cluster interconnects, ive tried clsetup and adding the interconnects via option4 and I remember configuring them during installation...
    Again any input would be greatly received...
    Thanks...
    Steve..

  • Configuring a HA service on Sun Cluster 2.2

    I am a Product Manager working with customers using Oracle software on Sun Cluster 2.2. My question is, how can I configure a service to bind to logical/virtual address, so as to make it available at the same address after failover? Are there cluster specific steps that I need to take in order to achieve this?

    In the OPS environment there is no use of the HA-Oracle agent, the same instance of the database is running on the different node. The failover is from the client side, because all nodes access the same shared disk space database. The tnsnames.ora files is modifed so that if a transaction fails the client will try the other nodes. The OPS environment also uses the Oracle's unix distributed lock mangage(UDLM), so there is some overhead issues.
    let me know if this the info you needed,
    Heath
    [email protected]

  • Sun Cluster 2.2 configuration

    Hi all,
    is it possible to configure Sun Cluster 2.2 to use "sqlplus" instead of "svrmgrl" to start an oracle db (9i)?
    thanks

    Oracle 9i is not supported on SC 2.2.
    Just replacing svrmgrl with 'sqlplus' will not be sufficient to make HA-Oracle work with Oracle 9i. You will most likely encounter problems in fault monitor, it will just fail.
    Oracle 8i fault monitor is likely to be incompatible with 9i server.
    If you really want to use Oracle 9i on SC2.2, you will be better of by writing a custom agent or moving to SC 3.x.

  • Sun cluster: scalable cluster configuration

    Hi all,
    is it possible to configure the scalable - load sharing cluster for SAP with oracle 10g database using only Sun Cluster 3.1 software with out oracle RAC.
    If yes how sun cluster manage the two database instance on different nodes?
    thanks and regards
    suj.

    Now, you would need the real SAP expert.
    If you want to have more than 1 DB instance (to reduce failover times) you need a parallel DB, in the case of Oracle this is Oracle RAC. If you have a single instance Oracle DB, this would be restarted in case of a failover.
    I think the only thing in SAP that is scalable is the app server. (I remember that there is also support for a replicated enqueue server, but that is a different issue) But if you had SAP (is this the CI only?) on one node and the DB on the other, you very probably would only have to restart the service that would failover in case of a node failure and not both.
    I recommend getting a more detailed answer by an SAP expert.
    Hartmut

  • Sun Cluster 3.0 MQ Series 5.2 configuration

    Hi All,
    we have to review MQ Series installation/configuration on 2 solaris 8 Clustered with Sun Cluster 3.0 machines. The present configuration has a global filesystem /var/mqm with one queue manager.
    According to Sun Cluster 3.1 dataservice for websphere MQ(5.3 ndr) there are 2 ways of filesystem layout
    FFS: with local qmgrs (data and log) at each cluster node
    GFS: with global filesystem qmgrs (data and log).
    Are there any special consideration about shmem and ipc directories in <qmgr>/data?
    Does this scenario also apply to 3.0 /5.2 ?
    Does FFS configuration allow persistant messages failover at takeover?
    Are there any dataservices/docs available for MQ on 3.0?
    Thanks in advance.

    To deploy multiple qmgrs requires /var/mqm to be mounted as a GFS. The reason for this is to overcome IPC key clashes. The recommended file system layout is as follows -> represents a symlink, assuming two qmgrs - qmgr1 & qmgr2
    Using FFS (recommended - /local/mqm etc.. are mounted as FFS with /etc/vfstab)
    /var/mqm -> /global/mqm
    /global/mqm/qmgrs/qmgr1 -> /local/mqm/qmgr/qmgr1
    /global/mqm/qmgrs/qmgr2 -> /local/mqm/qmgr/qmgr2
    /global/mqm/log/qmgr1 -> /local/mqm/log/qmgr1
    /global/mqm/log/qmgr2 -> /local/mqm/log/qmgr2
    Using GFS (mainly early SC3.0 as HAStoragePlus wasn't available until later on)
    All mounted as GFS with /etc/vfstab
    /var/mqm -> /global/mqm
    /global/mqm/qmgrs/qmgr1
    /global/mqm/qmgrs/qmgr2
    /global/mqm/log/qmgr1
    /global/mqm/log/qmgr2
    Finally, FFS (Failover File System) is recommend because, at present, whenever GFS is used for the qmgr & log files, MQ Series is unable to determine that the qmgr may have been started on another node. e.g. Assuming GFS, and MQ Series is started on Node A, it is possible (but don't do it) to start MQ Series on Node B.
    The Sun Cluster Agent provides some protection against this. Instead it's recommened to deploy FFS as above.
    The agent for WebSphere MQ for SC 3.1 is available and supported on SC3.0 update 3 as well as SC3.1. There is also a patch available for the WebSphere MQ Agent which deals with IPC cleanup, for single or multiple qmgrs.
    Docs available can be found at
    http://docs.sun.com/db/prod/7192#hic - Just select Sun Cluster Data Service for WebSphere MQ
    Finally, the above scenario also applies to SC3.0/5.2 as well as SC3.1/5.3 and either GFS/FFS allow for persistant messages to be available after a failover.
    Regards
    Neil

  • Configure iws on Sun cluster???

    I have installed sun cluster 3.1.On top of it I need to install iws(sunone web server).Does anyone have document pertaining to it?
    I tried docs.sun.com , the document there sound greek or latin to me
    Cheers

    Just to get you started:
    3) create the failover RG to hold the shared address.
    #scrgadm -a -g sa-rg (unique arbitrary RG name) -h prod-node1,prod-node2 (comma seperated list of nades that can host this RG, in the order you want it to fail over)
    again - #scrgadm -a -g sa-rg -h prod-node1,prod-node2
    4) add the network resource to the failover RG.
    # scrgadm -a -S (telling the cluster this is going to be a scalable resource, if it were failover you would use -L) -g sa-rg (the group we created in step #3) -l web-server (-l is for the hostname of the logical host. This name (web-server) needs to be specified in the /etc/hosts file on each node of the cluster. Even if a nodfe is not going to host the RG, it has to know about the LH (logical hosts) hostname!)
    again - #scrgadm -a -S -g sa-rg -l web-server
    5) create the scalable resource group that will run on all nodes.
    #scrgadm -a -g web-rg -y Maximum_primaries=2 -y Desired_primaries=2 -y RG_dependencies=sa-rg
    -y is an extension property. Most resources use standard properties, other "can" use extension properties, still others "must" have extension properties defined. Maximum_primaries says how many nodes you want instance to run on at the most. Desired_primaries is how many instances you want to run at the same time. For an eight node cluster, running other DS's you might say, Maximum_primaries=8 Desired_primaries=6 Which means an instance could run on any node in the cluste, but you want to try to make sure there are nodes available for your other resource so you only want to run 6 instance at any given time, leaving the other two nodes to run your other DS's.
    You could say Max=8 Desired=8 it's a matter of choice.
    6) create a storage resource to be used by the app. This tells the app where to go to find the software it needs to run or process.
    -a=add,-g=in the group,-j=resource name, needs to be unique and is arbitrary, -t resource type installed in pkg format earlier, and registered, -x= resource type extension property (a -y extension property could be used for a RG property or a RT property) -x is only for a RT property. /global/web is defined in the /etc/vfstab file with the mount options field specifying global,logging (at least global, maybe logging) (note you do not specify the DG, just mounts from storage supplied by the DG, because multiple RG's may use storage from the same DG)
    #scrgadm -a -g web-rg -j web-stor -t SUNW.HAStoragePlus (HAStoragePlus provides support only for global devices and file systems) -x Affinityon=false -x FileSystemMountPoints=/global/web
    7) create the app resource in the scalable RG.
    -a=add, -j new resource -g (in the group) web-rg (created in step #5) using the type -t SUNW.apache (defined in step #2, remember the pkg installed was SUNWscapc, SUNW.apache is a made up name we are using to use apache for possibly multiple resource groups. Each -j (resource name must be unique, and only used once) but each -t (resource type, allthough having a unique name from other RT's can be used over and over again in different resources of different RG's.) Bin_dir (self explanitory, where to go to get the binaries) Network_Resouces_Used=web-server (created in step #5, again is the hostname in the /etc/vfstab for the logical host, the name the clients are going to use to get to the resource.) Resource_Dependencies=web-stor (created in step #6) saying that apache-res depends on web-stor, so if web-stor is not online, don't bother trying to start apache because the binaries won't be there. They are supplied by the storage being online and /global/web being mounted up.
    #scrgadm -a -j apache-res -g web-rg -t SUNW.apache -x Bin_dir=/usr/apache/bin -y Scalable=True -y Network_Resources_Used=web-server -y Resource_dependencies=web-stor
    8) switch the failover group to activate it.
    #scswitch -z -g sa-rg
    9) switch the scalable RG to activate it.
    #scswitch -z -g web-rg
    10) make sure everything got started.
    #scstat -g
    11) connect to the newly, cluster started service.

  • Encountered ora-29701 during Sun Cluster for Oracle RAC 9.2.0.7 startup (UR

    Hi all,
    Need some help from all out there
    In our Sun Cluster 3.1 Data Service for Oracle RAC 9.2.0.7 (Solaris 9) configuration, my team had encountered
    ora-29701 *Unable to connect to Cluster Manager*
    during the startup of the Oracle RAC database instances on the Oracle RAC Server resources.
    We tried the attached workaround by Oracle. This workaround works well for the 1^st time but it doesn’t work anymore when the server is rebooted.
    Kindly help me to check whether anyone encounter the same problem as the above and able to resolve. Thanks.
    Bug No. 4262155
    Filed 25-MAR-2005 Updated 11-APR-2005
    Product Oracle Server - Enterprise Edition Product Version 9.2.0.6.0
    Platform Linux x86
    Platform Version 2.4.21-9.0.1
    Database Version 9.2.0.6.0
    Affects Platforms Port-Specific
    Severity Severe Loss of Service
    Status Not a Bug. To Filer
    Base Bug N/A
    Fixed in Product Version No Data
    Problem statement:
    ORA-29701 DURING DATABASE CREATION AFTER APPLYING 9.2.0.6 PATCHSET
    *** 03/25/05 07:32 am ***
    TAR:
    PROBLEM:
    Customer applied 9.2.0.6 patchset over 9.2.0.4 patchset.
    While creating the database, customer receives following error:
         ORA-29701: unable to connect to Cluster Manager
    However, if customer goes from 9.2.0.4 -> 9.2.0.5 -> 9.2.0.6, the problem does not occur.
    DIAGNOSTIC ANALYSIS:
    It seems that the problem is with libskgxn9.so shared library.
    For 9.2.0.4 -> 9.2.0.5 -> 9.2.0.6, the install log shows the following:
    installActions2005-03-22_03-44-42PM.log:,
    [libskgxn9.so->%ORACLE_HOME%/lib/libskgxn9.so 7933 plats=1=>[46]langs=1=> en,fr,ar,bn,pt_BR,bg,fr_CA,ca,hr,cs,da,nl,ar_EG,en_GB,et,fi,de,el,iw,hu,is,in, it,ja,ko,es,lv,lt,ms,es_MX,no,pl,pt,ro,ru,zh_CN,sk,sl,es_ES,sv,th,zh_TW, tr,uk,vi]]
    installActions2005-03-22_04-13-03PM.log:, [libcmdll.so ->%ORACLE_HOME%/lib/libskgxn9.so 64274 plats=1=>[46] langs=-554696704=>[en]]
    For 9.2.0.4 -> 9.2.0.6, install log shows:
    installActions2005-03-22_04-13-03PM.log:, [libcmdll.so ->%ORACLE_HOME%/lib/libskgxn9.so 64274 plats=1=>[46] langs=-554696704=>[en]] does not exist.
    This means that while patching from 9.2.0.4 -> 9.2.0.5, Installer copies the libcmdll.so library into libskgxn9.so, while patching from 9.2.0.4 -> 9.2.0.6 does not.
    ORACM is located in /app/oracle/ORACM which is different than ORACLE_HOME in customer's environment.
    WORKAROUND:
    Customer is using the following workaround:
    cd $ORACLE_HOME/rdbms/lib make -f ins_rdbms.mk rac_on ioracle ipc_udp
    RELATED BUGS:
    Bug 4169291

    Check if following MOS note helps.
    Series of ORA-7445 Errors After Applying 9.2.0.7.0 Patchset to 9.2.0.6.0 Database (Doc ID 373375.1)

  • SAP 7.0 on SUN Cluster 3.2 (Solaris 10 / SPARC)

    Dear All;
    i'm installing a two nodes cluster (SUN Cluster 3.2 / Solaris 10 / SPARC), for a HA SAP 7.0 / Oracle 10g DataBase
    SAP and Oracle softwares were successfully installed and i could successfully cluster the Oracle DB and it is tested and working fine.
    for the SAP i did the following configurations
    # clresource create -g sap-ci-res-grp -t SUNW.sap_ci_v2 -p SAPSID=PRD -p Ci_instance_id=01 -p Ci_services_string=SCS -p Ci_startup_script=startsap_01 -p Ci_shutdown_script=stopsap_01 -p resource_dependencies=sap-hastp-rs,ora-db-res sap-ci-scs-res
    # clresource create -g sap-ci-res-grp -t SUNW.sap_ci_v2 -p SAPSID=PRD -p Ci_instance_id=00 -p Ci_services_string=ASCS -p Ci_startup_script=startsap_00 -p Ci_shutdown_script=stopsap_00 -p resource_dependencies=sap-hastp-rs,or-db-res sap-ci-Ascs-res
    and when trying to bring the sap-ci-res-grp online # clresourcegroup online -M sap-ci-res-grp
    it executes the startsap scripts successfully as following
    Sun Microsystems Inc.     SunOS 5.10     Generic     January 2005
    stty: : No such device or address
    stty: : No such device or address
    Starting SAP-Collector Daemon
    11:04:57 04.06.2008 LOG: Effective User Id is root
    Starting SAP-Collector Daemon
    11:04:57 04.06.2008 LOG: Effective User Id is root
    * This is Saposcol Version COLL 20.94 700 - V3.72 64Bit
    * Usage: saposcol -l: Start OS Collector
    * saposcol -k: Stop OS Collector
    * saposcol -d: OS Collector Dialog Mode
    * saposcol -s: OS Collector Status
    * Starting collector (create new process)
    * This is Saposcol Version COLL 20.94 700 - V3.72 64Bit
    * Usage: saposcol -l: Start OS Collector
    * saposcol -k: Stop OS Collector
    * saposcol -d: OS Collector Dialog Mode
    * saposcol -s: OS Collector Status
    * Starting collector (create new process)
    saposcol on host eccprd01 started
    Starting SAP Instance ASCS00
    Startup-Log is written to /export/home/prdadm/startsap_ASCS00.log
    saposcol on host eccprd01 started
    Running /usr/sap/PRD/SYS/exe/run/startj2eedb
    Trying to start PRD database ...
    Log file: /export/home/prdadm/startdb.log
    Instance Service on host eccprd01 started
    Jun 4 11:05:01 eccprd01 SAPPRD_00[26054]: Unable to open trace file sapstartsrv.log. (Error 11 Resource temporarily unavailable) [ntservsserver.cpp 1863]
    /usr/sap/PRD/SYS/exe/run/startj2eedb completed successfully
    Starting SAP Instance SCS01
    Startup-Log is written to /export/home/prdadm/startsap_SCS01.log
    Instance Service on host eccprd01 started
    Jun 4 11:05:02 eccprd01 SAPPRD_01[26111]: Unable to open trace file sapstartsrv.log. (Error 11 Resource temporarily unavailable) [ntservsserver.cpp 1863]
    Instance on host eccprd01 started
    Instance on host eccprd01 started
    and the it repeats the following warnings on the /var/adm/messages till it fails to the other node
    Jun 4 12:26:22 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:25 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:25 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:28 eccprd01 last message repeated 1 time
    Jun 4 12:26:28 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:31 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:31 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:34 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:34 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:37 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:37 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:40 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:40 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:43 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:43 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:46 eccprd01 last message repeated 1 time
    Jun 4 12:26:46 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:49 eccprd01 last message repeated 1 time
    Jun 4 12:26:49 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:52 eccprd01 last message repeated 1 time
    Jun 4 12:26:52 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:55 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:55 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:26:58 eccprd01 last message repeated 1 time
    Jun 4 12:26:58 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:01 eccprd01 last message repeated 1 time
    Jun 4 12:27:01 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:04 eccprd01 last message repeated 1 time
    Jun 4 12:27:04 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:07 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:07 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:10 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:10 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:13 eccprd01 last message repeated 1 time
    Jun 4 12:27:13 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:16 eccprd01 last message repeated 1 time
    Jun 4 12:27:16 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:19 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:19 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:22 eccprd01 last message repeated 1 time
    Jun 4 12:27:22 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:25 eccprd01 last message repeated 1 time
    Jun 4 12:27:25 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:28 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:28 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:31 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:31 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:34 eccprd01 last message repeated 1 time
    Jun 4 12:27:34 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:37 eccprd01 last message repeated 1 time
    Jun 4 12:27:37 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:40 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:40 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:43 eccprd01 last message repeated 1 time
    Jun 4 12:27:43 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-scs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dispatcher to come up.
    Jun 4 12:27:46 eccprd01 last message repeated 1 time
    Jun 4 12:27:46 eccprd01 SC[SUNW.sap_ci_v2,sap-ci-res-grp,sap-ci-Ascs-res,sap_ci_svc_start]: [ID 183934 daemon.notice] Waiting for SAP Central Instance main dis
    can anyone one help me if there is any error on configurations or what is the cause of this problem.....thanks in advance
    ARSSES

    Hi all.
    I am having a similar issue with a Sun Cluster 3.2 and SAP 7.0
    Scenrio:
    Central Instance (not incluster) : Started on one node
    Dialog Instance (not in cluster): Started on the other node
    When I create the resource for SUNW.sap_as like
    clrs create --g sap-rg -t SUNW.sap_as .....etc etc
    in the /var/adm/messages I got lots of WAITING FOR DISPACHER TO COME UP....
    Then after timeout it gives up.
    Any clue? What does is try to connect or waiting for? I hve notest that it's something before the startup script....
    TIA

  • Creating Logical hostname in sun cluster

    Can someone tell me, what exactly logical hostname in sun cluster mean?
    For registering logical hostname resource in failoover group, what exactly i need to specify
    for example, i have two nodes in sun cluster , How to create or configure a logical hostanme and it should point to which IP Address ( Whether it should point to IP addresses of nodes in sun cluster). Can i get clarification on this?

    Thanks Thorsten for ur continue help...
    The output of clrs status abc_lg
    === Cluster Resources ===
    Resource Name Node Name State Status Message
    abc_lg node1 Offline Offline
    node2 Offline Offline
    The status is offline...
    the output of clresourcegroup status
    === Cluster Resource Groups ===
    Group Name Node Name Suspended Status
    abc_rg node1 No Unmanaged
    node2 No Unmanaged
    You say that the resource should de enabled after creating the resource.. I am using GDS and i am just following the steps he provided to acheive high availabilty (in developers guide...)
    I have 1) Logical hostname resorce.
    2) Application resource in my failover resource group
    When i bring online the failover resource group , what should my failover resource group status and the status of resource in my resource group

  • File System Sharing using Sun Cluster 3.1

    Hi,
    I need help on how to setup and configure the system to share a remote file system that is created on a SAN disk (SAN LUN ) between two Sun Solaris 10 servers.
    The files in the remote system should be read/writabe from both the solaris servers concurrently.
    As a security policy NFS mount is not allowed. Some one suggested it can be done by using Sun Cluster 3.1 agents on both servers. Any details on how I can do this using Sun Cluster 3.1 is really appreciated.
    thanks
    Suresh

    You could do this by installing Sun Cluster on both systems and then creating a global file system on the shared LUN. However, if there was significant write activity on both nodes, then the performance will not necessarily what you need.
    What is wrong with the security of NFS? If it is set up properly I don't think this should be a problem.
    The other option would be to use shared QFS, but without Sun Cluster.
    Regards,
    Tim
    ---

  • Sun Cluster with Netapps - iSCSI quorum and network port

    I am proposing Sun cluster with Netapps 3020C.
    May I know
    1) OS is Solaris 9. The SUN OSP says that we need to obtain an iSCSI license from Netapps. Is this the iSCSI initiator software for Solaris 9 to talk to the NAS quorum? Or do I need to purchased a 3rd party iSCSI initiator ?
    2) We provide 2 network ports for the Netapps private NAS LAN. Is it a must to cater another dedicated network port for the iSCSI communication with the quorum?
    3) If we need purchase a 3rd party iSCSI initiator, where can we get this? I have checked Qlogic and Cisco, they are both not suitable for my solution.
    Appreciate your help

    Hi,
    1) OS is Solaris 9. The SUN OSP says that we need to
    obtain an iSCSI license from Netapps. Is this the
    iSCSI initiator software for Solaris 9 to talk to the
    NAS quorum? Or do I need to purchased a 3rd party
    iSCSI initiator ?Have a look at http://docs.sun.com/app/docs/doc/817-7957/6mn8834r2?a=view
    I read the "Requirements When Configuring NAS Devices as Quorum Devices"
    section as this is the license for the iSCSI inititator software.
    So you need to enable iSCSI on the netapps box and need to install a package from netapps (NTAPclnas) on the cluster nodes.
    2) We provide 2 network ports for the Netapps
    private NAS LAN. Is it a must to cater another
    dedicated network port for the iSCSI communication
    with the quorum?Have a look at http://docs.sun.com/app/docs/doc/819-0580/6n30eahcc?a=view#ch4_quorum-9
    I don't read such a requirement there.
    3) If we need purchase a 3rd party iSCSI initiator,
    where can we get this? I have checked Qlogic and
    Cisco, they are both not suitable for my solution.
    Appreciate your helpI don't thibk you need such a 3rd party iSCSI initiator, unless this is stated in the above docs.
    Greets
    Thorsten

Maybe you are looking for