Availability of Sun Cluster 3.2

Has anyone some news about the release date and the new features of Sun Cluster 3.2. It was once announced by end of 2006.
Fritz

Hi Tim
I more or less expected this answer from you ;-)
We are planing to use the Sun Cluster to switch Zones / Containers as GDS between nodes. We have currently some installations with Sun Cluster 3.1, but we are now in the process of evaluating a framework to deploy Solaris Zones in a large scale. This would also include containers in a clustered environment. There Sun Cluster 3.2 seems to have some interesting new features.
Unfortunately I had not the resources to paritipizate in the beta program.
Regards
Fritz

Similar Messages

  • Beta Refresh Release Now Available!  Sun Cluster 3.2 Beta Program

    The Sun Cluster 3.2 Release team is pleased to announce a Beta Refresh release. This release is based on our latest and greatest build of Sun Cluster 3.2, build 70, which is close to the final Revenue Release build of the product.
    To apply for the Sun Cluster 3.2 Beta program, please visit:
    https://feedbackprograms.sun.com/callout/default.html?callid=%7B11B4E37C-D608-433B-AF69-07F6CD714AA1%7D
    or contact Eric Redmond <[email protected]>.
    New Features in Sun Cluster 3.2
    Ease of use
    * New Sun Cluster Object Oriented Command Set
    * Oracle RAC 10g improved integration and administration
    * Agent configuration wizards
    * Resources monitoring suspend
    * Flexible private interconnect IP address scheme
    Availability
    * Extended flexibility for fencing protocol
    * Disk path failure handling
    * Quorum Server
    * Cluster support for SMF services
    Flexibility
    * Solaris Container expanded support
    * HA ZFS
    * HDS TrueCopy campus cluster
    * Veritas Flashsnap Fast Mirror Resynchronization 4.1 and 5.0 option support
    * Multi-terabyte disk and EFI label support
    * Veritas Volume Replicator 5.0 support
    * Veritas Volume Manager 4.1 support on x86 platform
    * Veritas Storage Foundation 5.0 File System and Volume Manager
    OAMP
    * Live upgrade
    * Dual partition software swap (aka quantum leap)
    * Optional GUI installation
    * SNMP event MIB
    * Command logging
    * Workload system resource monitoring
    Note: Veritas 5.0 features are not supported with SC 3.2 Beta.
    Sun Cluster 3.2 beta supports the following Data Services
    * Apache (shipped with the Solaris OS)
    * DNS
    * NFS V3
    * Java Enterprise System 2005Q4: Application Server, Web Server, Message Queue, HADB

    Without speculating on the release date of Sun Cluster 3.x or even its feature list, I would like to understand what risk Sun would take when Sun Cluster would support ZFS as a failover filesystem? Once ZFS is part of Solaris 10, I am sure customers will want to use it in clustered environments.
    BTW: this means that even Veritas will have to do something about ZFS!!!
    If VCS is a much better option, it would be interesting to understand what features are missing from Sun Cluster to make it really competitive.
    Thanks
    Hartmut

  • Sun Cluster dataservice for Webspher MQ

    Is Sun Cluster dataservice for Websphere MQ available for
    Sun Cluster 3.0?
    Should a Sun Cluster 3.0 be upgraded to 3.1 in order to install the Websphere MQ dataservice?
    Thanks in advance.

    The Sun Cluster dataservice for WebSphere MQ is available for SC3.0 update 3 onwards.
    However the agent is only distributed on the SC3.1 Agents CD, please contact Sun for the CD or Agent.
    Updating to SC3.1 is not required. MQ 5.2/5.3 are supported, although only MQ 5.2 on Solaris 8, with MQ 5.3 on either Solaris 8 or 9.
    Regards
    Neil

  • IDS 5.0 and Sun-Cluster

    I want to make IDS 5.0 highly available using SunCluster. I have few questions about it.
    1. Can I install IDS on local disk and make it highly availbale. Sun-Cluster doc says it should be on shared disk.
    2. I have an already installed IDS I want to make it highly available using Sun-Cluster, what steps I should follow to achive this.

    I suppose that the answer is that it is not a simple task and it depends on the kind of cluster you want to deploy.
    I suggest that you carefully read the documentation of Sun Cluster and specifically the Directory Server specific parts.
    The way to do it is different with Sun Cluster 2 and Sun Cluster 3.0....
    Or you can request help from Sun Professional Services...
    Regards,
    Ludovic.

  • Sun Cluster 3.2  without share storage. (Sun StorageTek Availability Suite)

    Hi all.
    I have two node sun cluster.
    I am configured and installed AVS on this nodes. (AVS Remote mirror replication)
    AVS working fine. But I don't understand how integrate it in cluster.
    What did I do:
    Created remote mirror with AVS.
    v210-node1# sndradm -P
    /dev/rdsk/c1t1d0s1      ->      v210-node0:/dev/rdsk/c1t1d0s1
    autosync: on, max q writes: 4096, max q fbas: 16384, async threads: 2, mode: sync, group: AVS_TEST_GRP, state: replicating
    v210-node1# 
    v210-node0# sndradm -P
    /dev/rdsk/c1t1d0s1      <-      v210-node1:/dev/rdsk/c1t1d0s1
    autosync: on, max q writes: 4096, max q fbas: 16384, async threads: 2, mode: sync, group: AVS_TEST_GRP, state: replicating
    v210-node0#   Created resource group in Sun Cluster:
    v210-node0# clrg status avs_test_rg
    === Cluster Resource Groups ===
    Group Name       Node Name       Suspended      Status
    avs_test_rg      v210-node0      No             Offline
                     v210-node1      No             Online
    v210-node0#  Created SUNW.HAStoragePlus resource with AVS device:
    v210-node0# cat /etc/vfstab  | grep avs
    /dev/global/dsk/d11s1 /dev/global/rdsk/d11s1 /zones/avs_test ufs 2 no logging
    v210-node0#
    v210-node0# clrs show avs_test_hastorageplus_rs
    === Resources ===
    Resource:                                       avs_test_hastorageplus_rs
      Type:                                            SUNW.HAStoragePlus:6
      Type_version:                                    6
      Group:                                           avs_test_rg
      R_description:
      Resource_project_name:                           default
      Enabled{v210-node0}:                             True
      Enabled{v210-node1}:                             True
      Monitored{v210-node0}:                           True
      Monitored{v210-node1}:                           True
    v210-node0# In default all work fine.
    But if i need switch RG on second node - I have problem.
    v210-node0# clrs status avs_test_hastorageplus_rs
    === Cluster Resources ===
    Resource Name               Node Name    State     Status Message
    avs_test_hastorageplus_rs   v210-node0   Offline   Offline
                                v210-node1   Online    Online
    v210-node0# 
    v210-node0# clrg switch -n v210-node0 avs_test_rg
    clrg:  (C748634) Resource group avs_test_rg failed to start on chosen node and might fail over to other node(s)
    v210-node0#  If I change state in logging - all work.
    v210-node0# sndradm -C local -l
    Put Remote Mirror into logging mode? (Y/N) [N]: Y
    v210-node0# clrg switch -n v210-node0 avs_test_rg
    v210-node0# clrs status avs_test_hastorageplus_rs
    === Cluster Resources ===
    Resource Name               Node Name    State     Status Message
    avs_test_hastorageplus_rs   v210-node0   Online    Online
                                v210-node1   Offline   Offline
    v210-node0#  How can I do this without creating SC Agent for it?
    Anatoly S. Zimin

    Normally you use AVS to replicate data from one Solaris Cluster to another. Can you just clarify whether you are replicating to another cluster or trying to do it between a single cluster's nodes? If it is the latter, then this is not something that Sun officially support (IIRC) - rather it is something that has been developed in the open source community. As such it will not be documented in the Sun main SC documentation set. Furthermore, support and or questions for it should be directed to the author of the module.
    Regards,
    Tim
    ---

  • RAC 10g on Sun Cluster 3.1 U3 and Interconnect

    Hello,
    I have the following Interconnects on my Sun Cluster:
    ce5: flags=1008843<UP,BROADCAST,RUNNING,MULTICAST,PRIVATE,IPv4> mtu 1500 index 6
         inet 1.1.1.1 netmask ffffff80 broadcast 1.1.1.127
         ether 0:3:ba:95:fa:23
    ce5: flags=2008841<UP,RUNNING,MULTICAST,PRIVATE,IPv6> mtu 1500 index 6
         ether 0:3:ba:95:fa:23
         inet6 fe80::203:baff:fe95:fa23/10
    ce0: flags=1008843<UP,BROADCAST,RUNNING,MULTICAST,PRIVATE,IPv4> mtu 1500 index 7
         inet 1.1.0.129 netmask ffffff80 broadcast 1.1.0.255
         ether 0:3:ba:95:f9:97
    ce0: flags=2008841<UP,RUNNING,MULTICAST,PRIVATE,IPv6> mtu 1500 index 7
         ether 0:3:ba:95:f9:97
         inet6 fe80::203:baff:fe95:f997/10
    clprivnet0: flags=1009843<UP,BROADCAST,RUNNING,MULTICAST,MULTI_BCAST,PRIVATE,IPv4> mtu 1500 index 8
         inet 1.1.193.1 netmask ffffff00 broadcast 1.1.193.255
         ether 0:0:0:0:0:1
    In the Installation of RAC the routine will ask me which Interface I will use for RAC Interconnect and I do not know if it does not matter which Interface I choose, because I nevertheless in any case I have an SPOF.
    Can anybody help??
    Thank you very much

    Sorry for the late reply, but the interface to pick is the clprivnet0. This load-balances over the available private interconnects under the covers and so does not represent a single point of failure.
    Tim
    ---

  • Sun cluster patch for solaris 10 x86

    I have Solaris 10 6/06 installed on x4100 box with 2 node clustering using Sun Cluster 3.1 8/05. I just want to know is there any latest patches available for the OS to prevent cluster related bugs. what are they? My kernel patch is 118855-19.
    any inputs needed. let me know.

    Well, I would run S10 updatemanager and get the latest patches that way.
    Tim
    ---

  • Install of HA MaxDB in a Sun Cluster 3.2 environment

    Hi. I am installing SAP Content Server using MaxDB in a Sun Cluster 3.2 environment. According to the SUN doc, I must install the X_server as a scalable resource/service.  For a scalable resource, I have to make the disk available on both nodes of the cluster. I need to identify what are all the executables for the x_server so that I can make those mount points a NFS share.  From looking at the content of previous non-clustered installs, I am thinking I need to make the directories /sapdb/data and /sapdb/program NFS to isolate the x_server components.  Does anyone have experience with this setup ? I've placed a log with OSS but wanted to see if anyone else in the forum has done this before. Thanks for any input.

    Hello Dan,
    Please review the SAP Note No. 803452.
    X_server tool is the the part of the indepentent database software installation <independent_program_path>. Using NFS you could  set up the <independent_program_path> and <independent_data_path>  directories as share directories.
    Thank you and best regards, Natalia Khlopina

  • Information about Sun Cluster 3.1 5Q4 and Storage Foundation 4.1

    Hi,
    I have 2 Sunfire V440 with Solaris 9 last release 9/05 with last cluster patchs.. , Qlogic HBA fibre card on a seven disks shared on a Emc Clariion cx500. I have installed and configured Sun cluster 3.1 and Veritas Storage Foundation 4.1 MP1. My problems is when i run format wcommand on each node, I see the disks in a different order and veritas SF (4.1) is also picking up the disks in a different order.
    1. Storage Foundation 4.1 is compatible with Sun cluster 3.1 2005Q4?????
    2. Do you have a how to or other procedure for Storage foundation 4.1 with Sun Cluster 3.1.
    I'm very confuse with veritas Storage foundation
    Thanks!
    J-F Aubin

    This combination does not work today, but it will be available later.
    Since Sun and Veritas are two separate companies, it takes more
    time than expected to synchronize releases. Products supported by
    Sun for Sun Cluster installation undergo extensive testing, which also
    takes time.
    -- richard

  • Didadm: unable to determine hostname.  error on Sun cluster 4.0 - Solaris11

    Trying to install Sun Cluster 4.0 on Sun Solaris 11 (x86-64).
    iscs sharedi Quorum Disk are available in /dev/rdsk/ .. ran
    devfsadm
    cldevice populate
    But don't see DID devices getting populated in /dev/did.
    Also when scdidadm -L is issued getting the following error. Has any seen the same error ??
    - didadm: unable to determine hostname.
    Found in cluster 3.2 there was a Bug 6380956: didadm should exit with error message if it cannot determine the hostname
    The sun cluster command didadm, didadm -l in particular, requires the hostname to function correctly. It uses the standard C library function gethostname to achieve this.
    Early in the cluster boot, prior to the service svc:/system/identity:node coming online, gethostname() returns an empty string. This breaks didadm.
    Can anyone point me in the right direction to get past this issue with shared quorum disk DID.

    Let's step back a bit. First, what hardware are you installing on? Is it a supported platform or is it some guest VM? (That might contribute to the problems).
    Next, after you installed Solaris 11, did the system boot cleanly and all the services come up? (svcs -x). If it did boot cleanly, what did 'uname -n' return? Do commands like 'getent hosts <your_hostname>' work? If there are problems here, Solaris Cluster won't be able to get round them.
    If the Solaris install was clean, what were the results of the above host name commands after OSC was installed? Do the hostnames still resolve? If not, you need to look at why that is happening first.
    Regards,
    Tim
    ---

  • Wrong hostname setting after Sun Cluster failover

    Hi Gurus,
    our PI system has been setup to fail over in a sun cluster with a virtual hostname s280m (primary host s280 secondary host s281)
    The basis team set up the system profiles to use the virtual hostname, and I did all the steps in SAP Note 1052984 "Process Integration 7.1 High Availability" (my PI is 7.11)
    Now I believe to have substituted "s280m" in every spot where previously "s280" existed, but when I start the system on the DR box (s281), the java stack throws erros when starting. Both SCS01 and DVEBMGS00 work directories contain a file called dev_sldregs with the following error:
    Mon Apr 04 11:55:22 2011 Parsing XML document.
    Mon Apr 04 11:55:22 2011 Supplier Name: BCControlInstance
    Mon Apr 04 11:55:22 2011 Supplier Version: 1.0
    Mon Apr 04 11:55:22 2011 Supplier Vendor:
    Mon Apr 04 11:55:22 2011 CIM Model Version: 1.5.29
    Mon Apr 04 11:55:22 2011 Using destination file '/usr/sap/XP1/SYS/global/slddest.cfg'.
    Mon Apr 04 11:55:22 2011 Use binary key file '/usr/sap/XP1/SYS/global/slddest.cfg.key' for data decryption
    Mon Apr 04 11:55:22 2011 Use encryted destination file '/usr/sap/XP1/SYS/global/slddest.cfg' as data source
    Mon Apr 04 11:55:22 2011 HTTP trace: false
    Mon Apr 04 11:55:22 2011 Data trace: false
    Mon Apr 04 11:55:22 2011 Using destination file '/usr/sap/XP1/SYS/global/slddest.cfg'.
    Mon Apr 04 11:55:22 2011 Use binary key file '/usr/sap/XP1/SYS/global/slddest.cfg.key' for data decryption
    Mon Apr 04 11:55:22 2011 Use encryted destination file '/usr/sap/XP1/SYS/global/slddest.cfg' as data source
    Mon Apr 04 11:55:22 2011 ******************************
    Mon Apr 04 11:55:22 2011 *** Start SLD Registration ***
    Mon Apr 04 11:55:22 2011 ******************************
    Mon Apr 04 11:55:22 2011 HTTP open timeout     = 420 sec
    Mon Apr 04 11:55:22 2011 HTTP send timeout     = 420 sec
    Mon Apr 04 11:55:22 2011 HTTP response timeout = 420 sec
    Mon Apr 04 11:55:22 2011 Used URL: http://s280:50000/sld/ds
    Mon Apr 04 11:55:22 2011 HTTP open status: false - NI RC=0
    Mon Apr 04 11:55:22 2011 Failed to open HTTP connection!
    Mon Apr 04 11:55:22 2011 ****************************
    Mon Apr 04 11:55:22 2011 *** End SLD Registration ***
    Mon Apr 04 11:55:22 2011 ****************************
    notice it is using the wrong hostname (s280 instead of s280m). Where did I forget to change the hostname? Any ideas?
    thanks in advance,
    Peter

    Please note that the PI system is transparent about the Failover system used.
    When you configure the parameters against the mentioned note, this means that in case one of the nodes is down, the load will be sent to another system under the same Web Dispatcher/Load Balancer.
    When using the Solaris failover solution, it covers the whole environment, including the web dispatcher, database and all nodes.
    Therefore, please check the configuration as per the page below, which talks specifically about the Solaris failover solution for SAP usage:
    http://wikis.sun.com/display/SunCluster/InstallingandConfiguringSunClusterHAfor+SAP

  • Node shutdown in sun cluster

    I have a two node cluster configured for High availability..
    My resourcegroup is online on node1..
    so the resources, logical hostname resource and my application resource are online on node1.
    when the node1 is shutdown, the resource group is failovered to node2 and is online..But when the node1 is brought back, the logical hostname is plumbed up on node1 also.. So both nodes have the logical hostname plumbed up..(from ifconfig -a output)
    which is causing the problem.
    My question is "Does sun cluster check the status of resources in the resource group on the node where my resource group is offline" . If it does, what additional configuration is required.

    This is a pretty old post and you probably have the answer by now )or have abandoned all hope), but it seems to me that what you want is to reset the resource/resource group dependencies for node1.
    If node1 is coming online under logicalhostname without all the resources coming up, you just don't have the resource dependency set up. You can do this in the SunPlex Manager GUI pretty easily. That should make it so the node doesn't get added to the logicalhostname resource group until X dependencies are met (what X stands for is entirely up to you; I didn't see the resource you want to come up first listed.)

  • Multiple Oracle databases in Sun cluster 3.2 (without RAC setup)

    There are 2 Sun SPARC (Sun Fire T2000) servers with Solaris 10 (05/09) OS and Sun Cluster 3.2 software. We need two different Oracle databases & instances (Oracle 10g R2 without RAC) for an application. The first database is for Production and need to be configured in the first node & Shared storage disk and need high availability. This database should run from the second node if the first node fails. The second database is for Quality/Test and it is prefered to be running in the Second node for better load distribution. This DB doesnt require any failover.
    The shared storage is Sun SE 3510 FC and multiple LUNs can be created for different databases..
    Is it possible to configure two different resource groups (one for Quality and other for Production) and make the first node primary for Production RG, and the second node primary for Quality RG and thus distributing the load on 2 servers ? If possible, what special configuration required in Sun OS and Cluster side ?
    Appreciate if you can give some configuration procedures/documents for this multi-master cluster setup..

    You can configure two resource groups, such as:
    # clrg create -n node-a,node-b prod-rg
    # clrg create -n node-b qa-rgand you configure the required resources (disk groups / file systems, logical host, oracle listener, oracle server) as described within
    [http://docs.sun.com/app/docs/doc/819-2980?l=en&a=expand|http://docs.sun.com/app/docs/doc/819-2980?l=en&a=expand]
    Note that this is not really called a "multi-master" configuration - that has a specific meaning for a resource group (see [http://docs.sun.com/app/docs/doc/820-4682/babefcja?l=en&a=view|http://docs.sun.com/app/docs/doc/820-4682/babefcja?l=en&a=view] ) for details.
    With Solaris Cluster all nodes part of a cluster are considered active and can host resource groups. You can have any number of resource groups running, where a subset runs one one node, another subset on other nodes. The nodelist property of the resource group defines where it can run, the first node in the list is the preferred primary.
    You can even define resource group dependencies or affinities between the resource groups. Like you could define a negative affinity between qa-rg and prod-rg, such as if prod-rg needs to failover to node-b (since e.g. node-a died), it would offline qa-rg. Details for that kind of possibilities are described at [http://docs.sun.com/app/docs/doc/820-4682/ch14_resources_admin-35?l=en&a=view|http://docs.sun.com/app/docs/doc/820-4682/ch14_resources_admin-35?l=en&a=view].
    Regards
    Thorsten

  • TimesTen database in Sun Cluster environment

    Hi,
    Currently we have our application together with the TimesTen database installed at the customer on two different nodes (running on Sun Solaris 10). The second node acts as a backup to provide failover functionality, although right now only manual failover is supported.
    We are now looking into a hot-standby / high availability solution using Sun Cluster software. As understood from the documentation, applications can be 'plugged-in' to the Sun Cluster using Agents to monitor the application. Sun Cluster Agents should be already available for certain applications such as:
    # MySQL
    # Oracle 9i, 10g (HA and RAC)
    # Oracle 9iAS Application Server
    # PostgreSQL
    (See http://www.sun.com/software/solaris/cluster/faq.jsp#q_19)
    Our question is whether Sun Cluster Agents are already (freely) available for TimesTen? If so, where to find them. If not, should we write a specific Agent separately for TimesTen or handle database problems from the application.
    Does someone have any experience using TimesTen in a Sun Cluster environment?
    Thanks in advance!

    Yes, we use 2-way replication, but we don't use cache connect. The replication is created like this on both servers:
    create replication MYDB.REPSCHEME
    element SERVER01_DS datastore
    master MYDB on "SERVER01_REP"
    transmit nondurable
    subscriber MYDB on "SERVER02_REP"
    element SERVER02_DS datastore
    master MYDB on "SERVER02_REP"
    transmit nondurable
    subscriber MYDB on "SERVER01_REP"
    store MYDB on "SERVER01_REP"
    port 16004
    failthreshold 500
    store MYDB on "SERVER02_REP"
    port 16004
    failthreshold 500
    The application runs on SERVER01 and is standby on SERVER02. If an invalid state is detected in the application, the application on SERVER01 is stopped and the application on SERVER02 is started.
    In addition to this, we want to fail over if the database on the SERVER01 is in invalid state. What should we have monitored by the Clustering Agent to detect an invalid state in TT?

  • Sun Cluster 3.1  RIP or /etc/defaultroute

    Now built a cluster Sun cluster 3.1 without specifying a static route. Possible to specify a gateway to the file /etc/defaultroute? Problems may arise in the future?

    This is more or less depending on your choice. Both options are possible - either using a routing protocol, and the Solaris Cluster nodes consume those informations (which has the advantage of a centralized configuration), or the Solaris Cluster nodes get a static configuration. The correct config file is /etc/defaultrouter.
    More informations on how to configure routing on Solaris can be found at [http://docs.sun.com/app/docs/doc/816-4554/gcvjj?l=en&a=view|http://docs.sun.com/app/docs/doc/816-4554/gcvjj?l=en&a=view].
    In any case, the router infrastructure to be used by the Solaris Cluster nodes should be highly available. Also be aware that a cluster node itself must not be configured as a router itself.
    Note that simply adding something to /etc/defaultrouter does not yet make the configuration active. You would either need to also manually invoke the route command to add the default router, or you have to reboot (which you should do at some point, to test if your change really works).
    Regards
    Thorsten

Maybe you are looking for

  • How do I reinstall mail.app on OS 10.6.8 without an install disk

    Software update recently updated my software automatically to 10.6.8 on my MacBook 2.4 Ghz Intel Core Duo. Apparently because i organize my applications into folders inside the root Applications folder, the mail app was not updated properly. I put th

  • Need help in URL Service Data Control

    Hi Everyone, I'm working on JDev11gUp2, ADF-RC I am using URL Service Data Control to fetch data from CSV File and i got success but I need to know that is there any way to pass File Connection URL Dynamically at run time, As i know we have to provid

  • Error Destination in JMS Transport Configuration page (Proxy Services)

    Hi, Can anybody provides me the sample of Error Destination in JMS Transport Configuration page (Proxy Services) .I need to know the format. I am using "jms://<host:port[,host:port]*/factoryJndiName/destJndiName>", but it does not work. From Oracle w

  • JDBC Lookup  with BPM (no operation mapping)

    HI, In my scenario i have to use  JDBC lookup to get the data ,but  here i do not have operation mapping  ( this part i am doing  in BPM).. to select JDBC Look  up communication channel. In BPM i am facing  issue to select  lookup  parameters. -->In

  • Material  Creation with Classification Data

    Hi friends, I am technical guy, I need to know how to create material(MM01)  with classification data, if i want fetch data , where i will find that data ( means which tables). Thanks in Advance, regards, Bhaksar