Hardware checking in Sun Cluster environment

Hi! I wonder how I can check my hardware in Sun Cluster environment... I consider to use SunVTS, but it is not supported software to use with SC. Is there any other way to test HW which works under SC instead of OBP and POST testing?.. Or maybe there are some ways to use SunVTS in some cases, for instance after booting node in non-cluster mode or something else... I need to put HW to long-term stress testing but at that moment I can think only about SunVTS...

Thanks for your answer! Are there any restrictions to hardware classes or I can test any device in my server?.. There were some opinions in [this thread|http://forums.sun.com/thread.jspa?forumID=842&threadID=5418126] that I shouldn't put stress test on disk drives...

Similar Messages

  • TimesTen database in Sun Cluster environment

    Hi,
    Currently we have our application together with the TimesTen database installed at the customer on two different nodes (running on Sun Solaris 10). The second node acts as a backup to provide failover functionality, although right now only manual failover is supported.
    We are now looking into a hot-standby / high availability solution using Sun Cluster software. As understood from the documentation, applications can be 'plugged-in' to the Sun Cluster using Agents to monitor the application. Sun Cluster Agents should be already available for certain applications such as:
    # MySQL
    # Oracle 9i, 10g (HA and RAC)
    # Oracle 9iAS Application Server
    # PostgreSQL
    (See http://www.sun.com/software/solaris/cluster/faq.jsp#q_19)
    Our question is whether Sun Cluster Agents are already (freely) available for TimesTen? If so, where to find them. If not, should we write a specific Agent separately for TimesTen or handle database problems from the application.
    Does someone have any experience using TimesTen in a Sun Cluster environment?
    Thanks in advance!

    Yes, we use 2-way replication, but we don't use cache connect. The replication is created like this on both servers:
    create replication MYDB.REPSCHEME
    element SERVER01_DS datastore
    master MYDB on "SERVER01_REP"
    transmit nondurable
    subscriber MYDB on "SERVER02_REP"
    element SERVER02_DS datastore
    master MYDB on "SERVER02_REP"
    transmit nondurable
    subscriber MYDB on "SERVER01_REP"
    store MYDB on "SERVER01_REP"
    port 16004
    failthreshold 500
    store MYDB on "SERVER02_REP"
    port 16004
    failthreshold 500
    The application runs on SERVER01 and is standby on SERVER02. If an invalid state is detected in the application, the application on SERVER01 is stopped and the application on SERVER02 is started.
    In addition to this, we want to fail over if the database on the SERVER01 is in invalid state. What should we have monitored by the Clustering Agent to detect an invalid state in TT?

  • Hardware requirement for sun cluster

    Hi,
    I have 2 ultra 10 machines.
    can u pls advise me, what brand and model of ethernet cards are cheap and best for cluster interconnect
    and also please guide me what model and brand of scsi cards are cheap and best for connecting to my multipack to configure 2 node cluster
    Best Regards'
    Mallikarjun

    Also My ide drive is giving correctable data error. will format->analyse will correct the error on ide drive also?

  • Migrate Sun Cluster (+RACdisks to new hardware running Sun Cluster ( + RAC)

    Hello,
    We have old hardware (v490s) running Sun Cluster 3.2 + Oracle RAC 10.2.0.4.0 connected to SAN. We need to move to T4. Oracle advised against including new hardware into existing cluster, so we are planning on building a new cluster with T4's, same software (Solaris 10, Sun Cluster 3.2, RAC 10.2.0.4.0).
    When ready, we plan to shut down existing cluster, zone new cluster to existing disks and bring up everything on new hardware (simply stated).
    Will it work?
    Any gotchas - like need to clear disk ids or Sun Cluster panicking? RAC panicking? Any reference docs out there?
    Thanks
    user12961096

    Do we absolutely need that in our new setup or could we forgo that additional layer? Would Sun Cluster give us anything that the OS + RAC doesn't give us?Yes, Oracle Solaris Cluster does make things a lot easier. It looks after your device space and gives you consistent DID devices for CRS/RAC. It gives you the choice to use sQFS, raw metasets, or ASM. It has clprivnet which is a lot easier and performs better than an IPMP solution. The node failure detection time is <= 10 seconds which is quicker than CRS on it's own and it uses SCSI fencing instead of a STONITH approach. Finally, you have all the off the shelf agents that Solaris Cluster offers.
    However, if you are only doing RAC and you just want ASM and you don't need the last few seconds of failure detection that OSC gives you and you think STONITH is good enough for your fencing purposes, then CRS on its own is perfect. There are many, many deployments both with and without OSC, it's not a simple yes/no answer.
    Having worked for the Solaris Cluster group, I'm still slightly bias to including it rather than going without. Others have the alternate view! :-)
    Hope that helps,
    Tim
    ---

  • 11g r2 non rac using asm for sun cluster os (two node but non-rac)

    I am going to install grid installation for non-rac using asm for two node sun cluster environment..
    How to create candidate disk in solaris cluster (sparc os) to install grid home in asm.. please provide me the steps if anyone knows

    Please refer the thread Re: 11GR2 ASM in non-rac node not starting... failing with error ORA-29701
    and this doc http://docs.oracle.com/cd/E11882_01/install.112/e24616/presolar.htm#CHDHAAHE

  • Sun Cluster & 6130/6140 thru switch with cross-connections not supported?

    Hi:
    I noticed that the 6140 does not support cross-connecting the 2 controllers to 2 switches for higher availability when using Sun Cluster:
    http://docs.sun.com/source/819-7497-10/chapter3.html
    Does anyone know why this restriction is there?
    Thanks!

    Since there was no real answer to the question in this forum, I cross posted this issue to the cluster forum.
    See http://forum.java.sun.com/thread.jspa?threadID=5261282&tstart=0 for the full thread.
    Basically, the restriction against cross-connections is no longer valid and the documentation should be updated to remove the note.
    This is all a good thing, because I had my 6140's wired into my sun cluster environment via the 'cross-connections' method diagramed in figure 3-4. :-)

  • Sun cluster, 6140 and 'cross-connections'

    This was brought up in the storage forum by somebody else, but the responses never answered the original question:
    In the 6140 setup document located at http://docs.sun.com/source/819-7497-11/chapter3.html#50589714_93886 it shows two different ways to cable a host to a 6140 via a SAN switch. figure 3-3 and 3-4
    It states that the setup in 3-4 is not supported in a sun cluster environment.
    The problem is that given the acive/passive nature of the 6140, the setup shown in 3-4 is the obvious one to use since it prevents one from having to force all of the luns over in the event of an hba port or switch failure.
    To make life more interesting, the 6140 setup doc does not make not of what version of sun cluster it is not supported on. Or what the bugid is, or any informationto know if the restriction is still valid.
    So, does this restriction still exist? If so, for what version of sun cluster? What version of solaris?

    Thanks for the clarification.
    As an aside, it would be nice if in the future, the documentation could contain a bit more information than just a simple note saying 'this is not supported'. A reference to a bugid or info doc would go a long way in helping folks determine if the restriction is still valid.
    --john                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

  • Add two more hosts and 2 clusters to a Sun cluster

    How do I add 2 more hosts, each with 2 zones, to an existing Sun cluster environment?

    This is explained in the documentation:
    [http://docs.sun.com/app/docs/doc/820-4679/cacjggea?l=en&a=view|http://docs.sun.com/app/docs/doc/820-4679/cacjggea?l=en&a=view]
    Regards
    Thorsten

  • Running a Sun Cluster on one system using containers

    Hello,
    We would like to set up a test Sun Cluster Environment on one system using containers.
    Has any one done this ? Is there documentation in the sun community on how to do this ?
    Thank you
    Bradley Duncan
    Application Support
    City University of New York

    You can not install/configure a cluster just between non-global zones on the same physical node. The cluster software always needs to be installed in the global zone.
    What you can do is install Solaris Cluster just on one physical node (a single-node cluster) and then you can configure native non-global zones and configure resource groups to failover between those non-global zones on the same node.
    The resource group has a nodelist property, where you can configure node:zonea,node:zoneb - which will allow that RG to failover between zonea and zoneb on the same node.
    A good start if your purpose is just to have a learning/development/experiment environment.
    Regards
    Thorsten

  • Upgrading Solaris OS (9 to 10)  in sun cluster 3.1 environment

    Hi all ,
    I have to upgrade the solaris OS 9 to 10 in Sun cluster 3.1.
    Sun Cluster 3.1
    data service - Netbackup 5.1
    Questions:
    1 .Best ways to upgrade the Solaris 9 to 10 and the Problems while upgrading the OS?
    2 .Sun Trunking support in Sun Cluster 3.1?
    Regards
    Ramana

    Hi Ramana
    We had used the live upgrade for upgrading Solaris 9 to 10 and its the best method for less downtime and risk but you have to follow the proper procedure as it is not the same for normal solaris. Live upgrade with sun cluster is different . you have to take into consideration about global devices and veritas volume manager. while creating new boot environment.
    Thanks/Regards
    Sadiq

  • Install of HA MaxDB in a Sun Cluster 3.2 environment

    Hi. I am installing SAP Content Server using MaxDB in a Sun Cluster 3.2 environment. According to the SUN doc, I must install the X_server as a scalable resource/service.  For a scalable resource, I have to make the disk available on both nodes of the cluster. I need to identify what are all the executables for the x_server so that I can make those mount points a NFS share.  From looking at the content of previous non-clustered installs, I am thinking I need to make the directories /sapdb/data and /sapdb/program NFS to isolate the x_server components.  Does anyone have experience with this setup ? I've placed a log with OSS but wanted to see if anyone else in the forum has done this before. Thanks for any input.

    Hello Dan,
    Please review the SAP Note No. 803452.
    X_server tool is the the part of the indepentent database software installation <independent_program_path>. Using NFS you could  set up the <independent_program_path> and <independent_data_path>  directories as share directories.
    Thank you and best regards, Natalia Khlopina

  • Dependency between Sun Cluster 2.2 and Sun hardware ?

    I am evaluating Sun Cluster 2.2 for the usage as clustering software within our project.
    Can anybody say me, if there is a dependency between Sun Cluster 2.2 and Sun hardware ?
    Can I use Sun Cluster 2.2 with other hardware platforms running Sun Solaris 7 and Veritas Volume Manager ?

    I have had this discussion with fellow cluster admins, and we think a port of cluster 2.2 could be done to Intel running Solaris. And here is the BUT, BUT SUN does not support anything except a SPARC based cluster. So currently there is definately a depedancy bewteen the 2.2 software and the hardware.
    Hope this helps
    Heath

  • Migrating a Sun Cluster Running Oracle to New Hardware

    Has anyone attempted this? Essentially we are moving a Sun Cluster from one location to hardware at another location while maintaining the same node names. From what I can tell, I need to (on an install lan):
    1) Load the OS
    2) Configure the IPs
    3) Install Sun Cluster
    4) Install Oracle Parallel Server/RAC
    5) Restore the data on a per node basis
    6) Restore the shared data
    7) Adjust, tweak, and run
    Are there any pitfalls or suggestions on the approach? The shop is relatively new to clustering much less oracle clustering and the original cluster was installed by admins gone bye.

    I would say that Apple should be able to update your 36-months maintenance agreement with a OSXS 10.4 serial number.
    As far as I know, the structure of 10.3 and 10.4 serial numbers is different (wasn't the case between 10.2 and 10.3) so I'm short of a technical answer here.
    Maybe you could try :
    /System/Library/ServerSetup/serversetup -setServerSerialNumber xxxx-xxx-xxx-x-xxx-xxx-xxx-xxx-xxx-xxx-x
    in a Terminal window on the server. It's theorically the same as using Server Admin but maybe this could help.

  • Is Veritas- or Sun Cluster needed for RAC in a Solaris Environment

    Is a Veritas- or Sun Cluster needed for RAC in a Solaris Environment?
    Does anyone know, when OCFS will be available for Solaris?

    You don't need Veritas Cluster File System, but until OCFS comes out for Solaris you need to think about backups. If you've not got a backup solution that can integrate with rman for an SBT device then backups become more tricky.
    If you use ASM then you can take a backup to a "cluster filesystem" (although ASM is raw partitions think of it as a cluster filesystem), that both nodes can see. BUT you then need to get these to tape somehow, unless you've got NetBackup et al. that support RMAN and can backup direct to tape you're more stuck.
    Too many people don't think about this. You could create an NFS mount and backup to this from the nodes.

  • Sun Cluster + meta set shared disks -

    Guys, I am looking for some instructions that most sun administrators would mostly know i believe.
    I am trying to create some cluster resource groups and resources etc., but before that i am creating the file systems that is going to be used by two nodes in the sun cluster 3.2. we use SVM.
    I have some drives that i plan to use for this specific cluster resource group that is yet to be created.
    i know i have to create a metaset since thats how other resource groups in my environment are setup already so i will go with the same concept.
    # metaset -s TESTNAME
    Set name = TESTNAME, Set number = 5
    Host Owner
    server1
    server2
    Mediator Host(s) Aliases
    server1
    server2
    # metaset -s TESTNAME -a /dev/did/dsk/d15
    metaset: server1: TESTNAME: drive d15 is not common with host server2
    # scdidadm -L | grep d6
    6 server1:/dev/rdsk/c10t6005076307FFC4520000000000004133d0 /dev/did/rdsk/d6
    6 server2:/dev/rdsk/c10t6005076307FFC4520000000000004133d0 /dev/did/rdsk/d6
    # scdidadm -L | grep d15
    15 server1:/dev/rdsk/c10t6005076307FFC4520000000000004121d0 /dev/did/rdsk/d15
    Do you see what i am trying to say ? If i want to add d6 in the metaset it will go through fine, but not for d15 since it shows only against one node as you see from the scdidadm output above.
    Please Let me know how i share the drive d15 same as d6 with the other node too. thanks much for your help.
    -Param
    Edited by: paramkrish on Feb 18, 2010 11:01 PM

    Hi, Thanks for your reply. You got me wrong. I am not asking you to be liable for the changes you recommend since i know thats not reasonable while asking for help. I am aware this is not a support site but a forum to exchange information that people already are aware of.
    We have a support contract but that is only for the sun hardware and those support folks are somewhat ok when it comes to the Solaris and setup but not that experts. I will certainly seek their help when needed and thats my last option. Since i thought this problem that i see is possibly something trivial i quickly posted a question in this forum.
    We do have a test environment but that do not have two nodes but a 1 node with zone clusters. hence i dont get to see this similar problem in the test environment and also the "cldev populate" would be of no use as well to me if i try it in the test environment i think since we dont have two nodes.
    I will check the logs as you suggested and will get back if i find something. If you have any other thoughts feel free to let me know ( dont bother about the risks since i know i can take care of that ).
    -Param

Maybe you are looking for