Configuring Fail Over Clustering

I was assigned to configure the New DIA Line for our Data Center for Fail over (Back Up)
My Existing Line is - ISP ==} Router 2811 ==} Juniper SSG320 ==} Cisco Core Switch 6500
For Back Up We get additional line. 
How can I configure the Fail Over Clustering the new DIA Lines?
Please see attached my Network Diagram.
Thank you,
Michael  

Dear All,
My problem is solved by disabling antivirus.
thanks for the support
Sunil
SUNIL PATEL SYSTEM ADMINISTRATOR

Similar Messages

  • Is it possible to add hyper-V fail over clustering afterwards?

    Hi,
    We are testing Windows 2012R2 Hyper-V using only one stand alone host without fail over clustering now with few virtual machines. Is it possible to add fail over clustering afterwards and add second Hyper-V node and shared disk and move virtual
    machines there or do we have to install both nodes from scratch?
    ~ Jukka ~

    Hi Jukka,
    Inaddition, before you build hyper-v failover cluster please refer to these requirements within the article below :
    http://technet.microsoft.com/en-us/library/jj863389.aspx
    Best Regards
    Elton Ji
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Support for fail over/clustering?

    Is there any support for fail over/clustering Oracle on a two-node Linux cluster? Is it even possible to run Oracle on a two-node Linux cluster?

    To add more, I'm unsure whether an HA product is even needed once Oracle Releases Parallel Server. Also, is it possible to run Oracle w/o Parallel Server on a cluster? If so, what does that buy you?

  • Multiple types of database and fail over clustering

    Hi,
    I have a few questions here.
    1) Can I have 2 types of databases (eg: OLTP and OLAP)run at the same time on a same machine?
    2) Can I implement a cross fail over cluster in this situation? Meaning I have 2 machines with OLAP and OLTP database instances installed in them (replica of each other), 1st machine running OLTP and 2nd running OLAP. In the situation where one of machines fail, the passive instance on the other machine takes over (back to situation on question 1).
    Thanks
    Regards
    Lai Ling

    Dear All,
    My problem is solved by disabling antivirus.
    thanks for the support
    Sunil
    SUNIL PATEL SYSTEM ADMINISTRATOR

  • Failover cluster server - File Server role is clustered - Shadow copies do not seem to travel to other node when failing over

    Hi,
    New to 2012 and implementing a clustered environment for our File Services role.  Have got to a point where I have successfully configured the Shadow copy settings.
    Have a large (15tb) disk.  S:
    Have a VSS drive (volume shadow copy drive) V:
    Have successfully configured through Windows Explorer the Shadow copy settings.
    Created dependencies in Failcover Cluster Server console whereby S: depends on V:
    However, when I failover the resource and browse the Client Access Point share there are no entries under the "Previous Versions" tab. 
    When I visit the S: drive in windows explorer and open the Shadow copy dialogue box, there are entries showing the times and dates of the shadow copies ran when on the original node.  So the disk knows about the shadow copies that were ran on the
    original node but the "previous versions" tab has no entries to display.
    This is in a 2012 server (NOT R2 version).
    Can anyone explain what might be the reason?  Do I have an "issue" or is this by design?
    All help apprecieated!
    Kathy
    Kathleen Hayhurst Senior IT Support Analyst

    Hi,
    Please first check the requirements in following article:
    Using Shadow Copies of Shared Folders in a server cluster
    http://technet.microsoft.com/en-us/library/cc779378(v=ws.10).aspx
    Cluster-managed shadow copies can only be created in a single quorum device cluster on a disk with a Physical Disk resource. In a single node cluster or majority node set cluster without a shared cluster disk, shadow copies can only be created and managed
    locally.
    You cannot enable Shadow Copies of Shared Folders for the quorum resource, although you can enable Shadow Copies of Shared Folders for a File Share resource.
    The recurring scheduled task that generates volume shadow copies must run on the same node that currently owns the storage volume.
    The cluster resource that manages the scheduled task must be able to fail over with the Physical Disk resource that manages the storage volume.
    If you have any feedback on our support, please send to [email protected]

  • RAC using SUN Geo Clusters with Fail over

    Hi ,
    My customer is in the process of investigating and deploying Sun GeoClusters to fail over a RAC from one location to another, the distance between the primary and fail over site is 1200km, they are going to use TrueCopy to replicate the storage across the sites.
    I am in the process of gathering information and need to find out more detail and if any one has any knowledge of this software.If anybody knows about the clients who are using(some urls) the same please let me know.
    Regards
    Manoj

    TrueCopy is a way of replicating storage offsite. RAC works using a single source for the database. That means that RAC can not be used simultaneously at both locations with the files being used locally.
    If my memory serves me well, Hitachi TrueCopy was OSCP (oracle storage compatiblity program) certified, but the OSCP program seems to be discontinued per januari 2007 (see http://www.oracle.com/technology/deploy/availability/htdocs/oscp.html)
    That means that you can use TrueCopy to replicate the storage layer to another location (according to the OSCP note), and use the replicated storage to startup the RAC database in case of failover.

  • RAC - Oracle Grid Infrastructure configure failed

    Hi, am trying to install 2 node RAC on Oracle VMs. Before the installation during the -preinst check there were few issues which were resolved (ex user equivalence). After that during the installation process of the Grid it failed at step "Configure Oracle Grid Infrastructure for a cluster". After it failed at this step, subsequent steps too failed which I asked OUI to ignore and then I ran both the post installation scripts. And then ran post crsinst which failed. Pasting below the output of the root.sh script, post crsinst and other checks.
    [root@bsfrac01 grid]# sh root.sh
    Running Oracle 11g root.sh script...
    The following environment variables are set as:
    ORACLE_OWNER= oracle
    ORACLE_HOME= /u01/app/11.2/grid
    Enter the full pathname of the local bin directory: [usr/local/bin]:
    Copying dbhome to /usr/local/bin ...
    Copying oraenv to /usr/local/bin ...
    Copying coraenv to /usr/local/bin ...
    Creating /etc/oratab file...
    Entries will be added to the /etc/oratab file as needed by
    Database Configuration Assistant when a database is created
    Finished running generic part of root.sh script.
    Now product-specific root actions will be performed.
    2011-02-13 00:11:55: Parsing the host name
    2011-02-13 00:11:55: Checking for super user privileges
    2011-02-13 00:11:55: User has super user privileges
    Using configuration parameter file: /u01/app/11.2/grid/crs/install/crsconfig_params
    Creating trace directory
    LOCAL ADD MODE
    Creating OCR keys for user 'root', privgrp 'root'..
    Operation successful.
    root wallet
    root wallet cert
    root cert export
    peer wallet
    profile reader wallet
    pa wallet
    peer wallet keys
    pa wallet keys
    peer cert request
    pa cert request
    peer cert
    pa cert
    peer root cert TP
    profile reader root cert TP
    pa root cert TP
    peer pa cert TP
    pa peer cert TP
    profile reader pa cert TP
    profile reader peer cert TP
    peer user cert
    pa user cert
    Adding daemon to inittab
    CRS-4123: Oracle High Availability Services has been started.
    ohasd is starting
    CRS-2672: Attempting to start 'ora.gipcd' on 'bsfrac01'
    CRS-2672: Attempting to start 'ora.mdnsd' on 'bsfrac01'
    CRS-2676: Start of 'ora.mdnsd' on 'bsfrac01' succeeded
    CRS-2676: Start of 'ora.gipcd' on 'bsfrac01' succeeded
    CRS-2672: Attempting to start 'ora.gpnpd' on 'bsfrac01'
    CRS-2676: Start of 'ora.gpnpd' on 'bsfrac01' succeeded
    CRS-2672: Attempting to start 'ora.cssdmonitor' on 'bsfrac01'
    CRS-2676: Start of 'ora.cssdmonitor' on 'bsfrac01' succeeded
    CRS-2672: Attempting to start 'ora.cssd' on 'bsfrac01'
    CRS-2672: Attempting to start 'ora.diskmon' on 'bsfrac01'
    CRS-2676: Start of 'ora.diskmon' on 'bsfrac01' succeeded
    CRS-2676: Start of 'ora.cssd' on 'bsfrac01' succeeded
    CRS-2672: Attempting to start 'ora.ctssd' on 'bsfrac01'
    CRS-2676: Start of 'ora.ctssd' on 'bsfrac01' succeeded
    ASM created and started successfully.
    DiskGroup DATA1 created successfully.
    clscfg: -install mode specified
    Successfully accumulated necessary OCR keys.
    Creating OCR keys for user 'root', privgrp 'root'..
    Operation successful.
    CRS-2672: Attempting to start 'ora.crsd' on 'bsfrac01'
    CRS-2676: Start of 'ora.crsd' on 'bsfrac01' succeeded
    CRS-4256: Updating the profile
    Successful addition of voting disk 0ea2052d8a714fd7bf46d9d5c785483e.
    Successfully replaced voting disk group with +DATA1.
    CRS-4256: Updating the profile
    CRS-4266: Voting file(s) successfully replaced
    ## STATE File Universal Id File Name Disk group
    1. ONLINE 0ea2052d8a714fd7bf46d9d5c785483e (ORCL:DISK1) [DATA1]
    Located 1 voting disk(s).
    *Failed to rmtcopy "/tmp/filekRIMbG" to "/u01/app/11.2/grid/gpnp/manifest.txt" for nodes {bsfrac01,bsfrac02}, rc=256*
    *Failed to rmtcopy "/u01/app/11.2/grid/gpnp/bsfrac01/profiles/peer/profile.xml" to "/u01/app/11.2/grid/gpnp/profiles/peer/profile.xml" for nodes {bsfrac01,bsfrac02}, rc=256*
    rmtcopy aborted
    Failed to promote local gpnp setup to other cluster nodes
    CRS-2673: Attempting to stop 'ora.crsd' on 'bsfrac01'
    CRS-2677: Stop of 'ora.crsd' on 'bsfrac01' succeeded
    CRS-2673: Attempting to stop 'ora.asm' on 'bsfrac01'
    CRS-2677: Stop of 'ora.asm' on 'bsfrac01' succeeded
    CRS-2673: Attempting to stop 'ora.ctssd' on 'bsfrac01'
    CRS-2677: Stop of 'ora.ctssd' on 'bsfrac01' succeeded
    CRS-2673: Attempting to stop 'ora.cssdmonitor' on 'bsfrac01'
    CRS-2677: Stop of 'ora.cssdmonitor' on 'bsfrac01' succeeded
    CRS-2673: Attempting to stop 'ora.cssd' on 'bsfrac01'
    CRS-2677: Stop of 'ora.cssd' on 'bsfrac01' succeeded
    CRS-2673: Attempting to stop 'ora.gpnpd' on 'bsfrac01'
    CRS-2677: Stop of 'ora.gpnpd' on 'bsfrac01' succeeded
    CRS-2673: Attempting to stop 'ora.gipcd' on 'bsfrac01'
    CRS-2677: Stop of 'ora.gipcd' on 'bsfrac01' succeeded
    CRS-2673: Attempting to stop 'ora.mdnsd' on 'bsfrac01'
    CRS-2677: Stop of 'ora.mdnsd' on 'bsfrac01' succeeded
    Initial cluster configuration failed. See /u01/app/11.2/grid/cfgtoollogs/crsconfig/rootcrs_bsfrac01.log for details
    [root@bsfrac01 grid]#
    [oracle@bsfrac01 bin]$ ./cluvfy stage -post crsinst -n bsfrac01,bsfrac02 -verbose
    Performing post-checks for cluster services setup
    Checking node reachability...
    Check: Node reachability from node "bsfrac01"
    Destination Node Reachable?
    bsfrac01 yes
    bsfrac02 yes
    Result: Node reachability check passed from node "bsfrac01"
    Checking user equivalence...
    Check: User equivalence for user "oracle"
    Node Name Comment
    bsfrac01 passed
    bsfrac02 passed
    Result: User equivalence check passed for user "oracle"
    ERROR:
    PRKC-1094 : Failed to retrieve the active version of crs: {0}
    Checking time zone consistency...
    Time zone consistency check passed.
    ERROR:
    PRKC-1093 : Failed to retrieve the version of crs software on node "java.io.IOException: /u01/app/11.2.0/grid/bin/crsctl: not found
    " : {1}
    ERROR:
    Cluster manager integrity check failed
    PRVF-5434 : Cannot identify the current CRS software version
    UDev attributes check for OCR locations started...
    Result: UDev attributes check passed for OCR locations
    UDev attributes check for Voting Disk locations started...
    ERROR:
    PRVF-5197 : Failed to retrieve voting disk locationsPRKC-1092 : Failed to retrieve the location of votedisks: java.io.IOException: /u01/app/11.2.0/grid/bin/crsctl: not found
    Result: UDev attributes check failed for Voting Disk locations
    Check default user file creation mask
    Node Name Available Required Comment
    bsfrac01 0022 0022 passed
    bsfrac02 0022 0022 passed
    Result: Default user file creation mask check passed
    Checking cluster integrity...
    Node Name
    bsfrac01
    Cluster integrity check failed This check did not run on the following node(s):
    bsfrac02
    Checking OCR integrity...
    Checking the absence of a non-clustered configuration...
    All nodes free of non-clustered, local-only configurations
    ERROR:
    PRKC-1094 : Failed to retrieve the active version of crs: {0}
    ERROR:
    PRVF-5300 : Failed to retrieve active version for CRS on this node
    OCR integrity check failed
    Checking CRS integrity...
    ERROR:
    PRKC-1094 : Failed to retrieve the active version of crs: {0}
    ERROR:
    PRVF-5300 : Failed to retrieve active version for CRS on this node
    CRS integrity check failed
    OCR detected on ASM. Running ACFS Integrity checks...
    Starting check to see if ASM is running on all cluster nodes...
    PRVF-5137 : Failure while checking ASM status on node "bsfrac01"
    PRVF-5137 : Failure while checking ASM status on node "bsfrac02"
    Starting Disk Groups check to see if at least one Disk Group configured...
    PRVF-5112 : An Exception occurred while checking for Disk Groups
    PRVF-5114 : Disk Group check failed. No Disk Groups configured
    Task ACFS Integrity check failed
    Checking Oracle Cluster Voting Disk configuration...
    ERROR:
    PRKC-1093 : Failed to retrieve the version of crs software on node "java.io.IOException: /u01/app/11.2.0/grid/bin/crsctl: not found
    " : {1}
    ERROR:
    PRVF-5434 : Cannot identify the current CRS software version
    PRVF-5431 : Oracle Cluster Voting Disk configuration check failed
    Checking to make sure user "oracle" is not in "root" group
    Node Name Status Comment
    bsfrac01 does not exist passed
    bsfrac02 does not exist passed
    Result: User "oracle" is not part of "root" group. Check passed
    Post-check for cluster services setup was unsuccessful on all the nodes.
    [oracle@bsfrac01 bin]$ /u01/app/11.2/grid/bin/ocrcheck
    Status of Oracle Cluster Registry is as follows :
    Version : 3
    Total space (kbytes) : 262120
    Used space (kbytes) : 408
    Available space (kbytes) : 261712
    ID : 1671840043
    Device/File Name : +DATA1
    Device/File integrity check succeeded
    Device/File not configured
    Device/File not configured
    Device/File not configured
    Device/File not configured
    Cluster registry integrity check succeeded
    Logical corruption check bypassed due to non-privileged user
    ASM looks to be up and running..
    [oracle@bsfrac01 bin]$ /usr/sbin/oracleasm listdisks
    DISK1
    DISK2
    DISK3
    DISK4
    DISK5
    DISK6
    [oracle@bsfrac01 bin]$ /usr/sbin/oracleasm status
    Checking if ASM is loaded: yes
    Checking if /dev/oracleasm is mounted: yes
    Please help.

    before installation have u configure the private interconnect on both the nodes to same network adapter..
    for example on node 1 if the private interconnect is on eth0 then on the node 2 it should use eth0 only...
    for private interconnect use the hostonly option on both the nodes in the network configuration page of the vmware or virtual box..
    and for public network it can be bridged
    more over if you are installing on the laptop its good to configure the SSH using the OUI.. rather than doing it manually.. as it saves time
    the private and the public networks should not have same range of ip address. like if public address are like 192.168.2.222/255.255.255.0 and private address have to be different like 10.10.1.2/255.0.0.0 (this is just an example)
    have to configured the NTP.
    any ways try installing the oracle rac on virtual box follow the steps given the below website they are pretty straight forward...
    http://www.oracle-base.com/articles/11g/OracleDB11gR2RACInstallationOnOEL5UsingVirtualBox.php

  • Failing over Oracle connections in a pool

              Hi,
              This message is probably a bit out of context (I've already posted
              it to the JDBC group). I post here as well, since I guess it's
              the place where people have the most experience with clustering
              and HA. Original posting below...
              Could you please tell me whether, yes or no, connections to an
              Oracle database should fail over (when the database fails over
              to another machine)? I use Oracle's Transparent Application Failover
              (configured via Net8) with Weblogic 6 on Linux and Oracle 8.1.7
              on Solaris/SPARC.
              If this doesn't work in my configuration, is there any configuration
              where it should work? (Another version of Oracle, WLS, OS, ...)
              When I try TAF using the PetStore application, I get exceptions
              related to no being connected to the database.
              If TAF doesn't work with WebLogic, is there a way to work around
              the problem? Can I catch these exceptions and renew the connections
              in the pool? Or, what else is possible...?
              I'd appreciate any help. I'd like to demonstrate our HA product
              with WLS. If it doesn't work, I'll turn to iPlanet instead. Pity,
              I really like WLS!
              Thanks in advance for any help or advice!
              Regards, Frank Olsen
              

              Hi (Frank ;-)
              I got carried away a bit too fast...
              Some more testing shows that it doesn't work in all cases:
              - when someone is trying to check out the shopping cart when the
              the database fails (and fails over), I get exceptions once the
              databses has restarted on the backup node
              - the exceptions are related to some transactions being rolled
              back and Oracle stating that it couldn't safely replay the transactions
              - browsing the categories still works fine
              - all access to the shopping cart and sign-in/sign-out causes time-outs
              and exceptions
              Any ideas what may cause this problem, please?
              Regards,
              Frank Olsen
              "Frank Olsen" <[email protected]> wrote:
              >
              >Hi,
              >
              >TAF worked with WLS 6 on NT with the Oracle 8.1.7 client!
              >
              >Has anyone tested it on Solaris/SPARC?
              >
              >Regards,
              >Frank Olsen
              >
              >
              >
              >"Frank Olsen" <[email protected]> wrote:
              >>
              >>Hi,
              >>
              >>Most of my question below is still valid (in particular
              >>concerning
              >>whether TAF should work with WLS on some or all platforms
              >>and
              >>versions).
              >>
              >>However, when I tested TAF with the Oracle client (sqlplus)
              >>there
              >>also was no failover of the (one) connection. I then
              >checked
              >>the
              >>`V$SESSION' view and the colums related to failover showed
              >>that
              >>TAF was not correctly configured. Strange because I copied
              >>the
              >>`tnsnames.ora' parameters from the Oracle documentation
              >>for TAF.
              >>
              >>Has anyone managed to configure and use TAF, with or
              >without
              >>WLS?!
              >>
              >>Thanks in advance for your help!
              >>
              >>Regards,
              >>Frank Olsen
              >>
              >>
              >>"Frank Olsen" <[email protected]> wrote:
              >>>
              >>>Hi,
              >>>
              >>>This message is probably a bit out of context (I've
              >already
              >>>posted
              >>>it to the JDBC group). I post here as well, since I
              >guess
              >>>it's
              >>>the place where people have the most experience with
              >>clustering
              >>>and HA. Original posting below...
              >>>
              >>>----
              >>>
              >>>Could you please tell me whether, yes or no, connections
              >>>to an
              >>>Oracle database should fail over (when the database
              >fails
              >>>over
              >>>to another machine)? I use Oracle's Transparent Application
              >>>Failover
              >>>(configured via Net8) with Weblogic 6 on Linux and Oracle
              >>>8.1.7
              >>>on Solaris/SPARC.
              >>>
              >>>If this doesn't work in my configuration, is there any
              >>>configuration
              >>>where it should work? (Another version of Oracle,
              >WLS,
              >>>OS, ...)
              >>>
              >>>
              >>>When I try TAF using the PetStore application, I get
              >>exceptions
              >>>related to no being connected to the database.
              >>>
              >>>If TAF doesn't work with WebLogic, is there a way to
              >>work
              >>>around
              >>>the problem? Can I catch these exceptions and renew
              >the
              >>>connections
              >>>in the pool? Or, what else is possible...?
              >>>
              >>>I'd appreciate any help. I'd like to demonstrate our
              >>HA
              >>>product
              >>>with WLS. If it doesn't work, I'll turn to iPlanet instead.
              >>>Pity,
              >>>I really like WLS!
              >>>
              >>>Thanks in advance for any help or advice!
              >>>
              >>>Regards, Frank Olsen
              >>>
              >>
              >
              

  • Is my installation of SQL Server Fail Over cluster correct?

    I made a 2 node SQL Server 2012 fail over cluster but having some problems during installation so I wanted to know if the steps below I performed are correct.
    Hardware
    Node1 192.168.1.10
    Node2 192.168.1.11
    Added following entries in DNS
    cluster.domain.local 192.168.1.12 (for Windows Cluster)
    msdtc.domain.local 192.168.1.13 (for MSDTC)
    sql.domain.local 192.168.1.14 (for SQL Server Cluster)
    Cluster Storage
    Disk1 (for Quorum)
    Disk2 (for MSDTC
    Disk3 (for SQL Server)
    Now comes the installation. I am performing all these steps as DOMAIN ADMIN.
    1. First I installed clustering role on both nodes
    2. Then I ran fail over validation wizard on Node1 adding both nodes which went fine (there were some warnings)
    3. Then I made a Windows Cluster on Node1 using these two nodes. I gave the name and IP to this cluster which I wrote above i.e. cluster.domain.local 192.168.1.12
    4. Cluster was created and boths nodes are UP.
    Now I want to ask a question here. Is it best practice to perform the above operation using DOMAIN ADMIN? Or if I use a standard domain user account with local admin rights, will it work? If not then exactly what rights are required to perform this operation.
    5. Then I installed "Application Server" role on both Node1 and Node2 and also added "Distributed Transaction" feature
    6. Then I right clicked on Windows Cluster I created and added a new role/feature which is "DTC"
    7. I gave it the same name which I wrote above i.e. msdtc.domain.local 192.168.1.13
    8. MSDTC was created but when it tried to UP its service, it threw an error. Upon investigation it turns out the Windows Cluster cluster.domain.local doesn't have proper rights to created some objects in AD. I didn't know what rights to give so I gave it full
    permission and after that when I created MSDTC again, the service went up fine.
    So I want to know what rights does cluster.domain.com require to make MSDTC?
    Am I doing good so far?

    Hello,
    >>Then I made a Windows Cluster on Node1 using these two nodes. I gave the name and IP to this cluster which I wrote above i.e. cluster.domain.local 192.168.1.10
    Hello I suppose this IP was physical node IP windows cluster IP was 192.168.1.12  I suppose yo must have given this IP as windows cluster IP.10 and 11 are physical nodes in Cluster but 12 is Cluster IP .Correct me if I am wrong.
    Did you do failover and failback to check whether cluster is configured correctly or not ,If not please do it .
    >>Then I ran fail over validation wizard on Node1 adding both nodes which went fine (there were some warnings)
    Please remove warnings also ,it might cause issue.Not sure its correct every time but make sure cluster validation should be free of error and warning.
    >>Now I want to ask a question here. Is it best practice to perform the above operation using DOMAIN ADMIN?
    You can do it with domain admin account as this is required to create Cluster NAme object(CNO) in domain and local account might not have that right so I would say its ok.
    >>I gave it the same name which I wrote above i.e. msdtc.domain.local
    192.168.1.11
    again this IP is node 2 IP how can you give it to MSDTC.Use below link for reference
    http://blogs.msdn.com/b/cindygross/archive/2009/02/22/how-to-configure-dtc-for-sql-server-in-a-windows-2008-cluster.aspx
    Please mark this reply as the answer or vote as helpful, as appropriate, to make it useful for other readers

  • Load balancing / fail over

    Dear forum,
    Is there an other and more simple / "cheaper" way for load balancing / fail over then using RAC ? E.g. starting more instances againts the (same) database ?
    Thanks in advance,
    Michel
    Edited by: Michel77 on Dec 9, 2008 12:06 PM

    The answer would be highly platform and version dependent.
    On several platforms you have software implementing a 'shared nothing cluster'. This means you failover complete services to a second server, which is usually dormant.
    There is HP Service Guard, Suncluster, IBM HACMP, and Microsoft cluster, which are or can be configured as shared nothing clusters.
    For Microsoft there was or is, depending on version, Oracle Failsafe, on top of a MS Cluster.
    RAC can be configured in a shared nothing mode too. RAC itself is essentially active/active.
    On the other side, you can of course buy a second server, and use Dataguard to set up a standby database.
    For a physical standby database, which isn't enabled for more than 10 days, you don't pay a license.
    Setting up multiple instances on a single server is obviously not going to work. Those instances need to coordinate what they are doing. This is essentially the task of RAC.
    Hth
    Sybrand Bakker
    Senior Oracle DBA

  • Load-balancing and fail-over between web containers and EJB containers

    When web components and EJB components are run in different OC4J instances, can we achieve load-balancing and fail-over between web containers and EJB containers?
    null

    It looks like there is clustering, but not loadbalancing available for rmi
    from the rmi.xml configuration. The application will treat any ejbs on the
    cluster as one-to-one look-ups. Orion will go out and get the first ejb
    available on the cluster. See the docs on configuring rmi.xml (and also the
    note below).
    That is a kind-of failover, because if machine A goes down, and the
    myotherAejbs.jar are on machine B too, orion will go out and get the bean
    from machine B when it can't find machine A. But it doesn't go machine A
    then machine B for each remote instance of the bean. You could also specify
    the maximum number of instances of a bean, and as one machine gets "loaded",
    orion would go to the next available machine...but that's not really
    loadbalancing.
    That is, you can set up your web-apps with ejbs, but let all of the ejbs be
    remote="true" in the orion-application.xml file:
    <?xml version="1.0"?>
    <!DOCTYPE orion-application PUBLIC "-//Evermind//DTD J2EE Application
    runtime 1.2//EN" "http://www.orionserver.com/dtds/orion-application.dtd">
    <orion-application deployment-version="1.5.2">
    <ejb-module remote="true" path="myotherAejbs.jar" />
    <ejb-module remote="true" path="myotherBejbs.jar" />
    <ejb-module remote="true" path="myotherCejbs.jar" />
    &ltweb-module id="mysite" path="mysite.war" />
    ... other stuff ...
    </orion-application>In the rmi.xml you would define your clustering:
    <cluster host="230.0.0.1" id="123" password="123abc" port="9127"
    username="cluster-user" />
    Tag that is defined if the application is to be clustered. Used to set up
    a local multicast cluster. A username and password used for the servers to
    intercommunicate also needs to be specified.
    host - The multicast host/ip to transmit and receive cluster data on. The
    default is 230.0.0.1.
    id - The id (number) of this cluster node to identify itself with in the
    cluster. The default is based on local machine IP.
    password - The password configured for cluster access. Needs to match that
    of the other nodes in the cluster.
    port - The port to transmit and receive cluster data on. The default is
    9127.
    username - The username configured for cluster access. Needs to match that
    of the other nodes in the cluster.

  • Stateful bean not failing over

              I have a cluster of two servers and a Admin server. Both servers are running NT
              4 sp6 and WLS6 sp1.
              When I stop one of the servers, the client does n't automatically failover to
              the other server, instead it fails unable to contact server that has failed.
              My bean is configured to have its home clusterable and is a stateful bean. My
              client holds onto the remote interface, and makes calls through this. If Server
              B fails then it should automatically fail over to server A.
              I have tested my multicast address and all seems to be working fine between servers,
              my stateless bean work well, load balancing between servers nicely.
              Does anybody have any ideas, regarding what could be causing the stateful bean
              remote interface not to be providing failover info.
              Also is it true that you can have only one JMS destination queue/topic per cluster..The
              JMS cluster targeting doesn't work at the moment, so you need to deploy to individual
              servers?
              Thanks
              

    Did you enable stateful session bean replication in the
              weblogic-ejb-jar.xml?
              -- Rob
              Wayne Highland wrote:
              >
              > I have a cluster of two servers and a Admin server. Both servers are running NT
              > 4 sp6 and WLS6 sp1.
              > When I stop one of the servers, the client does n't automatically failover to
              > the other server, instead it fails unable to contact server that has failed.
              >
              > My bean is configured to have its home clusterable and is a stateful bean. My
              > client holds onto the remote interface, and makes calls through this. If Server
              > B fails then it should automatically fail over to server A.
              >
              > I have tested my multicast address and all seems to be working fine between servers,
              > my stateless bean work well, load balancing between servers nicely.
              >
              > Does anybody have any ideas, regarding what could be causing the stateful bean
              > remote interface not to be providing failover info.
              >
              > Also is it true that you can have only one JMS destination queue/topic per cluster..The
              > JMS cluster targeting doesn't work at the moment, so you need to deploy to individual
              > servers?
              >
              > Thanks
              Coming Soon: Building J2EE Applications & BEA WebLogic Server
              by Michael Girdley, Rob Woollen, and Sandra Emerson
              http://learnweblogic.com
              

  • NIC not failing Over in Cluster

    Hi there...I have configured 2 Node cluster with SoFS role...for VM Cluster and HA using Windows Server 2012 Data Center. Current set up is Host Server has 3 NICS (2 with Default Gateway setup (192.x.x.x), 3 NIC is for heartbeat 10.X.X.X). Configured CSV
    (can also see the shortcut in the C:\). Planning to setup few VMs pointing to the disk in the 2 separate storage servers (1 NIC in 192.x.x.x) and also have 2 NIC in 10.x.x.x network. I am able to install VM and point the disk to the share in the cluster volume
    1. 
    I have created 2 VM Switch for 2 separate Host server (using Hyper-V manager). When I test the functionality by taking Node 2, I can see the Disk Owner node is changing to Node 1, but the VM NIC 2 is not failing over automatically to VM NIC 1 (but I can
    see the VM NIC 1 is showing up un-selected in the VM Settings). when I go to the VM Settings > Network Adapter, I get error -
    An Error occurred for resource VM "VM Name". select the "information details" action to view events for this resource. The network adapter is configures to a switch which no longer exists or a resource
    pool that has been deleted or renamed (with configuration error in "Virtual Switch" drop down menu). 
    Can you please let me know any resolution to fix this issue...Hoping to hear from you.
    VT

    Hi,
    From your description “My another thing I would like to test is...I also would like to bring a disk down (right now, I have 2 disk - CSV and one Quorum disk) for that 2 node
    cluster. I was testing by bringing a csv disk down, the VM didnt failover” Are you trying to test the failover cluster now? If so, please refer the following related KB:
    Test the Failover of a Clustered Service or Application
    http://technet.microsoft.com/en-us/library/cc754577.aspx
    Hope this helps.
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Load balancing not happending but fail over is for Read only Entity beans

              The following are the configuration.
              Two NT servers with WL5.1 sp9 having only EJBs(Read only entity beans)
              One Client with WL5.1 sp9 having servlet/java application as
              EJB client.
              I am trying to make a call like findbyprimarykey in one of the
              entity bean. I could see the request is being directed only to the one of the
              server always. When I bring that server, fail over is happening to the other server.
              Here are the settings I have in the ejb-jar.xml :
                        <entity>
                             <ejb-name>device.StartHome</ejb-name>
                             <home>com.wl.api.device.StartHome</home>
                             <remote>com.wl.api.device.StartRemote</remote>
                             <ejb-class>com.wl.server.device.StartImpl</ejb-class>
                             <persistence-type>Bean</persistence-type>
                             <prim-key-class>java.lang.Long</prim-key-class>
                             <reentrant>False</reentrant>
                             <resource-ref>
                                  <res-ref-name>jdbc/wlPool</res-ref-name>
                                  <res-type>javax.sql.DataSource</res-type>
                                  <res-auth>Container</res-auth>
                             </resource-ref>
                        </entity>
              Here are the settings I have in the weblogic-ejb-jar.xml.
              <weblogic-enterprise-bean>
                        <ejb-name>device.StartHome</ejb-name>
                        <caching-descriptor>
                             <max-beans-in-cache>50</max-beans-in-cache>
                             <cache-strategy>Read-Only</cache-strategy>
                             <read-timeout-seconds>900</read-timeout-seconds>
                        </caching-descriptor>
                        <reference-descriptor>
                             <resource-description>
                                  <res-ref-name>jdbc/wlPool</res-ref-name>
                                  <jndi-name>weblogic.jdbc.pool.wlPool</jndi-name>
                             </resource-description>
                        </reference-descriptor>
                        <enable-call-by-reference>False</enable-call-by-reference>
                        <jndi-name>device.StartHome</jndi-name>
                   </weblogic-enterprise-bean>
              Am I doin any mistake in this?
              Any one's help is appreciated.
              Thanks
              Suresh
              

    we are using 5.1
              "Gene Chuang" <[email protected]> wrote in message
              news:[email protected]...
              > Colocation optimization occurs if your client resides in the same
              container (and also in the same
              > EAR for 6.0) as your ejbs.
              >
              > Gene
              >
              > "Suresh" <[email protected]> wrote in message
              news:[email protected]...
              > > Ok....the ejb-call-by-reference set to true is making the call to one
              server
              > > only. i am not sure why it is. I removed the property name and it
              works.
              > > Also I have one question, in our prduct environment, when i cache the
              ejb
              > > home it is not doing the load balancing. can any one help me for that.
              > > thanks
              > >
              > > Mike,
              > > From the sample pgm I sent, even from single client calls get load
              > > balanced.
              > >
              > > Suresh
              > >
              > >
              > > "Gene Chuang" <[email protected]> wrote in message
              > > news:[email protected]...
              > > > In WL, LoadBalancing will ONLY WORK if you reuse your EJBHome! Take
              your
              > > StartEndPointHome lookup
              > > > out of your for loop and see if this fixes your problem.
              > > >
              > > > I've seen this discussion in ejb-interest, and some other vendor
              (Borland,
              > > I believe it is), brings
              > > > up an interesting point: Clustering and LoadBalance is not in the
              J2EE
              > > specs, hence implementation
              > > > is totally up to the vendor. Weblogic loadbalances from the remote
              > > interfaces (EJBObject, EJBHome,
              > > > etc..), while Borland loadbalances from JNDI Context lookup.
              > > >
              > > > Let me suggest a third implmentation: loadbalance from BOTH Context
              > > lookup as well as stub method
              > > > invocation! Or create a smart replica-aware list manager which
              persists
              > > on the client thread
              > > > (ThreadLocal) and is aware of lookup/evocation history. Hence if I do
              the
              > > following in a client
              > > > hitting a 3 node cluster, I'll still get perfect round-robining
              regardless
              > > of what I do on the
              > > > client side:
              > > >
              > > > InitialContext ctxt = new InitialContext();
              > > > EJBHome myHome = ctxt.lookup(MY_BEAN);
              > > > myHome.findByPrimaryKey(pk); <== hits Node #1
              > > > myHome = ctxt.lookup(MY_BEAN);
              > > > myHome.findByPrimaryKey(pk); <== hits Node #2
              > > > myHome.findByPrimaryKey(pk); <== hits Node #3
              > > > myHome = ctxt.lookup(MY_BEAN);
              > > > myHome.findByPrimaryKey(pk); <== hits Node #1
              > > > ...
              > > >
              > > >
              > > > Gene
              > > >
              > > > "Suresh" <[email protected]> wrote in message
              > > news:[email protected]...
              > > > > Mike ,
              > > > >
              > > > > Do you have any reasons for the total number of machines to be 10.
              > > > >
              > > > > I tried with 7 machines.
              > > > >
              > > > >
              > > > > Here is my sample client java application running individual in the
              > > seven
              > > > > machines.
              > > > >
              > > > > StartEndPointHome =
              > > > > (StartEndPointHome)ctx.lookup("dev.StartEndPointHome");
              > > > > for(;;)
              > > > > {
              > > > > // logMsg(" --in loop "+currentTime);
              > > > > if (currentTime > nextRefereshTime)
              > > > > {
              > > > > logMsg("****- going to call");
              > > > > currentTime=getSystemTime();
              > > > > nextRefereshTime=currentTime+timeInterval;
              > > > > StartEndPointHome =
              > > > > (StartEndPointHome)ctx.lookup("dev.StartEndPointHome");
              > > > > long rndno=(long)(Math.random()*10)+range;
              > > > > logMsg(" going to call remotestub"+rndno);
              > > > > retVal =
              > > > >
              > >
              ((StartEndPointHome)getStartHome()).findByNumber("pe"+rndno+"_mportal_dsk36.
              > > > > mportal.com");
              > > > >
              > > > > logMsg("**++- called stub");
              > > > > }
              > > > >
              > > > >
              > > > >
              > > > > The range value is different for each of the machines in the
              cluster.
              > > > >
              > > > > If the first request starts at srv1, all request starts hitting the
              same
              > > > > server.
              > > > > If the first request starts at srv2, all request starts hitting the
              same
              > > > > server.
              > > > >
              > > > > I have the following for the url , user and pwd values for the
              context
              > > .
              > > > >
              > > > > public static String url="t3://10.11.12.14,10.11.12.117:8000";
              > > > > public static String user="guest";
              > > > > public static String password="guest";
              > > > >
              > > > >
              > > > >
              > > > > It would be great if you could help me.
              > > > >
              > > > > Thanks
              > > > > suresh
              > > > >
              > > > >
              > > > > "Mike Reiche" <[email protected]> wrote in message
              > > > > news:[email protected]...
              > > > > >
              > > > > > If you have only one client don't be surprised if you only hit one
              > > server.
              > > > > Try
              > > > > > running ten different clients and see if the hit the same server.
              > > > > >
              > > > > > Mike
              > > > > >
              > > > > >
              > > > > > "suresh" <[email protected]> wrote:
              > > > > > >
              > > > > > >The following are the configuration.
              > > > > > >
              > > > > > > Two NT servers with WL5.1 sp9 having only EJBs(Read only entity
              > > beans)
              > > > > > >
              > > > > > > One Client with WL5.1 sp9 having servlet/java application as
              > > > > > > EJB client.
              > > > > > >
              > > > > > >
              > > > > > >I am trying to make a call like findbyprimarykey in one of the
              > > > > > >entity bean. I could see the request is being directed only to
              the
              > > one
              > > > > > >of the
              > > > > > >server always. When I bring that server, fail over is happening
              to
              > > the
              > > > > > >other server.
              > > > > > >
              > > > > > >
              > > > > > >Here are the settings I have in the ejb-jar.xml :
              > > > > > > <entity>
              > > > > > > <ejb-name>device.StartHome</ejb-name>
              > > > > > > <home>com.wl.api.device.StartHome</home>
              > > > > > > <remote>com.wl.api.device.StartRemote</remote>
              > > > > > > <ejb-class>com.wl.server.device.StartImpl</ejb-class>
              > > > > > > <persistence-type>Bean</persistence-type>
              > > > > > > <prim-key-class>java.lang.Long</prim-key-class>
              > > > > > > <reentrant>False</reentrant>
              > > > > > > <resource-ref>
              > > > > > > <res-ref-name>jdbc/wlPool</res-ref-name>
              > > > > > > <res-type>javax.sql.DataSource</res-type>
              > > > > > > <res-auth>Container</res-auth>
              > > > > > > </resource-ref>
              > > > > > > </entity>
              > > > > > >
              > > > > > >
              > > > > > >Here are the settings I have in the weblogic-ejb-jar.xml.
              > > > > > >
              > > > > > ><weblogic-enterprise-bean>
              > > > > > > <ejb-name>device.StartHome</ejb-name>
              > > > > > >
              > > > > > > <caching-descriptor>
              > > > > > > <max-beans-in-cache>50</max-beans-in-cache>
              > > > > > > <cache-strategy>Read-Only</cache-strategy>
              > > > > > > <read-timeout-seconds>900</read-timeout-seconds>
              > > > > > > </caching-descriptor>
              > > > > > >
              > > > > > > <reference-descriptor>
              > > > > > > <resource-description>
              > > > > > > <res-ref-name>jdbc/wlPool</res-ref-name>
              > > > > > > <jndi-name>weblogic.jdbc.pool.wlPool</jndi-name>
              > > > > > > </resource-description>
              > > > > > > </reference-descriptor>
              > > > > > > <enable-call-by-reference>False</enable-call-by-reference>
              > > > > > > <jndi-name>device.StartHome</jndi-name>
              > > > > > > </weblogic-enterprise-bean>
              > > > > > >
              > > > > > >
              > > > > > >Am I doin any mistake in this?
              > > > > > >
              > > > > > >Any one's help is appreciated.
              > > > > > >Thanks
              > > > > > >Suresh
              > > > > >
              > > > >
              > > > >
              > > >
              > > >
              > >
              > >
              >
              >
              

  • Is there a way to config WLS to fail over from a primary RAC cluster to a DR RAC cluster?

    Here's the situation:
    We have two Oracle RAC clusters, one in a primary site, and the other in a DR site
    Although they run active/active using some sort of replication (Oracle Streams? not sure), we are being asked to use only the one currently being used as the primary to prevent latency & conflict issues
    We are using this only for read-only queries.
    We are not concerned with XA
    We're using WebLogic 10.3.5 with MultiDatasources, using the Oracle Thin driver (non-XA for this use case) for instances
    I know how to set up MultiDatasources for an individual RAC cluster, and I have been doing that for years.
    Question:
    Is there a way to configure MultiDatasources (mDS) in WebLogic to allow for automatic failover between the two clusters, or does the app have to be coded to failover from an mDS that's not working to one that's working (with preference to a currently labelled "primary" site).
    Note:
    We still want to have load balancing across the current "primary" cluster's members
    Is there a "best practice" here?

    Hi Steve,
    There are 2 ways to connect WLS to a Oracle RAC.
    1. Use the Oracle RAC service URL which contains the details of all the RAC nodes and the respective IP address and DNS.
    2. Connect to the primary cluster as you are currently doing and use a MDS to load-balance/failover between multiple nodes in the primary RAC (if applicable).
        In case of a primary RAC nodes failure and switch to DR RAC nodes, use WLST scripts to change the connection URL and restart the application to remove any old connections.
        Such DB fail-over tests can be conducted in a test/reference environment to set up the required log monitoring and subsequent steps to measure the timelines.
    Thanks,
    Souvik.

Maybe you are looking for

  • Is there a way to delete all links at once?

    I have a document with a couple hundred links. How can I bulk delete them all? I know I asked this same question a few years ago, but I can no longer find the "my stuff" link in the Adobe forums, and my previous post is not coming up when I search th

  • Is it possible to install Solaris 10 from Hard-Disk

    Linux support installation from HD, will it be possible in solaris?

  • Keyword List doesn't show the correct images when clicked

    I'm using LR 2.3RC on a Windows Vista Ultimate PC, although the same error occurs on LR2.2. I have a keyword structure "Karl > Family > People" with Karl having 33 images with this keyword. When I use the "Keyword List" on the right panel in the Libr

  • How to back-grade Camera RAW to 2.1?

    Can anyone explain how to back grade to camera raw support 2.1 verses the buggy 2.2 one? Luckily, I only updated my MBP and not my main work station but I'm in need of fixing my MBP for a location project I'm doing. I've found the former 2.1 update,

  • Creative SB0410 digital out only works for dd/dts, not for stereo or web videos...

    Creative SB0410 digital out only works for dd/dts, not for stereo or web videos... Hi all, Need some suggestions here. I have an old Creative sb0410 7.1 channel card and after reading few long threads, I am able to output dd/dts via 3.5mm to rca conn