3 Node Cluster with 2 node SQL Always_On

We currently have more than 20 always_on clusters i.e. all following the below configurations
Server 1, Server 2 (High Hardware Configuration i.e. 16 CPU & 30 GB RAM) & File-share Witness on a small box (2 CPU with 4 GB RAM). We have seen some situations where nodes not able to communicate with the file-share and taking the cluster resource
down
Question: Can we create a windows cluster with 2 nodes i.e. (2 big hardware configuration for SQL + 1 small hardware configuration) to make it a 3 node windows cluster and have SQL on only the first 2 nodes with Always_ON ?

Hi,
Could you apply the following updates and see if it will alleviate the symptoms?
System event log errors every 15 minutes on a Windows Server 2008 R2-based Cluster configured with Node Majority and File Share Witness quorum
http://support.microsoft.com/kb/2750820
A cluster resource remains offline after the "Run this resource in a separate Resource Monitor" option is disabled in Windows Server 2008
http://support.microsoft.com/default.aspx?scid=kb;EN-US;977837
In addition, you may update the NIC drivers and disable TCP chimney offload and check the result.
Information about the TCP Chimney Offload, Receive Side Scaling, and Network Direct Memory Access features in Windows Server 2008
http://support.microsoft.com/default.aspx?scid=kb;EN-US;951037
If still no progress, upload the cluster.log to your OneDrive and share the link here.
Understanding the Cluster Debug Log in 2008
http://blogs.technet.com/b/askcore/archive/2010/04/13/understanding-the-cluster-debug-log-in-2008.aspx
Thanks.
Tracy Cai
TechNet Community Support

Similar Messages

  • Is it possible to create a two node cluster with SQL server 2012 STD on node1 and SQL Server 2008 ENT on node2 on a Windows server 2012?

    Hi All,
    Is it possible to create a two node cluster with SQL server 2012 STD on node1 and SQL Server 2008 ENT on node2 on a Windows server 2012?
    aa

    I think you are confused with the way cluster is installed. Installing SQL on cluster is a two step process. On first node you create a new SQL cluster and on second node you need "AddNode" action.
    Same edition media has to be used in above two operation.
    you can install two separate instances with one enterprise and one standard.
    Balmukund Lakhani
    Please mark solved if I've answered your question, vote for it as helpful to help other users find a solution quicker
    This posting is provided "AS IS" with no warranties, and confers no rights.
    My Blog |
    Team Blog | @Twitter
    | Facebook
    Author: SQL Server 2012 AlwaysOn -
    Paperback, Kindle

  • How to add a new instance to a 2 node SQL 2008 R2 Cluster under the same networkname

    Hello All,
    I'm fairly new to sql and have just deployed a new 2 node SQL 2008 R2 cluster.
    During installation, I created a networkname SQLCLU and a named instance I01.
    I now would like to add another Instance (I02) under the same networkname.
    What is the proper procedure to do this?
    I don't believe that I should do this using the action: "New SQL Server failover cluster" from the setup menu, but rather using the action "New Installation or add features to an existing installation"
    With the second option, I'm however not sure how I should make this instance clustered..
    Should I also execute the "New Installation or add features to an existing installation" action on the 2nd node in the cluster?
    Many thanks for your  advice!
    Filip

    You cannot use the same network name for 2 instances. You need to use a different network name for the second instance I02
    A failover cluster instance contains:
    A combination of one or more disks in a Microsoft Cluster Service (MSCS) cluster group, also known as a resource group. Each resource group can contain at most one instance of SQL Server.
    A network name for the failover cluster instance.
    One or more IP addresses assigned to the failover cluster instance.
    One instance of SQL Server that includes SQL Server, SQL Server Agent, the Full-text Search (FTS) service, and Replication. You can install a failover cluster with SQL Server only, Analysis Services
    only, or SQL Server and Analysis Services
    http://msdn.microsoft.com/en-in/library/ms179410(v=sql.105).aspx

  • 3 Node Failover Cluster With iSCSI

    Is there any information available on the steps to create a 3 node failover cluster with iSCSI storage?  Is there a step-by-step guide?  I looked around but couldn't find much.  Thanks!

    Hi SCPSTech,
    The 3 node cluster create steps same with 2 node, you can refer the following step by step white paper create the 2 node cluster then add the another node to cluster.
     Configuring Failover Clusters with Windows Storage Server 2008
    http://blogs.technet.com/b/storageserver/archive/2009/12/17/configuring-failover-clusters-with-windows-storage-server-2008.aspx
    Add a Server to a Failover Cluster
    http://technet.microsoft.com/en-us/library/cc730998.aspx
    More information:
    Add or Remove Nodes in a SQL Server Failover Cluster (Setup)
    http://msdn.microsoft.com/en-us/library/ms191545.aspx
    Hope this helps.
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Creating 4 Node SQL Cluster

    Hi
    I need to create a 4 Node - SQL 2012 Cluster (all in active Mode). 
    Node-1 will be having 1 Instance, Node-2 will be having 2 Instance, Node-3 will be running with 1 instance, and Node-4 will be running with 6 SQL Instance.
    Let me know whether can I install all the 10 Instance on Node-1 and then can i go ahead and add the  node-2 by selecting 10 Instances which is already installed and then node-3 and node-4.
    or Let me know if any other recommended way of Setting up the Cluster.
    Raj

    Hi,
    In the Server 2012 Failover cluster the node maximum is 64, the SQL server also have the maximum node limit, such as SQL 2008 clustering depends on which version of the software
    you purchase, along with which version of the operating system you intend to use. Therefore for the rigid plan I suggest you ask the issue in SQL forum.
    More information:
    SQL Server forum:
    http://social.technet.microsoft.com/Forums/sqlserver/en-us/home?category=sqlserver
    Before Installing Failover Clustering
    http://msdn.microsoft.com/en-us/library/ms189910.aspx
    Thanks for your understanding and support.
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • [SOLVED] Can't add a node to the cluster with error (Exchange 2010 SP3 DAG Windows Server 2012)

    Hi there!
    I have a problem which makes me very angry already :)
    I have two servers Exchange 2010 SP3 with MB role started on Windows Server 2012. I decided to create a DAG.
    I have created the prestaged AD object for the cluster called msc-co-exc-01c, assigned necessary permissions and disabled it. Allowed through the Windows Firewall traffic between nodes and prepared the File Share Witness server.
    Then I have tried to add nodes. The first node has been added successfully, but the second node doesn't want to be added :). Now I can add only one node to the DAG. I tried to add different servers first, but only the first one was added.
    LOGS on the second nodes: 
    Application Log
    "Failed to initialize cluster with error 0x80004005." (MSExchangeIS)
    Failover Clustering Diagnostic Log
    "[VER] Could not read version data from database for node msc-co-exc-04v (id 1)."
    CMDLET Error:
    Summary: 1 item(s). 0 succeeded, 1 failed.
    Elapsed time: 00:06:21
    MSC-CO-EXC-02V
    Failed
    Error:
    A database availability group administrative operation failed. Error: The operation failed. CreateCluster errors may result from incorrectly configured static addresses. Error: An error occurred while attempting a cluster operation. Error: Cluster API '"AddClusterNode()
    (MaxPercentage=100) failed with 0x5b4. Error: This operation returned because the timeout period expired"' failed. [Server: msc-co-exc-04v.int.krls.ru]
    An Active Manager operation failed. Error An error occurred while attempting a cluster operation. Error: Cluster API '"AddClusterNode() (MaxPercentage=100) failed with 0x5b4. Error: This operation returned because the timeout period expired"' failed..
    This operation returned because the timeout period expired
    Click here for help... http://technet.microsoft.com/en-US/library/ms.exch.err.default(EXCHG.141).aspx?v=14.3.174.1&t=exchgf1&e=ms.exch.err.ExC9C315
    Warning:
    Network name 'msc-co-exc-01c' is not online. Please check that the IP address configuration for the database availability group is correct.
    Warning:
    The operation wasn't successful because an error was encountered. You may find more details in log file "C:\ExchangeSetupLogs\DagTasks\dagtask_2014-11-17_13-54-56.543_add-databaseavailabiltygroupserver.log".
    Exchange Management Shell command attempted:
    Add-DatabaseAvailabilityGroupServer -MailboxServer 'MSC-CO-EXC-02V' -Identity 'msc-co-exc-01c'
    Elapsed Time: 00:06:21
    UPD:
    when Exchange servers ran on the same Hyper-V node, the DAG is working well, but if I move one of VM to another node, It stops working.
    I have installed Wireshark and captured trafic of cluster interface. When DAG members on the same HV-node, there is inbound and outbound traffic on the cluster interface, but if I move one of DAG member to another node, in Wireshark I see only outbound traffic
    on both nodes.
    It's confused me, because there is normal connectivity between these DAG members through the main interface.
    Please, help me if you can.

    Hi, Jared! Thank you for the reply.
    Of course I did it already :) I have new info:
    when Exchange servers ran on the same Hyper-V node, the DAG is working well, but if I move one of VM to another node, It stops working.
    I have installed Wireshark and captured trafic of cluster interface. When DAG members on the same HV-node, there is inbound and outbound traffic on the cluster interface, but if I move one of DAG member to another node, in Wireshark I see only outbound traffic
    on both nodes.
    It's confused me, because there is normal connectivity between these DAG members through the main interface.

  • Can I create different Coherence nodes in the same cluster with defferent?

    Can I create different Coherence nodes in the same cluster with defferent cache-config.xml file ?
    Can a cache be distributed in these deffirent nodes?

    Yes. You can create different Coherence nodes in the same cluster with defferent cache-config.xml files as long as you use the same tangosol-coherence.xml file and the same tangosol-coherence-override.xml file. But you cannot store the cache data in the different nodes (started with different cache-config file). In other word, a node only create caches in their own's modes which are started with the same cache-config.xml file.
    See the following demo:
    I start a cache server using the cache config file examples-cache-server.xml. Then I start a storage-disabled cache console (cache client) using the cache config file coherence-cache-config.xml. Both of them using the same tangosol-coherence.xml file and the same tangosol-coherence-override.xml file.
    The cache server uses a cache service PartitionedPofCache. But the client side is using the Distributedcache service. The cluster address is same 224.3.5.2.
    The cluster name is also samme. They know each other.
    D:\coherence\lib>D:\examples\java\bin\run-cache-server.cmd
    D:\coherence\lib>D:\examples\java\bin\run-cache-server.cmd
    The system cannot find the file D:\coherence.
    The system cannot find the file C:\Oracle\Middleware\jdk160_11.
    2009-12-22 12:09:31.400/4.987 Oracle Coherence 3.5.2/463 <Info> (thread=main, member=n/a): Loaded operational configurat
    ion from resource "jar:file:/D:/coherence/lib/coherence.jar!/tangosol-coherence.xml"
    2009-12-22 12:09:31.450/5.037 Oracle Coherence 3.5.2/463 <Info> (thread=main, member=n/a): Loaded operational overrides
    from resource "jar:file:/D:/coherence/lib/coherence.jar!/tangosol-coherence-override-dev.xml"
    2009-12-22 12:09:31.470/5.057 Oracle Coherence 3.5.2/463 <D5> (thread=main, member=n/a): Optional configuration override
    "/tangosol-coherence-override.xml" is not specified
    2009-12-22 12:09:31.540/5.127 Oracle Coherence 3.5.2/463 <D5> (thread=main, member=n/a): Optional configuration override
    "/custom-mbeans.xml" is not specified
    Oracle Coherence Version 3.5.2/463
    Grid Edition: Development mode
    Copyright (c) 2000, 2009, Oracle and/or its affiliates. All rights reserved.
    2009-12-22 12:09:33.864/7.451 Oracle Coherence GE 3.5.2/463 <Info> (thread=main, member=n/a): Loaded cache configuration
    from "file:/D:/examples/java/resource/config/examples-cache-config.xml"
    2009-12-22 12:09:39.983/13.570 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=n/a): Service Cluster joined t
    he cluster with senior service member n/a
    2009-12-22 12:09:43.187/16.774 Oracle Coherence GE 3.5.2/463 <Info> (thread=Cluster, member=n/a): Created a new cluster
    "cluster:0xD3FB" with Member(Id=1, Timestamp=2009-12-22 12:09:38.06, Address=192.168.8.80:8088, MachineId=24656, Locatio
    n=process:144, Role=CoherenceServer, Edition=Grid Edition, Mode=Development, CpuCount=1, SocketCount=1) UID=0xC0A8085000
    000125B75D888C60501F98
    2009-12-22 12:09:43.508/17.095 Oracle Coherence GE 3.5.2/463 <D5> (thread=Invocation:Management, member=1): Service Mana
    gement joined the cluster with senior service member 1
    2009-12-22 12:09:46.582/20.169 Oracle Coherence GE 3.5.2/463 <D5> (thread=DistributedCache:PartitionedPofCache, member=1
    ): Service PartitionedPofCache joined the cluster with senior service member 1
    2009-12-22 12:09:46.672/20.259 Oracle Coherence GE 3.5.2/463 <Info> (thread=DistributedCache:PartitionedPofCache, member
    =1): Loading POF configuration from resource "file:/D:/examples/java/resource/config/examples-pof-config.xml"
    2009-12-22 12:09:46.702/20.289 Oracle Coherence GE 3.5.2/463 <Info> (thread=DistributedCache:PartitionedPofCache, member
    =1): Loading POF configuration from resource "jar:file:/D:/coherence/lib/coherence.jar!/coherence-pof-config.xml"
    2009-12-22 12:09:47.734/21.321 Oracle Coherence GE 3.5.2/463 <Info> (thread=main, member=1): Started DefaultCacheServer.
    SafeCluster: Name=cluster:0xD3FB
    Group{Address=224.3.5.2, Port=35463, TTL=4}
    MasterMemberSet
      ThisMember=Member(Id=1, Timestamp=2009-12-22 12:09:38.06, Address=192.168.8.80:8088, MachineId=24656, Location=process
    :144, Role=CoherenceServer)
      OldestMember=Member(Id=1, Timestamp=2009-12-22 12:09:38.06, Address=192.168.8.80:8088, MachineId=24656, Location=proce
    ss:144, Role=CoherenceServer)
      ActualMemberSet=MemberSet(Size=1, BitSetCount=2
        Member(Id=1, Timestamp=2009-12-22 12:09:38.06, Address=192.168.8.80:8088, MachineId=24656, Location=process:144, Rol
    e=CoherenceServer)
      RecycleMillis=120000
      RecycleSet=MemberSet(Size=0, BitSetCount=0
    Services
      TcpRing{TcpSocketAccepter{State=STATE_OPEN, ServerSocket=192.168.8.80:8088}, Connections=[]}
      ClusterService{Name=Cluster, State=(SERVICE_STARTED, STATE_JOINED), Id=0, Version=3.5, OldestMemberId=1}
      InvocationService{Name=Management, State=(SERVICE_STARTED), Id=1, Version=3.1, OldestMemberId=1}
      DistributedCache{Name=PartitionedPofCache, State=(SERVICE_STARTED), LocalStorage=enabled, PartitionCount=257, BackupCo
    unt=1, AssignedPartitions=257, BackupPartitions=0}
    2009-12-22 12:12:29.737/183.324 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=1): Member(Id=2, Timestamp=20
    09-12-22 12:12:29.541, Address=192.168.8.80:8089, MachineId=24656, Location=process:1188, Role=CoherenceConsole) joined
    Cluster with senior member 1
    2009-12-22 12:12:30.498/184.085 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=1): Member 2 joined Service M
    anagement with senior member 1
    2009-12-22 12:12:31.860/185.447 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=1): TcpRing: connecting to me
    mber 2 using TcpSocket{State=STATE_OPEN, Socket=Socket[addr=/192.168.8.80,port=8089,localport=2463]}
    2009-12-22 12:12:51.338/204.925 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=1): Member 2 joined Service D
    istributedCache with senior member 2The following command starts a cache client.
    D:\coherence\bin>coherence.cmd
    D:\coherence\bin>coherence.cmd
    ** Starting storage disabled console **
    java version "1.6.0_11"
    Java(TM) SE Runtime Environment (build 1.6.0_11-b03)
    Java HotSpot(TM) Server VM (build 11.0-b16, mixed mode)
    2009-12-22 12:12:21.054/3.425 Oracle Coherence 3.5.2/463 <Info> (thread=main, member=n/a): Loaded operational configurat
    ion from resource "jar:file:/D:/coherence/lib/coherence.jar!/tangosol-coherence.xml"
    2009-12-22 12:12:21.355/3.726 Oracle Coherence 3.5.2/463 <Info> (thread=main, member=n/a): Loaded operational overrides
    from resource "jar:file:/D:/coherence/lib/coherence.jar!/tangosol-coherence-override-dev.xml"
    2009-12-22 12:12:21.365/3.736 Oracle Coherence 3.5.2/463 <D5> (thread=main, member=n/a): Optional configuration override
    "/tangosol-coherence-override.xml" is not specified
    2009-12-22 12:12:21.415/3.786 Oracle Coherence 3.5.2/463 <D5> (thread=main, member=n/a): Optional configuration override
    "/custom-mbeans.xml" is not specified
    Oracle Coherence Version 3.5.2/463
    Grid Edition: Development mode
    Copyright (c) 2000, 2009, Oracle and/or its affiliates. All rights reserved.
    2009-12-22 12:12:29.316/11.687 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=n/a): Service Cluster joined t
    he cluster with senior service member n/a
    2009-12-22 12:12:29.356/11.727 Oracle Coherence GE 3.5.2/463 <Info> (thread=Cluster, member=n/a): Failed to satisfy the
    variance: allowed=16, actual=20
    2009-12-22 12:12:29.356/11.727 Oracle Coherence GE 3.5.2/463 <Info> (thread=Cluster, member=n/a): Increasing allowable v
    ariance to 17
    2009-12-22 12:12:29.807/12.178 Oracle Coherence GE 3.5.2/463 <Info> (thread=Cluster, member=n/a): This Member(Id=2, Time
    stamp=2009-12-22 12:12:29.541, Address=192.168.8.80:8089, MachineId=24656, Location=process:1188, Role=CoherenceConsole,
    Edition=Grid Edition, Mode=Development, CpuCount=1, SocketCount=1) joined cluster "cluster:0xD3FB" with senior Member(I
    d=1, Timestamp=2009-12-22 12:09:38.06, Address=192.168.8.80:8088, MachineId=24656, Location=process:144, Role=CoherenceS
    erver, Edition=Grid Edition, Mode=Development, CpuCount=1, SocketCount=1)
    2009-12-22 12:12:29.977/12.348 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=n/a): Member 1 joined Service
    Management with senior member 1
    2009-12-22 12:12:29.977/12.348 Oracle Coherence GE 3.5.2/463 <D5> (thread=Cluster, member=n/a): Member 1 joined Service
    PartitionedPofCache with senior member 1
    2009-12-22 12:12:30.578/12.949 Oracle Coherence GE 3.5.2/463 <D5> (thread=Invocation:Management, member=2): Service Mana
    gement joined the cluster with senior service member 1
    SafeCluster: Name=cluster:0xD3FB
    Group{Address=224.3.5.2, Port=35463, TTL=4}
    MasterMemberSet
      ThisMember=Member(Id=2, Timestamp=2009-12-22 12:12:29.541, Address=192.168.8.80:8089, MachineId=24656, Location=proces
    s:1188, Role=CoherenceConsole)
      OldestMember=Member(Id=1, Timestamp=2009-12-22 12:09:38.06, Address=192.168.8.80:8088, MachineId=24656, Location=proce
    ss:144, Role=CoherenceServer)
      ActualMemberSet=MemberSet(Size=2, BitSetCount=2
        Member(Id=1, Timestamp=2009-12-22 12:09:38.06, Address=192.168.8.80:8088, MachineId=24656, Location=process:144, Rol
    e=CoherenceServer)
        Member(Id=2, Timestamp=2009-12-22 12:12:29.541, Address=192.168.8.80:8089, MachineId=24656, Location=process:1188, R
    ole=CoherenceConsole)
      RecycleMillis=120000
      RecycleSet=MemberSet(Size=0, BitSetCount=0
    Services
      TcpRing{TcpSocketAccepter{State=STATE_OPEN, ServerSocket=192.168.8.80:8089}, Connections=[]}
      ClusterService{Name=Cluster, State=(SERVICE_STARTED, STATE_JOINED), Id=0, Version=3.5, OldestMemberId=1}
      InvocationService{Name=Management, State=(SERVICE_STARTED), Id=1, Version=3.1, OldestMemberId=1}
    Map (?):
    2009-12-22 12:12:49.505/31.906 Oracle Coherence GE 3.5.2/463 <Info> (thread=main, member=2): Loaded cache configuration
    from "jar:file:/D:/coherence/lib/coherence.jar!/coherence-cache-config.xml"
    2009-12-22 12:12:51.358/33.729 Oracle Coherence GE 3.5.2/463 <D5> (thread=DistributedCache, member=2): Service Distribut
    edCache joined the cluster with senior service member 2
    <distributed-scheme>
      <!--
      To use POF serialization for this partitioned service,
      uncomment the following section
      <serializer>
      <class-
      name>com.tangosol.io.pof.ConfigurablePofContext</class-
      name>
      </serializer>
      -->
      <scheme-name>example-distributed</scheme-name>
      <service-name>DistributedCache</service-name>
      <backing-map-scheme>
        <local-scheme>
          <scheme-ref>example-binary-backing-map</scheme-ref>
        </local-scheme>
      </backing-map-scheme>
      <autostart>true</autostart>
    </distributed-scheme>But when I try to store data into cache from the client side, it report error message: it's staorage-disabled. It shows that this cache console cannot store the data in the existing cache server because then using different cache config files.
    Map (ca3): cache ca2
    <distributed-scheme>
      <!--
      To use POF serialization for this partitioned service,
      uncomment the following section
      <serializer>
      <class-
      name>com.tangosol.io.pof.ConfigurablePofContext</class-
      name>
      </serializer>
      -->
      <scheme-name>example-distributed</scheme-name>
      <service-name>DistributedCache</service-name>
      <backing-map-scheme>
        <local-scheme>
          <scheme-ref>example-binary-backing-map</scheme-ref>
        </local-scheme>
      </backing-map-scheme>
      <autostart>true</autostart>
    </distributed-scheme>
    Map (ca2): put 1 one
    2009-12-22 14:00:04.999/6467.370 Oracle Coherence GE 3.5.2/463 <Error> (thread=main, member=2):
    java.lang.RuntimeException: Storage is not configured
            at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache$BinaryMap.onMissing
    Storage(DistributedCache.CDB:9)
            at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache$BinaryMap.ensureReq
    uestTarget(DistributedCache.CDB:34)
            at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache$BinaryMap.put(Distr
    ibutedCache.CDB:22)
            at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache$BinaryMap.put(Distr
    ibutedCache.CDB:1)
            at com.tangosol.util.ConverterCollections$ConverterMap.put(ConverterCollections.java:1541)
            at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache$ViewMap.put(Distrib
    utedCache.CDB:1)
            at com.tangosol.coherence.component.util.SafeNamedCache.put(SafeNamedCache.CDB:1)
            at com.tangosol.coherence.component.application.console.Coherence.processCommand(Coherence.CDB:581)
            at com.tangosol.coherence.component.application.console.Coherence.run(Coherence.CDB:39)
            at com.tangosol.coherence.component.application.console.Coherence.main(Coherence.CDB:3)
            at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
            at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
            at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
            at java.lang.reflect.Method.invoke(Method.java:597)
            at com.tangosol.net.CacheFactory.main(CacheFactory.java:1400)

  • How to make 2 Node File Cluster with SAS Disks

    Hello,
    I cant fine any detailed, specific information and answer on this question
    I have 2 Servers, Say HP, with 4 SAS disks each,  1 for OS (2012R2Data) and 3 available.
    I want to create fault tolerant SMB 3.0 share for my Hyper-v nodes, to make hyper-v FT cluster afterwards.
    So, I see inr equirements that SAS disks will work, but during creating cluster, I can not manage Cluster to see disks from both servers (total 6 free units)
    If it is made via ISCSi target, both servers have access to that targets and cluster sees all available through iSCSI disks, but how to mnake SAS disk available to other server?
    Is it possible what I need with this configuration?  is it a option to make each server iSCSI target and initiator? (but it would be way complicated and slow, I think).
    so, what I have misunderstood, how to make FT file Cluster with 2 servers with SAS drives?
    Any info about this SPECIFIC config would be welcomed, or any general step-by-step guides  (please, do not link me guides with other config, like iscsi or with additional servers, I have seen a lot of them :()
    thanks

    Hello,
    I cant fine any detailed, specific information and answer on this question
    I have 2 Servers, Say HP, with 4 SAS disks each,  1 for OS (2012R2Data) and 3 available.
    I want to create fault tolerant SMB 3.0 share for my Hyper-v nodes, to make hyper-v FT cluster afterwards.
    So, I see inr equirements that SAS disks will work, but during creating cluster, I can not manage Cluster to see disks from both servers (total 6 free units)
    If it is made via ISCSi target, both servers have access to that targets and cluster sees all available through iSCSI disks, but how to mnake SAS disk available to other server?
    Is it possible what I need with this configuration?  is it a option to make each server iSCSI target and initiator? (but it would be way complicated and slow, I think).
    so, what I have misunderstood, how to make FT file Cluster with 2 servers with SAS drives?
    Any info about this SPECIFIC config would be welcomed, or any general step-by-step guides  (please, do not link me guides with other config, like iscsi or with additional servers, I have seen a lot of them :()
    thanks
    You cannot do what you want with a Microsoft built-in tools. MSFT require you to take your SAS disks away from your servers, buy a SAS JBOD (better more then one to get advantage of a so-called "enclosure awareness" and avoid single point of failure) and
    configure Clustered Storage Spaces with now *external* SAS drives. See:
    How to Configure a Clustered Storage Space in Windows Server 2012
    http://blogs.msdn.com/b/clustering/archive/2012/06/02/10314262.aspx
    Prerequisites
             A minimum of three physical drives, with at least 4 gigabytes (GB) capacity each, are required to create a storage
    pool in a Failover Cluster.
             The clustered storage pool MUST be comprised
    of Serial Attached SCSI (SAS) connected physical disks. Layering any form of storage subsystem, whether an internal RAID card or an external RAID box, regardless of being directly connected or connected via a storage fabric, is not supported.
    Windows Server 2012 R2 had flexed out some limitations so now ReFS is supported (but useless as VMs cannot be integrity checked and protected) and now you can use parirty spaces (also useless as they are DOG slow with a typical VM workload dominated by small
    writes, say 4KB write initiates 256KB+ parity stripe read-modify-write update). Core requirements "SAS everywhere" is still there.
    So if you don't want to mess with SAS JBODs you may give a try to a virtual SAN solutions available on the market. They can cluster a pair of hosts (even with a single port SATA drives) w/o any external hardware, only Ethernet required. There are even free
    options available. See:
    Free Virtual SAN
    http://www.starwindsoftware.com/native-san-for-hyper-v-free-edition
    Also Google (or Bing?) for DataCore and SteelEye as they have very similar native (Windows-based) offerings. + there's bunch of a VM-running storage doing more or less the same.
    Hope this helped a bit :)
    StarWind VSAN [Virtual SAN] clusters Hyper-V without SAS, Fibre Channel, SMB 3.0 or iSCSI, uses Ethernet to mirror internally mounted SATA disks between hosts.

  • Sun Cluster with three nodes

    I need a manual or advices for introducing a third node in a RAC with Sun Cluster. I don't know if qourum votes readjust automatic or I have add new quorum votes manualy, if I have to add a thirn mediator in svm ... etc
    A lot of thanks and sorry for my english

    After you have added your nodes to the cluster you will need to expand the RGs node list to include the new nodes if you need the RG to run on them. This is not automatic. Something like:
    # clrg set -n <nodelist> <rg_name>
    Is what you need.
    I'm not sure I understand what you said about the quorum count. Only nodes and quorum devices (QD) or quorum servers (QS) get a vote, cabinets do not. So each node gets a vote and a QD/QS gets a vote count equal to the number of nodes it connects to minus 1. Thus with a two node cluster, you have 3 votes with one QD. With a 4 node cluster with one fully connected QD/QS, you have 7 votes (after re-adding it).
    Hope that helps,
    Tim
    P.S. <shameless plug> I can recommend a good book on the product: "Oracle Solaris Cluster Essentials" ;-)

  • SUn Cluster with multiple nodes

    Hi,
    I am planning to setup Sun Cluster with 3+ nodes.
    is it possible to configure sun cluster with just two network interfaces on each box. (One for Public access and second for cluster interconnection via dedicated switch)?
    Thanks in advance for the help.

    Yes, but only if the adapters support VLAN tagging.
    That way you would be able to create two tagged VLANs on each adapter, one for public and one for private traffic. You'd also need a switch that supported VLAN tagging, but I think these are fairly common now.
    If your NICs do not support tagging then you'd need to get another adapter to allow Sun Cluster to install.
    Regards,
    Tim
    ---

  • Do someone knows how to create a cluster with two nodes on oracle VM Ware

    dear members,
    can someone guide us through creating a cluster with two nodes on oracle VM. it will be nice if we can virtualiza RAC, atleast for education purposes.
    Thank you all in advance

    Here is an excellent article that explains how to do this:
    http://startoracle.com/2007/09/30/so-you-want-to-play-with-oracle-11gs-rac-heres-how/

  • Migrate data from a single node cluster to a single node cluster with CDOT 8.3

    Hi, I have two FAS2240-4 in the office, one for production and the second one intended to use it for DR, the FAS with 18 disk of 2TB each that is in production has 3 disk for the root aggregate and the rest for data, so I need to wipe the system in order to take advantage of the HDD partitioning for the entry platforms in CDOT 8.3. The Second system is fresh installed with CDOT 8.3 with the Root-Data HDD partitioning so now I want to move the Data and configuration to replace the first cluster with the second one.Some body gave me a procedure to backup the configuration and restore it at the destination cluster but I believe that is not complete. After finish the procedure I have all duplicated and not working properly, even with missed data, so the procedure is: ---------------------------------------------------
    SYSTEM CONFIGRUATION BACKUPS FOR CLUSTERED ONTAPSOURCE SYSTEM
    set -priv diagnostic
    cluster identiy show
    system configuration backup create -backup-name myBackup.7z -node node1 -backup-type cluster (Create a node backup of node1)
    system configuration backup upload -node node1 -backup myBackup.7z -destination ftp://[email protected] (Uploads a backup file to ftp)DESTINATION SYSTEM
    ## This would be done after setting up one node on the dr site, but prior to confguring the rest of the nodes
    ## you would basically stand up the system with ips and a base config, and load the config from the primary site
    ## and once it comes back up, you would setup & join the other nodes.
    set -priv diagnostic
    cluster identity show
    system configuration backup download -node node1 -backup myBackup.7z -destination ftp://[email protected]
    system configuration recovery cluster recreate -from backup -backup myBackup.7z
    system node reboot -node node1
    cluster identity show
    cluster identity modify -name newcluster
    cluster identity show
    system configutration recovery cluster modify -recovery-status complete----------------------------------------------------After doing this then I have problems with the folowing:
    - Volumes without properties
    - Network interfaces down and mixed with the old and new system
    - Aggregates with no info (This table is currently empty)- The node name is not able to rename it because it says that is already in use. I believe that I need more information in order to complete the procedure well.  I hope that you can help me with this. Best Regards, Miguel Valls

    In Non-Exa data, I followed below steps, I am sure same will work in exadata.
    Standby database creation using L0 backup restore
         While RMAN backup command on non-asm doesn't require change but restore in ASM requires you to add following kind of lines in restore script
              set newname for datafile 10 to '+Data02';
              set newname for tempfile 25 to '+Data10';
          After this you need to add
               restore database;
               switch datafile all;
               switch tempfile all;
    Standby database sync up using SCN based incremental backup
    Open Standby database in READ WRITE mode
    Convert DB from standalone to RAC
    Application doesn't need to be certified for Exadata, If your application runs on Oracle 11203 on RAC+ASM on OEL then it will run on Exadata as well.
    One important thing to remember is that, using physical migration method like this you won't be able to get EHCC (hybrid columnar compression) benefit unless you do the table move after the migration so most of the time logical methods like- data pump is preferred over physical method. In Exadata, these methods run very very fast. It might be less effort and less downtime compare to Physical standby method so give it a try.
    Thanks,
    Abhi

  • SCOM 2012 Reporting Components with a shared SQL cluster

    Hi, we are planning to install SCOM 2012 but i am a bit confused around the reporting components.
    We have a shared SQL Enterprise cluster with multiple instances which we would like to use for our OperationsMananger and OperationsManagerDW databases. However after trying a test SCOM deployment, installing the SCOM reporting components on a VM requires
    a local running copy of SQL reporting services.
    Is this true? It seems ludicrous to have to purchase an additional license of SQL just to run the reporting components of SCOM. We have spent a lot of money on our SQL cluster and we would like to use it. We dont want to install the SCOM reporting components
    on the shared SQL cluster.
    If this is true, can you tell me which license of SQL we require? Just to run SQL reporting services on a VM should just require a standard SQL license, right?
    Also if we have to purchase a SQL license for this standalone VM, would you recommend having the OperationsManagerDW database on the shared SQL cluster or on the VM where the standalone instance of SQL reporting services resides.
    Surely this is something that other people have an issue with? I would have thought most people would want to use a shared SQL cluster. Any help is much appreciated.

    Just to add that SQL Reporting Services isn't cluster aware so the best you could do is to install stand-alone onto a node of the cluster.
    To reinforce what Scott is saying, see this screenshot of the install of SCOM reporting:
    http://www.systemcentersolutions.co.uk/images/easyblog_images/42/scom/CTP2/Install/SQRS9.png
    Technically you can install SCOM reporting on a shared SQL Server Reporting Services instance but as Scott (and the screenshot) state - it will overwrite any existing security settings on existing reports.
    Scott is also correct that SQL Server Standard Edition is included as part of the licensing cost of System Center 2012 as long as the SQL instance only runs components of System Center.
    cheers
    Graham
    Regards Graham New System Center 2012 Blog! -
    http://www.systemcentersolutions.co.uk
    View OpsMgr tips and tricks at
    http://systemcentersolutions.wordpress.com/
    I have one SQL 2012 Server with a few instances.   1 for Scom, 1 for SCCM and 1 for Service Manager.  The instance shows up as blank for reporting services on the SCOM.   Do I have to install SQL on the SCOM box also?  Why Can't
    it see the instance on my 2012 SQL box?  It is in a different instance than SCCM.

  • SQL Server 2008R2 Active-Active cluster with standard edition?

    Hi all,
    can I have a 2 node cluster (Windows Server 2008R2 Datacenter Edition: node_1 and node_2) forming a failover cluster with 2 instances of SQL Server 2008R2 Standard Edition (instance_A and instance_B) with:
    instance_A running on node_1
    instance_B running on node_2

    Following thread is on similar topic:
    http://social.msdn.microsoft.com/Forums/sqlserver/en-US/450dfb42-635b-4c99-ae9f-6b0a23e3430b/sql-server-2008-r2-standard-edition-failover-cluster?forum=sqldatabaseengine
    Kalman Toth Database & OLAP Architect
    SQL Server 2014 Design & Programming
    New Book / Kindle: Exam 70-461 Bootcamp: Querying Microsoft SQL Server 2012

  • SharePoint Multi-Server Farm with LB and SQL Cluster.

    Helllo,
    We are setting up a Multi Server SharePoint 2013 Farm with a Load-Balancer. 
    We are also using SQL Server 2012, Two Node Cluster. 
    We have SQL Server Instance Running as a Service. Also I managed to install PowerPivot for SharePoint as a Instance. 
    Looking at various article's I have question if this Cluster Setup supported in SharePoint 2013. 
    Can the SharePoint Farm use this PowerPivot Instance as a failover service? 
    Or do I have to install Analysis services separately on Both SQL Nodes? 
    If so how do I configure Excel Services on SharePoint? 
    Thank you
    Sham

    PowerPivot is installed on the SharePoint server and is a scale-out service (you just install it on multiple SharePoint servers and make sure the Service Instance is started in SharePoint). Same with Excel Calc Services (but for this one you don't have
    to install anything, just start it).
    Trevor Seward
    Follow or contact me at...
    &nbsp&nbsp
    This post is my own opinion and does not necessarily reflect the opinion or view of Microsoft, its employees, or other MVPs.

Maybe you are looking for

  • My computer no longer has Stereo Mix / Wave Out Mix.

    I bought an HP G60-443cl notebook computer about less than two years ago. It had Vista installed on it. I was able to record direct audio using Hypercam, FRAPS, Audacity, etc...... Then I reinstalled Vista and now the only sound I'm able to record is

  • Can't choose cache folder in Bridge CS6 on a different drive/partition

    Hi, I just installed Photoshop CS6 and noticed a non-critical, but nevertheless annoying problem in Bridge (Platform: Win7 x64): When I try to relocate the cache folder (Menu: Edit -> Preferences -> Cache -> Location, then pressing the button "Choose

  • App got stuck while installing help

    I was trying to download an app and it got stuck while loading. I can't delete it. I have turned off and on my phone to no avail. I tried plugging in the phone to itunes. Help please

  • Margins are not the same when I print

    Help! I have written 80 pages and when I go to print, the pages have a ridiculous amount of margin space on the top, bottom and left sides of the paper. When I'm writing, everything looks right. It is once I get to Print Preview, I can see how off it

  • Do I need to call AS3_Release

    If I create an object in C++ code and return it in ActionScript code should I call AS3_Release before returning it? For example, I have the function in the *.gg file: public function makeThumbnail(...): Object      AS3_Val objDestByteArray = AS3_New(