Failover Cluster Core Resources question on a Windows 2008R2 three node cluster

We have a three node Windows 2008R2 cluster with SQL Server 2008 R2 as a clustered resource. There are three resource groups in this cluster 1) Available Storage 2) Cluster Group 3) SQL Server.  The Available Storage and SQL Server resource groups
reside on one node while the Cluster Group resides on another.  The only resources residing in the Cluster Resource Group is the Cluster name and IP.  I'd like to failover the Cluster Resource Group to be on the same node as everything else. 
I'm not sure what the implications are on doing this.  Failing over the Cluster Group shouldn't have any impact on the SQL Server Resource Group correct or would there be an interruption to SQL because of the failover of the Cluster Group.  It's
an critical application of which I'm trying to gather some information for a change request and I know I'm going to be asked if this impacts the production database and everybody using it.
Thanks
RG

No, that should not impact anything.  The cluster group is completely separate from the SQL group.
. : | : . : | : . tim

Similar Messages

  • Regarding Three node cluster

    Hi,
    Can i know how the High Availability feature works foe three node cluster.I configured two node cluster and it works fine, if one goes down the other runs the application. What in case of three node or higher topology cluters.
    Is the High availability feature confined to two node cluster , if not how it works for higher topology clusters.
    Can anyone help me on this.

    This assumes you use the agent builder to do this. If that is the case, you will need to put each of the generated packages on all of the nodes of the cluster. However, it's probably far easier to just create the resources from the command line. You only actually need to specify a start method, e.g
    # clrg create my-rg
    # clrs create -t SUNW.gds -g my-rg -p network_aware=false -p Start_command=/var/tmp/gds-start my-gds-rs
    where /var/tmp/gds-start in my case contains:
    #!/bin/ksh
    sleep 100000 &
    but in your case would contain the start commands for you application - which leaves some sort of process tree in place. If the process in the tree all fail, then the service would be restarted.
    Have a look at the other resources on the web about writing GDS services.
    Tim
    ---

  • Windows 2008r2 cluster dns issue

    We are running GW 8.0.2 not hot patches. System is running on a windows 2008r2 2 node cluster. When we failover the gwia from node 1 to node 2 we can no longer send email to the internet. We receive just fine, also imap and pop connections are working. We have checked firewall settings and they are the same for both nodes. We get 450 hosts unknown errors for all email going out. We think this is a dns issue, but can't find where the problem may be. We have checked the registry dns nameserver entries on both nodes and they match.
    Thanks,
    Bill

    On the GWIA, is the bind exclusive set on the addressing tab?
    The GWIA will send on the primary address of the box unless this box
    is checked - but it will listen on all IP addresses (this inbound will
    work but recieve will not)
    T
    On Fri, 01 Apr 2011 12:38:04 GMT, [email protected]: wrote:
    >We are running GW 8.0.2 not hot patches. System is running on a windows 2008r2 2 node cluster. When we failover the gwia from node 1 to node 2 we can no longer send email to the internet. We receive just fine, also imap and pop connections are working. We have checked firewall settings and they are the same for both nodes. We get 450 hosts unknown errors for all email going out. We think this is a dns issue, but can't find where the problem may be. We have checked the registry dns nameserver entries on both nodes and they match.
    >
    >Thanks,
    >Bill

  • Windows 2008R2 Failover Cluster node

    I am new to this company and have been asked to take over a task that the previous employee left unfinished.   I am "familiar" with Failover clustering but have not done any recently.   It looks like the previous employee evicted one
    of the 2 node cluster servers to replace the hardware.   I've setup the new hardware with Failover cluster but now need to setup the iSCSI volumes on this new node.   I have not done this before.  All I've done before was create NEW failover
    disks and those were Fibre Channel disks.  Can anyone give me advice or a link to a tutorial to setup these existing iSCSI volumes on the second node?  One more note:   the iSCSI volumes are on a Dell EqualLogic system and I've never worked
    on EqualLogic before either.

    Hi,
    If I doesn’t misunderstand it seems you are finding how to setup the fiber channel storage to the node, you can refer your hardware manual to setup the hardware connection,
    or call your hardware vendor for the more help, you can refer the following support scenario to configure your fiber channel storage:
    The Cluster service allows you to reconnect or replace a cable between the HBA and the switch, and then allows for the node to take ownership of the physical disk resource.
    The HBA driver must consider several issues for a Plug and Play rescan to occur. The HBA miniport driver must issue a "BusChangeDetected" notification when the cable is reconnected (an HBA port driver issues an "IoInvalidateDeviceRelations" notification) so
    that Windows is notified that a change has been made to the shared bus.
    The related KB:
    Removing the HBA cable on a server cluster
    http://support.microsoft.com/kb/294173/en-us
    Hope this helps.
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Which account is used to failover to passive node of Windows 2008 R2 2-node cluster?

    When I set up drive letter for shared drives in Windows 2008 R2 2-node cluster, I accidentally used a drive letter which has already mapped automatically due to the login I used.   After I run the cluster validation tests, all the drive letters
    for shared drives are changed.   I fixed this issue by disconnecting the automatic mapped drive and reassigned all the drive letters.  
    However, I wonder whether CNO is used to perform the failover so such issue will not occur actually.  Also, what happen if someone is logged on the passive node when the failover occur and any of the drive letters assigned for the shared drives
    is already mapped due to that person's login. 

    Hi Sebastian,
    Let me clarify my question.  For example, there are clustered drives W:, M:, and L:.   Active Directory Group Policy has policy for some users to have mapped drive M:.
    What will happen if failover occurs when one of the users having that policy is still logged on the passive node?
    What will happen if failover occurs but that user has already logged off the passive node?
    AKAIK, mapped drives are tied to the current logon user not for all users.  
    I don't know which account is actually used for the failover process.  CNO, VNO, the domain user account used to create the windows cluster, or ?
    Thanks.

  • Windows 2003 File Share 4 node Cluster: Does Cluster Resources need to be brought offline prior removing / unmapping any LUN's from SAN end?

    Hello All,
    Recently, on a 4 node Windows 2003 File Share Cluster, we encountered a problem where when trying to remove few shares (that were no longer in use) directly from SAN end crashed the entire Cluster (i.e., other shares also lost their SAN connectivity). I
    suppose the Cluster resources need to be brought offline prior removing it from SAN but I've been advised that these shares were not the root and instead a 'mount point' created within the share; and hence there is no need of bringing down any Cluster resources
    offline.
    Please can someone comment on the above and provide me detailed steps as to how we go about reclaiming SAN space from specific shares on a W2003 Cluster?
    p.s., let me know if you need any additional information.
    Thanks in advance.

    Hi Alex,
    The problem started when SAN Support reclaimed few storage LUNs by unmapping them from our clustered file servers.  When they reclaimed the unused LUNs, other SAN drives which were there also disappeared causing the unavailability of file shares.
    Internet access is not enabled on these servers. Servers in question are running 64-bit Windows Server 2003 Sp2. This is a four node file share cluster. When the unsued LUN's were pulled, the entire Cluster lost its SAN connectivity. Windows cluster service
    was not starting on any of  the cluster nodes. To resolve the problem all the four cluster nodes were rebooted after which cluster service started on all the cluster nodes and resources came online.
    Some of the events at the time of problem occurrence were,
    Event ID     : 57                                                      
    Raw Event ID : 57                                                      
    Record Nr.   : 25424072                                                
    Category     : None                                                    
    Source       : Ftdisk                                                  
    Type         : Warning                                                 
    Generated    : 19.07.2014 10:49:46                                     
    Written      : 19.07.2014 10:49:46                                     
    Machine      : ********                                             
    Message      : The system failed to flush data to the transaction log.   
    Corruption may occur.                                                  
    Event ID     : 1209   
    Raw Event ID : 1209                                                    
    Record Nr.   : 25424002                                                
    Category     : None                                                    
    Source       : ClusDisk                                                
    Type         : Error                                                   
    Generated    : 19.07.2014 10:49:10                                     
    Written      : 19.07.2014 10:49:10                                     
    Machine      : ***********                                             
    Message      : Cluster service is requesting a bus reset for device      
    \Device\ClusDisk0.                                                     
    Event ID     : 15     
    Raw Event ID : 15                                                      
    Record Nr.   : 25412958                                                
    Category     : None                                                    
    Source       : Disk                                                    
    Type         : Error                                                   
    Generated    : 11.07.2014 10:54:53                                     
    Written      : 11.07.2014 10:54:53                                     
    Machine      : *************                                            
    Message      : The device, \Device\Harddisk46\DR48, is not ready for access yet.                                                            
    Let me know if you need any additional info, many thanks.

  • Should one use MPIO and/or CSV in a Windows 2012 R2 guest cluster?

    Should one use MPIO and/or CSV in a Windows 2012 R2 guest cluster using VMware ESXi 5.5 presented Fiber LUN RDMs.
    If MPIO were implemented is there a preference for HW manufacturer DISM vs. MS DISM in a guest cluster?
    What partition size/offset is recommended for the MSR partition (currently set to 1000 MB) - unfortunately seeing storage validation error with failing block write at block 2048 (which in return may be related to VMware ESXi 5.5. disk partition layout)
    The current setup works without using MPIO (question is would it help overcome the current failing persistent SCSI-3 reservation warning.)
    What were the benefit of using CSV if any in a guest cluster? The Luns in scope would eventually hold SQL data and log files.
     Thanks for your input already.
    Sassan Karai

    Hi,
    Regardless of what type of the shared storage is failover cluster have to use the shared storage, the shared storage can redirect the failed node data to others node, accordingly
    the failover cluster can get the high availability.
    From you descripted error there must you choose the VMware® unsupported storage with failover cluster, please refer the following VMware® official KB then reconfirm your topology
    design is supported.
    Third party KB:
    VMware vSphere support for Microsoft clustering solutions on VMware products
    http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1037959
    The related KB:
    Failover Clustering Hardware Requirements and Storage Options
    http://technet.microsoft.com/zh-cn/library/jj612869.aspx
    More information:
    Validate Storage Spaces Persistent Reservation Test Results with Warning
    http://blogs.msdn.com/b/clustering/archive/2013/05/24/10421247.aspx
    Understanding Cluster Validation Tests: Storage
    http://technet.microsoft.com/en-us/library/cc771259.aspx#PersistentReservation
    Shared storage for Windows Failover Cluster with MPIO
    http://blogs.technet.com/b/storageserver/archive/2011/05/31/shared-storage-for-windows-failover-cluster-with-mpio.aspx
    Hope this helps.
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • 2 node cluster windows 2008 R2 cluster won't form

    Hi,
    I'm trying to build a 2 node cluster and having no luck at all.   My validation works great but when I go to create the cluster it fails on the forming cluster section after about 2 minutes.   I seem to have everything in order and can't
    figure out the problem.   Here is what is in the cluster log.
    000010b0.00001108::2010/08/26-20:43:54.067 INFO  [CHANNEL 192.168.210.236:~59824~] graceful close, status (of previous failure, may not indicate problem) ERROR_SUCCESS(0)
    000010b0.00001108::2010/08/26-20:43:54.067 ERR   [SM] Caught GracefulClose(1226)' because of 'channel to remote endpoint 192.168.210.236:~59824~ is closed' authenticating.
    000010b0.000012d8::2010/08/26-20:43:54.067 INFO  [SV] Authentication and authorization were successful
    000010b0.000012d8::2010/08/26-20:43:54.067 DBG   [SM] SrvCtxt initialized with package Kerberos, MaxTokenSize = 12000, RequiredCtxAttrib = 133122, HandShakeTimeout = 30000
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [SM] Handling auth handshake posted by thread id 4824
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [CHANNEL 192.168.210.236:~59824~]/recv: Attempting to perform I/O on closed stream.
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [CHANNEL 192.168.210.236:~59824~] Not closing handle because it is invalid.
    000010b0.00001108::2010/08/26-20:43:54.067 INFO  [CHANNEL 192.168.210.236:~59824~] graceful close, status (of previous failure, may not indicate problem) ERROR_SUCCESS(0)
    000010b0.00001108::2010/08/26-20:43:54.067 ERR   [SM] Caught GracefulClose(1226)' because of 'channel to remote endpoint 192.168.210.236:~59824~ is closed' authenticating.
    000010b0.000012d8::2010/08/26-20:43:54.067 DBG   [CHANNEL 192.168.210.236:~59824~] Not closing handle because it is invalid.
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  [CS] Service Stopping...
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  [CORE] Node quorum state is 'Not yet formed or joined a cluster'. Form/join status with other nodes is as follows:
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  [NODE] Node 1: Farthest reported progress joining with node iQCXSQL2 (id 2) is: Authenticate Initial Connection at time 2010/08/26-20:43:54.004: status 802 [SV] Authentication failed
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  [CORE] Node 1: Proposed View is <ViewChanged joiners=() downers=(1) newView=101() oldView=001(1) joiner=false form=false/>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  [RGP] Node 1: Stable_`0 => Opening`1
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  <class mscs::detail::ConsensusMessage>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <senderId> 1 </senderId>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <bestEpochSeen> 0 </bestEpochSeen>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <lastStableView> 001(1) </lastStableView>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <proposedView> 101() </proposedView>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <stage> Opening`1 </stage>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <inStage> () </inStage>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <inStagePrev> () </inStagePrev>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <joiners> () </joiners>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <trimmedNodes> (1) </trimmedNodes>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <innerScreen> () </innerScreen>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <pruningResult> () </pruningResult>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <matrix>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO      <ConnectivityMatrix>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO        <row id='0'00000000000000000000000000000000/>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO      </ConnectivityMatrix>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    </matrix>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <gemBlob> <counted_ptr p = 'nullptr'/> </gemBlob>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  </class mscs::detail::ConsensusMessage>
    000010b0.000006fc::2010/08/26-20:44:52.906 INFO  [RGP] Node 1: Timer Tick Started
    000010b0.0000106c::2010/08/26-20:44:52.906 DBG   [NODE] Node 1: eating message sent to the dead node 2
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  [RGP] Node 1: Just sent farewell regroup message
    000010b0.000010b8::2010/08/26-20:44:52.906 DBG   [NODE] Node 1: just about to send a message of size 0 to 1
    000010b0.000010b8::2010/08/26-20:44:52.906 DBG   [NODE] Node 1: message to node 1 sent
    000010b0.000010b8::2010/08/26-20:44:52.906 DBG   [NODE] Node 1: just about to send a message of size 361 to 1
    000010b0.000010b8::2010/08/26-20:44:52.906 DBG   [NODE] Node 1: message to node 1 sent
    000010b0.000006fc::2010/08/26-20:44:53.218 INFO  [RGP] Node 1: Timer Tick Started
    000010b0.0000106c::2010/08/26-20:44:53.420 INFO  [DM]: Shutting down, so unloading the cluster database.
    000010b0.0000106c::2010/08/26-20:44:53.420 INFO  [DM] Shutting down, so unloading the cluster database (waitForLock: true).
    000010b0.0000106c::2010/08/26-20:44:53.420 DBG   [DM] Unloading Hive, Key \Registry\Machine\Cluster, discardCurrentChanges true
    000010b0.0000106c::2010/08/26-20:44:53.420 INFO  [CS] Service Stopped...
    000010b0.0000106c::2010/08/26-20:44:53.420 INFO  [CS] About to exit service...
    Just to make sure I have given both nodes every permission under the sun.  I have tried creating the cluster computer in AD beforehand and disabling the computer and I have tried letting clustering services create the computer to no avail.  
    Please help!

    Here is the complete log Edwin.   Thank you.   
    000010b0.0000121c::2010/08/26-20:41:52.722 INFO  -----------------------------+ LOG BEGIN +-----------------------------
    000010b0.0000121c::2010/08/26-20:41:52.722 INFO  [CS] Starting clussvc as a service
    000010b0.0000121c::2010/08/26-20:41:52.722 INFO  [CS] cluster service logging level is 5
    000010b0.000012d0::2010/08/26-20:41:52.738 INFO  [CS] Creating cluster node <vector len='1'>
    000010b0.000012d0::2010/08/26-20:41:52.738 INFO      <item>ClusSvc</item>
    000010b0.000012d0::2010/08/26-20:41:52.738 INFO  </vector>
    000010b0.00000fd8::2010/08/26-20:41:52.738 INFO  [StartupConfig]: Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.738 INFO  [StartupConfig]: Local node name: iQCXSQL1.
    000010b0.00000fd8::2010/08/26-20:41:52.738 INFO  [StartupConfig]: Initializing NetFT adapter management information.
    000010b0.00000fd8::2010/08/26-20:41:52.738 INFO  [NETFT] Initializing NetFT driver manager
    000010b0.00000fd8::2010/08/26-20:41:52.738 INFO  [CORE] Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.738 INFO  [GEM] Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.738 INFO  [MM] Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.738 INFO  [MRR] Node 1: Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.738 INFO  [Causal] Node 1: Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.738 INFO  [EM] Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.738 INFO  [GUM] Node 1: Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.738 INFO  [DM] Node 1: Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.800 INFO  [CLFS] Initializing CLFS logger.
    000010b0.00000fd8::2010/08/26-20:41:52.800 DBG   [CLFS] Largest written record has size 131072 bytes
    000010b0.00000fd8::2010/08/26-20:41:52.800 DBG   [CLFS] Deleting old marshalling area.
    000010b0.00000fd8::2010/08/26-20:41:52.800 DBG   [CLFS] Marshaling area resized
    000010b0.00000fd8::2010/08/26-20:41:52.800 INFO  [TM] Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.800 INFO  [JPM] Node 1: Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.800 INFO  [VER]: Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.800 INFO  [CCM] Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.800 INFO  [FTI] Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.800 INFO  [ReM] Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.800 INFO  [HM]: Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.800 INFO  [NETFT] Starting NetFT eventing for FTI
    000010b0.00000fd8::2010/08/26-20:41:52.816 DBG   [NETFTAPI] received NsiInitialNotification
    000010b0.00000fd8::2010/08/26-20:41:52.816 DBG   [NETFTAPI] received NsiInitialNotification
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO  [NETFT] FTI NetFT event handler ready for events.
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO  [IM] Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO  [NM] Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO  [NSM] Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO  [CM] Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO  [RCM] Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO  [QUORUM] Node 1: Initializing.
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO  [STARTUP] Node 1: startup config
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO  <StartupConfig localNodeId='1' testMode='false' newCluster='true'
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO      <ClusterName> iqcxcluster </ClusterName>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO      <PaxosTag> 0:0:0 </PaxosTag>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO      <ConfiguredNodes> (1 2) </ConfiguredNodes>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO      <NodeWeights> (1 2) </NodeWeights>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO      <Nodes>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO          <Node name='iQCXSQL1' id='1'/>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO          <Node name='iQCXSQL2' id='2'/>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO      </Nodes>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO      <LocalHiveFile> C:\Windows\Cluster\CLUSDB </LocalHiveFile>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO      <LocalHiveKey> Cluster </LocalHiveKey>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO      <ParametersKey> System\\CurrentControlSet\\Services\ClusSvc\Parameters </ParametersKey>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO      <Virtual IPv6> fe80::c14f:8b1f:d265:9b42%22:~3343~ </Virtual IPv6>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO      <Virtual IPv4> 169.254.1.247:~3343~ </Virtual IPv4>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO      <NoVersionCheck> false </NoVersionCheck>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO      <Cookie> 3529ace1-67e4-4107-8940-c725883fab51 </Cookie>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO      <ClusterGuid> d4a97f31-178f-4726-993c-c0c5ff8407dc </ClusterGuid>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO      <RetainNetFTConnSecurity> false </RetainNetFTConnSecurity>
    000010b0.00000fd8::2010/08/26-20:41:52.816 INFO  </StartupConfig>
    000010b0.000012d0::2010/08/26-20:41:52.816 INFO  [CS] Node object creation completed.
    000010b0.000012d0::2010/08/26-20:41:52.816 INFO  [CS] Starting cluster node
    000010b0.000012d0::2010/08/26-20:41:52.816 INFO  [CS] Cluster Service started
    000010b0.000012d0::2010/08/26-20:41:52.816 INFO  [CS] The current time is 2010/08/26-16:41:52.816
    000010b0.000012d0::2010/08/26-20:41:52.816 INFO  [CS] Using virtual IPv6 Address fe80::c14f:8b1f:d265:9b42%22:~3343~
    000010b0.000012d0::2010/08/26-20:41:52.816 INFO  [CS] Using virtual IPv4 Address 169.254.1.247:~3343~
    000010b0.000012d0::2010/08/26-20:41:52.816 INFO  [CS] Max Licensed nodes is set 16
    000010b0.000012d0::2010/08/26-20:41:52.816 INFO  [NETFT] Disabling IP autoconfiguration on the NetFT adapter.
    000010b0.000012d0::2010/08/26-20:41:52.816 INFO  [NETFT] Disabling DHCP on the NetFT adapter.
    000010b0.000012d0::2010/08/26-20:41:52.816 DBG   [NETFT] Disabling DHCP on NetFT interface name ethernet_2.
    000010b0.000012d0::2010/08/26-20:41:52.816 INFO  [CS] Starting DM
    000010b0.000012d0::2010/08/26-20:41:52.816 INFO  [DM] Node 1: Reading quorum config
    000010b0.000012d0::2010/08/26-20:41:52.816 INFO  [DM] Node 1: new cluster. generating new hive
    000010b0.000012d0::2010/08/26-20:41:52.816 DBG   [DM] Unloading Hive, Key \Registry\Machine\Cluster, discardCurrentChanges true
    000010b0.000012d8::2010/08/26-20:41:52.816 DBG   [NETFTAPI] received NsiAddInstance  for 169.254.1.247
    000010b0.000010a0::2010/08/26-20:41:52.816 INFO  [CS] Disabling connection security.
    000010b0.000012d0::2010/08/26-20:41:52.816 INFO  [DM] Key \Registry\Machine\Cluster does not appear to be loaded (status STATUS_OBJECT_NAME_NOT_FOUND(c0000034))
    000010b0.000012d0::2010/08/26-20:41:52.816 DBG   [DM] Unloading Hive, Key \Registry\Machine\Cluster, discardCurrentChanges true
    000010b0.000012d0::2010/08/26-20:41:52.816 INFO  [DM] Key \Registry\Machine\Cluster does not appear to be loaded (status STATUS_OBJECT_NAME_NOT_FOUND(c0000034))
    000010b0.000012d0::2010/08/26-20:41:52.816 DBG   [DM] Creating Hive, Key Cluster, FilePath C:\Windows\Cluster\CLUSDB
    000010b0.000010a0::2010/08/26-20:41:52.816 INFO  [NETFT] Adding connection security permit filters for the NetFT adapter.
    000010b0.000010a0::2010/08/26-20:41:52.816 INFO  [NETFT] Adding FWPM filter Failover Cluster Permit Inbound Transport IPv4
    000010b0.000010a0::2010/08/26-20:41:52.831 INFO  [NETFT] Adding FWPM filter Failover Cluster Permit Inbound Transport IPv6
    000010b0.000010a0::2010/08/26-20:41:52.831 INFO  [NETFT] Adding FWPM filter Failover Cluster Permit Outbound Transport IPv4
    000010b0.000010a0::2010/08/26-20:41:52.831 INFO  [NETFT] Adding FWPM filter Failover Cluster Permit Outbound Transport IPv6
    000010b0.000010a0::2010/08/26-20:41:52.831 INFO  [NETFT] Adding FWPM filter Failover Cluster Permit ALE Accept IPv4 Universal
    000010b0.000010a0::2010/08/26-20:41:52.831 INFO  [NETFT] Adding FWPM filter Failover Cluster Permit ALE Accept IPv6 Universal
    000010b0.000010a0::2010/08/26-20:41:52.831 INFO  [NETFT] Adding FWPM filter Failover Cluster Permit ALE Accept IPv4 LIPS
    000010b0.000010a0::2010/08/26-20:41:52.831 INFO  [NETFT] Adding FWPM filter Failover Cluster Permit ALE Accept IPv6 LIPS
    000010b0.000010a0::2010/08/26-20:41:52.831 DBG   [CS] Done with connection security
    000010b0.000012d0::2010/08/26-20:41:52.847 DBG   [DM] Saving temp Hive
    000010b0.000012d0::2010/08/26-20:41:52.847 DBG   [DM] Unloading temp Hive
    000010b0.000012d0::2010/08/26-20:41:52.847 DBG   [DM] Deleting temp Hive
    000010b0.000012d0::2010/08/26-20:41:52.847 INFO  [DM] Loading Hive, Key Cluster, FilePath C:\Windows\Cluster\CLUSDB
    000010b0.000012d0::2010/08/26-20:41:52.878 INFO  [DM] Paxos Tag Read from Hive: 0:0:0
    000010b0.000012d0::2010/08/26-20:41:52.878 DBG   [DM] Beginning transaction
    000010b0.000012d0::2010/08/26-20:41:52.878 INFO  [QUORUM] unique node id for new node 1 is 00000000-0000-0000-0000-000000000001
    000010b0.000012d0::2010/08/26-20:41:52.878 INFO  [QUORUM] unique node id for new node 2 is 00000000-0000-0000-0000-000000000002
    000010b0.000012d0::2010/08/26-20:41:52.878 DBG   [CLFS] Deleting old marshalling area.
    000010b0.000012d0::2010/08/26-20:41:52.878 DBG   [CLFS] Marshaling area resized
    000010b0.000012d0::2010/08/26-20:41:52.878 DBG   [CLFS] Logger wrote 2820 bytes
    000010b0.000012d0::2010/08/26-20:41:52.878 DBG   [DM] Committed transaction 0:0:1
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [DM] Populating StartupConfig data from local hive.
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [DM] Node 1: startup configuration loaded:
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  <StartupConfig localNodeId='1' testMode='false' newCluster='true'
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <ClusterName> iqcxcluster </ClusterName>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <PaxosTag> 0:0:1 </PaxosTag>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <ConfiguredNodes> (1 2) </ConfiguredNodes>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <NodeWeights> (1 2) </NodeWeights>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <Nodes>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO          <Node name='iQCXSQL1' id='1'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO          <Node name='iQCXSQL2' id='2'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      </Nodes>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <LocalHiveFile> C:\Windows\Cluster\CLUSDB </LocalHiveFile>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <LocalHiveKey> Cluster </LocalHiveKey>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <ParametersKey> System\\CurrentControlSet\\Services\ClusSvc\Parameters </ParametersKey>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <Virtual IPv6> fe80::c14f:8b1f:d265:9b42%22:~3343~ </Virtual IPv6>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <Virtual IPv4> 169.254.1.247:~3343~ </Virtual IPv4>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <NoVersionCheck> false </NoVersionCheck>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <Cookie> 3529ace1-67e4-4107-8940-c725883fab51 </Cookie>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <ClusterGuid> d4a97f31-178f-4726-993c-c0c5ff8407dc </ClusterGuid>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <RetainNetFTConnSecurity> false </RetainNetFTConnSecurity>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  </StartupConfig>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [CS] Starting MM
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [MM] Starting.
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [GEM] Node 1: Resume (newView = 001(1), previousView = 000())
    000010b0.000012d0::2010/08/26-20:41:52.894 DBG   [GEM] Node 1: SendNextMessage
    000010b0.000012d0::2010/08/26-20:41:52.894 DBG   [GEM] Node 1: No more messages in queue
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [StartupConfig] Writing new StartupConfig entries to ObjectManager
    000010b0.000012d0::2010/08/26-20:41:52.894 DBG   [StartupConfig] Writing node iQCXSQL2 (id = 2) to ObjectManager
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [QUORUM] unique node id for new node 2 is 00000000-0000-0000-0000-000000000002
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [MQ-iQCXSQL2] Starting
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [MQ-iQCXSQL2] Clearing 0 unsent and 0 unacknowledged messages.
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  <ObjectManager>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO    <ROOT id='.root' name='.root'>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <GEMA id='GemAgent' name='GemAgent1'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <MRRA id='MrrAgent1' name='MrrAgent1'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <CAUS id='CausalAgent' name='CausalAgent'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <EVTM id='EventManager' name='EventManager'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <GUMA id='GumAgent1' name='GumAgent1'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <JPMA id='JpmAgent' name='JpmAgent'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <IMSC id='[IM-1] ' name='[IM-1] '/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <RPCS id='[RPC-1] ' name='[RPC-1] '/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <RCMA id='RcmAgent' name='RcmAgent'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <QRMA id='QuorumAgent1' name='QuorumAgent1'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <DCMA id='DcmAgent' name='DcmAgent'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <VSSR id='Cluster VSS Writer' name='Cluster VSS Writer'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <CLUS id='iqcxcluster' name='iqcxcluster'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <DEAD id='.Dead' name='.Dead'>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO        <NODE id='2' name='iQCXSQL2'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      </DEAD>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <LIVE id='.Live' name='.Live'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <JOIN id='.Joiners' name='.Joiners'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO      <MBRA id='MembershipAgent' name='MembershipAgent'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO    </ROOT>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  </ObjectManager>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [MM] Node 1: got a network stream to 1. Creating new NodeObject
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [QUORUM] unique node id for new node 1 is 00000000-0000-0000-0000-000000000001
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [MQ-iQCXSQL1] Starting
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [MQ-iQCXSQL1] Clearing 0 unsent and 0 unacknowledged messages.
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [NODE] Node 1: n1 node object adding stream
    000010b0.000012d0::2010/08/26-20:41:52.894 DBG   [NODE] Node 1: n1 node object got a channel
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [NODE] Node 1: Highest version with n1 = Major 0 Minor 0 Upgrade 0 ClusterVersion 0x00000000, lowest = Major 0 Minor 0 Upgrade 0 ClusterVersion 0x00000000
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [NODE] Node 1: Done processing new stream to n1.
    000010b0.00000e50::2010/08/26-20:41:52.894 INFO  [PULLER iQCXSQL1] Just about to start reading from <LoopbackStream name='iQCXSQL1'/>
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [CS] Starting QM
    000010b0.000012d0::2010/08/26-20:41:52.894 DBG   [NETFT] Halt Notification set
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [CS] Starting TM
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [NETFT] Starting NetFT eventing for TM
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [NETFT] TM NetFT event handler ready for events.
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [CS] Starting IM
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [NETFT] Starting NetFT eventing for IM
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [NETFT] IM NetFT event handler ready for events.
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [CS] Starting NM
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [CS] Starting api server
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [API] Api Onlined
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [API] UseProtseq ncalrpc => 0
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [API] UseProtseq ncadg_ip_tcp => 0
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [API] InqBindings => 0
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [API] RpcEpRegister => 0
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [API] RpcServerRegister => 0
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [API] Initializing
    000010b0.000012d0::2010/08/26-20:41:52.894 INFO  [API] Adding Mgmt API SPNs to node's Computer Object
    000010b0.000012d0::2010/08/26-20:41:52.972 INFO  [API] Successfully added Mgmt API SPNs to node's Computer Object
    000010b0.000012d0::2010/08/26-20:41:52.972 ERR   [API] DmQueryString failed to retrieve the security   descriptor status 2, default security descriptor will be used for authorizing client connections
    000010b0.000012d0::2010/08/26-20:41:52.972 INFO  [API] Successfully built the default cluster SD.
    000010b0.000012d0::2010/08/26-20:41:52.972 INFO  [CS] Bringing API Online
    000010b0.000012d0::2010/08/26-20:41:52.972 INFO  [CORE] Node 1: New View is <ViewChanged joiners=(1) downers=() newView=001(1) oldView=000() joiner=true form=true/> (Start Dispatch)
    000010b0.000012d0::2010/08/26-20:41:52.972 INFO  [MRR] Node 1: Process view 001(1)
    000010b0.000012d0::2010/08/26-20:41:52.972 INFO  [RCM] rcm::RcmAgent::ProcessView: Ignoring view change (won't attempt to move orphaned groups), since I'm not in RCM view yet.
    000010b0.000012d0::2010/08/26-20:41:52.972 DBG   [GEM] Node 1: message added to send queue
    000010b0.000012d0::2010/08/26-20:41:52.972 DBG   [GEM] Node 1: SendNextMessage
    000010b0.000012d0::2010/08/26-20:41:52.972 DBG   [GEM] Node 1: delivering message 1:1
    000010b0.000012d0::2010/08/26-20:41:52.972 INFO  [CORE] EmitReadyForState for node 1
    000010b0.000012d0::2010/08/26-20:41:52.972 INFO  [CORE] Issuing NeedStateViewChanged((1))
    000010b0.000012d0::2010/08/26-20:41:52.972 DBG   [GEM] Node 1: ProcessAck from node 1 (already acked ())
    000010b0.000012d0::2010/08/26-20:41:52.972 DBG   [GEM] Node 1: SendNextMessage
    000010b0.000012d0::2010/08/26-20:41:52.972 DBG   [GEM] Node 1: No more messages in queue
    000010b0.000012d0::2010/08/26-20:41:52.972 DBG   [NODE] Node 1: eating message sent to the dead node 2
    000010b0.000012d0::2010/08/26-20:41:52.972 INFO  [QUORUM] Node 1: I am the form/join coordinator (offline and lowest number node). Launching worker to request quorum config from other nodes
    000010b0.000012d0::2010/08/26-20:41:52.972 DBG   [MRR] Node 1: insert request record for RID 1
    000010b0.000012d0::2010/08/26-20:41:52.972 DBG   [MRR] SendRequest(<message action='qrm/request-info' GemId='0' target='QRMA' sender='-1'mrr='1'/>)
    000010b0.000012d0::2010/08/26-20:41:52.972 DBG   [CORE] Node 1: sending qrm/request-info to QRMA at iQCXSQL1
    000010b0.000010b8::2010/08/26-20:41:52.972 DBG   [NODE] Node 1: just about to send a message of size 109 to 1
    000010b0.000012d0::2010/08/26-20:41:52.972 INFO  [CORE] Node 1: New View is <ViewChanged joiners=(1) downers=() newView=001(1) oldView=000() joiner=true form=true/> (Stop Dispatch)
    000010b0.000012d0::2010/08/26-20:41:52.972 INFO  [CS] Starting HM Server
    000010b0.000010b8::2010/08/26-20:41:52.972 DBG   [NODE] Node 1: message to node 1 sent
    000010b0.000012d0::2010/08/26-20:41:52.972 INFO  [NODE] Node 1: New join with n2: stage: 'Listen for Connection'
    000010b0.000012d0::2010/08/26-20:41:52.972 INFO  [CS] Starting JPM (attempting connections to other cluster nodes)
    000010b0.000012d0::2010/08/26-20:41:52.972 DBG   [JPM] Node 1: Dead node mentioned iQCXSQL1
    000010b0.000012d0::2010/08/26-20:41:52.972 DBG   [JPM] Node 1: Dead node mentioned iQCXSQL2
    000010b0.000012d0::2010/08/26-20:41:52.972 INFO  [CS] Subscribing for VSS requests
    000010b0.000012d0::2010/08/26-20:41:52.972 INFO  [VSS] Initializing Cluster VSS Writer at the command of the VSS framework.
    000010b0.00000e50::2010/08/26-20:41:52.972 INFO  [QUORUM] Node 1: received request for quorum info. Replying with quorum state <QuorumConfig tag='0:0:1' set='(1 2)' weights='(1 2)'/>
    000010b0.00000e50::2010/08/26-20:41:52.972 DBG   [CORE] Node 1: sending mrr/reply to MRRA at iQCXSQL1
    000010b0.000010b8::2010/08/26-20:41:52.972 DBG   [NODE] Node 1: just about to send a message of size 0 to 1
    000010b0.000010b8::2010/08/26-20:41:52.972 DBG   [NODE] Node 1: message to node 1 sent
    000010b0.000012d8::2010/08/26-20:41:52.972 DBG   [JPM] Node 1: Building contact data for node iQCXSQL2.
    000010b0.000010b8::2010/08/26-20:41:52.972 DBG   [NODE] Node 1: just about to send a message of size 278 to 1
    000010b0.000010b8::2010/08/26-20:41:52.972 DBG   [NODE] Node 1: message to node 1 sent
    000010b0.000012d8::2010/08/26-20:41:52.972 INFO  [JPM] Node 1: Scheduling DNS query for node name iQCXSQL2
    000010b0.000012d8::2010/08/26-20:41:52.972 DBG   [JPM] Node 1: no valid contact endpoints yet for node iQCXSQL2.
    000010b0.00000e50::2010/08/26-20:41:52.972 DBG   [MRR] Node 1: processing reply <message action='mrr/reply' GemId='0' target='MRRA' sender='1'mrr='1'/>
    000010b0.00000e50::2010/08/26-20:41:52.972 DBG   [MRR] Node 1: Request RID 1 is completed
    000010b0.0000119c::2010/08/26-20:41:52.972 INFO  [QUORUM] Node 1: setting next best epoch to 0
    000010b0.0000119c::2010/08/26-20:41:52.972 DBG   [QC] Calculating quorum: Configured nodes count = 2, Node weights count = 2, Quorum witness configured = false, Intersection count = 1
    000010b0.0000119c::2010/08/26-20:41:52.972 INFO  [QUORUM] Node 1: Coordinator: no quorum yet. do nothing
    000010b0.000012d8::2010/08/26-20:41:52.972 INFO  [JPM] Node 1: Found possible contact address 192.168.210.236:~0~ for node iQCXSQL2 via DNS.
    000010b0.000012d0::2010/08/26-20:41:52.987 INFO  [CS] Starting the Watchdog Monitor
    000010b0.000012d0::2010/08/26-20:41:52.987 INFO  [WM] Monitoring the kernel-mode watchdog.
    000010b0.000012d0::2010/08/26-20:41:52.987 INFO  [CS] Reporting to SCM that cluster service has started.
    000010b0.000012d8::2010/08/26-20:41:53.019 INFO  [NM] Received request from client address ::1.
    000010b0.000012d8::2010/08/26-20:41:53.019 DBG   [API] Authenticated client--Client: IQCX\iqcxsql1 Interface: b97db8b2-4c63-11cf-bff6-08002be23f2f Server:
    [email protected] Level: RPC_C_AUTHN_LEVEL_PKT_PRIVACY Service: RPC_C_AUTHN_WINNT Protocol Sequence: ncacn_ip_tcp Client Address: ::1 Network Option: .
    000010b0.0000119c::2010/08/26-20:41:53.986 DBG   [JPM] Node 1: contacts size for node iQCXSQL2 is 1, current index 0
    000010b0.0000119c::2010/08/26-20:41:53.986 DBG   [JPM] Node 1: Trying to connect to node iQCXSQL2 (IP: 192.168.210.236:~0~)
    000010b0.0000119c::2010/08/26-20:41:53.986 DBG   [HM] Trying to connect to iQCXSQL2 at 192.168.210.236:~3343~
    000010b0.000012d8::2010/08/26-20:41:53.986 INFO  [CONNECT] 192.168.210.236:~3343~: Established connection to remote endpoint 192.168.210.236:~3343~.
    000010b0.000012d8::2010/08/26-20:41:53.986 INFO  [NODE] Node 1: New join with n2: stage: 'Attempt Initial Connection'
    000010b0.000012d8::2010/08/26-20:41:53.986 INFO  [SV] Securing route from (192.168.210.239:~49538~) to remote iQCXSQL2 (192.168.210.236:~3343~).
    000010b0.000012d8::2010/08/26-20:41:53.986 INFO  [SV] Got a new outgoing stream to iQCXSQL2 at 192.168.210.236:~3343~
    000010b0.000012d8::2010/08/26-20:41:53.986 DBG   [SM] Joiner: Initialized with SPN = iQCXSQL2, Package = Kerberos, RequiredCtxAttrib = 83990, HandShakeTimeout = 30000
    000010b0.000013f0::2010/08/26-20:41:53.986 DBG   [SM] Handling auth handshake posted by thread id 4824
    000010b0.000013f0::2010/08/26-20:41:53.986 DBG   [SM] Joiner: ISC returned status = 590610 output Blob size 1388
    000010b0.000013f0::2010/08/26-20:41:53.986 DBG   [SM] Joiner: Received SSPI blob from the Sponsor of size 96
    000010b0.000013f0::2010/08/26-20:41:53.986 DBG   [SM] Joiner: ISC returned status = 590610 output Blob size 1388
    000010b0.000013f0::2010/08/26-20:41:53.986 DBG   [SM] Joiner: Received SSPI blob from the Sponsor of size 96
    000010b0.000013f0::2010/08/26-20:41:53.986 DBG   [SM] Joiner: ISC returned status = -2146893022 output Blob size 12000
    000010b0.000012d8::2010/08/26-20:41:53.986 INFO  [NODE] Node 1: New join with n2: stage: 'Authenticate Initial Connection' status HrError(0x80090322) reason: '[SV] Authentication failed'
    000010b0.000012d8::2010/08/26-20:41:53.986 DBG   [CHANNEL 192.168.210.236:~3343~] Close().
    000010b0.000012d8::2010/08/26-20:41:53.986 WARN  cxl::ConnectWorker::operator (): HrError(0x80090322)' because of '[SV] Authentication or Authorization Failed'
    000010b0.00000ea8::2010/08/26-20:41:54.048 INFO  [ACCEPT] 0.0.0.0:~3343~: Accepted inbound connection from remote endpoint 192.168.210.236:~59820~.
    000010b0.000012d8::2010/08/26-20:41:54.048 INFO  [SV] Securing route from (192.168.210.239:~3343~) to remote  (192.168.210.236:~59820~).
    000010b0.000012d8::2010/08/26-20:41:54.048 INFO  [SV] Got a new incoming stream from 192.168.210.236:~59820~
    000010b0.000012d8::2010/08/26-20:41:54.048 DBG   [SM] SrvCtxt initialized with package Kerberos, MaxTokenSize = 12000, RequiredCtxAttrib = 165910, HandShakeTimeout = 30000
    000010b0.00001108::2010/08/26-20:41:54.048 DBG   [SM] Handling auth handshake posted by thread id 4824
    000010b0.00001108::2010/08/26-20:41:54.048 DBG   [SM] Sponsor: Received SSPI blob from the Joiner of size 1411
    000010b0.00001108::2010/08/26-20:41:54.048 DBG   [SM] Sponsor: SSPI ASC returned status = 590610
    000010b0.00001108::2010/08/26-20:41:54.048 DBG   [SM] Sponsor: Sending SSPI blob of size 96 to Joiner
    000010b0.00001108::2010/08/26-20:41:54.048 DBG   [SM] Sponsor: Received SSPI blob from the Joiner of size 1411
    000010b0.00001108::2010/08/26-20:41:54.048 DBG   [SM] Sponsor: SSPI ASC returned status = 590610
    000010b0.00001108::2010/08/26-20:41:54.048 DBG   [SM] Sponsor: Sending SSPI blob of size 96 to Joiner
    000010b0.00001108::2010/08/26-20:41:54.048 DBG   [SM] Sponsor: Received SSPI blob from the Joiner of size 12000
    000010b0.00001108::2010/08/26-20:41:54.048 DBG   [SM] Sponsor: SSPI ASC returned status = 590610
    000010b0.00001108::2010/08/26-20:41:54.048 DBG   [SM] Sponsor: Sending SSPI blob of size 115 to Joiner
    000010b0.00001108::2010/08/26-20:41:54.048 DBG   [CHANNEL 192.168.210.236:~59820~] Close().
    000010b0.00001108::2010/08/26-20:41:54.048 INFO  [CHANNEL 192.168.210.236:~59820~] graceful close, status (of previous failure, may not indicate problem) ERROR_SUCCESS(0)
    000010b0.00001108::2010/08/26-20:41:54.048 ERR   [SM] Caught GracefulClose(1226)' because of 'channel to remote endpoint 192.168.210.236:~59820~ is closed' authenticating.
    000010b0.000012d8::2010/08/26-20:41:54.048 INFO  [SV] Authentication and authorization were successful
    000010b0.000012d8::2010/08/26-20:41:54.048 DBG   [SM] SrvCtxt initialized with package Kerberos, MaxTokenSize = 12000, RequiredCtxAttrib = 133122, HandShakeTimeout = 30000
    000010b0.00001108::2010/08/26-20:41:54.048 DBG   [SM] Handling auth handshake posted by thread id 4824
    000010b0.00001108::2010/08/26-20:41:54.048 DBG   [CHANNEL 192.168.210.236:~59820~]/recv: Attempting to perform I/O on closed stream.
    000010b0.00001108::2010/08/26-20:41:54.048 DBG   [CHANNEL 192.168.210.236:~59820~] Not closing handle because it is invalid.
    000010b0.00001108::2010/08/26-20:41:54.048 INFO  [CHANNEL 192.168.210.236:~59820~] graceful close, status (of previous failure, may not indicate problem) ERROR_SUCCESS(0)
    000010b0.00001108::2010/08/26-20:41:54.048 ERR   [SM] Caught GracefulClose(1226)' because of 'channel to remote endpoint 192.168.210.236:~59820~ is closed' authenticating.
    000010b0.000012d8::2010/08/26-20:41:54.048 DBG   [CHANNEL 192.168.210.236:~59820~] Not closing handle because it is invalid.
    000010b0.00001310::2010/08/26-20:42:53.995 DBG   [JPM] Node 1: contacts size for node iQCXSQL2 is 1, current index 0
    000010b0.00001310::2010/08/26-20:42:53.995 DBG   [JPM] Node 1: Trying to connect to node iQCXSQL2 (IP: 192.168.210.236:~0~)
    000010b0.00001310::2010/08/26-20:42:53.995 DBG   [HM] Trying to connect to iQCXSQL2 at 192.168.210.236:~3343~
    000010b0.00001310::2010/08/26-20:42:53.995 INFO  [CONNECT] 192.168.210.236:~3343~: Established connection to remote endpoint 192.168.210.236:~3343~.
    000010b0.00001310::2010/08/26-20:42:53.995 INFO  [SV] Securing route from (192.168.210.239:~49561~) to remote iQCXSQL2 (192.168.210.236:~3343~).
    000010b0.00001310::2010/08/26-20:42:53.995 INFO  [SV] Got a new outgoing stream to iQCXSQL2 at 192.168.210.236:~3343~
    000010b0.00001310::2010/08/26-20:42:53.995 DBG   [SM] Joiner: Initialized with SPN = iQCXSQL2, Package = Kerberos, RequiredCtxAttrib = 83990, HandShakeTimeout = 30000
    000010b0.000013f0::2010/08/26-20:42:53.995 DBG   [SM] Handling auth handshake posted by thread id 4880
    000010b0.000013f0::2010/08/26-20:42:53.995 DBG   [SM] Joiner: ISC returned status = 590610 output Blob size 1388
    000010b0.000013f0::2010/08/26-20:42:53.995 DBG   [SM] Joiner: Received SSPI blob from the Sponsor of size 96
    000010b0.000013f0::2010/08/26-20:42:53.995 DBG   [SM] Joiner: ISC returned status = 590610 output Blob size 1388
    000010b0.000013f0::2010/08/26-20:42:53.995 DBG   [SM] Joiner: Received SSPI blob from the Sponsor of size 96
    000010b0.000013f0::2010/08/26-20:42:53.995 DBG   [SM] Joiner: ISC returned status = -2146893022 output Blob size 12000
    000010b0.00001310::2010/08/26-20:42:53.995 INFO  [NODE] Node 1: New join with n2: stage: 'Authenticate Initial Connection' status HrError(0x80090322) reason: '[SV] Authentication failed'
    000010b0.00001310::2010/08/26-20:42:53.995 DBG   [CHANNEL 192.168.210.236:~3343~] Close().
    000010b0.00001310::2010/08/26-20:42:53.995 WARN  cxl::ConnectWorker::operator (): HrError(0x80090322)' because of '[SV] Authentication or Authorization Failed'
    000010b0.00000ea8::2010/08/26-20:42:54.057 INFO  [ACCEPT] 0.0.0.0:~3343~: Accepted inbound connection from remote endpoint 192.168.210.236:~59822~.
    000010b0.00001310::2010/08/26-20:42:54.057 INFO  [SV] Securing route from (192.168.210.239:~3343~) to remote  (192.168.210.236:~59822~).
    000010b0.00001310::2010/08/26-20:42:54.057 INFO  [SV] Got a new incoming stream from 192.168.210.236:~59822~
    000010b0.00001310::2010/08/26-20:42:54.057 DBG   [SM] SrvCtxt initialized with package Kerberos, MaxTokenSize = 12000, RequiredCtxAttrib = 165910, HandShakeTimeout = 30000
    000010b0.00001108::2010/08/26-20:42:54.057 DBG   [SM] Handling auth handshake posted by thread id 4880
    000010b0.00001108::2010/08/26-20:42:54.057 DBG   [SM] Sponsor: Received SSPI blob from the Joiner of size 1411
    000010b0.00001108::2010/08/26-20:42:54.057 DBG   [SM] Sponsor: SSPI ASC returned status = 590610
    000010b0.00001108::2010/08/26-20:42:54.057 DBG   [SM] Sponsor: Sending SSPI blob of size 96 to Joiner
    000010b0.00001108::2010/08/26-20:42:54.057 DBG   [SM] Sponsor: Received SSPI blob from the Joiner of size 1411
    000010b0.00001108::2010/08/26-20:42:54.057 DBG   [SM] Sponsor: SSPI ASC returned status = 590610
    000010b0.00001108::2010/08/26-20:42:54.057 DBG   [SM] Sponsor: Sending SSPI blob of size 96 to Joiner
    000010b0.00001108::2010/08/26-20:42:54.057 DBG   [SM] Sponsor: Received SSPI blob from the Joiner of size 12000
    000010b0.00001108::2010/08/26-20:42:54.057 DBG   [SM] Sponsor: SSPI ASC returned status = 590610
    000010b0.00001108::2010/08/26-20:42:54.057 DBG   [SM] Sponsor: Sending SSPI blob of size 115 to Joiner
    000010b0.00001108::2010/08/26-20:42:54.057 DBG   [CHANNEL 192.168.210.236:~59822~] Close().
    000010b0.00001108::2010/08/26-20:42:54.057 INFO  [CHANNEL 192.168.210.236:~59822~] graceful close, status (of previous failure, may not indicate problem) ERROR_SUCCESS(0)
    000010b0.00001108::2010/08/26-20:42:54.057 ERR   [SM] Caught GracefulClose(1226)' because of 'channel to remote endpoint 192.168.210.236:~59822~ is closed' authenticating.
    000010b0.00001310::2010/08/26-20:42:54.057 INFO  [SV] Authentication and authorization were successful
    000010b0.00001310::2010/08/26-20:42:54.057 DBG   [SM] SrvCtxt initialized with package Kerberos, MaxTokenSize = 12000, RequiredCtxAttrib = 133122, HandShakeTimeout = 30000
    000010b0.00001108::2010/08/26-20:42:54.057 DBG   [SM] Handling auth handshake posted by thread id 4880
    000010b0.00001108::2010/08/26-20:42:54.057 DBG   [CHANNEL 192.168.210.236:~59822~]/recv: Attempting to perform I/O on closed stream.
    000010b0.00001108::2010/08/26-20:42:54.057 DBG   [CHANNEL 192.168.210.236:~59822~] Not closing handle because it is invalid.
    000010b0.00001108::2010/08/26-20:42:54.057 INFO  [CHANNEL 192.168.210.236:~59822~] graceful close, status (of previous failure, may not indicate problem) ERROR_SUCCESS(0)
    000010b0.00001108::2010/08/26-20:42:54.057 ERR   [SM] Caught GracefulClose(1226)' because of 'channel to remote endpoint 192.168.210.236:~59822~ is closed' authenticating.
    000010b0.00001310::2010/08/26-20:42:54.057 DBG   [CHANNEL 192.168.210.236:~59822~] Not closing handle because it is invalid.
    000010b0.000012d8::2010/08/26-20:43:54.004 DBG   [JPM] Node 1: contacts size for node iQCXSQL2 is 1, current index 0
    000010b0.000012d8::2010/08/26-20:43:54.004 DBG   [JPM] Node 1: Trying to connect to node iQCXSQL2 (IP: 192.168.210.236:~0~)
    000010b0.000012d8::2010/08/26-20:43:54.004 DBG   [HM] Trying to connect to iQCXSQL2 at 192.168.210.236:~3343~
    000010b0.000012d8::2010/08/26-20:43:54.004 INFO  [CONNECT] 192.168.210.236:~3343~: Established connection to remote endpoint 192.168.210.236:~3343~.
    000010b0.000012d8::2010/08/26-20:43:54.004 INFO  [SV] Securing route from (192.168.210.239:~49563~) to remote iQCXSQL2 (192.168.210.236:~3343~).
    000010b0.000012d8::2010/08/26-20:43:54.004 INFO  [SV] Got a new outgoing stream to iQCXSQL2 at 192.168.210.236:~3343~
    000010b0.000012d8::2010/08/26-20:43:54.004 DBG   [SM] Joiner: Initialized with SPN = iQCXSQL2, Package = Kerberos, RequiredCtxAttrib = 83990, HandShakeTimeout = 30000
    000010b0.000013f0::2010/08/26-20:43:54.004 DBG   [SM] Handling auth handshake posted by thread id 4824
    000010b0.000013f0::2010/08/26-20:43:54.004 DBG   [SM] Joiner: ISC returned status = 590610 output Blob size 1388
    000010b0.000013f0::2010/08/26-20:43:54.004 DBG   [SM] Joiner: Received SSPI blob from the Sponsor of size 96
    000010b0.000013f0::2010/08/26-20:43:54.004 DBG   [SM] Joiner: ISC returned status = 590610 output Blob size 1388
    000010b0.000013f0::2010/08/26-20:43:54.004 DBG   [SM] Joiner: Received SSPI blob from the Sponsor of size 96
    000010b0.000013f0::2010/08/26-20:43:54.004 DBG   [SM] Joiner: ISC returned status = -2146893022 output Blob size 12000
    000010b0.000012d8::2010/08/26-20:43:54.004 INFO  [NODE] Node 1: New join with n2: stage: 'Authenticate Initial Connection' status HrError(0x80090322) reason: '[SV] Authentication failed'
    000010b0.000012d8::2010/08/26-20:43:54.004 DBG   [CHANNEL 192.168.210.236:~3343~] Close().
    000010b0.000012d8::2010/08/26-20:43:54.004 WARN  cxl::ConnectWorker::operator (): HrError(0x80090322)' because of '[SV] Authentication or Authorization Failed'
    000010b0.00000ea8::2010/08/26-20:43:54.067 INFO  [ACCEPT] 0.0.0.0:~3343~: Accepted inbound connection from remote endpoint 192.168.210.236:~59824~.
    000010b0.000012d8::2010/08/26-20:43:54.067 INFO  [SV] Securing route from (192.168.210.239:~3343~) to remote  (192.168.210.236:~59824~).
    000010b0.000012d8::2010/08/26-20:43:54.067 INFO  [SV] Got a new incoming stream from 192.168.210.236:~59824~
    000010b0.000012d8::2010/08/26-20:43:54.067 DBG   [SM] SrvCtxt initialized with package Kerberos, MaxTokenSize = 12000, RequiredCtxAttrib = 165910, HandShakeTimeout = 30000
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [SM] Handling auth handshake posted by thread id 4824
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [SM] Sponsor: Received SSPI blob from the Joiner of size 1411
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [SM] Sponsor: SSPI ASC returned status = 590610
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [SM] Sponsor: Sending SSPI blob of size 96 to Joiner
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [SM] Sponsor: Received SSPI blob from the Joiner of size 1411
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [SM] Sponsor: SSPI ASC returned status = 590610
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [SM] Sponsor: Sending SSPI blob of size 96 to Joiner
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [SM] Sponsor: Received SSPI blob from the Joiner of size 12000
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [SM] Sponsor: SSPI ASC returned status = 590610
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [SM] Sponsor: Sending SSPI blob of size 115 to Joiner
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [CHANNEL 192.168.210.236:~59824~] Close().
    000010b0.00001108::2010/08/26-20:43:54.067 INFO  [CHANNEL 192.168.210.236:~59824~] graceful close, status (of previous failure, may not indicate problem) ERROR_SUCCESS(0)
    000010b0.00001108::2010/08/26-20:43:54.067 ERR   [SM] Caught GracefulClose(1226)' because of 'channel to remote endpoint 192.168.210.236:~59824~ is closed' authenticating.
    000010b0.000012d8::2010/08/26-20:43:54.067 INFO  [SV] Authentication and authorization were successful
    000010b0.000012d8::2010/08/26-20:43:54.067 DBG   [SM] SrvCtxt initialized with package Kerberos, MaxTokenSize = 12000, RequiredCtxAttrib = 133122, HandShakeTimeout = 30000
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [SM] Handling auth handshake posted by thread id 4824
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [CHANNEL 192.168.210.236:~59824~]/recv: Attempting to perform I/O on closed stream.
    000010b0.00001108::2010/08/26-20:43:54.067 DBG   [CHANNEL 192.168.210.236:~59824~] Not closing handle because it is invalid.
    000010b0.00001108::2010/08/26-20:43:54.067 INFO  [CHANNEL 192.168.210.236:~59824~] graceful close, status (of previous failure, may not indicate problem) ERROR_SUCCESS(0)
    000010b0.00001108::2010/08/26-20:43:54.067 ERR   [SM] Caught GracefulClose(1226)' because of 'channel to remote endpoint 192.168.210.236:~59824~ is closed' authenticating.
    000010b0.000012d8::2010/08/26-20:43:54.067 DBG   [CHANNEL 192.168.210.236:~59824~] Not closing handle because it is invalid.
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  [CS] Service Stopping...
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  [CORE] Node quorum state is 'Not yet formed or joined a cluster'. Form/join status with other nodes is as follows:
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  [NODE] Node 1: Farthest reported progress joining with node iQCXSQL2 (id 2) is: Authenticate Initial Connection at time 2010/08/26-20:43:54.004: status 802 [SV] Authentication failed
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  [CORE] Node 1: Proposed View is <ViewChanged joiners=() downers=(1) newView=101() oldView=001(1) joiner=false form=false/>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  [RGP] Node 1: Stable_`0 => Opening`1
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  <class mscs::detail::ConsensusMessage>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <senderId> 1 </senderId>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <bestEpochSeen> 0 </bestEpochSeen>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <lastStableView> 001(1) </lastStableView>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <proposedView> 101() </proposedView>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <stage> Opening`1 </stage>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <inStage> () </inStage>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <inStagePrev> () </inStagePrev>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <joiners> () </joiners>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <trimmedNodes> (1) </trimmedNodes>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <innerScreen> () </innerScreen>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <pruningResult> () </pruningResult>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <matrix>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO      <ConnectivityMatrix>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO        <row id='0'00000000000000000000000000000000/>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO      </ConnectivityMatrix>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    </matrix>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO    <gemBlob> <counted_ptr p = 'nullptr'/> </gemBlob>
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  </class mscs::detail::ConsensusMessage>
    000010b0.000006fc::2010/08/26-20:44:52.906 INFO  [RGP] Node 1: Timer Tick Started
    000010b0.0000106c::2010/08/26-20:44:52.906 DBG   [NODE] Node 1: eating message sent to the dead node 2
    000010b0.0000106c::2010/08/26-20:44:52.906 INFO  [RGP] Node 1: Just sent farewell regroup message
    000010b0.000010b8::2010/08/26-20:44:52.906 DBG   [NODE] Node 1: just about to send a message of size 0 to 1
    000010b0.000010b8::2010/08/26-20:44:52.906 DBG   [NODE] Node 1: message to node 1 sent
    000010b0.000010b8::2010/08/26-20:44:52.906 DBG   [NODE] Node 1: just about to send a message of size 361 to 1
    000010b0.000010b8::2010/08/26-20:44:52.906 DBG   [NODE] Node 1: message to node 1 sent
    000010b0.000006fc::2010/08/26-20:44:53.218 INFO  [RGP] Node 1: Timer Tick Started
    000010b0.0000106c::2010/08/26-20:44:53.420 INFO  [DM]: Shutting down, so unloading the cluster database.
    000010b0.0000106c::2010/08/26-20:44:53.420 INFO  [DM] Shutting down, so unloading the cluster database (waitForLock: true).
    000010b0.0000106c::2010/08/26-20:44:53.420 DBG   [DM] Unloading Hive, Key \Registry\Machine\Cluster, discardCurrentChanges true
    000010b0.0000106c::2010/08/26-20:44:53.420 INFO  [CS] Service Stopped...
    000010b0.0000106c::2010/08/26-20:44:53.420 INFO  [CS] About to exit service...

  • Server 2012 Failover cluster. Make two VMs stay on the same node

    We have a unique situation where i need two machines to stay on the same node. Its a 4 node cluster with 30+ resources but i want to make sure two boxes are ALWAYS on the same now. If one migrates to another node the second needs to follow. Is there
    a way to do this? 

    How an this KB help to stay the two vm's on the same node.
    With all do respect @justinv how could this helped you to your problem , your question was "We have a unique situation where i need two machines to stay on the same node. Its a 4 node cluster with 30+ resources but i want to make sure two boxes are ALWAYS
    on the same now"
    and the KB that elden showed you is for : "Failover clusters that are running inside of virtual machines (sometimes referred to as “guest clusters”) may have problems with nodes joining the cluster."
    @justinv cloud you tell us more about this did I misunderstand your question ?
    Greetings, Robert Smit Follow me @clustermvp http://robertsmit.wordpress.com/ “Please click "Vote As Helpful" if it is helpful for you and Proposed As Answer” Please remember to click “Mark as Answer” on the post that helps you
    I explained in one of my replies that my underlying issue was this exactly what the KB fixed..... A guest cluster failing when moved to different nodes. Thats the only reason why i wanted them on the same node to begin with. While this post didn't solve me
    original question is solved what my real problem was....

  • Switching resource group in 2 node cluster fails

    hi,
    i configured a 2 node cluster to provide high availability for my oracle DB 9.2.0.7
    i have created a resource and named it oracleha-rg,
    and i crated later the following resources
    oraclelh-rs for logical hostname
    hastp-rs for the HA storage resource
    oracle-server-rs for oracle resource
    and listener-rs for listener
    whenever i try to switch the resource group between nodes is gives me the following in dmesg:
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 224900 daemon.notice] launching method <hafoip_stop> for resource <oraclelh-rs>, resource group <oracleha-rg>, node <DB1>, timeout <300> seconds+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 784560 daemon.notice] resource oraclelh-rs status on node DB1 change to R_FM_UNKNOWN+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 922363 daemon.notice] resource oraclelh-rs status msg on node DB1 change to <Stopping>+
    +Feb  6 16:17:49 DB1 ip: [ID 678092 kern.notice] TCP_IOC_ABORT_CONN: local = 010.050.033.009:0, remote = 000.000.000.000:0, start = -2, end = 6+
    +Feb  6 16:17:49 DB1 ip: [ID 302654 kern.notice] TCP_IOC_ABORT_CONN: aborted 0 connection+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 784560 daemon.notice] resource oraclelh-rs status on node DB1 change to R_FM_OFFLINE+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 922363 daemon.notice] resource oraclelh-rs status msg on node DB1 change to <LogicalHostname offline.>+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 515159 daemon.notice] method <hafoip_stop> completed successfully for resource <oraclelh-rs>, resource group <oracleha-rg>, node <DB1>, time used: 0% of timeout <300 seconds>+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 443746 daemon.notice] resource oraclelh-rs state on node DB1 change to R_OFFLINE+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 224900 daemon.notice] launching method <hastorageplus_postnet_stop> for resource <hastp-rs>, resource group <oracleha-rg>, node <DB1>, timeout <1800> seconds+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 784560 daemon.notice] resource hastp-rs status on node DB1 change to R_FM_UNKNOWN+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 922363 daemon.notice] resource hastp-rs status msg on node DB1 change to <Stopping>+
    +Feb  6 16:17:49 DB1 SC[,SUNW.HAStoragePlus:8,oracleha-rg,hastp-rs,hastorageplus_postnet_stop]: [ID 843127 daemon.warning] Extension properties FilesystemMountPoints and GlobalDevicePaths and Zpools are empty.+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 515159 daemon.notice] method <hastorageplus_postnet_stop> completed successfully for resource <hastp-rs>, resource group <oracleha-rg>, node <DB1>, time used: 0% of timeout <1800 seconds>+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 443746 daemon.notice] resource hastp-rs state on node DB1 change to R_OFFLINE+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 784560 daemon.notice] resource hastp-rs status on node DB1 change to R_FM_OFFLINE+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 922363 daemon.notice] resource hastp-rs status msg on node DB1 change to <>+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 529407 daemon.error] resource group oracleha-rg state on node DB1 change to RG_OFFLINE_START_FAILED+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 529407 daemon.notice] resource group oracleha-rg state on node DB1 change to RG_OFFLINE+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 447451 daemon.notice] Not attempting to start resource group <oracleha-rg> on node <DB1> because this resource group has already failed to start on this node 2 or more times in the past 3600 seconds+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 447451 daemon.notice] Not attempting to start resource group <oracleha-rg> on node <DB2> because this resource group has already failed to start on this node 2 or more times in the past 3600 seconds+
    +Feb  6 16:17:49 DB1 Cluster.RGM.global.rgmd: [ID 674214 daemon.notice] rebalance: no primary node is currently found for resource group <oracleha-rg>.+
    +Feb  6 16:19:08 DB1 Cluster.RGM.global.rgmd: [ID 603096 daemon.notice] resource hastp-rs disabled.+
    +Feb  6 16:19:17 DB1 Cluster.RGM.global.rgmd: [ID 603096 daemon.notice] resource oraclelh-rs disabled.+
    +Feb  6 16:19:22 DB1 Cluster.RGM.global.rgmd: [ID 603096 daemon.notice] resource oracle-rs disabled.+
    +Feb  6 16:19:27 DB1 Cluster.RGM.global.rgmd: [ID 603096 daemon.notice] resource listener-rs disabled.+
    +Feb  6 16:19:51 DB1 Cluster.RGM.global.rgmd: [ID 529407 daemon.notice] resource group oracleha-rg state on node DB1 change to RG_OFF_PENDING_METHODS+
    +Feb  6 16:19:51 DB1 Cluster.RGM.global.rgmd: [ID 529407 daemon.notice] resource group oracleha-rg state on node DB2 change to RG_OFF_PENDING_METHODS+
    +Feb  6 16:19:51 DB1 Cluster.RGM.global.rgmd: [ID 224900 daemon.notice] launching method <bin/oracle_listener_fini> for resource <listener-rs>, resource group <oracleha-rg>, node <DB1>, timeout <30> seconds+
    +Feb  6 16:19:51 DB1 Cluster.RGM.global.rgmd: [ID 515159 daemon.notice] method <bin/oracle_listener_fini> completed successfully for resource <listener-rs>, resource group <oracleha-rg>, node <DB1>, time used: 0% of timeout <30 seconds>+
    +Feb  6 16:19:51 DB1 Cluster.RGM.global.rgmd: [ID 529407 daemon.notice] resource group oracleha-rg state on node DB1 change to RG_OFFLINE+
    +Feb  6 16:19:51 DB1 Cluster.RGM.global.rgmd: [ID 529407 daemon.notice] resource group oracleha-rg state on node DB2 change to RG_OFFLINE+
    and the resource group fails to switch...
    any help please?

    Hi,
    this forum is for Oracle Clusterware, not Solaris Cluster. You probably should close this thread and open your question in the corresponding Solaris Cluster forum, to get help.
    Regards
    Sebastian

  • Sun Cluster Core Conflict - on SUN Java install

    Hi
    We had a prototype cluster that we were playing with over two nodes.
    We decided to uninstall the cluster by putting node into single user mode and running scinstall -r.
    Afterwards we found that the Java Availability Suite was a little messed up - maybe because the kernel/registry had not been updated - it though the cluster and agent software was uninstalled and would not let us re-install. All the executabvles from /etc/cluster/bin had been removed from the nodes.
    So, On both nodes we ran the uninstall program from /var/sadm/prod/... and then selected cluster and agents to uninstall.
    On the first node, this completely removed the sun cluster compoenets and then allowed us to re-install the cluster software successfully.
    On the second node, for some reason, it has left behind the component "Sun Cluster Core", and will not allow us to remove it with the uninstall.
    When we try to re-install we get the following:
    "Conflict - incomplete version of Sun Cluster Core has been detected"
    In then points us to the sun cluster upgrade guide on sun.com.
    My question is - how do we 'clean up' this node and remove the sun cluster core so we can re-install the sun cluster software from scratch?
    I don't quite understand how this has been left behind....
    thanks in advance
    S1black.

    You can use prodreg directly to clean up when your de-install has gone bad.
    Use:
    # prodreg browse
    to list the products. You may need to recurse down into the individual items. The use:
    # prodreg unregister ...
    to unregister and pkgrm to remove the packages manually.
    That has worked for me in the past. Not sure if it is the 'official' way though!
    Regards,
    Tim
    ---

  • Cluster.exe  RESOURCE get error like 1214

    Dear Team,
    we have faced below issue when we install the ECC on windows cluster. Can you please suggest how may i resolve the error. please find my screen shot for your further clarifications.
    Execution of the command "cluster.exe RESOURCE "SAP IEP NetName" /PRIV
    Name= 10.23.18.67" finished with return code 1214. Output:

    Hi,
    This is SAP business one system administration forum. Please find correct forum and repost above discussion to get quick assistance.
    Please close this thread here with helpful answer.
    Thanks & Regards,
    Nagarajan

  • 1068 Error when bringing cluster name resource online

    Hello Guys, 
    I have configured 3 node cluster setup on windows 2012R2 machines. When I try bring the networker name resource online, I got following error and name resource failed become online 
    "Network Name resource 'Cluster Name' (with associated network name 'DEDUPCSV') has Kerberos Authentication support enabled. Failed to add required credentials to the LSA - the associated error code is '1068'."
    Can anyone please helpme to resolve this problem. 
    Thanks,

    Hi Ajith G,
    Could you confirm every node the “Network List Service” service is running or not, if this service stopped please start it, after you start this service if the issue still
    exist please run your cluster validation then post the warning and error part.
    More information:
    Event ID 1068 — File Share Resource Availability
    http://technet.microsoft.com/en-us/library/cc756219(v=ws.10).aspx
    Hope this helps.
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Is it possible to Migrate Live VMs from a Windows 2012 Hyper-V Cluster to a different 2012 R2 Cluster?

    At the moment I'm in a bit of dilemma because I know that Windows 2012 supports "Shared Nothing Live Migration" but recently I got to know that this feature is available for stand alone Hyper-V Servers. My Setup is that I have
    3 Servers running Windows Server 2012 Hyper-v Failover Cluster and I need to migrate everything to a new cluster running on Windows Server 2012 R2 and offcourse some VMs can't be turned off during migration so I need to do this live during production hours.
    Another note is that both clusters will be running on the different LUNs since each cluster has its own CSVs. I need to know if it is possible to migrate such VMs/Roles Live without have any downtime of the VMs/Roles?

    Yes, but one exception - live migration works only on "compatible cpu's". There actually is a setting in CPU for forcing compatibility
    mode. If that is not set, and you mvoe for example from AMD to Intel, then - this is not possible in live migration, cluster or not.

  • MSMQ Cluster in Windows 2008R2

    Did anyone have any issues on Clustering MSMQ on windows 2008?My team says theyr'e missing queues while setting up as cluster, and they decided to go  local MSMQ .
    Thanks

    Hello SubodhSaxena.
    I suggest you to ask this question in the Windows Server General Forum at
    http://social.technet.microsoft.com/Forums/en-US/home?forum=winservergen; I can move it there for you if you want.
    Bye.
    Luigi Bruno
    MCP, MCTS, MOS, MTA

Maybe you are looking for

  • How to print both letter and legal size docs at same time with two trays

    I have an HP Laser Jet P2055dn printer. How can I print both letter and legal docs at same time?

  • Managed bean can not be created. - dynamic menubar using ADF

    hi All, had to post the thread again to make sure it dint go un noticed! scenario: making dynamic menu navigation using ADFFaces, all steps followed as mentioned in SRDemo Sample. but still i get this error: error encountered: SEVERE:Managedbean menu

  • Multiple IQ databases on one host ---- NLS for BW-HANA

    Hi,  We are planning to build a Sybase IQ NLS landscape to support BW on HANA landscapes. We have thought of installing DEV/QA NLS (IQ) on one server and accordingly have created the file systems on the Linux host. My question is whether we have to i

  • Layer id to component id in Flex

    For designers is imperative to give individual names to layers, the thing is that these "names" in Flex continue as layer ids instead of being used as component ids (which would cut ALOT of work!!!). Please! Change this!!!!!

  • Function does not have a body

    Hello, i have a problem running this program. an error "function does not have a body" pops up, do you guys know how i can fix this? stop(); import flash.events.MouseEvent start_.addEventListener(MouseEvent.CLICK, onClick); function onClick(event:Mou