Windows NLB for SPF high availability

I have SPF in my organization as a bridge between VMM and WAP.  One of my design tenants is to configure everything in a high availability configuration and I am having trouble getting SPF to work in an HA environment.
If I set the binding on SPFs website to the NLB name and use the certificate that I have for it, WAP is unable to write anything to the SPF database.  If I set SPF back to no binding and use the certificate that is created for the machine itself, WAP
works exactly as it should.
The NLB that I am using is windows NLB and I have it configured in unicast /w all ports being forwarded to the SPF server.  I have just one NIC per server on my SPF instance. 
I understand that SPF is supported and have configured it using this guide... (link: http://www.hyper-v.nu/archives/mvaneijk/2013/06/system-center-2012-sp1-service-provider-foundation-high-availability/) It works from the server to the VMM server (using
the web test as prescribed) but gets an error 12 that references the provider section of the site when I try to create the SPF connection from WAP.
Does anybody have any ideas? 

thank you for the prompt reply.  The certificates that I am using are created by my enterprise CA which is automatically trusted by my WAP servers.  The servers all get a client/server authentication certificate as part of AD's auto-enrollment. 
I have created a custom certificate with just the name of the NLB.  Do i need to specify the actual server names of the SPF environment as SAN's in the certificate request?  Could that be the problem? 
Single server SPF uses a certificate that is trusted with the single server name
NLB environment uses a custom request certificate that has client/server EKU and the only name that is registered is the DNS name of the NLB's IP address.  This is the same name that I put in the binding for the spf website on my SPF servers. 
(this method is broken)

Similar Messages

  • Generating license for ISE high availability primary/secondary nodes

    We have two ISE servers that will act as primary/secondary in a high availability setup.
    The ISE 1.0.4 installation guide, page 93, mentions that "If you have two Cisco ISE nodes configured for high availability, then you must include both the primary and secondary Administration ISE node hardware and IDs in the license file."
    However, after entering the PAK in the licensing page, the only required fields are:
    - Primary Product ID
    - Primary Version ID
    - Primary Serial No
    In this case, how can i include both primary and secondry HW and IDs?
    Thanks in advance.

    I am refering you a Cisco ISE Nodes for High Availability configuration guide, Please check:
    http://www.cisco.com/en/US/docs/security/ise/1.0/user_guide/ise10_dis_deploy.html#wp1128454

  • I'm looking for Failover/High available solutions for IPS 4200 Series

    Hi all,
    I tried to find out Failover/High available solutions for IPS 4200 series,I didn't saw failover solutions in IPS guide document. Anybody can be help me!

    I do not know if this is documented anywhere, but I can tell you what I do. As long as the IPS 4200 has power, with the right software settings, the unit can fail such that it will pass traffic. Should the unit loose power, it does stop all traffic. I run a patch cable in parallel with the in line IPS unit, in the same VLAN, with a higher STP cost. Thus all traffic will traverse the IPS unit when possible, but should something happen to it, a $10 patch cable takes over.
    Mike

  • Migrating from 2010 windows nlb cas array to exchange 2013 using netscaler for HLB

    i currently have exchange 2010 sp3 setup as follows
    2 cas/ht using windows nlb for array.internal.com as the nlb name
    2 mailbox servers using 1 dag for replication between them.
    I would like to stand up 2 new CAS 2013 sp1 servers(2012 r2) and use our netscaler hlb to load balance and do ssl offloading.
    id like to use the hlb to load balance and ssl offload all possible traffic not just owa, i.e. activesync etc.
    the netscaler is running version 10.5.
    does anyone have any thoughts on how to perform such a migration?
    id like to make this as seamless as possible for the user, so no owa name change or anything like that.
    thx in advance for any help.

    Exchange 2013 CAS cannot be managed effectively without an Exchange 2013 mailbox server since the CAS won't run the Exchange 2013 management tools without it.  Your Exchange 2013 servers should simply have both roles.  There are few cases where
    splitting the roles has any business value.
    If what you're planning to do is within the same forest and organization, then it's a "transition", not a "migration".  What you are asking to do is easy.  Build the Exchange 2013 servers, configure load balancing, test
    access, and then switch the DNS records to point to them.
    Personally, I would skip the SSL offloading.  It's my opinion that it makes the solution more difficult to troubleshoot without providing any real benefit.
    Ed Crowley MVP "There are seldom good technological solutions to behavioral problems."

  • Question on replication/high availability designs

    We're currently trying to work out a design for a high-availability system using Oracle 9i release 2. Having gone through some of the Oracle whitepapers, it appears that the ideal architecture involves setting up 2 RAC sites using Dataguard to synchronize the data. However, due to time and financial constraints, we are only allowed to have 2 servers for hosting the databases, which are geographically separate from each other in prevention of natural disasters. Our app servers will use JDBC pools to connect to the databases.
    Our goal is to have both databases be the mirror image of each other at any given time, and the database must be working 24/7. We do have a primary and a secondary distinction between the two, so if the primary fails, we would like the secondary database to take over the tasks as needed.
    The ability to query existing data is mission critical. The ability to write/update the database is less important, however we do need the secondary to be able to process data input/updates when primary is down for a prolonged period of time, and have the ability to synchronize back with the primary site when it is back up again.
    My question now is which replication technology should we try to implement? I've looked into both Oracle Advanced Replication and Dataguard, each seems to have its own advantages and drawbacks:
    Replication - can easily switch between the two databases using multimaster implementation, however data recovery/synchronization may be difficult in case of failure, and possibly will lose data (pending implementation). There has been a few posts in this forum that suggested that replication should not really be considered as an option for high availability, why is that?
    Dataguard - zero data loss in failover/switchover, however manual intervention is required to initiate failover/switchover. Once the primary site fails over to the standby, the standby becomes the primary until DBA manually goes back in and switch the roles. In Oracle 10g release 2, seems that automatic failover is achieved through the use of an extra observer piece. There does not seem to be anyway to do this in Oracle 9i release 2.
    Being new to the implementation of high-availability systems, I am at somewhat of a loss at this point. Both implementations seem to be a possible candidate, but we will need to sacrifice some efforts for both of them also. Would anyone shine some light on this, maybe point out my misconceptions with Advanced Replication and Dataguard, and/or suggest a better architecture/technology to use? Any input is greatly appreciated, thanks in advance.
    Sincerely,
    Peter Tung

    Hi,
    It sounds as if you're talking about the DB_TXN_NOSYNC flag, rather than DB_NOSYNC.
    You mention that in general, you lose uncommitted transactions on system failure. I think what you mean is that you may lose some committed transactions on system failure. This is correct.
    It is also correct that if you use replication you can arrange to have clients have a copy of all committed transactions, so that if the master fails (and enough clients do not fail, of course) then the clients still have the transaction data, even when using DB_TXN_NOSYNC.
    This is a very common usage scenario for Berkeley DB replication/HA, used to achieve high throughput. You will want to pay attention to the configured ack policy, group size setting, setting of the 2SITE_STRICT option (if group size == 2).

  • Cisco Secure ACS and Windows NLB

    Hi,
    I have two ACS servers and have been trying unsuccessfully to setup Windows NLB for them. I can successful setup the NLB but ACS won't respond on the clustered IP. Other services running on the clustered IP will respond so I believe the NLB is working correctly.
    Has anyone had any success with ACS and Microsoft NLB? I can?t find any documentation to suggest that they are incompatible but I think this may be the case.
    Thanks,
    Neil

    Neil,
    ACS is not tested with NLB but if cluster hosts are attempting to communicate with the ACS using their clustered IP then ACS should reply.
    Do you see any hits on acs ? If you sniff the acs interface, what is the source IP address ? Is it clustered ip or clustered host IP ??
    Also on acs --->Network configuration add aaa client with host IP and clustered ip . Now see if acs responds to NLB.
    Regards,
    ~JG

  • TREX High availability

    Hi,
    We used virtual host name for our production TREX server. We used the instruction as per the installation guide. It is up and running now. I am setting up the TREX high availability in a different server using Veritas cluster service. The production node is down and the HA system is now mounted with the production file system. Is there any instruction on how to setup TREX HA?
    Thanks,
    RT

    Hi RT,
    I do not know about the context "Veritas Cluster".
    Otherwise, for TREX high availability, please check:
    [http://help.sap.com/saphelp_nw2004s/helpdata/EN/46/c1aec67b4f1826e10000000a1553f6/frameset.htm|http://help.sap.com/saphelp_nw2004s/helpdata/EN/46/c1aec67b4f1826e10000000a1553f6/frameset.htm]
    Best, Karsten

  • SAP POS - Xpress Server High availability

    Dear All,
    I have a requirement for Installing "High Availability" for SAP POS XPS server (Store server), Transnet server, & Transnet client.
    I have gone thru the possibilities of having Backup server for XPS store server, but the backup server supported by SAP has some major restrictions listed below.
    1. Store closing
    2. Terminal Closing
    3. EOD ( End of the day)
    4. Exchange with reciept
    5. Layaway sale.
    Please let me know whether their is a possibility for achieving high availability for store server which performs the above listed transactions too.
    Appreciate if any body can reply asap.
    Best regards
    Syed

    Hello Syed,
    Having High Availability for an Xpress is not standards SAP implementation and is quite u201Coverkillu201D. The use of the backup server is designed to allow the store to function and service itu2019s customers until a server can be rebuilt or repaired.
    BTW:
    -Layaways can be done in offline mode and it will post to the server when it becomes available
    -Returns can be done on the backup server if the customer implemented Returns Authorization

  • Highly available sendmail

    I note that the SC3.1 Release Notes state:
    Sun Cluster software currently does not have an HA Data Service for the sendmail(1M) subsystem. It is permitted to run sendmail on the individual cluster nodes, but the sendmail functionality will not be highly available, including the functionality of mail delivery and mail routing, queuing, or retry.
    Can a data service for sendmail be created, or is there something inherent in either sendmail or SC that prevents this? If so, what solutions exist for a highly available sendmail configuration that will not leave any queued messages in limbo?
    Thanks,
    Brian

    >
    Can a data service for sendmail be created, or is
    there something inherent in either sendmail or SC that
    prevents this? If so, what solutions exist for a
    highly available sendmail configuration that will not
    leave any queued messages in limbo?You create a redundant sendmail service by setting up several of them.
    DNS MX records ( Mail eXchange ) provides fallback routes and even load sharing
    between several sendmail servers. it is far too expensive to build a sendmail
    cluster service when you can solve the problem with two or three V100 ultrasparc II
    servers.
    //Lars

  • AGPM High Availability / Fault Tolerance

    I’m looking for any informations about Advanced Group Policy Management (AGPM) High Availability / Fault Tolerance.
    The most interesting is that, in introduction for AGPM Planning Guide is described section about
    Planning for AGPM high-availability and improved fault tolerance, but content not exists.
    Microsoft MVP: Virtual Machine | http://blog.porowski.pro |
    http://VirtualizationGeek.PL

    Hi,
    You may type the key words into the bing of TechNet website to search for the information you need. According to my search,
    the following two links inlcude the detailed information regarding AGPM. But less information regarding AGPM High Availability / Fault Tolerance can be found from TechNet website.
    Advanced Group Policy Management 4.0 Documents
     http://www.microsoft.com/downloads/en/details.aspx?FamilyID=58FD6CDA-0210-4DA0-91D2-8C3CC2817DF8&displaylang=en 
    Advanced Group Policy Management
     http://technet.microsoft.com/en-us/library/dd420466.aspx 
    Based on the current situation, I will also submit a feedback to Microsoft through our internal channel and requst more contents regarding
    AGPM High Availability / Fault Tolerance to be published. While we cannot guarantee that these contents will be added, we would like to ensure you that we will make every effort to see that they do.
    Arthur Li
     TechNet Subscriber Support 
    in forum
    If you have any feedback on our support, please contact
     [email protected] . 
    Please remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading the thread.

  • Windows Event Collector - Built-in options for load balancing and high availability ?

    Hello,
    I have a working collector. config is source initiated, and pushed by GPO.
    I would like to deploy a second collector for high availability and load balancing. What are the available options ? I have not found any guidance on TechNet articles.
    As a low cost option, is it fine to simply start using DNS round-robin with a common alias for both servers pushed as a collector name through GPO ?
    In my GPO Policy, if I individually declare both servers, events are forwarded twice, once for each server. Indeed it does cover high availability, but not really optimized.
    Thanks for your help.

    Hi,
    >>As a low cost option, is it fine to simply start using DNS round-robin with a common alias for both servers pushed as a collector name through GPO ?
    Based on the description, we can utilize DNS round robin to distribute workloads and increase fault tolerance. By default, DNS uses round robin to rotate the order of RR data returned in query answers where multiple RRs of the same type exist for a queried
    DNS domain name. This feature provides a simple method for load balancing client use of Web servers and other frequently queried multihomed computers. Besides, by default, DNS will perform round-robin rotation for all RR types.
    Regarding DNS round robin, the following article can be referred to for more information.
    Configuring round robin
    http://technet.microsoft.com/en-us/library/cc787484(v=ws.10).aspx
    TechNet Subscriber Support
    If you are TechNet Subscription user and have any feedback on our support quality, please send your feedback here.
    Best regards,
    Frank Shen

  • SharePoint Foundation 2013 - Can we use the foundation for intranet portal with high availability ( medium farm)

    Today I had requirement, where we have to use the SharePoint Foundation 2013 (free version) to build an intranet portal ( basic announcement , calendar , department site , document management - only check-in check-out / Version).
     Please help me regarding the license  and size limitations. ( I know the feature comparison of Standard / Enterprise) I just want to know only about the installation process and license.
    6 Server - 2 App / 2 Web / 2 DB cluster ( so total license 6 windows OS license , 2 SQL Server license and Guess no sharepoint licenes)

    Thanks Trevor,
    Is load balance service also comes in free license... So, in that case I can use SharePoint Foundation 2013 version for building a simple Intranet & DMS ( with limited functionality).  And for Workflow and content management we have to write code.
    Windows Network Load Balancing (the NLB feature) is included as part of Windows Server and would offer high availability for traffic bound to the SharePoint servers. WNLB can only associate with up to 4 servers.
    Trevor Seward
    Follow or contact me at...
    &nbsp&nbsp
    This post is my own opinion and does not necessarily reflect the opinion or view of Microsoft, its employees, or other MVPs.

  • Windows 2012 RDS - Session Host servers High Availability

    Hello Windows/Terminal server Champs,
    I am new middle of implementing RDS environment for one of my customer, Hope you could help me out.
    My customer has asked for HA for RDS session host where applications are published, and i have prepared below plan for server point of view.
     2 Session Host server, 1 webaccess, 1 License/connection
    Broker & 1 Gateway (DMZ).
     In first Phase, we are planning to target internal user
    who connect to Session host HA where these 2 servers will have application installed and internal user will use RDP to access these application.
    In second Phase we will be dealing with external Party who connect from external network where we are planning to integrate with NetIQ => gateway
    => Webaccess/Session host
     I have successfully installed and configured 2 Session
    Host, 1 license/Broker. 1 webAccess & 1 Gateway. But my main concern to have session Host High Available as it is hosting the application and most of the internal user going to use it. to configure it i am following http://technet.microsoft.com/en-us/library/cc753891.aspx  
    However most of the Architecture is change in RDS 2012. can you please help me out to setup the Session Host HA.
    Note: we can have only 1 Connection broker /Licensing server , 1 webacess server & 1 Gateway server, we cannot increase more server due to cost
    factor.
     thanks in advance.

    Yes, absolutely no problem in just using one connection broker in your environment as long as your customer understands the SPOF.
    the session hosts however aren't really what you would class HA - but to set them up so youhave reduancy you would use either Windows NLB, an external NLB device or windows dns round robin. My preferred option when using the connection broker is DNS round
    robin - where you give each server in the farm the same farm name dns entry - the connection broker then decides which server to allocate the session too.
    You must ensure your session host servers are identical in terms of software though - same software installed in the same paths on all the session host servers.
    if you use the 2012 deployment wizard through server manager roles the majority of the config is done for you.
    Regards,
    Denis Cooper
    MCITP EA - MCT
    Help keep the forums tidy, if this has helped please mark it as an answer
    My Blog
    LinkedIn:

  • 11.1.2 High Availability for HSS LCM

    Hello All,
    Has anyone out there successfully exported an LCM artifact (Native Users) through Shared Services console to the OS file system (<oracle_home>\user_projects\epmsystem1\import_export) when Shared Services is High Available through a load balancer?
    My current configuration is two load balanced Shared Services servers using the same HSS repository. Everything works perfectly everywhere except when I try to export anything through LCM from the HSS web console. I have shared out the import_export folder on server 1. And I have edited the oracle_home>\user_projects\epmsystem1\config\FoundationServices1\migration.properties file on the second server to point to the share on the first server. Here is a cut and paste of the migration.properties file from server2.
    grouping.size=100
    grouping.size_unknown_artifact_count=10000
    grouping.group_by_type=Y
    report.enabled=Y
    report.folder_path=<epm_oracle_instance>/diagnostics/logs/migration/reports
    fileSystem.friendlyNames=false
    msr.queue.size=200
    msr.queue.waittime=60
    group.count=10000
    double-encoding=true
    export.group.count = 30
    import.group.count = 10000
    filesystem.artifact.path=\\server1\import_export
    I perform an export of just the native users to the file system, the export fails stating and I find errors in the log stating the following.
    [2011-03-29T20:20:45.405-07:00] [EPMLCM] [NOTIFICATION] [EPMLCM-11001] [oracle.EPMLCM] [tid: 10] [ecid: 0000Iw4ll2fApIOpqg4EyY1D^e6D000000,0] [arg: /server1/import_export/msr/PKG_34.xml] Executing package file - /server1/import_export/msr/PKG_34.xml
    [2011-03-29T20:20:45.530-07:00] [EPMLCM] [ERROR] [EPMLCM-12097] [oracle.EPMLCM] [tid: 10] [ecid: 0000Iw4ll2fApIOpqg4EyY1D^e6D000000,0] [SRC_CLASS: com.hyperion.lcm.clu.CLUProcessor] [SRC_METHOD: execute:?] Cannot find the migration definition file in the specified file path - /server1/import_export/msr/PKG_34.xml.
    [2011-03-29T20:20:45.546-07:00] [EPMLCM] [NOTIFICATION] [EPMLCM-11000] [oracle.EPMLCM] [tid: 10] [ecid: 0000Iw4ll2fApIOpqg4EyY1D^e6D000000,0] [arg: Completed with failures.] Migration Status - Completed with failures.
    I go and look for the path that it is searching for "/server1/import_export/msr/PKG_34.xml" and find that this file does exist, it was infact created by the export process so I know that it is able to find the correct location, but it then says that it can't find it after the fact. I have tried creating mapped drives and editing the migration.properties file to reference the mapped drive but it gives me the same errors but with the new path. I also tried the below article that I found in support with no result. Please advise, thank you.
    Bug 11683769: LCM MIGRATIONS FAIL WHEN HSS IS SET UP FOR HA AND HSS IS STARTED AS A SERVICE
    Bug Attributes
    Type B - Defect Fixed in Product Version 11.1.2.0.000PSE
    Severity 3 - Minimal Loss of Service Product Version 11.1.2.0.00
    Status 90 - Closed, Verified by Filer Platform 233 - Microsoft Windows x64 (64-bit)
    Created 25-Jan-2011 Platform Version 2003 R2
    Updated 24-Feb-2011 Base Bug 11696634
    Database Version 2005
    Affects Platforms Generic
    Product Source Oracle
    Related Products
    Line - Family -
    Area - Product 4482 - Hyperion Lifecycle Management
    Hdr: 11683769 2005 SHRDSRVC 11.1.2.0.00 PRODID-4482 PORTID-233 11696634Abstract: LCM MIGRATIONS FAIL WHEN HSS IS SET UP FOR HA AND HSS IS STARTED AS A SERVICE*** 01/25/11 05:15 pm ***Hyperion Foundation Services is set up for high availability. Lifecycle management migrations fail when running Hyperion Foundation Services - Managed Server as a Windows Service. We applied the following configuration changes:Set up shared disk then we modify the migration.properties.filesystem.artifact.path=\\<servername>\LCM\import_export We are able to get the migrations to work if we set filesystem.artifact.path=V:\import_export (a maped drive to the shared disk) and we run Hyperion Foundation Services in console mode.*** 01/25/11 05:24 pm *** *** 01/28/11 03:32 pm *** (CHG: Sta->11)*** 01/28/11 03:32 pm *** (CHG: Comp. Ver-> 11.1.2 -> 11.1.2.0.00)*** 01/28/11 03:34 pm *** (CHG: Base Bug-> NULL -> 11696634)*** 02/02/11 01:53 pm *** (CHG: Sta->41)*** 02/02/11 01:53 pm ****** 02/02/11 05:18 pm ****** 02/03/11 10:30 am ****** 02/04/11 03:53 pm *** *** 02/04/11 03:53 pm *** (CHG: Sta->80 Asg->NEW OWNER)*** 02/10/11 08:49 am ****** 02/15/11 11:59 am ****** 02/15/11 04:56 pm ****** 02/16/11 09:58 am *** (CHG: Sta->90)*** 02/16/11 09:58 am ****** 02/16/11 02:12 pm *** (CHG: Sta->93)*** 02/24/11 06:14 pm *** (CHG: Sta->90)Back to top

    it is not possible to implement a kind of HA between two different appliances 3315 and 3395. 
    A node in HA can have the 3 persona. 
    Suppose Node A (Admin(primary), monitoring(Primary) and PSN). 
    Node B(Admin(Secondary), Monitoring(Secondary) and PSN). 
    If the Node A is unavailable, you will have to promote manually the Admin role to Primary. 
    Although the best way is to have
    Node A Admin(Primary), Monitoring(Secondary) and PSN
    Node B Admin(Secondary), Monitoring (Primary) and PSN. 
    Rate if helpful and Marked As correct if it is correct for the experts. 
    Regards,

  • Hyper-V 2012 High Availability using Windows Server 2012 File Server Storage

    Hi Guys,
    Need your expertise regarding hyper-v high availability. We setup 2 hyper-v 2012 host in our infra for our domain consolidation project. Unfortunately, we don't have hardware storage that is said to be a requirement that we can use to create a failover cluster
    for hyper-v host to implement HA. Here's the setup:
    Host1
    HP Proliant L380 G7
    Windows Server 2012 Std
    Hyper-V role, Failover Cluster Manager and File and Storage Services installed
    Host2
    Dell PowerEdge 2950
    Windows Server 2012 Std
    Hyper-V role, Failover Cluster Manager and File and Storage Services installed
    Storage
    Dell PowerEdge 6800
    Windows Server 2012 Std
    File and Storage Services installed
    I'm able to configure the new feature about Shared Nothing Live Migration - i'm able to move VM's back and fort between my hosts without storage. But this is planned and proactive approach. My concern is to have my hyper-v host to become highly available in
    the event of system failure. If my host1 dies, the VMs should go/move to host2 and vice versa. In setting this up, i believe i need to enable the failover clustering between my hyper-v hosts which i already did but upon validation, it says "No disks
    were found on which to perform cluster validation tests." Is it possible to cluster it using just a regular windows file server? I've read about SMB 3.0 and i've configured it as well i'm able to save VMs on my file server, but i don't think that my hyper-v
    hosts are already highly available.
    Any feedback and suggestions or recommendation is highly appreciated. Thanks in advance!

    Hi Guys,
    Need your expertise regarding hyper-v high availability. We setup 2 hyper-v 2012 host in our infra for our domain consolidation project. Unfortunately, we don't have hardware storage that is said to be a requirement that we can use to create a failover cluster
    for hyper-v host to implement HA. Here's the setup:
    Host1
    HP Proliant L380 G7
    Windows Server 2012 Std
    Hyper-V role, Failover Cluster Manager and File and Storage Services installed
    Host2
    Dell PowerEdge 2950
    Windows Server 2012 Std
    Hyper-V role, Failover Cluster Manager and File and Storage Services installed
    Storage
    Dell PowerEdge 6800
    Windows Server 2012 Std
    File and Storage Services installed
    I'm able to configure the new feature about Shared Nothing Live Migration - i'm able to move VM's back and fort between my hosts without storage. But this is planned and proactive approach. My concern is to have my hyper-v host to become highly available in
    the event of system failure. If my host1 dies, the VMs should go/move to host2 and vice versa. In setting this up, i believe i need to enable the failover clustering between my hyper-v hosts which i already did but upon validation, it says "No disks were
    found on which to perform cluster validation tests." Is it possible to cluster it using just a regular windows file server? I've read about SMB 3.0 and i've configured it as well i'm able to save VMs on my file server, but i don't think that my hyper-v hosts
    are already highly available.
    Any feedback and suggestions or recommendation is highly appreciated. Thanks in advance!
    Your shared storage is a single point of failure with this scenario so I would not consider the whole setup as a production configuration... Also setup is both slow (all I/O is travelling down the wire to storage server, running VMs from DAS is ages faster)
    and expensive (third server + extra Windows license). I would think twice about what you do and either deploy a built-in VM replication technologies (Hyper-V Replica) and apps built-in clustering features that does not require shared storage (SQL Server and
    Database Mirroring for example, BTW what workload do you run?) or use some third-party software creating fault tolerant shared storage from DAS or investing into physical shared storage hardware (HA one of course). 
    Hi VR38DETT,
    Thanks for responding. The hosts will cater a domain controller (on each host), Web filtering software (Websense), Anti-Virus (McAfee ePO), WSUS and an Auditserver as of the moment. Is the Hyper-V Replica somewhat give "high availability" to VMs or Hyper-V
    hosts? Also, is the cluster required in order to implement it? Haven't tried that but worth a try.

Maybe you are looking for

  • PO Awaiting for approval

    Hi We are on SRM 5.0 , configuring Extended classic scenario. After SC is approved , even if I assign the vendor while creating the SC , it always creates a incomplete PO and it is in Held status. When Purchaser Logs in and Process the PO, add the ve

  • Dead Screen

    After the new 2.2 update I experience a whole new phone.....A completely black screen. Oh, the phone works as well as all the other features but you just can't see it. You can't dial out or check out any of the cool apps. I'm not sure if this is beca

  • Is Photoshop available for Mac?  Is it part of Office?

    IS Photoshop available for Mac?  I am told I must have it to submit a book to Amazon's Create Space. Is Photoshop included in Office?

  • Dynamic SQl Query Select

    Hello Colleague, I have a table that contains multiple entries of TABLE NAME. I need to query the databse table and check if certain entry exist in that table. I Loop over the table containing the tablenames and try a select : Loop at lt_key into ls_

  • Can we do Dml statements on materialize views

    I want to know can we do Dml statements on materialize views .If yes, how