RD farm in High Availability

Hi all,
I am in the process of bilding a RD farm using 3 Server 2008 VMs as RDSH.
This is a temp and may only be used for max. 3 months.
Can Session Broker and NLB installed on RDSH servers?  
MS article regarding NLB states that
The Windows Server 2008-based server where you install the TS Session Broker role service does not have to be a terminal server or have Remote Desktop enabled.
http://technet.microsoft.com/en-us/library/cc771300(v=ws.10).aspx
I'm assuming this means RDSH can bea Session Broker as well?
But not sure whether we need dedicated NLB or we can use one of the RDSH's?
Please advise.
Thanks!

Hi,
Thank you for posting in Windows Server Forum.
From your description I can say that for 3 servers to have load balancing we must have TS Session Broker installed which will act as load balancer. 
NLB distributes traffic across several servers by using the TCP/IP networking protocol. Terminal Services Session Broker (TS Session Broker) is a role service that enables you to load balance sessions between terminal servers in a farm, and allows a user
to reconnect to an existing session in a load-balanced terminal server farm. TS Session Broker stores session state information that includes session IDs, their associated user names, and the name of the server where each session resides.
We can also use DNS Round Robin method to load balance different terminal server. Please check below article for information.
TS Session Broker Load Balancing Step-by-Step Guide
http://technet.microsoft.com/en-us/library/cc772418.aspx
Hope it helps!
Thanks.
Dharmesh Solanki

Similar Messages

  • Office Web Apps farm - how to make it high available

    Hi there,
    I have deployed OWAs for Lync 2013 with two servers in one farm. When I was testing, e.g. the case when other server is down I found out the following error:
    Log Name:      Microsoft Office Web Apps
    Source:        Office Web Apps
    Date:          17/03/2014 22:59:54
    Event ID:      8111
    Level:         Error
    Computer:      OWASrv02.domain.com
    Description:
    A Word or PowerPoint front end failed to communicate with backend machine
    http://OWASrv01:809/pptc/Viewing.svc
    Does that means, that OWAs cannot setup as high avaiable when using standalone OWAs only? Above error appeared when the server OWASrv01 was rebooting.
    Petri

    Hi Petri,
    For your scenario, you have achieve a  high performance for your Office Web Apps farm but not high availability. 
    For performing a high availability Office Web Apps farm, you can refer to the blog:
    http://blogs.technet.com/b/meamcs/archive/2013/03/27/office-web-apps-2013-multi-servers-nlb-installation-and-deployment-for-sharepoint-2013-step-by-step-guide.aspx 
    Thanks,
    Eric
    Forum Support
    Please remember to mark the replies as answers
    if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact [email protected]
    Eric Tao
    TechNet Community Support

  • Farm High Availability

    we have 2 FE, 2 APP and 1 db cluster server.
    How to make high availability of Farm....Central Amdin is hosted on App1.....if APP1 is down how we can make available Central Admin....both servers are running identical services...but i m unable to run central admin service on both server?

    In the event your CA server goes down then you can run PSConfig and enable it on the second server.
    However if you want to do it right then this article by Spence Harbar is a bit old but still the best guide i know of for load balancing CA:
    http://www.harbar.net/articles/spca.aspx

  • SharePoint Foundation 2013 - Can we use the foundation for intranet portal with high availability ( medium farm)

    Today I had requirement, where we have to use the SharePoint Foundation 2013 (free version) to build an intranet portal ( basic announcement , calendar , department site , document management - only check-in check-out / Version).
     Please help me regarding the license  and size limitations. ( I know the feature comparison of Standard / Enterprise) I just want to know only about the installation process and license.
    6 Server - 2 App / 2 Web / 2 DB cluster ( so total license 6 windows OS license , 2 SQL Server license and Guess no sharepoint licenes)

    Thanks Trevor,
    Is load balance service also comes in free license... So, in that case I can use SharePoint Foundation 2013 version for building a simple Intranet & DMS ( with limited functionality).  And for Workflow and content management we have to write code.
    Windows Network Load Balancing (the NLB feature) is included as part of Windows Server and would offer high availability for traffic bound to the SharePoint servers. WNLB can only associate with up to 4 servers.
    Trevor Seward
    Follow or contact me at...
    &nbsp&nbsp
    This post is my own opinion and does not necessarily reflect the opinion or view of Microsoft, its employees, or other MVPs.

  • SharePoint 2013 High Availability - BI Semantic

    SharePoint 2013 High Availability
    Hi all,
    Settings
    Two Sites connected via tunnel, no issue in networking – Single farm
    SQL2012, two DB with mirroring
    SharePoint 2013, two servers in the farm, each server is APP & WEB
    Issue, this has been designed against full site failure (e.g. losing connection to other site, which means losing 1 SQL and 1 SP server)
    When users try to view some reports from the BI page, they can’t , we have confirmed that SQL mirrors are failing correctly, the issue (we think to be to access the SharePoint reporting services DB which we have mirrored as well), any help will be great.
    Regards

    Okay, if you've validated that you have the 1Gbps and 1ms over 10 minutes average, then you're in a supported scenario. How did you configure the SSRS databases to use mirroring within SharePoint? With SSRS databases, you'll need to use PowerShell:
    $database = Get-SPDatabase | where {$_.Name -eq "ReportingServiceDbName"}
    $database.AddFailoverServiceInstance("InstanceName")
    $database.Update()
    Trevor Seward
    Follow or contact me at...
    &nbsp&nbsp
    This post is my own opinion and does not necessarily reflect the opinion or view of Microsoft, its employees, or other MVPs.

  • Windows 2012 RDS - Session Host servers High Availability

    Hello Windows/Terminal server Champs,
    I am new middle of implementing RDS environment for one of my customer, Hope you could help me out.
    My customer has asked for HA for RDS session host where applications are published, and i have prepared below plan for server point of view.
     2 Session Host server, 1 webaccess, 1 License/connection
    Broker & 1 Gateway (DMZ).
     In first Phase, we are planning to target internal user
    who connect to Session host HA where these 2 servers will have application installed and internal user will use RDP to access these application.
    In second Phase we will be dealing with external Party who connect from external network where we are planning to integrate with NetIQ => gateway
    => Webaccess/Session host
     I have successfully installed and configured 2 Session
    Host, 1 license/Broker. 1 webAccess & 1 Gateway. But my main concern to have session Host High Available as it is hosting the application and most of the internal user going to use it. to configure it i am following http://technet.microsoft.com/en-us/library/cc753891.aspx  
    However most of the Architecture is change in RDS 2012. can you please help me out to setup the Session Host HA.
    Note: we can have only 1 Connection broker /Licensing server , 1 webacess server & 1 Gateway server, we cannot increase more server due to cost
    factor.
     thanks in advance.

    Yes, absolutely no problem in just using one connection broker in your environment as long as your customer understands the SPOF.
    the session hosts however aren't really what you would class HA - but to set them up so youhave reduancy you would use either Windows NLB, an external NLB device or windows dns round robin. My preferred option when using the connection broker is DNS round
    robin - where you give each server in the farm the same farm name dns entry - the connection broker then decides which server to allocate the session too.
    You must ensure your session host servers are identical in terms of software though - same software installed in the same paths on all the session host servers.
    if you use the 2012 deployment wizard through server manager roles the majority of the config is done for you.
    Regards,
    Denis Cooper
    MCITP EA - MCT
    Help keep the forums tidy, if this has helped please mark it as an answer
    My Blog
    LinkedIn:

  • Load balancing and High Availability topology

    Our Forms 6i client-server application currently runs on Citrix farm of 20 Windows 2000 boxes (IBM Blade Servers 2 CPU and 2 Gig Memory).
    Application supports 2000 users.
    We are moving to AS 10g r2, forms 10g and the goal is to use same hardware, 20 Windows boxes (or less), for intranet web deployment.
    What will be our best choices for application Load balancing and High Availability?
    Hardware load balancer, Web Cache, mod-oc4j? Combinations?
    Any suggestions, best practices, your experience?

    Gerd, I understand, that you are running 10g web forms through the browser, but using Citrix for deployment. This means that in addition to Application Server and Forms runtime sessions, it will be separate browser session opened for each user. What the advantage of this configuration?
    Michael, we are aware, that Citrix is not supported by Oracle as a deployment platform. That only means that prior contacting Oracle Support we have to reproduce the problem in standard environment. It was never been a problem to reproduce problem :) We were using Citrix as a deployment platform for Forms 6i client/server for 4 years, but now we are forced to upgrade to 10g.
    We are familiar with various Load balancing options available. The question is which option is the most "workable" in our case.

  • SCOM 2012 R2 Reporting Services High Availability Configuration Setting

    Can u guys tell me how I can make a scom 2012 r2 report server high availability configuration kindly share the steps or screen of it.

    SCOM reporting feature is depend on SQL reporting and as a result you may refer to SQL reporting service high availability.
    For detail, pls. refer
    http://technet.microsoft.com/en-us/library/bb522745.aspx
    http://www.mssqltips.com/sqlservertip/2335/scale-out-sql-server-2008-r2-reporting-services-farm-using-nlb-part-1/
    http://www.mssqltips.com/sqlservertip/2336/scale-out-ssrs-r2-farm-using-windows-network-load-balancing-part-2/
    http://technet.microsoft.com/en-us/library/hh882437.aspx
    Roger

  • Topology.svc - Endpoints - Web Services High Availability

    Hi,
    I was recently performing some simple DRP tests before going to production and i faced some issues i never encountered before..
    (followed
    http://blogs.msdn.com/b/besidethepoint/archive/2011/02/19/how-i-learned-to-stop-worrying-and-love-the-sharepoint-topology-service.aspx for usefull commands related to enpoints)
    My farm:
    SP 2013 - CU August 2013
    2 WFE (WFE1, WFE2)
    2 App (App1, App2) : Most services started on both servers (UPSS on APP1, UPS on both) - Central Admin on Both.
    SQL Cluster
    At normal state the command : (Get-SPTopologyServiceApplicationProxy).ApplicationProxies | Format-List *
    returns > ServiceEndpointUri :
    https://app1:32844/Topology/topology.svc
    (if i'm not wrong, this topology.svc can run on only one server at a time)
    I stopped WFE1, no pb, the NBL (appliance) is doing the job.
    Then I stopped the App1 and started to have some issues. (most enpoints not balanced to app2)
    I run the job "Application Addresses Refresh Job"
    Or launch PS command : Start-SPTimerJob job-spconnectedserviceapplication-addressesrefresh
    Wait 20 sec.
    A few endpoints are now on APP2 (MMS, Search), it seems to work, i reached my web page.
    I ask a mate to try and he got the "sorry we encountered and error..." > Can't load user profile.
    I refreshed my browser and got the same error.......
    ULS review, I can see that some svc request (most about user, profiledbcacheService) are still on APP1 !!
    A failure was reported when trying to invoke a service application: EndpointFailure Process Name: w3wp Process ID: 6784 AppDomain Name: /LM/W3SVC/93617642/ROOT-1-130445830237578923 AppDomain ID: 2 Service Application Uri: urn:schemas-microsoft-com:sharepoint:service:e8315f8e5d7d4b1b90876e3b0043a4ae#authority=urn:uuid:164efb17f28c4d2d9702ce3e86f0c0e8&authority=https://app1:32844/Topology/topology.svc
    Active Endpoints: 1 Failed Endpoints:1 Affected Endpoint:
    http://app1:32843/e8315f8e5d7d4b1b90876e3b0043a4ae/ProfileService.svc
    The command (Get-SPTopologyServiceApplicationProxy).ApplicationProxies | Format-List *
    Still return me that my topology.svc is on
    https://app1:32844/Topology/topology.svc
    But App1 is down !!
    If my understanding is ok: Normally, the internal round robin loadbalancer (Application Discovery and Load Balancer Service, started on all servers, not configurable) should manage this.
    Application Addresses Refresh Job is running each 15 min. and refreshs available endpoints, using the topology.svc
    But, the topology.svc called is always on APP1 which is down !
    At this time, i haven't found why SharePoint is not detecting that APP1 is down and is not
    automatically recreating a topology service on another available server.....
    If you have any idea...your help is welcome :)
    Regards,
    O.P

    Hi  ,
    For achieving Web Services High Availability, you need to make sure service-applications have more than one server to service them to increase back-end resiliency to a SharePoint server dropping off the network.
    You can refer to the blog:
    http://blogs.msdn.com/b/sambetts/archive/2013/12/05/increasing-service-application-redundancy-high-availability-sharepoint.aspx
    Thanks,
    Eric
    Forum Support
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support,
    contact [email protected]
    Eric Tao
    TechNet Community Support

  • Multi Site SQL High Availability group

    Hi,
    We have our primary Data Centre where we have 2 MS SQL 2012 Enterprise Servers and 2 SharePoint 2013 Enterprise servers.
    We have SQL High Availability group implemented on our primary Site (Site A)
    Site A has subnet of 192.168.0.0.
    We recently added a new DR Site (Site B). Site B we have MS SQL 2012 Enterprise Servers.  Site B have a subnet of
    172.1.1.0
    Both sites are connected via a VPN Tunnel. MS SQL 2012 Enterprise Server on Site B has been added in have SQL High Availability
    group of Site A.
    SQL High Availability group have 2 IPs 192.168.0.32 and 172.1.1.32. SQL Listener have 2 IPs 192.168.0.33 and 172.1.1.33
    We want to make sure that if Site A completely down then Site B work as an active Site. But when Site A down we are unable
    to ping High Availability group and Site B is unable to work as active site. SQL and SharePoint services are completely down.
    Site A has AD(Primary Domain  Controller) and Site B has ADC(Additional Domain Controller).
     Site A has witness server.
    We are using Server 2012 Data Centre
    Please suggest.
    Farooq

    SharePoint is not the same as any other applications. The DR site has to be a completely different farm from your production. This means that the SharePoint_AdminContent and config databases on both farms are different and should not be included in the Availability
    Group (they do not support asynchronous Availability Group.) Only content databases and other SharePoint databases supported in an Availability Group should be included as per this
    TechNet article. Have a look at this
    blog post for your networking configuration.
    The reason your Windows cluster service goes down in the DR data center when the when your primary data center goes down is because the cluster has no majority votes. You have 4 available votes in your cluster - 3 in the primary data center in the form of
    the 2 Availability Group replicas and the witness server and 1 in the DR data center. If the Windows cluster does not have majority of votes, it automatically shuts down the cluster. This is by design. That is why when this situation happens, you need to force
    the cluster to start without a quorum as described in
    this article (Windows Server 2012 R2 introduced the concept of dynamic witness and dynamic quorum to address this concern.) By default, the IP address of the Availability Group listener name will be registered on all of the DNS servers if the cluster is
    running Windows Server 2012 and higher (this is the RegisterAllProvidersIP property that Samir mentioned.) However, you need to flush the DNS cache on the SharePoint web and application servers and point them to the correct IP address on the DR server after
    failover. This is because the default TTL value of the Availability Group listener name is 20 minutes. 
    It is important to define your DR strategies and objectives when designing the architecture to make sure that the solution will meet your goals. This will also help you automate some of the processes in the DR data center when failover or a disaster occurs.
    Edwin Sarmiento SQL Server MVP | Microsoft Certified Master
    Blog |
    Twitter | LinkedIn
    SQL Server High Availability and Disaster Recover Deep Dive Course

  • Best practice for highly available management / publishing servers

    I am testing a highly available appv 5.0 environment, which will deploy appv packages to a Xenapp farm.  I have two SQL 2012 servers configured as an availability group for the backend, and two publishing / management servers for the front end. 
    What is the best practice to configure the publishing / management servers for high availability?  Should I configure them as an NLB cluster, which I have tested and does seem to work, or should I just use the GPO to configure the clients to use both
    publishing servers, which I have also tested and appears to work?
    Thanks,
    Patrick Sullivan

    In App-V 5.0 the Management and Publishing Servers are hosted in IIS, so use the same approach for HA as you would any web application.
    If NLB is all that's available to you, then use that; otherwise I would recommend a proper load balancing solution such as Citrix NetScaler or KEMP LoadManager.
    Please remember to click "Mark as Answer" or "Vote as Helpful" on the post that answers your question (or click "Unmark as Answer" if a marked post does not actually
    answer your question). This can be beneficial to other community members reading the thread.
    This forum post is my own opinion and does not necessarily reflect the opinion or view of my employer, Microsoft, its employees, or other MVPs.
    Twitter:
    @stealthpuppy | Blog:
    stealthpuppy.com |
    The Definitive Guide to Delivering Microsoft Office with App-V

  • RD Web Access high availability

    I'm looking for information regarding how to deploy highly available 2012 RD web access servers. There seems to be little or no information online. Does anyone have any experience of setting this up? I would like to run two RD web access servers in an active/active
    model. Also can anyone give any insight into whether this role is happy to be installed on HA RDCB servers too?
    Many thanks

    Hi,
    Thank you for posting in Windows Server Forum.
    We can install the RDCB role to maintain the RDS Farm for RD Web. As per my knowledge we need to install this role on other server same on RDSH as need to use this server outside environment. We need to install RD Web and RD Gateway outside as we can access
    RD Web access page, RemoteApp from internet and then connect to RD Connection Broker which will redirect to an RDSH server which all connected by RD Gateway. 
    But sorry to disappoint you but there is no official document for RD Web HA. As we can use of DNS round robin or other load balancing techniques. Please check below article just for reference.
    RD Connection Broker High Availability in Windows Server 2012
    http://blogs.msdn.com/b/rds/archive/2012/06/27/rd-connection-broker-high-availability-in-windows-server-2012.aspx
    Step by Step Windows 2012 R2 Remote Desktop Services – Part 3
    http://msfreaks.wordpress.com/2013/12/26/windows-2012-r2-remote-desktop-services-part-3/
    Hope it helps!
    Thanks.
    Dharmesh Solanki
    TechNet Community Support

  • High availability for Web front END server

    Dear All
    I am unable to understand the high availability model for Web front end server
    I am currently working on MOSS 2007/IIS 7 but I think for all versions it will remain the same
    I am now running single WFE server and my installation mode allow for adding extra servers to the Farm
    now when I add Extra server what will happen next ? should I add extra web application and site collections ? will load balancing include lists, library items and workflows
    how this stuff will be stored on one database
    it's too vague to me so extra explanation will be appreciated
    Regards

    To get a fault tolerant environment you need to do two things.
    1) Add a second server that is running the the Microsoft SharePoint Foundation Web Service.  That's the service that responds to calls for the web site making the server a WFE.
    2) Implement some form of load balancing to distribute HTTP requests for SharePoint web sites to the two servers.  This can be done with something as simple as Windows Network Load Balancing or with dedicated hardware like an F5 Load balancer.  
    Once you've added a load balancer and a second server you don't need to do anything else.
    Paul Stork SharePoint Server MVP
    Principal Architect: Blue Chip Consulting Group
    Blog: http://dontpapanic.com/blog
    Twitter: Follow @pstork
    Please remember to mark your question as "answered" if this solves your problem.

  • Advice Requested - High Availability WITHOUT Failover Clustering

    We're creating an entirely new Hyper-V virtualized environment on Server 2012 R2.  My question is:  Can we accomplish high availability WITHOUT using failover clustering?
    So, I don't really have anything AGAINST failover clustering, and we will happily use it if it's the right solution for us, but to be honest, we really don't want ANYTHING to happen automatically when it comes to failover.  Here's what I mean:
    In this new environment, we have architected 2 identical, very capable Hyper-V physical hosts, each of which will run several VMs comprising the equivalent of a scaled-back version of our entire environment.  In other words, there is at least a domain
    controller, multiple web servers, and a (mirrored/HA/AlwaysOn) SQL Server 2012 VM running on each host, along with a few other miscellaneous one-off worker-bee VMs doing things like system monitoring.  The SQL Server VM on each host has about 75% of the
    physical memory resources dedicated to it (for performance reasons).  We need pretty much the full horsepower of both machines up and going at all times under normal conditions.
    So now, to high availability.  The standard approach is to use failover clustering, but I am concerned that if these hosts are clustered, we'll have the equivalent of just 50% hardware capacity going at all times, with full failover in place of course
    (we are using an iSCSI SAN for storage).
    BUT, if these hosts are NOT clustered, and one of them is suddenly switched off, experiences some kind of catastrophic failure, or simply needs to be rebooted while applying WSUS patches, the SQL Server HA will fail over (so all databases will remain up
    and going on the surviving VM), and the environment would continue functioning at somewhat reduced capacity until the failed host is restarted.  With this approach, it seems to me that we would be running at 100% for the most part, and running at 50%
    or so only in the event of a major failure, rather than running at 50% ALL the time.
    Of course, in the event of a catastrophic failure, I'm also thinking that the one-off worker-bee VMs could be replicated to the alternate host so they could be started on the surviving host if needed during a long-term outage.
    So basically, I am very interested in the thoughts of others with experience regarding taking this approach to Hyper-V architecture, as it seems as if failover clustering is almost a given when it comes to best practices and high availability.  I guess
    I'm looking for validation on my thinking.
    So what do you think?  What am I missing or forgetting?  What will we LOSE if we go with a NON-clustered high-availability environment as I've described it?
    Thanks in advance for your thoughts!

    Udo -
    Yes your responses are very helpful.
    Can we use the built-in Server 2012 iSCSI Target Server role to convert the local RAID disks into an iSCSI LUN that the VMs could access?  Or can that not run on the same physical box as the Hyper-V host?  I guess if the physical box goes down
    the LUN would go down anyway, huh?  Or can I cluster that role (iSCSI target) as well?  If not, do you have any other specific product suggestions I can research, or do I just end up wasting this 12TB of local disk storage?
    - Morgan
    That's a bad idea. First of all Microsoft iSCSI target is slow (it's non-cached @ server side). So if you really decided to use dedicated hardware for storage (maybe you do have a reason I don't know...) and if you're fine with your storage being a single
    point of failure (OK, maybe your RTOs and RPOs are fair enough) then at least use SMB share. SMB at least does cache I/O on both client and server sides and also you can use Storage Spaces as a back end of it (non-clustered) so read "write back flash cache
    for cheap". See:
    What's new in iSCSI target with Windows Server 2012 R2
    http://technet.microsoft.com/en-us/library/dn305893.aspx
    Improved optimization to allow disk-level caching
    Updated
    iSCSI Target Server now sets the disk cache bypass flag on a hosting disk I/O, through Force Unit Access (FUA), only when the issuing initiator explicitly requests it. This change can potentially improve performance.
    Previously, iSCSI Target Server would always set the disk cache bypass flag on all I/O’s. System cache bypass functionality remains unchanged in iSCSI Target Server; for instance, the file system cache on the target server is always bypassed.
    Yes you can cluster iSCSI target from Microsoft but a) it would be SLOW as there would be only active-passive I/O model (no real use from MPIO between multiple hosts) and b) that would require a shared storage for Windows Cluster. What for? Scenario was
    usable with a) there was no virtual FC so guest VM cluster could not use FC LUs and b) there was no shared VHDX so SAS could not be used for guest VM cluster as well. Now both are present so scenario is useless: just export your existing shared storage without
    any Microsoft iSCSI target and you'll be happy. For references see:
    MSFT iSCSI Target in HA mode
    http://technet.microsoft.com/en-us/library/gg232621(v=ws.10).aspx
    Cluster MSFT iSCSI Target with SAS back end
    http://techontip.wordpress.com/2011/05/03/microsoft-iscsi-target-cluster-building-walkthrough/
    Guest
    VM Cluster Storage Options
    http://technet.microsoft.com/en-us/library/dn440540.aspx
    Storage options
    The following tables lists the storage types that you can use to provide shared storage for a guest cluster.
    Storage Type
    Description
    Shared virtual hard disk
    New in Windows Server 2012 R2, you can configure multiple virtual machines to connect to and use a single virtual hard disk (.vhdx) file. Each virtual machine can access the virtual hard disk just like servers
    would connect to the same LUN in a storage area network (SAN). For more information, see Deploy a Guest Cluster Using a Shared Virtual Hard Disk.
    Virtual Fibre Channel
    Introduced in Windows Server 2012, virtual Fibre Channel enables you to connect virtual machines to LUNs on a Fibre Channel SAN. For more information, see Hyper-V
    Virtual Fibre Channel Overview.
    iSCSI
    The iSCSI initiator inside a virtual machine enables you to connect over the network to an iSCSI target. For more information, see iSCSI
    Target Block Storage Overviewand the blog post Introduction of iSCSI Target in Windows
    Server 2012.
    Storage requirements depend on the clustered roles that run on the cluster. Most clustered roles use clustered storage, where the storage is available on any cluster node that runs a clustered
    role. Examples of clustered storage include Physical Disk resources and Cluster Shared Volumes (CSV). Some roles do not require storage that is managed by the cluster. For example, you can configure Microsoft SQL Server to use availability groups that replicate
    the data between nodes. Other clustered roles may use Server Message Block (SMB) shares or Network File System (NFS) shares as data stores that any cluster node can access.
    Sure you can use third-party software to replicate 12TB of your storage between just a pair of nodes to create a fully fault-tolerant cluster. See (there's also a free offering):
    StarWind VSAN [Virtual SAN] for Hyper-V
    http://www.starwindsoftware.com/native-san-for-hyper-v-free-edition
    Product is similar to what VMware had just released for ESXi except it's selling for ~2 years so is mature :)
    There are other guys doing this say DataCore (more playing for Windows-based FC) and SteelEye (more about geo-cluster & replication). But you may want to give them a try.
    Hope this helped a bit :) 
    StarWind VSAN [Virtual SAN] clusters Hyper-V without SAS, Fibre Channel, SMB 3.0 or iSCSI, uses Ethernet to mirror internally mounted SATA disks between hosts.

  • 11.1.2 High Availability for HSS LCM

    Hello All,
    Has anyone out there successfully exported an LCM artifact (Native Users) through Shared Services console to the OS file system (<oracle_home>\user_projects\epmsystem1\import_export) when Shared Services is High Available through a load balancer?
    My current configuration is two load balanced Shared Services servers using the same HSS repository. Everything works perfectly everywhere except when I try to export anything through LCM from the HSS web console. I have shared out the import_export folder on server 1. And I have edited the oracle_home>\user_projects\epmsystem1\config\FoundationServices1\migration.properties file on the second server to point to the share on the first server. Here is a cut and paste of the migration.properties file from server2.
    grouping.size=100
    grouping.size_unknown_artifact_count=10000
    grouping.group_by_type=Y
    report.enabled=Y
    report.folder_path=<epm_oracle_instance>/diagnostics/logs/migration/reports
    fileSystem.friendlyNames=false
    msr.queue.size=200
    msr.queue.waittime=60
    group.count=10000
    double-encoding=true
    export.group.count = 30
    import.group.count = 10000
    filesystem.artifact.path=\\server1\import_export
    I perform an export of just the native users to the file system, the export fails stating and I find errors in the log stating the following.
    [2011-03-29T20:20:45.405-07:00] [EPMLCM] [NOTIFICATION] [EPMLCM-11001] [oracle.EPMLCM] [tid: 10] [ecid: 0000Iw4ll2fApIOpqg4EyY1D^e6D000000,0] [arg: /server1/import_export/msr/PKG_34.xml] Executing package file - /server1/import_export/msr/PKG_34.xml
    [2011-03-29T20:20:45.530-07:00] [EPMLCM] [ERROR] [EPMLCM-12097] [oracle.EPMLCM] [tid: 10] [ecid: 0000Iw4ll2fApIOpqg4EyY1D^e6D000000,0] [SRC_CLASS: com.hyperion.lcm.clu.CLUProcessor] [SRC_METHOD: execute:?] Cannot find the migration definition file in the specified file path - /server1/import_export/msr/PKG_34.xml.
    [2011-03-29T20:20:45.546-07:00] [EPMLCM] [NOTIFICATION] [EPMLCM-11000] [oracle.EPMLCM] [tid: 10] [ecid: 0000Iw4ll2fApIOpqg4EyY1D^e6D000000,0] [arg: Completed with failures.] Migration Status - Completed with failures.
    I go and look for the path that it is searching for "/server1/import_export/msr/PKG_34.xml" and find that this file does exist, it was infact created by the export process so I know that it is able to find the correct location, but it then says that it can't find it after the fact. I have tried creating mapped drives and editing the migration.properties file to reference the mapped drive but it gives me the same errors but with the new path. I also tried the below article that I found in support with no result. Please advise, thank you.
    Bug 11683769: LCM MIGRATIONS FAIL WHEN HSS IS SET UP FOR HA AND HSS IS STARTED AS A SERVICE
    Bug Attributes
    Type B - Defect Fixed in Product Version 11.1.2.0.000PSE
    Severity 3 - Minimal Loss of Service Product Version 11.1.2.0.00
    Status 90 - Closed, Verified by Filer Platform 233 - Microsoft Windows x64 (64-bit)
    Created 25-Jan-2011 Platform Version 2003 R2
    Updated 24-Feb-2011 Base Bug 11696634
    Database Version 2005
    Affects Platforms Generic
    Product Source Oracle
    Related Products
    Line - Family -
    Area - Product 4482 - Hyperion Lifecycle Management
    Hdr: 11683769 2005 SHRDSRVC 11.1.2.0.00 PRODID-4482 PORTID-233 11696634Abstract: LCM MIGRATIONS FAIL WHEN HSS IS SET UP FOR HA AND HSS IS STARTED AS A SERVICE*** 01/25/11 05:15 pm ***Hyperion Foundation Services is set up for high availability. Lifecycle management migrations fail when running Hyperion Foundation Services - Managed Server as a Windows Service. We applied the following configuration changes:Set up shared disk then we modify the migration.properties.filesystem.artifact.path=\\<servername>\LCM\import_export We are able to get the migrations to work if we set filesystem.artifact.path=V:\import_export (a maped drive to the shared disk) and we run Hyperion Foundation Services in console mode.*** 01/25/11 05:24 pm *** *** 01/28/11 03:32 pm *** (CHG: Sta->11)*** 01/28/11 03:32 pm *** (CHG: Comp. Ver-> 11.1.2 -> 11.1.2.0.00)*** 01/28/11 03:34 pm *** (CHG: Base Bug-> NULL -> 11696634)*** 02/02/11 01:53 pm *** (CHG: Sta->41)*** 02/02/11 01:53 pm ****** 02/02/11 05:18 pm ****** 02/03/11 10:30 am ****** 02/04/11 03:53 pm *** *** 02/04/11 03:53 pm *** (CHG: Sta->80 Asg->NEW OWNER)*** 02/10/11 08:49 am ****** 02/15/11 11:59 am ****** 02/15/11 04:56 pm ****** 02/16/11 09:58 am *** (CHG: Sta->90)*** 02/16/11 09:58 am ****** 02/16/11 02:12 pm *** (CHG: Sta->93)*** 02/24/11 06:14 pm *** (CHG: Sta->90)Back to top

    it is not possible to implement a kind of HA between two different appliances 3315 and 3395. 
    A node in HA can have the 3 persona. 
    Suppose Node A (Admin(primary), monitoring(Primary) and PSN). 
    Node B(Admin(Secondary), Monitoring(Secondary) and PSN). 
    If the Node A is unavailable, you will have to promote manually the Admin role to Primary. 
    Although the best way is to have
    Node A Admin(Primary), Monitoring(Secondary) and PSN
    Node B Admin(Secondary), Monitoring (Primary) and PSN. 
    Rate if helpful and Marked As correct if it is correct for the experts. 
    Regards,

Maybe you are looking for