High Availability design/upgrade

Hi,
We’re looking to upgrade our Production SQL Cluster infrastructure. Currently we are running on old hardware and software….We currently have two nodes in the cluster. 2-Proliant
DL380 Gen5 servers connected to our HP MSA2000 for shared storage(Direct Attached Storage) using a SAS cable with 1 terabyte. Windows 2008 R2 enterprise with Clustering turned-on and SQL 2005 server. The DB isn't that
big, but it is crucial to our company….  I was curious to get some opinions what would be our best options to upgrade….. We have two spare physical Proliant DL380 Gen 8 servers w/ 64 gigs of ram and total of 1.5 terabyte of worth of disk space. We would
have all the HD bay’s with HD’s therefore no need for a new HP MSA /direct attached storage….. So should we go with Windows 2012 Enterprise with SQL Enterprise and turnon SQL  Always on Availability Group? Or any other recommendations? Any suggestion
and comment would be greatly appreciated.

Hello,
SQL Server AlwaysOn Availability Groups requirement Windows Server Failover Clustering (WSFC) clusters.Each availability replica of a given availability group must reside on a different node of the same WSFC cluster
Based on your description, it seems that you had two node cluster with Windows Server 2008 R2. If so, you should install windows 2008 R2 SP1 for supporting  SQL Server AlwaysOn Availability Groups.
Reference:http://msdn.microsoft.com/en-us/library/ff878487.aspx 
Regards,
Fanny Liu
Fanny Liu
TechNet Community Support

Similar Messages

  • Wism2 in High Availability - FUS Upgrade

    Hello all,
    what is the procedure to be followed when upgrading the FUS on a WISM-2 in HA cluster?
    Is it the same procedure that applies when upgrading the wism's code in HA?
    Thank you
    Edited:
    The procedure is exactly the same as upgrading. First upgrade the image and then FUS.make sure that you are connected using console connection and definitely perform this within a maintenance window of at least 1 hour. 

    Hi,
    Thanks for sharing this info.
    Snipped from this guide also explains how its done:
    https://supportforums.cisco.com/discussion/12190806/wism2-high-availability-fus-upgrade
    The FUS image can be upgraded while the controllers have HA enabled. The secondary controller will 
    get upgraded just like it does when upgrading the regular code. However, when you initiate the reboot 
    on the primary controller both controllers will be unreachable until the FUS upgrade completes on both 
    the active and the standby in the HA pair. This process will take around 30 to 40 minutes to complete 
    just like in a non-HA FUS upgrade.

  • Best practice for High availability design, HSRP

    Hi,
    I am planning to create High Availability for LAN to WAN connectivity.
    But I want to know your opinion about the best way how to do this. I googled for a solution/best way how to do this, but I didn't found in my opinion right answer.
    The situation:
    I have 2 3945E Routers and 2 3560 switches. The design that I am planning to implement is below.
    The main goal is to have redundant connection, whatever one of the devices will fail. For example, if the R1 will fail, R2 should become active, if the SW1 will fail, the SW2 will take care about reachability and vice versa. The router 1 should be preferred always, if the link to ISP isn't down, because of greater bandwidth. So why am I drown 2 connections to 2 separate switches. If the SW1 will fail, I will still have a connection to WAN using R1 router.
    The Router interface should be configured with sub interfaces (preferred over secondary IP address of interface), because more than 10 subnets will be assigned to the LAN segment. The routers have 4 Gi ports.
    HSRP must be enabled on LAN side, because PC's on LAN must have redundant def. getaway.
    So, the question is - what is the best and preferred way to do this?
    In my opinion, I should use BVI and combine R1 routers 2 interfaces in to logical one and do the same for the R2.
    Next, turn the router in to L3 switch using IRB and then configure HSRP.
    What would be your preferred way to do this?

    Hi Audrius,
    I would suggest you to go with HSRP. GLBP you will use where you want load balance.
    I think the connectivity between your Routers (3945) and switches (3560) is gigabit connection which is high speed. So keep one physical link from your switches to each router and do HSRP on those router physical interfaces.
    In this way you will have high availability like if R1 fails then R2 will take over.
    Regarding the config see the below which I have for one of my Customer DC.
    ACTIVE:
    track 1 interface GigabitEthernet0/0 line-protocol
    track 2 interface GigabitEthernet0/0 line-protocol
    interface GigabitEthernet0/1
    ip address 10.10.10.12 255.255.255.0
    ip nat inside
    ip virtual-reassembly
    duplex full
    speed 100
    standby use-bia scope interface
    standby 0 ip 10.10.10.10
    standby 0 priority 110
    standby 0 preempt
    standby 0 authentication peter2mo
    standby 0 track 1 decrement 30
    standby 0 track 2 decrement 30
    STANDBY:
    track 1 interface GigabitEthernet0/0 line-protocol
    interface GigabitEthernet0/1
    ip address 10.10.10.11 255.255.255.0
    ip nat inside
    ip virtual-reassembly
    duplex full
    speed 100
    standby use-bia scope interface
    standby 0 ip 10.10.10.10
    standby 0 priority 90
    standby 0 authentication peter2mo
    standby 0 track 1 decrement 30
    Please rate the helpfull posts.
    Regards,
    Naidu.

  • Question on replication/high availability designs

    We're currently trying to work out a design for a high-availability system using Oracle 9i release 2. Having gone through some of the Oracle whitepapers, it appears that the ideal architecture involves setting up 2 RAC sites using Dataguard to synchronize the data. However, due to time and financial constraints, we are only allowed to have 2 servers for hosting the databases, which are geographically separate from each other in prevention of natural disasters. Our app servers will use JDBC pools to connect to the databases.
    Our goal is to have both databases be the mirror image of each other at any given time, and the database must be working 24/7. We do have a primary and a secondary distinction between the two, so if the primary fails, we would like the secondary database to take over the tasks as needed.
    The ability to query existing data is mission critical. The ability to write/update the database is less important, however we do need the secondary to be able to process data input/updates when primary is down for a prolonged period of time, and have the ability to synchronize back with the primary site when it is back up again.
    My question now is which replication technology should we try to implement? I've looked into both Oracle Advanced Replication and Dataguard, each seems to have its own advantages and drawbacks:
    Replication - can easily switch between the two databases using multimaster implementation, however data recovery/synchronization may be difficult in case of failure, and possibly will lose data (pending implementation). There has been a few posts in this forum that suggested that replication should not really be considered as an option for high availability, why is that?
    Dataguard - zero data loss in failover/switchover, however manual intervention is required to initiate failover/switchover. Once the primary site fails over to the standby, the standby becomes the primary until DBA manually goes back in and switch the roles. In Oracle 10g release 2, seems that automatic failover is achieved through the use of an extra observer piece. There does not seem to be anyway to do this in Oracle 9i release 2.
    Being new to the implementation of high-availability systems, I am at somewhat of a loss at this point. Both implementations seem to be a possible candidate, but we will need to sacrifice some efforts for both of them also. Would anyone shine some light on this, maybe point out my misconceptions with Advanced Replication and Dataguard, and/or suggest a better architecture/technology to use? Any input is greatly appreciated, thanks in advance.
    Sincerely,
    Peter Tung

    Hi,
    It sounds as if you're talking about the DB_TXN_NOSYNC flag, rather than DB_NOSYNC.
    You mention that in general, you lose uncommitted transactions on system failure. I think what you mean is that you may lose some committed transactions on system failure. This is correct.
    It is also correct that if you use replication you can arrange to have clients have a copy of all committed transactions, so that if the master fails (and enough clients do not fail, of course) then the clients still have the transaction data, even when using DB_TXN_NOSYNC.
    This is a very common usage scenario for Berkeley DB replication/HA, used to achieve high throughput. You will want to pay attention to the configured ack policy, group size setting, setting of the 2SITE_STRICT option (if group size == 2).

  • EBS High Availability design

    Hi Gurus,
    What is the best way of configuring HA on EBS (APPS) side the DB side is running on an Exadata machine.I have 2 Sparc T-4s that i can use for the APPS.MOS Notes appreciated.

    Solaris Cluster for EBS
    Oracle Solaris Cluster Data Service for Oracle E-Business Suite Guide
    October 2012, E29648-01
    http://docs.oracle.com/cd/E29086_01/html/E29648/index.html
    Using Oracle 11g Release 2 Real Application Clusters with Oracle E-Business Suite Release 12 [ID 823587.1]
    Interoperability Notes Oracle EBS R12 with Oracle Database 11gR2 (11.2.0.2) [ID 1367644.1]
    Oracle EBS R12 with Database Tier Only Platform on Oracle Database 11.2.0 [ID 456347.1]
    Check for Split configuration, to have APPS EBS on Solaris SPARC and DB on Linux

  • How to perform up high availability upgrade with database changes?

    Hi,
    Can someone point me to pertinent documentation for performing live upgrades? Specifically I'm looking for best practice for setting up and attaining High Availability especially during application upgrades which have db changes.
    Thanks in advance,
    Rich

    Hi,
    For internal clients, yes, your requirement can be met by setting the autodiscover site affinity by using the “AutoDiscoverSiteScope” parameter in Set-ClientAccessServer command:
    https://technet.microsoft.com/zh-cn/library/bb125157(v=exchg.150).aspx
    For external clients, I am afraid “No” because the “autodiscover.domain.com” only points to one CAS server.
    Thanks,
    Simon Wu
    TechNet Community Support

  • Exchange 2010 to Exchange 2013 Migration and Architect a resilient and high availability exchange setup

    Hi,
    I currently have a single Exchange 2010 Server that has all the roles supporting about 500 users. I plan to upgrade to 2013 and move to a four server HA Exchange setup (a CAS array with 2 Server as CAS servers  and one DAG with 2 mailbox Servers). My
    goal is to plan out the transition in steps with no downtime. Email is most critical with my company.
    Exchange 2010 is running SP3 on a Windows Server 2010 and a Separate Server as archive. In the new setup, rather than having a separate server for archiving, I am just going to put that on a separate partition.
    Here is what I have planned so far.
    1. Build out four Servers. 2 CAS and 2 Mailbox Servers. Mailbox Servers have 4 partitions each. One for OS. Second for DB. Third for Logs and Fourth for Archives.
    2. Prepare AD for exchange 2013.
    3. Install Exchange roles. CAS on two servers and mailbox on 2 servers. Add a DAG. Someone had suggested to me to use an odd number so 3 or 5. Is that a requirement?
    4. I am using a third party load balancer for CAS array instead of NLB so I will be setting up that.
    5. Do post install to ready up the new CAS. While doing this, can i use the same parameters as assigned on exchange 2010 like can i use the webmail URL for outlook anywhere, OAB etc.
    6. Once this is done. I plan to move a few mailboxes as test to the new mailbox servers or DAG.
    7. Testing outlook setups on new servers. inbound and outbound email tests.
    once this is done, I can migrate over and point all my MX records to the new servers.
    Please let me know your thoughts and what am I missing. I like to solidify a flowchart of all steps that I need to do before I start the migration. 
    thank you for your help in advance

    Hi,
    okay, you can use 4 virtual servers. But there is no need to deploy dedicated server roles (CAS + MBX). It is better to deploy multi-role Exchange servers, also virtual! You could install 2 multi-role servers and if the company growths, install another multi-role,
    and so on. It's much more simpler, better and less expensive.
    CAS-Array is only an Active Directory object, nothing more. The load balancer controls the sessions on which CAS the user will terminate. You can read more at
    http://blogs.technet.com/b/exchange/archive/2014/03/05/load-balancing-in-exchange-2013.aspx Also there is no session affinity required.
    First, build the complete Exchange 2013 architecture. High availability for your data is a DAG and for your CAS you use a load balancer.
    On channel 9 there is many stuff from MEC:
    http://channel9.msdn.com/search?term=exchange+2013
    Migration:
    http://geekswithblogs.net/marcde/archive/2013/08/02/migrating-from-microsoft-exchange-2010-to-exchange-2013.aspx
    Additional informations:
    http://exchangeserverpro.com/upgrading-to-exchange-server-2013/
    Hope this helps :-)

  • How to enable high availability on SQL Server 2005 with Windows Server 2008 Enterprise R2

    Dear Folks,
    I would like to ask you about this thing. I'm working for IT department for bank in Myanmar. Our bank have up to 96 branches across all of Myanmar including H.O. We are using Microsoft SQL Server 2005 with Windows Server 2008 for our banking
    information system. My main problem is having to backup and restore the database backup files every time the servers in branches goes down for whatever reasons. I want to deploy this feature of high availability and failover cluster using Windows Server 2008
    and SQL Server 2005. Our branches have 2 Servers. One is for Primary and other is for Backup. What I want to do is that, I want to change Backup Server to Primary Server whenever the Primary Server goes down for whatever reasons. All the working data and databases
    from Primary will immediately replicated into Backup Server along with all the IP information of Primary Server. Please give me step by step guide for this process.

    Try below
    http://blogs.msdn.com/b/cindygross/archive/2009/10/23/checklist-for-installing-sql-server-2005-as-a-clustered-instance.aspx
    I will recommend you to upgrade SQL server to newer version for support as well as flexibility.
    Regards,
    Vishal Patel
    Blog: http://vspatel.co.uk
    Site: http://lehrity.com

  • SharePoint 2013 High Availability - BI Semantic

    SharePoint 2013 High Availability
    Hi all,
    Settings
    Two Sites connected via tunnel, no issue in networking – Single farm
    SQL2012, two DB with mirroring
    SharePoint 2013, two servers in the farm, each server is APP & WEB
    Issue, this has been designed against full site failure (e.g. losing connection to other site, which means losing 1 SQL and 1 SP server)
    When users try to view some reports from the BI page, they can’t , we have confirmed that SQL mirrors are failing correctly, the issue (we think to be to access the SharePoint reporting services DB which we have mirrored as well), any help will be great.
    Regards

    Okay, if you've validated that you have the 1Gbps and 1ms over 10 minutes average, then you're in a supported scenario. How did you configure the SSRS databases to use mirroring within SharePoint? With SSRS databases, you'll need to use PowerShell:
    $database = Get-SPDatabase | where {$_.Name -eq "ReportingServiceDbName"}
    $database.AddFailoverServiceInstance("InstanceName")
    $database.Update()
    Trevor Seward
    Follow or contact me at...
    &nbsp&nbsp
    This post is my own opinion and does not necessarily reflect the opinion or view of Microsoft, its employees, or other MVPs.

  • Load balancing and High Availability topology

    Our Forms 6i client-server application currently runs on Citrix farm of 20 Windows 2000 boxes (IBM Blade Servers 2 CPU and 2 Gig Memory).
    Application supports 2000 users.
    We are moving to AS 10g r2, forms 10g and the goal is to use same hardware, 20 Windows boxes (or less), for intranet web deployment.
    What will be our best choices for application Load balancing and High Availability?
    Hardware load balancer, Web Cache, mod-oc4j? Combinations?
    Any suggestions, best practices, your experience?

    Gerd, I understand, that you are running 10g web forms through the browser, but using Citrix for deployment. This means that in addition to Application Server and Forms runtime sessions, it will be separate browser session opened for each user. What the advantage of this configuration?
    Michael, we are aware, that Citrix is not supported by Oracle as a deployment platform. That only means that prior contacting Oracle Support we have to reproduce the problem in standard environment. It was never been a problem to reproduce problem :) We were using Citrix as a deployment platform for Forms 6i client/server for 4 years, but now we are forced to upgrade to 10g.
    We are familiar with various Load balancing options available. The question is which option is the most "workable" in our case.

  • Multi Site SQL High Availability group

    Hi,
    We have our primary Data Centre where we have 2 MS SQL 2012 Enterprise Servers and 2 SharePoint 2013 Enterprise servers.
    We have SQL High Availability group implemented on our primary Site (Site A)
    Site A has subnet of 192.168.0.0.
    We recently added a new DR Site (Site B). Site B we have MS SQL 2012 Enterprise Servers.  Site B have a subnet of
    172.1.1.0
    Both sites are connected via a VPN Tunnel. MS SQL 2012 Enterprise Server on Site B has been added in have SQL High Availability
    group of Site A.
    SQL High Availability group have 2 IPs 192.168.0.32 and 172.1.1.32. SQL Listener have 2 IPs 192.168.0.33 and 172.1.1.33
    We want to make sure that if Site A completely down then Site B work as an active Site. But when Site A down we are unable
    to ping High Availability group and Site B is unable to work as active site. SQL and SharePoint services are completely down.
    Site A has AD(Primary Domain  Controller) and Site B has ADC(Additional Domain Controller).
     Site A has witness server.
    We are using Server 2012 Data Centre
    Please suggest.
    Farooq

    SharePoint is not the same as any other applications. The DR site has to be a completely different farm from your production. This means that the SharePoint_AdminContent and config databases on both farms are different and should not be included in the Availability
    Group (they do not support asynchronous Availability Group.) Only content databases and other SharePoint databases supported in an Availability Group should be included as per this
    TechNet article. Have a look at this
    blog post for your networking configuration.
    The reason your Windows cluster service goes down in the DR data center when the when your primary data center goes down is because the cluster has no majority votes. You have 4 available votes in your cluster - 3 in the primary data center in the form of
    the 2 Availability Group replicas and the witness server and 1 in the DR data center. If the Windows cluster does not have majority of votes, it automatically shuts down the cluster. This is by design. That is why when this situation happens, you need to force
    the cluster to start without a quorum as described in
    this article (Windows Server 2012 R2 introduced the concept of dynamic witness and dynamic quorum to address this concern.) By default, the IP address of the Availability Group listener name will be registered on all of the DNS servers if the cluster is
    running Windows Server 2012 and higher (this is the RegisterAllProvidersIP property that Samir mentioned.) However, you need to flush the DNS cache on the SharePoint web and application servers and point them to the correct IP address on the DR server after
    failover. This is because the default TTL value of the Availability Group listener name is 20 minutes. 
    It is important to define your DR strategies and objectives when designing the architecture to make sure that the solution will meet your goals. This will also help you automate some of the processes in the DR data center when failover or a disaster occurs.
    Edwin Sarmiento SQL Server MVP | Microsoft Certified Master
    Blog |
    Twitter | LinkedIn
    SQL Server High Availability and Disaster Recover Deep Dive Course

  • SCDPM 2012 R2 High Avalibility Design?

    Hello Everyone,
    Can we Install SCDPM 2012 R2 in High Availblity Mode?
    If yes What's the Design for SCDPM high availibility?
    Thnks.......

    Hello Everyone,
    Can we Install SCDPM 2012 R2 in High Availblity Mode?
    If yes What's the Design for SCDPM high availibility?
    Thnks.......
    Hi,
    There is not a High Availability solution for DPM such as clustering DPM servers. The options you do have are 
    Primary and Secondary DPM servers, DPM chaining, or
    Cyclic protection. Here is a breakdown of each:
    Primary to secondary protection: The database and replicas stored on a primary DPM server can be backed up to a secondary DPM server. If the primary server fails the secondary server continues to back up protected workloads.
    For more on Primary to secondary protection:
    http://technet.microsoft.com/en-us/library/jj244598.aspx
    DPM chaining: A chain of DPM servers provide protection, and each server protects the next one in in the chain. For example:
        DPM1 is protected by DPM2 (DPM1 is the primary and DPM2 is the secondary).
        DPM2 is protected by DPM3 (DPM2 is the primary and DPM3 is the secondary)
    To setup chaining:
    http://technet.microsoft.com/en-us/library/hh758194.aspx
    Cyclic protection: One DPM server is backed up by another DPM server, and vice versa. So that DPM1 protects DPM2, and likewise DPM2 protects DPM1.
    More on Cyclic protection:
    http://technet.microsoft.com/en-us/library/jj244597.aspx#BKMK_Cyclic
    I hope this adds some clarification around this.
    My Blog | www.buchatech.com | www.systemcenterportal.com
    If you found this post helpful, please give it a "Helpful" vote. If it answered your question, remember to mark it as an "Answer". This posting is provided "AS IS" with no warranties and confers no rights! Always test ANY suggestion
    in a test environment before implementing!

  • High Availability Options Without Enterprise Edition

    I think I've found the answer to this already but I just wanted to confirm that I'm getting this right. In SQL Server 2012, is there no way to implement a high-availability solution which doesn't require shared storage or use of a deprecated feature without
    having Enterprise Edition?
    At the moment we have databases hosted on a primary server running Windows Server 2012 Standard and SQL Server 2012 Standard.  We have a second, identical server, and the databases are mirrored between the two servers using Database Mirroring. 
    This works perfectly and meets all of our requirements.  The only possible problem is that we're looking at storing documents using FILESTREAM, which isn't available in Mirrored databases.
    From what I've read, Database Mirroring is deprecated in SQL Server 2012.  High Availability Groups, which sound great, are only available in the Enterprise Edition.  This feels like a real kick in the teeth from Microsoft.  We're stuck
    with either using a deprecated feature (Mirroring), which is only postponing the problem until the end of 2012's life cycle, or laying out significant amounts of money (in the tens of thousands of pounds GBP) to upgrade to Enterprise Edition, which we just
    don't have the budget for.
    Why couldn't Microsoft continue to offer a high availability option for businesses without deep pockets?  Do I have any options I'm not thinking of?  Shared storage is not an option (too costly, single point of failure, geographically separated
    datacentres).

    Thanks for all the feedback.
    I was forgetting that even data stored as FILESTREAM would need to be backed up using SQL backups, so I guess that's an issue either way, thank for reminding me, Sebastian.
    Geoff, I agree that replication isn't a viable HA solution.  FCI has lots going for it, but from what I can make out (and I am just reading up on this now so I could be wrong) requires either shared storage or a third-party tool to move the data from
    the Primary server's local storage to the local storage of the Secondaries.  It all seems overly complicated when Mirroring does exactly what I need it to do, merrily writing transactions to the mirrors at the same time as the primary databases. 
    Additionally, whilst you might be right about making the case for Enterprise Edition, it's difficult to make a clear case to non-technical people when we already have a working HA solution and document management.  Trying to explain the nuanced advantages
    of SQL Server full-text indexing (at present we use a third-party indexing product which stores the index discretely) and simplified querying when the documents are in the same database as their associated data as justification for spending tens of thousands
    of pounds is a challenge!
    David, that's useful to know, about the disadvantages of using FILESTREAM and how just storing the documents in a VARBINARY(MAX) column might actually give better performance, thank you.

  • WLC 5508 - High Availability

    I am upgrading the Wireless Infrastructure with two 5508 WLC.
    I am setting up High Availability, but I think is not quite working.
    Primary Controller = WLC1
    Secondary Controller = WLC2
    LAP = LAP1
    LAP1 has WLC2 as the primary controller for HA
    LAP1 has WLC1 as the secondary controller for HA
    While connected to LAP1, I shutdown WLC2. After ~ 20 seconds, LAP1 move to WLC1.
    I lost connection from LAP1
    Don't LAP1 should move with all its clients to WLC1?
    Am I missing something in my configuration?
    Thanks for all your help in advance

    Hello,
    Let me clarify high availability/failover here. As previously mentioned, you can set the heartbeat times to reduce WLC failure detection time. Once an AP determines it has lost the connection to it's controller, it will perform the discovery/join process to the configured Secondary WLC. During this process, the AP will need to verify configuration from the secondary WLC and the Radios will get reset.
    So, you will not experience a perfectly seamless transition when an AP moves from WLC1 to WLC2. Detection time with defaults is 30 second heartbeat, so I would expect anywhere from 15-45 second outage on a WLC failure scenario. The network will then be operational again and clients should be able to connect.
    Thanks,
    -Patrick Croak
    Wireless TAC

  • JSPM in high availability

    Dear Friends,
    I have installed PI 7.1 in High availability using Microsoft windows MNS based cluster. It was with initial patch level 04 of Netweaver 710.
    I have installed SCS, ASCS, Enque Replication Services etc in cluster mode. I have installed primary application server on Node 2 and Additional Application server on Node 1.
    Now I installed SP07 using JSPM from my Primary Application server i.e. Node 1. It went successfully and my support Pack level was upgraded to Sp07. But now when I open Java stack in Node 1, It shows only home page using url - http://10.6.4.178:50000
    After that if I open any link e.g. UME or NWA, it does not open and throughs following error :
    Service cannot be reached
    What has happened?
    URL http://10.6.4.178:50000/wsnavigator call was terminated because the corresponding service is not available.
    Note
    The termination occurred in system PIP with error code 404 and for the reason Not found.
    The selected virtual host was 0 .
    What can I do?
    Please select a valid URL.
    If it is a valid URL, check whether service /wsnavigator is active in transaction SICF.
    If you do not yet have a user ID, contact your system administrator.
    ErrorCode:ICF-NF-http-c:000-u:SAPSYS-l:E-i:sappicl1_PIP_00-v:0-s:404-r:Notfound
    HTTP 404 - Not found
    Your SAP Internet Communication Framework Team
    Java is started properly on both the nodes without any error. While it is working fine on Node 2 i.e. 10.6.4.179.
    Please suggest ! Do we need to do so special step for installing patch through JSM in High availability setup?
    Jitendra Tayal

    Hi Jitendra,
    While doing a patch upgrade with JSPM in HA scenario, you need to switch off the automatic switchover option because, once the components are deployed, the J2ee gets restarted. If the automatic failover is active, the service gets switched over to the other node.
    In this case, shutdown all the instances and stop the cluster.
    Restart the systems. Start the cluster in the original servers.
    Restart JSPM.
    It should work fine then.
    Let me know if it is helpful.
    Best Regards
    Raghu

Maybe you are looking for