Implementing High Availibility

Hi Friends,
I have to implement a high availability environment for our production database.
This production database is used for automation of plant, like getting input from PLCs and giving inputs to PLCs .
Following are the details of our present setup:
DATABASE Version: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 – Prod (Singel Instance) OS DETAILS:RHEL 3(2.3.21-27.EL)
HARDWARE PLATFORM:HP-Proliant DL380 G3
DISK SYSTEM :RAID 5
Memory Size :2GB
# of CPUs:2
Can any one guide me on this?
Regards.

You might find http://www.oracle.com/technology/deploy/availability/htdocs/maa.htm useful

Similar Messages

  • Do all the weblogic application servers need to share storage in order to implement high availability?

    I would like to understand if a shared storage is required for all application servers that will run weblogic + oracle identity manager.  I've been using the following oracle guides : http://docs.oracle.com/cd/E40329_01/doc.1112/e28391/iam.htm#BABEJGID and http://www.oracle.com/technetwork/database/availability/maa-deployment-blueprint-1735105.pdf and from my interpretation both talk about configuring all the application servers with access to a shared storage.  From an architecture standpoint, does this mean all the application servers need access to an external hard disk?  if shared storage is required what are the steps to implement it?
    Thanks,
    user12107187

    You can do it and it will work. But Fusion Middleware products have EDG to provide guidelines to help you to implement high availability.
    Having a shared storage will help you to recover from storage failure, otherwise this might be a point of failure.
    Enterprise Deployment Overview
    "An Oracle Fusion Middleware enterprise deployment:
    Considers various business service level agreements (SLA) to make high-availability best practices as widely applicable as possible
    Leverages database grid servers and storage grid with low-cost storage to provide highly resilient, lower cost infrastructure
    Uses results from extensive performance impact studies for different configurations to ensure that the high-availability architecture is optimally configured to perform and scale to business needs
    Enables control over the length of time to recover from an outage and the amount of acceptable data loss from a natural disaster
    Evolves with each Oracle version and is completely independent of hardware and operating system "
    Best Regards
    Luz

  • JCo Server High Availability

    How can I make the JCo server implement "High Availability" functionality. The SAP Server which the makes calls to the JCo Server is HA-aware. So if there is a failover, the SAP Server switches over to the other instance but the JCo server keeps sending the message "Server unavailable". Is there a solution for this problem.
    Thanks.

    Single Appliance not necessarily means Single Point of Failure, an appliance with HW Redundancy could handle failure and provide High Availability, if only configured well.
    Does Symantec BrightMail Appliance provide such redundancy configuration?
    You will have to ask their support or in a Symantec Forum.
    Twitter!: Please Note: My Posts are provided “AS IS” without warranty of any kind, either expressed or implied.

  • Ironport S160 high availability deployment

    Dear all,
    I would like to know how to implement high availability on a S160 ironport device.
    i have two S160 device but the user guide is not helping .please provide me a solutions.
    Thanks & Regards;
    Kunal

    Hi,
    have you managed to implement hugh availability with two S160 Ironport devices?
    Is it even possible? I could not find and info on the internet.

  • Highly Available SCOM 2012 R2 Implementation

    Hello All.............................I have following questions with regards to highly available implementation of SCOM 2012 R2:
    1.  Does it rely on Cluster?
    2.  Is shared disk required?
    3.  Is it active/passive or active/active implementation?
    4.  is the failover to the other server automated or manual?
    5.  Any special requirement with respect to active directory?
    6.  Rights required to service account that is setting up the environment?
    7.  Any link to blog/article for step by step instructions would be appreciated.
    Thanks in advance.

    Please see if these helps -
    http://technet.microsoft.com/en-us/library/hh316108.aspx
    http://blogs.technet.com/b/kevinholman/archive/2013/10/18/opsmgr-2012-r2-quickstart-deployment-guide.aspx
    http://adinermie.wordpress.com/2014/04/25/scom-2012-r2-data-access-service-das-high-availability-part-4-enable-nlb-on-each-management-server-add-a-host-to-nlb-cluster/
    Thanks, S K Agrawal

  • 11.1.2 High Availability for HSS LCM

    Hello All,
    Has anyone out there successfully exported an LCM artifact (Native Users) through Shared Services console to the OS file system (<oracle_home>\user_projects\epmsystem1\import_export) when Shared Services is High Available through a load balancer?
    My current configuration is two load balanced Shared Services servers using the same HSS repository. Everything works perfectly everywhere except when I try to export anything through LCM from the HSS web console. I have shared out the import_export folder on server 1. And I have edited the oracle_home>\user_projects\epmsystem1\config\FoundationServices1\migration.properties file on the second server to point to the share on the first server. Here is a cut and paste of the migration.properties file from server2.
    grouping.size=100
    grouping.size_unknown_artifact_count=10000
    grouping.group_by_type=Y
    report.enabled=Y
    report.folder_path=<epm_oracle_instance>/diagnostics/logs/migration/reports
    fileSystem.friendlyNames=false
    msr.queue.size=200
    msr.queue.waittime=60
    group.count=10000
    double-encoding=true
    export.group.count = 30
    import.group.count = 10000
    filesystem.artifact.path=\\server1\import_export
    I perform an export of just the native users to the file system, the export fails stating and I find errors in the log stating the following.
    [2011-03-29T20:20:45.405-07:00] [EPMLCM] [NOTIFICATION] [EPMLCM-11001] [oracle.EPMLCM] [tid: 10] [ecid: 0000Iw4ll2fApIOpqg4EyY1D^e6D000000,0] [arg: /server1/import_export/msr/PKG_34.xml] Executing package file - /server1/import_export/msr/PKG_34.xml
    [2011-03-29T20:20:45.530-07:00] [EPMLCM] [ERROR] [EPMLCM-12097] [oracle.EPMLCM] [tid: 10] [ecid: 0000Iw4ll2fApIOpqg4EyY1D^e6D000000,0] [SRC_CLASS: com.hyperion.lcm.clu.CLUProcessor] [SRC_METHOD: execute:?] Cannot find the migration definition file in the specified file path - /server1/import_export/msr/PKG_34.xml.
    [2011-03-29T20:20:45.546-07:00] [EPMLCM] [NOTIFICATION] [EPMLCM-11000] [oracle.EPMLCM] [tid: 10] [ecid: 0000Iw4ll2fApIOpqg4EyY1D^e6D000000,0] [arg: Completed with failures.] Migration Status - Completed with failures.
    I go and look for the path that it is searching for "/server1/import_export/msr/PKG_34.xml" and find that this file does exist, it was infact created by the export process so I know that it is able to find the correct location, but it then says that it can't find it after the fact. I have tried creating mapped drives and editing the migration.properties file to reference the mapped drive but it gives me the same errors but with the new path. I also tried the below article that I found in support with no result. Please advise, thank you.
    Bug 11683769: LCM MIGRATIONS FAIL WHEN HSS IS SET UP FOR HA AND HSS IS STARTED AS A SERVICE
    Bug Attributes
    Type B - Defect Fixed in Product Version 11.1.2.0.000PSE
    Severity 3 - Minimal Loss of Service Product Version 11.1.2.0.00
    Status 90 - Closed, Verified by Filer Platform 233 - Microsoft Windows x64 (64-bit)
    Created 25-Jan-2011 Platform Version 2003 R2
    Updated 24-Feb-2011 Base Bug 11696634
    Database Version 2005
    Affects Platforms Generic
    Product Source Oracle
    Related Products
    Line - Family -
    Area - Product 4482 - Hyperion Lifecycle Management
    Hdr: 11683769 2005 SHRDSRVC 11.1.2.0.00 PRODID-4482 PORTID-233 11696634Abstract: LCM MIGRATIONS FAIL WHEN HSS IS SET UP FOR HA AND HSS IS STARTED AS A SERVICE*** 01/25/11 05:15 pm ***Hyperion Foundation Services is set up for high availability. Lifecycle management migrations fail when running Hyperion Foundation Services - Managed Server as a Windows Service. We applied the following configuration changes:Set up shared disk then we modify the migration.properties.filesystem.artifact.path=\\<servername>\LCM\import_export We are able to get the migrations to work if we set filesystem.artifact.path=V:\import_export (a maped drive to the shared disk) and we run Hyperion Foundation Services in console mode.*** 01/25/11 05:24 pm *** *** 01/28/11 03:32 pm *** (CHG: Sta->11)*** 01/28/11 03:32 pm *** (CHG: Comp. Ver-> 11.1.2 -> 11.1.2.0.00)*** 01/28/11 03:34 pm *** (CHG: Base Bug-> NULL -> 11696634)*** 02/02/11 01:53 pm *** (CHG: Sta->41)*** 02/02/11 01:53 pm ****** 02/02/11 05:18 pm ****** 02/03/11 10:30 am ****** 02/04/11 03:53 pm *** *** 02/04/11 03:53 pm *** (CHG: Sta->80 Asg->NEW OWNER)*** 02/10/11 08:49 am ****** 02/15/11 11:59 am ****** 02/15/11 04:56 pm ****** 02/16/11 09:58 am *** (CHG: Sta->90)*** 02/16/11 09:58 am ****** 02/16/11 02:12 pm *** (CHG: Sta->93)*** 02/24/11 06:14 pm *** (CHG: Sta->90)Back to top

    it is not possible to implement a kind of HA between two different appliances 3315 and 3395. 
    A node in HA can have the 3 persona. 
    Suppose Node A (Admin(primary), monitoring(Primary) and PSN). 
    Node B(Admin(Secondary), Monitoring(Secondary) and PSN). 
    If the Node A is unavailable, you will have to promote manually the Admin role to Primary. 
    Although the best way is to have
    Node A Admin(Primary), Monitoring(Secondary) and PSN
    Node B Admin(Secondary), Monitoring (Primary) and PSN. 
    Rate if helpful and Marked As correct if it is correct for the experts. 
    Regards,

  • Hyper-V 2012 High Availability using Windows Server 2012 File Server Storage

    Hi Guys,
    Need your expertise regarding hyper-v high availability. We setup 2 hyper-v 2012 host in our infra for our domain consolidation project. Unfortunately, we don't have hardware storage that is said to be a requirement that we can use to create a failover cluster
    for hyper-v host to implement HA. Here's the setup:
    Host1
    HP Proliant L380 G7
    Windows Server 2012 Std
    Hyper-V role, Failover Cluster Manager and File and Storage Services installed
    Host2
    Dell PowerEdge 2950
    Windows Server 2012 Std
    Hyper-V role, Failover Cluster Manager and File and Storage Services installed
    Storage
    Dell PowerEdge 6800
    Windows Server 2012 Std
    File and Storage Services installed
    I'm able to configure the new feature about Shared Nothing Live Migration - i'm able to move VM's back and fort between my hosts without storage. But this is planned and proactive approach. My concern is to have my hyper-v host to become highly available in
    the event of system failure. If my host1 dies, the VMs should go/move to host2 and vice versa. In setting this up, i believe i need to enable the failover clustering between my hyper-v hosts which i already did but upon validation, it says "No disks
    were found on which to perform cluster validation tests." Is it possible to cluster it using just a regular windows file server? I've read about SMB 3.0 and i've configured it as well i'm able to save VMs on my file server, but i don't think that my hyper-v
    hosts are already highly available.
    Any feedback and suggestions or recommendation is highly appreciated. Thanks in advance!

    Hi Guys,
    Need your expertise regarding hyper-v high availability. We setup 2 hyper-v 2012 host in our infra for our domain consolidation project. Unfortunately, we don't have hardware storage that is said to be a requirement that we can use to create a failover cluster
    for hyper-v host to implement HA. Here's the setup:
    Host1
    HP Proliant L380 G7
    Windows Server 2012 Std
    Hyper-V role, Failover Cluster Manager and File and Storage Services installed
    Host2
    Dell PowerEdge 2950
    Windows Server 2012 Std
    Hyper-V role, Failover Cluster Manager and File and Storage Services installed
    Storage
    Dell PowerEdge 6800
    Windows Server 2012 Std
    File and Storage Services installed
    I'm able to configure the new feature about Shared Nothing Live Migration - i'm able to move VM's back and fort between my hosts without storage. But this is planned and proactive approach. My concern is to have my hyper-v host to become highly available in
    the event of system failure. If my host1 dies, the VMs should go/move to host2 and vice versa. In setting this up, i believe i need to enable the failover clustering between my hyper-v hosts which i already did but upon validation, it says "No disks were
    found on which to perform cluster validation tests." Is it possible to cluster it using just a regular windows file server? I've read about SMB 3.0 and i've configured it as well i'm able to save VMs on my file server, but i don't think that my hyper-v hosts
    are already highly available.
    Any feedback and suggestions or recommendation is highly appreciated. Thanks in advance!
    Your shared storage is a single point of failure with this scenario so I would not consider the whole setup as a production configuration... Also setup is both slow (all I/O is travelling down the wire to storage server, running VMs from DAS is ages faster)
    and expensive (third server + extra Windows license). I would think twice about what you do and either deploy a built-in VM replication technologies (Hyper-V Replica) and apps built-in clustering features that does not require shared storage (SQL Server and
    Database Mirroring for example, BTW what workload do you run?) or use some third-party software creating fault tolerant shared storage from DAS or investing into physical shared storage hardware (HA one of course). 
    Hi VR38DETT,
    Thanks for responding. The hosts will cater a domain controller (on each host), Web filtering software (Websense), Anti-Virus (McAfee ePO), WSUS and an Auditserver as of the moment. Is the Hyper-V Replica somewhat give "high availability" to VMs or Hyper-V
    hosts? Also, is the cluster required in order to implement it? Haven't tried that but worth a try.

  • Best practice for High availability design, HSRP

    Hi,
    I am planning to create High Availability for LAN to WAN connectivity.
    But I want to know your opinion about the best way how to do this. I googled for a solution/best way how to do this, but I didn't found in my opinion right answer.
    The situation:
    I have 2 3945E Routers and 2 3560 switches. The design that I am planning to implement is below.
    The main goal is to have redundant connection, whatever one of the devices will fail. For example, if the R1 will fail, R2 should become active, if the SW1 will fail, the SW2 will take care about reachability and vice versa. The router 1 should be preferred always, if the link to ISP isn't down, because of greater bandwidth. So why am I drown 2 connections to 2 separate switches. If the SW1 will fail, I will still have a connection to WAN using R1 router.
    The Router interface should be configured with sub interfaces (preferred over secondary IP address of interface), because more than 10 subnets will be assigned to the LAN segment. The routers have 4 Gi ports.
    HSRP must be enabled on LAN side, because PC's on LAN must have redundant def. getaway.
    So, the question is - what is the best and preferred way to do this?
    In my opinion, I should use BVI and combine R1 routers 2 interfaces in to logical one and do the same for the R2.
    Next, turn the router in to L3 switch using IRB and then configure HSRP.
    What would be your preferred way to do this?

    Hi Audrius,
    I would suggest you to go with HSRP. GLBP you will use where you want load balance.
    I think the connectivity between your Routers (3945) and switches (3560) is gigabit connection which is high speed. So keep one physical link from your switches to each router and do HSRP on those router physical interfaces.
    In this way you will have high availability like if R1 fails then R2 will take over.
    Regarding the config see the below which I have for one of my Customer DC.
    ACTIVE:
    track 1 interface GigabitEthernet0/0 line-protocol
    track 2 interface GigabitEthernet0/0 line-protocol
    interface GigabitEthernet0/1
    ip address 10.10.10.12 255.255.255.0
    ip nat inside
    ip virtual-reassembly
    duplex full
    speed 100
    standby use-bia scope interface
    standby 0 ip 10.10.10.10
    standby 0 priority 110
    standby 0 preempt
    standby 0 authentication peter2mo
    standby 0 track 1 decrement 30
    standby 0 track 2 decrement 30
    STANDBY:
    track 1 interface GigabitEthernet0/0 line-protocol
    interface GigabitEthernet0/1
    ip address 10.10.10.11 255.255.255.0
    ip nat inside
    ip virtual-reassembly
    duplex full
    speed 100
    standby use-bia scope interface
    standby 0 ip 10.10.10.10
    standby 0 priority 90
    standby 0 authentication peter2mo
    standby 0 track 1 decrement 30
    Please rate the helpfull posts.
    Regards,
    Naidu.

  • Windows 2012 RDS - Session Host servers High Availability

    Hello Windows/Terminal server Champs,
    I am new middle of implementing RDS environment for one of my customer, Hope you could help me out.
    My customer has asked for HA for RDS session host where applications are published, and i have prepared below plan for server point of view.
     2 Session Host server, 1 webaccess, 1 License/connection
    Broker & 1 Gateway (DMZ).
     In first Phase, we are planning to target internal user
    who connect to Session host HA where these 2 servers will have application installed and internal user will use RDP to access these application.
    In second Phase we will be dealing with external Party who connect from external network where we are planning to integrate with NetIQ => gateway
    => Webaccess/Session host
     I have successfully installed and configured 2 Session
    Host, 1 license/Broker. 1 webAccess & 1 Gateway. But my main concern to have session Host High Available as it is hosting the application and most of the internal user going to use it. to configure it i am following http://technet.microsoft.com/en-us/library/cc753891.aspx  
    However most of the Architecture is change in RDS 2012. can you please help me out to setup the Session Host HA.
    Note: we can have only 1 Connection broker /Licensing server , 1 webacess server & 1 Gateway server, we cannot increase more server due to cost
    factor.
     thanks in advance.

    Yes, absolutely no problem in just using one connection broker in your environment as long as your customer understands the SPOF.
    the session hosts however aren't really what you would class HA - but to set them up so youhave reduancy you would use either Windows NLB, an external NLB device or windows dns round robin. My preferred option when using the connection broker is DNS round
    robin - where you give each server in the farm the same farm name dns entry - the connection broker then decides which server to allocate the session too.
    You must ensure your session host servers are identical in terms of software though - same software installed in the same paths on all the session host servers.
    if you use the 2012 deployment wizard through server manager roles the majority of the config is done for you.
    Regards,
    Denis Cooper
    MCITP EA - MCT
    Help keep the forums tidy, if this has helped please mark it as an answer
    My Blog
    LinkedIn:

  • HTTP Server High Availability

    Hello All.
    I have a question regarding OC4J and HTTP server High Availability.
    I want to do something like the Figure 3-1 of the Oracle Application Server High Availability Guide 10.1.2. See this link
    http://download-east.oracle.com/docs/cd/B14099_11/core.1012/b14003/midtierdesc.htm#CIHCEDFC
    What I have now is the following:
    Three hosts
    Two of them are an OAS 10.1.2 which we already configured the Cluster and deployed our applications (used this tutorial: http://www.oracle.com/technology/obe/obe_as_1012/j2ee/deploy/j2eecluster/farmcluster.htm)
    Let's say this nodes are:
    - host1
    - host2
    The other one is the Oracle WebCache stand alone (will act as Load Balancer). We will call this
    - hostwc3
    We already configured the WebCache as Load Balancer and is working just fine. We also configured the session replication successful and work great with our applications.
    What we have not clear is the following:
    When a client try to visit http://hostwc3/application/ the LOAD BALANCER routes him to, let's say http://host1/application/ and in the browser's URL will not show the Virtual Server anymore (the webcache server) and will show the actual real Apache address (host1 )that is attending him. IF we "kill" on ENTIRE host1 (apache, oc4j, etc..) the clients WILL perceive the down and if they try to press F5, the will try to access to an Apache that doesn't is up and running.... The behavior expected is that the browser NEVER shows the actual Apache URL, so, when some apache goes down, the client do not disconnect (as it happens with and OC4J downfall ) and always works with the "virtual web server".
    I came up with some ideas but I want you Guys to give me an advice:
    - In Web Cache, do not route for load balancing to Apache, and route the Oc4J directly (Is this possible?)
    - Configure a HTTP Server Cluster, this means that we have to have a "Virtual Name"to the Apaches (two of them). Is this possible? how?
    - Use the rewrite mode of the Apache. Is this a good idea?
    - Any other idea how to fix the Apache "Single Point Of Failure" ?
    According with the figure 3-1 ( Link above ) we do can have HTTP Server in a cluster. But I have no idea how to manage it or configure it.
    Thanks in advance any help!

    You cannot point Outlook Anywhere to your DAG cluster IP address. It must be pointed to the actual IP address of either server.
    For no extra cost DNS round robin is the best you will get, but it does have some drawbacks as it may give the IP address of a server you have taken down for maintenance or the server has an issue.
    You could look to implement a load balancer but again if you are doing this for high availability then you want more than one load balancer in the cluster - otherwise you've just moved your single point of failure.
    Having your existing NAT and just remembering to update it to point to the other server during maintenance may suit your needs for now.
    If you can go into more detail about what the high availability your business is looking to achieve and the budget we can suggest the best method to meet those needs for the price point.
    Have a great day
    Oliver
    Oliver Moazzezi | Exchange MVP, MCSA:M, MCITP:Exchange 2010,Exchange 2013, BA (Hons) Anim | http://www.exchange2010.com | http://www.cobweb.com | http://twitter.com/OliverMoazzezi

  • Oracle EBS R12.1.3 High Availability with RAC 11g on VMWARE

    Dear All,
    Our customer requirement is of high availability and good utilization of hardware for EBS R12.13 implementation.As per our strategy we want to use Oracle VM(I think it doesn't require any licensing) , RAC 11gR2 and appstier load balancing.We have only two servers to achieve all these.One one node there will be two VMs one for dbtier and another one for appstier and same will be on second server.So this way we will be having four virtual machines on two physical nodes.For all these we will use OEL Linux as Operating system
    Please share your experiences , if above deployment model is correct or need further enhancements.
    Regards.

    Our customer requirement is of high availability and good utilization of hardware for EBS R12.13 implementation.As per our strategy we want to use Oracle VM(I think it doesn't require any licensing) , Please contact your Oracle Sales representative for license questions.
    RAC 11gR2 and appstier load balancing.We have only two servers to achieve all these.One one node there will be two VMs one for dbtier and another one for appstier and same will be on second server.So this way we will be having four virtual machines on two physical nodes.For all these we will use OEL Linux as Operating system
    Please share your experiences , if above deployment model is correct or need further enhancements.Please see these docs.
    Using Oracle VM Templates for Oracle E-Business Suite [ID 975734.1]
    Using Oracle VM with Oracle E-Business Suite Release 11i or Release 12 [ID 465915.1]
    Oracle E-Business Suite High Availability on Oracle VM [ID 986690.1]
    Certified Software on Oracle VM [ID 464754.1]
    Oracle VM General Frequently Asked Questions [ID 464756.1]
    Thanks,
    Hussein

  • SAP PI 7.1 High Availability, Active-Active or Active-Passive

    We are planning go to HA system for PI 7.11
    I've few doubts/questions regarding this.
    Can we make Active-Active clustering or Active-Passive? Or itu2019s our choice?
    If it is our choice, then is at Configuration level (Application level)? or hardware level?
    What are additional steps we need to take for Active-Active clustering? (Which we are looking for)
    What is recommended by SAP?
    Any Pros & Cons.
    Please reply me if you have any information regarding this, any information regards to this is greatly appreciated
    Thanks
    Sagar

    Hi Sagar,
    Please use Active-Passive for your SAP PI 7.1 High Availability system.
    For more details, please refer to the sdn link:
    High Availability
    ->High Availability FAQ  
    ->What is an "active-active cluster"?
    An active-active cluster consists of two independent server nodes at a minimum. The workload within a cluster resource is shared between the server nodes. If a cluster node crashes the processes are resumed by the remaining cluster nodes. An active-active cluster configuration means that a cluster resource is active on all cluster nodes. The aim of an active-active cluster is not only to provide high availability system but to distribute the workload between the cluster nodes. Applications with a very high workload like databases benefit from an active-active setup. Due to the SAP Central Services Instance is a light component an active-active setup does not make any sense. Therefore the SCS is implemented as an active-passive cluster resource.
    Best Regards
    Julia

  • Migrating Dimension Server from local installation to High Availability

    Hi Team,
    When Hyperion was installed on the client site (11.1.2.3), there was no requirement for EPMA to become highly available.
    Now that has become a requirement to make it highly available in the event that one of the virtual machine is unreachable.
    In the event that a server becomes unavailable, it is acceptable for a manual failover procedure to be done.
    What steps can be taken to prepare for such an event?
    Possible outcome:
    Configure another dimension server on the second Virtual machine and leave it manually switched off.
    Modify the IIS path to point to a clustered location accessible by both machines.
    In the event of a failover, manually bring up the dimension server on the other machine.
    This is yet to be implemented successfully.
    Current Scenerio:
    1 x Windows Machine hosting:
      Foundation Services
      Calculation Manager
      Performance Architect
      Planning
      Reporting and Analysis
      Repository location on a shared mount (Clustered filesystem)
      FDQM
      Essbase
      Analytic Provider Services
      Essbase Integration Services (not used)
      Essbase Studio (not used)
    1 x Linux Machine
      Essbase Server
      Arborpath on a shared mount
    Aim: Add 2 new Virtual Machines ( 1 x Windows App Server + 1 x Linux Essbase Server) for high availability.
    Progress:
      Windows Machine components clustered:
      Foundation Services
      Calculation Manager
      Planning
      Reporting and Analysis
      Essbase
      Analytic Provider Services
      Linux Essbase Machines:
      Configured Active/Passive using OPMN
    Challenge faced:
    Failover scenerio can be a manual list of steps to provide functionaility.
    Please provide support on introducing failover steps for dimension server.

    Remote Desktop/Preferences/Task Server
    Add the FQDN or I.P address of your server in this pane on your book(basically your pointing your book to your installed software on the other machine). Setup your server with the installs you want it to make, and administer it via your book (or any machines with another unlimited ARD) IF you want report generation, you would schedule the clients to send the info to the server. A task server can only do 2 things, install packages or change client settings, thats it.

  • Multi Site SQL High Availability group

    Hi,
    We have our primary Data Centre where we have 2 MS SQL 2012 Enterprise Servers and 2 SharePoint 2013 Enterprise servers.
    We have SQL High Availability group implemented on our primary Site (Site A)
    Site A has subnet of 192.168.0.0.
    We recently added a new DR Site (Site B). Site B we have MS SQL 2012 Enterprise Servers.  Site B have a subnet of
    172.1.1.0
    Both sites are connected via a VPN Tunnel. MS SQL 2012 Enterprise Server on Site B has been added in have SQL High Availability
    group of Site A.
    SQL High Availability group have 2 IPs 192.168.0.32 and 172.1.1.32. SQL Listener have 2 IPs 192.168.0.33 and 172.1.1.33
    We want to make sure that if Site A completely down then Site B work as an active Site. But when Site A down we are unable
    to ping High Availability group and Site B is unable to work as active site. SQL and SharePoint services are completely down.
    Site A has AD(Primary Domain  Controller) and Site B has ADC(Additional Domain Controller).
     Site A has witness server.
    We are using Server 2012 Data Centre
    Please suggest.
    Farooq

    SharePoint is not the same as any other applications. The DR site has to be a completely different farm from your production. This means that the SharePoint_AdminContent and config databases on both farms are different and should not be included in the Availability
    Group (they do not support asynchronous Availability Group.) Only content databases and other SharePoint databases supported in an Availability Group should be included as per this
    TechNet article. Have a look at this
    blog post for your networking configuration.
    The reason your Windows cluster service goes down in the DR data center when the when your primary data center goes down is because the cluster has no majority votes. You have 4 available votes in your cluster - 3 in the primary data center in the form of
    the 2 Availability Group replicas and the witness server and 1 in the DR data center. If the Windows cluster does not have majority of votes, it automatically shuts down the cluster. This is by design. That is why when this situation happens, you need to force
    the cluster to start without a quorum as described in
    this article (Windows Server 2012 R2 introduced the concept of dynamic witness and dynamic quorum to address this concern.) By default, the IP address of the Availability Group listener name will be registered on all of the DNS servers if the cluster is
    running Windows Server 2012 and higher (this is the RegisterAllProvidersIP property that Samir mentioned.) However, you need to flush the DNS cache on the SharePoint web and application servers and point them to the correct IP address on the DR server after
    failover. This is because the default TTL value of the Availability Group listener name is 20 minutes. 
    It is important to define your DR strategies and objectives when designing the architecture to make sure that the solution will meet your goals. This will also help you automate some of the processes in the DR data center when failover or a disaster occurs.
    Edwin Sarmiento SQL Server MVP | Microsoft Certified Master
    Blog |
    Twitter | LinkedIn
    SQL Server High Availability and Disaster Recover Deep Dive Course

  • High Availability Options Without Enterprise Edition

    I think I've found the answer to this already but I just wanted to confirm that I'm getting this right. In SQL Server 2012, is there no way to implement a high-availability solution which doesn't require shared storage or use of a deprecated feature without
    having Enterprise Edition?
    At the moment we have databases hosted on a primary server running Windows Server 2012 Standard and SQL Server 2012 Standard.  We have a second, identical server, and the databases are mirrored between the two servers using Database Mirroring. 
    This works perfectly and meets all of our requirements.  The only possible problem is that we're looking at storing documents using FILESTREAM, which isn't available in Mirrored databases.
    From what I've read, Database Mirroring is deprecated in SQL Server 2012.  High Availability Groups, which sound great, are only available in the Enterprise Edition.  This feels like a real kick in the teeth from Microsoft.  We're stuck
    with either using a deprecated feature (Mirroring), which is only postponing the problem until the end of 2012's life cycle, or laying out significant amounts of money (in the tens of thousands of pounds GBP) to upgrade to Enterprise Edition, which we just
    don't have the budget for.
    Why couldn't Microsoft continue to offer a high availability option for businesses without deep pockets?  Do I have any options I'm not thinking of?  Shared storage is not an option (too costly, single point of failure, geographically separated
    datacentres).

    Thanks for all the feedback.
    I was forgetting that even data stored as FILESTREAM would need to be backed up using SQL backups, so I guess that's an issue either way, thank for reminding me, Sebastian.
    Geoff, I agree that replication isn't a viable HA solution.  FCI has lots going for it, but from what I can make out (and I am just reading up on this now so I could be wrong) requires either shared storage or a third-party tool to move the data from
    the Primary server's local storage to the local storage of the Secondaries.  It all seems overly complicated when Mirroring does exactly what I need it to do, merrily writing transactions to the mirrors at the same time as the primary databases. 
    Additionally, whilst you might be right about making the case for Enterprise Edition, it's difficult to make a clear case to non-technical people when we already have a working HA solution and document management.  Trying to explain the nuanced advantages
    of SQL Server full-text indexing (at present we use a third-party indexing product which stores the index discretely) and simplified querying when the documents are in the same database as their associated data as justification for spending tens of thousands
    of pounds is a challenge!
    David, that's useful to know, about the disadvantages of using FILESTREAM and how just storing the documents in a VARBINARY(MAX) column might actually give better performance, thank you.

Maybe you are looking for

  • No video stream when calling 3rd party endpoint from Jabber client

    I use my own H323 client based on the H323Plus open source. It works very well with VCS and all endpoints we tried (a dozen or so from every vendor). We have a problem with Jabber Video though, across all versions from MOVI to 4.4.3. When my own clie

  • How do I save changes made to a sub vi automatically

    Hello Fellow Engineers! I have a doubt and welcome useful suggestions from all of you. I have the following problem: I am programming a vi that asks a user for his/her username and password. The user has an option of changing the password . When he/s

  • Help Converting old AS2 code to AS3

    I just updated an old flash file with some new photos.  I don't know much about Flash - enough to change the photos and hit export BUT now I'm getting errors.  I had a _root issue which I fixed thanks to this forum but I'm still getting these errors:

  • How to get the offset value in the Bex customer exit?

    Hello Friends, I have defined an offset on a variable in the query definition. I wish to capture this offset value  in the Bex variable customer exit . Does anyone how to get this?? Thanks, Gautam

  • Truncate table lock issue

    Hi,     In my SQL2000 stored procedure, there are 9 truncate table commands.      But the the stored procedure is take above 15 minutes to execute.      In this stored procedure, truncate table command has no problem.      But an insert query into th