Question on replication/high availability designs

We're currently trying to work out a design for a high-availability system using Oracle 9i release 2. Having gone through some of the Oracle whitepapers, it appears that the ideal architecture involves setting up 2 RAC sites using Dataguard to synchronize the data. However, due to time and financial constraints, we are only allowed to have 2 servers for hosting the databases, which are geographically separate from each other in prevention of natural disasters. Our app servers will use JDBC pools to connect to the databases.
Our goal is to have both databases be the mirror image of each other at any given time, and the database must be working 24/7. We do have a primary and a secondary distinction between the two, so if the primary fails, we would like the secondary database to take over the tasks as needed.
The ability to query existing data is mission critical. The ability to write/update the database is less important, however we do need the secondary to be able to process data input/updates when primary is down for a prolonged period of time, and have the ability to synchronize back with the primary site when it is back up again.
My question now is which replication technology should we try to implement? I've looked into both Oracle Advanced Replication and Dataguard, each seems to have its own advantages and drawbacks:
Replication - can easily switch between the two databases using multimaster implementation, however data recovery/synchronization may be difficult in case of failure, and possibly will lose data (pending implementation). There has been a few posts in this forum that suggested that replication should not really be considered as an option for high availability, why is that?
Dataguard - zero data loss in failover/switchover, however manual intervention is required to initiate failover/switchover. Once the primary site fails over to the standby, the standby becomes the primary until DBA manually goes back in and switch the roles. In Oracle 10g release 2, seems that automatic failover is achieved through the use of an extra observer piece. There does not seem to be anyway to do this in Oracle 9i release 2.
Being new to the implementation of high-availability systems, I am at somewhat of a loss at this point. Both implementations seem to be a possible candidate, but we will need to sacrifice some efforts for both of them also. Would anyone shine some light on this, maybe point out my misconceptions with Advanced Replication and Dataguard, and/or suggest a better architecture/technology to use? Any input is greatly appreciated, thanks in advance.
Sincerely,
Peter Tung

Hi,
It sounds as if you're talking about the DB_TXN_NOSYNC flag, rather than DB_NOSYNC.
You mention that in general, you lose uncommitted transactions on system failure. I think what you mean is that you may lose some committed transactions on system failure. This is correct.
It is also correct that if you use replication you can arrange to have clients have a copy of all committed transactions, so that if the master fails (and enough clients do not fail, of course) then the clients still have the transaction data, even when using DB_TXN_NOSYNC.
This is a very common usage scenario for Berkeley DB replication/HA, used to achieve high throughput. You will want to pay attention to the configured ack policy, group size setting, setting of the 2SITE_STRICT option (if group size == 2).

Similar Messages

  • Question about using DB_NOSYNC when using replication (high availability)

    Hi,
    I am using Berkeley db with replication (high availability ).
    I want to ask if i use the DB_NOSYNC flag would be any problem with the Durability of the system?
    I know that if there is no sync, data is not stored from memory to hard disk and i will lose uncommited transactions on system failure. Right?
    But i read in a thread that if you use Replication, all transactions pass to clients so there no loss if master server fails?
    I would like your knowledge about this issue.!

    Hi,
    It sounds as if you're talking about the DB_TXN_NOSYNC flag, rather than DB_NOSYNC.
    You mention that in general, you lose uncommitted transactions on system failure. I think what you mean is that you may lose some committed transactions on system failure. This is correct.
    It is also correct that if you use replication you can arrange to have clients have a copy of all committed transactions, so that if the master fails (and enough clients do not fail, of course) then the clients still have the transaction data, even when using DB_TXN_NOSYNC.
    This is a very common usage scenario for Berkeley DB replication/HA, used to achieve high throughput. You will want to pay attention to the configured ack policy, group size setting, setting of the 2SITE_STRICT option (if group size == 2).

  • Best practice for High availability design, HSRP

    Hi,
    I am planning to create High Availability for LAN to WAN connectivity.
    But I want to know your opinion about the best way how to do this. I googled for a solution/best way how to do this, but I didn't found in my opinion right answer.
    The situation:
    I have 2 3945E Routers and 2 3560 switches. The design that I am planning to implement is below.
    The main goal is to have redundant connection, whatever one of the devices will fail. For example, if the R1 will fail, R2 should become active, if the SW1 will fail, the SW2 will take care about reachability and vice versa. The router 1 should be preferred always, if the link to ISP isn't down, because of greater bandwidth. So why am I drown 2 connections to 2 separate switches. If the SW1 will fail, I will still have a connection to WAN using R1 router.
    The Router interface should be configured with sub interfaces (preferred over secondary IP address of interface), because more than 10 subnets will be assigned to the LAN segment. The routers have 4 Gi ports.
    HSRP must be enabled on LAN side, because PC's on LAN must have redundant def. getaway.
    So, the question is - what is the best and preferred way to do this?
    In my opinion, I should use BVI and combine R1 routers 2 interfaces in to logical one and do the same for the R2.
    Next, turn the router in to L3 switch using IRB and then configure HSRP.
    What would be your preferred way to do this?

    Hi Audrius,
    I would suggest you to go with HSRP. GLBP you will use where you want load balance.
    I think the connectivity between your Routers (3945) and switches (3560) is gigabit connection which is high speed. So keep one physical link from your switches to each router and do HSRP on those router physical interfaces.
    In this way you will have high availability like if R1 fails then R2 will take over.
    Regarding the config see the below which I have for one of my Customer DC.
    ACTIVE:
    track 1 interface GigabitEthernet0/0 line-protocol
    track 2 interface GigabitEthernet0/0 line-protocol
    interface GigabitEthernet0/1
    ip address 10.10.10.12 255.255.255.0
    ip nat inside
    ip virtual-reassembly
    duplex full
    speed 100
    standby use-bia scope interface
    standby 0 ip 10.10.10.10
    standby 0 priority 110
    standby 0 preempt
    standby 0 authentication peter2mo
    standby 0 track 1 decrement 30
    standby 0 track 2 decrement 30
    STANDBY:
    track 1 interface GigabitEthernet0/0 line-protocol
    interface GigabitEthernet0/1
    ip address 10.10.10.11 255.255.255.0
    ip nat inside
    ip virtual-reassembly
    duplex full
    speed 100
    standby use-bia scope interface
    standby 0 ip 10.10.10.10
    standby 0 priority 90
    standby 0 authentication peter2mo
    standby 0 track 1 decrement 30
    Please rate the helpfull posts.
    Regards,
    Naidu.

  • High Availability design/upgrade

    Hi,
    We’re looking to upgrade our Production SQL Cluster infrastructure. Currently we are running on old hardware and software….We currently have two nodes in the cluster. 2-Proliant
    DL380 Gen5 servers connected to our HP MSA2000 for shared storage(Direct Attached Storage) using a SAS cable with 1 terabyte. Windows 2008 R2 enterprise with Clustering turned-on and SQL 2005 server. The DB isn't that
    big, but it is crucial to our company….  I was curious to get some opinions what would be our best options to upgrade….. We have two spare physical Proliant DL380 Gen 8 servers w/ 64 gigs of ram and total of 1.5 terabyte of worth of disk space. We would
    have all the HD bay’s with HD’s therefore no need for a new HP MSA /direct attached storage….. So should we go with Windows 2012 Enterprise with SQL Enterprise and turnon SQL  Always on Availability Group? Or any other recommendations? Any suggestion
    and comment would be greatly appreciated.

    Hello,
    SQL Server AlwaysOn Availability Groups requirement Windows Server Failover Clustering (WSFC) clusters.Each availability replica of a given availability group must reside on a different node of the same WSFC cluster
    Based on your description, it seems that you had two node cluster with Windows Server 2008 R2. If so, you should install windows 2008 R2 SP1 for supporting  SQL Server AlwaysOn Availability Groups.
    Reference:http://msdn.microsoft.com/en-us/library/ff878487.aspx 
    Regards,
    Fanny Liu
    Fanny Liu
    TechNet Community Support

  • EBS High Availability design

    Hi Gurus,
    What is the best way of configuring HA on EBS (APPS) side the DB side is running on an Exadata machine.I have 2 Sparc T-4s that i can use for the APPS.MOS Notes appreciated.

    Solaris Cluster for EBS
    Oracle Solaris Cluster Data Service for Oracle E-Business Suite Guide
    October 2012, E29648-01
    http://docs.oracle.com/cd/E29086_01/html/E29648/index.html
    Using Oracle 11g Release 2 Real Application Clusters with Oracle E-Business Suite Release 12 [ID 823587.1]
    Interoperability Notes Oracle EBS R12 with Oracle Database 11gR2 (11.2.0.2) [ID 1367644.1]
    Oracle EBS R12 with Database Tier Only Platform on Oracle Database 11.2.0 [ID 456347.1]
    Check for Split configuration, to have APPS EBS on Solaris SPARC and DB on Linux

  • ACS 5.2 software high availability

    Hi,
      Can i setup primary/backup setup with ACS 5.2 software on Vmware ? Will replication/high availability happen just like if using two appliances ? I need to setup two ACS5.2 softwares on different Vmware machines.
    regards
    Joe

    Hi Joe,
    Yes that is possible. just make sure you meeting requirements mentioned on this document:
    http://www.cisco.com/en/US/docs/net_mgmt/cisco_secure_access_control_system/5.2/installation/guide/csacs_vmware.html
    Thanks
    Serge

  • Timesten high availability question

    I have a case presented here and wanted to know if it is actually possible to implement.
    Let us consider four nodes with timesten (11.2) installed in all of them. A datastore with the same name is created on each of the four servers. Two of these servers are in location A and the other two in location B. Servers in location A have replication defined between them and similarly servers in B have replication defined between them. But note that there is no replication defined between any server in location A with any server in location B.
    The basic idea of this entire setup is to maintain high availability of timesten at any point of time (in case of natural disasters, etc..) irrespective of the location of the servers
    Now, we have oracle software installed in four other systems. Two of these servers are in location A and the other two are in location B. Note that they are not installed on the same box as Timesten.
    Scenario 1:
    Question: Timesten in location A goes down, how is the high availability taken care of?
    Answer: Timesten in the other server in location A should come up and because of the replication process, this will solve the problem.
    Is this correct? I think it is.
    Scenario 2:
    Question: Timesten installed on both the nodes at location A go down, how is high availability taken care of?
    Answer: ?
    Please remember from above that timesten does not have a replication policy defined between any server in location A with any server in location B. The requirement says that we should be able to recover all the latest data that the nodes at location A had, by pulling it from oracle DB at location A and putting it into TT server in location B. I would like to know if it is possible to do this?

    Hello,
    Your approach is correct in designing a Disaster Recovery architecture for TimesTen and Oracle Database. TimesTen supports an Active-Standby pair topology that works well with integrating with the Oracle database within a particular site. However, like for any geographical based replication, it is recommended to configure replication across the WAN using the Oracle Database GoldenGate or Streams technologies in ASYNC mode for better throughput and efficiency. It is also recommended to compress replication traffic across the WAN between the Oracle databases.
    So while using the Oracle Database to replicate transactions across the WAN is the right thing to do (using Streams Replication or GoldenGate between the two Oracle databases (assuming using an Oracle RAC 2-node cluster in each site), you will not be able to guarantee that any transactions in site A has made it to site B. GoldenGate and Streams technologies have the ability to replicate the data bi-directionally. What this means is that when site A recovers, transactions that had been trapped there (either between TimesTen and the Oracle DB or in the Oracle DB transaction logs), will attempt to replicate again to site B, so it is important to set up a conflict detection/resolution approach which is possible to do in either GoldenGate or Streams.
    Note that Oracle Data Guard replication is not supported with TimesTen in such a configuration across the WAN where TimesTen datastores need to be maintained in both sites.
    To fully answer the question, however, we should get into the details of the type of cache group tables that you intend to use in TimesTen. If using TimesTen as just a read-only cache while all inserts/updates happen in the Oracle database, then OracleDB would be regarded as the database of record and it would be used to handles all changes while data changes get auto-refreshed from teh Oracle databases in each site into the respective TimesTen tables.
    If the application will be looking to take advantage of fast writes into TimesTen using AWT (async writethrough cache group tables), then it is recommended to configure those tables to be DYNAMIC AWT tables so that if a failover to site B takes place and the data is not in TimesTen (but it is in the Oracle Database), it will be automatically loaded on demand as needed from the Oracle database in site B. Note however that there are restrictions that exist with DYNAMIC load on demand cache groups that you need to look into to find out whether those would work in your application's case (particularly, load on demand works only if the where clause includes an equality predicate on the primary keys, foreigh keys, or unique indexes, etc...)
    To fully answer your question on non data loss across geographies, you'd have to use Synchronous replication between TimesTen and Oracle using Synchronous Writethrough Cache Groups and SYNC replication in Streams for example between the two geographies. Neither of those configurations are used to my knowledge in the field because they are very non optimal and carry huge response time expense, which slows down replication considerably and affects application response times.
    My assumption also is that the need for the Oracle database is because all data does not fit into memory. If the data does fit into memory, then you could also consider a pure timesten replication across the two sites using an active-standby pair on site A and a read-only subscriber on site B that would be made ACTIVE only in the case of a disaster on site A. Once site B takes over, you can also create an active-standby pair in site B based on the newly elected ACTIVE datastore in that site. In all these cases, it is recommended to use SYNC 2-SAFE replication between TimesTen datastores in the same site and ASYNC replication between the two sites.

  • High Availability (HA) questions

    Hi Gurus,
    First of all not sure is it a right forum to post this thread. But if anybody has any idea which forum, then I will do that again.
    But here is my scenarios and questions. Currently I am doing High Availability (HA) implementation using PowerHA & MSCS for SAp and NON-SAP environment. Using Oracle 10g possible datguard (still test is going how feasible is this), AIX 6.1 and Windows server. As per SAP docs, there are 3 categories for HA Testing.
    1. Before Failover u2013 Before Failing over to Passive Node
    2. After Failover u2013 After Failing over to Passive Node
    3. After Failback u2013 After Failing back to Active Node
    So I prepared my test plan for the said above categories for the system ECC, SRM, SCM, SCM-TM, SCM-Optimizer, BI, PI, EP MDM etc..
    Question(s) is/are:
    1. When switchover(or failover) from active to passive or when switchback(or failback) from passive to active, what transaction or command need to run to check all process or system is good. I mean initially I thought SGEN to run to do that. I am not sure SGEN is right one or not. If anybody has any idea, would be great. Although I am checking system consistency, ERS for any lock, message server, gateway etc. Other than that I mean while switchover suppose something running internally and do not know and did switchover. Then how will make sure nothing happened.
    2. What are the high availability functionality test for F5 Big-IP loadbalancer?
    3. What are the high availability functionality test plan for End-to-End process?
    4. What are the high availability test plan for VMWARE and WINDOWS environment?
    There are lots of questions, but the said above are initial to start and move on with HA.
    Thanks for effort and time.
    Krish

    Hi Krish,
    Answers to some questions.
    1. You can use JCMON tool to check the Java services are up after the switchover to any node ( either passive or active)
          For ABAP, you can just try to login and check if the services are up.
    2.  F5 - BIG IP Load Balancer is used to balance the load if you have multiple application servers i.e additional dialog instances. You need to refer F5 load balancer documents for the same.
    3.  End to End process test plan might change from client to client. So this should be drafted in conjunction with client and developers.
    4. VMWARE and WINDOWS - Not worked extensively on these platforms. So no comments on this
    Let me know if you need more information.
    Cheers...,
    Raghu

  • User experience question for High Availability

    Hi Experters,
    Not sure whether this is right forum to post this message.
    If anybody has already captured user experience for the failover environment, will appreciate their help on this. I will give overview of environment. Using PowerHA for ASCS/SCS failover, ORACLE Data Guard for Database failover.
    Trying to capture user experience for the failover environment
    if ASCS/SCS fail,
    App Server Fail (using F5 Load Balancer which should reroute, unless all app server fail),
    Database fail
    with following cases:
    1. User logged in, NO ACTIVITY. I believe NO IMPACT either ASCS/SCS fail or DB fail or App Server fail.
    2. User logged in and run a transaction before failover.
         What will happen in case of ASCS / SCS
         What will happen in case of DB failover
       and  what will happen in case of App server fail (NO High Availability, only redundant app servers.      MAX 4 app servers for each component)
    3. User logged in and run a transaction during failover.
       What will happen in case of ASCS / SCS
         What will happen in case of DB failover
       and  what will happen in case of App server fail (NO High Availability, only redundant app servers.      MAX 4 app servers for each component)
    Not sure which one possible or not. Some case thinking system will hang need to refresh. Some case hour glass and then come back once failover complete. and some case session closed completely.
    Thanks for your time and god bless the knowledge you have.
    Saroj

    just try to answer as much as I can (guess)
    > 1. User logged in, NO ACTIVITY. I believe NO IMPACT either ASCS/SCS fail or DB fail or App Server fail.
    DB failed and SCS failed won't have any impact to enduser, but if app server failed, the user session will lost, user will see a pop up error message at their SAPGUI.
    > 2. User logged in and run a transaction before failover.
    >      What will happen in case of ASCS / SCS
    user wont'be effected during failover if user is doing nothing but idle (replica enqueue is working before failover)
    >      What will happen in case of DB failover
    App server won't be able to do much thing but it's work processes into reonnecting status. it should resume (reconnect to DB) when failover is compoleted. So user should be able to continue the sessions.
    >    and  what will happen in case of App server fail (NO High Availability, only redundant app servers.      MAX 4 app servers for each component)
    user sessions in failed app server will be lost. However user should be able to logon agan if
    1)  logon via group, and
    2) within the group, there is at least one appl server alive.
    > 3. User logged in and run a transaction during failover.
    hanging or
    >    What will happen in case of ASCS / SCS
    if the transaction is using enqueue service, for example, then user will get error message.  otherwise, user won't be effected, e.g. if use is just search a list of order. user wont to able to logon via group during the failover.
    You also should be pepared for user connected through message server, .e.g. http request dispatching through message server directly , or via web dispatchr. user wont be able to connect during the failover.
    >      What will happen in case of DB failover
    user will get error message, transaction will be aborted.
    >    and  what will happen in case of App server fail (NO High Availability, only redundant app servers.      MAX 4 app servers for each component)
    very similar case  as case 2.

  • Cisco Prime Infrastructure 2.1 High availability Question

    Hello All,
    I am configuring high availability for two prime infrastructure 2.1 servers. I have configured manual HA between the servers. I need to know what will be the configuration in the devices ( switches,routers etc.) for proper working of the HA. For example
    Should we need to configure both the prime infra servers as snmp hosts in the devices??. If we have to when an event happens the switch well unnecessarily send the traps to the secondary even when the primary is alive??. 
    If anyone has a copy of the configuration of such a set up please share it with me. 
    Thanks and Regards
    Shabeeb

    Hi Shabeeb,
    You are correct on that part that unnecessarily devices will  try to send traps to the secondary server if you specify that in the device's config. I don't think it should be a concern , this is expected.
    otherwise you need to configure them later once the PI server fail over to secondary .
    If you have any other doubt ,kindly ask.
    Thanks-
    Afroz
    ***Ratings Encourages Contributors ***

  • Advice Requested - High Availability WITHOUT Failover Clustering

    We're creating an entirely new Hyper-V virtualized environment on Server 2012 R2.  My question is:  Can we accomplish high availability WITHOUT using failover clustering?
    So, I don't really have anything AGAINST failover clustering, and we will happily use it if it's the right solution for us, but to be honest, we really don't want ANYTHING to happen automatically when it comes to failover.  Here's what I mean:
    In this new environment, we have architected 2 identical, very capable Hyper-V physical hosts, each of which will run several VMs comprising the equivalent of a scaled-back version of our entire environment.  In other words, there is at least a domain
    controller, multiple web servers, and a (mirrored/HA/AlwaysOn) SQL Server 2012 VM running on each host, along with a few other miscellaneous one-off worker-bee VMs doing things like system monitoring.  The SQL Server VM on each host has about 75% of the
    physical memory resources dedicated to it (for performance reasons).  We need pretty much the full horsepower of both machines up and going at all times under normal conditions.
    So now, to high availability.  The standard approach is to use failover clustering, but I am concerned that if these hosts are clustered, we'll have the equivalent of just 50% hardware capacity going at all times, with full failover in place of course
    (we are using an iSCSI SAN for storage).
    BUT, if these hosts are NOT clustered, and one of them is suddenly switched off, experiences some kind of catastrophic failure, or simply needs to be rebooted while applying WSUS patches, the SQL Server HA will fail over (so all databases will remain up
    and going on the surviving VM), and the environment would continue functioning at somewhat reduced capacity until the failed host is restarted.  With this approach, it seems to me that we would be running at 100% for the most part, and running at 50%
    or so only in the event of a major failure, rather than running at 50% ALL the time.
    Of course, in the event of a catastrophic failure, I'm also thinking that the one-off worker-bee VMs could be replicated to the alternate host so they could be started on the surviving host if needed during a long-term outage.
    So basically, I am very interested in the thoughts of others with experience regarding taking this approach to Hyper-V architecture, as it seems as if failover clustering is almost a given when it comes to best practices and high availability.  I guess
    I'm looking for validation on my thinking.
    So what do you think?  What am I missing or forgetting?  What will we LOSE if we go with a NON-clustered high-availability environment as I've described it?
    Thanks in advance for your thoughts!

    Udo -
    Yes your responses are very helpful.
    Can we use the built-in Server 2012 iSCSI Target Server role to convert the local RAID disks into an iSCSI LUN that the VMs could access?  Or can that not run on the same physical box as the Hyper-V host?  I guess if the physical box goes down
    the LUN would go down anyway, huh?  Or can I cluster that role (iSCSI target) as well?  If not, do you have any other specific product suggestions I can research, or do I just end up wasting this 12TB of local disk storage?
    - Morgan
    That's a bad idea. First of all Microsoft iSCSI target is slow (it's non-cached @ server side). So if you really decided to use dedicated hardware for storage (maybe you do have a reason I don't know...) and if you're fine with your storage being a single
    point of failure (OK, maybe your RTOs and RPOs are fair enough) then at least use SMB share. SMB at least does cache I/O on both client and server sides and also you can use Storage Spaces as a back end of it (non-clustered) so read "write back flash cache
    for cheap". See:
    What's new in iSCSI target with Windows Server 2012 R2
    http://technet.microsoft.com/en-us/library/dn305893.aspx
    Improved optimization to allow disk-level caching
    Updated
    iSCSI Target Server now sets the disk cache bypass flag on a hosting disk I/O, through Force Unit Access (FUA), only when the issuing initiator explicitly requests it. This change can potentially improve performance.
    Previously, iSCSI Target Server would always set the disk cache bypass flag on all I/O’s. System cache bypass functionality remains unchanged in iSCSI Target Server; for instance, the file system cache on the target server is always bypassed.
    Yes you can cluster iSCSI target from Microsoft but a) it would be SLOW as there would be only active-passive I/O model (no real use from MPIO between multiple hosts) and b) that would require a shared storage for Windows Cluster. What for? Scenario was
    usable with a) there was no virtual FC so guest VM cluster could not use FC LUs and b) there was no shared VHDX so SAS could not be used for guest VM cluster as well. Now both are present so scenario is useless: just export your existing shared storage without
    any Microsoft iSCSI target and you'll be happy. For references see:
    MSFT iSCSI Target in HA mode
    http://technet.microsoft.com/en-us/library/gg232621(v=ws.10).aspx
    Cluster MSFT iSCSI Target with SAS back end
    http://techontip.wordpress.com/2011/05/03/microsoft-iscsi-target-cluster-building-walkthrough/
    Guest
    VM Cluster Storage Options
    http://technet.microsoft.com/en-us/library/dn440540.aspx
    Storage options
    The following tables lists the storage types that you can use to provide shared storage for a guest cluster.
    Storage Type
    Description
    Shared virtual hard disk
    New in Windows Server 2012 R2, you can configure multiple virtual machines to connect to and use a single virtual hard disk (.vhdx) file. Each virtual machine can access the virtual hard disk just like servers
    would connect to the same LUN in a storage area network (SAN). For more information, see Deploy a Guest Cluster Using a Shared Virtual Hard Disk.
    Virtual Fibre Channel
    Introduced in Windows Server 2012, virtual Fibre Channel enables you to connect virtual machines to LUNs on a Fibre Channel SAN. For more information, see Hyper-V
    Virtual Fibre Channel Overview.
    iSCSI
    The iSCSI initiator inside a virtual machine enables you to connect over the network to an iSCSI target. For more information, see iSCSI
    Target Block Storage Overviewand the blog post Introduction of iSCSI Target in Windows
    Server 2012.
    Storage requirements depend on the clustered roles that run on the cluster. Most clustered roles use clustered storage, where the storage is available on any cluster node that runs a clustered
    role. Examples of clustered storage include Physical Disk resources and Cluster Shared Volumes (CSV). Some roles do not require storage that is managed by the cluster. For example, you can configure Microsoft SQL Server to use availability groups that replicate
    the data between nodes. Other clustered roles may use Server Message Block (SMB) shares or Network File System (NFS) shares as data stores that any cluster node can access.
    Sure you can use third-party software to replicate 12TB of your storage between just a pair of nodes to create a fully fault-tolerant cluster. See (there's also a free offering):
    StarWind VSAN [Virtual SAN] for Hyper-V
    http://www.starwindsoftware.com/native-san-for-hyper-v-free-edition
    Product is similar to what VMware had just released for ESXi except it's selling for ~2 years so is mature :)
    There are other guys doing this say DataCore (more playing for Windows-based FC) and SteelEye (more about geo-cluster & replication). But you may want to give them a try.
    Hope this helped a bit :) 
    StarWind VSAN [Virtual SAN] clusters Hyper-V without SAS, Fibre Channel, SMB 3.0 or iSCSI, uses Ethernet to mirror internally mounted SATA disks between hosts.

  • Oracle Berkeley DB Java Edition High Availability (White Paper)

    Hi all,
    I've just read Oracle Berkeley DB Java Edition High Availability White Paper
    http://www.oracle.com/technetwork/database/berkeleydb/berkeleydb-je-ha-whitepaper-132079.pdf
    In section "Time Consistency Policy" (Page 18) it is written:
    "Setting a lag period that is too small, given the load and available hardware resources, could result in
    frequent timeout exceptions and reduce a replica's availability for read operations. It could also increase
    the latency associated with read requests, as the replica makes the read transaction wait so that it can
    catch up in the replication stream."
    Can you tell me why those read operations will not be taken by the master ?
    Why will we have frequent timeout ?
    Why should read transaction wait instead of being redirect to the master ?
    Why should it reduce replica's availability for read operations ?
    Thanks

    Please post this question on the Berkeley DB Java Edition (BDB JE) forum Berkeley DB Java Edition. This is the Berkeley DB Core (BDB) forum.
    Thanks,
    Andrei

  • HTTP Server High Availability

    Hello All.
    I have a question regarding OC4J and HTTP server High Availability.
    I want to do something like the Figure 3-1 of the Oracle Application Server High Availability Guide 10.1.2. See this link
    http://download-east.oracle.com/docs/cd/B14099_11/core.1012/b14003/midtierdesc.htm#CIHCEDFC
    What I have now is the following:
    Three hosts
    Two of them are an OAS 10.1.2 which we already configured the Cluster and deployed our applications (used this tutorial: http://www.oracle.com/technology/obe/obe_as_1012/j2ee/deploy/j2eecluster/farmcluster.htm)
    Let's say this nodes are:
    - host1
    - host2
    The other one is the Oracle WebCache stand alone (will act as Load Balancer). We will call this
    - hostwc3
    We already configured the WebCache as Load Balancer and is working just fine. We also configured the session replication successful and work great with our applications.
    What we have not clear is the following:
    When a client try to visit http://hostwc3/application/ the LOAD BALANCER routes him to, let's say http://host1/application/ and in the browser's URL will not show the Virtual Server anymore (the webcache server) and will show the actual real Apache address (host1 )that is attending him. IF we "kill" on ENTIRE host1 (apache, oc4j, etc..) the clients WILL perceive the down and if they try to press F5, the will try to access to an Apache that doesn't is up and running.... The behavior expected is that the browser NEVER shows the actual Apache URL, so, when some apache goes down, the client do not disconnect (as it happens with and OC4J downfall ) and always works with the "virtual web server".
    I came up with some ideas but I want you Guys to give me an advice:
    - In Web Cache, do not route for load balancing to Apache, and route the Oc4J directly (Is this possible?)
    - Configure a HTTP Server Cluster, this means that we have to have a "Virtual Name"to the Apaches (two of them). Is this possible? how?
    - Use the rewrite mode of the Apache. Is this a good idea?
    - Any other idea how to fix the Apache "Single Point Of Failure" ?
    According with the figure 3-1 ( Link above ) we do can have HTTP Server in a cluster. But I have no idea how to manage it or configure it.
    Thanks in advance any help!

    You cannot point Outlook Anywhere to your DAG cluster IP address. It must be pointed to the actual IP address of either server.
    For no extra cost DNS round robin is the best you will get, but it does have some drawbacks as it may give the IP address of a server you have taken down for maintenance or the server has an issue.
    You could look to implement a load balancer but again if you are doing this for high availability then you want more than one load balancer in the cluster - otherwise you've just moved your single point of failure.
    Having your existing NAT and just remembering to update it to point to the other server during maintenance may suit your needs for now.
    If you can go into more detail about what the high availability your business is looking to achieve and the budget we can suggest the best method to meet those needs for the price point.
    Have a great day
    Oliver
    Oliver Moazzezi | Exchange MVP, MCSA:M, MCITP:Exchange 2010,Exchange 2013, BA (Hons) Anim | http://www.exchange2010.com | http://www.cobweb.com | http://twitter.com/OliverMoazzezi

  • SCDPM 2012 R2 High Avalibility Design?

    Hello Everyone,
    Can we Install SCDPM 2012 R2 in High Availblity Mode?
    If yes What's the Design for SCDPM high availibility?
    Thnks.......

    Hello Everyone,
    Can we Install SCDPM 2012 R2 in High Availblity Mode?
    If yes What's the Design for SCDPM high availibility?
    Thnks.......
    Hi,
    There is not a High Availability solution for DPM such as clustering DPM servers. The options you do have are 
    Primary and Secondary DPM servers, DPM chaining, or
    Cyclic protection. Here is a breakdown of each:
    Primary to secondary protection: The database and replicas stored on a primary DPM server can be backed up to a secondary DPM server. If the primary server fails the secondary server continues to back up protected workloads.
    For more on Primary to secondary protection:
    http://technet.microsoft.com/en-us/library/jj244598.aspx
    DPM chaining: A chain of DPM servers provide protection, and each server protects the next one in in the chain. For example:
        DPM1 is protected by DPM2 (DPM1 is the primary and DPM2 is the secondary).
        DPM2 is protected by DPM3 (DPM2 is the primary and DPM3 is the secondary)
    To setup chaining:
    http://technet.microsoft.com/en-us/library/hh758194.aspx
    Cyclic protection: One DPM server is backed up by another DPM server, and vice versa. So that DPM1 protects DPM2, and likewise DPM2 protects DPM1.
    More on Cyclic protection:
    http://technet.microsoft.com/en-us/library/jj244597.aspx#BKMK_Cyclic
    I hope this adds some clarification around this.
    My Blog | www.buchatech.com | www.systemcenterportal.com
    If you found this post helpful, please give it a "Helpful" vote. If it answered your question, remember to mark it as an "Answer". This posting is provided "AS IS" with no warranties and confers no rights! Always test ANY suggestion
    in a test environment before implementing!

  • HANA High Availability System Vs Storage Vs VIP failover

    Dear Experts,
    Hope your all doing great. I would like to seek your expertise on HANA high availability best practice. We have been deciding to use TDI for BW on HANA. The next big question for us is how make it available atleast 99.99%.
    I was going through multiple documents, SDN forums, etc..but would like to see how the experts are performing in real time.
    My view -
    Virtual IP failover is a common HA practice which have been used to failover CI / DB hosts depends on failure/maintenance. In this case, both nodes can be used to run app servers.
    System replication - HANA based required secondary standby node, which doesn't accept user requests, but replicate the database from primary using logs after initial data snapshot either synchronous or Asynchronous. (Can be used as HA or DR - if servers are between different data centers).
    Storage replication - HANA based required secondary standby node, which replicates SAN for HA/DR.
    Could you please provide your expertise method you followed for HANA HA and what are the pros, cons and challenges that you have faced or facing.
    Thanks
    Yoga

    Thanks forbrich
    Do you know any specific doc that describes the installation and configuration steps of 10g RAC on NAS? If possible, can you provide some link that I could use to perform this task?
    I have done RAC installations on SAN without any problems and its something I'm fairly experienced with. With NAS I am not really comfortable since I can't seem to find any documentation that describes step by step installation procedure or guidelines for that matter.
    Thank you for your input
    Best Regards
    Abbas

Maybe you are looking for

  • Is it possible to sort a playlist alphabetically...

    Is it possible to sort a PLAYLIST or ARTIST songlist alphabetically in itunes and retain that order on ipod (Classic)? Some of my playlists or artist's songlists have literally hundreds of songs and it is cumbersome to try to locate a particular song

  • Sharing iPhoto Library

    I've got my iPhoto 8 library on my iMac and can share it to my PowerBook G3 (running iPhoto 6). I tried to run a slide show from my PowerBook to my 42 inch Plasma screen TV. Got the message that slide show required millions of colours; but when I sel

  • Registry name and value to check if server has internet connection

    Hi I want to check that if server has internet connection though registry value. Can any one help me the path of registry to check if server has an internet connection or not

  • RBS Third Party Provider with Sharepoint

    Our SharePoint content database size is increasing day by day...we are assuming it will reach upto 1 TB. currenty the size is 136 GB. We are launching the application in different location. My company plans to use RBS for this. In technet website, fo

  • Trying to account for dozens of unidentified TB on SSD

    I am working  with a late 2009 27" 2.66Ghz iMac running OS X 10.9.3 It is fitted with 16GB RAM. the Operating system and applications are on a 240Gb OWC 3G SSD. Issue: When I go to SSD > users > user name and use the 'get info" tool it tells me that