Exalogic VDC Live Migration, DRS, and HA functionality?

Greetings,
Could you please help us understand the Exalogic VDC Live Migration, DRS, and HA functionality compared to Oracle VM? For example, dows Exalogic now support Live Migration? If no, why? Does Exalogic now support DRS? If no why? Does Exalogic support Oracle VM / OCFS2 HA? If no why?
Thank you,
Roddy

Thanks a lot for so fast info and the tips.
I don't have VIP configured, but in my test only "virtual machine server", which doesn't play a role of pool master, is failed. I have 2 computers in the cluster: one is pool master, utility server, vm server and the second one is only vm server. That second one is failed in my test, so pool master is still running and VIP should not be important I think.
Is it true?
Anyway I'll configure that VIP and retest it.
There is no hardware resources issues as well. Wizards didn't allow me to create vm (HA enabled) with more CPU/RAM assigned than one machine can serve.
Also I can not find in logs any error, only that one, when I was trying to power on/off the vm in improper state (when server was down, statuses were unreachable/unknown in the console, log was saying still RUNNING):
"2010-07-19 14:55:03" ERROR=> start_vm: failed! vm('/OVS/running_pool/20_vwin02') <Exception: vm('/OVS/running_pool/20_vwin02') status ('RUNNING') not in ('DOWN', 'ERROR')>
So, I'm still finding questions:
1.) Is really needed VIP when only VM server (not pool master) is failed?
2.) Only for pool master is it supported (automatic vm restart when server failed)?
Kind regards,
Greg

Similar Messages

  • Hyper-V Failover Cluster Live migration over Virtual Adapter

    Hello,
    Currently we have a test environment 3 HyperHost wich have 2 Nic teams.  LAN team and team Live/Mngt.
    In Hyper-V we created a Virtual Switch (Wich is Nic team Live/Mngt)
    We want to seperate the Mngt and live with Vlans. To do this we created 2 virtual adapters. Mngt and Live, assigned ip adresses and 2 vlans(Mngt 10, Live 20)
    Now here is our problem: In Failover cluster you cannot select the Virtual Adapter(live). Only the Virtual Switch wich both are on. Meaning live migration simple uses the vSwitch instead of the Virtual adapter.
    Either its not possible to seperate Live migration with a Vlan this way or maybe there are Powershell commands to add Live migration to a Virtual adapter?
    Greetings Selmer

    It can be done in PowerShell but it's not intuitive.
    In Failover Cluster Manager, right-click Networks and you'll find this:
    Click Live Migration Settings and you'll find this:
    Checked networks are allowed to carry traffic. Networks higher in the list are preferred.
    Eric Siron
    Altaro Hyper-V Blog
    I am an independent blog contributor, not an Altaro employee. I am solely responsible for the content of my posts.

  • Shared nothing live migration wrong network

    Hey,
    I have a problem when doing shared nothing live migration between clusters using SCVMM 2012 R2.
    The transfer goes ok. But It choose to use my team of 2 x 1GB nic (routable) instead of the team with 2x10GB (non routable, but opened between all hosts).
    I have set vmmigrationsubnet to correct net. Check live migration setting, and that port 6600 is listening on the correct IP.
    But still it chooses to use the 1GB net.
    Anyone got any idea what I can do next?
    //Johan Runesson

    Do you have only your live migration network defined as such on the clusters or do you have both defined?  What are the IP networks on the live migration networks on each cluster?
    .:|:.:|:. tim

  • Hyper-v 2012 R2 Live migration issue in 2003 Domain function Level

    hi Team ,
    i recently build 2012 R2 Hyper-v Cluster with three node. Everrything working fine with out any issue . Cluster working also fine. Later i came across one issue when tried to Live migration virtual machine from one host to another . it failed all the time
    while quick migration is working . i gone through few articles and find it is known issue with hyper-v 2012 R2 where domain functional level is set to 2003 . although they have provided Hotfix but no solution.
    http://support.microsoft.com/kb/2838043
    Please let me know if any one face similar issue and able to resolve by any hotfix. My host are updated .
    Thanks
    Ravindra
    Ravi

    Hi Ravi1987,
    The KB2838043 is applied for Server 2012 node, Could you offer us the related cluster error event id, or you can refer the following article to check your cluster
    network binding order is correct or not.
    Configuring Windows Failover Cluster Networks
    http://blogs.technet.com/b/askcore/archive/2014/02/20/configuring-windows-failover-cluster-networks.aspx
    You can try to install recommended hotfixes and updates for Windows Server 2012 R2-based failover clusters first, then monitor this issue again.
    The KB download:
    Recommended hotfixes and updates for Windows Server 2012 R2-based failover clusters
    http://support.microsoft.com/kb/2920151
    More information:
    Windows Server 2008 R2 Live Migration – “The devil may be in the networking details.”
    http://blogs.technet.com/b/askcore/archive/2009/12/10/windows-server-2008-r2-live-migration-the-devil-may-be-in-the-networking-details.aspx
    I’m glad to be of help to you!

  • Cluster node reboot and Quick Migration of VMs instead of Live Migration...

    Hi to all,
    how can one configure a Windows Server 2012 multi-node failover cluster, that vms are migrated per Live Migration and NOT per Quick Migration, if one node of the failover cluster will be rebooted.
    Thanks in advance
    Joerg

    Hi Aidan,
    only for the record:
    We get the requested functionality - Live migrate all VMs on reboot without first pausing the cluster- when we do the following:
    Change the value of HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Control\PreshutdownOrder
    from the default
    vmms
    wuauserv
    gpsvc
    trustedinstall
    to
    clussvc
    vmms
    wuauserv
    gpsvc
    trustedinstall
    Now the cluster service stops at first, if we Trigger a reboot and all VMs migrate as configured per MoveTypeThreshold cluster setting.
    Greetings
    Joerg

  • Hyper V Lab and Live Migration

    Hi Guys,
    I have 2 Hyper V hosts I am setting up in a lab environment. Initially, I successfully setup a 2 node cluster with CSVs which allowed me to do Live Migrations. 
    The problem I have is my shared storage is a bit of a cheat as I have one disk assigned in each host and each host has starwinds virtual SAN installed. The hostA has an iscsi connection to hostB storage and visa versa.
    The issue this causes is when the hosts shutdown (because this is a lab its only on when required), the cluster is in a mess when it starts up eg VMs missing etc. I can recover from it but it takes time. I tinkered with the HA settings and the VM settings
    so they restarted/ didnt restart etc but with no success.
    My question is can I use something like SMB3 shared storage on one of the hosts to perform live migrations but without a full on cluster? I know I can do Shared Nothing Live Migrations but this takes time.
    Any ideas on a better solution (rather than actually buying proper shared storage ;-) ) Or if shared storage is the only option to do this cleanly, what would people recommend bearing in mind I have SSDs in the hyper V hosts.
    Hope all that makes sense

    Hi Sir,
    >>I have 2 Hyper V hosts I am setting up in a lab environment. Initially, I successfully setup a 2 node cluster with CSVs which allowed me to do Live Migrations. 
    As you mentioned , you have 2 hyper-v host and use starwind to provide ISCSI target (this is same as my first lab environment ) , then I realized that I need one or more host to simulate more production scenario . 
    But ,if you have more physical computers you may try other's progects .
    Also please refer to this thread :
    https://social.technet.microsoft.com/Forums/windowsserver/en-US/e9f81a9e-0d50-4bca-a24d-608a4cce20e8/2012-r2-hyperv-cluster-smb-30-sofs-share-permissions-issues?forum=winserverhyperv
    Best Regards
    Elton Ji
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Live Migration and private network

    Is it a best practice to put up a Private Network beetween the nodes in a pool (reserving a few network cards and switch ports for it), to have a dedicated network for the traffic generated e.g. by live migration and/or ocfs2 heartbeat? I was wondering why such setup is generally recommended in other virtualization solutions, but apparently it's not considered strictly necessary in OVM... Why? Are there any docs regarding this? I couldn't find any.
    Thanks!

    Hi Roynor,
    regarding the physical separation beetween management+hypervisor and the guest VMs, it's now implemented and working...
    My next doubt on the list of doubts :-) at this point is:
    I could easily set up ONE MORE dedicated bond, create a Bridge with a private IP on it on each server (e.g. 10.xxx.xxx.xxx), and then create a Private VLAN completely insulated from the rest of the world.
    I'd be putting the physical switch ports where the Private Bonds/Bridges belong to on the same VLAN ID.
    But:
    - How can I be sure that this network WILL be actually used by the relevant traffic? If I'm not wrong, when you set up e.g. a physical RAC cluster, at a certain point you are prompted to choose what network to use for the Heartbeat (and it will be marked as PRIVATE), and what network will be used by clients traffic (PUBLIC).
    In Oracle VM such setting does not exist... Neither during installation, nor in VM Manager, nowhere.
    - Apart from Security, I'm doubting that during heavy VMs migration problems could arise, because if the network gets saturated, there are chances that the OCFS2 heartbeat would be somehow "lost", therefore messing up HA etc. This is at least the reason why in a RAC setup a private network is highly recommended.
    - I finally found that doc you mention from IBM (thanks for pointing it out!) but my opinion is that THEIR INTENTION was to separate the traffic at the same way I'd like to, but there is simply NO PROOF that such setup would work... They do not mention where you can specify what traffic you want to be on what network...
    This is a very important point... I'm wondering why this lack of information.
    Thanks for your feedback, btw
    Edited by: rlomba on Dec 17, 2009 6:16 AM

  • When setting up converged network in VMM cluster and live migration virtual nics not working

    Hello Everyone,
    I am having issues setting up converged network in VMM.  I have been working with MS engineers to no avail.  I am very surprised with the expertise of the MS engineers.  They had no idea what a converged network even was.  I had way more
    experience then these guys and they said there was no escalation track so I am posting here in hopes of getting some assistance.
    Everyone including our consultants says my setup is correct. 
    What I want to do:
    I have servers with 5 nics and want to use 3 of the nics for a team and then configure cluster, live migration and host management as virtual network adapters.  I have created all my logical networks, port profile with the uplink defined as team and
    networks selected.  Created logical switch and associated portprofle.  When I deploy logical switch and create virtual network adapters the logical switch works for VMs and my management nic works as well.  Problem is that the cluster and live
    migration virtual nics do not work.  The correct Vlans get pulled in for the corresponding networks and If I run get-vmnetworkadaptervlan it shows cluster and live migration in vlans 14 and 15 which is correct.  However nics do not work at all.
    I finally decided to do this via the host in powershell and everything works fine which means this is definitely an issue with VMM.  I then imported host into VMM again but now I cannot use any of the objects I created and VMM and have to use standard
    switch.
    I am really losing faith in VMM fast. 
    Hosts are 2012 R2 and VMM is 2012 R2 all fresh builds with latest drivers
    Thanks

    Have you checked our whitepaper http://gallery.technet.microsoft.com/Hybrid-Cloud-with-NVGRE-aa6e1e9a for how to configure this through VMM?
    Are you using static IP address assignment for those vNICs?
    Are you sure your are teaming the correct physical adapters where the VLANs are trunked through the connected ports?
    Note; if you create the teaming configuration outside of VMM, and then import the hosts to VMM, then VMM will not recognize the configuration. 
    The details should be all in this whitepaper.
    -kn
    Kristian (Virtualization and some coffee: http://kristiannese.blogspot.com )

  • 2012 R2 Cluster and Live Migration

    6 Node Cluster with Server 2012 R2, all VM's are Server 2012 R2
    4 Fiber SAN's
    Moving and Live Migration worked fine in FailOver Cluster Manager
    But this time we were trying do it with SCVMM 2012 R2 and just move one VM (Gen2)
    Of course it failed at 99%
    Error (12711)
    VMM cannot complete the WMI operation on the server (whatever) because of an error: [MSCluster_Resource.Name=&quot;SCVMM VMHost&quot;] The cluster resource could not be found.
    The cluster resource could not be found (0x138F)
    Recommended Action
    Resolve the issue and then try the operation again.
    How do I fix this? the VM is still running. The two vhdx files it was moving are smaller then orginal's , but it change the configuration file to point to new ones, which are bad.
    it says I can Repair it... Redo or Undo....of course neither of those options work.
    Wait for the object to be updated automatically by the next periodic Live migrate storage of virtual machine vmhost from whatever to whatever job.
    ID: 1708
    Cluster has no errors, SAN's have no errors, CSV have no errors. the machine running scvmm is VM running on the cluster

    How did you create this VM? if this is created outside of VMM, I recommend doing a manual refresh of the VM first to ensure that VMM can read its attributes. Then retry the operation.
    Btw, are the VMs using diff disk? any checkpoints associated with them?
    Kristian (Virtualization and some coffee: http://kristiannese.blogspot.com )

  • Live Migration failed using virtual HBA's and Guest Clustering

    Hi,
    We have a Guest Cluster Configuration on top of an Hyper-V Cluster. We are using Windows 2012 and Fiber Channel shared storage.
    The problem is regarding Live Migration. Some times when we move a virtual machine from node A to node B everything goes well but when we try to move back to node A Live Migration fails. What we can see is that when we move the VM from node A to B and Live
    Migration completes successfully the virtual ports remain active on node A, so when we try to move back from B to A Live Migration fails because the virtual ports are already there.
    This doesn't happen every time.
    We have checked the zoning between Host Cluster Hyper-V and the SAN, the mapping between physical HBA's and the vSAN's on the Hyper-V and everything is ok.
    Our doubt is, what is the best practice for zoning the vHBA on the VM's and our Fabric? We setup our zoning using an alias for the vHBA 1 and the two WWN (A and B) on the same object and an alias for the vHBA 2 and the correspondent WWN (A and B). Is it
    better to create an alias for vHBA 1 -> A (with WWN A) and other alias for vHBA 1 -> B (with WWN B)? 
    The guest cluster VM's have 98GB of RAM each. Could it be a time out issue when Live Migration happen's and the virtual ports remain active on the source node? When everything goes well, the VM moves from node A with vHBA WWN A to node B and stays there
    with vHBA WWN B. On the source node the virtual ports should be removed automatically when the Live Migration completes. And that is the issue... sometimes the virtual ports (WWN A) stay active on the source node and when we try to move back the VM Live Migration
    fails.
    I hope You may understand the issue.
    Regards,
    Carlos Monteiro.

    Hi ,
    Hope the following link may help.
    To support live migration of virtual machines across Hyper-V hosts while maintaining Fibre Channel connectivity, two WWNs are configured for each virtual Fibre Channel adapter: Set A and Set B. Hyper-V automatically alternates between the Set A and Set B
    WWN addresses during a live migration. This ensures that all LUNs are available on the destination host before the migration and that no downtime occurs during the migration.
    Hyper-V Virtual Fibre Channel Overview
    http://technet.microsoft.com/en-us/library/hh831413.aspx
    More information:
    Hyper-V Virtual Fibre Channel Troubleshooting Guide
    http://social.technet.microsoft.com/wiki/contents/articles/18698.hyper-v-virtual-fibre-channel-troubleshooting-guide.aspx
    Hyper-V Virtual Fibre Channel Design Guide
    http://blogs.technet.com/b/privatecloud/archive/2013/07/23/hyper-v-virtual-fibre-channel-design-guide.aspx
    Hyper-V virtual SAN
    http://salworx.blogspot.co.uk/
    Thanks.
    Please remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading the thread.

  • Live Migration and Remote Desktop (Terminal Services)

    From what I've read about live migration, it seems that network connections are maintained after the migration is complete.  If I am migrating a terminal server, will the user sessions be maintained or will they need to reconnect?
    Eric Logsdon
    Eric Logsdon Cooperative Technologies, Inc.

    We currently have multiple Citrix servers housed on Hyper-V clusters and have at time migrated a VM with anywhere between 10- 40 users on it without any noticeable interruption of service.  From my experience it works as expected, but Citrix has
    session reliability which helps through periodic network interruptions.  Testing certain apps will be the key.  Also, to be safe, moving a fullly loaded Terminal Server in the middle may not be the prime time when you want to do this on a regular
    basis.  In my 24/7 organization the addition of live migration has moved my maintenance window from 1am - 3am to 7pm - 10pm.
    Rob McShinsky (www.VirtuallyAware.com)
    VirtuallyAware - Experiences in a Virtual World (Microsoft MVP - Virtual Machine)

  • Oracle VM (XEN) and Live Migration throws PCI error

    First off, anyone using a HS21 XM blade with Oracle VM? Anyone attempted a live-migration, does it work?
    When attempting a live migration on a HS21 XM blade we get the following errors and the host hangs:
    102 I Blade_02 08/10/08, 08:46:38 (CADCOVMA02) Blade reboot
    103 E Blade_02 08/10/08, 08:46:26 (CADCOVMA02) Software critical interrupt.
    104 E Blade_02 08/10/08, 08:46:21 (CADCOVMA02) SMI Hdlr: 00150700 PERR: Slave signaled parity error B/D/F/Slot=07000300 Vend=8086 Dev=3 0C Status=8000
    105 E Blade_02 08/10/08, 08:46:19 (CADCOVMA02) SMI Hdlr: 00150900 SERR/PERR Detected on PCI bus
    106 E Blade_02 08/10/08, 08:46:19 (CADCOVMA02) SMI Hdlr: 00150700 PERR: Slave signaled parity error B/D/F/Slot=00020000 Vend=8086 Dev=25F7 Statu
    107 E Blade_02 08/10/08, 08:46:18 (CADCOVMA02) PCI PERR: parity error.
    108 E Blade_02 08/10/08, 08:46:17 (CADCOVMA02) SMI Hdlr: 00150900 SERR/PERR Detected on PCI bus
    109 E Blade_02 08/10/08, 08:46:17 (CADCOVMA02) SMI Hdlr: 00150400 SERR: Device Signaled SERR B/D/F/Slot=07000300 Vend=8086 Dev=350C Sta us=4110
    110 E Blade_02 08/10/08, 08:46:16 (CADCOVMA02) SMI Hdlr: 00150400 SERR: Device Signaled SERR B/D/F/Slot=00020000 Vend=8086 Dev=25F7 Status=4110
    111 E Blade_02 08/10/08, 08:46:14 (CADCOVMA02) PCI system error.
    any ideas? I've called IBM support but their options are reseat the hardware or replace it. this happens on more than one blade so we're assuming it has something to do with oracle VM. Thanks!

    Hi Eduardo,
    How are things going ?
    Best Regards
    Elton Ji
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Best Configuration for T4-4 and live migration

    I have several T4-4 servers with Solaris 11.1
    What kind of settings I need for production environments with OVM live migration?
    Should I use OVM manager for rugged environments?
    Should I set two io-domain on each server?
    Considering that I do live migration, for guess ldom system disk should I configure (publish) one big LUN to all servers or just a separate LUN for each ldom guest.

    For the first questions: it's best to have separate LUNs or other backend disks for each guest - otherwise how would you plan to share them across different guests and systems? Separate LUNs also is good for performance by encouraging parallel I/O. Choice of using an additional I/O domain is up to you, but it's very typical for production users because of the resiliency it adds.
    For advice on live migration, seehttps://blogs.oracle.com/jsavit/entry/best_practices_live_migration
    Other blog posts after that discuss availability and performance.
    Hope that helps, Jeff

  • LDOMs, Solaris zones and Live Migration

    Hi all,
    If you are planning to use Solaris zones inside a LDOM and using an external zpool as Solaris zone disk, wouldn't this break one of the requirements for being able to do a Live Migration ? If so, do you have any ideas on how to use Solaris zones inside an LDOM and at the same time be able to do a Live Migration or is it impossible ? I know this may sound as a bad idea but I would very much like to know if it is doable.

    Thanks,
    By external pool I am thinking of the way you probably are doing it, separate LUNs mirrored in a zpool for the zones coming from two separate IO/Service domains. So even if this zpool exist inside the LDOM as zone storage this will not prevent LM ? That's good news. The requirement "no zpool if Live Migration" must then only be valid for the LDOM storage itself and not for storage attached to the running LDOM. I am also worried about a possible performance penalty introducing an extra layer of virtualisation. Have you done any tests regarding this ?

  • LIve migration trigger on Resource crunch in Hyper-v 2012 or R2

    Like VMware DRS , who can trigger Migration of virtual machine from 1 datastore to another or 1 host to another. do we have same kind of machanism in Hyper-v 2012 also.
    i have five HYper-v 2012 host in single cluster. now i want if any host face resource crucnh of memory than it migrate the virtual machine on another host.
    Thanks ravinder
    Ravi

    SCVMM has a feature called Dynamic Optimization.
    Dynamic Optimization can be configured on a host group, to migrate virtual machines within host clusters with a specified frequency and aggressiveness. Aggressiveness determines the amount of load imbalance that is required to initiate a migration during
    Dynamic Optimization. By default, virtual machines are migrated every 10 minutes with medium aggressiveness. When configuring frequency and aggressiveness for Dynamic Optimization, an administrator should factor in the resource cost of additional migrations
    against the advantages of balancing load among hosts in a host cluster. By default, a host group inherits Dynamic Optimization settings from its parent host group.
    Dynamic Optimization can be set up for clusters with two or more nodes. If a host group contains stand-alone hosts or host clusters that do not support live migration, Dynamic Optimization is not performed on those hosts. Any hosts that are in maintenance
    mode also are excluded from Dynamic Optimization. In addition, VMM only migrates highly available virtual machines that use shared storage. If a host cluster contains virtual machines that are not highly available, those virtual machines are not migrated during
    Dynamic Optimization.
    http://technet.microsoft.com/en-us/library/gg675109.aspx
    Cheers !
    Optimism is the faith that leads to achievement. Nothing can be done without hope and confidence.
    InsideVirtualization.com

Maybe you are looking for

  • Found iCal events in Time Machine.  Now Lion only allowing me to import one event at a time.

    Upgrade my macbook pro over the weekend to Snow Leopard and then Lion.   I lost all my the majority of my iCal events.   Went to Apple store and they said that was rare and not sure how to restore my last back up from Time Machine on an external driv

  • Need help backing up contacts from deleted exchange email address.

    I had a microsoft exchange email that has been deleted. All contacts have been lost however they are still on my ipad as I have not changed any settings. Bc they are apart of exchange server they wont save to cloud. Any ideas how I can save them and

  • General place for newbie datamodeler questions?

    With the production release of SQL Developer Data Modeler I'd assume that this forum will draw a bunch of new users. Is there a better place to ask some pretty simple, newbie like questions that won't bother the more seasoned members? For instance, I

  • IPhoto can't click on on options (after OPT-CMD) startup

    I am using iPhoto 9.3.2 (670.63) under Mountain Lion (10.8.1). The thumbnails are showing up wrong pictures so I do a OPT-CMD startup of iPhoto. After iPhoto pop up with the dialog box to select the "rebuild thumbnail" option, I cannot click the blue

  • Reviewers unable to open AdobeCaptivateReviewer.air

    I sent AdobeCaptivateReviewer.air to my reviewers, along with the navigation path to the shared drive where the .crev file is posted. My reviewers all have the same result when they try to open the AIR application. They are asked what program to use