Live migration to HA failed leaving VHD on local storage and VM in cluster = Unsupported Cluster Configuration

Hi all
Fun one here, I've been moving non-HA VMs to a HA and everything has been working perfectly until now.  All this is being performed on Hyper-V 2012R2, Windows Server 2012R2 and VMM 2012R2.
For some reason on the VMs failed the migration with an error 10608 "Cannot create or update a highly available virtual machine because Virtual Machine Manager could not locate or access Drive:\Folder"  The odd thing is the drive\folder is
a local storage one and I selected a CSV in the migration wizard.
The net result is that the VM is half configured into the cluster but the VHD is still on local storage.  Hence the "unsupported cluster configuration" error.
The question is how do I roll back? I either need to get the VM out of the cluster and back into a non-HA state or move the VHD onto the CSV.  Not sure if the latter is really a option.
I've foolishly clicked "Ignore" on the repair so now I can't use the "undo" option (brain fade moment on my part).
Any help gratefully received as I'm a bit stuck with this.
Thanks
Rob

Hi Simar
Thanks for the advice, I've now got the VM back in a stable state and running HA.
Just to finish off the thread for future I did the following
- Shutdown the VM
- Remove the VM from the Failover Cluster Manager (as you say this did leave the VM configuration intact)
- I was unable to import the VM as per your instructions so I copied the VHD to another folder on the local storage and took a note of the VM configuration.
- Deleted the VM from VMM so this removed all the configuration details/old VHD.
- Built a new VM using the details I saved from the point above
- Copied the VHD into the new VMs folder and attached it to the VM.
- Started it up and reconfigured networking
- Use VMM to make the VM HA.
I believe I have found the reason for the initial error, it appears there was a empty folder in the Snapshot folder, probably from an old Checkpoint that hadn't cleaned up properly when it was deleted.
The system is up and running now so thanks again for the advice.
Rob

Similar Messages

  • Adobe Flash FAIL:  Adobe Flash Player local storage settings incorrect.  Module 'Resume' feature may not work on this computer.

    Using a Windows 2012 RDS Environment, we have users connecting to a CPD website, and as part of the CPD they need to run a systems checker.  When they run the systems checker they get the following error message: "Adobe Flash FAIL:  Adobe Flash Player local storage settings incorrect.  Module 'Resume' feature may not work on this computer". All users are connecting to this environment with Windows CE Clients,I have checked the setting on Adobe Flash and they seem correct but as each user has its own profile on the RDS session, is there something that I should be setting for each user. I have added the website to the trusted sites and it has made no difference.   Any ideas

    It sounds like what's happening is that Flash Player can't write or read from the local shared objects in the user's redirected home directory because we disallow traversing junctions in the broker process.  This behavior was disabled to address a vulnerability identified in some of John Forshaw's research into the IE broker last year.
    You can enable this behavior by adding the following setting to mms.cfg:
    EnableInsecureJunctionBehavior=1
    That said, you can probably gather from the name of the flag that we don't really recommend this approach, and disable this attack surface by default.  There's some risk that a network attacker could craft content that abuses fundamental issues with how Windows handles Junctions to write to arbitrary locations.
    Unfortunately, there's not a simple or easy workaround that I'm aware of (but it's been ages since I've administered a Windows domain) for this kind of NAS/SAN-backed terminal server environment where Flash is not able to access \Users\<user>\AppData\Roaming\Macromedia\Flash Player\ without traversing a junction.

  • Failed live migration from HyperV2012 to Hyperv2012R2 cluster

    I’m at a loss on something and I was wondering if you could point me in the right direction.    I tried to migrate a server from one of my 2012 hosts to a brand new verified Hyperv 2012 R2 cluster.  I had already succesfully migrated 5 others
    from there but this one Sql 2012 guest gave me a problem.
    When I tried to live migrate it, it failed.  It said the hardware on the destination computer was not compatible.  I did some research and one blog said that sometimes you migrate to different hardware and it fails, like processors.  The guy
    on the blog said to check the box under processor\compatibility to put the virtual processor in compatibility mode.  I did that and tried again but it failed again.  This blog said something about resource groups.  The virt was “locked” at that
    point.
    I got frustrated so I just turned off the guest and copied the disks and rebuilt it on Aag with the same memory processor setting.  I thought I was fine, but I looked at it the next day and in the failover cluster manager it shows the machine is “off”. 
    BUT the VM is actually running because the server is working. It’s a database server and the sites using it are up, and I can RDC to it.  So I’m afraid to touch it.
    And these are the errors from the event log
    'Virtual Machine Configuration DEVDB1' failed to register the virtual machine with the virtual machine management service.
    The Virtual Machine Management Service failed to register the configuration for the virtual machine '62AAD2B5-8E03-4B59-84E0-D52CBF36934B' at 'C:\ClusterStorage\Volume2\Devdb1\DEVDB1': The
    system cannot find the file specified. (0x80070002). If the virtual machine is managed by a failover cluster, ensure that the file is located at a path that is accessible to other nodes of the cluster.
    Cluster resource 'Virtual Machine Configuration DEVDB1' of type 'Virtual Machine Configuration' in clustered role 'DEVDB1' failed. The error code was '0x2' ('The system cannot find the file
    specified.').
    Based on the failure policies for the resource and role, the cluster service may try to bring the resource online on this node or move the group to another node of the cluster and then restart
    it.  Check the resource and group state using Failover Cluster Manager or the Get-ClusterResource Windows PowerShell cmdlet.
    I think I moved it while it was in a funky state.  I also think I was supposed to “export” the vm first.  I guess I’m a newbie but I’m not sure what to do to fix it.  Any advice is greatly appreciated.

    I think i found at least part of the problem but i'm not sure.  My new VM's configuration files and directory are called this.
    D79490DB-48F9-40A4-9540-53F2532D3F7F
    D79490DB-48F9-40A4-9540-53F2532D3F7F.xml
    Not 62AAD2B5-8E03-4B59-84E0-D52CBF36934B
    Still not sure what to do about that though.

  • Hyper-V live migration failed

    There is Hyper-V cluster with 2 nodes. Windows Server 2012 R2 is used as operating system.
    Trying to live migrate test VM from node 1 to node 2 and get error 21502:
    Live migration of 'Virtual Machine test' failed.
    'Virtual Machine test' failed to fixup network settings. Verify VM settings and update them as necessary.
    VM has Network Adapter connected to Virtual switch. This vSwitch has Private network as connection type.
    If I set virtual switch property to "Not connected" in Network Adapter settings of VM I get successful migration.
    All VM's that are not connected to any private networks (virtual switches with private network connection type) can be live migrated without any issues.
    Is there any official reference related to Hyper-V live migration of VM's that have "private network" connection type?

    I can Live Migrate virtual machines with adapters on private switches without error. Aside from having the wrong name, the only way I can get it to fail is if I make the switch on one host use a different QoS minimum mode than the other and
    enable QoS on the virtual adapter. Even then I get a different message than what you're getting. I only get that one with differently named switches.
    There is a PowerShell cmdlet available to see why a guest won't run on another host.
    Here's an example of its usage.
    There's a way to use it to get it to Live Migrate.
    But there is no way to truly Live Migrate three virtual machines in perfect lockstep. Even if you figure out whatever is preventing you from migrating these machines, there will still be periods during Live Migration where they can't communicate across that
    private network. You also can't guarantee that all these guests will always be running on the same host without preventing Live Migration in the first place. This is why there really isn't anyone doing what you're trying to do. I suggest you consider another
    isolation solution, like VLANs.
    Eric Siron Altaro Hyper-V Blog
    I am an independent blog contributor, not an Altaro employee. I am solely responsible for the content of my posts.
    "Every relationship you have is in worse shape than you think."

  • How to Fix: Error (10698) The virtual machine () could not be live migrated to the virtual machine host () using this cluster configuration.

    I am unable to live migrate via SCVMM 2012 R2 to one Host in our 5 node cluster.  The job fails with the errors below.
    Error (10698)
    The virtual machine () could not be live migrated to the virtual machine host () using this cluster configuration.
    Recommended Action
    Check the cluster configuration and then try the operation again.
    Information (11037)
    There currently are no network adapters with network optimization available on host.
    The host properties indicate network optimization is available as indicated in the screen shot below.
    Any guidance on things to check is appreciated.
    Thanks,
    Glenn

    Here is a snippet of the cluster log when from the current VM owner node of the failed migration:
    00000e50.000025c0::2014/02/03-13:16:07.495 INFO  [RHS] Resource Virtual Machine Configuration VMNameHere called SetResourceLockedMode. LockedModeEnabled0, LockedModeReason0.
    00000b6c.00001a9c::2014/02/03-13:16:07.495 INFO  [RCM] HandleMonitorReply: LOCKEDMODE for 'Virtual Machine Configuration VMNameHere', gen(0) result 0/0.
    00000e50.000025c0::2014/02/03-13:16:07.495 INFO  [RHS] Resource Virtual Machine VMNameHere called SetResourceLockedMode. LockedModeEnabled0, LockedModeReason0.
    00000b6c.00001a9c::2014/02/03-13:16:07.495 INFO  [RCM] HandleMonitorReply: LOCKEDMODE for 'Virtual Machine VMNameHere', gen(0) result 0/0.
    00000b6c.00001a9c::2014/02/03-13:16:07.495 INFO  [RCM] HandleMonitorReply: INMEMORY_NODELOCAL_PROPERTIES for 'Virtual Machine VMNameHere', gen(0) result 0/0.
    00000b6c.000020ec::2014/02/03-13:16:07.495 INFO  [GEM] Node 3: Sending 1 messages as a batched GEM message
    00000e50.000025c0::2014/02/03-13:16:07.495 INFO  [RES] Virtual Machine Configuration <Virtual Machine Configuration VMNameHere>: Current state 'MigrationSrcWaitForOffline', event 'MigrationSrcCompleted', result 0x8007274d
    00000e50.000025c0::2014/02/03-13:16:07.495 INFO  [RES] Virtual Machine Configuration <Virtual Machine Configuration VMNameHere>: State change 'MigrationSrcWaitForOffline' -> 'Online'
    00000e50.000025c0::2014/02/03-13:16:07.495 INFO  [RES] Virtual Machine <Virtual Machine VMNameHere>: Current state 'MigrationSrcOfflinePending', event 'MigrationSrcCompleted', result 0x8007274d
    00000e50.000025c0::2014/02/03-13:16:07.495 INFO  [RES] Virtual Machine <Virtual Machine VMNameHere>: State change 'MigrationSrcOfflinePending' -> 'Online'
    00000e50.00002080::2014/02/03-13:16:07.510 ERR   [RES] Virtual Machine <Virtual Machine VMNameHere>: Live migration of 'Virtual Machine VMNameHere' failed.
    Virtual machine migration operation for 'VMNameHere' failed at migration source 'SourceHostNameHere'. (Virtual machine ID 6901D5F8-B759-4557-8A28-E36173A14443)
    The Virtual Machine Management Service failed to establish a connection for a Virtual Machine migration with host 'DestinationHostNameHere': No connection could be made because the tar
    00000e50.00002080::2014/02/03-13:16:07.510 ERR   [RHS] Resource Virtual Machine VMNameHere has cancelled offline with error code 10061.
    00000b6c.000020ec::2014/02/03-13:16:07.510 INFO  [RCM] HandleMonitorReply: OFFLINERESOURCE for 'Virtual Machine VMNameHere', gen(0) result 0/10061.
    00000b6c.000020ec::2014/02/03-13:16:07.510 INFO  [RCM] Res Virtual Machine VMNameHere: OfflinePending -> Online( StateUnknown )
    00000b6c.000020ec::2014/02/03-13:16:07.510 INFO  [RCM] TransitionToState(Virtual Machine VMNameHere) OfflinePending-->Online.
    00000b6c.00001a9c::2014/02/03-13:16:07.510 INFO  [GEM] Node 3: Sending 1 messages as a batched GEM message
    00000b6c.000020ec::2014/02/03-13:16:07.510 INFO  [RCM] rcm::QueuedMovesHolder::VetoOffline: (VMNameHere with flags 0)
    00000b6c.000020ec::2014/02/03-13:16:07.510 INFO  [RCM] rcm::QueuedMovesHolder::RemoveGroup: (VMNameHere) GroupBeingMoved: false AllowMoveCancel: true NotifyMoveFailure: true
    00000b6c.000020ec::2014/02/03-13:16:07.510 INFO  [RCM] VMNameHere: Removed Flags 4 from StatusInformation. New StatusInformation 0
    00000b6c.000020ec::2014/02/03-13:16:07.510 INFO  [RCM] rcm::RcmGroup::CancelClusterGroupOperation: (VMNameHere)
    00000b6c.00001a9c::2014/02/03-13:16:07.510 INFO  [GEM] Node 3: Sending 1 messages as a batched GEM message
    00000b6c.000021a8::2014/02/03-13:16:07.510 INFO  [GUM] Node 3: executing request locally, gumId:3951, my action: /dm/update, # of updates: 1
    00000b6c.000021a8::2014/02/03-13:16:07.510 INFO  [GEM] Node 3: Sending 1 messages as a batched GEM message
    00000b6c.00001a9c::2014/02/03-13:16:07.510 INFO  [GEM] Node 3: Sending 1 messages as a batched GEM message
    00000b6c.000022a0::2014/02/03-13:16:07.510 INFO  [RCM] moved 0 tasks from staging set to task set.  TaskSetSize=0
    00000b6c.000022a0::2014/02/03-13:16:07.510 INFO  [RCM] rcm::RcmPriorityManager::StartGroups: [RCM] done, executed 0 tasks
    00000b6c.00000dd8::2014/02/03-13:16:07.510 INFO  [RCM] ignored non-local state Online for group VMNameHere
    00000b6c.000021a8::2014/02/03-13:16:07.526 INFO  [GUM] Node 3: executing request locally, gumId:3952, my action: /dm/update, # of updates: 1
    00000b6c.000021a8::2014/02/03-13:16:07.526 INFO  [GEM] Node 3: Sending 1 messages as a batched GEM message
    00000b6c.000018e4::2014/02/03-13:16:07.526 INFO  [RCM] HandleMonitorReply: INMEMORY_NODELOCAL_PROPERTIES for 'Virtual Machine VMNameHere', gen(0) result 0/0.
    No entry is made on the cluster log of the destination node. 
    To me this means the nodes cannot talk to each other, but I don’t know why.  
    They are on the same domain.  Their server names resolve properly and they can ping eachother both by name and IP.

  • Oracle vm 3.1.1 ( kernel 2.6.39-200.1.9.el5uek ) : Live migration breaks

    Hello,
    we are upgrading with last update of OVS the servers.. After we upgraded one server (called A) from 2.6.39-200.1.1.el5uek to 2.6.39-200.1.9.el5uek with certified yum repository from Oracle, the live migration not working anymore in correct mode.
    If I will migrate one guest from another server to server A or vice versa , the results are the same, the 3%-10% of packets dropped. Is it a normal behaviour , if the kernel are different ? Or are this kernel/driver/xen bugged ?
    Obviously the version of OVS are always 3.1.1 and the oracle vm the 3.1.1 build 478, and previously the live migration always worked well. No errors are visible and the job gone well.
    Kind Regards
    Edited by: user10717184 on Oct 29, 2012 12:46 AM

    I try to migrate with xm command but the problem not disappear .
    The xm command not give any result code. it finished correctly, by the way or we lost 3-10% of packets or stop pinging .
    [root@******** ~]# xm migrate -l ****UUID*** ****SERVER_OVS_NAME***
    [root@******** ~]# echo $?
    0
    Now the server have both the new kernel, but it continues to have the problem. The strange thing is that if you return to previous server OVS, the pinging restart, sometime.

  • Losing pings on live migration

    I currently have a 2 node OVM 3.1.1 cluster, fully patched, and have noticed that I lose network connectivity while performing a live migration. I have separate networks for VM's and for live migration. I am new to OVM so I do not know if this is typical or not. It appears that every time I migrate a VM, the client will lose connectivity to the network for anywhere from 5 - 20 seconds.

    Are you talking about lost pings to the guest VM that is live migrated? If yes, that I'd suppose that to be normal behaviour, since there has to be some interruption, once the memory contents is finally synchronized between the source and the target VM server.
    Depending on much RAM is used by the running VM and the speed of your network, this "outage" might vary in time, but there will surely always be some time span where pings to the VM that is being live migrated get lost.

  • Live migration from oracle 8.0.3.0.0 to oracle 10g

    Hello everyone,
    I have Oracle 8.0.3.0.0 running on Windows NT platform. I wish to migrate from Oracle 8.0.3.0.0 to Oracle 10g. And I wish to do it without shutting down my Oracle 8 server. Sort of live migration. Since transaction rate is very high. And putting down the server would mean hanging up whole work. Is it possible. If yes what all hardware and software requirements are there? Or Import and Export is the only option,
    In anticipation of your guidlines,
    Durgesh
    919416228922

    Durgesh,
    <br>No, you cannot migrate directly from 8.0.3 to 10g (either 10gR1 and 10gR2). Read the upgrade guide which concern your target 10g release :</br>
    <br>Oracle Database Upgrade Guide 10g Release 2 (10.2)</br>
    <br>Oracle Database Upgrade Guide 10g Release 1 (10.1)</br>
    <br>Nicolas.</br>

  • Server 2012 r2 live migration fails with hardware error

    Hello all, we just upgraded one of our hyper v hosts from server 2012 to server 2012 r2; previously we had live replication setup between it and another box on the network which was also running server 2012. After installing server 2012 r2 when a live migration
    is attempted we get the message:
    "The virtual machine cannot be moved to the destination computer. The hardware on the destination computer is not compatible with the hardware requirements of this virtual machine. Virtual machine migration failed at migration source."
    The servers in question are both dell, currently we have a poweredge r910 running server 2012 and a poweredge r900 running server 2012 r2. The section under processor for "migrate to a physical computer using a different processor" is already checked
    and this same vm was successfully being live replicated before the upgrade to server 2012 r2. What would have changed around hardware requirements?
    We are migrating from server 2012 on the poweredge r910 to server 2012 r2 on the poweredge r900. Also When I say this was an upgrade, we did a full re install and wiped out the installation of server 2012 and installed server 2012 r2, this was not an upgrade
    installation.

    The only cause I’ve seen so far is virtual switches being named differently. I do remember that one of our VMs didn’t move, but we simply bypassed this problem, using one-time backup (VeeamZIP, more specifically).
    If it’s one-time operation you can use the same procedure for the VMs in question -> backup and restore them at new server.
    Kind regards, Leonardo.

  • Live Migration Fails with error Synthetic FiberChannel Port: Failed to finish reserving resources on an VM using Windows Server 2012 R2 Hyper-V

    Hi, I'm currently experiencing a problem with some VMs in a Hyper-V 2012 R2 failover cluster using Fiber Channel adapters with Virtual SAN configured on the hyper-v hosts.
    I have read several articles about this issues like this ones:
    https://social.technet.microsoft.com/Forums/windowsserver/en-US/baca348d-fb57-4d8f-978b-f1e7282f89a1/synthetic-fibrechannel-port-failed-to-start-reserving-resources-with-error-insufficient-system?forum=winserverhyperv
    http://social.technet.microsoft.com/wiki/contents/articles/18698.hyper-v-virtual-fibre-channel-troubleshooting-guide.aspx
    But haven't been able to fix my issue.
    The Virtual SAN is configured on every hyper-v host node in the cluster. And every VM has 2 fiber channel adapters configured.
    All the World Wide Names are configured both on the FC Switch as well as the FC SAN.
    All the drivers for the FC Adapter in the Hyper-V Hosts have been updated to their latest versions.
    The strange thing is that the issue is not affecting all of the VMs, some of the VMs with FC adapters configured are live migrating just fine, others are getting this error.
    Quick migration works without problems.
    We even tried removing and creating new FC Adapters on a VM with problems, we had to configure the switch and SAN with the new WWN names and all, but ended up having the same problem.
    At first we thought is was related to the hosts, but since some VMs do work live migrating with FC adapters we tried migrating them on every host, everything worked well.
    My guess is that it has to be something related to the VMs itself but I haven't been able to figure out what is it.
    Any ideas on how to solve this is deeply appreciated.
    Thank you!
    Eduardo Rojas

    Hi Eduardo,
    How are things going ?
    Best Regards
    Elton Ji
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Live Migration failed using virtual HBA's and Guest Clustering

    Hi,
    We have a Guest Cluster Configuration on top of an Hyper-V Cluster. We are using Windows 2012 and Fiber Channel shared storage.
    The problem is regarding Live Migration. Some times when we move a virtual machine from node A to node B everything goes well but when we try to move back to node A Live Migration fails. What we can see is that when we move the VM from node A to B and Live
    Migration completes successfully the virtual ports remain active on node A, so when we try to move back from B to A Live Migration fails because the virtual ports are already there.
    This doesn't happen every time.
    We have checked the zoning between Host Cluster Hyper-V and the SAN, the mapping between physical HBA's and the vSAN's on the Hyper-V and everything is ok.
    Our doubt is, what is the best practice for zoning the vHBA on the VM's and our Fabric? We setup our zoning using an alias for the vHBA 1 and the two WWN (A and B) on the same object and an alias for the vHBA 2 and the correspondent WWN (A and B). Is it
    better to create an alias for vHBA 1 -> A (with WWN A) and other alias for vHBA 1 -> B (with WWN B)? 
    The guest cluster VM's have 98GB of RAM each. Could it be a time out issue when Live Migration happen's and the virtual ports remain active on the source node? When everything goes well, the VM moves from node A with vHBA WWN A to node B and stays there
    with vHBA WWN B. On the source node the virtual ports should be removed automatically when the Live Migration completes. And that is the issue... sometimes the virtual ports (WWN A) stay active on the source node and when we try to move back the VM Live Migration
    fails.
    I hope You may understand the issue.
    Regards,
    Carlos Monteiro.

    Hi ,
    Hope the following link may help.
    To support live migration of virtual machines across Hyper-V hosts while maintaining Fibre Channel connectivity, two WWNs are configured for each virtual Fibre Channel adapter: Set A and Set B. Hyper-V automatically alternates between the Set A and Set B
    WWN addresses during a live migration. This ensures that all LUNs are available on the destination host before the migration and that no downtime occurs during the migration.
    Hyper-V Virtual Fibre Channel Overview
    http://technet.microsoft.com/en-us/library/hh831413.aspx
    More information:
    Hyper-V Virtual Fibre Channel Troubleshooting Guide
    http://social.technet.microsoft.com/wiki/contents/articles/18698.hyper-v-virtual-fibre-channel-troubleshooting-guide.aspx
    Hyper-V Virtual Fibre Channel Design Guide
    http://blogs.technet.com/b/privatecloud/archive/2013/07/23/hyper-v-virtual-fibre-channel-design-guide.aspx
    Hyper-V virtual SAN
    http://salworx.blogspot.co.uk/
    Thanks.
    Please remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading the thread.

  • Live Migration Failed After Yum Update on 2.2.2

    Hi,
    I've live migrated vm's from node2 to node1 (master server). Put node2 into maintenance mode, then reconfigure ntp.conf to sync with our new ntp server. While in maintenance mode, I've also ran yum update which updated the kernel from 2.6.18-128.2.1.4.37.el5xen to 2.6.18-128.2.1.4.44.el5xen. In the process, xen was also updated from 3.4.0-0.1.32.el5 to 3.4.0-0.1.39.el5.
    After rebooting, I've put node2 back into active mode. But now I can't live migrate the vm's back into node 2. From node1, ovs_operations.log:
    "2012-06-23 11:03:28" INFO=> migrate_vm: vm('/OVS/running_pool/1210_vm05') start...
    "2012-06-23 11:03:29" INFO=> xen_migrate_vm: migrate with ssl enabled failed, do failover(no ssl). vm('/var/ovs/mount/54DA5753709A48B3BFAEE65C2EAECCE0/running_pool/1210_vm05') -> tgt_srv('node2')
    "2012-06-23 11:03:30" ERROR=> xen_migrate_vm: failed. vm('/var/ovs/mount/54DA5753709A48B3BFAEE65C2EAECCE0/running_pool/1210_vm05') -> tgt_srv('node2') =><Exception: xen_migrate_vm: migrate without ssl failed either.>
    From node2 ovs_operations.log:
    "2012-06-23 11:03:30" INFO=> ha_join_dlm_domain: =>success
    "2012-06-23 11:03:30" ERROR=> ha_set_dlm_lock:failed. lock('6f1cb211-ecef-4ac6-af2b-091fc6fd5966') name('1210_vm05')=> <Exception: create lock('/dlm/ovm/f1cb211ecef4ac6af2b091fc6fd5966') failed. <OSError: [Errno 26] Text file busy: '/dlm/ovm/f1cb211ecef4ac6af2b091fc6fd5966'>
    Is this due to the different xen versions after the update? How can I live migrate the vm's back to node2 so that I can take node1 into maintenance mode to reconfigure ntp.conf and run yum update?
    Thanks!

    The next thing to try is to work though the steps in [[Error loading web sites]]. Though if this is happening on all of your computers it might have something to do with your internet connection. Maybe resetting that will help.

  • Storage live migration leaves files in the old location

    I have Hyper-v cluster and a new Scale-put file server cluster. I am in the process of storage live migrating my vms into the SOFS cluster. Some Vms migrate over fine and other leave references to the old storage location when the migration completes.
    I have configured constrained delegation between the Hyper-V host and also SMB delegation on the sofs nodes.
    Any ideas on why it is not the storage live migration is not completely migrating over to the new location?
    thanks so much for your time

    Yes i have all updates on sofs cluster nodes a hyper-v cluster nodes.
    Its weird that  hard drive vm property shows new location,  but seems that the vm xml config files are actually still referenced and being used in vm properties.
    Thanks for the help

  • Live Migration Failed while Quick Migration is Ok...Virtual machine with synthetic FC HBA !

    When I migration the virtual machine with synthetic FC HBA  in windows server 2012 R2 Cluster,it fails
    but I do it in the style of quick migraton ,it secceed!
    The error event here
    Live migration of 'Virtual Machine PTSCSQL01' failed.
    Virtual machine migration operation for 'PTSCSQL01' failed at migration destination 'PTCLS0106'. (Virtual machine ID B8FBDE64-FF97-4E9B-BC40-6DCFA09B31BE)
    'PTSCSQL01' Synthetic FibreChannel Port: Failed to finish reserving resources with Error 'Unspecified error' (0x80004005). (Virtual machine ID B8FBDE64-FF97-4E9B-BC40-6DCFA09B31BE)
    'PTSCSQL01' Synthetic FibreChannel Port: Failed to finish reserving resources with Error 'Unspecified error' (0x80004005). (Virtual machine ID B8FBDE64-FF97-4E9B-BC40-6DCFA09B31BE)
    My virtual machine's  synthetic FC HBA setting here
    做微软的先行者,享受用户体验

    Yes, definitely check your zoning/masking.  Remember that with vHBA you have twice as many WWPNs to account for.  Performing a live migration makes use of both pairs during the transfer from one host to the other - one set is active on the machine
    currently running and the second set is used to ensure connectivity on the destination.  So if you are using Address Set A on Host1, Host2 will try to set up the fibre channel connection using Address Set B.  If you do a quick migration, you would
    continue to use the same Address Set on the second host.  That's why you most likely need to check your zoning/masking for the alternate set.
    . : | : . : | : . tim

  • Live Migration - Failed to establish a connection because of an unsupported protocol version (protocol version 5.0)

    I'm getting a not the disk and everything) back to the host I originally had it on.  Here is my setup:
    Host 1 - Microsoft Storage Server 2012 Standard (NAS)
    Host 2 - Microsoft Hyper-V Server 2012 (All connected virtual hard drives use the SMB 3.0 protocol to connect to the NAS - Host 1)
    Delegation is setup on both of these hosts to include:
    -Trust this computer, Use Kerberos only, Services - cifs, Hyper-V Replication, Microsoft Virtual Console Service, Microsoft Virtual System Migration Service
    Live Migration is setup on both of these hosts to include:
    -Enable incoming and outgoing live migrations using Kerberos
    -Use any available network for live migration
    I was able to migrate the VM from host 1 to host 2 with no problems.  But when I try to migrate the VM host 1 that's when I encounter the error. 
    I saw a KB article similar to the issue I'm having but it only effects Server 2008 R2 (KB2705759)
    I've rebooted both hosts but still get the error.  The VM that is on host 2 has it's hard drive on host 1 run's fine so there are no connection issues there. 
    I've searched the Internet for this issue and keep coming up empty handed.  I even named the virtual switch on both hosts the same to see if that would work.  Please help!  I'd rather not do an export and import of the VM because I'm evaluating
    this product so I want to make sure it works before I put it into production. 

    Hi PPlbrad,
    Agree with Shabarinath .
    We can just live migration the VM from low version (2012) to high version (2012R2) but the reverse is not ture .
    So do hyper-v replica (just for updating )
    Best Regards
    Elton JI
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

Maybe you are looking for

  • Data Not updated in BDC

    hi alll 1.   i have written below coding for vendor master update. i have two records in text file, in that first record not updated. while updating its shows error. but second record updated. how can i solve. plse help me 2.  I want to show the erro

  • How Can I Check if my Windows Phone is Refurbished...

    Hey, Im going to buy the lumia 920 and the price that the store gave to me sounds little too much *cheap* So i want to confirm that the phone is Factory-new and not Refurbished while im buying the phone. How can i do that? There's any kind of code? c

  • Where's the beef? I mean the audio...?

    When I export a movie from FCP with audio (it gives me the option) and then import it to LT the audio channels do not appear. What am I missing? Also, is there any upgrade for LT with newer Live Fonts and other stuff? Thanks, Carlos

  • The radio changes stations on its own

    The radio just changes stations by itself - it is irritating...i'm trying to study to nice, mellow music and all the sudden I get celtic fling music.....which is fine if I'm drinking a guiness, not so good if I'm trying to concentrate!

  • Is canon rebel t3 comes with factory warranty ?

    I brought this camera on NOv 28 2013. And battery kind of low now. I need new battery. I'm wondering if it's cover by warranty before buying new one.