Oracle VM (XEN) and Live Migration throws PCI error

First off, anyone using a HS21 XM blade with Oracle VM? Anyone attempted a live-migration, does it work?
When attempting a live migration on a HS21 XM blade we get the following errors and the host hangs:
102 I Blade_02 08/10/08, 08:46:38 (CADCOVMA02) Blade reboot
103 E Blade_02 08/10/08, 08:46:26 (CADCOVMA02) Software critical interrupt.
104 E Blade_02 08/10/08, 08:46:21 (CADCOVMA02) SMI Hdlr: 00150700 PERR: Slave signaled parity error B/D/F/Slot=07000300 Vend=8086 Dev=3 0C Status=8000
105 E Blade_02 08/10/08, 08:46:19 (CADCOVMA02) SMI Hdlr: 00150900 SERR/PERR Detected on PCI bus
106 E Blade_02 08/10/08, 08:46:19 (CADCOVMA02) SMI Hdlr: 00150700 PERR: Slave signaled parity error B/D/F/Slot=00020000 Vend=8086 Dev=25F7 Statu
107 E Blade_02 08/10/08, 08:46:18 (CADCOVMA02) PCI PERR: parity error.
108 E Blade_02 08/10/08, 08:46:17 (CADCOVMA02) SMI Hdlr: 00150900 SERR/PERR Detected on PCI bus
109 E Blade_02 08/10/08, 08:46:17 (CADCOVMA02) SMI Hdlr: 00150400 SERR: Device Signaled SERR B/D/F/Slot=07000300 Vend=8086 Dev=350C Sta us=4110
110 E Blade_02 08/10/08, 08:46:16 (CADCOVMA02) SMI Hdlr: 00150400 SERR: Device Signaled SERR B/D/F/Slot=00020000 Vend=8086 Dev=25F7 Status=4110
111 E Blade_02 08/10/08, 08:46:14 (CADCOVMA02) PCI system error.
any ideas? I've called IBM support but their options are reseat the hardware or replace it. this happens on more than one blade so we're assuming it has something to do with oracle VM. Thanks!

Hi Eduardo,
How are things going ?
Best Regards
Elton Ji
We
are trying to better understand customer views on social support experience, so your participation in this
interview project would be greatly appreciated if you have time.
Thanks for helping make community forums a great place.

Similar Messages

  • When setting up converged network in VMM cluster and live migration virtual nics not working

    Hello Everyone,
    I am having issues setting up converged network in VMM.  I have been working with MS engineers to no avail.  I am very surprised with the expertise of the MS engineers.  They had no idea what a converged network even was.  I had way more
    experience then these guys and they said there was no escalation track so I am posting here in hopes of getting some assistance.
    Everyone including our consultants says my setup is correct. 
    What I want to do:
    I have servers with 5 nics and want to use 3 of the nics for a team and then configure cluster, live migration and host management as virtual network adapters.  I have created all my logical networks, port profile with the uplink defined as team and
    networks selected.  Created logical switch and associated portprofle.  When I deploy logical switch and create virtual network adapters the logical switch works for VMs and my management nic works as well.  Problem is that the cluster and live
    migration virtual nics do not work.  The correct Vlans get pulled in for the corresponding networks and If I run get-vmnetworkadaptervlan it shows cluster and live migration in vlans 14 and 15 which is correct.  However nics do not work at all.
    I finally decided to do this via the host in powershell and everything works fine which means this is definitely an issue with VMM.  I then imported host into VMM again but now I cannot use any of the objects I created and VMM and have to use standard
    switch.
    I am really losing faith in VMM fast. 
    Hosts are 2012 R2 and VMM is 2012 R2 all fresh builds with latest drivers
    Thanks

    Have you checked our whitepaper http://gallery.technet.microsoft.com/Hybrid-Cloud-with-NVGRE-aa6e1e9a for how to configure this through VMM?
    Are you using static IP address assignment for those vNICs?
    Are you sure your are teaming the correct physical adapters where the VLANs are trunked through the connected ports?
    Note; if you create the teaming configuration outside of VMM, and then import the hosts to VMM, then VMM will not recognize the configuration. 
    The details should be all in this whitepaper.
    -kn
    Kristian (Virtualization and some coffee: http://kristiannese.blogspot.com )

  • 2012 R2 Cluster and Live Migration

    6 Node Cluster with Server 2012 R2, all VM's are Server 2012 R2
    4 Fiber SAN's
    Moving and Live Migration worked fine in FailOver Cluster Manager
    But this time we were trying do it with SCVMM 2012 R2 and just move one VM (Gen2)
    Of course it failed at 99%
    Error (12711)
    VMM cannot complete the WMI operation on the server (whatever) because of an error: [MSCluster_Resource.Name="SCVMM VMHost"] The cluster resource could not be found.
    The cluster resource could not be found (0x138F)
    Recommended Action
    Resolve the issue and then try the operation again.
    How do I fix this? the VM is still running. The two vhdx files it was moving are smaller then orginal's , but it change the configuration file to point to new ones, which are bad.
    it says I can Repair it... Redo or Undo....of course neither of those options work.
    Wait for the object to be updated automatically by the next periodic Live migrate storage of virtual machine vmhost from whatever to whatever job.
    ID: 1708
    Cluster has no errors, SAN's have no errors, CSV have no errors. the machine running scvmm is VM running on the cluster

    How did you create this VM? if this is created outside of VMM, I recommend doing a manual refresh of the VM first to ensure that VMM can read its attributes. Then retry the operation.
    Btw, are the VMs using diff disk? any checkpoints associated with them?
    Kristian (Virtualization and some coffee: http://kristiannese.blogspot.com )

  • Best Configuration for T4-4 and live migration

    I have several T4-4 servers with Solaris 11.1
    What kind of settings I need for production environments with OVM live migration?
    Should I use OVM manager for rugged environments?
    Should I set two io-domain on each server?
    Considering that I do live migration, for guess ldom system disk should I configure (publish) one big LUN to all servers or just a separate LUN for each ldom guest.

    For the first questions: it's best to have separate LUNs or other backend disks for each guest - otherwise how would you plan to share them across different guests and systems? Separate LUNs also is good for performance by encouraging parallel I/O. Choice of using an additional I/O domain is up to you, but it's very typical for production users because of the resiliency it adds.
    For advice on live migration, seehttps://blogs.oracle.com/jsavit/entry/best_practices_live_migration
    Other blog posts after that discuss availability and performance.
    Hope that helps, Jeff

  • Hyper V Lab and Live Migration

    Hi Guys,
    I have 2 Hyper V hosts I am setting up in a lab environment. Initially, I successfully setup a 2 node cluster with CSVs which allowed me to do Live Migrations. 
    The problem I have is my shared storage is a bit of a cheat as I have one disk assigned in each host and each host has starwinds virtual SAN installed. The hostA has an iscsi connection to hostB storage and visa versa.
    The issue this causes is when the hosts shutdown (because this is a lab its only on when required), the cluster is in a mess when it starts up eg VMs missing etc. I can recover from it but it takes time. I tinkered with the HA settings and the VM settings
    so they restarted/ didnt restart etc but with no success.
    My question is can I use something like SMB3 shared storage on one of the hosts to perform live migrations but without a full on cluster? I know I can do Shared Nothing Live Migrations but this takes time.
    Any ideas on a better solution (rather than actually buying proper shared storage ;-) ) Or if shared storage is the only option to do this cleanly, what would people recommend bearing in mind I have SSDs in the hyper V hosts.
    Hope all that makes sense

    Hi Sir,
    >>I have 2 Hyper V hosts I am setting up in a lab environment. Initially, I successfully setup a 2 node cluster with CSVs which allowed me to do Live Migrations. 
    As you mentioned , you have 2 hyper-v host and use starwind to provide ISCSI target (this is same as my first lab environment ) , then I realized that I need one or more host to simulate more production scenario . 
    But ,if you have more physical computers you may try other's progects .
    Also please refer to this thread :
    https://social.technet.microsoft.com/Forums/windowsserver/en-US/e9f81a9e-0d50-4bca-a24d-608a4cce20e8/2012-r2-hyperv-cluster-smb-30-sofs-share-permissions-issues?forum=winserverhyperv
    Best Regards
    Elton Ji
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • LDOMs, Solaris zones and Live Migration

    Hi all,
    If you are planning to use Solaris zones inside a LDOM and using an external zpool as Solaris zone disk, wouldn't this break one of the requirements for being able to do a Live Migration ? If so, do you have any ideas on how to use Solaris zones inside an LDOM and at the same time be able to do a Live Migration or is it impossible ? I know this may sound as a bad idea but I would very much like to know if it is doable.

    Thanks,
    By external pool I am thinking of the way you probably are doing it, separate LUNs mirrored in a zpool for the zones coming from two separate IO/Service domains. So even if this zpool exist inside the LDOM as zone storage this will not prevent LM ? That's good news. The requirement "no zpool if Live Migration" must then only be valid for the LDOM storage itself and not for storage attached to the running LDOM. I am also worried about a possible performance penalty introducing an extra layer of virtualisation. Have you done any tests regarding this ?

  • Live Migration Fails with error Synthetic FiberChannel Port: Failed to finish reserving resources on an VM using Windows Server 2012 R2 Hyper-V

    Hi, I'm currently experiencing a problem with some VMs in a Hyper-V 2012 R2 failover cluster using Fiber Channel adapters with Virtual SAN configured on the hyper-v hosts.
    I have read several articles about this issues like this ones:
    https://social.technet.microsoft.com/Forums/windowsserver/en-US/baca348d-fb57-4d8f-978b-f1e7282f89a1/synthetic-fibrechannel-port-failed-to-start-reserving-resources-with-error-insufficient-system?forum=winserverhyperv
    http://social.technet.microsoft.com/wiki/contents/articles/18698.hyper-v-virtual-fibre-channel-troubleshooting-guide.aspx
    But haven't been able to fix my issue.
    The Virtual SAN is configured on every hyper-v host node in the cluster. And every VM has 2 fiber channel adapters configured.
    All the World Wide Names are configured both on the FC Switch as well as the FC SAN.
    All the drivers for the FC Adapter in the Hyper-V Hosts have been updated to their latest versions.
    The strange thing is that the issue is not affecting all of the VMs, some of the VMs with FC adapters configured are live migrating just fine, others are getting this error.
    Quick migration works without problems.
    We even tried removing and creating new FC Adapters on a VM with problems, we had to configure the switch and SAN with the new WWN names and all, but ended up having the same problem.
    At first we thought is was related to the hosts, but since some VMs do work live migrating with FC adapters we tried migrating them on every host, everything worked well.
    My guess is that it has to be something related to the VMs itself but I haven't been able to figure out what is it.
    Any ideas on how to solve this is deeply appreciated.
    Thank you!
    Eduardo Rojas

    Hi Eduardo,
    How are things going ?
    Best Regards
    Elton Ji
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • MSS Requisitions is not moving to Recruiter's POWL and system is throwing the errors.

    Hi everyone,
    Hope all are doing in good spirit.
    I have an issue in regards with Requisition Creation and its related process.
    I am working in Standalone architecture where Recruiting server is in another instance apart from HR system.
    We are using WS45600028 standard workflow.
    When Manager submits the Requisition to higher level manager to approve, its done successfully and the notification is going to the Manager. But after approval, Requisition is not going to Requiter's POWL queries.
    I have maintained Support Group, Lead Recruiter in it and assigned this to Requisition.
    I have checked HRP5131 table. But there is no SOBID which sould be flaged as Lead Recruiter.
    When Higher level Manager approves the Requisition, It is successful bu there is an error in slg1.
    Error: Error when ending the Work item 0000007704 in HR system.
    The system returned the call of interface SAP_WAPI_WRITE_CONTAINER in this context
    System Response:
    Call ended with an error. The function was not executed.
    Procedure: Remove call.
    At the same system is throwing the error for the User (This user is used for RFC creation using with transaction SM59): No candidate for this User.
    Details of error: The error occurred in program CL_HRRCF_CANDIDATE_ADMIN_BL===CM00B line 77
    Procedure for system administration:
    Check the context in which the error occurs.
    Since it it probably program error and not an input error, you should contact SAP.
    Can you please look into the above issue and address me if you have any solution for the same.
    Any help would be greatly appreciated.
    Thank you.
    Regards,
    Sekhar

    Hi Nicloe,
    Thank you very much for your response.
    I have done the data transfer from HR system to E-Rec system through ALE iDoc using PFAL.
    All the is transferred and corresponding NA, CP, BP generated in E-Rec for P.
    All the users(Recruiter, Managers and Recruiting Admin) data have transferred to E-Rec System.
    But when Manager is creating the Requisition Request, it going to his higher level manager for approval and its been approved successfully. Later it is not updating the Recruiter's POWL.
    In DB table HRP5125 Requisition data is created but with only few details like Start and End date, Object ID. in E-Rec system.
    In HRP5131, NB related data is there but there is no SOBID for lead Recruiter.
    Immediately i am getting errors in both HR and E-Rec systems as follows after the Requisition got approved.
    In HR system, Error when ending the Work item 0000007704. Workflow completed successfully but still here is an error.
    In E-Rec systemThe error occurred in program CL_HRRCF_CANDIDATE_ADMIN_BL===CM00B line 77.
    And system says, Since it it probably program error and not an input error, you should contact SAP.
    I heard that there might be an inconsistency in NA, BP, CP when we do PFAL transaction for data transfers. Can you suggest me to raise SAP ticket in this context?
    Or Did i miss any configuration in this connection?
    What can be done now to fix this? I am working EHP 5 and SP level is 11.
    Can you please let me know. Kindly help me out of it as i am struck with this since very long time.
    Thank you.
    Regards,
    Sekhar

  • Installing CF10, in Configuration and Settings Migration Wizard - getting error: Invalid Password

    Installing CF10, in Configuration and Settings Migration Wizard, in browserhttp://127.0.0.1/CFIDE/administrator/enter.cfm,
    ColdFusion has been successfully installed. This wizard will guide you through the remaining server configuration steps and, if applicable, migrate settings from a previous version of ColdFusion.
    To guarantee the security of your server, please enter your ColdFusion Administrator password.
    - getting error: Invalid Password. It won't take the password I provided in previous "Set Admin password" screen. Help?

    Hi,
    I am having the issue with the URL on the Configuration ans Setting Migration Wizard pointing to the following URL:
    http://localhost/CFIDE/administrator/enter.cfm
    Receiving a message:
    HTTP Error 404.0 - Not Found
    The resource you are looking for has been removed, had its name changed, or is temporarily unavailable.
    The following is the information in the error page:
    Module
    IIS Web Core
    Notification
    MapRequestHandler
    Handler
    ISAPI-dll
    Error Code
    0x80070002
    Requested URL
    http://localhost:80/jakarta/isapi_redirect.dll
    Physical Path
    D:\ColdFusion10\cfusion\wwwroot\jakarta\isapi_redirect.dll
    Logon Method
    Anonymous
    Logon User
    Anonymous
    I could not find a directory called jakarta.
    Any ideas?
    Thanks,
    Mike

  • Hyper-V host crash and live migration - what happens?

    Hello there,
    it would be great if you could help me understand how live migrationworks on a sudden host crash of a cluster node.
    I know that the cluster functionality mounts the vhd file on another host and runs the VM.
    But what happens to the VM itself? Does it crash too and is simply restarted on another host?
    To my knowledge the RAM of the VM is run in the RAM of the host, wo I would assume that is all lost on a host crash?
    Thanks for any help understanding this,
    Best regards, Marcus

    In a sudden crash, there is no time for the state of any running VMs to move to another system.  Therefore, the VM is restarted on another node of the cluster.  Think of the host crashing as being the same as pulling the power cable on the VM.
    . : | : . : | : . tim

  • Oracle instant Client and unixODBC for Linux has Error

    Hi experts,
    I have two Redhat EL 5. On the first Rehat(xxx.xxx.xxx.121 --> called Redhat1) I have installed Oracle 10g database server. On the second Redhat(xxx.xxx.xxx.123 --> called Redhat2) I want to connect to Oracle database server on xxx.xxx.xxx.121 via ODBC. But it does not work. Pls help me!
    I have done steps as following:
    Login to Redhat2 as root and type commands:
    1)Install Instant Client + sqlplus
    # rpm -ivh oracle-instantclient-basic-10.2.0.4-1.i386.rpm
    # rpm -ivh oracle-instantclient-sqlplus-10.2.0.4-1.i386.rpm
    2)Setting environment variable
    # LD_LIBRARY_PATH=/usr/lib/oracle/10.2.0.4/client/lib:${LD_LIBRARY_PATH}; export LD_LIBRARY_PATH
    # SQLPATH=/usr/lib/oracle/10.2.0.4/client/lib:${SQLPATH}; export SQLPATH
    3)Testing connection to Redhat1(Oracle DB server) using sqlplus
    # sqlplus abc/[email protected]:1521/lab2
    Connected.
    5)Begin to install unixODBC
    # rpm -ivh oracle-instantclient-odbc-10.2.0.4-1.i386.rpm
    # rpm -ivh unixODBC-2.2.11-7.1.i386.rpm
    # rpm -ivh unixODBC-devel-2.2.11-7.1.i386.rpm
    # rpm -ivh unixODBC-kde-2.2.11-7.1.i386.rpm
    6)After installation above packages I check for existing of odbc.ini and odbcinst.ini in /etc --> ok
    # ls /etc | grep odbc
    odbc.ini
    odbcinst.ini
    7) Tar unixODBC-2.2.14-linux-x86-32.tar.gz and copy file to /usr/local/bin, /usr/local/include, /usr/local/lib
    # tar -xzf unixODBC-2.2.14-linux-x86-32.tar.gz
    # ls /usr/local/bin/
    dltest isql iusql odbc_config odbcinst
    # ls /usr/local/include
    autotest.h odbcinstext.h odbcinst.h sqlext.h sql.h sqltypes.h sqlucode.h unixodbc_conf.h uodbc_extras.h uodbc_stats.h
    # ls /usr/local/lib
    libboundparam.la libgtrtst.la libodbccr.la libodbcinst.la libodbc.la
    libboundparam.so libgtrtst.so libodbccr.so libodbcinst.so libodbc.so
    libboundparam.so.1 libgtrtst.so.1 libodbccr.so.1 libodbcinst.so.1 libodbc.so.1
    libboundparam.so.1.0.0 libgtrtst.so.1.0.0 libodbccr.so.1.0.0 libodbcinst.so.1.0.0 libodbc.so.1.0.0
    # NLS_LANG=AMERICAN_AMERICA.UTF8; export NLS_LANG
    # TNS_ADMIN=/root; export TNS_ADMIN
    # ODBCINI=/etc/odbc.ini; export ODBCINI
    7)Retest sqlplus to Redhad1(Oracle database server) with local tnsname
    # sqlplus vonphoto/vonphoto@lab2
    SQL*Plus: Release 10.2.0.4.0 - Production on Sat Mar 27 03:20:55 2010
    Copyright (c) 1982, 2007, Oracle. All Rights Reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Production
    With the Partitioning, OLAP and Data Mining options
    8) Edit odbc.ini and odbcinst.ini
    ---odbcinst.ini
    # Included in the unixODBC package
    [Oracle 10g ODBC driver]
    Description = Oracle ODBC driver for Oracle 10g
    Driver = /usr/lib/oracle/10.2.0.4/client/lib/libsqora.so.10.1
    Setup =
    FileUsage =
    CPTimeout =
    CPReuse =
    ---odbc.ini
    [ora_dns]
    Driver = Oracle 10g ODBC driver
    DSN = ora_dns
    ServerName = xxx.xxx.xxx.121
    UserID = vonphoto
    Application Attributes = T
    Attributes = W
    BatchAutocommitMode = IfAllSuccessful
    BindAsFLOAT = F
    CloseCursor = F
    DisableDPM = F
    DisableMTS = T
    EXECSchemaOpt =
    EXECSyntax = T
    Failover = T
    FailoverDelay = 10
    FailoverRetryCount = 10
    FetchBufferSize = 64000
    ForceWCHAR = F
    Lobs = T
    Longs = T
    MetadataIdDefault = F
    QueryTimeout = T
    ResultSets = T
    SQLGetData extensions = F
    Translation DLL =
    Translation Option = 0
    DisableRULEHint = T
    TraceFile = /backup/sql.log
    Trace = Yes
    9)Testing the driver and the data source
    # odbcinst -q -d
    [Oracle 10g ODBC driver]
    # odbcinst -q -s
    [ora_dns]
    10) Test connection via ODBC from Redhat2(xxx.xxx.xxx.123) to Redhat1(xxx.xxx.xxx.121)
    # isql -v ora_dns
    isql: error while loading shared libraries: libreadline.so.3: cannot open shared object file: No such file or directory
    I have spent 4 days to search on the internet. But it still error.
    Pls help me.

    Apologies - thought it was a database to database connection that you were trying to create.
    isql: error while loading shared libraries: libreadline.so.3: cannot open shared object file: No such file or directoryThe readline library is in package readline-4.3-13 on RHEL4 and in readline-5.1-1.1 for RHEL5 on my servers (minor version number are of course dependent on the kernel version used).
    This is also not an ODBC or Oracle issue. It seems that isql is using rlwrap or similar - this requires the readline library. And possibly LD_LIBRARY_PATH to be set too.

  • Upgraded to 64 bit Oracle DataAccess.dll and now more open cursors errors

    Hi, We recently converted from .net framework 1.1 application to .net 4.0 framework for 64 bit. We installed the 64 bit version of Oracle DataAccess dll. It now seems like we are getting many more open cursor error messages in our application. The open_cursor value set in the database has not changed and other than the 64 bit Oracle and .net 4.0 framework nothing has changed. I was just curious if anyone else that has converted to the 64 bit Oracle DataAccess dll has experienced an increased number of open cursor errors? If so, what was the resolution.
    I apprecieate any feedback.
    Thanks

    What do you have /32BIT flag set to in the assembly? Ie, was it compiled with AnyCPU, x86, etc? You can use corflags.exe with the .net sdk to find out.
    Prior to VS2010, the default was AnyCPU which meant it ran as 64 bit on 64 bit OS, and 32 bit on 32 bit OS.
    As of VS2010, the IDE defaults to x86, which means it runs as 32 bit no matter what.
    32 bit apps need 32 bit odp/client. 64 bit apps need 64 bit odp/client.
    The error you're getting usually means you have the wrong bits (32/64) of the Oracle client installed.
    Assuming you have the wrong client bits installed, you could either
    a) make the app run as the other bits by setting/clearing the 32BIT flag (also by using corflags.exe)
    b) install the other bits of Oracle client software. a 32 bit home and a 64 bit home play rather nicely together for the most part, but you need to install them into separate homes.
    Hope it helps,
    Greg

  • Keynote and iPhone Remote throwing an error

    They remote works great when I open keynote and then use the remote app. But once I stop the slide show, and try and start the slide show again I get an error. When that happens it won't even let me start the slide show manually. I have to shut keynote down and restart it.
    Any ideas?

    Hi,
    Please follow the following procedure.
    Goto Transaction SICF->SAP->BC->SAP->Application name(Right click and activate the application)
    Hope this solves the issue, let me know if any queries.
    Regards,
    Rajani

  • Workflow manager IIS site throws a 403 Forbidden error and published workflows throw an error 400 bad request

    Hello All , Request your help on the below issue ,
    We have a single server farm SP 2013 where workflow was working as expected but its broken now , Not sure
    what caused this. Workflow manager is installed alongside Service Bus. Workflow Manager client is installed on the same SP server.
    We are able to create and publish the workflows from SP Designer without errors but when we start the workflow from SharePoint list we get an error " Sorry something went wrong". We have tried creating new workflows but the same problem.
    We receive the below error in event viewer.
    Error processing management request. Method: PUT, RequestUri: http://server URL:12291/$Workflows/89c767d2-82d2-4bd9-8e5e-dcace81d844f, Error: System.ArgumentException: The remote server returned an error: (400) Bad Request. ---> System.Net.WebException:
    The remote server returned an error: (400) Bad Request.
       at System.Net.HttpWebRequest.EndGetResponse(IAsyncResult asyncResult)
       at Microsoft.ServiceBus.Messaging.ServiceBusResourceOperations.GetAsyncResult`1.EndGetResponse(GetAsyncResult`1 thisPtr, IAsyncResult ar)
       at Microsoft.ServiceBus.Messaging.IteratorAsyncResult`1.StepCallback(IAsyncResult result)
       --- End of inner exception stack trace ---
    Server stack trace: 
    Exception rethrown at [0]: 
       at Microsoft.Workflow.Common.AsyncResult.End[TAsyncResult](IAsyncResult result)
       at Microsoft.Workflow.Gateway.HttpAsyncResult.End[TAsyncResult](IAsyncResult result)
       at Microsoft.Workflow.Gateway.WorkflowServiceGateway.<>c__DisplayClass6a.<EndCreateOrUpdateWorkflow>b__69()
       at Microsoft.Workflow.Gateway.WorkflowServiceGateway.ExceptionFilter[T](Func`1 body)
       at System.Threading.Tasks.TaskFactory`1.FromAsyncCoreLogic(IAsyncResult iar, Func`2 endFunction, Action`1 endAction, Task`1 promise, Boolean requiresSynchronization)
    ULS logs give below error 
    Error 1 - Microsoft.Workflow.Client.InternalServerException: An internal error occured. For more details, please see the server logs. HTTP headers received from the server - ActivityId: ba6e77cb-aaae-41a0-9cc8-1ea817d0045e. NodeId: Servername.
    Scope: /SharePoint/default/bc694777-7bbb-473e-ab2d-20087221a97a/5229793b-31a1-44b1-9b0b-374f38b762fd. Client ActivityId : a6daec9c-7139-7034-b4ee-edeb1f96ec2a. ---> System.Net.WebException: The remote server returned an error: (500) Internal Server Error.
     at Microsoft.Workflow.Common.AsyncResult.End[TAsyncResult](IAsyncResult result)    
     at Microsoft.Workflow.Client.HttpGetResponseAsyncResult`1.End(IAsyncResult result)    
     at Microsoft.Workflow.Client.ClientHelpers.SendRequest[T](HttpWebRequest request, T content)     -
     -- End of inner exception stack trace ---    
     at Microsoft.Workflow.Client.ClientHelpers.SendRequest[T](HttpWebRequest request, T content)    
     at Microsoft.Workflow.Client.WorkflowManagementClient.SendRequest[T](HttpWebRequest request, T content)    
     at Microsoft.Workflow.Client.WorkflowManager.Publish(WorkflowDescription description, Boolean terminateActiveInstances)    
     at Microsoft.SharePoint.WorkflowServices.FabricWorkflowManagementClient.PublishWorkflow(String serviceGroupName, WorkflowDescription workflow)    
     at Microsoft.SharePoint.WorkflowServices.WorkflowProxy.CreateDefinitionSubscription(String subscriptionName, String eventSource, List`1 eventTypes, String definitionName, IDictionary`2 metadata, WorkflowEventContext eventContext) StackTrace: 
     at Microsoft.Office.Server.Native.dll: (sig=35339f9c-8871-4ef1-9118-9c1bd0cb6598|2|microsoft.office.server.native.pdb, offset=131CE)
     at Microsoft.Office.Server.Native.dll: (offset=21B4D)
    Error 2 - Error publishing workflow subscription (republish or retry publish) information: Microsoft.Workflow.Client.InternalServerException: An internal error occured. For more details, please see the server logs. HTTP headers received
    from the server - ActivityId: ba6e77cb-aaae-41a0-9cc8-1ea817d0045e. NodeId:
    ServerName. Scope: /SharePoint/default/bc694777-7bbb-473e-ab2d-20087221a97a/5229793b-31a1-44b1-9b0b-374f38b762fd. Client ActivityId : a6daec9c-7139-7034-b4ee-edeb1f96ec2a. ---> System.Net.WebException: The remote server returned an error: (500)
    Internal Server Error.    
     at Microsoft.Workflow.Common.AsyncResult.End[TAsyncResult](IAsyncResult result)    
     at Microsoft.Workflow.Client.HttpGetResponseAsyncResult`1.End(IAsyncResult result)    
     at Microsoft.Workflow.Client.ClientHelpers.SendRequest[T](HttpWebRequest request, T content)     -
     -- End of inner exception stack trace ---    
     at Microsoft.Workflow.Client.ClientHelpers.SendRequest[T](HttpWebRequest request, T content)    
     at Microsoft.Workflow.Client.WorkflowManagementClient.SendRequest[T](HttpWebRequest request, T content)    
     at Microsoft.Workflow.Client.WorkflowManager.Publish(WorkflowDescription description, Boolean terminateActiveInstances)    
     at Microsoft.SharePoint.WorkflowServices.FabricWorkflowManagementClient.PublishWorkflow(String serviceGroupName, WorkflowDescription workflow)    
     at Microsoft.SharePoint.WorkflowServices.WorkflowProxy.CreateDefinitionSubscription(String subscriptionName, String eventSource, List`1 eventTypes, String definitionName, IDictionary`2 metadata, WorkflowEventContext eventContext)    
     at Microsoft.SharePoint.WorkflowServices.FabricSubscriptionService.<>c__DisplayClass1.<CreateDefinitionSubscription>b__0()    
     at Microsoft.SharePoint.WorkflowServices.WorkflowServiceContextExtensions.InvokeWithEcosystemRetry(WorkflowServicesContext context, EcosystemRequiredMethod method)    
     at Microsoft.SharePoint.WorkflowServices.FabricSubscriptionService.CreateDefinitionSubscription(String subscriptionName, Guid eventSourceId, String eventSource, List`1 eventNames, String definitionName, IDictionary`2 metadata, WorkflowEventContext eventContext)
     at Microsoft.SharePoint.WorkflowServices.WorkflowSubscriptionStorageEventReceiver.ItemUpdating(SPItemEventProperties properties)
    Error3 - Microsoft.Workflow.Client.InternalServerException: An internal error occured. For more details, please see the server logs. HTTP headers received from the server - ActivityId: ba6e77cb-aaae-41a0-9cc8-1ea817d0045e. NodeId: ServerName.
    Scope: /SharePoint/default/bc694777-7bbb-473e-ab2d-20087221a97a/5229793b-31a1-44b1-9b0b-374f38b762fd. Client ActivityId : a6daec9c-7139-7034-b4ee-edeb1f96ec2a. ---> System.Net.WebException: The remote server returned an error: (500) Internal Server Error.
     at  SPRequest.AddOrUpdateItem: UserPrincipalName=i:0).w|s-1-5-21-182846760-3036401646-508972453-351846, AppPrincipalName= ,bstrUrl=http://sharepointsite ,bstrListName={48A21758-9123-48F0-9BDE-9EB632B07BC6} ,bAdd=False ,bSystemUpdate=False ,bPreserveItemVersion=False
    ,bPreserveItemUIVersion=False ,bUpdateNoVersion=False ,pbstrNewDocId=00000000-0000-0000-0000-000000000000 ,bHasNewDocId=False ,bstrVersion=16 ,bCheckOut=False ,bCheckin=False ,bUnRestrictedUpdateInProgress=True ,bMigration=False ,bPublish=False ,bstrFileName=<null>
    Any help appreciated. Thanks in advance.

    Hi,
    As I understand, you encountered the issue when you published a workflow on SharePoint 2013 workflow platform.
    I wonder does it work well before when using 2013 workflow platform? If it works before, did you install any updates or change the configuration to the workflow related settings?
    If this is the first time after you installed workflow manager 2013, then I’d recommend you try re-registering workflow service per the link below and post the result:
    http://technet.microsoft.com/en-us/library/jj663115(v=office.15).aspx
    http://technet.microsoft.com/en-us/library/jj658588(v=office.15).aspx
    From the message you accessed workflow host uri, please make sure the account to wfsetup and wfservice account are both in wfadmins group.
    http://blogs.msdn.com/b/briangre/archive/2013/02/20/least-privilege-configuration-for-windows-azure-workflow-with-sharepoint-2013.aspx
    Regards,
    Rebecca Tu
    TechNet Community Support

  • Live Migration failed using virtual HBA's and Guest Clustering

    Hi,
    We have a Guest Cluster Configuration on top of an Hyper-V Cluster. We are using Windows 2012 and Fiber Channel shared storage.
    The problem is regarding Live Migration. Some times when we move a virtual machine from node A to node B everything goes well but when we try to move back to node A Live Migration fails. What we can see is that when we move the VM from node A to B and Live
    Migration completes successfully the virtual ports remain active on node A, so when we try to move back from B to A Live Migration fails because the virtual ports are already there.
    This doesn't happen every time.
    We have checked the zoning between Host Cluster Hyper-V and the SAN, the mapping between physical HBA's and the vSAN's on the Hyper-V and everything is ok.
    Our doubt is, what is the best practice for zoning the vHBA on the VM's and our Fabric? We setup our zoning using an alias for the vHBA 1 and the two WWN (A and B) on the same object and an alias for the vHBA 2 and the correspondent WWN (A and B). Is it
    better to create an alias for vHBA 1 -> A (with WWN A) and other alias for vHBA 1 -> B (with WWN B)? 
    The guest cluster VM's have 98GB of RAM each. Could it be a time out issue when Live Migration happen's and the virtual ports remain active on the source node? When everything goes well, the VM moves from node A with vHBA WWN A to node B and stays there
    with vHBA WWN B. On the source node the virtual ports should be removed automatically when the Live Migration completes. And that is the issue... sometimes the virtual ports (WWN A) stay active on the source node and when we try to move back the VM Live Migration
    fails.
    I hope You may understand the issue.
    Regards,
    Carlos Monteiro.

    Hi ,
    Hope the following link may help.
    To support live migration of virtual machines across Hyper-V hosts while maintaining Fibre Channel connectivity, two WWNs are configured for each virtual Fibre Channel adapter: Set A and Set B. Hyper-V automatically alternates between the Set A and Set B
    WWN addresses during a live migration. This ensures that all LUNs are available on the destination host before the migration and that no downtime occurs during the migration.
    Hyper-V Virtual Fibre Channel Overview
    http://technet.microsoft.com/en-us/library/hh831413.aspx
    More information:
    Hyper-V Virtual Fibre Channel Troubleshooting Guide
    http://social.technet.microsoft.com/wiki/contents/articles/18698.hyper-v-virtual-fibre-channel-troubleshooting-guide.aspx
    Hyper-V Virtual Fibre Channel Design Guide
    http://blogs.technet.com/b/privatecloud/archive/2013/07/23/hyper-v-virtual-fibre-channel-design-guide.aspx
    Hyper-V virtual SAN
    http://salworx.blogspot.co.uk/
    Thanks.
    Please remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading the thread.

Maybe you are looking for

  • I tried to update iTunes and it wouldn't let me install it

    iTunes wanted me to update it (it was working fine before) and I downloaded it and it gave me an error message and wouldn't let me start iTunes up. It said I could retry so I redownloaded iTunes from apple.com and tried over and over again but it did

  • Checking ABAP Engine

    Hello, I want to know how to find whether ABAP engine is working in SAP systems or not. Kindly guide me. Advace Thanks.

  • Maximum number of accounts allowed in Mail?

    Here's a unique one for you: Is there a maximum number of email accounts you can set Mail up to receive? I'm involved in three businesses, which means that between my personal accounts and professional ones, there are 25-30 different email accounts t

  • Oracle Forms 11g / frmall.jar or frm90all.jar installation

    I need to import Oracle.Forms.properties.ID in one of my JDeveloper application. I am using JDeveloper 10.1.3.5. I need frmall.jar/frm90all.jar to make the above import statement work in JDeveloper if i am not mistaking.Basically i need to add either

  • [Update] No signal from keystroke

    Hello Guys All of sudden I can't raise or lower my screen brightness via the media key anymore. Sometimes it works, but most times it wont. When I ran xev and pressed the corresponding keys(Thinkpad x230t so it's fn+f8 which produces the XF86MonBrigh