ACS Cluster with non-identical equipments

Hi,
On ACS, is it possible to realize a cluster A/A or A/S with two  non-identical equipments : a physical appliance and virtual ACS ?
Best regards

I have that design working. the virtual is primary and the appliance is secondary. primary instance where you do the config and Logs, secondary do the hard work of authenticating.
hope this help
Antero

Similar Messages

  • Server pool with non-identical HW/CPU

    Hi,
    I know that the installation manual requests, that all VM servers in one pool are based on the same hardware. Now, will OVM refuse to add new VM servers that are based on a different hardware, especially with different CPUs? I do have E5506 and E5620/30 CPUs in my various hosts.
    Precisely: the new OVM 3.0.1 server pool consists of hosts that are E5506 equipped, where as the old OVM 2.2.3 server pool consists of hosts that are E5620/30 equipped.
    The reason I am asking is, that I have setup a new OVM 3.0.1 cluster and I have already begun to import my OVM templates as well as a couple of guests to the new cluster. Alas, some of my OVM 2.2. guests require more RAM than my current VM servers offer, so I wanted to setup one of my OVM 2.2 VM servers to join the new server pool, so I can reduce the downtimes as much as possible.
    Cheers

    I have just tried this and it doesn't seem to work as I expected. I added another VM server to my server pool that has a E5630 CPU instead of a 5502 one and I neither can I migrate a guest from the Unassignes Virtual Machines folder, nor can I deploy a template to the new VM server, it either makes it to one of the first two VM servers.
    Edited by: budachst on Sep 9, 2011 12:50 PM
    Seems that my 2nd LUN where the storage repo resides doesn't get mounted on my newly added VM server, which explains nicely why no gues will make it onto the new VM server. I have verified that the LUN is visible on the new host, but adding the new server to the server pool only mounts the poolsfsmnt volume while not mounting the storage repo.
    What can I do about that?

  • ACS 5.3 Authorization problem with using Identity Groups in Access Policy Rule

    Hello guys, I am found a problem which I can't solve regarding authorization with using Identity Groups in Access Policy rule.
    ACS version: 5.3.0.40.6 (internal build B.839)
    I have very simple RADIUS Authorization rule which authorize user on behalf of right Identity Group.
    Requested Identity Group exist
    Testing user is created in Internal Users and has assigned requested Identity Group
    Radius Access Policy: 
    Authentication against Identity Store Sequence, where authorization server is external RSA SecurID device and additional attributes retrieval is configured from Internal Users.
    Authorization is very simple – One Rule with only one Condition which is: Identity Group - in - Requested_Testing_Rule. Then Default rule is set to Deny.
    When I will try login with my testing user then authentication against RSA SecurID is OK, but authorization will be denied by Default rule – It looks like my Rule with Identity Group is totally omitted.
    I am managing several other ACS servers (version 5.3 but with older patches) where similar rules are working without problem.
    What I am tested:
    Remove testing user and create his account again.
    Rename Identity Group
    Use another Identity Group
    Remove Access Policy rule and create it again
    Use Compound Condition: System:Identity Group
    Use Compound Condition: System:UserID instead of Identity Group in Rule (it is working without problem)
    Do you have any idea where problem can be?

    OK guys, it started working yesterday without any configuration change. Maybe it was some database inconsistence wich was solved by ACS itself.

  • Identity SSO API with non-web based appilcations

    hi,
    i can appreciate hwo this works with cookies etc for web based applications that use the api or one of the agents on apache etc.
    but how does it go with non web based java and say windows applications?
    can anyone point me to some docs?
    thanks

    I don't work for Sun but here are my thoughts
    1. Yes, if you don't want to use the AM SDK then the
    XML auth service is the way to go.
    2 & 3. dunno
    4. I think if you pass around the SSOToken ID
    generated by AM then any application can issue a SAML
    query to see if the session is still valid and get
    identity/auth attributes back
    5. I think SAML 2.0 supports authentication and
    single signoff
    6. If you are doing a lot of thick client apps you
    should use kerberos instead of AM web based
    Authentication. AM supports kerberos authentication,
    most modern browsers support SPEGNO for passing
    credentials to web server, AD supports Kerberos, and
    even Solaris 10 comes with a free KDC built into the
    OS. There is plenty of documentation around kerberos
    and the JDK out of the box supports GSS-API for
    Kerberos authenticationThank you for your feedback.
    We looked at the use of kerberos as well, but this is not really an option as we are dealing with fat clients installed on desktops of clients. So these desktops do not fall within our span control (multiple domain controllers etc.).
    Regards,
    Thomas

  • Sun cluster 3.20, live upgrade with non-global zones

    I have a two node cluster with 4 HA-container resource groups holding 4 non-global zones running Sol 10 8/07 u4 which I would upgrade to sol10 u6 10/8. The root fileystem of the non-global zones is ZFS and on shared SAN disks so that can be failed over.
    For the LIve upgrade I need to convert the root ZFS to UFS which should be straight forward.
    The tricky stuff is going to be performing a live upgrade on non-global zones as their root fs is on the shared disk. I have a free internal disk on each of thenodes for ABE environments. But when I run the lucreate command is it going put the ABE of the zones on the internal disk as well or can i specifiy the location ABE for non-global zones. Ideally I want this to be shared disk
    Any assistance gratefully received

    Hi,
    I am not sure whether this document:
    http://wikis.sun.com/display/BluePrints/Maintaining+Solaris+with+Live+Upgrade+and+Update+On+Attach
    has been on the list of docs you found already.
    If you click on the download link, it won't work. But if you use the Tools icon on the upper right hand corner and click on attachements, you'll find the document. Its content is solely based on configurations with ZFS as root and zone root, but should have valuable information for other deployments as well.
    Regards
    Hartmut

  • Share the cluster with a non-SAP application

    Hi,
    We need to  install the SAP Systems on existing Oracle Cluster Environment (10g).
    The scenario is there is an existing Oracle cluster which the client is using for different applications,we need to use the same existing Oracle cluster (RAC) to setup the SAP Systems.
    So what we would like to do is to share an already existing non-SAP application on RAC with a new ( to be installed) SAP System.
    Can someone guide me what are the pre-requisites and process for this kind of installation? The SAP Installation is on Linux environment.
    Thanks and regards,
    Silvio

    Hi James
    have a look on these links for <b>Integration with Non-SAP Legacy Systems </b>
    <b>
    How To…Connect non-SAP RFCPrograms with SAP Netweaver  Exchange Infrastructure</b>
    Refer this PDF....
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/d0a8949b-098d-2910-2b8e-b5e9b7daeb26
    <b>Integration With Microsoft</b>
    http://help.sap.com/saphelp_nw04s/helpdata/en/75/b8a742d2efc611e10000000a1550b0/frameset.htm
    SAP XI 3.0 help says:
    You can use the IDoc adapter to connect SAP systems (as of release 3.1) and <b>non-SAP systems with the Integration Server using native IDoc structures</b>
    <b>Integration SAP EM 4.0 and SAP ICH 4.0 via SAP XI </b> Pdf
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/50acad90-0201-0010-43ba-f8fc18ebb6ba
    Integration of SAP CRM with non-SAP-Systems
    Using XI, you can make use of the JDBC receiver adapter at the non-sap system since i believe its a database.
    http://help.sap.com/saphelp_nw04/helpdata/en/64/ce4e886334ec4ea7c2712e11cc567c/content.htm
    <b>Please go through the following Blog</b>
    /people/swaroopa.vishwanath/blog/2007/01/22/ale-configuration-for-pushing-idocs-from-sap-to-xi
    Thanks !!
    Questions are welcome here!!
    Also mark helpful answers
    Thanks,
    Abhishek Agrahari

  • Migrate data from a single node cluster to a single node cluster with CDOT 8.3

    Hi, I have two FAS2240-4 in the office, one for production and the second one intended to use it for DR, the FAS with 18 disk of 2TB each that is in production has 3 disk for the root aggregate and the rest for data, so I need to wipe the system in order to take advantage of the HDD partitioning for the entry platforms in CDOT 8.3. The Second system is fresh installed with CDOT 8.3 with the Root-Data HDD partitioning so now I want to move the Data and configuration to replace the first cluster with the second one.Some body gave me a procedure to backup the configuration and restore it at the destination cluster but I believe that is not complete. After finish the procedure I have all duplicated and not working properly, even with missed data, so the procedure is: ---------------------------------------------------
    SYSTEM CONFIGRUATION BACKUPS FOR CLUSTERED ONTAPSOURCE SYSTEM
    set -priv diagnostic
    cluster identiy show
    system configuration backup create -backup-name myBackup.7z -node node1 -backup-type cluster (Create a node backup of node1)
    system configuration backup upload -node node1 -backup myBackup.7z -destination ftp://[email protected] (Uploads a backup file to ftp)DESTINATION SYSTEM
    ## This would be done after setting up one node on the dr site, but prior to confguring the rest of the nodes
    ## you would basically stand up the system with ips and a base config, and load the config from the primary site
    ## and once it comes back up, you would setup & join the other nodes.
    set -priv diagnostic
    cluster identity show
    system configuration backup download -node node1 -backup myBackup.7z -destination ftp://[email protected]
    system configuration recovery cluster recreate -from backup -backup myBackup.7z
    system node reboot -node node1
    cluster identity show
    cluster identity modify -name newcluster
    cluster identity show
    system configutration recovery cluster modify -recovery-status complete----------------------------------------------------After doing this then I have problems with the folowing:
    - Volumes without properties
    - Network interfaces down and mixed with the old and new system
    - Aggregates with no info (This table is currently empty)- The node name is not able to rename it because it says that is already in use. I believe that I need more information in order to complete the procedure well.  I hope that you can help me with this. Best Regards, Miguel Valls

    In Non-Exa data, I followed below steps, I am sure same will work in exadata.
    Standby database creation using L0 backup restore
         While RMAN backup command on non-asm doesn't require change but restore in ASM requires you to add following kind of lines in restore script
              set newname for datafile 10 to '+Data02';
              set newname for tempfile 25 to '+Data10';
          After this you need to add
               restore database;
               switch datafile all;
               switch tempfile all;
    Standby database sync up using SCN based incremental backup
    Open Standby database in READ WRITE mode
    Convert DB from standalone to RAC
    Application doesn't need to be certified for Exadata, If your application runs on Oracle 11203 on RAC+ASM on OEL then it will run on Exadata as well.
    One important thing to remember is that, using physical migration method like this you won't be able to get EHCC (hybrid columnar compression) benefit unless you do the table move after the migration so most of the time logical methods like- data pump is preferred over physical method. In Exadata, these methods run very very fast. It might be less effort and less downtime compare to Physical standby method so give it a try.
    Thanks,
    Abhi

  • Help with seting up a Data Sorce can't be created with non-existent Pool

    I am wanting to use an Oracle 9i with WebLogic 7
    I have the following in my config.xml:
    <JDBCConnectionPool DriverName="oracle.jdbc.driver.OracleDriver"
    Name="Thin.Pool" Password="{3DES}C3xDZIWIABA="
    Properties="user=SYSTEM" TestTableName="OID"
    URL="jdbc:oracle:thin:@localhost:1521:DB_SID"/>
    <JDBCDataSource JNDIName="DB_DS" Name="DB_DS" PoolName="Thin.Pool"/>
    The console seems happy, no error mesages but in the log I get:
    ####<Mar 31, 2003 6:33:45 PM MST> <Info> <HTTP> <blue> <GameServe>
    <ExecuteThread: '1' for queue: '__weblogic_admin_html_queue'> <kernel
    identity> <> <101047>
    <[ServletContext(id=4110316,name=console,context-path=/console)]
    FileServlet: Using standard I/O>
    ####<Mar 31, 2003 6:35:37 PM MST> <Info> <JDBC> <blue> <GameServe>
    <ExecuteThread: '1' for queue: '__weblogic_admin_html_queue'> <kernel
    identity> <> <001082> <Creating Data Source named DB_DS for pool
    Thin.Pool>
    ####<Mar 31, 2003 6:35:37 PM MST> <Error> <JDBC> <blue> <GameServe>
    <ExecuteThread: '1' for queue: '__weblogic_admin_html_queue'> <kernel
    identity> <> <001059> <Error during Data Source creation:
    weblogic.common.ResourceException: DataSource(DB_DS) can't be created
    with non-existent Pool (connection or multi) (Thin.Pool)
         at weblogic.jdbc.common.internal.JdbcInfo.validateConnectionPool(JdbcInfo.java:127)
         at weblogic.jdbc.common.internal.JdbcInfo.startDataSource(JdbcInfo.java:260)
         at weblogic.jdbc.common.internal.JDBCService.addDeploymentx(JDBCService.java:293)
         at weblogic.jdbc.common.internal.JDBCService.addDeployment(JDBCService.java:270)
         at weblogic.management.mbeans.custom.DeploymentTarget.addDeployment(DeploymentTarget.java:375)
         at weblogic.management.mbeans.custom.DeploymentTarget.addDeployment(DeploymentTarget.java:154)
         at java.lang.reflect.Method.invoke(Native Method)
         at weblogic.management.internal.DynamicMBeanImpl.invokeLocally(DynamicMBeanImpl.java:732)
         at weblogic.management.internal.DynamicMBeanImpl.invoke(DynamicMBeanImpl.java:714)
         at weblogic.management.internal.ConfigurationMBeanImpl.invoke(ConfigurationMBeanImpl.java:417)
         at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1557)
         at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1525)
         at weblogic.management.internal.RemoteMBeanServerImpl.invoke(RemoteMBeanServerImpl.java:952)
         at weblogic.management.internal.ConfigurationMBeanImpl.updateConfigMBeans(ConfigurationMBeanImpl.java:578)
         at weblogic.management.internal.ConfigurationMBeanImpl.invoke(ConfigurationMBeanImpl.java:419)
         at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1557)
         at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1525)
         at weblogic.management.internal.RemoteMBeanServerImpl.invoke(RemoteMBeanServerImpl.java:952)
         at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:470)
         at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:198)
         at $Proxy16.addDeployment(Unknown Source)
         at weblogic.management.internal.DynamicMBeanImpl.unprotectedUpdateDeployments(DynamicMBeanImpl.java:1784)
         at weblogic.management.internal.DynamicMBeanImpl.access$0(DynamicMBeanImpl.java:1737)
         at weblogic.management.internal.DynamicMBeanImpl$1.run(DynamicMBeanImpl.java:1715)
         at weblogic.security.service.SecurityServiceManager.runAs(SecurityServiceManager.java:780)
         at weblogic.management.internal.DynamicMBeanImpl.updateDeployments(DynamicMBeanImpl.java:1711)
         at weblogic.management.internal.DynamicMBeanImpl.setAttribute(DynamicMBeanImpl.java:1035)
         at weblogic.management.internal.ConfigurationMBeanImpl.setAttribute(ConfigurationMBeanImpl.java:353)
         at com.sun.management.jmx.MBeanServerImpl.setAttribute(MBeanServerImpl.java:1358)
         at com.sun.management.jmx.MBeanServerImpl.setAttribute(MBeanServerImpl.java:1333)
         at weblogic.management.internal.RemoteMBeanServerImpl.setAttribute(RemoteMBeanServerImpl.java:898)
         at weblogic.management.internal.MBeanProxy.setAttribute(MBeanProxy.java:324)
         at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:193)
         at $Proxy13.setTargets(Unknown Source)
         at java.lang.reflect.Method.invoke(Native Method)
         at weblogic.management.console.info.FilteredMBeanAttribute.doSet(FilteredMBeanAttribute.java:92)
         at weblogic.management.console.actions.mbean.DoEditMBeanAction.perform(DoEditMBeanAction.java:145)
         at weblogic.management.console.actions.internal.ActionServlet.doAction(ActionServlet.java:171)
         at weblogic.management.console.actions.internal.ActionServlet.doPost(ActionServlet.java:85)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:760)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
         at weblogic.servlet.internal.ServletStubImpl$ServletInvocationAction.run(ServletStubImpl.java:1058)
         at weblogic.servlet.internal.ServletStubImpl.invokeServlet(ServletStubImpl.java:401)
         at weblogic.servlet.internal.ServletStubImpl.invokeServlet(ServletStubImpl.java:306)
         at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:5445)
         at weblogic.security.service.SecurityServiceManager.runAs(SecurityServiceManager.java:780)
         at weblogic.servlet.internal.WebAppServletContext.invokeServlet(WebAppServletContext.java:3105)
         at weblogic.servlet.internal.ServletRequestImpl.execute(ServletRequestImpl.java:2588)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:213)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:189)
    Why does it say:
    can't be created with non-existent Pool
    Thanks,

    Add "Targets" attribute to the connection pool. You may
    get an idea how it looks like by searching config.xml
    for "targets". If target servers are not set, the pool won't be
    deployed and can not be used to a create datasource.
    Regards,
    Slava Imeshev
    "BBaker" <[email protected]> wrote in message
    news:[email protected]...
    I am wanting to use an Oracle 9i with WebLogic 7
    I have the following in my config.xml:
    <JDBCConnectionPool DriverName="oracle.jdbc.driver.OracleDriver"
    Name="Thin.Pool" Password="{3DES}C3xDZIWIABA="
    Properties="user=SYSTEM" TestTableName="OID"
    URL="jdbc:oracle:thin:@localhost:1521:DB_SID"/>
    <JDBCDataSource JNDIName="DB_DS" Name="DB_DS" PoolName="Thin.Pool"/>
    The console seems happy, no error mesages but in the log I get:
    ####<Mar 31, 2003 6:33:45 PM MST> <Info> <HTTP> <blue> <GameServe>
    <ExecuteThread: '1' for queue: '__weblogic_admin_html_queue'> <kernel
    identity> <> <101047>
    <[ServletContext(id=4110316,name=console,context-path=/console)]
    FileServlet: Using standard I/O>
    ####<Mar 31, 2003 6:35:37 PM MST> <Info> <JDBC> <blue> <GameServe>
    <ExecuteThread: '1' for queue: '__weblogic_admin_html_queue'> <kernel
    identity> <> <001082> <Creating Data Source named DB_DS for pool
    Thin.Pool>
    ####<Mar 31, 2003 6:35:37 PM MST> <Error> <JDBC> <blue> <GameServe>
    <ExecuteThread: '1' for queue: '__weblogic_admin_html_queue'> <kernel
    identity> <> <001059> <Error during Data Source creation:
    weblogic.common.ResourceException: DataSource(DB_DS) can't be created
    with non-existent Pool (connection or multi) (Thin.Pool)
    atweblogic.jdbc.common.internal.JdbcInfo.validateConnectionPool(JdbcInfo.java:
    127)
    atweblogic.jdbc.common.internal.JdbcInfo.startDataSource(JdbcInfo.java:260)
    atweblogic.jdbc.common.internal.JDBCService.addDeploymentx(JDBCService.java:29
    3)
    atweblogic.jdbc.common.internal.JDBCService.addDeployment(JDBCService.java:270
    atweblogic.management.mbeans.custom.DeploymentTarget.addDeployment(DeploymentT
    arget.java:375)
    atweblogic.management.mbeans.custom.DeploymentTarget.addDeployment(DeploymentT
    arget.java:154)
    at java.lang.reflect.Method.invoke(Native Method)
    atweblogic.management.internal.DynamicMBeanImpl.invokeLocally(DynamicMBeanImpl
    .java:732)
    atweblogic.management.internal.DynamicMBeanImpl.invoke(DynamicMBeanImpl.java:7
    14)
    atweblogic.management.internal.ConfigurationMBeanImpl.invoke(ConfigurationMBea
    nImpl.java:417)
    atcom.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1557)
    atcom.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1525)
    atweblogic.management.internal.RemoteMBeanServerImpl.invoke(RemoteMBeanServerI
    mpl.java:952)
    atweblogic.management.internal.ConfigurationMBeanImpl.updateConfigMBeans(Confi
    gurationMBeanImpl.java:578)
    atweblogic.management.internal.ConfigurationMBeanImpl.invoke(ConfigurationMBea
    nImpl.java:419)
    atcom.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1557)
    atcom.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1525)
    atweblogic.management.internal.RemoteMBeanServerImpl.invoke(RemoteMBeanServerI
    mpl.java:952)
    at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:470)
    at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:198)
    at $Proxy16.addDeployment(Unknown Source)
    atweblogic.management.internal.DynamicMBeanImpl.unprotectedUpdateDeployments(D
    ynamicMBeanImpl.java:1784)
    atweblogic.management.internal.DynamicMBeanImpl.access$0(DynamicMBeanImpl.java
    :1737)
    atweblogic.management.internal.DynamicMBeanImpl$1.run(DynamicMBeanImpl.java:17
    15)
    atweblogic.security.service.SecurityServiceManager.runAs(SecurityServiceManage
    r.java:780)
    atweblogic.management.internal.DynamicMBeanImpl.updateDeployments(DynamicMBean
    Impl.java:1711)
    atweblogic.management.internal.DynamicMBeanImpl.setAttribute(DynamicMBeanImpl.
    java:1035)
    atweblogic.management.internal.ConfigurationMBeanImpl.setAttribute(Configurati
    onMBeanImpl.java:353)
    atcom.sun.management.jmx.MBeanServerImpl.setAttribute(MBeanServerImpl.java:135
    8)
    atcom.sun.management.jmx.MBeanServerImpl.setAttribute(MBeanServerImpl.java:133
    3)
    atweblogic.management.internal.RemoteMBeanServerImpl.setAttribute(RemoteMBeanS
    erverImpl.java:898)
    atweblogic.management.internal.MBeanProxy.setAttribute(MBeanProxy.java:324)
    at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:193)
    at $Proxy13.setTargets(Unknown Source)
    at java.lang.reflect.Method.invoke(Native Method)
    atweblogic.management.console.info.FilteredMBeanAttribute.doSet(FilteredMBeanA
    ttribute.java:92)
    atweblogic.management.console.actions.mbean.DoEditMBeanAction.perform(DoEditMB
    eanAction.java:145)
    atweblogic.management.console.actions.internal.ActionServlet.doAction(ActionSe
    rvlet.java:171)
    atweblogic.management.console.actions.internal.ActionServlet.doPost(ActionServ
    let.java:85)
    at javax.servlet.http.HttpServlet.service(HttpServlet.java:760)
    at javax.servlet.http.HttpServlet.service(HttpServlet.java:853)
    atweblogic.servlet.internal.ServletStubImpl$ServletInvocationAction.run(Servle
    tStubImpl.java:1058)
    atweblogic.servlet.internal.ServletStubImpl.invokeServlet(ServletStubImpl.java
    :401)
    atweblogic.servlet.internal.ServletStubImpl.invokeServlet(ServletStubImpl.java
    :306)
    atweblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(W
    ebAppServletContext.java:5445)
    atweblogic.security.service.SecurityServiceManager.runAs(SecurityServiceManage
    r.java:780)
    atweblogic.servlet.internal.WebAppServletContext.invokeServlet(WebAppServletCo
    ntext.java:3105)
    atweblogic.servlet.internal.ServletRequestImpl.execute(ServletRequestImpl.java
    :2588)
    at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:213)
    at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:189)
    Why does it say:
    can't be created with non-existent Pool
    Thanks,

  • [SOLVED] Can't add a node to the cluster with error (Exchange 2010 SP3 DAG Windows Server 2012)

    Hi there!
    I have a problem which makes me very angry already :)
    I have two servers Exchange 2010 SP3 with MB role started on Windows Server 2012. I decided to create a DAG.
    I have created the prestaged AD object for the cluster called msc-co-exc-01c, assigned necessary permissions and disabled it. Allowed through the Windows Firewall traffic between nodes and prepared the File Share Witness server.
    Then I have tried to add nodes. The first node has been added successfully, but the second node doesn't want to be added :). Now I can add only one node to the DAG. I tried to add different servers first, but only the first one was added.
    LOGS on the second nodes: 
    Application Log
    "Failed to initialize cluster with error 0x80004005." (MSExchangeIS)
    Failover Clustering Diagnostic Log
    "[VER] Could not read version data from database for node msc-co-exc-04v (id 1)."
    CMDLET Error:
    Summary: 1 item(s). 0 succeeded, 1 failed.
    Elapsed time: 00:06:21
    MSC-CO-EXC-02V
    Failed
    Error:
    A database availability group administrative operation failed. Error: The operation failed. CreateCluster errors may result from incorrectly configured static addresses. Error: An error occurred while attempting a cluster operation. Error: Cluster API '"AddClusterNode()
    (MaxPercentage=100) failed with 0x5b4. Error: This operation returned because the timeout period expired"' failed. [Server: msc-co-exc-04v.int.krls.ru]
    An Active Manager operation failed. Error An error occurred while attempting a cluster operation. Error: Cluster API '"AddClusterNode() (MaxPercentage=100) failed with 0x5b4. Error: This operation returned because the timeout period expired"' failed..
    This operation returned because the timeout period expired
    Click here for help... http://technet.microsoft.com/en-US/library/ms.exch.err.default(EXCHG.141).aspx?v=14.3.174.1&t=exchgf1&e=ms.exch.err.ExC9C315
    Warning:
    Network name 'msc-co-exc-01c' is not online. Please check that the IP address configuration for the database availability group is correct.
    Warning:
    The operation wasn't successful because an error was encountered. You may find more details in log file "C:\ExchangeSetupLogs\DagTasks\dagtask_2014-11-17_13-54-56.543_add-databaseavailabiltygroupserver.log".
    Exchange Management Shell command attempted:
    Add-DatabaseAvailabilityGroupServer -MailboxServer 'MSC-CO-EXC-02V' -Identity 'msc-co-exc-01c'
    Elapsed Time: 00:06:21
    UPD:
    when Exchange servers ran on the same Hyper-V node, the DAG is working well, but if I move one of VM to another node, It stops working.
    I have installed Wireshark and captured trafic of cluster interface. When DAG members on the same HV-node, there is inbound and outbound traffic on the cluster interface, but if I move one of DAG member to another node, in Wireshark I see only outbound traffic
    on both nodes.
    It's confused me, because there is normal connectivity between these DAG members through the main interface.
    Please, help me if you can.

    Hi, Jared! Thank you for the reply.
    Of course I did it already :) I have new info:
    when Exchange servers ran on the same Hyper-V node, the DAG is working well, but if I move one of VM to another node, It stops working.
    I have installed Wireshark and captured trafic of cluster interface. When DAG members on the same HV-node, there is inbound and outbound traffic on the cluster interface, but if I move one of DAG member to another node, in Wireshark I see only outbound traffic
    on both nodes.
    It's confused me, because there is normal connectivity between these DAG members through the main interface.

  • How to fix "Modifying a column with the 'Identity' pattern is not supported"

    When doing Code First Migrations my mobile service always errors in the seed method with: 'Modifying a column with the 'Identity' pattern is not supported. Column: 'CreatedAt'. Table: 'CodeFirstDatabaseSchema.Methodology' for the CreatedAt column. All my
    models inherit from EntityData. 
    // Sample model
    using System;
    using System.Collections.Generic;
    using System.ComponentModel.DataAnnotations;
    using Microsoft.WindowsAzure.Mobile.Service;
    namespace sbp_ctService.Models
    public class Methodology : EntityData
    public Methodology()
    this.Scenarioes = new List<Scenario>();
    public string Id { get; set; }
    [Required]
    [StringLength(50)]
    public string EntryMethod { get; set; }
    [Required]
    [StringLength(50)]
    public string TestDirection { get; set; }
    [Required]
    [StringLength(50)]
    public string PassCriteria { get; set; }
    [Required]
    [StringLength(50)]
    public string Dependency { get; set; }
    public bool ExtraInfo { get; set; }
    public virtual ICollection<Scenario> Scenarioes { get; set; }
    And in my Configuration.cs file during an update here's my seed method:
    protected override void Seed(sbp_ctService.Models.sbp_ctContext context)
    // This method will be called after migrating to the latest version.
    context.Methodologies.AddOrUpdate(
    m => m.Id,
    new Methodology { Id = "Methodology1", EntryMethod = "P/F", PassCriteria = "P/F", Dependency = "None", ExtraInfo = false, TestDirection = "Round" },
    new Methodology { Id = "Methodology2", EntryMethod = "P/F", PassCriteria = "Best", Dependency = "None", ExtraInfo = false, TestDirection = "Round" },
    new Methodology { Id = "Methodology3", EntryMethod = "P/F", PassCriteria = "Best", Dependency = "None", ExtraInfo = false, TestDirection = "In/Out" },
    new Methodology { Id = "Methodology4", EntryMethod = "P/F", PassCriteria = "Best", Dependency = "None", ExtraInfo = false, TestDirection = "Out" }
    For some reason on an update the CreatedAt field is created and given a value of null. So of course on an insert/update it will error because CreatedAt is an Identity field.
    I've tried to configure the modelBuilder in my context to tell it that CreatedAt is an identity field, but that still doesn't work.
    modelBuilder.Entity<Methodology>()
    .Property(m => m.CreatedAt)
    .HasDatabaseGeneratedOption(DatabaseGeneratedOption.Identity);
    So far the only way to fix this is by commenting out my Seed data, but it's not a fix. I've seen other solutions where you can force it to not serialize certain fields, but I don't know if those solutions apply.

    So I think this occurs because you might have created the database (Code-first) with POCOs that didn't have the CreatedAt field in them. I think that's what I did and the easiest way to fix it for me was to delete my database and re-create it with my POCOs
    inheriting from Entity Data from the very beginning. We were still in development so it worked out for us but I know some people might not be able to do that. Here's what my table looks like after it was created correctly:
    USE [database_name]
    GO
    /****** Object: Table [sbp_ct].[Methodologies] Script Date: 2/24/2015 9:48:45 PM ******/
    SET ANSI_NULLS ON
    GO
    SET QUOTED_IDENTIFIER ON
    GO
    CREATE TABLE [schema_name].[Methodologies] (
    [Id] NVARCHAR (128) NOT NULL,
    [EntryMethod] NVARCHAR (50) NOT NULL,
    [TestDirection] NVARCHAR (50) NOT NULL,
    [PassCriteria] NVARCHAR (50) NOT NULL,
    [Dependency] NVARCHAR (50) NOT NULL,
    [ExtraInfo] BIT NOT NULL,
    [Version] ROWVERSION NOT NULL,
    [CreatedAt] DATETIMEOFFSET (7) NULL,
    [UpdatedAt] DATETIMEOFFSET (7) NULL,
    [Deleted] BIT NOT NULL,
    [Name] NVARCHAR (MAX) NULL
    GO
    CREATE CLUSTERED INDEX [IX_CreatedAt]
    ON [schema_name].[Methodologies]([CreatedAt] ASC);
    GO
    ALTER TABLE [schema_name].[Methodologies]
    ADD CONSTRAINT [PK_schema_name.Methodologies] PRIMARY KEY NONCLUSTERED ([Id] ASC);
    Does yours look something like that?

  • Live upgrade - solaris 8/07 (U4) , with non-global zones and SC 3.2

    Dears,
    I need to use live upgrade for SC3.2 with non-global zones, solaris 10 U4 to Solaris 10 10/09 (latest release) and update the cluster to 3.2 U3.
    i dont know where to start, i've read lots of documents, but couldn't find one complete document to cover the whole process.
    i know that upgrade for solaris 10 with non-global zones is supported since my soalris 10 release, but i am not sure if its supported with SC.
    Appreciate your help

    Hi,
    I am not sure whether this document:
    http://wikis.sun.com/display/BluePrints/Maintaining+Solaris+with+Live+Upgrade+and+Update+On+Attach
    has been on the list of docs you found already.
    If you click on the download link, it won't work. But if you use the Tools icon on the upper right hand corner and click on attachements, you'll find the document. Its content is solely based on configurations with ZFS as root and zone root, but should have valuable information for other deployments as well.
    Regards
    Hartmut

  • Guest two node Failover Cluster with SAS HBA and Windows 2012 R1

    Hi all, i have two IBM x3560 brand new servers with V3700 IBM Storage. The Servers are connected to the storage through four SAS HBA adabters (two HBA's on each server). I want to create a two node guest Fileserver Failover Cluster. I can present the
    LUN's to the guest machines, but when i 'm running the cluster creation wizard it can't see any disk. I can see the disks on disk management console. Is there any way to achive this (the cluster creation) using my SAS HBA presented
    disks, or i have to use iSCSI to present the disks to my cluster?
    Thank you in advance, George
      

    Hi all, i have two IBM x3560 brand new servers with V3700 IBM Storage. The Servers are connected to the storage through four SAS HBA adabters (two HBA's on each server). I want to create a two node guest Fileserver Failover Cluster. I can present the
    LUN's to the guest machines, but when i 'm running the cluster creation wizard it can't see any disk. I can see the disks on disk management console. Is there any way to achive this (the cluster creation) using my SAS HBA presented
    disks, or i have to use iSCSI to present the disks to my cluster?
    Thank you in advance, George
    1) Update to R2 and use shared VHDX which is a better way to go. See:
    Shared VHDX
    http://blogs.technet.com/b/storageserver/archive/2013/11/25/shared-vhdx-files-my-favorite-new-feature-in-windows-server-2012-r2.aspx
    Clustering
    Options
    http://blogs.technet.com/b/josebda/archive/2013/07/31/windows-server-2012-r2-storage-step-by-step-with-storage-spaces-smb-scale-out-and-shared-vhdx-virtual.aspx
    2) If you want to stick with non-R2 (which is a BAD idea b/c tons of reasons) you can spawn an iSCSI target on top of your storage, make it clustered and make it provide LUs to your guest VMs. See:
    iSCSI Target in Failover
    http://technet.microsoft.com/en-us/library/gg232632(v=ws.10).aspx
    iSCSI Target Failover Step-by-Step
    http://techontip.wordpress.com/2011/05/03/microsoft-iscsi-target-cluster-building-walkthrough/
    3) Use third-party software providing clustered storage (active-active) out-of-box. 
    I would strongly recommend to upgrade to R2 and use shared VHDX.
    StarWind VSAN [Virtual SAN] clusters Hyper-V without SAS, Fibre Channel, SMB 3.0 or iSCSI, uses Ethernet to mirror internally mounted SATA disks between hosts.

  • Cluster with only 2 NICs, is it possible?

    I am trying to setup a non-production 2 node cluster, problem is I only have 2 NIC ports avail since I can't spend any money on it. Is it possible to work around that and get the second node to join the cluster with one public and one private interface? Not having any luck and am curious if anyone has done the same and knows of some sort of hack to get around it? When node 2 first boots and tries to join the cluster, it is unable to communicate over the cluster interconnect. I understand this isn't the desired setup and voids the whole HA scenario, but this is just a proof of concept before moving a new production app over to SC3.2, which will not have this limitation.
    Thanks in advance for any input.

    A little more research in the interconnect not working I've found two common causes that people have discovered, ipfilter and kernel patch 138888-01. I am not using ipfilter, the service is disabled, and as I understand it, kernel patch 138888-01 and above only cause this issue if connected via a switch. My cluster nodes are connected directly by a crossover cable. Is this still possibly my issue, or should I keep digging? One node is a T2000 with e1000g interfaces, so that also lines up.
    Thanks again..

  • Create XA MultiDataSource with Non-XA DataSource(support global txn)

    I have two datasources DataSource1 and DataSource2. Since my driver doesnt support XA, I am Emulating XA for these two data sources.
    Now, I want to create a Multi Data Source and add these two data sources. Is there anyway I can Emulate XA for this Multi Data Source?

    Thanks for your quick reply Joe.
    The target for this multipool is single managed server(not a cluster). wee have a lot of resources like MQ, AQ,ORacle etc in the transaction. This datasource is for Tandem jdbc. We dont have XA jdbc driver for this and thats the reason we are emulating the XA.
    We were using weblogic 9.2 and there we were able to setup XA Multipool with non xa pools (with XA emulated) without any issues. and now we are migrating to weblogic 9.2
    This is what i did in weblogic 9.2 console.
    1. Created DataSource1 and DataSource2. Both have GloablTrasactionSupport enabled with Emulate XA
    2. Created a MultiDataSource. During creation, I selected non-XA driver and from the data source page, I added DataSource1 and DataSource2. (If I select XA driver, then the datasource selection page wont show DataSource1 and DataSource2 since it is non-XA)
    The MultiDataSource is deployed without any issues. but when my application try to get a connection from MultiDataSource, it is throwing below exception.
    Sep 8, 2008 1:09:03 AM CDT> <Error> <EJB> <BEA-010080> <An error occurred while attempting to process a message inside a message-driven bean: weblogic.transaction.RollbackException: Could not prepare resource 'weblogic.jdbc.wrapper.JTSXAResourceImpl
    JDBC driver does not support XA, hence cannot be a participant in two-phase commit. To force this participation, set the GlobalTransactionsProtocol attribute to LoggingLastResource (recommended) or EmulateTwoPhaseCommit for the Data Source = Multi_DataSource_XA
    Exception is : weblogic.transaction.RollbackException: Could not prepare resource 'weblogic.jdbc.wrapper.JTSXAResourceImpl
    JDBC driver does not support XA, hence cannot be a participant in two-phase commit. To force this participation, set the GlobalTransactionsProtocol attribute to LoggingLastResource (recommended) or EmulateTwoPhaseCommit for the Data Source = Multi_DataSource_XA
         at weblogic.transaction.internal.TransactionImpl.throwRollbackException(TransactionImpl.java:1782)
         at weblogic.transaction.internal.ServerTransactionImpl.internalCommit(ServerTransactionImpl.java:331)
         at weblogic.transaction.internal.ServerTransactionImpl.commit(ServerTransactionImpl.java:227)
         at weblogic.ejb.container.internal.MDListener.execute(MDListener.java:463)
         at weblogic.ejb.container.internal.MDListener.transactionalOnMessage(MDListener.java:335)
         at weblogic.ejb.container.internal.JMSMessagePoller.processOneMessage(JMSMessagePoller.java:320)
         at weblogic.ejb.container.internal.JMSMessagePoller.pollContinuously(JMSMessagePoller.java:394)
         at weblogic.ejb.container.internal.JMSMessagePoller.pollForParent(JMSMessagePoller.java:517)
         at weblogic.ejb.container.internal.JMSMessagePoller.run(JMSMessagePoller.java:533)
         at java.lang.Thread.run()V(Unknown Source)
    >
    Jiji

  • ACS Integration with Microsoft Active Directory Services

    Hello Everyone,
    I've been tasked to design the integration of ACS with MS AD. What I want to know is the below assuming I have a software ACS or a ACS device and the protocol for authentication is Radius
    - What is the criteria for the AD to integrate with ACS software of appliance
    - Should that AD be hosted on the domain controller or not?
    - If not, on what (Domain Controller, Tree, Forest, Branch, Flower, Fruit  ) should the AD be hosted on?
    - What will I have to do to authenticate users logging into Cisco Security Manager with ACS integrated with AD?
    - Are there any other dependencies that I will have to categorically mention in my design document?
    Thanks,
    Rishi

    In ACS v5.x, there is a screen for integrating the ACS with AD. 
         (Users and Identity Stores > External Identity Stores > Active Directory)
    Just enter the local domain name (domain.com) and a valid AD administrator account username and password, and the ACS will connect to the domain.  This allows you to use existing AD credentials to login and administer your network devices. 
    Tying the ACS to AD really only takes one screen and less than a minute, but you will still have to tell the ACS which AD groups get which permissions (for example, read-only or read-write access), and you will have to setup a search sequence (Users and Identity Stores > Identity Store Sequences) to tell ACS to first look at AD for credentials, then check the local ACS user database for valid accounts.  The permissions part is still fairly quick, and it only takes me about 45 minutes to build an ACS from scratch including all AD integration and custom RADIUS attributes for some of our devices. 
    The authentication would occur like this:
    User SSH/telnet/console to device
    Device contacts ACS using TACACS or RADIUS
    User receives login prompt and enters AD credentials
    Devices sends credentials to ACS
    ACS validates credentials in AD
    ACS sends authentication OK message to Device
    Device logs user in.
    Command Authorization looks something like this:
    User enters a command
    Device sends command authorization request to ACS
    ACS looks at which AD group the user belongs to and looks up permissions configured in ACS for that group
    Based on the permissions you have assigned, ACS either sends an allow or deny message to the Device
    Device allows or denies the user command.
    Criteria:  We use an ACS 5.2 virtual machine and have had it work perfectly with Server 2003 and Server 2008.
    AD is hosted on our local domain controller (Bonus:  no planting of flowers required!)
    Dependencies: 
    Issue:  The Device looks to ACS.  ACS looks to AD.  If AD fails, users cannot use their AD credentials to login.
              Device ---> ACS ---> AD
    Solution:  Configure the Device to look at ACS first, then a local table if ACS is not available.  Also, configure the ACS to look at AD first, then a local ACS account list if AD is not available.  (You can configure local user accounts on the Device and in the ACS) 
              Device ---> ACS ---> AD
              Device ---> ACS ---> AD ---> ACS local
              Device ---> ACS ---> AD ---> ACS local ---> Device local
    The new version of Cisco ACS is UNIX-based, and you can download a free trial to load up and try before you buy.  It is far FAR superior to the old ACS v3.3 that we had for years.
    I hope this helps for your design document!
    --Chris

Maybe you are looking for

  • (CES2014)MSI gaming notebook photo gallery

    MSI Computer Corp, leading manufacturer of computer hardware products and solutions, unveils its 2014 lineup during CES in Las Vegas. On display at the MGM Skyline Marquee Suites will be MSI's next-generation GT, GE and GX gaming notebooks, Quadro wo

  • Itemrenderer in datagrid

    i've the following datagrid code: <mx:DataGrid  id="empdg"   dataProvider="{employeeService.lastResult.employees.employee}"   width="40%" editable="true" itemClick="addRow(event)"> <mx:columns> <mx:DataGridColumn headerText="Reportees" editable="fals

  • Playing Call of Duty II on my new MacBook Pro... Not Good!

    Hey Just bought a brand new MacBook Pro with ALL the trimmings - and Call of Duty II makes my mac run like I was playing on a Commodor 64. Does anybody had the same experience - and have you found patches or updates that fixes this problem? Thanks! M

  • Variant -Variables

    Hi, This is Urgent ..please I have a requirement like .. I know the VARIABLE name , i need to findout the VARIANT names which are using this VARIABLE. Ex: Variable name = LAST_WEEK . Please let me know how to findout the VARIANT names which is using

  • Ridiculously slow performance with WRT320N

    Hello all.  My name is Ben Herrmann.  I'm running a Windows 7 64 Bit machine and am using a Motorola cable modem.  I have the WRT320N wireless router attached.  My PC is connected via ethernet cable to the machine - but the other two laptops are conn