FrameMaker Server 8 in a cluster server

Does anyone have any experience with FrameMaker Server 8 (or previous versions) running in a cluster server?
Also, if anybody has implemented FrameMaker Server 8 running unattended in any kind of server --be it a High Performance, High Availability or any other configuration-- please feel free to comment.
Thanks
Juan Souza

Arnis,
First of all, thanks for replying.
I am aware that the only difference between FM and FM Server is the licensing. Adobe requires a Server license for any unattended use of FrameMaker.
The server environment, and specifically a clustered server, is a client requirement. They are especially sensitive about their High Availability setup. Since FrameMaker isn't "cluster-aware", we have had to establish control procedures to continue any process that has been interrupted when the active server in the cluster is either shut down or moved into passive mode.
We are producing PDF, RTF and printed output, processing XML streams. We use FrameMaker Server 8, driven by FrameScript scripts. Given your experience, do you see anything wrong with our setup? There is a budgetary issue which doesn't allow the Miramo product to be an option. Also, we do not build documents from database query/report procedures.
Again, thanks for your input.
Juan

Similar Messages

  • The cluster server element 441305751 is not added in the EP cluster context

    Hello Experts,
    We are using load balancer to distribute the load across of our four portal servers nodes. Lately, two of our nodes are behavior very strangely; it continuously stopped working momentarily and it started back up again automatically by it self. And the users usually received a message u201C500 internal server erroru201D.
    #1.5 #001372559B8800280000000200002D940004692BC78C988A#1241536659228#com.sap.engine.services.prtbridge##com.sap.engine.services.prtbridge####n/a##defe75c0398711debfb1001372559b88#SAPEngine_System_Thread[impl:5]_55##0#0#Error#1#/System/Server#Plain###
    The cluster server element 441305751 is not added in the EP cluster context,
       because prtbridge on 441305751 is stopped  or not yet started (maybe the server is starting or shuting down).
    Any helps would be appreciated,
    davidn

    HI David,
    Did you figure out why it happened? we have same exact problem and same log was written, could you please share your solution.
    Thanks

  • Solaris Cluster Server Homogeneous System Copy

    Hi,
    We have SAP ECC 6.0 on Solaris Sparc server. Our database is Oracle 10.2.0.2. We will buy 2 new solaris server and we want to use these server as cluster. We want to make homogeneous system copy from current server to new cluster server. How can we make homogeneous copy from current server to new cluster server? Firts of all do we have to install new ECC 6.0 on cluster server and then will we make homogeneous system copy?
    Also how can we install ECC 6.0 on Solaris Cluster server? Will we have install ECC 6.0 seperatly for each server of cluster unit? I have SAP installation document but it is not clear. Do you have another document for SAP installation on cluster server? Please help us about these issues.
    Best regards.

    First to move your ECC system from your current server to the new server, yes you do need to homogenous copy. There are other methods like storage sub-system level copy, but it is not supported. Since your system is ABAP, you can setup standby database on the new server and use dataguard to replicate database. You can manual move profiles, binaries and other application file systems manually since they are on external storage. You should change your profiles to use logical hostname instead of the physical hostname though Solaris 10 gives you the ability to move zone/containers with same hostname.
    I am assuming you have distributed system, ie database and SAP separate. You can setup RAC for database HA. For SAP, you can use sun cluster or veritas cluster to move SAP central instance with the logical hostname to the 2nd standby server.
    You can also separate ASCS from SAP central instance and setup failover for the CS instance with the cluster software. Check notes 821904, 870652. You can take another step in HA with ENQ replication (see http://help.sap.com/saphelp_nw70/helpdata/EN/36/67973c3f5aff39e10000000a114084/content.htm)
    -Regards

  • Install SharePoint 2013 database on SQL server 2012 SP1 cluster

    Good Day all,
    We are trying to install SharePoint 2013 database on a SQL server 2012 Enterprise with SP1 cluster (configured for mixed mode) but the connection is not working (times out). We can telent the SQL cluster network name on port 1433 and we made
    a UDL connection file, test the connection via windows authentication and it is working, but if we use the sql authentication it is not working.
    Also we have two SharePoint 2013 servers connected via NLB.
    Any idea about this issue guys?  

    event if 3355 means 3 things:
    The SQL Server might be offline.
    The SharePoint Foundation database access account might not have the necessary permissions to communicate with the SQL Server.
    A firewall that runs on either the local server or on SQL Server might be blocking network communications.
    Big question, how you trying to connect SQL  windows or sql authentication? is it throwing error on Config wizard or somewhere else?
    Please remember to mark your question as answered &Vote helpful,if this solves/helps your problem. ****************************************************************************************** Thanks -WS MCITP(SharePoint 2010, 2013) Blog: http://wscheema.com/blog

  • Sql developer won't connect with SQL server database on cluster

    Hello,
    I have a microsoft SQL server database on a windows cluster. Name for this cluster is like XXX/YYY
    When I try to connect to this database with SQL developer, I get a failure, mentioning:
    "Status : Failure -Cannot connect to Miscrosoft SQL Server on XXX/YYY"
    As far as I know, I have installed the correct jtds driver (version 1.2.5), since I can connect to sql server databases on single server hosts.
    Does anyone know how to connect to databases on windows clusters?
    Edited by: Ingmar on Jan 16, 2012 4:08 PM
    Edited by: Ingmar on Jan 16, 2012 4:22 PM

    Hi Ingmar,
    Try this is in the port box
    1433/dbname;instance=YYY
    dbname can be omitted
    -Turloch
    SQL Developer team
    1/Giving an instance name in SQLDeveloper connection to SQL Server (from August 2011)
    MS SQL Server Connection
    The correct syntax to deal with this is in the port box
    1433/dbname;instance=instance
    dbname can be omitted
    2/cluster configuration (uses instance name)
    http://stackoverflow.com/questions/1160024/jdbc-connect-string-for-sql-server-cluster
    where vvv\iii is the cluster address: jdbc address is:
    "jdbc:sqlserver://vvv;instanceName=iii"
    3/jtds url details available in
    http://jtds.sourceforge.net/faq.html

  • Error reading Web application occurs when starting one server in the cluster

    Hi All,
              I have configured two cluster servers in the win 2000. Admin server also win
              2000. One cluster server joins the cluster while other gives following error
              and starts. But it seems that it does not join to the cluster view because
              prevoius server only serves the requests. I have attached the log also.
              <> <101062> <[HTTP synergyserver] Error reading Web application
              "D:\bea\wlserver6.0\.\config\bd2kadmindomain\applications\.wl_temp_do_not_de
              lete_synergyserver\wl_local_comp12863.war">
              java.net.UnknownHostException: java.sun.com
              at java.net.InetAddress.getAllByName0(InetAddress.java:571)
              at java.net.InetAddress.getAllByName0(InetAddress.java:540)
              at java.net.InetAddress.getAllByName(InetAddress.java:533)
              at weblogic.net.http.HttpClient.openServer(HttpClient.java:159)
              at weblogic.net.http.HttpClient.openServer(HttpClient.java:221)
              at weblogic.net.http.HttpClient.<init>(HttpClient.java:85)
              at
              weblogic.net.http.HttpURLConnection.getHttpClient(HttpURLConnection.java:109
              at
              weblogic.net.http.HttpURLConnection.getInputStream(HttpURLConnection.java:30
              1)
              at java.net.URL.openStream(URL.java:798)
              at
              weblogic.apache.xerces.readers.DefaultReaderFactory.createReader(DefaultRead
              erFactory.java:149)
              at
              weblogic.apache.xerces.readers.DefaultEntityHandler.startReadingFromExternal
              Entity(DefaultEntityHandler.java:775)
              at
              weblogic.apache.xerces.readers.DefaultEntityHandler.startReadingFromExternal
              Subset(DefaultEntityHandler.java:570)
              at
              weblogic.apache.xerces.framework.XMLDTDScanner.scanDoctypeDecl(XMLDTDScanner
              .java:1131)
              at
              weblogic.apache.xerces.framework.XMLDocumentScanner.scanDoctypeDecl(XMLDocum
              entScanner.java:2177)
              at
              weblogic.apache.xerces.framework.XMLDocumentScanner.access$0(XMLDocumentScan
              ner.java:2133)
              at
              weblogic.apache.xerces.framework.XMLDocumentScanner$PrologDispatcher.dispatc
              h(XMLDocumentScanner.java:882)
              at
              weblogic.apache.xerces.framework.XMLDocumentScanner.parseSome(XMLDocumentSca
              nner.java:380)
              at weblogic.apache.xerces.framework.XMLParser.parse(XMLParser.java:900)
              at
              weblogic.apache.xerces.jaxp.DocumentBuilderImpl.parse(DocumentBuilderImpl.ja
              va:123)
              at
              weblogic.servlet.internal.dd.DescriptorLoader.<init>(DescriptorLoader.java:1
              78)
              at weblogic.servlet.internal.HttpServer.loadWARContext(HttpServer.java:446)
              at weblogic.servlet.internal.HttpServer.loadWebApp(HttpServer.java:404)
              at weblogic.j2ee.WebAppComponent.deploy(WebAppComponent.java:74)
              at weblogic.j2ee.Application.addComponent(Application.java:133)
              at weblogic.j2ee.J2EEService.addDeployment(J2EEService.java:115)
              at
              weblogic.management.mbeans.custom.DeploymentTarget.addDeployment(DeploymentT
              arget.java:327)
              at
              weblogic.management.mbeans.custom.DeploymentTarget.addDeployment(DeploymentT
              arget.java:143)
              at
              weblogic.management.mbeans.custom.WebServer.addWebDeployment(WebServer.java:
              76)
              at java.lang.reflect.Method.invoke(Native Method)
              at
              weblogic.management.internal.DynamicMBeanImpl.invokeLocally(DynamicMBeanImpl
              .java:562)
              at
              weblogic.management.internal.DynamicMBeanImpl.invoke(DynamicMBeanImpl.java:5
              48)
              at
              weblogic.management.internal.ConfigurationMBeanImpl.invoke(ConfigurationMBea
              nImpl.java:285)
              at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1555)
              at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1523)
              at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:439)
              at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:180)
              at $Proxy40.addWebDeployment(Unknown Source)
              at
              weblogic.management.configuration.WebServerMBean_CachingStub.addWebDeploymen
              t(WebServerMBean_CachingStub.java:1012)
              at
              weblogic.management.mbeans.custom.DeploymentTarget.addDeployment(DeploymentT
              arget.java:313)
              at
              weblogic.management.mbeans.custom.DeploymentTarget.addDeployments(Deployment
              Target.java:277)
              at
              weblogic.management.mbeans.custom.DeploymentTarget.updateServerDeployments(D
              eploymentTarget.java:232)
              at
              weblogic.management.mbeans.custom.DeploymentTarget.updateDeployments(Deploym
              entTarget.java:192)
              at java.lang.reflect.Method.invoke(Native Method)
              at
              weblogic.management.internal.DynamicMBeanImpl.invokeLocally(DynamicMBeanImpl
              .java:562)
              at
              weblogic.management.internal.DynamicMBeanImpl.invoke(DynamicMBeanImpl.java:5
              48)
              at
              weblogic.management.internal.ConfigurationMBeanImpl.invoke(ConfigurationMBea
              nImpl.java:285)
              at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1555)
              at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1523)
              at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:439)
              at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:180)
              at $Proxy0.updateDeployments(Unknown Source)
              at
              weblogic.management.configuration.ServerMBean_CachingStub.updateDeployments(
              ServerMBean_CachingStub.java:2299)
              at
              weblogic.management.mbeans.custom.ApplicationManager.startConfigManager(Appl
              icationManager.java:240)
              at
              weblogic.management.mbeans.custom.ApplicationManager.start(ApplicationManage
              r.java:122)
              at java.lang.reflect.Method.invoke(Native Method)
              at
              weblogic.management.internal.DynamicMBeanImpl.invokeLocally(DynamicMBeanImpl
              .java:562)
              at
              weblogic.management.internal.DynamicMBeanImpl.invoke(DynamicMBeanImpl.java:5
              48)
              at
              weblogic.management.internal.ConfigurationMBeanImpl.invoke(ConfigurationMBea
              nImpl.java:285)
              at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1555)
              at com.sun.management.jmx.MBeanServerImpl.invoke(MBeanServerImpl.java:1523)
              at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:439)
              at weblogic.management.internal.MBeanProxy.invoke(MBeanProxy.java:180)
              at $Proxy9.start(Unknown Source)
              at
              weblogic.management.configuration.ApplicationManagerMBean_CachingStub.start(
              ApplicationManagerMBean_CachingStub.java:435)
              at weblogic.management.Admin.startApplicationManager(Admin.java:1033)
              at weblogic.management.Admin.finish(Admin.java:493)
              at weblogic.t3.srvr.T3Srvr.start(T3Srvr.java:429)
              at weblogic.t3.srvr.T3Srvr.run(T3Srvr.java:170)
              at weblogic.Server.main(Server.java:35)
              Each cluster servers domain names are different i.e. not "mydomain". The
              file its complaining is in the specified directory and it has proper
              priviledges.
              If anyone has an idea please respond.
              Thanks
              Nalika
              [synergyserver.log]
              

    You're getting that probably because the WL instance was not shut down properly.
    If that's the case, you'll need to remove an ldap lock file (with a .lok extension) in the directory of ldap files (under the server dir) .

  • How to install & confgiure Oracle9i on Veritas Cluster Server

    Hi,
    I want to know how to install oracle9i on Veritas Cluster Server having OS as Solaris 9. Also How to configure oracle in VCS so that it can be failover.
    I really appreciate the replies.
    Thanks,
    Shrikant

    I want to install following component
    Oracle 9i form server & report server
    Oracle HTTP Server in SSL mode
    Oracle As Web Cache
    Regards
    Arvind

  • Failover cluster server - File Server role is clustered - Shadow copies do not seem to travel to other node when failing over

    Hi,
    New to 2012 and implementing a clustered environment for our File Services role.  Have got to a point where I have successfully configured the Shadow copy settings.
    Have a large (15tb) disk.  S:
    Have a VSS drive (volume shadow copy drive) V:
    Have successfully configured through Windows Explorer the Shadow copy settings.
    Created dependencies in Failcover Cluster Server console whereby S: depends on V:
    However, when I failover the resource and browse the Client Access Point share there are no entries under the "Previous Versions" tab. 
    When I visit the S: drive in windows explorer and open the Shadow copy dialogue box, there are entries showing the times and dates of the shadow copies ran when on the original node.  So the disk knows about the shadow copies that were ran on the
    original node but the "previous versions" tab has no entries to display.
    This is in a 2012 server (NOT R2 version).
    Can anyone explain what might be the reason?  Do I have an "issue" or is this by design?
    All help apprecieated!
    Kathy
    Kathleen Hayhurst Senior IT Support Analyst

    Hi,
    Please first check the requirements in following article:
    Using Shadow Copies of Shared Folders in a server cluster
    http://technet.microsoft.com/en-us/library/cc779378(v=ws.10).aspx
    Cluster-managed shadow copies can only be created in a single quorum device cluster on a disk with a Physical Disk resource. In a single node cluster or majority node set cluster without a shared cluster disk, shadow copies can only be created and managed
    locally.
    You cannot enable Shadow Copies of Shared Folders for the quorum resource, although you can enable Shadow Copies of Shared Folders for a File Share resource.
    The recurring scheduled task that generates volume shadow copies must run on the same node that currently owns the storage volume.
    The cluster resource that manages the scheduled task must be able to fail over with the Physical Disk resource that manages the storage volume.
    If you have any feedback on our support, please send to [email protected]

  • Which role do I need DFS or File server on fail over cluster server 2012 R2?

    what I want to achieve is that I want to share all my user data files in a central location and to be highly available all the time whether it's a general share or folder redirection data. BUT I'm a bit confused;  I have fail over cluster  set-up
    on server 2012, now I would like to add DFS as a role but than we have another role called File server and virtually it does the same thing as DFS? Means it creates a namespace share that can be access even one of the nodes goes down. Now I am thinking is
    that DFS does the replication between two physical location but fail over cluster works slightly differently  and with file server it pretty much does the same thing except for replicating data from one drive to another. Now what do you suggest I do or
    did I get the concept wrong like a noob?

    DFS and Failover Clustering for file shares provides a similar end result for file access, but they are significantly different implementations.
    Clustering provides high availability to files by presenting shared access to set a files served from a cluster.  With 2012 R2 Microsoft added the ability to create a Scale-out File Server that even allows all nodes of the cluster to server access to
    the files for a higher level of performance and other great things.  Bottom line with Failover Clusters for files is that there is a single copy of the file presented from the cluster.
    DFS on the other hand provides high availability to files by presenting multiple copies of the file by making a copy in two or more locations and presenting a naming space that allows access to the file through any of the network paths.  DFS works very
    well for files that are primarily read-only.  When you get into a situation where there is a lot of updating of the shared files, DFS is not a very good solution.  There are ways to implement DFS for read/write files, but it generally requires a
    good knowledge of how the files are used and how you want to manage them.
    The key to answering your question comes in your first sentence "I want to share all my user data files in a central location and to be highly available all the time".  My initial reaction to this is that central location means Failover Cluster
    - there is only a single copy of the file.  However, "all the time" can be compromised by network failures to the central site.  Remote sites would not have access if they can't access the central site.  DFS provides the ability to
    have copies remotely, but then if you allow updating at multiple sites, you have to manage the merging of the changes, among other things.
    . : | : . : | : . tim

  • Server 2012 Failover Cluster No Disks available / iSCSI

    Hi All,
    I am testing out the Failover Clustering on Windows Server 2012 with hopes of winding up with a clustered File Server once I am done. 
    I am starting with a single node in the cluster for testing purposes; I have connected to this cluster a single iSCSI LUN that is 100GB in size.
    When I right click on Storage -> Disks  and then click 'Add Disk', I get No disks suitable for cluster disks were found.
    I get this, even if I add a second server to the cluster, and connect it to the iSCSI drive as well.
    Any ideas?

    Hi All,
    I am testing out the Failover Clustering on Windows Server 2012 with hopes of winding up with a clustered File Server once I am done. 
    I am starting with a single node in the cluster for testing purposes; I have connected to this cluster a single iSCSI LUN that is 100GB in size.
    When I right click on Storage -> Disks  and then click 'Add Disk', I get No disks suitable for cluster disks were found.
    I get this, even if I add a second server to the cluster, and connect it to the iSCSI drive as well.
    Any ideas?
    For testing purpose you'd better spawn a set of VMs on a single physical Hyper-V host and use shared VHDX as a back clusterd storage. That would be both much easier and much faster then what you do. + it would be trivial move one of the VMs to another physical
    host, shared VHDX to CSV on a shared storage and go from Test & Development to production :) See:
    Shared VHDX
    http://blogs.technet.com/b/storageserver/archive/2013/11/25/shared-vhdx-files-my-favorite-new-feature-in-windows-server-2012-r2.aspx
    Virtual File Server with Shared VHDX
    http://www.aidanfinn.com/?p=15145
    Guest
    VM Cluster with Shared VHDX
    http://technet.microsoft.com/en-us/library/dn265980.aspx
    For a pure iSCSI scenario you may try this step-by-step guide (just skip StarWind config as you do have a shared storage already with your SAN). See:
    Configuring HA File Server on Windows Server 2012 for SMB NAS
    http://www.starwindsoftware.com/configuring-ha-file-server-on-windows-server-2012-for-smb-nas
    Hope this helped a bit :)
    StarWind VSAN [Virtual SAN] clusters Hyper-V without SAS, Fibre Channel, SMB 3.0 or iSCSI, uses Ethernet to mirror internally mounted SATA disks between hosts.

  • How to Perform Forced Manual Failover of Availability Group (SQL Server) and WSFC (Windows Server Failover Cluster)

    I have a scenario with the three nodes with server 2012 standard, each running an instance of SQL Server 2012 enterprise, participate in a
    single Windows Server Failover Cluster (WSFC) that spans two data centers.
    If the nodes in the primary data center are unavailable due to data center outage. Then how I can able to access node in the WSFC (Windows Server Failover Cluster) in the secondary disaster recovery data center automatically with some script.
    I want to write script that can be able to check primary data center by pinging some IP after every 5 or 10 minutes.
    If that IP is unable to respond then script can be able to Perform Forced Manual Failover of Availability Group (SQL Server) and WSFC (Windows Server Failover Cluster)
    Can you please guide me for script writing for automatic failover in case of primary datacenter outage?

    please post you question on failover clusters in the cluster forum.  THey will explain how this works and point you at scipts.
    You should also look in the Gallery for cluster management scripts.
    ¯\_(ツ)_/¯

  • How to Perform Forced Manual Failover of Availability Group (SQL Server) and WSFC (Windows Server Failover Cluster) with scrpiting

    I have a scenario with the three nodes with server 2012 standard, each running an instance of SQL Server 2012 enterprise, participate in a
    single Windows Server Failover Cluster (WSFC) that spans two data centers.
    If the nodes in the primary data center are unavailable due to data center outage. Then how I can able to access node in the WSFC (Windows Server Failover Cluster) in the secondary disaster recovery data center automatically with some script.
    I want to write script that can be able to check primary data center by pinging some IP after every 5 or 10 minutes.
    If that IP is unable to respond then script can be able to Perform Forced Manual Failover of Availability Group (SQL Server) and WSFC (Windows Server Failover Cluster)
    Can you please guide me for script writing for automatic failover in case of primary datacenter outage?

    You are trying to implement manually what should be happening automatically in the cluster. If the primary SQL Server becomes unavailable in the data center, it should fail over to the secondary SQL Server automatically.  Is that not working?
    You also might want to run this configuration by some SQL experts.  I am not a SQL expert, but if you have both hosts in the data center in a cluster, there is no need for replication between those two nodes as they would be accessing
    the database from some form of shared storage.  Then it looks like you are trying to implement Always On to the DR site.  I'm not sure you can mix both types of failover in a single configuration.
    FYI, it would make more sense to establish a file share witness in your DR site instead of placing a third node in the data center for Node Majority quorum.
    . : | : . : | : . tim

  • A Server in a cluster gone down event

    Hi,
              Looks like this question was asked indirectly earlier.
              I see that when 1 server is a cluster goes down, the others servers get
              a 'weblogic.rjvm.PeerGoneException'.
              Is there a way by which i can subscribe to the
              'weblogic.rjvm.PeerGoneEvent' using 'EventRegistrationDef'.
              I am using 5.1 and looking for a way at solving the isssue of Weblogic
              removing my non RMI Object from all other servers (JNDI tree) when the
              host server goes down.
              Thanks
              Mario
              

    The API is not public.
              This non rmi object that is living in JNDI, sounds like out of band data.
              If this object doesn't contain any state you can bind it from all the
              servers but don't replicate the bindings.
              Hope this helps.
              --- Prasad
              Mario Briggs wrote:
              > Hi,
              > Looks like this question was asked indirectly earlier.
              >
              > I see that when 1 server is a cluster goes down, the others servers get
              > a 'weblogic.rjvm.PeerGoneException'.
              > Is there a way by which i can subscribe to the
              > 'weblogic.rjvm.PeerGoneEvent' using 'EventRegistrationDef'.
              >
              > I am using 5.1 and looking for a way at solving the isssue of Weblogic
              > removing my non RMI Object from all other servers (JNDI tree) when the
              > host server goes down.
              >
              > Thanks
              > Mario
              

  • Deploy jw webservice to cluster server, always point to 9081

    i have cluster server(9081, 9082, 9083), when i deploy the helloword webservice generated by jwsc http://download.oracle.com/docs/cd/E12840_01/wls/docs103/webserv/use_cases.html,
    the dynamical wsdl always point to 9081,<s2:address location="http://localhost:9081/HelloWorldImpl/HelloWorldService" /> even when i invoke the service on 9082, http://localhost:9082/HelloWorldImpl/HelloWorldService?WSDL
    i tried generate webservice by axis it works fine. i gooled, couldn't find solution. thx.

    Thanks for looking into it. Clean cache doesn't work for me. Since it works fine for you. I think the issue might related to create/deploy ear file. i don't have cluster setup on my local, so i created ear file and deployed to our DEV(unix) server thru admin console. the web-uri in application.xml generated by the jwsc contains the path examples/webservices/hello_world/HelloWorldImpl.war, instead of HelloWorldImpl.war. i have to manually updated the uri, otherwise i got "not be able to identifiy application path ......war" during install. Since Axis works fine cluster config should be ok. Do you mind share how you config the cluster and how to generate the ear file? this are the two possible places i can think of. i also tried set FrontEnd HTTP Port for each server, still didn't work.
    May i know are you tech support from weblogic? i am trying to find way to create ticket for weblogic tech support for this issue. thx.

  • SQL Server 2012 Failover Cluster configuration questions

    Hi,
    I have few questions on , SQL Server 2012 Failover cluserting pleasse provide suggestions:
    1) In SQL Server 2012 is there a configuration for active / passive Failover Cluster installation? If so how is it done? if you could provide any links or articles that would help.
    OR 
    Is this been replaced by Always ON Availability Groups?
    2) Also in our environment we have done active/active installation but at a time my understanding is only node has the ownership to the shared storage versus both the nodes, is that correct? If not please provide an explanation?
    Any additional information would be valuable in clearing my doubts?
    Thank youy
    Malini=

    Hi malinisethi,
    If you install SQL Server in a cluster and configure Active-Passive cluster. In the first node, select “New SQL Server Failover Cluster installation” option , when we are installing Active-Passive cluster, we have to specify one virtual/network name. (Note:
    For Active-Active clustering you have specify different network names as per the number of nodes). on the other nodes is similar to installing on the first node except that we have select the Add Node to a SQL Server failover cluster option from the initial
    menu. For more information, there is a similar issue about SQL Server 2008 Active-Passive/Active-Active cluster installation, you can review the following article.
    http://sqldbpool.com/2009/10/07/sql-server-2008-active-passive-cluster-installation/
    About Active/Active SQL Cluster, two clustered SQL Server instance are created in different nodes. Then apply Active /Active configuration of both instance. There is an example about creating an Active/Active SQL Cluster using Hyper-V. you can review the
    following article.
    http://blogs.msdn.com/b/momalek/archive/2012/04/11/creating-an-active-active-sql-cluster-using-hyper-v-part2-the-clustered-instances.aspx
    Regards,
    Sofiya Li
    Sofiya Li
    TechNet Community Support

Maybe you are looking for

  • Calculate stock of material on a particular date ..very urgent (have points

    Hi all, My requirement is to calulate stock of a material on a particular date.. See. I need a function mudule that will accept.. matnr, date, werks....n give me stock as output.. Please help me reg. this.. Have ur points...

  • Is it possible to Clone system dr.230Gb 7200rpm drive to 150Gb 10,000 rpm?

    For some reason i remember reading a artical stating that you cannot clone from bigger drive to smaller if you do it does something with the system id and you will have to re-enter serials for your software Is this true if so is there a work around?

  • UploadBean with JBoss Problem

    Dear Experts I am currently creating a web page which allows the customers to upload csv file. I serached through the net and found that people are using "UploadBean" for this thus I downloaded the uploadbean jar file and imported it to my Jboss libr

  • 3502 access point will not connect to 4400 Series WLAN controller

    I recently upgraded our controllers to the latest version 7 software, as I read this was one of the requirements to get them to connect.  But I am not having any luck getting into a controller.  Normally I plug them in to the network, they pop into t

  • Enhancement for Date changes on IW52

    Hi All,    I have been looking for an exit or a BADI in the forum for transaction IW51 or IW52, and I couldn't find one which meets the requirement. The requirement is: On the IW51/IW52 screen, when user changes the Begin Date and hits the 'Enter' ke