Test Cluster...

HI dev2dev
          I have 2 bea box ,so i have clustered
          First Machine
          One Admin Server
          One Managed Server
          Second Machine
          One Managed Server
          Weblogic Console says both the machine's are in cluster
          1.How to Test,Wheather The Machines are in Cluster correctly
          2.I tried Deploy my EAR and it deployed successfuly in the cluster.I was able the access to the deployed EAR
          then i tried to stop Admin server and test for the EAR deployed.I was NOT able to the access the deployed EAR
          3.Let me know how to test the ear is deployed in managed server also.
          Thanks in Adv
          maria

Hi,
          In weblogic, the load balancer is deployed onto your
          Admin server (ie. web application containing
          HttpClusterServlet) & your application EAR is deployed
          onto individual managed servers.
          By restarting the admin server, you are shutting off the
          load balancer (which is the gateway to both the managed servers). No wonder shutting it off does not allow you access to your EAR.
          Instead, shutdown one of your managed servers & then try
          accessing your EAR (application). You will notice (from application logs) that all your requests are going to the other managed server (which is up & running).
          That would be evident enough that your cluster is up & running.
          Regards
          Sandeep

Similar Messages

  • Cluster Shared Volume disappeared after taking the volume offline for Validation Tests.

    Hi,
    After an unknown issue with one of our Hyper-V 4 Node cluster running on Server 2008 R2 SP1 with fibre channel NEC D3-10 SAN Storage all our cluster shared volumes were in redirecting mode and I was unable to get them back online. Only after rebooting all
    the nodes one by one the disks came back online. Eventlog messages indicated that I had to test my cluster validation. After shutting down all the virtual machines I set all the cluster shared volumes offline and started the complete validation test. The following
    warnings/errors appeared during the test.
    An error occurred while executing the test.
    An error occurred retrieving the
    disk information for the resource 'VSC2_DATA_H'.
    Element not found (Validate Volume Consistency Test)
    Cluster disk 4 is a Microsoft MPIO based disk
    Cluster disk 4 from node has 4 usable path(s) to storage target
    Cluster disk 4 from node has 4 usable path(s) to storage target
    Cluster disk 4 is not managed by Microsoft MPIO from node
    Cluster disk 4 is not managed by Microsoft MPIO from node (Validate Microsoft MPIO-based disks test)
    SCSI page 83h VPD descriptors for cluster disk 4 and 5 match (Validate SCSI device Vital Product Data (VPD) test)
    After the test the cluster shared volume was disappeared (the resource is online).
    Cluster events that are logged
    Cluster physical disk resource 'DATA_H' cannot be brought online because the associated disk could not be found. The expected signature of the disk was '{d6e6a1e0-161e-4fe2-9ca0-998dc89a6f25}'. If the disk was replaced or restored, in the Failover Cluster
    Manager snap-in, you can use the Repair function (in the properties sheet for the disk) to repair the new or restored disk. If the disk will not be replaced, delete the associated disk resource. (Event 1034)
    Cluster disk resource found the disk identifier to be stale. This may be expected if a restore operation was just performed or if this cluster uses replicated storage. The DiskSignature or DiskUniqueIds property for the disk resource has been corrected.
    (Event 1568)
    In disk management the disk is unallocated, unknown, Reserved. When the resource is on one node and i open disk management i get the warning that i have to initialize the disk. I did not do this yet.
    Reading from other posts i think that the partition table got corrupted but i have no idea how to get it back. I found the following information but it's not enough for me to go ahead with: Using a tool like TestDisk to rewrite the partition table. then
    rewriting the uniqueID to the disk brought everything back. But still no explaination as to why we had our "High Availability" Fail Over cluster down for nearly 2 Days. This happened to us twice within the past week.
    Anybody that an idea how to solve this? I think my data is still intact.
    Thanx for taking the time to read this.
    DJITS.

    Hi,
    Error information you provided indicate disk connection failure issue, please confirm shared disk 4 is available:
    To review hardware, connections, and configuration of a disk in cluster storage:
    On each node in the cluster, open Disk Management (which is in Server Manager under Storage) and see if the disk is visible from one of the nodes (it should be visible from one node but not multiple nodes). If it is visible to
    a node, continue to the next step. If it is not visible from any node, still in Disk Management on a node, right-click any volume, click Properties, and then click the Hardware tab. Click the listed disks or LUNs to see if all expected disks or LUNs appear.
    If they do not, check cables, multi-path software, and the storage device, and correct any issues that are preventing one or more disks or LUNs from appearing. If this corrects the overall problem, skip all the remaining steps and procedures.
    Review the event log for any events that indicate problems with the disk. If an event provides information about the disk signature expected by the cluster, save this information and skip to the last step in this procedure.
    To open the failover cluster snap-in, click Start, click Administrative Tools, and then click Failover Cluster Management. If the User Account Control dialog box appears, confirm that the action it displays is what you want, and
    then click Continue.
    In the Failover Cluster Management snap-in, if the cluster you want to manage is not displayed, in the console tree, right-click Failover Cluster Management, click Manage a Cluster, and then select or specify the cluster that
    you want.
    If the console tree is collapsed, expand the tree under the cluster you want to manage, and then click Storage.
    In the center pane, find the disk resource whose configuration you want to check, and record the exact name of the resource for use in a later step.
    Click Start, point to All Programs, click Accessories, right-click Command Prompt, and then click Run as administrator.
    Type:
    CLUSTER RESOURCE DiskResourceName /PRIV >path\filename.TXT
    For DiskResourceName, type the name of the disk resource, and for path\filename, type a path and a new filename of your choosing.
    Locate the file you created in the previous step and open it. For a master boot record (MBR) disk, look in the file for DiskSignature. For a GPT disk, look in the file for DiskIdGuid.
    Use the software for your storage to determine whether the signature of the disk matches either the DiskSignature or DiskIdGuid for the disk resource. If it does not, use the following procedure to repair the disk configuration.
    For more information please refer to following MS articles:
    Event ID 1034 — Cluster Storage Functionality
    http://technet.microsoft.com/en-us/library/cc756229(v=WS.10).aspx
    Hope this helps!
    TechNet Subscriber Support
    If you are
    TechNet Subscription user and have any feedback on our support quality, please send your feedback
    here.
    Lawrence
    TechNet Community Support

  • Reading data from a Table Cluster sans SAP code

    I need to read data at the database level from a cluster table.  That is to say, I want to decompress the VARDATA column that is in
    the database table and interpreet it with my own code.
    I've read all the SAP help documents on the subject, and I have created a test Table Cluster (the physical object) and associated test Cluster Tables (the logical tables) with the Table Cluster, but the data is compressed at the at the database level and incomprehensible.
    I realize one can use ABAP and OPEN SQL to read the table, because that goes through the SAP database interface.  I could
    also call a BAPI from an external program, but I don't want to do that.
    I need to read the data directly from the database without going through the SAP database interface or calling any SAP code.  I
    am willing to write the required code in another language, I just need to know how the data is compressed in the cluster, so I can
    decompress it.
    1. Does anyone know how to decode the information stored in a table cluster  WITHOUT using SAP code? 
    2. If no one knows how to do (1), does anyone know what compression algorithm SAP uses when storing data in the table cluster?
    3. Is there a specification or layout that explains exactly how the VARDATA column in a table cluster is structured, so that once I
    decompress it, I can figure out what data belonged to which table and column?
    If you even have a guess as to the compression algorithm or layout of VARDATA, I would be interested to hear it.
    Thank You,
    Larry Smith

    Hi All,
    The question is still open, waiting for points to be awarded.  Fidel's answer didn't really help.  I already know that I currently can't read the cluster, that's why I posted the question. 
    I would think SAP uses a known compression algorithm for the clusters, so the most important thing is finding out what the algorithm is, and finding out if SAP does anything weird to the compression headers.
    Someone has to know out there!
    Thank You,
    Larry Smith

  • Error when running weblogic 10 in cluster

    I am testing out weblogic 10 and when I run it with a single managed server and an admin server then everything works fine. When I run it in a cluster with a valid test cluster license, I get the following error. Does anyone know what the cause of this error is and how to fix it? Thanks for the help.
              <Sep 14, 2007 7:06:03 PM UTC> <Warning> <RMI> <BEA-080003> <RuntimeException thrown by rmi server: weblogic.rmi.internal.dgc.DGCServerImpl.renewLease([I)
              java.lang.ClassCastException: weblogic.rjvm.ImmutableServiceContext.
              java.lang.ClassCastException: weblogic.rjvm.ImmutableServiceContext
              at weblogic.rjvm.MsgAbbrevInputStream.readClassDescriptor(MsgAbbrevInputStream.java:369)
              at weblogic.utils.io.ChunkedObjectInputStream$NestedObjectInputStream.readClassDescriptor(ChunkedObjectInputStream.java:265)
              at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1534)
              at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1466)
              at java.io.ObjectInputStream.readArray(ObjectInputStream.java:1591)
              at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1299)
              at java.io.ObjectInputStream.readObject(ObjectInputStream.java:348)
              at weblogic.utils.io.ChunkedObjectInputStream.readObject(ChunkedObjectInputStream.java:195)
              at weblogic.rjvm.MsgAbbrevInputStream.readObject(MsgAbbrevInputStream.java:565)
              at weblogic.utils.io.ChunkedObjectInputStream.readObject(ChunkedObjectInputStream.java:191)
              at weblogic.rmi.internal.dgc.DGCServerImpl_WLSkel.invoke(Unknown Source)
              at weblogic.rmi.internal.BasicServerRef.invoke(BasicServerRef.java:589)
              at weblogic.rmi.internal.BasicServerRef$1.run(BasicServerRef.java:479)
              at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:363)
              at weblogic.security.service.SecurityManager.runAs(Unknown Source)
              at weblogic.rmi.internal.BasicServerRef.handleRequest(BasicServerRef.java:475)
              at weblogic.rmi.internal.BasicServerRef.access$300(BasicServerRef.java:59)
              at weblogic.rmi.internal.BasicServerRef$BasicExecuteRequest.run(BasicServerRef.java:1016)
              at weblogic.work.SelfTuningWorkManagerImpl.schedule(SelfTuningWorkManagerImpl.java:126)
              at weblogic.rmi.internal.BasicServerRef.dispatch(BasicServerRef.java:321)
              at weblogic.rmi.internal.BasicServerRef.dispatch(BasicServerRef.java:918)
              at weblogic.rjvm.RJVMImpl.dispatchRequest(RJVMImpl.java:1084)
              at weblogic.rjvm.RJVMImpl.dispatch(RJVMImpl.java:1001)
              at weblogic.rjvm.ConnectionManagerServer.handleRJVM(ConnectionManagerServer.java:230)
              at weblogic.rjvm.ConnectionManager.dispatch(ConnectionManager.java:877)
              at weblogic.rjvm.MsgAbbrevJVMConnection.dispatch(MsgAbbrevJVMConnection.java:446)
              at weblogic.rjvm.t3.MuxableSocketT3.dispatch(MuxableSocketT3.java:368)
              at weblogic.socket.AbstractMuxableSocket.dispatch(AbstractMuxableSocket.java:383)
              at weblogic.socket.SocketMuxer.readReadySocketOnce(SocketMuxer.java:872)
              at weblogic.socket.SocketMuxer.readReadySocket(SocketMuxer.java:818)
              at weblogic.socket.EPollSocketMuxer.dataReceived(EPollSocketMuxer.java:192)
              at weblogic.socket.EPollSocketMuxer.processSockets(EPollSocketMuxer.java:174)
              at weblogic.socket.SocketReaderRequest.run(SocketReaderRequest.java:29)
              at weblogic.socket.SocketReaderRequest.execute(SocketReaderRequest.java:42)
              at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:145)
              at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:117)
              >

    We solved this by increasing the number of open files, by adding the following in the /etc/security/limits.conf
    * soft nofile 50000
    * hard nofile 50000

  • Jms in cluster / load balancing and failover

              Did I get it right ???
              I have 1 admin server and 4 managed servers in a 2 clusters, a development cluster
              and a test cluster.
              I now want to have loadbalancing with my jms server and I want to be able to migrate
              my jms server in case of failer.
              for each cluster I have created
              connectionFactory targeted to the cluster
              Distributed destination with 2 queue members
              One JMSServer migratable targeted on managed server 1, with destination 1 from
              the ditributed destination
              One JMSServer migratable targeted on managed server 2, with destination 2 from
              the ditributed destination
              I expect this to make loadbalancing between the 2 servers in the cluster, and
              I can migrate the jms server if one of the server fails to the running server.
              One thing is now.....If one server fails and I migrate the jms server to the other
              server that is running, and I then restart the server that was down, what is then
              happening, do I then have 3 jms servers ???
              [config.xml]
              

              "Kris" <[email protected]> wrote in message news:[email protected]...
              >
              > "Kawaljit Singh Sunny" <[email protected]> wrote:
              > >
              > >"Kris" <[email protected]> wrote in message
              > >news:[email protected]...
              > >>
              > >> Did I get it right ???
              > >>
              > >> I have 1 admin server and 4 managed servers in a 2 clusters, a
              development
              > >cluster
              > >> and a test cluster.
              > >> I now want to have loadbalancing with my jms server and I want to be
              > >able
              > >to migrate
              > >> my jms server in case of failer.
              > >>
              > >> for each cluster I have created
              > >> connectionFactory targeted to the cluster
              > >> Distributed destination with 2 queue members
              > >> One JMSServer migratable targeted on managed server 1, with destination
              > >1
              > >from
              > >> the ditributed destination
              > >> One JMSServer migratable targeted on managed server 2, with destination
              > >2
              > >from
              > >> the ditributed destination
              > >
              > >If the server where your JMSConnections are loadBalanced to goes down,
              > >the
              > >producers and consumers using this JMSConnection are closed.
              > >You have to recreate these producers and consumers.
              > >If the server where your Destination resides goes dow, the consumers
              > >are
              > >closed.
              > >If the producers JMSConnection is not on this server, the producer stays
              > >up.
              > >
              > >>
              > >>
              > >> I expect this to make loadbalancing between the 2 servers in the
              cluster,
              > >and
              > >> I can migrate the jms server if one of the server fails to the running
              > >server.
              > >>
              > >> One thing is now.....If one server fails and I migrate the jms server
              > >to
              > >the other
              > >> server that is running, and I then restart the server that was down,
              > >what
              > >is then
              > >> happening, do I then have 3 jms servers ???
              > >
              > >No you still have 2 JMSServers. JMS Migration is manual.
              > >
              > >>
              >
              > you say : No you still have 2 JMSServers. JMS Migration is manual.
              >
              > But if I manual migrate the jmsserver that was down to the running wls
              server,
              > that already have one jms server running, this wls server must then have 2
              jms
              > servers. And I boot the wls server that hosted the jms server that was
              down, this
              > will now have a running jms server. isn't that 3 jms servers ?
              Once you migrate a JMSServer from a WeblogicServer1 to WeblogicServer2,
              and then you boot WeblogicServer1, this JMSServer which was migrated should
              NOT be on WeblogicServer1.
              (You have migrated the JMSServer from WeblogicServer1 to WeblogicServer2)
              >
              > But I was thinking about that I could spare the migration part. If I have
              2 wls
              > servers and a jms server on each of them, and a destributed destination
              with 2
              > queue members that are persistent in a database. If a wls og just a jms
              server
              > goes down, I just have to reboot the server and it will run again. This
              way I
              > dont have to think about migration, or what ?
              Yes that is true.
              Irrespective of whether you have migration or not,
              only thing you need to do take care is to reconnect to weblogic server, if
              the the server where your JMSConnection is loadBalanced to goes down.
              There is no failover of JMSConnections. Producers inside this JMSConnection
              will be closed. You will have to create a new JMSConnection and a new
              Producer and continue with your production of JMS Messages.
              -sunny
              

  • Two node cluster - disk not responding to selection

    I'm building 2 node cluster (Solaris 10/SC3.2) on Dell's 1950/PERC6i servers with quorum as a virtual server. Because I need to introduce quorum server to the cluster - my cluster nodes are still in install mode.
    I have tried to add quorum using scsetup or clsetup but I'm getting always the same message:
    root@node01:~# scsetup
    Failed to get node zone list
    Failed to get node zone list
        This program has detected that the cluster "installmode" attribute is
        still enabled. As such, certain initial cluster setup steps will be
        performed at this time. This includes adding any necessary quorum
        devices, then resetting both the quorum vote counts and the
        "installmode" property.
        Please do not proceed if any additional nodes have yet to join the
        cluster.
        Is it okay to continue (yes/no) [yes]?  yes
    Unable to establish the list of cluster nodes.
    Press Enter to continue:Also the most imortant issue is that immediately after restart of the first node during scinstall procedure, I started to getting those follwing messages:
    Feb  4 17:33:20 node01 scsi: [ID 107833 kern.warning] WARNING: /pci@0,0/pci8086,25e3@3/pci1028,1f0c@0/sd@0,0 (sd3):
    Feb  4 17:33:20 node01      disk not responding to selection
    Feb  4 17:33:21 node01 scsi: [ID 107833 kern.warning] WARNING: /pci@0,0/pci8086,25e3@3/pci1028,1f0c@0/sd@0,0 (sd3):
    Feb  4 17:33:21 node01      disk not responding to selection
    Feb  4 17:26:46 node02 scsi: [ID 107833 kern.warning] WARNING: /pci@0,0/pci8086,25e3@3/pci1028,1f0c@0/sd@1,0 (sd4):
    Feb  4 17:26:46 node02      disk not responding to selection
    Feb  4 17:26:46 node02 scsi: [ID 107833 kern.warning] WARNING: /pci@0,0/pci8086,25e3@3/pci1028,1f0c@0/sd@0,0 (sd3):
    Feb  4 17:26:46 node02      disk not responding to selectionBoth nodes are extreamly slow, I could use only telnet to login because of long timeouts for many services.
    Here is output from: cfgadm -l
    root@node01:~# cfgadm -al
    Ap_Id                          Type         Receptacle   Occupant     Condition
    c0                             scsi-bus     connected    configured   unknown
    c0::dsk/c0t0d0                 disk         connected    configured   unknown
    c0::dsk/c0t1d0                 disk         connected    configured   unknown
    c4                             fc           connected    unconfigured unknown
    c5                             fc           connected    unconfigured unknown
    usb0/1                         unknown      empty        unconfigured ok
    usb0/2                         unknown      empty        unconfigured ok
    usb1/1                         unknown      empty        unconfigured ok
    usb1/2                         unknown      empty        unconfigured ok
    usb2/1                         unknown      empty        unconfigured ok
    usb2/2                         unknown      empty        unconfigured ok
    usb3/1                         unknown      empty        unconfigured ok
    usb3/2                         unknown      empty        unconfigured ok
    usb4/1                         usb-hub      connected    configured   ok
    usb4/1.1                       usb-device   connected    configured   ok
    usb4/1.2                       usb-device   connected    configured   ok
    usb4/2                         unknown      empty        unconfigured ok
    usb4/3                         unknown      empty        unconfigured ok
    usb4/4                         unknown      empty        unconfigured ok
    usb4/5                         usb-hub      connected    configured   ok
    usb4/5.1                       unknown      empty        unconfigured ok
    usb4/5.2                       unknown      empty        unconfigured ok
    usb4/5.3                       unknown      empty        unconfigured ok
    usb4/5.4                       unknown      empty        unconfigured ok
    usb4/6                         unknown      empty        unconfigured ok
    usb4/7                         unknown      empty        unconfigured ok
    usb4/8                         unknown      empty        unconfigured okHow to solve those two problems? This one with SCSI issue and a problem with node list...
    Best regards,
    Vladimir

    During a last weekend I have reinstalled both nodes. I rebuild virtual disks (LUNs) under PERC 6i controller. I have chosen RAID0 instead of RAID1 as it was before. Still I'm not sure did RAID0 help me or just a rebuilding disks. MegaCli tool from LSI could not help me and tell what was wrong with disk/partitions that I used with first try. For MegaCli status of controller was without errors.
    Probably something was wrong with this partitioning. I used also as in first try Solaris 10 with latest patches 01/2009 and Sun Cluster 3.2 also with latest security patch.
    After reinstalling everything is almost fine :-). The only difference with other productive clusters (the patch level of Solaris & SUN Cluster is not as not this test cluster) is that cacao container agent is offline:
    offline        Feb_07   svc:/application/management/common-agent-container-1:default
    because of that I have following service as disabled:
    disabled       Feb_07   svc:/system/cluster/rgm:default
    Does anyone knows, how serious is this? And how to enable now svc:/application/management/common-agent-container-1:default?
    Here is svcs -xv output:
    svc:/application/print/server:default (LP print server)
    State: disabled since Sat Feb 07 10:42:05 2009
    Reason: Disabled by an administrator.
       See: http://sun.com/msg/SMF-8000-05
       See: man -M /usr/share/man -s 1M lpsched
    Impact: 2 dependent services are not running:
            svc:/application/print/rfc1179:default
            svc:/application/print/ipp-listener:default
    svc:/system/cluster/rgm:default (Resource Group Manager Daemon)
    State: disabled since Sat Feb 07 10:42:06 2009
    Reason: Disabled by an administrator.
       See: http://sun.com/msg/SMF-8000-05
    Impact: 1 dependent service is not running:
            svc:/application/management/common-agent-container-1:default
    svc:/system/cluster/scsymon-srv:default (Sun Cluster SyMON Server Daemon)
    State: offline since Sat Feb 07 10:42:07 2009
    Reason: Dependency svc:/application/management/sunmcagent:default is absent.
       See: http://sun.com/msg/SMF-8000-E2
    Impact: This service is not running.Regards,
    Vladimir

  • What are best practices for rolling out cluster upgrade?

    Hello,
    I am looking for your input on the approaches to implement Production RAC upgrade without having a spare instance of RAC servers. We have a 2-node database RAC 11.1 on Production. We are planning to upgrade to 11.2 but only have a single database that the pre-Production integration can be verified on. Our concern is that the integration may behave differently on RAC vs. the signle database instance. How everybody else approaches this problem?

    you want to test RAC upgrade on NON RAC database. If you ask me that is a risk but it depends on may things
    Application configuration - If your application is configured for RAC, FAN etc. you cannot test it on non RAC systems
    Cluster upgrade - If your standalone database is RAC one node you can probably test your cluster upgrade there. If you have non RAC database then you will not be able to test cluster upgrade or CRS
    Database upgrade - There are differences when you upgrade RAC vs non RAC database which you will not be able to test
    I think the best way for you is to convert your standalone database to RAC one node database and test it. that will take you close to multi node RAC

  • Creating a Two Node Cluster

    Good afternoon,
    I'm looking to build/create an inexpensive two node cluster. I have a SLES11SP1 server that is running XEN as a virtual hosting server, I run about five servers in a virtual environment. I have three USB drives set up to host my Guest servers, what I would like to do is to purchase another USB drive, so that I can use that as an iscsi/SAN server location, it would be 2TB in size, and then I would build my new servers making them cluster enabled. They should then be able to "see" the SAN/iscsi/storage location.
    Does anyone have any further suggestions?
    Thanks
    -DS

    Originally Posted by gleach1
    while I wouldn't recommend running servers off USB drives (unless this is for testing), I don't see any issue in doing it but the performance may not be fantastic
    I really appreciate your responding back and assisting in this testing. This cluster is for testing only, I have a small server farm in my basement, and I have a customer that is using clustering and I would like to at least be able to have something to test with
    I've actually run a test cluster off a USB drive using vmware workstation before and it seemed to run fine just as a test system
    If you set up your xen host as an iscsi server, use the USB disk as the storage you present to the guests and set them up with iscsi initiators it should work like any other iscsi san would, obviously a touch slower...
    Is there some documentation that explains how to do the iscsi server setup that you described? I have the "Configuring Novell Cluster Services in a XEN Virtualization Environment" but I really don't see anything about the iscsi initiator setup, I was going to add the USB drive as a /storage volume on the XEN host, and then point the cluster to that? I will also be adding a third card to handle the clustering network.

  • Error received when running weblogic in cluster

    I am trying out clusters to see what the benefit would be to our application but I keep getting an error in the background. I have obtained a test cluster license and it is loading it up on startup. Does anyone know what this error is or how to fix it? Thanks for the help.
    <Sep 14, 2007 7:06:03 PM UTC> <Warning> <RMI> <BEA-080003> <RuntimeException thrown by rmi server: weblogic.rmi.internal.dgc.DGCServerImpl.renewLease([I)
    java.lang.ClassCastException: weblogic.rjvm.ImmutableServiceContext.
    java.lang.ClassCastException: weblogic.rjvm.ImmutableServiceContext
    at weblogic.rjvm.MsgAbbrevInputStream.readClassDescriptor(MsgAbbrevInputStream.java:369)
    at weblogic.utils.io.ChunkedObjectInputStream$NestedObjectInputStream.readClassDescriptor(ChunkedObjectInputStream.java:265)
    at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1534)
    at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1466)
    at java.io.ObjectInputStream.readArray(ObjectInputStream.java:1591)
    at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1299)
    at java.io.ObjectInputStream.readObject(ObjectInputStream.java:348)
    at weblogic.utils.io.ChunkedObjectInputStream.readObject(ChunkedObjectInputStream.java:195)
    at weblogic.rjvm.MsgAbbrevInputStream.readObject(MsgAbbrevInputStream.java:565)
    at weblogic.utils.io.ChunkedObjectInputStream.readObject(ChunkedObjectInputStream.java:191)
    at weblogic.rmi.internal.dgc.DGCServerImpl_WLSkel.invoke(Unknown Source)
    at weblogic.rmi.internal.BasicServerRef.invoke(BasicServerRef.java:589)
    at weblogic.rmi.internal.BasicServerRef$1.run(BasicServerRef.java:479)
    at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:363)
    at weblogic.security.service.SecurityManager.runAs(Unknown Source)
    at weblogic.rmi.internal.BasicServerRef.handleRequest(BasicServerRef.java:475)
    at weblogic.rmi.internal.BasicServerRef.access$300(BasicServerRef.java:59)
    at weblogic.rmi.internal.BasicServerRef$BasicExecuteRequest.run(BasicServerRef.java:1016)
    at weblogic.work.SelfTuningWorkManagerImpl.schedule(SelfTuningWorkManagerImpl.java:126)
    at weblogic.rmi.internal.BasicServerRef.dispatch(BasicServerRef.java:321)
    at weblogic.rmi.internal.BasicServerRef.dispatch(BasicServerRef.java:918)
    at weblogic.rjvm.RJVMImpl.dispatchRequest(RJVMImpl.java:1084)
    at weblogic.rjvm.RJVMImpl.dispatch(RJVMImpl.java:1001)
    at weblogic.rjvm.ConnectionManagerServer.handleRJVM(ConnectionManagerServer.java:230)
    at weblogic.rjvm.ConnectionManager.dispatch(ConnectionManager.java:877)
    at weblogic.rjvm.MsgAbbrevJVMConnection.dispatch(MsgAbbrevJVMConnection.java:446)
    at weblogic.rjvm.t3.MuxableSocketT3.dispatch(MuxableSocketT3.java:368)
    at weblogic.socket.AbstractMuxableSocket.dispatch(AbstractMuxableSocket.java:383)
    at weblogic.socket.SocketMuxer.readReadySocketOnce(SocketMuxer.java:872)
    at weblogic.socket.SocketMuxer.readReadySocket(SocketMuxer.java:818)
    at weblogic.socket.EPollSocketMuxer.dataReceived(EPollSocketMuxer.java:192)
    at weblogic.socket.EPollSocketMuxer.processSockets(EPollSocketMuxer.java:174)
    at weblogic.socket.SocketReaderRequest.run(SocketReaderRequest.java:29)
    at weblogic.socket.SocketReaderRequest.execute(SocketReaderRequest.java:42)
    at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:145)
    at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:117)
    >

    I am trying out clusters to see what the benefit would be to our application but I keep getting an error in the background. I have obtained a test cluster license and it is loading it up on startup. Does anyone know what this error is or how to fix it? Thanks for the help.
    <Sep 14, 2007 7:06:03 PM UTC> <Warning> <RMI> <BEA-080003> <RuntimeException thrown by rmi server: weblogic.rmi.internal.dgc.DGCServerImpl.renewLease([I)
    java.lang.ClassCastException: weblogic.rjvm.ImmutableServiceContext.
    java.lang.ClassCastException: weblogic.rjvm.ImmutableServiceContext
    at weblogic.rjvm.MsgAbbrevInputStream.readClassDescriptor(MsgAbbrevInputStream.java:369)
    at weblogic.utils.io.ChunkedObjectInputStream$NestedObjectInputStream.readClassDescriptor(ChunkedObjectInputStream.java:265)
    at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1534)
    at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1466)
    at java.io.ObjectInputStream.readArray(ObjectInputStream.java:1591)
    at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1299)
    at java.io.ObjectInputStream.readObject(ObjectInputStream.java:348)
    at weblogic.utils.io.ChunkedObjectInputStream.readObject(ChunkedObjectInputStream.java:195)
    at weblogic.rjvm.MsgAbbrevInputStream.readObject(MsgAbbrevInputStream.java:565)
    at weblogic.utils.io.ChunkedObjectInputStream.readObject(ChunkedObjectInputStream.java:191)
    at weblogic.rmi.internal.dgc.DGCServerImpl_WLSkel.invoke(Unknown Source)
    at weblogic.rmi.internal.BasicServerRef.invoke(BasicServerRef.java:589)
    at weblogic.rmi.internal.BasicServerRef$1.run(BasicServerRef.java:479)
    at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:363)
    at weblogic.security.service.SecurityManager.runAs(Unknown Source)
    at weblogic.rmi.internal.BasicServerRef.handleRequest(BasicServerRef.java:475)
    at weblogic.rmi.internal.BasicServerRef.access$300(BasicServerRef.java:59)
    at weblogic.rmi.internal.BasicServerRef$BasicExecuteRequest.run(BasicServerRef.java:1016)
    at weblogic.work.SelfTuningWorkManagerImpl.schedule(SelfTuningWorkManagerImpl.java:126)
    at weblogic.rmi.internal.BasicServerRef.dispatch(BasicServerRef.java:321)
    at weblogic.rmi.internal.BasicServerRef.dispatch(BasicServerRef.java:918)
    at weblogic.rjvm.RJVMImpl.dispatchRequest(RJVMImpl.java:1084)
    at weblogic.rjvm.RJVMImpl.dispatch(RJVMImpl.java:1001)
    at weblogic.rjvm.ConnectionManagerServer.handleRJVM(ConnectionManagerServer.java:230)
    at weblogic.rjvm.ConnectionManager.dispatch(ConnectionManager.java:877)
    at weblogic.rjvm.MsgAbbrevJVMConnection.dispatch(MsgAbbrevJVMConnection.java:446)
    at weblogic.rjvm.t3.MuxableSocketT3.dispatch(MuxableSocketT3.java:368)
    at weblogic.socket.AbstractMuxableSocket.dispatch(AbstractMuxableSocket.java:383)
    at weblogic.socket.SocketMuxer.readReadySocketOnce(SocketMuxer.java:872)
    at weblogic.socket.SocketMuxer.readReadySocket(SocketMuxer.java:818)
    at weblogic.socket.EPollSocketMuxer.dataReceived(EPollSocketMuxer.java:192)
    at weblogic.socket.EPollSocketMuxer.processSockets(EPollSocketMuxer.java:174)
    at weblogic.socket.SocketReaderRequest.run(SocketReaderRequest.java:29)
    at weblogic.socket.SocketReaderRequest.execute(SocketReaderRequest.java:42)
    at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:145)
    at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:117)
    >

  • Cluster Validation error.

    I have 2012 R2 Failover cluster with two nodes.
    When I try to add third node and run Validation test, I've got error in Validation Report:
    "Two disks have been found on node Hyperv01.local.com with duplicate disk signatures or disk GUIDs. The disks involved are physical disk 2 and physical disk 12. There may be a Multipath I/O (MPIO) problem such as the software is not installed or is not
    working properly. If MPIO is not involved, or MPIO has been verified as working, you must either mask one of these disks off at this node, or run validation and specify a disk list that includes only one of these disks, for example by using the Test-Cluster
    cmdlet in Windows PowerShell."
    But disk signature and disk GUIDs are different.
    All disks connected via ISCSI.
    Can anyone please help?
    Thank you.

    Hi oxyGEN_ua,
    Please confirm your storage compatible with server 2012r2 and your do not use the disk which marked as active and please confirm all your node have the same update and hotfix,
    please install the following update.
    Recommended hotfixes and updates for Windows Server 2012 R2-based failover clusters
    https://support.microsoft.com/kb/2920151/en-us?wa=wsignin1.0
    I’m glad to be of help to you!
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • How can I programatically change a reference to an item in a cluster?

    Hello!
    I'm trying to programatically change the text background color of the elements in a large cluster.
    I'm using LabVIEW 2010 SP1.  Attached is a test vi that I'm expermimenting with.  The test cluster has 6 elements and the final cluster will have 140.
    All elements will be the same data type.
    Based on the rate of change of the element value, I have to change the background color.
    Instead of creating a reference for each element in the cluster (as in my test vi), I'm looking for a way of changing the reference programatically.  Is there a way?
    I'm also considering building an array of references as a brute force alternative but I would like to see if there is something more elegant / efficient first!
    Am I making sense?
    thanks,
    mike
    Solved!
    Go to Solution.
    Attachments:
    cluster_color_change.vi ‏11 KB

    Mike,
    You can get references to the elements inside the cluster from the cluster reference using property node: All Objects[].  This is an array of references to the controls inside the cluster.
    Another way which might have advantages is to put a color box indicator behind each numeric.  Color the numeric backgrounds Transparent.  Then write the color to the color box.  I did a quick modification to your VI to show this possibility.  I did not adjust the size of the color box to match the numeric to emphasize what I did.
    Lynn
    Another note: If any of your users are color blind some of the combinations (black on red) might not be very visible.
    Attachments:
    cluster_color_change.2.vi ‏9 KB

  • Using orchestrator to create failover cluster. out of options

    Hi all,
    I am trying to build failover cluster using orchestrator 2012 R2 and then services on the cluster.
    However, I am so stuck now at the first stage:getting the cluster online.
    It is very basic, 2 nodes 2008R2 servers, shared storage for quorum, a virtual name and an IP.
    I try to put as little things as possible at start. I added the failover clustering feature into the orchestrator to avoid the double hop problem. I found that I can create the cluster with a command like this.
    C:\Windows\System32\WindowsPowerShell\v1.0\powershell.exe
    -NonInteractive -Command {Import-Module 'C:\Windows\system32\WindowsPowerShell\1.0\Modules\FailoverClusters\FailoverClusters.psd1';New-Cluster -name clustername -node servername1,servername2 -staticaddress 1.1.1.1}
    I use the powershell.exe to avoid the 32bit powershell problem from orchestartor.
    I can successfully create the cluster using this same command if I run it manually inside my powershell prompt and it usually took 20secs to complete and shows the progress within powershell prompt.
    but when I put in the Run .Net Script and powershell in my orchestrator and run it. It shown success without give any warning/error but unfortunately I found that there are errors in the event log, as below
    Could not retrieve the network topology for the nodes.
    Security must be initialized before any interfaces are marshalled or unmarshalled. It cannot be changed once initialized
    I'd like to say that if I replace "New-Cluster" with "Test-Cluster". It runs normally and perform all tests with a report as output.
    Tried to turn off UAC (admin approval mode for all administrators)
    I’ve also created the disabeld cluster AD computer object with enough permissions before running the command (once again. I can create the cluster manually running the command)
    Also, I’ve set the orchestrator to run the runbook service using my user account. So the credential used for manual execution and orchestrator should be the same (the account is local administrators group member to both node)
    P.S. I found that the error Security must be initialized before any interfaces are marshalled or
    unmarshalled is not specific for Orchestrator + powershell but also for other products.
    Please help, thanks

    Turn out I need to use the PowerShell OIP from
    link
    Setup CredSSP authentication
    Then use the following script to call powershell 3.0
    $ABC = powershell {Import-Module ‘C:\Windows\system32\WindowsPowerShell\1.0\Modules\FailoverClusters\FailoverClusters.psd1′;New-Cluster -name clustername -node servername1,servername2 -staticaddress 1.1.1.1}
    finally can run the activity without issue

  • Cluster not load-balancing, ideas?

    I've been struggling to identify why my JMS producers are not load-balancing against a remote cluster.
              I've ruled out the producer as being the problem (I see the same non-load-balancing behavior regardless of what I use to create messages - Hermes, ALSB, simple Java producer...) I also don't think the JMS Connection Factory config is the problem, judging by the help I've received from folks over on the jms forum.
              I believe something is wrong with our cluster setup because in addition to the problem I just mentioned, we also are not seeing JNDI entries propagate to all managed servers - for example, if I create one jms queue on m1, that queue does not appear in the jndi tree on m2.
              I've been trying to find any documentation on what settings I should look at to verify the cluster configuration. If I go through the WLS console and look at the Cluster settings, I see both managed servers there, is there some other place that the configuration could be messed up?
              Added 6/11, 9:30 am:
              We're focusing on multicast now as the most likely problem. Can anyone tell me whether clusters on the same multicast address but different ports will interfere with each other? It looks like the infrastructure team has set up 5 clusters like that (same multicast address in each cluster, but different ports).
              We've got a ticket open with BEA but it's been two weeks now and nothing except requests for more information.
              Any ideas/help are much appreciated!
              Meghan
              Edited by pietila at 06/11/2008 7:38 AM

    Meghan Pietila wrote:
              > I've been struggling to identify why my JMS producers are not load-balancing against a remote cluster.
              >
              > I've ruled out the producer as being the problem (I see the same non-load-balancing behavior regardless of what I use to create messages - Hermes, ALSB, simple Java producer...) I also don't think the JMS Connection Factory config is the problem, judging by the help I've received from folks over on the jms forum.
              >
              > I believe something is wrong with our cluster setup because in addition to the problem I just mentioned, we also are not seeing JNDI entries propagate to all managed servers - for example, if I create one jms queue on m1, that queue does not appear in the jndi tree on m2.
              >
              > I've been trying to find any documentation on what settings I should look at to verify the cluster configuration. If I go through the WLS console and look at the Cluster settings, I see both managed servers there, is there some other place that the configuration could be messed up?
              >
              > Added 6/11, 9:30 am:
              > We're focusing on multicast now as the most likely problem. Can anyone tell me whether clusters on the same multicast address but different ports will interfere with each other? It looks like the infrastructure team has set up 5 clusters like that (same multicast address in each cluster, but different ports).
              >
              > We've got a ticket open with BEA but it's been two weeks now and nothing except requests for more information.
              >
              > Any ideas/help are much appreciated!
              >
              > Meghan
              >
              > --
              > Edited by pietila at 06/11/2008 7:38 AM
              You could be right. I think we have had problems where the same IP but
              different ports were used for multicast. This is on 8.1 though.
              I think as a rule, it's best to have a different ip and port for each
              cluster.
              Also - can you be sure that no one else is using the multicast addresses
              on the network for anything else - we had someone bring up a test
              cluster using our addresses which caused a few issues and took a while
              to find! We also have security cameras which also use multicast, which
              if they are using the same address/port can cause issues!
              We're using 239.192.1.4:8001 for one cluster and 239.192.1.3:7001 for
              the other - I think it's best to keep those as different as you can.
              In 8.1, there is also the multicast monitor utility - there's a support
              pattern on e-support on how to diagnose it. I've found this useful in
              the past when I've suspected a cluster issue.
              https://support.bea.com/application_content/product_portlets/support_patterns/wls/MulticastErrorsPattern.html
              Check also that you're using a valid range for the address - we weren't
              for a while and had odd problems from time to time.
              There are also cluster debug flags available which you'll see listed in
              the support document.
              Are you seeing dropped multicast packets?
              Hope that helps.
              Pete

  • Unable to Validate a Cluster Configuration. The operation has failed. The action validate a configuration did not complete .

    There is an error in XML document (5, 73).  
    Attempt by method
    Microsoft.Xml.Serialzation.GeneratedAssembly.XmlSerialzationReaderClusterPrep.Config.Read4_As...Bolean) to access method
    MS.Internal.ServerClusters.Validation.TestAssemblyCollection.Add(MS.Internal.ServerClusters.V....Failed

    Resolution:
    1.) Copy the 2 Config Files attached to $PSHome (typically C:\Windows\System32\WindowsPowerShell\v1.0) – Done !
    2.) Copy powershell.exe to PS.exe and create a config file
    - cd $PSHome (which is the path where the powershell binaries live)
    - copy powershell.exe PS.exe
    - create a config file PS.exe.config with the content below
    Now you have the possibility to open a “normal” PS 4.0 Window and run the following command: 
    ps.exe –command “Test-Cluster –li”
    You can run this command from a powershell script as well.
    This way you don´t touch your existing PS4.0 environment and don´t need to do extra testing of all scripts, but can run the Test-Cluster cmd-let on the clusters as expected.
    More information:
    Powershell.exe.config will be read, if a Powershell Window is opened.
    Powershell_ise.exe.config will be read if the „Integrated Scripting Environment“ Powershell_ise.exe is opened.
    The Content of the Config Files is as follows and can be edited with notepad
    <?xml
    version="1.0"
    encoding="utf-8" ?>
    <configuration>
     <system.xml.serialization>
       <xmlSerializer
    useLegacySerializerGeneration="true"/>
     </system.xml.serialization>
    </configuration>
    In the config file powershell is advised to execute the legacy serialization code as it has been done in Powershell 2.0.
    If you set the value of the variable <xmlSerializer
    useLegacySerializerGeneration="false"/>
    to false,
    powershell 4.0 would behave as if you would not have created a config file.

  • Testing after Upgrade

    Is there any Testing to do done by Basis Consultant after Upgradation. I guess we technical people doesnt have thing to do in testing. If yes please let me provide few guides.
    Thanks & Regaards
    Raghu Nandana Rao. S

    some example then
    - test backup recovery again
    - If there was a lot customer background jobs before, check all processes still operating correctly.
    - role and profile --> if not done already
    - test cluster again if used before
    - all suggested post processing included in formal guide from SAP marketplace
    - depending on people, some do the tasks using OS scheduler, if yes check them also

Maybe you are looking for

  • Application crashing regularly

    Hi, Strange issue since months on my MacBookPro. During my Mac utilisation, applications are regularly crashing (one at a time). It happens with Cord, Skype, Safari, Firefox, ITunes, ... Today (I am working from 1 hour) and Safari crashes 4 times alr

  • GE70 2OE: How to successfully update to Windows 8.1

    Hi all.  I have successfully upgraded my MSI GE70 2OE (non-raid sku) to Windows 8.1.  It is a little bit of a headache, but I will cover what I did in order to get it to work*. 1) Download Windows 8.1 from the Windows Store.  It will download and aut

  • Status of an action PPF

    Hi, I used an action definition and it is called successfully when I save a document. But the status of this action is generated in error (red icon) although rp_status = 0 in my implementation. I want to make this action as unprocessed (Yellow icon).

  • Error Trying to Create Capture Media Error 80004005

    Hey everyone need some help here.  I was at TechEd last week, in which I did the hands-on-lab on using boot media to capture a reference PC, can't remember the name.  Anyway, here is my problem, I tried to do the exact same set up on our setup and it

  • My Macbook pro will not power on.

    I shut my screen to move to the nearest outlet, and once I plugged in the computer and re-opened the screen the computer would not turn on.  It shows no signs of activity when I press the power button.  I have tried resetting the SMC with no results.