J_security_check not working in a cluster

          We have a cluster set up using two machines (Sun fire 280R) two instances of weblogic
          on each machine (and a third for admin on machine 1) the cluster starts up fine,
          admin server taken off line, the web app (.war) is distributed and accessable
          across the cluster. But when a user tries to log in (using FORM based auth, posting
          to j_security_check) nothing happens, just returned to the login page, no errors
          no output no nothing. If we remove the cluster and just run the app on a single
          server (the admin server) it works fine. Even if we run the app on another managed
          server (not part of a cluster) it still doesn't work. It only works on the admin
          server, we are using a RDBMS Realm to authenticate, we have checked the obvious,
          DB Pool deployed to cluster etc.
          Replies very much appriceated
          Thanks
          Gareth
          

Hello Gareth,
          Iam facing exact same issue as well. Its a WL 6.1 sp2 cluster. Iam
          using HttpClusterServlet on WL as proxy. Also using RDMBS realm for
          Authentication. Like you said after I type userid/password click o.k
          nothing happens. No errors no messages! its been frustrating as I was
          not sure if my configuration was right or not! Do let me know if you
          should find any solution to this!
          thanks!
          -priya
          p.s: I also notice that if i keep trying 3, 4 or 5 times and finally
          do get in, the httpSession is all confused. Some time its starts 2
          sessions! Sometimes is does not replicate properly. And most of the
          time if I restart the cluster server it goes works for once or twice
          and does'nt work after that!
          "Gareth James" <[email protected]> wrote in message news:<[email protected]>...
          > We have a cluster set up using two machines (Sun fire 280R) two instances of weblogic
          > on each machine (and a third for admin on machine 1) the cluster starts up fine,
          > admin server taken off line, the web app (.war) is distributed and accessable
          > across the cluster. But when a user tries to log in (using FORM based auth, posting
          > to j_security_check) nothing happens, just returned to the login page, no errors
          > no output no nothing. If we remove the cluster and just run the app on a single
          > server (the admin server) it works fine. Even if we run the app on another managed
          > server (not part of a cluster) it still doesn't work. It only works on the admin
          > server, we are using a RDBMS Realm to authenticate, we have checked the obvious,
          > DB Pool deployed to cluster etc.
          >
          > Replies very much appriceated
          >
          > Thanks
          >
          > Gareth
          

Similar Messages

  • Failover not working in a cluster environment asks Relogin to Application

    Hi,
    I have setup cluster on weblogic and deployed CRM (ejb/ear) application.
    When I stop managed server which is handling request other managed server picks the request but it asks for relogin.
    I think failover is not working properly.
    Could you please help?
    Thanks,
    Vishal

    Hi Kalyan,
    Please See below Diagnostics:
    Managed Server1:
    ===================================================
    Date: Tue Sep 04 18:33:12 IST 2012
    DdCompnt=11,SessionContext=7,CurrJvmHeapMB=235,Bos=2,JForm=1,MaxBoRowCount=493,JApp=13,JField=1,JDdField=1,JSession=6,Sessions=7,JBaseBo=2,Misc=74,JPTrace=1,LpmErrors=1,Forms=1,Active Sessions=6,JDdTable=1021,MaxJvmHeapMB=467,MaxBos=18,BoRowCount=33,JError=1,JFields=1,UtlRecs=175,FlexAttr=-5,UserContext=3
    Managed Server2:
    ===================================================
    Date: Tue Sep 04 18:33:20 IST 2012
    DdCompnt=11,SessionContext=8,CurrJvmHeapMB=226,JForm=1,JApp=13,JSession=7,Sessions=8,Misc=148,JPTrace=1,Forms=1,Active Sessions=7,JDdTable=1124,MaxJvmHeapMB=472,UtlRecs=81,FlexAttr=-5,UserContext=4
    Free Java Heap Memory: 258400 KB
    Total Java Heap Memory: 495779 KB
    I tried to open many sessions. Its assigning randomly, but when I kill Managed Server2, its corresponding sessions are not working.
    System Error error executing LPoms.getUserContext()
    Also I found following warning in Weblogic logs.
    <Sep 4, 2012 6:26:15 PM IST> <Warning> <Socket> <BEA-000402> <There are: 5 active sockets, but the maximum number of socket reader threads allowed by the configuration is: 4. The configuration may need altered.>
    Thanks for your immediate help.

  • CAMs does not work in a cluster

    Tell  me please, I have two Cams worked in a cluster on version 4.1.8, after  updating to version 4.8.0, interfaces, synchronization can not see each  other and the tab failover each CAM neighbor is not active (dead)
    Sincerely, Alexander

    Alexander,
    Check your certificates. Install the same certificate/private key pair on both CAMs.
    More details here:
    https://supportforums.cisco.com/docs/DOC-12609
    Also other certificate related documents are available here:
    https://supportforums.cisco.com/community/netpro/security/others?view=documents
    HTH,
    Faisal
    If you find this post helpful, please rate so others can find the answer easily

  • In-memory replication of http session is not working in BEA7 cluster

              Hi everyone,
              I have 3 managed servers in Bea7.0 SP4 in a cluster. The client requests are sent
              through apache web server. I have given cluster address as URL in httpd.conf of
              apache server which sends the client requests for dynamic pages such as JSPs and
              servlets to the weblogic cluster.
              Load balancing is working fine. I ensured this from the log files of all the 3
              servers. All the 3 servers are getting different client requests and thus load
              balancing is working.
              Now, I wanted to achieve Fail-over. I do not think that i should use proxy plug-in
              for this. I feel the cluster itself will handle fail-over provided i make the
              http session as memory replicated.
              I updated the weblogic.xml with the following entry :
              <session-descriptor>
              <param-name>PersistentStoreType</param-name>
              <param-value>replicated</param-value>
              </session-param>
              </session-descriptor>
              I guess this is sufficient to make the http session as cluster aware.
              But when I shutdown server1, the user connected to server1 will be kicked out
              of the session and come to login page through server2 or server3 which are running
              fine.
              Could anyone help me to achieve http session as cluster aware. Does it indicate
              that I have to go for WLS proxy – HttpClusterServlet to achieve fail over for
              http session ?
              BTW, for your info, i am using setAttribute() and getAttribute() while manipulating
              the session.
              thanks in advance.
              

              Hi Ryan,
              Thanks for ur valuable input.
              I can see failover working.
              But, I can not continue with the same session in my application.
              I printed session Ids before and after failover, I found both are different.
              I guess session replication is a responsibility of weblogic/apache plugin.
              If not please let me know which all settings I should do to make failover working?
              Thanks again.
              Plad
              "ryan upton" <ryanjupton at learningvoyage dot com> wrote:
              >Plad,
              >
              >Are you trying to gracefully shut down the server? If you are then the
              >problem that you say you can't identify is simply the server's default
              >behavior which is to wait for all non-replicated sessions to be dropped
              >or
              >timed out before killing the process. Try forcing the shutdown: kill
              >-9 the
              >PID or CTRL-C if you started the server from the command line. You can
              >also
              >check the ``Ignore Sessions During Shutdown" checkbox under the server's
              >control tab in the admin console, this should allow you to shut down
              >gracefully without waiting for session timeout. BTW your sequence is
              >off
              >in #5 below, the replication doesn't occur upon failure, the replication
              >has
              >already happened once you created the session object on the first server,
              >I
              >think maybe you're confusing replication with failover.
              >
              >~RU
              >
              >"Plad" <[email protected]> wrote in message
              >news:[email protected]...
              >>
              >> Hi,
              >> I have 2 managed servers in a cluster.
              >>
              >> 1. I have got a DNS name configured which maps to these 2 managed server's
              >IP
              >> addresses.
              >> 2. I can browse my site using this DNS name.
              >> In HTTPD.conf I have :
              >>
              >> ServerName dev.a.b.net
              >>
              >> <IfModule mod_weblogic.c>
              >> WebLogicCluster 10.1.38.232:7023,10.1.34.51:7023
              >> MatchExpression *.*
              >> </IfModule>
              >>
              >> LoadModule weblogic_module modules/mod_wl_20.so
              >>
              >> 3. I have adeded session descriptor in weblogic.xml , also enabled
              >proxy
              >plugin
              >> in weblogic console.
              >>
              >> 4. I tested accessing my application using DNS url after shutting down
              >alternatively
              >> each manaed server. I can access application.
              >>
              >> 5. Now, problem comes when I access a managed server1 , keeping server2
              >down.
              >> I am able to access my application.
              >> Now, I start the server2.
              >> (Here I am supposing that replication should occur)
              >> Then I am shutting down server1.
              >> But, this time the server log shows me following:
              >>
              >>
              >> 9:58:51 AM GMT+05:30 NOTICE Web application(s) chlist still have
              >non-replicated
              >> sessions after 2 minutes of initiating SUSPEND. Waiting for non-replicated
              >sessions
              >> to finish.
              >> 10:00:51 AM GMT+05:30 NOTICE Web application(s) chlist still have
              >non-replicated
              >> sessions after 4 minutes of initiating SUSPEND. Waiting for non-replicated
              >sessions
              >> to finish.
              >>
              >> I am unable to make out where the problem is?
              >> Can it be a problem of Liecense? Is there any specialcluster liecense
              >for
              >weblogic8?
              >>
              >> Hoping to get replies.
              >> Thanx.
              >> Plad
              >>
              >> "ryan upton" <ryanjupton at learningvoyage dot com> wrote:
              >> >See my reply to your first post, but I've also added a few comments
              >here.
              >> >
              >> >"jyothi" <[email protected]> wrote in message
              >> >news:[email protected]...
              >> >>
              >> >> I guess someone from bea support team only can answer both your
              >question
              >> >and mine.
              >> >> As per my knowledge, we do not need to do any setup at Apache
              >side
              >> >regarding
              >> >> cluster other than mentioning cluster address as URL while
              >contacting
              >> >WLS
              >> >> from apache.
              >> >>
              >> >> I hope someone from Bea, will help us. I do not think that we
              >> >go for
              >> >WLS
              >> >> proxy plug-in using HttpClusterServlet for making session replication.
              >> > I
              >> >strongly
              >> >> feel that the cluster itself be able to manage the fail-over of
              >> >http
              >> >sessions
              >> >> provided we put the entry "PersistentStoreType" in weblogic.xml
              >> >regarding
              >> >> the session replication.
              >> >>
              >> >
              >> >The cluster does handle the management of Sessions. The clustered
              >> >applications still create the Session objects and the cluster manages
              >> >them
              >> >as per your deployment descriptor settings (replicated, JDBC, File)
              >however
              >> >the proxy has to be aware of which server the client has an affinity
              >> >for
              >> >(only with replicated sessions) and it does that by reading a cookie
              >> >passed
              >> >back from the server that handled the initial request and created
              >the
              >> >primary session object. The proxy has a list of both the primary
              >and
              >> >secondary server locations from this cookie that it can use to failover
              >> >the
              >> >request if the primary server fails. Clusters _DO NOT_ failover nor
              >> >do they
              >> >load balance, that's the job of your proxy, whether you're using the
              >> >HTTPClusterServlet, WLS Plug-in or a more sophisticated hardware load
              >> >balancer like Big IPs F5
              >> >
              >> >> jyothi
              >> >>
              >> >
              >> >~RU
              >> >
              >> >
              >>
              >
              >
              

  • Event in maintainance view not working in View cluster

    Hello All,
                I have created a view for a student table, which contains three fields. The third field should be automatically incremented, based on the entries in the first two fields. The functionality is working fine in the view created, as i have used the event01 in
    table maintainance generator->environment->events; with the codes for incrementing the third field automatically.
               After getting this functionality working, i have created a view cluster with the particular view also.
    The problem now is that, when i try to enter values to the table via view in the view cluster, the same event is not functioning for auto-incrementing. While saving the entries, the third field is remaining blank.
    The internal table 'total' which i am using in the view is displaying the table entries in the view created, but the same table is showing null value in the view cluster.
    Can anybody guide me in this?
    Thanks in Advance,
    Shino.

    Hi Shino,
    Let`s say you have a table ZTEST_TABLE1. You create a maintenance view on it ZTEST_VIEW1. You generate the maintenance dialog for the view with Function Group ZTEST_FUGR1. After this you will be able to maintain the table from SM30 using view ZTEST_VIEW1.
    You create view cluster ZTEST_CLUSTER1 and include your view ZTEST_VIEW1 in it. After this you can maintain the table from the view cluster as well. You then create a subroutine pool for the event routines like ZTEST_PROG1. In this program you include the program I mentioned: INCLUDE lsvcmcod.
    Now you create the event 04 - Before saving the data in the database. You specify the routine name eg BEFORE_SAVE and assign the subroutine pool as the main program. There you create the FORM. Example code below.
    I suggest to check the SAP Library, there you`ll find more details as well.
    http://help.sap.com/saphelp_nw04/helpdata/en/62/c302c7de8e11d1a5960000e82deaaa/frameset.htm
    FORM before_save.
      DATA: viewname TYPE vclstruc-object,
            error_flag TYPE vcl_flag_type,
            header TYPE vimdesc,
            wa_total(1000) TYPE c,
            wa_extract(1000) TYPE c.
      FIELD-SYMBOLS: <view>    TYPE ANY,
                     <view_x>  TYPE x,
                     <ent_x>   TYPE x,
                     <viewkey> TYPE x.
    " assign data of current view to global field symbols (<vcl_total>, <vcl_extract> etc.)
      viewname = 'ZTEST_VIEW1'.
      PERFORM vcl_set_table_access_for_obj USING    viewname
                                           CHANGING error_flag.
    " get view name (if you`ll use several views to make it dynamic)
      READ TABLE <vcl_header> INTO header INDEX 1.
    " assign field symbols
      ASSIGN: wa_total                TO <view>    CASTING TYPE (header-maintview),
              <view>                  TO <view_x>  CASTING,
              <view_x>(header-keylen) TO <viewkey>.
    " <vcl_extract> is not sorted
      SORT <vcl_extract>.
      LOOP AT <vcl_total> INTO wa_total.
    " your checks and modifications on the record using <fs_view>
    "    IF <fs_view>-fieldxy...
    " apply the changes to the EXTRACT table as well
        READ TABLE <vcl_extract> INTO wa_extract WITH KEY <viewkey> BINARY SEARCH.
        IF sy-subrc = 0.
    "     apply your changes to <vcl_extract>
        ENDIF.
      ENDLOOP.
    ENDFORM.                    "before_save

  • J_security_check not working

    Hi,
    I am using a simple login form to test the j_security functionality. The username
    is j_username , the password is j_password and the action url is j_security_check.
    However it is not authenticating any user id and password
    existing in my weblogic server. The weblogic server is wl6.0 sp2. All other configurations
    are default configurations. I am trying to test with those login and passwords
    , I what I can see through weblogic console(weblogic ->security-->users).
    Am I missing something ?
    Thx

    Hi,
    I am using a simple login form to test the j_security functionality. The username
    is j_username , the password is j_password and the action url is j_security_check.
    However it is not authenticating any user id and password
    existing in my weblogic server. The weblogic server is wl6.0 sp2. All other configurations
    are default configurations. I am trying to test with those login and passwords
    , I what I can see through weblogic console(weblogic ->security-->users).
    Am I missing something ?
    Thx

  • Apache Plugin is not working with Cluster Entry in httpd.conf

    Hi Guys,
    We have a very particular issue that is happening here in our production environments. Before explaining the issue to you, I will just give a brief of our architecture.
    We are using Weblogic Portal Platform SP4.
    Our current Production Domain Consists of 1 Admin Server and 2 Managed Servers. The managed servers are in a cluster.
    We have two Linux boxes on which the weblogic servers are running. These two machines have IPs: x.x.x.58 and x.x.x.59. Now, the Admin Server and managed Server 1 are started on the machine x.x.x.58 and managed server 2 is started on the machine x.x.x.59. Both the managedServers listen on port 8007.
    This entire setup mentioned in the above paragraph is at a location A (lets say).
    Now, the requests are proxied to the managed server cluster using BEA's Apache plugin residing on Apache. Apache is running on two Linux boxes that are present at a location B.
    Location B is connected to Location A by a dedicated WAN Link.
    Now, before the two Apache servers there is a physical Load Balancer that balances the requests between the two Apache boxes.
    This means that whenever there is a request from the Internet to the weblogic cluster, the load balancer first sends it to either of the Apache servers and then the BEA Apache plugin on that Apache forwards the request via the 2 mbps WAN link to the weblogic cluster.
    Now following is the description of the issue we are facing. We donot know whether the issue is with the plugin or with weblogic server.
    In order that the plugin load balances the requests between the servers in the cluster and for failover, the following entry was put in httpd.conf on both the Apache Servers.
    WebLogicCluster y.y.y.11:8007,y.y.y.13:8007
    Note: Because of the WAN link there is natting used. This implies y.y.y.11 is natted to x.x.x.58 and y.y.y.13 is natted to x.x.x.59.
    Here in lies the issue. When we use this plugin entry the following issue occurs.
    1.     First we go to the home page and login.
    2.     The user is then redirected to a Portal Desktop fitting those credentials.
    3.     Now on this Portal Desktop, invariably whenever we click any of the desktop's links, an HTTP 403 is thrown. This error is thrown because on the desktop there are entitlements so that it cannot be accessed without a user logging in with credentials. But this error indicates that the server is losing the session and entitlement details.
    Why does this happen?
    As a result, we cannot use that plugin entry. The current entry we have is
    WebLogicHost y.y.y.11
    WebLogicPort 8007
    As you can see, here we are only redirecting to one managed server. This entry is not what we want, obviously because if that server fails, the plugin wont know which other server to pass the request on to. Also in this case there is no load balancing.
    Our issues our,
    1. How to resolve this plugin entry issue. How can we use the cluster entry without encountering this HTTP 403 Error?
    3. Consider the following sequence of steps.
    - Request for the home page is served by Apache Server 1
    - User logins in, the login request is sent by the Load Balancer to Apache Server 1 and plugin on Apache Server 1 sends it to the cluster via WAN l ink. Then the internal mechanism works with creating the session id and primary server and secondary server and logs the user in and sends the response via Apache Server 1.
    - Now, the user clicks a link on the desktop. The request now goes to the load balancer and it sends it to Apache Server 2.
    The question is now does the plugin on Apache server 2 correctly understand and send the request to the primary server of that session.
    Pls help me as to giving me some direction in resolving this issue

    From your Xorg.0.log:
    (II) fglrx(0): DRIScreenInit done
    (II) fglrx(0): Kernel Module Version Information:
    (II) fglrx(0): Name: fglrx
    (II) fglrx(0): Version: 8.49.7
    (II) fglrx(0): Date: May 12 2008
    (II) fglrx(0): Desc: ATI FireGL DRM kernel module
    (WW) fglrx(0): Kernel Module version does *not* match driver.
    (EE) fglrx(0): incompatible kernel module detected - HW accelerated OpenGL will not work
    You have to quit X, then rmmod and modprobe the kernel module - you're still using the old one.  (a reboot would technically fix it too)
    # rmmod fglrx
    # modprobe fglrx
    # dmesg | tail
    The dmesg output should say you've successfully loaded the 8.501 driver.

  • When setting up converged network in VMM cluster and live migration virtual nics not working

    Hello Everyone,
    I am having issues setting up converged network in VMM.  I have been working with MS engineers to no avail.  I am very surprised with the expertise of the MS engineers.  They had no idea what a converged network even was.  I had way more
    experience then these guys and they said there was no escalation track so I am posting here in hopes of getting some assistance.
    Everyone including our consultants says my setup is correct. 
    What I want to do:
    I have servers with 5 nics and want to use 3 of the nics for a team and then configure cluster, live migration and host management as virtual network adapters.  I have created all my logical networks, port profile with the uplink defined as team and
    networks selected.  Created logical switch and associated portprofle.  When I deploy logical switch and create virtual network adapters the logical switch works for VMs and my management nic works as well.  Problem is that the cluster and live
    migration virtual nics do not work.  The correct Vlans get pulled in for the corresponding networks and If I run get-vmnetworkadaptervlan it shows cluster and live migration in vlans 14 and 15 which is correct.  However nics do not work at all.
    I finally decided to do this via the host in powershell and everything works fine which means this is definitely an issue with VMM.  I then imported host into VMM again but now I cannot use any of the objects I created and VMM and have to use standard
    switch.
    I am really losing faith in VMM fast. 
    Hosts are 2012 R2 and VMM is 2012 R2 all fresh builds with latest drivers
    Thanks

    Have you checked our whitepaper http://gallery.technet.microsoft.com/Hybrid-Cloud-with-NVGRE-aa6e1e9a for how to configure this through VMM?
    Are you using static IP address assignment for those vNICs?
    Are you sure your are teaming the correct physical adapters where the VLANs are trunked through the connected ports?
    Note; if you create the teaming configuration outside of VMM, and then import the hosts to VMM, then VMM will not recognize the configuration. 
    The details should be all in this whitepaper.
    -kn
    Kristian (Virtualization and some coffee: http://kristiannese.blogspot.com )

  • SAP ECC 6.0 SR3 Cluster failover not working in AIX with DB2 UDB V9.1 FP6

    Hi Gurus,
    We have installed the SAP ECC 6.0 SR3 High Availability  with DB2 UDB V9.1 FP6 in AIX cluster environment.
    After installation we are doing the cluster fail test.
    Node A
    Application Server
    Mount Points:
    /sapmnt/<SID>
    /usr/sap/<SID>
    /usr/sap/trans
    Node B
    Database Server
    Mount Points:
    /db2//<SID>
    The procedure followed to do the cluster failover:
    We have down the cluster on Node A and all the resources of the Node A has been moved to Node B.
    On Node B when we issued a command to start the SAP. It says u201Cno start profiles foundu201D
    WE have down the cluster on Node B and  moved the Resource from Node B to Node A .  There the db2 User IDu2019s are not available. We have crated the user Idu2019s manually on Node A. however it did not work.
    Please suggest the procedure to start the sap in cluster failover.
    Best Regards
    Sija

    Hi Sija,
    Can i have detailed scenario in your cluster configuration.
    Means you are saying that going to start cluster package manually, if it is right please make sure that you had the same copy of start, instance profiles of NodeA to Node B. Means you need to maintain two startup, two instance profiles for both nodes. In a normal situation it will picik the profile of node A to start databse from A node. But in a failover situation it will not pick node A profile to start, it should pick Node B s profiles.
    Just make a copy from node A and change the profile name accordingly to Node b. Then try to restart.
    Regards
    Nick Loy

  • Cluster Example shipped with 7.0 SP4 does not work !!!

              I have tried and tried to setup a cluster example using the instructions shipped
              in the cluster example and it does not work. When I call the Session.jsp via
              the proxy URL I get the following output in the proxy log.
              <Mon Apr 26 08:49:39 BST 2004>: Start connection timeout scheduler
              <Mon Apr 26 08:49:39 BST 2004>: GenericProxyServelt: init()
              <Mon Apr 26 08:49:39 BST 2004>: HttpClusterServlet:init()
              <Mon Apr 26 08:49:43 BST 2004>: ===New Request===GET /ClusterTest/Session.jsp
              HT
              TP/1.1
              <Mon Apr 26 08:49:43 BST 2004>: Found session from url: null
              <Mon Apr 26 08:49:43 BST 2004>: attempt #0 out of a max of 5
              <Mon Apr 26 08:49:43 BST 2004>: #### Trying to connect with server cl1!9001!443
              <Mon Apr 26 08:49:43 BST 2004>: Failed to connect: cl1!9001!443
              <Mon Apr 26 08:49:43 BST 2004>: marked bad: cl1!9001!443
              <Mon Apr 26 08:49:43 BST 2004>: attempt #0 out of a max of 5
              <Mon Apr 26 08:49:43 BST 2004>: #### Trying to connect with server cl2!9002!443
              <Mon Apr 26 08:49:43 BST 2004>: Failed to connect: cl2!9002!443
              <Mon Apr 26 08:49:43 BST 2004>: marked bad: cl2!9002!443
              <Mon Apr 26 08:49:43 BST 2004>: attempt #0 out of a max of 5
              <Mon Apr 26 08:49:43 BST 2004>: === whole list is bad, reverting to the all serv
              ers list ===
              <Mon Apr 26 08:49:43 BST 2004>: #### Trying to connect with server cl1!9001!443
              <Mon Apr 26 08:49:43 BST 2004>: Failed to connect: cl1!9001!443
              <Mon Apr 26 08:49:43 BST 2004>: marked bad: cl1!9001!443
              <Mon Apr 26 08:49:43 BST 2004>: Sleeping for 2 secs .....
              <Mon Apr 26 08:49:45 BST 2004>: attempt #1 out of a max of 5
              <Mon Apr 26 08:49:45 BST 2004>: #### Trying to connect with server cl2!9002!443
              <Mon Apr 26 08:49:45 BST 2004>: Failed to connect: cl2!9002!443
              <Mon Apr 26 08:49:45 BST 2004>: marked bad: cl2!9002!443
              <Mon Apr 26 08:49:45 BST 2004>: Sleeping for 2 secs .....
              <Mon Apr 26 08:49:47 BST 2004>: attempt #2 out of a max of 5
              <Mon Apr 26 08:49:47 BST 2004>: === whole list is bad, reverting to the all serv
              ers list ===
              <Mon Apr 26 08:49:47 BST 2004>: #### Trying to connect with server cl2!9002!443
              <Mon Apr 26 08:49:47 BST 2004>: Failed to connect: cl2!9002!443
              <Mon Apr 26 08:49:47 BST 2004>: marked bad: cl2!9002!443
              <Mon Apr 26 08:49:47 BST 2004>: Sleeping for 2 secs .....
              <Mon Apr 26 08:49:49 BST 2004>: attempt #3 out of a max of 5
              <Mon Apr 26 08:49:49 BST 2004>: #### Trying to connect with server cl1!9001!443
              <Mon Apr 26 08:49:49 BST 2004>: Failed to connect: cl1!9001!443
              <Mon Apr 26 08:49:49 BST 2004>: marked bad: cl1!9001!443
              <Mon Apr 26 08:49:49 BST 2004>: Sleeping for 2 secs .....
              <Mon Apr 26 08:49:51 BST 2004>: attempt #4 out of a max of 5
              <Mon Apr 26 08:50:17 BST 2004>: #### Trying to connect with server cl2!9002!443
              <Mon Apr 26 08:50:17 BST 2004>: Failed to connect: cl2!9002!443
              <Mon Apr 26 08:50:17 BST 2004>: marked bad: cl2!9002!443
              <Mon Apr 26 08:50:17 BST 2004>: Sleeping for 2 secs .....
              <Mon Apr 26 08:50:19 BST 2004>: attempt #5 out of a max of 5
              <Mon Apr 26 08:50:19 BST 2004>: #### Trying to connect with server cl1!9001!443
              <Mon Apr 26 08:50:19 BST 2004>: Failed to connect: cl1!9001!443
              <Mon Apr 26 08:50:19 BST 2004>: marked bad: cl1!9001!443
              <Mon Apr 26 08:50:19 BST 2004>: Tried all servers but didn't succeed
              Any ideas ?
              For starters I think the documentation is wrong.
              It states
              servlet weblogic.servlet.internal.HttpClusterServlet
              yet in the weblogic upgrade 6.0 to 7.0 documentation it says replace this with
              weblogic.servlet.proxy.HttpClusterServlet
              Thanks
              Justin
              

    Have you tried deleting the iDVD preference file, com.apple. iDVD.plist, that resides in your User/Library/Preferences folder?
    OT

  • WL Cluster: install of weblogic is not working properly

    Cluster install of weblogic is not working properly. We are getting kicked out every now and then
    from the portal when both PIA1 and PIA2 nodes are up. When we shut down one meaning switch off clustering
    then we are fine(Portal becomes very stable).
    env setup:
    Weblogic 9.2 with MP1-p5 and all the patches on Oracle Enterprise Linux 4 32 bit
    PTools version - 8.49.08
    Here is the message at the Kernel level if that any help:
    Oct 7 10:43:22 hrrecweb01 kernel: Badness in local_bh_enable at kernel/softirq. c:141
    Oct 7 10:43:23 hrrecweb01 kernel: [<c0121178>] local_bh_enable+0x3f/0x62
    Oct 7 10:43:23 hrrecweb01 kernel: [<c02177b9>] skb_checksum+0x133/0x25e
    Oct 7 10:43:23 hrrecweb01 kernel: [<c0250f36>] udp_poll+0x66/0x113
    Oct 7 10:43:23 hrrecweb01 kernel: [<c02135e1>] sock_poll+0x19/0x1d
    Oct 7 10:43:23 hrrecweb01 kernel: [<c016d777>] do_pollfd+0x42/0x7d
    HRRECWEB01 is WebServer name
    any thoughts? Is there a kernel variable or settings for linux that I need to set for
    cluster weblogic install?
    Thanks/Thaya

    Kunal,
    Have you tried clearing the browser cache?
    -Arun

  • WLS6sp2 Hot deployment in cluster does not work

              I have two WLS6sp2 servers in cluster in two different machines. There is an EJB
              deployed in the cluster that works fine. Now I go to the console and I undeploy
              it: the servers show a message of "ejb undeployed". Now I activate the "deploy"
              checkbox again: the following error appears:
              weblogic.ejb20.EJBDeploymentException: Error deploying stateless EJB; nested exception
              is:
              javax.naming.NameAlreadyBoundException: Can't rebind anything but a replica-aware
              stub to a name that is currently bound to a replica-aware stub; remaining name
              javax.naming.NameAlreadyBoundException: Can't rebind anything but a replica-aware
              stub to a name that is currently bound to a replica-aware stub; remaining name
              The object is obviously clusterable (it is a stateless EJB with "home-is-clusterable=true").
              It has been deployed to the cluster target. When is working it is load balanced
              between both servers. It deploys perfectly when starting up both servers, but
              hot deployment does not work.
              This works fine when only one server is up.
              ¿Is not possible to do hot deployment in cluster?
              

    Hm, where are you seeing the javax.naming.NameAlreadyBoundException? In one of your managed
              servers, in all of your managed server, or in your admin server?
              If you see this in your admin server, then this means you got your admin joined in the cluster,
              which is not good. Remove it from the cluster if this is so.
              Gene
              "David Agudo" <[email protected]> wrote in message news:[email protected]...
              >
              > Yes, target="mycluster".
              >
              > "Gene Chuang" <[email protected]> wrote:
              > >Did you target the ejb to the cluster, or to each individual Node in
              > >your cluster? There is a
              > >difference...
              > >
              > >Gene
              > >
              > >"David Agudo" <[email protected]> wrote in message news:[email protected]...
              > >>
              > >> I have two WLS6sp2 servers in cluster in two different machines. There
              > >is an EJB
              > >> deployed in the cluster that works fine. Now I go to the console and
              > >I undeploy
              > >> it: the servers show a message of "ejb undeployed". Now I activate
              > >the "deploy"
              > >> checkbox again: the following error appears:
              > >>
              > >> weblogic.ejb20.EJBDeploymentException: Error deploying stateless EJB;
              > >nested exception
              > >> is:
              > >> javax.naming.NameAlreadyBoundException: Can't rebind anything but a
              > >replica-aware
              > >> stub to a name that is currently bound to a replica-aware stub; remaining
              > >name
              > >> ''
              > >> javax.naming.NameAlreadyBoundException: Can't rebind anything but a
              > >replica-aware
              > >> stub to a name that is currently bound to a replica-aware stub; remaining
              > >name
              > >> ''
              > >>
              > >> The object is obviously clusterable (it is a stateless EJB with "home-is-clusterable=true").
              > >> It has been deployed to the cluster target. When is working it is load
              > >balanced
              > >> between both servers. It deploys perfectly when starting up both servers,
              > >but
              > >> hot deployment does not work.
              > >>
              > >> This works fine when only one server is up.
              > >>
              > >> ¿Is not possible to do hot deployment in cluster?
              > >>
              > >
              > >
              >
              

  • Cluster does not work after a while

    Hi :
              I have 2 clustered EJB server, the IP address is 192.168.0.226 and
              192.168.0.227.
              A servlet server is calling these two EJB servers with
              t3://192.168.0.226,192.168.0.227:7001
              All three machine use weblogic 5.1 sp9, Win2K Advanced Server.
              If I restart weblogic service, then the cluster works fine. but after while
              the cluster does not work, all the client request goto only one machine, in
              most case it's 227.
              It has no use even if I restart the service in the losted server. I have to
              restart all the server.
              The weblogic.log in the losted server is like this :
              Thu Apr 26 00:04:16 GMT 2001:<I> <RJVM> Signaling
              peer -6817319611378695685S192.168.0.204:[7001,7001,7002,7002,7001,-1] gone:
              weblogic.rjvm.PeerGoneException:
              - with nested exception:
              [java.io.EOFException]
              Thu Apr 26 00:04:47 GMT 2001:<I> <RJVM> Signaling
              peer -2123734719233546013S192.168.0.227:[7001,7001,7002,7002,7001,-1] gone:
              weblogic.rjvm.PeerGoneException:
              - with nested exception:
              [java.io.EOFException]
              Thu Apr 26 00:14:16 GMT 2001:<E> <ConMan> Attempt to sendMsg using a closed
              connection
              Thu Apr 26 00:14:16 GMT 2001:<E> <RJVM> Exception on send :
              weblogic.rmi.ConnectException: Attempt to sendMsg using a closed connection
              Please help me
              thanks
              andrew
              

    Are you running with any proxy? Is the servlet server also the static httpd
              server?
              What type of network equipment are you using? If its high-end (Cisco) make sure
              all ports are set to 100/Full.
              Also, try setting the SendDelay slightly higher (ie. 25, 30, 35). Experiment
              with different values.
              weblogic.cluster.multicastSendDelay=25
              andrew wrote:
              > Mike, thanks for reply
              > The 192.168.0.204 is the servlet server.
              > I changed the NICs to use 100Mbps/Full duplex. it looks better: the error
              > msg is less than before, the it still happened.
              > Any suggestion?
              > thanks
              > andrew
              >
              > "Mike Kincer" <[email protected]> wrote in message
              > news:[email protected]...
              > > Which box is 192.168.0.204 ??
              > > It is communicating via multicast, which I would say is suspect.
              > > Otherwise, I'd say you have some network issues.
              > > Make sure all switch ports and all NICS are NOT set to "auto" config.
              > Select
              > > 100Mbps/Full duplex on all ports.
              > >
              > > andrew wrote:
              > >
              > > > Hi :
              > > > I have 2 clustered EJB server, the IP address is 192.168.0.226 and
              > > > 192.168.0.227.
              > > > A servlet server is calling these two EJB servers with
              > > > t3://192.168.0.226,192.168.0.227:7001
              > > >
              > > > All three machine use weblogic 5.1 sp9, Win2K Advanced Server.
              > > >
              > > > If I restart weblogic service, then the cluster works fine. but after
              > while
              > > > the cluster does not work, all the client request goto only one machine,
              > in
              > > > most case it's 227.
              > > >
              > > > It has no use even if I restart the service in the losted server. I have
              > to
              > > > restart all the server.
              > > >
              > > > The weblogic.log in the losted server is like this :
              > > >
              > > > Thu Apr 26 00:04:16 GMT 2001:<I> <RJVM> Signaling
              > > > peer -6817319611378695685S192.168.0.204:[7001,7001,7002,7002,7001,-1]
              > gone:
              > > > weblogic.rjvm.PeerGoneException:
              > > > - with nested exception:
              > > > [java.io.EOFException]
              > > > Thu Apr 26 00:04:47 GMT 2001:<I> <RJVM> Signaling
              > > > peer -2123734719233546013S192.168.0.227:[7001,7001,7002,7002,7001,-1]
              > gone:
              > > > weblogic.rjvm.PeerGoneException:
              > > > - with nested exception:
              > > > [java.io.EOFException]
              > > > Thu Apr 26 00:14:16 GMT 2001:<E> <ConMan> Attempt to sendMsg using a
              > closed
              > > > connection
              > > > Thu Apr 26 00:14:16 GMT 2001:<E> <RJVM> Exception on send :
              > > > weblogic.rmi.ConnectException: Attempt to sendMsg using a closed
              > connection
              > > >
              > > > Please help me
              > > > thanks
              > > > andrew
              > >
              > > --
              > > /\/\i|<e
              > >
              > > Mike Kincer
              > > Solutions Developer/Engineer
              > > Atlas Commerce "ebusiness evolved"
              > > Office phone: +1-607-741-8877
              > > mailto:[email protected] [http://www.atlascommerce.com]
              > >
              > >
              /\/\i|<e
              Mike Kincer
              Solutions Developer/Engineer
              Atlas Commerce "ebusiness evolved"
              Office phone: +1-607-741-8877
              mailto:[email protected] [http://www.atlascommerce.com]
              

  • Sun Cluster and EMC AX150 does not work

    Hi All,
    I tried to setup a Sun Cluster 3.2 (2x T2000) with an EMC AX 150 Storage. the device paths are highly instable and breaking down time and again. Tried to install Powerpath software of EMC but this time the did-devices are not recognized by Sun Cluster.
    Any help will be appriciated.
    P.S: does anybody know whether a MSA 1000 will work with Sun Cluster 3.2.
    Regards

    Looking at the EMC link on http://www.sun.com/software/cluster/osp/ it doesn't look like these storage systems have been tested by EMC and so may or may not work.
    You could try using Sun's MPxIO rather than PowerPath. MPxIO is included in Solaris for free. Again, it may or may not work. You are unlikely to get support on either configuration until EMC has done the prerequisite engineering testing to resolve the sort of issues you are currently seeing.
    Regards,
    Tim
    ---

  • Cluster Health Monitor Links not working

    Is anyone else having an issue downloading the cluster health monitor?
    I keep receiving content not found. I have a pre 11.2 database that I would like to give this a test run on.
    Thank you

    user7002963 wrote:
    Is anyone else having an issue downloading the cluster health monitor?
    I keep receiving content not found. I have a pre 11.2 database that I would like to give this a test run on.
    Thank youThats right, i tried and its not working.
    Possible workaround: Call oracle global support. You'll get the link in google.
    Or
    Raise an SR with support.oracle.com if you have valid CSI
    The latest copy of Cluster Health Monitor (IPD/OS) is always with the install image if the version is 11.2.0.2 or greater

Maybe you are looking for

  • Server-side workflows against Service Link updated parameter/data

    Is there a way to attach any server-side workflows to service link inbound updates/data other than what is available via parameter mapping and prebuilt functions? I need to manipulate a piece of data that is being updated in to a field by service lin

  • HT204266 How do I submit a review of an app?

    I want to submit a review of an app? Where and how?

  • [SOLVED] Root partition wrongly assumed to be encrypted.

    Hi, I'm trying to install arch using newest iso image (2012.10.06) in a following manner: gpt partition on /dev/sda1 root on btrfs on /dev/sda2 var on btrfs on /dev/sda3 home on btrfs on /dev/mapper/crypt (LUKS) on /dev/sda4, but I have a problem aft

  • How to solve this error of NI report generation?

    The following error occurred when I program with the NI report genenration toolkit.How  can I solve the error? as the picture Attachments: 2.JPG ‏50 KB 1.JPG ‏91 KB

  • Logic vs. express

    I am new to digital recording. Currently using garageband with a cheap mixer I want to step up to express or logic. I want to do this for as little money as possible. The plan is to get a presonus firestudio package which includes the interface, MXL