ZFS ZIL and Clustering

Hi,
we are thinking about using a local RAM-Disks or local connected SSDs Disk for the ZIL, the other storage devices in the pool will be SAN disc connected via FC.
If we use this in a cluster (or use export / import to move the pool to another node in case of a node failure) we will loose the ZIL, eventually before its content has been commited to the SAN Disks.
So the question is, what happens to a ZFS pool if it losses its ZIL, specifically if we are storing Oracle data files and logs in the pool?
Fritz

You could lose data that has been declared commited to the application. The filesystem itself will still be consistent on disk, but could be out-of-date with respect to the application.
For a database, this seems particularly dangerous.
Darren

Similar Messages

  • Session Failover and Clustering

              Let's say that we have two WebServers (NES) with the weblogic plugin (say WS1 and WS2) and a cluster with two WebLogicCommerce AppServers (say AS1 and AS2). Let's assume that each WebServer and AppServer runs on its own machine (total: 4 machines). Now, let's assume that the WebServer "obj.conf" files (on both WS1 and WS2) are setup so that they point to the servers in the cluster (WebLogicCluster="AS1:7601,AS2:7601").
              When a new request comes in to one of the WebServers (say WS1), the plugin will route it to one of the AppServers using Round-Robin (say AS1). A session will now be initiated in AS1 and it sends a response back to the client.
              Question 1: How does the other proxy in WS2 know that all future requests for this client need to be forwarded to AS1?
              Question 2: For failover, does the cluster automatically replicate the session state existing in AS1 onto AS2 before sending the response (does AS2 automatically become the secondary)?
              Now let's assume that AS1 crashes/dies. When the next request from the client comes to WS1 or WS2, they will forward it to AS1 (assuming that WS2 knows about the client session in AS1) . Since AS1 has crashed, will the client eventually get a timeout error message?
              Question 3: To ensure that the session failover happens so that AS2 gets the request instead (becomes the primary), do we need to setup a WebLogic Proxy Server? If so, why can't the plugins for NES provide the failover themselves?
              Thank you very much for your help!
              Giri
              

              Thank you very much for your responses. It has been very helpful and I am clear on the session/clustering stuff. I have new questions on EJB and clustering which I will post as a separate thread.
              Giri
              "Jason Rosenberg" <[email protected]> wrote:
              >And also, if the browser has cookies disabled, it is important for
              >the app server to embed the WebLogicSession info via url rewriting,
              >otherwise the proxy or NES will not be able to route the session
              >properly.
              >
              >So, in all http responses, be sure to pass the url string through
              >response.encodeURL(). This will do the right thing depending
              >on whether cookies are enabled or not.
              >
              >I've just only recently figured this out. Haven't actually tried it
              >all out yet, so forgive me if it is not quite this simple, but this
              >seems to be the gist of it...
              >
              >Jason
              >
              >
              >"Justin James" <[email protected]> wrote in message news:[email protected]...
              >>
              >> Giri,
              >>
              >> I'm not a weblogic representative, but I tried to replicate this proxing service inside a load balancing switch(BigIP) and I
              >discovered a few things. The weblogic server sets a cookie (WebLogicSession)that the webserver plugin uses to manage the proxying.
              >The cookie (found in the HTTP header information) contains encoded information about the primary and secondary application servers
              >that the session is bound too. Any web server can read the cookie to determine how to dispatch the request to the primary server.
              >If the primary server does not respond, the request is forwarded to the secondary server by the plugin. Regardless of cluster size,
              >the session is replicated to only one other server.
              >>
              >> <[email protected]> wrote:
              >> >Giri Alwar wrote:
              >> >
              >> >> I need a couple of clarifications. First with regard to Question 1, I understand that plugins provide load balancing and
              >failover but what I really was asking is how the plugin in WS2 knows that a session for the client has been initiated in AS1 as a
              >result of WS1 sending the initial request to AS1. If WS2 gets a future request from the client, it needs to know this to send the
              >request to AS1. Does the plugin talk to the cluster to find out if there is a primary and who it is?
              >> >>
              >> >
              >> >> I should have clarified that my other questions pertain to in-memory replication. If I do not persist the session in a database
              >then does the client get an error message (timeout) when AS1 goes down (assuming we use NES with the WebLogic plugin)?
              >> >
              >> >Plugins' know how to route requests based on cookies. If it can't reach the primary server it will automatically try secondary.
              >In your case it doesn't matter if it reaches to proxy 1 or proxy 2, it is still the same.
              >> >
              >> >- Prasad
              >> >
              >> >> To prevent this error message and achieve failover, do I need to use WebLogic as the proxy server? If so, why isn't the NES
              >plugin doing this?
              >> >
              >> >> Thanks.
              >> >> Giri
              >> >>
              >> >> Prasad Peddada <[email protected]> wrote:
              >> >> >Giri Alwar wrote:
              >> >> >
              >> >> >> Let's say that we have two WebServers (NES) with the weblogic plugin (say WS1 and WS2) and a cluster with two
              >WebLogicCommerce AppServers (say AS1 and AS2). Let's assume that each WebServer and AppServer runs on its own machine (total: 4
              >machines). Now, let's assume that the WebServer "obj.conf" files (on both WS1 and WS2) are setup so that they point to the servers
              >in the cluster (WebLogicCluster="AS1:7601,AS2:7601").
              >> >> >>
              >> >> >> When a new request comes in to one of the WebServers (say WS1), the plugin will route it to one of the AppServers using
              >Round-Robin (say AS1). A session will now be initiated in AS1 and it sends a response back to the client.
              >> >> >>
              >> >> >> Question 1: How does the other proxy in WS2 know that all future requests for this client need to be forwarded to AS1?
              >> >> >
              >> >> > Plugin takes care of load balancing and failover, it is all transparent to the client.
              >> >> >
              >> >> >
              >> >> >> Question 2: For failover, does the cluster automatically replicate the session state existing in AS1 onto AS2 before sending
              >the response (does AS2 automatically become the secondary)?
              >> >> >
              >> >> > If you have only two yes it is automatically your secondary. Yes, replication is synchronous.
              >> >> >
              >> >> >>
              >> >> >> Now let's assume that AS1 crashes/dies. When the next request from the client comes to WS1 or WS2, they will forward it to
              >AS1 (assuming that WS2 knows about the client session in AS1) . Since AS1 has crashed, will the client eventually get a timeout
              >error message?
              >> >> >
              >> >> > If you are using some kind of persistence then you will be able to retrieve the session information and since the server
              >is not available the request will automatically failover.
              >> >> >
              >> >> >> Question 3: To ensure that the session failover happens so that AS2 gets the request instead (becomes the primary), do we
              >need to setup a WebLogic Proxy Server? If so, why can't the plugins for NES provide the failover themselves?
              >> >> >
              >> >> > No, you need only one. Either NES or weblogic proxy.
              >> >> >
              >> >> >
              >> >> >> Thank you very much for your help!
              >> >> >> Giri
              >> >> >
              >> >> >- Prasad
              >> >> >
              >> >
              >>
              >
              >
              

  • ZFS 7320c and T4-2 server mount points for NFS

    Hi All,
    We have an Oracle ZFS 7320c and T4-2 servers. Apart from the on-board 1 GB Ethernet, we also have a 10 Gbe connectivity between the servers and the storage
    configured as 10.0.0.0/16 network.
    We have created a few NFS shares but unable to mount them automatically after reboot inside Oracle VM Server for SPARC guest domains.
    The following document helped us in configuration:
    Configure and Mount NFS shares from SUN ZFS Storage 7320 for SPARC SuperCluster [ID 1503867.1]
    However, we can manually mount the file systems after reaching run level 3.
    The NFS mount points are /orabackup and /stage and the entries in /etc/vfstab are as follows:
    10.0.0.50:/export/orabackup - /orabackup nfs - yes rw,bg,hard,nointr,rsize=131072,wsize=131072,proto=tcp,vers=3
    10.0.0.50:/export/stage - /stage nfs - yes rw,bg,hard,nointr,rsize=131072,wsize=131072,proto=tcp,vers=3
    On the ZFS storage, the following are the properties for shares:
    zfsctrl1:shares> select nfs_prj1
    zfsctrl1:shares nfs_prj1> show
    Properties:
    aclinherit = restricted
    aclmode = discard
    atime = true
    checksum = fletcher4
    compression = off
    dedup = false
    compressratio = 100
    copies = 1
    creation = Sun Jan 27 2013 11:17:17 GMT+0000 (UTC)
    logbias = latency
    mountpoint = /export
    quota = 0
    readonly = false
    recordsize = 128K
    reservation = 0
    rstchown = true
    secondarycache = all
    nbmand = false
    sharesmb = off
    sharenfs = on
    snapdir = hidden
    vscan = false
    sharedav = off
    shareftp = off
    sharesftp = off
    sharetftp =
    pool = oocep_pool
    canonical_name = oocep_pool/local/nfs_prj1
    default_group = other
    default_permissions = 700
    default_sparse = false
    default_user = nobody
    default_volblocksize = 8K
    default_volsize = 0
    exported = true
    nodestroy = false
    space_data = 43.2G
    space_unused_res = 0
    space_unused_res_shares = 0
    space_snapshots = 0
    space_available = 3.97T
    space_total = 43.2G
    origin =
    Shares:
    Filesystems:
    NAME SIZE MOUNTPOINT
    orabackup 31K /export/orabackup
    stage 43.2G /export/stage
    Children:
    groups => View per-group usage and manage group
    quotas
    replication => Manage remote replication
    snapshots => Manage snapshots
    users => View per-user usage and manage user quotas
    zfsctrl1:shares nfs_prj1> select orabackup
    zfsctrl1:shares nfs_prj1/orabackup> show
    Properties:
    aclinherit = restricted (inherited)
    aclmode = discard (inherited)
    atime = true (inherited)
    casesensitivity = mixed
    checksum = fletcher4 (inherited)
    compression = off (inherited)
    dedup = false (inherited)
    compressratio = 100
    copies = 1 (inherited)
    creation = Sun Jan 27 2013 11:17:46 GMT+0000 (UTC)
    logbias = latency (inherited)
    mountpoint = /export/orabackup (inherited)
    normalization = none
    quota = 200G
    quota_snap = true
    readonly = false (inherited)
    recordsize = 128K (inherited)
    reservation = 0
    reservation_snap = true
    rstchown = true (inherited)
    secondarycache = all (inherited)
    shadow = none
    nbmand = false (inherited)
    sharesmb = off (inherited)
    sharenfs = sec=sys,rw,[email protected]/16:@10.0.0.218/16:@10.0.0.215/16:@10.0.0.212/16:@10.0.0.209/16:@10.0.0.206/16:@10.0.0.13/16:@10.0.0.200/16:@10.0.0.203/16
    snapdir = hidden (inherited)
    utf8only = true
    vscan = false (inherited)
    sharedav = off (inherited)
    shareftp = off (inherited)
    sharesftp = off (inherited)
    sharetftp = (inherited)
    pool = oocep_pool
    canonical_name = oocep_pool/local/nfs_prj1/orabackup
    exported = true (inherited)
    nodestroy = false
    space_data = 31K
    space_unused_res = 0
    space_snapshots = 0
    space_available = 200G
    space_total = 31K
    root_group = other
    root_permissions = 700
    root_user = nobody
    origin =
    zfsctrl1:shares nfs_prj1> select stage
    zfsctrl1:shares nfs_prj1/stage> show
    Properties:
    aclinherit = restricted (inherited)
    aclmode = discard (inherited)
    atime = true (inherited)
    casesensitivity = mixed
    checksum = fletcher4 (inherited)
    compression = off (inherited)
    dedup = false (inherited)
    compressratio = 100
    copies = 1 (inherited)
    creation = Tue Feb 12 2013 11:28:27 GMT+0000 (UTC)
    logbias = latency (inherited)
    mountpoint = /export/stage (inherited)
    normalization = none
    quota = 100G
    quota_snap = true
    readonly = false (inherited)
    recordsize = 128K (inherited)
    reservation = 0
    reservation_snap = true
    rstchown = true (inherited)
    secondarycache = all (inherited)
    shadow = none
    nbmand = false (inherited)
    sharesmb = off (inherited)
    sharenfs = sec=sys,rw,[email protected]/16:@10.0.0.218/16:@10.0.0.215/16:@10.0.0.212/16:@10.0.0.209/16:@10.0.0.206/16:@10.0.0.203/16:@10.0.0.200/16
    snapdir = hidden (inherited)
    utf8only = true
    vscan = false (inherited)
    sharedav = off (inherited)
    shareftp = off (inherited)
    sharesftp = off (inherited)
    sharetftp = (inherited)
    pool = oocep_pool
    canonical_name = oocep_pool/local/nfs_prj1/stage
    exported = true (inherited)
    nodestroy = false
    space_data = 43.2G
    space_unused_res = 0
    space_snapshots = 0
    space_available = 56.8G
    space_total = 43.2G
    root_group = root
    root_permissions = 755
    root_user = root
    origin =
    Can anybody please help?
    Regards.

    try this:
    svcadm enable nfs/clientcheers
    bjoern

  • Rac and clustering setup

    Hi all,
    I want to know abt RAC and clustering.Can any one describe me abt these two things.
    I got two instances running in 2 AIX machines of size 25 gb each. im zero in this. can any one help me regarding this
    first of all i do no what is RAC and Clustering
    Regards
    Elamaran

    2 instances (or more) will indeed be reading and writing to the same database files. When one instance fails, the other one takes over. Therefore availability is ensured.
    look here fore info on RAC:
    http://download-uk.oracle.com/docs/cd/B19306_01/rac.102/b14197/toc.htm

  • Configuring Servers and Clusters

    Hi All,
    Whenever I am creatin HFM application,it is asking server,I have tried to "Configuring Servers and Clusters" ,but I am unable to coniger,please assist me.
    Regards,
    Prabhu

    For your second question I think it is E. Coherence Web Edition
    Because Coherence provides the ability to scale your application and also offers high availability by replicating the session objects to coherence nodes/caches.
    Thanks,
    Vijaya

  • Difference between Zfs volume and filesystem ?

    Hi,
    Does any one know the difference between Zfs volume and Zfs filesystem?
    On one of the existing nodes i saw the following enries for two times....
    root@node11> zfs get all rpool/dump
    NAME PROPERTY VALUE SOURCE
    rpool/dump  type                  volume -
    rpool/dump creation Thu Feb 18 13:55 2010 -
    rpool/dump used 1.00G -
    rpool/dump available 261G -
    rpool/dump referenced 1.00G -
    rpool/dump compressratio 1.00x -
    rpool/dump reservation none default
    rpool/dump volsize 1G -
    rpool/dump volblocksize 128K -
    root@node11> zfs get all rpool/ROOT/firstbe/opt/SMAW
    NAME PROPERTY VALUE SOURCE
    rpool/ROOT/firstbe/opt/SMAW  type                  filesystem -
    rpool/ROOT/firstbe/opt/SMAW creation Thu Feb 18 14:03 2010 -
    rpool/ROOT/firstbe/opt/SMAW used 609M -
    rpool/ROOT/firstbe/opt/SMAW available 264G -
    rpool/ROOT/firstbe/opt/SMAW referenced 609M -
    rpool/ROOT/firstbe/opt/SMAW compressratio 1.00x -
    rpool/ROOT/firstbe/opt/SMAW mounted yes -
    rpool/ROOT/firstbe/opt/SMAW quota none default
    rpool/ROOT/firstbe/opt/SMAW reservation 4G local
    rpool/ROOT/firstbe/opt/SMAW recordsize 128K default
    rpool/ROOT/firstbe/opt/SMAW mountpoint /opt/SMAW inherited from rpool/ROOT/firstbe
    root@node11> zfs list
    NAME USED AVAIL REFER MOUNTPOINT
    rpool/dump 1.00G 261G 1.00G -
    rpool/ROOT/firstbe/opt/SMAW 609M 264G 609M /opt/SMAW
    Regards,
    Nitin K

    nitin.k wrote:
    Hi,
    Does any one know the difference between Zfs volume and Zfs filesystem?A volume is a block device. A filesystem is a mounted point for file access.
    For most users, a volume isn't normally necessary except for 'dump'.
    Darren

  • Multiple call managers and clusters sharing one application server

    Hi
    I need to program my application such that it can be shared by multiple call managers and clusters. That is, I can only have one application server, and as a result I need to tell my application from which call manager / cluster this "request" is coming from.
    My question is how do we do that? do we have access to some parameters / variables, similar to "#DEVICENAME#", which can be used to pass the call manager identity to the application?
    TIA

    Use multiple instances of your application with different URIs (assuming the instances will not have to talk to each other), then, when configuring the service on each cluster, simply reference the specific URI of the application instance for that cluster. eg: Using Tomcat to host the PhoneMessenger application for Cluster1 and Cluster2, Deploy the same application twice: PhoneMessengerForCluster1 and PhoneMessengerForCluster2. The service URL on Cluster1 might be: http://app_server/PhoneMessengerForCluster1/servlet/PhoneMessenger , and for Cluster2 it could be http://app_server/PhoneMessengerForCluster2/servlet/PhoneMessenger

  • How to decide how big my ZIL and Cache device should be?

    Hi all,
    I have a multiple LUN connected to my server with different size.
    1.
    If I want to add ZIL to the pool:
         How to calculate the ZIL (size) to fit different Pool size?
         it is better to mirror the ZIL? If the ZIL devices is from the SAN storage?
    2.
    Same as Cache device, how big  should I give?
    3.
    Cache device is recommended to use SSD. However, if my SSD is not the local SSD but it is a SSD in the SAN storage which is the same storage of the pool. Is it useless to give a cache device to the pool? As it limited to the fiber channel throughput.

    Hi,
    Good questions and to recap these performance features:
    Separate log devices (ZIL) are good for improving synchronous write performance
    Separate cache devices (L2ARC) are good for improving read performance
    In general, we recommend the following:
    1. Use SSDs for both ZIL or cache otherwise you won't see the performance boosts when using HDDs.
    2. The cache device size should equal your application's warm working set size.
    3. General log sizing recommendations are here:
    Creating and Destroying ZFS Storage Pools - Oracle Solaris 11.1 Administration: ZFS File Systems
    Creating a ZFS Storage Pool With Log Devices
    A more specific case is for the Oracle db redo log, where the recommendation is 15 seconds of redo
    log activity x 2 or 300 MB.
    4. Local attached SSDs as log or cache devices will perform much better than if they are attached
    through a SAN array.
    5. Mirrored log devices are recommended, but unnecessary for cache devices.
    Thanks, Cindy

  • Apex Listener and Clustering - WebLogic Server

    Has anybody worked with the apex listener on a clustered WebLogic environment with a load balancer? We are having lots of connection issues, and I am wondering if this could be part of the problem.
    Thanks,
    Michelle

    Someone Kindly try to give me reply for my issue .
    Thanks& Regards,
    Nitesh pareek.

  • WSS security and clustering  in SOA Suite 11g

    We are calling one composite from another as a reference. Both are on the same clustered domain.
    We are using AIA WSDLs so there is a default WSS username password token as policy.
    While on the single server domain, we do not need to pass the credentials since request was going on the same Weblogic server and it was running fine.
    But when we moved to the clustered domain, we changed the endpoints in composites to an external hardware load balancer and now it needs the sercurity credentials because the request is coming via the load balancer running on the different machine.
    I have two queries:
    1. Is there a way we can skip external load balancer while calling composites within the same clustered domain and still be able to route the requests to all managed servers instead of only one? (may be there is a way we can provide multuple URLs in the composite somehow or any other setting on cluster which would enable us to send all the requests to a common hostname which will route it to multiple managed servers in round robin)
    2. If above is not possible then we will have to use the external load balancer, in which case I need to set username password tokens while calling different composites. I need help in configuring this. It would be nice if anyone can provide me a link to a sample may be.
    Thanks

    In 11g, Oracle WSM Agent has been integrated into Oracle WebLogic Server. Please refer below link for more details -
    http://docs.oracle.com/cd/E15523_01/web.1111/b32511/intro.htm
    Regards,
    Anuj

  • Configuart​ion data and clusters

    When coding for test rigs, it is normally required to save a large amount of configuration parameters that will be used during the test sequence. And these have to be saved in the hard disk.
    The only option seems to be to save these parameters in clusters and write to disk. I am talking of something like 40 to 60 values - single precision, boolean, bytes and even strings all mixed.
    And when the test is running, these clusters are read from disk, unbundled and used in various functions for comparison, limit values and the like.
    Obviously this involves intensive wiring to bundle the clusters and unbundle them. Are there any other options that are more elegant ?
    Thanks
    Raghunathan
    PS : I had posted the same query today morning but somehow it had vanished ! Hence posting again.
    Raghunathan
    LV2012 to Automate Hydraulic Test rigs.

    I think there are various ways to go about that. It is not completely clear to me what is bugging you the most, but it seems like you think it takes too much code to manage configuration data. First of all, I would like to point out the following: when using datalog or configuration file VI's, you do not need to cast any data: these VI's will do that automatically for you. Datalog files seem to be the most powerful because you can write and read complete complex clusters to file and read complete clusters as well.
    here I got distracted by somebody and got a bunch of beers in lovely downtown Amsterdam....
    Back behind my computer I notice your question got answered already. I would like to add a suggestion to the one VI solution:
    Use two enums: one for read or write and one for all the "properies"/configurations. This will only work if you don't have to many different data types though. It is the cleanest solution I can think of, loading all settings into an uninitialized shift register (at startup) and calling that vi anywhere in your code where you need to read or write values. If you have many diffrent datatypes, you may consider embedding this vi inside new vi's that just do more specific tasks.
    Or maybe I just had to much to drink..
    Aart-Jan

  • Configuration data and clusters

    While coding in LV for product test rig applications there is this need to save configuration data to disk and retrieve it while running the test. A typical scenario to test product "AA" will be :
    1. There are 10 different variations of the basic product AA_1 to AA_10.
    2. Each variant has 40 odd parameters that can vary.
    3. User enters the values for these variants and will have to be saved on disk.
    4. When a particular model is tested, all 40 parameters for it have to be read from disk and used in the various tests.
    In such events the only option seems to be :
    (*) Collect all the user entered parameters into clusters. If there are totally 8 tests for each model, then have 8 clusters and bundle parameters pertaining to each test into that cluster.
    (*) Write the clusters to disk in a bit stream using "Write File.VI"
    (*) When running the test, retrieve the parameters from disk and load them into the relevant clusters back. Unbundle by name and use each of the parameters as required. And if the same parameter is required to be used in many functions in the same test, then assign it to a local variable (my beloved ;-)) and use it.
    I would like experienced LV professionals to comment on this method.
    Handling the clusters and wiring them for large data storing and retrieval calls for heavy wiring. Is there no way out ?
    Thanks
    Raghunathan
    Raghunathan
    LV2012 to Automate Hydraulic Test rigs.

    Hi Raghunathan,
    Since you have to deal with so many elements of data, it looks like clusters will be your best method. It will simplify your file I/O and will also help you keep the data organized. I'm not sure why you must write the data to disk, but if it is necessary, clusters will be helpful. It's true that there will be many wires when you bundle and unbundle those clusters, but that will happen no matter how you deal with large numbers of parameters. It will probably make your code more readable if you write subVIs to take the clusters as input, and then break apart the clusters within the subVIs.
    I hope this is helpful!
    Megan B.
    National Instruments

  • Need recommendation - iis with two physical servers and clusters

    Hello,
    I was hoping to get a recommendation/opinion or two on the configuration I've inherited. I'm looking
    at the following configuration using weblogic 8.1 sp4 server (to be upgraded to sp6) running on two physical machines:
    MachineA:
    Domain: myDomain
    Cluster: cluster1
    Servers: Admin Server:9001 (not part of cluster)
    ManagedServer1:7001
    ManagedServer2:7010
    MachineB:
    Domain: myDomain
    Cluster: cluster2
    Servers: Admin Server:9001 (not part of cluster)
    ManagedServer3:7001
    ManagedServer4:7010
    On MachineA: IIS Proxy Plugin with parameter in .ini file
    WeblogicCluster: MachineA:7001, MachineA:7010, MachineB:7001,MachineB7010
    Naturally, the IIS is acting as a round robin load balancer..
    The domains were created using the wizard on both machines and basic domain template..
    For proper failover to be taken advantage of, does something different need to be configured? Or is it fine?
    And if so, does group replication have something to do with it?
    Should either of the clusters be Managed Servers across physical machines?
    or should this just be one cluster of 4 managed servers across two machines?
    I really am not sure on any of this, even though i've read through much of the documentation for clustering..twice.
    I should mention that there is only a .war file deployed with 4 connection pools and 4 Data Sources..No EJBs or JMS
    Would really appreciate any feedback or opinions/advice..
    Thanks

    This is what I'd do
    MachineA:
    Domain: myDomain
    Cluster: cluster1
    Servers: Admin Server:9001 (not part of cluster)
    ManagedServer1:7010
    ManagedServer2:7020
    MachineB:
    Domain: myDomain
    Cluster: cluster2
    Servers: Admin Server:9001 (not part of cluster)
    ManagedServer3:8010
    ManagedServer4:8020
    I just changed the port numbers so they flow and have a pattern - 70* port numbers belong to cluster 1, 80* port numbers belong to cluster 2. No technical benefit apart from it makes it tidier in my opinion.
    Port 9001 for the admin server is normally the domain wide admin port - you could be using that, so that's fine, but my admin server is running 7001.
    You also only need one admin server per domain, this can run on MachineA
    Replication groups are used to help WLS figure out where to place the secondary http session. It will try servers on a different machine/replication group first, if none available, it will create secondary on the same machine.
    What you will end up with is a domain with 4 managed servers in it. The admin server will run on machineA, say using port 7001, each of the managed servers will contact the admin server to download it's config.
    The cluster1 will contain 2 managed servers and the same for cluster2.
    providing you have your cluster multicast addresses set up and the cluster is working, that should be all you need to do for the session failover.
    There's some BEA sample code somewhere which will deploy a web-app and show which managed server you've attached to. If you find that and then connect through IIS, shut down the server that the JSP tells you, you should be able to see the session fail-over nicely.
    Hope that helps,
    Pete

  • Zfs snapshots and booting ...

    Hello,
    In solaris 9, filesystem snapshots did not survive reboots. Do zfs snapshots in solaris 10 persist across reboots ?
    Can I boot off of a zfs partition ?
    thanks.

    Does this mean that when new machines appear with zfs
    support, OR when I can update my PROM, that I will be
    able to boot a zfs partition ?ZFS isn't out yet, so your question is premature. We'll get a look at it within a few weeks, hopefully.
    However, a few months ago it was widely reported by the developers that the initial release would not have boot support. Who knows if this has changed or not.
    I don't see any particular reason that PROM or hardware support is required, it should just need a bootloader that understands ZFS. I don't think that there's any UFS support in the existing proms. Just stuff that understands the VTOC label and how to load and execute a few blocks from a particular slice.
    Darren

  • Load balancing and clustering in sharepoint

    Hi,
    Still i am confuse about load balancing with cluster in sharepoint level.please let me know how to install and troubleshoot,how it is work, Advantage, etc..
    Thanks,
    Inguru

    Hi Inguru,
    Per my knowledge, SharePoint only supports Load balancing and SQL support clustering.
    Here is a similar thread for you to take a look:
    http://social.technet.microsoft.com/Forums/en-US/2b20d1d5-de35-486e-9b0e-37222a307615/clustering-and-load-balancing?forum=sharepointgeneralprevious
    To configure load balancing for SharePoint, please follow the links below:
    http://blogs.technet.com/b/praveenh/archive/2010/12/17/setting-up-load-balancing-on-a-sharepoint-farm-running-on-windows-server-2008.aspx
    http://community.bamboosolutions.com/blogs/sharepoint-2013/archive/2014/01/07/network-load-balancing-for-sharepoint-2013-part-three-installing-network-load-balancing-on-wfe1-in-a-three-server-farm.aspx
    Best regards.
    Thanks
    Victoria Xia
    TechNet Community Support

Maybe you are looking for

  • Error while previewing report  in Mobile

    Hi Experts, one of my client required to view reports in mobile for that i have done crystal report but while previewing it was throwing an error as attatched screen shot how to resolvwe this issue

  • How can I drag files from my macbook to my external hard drive?

    Hi! I can't seem to drag files to my hard drive using my macbook pro? How can I do that?

  • Nano's only partially shuffling

    My 2nd gen. Nano was working fine until about two days ago, when the shuffle function on it decided to partially die. When it's on shuffle, it plays about four or five songs in non-shuffled order (almost always from the same album). Then it shuffles

  • Naming audio files

    This one really bugs me. How do I name my audio tracks that are saved to the hard disk the same as the name on the arange track. I set my path audio name, drums for example. Then I option click on the track name in the arrange, name all instruments,

  • Are there any vi's that model ac induction machines on this site?

    I am making a vi to model an AC induction machine given the mechanical torque (Tm) or the shaft velocity (omega_r). If this already exists I won't add my own (because it is rudimentery). It seems like it should be here, but I can't find it. The begin