Server 2012 R2 Hyper-v two node cluster network connections help please

I am trying to do a hyper-v failover clustering project and we have following hardwares:
2x HP DL 380 G8 with 8 LAN ports
1x HP MSA 2000 with dual controller (each controller has 2 iSCSI LAN ports)
2x 24 port gig switch (HP Procurve)
I don’t know whether this is a best practice but I would like to setup separate network for live migration, iSCSI, heartbeat & data/management. So far I have installed Server 2012 R2 datacenter on both servers.
MSA Controller A port 1 is connected to Switch 1 – Vlan 1
MSA Controller A port 2 is connected to Switch 2 – Vlan 2
MSA Controller B port 1 is connected to Switch 1 – Vlan 1
MSA Controller B port 2 is connected to Switch 2 – Vlan 2
For each server is this how network connection should be? Ports will be teamed.
2 ports for iSCSI – vlan 3
2 ports for Live Migration – vlan 4
2 ports for Heartbeat – vlan 5
2 ports for Management/VM – vlan 6
Please advise if there is other better options.
Any help would be much appreciated, thank you.

Watch this webcast from TechEd, it will answer your questions:
http://channel9.msdn.com/Events/TechEd/NorthAmerica/2013/MDC-B337#fbid=?hashlink=fbid
Thanks!
Elden

Similar Messages

  • SQL Server 2012 standard edition on 2 node cluster

    I have an environment with three SQL instances running an application. Two of the instances are enterprise edition and one is standard. I am looking to move all instances up to SQL Server 2012 and hoping to put them on a 3 node Windows Failover Cluster.
    I understand that standard edition is limited to two nodes. What I want to know is what my options are?
    Can install standard edition on the cluster but only on two of the three nodes?
    Do I need to create a second Windows Cluster? Is this even possible?
    Regards, Matt Bowler MCITP, My Blog

    Can install standard edition on the cluster but only on two of the three nodes?
    Yes. That's possible. SQL Setup would block you to add third node. It doesn't check how many nodes we have in windows cluster.
    Balmukund Lakhani
    Please mark solved if I've answered your question, vote for it as helpful to help other users find a solution quicker
    This posting is provided "AS IS" with no warranties, and confers no rights.
    My Blog |
    Team Blog | @Twitter
    | Facebook
    Author: SQL Server 2012 AlwaysOn -
    Paperback, Kindle

  • Server 2012 R2 Hyper-V Cluster, VM blue screens after migration between nodes.

    I currently have a two node Server 2012 R2 Hyper-v Cluster (fully patched) with a Windows Server 2012 R2 Iscsi target.
    The VMs run fine all day long, but when I try to do a live/quick migration the VM blue screens after about 20 minutes. The blue reports back about a “Critical_Structure_Corruption”.
    I’m being to think it might be down to CPU, as one system has an E-2640v2 and the other one has an E5-2670v3. Should I be able to migrate between these two systems with these type of CPU?
    Tim

    Sorry Tim, is that all 50 blue screen if live migrated?
    Are they all on the latest integration services? Does a cluster validation complete successfully? Are the hosts patched to the same level?
    The fact that if you power them off then migrate them and they boot fine, does point to a processor incompatibility and the memory BIN file is not accepted on the new host.
    Bit of a long shot but the only other thing I can think of off the top of my head if the compatibility option is checked, is checking the location of the BIN file when the VM is running to make sure its in the same place as the VHD\VHDX in the
    CSV storage where the VM is located and not on the local host somewhere like C:\program data\.... that is stopping it being migrated to the new host when the VM is live migrated.
    Kind Regards
    Michael Coutanche
    Blog:   
    Twitter:   LinkedIn:
    Note: Posts are provided “AS IS” without warranty of any kind, either expressed or implied, including but not limited to the implied warranties of merchantability and/or fitness for a particular purpose.

  • Is it possible to create a two node cluster with SQL server 2012 STD on node1 and SQL Server 2008 ENT on node2 on a Windows server 2012?

    Hi All,
    Is it possible to create a two node cluster with SQL server 2012 STD on node1 and SQL Server 2008 ENT on node2 on a Windows server 2012?
    aa

    I think you are confused with the way cluster is installed. Installing SQL on cluster is a two step process. On first node you create a new SQL cluster and on second node you need "AddNode" action.
    Same edition media has to be used in above two operation.
    you can install two separate instances with one enterprise and one standard.
    Balmukund Lakhani
    Please mark solved if I've answered your question, vote for it as helpful to help other users find a solution quicker
    This posting is provided "AS IS" with no warranties, and confers no rights.
    My Blog |
    Team Blog | @Twitter
    | Facebook
    Author: SQL Server 2012 AlwaysOn -
    Paperback, Kindle

  • Migrate from server core 2008 r2 hyper-v with failover cluster volumes to server core 2012 r2 hyper-v with failover cluster volumes on new san hardware

    We are getting ready to migrate from server core 2008 r2 hyper-v with failover cluster volumes on an iscsi san to server core 2012 r2 hyper-v with failover cluster volumes on a new iscsi san.
    I've been searching for a "best practices" article for this but have been coming up short.  The information I have found either pertains to migrating from 2008 r2 to 2012 r2 with failover cluster volumes on the same hardware, or migrating
    to different hardware without failover cluster volumes.
    If there is anyone out there that has completed a similar migration, it would be great to hear any feedback you may have on your experiences.
    Currently, my approach is as follows:
    1. Configure new hyper-v with failover cluster volumes on new SAN with new 2012 r2 hostnodes and 2012 r2 management server
    2. Turn off the virtual machines on old 2008 r2 hyper-v hostnodes
    3. Stop the VMMS service on the 2008 r2 hostnodes
    4. copy the virtual machine files and folders over to the new failover cluster volumes
    5. Import vm's into server 2012 r2 hyper-v.
    Any feedback on the opertain I have in mind would be helpful.
    Thank you,
    Rob

    Hi Rob,
    Yes , I agree with that "file copy " can achieve migration.
    Also you can try "copy cluster wizard " :
    https://technet.microsoft.com/en-us/library/dn530779.aspx
    Best Regards,
    Elton Ji
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact [email protected] .

  • Testing ha-nfs in two node cluster (cannot statvfs /global/nfs: I/O error )

    Hi all,
    I am testing HA-NFS(Failover) on two node cluster. I have sun fire v240 ,e250 and Netra st a1000/d1000 storage. I have installed Solaris 10 update 6 and cluster packages on both nodes.
    I have created one global file system (/dev/did/dsk/d4s7) and mounted as /global/nfs. This file system is accessible form both the nodes. I have configured ha-nfs according to the document, Sun Cluster Data Service for NFS Guide for Solaris, using command line interface.
    Logical host is pinging from nfs client. I have mounted there using logical hostname. For testing purpose I have made one machine down. After this step files tem is giving I/O error (server and client). And when I run df command it is showing
    df: cannot statvfs /global/nfs: I/O error.
    I have configured with following commands.
    #clnode status
    # mkdir -p /global/nfs
    # clresourcegroup create -n test1,test2 -p Pathprefix=/global/nfs rg-nfs
    I have added logical hostname,ip address in /etc/hosts
    I have commented hosts and rpc lines in /etc/nsswitch.conf
    # clreslogicalhostname create -g rg-nfs -h ha-host-1 -N
    sc_ipmp0@test1, sc_ipmp0@test2 ha-host-1
    # mkdir /global/nfs/SUNW.nfs
    Created one file called dfstab.user-home in /global/nfs/SUNW.nfs and that file contains follwing line
    share -F nfs –o rw /global/nfs
    # clresourcetype register SUNW.nfs
    # clresource create -g rg-nfs -t SUNW.nfs ; user-home
    # clresourcegroup online -M rg-nfs
    Where I went wrong? Can any one provide document on this?
    Any help..?
    Thanks in advance.

    test1#  tail -20 /var/adm/messages
    Feb 28 22:28:54 testlab5 Cluster.SMF.DR: [ID 344672 daemon.error] Unable to open door descriptor /var/run/rgmd_receptionist_door
    Feb 28 22:28:54 testlab5 Cluster.SMF.DR: [ID 801855 daemon.error]
    Feb 28 22:28:54 testlab5 Error in scha_cluster_get
    Feb 28 22:28:54 testlab5 Cluster.scdpmd: [ID 489913 daemon.notice] The state of the path to device: /dev/did/rdsk/d5s0 has changed to OK
    Feb 28 22:28:54 testlab5 Cluster.scdpmd: [ID 489913 daemon.notice] The state of the path to device: /dev/did/rdsk/d6s0 has changed to OK
    Feb 28 22:28:58 testlab5 svc.startd[8]: [ID 652011 daemon.warning] svc:/system/cluster/scsymon-srv:default: Method "/usr/cluster/lib/svc/method/svc_scsymon_srv start" failed with exit status 96.
    Feb 28 22:28:58 testlab5 svc.startd[8]: [ID 748625 daemon.error] system/cluster/scsymon-srv:default misconfigured: transitioned to maintenance (see 'svcs -xv' for details)
    Feb 28 22:29:23 testlab5 Cluster.RGM.rgmd: [ID 537175 daemon.notice] CMM: Node e250 (nodeid: 1, incarnation #: 1235752006) has become reachable.
    Feb 28 22:29:23 testlab5 Cluster.RGM.rgmd: [ID 525628 daemon.notice] CMM: Cluster has reached quorum.
    Feb 28 22:29:23 testlab5 Cluster.RGM.rgmd: [ID 377347 daemon.notice] CMM: Node e250 (nodeid = 1) is up; new incarnation number = 1235752006.
    Feb 28 22:29:23 testlab5 Cluster.RGM.rgmd: [ID 377347 daemon.notice] CMM: Node testlab5 (nodeid = 2) is up; new incarnation number = 1235840337.
    Feb 28 22:37:15 testlab5 Cluster.CCR: [ID 499775 daemon.notice] resource group rg-nfs added.
    Feb 28 22:39:05 testlab5 Cluster.RGM.rgmd: [ID 375444 daemon.notice] 8 fe_rpc_command: cmd_type(enum):<5>:cmd=<null>:tag=<>: Calling security_clnt_connect(..., host=<testlab5>, sec_type {0:WEAK, 1:STRONG, 2:DES} =<1>, ...)
    Feb 28 22:39:05 testlab5 Cluster.CCR: [ID 491081 daemon.notice] resource ha-host-1 removed.
    Feb 28 22:39:17 testlab5 Cluster.RGM.rgmd: [ID 375444 daemon.notice] 8 fe_rpc_command: cmd_type(enum):<5>:cmd=<null>:tag=<>: Calling security_clnt_connect(..., host=<testlab5>, sec_type {0:WEAK, 1:STRONG, 2:DES} =<1>, ...)
    Feb 28 22:39:17 testlab5 Cluster.CCR: [ID 254131 daemon.notice] resource group nfs-rg removed.
    Feb 28 22:39:30 testlab5 Cluster.RGM.rgmd: [ID 224900 daemon.notice] launching method <hafoip_validate> for resource <ha-host-1>, resource group <rg-nfs>, node <testlab5>, timeout <300> seconds
    Feb 28 22:39:30 testlab5 Cluster.RGM.rgmd: [ID 375444 daemon.notice] 8 fe_rpc_command: cmd_type(enum):<1>:cmd=</usr/cluster/lib/rgm/rt/hafoip/hafoip_validate>:tag=<rg-nfs.ha-host-1.2>: Calling security_clnt_connect(..., host=<testlab5>, sec_type {0:WEAK, 1:STRONG, 2:DES} =<1>, ...)
    Feb 28 22:39:30 testlab5 Cluster.RGM.rgmd: [ID 515159 daemon.notice] method <hafoip_validate> completed successfully for resource <ha-host-1>, resource group <rg-nfs>, node <testlab5>, time used: 0% of timeout <300 seconds>
    Feb 28 22:39:30 testlab5 Cluster.CCR: [ID 973933 daemon.notice] resource ha-host-1 added.

  • Creating a Two Node Cluster

    Good afternoon,
    I'm looking to build/create an inexpensive two node cluster. I have a SLES11SP1 server that is running XEN as a virtual hosting server, I run about five servers in a virtual environment. I have three USB drives set up to host my Guest servers, what I would like to do is to purchase another USB drive, so that I can use that as an iscsi/SAN server location, it would be 2TB in size, and then I would build my new servers making them cluster enabled. They should then be able to "see" the SAN/iscsi/storage location.
    Does anyone have any further suggestions?
    Thanks
    -DS

    Originally Posted by gleach1
    while I wouldn't recommend running servers off USB drives (unless this is for testing), I don't see any issue in doing it but the performance may not be fantastic
    I really appreciate your responding back and assisting in this testing. This cluster is for testing only, I have a small server farm in my basement, and I have a customer that is using clustering and I would like to at least be able to have something to test with
    I've actually run a test cluster off a USB drive using vmware workstation before and it seemed to run fine just as a test system
    If you set up your xen host as an iscsi server, use the USB disk as the storage you present to the guests and set them up with iscsi initiators it should work like any other iscsi san would, obviously a touch slower...
    Is there some documentation that explains how to do the iscsi server setup that you described? I have the "Configuring Novell Cluster Services in a XEN Virtualization Environment" but I really don't see anything about the iscsi initiator setup, I was going to add the USB drive as a /storage volume on the XEN host, and then point the cluster to that? I will also be adding a third card to handle the clustering network.

  • New Howto guide available: HOW to INSTALL and CONFIGURE A TWO-NODE CLUSTER

    Hi,
    I am not sure that this has been posted or not. But for all of us who are too lazy (like myself) to read the full set of documentation, or who have not enough time, please have a look at this short document with screen shots that explains how to setup a simple two-node cluster.
    http://www.sun.com/software/solaris/howtoguides/twonodecluster.jsp
    I still recommend to browse through the "Sun Cluster Concepts Guide for Solaris OS" http://docs.sun.com/app/docs/doc/819-0421
    Have fun
    Hartmut

    I am afraid there is no such document. But I can try to give you some hints.
    N+1 can be seen as two things:
    - the phsical setup of the cluster and
    - the logical configuration of services
    A physical N+1 setup would mean, that you have one server that is connected to all shared storage, but the other N nodes are only connected to the shared storage they need.
    This configuration would also need the logical N+1 setup, namely that all HA services could only run on 2 nodes, their primary node and as a backup this one dedicated node that is connected to all shared storage.
    If you do not have the physical N+1 setup, you could easily have a logical N+1 setup, by having a N*N topology, j.e. all nodes connected to all storage, but logically each service again would have only 2 nodes on its nodelist.
    It seems that the second option is the one that is used more often, as ift offers you the ability to reconfigure things in case of failures.
    With regards to asymmetric and symmetric I am not quite sure what you mean: And I do not think that these terms are used in the SC docs.
    What I can think of is, that asymmetric means that you have one node being active with 1 or more HA services and have one inactive secondary node as a backup.
    Regards
    Hartmut

  • (Request for:) Best practices for setting up a new Windows Server 2012 r2 Hyper-V Virtualized AD DC

    Could you please share your best practices for setting up a new Windows Server 2012 r2 Hyper-V Virtualized AD DC, that will be running on a new WinSrv 2012 r2 host server.   (This
    will be for a brand new network setup, new forest, domain, etc.)
    Specifically, your best practices regarding:
    the sizing of non virtual and virtual volumes/partitions/drives,  
    the use of sysvol, logs, & data volumes/drives on hosts & guests,
    RAID levels for the host and the guest(s),  
    IDE vs SCSI and drivers both non virtual and virtual and the booting there of,  
    disk caching settings on both host and guests.  
    Thanks so much for any information you can share.

    A bit of non essential additional info:
    We are small to midrange school district who, after close to 20 years on Novell networks, have decided to design and create a new Microsoft network and migrate all of our data and services
    over to the new infrastructure .   We are planning on rolling out 2012 r2 servers with as much Hyper-v virtualization as possible.
    During the last few weeks we have been able to find most of the information we need to undergo this project, and most of the information was pretty solid with little ambiguity, except for
    information regarding virtualizing the DCs, which as been a bit inconsistent.
    Yes, we have read all the documents that most of these posts tend point to, but found some, if not most are still are referring to performing this under Srvr 2008 r2, and haven’t really
    seen all that much on Srvr2012 r2.
    We have read these and others:
    Introduction to Active Directory Domain Services (AD DS) Virtualization (Level 100), 
    Virtualized Domain Controller Technical Reference (Level 300),
    Virtualized Domain Controller Cloning Test Guidance for Application Vendors,
    Support for using Hyper-V Replica for virtualized domain controllers.
    Again, thanks for any information, best practices, cookie cutter or otherwise that you can share.
    Chas.

  • Server 2012 r2 Hyper-v guest network speed with 10 GB NIC's

    Hi
    Setting up a Server 2012 r2 Hyper-v host with an onboard 4 port 1 GB nic and a offboard dual port 10 GB nic. We are using the 1gb quad port for management. The 10 Gb we intend using for hyper-v guest use however we don't seem to be able to achieve very high
    speeds though the 10 GB adapter when assigned to a virtual switch. Testing the 10 Gb nic from the host confirms that we can transfer/get 10 GB speeds. When assigned to a virtual switch however the speed drops to about 1-2 Gbps. Has anybody seen anything like
    this ?. We've tried a variety of settings and i can confirm that the network traffic is using the 10 GB nic.

    Hi,
    I am almost certain what you see is by design. I don't want to go into to much detail because I'm in a bit of a hurry. But allow me to explain.
    Suppose you have a 10Gb NIC interface and an multi-core CPU. When you have RSS (Receive-Side Scaling) disabled a single core can normally handing
    around 2.5-4.3Gbps. When you enable RSS the data streams/traffic is spread across multiple cores.
    But RSS is a network optimization feature for the Management OS, not for a vSwitch. Instead you should use VMQ (Virtual Machine Queuing). VMQ is
    the equilavant of RSS, but then intended for VMs. But... there is a difference. VMQ does not spread single data streams/traffic across multiple cores. Instead is (dynamically) assigns each vmNIC its own core. So... you don't use multiple-cores at a time for
    a single vmNIC. You can't use RSS and VMQ at the same time. Keep that in mind.
    But... you can accomplish this by enabling vRSS as well. That's another story.
    If you watch the following video, at the end you know all about VMQ.
    Hyper-V dVMQ
    http://www.hypervrockstar.com/tag/dvmq/
    Boudewijn Plomp | BPMi Infrastructure & Security
    This posting is provided "AS IS" with no warranties, and confers no rights. Please remember, if you see a post that helped you please click "Vote as Helpful", and if it answered your question, please click "Mark as Answer".

  • Creation of diskset in Two node cluster

    Hi All ,
    I have created one diskset in solaris 9 using SVM in two node cluster.
    After diskset creation, I mounted the diskset in a primary node in the mount point /test. But, the disk set is mounting on both the nodes.
    I created this diskset for failover purposes, if one node goes down the other node will take care.
    My idea is to create a failover resource (diskset resources) in the two node cluster.
    Below are steps used for creating the disk set.
    root@host2# /usr/cluster/bin/scdidadm -L d8
    8        host1:/dev/rdsk/c1t9d0   /dev/did/rdsk/d8
    8        host2:/dev/rdsk/c1t9d0   /dev/did/rdsk/d8
    metaset -s diskset -a -h host2 host1
    metaset -s diskset -a -m host2 host1
    metaset -s diskset -a /dev/did/rdsk/d8
    metainit -s diskset d40 1 1 /dev/did/dsk/d8s0
    newfs /dev/md/diskset/rdsk/d40
    mount /dev/md/diskset/dsk/d40 /test
    root@host2# metaset -s diskset
    Set name = diskset, Set number = 1
    Host                Owner
      host2                  Yes
      host1
    Mediator Host(s)    Aliases
      host2
      host1
    Driv Dbase
    d8   YesPlease let me know how to mount the disk set in one node.
    If i am wrong, please correct me.
    Regards,
    R. Rajesh Kannan.

    The file system will only mount on both (all) nodes if you mount it globally, i.e with the global flag or if there is an entry in /etc/vfstab that has a global option.
    Given your output below, I would guess you have a global mount for /test defined in /etc/vfstab.
    Regards,
    Tim
    ---

  • Do I need 3 machines for Sun Two-Node Cluster HA for Apache?

    The vendor is trying to set up a Sun Cluster HA for Apache and said it requires a dedicated machine
    for monitoring purpose(The Sun Cluster HA for Apache probe??). Is this monitoring machine mandatory and have to be dedicated for that purpose? And can I use 1 machine to monitor more than 1 cluster? Because the vendor requires 2 extra machine to monitor 2 set of 2-node cluster..and the document I read from Sun only mentions the 3rd machine is required for admin console.
    Also, if I only have 2 node cluster, can I configure Apache as a scalable + failover service at the same time or I have to choose either one? Thanks!!

    Incorrect. You can set up a Solaris Cluster on a single node if you wish although two nodes give you redundancy. The probe for Apache runs from the node on which Apache is executing.
    If you have a two node cluster you could have multiple implementations of Apache: fail-over and scalable. You can mix them until you run out of resources: CPU, networking bandwidth, IP addresses, etc.
    Tim
    ---

  • How to install Enterprise manager on two node cluster

    Hi,
    I am new to Oracle working on 10g
    can anyone post a document or link for step by step installation of oracel 10g enterprise manager on a two node cluster?
    Thanks in advance.

    userArshad wrote:
    Thanks Hans.
    To be specific I have to install Database control .that means I need to use the GUI .
    Also I could see database control and DBconsole service on on 1st node but I cannot find that in 2nd node.That is the normal configuration - you use the DB Control from node 1 and it gets the information using the agent on Node 2. The DB Console then displays the information for both nodes.
    Please note that the DB Control will place a relatively heavy load on each machine, which is why we want to restrict our load to the agent. It is for that reason that we prefer to use the Grid Control whenever possible - it provides the same mechanism and screens, but the JEE platform is on a separate machine and all you have on the RAC nodes is the agent.
    But if you really want to run the DB Console from each node (again, not normally needed) you need to use emca to configure it on the second node. Metalink note 863631.1 should help you with that.

  • Simple two node Cluster Install - Hung after reboot of first node

    Hello,
    Over the past couple of days I have tried to install a simple two node cluster using two identical SunFire X4200s, firstly following the recipe in: http://www.sun.com/software/solaris/howtoguides/twonodecluster.jsp
    and when that failed referring to http://docs.sun.com/app/docs/doc/819-0912 and http://docs.sun.com/app/docs/doc/819-2970.
    I am trying to keep the install process as simple as possible, no switch, just back to back connections for the internal networking (node1 e1000g0 <--> node2 e1000g0, node1 e1000g1 <--> node2 e1000g1)
    I ran the installer on both X4200s with default answers. This went through smoothly without problems.
    I ran scinstall on node1, first time through, choosing "typical" as suggested in the how to guide. Everything goes OK (no errors) node2 reboots, but node1 just sits there waiting for node2, no errors, nothing....
    I also tried rerunning scinstall choosing "Custom", and then selecting the no switch option. Same thing happened.
    I must be doing something stupid, it's such a simple setup! Any ideas??
    Here's the final screen from node1 (dcmds0) in both cases:
    Cluster Creation
    Log file - /var/cluster/logs/install/scinstall.log.940
    Checking installation status ... done
    The Sun Cluster software is installed on "dcmds0".
    The Sun Cluster software is installed on "dcmds1".
    Started sccheck on "dcmds0".
    Started sccheck on "dcmds1".
    sccheck completed with no errors or warnings for "dcmds0".
    sccheck completed with no errors or warnings for "dcmds1".
    Configuring "dcmds1" ... done
    Rebooting "dcmds1" ...
    Output from scconf on node2 (dcmds1):
    bash-3.00# scconf -p
    Cluster name: dcmdscluster
    Cluster ID: 0x47538959
    Cluster install mode: enabled
    Cluster private net: 172.16.0.0
    Cluster private netmask: 255.255.248.0
    Cluster maximum nodes: 64
    Cluster maximum private networks: 10
    Cluster new node authentication: unix
    Cluster authorized-node list: dcmds0 dcmds1
    Cluster transport heart beat timeout: 10000
    Cluster transport heart beat quantum: 1000
    Round Robin Load Balancing UDP session timeout: 480
    Cluster nodes: dcmds1
    Cluster node name: dcmds1
    Node ID: 1
    Node enabled: yes
    Node private hostname: clusternode1-priv
    Node quorum vote count: 1
    Node reservation key: 0x4753895900000001
    Node zones: <NULL>
    CPU shares for global zone: 1
    Minimum CPU requested for global zone: 1
    Node transport adapters: e1000g0 e1000g1
    Node transport adapter: e1000g0
    Adapter enabled: no
    Adapter transport type: dlpi
    Adapter property: device_name=e1000g
    Adapter property: device_instance=0
    Adapter property: lazy_free=1
    Adapter property: dlpi_heartbeat_timeout=10000
    Adapter property: dlpi_heartbeat_quantum=1000
    Adapter property: nw_bandwidth=80
    Adapter property: bandwidth=70
    Adapter port names: <NULL>
    Node transport adapter: e1000g1
    Adapter enabled: no
    Adapter transport type: dlpi
    Adapter property: device_name=e1000g
    Adapter property: device_instance=1
    Adapter property: lazy_free=1
    Adapter property: dlpi_heartbeat_timeout=10000
    Adapter property: dlpi_heartbeat_quantum=1000
    Adapter property: nw_bandwidth=80
    Adapter property: bandwidth=70
    Adapter port names: <NULL>
    Cluster transport switches: <NULL>
    Cluster transport cables
    Endpoint Endpoint State
    Quorum devices: <NULL>
    Rob.

    I have found out why the install hung - this needs to be added into the install guide(s) at once!! - It's VERY frustrating when an install guide is incomplete!
    The solution is posted in the HA-Cluster OpenSolaris forums at:
    http://opensolaris.org/os/community/ha-clusters/ohac/Documentation/SCXdocs/relnotes/#bugs
    In particular, my problem was that I selected to make my Solaris install secure (A good idea, I thought!). Unfortunately, this stops Sun Cluster from working. To fix the problem you need to perform the following steps on each secured node:
    Problem Summary: During Solaris installation, the setting of a restricted network profile disables external access to network services that Sun Cluster functionality uses, ie: The RPC communication service, which is required for cluster communication
    Workaround: Restore external access to RPC communication.
    Perform the following commands to restore external access to RPC communication.
    # svccfg
    svc:> select network/rpc/bind
    svc:/network/rpc/bind> setprop config/local_only=false
    svc:/network/rpc/bind> quit
    # svcadm refresh network/rpc/bind:default
    # svcprop network/rpc/bind:default | grep local_only
    Once I applied these commands, the install process continued ... AT LAST!!!
    Rob.

  • Why does my laptop use a proxy server for mozilla or chrome when i check in internet explorer it sais my system isnt but it is it jumps back to a proxy server in mozille when i tick no proxy ?? help please im not wizzy thanks

    Why does my laptop use a proxy server for mozilla or chrome when i check in internet explorer it sais my system isnt but it is it jumps back to a proxy server in mozille when i tick no proxy ?? help please im not wizzy thanks

    Start Firefox in <u>[[Safe Mode]]</u> to check if one of the extensions or if hardware acceleration is causing the problem (switch to the DEFAULT theme: Firefox (Tools) > Add-ons > Appearance/Themes).
    *Don't make any changes on the Safe mode start window.
    *https://support.mozilla.com/kb/Safe+Mode
    See also:
    *http://kb.mozillazine.org/Preferences_not_saved
    *https://support.mozilla.com/kb/Preferences+are+not+saved

Maybe you are looking for