Best Practices - VMware ESX 4.0 in a Cisco Environment?

Hello,
I'm presently designing a VMware ESX 4.0 deployment and integrating it with our Cisco environment.  I've found the following document:
http://www.cisco.com/en/US/docs/solutions/Enterprise/Data_Center/vmware/VMware.html "VMware Infrastructure 3 in a Cisco Network Environment" and I was just wondering if there was a newer document applicable to ESX 4.x or if these best practices still applied?
I'm particularly interested in proper vlan design for the various port-groups with in ESX and etherchannel configuration between ESX hosts and Cisco switches.
Thanks,
Rob

Well, in that this is a Storage group, I'll answer froma storage noetworking point of view.
ESX hosts are no different to any other host,  Just stick with the standard best practice of single initiator zoning and you'll be fine.
As a slight aside, from an array point of view, I've tended to configure all the pWWN's of the whole cluster into one "host" definition, as this makes LUN mapping easier.
Steven

Similar Messages

  • Best Practice to generate UUIDs in a Cluster-Server Environment

    Hi all,
    I just need some inputs over the best practices to generate UUIDs in typical internet world where there are multiple servers/JVMs involved for load balancing or traffic distribution etc. I know JAVA is shipped with very efficient UUID generator API.
    But still that doesn't solve the issue in multiple server environment.
    For the discussion sake lets assume I need it to be unique over the setup than a near unique.
    How do you guys approach it?
    Thanks you all in advance.

    codeNombre wrote:
    jverd wrote:
    codeNombre wrote:
    Thanks jverd,
    So adding to the theory of "distinguishing all possible servers" in addition to UUID over each server would be the way to go.If you're unreasonably paranoid, sure.I think its a common problem and there is a big number of folks who might still be bugged about the "relative uniqueness" of UUID in long run. People who don't understand probability and scale, sure.
    Again coming back to my original problem in an "internet world", shouldn't the requirement like unique id between different servers be dealt with generating the UUID's at a layer before entering into the multi-server setup. Where would that be? I don't have the answer..Again, that is the POINT of the UUID class--so that you can generate as many IDs as you want and still be confident that nobody anywhere is in the world has ever generated any of those same IDs. However, if your requirements say UUID is not good enough, then you need to define what is, and that means having a lot of foresight as to how this system will evolve and how long it will live, AND having total control over some aspect of your servers, AND having a process that is so good that it's LESS LIKELY for a human to screw up and re-use a "unique" server ID than the probabilities I presented in my previous post.

  • Best Practice for FlexConnect Wireless roaming in MediaNet environment?

    Hello!
    Current Cisco best practice recommendations for enterprise MediaNet design, specify that VLANs be local to a switch / switch stack (i.e., to limit the scope of spanning-tree). 
    In the wireless world, this causes problems if you want users while roaming to keep real-time applications up and running.  Every time they connect to a new AP on a different VLAN, then they will need to get a new IP address, which interrupts real-time apps. 
    So...best practice for LAN users causes real problems for wireless users.
    I thought I'd post here in case there's a best practice for implementing wireless roaming in a routed environment that we might have missed so far!
    We have a failover pair of FlexConnect 7510s, btw, configured for local switching for Internal users, and central switching with an anchor controller on the DMZ for Guest users.
    Thanks,
    Deb

    Thanks for your replies, Stephen and JSnyder.
    The situation here is that the original design engineer is no longer here, and the original design was not MediaNet-friendly, in that it had a very few /20 subnets bridged over entire large sites. 
    These several large sites (with a few hundred wireless users per site), are connected to an HQ location (where the 7510s in failover mode are installed) via 1G ethernet hand-offs (MPLS at the WAN provider).  The 7510s are new, and are replacing older contollers at the HQ location. 
    The internal employee wireless users use resources both local to their site, as well as centralized resources.  There are at least as many Guest wireless users per site as there are internal employee users, and the service to them consists of Internet traffic only.  (When moved to the 7510s, their traffic will continue to be centrally switched and carried to an anchor controller in the DMZ.) 
    (1) So, going local mode seems impractical due to the sheer number of users whose traffic bound for their local site would be traversing the WAN twice.  Too much bandwidth would be used.  So, that implies the need to use Flex / HREAP mode instead.
    (2) However, re-designing each site's IP environment for MediaNet would suggest to go routed to the closet.  However, this breaks seamless roaming for users....
    So, this conundrum is why I thought I'd post here, and see if there was some other cool / nifty solution I wasn't yet aware of. 
    The only other (possibly friendly to both needs) solution I'd thought of was to GRE tunnel a subnet from each closet to the collapsed Core / Disti switch at each site.  Unfortunately, GRE tunnels are not supported in the rev of IOS on the present equipment, and so it isn't possible to try this idea.
    Another "blue sky" idea I had (not for this customer, but possibly elsewhere in the future), is to use LAN switches such as 3850s that have WLC functionality built-in.  I haven't yet worked with the WLC s/w available on those, but I was thinking it looks like they could be put into a mobility group, and L3 user roaming between them might then work.  Do you happen to know if this might be a workable solution to the overall big-picture problem? 
    Thanks again for taking the time and trouble to reply!
    Deb

  • What is the best practice to display info of completed task in process flow

    Hi all,
    I'm starting to study BPM modeling with CE7.1 EHP1. Thanks to the tutorial and example on SDN site and I can easily build my own process in NWDS and deploy to server, start it, finish it.
    I like the new runtime which can show a BPMN diagram to the processors. However, I can't find a way to let the follow up processor to review the task result completed in previous step. I'm more familiar with Guided Procedure, and know there is "Display Callable Object" which can used to show some info of a completed task when the processor/owner/admin/overseer click on a completed task.  Where is the feature in BPM ? What is the best practice to show such task information in BPM environment.
    For example, A multiple level approval process, the higher level approver need to know the comment written by the previous approver. Can he read this information from process flow ?
    I think it is very important feature for a BPM platform. In Guided Procedure, such requirement can be done with Display Callable Object + View Permission, and you just need some coding for the UI. If BPM is superior to GP, I think there must be a way to achieve this, I just do not know how ?
    Can anyone shed me some light on it ?

    Oliver,
    Thanks for your quick reply.
    Yes, Notes and Attachment CAN BE USED for the purpose. But I'm still looking for a more elegant solution.
    With the solution of using Notes/Attachment, the processor need to give input at two places : the task UI and Note/Attach , with similar or same data. It is really annoying.
    Is there any SAP BPM real-world deployment ? None of customer has the requirement ?

  • Best practice for assigning permissions

    Good morning,
    I am trying redo permissions on our shared folders, and want to incorporate some sort of best practice and be security conscious.
    The current environment is permissions is assigned directly to the folder, and it is usually domain users :(.
    I have a multi-domain environment, I want to know what is the best way to handle permissions, so for instance I have a folder called
    \\ITserver01\ITtest, what kind of naming scheme do you give? I was thinking about maybe ITserver01_ITtest_RW as an example...
    Also do I have to create a domain local r/w and R/o group and a universal group r/w and r/o, since I cannot assign place users directly in the domain local account?
    Chad

    Best practices (esp in naming schemes) depends a bit on the corporate culture and standard procedures. However, we put users in domain local groups based on their role. Those groups would be made a member of a domain group that is used to grant access to
    local resources. and then make those resource access domain groups are member of local groups on the server.
    For example, If I have a server 'test', then there is a domain group called 'test administrators' and that group is then a member of the local admins group of the test server. And one of the members of the 'test administrators' group would be the 'site domain
    admins' group.
    For your example,  ITserver01_ITtest_RW would be a domain local group. And you would not put users in it directly, but user groups.  Users are in groups like 'Site helpdesk admins' or whtever. Something that defines their role in the orgnization.
    And then you would put the  'Site helpdesk admins' as member in the ITserver01_ITtest_RW group.
    Does that make sense?

  • Best Practices for Team Development

    Hi!
    I've a question regarding a project using JHeadstart 10.1.2. I've a small team that will work on a new project.
    1. Which are the best practices to handle a project in a team environment? Besides using a version control system and so on. Is it advisable to use multiples Application Structures Files? One for each developer and for each "subapplication".
    2. A component of the system is a kind of "Task Management System" that needs to be available across all the "sub applications" Which is the best way to share that component between the applications?
    Best regards,
    Gerardo

    Gerardo,
    Yes, we usually create separate application structure files for sub systems. Each subsystem is then accessed through a global button, so the subsystems can keep their own generated menu structure. This technique is also used in the JHeadstart Demo application, shipped with release 10.1.2.1 or higher.
    For more info on team-based development, see this thread:
    Re: Jheadstart, JDeveloper and CVS and TeamWork
    Please create a new thread for your other question.
    Steven Davelaar,
    JHeadstart Team.

  • Best Practices for Staging

    Hey Guys!
    We are currently evaluating Oracle Enterprise Pack for Eclipse with one of our new projects.
    As we are moving from development phase to QA, we are looking for ways to standardize the staging process. Up to now we were unable to figure out how or where OEPE creates the artifacts (war & ear files) before they are deployed to the remote target environment (Weblogic Server) for tests during development. We had assumed that OEPE uses a simple ANT file or something similar we could use for our staging environments.
    Is there a Best Practice for Staging? What does Oracle recommend?
    Thanks in advance - Alex

    Hi,
    According to your post, my understanding is that you wanted to know the best practices to deploy SharePoint Portal in different SharePoint environment.
    If the site is not existing in the production server, we can save the site from the development server, and then import it to the production server.
    But if the site is already existing in the production server, we should follow these steps to just add the taxonomy and content types to the production server:
    Save the site from Dev as a template
    Import the template as solution in Visual Studio
    Remove unnecessary items from the solution(Please pay more      attention on it. If a content type/list... in the solution is existing in      the production site too, it will replace the
    same object existing in the      production after deployment)
    Package the solution
    Deploy the solution in the production
    For more detailed, please see:
    http://ahmedmadany.wordpress.com/2012/12/30/importing-sharepoint-solution-package-wsp-into-visual-studio-2010/
    There is a similar thread for your reference.
    http://social.technet.microsoft.com/Forums/en-US/7dcf61a8-1af2-4f83-a04c-ff6c439e8268/best-practices-guide-for-deploying-sharepoint-2010-from-dev-to-test-to-production?forum=sharepointgeneralprevious
    Thanks & Regards,
    Jason
    Jason Guo
    TechNet Community Support

  • Question about Best Practices - Redwood Landscape/Object Naming Conventions

    Having reviewed documentation and posts, I find that there is not that much information available in regards to best practices for the Redwood Scheduler in a SAP environment. We are running the free version.
    1) The job scheduling for SAP reference book (SAP Press) recommends multiple Redwood installations and using export/import to move jobs and other redwood objects from say DEV->QAS->PROD. Presentations from the help.sap.com Web Site show the Redwood Scheduler linked to Solution Manager and handling job submissions for DEV-QAS-PROD. Point and Shoot (just be careful where you aim!) functionality is described as an advantage for the product. There is a SAP note (#895253) on making Redwood highly available. I am open to comments inputs and suggestions on this issue based on SAP client experiences.
    2) Related to 1), I have not seen much documentation on Redwood object naming conventions. I am interested in hearing how SAP clients have dealt with Redwood object naming (i.e. applications, job streams, scripts, events, locks). To date, I have seen in a presentation where customer objects are named starting with Z_. I like to include the object type in the name (e.g. EVT - Event, CHN - Job Chain, SCR - Script, LCK - Lock) keeping in mind the character length limitation of 30 characters. I also have an associated issue with Event naming given that we have 4 environments (DEV, QA, Staging, PROD). Assuming that we are not about to have one installation per environment, then we need to include the environment in the event name. The downside here is that we lose transportability for the job stream. We need to modify the job chain to wait for a different event name when running in a different environment. Comments?

    Hi Paul,
    As suggested in book u2018job scheduling for SAP from SAPu2019 press it is better to have multiple instances of Cronacle version (at least 2 u2013 one for development & quality and other separate one for production. This will have no confusion).
    Regarding transporting / replicating of the object definitions - it is really easy to import and export the objects like Events, Job Chain, Script, Locks etc. Also it is very easy and less time consuming to create a fresh in each system. Only complicated job chains creation can be time consuming.
    In normal cases the testing for background jobs mostly happens only in SAP quality instance and then the final scheduling in production. So it is very much possible to just export the verified script / job chain form Cronacle quality instance and import the same in Cronacle production instance (use of Cronacle shell is really recommended for fast processing)
    Regarding OSS note 895253 u2013 yes it is highly recommended to keep your central repository, processing server and licencing information on highly available clustered environment. This is very much required as Redwood Cronacle acts as central job scheduler in your SAP landscape (with OEM version).
    As you have confirmed, you are using OEM and hence you have only one process server.
    Regarding the conventions for names, it is recommended to create a centrally accessible naming convention document and then follow it. For example in my company we are using the naming convention for the jobs as Z_AAU_MM_ZCHGSTA2_AU01_LSV where A is for APAC region, AU is for Australia (country), MM is for Materials management and then ZCHGSTA2_AU01_LSV is the free text as provided by batch job requester.
    For other Redwood Cronacle specific objects also you can derive naming conventions based on SAP instances like if you want all the related scripts / job chains to be stored in one application, its name can be APPL_<logical name of the instance>.
    So in a nutshell, it is highly recommend
    Also the integration of SAP solution manager with redwood is to receive monitoring and alerting data and to pass the Redwood Cronacle information to SAP SOL MAN to create single point of control. You can find information on the purpose of XAL and XMW interfaces in Cronacle help (F1). 
    Hope this answers your queries. Please write if you need some more information / help in this regard.
    Best regards,
    Vithal

  • Best practices for deployment from Dev /Staging /Production in SharePoint ?

    Hi All,
    What is a best practices to deploy SharePoint Portal to dev / staging / Production.
    I have custom solution deployed using WSP file. But I have done some changes using sharepoint designer.
    Like as Designer workflow, master pages etc.
    How can I deploy my document libraries and list to dev to prod using best practices?
    Thanks
    Balaji More

    Hi,
    According to your post, my understanding is that you wanted to know the best practices to deploy SharePoint Portal in different SharePoint environment.
    If the site is not existing in the production server, we can save the site from the development server, and then import it to the production server.
    But if the site is already existing in the production server, we should follow these steps to just add the taxonomy and content types to the production server:
    Save the site from Dev as a template
    Import the template as solution in Visual Studio
    Remove unnecessary items from the solution(Please pay more      attention on it. If a content type/list... in the solution is existing in      the production site too, it will replace the
    same object existing in the      production after deployment)
    Package the solution
    Deploy the solution in the production
    For more detailed, please see:
    http://ahmedmadany.wordpress.com/2012/12/30/importing-sharepoint-solution-package-wsp-into-visual-studio-2010/
    There is a similar thread for your reference.
    http://social.technet.microsoft.com/Forums/en-US/7dcf61a8-1af2-4f83-a04c-ff6c439e8268/best-practices-guide-for-deploying-sharepoint-2010-from-dev-to-test-to-production?forum=sharepointgeneralprevious
    Thanks & Regards,
    Jason
    Jason Guo
    TechNet Community Support

  • RDS Gateway Best practices Dual-Homed?

    Good Day,
    I am wondering what is a typical amount of time others see when end users launch a RemoteApp session that goes through the RDS Gateway.
    Our two RDS Gateway servers (entire environment is W2k12R2) seem slow to me. They are both Dual-Homed and with a Nic on the DMZ and Internal side of the network. Maybe I would be better off disabling the Internal Nics and reconfiguring the firewall rules
    so that everything routes through the DMZ nic?
    Steve J.

    Hi Steve,
    Thank you for posting in Windows Server Forum.
    Best practice for any server is depends on your environment scenario as you need to decide whether to place the gateway in DMZ or allow 443 to be opened to the internal network. Placing the RDS Gateway in DMZ is more secure, you can have more information from
    beneath article.
    RD Gateway deployment in a perimeter network & Firewall rules
    http://blogs.msdn.com/b/rds/archive/2009/07/31/rd-gateway-deployment-in-a-perimeter-network-firewall-rules.aspx
    For more detail and understand RD Gateway refer this article.
    Hope it helps!
    Thanks.
    Dharmesh Solanki
    TechNet Community Support
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Support, contact [email protected]

  • Please recommend a best practices book

    Can someone recommend a core java best practices book.
    I have read several books and yet when someone says use a tagging interface... I just look at them and wonder what they are talking about.. then I google and find out.. There may be several such terms that I am not aware of... How do I learn about these terms like POJOs etc.. People keep using words like POJOs and then I find out its nothing but a java class object.. but how do I know that?
    Someone once said "You know only 5%, 20% you know you dont know, 75% you dont even know that you dont know".. Which book can give me that 75% ?
    I know basic datastructures etc.. but when to apply what is what I want to learn.
    A book which talks about when to use which datastructure, good usage of interface examples, and best practices that are used in real application production environment. Can someone please recommend a good book like that ? Is there any ? Real cases?

    Maybe this should be a different thread but how do
    programmers organize folders for projects?
    Every company has a common java classes shared by
    programmers I presume? That must be located in a
    central access place and I am curious how they
    organize folders?Folders? Maybe. Version control system, with a "commons" tree? Better.
    Have a look at Matt Raible's AppFuse. It's a fine example organization for a project.
    As for me, I do something like this:
    I build with Ant: http://ant.apache.org
    /source is the directory where anything I check out of version control will go.
    Under that:
    /java - .java source files.
    /lib - 3rd party JARs (e.g., JDBC drivers, etc.)
    /jsp - JSPs
    /etc - deployment descriptors, like web.xml, etc.
    /xml - XML files
    /xsl - XSL stylesheets
    /html - HTML files
    /css - cascading stylesheets
    /sql - SQL scripts for creating, altering, populating relational tables.
    %

  • Enterpise Best Practices for iPad

    Is anyone aware of any documentation idnetifying best practices for securely deploying iPads in an enterprise environment?

    There is some information out there, though not as much as I think we are typically used to for enterprise environments. (It is a consumer device, and Apple is a consumer-driven company, and I don't fault them for that one bit.)
    Here is some documentation from Apple:
    http://www.apple.com/support/ipad/enterprise/
    Also, Jamf Software has some information regarding their Casper suite.
    We don't use it yet at my workplace, but I have heard good things about them.
    http://www.jamfsoftware.com/solutions/mobile-device-management
    Edit:
    And welcome to the forums!
    Message was edited by: tibor.moldovan

  • Lun Size best practice for UC apps and VMWare?

    Hi,
    We have UCS manager v2.1 with FI 6248 direct FC attached to NetApp with plenty of storage.
    Per following doc, Lun size for UC apps should be 500GB - 1.5TB and 4 to 8 VMs per Lun.
    http://docwiki.cisco.com/wiki/UC_Virtualization_Storage_System_Design_Requirements#Best_Practices_for_Storage_Array_LUNs_for_Unified_Communications_Applications
    We have four B200M3 blades and 3 to 4 UC apps (CUCM, Unity, UCCX) will be hosted on each blade. May add more VM the blades in the future.
    I am thinking four 1 TB Luns and one for each blades. (actually 8 Luns in toal, 4 boot luns for ESXi and 4 for UC apps).
    What is the best practice (or common deployment) to create Lun size and design?
    Thanks,
    Harry

    UC apps need low IO,nothing special,Reference vmware LUN design is ok.

  • Dell MD3620i connect to vmware - best practices

    Hello Community,
    I've purchased a Dell MD3620i with 2 x ports 10Gbase-T Ethernet on each controller (2 x controllers).
    My vmware environment consists of 2 x ESXi hosts (each with 2ports x 1Gbase-T) and an HP Lefthand storage( also 1Gbase-T). The switches I have are Cisco3750 that have only 1Gbase-T Ethernets.
    I'm going to replace this HP Storage with a DELL storage.
    As I have never worked with DELL storages, I need your help to answer my questions:
    1. What is teh best practices to connect vmware hosts to the Dell MD3620i ?
    2. what is the process to create a LUNs?
    3. Can I create multiply LUNs on only one disk group? or is the best practice to create one LUN on one disk group?
    4. How to set iSCSI 10GBase-T ports working on 1Gbps switch?
    5. Is the best practice to connect the Dell MD3620i directly to the vmware Hosts without switch?
    6. The old iscsi on HP storage is in a different network, can I do vmotion to move all virtual machines from one iSCSI network to another and then change the iSCSI IP addresses on vmware hosts without virtual machines interruption?
    7. Can I bundle two iSCSI ports to one 2Gbps interface and conenct to the switch? I'm using two switches, so I want connect each controller to each switch by bounding their interfaces to 2Gbps. My Question is, would be controller switched over to another controller if the Ethernet link falls on the switch?(in case one switch is rebooting)
    tahnks in advanse!

    TCP/IP basics: A computer cannot connect to 2 different (isolated) networks (e.g. 2 directly-attached cables between the server and a SAN's iSCSI port) that share the same subnet.
    Data corruption is highly unlikely if you were to share the same vlan for iSCSI, however, performance and overall reliability would be impacted.
    With a MD3620i, here are a few setup scenarios using the factory default subnets (and for direct-attached setups I had to add 4 additional subnets):
    Single switch (not recommended as the switch becomes your single point of failure):
    Controller 0:
    iSCSI port 0: 192.168.130.101
    iSCSI port 1: 192.168.131.101
    iSCSI port 2: 192.168.132.101
    iSCSI port 4: 192.168.133.101
    Controller 1:
    iSCSI port 0: 192.168.130.102
    iSCSI port 1: 192.168.131.102
    iSCSI port 2: 192.168.132.102
    iSCSI port 4: 192.168.133.102
    Server 1:
    iSCSI NIC 0: 192.168.130.110
    iSCSI NIC 1: 192.168.131.110
    iSCSI NIC 2: 192.168.132.110
    iSCSI NIC 3: 192.168.133.110
    Server 2: <end in 120>
    All ports plug into that 1 switch (obviously).
    If you only want to use 2 NICs for iSCSI, have server 1 use the 130 and 131 subnet, and server 2 use 132 and 133, server 3 then uses 130 and 131 again. This spreads the IO load between the iSCSI ports on the SAN.
    Dual switches (one VLAN for all the iSCSI ports on that switch though):
    NOTE: Do NOT link the switches together. This helps prevent issues that occur on one switch from affecting the other switch.
    Controller 0:
    iSCSI port 0: 192.168.130.101 -> To Switch 1
    iSCSI port 1: 192.168.131.101 -> To Switch 2
    iSCSI port 2: 192.168.132.101 -> To Switch 1
    iSCSI port 4: 192.168.133.101 -> To Switch 2
    Controller 1:
    iSCSI port 0: 192.168.130.102 -> To Switch 1
    iSCSI port 1: 192.168.131.102 -> To Switch 2
    iSCSI port 2: 192.168.132.102 -> To Switch 1
    iSCSI port 4: 192.168.133.102 -> To Switch 2
    Server 1:
    iSCSI NIC 0: 192.168.130.110 -> To Switch 1
    iSCSI NIC 1: 192.168.131.110 -> To Switch 2
    iSCSI NIC 2: 192.168.132.110 -> To Switch 1
    iSCSI NIC 3: 192.168.133.110 -> To Switch 2
    Server 2: <end in 120>
    Same note about using just 2 NICs per server for iSCSI. In this setup each server will still use both switches so that a switch failure should not take any of your servers' iSCSI connectivity down.
    Quad switches (or 2 VLANs on each of the 2 switches above):
    iSCSI port 0: 192.168.130.101 -> To Switch 1
    iSCSI port 1: 192.168.131.101 -> To Switch 2
    iSCSI port 2: 192.168.132.101 -> To Switch 3
    iSCSI port 4: 192.168.133.101 -> To Switch 4
    Controller 1:
    iSCSI port 0: 192.168.130.102 -> To Switch 1
    iSCSI port 1: 192.168.131.102 -> To Switch 2
    iSCSI port 2: 192.168.132.102 -> To Switch 3
    iSCSI port 4: 192.168.133.102 -> To Switch 4
    Server 1:
    iSCSI NIC 0: 192.168.130.110 -> To Switch 1
    iSCSI NIC 1: 192.168.131.110 -> To Switch 2
    iSCSI NIC 2: 192.168.132.110 -> To Switch 3
    iSCSI NIC 3: 192.168.133.110 -> To Switch 4
    Server 2: <end in 120>
    In this case using 2 NICs per server means the first server uses the first 2 switches and the second server uses the second set of switches.
    Direct attach:
    iSCSI port 0: 192.168.130.101 -> To server iSCSI NIC 1 (on an example IP of 192.168.130.110)
    iSCSI port 1: 192.168.131.101 -> To server iSCSI NIC 2 (on an example IP of 192.168.131.110)
    iSCSI port 2: 192.168.132.101 -> To server iSCSI NIC 3 (on an example IP of 192.168.132.110)
    iSCSI port 4: 192.168.133.101 -> To server iSCSI NIC 4 (on an example IP of 192.168.133.110)
    Controller 1:
    iSCSI port 0: 192.168.134.102 -> To server iSCSI NIC 5 (on an example IP of 192.168.134.110)
    iSCSI port 1: 192.168.135.102 -> To server iSCSI NIC 6 (on an example IP of 192.168.135.110)
    iSCSI port 2: 192.168.136.102 -> To server iSCSI NIC 7 (on an example IP of 192.168.136.110)
    iSCSI port 4: 192.168.137.102 -> To server iSCSI NIC 8 (on an example IP of 192.168.137.110)
    I left controller 1 on the "102" IPs for easier future changing back to just 4 subnets.

  • Best practice for designing a print enviroment​?

    Greetings,
    If there is a better location for this, please let me know.
    Goal:
    Redesign and redeploy my print enviroment with best practices in mind.
    Overview: VMWare enviroment running 2008 R2,  with ~200 printers. I have a majority of HP printers ranging from 10 years old to brand new. Laserjets, MFPs, OfficeJets, etc.. etc.. in addition to Konica, Xerox, and Savin copiers. Many of our printer models aren't support in 2008, let alone x64.
    Our future goals include eprint services, as well as a desire to manage print quality, and consumition levels through something like Web Jetadmin.
    Presently we have a 2003 x86 server running our very old printers and until 6 months ago the rest on a single 2008r2 x64 server. We ended up not giving it the attention of detail it needed and the drivers became very cluttered, this lead to a single UPD PCL6 update that ended up corrupting several drivers across the UPD PCL 5 and 6 spectrum. At that time we brought up a second 2008r2 server and began to migrate those affected. In some instances we were forced to manually delete the drivers off of the clients system32->Spool->Driver and reinstall.
    I haven't had much luck finding good best practice information and figured I'd ask. Some documents I came across suggested that I should isolate a Universal driver to a single server, such as 3 servers for PCL5, PCL6, and PS. Then there is the need to deal with my various copiers.
    I appreciate your advice, thank you!
    This question was solved.
    View Solution.

    This forum is focused on consumer level products.  For your question you may have better results in the HP Enterprise forum here.
    Bob Headrick,  HP Expert
    I am not an employee of HP, I am a volunteer posting here on my own time.
    If your problem is solved please click the "Accept as Solution" button ------------V
    If my answer was helpful please click the "Thumbs Up" to say "Thank You"--V

Maybe you are looking for

  • Is it a Bug in Java XML JAXP?!

    i have a sample XML file ========================== <?xml version='1.0' encoding='utf-8'?> <aaa>      <bbb>           <ccc>                Hi All           </ccc>      </bbb> </aaa> <?execute.somecde?> When transofrming the DOMSource to XML StreamRes

  • Xfce4 programs not shown in menu!?

    I installed xfce4 on 2 pcs. On both no xfce4 programs are shown in the menu of the taskbar. Internet,Office,... programs are alls shown, but the others like xfce4-settings-manager,... not. Does somebody know, how I can fix this?

  • Setting Filter on DNL_PLANT

    Hello, We would like to restrict the plants being created in the CRM.  I have tried setting a filter in DNL_PLANT but it doesn't allow me to add the table and field. Has anyone done this before and could you explain how you did it. Your help is appre

  • Itunes update locks phone to usb/itunes logo screen

    So for the past year whenever apple has an update it messes with my phone. This as I said first happened about a year ago. I updated my phone and in the middle of doing so itunes errored out and I was unable to continue with the update. My phone was

  • Raw device in ASM

    Hi, I am assigning raw devices to ASM on IBM-AIX 5.2. But the devices are not visible in ASM. Please somebody help me how to identify the disks in ASM or from AIX to check whether the disks are accessiable. I am able to ls -lrt and find the disk. but