Bundling Best Practices

Good morning All,
I know that this has been asked for before, and my apologies for doing so. However; I would like to provide my customer something that can describe Best Practices when it comes to building, deploying bundles, and policies (which can be handled in another posting). Especially in relation to the placement of these Bundles, whether they should be at the Device, User, or Workstation level (if an enterprise wide application or Policy)
There is a debate within the organization, that Policies, Bundles, should be deployed out to all geographic locations, instead of just being Deployed from the Top level down. In this case they have Group Policies which go out to every PC in the organization, and instead of doing that from the Workstation level, they deploy these Policy's from each geographic local. Which seems to not be according to best practices, or at least what I have always assumed to be Best Practices.
I have seen this from the Documentation site, http://www.novell.com/documentation/zenworks11/ - System Planning, Deployment, and Best Practices Guide, and I have found some others looking to put together a packet for the customer.
Thank you,
-DS

I know that the University of Uppsala has written up some, but they are
in Swedish. Anyway, as for your question, I would say that it depends.
Your one main goal should be management by exception.
Are policies, bundles etc different based on geography?
Anders Gustafsson (NKP)
The Aaland Islands (N60 E20)
Have an idea for a product enhancement? Please visit:
http://www.novell.com/rms

Similar Messages

  • What are Resource Bundle Best Practices techniques for Enterprise App?

    Regarding JDeveloper: 11.1.1.6.0, Studio Edition
    I was wondering if someone could provide advice on Best Practices for managing Resource Bundles for an international Enterprise Application.
    I have been reading textbooks and throughout the web, and I can find different options available. And I can find cautionary tales to get it right at the beginning of Development, but I cannot find Best Practices suggestions.
    For instance:
    - Should I use XLIFF Resource Bundle, Properties Bundle, or List Resource Bundle?
    - What are the benefits and disadvantages of storing the Key/Value pairs in the database?
    - It seems that storing in the db would make maintenance easier, because applications do not need to be redeployed, but would they be slower?
    - One textbook indicates that "One Bundle per Project" is preferred for ViewController Project, and "One Bundle Per File" is preferred for Model Project. However, I cannot help but think if the whole Enterprise used just one Resource Bundle, it would save typing cust_id/Customer Number in 10 different Bundles.
    - One text indicates how to maintain translated versions of Access Keys, if the Bundle is a Properties Bundle, but provides no assistance for other Resource Bundles.
    Advice regarding Best Practices would be quite helpful.
    Sincerely,
    Arie

    Anyone?

  • Locale Resource Bundle Best Practice

    Hi
    I have a Flex application that loads it's locale resource bundle from a service.
    So when the httpservice loads the bundle it populates an instance of a class "I18NBundle" that contains all the bundle properties. For instance you could do:
    i18nBundle.hello_message and it would return "Hello"
    I'm currently using Cairngorm for this project so this i18nBundle instance it's on the model locator.
    What I'm seeing and that I don't like is that for a component be able to get the bundle it must access to the model locator and then to the i18nBundle.
    Instead of that what I would like is that each component doesn't rely on this for getting the bundle.
    I guess I could create a "bundle" property on each component class and then pass it the reference to the bundle when it is instanciated. It seems it could be messy and a difficult task to initialize this property in some cases for instance on a datagrid cell renderer.
    Other could be transforming the I18NBundle class into a singleton and then when the service response it's recieved the singleton it's initialized. Since all application components should/must access the same locale bundle I guess this could be a better option.
    What do you think about this?
    Do you think there is some a better way to achieve this with Flex?
    Any opinion or recommendation would be appreciatted.
    (I don't wish to use the Adobe Flex proposal of having the properties file on the Flex project and the compiling them into swf.)
    thanks in advance.
    Polaco.
    ps: If you think I haven't expressed myself correctly please let me know and I will rewrite it.

    I have managed to extend IResourceBundle and added it to ResourceManager.
    The only problem now is that my bundle does not represent a language and it doesn't need a name either.
    Since the localization part is done on the webapplication and then the apporpiate bundle returned in the request's response.
    So it's locale attribute value is "".
    And it's name is "".
    I can display a property correctly if I use the following code:
    (resourceManager.getResourceBundle('','')).content.helloMessage
    but I doesn't work if I try to retrieve it like:
    resourceManager.getString("", "helloMessage");
    any ideas ?
    thanks

  • Best practice using clusters to create queue/notifier/bundles ?

    I have in a block diagram a queue, notifier and several instances of bundle cluster
    that all use the same data structure.   There is a cluster typedef for the data structure.
    Of course, each of these objects (define queue, define notifier, bundle)
    need to know how the cluster is defined.
    What is considered best practice?
    1) create a dummy instance of the cluster everywhere the data structure
        definition is needed (and hide them all on the FP)
    2) create only one instance and wire it to all the places it is needed
       But there is no data flow on this wire -- it's only the cluster *definition*
       that is being used, so this would seem to clutter up the BD unnecessarily.
    3) Create only one control instance of the cluster and use local variables
        everywhere else the cluster definition is needed.  It's _value_ is never
        assigned or read so there's no issue with race conditions.
    4) Some other way?
    If you had to clean up someone else's code, how would you hope to
    see this handled?
    It occurred to me in the course of writing this that where I have
    "unbundle... code ... bundle" I could wire the original bundle to 
    both "unbundle" and "bundle" -- but would this be too confusing
    and clutter the BD with unnecessary wire ?
    Thanks and Best Regards,
    -- J.
    Message Edited by JeffBuckles on 11-06-2008 03:17 PM
    Solved!
    Go to Solution.

    Discovered a rather unpleasant side-effect of using the typedef constant cluster this way. Since I'm using the typedef cluster constant only to communicate the definition of the cluster--not for dataflow--I resize it to the same size and shape as a control terminal.  This is shown in the first image, clean_cluster.jpg.   I was also careful to set "Autosizing->None" on all instances of the constant.   However, when I updated the typedef, all instances of the constant resized!  That's dozens of instances across many different VIs.   Why was "Autosizing->None" not respected?  Is there a better way to do this so that my BD doesn't get messed up if I have to update a typedef such as this one?
    Thanks and Best Regards,
    J. 
    Attachments:
    clean_cluster.png ‏2 KB
    Cluster_Updated.png ‏3 KB

  • Resource Bundles and Embedded fonts - best practice

    Hello,
    I am in digging into creating a localized app and I would
    like to use embedded fonts.
    Ideally, I would like to have two locales (for example), each
    with a different embedded font and/or unicode range.
    For example, how do I set up my localized app so that EN uses
    Arial (Latin Range), and JP uses Arial Unicode MS with Japanese
    Kanji unicode range ?
    Note that I do know how to embed fonts with different ranges,
    I just don't know how to properly embed them into resource bundles
    and access them easily in style sheets.
    Thanks!
    -Daniel

    Hello!
    This is a very pertinent question, however as many things in life there is no one size fits all here.
    We basically recommend, as best practice, to allocate for each specific context only the estimated needed resources. These values should always come from a previous study on the network patterns/load.
    To accomodate for growth and scalability it is strongly advised to initially keep as many resources reserved as possible and allocate the unused resources as needed. To accomplish this goal, you should created a reserved resource class, as you did already, with a guarantee of 20 to 40 percent of all ACE resources and configure a virtual context solely with the purpose of ensuring that these resources are reserved.
    As you might already know ACE protects resources in use, this means that when decreasing a context's resources, the resources must be unused before then can be reused by other context. Although it is possible to decrease the resource allocations in real time, it typically requires additional overhead to clear any used resources before reducing them.
    Based on the traffic patterns, number of connections, throughput, concurrent SSL connections , etc, for each of the sites you will be deploying you will have a better idea on what might be the estimated needed resources and then assign them to each of the contexts. Thus this is something that greatly depends on customer's network environment.
    Hope this helps to clarify your doubts.

  • Best practices for building menus using resource bundles?

    Greetings; I am curious to find out what the current best practices people are using to build menus/menu bars using resource bundles, specifically ListResourceBundle.
    What I am trying to figure out is how best to write my Swing application so it does not need to know what menu items it needs to grab from the resource bundle.
    The only idea I have come up with is this:
    class MyBundle extends ListResourceBundle {
    private Object[][] contents = {
            {"menubar", { {"menu.file.item", "blah"}, ....} }
    }Inside the GUI class:
    Object[][] menubar = resourceBundle.getObject("menubar");I would then iterate over the menu bar items and build the menu. I would have to use a naming scheme and then parse appropriately to know when to start a new menu, when a submenu occurs, etc.
    Is this the common practice, or does anyone know of a more clever way of doing this? I've searched various FAQs and googled about, but I have yet to come across any sort of tutorial or page that covers this.

    Anyone have any input on this? Am I close to the solution people are
    using out in real production environments?

  • SAP to Non-SAP Integration best Practices

    Hi Folks,
    Recently I demonstrated to few of my managers the integration of our SAP ISU with a 3rd Party MDUS System via SAP PI. A question which was repeatedly asked is 'Why SAP PI'? Isn't there any other way to do it? They did mention BAPIs and doing things directly in ABAP but I couldn't really answer as to how weigh one on the other in this particular scenario.
    I do know that there are standard ES Bundles for achieving integration with 3rd Party Systems via SAP PI, We can do the interface and message mappings but
    is it possible to achieve this integration with the 3rd Party MDUS System without using PI?
    3rd party MDUS can only integrate via its web Services so how would they get called?
    Whats the trade-off in case of Performance, Development Cost?
    I am looking for best practices, recommendations, trade-offs and possibilities. Your input is very much appreciated.
    Regards,
    Adil Khalil

    Hi Adil,
    The below blog might be useful
    Consuming Services with ABAP
    regards,
    Harish

  • Best practice for TM on AEBS with multiple macs

    Like many others, I just plugged a WD 1TB drive (mac ready) into the AEBS and started TM.
    But in reading here and elsewhere I'm realizing that there might be a better way.
    I'd like suggestions for best practices on how to setup the external drive.
    The environment is...
    ...G4 Mac mini, 10.4 PPC - this is the system I'm moving from, it has all iPhotos, iTunes, and it being left untouched until I get all the TM/backup setup and tested. But it will got to 10.5 eventually.
    ...Intel iMac, 10.5 soon to be 10.6
    ...Intel Mac mini, 10.5, soon to be 10.6
    ...AEBS with (mac ready) WD-1TB usb attached drive.
    What I'd like to do...
    ...use the one WD-1TB drive for all three backups, AND keep a copy of system and iLife DVD's to recover from.
    From what I'm reading, I should have a separate partition for each mac's TM to backup to.
    The first question is partitioning... disk utility see's my iMac's internal HD&DVD, but doesn't see the WD-1TB on the AEBS. (when TM is activity it will appear in disk utility, but when TM ends, it drops off the disk utility list).
    I guess I have to connect it via USB to the iMac for the partitioning, right?
    I've also read the benefits of keeping a copy of the install DVD's on the external drive... but this raises more questions.
    How do I get an image of the install DVD onto the 1TB drive?
    How do I do that? (install?, ISO image?, straight copy?)
    And what about the 2nd disk (for iLife?) - same partition, a different one, ISO image, straight copy?
    Can I actually boot from the external WD 1TB while it it connected to the AEBS, or do I have to temporarily plug it in via USB?
    And if I have to boot the O/S from USB, once I load it and it wants to restore from the TM, do I leave it USB or move it to the AEBS? (I've heard the way the backups are created differ local vs network)>
    I know its a lot of question but here are the two objectives...
    1. Use TM in typical fashion, to recover the occasion deleted file.
    2. The ability to perform a bare-metal point-in-time recovery (not always to the very last backup, but sometimes to a day or two before.)

    dmcnish wrote:
    From what I'm reading, I should have a separate partition for each mac's TM to backup to.
    Hi, and welcome to the forums.
    You can, but you really only need a separate partition for the Mac that's backing-up directly. It won't have a Sparse Bundle, but a Backups.backupdb folder, and if you ever have or want to delete all of them (new Mac, certain hardware repairs, etc.) you can just erase the partition.
    The first question is partitioning... disk utility see's my iMac's internal HD&DVD, but doesn't see the WD-1TB on the AEBS. (when TM is activity it will appear in disk utility, but when TM ends, it drops off the disk utility list).
    I guess I have to connect it via USB to the iMac for the partitioning, right?
    Right.
    I've also read the benefits of keeping a copy of the install DVD's on the external drive... but this raises more questions.
    Can I actually boot from the external WD 1TB while it it connected to the AEBS, or do I have to temporarily plug it in via USB?
    I don't think so. I've never tried it, but even if it works, it will be very slow. So connect via F/W or USB (the PPC Mac probably can't boot from USB, but the Intels can).
    And if I have to boot the O/S from USB, once I load it and it wants to restore from the TM, do I leave it USB or move it to the AEBS? (I've heard the way the backups are created differ local vs network)
    That's actually two different questions. To do a full system restore, you don't load OSX at all, but you do need the Leopard Install disc, because it has the installer. See item #14 of the Frequently Asked Questions *User Tip* at the top of this forum.
    If for some reason you do install OSX, then you can either "transfer" (as part of the installation) or "Migrate" (after restarting, via the Migration Assistant app in your Applications/Utilities folder) from your TM backups. See the *Erase, Install, & Migrate* section of the Glenn Carter - Restoring Your Entire System / Time Machine *User Tip* at the top of this forum.
    In either case, If the backups were done wirelessly, you must transfer/migrate wirelessly (although you can speed it up by connecting via Ethernet).

  • Best practice for managing a Windows 7 deployment with both 32-bit and 64-bit?

    What is the best practice for creating and organizing deployment shares in MDT for a Windows 7 deployment that has mostly 32-bit computers, but a few 64-bit computers as well? Is it better to create a single deployment share for Windows 7 and include both
    versions, or is it better to create two separate deployment shares? And what about 32-bit and 64-bit versions of applications?
    I'm currently leaning towards creating two separate deployment shares, just so that I don't have to keep typing (x86) and (x64) for every application I import, as well as making it easier when choosing applications in the Lite Touch installation. But I know
    each deployment share has the option to create both an x86 and x64 boot image, so that's why I am confused. 

    Supporting two task sequences is way easier than supporting two shares. Two shares means two boot media, or maintaining a method of directing the user to one or the other. Everything needs to be imported or configured twice. Not to mention doubling storage
    space. MDT is designed to have multiple task sequences, why wouldn't you use them?
    Supporting multiple task sequences can be a pain, but not bad once you get a system. Supporting app installs intelligently is a large part of that. We have one folder per app install, with a wrapper vbscript that handles OS detection. If there are separate
    binaries, they are placed in x86 and x64 subfolders. Everything runs from one folder via the same command, "cscript install.vbs". So, import once, assign once, and forget it. Its the same install package we use for Altiris, and we'll be using a Powershell
    version of it when we fully migrate to SCCM.
    Others handle x86 and x64 apps separately, and use the MDT app details to select what platform the app is meant for. I've done that, but we have a template for the vbscript wrapper and its a standard process, I believe its easier. YMMV.
    Once you get your apps into MDT, create bundles. Core build bundle, core deploy bundle, Laptop deploy bundle, etcetera. Now you don't have to assign twenty apps to both task sequences, just one bundle. When you replace one app in the bundle, all TS'es are
    updated automatically. Its kind of the same mentality as active directory. Users, groups and resources = apps, bundles and task sequences.
    If you have separate build and deploy shares in your lab, great. If not, separate your apps into build and deploy folders in your lab MDT share. Use a selection profile to upload only your deploy side to production. In fact I separate everything (except
    drivers) into Build and deploy folders on my lab server. Don't mix build and deploy, and don't mix Lab/QA and production. I also keep a "Retired" folder. When I replace an app, TS, OS, etcetera, I move it to the retired folder and append "RETIRED - " to the
    front of it  so I can instantly spot it if it happens to show up somewhere it shouldn't.
    To me, the biggest "weakness" of MDT is its flexibility. There's literally a dozen different ways to do everything, and there's no fences to keep you on the path. If you don't create some sort of organization for yourself, its very easy to get lost as things
    get complicated. Tossing everything into one giant bucket will have you pulling your hair out.

  • Best Practice for CQ Updates in complex installations (clustering, replication)?

    Hi everybody,
    we are planning a production setup of CQ 5.5 with an authoring cluster replicating to 4 publisher instances. We were wondering what the best update process looks like in a scenario like this. Let's say, we need to install the latest CQ 5 Update - which we actually have to -:
    Do we need to do this on every single instance, or can replication be utilized to distribute updates?
    If updating a cluster - same question: one instance at a time? Just one, and the cluster does the rest?
    The question is really: can update packages (official or custom) be automatically distributed to multiple instances? If yes, is there a "best practice" way to do this?
    Thanks for any help on this!
    Henning

    Hi Henning,
    The CQ5.5 servicepacks are distributed as CRX packages. You can replicate these packages and on the publishs they are unpacked and installed.
    In a cluster the situation is different: You have only 1 repository. So when you have installed the servicepack on one node, the new versions of bundles and other stuff is unpacked to the repository (most likely to /libs). Then the magic (essentially the JcrInstaller) takes care, that the bundles are extracted to started.
    I would not recommend to activate the service pack in a production environment, because then all publishs will be updated the same time. And as a restart is required, you might encounter downtimes. Of course you can make it work when you play with the replication agents :-)
    cheers,
    Jörg

  • Best practice on dynamically changing drop down in page fragment

    I have a search form, which is a page fragmant that is shared across many pages. This form allows users to selct a field from the drop down list, and search for a particular value, as seen in the screenshot here:
    http://photos1.blogger.com/blogger2/1291/174258525936074/1600/expanded.5.gif
    Please note that the search options are part of page fragmant embedded within a page - so that I can re-use it across many pages.
    The drop down list changes,based on the page this fragment is embedded. For users page, it will be last name, first name, etc. For parts page, it will be part number, part desc., etc.
    Here is my code:
              Iterator it=getTableRowGroup1().getTableColumnChildren();
            Option options[]=new Option[getTableRowGroup1().getColumnCount()];
            int i=0;
            while (it.hasNext()){
                TableColumn column=(TableColumn)it.next();
                if (column.getSort()!=null){
                    options=new Option(column.getSort().toString(), column.getHeaderText());
    }else{
    options[i]=new Option(column.getHeaderText(), column.getHeaderText());
    i++;
    search search=(search)getBean("search");
    search.getSearchDDFieldsDefaultOptions().setOptions(options);
    This code works, but it gives me all fields of the table available in the drop down. I want to be able to pick and choose. May be have an external properties file associated with each page, where I can list the fields available for search drop down??
    What is the best practice to load the list of options available for drop down on each page? (i.e last name, first name, etc.) I can have the code embedded in prerender of each page, or use sort of a resouce bundle for each page or may be use a bean?

    I have to agree with Pixlor and here's why:
    http://www.losingfight.com/blog/2006/08/11/the-sordid-tale-of-mm_menufw_menujs/
    and another:
    http://apptools.com/rants/menus.php
    Don't waste your time on them, you'll only end up pulling your hair out  :-)
    Nadia
    Adobe® Community Expert : Dreamweaver
    Unique CSS Templates |Tutorials |SEO Articles
    http://www.DreamweaverResources.com
    Book: Ultimate CSS Reference
    http://www.sitepoint.com/launch/005dfd4/3/133
    http://twitter.com/nadiap

  • Best practice recommendation--BC set

    Dear friends,
       I am using BC set concept to capture my configurations. I am a PP Consultant.
    Let us consider, one scenario, Configure plant parameters in t.code OPPQ.
    My requirement is:
    A.   Define Floats: (Schedule Margin key)
    SM key: 001
    opening period: 1day
    Float before prod: 2day
    Float After prod: 1 day
    Release period: 1 day
    B.   Number range
    Manitain internal number range as: 10-from:010000000999999999. (for planned orders)
    This is my configuration requirement.
    Method M1:
    Name of the BC set: ZBC_MRP1
    while creating BC set first time, while defining floats, i have wrongly captured/ activated the opening periodas 100, instead of 001. But i have correctly captured the value for number range (for my planned orders)
    Now if u see the activation log for my BC set, my BC set is in "GREEN" light--Version1, successfully activated, but activated values are wrong)
    So, i want to change my BC set values. Now i want to reactivate my BC set with correct value. Now i am again activating the same BC set with corret value of opening period (Value as 001 ). After reactivating the BC set, if i get into my BC set activation log, one more version (version 2) has appeared with "GREEN" light.
    So in my activation log, two BC sets are visible.
    If i activate Version 1---wrong values will be updated in configuration
    If i activate Version 2---corrrect values will be activated in configurations
    But both versions can be activated at any point of time. The latest activated version will be alwyas in top.
    <b>So method 1 (M1) talks about, with one BC set name, maintain different versions of BC set.</b>...Based on your requirement activate the versions
    Method 2 (M2)
    Instead of creating versions within a same BC set, create one more BC set to capture new values.
    So if i activate second BC set, the configuration will be updated.
    Please suggest me, which method is best practice( M1 or M2)?
    Thanks
    Senthil

    I am familiar with resource bundles, but wonder if there is a better approach within
    JDeveloper. Resourcebundles are the java-native way of handling locale-specific texts.
    Are there any plans to enhance this area in 9.0.3? For BC4J, in 9.0.3, all control-hints and custom-validation messages (new feature) are generated in resource-bundles rather than xml-files to make it easier to "extend" for multiple locales.

  • Best Practice: Usage of the ABAP Packages Concept?

    Hi SDN folks,
      I've just started on a new project - I have significant ABAP development experience (15 years+) - but one thing that I have never seen used correctly is the Package concept in ABAP - for any of the projects that I have worked on.
    I would like to define some best practices - about when we should create packages - and about how they should be structured.
    My understanding of the package concept is that they allow you to bundle all of the related objects of a piece of development work together. In previous projects - and almost every project I have ever worked on - we just have packages ZBASIS, ZMM, ZSD, ZFI and so on. But this to me is a very crude usage of packages, and really it seems that we have not moved on passed the 4.6 usage of the old development class concept - and it means that packages do not really add much value.
    I read in the SAP PRESS Next Generation ABAP book (Thomas Ljung, Rich Hellman) (I only have the 1st edition) - that we should use packages for defining separation of concern for an application. So it seems there they are recommending that for each and every application we write - we define at the least 3 packages - one for Model, one for Controller and one for view based objects. It occurs to me that following this approach will lead to a tremendous number of packages over the life cycle of an implementation, which could potentially lead to confusion - and so also add little value. Is this really the best practice approach? Has anyone tried this approach across a full blown implementation?
    As we are starting a new implementation - we will be running with 7 EHP2 and I would really like to get the most out of the functionality that is provided. I wonder what others have for experience in the definition of packages.
    One possible usage occurs to me that you could define the packages as a mirror image of the application business object class hierarchy (see below). But perhaps this is overcomplicating their usage - and would lead to issues later in terms of transportation conflicts etc.:
                                          ZSD
                                            |
                    ZSOrder    ZDelivery   ZBillingDoc
    Does anyone have any good recommendations for the usage of the ABAP Package concept - from real life project experience?
    All contributions are most welcome - although please refrain from sending links on how to create packages in SE80
    Kind Regards,
    Julian

    Hi Julian,
    I have struggled with the same questions you are addressing. On a previous project we tried to model based on packages, but during the course of the project we encountered some problems that grew overtime. The main problems were:
    1. It is hard to enforce rules on package assignments
    2. With multiple developers on the project and limited time we didn't have time to review package assignment
    3. Devopelers would click away warnings that an object was already part of another project and just continue
    4. After go-live the maintenance partner didn't care.
    So, my experience is is that it is a nice feature, but only from a high level design point of view. In real life it will get messy and above all, it doesn't add much value to the development. On my neew assignment we are just working with packages based on functional area and that works just fine.
    Roy

  • Best practice for heirachical data

    First off, I have to say that JMX in Java 6 is terrific stuff. Bundling jconsole in with Java has made JMX adoption so much easier for us.
    Now, to my question. We have read-only hierarchical data (think a DOM tree) that we would like to publish via JMX. What is the best practice? We see two possibilities:
    1. Publish each node of the tree with it's own object name and type. This will allow jconsole to display the information in the tree control.
    2. Publish just the root of the tree with an object name and type and then use CompositeType to describe the nodes of the tree. This means you look at the tree in the "Attribute Value" panel of jconsole.
    Is there any best practices for such data? We have implemented #2 and it works but we are wondering if long term this might lead to unforeseen consequences.
    Thanks in advance.
    --Marty                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

    Path,
    I did go with #1 and it worked out great. Every node in our tree is an ObjectName node. Works very well for us.
    --Marty                                                                                                                                                                                                                                                                   

  • SLES OES Patching using ZCM - Best Practice?

    Have recently updated an underused ZLM 7.3 to ZCM 11 and have started to utilise it to perform patching of SLES 10 SP3 and OES 2.
    At present I have a Novell subscription that downloads from novell NU as Patches with
    Create Category based Bundle Groups on.
    I hadn't at this point got my head around the sandbox but now want to utilise this as I don't want to implement it willy nilly and have patches deployed and installed without testing first.
    So first question is whether there is a best practice document for using ZCM 11 for patching SLES 10 and Sles/oes boxes that describes how best to use the sandbox.
    Next question is whether sandbox applies to the patches or only monolithic bundles
    At the end of the day it would be great to have a best practice document on how to utilise ZCM with change control over keeping SLES10/OES boxes patched up to date

    Hello,
    updates via ZCM is also a problem for me.
    At least i could give a little response from testing.
    As stated in the oes2 documentation patches had to be used for updates. At that point i had the problem that the subscription only created a bundle group which included the patches.
    The monolithic bundle seems to be an ordinary update (zcm docu states rpm -U ...) and so should not be used for oes2 servers.
    I found out that its ok to rename the patch bundle group created by the subscription to fix the patchlist. I use bundle names which include the date to keep updates from sles and oes in sync.
    So you should be able to do updates with these bundlegroups without problem but it is not very nice because each patch bundle had its on log report and entry in the bundle list.
    So to keep a global view is not very easy.
    The other problem is that there is no official documentation neither on the oes2 or the zcm site how to correct update/patch a sles or oes system with zcm (without patchlink).
    Hope this will be included soon as all linux satellite server will have the problem that this is the only way to get updates without direkt connection to the internet - at least as far as i can see.

Maybe you are looking for

  • Mail 3.1; Save PDF to web receipts folder

    When I am mailed passwords for downloaded software, I need to save such important information in a preferred place. The web receipts folder is a perfect place in this instance. However, I can't seem to get my Mail application to successfully complete

  • I have a problem with Project Administrator Menu

    I'm trying to cut some projects with Premiere CS4 for free disk space and, though I try, the new projects have the same size as the originals. I use the Create new project cut option and I use only the final assembly sequence with the video exclusive

  • Windows XP Still Camera empty

    Hi all, Following problem: I have an entry "digital till camera" in my device manager and workspace. But wenn I open it in the workspace, it is empty. When I tell windows to transfer images from the camera to the computer, it says, that it can not fi

  • Exporting JSF/JSP to Excel

    Hi, I have a JSP(JSF) containing only 1 table with data coming from a database. I use: <jsp:directive.page contentType="application/vnd.ms-excel"> to enable the users export these data to Excel. This is working fine on Win2K using Office 2K, but it g

  • Macros in msword

    HI all can anybody tell how to find macro coding in msword, i need to use it in ole, very urgent .