Subclass design problems/best practices

Hello gurus -
I have a question problem regarding the domain objects I'm sticking in my cache. I have a Product object - and would like to create a few subclasses - say BookProduct and MovieProduct (along with the standard Product objects). These really need to be contained in the same cache. The issue/concern here is that both subclasses have attributes that I'd like to index AND query on.
When I try to create an index on the subclasses attributes when there are just "standard" products - I get the following error (which only exists on one of the subclasses):
2010-10-20 11:08:43.280/227.055 Oracle Coherence GE 3.5.2/463 <Error> (thread=DistributedCache:MyCache, member=2): Exception occured during index rebuild: java.lang.RuntimeException: Missing or inaccessible method: com.test.domain.product.Product.getAuthors()
So I'm not sure the indexing is working or stopping once it hits this exception.
Furthermore, I get a similar error when attempting to Filter based on that attribute. So if I want to add the following filter:
Filter filter = new ContainsAnyFilter( "getAuthors", authors );
I will receive the following exception:
Caused by: Portable(java.lang.RuntimeException): Missing or inaccessible method: com.test.domain.product.Product.getAuthors()
What is considered the best practices for this assuming these really should be part of the same names cache? Should I attempt to subclass the extractors to "inspect" the Object for its class type during indexing or applying filters? Or should I just add all the attribute in the BookProduct and MovieProduct into the parent object and just forget about subclassing? That seems to have a pretty high "yuck" factor to me. I'm assuming people have run into this issue before and am looking for some best practices or perhaps something that deals with this that I'm missing. We're currently using Coherence 3.5.2. Not sure if it matters, but we are using the POF format for serialization.
Thanks!
Chris

Hi Chris,
I had a similar problem. The way I solved it was to use a subclass of the ChainedExtractor that catches all RuntimeException occurring during the extraction like the following:
* {@link ChainedExtractor} that catches any exceptions during extraction and returns null instead.
* Use this for cases where you're not certain that an object contains that necessary methods to be extracted.
* F.e. an object in the cache does not contain the getSomeProperty() method. However other objects do.
* When these are put together in the same cache we might want to use a {@link ChainedExtractor} like the following:
* new ChainedExtractor("getSomeProperty.getSomeNestedProperty"). However this will result in a RuntimeException for those entries that
* don't contain the object with the someProperty. Using the this class instead won't result in the exception.
public class SafeChainedExtractor extends ChainedExtractor
     public SafeChainedExtractor()
          super();
     public SafeChainedExtractor( String sMethod )
          super( sMethod );
     @Override
     public Object extract( Object entry )
          try
               return super.extract( entry );
          catch(RuntimeException e)
               return null;
     @Override
     public Object extractFromEntry( Entry entry )
          try
               return super.extractFromEntry( entry );
          catch(RuntimeException e)
               return null;
}For all indexes and filters we then use extractors that subclassed the SafeChainedExtractor like the following:
public class NestedPropertyExtractor extends SafeChainedExtractor
     private static final long serialVersionUID = 1L;
     public NestedPropertyExtractor()
          super("getSomeProperty.getSomeNestedProperty");
//adding an index:
myCache.addIndex( new NestedPropertyExtractor(), false, null );
//using a filter:
myCache.keySet(new EqualsFilter(new NestedPropertyExtractor(), "myNestedProperty"));This way, the extractor will just return null when a property doesn't exist on the target class.
Regards
Jan

Similar Messages

  • Design Patterns/Best Practices etc...

    fellow WLI gurus,
    I am looking for design patterns/best practices especially in EAI / WLI.
    Books ? Links ?
    With patterns/best practices I mean f.i.
    * When to use asynchronous/synchronous application view calls
    * where to do validation (if your connecting 2 EIS, both EIS, only in WLI,
    * what if an EIS is unavailable? How to handle this in your workflow?
    * performance issues
    Anyone want to share his/her thoughts on this ?
    Kris

              Hi.
              I recently bought WROX Press book Professional J2EE EAI, which discusses Enterprise
              Integration. Maybe not on a Design Pattern-level (if there is one), but it gave
              me a good overview and helped me make some desig decisions. I´m not sure if its
              technical enough for those used to such decisions, but it proved useful to me.
              http://www.wrox.com/ACON11.asp?WROXEMPTOKEN=87620ZUwNF3Eaw3YLdhXRpuVzK&ISBN=186100544X
              HTH
              Oskar
              

  • Process modelling : Design patterns & Best Practices

    Hi
    Could some one please suggest/share any technical information or documents tha's related to 'Process modelling - Design Patterns & Best Practices'
    Thanks in Advance
    Santosh K.
    Edited by: Santosh539 on Jul 29, 2010 4:07 PM

    Hi Santosh,
    There is no specific site with all the information you asked for.
    But I think these links would be helpful...
    on Work Flow Patterns: http://www.workflowpatterns.com/
    on BPM Service Pattern: http://enterprisearchitecture.nih.gov/ArchLib/AT/TA/WorkflowServicePattern.htm
    HTH
    Sharma

  • Design Pattern / Best Practice Question

    Hi,
    I have been using Flex for a while now, but there is a
    scenario which I still have not found a solution I'm entirely happy
    with. I'm wondering if anyone else out there might have suggestions
    on a design pattern or best practice.
    Suppose I have a view which depends on model data which
    resides in some back end systems. That model data may or may not
    have been loaded (e.g. via a web service or remote object call) at
    the time the view is displayed.
    I don't know if the user will ever visit this part of the
    application so I would prefer to defer retrieval of the data until
    the user actually navigates to this view. Or I want to retrieve the
    data each time the view is displayed because the data is dynamic
    and could change between one presentation of the view and the next.
    Because the data comes from several systems, I cannot simply
    make one service call and display the view when it completes and
    all the data is available. I need to call several services which
    could complete in any order but I only want to display my view
    after I know all of them have completed and all of the model data
    is available. Otherwise, I can present the user an incomplete view
    (e.g. some combo boxes are empty until the corresponding service
    call to get the data completes).
    The solution I like best so far is to dispatch a single event
    (I am using Cairngorm) handled by a single command which acts as
    the caller and responder for all of the services. This command then
    remembers which responses it has received and dispatches another
    event to navigate to the view once all the results have returned.
    If the services being called are used in different
    combinations on different screens, this results in proliferation of
    events and commands. An event and command for each service and
    additional events and commands to bundle the services and the
    handling of their responses in the right combinations for each of
    the views.
    Another approach is to have some helper class listen for all
    of the model changes and only display the view when the model
    enters some state that is acceptable. It is sometimes difficult to
    determine just by looking at the model whether it is in the right
    state (e.g. how can I tell that a collection is the new collection
    that should just have been requested versus an old one lingering
    from a previous call). The logic required can get kind of
    convoluted and brittle.
    Basically, all of the solutions I've come up with so far seem
    less than ideal and a little hackish. I keep thinking there is some
    elegant solution out there that I am just missing ... but so far,
    no luck finding it. Thoughts?
    Thanks.
    Bill

    i think a service class is right - to coordinate your calls.
    i would have 1 event per call (so you could listen to individual
    responses if you wanted to).
    then i would use a flag. if you want to check for staleness,
    you would probably want two objects to map your service flag to
    lastRequested and lastCompleted. when you check, check if it's
    completed, and if it's not stale and that your lastRequested is
    less than lastCompleted (meaning that you're not currently waiting,
    i.e. you've returned since making a request). then make the request
    and update your lastRequested.
    here's a snippet of what i mean.
    ./paul
    public static const SVC1_LOADED:int = 1;
    public static const SVC2_LOADED:int = 2;
    public static const SVC3_LOADED:int = 4;
    public static const SVCALL_LOADED:int = 7;
    private var completedFlag:int = 0;
    then each call would have it's own callback.
    private function onSvc1Complete( evt:Event):void {
    completedFlag |= SVC1_LOADED;
    lastCompleted[ SVC1_LOADED ] = getTimer();
    dispatchEvent( new Event("svc1complete") );
    checkDone();
    private function checkDone():void{
    if( completedFlag == SVCALL_LOADED )
    dispatchEvent(new Event( "allLoaded" ));

  • Network Design Review - Best Practices

    Looking to start a discussion around best practices for inbound network design at the core. 
    The planned devices are as followings:
    Edge Routing / DMVPN - Cisco 2951
    Cisco UCM / IP Phone VPN Concentrator - Cisco ASA 5512-X
    Cisco AnyConnect SSL Client Concentrator - Cisco ASA 5515-X
    Cisco FirePower / IPS Device - Cisco ASA 5515-X
    The plan is as follows:
    All traffic enters through the 2951. 
    DMVPN traffic will go directly to the FirePower Device and then to the core network.
    IP Phones will pass-through 2951, enter 5512-X for VPN, go to FirePower and then to the core network.
    AnyConnect Clients will pass-through 2951, enter 5515-X for VPN, go to FirePower and then to the core network. 
    Wondering if anyone else has completed a similar setup and any issues you may have fun into. 
    Basic diagram attached. 
    Thanks!

    There really isn't a true two factor authentication you can just do with radius unless its ISE and your doing EAP Chaining.  One way that is a workaround and works with ACS or ISE is to use "Was machine authenticated".  This again only works for Domain Computers.  How Microsoft works:) is you have a setting for user or computer... this does not mean user AND computer.  So when a windows machine boots up, it will sen its system name first and then the user credentials.  System name or machine authentication only happens once and that is during the boot up.  User happens every time there is a full authentication that has to happen.
    Check out these threads and it explains it pretty well.
    https://supportforums.cisco.com/message/3525085#3525085
    https://supportforums.cisco.com/thread/2166573
    Thanks,
    Scott
    Help out other by using the rating system and marking answered questions as "Answered"

  • Populating users and groups - design considerations/best practice

    We are currently running a 4.5 Portal in production. We are doing requirements/design for the 5.0 upgrade.
    We currently have a stored procedure that assigns users to the appropriate groups based on the domain info and role info from an ERP database after they are imported and synched up by the authentication source.
    We need to migrate this functionality to the 5.0 portal. We are debating whether to provide this functionality by doing this process via a custom Profile Web service. It was recommended during ADC and other presentation that we should stay away from using the database security/membership tables in the database directy and use the EDK/PRC instead.
    Please advise on the best way to approach(With details) this issue. We need to finalize the best approach to take asap.
    Thanks.
    Vanita

    So the best way to do this is to write a custom Authentication Web Service.  Database customizations can do much more damage and the EDK/PRC/API are designed to prevent inconsistencies and problems.
    Along those lines they also make it really easy to rationalize data from multiple backend systems into an orgainzation you'd like for your portal.  For example you could write a Custom Authentication Source that would connect to your NT Domain and get all the users and groups, then connect to your ERP system and do the same work your stored procedure would do.  It can then present this information to the portal in the way that the portal expects and let the portal maintain its own database and information store.
    Another solution is to write an External Operation that encapsulates the logic in your stored procedure but uses the PRC/Server API to manipulate users and group memberships.  I suggest you use the PRC interface since the Server API may change in subtle ways from release to release and is not as well documented.
    Either of these solutions would be easier in the long term to maintain than a database stored procedure.
    Hope this helps,
    -Akash

  • IOS Timer Problem - Best practice?

    Hey there,
    I'm using a global "idle timer" and for some reason, after the app is started the first time the timer event is launched it will launch dozens of times.
    After 1..2 regular events all becomes normal. Any idea or tips?
    Using AIR SDK 2.7
    Best,
    Cedric

    Hello,
    I wrote this workaround for Timer et setTimeOut:
    https://github.com/jonasmonnier/Mobilo/tree/master/src/com/mobilo/time
    A Timer :
    var id:int = Interval.create(method, delay, repeatCount);
    A Timeout :
    var id:int = Timeout.create(method, delay, params);
    Examples here :
    https://github.com/jonasmonnier/Mobilo/tree/master/src/test
    They are all based on my Tick class to use an unique ENTER_FRAME

  • OID DIT tree design. Best practice.

    Can I extend the orcluser object class to include all my application related attributes and define a new object class appusr to define all the user attributes. Similarly I have extended the OrclGroup to appgroup class. And I have configured "User Object Classes" in the User Entry management in OIDDAS to look into my new class and attributes. I didn't modify the defalut "User Search Context" and "Group Search Context". Is this the correct approach if we need to extend application specific user and group information. If so, will this arrangement work well for ALL the associated 9iAS components like Portal, SSO etc...
    Apart from the userpassword reset issue for upcoming releases, is there any portability/compatability issues with version upgrades?

    Hi,
    don't have the answer to your question but you might find some guidlines in the 'Oracle9i Directory Service Integration and Deployment Guide' and the chapter about 'Deploying Oracle Products with Oracle Internet Directory'
    http://download-west.oracle.com/docs/cd/B10501_01/network.920/a96579/products.htm#1008968
    --Olaf                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

  • Best Practice for Designing Database Tables?

    Hi,
    I work at a company for tracking devices (GPS Devices). Our SQL Server database is designed to have a table for each device we sell, currently there is 2500 tables in our database and they all have the same columns they only differ in table name. Each device
    sends about 4K records per day.
    currently each table hold from 10K records to 300K records
    What is the best practice to design a database in this situation? 
    When accessing database from a C# application, which is better to use, direct SQL commands or views? 
    a detailed description about what is best to do in such scenario would be great. 
    Thanks in advance.
    Edit:
    Tables columns are:
    [MessageID]
          ,[MessageUnit]
          ,[MessageLong]
          ,[MessageLat]
          ,[MessageSpeed]
          ,[MessageTime]
          ,[MessageDate]
          ,[MessageHeading]
          ,[MessageSatNumber]
          ,[MessageInput]
          ,[MessageCreationDate]
          ,[MessageInput2]
          ,[MessageInput3]
          ,[MessageIO]

    Hello Louis, thank you so much for your informative post. I'll describe in detail what situations I came through my 9 months of work in the company (working as a software engineer, but I am planning to take over database maintenance since no one is maintaining
    it right now and I cannot do anything else in the code to make it faster)
    At every end of the month our clients generate report for the previous month for all their cars, some clients have 100+ cars, and some have few. This is when real issue start, they are calling their data from our server through internet while having 2000
    unit sending data to our server, they keep on getting read time out since SQL Server gives priority to insert and hold all select commands. I solved it temporary in the code using "Read Uncommitted" once I initialize a connection through C#. 
    The other issue is generating reports for a month or two takes lots of time when selecting 100+ units. Thats what I want to solve, the problem is the one who wrote the C# app used hard coded SQL Statements
    AND
    the company is refusing to upgrade from SQL Server 2003 and Windows Server 2003. 
    Now talking about reports, there are summary reports, stops reports, zone reports ..etc most of them depend usually on at least MessageTime, MessageDate, MessageSpeed, MessageIO and MessageSatNumber.
    So from your post I conclude that for now I need to set snapshots so that select statements don't get kicked out in favor for insert commands, but does SQL Server automatically select from the snapshots or do I have to tell it to do so? 
    Other than proper indexing what else I need? Tom
    Phillips suggested Table partitioning but I don't think it is needed in my case since our database size is 78GB
    When I run code analysis on the app, Visual Studio tells me I better use stored procedures, views than using hard coded Select Statements, what difference will this bring me when talking about performance?
    Thanks in advance. 

  • Best practice in dreamweaver based web design

    Hi DW Gurus
    I have been learning dreamweaver at work and at home for a while now and I am very interested in getting some details about best practices.
    I am aware that my designs are still quite simple and don't expand dreamweaver anyway near its abilities, however as a novice I am sticking with DW CS4 and trying to developing foundation skills before trying to move on to greater adventures in web design.  My dream is to eventually leave my employer and start my own freelance design business that I can manage while I am travelling in SE Asia (yeah I know - dream on .....). I am saving for my CS4 master suite license - my current one belongs to my employer.
    In particular I have just started to do some free design a community organisation - I am trying to develop a portfolio of sites that I can show new clients (with money) what they could expect for their investments.
    I am interested in getting some ideas about work flows - design standards - whether designers use the dreamweaver layouts - whether to code or to use the WYSIWYG.
    Anyone care to share how they plan and go about their design work.
    In my past life I was a SQL developer and a C#.net programmer - in those feilds there are standards for code and applications . Are there industry standards that would point to good or better designing?
    Thanks for your time and interest in my topic,
    Respect,
    Doug

    The first thing you need to learn is html and css, without knowing the programming language will only make DW a hard tool to use.  You need to learn coding to WC3 standards.
    FREE HTML & CSS Tutorials  - http://w3schools.com/
    http://reference.sitepoint.com/css
    http://reference.sitepoint.com/html
    Validation tools that you will need to bookmark.
    HTML Validator - http://validator.w3.org
    CSS Validator - http://jigsaw.w3.org/css-validator/
    Hopefully you use Firefox, and if you do, the web developer toolbar is essential:
    http://chrispederick.com/work/web-developer/
    Yes, you can rely somewhat on DW's layout mode (or what features are left of it), but this will only lead to problems when it comes time to troubleshoot any problem pages that do not render correctly across the various browsers.
    A paying client will expect a well designed, fully functional website and you need to have the programming skills to accomplish this for them.
    Not sure how far advanced you are but the Adobe site has quite a few CSS based tutorials - and I would strongly urge you to go through the series listed below - to get up to speed on the correct way to lay out a page.
    Creating your first website (series)
    http://help.adobe.com/en_US/Dreamweaver/10.0_Using/WS42d4a1c0291fbe4e59147ede1232ff9686c-8 000.html
    LIST OF CSS TUTORIALS ON ADOBE SITE:
    http://www.adobe.com/devnet/dreamweaver/css.html
    If you use Fireworks as a tool to create your layouts, then the following tutorial may be of benefit:
    TAKING FIREWORKS COMP TO DREAMWEAVER:
    http://www.adobe.com/devnet/dreamweaver/articles/dw_fw_css_pt1.html
    Good luck in your future endeavours 
    Nadia
    Adobe Community Expert : Dreamweaver
    Unique CSS Templates | Tutorials | SEO Articles
    http://www.DreamweaverResources.com
    Web Design & Development
    http://www.perrelink.com.au
    http://twitter.com/nadiap

  • OBIEE Best Practice Data Model/Repository Design for Objectives/Targets

    Hello World!
    We are faced with a design question that has become somewhat difficult and we need some help. We want to be able to compare side-by-side actual measures with their corresponding objectives/targets. Sounds simple. But, our objectives are static (not able to be aggregated) with multi-dimensionality and multi-levels. We need some best practice tips on how to design our data model and repository properly so that we can see the objective/target for a measure regardless of the dimensions that are used in the criteria and regardless of the level.
    Here is some more details:
    Example of existing objective table.
    Dimension1
    Dimension2
    Dimension3
    Obj1
    Obj2
    Quarter
    NULL
    NULL
    NULL
    .99
    1.8
    1Q13
    DIM1VAL1
    NULL
    NULL
    .99
    2.4
    1Q13
    DIM1VAL1
    DIM2VAL1
    NULL
    .98
    2.41
    1Q13
    DIM1VAL1
    DIM2VAL1
    DIM3VAL1
    .97
    2.3
    1Q13
    DIM1VAL1
    NULL
    DIM3VAL1
    .96
    1.9
    1Q13
    NULL
    DIM2VAL1
    NULL
    .97
    2.2
    1Q13
    NULL
    DIM2VAL1
    DIM3VAL1
    .95
    2.0
    1Q13
    NULL
    NULL
    DIM3VAL1
    .94
    3.1
    1Q13
    - Right now we have quarterly objectives set using 3 different dimensions. So, if an author were to add one or more (or zero) dimensions to their criteria for a given measure they could get back a different objective. They could add Dimension1 and get 99%. They could add Dimension1 and Dimension2 and get 98%. They could add all three dimensions and get 97%. They could add zero dimensions (highest grain) and get 99%. Using our existing structure if we were to add a new dimension to the mix the possible combinations would grow dramatically. (Not flexible)
    - We would like our final solution to be flexible enough so that we could view objectives with altogether different dimensions and possibly get different objectives.
    - We currently have 3 fact tables with 3+ conformed dimension tables and a few unique dimension tables.
    Could anyone share a similar situation where you have implemented a data model structure with the proper repository joins to handle showing side-by-side objectives/targets where the objectives were static and could be displayed at differing levels with flexible dimensions as described?
    Any help would be greatly appreciated.

    hi..yes this suggestion is nice...first configure the sensors(activity or variable) ..then configure the sensor action as a JMS Topic which will in turn insert the data into a DB..Or when u configure the sensor action as a DB..then the data goes to Oracle Reports schema..if there is any chance of altering the DB..i mean if there is any chance by changing config files so that the data doesnt go to that Reports schema and goes to a custom schema created by any User....i dont know if it can b done...my problem is wen i m configuring the jms Topic for sensor actions..i see blank data coming..for sm reason or the other the data is not getting posted ...i have used a esb ..a routing service based on the schema which i am monitoring...can any1 help?

  • Best practice game design

    Having decided to use java to develop what will be for the most part a 2D (overhead) tile-based RPG game I'm wondering what the best practice is for designing a game like this.
    At the minute I'm intending on using a 'state' system, varying game states (main game, shop, battle, world map etc., state transitions and single use states to create further immersion in the game) implement a State interface that defines how the main game engine interacts with the state (mainly a tick method called on each game loop). When the program wants to switch the game state it makes a call to the main engine, something like setState(State).
    Has anyone written an RPG along these lines? What kind of memory usage did you have, and did you encounter any problems along the way with this design?

    So far, I've identified two perspectives.
    The first is to consider yourself as a sort of "overmind" that tracks each object within the game as a seperate entity. The primary advantage seems to be that it is trivial to add objects as you go along. The primary disadvantage is dealing with all the message passing.
    For example, say you are fighting a horde of goblins when a dragon flies overhead and breathes fire on the battlefield. You would have to call each goblin instance, and have that particular goblin check to see if it dies, and if so, call its death routine. You would have to call your own player object, and have it check to see if you are carrying potions. Each potion then would have to check to see if it boils and explodes in reaction to the dragon breath. And so on. But it would be trivial to add an orc to the battle, or change the dragon into an evil wizard.
    This consumes quite a bit of memory (for all the objects and references), and it can be a little tricky incorporating it into the game loop.
    The other aspect is to code the player as the "center of the universe" and everything happens with respect to that player. It becomes easier find relevant objects and interact with them, but it becomes much harder to just drop an object in place.
    Using the prior example, the dragon becomes your combatant. Damage is done to a goblin in the array and it dies. The next enemy in your array dies. Do so much damage to your inventory and remove them as appropriate. Keeping track of what's going on appears to be much more straightforward, but it would be harder to add the evil wizard to the mix.
    This consumes less memory - the game is only worried about things that happen around you as opposed to well... a whole bunch of "entities" at the same time, and is significantly faster and easier to add to the game loop.
    State machines tend to follow this approach, event driven games tend to follow the prior approach.
    Obviously, I don't code games for a living. I would think that most games these days are somewhat of a mish-mash. Use a basic state machine as in the FF7 example to keep track of broad activities, and within each activity, let the objects manage themselves in response to events.
    For example, each game tick could switch among movement activities, "peaceful magic" activities, battle activities, clean up activities (such as death and xp rewards), and finally general inventory and shop activities, skipping steps as appropriate.
    And say within the movement event, you could have an monster object scream and add itself to the battle mode queue if it detects your player object within such and such a distance.
    Anyway, I've been toying with the idea of coding or volunteernig to work on an RPG as well, so I'd be interested in hearing from you guys on how to organize things too.

  • MPLS Design Best Practices for SP

    When deploying a new MPLS backbone for a Service Provider, what will be consider the best practices in general? For example what about the following list and any other items:
    - Define the Internet as a VRF?
    - Use private ASNs?
    - Define a VRF per special service?
    - Use at least two route reflectors?
    - Use OSPF as IGP?
    - Limit the CE-PE routing support to OSPF and BGP?
    What will be the best approach for management of the devices? A management VRF or the nodes to natively be on a management network?
    What to consider when designing from scratch?

    William some recommended practises, although you can point out your specific constraints in adopting any.
    - Define the Internet as a VRF?
    (Yes Logical speperation is the way to go.)
    - Use private ASNs?
    (No, use a Public AS, you may have to peer outside your AS in a VRF with other AS's)
    - Define a VRF per special service?
    (This is Perfect , Logical Seperation)
    - Use at least two route reflectors?
    (Right, atleast 2 and above that depends on the size of your network)
    - Use OSPF as IGP?
    (I dont see any problems with OSPF in scaling for big networks)
    - Limit the CE-PE routing support to OSPF and BGP?
    (This aspect shouldnt impact much really, you can very well support all the protocols, as its more of serving your customers, rather than dictating the conditions.
    Yes have a seperate VRF for Device Managements (also give a thought for a management subnet, which would be unique across your network)
    You should generally start with a overview topology, introdcution of the objectives. And then go ahead with the suggested phy topo,
    And then move on to the logical services, beggining from Core IGP, then core BGP, and then all the add on protocols, multicast , MPLS TE etc/. Then you can cover specilized service and their logic and description in the end.
    Pretty much, just simply think of building out right from scratch that is Physical Layer and Move to Layer 2 and then Layer 3 Layer 4 .
    So basically you doc should be index in a manner following the sequence of the OSI layers, this gives a good flow to the doc. And rest remains is the description of the logic used in each service or deployment method, that would be your skill.
    HTH-Cheers,
    Swaroop

  • Any best practices on workflow design??

    I feel difficult when migrating applications from DEV->TEST->PROD.
           This is because I have created web services in .net.
           So, for each migration, I am suppose to change all the WSDL links in all forms and Workflows.
           Currently:
              I do open the processes in notepad replace all wsdl connection with TEST/PROD connections.
              I do open each form and goto XML Source and replace all WSDl strings.
            Is there any other best practice(s) to do that?
    Nith

    I followed a different approach for one of my project as:
    created several form variables which hold the WSDL for each different servers.
    Calling the web service from javascript code (instead of data connections).
    This solves the problem but increases the development overhead.
    http://groups.google.com/group/livecycle/web/form%20variables.PNG
    Another way is to design all your web services within adobe itself. In this case we will have the same host name (localhost) forever which doesn't require any modification throughout its lifetime.
    Nith

  • Best practice for GSS design

    Please advice as to what records needs to go in Public DNS server in a scenario where i have url say x.y.com which is listed in the Domain List of the GSS-P, sot that GSS-P or GSS-S can handout the respective external VIP to the clients requesting the url in case one of the GSS/site (GSS_P and GSS-S) goes unavailable
    Please also specify the communication path of a client accessing x.y.com.
    Advice the best practice
    Thanks in advance
    ~EM

    Hi,
    I am new to GSS. I would appreciate if some can help me with the deisgn. I want to know if I need to put the GSS inline after the inernet facing firewall and befor the ACE module. OR use it as one arm mode. Trying to figure out the best fit in the design.
    FWSM1 >>> GSS >>> ACE
    or
    just put the GSS as one arm mode between the FWSM1 >>> ACE
                                                                                                         |
                                                                                                    GSS
    Thanks in advance,
    Nav

Maybe you are looking for

  • IPod Nano 2G with 1Gb of "Other" stuff!?

    Hi, I gave my iPod Nano (2G) to my daughter recently. Before I did, I restored it and put a bunch of her music onto it. It needed to be charged, so I plugged it in last night only to find all the music gone. I charged it and replaced all the music fi

  • Problem in JAXB for processing XML files

    hello I have been working on a project where i need to process data in XML format. the flow goes thus I have 28 data elements that i need to represent as a XML so i compile the schema files and generate the class files for each of the tags and thus i

  • Problem in calling otf to pdf

    Hi, i had added a code in which there is conversion of the otf format to pdf ,i had done the syntax check which says the program is right ,when i execute it ,it work ok but at the when come back from the output screen it is giving the run time error

  • What is the next certification

    After doing my OCP certification what is the next one. I mean what would be my next step. And what are recommanded books for that certification. Taimoor

  • Can Someone Tell me how to NETWORK my Mac to Windows?

    I am using a Macbook Pro and Windows (Windows 7) and assume its similar to networking with Vista (which I also do not know how to do). Is this the best place to post about networking? Can somebody help us? Thanks.