"Source used" item best practice

Hello,
How can I configure an item with source used "Always, replacing any..." and source type="Database Column" and STILL not loose the entered value every time the page comes back (items are by default repopulated with DB column values, which make me loose all entered values),
The reason why the page comes back before a submit is because I have some "Select List with Redirect" which dynamically populate other "select list" (dynamic LOV)
How can solve/workaround this issue? Remember, I want all entered values NOT be reset by db column values
Thanks for your help!
Sam

Sam,
I've begun to use AJAX for this type of select list. Solves the caching issues nicely and doesn't cause an entire page refresh as the page isn't submitted until all the data is entered. Carl Backstrom has an example in his samples application - http://apex.oracle.com/pls/otn/f?p=11933:37
Make that Denes Kubicek's - http://deneskubicek.blogspot.com/2008/04/cascading-select-list-in-tabular-form.html
Dave
Message was edited by:
dsteinich
Message was edited by:
dsteinich

Similar Messages

  • Source Used Item Attribute

    Hi All,
    I am using HTML DB 2.0. I am a new user of HTMLDB. Please explain the difference between 2 Source Used Item Attribute
    1. Always, replacing an existing value in session state
    2. Only when current value in session state is null.
    If there is already a thread on this, please give me the link.
    Thanks in advance
    Monika

    So are you saying that this behaviour is different in v1.6?Yes. Otherwise the PPR stuff wouldn't work when those requests referenced uncommitted session state.
    You mean, "will NOT be able to ..."No, it will be able to see the value of the item (in uncommitted session state). That's the change from 1.6.
    What is an example of a "session-state committing source method"? Code that not only sets the source of the item for display but as a side-effect cause update_cache_with_write to occur for that item or for any other item(s). For example a PL/SQL Function source type that returns the source value and also does :P1_ITEM := 'value'; .
    Scott
    Message was edited by:
    sspadafo

  • Templates & Library items - best practice

    I've been having a bit of a heated discussion about how
    templates and library items should be used on medium to large
    websites.
    I'm of the opinion that you should have 1 template (or as few
    as possible) and any differences be managed with Library items.
    My colleague's argument is you should have a different
    template for each section, so for example 1 for the news, 1 for the
    features section (for magazine sites) so that if you have a banner
    section wide you change that template, as opposed to the way I'm
    used to which would be 1 template with multiple library items in
    the editable regions of the template, my reasoning for this is the
    large amount of html used for each template (like rollovers) can
    become inconsistent when there's more than one template. Like if
    you change something site wide you have to remember to change all
    of the templates, if you do this with a simple copy past DW can
    still make corrupt code if you use multiple templates.
    So basically I was wondering if there was a standard or best
    practice way of doing this.
    Thanks

    Your colleague is mistaken IN MY OPINION. I think you are on
    the right
    track *unless* your section pages are *SO* different in
    layout that you
    cannot conveniently include all of the
    alterations/permutations in a single
    template page. Further, my current workflow is to rely
    heavily on
    server-side includes within template pages, as described here
    Templates do make your life much easier. I use the following
    scheme....
    First, I mentally separate the page layout into three
    sections:
    1. Stuff that will not change for the life of the site (i.e.,
    the basic
    structural elements)
    2. Stuff that *could* change from time to time (e.g.,
    navigation elements,
    burst advertisements, section-specific navigation, etc.)
    3. Stuff that *will* change from one page to the next
    Then I create a template containing all class1 elements. Next
    I create
    server-side include files containing all class 2 elements and
    place them on
    the template as needed. Note - some of the class 2 elements
    may be
    "section-specific elements", and their placement on the
    template will be
    subject to the next item. Finally, I insert editable regions
    to cover the
    class 3 items, INCLUDING the section-specific navigation.
    This allows me to just cookie-cut the rest of the site. I
    estimate that
    even for fairly large sites, about 80% of my work goes into
    planning and
    creating this template file.
    For what it's worth, I never use Library items....
    Murray --- ICQ 71997575
    Adobe Community Expert
    (If you *MUST* email me, don't LAUGH when you do so!)
    ==================
    http://www.dreamweavermx-templates.com
    - Template Triage!
    http://www.projectseven.com/go
    - DW FAQs, Tutorials & Resources
    http://www.dwfaq.com - DW FAQs,
    Tutorials & Resources
    http://www.macromedia.com/support/search/
    - Macromedia (MM) Technotes
    ==================
    "tefnut" <[email protected]> wrote in
    message
    news:[email protected]...
    > I've been having a bit of a heated discussion about how
    templates and
    > library
    > items should be used on medium to large websites.
    > I'm of the opinion that you should have 1 template (or
    as few as possible)
    > and
    > any differences be managed with Library items.
    > My colleague's argument is you should have a different
    template for each
    > section, so for example 1 for the news, 1 for the
    features section (for
    > magazine sites) so that if you have a banner section
    wide you change that
    > template, as opposed to the way I'm used to which would
    be 1 template with
    > multiple library items in the editable regions of the
    template, my
    > reasoning
    > for this is the large amount of html used for each
    template (like
    > rollovers)
    > can become inconsistent when there's more than one
    template. Like if you
    > change
    > something site wide you have to remember to change all
    of the templates,
    > if you
    > do this with a simple copy past DW can still make
    corrupt code if you use
    > multiple templates.
    >
    > So basically I was wondering if there was a standard or
    best practice way
    > of
    > doing this.
    >
    > Thanks
    >

  • SLES OES Patching using ZCM - Best Practice?

    Have recently updated an underused ZLM 7.3 to ZCM 11 and have started to utilise it to perform patching of SLES 10 SP3 and OES 2.
    At present I have a Novell subscription that downloads from novell NU as Patches with
    Create Category based Bundle Groups on.
    I hadn't at this point got my head around the sandbox but now want to utilise this as I don't want to implement it willy nilly and have patches deployed and installed without testing first.
    So first question is whether there is a best practice document for using ZCM 11 for patching SLES 10 and Sles/oes boxes that describes how best to use the sandbox.
    Next question is whether sandbox applies to the patches or only monolithic bundles
    At the end of the day it would be great to have a best practice document on how to utilise ZCM with change control over keeping SLES10/OES boxes patched up to date

    Hello,
    updates via ZCM is also a problem for me.
    At least i could give a little response from testing.
    As stated in the oes2 documentation patches had to be used for updates. At that point i had the problem that the subscription only created a bundle group which included the patches.
    The monolithic bundle seems to be an ordinary update (zcm docu states rpm -U ...) and so should not be used for oes2 servers.
    I found out that its ok to rename the patch bundle group created by the subscription to fix the patchlist. I use bundle names which include the date to keep updates from sles and oes in sync.
    So you should be able to do updates with these bundlegroups without problem but it is not very nice because each patch bundle had its on log report and entry in the bundle list.
    So to keep a global view is not very easy.
    The other problem is that there is no official documentation neither on the oes2 or the zcm site how to correct update/patch a sles or oes system with zcm (without patchlink).
    Hope this will be included soon as all linux satellite server will have the problem that this is the only way to get updates without direkt connection to the internet - at least as far as i can see.

  • Using cache - best practice

    We created a banner using toplink. The buttons (images) are all stored in the database. The banner is in 300+ pages and they cange depending on some business rules. I want the best performance available. Yet, I want to make sure that once a record is updated so is the banner. What is the best cache option to use? Also, where should I set the cache? If I click Toplink Mapping (I'm using Jdeveloper 10.12), should I double-click on the specific mapping? I only see two options:
    - default (checked)
    - always refresh
    -only refresh if newer version
    Is there some type of "best practices" of using Toplink's cache?
    Thanks,
    Marcelo

    Hello Marcelo,
    Can't be sure exactly, but are you modifying the database outside of TopLink? This would explain why the cached data is stale. If so, what is needed is a strategy to handle refreshing the cache once changes are known be made outside TopLink, or revise the caching strategy being used. This could be as easy as calling session.refreshObject(staleObject), or configuring specific class descriptors to always refresh when queried.
    Since this topic is rather large and application dependent, I'd recommend looking over the 10.1.3 docs:
    http://download-west.oracle.com/docs/cd/B25221_01/web.1013/b13593/cachun.htm#CHEJAEBH
    There are also a few other threads that have good discussions on how to avoid stale data, such as:
    Re: Only Refresh If Newer Version and ReadObjectQuery
    Best Regards,
    Chris

  • How to use KNOA - best practices

    I am looking for best practices in implementing KNOA. How is the tool being used? How do we make sure it yelds value? Your input is appreciated!

    Hi Shishir,
    https://websmp104.sap-ag.de/instguides
    Also:
    help.sap.com\Sap Best Practices\Getting Started R/3
    Patricio.

  • REACh Implementation using SAP Best Practices for Chemicals

    Hello,
    We are working on implementation of SAP REACh Compliance in to Best Practices for Chemicals. Currently we are extending the Best Practices scenario for "SVT" to  cover REACh Compliance.
    Do we have a preconfigured scenario available with list of building blocks to be activated for REACh Compliance to work?  Are there any plans to release the same in future?
    Thanks
    Jayakumar

    Chemicals best Practice support on REACH  is currently limited to SVT scenario you mentioned in your question. Having Best practice support for SAP REACH Compliance is a good idea of course but BP community needs to decide to do so and needs to take action. This has not be done so far and would currently be difficult to realize since there is simply no REACH best practice yet... All companies are more or less still in learning mode about how to find the best way to implement REACH legislation best not to forget  the that REACH as such (its guidance documents) are still changing too.
    Kind rgds
    MArko

  • List Item Best Practice

    I have created an unordered list.  Instead of one of the standard CSS pics available, I just used a simple hyphen.  Given that it's text and it's part of the text, will that effect any search results in google or any other search engine?  Should I make the hyphen a picture and use that instead of the simple text?
    Thanks for the input!

    I agree with Joe, and think that a background image might be best for you:
    ul {
        list-style: none;
    li {
        background: url(dash.gif);
        padding-left: 10px;
    E. Michael Brandt
    www.divahtml.com
    www.divahtml.com/products/scripts_dreamweaver_extensions.php
    Standards-compliant scripts and Dreamweaver Extensions
    www.valleywebdesigns.com/vwd_Vdw.asp
    JustSo PictureWindow
    JustSo PhotoAlbum, et alia

  • JSP Best Practices and Oracle Report

    Hello,
    I am writing an application that obtains information from the user using a JSP/HTML form and then submitted to a database, the JSP page is setup using JSP Best Practices in which the SQL statments, database connectivity information, and most of the Java source code in a java bean/java class. I want to use Oracle Reports to call this bean, and generate a JSP page displaying the information the user requested from the database. Would you please offer me guidance for setting this up.
    Thank you,
    Michelle

    JSP Best Practices.
    More JSP Best Practices
    But the most important Best Practice has already been given in this thread: use JSP pages for presentation only.

  • Best practice for moving from a G5 to a new Mac with SL

    I am receiving my new iMac today (27") and am very excited
    However I want to move over using the best practices to assure that I remain excited and not frustrated
    My initial thoughts are to boot it up and doe the initial set up - to move my iPhoto library over and to use migration assistance to move the rest of my data files
    Then to install all of the extra software that I can find the packages for from the original installation disks
    And then finally to use migration assistant again to move over any software that I can not find original disks for (I've moved from Mac to Mac to Mac over and over and some of the software goes back to OS 9 (and won't run anymore I guess)
    Is this a good way
    OR
    will I mess up doing it this way
    OR
    am I spending far too much time worrying about moving old problems over and would be better off to just turn MA loose and let it do its thing form the beginning?
    BTW - mail crashes a lot on my existing system - pretty much everything else seems ok - except iPhoto is slow - hoping that the new Intel dual core will help that
    LN

    Migration Assistant is not a general file moving tool. MA will migrate your Applications and Home folders transferring only your third-party applications. MA will transfer any application support folders required by your applications, your preferences, and network setup. You do not have a choice of what will be migrated other than the above. MA cannot determine whether anything transferred is compatible with Snow Leopard. I recommend you look at the following:
    A Basic Guide for Migrating to Intel-Macs
    If you are migrating a PowerPC system (G3, G4, or G5) to an Intel-Mac be careful what you migrate. Keep in mind that some items that may get transferred will not work on Intel machines and may end up causing your computer's operating system to malfunction.
    Rosetta supports "software that runs on the PowerPC G3, G4, or G5 processor that are built for Mac OS X". This excludes the items that are not universal binaries or simply will not work in Rosetta:
    Classic Environment, and subsequently any Mac OS 9 or earlier applications
    Screensavers written for the PowerPC
    System Preference add-ons
    All Unsanity Haxies
    Browser and other plug-ins
    Contextual Menu Items
    Applications which specifically require the PowerPC G5
    Kernel extensions
    Java applications with JNI (PowerPC) libraries
    See also What Can Be Translated by Rosetta.
    In addition to the above you could also have problems with migrated cache files and/or cache files containing code that is incompatible.
    If you migrate a user folder that contains any of these items, you may find that your Intel-Mac is malfunctioning. It would be wise to take care when migrating your systems from a PowerPC platform to an Intel-Mac platform to assure that you do not migrate these incompatible items.
    If you have problems with applications not working, then completely uninstall said application and reinstall it from scratch. Take great care with Java applications and Java-based Peer-to-Peer applications. Many Java apps will not work on Intel-Macs as they are currently compiled. As of this time Limewire, Cabos, and Acquisition are available as universal binaries. Do not install browser plug-ins such as Flash or Shockwave from downloaded installers unless they are universal binaries. The version of OS X installed on your Intel-Mac comes with special compatible versions of Flash and Shockwave plug-ins for use with your browser.
    The same problem will exist for any hardware drivers such as mouse software unless the drivers have been compiled as universal binaries. For third-party mice the current choices are USB Overdrive or SteerMouse. Contact the developer or manufacturer of your third-party mouse software to find out when a universal binary version will be available.
    Also be careful with some backup utilities and third-party disk repair utilities. Disk Warrior 4.1, TechTool Pro 4.6.1, SuperDuper 2.5, and Drive Genius 2.0.2 work properly on Intel-Macs with Leopard. The same caution may apply to the many "maintenance" utilities that have not yet been converted to universal binaries. Leopard Cache Cleaner, Onyx, TinkerTool System, and Cocktail are now compatible with Leopard.
    Before migrating or installing software on your Intel-Mac check MacFixit's Rosetta Compatibility Index.
    Additional links that will be helpful to new Intel-Mac users:
    Intel In Macs
    Apple Guide to Universal Applications
    MacInTouch List of Compatible Universal Binaries
    MacInTouch List of Rosetta Compatible Applications
    MacUpdate List of Intel-Compatible Software
    Transferring data with Setup Assistant - Migration Assistant FAQ
    Because Migration Assistant isn't the ideal way to migrate from PowerPC to Intel Macs, using Target Disk Mode, copying the critical contents to CD and DVD, an external hard drive, or networking
    will work better when moving from PowerPC to Intel Macs. The initial section below discusses Target Disk Mode. It is then followed by a section which discusses networking with Macs that lack Firewire.
    If both computers support the use of Firewire then you can use the following instructions:
    1. Repair the hard drive and permissions using Disk Utility.
    2. Backup your data. This is vitally important in case you make a mistake or there's some other problem.
    3. Connect a Firewire cable between your old Mac and your new Intel Mac.
    4. Startup your old Mac in Target Disk Mode.
    5. Startup your new Mac for the first time, go through the setup and registration screens, but do NOT migrate data over. Get to your desktop on the new Mac without migrating any new data over.
    If you are not able to use a Firewire connection (for example you have a Late 2008 MacBook that only supports USB:)
    1. Set up a local home network: Creating a small Ethernet Network.
    2. If you have a MacBook Air or Late 2008 MacBook see the following:
    MacBook (13-inch, Aluminum, Late 2008) and MacBook Pro (15-inch, Late 2008)- Migration Tips and Tricks;
    MacBook (13-inch, Aluminum, Late 2008) and MacBook Pro (15-inch, Late 2008)- What to do if migration is unsuccessful;
    MacBook Air- Migration Tips and Tricks;
    MacBook Air- Remote Disc, Migration, or Remote Install Mac OS X and wireless 802.11n networks.
    Copy the following items from your old Mac to the new Mac:
    In your /Home/ folder: Documents, Movies, Music, Pictures, and Sites folders.
    In your /Home/Library/ folder:
    /Home/Library/Application Support/AddressBook (copy the whole folder)
    /Home/Library/Application Support/iCal (copy the whole folder)
    Also in /Home/Library/Application Support (copy whatever else you need including folders for any third-party applications)
    /Home/Library/Keychains (copy the whole folder)
    /Home/Library/Mail (copy the whole folder)
    /Home/Library/Preferences/ (copy the whole folder)
    /Home /Library/Calendars (copy the whole folder)
    /Home /Library/iTunes (copy the whole folder)
    /Home /Library/Safari (copy the whole folder)
    If you want cookies:
    /Home/Library/Cookies/Cookies.plist
    /Home/Library/Application Support/WebFoundation/HTTPCookies.plist
    For Entourage users:
    Entourage is in /Home/Documents/Microsoft User Data
    Also in /Home/Library/Preferences/Microsoft
    Credit goes to Macjack for this information.
    If you need to transfer data for other applications please ask the vendor or ask in the Discussions where specific applications store their data.
    5. Once you have transferred what you need restart the new Mac and test to make sure the contents are there for each of the applications.
    Written by Kappy with additional contributions from a brody.
    Revised 1/6/2009
    In general you are better off reinstalling any third-party software that is PPC-only. Otherwise update your software so it's compatible with Snow Leopard.
    Do not transfer any OS 9 software because it's unsupported. You can transfer documents you want to keep.
    Buy an external hard drive to use for backup.

  • Best Practices Pharma-MM Docs & process Over view

    Hi All,
            Could anybody provide the Pharma Industry processes-Typical MM Point of view.
    Also if any additional useful docs,which can be useful for implemenating Pharma best Practices package in a Industry.
    Thanks
    JP

    Hi Jyoti Prakash,
    The following link will guide and takes you though in detail the SAP Best Practices for Pharmaceuticals. 
    http://help.sap.com/bp_pharmav1600/Pharma_US/index.htm
    This will be the best available source for pharma best practices in line with SAP.
    Hope this will be very useful to you.  Please confirm
    Regards
    R. Senthil Mareeswaran.

  • Best practices about JTables.

    Hi,
    I'm programming in Java since 5 months ago. Now I'm developing an application that uses tables to present information from a database. This is my first time handling tables in Java. I've read Sun's Swing tutorial about JTable, and several information on other websites, but they limit to table's syntax and not in best practices.
    So I decided what I think is a proper way to handle data from a table, but I'm not sure that is the best way.Let me tell you the general steps I'm going through:
    1) I query employee data from Java DB (using EclipseLink JPA), and load it in an ArrayList.
    2) I use this list to create the JTable, prior transformation to an Object[][] and feeding this into a custom TableModel.
    3) From now on, if I need to search an object on the table, I search it on the list and then with the resulting index, I get it from the table. This is possible because I keep the same row order on the table and on the list.
    4) If I need to insert an item on the table, I do it also on the list, and so forth if I'd need to remove or modify an element.
    Is the technique I'm using a best practice? I'm not sure that having to keep synchronized the table with the list is the better way to handle this, but I don't know how I'd deal just with the table, for instance to efficiently search an item or to sort the table, without doing that first on a list.
    Are there any best practices in dealing with tables?
    Thank you!
    Francisco.

    Hi Joachim,
    What I'm doing now is extending DefaultTableModel instead of implementing AbstractTableModel. This is to save implementing methods I don't need and because I inherit methods like addRow from DefaultTableModel. Let me paste the private class:
    protected class MyTableModel extends DefaultTableModel {
            private Object[][] datos;
            public MyTableModel(Object[][] datos, Object[] nombreColumnas) {
                super(datos, nombreColumnas);
                this.datos = datos;
            @Override
            public boolean isCellEditable(int fila, int columna) {
                return false;
            @Override
            public Class getColumnClass(int col) {
                return getValueAt(0, col).getClass();
        }What you are suggesting me, if I well understood, is to register MyTableModel as a ListSelectionListener, so changes on the List will be observed by the table? In that case, if I add, change or remove an element from the list, I could add, change or remove that element from the table.
    Another question: is it possible to only use the list to create the table, but then managing everything just with the table, without using a list?
    Thanks.
    Francisco.

  • EFashion sample Universes and best practices?

    Hi experts,
    Do you all think that the eFashion sample Universe was developed based on the best practices of Universe design? Below is one of my questions/problems:
    Universe is designed to hide technical details and answer all valid business questions (queries/reports). For non-sense questions, it will show 'incompatible' etc. In the eFashion sample, I tried to compose a query to answer "for a period of time, e.g. from 2008.5 to 2008.9, in each week for each product (article), it's MSRP (sales price) and sold_price and margin and quantity_sold and promotion_flag". I grabed the Product.SKUnumber, week from Time period, Unit Price MSRP from Product, Sold at (unit price) from Product, Promotions.promotion, Margin and Quantity sold from Measures into the Query Panel. It gives me 'incompatible' error message when I try to run it. I think the whole sample (from database data model to Universe schema structure/joins...) is flawed. In the Product_promotion_facts table, it seems that if a promotion lasts for more than one week, the weekid will be the starting week and duration will indicate how long it lasts. In this design, to answer "what promotions run in what weeks" will not be easy because you need to join the Product_promotion_facts with Time dimention using "time.weekid between p_prom.weekid and p_prom.weekid+duration" assuming weekid is in sequence, instead of simple "time.weekid=p_prom.weekid".  The weekid joins between Shop_fact and product_promotion_facts and Calendar_year_lookup are very confusing because one is about "the week the sales happened" and the other "the week the promotion started". No tools can smart enough to resolve this ambitious automatically. Then the shortcut join between shop_facts and product_promotion_facts. it's based on the articleid alone. obviously the two have to be joined on both article and time (using between/and, not the simple weekid=weekid in this design), otherwise the join doesn't make sense (a sale of one article on one day joins to all the promotions to this article of all time?).
    What do you think?
    thanks.
    Edward

    You seem to have the idea that finding out whether a project uses "best practices" is the same as finding out whether a car is blue. Or perhaps you think there is a standards board somewhere which reviews projects for the use of "best practices".
    Well, it isn't like that. The most cynical viewpoint is that "best practices" is simply an advertising slogan used by IT consultants to make them appear competent to their prospective clients. But basically it's a value judgement. For example using Hibernate may be a good thing to do in many projects, but there are projects where it would not be a good thing to do. So you can't just say that using Hibernate is a "best practice".
    However it's always a good idea to keep your source code in a repository (CVS, Subversion, git, etc.) so I think most people would call that a "best practice". And you could talk about software development techniques, but "best practice" for a team of three is very different from "best practice" for a team of 250.
    So you aren't going to get a one-paragraph description of what features you should stick in your project to qualify as "best practices". And you aren't going to get a checklist off the web whereby you can rate yourself for "best practices" either. Or if you do, you'll find that the "best practice" involves buying something from the people who provided the checklist.

  • Adobe Air SQLite Concurrency/Best Practices

    Hello Everyone,
      I posted this on Stackoverflow, but I was wondering if there were any I guess design patterns or hints concerning the usage of a SQLite DB.  I'm asking this mostly because with the Adobe Air sqlite implementation, you can't modify the timeout, so you have to re-architect your application to try and retry your sql statements.
    I am attempting to use a SQLite DB with an AIR project and there seem to be a few questions that I just can't seem to find the answers to on the internet, especially ones concerning best practices.
    The first is about having multiple connections in the same application instance.  Some application processes require the use of a db for a split second, and would better be served by a synchronous type code flow (try{}catch(){}).  Some processes may take awhile and would be better served by an asynchronous code flow with an AsyncResponder.  Because of this, I have thought about having a ConnectionPoolManager which would have multiple instances of async and one instance of my sync SQLConnection classes.  Is this a decent, good or absolutely atrocious idea ?  This leads me to my next point.
    Is there any issue of having multiple async connections running their async statements in parallel ?  I've seen some people complain about the db being locked, and I've read that only one statement can perform a write at a time.   What happens if a write statement gets called to execute(new Async...) while another one is busy ?  Will an error be thrown, or will it wait for a timeout ?
    Any clarification about this issue would be greatly appreciated.  I keep running into sources that kind of answer one-off questions but not my question in particular.  Usually I would go test it myself, but I'm worried I may not be able to reproduce the pitfall error that will only manifest itself at heavier usage levels.

    I'd suggest checking out the following links for tips and
    source code for best practices on SQLite:
    http://www.adobe.com/devnet/air/flex/articles/air_sql_operations.html
    http://www.peterelst.com/blog/2008/04/07/introduction-to-sqlite-in-adobe-air/
    - Rob

  • Adobe "Best Practices" aren't practical to implement!

    Maybe I'm just venting, but seriously -- what the heck guys
    (Adobe)?
    I've been pretty forgiving in general about this kind of
    stuff but the list of annoying issues I've been finding is getting
    a little long. It seems like every time I make a serious effort to
    implement Adobe's "best practices" form validation functionality,
    it takes a massive amount of time unless you're doing it EXACTLY
    like the example presented here:
    http://www.adobe.com/devnet/flex/quickstart/validating_data/#section3
    And even here, there's *alot* of repetitive (and
    bug-vulnerable) code going on here if you're doing multiple forms,
    or forms of any modest level complexity. God help you if you're
    trying to use these "best practices" using an editable DataGrid
    with validation assistance in a popup window...
    Is this an outdated article or something? Are there any
    useful examples of how to do this more effectively (maybe with some
    code reuse?), that include less commonly (but still important)
    features required for accessibilty (disabling buttons seems to
    screw with tab navigation) or various types of custom
    functionality?
    I know this is like complaining that the air conditioning in
    my private jet is hard to control, but if Adobe's gonna recommend
    doing this, it shouldn't take hours to do, right?

    This may be a good starting to to work on a better solution.
    Seems to take care of the problem of repetitive code at least.
    http://blog.wheelerstreet.com/?p=123
    Hopefully someone from Adobe will address these issues, and
    improve on their documentation and examples. I find myself going
    through the Flex SDK source code sometimes trying to figure out why
    something doesn't work (DataGrid stuff in particular) or how to get
    a certain behavior.

Maybe you are looking for