Best practices to include client libraries used at component level

How to include component level resources, while following the best practices.
Ex:
I am looking at the geometrixx media site in CQ 5.6. In the some of components, ex: 2-col-article-summary we have a client library defined under the component.
/apps/geometrixx-media/components/2-col-article-summary
-2-col-article-summary.jsp
-clientlibs
  -css
  -css.txt
if i look at the categories of the clientlib, it is defined as follows
categories String[] apps.geometrixx-media, apps.geometrixx-media.2-col-article-summary
The only place this client library is included is in the head.jsp of main page level component.
    <cq:includeClientLib categories="apps.geometrixx-media.all"/> - this in turn embeds the apps.geometrixx-media
my questions are as follows
1) Why do we have two categories for the clientlib, if ithe second category name( apps.geometrixx-media.2-col-article-summary) is not being used. Is there is some other usage for this i am missing.
2) Also these set of css is always included no matter whether a specific component is added to the page or not.
3)  I could use the following to include the client at the component level, but this will cause unnecessary <script> and <link> elements at the component level mark up.
   <cq:includeClientLib categories="apps.geometrixx-media.2-col-article-summary"/>
Essentially  i am trying to understand, how to include a specific component level resources while following the best practices

Hi,
I dont have CQ5.6 setup so could not see referring example but by looking at description i can say
Ans 1. The client library can be invoked directly through tag lib as <cq:includeClientLib categories="category name"/> but it can also be invoked when you add "dependencies" property to client library folder and in that case it resolve all the dependencies first. So to answer your quesiton by looking at client library folder configuration you can not say that specific category has not been used any where or not invoked.
Ans 2. Invoking client library folder completely depends on your code where you are placing call for client library using <cq:includeClientLib> tag and dependencies configuration. So you have to dig it more to trace out all the calls (also default css/js loads)
Ans 3. Correct. So accomplish that best way to manage client library at component level and give it a unique name which can not be invoked any where neither through <cq:includeClientLib> call nor through dependenies configuration. This way you can avoide overridding of same library files. (Better to manage proper hierarchy of library)
Hope it gives you some idea .
Thanks,
Pawan

Similar Messages

  • Is Adobe Connect part of Adobe Creative Cloud? Are there any best practices ideas from people who use Connect and Creative Cloud?

    Is Adobe Connect part of Adobe Creative Cloud? Are there any best practices ideas from people who use Connect and Creative Cloud?
    I have an Adobe Connect account and I'm are also in the early stages of developing a webinar. I am looking for any tips and advice from anyone who uses both of these services.

    As the £27, was an introductory offer. Upon the completion of one year, the price will change to the normal creative cloud cost which is at £46.88. However if you have the previous versions of the creative suites like CS 3, 4, 5, 5.5 or the CS 6. You can avail the offer at £27.34 per month incl. VAT. However this Requires annual commitment; billed monthly.

  • Best Practices for SAP, connections to use without BW

    Hello,
    Could you help me to solve a problem with Best Practices 4.31?
    SAP Integration Kit XI 3.1 SP3 has 4 options  :
    u2022     SAP Infosets
    u2022     SAP BW MDX Query
    u2022     SAP BW Query
    u2022     Table, cluster ou function
    When I install SAP Integration Kit XI 3.1 SP3 for serveur, I have one additional service in my CCM,  BW Publisher 12. This service is from SAP BW.
    From the article /people/glen.spalding/blog/2010/08/04/the-full-montypart-13bobj-integration-kit-sp3-install-configure and installation guide for IIntegration Kit XI 3.1 SP3 I've learned that I need SAP BW to use Infosets dans Crystal Reports.
    Dans notre installation nous avons
    u2022     SAP ERP système without BW,
    u2022     BOBJ Server Setup,
    u2022     BOBJ Edge Integration Kit for SAP XI 3.1 SP3 server
    u2022     BOBJ Client Setup,
    u2022     BOBJ Edge Integration Kit for SAP XI 3.1 SP3 client
    u2022     Crystal Reports 2008
    u2022     Xcelsius 2008
    I've got Best Practices reports via « Table, cluster ou fonction » connection. There is no problem with this type of connection !
    I suppose that « SAP Infosets », « SAP BW MDX Query » et « SAP BW Query » connections should be used with SAP BW.
    Could you give me your opinion about this question?
    Thanks beforehand,
    Malika

    Hi,
    the InfoSets are a connection option for the classic InfoSets in an ERP system. Would suggest you take a look at the user guide for the SAP Integration Kit
    ingo

  • Best Practices on OWB/ODI when using Asynchronous Distributed HotLog Mode

    Hello OWB/ODI:
    I want to get some advice on best practices when implementing OWB/ODI mappings to handle Oracle Asynchronous Distributed HotLog CDC (change data capture), specifically for “updates”.
    Under Asynchronous Distributed HotLog mode, if a record is changed in a given source table, only the column that has been changed is populated in the CDC table with the old and new value, and all other columns with the exception of the keys are populated with NULL values.
    In order to process this update with an OWB or ODI mapping, I need to compare the old value (UO) against the new value (UN) in the CDC table. If both the old and the new value are NOT the same, then this is the updated column. If both the old and the new value are NULL, then this column was not updated.
    Before I apply a row-update to my destination table, I need to figure out the current value of those columns that have not been changed, and replace the NULL values with its current value. Otherwise, my row-update will replace with nulls those columns that its value has not been changed. This is where I am looking for an advise on best practices. Here are the possible 2 solutions I can come up with, unless you guys have a better suggestion on how to handle “updates”:
    About My Environment: My destination table(s) are part of a dimensional DW database. My only access to the source database is via Asynchronous Distributed HotLog mode. To build the datawarehouse, I will create initial mappings in OWB or ODI that will replicate the source tables into staging tables. Then, I will create another set of mappings to transform and load the data from the staging tables into the dimension tables.
    Solution #1: Use the staging tables as lookup tables when working with “updates”:
    1.     Create an exact copy of the source tables into a staging environment. This is going to be done with the initial mappings.
    2.     Once the initial DW database is built, keep the staging tables.
    3.     Create mappings to maintain the staging tables using as source the CDC tables.
    4.     The staging tables will always be in sync with the source tables.
    5.     In the dimension load mapping, “join” the staging tables, and identify “inserts”, “updates”, and “deletes”.
    6.     For “updates”, use the staging tables as lookup tables to get the current value of the column(s) that have not been changed.
    7.     Apply the updates in the dimension tables.
    Solution #2: Use the dimension tables as lookup tables when working with “updates”:
    1.     Delete the content of the staging tables once the initial datawarehouse database has been built.
    2.     Use the empty staging tables as a place to process the CDC records
    3.     Create mappings to insert CDC records into the staging tables.
    4.     The staging tables will only contain CDC records (i.e. new records, updated records, and deleted records)
    8.     In the dimension load mapping, “outer join” the staging tables, and identify “inserts”, “updates”, and “deletes”.
    5.     For “updates”, use the dimension tables as lookup tables to get the current value of a column(s) that has not been changed.
    6.     Apply the updates in the dimension tables.
    Solution #1 uses staging tables as lookup tables. It requires extra space to store copies of source tables in a staging environment, and the dimension load mappings may take longer to run because the staging tables may contain many records that may never change.
    Solution #2 uses the dimension tables as both the lookup tables as well as the destination tables for the “updates”. Notice that the dimension tables will be updated with the “updates” AFTER they are used as lookup tables.
    Any other approach that you guys may suggest? Do you see any other advantage or disadvantage against any of the above solutions?
    Any comments will be appreciated.
    Thanks.

    hi,
    can you please tell me how to make the JDBC call. I triedit as:
    1. TopicConnectionFactory tc_fact = AQjmsFactory.getTopicConnectionFactory(host, SID, Integer.parseInt(port), "jdbc:oracle:thin");
    and
    2. TopicConnectionFactory tc_fact = AQjmsFactory.getTopicConnectionFactory(host, SID, Integer.parseInt(port), "thin");
    -as given in http://www.acs.ilstu.edu/docs/oracle/server.101/b10785/jm_opers.htm#CIHJHHAD
    The 1st one is giving the error:
    Caused by: oracle.jms.AQjmsException: JMS-135: Driver jdbc:oracle:thin not supported
    at oracle.jms.AQjmsError.throwEx(AQjmsError.java:330)
    at oracle.jms.AQjmsTopicConnectionFactory.<init>(AQjmsTopicConnectionFactory.java:96)
    at oracle.jms.AQjmsFactory.getTopicConnectionFactory(AQjmsFactory.java:240)
    at com.ivy.jms.JMSTopicDequeueHandler.init(JMSTopicDequeueHandler.java:57)
    The 2nd one is erroring out:
    oracle.jms.AQjmsException: JMS-225: Invalid JDBC driver - OCI driver must be used for this operation
    at oracle.jms.AQjmsError.throwEx(AQjmsError.java:288)
    at oracle.jms.AQjmsConsumer.dequeue(AQjmsConsumer.java:1307)
    at oracle.jms.AQjmsConsumer.receiveFromAQ(AQjmsConsumer.java:1028)
    at oracle.jms.AQjmsConsumer.receiveFromAQ(AQjmsConsumer.java:951)
    at oracle.jms.AQjmsConsumer.receiveFromAQ(AQjmsConsumer.java:929)
    at oracle.jms.AQjmsConsumer.receive(AQjmsConsumer.java:781)
    at com.ivy.jms.JMSTopicDequeueHandler.receiveMessages(JMSTopicDequeueHandler.java:115)
    at com.ivy.jms.JMSManager.run(JMSManager.java:90)
    at java.lang.Thread.run(Thread.java:619)
    Is anything else beyond this is required??? please help. :(
    oracle: 10g R4
    linux environment and java is trying to do AQjmsFactory.getTopicConnectionFactory(...); Java machine is diffarent from the database and no oracle client is to be installed on java machine.
    The same code is working fine when i use oc8i instead of thin drivers and run it on db machine.
    ravi

  • Best practice for Video over IP using ISDN WAN

    I am looking for the best practice to ensure that the WAN has suffient active ISDN channels to support the video conference connection.
    Reliance on load threshold either -
    Takes to long for the ISDN calls to establish causing the problems for video setup
    - or is too fast to place additional ISDN calls when only data is using the line
    What I need is for the ISDN calls to be pre-established just prior to the video call. Have done this in the past with the "ppp multilink links minimum commmand but this manual intervention isn't the preferred option in this case
    thanks

    This method is as secure as the password: an attacker can see
    the hashed value, and you must assume that they know what has been
    hashed, with what algorithm. Therefore, the challenge in attacking
    this system is simply to hash lots of passwords until you get one
    that gives the same value. Rainbow tables may make this easier than
    you assume.
    Why not use SSL to send the login request? That encrypts the
    entire conversation, making snooping pointless.
    You should still MD5 the password so you don't have to store
    it unencrypted on the server, but that's a side issue.

  • Best practice for including additional DLLs/data files with plug-in

    Hi,
    Let's say I'm writing a plug-in which calls code in additional DLLs, and I want to ship these DLLs as part of the plug-in.  I'd like to know what is considered "best practice" in terms of whether this is ok  (assuming of course that the un-installer is set up to remove them correctly), and if so, where is the best place to put the DLLs.
    Is it considered ok at all to ship additional DLLs, or should I try and statically link everything?
    If it's ok to ship additional DLLs, should I install them in the same folder as the plug-in DLL (e.g. the .8BF or whatever), in a subfolder of the plug-in folder or somewhere else?
    (I have the same question about shipping additional files too, such as data or resource files.)
    Thanks
                             -Matthew

    Brother wrote:
    I probably should have posed this question here before I delved into writing Java to get data for reports, but better late than never.
    Our ERP is written in COBOL. We have a third party ODBC which allows us to access data using a version of SQL. I have several Java sources compiled in my database that access the data and return something relevant. The Java sources are written in a procedural style rather than taking advantage of object oriented programming with attributes and methods.
    OO is a choice not a mandate. Using Java in a procedural way is certainly not ideal but given that it is existing code I would look more into whether is well written procedural code rather than looking at the lack of OO.
    Now that I am becoming more comfortable with the Java language, I would greatly appreciate any feedback as to best practices for incorporating Java into my database.
    My guess is that it would be helpful to model the ERP "tables" with Java classes that would have attributes, which correspond to the fields, and methods to return the attributes in an appropriate way. Does that sound reasonable? If so, is there a way to automate the task of modeling the tables? If not reasonable, what would you recommend?Normally you create a data model driven by business need. You then implement using whatever means seem expedient in terms of other business constraints to closely model that data model.
    It is often the case that there is a strong correlation between data models and tables but certainly in my experience it is rare when there are not other needs driven by the data model (such as how foreign keys and link tables are implemented and used.)

  • Best Practice For Referencing JPEG Path Using Servlets To Prepare HTML IMG

    I am migrating a legacy app from Tomcat 5 to Weblogic 11g (10.3). In the legacy app, servlets write HTML that uses relative paths for <IMG src="../images/img.jpeg"> and <script src="../javascript/js.js">. The app is deployed as an exploded archive. Unfortunately, none of the images or script are being loaded. I've tried using http://serverIP:portNum/contextName/images/img.jpeg"> but it doesn't work. I've also checked to get the name of the context in the servlet and it's the root context. Could it have something to do with me having to append .war onto the application when I deploy it? Would it help if I deployed it as a war inside an ear? Basically, I want some best practices for doing this on Weblogic 11g. There are a lot of images and javascript and I'm really hoping they don't have to be inserted using ClassLoader.getResourceAsStream()... thank you.

    Best Practice:
    1. Move Static files like images, css, java scripts to a web server infrastructure if available.
    2. If this is not your case, then please send your directory information how you have packaged your EAR. I can advice :)

  • Best practices for sharing iPhoto libraries across multiple Macs

    I have a pair of iMacs running 10.6.x. and iPhoto 11 connected via Ethernet to an Airport Extreme. I am looking for a way to share my iPhoto library between the two Macs.
    I found an Apple Support note on the topic (http://support.apple.com/kb/HT1198) which states that a disk image is a supported location for a shared iPhoto library (with limitations acceptable to me) and it strikes me that this would be a way of hosting a shared iPhoto library onto a NAS device.
    Am I missing a simpler solution or, more importantly, am I missing any blindingly obvious caveats? I'd love to hear from anybody who has tried this (successfully or otherwise) or anybody who has a better idea. I haven't bought a NAS device yet, so I'm open to alternative suggestions.
    Specific requirements:
    Members of my family use either one of a pair of iMacs.
    The only user who edits the iPhoto library is me and I only need read/write access from one machine.
    iPhoto library access limited to one user at a time is acceptable and practical in my case.

    Easiest:
    If the other users only get to see the Library then just use iPhoto Sharing.
    Enable sharing in the iPhoto Preferences on the Host machine, then go to the other macgine and in the same location enable 'Look for Shared Libraries'
    Other forms of sharing will give the users read/write access to the Library.
    A strong warning: If you're trying to edit the Library (that is, make albums, move photos around, keyword, make books or slideshows etc.) or edit individual photos in it via Wireless be very careful. Dropouts are a common fact of wireless networking, and should one occur while the app is writing to the database then your Library will be damaged. Simply, I would not do this with my Libraries. 
    Regards
    TD

  • Best practice for roaming clients

    This pertains to a WORKGROUP setup - NOT ACTIVE DIRECTORY
    We are a small business with 40-50 users and a WSUS installation on Windows Server 2008 R2 (64 bit).  Most of the clients are locally connected to the network, but there are a few that use laptops and are seldom connected to the network.
    Is there a way that we could open a port through the firewall to allow users access to the WSUS server - so that they could connect, update and disconnect 'relatively' quickly?  If so, can anyone offer a suggestion on the best & safest way to accomplish
    this?
    Thanks much

    Ideally, yes, it would be much simpler for laptops &c. to update themselves but, for security reasons, we need to be able to assure that systems connecting to our network are compliant for both Windows updates and for Anti-Virus.
    Thanks for the response.
    Personally I think you have the cart before the horse. You have a 50-user network without Active Directory and you're so worried about AV/Patch compliance that you're trying to publish a WSUS server to the Internet. Without centralized management of network
    access, patch compliance is almost a waste of time.
    Yes, you can publish the WSUS server to Internet-based clients, but there are a myriad of reasons you don't want to do that, and most of them are spelled S-E-C-U-R-I-T-Y.
    To Don's point, your best option for those offsite clients is to enable Automatic Updates and let them be patched with Security Updates automatically and immediately.
    Back on the compliance issue -- if you need to implement compliance to the point of ensuring that mobile systems are patched and have up-to-date AV/AM software, then you need to implement Network Access Policies, which requires Windows Server 2008 R2 (at
    least) and Active Directory.
    Lawrence Garvin, M.S., MCSA, MCITP:EA, MCDBA
    SolarWinds Head Geek
    Microsoft MVP - Software Packaging, Deployment & Servicing (2005-2014)
    My MVP Profile: http://mvp.microsoft.com/en-us/mvp/Lawrence%20R%20Garvin-32101
    http://www.solarwinds.com/gotmicrosoft
    The views expressed on this post are mine and do not necessarily reflect the views of SolarWinds.

  • SX20 and SX10 best practices for small conference rooms using built in Display Speakers

    Hi, 
    Im planning to deploy some small meeting rooms using SX10 and SX20 codecs, I was wondering if someone could give me some key points to consider: 
    1. Display recommendation, nice built in speakers, low delay (goog performance with codec echo canceller) I wish to use a 60 - 70 inch Led TV (not yet defined which brand and model)  I would like to receive some feedback about nice performance displays with nice audio and echo cancellation performance based on your experience. 
    2. What would be the drawbacks of using SX10 and SX20 in larger meeting rooms maybe 10 or 12 people?
    3. When to use external microphone? or when is recommended to use only  internal microphone?
    Thanks in advance for your help. 
    Best Regards,  have a nice day!

    Pretty well any modern display should work well - different people are going to have different ideas about whether the speakers are any good or not, so you'll have to listen to a few yourself and make up your mid what you think is "better".  The echo cancellation can be tweaked manually on the codec to suit the display if required (rather than leaving it on the "auto" setting which, in our installations, has caused more troubles than it's worth).
    The main drawback of using the SX10, or the cheaper of the range of the SX20s, is the camera and the amount of zoom.  In a larger room, you're not going to get a good close up view with a 2.5x camera.  I'd suggest for a bigger room, you look at the 12x cameras.
    As a general rule - use additional microphone(s) when you can't get someone within a 3m radius one of the other microphones.  So in a larger room, you may need many.
    Wayne
    Please remember to rate responses and to mark your question as answered if appropriate.

  • Best practice for moving portal solution using content db from UAT to PROD

    Hi,
     Would like to know can we backup the database from UAT env. and restore the  same to  PROD. if all of my functionality is working fine in UAT env.
    I have event receivers[web level features], site collection level features,custom web parts, custom permissions, saved site templates, custom discussion forums etc.
    Assuming that I have my custom solution deployed on the Prod. which will activate features for those web parts and my custom application page features.
    Is there any issues I can anticipate in PROD.env, if i perform this activity.
    or
    Is this approach not recommended by Microsoft ? if yes , whats the best approach for deploying portal solution in PROD?
    Should I create teh web application, site collections, everything in PROD.from scratch.
    any links regarding this approach and the bext practices / helpful info is appreciated.

    Thanks Trveor for the reply.
    so, I can go ahead and  create the web applns, site collections and  deploy my web parts, item event receivers, appln pages and my timer jobs in UAT and take the  backup of the same and restore it in PROD env.
    But, i ahve a doubt here , as I have few site pages created it in my site template and when i take the backupof this web apppln's content db --- [ i think i can take the backup of web appln content db through power shell] ---- 
    will the site pages also be part of this backup?
    I had some experience in prev.version of SP, wherein i have few site pages and saved site template I have taken the backup of the  web appln and  restore it in another farm and  associate the restored content db to the
    newly created web appln in the targeted farm.
    But when I navigated to thsoe restored site pages, it gave me "resource not found /file not found " error.
     I had  deployed the custom web parts as a custom wsp and added into those site pages.
    and it failed to load those web parts UI.
    I was not sute whether this happened because of backup or restore from source  spfarm to the  targeted sp farm .

  • Best practices for including CSS on publishing

    In order to have completely self-contained project folders, I have css in my project folder.   I don't understand why you can include .js Scripts, and have them publish automatically yet you can't do the same for css scripts.  This is truly baffling since css is at least as fundamental to an html project as js.
    What I want is to be able to include a .css file in my project, so that when the project is published, the css file gets copied into the published folder along with the images and js.  There's really no reason why images and js should be copied but not css.  Also, this would include saving as a template.

    Well the problem is not loading the CSS files.  That can easily be done.  The problem is using the local css file, and having that file COPIED to the publish folder.  This is important because....
    If you edit an image, the image will get copied on publish
    If you edit a locally included .js file, the local verson will get copied on publish
    If you make any edit to the project, the edits will show up in the published files.
    If you make an edit to a local css file, it WILL NOT be copied on publish.
    So if you have several existing example projects, and you open one up, make some changes and publish it to a new location, everything will be there EXCEPT the css.  That is very odd behavior for an html5 creation app.
    If there is possible interferance, that all the more reason to have Edge adding css to projects, so it can notice such interferances.

  • Best practice for sending client ai files

    If a client requests ai files...and they don't have Illustrator - but they want the orig files for printers and vendors, should I save the file using legacy or just save as a normal cs4 ai file?
    Thx.
    Chemol

    in this case it's just a simple 2-colour logo (same file you helped me with earlier) - no effects, nothing to be changed. Nothing mentioned in the contract, I just want them to have the original file so they can provide it to print providers if need be in the future for whatever projects they use it for.
    Just a simple situation, wanting to provide most versatile format for them in the long run. I have CS4, and have been told by one past client (for a signage project) that the printer was unable to open the file - I think they had an older version of AI....
    Just curious - any insight would be helpful.
    Thx.

  • Whats the best practice to simply manage data using php sql?

    Hi
    I'm trying to follow some tutorials but there is too much old material and I'm getting confused.
    This is basic. I have a db and I want to movement data from flex.
    So, this is how I'm trying to do so. As a newbie, I will try to be clear.
    I have created a simple table called "teste" using myphpadmin with only 2 fields for testing.  (id, name).
    I have created the services automaticly using flex to generate a php code.
    This code is a simple exercise with buttons to add, update, and delete a Item.
    Although there is a lot of auto-generated code, I had a lot of work (due to my ignorance) to make it work.
    It fairly works, but I know veteran users would make it better, smarter, shorter, and ALL I WANT is a better (and simple) example to follow
    Please.
    Thanks in advance
    ps. I have set my table field "name" to the 'utf8_unicode_ci' while creating it and I it seems the adobe's generated php can't handle latin chars.
    THE CODE AS FOLLOWS:
    <?xml version="1.0" encoding="utf-8"?>
    <s:Application xmlns:fx="http://ns.adobe.com/mxml/2009"
                      xmlns:s="library://ns.adobe.com/flex/spark"
                      xmlns:mx="library://ns.adobe.com/flex/mx"
                      minWidth="955" minHeight="600"
                      xmlns:valueObjects="valueObjects.*"
                      xmlns:testeservice="services.testeservice.*"
                      creationComplete="application1_creationCompleteHandler()" >
    <fx:Script>
              <![CDATA[
                   import mx.collections.ArrayCollection;
                   import mx.controls.Alert;
                   import mx.events.FlexEvent;
                   import mx.events.ListEvent;               
                   import spark.events.IndexChangeEvent;
                   import spark.events.TextOperationEvent;               
                   //selected item
                   public var selId:int;
                   //------GET ITEM (init)------------------------------------------------------
                   protected function application1_creationCompleteHandler():void
                        tx_edit.addEventListener(KeyboardEvent.KEY_UP,parallelEdit)
                        getAllTesteResult.token = testeService.getAllTeste();                    
                   //------SELECTED ITEM ID------------------------------------------------------          
                   protected function list_changeHandler(event:IndexChangeEvent):void
                        selId = event.currentTarget.selectedItem.id;
                        lb_selectedId.text = "sel id: "+ selId;
                        if(tb_edit.selected) tx_edit.text = event.currentTarget.selectedItem.name;
                   //------ADD ITEM-----------------------------------------------------
                   protected function button_clickHandler(event:MouseEvent):void
                        var teste2:Teste     =     new Teste();
                        teste2.name                = nameTextInput.text;
                        createTesteResult.token = testeService.createTeste(teste2);
                        application1_creationCompleteHandler();
                   //------UPDATE ITEM (in parallel)-----------------------------------------------------
                   private function parallelEdit(e:KeyboardEvent):void
                        if(!isNaN(selId) && selId > 0){
                             if(tb_edit.selected){
                                  //uptadating
                                  teste2.id                     = selId;
                                  teste2.name                = tx_edit.text;
                                  updateTesteResult.token = testeService.updateTeste(teste2);
                   //------UPDATE ITEM (button click)-----------------------------------------------------
                   protected function button3_clickHandler(event:MouseEvent):void
                        teste2.id                     = parseInt(idTextInput2.text);
                        teste2.name                = nameTextInput2.text;
                        updateTesteResult.token = testeService.updateTeste(teste2);
                   //------DELETE ITEM------------------------------------------------------     
                   protected function button2_clickHandler(event:MouseEvent):void
                        if(!isNaN(selId) && selId > 0)     deleteTesteResult.token = testeService.deleteTeste(selId);
              ]]>
         </fx:Script>
         <fx:Declarations>
              <!-- this part was mostly auto generated -->
              <valueObjects:Teste id="teste"/>
              <testeservice:TesteService id="testeService" fault="Alert.show(event.fault.faultString + '\n' + event.fault.faultDetail)" showBusyCursor="true"/>
              <s:CallResponder id="createTesteResult"/>          
              <s:CallResponder id="getAllTesteResult"/>
              <s:CallResponder id="deleteTesteResult"/>
              <s:CallResponder id="updateTesteResult"/>
              <valueObjects:Teste id="teste2"/> 
         </fx:Declarations>
              <!-- this are just visual objects. Renamed only necessary for this example
                    most lines were auto-generated-->
         <!--Create form -->
         <mx:Form defaultButton="{button}">
              <mx:FormItem label="Name">
                   <s:TextInput id="nameTextInput" text="{teste.name}"/>
              </mx:FormItem>
              <s:Button label="CreateTeste" id="button" click="button_clickHandler(event)"/>
         </mx:Form>     
         <!--Create Result -->
         <mx:Form x="10" y="117">
              <mx:FormItem label="CreateTeste">
                   <s:TextInput id="createTesteTextInput" text="{createTesteResult.lastResult as int}" />
              </mx:FormItem>
         </mx:Form>
         <!--List -->
         <s:List x="10" y="179" id="list" labelField="name" change="list_changeHandler(event)">
              <s:AsyncListView  id="anta" list="{getAllTesteResult.lastResult}" />
         </s:List>
         <!--Update in parallel -->
         <s:Label x="147" y="179" id="lb_selectedId"/>
         <s:Button x="147" y="222" label="Remove" id="button2" click="button2_clickHandler(event)"/>
         <s:TextInput x="225" y="257" id="tx_edit"/>
         <s:ToggleButton x="147" y="257" label="Edit" id="tb_edit"  />
         <!--Update with button click -->
         <mx:Form defaultButton="{button3}" x="220" y="0">
              <mx:FormItem label="Id">
                   <s:TextInput id="idTextInput2" text="{teste2.id}"/>
              </mx:FormItem>
              <mx:FormItem label="Name">
                   <s:TextInput id="nameTextInput2" text="{teste2.name}"/>
              </mx:FormItem>
              <s:Button label="UpdateTeste" id="button3" click="button3_clickHandler(event)"/>
         </mx:Form>
    </s:Application>

    Sorry - I had to read up on what a base table block was.
    I think I am. In this particular form I have one block - a customer block which has its items pulled in from the customer table, no problems there, it works fine.
    My problem is I need to make it as user friendly as possible, if the user inputs the wrong customer ID (or does not know their name) I want them to be able to scroll through the customer list using an up or down button.
    When do you have the commit_form statement run, (the OK button) after every change or after a block of changes ? Commit writes the changes to the database giving other users access to that data right ?! How often should it be used ?

  • As per best practice which Group should we use AD or SharePoint Group?

    Hi,
    Please help.
    thanks
    srabon

    Hi Srabon ,
    Here are the benefits and limitations of both ,decide as per your environment and requirements
    SharePoint Groups:
              Benefits – The definition and use of SharePoint Groups is under the control of the SharePoint administrators.  SharePoint is a more dynamic environment
    where Sites and Lists are created for short term use by teams as part of a project.  SharePoint Groups provide the flexibility to be created when needed and then removed just as easily when the project is completed.
              Limitations – The scope of a SharePoint group is limited to a Site Collection.  So, if I want to give
    access to a group of users to more than one SharePoint Site Collection, I need to create that group in each of the Site Collections.  If I do that, I end up having to manage membership across each of those SharePoint Groups separately.
    Active Directory Groups:
    Benefits – As the foundation of the operating system and Exchange, most companies are familiar with Active Directory Groups.  They are used as security groups for access to network devices
    and other infrastructure and applications.  They are also used to manage distribution lists in Exchange. 
    Limitations – Active Directory Group management (adding/removing users) is usually controlled by the Network group within IT.   Due to the impact
    of these groups, IT departments usually have very strong controls and policies around membership requests and changes.  These controls make it difficult for the SharePoint administrators to be responsive to the dynamic nature of SharePoint – meeting the
    needs of a collaborative use model.
    here is a good article you can refer to :
    http://www.metalogix.com/Blog/blog-article/13-12-16/SharePoint_Groups_or_Active_Directory_Groups
    Regards,
    subhash reddy

Maybe you are looking for