Best Practices - Relational to XML?

Oracle 10gR2 Enterprise Edition 10.2.0.4.0
Read the OpenWorld 2009 Oracle XML DB Design Guidelines, paraphrased as:
- Use Case:
-- Data already in relational form
-- Need to generate different XML shape for presentation, exchange, reporting
- Recommendations:
-- Do Use XMLElement(), XMLForest(), XMLAgg() SQL/XML to define views
-- Do Use XQuery with ora:view() for complex XML report generation
-- Don't use use DBMS_XMLGEN(), DBMS_XMLQUERY(), XSU
--- Lower perf, less declarative
Although, Mark Drake's post in here Performance problems with XMLTABLE and XMLQUERY involving relational data has me wondering if the "Do Use XQuery with ora:view() for complex XML report generation" point has some caveats?
I'm starting to build a solution to send relational data as XML to another system, where the relational data consists of several joined tables, including many one-to-one joins and a few one-to-many joins (including a pair of one-to-many joins where the rows from two joined tables must be interleaved). ( Note - To date, all of our XML processing has been for presentation and has been done on Windows .NET client apps. )
Have built some simple SQL/XML views with XMLElement(), XMLAttributes(), XMLForest(), ... to begin building my SQL/XML and XML DB experience base, and am now working on an XQuery-based solution for the real solution.
Questions:
1. For performance with multiple 1:1 joined tables, should SQL VIEWs be preferred to building multiple FLOWR joins in XQuery?
2. Can a relational tables with CLOB columns be referenced in the XQuery for clause of an SELECT XMLQuery() query?
Note: Q2 is due to performance issues I encountered - when I added an XQuery for clause referencing a relational table containing several CLOB columns, the SELECT XMLQuery() query never returned.
Alex

stupid remark removed

Similar Messages

  • Looking for customization best practices related to Timezone feature in EBS

    Hi all,
    We are updating our 11.5.10.2 env to deploy the TImezone feature and i'm anticipating some issues for the customized code as i'm not sure the developers followed best practices about handling dates.
    Is there any documents where i can find the best practices to follow for customized coding in order to be compliant with the EBS Timezone feature ?
    Like for instance, Date APIs usage or the work to do to convert Date data for interfaces ?
    Thanks in advance
    Chris

    Is there any documents where i can find the best practices to follow for customized coding in order to be compliant with the EBS Timezone feature ?
    Like for instance, Date APIs usage or the work to do to convert Date data for interfaces ?I do not think such a doc exists.
    You may review these docs and see if it helps.
    User Preferred Time Zone Support Guidelines for Applications 11i10CU2 [ID 330075.1]
    User-Preferred Time Zone Support in Oracle E-Business Suite Release 12 [ID 402650.1]
    Globalization Guide for Oracle Applications Release 12 [ID 393861.1]
    Also, please log a SR to confirm the existence of the doc you are looking for.
    Thanks,
    Hussein

  • Best Practice: SAPGUI Version and Patch Upgrades

    Hello -
    Does anyone have some thoughts/information on best practices relating to SAPGUI version and patch upgrades.
    Obviously, sometimes upgrades are forced upon us (e.g. 7.10 for Vista) and in other cases it may just be considered "nice to have".
    Either way - it always signifies regression test and deployment effort.  How do we balance the benefit and cost?
    Thanks, Steve

    Hi Steve,
    you're right for the first part, yes, we (usually) patch twice a year.
    Now for the rest
    An uninstall will only happen on release changes (6.20->6.40->7.10), i. e. about every 4-5 years as SAP releases them.
    Patches are applied to the installation server and the setup on the client will only update changed program parts. For example, upgrading 6.40->7.10 took about 10 minutes (incl. uninstall), applying patch 1 less than 5 minutes.
    I recommend, you read the "SAP Frontend Installation Guide - 7.10" which you find at SMP alias sapgui. Navigate to  Media Library - Literature. It explains setup of the installation server (sounds like a big thing, but ain't much more than creating a share), creating packages, applying updates etc.
    Peter
    Points always appreciated

  • Switching Best Practice - Spanning Tree andEtherchannel

    Dear All,
    Regarding best practice related to Spanning Tree and Etherchannel, we have decided to configure following.
    1. Manually configure STP Root Bridge.
    2. On end ports, enable portfast and bpduguard.
    3. On ports connecting to other switches enable root guard.
    In etherchannel config, we have kept mode on on both side, need to change to Active and desirable as I have read that mode on may create loops? Please let me know if this is OK and suggest if something missing.
    Thank You,
    Abhisar.

    Hi Abhisar,
    Regarding your individual decisions: Manually configuring the Root Bridge is a natural thing to do. You should never leave your network just pick up a root switch based on default switch settings.
    On end ports, using PortFast and BPDU Guard is a must especially if you are running Rapid PVST+ or MSTP.
    Regarding the Root Guard on ports to other switches - this is something I do not recommend. The Root Guard is a protective mechanism in situations when your network and the network of your customer need to form a single STP domain, yet you want to have the STP Root Bridge in your network part and you do not want your customer to take over this root switch selection. In these cases, you would put the Root Guard on ports toward the customer. However, inside your own network, using Root Guard is a questionable practice. Your network can be considered trustworthy and there is no rogue root switch to protect against. Using Root Guard in your own network could cause your network to be unable to converge on a new workable spanning tree if any of the primary links failed, and it would also prevent your network from converging to a secondary root switch if the primary root switch failed entirely. Therefore, I personally see no reason to use Root Guard inside your own network - on the contrary, I am concerned that it would basically remove the possibility of your network to actually utilize the redundant links and switches.
    Regarding EtherChannels - yes, you are right, using the on mode can, under circumstances, lead to permanent switching loops. EtherChannel is one of few technologies in which I wholeheartedly recommend on relying on a signalling protocol to set it up, as opposed to configuring it manually. The active mode is my preferred mode, as it utilizes the open LACP to signal the creation of an EtherChannel, and setting both ends of a link to active helps to bring up the EtherChannel somewhat faster.
    If you are using fiber links between switches, I recommend running UDLD on them to be protected against issues caused by uni-directional links. UDLD is not helpful on copper ports and is not recommended to be run on them. However, I strongly recommend running Loop Guard configured globally with the spanning-tree loopguard default. Loop Guard can, and should, be run regardless of UDLD, and they can be used both as they nicely complement each other.
    My $0.02...
    Best regards,
    Peter

  • Best practices for additonal storage devices in a clustered container

    I am wondering what is considered as best practices related to adding more storage devices. I have a Solaris Cluster 3.2, a failover resource group with a SZBT resource and a HASP resource for the zone root path. Now i have to add more storage devices (for Oracle to the zone). The additional storage devices are in the same metaset as the zone root. I see the following options:
    -> Add it to the zone (simple not managed by HASP)
    -> Add it to the existing HASP for the RG and lofs mount it in the Zone (managed by HASP, but needs lofs mount)
    Are there other options, and what is considered best practices?
    Fritz

    Hi Fritz,
    no matter which option you take it should resoult in lofs mounts.
    I in person would not mix the concepts, you started with HASP, so I would continue with HASP.
    To achieve this you have to take three steps:
    1 add the file systems to the HASP resource,
    2. create the mountpoints in the local zone
    3 add the file system to the Mounts variable in the SCZBT's parameter file.
    If you do not want to reboot the zones, you can mount the loopback files manually, but the safest option ( you will not mess up with typos in th mount pints and realize it later) would be to disable and reenable th sczbt resource.
    Cheers
    Detlef

  • SLD Best Practices

    Hi
    From where can I get the best practices related to SLD?
    My XI system Landscape contains following systems.
    Sandbox
    Development
    Quality
    Production
    I am considering following options:
    Individual SLD for each system.
    One SLD for Sandbox, Second for Development and Quality and Third for Production.
    One for Sandbox, Development and Quality and Second for Production.
    I have some advantages and disadvantages in my mind regarding each approach however I would request forum to please share their experiences.
    Further to this there is Solution Manager also in the system which has its own SLD. Should that be utilized or should that be kept aloof from the XI system.

    Hi,
    My opinion, one SLD for each environment is good.
    Just go thru following links -
    SLD
    Central SLD vs SLD for each XI instance
    SLD best practices
    SAP NW 2004(s) SOLMAN/SLD/SLM best practice
    SLD best practices
    Check how to guides from service.sap.com.
    Hope this helps..
    Regards,
    Moorthy

  • SEM BPS/IP Best Practice

    Hi Guru's,
    Regards, Plz any body can give some links to Best Practice of SEM BPS/IP would be great full.
    many thanks
    Mohan

    Hi,
    There is not much best practice related to BPS an IP but  you may check the ones below.
    http://help.sap.com/bp_biv335/BI_EN/html/BW/SalesPlan.htm
    http://help.sap.com/bp_biv235/bi_en/html/BW/CostCtrPlan.htm
    Regards.

  • CRM 5.2 Best Practices

    Hi All;
    I'm basically looking for Best Practices related to the new CRM 5.2. I need almost everything starting from Internet Sales Best Practices landing to activity and opportunity management.
    Does anybody have documentation i could be interested in?
    Thank you in advance.

    Thank you Mike.
    If it's possible i would like to have some sort of scenarios ready tailored on the new CRM 5.2... I'm preparing for my company some kind of demos where we wanna show the potentiality of the new UI.
    I would lilke something more specific or a way how to get that...
    Thank you again

  • Seeking advice on Best Practices for XML Storage Options - XMLTYPE

    Sparc64
    11.2.0.2
    During OOW12 I tried to attend every xml session I could. There was one where a Mr. Drake was explaining something about not using clob
    as an attribute to storing the xml and that "it will break your application."
    We're moving forward with storing the industry standard invoice in an xmltype column, but Im not concerned that our table definition is not what was advised:
    --i've dummied this down to protect company assets
      CREATE TABLE "INVOICE_DOC"
       (     "INVOICE_ID" NUMBER NOT NULL ENABLE,
         "DOC" "SYS"."XMLTYPE"  NOT NULL ENABLE,
         "VERSION" VARCHAR2(256) NOT NULL ENABLE,
         "STATUS" VARCHAR2(256),
         "STATE" VARCHAR2(256),
         "USER_ID" VARCHAR2(256),
         "APP_ID" VARCHAR2(256),
         "INSERT_TS" TIMESTAMP (6) WITH LOCAL TIME ZONE,
         "UPDATE_TS" TIMESTAMP (6) WITH LOCAL TIME ZONE,
          CONSTRAINT "FK_####_DOC_INV_ID" FOREIGN KEY ("INVOICE_ID")
                 REFERENCES "INVOICE_LO" ("INVOICE_ID") ENABLE
       ) SEGMENT CREATION IMMEDIATE
    INITRANS 20  
    TABLESPACE "####_####_DATA"
           XMLTYPE COLUMN "DOC" STORE AS BASICFILE CLOB  (
      TABLESPACE "####_####_DATA"  XMLTYPE COLUMN "DOC" STORE AS BASICFILE CLOB  (
      TABLESPACE "####_####_DATA" ENABLE STORAGE IN ROW CHUNK 16384 RETENTION
      NOCACHE LOGGING
      STORAGE(INITIAL 81920 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645
      PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_FLASH_CACHE DEFAULT))
    XMLSCHEMA "http://mycompanynamehere.com/xdb/Invoice###.xsd" ELEMENT "Invoice" ID #####"
    {code}
    What is a best practice for this type of table?  Yes, we intend on registering the schema against an xsd.
    Any help/advice would be appreciated.
    -abe                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

    Hi,
    I suggest you read this paper : Oracle XML DB : Choosing the Best XMLType Storage Option for Your Use Case
    It is available on the XML DB home page along with other documents you may be interested in.
    To sum up, the storage method you need depends on the requirement, i.e. how XML data is accessed.
    There was one where a Mr. Drake was explaining something about not using clob as an attribute to storing the xml and that "it will break your application."I think the message Mark Drake wanted to convey is that CLOB storage is now deprecated and shouldn't be used anymore (though still supported for backward compatibility).
    The default XMLType storage starting with version 11.2.0.2 is now Binary XML, a posted-parsed binary format that optimizes both storage size and data access (via XQuery), so you should at least use it instead of the BASICFILE CLOB.
    Schema-based Binary XML is also available, it adds another layer of "awareness" for Oracle to manage instance documents.
    To use this feature, the XML schema must be registered with "options => dbms_xmlschema.REGISTER_BINARYXML".
    The other common approach for schema-based XML is Object-Relational storage.
    BTW... you may want to post here next time, in the dedicated forum : {forum:id=34}
    Mark Drake is one of the regular user, along with Marco Gralike you've probably seen too at OOW.
    Edited by: odie_63 on 18 oct. 2012 21:55

  • Best Practices:: How to generate XML file from a ResultSet

    Hi all,
    Could someone please suggest the best practices of how to generate an XML file from a resultset? I am developing a web application in Java with Oracle database and one of my tasks is to generate an XML file when the user, for example, click a "download as XML" button on the JSP. The application is basically like an Order with line items. I am using Struts and my first thought has been to have an action class which will extend struts's DownloadAction and through StAX's Iterator API to create an XML file. I intend to have a POJO which will have properties of all columns of my order and line items tables so that for each order I get all line items and:
    1. Write order details then
    2. Through an iterator write line items of that order to an XML file.
    I will greatly appreciate for comments or suggestions on the best way to do this through any pointers on the Web.
    alex

    Use a OracleWebRowSet in which an XML representation of the result set may be obtained.
    http://www.oracle.com/technology/sample_code/tech/java/sqlj_jdbc/files/oracle10g/webrowset/Readme.html
    http://download.oracle.com/docs/cd/B28359_01/java.111/b31224/jcrowset.htm

  • Best practice "changing several related objects via BDT" (Business Data Toolset) / Mehrere verbundene Objekte per BDT ändern

    Hallo,
    I want to start a
    discussion, to find a best practice method to change several related master
    data objects via BDT. At the moment we are faced with miscellaneous requirements,
    where we have a master data object which uses BDT framework for maintenance (in
    our case an insured objects). While changing or creating the insured objects a
    several related objects e.g. Business Partner should also be changed or
    created. So am searching for a best practices approach how to implement such a
    solution.
    One Idea was to so call a
    report via SUBMIT AND RETURN in Event DSAVC or DSAVE. Unfortunately this implementation
    method has only poor options to handle errors. Second it is also hard to keep LUW
    together.
    Another idea is to call an additional
    BDT instance in the DCHCK-event via FM BDT_INSTANCE_SELECT and the parameters
    iv_xpush_classic = ‘X’ and iv_xpop_classic = ‘X’. At this time we didn’t get
    this solution working correctly, because there is always something missing
    (e.g. global memory is not transferred correctly between the two BDT instances).
    So hopefully you can report
    about your implementations to find a best practice approach for facing such
    requirements.
    Hallo
    ich möchte an der Stelle eine Diskussion starten um einen Best Practice
    Ansatz zu finden, der eine BDT Implementierung/Erweiterung beschreibt, bei der
    verschiedene abhängige BDT-Objekte geändert werden. Momentan treffen bei uns
    mehrere Anforderungen an, bei deinen Änderungen eines BDT Objektes an ein
    anderes BDT Objekte vererbt werden sollen. Sprich es sollen weitere Objekte geänderte
    werden, wenn ein Objekt (in unserem Fall ein Versicherungsvertrag) angelegt
    oder geändert wird (zum Beispiel ein Geschäftspartner)
    Die erste unserer Ideen war es, im Zeitpunkt DSAVC oder DSAVE einen
    Report per SUBMIT AND RETURN aufzurufen. Dieser sollte dann die abhängigen Änderungen
    durchführen. Allerdings gibt es hier Probleme mit der Fehlerbehandlung, da
    diese asynchrone stattfinden muss. Weiterhin ist es auch schwer die Konsistenz der
    LUW zu garantieren.
    Ein anderer Ansatz den wir verfolgt hatten, war im Zeitpunkt
    DCHCK per FuBA BDT_INSTANCE_SELECT und den Parameter iv_xpush_classic = ‘X’ and
    iv_xpop_classic = ‘X’ eine neue BDT Instanz zu erzeugen. Leider konnten wir diese
    Lösung nicht endgültig zum Laufen bekommen, da es immer Probleme beim
    Übertragen der globalen Speicher der einzelnen BDT Instanzen gab.
    Ich hoffe Ihr könnt hier eure Implementierungen kurz beschreiben, dass wir
    eine Best Practice Ansatz für das Thema finden können
    BR/VG
    Dominik

  • [CS5.5/6] - XML / Data Merge questions & Best practice.

    Fellow Countrymen (and women),
    I work as a graphic designer for a large outlet chain retailer which is constantly growing our base of centers.  This growth has brought a workload that used to be manageable with but two people to a never ending sprint with five.  Much of what we do is print, which is not my forte, but is also generally a disorganized, ad-hoc affair into which I am wading to try to help reduce overall strain.
    Upon picking up InDesign I noted the power of the simple Data Merge function and have added it to our repetoire in mass merging data sources.  There are some critical failures I see in this as a tool going forward for our purposes, however:
    1) Data Merge cannot handle information stored and categorized in a singular column well.  As an example we have centers in many cities, and each center has its own list of specific stores.  Data merge cannot handle a single column, or even multiple column list of these stores very easily and has forced us into some manual operations to concatenate the data into one cell and then, using delimiter characters, find and replace hard returns to seperate them.
    2) Data Merge offers no method of alternate alignment of data, or selection by ranges.  That is to say:  I cannot tell Data merge to start at Cell1 in one column, and in another column select say... Cell 42 as the starting point.
    3) Data merge only accepts data organized in a very specific, and generally inflexible pattern.
    These are just a few limitations.
    ON TO MY ACTUAL DILEMMA aka Convert to XML or not?
    Recently my coworker has suggested we move toward using XML as a repository / delivery system that helps us quickly get data from our SQL database into a usable form in InDesign. 
    I've watched some tutorials on Lynda.com and havent yet seen a clear answer to a very simple question:
    "Can XML help to 'merge' large, dynamic, data sets like a list of 200 stores per center over 40 centers based off of a single template file?"
    What I've seen is that I would need to manually duplicate pages, linking the correct XML entry as I go rather than the program generating a set of merged pages like that from Data Merge with very little effort on my part.  Perhaps setting up a master page would allow for easy drag and drop fields for my XML data?
    I'm not an idiot, I'm simply green with this -- and it's kind of scary because I genuinely want us to proceed forward with the most flexible, reliable, trainable and sustainable solution.  A tall order, I know.  Correct me if I'm wrong, but XML is that beast, no?
    Formatting the XML
    Currently I'm afraid our XML feed for our centers isnt formatted correctly with the current format looking as such:
    <BRANDS>
         <BRAND>
              • BrandID = xxxx
              [Brand Name]
              [Description]
              [WebMoniker]
              <CATEGORIES>
                   <CATEGORY>
                        • xmlns = URL
                        • WebMoniker = category_type
              <STORES>
                   <STORE>
                        • StoreID = ID#
                        • CenterID = ID#
    I dont think this is currently usable because if I wanted to create a list of stores from a particular center, that information is stored as an attribute of the <Store> tag, buried deep within the data, making it impossible to 'drag-n-drop'. 
    Not to mention much of the important data is held in attributes rather than text fields which are children of the tag.
    Im thinking of proposing the following organizational layout:
    <CENTERS>
         <CENTER>
         [Center_name]
         [Center_location]
              <CATEGORIES>
                   <CATEGORY>
                        [Category_Type]
                        <BRANDS>
                             <BRAND>
                                  [Brand_name]
    My thought is that if I have the <CENTER> tag then I can simply drag that into a frame and it will auto populate all of the brands by Category (as organized in the XML) for that center into the frame.
    Why is this important?
    This is used on multiple documents in different layout styles, and since our store list is ever changes as leases end or begin, over 40 centers this becomes a big hairy monster.  We want this to be as automated as possible, but I'd settle for a significant amount of dragging and dropping as long as it is simple and straightforward.  I have a high tollerance for druding through code and creating work arounds but my co-workers do not.  This needs to be a system that is repeatable and understandable and needs to be able to function whether I'm here or not -- Mainly because I would like to step away from the responsibility of setting it up every time
    I'd love to hear your raw, unadulterated thoughts on the subject of Data merge and XML usage to accomplish these sorts of tasks.  What are your best practices and how would you / do you accomplish these operations?
    Regards-
    Robert

    From what I've gleaned through watching Lynda tutorials on the subject is that what I'm hoping to do is indeed possible.
    Peter, I dont disagree with you that there is a steep learning curve for me as the instigator / designer of this method for our team, but in terms of my teammates and end-users that will be softened considerably.  Even so I'm used to steep learning curves and the associated frustrations -- but I cope well with new learning and am self taught in many tools and programs.
    Flow based XML structures:
    It seems as though as long as the initial page is set up correctly using imported XML, individual data records that cascade in a logical fashion can be flowed automatically into new pages.  Basically what you do is to create an XML based layout with the dynamic portion you wish to flow in a single frame, apply paragraph styles to the different tags appropriately and then after deleting unused records, reimport the XML with some specific boxes checked (depending on how you wish to proceed).
    From there simply dragging the data root into the frame will cause overset text as it imports all the XML information into the frame.  Assuming that everything is cascaded correctly using auto-flow will cause new pages to be automatically generated with the tags correctly placed in a similar fashion to datamerge -- but far more powerful and flexible. 
    The issue then again comes down to data organization in the XML file.  In order to use this method the data must be organized in the same order in which it will be displayed.  For example if I had a Lastname field, and a Firstname field in that order, I could not call the Firstname first without faulting the document using the flow method.  I could, however, still drag and drop content from each tag into the frame and it would populate correctly regardless of the order of appearance in the XML.
    Honestly either method would be fantastic for our current set of projects, however the flow method may be particularly useful in jobs that would require more than 40 spreads or simple layouts with huge amounts of data to be merged.

  • When to use unattend.xml in task sequence - best practice?

    Hi, I've tried researching this but not found an answer to my specific query.
    We have ConfigMgr 2012 R2 with MDT 2013 although I don't think this is an MDT specific question.
    I'm trying to create a Build and Capture task sequence for our Windows Server 2008 R2 and Server 2012 /2012R2 server builds utilising an UNATTEND.XML file to make some customisations that can be deployed for every build afterwards in a Deployment Task Sequence.
    Specifically the addition of some Windows Features like SNMP and it's configuration and the addition of the Telnet Client. There are other bits like language settings and configuration items but I'm specifically interested in the Features part for my question.
    In CM 2012R2 you now have the option under the "Apply Operating System" to use a captured image or an original installation source. However they work differently if you specify the use of the same unattended answer file.
    The "image" deployment ignores all of the "add features" sections of the XML file and the "installation source" loses the  configuration options from SNMP from the XML file. When you then deploy the captured image using
    the same unattend.xml again the one from the "installer" now has all the SNMP features required and the one from the "image" is still missing everything.
    So my question is as follows.
    What is best practice for specifying an unattend.xml file in a task sequence. Is it in the build and capture TS or in the Deployment TS ?
    or
    Do I need multiple XML files, one for build and capture with some bits in and another for deployment with the rest in?
    or
    Should I be doing something else?
    Although this is specifically asking about Server O/S we will be using the same methodology for Windows 7 deployment.

    In this case DISM is only used to add the actual features... for configuration you could use a simple script that runs afterwards. Sample registry file:
    SAMPLE REG FILE - HKLM-SNMP.reg
    Windows Registry Editor Version 5.00
    [HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\services\SNMP\Parameters]
    "NameResolutionRetries"=dword:00000010
    "EnableAuthenticationTraps"=dword:00000001
    [HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\services\SNMP\Parameters\PermittedManagers]
    [HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\services\SNMP\Parameters\RFC1156Agent]
    "sysServices"=dword:0000004f
    "sysLocation"=""
    "sysContact"=""
    [HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\services\SNMP\Parameters\TrapConfiguration]
    [HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\services\SNMP\Parameters\TrapConfiguration\public]
    "1"="127.0.0.1"
    [HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\services\SNMP\Parameters\ValidCommunities]
    "public"=dword:00000004
    Sample batch file:
    SAMPLE SCRIPT FILE - ConfigureSNMPService.bat
    @ECHO OFF
    net stop "SNMP Service"
    regedit /s HKLM-SNMP.reg
    net start "SNMP Service"
    Also some settings for SNMP can be controlled through group policy:
    http://serverfault.com/questions/285762/group-policy-for-multiple-snmp-permitted-managers

  • Best practice for optimizing processing of big XMLs?

    All,
    What is the best practice when dealing with large XML files ... (say couple of MBs).
    Instead of having to read the file from the file system everytime a static method is run, what would be the best way in which the program reads the file once and then keeps it in memory. So the next time it would not have to read and parse it all over again?
    Currently my code just read the file in the static method like ...
    public static String doOperation(String path,...) throws Exception
    try{           
    String masterFile = path+"configfile.xml";
    Document theFile = (Document)getDocument(masterFile);
    Element root = theFile.getDocumentElement();
    NodeList nl = root.getChildNodes();
    // ... operations on file
    Optimization tips and tricks most appreciated :-)
    Thanks,
    David

    The best practice for multi-megabyte XML files is not to have them at all.
    However if you must, presumably you don't need all of the information in your XML, repeatedly. Or do you? If you need a little bit of it here, then another little bit of it there, then yet another little bit of it later, then you shouldn't have stored your data in one big XML.
    Sorry if that sounds unhelpful, but I'm having trouble imagining a scenario when you need all the data in an XML document repeatedly. Perhaps you could expand on your design?
    PC&#178;

  • Info on best practices to add jar references in server.xml

    Hi
    I want to add some jar references (pcl.jar & struts.jar) in server.xml.
    Can someone let me know if this can be added to <shared-library name="global.tag.libraries" version="1.0" library-compatible="true"> ?
    What is the best practice to add such entries in server.xml?
    Thanks
    Badri

    If you want to use it in BPEL it should be placed in oracle.bpel.common
    cheers
    James

Maybe you are looking for

  • Songs getting mixed with each other

    Say 2 of the songs in my library are Song1 and Song2. On my iPhone Song1 starts actually as Song1 but the last part of it is Song2. This problem does not happen in my pc or imac itunes libraries and it does not happen with all songs. I have reset my

  • ENCODER 2.5 Issue on XP

    Using Live Media Encoder version 2.5 (also tried ver 3.0 with same result) running on XP via bootcamp on a Mac. On open looks good but as soon as I go and change ANY setting within it locks up/freezes the application and must close it down. Any one r

  • Google Domain E-Mail sync

    I am trying to create some documentation for students at the college I am currently employed at. I do not have access to student Id's or passwords. We are currently using google to provide e-mail for our students. We are using a seperate domain, mean

  • How to upload data for me01 using lsmw recording method

    *dear expert please tell me the complete procedure for how to upload data in me01 transaction using lsmw recording method. Moderator Message: Duplicate post locked. Edited by: Vinod Kumar on May 8, 2011 7:56 PM

  • Saving added attachments in emails

    Can anybody tell me how to save these attachments on internal/added memory?