Tuning & Testing

Hi,
How to do Performance Tuning, Integration Testing, Unit Testing, Dump Analysis in ABAP?
anybody got screen shots / with example?
Thanks
Seshu

HI ,
Performance Tuning..
1. Check the SELECT STatements
   a. Specify necessary conditions to retrieve the data
   b. Specify all the key fields if known.
   c. Make sure that the tables are not empty when used 
      in SELECT --- FOR ALL ENTRIES.
   d. Always clear/ refresh variables and tables if not 
      required any more in the program.
   e. Use the INTO TABLE instead of INTO CORRESPONDING
      FIELDS OF TABLE in a SELECT statement.
2. Check the Nested Loops in the programs.
3. Avoid SELECT inside LOOPS.
   Instead of using SELECT inside loops...use the
   appropriate JOINS to select records from two or more 
   tables.
4. When using the CONTROLS, optimally use the FLUSH
   Commands.
5. Read the tables by Binary Search Method specifying
   the Keys used for Sorting that table.
6. Use of right kind of Internal tables (Standard,  
   Sorted, and Hash tabes) is very important.
Integration Testing.
It is always good practise test the individual applications thoroughly before going for integration testing. The application accessing all the submodules is tested only when the individual components are working correctly.
Unit Testing.
1. Initially check the coding standards specified by  
   your company/ client.
2. Check the performance tuning techniques specified in
    above.
3. Check the application specific conditions.
4. Check for all the errors and the corresponding error
   handling conitions.
5. Do the negative testing, that enter incorrect data
   to process records.
Regards,
Vara

Similar Messages

  • Some Thoughts On An OWB Performance/Testing Framework

    Hi all,
    I've been giving some thought recently to how we could build a performance tuning and testing framework around Oracle Warehouse Builder. Specifically, I'm looking at was in which we can use some of the performance tuning techniques described in Cary Millsap/Jeff Holt's book "Optimizing Oracle Performance" to profile and performance tune mappings and process flows, and to use some of the ideas put forward in Kent Graziano's Agile Methods in Data Warehousing paper http://www.rmoug.org/td2005pres/graziano.zip and Steven Feuernstein's utPLSQL project http://utplsql.sourceforge.net/ to provide an agile/test-driven way of developing mappings, process flows and modules. The aim of this is to ensure that the mappings we put together are as efficient as possible, work individually and together as expected, and are quick to develop and test.
    At the moment, most people's experience of performance tuning OWB mappings is firstly to see if it runs set-based rather than row-based, then perhaps to extract the main SQL statement and run an explain plan on it, then check to make sure indexes etc are being used ok. This involves a lot of manual work, doesn't factor in the data available from the wait interface, doesn't store the execution plans anywhere, and doesn't really scale out to encompass entire batches of mapping (process flows).
    For some background reading on Cary Millsap/Jeff Holt's approach to profiling and performance tuning, take a look at http://www.rittman.net/archives/000961.html and http://www.rittman.net/work_stuff/extended_sql_trace_and_tkprof.htm. Basically, this approach traces the SQL that is generated by a batch file (read: mapping) and generates a file that can be later used to replay the SQL commands used, the explain plans that relate to the SQL, details on what wait events occurred during execution, and provides at the end a profile listing that tells you where the majority of your time went during the batch. It's currently the "preferred" way of tuning applications as it focuses all the tuning effort on precisely the issues that are slowing your mappings down, rather than database-wide issues that might not be relevant to your mapping.
    For some background information on agile methods, take a look at Kent Graziano's paper, this one on test-driven development http://c2.com/cgi/wiki?TestDrivenDevelopment , this one http://martinfowler.com/articles/evodb.html on agile database development, and the sourceforge project for utPLSQL http://utplsql.sourceforge.net/. What this is all about is having a development methodology that builds in quality but is flexible and responsive to changes in customer requirements. The benefit of using utPLSQL (or any unit testing framework) is that you can automatically check your altered mappings to see that they still return logically correct data, meaning that you can make changes to your data model and mappings whilst still being sure that it'll still compile and run.
    Observations On The Current State of OWB Performance Tuning & Testing
    At present, when you build OWB mappings, there is no way (within the OWB GUI) to determine how "efficient" the mapping is. Often, when building the mapping against development data, the mapping executes quickly and yet when run against the full dataset, problems then occur. The mapping is built "in isolation" from its effect on the database and there is no handy tool for determining how efficient the SQL is.
    OWB doesn't come with any methodology or testing framework, and so apart from checking that the mapping has run, and that the number of rows inserted/updated/deleted looks correct, there is nothing really to tell you whether there are any "logical" errors. Also, there is no OWB methodology for integration testing, unit testing, or any other sort of testing, and we need to put one in place. Note - OWB does come with auditing, error reporting and so on, but there's no framework for guiding the user through a regime of unit testing, integration testing, system testing and so on, which I would imagine more complete developer GUIs come with. Certainly there's no built in ability to use testing frameworks such as utPLSQL, or a part of the application that let's you record whether a mapping has been tested, and changes the test status of mappings when you make changes to ones that they are dependent on.
    OWB is effectively a code generator, and this code runs against the Oracle database just like any other SQL or PL/SQL code. There is a whole world of information and techniques out there for tuning SQL and PL/SQL, and one particular methodology that we quite like is the Cary Millsap/Jeff Holt "Extended SQL Trace" approach that uses Oracle diagnostic events to find out exactly what went on during the running of a batch of SQL commands. We've been pretty successful using this approach to tune customer applications and batch jobs, and we'd like to use this, together with the "Method R" performance profiling methodology detailed in the book "Optimising Oracle Performance", as a way of tuning our generated mapping code.
    Whilst we want to build performance and quality into our code, we also don't want to overburden developers with an unwieldy development approach, because what we'll know will happen is that after a short amount of time, it won't get used. Given that we want this framework to be used for all mappings, it's got to be easy to use, cause minimal overhead, and have results that are easy to interpret. If at all possible, we'd like to use some of the ideas from agile methodologies such as eXtreme Programming, SCRUM and so on to build in quality but minimise paperwork.
    We also recognise that there are quite a few settings that can be changed at a session and instance level, that can have an effect on the performance of a mapping. Some of these include initialisation parameters that can change the amount of memory assigned to the instance and the amount of memory subsequently assigned to caches, sort areas and the like, preferences that can be set so that indexes are preferred over table scans, and other such "tweaks" to the Oracle instance we're working with. For reference, the version of Oracle we're going to use to both run our code and store our data is Oracle 10g 10.1.0.3 Enterprise Edition, running on Sun Solaris 64-bit.
    Some initial thoughts on how this could be accomplished
    - Put in place some method for automatically / easily generating explain plans for OWB mappings (issue - this is only relevant for mappings that are set based, and what about pre- and post- mapping triggers)
    - Put in place a method for starting and stopping an event 10046 extended SQL trace for a mapping
    - Put in place a way of detecting whether the explain plan / cost / timing for a mapping changes significantly
    - Put in place a way of tracing a collection of mappings, i.e. a process flow
    - The way of enabling tracing should either be built in by default, or easily added by the OWB developer. Ideally it should be simple to switch it on or off (perhaps levels of event 10046 tracing?)
    - Perhaps store trace results in a repository? reporting? exception reporting?
    at an instance level, come up with some stock recommendations for instance settings
    - identify the set of instance and session settings that are relevant for ETL jobs, and determine what effect changing them has on the ETL job
    - put in place a regime that records key instance indicators (STATSPACK / ASH) and allows reports to be run / exceptions to be reported
    - Incorporate any existing "performance best practices" for OWB development
    - define a lightweight regime for unit testing (as per agile methodologies) and a way of automating it (utPLSQL?) and of recording the results so we can check the status of dependent mappings easily
    other ideas around testing?
    Suggested Approach
    - For mapping tracing and generation of explain plans, a pre- and post-mapping trigger that turns extended SQL trace on and off, places the trace file in a predetermined spot, formats the trace file and dumps the output to repository tables.
    - For process flows, something that does the same at the start and end of the process. Issue - how might this conflict with mapping level tracing controls?
    - Within the mapping/process flow tracing repository, store the values of historic executions, have an exception report that tells you when a mapping execution time varies by a certain amount
    - get the standard set of preferred initialisation parameters for a DW, use these as the start point for the stock recommendations. Identify which ones have an effect on an ETL job.
    - identify the standard steps Oracle recommends for getting the best performance out of OWB (workstation RAM etc) - see OWB Performance Tips http://www.rittman.net/archives/001031.html and Optimizing Oracle Warehouse Builder Performance http://www.oracle.com/technology/products/warehouse/pdf/OWBPerformanceWP.pdf
    - Investigate what additional tuning options and advisers are available with 10g
    - Investigate the effect of system statistics & come up with recommendations.
    Further reading / resources:
    - Diagnosing Performance Problems Using Extended Trace" Cary Millsap
    http://otn.oracle.com/oramag/oracle/04-jan/o14tech_perf.html
    - "Performance Tuning With STATSPACK" Connie Dialeris and Graham Wood
    http://www.oracle.com/oramag/oracle/00-sep/index.html?o50tun.html
    - "Performance Tuning with Statspack, Part II" Connie Dialeris and Graham Wood
    http://otn.oracle.com/deploy/performance/pdf/statspack_tuning_otn_new.pdf
    - "Analyzing a Statspack Report: A Guide to the Detail Pages" Connie Dialeris and Graham Wood
    http://www.oracle.com/oramag/oracle/00-nov/index.html?o60tun_ol.html
    - "Why Isn't Oracle Using My Index?!" Jonathan Lewis
    http://www.dbazine.com/jlewis12.shtml
    - "Performance Tuning Enhancements in Oracle Database 10g" Oracle-Base.com
    http://www.oracle-base.com/articles/10g/PerformanceTuningEnhancements10g.php
    - Introduction to Method R and Hotsos Profiler (Cary Millsap, free reg. required)
    http://www.hotsos.com/downloads/registered/00000029.pdf
    - Exploring the Oracle Database 10g Wait Interface (Robin Schumacher)
    http://otn.oracle.com/pub/articles/schumacher_10gwait.html
    - Article referencing an OWB forum posting
    http://www.rittman.net/archives/001031.html
    - How do I inspect error logs in Warehouse Builder? - OWB Exchange tip
    http://www.oracle.com/technology/products/warehouse/pdf/Cases/case10.pdf
    - What is the fastest way to load data from files? - OWB exchange tip
    http://www.oracle.com/technology/products/warehouse/pdf/Cases/case1.pdf
    - Optimizing Oracle Warehouse Builder Performance - Oracle White Paper
    http://www.oracle.com/technology/products/warehouse/pdf/OWBPerformanceWP.pdf
    - OWB Advanced ETL topics - including sections on operating modes, partition exchange loading
    http://www.oracle.com/technology/products/warehouse/selfserv_edu/advanced_ETL.html
    - Niall Litchfield's Simple Profiler (a creative commons-licensed trace file profiler, based on Oracle Trace Analyzer, that displays the response time profile through HTMLDB. Perhaps could be used as the basis for the repository/reporting part of the project)
    http://www.niall.litchfield.dial.pipex.com/SimpleProfiler/SimpleProfiler.html
    - Welcome to the utPLSQL Project - a PL/SQL unit testing framework by Steven Feuernstein. Could be useful for automating the process of unit testing mappings.
    http://utplsql.sourceforge.net/
    Relevant postings from the OTN OWB Forum
    - Bulk Insert - Configuration Settings in OWB
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=291269&tstart=30&trange=15
    - Default Performance Parameters
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=213265&message=588419&q=706572666f726d616e6365#588419
    - Performance Improvements
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=270350&message=820365&q=706572666f726d616e6365#820365
    - Map Operator performance
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=238184&message=681817&q=706572666f726d616e6365#681817
    - Performance of mapping with FILTER
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=273221&message=830732&q=706572666f726d616e6365#830732
    - Poor mapping performance
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=275059&message=838812&q=706572666f726d616e6365#838812
    - Optimizing Mapping Performance With OWB
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=269552&message=815295&q=706572666f726d616e6365#815295
    - Performance of mapping with FILTER
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=273221&message=830732&q=706572666f726d616e6365#830732
    - Performance of the OWB-Repository
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=66271&message=66271&q=706572666f726d616e6365#66271
    - One large JOIN or many small ones?
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=202784&message=553503&q=706572666f726d616e6365#553503
    - NATIVE PL SQL with OWB9i
    http://forums.oracle.com/forums/thread.jsp?forum=57&thread=270273&message=818390&q=706572666f726d616e6365#818390
    Next Steps
    Although this is something that I'll be progressing with anyway, I'd appreciate any comment from existing OWB users as to how they currently perform performance tuning and testing. Whilst these are perhaps two distinct subject areas, they can be thought of as the core of an "OWB Best Practices" framework and I'd be prepared to write the results up as a freely downloadable whitepaper. With this in mind, does anyone have an existing best practices for tuning or testing, have they tried using SQL trace and TKPROF to profile mappings and process flows, or have you used a unit testing framework such as utPLSQL to automatically test the set of mappings that make up your project?
    Any feedback, add it to this forum posting or send directly through to me at [email protected]. I'll report back on a proposed approach in due course.

    Hi Mark,
    interesting post, but I think you may be focusing on the trees, and losing sight of the forest.
    Coincidentally, I've been giving quite a lot of thought lately to some aspects of your post. They relate to some new stuff I'm doing. Maybe I'll be able to answer in more detail later, but I do have a few preliminary thoughts.
    1. 'How efficient is the generated code' is a perennial topic. There are still some people who believe that a code generator like OWB cannot be in the same league as hand-crafted SQL. I answered that question quite definitely: "We carefully timed execution of full-size runs of both the original code and the OWB versions. Take it from me, the code that OWB generates is every bit as fast as the very best hand-crafted and fully tuned code that an expert programmer can produce."
    The link is http://www.donnapkelly.pwp.blueyonder.co.uk/generated_code.htm
    That said, it still behooves the developer to have a solid understanding of what the generated code will actually do, such as how it will take advantage of indexes, and so on. If not, the developer can create such monstrosities as lookups into an un-indexed field (I've seen that).
    2. The real issue is not how fast any particular generated mapping runs, but whether or not the system as a whole is fit for purpose. Most often, that means: does it fit within its batch update window? My technique is to dump the process flow into Microsoft Project, and then to add the timings for each process. That creates a Critical Path, and then I can visually inspect it for any bottleneck processes. I usually find that there are not more than one or two dogs. I'll concentrate on those, fix them, and re-do the flow timings. I would add this: the dogs I have seen, I have invariably replaced. They were just garbage, They did not need tuning at all - just scrapping.
    Gee, but this whole thing is minimum effort and real fast! I generally figure that it takes maybe a day or two (max) to soup up system performance to the point where it whizzes.
    Fact is, I don't really care whether there are a lot of sub-optimal processes. All I really care about is performance of the system as a whole. This technique seems to work for me. 'Course, it depends on architecting the thing properly in the first place. Otherwise, no amount of tuning of going to help worth a darn.
    Conversely (re. my note about replacing dogs) I do not think I have ever tuned a piece of OWB-generated code. Never found a need to. Not once. Not ever.
    That's not to say I do not recognise the value of playing with deployment configuration parameters. Obviously, I set auditing=none, and operating mode=set based, and sometimes, I play with a couple of different target environments to fool around with partitioning, for example. Nonetheless, if it is not a switch or a knob inside OWB, I do not touch it. This is in line with my dictat that you shall use no other tool than OWB to develop data warehouses. (And that includes all documentation!). (OK, I'll accept MS Project)
    Finally, you raise the concept of a 'testing framework'. This is a major part of what I am working on at the moment. This is a tough one. Clearly, the developer must unit test each mapping in a design-model-deploy-execute cycle, paying attention to both functionality and performance. When the developer is satisifed, that mapping will be marked as 'done' in the project workbook. Mappings will form part of a stream, executed as a process flow. Each process flow will usually terminate in a dimension, a fact, or an aggregate. Each process flow will be tested as an integrated whole. There will be test strategies devised, and test cases constructed. There will finally be system tests, to verify the validity of the system as a production-grade whole. (stuff like recovery/restart, late-arriving data, and so on)
    For me, I use EDM (TM). That's the methodology I created (and trademarked) twenty years ago: Evolutionary Development Methodology (TM). This is a spiral methodology based around prototyping cycles within Stage cycles within Release cycles. For OWB, a Stage would consist (say) of a Dimensional update. What I am trying to now is to graft this within a traditional waterfall methodology, and I am having the same difficulties I had when I tried to do it then.
    All suggestions on how to do that grafting gratefully received!
    To sum up, I 'm kinda at a loss as to why you want to go deep into OWB-generated code performance stuff. Jeepers, architect the thing right, and the code runs fast enough for anyone. I've worked on ultra-large OWB systems, including validating the largest data warehouse in the UK. I've never found any value in 'tuning' the code. What I'd like you to comment on is this: what will it buy you?
    Cheers,
    Donna
    http://www.donnapkelly.pwp.blueyonder.co.uk

  • Maximize server uptime - memory leak?

    Since we stabilized the newly introduced WLI 8.1 application we are now fine tuning the JVM. We are facing some kind of memory leak which forces us to reboot the WLS instances daily.
    I'm now asked to identify some strategies how we could let the WLI instances run for longer than 1 day. My goal is 7 days, so that the machines must only be touched once a week. The relevant JVM settings are like this:
    -Xms2048m -Xmx2048m -Xmanagement -Djrockit.managementserver.port=30011 -XgcPrio:throughput
    I've choosen the "throughput" strategy as we have here a systems which acts asynchronously most of the time (WLI). I have attached two JRA records. The first ("1day") shows a system which has an uptime of now 16 hours. The heap utilization is almost all the time between 90%-100%. Things get worse after a while. In the consequence we are rebooting the server when we see more and more timeout-exceptions in our WLI layer and heavy GC activity ("average time spent in GC")
    The other record ("leakdetector") shows a system which is now running for almost 3 hours. Here I connected the memory leak detector. The graph looks a bit better / more balanced. Means, that after 3 hours the avg heap util remains between the margins of 40-50 %. The trend of the bottom margin however clearly indicates an increasing memory footprint.
    To understand this in more detail I started using the memory leak analyzer. So far I got no benefit from using the tool, though it looks very impressive. However, when I connect the memory leak analyzer I can observe some strange heap graph changes. I suppose the memory leak detector has some impact onto the GC strategies, doesn't it? My heap graph looks totally different with no leak detector attached. What can be the reason on this?
    Also I'm quite a bit confused why the "Growth" indicator always stays at 0 bytes/sec, even when I can see that objects are getting bigger and bigger. What is the secret here?
    i appreciate every comment you have on my case,
    thanks a lot

    You might also look at approaching the problem for a different perspective. Which transactions, requests, components are creating the most amount of memory? What concurrent requests where active when the memory jumped or an OOME was thrown. I have found this approach effective in getting a good idea on where to focus memory diagnostics after ruling of course a resource capacity issue.
    The following article discusses the difference between a leak and a capacity problem.
    http://www.jinspired.com/products/jxinsight/outofmemoryexceptions.html
    Also you should try high level monitoring to see whether there are global patterns in metric data that provides clues. This blog entry shows what is possible with professional performance management tools with visualizations going beyond pie charts and table views.
    Beautiful Evidence: Metric Monitoring
    http://blog.jinspired.com/?p=33
    There are also many low level memory inspection tools on the market that might help quickly navigate the heap and identify the problem though I think the JRA tool has probably most of the same features.
    Regards,
    William Louth
    JXInsight Product Architect
    CTO, JInspired
    "Java EE tuning, testing, tracing, and monitoring with JXInsight"
    http://www.jinspired.com

  • Anyone running Portal with NOARCHIVELOG mode in the Oracle DB?

    Although we have several high-volume SQL Server customers running their DB in simple recovery mode (no transaction logging) with success, I am trying to determine if anyone is doing the equivalent with Oracle DB.
    Even if you aren't using NOARCHIVELOG mode, please let me know your thoughts on performance, especially since the perception exists that the Portal is less performant on Oracle DB than on SQL Server. If that perception has any basis in reality, there should be an even greater potential for performance gain using Oracle DB, through intelligent tuning.
    The cons of NOARCHIVELOG were pointed out to me by an excellent DBA:
    1. Lose the ability to do hot online physically backup, (both RMAN and Manual)
    2. Lose the ability to do point in time recovery of your database (irrelevant to the Portal)
    3. When something bad happens to your DB, you are guaranteed to lose some data, because you can only recover to last good backup. (This is either acceptable to the business or not. Usually it is.)
    Thanks for your thoughts!

    There are no dumb questions. 'Better' or 'worse' can mean different things from different angles. I won't recommend something to a customer that I saw in an article somewhere, or gained knowledge via hearsay. I have to have tried it successfully with a customer, usually though iterative tuning, testing and monitoring cycles.
    With that said, I see nothing wrong with using NOARCHIVELOG mode for the portal, and am looking to find someone with experience using it with the portal; and I don't mean a portal with four communities and a scant thousand users. I am looking for something more on the scale of our mutual customer who had about 15 front end portal servers, and about 15,000 hits per hour. Who, incidentally, is happily using SQL Server in simple recovery mode, and was game to try it out in the first place.

  • Weblogic tangosol integration

    Hi folks,
    Has any body successfully integrated kodo with tangosol in a clustered enviornment.
    i have some problems with this.
    I am getting a class not found exception.
    Caused by: java.lang.NoClassDefFoundError: com/tangosol/net/NamedCache
    at java.lang.Class.getDeclaredMethods0(Native Method)
    complete stack trace is attached.
    Here is what we did with our application.
    Modified persistence.xml file to include the property <property name="kodo.DataCache" value="tangosol(TangosolCacheName=kodo)"/>
    And added the coherence library to the domain lib.
    We also tried putting the tangosol.jar which contains the com.tangosol.net.NamedCache class file in the application lib.
    Still the same error is occurring.
    Any help is highly appreciated.
    Rgds
    Vivek
    [12 May 2008 15:33:11,627] ERROR [[STANDBY] ExecuteThread: '2' for queue: 'weblogic.kernel.Default (self-tuning)'] test.TestSessionBean:88 - Exception
    <1.0.0 nonfatal general error> org.apache.openjpa.persistence.PersistenceException: [kodo.datacache.TangosolDataCache@1f75468:null].TangosolCacheName = kodo
    at org.apache.openjpa.kernel.AbstractBrokerFactory.newBroker(AbstractBrokerFactory.java:193)
    at org.apache.openjpa.kernel.DelegatingBrokerFactory.newBroker(DelegatingBrokerFactory.java:142)
    at org.apache.openjpa.persistence.EntityManagerFactoryImpl.createEntityManager(EntityManagerFactoryImpl.java:190)
    at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:74)
    at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:36)
    at org.apache.openjpa.persistence.EntityManagerFactoryImpl.createEntityManager(EntityManagerFactoryImpl.java:143)
    at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:70)
    at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:36)
    at com.git.test.TestSessionBean.<init>(TestSessionBean.java:54)
    at com.git.test.TestSessionBean_eb2d8g_Impl.<init>(TestSessionBean_eb2d8g_Impl.java:45)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
    at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
    at java.lang.reflect.Constructor.newInstance(Constructor.java:494)
    at java.lang.Class.newInstance0(Class.java:350)
    at java.lang.Class.newInstance(Class.java:303)
    at weblogic.ejb.container.injection.EjbComponentCreatorImpl.getBean(EjbComponentCreatorImpl.java:73)
    at weblogic.ejb.container.manager.BaseEJBManager.createNewBeanInstance(BaseEJBManager.java:216)
    at weblogic.ejb.container.manager.BaseEJBManager.allocateBean(BaseEJBManager.java:233)
    at weblogic.ejb.container.manager.StatelessManager.createBean(StatelessManager.java:305)
    at weblogic.ejb.container.pool.StatelessSessionPool.createBean(StatelessSessionPool.java:173)
    at weblogic.ejb.container.pool.Pool.createInitialBeans(Pool.java:295)
    at weblogic.ejb.container.manager.StatelessManager.initializePool(StatelessManager.java:446)
    at weblogic.ejb.container.deployer.EJBDeployer.initializePools(EJBDeployer.java:1635)
    at weblogic.ejb.container.deployer.EJBDeployer.start(EJBDeployer.java:1464)
    at weblogic.ejb.container.deployer.EJBModule.start(EJBModule.java:513)
    at weblogic.application.internal.flow.ModuleListenerInvoker.start(ModuleListenerInvoker.java:117)
    at weblogic.application.internal.flow.ModuleStateDriver$3.next(ModuleStateDriver.java:204)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:26)
    at weblogic.application.internal.flow.ModuleStateDriver.start(ModuleStateDriver.java:60)
    at weblogic.application.internal.flow.StartModulesFlow.activate(StartModulesFlow.java:26)
    at weblogic.application.internal.BaseDeployment$2.next(BaseDeployment.java:635)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:26)
    at weblogic.application.internal.BaseDeployment.activate(BaseDeployment.java:212)
    at weblogic.application.internal.DeploymentStateChecker.activate(DeploymentStateChecker.java:154)
    at weblogic.deploy.internal.targetserver.AppContainerInvoker.activate(AppContainerInvoker.java:80)
    at weblogic.deploy.internal.targetserver.operations.AbstractOperation.activate(AbstractOperation.java:566)
    at weblogic.deploy.internal.targetserver.operations.ActivateOperation.activateDeployment(ActivateOperation.java:136)
    at weblogic.deploy.internal.targetserver.operations.ActivateOperation.doCommit(ActivateOperation.java:104)
    at weblogic.deploy.internal.targetserver.operations.StartOperation.doCommit(StartOperation.java:139)
    at weblogic.deploy.internal.targetserver.operations.AbstractOperation.commit(AbstractOperation.java:320)
    at weblogic.deploy.internal.targetserver.DeploymentManager.handleDeploymentCommit(DeploymentManager.java:816)
    at weblogic.deploy.internal.targetserver.DeploymentManager.activateDeploymentList(DeploymentManager.java:1223)
    at weblogic.deploy.internal.targetserver.DeploymentManager.handleCommit(DeploymentManager.java:434)
    at weblogic.deploy.internal.targetserver.DeploymentServiceDispatcher.commit(DeploymentServiceDispatcher.java:161)
    at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer.doCommitCallback(DeploymentReceiverCallbackDeliverer.java:181)
    at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer.access$100(DeploymentReceiverCallbackDeliverer.java:12)
    at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer$2.run(DeploymentReceiverCallbackDeliverer.java:67)
    at weblogic.work.SelfTuningWorkManagerImpl$WorkAdapterImpl.run(SelfTuningWorkManagerImpl.java:464)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:200)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:172)
    Caused by: org.apache.openjpa.lib.util.ParseException: [kodo.datacache.TangosolDataCache@1f75468:null].TangosolCacheName = kodo
    at org.apache.openjpa.lib.util.Options.setInto(Options.java:234)
    at org.apache.openjpa.lib.util.Options.setInto(Options.java:187)
    at org.apache.openjpa.lib.conf.Configurations.configureInstance(Configurations.java:421)
    at org.apache.openjpa.lib.conf.Configurations.configureInstance(Configurations.java:375)
    at org.apache.openjpa.lib.conf.PluginListValue.instantiate(PluginListValue.java:93)
    at org.apache.openjpa.lib.conf.ObjectValue.instantiate(ObjectValue.java:79)
    at kodo.datacache.KodoDataCacheManager.instantiateDataCaches(KodoDataCacheManager.java:77)
    at kodo.datacache.KodoDataCacheManager.initialize(KodoDataCacheManager.java:49)
    at org.apache.openjpa.conf.OpenJPAConfigurationImpl.getDataCacheManagerInstance(OpenJPAConfigurationImpl.java:588)
    at org.apache.openjpa.kernel.AbstractBrokerFactory.newBroker(AbstractBrokerFactory.java:169)
    ... 50 more
    Caused by: java.lang.NoClassDefFoundError: com/tangosol/net/NamedCache
    at java.lang.Class.getDeclaredMethods0(Native Method)
    at java.lang.Class.privateGetDeclaredMethods(Class.java:2395)
    at java.lang.Class.privateGetPublicMethods(Class.java:2519)
    at java.lang.Class.getMethods(Class.java:1406)
    at org.apache.openjpa.lib.util.Options.matchOptionToMember(Options.java:310)
    at org.apache.openjpa.lib.util.Options.setInto(Options.java:210)
    ... 59 more

    Hi folks,
    Has any body successfully integrated kodo with tangosol in a clustered enviornment.
    i have some problems with this.
    I am getting a class not found exception.
    Caused by: java.lang.NoClassDefFoundError: com/tangosol/net/NamedCache
    at java.lang.Class.getDeclaredMethods0(Native Method)
    complete stack trace is attached.
    Here is what we did with our application.
    Modified persistence.xml file to include the property <property name="kodo.DataCache" value="tangosol(TangosolCacheName=kodo)"/>
    And added the coherence library to the domain lib.
    We also tried putting the tangosol.jar which contains the com.tangosol.net.NamedCache class file in the application lib.
    Still the same error is occurring.
    Any help is highly appreciated.
    Rgds
    Vivek
    [12 May 2008 15:33:11,627] ERROR [[STANDBY] ExecuteThread: '2' for queue: 'weblogic.kernel.Default (self-tuning)'] test.TestSessionBean:88 - Exception
    <1.0.0 nonfatal general error> org.apache.openjpa.persistence.PersistenceException: [kodo.datacache.TangosolDataCache@1f75468:null].TangosolCacheName = kodo
    at org.apache.openjpa.kernel.AbstractBrokerFactory.newBroker(AbstractBrokerFactory.java:193)
    at org.apache.openjpa.kernel.DelegatingBrokerFactory.newBroker(DelegatingBrokerFactory.java:142)
    at org.apache.openjpa.persistence.EntityManagerFactoryImpl.createEntityManager(EntityManagerFactoryImpl.java:190)
    at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:74)
    at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:36)
    at org.apache.openjpa.persistence.EntityManagerFactoryImpl.createEntityManager(EntityManagerFactoryImpl.java:143)
    at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:70)
    at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:36)
    at com.git.test.TestSessionBean.<init>(TestSessionBean.java:54)
    at com.git.test.TestSessionBean_eb2d8g_Impl.<init>(TestSessionBean_eb2d8g_Impl.java:45)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
    at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
    at java.lang.reflect.Constructor.newInstance(Constructor.java:494)
    at java.lang.Class.newInstance0(Class.java:350)
    at java.lang.Class.newInstance(Class.java:303)
    at weblogic.ejb.container.injection.EjbComponentCreatorImpl.getBean(EjbComponentCreatorImpl.java:73)
    at weblogic.ejb.container.manager.BaseEJBManager.createNewBeanInstance(BaseEJBManager.java:216)
    at weblogic.ejb.container.manager.BaseEJBManager.allocateBean(BaseEJBManager.java:233)
    at weblogic.ejb.container.manager.StatelessManager.createBean(StatelessManager.java:305)
    at weblogic.ejb.container.pool.StatelessSessionPool.createBean(StatelessSessionPool.java:173)
    at weblogic.ejb.container.pool.Pool.createInitialBeans(Pool.java:295)
    at weblogic.ejb.container.manager.StatelessManager.initializePool(StatelessManager.java:446)
    at weblogic.ejb.container.deployer.EJBDeployer.initializePools(EJBDeployer.java:1635)
    at weblogic.ejb.container.deployer.EJBDeployer.start(EJBDeployer.java:1464)
    at weblogic.ejb.container.deployer.EJBModule.start(EJBModule.java:513)
    at weblogic.application.internal.flow.ModuleListenerInvoker.start(ModuleListenerInvoker.java:117)
    at weblogic.application.internal.flow.ModuleStateDriver$3.next(ModuleStateDriver.java:204)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:26)
    at weblogic.application.internal.flow.ModuleStateDriver.start(ModuleStateDriver.java:60)
    at weblogic.application.internal.flow.StartModulesFlow.activate(StartModulesFlow.java:26)
    at weblogic.application.internal.BaseDeployment$2.next(BaseDeployment.java:635)
    at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:26)
    at weblogic.application.internal.BaseDeployment.activate(BaseDeployment.java:212)
    at weblogic.application.internal.DeploymentStateChecker.activate(DeploymentStateChecker.java:154)
    at weblogic.deploy.internal.targetserver.AppContainerInvoker.activate(AppContainerInvoker.java:80)
    at weblogic.deploy.internal.targetserver.operations.AbstractOperation.activate(AbstractOperation.java:566)
    at weblogic.deploy.internal.targetserver.operations.ActivateOperation.activateDeployment(ActivateOperation.java:136)
    at weblogic.deploy.internal.targetserver.operations.ActivateOperation.doCommit(ActivateOperation.java:104)
    at weblogic.deploy.internal.targetserver.operations.StartOperation.doCommit(StartOperation.java:139)
    at weblogic.deploy.internal.targetserver.operations.AbstractOperation.commit(AbstractOperation.java:320)
    at weblogic.deploy.internal.targetserver.DeploymentManager.handleDeploymentCommit(DeploymentManager.java:816)
    at weblogic.deploy.internal.targetserver.DeploymentManager.activateDeploymentList(DeploymentManager.java:1223)
    at weblogic.deploy.internal.targetserver.DeploymentManager.handleCommit(DeploymentManager.java:434)
    at weblogic.deploy.internal.targetserver.DeploymentServiceDispatcher.commit(DeploymentServiceDispatcher.java:161)
    at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer.doCommitCallback(DeploymentReceiverCallbackDeliverer.java:181)
    at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer.access$100(DeploymentReceiverCallbackDeliverer.java:12)
    at weblogic.deploy.service.internal.targetserver.DeploymentReceiverCallbackDeliverer$2.run(DeploymentReceiverCallbackDeliverer.java:67)
    at weblogic.work.SelfTuningWorkManagerImpl$WorkAdapterImpl.run(SelfTuningWorkManagerImpl.java:464)
    at weblogic.work.ExecuteThread.execute(ExecuteThread.java:200)
    at weblogic.work.ExecuteThread.run(ExecuteThread.java:172)
    Caused by: org.apache.openjpa.lib.util.ParseException: [kodo.datacache.TangosolDataCache@1f75468:null].TangosolCacheName = kodo
    at org.apache.openjpa.lib.util.Options.setInto(Options.java:234)
    at org.apache.openjpa.lib.util.Options.setInto(Options.java:187)
    at org.apache.openjpa.lib.conf.Configurations.configureInstance(Configurations.java:421)
    at org.apache.openjpa.lib.conf.Configurations.configureInstance(Configurations.java:375)
    at org.apache.openjpa.lib.conf.PluginListValue.instantiate(PluginListValue.java:93)
    at org.apache.openjpa.lib.conf.ObjectValue.instantiate(ObjectValue.java:79)
    at kodo.datacache.KodoDataCacheManager.instantiateDataCaches(KodoDataCacheManager.java:77)
    at kodo.datacache.KodoDataCacheManager.initialize(KodoDataCacheManager.java:49)
    at org.apache.openjpa.conf.OpenJPAConfigurationImpl.getDataCacheManagerInstance(OpenJPAConfigurationImpl.java:588)
    at org.apache.openjpa.kernel.AbstractBrokerFactory.newBroker(AbstractBrokerFactory.java:169)
    ... 50 more
    Caused by: java.lang.NoClassDefFoundError: com/tangosol/net/NamedCache
    at java.lang.Class.getDeclaredMethods0(Native Method)
    at java.lang.Class.privateGetDeclaredMethods(Class.java:2395)
    at java.lang.Class.privateGetPublicMethods(Class.java:2519)
    at java.lang.Class.getMethods(Class.java:1406)
    at org.apache.openjpa.lib.util.Options.matchOptionToMember(Options.java:310)
    at org.apache.openjpa.lib.util.Options.setInto(Options.java:210)
    ... 59 more

  • Coherence exception in JPA

    Hi,
    I am using open jpa for persistence in my clustered weblogic server envirnment. I am using coherence for locking support and distributed cache mgmt.
    But while creating entity manager from the Persistence.xml file, i am getting an error, for which the base exception is thrown from the tangosol classes.
    Here is a complete stack trace of the error.
    [21 May 2008 18:33:36,650] ERROR [[ACTIVE] ExecuteThread: '0' for queue:
    'weblogic.kernel.Default (self-tuning)'] test.TestSessionBean:83 - Exception
    <1.0.0 nonfatal general error> org.apache.openjpa.persistence.PersistenceException: (Wrapped: Failed to load the factory) null
         at org.apache.openjpa.kernel.AbstractBrokerFactory.newBroker(AbstractBrokerFactory.java:193)
         at org.apache.openjpa.kernel.DelegatingBrokerFactory.newBroker(DelegatingBrokerFactory.java:142)
         at org.apache.openjpa.persistence.EntityManagerFactoryImpl.createEntityManager(EntityManagerFactoryImpl.java:190)
         at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:74)
         at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:36)
         at org.apache.openjpa.persistence.EntityManagerFactoryImpl.createEntityManager(EntityManagerFactoryImpl.java:143)
         at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:70)
         at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:36)
         at com.git.test.TestSessionBean.<init>(TestSessionBean.java:49)
         at com.git.test.TestSessionBean_eb2d8g_Impl.<init>(TestSessionBean_eb2d8g_Impl.java:45)
         at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
         at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
         at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
         at java.lang.reflect.Constructor.newInstance(Constructor.java:494)
         at java.lang.Class.newInstance0(Class.java:350)
         at java.lang.Class.newInstance(Class.java:303)
         at weblogic.ejb.container.injection.EjbComponentCreatorImpl.getBean(EjbComponentCreatorImpl.java:73)
         at weblogic.ejb.container.manager.BaseEJBManager.createNewBeanInstance(BaseEJBManager.java:216)
         at weblogic.ejb.container.manager.BaseEJBManager.allocateBean(BaseEJBManager.java:233)
         at weblogic.ejb.container.manager.StatelessManager.createBean(StatelessManager.java:305)
         at weblogic.ejb.container.pool.StatelessSessionPool.createBean(StatelessSessionPool.java:173)
         at weblogic.ejb.container.pool.Pool.createInitialBeans(Pool.java:295)
         at weblogic.ejb.container.manager.StatelessManager.initializePool(StatelessManager.java:446)
         at weblogic.ejb.container.deployer.EJBDeployer.initializePools(EJBDeployer.java:1635)
         at weblogic.ejb.container.deployer.EJBDeployer.start(EJBDeployer.java:1464)
         at weblogic.ejb.container.deployer.EJBModule.start(EJBModule.java:513)
         at weblogic.application.internal.flow.ModuleListenerInvoker.start(ModuleListenerInvoker.java:117)
         at weblogic.application.internal.flow.ModuleStateDriver$3.next(ModuleStateDriver.java:204)
         at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:26)
         at weblogic.application.internal.flow.ModuleStateDriver.start(ModuleStateDriver.java:60)
         at weblogic.application.internal.flow.StartModulesFlow.activate(StartModulesFlow.java:26)
         at weblogic.application.internal.BaseDeployment$2.next(BaseDeployment.java:635)
         at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:26)
         at weblogic.application.internal.BaseDeployment.activate(BaseDeployment.java:212)
         at weblogic.application.internal.DeploymentStateChecker.activate(DeploymentStateChecker.java:154)
         at weblogic.deploy.internal.targetserver.AppContainerInvoker.activate(AppContainerInvoker.java:80)
         at weblogic.deploy.internal.targetserver.BasicDeployment.activate(BasicDeployment.java:182)
         at weblogic.deploy.internal.targetserver.BasicDeployment.activateFromServerLifecycle(BasicDeployment.java:359)
         at weblogic.management.deploy.internal.DeploymentAdapter$1.doActivate(DeploymentAdapter.java:51)
         at weblogic.management.deploy.internal.DeploymentAdapter.activate(DeploymentAdapter.java:196)
         at weblogic.management.deploy.internal.AppTransition$2.transitionApp(AppTransition.java:30)
         at weblogic.management.deploy.internal.ConfiguredDeployments.transitionApps(ConfiguredDeployments.java:233)
         at weblogic.management.deploy.internal.ConfiguredDeployments.activate(ConfiguredDeployments.java:169)
         at weblogic.management.deploy.internal.ConfiguredDeployments.deploy(ConfiguredDeployments.java:123)
         at weblogic.management.deploy.internal.DeploymentServerService.resume(DeploymentServerService.java:173)
         at weblogic.management.deploy.internal.DeploymentServerService.start(DeploymentServerService.java:89)
         at weblogic.t3.srvr.SubsystemRequest.run(SubsystemRequest.java:64)
         at weblogic.work.ExecuteThread.execute(ExecuteThread.java:200)
         at weblogic.work.ExecuteThread.run(ExecuteThread.java:172)
    Caused by: (Wrapped: Failed to load the factory) java.lang.reflect.InvocationTargetException
         at com.tangosol.util.Base.ensureRuntimeException(Base.java:283)
         at com.tangosol.net.CacheFactory.getConfigurableCacheFactory(CacheFactory.java:610)
         at com.tangosol.net.CacheFactory.getCache(CacheFactory.java:689)
         at com.tangosol.net.CacheFactory.getCache(CacheFactory.java:667)
         at kodo.datacache.TangosolDataCache.newTangosolCache(TangosolDataCache.java:251)
         at kodo.datacache.TangosolDataCache.endConfiguration(TangosolDataCache.java:224)
         at org.apache.openjpa.lib.conf.Configurations.configureInstance(Configurations.java:450)
         at org.apache.openjpa.lib.conf.Configurations.configureInstance(Configurations.java:375)
         at org.apache.openjpa.lib.conf.PluginListValue.instantiate(PluginListValue.java:93)
         at org.apache.openjpa.lib.conf.ObjectValue.instantiate(ObjectValue.java:79)
         at kodo.datacache.KodoDataCacheManager.instantiateDataCaches(KodoDataCacheManager.java:77)
         at kodo.datacache.KodoDataCacheManager.initialize(KodoDataCacheManager.java:49)
         at org.apache.openjpa.conf.OpenJPAConfigurationImpl.getDataCacheManagerInstance(OpenJPAConfigurationImpl.java:588)
         at org.apache.openjpa.kernel.AbstractBrokerFactory.newBroker(AbstractBrokerFactory.java:169)
         ... 48 more

    Hi,
    I am using open jpa for persistence in my clustered weblogic server envirnment. I am using coherence for locking support and distributed cache mgmt.
    But while creating entity manager from the Persistence.xml file, i am getting an error, for which the base exception is thrown from the tangosol classes.
    Here is a complete stack trace of the error.
    [21 May 2008 18:33:36,650] ERROR [[ACTIVE] ExecuteThread: '0' for queue:
    'weblogic.kernel.Default (self-tuning)'] test.TestSessionBean:83 - Exception
    <1.0.0 nonfatal general error> org.apache.openjpa.persistence.PersistenceException: (Wrapped: Failed to load the factory) null
         at org.apache.openjpa.kernel.AbstractBrokerFactory.newBroker(AbstractBrokerFactory.java:193)
         at org.apache.openjpa.kernel.DelegatingBrokerFactory.newBroker(DelegatingBrokerFactory.java:142)
         at org.apache.openjpa.persistence.EntityManagerFactoryImpl.createEntityManager(EntityManagerFactoryImpl.java:190)
         at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:74)
         at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:36)
         at org.apache.openjpa.persistence.EntityManagerFactoryImpl.createEntityManager(EntityManagerFactoryImpl.java:143)
         at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:70)
         at kodo.persistence.KodoEntityManagerFactoryImpl.createEntityManager(KodoEntityManagerFactoryImpl.java:36)
         at com.git.test.TestSessionBean.<init>(TestSessionBean.java:49)
         at com.git.test.TestSessionBean_eb2d8g_Impl.<init>(TestSessionBean_eb2d8g_Impl.java:45)
         at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
         at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
         at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
         at java.lang.reflect.Constructor.newInstance(Constructor.java:494)
         at java.lang.Class.newInstance0(Class.java:350)
         at java.lang.Class.newInstance(Class.java:303)
         at weblogic.ejb.container.injection.EjbComponentCreatorImpl.getBean(EjbComponentCreatorImpl.java:73)
         at weblogic.ejb.container.manager.BaseEJBManager.createNewBeanInstance(BaseEJBManager.java:216)
         at weblogic.ejb.container.manager.BaseEJBManager.allocateBean(BaseEJBManager.java:233)
         at weblogic.ejb.container.manager.StatelessManager.createBean(StatelessManager.java:305)
         at weblogic.ejb.container.pool.StatelessSessionPool.createBean(StatelessSessionPool.java:173)
         at weblogic.ejb.container.pool.Pool.createInitialBeans(Pool.java:295)
         at weblogic.ejb.container.manager.StatelessManager.initializePool(StatelessManager.java:446)
         at weblogic.ejb.container.deployer.EJBDeployer.initializePools(EJBDeployer.java:1635)
         at weblogic.ejb.container.deployer.EJBDeployer.start(EJBDeployer.java:1464)
         at weblogic.ejb.container.deployer.EJBModule.start(EJBModule.java:513)
         at weblogic.application.internal.flow.ModuleListenerInvoker.start(ModuleListenerInvoker.java:117)
         at weblogic.application.internal.flow.ModuleStateDriver$3.next(ModuleStateDriver.java:204)
         at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:26)
         at weblogic.application.internal.flow.ModuleStateDriver.start(ModuleStateDriver.java:60)
         at weblogic.application.internal.flow.StartModulesFlow.activate(StartModulesFlow.java:26)
         at weblogic.application.internal.BaseDeployment$2.next(BaseDeployment.java:635)
         at weblogic.application.utils.StateMachineDriver.nextState(StateMachineDriver.java:26)
         at weblogic.application.internal.BaseDeployment.activate(BaseDeployment.java:212)
         at weblogic.application.internal.DeploymentStateChecker.activate(DeploymentStateChecker.java:154)
         at weblogic.deploy.internal.targetserver.AppContainerInvoker.activate(AppContainerInvoker.java:80)
         at weblogic.deploy.internal.targetserver.BasicDeployment.activate(BasicDeployment.java:182)
         at weblogic.deploy.internal.targetserver.BasicDeployment.activateFromServerLifecycle(BasicDeployment.java:359)
         at weblogic.management.deploy.internal.DeploymentAdapter$1.doActivate(DeploymentAdapter.java:51)
         at weblogic.management.deploy.internal.DeploymentAdapter.activate(DeploymentAdapter.java:196)
         at weblogic.management.deploy.internal.AppTransition$2.transitionApp(AppTransition.java:30)
         at weblogic.management.deploy.internal.ConfiguredDeployments.transitionApps(ConfiguredDeployments.java:233)
         at weblogic.management.deploy.internal.ConfiguredDeployments.activate(ConfiguredDeployments.java:169)
         at weblogic.management.deploy.internal.ConfiguredDeployments.deploy(ConfiguredDeployments.java:123)
         at weblogic.management.deploy.internal.DeploymentServerService.resume(DeploymentServerService.java:173)
         at weblogic.management.deploy.internal.DeploymentServerService.start(DeploymentServerService.java:89)
         at weblogic.t3.srvr.SubsystemRequest.run(SubsystemRequest.java:64)
         at weblogic.work.ExecuteThread.execute(ExecuteThread.java:200)
         at weblogic.work.ExecuteThread.run(ExecuteThread.java:172)
    Caused by: (Wrapped: Failed to load the factory) java.lang.reflect.InvocationTargetException
         at com.tangosol.util.Base.ensureRuntimeException(Base.java:283)
         at com.tangosol.net.CacheFactory.getConfigurableCacheFactory(CacheFactory.java:610)
         at com.tangosol.net.CacheFactory.getCache(CacheFactory.java:689)
         at com.tangosol.net.CacheFactory.getCache(CacheFactory.java:667)
         at kodo.datacache.TangosolDataCache.newTangosolCache(TangosolDataCache.java:251)
         at kodo.datacache.TangosolDataCache.endConfiguration(TangosolDataCache.java:224)
         at org.apache.openjpa.lib.conf.Configurations.configureInstance(Configurations.java:450)
         at org.apache.openjpa.lib.conf.Configurations.configureInstance(Configurations.java:375)
         at org.apache.openjpa.lib.conf.PluginListValue.instantiate(PluginListValue.java:93)
         at org.apache.openjpa.lib.conf.ObjectValue.instantiate(ObjectValue.java:79)
         at kodo.datacache.KodoDataCacheManager.instantiateDataCaches(KodoDataCacheManager.java:77)
         at kodo.datacache.KodoDataCacheManager.initialize(KodoDataCacheManager.java:49)
         at org.apache.openjpa.conf.OpenJPAConfigurationImpl.getDataCacheManagerInstance(OpenJPAConfigurationImpl.java:588)
         at org.apache.openjpa.kernel.AbstractBrokerFactory.newBroker(AbstractBrokerFactory.java:169)
         ... 48 more

  • Any good book on clustering

    Is there any book available for clustering, which would give details
              like how the beans are to be coded, what are the pitfalls, performance
              tuning, testing the failover etc etc..
              I am new to clustering and am wondering how to go about implementing
              this in our existing project (which does not have clustering currently).
              BEA site does not give much detail on the clustering aspect.
              Any help would be appreciated.
              Santosh
              

    Thanks A Lot,
    I am a beginner - not to Java but to Java 3D - I want to know what it has to offer me...something like - if I know Java 3D I can do...???
    Thanks
    --jj                                                                                                                                                                                                                                                                                                                                           

  • JRockit Java 5 Performance Issue

    Hi,
    I thought the JRockit team might be interested in this insight article that discusses the performance impact of GC and temporary object allocations in Sun and JRockits Exception and Thread (CallStack) API's. During the creation of the article we discovered a significant regression in the performance of the JRockit JVM for exceptions and thread callstacks.
    http://www.jinspired.com/products/jdbinsight/callstackbenchmark.html
    Regards,
    William Louth
    JXInsight Product Arhcitect
    CTO, JInspired
    "J*EE tuning, testing and tracing with JXInsight"
    http://www.jinspired.com

    Hi,
    I think a clue is the plural form of the method signature: jrockit/vm/Threads.dumpThreads
    Regards,
    William
    at jrockit/management/jvmpi/Tracker.nativeAllocSite(Ljava/lang/Object;)V(Native Method)
    at jrockit/management/jvmpi/Tracker.allocSite(Ljava/lang/Object;)V(Unknown Source)[inlined]
    at java/nio/CharBuffer.wrap([CII)Ljava/nio/CharBuffer;(Unknown Source)[inlined]
    at java/nio/CharBuffer.wrap([C)Ljava/nio/CharBuffer;(Unknown Source)[inlined]
    at java/lang/StringCoding$CharsetSD.decode([BII)[C(Unknown Source)[optimized]
    at java/lang/StringCoding.decode(Ljava/lang/String;[BII)[C(Unknown Source)
        at java/lang/String.<init>([BIILjava/lang/String;)V(Unknown Source)[inlined]
    at java/lang/String.<init>([BLjava/lang/String;)V(Unknown Source)[inlined]
    at jrockit/vm/Memory$DumpReader.readIString(Z)Ljava/lang/String;(Unknown Source)[inlined]
    at jrockit/vm/Threads.dumpThreads(Ljrockit/vm/Threads$ThreadDumpCallback;[Ljava/lang/Thread;IZ)V(Unknown Source)[optimized]
    at jrockit/vm/Threads.dumpThreads(Ljrockit/vm/Threads$ThreadDumpCallback;[Ljava/lang/Thread;Z)V(Unknown Source)
        at jrockit/vm/Threads.dumpThreads([Ljava/lang/Thread;)[[Ljava/lang/StackTraceElement;(Unknown Source)
        at java/lang/Thread.getStackTrace()[Ljava/lang/StackTraceElement;(Unknown Source)
        at Benchmark.execute()V(Benchmark.java:48)
        at Benchmark.call(I)V(Benchmark.java:53)
        at Benchmark.call(I)V(Benchmark.java:55)
        at Benchmark.call(I)V(Benchmark.java:55)
        at Benchmark.call(I)V(Benchmark.java:55)
        at Benchmark.call(I)V(Benchmark.java:55)                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

  • Option ReTest Sequence Failure

    Hi All,
    I'm quite new in test stand and I need some help on my test program.
    I have a local sequence created on my main sequence. I shall call this sequence as "tuning".
    This "tuning" sequence actualy tests if the unit is tuned properly.
    In most cases the first tuning always fail. So I want to give them the option to retune it and redo the sequence again.
    I can do a pass fail count on the looping properties for example "Stop after 1 pass or a maximum of 10 Iterations" in which to my understanding
    is the operator can tune it upto 10 ten times before it could be considered a fail unit.
    My problem is that sometimes after one tuning test the operator can tell right away that it is a helpless case and it can not be tuned. So i dont want them
    to wait until ten times before failing a unit since the sequence cycle time is quite long.
    I want to give them an option to fail the unit right away if they think its a failure maybe after trying 4-5 times.
    This where my question goes::: how do i approach it?
    Advise will be gladly appreciated.
    Best Regards,Ray

    This is what I have done.. I don't know where I'm getting wrong..
    Attachments:
    11.seq ‏9 KB

  • Managing Coherence via JMX for web and non web app

    I read your posting on,
    Q: Where can I get an advanced example of managing Coherence via JMX?
    and downloaded the jar file.
    It works for the web app but not for the server application.
    My configuratiion is two weblogic server with two tangosol nodes and two independent tangosol server nodes running in a cluster. That is 4 tangosol cluster are running. The jar file allows me to see only one node that is running on the web logic server. How can I view the other 3 nodes?
    Thanks.

    Hi,
    JInspired recently released a special update to JXInsight which includes support for Tangosol's Coherence product. The release provides 3 important features for Coherence customers:
    1. Coherence JMX-to-JXInsight metric mapping
    2. Coherence JVMInsight extension (remote inspection of cache runtime state)
    3. Coherence Cache trace extension
    The following links provide additional information entries:
    http://blog.jinspired.com/?p=13
    http://blog.jinspired.com/?p=12
    http://forums.jinspired.com/viewtopic.php?t=8
    One large benefit to using JXInsight for monitoring Coherence is that you could cache associate activity with other layers in your application stack (JSP, EJB, JMS, JCA, Hibernate, JDO,...) including contextual processing steps.
    Within our installation root directory there is a sample/coherence directory containing a sample script to test with the Coherence product.
    Kind regards,
    William Louth
    CTO, JInspired
    "Java EE tuning, testing, tracing, and monitoring with JXInsight"
    http://www.jinspired.com

  • Stress test tool and guidelines for tuning ADF 11g

    Hi all,
    I am working our ADF 11g application on load & stress testing. The load test works fine with some millions of full dummy rows (this generated data tool is easily written by ourself). The tunning only was done with only indexing. But with stress I'm concerning with RAM and other tunning parameters such IO and share pool from OS to DB and to WL servers and how to tune it for around 200-300 concurrent connections. I'm studying Jmeter for this purpose but still not get the answer. I have tested with Oracle Openscript but it worked very heavily slow :(.
    The question here is that do you have any experiences on tuning parameters for such requirements and what is the best automatic test tool for testing multi-connections, please advise me.
    Thank you very much.

    Oracle Application Testing Suite has specific support for ADF testing:
    http://www.oracle.com/technetwork/oem/app-test/index.html
    http://www.testingreflections.com/node/view/8538
    Some info on JMeter: http://www.connotea.org/user/jdeveloper/tag/jmeter
    Another option: http://www.yenlo.nl/harryvanoosten/2010/03/10/load-testing-an-adf-11g-application/
    Some entries about tuning: http://www.connotea.org/user/jdeveloper/tag/tuning

  • A SQL tuning issue-sql runs much slower in test than in production?

    Hi Buddies,
    I am working on a sql tuning issue. A sql runs much slower in test than in production.
    I compared the two explain plans in test and production
    seems in test, CBO refuses to use index SUBLEDGER_ENTRY_I2.
    we rebuile it and re-gether that index statistcs. run, still slow..
    I compared the init.ora parameters like hash_area_size, sort_area_size in test, they are same as production.
    I wonder if any expert friend can show some light.
    in production,
    SQL> set autotrace traceonly
    SQL> SELECT rpt_horizon_subledger_entry_vw.onst_offst_cd,
    2 rpt_horizon_subledger_entry_vw.bkng_prd,
    3 rpt_horizon_subledger_entry_vw.systm_afflt_cd,
    4 rpt_horizon_subledger_entry_vw.jrnl_id,
    5 rpt_horizon_subledger_entry_vw.ntrl_accnt_cd,
    6 rpt_horizon_subledger_entry_vw.gnrl_ldgr_chrt_of_accnt_nm,
    7 rpt_horizon_subledger_entry_vw.lgl_entty_brnch_cd,
    8 rpt_horizon_subledger_entry_vw.crprt_melob_cd AS corp_mlb_cd,
    rpt_horizon_subledger_entry_vw.onst_offst_cd, SUM (amt) AS amount
    9 10 FROM rpt_horizon_subledger_entry_vw
    11 WHERE rpt_horizon_subledger_entry_vw.bkng_prd = '092008'
    12 AND rpt_horizon_subledger_entry_vw.jrnl_id = 'RCS0002100'
    13 AND rpt_horizon_subledger_entry_vw.systm_afflt_cd = 'SAFF01'
    14 GROUP BY rpt_horizon_subledger_entry_vw.onst_offst_cd,
    15 rpt_horizon_subledger_entry_vw.bkng_prd,
    16 rpt_horizon_subledger_entry_vw.systm_afflt_cd,
    17 rpt_horizon_subledger_entry_vw.jrnl_id,
    18 rpt_horizon_subledger_entry_vw.ntrl_accnt_cd,
    19 rpt_horizon_subledger_entry_vw.gnrl_ldgr_chrt_of_accnt_nm,
    20 rpt_horizon_subledger_entry_vw.lgl_entty_brnch_cd,
    21 rpt_horizon_subledger_entry_vw.crprt_melob_cd,
    22 rpt_horizon_subledger_entry_vw.onst_offst_cd;
    491 rows selected.
    Execution Plan
    0 SELECT STATEMENT Optimizer=CHOOSE (Cost=130605 Card=218764 B
    ytes=16407300)
    1 0 SORT (GROUP BY) (Cost=130605 Card=218764 Bytes=16407300)
    2 1 VIEW OF 'RPT_HORIZON_SUBLEDGER_ENTRY_VW' (Cost=129217 Ca
    rd=218764 Bytes=16407300)
    3 2 SORT (UNIQUE) (Cost=129217 Card=218764 Bytes=35877296)
    4 3 UNION-ALL
    5 4 HASH JOIN (Cost=61901 Card=109382 Bytes=17719884)
    6 5 TABLE ACCESS (FULL) OF 'GNRL_LDGR_CHRT_OF_ACCNT'
    (Cost=2 Card=111 Bytes=3774)
    7 5 HASH JOIN (Cost=61897 Card=109382 Bytes=14000896
    8 7 TABLE ACCESS (FULL) OF 'SUBLEDGER_CHART_OF_ACC
    OUNT' (Cost=2 Card=57 Bytes=1881)
    9 7 HASH JOIN (Cost=61893 Card=109382 Bytes=103912
    90)
    10 9 TABLE ACCESS (FULL) OF 'HORIZON_LINE' (Cost=
    34 Card=4282 Bytes=132742)
    11 9 HASH JOIN (Cost=61833 Card=109390 Bytes=7000
    960)
    12 11 TABLE ACCESS (BY INDEX ROWID) OF 'SUBLEDGE
    R_ENTRY' (Cost=42958 Card=82076 Bytes=3611344)
    13 12 INDEX (RANGE SCAN) OF 'SUBLEDGER_ENTRY_I
    2' (NON-UNIQUE) (Cost=1069 Card=328303)
    14 11 TABLE ACCESS (FULL) OF 'HORIZON_SUBLEDGER_
    LINK' (Cost=14314 Card=9235474 Bytes=184709480)
    15 4 HASH JOIN (Cost=61907 Card=109382 Bytes=18157412)
    16 15 TABLE ACCESS (FULL) OF 'GNRL_LDGR_CHRT_OF_ACCNT'
    (Cost=2 Card=111 Bytes=3774)
    17 15 HASH JOIN (Cost=61903 Card=109382 Bytes=14438424
    18 17 TABLE ACCESS (FULL) OF 'SUBLEDGER_CHART_OF_ACC
    OUNT' (Cost=2 Card=57 Bytes=1881)
    19 17 HASH JOIN (Cost=61899 Card=109382 Bytes=108288
    18)
    20 19 TABLE ACCESS (FULL) OF 'HORIZON_LINE' (Cost=
    34 Card=4282 Bytes=132742)
    21 19 HASH JOIN (Cost=61838 Card=109390 Bytes=7438
    520)
    22 21 TABLE ACCESS (BY INDEX ROWID) OF 'SUBLEDGE
    R_ENTRY' (Cost=42958 Card=82076 Bytes=3939648)
    23 22 INDEX (RANGE SCAN) OF 'SUBLEDGER_ENTRY_I
    2' (NON-UNIQUE) (Cost=1069 Card=328303)
    24 21 TABLE ACCESS (FULL) OF 'HORIZON_SUBLEDGER_
    LINK' (Cost=14314 Card=9235474 Bytes=184709480)
    Statistics
    25 recursive calls
    18 db block gets
    343266 consistent gets
    370353 physical reads
    0 redo size
    15051 bytes sent via SQL*Net to client
    1007 bytes received via SQL*Net from client
    34 SQL*Net roundtrips to/from client
    1 sorts (memory)
    1 sorts (disk)
    491 rows processed
    in test
    SQL> set autotrace traceonly
    SQL> SELECT rpt_horizon_subledger_entry_vw.onst_offst_cd,
    2 rpt_horizon_subledger_entry_vw.bkng_prd,
    3 rpt_horizon_subledger_entry_vw.systm_afflt_cd,
    4 rpt_horizon_subledger_entry_vw.jrnl_id,
    5 rpt_horizon_subledger_entry_vw.ntrl_accnt_cd,
    rpt_horizon_subledger_entry_vw.gnrl_ldgr_chrt_of_accnt_nm,
    6 7 rpt_horizon_subledger_entry_vw.lgl_entty_brnch_cd,
    8 rpt_horizon_subledger_entry_vw.crprt_melob_cd AS corp_mlb_cd,
    9 rpt_horizon_subledger_entry_vw.onst_offst_cd, SUM (amt) AS amount
    10 FROM rpt_horizon_subledger_entry_vw
    11 WHERE rpt_horizon_subledger_entry_vw.bkng_prd = '092008'
    12 AND rpt_horizon_subledger_entry_vw.jrnl_id = 'RCS0002100'
    AND rpt_horizon_subledger_entry_vw.systm_afflt_cd = 'SAFF01'
    13 14 GROUP BY rpt_horizon_subledger_entry_vw.onst_offst_cd,
    15 rpt_horizon_subledger_entry_vw.bkng_prd,
    16 rpt_horizon_subledger_entry_vw.systm_afflt_cd,
    17 rpt_horizon_subledger_entry_vw.jrnl_id,
    18 rpt_horizon_subledger_entry_vw.ntrl_accnt_cd,
    rpt_horizon_subledger_entry_vw.gnrl_ldgr_chrt_of_accnt_nm,
    rpt_horizon_subledger_entry_vw.lgl_entty_brnch_cd,
    rpt_horizon_subledger_entry_vw.crprt_melob_cd,
    rpt_horizon_subledger_entry_vw.onst_offst_cd; 19 20 21 22
    no rows selected
    Execution Plan
    0 SELECT STATEMENT Optimizer=CHOOSE (Cost=92944 Card=708 Bytes
    =53100)
    1 0 SORT (GROUP BY) (Cost=92944 Card=708 Bytes=53100)
    2 1 VIEW OF 'RPT_HORIZON_SUBLEDGER_ENTRY_VW' (Cost=92937 Car
    d=708 Bytes=53100)
    3 2 SORT (UNIQUE) (Cost=92937 Card=708 Bytes=124962)
    4 3 UNION-ALL
    5 4 HASH JOIN (Cost=46456 Card=354 Bytes=60180)
    6 5 TABLE ACCESS (FULL) OF 'SUBLEDGER_CHART_OF_ACCOU
    NT' (Cost=2 Card=57 Bytes=1881)
    7 5 NESTED LOOPS (Cost=46453 Card=354 Bytes=48498)
    8 7 HASH JOIN (Cost=11065 Card=17694 Bytes=1362438
    9 8 HASH JOIN (Cost=27 Card=87 Bytes=5133)
    10 9 TABLE ACCESS (FULL) OF 'HORIZON_LINE' (Cos
    t=24 Card=87 Bytes=2175)
    11 9 TABLE ACCESS (FULL) OF 'GNRL_LDGR_CHRT_OF_
    ACCNT' (Cost=2 Card=111 Bytes=3774)
    12 8 TABLE ACCESS (FULL) OF 'HORIZON_SUBLEDGER_LI
    NK' (Cost=11037 Card=142561 Bytes=2566098)
    13 7 TABLE ACCESS (BY INDEX ROWID) OF 'SUBLEDGER_EN
    TRY' (Cost=2 Card=1 Bytes=60)
    14 13 INDEX (UNIQUE SCAN) OF 'SUBLEDGER_ENTRY_PK'
    (UNIQUE) (Cost=1 Card=1)
    15 4 HASH JOIN (Cost=46456 Card=354 Bytes=64782)
    16 15 TABLE ACCESS (FULL) OF 'SUBLEDGER_CHART_OF_ACCOU
    NT' (Cost=2 Card=57 Bytes=1881)
    17 15 NESTED LOOPS (Cost=46453 Card=354 Bytes=53100)
    18 17 HASH JOIN (Cost=11065 Card=17694 Bytes=1362438
    19 18 HASH JOIN (Cost=27 Card=87 Bytes=5133)
    20 19 TABLE ACCESS (FULL) OF 'HORIZON_LINE' (Cos
    t=24 Card=87 Bytes=2175)
    21 19 TABLE ACCESS (FULL) OF 'GNRL_LDGR_CHRT_OF_
    ACCNT' (Cost=2 Card=111 Bytes=3774)
    22 18 TABLE ACCESS (FULL) OF 'HORIZON_SUBLEDGER_LI
    NK' (Cost=11037 Card=142561 Bytes=2566098)
    23 17 TABLE ACCESS (BY INDEX ROWID) OF 'SUBLEDGER_EN
    TRY' (Cost=2 Card=1 Bytes=73)
    24 23 INDEX (UNIQUE SCAN) OF 'SUBLEDGER_ENTRY_PK'
    (UNIQUE) (Cost=1 Card=1)
    Statistics
    1134 recursive calls
    0 db block gets
    38903505 consistent gets
    598254 physical reads
    60 redo size
    901 bytes sent via SQL*Net to client
    461 bytes received via SQL*Net from client
    1 SQL*Net roundtrips to/from client
    34 sorts (memory)
    0 sorts (disk)
    0 rows processed
    Thanks a lot in advance
    Jerry

    Hi
    Basically there are two kinds of tables
    - fact
    - lookup
    The number of records in a lookup table is usually small.
    The number of records in a fact table is usually huge.
    However, in test systems the number of records in a fact table is often also small.
    This results in different execution plans.
    I notice again you don't post version and platform info, and you didn't make sure your explain is properly idented
    Please read the FAQ to make sure it is properly idented.
    Also using the word 'buddies' is as far as I am concerned nearing disrespect and rudeness.
    Sybrand Bakker
    Senior Oracle DBA

  • When will Hub automatically select fewer than 2500 segments for tuning and/or testing?

    I have two systems (FR/IT) with 64K+ total segments uploaded.
    When I specify "Auto" for tuning and testing, the system only extracts 500 (or some other number < 2500) tuning and testing segments from the TMX files that I uploaded. On visual inspection of the subsequently downloaded training, tuning,
    and test corpora, I see nothing unusual:  alignment is OK, repeated and partial segments are not very frequent, etc.
    Similar content extracted in the same way from the same WorldServer db for other languages (PE, ES, NL, ID) works fine.
    When will Hub automatically select fewer than 2500 segments for tuning and/or testing?
    Thanks!

    Hi Mike,
    Here is how auto-selection of Testing and Tuning data works in the Hub.
    1. If a segment in tuning data is identical to a segment in training data, the segment is removed from training data.
    2. If a segment in tuning data is identical to a segment in testing data, the segment is removed from testing data.
    3. If a segment in tuning data and testing data is identical to a segment in training data, the segment is removed from training and testing data.
    4. If a segment in testing data is identical to a segment in training data, the segment is removed from training data.
    Hub auto selects 2500 sentences and the data under goes above adjustments. As a result you will see fewer than 2500 sentences.
    Additional details can be found in hub user guide:https://hub.microsofttranslator.com/Help/Download/Microsoft%20Translator%20Hub%20User%20Guide.pdf
    Please let us know if you need any additional information.
    Thank you

  • Beta Testing Continues for "Oracle Database 11g: Performance Tuning" Exam

    !http://blogs.oracle.com/certification/2009-0204-1123.gif!
    The beta period continues for the <strong>Oracle Database 11g: Performance Tuning certification exam (1Z1-054)</strong>, which is a single exam requirement for Oracle 11g DBA OCPs to earn the Oracle Database 11g Performance Tuning Certified Expert (OCE) certification. Read more on the <a href="http://bit.ly/UGJh4">Oracle Certification Blog</a>.</p>

    Hello Hussein
    Yes true, I remember it for the OCE and Linux exams they rescheduled the end date several times. As far as I know it is related to the number of participants and the given feedback.
    I've also participated to several other exams, and I must admit that it is a long and hard process to get through. When I got the feedback 10 weeks after the beta period closure, I had to review nearly all the topics to get the exams passed the second time. But this it is a cheap and good exam preparation.
    What about you Hussein? Do you think that's trivial?
    Cheers,
    Hub

  • Beta Testing Ends For "Oracle Database 11g: Performance Tuning" Exam

    !http://blogs.oracle.com/certification/2009-0204-1122.gif!
    <p>The beta period ended on Aug 31, 2009 for the <strong>Oracle Database 11g: Performance Tuning</strong> certification exam (1Z1-054), which is a single exam requirement for Oracle 11g DBA OCPs to earn the Oracle Database 11g Performance Tuning Certified Expert (OCE) certification.</p>
    <p align="justify">Full details are posted on the Oracle Certification Blog.</p>

    Hello Hussein
    Yes true, I remember it for the OCE and Linux exams they rescheduled the end date several times. As far as I know it is related to the number of participants and the given feedback.
    I've also participated to several other exams, and I must admit that it is a long and hard process to get through. When I got the feedback 10 weeks after the beta period closure, I had to review nearly all the topics to get the exams passed the second time. But this it is a cheap and good exam preparation.
    What about you Hussein? Do you think that's trivial?
    Cheers,
    Hub

Maybe you are looking for

  • I do not have a delete option for files or folders in Adobe Creative Cloud

    I do not have a delete option for files or folders in Adobe Creative Cloud im looking and looking.... im stumped..... 4 weeks now.... kai

  • Is this possible.........? connect to rtmp from XML

    AS3 Code var myXML:XML; var myLoader:URLLoader = new URLLoader(); myLoader.load(new URLRequest("xmlfile.xml")); myLoader.addEventListener(Event.COMPLETE, processXML); function processXML(e:Event):void           myXML = new XML(e.target.data);        

  • Problems with the side volume buttons

    I can no longer control the ringer volume with the side buttons.  The buttons will work if the headphones are plugged in.  Any ideas?

  • Erratic behavior

    I am having a lot of problems with scripts from a book tutorial. The book is PHP 5 Advanced by Larry Ullman. I've worked through two of his other books with no problems. But with this book I am getting a lot of errors and erratic behavior. With a hea

  • Default Data Is Executed When in the GTS Interface

    Hello, I just want to confirm with the group of when the Default Data is executed in the GTS Interface.  When the interface is running on the GTS-side, there are several user exits (BaDIs) that are available. Does anyone know if the Default Data is e