LARGE POOL and RMAN

Hi,
Because I use RMAN I decided to configure Large Pool. It is said, that RMAN should use this area if it's defined. But im my system statistics for LargePool are constant - there is always only empty space, no blocks are allocated by RMAN. Could someone tell me what's wrong? Why my RMAN still uses shared pool and no large?
Regards,
Kuba

that was an interesting question, after checking in metalink this seems to be myth that if you configure large_pool rman will use it. (the problem with oracle documentation is it conveniently forgets to qualify the conditions).
"The LARGE_POOL is only used if I/O Slaves are specified by one or both of the initialization parameters DBWR_IO_SLAVES or BACKUP_TAPE_IO_SLAVES."
so in your case may be you have not using these parameters, that maybe the reason why RMAN is not using Large_pool.
check metalink note : 134214.1 for more RMAN myths.
which leaves us wondering...why can't these information be made available in the public documentation in the form of short notes or caveats???

Similar Messages

  • Large pool and AQ

    Hello All,
    I'm using Oracle 10.2.0.3 under Linux IA 32Bits.
    I've got a big problem with my large pool that keep increasing as I'm using Advance queuing. After a while, errors 4031 are raised.
    Can someone explain how to monitor large pool usage. Espaecially I would like to know wich process/ session consumes large pool.
    Thanks in advance for your help

    that was an interesting question, after checking in metalink this seems to be myth that if you configure large_pool rman will use it. (the problem with oracle documentation is it conveniently forgets to qualify the conditions).
    "The LARGE_POOL is only used if I/O Slaves are specified by one or both of the initialization parameters DBWR_IO_SLAVES or BACKUP_TAPE_IO_SLAVES."
    so in your case may be you have not using these parameters, that maybe the reason why RMAN is not using Large_pool.
    check metalink note : 134214.1 for more RMAN myths.
    which leaves us wondering...why can't these information be made available in the public documentation in the form of short notes or caveats???

  • Large pool usage and advanced queueing

    Hello All,
    I'm using Oracle 10.2.0.3 under Linux IA 32Bits.
    I've got a big problem with my large pool that keep increasing as I'm using Advance queuing. After a while, errors 4031 are raised.
    Can someone explain how to monitor large pool usage. Espaecially I would like to know wich process/ session consumes large pool.
    Thanks in advance for your help

    Whilst Daniel does his 'Twenty Questions' trick on you (he has a valid point: OS and Oracle Version are ALWAYS useful to know... but there's no need to pin it down to four decimal places for such a simple question!), I'll have a go at actually providing an answer.
    A number of things might be going on.
    First, do a google search on "joxs heap init" and you'll find suggestions that setting or increasing the JAVA_POOL_SIZE can fix the problem. See this for example, which has exactly your problem:
    http://www.mail-archive.com/[email protected]/msg36154.html
    Second, a lot of things that would use the Large Pool if it exists can instead make use of the shared pool if it doesn't -but they fragment the thing mightily as they do (the Large Pool was invented precisely to provide a chunk of memory that doesn't age out, only gets re-used and thus never fragments). So something (an RMAN backup, for example) might work without a large pool, if there's room in the shared pool; but if the shared pool is already fragmented or tight, then it will error with a 4031. Allocate it a bit of Large Pool, however, and it prefers to use that, finds the memory it needs is now available and therefore works fine.
    The large pool is used for RMAN, shared server, I/O slaves and parallel slaves. Who knows precisely what's going on in an export that might be touching on at least 2 of those things without you explicitly asking for it to do so?! I mean that it's possible it's doing a bit of 'internal parallelism' to make itself efficient, for example, and failing to find the memory it needs for inter-parallel-slave communications. Just a possibility and a bit of speculation on my part.
    I suppose the next experiment would be: take Large Pool back down to zero and then set Java Pool to 15M instead (note you should really be allocating in units of 4M these days, though Oracle will round up anyway).

  • Alter system flushed shared pool in RMAN backup

    Hi,
    I am trying to take RMAN backup of 11.2.0.1 Database in IBM AIX 6.1 server.
    The RMAN is hanging .
    Though the backup gets completed, The channels allocated doesnt get released and the RMAN gets hanging.
    In earlier RMAN backup Scripts,
    the DBA was using alter system flush shared pool in RMAN backup script and the backup was getting succesful.
    Now my question is , is using ALTER SYSTEM FLUSH SHARED POOL have any performance impact on the database.
    Regards,
    TEJAS

    TEJAS_DBA wrote:
    Hi,
    I am trying to take RMAN backup of 11.2.0.1 Database in IBM AIX 6.1 server.
    The RMAN is hanging .
    Though the backup gets completed, The channels allocated doesnt get released and the RMAN gets hanging.Are you setting the large pool? If you don't, then rman uses the shared pool. Read about tuning rman performance in the docs.
    >
    In earlier RMAN backup Scripts,
    the DBA was using alter system flush shared pool in RMAN backup script and the backup was getting succesful.
    Now my question is , is using ALTER SYSTEM FLUSH SHARED POOL have any performance impact on the database.Yes, you are allowing the components in there to be loaded in the random order of whatever is called first. This may have a good impact if you had some fragmentation in there, or it could be mildly bad if everything was well sorted, or it could be very bad if you are unlucky or have some pattern of invalidations or should be pinning something or who-knows-what. It generally is considered not a good thing to do as a habit. You wind up with [url http://tkyte.blogspot.com/2012/05/another-debugging-story.html]rainy Monday scenarios.
    Edit: I notice there are some bugs, including very slow performance when using a catalog. When you say hang, how long are you waiting? Have you considered current patches?
    Edited by: jgarry on Aug 8, 2012 11:09 AM

  • Session pooling and statement handles

    Hi there,
    I have a large multi-threaded application (perhaps >100 threads). Each thread is continuously processing events (very high volumes) which involves some manipulation and some database operations (from a fixed set of possible operations).
    I am using session pooling but what I want to know is, Should I:
    (a) Prepare my fixed set of statement handles up front at program start-up when I'm creating the session pool and then reuse the statement handles in each event processing thread (also, is this thread safe ? even if it is, all threads would be contending on the same statement handles)
    or
    (b) Prepare the statement handle for each event which presumably will exploit the statement cache on the session pool. This would also mean not having statement handles shared between threads thus removing any thread contention issues.
    I think (b) is the option for me, but does anyone have any thoughts ?

    With a), one would think it's OK, but I would hate to find out that it's not thread safe by accident.
    But anyway, with b) the cost of allocating private statement handles in each thread seems very low. The memory required for the statement handle plus its bind and define handles could very well be below 8k per statement. If you've got say, 5 statements * 100 threads, you're only looking at around 4MB overall.
    Finally, you might want to make sure that the session pool statement cache is working by checking the values for 'executions' and 'parse_calls' in V$SQL for your statements.

  • Calculating Large Pool value

    Hi,
    I need to determine the 'large pool' size to be implemented for using RMAN. (We already have RMAN running). I need to know what parameters I need to set in init.ora for large_pool to work efficiently in conjunction with RMAN. How do I calculate the required size. I found some material on this, here, but it's confusing.
    By default LARGE POOL = 0, unless PARALLEL_AUTOMATIC_TUNING = TRUE, (in which case LP will be set automatically). Also, set LARGE_POOL_MIN_ALLOC to a non-zero value to enable a minimum allocation with LP.
    When I read document (ID 73079.1) it gets more confusing, as whether to take MULTIBLOCK-READ-COUNT value or DB_FILE_DIRECT_IO_COUNT into consideration. Also, what is the number of datafiles, (is it the same as db_files in init.ora ??)
    Thanks in advance, Sreedhar
    [email protected]
    null

    Hi ...
    Take care about use LARGE_POOL if your database also use MTS . All MTS services will be deployed at the large_pool.
    Just consider it ...
    Regards,
    Jose Ramon

  • Where we can use pooled and cluster tables.

    Hi Experts,
    I have read all the threads and tutorials. all are telling like the difference between pooled and clulster tables are
    Pooled table
    It can hold small amount of data ( 10 to 100 records) and large number of tables( 100 to 1000 tables). It has fixed structure like Tabname, Varkey, Datain, Vardata.
    Cluter tables
    It can hold large amount of data but small number of tables. It is also have fixed strucuter like CLKEY1,CLKEY2 , CLKEY3...., Pageno, Timestamp, Pagelg, Vardata.
    But i want which situation we can choose pooled table and which situation we can choose cluster table. In general way or Business way.
    Thank you,
    Srinivas M
    Edited by: Srinivas Marapureddy on May 11, 2010 4:39 PM

    >
    El Tony wrote:
    > Those table to manage big data volume, for example BSEG is a cluster table, this help to database for store the data in little amount. Instead the Pooled Table to store the table in other table, so that this help to database for store the data too but in more little amount.
    >
    > In Background ABAP to control this, because there are source code that help to show the data complete, but in the databse the data is store with alphanumerics. This is a advantage for the database.
    Let me correct you, the data is not stored in alphanumeric but in RAW format in pool & cluster tables.
    SAP says pool & cluster tables are created so that data from different tables belonging containing similar kind of data are clubbed together. I am not sure if storing data in RAW format compresses its size but SAP recommends storing business data in transperant tables.
    Quote from SAP documentation:
    "Data of commercial relevance is usually stored in transparent tables."
    If data compression were a significant point SAP would have stored data in table clusters & not in transperant tables.
    Comments are welcome.
    BR,
    Suhas

  • Large pool processess

    hai
    For what type of I/O server processess the large pool is being used.

    Hi,
    On documentation, the large pool allocation heap is used in shared server systems for session memory, by parallel execution for message buffers, and by backup processes for disk I/O buffers.
    http://download-east.oracle.com/docs/cd/B19306_01/server.102/b14237/initparams090.htm#sthref379
    Cheers

  • How do I create my own favorite template for DVD slideshows? I used to be able to select this from pulldown menu, but cannot now do so. I am directed straight to templates, which take more memory. I have a large slideshow, and need all the space I can get

    First, how do I create my own favorite theme template for DVD slideshows? I used to be able to select this from pulldown menu, but cannot now do so. I am directed straight to already existing themes, which take more memory. I have a large slideshow, and need all the space I can get. I just want to use a picture as my DVD cover, and then insert a slideshow. Also, when I try to burn my 8.5gb double sided slideshow, all that burns is the music. It is a large slideshow, a memorial on the life of my now deceased brother. This means a lot to me and to my family, and I am having so much trouble trying to burn it. I have gone into Project View and selected appropriately. The bar shows I have room to burn this DVD, but it does not burn.  I have burned so many DVDs in the past, but this one just will not burn. I am so confused at this point. I will say this is the first 8.5gb I have attempted to create and burn. My specs list a 7.7gb or 4.7gb as operable....but there are no 7.7gb dvds. I had to purchase 8.5gb. Help? What am I doing wrong? I have spent so much time on this, and just cannot figure it out.

    Final Cut is a separate, higher end video editor.  The pro version of iMovie.
    Give iPhoto a look at for creating the slideshow.  It's easy to assemble the photos in an album in iPhoto, put them in the order you want and then make a slideshow of them.  You can select from various themes and transitions between slides and add music from your iTunes library.
    When you have the slidshow as you want use the Export button at the bottom of the iPhoto window and export with Size = Medium or Large.
    Save the resulting Quicktime movie file in your Movies folder.
    Next, open iDVD, choose your theme and drag the QT movie file into the menu window being careful to avoid any drop zones.
    Then follow this workflow to help assure the best qualty video DVD:
    Once you have the project as you want it save it as a disk image via the File ➙ Save as Disk Image  menu option. This will separate the encoding process from the burn process. 
    To check the encoding mount the disk image, launch DVD Player and play it.  If it plays OK with DVD Player the encoding is good.
    Then burn to disk with Disk Utility or Toast at the slowest speed available (2x-4x) to assure the best burn quality.  Always use top quality media:  Verbatim, Maxell or Taiyo Yuden DVD-R are the most recommended in these forums.
    The reason I suggest iPhoto is that I find it much easier to use than iMovie (except for the older iMovie 6 HD version).  Personal preferences showing here.

  • Difference between using app server connection pooling and using the driver

    Hi all,
    How to get connection pooling with out application server and tomcat also?
    What is the difference between using app server connection pooling and using the driver supported connection pooling?
    Regards,
    Murali

    maybe the performance of App server pool is better than the JDBC pool,
    for you don't know wether the implementation of the JDBC interface is good or bad.

  • Connection pool and Connection factory difference?

    Hi,
    Can someone clarify the difference between connection pool and connection factory.
    I know connection pool is some thing like a pool of database connection manitained at one place so that an application or client can use it when it requires.
    Then wat exactly is connection factory..say in 10g we always put a connection factory name in all adapters and acces the data sources like DB ,FTP,AQ ..etc
    can someone clarify??
    regards,

    To conserve system resources and to improve the performance of transactional applications, WebLogic allows you to define a pool of client connections (generally database but may be a FTP,FILE etc. as well)
    A Connection Factory object encapsulates a set of connection configuration parameters that has been defined by an administrator. A client uses it to create a connection with a service provider (generally JMS).
    Regards,
    Anuj

  • Difference between connection pooling and simple connection

    Anybody please tell me what is the Difference between connection pooling and simple connection and also where we define connection pooling and how.
    Thanks
    Please reply soon
    amitindia

    Creating and closing connection to the database is a relatively slow process. Equally connections use database resources so you can't just open as many as you want.
    A connection pool maintains a number of open connections throughout the lifetime of the application. Instead of opening and closing the connections your application just "borrows" them from the pool when they're needed.
    If the pool runs out of connections it will usually create new ones as needed until you reach some predefined upper limit.
    A good connection pool will also manage connections which have failed for any reason, and report code which fails to return connections to the pool (ie connection leaks).
    The number of connections created at startup is referred to as the "low water mark" and the maximum number that the pool will allow to be opened at any given time is the "high water mark". If no connections are available client code will generally block until one is released.
    If by "defined" you mean where can you get a working implementation then your database or application server vendor is likely to provide one. Also the Apache Jakarta Commons includes [url http://jakarta.apache.org/commons/dbcp/]an implementation which you can use with any JDBC driver.
    Dave.

  • Is connection pooling and sharing available on Oracle 9i RDBMS ?

    Hello,
    I would like to connect from oracle to sql server through db link and ODBC (Heterogenous connectivity). But every session in oracle launch session in sql server. Is it possible to have connection pooling and sharing from Oracle RDBMS level ? I need one solution : when (for example) i run 100 sessions in Oracle i would like to see 10 sessions on sql server. I would like remain 90 sessions from oracle to be queued.
    I would mention that i was using Heterogenous connectivity with multithreading agent (agtctl) without success.
    appreciate any help :-)

    There are two concepts you could evaluate, but they are mainly used for connections to the database, not to sqlserver, but if you can make them work with heterogeneous connectivity this could help:
    Connection Pooling. When many sessions are connecting to the same database, it could be that some of them remain idle. Oracle can detect them and timed them out letting another session to enter into the database, letting the idle session remain open without closing its session. This is configured by means of the shared server architecture.
    Session Multiplexing. Session multiplexing allows the same bandwidth to be used by many sessions at the same time, this enables the server to use fewer network connection endpoints for incoming request. In order for you to configure session multiplexing you have to configure connection manager.
    Ref. Oracle® Database Net Services Administrator's Guide
    10g Release 2 (10.2)
    Part Number B14212-02
    ~ Madrid
    http://hrivera99.blogspot.com/

  • Connection pooling and auditing on an oracle database

    Integration of a weblogic application with an oracle backend,
    Connection pooling, and auditing ,2 conflicting requirements ?
    Problem statement :
    We are in the process of maintaining a legacy client server application where
    the client is
    written in PowerBuilder and the backend is using an Oracle database.
    Almost all business logic is implemented in stored procedures on the database.
    When working in client/server mode ,1 PowerBuilder User has a one-to-one relation
    with
    a connection(session) on the oracle database.
    It is a requirement that the database administrator must see the real user connected
    to the database
    and NOT some kind of superuser, therefore in the PowerBuilder app each user connects
    to the database
    with his own username.(Each user is configured on the database via a seperate
    powerbuilder security app).
    For the PowerBuilder app all is fine and this app can maintain conversional state(setting
    and
    reading of global variables in oracle packages).
    The management is pushing for web-based application where we will be using bea
    weblogic appserver(J2EE based).
    We have build an business app which is web-based and accessing the same oracle
    backend app as
    the PowerBuilder app is doing.
    The first version of this web-based app is using a custom build connector(based
    on JCA standard and
    derived from a template provided by the weblogic integration installation).
    This custom build connector is essentially a combination of a custom realm in
    weblogic terms
    and a degraded connection pool , where each web session(browser) has a one-to-one
    relation
    with the back end database.
    The reason that this custom connector is combining the security functionality
    and the pooling
    functionality , is because each user must be authenticated against the oracle
    database(security requirement)
    and NOT against a LDAP server, and we are using a statefull backend(oracle packages)
    which would make it
    difficult to reuse connections.
    A problem that surfaced while doing heavy loadtesting with the custom connector,
    is that sometimes connections are closed and new ones made in the midst of a transaction.
    If you imagine a scenario where a session bean creates a business entity ,and
    the session bean
    calls 1 entity bean for the header and 1 entity bean for the detail, then the
    header and detail
    must be created in the same transaction AND with the same connection(there is
    a parent-child relationship
    between header and detail enforced on the back end database via Primary and Foreing
    Keys).
    We have not yet found why weblogic is closing the connection!
    A second problem that we are experincing with the custom connector, is the use
    of CMP(container managed persistence)
    within entity beans.
    The J2EE developers state that the use of CMP decreases the develoment time and
    thus also maintenance costs.
    We have not yet found a way to integrate a custom connector with the CMP persistence
    scheme !
    In order to solve our loadtesting and CMP persistence problems i was asked to
    come up with a solution
    which should not use a custom connector,but use standard connection pools from
    weblogic.
    To resolve the authentication problem on weblogic i could make a custom realm
    which connects to the
    backend database with the username and password, and if the connection is ok ,
    i could consider this
    user as authenticated in weblogic.
    That still leaves me with the problem of auditing and pooling.
    If i were to use a standard connection pool,then all transaction made in the oracle
    database
    would be done by a pool user or super user, a solution which will be rejected
    by our local security officer,
    because you can not see which real user made a transaction in the database.
    I could still use the connection pool and in the application , advise the application
    developers
    to set an oracle package variable with the real user, then on arrival of the request
    in the database,
    the logic could use this package variable to set the transaction user.
    There are still problems with this approach :
    - The administrator of the database can still not see who is connected , he will
    only see the superuser connection.
    - This scheme can not be used when you want to use CMP persistence , since it
    is weblogic who will generate the code
    to access the database.
    I thought i had a solution when oracle provided us with a connection pool known
    as OracleOCIConnectionPool
    where there is a connection made by a superuser, but where sessions are multiplexed
    over this physical pipe with the real user.
    I can not seem to properly integrate this OCI connectionpool into weblogic.
    When using this pool , and we are coming into a bean (session or entity bean)
    weblogic is wrapping
    this pool with it's own internal Datasource and giving me back a connection of
    the superuser, but not one for the real user,
    thus setting me with my back to the wall again.
    I would appreciate if anyone had experienced the same problem to share a possible
    solution with us
    in order to satisfy all requirements(security,auditing,CMP).
    Many Thanks
    Blyau Gino
    [email protected]

    Hi Blyau,
    As Joe has already provided some technical advice,
    I'll try to say something on engineering process level.
    While migrating an application from one technology to
    other, like client-server to n-tier in you case, customers and
    stakeholders want to push into the new system as many old
    requirements as possible. This approach is AKA "we must
    have ALL of the features of the old system". Mostly it happens
    because they don't know what they want. Ad little understanding
    of abilities of the new technology, and you will get a requirement
    like the one you have in you hands.
    I think "DBA must see real user" is one of those. For this
    type of requirements it can make sense to try to drop it,
    or to understand its nature and suggest alternatives. In this
    particular case it can be a system that logs user names,
    login and logout times.
    Blind copying of old features into an incompatible new architecture
    may endanger the whole project and can result in its failure.
    Hope this helps.
    Regards,
    Slava Imeshev
    "Blyau Gino" <[email protected]> wrote in message
    news:[email protected]...
    >
    Integration of a weblogic application with an oracle backend,
    Connection pooling, and auditing ,2 conflicting requirements ?
    Problem statement :
    We are in the process of maintaining a legacy client server applicationwhere
    the client is
    written in PowerBuilder and the backend is using an Oracle database.
    Almost all business logic is implemented in stored procedures on thedatabase.
    When working in client/server mode ,1 PowerBuilder User has a one-to-onerelation
    with
    a connection(session) on the oracle database.
    It is a requirement that the database administrator must see the real userconnected
    to the database
    and NOT some kind of superuser, therefore in the PowerBuilder app eachuser connects
    to the database
    with his own username.(Each user is configured on the database via aseperate
    powerbuilder security app).
    For the PowerBuilder app all is fine and this app can maintainconversional state(setting
    and
    reading of global variables in oracle packages).
    The management is pushing for web-based application where we will be usingbea
    weblogic appserver(J2EE based).
    We have build an business app which is web-based and accessing the sameoracle
    backend app as
    the PowerBuilder app is doing.
    The first version of this web-based app is using a custom buildconnector(based
    on JCA standard and
    derived from a template provided by the weblogic integrationinstallation).
    This custom build connector is essentially a combination of a custom realmin
    weblogic terms
    and a degraded connection pool , where each web session(browser) has aone-to-one
    relation
    with the back end database.
    The reason that this custom connector is combining the securityfunctionality
    and the pooling
    functionality , is because each user must be authenticated against theoracle
    database(security requirement)
    and NOT against a LDAP server, and we are using a statefull backend(oraclepackages)
    which would make it
    difficult to reuse connections.
    A problem that surfaced while doing heavy loadtesting with the customconnector,
    >
    is that sometimes connections are closed and new ones made in the midst ofa transaction.
    If you imagine a scenario where a session bean creates a business entity,and
    the session bean
    calls 1 entity bean for the header and 1 entity bean for the detail, thenthe
    header and detail
    must be created in the same transaction AND with the same connection(thereis
    a parent-child relationship
    between header and detail enforced on the back end database via Primaryand Foreing
    Keys).
    We have not yet found why weblogic is closing the connection!
    A second problem that we are experincing with the custom connector, is theuse
    of CMP(container managed persistence)
    within entity beans.
    The J2EE developers state that the use of CMP decreases the develomenttime and
    thus also maintenance costs.
    We have not yet found a way to integrate a custom connector with the CMPpersistence
    scheme !
    In order to solve our loadtesting and CMP persistence problems i was askedto
    come up with a solution
    which should not use a custom connector,but use standard connection poolsfrom
    weblogic.
    To resolve the authentication problem on weblogic i could make a customrealm
    which connects to the
    backend database with the username and password, and if the connection isok ,
    i could consider this
    user as authenticated in weblogic.
    That still leaves me with the problem of auditing and pooling.
    If i were to use a standard connection pool,then all transaction made inthe oracle
    database
    would be done by a pool user or super user, a solution which will berejected
    by our local security officer,
    because you can not see which real user made a transaction in thedatabase.
    I could still use the connection pool and in the application , advise theapplication
    developers
    to set an oracle package variable with the real user, then on arrival ofthe request
    in the database,
    the logic could use this package variable to set the transaction user.
    There are still problems with this approach :
    - The administrator of the database can still not see who is connected ,he will
    only see the superuser connection.
    - This scheme can not be used when you want to use CMP persistence , sinceit
    is weblogic who will generate the code
    to access the database.
    I thought i had a solution when oracle provided us with a connection poolknown
    as OracleOCIConnectionPool
    where there is a connection made by a superuser, but where sessions aremultiplexed
    over this physical pipe with the real user.
    I can not seem to properly integrate this OCI connectionpool intoweblogic.
    When using this pool , and we are coming into a bean (session or entitybean)
    weblogic is wrapping
    this pool with it's own internal Datasource and giving me back aconnection of
    the superuser, but not one for the real user,
    thus setting me with my back to the wall again.
    I would appreciate if anyone had experienced the same problem to share apossible
    solution with us
    in order to satisfy all requirements(security,auditing,CMP).
    Many Thanks
    Blyau Gino
    [email protected]

  • Connection Pool and Database Sessions

    Hi,
    Is there any way to use the connection pool or Datasource while connecting to database?If I am using a stateless sesssion bean and using a Data Access layer which just creates a database session to write the persistence toplink objects how I can make use of application server connection pool?
    Thanks,
    Vinod

    Hi,
    Is there any way to use the connection pool or Datasource while connecting to database?If I am using a stateless sesssion bean and using a Data Access layer which just creates a database session to write the persistence toplink objects how I can make use of application server connection pool?Hi Vinod,
    Yes, TopLink allows you to use the app server's connection pooling and transaction services. Chapter 2 of the Oracle9iAS TopLink Foundation Library Guide provides details as do the TopLink examples. The easiest way to set this up is by using the sessions.xml file. The sample XML below is from the file <toplink903>\examples\ias\examples\ejb\sessionbean\sessions.xml. Here we are adding the datasource defined in OC4J and specifying that we are using the OC4J transaction controller also.
    <login>
    <user-name>sa</user-name>
    <password></password>
    <datasource>java:comp/env/jdbc/ejbJTSDataSource</datasource>
    <uses-external-transaction-controller>true</uses-external-transaction-controller>
    <uses-external-connection-pool>true</uses-external-connection-pool>
    </login>
    <external-transaction-controller-class>oracle.toplink.jts.oracle9i.Oracle9iJTSExternalTransactionController</external-transaction-controller-class>
    When using this approach you need to change your TopLink code slightly in the EJB methods:
    a. Acquire the ACTIVE unit of work from the server
    session (again, see the EmployeeSessionEJB code
    example) with something like:
    UnitOfWork uow = clientSession.getActiveUnitOfWork();
    b. Calls to uow.commit() can be ommitted or commented out
    because the EJB will handle this. Note that of course
    the methods you create in the EJB that are using this
    approach must have TX Required (default).
    Hope this helps.
    Pete

Maybe you are looking for

  • Arrow buttons missing from the scrollbars!

    Does anyone know of anyway including a third party extension to return arrow buttons to the scrollbrs? Lion has what is a fatal flaw to me. There are no arrow buttons in the scrollbars. Not even an option for them! I am aware of all the other ways to

  • FSCM 9.0: COBOL Process GLPPPOST does not post Journal Entry correctly

    Folks, Hello. I am workin on FSCM 9.0 General Ledger module. "Edit Journal" is AE program that works correctly. But "Post Journal" is a COBOL program that does not works correctly. COBOL process GLPPPOST posts only debit accounts of a Journal entry i

  • JMS / MBD Best Practices

    Hello All, There is currently a dispute in my camp about the limit of "responsibility" to place on a single MDB. In this case, the argument surrounds whether it is "good practice" to allow a MDB's onMessage function to: - Open a connection to a servl

  • Bex to portal connectivity

    Hi all when i am executing a query in BEx, it is firing a wrong target address. where i remove some text from that URL string , i am able to connect to EP well, the location path which is firing when executing is actually here "ephdev" appearing two

  • FIRMWARE N MOBILE SECURITY 4.0 ON N-70

    I did a firmware upgrade yesterday and it went fine till I noticed that the Symantec Mobile Security 4.0 which was pre-installed on my phone could not not be found anywhere on the phone. I can locate it on the memory card through Application Manager