Question on autcommit of transaction and database configuration

I have a ase 12.5 with following setting on a dev server:
select into/bulkcopy/pllsort : true
ddl in tran: true
when run a pb app without autocommit setting. It is fine.
but on my production server, same db have different option setting:
select into/bulkcopy/pllsort : false
ddl in tran: false
then run same app, the transaction from pb app not commit, cause lot of lock and crash app.
So what the relationship between pb transaction autocommit attribute and sybase ase database dboptions setting?

Hi Kent;
  FWIW: Check the "locking scheme) in development vs production. You may have page level in one DB vs row locking enabled in the other environment. This can make a world of difference in the locking conflicts when you run your PB application.
Regards ... Chris

Similar Messages

  • Transactions and database locks

    Hi,
                   We use Weblogic 4.5.1 on Windows NT 4.0 with Oracle 8.0.5. Our database
              isolation is set to TRANSACTION_READ_COMMITTED. I have an entity bean with
              TX_REQUIRED & TRANSACTION_READ_COMMITTED settings. If my client creates a
              transaction, and starts calling methods on this entity bean, is the
              corresponding database row locked for the duration of the transaction? We
              have concurrent SQl-plus sessions going on and we want make sure there is
              no data corruption. If the row is not locked, is it ok for me to explicitly
              lock it from inside my entity bean?
              Thanks,
              Srini.
              

    Hi. This should have been posted to the EJB or JDBC group, but I'll take it.
              This is an Oracle question. If you have a transaction as you've described,
              then the behavior will be exactly as if you had multiple SQL-PLUS sessions,
              and in one of them, you did:
              SQL> BEGIN;
              -- do what your bean would do;
              SQL> COMMIT;
              You can test this there. In general, you'll find that Oracle's optimistic locking
              will allow any number of simutaneous transactions to access a given row
              at one time. Oracle does not lock the real data while a transaction is ongoing,
              instead making a copy for the client to work off of. At commit time, depending
              on the isolation level semantics, some or all of the transactions may fail when
              Oracle tries to update the real data from the per-session private data.
              I would council against running with SERIALIZABLE mode because there
              is a serious bug in Oracle, where serializable transactions may fail silently.
              Details on request.
              Joe
              Srini wrote:
              > Hi,
              > We use Weblogic 4.5.1 on Windows NT 4.0 with Oracle 8.0.5. Our database
              > isolation is set to TRANSACTION_READ_COMMITTED. I have an entity bean with
              > TX_REQUIRED & TRANSACTION_READ_COMMITTED settings. If my client creates a
              > transaction, and starts calling methods on this entity bean, is the
              > corresponding database row locked for the duration of the transaction? We
              > have concurrent SQl-plus sessions going on and we want make sure there is
              > no data corruption. If the row is not locked, is it ok for me to explicitly
              > lock it from inside my entity bean?
              >
              > Thanks,
              > Srini.
              PS: Folks: BEA WebLogic is in S.F., and now has some entry-level positions for
              people who want to work with Java and E-Commerce infrastructure products. Send
              resumes to [email protected]
              The Weblogic Application Server from BEA
              JavaWorld Editor's Choice Award: Best Web Application Server
              Java Developer's Journal Editor's Choice Award: Best Web Application Server
              Crossroads A-List Award: Rapid Application Development Tools for Java
              Intelligent Enterprise RealWare: Best Application Using a Component Architecture
              http://weblogic.beasys.com/press/awards/index.htm
              

  • Question about model driven dev and RDS configuring

    I am testing the model driven development that is brought up by Flash Builder4 and LCDS3 recently, and it went pretty well. However i've noticed that the configuration of the RDS for development mode is not recommanded for a production environment.
    My first quetion is can I simply disable the RDS on the server when I want to deploy my app to a production environment and still be able to use all of the features such as data management service and remote service etc.? I've tried to do so but it gave me a 404 page. I felt it was a little bit confusing when you know you DO have a html template located in the right location. If RDS can be disabled without interfering the functionality of the app, where can I find a doc to follow?
    My second question is if the RDS is a must, then how to configure the security setting? I knew some of the requirements by visiting this url:http://help.adobe.com/en_US/LiveCycleDataServicesES/3.1/Developing/WS6822bf3ec1c7b2e9-30fb 9b22126b9ca84aa-8000.html,
    however, when I followed the steps and test it, I got a 400 error in FB4 RDS configuration page.
    Since I am testing this under Tomcat 5.5, I am not sure where should I put the flex-tomcat-common.jar and flex-tomcat-server.jar file in, cause the dir structure of Tomcat 5.5 is a little bit different than that of Tomcat 6.0. Then I put these files under both of the tomcat 5.5\common\lib folder and the tomcat 5.5\server\lib folder. I've doubled checked the the catalina.properties file which is located under tomcat5.5\conf folder and I see the following environment param has been there : ${catalina.home}/common/lib/*.jar.
    The other configuration I made that differs from the doc was: I put <valve classname="flex.messaging.security.TomcatValve"></valve> into the context.xml, which is under the folder Tomcat 5.5/conf. Since I put the Resource definition in the same file, I guess this should not be the problem.
    Does anyone has any successful experience on setting this up?
    Any hint will be appreciated.

    Thanks for the quick reply.
    I solved my first problem, it was because I accidentally commented out part of the RDS configuration and left the servlet-mapping part un-commented. After I generate all the client and server codes and re-deploy them to the environment, I finally got my running demo with RDS disabled.
    For the second problem, I am still stucking in. I've tried to switch to the Tomcat 6.0 and follow both of the LCDS installation doc and the LCDS3.1_using doc. None could work me out. I constantly get the error message said that "The RDS server was successfully contacted, but your security credentials were invalid". So after struggling a whole day, I found the following statement, which is appeared in the ColdFusion configuraton doc:"Note: If you are using ColdFusion MX 7 or earlier, the message "The RDS server was successfully contacted, but your security credentials were invalid," appears. The message indicates that the password was not validated, even if it is correct. Click OK to close the message. ". That's interesting cause it seems it was exact the situation I was encountered. However, in my case, ignoring the error message does not give me a successful RDS connection.
    For your information, I am pasting the configurations I had on my machine and maybe some ppl could give me a correction.
    service-config.xml:
    <security>
      <login-command class="flex.messaging.security.TomcatLoginCommand" server="Tomcat"/>       
            <!-- Uncomment the correct app server
            <login-command class="flex.messaging.security.JRunLoginCommand" server="JRun"/>
            <login-command class="flex.messaging.security.WeblogicLoginCommand" server="Weblogic"/>
            <login-command class="flex.messaging.security.WebSphereLoginCommand" server="WebSphere"/>
            -->
            <security-constraint id="basic-read-access">
                <auth-method>Custom</auth-method>
                <roles>
                    <role>guests</role>
                    <role>accountants</role>
                    <role>employees</role>
                    <role>managers</role>
                    <role>rds</role>
                </roles>
            </security-constraint>
        </security>
    tomcat-user.xml:
    <tomcat-users>
      <role rolename="rds"/>
      <user username="rdsuser" password="rdspassword" roles="rds"/>
    <!--
      <role rolename="tomcat"/>
      <role rolename="role1"/>
      <user username="tomcat" password="tomcat" roles="tomcat"/>
      <user username="both" password="tomcat" roles="tomcat,role1"/>
      <user username="role1" password="tomcat" roles="role1"/>
    -->
    </tomcat-users>
    web.xml:
    <servlet>
            <servlet-name>RDSDispatchServlet</servlet-name>
        <display-name>RDSDispatchServlet</display-name>
            <servlet-class>flex.rds.server.servlet.FrontEndServlet</servlet-class>
            <init-param>
          <param-name>useAppserverSecurity</param-name>
          <param-value>true</param-value>
        </init-param>
            <load-on-startup>10</load-on-startup>
        </servlet>
        <servlet-mapping id="RDS_DISPATCH_MAPPING">
            <servlet-name>RDSDispatchServlet</servlet-name>
            <url-pattern>/CFIDE/main/ide.cfm</url-pattern>
        </servlet-mapping>
    lcds.xml:
    <Valve className="flex.messaging.security.TomcatValve"/>
    <Transaction factory="org.objectweb.jotm.UserTransactionFactory" jotm.timeout="60"/>
    <Resource name="jdbc/codedata" auth="Container"
       type="javax.sql.DataSource"
       username="root"
       password="******"
       driverClassName="com.mysql.jdbc.Driver"
       url="jdbc:mysql://localhost:3306/codedata?autoReconnect=true"/>
       <Resource name="jdbc/atlantis" auth="Container"
       type="javax.sql.DataSource"
       username="root"
       password="******"
       driverClassName="com.mysql.jdbc.Driver"
       url="jdbc:mysql://localhost:3306/atlantis?autoReconnect=true"/> 
    Note that both of the flex-tomcat-common.jar and flex-tomcat-server.jar were already under Tomcat 6.0/lib folder.

  • Re: Transactions and Locking Rows for Update

    Dale,
    Sounds like you either need an "optimistic locking" scheme, usually
    implemented with timestamps at the database level, or a concurrency manager.
    A concurrency manager registers objects that may be of interest to multiple
    users in a central location. It takes care of notifying interested parties
    (i.e., clients,) of changes made to those objects, using a "notifier" pattern.
    The optimistic locking scheme is relatively easy to implement at the
    database level, but introduces several problems. One problem is that the
    first person to save their changes "wins" - every one else has to discard
    their changes. Also, you now have business policy effectively embedded in
    the database.
    The concurrency manager is much more flexible, and keeps the policy where
    it probably belongs. However, it is more complex, and there are some
    implications to performance when you get to the multiple-thousand-user
    range because of its event-based nature.
    Another pattern of lock management that has been implemented is a
    "key-based" lock manager that does not use events, and may be more
    effective at managing this type of concurrency for large numbers of users.
    There are too many details to go into here, but I may be able to give you
    more ideas in a separate note, if you want.
    Don
    At 04:48 PM 6/5/97 PDT, Dale "V." Georg wrote:
    I have a problem in the application I am currently working on, which it
    seems to me should be easily solvable via appropriate use of transactions
    and database locking, but I'm having trouble figuring out exactly how to
    do it. The database we are using is Oracle 7.2.
    The scenario is as follows: We have a window where the user picks an
    object from a dropdown list. Some of the object's attributes are then
    displayed in that window, and the user then has the option of editing
    those attributes, and at some point hitting the equivalent of a 'save'button
    to write the changes back to the database. So far, so good. Now
    introduce a second user. If user #1 and user #2 both happen to pull up
    the same object and start making changes to it, user #1 could write back
    to the database and then 15 seconds later user #2 could write back to the
    database, completely overlaying user #1's changes without ever knowing
    they had happened. This is not good, particularly for our application
    where editing the object causes it to progress from one state to the next,
    and multiple users trying to edit it at the same time spells disaster.
    The first thing that came to mind was to do a select with intent to update,
    i.e. 'select * from table where key = 'somevalue' with update'. This way
    the next user to try to select from the table using the same key would not
    be able to get it. This would prevent multiple users from being able to
    pull the same object up on their screens at the same time. Unfortunately,
    I can think of a number of problems with this approach.
    For one thing, the lock is only held for the duration of the transaction, so
    I would have to open a Forte transaction, do the select with intent to
    update, let the user modify the object, then when they saved it back again
    end the transaction. Since a window is driven by the event loop I can't
    think of any way to start a transaction, let the user interact with the
    window, then end the transaction, short of closing and re-opening the
    window. This would imply having a separate window specifically for
    updating the object, and then wrapping the whole of that window's event
    loop in a transaction. This would be a different interface than we wanted
    to present to the users, but it might still work if not for the next issue.
    The second problem is that we are using a pooled DBSession approach
    to connecting to the database. There is a single Oracle login account
    which none of the users know the password to, and thus the users
    simply share DBSession resources. If one user starts a transaction
    and does a select with intent to update on one DBSession, then another
    user starts a transaction and tries to do the same thing on the same
    DBSession, then the second user will get an error out of Oracle because
    there's already an open transaction on that DBSession.
    At this point, I am still tossing ideas around in my head, but after
    speaking with our Oracle/Forte admin here, we came to the conclusion
    that somebody must have had to address these issues before, so I
    thought I'd toss it out and see what came back.
    Thanks in advance for any ideas!
    Dale V. Georg
    Indus Consultancy Services [email protected]
    Mack Trucks, Inc. [email protected]
    >
    >
    >
    >
    ====================================
    Don Nelson
    Senior Consultant
    Forte Software, Inc.
    Denver, CO
    Corporate voice mail: 510-986-3810
    aka: [email protected]
    ====================================
    "I think nighttime is dark so you can imagine your fears with less
    distraction." - Calvin

    We have taken an optimistic data locking approach. Retrieved values are
    stored as initial values; changes are stored seperately. During update, key
    value(s) or the entire retieved set is used in a where criteria to validate
    that the data set is still in the initial state. This allows good decoupling
    of the data access layer. However, optimistic locking allows multiple users
    to access the same data set at the same time, but then only one can save
    changes, the rest would get an error message that the data had changed. We
    haven't had any need to use a pessimistic lock.
    Pessimistic locking usually involves some form of open session or DBMS level
    lock, which we haven't implemented for performance reasons. If we do find the
    need for a pessimistic lock, we will probably use cached data sets that are
    checked first, and returned as read-only if already in the cache.
    -DFR
    Dale V. Georg <[email protected]> on 06/05/97 03:25:02 PM
    To: Forte User Group <[email protected]> @ INTERNET
    cc: Richards* Debbie <[email protected]> @ INTERNET, Gardner*
    Steve <[email protected]> @ INTERNET
    Subject: Transactions and Locking Rows for Update
    I have a problem in the application I am currently working on, which it
    seems to me should be easily solvable via appropriate use of transactions
    and database locking, but I'm having trouble figuring out exactly how to
    do it. The database we are using is Oracle 7.2.
    The scenario is as follows: We have a window where the user picks an
    object from a dropdown list. Some of the object's attributes are then
    displayed in that window, and the user then has the option of editing
    those attributes, and at some point hitting the equivalent of a 'save' button
    to write the changes back to the database. So far, so good. Now
    introduce a second user. If user #1 and user #2 both happen to pull up
    the same object and start making changes to it, user #1 could write back
    to the database and then 15 seconds later user #2 could write back to the
    database, completely overlaying user #1's changes without ever knowing
    they had happened. This is not good, particularly for our application
    where editing the object causes it to progress from one state to the next,
    and multiple users trying to edit it at the same time spells disaster.
    The first thing that came to mind was to do a select with intent to update,
    i.e. 'select * from table where key = 'somevalue' with update'. This way
    the next user to try to select from the table using the same key would not
    be able to get it. This would prevent multiple users from being able to
    pull the same object up on their screens at the same time. Unfortunately,
    I can think of a number of problems with this approach.
    For one thing, the lock is only held for the duration of the transaction, so
    I would have to open a Forte transaction, do the select with intent to
    update, let the user modify the object, then when they saved it back again
    end the transaction. Since a window is driven by the event loop I can't
    think of any way to start a transaction, let the user interact with the
    window, then end the transaction, short of closing and re-opening the
    window. This would imply having a separate window specifically for
    updating the object, and then wrapping the whole of that window's event
    loop in a transaction. This would be a different interface than we wanted
    to present to the users, but it might still work if not for the next issue.
    The second problem is that we are using a pooled DBSession approach
    to connecting to the database. There is a single Oracle login account
    which none of the users know the password to, and thus the users
    simply share DBSession resources. If one user starts a transaction
    and does a select with intent to update on one DBSession, then another
    user starts a transaction and tries to do the same thing on the same
    DBSession, then the second user will get an error out of Oracle because
    there's already an open transaction on that DBSession.
    At this point, I am still tossing ideas around in my head, but after
    speaking with our Oracle/Forte admin here, we came to the conclusion
    that somebody must have had to address these issues before, so I
    thought I'd toss it out and see what came back.
    Thanks in advance for
    any
    ideas!
    Dale V. Georg
    Indus Consultancy Services [email protected]
    Mack Trucks, Inc. [email protected]
    ------ Message Header Follows ------
    Received: from pebble.Sagesoln.com by notes.bsginc.com
    (PostalUnion/SMTP(tm) v2.1.9c for Windows NT(tm))
    id AA-1997Jun05.162418.1771.334203; Thu, 05 Jun 1997 16:24:19 -0500
    Received: (from sync@localhost) by pebble.Sagesoln.com (8.6.10/8.6.9) id
    NAA11825 for forte-users-outgoing; Thu, 5 Jun 1997 13:47:58 -0700
    Received: (from uucp@localhost) by pebble.Sagesoln.com (8.6.10/8.6.9) id
    NAA11819 for <[email protected]>; Thu, 5 Jun 1997 13:47:56 -0700
    Received: from unknown(207.159.84.4) by pebble.sagesoln.com via smap (V1.3)
    id sma011817; Thu Jun 5 13:47:43 1997
    Received: from tes0001.macktrucks.com by relay.macktrucks.com
    via smtpd (for pebble.sagesoln.com [206.80.24.108]) with SMTP; 5 Jun
    1997 19:35:31 UT
    Received: from dale by tes0001.macktrucks.com (SMI-8.6/SMI-SVR4)
    id QAA04637; Thu, 5 Jun 1997 16:45:51 -0400
    Message-ID: <[email protected]>
    Priority: Normal
    To: Forte User Group <[email protected]>
    Cc: "Richards," Debbie <[email protected]>,
    "Gardner," Steve <[email protected]>
    MIME-Version: 1.0
    From: Dale "V." Georg <[email protected]>
    Subject: Transactions and Locking Rows for Update
    Date: Thu, 05 Jun 97 16:48:37 PDT
    Content-Type: text/plain; charset=US-ASCII; X-MAPIextension=".TXT"
    Content-Transfer-Encoding: quoted-printable
    Sender: [email protected]
    Precedence: bulk
    Reply-To: Dale "V." Georg <[email protected]>

  • Transactions and Locking Rows for Update

    I have a problem in the application I am currently working on, which it
    seems to me should be easily solvable via appropriate use of transactions
    and database locking, but I'm having trouble figuring out exactly how to
    do it. The database we are using is Oracle 7.2.
    The scenario is as follows: We have a window where the user picks an
    object from a dropdown list. Some of the object's attributes are then
    displayed in that window, and the user then has the option of editing
    those attributes, and at some point hitting the equivalent of a 'save' button
    to write the changes back to the database. So far, so good. Now
    introduce a second user. If user #1 and user #2 both happen to pull up
    the same object and start making changes to it, user #1 could write back
    to the database and then 15 seconds later user #2 could write back to the
    database, completely overlaying user #1's changes without ever knowing
    they had happened. This is not good, particularly for our application
    where editing the object causes it to progress from one state to the next,
    and multiple users trying to edit it at the same time spells disaster.
    The first thing that came to mind was to do a select with intent to update,
    i.e. 'select * from table where key = 'somevalue' with update'. This way
    the next user to try to select from the table using the same key would not
    be able to get it. This would prevent multiple users from being able to
    pull the same object up on their screens at the same time. Unfortunately,
    I can think of a number of problems with this approach.
    For one thing, the lock is only held for the duration of the transaction, so
    I would have to open a Forte transaction, do the select with intent to
    update, let the user modify the object, then when they saved it back again
    end the transaction. Since a window is driven by the event loop I can't
    think of any way to start a transaction, let the user interact with the
    window, then end the transaction, short of closing and re-opening the
    window. This would imply having a separate window specifically for
    updating the object, and then wrapping the whole of that window's event
    loop in a transaction. This would be a different interface than we wanted
    to present to the users, but it might still work if not for the next issue.
    The second problem is that we are using a pooled DBSession approach
    to connecting to the database. There is a single Oracle login account
    which none of the users know the password to, and thus the users
    simply share DBSession resources. If one user starts a transaction
    and does a select with intent to update on one DBSession, then another
    user starts a transaction and tries to do the same thing on the same
    DBSession, then the second user will get an error out of Oracle because
    there's already an open transaction on that DBSession.
    At this point, I am still tossing ideas around in my head, but after
    speaking with our Oracle/Forte admin here, we came to the conclusion
    that somebody must have had to address these issues before, so I
    thought I'd toss it out and see what came back.
    Thanks in advance for any ideas!
    Dale V. Georg
    Indus Consultancy Services [email protected]
    Mack Trucks, Inc. [email protected]
    [email protected]------------------

    I have a problem in the application I am currently working on, which it
    seems to me should be easily solvable via appropriate use of transactions
    and database locking, but I'm having trouble figuring out exactly how to
    do it. The database we are using is Oracle 7.2.
    The scenario is as follows: We have a window where the user picks an
    object from a dropdown list. Some of the object's attributes are then
    displayed in that window, and the user then has the option of editing
    those attributes, and at some point hitting the equivalent of a 'save' button
    to write the changes back to the database. So far, so good. Now
    introduce a second user. If user #1 and user #2 both happen to pull up
    the same object and start making changes to it, user #1 could write back
    to the database and then 15 seconds later user #2 could write back to the
    database, completely overlaying user #1's changes without ever knowing
    they had happened. This is not good, particularly for our application
    where editing the object causes it to progress from one state to the next,
    and multiple users trying to edit it at the same time spells disaster.
    The first thing that came to mind was to do a select with intent to update,
    i.e. 'select * from table where key = 'somevalue' with update'. This way
    the next user to try to select from the table using the same key would not
    be able to get it. This would prevent multiple users from being able to
    pull the same object up on their screens at the same time. Unfortunately,
    I can think of a number of problems with this approach.
    For one thing, the lock is only held for the duration of the transaction, so
    I would have to open a Forte transaction, do the select with intent to
    update, let the user modify the object, then when they saved it back again
    end the transaction. Since a window is driven by the event loop I can't
    think of any way to start a transaction, let the user interact with the
    window, then end the transaction, short of closing and re-opening the
    window. This would imply having a separate window specifically for
    updating the object, and then wrapping the whole of that window's event
    loop in a transaction. This would be a different interface than we wanted
    to present to the users, but it might still work if not for the next issue.
    The second problem is that we are using a pooled DBSession approach
    to connecting to the database. There is a single Oracle login account
    which none of the users know the password to, and thus the users
    simply share DBSession resources. If one user starts a transaction
    and does a select with intent to update on one DBSession, then another
    user starts a transaction and tries to do the same thing on the same
    DBSession, then the second user will get an error out of Oracle because
    there's already an open transaction on that DBSession.
    At this point, I am still tossing ideas around in my head, but after
    speaking with our Oracle/Forte admin here, we came to the conclusion
    that somebody must have had to address these issues before, so I
    thought I'd toss it out and see what came back.
    Thanks in advance for any ideas!
    Dale V. Georg
    Indus Consultancy Services [email protected]
    Mack Trucks, Inc. [email protected]
    [email protected]------------------

  • Database Configuration Assistant fails during instalation of Oracle EE 8.1.7

    My machine runs RH 6.2 and I have tried to install it Oracle. After executing all pre-instalation steps, I run ./runInstall and instalation process goes on. I select Typical Instalation and instalation proceeds until 100%, when OUI tries to configure Net 8 Assistant and Database Configuration Assistant. Both applications fail in this phase. Unfortunately, OUI shows a message (Database Configuration Assistant) informing that /u01/.../JRE/bin/jre does not exist. However, I have installed JDK 118 and JRE link seems correct, so that I can access JDK directory throught JRE link.
    HELP, HELP!!!
    Cheers,
    Andre

    some good installation guides to check out:
    http://www.akadia.com/services/ora_linux_install_901.html
    http://www.e-gineer.com/instructions/install-oracle9i-on-linux.phtml
    good luck!
    Jenny

  • Database Configuration Assistant fails during installation Oracle EE 8.1.7

    My machine runs RH 6.2 and I have tried to install it Oracle. After executing all pre-instalation steps, I run ./runInstall and instalation process goes on. I select Typical Instalation and instalation proceeds until 100%, when OUI tries to configure Net 8 Assistant and Database Configuration Assistant. Both applications fail in this phase. Unfortunately, OUI shows a message (for Database Configuration Assistant) informing that /u01/.../JRE/bin/jre does not exist. However, I have installed JDK 118 and JRE link seems correct, so that I can access JDK directory throught JRE link.
    HELP, HELP!!!
    Cheers,
    Andre

    some good installation guides to check out:
    http://www.akadia.com/services/ora_linux_install_901.html
    http://www.e-gineer.com/instructions/install-oracle9i-on-linux.phtml
    good luck!
    Jenny

  • Is it possible to configure transaction and merge replication one database as a publisher?

    Hi All,
    We have a requirement to configure replication between the servers and its plan is as fallows:
    A------>B -- One way transactional replication ---  Already Configured
    A------> C --Merge Replication (Both ways)--- This is in plan
    1) Our requirement to configure Merge replication is to allow multiple users to access both publisher and subscriber databases. Configuring Merge replication with a combination of Transactional replication is correct or do we end up facing issues?
    2) A to B transactional replication is already configured so if we configure merge replication between A to C will effect existing replication?
    Please let us know if you need any details on this. Thank You.
    Grateful to your time and support. Regards, Shiva

    Hi Sir,
    Thanks for the information. But I have small doubt, "best to use merge all the way" is that mean you recommend to use Merge replication for all the servers.
    A------>B --
    Merge Replication (Both ways) 
    A------> C --Merge Replication (Both ways)
    Please correct me if I am wrong
    Grateful to your time and support. Regards, Shiva

  • Standby Database Configuration with LGWR and ASYNC

    Hi,
    I am running a standby database configuration on a 100MBit LAN
    with the following set-up:
    log_archive_dest_2='service=standby mandatory reopen=300 lgwr
    async=2000'
    The system is handling a lot of very small transactions, not
    involving large amounts of data.
    My questions are:
    - What are the potential problems in using a small ASYNC buffer
    like the one above?
    - Does a larger ASYNC buffer influence the latency in copying
    changes from the production database to the standby database -
    will it buffer more changes before sending them to the standby
    database?

    Murlib,
    I have few more doubt-
    Our requirement is to configure a Standby( Physical-MAXIMUM PERFORMANCE
    mode) in a place, which is 600 KM away from our primary destination.
    Currently our LAN network traffic rate is 100 Mbps. but this traffic is
    virtually reduced to 1Mbps out side our LAN.
    Our Production Database is 24X7 and Its generating 17 GB Archive files every
    day.
    Since the net work traffic is slow i think, it will create some log gaps,and
    also we couldn't do a point in time recovery.
    We are configuring a Standby, here inside our LAN in Managed Recovery Mode
    for recovery purpose and will keep a Standby there in remote place for
    Reports.
    and it will be recover it in every morning.If iam following this procedure,
    my
    log_archive_dest_2 ='service=stby ARCH NOAFFIRM' ( which is the standby here
    inside our LAN and should be in MANAGED RECOVERY mode)
    and i need to configure the parameters for standby in my remote location.So
    my doubt is-
    Shall i need to configure "log_archive_dest_n" in the parameterlist of my
    Primary for that remote Standby ?
    I think for manual recovery we can aviod that.But we need to eonnect it
    thrrough Oracle Net
    Can you please tell me the essential PRIMARY parameter list entries for this
    kind of remote standby , recovering in a manual mode ?
    i think the following parameter should be there -
    FAL_SERVER
    FAL_CLIENT
    DB_FILE_NAME_CONVERT
    LOG_FILE_NAME_CONVERT
    STANDBY_FILE_MANAGEMENT=AUTO
    STANDBY_ARCHIVE_DEST
    Thanks and Regards,
    Raj

  • Transaction Sync and Database Size

    Hello,
    We're using BDB (via the Java interface) as the persistent store for a messaging platform. In order to achieve high performance, the transactional operations are configured to not sync, i.e., TransactionConfig.setSync(false) . While we do achieve better performance, the size of the database does seem rather large. We checkpoint on a periodic basis, and each time we checkpoint, the size of the database grows, even though records (messages in our world) are being deleted. So, if I were to add, say 10000 records, delete all of them and then checkpoint, the size of the database would actually grow! In addition, the database file, while being large, is also very sparse - a 30GB file when compressed reduces in size to 0.5 GB.
    We notice that if we configure our transactional operations to sync, the size is much smaller, and stays constant, i.e., if I were to insert and subsequently delete 10000 records into a database whose file is X MB, the size of the database file after the operations would be roughly X MB.
    I understand that transaction logs are applied to the database when we checkpoint, but should I be configuring the behaviour of the checkpointing (via CheckpoinConfig )?
    Also, I am checkpointing periodically from a separate thread. Does BDB itself spawn any threads for checkpointing?
    Our environment is as follows:
    RedHat EL 2.6.9-34.ELsmp
    Java 1.5
    BDB 4.5.20
    Thanks much in advance,
    Prashanth

    Hi Prashanth,
    If your delete load is high, your workload should benefit from setting the DB_REVSPLITOFF flag, which keeps around the structure of the btree regardless of records being deleted. The result should be less splits and merges, and is therefore better concurrency.
    Here you can find some documentation that should help you:
    Access method tuning: http://www.oracle.com/technology/documentation/berkeley-db/db/ref/am_misc/tune.html
    Transaction tuning: http://www.oracle.com/technology/documentation/berkeley-db/db/ref/transapp/tune.html
    If you are filling the cache with dirty pages, you can indeed call checkpoint() periodically in the application, or you can create a memp_trickle thread. See the following sections of the documentation:
    - Javadoc: Environment -> trickleCacheWrite" http://www.oracle.com/technology/documentation/berkeley-db/db/java/com/sleepycat/db/Environment.html#trickleCacheWrite(int)
    Some related thread for the "database size issue", can be found here: http://forums.oracle.com/forums/thread.jspa?threadID=534371&tstart=0
    Bogdan Coman

  • Transaction Database Configuration option

    Hi:
    When I tried to install Oracle 9.2.0.1.0 with Transaction Database Configuration option, the assistant process got me the following error message ...
    Invalid entry size (expected 152702976 but got 152703068 bytes) ....
    It's appear that this happened when assistant is trying to clone the EXAMPLE01.DBF datafile ...
    Does any body know any clue or hint to fix it ??
    Thanks and Regards
    Angel Castro
    Intergraph Mexico

    prob. a known bug.
    Check on metalink or set up a GP database and retune it.

  • Ocfs2 configuration require crs and database shutdown

    Hi,
    We are in the process of adding a node to the cluster.
    This is a 3 node cluster. we are going to add a 4 node to the cluster.
    ocr , voting disk are kept on ocfs2 file system. when we are trying to configure ocfs2.
    we were not able to mount ocfs2 device 4th node. we install all the ocfs2 required which was not there on 4 node.
    we can see the entries of 3 node in ocfs2 configure file we have make the 4th node using ocfs2cosole view.
    node:
    ip_port = 7777
    ip_address = 192.168.11.97
    number = 0
    name = rscorc07a.corp.shire.com
    cluster = ocfs2
    node:
    ip_port = 7777
    ip_address = 192.168.11.99
    number = 1
    name = rscorc07b.corp.shire.com
    cluster = ocfs2
    node:
    ip_port = 7777
    ip_address = 192.168.11.101
    number = 2
    name = rscorc07c.corp.shire.com
    cluster = ocfs2
    node:
    ip_port = 7777
    ip_address = 192.168.11.103
    number = 3
    name = rsdorc07d.corp.shire.com
    cluster = ocfs2
    cluster:
    node_count = 4
    name = ocfs2
    we need to add the 4th node entry on all the nodes.
    My question is does this require shutdown of database or crs or we can configure ocfs2 and add ocfs online(when the crs and database is up)
    Thanks and Regards,

    Hi All,
    The crs and database dont need to be shut while configuring ocfs2 for node addition.
    We took a help from oracle support.
    They asked us to increase the min_free_kbytes to 10000 kernel parameter is sysctl.conf and reproduce the issue.
    we have increased it to 1000 and we will again try to add the 4th node and let you know what is the outcome.
    Thanks

  • How can i create a new and tableless database using database configuration

    How can i create a new and tableless database using database configuration

    How can i create a new and tableless database using database configuration
    Just don't install the sample schemas. See the installation guide
    http://docs.oracle.com/cd/E11882_01/server.112/e10831/installation.htm
    Using the Database Configuration Assistant
    When you install Oracle Database with the Oracle Universal Installer, the sample schemas are installed by default if you select the Basic Installation option. Selecting the sample schemas option installs all five schemas (HR, OE, PM, IX, and SH) in the database. If you choose not to install the sample schemas at that time, you can add them later by following the instructions in section "Manually Installing Sample Schemas".
    Choose a custom install and don't install the sample schemas.
    All other schems/tables installed are REQUIRED by Oracle

  • ERROR: Exception occured while encrypting the configuration and database

    I'm facing below issue/error during the OIM 11g R2 configuration (fresh install).  Resolutions from other blog with same error (DOMAIN_HOME misconfigured) isn't helping in my case.
    Thanks for your help
    updateMLSLocale:ORACLE_HOME :/fmw/Oracle_IDM1
    updateMLSLocale:LOCALE_PROPERTIES_FILE :/fmw/Oracle_IDM1/inventory/Scripts/ext/jlib/oim/OIMLocales.properties
    java.lang.Exception: Exception occured while encrypting the configuration and database
      at oracle.as.install.oim.config.util.EncryptConfigurationAndDB.encryptConfigurationAndDatbase(EncryptConfigurationAndDB.java:239)
      at oracle.as.install.oim.config.OIMConfigManager.encryptDB(OIMConfigManager.java:1035)
      at oracle.as.install.oim.config.OIMConfigManager.configureOIM(OIMConfigManager.java:891)
      at oracle.as.install.oim.config.OIMConfigManager.doExecute(OIMConfigManager.java:583)
      at oracle.as.install.engine.modules.configuration.client.ConfigAction.execute(ConfigAction.java:371)
      at oracle.as.install.engine.modules.configuration.action.TaskPerformer.run(TaskPerformer.java:88)
      at oracle.as.install.engine.modules.configuration.action.TaskPerformer.startConfigAction(TaskPerformer.java:105)
      at oracle.as.install.engine.modules.configuration.action.ActionRequest.perform(ActionRequest.java:15)
      at oracle.as.install.engine.modules.configuration.action.RequestQueue.perform(RequestQueue.java:64)
      at oracle.as.install.engine.modules.configuration.standard.StandardConfigActionManager.start(StandardConfigActionManager.java:160)
      at oracle.as.install.engine.modules.configuration.boot.ConfigurationExtension.kickstart(ConfigurationExtension.java:81)
      at oracle.as.install.engine.modules.configuration.ConfigurationModule.run(ConfigurationModule.java:86)
      at java.lang.Thread.run(Thread.java:662)
    Caused by: java.lang.Exception: Exception occured while encrypting the database
      at oracle.as.install.oim.config.util.EncryptDataBase.encryptDBContent(EncryptDataBase.java:159)
      at oracle.as.install.oim.config.util.EncryptConfigurationAndDB.encryptConfigurationAndDatbase(EncryptConfigurationAndDB.java:230)
      ... 12 more
    Caused by: java.lang.Exception: Exception occured in updateMLSLocale method while updating Locale to OIM DB
      at oracle.as.install.oim.config.util.EncryptDataBase.updateMLSLocale(EncryptDataBase.java:318)
      at oracle.as.install.oim.config.util.EncryptDataBase.encryptDBContent(EncryptDataBase.java:125)
      ... 13 more
    Caused by: java.sql.SQLIntegrityConstraintViolationException: ORA-00001: unique constraint (DEV_OIM.UK_MLS_LOCALE_MLS_LOCALE_CODE) violated
      at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:462)
      at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:405)
      at oracle.jdbc.driver.T4C8Oall.processError(T4C8Oall.java:931)
      at oracle.jdbc.driver.T4CTTIfun.receive(T4CTTIfun.java:481)
      at oracle.jdbc.driver.T4CTTIfun.doRPC(T4CTTIfun.java:205)
      at oracle.jdbc.driver.T4C8Oall.doOALL(T4C8Oall.java:548)
      at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:217)
      at oracle.jdbc.driver.T4CPreparedStatement.executeForRows(T4CPreparedStatement.java:1115)
      at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1488)
      at oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3769)
      at oracle.jdbc.driver.OraclePreparedStatement.executeUpdate(OraclePreparedStatement.java:3904)
      at oracle.jdbc.driver.OraclePreparedStatementWrapper.executeUpdate(OraclePreparedStatementWrapper.java:1512)
      at oracle.as.install.oim.config.util.EncryptDataBase.updateMLSLocale(EncryptDataBase.java:310)
      ... 14 more

    Hi
    I faced this issue before ,Reinstall is the option you have .Verify the version of RCU before you start creating schema .Set all Pre DB setting ,hostname and IP Address ,If DB and OIM  are in ifferent machines check pinging from both the sides .
    Please Drop all OLD schema ,Create a New "Prefix" for fresh installation , don't use old schema .
    Let me know .
    Thanks,
    Ari

  • Can't use Database Configuration Assistants and Net8.

    I installed 8.1.5 on win2000, in process of installation of Configuration of Net8 and Database Assistants are fail and error message is no class.
    Please help me how to solve this problem.
    thanks

    Hi,
    Please refer the note . 77627.1 through Metalink Advanced Search for the Oracle
    Server support matrix on Windows 2000,which mentions that 8.1.5 is not supported with
    Win 2000 and only from 8.1.6 it is supported.
    Regards
    Anand

Maybe you are looking for