Oracle dump transaction log like Sybase

Hello
In Sybase we use the dump transaction log command for the needed databases that dumps whatever is in the transaction log of that db to file system, after that dump file is copied over to another database server and loads the other database, this is like a replication
Does oracle have something similar to this?
Many thanks

the closest you'll get to what you are looking for is a dataguard setup. Dataguard will manage the shipment of the logs on it's own and keep things in sync.
There is another option of Oracle Advanced Replication but you will need to review if that really meets your needs. The following is simply an overview of it:
http://www.orafaq.com/wiki/Advanced_Replication_FAQ
Or GoldenGate:
http://www.oracle.com/technetwork/middleware/goldengate/overview/index.html
Edited by: rui_catcuddler on Oct 20, 2010 7:54 AM

Similar Messages

  • Oracle 9i and Oracle 10g transaction log shipping

    Hi,
    We have Oracle 9i and we use the transaction log shipping mechanism to transport our transaction log files to our DR site. We then import these against the database at the remote site.
    Our telecomms infrastructure is pretty limited so this is a manual process.
    We are looking at upgrading to 10G. I believe that with 10g you have to use dataguard, or is there a way to mimic the behavior of 9i that would allow us to transport and apply the transaction logs manually?
    Thanks
    Andrew

    You can try setting the SGA to a low value and bring up both the databases. I don't think it should be too slow provided you are not running other windows programs.
    If you are really interested in trying out new products you can also explore the option of installing VMware, creating virtual machines & installing Linux, and then playing with the different Oracle products. Doing this will at least keep your main windows operating system clean.
    You may want to check out my blog post on Build your own oracle test lab
    Cheers !!!
    Ashish Agarwal
    http://www.asagarwal.com

  • Cannot write to transaction log "C:\Program Files (x86)\SAP BusinessObjects\sqlanywhere\database\BI4_Audit.log

    Hi friends,
    My server Intelligence Agent (SIA) can not start because the database service "SQLAnywhereForBI" can't start also. I got the following error :
    "I . 08/09 20:35:06. A read failed with error code: (1392), Le fichier ou le répertoire est endommagé et illisible.
    E. 08/09 20:35:06. Fatal error:  cannot write to transaction log "C:\Program Files (x86)\SAP BusinessObjects\sqlanywhere\database\BI4_Audit.log"
    E. 08/09 20:35:06. unable to start database "C:\Program Files (x86)\SAP BusinessObjects\sqlanywhere\database\BI4_CMS.db"
    E. 08/09 20:35:06. Error writing to transaction log file
    I. 08/09 20:35:06. Database server shutdown due to startup error "
    inside the database log file.
    Please, can you help me

    I found the solution by following the advice given on the following forum:
    http://evtechnologies.com/transaction-logs-on-sybase-sql-anywhere-and-sap-​​businessobjects-bi-4-1
    In fact, I crushed the BI4_Audit.db and BI4_Audit.log files and I replaced with others that I got from another machine where I installed BO again and where the files are not corrupted . After I logged in to the CMS database by executing the command in the command line:
    dbisql -c "UID = DBA; PWD = mypassword; BI4 Server =; DBF = C: \ Program Files (x86) \ SAP BusinessObjects \ sqlanywhere \ database \ BI4_CMS.db."
    Once connected, I start the command:
    alter database 'C: \ Program Files (x86) \ SAP BusinessObjects \ sqlanywhere \ database \ BI4_Audit.db' alter log off;
    The query runs successfully.
    And that's good, I can be connected to BO smoothly.
    Thank you again Eric

  • Exchange 2010 DAG backup & Transaction logs

    Hi, 
    What is Microsoft recommended best practise for Exchange DAG group backup in an environment where there are Active & multiple (2-3) passive copies of the databases?
    Is it a good practice to backup transaction logs as frequently as possible in addition to daily full backup ? This I belive will allow to restore the DB to the latest possible state
    using last good full backup & transaction logs (Like restoring SQL databases)
    Thanks

    Hi,
    Windows Server Backup can't backup passive copy. If you want to backup both active and passive copies, you need to use DPM or other third party.
    Here is a similar thread for your reference.
    Exchange 2010 DAG Backup Best Practices
    http://social.technet.microsoft.com/Forums/exchange/en-US/269c195f-f7d7-488c-bb2e-98b98c7e8325/exchange-2010-dag-backup-best-practices
    Besides, here is a related blog below which may help you.
    Backup issues and limitations with Exchange 2010 and DAG
    http://blogs.technet.com/b/ehlro/archive/2010/02/13/backup-issues-and-limitations-with-exchange-2010-and-dag.aspx
    Hope this helps.
    Best regards,
    Belinda
    Belinda Ma
    TechNet Community Support

  • Restore ASE - Dump of open transaction-log required?

    Hi experts,
    I am still doing some restore tests.
    What about the following case.
    Last transaction log was done at 1 o'clock and next will be at 4 o'clock.
    At 3 o'clock we detect that we have to restore two 2 o'clock.
    So for this restore, I need the transaction log which isn't dumped yet.
    My question is, do I have to dump the current transaction log also to a file for the restore procedure?
    Or is there another way to included in the restore the current-log file?
    In other words, when will the log-file be touched first?
    After "online database" command?
    If so, I can also do the restore using the original-logfile, right?
    Kind regards

    Christian,
    You are right.
    Let me tell you what is the practice I reccommend to follow:
    1. Take full backup daily during your off business hours if you have the infrastructure like tape/disk or SAN and the data is very critical may be production or development
    2 During the business hours take hourly or half an hour once take the transaction backup may be between 9-6 as per your time zone :)
    3 This mostly helps you to minimise the tran log loss.
    4 As you have the week end reorg and update stats running I prefer just before start of production hours on Monday take a full backup and keep it safe so that the data is super clean and secure
    If there is any confusion let me know I will explain you still clearly in simpler words
    PS:One full backup per day is fine if you can preserve and retain for 7-10 days and delete it later point if you don't need it and don't have infrastructure and disk cost problems :P ;)
    Cheers
    Kiran K Adharapuram

  • Oracle DB equivalent of SQL Server's Simple Transaction Logging mode?

    G'Day Experts !
    Was wondering if Oracle DB has the functional equivalent of the 'simple' transaction logging available in SQL Server?
    Would this be availabe at the schema level, or would it have to be the entire instance?
    I'm asking because the WebCenter Interaction portal and related services has no practical use for point-in-time rollbacks. The portal uses discreet event boundaries which unfortunately do not map into the relational world.
    Thanks!
    Rob in Vermont

    Plumtree wrote:
    G'Day Experts !
    Was wondering if Oracle DB has the functional equivalent of the 'simple' transaction logging available in SQL Server?
    Would this be availabe at the schema level, or would it have to be the entire instance?
    I'm asking because the WebCenter Interaction portal and related services has no practical use for point-in-time rollbacks. The portal uses discreet event boundaries which unfortunately do not map into the relational world.
    Thanks!
    Rob in VermontHi Rob
    I assume you are referring to the simple recovery model, i.e lose everything since last backup. Oracle's equivalent of that is to run a database in NOARCHIVELOG mode. It applies to the database rather than the instance, though you probably intended database when you said instance.
    Niall Litchfield
    http://www.orawin.info/

  • ASE - Dump Database  (Transaction Log deleted?)

    Hi experts,
    I have a question.
    Will the transaction log be deleted reseted to the time of a database dump is finished?
    Hopefully it will only be deleted by dumping the transaction log itself, right?
    Kind regards

    Hello Christian,
    Don't worry we have set the options to the databases which are getting installed through business suites.
    However I will provide you the detailed dB details and their respective option which yoibcan cross verify as well from the isql command
    sp_helpdb
    go
    And you can very well enable or disable the options set as well using
    sp_dboption "<SID>","<option_name>",true/false
    go
    Also if you needed the detailed dB wise default dB options that to be set let me know I will attach the details as well ;)
    Regards
    Kiran K Adharapuram

  • Is the transaction log is changed when i execute select statemetns

    hello
    i want to ask about updateing the transaction log: is it modified when a select statement is executed
    is it changed when update,delete or insert statements are executed ?

    In general SELECT statement don't write redo but if you use SELECT with FOR UPDATE clause, then redo is generated and written to redo logs or in special cases like delayed block cleanout. INSERT, UPDATE and DELETE write redo, yes.
    Note that Oracle right name to use is redo log and not transaction log which a Sybase/MS SQL Server concept.

  • Audit Vault 12.1.1 error creating audit trail with TRANSACTION LOG

    Hi,
    i installed AV 12.1.1 , the DB target is with Data Guard.
    when i run the script oracle_user_setup with the mode REDO_COLL the final message is that was succesfull , but when i go to the AV console and try to create an audit trail with TRANSACTION LOG the AV console shows me an error and the log shows me this :
    [2013-10-16T03:37:18.593-05:00] [collfwk] [ERROR] [] [] [tid: 10] [ecid: 192.168.56.8:78800:1381912639433:0,0] RedoCollector : runSourceScript : Error while running script on source for REDO collector.
    [2013-10-16T03:37:19.528-05:00] [collfwk] [ERROR] [] [] [tid: 10] [ecid: 192.168.56.8:78800:1381912639433:0,0] OAV-8004: Failed to start collector {0}:{1}CollectionFactory : createCollection : Exception while creating collection. [[
    Failed to start collector {0}:{1}
                    at oracle.av.platform.agent.collfwk.impl.redo.RedoCollector.runSourceScript(RedoCollector.java:816)
                    at oracle.av.platform.agent.collfwk.impl.redo.RedoCollector.sourceSetup(RedoCollector.java:579)
                    at oracle.av.platform.agent.collfwk.impl.redo.RedoCollector.setup(RedoCollector.java:454)
                    at oracle.av.platform.agent.collfwk.impl.redo.RedoCollector.startCollector(RedoCollector.java:216)
                    at oracle.av.platform.agent.collfwk.impl.redo.RedoCollectorManager.startTrail(RedoCollectorManager.java:199)
                    at oracle.av.platform.agent.collfwk.impl.factory.CollectionFactory.createCollection(CollectionFactory.java:504)
                    at oracle.av.platform.agent.collfwk.impl.factory.CollectionFactory.createCollection(CollectionFactory.java:354)
                    at oracle.av.platform.agent.StartTrailCommandHandler.processMessage(StartTrailCommandHandler.java:63)
                    at oracle.av.platform.agent.AgentController.processMessage(AgentController.java:325)
                    at oracle.av.platform.agent.AgentController$MessageListenerThread.run(AgentController.java:1859)
                    at java.lang.Thread.run(Thread.java:679)
    Nested Exception:
    java.sql.SQLSyntaxErrorException: ORA-01031: insufficient privileges
    ORA-06512: at line 1
                    at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:445)
                    at oracle.jdbc.driver.T4CTTIoer.processError(T4CTTIoer.java:396)
                    at oracle.jdbc.driver.T4C8Oall.processError(T4C8Oall.java:879)
                    at oracle.jdbc.driver.T4CTTIfun.receive(T4CTTIfun.java:450)
                    at oracle.jdbc.driver.T4CTTIfun.doRPC(T4CTTIfun.java:192)
                    at oracle.jdbc.driver.T4C8Oall.doOALL(T4C8Oall.java:531)
                    at oracle.jdbc.driver.T4CPreparedStatement.doOall8(T4CPreparedStatement.java:207)
                    at oracle.jdbc.driver.T4CPreparedStatement.executeForRows(T4CPreparedStatement.java:1044)
                    at oracle.jdbc.driver.OracleStatement.doExecuteWithTimeout(OracleStatement.java:1329)
                    at oracle.jdbc.driver.OraclePreparedStatement.executeInternal(OraclePreparedStatement.java:3584)
                    at oracle.jdbc.driver.OraclePreparedStatement.execute(OraclePreparedStatement.java:3685)
                    at oracle.jdbc.driver.OraclePreparedStatementWrapper.execute(OraclePreparedStatementWrapper.java:1376)
                    at sun.reflect.GeneratedMethodAccessor9.invoke(Unknown Source)
                    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
                    at java.lang.reflect.Method.invoke(Method.java:616)
                    at oracle.ucp.jdbc.proxy.StatementProxyFactory.invoke(StatementProxyFactory.java:230)
                    at oracle.ucp.jdbc.proxy.PreparedStatementProxyFactory.invoke(PreparedStatementProxyFactory.java:124)
                    at $Proxy2.execute(Unknown Source)
                    at oracle.av.platform.agent.collfwk.impl.redo.RedoCollector.runSourceScript(RedoCollector.java:747)
                    at oracle.av.platform.agent.collfwk.impl.redo.RedoCollector.sourceSetup(RedoCollector.java:579)
                    at oracle.av.platform.agent.collfwk.impl.redo.RedoCollector.setup(RedoCollector.java:454)
                    at oracle.av.platform.agent.collfwk.impl.redo.RedoCollector.startCollector(RedoCollector.java:216)
                    at oracle.av.platform.agent.collfwk.impl.redo.RedoCollectorManager.startTrail(RedoCollectorManager.java:199)
                    at oracle.av.platform.agent.collfwk.impl.factory.CollectionFactory.createCollection(CollectionFactory.java:504)
                    at oracle.av.platform.agent.collfwk.impl.factory.CollectionFactory.createCollection(CollectionFactory.java:354)
                    at oracle.av.platform.agent.StartTrailCommandHandler.processMessage(StartTrailCommandHandler.java:63)
                    at oracle.av.platform.agent.AgentController.processMessage(AgentController.java:325)
                    at oracle.av.platform.agent.AgentController$MessageListenerThread.run(AgentController.java:1859)
                    at java.lang.Thread.run(Thread.java:679)
    i don't understand why the issue because the user has the privileges given by the script and i tried with grant as sysdba but without any result
    i don't understand what are the privileges that the collector needs.
    any idea?
    thnks for any help

    Hi
    Just run the script $AV_AGENT/av/plugins/com.oracle.av.plugin.oracle/config/oracle_user_setup.sql  USER_NAME REDO_COLL
    This will grant the user some privileges and roles like DBA and CREATE Database Link
    I hope this answer your question
    Thanks
    Ahmed Moustafa

  • Viewing a transaction log for a query

    Hi all,
    I have an application which has a query in some functionality of it.
    I want to check how many times in a month that query is referenced or used.
    I hope oracle maintains a transaction log or something of this sort.
    Is there a way to analyze this?
    Thanks in advance.
    BRK

    > I want to check how many times in a month that query is referenced or used.
    Simple answer, no.
    Yes, Oracle do have a transaction (redo and undo) logs. Yes, Oracle has advance features like flashback queries. But keeping track on how many times a query (SQL) has been used in a month? Not the best of ideas. I have SQLs that are run over a billion times per month. I do not have the storage space for Oracle to keep track on just when these were run, with that values, what the performance were, etc.
    Remember that on a busy system that runs 1000's of SQLs per second, there is no time to waste on maintaining something like a query log.
    > Is there a way to analyze this
    Simple answer, yes. The SQL Shared Pool will tell you how many times a query (cursor) in the pool has been executed - assuming that it has not been aged out of the pool as being "old and cold". AWR reporting can be used. Etc.
    The real question is what do you want to achieve with this analysis? The number of times a SQL is executed is meaningless on its own - additional measures are required for meaningful analysis.

  • Oracle dump file

    Hi all,
    I have tried do a full export in oracle 10g2 in server A then follow by a full import do another 10g2 in server B successfully.
    But when i insert a row the10g2 server A and do a full export, follow by bringing the dump file to the database in server B and do full import again using the dump file. I found out that the row inserted does not reflect in the database in server B. Is there any way i can do so that the inserted row can be reflected in serverB?
    imp username/password file=dump.dmp full=y log=import.log
    Thanks
    cheng

    thanks hemant,
    ignore=y will ignore the prensence of the table. I tried it out, apprantly it run slowly than i do the inital full import.
    Other than ignore=y, is there other way of append the data?
    Regarding this: To be safe you could either
    a. TRUNCATE all the tables in the target "B" database
    OR
    b. DROP all the tables in the target "B" database (or drop the whole schema)
    before you begin the import.
    Sorry i quite new to Oracle database.
    b.) there a command to drop the whole schema? so that i can do a full import.
    a.) Honestly, i also do not know what is truncate, maybe i can go find out from google.
    Regarding this: NOTE : Since you are copying specific schemas, you don't need to be using the FULL=Y option but OWNER= in the Export and, optionally, FROMUSER=, TOUSER= in the Import.
    So instead of of a full export/import. i should do sth like this.
    exp username/password file=dump.dmp log=export.log owner=username
    imp username/password file=dump.dmp log=import.log fromuser=username, touser=username
    eventhough the username is the same for both the database?

  • Oracle 9i Transaction Performance (Real World vs. TPC)?

    Hello:
    I was curious if anybody reading this forum has a server/disk architecture that has achieved over 10,000 Oracle Database transactions (with logging enabled) per second?
    If so, can you please supply me with the clustered or standalone server and disk farm (NAS, SAN, DAS) details.
    Thanks,
    Robert
    [email protected]
    310-536-0018 x124
    http://www.imperialtech.com

    Hi,
    have a look at http://publications.uu.se/theses/.
    This site is now up for nearly one year. We're using Oracle 8.1.7, Apache 1.3.24 and Tomcat 4.0.3. We've got about 4500 hits per day with a maximum up to 9000 hits.
    Regards Uwe Hi there Uwe,
    Given you are running this in a production environment, are you running the XSQLRequest object or are you running the servlet directly. Running the servlet directly in particular, is just not production ready, unless I am missing something.
    1. There are no direct filesystem logging capabilities in the servlet. Doing inserts into a db table for the purposes of logging, is not adequate as the critical points in the servlet operation is either connecting to the database or in the post query phase during the xml transform.
    2. Errors like invalid db connections, timeouts, invalid xsql/xsl files, ie All xsql-xxxx errors other than query errors are not trappable from the servlet either via xml or otherwise. They get written directly to the Output stream as text.
    Using the XSQLRequest object programmatically, we can work around some of these limitations, but that defeats the purpose and the otherwise ease of use of this servlet in a XML envirnoment.
    - Manish

  • Transaction logs off on  MSSQL Server-2008

    Hi,
      I want to off the transaction logs in mssql server-2008 like archive mode off in oracle,auto over write on in maxdb database.
    Thanku

    You cannot stop.!
    Please read sql Architecture.
    The transaction log of the SAP database records all changes made to the database. It may never be deleted and must be backed up separately. Transaction log backups save the log files. They are mandatory when you use the Full or Bulk-Logged Recovery Model since they are needed.
    Read this link.
    [http://help.sap.com/saphelp_nwmobile71/helpdata/en/f2/31ad41810c11d288ec0000e8200722/content.htm]
    Thanks,
    Siva

  • Turn off the Transaction log

    All,
    My client copy is really big.
    How can I switch off the Transaction logs in SQL server whie doing client copy.
    I DO NOT want Trans log keep increasing.
    In oracle, I can turn off the archive logs.
    Please advise.
    Thanks again and hope to hear from you soon.
    From,
    PT.

    Hello,
    the log backup on simple mode is not necessary. The full database after switching back to full is a must.
    Please keep in mind, that even running on simple mode the log can be filled up, as all transactions are still written to the log. Commited transaction then can truncated from the log. But when you run a hugh transaction like a client copy, the log might grow as well. The log will be freed once the transaction commits or rolls back. And no, you can't split a client copy in several transactions.
    Best regards
      Clas

  • System Crash after transactional log filled filesystem

    Dear gurus,
    We have an issue in our PRD system under FlexFrame platform. We SAP NW 7.4 (SP03) with ASE 15.7.0.042 (SuSe SLES 11 SP1) running as BW system.
    While uploading data from ERP system, the transactional log was filled. We can see in <SID>.log:
    Can't allocate space for object 'syslogs' in database '<SID>' because 'logsegment' segment is full/has no free extents. If you ran out of space in syslogs, dump the transaction log. Otherwise, use ALTER DATABASE to increase the size of the segment.
    After this, we increase the transactional log (disk resize). Then, executed ALTER DATABASE <SID> log on <LOGDEVICE> = '<size>'
    While ALTER is running the log filesystem was filled (100%), after this, <SID>.log began to grow tremendously.
    We stopped Sybase and now, when we try start it all FF node will be down. The filesystem has free space (around 10 GB)
    Could you help us?
    Add: We think that a posible solution could be to delete the transactional log due to the fact that we understand that the failure is related to this log (maybe corrupted?)
    Regards

    ====================
    00:0008:00000:00009:2014/06/26 15:49:37.09 server  Checkpoint process detected hardware error writing logical page '2854988', device 5, virtual page 6586976 for dbid 4, cache 'log cache'. It will sleep until write completes successfully.
    00:0010:00000:00000:2014/06/26 15:49:37.10 kernel  sddone: write error on virtual disk 5 block 6586976:
    00:0010:00000:00000:2014/06/26 15:49:37.10 kernel  sddone: No space left on device
    00:0008:00000:00009:2014/06/26 15:49:37.10 server  bufwritedes: write error detected - spid=9, ppage=2854988, bvirtpg=(device 5, page 6586976), db id=4
    =======================
    1 - check to make sure the filesystem that device #5 (vdevno=5) sits on is not full; make sure filesystem is large enough to hold the entire defined size of device #5; make sure no other processes are writing to said filesystem
    2 - have your OS/disk admin(s) make sure the disk fragment(s) underlying device #5's filesystem isn't referenced by other filesystems and/or raw device definitions

Maybe you are looking for