Oracle Streams question

Can someone direct me to where I might find a discussion forum on Oracle Streams? thank you.

You can go to OTN-->Discussion Forum-->search for a "keyword". When you see the list of posted questions, click on "Go to top of list" and you'll find:
Streams
Good luck.
cb

Similar Messages

  • Complex Oracle Streams issue - Update conflicts

    This is for Oracle streams replication on 11g r2.
    I am facing update conflicts in a table. The conflict arise due to technical and business logic issue. The business logic will pass through the replication/apply process successfully but we want to arrest and resolve it before replication for our requirements. These are typically a bit complex cases and we are exploring the possibility of having both DML handlers and Error handlers. The DML handlers will take care of business logic conflicts and Error handler for technical issues before pushing it to Error queue by Streams. Based on our understanding and verification, we found a limitation to configure both procedure DML handler and Error handler together for the same table operation.
    Statement handlers can not be used for our conflict scenarios.
    Following are my questions:
    1. Have anyone implemented or faced such a scenario in their real time application? If yes, can you please share some insights or inputs?
    2. Is there a custom way to handle this complex problem of configuring both DML and Error handler?
    3. Is there any alternative possible way to resolve this situation at Oracle streams environment with other handlers?

    Dear All
    I too have a similar requirement. Could anyone help with one?
    We can handle the error-ing transactions via Error Handler procedures.
    But we can not configure the DML handler procedure for transactions that are successfully replicated. STreams does not allow us to configure a handler for this. Is there any other handler / procedures / hooks in streams where we can implement the desired functionality - which includes changing the values in the LCR before invoking lcr.execute() and we should be able to discard the LCR also if required.
    Regards
    Velmurugan
    Edited by: 982387 on Jan 16, 2013 11:25 PM
    Edited by: 982387 on Jan 16, 2013 11:27 PM

  • Oracle streams configuration

    Hi,
    Our organization is planning to implement Oracle streams. I have couple of fundamental questions:
    1. Can you configure Oracle Streams while the source database is up and running?
    2. I think the answer is yes, but please confirm that the LCR data can be extracted and transformed before applying to the target
    3. Is there any performance impact on the source database if Oracle Streams is enabled? If yes, how much?
    That’s all for now…
    Thanks…

    See my answers inline.
    1. Can you configure Oracle Streams while the source database is up and running?
    YES. Your database must be up and running to create streams processes.
    2. I think the answer is yes, but please confirm that the LCR data can be extracted and transformed before applying to the target
    You have to instantiate the target database so that there source and target are in sync.
    3. Is there any performance impact on the source database if Oracle Streams is enabled? If yes, how much?
    There is a minimal overhead of running the capture process on the source. But this should be negligable. This also depends upon what is being streamed i.e. LOB's are being streamed, high transaction volume etc.

  • Oracle Streams 10gR2 Schedule Propagation or Application

    Hi all,
    Is there a way to schedule the propagation and application when configuring Oracle Streams?, I mean, I don't want to execute online replication because I have other object outside Oracle that need to replicate withing the db to have my aplication (red only) in sync.
    So, where I can do that.
    BEGIN
    DBMS_STREAMS_ADM.ADD_SCHEMA_RULES(
    schema_name => 'shm',
    streams_type => 'apply',
    streams_name => 'apply_from_db1',
    queue_name => 'strmadmin.from_db1',
    include_dml => true,
    include_ddl => true,
    source_database => 'db1.world',
    inclusion_rule => true);
    END;
    BEGIN
    DBMS_STREAMS_ADM.ADD_SCHEMA_PROPAGATION_RULES(
    schema_name => 'shm',
    streams_name => 'db1_to_db2',
    source_queue_name => 'strmadmin.captured_db1',
    destination_queue_name => '[email protected]',
    include_dml => true,
    include_ddl => true,
    source_database => 'db1.world',
    inclusion_rule => true,
    queue_to_queue => true);
    END;
    /

    Hello, Am I not being clear with my question?. Or maybe I'm doing everything in the wrong way about Stream propagation and application processes?.
    I looked into all documentation available and I can't find how to schedule the processes ... what's the part I'm missundertanding?.
    AS long I have other non Oracle objects (file system objects, we say some ECM file system objects) to replicate with the data schema, I can't replicate (automatically) all the changes occurred at source table's schema to the destiny database schema. So, I'm replicating ECM file systems objects within other external tool, and It has Windows schedule integration ... but how I can schedule propagation and/or application processes in the a 2-way Streams environment.
    Please, I need a hint from somebody.
    Thanks.

  • Oracle Stream Replication 10G

    Hi.
    We are planing to use 10G stream replication for our production enviorment. But we are concenred over few things before we move ahead with option. Our prior experience with 9i stream replication was not that good.
    9i does'nt functionality to do some advance maintenance task that could required on production enviorment. We are hoping 10G would be better option to go with streams.
    However, following are few questions that we have regarding this product. Request your valuable inputs on Oracle 10G streams replication.
    -->How effective this tool is compared to prior version 9i.
    -->How stable Streams are with 10G? Are still there any know critical issues with stream replication 10G for e.g bugs or patches that are under review e.g?
    -->For replication of production enviorment, would it be an ideal choice compared to third party vendor like golden Gate or shareplex.
    You reponse would be greatly appreciated.
    Thnks
    Ratheesh

    Hi Rateesh,
    -->How effective this tool is compared to prior version 9i.Very good. I use Streams in many production environments, even for bidirectional replication. Streams is very stable in 10g.
    For replication of production enviorment, would it be an ideal choice compared to third party vendor like golden Gate or shareplex.It depends on several factors. It's better than GG and Shareplex by far (IMHO).
    Madhu Tumma has some great Streams notes on this in his book "Oracle Streams":
    http://www.rampant-books.com/book_2004_2_streams.htm
    Hope this helps. . . .
    Donald K. Burleson
    Oracle Press author

  • Oracle stream two ways replication

    Hi all,
    I have a primary database and standby database. I'm using Oracle Stream for replication.
    I have a question.
    Can my standby database becomes primary and primary database becomes standby?
    Thank you!
    Dan.

    Thanks Arun,
    I've configured Bi-directional Replication on DB-11G-R2 by command and configured 1 way replication on Enterprise Manager but i failed to configure Bi-directional Replication on Enterprise Manager.
    Do you have document that guides steps by steps configuration Bi-directional Replication on Enterprise Manager?
    Brds,
    Dan.

  • Oracle streams running problems

    Hello
    Can anybody tell me pls, what was wrong ?.
    i just created the oracle streams on the both sides ( db1 und db2 ) via example on the metalinks. It worked , everthing looks well. But in the next day, after i updated a row on the source side, it 's been not transfern to another side.
    In the streams_queue of the scheme strmadmin the start_time still stayed in the past. There is not the currenday . ??
    Do i missed something ?
    thk
    duc

    Please ask your question in oracle streams forum.
    Streams

  • Support oracle stream feature on 12c

    Oracle doc says the oracle stream on 12c is deprecated, but it doesn't say desupported. My question is "still can I use oracle stream on 12c and get tech support from oracle"?

    Yes, all Oracle databases participating in an Oracle Streams environment must be at Oracle9i Release 2.

  • Oracle Streams and CLOB column

    Hi there,
    We are using "Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit". My question is "Does Oracle Streams captures, propagates (source capture method) and applies CLOB column changes?"
    If yes, is this default behavior? Can we tell Streams to exclude the CLOB column from the whole (capture-propage-apply) process?
    Thanks in advance!

    You can exclude columns via a rule (dbms_streams_adm.delete_column).
    CLOBs are captured.
    http://download.oracle.com/docs/cd/E11882_01/server.112/e17069/strms_capture.htm#i1006263

  • Oracle Streams b/w MS-Access 2007 and Oracle 10g.

    Can we set up Oracle Streams between MS-Access 2007 and Oracle 10g? Ms-Access as source and Oracle 10g as destination database. If so, can any one please give me little heads up with supported doc's or any source of info.

    Help Help....!!!

  • Doubt in Oracle streams

    Doubt in Oracle streams .Can you please help me in understanding the terms
    1. Message
    2.User defined Event
    3. Event
    4.Rules
    5.Oracle supplied PL/SQL packages
    6.Subscriber,Consumer

    Hi
    Message
    A message is the smallest unit of information that is inserted into and retrieved from a queue.
    Queue
    A queue is repository for messages. Queues are stored in queue tables
    Enqueue
    To place a message in queue
    Dequeue
    To comsume a message
    Agent
    An agent is a end user or the application uses a queue
    Thanks
    Venkat

  • Help on Oracle streams 11g configuration

    Hi Streams experts
    Can you please validate the following creation process steps ?
    What is need to have streams doing is a one way replication of the AR
    schema from a database to another database. Both DML and DDL shall do
    the replication of the data.
    Help on Oracle streams 11g configuration. I would also need your help
    on the maintenance steps, controls and procedures
    2 databases
    1 src as source database
    1 dst as destination database
    replication type 1 way of the entire schema FaeterBR
    Step 1. Set all databases in archivelog mode.
    Step 2. Change initialization parameters for Streams. The Streams pool
    size and NLS_DATE_FORMAT require a restart of the instance.
    SQL> alter system set global_names=true scope=both;
    SQL> alter system set undo_retention=3600 scope=both;
    SQL> alter system set job_queue_processes=4 scope=both;
    SQL> alter system set streams_pool_size= 20m scope=spfile;
    SQL> alter system set NLS_DATE_FORMAT=
    'YYYY-MM-DD HH24:MI:SS' scope=spfile;
    SQL> shutdown immediate;
    SQL> startup
    Step 3. Create Streams administrators on the src and dst databases,
    and grant required roles and privileges. Create default tablespaces so
    that they are not using SYSTEM.
    ---at the src
    SQL> create tablespace streamsdm datafile
    '/u01/product/oracle/oradata/orcl/strepadm01.dbf' size 100m;
    ---at the replica:
    SQL> create tablespace streamsdm datafile
    ---at both sites:
    '/u02/oracle/oradata/str10/strepadm01.dbf' size 100m;
    SQL> create user streams_adm
    identified by streams_adm
    default tablespace strepadm01
    temporary tablespace temp;
    SQL> grant connect, resource, dba, aq_administrator_role to
    streams_adm;
    SQL> BEGIN
    DBMS_STREAMS_AUTH.GRANT_ADMIN_PRIVILEGE (
    grantee => 'streams_adm',
    grant_privileges => true);
    END;
    Step 4. Configure the tnsnames.ora at each site so that a connection
    can be made to the other database.
    Step 5. With the tnsnames.ora squared away, create a database link for
    the streams_adm user at both SRC and DST. With the init parameter
    global_name set to True, the db_link name must be the same as the
    global_name of the database you are connecting to. Use a SELECT from
    the table global_name at each site to determine the global name.
    SQL> select * from global_name;
    SQL> connect streams_adm/streams_adm@SRC
    SQL> create database link DST
    connect to streams_adm identified by streams_adm
    using 'DST';
    SQL> select sysdate from dual@DST;
    SLQ> connect streams_adm/streams_adm@DST
    SQL> create database link SRC
    connect to stream_admin identified by streams_adm
    using 'SRC';
    SQL> select sysdate from dual@SRC;
    Step 6. Control what schema shall be replicated
    FaeterBR is the schema to be replicated
    Step 7. Add supplemental logging to the FaeterBR schema on all the
    tables?
    SQL> Alter table FaeterBR.tb1 add supplemental log data
    (ALL) columns;
    SQL> alter table FaeterBR.tb2 add supplemental log data
    (ALL) columns;
    etc...
    Step 8. Create Streams queues at the primary and replica database.
    ---at SRC (primary):
    SQL> connect stream_admin/stream_admin@ORCL
    SQL> BEGIN
    DBMS_STREAMS_ADM.SET_UP_QUEUE(
    queue_table => 'streams_adm.FaeterBR_src_queue_table',
    queue_name => 'streams_adm.FaeterBR_src__queue');
    END;
    ---At DST (replica):
    SQL> connect stream_admin/stream_admin@STR10
    SQL> BEGIN
    DBMS_STREAMS_ADM.SET_UP_QUEUE(
    queue_table => 'stream_admin.FaeterBR_dst_queue_table',
    queue_name => 'stream_admin.FaeterBR_dst_queue');
    END;
    Step 9. Create the capture process on the source database (SRC).
    SQL> BEGIN
    DBMS_STREAMS_ADM.ADD_SCHEMA_RULES(
    schema_name =>'FaeterBR',
    streams_type =>'capture',
    streams_name =>'FaeterBR_src_capture',
    queue_name =>'FaeterBR_src_queue',
    include_dml =>true,
    include_ddl =>true,
    include_tagged_lcr =>false,
    source_database => NULL,
    inclusion_rule => true);
    END;
    Step 10. Instantiate the FaeterBR schema at DST. by doing export
    import : Can I use now datapump to do that ?
    ---AT SRC:
    exp system/superman file=FaeterBR.dmp log=FaeterBR.log
    object_consistent=y owner=FaeterBR
    ---AT DST:
    ---Create FaeterBR tablespaces and user:
    create tablespace FaeterBR_datafile
    '/u02/oracle/oradata/str10/FaeterBR_01.dbf' size 100G;
    create tablespace ws_app_idx datafile
    '/u02/oracle/oradata/str10/FaeterBR_01.dbf' size 100G;
    create user FaeterBR identified by FaeterBR_
    default tablespace FaeterBR_
    temporary tablespace temp;
    grant connect, resource to FaeterBR;
    imp system/123db file=FaeterBR_.dmp log=FaeterBR.log fromuser=FaeterBR
    touser=FaeterBR streams_instantiation=y
    Step 11. Create a propagation job at the source database (SRC).
    SQL> BEGIN
    DBMS_STREAMS_ADM.ADD_SCHEMA_PROPAGATION_RULES(
    schema_name =>'FaeterBR',
    streams_name =>'FaeterBR_src_propagation',
    source_queue_name =>'stream_admin.FaeterBR_src_queue',
    destination_queue_name=>'stream_admin.FaeterBR_dst_queue@dst',
    include_dml =>true,
    include_ddl =>true,
    include_tagged_lcr =>false,
    source_database =>'SRC',
    inclusion_rule =>true);
    END;
    Step 12. Create an apply process at the destination database (DST).
    SQL> BEGIN
    DBMS_STREAMS_ADM.ADD_SCHEMA_RULES(
    schema_name =>'FaeterBR',
    streams_type =>'apply',
    streams_name =>'FaeterBR_Dst_apply',
    queue_name =>'FaeterBR_dst_queue',
    include_dml =>true,
    include_ddl =>true,
    include_tagged_lcr =>false,
    source_database =>'SRC',
    inclusion_rule =>true);
    END;
    Step 13. Create substitution key columns for äll the tables that
    haven't a primary key of the FaeterBR schema on DST
    The column combination must provide a unique value for Streams.
    SQL> BEGIN
    DBMS_APPLY_ADM.SET_KEY_COLUMNS(
    object_name =>'FaeterBR.tb2',
    column_list =>'id1,names,toys,vendor');
    END;
    Step 14. Configure conflict resolution at the replication db (DST).
    Any easier method applicable the schema?
    DECLARE
    cols DBMS_UTILITY.NAME_ARRAY;
    BEGIN
    cols(1) := 'id';
    cols(2) := 'names';
    cols(3) := 'toys';
    cols(4) := 'vendor';
    DBMS_APPLY_ADM.SET_UPDATE_CONFLICT_HANDLER(
    object_name =>'FaeterBR.tb2',
    method_name =>'OVERWRITE',
    resolution_column=>'FaeterBR',
    column_list =>cols);
    END;
    Step 15. Enable the capture process on the source database (SRC).
    BEGIN
    DBMS_CAPTURE_ADM.START_CAPTURE(
    capture_name => 'FaeterBR_src_capture');
    END;
    Step 16. Enable the apply process on the replication database (DST).
    BEGIN
    DBMS_APPLY_ADM.START_APPLY(
    apply_name => 'FaeterBR_DST_apply');
    END;
    Step 17. Test streams propagation of rows from source (src) to
    replication (DST).
    AT ORCL:
    insert into FaeterBR.tb2 values (
    31000, 'BAMSE', 'DR', 'DR Lejetoej');
    AT STR10:
    connect FaeterBR/FaeterBR
    select * from FaeterBR.tb2 where vendor= 'DR Lejetoej';
    Any other test that can be made?

    Check the metalink doc 301431.1 and validate
    How To Setup One-Way SCHEMA Level Streams Replication [ID 301431.1]
    Oracle Server Enterprise Edition - Version: 10.1.0.2 to 11.1.0.6
    Cheers.

  • Oracle streams configuration problem

    Hi all,
    i'm trying to configure oracle stream to my source database (oracle 9.2) and when i execute the package DBMS_LOGMNR_D.SET_TABLESPACE('LOGMNRTS'); i got an error bellow:
    ERROR at line 1:
    ORA-01353: existing Logminer session
    ORA-06512: at "SYS.DBMS_LOGMNR_D", line 2238
    ORA-06512: at line 1
    When checking some docs, they told i have to destraoy all logminer session, and i verify to v$session view and cannot identify logminer session. If someone can help me because i need this sttream tools for schema synchronization of my production database and datawarehouse database.
    That i want is how to destroy or stop logminer session.
    Thnaks for your help
    regards
    raitsarevo

    Thanks Werner, it's ok now my problem is solved and here bellow the output of your script.
    I profit if you have some docs or some advise for my database schema synchronisation, is using oracle sctrems is the best or can i use anything else but not Dataguard concept or standby database because i only want to apply DMl changes not DDL. If you have some docs for Oracle streams and especially for schema synchronization not tables.
    many thanks again, and please send to my email address [email protected] if needed
    ABILLITY>DELETE FROM system.logmnr_uid$;
    1 row deleted.
    ABILLITY>DELETE FROM system.logmnr_session$;
    1 row deleted.
    ABILLITY>DELETE FROM system.logmnrc_gtcs;
    0 rows deleted.
    ABILLITY>DELETE FROM system.logmnrc_gtlo;
    13 rows deleted.
    ABILLITY>EXECUTE DBMS_LOGMNR_D.SET_TABLESPACE('LOGMNRTS');
    PL/SQL procedure successfully completed.
    regards
    raitsarevo

  • Photo stream questions, sharing, etc.

    Shared photo stream questions
    Camera Roll (CR), Photo Stream (PS)
    If I create a shared PS and share it with a friend, is this temporary like the main PS (most recent 1000 images)?
    For example, if I put one photo in a shared PS and then take 1000 new photos (but I don't add it to the shared PS), presumably it will disappear from my main PS, but will the original photo disappear from the shared PS, too?
    I noticed that each month I have a new monthly event titled, for example, "August photo stream", etc. Are these temporary, too? If I take 1000 photos in September, will the August PS event start to lose images, or will this only occur in the main PS section?
    Apple's web site says "iCloud stores new photos for 30 days, so you have plenty of time to connect your iOS device to Wi-Fi and make sure you always have your most recent shots handy." What does this mean? If I take a photo on iPhone it's permanently in CR (until I sync with iTunes on OS X and then iPhoto offers to remove them from CR), but they're also in PS, which is temporary, right? If so, as long as CR is permanent, how big an issue is the 30 day rule?
    Here's where I'm (further) confused...
    On iPad, my PS section (the tab on bottom) only has 30, relatively recent photos. Each photo was taken with iPhone.
    However, on iPhone, my PS section has 547 images. Some of these were taken as screen shots from iPad (even though they're not on iPad's PS??) Some of the images were "Save As... "  image files from OS X, some were photos imported to desktop iPhoto from a camera. Many of the images were taken years ago. On desktop iPhoto, I have over 3000 images, but the images on iPhone's PS do not appear to be the most recent 1000 images imported to iPhoto.
    So why did iPad's screen shots disappear from iPads PS yet appear on iPhone's PS?
    Why aren't the two iOS device's PS sections the same?
    Is this confusing to everyone, or am I just dense?
    Thanks!

    You are most likely using the same apple id under settings>iCloud. If you just want to stop the photo sharing you can open settings>iCloud and turn off photostream. If you want to keep using photostream and still not share you need to set up an alternate apple id to use under settings>icloud

  • Oracle Streaming in same database - 10g

    Oracle Streaming in 10g database
    Posted: Sep 21, 2006 4:25 AM Reply
    Hello,
    I am trying do streaming at table level for all dml changes from one schema(source) to another schema (target) controlled by admin schema (stream_admin). I have all these schemas in same database(10g-enterprise edition).
    As given in documentations , I created
    1. two queues(in_Q and out_Q) in Stream_admin,
    2. two process( capture and apply process) in Stream_admin ,
    3. A propagation rule for propagation between in_Q and out_Q,
    4. did instantiation,
    5. started both capture and apply process.
    Having done that , I insert in source table and check for the same in target but alas!! nothing happens. I fail to achieve streaming.
    I am not getting any error. Neither in process , nor in propagation or queues. And all queues,rules and process are enabled.
    Please help.

    datapump uses dbms_metadata extensively.
    Problem is twofold
    - the amount of data
    - why on earth do you need to 'copy' these 1.2 Tb a 'number of times'
    One would expect you would have tested the upgrade on a smaller test database. and you wouldn't need to fix bugs in a 1.2 Tb database.
    A better idea would be to duplicate the complete database using RMAN. This doesn't perform conventional INSERTs and doesn't create redo.
    Sybrand Bakker
    Senior Oracle DBA

Maybe you are looking for

  • IPod Touch 1g not recognized in Windows and iTunes

    Hi all i have a old iPod Touch 1g at home. the problem is that the device is death. I can hold Home and Pwr button for a long time but nothing happens. Even when i connect it to a Pc running XP, Windows 7 or a Mac the screen stays black and no error

  • Issue with Vendor Payment Terms

    Hi Gurus, We're facing an issue with the Payment Terms of a Vendor in our SRM system. In R/3 the Vendor has Payment Term 'A' but when we create a PO for this vendor it appears with Payment Term 'B' (PT from SRM I think) The correct PT is 'A' (the one

  • Swnconfig exclude outbox workitem

    Hi all. I have scheduled the SWNCONFIG to send out approval notification every 10 minutes in my system. However, users might manually approve them from SBWP. After they had approve the document, the workitem moved from INBOX to OUTBOX. Now, how do I

  • Open tabs not syncing.

    ''Duplicate post, continue here - [https://support.mozilla.com/en-US/questions/794304]'' I would like to be able to sync my open tabs between my pc and my laptop. These are both used at home. When watching tv, I use the laptop and when I find somethi

  • Invited user doesn't get email

    I'm setting up an InContext site for the first time, and in order to test the "Publishers" experience before offering it to the client to test, I invited myself to be a publisher, using a completely different email address. I never got an email invit