Configuring two_database replication environment

I have the (HR) schema on both databases but the procedure maintain_schemas is failing whenever I execute it.
If I want to configure a one-way replication, what should I do? Where should I create the apply process, the local capture process, the propagation i.e. on the source database or on the destination database?

Hi,
Did you follow the steps on how to create the Streams Administrator and grant all the privileges? Check the documentation to see if you are lacking of some privilege.
What is the error you're getting?
Cheers

Similar Messages

  • Database recovery in a replication environment

    Hi everybody,
    I've configured a replication environment with Oracle 11g (Multimaster synchronous replication). Apparently it works fine, when replication is started any change in a table that belongs in a node is replicated in the rest of the nodes.
    One of the tests I'm making is the following: Try to see what happens if one of the nodes drops. Meanwhile one of the nodes is no avalaible, any modification in an avalaible node is not allowed (as I expected). I'm trying to see if the database service can be avalaible meanwhile a node is missing, although the replications is not running.
    What I've experienced is replication cannot be suspended until the the node was down is restarted.
    Is this the usual behaviour or can I set any parameters to avoid this? In other words, is it possible the database service is running meanwhile one of the nodes is down? The application I'm working on needs this type of replication, buy also needs the service is running always?
    What can I do to make possible the service is running even when a node falls?
    Thanks in advance

    Yes, I've considered an asynchronous replication, with a short period of time (it should be almost synchronous).
    I've tested that, if one of the nodes drops, database service is not available. User intervention is required to stop replication and let the system go on without replication.
    I could be a great idea that, althoug replication is not working, database service is still running and, when replication is restarted, system continue as in the beginning.
    When I test the asynchronous choice I'll answer this post.
    Best wishes,
    Payo

  • Message filtering in propagation process (stream replication environment)

    Hi!
    We have fine configured stream replication in star topology:
    ORCL2 <=> ORCL1 <=> ORCL3
    Where the ORCL1 is "headquarters" and there is no message flow between ORCL2 and ORCL3.
    For some reason we want to filter messages in propagation processes, e.g. DML captured on ORCL1 should be replicated only to ORCL2 or only to ORCL3. There is one propagation process for each "satellite" database.
    To solve this problem I have written function:
    FUNCTION Replicate_Lcr (
    p_lcr IN SYS.lcr$_row_record)
    RETURN VARCHAR2 IS
    which will be making a decision whether to pass the message (return 'Y') or not (return 'N').
    But there is problem: rule is evaluated and function is executed (there is insert into 'stream_log_lcr' table) but value of the expression seems to be 'FALSE' and message (LCR) is not beeing sent to ORCL2 (or to ORCL3).
    When I remove function 'Replicate_Lcr' from propagation rule condition then every message captured by capture process on ORCL1 reaches destination database (ORCL2 or ORCL3).
    The second observation is that, if I run the same code on ORCL2 or ORCL3 then everything seems to be OK: there is insert into 'stream_log_lcr' table and DML captured on ORCL2 (or ORCL3) appears in ORCL1 ("headquarters").
    I suppose that this could be problem with other version of database on "headquarters" (ORCL1) or configuration issues.
    I will appreciate every suggestion.
    Databases:
    ORCL1: 64-bit Windows, ver. 10.2.0.4.0, Windows 2008 server
    ORCL2: 32-bit Windows, ver. 10.2.0.1.0, Windows XP
    ORCL3: 32-bit Windows, ver. 10.2.0.1.0, Windows XP
    SQL code run on ORCL1:
    CREATE TABLE stream_log_lcr (
    data DATE NOT NULL,
    msg SYS.lcr$_row_record NOT NULL
    -- simplified, always return 'Y'
    CREATE OR REPLACE FUNCTION Replicate_Lcr (
    p_lcr IN SYS.lcr$_row_record)
    RETURN VARCHAR2 IS
    PRAGMA AUTONOMOUS_TRANSACTION;
    BEGIN
    IF p_lcr IS NOT NULL
    THEN
    INSERT INTO stream_log_lcr
    VALUES (SYSDATE,
    p_lcr);
    COMMIT;
    END IF;
    RETURN 'Y';
    END;
    -- create propagation process with above function in rule condition
    BEGIN
    DBMS_STREAMS_ADM.add_schema_propagation_rules
    (schema_name => 'data_schema',
    streams_name => 'primary_to_secondary2',
    source_queue_name => 'strmadmin.capture_primary',
    destination_queue_name => 'strmadmin.from_primary@ORCL2',
    include_dml => TRUE,
    include_ddl => TRUE,
    source_database => 'ORCL1',
    and_condition => ' strmadmin.Replicate_Lcr(:dml) = ''Y'' ',
    inclusion_rule => TRUE,
    queue_to_queue => TRUE);
    END;
    -- check if function 'Replicate_Lcr' was evoked:
    SELECT * FROM stream_log_lcr ORDER BY data;

    hi porzer,
    In Propagation process ( source) also 0 errors. But in apply ( dest ) under statistics under server status is displaying as IDLE. And Coordinator status is APPLYING. In Capture (source) no error. In Applying ( dest ) no error. what else i can do please?
    up to now what i did i am telling:
    I had 2 databases had one table same table. one database i changed the mode to ARCHIVELOG mode. Another database is in NOARCHIVELOG mode only. In first database setup streams i run. And i run manually 2 .sql files and one .dat file like this :
    SQL>@e:\oracle\product\10.2.0\client_2\sysman\report\OTEST_ADMIN_NON_OMS_SETUP.sql
    SQL>host e:\oracle\product\10.2.0\client_2\sysman\report\OTEST_ADMIN_NON_OMS_exportimport.bat
    SQL>@e:\oracle\product\10.2.0\client_2\sysman\report\OTEST_ADMIN_NON_OMS_startup.sql
    Any thing else i can do? i didn't have metalink registration.i hope i am not boring you.
    Thanks in advance.

  • How to configure multiple dev environment in single IP

    HI,
    I have a WLS running on a Windows NT machine. I want to configure multiple partition/environment
    in that single WLS so that each user will have their own partition/environment to
    play around. Once they are happy with their EJBs then we shall move them to a testing
    environment. Is that possible in WLS??. Can anyone please guide me to a good documentation
    about this one.
    Many Thanks
    Vasan.
    PS: I dont want a clustering technology all i want is that in a single ip i wanna
    multiple development environment.

    Tom,
    looks like you are specifying http protocol in
    sky.framework.ejb.SkySessionBean.getInitialContext method.
    Change it to t3 or, if the JNDI lookup is in the same JVM, don't specify
    PROVIDER_URL. If you have to use http tunneling due to any reason , you have to
    enable HTTP Tunneling in server to which initial context is created. It is
    disabled by default in 60. You can do so via console. Please refer to
    documentation , let me know if you need additional help.
    Viresh Garg
    Principal Developer Relations Engineer
    BEA Systems
    Tom Purcell wrote:
    Kevin,
    You said that with 6.0 the following happens:
    On the second server I get an error when I do
    http://host:secondServerPort/index.html
    Is this the error you get on the Admin Server:
    <Feb 28, 2001 12:42:24 PM EST> <Info> <HTTPTunneling> <Login rejected with
    code: 'Failed', reas
    on: You are trying to use http tunneling
    The server will not let you connect'>
    <Feb 28, 2001 12:42:24 PM EST> <Error> <HTTP> <Problem dispatching servlet
    request
    weblogic.common.T3Exception: You are trying to use http tunneling
    The server will not let you connect
    at weblogic.socket.TunnelLogin.rejectConnection(TunnelLogin.java:87)
    at weblogic.socket.TunnelLogin.service(TunnelLogin.java:59)
    at
    weblogic.servlet.internal.MuxableSocketHTTP.dispatch(MuxableSocketHTTP.java:
    467)
    at
    weblogic.socket.NTSocketMuxer.processSockets(NTSocketMuxer.java:638)
    at
    weblogic.socket.SocketReaderRequest.execute(SocketReaderRequest.java:23)
    at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:137)
    at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:120)
    >
    And on the Managed Server:
    <Feb 28, 2001 12:42:24 PM EST> <Error> <ConnectionManager> <Could not
    connect to: '0S:127.0.0.1
    :[-1,7001,-1,-1,-1,-1,-1]'
    java.net.ProtocolException: Tunneling result unspecified - is the HTTP
    server at host: 'localho
    st' and port: '7001' a WebLogic Server?
    at
    weblogic.rjvm.http.HTTPClientJVMConnection.handleNullResponse(HTTPClientJVMC
    onnectio
    n.java:152)
    at
    weblogic.rjvm.http.HTTPClientJVMConnection.connect(HTTPClientJVMConnection.j
    ava:194)
    at
    weblogic.rjvm.http.HTTPClientJVMConnection.createConnection(HTTPClientJVMCon
    nection.
    java:98)
    at weblogic.rjvm.Protocol.createConnection(Protocol.java:207)
    at
    weblogic.rjvm.ConnectionManager.findOrCreateConnection(ConnectionManager.jav
    a:997)
    at
    weblogic.rjvm.ConnectionManager.bootstrap(ConnectionManager.java:364)
    at
    weblogic.rjvm.ConnectionManager.bootstrap(ConnectionManager.java:330)
    at
    weblogic.rjvm.RJVMManager.findOrCreateRemoteInternal(RJVMManager.java:228)
    at weblogic.rjvm.RJVMManager.findOrCreate(RJVMManager.java:188)
    at
    weblogic.rjvm.RJVMFinder.findOrCreateRemoteServer(RJVMFinder.java:180)
    at weblogic.rjvm.RJVMFinder.findOrCreate(RJVMFinder.java:149)
    at weblogic.rjvm.ServerURL.findOrCreateRJVM(ServerURL.java:207)
    at
    weblogic.jndi.WLInitialContextFactoryDelegate.getInitialContext(WLInitialCon
    textFact
    oryDelegate.java:232)
    at
    weblogic.jndi.WLInitialContextFactoryDelegate.getInitialContext(WLInitialCon
    textFact
    oryDelegate.java:174)
    at
    weblogic.jndi.WLInitialContextFactory.getInitialContext(WLInitialContextFact
    ory.java
    :123)
    at
    javax.naming.spi.NamingManager.getInitialContext(NamingManager.java:668)
    at
    javax.naming.InitialContext.getDefaultInitCtx(InitialContext.java:246)
    at javax.naming.InitialContext.init(InitialContext.java:222)
    at javax.naming.InitialContext.<init>(InitialContext.java:198)
    at
    sky.framework.ejb.SkySessionBean.getInitialContext(SkySessionBean.java:112)
    at
    sky.framework.ejb.SkySessionBean.setSessionContext(SkySessionBean.java:94)
    at
    sky.ejb.session.skyreferencesession.SkyReferenceSessionBeanImpl.setSessionCo
    ntext(Sk
    yReferenceSessionBeanImpl.java:87)
    at
    weblogic.ejb20.internal.StatelessEJBHome.createBean(StatelessEJBHome.java:22
    9)
    at
    weblogic.ejb20.pool.StatelessSessionPool.createBean(StatelessSessionPool.jav
    a:90)
    at
    weblogic.ejb20.pool.StatelessSessionPool.getBean(StatelessSessionPool.java:7
    5)
    at
    weblogic.ejb20.manager.StatelessManager.preInvoke(StatelessManager.java:67)
    at
    weblogic.ejb20.internal.BaseEJBObject.preInvoke(BaseEJBObject.java:113)
    at
    weblogic.ejb20.internal.StatelessEJBObject.preInvoke(StatelessEJBObject.java
    :55)
    at
    sky.ejb.session.skyreferencesession.SkyReferenceSessionBeanEOImpl.setDataSou
    rce(SkyR
    eferenceSessionBeanEOImpl.java:22)
    at
    sky.ejb.session.skyreferencesession.SkyReferenceSessionBeanEOImpl_WLSkel.inv
    oke(SkyR
    eferenceSessionBeanEOImpl_WLSkel.java:161)
    at
    weblogic.rmi.internal.BasicServerAdapter.invoke(BasicServerAdapter.java:373)
    at
    weblogic.rmi.cluster.ReplicaAwareServerRef.invoke(ReplicaAwareServerRef.java
    :128)
    at
    weblogic.rmi.internal.BasicServerAdapter.invoke(BasicServerAdapter.java:237)
    at
    weblogic.rmi.internal.BasicRequestHandler.handleRequest(BasicRequestHandler.
    java:103
    at
    weblogic.rmi.internal.BasicExecuteRequest.execute(BasicExecuteRequest.java:1
    7)
    at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:137)
    at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:120)
    >
    Thanks,
    Tom
    <Feb 28, 2001 12:42:24 PM EST> <Info> <ConnectionManager> <Bootstrap unable
    to get a: 'Protocol
    : 'http'' connection to: 'localhost/127.0.0.1' on port: '7001'>
    "Kevin Apte" <[email protected]> wrote in message
    news:[email protected]...
    This advice applies to WL5.1. I have been trying to configure a
    multiinstance Weblogic with no success. I want multiple copies
    of the sample petStore available for test by multiple users.
    I think in 6.0
    You would need to open an admin server
    Open the configuration file for say PetStore.
    Clone the petStore- call it petStore2. .
    Create a copy of the startup script and modify it to start PetStore2.
    After doing this only one config.xml is created for both servers.
    I am able to start the first server and view the petStore application,
    On the second server I get an error when I do
    http://host:secondServerPort/index.html
    How does one create multiple instances in WL6.0 using the console
    or using a XML editor ?
    Jesse E Tilly <[email protected]> wrote:
    [This followup was posted to weblogic.developer.interest.management
    and
    a copy was sent to the cited author.]
    In article <[email protected]>, [email protected]
    says...
    HI,
    I have a WLS running on a Windows NT machine. I wantto configure multiple partition/environment
    in that single WLS so that each user will have theirown partition/environment to
    play around. Once they are happy with their EJBs thenwe shall move them to a testing
    environment. Is that possible in WLS??. Can anyoneplease guide me to a good documentation
    about this one.
    Many Thanks
    Vasan.
    PS: I dont want a clustering technology all i wantis that in a single ip i wanna
    multiple development environment.
    NT isn't a multiuser environment. I will assume that
    you mean some
    Win32 box sitting in a room where people can copy their
    classes/EJBs and
    test their stuff. If this is the case, the setup is easy.
    Create a directory and startup script for every developer.
    Have a
    global properties file in the same directory as the scripts
    so that
    common settings are, well, common. Tell each user to
    modify the various
    parts of the properties files in their directories to
    deploy EJBs and
    register servlets, etc etc.
    In each startup script, you will have to specify the server
    directory
    and the binding port for each person. Pick a port (something
    easy to
    remember and high, like 30000, 30001, etc etc) for each
    person. Why?
    Well, when they connect to the app server, they'll know
    they're hitting
    only their server. Also, TCP/IP won't let more than one
    JVM bind to the
    same port, so if you want simultaneous instances (> 4
    you better have
    one kick ass Win32 box) you have to do this.
    Hope this solves your problem.
    JT
    Senior Developer, HotelTools INC
    [email protected]

  • Multiversions in Materialized View Replication Environment

    Dear friends
    In my company we have a matarialized view environment where A materialized view, in snapshot site, contains a complete or partial copy of a target master - master site.
    This environment is in Oracle 8i Enterprise Edition and now we are increasing it. We are shrinking our budget and then i'm research an solution more cheaper, i found Oracle 10G Database Standard Edition One, an solution desinged for small to medium environments. My thought is to use Oracle 10G SE1 for news snapshot sites.
    Can someone talk about the user of differents versions (8i mastersite X 10G snapshotsite) in a Materialized View Replication Environment, is it possible? must i upgrade all my environment replication for 10G or to use multiversions ?
    Thanks !!

    10g to 8i can work, assuming
    - 8i = 8.1.7
    - You are applying at least the 8.1.7.4 patchset to the master (8i) or you are applying the latest patchset to the slave (10g). Realistically, you probably want to do both.
    If you are talking about using the standard edition, I assume that you are only doing basic replication, not multi-master replication. I also assume you are aware of the limitations on hardware that come with Standard Edition 1-- I don't recall if you are limited to machines that have a capacity for 2 or 4 CPU's. Also, be aware that Oracle considers multi-core CPU's as multiple CPU's.
    Justin
    Distributed Database Consulting, Inc.
    http://www.ddbcinc.com/askDDBC

  • Preventing Item On Hand reservation in a configure to order environment for multi org same OU environment

    I have a scenario in a configure to order environment where I want to automatically creating flow schedules based on sales order demand. I have two orgs in same OU.
    1. How to setup sales order line to generate flow schedules in the manufacturing org and then ship the finished goods from the shipping org. I believe this can be done from sourcing rules to populate Warehouse field in sales order line with the shipping org and create Flow schedule in another manufacturing org. Please advise ?
    2. Once flow schedule is completed I want to prevent the reservation of Finished good on hand against Sales order so that the On hand could be transferred to shipping org and shipped. Please advise ?
    Thanks for your inputs.
    Best,
    Rajesh

    Pl do not post duplicates - Preventing Item On Hand reservation in a configure to order environment for multi org same OU environment

  • If the production database i down, how to switch to standby database using the replication environment in ASsE

    This question is in Business Continuity Plan (BCP)
    if the production database is down, how to switch to standby database using the replication environment in ASE

    Assuming you're referring to Sybase's Replication Server (repserver) product ...
    Repserver moves data changes from one database to another.
    Repserver can't 'switch' your application to use a different database; something outside of repserver has to do the actual switching.
    If your application happens to be connecting to your ASE via Sybase's OpenSwitch (or similar) product, then OpenSwitch would be responsible for switching your application to use the standby database.  Whether you have OpenSwitch perform the switch automatically, or via a manual command, is up to your BCP requirements.
    If your application connects directly to ASE then you'll have to decide how/when to get your application to switch to the standby database.  In the simplest form this would require shutting down the application, making sure replication is quiesced (ie, all in-flight transactions have been applied to the standby database), modify the interfaces/sql.ini file (used by your application) to point to the standby database, make any customized changes to the standby dataserver/database (eg, unlock application logins, take databases out of 'dbo use only' mode, etc), and startup the application (at this point it should connect to the standby database per the changes to the interfaces/sql.ini file).

  • Replication environment

    I'm trying to create replication environment with one master site and one materialized view site in my test databases on the same node. I followed step-by-step the example in the Replication Management API Reference book ( from June 2001 ). Replication environment and replication object is created successfully but replication not runing properly and dont send any error message. My ORACLE Server 9i (9.0.1.1) Enterprise Edition for Windows NT/2000/XP Pro runing on Win2000 Advanced server.
    Thanks.
    Attila Fintor

    Have you set the system variable job_queue_processes to be higher than 0
    e.g. alter system set job_queue_processes = 2;
    I had this problem and searched for a while till i found the reason.
    If this is not set correctly, manual refresh will work but no automatic
    refresh may run since no backgraound processes are allowed!

  • Replication environment data inconsistency...

    Gurus,
    I have multi master replication environment there is one schema SGI on two databases LLU and BER data in LLU is correct and data in BER is incorrect.
    I need to drop SGI schema in BER and import export dump from LLU, but I don't want to break replication environment and rebuild it.
    I have checked DBMS_RECTIFIER_DIFF given in metalink but somehow could not find concrete example that would help me use it correctly.
    Does anyone have any idea about it? Or any other suggestion on how can I achieve my desired goal?
    Please advice.

    Gurus,
    I have multi master replication environment there is
    one schema SGI on two databases LLU and BER data in
    LLU is correct and data in BER is incorrect.
    I need to drop SGI schema in BER and import export
    dump from LLU, but I don't want to break replication
    environment and rebuild it.
    If you drop the schema you will destroy your replication environment anyway. In that case you have more work to do in order to cleanup the replication environment at your other master site.
    I don't think you want to do that!
    My personal opinion is that you should just remove from the replication environment the master replication objects that belong to the schema you want to refresh. Refresh the data (using import utility) and then create the replication objects again.
    I don't say it is easy, but I think it is the right way to go. Don't be lazy. Follow the rules written in the docs and you should be fine.
    The high level picture of the steps you have to do is:
    - Make sure there are no deferred transactions waiting to be applied. If there are, deal with them first.
    - From your master definition site suspend the replication activity (using dbms_repcat.suspend_master_activity)
    - From your master definition site, remove the master rep objects you want to refresh. (using dbms_repcat.drop_master_repobject)
    - Import the data in SGI BER
    - From your master definition site, create the master rep objects again using (DBMS_REPCAT.CREATE_MASTER_REPOBJECT and make sure you use copy_rows => FALSE and use_existing_object => TRUE)
    - Resume the master activity (using dbms_repcat.resume_master_activity)
    and that is it.
    And again, make sure you don't have deferred transactions before you start this.
    Read the docs until you fully understand what you have to do.
    In order to feel more comfortable, I would suggest you to test this approach first.
    I hope my advice will help you to find the right approach.
    P.S. Hermant's suggestion is also worth considering (the link he provided). The approach that I mentioned may require more work from your side, but it also gives you more control and flexibility so you would know what is going on in every moment.
    Message was edited by:
    tekicora

  • Character set conversion in advanced replication environment 10g

    we have a replication environment with 5 databases using oracle 10g. one master definition site, 4 master sites. we want to change database character set in replication environment.
    Does somebody has experience about database character set conversion in replication environment.
    Thanks

    842106 wrote:
    Hi All,
    I am currently facing a strange issue while loading a csv into oracle BE table.
    The data is coming as Técnicas but I need to store it as Técnicas . I think there is some character set issue for which data is coming as in distorted manner .
    I am using oracle 10G enterprise edition .
    Any help is greatly appreciated.How do I ask a question on the forums?
    SQL and PL/SQL FAQ
    what do you observe when you inspect the actual content of the CSV file?

  • Steps by Step document for configuration Merge Replication in 2005

    Hi ,
    Can anyone provide link to configure Merge Replication in 2005 with print screen.Also let me know the difference between Transcation level and Merge Replication.
    regards
    Vijay

    Here you have one tutorial:
    http://www.sqlshack.com/sql-server-replications/
    Use merge replication if you want to reflect changes from the subscriber to the publisher or viceversa.
    Use transactional replication if you want changes from the publisher to the subscriber.
    You can use transactional databases with Oracle databases.
    http://www.sqlshack.com/sql-server-replications/
    http://www.sqlshack.com/sql-server-replications/
    http://www.sqlshack.com/sql-server-replications/
    http://www.sqlshack.com/sql-server-replications/
    http://www.sqlshack.com/sql-server-replications/
    MVP MCT MCTS Daniel Calbimonte
    http://elpaladintecnologico.blogspot.com

  • SSO Configuration in UNIX Environment

    Hi,
    We are planning to impliment SSO between 2004s Portal with R/3, BW, and host of other non SAP application which supports SSO. All the applications are hosted in UNIX environment.
    Could anyone help us in understanding the pros and cons of having SSO in unix envrionment. Also you could give us any document (other than help.sap.com) which guides us in configuring will be of great help.
    Thanks
    Vivek Raj.

    Hi,
    There are no pros and cons for <b>SSO on Unix</b>. Needless to mention, the pros and cons of different operating systems supported by SAP.
    The steps involved to configure SSO remains the same for any of the SAP backend systems:
    1) Create a system on Portal and give a system alias to it.
    2) Download the verify.der file from Portal (System Administration -> System Configuration -> Keystore Administration)
    3) Log on to the backend SAP system.
    4) Use transaction STRUSTSSO2 and import the Portal certifcate (verify.der) to the SAP R3 or SAP BW or SAP CRM or for that matter any SAP system with ABAP stack.
    5) Give the values for EP system ID, client and language. Refer to this link : http://help.sap.com/saphelp_nw2004s/helpdata/en/78/f1a8490e7011d6999500508b6b8a93/content.htm
    6) Configure the backend SAP systems to accept and verify the logon ticket. Use RZ10 and set the parameters : login/accept_sso2_ticket=1 and login/create_sso2_ticket=1
    7) Restart your R3 or BW system.
    For non-SAP systems, can you just let me know which systems are you referring to ?
    Incase you need a document for SSO configuration, do send me your email address. You can also find documents in SDN Library.
    Regards,
    Sunil

  • RMAN Configuration in DG  Environment

    Hi,
    I have planned a backup strategy for my Oracle 10g Database using RMAN . Physical standby is used for DG . I wanted to know how can I create a physically seperate catalog server for the above purpose.
    OS Version- Windows Server 2003
    Regards
    Vinita

    Hello,
    About Standby Database Registration
    In a Data Guard environment, the primary and standby databases share the same DBID and database name. To be eligible for registration in the recovery catalog, each database in the Data Guard environment must have different DB_UNIQUE_NAME values. The DB_UNIQUE_NAME parameter for a database is set in its initialization parameter file.
    If you use RMAN in a Data Guard environment, then you can use the REGISTER DATABASE command only for the primary database. You can use the following techniques to register a standby database in the recovery catalog:
    When you connect to a standby database as TARGET, RMAN automatically registers the database in the recovery catalog.
    When you run the CONFIGURE DB_UNIQUE_NAME command for a standby database that is not known to the recovery catalog, RMAN automatically registers this standby database as long as its primary database is registered.
    http://download-uk.oracle.com/docs/cd/B28359_01/server.111/b28294/rman.htm#SBYDB04700
    Regards,
    Vineeth

  • EntSSO configuration in clustered environment

    Hi All
    I am trying to use SSO Applications to store credentials and use it on Send / Receive ports. It worked fine on the development machine which is a standalone box. But while moving it to a higher clustered environment, I am facing some issues. 
    BTS_Service is a BT service account which is added to both BTS_SSO_Admins and BTS_SSO_Affiliate_Admins
    BTS_SSO_Service is an SSO Service account which is running the ENTSSO service and its also added to BTS_SSO_Admins and BTS_SSO_Affiliate_Admins
    Logged in using BTS_Service account
    Using SSO Administration - Created SSO Application as individual, allow tickets true.
    Using SSO Client Utility - Added credentials to the SSO Application 
    Now logged back in using the normal user credentials
    Ticketing component and pipeline is already deployed (tested in standalone - working fine)
    On a WCF Send port, used the Ticketing Send pipeline, and on credentials tab, selected the SSO Application created in Step 2. 
    while running, I get the below error
    There was a failure executing the send pipeline: "BizTalk_Server_Project1.SendPipelineSSOTicketRetriver, XXX.Pipeline.SSOTicketRetriever, Version=1.0.0.0, Culture=neutral, PublicKeyToken=sf343fs3t34daf8" Source: "SSO Ticket Retriever"
    Send Port: "XXXRequestResponse_WCFCustom" URI: "http://XXXXXXXXXXXXXX/XXXXXXXXXXXXXXX" Reason: Access denied.
    I have referred to Richard Seroter's post .
    http://blogs.msdn.com/b/richardbpi/archive/2005/08/09/building-and-executing-a-biztalk-single-sign-on-scenario.aspx
    Another big difference i see is that when in my development machine, i have windows accounts setup in BizTalk configuration under Enterprise SSO tab as below. while in the Clustered environment where I am facing the issue, windows accounts are not added. 
    Could that be the issue?
    mayur macwan

    Hi Mayur,
    Since you are working in cluster environment might be the SSO admin is configured in this clustered node. Also, check whether this cluster node contains SSO admin group. If yes, add your credential into this group. Since SSO denied access is popping up I
    believe you first try to add your credential in BizTalk SSO group (might be present in different cluster group).
    Thanks

  • ISE Configuration in Distributed Environment

                      Hi All,
    I have quick questions about  ISE deployment in Distributed environment, as i have purchased 2 X Cisco ise 3395 - For Data Center and 3 X Cisco ISE 3355 for remote location with 3500 Base licences and 500 Advance licences.
    i have some questions on this deployment
    i will install 1 3395 in Primary Datacenter and other 3395 in Our secondary Data center as Primary admin+Primary Monitoring and Secondary Admin+Secondary Monitoring
    and each 3355 will get installed in Remote location as policy server, My Question is it this will be correct deployment?
    or while configuring 3395 do i need to configure Policy server as well in addition to Primary admin and monitoring?
    or please suggest me best deployment stratagy!
    Thanks,
    Sachin

    Thanks for the reply,
    all three sites are connected in MPLS with 100MB redundant band width
    we are have 2 data center one is primary and other is secondary. and all client locations are connected with 100 Meg links where i am planning to install 3355 which will act as authentication server.
    but now my question is
    3395 - Primary Admin+Primary Monitoring - Primary DC
    3395 - Secondary Admin+ Secondary Monitoring - Secondary DC
    3355- will say for one remote location(PSN)
    3355- Second remote Location(PSN)
    3355- third Remote location (PSN)
    thanks,
    Sachin

Maybe you are looking for