Best way to Rebuild multi-master replication

I have a site (A) which has been modified (tables and columns and such) that was used as the master replication site.
I have another site (B) which was the target site.  The tables in site B have not been changed.
The replication is functioning correctly (as the columns and tables were only additions).
All the replication is asynchronous and is divided into to groups (one that refreshes every 2 minutes and one that refreshes every 15 minutes)
I want to bring site B up to date (move the changes from site A into site B).
I know that the data in the added columns has not been utilized as the new software was not in production yet.
I know I will need to rebuild the replication groups to include the missing columns, and I presume I will need to stop the replication (I will disable the PUSH so I can still work with the master database).
I will also need to add the new tables to the replication groups.
I can rebuild the database entirely using an export and import (after cleaning out the existing database in site B).
The question I have is efficiency, would I be better to just drop all the replication and start from scratch using the export?
Could I simply add the tables and columns to site B and rebuild the replication?  Will I run into issues when I try to rebuild the replication if the job remains broken?
Or can I rebuild site B (after breaking the job) using the export, then rebuild all the replication groups and re-enable replication (I know the scn numbers need to align but I am not sure if it will cause issues with the pending replication in the queue once I break the job).
I am pretty much free to do what is needed (even though this is a production environment, the replication is not critical - I just cannot deny access to the master site for the users)
The "former" DBA did not understand replication apparently and I have inherited this mess, while I understand how to do most of the things in this situation, I am not so sure of the best (or fastest) way.
Forgot to mention the environment Oracle 11g R2 running on Linux (Ubuntu 12.04 LTS)
Message was edited by: MrOracle

There was a good discussion on the same subject in one of the oracle forum.
Q] I have an application where data is never deleted, I have just rebuilt an
index which was previously 4gb in size and its now 3gb in size. Can anybody
explain?
A1] The first thing that comes to mind is that (assuming what was taught
on the Oracle Fundementals and tuning course I went on with Oracle
University last year is true[1]) when an index leaf block that is not
the right most block is filled up the next entry to be put in results
in a 50-50 split, two blocks are created one containing the lower 50%
of the entries and the other containing the higher 50%, both half
full.. Assumign that the indexed field isn't an increasing key type
value (i.e. it is possible to insert a value that will be lower
(futher left) than an earlier value) that will mean that your index
leaf blocks will mostly be between 50% and 100% full. Assuming a
reasonably random distribution of data your average block would be
around 75% full (split the difference between 50% and 100%). When you
rebuild the index the blocks are (with the same caveat as above)
repacked to 100% (or upto PCTFREE, I forget which) full. this would
explain why the size of the index has changed to 75% of it's previous
value. The same amount of data, it's just packed more tightly.
I don't have a database to hand to check this but it does seem to fit
the situation and my reccollection of how I'm told Oracle handles
indexes.
Stephen
[1] I'm told, and have seen, that it's not always a given. Maybe
there's a conference paper in somone's future entitled something like
"Lies Oracle Told Me: The gap between training and reality"
A2] If you look at Stephens post, then a reduction from free space of 30% in the blocks to 10% in the blocks (assuming pctfree =3D 10), gives around 20% gain back or around 800m from a 4G index. This would fit in quite nicely with what you are observing.
Jaffar

Similar Messages

  • Problem in multi master replication creation using DBA Studio -- replication

    Hi,
    I am trying to create multi master replication using DBA studio but facing following problem at the time of master group creation.
    ORA-04052 error occured when looking up remote object SYS.SYS@CBOLDATA
    ORA-00604 error occured at recursive SQL level 2
    ORA-01017 invalid username/password, logon denied
    ORA-02063 preceding line from CBOLDATA
    If you want to know how I am trying the whole thing then here is the way I am doing comfiguration.
    First I have created master site which is created successfully.
    Here I have used two database named UPP817 & CBOLDATA
    Step of master site creation is followed like this.
    1. Master Site addition (Added both UPP817 & CBOLDATA using SYSTEM username)
    2. Default User (No change is done, Default schema and password taken)
    3. Master Site Schema (Added SCOTT/tiger)
    4. Scheduled Link (No change is done, Default values taken)
    5. Purge (No change is done, Default values taken)
    6. Finished successfully
    Master group creation there is three option available
    1. General (Only value of Name is given)
    2. Object (No object added)
    3. Master Sites (It takes UPP817 by default as a master defination site then as a Master Site I have added CBOLDATA)
    when I pressed on Create It gives me the error listed above.
    I really appreceate your help
    Mukesh

    Create public database link at master site database for CBOLDATA.
    Also see that you are using SNMP protocol and Oracle Pipe at both
    databases. This will be usefull while making replication from remote
    place.
    I hope this will work.
    regards
    Avinash Jadhav

  • Multi-master Replication and Metadirectory 5.0

    The Metadirectory 5.0 documentation states that it cannot work with a directory server configured for multi-master replication. We need to use Metadirectory since we are integrating the Directory Server with other systems. Does this mean that we'll be forced away from MMR configuration? What are some of the alternatives? Does iPlanet have any plans for supporting MMR in future versions of Metadirectory?

    I think you can enable the retro changelog on a consumer replica. I'm pretty sure that works.
    You might be able to enable it on a hub. You also might get it to work on a master, but the changelog on the master will contain some replication housekeeping changes that may confuse Meta Directory. I'm not sure what they are but they are obvious if you look at the changelog. If you can figure out some way to screen those changes out from Meta Directory, you might be able to use it.

  • Two problems after reinitializing multi-master replication.

    Hi,
    I recently set up multi-master replication between two machines, A and B, running iDS 5.1. Everything works well until I disable replication and re-enable replication. Here's the procedure that I follow:
    1. Delete agreement on machine A and B.
    2. Disable the replica on machine A and B.
    3. Disable the changelog on machine A and B.
    4. Enable the changelog on machine A and B.
    5. Enable the replica on machine A and B.
    6. Create replication agreement on A and B.
    7. Initialize B from A.
    Problem #1:
    After I follow this procedure, updates to A result in the error:
    <B>
    Data required to update replica "cn=rep1, cn=replica, cn="o=sylantro.com", cn=mapping tree, cn=config (host rosemary.sylantro.com, port 389)" has been purged. The replica must be reinitialized.
    [07/Feb/2002:10:43:28 -0800] NSMMReplicationPlugin - Replication: Incremental update failed and requires administrator action
    </B>
    At this point if I re-initialize B from A, the problem ceases and everything is fine.
    This problem still occurs even if I restart the server before Step 4.
    Problem #2.
    When I assign machine A the old replicaId of machine B and assign machine B the old replicaID of machine A in Step 5, I get the following error.
    <B>
    [07/Feb/2002:10:06:02 -0800] NSMMReplicationPlugin - Unable to aquire replica "cn=rep1, cn=replica, cn="o=s
    ylantro.com", cn=mapping tree, cn=config (host wasrocky.sylantro.com, port 389)": the replica has the same Replica ID as this one. Replication is aborting.
    </B>
    The server looks like it is remembering its old replicaID.
    This problem still occurs even if I restart the server before Step 4.
    Any insight would be greatly appreciated.
    Thanks!
    Gil

    For the problem #1, the error message is normal since you have actually remove the changelog. However, once you have done an initialization, Replication should work.
    For the problem #2, the message is correct, because, although you have changed the ReplicaId on both server, your data still contain references to the previous replicaId both on Server A and Server B.
    In general, Replica ID uniquely identify a Master in a replication topology. It should not change. The replicaID is also used in the timestamps associated with each change and they guarantee that the Replication can always distinguish 2 changes, and also can detect time synchronization issues between masters.
    One way to avoid the Replica ID issue when changing ReplicaIds, is to reload the data on one server and then re-initialize the second one from the first one.
    Regards,
    Ludovic.

  • Problem with Multi Master Replication

    Hello All,
    I've setup a multi-master replication with no consumers. i.e i'm having 2 suppliers which should update each other. The setup seems to be fine since the initialization of one supplier by other works very fine. But i couldn't get the synchronization btwn the suppliers get worked. I noticed in the error log that the syn-scan request arrived, but ignored. What are the possibilities of this error ?
    Please help me with this regard.
    Thanks in advance,
    Rajesh

    Hello All,
    Rich, you have been a support to most of us in the group(indeed much to my help)...Its splendid work....
    My problems disappeared after applying the Service pack ....the service pack in fact is mainly to sort out the replication issues.
    Advice from my experience - The patch may be more than enough for most of the replication issues.
    One observation - i had the replica busy error, but i didn't have to restart the replica as suggested by some of the previous threads. Seems the service pack did some fix for it.
    Thank you all,
    Best Regards,
    Rajesh

  • Basic Multi Master Replication doesn't work!

    Hi all,
    <br><br>
    I am studying Oracle replication and I tried to apply multi master replication MMR between two databases: WinXp12 (master definition) and WinXp11.<br>
    After successfully doing all the steps in the code below to replicate table SCOTT.DEPT, I inserted a row in the table in WinXp12. But I didn't see the inserted row in the remote site WinXp11.<br><br>
    Where wrong was I? Anything missed?<br><br>
    By the way, deferror table contains no row and DBA_JOBS has no failures.<br><br>
    Thanks in advance.
    <br><br>
    CREATE USER REPADMIN IDENTIFIED BY REPADMIN;
    GRANT CONNECT, RESOURCE, CREATE DATABASE LINK TO REPADMIN;
    EXECUTE DBMS_REPCAT_ADMIN.GRANT_ADMIN_ANY_SCHEMA('REPADMIN');
    GRANT COMMENT ANY TABLE TO REPADMIN;
    GRANT LOCK ANY TABLE TO REPADMIN;
    EXECUTE DBMS_DEFER_SYS.REGISTER_PROPAGATOR('REPADMIN');
    CONN REPADMIN/REPADMIN
    CREATE DATABASE LINK WINXP11 CONNECT TO REPADMIN IDENTIFIED BY REPADMIN USING 'WINXP11';
    SELECT SYSDATE FROM DUAL@WINXP11 ;
    -- Add jobs to WINXP11
    CONNECT REPADMIN/REPADMIN@WINXP11
    BEGIN
      DBMS_DEFER_SYS.SCHEDULE_PUSH(
        DESTINATION => 'WINXP12',
        INTERVAL => 'SYSDATE + 1/(60*24)',
        NEXT_DATE => SYSDATE,
        STOP_ON_ERROR => FALSE,
        DELAY_SECONDS => 0,
        PARALLELISM => 1);
    END;
    BEGIN
    DBMS_DEFER_SYS.SCHEDULE_PURGE(
      NEXT_DATE => SYSDATE,
      INTERVAL => 'SYSDATE + 1/24',
      DELAY_SECONDS => 0,
      ROLLBACK_SEGMENT => '');
    END;
    -- ADD JOBS TO WinXP12
    CONNECT REPADMIN/REPADMIN@WINXP12
    BEGIN
      DBMS_DEFER_SYS.SCHEDULE_PUSH(
        DESTINATION => 'WINXP11',
        INTERVAL => 'SYSDATE + 1/(60*24)',
        NEXT_DATE => SYSDATE,
        STOP_ON_ERROR => FALSE,
        DELAY_SECONDS => 0,
        PARALLELISM => 1);
    END;
    BEGIN
    DBMS_DEFER_SYS.SCHEDULE_PURGE(
      NEXT_DATE => SYSDATE,
      INTERVAL => 'SYSDATE + 1/24',
      DELAY_SECONDS => 0,
      ROLLBACK_SEGMENT => '');
    END;
    BEGIN
       DBMS_REPCAT.CREATE_MASTER_REPGROUP(
         GNAME => '"MGROUP1"',
         QUALIFIER => '',
         GROUP_COMMENT => '');
    END;
    BEGIN
       DBMS_REPCAT.CREATE_MASTER_REPOBJECT(
         GNAME => '"MGROUP1"',
         TYPE => 'TABLE',
         ONAME => '"DEPT',
         SNAME => '"SCOTT"');
    END;
    -- Generate Replication Support
    BEGIN
       DBMS_REPCAT.GENERATE_REPLICATION_SUPPORT(
         SNAME => '"SCOTT"',
         ONAME => '"DEPT"',
         TYPE => 'TABLE',
         MIN_COMMUNICATION => TRUE,
         GENERATE_80_COMPATIBLE => FALSE);
    END;
    SELECT * FROM DBA_REPCATLOG ;
    -- NO ERROR
    BEGIN
    DBMS_REPCAT.RESUME_MASTER_ACTIVITY(
    GNAME => '"MGROUP1"');
    END;
    BEGIN
    DBMS_REPCAT.SUSPEND_MASTER_ACTIVITY(
    GNAME => '"MGROUP1"');
    END;
    BEGIN
    DBMS_REPCAT.ADD_MASTER_DATABASE(
    GNAME => '"MGROUP1"',
    MASTER => 'WINXP11');
    END;
    -- Restart replication support:
    BEGIN
    DBMS_REPCAT.RESUME_MASTER_ACTIVITY(
    GNAME => '"MGROUP1"');
    END;
    -- here I could see in WinXP11 the tables created in SCOTT schema with data
    -- in WinXP12 I successfully issued the command
    insert into dept values ( 44,'text',null);
    -- I don't see the data in WinXP11
    -- No rows in deferror
    -- dba_jobs shows that there is not broken job

    Hi!
    You will need to create a public db link and a private db link connecting to the public link.
    CONN / AS SYSDBA
    CREATE PUBLIC DATABASE LINK WINXP11 USING 'WINXP11';
    CONN REPADMIN/REPADMIN
    CREATE DATABASE LINK WINXP11 CONNECT TO REPADMIN IDENTIFIED BY REPADMIN;
    SELECT SYSDATE FROM DUAL@WINXP11;
    Regards,
    PP

  • Regarding multi master replication

    hi
    i am getting below error while i am doing multi master replication.
    I had master 1 with some entries and now i am doing replication with master-2 which doesn't have any entries.
    [21/May/2007:16:52:15 -0700] - import userRoot: Import complete. Processed 258 entries in 6 seconds. (43.00 entries/sec)
    [21/May/2007:16:52:15 -0700] - INFORMATION - NSMMReplicationPlugin - conn=-1 op=-1 msgId=-1 - multimaster_be_state_change: replica dc=homedepot,dc=com is coming online; enabling replication
    [21/May/2007:16:52:15 -0700] - INFORMATION - NSMMReplicationPlugin - conn=-1 op=-1 msgId=-1 - replica_reload_ruv: Warning: new data for replica dc=homedepot,dc=com does not match the data in the changelog.
    Recreating the changelog file. This could affect replication with replica's consumers in which case the consumers should be reinitialized.
    [21/May/2007:16:52:15 -0700] - INFORMATION - NSMMReplicationPlugin - conn=-1 op=-1 msgId=-1 - This supplier for replica dc=homedepot,dc=com will immediately start accepting client updates
    [21/May/2007:16:52:15 -0700] - INFORMATION - NSMMReplicationPlugin - conn=-1 op=-1 msgId=-1 - Replica (dc=homedepot,dc=com) has been initialized by total protocol as full replica
    [21/May/2007:16:52:16 -0700] - WARNING<10276> - Incremental Protocol - conn=-1 op=-1 msgId=-1 - Replication inconsistency Consumer Replica "alpha.ad.com:19940/dc=homedepot,dc=com" has a different data version. It may have not been initialized yet.
    i initialized both replicas.
    any help is appreciated .

    As Ludovic mentioned, after you load data to the first master server, you have to initialize your second server (either it's master or read-only replica) in two ways:
    1) offline (with server down)
    in server 1: do db2ldif -r and save the output to file
    in server 2: do ldif2db using the saved file
    2) online: using GUI
    I don't think that you need copy over the user schema, which should be able to replicated during replication. (please correct if wrong)

  • Latency at the apply side in multi master replication

    Hi Gurus,
    We are facing some multi master replication latency always at one side . Ex: Site1 -> Site 3 . Transactions from site 1 always gets applied on Site 3 with some latency though the latency doesnt happen while applying from Site 1 - Site 2 for the same transaction originating at site 1 in 3 site(3 way) multi master replication environment.
    We have investigated into it and now looking to check the system replication tables which would be involved at the apply side (ex:- Site 3).
    Could someone pls let me know all the system replication tables that would be involved at the apply side which can impact the latency .I know few of them but dont know the all of them.
    System.def$_AQERROR
    System.def$_ERROR
    System.def$_ORIGIN
    Thanks

    I would say that 50, 50 and 75 are a Very Large number of Job Queue processes. Do you really have that many jobs that need to run concurrently ?
    Since Advanced Replication Queues are maintained in only a small set of tables you might end up having "buffer busy waits" or "read by other session" waits or latch waits.
    BTW, what other factors did you eliminate before deciding to look at the Replication tables ?
    See the documentation on monitoring performance in replication :
    http://download.oracle.com/docs/cd/B10501_01/server.920/a96568/rarmonit.htm#35535
    If you want to look at the "tables" start with the Replication Data Dictionary Reference at
    http://download.oracle.com/docs/cd/B10501_01/server.920/a96568/rarpart4.htm#435986
    and then drill down through the View definitions to the underlying base tables.

  • I am trying to rebuild my iPhoto library and noticed my backup contains aliases (pointers?) and not the actual file. What's the best way to rebuild my library?

    I am trying to rebuild my iPhoto library and noticed my backup contains aliases (pointers?) and not the actual file. What's the best way to rebuild my library?
    Facts:
    In moving to a new iMac, I copied the iPhoto library to an external HDD assuming that I would point the new iMac to the backed up iPhoto Library
    All worked fine when I pointed the new library but noticed that some folders contained aliases and not the original file. So when I attempt to open that photo it can't find it because the alias is pointing to another drive.
    I do have all original photos from a couple of external HDDs. In the folders titled, "Originals" (from older versions of iPhoto) and "Masters" (from current iPhoto)
    I'm thinking I can create a new folder and drop the original files and make that my new iPhoto library. Is there a better way to rebuild my library? I do not want to create any future aliases.
    Thanks in advance for any help!

    do you have a strongly recommended default "managed" library (the iPhoto preference to "copy imported items to the iPhoto library is in its checked state) or a referenced library - you have unchecked that option?
    It sounds like you have a referenced library and are now experiancing one of the very siginificant drawbacks of a referenced library and one of the many reasons they are strongly not recommended
    Also note that iPhoto '11 may use alises in the originals folder as part of the upgrade
    It is important that we understand exactly what you have and what is not sorking - what error messages you are getting
    You must NEVER make any changes of any sort to the structure of content of the iPhoto library - there are no user servicable parts in it  --  and you can not rebuild yoru librtary - only iPhoto ir iPhoto Library Manager - http://www.fatcatsoftware.com/iplm/ -  can rebuild a library unless you are a SQL programmer and understand the structure that iPhoto uses
    LN

  • Multi Master Replication - Only works for some tables?

    I have a multi master replication between a 9i and an 816 database.
    All the tables are in the same tablespace and have the same owner. The replication user has full privs on all the tables.
    When setting up the replication some tables create properly, but others fail with a "table not found" error.
    Ideas anyone ?
    Andrew

    You said that you have a 9i replicated with a 816.
    I try the same thing but with two 9i enterprise version databases, free downloaded from www.oracle.com.
    when i ran
    exec dbms_repcat.add_master_database(gname=>'groupname', master=>'replica_link')
    this error appears
    ERROR at line 1:
    ORA-23375: feature is incompatible with database version at replica_link
    ORA-06512: at "SYS.DBMS_SYS_ERROR", line 86
    ORA-06512: at "SYS.DBMS_REPCAT_MAS", line 2159
    ORA-06512: at "SYS.DBMS_REPCAT", line 146
    ORA-06512: at line 1
    please help me if u have any idea.

  • Multi-Master Replication Limits

    I'm starting new project and would like ask your opinion. They like to have multi-master replication of a single table between Chicago and Denver. A record could be written in Denver, then same row read in Chicago. The opposite is also true. Replication delay should be less than 100 ms. Load is 100 TSP
    I'm not even sure is possible to replicate that fast. What technology should I look at? I heard multi-master replication is not recommended any more.
    Many Thanks,
    Ed Jimenez

    Wich database version are you using?

  • Multi-Master Replication over the WAN

    DS verison: 5.1 sp1
    Did any one implement multi-master replication across the WAN or different IP subnets?

    Sun does mention about DS 5.2 being better than DS 5.1 in WAN based Muli-Master replication with respect to replication performance. I wanted to see if any one out there had implemented (or even played with this topology in their labs) it with out any major hickups.
    Thank you!

  • Best way to handle multi-page documents?

    Hi,
    I'm creating a multi-page editor on top of TLF and I'd like to hear some opinions about the best way to handle multi-pade documents.
    I'm basicaly considering two options:
    1.) A single RichEditableText and several ContainerControllers.
    2.) Several RichEditableText.
    Tks.

    I'm not sure you can create a RichEditableText and give it multiple ContainerControllers. If you use multiple RichEditableText
    components, then you will have to decide in advance which text goes in which one.
    We have posted sample code for doing this in straight ActionScript, which you could presumably also host inside a Flex container. See:
    http://blogs.adobe.com/tlf/examples/
    Look for "ActionScript Pagination Example".
    - robin

  • Procedure to create Multi Master Replication

    Hi,
    can anybody help on this. How to do Multi Master replication from scarch . I used to do single master . this is the first time iam trying to do Multi-Master Replication . I am looking for doc, iam unable to find . Is ther anybody can tell how to do Multi-Master little in detail. I installed my Software and i loaded my company schema and the data on one server. I have another server ready to install software and load data.

    with regards to the post "RTFM"..
    in my experience the sunone/iplanet manuals are very badly written and in some cases just plain wrong..
    sure start with reading the manuals, then make sure you also read the release notes and then you can still be confused in some cases....
    the section on mmr in the manual has a number of mistakes last time i looked..
    one reason why i think community sites, and generally available "support" are such prime requisites in choosing software...
    just my 2 cents

  • Multi-master replication questions for iPlanet 5.0, gurus out there?

    hi:
    I'm using iPlanet Dir Server 5.0 and I note that many gurus out there has
    been able
    to get this to work, that's good, but I have yet to. I have several
    questions, maybe
    someone can spend a few minutes and save me hours...
    I have a suffix called dc=calient,dc=net. I followed the suggestions in
    the
    iPlanet install guide and created 2 directory servers
    a) suffix o=NetscapeRoot, at some arbitrary port, 4601
    b) suffix dc=calient,dc=net, at the usual port 389.
    All my searches/create/delete work fine. However, when I try to replicate
    with multi-master between 2 machines, I keep getting into problems.
    Here's one set of questions...
    Q1: do people out there really split their tree from the o=NetscapeRoot
    tree?
    Q2: The admin guide says the the unit of replication is a database, and
    that each replication can only have 1 suffix. Is this true? Can
    a replicated db have more than 1 suffix?
    Q3: If I also want to replicate the o=NetscapeRoot tree, I have to set
    up yet 2 more replication agreements. Isn't this more work? If
    I just lump the 2 suffixes together, wouldn't it be easier? But would
    it work?
    Q4: I followed the instructions to enable replicas on the masters.
    But then I tried to create this cn=Replication Manager, cn=config
    object.
    But what is the object class of this entry? An iPlanet user has uid
    as its RDN... I tried a person object class, and I added a password.
    But then I keep getting error code 32, object not found in the error
    log. What gives? such as
    WARNING: 'get_entry' can't find entry 'cn=replication
    manager,cn=config', err 32
    Q5: Also, are there any access control issues with this cn=Replication
    Manager,
    cn=config object? By this I mean, I cannot seem to see this object
    using
    ldapsearch, I can only see cn=SNMP, cn=config. Also, do I have
    to give all access via aci to my suffix dc=calient,dc=net? Also,
    given the fact that my o=NetscapeRoot tree is at a different port (say
    4601),
    not 389, could this be an issue?
    Q6: when replication fails, should the Dir Server still come up? Mine does
    not anymore
    which is strange. I keep getting things like this in my log file
    [08/Nov/2001:21:49:13 -0800] NSMMReplicationPlugin - Could not send consumer
    mufasa.chromisys.com:389 the bind request
    [08/Nov/2001:21:49:13 -0800] NSMMReplicationPlugin - Failed to connect to
    replication consumer mufasa.chromisys.com:389
    But why shouldn't the dir server itself come up even if replication
    fails?
    steve

    Hi Steve,
    First, please read the 'Deployment Guide'. I think that is easier to
    understand when you want to setup multi-master replication. The
    'Administrator's Guide' gives you step-by-step instructions, but it may
    not help you to understand how to design your directory services.
    Stephen Tsun wrote:
    I have a suffix called dc=calient,dc=net. I followed the suggestions in
    the
    iPlanet install guide and created 2 directory servers
    a) suffix o=NetscapeRoot, at some arbitrary port, 4601
    b) suffix dc=calient,dc=net, at the usual port 389.
    All my searches/create/delete work fine. However, when I try to replicate
    with multi-master between 2 machines, I keep getting into problems.I don't understand something: which backend do you want to replicate?
    The one holding 'o=NetscapeRoot' or the one holding 'dc=calient,dc=net'?
    Do you want to setup replication between these two instances of the
    directory server (i.e. between port 4601 and 389 in your example)?
    Q1: do people out there really split their tree from the o=NetscapeRoot
    tree?If you have multiple directory servers installed in your environment, it
    is probably worth dedicating (at least) one directory server for the
    o=netscaperoot tree.
    Q2: The admin guide says the the unit of replication is a database, and
    that each replication can only have 1 suffix. Is this true? Can
    a replicated db have more than 1 suffix?Well, it is normal, since in iDS 5.x you have 1 suffix per database.
    You can, however, replicate multiple databases.
    Q3: If I also want to replicate the o=NetscapeRoot tree, I have to set
    up yet 2 more replication agreements. Isn't this more work? If
    I just lump the 2 suffixes together, wouldn't it be easier? But would
    it work?You can't lump the 2 suffixes together, because each backend has 1
    suffix associated with.
    Q4: I followed the instructions to enable replicas on the masters.
    But then I tried to create this cn=Replication Manager, cn=config
    object.
    But what is the object class of this entry?Usually, it is organizationalperson or inetorgperson. In most of the
    cases you want an objectclass which can have userPassword attribute.
    An iPlanet user has uid
    as its RDN... I tried a person object class, and I added a password.
    But then I keep getting error code 32, object not found in the error
    log. What gives? such asYou must have misconfigured something. Or perhaps, it is not
    cn=replication manager, cn=config, but 'uid=replication manager,cn=config'
    Q5: Also, are there any access control issues with this cn=Replication
    Manager,
    cn=config object? By this I mean, I cannot seem to see this object
    using
    ldapsearch, I can only see cn=SNMP, cn=config.The configuration tree is protected by ACIs, so you can not see them
    using anonymous BINDs. Try binding as 'directory manager' and you will
    find your entry.
    Also, do I have
    to give all access via aci to my suffix dc=calient,dc=net?For what purpose? For replication, it is enough to set user DN in the
    replication agreement and this user can update the replicated backend.
    Q6: when replication fails, should the Dir Server still come up?Yes.
    Bertold

Maybe you are looking for