Triggering replication

How does the coherence framework decide when replication should be triggered on a cached object ? Do we have to invoke put() to ensure that any changes to the object (if any) are replicated to peer caches ?
     The objects we wish to cache have large object graphs, of which any of the underlying elements may change based upon events in our system, or based upon user interaction. The object is heavy on reads and heavy on writes.
     Thanks.

How does the coherence framework decide when     > replication should be triggered on a cached object ?
     > Do we have to invoke put() to ensure that any
     > changes to the object (if any) are replicated to
     > peer caches ?
     When using the Coherence API, yes.
     When using Coherence*Web session management, no.
     > The objects we wish to cache have large object
     > graphs, of which any of the underlying elements may
     > change based upon events in our system, or based upon
     > user interaction. The object is heavy on reads and
     > heavy on writes.
     Objects with large object graphs are not optimal for Coherence to deal with.
     You can use entry processors (see the InvocableMap API) to leave them "in place" and access/manipulate them remotely. That is optimal for heavy writes, and (due to the number of changes) acceptable for heavy reads.
     Breaking up the graph, if conceptually possible, would be a more optimal approach. Are you already working with an account manager at Tangosol, since they can allocate appropriate engineering resources to help you with this? Would you want to set up a discussion with one of our architects to determine an approach?
     Peace,
     Cameron Purdy
     Tangosol Coherence: The Java Data Grid

Similar Messages

  • Transactional replication triggers disabled after server restart

    Hello Everyone,
    I have been facing this issue since couple of days. I have configured transactional replication with updatable subscriber in sql server 2008 R2.  It was working fine till date. Recently we are facing an issue where subscriber whenever restarts, one
    of the table's replication triggers(insert, update and delete) are disabled automatically. We have to manually enable all the triggers for that table. Other tables are working fine. Please help. Thanks in advance.
    Regards,
    Sai

    Have you checked information from replication monitor?
    You should disable all triggers because you don't want SQL Server to fire the trigger on the Subscriber database when the database receives an insert, update, or delete operation. Because the trigger has already fired on the Publisher database, firing the trigger
    on the Subscriber database might repeat the transaction.
    http://technet.microsoft.com/en-us/library/ms151718(v=sql.110).aspx
    http://msdn.microsoft.com/en-us/library/ms152529.aspx

  • Missing privilege:NO TRIGGERS FOR REPLICATION USER

    Hello,
    I'm configuring the Synchronization Manager following the steps on the SAP MaxDB Library 7.6. I've done every step until the activation of the Replication Units, where the Synchronization Manager GUI give me the following error:
    +Create MaxDB Trigger
    com.sap.dbtech.jdbc.exceptions.DatabaseException: [-5001]: Missing privilege:NO TRIGGERS FOR REPLICATION USER
         at com.sap.dbtech.jdbc.packet.ReplyPacket.createException(ReplyPacket.java:65)
         at com.sap.dbtech.jdbc.ConnectionSapDB.throwSQLError(ConnectionSapDB.java:1061)
         at com.sap.dbtech.jdbc.ConnectionSapDB.execute(ConnectionSapDB.java:689)
         at com.sap.dbtech.jdbc.ConnectionSapDB.execute(ConnectionSapDB.java:563)
         at com.sap.dbtech.jdbc.StatementSapDB.sendCommand(StatementSapDB.java:855)
         at com.sap.dbtech.jdbc.StatementSapDB.sendSQL(StatementSapDB.java:919)
         at com.sap.dbtech.jdbc.StatementSapDB.execute(StatementSapDB.java:266)
         at com.sap.dbtech.jdbc.StatementSapDB.execute(StatementSapDB.java:234)
         at com.sap.dbtech.jdbc.trace.Statement.execute(Statement.java:79)
         at com.sap.sdb.syncMan.util.SQLBuilder.markMaxDBTableForRep(SQLBuilder.java:301)
         at com.sap.sdb.syncMan.design.DesignSQLLayer.installTableOfParticipant(DesignSQLLayer.java:2427)
         at com.sap.sdb.syncMan.design.DesignSQLLayer.installReplicationUnit(DesignSQLLayer.java:1493)
         at com.sap.sdb.syncMan.gui.jface_controls.UnitControl.internalActivateParticipant(UnitControl.java:1165)
         at com.sap.sdb.syncMan.gui.jface_controls.UnitControl.internalActivateSelectedParticipant(UnitControl.java:1067)
         at com.sap.sdb.syncMan.gui.jface_controls.UnitControl.access$000(UnitControl.java:98)
         at com.sap.sdb.syncMan.gui.jface_controls.UnitControl$WorkerThread.run(UnitControl.java:125)+
    Question:
    1. What privilege should I give to the Synchronization User? It's already a DBA user.
    2. Could it be a library problem?. According to the documentation installed with the application, the Mysql connector jdbc should be used instead of the sapdbc.jar; but doing this the Synchronization Manager GUI doesn't start.
    Regards!
    MY

    Hello,
    you can create the replication user like this.
    "create user dbservice password <your password> dba not exclusive replication"
    Did you grant the tables you want to replicate to the replication user?
    The replication user itself must not own triggers or tables. It owns special system triggers and repliaction meta tables.
    Best Regards
      Wolfgang

  • Help needed for replication environment. atleast get me good URL links.

    Hi friends,
    Our database is having nearly 150 tables and 200 views, 10 database procedures, 5 database triggers, 5 DBMS_JOB, and 15 users are using. it is ORACLE 8i under win 2000 environemnt. daily 5000 new records are inserted into the database.
    now i want to have a standby database, read and write complete replication of master database, in this following scenario.
    all the clients are conneting uisng connect string "terminal", if any thing happens to the "terminal" computer, the users could be able able to connect using "terminal1" lets say this "terminal1" is the stand by database and it should be functional as original.
    any body can give documentation, easy methods, or step by step documentation. we have access to metalink also. but failed to get easy methods to achieve this.
    thanks in advance

    is a very complicated thing you want to do: it is what Oracle call the Standby database or Dataguard - it's part of the Enterprise Edition, but I don't know whether Oracle charge extra for it. If you are not an experienced Oracle DBA (and, forgive me, but some of your other posts suggest you are not) you do not want to try doing this yourself. Pay Oracle's wergild and get their solution for this.
    Replication is really not suitable for this sort of process as it doesn't function well in real-time; this means you cannot guarantee keeping the two databases usably up-to-date. That causes a problem later on when you need to switch to your back-up database. And again when your original database is back online. Plus replication is flaky and has performance costs (voice of experience).
    Cheers, APC

  • Using database link in my triggers,urgent

    There's DB A on ServerA,DB B on ServerB.Table TA in ServerA's DB and TB on ServerB's DB.My purpose is to synchronize usera.TA with userb.TB real time.
    In order to realize it,I created dblink AtoB from A to B using command "create database link AtoB.domain connect to userb identified by userb using 'BTNSname';",it's no problem for I have test on it :select count(*) from [email protected];It's ok to get the result.
    But when I create trigger on table TA by using the dblink AtoB.domain to update,delete and insert datas to TB,the error "PLS-00201: identifier '[email protected]' must be declared" occurs while compiling.
    The oracle version on the two servers is 8.1.7.
    It happens on the production envrionment,but in our lab,there's no problem.The difference between them is that it is cluster structure in production one.
    I tried several times,but failed,just wondered if some privilege was not right?But I can do selection from this dblink,only in trigger there's compilation errors.
    Really hope that someone could give light on it.
    Thank you in advance!
    In addition,I did some modification in my triggers to have a try.
    I created a synonym for the remote table,and the trigger then updates the synonym instead of the remote table [email protected] .But also failed.
    The same error,saying the synonym is not declared.
    Any good way to realize the same thing?

    Thank you for your answering first.
    I don't use replication because the destination table's structure is not exactly the same as the source one.
    --The selection was done in the production,connection to the source schema,then select through the dblink from the source to the dest schema,the trigger is created in source schema on source table to handle datas in destination schema's table.In lab,all is ok,no such a problem.
    --I created two dblinks,one is connect to source as system user,created a public dblink "create public database link AtoB.domain using 'BTNSname';";the other is connect as UserA,"create database link AtoB.domain connect to UserB identified by UserB using 'BTNSname';".                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • No Material Master replication from R/3 to CRM

    Hallo guys,
    i've got no replication of material master, haven't tried any other object except customizing but customizing is triggered from CRM and was replicated fine. Can anybody help me how can i debug the replication process from R/3?
    Thanks for your help in advance.

    Dmitry,
    You first need to determine where your errors are happening?
    1.  Do you have blocked queues on CRM?  If you see a blocked inbound/outbound queue on CRM via SMQ1 or SMQ2, then you can try to rerun the BDOC message in the queue.
    2.  Is there tax code customizing not downloaded correctly into CRM?  This sometimes can hang up your download of materials?
    3.  Have you looked at SMW01 after the load to see if you have any BDOC's in error, the queues were empty?
    4.  What is the status of the initial load R3AM1?
    I would hold off from debugging until I figure out where in the download process that I am receiving the error.  Once I know that fact them I can debug a single download.  You will have to set a filter on the material object to a single material and attempt an initial download.  When you perform the initial download you will need all queues deregistered on the CRM side, so that you manually execute the BDOC import into CRM. 
    The last step will require placing a breakpoint into the validation function module for material.
    I believe validation function module is: COM_PRODUCT_MAT_VALIDATE.  If you have questions along the way of debugging the flow, please reply back and I will try to see if I can answer them.  It has been a little while since I had to debug our middleware downloads.
    Good Luck,
    Stephen

  • Z table middleware replication from R/3 to CRM

    Looking for a CRM Middleware Guru out there !!!
    I'm trying to replicate a simple R/3 table (list of countries in my example below. The replication must include both initial and delta downloads.
    My problem is the following:
    When doing an initial load using Transaction R3AS the BDOC is created correctly.  
    The traffic lights are green in transaction SMW01 with a status of Confirmed (fully processed). 
    The segments of the BDOC are populated with the correct value from R3 <b>BUT</b> <b>the table on CRM is not populated.</b>
    Here is what I have done so far:
    1-Created table on R3 ZCOUNTRY
    2-Created mapping Function on R3 Z_CRS_MAP_ZCOUNTRY_BAPIMCTS
    3-Created extract Function on R3 ZCOUNTRY_EXTRACT
    4-Created an entry in Table CRMSUBTAB for Function ZCOUNTRY_EXTRACT
    5-Created Receiving Module in R3 Z_CRM_COUNTRY_COLLECT_DATA triggered on SM30 for the Delta Load
    6-Created table on CRM ZCOUNTRY
    7-Created Bdoc on CRM ZCOUNTR_MESG
    8-Created validation Function on CRM ZCRM_ZCOUNTR_MESG_VAL
    9-Add an entry in SMW3BDOCIF for BDOC ZCOUNTR_MESG and Validation Function ZCRM_ZCOUNTR_MESG_VAL
    10-Created mapping Function on CRM ZMAP_COUNTRY_BAPIMTCS_TO_MBDOC
    11-Created MW Adapter Object in CRM ZCRMCOUNTRY
    12-Created a replication Object on CRM ZCOUNTR_MESG
    13-Created a publication on CRM ZCOUNTRY
    I'm not sure why the CRM does not get populated. Could it have anything to do with "GUID" keys?
    The 2 fields on my bespoke table are:
    -Country
    -Continent
    In my Root Data Structure for the MBDOC do I need to add GUID as the first fields i.o.w.
    -GUID
    -Country
    -Continent
    When I choose my segment fields do I choose (?):
    -GUID
    -Country
    -Continent
    In my CRM Mapping function how do I populate the GUID field? (see the code below in the mapping function).  How do I get a value into lv_sfakey and does this have anything to do with the GUID mentioned above.
    MOVE lv_sfakey TO segment_line-guid.
    AND
    MOVE lv_sfakey TO ext_line-data-guid.
    Thanks for any help
    Manoj

    Hi Manoj,
    I have a similar situation as you have described. However my problem is that when i start the initial load, the status in the RFC inbound queue of CRM (smq2) shows stop and the oject shows the status running. I dont know how to debug the
    Validation FM and Mapping FM written for the same. Can you please explain me the meaning of the status shown in the crm queue.
    Thanks,
    Vaishali

  • How to stop the replication of a particular business partner from sap crm to sap ecc - BDOC

    Hi experts,
    I need to know the procedure to stop replicating a particular BP #1234(example) from CRM to ECC. I don't want this BP to be replicated in ECC as in ECC this BP will not be maintained.
    BDOC failure:
    In my current scenario BDOCs are failing with error like "No BP exist". Since this BP #1234 itself is not present in ECC, the error is obvious.
    This BP #1234 is defined in CRM as an Org. Unit. We don't maintain Org. Unit in ECC. All other BP are saved under this BP(#1234).
    Issue:
    So whenever a new BP(example #ABC) is created, the BUPA_REL type BDOC fails with error "BP with number #1234 doesnot exist.
    I don't want this to happen. So what i found if we can stop the replication of this particular BP #1234, then if we make any change to the relation BP, this error should not come.
    Please help me to resolve this BDOC error.
    regards,
    Vicky

    Hi Ganesh,
    1st we create a Organization structure through tcode-PPOSA_CRM.
    A business Partner is created which represent the Organization Structure partner ID.
    2nd we create a BP through and do a Org. assignment in tcode - PPOSA_CRM.
    A business Partner is created which represent the any customer partner ID.
    As soon as you do an assignment,assigning the customer BP to the org BP, a BUPA_REL BDoc flow but it fails.
    Error we get in the BDoc: "BP XXXX doesnot exist"
    Reason: Since this Org BP is not present in ECC, the BDoc fails in CRM.
    Possible Resolution 1:
    1. I created this Org. BP in ECC through CRMM_BUPA_MAP
    2. The BP got created.
    3.Then i tried to assign a new customer BP to Org BP to check if the issue is resolved, but we get another type of error 'BP is not defined as internal Org. Unit".
    I don't know how to define BP as an internal Org. Unit in ECC?
    Can you please help me in this.
    Possible Resolution 2:
    If i can stop this particular BUPA_REL BDoc which contains this Org.unit BP, from getting replicated, means when i assign the customer BP to Org.Unit BP, BUPA_REL should not be triggered.
    I don't know if this is possible or not, but if this is possible, i would consider this a last option to resolve this issue.
    Please let me know if this helps you to understand the issue i am facing?
    regards,
    Vicky

  • Replication of Z table from CRM to R/3 - No mBDoc Created

    I need to transfer the contents of a bespoke customer table from CRM into R/3, off the back of delta changes being made to the CRM table.  To help us to achieve this we have performed the following steps so far:
    1. Created the customer table in both systems.
    2. Created a new messaging BDoc in CRM and linked it to the R/3 Site Type.
    3. Created a new mapping module in CRM that takes the data from the BDoc and maps it to the BAPI structure.
    4. Created a new Adapter Object that links to my BDoc, contains the new customer table as the source table in CRM and contains the mapping Module mentioned above. 
    5. Created a new Replication Object based on my new BDoc.
    6. Created a new Publication and assigned it to the Replication Object.
    7. Created a new Subscription and assigned to the Publication and Replication Object. Also assigned it to my R/3 site.
    On the R/3 side we have created a mapping module to map the data from the BAPI structure into the equivalent R/3 table. We also have an entry in table CRMSUBTAB.
    However when I insert an entry in the customer table in CRM no BDoc is being created. In fact I cannot see anything at all in the system that indicates that it has even tried to capture the change and invoke the Middleware process.
    What am I doing wrong?
    Do I need anything else (some sort of delta program?) that picks up the parameters from the table update and feed them into my process?  The literature that I have found (and it is not much) does not mention anything like this though.
    Any help would be greatly appreciated as this is now a very urgent requirement.
    Regards
    Ian
    Edited by: IAN HAWLEY on Aug 21, 2008 9:42 AM

    Ian,
    I do expected the follow up questions. Check my explanation below and hope it will answer your queries:
    1. I assume all of the activities performed to date are still valid to supplement your solution, e.g. the BDoc, Replication Object, Publication and Subscription details?
    2. The R/3 to CRM Mapping Module. Is this required to allow messages to be sent back from R/3 to update the BDoc, e.g. a sort of validation to prove that the posting has completed ok?
    FM ZMAP_BAPIMTCS_TO_MBDOC in CRM, to map the BAPIMTCS format data and build the BDoc. This BAPIMTCS format is a temporary one and is not the final data format, that is taken to ECC. This function module also takes care of receiving the response message from ECC, once the BDoc data reaches and updates in ECC. If there is any error occured during the updation, it is captured in the error table of the BDoc and the status of BDoc is set to 'Error'. If no error occurs, the status of BDoc is set to 'Confirmed'.
    3. The Extractor Module in CRM. Does this get the data out of the table and will it work for deltas?
    Yes , It should work for Delta too.The delta load makes use of the same program and flow for Initial load (SMOF_DOWNLOAD)
    4. CRMSUBTAB in CRM. I knew that we populated this in R/3, I did not realise we would need it in CRM as I assumed it was R/3 specific.
    5. You list the sequences of FM calls at the end. I was confused in the order. As we are initiating data to be sent from CRM to R/3 should some of these be in the reverse order, e.g. ZMAP_MBDOC_TO BAPIMTCS would be called before ZMAPBAPIMTCS_TO_MBOC as would pass data into the BDoc to send it to R/3 before we then received an update message back?
    Step 1: Z_EXTRACT_MODULE  will be called.( It calles ZPICK_DATA_FROM_CRM). This function module calls the standard function module CRS_SEND_TO_SERVER the one triggers the other function modules
    Setp 2: Creat function module ZDATA_TO_BAPIMTCS(missed to mention earlier) in CRM, to map the data in the final internal table to BAPIMTCS format. This format is temporary and will be used to build the BDoc data.
    Step 3: Created function module ZMAP_BAPIMTCS_TO_MBDOC in CRM, to map the BAPIMTCS format data and build the BDoc. This BAPIMTCS format is a temporary one and is not the final data format, that is taken to ECC. This function module also takes care of receiving the response message from ECC, once the BDoc data reaches and updates in ECC. If there is any error occured during the updation, it is captured in the error table of the BDoc and the status of BDoc is set to 'Error'. If no error occurs, the status of BDoc is set to 'Confirmed'
    Step 4: Created function module ZMAP_MBDOC_TO_BAPIMTCS in CRM, to build the final BAPIMTCS structure from the BDOC. This BAPIMTCS is the final data structure that goes to ECC. The table name, objectkey, relkey that is relevant for the BAPIMTCS, is filled in this function module..
    Step 5: Created function module Z_LOAD_PROXY_FINAL in ECC, to receive the data from CRM. The BAPIMTCS data is received and mapped to local internal tables and then updates to custom tables through the function module Z_UPDATE_ECC. The errors if any are captured in this function module and returned back to CRM using the standard function module CRS_SEND_TO_SERVER.
    To reduce the load on the interface, at the final stage, it was decided to fetch the data completely in ECC. So the incoming  data from CRM is ignored and is fetched completely from ECC tables.
    6. Is there a test FM available for the extract, e.g. is CRM_SAMPLE_EXTRACT_MODULE the one to copy?
        No, You have to develop this Extractor FM say ZPICK_DATA_FROM_CRM and should be called in Z_EXTRACT_MODULE.
    Apologize for any spelling errors, as I too running to meeting.
    Update me the status.
    Bobby
    Edited by: Bobby on Aug 22, 2008 2:13 PM

  • Event ID: 5014, 5004 The DFS Replication Service is stopping communication with partner / Error 1726 (The remote procedure call failed.)

    I'm replicating between two servers in two sites (Server A - Server 2012 R2 STD, Server B - Server 2008 R2) over a VPN (Sonicwall Firewall).  Though the initial replication seems to be
    happening it is very slow (the folder in question is less than 3GB).  I'm seeing these in the event viewer every few minutes:
    The DFS Replication service is stopping communication with partner PPIFTC for replication group FTC due to an error. The service will retry the connection periodically.
    Additional Information:
    Error: 1726 (The remote procedure call failed.)
    and then....
    The DFS Replication service successfully established an inbound connection with partner PPIFTC for replication group FTC.
    Here are all my troubleshooting steps (keep in mind that our VPN is going through a SonicWall <--I increased the TCP timeout to 24 hours):
    -Increased TCP Timeout to 24 hours 
    -Added the following values on both sending and receiving members and rebooted server
    HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Services\Tcpip\Parameters
    Value =DisableTaskOffload
    Type = DWORD
    Data = 1
    Value =EnableTCPChimney
    Type = DWORD
    Data = 0
    Value =EnableTCPA
    Type = DWORD
    Data = 0
    Value =EnableRSS
    Type = DWORD
    Data = 0
    ---------------------------------more troubleshooting--------------------------
    -Disabled AntiVirus on both members
    -Made sure DFSR TCP ports 135 & 5722 are open
    -Installed all hotfixes for 2008 R2 (http://support.microsoft.com/kb/968429) and rebooted
    -Ran NETSTAT –ANOBP TCP and the DFS executable results are listed below:
    Sending Member:
    [DFSRs.exe]
      TCP    10.x.x.x:53            0.0.0.0:0             
    LISTENING       1692
    [DFSRs.exe]
      TCP    10.x.x.x:54669        
    10.x.x.x:5722          TIME_WAIT       0
      TCP    10.x.x.x:54673        
    10.x.x.x:5722          ESTABLISHED     1656
     [DFSRs.exe]
      TCP    10.x.x.x:64773        
    10.x.x.x:389           ESTABLISHED     1692
    [DFSRs.exe]
      TCP    10.x.x.x:64787        
    10.x.x.x:389           ESTABLISHED     1656
     [DFSRs.exe]
      TCP    10.x.x.x:64795        
    10.x.x.x:389           ESTABLISHED     2104
    Receiving Member:
    [DFSRs.exe]
      TCP    10.x.x.x:56683        
    10.x.x.x:389           ESTABLISHED     7472
     [DFSRs.exe]
      TCP    10.x.x.x:57625        
    10.x.x.x:54886         ESTABLISHED     2808
    [DFSRs.exe]
      TCP    10.x.x.x:61759        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61760        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61763        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61764        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61770        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61771        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61774        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61775        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61776        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61777        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61778        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61779        
    10.x.x.x:57625         TIME_WAIT       0
      TCP    10.x.x.x:61784        
    10.x.x.x:52757         ESTABLISHED     7472
    [DFSRs.exe]
      TCP    10.x.x.x:63661        
    10.x.x.x:63781         ESTABLISHED     4880
    ------------------------------more troubleshooting--------------------------
    -Increased Staging to 32GB
    -Opened the ADSIedit.msc console to verify the "Authenticated Users" is set with the default READ permission on the following object:
    a. The computer object of the DFS server
    b. The DFSR-LocalSettings object under the DFS server computer object
    -Ran
    ping <var>10.x.x.x</var> -f -l 1472 and got replies back from both servers
    -AD replication is successful on all partners
    -Nslookup is working so DNS is working
    -Updated NIC drivers on both servers
    - I ran the following to set the Primary Member:
    dfsradmin Membership Set /RGName:<replication group name> /RFName:<replicated folder name> /MemName:<primary member> /IsPrimary:True
    Then Dfsrdiag Pollad /Member:<member name>
    I'm seeing these errors in the dfsr logs:
    20141014 19:28:17.746 9116 SRTR   957 [WARN] SERVER_EstablishSession Failed to establish a replicated folder session. connId:{45C8C309-4EDD-459A-A0BB-4C5FACD97D44} csId:{7AC7917F-F96F-411B-A4D8-6BB303B3C813}
    Error:
    + [Error:9051(0x235b) UpstreamTransport::EstablishSession upstreamtransport.cpp:808 9116 C The content set is not ready]
    + [Error:9051(0x235b) OutConnection::EstablishSession outconnection.cpp:532 9116 C The content set is not ready]
    + [Error:9051(0x235b) OutConnection::EstablishSession outconnection.cpp:471 9116 C The content set is not ready]
    ---------------------------------------more troubleshooting-----------------------------
    I've done a lot of research on the Internet and most of it is pointing to the same stuff I've tried.  Does anyone have any other suggestions?  Maybe I need to look somewhere
    else on the server side or firewall side? 
    I tried replicating from a 2012 R2 server to another 2012 server and am getting the same events in the event log so maybe it's not a server issue. 
    Some other things I'm wondering:
    -Could it be the speed of the NICs?  Server A is a 2012 Server that has Hyper-V installed.  NIC teaming was initially setup and since Hyper-V is installed the NIC is a "vEthernet
    (Microsoft Network Adapter Multiplexor Driver Virtual Switch) running at a speed of 10.0Gbps whereas Server B is running a single NIC at 1.0Gbps
    -Could occasional ping timeout's cause the issue?  From time to time I get a timeout but it's not as often as the events I'm seeing.  I'm getting 53ms pings.  The folder
    is only 3 GB so it shouldn't take that long to replicate but it's been days.  The schedule I have set for replication is mostly all day except for our backup times which start at 11pm-5am.  Throughout the rest of the time I have it set anywhere from
    4Mbps to 64 Kbps.  Server A is on a 5mb circuit and Server B is on a 10mb circuit. 

    I'm seeing the same errors, all servers are running 2008 R2 x64. Across multiple sites, VPN is steady and reliably.
    185 events from 12:28:21 to 12:49:25
    Events are for all five servers (one per office, five total offices, no two in the same city, across three states).
    Events are not limited to one replication group. I have quite a few replication groups, so I don't know for sure but I'm running under the reasonable assumption that none are spared.
    Reminder from original post (and also, yes, same for me), the error is: Error: 1726 (The remote procedure call failed.)
    Some way to figure out what code triggers an Event ID 5014, and what code therein specifies an Error 1726, would extremely helpful. Trying random command line/registry changes on live servers is exceptionally unappealing.
    Side note, 1726 is referenced here:
    https://support.microsoft.com/kb/976442?wa=wsignin1.0
    But it says, "This RPC connection problem may be caused by an unstable WAN connection." I don't believe this is the case for my system.
    It also says...
    For most RPC connection problems, the DFS Replication service will try to obtain the files again without logging a warning or an error in the DFS Replication log. You can capture the network trace to determine whether the cause of the problem is at the network
    layer. To examine the TCP ports that the DFS Replication service is using on replication partners, run the following command in a
    Command Prompt window:
    NETSTAT –ANOBP TCP
    This returns all open TCP connections. The connections in question are "DFSRs.exe", which the command won't let you filter for.
    Instead, I used the NETSTAT command as advertised, dumping output to info.txt:
    NETSTAT -ANOBP TCP >> X:\info.txt
    Then I opened Excel and manually opened the .TXT for the open wizard. I chose fixed-width fields based on the first row for each result, and then added a column:
    =IF(A3="Can not", "Can not obtain ownership information", IF(LEFT(A3,1) = "[", A3&B3&C3, ""))
    Dragging this down through the entire file let me see that row (Row F) as the file name. Some anomalies were present but none impacted DFSrs.exe results.
    Finally, you can sort/filter (I sorted because I like being able to see everything, should I choose to) to get just the results you need, with the partial rows removed from the result set, or bumped to the end.
    My server had 125 connections open.
    That is a staggering number of connections to review, and I feel like I'm looking for a needle in a haystack.
    I'll see if I can find anything useful out, but a better solution would be most wonderful.

  • Hetrogenous replication in sql server 2008

    I am trying to setup a transaction rpelication between a Oracle publisher (version 10.2) and sql server 2008 sp1 on windows 2008, 64 bit AMD Processor.
    I have installed the oracle client 11.0.1 and followed the permissions for the login on the orcale server.
    I was able to add the oracle publusher.
    I am tryign to create the publication and atricles and i get the following error.
    TITLE: Publication Properties
    Cannot add article 'CACHE_LEVEL_LU'.
    ADDITIONAL INFORMATION:
    An exception occurred while executing a Transact-SQL statement or batch. (Microsoft.SqlServer.ConnectionInfo)
    Cannot publish the table 'SAT.CACHE_LEVEL_LU' from the Publisher 'CD15'. Verify connection information and ensure that you can connect to the Publisher through a tool like SQL*PLUS. Ensure that the replication administrative user schema has the required permissions.
    Cannot add the Publisher triggers and the article log table to the Oracle Publisher for the article 'CACHE_LEVEL_LU'. Verify connection information and ensure that you can connect to the Publisher through a tool like SQL*PLUS. Ensure that the replication administrative user schema has the required permissions.
    Changed database context to 'distribution'.
    OLE DB provider "OraOLEDB.ORACLE" for linked server "CD15" returned message "ORA-00936: missing expression
    ORA-06512: at "BIA_REPL.HREPL", line 1722
    ORA-06512: at line 1".
    Error: 7215, Sev: 17, State: 1, Msg: Could not execute statement on remote server 'CD15'. (Microsoft SQL Server, Error: 21790)
    For help, click: http://go.microsoft.com/fwlink?ProdName=Microsoft+SQL+Server&ProdVer=10.00.2531&EvtSrc=MSSQLServer&EvtID=21790&LinkId=20476
    BUTTONS:
    OK
    Any idea?
    Do you guys knwo fi the replication between Oracle and sql works?
    thanks
    uday

    Did you ever find a solution here? I'm running into the exact same issue with my attempted replication setup - I'm hopeful that you were able to solve this!

  • 1:N Replication Scenario Approach

    Hi,
    We plan to split HANA and like to consider the parallel run option whereby we build the new hardware and do a migration/restore from existing Production onto the new box. Then we like to switch on SLT for both the existing and new HANA boxes and run the models in parallel to compare the results before we switch off the old one.
    This would mean that we need to setup new SLT triggers for the new server whilst the existing triggers are still running.The current configuration for SLT -- Allow Multiple Usage is not selected.How we can perform this in our landscape?What would be the best approach?

    Hi Roy,
    look at this note. it describes how to setup 1:n once you already created the config without "multiple usage" flagged.
    Best,
    Tobias
    SAP Note 1898479:
    SLT replication: Redefinition of existing DB triggers

  • Problem on Material Replication from R/3 to EBP

    Hi friends,
        I am having a problem on Material replication from R/3 to EBP. I had configured all the required settings in EBP as well as in R/3. Whenever we are starting initial loading R3AS I am getting outbound queues in EBP instead of R/3. I am getting the error like 'The current application triggered a termination'. I could able to replicate Vendor and Locations from R/3 to EBP. If anybody can solve this problem.
    Thanx in advance.
    Raj

    Hi Ahmed,
        Could you please elaborate. I am getting runtime error as 'Message_Type_X' (The current application triggered as termination with a short dump).
    Thanks
    Raj

  • How to trigger replication from CRM to ECC when changing custom z-field?

    Hi all,
    I've created a z-field in CRM which matches the field KNVV-KLABC in ECC. I've created the needed user exits to replicate from ECC to CRM and vice versa. The functions are called and update successfully. So actually everything works fine except for one thing...
    When I change the z-field in CRM - and only that one field - the replication is not triggered hence KNVV-KLABC is not updated. When I however change my z-field together with another pricing field (e.g. Customer Group) then the replication runs and both fields are updated in ECC.
    How can I trigger the replication if only my z-field is changed?
    Best regards,
    Melanie

    Hi there,
    Thanks for your replies.
    I've solved this myself a couple of days ago. However I did it in a different way as suggested here. I will post my solution as soon as I find the time to do the documentation.
    Maybe as a hint if someone runs into this; adding a field to the pricing structure of the BP sales areas (SET0030 if you need to know) seems strangely enough "not possible" - I mean it's "not ready yet".
    AET f.i. doesn't allow you to add a field there and with EEWB you cannot choose to replicate it to R3. Thus for my solution I had to change SAP Standard unfortunatelly. Also the my customer wants to be able to search for this new field on the Account Search Page - so many steps were needed to achieve all this. As I mentioned I shall present my solution as soon as I get the time for it. If I should forget and you need to know how I did it you may reply to this thread as I'm watching it so I'll be reminded.
    Best regards,
    Melanie

  • Transactional replication very slow with indexes on Subscriber table

    I have setup Transactional Replication for one of our databases where one table with about 5mln records is replicated to a Subsriber database. With every replication about 500-600.000 changed records are send to the Subscriber.
    Since one month I see very strange behaviour when I add about 10 indexes to the Subscriber table. As soon as I have added the indexes replication speed becomes extremely slow (almost 3 hours for 600k records). As soon as I remove the indexes the replication
    is again very fast, about 3 minutes for the same amount of records.
    I've searched a lot on the internet to solve this issue but can't find any explaination for this strange behaviour after adding the indexes. As far as I know it doesn't have to be a problem to add indexes to a Subscriber table, and it hasn't been before on
    another replication configuration we use.
    Some information from the Replication Log:
    With indexes on the Subscriber table
    Total Run Time (ms) : 9589938 Total Work Time : 9586782
    Total Num Trans : 3 Num Trans/Sec : 0.00
    Total Num Cmds : 616245 Num Cmds/Sec : 64.28
    Total Idle Time : 0 
    Writer Thread Stats
    Total Number of Retries : 0 
    Time Spent on Exec : 9580752 
    Time Spent on Commits (ms): 2687 Commits/Sec : 0.00
    Time to Apply Cmds (ms) : 9586782 Cmds/Sec : 64.28
    Time Cmd Queue Empty (ms) : 5499 Empty Q Waits > 10ms: 172
    Total Time Request Blk(ms): 5499 
    P2P Work Time (ms) : 0 P2P Cmds Skipped : 0
    Reader Thread Stats
    Calls to Retrieve Cmds : 2 
    Time to Retrieve Cmds (ms): 10378 Cmds/Sec : 59379.94
    Time Cmd Queue Full (ms) : 9577919 Full Q Waits > 10ms : 6072
    Without indexes on the Subscriber table
    Total Run Time (ms) : 89282 Total Work Time : 88891
    Total Num Trans : 3 Num Trans/Sec : 0.03
    Total Num Cmds : 437324 Num Cmds/Sec : 4919.78
    Total Idle Time : 0 
    Writer Thread Stats
    Total Number of Retries : 0 
    Time Spent on Exec : 86298 
    Time Spent on Commits (ms): 282 Commits/Sec : 0.03
    Time to Apply Cmds (ms) : 88891 Cmds/Sec : 4919.78
    Time Cmd Queue Empty (ms) : 1827 Empty Q Waits > 10ms: 113
    Total Time Request Blk(ms): 1827 
    P2P Work Time (ms) : 0 P2P Cmds Skipped : 0
    Reader Thread Stats
    Calls to Retrieve Cmds : 2 
    Time to Retrieve Cmds (ms): 2812 Cmds/Sec : 155520.63
    Time Cmd Queue Full (ms) : 86032 Full Q Waits > 10ms : 4026
    Can someone please help me with this issue? Any ideas? 
    Pim 

    Hi Megens:
    Insert statement might be slow with not only indexes and few others things too
    0) SQL DB Blocking during inserts
    1) If any insert triggers are existed
    2) Constraints - if any
    3) Index fragmentation
    4) Page splits / fill factor
    Without indexes inserts will be fast because, each time when new row going to insert to the table, SQL Server will do
    1) it will check for the room, if no room page splits will happen and record will placed at right place
    2) Once the record updated all the index should be update
    3) all these extra update work will cause can make insert statement bit slow
    Its better to have index maintenance jobs frequently to avoid fragmentation.
    If every thing is clear on SQL Server Side, you need look up on DISK IO, N/W Latency between the servers and so on
    Thanks,
    Thanks, Satish Kumar. Please mark as this post as answered if my anser helps you to resolves your issue :)

Maybe you are looking for