Multimaster Replication problem

Hello,
I've configured a multimaster replication environment with two master sites and a master group whith both of them and asynchronous replication. My problem is that in the master def the master group appeared stopped and I don´t know how activate it, and far over, the administrative requests in my replicated databases aren't executed by themselves, in the masterdef I can't depure them but I've to do it explicitily and in the other master site there is no way to get this requests dissapear.
The main problem is that in the secondary master site the replication status is always quiesced whereas in the masterdef the status is normal if there is no objects in the master group and quiesced when I add some objects even though I generate the replication support for all the new objects.
Any help would be very useful because I'm beginning with replication and I've no found any information in the 2 manuals of Oracle about this topic.
Thanks in advance.
Mabel

you can find reason from dba_repcatlog view. I will suggest that you must recreate whole environment and then create your replication freshly. remeber that you replicated table must have primary key coulmn or you have configured suitable conflict resolution method for these tables. Moreover after generating replication support for the objects check dba_repcatlog view. this view should be empty and there is no error located in this view. if there is any error see message from this view and solve it and regenerate you replication support for the object and again check. when you complete all error then resume your replication activity.

Similar Messages

  • Synchronous multimaster replication Backup strategy using RMAN

    Hi all,
    I am using synchronous multimaster replication. my question is backup strategy has to perofirm in all the master site or only in any of 1 mastersite? what all problems will encounter in RMAn backup in my scenario please help me out with your suggestion
    nagaraj.K
    [email protected]

    You ask: "I want to configure backup strategy using RMAN. any one can help me that"
    And the answer is clearly no we can not.
    An RMAN backup strategy depends on your SLA (Service Level Agreement) with your customers that you didn't post. What is your down-time window? What is your skill set?
    You ask: "How to configure for RMAN Incremental backup?"
    Read the docs at http://tahiti.oracle.com for your version (which you didn't mention).
    You ask: "What will be backup space and there size ?"
    We have no idea. Are you going to create an image copy or a backup set? Read the docs and learn the difference. Are you going to turn on compression? Are you going to back up only changed blocks? We don't know.
    You ask: "how to manage growing online archiving files?"
    Again we can't help you without knowing your SLA. How many MG/GB are they? What period of time do you need to retain them to meet your SLA? When you figure this out back up to tape those you don't need to keep on the server.
    You ask: "how to manage growing data and there disk space?"
    This is one we can answer: BUY MORE DISK! That was easy.
    You ask: "How we can give good performance our CPU and memory?"
    Do you really expect that we can with zero knowledge of your environment, your version, your application, etc. distill into a few short paragraphs the collective wisdom of Cary Millsap, Jonathan Lewis, and the rest of the Oak Table Network membership?
    Of course we can not. So what you need to do is buy Cary's book, Jonathan's book, and take their classes.
    You ask: "we need keep all archive log in backup files or we need to remove old archive files?"
    Remove the old ones after being absolutely sure you have a good backup.
    You ask: "where we can take backup tape drive,SAN,disk removable hard disk? which one is better?"
    No one solution is better than the other. They are all adequate.

  • Comparision between Multimaster replication and data guard

    Hi,
    I have some questions regarding Multimaster Replication and Data Guard. Like
    *1.)* I have a web site having database in oracle 10.2.0 and multimaster replication configured on it. But whenever i need to change the structure of the tables i
    need to stop replication. So my web site is unavailable but my first priority is availability. So will it be useful for me to configure data guard.
    *2.)* I have configured data guard for testing . My Database is Named as Gard and physical standby database is like stan.
    it is working fine.
    suppose my main database server get corrupted i have no option to start main database server. Now i left with only standby server how i can start standby server
    as database main server.
    *3.)* Is there any if main database stop working Standby database start working as main without dba intervention.
    Thanks
    Umesh
    Edited by: Umesh Sharma on Jan 14, 2009 4:17 AM

    Hi Umesh,
    first of all you should be aware that Dataguard and Replication are two completely different things.
    While the database where you replicate to is always up and open in read write mode, a physical standby can either be recovering or open in read only mode.
    1.) I think from what you are telling you probably should consider using dataguard with a logical standby database, but be aware that there will be some limitations regarding datatypes.
    2.) you can do a failover or a switchover, depending on your database version switchback may not be possible
    3.) have a look at dataguard broker, you can use it to automate the failover
    Best regards,
    PP

  • How can i solve this replication problem between TT and ORACLE

    Hi
    I have an application that using AWT cashgroup implement the replication between TT (7.0.6.7) and ORACLE(10g);
    but i encounter this problem:
    16:16:50.01 Err : REP: 2302682: ABM_BAL_WH:meta.c(4588): TT5259: Failed to store Awt runtime information for datastore /abm_wh/abm_bal_ttdata/abm_bal_wh on Oracle.
    16:16:50.02 Err : REP: 2302682: ABM_BAL_WH:meta.c(4588): TT5107: TT5107: Oracle(OCI) error in OCIStmtExecute(): ORA-08177: can't serialize access for this transaction rc = -1 -- file "bdbStmt.c", lineno 3726, procedure "ttBDbStmtExecute()"
    16:16:50.02 Err : REP: 2302682: ABM_BAL_WH:receiver.c(5612): TT16187: Transaction 1316077016/357692526; Error: transient 0, permanent 1
    the isolation level of my date store is read-committed ,and the sys.ODBC.INI file is also set Isolation=1(readcommitted mode)
    so ,I still wonder how the error ORA-08177!
    how can i solve this replication problem?
    thank you.

    I suspect this is failing on an UPDATE to the tt_03_reppeers table on Oracle. I would guess the TT repagent has to temporarily use serializable isolation when updating this table. Do you have any other datastores with AWT cachegroups propagating into the same Oracle database? Or can you identify if some other process is preventing the repagent from using serializable isolation? If you google ORA-08177 there seem to be ways out there to narrow down what's causing the contention.

  • Load Balancing and Asynchronous Multimaster Replication

    We are planning a new project with 2 server lines for load balancing reasons. We also plan to use sessions being valid for 30 minutes. Each of the 2 server lines has its own Oracle database (identical tables) for the transactions. Within one session the load balancing sends the user always to the same server line.
    Independant transactions are executed on both server lines.
    Is Asynchronous Multimaster Replication the adequate solution to have the two databases synchronized, what has to be considered (network, memory, cpu) or what else can (has to) be used ?
    We want to avoid a RAC solution because of the high costs.

    The two options that come to mind here are asynchronous multimaster replication and Oracle Streams.
    - What version of the Oracle database will you be using?
    - What sort of transaction volume do you anticipate? How much redo generation?
    - What sort of connection exists between the two machines?
    - Is there an upper limit on the time that can be taken to replicate a transaction from one machine to the other?
    Justin
    Distributed Database Consulting, Inc.
    http://www.ddbcinc.com/askDDBC

  • Session in-memory replication problem

    Hi,
              I am running into some cluster HttpSession replication problems. Here is
              the scenario where replication fails (all servers mentioned here are a
              part of a cluster).
              1a - 2 Weblogic servers (A&B) are running - no users logged in,
              2a - user logs in and a new session in server A is created.
              3a - after several interactions, server A is killed.
              4a - after user makes susequent request, Weblogic correctly fails over
              to server B
              Problem: Not entire session data is replicated. The authentication info
              seems to
              be replicated correctly but there are some collections in the session of
              server A
              that did not make it to the session in server B.
              The interesting part is this: If there is only one server A running to
              begin with and a user
              interacts with it for a while and only then server B is started, when
              after server B starts up
              server A dies - the entire session (which is exactly the same as in the
              failing scenario) is
              corretly replicated in B, including collections that were missing in the
              failing scenario.
              How can this be possible ????
              Thanks for any info on this one - it really puzzles me.
              Andrew
              

    Yes, you are on the right track. Everytime you modify the object you should call
              putValue. We will make it more clear in the docs.
              - Prasad
              Andrzej Porebski wrote:
              > Everything is Serilizable. I get no exceptions. I did however read some old
              > posts regarding
              > session replication and I hope I found an answer. It basically seems to boil
              > down to what
              > triggers session sync-up between servers. In my case , I store an object into
              > session and
              > later on manipulate that object directly wihotu session involvment and the
              > results of those manipulations
              > are not replicated - no wonder if HttpSession's putValue method is the only
              > trigger.
              > Am i on the right track here?
              >
              > -Andrew
              >
              > Prasad Peddada wrote:
              >
              > > Do you have non serializable data by any chance?
              > >
              > > - Prasad
              > >
              > > Andrzej Porebski wrote:
              > >
              > > > Hi,
              > > > I am running into some cluster HttpSession replication problems. Here is
              > > > the scenario where replication fails (all servers mentioned here are a
              > > > part of a cluster).
              > > > 1a - 2 Weblogic servers (A&B) are running - no users logged in,
              > > > 2a - user logs in and a new session in server A is created.
              > > > 3a - after several interactions, server A is killed.
              > > > 4a - after user makes susequent request, Weblogic correctly fails over
              > > > to server B
              > > >
              > > > Problem: Not entire session data is replicated. The authentication info
              > > > seems to
              > > > be replicated correctly but there are some collections in the session of
              > > > server A
              > > > that did not make it to the session in server B.
              > > >
              > > > The interesting part is this: If there is only one server A running to
              > > > begin with and a user
              > > > interacts with it for a while and only then server B is started, when
              > > > after server B starts up
              > > > server A dies - the entire session (which is exactly the same as in the
              > > > failing scenario) is
              > > > corretly replicated in B, including collections that were missing in the
              > > > failing scenario.
              > > >
              > > > How can this be possible ????
              > > >
              > > > Thanks for any info on this one - it really puzzles me.
              > > >
              > > > Andrew
              > >
              > > --
              > > Cheers
              > >
              > > - Prasad
              >
              > --
              > -------------------------------------------------------------
              > Andrzej Porebski
              > Sailfish Systems, Ltd. Phone 1 + (212) 607-3061
              > 44 Wall Street, 17th floor Fax: 1 + (212) 607-3075
              > New York, NY 10005
              > -------------------------------------------------------------
              

  • Maximum number of nodes in Multimaster Replication

    Has anyone setup a multimaster replication environment on 8.1.7 with more than 11 nodes?
    What could be the maximum number of nodes for reliable operation?

    Hi,
    With Oracle 8.1 you can have upto 9; but I think you can have under one of the replicated DB another replicated DB. tree like tructure.
    Best,
    EA

  • Multimaster replication error

    Hi,
    Well I am a newbie to multimaster replication. I tried to replication few tables and indexes of hr schema to other database but I am receiving the following errors:
    ID REQUEST ERRNUM MESSAGE
    43 GENERATE_INTERNAL_PKG_SUPPORT -23308 ORA-23308: object HR.EMPLOYEES does not exist or is inv
    121 GENERATE_INTERNAL_PKG_SUPPORT -23308 ORA-23308: object HR.DEPARTMENTS does not exist
    41 GENERATE_INTERNAL_PKG_SUPPORT -23308 ORA-23308: object HR.DEPARTMENTS does not exist or is i
    141 GENERATE_INTERNAL_PKG_SUPPORT -23308 ORA-23308: object HR.DEPARTMENTS does not exist
    161 GENERATE_INTERNAL_PKG_SUPPORT -23308 ORA-23308: object HR.EMPLOYEES does not exist or
    21 CREATE_MASTER_REPOBJECT -942 ORA-00942: table or view does not exist
    22 CREATE_MASTER_REPOBJECT -942 ORA-00942: table or view does not exist
    123 GENERATE_INTERNAL_PKG_SUPPORT -23308 ORA-23308: object HR.EMPLOYEES does not exist or
    24 CREATE_MASTER_REPOBJECT -942 ORA-00942: table or view does not exist
    25 CREATE_MASTER_REPOBJECT -942 ORA-00942: table or view does not exist
    26 CREATE_MASTER_REPOBJECT -942 ORA-00942: table or view does not exist
    ID REQUEST ERRNUM MESSAGE
    27 CREATE_MASTER_REPOBJECT -942 ORA-00942: table or view does not exist
    28 CREATE_MASTER_REPOBJECT -942 ORA-00942: table or view does not exist
    29 CREATE_MASTER_REPOBJECT -942 ORA-00942: table or view does not exist
    30 CREATE_MASTER_REPOBJECT -942 ORA-00942: table or view does not exist
    31 CREATE_MASTER_REPOBJECT -942 ORA-00942: table or view does not exist
    47 GENERATE_INTERNAL_PKG_SUPPORT -23308 ORA-23308: object HR.JOB_HISTORY does not exist or is i
    127 GENERATE_INTERNAL_PKG_SUPPORT -23308 ORA-23308: object HR.JOB_HISTORY does not exist
    Kindly, help me regarding this issue.
    Thank you for your cooperation.
    Kind Regards,
    Adnan Hamdus Salam

    You will see "replica busy" errors from time to time. The replication protocol is "one at a time" - if a consumer has more than 1 supplier, only 1 supplier may update the consumer at a time - the other suppliers get a busy signal during the update. Note that updates to 1 supplier may be sent to another supplier and then to the consumer if the first supplier gets a busy signal, so all of your updates should get through pretty quickly.
    The most important thing is - do your updates get through to all consumers? Do they get through quickly?

  • In a MultiMaster Replication to not propagate changes on Second DB

    Hello,
    I've setup a multimaster replication environment with DB1 and DB2.
    DB1 replicates all changes on some tables to DB2. But i don't want to DB2 to propagate changes to DB1. So, i didn't create a scheduled link on DB2 to propagate changes to DB1. But deferred transaction queue is getting too large on DB2. How can i settle the DB2 to not to put the changes to deferred transactions?
    Regards

    Replicating from DB1 to DB2 and allowing local updates on DB2 without propagating them back to DB1 is done with WRITABLE MATERIALIZED VIEWS, not MultiMaster Replication.
    Changes made at DB2 are "local" until they are overwritten by changes from DB1.
    See http://download.oracle.com/docs/cd/B19306_01/server.102/b14226/repoverview.htm#sthref57
    Hemant K Chitale
    http://hemantoracledba.blogspot.com

  • DFS Replication Problem

    Hi Friends,
    I have windows server 2003 domain at two location before somw month back its replication data
    and its working fine but now i unable to see replicate data i mean i thing having replication problem
    i gotted some evint id error on server
    Event id error :- 5002  , 4202  , 1925 ,13568
    Please help me .
    Thanks,
    Madhukar

    The 4202 is staging quota size is too small.  
    Run these 2 Power Shell commands to determine the correct Staging Quota size:
    $big32 = Get-ChildItem DriveLetter:\FolderName -recurse | Sort-Object length -descending | select-object -first 32 | measure-object -property length –sum
    $big32.sum /1gb
    Take that resulting number, round it up to the nearest whole integer, mulitply that times 1024 and enter that number on the Staging tab of the Properties of a replicated folder in DFS Mgt.
    More info here:
    http://blogs.technet.com/b/askds/archive/2007/10/05/top-10-common-causes-of-slow-replication-with-dfsr.aspx
    Run this command to tell you the status of Replication:
    wmic /namespace:\\root\microsoftdfs path DfsrReplicatedFolderInfo get replicatedFolderName, State
    0: Uninitialized
    1: Initialized
    2: Initial Sync
    3: Auto Recovery
    4: Normal
    5: In Error
    Let us know how that goes.

  • Is schema replicated in multimaster replication scenario?

    hellos
    I am asking basic question. I have set up a multimaster replication (ids5.1sp2) on database c=fi.. this works well.
    however, in the future I know that more objectClasses and attributes will be added to the schema.
    My simple question.. is do I have to amend the 99user.ldif files on BOTH ids5.1 servers or can I rely on the schema being replicated.
    thanks.

    Schema is supposed to replicate in all replication scenario's
    The documentation states
    "The best way to guarantee schema consistency is to make schema modifications on a single master server, even in the case of a multi-master replication environment.
    Schema replication happens automatically. If replication has been configured between a supplier and a consumer, schema replication will happen by default."
    However a few people in this forum have complained that this replication doesn't always work - so be sure to check to make sure the schema does in fact replicate.
    For more information on schema replication check out this link:
    http://docs.sun.com/source/816-5609-10/rep.htm#1063947
    It is at the bottom of the page ...

  • I have replication problems.

    I have a multimaster replication environment
    A server A is used for R/W which has replication agreement with server B (kind of backup).
    Changes form Server B to server A are replicated without anyproblems.
    But those from server A to server B are not.
    Heres is the error message I have can anyone help me?
    [17/Oct/2003:17:45:26 +0200] NSMMReplicationPlugin - cl5CheckpointMain: checkpoint failed, db error - -30989 DBRUNRECOVERY: Fatal error, run database recovery
    [17/Oct/2003:17:45:27 +0200] NSMMReplicationPlugin - cl5CheckpointMain: checkpoint failed, db error - -30989 DBRUNRECOVERY: Fatal error, run database recovery
    [17/Oct/2003:17:45:27 +0200] NSMMReplicationPlugin - cl5CheckpointMain: checkpoint failed, db error - -30989 DBRUNRECOVERY: Fatal error, run database recovery
    [17/Oct/2003:17:45:28 +0200] NSMMReplicationPlugin - cl5CheckpointMain: checkpoint failed, db error - -30989 DBRUNRECOVERY: Fatal error, run database recovery
    [17/Oct/2003:17:45:28 +0200] NSMMReplicationPlugin - cl5CheckpointMain: checkpoint failed, db error - -30989 DBRUNRECOVERY: Fatal error, run database recovery
    Thanks

    Heres is my new error message on the server whoch cannot get replicated
    [20/Oct/2003:10:15:04 +0200] NSMMReplicationPlugin - repl_set_mtn_referrals: could not set referrals for replica o=customer: 20
    [20/Oct/2003:10:15:43 +0200] - add value to attribute type nsslapd-referral in entry cn=\22o=customer\22,cn=mapping tree,cn=config failed: duplicate value
    [20/Oct/2003:10:15:43 +0200] NSMMReplicationPlugin - repl_set_mtn_referrals: could not set referrals for replica o=customer: 20
    [20/Oct/2003:10:15:51 +0200] - add value to attribute type nsslapd-referral in entry cn=\22o=customer\22,cn=mapping tree,cn=config failed: duplicate value
    Thanks for your help

  • Multimaster replication survival & overhead

    Hi
    We are planning a system which should have a very high survivability rate. To this aim we consider using 3 oracle servers (8.1.7), located in 3 different sites, replicated in an asynchronous multimaster configuration, with a short refresh time. The nature of the application is that many small update transactions are happening constantly.
    Several questions:
    1. We are trying to assess the required bandwidth of the site-to-site network connection we will need to support this configuration. What is the approx. network overhead of a small (~500 bytes) transaction?
    2. Provided we come up with a smart enough conflict resolving mechanism, will this system survive? It is vital that committed transactions will never, ever, be lost.
    3. Can the 3 db architecture can be left in an inconsistent state? For example, can it be that a committed transaction in database A, is propagated to database B, and then A fails before it can be propagated to database C? leaving inconsistencies in working databases B & C (atleast until A is restored).
    4. How difficult would be to return such a system to normal state after one (or more) sites fails?
    Thanks,
    Idan

    <BLOCKQUOTE><font size="1" face="Verdana, Arial, Helvetica">quote:</font><HR>Originally posted by Patricia McElroy ([email protected]):
    This requires a bit more detail - what indicates that replication is hung... is there anything in any of the database logs that indicates a problem. You indicate that even changes to the local database can not be made - is this true also for non-replicated tables? Also, Run "Select * from dba_repprop;" and confirm that the method for propagation is indeed ASYNCHRONOUS. <HR></BLOCKQUOTE>
    thanks patricia
    here is the details:
    1. i am having 2 DB servers at remote locations running oracle 8.1.6 on solaris 2.7
    2. When thre is a link fails between these two server, i am not able to put the job in a deffered que transactions though i have scheduled job and link as well.
    3. it's not allowing me to commit insert statement on local server,when the link is down.
    null

  • In memory replication problems when I bring up a new server

              I've got in memory replication set up for 6.1. It works fine if I have 2 servers
              up and 1 goes down.
              However, if I have 1 server up and a bring a second server up, the sessions blow
              out.
              E.g. I've got server A and server B.
              Both are up, both have sessions. As new sessions come in, they are replicated over
              to the other server.
              now I bring server B down. All sessions on B fail over to A.
              so far so good.
              However when I bring server A back up some of the sessions fail as soon as the server
              is back up.
              Is this a configuration issue, is this a know problem?
              This worked fine in weblogic 5.1. In 5.1 when I brought an instance back up, everything
              worked fine.
              

              It turns out the problem was caused by using an old version of the Apache Plugin.
              This problem occurred while using the 5.1 apache plugin with WLS 6.1.
              Once we realized we were using the wrong plugin and swithced to the 6.1 plugin, the
              problem went away.
              

  • JNDI replication problems in WebLogic cluster.

    I need to implement a replicable property in the cluster: each server could
    update it and new value should be available for all cluster. I tried to bind
    this property to JNDI and got several problems:
    1) On each rebinding I got error messages:
    <Nov 12, 2001 8:30:08 PM PST> <Error> <Cluster> <Conflict start: You tried
    to bind an object under the name example.TestName in the jndi tree. The
    object you have bound java.util.Date from 10.1.8.114 is non clusterable and
    you have tried to bind more than once from two or more servers. Such objects
    can only deployed from one server.>
    <Nov 12, 2001 8:30:18 PM PST> <Error> <Cluster> <Conflict Resolved:
    example.TestName for the object java.util.Date from 10.1.9.250 under the
    bind name example.TestName in the jndi tree.>
    As I understand this is a designed behavior for non-RMI objects. Am I
    correct?
    2) Replication is still done, but I got randomly results: I bind object to
    server 1, get it from server 2 and they are not always the same even with
    delay between operation in several seconds (tested with 0-10 sec.) and while
    it lookup returns old version after 10 sec, second attempt without delay
    could return correct result.
    Any ideas how to ensure correct replication? I need lookup to return the
    object I bound on different sever.
    3) Even when lookup returns correct result, Admin Console in
    Server->Monitoring-> JNDI Tree shows an error for bound object:
    Exception
    javax.naming.NameNotFoundException: Unable to resolve example. Resolved: ''
    Unresolved:'example' ; remaining name ''
    My configuration: admin server + 3 managed servers in a cluster.
    JNDI bind and lookup is done from stateless session bean. Session is
    clusterable and deployed to all servers in cluster. Client invokes session
    methods throw t3 protocol directly on servers.
    Thank you for any help.

    It is not a good idea to use JNDI to replicate application data. Did you consider
    using JMS for this? Or JavaGroups (http://sourceforge.net/projects/javagroups/) -
    there is an example of distibuted hashtable in examples.
    Alex Rogozinsky <[email protected]> wrote:
    I need to implement a replicable property in the cluster: each server could
    update it and new value should be available for all cluster. I tried to bind
    this property to JNDI and got several problems:
    1) On each rebinding I got error messages:
    <Nov 12, 2001 8:30:08 PM PST> <Error> <Cluster> <Conflict start: You tried
    to bind an object under the name example.TestName in the jndi tree. The
    object you have bound java.util.Date from 10.1.8.114 is non clusterable and
    you have tried to bind more than once from two or more servers. Such objects
    can only deployed from one server.>
    <Nov 12, 2001 8:30:18 PM PST> <Error> <Cluster> <Conflict Resolved:
    example.TestName for the object java.util.Date from 10.1.9.250 under the
    bind name example.TestName in the jndi tree.>
    As I understand this is a designed behavior for non-RMI objects. Am I
    correct?
    2) Replication is still done, but I got randomly results: I bind object to
    server 1, get it from server 2 and they are not always the same even with
    delay between operation in several seconds (tested with 0-10 sec.) and while
    it lookup returns old version after 10 sec, second attempt without delay
    could return correct result.
    Any ideas how to ensure correct replication? I need lookup to return the
    object I bound on different sever.
    3) Even when lookup returns correct result, Admin Console in
    Server->Monitoring-> JNDI Tree shows an error for bound object:
    Exception
    javax.naming.NameNotFoundException: Unable to resolve example. Resolved: ''
    Unresolved:'example' ; remaining name ''
    My configuration: admin server + 3 managed servers in a cluster.
    JNDI bind and lookup is done from stateless session bean. Session is
    clusterable and deployed to all servers in cluster. Client invokes session
    methods throw t3 protocol directly on servers.
    Thank you for any help.--
    Dimitri

Maybe you are looking for

  • Please add a "link" or "share" button to replies / answers for easier referencing and linking.

    Currently, the only ways to link to a specific reply or answer are: Hunt through the poster's activity page and look for the link to the post in question, or Inspect the post's HTML, extract the ID from the enclosing <li> tag, then manually construct

  • Problem with Remove Instance

    Hi, I have PDF form 2-3 pages long with a Tabular Data View. It contains almost 50 records per per page. Now I am try to remove a particular Instance and its working fine I am uisng details._dataForm.removeInstance(this.parent.index). But the issue i

  • Subroutine program its display the text at line item as it's service order

    Hi All, Am looking for subroutine program its display the text at line item as it's service order, I want to view the text in sapscript in the tcode ME9F of the order text. Ex:- in tcode ME23N, select PO & item details in Account Assignment tab, ORDE

  • New to sap enhancements..

    Hi ,    I am new to sap user exits . Can any give the name of the book which gives full information about exits ? And can any one give interview questions on User exits ? and Real time Examples like 2 to 3 tasks on User Exits ? Please Urgent.

  • How to control the mesh tool?

    I am wanting to use the mesh tool but I need to know how to control my lines. I want a starting point and an end point but when I click the starting point is decided by me but the end point is decided on its own, and it is not the end point I want. I