Necessary configurations required to generate an IDOC

Hi Friends,
I want to know the necessary configuration that has to be done , inorder to generate an IDOC when a particular record is saved.
for example when I tried to change the QTY in Me22(PO), an IDOC, orders05 need tobe generated, which is not getting generated. Please let me know the configurations needed to generate an IDOC.
Thnaks in Advance,
Meghna.

Hi,
you need to call the outbound processing program for IDOC,
The system generates an IDoc of type ORDERS01, which is not yet transferred to the port (output mode Collect IDocs) for Tcode ME11.
Similarly you need to call the IDOC processing Function Module in User Exit of Tcode ME22
Refer
http://help.sap.com/saphelp_nw04/helpdata/en/dc/6b7c9f43d711d1893e0000e8323c4f/frameset.htm
http://help.sap.com/saphelp_nw04/helpdata/en/0b/2a60c8507d11d18ee90000e8366fc2/frameset.htm
http://help.sap.com/saphelp_nw04/helpdata/en/f2/660b378350c976e10000009b38f839/frameset.htm
/people/swaroopa.vishwanath/blog/2007/01/22/ale-configuration-for-pushing-idocs-from-sap-to-xi
/people/prateek.shah/blog/2005/06/08/introduction-to-idoc-xi-file-scenario-and-complete-walk-through-for-starters
idoc settings /people/ravikumar.allampallam/blog/2005/02/23/configuration-steps-required-for-posting-idocsxi
Also
/people/sravya.talanki2/blog/2005/10/27/idoc146s-not-reaching-xi133-not-posted-in-the-receiver-sap-systems133
IDOc testing - /people/suraj.sr/blog/2005/12/29/generate-test-case-for-an-idoc-scenario
Thanks
Swarup

Similar Messages

  • Requirement to generate new IDOC'S

    Hi All,
    I have a flat file with customer legacy data .
    Using the data from the file, I  need to generate IDOCS and populate the respective database tables.
    Please could someone guide me as to how we could do this.
    Regards,
    Sushanth Hulkod

    Hi,
    1.First append the file data into an internal table Using GUI_UPLOAD.
    2. Find the appropriate idoc type  and  fill the idoc segments with the internal table data .
    3. pass the control data like legacy system details sender, partner profile, receiver to control idoc .
    4. Finally call the Function module 'MASTER_IDOC_DISTRIBUTE '.
    Edited by: Archana.T on Dec 15, 2009 8:19 AM

  • Generate outbound idoc for Customer Master

    Hi Developers,
    I have requirement to generate outbound IDOC for Customer Master Data. I need to send customer master data through an Outbound IDOC. Can we generate an outbound DEBMAS06 idoc? Can someone help me finding a solution to this problem?
    Thanks in advance for helping.
    Regards,
    Gajendra

    Hi,
    If you have done all the required settings (logical system, partner profiles, distribution model), then either transaction BD10 / BD14 will help you generate the IDOC.
    If you want to send the idocs automatically, then you will to configure some program.
    Unlike transaction data, Master data IDOCS are not triggered immediately, you will have to do the required configuration in SPRO.
    Transaction SPRO
    SAP Customizing Implementation Guide -->
    SAP Web application Server --> Application Link Enabling --> Modeling and Implementing Business Processes -->
    Master Data distribution --> Replication of Modified Data --> Creating IDOCS from Change Pointers
    Take a look at this thread.
    Re: automatic Idoc outbound
    Regards,
    Ravi
    Note : Please mark the helpful answers
    Message was edited by: Ravikumar Allampallam

  • Configurations required in SLD,IR,ID,SM59,BD64 to Start with IDOCS

    Hello all,
    Can You Post me the Above details.
    thanks,
    Srinivasa

    Hi Srinivas,
    Let us explore the need for each steps and detail it now...
    <b>Step-1 :</b> You need RFC destinations for placing the metadata of the IDoc.
    Use transaction SM59 to maintain an RFC destination for the IDoc sender/receiver system.
    This RFC destination is used to retrieve the IDoc metadata from the sender system.
    The IDoc adapter needs these metadata to create the corresponding IDoc-XML message from the RFC stream.
    <b>Step-2 :</b> Next You need IDx1 for getting IDoc metadata from the Sender system.
    Use transaction IDX1 to assign a port (RFC destination) to the system that contains the metadata of the IDoc types.
    <b>Note:</b>
    There is a mechanism on the Integration Server that uses this RFC destination to retrieve and cache the metadata at runtime if it is not yet available on the Integration Server.
    <b>Nut-Shell:</b>
    Since IDoc metadata is cross-client, you should only assign one port for each system.
    If several ports are assigned, ensure that they are all working.
    This is Done in Partner Profile.
    If you are not able to load the metadata from this system (because of administrative or security-related restrictions, for example),
    you can also load it from a reference system (for example, a test system) and use transaction IDX2 to assign it to your production system afterwards.
    For non-SAP systems, either the ports in transaction IDX1 have to point to a reference SAP system, or you copy the metadata with transaction IDX2.
    I come to mean that if you double click the OutBound Parameter in the Partner Profile you will be allowed to choose the Idoc type like Matmas,Debmas blabla..
    There after providing the receiver destination in that partner profile which is nothing but your Maintaining Metadata through a port,
    which on double click leads to RFC destination alloted with a port to transfer its data from sender to receiver Server.
    <b>Step-3 :</b> Goto SLD and do the following.
    Access the System Landscape Directory (SLD) to maintain the technical systems for the sender and receiver business systems of your system landscape.
    You have to define a technical system (and a client in the case of SAP systems) to which your business system belongs.
    When you define a client number, you also have to specify the corresponding logical system name in the Technical System Browser.
    <b>Note:</b>
    The technical system configuration is not required if your business system is configured as a data supplier for the SLD.
    <b>Step-4 :</b> Goto Integration Directory and do the folllowing.
    Access the Integration Directory to define your business systems as services without party.
    a) Either you can choose from the Objects Business Service where your client is defined for it.
       With this Selected Business Service you just rightclick and choose ASSIGN TO BUSINESS SCENARIO.
    b) Second choice is to choose the business system which was done in SLD.
    Do the receiver determination and interface determination but since the adapter is  not required to choose for Idoc dont do sender determination.
    <b>Note:</b>
    The IDoc adapter only uses the service definition (business system) together with the corresponding adapter-specific identifiers in the Integration Directory.
    The maintenance of your IDoc sender system in the SLD is therefore not sufficient.
    It is recommended that you assign your business system definition retrieved from the SLD to a service in the Integration Directory.
    <b>Nut-Shell:</b>
    Using this information about the system ID, client, and logical system name for a specific business system, the IDoc adapter is able to specify the corresponding service in the XML header. Routing is then based on service names.
    You need the
    &#61601;  logical system, SAP system ID, and client for an IDoc receiver SAP system
    &#61601;  SAP system ID and client for an SAP system
    &#61601;  logical system for a non-SAP system.
    This means that the business systems used for the routing definitions in the IDoc-XML message header are retrieved from the adapter-specific identifiers of the service definitions in the Integration Directory,
    where for each business system client, the corresponding system ID, client, and logical system name is defined.
    <b>Caution:</b>
    Use the import function to retrieve the adapter-specific identifiers for a service (business system) to avoid double maintenance in the SLD and Integration Directory.
    <b>Step-5 :</b> Steps in Integration Repository.
    Export the IDoc in Imported Objects.
    Double click the Idoc to view the structure.
    This Idoc type is decided in the partner Profile.
    See step-3 for further details.
    Create the Interface Objects like
    1.Data Type for Receiver File.
    2.Message type which wrapps this Data type.
    3.Interface Type.
      a) For Sender Idoc just drag the Idoc name which you have loaDED IT IN YOUR imported objects.
      b) For Receiver File It just wrapps the Message type of the file.
    4.Message Mapping:
      The Fields selected in the data type for the file should be the mandatory fields in the Idoc.
      Since the Idoc is in the sender side,it just performs similar operation like selecting things from  a database.
      Begin and Other Attributes are neednt be worried to map.
      The Mandatory fields which was chosen will need to be mapped with the target file.
    5.Interface Map:
      Just Wrapps the Message mapping.
    <b>Step-6 :</b> BD64 is to Distribute the Idoc  to Various Acceptors.
    1.Create Model View.
    2.Add Message Type to it.
    3.So that it displays your model view with your Message type like MATMAS.
    4.Goto Environment and click Generate Partner Profiles.
    5.Executing it,you will get a log indicating the Partner,port and Outbound Parameters.
    6.After Doing this goto Edit Menu available in the tool Bar and give Distribute.
    7.After distributing it ,it shows your logical system name along with the technical name.
    8.Now to fill the IDOCs with Data we have to go to Transaction bd10.
    9.We can give our own selection criteria’s and select the list of materials accordingly.
    10.Now to check whether the IDOCs has been correctly created and sent to the receiving system (in this case it is the SAP XI Server)
    11.we have to check it out in Transaction WE05.
    12.Now we can go to the Location we specified in the Configuration where the file will be created.

  • Flat File to IDOC Mapping requirement to generate Multiple Segments

    Hi Experts,
    I got a requirement were i have 2 records in a file and i need to generate 2 IDOCs  with  multiple segments in it.
    FILE :
    10/01/2010     101  KRNA     ic_quantity          30-0257     3526     1     1     ea     110000     10
    10/01/2010     101     KRNA     ic_quantity          90-0005     3526     1     2     ea     110000     10
    Idoc should generate 2 IDOCs with multiple segments as shown below
    I have imported the IDOC and changed the occurrence to " unbounded "
    The Basic  IDOC Type :  WMMBID02
    I need to generate Multiple segments of  E1MBXY1
    i.e..,  First IDOC should contain two  E1MBXY1 segments
             Second IDOC should contain Four  E1MBXY1 segments
    IDOC1 :   WMMBID02
    Segment :   E1MBXY1( 2 segments)                                                       
    10/01/2010     101     KRNA     ic_quantity          30-0257     3526     1     1     ea     110000     10
    10/01/2010     101     KRNA     ic_quantity          90-0005     3526     1     2     ea     110000     10
    IDOC2 : WMMBID02
    Segment :  E1MBXY1 ( 4 segments)                                                  
    10/01/2010     101     KRNA     ic_quantity          30-0257     3526     1     1     ea     110000     10
    10/01/2010     101     KRNA     ic_quantity          30-0257     3521     1     1     ea     110000     10
    10/01/2010     101     KRNA     ic_quantity          90-0005     3526     1     2     ea     110000     10
    10/01/2010     101     KRNA      ic_quantity          90-0005     3521     1     2     ea     110000     10
    Can anyone suggest me how to generate IDOCs with multiple segments
    what are multiple ways of generating it
    Whether it can be achieved using Multi-mapping or I need go for UDF
    If any one has done has done this type of requirement ,please share the points.
    Thanks
    Sai

    Basically you need to generate idoc per record in the flat file. During fcc conversion you convert flat file to xml structure at the sender side. In the mapping use xml file structure source and idoc as receiver structure. You just export idoc and update the idoc segment 1 to unbounded.  Please follow the michael blog for file to idoc multimapping without bpm. Yes without bpm it is possible.
    see this link... This will answer your requirement.
    https://wiki.sdn.sap.com/wiki/display/XI/File%20to%20Multiple%20IDOC%20Splitting%20without%20BPM
    >Whether it can be achieved using Multi-mapping or I need go for UDF
    you dont need udf for this.

  • Configurations required for IDOC to File

    Can any body suggest me the configuration required for integrating remote R/3 system with the XI Server.
    I want to know the following.
    1) How and where to configure R/3 with the XI server.
    2) What are the configurations required for Outbound and Inbound IDocs in R/3 and XI Servers.

    Hi Rajeshwar,
    take a look over here:
    https://www.sdn.sap.com/sdn/weblogs.sdn?blog=/weblogs/topic/16
    you'll find at least 5 weblogs that show the idoc configuration
    Regards,
    michal
    XI FAQ - http://www.bcc.com.pl/netweaver/sapxifaq

  • CONCURRENT MANAGER SETUP AND CONFIGURATION REQUIREMENTS IN AN 11I RAC ENVIR

    제품 : AOL
    작성날짜 : 2004-05-13
    PURPOSE
    RAC-PCP 구성에 대한 Setup 사항을 기술한 문서입니다.
    PCP 구현은 CM의 workload 분산, Failover등을 목적으로 합니다.
    Explanation
    Failure sceniro 는 다음 3가지로 구분해 볼수 있습니다.
    1. The database instance that supports the CP, Applications, and Middle-Tier
    processes such as Forms, or iAS can fail.
    2. The Database node server that supports the CP, Applications, and Middle-
    Tier processes such as Forms, or iAS can fail.
    3. The Applications/Middle-Tier server that supports the CP (and Applications)
    base can fail.
    아래부분은 CM,AP 구성과
    CM과 GSM(Global Service Management)과의 관계를 설명하고 있습니다.
    The concurrent processing tier can reside on either the Applications, Middle-
    Tier, or Database Tier nodes. In a single tier configuration, non PCP
    environment, a node failure will impact Concurrent Processing operations do to
    any of these failure conditions. In a multi-node configuration the impact of
    any these types of failures will be dependent upon what type of failure is
    experienced, and how concurrent processing is distributed among the nodes in
    the configuration. Parallel Concurrent Processing provides seamless failover
    for a Concurrent Processing environment in the event that any of these types of
    failures takes place.
    In an Applications environment where the database tier utilizes Listener (
    server) load balancing is implemented, and in a non-load balanced environment,
    there are changes that must be made to the default configuration generated by
    Autoconfig so that CP initialization, processing, and PCP functionality are
    initiated properly on their respective/assigned nodes. These changes are
    described in the next section - Concurrent Manager Setup and Configuration
    Requirements in an 11i RAC Environment.
    The current Concurrent Processing architecture with Global Service Management
    consists of the following processes and communication model, where each process
    is responsible for performing a specific set of routines and communicating with
    parent and dependent processes.
    아래 내용은 PCP환경에서 ICM, FNDSM, IM, Standard Manager의 역활을 설명하고
    있습니다.
    Internal Concurrent Manager (FNDLIBR process) - Communicates with the Service
    Manager.
    The Internal Concurrent Manager (ICM) starts, sets the number of active
    processes, monitors, and terminates all other concurrent processes through
    requests made to the Service Manager, including restarting any failed processes.
    The ICM also starts and stops, and restarts the Service Manager for each node.
    The ICM will perform process migration during an instance or node failure.
    The ICM will be
    active on a single node. This is also true in a PCP environment, where the ICM
    will be active on at least one node at all times.
    Service Manager (FNDSM process) - Communicates with the Internal Concurrent
    Manager, Concurrent Manager, and non-Manager Service processes.
    The Service Manager (SM) spawns, and terminates manager and service processes (
    these could be Forms, or Apache Listeners, Metrics or Reports Server, and any
    other process controlled through Generic Service Management). When the ICM
    terminates the SM that
    resides on the same node with the ICM will also terminate. The SM is ?hained?
    to the ICM. The SM will only reinitialize after termination when there is a
    function it needs to perform (start, or stop a process), so there may be
    periods of time when the SM is not active, and this would be normal. All
    processes initialized by the SM
    inherit the same environment as the SM. The SM environment is set by APPSORA.
    env file, and the gsmstart.sh script. The TWO_TASK used by the SM to connect
    to a RAC instance must match the instance_name from GV$INSTANCE. The apps_<sid>
    listener must be active on each CP node to support the SM connection to the
    local instance. There
    should be a Service Manager active on each node where a Concurrent or non-
    Manager service process will reside.
    Internal Monitor (FNDIMON process) - Communicates with the Internal Concurrent
    Manager.
    The Internal Monitor (IM) monitors the Internal Concurrent Manager, and
    restarts any failed ICM on the local node. During a node failure in a PCP
    environment the IM will restart the ICM on a surviving node (multiple ICM's may
    be started on multiple nodes, but only the first ICM started will eventually
    remain active, all others will gracefully terminate). There should be an
    Internal Monitor defined on each node
    where the ICM may migrate.
    Standard Manager (FNDLIBR process) - Communicates with the Service Manager and
    any client application process.
    The Standard Manager is a worker process, that initiates, and executes client
    requests on behalf of Applications batch, and OLTP clients.
    Transaction Manager - Communicates with the Service Manager, and any user
    process initiated on behalf of a Forms, or Standard Manager request. See Note:
    240818.1 regarding Transaction Manager communication and setup requirements for
    RAC.
    Concurrent Manager Setup and Configuration Requirements in an 11i RAC
    Environment
    PCP를 사용하기위한 기본적인 Setup 절차를 설명하고 있습니다.
    In order to set up Setup Parallel Concurrent Processing Using AutoConfig with
    GSM,
    follow the instructions in the 11.5.8 Oracle Applications System Administrators
    Guide
    under Implementing Parallel Concurrent Processing using the following steps:
    1. Applications 11.5.8 and higher is configured to use GSM. Verify the
    configuration on each node (see WebIV Note:165041.1).
    2. On each cluster node edit the Applications Context file (<SID>.xml), that
    resides in APPL_TOP/admin, to set the variable <APPLDCP oa_var="s_appldcp">
    ON </APPLDCP>. It is normally set to OFF. This change should be performed
    using the Context Editor.
    3. Prior to regenerating the configuration, copy the existing tnsnames.ora,
    listener.ora and sqlnet.ora files, where they exist, under the 8.0.6 and iAS
    ORACLE_HOME locations on the each node to preserve the files (i.e./<some_
    directory>/<SID>ora/$ORACLE_HOME/network/admin/<SID>/tnsnames.ora). If any of
    the Applications startup scripts that reside in COMMON_TOP/admin/scripts/<SID>
    have been modified also copy these to preserve the files.
    4. Regenerate the configuration by running adautocfg.sh on each cluster node as
    outlined in Note:165195.1.
    5. After regenerating the configuration merge any changes back into the
    tnsnames.ora, listener.ora and sqlnet.ora files in the network directories,
    and the startup scripts in the COMMON_TOP/admin/scripts/<SID> directory.
    Each nodes tnsnames.ora file must contain the aliases that exist on all
    other nodes in the cluster. When merging tnsnames.ora files ensure that each
    node contains all other nodes tnsnames.ora entries. This includes tns
    entries for any Applications tier nodes where a concurrent request could be
    initiated, or request output to be viewed.
    6. In the tnsnames.ora file of each Concurrent Processing node ensure that
    there is an alias that matches the instance name from GV$INSTANCE of each
    Oracle instance on each RAC node in the cluster. This is required in order
    for the SM to establish connectivity to the local node during startup. The
    entry for the local node will be the entry that is used for the TWO_TASK in
    APPSORA.env (also in the APPS<SID>_<HOSTNAME>.env file referenced in the
    Applications Listener [APPS_<SID>] listener.ora file entry "envs='MYAPPSORA=<
    some directory>/APPS<SID>_<HOSTNAME>.env)
    on each node in the cluster (this is modified in step 12).
    7. Verify that the FNDSM_<SID> entry has been added to the listener.ora file
    under the 8.0.6 ORACLE_HOME/network/admin/<SID> directory. See WebiV Note:
    165041.1 for instructions regarding configuring this entry. NOTE: With the
    implementation of GSM the 8.0.6 Applications, and 9.2.0 Database listeners
    must be active on all PCP nodes in the cluster during normal operations.
    8. AutoConfig will update the database profiles and reset them for the node
    from which it was last run. If necessary reset the database profiles back to
    their original settings.
    9. Ensure that the Applications Listener is active on each node in the cluster
    where Concurrent, or Service processes will execute. On each node start the
    database and Forms Server processes as required by the configuration that
    has been implemented.
    10. Navigate to Install > Nodes and ensure that each node is registered. Use
    the node name as it appears when executing a nodename?from the Unix prompt on
    the server. GSM will add the appropriate services for each node at startup.
    11. Navigate to Concurrent > Manager > Define, and set up the primary and
    secondary node names for all the concurrent managers according to the
    desired configuration for each node workload. The Internal Concurrent
    Manager should be defined on the primary PCP node only. When defining the
    Internal Monitor for the secondary (target) node(s), make the primary node (
    local node) assignment, and assign a secondary node designation to the
    Internal Monitor, also assign a standard work shift with one process.
    12. Prior to starting the Manager processes it is necessary to edit the APPSORA.
    env file on each node in order to specify a TWO_TASK entry that contains
    the INSTANCE_NAME parameter for the local nodes Oracle instance, in order
    to bind each Manager to the local instance. This should be done regardless
    of whether Listener load balancing is configured, as it will ensure the
    configuration conforms to the required standards of having the TWO_TASK set
    to the instance name of each node as specified in GV$INSTANCE. Start the
    Concurrent Processes on their primary node(s). This is the environment
    that the Service Manager passes on to each process that it initializes on
    behalf of the Internal Concurrent Manager. Also make the same update to
    the file referenced by the Applications Listener APPS_<SID> in the
    listener.ora entry "envs='MYAPPSORA= <some directory>/APPS<SID>_<HOSTNAME>.
    env" on each node.
    13. Navigate to Concurrent > Manager > Administer and verify that the Service
    Manager and Internal Monitor are activated on the secondary node, and any
    other addititional nodes in the cluster. The Internal Monitor should not be
    active on the primary cluster node.
    14. Stop and restart the Concurrent Manager processes on their primary node(s),
    and verify that the managers are starting on their appropriate nodes. On
    the target (secondary) node in addition to any defined managers you will
    see an FNDSM process (the Service Manager), along with the FNDIMON process (
    Internal Monitor).
    Reference Documents
    Note 241370.1

    What is your database version? OS?
    We are using VCP suite for Planning Purpose. We are using VCP environment (12.1.3) in Decentralized structure connecting to 3 differect source environment ( consisting 11i and R12). As per the Oracle Note {RAC Configuration Setup For Running MRP Planning, APS Planning, and Data Collection Processes [ID 279156]} we have implemented RAC in our test environment to get better performance.
    But after doing all the setups and concurrent programs assignment to different nodes, we are seeing huge performance issue. The Complete Collection which takes generally on an avg 180 mins in Production, is taking more than 6 hours to complete in RAC.
    So I would like to get suggestion from this forum, if anyone has implemented RAC in pure VCP (decentralized) environment ? Will there be any improvement if we make our VCP Instance in RAC ?Do you PCP enabled? Can you reproduce the issue when you stop the CM?
    Have you reviewed these docs?
    Value Chain Planning - VCP - Implementation Notes & White Papers [ID 280052.1]
    Concurrent Processing - How To Ensure Load Balancing Of Concurrent Manager Processes In PCP-RAC Configuration [ID 762024.1]
    How to Setup and Run Data Collections [ID 145419.1]
    12.x - Latest Patches and Installation Requirements for Value Chain Planning (aka APS Advanced Planning & Scheduling) [ID 746824.1]
    APSCHECK.sql Provides Information Needed for Diagnosing VCP and GOP Applications Issues [ID 246150.1]
    Thanks,
    Hussein

  • Problem in generating an IDOC from Output type

    Hi Experts,
    I am trying to generate an IDOC of basic type "OILSHI01" using the custom output type. in my scenario the output type gets issued , but IDOC does not get generated , i have checked the configuration in NACE transaction .
    Can you please help me understand what can be other reason why this IDOC does not get generated?
    Regards,
    AshiSh Shah

    Hello,
    Did you try to debug the idoc generation program & where it fails to put together the internal data into IDoc?
    That should help you identify problem.
    If  configuration settings is correct ,port settings are correct & partner profiles are correctly mapped for necessary Idoc creation,it should work.
    Hope this helps.
    Thanks.

  • Generating an IDOC with all segments filled every time

    IDOC Experts,
    I am new to IDOC/ALE and I am struggling with the following situation. 
    I am trying to create an ALE interface with a Transportation Management System.  I have been able to configure ALE and IDOC to generate an IDOC XML file for message type DEBMAS.  I have activated change pointers and I am seeing the changed customers when I run RBDMIDOC. 
    The problem is the external system will require the complete data every time the customer master is changed.  Right now I am only seeing the changed segments in the IDOC file. 
    Is it possible to configure SAP to populate all of the segments every time a DEBMAS IDOC is generated?
    Thank you in advance for any direction you can provide.
    Brian Snyder

    Hi Brian,
    I tested my system and it does not work,I guess we can't set up mandatory.
    I think if you know any user exit for that application,then fill the segments programtically.
    Sorry for wrong information what i  provided.
    Thanks
    Seshu

  • Creating an Invoice and generating and IDoc

    Hi All
    Can anyone let me know what are the prerequisites for creating an invoice, the transaction for creating an invoice and steps to create an Invoice using the transaction. I also want to know how to generate an Idoc from the Invoice related transaction screen.
    thanks in advance

    I mentioned very briefly about how to create outbound INVOIC. "Reward point if useful".
    Complete configuration information.
    1) In the Command field, enter transaction V/54 and choose Enter (or in
    the IMG, choose Sales and Distribution>Basic Functions>Output
    Control>Output Determination>Output Determination Using the
    Condition Technique-->Maintain Output Determination for Billing
    Documents-->Maintain access sequences).
    2) On the Change View “Access Sequence (Output Billing)”.
    3) Choose the desired access sequence (for example, YPCC, Sales
    organization).
    --> Choose Accesses.
    4) On the Change View “Accesses (Output Billing)”.
      a)Add the Sales Organization/Customer Number access above the
        existing Sales Org. access.
      b)In the Requirement column, enter a requirement.
      c) Select the Exclusive checkbox for the first access.
      d) Choose Save.
    5) EDI Specific configuration.
    Configuring for EDI involves the following two steps:
    -->Condition record (output determination)
    -->Outbound partner profile
    Condition record(Output determination)
    This section outlines the configuration steps required for each
    customer you wish to send an invoice via EDI.
    Creating a condition record for the specific customer sets up output
    determination to generate an EDI message whenever an invoice is
    created for that customer.
    1)-->To access the Create Output – Condition Records screen, in the Command
    field, enter transaction VV31 and choose Enter (or choose Logistics -->
    Sales and distribution --> Master data --> Output --> Billing document -->
    Create).
    2)--> On the Create Output – Condition Records: Billing screen:
         a) Enter RD00 in Output type.
         b)Choose Key combination.
    3) On the Key Combination popup window:
         a) Select Sales Organization/Customer Number.
         b) Choose Enter.
    4) On the Create Condition Records (Invoice): Fast Entry screen:
         a)Enter the customer number in Customer.
         b)Enter RE in PartF.
         c)Enter 6 in Medium.
         d)Enter 4 in Time.
         e)Enter EN in Language.
         f)Choose Save.
    5)Maintain partner profile.
    The setup of a partner profile for each specific EDI customer allows
    you to define the EDI-specific information the system needs when
    generating the outbound IDocs.
    Go to transaction WE20.
    a)Select the customer (partner) from the list.
    b)Choose outbound parameter.
    c) MG On the New Entries: Details of Added Entries screen:
    6)Enter RE in Partn.funct.
         a) Enter INVOIC in Message type.
         b)Enter the port name in Receiver port.
         c) Under Output mode, select Transfer IDoc immed.
         d) Enter INVOIC02 in Basic type.
         e)Enter US in Typ.
         f) Enter EN in Lang.
         h) Enter the user ID in ID.
    On the Change View “Partner profiles: message Control”: Overview
    screen, choose New entries.
    -->On the New Entries: Details of Added Entries screen:
         a)Enter V3 in Application.
         b) Enter RD00 in Message type.
         c) Enter SD09 in Process code.
         d)Choose Save.
    -->Choose the EDI standard tab.
    Thanks,
    Narayan

  • How to create an extension for generated type idoc

    Hi all,
    Can u please look into this
    How to create an extension for generated type idoc ?  for example Idoc type BATMAS02

    Enter transaction WE30 (ALE->Extension-> IDOC types->Maintain Idoc type)
    - Type in your name of the extended IDOC type (usually starting with 'Z') and click on the Basic IDoc type, click the create icon.
    - Click on Create new and enter a description and press enter.
    - Click on ZIDOCTYPE01 and then on the Create icon.
    - Enter ZIDOCTYPE as the segment type, click on Segment Editor.
    - Enter a description for your segment type and create.
    - Enter a description for your segment, enter each field required in your IDoc and press enter to validate.
    - Save and generate, press back
    - To release the segment choose Goto, Release from the menu.
    - Check the box on the line of your segment.
    - Save, back and enter.
    - Your Idoc type structure should be displayed with your new segment.
    - Save and back.
    - To release the Idoc type choose Extras, Release type from the menu and Yes.
    ALE FUNCTION MODULE ENHANCEMENTS
    Having extended the IDOC type to contain additional fields for an inbound or outbound application, you now want to enhance ALE function modules for populating the additional segment on the outbound or applying the additional segment data on the inbound application.
    The core working code for ALE processes for a given application area is always encapsulated in ABAP/4 function modules. These function modules are associated with such control information as message types and process codes. So the ALE process checks this control information and derives the name of the function module to invoke for that particular IDOC processing from certain database tables. These function modules contain objects known as customer functions, which can be considered SAP Enhanced user exits. A function module is called at a particular point during the processing of the main program or function module, and it can be used to influence data processing at that point by adding code to the customer function. The customer function behaves like a normal function module and has import and export parameters, tables (internal tables) statement, and exception processing. Unlike a conventional user exit, customer functions give you the ability to modify only data available to you by the function moduleâs parameters and internal tables. While most ALE/EDI function modules are supported by customer functions, there are ALE/EDI processes that still use conventional user exits. There are a few ways to determine which function module to enhance for a given message type/process code:
    • For master data distribution, from SALE go to Extensions -> Master data distribution -> Setup additional data for message types. Search for message type DEBMAS in this example. You see an entry for DEBMAS associated with function module MASTERIDOC_CREATE_SMD_DEBMAS. This data is stored on table TBDME. The function module names for all master data message types follow this pattern: MASTERIDOC_CREATE_SMD_messagetype. This function module calls another function module of name MASTERIDOC_CREATE_DEBMAS or MASTERIDOC_CREATE_messagetype. Search for the words customer function, and you find several hits that can be used to add code to the function module.
    • From WEDI got to Control -> Inbound process codes -> Inbound with ALE service -> Processing by function module (transaction WE42), or from WEDI go to Control -> Outbound process codes -> Outbound with ALE service -> With function module (transaction WE41). There will be function modules associated with the process codes. For inbound, the function modules usually follow this pattern: IDOC_INPUT_messagetype: for example, IDOC_INPUT_CHRMAS for inbound characteristics master.
    • Use transaction WE57 or from WEDI go to Development -> Message/Application Object. The entries list the function module, Business Object, message type, and IDOC type that are used for inbound ALE/EDI interfaces.
    Customer functions are not specific only to ALE and EDI but also to all programs/modules in SAP R/3. Customer function is a SAP enhancement component; the other two types are menu and screen enhancements.
    All customer function exits are maintained in SAP enhancements and are found by using transaction SMOD. After executing transaction SMOD, pull down (F4) on the enhancement name field, and execute again. This provides you with a list of all SAP enhancements available. SAP enhancements are grouped by development class pertaining to an application area. Choose Application development R/3 SD master data distribution for development class VSV to lead to a screen that lists VSV00001 as an enhancement (see Figure 5). Press Component +/- to display its function exit components. There are four possible components listed, all of which are function exits (and are function modules) that are called from the ALE function modules in the form Call Customer Function Î001â. This is a special occurrence of the ABAP statement Call. Go to item Exit_SAPLVV01_ 001, which you need to enhance for the Customer Master outbound example of an IDOC extension. In the ALE-function module MASTERIDOC_CREATE_DEBMAS, the statement CALL Customer Function 001 is translated in the background to call component EXIT_SAPLVV01_001. Although this function exit can be edited using transaction SE37, you will use a simpler approach.
    When you use SAP enhancements and their components, you manage them with an SAP object known as a project, which is like an envelope containing the selected enhancements and their components. A project can be used to control the execution of components and to transport them to other clients and instances in SAP. Basically, the process involves creating a project, including enhancements and components that are to be enhanced, editing the components, and then activating the project. The following process creates a project for our example Customer Master IDOC extension:
    • Execute transaction CMOD.
    • Enter name of project, say CSTMAST1.
    • Click on Create.
    • Enter a description of the project.
    • Save.
    • Click on SAP Enhancements.
    • Enter VSV00001 for Enhancement.
    • Save.
    Once youâve created the project, edit the function exit components and activate the project. Remember that the code in the function exit enhancement will execute only if the project is activated. In fact, this is a convenient SAP enhancements feature, whereby the work in progress (developing code in the customer function) will not affect users of that application. When the code is completed, the project can be activated so the enhanced functionality takes effect. It can also be deactivated for maintenance.
    As mentioned earlier, customer functions (function exits) are embedded in ALE function modules and can be used to influence the creation and modification of IDOC data on an outbound application or to post additional or modified IDOC data to an inbound R/3 application. Function exits are similar to regular function modules, with import/export parameters, tables (internal tables), and exceptions.
    The two important factors to consider while developing the customer function are:
    1. The point in the ALE function module where the function exit occurs
    2. The data made available by the customer function that can be modified or posted to the R/3 application, based on the direction.
    Because some function modules have several customer functions, it is critical to choose the function exit best suited for that particular enhancement. Do not attempt to perform activities that the function exit is not designed for. The importance of this point is illustrated by the following description of enhancing function modules for outbound and inbound ALE interfaces.
    Outbound interfaces. In an outbound ALE interface you use function exits (customer functions) to populate additional segments created by an IDOC extension or to modify the existing IDOC data segments as per business requirements. Previously, you identified that enhancement VSV00001 has a component EXIT_SAPLVV01_001 (function exit), which can be used for populating the additional data segment Z1SADRX that you created in the IDOC extension ZDEBMASX (IDOC type ZDEBMASZ, based on Basic IDOC type DEBMAS02). You also learned that the ALE function module that calls this function exit is MASTERIDOC_CREATE_DEBMAS, which has a statement Call Customer Function 001.
    Browse the function module MASTERIDOC_CREATE_DEBMAS using transaction SE37. You will find that this customer function is invoked for every segment of IDOC type DEBMAS02. In fact, the function exit is called soon after the creation of an existing segment has been populated with data and appended to the IDOC data table (internal table). Also, the function exit is exporting the message type, IDOC type, and the segment name and is importing the IDOC extension type. It is also passing the IDOC data internal table. This indicates that the ALE function module is allowing you to populate additional segments for every existing segment and modify the existing segmentâs data.
    Letâs write ABAP/4 code to accomplish the task of populating IDOC segment Z1SADRX with a contact personâs business address:
    • From SE37, display function module MASTERIDOC_CREATE_ DEBMAS.
    • Find Customer Function 001.
    • Double-click on 001.
    • The function EXIT_SAPLVV01_001 will be displayed.
    • Double-click on INCLUDE ZXVSVU01.
    • You will be asked to create a new include object. Proceed as desired.
    • Enter code (as in Listing 1).
    • Be sure to perform a main program check (Function Module -> Check -> main program) and extended program check (Function module -> Check -> Extended check).
    Now that you have extended the IDOC and enhanced the ALE function module based on the requirements for the contact personâs business address on the Customer Master, letâs test the interface. You should create a logical system and define a port for this interface. You should also configure the Customer Distribution Model to indicate that message type DEBMAS is being distributed to this logical system. The only difference in configuration between a regular outbound ALE interface and an enhanced one is the partner profile definition. While maintaining the outbound parameters of the partner profile, make sure the IDOC type is ZDEBMASZ. The fields for Basic IDOC type and extension type are automatically populated with DEBMAS02 and ZDEBMASX, respectively.
    To maintain the contact personâs business address of a customer:
    • Use transaction BD12 or from BALE go to Master Data ->Customer -> Send and send that Customer Master record by executing the transaction after filling in the relevant fields such as customer number, message type, and logical system.
    • Use transaction WE02 or WE05 to verify the IDOC created. You should see the new segment Z1SADRX populated with the correct data.
    With SAP releases below 4.5B, you cannot capture changes to business address through change pointers because a change document object is not available for capturing business address changes, and also earlier releases have not been configured to write change documents for a contact personâs business address. If you would like this functionality, you can either create change document objects, generate function modules to create change documents, and perform ALE configuration to tie it in, or make a cosmetic change to the contact person screen data while changing the contact personâs business address so that it gets captured as a change to the Customer Master. Subsequently, the ALE enhancement that you performed captures the contact personâs business address.
    Inbound interfaces. The process for enhancing inbound ALE interfaces is similar for outbound, with a few exceptions; specifically in the coding of customer functions (function exits) for the ALE/EDI function modules.
    The first step is to create an IDOC extension for the specific Basic IDOC type by adding new segments at the appropriate hierarchy level: that is, associated to the relevant existing segment. Populate the data fields on the new segments with application data by the translator or external system/program before importing them into the R/3 System. Then, find the ALE function module that is invoked by the inbound processing. By browsing through the code or reading the documentation on the function exit enhancements using the SMOD transaction, identify the function exit in which you should place your code. The technique used in the code to post the additional or modified IDOC data to the application can vary based on the application rules and requirements, the data available at that point in processing, and the application function modules available to update the application tables. It is important to search first for application modules that process the data and see if they can be called within the function exit. If the additional data in the extended segments in specific to a custom table or resides in nonkey fields of a single or small set of tables, you may be able to update it directly by SQL statements in the function exit. This approach should be carefully evaluated and is certainly not highly recommended.
    Another option is to use Call Transaction from within the function exit to process the additional data. For example, in the case of message type WMMBXY for inbound goods movements from a warehouse management system, the standard interface creates batches for materials, but does not update its characteristics. In such a case, you can use Call Transaction MSC1 to create the batch and assign characteristic values to it from within the function exit provided.
    regards,
    srinivas

  • Is it possible to generate outbound idocs for FI invoices? How?

    Is it possible to generate Outbound idocs for FI invoices? How? or else i have to write program to get data in file and pass it to translator. Please help.

    Dear Gabriel
    Before letting you know the possibility, can you please let me know why you want such a requirement, since standard has not provided such copy control, to my knowledge.
    As you would know order type KB is for consignment fillup where you are just transferring the goods from FG unrestricted to your own warehouse and that is why in standard you can see the copy control KB --> F8 which means you can raise proforma against delivery and not against fillup order.
    So think of whether really you wanted to raise proforma against fillup order which is NOT recommended by SAP.  However, still if you feel that the requirement is so to have, you can make the settings in VTFA.
    thanks
    G. Lakshmipathi

  • Generating an IDOC for each Order present in the file (File to IDOC)

    Hi ,
    I have a requirement where in i have to generate an idoc for each orders present in a file . Initially i had implemented 1:N mapping by changing the occurence of idoc to 1..unbounded & using it as external definition i was able to generate multiple idocs.
    But now the requirement is that even if one record contains invalid data , then XI shoul process remaining records(which has  valid data) & then raise an alert for the invalid record . I have implemented this approach by using BPM (splitting the message) & the sending out each record as a single message to (1:1 mapping between file & IDOC).
    But now they dont want me to use an BPM . They want me to come up with an approach using 1:N mapping only where in i can make sure that the mappig doesnt fail even if there is one invalid record ,they want me to process remaining records & raise an alert for the invalid one .
    As far as i know this cant be done without BPM . Please let me know any valuable suggestions or inputs if there is any
    Regards
    Vinay P.

    Hi Vinay,
    Had a similar requirement few days back..
    What we did was to do a multimapping of 1:N where N = N1 and N2, N1 is the Idocs which you are doing today, N2 is the set of invalid records...
    Now we have written this as a file and then created another interface which will pick the file and based on the name of file we decided the To address. We have used mail package and this interface (file to mail ) was made as a generic one so that it can be used across many interfaces...
    filename and To address relationship was maintained as a value mapping in ID... You can have a thought of this approach and try to send the invalid records as a mail rather than alert.
    Regards
    Suraj

  • Generate DELVRY  Idoc when Post goods issue is done in SAP

    I need to generate an Idoc when post goods issue is done in SAP (outbound to SAP) .I understood that I can do this using the message control .The message type DESADV with the process code DELV will support this.
    But I didn’t understand how to trigger the Idoc.In the current scenario our Company is doing the post goods issue in SAP for the in three different ways.
    1.Manually doing the post goods issue through the transaction VL02N
    2. an Inbound Interface (using Idoc)  which will do Post goods Issue in SAP
    3.Another Interface using BAPI, which will, do post goods issue in SAP
    I have to run my outbound interface when the post goods issue is done by all the above three ways.  Can any one help me out in this?

    You have to configure the output first .
    You can use output type LAVA ( avaialble in Std SAP ) .
    ( Tcode is NACE , Select Application V2 -> Outputtypes ).
    ( Create condition records suitable to generate an output)
    Next is ALE.
    Setup a distribution Model ( for DESADV ) and all other setup for ALE i.e. Distribute Dist. Model / Generate Partner Profile .
    In the Partner profile in WE20 select your partner system. Double CLick on Outbound Message DESADV ( this wd be there after you generate partner profile , if not add it here in outbound ) . On this screen goto tab MESSAGE CONTROL .
    Application - V2
    Message Type - LAVA
    Process Code - DELV.
    Now you are ready to go. The moment a PGI is done in any form an output LAVA would be proposed and processed using process code DELV.
    Cheers.

  • Generating an IDOC for Message type PROJECT in Change pointer?

    Hi,
    I am trying to generate an IDOC. for the Message Type PROJECT(Actually this msg. type PROJECT is for BAPI method) in CHANGE POINTERS but it is possible Error is coming, is there any other method to use the message type in change pointer technic.
    Thanks,
    Vinayak,
    Message was edited by: vinayaga sundaram

    Hi Vinay,
    Had a similar requirement few days back..
    What we did was to do a multimapping of 1:N where N = N1 and N2, N1 is the Idocs which you are doing today, N2 is the set of invalid records...
    Now we have written this as a file and then created another interface which will pick the file and based on the name of file we decided the To address. We have used mail package and this interface (file to mail ) was made as a generic one so that it can be used across many interfaces...
    filename and To address relationship was maintained as a value mapping in ID... You can have a thought of this approach and try to send the invalid records as a mail rather than alert.
    Regards
    Suraj

Maybe you are looking for

  • Set and Get parameters using SAP memory.

    I have an ALV grid list that I want to transfer to a detail report when the user doubleclicks.  I have set up a parameter transaction associated with the detail report.  My problem is the transfer does not take place.  After the "CALL TRANSACTION" th

  • What to do with orignal install

    Hey when? install the new beta driver can i still install all the other stuff from the install cd?

  • How to import updated version of APEX app without losing saved IR's?

    Oracle XE 11.2 Apex 4.1 Linux Red Hat 5.4 ============ Does anyone have suggestions on how to import updated version of APEX app without losing saved IR's? Also, I want to maintain app id from update to update (from revision to revision). In fact, I

  • Redirecting Readonly Queries to ADG or RAC

    Hi All, We are planning to redirect readonly queries to a seperate Database. Alhough we have ADG( Active Data Guard) due to cost constraints we cannot afford that. So we are planning to redirect ready only queries to RAC Node #4. As we have a four no

  • Incompatible with VTS 1

    Hi Guys, Im trying to burn a DVD from DVD Studio, I have an hour long video thats a MPEG 2 - Dolby 2.0 sound however when i click burn i get an error saying "Main Menu: Incompatible with VTS 1 "Features: Incompatible with VTS 1"