ALE - Idoc not getting processed automatically, instead going fine in Debug

Hello All,
I have a problem with Inbound idoc processing.
IDoc settings were made 'immediate processing' in WE20, and the process code is set correctly and the attached of Z-Fun. module to process the idocs to process code.
The error in inbound side is as 'Function module not allowed: APPL_IDOC_INPUTI' 51 status. Am sure i have to do some thing with ALE settings.
Can some body help me with correction step or sequence of steps  for this ALE process.
Thanks & Regards, Jilan

Hi Jilan,
Plz follow the steps:
ALE Configuration:
•     WE30 - IDOC type creation
•     WE31 - Create segment
•     WE81 - Message type creation
•     WE82 - Link IDOC type to Message type
•     SE37   - Create Inbound Function Module
•     BD51  - Maintain entry for Function Module
•     WE57 - Maintain
•     BD57  - Link Function Module, IDOC type and Message type
•     WE42 – Create Inbound Process Code
•     BD67  - Link Process code to Function Module
•     WE20 - Create Partner Profile
•     BD64  -  Display Distribution Model
•     WE02 -  IDoc List, Display all Inbound/Outbound IDocs
•     WE14 -  Outbound Processing of IDoc’s
•     BD20 -  Inbound Processing of IDoc’s
1st Step:   Create  a Segment  ( WE31)
         Segment is a structure for passing data in IDoc. It actually contains the IDoc data, just like the DDIC table/structure. Segment is created with all the required fields. Then Save it.  But, to actually use this Segment, you have to Release the Segment, otherwise u can’t use the Segment, by menu, EDIT--> Set Release.
         Now, if you want to do some change to this Segment, u cant, unless & until u Cancel the Release, by Edit--> Cancel Release.
2nd Step:   Create IDOC Type  ( WE30 )
        After creating the Segment, now we have to create the IDoc Type. IDOC Type is like an envelop of a letter, which contains the data inside it, & also some more information like address. IDoc type can be Basic Or Extended.
       Basic IDoc Type:  Using some SAP existing IDOC type (ex. MATMAS) or Custom IDOC type.
       Extension IDOC Type:  When we need some fields extra to an existing IDOC type, then we can extend that Basic Idoc Type by another Segment. This is called Extended idoc type
3rd Step:   Create Message Type (WE81)
     Message Type is like the Postman for sending the Letter.
  4th Step:   Attach Message Type to the IDOC Type (WE82)
5th Step:   Create a Function Module (SE37)
       Write the Processing logic in a Function Module.
6th Step:   Mention the IDOC Type, i.e, 0/1/2 (BD51)
7th Step:   Assign the Message Type, IDOC Type & Function Module (WE57)
8th Step:   Create a Custom Process Code (WE42)
9th Step:   Attach the Function Module to the Process Code (BD67)
10th Step: Execute Inbound IDOC’ s (WE19)
Example Function Module: IDOC_INPUT_ORDERS (Standard FM to create Sales Order).
reward is useful,
regards,
Tanmay

Similar Messages

  • IDOCS Error - Not getting processed automatically

    Hi All,
    We are loading hierarchy for a Product from R/3 system using the standard datasource.
    When we execute the info package, IDOCs are not getting processed automatically.
    We are facing the below error message.
    Error when updating Idocs in Business Information Warehouse
    Diagnosis
    Errors have been reported in Business Information Warehouse during IDoc update:
    No status record was passed to ALE by the applicat
    System Response
    Some IDocs have error status.
    Procedure
    Check the IDocs in Business Information Warehouse . You do this using the extraction monitor.
    Error handling:
    How you resolve the errors depends on the error message you get.
    But when we goto BD87 and process the IDOCs manually, these are getting posted and the hierarchy is loading.
    Can someone please guide me on what is the issue with the IDOCs and how to make them to post automatically.
    Thanks in Advance
    Regards,
    Sachin

    Hi,
    This will happen due to Non-updated IDOCu2019s in the Source system i.e., occurs whenever LUWu2019s are not transferred from the source system to the destination system. If you look at RSMO of the status tab, the error message would appear like u201CtRFC Error in Source Systemu201D or u201CtRFC Error in Data Warehouseu201D or simply u201CtRFC Erroru201D depending on the system from where data is being extracted. Sometimes IDOC are also stuck on R/3 side as there were no processors available to process them. The solution for this Execute LUWu2019s manually. Go to the menu Environment -> Transact. RFC -> In the Source System from RSMO which will asks to login into the source system. The u201CStatus Textu201D for stucked LUWu2019s may be Transaction Recorded or Transaction waiting. Once you encounter this type of status Execute LUWu2019s manually using u201CF6u201D or Editexecute LUWu2019s(F6).Just keep on refreshing until you get the status u201CTransaction is executingu201D in the Production system. We can even see the stuck IDOCu2019c in Transaction BD87 also.Now the data will be pulled into the BW.
    Hope it helps a lot.
    Thanks and Regards,
    Kamesham

  • Article IDoc not getting processed in PI for particle article type

    Hello All,
         I have an strange behaviour in my system.
    i.e 1, we are creating an article IDOC by using BD10 for particulat article type.
    2, IDOC created succesfully , then
    3, Process the IDOC in BD87 to push the IDOC to PI system .
    4, But i could not see the message in PI that too only for one article type.
    I'm wondering why this is not getting processed only for one article type and not for others.
    we also checked the basic filter for the message type in dist model and even i could not see any condition in PI also,
      Is there any thing else i should check? can you please help us on this.
    Regards , Sethu.

    Hi,
    Try to reimport the metadata of IDOC into PI in idx2. Also check the port definition in ECC and whether the segment is released or not. Please go through the below discussions it may help you.
    EDISDEF: Port XXX segment defn E2EDK35 in IDoc type ORDERS05 CIM type
    IDoc 0000000000181226 was saved but cannot or should not be sent
    Regards,
    Priyanka

  • RSEOUT00 - Idoc not getting processed

    Hi All,
    I have few idocs in prod.system with status 30 and they are not getting processed using RSEOUT00.
    I tried using back ground job / forgrond / BD87 also but they are not getting converted to 03 status.
    receiver port is xmlgen.
    Any thing like idoc lock ? or something other that is preventing it process ?
    Please let me know if anything.

    check out the multiple threads in SDN regarding this issue
    idoc remains random in status 30
    some of the common pointers looks to be locked IDOC (check in SM12), SM58 to see if anything is in the RFC queue (though I doubt that would be the case)..
    How are you generating your idoc? standard FM or MASTER_IDOC_DISTRIBUTE?

  • IDoc not getting processed in Real Time job

    Hello,
    The configurations are done as per the below link to receive iDoc from SAP in BODS.
    How to Receive IDOC in BODS from SAP (Outbound to SAP)
    An iDoc is sent and Real Time Services on management console shows that request is received successfully. But I cannot see output xml file created. When I checked the real time services and client interfaces, I could see all are in green.
    But the real time job status shows its in running state. It has been in running state for 2 days. Trace shows as below. There is no error message.
    14.2) 04-24-14 18:47:41 (18417:1140171104)      JOB: Optimizing job <JOB_BODS_RT_IDocMatmas01>.
    (14.2) 04-24-14 18:47:47 (18417:1140171104)      JOB: Job <JOB_BODS_RT_IDocMatmas01> is started.
    (14.2) 04-24-14 18:47:47 (18417:1140171104) DATAFLOW: Data flow <DF_RT_IDocMatmas01> is initialized.
    (14.2) 04-24-14 18:47:47 (18417:1140171104) DATAFLOW: Data flow <DF_RT_IDocMatmas01> using IN MEMORY Cache.
    (14.2) 04-24-14 18:47:48 (18417:1140171104) DATAFLOW: Data flow <ActaDefaultNotificationSender> is initialized.
    (14.2) 04-24-14 18:47:49 (18417:1140171104) DATAFLOW: Data flow <ActaDefaultNotificationSender> using IN MEMORY Cache.
    Can anyone please let me know if there is anything missing in the configuration? How do I resolve this issue?
    DS version: 4.2

    Hi,
    Try to reimport the metadata of IDOC into PI in idx2. Also check the port definition in ECC and whether the segment is released or not. Please go through the below discussions it may help you.
    EDISDEF: Port XXX segment defn E2EDK35 in IDoc type ORDERS05 CIM type
    IDoc 0000000000181226 was saved but cannot or should not be sent
    Regards,
    Priyanka

  • Automatic update of number ranges in FBN1 is not getting processed

    Automatic update of number ranges in FBN1 (current status) level is not getting processed.
    Tried posting a customer payment document in the system though number ranges are maintained for the year 2015 the number ranges are not overlapped last year all documents are closed but still the numbers are not been picked automatically in FBN1 kindly help in this issue

    Hello Rajesh
    Please verify if number ranges are maintained correctly, verify same using transaction OBA7. Also ensure that external assignment is unchecked in FBN1. Share screenshots if possible.
    Regards
    Aniket

  • IDOC status 64 not getting processed.

    Hi Gurus,
          I have a problem where IDOC's with status 64 are not getting processed via background job. the IDOCs weere getting processed with the background job till today evening, but suddenly no changes were made and the processing of IDOC's with status 64 has stopped and they keep filling up, when i checked the log of the Background job it says finished, and the log has the following information,
    17.09.2009 19:05:23 Job started                                                                   00           516          S
    17.09.2009 19:05:23 Step 001 started (program RBDAPP01, variant POS_IDOCS, user ID SAPAUDIT)      00           550          S
    17.09.2009 19:05:25 No data could be selected.                                                    B1           083          I
    17.09.2009 19:05:25 Job finished                                                                  00           517          S
    Kindly advise on this.
    Regards,
    Riaz.

    When i process the IDOC's using BD87, they gets processed wiwth out aany issues no dump is displayed, idoc gets a status 53 after processing via BD87.
    There is another problem how ever there are other jobs scheduled to run in the background other than this background job which are getting cancelled and have two different error logs for different jobs they are as follows.
    1) 18.09.2009 02:00:06 Job started                                             00           516          S
    18.09.2009 02:00:06 Logon not possible (error in license check)             00           179          E
    18.09.2009 02:00:06 Job cancelled after system exception ERROR_MESSAGE      00           564          A
    2) 18.09.2009 00:55:27 Job started                                                          00           516          S
    18.09.2009 00:55:27 Step 001 started (program RSCOLL00, variant , user ID JGERARDO)      00           550          S
    18.09.2009 00:55:29 Clean_Plan:Gestartet von RSDBPREV                                   DB6PM         000          S
    18.09.2009 00:55:29 Clean_Plan:beendet                                                  DB6PM         000          S
    18.09.2009 01:00:52 ABAP/4 processor: DBIF_RSQL_SQL_ERROR                                00           671          A
    18.09.2009 01:00:52 Job cancelled                                                        00           518          A
    This is a production server and the Licence is valid till 31.12.9999, and has no issues with the license.

  • Open Item not getting Cleared automatically

    Dear Sir,
    For a GL "Cenvat Clearing" , we have ticked the option "open Item Management" in FS00 .
    Our business scenario is as below :
    After MIGO , we do excise posting using J1IEX and here in the posted Fi document the GL "Cenvat Clearing" gets Credited . Subsequently during the MIRO this GL i.e "Cenvat Claering" gets Debited .
    My problem is that when I see in FBL3n for this GL the relevant transaction still appear under the open Items . Pl guide me , as what is the reason for this entry not getting cleared automatically .
    Regards
    B Mittal
    Regar

    PO number cannot be taken as sort feild  hence the value is in blank always.
    You need to identify some value for this feild and as a process you need to tell your users to incorporate the same at the time of doing MIRO.
    Or other wise you can take out the assignment of sort key to this GL account, and as a process you can instruct your user to use reference feild and input a value. If you fell that this is a tedious process,
    the Identify another feild that is common for both debit and credit line items and try to post with a common valuie using, one ABAP program which will update this feild. Mind that this feild need to be in the OB74.
    reward if useful
    sarma

  • Flash Player not getting updated automatically on OS X

    Hi, I have Mavericks installed and my Flash Player "Updates" settings are set to "Allow Adobe to install updates (recommended)". But still, the Flash Player is not getting updated automatically.
    When I click the "Check Now" button, a dialog box appears notifying an update is available and it takes me to the Flash Player download page. Isn't this supposed to download it and update in the background instead?

    Hi,
    Can you please go to your System Preferences -> Flash Player icon -> Updates tab or Advanced tab (If there is no Updates tab) and post a screenshot here using How to post a screenshot in the forum
    -Varun

  • Why SO Header not get closed automatically after all SO lines get closed?

    Hi all :
    Do you know the reason, why some SO Header still as Booked Status, even All the SO lines are already closed ?
    If I see the Workflow status the SO Header is in Status = Deferred and Activity = Wait
    And I can close it manually if I rewind the Activity in the Workflow, but I want to know why the SOs not get closed automatically.
    Thank you in advance
    Edited by: Daniel on Jan 18, 2011 8:41 AM

    Have you scheduled the concurrent request to use Workflow Background Processor to Process OEOH ? What is the frequency? Has it run after the lines got closed?
    Also, check the value of the WAIT attribute in the OEOH workflow? By default, it is set to close headers at the end of month.
    For more detail please see Note:141919.1 on Metalink.
    Also refer to metalink document # 397364.1 which details how to change the WAIT attributes.
    Hope this helps,
    Sandeep Gandhi

  • Savings Plan Contribution is not getting processed in Payroll

    Hello Friends,
    We are in the process of upgrading our system from 4.0 to 4.7, After correcting the schema as per the release/OSS notes, we are getting Savings Plan Contribution as blank (Retirement Savings).
    Cumulation values are coming thru right but current period contribution is missing.
    Earlier we had P0169 with P61, now we do not see processing class(PCLS)61 in the PCLS list. Our earlier wage types were referring to PCLS 61, I am not sure if this is causing an issue.
    Please let me know if you have any faced similar situation and have an answer.
    Thanks & Best Regards
    ravi

    Ravi,
    Appriciate if you could share the solution for your problem in savings plan contribution is not getting processed in the payroll. As I am facing the same kind od problem for my new configured benefit plan ROTH 401k plan.
    Please let me know the solution.
    Thanks & Regards,
    Sunita

  • Bulky Message are not getting processed

    Hi All,
    Recently we have upgraded our XI systems to EHP4.
    After upgrade Bulky files are not getting processed into XI system. Small files are getting processed successfully.
    Earlier we had faced same issue at that time we have maintained some parameters in transaction RZ10, RZ11, SXMB_ADM-->
    Integration Configuration--> HTTP time out parameter.
    Same parameters are still there. But now the message is not getting processed in XI when we check into sxi_monitor. Message mapping is also not seems to be a problem as we havnt put any bulky logic over there.
    The message is not getting processed ahead of  Receiver Grouping .  Please find attached log from Performanceheader
    <?xml version="1.0" encoding="UTF-8" standalone="yes" ?>
    - <!--  Receiver Grouping
      -->
    - <SAP:PerformanceHeader xmlns:SAP="http://sap.com/xi/XI/Message/30">
    - <SAP:RunTimeItem>
      <SAP:Name type="ADAPTER_IN">INTEGRATION_ENGINE_HTTP_ENTRY</SAP:Name>
      <SAP:Timestamp type="begin" host="evoxicqab">20091210134229.930472</SAP:Timestamp>
      </SAP:RunTimeItem>
    - <SAP:RunTimeItem>
      <SAP:Name type="ADAPTER_IN">INTEGRATION_ENGINE_HTTP_ENTRY</SAP:Name>
      <SAP:Timestamp type="end" host="evoxicqab">20091210134232.052577</SAP:Timestamp>
      </SAP:RunTimeItem>
    - <SAP:RunTimeItem>
      <SAP:Name type="CORE">INTEGRATION_ENGINE</SAP:Name>
      <SAP:Timestamp type="begin" host="evoxicqab">20091210134232.059576</SAP:Timestamp>
      </SAP:RunTimeItem>
    - <SAP:RunTimeItem>
      <SAP:Name type="CORE">INTEGRATION_ENGINE</SAP:Name>
      <SAP:Timestamp type="end" host="evoxicqab">20091210134232.07151</SAP:Timestamp>
      </SAP:RunTimeItem>
    - <SAP:RunTimeItem>
      <SAP:Name type="DBQUEUE">DB_ENTRY_QUEUING</SAP:Name>
      <SAP:Timestamp type="begin" host="evoxicqab">20091210134232.071518</SAP:Timestamp>
      </SAP:RunTimeItem>
    - <SAP:RunTimeItem>
      <SAP:Name type="DBQUEUE">DB_ENTRY_QUEUING</SAP:Name>
      <SAP:Timestamp type="end" host="evoxicqab">20091210134237.239947</SAP:Timestamp>
      </SAP:RunTimeItem>
    - <SAP:RunTimeItem>
      <SAP:Name type="PLSRV">PLSRV_RECEIVER_DETERMINATION</SAP:Name>
      <SAP:Timestamp type="begin" host="evoxicqab">20091210134237.241179</SAP:Timestamp>
      </SAP:RunTimeItem>
    - <SAP:RunTimeItem>
      <SAP:Name type="PLSRV">PLSRV_RECEIVER_DETERMINATION</SAP:Name>
      <SAP:Timestamp type="end" host="evoxicqab">20091210134237.250385</SAP:Timestamp>
      </SAP:RunTimeItem>
    - <SAP:RunTimeItem>
      <SAP:Name type="PLSRV">PLSRV_INTERFACE_DETERMINATION</SAP:Name>
      <SAP:Timestamp type="begin" host="evoxicqab">20091210134239.999045</SAP:Timestamp>
      </SAP:RunTimeItem>
    - <SAP:RunTimeItem>
      <SAP:Name type="PLSRV">PLSRV_INTERFACE_DETERMINATION</SAP:Name>
      <SAP:Timestamp type="end" host="evoxicqab">20091210134240.001292</SAP:Timestamp>
      </SAP:RunTimeItem>
    - <SAP:RunTimeItem>
      <SAP:Name type="PLSRV">PLSRV_RECEIVER_MESSAGE_SPLIT</SAP:Name>
      <SAP:Timestamp type="begin" host="evoxicqab">20091210134240.001413</SAP:Timestamp>
      </SAP:RunTimeItem>
    - <SAP:RunTimeItem>
      <SAP:Name type="PLSRV">PLSRV_RECEIVER_MESSAGE_SPLIT</SAP:Name>
      <SAP:Timestamp type="end" host="evoxicqab">20091210134240.026156</SAP:Timestamp>
      </SAP:RunTimeItem>
    - <SAP:RunTimeItem>
      <SAP:Name type="DBQUEUE">DB_SPLITTER_QUEUING</SAP:Name>
      <SAP:Timestamp type="begin" host="evoxicqab">20091210134240.026164</SAP:Timestamp>
      </SAP:RunTimeItem>
      </SAP:PerformanceHeader>

    Hi
    Have you checked this thread, same discussion here
    Performance of XI Interfaces
    Also check this blog
    /people/michal.krawczyk2/blog/2006/06/08/xi-timeouts-timeouts-timeouts
    Regards
    Ramesh

  • Purchase order output type NEU is not get assigned automatically

    Dear All,
    when i am creting purchase order the output type NEU is not get assigned automatically.
    I have maintained message condn record for NEU and for specific PO doc type NB.
    same problame i was facing for the inventory docuemnts but i was resolved by running one SAP std program so please guide me if any such program is available for Purchasing documents also or any other setting i am missing to make.
    Pl suggest
    Regards,
    Vimlesh

    Hi
    Please check the following in customization.Dont check in NACE you may miss something like fine tune control
    1) Check the Message determination procedure is having the NEU output type
    2) Check the std Message determination procedure is assigned to PO
    3) Check the output type for the Access sequence, default data for transmission medium and partner function VN, Language & Form assigned to it
    4) Check the Access sequence for the condition tables
    5) Check the condition record in MN04/MN05
    6) Check the purchasing group has printer assignement
    7) Check the fine tuen control for operation 1 & 2 for message type.
    Hope your problem now solved.
    Thanks /karthik
    Edited by: Karthik on Aug 13, 2009 5:29 PM

  • WCS:- WLC Alarm not getting cleared automatically on WCS

    ALarm TAB:-
    ==========
    Severity: Message:
    Critical ------------------------------Port 2 is down on Controller.
    This Alarm is not getting cleared automatically, Where as the alarms generated for AP's are getting cleared, I would like to know is this the default behaviour or we need to look into some thing here. To clear this i need to manually delete the alarm.
    Any help would be greatly appriciated.

    Alarms are automatically cleared after 7 days. Clear the alarms manually if you want to clear the alarms before 7 days. For more information refer http://www.cisco.com/en/US/products/ps6305/products_qanda_item09186a00807a60f0.shtml#qa11.

  • Snapshots not getting generated automatically

    Hi all,
    I want to generate an ADDM report. For that, I am running the addmrpt.sql script in $ORACLE_HOME/rdbms/admin
    While running, i am being asked for the begin_snap ID and end_snap ID. The problem is I have only one snap_id. The snapshots are not getting generated automatically for every 1 hr, as they shd be. I have to manually create a snapshot and run the ADDM script. I know that the database has to be open for atleast next 1 hr for getting the snapshot. I have not closed my database for the whole day.
    Now, I want to know what exactly I shd be looking at, or rather be doing to get my snapshots generate automatically.
    Please help me in this regard.
    Thanx in advance....
    Regards,
    ORA_SRI

    Hmm, you probably want to ask questions about ADDM in the [Database General Forum|http://forums.oracle.com/forums/forum.jspa?forumID=61]

Maybe you are looking for

  • Solaris 10 net install doesn't start

    Hi All, I'm trying to install Solaris 10 through a network installation guided by this document: http://www-uxsup.csx.cam.ac.uk/pub/doc/sun/solaris10/solaris10_network_installation.pdf I've did everything as described in chapter 6 and when try to boo

  • ORA-1652: unable to extend temp segment

    Hi, I am new here and i am not sure if my question is for that section but will be happy to get some answers. I found ORA-1652: unable to extend temp segment by 128 in tablespace PSAPTEMP in alert log and checked the space in PSAPTEMP. TABLESPACE_NAM

  • EEM on port error-disable due to UDLD

    I would like to ask from the EEM community for some assistance in regards to crafting an EEM script for an ME-3800X.  We have one or two appliances whereby the only uplink (TenGigabitEthernet0/1 and/or TenGigabitEthernet0/2) goes into error-disable a

  • Does soundtrack pro 3 take advantage of 64bit processing?

    Does the Soundtrack Pro 3 application run at 32bit or 64bit or can it switch between both and if so where is the setting? I am not referring to audio bit depth (16bit or 24bit) My question is in regards to the application running in 32bit mode or 64b

  • Overview of Identity Management & Single Sign On

    Guys, Im newbies here, would like to get an overview of Identity Management & Single Sign On and what's all about. Hopefully can get some sort of beginner explaination to start. Thanks.